1
First pullreq for arm of the 4.1 series, since I'm back from
1
Mostly this is RTH's memtag series, but there are also some cleanups
2
holiday now. This is mostly my M-profile FPU series and Philippe's
2
from Philippe.
3
devices.h cleanup. I have a pile of other patchsets to work through
4
in my to-review folder, but 42 patches is definitely quite
5
big enough to send now...
6
3
7
thanks
4
thanks
8
-- PMM
5
-- PMM
9
6
10
The following changes since commit 413a99a92c13ec408dcf2adaa87918dc81e890c8:
7
The following changes since commit 10f7ffabf9c507fc02382b89912003b1c43c3231:
11
8
12
Add Nios II semihosting support. (2019-04-29 16:09:51 +0100)
9
Merge remote-tracking branch 'remotes/mcayland/tags/qemu-macppc-20200626' into staging (2020-06-26 12:14:18 +0100)
13
10
14
are available in the Git repository at:
11
are available in the Git repository at:
15
12
16
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20190429
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200626
17
14
18
for you to fetch changes up to 437cc27ddfded3bbab6afd5ac1761e0e195edba7:
15
for you to fetch changes up to c7459633baa71d1781fde4a245d6ec9ce2f008cf:
19
16
20
hw/devices: Move SMSC 91C111 declaration into a new header (2019-04-29 17:57:21 +0100)
17
target/arm: Enable MTE (2020-06-26 14:32:24 +0100)
21
18
22
----------------------------------------------------------------
19
----------------------------------------------------------------
23
target-arm queue:
20
target-arm queue:
24
* remove "bag of random stuff" hw/devices.h header
21
* hw/arm/aspeed: improve QOM usage
25
* implement FPU for Cortex-M and enable it for Cortex-M4 and -M33
22
* hw/misc/pca9552: trace GPIO change events
26
* hw/dma: Compile the bcm2835_dma device as common object
23
* target/arm: Implement ARMv8.5-MemTag for system emulation
27
* configure: Remove --source-path option
28
* hw/ssi/xilinx_spips: Avoid variable length array
29
* hw/arm/smmuv3: Remove SMMUNotifierNode
30
24
31
----------------------------------------------------------------
25
----------------------------------------------------------------
32
Eric Auger (1):
26
Philippe Mathieu-Daudé (12):
33
hw/arm/smmuv3: Remove SMMUNotifierNode
27
hw/arm/aspeed: Remove extraneous MemoryRegion object owner
28
hw/arm/aspeed: Rename AspeedBoardState as AspeedMachineState
29
hw/arm/aspeed: QOM'ify AspeedMachineState
30
hw/i2c/core: Add i2c_try_create_slave() and i2c_realize_and_unref()
31
hw/misc/pca9552: Rename 'nr_leds' as 'pin_count'
32
hw/misc/pca9552: Rename generic code as pca955x
33
hw/misc/pca9552: Add generic PCA955xClass, parent of TYPE_PCA9552
34
hw/misc/pca9552: Add a 'description' property for debugging purpose
35
hw/misc/pca9552: Trace GPIO High/Low events
36
hw/arm/aspeed: Describe each PCA9552 device
37
hw/misc/pca9552: Trace GPIO change events
38
hw/misc/pca9552: Model qdev output GPIOs
34
39
35
Peter Maydell (28):
40
Richard Henderson (45):
36
hw/ssi/xilinx_spips: Avoid variable length array
41
target/arm: Add isar tests for mte
37
configure: Remove --source-path option
42
target/arm: Improve masking of SCR RES0 bits
38
target/arm: Make sure M-profile FPSCR RES0 bits are not settable
43
target/arm: Add support for MTE to SCTLR_ELx
39
hw/intc/armv7m_nvic: Allow reading of M-profile MVFR* registers
44
target/arm: Add support for MTE to HCR_EL2 and SCR_EL3
40
target/arm: Implement dummy versions of M-profile FP-related registers
45
target/arm: Rename DISAS_UPDATE to DISAS_UPDATE_EXIT
41
target/arm: Disable most VFP sysregs for M-profile
46
target/arm: Add DISAS_UPDATE_NOCHAIN
42
target/arm: Honour M-profile FP enable bits
47
target/arm: Add MTE system registers
43
target/arm: Decode FP instructions for M profile
48
target/arm: Add MTE bits to tb_flags
44
target/arm: Clear CONTROL_S.SFPA in SG insn if FPU present
49
target/arm: Implement the IRG instruction
45
target/arm: Handle SFPA and FPCA bits in reads and writes of CONTROL
50
target/arm: Revise decoding for disas_add_sub_imm
46
target/arm/helper: don't return early for STKOF faults during stacking
51
target/arm: Implement the ADDG, SUBG instructions
47
target/arm: Handle floating point registers in exception entry
52
target/arm: Implement the GMI instruction
48
target/arm: Implement v7m_update_fpccr()
53
target/arm: Implement the SUBP instruction
49
target/arm: Clear CONTROL.SFPA in BXNS and BLXNS
54
target/arm: Define arm_cpu_do_unaligned_access for user-only
50
target/arm: Clean excReturn bits when tail chaining
55
target/arm: Implement LDG, STG, ST2G instructions
51
target/arm: Allow for floating point in callee stack integrity check
56
target/arm: Implement the STGP instruction
52
target/arm: Handle floating point registers in exception return
57
target/arm: Restrict the values of DCZID.BS under TCG
53
target/arm: Move NS TBFLAG from bit 19 to bit 6
58
target/arm: Simplify DC_ZVA
54
target/arm: Overlap VECSTRIDE and XSCALE_CPAR TB flags
59
target/arm: Implement the LDGM, STGM, STZGM instructions
55
target/arm: Set FPCCR.S when executing M-profile floating point insns
60
target/arm: Implement the access tag cache flushes
56
target/arm: Activate M-profile floating point context when FPCCR.ASPEN is set
61
target/arm: Move regime_el to internals.h
57
target/arm: New helper function arm_v7m_mmu_idx_all()
62
target/arm: Move regime_tcr to internals.h
58
target/arm: New function armv7m_nvic_set_pending_lazyfp()
63
target/arm: Add gen_mte_check1
59
target/arm: Add lazy-FP-stacking support to v7m_stack_write()
64
target/arm: Add gen_mte_checkN
60
target/arm: Implement M-profile lazy FP state preservation
65
target/arm: Implement helper_mte_check1
61
target/arm: Implement VLSTM for v7M CPUs with an FPU
66
target/arm: Implement helper_mte_checkN
62
target/arm: Implement VLLDM for v7M CPUs with an FPU
67
target/arm: Add helper_mte_check_zva
63
target/arm: Enable FPU for Cortex-M4 and Cortex-M33
68
target/arm: Use mte_checkN for sve unpredicated loads
69
target/arm: Use mte_checkN for sve unpredicated stores
70
target/arm: Use mte_check1 for sve LD1R
71
target/arm: Tidy trans_LD1R_zpri
72
target/arm: Add arm_tlb_bti_gp
73
target/arm: Add mte helpers for sve scalar + int loads
74
target/arm: Add mte helpers for sve scalar + int stores
75
target/arm: Add mte helpers for sve scalar + int ff/nf loads
76
target/arm: Handle TBI for sve scalar + int memory ops
77
target/arm: Add mte helpers for sve scatter/gather memory ops
78
target/arm: Complete TBI clearing for user-only for SVE
79
target/arm: Implement data cache set allocation tags
80
target/arm: Set PSTATE.TCO on exception entry
81
target/arm: Always pass cacheattr to get_phys_addr
82
target/arm: Cache the Tagged bit for a page in MemTxAttrs
83
target/arm: Create tagged ram when MTE is enabled
84
target/arm: Add allocation tag storage for system mode
85
target/arm: Enable MTE
64
86
65
Philippe Mathieu-Daudé (13):
87
include/hw/arm/aspeed.h | 12 +-
66
hw/dma: Compile the bcm2835_dma device as common object
88
include/hw/i2c/i2c.h | 2 +
67
hw/arm/aspeed: Use TYPE_TMP105/TYPE_PCA9552 instead of hardcoded string
89
include/hw/misc/pca9552.h | 16 +-
68
hw/arm/nseries: Use TYPE_TMP105 instead of hardcoded string
90
target/arm/cpu.h | 50 +-
69
hw/display/tc6393xb: Remove unused functions
91
target/arm/helper-a64.h | 16 +
70
hw/devices: Move TC6393XB declarations into a new header
92
target/arm/helper-sve.h | 488 ++++++++++++++
71
hw/devices: Move Blizzard declarations into a new header
93
target/arm/helper.h | 2 +
72
hw/devices: Move CBus declarations into a new header
94
target/arm/internals.h | 153 ++++-
73
hw/devices: Move Gamepad declarations into a new header
95
target/arm/translate-a64.h | 5 +
74
hw/devices: Move TI touchscreen declarations into a new header
96
target/arm/translate.h | 23 +-
75
hw/devices: Move LAN9118 declarations into a new header
97
hw/arm/aspeed.c | 46 +-
76
hw/net/ne2000-isa: Add guards to the header
98
hw/arm/virt.c | 55 +-
77
hw/net/lan9118: Export TYPE_LAN9118 and use it instead of hardcoded string
99
hw/i2c/core.c | 18 +-
78
hw/devices: Move SMSC 91C111 declaration into a new header
100
hw/misc/pca9552.c | 216 +++++--
101
target/arm/cpu.c | 81 ++-
102
target/arm/cpu64.c | 5 +
103
target/arm/helper-a64.c | 94 +--
104
target/arm/helper.c | 423 ++++++++++---
105
target/arm/m_helper.c | 11 +-
106
target/arm/mte_helper.c | 906 ++++++++++++++++++++++++++
107
target/arm/op_helper.c | 16 +
108
target/arm/sve_helper.c | 616 ++++++++++++++----
109
target/arm/tlb_helper.c | 13 +-
110
target/arm/translate-a64.c | 657 ++++++++++++++++---
111
target/arm/translate-sve.c | 1366 ++++++++++++++++++++++++++--------------
112
target/arm/translate-vfp.inc.c | 4 +-
113
target/arm/translate.c | 16 +-
114
hw/misc/trace-events | 4 +
115
target/arm/Makefile.objs | 1 +
116
29 files changed, 4391 insertions(+), 924 deletions(-)
117
create mode 100644 target/arm/mte_helper.c
79
118
80
configure | 10 +-
81
hw/dma/Makefile.objs | 2 +-
82
include/hw/arm/omap.h | 6 +-
83
include/hw/arm/smmu-common.h | 8 +-
84
include/hw/devices.h | 62 ---
85
include/hw/display/blizzard.h | 22 ++
86
include/hw/display/tc6393xb.h | 24 ++
87
include/hw/input/gamepad.h | 19 +
88
include/hw/input/tsc2xxx.h | 36 ++
89
include/hw/misc/cbus.h | 32 ++
90
include/hw/net/lan9118.h | 21 +
91
include/hw/net/ne2000-isa.h | 6 +
92
include/hw/net/smc91c111.h | 19 +
93
include/qemu/typedefs.h | 1 -
94
target/arm/cpu.h | 95 ++++-
95
target/arm/helper.h | 5 +
96
target/arm/translate.h | 3 +
97
hw/arm/aspeed.c | 13 +-
98
hw/arm/exynos4_boards.c | 3 +-
99
hw/arm/gumstix.c | 2 +-
100
hw/arm/integratorcp.c | 2 +-
101
hw/arm/kzm.c | 2 +-
102
hw/arm/mainstone.c | 2 +-
103
hw/arm/mps2-tz.c | 3 +-
104
hw/arm/mps2.c | 2 +-
105
hw/arm/nseries.c | 7 +-
106
hw/arm/palm.c | 2 +-
107
hw/arm/realview.c | 3 +-
108
hw/arm/smmu-common.c | 6 +-
109
hw/arm/smmuv3.c | 28 +-
110
hw/arm/stellaris.c | 2 +-
111
hw/arm/tosa.c | 2 +-
112
hw/arm/versatilepb.c | 2 +-
113
hw/arm/vexpress.c | 2 +-
114
hw/display/blizzard.c | 2 +-
115
hw/display/tc6393xb.c | 18 +-
116
hw/input/stellaris_input.c | 2 +-
117
hw/input/tsc2005.c | 2 +-
118
hw/input/tsc210x.c | 4 +-
119
hw/intc/armv7m_nvic.c | 261 +++++++++++++
120
hw/misc/cbus.c | 2 +-
121
hw/net/lan9118.c | 3 +-
122
hw/net/smc91c111.c | 2 +-
123
hw/ssi/xilinx_spips.c | 6 +-
124
target/arm/cpu.c | 20 +
125
target/arm/helper.c | 873 +++++++++++++++++++++++++++++++++++++++---
126
target/arm/machine.c | 16 +
127
target/arm/translate.c | 150 +++++++-
128
target/arm/vfp_helper.c | 8 +
129
MAINTAINERS | 7 +
130
50 files changed, 1595 insertions(+), 235 deletions(-)
131
delete mode 100644 include/hw/devices.h
132
create mode 100644 include/hw/display/blizzard.h
133
create mode 100644 include/hw/display/tc6393xb.h
134
create mode 100644 include/hw/input/gamepad.h
135
create mode 100644 include/hw/input/tsc2xxx.h
136
create mode 100644 include/hw/misc/cbus.h
137
create mode 100644 include/hw/net/lan9118.h
138
create mode 100644 include/hw/net/smc91c111.h
139
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
I'm confused by this code, 'bmc' is created as:
4
5
bmc = g_new0(AspeedBoardState, 1);
6
7
Then we use it as QOM owner for different MemoryRegion objects.
8
But looking at memory_region_init_ram (similarly for ROM):
9
10
void memory_region_init_ram(MemoryRegion *mr,
11
struct Object *owner,
12
const char *name,
13
uint64_t size,
14
Error **errp)
15
{
16
DeviceState *owner_dev;
17
Error *err = NULL;
18
19
memory_region_init_ram_nomigrate(mr, owner, name, size, &err);
20
if (err) {
21
error_propagate(errp, err);
22
return;
23
}
24
/* This will assert if owner is neither NULL nor a DeviceState.
25
* We only want the owner here for the purposes of defining a
26
* unique name for migration. TODO: Ideally we should implement
27
* a naming scheme for Objects which are not DeviceStates, in
28
* which case we can relax this restriction.
29
*/
30
owner_dev = DEVICE(owner);
31
vmstate_register_ram(mr, owner_dev);
32
}
33
34
The expected assertion is not triggered ('bmc' is not NULL neither
35
a DeviceState).
36
37
'bmc' structure is defined as:
38
39
struct AspeedBoardState {
40
AspeedSoCState soc;
41
MemoryRegion ram_container;
42
MemoryRegion max_ram;
43
};
44
45
What happens is when using 'OBJECT(bmc)', the QOM macros cast the
46
memory pointed by bmc, which first member is 'soc', which is
47
initialized ...:
48
49
object_initialize_child(OBJECT(machine), "soc",
50
&bmc->soc, amc->soc_name);
51
52
The 'soc' object is indeed a DeviceState, so the assertion passes.
53
54
Since this is fragile and only happens to work by luck, remove the
55
dangerous OBJECT(bmc) owner argument.
56
57
Note, this probably breaks migration for this machine.
58
59
Reviewed-by: Cédric Le Goater <clg@kaod.org>
60
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
61
Message-id: 20200623072132.2868-2-f4bug@amsat.org
62
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
63
---
64
hw/arm/aspeed.c | 6 +++---
65
1 file changed, 3 insertions(+), 3 deletions(-)
66
67
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/aspeed.c
70
+++ b/hw/arm/aspeed.c
71
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
72
* needed by the flash modules of the Aspeed machines.
73
*/
74
if (ASPEED_MACHINE(machine)->mmio_exec) {
75
- memory_region_init_alias(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
76
+ memory_region_init_alias(boot_rom, NULL, "aspeed.boot_rom",
77
&fl->mmio, 0, fl->size);
78
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
79
boot_rom);
80
} else {
81
- memory_region_init_rom(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
82
+ memory_region_init_rom(boot_rom, NULL, "aspeed.boot_rom",
83
fl->size, &error_abort);
84
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
85
boot_rom);
86
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
87
if (machine->kernel_filename && sc->num_cpus > 1) {
88
/* With no u-boot we must set up a boot stub for the secondary CPU */
89
MemoryRegion *smpboot = g_new(MemoryRegion, 1);
90
- memory_region_init_ram(smpboot, OBJECT(bmc), "aspeed.smpboot",
91
+ memory_region_init_ram(smpboot, NULL, "aspeed.smpboot",
92
0x80, &error_abort);
93
memory_region_add_subregion(get_system_memory(),
94
AST_SMP_MAILBOX_BASE, smpboot);
95
--
96
2.20.1
97
98
diff view generated by jsdifflib
1
The only "system register" that M-profile floating point exposes
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
via the VMRS/VMRS instructions is FPSCR, and it does not have
3
the odd special case for rd==15. Add a check to ensure we only
4
expose FPSCR.
5
2
3
To have a more consistent naming, rename AspeedBoardState
4
as AspeedMachineState.
5
6
Suggested-by: Cédric Le Goater <clg@kaod.org>
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Message-id: 20200623072132.2868-3-f4bug@amsat.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20190416125744.27770-5-peter.maydell@linaro.org
9
---
11
---
10
target/arm/translate.c | 19 +++++++++++++++++--
12
include/hw/arm/aspeed.h | 4 ++--
11
1 file changed, 17 insertions(+), 2 deletions(-)
13
hw/arm/aspeed.c | 20 ++++++++++----------
14
2 files changed, 12 insertions(+), 12 deletions(-)
12
15
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
14
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/translate.c
18
--- a/include/hw/arm/aspeed.h
16
+++ b/target/arm/translate.c
19
+++ b/include/hw/arm/aspeed.h
17
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
20
@@ -XXX,XX +XXX,XX @@
18
}
21
19
}
22
#include "hw/boards.h"
20
} else { /* !dp */
23
21
+ bool is_sysreg;
24
-typedef struct AspeedBoardState AspeedBoardState;
22
+
25
+typedef struct AspeedMachineState AspeedMachineState;
23
if ((insn & 0x6f) != 0x00)
26
24
return 1;
27
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
25
rn = VFP_SREG_N(insn);
28
#define ASPEED_MACHINE(obj) \
26
+
29
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineClass {
27
+ is_sysreg = extract32(insn, 21, 1);
30
const char *spi_model;
28
+
31
uint32_t num_cs;
29
+ if (arm_dc_feature(s, ARM_FEATURE_M)) {
32
uint32_t macs_mask;
30
+ /*
33
- void (*i2c_init)(AspeedBoardState *bmc);
31
+ * The only M-profile VFP vmrs/vmsr sysreg is FPSCR.
34
+ void (*i2c_init)(AspeedMachineState *bmc);
32
+ * Writes to R15 are UNPREDICTABLE; we choose to undef.
35
} AspeedMachineClass;
33
+ */
36
34
+ if (is_sysreg && (rd == 15 || (rn >> 1) != ARM_VFP_FPSCR)) {
37
35
+ return 1;
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
36
+ }
39
index XXXXXXX..XXXXXXX 100644
37
+ }
40
--- a/hw/arm/aspeed.c
38
+
41
+++ b/hw/arm/aspeed.c
39
if (insn & ARM_CP_RW_BIT) {
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
40
/* vfp->arm */
43
.board_id = -1, /* device-tree-only board */
41
- if (insn & (1 << 21)) {
44
};
42
+ if (is_sysreg) {
45
43
/* system register */
46
-struct AspeedBoardState {
44
rn >>= 1;
47
+struct AspeedMachineState {
45
48
AspeedSoCState soc;
46
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
49
MemoryRegion ram_container;
47
}
50
MemoryRegion max_ram;
48
} else {
51
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
49
/* arm->vfp */
52
50
- if (insn & (1 << 21)) {
53
static void aspeed_machine_init(MachineState *machine)
51
+ if (is_sysreg) {
54
{
52
rn >>= 1;
55
- AspeedBoardState *bmc;
53
/* system register */
56
+ AspeedMachineState *bmc;
54
switch (rn) {
57
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
58
AspeedSoCClass *sc;
59
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
60
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
61
int i;
62
NICInfo *nd = &nd_table[0];
63
64
- bmc = g_new0(AspeedBoardState, 1);
65
+ bmc = g_new0(AspeedMachineState, 1);
66
67
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
68
4 * GiB);
69
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
70
arm_load_kernel(ARM_CPU(first_cpu), machine, &aspeed_board_binfo);
71
}
72
73
-static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
74
+static void palmetto_bmc_i2c_init(AspeedMachineState *bmc)
75
{
76
AspeedSoCState *soc = &bmc->soc;
77
DeviceState *dev;
78
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
79
object_property_set_int(OBJECT(dev), 110000, "temperature3", &error_abort);
80
}
81
82
-static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
83
+static void ast2500_evb_i2c_init(AspeedMachineState *bmc)
84
{
85
AspeedSoCState *soc = &bmc->soc;
86
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
87
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
88
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
89
}
90
91
-static void ast2600_evb_i2c_init(AspeedBoardState *bmc)
92
+static void ast2600_evb_i2c_init(AspeedMachineState *bmc)
93
{
94
/* Start with some devices on our I2C busses */
95
ast2500_evb_i2c_init(bmc);
96
}
97
98
-static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
99
+static void romulus_bmc_i2c_init(AspeedMachineState *bmc)
100
{
101
AspeedSoCState *soc = &bmc->soc;
102
103
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
104
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
105
}
106
107
-static void swift_bmc_i2c_init(AspeedBoardState *bmc)
108
+static void swift_bmc_i2c_init(AspeedMachineState *bmc)
109
{
110
AspeedSoCState *soc = &bmc->soc;
111
112
@@ -XXX,XX +XXX,XX @@ static void swift_bmc_i2c_init(AspeedBoardState *bmc)
113
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 12), "tmp105", 0x4a);
114
}
115
116
-static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
117
+static void sonorapass_bmc_i2c_init(AspeedMachineState *bmc)
118
{
119
AspeedSoCState *soc = &bmc->soc;
120
121
@@ -XXX,XX +XXX,XX @@ static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
122
123
}
124
125
-static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
126
+static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
127
{
128
AspeedSoCState *soc = &bmc->soc;
129
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
55
--
130
--
56
2.20.1
131
2.20.1
57
132
58
133
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
AspeedMachineState seems crippled. We use incorrectly 2
4
different structures to do the same thing. Merge them
5
altogether:
6
- Move AspeedMachine fields to AspeedMachineState
7
- AspeedMachineState is now QOM
8
- Remove unused AspeedMachine structure
9
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Message-id: 20200623072132.2868-4-f4bug@amsat.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
include/hw/arm/aspeed.h | 8 +-------
16
hw/arm/aspeed.c | 11 +++++++----
17
2 files changed, 8 insertions(+), 11 deletions(-)
18
19
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/include/hw/arm/aspeed.h
22
+++ b/include/hw/arm/aspeed.h
23
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineState AspeedMachineState;
24
25
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
26
#define ASPEED_MACHINE(obj) \
27
- OBJECT_CHECK(AspeedMachine, (obj), TYPE_ASPEED_MACHINE)
28
-
29
-typedef struct AspeedMachine {
30
- MachineState parent_obj;
31
-
32
- bool mmio_exec;
33
-} AspeedMachine;
34
+ OBJECT_CHECK(AspeedMachineState, (obj), TYPE_ASPEED_MACHINE)
35
36
#define ASPEED_MAC0_ON (1 << 0)
37
#define ASPEED_MAC1_ON (1 << 1)
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
};
44
45
struct AspeedMachineState {
46
+ /* Private */
47
+ MachineState parent_obj;
48
+ /* Public */
49
+
50
AspeedSoCState soc;
51
MemoryRegion ram_container;
52
MemoryRegion max_ram;
53
+ bool mmio_exec;
54
};
55
56
/* Palmetto hardware value: 0x120CE416 */
57
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
58
59
static void aspeed_machine_init(MachineState *machine)
60
{
61
- AspeedMachineState *bmc;
62
+ AspeedMachineState *bmc = ASPEED_MACHINE(machine);
63
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
64
AspeedSoCClass *sc;
65
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
66
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
67
int i;
68
NICInfo *nd = &nd_table[0];
69
70
- bmc = g_new0(AspeedMachineState, 1);
71
-
72
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
73
4 * GiB);
74
memory_region_add_subregion(&bmc->ram_container, 0, machine->ram);
75
@@ -XXX,XX +XXX,XX @@ static const TypeInfo aspeed_machine_types[] = {
76
}, {
77
.name = TYPE_ASPEED_MACHINE,
78
.parent = TYPE_MACHINE,
79
- .instance_size = sizeof(AspeedMachine),
80
+ .instance_size = sizeof(AspeedMachineState),
81
.instance_init = aspeed_machine_instance_init,
82
.class_size = sizeof(AspeedMachineClass),
83
.class_init = aspeed_machine_class_init,
84
--
85
2.20.1
86
87
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Suggested-by: Markus Armbruster <armbru@redhat.com>
3
Extract i2c_try_create_slave() and i2c_realize_and_unref()
4
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
from i2c_create_slave().
5
Message-id: 20190412165416.7977-3-philmd@redhat.com
5
We can now set properties on a I2CSlave before it is realized.
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
7
This is in line with the recent qdev/QOM changes merged
8
in commit 6675a653d2e.
9
10
Reviewed-by: Corey Minyard <cminyard@mvista.com>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Reviewed-by: Markus Armbruster <armbru@redhat.com>
14
Tested-by: Cédric Le Goater <clg@kaod.org>
15
Message-id: 20200623072723.6324-2-f4bug@amsat.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
17
---
9
hw/arm/nseries.c | 3 ++-
18
include/hw/i2c/i2c.h | 2 ++
10
1 file changed, 2 insertions(+), 1 deletion(-)
19
hw/i2c/core.c | 18 ++++++++++++++++--
20
2 files changed, 18 insertions(+), 2 deletions(-)
11
21
12
diff --git a/hw/arm/nseries.c b/hw/arm/nseries.c
22
diff --git a/include/hw/i2c/i2c.h b/include/hw/i2c/i2c.h
13
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/arm/nseries.c
24
--- a/include/hw/i2c/i2c.h
15
+++ b/hw/arm/nseries.c
25
+++ b/include/hw/i2c/i2c.h
16
@@ -XXX,XX +XXX,XX @@
26
@@ -XXX,XX +XXX,XX @@ int i2c_send(I2CBus *bus, uint8_t data);
17
#include "hw/boards.h"
27
uint8_t i2c_recv(I2CBus *bus);
18
#include "hw/i2c/i2c.h"
28
19
#include "hw/devices.h"
29
DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr);
20
+#include "hw/misc/tmp105.h"
30
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr);
21
#include "hw/block/flash.h"
31
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp);
22
#include "hw/hw.h"
32
23
#include "hw/bt.h"
33
/* lm832x.c */
24
@@ -XXX,XX +XXX,XX @@ static void n8x0_i2c_setup(struct n800_s *s)
34
void lm832x_key_event(DeviceState *dev, int key, int state);
25
qemu_register_powerdown_notifier(&n8x0_system_powerdown_notifier);
35
diff --git a/hw/i2c/core.c b/hw/i2c/core.c
26
36
index XXXXXXX..XXXXXXX 100644
27
/* Attach a TMP105 PM chip (A0 wired to ground) */
37
--- a/hw/i2c/core.c
28
- dev = i2c_create_slave(i2c, "tmp105", N8X0_TMP105_ADDR);
38
+++ b/hw/i2c/core.c
29
+ dev = i2c_create_slave(i2c, TYPE_TMP105, N8X0_TMP105_ADDR);
39
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_i2c_slave = {
30
qdev_connect_gpio_out(dev, 0, tmp_irq);
40
}
41
};
42
43
-DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
44
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr)
45
{
46
DeviceState *dev;
47
48
dev = qdev_new(name);
49
qdev_prop_set_uint8(dev, "address", addr);
50
- qdev_realize_and_unref(dev, &bus->qbus, &error_fatal);
51
+ return dev;
52
+}
53
+
54
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp)
55
+{
56
+ return qdev_realize_and_unref(dev, &bus->qbus, errp);
57
+}
58
+
59
+DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
60
+{
61
+ DeviceState *dev;
62
+
63
+ dev = i2c_try_create_slave(name, addr);
64
+ i2c_realize_and_unref(dev, bus, &error_fatal);
65
+
66
return dev;
31
}
67
}
32
68
33
--
69
--
34
2.20.1
70
2.20.1
35
71
36
72
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
The PCA9552 device does not expose LEDs, but simple pins
4
to connnect LEDs to. To be clearer with the device model,
5
rename 'nr_leds' as 'pin_count'.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-3-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 2 +-
14
hw/misc/pca9552.c | 10 +++++-----
15
2 files changed, 6 insertions(+), 6 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
22
23
uint8_t regs[PCA9552_NR_REGS];
24
uint8_t max_reg;
25
- uint8_t nr_leds;
26
+ uint8_t pin_count;
27
} PCA9552State;
28
29
#endif
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
35
{
36
int i;
37
38
- for (i = 0; i < s->nr_leds; i++) {
39
+ for (i = 0; i < s->pin_count; i++) {
40
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
41
uint8_t input_shift = (i % 8);
42
uint8_t config = pca9552_pin_get_config(s, i);
43
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
44
error_setg(errp, "%s: error reading %s", __func__, name);
45
return;
46
}
47
- if (led < 0 || led > s->nr_leds) {
48
+ if (led < 0 || led > s->pin_count) {
49
error_setg(errp, "%s invalid led %s", __func__, name);
50
return;
51
}
52
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
53
error_setg(errp, "%s: error reading %s", __func__, name);
54
return;
55
}
56
- if (led < 0 || led > s->nr_leds) {
57
+ if (led < 0 || led > s->pin_count) {
58
error_setg(errp, "%s invalid led %s", __func__, name);
59
return;
60
}
61
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
62
* PCA955X device
63
*/
64
s->max_reg = PCA9552_LS3;
65
- s->nr_leds = 16;
66
+ s->pin_count = 16;
67
68
- for (led = 0; led < s->nr_leds; led++) {
69
+ for (led = 0; led < s->pin_count; led++) {
70
char *name;
71
72
name = g_strdup_printf("led%d", led);
73
--
74
2.20.1
75
76
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Reviewed-by: Markus Armbruster <armbru@redhat.com>
3
Various code from the PCA9552 device model is generic to the
4
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
PCA955X family. We'll split the generic code in a base class
5
Message-id: 20190412165416.7977-12-philmd@redhat.com
5
in the next commit. To ease review, first do a dumb renaming.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-4-f4bug@amsat.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
12
---
8
include/hw/net/lan9118.h | 2 ++
13
include/hw/misc/pca9552.h | 10 ++---
9
hw/arm/exynos4_boards.c | 3 ++-
14
hw/misc/pca9552.c | 80 +++++++++++++++++++--------------------
10
hw/arm/mps2-tz.c | 3 ++-
15
2 files changed, 45 insertions(+), 45 deletions(-)
11
hw/net/lan9118.c | 1 -
16
12
4 files changed, 6 insertions(+), 3 deletions(-)
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
13
14
diff --git a/include/hw/net/lan9118.h b/include/hw/net/lan9118.h
15
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
16
--- a/include/hw/net/lan9118.h
19
--- a/include/hw/misc/pca9552.h
17
+++ b/include/hw/net/lan9118.h
20
+++ b/include/hw/misc/pca9552.h
18
@@ -XXX,XX +XXX,XX @@
21
@@ -XXX,XX +XXX,XX @@
19
#include "hw/irq.h"
22
#include "hw/i2c/i2c.h"
20
#include "net/net.h"
23
21
24
#define TYPE_PCA9552 "pca9552"
22
+#define TYPE_LAN9118 "lan9118"
25
-#define PCA9552(obj) OBJECT_CHECK(PCA9552State, (obj), TYPE_PCA9552)
23
+
26
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
24
void lan9118_init(NICInfo *, uint32_t, qemu_irq);
27
28
-#define PCA9552_NR_REGS 10
29
+#define PCA955X_NR_REGS 10
30
31
-typedef struct PCA9552State {
32
+typedef struct PCA955xState {
33
/*< private >*/
34
I2CSlave i2c;
35
/*< public >*/
36
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
37
uint8_t len;
38
uint8_t pointer;
39
40
- uint8_t regs[PCA9552_NR_REGS];
41
+ uint8_t regs[PCA955X_NR_REGS];
42
uint8_t max_reg;
43
uint8_t pin_count;
44
-} PCA9552State;
45
+} PCA955xState;
25
46
26
#endif
47
#endif
27
diff --git a/hw/arm/exynos4_boards.c b/hw/arm/exynos4_boards.c
48
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
28
index XXXXXXX..XXXXXXX 100644
49
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/arm/exynos4_boards.c
50
--- a/hw/misc/pca9552.c
30
+++ b/hw/arm/exynos4_boards.c
51
+++ b/hw/misc/pca9552.c
31
@@ -XXX,XX +XXX,XX @@
52
@@ -XXX,XX +XXX,XX @@
32
#include "hw/arm/arm.h"
53
33
#include "exec/address-spaces.h"
54
static const char *led_state[] = {"on", "off", "pwm0", "pwm1"};
34
#include "hw/arm/exynos4210.h"
55
35
+#include "hw/net/lan9118.h"
56
-static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
36
#include "hw/boards.h"
57
+static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
37
58
{
38
#undef DEBUG
59
uint8_t reg = PCA9552_LS0 + (pin / 4);
39
@@ -XXX,XX +XXX,XX @@ static void lan9215_init(uint32_t base, qemu_irq irq)
60
uint8_t shift = (pin % 4) << 1;
40
/* This should be a 9215 but the 9118 is close enough */
61
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
41
if (nd_table[0].used) {
62
return extract32(s->regs[reg], shift, 2);
42
qemu_check_nic_model(&nd_table[0], "lan9118");
63
}
43
- dev = qdev_create(NULL, "lan9118");
64
44
+ dev = qdev_create(NULL, TYPE_LAN9118);
65
-static void pca9552_update_pin_input(PCA9552State *s)
45
qdev_set_nic_properties(dev, &nd_table[0]);
66
+static void pca955x_update_pin_input(PCA955xState *s)
46
qdev_prop_set_uint32(dev, "mode_16bit", 1);
67
{
47
qdev_init_nofail(dev);
68
int i;
48
diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c
69
49
index XXXXXXX..XXXXXXX 100644
70
for (i = 0; i < s->pin_count; i++) {
50
--- a/hw/arm/mps2-tz.c
71
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
51
+++ b/hw/arm/mps2-tz.c
72
uint8_t input_shift = (i % 8);
52
@@ -XXX,XX +XXX,XX @@
73
- uint8_t config = pca9552_pin_get_config(s, i);
53
#include "hw/arm/armsse.h"
74
+ uint8_t config = pca955x_pin_get_config(s, i);
54
#include "hw/dma/pl080.h"
75
55
#include "hw/ssi/pl022.h"
76
switch (config) {
56
+#include "hw/net/lan9118.h"
77
case PCA9552_LED_ON:
57
#include "net/net.h"
78
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
58
#include "hw/core/split-irq.h"
79
}
59
80
}
60
@@ -XXX,XX +XXX,XX @@ static MemoryRegion *make_eth_dev(MPS2TZMachineState *mms, void *opaque,
81
61
* except that it doesn't support the checksum-offload feature.
82
-static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
83
+static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
84
{
85
switch (reg) {
86
case PCA9552_INPUT0:
87
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
88
}
89
}
90
91
-static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
92
+static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
93
{
94
switch (reg) {
95
case PCA9552_PSC0:
96
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
97
case PCA9552_LS2:
98
case PCA9552_LS3:
99
s->regs[reg] = data;
100
- pca9552_update_pin_input(s);
101
+ pca955x_update_pin_input(s);
102
break;
103
104
case PCA9552_INPUT0:
105
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
106
* after each byte is sent to or received by the device. The index
107
* rollovers to 0 when the maximum register address is reached.
108
*/
109
-static void pca9552_autoinc(PCA9552State *s)
110
+static void pca955x_autoinc(PCA955xState *s)
111
{
112
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
113
uint8_t reg = s->pointer & 0xf;
114
@@ -XXX,XX +XXX,XX @@ static void pca9552_autoinc(PCA9552State *s)
115
}
116
}
117
118
-static uint8_t pca9552_recv(I2CSlave *i2c)
119
+static uint8_t pca955x_recv(I2CSlave *i2c)
120
{
121
- PCA9552State *s = PCA9552(i2c);
122
+ PCA955xState *s = PCA955X(i2c);
123
uint8_t ret;
124
125
- ret = pca9552_read(s, s->pointer & 0xf);
126
+ ret = pca955x_read(s, s->pointer & 0xf);
127
128
/*
129
* From the Specs:
130
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_recv(I2CSlave *i2c)
131
__func__);
132
}
133
134
- pca9552_autoinc(s);
135
+ pca955x_autoinc(s);
136
137
return ret;
138
}
139
140
-static int pca9552_send(I2CSlave *i2c, uint8_t data)
141
+static int pca955x_send(I2CSlave *i2c, uint8_t data)
142
{
143
- PCA9552State *s = PCA9552(i2c);
144
+ PCA955xState *s = PCA955X(i2c);
145
146
/* First byte sent by is the register address */
147
if (s->len == 0) {
148
s->pointer = data;
149
s->len++;
150
} else {
151
- pca9552_write(s, s->pointer & 0xf, data);
152
+ pca955x_write(s, s->pointer & 0xf, data);
153
154
- pca9552_autoinc(s);
155
+ pca955x_autoinc(s);
156
}
157
158
return 0;
159
}
160
161
-static int pca9552_event(I2CSlave *i2c, enum i2c_event event)
162
+static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
163
{
164
- PCA9552State *s = PCA9552(i2c);
165
+ PCA955xState *s = PCA955X(i2c);
166
167
s->len = 0;
168
return 0;
169
}
170
171
-static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
172
+static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
173
void *opaque, Error **errp)
174
{
175
- PCA9552State *s = PCA9552(obj);
176
+ PCA955xState *s = PCA955X(obj);
177
int led, rc, reg;
178
uint8_t state;
179
180
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
181
* reading the INPUTx reg
62
*/
182
*/
63
qemu_check_nic_model(nd, "lan9118");
183
reg = PCA9552_LS0 + led / 4;
64
- mms->lan9118 = qdev_create(NULL, "lan9118");
184
- state = (pca9552_read(s, reg) >> (led % 8)) & 0x3;
65
+ mms->lan9118 = qdev_create(NULL, TYPE_LAN9118);
185
+ state = (pca955x_read(s, reg) >> (led % 8)) & 0x3;
66
qdev_set_nic_properties(mms->lan9118, nd);
186
visit_type_str(v, name, (char **)&led_state[state], errp);
67
qdev_init_nofail(mms->lan9118);
187
}
68
188
69
diff --git a/hw/net/lan9118.c b/hw/net/lan9118.c
189
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
70
index XXXXXXX..XXXXXXX 100644
190
((state & 0x3) << (led_num << 1));
71
--- a/hw/net/lan9118.c
191
}
72
+++ b/hw/net/lan9118.c
192
73
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_lan9118_packet = {
193
-static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
194
+static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
195
void *opaque, Error **errp)
196
{
197
- PCA9552State *s = PCA9552(obj);
198
+ PCA955xState *s = PCA955X(obj);
199
Error *local_err = NULL;
200
int led, rc, reg, val;
201
uint8_t state;
202
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
203
}
204
205
reg = PCA9552_LS0 + led / 4;
206
- val = pca9552_read(s, reg);
207
+ val = pca955x_read(s, reg);
208
val = pca955x_ledsel(val, led % 4, state);
209
- pca9552_write(s, reg, val);
210
+ pca955x_write(s, reg, val);
211
}
212
213
static const VMStateDescription pca9552_vmstate = {
214
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription pca9552_vmstate = {
215
.version_id = 0,
216
.minimum_version_id = 0,
217
.fields = (VMStateField[]) {
218
- VMSTATE_UINT8(len, PCA9552State),
219
- VMSTATE_UINT8(pointer, PCA9552State),
220
- VMSTATE_UINT8_ARRAY(regs, PCA9552State, PCA9552_NR_REGS),
221
- VMSTATE_I2C_SLAVE(i2c, PCA9552State),
222
+ VMSTATE_UINT8(len, PCA955xState),
223
+ VMSTATE_UINT8(pointer, PCA955xState),
224
+ VMSTATE_UINT8_ARRAY(regs, PCA955xState, PCA955X_NR_REGS),
225
+ VMSTATE_I2C_SLAVE(i2c, PCA955xState),
226
VMSTATE_END_OF_LIST()
74
}
227
}
75
};
228
};
76
229
77
-#define TYPE_LAN9118 "lan9118"
230
static void pca9552_reset(DeviceState *dev)
78
#define LAN9118(obj) OBJECT_CHECK(lan9118_state, (obj), TYPE_LAN9118)
231
{
79
232
- PCA9552State *s = PCA9552(dev);
80
typedef struct {
233
+ PCA955xState *s = PCA955X(dev);
234
235
s->regs[PCA9552_PSC0] = 0xFF;
236
s->regs[PCA9552_PWM0] = 0x80;
237
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
238
s->regs[PCA9552_LS2] = 0x55;
239
s->regs[PCA9552_LS3] = 0x55;
240
241
- pca9552_update_pin_input(s);
242
+ pca955x_update_pin_input(s);
243
244
s->pointer = 0xFF;
245
s->len = 0;
246
}
247
248
-static void pca9552_initfn(Object *obj)
249
+static void pca955x_initfn(Object *obj)
250
{
251
- PCA9552State *s = PCA9552(obj);
252
+ PCA955xState *s = PCA955X(obj);
253
int led;
254
255
/* If support for the other PCA955X devices are implemented, these
256
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
257
char *name;
258
259
name = g_strdup_printf("led%d", led);
260
- object_property_add(obj, name, "bool", pca9552_get_led, pca9552_set_led,
261
+ object_property_add(obj, name, "bool", pca955x_get_led, pca955x_set_led,
262
NULL, NULL);
263
g_free(name);
264
}
265
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
266
DeviceClass *dc = DEVICE_CLASS(klass);
267
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
268
269
- k->event = pca9552_event;
270
- k->recv = pca9552_recv;
271
- k->send = pca9552_send;
272
+ k->event = pca955x_event;
273
+ k->recv = pca955x_recv;
274
+ k->send = pca955x_send;
275
dc->reset = pca9552_reset;
276
dc->vmsd = &pca9552_vmstate;
277
}
278
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
279
static const TypeInfo pca9552_info = {
280
.name = TYPE_PCA9552,
281
.parent = TYPE_I2C_SLAVE,
282
- .instance_init = pca9552_initfn,
283
- .instance_size = sizeof(PCA9552State),
284
+ .instance_init = pca955x_initfn,
285
+ .instance_size = sizeof(PCA955xState),
286
.class_init = pca9552_class_init,
287
};
288
289
-static void pca9552_register_types(void)
290
+static void pca955x_register_types(void)
291
{
292
type_register_static(&pca9552_info);
293
}
294
295
-type_init(pca9552_register_types)
296
+type_init(pca955x_register_types)
81
--
297
--
82
2.20.1
298
2.20.1
83
299
84
300
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Reviewed-by: Thomas Huth <thuth@redhat.com>
3
Extract the code common to the PCA955x family in PCA955xClass,
4
Reviewed-by: Markus Armbruster <armbru@redhat.com>
4
keeping the PCA9552 specific parts into pca9552_class_init().
5
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
5
Remove the 'TODO' comment added in commit 5141d4158cf.
6
Message-id: 20190412165416.7977-7-philmd@redhat.com
6
7
Suggested-by: Cédric Le Goater <clg@kaod.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-5-f4bug@amsat.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
13
---
9
include/hw/devices.h | 14 --------------
14
include/hw/misc/pca9552.h | 6 ++--
10
include/hw/misc/cbus.h | 32 ++++++++++++++++++++++++++++++++
15
hw/misc/pca9552.c | 66 ++++++++++++++++++++++++++++-----------
11
hw/arm/nseries.c | 1 +
16
2 files changed, 51 insertions(+), 21 deletions(-)
12
hw/misc/cbus.c | 2 +-
17
13
MAINTAINERS | 1 +
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
14
5 files changed, 35 insertions(+), 15 deletions(-)
15
create mode 100644 include/hw/misc/cbus.h
16
17
diff --git a/include/hw/devices.h b/include/hw/devices.h
18
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/devices.h
20
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/devices.h
21
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ void tsc2005_set_transform(void *opaque, MouseTransformInfo *info);
22
/* stellaris_input.c */
23
void stellaris_gamepad_init(int n, qemu_irq *irq, const int *keycode);
24
25
-/* cbus.c */
26
-typedef struct {
27
- qemu_irq clk;
28
- qemu_irq dat;
29
- qemu_irq sel;
30
-} CBus;
31
-CBus *cbus_init(qemu_irq dat_out);
32
-void cbus_attach(CBus *bus, void *slave_opaque);
33
-
34
-void *retu_init(qemu_irq irq, int vilma);
35
-void *tahvo_init(qemu_irq irq, int betty);
36
-
37
-void retu_key_event(void *retu, int state);
38
-
39
#endif
40
diff --git a/include/hw/misc/cbus.h b/include/hw/misc/cbus.h
41
new file mode 100644
42
index XXXXXXX..XXXXXXX
43
--- /dev/null
44
+++ b/include/hw/misc/cbus.h
45
@@ -XXX,XX +XXX,XX @@
46
+/*
47
+ * CBUS three-pin bus and the Retu / Betty / Tahvo / Vilma / Avilma /
48
+ * Hinku / Vinku / Ahne / Pihi chips used in various Nokia platforms.
49
+ * Based on reverse-engineering of a linux driver.
50
+ *
51
+ * Copyright (C) 2008 Nokia Corporation
52
+ * Written by Andrzej Zaborowski
53
+ *
54
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
55
+ * See the COPYING file in the top-level directory.
56
+ */
57
+
58
+#ifndef HW_MISC_CBUS_H
59
+#define HW_MISC_CBUS_H
60
+
61
+#include "hw/irq.h"
62
+
63
+typedef struct {
64
+ qemu_irq clk;
65
+ qemu_irq dat;
66
+ qemu_irq sel;
67
+} CBus;
68
+
69
+CBus *cbus_init(qemu_irq dat_out);
70
+void cbus_attach(CBus *bus, void *slave_opaque);
71
+
72
+void *retu_init(qemu_irq irq, int vilma);
73
+void *tahvo_init(qemu_irq irq, int betty);
74
+
75
+void retu_key_event(void *retu, int state);
76
+
77
+#endif
78
diff --git a/hw/arm/nseries.c b/hw/arm/nseries.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/hw/arm/nseries.c
81
+++ b/hw/arm/nseries.c
82
@@ -XXX,XX +XXX,XX @@
22
@@ -XXX,XX +XXX,XX @@
83
#include "hw/i2c/i2c.h"
23
#include "hw/i2c/i2c.h"
84
#include "hw/devices.h"
24
85
#include "hw/display/blizzard.h"
25
#define TYPE_PCA9552 "pca9552"
86
+#include "hw/misc/cbus.h"
26
-#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
87
#include "hw/misc/tmp105.h"
27
+#define TYPE_PCA955X "pca955x"
88
#include "hw/block/flash.h"
28
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA955X)
89
#include "hw/hw.h"
29
90
diff --git a/hw/misc/cbus.c b/hw/misc/cbus.c
30
#define PCA955X_NR_REGS 10
31
+#define PCA955X_PIN_COUNT_MAX 16
32
33
typedef struct PCA955xState {
34
/*< private >*/
35
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
36
uint8_t pointer;
37
38
uint8_t regs[PCA955X_NR_REGS];
39
- uint8_t max_reg;
40
- uint8_t pin_count;
41
} PCA955xState;
42
43
#endif
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
91
index XXXXXXX..XXXXXXX 100644
45
index XXXXXXX..XXXXXXX 100644
92
--- a/hw/misc/cbus.c
46
--- a/hw/misc/pca9552.c
93
+++ b/hw/misc/cbus.c
47
+++ b/hw/misc/pca9552.c
94
@@ -XXX,XX +XXX,XX @@
48
@@ -XXX,XX +XXX,XX @@
95
#include "qemu/osdep.h"
49
* https://www.nxp.com/docs/en/application-note/AN264.pdf
96
#include "hw/hw.h"
50
*
97
#include "hw/irq.h"
51
* Copyright (c) 2017-2018, IBM Corporation.
98
-#include "hw/devices.h"
52
+ * Copyright (c) 2020 Philippe Mathieu-Daudé
99
+#include "hw/misc/cbus.h"
53
*
100
#include "sysemu/sysemu.h"
54
* This work is licensed under the terms of the GNU GPL, version 2 or
101
55
* later. See the COPYING file in the top-level directory.
102
//#define DEBUG
56
@@ -XXX,XX +XXX,XX @@
103
diff --git a/MAINTAINERS b/MAINTAINERS
57
#include "qapi/error.h"
104
index XXXXXXX..XXXXXXX 100644
58
#include "qapi/visitor.h"
105
--- a/MAINTAINERS
59
106
+++ b/MAINTAINERS
60
+typedef struct PCA955xClass {
107
@@ -XXX,XX +XXX,XX @@ F: hw/input/tsc2005.c
61
+ /*< private >*/
108
F: hw/misc/cbus.c
62
+ I2CSlaveClass parent_class;
109
F: hw/timer/twl92230.c
63
+ /*< public >*/
110
F: include/hw/display/blizzard.h
64
+
111
+F: include/hw/misc/cbus.h
65
+ uint8_t pin_count;
112
66
+ uint8_t max_reg;
113
Palm
67
+} PCA955xClass;
114
M: Andrzej Zaborowski <balrogg@gmail.com>
68
+
69
+#define PCA955X_CLASS(klass) \
70
+ OBJECT_CLASS_CHECK(PCA955xClass, (klass), TYPE_PCA955X)
71
+#define PCA955X_GET_CLASS(obj) \
72
+ OBJECT_GET_CLASS(PCA955xClass, (obj), TYPE_PCA955X)
73
+
74
#define PCA9552_LED_ON 0x0
75
#define PCA9552_LED_OFF 0x1
76
#define PCA9552_LED_PWM0 0x2
77
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
78
79
static void pca955x_update_pin_input(PCA955xState *s)
80
{
81
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
82
int i;
83
84
- for (i = 0; i < s->pin_count; i++) {
85
+ for (i = 0; i < k->pin_count; i++) {
86
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
87
uint8_t input_shift = (i % 8);
88
uint8_t config = pca955x_pin_get_config(s, i);
89
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
90
*/
91
static void pca955x_autoinc(PCA955xState *s)
92
{
93
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
94
+
95
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
96
uint8_t reg = s->pointer & 0xf;
97
98
- reg = (reg + 1) % (s->max_reg + 1);
99
+ reg = (reg + 1) % (k->max_reg + 1);
100
s->pointer = reg | PCA9552_AUTOINC;
101
}
102
}
103
@@ -XXX,XX +XXX,XX @@ static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
104
static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
105
void *opaque, Error **errp)
106
{
107
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
108
PCA955xState *s = PCA955X(obj);
109
int led, rc, reg;
110
uint8_t state;
111
@@ -XXX,XX +XXX,XX @@ static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
112
error_setg(errp, "%s: error reading %s", __func__, name);
113
return;
114
}
115
- if (led < 0 || led > s->pin_count) {
116
+ if (led < 0 || led > k->pin_count) {
117
error_setg(errp, "%s invalid led %s", __func__, name);
118
return;
119
}
120
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
121
static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
122
void *opaque, Error **errp)
123
{
124
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
125
PCA955xState *s = PCA955X(obj);
126
Error *local_err = NULL;
127
int led, rc, reg, val;
128
@@ -XXX,XX +XXX,XX @@ static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
129
error_setg(errp, "%s: error reading %s", __func__, name);
130
return;
131
}
132
- if (led < 0 || led > s->pin_count) {
133
+ if (led < 0 || led > k->pin_count) {
134
error_setg(errp, "%s invalid led %s", __func__, name);
135
return;
136
}
137
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
138
139
static void pca955x_initfn(Object *obj)
140
{
141
- PCA955xState *s = PCA955X(obj);
142
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
143
int led;
144
145
- /* If support for the other PCA955X devices are implemented, these
146
- * constant values might be part of class structure describing the
147
- * PCA955X device
148
- */
149
- s->max_reg = PCA9552_LS3;
150
- s->pin_count = 16;
151
-
152
- for (led = 0; led < s->pin_count; led++) {
153
+ assert(k->pin_count <= PCA955X_PIN_COUNT_MAX);
154
+ for (led = 0; led < k->pin_count; led++) {
155
char *name;
156
157
name = g_strdup_printf("led%d", led);
158
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
159
}
160
}
161
162
-static void pca9552_class_init(ObjectClass *klass, void *data)
163
+static void pca955x_class_init(ObjectClass *klass, void *data)
164
{
165
- DeviceClass *dc = DEVICE_CLASS(klass);
166
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
167
168
k->event = pca955x_event;
169
k->recv = pca955x_recv;
170
k->send = pca955x_send;
171
+}
172
+
173
+static const TypeInfo pca955x_info = {
174
+ .name = TYPE_PCA955X,
175
+ .parent = TYPE_I2C_SLAVE,
176
+ .instance_init = pca955x_initfn,
177
+ .instance_size = sizeof(PCA955xState),
178
+ .class_init = pca955x_class_init,
179
+ .abstract = true,
180
+};
181
+
182
+static void pca9552_class_init(ObjectClass *oc, void *data)
183
+{
184
+ DeviceClass *dc = DEVICE_CLASS(oc);
185
+ PCA955xClass *pc = PCA955X_CLASS(oc);
186
+
187
dc->reset = pca9552_reset;
188
dc->vmsd = &pca9552_vmstate;
189
+ pc->max_reg = PCA9552_LS3;
190
+ pc->pin_count = 16;
191
}
192
193
static const TypeInfo pca9552_info = {
194
.name = TYPE_PCA9552,
195
- .parent = TYPE_I2C_SLAVE,
196
- .instance_init = pca955x_initfn,
197
- .instance_size = sizeof(PCA955xState),
198
+ .parent = TYPE_PCA955X,
199
.class_init = pca9552_class_init,
200
};
201
202
static void pca955x_register_types(void)
203
{
204
+ type_register_static(&pca955x_info);
205
type_register_static(&pca9552_info);
206
}
207
115
--
208
--
116
2.20.1
209
2.20.1
117
210
118
211
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Add an entries the Blizzard device in MAINTAINERS.
3
Add a description field to distinguish between multiple devices.
4
4
5
Reviewed-by: Thomas Huth <thuth@redhat.com>
5
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Reviewed-by: Markus Armbruster <armbru@redhat.com>
6
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
7
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20190412165416.7977-6-philmd@redhat.com
8
Message-id: 20200623072723.6324-6-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
include/hw/devices.h | 7 -------
11
include/hw/misc/pca9552.h | 1 +
12
include/hw/display/blizzard.h | 22 ++++++++++++++++++++++
12
hw/misc/pca9552.c | 18 ++++++++++++++++++
13
hw/arm/nseries.c | 1 +
13
2 files changed, 19 insertions(+)
14
hw/display/blizzard.c | 2 +-
15
MAINTAINERS | 2 ++
16
5 files changed, 26 insertions(+), 8 deletions(-)
17
create mode 100644 include/hw/display/blizzard.h
18
14
19
diff --git a/include/hw/devices.h b/include/hw/devices.h
15
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
20
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
21
--- a/include/hw/devices.h
17
--- a/include/hw/misc/pca9552.h
22
+++ b/include/hw/devices.h
18
+++ b/include/hw/misc/pca9552.h
23
@@ -XXX,XX +XXX,XX @@ void tsc2005_set_transform(void *opaque, MouseTransformInfo *info);
19
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
24
/* stellaris_input.c */
20
uint8_t pointer;
25
void stellaris_gamepad_init(int n, qemu_irq *irq, const int *keycode);
21
26
22
uint8_t regs[PCA955X_NR_REGS];
27
-/* blizzard.c */
23
+ char *description; /* For debugging purpose only */
28
-void *s1d13745_init(qemu_irq gpio_int);
24
} PCA955xState;
29
-void s1d13745_write(void *opaque, int dc, uint16_t value);
25
30
-void s1d13745_write_block(void *opaque, int dc,
26
#endif
31
- void *buf, size_t len, int pitch);
27
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
32
-uint16_t s1d13745_read(void *opaque, int dc);
33
-
34
/* cbus.c */
35
typedef struct {
36
qemu_irq clk;
37
diff --git a/include/hw/display/blizzard.h b/include/hw/display/blizzard.h
38
new file mode 100644
39
index XXXXXXX..XXXXXXX
40
--- /dev/null
41
+++ b/include/hw/display/blizzard.h
42
@@ -XXX,XX +XXX,XX @@
43
+/*
44
+ * Epson S1D13744/S1D13745 (Blizzard/Hailstorm/Tornado) LCD/TV controller.
45
+ *
46
+ * Copyright (C) 2008 Nokia Corporation
47
+ * Written by Andrzej Zaborowski
48
+ *
49
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
50
+ * See the COPYING file in the top-level directory.
51
+ */
52
+
53
+#ifndef HW_DISPLAY_BLIZZARD_H
54
+#define HW_DISPLAY_BLIZZARD_H
55
+
56
+#include "hw/irq.h"
57
+
58
+void *s1d13745_init(qemu_irq gpio_int);
59
+void s1d13745_write(void *opaque, int dc, uint16_t value);
60
+void s1d13745_write_block(void *opaque, int dc,
61
+ void *buf, size_t len, int pitch);
62
+uint16_t s1d13745_read(void *opaque, int dc);
63
+
64
+#endif
65
diff --git a/hw/arm/nseries.c b/hw/arm/nseries.c
66
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
67
--- a/hw/arm/nseries.c
29
--- a/hw/misc/pca9552.c
68
+++ b/hw/arm/nseries.c
30
+++ b/hw/misc/pca9552.c
69
@@ -XXX,XX +XXX,XX @@
70
#include "hw/boards.h"
71
#include "hw/i2c/i2c.h"
72
#include "hw/devices.h"
73
+#include "hw/display/blizzard.h"
74
#include "hw/misc/tmp105.h"
75
#include "hw/block/flash.h"
76
#include "hw/hw.h"
77
diff --git a/hw/display/blizzard.c b/hw/display/blizzard.c
78
index XXXXXXX..XXXXXXX 100644
79
--- a/hw/display/blizzard.c
80
+++ b/hw/display/blizzard.c
81
@@ -XXX,XX +XXX,XX @@
31
@@ -XXX,XX +XXX,XX @@
82
#include "qemu/osdep.h"
32
#include "qemu/osdep.h"
83
#include "qemu-common.h"
33
#include "qemu/log.h"
84
#include "ui/console.h"
34
#include "qemu/module.h"
85
-#include "hw/devices.h"
35
+#include "hw/qdev-properties.h"
86
+#include "hw/display/blizzard.h"
36
#include "hw/misc/pca9552.h"
87
#include "ui/pixel_ops.h"
37
#include "hw/misc/pca9552_regs.h"
88
38
#include "migration/vmstate.h"
89
typedef void (*blizzard_fn_t)(uint8_t *, const uint8_t *, unsigned int);
39
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
90
diff --git a/MAINTAINERS b/MAINTAINERS
40
}
91
index XXXXXXX..XXXXXXX 100644
41
}
92
--- a/MAINTAINERS
42
93
+++ b/MAINTAINERS
43
+static void pca955x_realize(DeviceState *dev, Error **errp)
94
@@ -XXX,XX +XXX,XX @@ M: Peter Maydell <peter.maydell@linaro.org>
44
+{
95
L: qemu-arm@nongnu.org
45
+ PCA955xState *s = PCA955X(dev);
96
S: Odd Fixes
46
+
97
F: hw/arm/nseries.c
47
+ if (!s->description) {
98
+F: hw/display/blizzard.c
48
+ s->description = g_strdup("pca-unspecified");
99
F: hw/input/lm832x.c
49
+ }
100
F: hw/input/tsc2005.c
50
+}
101
F: hw/misc/cbus.c
51
+
102
F: hw/timer/twl92230.c
52
+static Property pca955x_properties[] = {
103
+F: include/hw/display/blizzard.h
53
+ DEFINE_PROP_STRING("description", PCA955xState, description),
104
54
+ DEFINE_PROP_END_OF_LIST(),
105
Palm
55
+};
106
M: Andrzej Zaborowski <balrogg@gmail.com>
56
+
57
static void pca955x_class_init(ObjectClass *klass, void *data)
58
{
59
+ DeviceClass *dc = DEVICE_CLASS(klass);
60
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
61
62
k->event = pca955x_event;
63
k->recv = pca955x_recv;
64
k->send = pca955x_send;
65
+ dc->realize = pca955x_realize;
66
+ device_class_set_props(dc, pca955x_properties);
67
}
68
69
static const TypeInfo pca955x_info = {
107
--
70
--
108
2.20.1
71
2.20.1
109
72
110
73
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
This commit finally deletes "hw/devices.h".
3
Add a trivial representation of the PCA9552 GPIOs.
4
4
5
Reviewed-by: Markus Armbruster <armbru@redhat.com>
5
Example booting obmc-phosphor-image:
6
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
6
7
Message-id: 20190412165416.7977-13-philmd@redhat.com
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_status
8
1592689902.327837:pca955x_gpio_status pca-unspecified GPIOs 0-15 [*...............]
9
1592689902.329934:pca955x_gpio_status pca-unspecified GPIOs 0-15 [**..............]
10
1592689902.330717:pca955x_gpio_status pca-unspecified GPIOs 0-15 [***.............]
11
1592689902.331431:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****............]
12
1592689902.332163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*..]
13
1592689902.332888:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........**.]
14
1592689902.333629:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
15
1592690032.793289:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
16
1592690033.303163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
17
1592690033.812962:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
18
1592690034.323234:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
19
1592690034.832922:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
20
21
We notice the GPIO #14 (front-power LED) starts to blink.
22
23
This LED is described in the witherspoon device-tree [*]:
24
25
front-power {
26
retain-state-shutdown;
27
default-state = "keep";
28
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
29
};
30
31
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
32
33
Suggested-by: Cédric Le Goater <clg@kaod.org>
34
Reviewed-by: Cédric Le Goater <clg@kaod.org>
35
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
36
Tested-by: Cédric Le Goater <clg@kaod.org>
37
Message-id: 20200623072723.6324-7-f4bug@amsat.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
38
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
39
---
10
include/hw/devices.h | 11 -----------
40
hw/misc/pca9552.c | 39 +++++++++++++++++++++++++++++++++++++++
11
include/hw/net/smc91c111.h | 19 +++++++++++++++++++
41
hw/misc/trace-events | 3 +++
12
hw/arm/gumstix.c | 2 +-
42
2 files changed, 42 insertions(+)
13
hw/arm/integratorcp.c | 2 +-
14
hw/arm/mainstone.c | 2 +-
15
hw/arm/realview.c | 2 +-
16
hw/arm/versatilepb.c | 2 +-
17
hw/net/smc91c111.c | 2 +-
18
8 files changed, 25 insertions(+), 17 deletions(-)
19
delete mode 100644 include/hw/devices.h
20
create mode 100644 include/hw/net/smc91c111.h
21
43
22
diff --git a/include/hw/devices.h b/include/hw/devices.h
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
23
deleted file mode 100644
24
index XXXXXXX..XXXXXXX
25
--- a/include/hw/devices.h
26
+++ /dev/null
27
@@ -XXX,XX +XXX,XX @@
28
-#ifndef QEMU_DEVICES_H
29
-#define QEMU_DEVICES_H
30
-
31
-/* Devices that have nowhere better to go. */
32
-
33
-#include "hw/hw.h"
34
-
35
-/* smc91c111.c */
36
-void smc91c111_init(NICInfo *, uint32_t, qemu_irq);
37
-
38
-#endif
39
diff --git a/include/hw/net/smc91c111.h b/include/hw/net/smc91c111.h
40
new file mode 100644
41
index XXXXXXX..XXXXXXX
42
--- /dev/null
43
+++ b/include/hw/net/smc91c111.h
44
@@ -XXX,XX +XXX,XX @@
45
+/*
46
+ * SMSC 91C111 Ethernet interface emulation
47
+ *
48
+ * Copyright (c) 2005 CodeSourcery, LLC.
49
+ * Written by Paul Brook
50
+ *
51
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
52
+ * See the COPYING file in the top-level directory.
53
+ */
54
+
55
+#ifndef HW_NET_SMC91C111_H
56
+#define HW_NET_SMC91C111_H
57
+
58
+#include "hw/irq.h"
59
+#include "net/net.h"
60
+
61
+void smc91c111_init(NICInfo *, uint32_t, qemu_irq);
62
+
63
+#endif
64
diff --git a/hw/arm/gumstix.c b/hw/arm/gumstix.c
65
index XXXXXXX..XXXXXXX 100644
45
index XXXXXXX..XXXXXXX 100644
66
--- a/hw/arm/gumstix.c
46
--- a/hw/misc/pca9552.c
67
+++ b/hw/arm/gumstix.c
47
+++ b/hw/misc/pca9552.c
68
@@ -XXX,XX +XXX,XX @@
69
#include "hw/arm/pxa.h"
70
#include "net/net.h"
71
#include "hw/block/flash.h"
72
-#include "hw/devices.h"
73
+#include "hw/net/smc91c111.h"
74
#include "hw/boards.h"
75
#include "exec/address-spaces.h"
76
#include "sysemu/qtest.h"
77
diff --git a/hw/arm/integratorcp.c b/hw/arm/integratorcp.c
78
index XXXXXXX..XXXXXXX 100644
79
--- a/hw/arm/integratorcp.c
80
+++ b/hw/arm/integratorcp.c
81
@@ -XXX,XX +XXX,XX @@
82
#include "qemu-common.h"
83
#include "cpu.h"
84
#include "hw/sysbus.h"
85
-#include "hw/devices.h"
86
#include "hw/boards.h"
87
#include "hw/arm/arm.h"
88
#include "hw/misc/arm_integrator_debug.h"
89
+#include "hw/net/smc91c111.h"
90
#include "net/net.h"
91
#include "exec/address-spaces.h"
92
#include "sysemu/sysemu.h"
93
diff --git a/hw/arm/mainstone.c b/hw/arm/mainstone.c
94
index XXXXXXX..XXXXXXX 100644
95
--- a/hw/arm/mainstone.c
96
+++ b/hw/arm/mainstone.c
97
@@ -XXX,XX +XXX,XX @@
98
#include "hw/arm/pxa.h"
99
#include "hw/arm/arm.h"
100
#include "net/net.h"
101
-#include "hw/devices.h"
102
+#include "hw/net/smc91c111.h"
103
#include "hw/boards.h"
104
#include "hw/block/flash.h"
105
#include "hw/sysbus.h"
106
diff --git a/hw/arm/realview.c b/hw/arm/realview.c
107
index XXXXXXX..XXXXXXX 100644
108
--- a/hw/arm/realview.c
109
+++ b/hw/arm/realview.c
110
@@ -XXX,XX +XXX,XX @@
111
#include "hw/sysbus.h"
112
#include "hw/arm/arm.h"
113
#include "hw/arm/primecell.h"
114
-#include "hw/devices.h"
115
#include "hw/net/lan9118.h"
116
+#include "hw/net/smc91c111.h"
117
#include "hw/pci/pci.h"
118
#include "net/net.h"
119
#include "sysemu/sysemu.h"
120
diff --git a/hw/arm/versatilepb.c b/hw/arm/versatilepb.c
121
index XXXXXXX..XXXXXXX 100644
122
--- a/hw/arm/versatilepb.c
123
+++ b/hw/arm/versatilepb.c
124
@@ -XXX,XX +XXX,XX @@
125
#include "cpu.h"
126
#include "hw/sysbus.h"
127
#include "hw/arm/arm.h"
128
-#include "hw/devices.h"
129
+#include "hw/net/smc91c111.h"
130
#include "net/net.h"
131
#include "sysemu/sysemu.h"
132
#include "hw/pci/pci.h"
133
diff --git a/hw/net/smc91c111.c b/hw/net/smc91c111.c
134
index XXXXXXX..XXXXXXX 100644
135
--- a/hw/net/smc91c111.c
136
+++ b/hw/net/smc91c111.c
137
@@ -XXX,XX +XXX,XX @@
48
@@ -XXX,XX +XXX,XX @@
138
#include "qemu/osdep.h"
49
#include "qemu/osdep.h"
139
#include "hw/sysbus.h"
140
#include "net/net.h"
141
-#include "hw/devices.h"
142
+#include "hw/net/smc91c111.h"
143
#include "qemu/log.h"
50
#include "qemu/log.h"
144
/* For crc32 */
51
#include "qemu/module.h"
145
#include <zlib.h>
52
+#include "qemu/bitops.h"
53
#include "hw/qdev-properties.h"
54
#include "hw/misc/pca9552.h"
55
#include "hw/misc/pca9552_regs.h"
56
#include "migration/vmstate.h"
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
+#include "trace.h"
60
61
typedef struct PCA955xClass {
62
/*< private >*/
63
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
64
return extract32(s->regs[reg], shift, 2);
65
}
66
67
+/* Return INPUT status (bit #N belongs to GPIO #N) */
68
+static uint16_t pca955x_pins_get_status(PCA955xState *s)
69
+{
70
+ return (s->regs[PCA9552_INPUT1] << 8) | s->regs[PCA9552_INPUT0];
71
+}
72
+
73
+static void pca955x_display_pins_status(PCA955xState *s,
74
+ uint16_t previous_pins_status)
75
+{
76
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
77
+ uint16_t pins_status, pins_changed;
78
+ int i;
79
+
80
+ pins_status = pca955x_pins_get_status(s);
81
+ pins_changed = previous_pins_status ^ pins_status;
82
+ if (!pins_changed) {
83
+ return;
84
+ }
85
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_STATUS)) {
86
+ char *buf = g_newa(char, k->pin_count + 1);
87
+
88
+ for (i = 0; i < k->pin_count; i++) {
89
+ if (extract32(pins_status, i, 1)) {
90
+ buf[i] = '*';
91
+ } else {
92
+ buf[i] = '.';
93
+ }
94
+ }
95
+ buf[i] = '\0';
96
+ trace_pca955x_gpio_status(s->description, buf);
97
+ }
98
+}
99
+
100
static void pca955x_update_pin_input(PCA955xState *s)
101
{
102
PCA955xClass *k = PCA955X_GET_CLASS(s);
103
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
104
105
static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
106
{
107
+ uint16_t pins_status;
108
+
109
switch (reg) {
110
case PCA9552_PSC0:
111
case PCA9552_PWM0:
112
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
113
case PCA9552_LS1:
114
case PCA9552_LS2:
115
case PCA9552_LS3:
116
+ pins_status = pca955x_pins_get_status(s);
117
s->regs[reg] = data;
118
pca955x_update_pin_input(s);
119
+ pca955x_display_pins_status(s, pins_status);
120
break;
121
122
case PCA9552_INPUT0:
123
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
124
index XXXXXXX..XXXXXXX 100644
125
--- a/hw/misc/trace-events
126
+++ b/hw/misc/trace-events
127
@@ -XXX,XX +XXX,XX @@ via1_adb_poll(uint8_t data, const char *vadbint, int status, int index, int size
128
# grlib_ahb_apb_pnp.c
129
grlib_ahb_pnp_read(uint64_t addr, uint32_t value) "AHB PnP read addr:0x%03"PRIx64" data:0x%08x"
130
grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx64" data:0x%08x"
131
+
132
+# pca9552.c
133
+pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
146
--
134
--
147
2.20.1
135
2.20.1
148
136
149
137
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Reviewed-by: Thomas Huth <thuth@redhat.com>
3
We have 2 distinct PCA9552 devices. Set their description
4
to distinguish them when looking at the trace events.
5
6
Description name taken from:
7
https://github.com/open-power/witherspoon-xml/blob/master/witherspoon.xml
8
4
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Reviewed-by: Cédric Le Goater <clg@kaod.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Corey Minyard <cminyard@mvista.com>
5
Reviewed-by: Markus Armbruster <armbru@redhat.com>
12
Reviewed-by: Markus Armbruster <armbru@redhat.com>
6
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
13
Tested-by: Cédric Le Goater <clg@kaod.org>
7
Message-id: 20190412165416.7977-2-philmd@redhat.com
14
Message-id: 20200623072723.6324-8-f4bug@amsat.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
16
---
10
hw/arm/aspeed.c | 13 +++++++++----
17
hw/arm/aspeed.c | 13 +++++++++----
11
1 file changed, 9 insertions(+), 4 deletions(-)
18
1 file changed, 9 insertions(+), 4 deletions(-)
12
19
13
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
20
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
14
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
15
--- a/hw/arm/aspeed.c
22
--- a/hw/arm/aspeed.c
16
+++ b/hw/arm/aspeed.c
23
+++ b/hw/arm/aspeed.c
17
@@ -XXX,XX +XXX,XX @@
24
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
18
#include "hw/arm/aspeed_soc.h"
25
{
19
#include "hw/boards.h"
20
#include "hw/i2c/smbus_eeprom.h"
21
+#include "hw/misc/pca9552.h"
22
+#include "hw/misc/tmp105.h"
23
#include "qemu/log.h"
24
#include "sysemu/block-backend.h"
25
#include "hw/loader.h"
26
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
27
eeprom_buf);
28
29
/* The AST2500 EVB expects a LM75 but a TMP105 is compatible */
30
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 7), "tmp105", 0x4d);
31
+ i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 7),
32
+ TYPE_TMP105, 0x4d);
33
34
/* The AST2500 EVB does not have an RTC. Let's pretend that one is
35
* plugged on the I2C bus header */
36
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
37
AspeedSoCState *soc = &bmc->soc;
26
AspeedSoCState *soc = &bmc->soc;
38
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
27
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
39
28
+ DeviceState *dev;
40
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), "pca9552", 0x60);
29
41
+ i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
30
/* Bus 3: TODO bmp280@77 */
42
+ 0x60);
31
/* Bus 3: TODO max31785@52 */
32
/* Bus 3: TODO dps310@76 */
33
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
34
- 0x60);
35
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
36
+ qdev_prop_set_string(dev, "description", "pca1");
37
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3),
38
+ &error_fatal);
43
39
44
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
40
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
45
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
41
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
46
42
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
47
/* The Witherspoon expects a TMP275 but a TMP105 is compatible */
48
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 9), "tmp105", 0x4a);
49
+ i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 9), TYPE_TMP105,
50
+ 0x4a);
51
52
/* The witherspoon board expects Epson RX8900 I2C RTC but a ds1338 is
53
* good enough */
54
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
55
43
56
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
44
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
57
eeprom_buf);
45
eeprom_buf);
58
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "pca9552",
46
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
59
+ i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
47
- 0x60);
60
0x60);
48
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
49
+ qdev_prop_set_string(dev, "description", "pca0");
50
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11),
51
+ &error_fatal);
52
/* Bus 11: TODO ucd90160@64 */
61
}
53
}
62
54
63
--
55
--
64
2.20.1
56
2.20.1
65
57
66
58
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Emit a trace event when a GPIO change its state.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_change
8
1592690552.687372:pca955x_gpio_change pca1 GPIO id:0 status: 0 -> 1
9
1592690552.690169:pca955x_gpio_change pca1 GPIO id:1 status: 0 -> 1
10
1592690552.691673:pca955x_gpio_change pca1 GPIO id:2 status: 0 -> 1
11
1592690552.696886:pca955x_gpio_change pca1 GPIO id:3 status: 0 -> 1
12
1592690552.698614:pca955x_gpio_change pca1 GPIO id:13 status: 0 -> 1
13
1592690552.699833:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
14
1592690552.700842:pca955x_gpio_change pca1 GPIO id:15 status: 0 -> 1
15
1592690683.841921:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
16
1592690683.861660:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
17
1592690684.371460:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
18
1592690684.882115:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
19
1592690685.391411:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
20
1592690685.901391:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
21
1592690686.411678:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
22
1592690686.921279:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
23
24
We notice the GPIO #14 (front-power LED) starts to blink.
25
26
This LED is described in the witherspoon device-tree [*]:
27
28
front-power {
29
retain-state-shutdown;
30
default-state = "keep";
31
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
32
};
33
34
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
35
36
Reviewed-by: Cédric Le Goater <clg@kaod.org>
37
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
38
Tested-by: Cédric Le Goater <clg@kaod.org>
39
Message-id: 20200623072723.6324-9-f4bug@amsat.org
40
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
41
---
42
hw/misc/pca9552.c | 15 +++++++++++++++
43
hw/misc/trace-events | 1 +
44
2 files changed, 16 insertions(+)
45
46
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/hw/misc/pca9552.c
49
+++ b/hw/misc/pca9552.c
50
@@ -XXX,XX +XXX,XX @@ static void pca955x_display_pins_status(PCA955xState *s,
51
buf[i] = '\0';
52
trace_pca955x_gpio_status(s->description, buf);
53
}
54
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_CHANGE)) {
55
+ for (i = 0; i < k->pin_count; i++) {
56
+ if (extract32(pins_changed, i, 1)) {
57
+ unsigned new_state = extract32(pins_status, i, 1);
58
+
59
+ /*
60
+ * We display the state using the PCA logic ("active-high").
61
+ * This is not the state of the LED, which signal might be
62
+ * wired "active-low" on the board.
63
+ */
64
+ trace_pca955x_gpio_change(s->description, i,
65
+ !new_state, new_state);
66
+ }
67
+ }
68
+ }
69
}
70
71
static void pca955x_update_pin_input(PCA955xState *s)
72
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/misc/trace-events
75
+++ b/hw/misc/trace-events
76
@@ -XXX,XX +XXX,XX @@ grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx6
77
78
# pca9552.c
79
pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
80
+pca955x_gpio_change(const char *description, unsigned id, unsigned prev_state, unsigned current_state) "%s GPIO id:%u status: %u -> %u"
81
--
82
2.20.1
83
84
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
The PCA9552 has 16 GPIOs which can be used as input,
4
output or PWM mode. QEMU models the output GPIO with
5
the qemu_irq type. Let the device expose the 16 GPIOs
6
to allow us to later connect LEDs to these outputs.
7
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-10-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 1 +
15
hw/misc/pca9552.c | 6 ++++++
16
2 files changed, 7 insertions(+)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
23
uint8_t pointer;
24
25
uint8_t regs[PCA955X_NR_REGS];
26
+ qemu_irq gpio[PCA955X_PIN_COUNT_MAX];
27
char *description; /* For debugging purpose only */
28
} PCA955xState;
29
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@
35
#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
+#include "hw/irq.h"
39
#include "migration/vmstate.h"
40
#include "qapi/error.h"
41
#include "qapi/visitor.h"
42
@@ -XXX,XX +XXX,XX @@ static void pca955x_update_pin_input(PCA955xState *s)
43
44
switch (config) {
45
case PCA9552_LED_ON:
46
+ qemu_set_irq(s->gpio[i], 1);
47
s->regs[input_reg] |= 1 << input_shift;
48
break;
49
case PCA9552_LED_OFF:
50
+ qemu_set_irq(s->gpio[i], 0);
51
s->regs[input_reg] &= ~(1 << input_shift);
52
break;
53
case PCA9552_LED_PWM0:
54
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
55
56
static void pca955x_realize(DeviceState *dev, Error **errp)
57
{
58
+ PCA955xClass *k = PCA955X_GET_CLASS(dev);
59
PCA955xState *s = PCA955X(dev);
60
61
if (!s->description) {
62
s->description = g_strdup("pca-unspecified");
63
}
64
+
65
+ qdev_init_gpio_out(dev, s->gpio, k->pin_count);
66
}
67
68
static Property pca955x_properties[] = {
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
1
Move the NS TBFLAG down from bit 19 to bit 6, which has not
1
From: Richard Henderson <richard.henderson@linaro.org>
2
been used since commit c1e3781090b9d36c60 in 2015, when we
3
started passing the entire MMU index in the TB flags rather
4
than just a 'privilege level' bit.
5
2
6
This rearrangement is not strictly necessary, but means that
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
we can put M-profile-only bits next to each other rather
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
than scattered across the flag word.
5
Message-id: 20200626033144.790098-2-richard.henderson@linaro.org
9
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20190416125744.27770-17-peter.maydell@linaro.org
13
---
7
---
14
target/arm/cpu.h | 11 ++++++-----
8
target/arm/cpu.h | 10 ++++++++++
15
1 file changed, 6 insertions(+), 5 deletions(-)
9
1 file changed, 10 insertions(+)
16
10
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
13
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
14
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_ANY, BE_DATA, 23, 1)
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_bti(const ARMISARegisters *id)
22
FIELD(TBFLAG_A32, THUMB, 0, 1)
16
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, BT) != 0;
23
FIELD(TBFLAG_A32, VECLEN, 1, 3)
17
}
24
FIELD(TBFLAG_A32, VECSTRIDE, 4, 2)
18
25
+/*
19
+static inline bool isar_feature_aa64_mte_insn_reg(const ARMISARegisters *id)
26
+ * Indicates whether cp register reads and writes by guest code should access
20
+{
27
+ * the secure or nonsecure bank of banked registers; note that this is not
21
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) != 0;
28
+ * the same thing as the current security state of the processor!
22
+}
29
+ */
23
+
30
+FIELD(TBFLAG_A32, NS, 6, 1)
24
+static inline bool isar_feature_aa64_mte(const ARMISARegisters *id)
31
FIELD(TBFLAG_A32, VFPEN, 7, 1)
25
+{
32
FIELD(TBFLAG_A32, CONDEXEC, 8, 8)
26
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) >= 2;
33
FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
27
+}
34
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
28
+
35
* checks on the other bits at runtime
29
static inline bool isar_feature_aa64_pmu_8_1(const ARMISARegisters *id)
36
*/
30
{
37
FIELD(TBFLAG_A32, XSCALE_CPAR, 17, 2)
31
return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, PMUVER) >= 4 &&
38
-/* Indicates whether cp register reads and writes by guest code should access
39
- * the secure or nonsecure bank of banked registers; note that this is not
40
- * the same thing as the current security state of the processor!
41
- */
42
-FIELD(TBFLAG_A32, NS, 19, 1)
43
/* For M profile only, Handler (ie not Thread) mode */
44
FIELD(TBFLAG_A32, HANDLER, 21, 1)
45
/* For M profile only, whether we should generate stack-limit checks */
46
--
32
--
47
2.20.1
33
2.20.1
48
34
49
35
diff view generated by jsdifflib
1
The TailChain() pseudocode specifies that a tail chaining
1
From: Richard Henderson <richard.henderson@linaro.org>
2
exception should sanitize the excReturn all-ones bits and
3
(if there is no FPU) the excReturn FType bits; we weren't
4
doing this.
5
2
3
Protect reads of aa64 id registers with ARM_CP_STATE_AA64.
4
Use this as a simpler test than arm_el_is_aa64, since EL3
5
cannot change mode.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-3-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20190416125744.27770-14-peter.maydell@linaro.org
9
---
11
---
10
target/arm/helper.c | 8 ++++++++
12
target/arm/helper.c | 15 ++++++++-------
11
1 file changed, 8 insertions(+)
13
1 file changed, 8 insertions(+), 7 deletions(-)
12
14
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
16
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ static void v7m_exception_taken(ARMCPU *cpu, uint32_t lr, bool dotailchain,
19
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
18
qemu_log_mask(CPU_LOG_INT, "...taking pending %s exception %d\n",
20
uint32_t valid_mask = 0x3fff;
19
targets_secure ? "secure" : "nonsecure", exc);
21
ARMCPU *cpu = env_archcpu(env);
20
22
21
+ if (dotailchain) {
23
- if (arm_el_is_aa64(env, 3)) {
22
+ /* Sanitize LR FType and PREFIX bits */
24
+ if (ri->state == ARM_CP_STATE_AA64) {
23
+ if (!arm_feature(env, ARM_FEATURE_VFP)) {
25
value |= SCR_FW | SCR_AW; /* these two bits are RES1. */
24
+ lr |= R_V7M_EXCRET_FTYPE_MASK;
26
valid_mask &= ~SCR_NET;
27
+
28
+ if (cpu_isar_feature(aa64_lor, cpu)) {
29
+ valid_mask |= SCR_TLOR;
25
+ }
30
+ }
26
+ lr = deposit32(lr, 24, 8, 0xff);
31
+ if (cpu_isar_feature(aa64_pauth, cpu)) {
27
+ }
32
+ valid_mask |= SCR_API | SCR_APK;
28
+
33
+ }
29
if (arm_feature(env, ARM_FEATURE_V8)) {
34
} else {
30
if (arm_feature(env, ARM_FEATURE_M_SECURITY) &&
35
valid_mask &= ~(SCR_RW | SCR_ST);
31
(lr & R_V7M_EXCRET_S_MASK)) {
36
}
37
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
38
valid_mask &= ~SCR_SMD;
39
}
40
}
41
- if (cpu_isar_feature(aa64_lor, cpu)) {
42
- valid_mask |= SCR_TLOR;
43
- }
44
- if (cpu_isar_feature(aa64_pauth, cpu)) {
45
- valid_mask |= SCR_API | SCR_APK;
46
- }
47
48
/* Clear all-context RES0 bits. */
49
value &= valid_mask;
32
--
50
--
33
2.20.1
51
2.20.1
34
52
35
53
diff view generated by jsdifflib
1
Handle floating point registers in exception return.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
This corresponds to pseudocode functions ValidateExceptionReturn(),
3
ExceptionReturn(), PopStack() and ConsumeExcStackFrame().
4
2
3
This does not attempt to rectify all of the res0 bits, but does
4
clear the mte bits when not enabled. Since there is no high-part
5
mapping of SCTLR, aa32 mode cannot write to these bits.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-4-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20190416125744.27770-16-peter.maydell@linaro.org
8
---
11
---
9
target/arm/helper.c | 142 +++++++++++++++++++++++++++++++++++++++++++-
12
target/arm/helper.c | 23 +++++++++++++++++------
10
1 file changed, 141 insertions(+), 1 deletion(-)
13
1 file changed, 17 insertions(+), 6 deletions(-)
11
14
12
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
15
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
16
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
19
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
17
bool rettobase = false;
20
{
18
bool exc_secure = false;
21
ARMCPU *cpu = env_archcpu(env);
19
bool return_to_secure;
22
20
+ bool ftype;
23
+ if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
21
+ bool restore_s16_s31;
24
+ /* M bit is RAZ/WI for PMSA with no MPU implemented */
22
25
+ value &= ~SCTLR_M;
23
/* If we're not in Handler mode then jumps to magic exception-exit
24
* addresses don't have magic behaviour. However for the v8M
25
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
26
excret);
27
}
28
29
+ ftype = excret & R_V7M_EXCRET_FTYPE_MASK;
30
+
31
+ if (!arm_feature(env, ARM_FEATURE_VFP) && !ftype) {
32
+ qemu_log_mask(LOG_GUEST_ERROR, "M profile: zero FTYPE in exception "
33
+ "exit PC value 0x%" PRIx32 " is UNPREDICTABLE "
34
+ "if FPU not present\n",
35
+ excret);
36
+ ftype = true;
37
+ }
26
+ }
38
+
27
+
39
if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
28
+ /* ??? Lots of these bits are not implemented. */
40
/* EXC_RETURN.ES validation check (R_SMFL). We must do this before
29
+
41
* we pick which FAULTMASK to clear.
30
+ if (ri->state == ARM_CP_STATE_AA64 && !cpu_isar_feature(aa64_mte, cpu)) {
42
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
31
+ if (ri->opc1 == 6) { /* SCTLR_EL3 */
43
*/
32
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF | SCTLR_ATA);
44
write_v7m_control_spsel_for_secstate(env, return_to_sp_process, exc_secure);
45
46
+ /*
47
+ * Clear scratch FP values left in caller saved registers; this
48
+ * must happen before any kind of tail chaining.
49
+ */
50
+ if ((env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_CLRONRET_MASK) &&
51
+ (env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK)) {
52
+ if (env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_LSPACT_MASK) {
53
+ env->v7m.sfsr |= R_V7M_SFSR_LSERR_MASK;
54
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
55
+ qemu_log_mask(CPU_LOG_INT, "...taking SecureFault on existing "
56
+ "stackframe: error during lazy state deactivation\n");
57
+ v7m_exception_taken(cpu, excret, true, false);
58
+ return;
59
+ } else {
33
+ } else {
60
+ /* Clear s0..s15 and FPSCR */
34
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF0 | SCTLR_TCF |
61
+ int i;
35
+ SCTLR_ATA0 | SCTLR_ATA);
62
+
63
+ for (i = 0; i < 16; i += 2) {
64
+ *aa32_vfp_dreg(env, i / 2) = 0;
65
+ }
66
+ vfp_set_fpscr(env, 0);
67
+ }
36
+ }
68
+ }
37
+ }
69
+
38
+
70
if (sfault) {
39
if (raw_read(env, ri) == value) {
71
env->v7m.sfsr |= R_V7M_SFSR_INVER_MASK;
40
/* Skip the TLB flush if nothing actually changed; Linux likes
72
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
41
* to do a lot of pointless SCTLR writes.
73
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
42
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
74
}
43
return;
75
}
44
}
76
45
77
+ if (!ftype) {
46
- if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
78
+ /* FP present and we need to handle it */
47
- /* M bit is RAZ/WI for PMSA with no MPU implemented */
79
+ if (!return_to_secure &&
48
- value &= ~SCTLR_M;
80
+ (env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_LSPACT_MASK)) {
49
- }
81
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
50
-
82
+ env->v7m.sfsr |= R_V7M_SFSR_LSERR_MASK;
51
raw_write(env, ri, value);
83
+ qemu_log_mask(CPU_LOG_INT,
52
- /* ??? Lots of these bits are not implemented. */
84
+ "...taking SecureFault on existing stackframe: "
85
+ "Secure LSPACT set but exception return is "
86
+ "not to secure state\n");
87
+ v7m_exception_taken(cpu, excret, true, false);
88
+ return;
89
+ }
90
+
53
+
91
+ restore_s16_s31 = return_to_secure &&
54
/* This may enable/disable the MMU, so do a TLB flush. */
92
+ (env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_TS_MASK);
55
tlb_flush(CPU(cpu));
93
+
56
94
+ if (env->v7m.fpccr[return_to_secure] & R_V7M_FPCCR_LSPACT_MASK) {
95
+ /* State in FPU is still valid, just clear LSPACT */
96
+ env->v7m.fpccr[return_to_secure] &= ~R_V7M_FPCCR_LSPACT_MASK;
97
+ } else {
98
+ int i;
99
+ uint32_t fpscr;
100
+ bool cpacr_pass, nsacr_pass;
101
+
102
+ cpacr_pass = v7m_cpacr_pass(env, return_to_secure,
103
+ return_to_priv);
104
+ nsacr_pass = return_to_secure ||
105
+ extract32(env->v7m.nsacr, 10, 1);
106
+
107
+ if (!cpacr_pass) {
108
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE,
109
+ return_to_secure);
110
+ env->v7m.cfsr[return_to_secure] |= R_V7M_CFSR_NOCP_MASK;
111
+ qemu_log_mask(CPU_LOG_INT,
112
+ "...taking UsageFault on existing "
113
+ "stackframe: CPACR.CP10 prevents unstacking "
114
+ "FP regs\n");
115
+ v7m_exception_taken(cpu, excret, true, false);
116
+ return;
117
+ } else if (!nsacr_pass) {
118
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, true);
119
+ env->v7m.cfsr[M_REG_S] |= R_V7M_CFSR_INVPC_MASK;
120
+ qemu_log_mask(CPU_LOG_INT,
121
+ "...taking Secure UsageFault on existing "
122
+ "stackframe: NSACR.CP10 prevents unstacking "
123
+ "FP regs\n");
124
+ v7m_exception_taken(cpu, excret, true, false);
125
+ return;
126
+ }
127
+
128
+ for (i = 0; i < (restore_s16_s31 ? 32 : 16); i += 2) {
129
+ uint32_t slo, shi;
130
+ uint64_t dn;
131
+ uint32_t faddr = frameptr + 0x20 + 4 * i;
132
+
133
+ if (i >= 16) {
134
+ faddr += 8; /* Skip the slot for the FPSCR */
135
+ }
136
+
137
+ pop_ok = pop_ok &&
138
+ v7m_stack_read(cpu, &slo, faddr, mmu_idx) &&
139
+ v7m_stack_read(cpu, &shi, faddr + 4, mmu_idx);
140
+
141
+ if (!pop_ok) {
142
+ break;
143
+ }
144
+
145
+ dn = (uint64_t)shi << 32 | slo;
146
+ *aa32_vfp_dreg(env, i / 2) = dn;
147
+ }
148
+ pop_ok = pop_ok &&
149
+ v7m_stack_read(cpu, &fpscr, frameptr + 0x60, mmu_idx);
150
+ if (pop_ok) {
151
+ vfp_set_fpscr(env, fpscr);
152
+ }
153
+ if (!pop_ok) {
154
+ /*
155
+ * These regs are 0 if security extension present;
156
+ * otherwise merely UNKNOWN. We zero always.
157
+ */
158
+ for (i = 0; i < (restore_s16_s31 ? 32 : 16); i += 2) {
159
+ *aa32_vfp_dreg(env, i / 2) = 0;
160
+ }
161
+ vfp_set_fpscr(env, 0);
162
+ }
163
+ }
164
+ }
165
+ env->v7m.control[M_REG_S] = FIELD_DP32(env->v7m.control[M_REG_S],
166
+ V7M_CONTROL, FPCA, !ftype);
167
+
168
/* Commit to consuming the stack frame */
169
frameptr += 0x20;
170
+ if (!ftype) {
171
+ frameptr += 0x48;
172
+ if (restore_s16_s31) {
173
+ frameptr += 0x40;
174
+ }
175
+ }
176
/* Undo stack alignment (the SPREALIGN bit indicates that the original
177
* pre-exception SP was not 8-aligned and we added a padding word to
178
* align it, so we undo this by ORing in the bit that increases it
179
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
180
*frame_sp_p = frameptr;
181
}
182
/* This xpsr_write() will invalidate frame_sp_p as it may switch stack */
183
- xpsr_write(env, xpsr, ~XPSR_SPREALIGN);
184
+ xpsr_write(env, xpsr, ~(XPSR_SPREALIGN | XPSR_SFPA));
185
+
186
+ if (env->v7m.secure) {
187
+ bool sfpa = xpsr & XPSR_SFPA;
188
+
189
+ env->v7m.control[M_REG_S] = FIELD_DP32(env->v7m.control[M_REG_S],
190
+ V7M_CONTROL, SFPA, sfpa);
191
+ }
192
193
/* The restored xPSR exception field will be zero if we're
194
* resuming in Thread mode. If that doesn't match what the
195
--
57
--
196
2.20.1
58
2.20.1
197
59
198
60
diff view generated by jsdifflib
1
Handle floating point registers in exception entry.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
This corresponds to the FP-specific parts of the pseudocode
3
functions ActivateException() and PushStack().
4
2
5
We defer the code corresponding to UpdateFPCCR() to a later patch.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-5-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20190416125744.27770-11-peter.maydell@linaro.org
10
---
7
---
11
target/arm/helper.c | 98 +++++++++++++++++++++++++++++++++++++++++++--
8
target/arm/helper.c | 14 +++++++++++---
12
1 file changed, 95 insertions(+), 3 deletions(-)
9
1 file changed, 11 insertions(+), 3 deletions(-)
13
10
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
13
--- a/target/arm/helper.c
17
+++ b/target/arm/helper.c
14
+++ b/target/arm/helper.c
18
@@ -XXX,XX +XXX,XX @@ static void v7m_exception_taken(ARMCPU *cpu, uint32_t lr, bool dotailchain,
15
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
19
switch_v7m_security_state(env, targets_secure);
16
if (cpu_isar_feature(aa64_pauth, cpu)) {
20
write_v7m_control_spsel(env, 0);
17
valid_mask |= SCR_API | SCR_APK;
21
arm_clear_exclusive(env);
22
+ /* Clear SFPA and FPCA (has no effect if no FPU) */
23
+ env->v7m.control[M_REG_S] &=
24
+ ~(R_V7M_CONTROL_FPCA_MASK | R_V7M_CONTROL_SFPA_MASK);
25
/* Clear IT bits */
26
env->condexec_bits = 0;
27
env->regs[14] = lr;
28
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
29
uint32_t xpsr = xpsr_read(env);
30
uint32_t frameptr = env->regs[13];
31
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
32
+ uint32_t framesize;
33
+ bool nsacr_cp10 = extract32(env->v7m.nsacr, 10, 1);
34
+
35
+ if ((env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK) &&
36
+ (env->v7m.secure || nsacr_cp10)) {
37
+ if (env->v7m.secure &&
38
+ env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_TS_MASK) {
39
+ framesize = 0xa8;
40
+ } else {
41
+ framesize = 0x68;
42
+ }
43
+ } else {
44
+ framesize = 0x20;
45
+ }
46
47
/* Align stack pointer if the guest wants that */
48
if ((frameptr & 4) &&
49
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
50
xpsr |= XPSR_SPREALIGN;
51
}
52
53
- frameptr -= 0x20;
54
+ xpsr &= ~XPSR_SFPA;
55
+ if (env->v7m.secure &&
56
+ (env->v7m.control[M_REG_S] & R_V7M_CONTROL_SFPA_MASK)) {
57
+ xpsr |= XPSR_SFPA;
58
+ }
59
+
60
+ frameptr -= framesize;
61
62
if (arm_feature(env, ARM_FEATURE_V8)) {
63
uint32_t limit = v7m_sp_limit(env);
64
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
65
v7m_stack_write(cpu, frameptr + 24, env->regs[15], mmu_idx, false) &&
66
v7m_stack_write(cpu, frameptr + 28, xpsr, mmu_idx, false);
67
68
+ if (env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK) {
69
+ /* FPU is active, try to save its registers */
70
+ bool fpccr_s = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_S_MASK;
71
+ bool lspact = env->v7m.fpccr[fpccr_s] & R_V7M_FPCCR_LSPACT_MASK;
72
+
73
+ if (lspact && arm_feature(env, ARM_FEATURE_M_SECURITY)) {
74
+ qemu_log_mask(CPU_LOG_INT,
75
+ "...SecureFault because LSPACT and FPCA both set\n");
76
+ env->v7m.sfsr |= R_V7M_SFSR_LSERR_MASK;
77
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
78
+ } else if (!env->v7m.secure && !nsacr_cp10) {
79
+ qemu_log_mask(CPU_LOG_INT,
80
+ "...Secure UsageFault with CFSR.NOCP because "
81
+ "NSACR.CP10 prevents stacking FP regs\n");
82
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, M_REG_S);
83
+ env->v7m.cfsr[M_REG_S] |= R_V7M_CFSR_NOCP_MASK;
84
+ } else {
85
+ if (!(env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_LSPEN_MASK)) {
86
+ /* Lazy stacking disabled, save registers now */
87
+ int i;
88
+ bool cpacr_pass = v7m_cpacr_pass(env, env->v7m.secure,
89
+ arm_current_el(env) != 0);
90
+
91
+ if (stacked_ok && !cpacr_pass) {
92
+ /*
93
+ * Take UsageFault if CPACR forbids access. The pseudocode
94
+ * here does a full CheckCPEnabled() but we know the NSACR
95
+ * check can never fail as we have already handled that.
96
+ */
97
+ qemu_log_mask(CPU_LOG_INT,
98
+ "...UsageFault with CFSR.NOCP because "
99
+ "CPACR.CP10 prevents stacking FP regs\n");
100
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE,
101
+ env->v7m.secure);
102
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_NOCP_MASK;
103
+ stacked_ok = false;
104
+ }
105
+
106
+ for (i = 0; i < ((framesize == 0xa8) ? 32 : 16); i += 2) {
107
+ uint64_t dn = *aa32_vfp_dreg(env, i / 2);
108
+ uint32_t faddr = frameptr + 0x20 + 4 * i;
109
+ uint32_t slo = extract64(dn, 0, 32);
110
+ uint32_t shi = extract64(dn, 32, 32);
111
+
112
+ if (i >= 16) {
113
+ faddr += 8; /* skip the slot for the FPSCR */
114
+ }
115
+ stacked_ok = stacked_ok &&
116
+ v7m_stack_write(cpu, faddr, slo, mmu_idx, false) &&
117
+ v7m_stack_write(cpu, faddr + 4, shi, mmu_idx, false);
118
+ }
119
+ stacked_ok = stacked_ok &&
120
+ v7m_stack_write(cpu, frameptr + 0x60,
121
+ vfp_get_fpscr(env), mmu_idx, false);
122
+ if (cpacr_pass) {
123
+ for (i = 0; i < ((framesize == 0xa8) ? 32 : 16); i += 2) {
124
+ *aa32_vfp_dreg(env, i / 2) = 0;
125
+ }
126
+ vfp_set_fpscr(env, 0);
127
+ }
128
+ } else {
129
+ /* Lazy stacking enabled, save necessary info to stack later */
130
+ /* TODO : equivalent of UpdateFPCCR() pseudocode */
131
+ }
132
+ }
133
+ }
134
+
135
/*
136
* If we broke a stack limit then SP was already updated earlier;
137
* otherwise we update SP regardless of whether any of the stack
138
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
139
140
if (arm_feature(env, ARM_FEATURE_V8)) {
141
lr = R_V7M_EXCRET_RES1_MASK |
142
- R_V7M_EXCRET_DCRS_MASK |
143
- R_V7M_EXCRET_FTYPE_MASK;
144
+ R_V7M_EXCRET_DCRS_MASK;
145
/* The S bit indicates whether we should return to Secure
146
* or NonSecure (ie our current state).
147
* The ES bit indicates whether we're taking this exception
148
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
149
if (env->v7m.secure) {
150
lr |= R_V7M_EXCRET_S_MASK;
151
}
18
}
152
+ if (!(env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK)) {
19
+ if (cpu_isar_feature(aa64_mte, cpu)) {
153
+ lr |= R_V7M_EXCRET_FTYPE_MASK;
20
+ valid_mask |= SCR_ATA;
154
+ }
21
+ }
155
} else {
22
} else {
156
lr = R_V7M_EXCRET_RES1_MASK |
23
valid_mask &= ~(SCR_RW | SCR_ST);
157
R_V7M_EXCRET_S_MASK |
24
}
25
@@ -XXX,XX +XXX,XX @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
26
if (cpu_isar_feature(aa64_pauth, cpu)) {
27
valid_mask |= HCR_API | HCR_APK;
28
}
29
+ if (cpu_isar_feature(aa64_mte, cpu)) {
30
+ valid_mask |= HCR_ATA | HCR_DCT | HCR_TID5;
31
+ }
32
}
33
34
/* Clear RES0 bits. */
35
value &= valid_mask;
36
37
- /* These bits change the MMU setup:
38
+ /*
39
+ * These bits change the MMU setup:
40
* HCR_VM enables stage 2 translation
41
* HCR_PTW forbids certain page-table setups
42
- * HCR_DC Disables stage1 and enables stage2 translation
43
+ * HCR_DC disables stage1 and enables stage2 translation
44
+ * HCR_DCT enables tagging on (disabled) stage1 translation
45
*/
46
- if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC)) {
47
+ if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC | HCR_DCT)) {
48
tlb_flush(CPU(cpu));
49
}
50
env->cp15.hcr_el2 = value;
158
--
51
--
159
2.20.1
52
2.20.1
160
53
161
54
diff view generated by jsdifflib
1
The M-profile FPCCR.ASPEN bit indicates that automatic floating-point
1
From: Richard Henderson <richard.henderson@linaro.org>
2
context preservation is enabled. Before executing any floating-point
3
instruction, if FPCCR.ASPEN is set and the CONTROL FPCA/SFPA bits
4
indicate that there is no active floating point context then we
5
must create a new context (by initializing FPSCR and setting
6
FPCA/SFPA to indicate that the context is now active). In the
7
pseudocode this is handled by ExecuteFPCheck().
8
2
9
Implement this with a new TB flag which tracks whether we
3
Emphasize that the is_jmp option exits to the main loop.
10
need to create a new FP context.
11
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-6-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20190416125744.27770-20-peter.maydell@linaro.org
15
---
9
---
16
target/arm/cpu.h | 2 ++
10
target/arm/translate.h | 14 ++++++++------
17
target/arm/translate.h | 1 +
11
target/arm/translate-a64.c | 8 ++++----
18
target/arm/helper.c | 13 +++++++++++++
12
target/arm/translate-vfp.inc.c | 4 ++--
19
target/arm/translate.c | 29 +++++++++++++++++++++++++++++
13
target/arm/translate.c | 12 ++++++------
20
4 files changed, 45 insertions(+)
14
4 files changed, 20 insertions(+), 18 deletions(-)
21
15
22
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/cpu.h
25
+++ b/target/arm/cpu.h
26
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, NS, 6, 1)
27
FIELD(TBFLAG_A32, VFPEN, 7, 1)
28
FIELD(TBFLAG_A32, CONDEXEC, 8, 8)
29
FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
30
+/* For M profile only, set if we must create a new FP context */
31
+FIELD(TBFLAG_A32, NEW_FP_CTXT_NEEDED, 19, 1)
32
/* For M profile only, set if FPCCR.S does not match current security state */
33
FIELD(TBFLAG_A32, FPCCR_S_WRONG, 20, 1)
34
/* For M profile only, Handler (ie not Thread) mode */
35
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
36
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/translate.h
18
--- a/target/arm/translate.h
38
+++ b/target/arm/translate.h
19
+++ b/target/arm/translate.h
39
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
40
bool v8m_secure; /* true if v8M and we're in Secure mode */
21
41
bool v8m_stackcheck; /* true if we need to perform v8M stack limit checks */
22
/* is_jmp field values */
42
bool v8m_fpccr_s_wrong; /* true if v8M FPCCR.S != v8m_secure */
23
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
43
+ bool v7m_new_fp_ctxt_needed; /* ASPEN set but no active FP context */
24
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
44
/* Immediate value in AArch32 SVC insn; must be set if is_jmp == DISAS_SWI
25
+/* CPU state was modified dynamically; exit to main loop for interrupts. */
45
* so that top level loop can generate correct syndrome information.
26
+#define DISAS_UPDATE_EXIT DISAS_TARGET_1
46
*/
27
/* These instructions trap after executing, so the A32/T32 decoder must
47
diff --git a/target/arm/helper.c b/target/arm/helper.c
28
* defer them until after the conditional execution state has been updated.
29
* WFI also needs special handling when single-stepping.
30
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
31
* custom end-of-TB code)
32
*/
33
#define DISAS_BX_EXCRET DISAS_TARGET_8
34
-/* For instructions which want an immediate exit to the main loop,
35
- * as opposed to attempting to use lookup_and_goto_ptr. Unlike
36
- * DISAS_UPDATE this doesn't write the PC on exiting the translation
37
- * loop so you need to ensure something (gen_a64_set_pc_im or runtime
38
- * helper) has done so before we reach return from cpu_tb_exec.
39
+/*
40
+ * For instructions which want an immediate exit to the main loop, as opposed
41
+ * to attempting to use lookup_and_goto_ptr. Unlike DISAS_UPDATE_EXIT, this
42
+ * doesn't write the PC on exiting the translation loop so you need to ensure
43
+ * something (gen_a64_set_pc_im or runtime helper) has done so before we reach
44
+ * return from cpu_tb_exec.
45
*/
46
#define DISAS_EXIT DISAS_TARGET_9
47
48
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
48
index XXXXXXX..XXXXXXX 100644
49
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/helper.c
50
--- a/target/arm/translate-a64.c
50
+++ b/target/arm/helper.c
51
+++ b/target/arm/translate-a64.c
51
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
52
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
52
flags = FIELD_DP32(flags, TBFLAG_A32, FPCCR_S_WRONG, 1);
53
gen_helper_msr_i_daifclear(cpu_env, t1);
54
tcg_temp_free_i32(t1);
55
/* For DAIFClear, exit the cpu loop to re-evaluate pending IRQs. */
56
- s->base.is_jmp = DISAS_UPDATE;
57
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
58
break;
59
60
default:
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
63
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
64
/* I/O operations must end the TB here (whether read or write) */
65
- s->base.is_jmp = DISAS_UPDATE;
66
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
53
}
67
}
54
68
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
55
+ if (arm_feature(env, ARM_FEATURE_M) &&
69
/*
56
+ (env->v7m.fpccr[env->v7m.secure] & R_V7M_FPCCR_ASPEN_MASK) &&
70
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
57
+ (!(env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK) ||
71
* but allow this to be suppressed by the register definition
58
+ (env->v7m.secure &&
72
* (usually only necessary to work around guest bugs).
59
+ !(env->v7m.control[M_REG_S] & R_V7M_CONTROL_SFPA_MASK)))) {
73
*/
60
+ /*
74
- s->base.is_jmp = DISAS_UPDATE;
61
+ * ASPEN is set, but FPCA/SFPA indicate that there is no active
75
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
62
+ * FP context; we must create a new FP context before executing
76
}
63
+ * any FP insn.
77
}
64
+ */
78
65
+ flags = FIELD_DP32(flags, TBFLAG_A32, NEW_FP_CTXT_NEEDED, 1);
79
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
66
+ }
80
gen_goto_tb(dc, 1, dc->base.pc_next);
67
+
81
break;
68
*pflags = flags;
82
default:
69
*cs_base = 0;
83
- case DISAS_UPDATE:
84
+ case DISAS_UPDATE_EXIT:
85
gen_a64_set_pc_im(dc->base.pc_next);
86
/* fall through */
87
case DISAS_EXIT:
88
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
89
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate-vfp.inc.c
91
+++ b/target/arm/translate-vfp.inc.c
92
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
93
* this to be the last insn in the TB).
94
*/
95
if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
96
- s->base.is_jmp = DISAS_UPDATE;
97
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
98
gen_io_start();
99
}
100
gen_helper_v7m_preserve_fp_state(cpu_env);
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
102
tcg_temp_free_i32(fptr);
103
104
/* End the TB, because we have updated FP control bits */
105
- s->base.is_jmp = DISAS_UPDATE;
106
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
107
return true;
70
}
108
}
71
diff --git a/target/arm/translate.c b/target/arm/translate.c
109
diff --git a/target/arm/translate.c b/target/arm/translate.c
72
index XXXXXXX..XXXXXXX 100644
110
index XXXXXXX..XXXXXXX 100644
73
--- a/target/arm/translate.c
111
--- a/target/arm/translate.c
74
+++ b/target/arm/translate.c
112
+++ b/target/arm/translate.c
75
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
113
@@ -XXX,XX +XXX,XX @@ static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
76
/* Don't need to do this for any further FP insns in this TB */
114
tcg_temp_free_i32(tcg_tgtmode);
77
s->v8m_fpccr_s_wrong = false;
115
tcg_temp_free_i32(tcg_regno);
78
}
116
tcg_temp_free_i32(tcg_reg);
79
+
117
- s->base.is_jmp = DISAS_UPDATE;
80
+ if (s->v7m_new_fp_ctxt_needed) {
118
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
81
+ /*
119
}
82
+ * Create new FP context by updating CONTROL.FPCA, CONTROL.SFPA
120
83
+ * and the FPSCR.
121
static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
84
+ */
122
@@ -XXX,XX +XXX,XX @@ static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
85
+ TCGv_i32 control, fpscr;
123
tcg_temp_free_i32(tcg_tgtmode);
86
+ uint32_t bits = R_V7M_CONTROL_FPCA_MASK;
124
tcg_temp_free_i32(tcg_regno);
87
+
125
store_reg(s, rn, tcg_reg);
88
+ fpscr = load_cpu_field(v7m.fpdscr[s->v8m_secure]);
126
- s->base.is_jmp = DISAS_UPDATE;
89
+ gen_helper_vfp_set_fpscr(cpu_env, fpscr);
127
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
90
+ tcg_temp_free_i32(fpscr);
128
}
91
+ /*
129
92
+ * We don't need to arrange to end the TB, because the only
130
/* Store value to PC as for an exception return (ie don't
93
+ * parts of FPSCR which we cache in the TB flags are the VECLEN
131
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
94
+ * and VECSTRIDE, and those don't exist for M-profile.
132
tcg_temp_free_i32(tmp);
95
+ */
96
+
97
+ if (s->v8m_secure) {
98
+ bits |= R_V7M_CONTROL_SFPA_MASK;
99
+ }
100
+ control = load_cpu_field(v7m.control[M_REG_S]);
101
+ tcg_gen_ori_i32(control, control, bits);
102
+ store_cpu_field(control, v7m.control[M_REG_S]);
103
+ /* Don't need to do this for any further FP insns in this TB */
104
+ s->v7m_new_fp_ctxt_needed = false;
105
+ }
106
}
133
}
107
134
tcg_temp_free_i32(addr);
108
if (extract32(insn, 28, 4) == 0xf) {
135
- s->base.is_jmp = DISAS_UPDATE;
109
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
136
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
110
regime_is_secure(env, dc->mmu_idx);
137
}
111
dc->v8m_stackcheck = FIELD_EX32(tb_flags, TBFLAG_A32, STACKCHECK);
138
112
dc->v8m_fpccr_s_wrong = FIELD_EX32(tb_flags, TBFLAG_A32, FPCCR_S_WRONG);
139
/* Generate a label used for skipping this instruction */
113
+ dc->v7m_new_fp_ctxt_needed =
140
@@ -XXX,XX +XXX,XX @@ static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
114
+ FIELD_EX32(tb_flags, TBFLAG_A32, NEW_FP_CTXT_NEEDED);
141
}
115
dc->cp_regs = cpu->cp_regs;
142
if (a->E != (s->be_data == MO_BE)) {
116
dc->features = env->features;
143
gen_helper_setend(cpu_env);
117
144
- s->base.is_jmp = DISAS_UPDATE;
145
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
146
}
147
return true;
148
}
149
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
150
break;
151
case DISAS_NEXT:
152
case DISAS_TOO_MANY:
153
- case DISAS_UPDATE:
154
+ case DISAS_UPDATE_EXIT:
155
gen_set_pc_im(dc, dc->base.pc_next);
156
/* fall through */
157
default:
158
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
159
case DISAS_JUMP:
160
gen_goto_ptr();
161
break;
162
- case DISAS_UPDATE:
163
+ case DISAS_UPDATE_EXIT:
164
gen_set_pc_im(dc, dc->base.pc_next);
165
/* fall through */
166
default:
118
--
167
--
119
2.20.1
168
2.20.1
120
169
121
170
diff view generated by jsdifflib
1
Correct the decode of the M-profile "coprocessor and
1
From: Richard Henderson <richard.henderson@linaro.org>
2
floating-point instructions" space:
3
* op0 == 0b11 is always unallocated
4
* if the CPU has an FPU then all insns with op1 == 0b101
5
are floating point and go to disas_vfp_insn()
6
2
7
For the moment we leave VLLDM and VLSTM as NOPs; in
3
Add an option that writes back the PC, like DISAS_UPDATE_EXIT,
8
a later commit we will fill in the proper implementation
4
but does not exit back to the main loop.
9
for the case where an FPU is present.
10
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-7-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20190416125744.27770-7-peter.maydell@linaro.org
14
---
10
---
15
target/arm/translate.c | 26 ++++++++++++++++++++++----
11
target/arm/translate.h | 2 ++
16
1 file changed, 22 insertions(+), 4 deletions(-)
12
target/arm/translate-a64.c | 3 +++
13
target/arm/translate.c | 4 ++++
14
3 files changed, 9 insertions(+)
17
15
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.h
19
+++ b/target/arm/translate.h
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
21
* return from cpu_tb_exec.
22
*/
23
#define DISAS_EXIT DISAS_TARGET_9
24
+/* CPU state was modified dynamically; no need to exit, but do not chain. */
25
+#define DISAS_UPDATE_NOCHAIN DISAS_TARGET_10
26
27
#ifdef TARGET_AARCH64
28
void a64_translate_init(void);
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
30
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/translate-a64.c
32
+++ b/target/arm/translate-a64.c
33
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
34
case DISAS_EXIT:
35
tcg_gen_exit_tb(NULL, 0);
36
break;
37
+ case DISAS_UPDATE_NOCHAIN:
38
+ gen_a64_set_pc_im(dc->base.pc_next);
39
+ /* fall through */
40
case DISAS_JUMP:
41
tcg_gen_lookup_and_goto_ptr();
42
break;
18
diff --git a/target/arm/translate.c b/target/arm/translate.c
43
diff --git a/target/arm/translate.c b/target/arm/translate.c
19
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/translate.c
45
--- a/target/arm/translate.c
21
+++ b/target/arm/translate.c
46
+++ b/target/arm/translate.c
22
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
47
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
23
case 6: case 7: case 14: case 15:
48
case DISAS_NEXT:
24
/* Coprocessor. */
49
case DISAS_TOO_MANY:
25
if (arm_dc_feature(s, ARM_FEATURE_M)) {
50
case DISAS_UPDATE_EXIT:
26
- /* We don't currently implement M profile FP support,
51
+ case DISAS_UPDATE_NOCHAIN:
27
- * so this entire space should give a NOCP fault, with
52
gen_set_pc_im(dc, dc->base.pc_next);
28
- * the exception of the v8M VLLDM and VLSTM insns, which
53
/* fall through */
29
- * must be NOPs in Secure state and UNDEF in Nonsecure state.
54
default:
30
+ /* 0b111x_11xx_xxxx_xxxx_xxxx_xxxx_xxxx_xxxx */
55
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
31
+ if (extract32(insn, 24, 2) == 3) {
56
case DISAS_TOO_MANY:
32
+ goto illegal_op; /* op0 = 0b11 : unallocated */
57
gen_goto_tb(dc, 1, dc->base.pc_next);
33
+ }
58
break;
34
+
59
+ case DISAS_UPDATE_NOCHAIN:
35
+ /*
60
+ gen_set_pc_im(dc, dc->base.pc_next);
36
+ * Decode VLLDM and VLSTM first: these are nonstandard because:
61
+ /* fall through */
37
+ * * if there is no FPU then these insns must NOP in
62
case DISAS_JUMP:
38
+ * Secure state and UNDEF in Nonsecure state
63
gen_goto_ptr();
39
+ * * if there is an FPU then these insns do not have
64
break;
40
+ * the usual behaviour that disas_vfp_insn() provides of
41
+ * being controlled by CPACR/NSACR enable bits or the
42
+ * lazy-stacking logic.
43
*/
44
if (arm_dc_feature(s, ARM_FEATURE_V8) &&
45
(insn & 0xffa00f00) == 0xec200a00) {
46
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
47
/* Just NOP since FP support is not implemented */
48
break;
49
}
50
+ if (arm_dc_feature(s, ARM_FEATURE_VFP) &&
51
+ ((insn >> 8) & 0xe) == 10) {
52
+ /* FP, and the CPU supports it */
53
+ if (disas_vfp_insn(s, insn)) {
54
+ goto illegal_op;
55
+ }
56
+ break;
57
+ }
58
+
59
/* All other insns: NOCP */
60
gen_exception_insn(s, 4, EXCP_NOCP, syn_uncategorized(),
61
default_exception_el(s));
62
--
65
--
63
2.20.1
66
2.20.1
64
67
65
68
diff view generated by jsdifflib
1
The M-profile FPCCR.S bit indicates the security status of
1
From: Richard Henderson <richard.henderson@linaro.org>
2
the floating point context. In the pseudocode ExecuteFPCheck()
2
3
function it is unconditionally set to match the current
3
This is TFSRE0_EL1, TFSR_EL1, TFSR_EL2, TFSR_EL3,
4
security state whenever a floating point instruction is
4
RGSR_EL1, GCR_EL1, GMID_EL1, and PSTATE.TCO.
5
executed.
5
6
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Implement this by adding a new TB flag which tracks whether
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
FPCCR.S is different from the current security state, so
8
Message-id: 20200626033144.790098-8-richard.henderson@linaro.org
9
that we only need to emit the code to update it in the
10
less-common case when it is not already set correctly.
11
12
Note that we will add the handling for the other work done
13
by ExecuteFPCheck() in later commits.
14
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
17
Message-id: 20190416125744.27770-19-peter.maydell@linaro.org
18
---
10
---
19
target/arm/cpu.h | 2 ++
11
target/arm/cpu.h | 4 ++
20
target/arm/translate.h | 1 +
12
target/arm/internals.h | 9 ++++
21
target/arm/helper.c | 5 +++++
13
target/arm/helper.c | 94 ++++++++++++++++++++++++++++++++++++++
22
target/arm/translate.c | 20 ++++++++++++++++++++
14
target/arm/translate-a64.c | 21 +++++++++
23
4 files changed, 28 insertions(+)
15
4 files changed, 128 insertions(+)
24
16
25
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
26
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/cpu.h
19
--- a/target/arm/cpu.h
28
+++ b/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
29
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, NS, 6, 1)
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
30
FIELD(TBFLAG_A32, VFPEN, 7, 1)
22
uint64_t pmccfiltr_el0; /* Performance Monitor Filter Register */
31
FIELD(TBFLAG_A32, CONDEXEC, 8, 8)
23
uint64_t vpidr_el2; /* Virtualization Processor ID Register */
32
FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
24
uint64_t vmpidr_el2; /* Virtualization Multiprocessor ID Register */
33
+/* For M profile only, set if FPCCR.S does not match current security state */
25
+ uint64_t tfsr_el[4]; /* tfsre0_el1 is index 0. */
34
+FIELD(TBFLAG_A32, FPCCR_S_WRONG, 20, 1)
26
+ uint64_t gcr_el1;
35
/* For M profile only, Handler (ie not Thread) mode */
27
+ uint64_t rgsr_el1;
36
FIELD(TBFLAG_A32, HANDLER, 21, 1)
28
} cp15;
37
/* For M profile only, whether we should generate stack-limit checks */
29
38
diff --git a/target/arm/translate.h b/target/arm/translate.h
30
struct {
39
index XXXXXXX..XXXXXXX 100644
31
@@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu);
40
--- a/target/arm/translate.h
32
#define PSTATE_SS (1U << 21)
41
+++ b/target/arm/translate.h
33
#define PSTATE_PAN (1U << 22)
42
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
34
#define PSTATE_UAO (1U << 23)
43
bool v7m_handler_mode;
35
+#define PSTATE_TCO (1U << 25)
44
bool v8m_secure; /* true if v8M and we're in Secure mode */
36
#define PSTATE_V (1U << 28)
45
bool v8m_stackcheck; /* true if we need to perform v8M stack limit checks */
37
#define PSTATE_C (1U << 29)
46
+ bool v8m_fpccr_s_wrong; /* true if v8M FPCCR.S != v8m_secure */
38
#define PSTATE_Z (1U << 30)
47
/* Immediate value in AArch32 SVC insn; must be set if is_jmp == DISAS_SWI
39
diff --git a/target/arm/internals.h b/target/arm/internals.h
48
* so that top level loop can generate correct syndrome information.
40
index XXXXXXX..XXXXXXX 100644
49
*/
41
--- a/target/arm/internals.h
42
+++ b/target/arm/internals.h
43
@@ -XXX,XX +XXX,XX @@ static inline uint32_t aarch64_pstate_valid_mask(const ARMISARegisters *id)
44
if (isar_feature_aa64_uao(id)) {
45
valid |= PSTATE_UAO;
46
}
47
+ if (isar_feature_aa64_mte(id)) {
48
+ valid |= PSTATE_TCO;
49
+ }
50
51
return valid;
52
}
53
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
54
55
#endif /* !CONFIG_USER_ONLY */
56
57
+/*
58
+ * The log2 of the words in the tag block, for GMID_EL1.BS.
59
+ * The is the maximum, 256 bytes, which manipulates 64-bits of tags.
60
+ */
61
+#define GMID_EL1_BS 6
62
+
63
#endif
50
diff --git a/target/arm/helper.c b/target/arm/helper.c
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
51
index XXXXXXX..XXXXXXX 100644
65
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/helper.c
66
--- a/target/arm/helper.c
53
+++ b/target/arm/helper.c
67
+++ b/target/arm/helper.c
54
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
68
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
55
flags = FIELD_DP32(flags, TBFLAG_A32, STACKCHECK, 1);
69
{ K(3, 0, 1, 2, 0), K(3, 4, 1, 2, 0), K(3, 5, 1, 2, 0),
56
}
70
"ZCR_EL1", "ZCR_EL2", "ZCR_EL12", isar_feature_aa64_sve },
57
71
58
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY) &&
72
+ { K(3, 0, 5, 6, 0), K(3, 4, 5, 6, 0), K(3, 5, 5, 6, 0),
59
+ FIELD_EX32(env->v7m.fpccr[M_REG_S], V7M_FPCCR, S) != env->v7m.secure) {
73
+ "TFSR_EL1", "TFSR_EL2", "TFSR_EL12", isar_feature_aa64_mte },
60
+ flags = FIELD_DP32(flags, TBFLAG_A32, FPCCR_S_WRONG, 1);
74
+
61
+ }
75
/* TODO: ARMv8.2-SPE -- PMSCR_EL2 */
62
+
76
/* TODO: ARMv8.4-Trace -- TRFCR_EL2 */
63
*pflags = flags;
77
};
64
*cs_base = 0;
78
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
65
}
79
};
66
diff --git a/target/arm/translate.c b/target/arm/translate.c
80
#endif /*CONFIG_USER_ONLY*/
67
index XXXXXXX..XXXXXXX 100644
81
68
--- a/target/arm/translate.c
82
+static CPAccessResult access_aa64_tid5(CPUARMState *env, const ARMCPRegInfo *ri,
69
+++ b/target/arm/translate.c
83
+ bool isread)
70
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
84
+{
85
+ if ((arm_current_el(env) < 2) && (arm_hcr_el2_eff(env) & HCR_TID5)) {
86
+ return CP_ACCESS_TRAP_EL2;
87
+ }
88
+
89
+ return CP_ACCESS_OK;
90
+}
91
+
92
+static CPAccessResult access_mte(CPUARMState *env, const ARMCPRegInfo *ri,
93
+ bool isread)
94
+{
95
+ int el = arm_current_el(env);
96
+
97
+ if (el < 2 &&
98
+ arm_feature(env, ARM_FEATURE_EL2) &&
99
+ !(arm_hcr_el2_eff(env) & HCR_ATA)) {
100
+ return CP_ACCESS_TRAP_EL2;
101
+ }
102
+ if (el < 3 &&
103
+ arm_feature(env, ARM_FEATURE_EL3) &&
104
+ !(env->cp15.scr_el3 & SCR_ATA)) {
105
+ return CP_ACCESS_TRAP_EL3;
106
+ }
107
+ return CP_ACCESS_OK;
108
+}
109
+
110
+static uint64_t tco_read(CPUARMState *env, const ARMCPRegInfo *ri)
111
+{
112
+ return env->pstate & PSTATE_TCO;
113
+}
114
+
115
+static void tco_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t val)
116
+{
117
+ env->pstate = (env->pstate & ~PSTATE_TCO) | (val & PSTATE_TCO);
118
+}
119
+
120
+static const ARMCPRegInfo mte_reginfo[] = {
121
+ { .name = "TFSRE0_EL1", .state = ARM_CP_STATE_AA64,
122
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 1,
123
+ .access = PL1_RW, .accessfn = access_mte,
124
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[0]) },
125
+ { .name = "TFSR_EL1", .state = ARM_CP_STATE_AA64,
126
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 0,
127
+ .access = PL1_RW, .accessfn = access_mte,
128
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[1]) },
129
+ { .name = "TFSR_EL2", .state = ARM_CP_STATE_AA64,
130
+ .opc0 = 3, .opc1 = 4, .crn = 5, .crm = 6, .opc2 = 0,
131
+ .access = PL2_RW, .accessfn = access_mte,
132
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[2]) },
133
+ { .name = "TFSR_EL3", .state = ARM_CP_STATE_AA64,
134
+ .opc0 = 3, .opc1 = 6, .crn = 5, .crm = 6, .opc2 = 0,
135
+ .access = PL3_RW,
136
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[3]) },
137
+ { .name = "RGSR_EL1", .state = ARM_CP_STATE_AA64,
138
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 5,
139
+ .access = PL1_RW, .accessfn = access_mte,
140
+ .fieldoffset = offsetof(CPUARMState, cp15.rgsr_el1) },
141
+ { .name = "GCR_EL1", .state = ARM_CP_STATE_AA64,
142
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 6,
143
+ .access = PL1_RW, .accessfn = access_mte,
144
+ .fieldoffset = offsetof(CPUARMState, cp15.gcr_el1) },
145
+ { .name = "GMID_EL1", .state = ARM_CP_STATE_AA64,
146
+ .opc0 = 3, .opc1 = 1, .crn = 0, .crm = 0, .opc2 = 4,
147
+ .access = PL1_R, .accessfn = access_aa64_tid5,
148
+ .type = ARM_CP_CONST, .resetvalue = GMID_EL1_BS },
149
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
150
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
151
+ .type = ARM_CP_NO_RAW,
152
+ .access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
153
+ REGINFO_SENTINEL
154
+};
155
+
156
+static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
157
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
158
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
159
+ .type = ARM_CP_CONST, .access = PL0_RW, },
160
+ REGINFO_SENTINEL
161
+};
162
#endif
163
164
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
165
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
71
}
166
}
72
}
167
}
73
168
#endif /*CONFIG_USER_ONLY*/
74
+ if (arm_dc_feature(s, ARM_FEATURE_M)) {
169
+
75
+ /* Handle M-profile lazy FP state mechanics */
170
+ /*
76
+
171
+ * If full MTE is enabled, add all of the system registers.
77
+ /* Update ownership of FP context: set FPCCR.S to match current state */
172
+ * If only "instructions available at EL0" are enabled,
78
+ if (s->v8m_fpccr_s_wrong) {
173
+ * then define only a RAZ/WI version of PSTATE.TCO.
79
+ TCGv_i32 tmp;
174
+ */
80
+
175
+ if (cpu_isar_feature(aa64_mte, cpu)) {
81
+ tmp = load_cpu_field(v7m.fpccr[M_REG_S]);
176
+ define_arm_cp_regs(cpu, mte_reginfo);
82
+ if (s->v8m_secure) {
177
+ } else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
83
+ tcg_gen_ori_i32(tmp, tmp, R_V7M_FPCCR_S_MASK);
178
+ define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
179
+ }
180
#endif
181
182
if (cpu_isar_feature(any_predinv, cpu)) {
183
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
184
index XXXXXXX..XXXXXXX 100644
185
--- a/target/arm/translate-a64.c
186
+++ b/target/arm/translate-a64.c
187
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
188
s->base.is_jmp = DISAS_UPDATE_EXIT;
189
break;
190
191
+ case 0x1c: /* TCO */
192
+ if (dc_isar_feature(aa64_mte, s)) {
193
+ /* Full MTE is enabled -- set the TCO bit as directed. */
194
+ if (crm & 1) {
195
+ set_pstate_bits(PSTATE_TCO);
84
+ } else {
196
+ } else {
85
+ tcg_gen_andi_i32(tmp, tmp, ~R_V7M_FPCCR_S_MASK);
197
+ clear_pstate_bits(PSTATE_TCO);
86
+ }
198
+ }
87
+ store_cpu_field(tmp, v7m.fpccr[M_REG_S]);
199
+ t1 = tcg_const_i32(s->current_el);
88
+ /* Don't need to do this for any further FP insns in this TB */
200
+ gen_helper_rebuild_hflags_a64(cpu_env, t1);
89
+ s->v8m_fpccr_s_wrong = false;
201
+ tcg_temp_free_i32(t1);
202
+ /* Many factors, including TCO, go into MTE_ACTIVE. */
203
+ s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
204
+ } else if (dc_isar_feature(aa64_mte_insn_reg, s)) {
205
+ /* Only "instructions accessible at EL0" -- PSTATE.TCO is WI. */
206
+ s->base.is_jmp = DISAS_NEXT;
207
+ } else {
208
+ goto do_unallocated;
90
+ }
209
+ }
91
+ }
210
+ break;
92
+
211
+
93
if (extract32(insn, 28, 4) == 0xf) {
212
default:
94
/*
213
do_unallocated:
95
* Encodings with T=1 (Thumb) or unconditional (ARM):
214
unallocated_encoding(s);
96
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
97
dc->v8m_secure = arm_feature(env, ARM_FEATURE_M_SECURITY) &&
98
regime_is_secure(env, dc->mmu_idx);
99
dc->v8m_stackcheck = FIELD_EX32(tb_flags, TBFLAG_A32, STACKCHECK);
100
+ dc->v8m_fpccr_s_wrong = FIELD_EX32(tb_flags, TBFLAG_A32, FPCCR_S_WRONG);
101
dc->cp_regs = cpu->cp_regs;
102
dc->features = env->features;
103
104
--
215
--
105
2.20.1
216
2.20.1
106
217
107
218
diff view generated by jsdifflib
1
The M-profile architecture floating point system supports
1
From: Richard Henderson <richard.henderson@linaro.org>
2
lazy FP state preservation, where FP registers are not
3
pushed to the stack when an exception occurs but are instead
4
only saved if and when the first FP instruction in the exception
5
handler is executed. Implement this in QEMU, corresponding
6
to the check of LSPACT in the pseudocode ExecuteFPCheck().
7
2
3
Cache the composite ATA setting.
4
5
Cache when MTE is fully enabled, i.e. access to tags are enabled
6
and tag checks affect the PE. Do this for both the normal context
7
and the UNPRIV context.
8
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20200626033144.790098-9-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20190416125744.27770-24-peter.maydell@linaro.org
11
---
13
---
12
target/arm/cpu.h | 3 ++
14
target/arm/cpu.h | 12 ++++++++----
13
target/arm/helper.h | 2 +
15
target/arm/internals.h | 18 +++++++++++++++++
14
target/arm/translate.h | 1 +
16
target/arm/translate.h | 5 +++++
15
target/arm/helper.c | 112 +++++++++++++++++++++++++++++++++++++++++
17
target/arm/helper.c | 40 ++++++++++++++++++++++++++++++++++++++
16
target/arm/translate.c | 22 ++++++++
18
target/arm/translate-a64.c | 4 ++++
17
5 files changed, 140 insertions(+)
19
5 files changed, 75 insertions(+), 4 deletions(-)
18
20
19
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
20
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.h
23
--- a/target/arm/cpu.h
22
+++ b/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
23
@@ -XXX,XX +XXX,XX @@
25
@@ -XXX,XX +XXX,XX @@ typedef ARMCPU ArchCPU;
24
#define EXCP_NOCP 17 /* v7M NOCP UsageFault */
26
* | | | TBFLAG_A32 | |
25
#define EXCP_INVSTATE 18 /* v7M INVSTATE UsageFault */
27
* | | +-----+----------+ TBFLAG_AM32 |
26
#define EXCP_STKOF 19 /* v8M STKOF UsageFault */
28
* | TBFLAG_ANY | |TBFLAG_M32| |
27
+#define EXCP_LAZYFP 20 /* v7M fault during lazy FP stacking */
29
- * | | +-+----------+--------------|
28
/* NB: add new EXCP_ defines to the array in arm_log_exception() too */
30
- * | | | TBFLAG_A64 |
29
31
- * +--------------+---------+---------------------------+
30
#define ARMV7M_EXCP_RESET 1
32
- * 31 20 15 0
31
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, NS, 6, 1)
33
+ * | +-----------+----------+--------------|
32
FIELD(TBFLAG_A32, VFPEN, 7, 1)
34
+ * | | TBFLAG_A64 |
33
FIELD(TBFLAG_A32, CONDEXEC, 8, 8)
35
+ * +--------------+-------------------------------------+
34
FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
36
+ * 31 20 0
35
+/* For M profile only, set if FPCCR.LSPACT is set */
37
*
36
+FIELD(TBFLAG_A32, LSPACT, 18, 1)
38
* Unless otherwise noted, these bits are cached in env->hflags.
37
/* For M profile only, set if we must create a new FP context */
39
*/
38
FIELD(TBFLAG_A32, NEW_FP_CTXT_NEEDED, 19, 1)
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, BT, 9, 1)
39
/* For M profile only, set if FPCCR.S does not match current security state */
41
FIELD(TBFLAG_A64, BTYPE, 10, 2) /* Not cached. */
40
diff --git a/target/arm/helper.h b/target/arm/helper.h
42
FIELD(TBFLAG_A64, TBID, 12, 2)
43
FIELD(TBFLAG_A64, UNPRIV, 14, 1)
44
+FIELD(TBFLAG_A64, ATA, 15, 1)
45
+FIELD(TBFLAG_A64, TCMA, 16, 2)
46
+FIELD(TBFLAG_A64, MTE_ACTIVE, 18, 1)
47
+FIELD(TBFLAG_A64, MTE0_ACTIVE, 19, 1)
48
49
/**
50
* cpu_mmu_index:
51
diff --git a/target/arm/internals.h b/target/arm/internals.h
41
index XXXXXXX..XXXXXXX 100644
52
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/helper.h
53
--- a/target/arm/internals.h
43
+++ b/target/arm/helper.h
54
+++ b/target/arm/internals.h
44
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_2(v7m_blxns, void, env, i32)
55
@@ -XXX,XX +XXX,XX @@ static inline int exception_target_el(CPUARMState *env)
45
56
return target_el;
46
DEF_HELPER_3(v7m_tt, i32, env, i32, i32)
57
}
47
58
48
+DEF_HELPER_1(v7m_preserve_fp_state, void, env)
59
+/* Determine if allocation tags are available. */
60
+static inline bool allocation_tag_access_enabled(CPUARMState *env, int el,
61
+ uint64_t sctlr)
62
+{
63
+ if (el < 3
64
+ && arm_feature(env, ARM_FEATURE_EL3)
65
+ && !(env->cp15.scr_el3 & SCR_ATA)) {
66
+ return false;
67
+ }
68
+ if (el < 2
69
+ && arm_feature(env, ARM_FEATURE_EL2)
70
+ && !(arm_hcr_el2_eff(env) & HCR_ATA)) {
71
+ return false;
72
+ }
73
+ sctlr &= (el == 0 ? SCTLR_ATA0 : SCTLR_ATA);
74
+ return sctlr != 0;
75
+}
49
+
76
+
50
DEF_HELPER_2(v8m_stackcheck, void, env, i32)
77
#ifndef CONFIG_USER_ONLY
51
78
52
DEF_HELPER_4(access_check_cp_reg, void, env, ptr, i32, i32)
79
/* Security attributes for an address, as returned by v8m_security_lookup. */
53
diff --git a/target/arm/translate.h b/target/arm/translate.h
80
diff --git a/target/arm/translate.h b/target/arm/translate.h
54
index XXXXXXX..XXXXXXX 100644
81
index XXXXXXX..XXXXXXX 100644
55
--- a/target/arm/translate.h
82
--- a/target/arm/translate.h
56
+++ b/target/arm/translate.h
83
+++ b/target/arm/translate.h
57
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
84
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
58
bool v8m_stackcheck; /* true if we need to perform v8M stack limit checks */
85
ARMMMUIdx mmu_idx; /* MMU index to use for normal loads/stores */
59
bool v8m_fpccr_s_wrong; /* true if v8M FPCCR.S != v8m_secure */
86
uint8_t tbii; /* TBI1|TBI0 for insns */
60
bool v7m_new_fp_ctxt_needed; /* ASPEN set but no active FP context */
87
uint8_t tbid; /* TBI1|TBI0 for data */
61
+ bool v7m_lspact; /* FPCCR.LSPACT set */
88
+ uint8_t tcma; /* TCMA1|TCMA0 for MTE */
62
/* Immediate value in AArch32 SVC insn; must be set if is_jmp == DISAS_SWI
89
bool ns; /* Use non-secure CPREG bank on access */
63
* so that top level loop can generate correct syndrome information.
90
int fp_excp_el; /* FP exception EL or 0 if enabled */
64
*/
91
int sve_excp_el; /* SVE exception EL or 0 if enabled */
92
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
93
bool unpriv;
94
/* True if v8.3-PAuth is active. */
95
bool pauth_active;
96
+ /* True if v8.5-MTE access to tags is enabled. */
97
+ bool ata;
98
+ /* True if v8.5-MTE tag checks affect the PE; index with is_unpriv. */
99
+ bool mte_active[2];
100
/* True with v8.5-BTI and SCTLR_ELx.BT* set. */
101
bool bt;
102
/* True if any CP15 access is trapped by HSTR_EL2 */
65
diff --git a/target/arm/helper.c b/target/arm/helper.c
103
diff --git a/target/arm/helper.c b/target/arm/helper.c
66
index XXXXXXX..XXXXXXX 100644
104
index XXXXXXX..XXXXXXX 100644
67
--- a/target/arm/helper.c
105
--- a/target/arm/helper.c
68
+++ b/target/arm/helper.c
106
+++ b/target/arm/helper.c
69
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_blxns)(CPUARMState *env, uint32_t dest)
107
@@ -XXX,XX +XXX,XX @@ static int aa64_va_parameter_tbid(uint64_t tcr, ARMMMUIdx mmu_idx)
70
g_assert_not_reached();
108
}
71
}
109
}
72
110
73
+void HELPER(v7m_preserve_fp_state)(CPUARMState *env)
111
+static int aa64_va_parameter_tcma(uint64_t tcr, ARMMMUIdx mmu_idx)
74
+{
112
+{
75
+ /* translate.c should never generate calls here in user-only mode */
113
+ if (regime_has_2_ranges(mmu_idx)) {
76
+ g_assert_not_reached();
114
+ return extract64(tcr, 57, 2);
115
+ } else {
116
+ /* Replicate the single TCMA bit so we always have 2 bits. */
117
+ return extract32(tcr, 30, 1) * 3;
118
+ }
77
+}
119
+}
78
+
120
+
79
uint32_t HELPER(v7m_tt)(CPUARMState *env, uint32_t addr, uint32_t op)
121
ARMVAParameters aa64_va_parameters(CPUARMState *env, uint64_t va,
122
ARMMMUIdx mmu_idx, bool data)
80
{
123
{
81
/* The TT instructions can be used by unprivileged code, but in
124
@@ -XXX,XX +XXX,XX @@ static uint32_t rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
82
@@ -XXX,XX +XXX,XX @@ pend_fault:
125
}
83
return false;
126
}
84
}
127
85
128
+ if (cpu_isar_feature(aa64_mte, env_archcpu(env))) {
86
+void HELPER(v7m_preserve_fp_state)(CPUARMState *env)
129
+ /*
87
+{
130
+ * Set MTE_ACTIVE if any access may be Checked, and leave clear
88
+ /*
131
+ * if all accesses must be Unchecked:
89
+ * Preserve FP state (because LSPACT was set and we are about
132
+ * 1) If no TBI, then there are no tags in the address to check,
90
+ * to execute an FP instruction). This corresponds to the
133
+ * 2) If Tag Check Override, then all accesses are Unchecked,
91
+ * PreserveFPState() pseudocode.
134
+ * 3) If Tag Check Fail == 0, then Checked access have no effect,
92
+ * We may throw an exception if the stacking fails.
135
+ * 4) If no Allocation Tag Access, then all accesses are Unchecked.
93
+ */
136
+ */
94
+ ARMCPU *cpu = arm_env_get_cpu(env);
137
+ if (allocation_tag_access_enabled(env, el, sctlr)) {
95
+ bool is_secure = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_S_MASK;
138
+ flags = FIELD_DP32(flags, TBFLAG_A64, ATA, 1);
96
+ bool negpri = !(env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_HFRDY_MASK);
139
+ if (tbid
97
+ bool is_priv = !(env->v7m.fpccr[is_secure] & R_V7M_FPCCR_USER_MASK);
140
+ && !(env->pstate & PSTATE_TCO)
98
+ bool splimviol = env->v7m.fpccr[is_secure] & R_V7M_FPCCR_SPLIMVIOL_MASK;
141
+ && (sctlr & (el == 0 ? SCTLR_TCF0 : SCTLR_TCF))) {
99
+ uint32_t fpcar = env->v7m.fpcar[is_secure];
142
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE_ACTIVE, 1);
100
+ bool stacked_ok = true;
143
+ }
101
+ bool ts = is_secure && (env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_TS_MASK);
144
+ }
102
+ bool take_exception;
145
+ /* And again for unprivileged accesses, if required. */
103
+
146
+ if (FIELD_EX32(flags, TBFLAG_A64, UNPRIV)
104
+ /* Take the iothread lock as we are going to touch the NVIC */
147
+ && tbid
105
+ qemu_mutex_lock_iothread();
148
+ && !(env->pstate & PSTATE_TCO)
106
+
149
+ && (sctlr & SCTLR_TCF0)
107
+ /* Check the background context had access to the FPU */
150
+ && allocation_tag_access_enabled(env, 0, sctlr)) {
108
+ if (!v7m_cpacr_pass(env, is_secure, is_priv)) {
151
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE0_ACTIVE, 1);
109
+ armv7m_nvic_set_pending_lazyfp(env->nvic, ARMV7M_EXCP_USAGE, is_secure);
152
+ }
110
+ env->v7m.cfsr[is_secure] |= R_V7M_CFSR_NOCP_MASK;
153
+ /* Cache TCMA as well as TBI. */
111
+ stacked_ok = false;
154
+ flags = FIELD_DP32(flags, TBFLAG_A64, TCMA,
112
+ } else if (!is_secure && !extract32(env->v7m.nsacr, 10, 1)) {
155
+ aa64_va_parameter_tcma(tcr, mmu_idx));
113
+ armv7m_nvic_set_pending_lazyfp(env->nvic, ARMV7M_EXCP_USAGE, M_REG_S);
114
+ env->v7m.cfsr[M_REG_S] |= R_V7M_CFSR_NOCP_MASK;
115
+ stacked_ok = false;
116
+ }
156
+ }
117
+
157
+
118
+ if (!splimviol && stacked_ok) {
158
return rebuild_hflags_common(env, fp_el, mmu_idx, flags);
119
+ /* We only stack if the stack limit wasn't violated */
120
+ int i;
121
+ ARMMMUIdx mmu_idx;
122
+
123
+ mmu_idx = arm_v7m_mmu_idx_all(env, is_secure, is_priv, negpri);
124
+ for (i = 0; i < (ts ? 32 : 16); i += 2) {
125
+ uint64_t dn = *aa32_vfp_dreg(env, i / 2);
126
+ uint32_t faddr = fpcar + 4 * i;
127
+ uint32_t slo = extract64(dn, 0, 32);
128
+ uint32_t shi = extract64(dn, 32, 32);
129
+
130
+ if (i >= 16) {
131
+ faddr += 8; /* skip the slot for the FPSCR */
132
+ }
133
+ stacked_ok = stacked_ok &&
134
+ v7m_stack_write(cpu, faddr, slo, mmu_idx, STACK_LAZYFP) &&
135
+ v7m_stack_write(cpu, faddr + 4, shi, mmu_idx, STACK_LAZYFP);
136
+ }
137
+
138
+ stacked_ok = stacked_ok &&
139
+ v7m_stack_write(cpu, fpcar + 0x40,
140
+ vfp_get_fpscr(env), mmu_idx, STACK_LAZYFP);
141
+ }
142
+
143
+ /*
144
+ * We definitely pended an exception, but it's possible that it
145
+ * might not be able to be taken now. If its priority permits us
146
+ * to take it now, then we must not update the LSPACT or FP regs,
147
+ * but instead jump out to take the exception immediately.
148
+ * If it's just pending and won't be taken until the current
149
+ * handler exits, then we do update LSPACT and the FP regs.
150
+ */
151
+ take_exception = !stacked_ok &&
152
+ armv7m_nvic_can_take_pending_exception(env->nvic);
153
+
154
+ qemu_mutex_unlock_iothread();
155
+
156
+ if (take_exception) {
157
+ raise_exception_ra(env, EXCP_LAZYFP, 0, 1, GETPC());
158
+ }
159
+
160
+ env->v7m.fpccr[is_secure] &= ~R_V7M_FPCCR_LSPACT_MASK;
161
+
162
+ if (ts) {
163
+ /* Clear s0 to s31 and the FPSCR */
164
+ int i;
165
+
166
+ for (i = 0; i < 32; i += 2) {
167
+ *aa32_vfp_dreg(env, i / 2) = 0;
168
+ }
169
+ vfp_set_fpscr(env, 0);
170
+ }
171
+ /*
172
+ * Otherwise s0 to s15 and FPSCR are UNKNOWN; we choose to leave them
173
+ * unchanged.
174
+ */
175
+}
176
+
177
/* Write to v7M CONTROL.SPSEL bit for the specified security bank.
178
* This may change the current stack pointer between Main and Process
179
* stack pointers if it is done for the CONTROL register for the current
180
@@ -XXX,XX +XXX,XX @@ static void arm_log_exception(int idx)
181
[EXCP_NOCP] = "v7M NOCP UsageFault",
182
[EXCP_INVSTATE] = "v7M INVSTATE UsageFault",
183
[EXCP_STKOF] = "v8M STKOF UsageFault",
184
+ [EXCP_LAZYFP] = "v7M exception during lazy FP stacking",
185
};
186
187
if (idx >= 0 && idx < ARRAY_SIZE(excnames)) {
188
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
189
return;
190
}
191
break;
192
+ case EXCP_LAZYFP:
193
+ /*
194
+ * We already pended the specific exception in the NVIC in the
195
+ * v7m_preserve_fp_state() helper function.
196
+ */
197
+ break;
198
default:
199
cpu_abort(cs, "Unhandled exception 0x%x\n", cs->exception_index);
200
return; /* Never happens. Keep compiler happy. */
201
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
202
flags = FIELD_DP32(flags, TBFLAG_A32, NEW_FP_CTXT_NEEDED, 1);
203
}
204
205
+ if (arm_feature(env, ARM_FEATURE_M)) {
206
+ bool is_secure = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_S_MASK;
207
+
208
+ if (env->v7m.fpccr[is_secure] & R_V7M_FPCCR_LSPACT_MASK) {
209
+ flags = FIELD_DP32(flags, TBFLAG_A32, LSPACT, 1);
210
+ }
211
+ }
212
+
213
*pflags = flags;
214
*cs_base = 0;
215
}
159
}
216
diff --git a/target/arm/translate.c b/target/arm/translate.c
160
161
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
217
index XXXXXXX..XXXXXXX 100644
162
index XXXXXXX..XXXXXXX 100644
218
--- a/target/arm/translate.c
163
--- a/target/arm/translate-a64.c
219
+++ b/target/arm/translate.c
164
+++ b/target/arm/translate-a64.c
220
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
165
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
221
if (arm_dc_feature(s, ARM_FEATURE_M)) {
166
dc->mmu_idx = core_to_aa64_mmu_idx(core_mmu_idx);
222
/* Handle M-profile lazy FP state mechanics */
167
dc->tbii = FIELD_EX32(tb_flags, TBFLAG_A64, TBII);
223
168
dc->tbid = FIELD_EX32(tb_flags, TBFLAG_A64, TBID);
224
+ /* Trigger lazy-state preservation if necessary */
169
+ dc->tcma = FIELD_EX32(tb_flags, TBFLAG_A64, TCMA);
225
+ if (s->v7m_lspact) {
170
dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
226
+ /*
171
#if !defined(CONFIG_USER_ONLY)
227
+ * Lazy state saving affects external memory and also the NVIC,
172
dc->user = (dc->current_el == 0);
228
+ * so we must mark it as an IO operation for icount.
173
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
229
+ */
174
dc->bt = FIELD_EX32(tb_flags, TBFLAG_A64, BT);
230
+ if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
175
dc->btype = FIELD_EX32(tb_flags, TBFLAG_A64, BTYPE);
231
+ gen_io_start();
176
dc->unpriv = FIELD_EX32(tb_flags, TBFLAG_A64, UNPRIV);
232
+ }
177
+ dc->ata = FIELD_EX32(tb_flags, TBFLAG_A64, ATA);
233
+ gen_helper_v7m_preserve_fp_state(cpu_env);
178
+ dc->mte_active[0] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE_ACTIVE);
234
+ if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
179
+ dc->mte_active[1] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE0_ACTIVE);
235
+ gen_io_end();
180
dc->vec_len = 0;
236
+ }
181
dc->vec_stride = 0;
237
+ /*
182
dc->cp_regs = arm_cpu->cp_regs;
238
+ * If the preserve_fp_state helper doesn't throw an exception
239
+ * then it will clear LSPACT; we don't need to repeat this for
240
+ * any further FP insns in this TB.
241
+ */
242
+ s->v7m_lspact = false;
243
+ }
244
+
245
/* Update ownership of FP context: set FPCCR.S to match current state */
246
if (s->v8m_fpccr_s_wrong) {
247
TCGv_i32 tmp;
248
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
249
dc->v8m_fpccr_s_wrong = FIELD_EX32(tb_flags, TBFLAG_A32, FPCCR_S_WRONG);
250
dc->v7m_new_fp_ctxt_needed =
251
FIELD_EX32(tb_flags, TBFLAG_A32, NEW_FP_CTXT_NEEDED);
252
+ dc->v7m_lspact = FIELD_EX32(tb_flags, TBFLAG_A32, LSPACT);
253
dc->cp_regs = cpu->cp_regs;
254
dc->features = env->features;
255
256
--
183
--
257
2.20.1
184
2.20.1
258
185
259
186
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Markus Armbruster <armbru@redhat.com>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20190412165416.7977-10-philmd@redhat.com
5
Message-id: 20200626033144.790098-10-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
7
---
8
include/hw/devices.h | 3 ---
8
target/arm/helper-a64.h | 2 ++
9
include/hw/net/lan9118.h | 19 +++++++++++++++++++
9
target/arm/internals.h | 5 +++
10
hw/arm/kzm.c | 2 +-
10
target/arm/mte_helper.c | 72 ++++++++++++++++++++++++++++++++++++++
11
hw/arm/mps2.c | 2 +-
11
target/arm/translate-a64.c | 18 ++++++++++
12
hw/arm/realview.c | 1 +
12
target/arm/Makefile.objs | 1 +
13
hw/arm/vexpress.c | 2 +-
13
5 files changed, 98 insertions(+)
14
hw/net/lan9118.c | 2 +-
14
create mode 100644 target/arm/mte_helper.c
15
7 files changed, 24 insertions(+), 7 deletions(-)
16
create mode 100644 include/hw/net/lan9118.h
17
15
18
diff --git a/include/hw/devices.h b/include/hw/devices.h
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
19
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/devices.h
18
--- a/target/arm/helper-a64.h
21
+++ b/include/hw/devices.h
19
+++ b/target/arm/helper-a64.h
22
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autda, TCG_CALL_NO_WG, i64, env, i64, i64)
23
/* smc91c111.c */
21
DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
24
void smc91c111_init(NICInfo *, uint32_t, qemu_irq);
22
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
25
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
-/* lan9118.c */
24
+
27
-void lan9118_init(NICInfo *, uint32_t, qemu_irq);
25
+DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
28
-
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
27
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/internals.h
29
+++ b/target/arm/internals.h
30
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
31
*/
32
#define GMID_EL1_BS 6
33
34
+static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
35
+{
36
+ return deposit64(ptr, 56, 4, rtag);
37
+}
38
+
29
#endif
39
#endif
30
diff --git a/include/hw/net/lan9118.h b/include/hw/net/lan9118.h
40
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
31
new file mode 100644
41
new file mode 100644
32
index XXXXXXX..XXXXXXX
42
index XXXXXXX..XXXXXXX
33
--- /dev/null
43
--- /dev/null
34
+++ b/include/hw/net/lan9118.h
44
+++ b/target/arm/mte_helper.c
35
@@ -XXX,XX +XXX,XX @@
45
@@ -XXX,XX +XXX,XX @@
36
+/*
46
+/*
37
+ * SMSC LAN9118 Ethernet interface emulation
47
+ * ARM v8.5-MemTag Operations
38
+ *
48
+ *
39
+ * Copyright (c) 2009 CodeSourcery, LLC.
49
+ * Copyright (c) 2020 Linaro, Ltd.
40
+ * Written by Paul Brook
41
+ *
50
+ *
42
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
51
+ * This library is free software; you can redistribute it and/or
43
+ * See the COPYING file in the top-level directory.
52
+ * modify it under the terms of the GNU Lesser General Public
53
+ * License as published by the Free Software Foundation; either
54
+ * version 2.1 of the License, or (at your option) any later version.
55
+ *
56
+ * This library is distributed in the hope that it will be useful,
57
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
58
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
59
+ * Lesser General Public License for more details.
60
+ *
61
+ * You should have received a copy of the GNU Lesser General Public
62
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
44
+ */
63
+ */
45
+
64
+
46
+#ifndef HW_NET_LAN9118_H
65
+#include "qemu/osdep.h"
47
+#define HW_NET_LAN9118_H
66
+#include "cpu.h"
67
+#include "internals.h"
68
+#include "exec/exec-all.h"
69
+#include "exec/cpu_ldst.h"
70
+#include "exec/helper-proto.h"
48
+
71
+
49
+#include "hw/irq.h"
50
+#include "net/net.h"
51
+
72
+
52
+void lan9118_init(NICInfo *, uint32_t, qemu_irq);
73
+static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
74
+{
75
+ if (exclude == 0xffff) {
76
+ return 0;
77
+ }
78
+ if (offset == 0) {
79
+ while (exclude & (1 << tag)) {
80
+ tag = (tag + 1) & 15;
81
+ }
82
+ } else {
83
+ do {
84
+ do {
85
+ tag = (tag + 1) & 15;
86
+ } while (exclude & (1 << tag));
87
+ } while (--offset > 0);
88
+ }
89
+ return tag;
90
+}
53
+
91
+
54
+#endif
92
+uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
55
diff --git a/hw/arm/kzm.c b/hw/arm/kzm.c
93
+{
94
+ int rtag;
95
+
96
+ /*
97
+ * Our IMPDEF choice for GCR_EL1.RRND==1 is to behave as if
98
+ * GCR_EL1.RRND==0, always producing deterministic results.
99
+ */
100
+ uint16_t exclude = extract32(rm | env->cp15.gcr_el1, 0, 16);
101
+ int start = extract32(env->cp15.rgsr_el1, 0, 4);
102
+ int seed = extract32(env->cp15.rgsr_el1, 8, 16);
103
+ int offset, i;
104
+
105
+ /* RandomTag */
106
+ for (i = offset = 0; i < 4; ++i) {
107
+ /* NextRandomTagBit */
108
+ int top = (extract32(seed, 5, 1) ^ extract32(seed, 3, 1) ^
109
+ extract32(seed, 2, 1) ^ extract32(seed, 0, 1));
110
+ seed = (top << 15) | (seed >> 1);
111
+ offset |= top << i;
112
+ }
113
+ rtag = choose_nonexcluded_tag(start, offset, exclude);
114
+ env->cp15.rgsr_el1 = rtag | (seed << 8);
115
+
116
+ return address_with_allocation_tag(rn, rtag);
117
+}
118
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
56
index XXXXXXX..XXXXXXX 100644
119
index XXXXXXX..XXXXXXX 100644
57
--- a/hw/arm/kzm.c
120
--- a/target/arm/translate-a64.c
58
+++ b/hw/arm/kzm.c
121
+++ b/target/arm/translate-a64.c
59
@@ -XXX,XX +XXX,XX @@
122
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
60
#include "qemu/error-report.h"
123
return clean;
61
#include "exec/address-spaces.h"
124
}
62
#include "net/net.h"
125
63
-#include "hw/devices.h"
126
+/* Insert a zero tag into src, with the result at dst. */
64
+#include "hw/net/lan9118.h"
127
+static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
65
#include "hw/char/serial.h"
128
+{
66
#include "sysemu/qtest.h"
129
+ tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
67
130
+}
68
diff --git a/hw/arm/mps2.c b/hw/arm/mps2.c
131
+
132
typedef struct DisasCompare64 {
133
TCGCond cond;
134
TCGv_i64 value;
135
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
136
case 3: /* SDIV */
137
handle_div(s, true, sf, rm, rn, rd);
138
break;
139
+ case 4: /* IRG */
140
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
141
+ goto do_unallocated;
142
+ }
143
+ if (s->ata) {
144
+ gen_helper_irg(cpu_reg_sp(s, rd), cpu_env,
145
+ cpu_reg_sp(s, rn), cpu_reg(s, rm));
146
+ } else {
147
+ gen_address_with_allocation_tag0(cpu_reg_sp(s, rd),
148
+ cpu_reg_sp(s, rn));
149
+ }
150
+ break;
151
case 8: /* LSLV */
152
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
153
break;
154
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
69
index XXXXXXX..XXXXXXX 100644
155
index XXXXXXX..XXXXXXX 100644
70
--- a/hw/arm/mps2.c
156
--- a/target/arm/Makefile.objs
71
+++ b/hw/arm/mps2.c
157
+++ b/target/arm/Makefile.objs
72
@@ -XXX,XX +XXX,XX @@
158
@@ -XXX,XX +XXX,XX @@ obj-$(CONFIG_SOFTMMU) += psci.o
73
#include "hw/timer/cmsdk-apb-timer.h"
159
obj-$(TARGET_AARCH64) += translate-a64.o helper-a64.o
74
#include "hw/timer/cmsdk-apb-dualtimer.h"
160
obj-$(TARGET_AARCH64) += translate-sve.o sve_helper.o
75
#include "hw/misc/mps2-scc.h"
161
obj-$(TARGET_AARCH64) += pauth_helper.o
76
-#include "hw/devices.h"
162
+obj-$(TARGET_AARCH64) += mte_helper.o
77
+#include "hw/net/lan9118.h"
78
#include "net/net.h"
79
80
typedef enum MPS2FPGAType {
81
diff --git a/hw/arm/realview.c b/hw/arm/realview.c
82
index XXXXXXX..XXXXXXX 100644
83
--- a/hw/arm/realview.c
84
+++ b/hw/arm/realview.c
85
@@ -XXX,XX +XXX,XX @@
86
#include "hw/arm/arm.h"
87
#include "hw/arm/primecell.h"
88
#include "hw/devices.h"
89
+#include "hw/net/lan9118.h"
90
#include "hw/pci/pci.h"
91
#include "net/net.h"
92
#include "sysemu/sysemu.h"
93
diff --git a/hw/arm/vexpress.c b/hw/arm/vexpress.c
94
index XXXXXXX..XXXXXXX 100644
95
--- a/hw/arm/vexpress.c
96
+++ b/hw/arm/vexpress.c
97
@@ -XXX,XX +XXX,XX @@
98
#include "hw/sysbus.h"
99
#include "hw/arm/arm.h"
100
#include "hw/arm/primecell.h"
101
-#include "hw/devices.h"
102
+#include "hw/net/lan9118.h"
103
#include "hw/i2c/i2c.h"
104
#include "net/net.h"
105
#include "sysemu/sysemu.h"
106
diff --git a/hw/net/lan9118.c b/hw/net/lan9118.c
107
index XXXXXXX..XXXXXXX 100644
108
--- a/hw/net/lan9118.c
109
+++ b/hw/net/lan9118.c
110
@@ -XXX,XX +XXX,XX @@
111
#include "hw/sysbus.h"
112
#include "net/net.h"
113
#include "net/eth.h"
114
-#include "hw/devices.h"
115
+#include "hw/net/lan9118.h"
116
#include "sysemu/sysemu.h"
117
#include "hw/ptimer.h"
118
#include "qemu/log.h"
119
--
163
--
120
2.20.1
164
2.20.1
121
165
122
166
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
The current Arm ARM has adjusted the official decode of
4
"Add/subtract (immediate)" so that the shift field is only bit 22,
5
and bit 23 is part of the op1 field of the parent category
6
"Data processing - immediate".
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20200626033144.790098-11-richard.henderson@linaro.org
11
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/translate-a64.c | 23 ++++++++---------------
16
1 file changed, 8 insertions(+), 15 deletions(-)
17
18
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/translate-a64.c
21
+++ b/target/arm/translate-a64.c
22
@@ -XXX,XX +XXX,XX @@ static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
23
/*
24
* Add/subtract (immediate)
25
*
26
- * 31 30 29 28 24 23 22 21 10 9 5 4 0
27
- * +--+--+--+-----------+-----+-------------+-----+-----+
28
- * |sf|op| S| 1 0 0 0 1 |shift| imm12 | Rn | Rd |
29
- * +--+--+--+-----------+-----+-------------+-----+-----+
30
+ * 31 30 29 28 23 22 21 10 9 5 4 0
31
+ * +--+--+--+-------------+--+-------------+-----+-----+
32
+ * |sf|op| S| 1 0 0 0 1 0 |sh| imm12 | Rn | Rd |
33
+ * +--+--+--+-------------+--+-------------+-----+-----+
34
*
35
* sf: 0 -> 32bit, 1 -> 64bit
36
* op: 0 -> add , 1 -> sub
37
* S: 1 -> set flags
38
- * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
39
+ * sh: 1 -> LSL imm by 12
40
*/
41
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
42
{
43
int rd = extract32(insn, 0, 5);
44
int rn = extract32(insn, 5, 5);
45
uint64_t imm = extract32(insn, 10, 12);
46
- int shift = extract32(insn, 22, 2);
47
+ bool shift = extract32(insn, 22, 1);
48
bool setflags = extract32(insn, 29, 1);
49
bool sub_op = extract32(insn, 30, 1);
50
bool is_64bit = extract32(insn, 31, 1);
51
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
53
TCGv_i64 tcg_result;
54
55
- switch (shift) {
56
- case 0x0:
57
- break;
58
- case 0x1:
59
+ if (shift) {
60
imm <<= 12;
61
- break;
62
- default:
63
- unallocated_encoding(s);
64
- return;
65
}
66
67
tcg_result = tcg_temp_new_i64();
68
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
69
case 0x20: case 0x21: /* PC-rel. addressing */
70
disas_pc_rel_adr(s, insn);
71
break;
72
- case 0x22: case 0x23: /* Add/subtract (immediate) */
73
+ case 0x22: /* Add/subtract (immediate) */
74
disas_add_sub_imm(s, insn);
75
break;
76
case 0x24: /* Logical (immediate) */
77
--
78
2.20.1
79
80
diff view generated by jsdifflib
1
Implement the VLLDM instruction for v7M for the FPU present cas.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-12-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20190416125744.27770-26-peter.maydell@linaro.org
6
---
7
---
7
target/arm/helper.h | 1 +
8
target/arm/helper-a64.h | 1 +
8
target/arm/helper.c | 54 ++++++++++++++++++++++++++++++++++++++++++
9
target/arm/internals.h | 9 +++++++
9
target/arm/translate.c | 2 +-
10
target/arm/mte_helper.c | 10 ++++++++
10
3 files changed, 56 insertions(+), 1 deletion(-)
11
target/arm/translate-a64.c | 51 ++++++++++++++++++++++++++++++++++++++
12
4 files changed, 71 insertions(+)
11
13
12
diff --git a/target/arm/helper.h b/target/arm/helper.h
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/helper.h
16
--- a/target/arm/helper-a64.h
15
+++ b/target/arm/helper.h
17
+++ b/target/arm/helper-a64.h
16
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_3(v7m_tt, i32, env, i32, i32)
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
17
DEF_HELPER_1(v7m_preserve_fp_state, void, env)
19
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
18
20
19
DEF_HELPER_2(v7m_vlstm, void, env, i32)
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
20
+DEF_HELPER_2(v7m_vlldm, void, env, i32)
22
+DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
21
23
diff --git a/target/arm/internals.h b/target/arm/internals.h
22
DEF_HELPER_2(v8m_stackcheck, void, env, i32)
23
24
diff --git a/target/arm/helper.c b/target/arm/helper.c
25
index XXXXXXX..XXXXXXX 100644
24
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/helper.c
25
--- a/target/arm/internals.h
27
+++ b/target/arm/helper.c
26
+++ b/target/arm/internals.h
28
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_vlstm)(CPUARMState *env, uint32_t fptr)
27
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
29
g_assert_not_reached();
28
*/
30
}
29
#define GMID_EL1_BS 6
31
30
32
+void HELPER(v7m_vlldm)(CPUARMState *env, uint32_t fptr)
31
+/* We associate one allocation tag per 16 bytes, the minimum. */
32
+#define LOG2_TAG_GRANULE 4
33
+#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
34
+
35
+static inline int allocation_tag_from_addr(uint64_t ptr)
33
+{
36
+{
34
+ /* translate.c should never generate calls here in user-only mode */
37
+ return extract64(ptr, 56, 4);
35
+ g_assert_not_reached();
36
+}
38
+}
37
+
39
+
38
uint32_t HELPER(v7m_tt)(CPUARMState *env, uint32_t addr, uint32_t op)
40
static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
39
{
41
{
40
/* The TT instructions can be used by unprivileged code, but in
42
return deposit64(ptr, 56, 4, rtag);
41
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_vlstm)(CPUARMState *env, uint32_t fptr)
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
42
env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_FPCA_MASK;
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
48
49
return address_with_allocation_tag(rn, rtag);
43
}
50
}
44
51
+
45
+void HELPER(v7m_vlldm)(CPUARMState *env, uint32_t fptr)
52
+uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
53
+ int32_t offset, uint32_t tag_offset)
46
+{
54
+{
47
+ /* fptr is the value of Rn, the frame pointer we load the FP regs from */
55
+ int start_tag = allocation_tag_from_addr(ptr);
48
+ assert(env->v7m.secure);
56
+ uint16_t exclude = extract32(env->cp15.gcr_el1, 0, 16);
57
+ int rtag = choose_nonexcluded_tag(start_tag, tag_offset, exclude);
49
+
58
+
50
+ if (!(env->v7m.control[M_REG_S] & R_V7M_CONTROL_SFPA_MASK)) {
59
+ return address_with_allocation_tag(ptr + offset, rtag);
60
+}
61
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
62
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate-a64.c
64
+++ b/target/arm/translate-a64.c
65
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
66
tcg_temp_free_i64(tcg_result);
67
}
68
69
+/*
70
+ * Add/subtract (immediate, with tags)
71
+ *
72
+ * 31 30 29 28 23 22 21 16 14 10 9 5 4 0
73
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
74
+ * |sf|op| S| 1 0 0 0 1 1 |o2| uimm6 |o3| uimm4 | Rn | Rd |
75
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
76
+ *
77
+ * op: 0 -> add, 1 -> sub
78
+ */
79
+static void disas_add_sub_imm_with_tags(DisasContext *s, uint32_t insn)
80
+{
81
+ int rd = extract32(insn, 0, 5);
82
+ int rn = extract32(insn, 5, 5);
83
+ int uimm4 = extract32(insn, 10, 4);
84
+ int uimm6 = extract32(insn, 16, 6);
85
+ bool sub_op = extract32(insn, 30, 1);
86
+ TCGv_i64 tcg_rn, tcg_rd;
87
+ int imm;
88
+
89
+ /* Test all of sf=1, S=0, o2=0, o3=0. */
90
+ if ((insn & 0xa040c000u) != 0x80000000u ||
91
+ !dc_isar_feature(aa64_mte_insn_reg, s)) {
92
+ unallocated_encoding(s);
51
+ return;
93
+ return;
52
+ }
94
+ }
53
+
95
+
54
+ /* Check access to the coprocessor is permitted */
96
+ imm = uimm6 << LOG2_TAG_GRANULE;
55
+ if (!v7m_cpacr_pass(env, true, arm_current_el(env) != 0)) {
97
+ if (sub_op) {
56
+ raise_exception_ra(env, EXCP_NOCP, 0, 1, GETPC());
98
+ imm = -imm;
57
+ }
99
+ }
58
+
100
+
59
+ if (env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_LSPACT_MASK) {
101
+ tcg_rn = cpu_reg_sp(s, rn);
60
+ /* State in FP is still valid */
102
+ tcg_rd = cpu_reg_sp(s, rd);
61
+ env->v7m.fpccr[M_REG_S] &= ~R_V7M_FPCCR_LSPACT_MASK;
103
+
104
+ if (s->ata) {
105
+ TCGv_i32 offset = tcg_const_i32(imm);
106
+ TCGv_i32 tag_offset = tcg_const_i32(uimm4);
107
+
108
+ gen_helper_addsubg(tcg_rd, cpu_env, tcg_rn, offset, tag_offset);
109
+ tcg_temp_free_i32(tag_offset);
110
+ tcg_temp_free_i32(offset);
62
+ } else {
111
+ } else {
63
+ bool ts = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_TS_MASK;
112
+ tcg_gen_addi_i64(tcg_rd, tcg_rn, imm);
64
+ int i;
113
+ gen_address_with_allocation_tag0(tcg_rd, tcg_rd);
65
+ uint32_t fpscr;
66
+
67
+ if (fptr & 7) {
68
+ raise_exception_ra(env, EXCP_UNALIGNED, 0, 1, GETPC());
69
+ }
70
+
71
+ for (i = 0; i < (ts ? 32 : 16); i += 2) {
72
+ uint32_t slo, shi;
73
+ uint64_t dn;
74
+ uint32_t faddr = fptr + 4 * i;
75
+
76
+ if (i >= 16) {
77
+ faddr += 8; /* skip the slot for the FPSCR */
78
+ }
79
+
80
+ slo = cpu_ldl_data(env, faddr);
81
+ shi = cpu_ldl_data(env, faddr + 4);
82
+
83
+ dn = (uint64_t) shi << 32 | slo;
84
+ *aa32_vfp_dreg(env, i / 2) = dn;
85
+ }
86
+ fpscr = cpu_ldl_data(env, fptr + 0x40);
87
+ vfp_set_fpscr(env, fpscr);
88
+ }
114
+ }
89
+
90
+ env->v7m.control[M_REG_S] |= R_V7M_CONTROL_FPCA_MASK;
91
+}
115
+}
92
+
116
+
93
static bool v7m_push_stack(ARMCPU *cpu)
117
/* The input should be a value in the bottom e bits (with higher
94
{
118
* bits zero); returns that value replicated into every element
95
/* Do the "set up stack frame" part of exception entry,
119
* of size e in a 64 bit integer.
96
diff --git a/target/arm/translate.c b/target/arm/translate.c
120
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
97
index XXXXXXX..XXXXXXX 100644
121
case 0x22: /* Add/subtract (immediate) */
98
--- a/target/arm/translate.c
122
disas_add_sub_imm(s, insn);
99
+++ b/target/arm/translate.c
123
break;
100
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
124
+ case 0x23: /* Add/subtract (immediate, with tags) */
101
TCGv_i32 fptr = load_reg(s, rn);
125
+ disas_add_sub_imm_with_tags(s, insn);
102
126
+ break;
103
if (extract32(insn, 20, 1)) {
127
case 0x24: /* Logical (immediate) */
104
- /* VLLDM */
128
disas_logic_imm(s, insn);
105
+ gen_helper_v7m_vlldm(cpu_env, fptr);
129
break;
106
} else {
107
gen_helper_v7m_vlstm(cpu_env, fptr);
108
}
109
--
130
--
110
2.20.1
131
2.20.1
111
132
112
133
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-13-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 15 +++++++++++++++
9
1 file changed, 15 insertions(+)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
16
cpu_reg_sp(s, rn));
17
}
18
break;
19
+ case 5: /* GMI */
20
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
21
+ goto do_unallocated;
22
+ } else {
23
+ TCGv_i64 t1 = tcg_const_i64(1);
24
+ TCGv_i64 t2 = tcg_temp_new_i64();
25
+
26
+ tcg_gen_extract_i64(t2, cpu_reg_sp(s, rn), 56, 4);
27
+ tcg_gen_shl_i64(t1, t1, t2);
28
+ tcg_gen_or_i64(cpu_reg(s, rd), cpu_reg(s, rm), t1);
29
+
30
+ tcg_temp_free_i64(t1);
31
+ tcg_temp_free_i64(t2);
32
+ }
33
+ break;
34
case 8: /* LSLV */
35
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
36
break;
37
--
38
2.20.1
39
40
diff view generated by jsdifflib
1
The M-profile CONTROL register has two bits -- SFPA and FPCA --
1
From: Richard Henderson <richard.henderson@linaro.org>
2
which relate to floating-point support, and should be RES0 otherwise.
3
Handle them correctly in the MSR/MRS register access code.
4
Neither is banked between security states, so they are stored
5
in v7m.control[M_REG_S] regardless of current security state.
6
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-14-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20190416125744.27770-9-peter.maydell@linaro.org
10
---
7
---
11
target/arm/helper.c | 57 ++++++++++++++++++++++++++++++++++++++-------
8
target/arm/translate-a64.c | 24 ++++++++++++++++++++++--
12
1 file changed, 49 insertions(+), 8 deletions(-)
9
1 file changed, 22 insertions(+), 2 deletions(-)
13
10
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
15
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
13
--- a/target/arm/translate-a64.c
17
+++ b/target/arm/helper.c
14
+++ b/target/arm/translate-a64.c
18
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
15
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
19
return xpsr_read(env) & mask;
16
*/
20
break;
17
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
21
case 20: /* CONTROL */
18
{
22
- return env->v7m.control[env->v7m.secure];
19
- unsigned int sf, rm, opcode, rn, rd;
23
+ {
20
+ unsigned int sf, rm, opcode, rn, rd, setflag;
24
+ uint32_t value = env->v7m.control[env->v7m.secure];
21
sf = extract32(insn, 31, 1);
25
+ if (!env->v7m.secure) {
22
+ setflag = extract32(insn, 29, 1);
26
+ /* SFPA is RAZ/WI from NS; FPCA is stored in the M_REG_S bank */
23
rm = extract32(insn, 16, 5);
27
+ value |= env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK;
24
opcode = extract32(insn, 10, 6);
28
+ }
25
rn = extract32(insn, 5, 5);
29
+ return value;
26
rd = extract32(insn, 0, 5);
30
+ }
27
31
case 0x94: /* CONTROL_NS */
28
- if (extract32(insn, 29, 1)) {
32
/* We have to handle this here because unprivileged Secure code
29
+ if (setflag && opcode != 0) {
33
* can read the NS CONTROL register.
30
unallocated_encoding(s);
34
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
35
if (!env->v7m.secure) {
36
return 0;
37
}
38
- return env->v7m.control[M_REG_NS];
39
+ return env->v7m.control[M_REG_NS] |
40
+ (env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK);
41
}
42
43
if (el == 0) {
44
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
45
*/
46
uint32_t mask = extract32(maskreg, 8, 4);
47
uint32_t reg = extract32(maskreg, 0, 8);
48
+ int cur_el = arm_current_el(env);
49
50
- if (arm_current_el(env) == 0 && reg > 7) {
51
- /* only xPSR sub-fields may be written by unprivileged */
52
+ if (cur_el == 0 && reg > 7 && reg != 20) {
53
+ /*
54
+ * only xPSR sub-fields and CONTROL.SFPA may be written by
55
+ * unprivileged code
56
+ */
57
return;
31
return;
58
}
32
}
59
33
60
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
34
switch (opcode) {
61
env->v7m.control[M_REG_NS] &= ~R_V7M_CONTROL_NPRIV_MASK;
35
+ case 0: /* SUBP(S) */
62
env->v7m.control[M_REG_NS] |= val & R_V7M_CONTROL_NPRIV_MASK;
36
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
63
}
37
+ goto do_unallocated;
64
+ /*
38
+ } else {
65
+ * SFPA is RAZ/WI from NS. FPCA is RO if NSACR.CP10 == 0,
39
+ TCGv_i64 tcg_n, tcg_m, tcg_d;
66
+ * RES0 if the FPU is not present, and is stored in the S bank
40
+
67
+ */
41
+ tcg_n = read_cpu_reg_sp(s, rn, true);
68
+ if (arm_feature(env, ARM_FEATURE_VFP) &&
42
+ tcg_m = read_cpu_reg_sp(s, rm, true);
69
+ extract32(env->v7m.nsacr, 10, 1)) {
43
+ tcg_gen_sextract_i64(tcg_n, tcg_n, 0, 56);
70
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_FPCA_MASK;
44
+ tcg_gen_sextract_i64(tcg_m, tcg_m, 0, 56);
71
+ env->v7m.control[M_REG_S] |= val & R_V7M_CONTROL_FPCA_MASK;
45
+ tcg_d = cpu_reg(s, rd);
72
+ }
46
+
73
return;
47
+ if (setflag) {
74
case 0x98: /* SP_NS */
48
+ gen_sub_CC(true, tcg_d, tcg_n, tcg_m);
75
{
49
+ } else {
76
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
50
+ tcg_gen_sub_i64(tcg_d, tcg_n, tcg_m);
77
env->v7m.faultmask[env->v7m.secure] = val & 1;
78
break;
79
case 20: /* CONTROL */
80
- /* Writing to the SPSEL bit only has an effect if we are in
81
+ /*
82
+ * Writing to the SPSEL bit only has an effect if we are in
83
* thread mode; other bits can be updated by any privileged code.
84
* write_v7m_control_spsel() deals with updating the SPSEL bit in
85
* env->v7m.control, so we only need update the others.
86
* For v7M, we must just ignore explicit writes to SPSEL in handler
87
* mode; for v8M the write is permitted but will have no effect.
88
+ * All these bits are writes-ignored from non-privileged code,
89
+ * except for SFPA.
90
*/
91
- if (arm_feature(env, ARM_FEATURE_V8) ||
92
- !arm_v7m_is_handler_mode(env)) {
93
+ if (cur_el > 0 && (arm_feature(env, ARM_FEATURE_V8) ||
94
+ !arm_v7m_is_handler_mode(env))) {
95
write_v7m_control_spsel(env, (val & R_V7M_CONTROL_SPSEL_MASK) != 0);
96
}
97
- if (arm_feature(env, ARM_FEATURE_M_MAIN)) {
98
+ if (cur_el > 0 && arm_feature(env, ARM_FEATURE_M_MAIN)) {
99
env->v7m.control[env->v7m.secure] &= ~R_V7M_CONTROL_NPRIV_MASK;
100
env->v7m.control[env->v7m.secure] |= val & R_V7M_CONTROL_NPRIV_MASK;
101
}
102
+ if (arm_feature(env, ARM_FEATURE_VFP)) {
103
+ /*
104
+ * SFPA is RAZ/WI from NS or if no FPU.
105
+ * FPCA is RO if NSACR.CP10 == 0, RES0 if the FPU is not present.
106
+ * Both are stored in the S bank.
107
+ */
108
+ if (env->v7m.secure) {
109
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_SFPA_MASK;
110
+ env->v7m.control[M_REG_S] |= val & R_V7M_CONTROL_SFPA_MASK;
111
+ }
112
+ if (cur_el > 0 &&
113
+ (env->v7m.secure || !arm_feature(env, ARM_FEATURE_M_SECURITY) ||
114
+ extract32(env->v7m.nsacr, 10, 1))) {
115
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_FPCA_MASK;
116
+ env->v7m.control[M_REG_S] |= val & R_V7M_CONTROL_FPCA_MASK;
117
+ }
51
+ }
118
+ }
52
+ }
53
+ break;
54
case 2: /* UDIV */
55
handle_div(s, false, sf, rm, rn, rd);
119
break;
56
break;
120
default:
121
bad_reg:
122
--
57
--
123
2.20.1
58
2.20.1
124
59
125
60
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Use the same code as system mode, so that we generate the same
4
exception + syndrome for the unaligned access.
5
6
For the moment, if MTE is enabled so that this path is reachable,
7
this would generate a SIGSEGV in the user-only cpu_loop. Decoding
8
the syndrome to produce the proper SIGBUS will be done later.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-15-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/cpu.c | 2 +-
16
target/arm/tlb_helper.c | 4 ++--
17
2 files changed, 3 insertions(+), 3 deletions(-)
18
19
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.c
22
+++ b/target/arm/cpu.c
23
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
24
cc->tlb_fill = arm_cpu_tlb_fill;
25
cc->debug_excp_handler = arm_debug_excp_handler;
26
cc->debug_check_watchpoint = arm_debug_check_watchpoint;
27
-#if !defined(CONFIG_USER_ONLY)
28
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
29
+#if !defined(CONFIG_USER_ONLY)
30
cc->do_transaction_failed = arm_cpu_do_transaction_failed;
31
cc->adjust_watchpoint_address = arm_adjust_watchpoint_address;
32
#endif /* CONFIG_TCG && !CONFIG_USER_ONLY */
33
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
34
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/tlb_helper.c
36
+++ b/target/arm/tlb_helper.c
37
@@ -XXX,XX +XXX,XX @@
38
#include "internals.h"
39
#include "exec/exec-all.h"
40
41
-#if !defined(CONFIG_USER_ONLY)
42
-
43
static inline uint32_t merge_syn_data_abort(uint32_t template_syn,
44
unsigned int target_el,
45
bool same_el, bool ea,
46
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
47
arm_deliver_fault(cpu, vaddr, access_type, mmu_idx, &fi);
48
}
49
50
+#if !defined(CONFIG_USER_ONLY)
51
+
52
/*
53
* arm_cpu_do_transaction_failed: handle a memory system error response
54
* (eg "no device/memory present at address") by raising an external abort
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
1
Implement the VLSTM instruction for v7M for the FPU present case.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-16-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20190416125744.27770-25-peter.maydell@linaro.org
6
---
7
---
7
target/arm/cpu.h | 2 +
8
target/arm/helper-a64.h | 7 ++
8
target/arm/helper.h | 2 +
9
target/arm/helper.h | 2 +
9
target/arm/helper.c | 84 ++++++++++++++++++++++++++++++++++++++++++
10
target/arm/mte_helper.c | 194 +++++++++++++++++++++++++++++++++++++
10
target/arm/translate.c | 15 +++++++-
11
target/arm/op_helper.c | 16 +++
11
4 files changed, 102 insertions(+), 1 deletion(-)
12
target/arm/translate-a64.c | 172 +++++++++++++++++++++++++++++++-
13
5 files changed, 386 insertions(+), 5 deletions(-)
12
14
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
14
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.h
17
--- a/target/arm/helper-a64.h
16
+++ b/target/arm/cpu.h
18
+++ b/target/arm/helper-a64.h
17
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
18
#define EXCP_INVSTATE 18 /* v7M INVSTATE UsageFault */
20
19
#define EXCP_STKOF 19 /* v8M STKOF UsageFault */
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
20
#define EXCP_LAZYFP 20 /* v7M fault during lazy FP stacking */
22
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
21
+#define EXCP_LSERR 21 /* v8M LSERR SecureFault */
23
+DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
22
+#define EXCP_UNALIGNED 22 /* v7M UNALIGNED UsageFault */
24
+DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
23
/* NB: add new EXCP_ defines to the array in arm_log_exception() too */
25
+DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
24
26
+DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
25
#define ARMV7M_EXCP_RESET 1
27
+DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
28
+DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
29
+DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
26
diff --git a/target/arm/helper.h b/target/arm/helper.h
30
diff --git a/target/arm/helper.h b/target/arm/helper.h
27
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/helper.h
32
--- a/target/arm/helper.h
29
+++ b/target/arm/helper.h
33
+++ b/target/arm/helper.h
30
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_3(v7m_tt, i32, env, i32, i32)
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
31
35
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
32
DEF_HELPER_1(v7m_preserve_fp_state, void, env)
36
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
33
37
34
+DEF_HELPER_2(v7m_vlstm, void, env, i32)
38
+DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
35
+
39
+
36
DEF_HELPER_2(v8m_stackcheck, void, env, i32)
40
DEF_HELPER_1(vfp_get_fpscr, i32, env)
37
41
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
38
DEF_HELPER_4(access_check_cp_reg, void, env, ptr, i32, i32)
42
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
40
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/helper.c
45
--- a/target/arm/mte_helper.c
42
+++ b/target/arm/helper.c
46
+++ b/target/arm/mte_helper.c
43
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_preserve_fp_state)(CPUARMState *env)
47
@@ -XXX,XX +XXX,XX @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
44
g_assert_not_reached();
48
return tag;
45
}
49
}
46
50
47
+void HELPER(v7m_vlstm)(CPUARMState *env, uint32_t fptr)
51
+/**
48
+{
52
+ * allocation_tag_mem:
49
+ /* translate.c should never generate calls here in user-only mode */
53
+ * @env: the cpu environment
50
+ g_assert_not_reached();
54
+ * @ptr_mmu_idx: the addressing regime to use for the virtual address
51
+}
55
+ * @ptr: the virtual address for which to look up tag memory
52
+
56
+ * @ptr_access: the access to use for the virtual address
53
uint32_t HELPER(v7m_tt)(CPUARMState *env, uint32_t addr, uint32_t op)
57
+ * @ptr_size: the number of bytes in the normal memory access
58
+ * @tag_access: the access to use for the tag memory
59
+ * @tag_size: the number of bytes in the tag memory access
60
+ * @ra: the return address for exception handling
61
+ *
62
+ * Our tag memory is formatted as a sequence of little-endian nibbles.
63
+ * That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
64
+ * tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
65
+ * for the higher addr.
66
+ *
67
+ * Here, resolve the physical address from the virtual address, and return
68
+ * a pointer to the corresponding tag byte. Exit with exception if the
69
+ * virtual address is not accessible for @ptr_access.
70
+ *
71
+ * The @ptr_size and @tag_size values may not have an obvious relation
72
+ * due to the alignment of @ptr, and the number of tag checks required.
73
+ *
74
+ * If there is no tag storage corresponding to @ptr, return NULL.
75
+ */
76
+static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
77
+ uint64_t ptr, MMUAccessType ptr_access,
78
+ int ptr_size, MMUAccessType tag_access,
79
+ int tag_size, uintptr_t ra)
80
+{
81
+ /* Tag storage not implemented. */
82
+ return NULL;
83
+}
84
+
85
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
54
{
86
{
55
/* The TT instructions can be used by unprivileged code, but in
87
int rtag;
56
@@ -XXX,XX +XXX,XX @@ static void v7m_update_fpccr(CPUARMState *env, uint32_t frameptr,
88
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
89
90
return address_with_allocation_tag(ptr + offset, rtag);
91
}
92
+
93
+static int load_tag1(uint64_t ptr, uint8_t *mem)
94
+{
95
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
96
+ return extract32(*mem, ofs, 4);
97
+}
98
+
99
+uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
100
+{
101
+ int mmu_idx = cpu_mmu_index(env, false);
102
+ uint8_t *mem;
103
+ int rtag = 0;
104
+
105
+ /* Trap if accessing an invalid page. */
106
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
107
+ MMU_DATA_LOAD, 1, GETPC());
108
+
109
+ /* Load if page supports tags. */
110
+ if (mem) {
111
+ rtag = load_tag1(ptr, mem);
112
+ }
113
+
114
+ return address_with_allocation_tag(xt, rtag);
115
+}
116
+
117
+static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
118
+{
119
+ if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
120
+ arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
121
+ cpu_mmu_index(env, false), ra);
122
+ g_assert_not_reached();
123
+ }
124
+}
125
+
126
+/* For use in a non-parallel context, store to the given nibble. */
127
+static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
128
+{
129
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
130
+ *mem = deposit32(*mem, ofs, 4, tag);
131
+}
132
+
133
+/* For use in a parallel context, atomically store to the given nibble. */
134
+static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
135
+{
136
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
137
+ uint8_t old = atomic_read(mem);
138
+
139
+ while (1) {
140
+ uint8_t new = deposit32(old, ofs, 4, tag);
141
+ uint8_t cmp = atomic_cmpxchg(mem, old, new);
142
+ if (likely(cmp == old)) {
143
+ return;
144
+ }
145
+ old = cmp;
146
+ }
147
+}
148
+
149
+typedef void stg_store1(uint64_t, uint8_t *, int);
150
+
151
+static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
152
+ uintptr_t ra, stg_store1 store1)
153
+{
154
+ int mmu_idx = cpu_mmu_index(env, false);
155
+ uint8_t *mem;
156
+
157
+ check_tag_aligned(env, ptr, ra);
158
+
159
+ /* Trap if accessing an invalid page. */
160
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
161
+ MMU_DATA_STORE, 1, ra);
162
+
163
+ /* Store if page supports tags. */
164
+ if (mem) {
165
+ store1(ptr, mem, allocation_tag_from_addr(xt));
166
+ }
167
+}
168
+
169
+void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
170
+{
171
+ do_stg(env, ptr, xt, GETPC(), store_tag1);
172
+}
173
+
174
+void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
175
+{
176
+ do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
177
+}
178
+
179
+void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
180
+{
181
+ int mmu_idx = cpu_mmu_index(env, false);
182
+ uintptr_t ra = GETPC();
183
+
184
+ check_tag_aligned(env, ptr, ra);
185
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
186
+}
187
+
188
+static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
189
+ uintptr_t ra, stg_store1 store1)
190
+{
191
+ int mmu_idx = cpu_mmu_index(env, false);
192
+ int tag = allocation_tag_from_addr(xt);
193
+ uint8_t *mem1, *mem2;
194
+
195
+ check_tag_aligned(env, ptr, ra);
196
+
197
+ /*
198
+ * Trap if accessing an invalid page(s).
199
+ * This takes priority over !allocation_tag_access_enabled.
200
+ */
201
+ if (ptr & TAG_GRANULE) {
202
+ /* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
203
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
204
+ TAG_GRANULE, MMU_DATA_STORE, 1, ra);
205
+ mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
206
+ MMU_DATA_STORE, TAG_GRANULE,
207
+ MMU_DATA_STORE, 1, ra);
208
+
209
+ /* Store if page(s) support tags. */
210
+ if (mem1) {
211
+ store1(TAG_GRANULE, mem1, tag);
212
+ }
213
+ if (mem2) {
214
+ store1(0, mem2, tag);
215
+ }
216
+ } else {
217
+ /* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
218
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
219
+ 2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
220
+ if (mem1) {
221
+ tag |= tag << 4;
222
+ atomic_set(mem1, tag);
223
+ }
224
+ }
225
+}
226
+
227
+void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
228
+{
229
+ do_st2g(env, ptr, xt, GETPC(), store_tag1);
230
+}
231
+
232
+void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
233
+{
234
+ do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
235
+}
236
+
237
+void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
238
+{
239
+ int mmu_idx = cpu_mmu_index(env, false);
240
+ uintptr_t ra = GETPC();
241
+ int in_page = -(ptr | TARGET_PAGE_MASK);
242
+
243
+ check_tag_aligned(env, ptr, ra);
244
+
245
+ if (likely(in_page >= 2 * TAG_GRANULE)) {
246
+ probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
247
+ } else {
248
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
249
+ probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
250
+ }
251
+}
252
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
253
index XXXXXXX..XXXXXXX 100644
254
--- a/target/arm/op_helper.c
255
+++ b/target/arm/op_helper.c
256
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
257
return ((uint32_t)x >> shift) | (x << (32 - shift));
57
}
258
}
58
}
259
}
59
260
+
60
+void HELPER(v7m_vlstm)(CPUARMState *env, uint32_t fptr)
261
+void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
61
+{
262
+ uint32_t access_type, uint32_t mmu_idx,
62
+ /* fptr is the value of Rn, the frame pointer we store the FP regs to */
263
+ uint32_t size)
63
+ bool s = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_S_MASK;
264
+{
64
+ bool lspact = env->v7m.fpccr[s] & R_V7M_FPCCR_LSPACT_MASK;
265
+ uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
65
+
266
+ uintptr_t ra = GETPC();
66
+ assert(env->v7m.secure);
267
+
67
+
268
+ if (likely(size <= in_page)) {
68
+ if (!(env->v7m.control[M_REG_S] & R_V7M_CONTROL_SFPA_MASK)) {
269
+ probe_access(env, ptr, size, access_type, mmu_idx, ra);
270
+ } else {
271
+ probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
272
+ probe_access(env, ptr + in_page, size - in_page,
273
+ access_type, mmu_idx, ra);
274
+ }
275
+}
276
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
277
index XXXXXXX..XXXXXXX 100644
278
--- a/target/arm/translate-a64.c
279
+++ b/target/arm/translate-a64.c
280
@@ -XXX,XX +XXX,XX @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
281
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
282
}
283
284
+static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
285
+ MMUAccessType acc, int log2_size)
286
+{
287
+ TCGv_i32 t_acc = tcg_const_i32(acc);
288
+ TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
289
+ TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
290
+
291
+ gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
292
+ tcg_temp_free_i32(t_acc);
293
+ tcg_temp_free_i32(t_idx);
294
+ tcg_temp_free_i32(t_size);
295
+}
296
+
297
typedef struct DisasCompare64 {
298
TCGCond cond;
299
TCGv_i64 value;
300
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
301
}
302
}
303
304
+/*
305
+ * Load/Store memory tags
306
+ *
307
+ * 31 30 29 24 22 21 12 10 5 0
308
+ * +-----+-------------+-----+---+------+-----+------+------+
309
+ * | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
310
+ * +-----+-------------+-----+---+------+-----+------+------+
311
+ */
312
+static void disas_ldst_tag(DisasContext *s, uint32_t insn)
313
+{
314
+ int rt = extract32(insn, 0, 5);
315
+ int rn = extract32(insn, 5, 5);
316
+ uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
317
+ int op2 = extract32(insn, 10, 2);
318
+ int op1 = extract32(insn, 22, 2);
319
+ bool is_load = false, is_pair = false, is_zero = false;
320
+ int index = 0;
321
+ TCGv_i64 addr, clean_addr, tcg_rt;
322
+
323
+ /* We checked insn bits [29:24,21] in the caller. */
324
+ if (extract32(insn, 30, 2) != 3) {
325
+ goto do_unallocated;
326
+ }
327
+
328
+ /*
329
+ * @index is a tri-state variable which has 3 states:
330
+ * < 0 : post-index, writeback
331
+ * = 0 : signed offset
332
+ * > 0 : pre-index, writeback
333
+ */
334
+ switch (op1) {
335
+ case 0:
336
+ if (op2 != 0) {
337
+ /* STG */
338
+ index = op2 - 2;
339
+ break;
340
+ }
341
+ goto do_unallocated;
342
+ case 1:
343
+ if (op2 != 0) {
344
+ /* STZG */
345
+ is_zero = true;
346
+ index = op2 - 2;
347
+ } else {
348
+ /* LDG */
349
+ is_load = true;
350
+ }
351
+ break;
352
+ case 2:
353
+ if (op2 != 0) {
354
+ /* ST2G */
355
+ is_pair = true;
356
+ index = op2 - 2;
357
+ break;
358
+ }
359
+ goto do_unallocated;
360
+ case 3:
361
+ if (op2 != 0) {
362
+ /* STZ2G */
363
+ is_pair = is_zero = true;
364
+ index = op2 - 2;
365
+ break;
366
+ }
367
+ goto do_unallocated;
368
+
369
+ default:
370
+ do_unallocated:
371
+ unallocated_encoding(s);
69
+ return;
372
+ return;
70
+ }
373
+ }
71
+
374
+
72
+ /* Check access to the coprocessor is permitted */
375
+ if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
73
+ if (!v7m_cpacr_pass(env, true, arm_current_el(env) != 0)) {
376
+ goto do_unallocated;
74
+ raise_exception_ra(env, EXCP_NOCP, 0, 1, GETPC());
377
+ }
75
+ }
378
+
76
+
379
+ if (rn == 31) {
77
+ if (lspact) {
380
+ gen_check_sp_alignment(s);
78
+ /* LSPACT should not be active when there is active FP state */
381
+ }
79
+ raise_exception_ra(env, EXCP_LSERR, 0, 1, GETPC());
382
+
80
+ }
383
+ addr = read_cpu_reg_sp(s, rn, true);
81
+
384
+ if (index >= 0) {
82
+ if (fptr & 7) {
385
+ /* pre-index or signed offset */
83
+ raise_exception_ra(env, EXCP_UNALIGNED, 0, 1, GETPC());
386
+ tcg_gen_addi_i64(addr, addr, offset);
84
+ }
387
+ }
85
+
388
+
86
+ /*
389
+ if (is_load) {
87
+ * Note that we do not use v7m_stack_write() here, because the
390
+ tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
88
+ * accesses should not set the FSR bits for stacking errors if they
391
+ tcg_rt = cpu_reg(s, rt);
89
+ * fail. (In pseudocode terms, they are AccType_NORMAL, not AccType_STACK
392
+ if (s->ata) {
90
+ * or AccType_LAZYFP). Faults in cpu_stl_data() will throw exceptions
393
+ gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
91
+ * and longjmp out.
394
+ } else {
92
+ */
395
+ clean_addr = clean_data_tbi(s, addr);
93
+ if (!(env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_LSPEN_MASK)) {
396
+ gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
94
+ bool ts = env->v7m.fpccr[M_REG_S] & R_V7M_FPCCR_TS_MASK;
397
+ gen_address_with_allocation_tag0(tcg_rt, addr);
95
+ int i;
398
+ }
96
+
399
+ } else {
97
+ for (i = 0; i < (ts ? 32 : 16); i += 2) {
400
+ tcg_rt = cpu_reg_sp(s, rt);
98
+ uint64_t dn = *aa32_vfp_dreg(env, i / 2);
401
+ if (!s->ata) {
99
+ uint32_t faddr = fptr + 4 * i;
402
+ /*
100
+ uint32_t slo = extract64(dn, 0, 32);
403
+ * For STG and ST2G, we need to check alignment and probe memory.
101
+ uint32_t shi = extract64(dn, 32, 32);
404
+ * TODO: For STZG and STZ2G, we could rely on the stores below,
102
+
405
+ * at least for system mode; user-only won't enforce alignment.
103
+ if (i >= 16) {
406
+ */
104
+ faddr += 8; /* skip the slot for the FPSCR */
407
+ if (is_pair) {
408
+ gen_helper_st2g_stub(cpu_env, addr);
409
+ } else {
410
+ gen_helper_stg_stub(cpu_env, addr);
105
+ }
411
+ }
106
+ cpu_stl_data(env, faddr, slo);
412
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
107
+ cpu_stl_data(env, faddr + 4, shi);
413
+ if (is_pair) {
108
+ }
414
+ gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
109
+ cpu_stl_data(env, fptr + 0x40, vfp_get_fpscr(env));
415
+ } else {
110
+
416
+ gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
111
+ /*
112
+ * If TS is 0 then s0 to s15 and FPSCR are UNKNOWN; we choose to
113
+ * leave them unchanged, matching our choice in v7m_preserve_fp_state.
114
+ */
115
+ if (ts) {
116
+ for (i = 0; i < 32; i += 2) {
117
+ *aa32_vfp_dreg(env, i / 2) = 0;
118
+ }
417
+ }
119
+ vfp_set_fpscr(env, 0);
418
+ } else {
120
+ }
419
+ if (is_pair) {
121
+ } else {
420
+ gen_helper_st2g(cpu_env, addr, tcg_rt);
122
+ v7m_update_fpccr(env, fptr, false);
421
+ } else {
123
+ }
422
+ gen_helper_stg(cpu_env, addr, tcg_rt);
124
+
423
+ }
125
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_FPCA_MASK;
424
+ }
126
+}
425
+ }
127
+
426
+
128
static bool v7m_push_stack(ARMCPU *cpu)
427
+ if (is_zero) {
428
+ TCGv_i64 clean_addr = clean_data_tbi(s, addr);
429
+ TCGv_i64 tcg_zero = tcg_const_i64(0);
430
+ int mem_index = get_mem_index(s);
431
+ int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
432
+
433
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
434
+ MO_Q | MO_ALIGN_16);
435
+ for (i = 8; i < n; i += 8) {
436
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
437
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
438
+ }
439
+ tcg_temp_free_i64(tcg_zero);
440
+ }
441
+
442
+ if (index != 0) {
443
+ /* pre-index or post-index */
444
+ if (index < 0) {
445
+ /* post-index */
446
+ tcg_gen_addi_i64(addr, addr, offset);
447
+ }
448
+ tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
449
+ }
450
+}
451
+
452
/* Loads and stores */
453
static void disas_ldst(DisasContext *s, uint32_t insn)
129
{
454
{
130
/* Do the "set up stack frame" part of exception entry,
455
@@ -XXX,XX +XXX,XX @@ static void disas_ldst(DisasContext *s, uint32_t insn)
131
@@ -XXX,XX +XXX,XX @@ static void arm_log_exception(int idx)
456
case 0x0d: /* AdvSIMD load/store single structure */
132
[EXCP_INVSTATE] = "v7M INVSTATE UsageFault",
457
disas_ldst_single_struct(s, insn);
133
[EXCP_STKOF] = "v8M STKOF UsageFault",
134
[EXCP_LAZYFP] = "v7M exception during lazy FP stacking",
135
+ [EXCP_LSERR] = "v8M LSERR UsageFault",
136
+ [EXCP_UNALIGNED] = "v7M UNALIGNED UsageFault",
137
};
138
139
if (idx >= 0 && idx < ARRAY_SIZE(excnames)) {
140
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
141
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, env->v7m.secure);
142
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_STKOF_MASK;
143
break;
458
break;
144
+ case EXCP_LSERR:
459
- case 0x19: /* LDAPR/STLR (unscaled immediate) */
145
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
460
- if (extract32(insn, 10, 2) != 0 ||
146
+ env->v7m.sfsr |= R_V7M_SFSR_LSERR_MASK;
461
- extract32(insn, 21, 1) != 0) {
147
+ break;
462
+ case 0x19:
148
+ case EXCP_UNALIGNED:
463
+ if (extract32(insn, 21, 1) != 0) {
149
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, env->v7m.secure);
464
+ disas_ldst_tag(s, insn);
150
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_UNALIGNED_MASK;
465
+ } else if (extract32(insn, 10, 2) == 0) {
151
+ break;
466
+ disas_ldst_ldapr_stlr(s, insn);
152
case EXCP_SWI:
467
+ } else {
153
/* The PC already points to the next instruction. */
468
unallocated_encoding(s);
154
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SVC, env->v7m.secure);
469
- break;
155
diff --git a/target/arm/translate.c b/target/arm/translate.c
470
}
156
index XXXXXXX..XXXXXXX 100644
471
- disas_ldst_ldapr_stlr(s, insn);
157
--- a/target/arm/translate.c
472
break;
158
+++ b/target/arm/translate.c
473
default:
159
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
474
unallocated_encoding(s);
160
if (!s->v8m_secure || (insn & 0x0040f0ff)) {
161
goto illegal_op;
162
}
163
- /* Just NOP since FP support is not implemented */
164
+
165
+ if (arm_dc_feature(s, ARM_FEATURE_VFP)) {
166
+ TCGv_i32 fptr = load_reg(s, rn);
167
+
168
+ if (extract32(insn, 20, 1)) {
169
+ /* VLLDM */
170
+ } else {
171
+ gen_helper_v7m_vlstm(cpu_env, fptr);
172
+ }
173
+ tcg_temp_free_i32(fptr);
174
+
175
+ /* End the TB, because we have updated FP control bits */
176
+ s->base.is_jmp = DISAS_UPDATE;
177
+ }
178
break;
179
}
180
if (arm_dc_feature(s, ARM_FEATURE_VFP) &&
181
--
475
--
182
2.20.1
476
2.20.1
183
477
184
478
diff view generated by jsdifflib
1
Enforce that for M-profile various FPSCR bits which are RES0 there
1
From: Richard Henderson <richard.henderson@linaro.org>
2
but have defined meanings on A-profile are never settable. This
3
ensures that M-profile code can't enable the A-profile behaviour
4
(notably vector length/stride handling) by accident.
5
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-17-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20190416125744.27770-2-peter.maydell@linaro.org
9
---
7
---
10
target/arm/vfp_helper.c | 8 ++++++++
8
target/arm/translate-a64.c | 29 ++++++++++++++++++++++++++---
11
1 file changed, 8 insertions(+)
9
1 file changed, 26 insertions(+), 3 deletions(-)
12
10
13
diff --git a/target/arm/vfp_helper.c b/target/arm/vfp_helper.c
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
14
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/vfp_helper.c
13
--- a/target/arm/translate-a64.c
16
+++ b/target/arm/vfp_helper.c
14
+++ b/target/arm/translate-a64.c
17
@@ -XXX,XX +XXX,XX @@ void HELPER(vfp_set_fpscr)(CPUARMState *env, uint32_t val)
15
@@ -XXX,XX +XXX,XX @@ static void disas_ld_lit(DisasContext *s, uint32_t insn)
18
val &= ~FPCR_FZ16;
16
* +-----+-------+---+---+-------+---+-------+-------+------+------+
17
*
18
* opc: LDP/STP/LDNP/STNP 00 -> 32 bit, 10 -> 64 bit
19
- * LDPSW 01
20
+ * LDPSW/STGP 01
21
* LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
22
* V: 0 -> GPR, 1 -> Vector
23
* idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
24
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
25
bool is_signed = false;
26
bool postindex = false;
27
bool wback = false;
28
+ bool set_tag = false;
29
30
TCGv_i64 clean_addr, dirty_addr;
31
32
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
33
34
if (is_vector) {
35
size = 2 + opc;
36
+ } else if (opc == 1 && !is_load) {
37
+ /* STGP */
38
+ if (!dc_isar_feature(aa64_mte_insn_reg, s) || index == 0) {
39
+ unallocated_encoding(s);
40
+ return;
41
+ }
42
+ size = 3;
43
+ set_tag = true;
44
} else {
45
size = 2 + extract32(opc, 1, 1);
46
is_signed = extract32(opc, 0, 1);
47
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
48
return;
19
}
49
}
20
50
21
+ if (arm_feature(env, ARM_FEATURE_M)) {
51
- offset <<= size;
22
+ /*
52
+ offset <<= (set_tag ? LOG2_TAG_GRANULE : size);
23
+ * M profile FPSCR is RES0 for the QC, STRIDE, FZ16, LEN bits
53
24
+ * and also for the trapped-exception-handling bits IxE.
54
if (rn == 31) {
25
+ */
55
gen_check_sp_alignment(s);
26
+ val &= 0xf7c0009f;
56
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
57
if (!postindex) {
58
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
59
}
60
- clean_addr = clean_data_tbi(s, dirty_addr);
61
62
+ if (set_tag) {
63
+ if (!s->ata) {
64
+ /*
65
+ * TODO: We could rely on the stores below, at least for
66
+ * system mode, if we arrange to add MO_ALIGN_16.
67
+ */
68
+ gen_helper_stg_stub(cpu_env, dirty_addr);
69
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
70
+ gen_helper_stg_parallel(cpu_env, dirty_addr, dirty_addr);
71
+ } else {
72
+ gen_helper_stg(cpu_env, dirty_addr, dirty_addr);
73
+ }
27
+ }
74
+ }
28
+
75
+
29
/*
76
+ clean_addr = clean_data_tbi(s, dirty_addr);
30
* We don't implement trapped exception handling, so the
77
if (is_vector) {
31
* trap enable bits, IDE|IXE|UFE|OFE|DZE|IOE are all RAZ/WI (not RES0!)
78
if (is_load) {
79
do_fp_ld(s, rt, clean_addr, size);
32
--
80
--
33
2.20.1
81
2.20.1
34
82
35
83
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
We can simplify our DC_ZVA if we recognize that the largest BS
4
that we actually use in system mode is 64. Let us just assert
5
that it fits within TARGET_PAGE_SIZE.
6
7
For DC_GVA and STZGM, we want to be able to write whole bytes
8
of tag memory, so assert that BS is >= 2 * TAG_GRANULE, or 32.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-18-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/cpu.c | 24 ++++++++++++++++++++++++
16
1 file changed, 24 insertions(+)
17
18
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.c
21
+++ b/target/arm/cpu.c
22
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
23
}
24
#endif
25
26
+ if (tcg_enabled()) {
27
+ int dcz_blocklen = 4 << cpu->dcz_blocksize;
28
+
29
+ /*
30
+ * We only support DCZ blocklen that fits on one page.
31
+ *
32
+ * Architectually this is always true. However TARGET_PAGE_SIZE
33
+ * is variable and, for compatibility with -machine virt-2.7,
34
+ * is only 1KiB, as an artifact of legacy ARMv5 subpage support.
35
+ * But even then, while the largest architectural DCZ blocklen
36
+ * is 2KiB, no cpu actually uses such a large blocklen.
37
+ */
38
+ assert(dcz_blocklen <= TARGET_PAGE_SIZE);
39
+
40
+ /*
41
+ * We only support DCZ blocksize >= 2*TAG_GRANULE, which is to say
42
+ * both nibbles of each byte storing tag data may be written at once.
43
+ * Since TAG_GRANULE is 16, this means that blocklen must be >= 32.
44
+ */
45
+ if (cpu_isar_feature(aa64_mte, cpu)) {
46
+ assert(dcz_blocklen >= 2 * TAG_GRANULE);
47
+ }
48
+ }
49
+
50
qemu_init_vcpu(cs);
51
cpu_reset(cs);
52
53
--
54
2.20.1
55
56
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Now that we know that the operation is on a single page,
4
we need not loop over pages while probing.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-19-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-a64.c | 94 +++++++++++------------------------------
12
1 file changed, 25 insertions(+), 69 deletions(-)
13
14
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.c
17
+++ b/target/arm/helper-a64.c
18
@@ -XXX,XX +XXX,XX @@ void HELPER(dc_zva)(CPUARMState *env, uint64_t vaddr_in)
19
* (which matches the usual QEMU behaviour of not implementing either
20
* alignment faults or any memory attribute handling).
21
*/
22
-
23
- ARMCPU *cpu = env_archcpu(env);
24
- uint64_t blocklen = 4 << cpu->dcz_blocksize;
25
+ int blocklen = 4 << env_archcpu(env)->dcz_blocksize;
26
uint64_t vaddr = vaddr_in & ~(blocklen - 1);
27
+ int mmu_idx = cpu_mmu_index(env, false);
28
+ void *mem;
29
+
30
+ /*
31
+ * Trapless lookup. In addition to actual invalid page, may
32
+ * return NULL for I/O, watchpoints, clean pages, etc.
33
+ */
34
+ mem = tlb_vaddr_to_host(env, vaddr, MMU_DATA_STORE, mmu_idx);
35
36
#ifndef CONFIG_USER_ONLY
37
- {
38
+ if (unlikely(!mem)) {
39
+ uintptr_t ra = GETPC();
40
+
41
/*
42
- * Slightly awkwardly, QEMU's TARGET_PAGE_SIZE may be less than
43
- * the block size so we might have to do more than one TLB lookup.
44
- * We know that in fact for any v8 CPU the page size is at least 4K
45
- * and the block size must be 2K or less, but TARGET_PAGE_SIZE is only
46
- * 1K as an artefact of legacy v5 subpage support being present in the
47
- * same QEMU executable. So in practice the hostaddr[] array has
48
- * two entries, given the current setting of TARGET_PAGE_BITS_MIN.
49
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
50
+ * the original pointer for an invalid page. But watchpoints require
51
+ * that we probe the actual space. So do both.
52
*/
53
- int maxidx = DIV_ROUND_UP(blocklen, TARGET_PAGE_SIZE);
54
- void *hostaddr[DIV_ROUND_UP(2 * KiB, 1 << TARGET_PAGE_BITS_MIN)];
55
- int try, i;
56
- unsigned mmu_idx = cpu_mmu_index(env, false);
57
- TCGMemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
58
+ (void) probe_write(env, vaddr_in, 1, mmu_idx, ra);
59
+ mem = probe_write(env, vaddr, blocklen, mmu_idx, ra);
60
61
- assert(maxidx <= ARRAY_SIZE(hostaddr));
62
-
63
- for (try = 0; try < 2; try++) {
64
-
65
- for (i = 0; i < maxidx; i++) {
66
- hostaddr[i] = tlb_vaddr_to_host(env,
67
- vaddr + TARGET_PAGE_SIZE * i,
68
- 1, mmu_idx);
69
- if (!hostaddr[i]) {
70
- break;
71
- }
72
- }
73
- if (i == maxidx) {
74
- /*
75
- * If it's all in the TLB it's fair game for just writing to;
76
- * we know we don't need to update dirty status, etc.
77
- */
78
- for (i = 0; i < maxidx - 1; i++) {
79
- memset(hostaddr[i], 0, TARGET_PAGE_SIZE);
80
- }
81
- memset(hostaddr[i], 0, blocklen - (i * TARGET_PAGE_SIZE));
82
- return;
83
- }
84
+ if (unlikely(!mem)) {
85
/*
86
- * OK, try a store and see if we can populate the tlb. This
87
- * might cause an exception if the memory isn't writable,
88
- * in which case we will longjmp out of here. We must for
89
- * this purpose use the actual register value passed to us
90
- * so that we get the fault address right.
91
+ * The only remaining reason for mem == NULL is I/O.
92
+ * Just do a series of byte writes as the architecture demands.
93
*/
94
- helper_ret_stb_mmu(env, vaddr_in, 0, oi, GETPC());
95
- /* Now we can populate the other TLB entries, if any */
96
- for (i = 0; i < maxidx; i++) {
97
- uint64_t va = vaddr + TARGET_PAGE_SIZE * i;
98
- if (va != (vaddr_in & TARGET_PAGE_MASK)) {
99
- helper_ret_stb_mmu(env, va, 0, oi, GETPC());
100
- }
101
+ for (int i = 0; i < blocklen; i++) {
102
+ cpu_stb_mmuidx_ra(env, vaddr + i, 0, mmu_idx, ra);
103
}
104
- }
105
-
106
- /*
107
- * Slow path (probably attempt to do this to an I/O device or
108
- * similar, or clearing of a block of code we have translations
109
- * cached for). Just do a series of byte writes as the architecture
110
- * demands. It's not worth trying to use a cpu_physical_memory_map(),
111
- * memset(), unmap() sequence here because:
112
- * + we'd need to account for the blocksize being larger than a page
113
- * + the direct-RAM access case is almost always going to be dealt
114
- * with in the fastpath code above, so there's no speed benefit
115
- * + we would have to deal with the map returning NULL because the
116
- * bounce buffer was in use
117
- */
118
- for (i = 0; i < blocklen; i++) {
119
- helper_ret_stb_mmu(env, vaddr + i, 0, oi, GETPC());
120
+ return;
121
}
122
}
123
-#else
124
- memset(g2h(vaddr), 0, blocklen);
125
#endif
126
+
127
+ memset(mem, 0, blocklen);
128
}
129
--
130
2.20.1
131
132
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-20-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-a64.h | 3 ++
9
target/arm/translate.h | 2 +
10
target/arm/mte_helper.c | 84 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-a64.c | 72 ++++++++++++++++++++++++++++----
12
4 files changed, 153 insertions(+), 8 deletions(-)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
19
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
20
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
21
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
22
+DEF_HELPER_FLAGS_2(ldgm, TCG_CALL_NO_WG, i64, env, i64)
23
+DEF_HELPER_FLAGS_3(stgm, TCG_CALL_NO_WG, void, env, i64, i64)
24
+DEF_HELPER_FLAGS_3(stzgm_tags, TCG_CALL_NO_WG, void, env, i64, i64)
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/translate.h
28
+++ b/target/arm/translate.h
29
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
30
* < 0, set by the current instruction.
31
*/
32
int8_t btype;
33
+ /* A copy of cpu->dcz_blocksize. */
34
+ uint8_t dcz_blocksize;
35
/* True if this page is guarded. */
36
bool guarded_page;
37
/* Bottom two bits of XScale c15_cpar coprocessor access control reg */
38
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/target/arm/mte_helper.c
41
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@ void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
43
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
44
}
45
}
46
+
47
+#define LDGM_STGM_SIZE (4 << GMID_EL1_BS)
48
+
49
+uint64_t HELPER(ldgm)(CPUARMState *env, uint64_t ptr)
50
+{
51
+ int mmu_idx = cpu_mmu_index(env, false);
52
+ uintptr_t ra = GETPC();
53
+ void *tag_mem;
54
+
55
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
56
+
57
+ /* Trap if accessing an invalid page. */
58
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD,
59
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
60
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
61
+
62
+ /* The tag is squashed to zero if the page does not support tags. */
63
+ if (!tag_mem) {
64
+ return 0;
65
+ }
66
+
67
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
68
+ /*
69
+ * We are loading 64-bits worth of tags. The ordering of elements
70
+ * within the word corresponds to a 64-bit little-endian operation.
71
+ */
72
+ return ldq_le_p(tag_mem);
73
+}
74
+
75
+void HELPER(stgm)(CPUARMState *env, uint64_t ptr, uint64_t val)
76
+{
77
+ int mmu_idx = cpu_mmu_index(env, false);
78
+ uintptr_t ra = GETPC();
79
+ void *tag_mem;
80
+
81
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
82
+
83
+ /* Trap if accessing an invalid page. */
84
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
85
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
86
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
87
+
88
+ /*
89
+ * Tag store only happens if the page support tags,
90
+ * and if the OS has enabled access to the tags.
91
+ */
92
+ if (!tag_mem) {
93
+ return;
94
+ }
95
+
96
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
97
+ /*
98
+ * We are storing 64-bits worth of tags. The ordering of elements
99
+ * within the word corresponds to a 64-bit little-endian operation.
100
+ */
101
+ stq_le_p(tag_mem, val);
102
+}
103
+
104
+void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
105
+{
106
+ uintptr_t ra = GETPC();
107
+ int mmu_idx = cpu_mmu_index(env, false);
108
+ int log2_dcz_bytes, log2_tag_bytes;
109
+ intptr_t dcz_bytes, tag_bytes;
110
+ uint8_t *mem;
111
+
112
+ /*
113
+ * In arm_cpu_realizefn, we assert that dcz > LOG2_TAG_GRANULE+1,
114
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway,
115
+ * to make sure that we can access one complete tag byte here.
116
+ */
117
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
118
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
119
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
120
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
121
+ ptr &= -dcz_bytes;
122
+
123
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, dcz_bytes,
124
+ MMU_DATA_STORE, tag_bytes, ra);
125
+ if (mem) {
126
+ int tag_pair = (val & 0xf) * 0x11;
127
+ memset(mem, tag_pair, tag_bytes);
128
+ }
129
+}
130
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
131
index XXXXXXX..XXXXXXX 100644
132
--- a/target/arm/translate-a64.c
133
+++ b/target/arm/translate-a64.c
134
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
135
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
136
int op2 = extract32(insn, 10, 2);
137
int op1 = extract32(insn, 22, 2);
138
- bool is_load = false, is_pair = false, is_zero = false;
139
+ bool is_load = false, is_pair = false, is_zero = false, is_mult = false;
140
int index = 0;
141
TCGv_i64 addr, clean_addr, tcg_rt;
142
143
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
144
if (op2 != 0) {
145
/* STG */
146
index = op2 - 2;
147
- break;
148
+ } else {
149
+ /* STZGM */
150
+ if (s->current_el == 0 || offset != 0) {
151
+ goto do_unallocated;
152
+ }
153
+ is_mult = is_zero = true;
154
}
155
- goto do_unallocated;
156
+ break;
157
case 1:
158
if (op2 != 0) {
159
/* STZG */
160
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
161
/* ST2G */
162
is_pair = true;
163
index = op2 - 2;
164
- break;
165
+ } else {
166
+ /* STGM */
167
+ if (s->current_el == 0 || offset != 0) {
168
+ goto do_unallocated;
169
+ }
170
+ is_mult = true;
171
}
172
- goto do_unallocated;
173
+ break;
174
case 3:
175
if (op2 != 0) {
176
/* STZ2G */
177
is_pair = is_zero = true;
178
index = op2 - 2;
179
- break;
180
+ } else {
181
+ /* LDGM */
182
+ if (s->current_el == 0 || offset != 0) {
183
+ goto do_unallocated;
184
+ }
185
+ is_mult = is_load = true;
186
}
187
- goto do_unallocated;
188
+ break;
189
190
default:
191
do_unallocated:
192
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
193
return;
194
}
195
196
- if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
197
+ if (is_mult
198
+ ? !dc_isar_feature(aa64_mte, s)
199
+ : !dc_isar_feature(aa64_mte_insn_reg, s)) {
200
goto do_unallocated;
201
}
202
203
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
204
tcg_gen_addi_i64(addr, addr, offset);
205
}
206
207
+ if (is_mult) {
208
+ tcg_rt = cpu_reg(s, rt);
209
+
210
+ if (is_zero) {
211
+ int size = 4 << s->dcz_blocksize;
212
+
213
+ if (s->ata) {
214
+ gen_helper_stzgm_tags(cpu_env, addr, tcg_rt);
215
+ }
216
+ /*
217
+ * The non-tags portion of STZGM is mostly like DC_ZVA,
218
+ * except the alignment happens before the access.
219
+ */
220
+ clean_addr = clean_data_tbi(s, addr);
221
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
222
+ gen_helper_dc_zva(cpu_env, clean_addr);
223
+ } else if (s->ata) {
224
+ if (is_load) {
225
+ gen_helper_ldgm(tcg_rt, cpu_env, addr);
226
+ } else {
227
+ gen_helper_stgm(cpu_env, addr, tcg_rt);
228
+ }
229
+ } else {
230
+ MMUAccessType acc = is_load ? MMU_DATA_LOAD : MMU_DATA_STORE;
231
+ int size = 4 << GMID_EL1_BS;
232
+
233
+ clean_addr = clean_data_tbi(s, addr);
234
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
235
+ gen_probe_access(s, clean_addr, acc, size);
236
+
237
+ if (is_load) {
238
+ /* The result tags are zeros. */
239
+ tcg_gen_movi_i64(tcg_rt, 0);
240
+ }
241
+ }
242
+ return;
243
+ }
244
+
245
if (is_load) {
246
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
247
tcg_rt = cpu_reg(s, rt);
248
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
249
dc->vec_stride = 0;
250
dc->cp_regs = arm_cpu->cp_regs;
251
dc->features = env->features;
252
+ dc->dcz_blocksize = arm_cpu->dcz_blocksize;
253
254
/* Single step state. The code-generation logic here is:
255
* SS_ACTIVE == 0:
256
--
257
2.20.1
258
259
diff view generated by jsdifflib
1
For v8M floating point support, transitions from Secure
1
From: Richard Henderson <richard.henderson@linaro.org>
2
to Non-secure state via BLNS and BLXNS must clear the
3
CONTROL.SFPA bit. (This corresponds to the pseudocode
4
BranchToNS() function.)
5
2
3
Like the regular data cache flushes, these are nops within qemu.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-21-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20190416125744.27770-13-peter.maydell@linaro.org
9
---
9
---
10
target/arm/helper.c | 4 ++++
10
target/arm/helper.c | 65 +++++++++++++++++++++++++++++++++++++++++++++
11
1 file changed, 4 insertions(+)
11
1 file changed, 65 insertions(+)
12
12
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
16
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_bxns)(CPUARMState *env, uint32_t dest)
17
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
18
/* translate.c should have made BXNS UNDEF unless we're secure */
18
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
19
assert(env->v7m.secure);
19
.type = ARM_CP_NO_RAW,
20
20
.access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
21
+ if (!(dest & 1)) {
21
+ { .name = "DC_IGVAC", .state = ARM_CP_STATE_AA64,
22
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_SFPA_MASK;
22
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 3,
23
+ }
23
+ .type = ARM_CP_NOP, .access = PL1_W,
24
switch_v7m_security_state(env, dest & 1);
24
+ .accessfn = aa64_cacheop_poc_access },
25
env->thumb = 1;
25
+ { .name = "DC_IGSW", .state = ARM_CP_STATE_AA64,
26
env->regs[15] = dest & ~1;
26
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 4,
27
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_blxns)(CPUARMState *env, uint32_t dest)
27
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
28
*/
28
+ { .name = "DC_IGDVAC", .state = ARM_CP_STATE_AA64,
29
write_v7m_exception(env, 1);
29
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 5,
30
+ .type = ARM_CP_NOP, .access = PL1_W,
31
+ .accessfn = aa64_cacheop_poc_access },
32
+ { .name = "DC_IGDSW", .state = ARM_CP_STATE_AA64,
33
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 6,
34
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
35
+ { .name = "DC_CGSW", .state = ARM_CP_STATE_AA64,
36
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 4,
37
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
38
+ { .name = "DC_CGDSW", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 6,
40
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
41
+ { .name = "DC_CIGSW", .state = ARM_CP_STATE_AA64,
42
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 4,
43
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
44
+ { .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
45
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
46
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
47
REGINFO_SENTINEL
48
};
49
50
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
51
.type = ARM_CP_CONST, .access = PL0_RW, },
52
REGINFO_SENTINEL
53
};
54
+
55
+static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
56
+ { .name = "DC_CGVAC", .state = ARM_CP_STATE_AA64,
57
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 3,
58
+ .type = ARM_CP_NOP, .access = PL0_W,
59
+ .accessfn = aa64_cacheop_poc_access },
60
+ { .name = "DC_CGDVAC", .state = ARM_CP_STATE_AA64,
61
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 5,
62
+ .type = ARM_CP_NOP, .access = PL0_W,
63
+ .accessfn = aa64_cacheop_poc_access },
64
+ { .name = "DC_CGVAP", .state = ARM_CP_STATE_AA64,
65
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 3,
66
+ .type = ARM_CP_NOP, .access = PL0_W,
67
+ .accessfn = aa64_cacheop_poc_access },
68
+ { .name = "DC_CGDVAP", .state = ARM_CP_STATE_AA64,
69
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 5,
70
+ .type = ARM_CP_NOP, .access = PL0_W,
71
+ .accessfn = aa64_cacheop_poc_access },
72
+ { .name = "DC_CGVADP", .state = ARM_CP_STATE_AA64,
73
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 3,
74
+ .type = ARM_CP_NOP, .access = PL0_W,
75
+ .accessfn = aa64_cacheop_poc_access },
76
+ { .name = "DC_CGDVADP", .state = ARM_CP_STATE_AA64,
77
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 5,
78
+ .type = ARM_CP_NOP, .access = PL0_W,
79
+ .accessfn = aa64_cacheop_poc_access },
80
+ { .name = "DC_CIGVAC", .state = ARM_CP_STATE_AA64,
81
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 3,
82
+ .type = ARM_CP_NOP, .access = PL0_W,
83
+ .accessfn = aa64_cacheop_poc_access },
84
+ { .name = "DC_CIGDVAC", .state = ARM_CP_STATE_AA64,
85
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
86
+ .type = ARM_CP_NOP, .access = PL0_W,
87
+ .accessfn = aa64_cacheop_poc_access },
88
+ REGINFO_SENTINEL
89
+};
90
+
91
#endif
92
93
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
94
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
95
*/
96
if (cpu_isar_feature(aa64_mte, cpu)) {
97
define_arm_cp_regs(cpu, mte_reginfo);
98
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
99
} else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
100
define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
101
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
30
}
102
}
31
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_SFPA_MASK;
103
#endif
32
switch_v7m_security_state(env, 0);
104
33
env->thumb = 1;
34
env->regs[15] = dest;
35
--
105
--
36
2.20.1
106
2.20.1
37
107
38
108
diff view generated by jsdifflib
1
If the floating point extension is present, then the SG instruction
1
From: Richard Henderson <richard.henderson@linaro.org>
2
must clear the CONTROL_S.SFPA bit. Implement this.
3
2
4
(On a no-FPU system the bit will always be zero, so we don't need
3
We will shortly need this in mte_helper.c as well.
5
to make the clearing of the bit conditional on ARM_FEATURE_VFP.)
6
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-22-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20190416125744.27770-8-peter.maydell@linaro.org
10
---
9
---
11
target/arm/helper.c | 1 +
10
target/arm/internals.h | 36 ++++++++++++++++++++++++++++++++++++
12
1 file changed, 1 insertion(+)
11
target/arm/helper.c | 36 ------------------------------------
12
2 files changed, 36 insertions(+), 36 deletions(-)
13
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_pan(CPUARMState *env, ARMMMUIdx mmu_idx)
19
}
20
}
21
22
+/* Return the exception level which controls this address translation regime */
23
+static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
24
+{
25
+ switch (mmu_idx) {
26
+ case ARMMMUIdx_E20_0:
27
+ case ARMMMUIdx_E20_2:
28
+ case ARMMMUIdx_E20_2_PAN:
29
+ case ARMMMUIdx_Stage2:
30
+ case ARMMMUIdx_E2:
31
+ return 2;
32
+ case ARMMMUIdx_SE3:
33
+ return 3;
34
+ case ARMMMUIdx_SE10_0:
35
+ return arm_el_is_aa64(env, 3) ? 1 : 3;
36
+ case ARMMMUIdx_SE10_1:
37
+ case ARMMMUIdx_SE10_1_PAN:
38
+ case ARMMMUIdx_Stage1_E0:
39
+ case ARMMMUIdx_Stage1_E1:
40
+ case ARMMMUIdx_Stage1_E1_PAN:
41
+ case ARMMMUIdx_E10_0:
42
+ case ARMMMUIdx_E10_1:
43
+ case ARMMMUIdx_E10_1_PAN:
44
+ case ARMMMUIdx_MPrivNegPri:
45
+ case ARMMMUIdx_MUserNegPri:
46
+ case ARMMMUIdx_MPriv:
47
+ case ARMMMUIdx_MUser:
48
+ case ARMMMUIdx_MSPrivNegPri:
49
+ case ARMMMUIdx_MSUserNegPri:
50
+ case ARMMMUIdx_MSPriv:
51
+ case ARMMMUIdx_MSUser:
52
+ return 1;
53
+ default:
54
+ g_assert_not_reached();
55
+ }
56
+}
57
+
58
/* Return the FSR value for a debug exception (watchpoint, hardware
59
* breakpoint or BKPT insn) targeting the specified exception level.
60
*/
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
61
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
index XXXXXXX..XXXXXXX 100644
62
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
63
--- a/target/arm/helper.c
17
+++ b/target/arm/helper.c
64
+++ b/target/arm/helper.c
18
@@ -XXX,XX +XXX,XX @@ static bool v7m_handle_execute_nsc(ARMCPU *cpu)
65
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
19
qemu_log_mask(CPU_LOG_INT, "...really an SG instruction at 0x%08" PRIx32
66
}
20
", executing it\n", env->regs[15]);
67
#endif /* !CONFIG_USER_ONLY */
21
env->regs[14] &= ~1;
68
22
+ env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_SFPA_MASK;
69
-/* Return the exception level which controls this address translation regime */
23
switch_v7m_security_state(env, true);
70
-static uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
24
xpsr_write(env, 0, XPSR_IT);
71
-{
25
env->regs[15] += 4;
72
- switch (mmu_idx) {
73
- case ARMMMUIdx_E20_0:
74
- case ARMMMUIdx_E20_2:
75
- case ARMMMUIdx_E20_2_PAN:
76
- case ARMMMUIdx_Stage2:
77
- case ARMMMUIdx_E2:
78
- return 2;
79
- case ARMMMUIdx_SE3:
80
- return 3;
81
- case ARMMMUIdx_SE10_0:
82
- return arm_el_is_aa64(env, 3) ? 1 : 3;
83
- case ARMMMUIdx_SE10_1:
84
- case ARMMMUIdx_SE10_1_PAN:
85
- case ARMMMUIdx_Stage1_E0:
86
- case ARMMMUIdx_Stage1_E1:
87
- case ARMMMUIdx_Stage1_E1_PAN:
88
- case ARMMMUIdx_E10_0:
89
- case ARMMMUIdx_E10_1:
90
- case ARMMMUIdx_E10_1_PAN:
91
- case ARMMMUIdx_MPrivNegPri:
92
- case ARMMMUIdx_MUserNegPri:
93
- case ARMMMUIdx_MPriv:
94
- case ARMMMUIdx_MUser:
95
- case ARMMMUIdx_MSPrivNegPri:
96
- case ARMMMUIdx_MSUserNegPri:
97
- case ARMMMUIdx_MSPriv:
98
- case ARMMMUIdx_MSUser:
99
- return 1;
100
- default:
101
- g_assert_not_reached();
102
- }
103
-}
104
-
105
uint64_t arm_sctlr(CPUARMState *env, int el)
106
{
107
/* Only EL0 needs to be adjusted for EL1&0 or EL2&0. */
26
--
108
--
27
2.20.1
109
2.20.1
28
110
29
111
diff view generated by jsdifflib
1
Add a new helper function which returns the MMU index to use
1
From: Richard Henderson <richard.henderson@linaro.org>
2
for v7M, where the caller specifies all of the security
3
state, privilege level and whether the execution priority
4
is negative, and reimplement the existing
5
arm_v7m_mmu_idx_for_secstate_and_priv() in terms of it.
6
2
7
We are going to need this for the lazy-FP-stacking code.
3
We will shortly need this in mte_helper.c as well.
8
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-23-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20190416125744.27770-21-peter.maydell@linaro.org
12
---
9
---
13
target/arm/cpu.h | 7 +++++++
10
target/arm/internals.h | 9 +++++++++
14
target/arm/helper.c | 14 +++++++++++---
11
target/arm/helper.c | 9 ---------
15
2 files changed, 18 insertions(+), 3 deletions(-)
12
2 files changed, 9 insertions(+), 9 deletions(-)
16
13
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
18
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
16
--- a/target/arm/internals.h
20
+++ b/target/arm/cpu.h
17
+++ b/target/arm/internals.h
21
@@ -XXX,XX +XXX,XX @@ static inline int arm_mmu_idx_to_el(ARMMMUIdx mmu_idx)
18
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
22
}
19
}
23
}
20
}
24
21
25
+/*
22
+/* Return the TCR controlling this translation regime */
26
+ * Return the MMU index for a v7M CPU with all relevant information
23
+static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
27
+ * manually specified.
24
+{
28
+ */
25
+ if (mmu_idx == ARMMMUIdx_Stage2) {
29
+ARMMMUIdx arm_v7m_mmu_idx_all(CPUARMState *env,
26
+ return &env->cp15.vtcr_el2;
30
+ bool secstate, bool priv, bool negpri);
27
+ }
28
+ return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
29
+}
31
+
30
+
32
/* Return the MMU index for a v7M CPU in the specified security and
31
/* Return the FSR value for a debug exception (watchpoint, hardware
33
* privilege state.
32
* breakpoint or BKPT insn) targeting the specified exception level.
34
*/
33
*/
35
diff --git a/target/arm/helper.c b/target/arm/helper.c
34
diff --git a/target/arm/helper.c b/target/arm/helper.c
36
index XXXXXXX..XXXXXXX 100644
35
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/helper.c
36
--- a/target/arm/helper.c
38
+++ b/target/arm/helper.c
37
+++ b/target/arm/helper.c
39
@@ -XXX,XX +XXX,XX @@ int fp_exception_el(CPUARMState *env, int cur_el)
38
@@ -XXX,XX +XXX,XX @@ static inline uint64_t regime_ttbr(CPUARMState *env, ARMMMUIdx mmu_idx,
40
return 0;
39
41
}
40
#endif /* !CONFIG_USER_ONLY */
42
41
43
-ARMMMUIdx arm_v7m_mmu_idx_for_secstate_and_priv(CPUARMState *env,
42
-/* Return the TCR controlling this translation regime */
44
- bool secstate, bool priv)
43
-static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
45
+ARMMMUIdx arm_v7m_mmu_idx_all(CPUARMState *env,
44
-{
46
+ bool secstate, bool priv, bool negpri)
45
- if (mmu_idx == ARMMMUIdx_Stage2) {
47
{
46
- return &env->cp15.vtcr_el2;
48
ARMMMUIdx mmu_idx = ARM_MMU_IDX_M;
47
- }
49
48
- return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
50
@@ -XXX,XX +XXX,XX @@ ARMMMUIdx arm_v7m_mmu_idx_for_secstate_and_priv(CPUARMState *env,
49
-}
51
mmu_idx |= ARM_MMU_IDX_M_PRIV;
50
-
52
}
51
/* Convert a possible stage1+2 MMU index into the appropriate
53
52
* stage 1 MMU index
54
- if (armv7m_nvic_neg_prio_requested(env->nvic, secstate)) {
53
*/
55
+ if (negpri) {
56
mmu_idx |= ARM_MMU_IDX_M_NEGPRI;
57
}
58
59
@@ -XXX,XX +XXX,XX @@ ARMMMUIdx arm_v7m_mmu_idx_for_secstate_and_priv(CPUARMState *env,
60
return mmu_idx;
61
}
62
63
+ARMMMUIdx arm_v7m_mmu_idx_for_secstate_and_priv(CPUARMState *env,
64
+ bool secstate, bool priv)
65
+{
66
+ bool negpri = armv7m_nvic_neg_prio_requested(env->nvic, secstate);
67
+
68
+ return arm_v7m_mmu_idx_all(env, secstate, priv, negpri);
69
+}
70
+
71
/* Return the MMU index for a v7M CPU in the specified security state */
72
ARMMMUIdx arm_v7m_mmu_idx_for_secstate(CPUARMState *env, bool secstate)
73
{
74
--
54
--
75
2.20.1
55
2.20.1
76
56
77
57
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform a single logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-24-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/internals.h | 8 +++
14
target/arm/translate-a64.h | 2 +
15
target/arm/mte_helper.c | 8 +++
16
target/arm/translate-a64.c | 100 ++++++++++++++++++++++++++++---------
17
5 files changed, 95 insertions(+), 24 deletions(-)
18
19
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper-a64.h
22
+++ b/target/arm/helper-a64.h
23
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
24
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
25
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
27
+DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
28
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
29
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
30
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
31
diff --git a/target/arm/internals.h b/target/arm/internals.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/internals.h
34
+++ b/target/arm/internals.h
35
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
36
#define LOG2_TAG_GRANULE 4
37
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
38
39
+/* Bits within a descriptor passed to the helper_mte_check* functions. */
40
+FIELD(MTEDESC, MIDX, 0, 4)
41
+FIELD(MTEDESC, TBI, 4, 2)
42
+FIELD(MTEDESC, TCMA, 6, 2)
43
+FIELD(MTEDESC, WRITE, 8, 1)
44
+FIELD(MTEDESC, ESIZE, 9, 5)
45
+FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
46
+
47
static inline int allocation_tag_from_addr(uint64_t ptr)
48
{
49
return extract64(ptr, 56, 4);
50
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/translate-a64.h
53
+++ b/target/arm/translate-a64.h
54
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
55
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
56
unsigned int imms, unsigned int immr);
57
bool sve_access_check(DisasContext *s);
58
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
59
+ bool tag_checked, int log2_size);
60
61
/* We should have at some point before trying to access an FP register
62
* done the necessary access check, so assert that
63
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/mte_helper.c
66
+++ b/target/arm/mte_helper.c
67
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
68
memset(mem, tag_pair, tag_bytes);
69
}
70
}
71
+
72
+/*
73
+ * Perform an MTE checked access for a single logical or atomic access.
74
+ */
75
+uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
76
+{
77
+ return ptr;
78
+}
79
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
80
index XXXXXXX..XXXXXXX 100644
81
--- a/target/arm/translate-a64.c
82
+++ b/target/arm/translate-a64.c
83
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
84
}
85
86
/*
87
- * Return a "clean" address for ADDR according to TBID.
88
- * This is always a fresh temporary, as we need to be able to
89
- * increment this independently of a dirty write-back address.
90
+ * Handle MTE and/or TBI.
91
+ *
92
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
93
+ * for the tag to be present in the FAR_ELx register. But for user-only
94
+ * mode we do not have a TLB with which to implement this, so we must
95
+ * remove the top byte now.
96
+ *
97
+ * Always return a fresh temporary that we can increment independently
98
+ * of the write-back address.
99
*/
100
+
101
static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
102
{
103
TCGv_i64 clean = new_tmp_a64(s);
104
- /*
105
- * In order to get the correct value in the FAR_ELx register,
106
- * we must present the memory subsystem with the "dirty" address
107
- * including the TBI. In system mode we can make this work via
108
- * the TLB, dropping the TBI during translation. But for user-only
109
- * mode we don't have that option, and must remove the top byte now.
110
- */
111
#ifdef CONFIG_USER_ONLY
112
gen_top_byte_ignore(s, clean, addr, s->tbid);
113
#else
114
@@ -XXX,XX +XXX,XX @@ static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
115
tcg_temp_free_i32(t_size);
116
}
117
118
+/*
119
+ * For MTE, check a single logical or atomic access. This probes a single
120
+ * address, the exact one specified. The size and alignment of the access
121
+ * is not relevant to MTE, per se, but watchpoints do require the size,
122
+ * and we want to recognize those before making any other changes to state.
123
+ */
124
+static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
125
+ bool is_write, bool tag_checked,
126
+ int log2_size, bool is_unpriv,
127
+ int core_idx)
128
+{
129
+ if (tag_checked && s->mte_active[is_unpriv]) {
130
+ TCGv_i32 tcg_desc;
131
+ TCGv_i64 ret;
132
+ int desc = 0;
133
+
134
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, core_idx);
135
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
136
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
137
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
138
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_size);
139
+ tcg_desc = tcg_const_i32(desc);
140
+
141
+ ret = new_tmp_a64(s);
142
+ gen_helper_mte_check1(ret, cpu_env, tcg_desc, addr);
143
+ tcg_temp_free_i32(tcg_desc);
144
+
145
+ return ret;
146
+ }
147
+ return clean_data_tbi(s, addr);
148
+}
149
+
150
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
151
+ bool tag_checked, int log2_size)
152
+{
153
+ return gen_mte_check1_mmuidx(s, addr, is_write, tag_checked, log2_size,
154
+ false, get_mem_index(s));
155
+}
156
+
157
typedef struct DisasCompare64 {
158
TCGCond cond;
159
TCGv_i64 value;
160
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap(DisasContext *s, int rs, int rt,
161
if (rn == 31) {
162
gen_check_sp_alignment(s);
163
}
164
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
165
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size);
166
tcg_gen_atomic_cmpxchg_i64(tcg_rs, clean_addr, tcg_rs, tcg_rt, memidx,
167
size | MO_ALIGN | s->be_data);
168
}
169
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt,
170
if (rn == 31) {
171
gen_check_sp_alignment(s);
172
}
173
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
174
+
175
+ /* This is a single atomic access, despite the "pair". */
176
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size + 1);
177
178
if (size == 2) {
179
TCGv_i64 cmp = tcg_temp_new_i64();
180
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
181
if (is_lasr) {
182
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
183
}
184
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
185
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
186
+ true, rn != 31, size);
187
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, false);
188
return;
189
190
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
191
if (rn == 31) {
192
gen_check_sp_alignment(s);
193
}
194
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
195
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
196
+ false, rn != 31, size);
197
s->is_ldex = true;
198
gen_load_exclusive(s, rt, rt2, clean_addr, size, false);
199
if (is_lasr) {
200
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
201
gen_check_sp_alignment(s);
202
}
203
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
204
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
205
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
206
+ true, rn != 31, size);
207
do_gpr_st(s, cpu_reg(s, rt), clean_addr, size, true, rt,
208
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
209
return;
210
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
211
if (rn == 31) {
212
gen_check_sp_alignment(s);
213
}
214
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
215
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
216
+ false, rn != 31, size);
217
do_gpr_ld(s, cpu_reg(s, rt), clean_addr, size, false, false, true, rt,
218
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
219
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
220
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
221
if (is_lasr) {
222
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
223
}
224
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
225
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
226
+ true, rn != 31, size);
227
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, true);
228
return;
229
}
230
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
231
if (rn == 31) {
232
gen_check_sp_alignment(s);
233
}
234
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
235
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
236
+ false, rn != 31, size);
237
s->is_ldex = true;
238
gen_load_exclusive(s, rt, rt2, clean_addr, size, true);
239
if (is_lasr) {
240
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
241
bool iss_valid = !is_vector;
242
bool post_index;
243
bool writeback;
244
+ int memidx;
245
246
TCGv_i64 clean_addr, dirty_addr;
247
248
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
249
if (!post_index) {
250
tcg_gen_addi_i64(dirty_addr, dirty_addr, imm9);
251
}
252
- clean_addr = clean_data_tbi(s, dirty_addr);
253
+
254
+ memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
255
+ clean_addr = gen_mte_check1_mmuidx(s, dirty_addr, is_store,
256
+ writeback || rn != 31,
257
+ size, is_unpriv, memidx);
258
259
if (is_vector) {
260
if (is_store) {
261
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
262
}
263
} else {
264
TCGv_i64 tcg_rt = cpu_reg(s, rt);
265
- int memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
266
bool iss_sf = disas_ldst_compute_iss_sf(size, is_signed, opc);
267
268
if (is_store) {
269
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn,
270
ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
271
272
tcg_gen_add_i64(dirty_addr, dirty_addr, tcg_rm);
273
- clean_addr = clean_data_tbi(s, dirty_addr);
274
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, true, size);
275
276
if (is_vector) {
277
if (is_store) {
278
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn,
279
dirty_addr = read_cpu_reg_sp(s, rn, 1);
280
offset = imm12 << size;
281
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
282
- clean_addr = clean_data_tbi(s, dirty_addr);
283
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, rn != 31, size);
284
285
if (is_vector) {
286
if (is_store) {
287
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
288
if (rn == 31) {
289
gen_check_sp_alignment(s);
290
}
291
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
292
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), false, rn != 31, size);
293
294
if (o3_opc == 014) {
295
/*
296
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pac(DisasContext *s, uint32_t insn,
297
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
298
299
/* Note that "clean" and "dirty" here refer to TBI not PAC. */
300
- clean_addr = clean_data_tbi(s, dirty_addr);
301
+ clean_addr = gen_mte_check1(s, dirty_addr, false,
302
+ is_wback || rn != 31, size);
303
304
tcg_rt = cpu_reg(s, rt);
305
do_gpr_ld(s, tcg_rt, clean_addr, size, /* is_signed */ false,
306
--
307
2.20.1
308
309
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform multiple logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-25-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/translate-a64.h | 2 ++
14
target/arm/mte_helper.c | 8 +++++
15
target/arm/translate-a64.c | 71 +++++++++++++++++++++++++++++---------
16
4 files changed, 66 insertions(+), 16 deletions(-)
17
18
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/helper-a64.h
21
+++ b/target/arm/helper-a64.h
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
25
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
26
+DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
27
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
28
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
29
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
30
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/translate-a64.h
33
+++ b/target/arm/translate-a64.h
34
@@ -XXX,XX +XXX,XX @@ bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
35
bool sve_access_check(DisasContext *s);
36
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
37
bool tag_checked, int log2_size);
38
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
39
+ bool tag_checked, int count, int log2_esize);
40
41
/* We should have at some point before trying to access an FP register
42
* done the necessary access check, so assert that
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
48
{
49
return ptr;
50
}
51
+
52
+/*
53
+ * Perform an MTE checked access for multiple logical accesses.
54
+ */
55
+uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
56
+{
57
+ return ptr;
58
+}
59
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/target/arm/translate-a64.c
62
+++ b/target/arm/translate-a64.c
63
@@ -XXX,XX +XXX,XX @@ TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
64
false, get_mem_index(s));
65
}
66
67
+/*
68
+ * For MTE, check multiple logical sequential accesses.
69
+ */
70
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
71
+ bool tag_checked, int log2_esize, int total_size)
72
+{
73
+ if (tag_checked && s->mte_active[0] && total_size != (1 << log2_esize)) {
74
+ TCGv_i32 tcg_desc;
75
+ TCGv_i64 ret;
76
+ int desc = 0;
77
+
78
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
79
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
80
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
81
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
82
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_esize);
83
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, total_size);
84
+ tcg_desc = tcg_const_i32(desc);
85
+
86
+ ret = new_tmp_a64(s);
87
+ gen_helper_mte_checkN(ret, cpu_env, tcg_desc, addr);
88
+ tcg_temp_free_i32(tcg_desc);
89
+
90
+ return ret;
91
+ }
92
+ return gen_mte_check1(s, addr, is_write, tag_checked, log2_esize);
93
+}
94
+
95
typedef struct DisasCompare64 {
96
TCGCond cond;
97
TCGv_i64 value;
98
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
99
}
100
}
101
102
- clean_addr = clean_data_tbi(s, dirty_addr);
103
+ clean_addr = gen_mte_checkN(s, dirty_addr, !is_load,
104
+ (wback || rn != 31) && !set_tag,
105
+ size, 2 << size);
106
+
107
if (is_vector) {
108
if (is_load) {
109
do_fp_ld(s, rt, clean_addr, size);
110
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
111
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
112
MemOp endian = s->be_data;
113
114
- int ebytes; /* bytes per element */
115
+ int total; /* total bytes */
116
int elements; /* elements per vector */
117
int rpt; /* num iterations */
118
int selem; /* structure elements */
119
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
120
endian = MO_LE;
121
}
122
123
- /* Consecutive little-endian elements from a single register
124
+ total = rpt * selem * (is_q ? 16 : 8);
125
+ tcg_rn = cpu_reg_sp(s, rn);
126
+
127
+ /*
128
+ * Issue the MTE check vs the logical repeat count, before we
129
+ * promote consecutive little-endian elements below.
130
+ */
131
+ clean_addr = gen_mte_checkN(s, tcg_rn, is_store, is_postidx || rn != 31,
132
+ size, total);
133
+
134
+ /*
135
+ * Consecutive little-endian elements from a single register
136
* can be promoted to a larger little-endian operation.
137
*/
138
if (selem == 1 && endian == MO_LE) {
139
size = 3;
140
}
141
- ebytes = 1 << size;
142
- elements = (is_q ? 16 : 8) / ebytes;
143
-
144
- tcg_rn = cpu_reg_sp(s, rn);
145
- clean_addr = clean_data_tbi(s, tcg_rn);
146
- tcg_ebytes = tcg_const_i64(ebytes);
147
+ elements = (is_q ? 16 : 8) >> size;
148
149
+ tcg_ebytes = tcg_const_i64(1 << size);
150
for (r = 0; r < rpt; r++) {
151
int e;
152
for (e = 0; e < elements; e++) {
153
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
154
155
if (is_postidx) {
156
if (rm == 31) {
157
- tcg_gen_addi_i64(tcg_rn, tcg_rn, rpt * elements * selem * ebytes);
158
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
159
} else {
160
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
161
}
162
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
163
int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
164
bool replicate = false;
165
int index = is_q << 3 | S << 2 | size;
166
- int ebytes, xs;
167
+ int xs, total;
168
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
169
170
if (extract32(insn, 31, 1)) {
171
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
172
return;
173
}
174
175
- ebytes = 1 << scale;
176
-
177
if (rn == 31) {
178
gen_check_sp_alignment(s);
179
}
180
181
+ total = selem << scale;
182
tcg_rn = cpu_reg_sp(s, rn);
183
- clean_addr = clean_data_tbi(s, tcg_rn);
184
- tcg_ebytes = tcg_const_i64(ebytes);
185
186
+ clean_addr = gen_mte_checkN(s, tcg_rn, !is_load, is_postidx || rn != 31,
187
+ scale, total);
188
+
189
+ tcg_ebytes = tcg_const_i64(1 << scale);
190
for (xs = 0; xs < selem; xs++) {
191
if (replicate) {
192
/* Load and replicate to all elements */
193
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
194
195
if (is_postidx) {
196
if (rm == 31) {
197
- tcg_gen_addi_i64(tcg_rn, tcg_rn, selem * ebytes);
198
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
199
} else {
200
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
201
}
202
--
203
2.20.1
204
205
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-26-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 48 +++++++++++++++
11
target/arm/mte_helper.c | 132 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 179 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, WRITE, 8, 1)
19
FIELD(MTEDESC, ESIZE, 9, 5)
20
FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
21
22
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
23
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
24
+ uint64_t ptr, uintptr_t ra);
25
+
26
static inline int allocation_tag_from_addr(uint64_t ptr)
27
{
28
return extract64(ptr, 56, 4);
29
@@ -XXX,XX +XXX,XX @@ static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
30
return deposit64(ptr, 56, 4, rtag);
31
}
32
33
+/* Return true if tbi bits mean that the access is checked. */
34
+static inline bool tbi_check(uint32_t desc, int bit55)
35
+{
36
+ return (desc >> (R_MTEDESC_TBI_SHIFT + bit55)) & 1;
37
+}
38
+
39
+/* Return true if tcma bits mean that the access is unchecked. */
40
+static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
41
+{
42
+ /*
43
+ * We had extracted bit55 and ptr_tag for other reasons, so fold
44
+ * (ptr<59:55> == 00000 || ptr<59:55> == 11111) into a single test.
45
+ */
46
+ bool match = ((ptr_tag + bit55) & 0xf) == 0;
47
+ bool tcma = (desc >> (R_MTEDESC_TCMA_SHIFT + bit55)) & 1;
48
+ return tcma && match;
49
+}
50
+
51
+/*
52
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
53
+ * for the tag to be present in the FAR_ELx register. But for user-only
54
+ * mode, we do not have a TLB with which to implement this, so we must
55
+ * remove the top byte.
56
+ */
57
+static inline uint64_t useronly_clean_ptr(uint64_t ptr)
58
+{
59
+ /* TBI is known to be enabled. */
60
+#ifdef CONFIG_USER_ONLY
61
+ ptr = sextract64(ptr, 0, 56);
62
+#endif
63
+ return ptr;
64
+}
65
+
66
+static inline uint64_t useronly_maybe_clean_ptr(uint32_t desc, uint64_t ptr)
67
+{
68
+#ifdef CONFIG_USER_ONLY
69
+ int64_t clean_ptr = sextract64(ptr, 0, 56);
70
+ if (tbi_check(desc, clean_ptr < 0)) {
71
+ ptr = clean_ptr;
72
+ }
73
+#endif
74
+ return ptr;
75
+}
76
+
77
#endif
78
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/target/arm/mte_helper.c
81
+++ b/target/arm/mte_helper.c
82
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
83
}
84
}
85
86
+/* Record a tag check failure. */
87
+static void mte_check_fail(CPUARMState *env, int mmu_idx,
88
+ uint64_t dirty_ptr, uintptr_t ra)
89
+{
90
+ ARMMMUIdx arm_mmu_idx = core_to_aa64_mmu_idx(mmu_idx);
91
+ int el, reg_el, tcf, select;
92
+ uint64_t sctlr;
93
+
94
+ reg_el = regime_el(env, arm_mmu_idx);
95
+ sctlr = env->cp15.sctlr_el[reg_el];
96
+
97
+ switch (arm_mmu_idx) {
98
+ case ARMMMUIdx_E10_0:
99
+ case ARMMMUIdx_E20_0:
100
+ el = 0;
101
+ tcf = extract64(sctlr, 38, 2);
102
+ break;
103
+ default:
104
+ el = reg_el;
105
+ tcf = extract64(sctlr, 40, 2);
106
+ }
107
+
108
+ switch (tcf) {
109
+ case 1:
110
+ /*
111
+ * Tag check fail causes a synchronous exception.
112
+ *
113
+ * In restore_state_to_opc, we set the exception syndrome
114
+ * for the load or store operation. Unwind first so we
115
+ * may overwrite that with the syndrome for the tag check.
116
+ */
117
+ cpu_restore_state(env_cpu(env), ra, true);
118
+ env->exception.vaddress = dirty_ptr;
119
+ raise_exception(env, EXCP_DATA_ABORT,
120
+ syn_data_abort_no_iss(el != 0, 0, 0, 0, 0, 0, 0x11),
121
+ exception_target_el(env));
122
+ /* noreturn, but fall through to the assert anyway */
123
+
124
+ case 0:
125
+ /*
126
+ * Tag check fail does not affect the PE.
127
+ * We eliminate this case by not setting MTE_ACTIVE
128
+ * in tb_flags, so that we never make this runtime call.
129
+ */
130
+ g_assert_not_reached();
131
+
132
+ case 2:
133
+ /* Tag check fail causes asynchronous flag set. */
134
+ mmu_idx = arm_mmu_idx_el(env, el);
135
+ if (regime_has_2_ranges(mmu_idx)) {
136
+ select = extract64(dirty_ptr, 55, 1);
137
+ } else {
138
+ select = 0;
139
+ }
140
+ env->cp15.tfsr_el[el] |= 1 << select;
141
+ break;
142
+
143
+ default:
144
+ /* Case 3: Reserved. */
145
+ qemu_log_mask(LOG_GUEST_ERROR,
146
+ "Tag check failure with SCTLR_EL%d.TCF%s "
147
+ "set to reserved value %d\n",
148
+ reg_el, el ? "" : "0", tcf);
149
+ break;
150
+ }
151
+}
152
+
153
/*
154
* Perform an MTE checked access for a single logical or atomic access.
155
*/
156
+static bool mte_probe1_int(CPUARMState *env, uint32_t desc, uint64_t ptr,
157
+ uintptr_t ra, int bit55)
158
+{
159
+ int mem_tag, mmu_idx, ptr_tag, size;
160
+ MMUAccessType type;
161
+ uint8_t *mem;
162
+
163
+ ptr_tag = allocation_tag_from_addr(ptr);
164
+
165
+ if (tcma_check(desc, bit55, ptr_tag)) {
166
+ return true;
167
+ }
168
+
169
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
170
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
171
+ size = FIELD_EX32(desc, MTEDESC, ESIZE);
172
+
173
+ mem = allocation_tag_mem(env, mmu_idx, ptr, type, size,
174
+ MMU_DATA_LOAD, 1, ra);
175
+ if (!mem) {
176
+ return true;
177
+ }
178
+
179
+ mem_tag = load_tag1(ptr, mem);
180
+ return ptr_tag == mem_tag;
181
+}
182
+
183
+/*
184
+ * No-fault version of mte_check1, to be used by SVE for MemSingleNF.
185
+ * Returns false if the access is Checked and the check failed. This
186
+ * is only intended to probe the tag -- the validity of the page must
187
+ * be checked beforehand.
188
+ */
189
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr)
190
+{
191
+ int bit55 = extract64(ptr, 55, 1);
192
+
193
+ /* If TBI is disabled, the access is unchecked. */
194
+ if (unlikely(!tbi_check(desc, bit55))) {
195
+ return true;
196
+ }
197
+
198
+ return mte_probe1_int(env, desc, ptr, 0, bit55);
199
+}
200
+
201
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
202
+ uint64_t ptr, uintptr_t ra)
203
+{
204
+ int bit55 = extract64(ptr, 55, 1);
205
+
206
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
207
+ if (unlikely(!tbi_check(desc, bit55))) {
208
+ return ptr;
209
+ }
210
+
211
+ if (unlikely(!mte_probe1_int(env, desc, ptr, ra, bit55))) {
212
+ int mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
213
+ mte_check_fail(env, mmu_idx, ptr, ra);
214
+ }
215
+
216
+ return useronly_clean_ptr(ptr);
217
+}
218
+
219
uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
220
{
221
- return ptr;
222
+ return mte_check1(env, desc, ptr, GETPC());
223
}
224
225
/*
226
--
227
2.20.1
228
229
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-27-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 2 +
11
target/arm/mte_helper.c | 165 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 166 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
19
bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
20
uint64_t mte_check1(CPUARMState *env, uint32_t desc,
21
uint64_t ptr, uintptr_t ra);
22
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
23
+ uint64_t ptr, uintptr_t ra);
24
25
static inline int allocation_tag_from_addr(uint64_t ptr)
26
{
27
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/mte_helper.c
30
+++ b/target/arm/mte_helper.c
31
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
32
/*
33
* Perform an MTE checked access for multiple logical accesses.
34
*/
35
+
36
+/**
37
+ * checkN:
38
+ * @tag: tag memory to test
39
+ * @odd: true to begin testing at tags at odd nibble
40
+ * @cmp: the tag to compare against
41
+ * @count: number of tags to test
42
+ *
43
+ * Return the number of successful tests.
44
+ * Thus a return value < @count indicates a failure.
45
+ *
46
+ * A note about sizes: count is expected to be small.
47
+ *
48
+ * The most common use will be LDP/STP of two integer registers,
49
+ * which means 16 bytes of memory touching at most 2 tags, but
50
+ * often the access is aligned and thus just 1 tag.
51
+ *
52
+ * Using AdvSIMD LD/ST (multiple), one can access 64 bytes of memory,
53
+ * touching at most 5 tags. SVE LDR/STR (vector) with the default
54
+ * vector length is also 64 bytes; the maximum architectural length
55
+ * is 256 bytes touching at most 9 tags.
56
+ *
57
+ * The loop below uses 7 logical operations and 1 memory operation
58
+ * per tag pair. An implementation that loads an aligned word and
59
+ * uses masking to ignore adjacent tags requires 18 logical operations
60
+ * and thus does not begin to pay off until 6 tags.
61
+ * Which, according to the survey above, is unlikely to be common.
62
+ */
63
+static int checkN(uint8_t *mem, int odd, int cmp, int count)
64
+{
65
+ int n = 0, diff;
66
+
67
+ /* Replicate the test tag and compare. */
68
+ cmp *= 0x11;
69
+ diff = *mem++ ^ cmp;
70
+
71
+ if (odd) {
72
+ goto start_odd;
73
+ }
74
+
75
+ while (1) {
76
+ /* Test even tag. */
77
+ if (unlikely((diff) & 0x0f)) {
78
+ break;
79
+ }
80
+ if (++n == count) {
81
+ break;
82
+ }
83
+
84
+ start_odd:
85
+ /* Test odd tag. */
86
+ if (unlikely((diff) & 0xf0)) {
87
+ break;
88
+ }
89
+ if (++n == count) {
90
+ break;
91
+ }
92
+
93
+ diff = *mem++ ^ cmp;
94
+ }
95
+ return n;
96
+}
97
+
98
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
99
+ uint64_t ptr, uintptr_t ra)
100
+{
101
+ int mmu_idx, ptr_tag, bit55;
102
+ uint64_t ptr_last, ptr_end, prev_page, next_page;
103
+ uint64_t tag_first, tag_end;
104
+ uint64_t tag_byte_first, tag_byte_end;
105
+ uint32_t esize, total, tag_count, tag_size, n, c;
106
+ uint8_t *mem1, *mem2;
107
+ MMUAccessType type;
108
+
109
+ bit55 = extract64(ptr, 55, 1);
110
+
111
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
112
+ if (unlikely(!tbi_check(desc, bit55))) {
113
+ return ptr;
114
+ }
115
+
116
+ ptr_tag = allocation_tag_from_addr(ptr);
117
+
118
+ if (tcma_check(desc, bit55, ptr_tag)) {
119
+ goto done;
120
+ }
121
+
122
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
123
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
124
+ esize = FIELD_EX32(desc, MTEDESC, ESIZE);
125
+ total = FIELD_EX32(desc, MTEDESC, TSIZE);
126
+
127
+ /* Find the addr of the end of the access, and of the last element. */
128
+ ptr_end = ptr + total;
129
+ ptr_last = ptr_end - esize;
130
+
131
+ /* Round the bounds to the tag granule, and compute the number of tags. */
132
+ tag_first = QEMU_ALIGN_DOWN(ptr, TAG_GRANULE);
133
+ tag_end = QEMU_ALIGN_UP(ptr_last, TAG_GRANULE);
134
+ tag_count = (tag_end - tag_first) / TAG_GRANULE;
135
+
136
+ /* Round the bounds to twice the tag granule, and compute the bytes. */
137
+ tag_byte_first = QEMU_ALIGN_DOWN(ptr, 2 * TAG_GRANULE);
138
+ tag_byte_end = QEMU_ALIGN_UP(ptr_last, 2 * TAG_GRANULE);
139
+
140
+ /* Locate the page boundaries. */
141
+ prev_page = ptr & TARGET_PAGE_MASK;
142
+ next_page = prev_page + TARGET_PAGE_SIZE;
143
+
144
+ if (likely(tag_end - prev_page <= TARGET_PAGE_SIZE)) {
145
+ /* Memory access stays on one page. */
146
+ tag_size = (tag_byte_end - tag_byte_first) / (2 * TAG_GRANULE);
147
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, total,
148
+ MMU_DATA_LOAD, tag_size, ra);
149
+ if (!mem1) {
150
+ goto done;
151
+ }
152
+ /* Perform all of the comparisons. */
153
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, tag_count);
154
+ } else {
155
+ /* Memory access crosses to next page. */
156
+ tag_size = (next_page - tag_byte_first) / (2 * TAG_GRANULE);
157
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, next_page - ptr,
158
+ MMU_DATA_LOAD, tag_size, ra);
159
+
160
+ tag_size = (tag_byte_end - next_page) / (2 * TAG_GRANULE);
161
+ mem2 = allocation_tag_mem(env, mmu_idx, next_page, type,
162
+ ptr_end - next_page,
163
+ MMU_DATA_LOAD, tag_size, ra);
164
+
165
+ /*
166
+ * Perform all of the comparisons.
167
+ * Note the possible but unlikely case of the operation spanning
168
+ * two pages that do not both have tagging enabled.
169
+ */
170
+ n = c = (next_page - tag_first) / TAG_GRANULE;
171
+ if (mem1) {
172
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, c);
173
+ }
174
+ if (n == c) {
175
+ if (!mem2) {
176
+ goto done;
177
+ }
178
+ n += checkN(mem2, 0, ptr_tag, tag_count - c);
179
+ }
180
+ }
181
+
182
+ /*
183
+ * If we failed, we know which granule. Compute the element that
184
+ * is first in that granule, and signal failure on that element.
185
+ */
186
+ if (unlikely(n < tag_count)) {
187
+ uint64_t fail_ofs;
188
+
189
+ fail_ofs = tag_first + n * TAG_GRANULE - ptr;
190
+ fail_ofs = ROUND_UP(fail_ofs, esize);
191
+ mte_check_fail(env, mmu_idx, ptr + fail_ofs, ra);
192
+ }
193
+
194
+ done:
195
+ return useronly_clean_ptr(ptr);
196
+}
197
+
198
uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
199
{
200
- return ptr;
201
+ return mte_checkN(env, desc, ptr, GETPC());
202
}
203
--
204
2.20.1
205
206
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Thomas Huth <thuth@redhat.com>
3
Use a special helper for DC_ZVA, rather than the more
4
Reviewed-by: Markus Armbruster <armbru@redhat.com>
4
general mte_checkN.
5
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
5
6
Message-id: 20190412165416.7977-11-philmd@redhat.com
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-28-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
10
---
9
include/hw/net/ne2000-isa.h | 6 ++++++
11
target/arm/helper-a64.h | 1 +
10
1 file changed, 6 insertions(+)
12
target/arm/mte_helper.c | 106 +++++++++++++++++++++++++++++++++++++
13
target/arm/translate-a64.c | 16 +++++-
14
3 files changed, 122 insertions(+), 1 deletion(-)
11
15
12
diff --git a/include/hw/net/ne2000-isa.h b/include/hw/net/ne2000-isa.h
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
13
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
14
--- a/include/hw/net/ne2000-isa.h
18
--- a/target/arm/helper-a64.h
15
+++ b/include/hw/net/ne2000-isa.h
19
+++ b/target/arm/helper-a64.h
16
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
17
* This work is licensed under the terms of the GNU GPL, version 2 or later.
21
18
* See the COPYING file in the top-level directory.
22
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
19
*/
23
DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
20
+
24
+DEF_HELPER_FLAGS_3(mte_check_zva, TCG_CALL_NO_WG, i64, env, i32, i64)
21
+#ifndef HW_NET_NE2K_ISA_H
25
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+#define HW_NET_NE2K_ISA_H
26
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
+
27
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
24
#include "hw/hw.h"
28
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
25
#include "hw/qdev.h"
29
index XXXXXXX..XXXXXXX 100644
26
#include "hw/isa/isa.h"
30
--- a/target/arm/mte_helper.c
27
@@ -XXX,XX +XXX,XX @@ static inline ISADevice *isa_ne2000_init(ISABus *bus, int base, int irq,
31
+++ b/target/arm/mte_helper.c
28
}
32
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
29
return d;
33
{
34
return mte_checkN(env, desc, ptr, GETPC());
30
}
35
}
31
+
36
+
32
+#endif
37
+/*
38
+ * Perform an MTE checked access for DC_ZVA.
39
+ */
40
+uint64_t HELPER(mte_check_zva)(CPUARMState *env, uint32_t desc, uint64_t ptr)
41
+{
42
+ uintptr_t ra = GETPC();
43
+ int log2_dcz_bytes, log2_tag_bytes;
44
+ int mmu_idx, bit55;
45
+ intptr_t dcz_bytes, tag_bytes, i;
46
+ void *mem;
47
+ uint64_t ptr_tag, mem_tag, align_ptr;
48
+
49
+ bit55 = extract64(ptr, 55, 1);
50
+
51
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
52
+ if (unlikely(!tbi_check(desc, bit55))) {
53
+ return ptr;
54
+ }
55
+
56
+ ptr_tag = allocation_tag_from_addr(ptr);
57
+
58
+ if (tcma_check(desc, bit55, ptr_tag)) {
59
+ goto done;
60
+ }
61
+
62
+ /*
63
+ * In arm_cpu_realizefn, we asserted that dcz > LOG2_TAG_GRANULE+1,
64
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway, to make
65
+ * sure that we can access one complete tag byte here.
66
+ */
67
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
68
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
69
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
70
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
71
+ align_ptr = ptr & -dcz_bytes;
72
+
73
+ /*
74
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
75
+ * the original pointer for an invalid page. But watchpoints require
76
+ * that we probe the actual space. So do both.
77
+ */
78
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
79
+ (void) probe_write(env, ptr, 1, mmu_idx, ra);
80
+ mem = allocation_tag_mem(env, mmu_idx, align_ptr, MMU_DATA_STORE,
81
+ dcz_bytes, MMU_DATA_LOAD, tag_bytes, ra);
82
+ if (!mem) {
83
+ goto done;
84
+ }
85
+
86
+ /*
87
+ * Unlike the reasoning for checkN, DC_ZVA is always aligned, and thus
88
+ * it is quite easy to perform all of the comparisons at once without
89
+ * any extra masking.
90
+ *
91
+ * The most common zva block size is 64; some of the thunderx cpus use
92
+ * a block size of 128. For user-only, aarch64_max_initfn will set the
93
+ * block size to 512. Fill out the other cases for future-proofing.
94
+ *
95
+ * In order to be able to find the first miscompare later, we want the
96
+ * tag bytes to be in little-endian order.
97
+ */
98
+ switch (log2_tag_bytes) {
99
+ case 0: /* zva_blocksize 32 */
100
+ mem_tag = *(uint8_t *)mem;
101
+ ptr_tag *= 0x11u;
102
+ break;
103
+ case 1: /* zva_blocksize 64 */
104
+ mem_tag = cpu_to_le16(*(uint16_t *)mem);
105
+ ptr_tag *= 0x1111u;
106
+ break;
107
+ case 2: /* zva_blocksize 128 */
108
+ mem_tag = cpu_to_le32(*(uint32_t *)mem);
109
+ ptr_tag *= 0x11111111u;
110
+ break;
111
+ case 3: /* zva_blocksize 256 */
112
+ mem_tag = cpu_to_le64(*(uint64_t *)mem);
113
+ ptr_tag *= 0x1111111111111111ull;
114
+ break;
115
+
116
+ default: /* zva_blocksize 512, 1024, 2048 */
117
+ ptr_tag *= 0x1111111111111111ull;
118
+ i = 0;
119
+ do {
120
+ mem_tag = cpu_to_le64(*(uint64_t *)(mem + i));
121
+ if (unlikely(mem_tag != ptr_tag)) {
122
+ goto fail;
123
+ }
124
+ i += 8;
125
+ align_ptr += 16 * TAG_GRANULE;
126
+ } while (i < tag_bytes);
127
+ goto done;
128
+ }
129
+
130
+ if (likely(mem_tag == ptr_tag)) {
131
+ goto done;
132
+ }
133
+
134
+ fail:
135
+ /* Locate the first nibble that differs. */
136
+ i = ctz64(mem_tag ^ ptr_tag) >> 4;
137
+ mte_check_fail(env, mmu_idx, align_ptr + i * TAG_GRANULE, ra);
138
+
139
+ done:
140
+ return useronly_clean_ptr(ptr);
141
+}
142
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
143
index XXXXXXX..XXXXXXX 100644
144
--- a/target/arm/translate-a64.c
145
+++ b/target/arm/translate-a64.c
146
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
147
return;
148
case ARM_CP_DC_ZVA:
149
/* Writes clear the aligned block of memory which rt points into. */
150
- tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
151
+ if (s->mte_active[0]) {
152
+ TCGv_i32 t_desc;
153
+ int desc = 0;
154
+
155
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
156
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
157
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
158
+ t_desc = tcg_const_i32(desc);
159
+
160
+ tcg_rt = new_tmp_a64(s);
161
+ gen_helper_mte_check_zva(tcg_rt, cpu_env, t_desc, cpu_reg(s, rt));
162
+ tcg_temp_free_i32(t_desc);
163
+ } else {
164
+ tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
165
+ }
166
gen_helper_dc_zva(cpu_env, tcg_rt);
167
return;
168
default:
33
--
169
--
34
2.20.1
170
2.20.1
35
171
36
172
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Since uWireSlave is only used in this new header, there is no
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
need to expose it via "qemu/typedefs.h".
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
5
Message-id: 20200626033144.790098-29-richard.henderson@linaro.org
6
Reviewed-by: Markus Armbruster <armbru@redhat.com>
7
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
8
Message-id: 20190412165416.7977-9-philmd@redhat.com
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
7
---
11
include/hw/arm/omap.h | 6 +-----
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
12
include/hw/devices.h | 15 ---------------
9
1 file changed, 33 insertions(+), 28 deletions(-)
13
include/hw/input/tsc2xxx.h | 36 ++++++++++++++++++++++++++++++++++++
14
include/qemu/typedefs.h | 1 -
15
hw/arm/nseries.c | 2 +-
16
hw/arm/palm.c | 2 +-
17
hw/input/tsc2005.c | 2 +-
18
hw/input/tsc210x.c | 4 ++--
19
MAINTAINERS | 2 ++
20
9 files changed, 44 insertions(+), 26 deletions(-)
21
create mode 100644 include/hw/input/tsc2xxx.h
22
10
23
diff --git a/include/hw/arm/omap.h b/include/hw/arm/omap.h
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
24
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
25
--- a/include/hw/arm/omap.h
13
--- a/target/arm/translate-sve.c
26
+++ b/include/hw/arm/omap.h
14
+++ b/target/arm/translate-sve.c
27
@@ -XXX,XX +XXX,XX @@
15
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
28
#include "exec/memory.h"
16
int len_remain = len % 8;
29
# define hw_omap_h        "omap.h"
17
int nparts = len / 8 + ctpop8(len_remain);
30
#include "hw/irq.h"
18
int midx = get_mem_index(s);
31
+#include "hw/input/tsc2xxx.h"
19
- TCGv_i64 addr, t0, t1;
32
#include "target/arm/cpu-qom.h"
20
+ TCGv_i64 dirty_addr, clean_addr, t0, t1;
33
#include "qemu/log.h"
21
34
22
- addr = tcg_temp_new_i64();
35
@@ -XXX,XX +XXX,XX @@ qemu_irq *omap_mpuio_in_get(struct omap_mpuio_s *s);
23
- t0 = tcg_temp_new_i64();
36
void omap_mpuio_out_set(struct omap_mpuio_s *s, int line, qemu_irq handler);
24
+ dirty_addr = tcg_temp_new_i64();
37
void omap_mpuio_key(struct omap_mpuio_s *s, int row, int col, int down);
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
38
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
39
-struct uWireSlave {
27
+ tcg_temp_free_i64(dirty_addr);
40
- uint16_t (*receive)(void *opaque);
28
41
- void (*send)(void *opaque, uint16_t data);
29
- /* Note that unpredicated load/store of vector/predicate registers
42
- void *opaque;
30
+ /*
43
-};
31
+ * Note that unpredicated load/store of vector/predicate registers
44
struct omap_uwire_s;
32
* are defined as a stream of bytes, which equates to little-endian
45
void omap_uwire_attach(struct omap_uwire_s *s,
33
- * operations on larger quantities. There is no nice way to force
46
uWireSlave *slave, int chipselect);
34
- * a little-endian load for aarch64_be-linux-user out of line.
47
diff --git a/include/hw/devices.h b/include/hw/devices.h
35
- *
48
index XXXXXXX..XXXXXXX 100644
36
+ * operations on larger quantities.
49
--- a/include/hw/devices.h
37
* Attempt to keep code expansion to a minimum by limiting the
50
+++ b/include/hw/devices.h
38
* amount of unrolling done.
51
@@ -XXX,XX +XXX,XX @@
39
*/
52
/* Devices that have nowhere better to go. */
40
if (nparts <= 4) {
53
41
int i;
54
#include "hw/hw.h"
42
55
-#include "ui/console.h"
43
+ t0 = tcg_temp_new_i64();
56
44
for (i = 0; i < len_align; i += 8) {
57
/* smc91c111.c */
45
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
58
void smc91c111_init(NICInfo *, uint32_t, qemu_irq);
46
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
59
@@ -XXX,XX +XXX,XX @@ void smc91c111_init(NICInfo *, uint32_t, qemu_irq);
47
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
60
/* lan9118.c */
48
tcg_gen_st_i64(t0, cpu_env, vofs + i);
61
void lan9118_init(NICInfo *, uint32_t, qemu_irq);
49
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
62
50
}
63
-/* tsc210x.c */
51
+ tcg_temp_free_i64(t0);
64
-uWireSlave *tsc2102_init(qemu_irq pint);
52
} else {
65
-uWireSlave *tsc2301_init(qemu_irq penirq, qemu_irq kbirq, qemu_irq dav);
53
TCGLabel *loop = gen_new_label();
66
-I2SCodec *tsc210x_codec(uWireSlave *chip);
54
TCGv_ptr tp, i = tcg_const_local_ptr(0);
67
-uint32_t tsc210x_txrx(void *opaque, uint32_t value, int len);
55
68
-void tsc210x_set_transform(uWireSlave *chip,
56
+ /* Copy the clean address into a local temp, live across the loop. */
69
- MouseTransformInfo *info);
57
+ t0 = clean_addr;
70
-void tsc210x_key_event(uWireSlave *chip, int key, int down);
58
+ clean_addr = tcg_temp_local_new_i64();
59
+ tcg_gen_mov_i64(clean_addr, t0);
60
+ tcg_temp_free_i64(t0);
61
+
62
gen_set_label(loop);
63
64
- /* Minimize the number of local temps that must be re-read from
65
- * the stack each iteration. Instead, re-compute values other
66
- * than the loop counter.
67
- */
68
+ t0 = tcg_temp_new_i64();
69
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
70
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
71
+
72
tp = tcg_temp_new_ptr();
73
- tcg_gen_addi_ptr(tp, i, imm);
74
- tcg_gen_extu_ptr_i64(addr, tp);
75
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
71
-
76
-
72
-/* tsc2005.c */
77
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
73
-void *tsc2005_init(qemu_irq pintdav);
74
-uint32_t tsc2005_txrx(void *opaque, uint32_t value, int len);
75
-void tsc2005_set_transform(void *opaque, MouseTransformInfo *info);
76
-
78
-
77
#endif
79
tcg_gen_add_ptr(tp, cpu_env, i);
78
diff --git a/include/hw/input/tsc2xxx.h b/include/hw/input/tsc2xxx.h
80
tcg_gen_addi_ptr(i, i, 8);
79
new file mode 100644
81
tcg_gen_st_i64(t0, tp, vofs);
80
index XXXXXXX..XXXXXXX
82
tcg_temp_free_ptr(tp);
81
--- /dev/null
83
+ tcg_temp_free_i64(t0);
82
+++ b/include/hw/input/tsc2xxx.h
84
83
@@ -XXX,XX +XXX,XX @@
85
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
84
+/*
86
tcg_temp_free_ptr(i);
85
+ * TI touchscreen controller
87
}
86
+ *
88
87
+ * Copyright (c) 2006 Andrzej Zaborowski
89
- /* Predicate register loads can be any multiple of 2.
88
+ * Copyright (C) 2008 Nokia Corporation
90
+ /*
89
+ *
91
+ * Predicate register loads can be any multiple of 2.
90
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
92
* Note that we still store the entire 64-bit unit into cpu_env.
91
+ * See the COPYING file in the top-level directory.
93
*/
92
+ */
94
if (len_remain) {
93
+
95
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
94
+#ifndef HW_INPUT_TSC2XXX_H
96
-
95
+#define HW_INPUT_TSC2XXX_H
97
+ t0 = tcg_temp_new_i64();
96
+
98
switch (len_remain) {
97
+#include "hw/irq.h"
99
case 2:
98
+#include "ui/console.h"
100
case 4:
99
+
101
case 8:
100
+typedef struct uWireSlave {
102
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
101
+ uint16_t (*receive)(void *opaque);
103
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx,
102
+ void (*send)(void *opaque, uint16_t data);
104
+ MO_LE | ctz32(len_remain));
103
+ void *opaque;
105
break;
104
+} uWireSlave;
106
105
+
107
case 6:
106
+/* tsc210x.c */
108
t1 = tcg_temp_new_i64();
107
+uWireSlave *tsc2102_init(qemu_irq pint);
109
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL);
108
+uWireSlave *tsc2301_init(qemu_irq penirq, qemu_irq kbirq, qemu_irq dav);
110
- tcg_gen_addi_i64(addr, addr, 4);
109
+I2SCodec *tsc210x_codec(uWireSlave *chip);
111
- tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW);
110
+uint32_t tsc210x_txrx(void *opaque, uint32_t value, int len);
112
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUL);
111
+void tsc210x_set_transform(uWireSlave *chip, MouseTransformInfo *info);
113
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
112
+void tsc210x_key_event(uWireSlave *chip, int key, int down);
114
+ tcg_gen_qemu_ld_i64(t1, clean_addr, midx, MO_LEUW);
113
+
115
tcg_gen_deposit_i64(t0, t0, t1, 32, 32);
114
+/* tsc2005.c */
116
tcg_temp_free_i64(t1);
115
+void *tsc2005_init(qemu_irq pintdav);
117
break;
116
+uint32_t tsc2005_txrx(void *opaque, uint32_t value, int len);
118
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
117
+void tsc2005_set_transform(void *opaque, MouseTransformInfo *info);
119
g_assert_not_reached();
118
+
120
}
119
+#endif
121
tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
120
diff --git a/include/qemu/typedefs.h b/include/qemu/typedefs.h
122
+ tcg_temp_free_i64(t0);
121
index XXXXXXX..XXXXXXX 100644
123
}
122
--- a/include/qemu/typedefs.h
124
- tcg_temp_free_i64(addr);
123
+++ b/include/qemu/typedefs.h
125
- tcg_temp_free_i64(t0);
124
@@ -XXX,XX +XXX,XX @@ typedef struct RAMBlock RAMBlock;
126
+ tcg_temp_free_i64(clean_addr);
125
typedef struct Range Range;
127
}
126
typedef struct SHPCDevice SHPCDevice;
128
127
typedef struct SSIBus SSIBus;
129
/* Similarly for stores. */
128
-typedef struct uWireSlave uWireSlave;
129
typedef struct VirtIODevice VirtIODevice;
130
typedef struct Visitor Visitor;
131
typedef void SaveStateHandler(QEMUFile *f, void *opaque);
132
diff --git a/hw/arm/nseries.c b/hw/arm/nseries.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/hw/arm/nseries.c
135
+++ b/hw/arm/nseries.c
136
@@ -XXX,XX +XXX,XX @@
137
#include "ui/console.h"
138
#include "hw/boards.h"
139
#include "hw/i2c/i2c.h"
140
-#include "hw/devices.h"
141
#include "hw/display/blizzard.h"
142
+#include "hw/input/tsc2xxx.h"
143
#include "hw/misc/cbus.h"
144
#include "hw/misc/tmp105.h"
145
#include "hw/block/flash.h"
146
diff --git a/hw/arm/palm.c b/hw/arm/palm.c
147
index XXXXXXX..XXXXXXX 100644
148
--- a/hw/arm/palm.c
149
+++ b/hw/arm/palm.c
150
@@ -XXX,XX +XXX,XX @@
151
#include "hw/arm/omap.h"
152
#include "hw/boards.h"
153
#include "hw/arm/arm.h"
154
-#include "hw/devices.h"
155
+#include "hw/input/tsc2xxx.h"
156
#include "hw/loader.h"
157
#include "exec/address-spaces.h"
158
#include "cpu.h"
159
diff --git a/hw/input/tsc2005.c b/hw/input/tsc2005.c
160
index XXXXXXX..XXXXXXX 100644
161
--- a/hw/input/tsc2005.c
162
+++ b/hw/input/tsc2005.c
163
@@ -XXX,XX +XXX,XX @@
164
#include "hw/hw.h"
165
#include "qemu/timer.h"
166
#include "ui/console.h"
167
-#include "hw/devices.h"
168
+#include "hw/input/tsc2xxx.h"
169
#include "trace.h"
170
171
#define TSC_CUT_RESOLUTION(value, p)    ((value) >> (16 - (p ? 12 : 10)))
172
diff --git a/hw/input/tsc210x.c b/hw/input/tsc210x.c
173
index XXXXXXX..XXXXXXX 100644
174
--- a/hw/input/tsc210x.c
175
+++ b/hw/input/tsc210x.c
176
@@ -XXX,XX +XXX,XX @@
177
#include "audio/audio.h"
178
#include "qemu/timer.h"
179
#include "ui/console.h"
180
-#include "hw/arm/omap.h"    /* For I2SCodec and uWireSlave */
181
-#include "hw/devices.h"
182
+#include "hw/arm/omap.h" /* For I2SCodec */
183
+#include "hw/input/tsc2xxx.h"
184
185
#define TSC_DATA_REGISTERS_PAGE        0x0
186
#define TSC_CONTROL_REGISTERS_PAGE    0x1
187
diff --git a/MAINTAINERS b/MAINTAINERS
188
index XXXXXXX..XXXXXXX 100644
189
--- a/MAINTAINERS
190
+++ b/MAINTAINERS
191
@@ -XXX,XX +XXX,XX @@ F: hw/input/tsc2005.c
192
F: hw/misc/cbus.c
193
F: hw/timer/twl92230.c
194
F: include/hw/display/blizzard.h
195
+F: include/hw/input/tsc2xxx.h
196
F: include/hw/misc/cbus.h
197
198
Palm
199
@@ -XXX,XX +XXX,XX @@ L: qemu-arm@nongnu.org
200
S: Odd Fixes
201
F: hw/arm/palm.c
202
F: hw/input/tsc210x.c
203
+F: include/hw/input/tsc2xxx.h
204
205
Raspberry Pi
206
M: Peter Maydell <peter.maydell@linaro.org>
207
--
130
--
208
2.20.1
131
2.20.1
209
132
210
133
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Markus Armbruster <armbru@redhat.com>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20190412165416.7977-8-philmd@redhat.com
5
Message-id: 20200626033144.790098-30-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
7
---
8
include/hw/devices.h | 3 ---
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
9
include/hw/input/gamepad.h | 19 +++++++++++++++++++
9
1 file changed, 33 insertions(+), 28 deletions(-)
10
hw/arm/stellaris.c | 2 +-
11
hw/input/stellaris_input.c | 2 +-
12
MAINTAINERS | 1 +
13
5 files changed, 22 insertions(+), 5 deletions(-)
14
create mode 100644 include/hw/input/gamepad.h
15
10
16
diff --git a/include/hw/devices.h b/include/hw/devices.h
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
17
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/devices.h
13
--- a/target/arm/translate-sve.c
19
+++ b/include/hw/devices.h
14
+++ b/target/arm/translate-sve.c
20
@@ -XXX,XX +XXX,XX @@ void *tsc2005_init(qemu_irq pintdav);
15
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
21
uint32_t tsc2005_txrx(void *opaque, uint32_t value, int len);
16
int len_remain = len % 8;
22
void tsc2005_set_transform(void *opaque, MouseTransformInfo *info);
17
int nparts = len / 8 + ctpop8(len_remain);
23
18
int midx = get_mem_index(s);
24
-/* stellaris_input.c */
19
- TCGv_i64 addr, t0;
25
-void stellaris_gamepad_init(int n, qemu_irq *irq, const int *keycode);
20
+ TCGv_i64 dirty_addr, clean_addr, t0;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
/* Note that unpredicated load/store of vector/predicate registers
30
* are defined as a stream of bytes, which equates to little-endian
31
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
32
if (nparts <= 4) {
33
int i;
34
35
+ t0 = tcg_temp_new_i64();
36
for (i = 0; i < len_align; i += 8) {
37
tcg_gen_ld_i64(t0, cpu_env, vofs + i);
38
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
39
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
40
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
41
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
42
}
43
+ tcg_temp_free_i64(t0);
44
} else {
45
TCGLabel *loop = gen_new_label();
46
- TCGv_ptr t2, i = tcg_const_local_ptr(0);
47
+ TCGv_ptr tp, i = tcg_const_local_ptr(0);
48
+
49
+ /* Copy the clean address into a local temp, live across the loop. */
50
+ t0 = clean_addr;
51
+ clean_addr = tcg_temp_local_new_i64();
52
+ tcg_gen_mov_i64(clean_addr, t0);
53
+ tcg_temp_free_i64(t0);
54
55
gen_set_label(loop);
56
57
- t2 = tcg_temp_new_ptr();
58
- tcg_gen_add_ptr(t2, cpu_env, i);
59
- tcg_gen_ld_i64(t0, t2, vofs);
26
-
60
-
27
#endif
61
- /* Minimize the number of local temps that must be re-read from
28
diff --git a/include/hw/input/gamepad.h b/include/hw/input/gamepad.h
62
- * the stack each iteration. Instead, re-compute values other
29
new file mode 100644
63
- * than the loop counter.
30
index XXXXXXX..XXXXXXX
64
- */
31
--- /dev/null
65
- tcg_gen_addi_ptr(t2, i, imm);
32
+++ b/include/hw/input/gamepad.h
66
- tcg_gen_extu_ptr_i64(addr, t2);
33
@@ -XXX,XX +XXX,XX @@
67
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
34
+/*
68
- tcg_temp_free_ptr(t2);
35
+ * Gamepad style buttons connected to IRQ/GPIO lines
69
-
36
+ *
70
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
37
+ * Copyright (c) 2007 CodeSourcery.
71
-
38
+ * Written by Paul Brook
72
+ t0 = tcg_temp_new_i64();
39
+ *
73
+ tp = tcg_temp_new_ptr();
40
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
74
+ tcg_gen_add_ptr(tp, cpu_env, i);
41
+ * See the COPYING file in the top-level directory.
75
+ tcg_gen_ld_i64(t0, tp, vofs);
42
+ */
76
tcg_gen_addi_ptr(i, i, 8);
77
+ tcg_temp_free_ptr(tp);
43
+
78
+
44
+#ifndef HW_INPUT_GAMEPAD_H
79
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
45
+#define HW_INPUT_GAMEPAD_H
80
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
46
+
81
+ tcg_temp_free_i64(t0);
47
+#include "hw/irq.h"
82
48
+
83
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
49
+/* stellaris_input.c */
84
tcg_temp_free_ptr(i);
50
+void stellaris_gamepad_init(int n, qemu_irq *irq, const int *keycode);
85
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
51
+
86
52
+#endif
87
/* Predicate register stores can be any multiple of 2. */
53
diff --git a/hw/arm/stellaris.c b/hw/arm/stellaris.c
88
if (len_remain) {
54
index XXXXXXX..XXXXXXX 100644
89
+ t0 = tcg_temp_new_i64();
55
--- a/hw/arm/stellaris.c
90
tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
56
+++ b/hw/arm/stellaris.c
91
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
57
@@ -XXX,XX +XXX,XX @@
92
58
#include "hw/sysbus.h"
93
switch (len_remain) {
59
#include "hw/ssi/ssi.h"
94
case 2:
60
#include "hw/arm/arm.h"
95
case 4:
61
-#include "hw/devices.h"
96
case 8:
62
#include "qemu/timer.h"
97
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
63
#include "hw/i2c/i2c.h"
98
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx,
64
#include "net/net.h"
99
+ MO_LE | ctz32(len_remain));
65
@@ -XXX,XX +XXX,XX @@
100
break;
66
#include "sysemu/sysemu.h"
101
67
#include "hw/arm/armv7m.h"
102
case 6:
68
#include "hw/char/pl011.h"
103
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUL);
69
+#include "hw/input/gamepad.h"
104
- tcg_gen_addi_i64(addr, addr, 4);
70
#include "hw/watchdog/cmsdk-apb-watchdog.h"
105
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUL);
71
#include "hw/misc/unimp.h"
106
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
72
#include "cpu.h"
107
tcg_gen_shri_i64(t0, t0, 32);
73
diff --git a/hw/input/stellaris_input.c b/hw/input/stellaris_input.c
108
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUW);
74
index XXXXXXX..XXXXXXX 100644
109
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUW);
75
--- a/hw/input/stellaris_input.c
110
break;
76
+++ b/hw/input/stellaris_input.c
111
77
@@ -XXX,XX +XXX,XX @@
112
default:
78
*/
113
g_assert_not_reached();
79
#include "qemu/osdep.h"
114
}
80
#include "hw/hw.h"
115
+ tcg_temp_free_i64(t0);
81
-#include "hw/devices.h"
116
}
82
+#include "hw/input/gamepad.h"
117
- tcg_temp_free_i64(addr);
83
#include "ui/console.h"
118
- tcg_temp_free_i64(t0);
84
119
+ tcg_temp_free_i64(clean_addr);
85
typedef struct {
120
}
86
diff --git a/MAINTAINERS b/MAINTAINERS
121
87
index XXXXXXX..XXXXXXX 100644
122
static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
88
--- a/MAINTAINERS
89
+++ b/MAINTAINERS
90
@@ -XXX,XX +XXX,XX @@ M: Peter Maydell <peter.maydell@linaro.org>
91
L: qemu-arm@nongnu.org
92
S: Maintained
93
F: hw/*/stellaris*
94
+F: include/hw/input/gamepad.h
95
96
Versatile Express
97
M: Peter Maydell <peter.maydell@linaro.org>
98
--
123
--
99
2.20.1
124
2.20.1
100
125
101
126
diff view generated by jsdifflib
1
In the stripe8() function we use a variable length array; however
1
From: Richard Henderson <richard.henderson@linaro.org>
2
we know that the maximum length required is MAX_NUM_BUSSES. Use
3
a fixed-length array and an assert instead.
4
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-31-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Francisco Iglesias <frasse.iglesias@gmail.com>
8
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
9
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
10
Message-id: 20190328152635.2794-1-peter.maydell@linaro.org
11
---
7
---
12
hw/ssi/xilinx_spips.c | 6 ++++--
8
target/arm/translate-sve.c | 6 ++++--
13
1 file changed, 4 insertions(+), 2 deletions(-)
9
1 file changed, 4 insertions(+), 2 deletions(-)
14
10
15
diff --git a/hw/ssi/xilinx_spips.c b/hw/ssi/xilinx_spips.c
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
16
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/ssi/xilinx_spips.c
13
--- a/target/arm/translate-sve.c
18
+++ b/hw/ssi/xilinx_spips.c
14
+++ b/target/arm/translate-sve.c
19
@@ -XXX,XX +XXX,XX @@ static void xlnx_zynqmp_qspips_reset(DeviceState *d)
15
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
20
16
unsigned esz = dtype_esz[a->dtype];
21
static inline void stripe8(uint8_t *x, int num, bool dir)
17
unsigned msz = dtype_msz(a->dtype);
22
{
18
TCGLabel *over = gen_new_label();
23
- uint8_t r[num];
19
- TCGv_i64 temp;
24
- memset(r, 0, sizeof(uint8_t) * num);
20
+ TCGv_i64 temp, clean_addr;
25
+ uint8_t r[MAX_NUM_BUSSES];
21
26
int idx[2] = {0, 0};
22
/* If the guarding predicate has no bits set, no load occurs. */
27
int bit[2] = {0, 7};
23
if (psz <= 8) {
28
int d = dir;
24
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
29
25
/* Load the data. */
30
+ assert(num <= MAX_NUM_BUSSES);
26
temp = tcg_temp_new_i64();
31
+ memset(r, 0, sizeof(uint8_t) * num);
27
tcg_gen_addi_i64(temp, cpu_reg_sp(s, a->rn), a->imm << msz);
28
- tcg_gen_qemu_ld_i64(temp, temp, get_mem_index(s),
29
+ clean_addr = gen_mte_check1(s, temp, false, true, msz);
32
+
30
+
33
for (idx[0] = 0; idx[0] < num; ++idx[0]) {
31
+ tcg_gen_qemu_ld_i64(temp, clean_addr, get_mem_index(s),
34
for (bit[0] = 7; bit[0] >= 0; bit[0]--) {
32
s->be_data | dtype_mop[a->dtype]);
35
r[idx[!d]] |= x[idx[d]] & 1 << bit[d] ? 1 << bit[!d] : 0;
33
34
/* Broadcast to *all* elements. */
36
--
35
--
37
2.20.1
36
2.20.1
38
37
39
38
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Markus Armbruster <armbru@redhat.com>
3
Move the variable declarations to the top of the function,
4
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
but do not create a new label before sve_access_check.
5
Message-id: 20190412165416.7977-5-philmd@redhat.com
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-32-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
include/hw/devices.h | 6 ------
11
target/arm/translate-sve.c | 12 +++++++-----
9
include/hw/display/tc6393xb.h | 24 ++++++++++++++++++++++++
12
1 file changed, 7 insertions(+), 5 deletions(-)
10
hw/arm/tosa.c | 2 +-
11
hw/display/tc6393xb.c | 2 +-
12
MAINTAINERS | 1 +
13
5 files changed, 27 insertions(+), 8 deletions(-)
14
create mode 100644 include/hw/display/tc6393xb.h
15
13
16
diff --git a/include/hw/devices.h b/include/hw/devices.h
14
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
17
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/devices.h
16
--- a/target/arm/translate-sve.c
19
+++ b/include/hw/devices.h
17
+++ b/target/arm/translate-sve.c
20
@@ -XXX,XX +XXX,XX @@ void *tahvo_init(qemu_irq irq, int betty);
18
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RQ_zpri(DisasContext *s, arg_rpri_load *a)
21
19
/* Load and broadcast element. */
22
void retu_key_event(void *retu, int state);
20
static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
23
21
{
24
-/* tc6393xb.c */
22
- if (!sve_access_check(s)) {
25
-typedef struct TC6393xbState TC6393xbState;
23
- return true;
26
-TC6393xbState *tc6393xb_init(struct MemoryRegion *sysmem,
24
- }
27
- uint32_t base, qemu_irq irq);
28
-qemu_irq tc6393xb_l3v_get(TC6393xbState *s);
29
-
25
-
30
#endif
26
unsigned vsz = vec_full_reg_size(s);
31
diff --git a/include/hw/display/tc6393xb.h b/include/hw/display/tc6393xb.h
27
unsigned psz = pred_full_reg_size(s);
32
new file mode 100644
28
unsigned esz = dtype_esz[a->dtype];
33
index XXXXXXX..XXXXXXX
29
unsigned msz = dtype_msz(a->dtype);
34
--- /dev/null
30
- TCGLabel *over = gen_new_label();
35
+++ b/include/hw/display/tc6393xb.h
31
+ TCGLabel *over;
36
@@ -XXX,XX +XXX,XX @@
32
TCGv_i64 temp, clean_addr;
37
+/*
33
38
+ * Toshiba TC6393XB I/O Controller.
34
+ if (!sve_access_check(s)) {
39
+ * Found in Sharp Zaurus SL-6000 (tosa) or some
35
+ return true;
40
+ * Toshiba e-Series PDAs.
36
+ }
41
+ *
42
+ * Copyright (c) 2007 Hervé Poussineau
43
+ *
44
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
45
+ * See the COPYING file in the top-level directory.
46
+ */
47
+
37
+
48
+#ifndef HW_DISPLAY_TC6393XB_H
38
+ over = gen_new_label();
49
+#define HW_DISPLAY_TC6393XB_H
50
+
39
+
51
+#include "exec/memory.h"
40
/* If the guarding predicate has no bits set, no load occurs. */
52
+#include "hw/irq.h"
41
if (psz <= 8) {
53
+
42
/* Reduce the pred_esz_masks value simply to reduce the
54
+typedef struct TC6393xbState TC6393xbState;
55
+
56
+TC6393xbState *tc6393xb_init(struct MemoryRegion *sysmem,
57
+ uint32_t base, qemu_irq irq);
58
+qemu_irq tc6393xb_l3v_get(TC6393xbState *s);
59
+
60
+#endif
61
diff --git a/hw/arm/tosa.c b/hw/arm/tosa.c
62
index XXXXXXX..XXXXXXX 100644
63
--- a/hw/arm/tosa.c
64
+++ b/hw/arm/tosa.c
65
@@ -XXX,XX +XXX,XX @@
66
#include "hw/hw.h"
67
#include "hw/arm/pxa.h"
68
#include "hw/arm/arm.h"
69
-#include "hw/devices.h"
70
#include "hw/arm/sharpsl.h"
71
#include "hw/pcmcia.h"
72
#include "hw/boards.h"
73
+#include "hw/display/tc6393xb.h"
74
#include "hw/i2c/i2c.h"
75
#include "hw/ssi/ssi.h"
76
#include "hw/sysbus.h"
77
diff --git a/hw/display/tc6393xb.c b/hw/display/tc6393xb.c
78
index XXXXXXX..XXXXXXX 100644
79
--- a/hw/display/tc6393xb.c
80
+++ b/hw/display/tc6393xb.c
81
@@ -XXX,XX +XXX,XX @@
82
#include "qapi/error.h"
83
#include "qemu/host-utils.h"
84
#include "hw/hw.h"
85
-#include "hw/devices.h"
86
+#include "hw/display/tc6393xb.h"
87
#include "hw/block/flash.h"
88
#include "ui/console.h"
89
#include "ui/pixel_ops.h"
90
diff --git a/MAINTAINERS b/MAINTAINERS
91
index XXXXXXX..XXXXXXX 100644
92
--- a/MAINTAINERS
93
+++ b/MAINTAINERS
94
@@ -XXX,XX +XXX,XX @@ F: hw/misc/mst_fpga.c
95
F: hw/misc/max111x.c
96
F: include/hw/arm/pxa.h
97
F: include/hw/arm/sharpsl.h
98
+F: include/hw/display/tc6393xb.h
99
100
SABRELITE / i.MX6
101
M: Peter Maydell <peter.maydell@linaro.org>
102
--
43
--
103
2.20.1
44
2.20.1
104
45
105
46
diff view generated by jsdifflib
1
Implement the code which updates the FPCCR register on an
1
From: Richard Henderson <richard.henderson@linaro.org>
2
exception entry where we are going to use lazy FP stacking.
3
We have to defer to the NVIC to determine whether the
4
various exceptions are currently ready or not.
5
2
3
Introduce an lvalue macro to wrap target_tlb_bit0.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-33-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Message-id: 20190416125744.27770-12-peter.maydell@linaro.org
8
---
9
---
9
target/arm/cpu.h | 14 +++++++++
10
target/arm/cpu.h | 13 +++++++++++++
10
hw/intc/armv7m_nvic.c | 34 ++++++++++++++++++++++
11
target/arm/helper.c | 2 +-
11
target/arm/helper.c | 67 ++++++++++++++++++++++++++++++++++++++++++-
12
target/arm/translate-a64.c | 2 +-
12
3 files changed, 114 insertions(+), 1 deletion(-)
13
3 files changed, 15 insertions(+), 2 deletions(-)
13
14
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
17
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
18
@@ -XXX,XX +XXX,XX @@ void armv7m_nvic_acknowledge_irq(void *opaque);
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
19
* (Ignoring -1, this is the same as the RETTOBASE value before completion.)
20
/* Shared between translate-sve.c and sve_helper.c. */
20
*/
21
extern const uint64_t pred_esz_masks[4];
21
int armv7m_nvic_complete_irq(void *opaque, int irq, bool secure);
22
22
+/**
23
+/* Helper for the macros below, validating the argument type. */
23
+ * armv7m_nvic_get_ready_status(void *opaque, int irq, bool secure)
24
+static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
24
+ * @opaque: the NVIC
25
+ * @irq: the exception number to mark pending
26
+ * @secure: false for non-banked exceptions or for the nonsecure
27
+ * version of a banked exception, true for the secure version of a banked
28
+ * exception.
29
+ *
30
+ * Return whether an exception is "ready", i.e. whether the exception is
31
+ * enabled and is configured at a priority which would allow it to
32
+ * interrupt the current execution priority. This controls whether the
33
+ * RDY bit for it in the FPCCR is set.
34
+ */
35
+bool armv7m_nvic_get_ready_status(void *opaque, int irq, bool secure);
36
/**
37
* armv7m_nvic_raw_execution_priority: return the raw execution priority
38
* @opaque: the NVIC
39
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
40
index XXXXXXX..XXXXXXX 100644
41
--- a/hw/intc/armv7m_nvic.c
42
+++ b/hw/intc/armv7m_nvic.c
43
@@ -XXX,XX +XXX,XX @@ int armv7m_nvic_complete_irq(void *opaque, int irq, bool secure)
44
return ret;
45
}
46
47
+bool armv7m_nvic_get_ready_status(void *opaque, int irq, bool secure)
48
+{
25
+{
49
+ /*
26
+ return x;
50
+ * Return whether an exception is "ready", i.e. it is enabled and is
51
+ * configured at a priority which would allow it to interrupt the
52
+ * current execution priority.
53
+ *
54
+ * irq and secure have the same semantics as for armv7m_nvic_set_pending():
55
+ * for non-banked exceptions secure is always false; for banked exceptions
56
+ * it indicates which of the exceptions is required.
57
+ */
58
+ NVICState *s = (NVICState *)opaque;
59
+ bool banked = exc_is_banked(irq);
60
+ VecInfo *vec;
61
+ int running = nvic_exec_prio(s);
62
+
63
+ assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
64
+ assert(!secure || banked);
65
+
66
+ /*
67
+ * HardFault is an odd special case: we always check against -1,
68
+ * even if we're secure and HardFault has priority -3; we never
69
+ * need to check for enabled state.
70
+ */
71
+ if (irq == ARMV7M_EXCP_HARD) {
72
+ return running > -1;
73
+ }
74
+
75
+ vec = (banked && secure) ? &s->sec_vectors[irq] : &s->vectors[irq];
76
+
77
+ return vec->enabled &&
78
+ exc_group_prio(s, vec->prio, secure) < running;
79
+}
27
+}
80
+
28
+
81
/* callback when external interrupt line is changed */
29
+/*
82
static void set_irq_level(void *opaque, int n, int level)
30
+ * Lvalue macros for ARM TLB bits that we must cache in the TCG TLB.
83
{
31
+ * Using these should be a bit more self-documenting than using the
32
+ * generic target bits directly.
33
+ */
34
+#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
35
+
36
/*
37
* Naming convention for isar_feature functions:
38
* Functions which test 32-bit ID registers should have _aa32_ in
84
diff --git a/target/arm/helper.c b/target/arm/helper.c
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
85
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
86
--- a/target/arm/helper.c
41
--- a/target/arm/helper.c
87
+++ b/target/arm/helper.c
42
+++ b/target/arm/helper.c
88
@@ -XXX,XX +XXX,XX @@ static void v7m_exception_taken(ARMCPU *cpu, uint32_t lr, bool dotailchain,
43
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
89
env->thumb = addr & 1;
44
}
45
/* When in aarch64 mode, and BTI is enabled, remember GP in the IOTLB. */
46
if (aarch64 && guarded && cpu_isar_feature(aa64_bti, cpu)) {
47
- txattrs->target_tlb_bit0 = true;
48
+ arm_tlb_bti_gp(txattrs) = true;
49
}
50
51
if (cacheattrs != NULL) {
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
53
index XXXXXXX..XXXXXXX 100644
54
--- a/target/arm/translate-a64.c
55
+++ b/target/arm/translate-a64.c
56
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
57
* table entry even for that case.
58
*/
59
return (tlb_hit(entry->addr_code, addr) &&
60
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs.target_tlb_bit0);
61
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
62
#endif
90
}
63
}
91
64
92
+static void v7m_update_fpccr(CPUARMState *env, uint32_t frameptr,
93
+ bool apply_splim)
94
+{
95
+ /*
96
+ * Like the pseudocode UpdateFPCCR: save state in FPCAR and FPCCR
97
+ * that we will need later in order to do lazy FP reg stacking.
98
+ */
99
+ bool is_secure = env->v7m.secure;
100
+ void *nvic = env->nvic;
101
+ /*
102
+ * Some bits are unbanked and live always in fpccr[M_REG_S]; some bits
103
+ * are banked and we want to update the bit in the bank for the
104
+ * current security state; and in one case we want to specifically
105
+ * update the NS banked version of a bit even if we are secure.
106
+ */
107
+ uint32_t *fpccr_s = &env->v7m.fpccr[M_REG_S];
108
+ uint32_t *fpccr_ns = &env->v7m.fpccr[M_REG_NS];
109
+ uint32_t *fpccr = &env->v7m.fpccr[is_secure];
110
+ bool hfrdy, bfrdy, mmrdy, ns_ufrdy, s_ufrdy, sfrdy, monrdy;
111
+
112
+ env->v7m.fpcar[is_secure] = frameptr & ~0x7;
113
+
114
+ if (apply_splim && arm_feature(env, ARM_FEATURE_V8)) {
115
+ bool splimviol;
116
+ uint32_t splim = v7m_sp_limit(env);
117
+ bool ign = armv7m_nvic_neg_prio_requested(nvic, is_secure) &&
118
+ (env->v7m.ccr[is_secure] & R_V7M_CCR_STKOFHFNMIGN_MASK);
119
+
120
+ splimviol = !ign && frameptr < splim;
121
+ *fpccr = FIELD_DP32(*fpccr, V7M_FPCCR, SPLIMVIOL, splimviol);
122
+ }
123
+
124
+ *fpccr = FIELD_DP32(*fpccr, V7M_FPCCR, LSPACT, 1);
125
+
126
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, S, is_secure);
127
+
128
+ *fpccr = FIELD_DP32(*fpccr, V7M_FPCCR, USER, arm_current_el(env) == 0);
129
+
130
+ *fpccr = FIELD_DP32(*fpccr, V7M_FPCCR, THREAD,
131
+ !arm_v7m_is_handler_mode(env));
132
+
133
+ hfrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_HARD, false);
134
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, HFRDY, hfrdy);
135
+
136
+ bfrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_BUS, false);
137
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, BFRDY, bfrdy);
138
+
139
+ mmrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_MEM, is_secure);
140
+ *fpccr = FIELD_DP32(*fpccr, V7M_FPCCR, MMRDY, mmrdy);
141
+
142
+ ns_ufrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_USAGE, false);
143
+ *fpccr_ns = FIELD_DP32(*fpccr_ns, V7M_FPCCR, UFRDY, ns_ufrdy);
144
+
145
+ monrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_DEBUG, false);
146
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, MONRDY, monrdy);
147
+
148
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
149
+ s_ufrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_USAGE, true);
150
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, UFRDY, s_ufrdy);
151
+
152
+ sfrdy = armv7m_nvic_get_ready_status(nvic, ARMV7M_EXCP_SECURE, false);
153
+ *fpccr_s = FIELD_DP32(*fpccr_s, V7M_FPCCR, SFRDY, sfrdy);
154
+ }
155
+}
156
+
157
static bool v7m_push_stack(ARMCPU *cpu)
158
{
159
/* Do the "set up stack frame" part of exception entry,
160
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
161
}
162
} else {
163
/* Lazy stacking enabled, save necessary info to stack later */
164
- /* TODO : equivalent of UpdateFPCCR() pseudocode */
165
+ v7m_update_fpccr(env, frameptr + 0x20, true);
166
}
167
}
168
}
169
--
65
--
170
2.20.1
66
2.20.1
171
67
172
68
diff view generated by jsdifflib
1
In the v7M architecture, if an exception is generated in the process
1
From: Richard Henderson <richard.henderson@linaro.org>
2
of doing the lazy stacking of FP registers, the handling of
3
possible escalation to HardFault is treated differently to the normal
4
approach: it works based on the saved information about exception
5
readiness that was stored in the FPCCR when the stack frame was
6
created. Provide a new function armv7m_nvic_set_pending_lazyfp()
7
which pends exceptions during lazy stacking, and implements
8
this logic.
9
2
10
This corresponds to the pseudocode TakePreserveFPException().
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
11
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-34-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20190416125744.27770-22-peter.maydell@linaro.org
15
---
10
---
16
target/arm/cpu.h | 12 ++++++
11
target/arm/cpu.h | 1 +
17
hw/intc/armv7m_nvic.c | 96 +++++++++++++++++++++++++++++++++++++++++++
12
target/arm/helper-sve.h | 58 ++++++++++
18
2 files changed, 108 insertions(+)
13
target/arm/internals.h | 6 +
14
target/arm/sve_helper.c | 218 ++++++++++++++++++++++++++++++-------
15
target/arm/translate-sve.c | 186 ++++++++++++++++++++++---------
16
5 files changed, 378 insertions(+), 91 deletions(-)
19
17
20
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/cpu.h
20
--- a/target/arm/cpu.h
23
+++ b/target/arm/cpu.h
21
+++ b/target/arm/cpu.h
24
@@ -XXX,XX +XXX,XX @@ void armv7m_nvic_set_pending(void *opaque, int irq, bool secure);
22
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
25
* a different exception).
23
* generic target bits directly.
26
*/
24
*/
27
void armv7m_nvic_set_pending_derived(void *opaque, int irq, bool secure);
25
#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
28
+/**
26
+#define arm_tlb_mte_tagged(x) (typecheck_memtxattrs(x)->target_tlb_bit1)
29
+ * armv7m_nvic_set_pending_lazyfp: mark this lazy FP exception as pending
27
30
+ * @opaque: the NVIC
28
/*
31
+ * @irq: the exception number to mark pending
29
* Naming convention for isar_feature functions:
32
+ * @secure: false for non-banked exceptions or for the nonsecure
30
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
33
+ * version of a banked exception, true for the secure version of a banked
31
index XXXXXXX..XXXXXXX 100644
34
+ * exception.
32
--- a/target/arm/helper-sve.h
35
+ *
33
+++ b/target/arm/helper-sve.h
36
+ * Similar to armv7m_nvic_set_pending(), but specifically for exceptions
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ld1sds_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
+ * generated in the course of lazy stacking of FP registers.
35
DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
DEF_HELPER_FLAGS_4(sve_ld1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
38
+DEF_HELPER_FLAGS_4(sve_ld1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_4(sve_ld2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ld3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_ld4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ld1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_4(sve_ld2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ld3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_ld4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+
48
+DEF_HELPER_FLAGS_4(sve_ld1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ld2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_ld3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ld4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+
53
+DEF_HELPER_FLAGS_4(sve_ld1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_4(sve_ld2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_ld3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ld4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+
58
+DEF_HELPER_FLAGS_4(sve_ld1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
59
+DEF_HELPER_FLAGS_4(sve_ld2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_ld3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_ld4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_ld1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_ld2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ld3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_ld4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ld1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_ld2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ld3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_4(sve_ld4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_4(sve_ld1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_4(sve_ld1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_4(sve_ld1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_4(sve_ld1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_4(sve_ld1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_4(sve_ld1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
+
80
+DEF_HELPER_FLAGS_4(sve_ld1hsu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ld1hdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ld1hss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ld1hds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+
85
+DEF_HELPER_FLAGS_4(sve_ld1hsu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ld1hdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_4(sve_ld1hss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_4(sve_ld1hds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
89
+
90
+DEF_HELPER_FLAGS_4(sve_ld1sdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
91
+DEF_HELPER_FLAGS_4(sve_ld1sds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
92
+
93
+DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ld1sds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
95
+
96
DEF_HELPER_FLAGS_4(sve_ldff1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
97
DEF_HELPER_FLAGS_4(sve_ldff1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
98
DEF_HELPER_FLAGS_4(sve_ldff1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
99
diff --git a/target/arm/internals.h b/target/arm/internals.h
100
index XXXXXXX..XXXXXXX 100644
101
--- a/target/arm/internals.h
102
+++ b/target/arm/internals.h
103
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
104
#define LOG2_TAG_GRANULE 4
105
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
106
107
+/*
108
+ * The SVE simd_data field, for memory ops, contains either
109
+ * rd (5 bits) or a shift count (2 bits).
38
+ */
110
+ */
39
+void armv7m_nvic_set_pending_lazyfp(void *opaque, int irq, bool secure);
111
+#define SVE_MTEDESC_SHIFT 5
40
/**
112
+
41
* armv7m_nvic_get_pending_irq_info: return highest priority pending
113
/* Bits within a descriptor passed to the helper_mte_check* functions. */
42
* exception, and whether it targets Secure state
114
FIELD(MTEDESC, MIDX, 0, 4)
43
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
115
FIELD(MTEDESC, TBI, 4, 2)
116
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
44
index XXXXXXX..XXXXXXX 100644
117
index XXXXXXX..XXXXXXX 100644
45
--- a/hw/intc/armv7m_nvic.c
118
--- a/target/arm/sve_helper.c
46
+++ b/hw/intc/armv7m_nvic.c
119
+++ b/target/arm/sve_helper.c
47
@@ -XXX,XX +XXX,XX @@ void armv7m_nvic_set_pending_derived(void *opaque, int irq, bool secure)
120
@@ -XXX,XX +XXX,XX @@ static void sve_cont_ldst_watchpoints(SVEContLdSt *info, CPUARMState *env,
48
do_armv7m_nvic_set_pending(opaque, irq, secure, true);
121
#endif
49
}
122
}
50
123
51
+void armv7m_nvic_set_pending_lazyfp(void *opaque, int irq, bool secure)
124
+typedef uint64_t mte_check_fn(CPUARMState *, uint32_t, uint64_t, uintptr_t);
125
+
126
+static inline QEMU_ALWAYS_INLINE
127
+void sve_cont_ldst_mte_check_int(SVEContLdSt *info, CPUARMState *env,
128
+ uint64_t *vg, target_ulong addr, int esize,
129
+ int msize, uint32_t mtedesc, uintptr_t ra,
130
+ mte_check_fn *check)
52
+{
131
+{
53
+ /*
132
+ intptr_t mem_off, reg_off, reg_last;
54
+ * Pend an exception during lazy FP stacking. This differs
133
+
55
+ * from the usual exception pending because the logic for
134
+ /* Process the page only if MemAttr == Tagged. */
56
+ * whether we should escalate depends on the saved context
135
+ if (arm_tlb_mte_tagged(&info->page[0].attrs)) {
57
+ * in the FPCCR register, not on the current state of the CPU/NVIC.
136
+ mem_off = info->mem_off_first[0];
58
+ */
137
+ reg_off = info->reg_off_first[0];
59
+ NVICState *s = (NVICState *)opaque;
138
+ reg_last = info->reg_off_split;
60
+ bool banked = exc_is_banked(irq);
139
+ if (reg_last < 0) {
61
+ VecInfo *vec;
140
+ reg_last = info->reg_off_last[0];
62
+ bool targets_secure;
63
+ bool escalate = false;
64
+ /*
65
+ * We will only look at bits in fpccr if this is a banked exception
66
+ * (in which case 'secure' tells us whether it is the S or NS version).
67
+ * All the bits for the non-banked exceptions are in fpccr_s.
68
+ */
69
+ uint32_t fpccr_s = s->cpu->env.v7m.fpccr[M_REG_S];
70
+ uint32_t fpccr = s->cpu->env.v7m.fpccr[secure];
71
+
72
+ assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
73
+ assert(!secure || banked);
74
+
75
+ vec = (banked && secure) ? &s->sec_vectors[irq] : &s->vectors[irq];
76
+
77
+ targets_secure = banked ? secure : exc_targets_secure(s, irq);
78
+
79
+ switch (irq) {
80
+ case ARMV7M_EXCP_DEBUG:
81
+ if (!(fpccr_s & R_V7M_FPCCR_MONRDY_MASK)) {
82
+ /* Ignore DebugMonitor exception */
83
+ return;
84
+ }
141
+ }
85
+ break;
142
+
86
+ case ARMV7M_EXCP_MEM:
143
+ do {
87
+ escalate = !(fpccr & R_V7M_FPCCR_MMRDY_MASK);
144
+ uint64_t pg = vg[reg_off >> 6];
88
+ break;
145
+ do {
89
+ case ARMV7M_EXCP_USAGE:
146
+ if ((pg >> (reg_off & 63)) & 1) {
90
+ escalate = !(fpccr & R_V7M_FPCCR_UFRDY_MASK);
147
+ check(env, mtedesc, addr, ra);
91
+ break;
148
+ }
92
+ case ARMV7M_EXCP_BUS:
149
+ reg_off += esize;
93
+ escalate = !(fpccr_s & R_V7M_FPCCR_BFRDY_MASK);
150
+ mem_off += msize;
94
+ break;
151
+ } while (reg_off <= reg_last && (reg_off & 63));
95
+ case ARMV7M_EXCP_SECURE:
152
+ } while (reg_off <= reg_last);
96
+ escalate = !(fpccr_s & R_V7M_FPCCR_SFRDY_MASK);
97
+ break;
98
+ default:
99
+ g_assert_not_reached();
100
+ }
153
+ }
101
+
154
+
102
+ if (escalate) {
155
+ mem_off = info->mem_off_first[1];
103
+ /*
156
+ if (mem_off >= 0 && arm_tlb_mte_tagged(&info->page[1].attrs)) {
104
+ * Escalate to HardFault: faults that initially targeted Secure
157
+ reg_off = info->reg_off_first[1];
105
+ * continue to do so, even if HF normally targets NonSecure.
158
+ reg_last = info->reg_off_last[1];
106
+ */
159
+
107
+ irq = ARMV7M_EXCP_HARD;
160
+ do {
108
+ if (arm_feature(&s->cpu->env, ARM_FEATURE_M_SECURITY) &&
161
+ uint64_t pg = vg[reg_off >> 6];
109
+ (targets_secure ||
162
+ do {
110
+ !(s->cpu->env.v7m.aircr & R_V7M_AIRCR_BFHFNMINS_MASK))) {
163
+ if ((pg >> (reg_off & 63)) & 1) {
111
+ vec = &s->sec_vectors[irq];
164
+ check(env, mtedesc, addr, ra);
112
+ } else {
165
+ }
113
+ vec = &s->vectors[irq];
166
+ reg_off += esize;
114
+ }
167
+ mem_off += msize;
115
+ }
168
+ } while (reg_off & 63);
116
+
169
+ } while (reg_off <= reg_last);
117
+ if (!vec->enabled ||
118
+ nvic_exec_prio(s) <= exc_group_prio(s, vec->prio, secure)) {
119
+ if (!(fpccr_s & R_V7M_FPCCR_HFRDY_MASK)) {
120
+ /*
121
+ * We want to escalate to HardFault but the context the
122
+ * FP state belongs to prevents the exception pre-empting.
123
+ */
124
+ cpu_abort(&s->cpu->parent_obj,
125
+ "Lockup: can't escalate to HardFault during "
126
+ "lazy FP register stacking\n");
127
+ }
128
+ }
129
+
130
+ if (escalate) {
131
+ s->cpu->env.v7m.hfsr |= R_V7M_HFSR_FORCED_MASK;
132
+ }
133
+ if (!vec->pending) {
134
+ vec->pending = 1;
135
+ /*
136
+ * We do not call nvic_irq_update(), because we know our caller
137
+ * is going to handle causing us to take the exception by
138
+ * raising EXCP_LAZYFP, so raising the IRQ line would be
139
+ * pointless extra work. We just need to recompute the
140
+ * priorities so that armv7m_nvic_can_take_pending_exception()
141
+ * returns the right answer.
142
+ */
143
+ nvic_recompute_state(s);
144
+ }
170
+ }
145
+}
171
+}
146
+
172
+
147
/* Make pending IRQ active. */
173
+typedef void sve_cont_ldst_mte_check_fn(SVEContLdSt *info, CPUARMState *env,
148
void armv7m_nvic_acknowledge_irq(void *opaque)
174
+ uint64_t *vg, target_ulong addr,
175
+ int esize, int msize, uint32_t mtedesc,
176
+ uintptr_t ra);
177
+
178
+static void sve_cont_ldst_mte_check1(SVEContLdSt *info, CPUARMState *env,
179
+ uint64_t *vg, target_ulong addr,
180
+ int esize, int msize, uint32_t mtedesc,
181
+ uintptr_t ra)
182
+{
183
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
184
+ mtedesc, ra, mte_check1);
185
+}
186
+
187
+static void sve_cont_ldst_mte_checkN(SVEContLdSt *info, CPUARMState *env,
188
+ uint64_t *vg, target_ulong addr,
189
+ int esize, int msize, uint32_t mtedesc,
190
+ uintptr_t ra)
191
+{
192
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
193
+ mtedesc, ra, mte_checkN);
194
+}
195
+
196
+
197
/*
198
* Common helper for all contiguous 1,2,3,4-register predicated stores.
199
*/
200
static inline QEMU_ALWAYS_INLINE
201
void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
202
uint32_t desc, const uintptr_t retaddr,
203
- const int esz, const int msz, const int N,
204
+ const int esz, const int msz, const int N, uint32_t mtedesc,
205
sve_ldst1_host_fn *host_fn,
206
- sve_ldst1_tlb_fn *tlb_fn)
207
+ sve_ldst1_tlb_fn *tlb_fn,
208
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
149
{
209
{
210
const unsigned rd = simd_data(desc);
211
const intptr_t reg_max = simd_oprsz(desc);
212
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
213
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
214
BP_MEM_READ, retaddr);
215
216
- /* TODO: MTE check. */
217
+ /*
218
+ * Handle mte checks for all active elements.
219
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
220
+ */
221
+ if (mte_check_fn && mtedesc) {
222
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
223
+ mtedesc, retaddr);
224
+ }
225
226
flags = info.page[0].flags | info.page[1].flags;
227
if (unlikely(flags != 0)) {
228
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
229
}
230
}
231
232
-#define DO_LD1_1(NAME, ESZ) \
233
-void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
234
- target_ulong addr, uint32_t desc) \
235
-{ \
236
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
237
- sve_##NAME##_host, sve_##NAME##_tlb); \
238
+static inline QEMU_ALWAYS_INLINE
239
+void sve_ldN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
240
+ uint32_t desc, const uintptr_t ra,
241
+ const int esz, const int msz, const int N,
242
+ sve_ldst1_host_fn *host_fn,
243
+ sve_ldst1_tlb_fn *tlb_fn)
244
+{
245
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
246
+ int bit55 = extract64(addr, 55, 1);
247
+
248
+ /* Remove mtedesc from the normal sve descriptor. */
249
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
250
+
251
+ /* Perform gross MTE suppression early. */
252
+ if (!tbi_check(desc, bit55) ||
253
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
254
+ mtedesc = 0;
255
+ }
256
+
257
+ sve_ldN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
258
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
259
}
260
261
-#define DO_LD1_2(NAME, ESZ, MSZ) \
262
-void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
263
- target_ulong addr, uint32_t desc) \
264
-{ \
265
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
266
- sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
267
-} \
268
-void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
269
- target_ulong addr, uint32_t desc) \
270
-{ \
271
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
272
- sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
273
+#define DO_LD1_1(NAME, ESZ) \
274
+void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
275
+ target_ulong addr, uint32_t desc) \
276
+{ \
277
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, 0, \
278
+ sve_##NAME##_host, sve_##NAME##_tlb, NULL); \
279
+} \
280
+void HELPER(sve_##NAME##_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
284
+ sve_##NAME##_host, sve_##NAME##_tlb); \
285
+}
286
+
287
+#define DO_LD1_2(NAME, ESZ, MSZ) \
288
+void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
289
+ target_ulong addr, uint32_t desc) \
290
+{ \
291
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
292
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb, NULL); \
293
+} \
294
+void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
295
+ target_ulong addr, uint32_t desc) \
296
+{ \
297
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
298
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb, NULL); \
299
+} \
300
+void HELPER(sve_##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
301
+ target_ulong addr, uint32_t desc) \
302
+{ \
303
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
304
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
305
+} \
306
+void HELPER(sve_##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
307
+ target_ulong addr, uint32_t desc) \
308
+{ \
309
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
310
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
311
}
312
313
DO_LD1_1(ld1bb, MO_8)
314
@@ -XXX,XX +XXX,XX @@ DO_LD1_2(ld1dd, MO_64, MO_64)
315
#undef DO_LD1_1
316
#undef DO_LD1_2
317
318
-#define DO_LDN_1(N) \
319
-void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
320
- target_ulong addr, uint32_t desc) \
321
-{ \
322
- sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
323
- sve_ld1bb_host, sve_ld1bb_tlb); \
324
+#define DO_LDN_1(N) \
325
+void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
326
+ target_ulong addr, uint32_t desc) \
327
+{ \
328
+ sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, 0, \
329
+ sve_ld1bb_host, sve_ld1bb_tlb, NULL); \
330
+} \
331
+void HELPER(sve_ld##N##bb_r_mte)(CPUARMState *env, void *vg, \
332
+ target_ulong addr, uint32_t desc) \
333
+{ \
334
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
335
+ sve_ld1bb_host, sve_ld1bb_tlb); \
336
}
337
338
-#define DO_LDN_2(N, SUFF, ESZ) \
339
-void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
340
- target_ulong addr, uint32_t desc) \
341
-{ \
342
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
343
- sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
344
-} \
345
-void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
346
- target_ulong addr, uint32_t desc) \
347
-{ \
348
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
349
- sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
350
+#define DO_LDN_2(N, SUFF, ESZ) \
351
+void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
352
+ target_ulong addr, uint32_t desc) \
353
+{ \
354
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
355
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb, NULL); \
356
+} \
357
+void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
358
+ target_ulong addr, uint32_t desc) \
359
+{ \
360
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
361
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb, NULL); \
362
+} \
363
+void HELPER(sve_ld##N##SUFF##_le_r_mte)(CPUARMState *env, void *vg, \
364
+ target_ulong addr, uint32_t desc) \
365
+{ \
366
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
367
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
368
+} \
369
+void HELPER(sve_ld##N##SUFF##_be_r_mte)(CPUARMState *env, void *vg, \
370
+ target_ulong addr, uint32_t desc) \
371
+{ \
372
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
373
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
374
}
375
376
DO_LDN_1(2)
377
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
378
index XXXXXXX..XXXXXXX 100644
379
--- a/target/arm/translate-sve.c
380
+++ b/target/arm/translate-sve.c
381
@@ -XXX,XX +XXX,XX @@ static const uint8_t dtype_esz[16] = {
382
};
383
384
static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
385
- int dtype, gen_helper_gvec_mem *fn)
386
+ int dtype, uint32_t mte_n, bool is_write,
387
+ gen_helper_gvec_mem *fn)
388
{
389
unsigned vsz = vec_full_reg_size(s);
390
TCGv_ptr t_pg;
391
TCGv_i32 t_desc;
392
- int desc;
393
+ int desc = 0;
394
395
- /* For e.g. LD4, there are not enough arguments to pass all 4
396
+ /*
397
+ * For e.g. LD4, there are not enough arguments to pass all 4
398
* registers as pointers, so encode the regno into the data field.
399
* For consistency, do this even for LD1.
400
+ * TODO: mte_n check here while callers are updated.
401
*/
402
- desc = simd_desc(vsz, vsz, zt);
403
+ if (mte_n && s->mte_active[0]) {
404
+ int msz = dtype_msz(dtype);
405
+
406
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
407
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
408
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
409
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
410
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
411
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
412
+ desc <<= SVE_MTEDESC_SHIFT;
413
+ }
414
+ desc = simd_desc(vsz, vsz, zt | desc);
415
t_desc = tcg_const_i32(desc);
416
t_pg = tcg_temp_new_ptr();
417
418
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
419
static void do_ld_zpa(DisasContext *s, int zt, int pg,
420
TCGv_i64 addr, int dtype, int nreg)
421
{
422
- static gen_helper_gvec_mem * const fns[2][16][4] = {
423
- /* Little-endian */
424
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
425
+ static gen_helper_gvec_mem * const fns[2][2][16][4] = {
426
+ { /* mte inactive, little-endian */
427
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
428
gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
429
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
430
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
431
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
432
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
433
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
434
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
435
436
- { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
437
- { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
438
- gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
439
- { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
440
- { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
441
+ { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
442
+ { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
443
+ gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
444
+ { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
445
+ { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
446
447
- { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
448
- { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
449
- { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
450
- gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
451
- { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
452
+ { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
453
+ { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
454
+ { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
455
+ gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
456
+ { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
457
458
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
459
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
460
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
461
- { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
462
- gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
463
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
464
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
465
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
466
+ { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
467
+ gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
468
469
- /* Big-endian */
470
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
471
- gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
472
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
473
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
474
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
475
+ /* mte inactive, big-endian */
476
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
477
+ gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
478
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
479
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
480
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
481
482
- { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
483
- { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
484
- gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
485
- { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
486
- { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
487
+ { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
488
+ { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
489
+ gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
490
+ { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
491
+ { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
492
493
- { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
494
- { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
495
- { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
496
- gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
497
- { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
498
+ { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
499
+ { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
500
+ { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
501
+ gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
502
+ { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
503
504
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
505
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
506
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
507
- { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
508
- gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } }
509
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
510
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
511
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
512
+ { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
513
+ gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } } },
514
+
515
+ { /* mte active, little-endian */
516
+ { { gen_helper_sve_ld1bb_r_mte,
517
+ gen_helper_sve_ld2bb_r_mte,
518
+ gen_helper_sve_ld3bb_r_mte,
519
+ gen_helper_sve_ld4bb_r_mte },
520
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
521
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
522
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
523
+
524
+ { gen_helper_sve_ld1sds_le_r_mte, NULL, NULL, NULL },
525
+ { gen_helper_sve_ld1hh_le_r_mte,
526
+ gen_helper_sve_ld2hh_le_r_mte,
527
+ gen_helper_sve_ld3hh_le_r_mte,
528
+ gen_helper_sve_ld4hh_le_r_mte },
529
+ { gen_helper_sve_ld1hsu_le_r_mte, NULL, NULL, NULL },
530
+ { gen_helper_sve_ld1hdu_le_r_mte, NULL, NULL, NULL },
531
+
532
+ { gen_helper_sve_ld1hds_le_r_mte, NULL, NULL, NULL },
533
+ { gen_helper_sve_ld1hss_le_r_mte, NULL, NULL, NULL },
534
+ { gen_helper_sve_ld1ss_le_r_mte,
535
+ gen_helper_sve_ld2ss_le_r_mte,
536
+ gen_helper_sve_ld3ss_le_r_mte,
537
+ gen_helper_sve_ld4ss_le_r_mte },
538
+ { gen_helper_sve_ld1sdu_le_r_mte, NULL, NULL, NULL },
539
+
540
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
541
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
542
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
543
+ { gen_helper_sve_ld1dd_le_r_mte,
544
+ gen_helper_sve_ld2dd_le_r_mte,
545
+ gen_helper_sve_ld3dd_le_r_mte,
546
+ gen_helper_sve_ld4dd_le_r_mte } },
547
+
548
+ /* mte active, big-endian */
549
+ { { gen_helper_sve_ld1bb_r_mte,
550
+ gen_helper_sve_ld2bb_r_mte,
551
+ gen_helper_sve_ld3bb_r_mte,
552
+ gen_helper_sve_ld4bb_r_mte },
553
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
554
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
555
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
556
+
557
+ { gen_helper_sve_ld1sds_be_r_mte, NULL, NULL, NULL },
558
+ { gen_helper_sve_ld1hh_be_r_mte,
559
+ gen_helper_sve_ld2hh_be_r_mte,
560
+ gen_helper_sve_ld3hh_be_r_mte,
561
+ gen_helper_sve_ld4hh_be_r_mte },
562
+ { gen_helper_sve_ld1hsu_be_r_mte, NULL, NULL, NULL },
563
+ { gen_helper_sve_ld1hdu_be_r_mte, NULL, NULL, NULL },
564
+
565
+ { gen_helper_sve_ld1hds_be_r_mte, NULL, NULL, NULL },
566
+ { gen_helper_sve_ld1hss_be_r_mte, NULL, NULL, NULL },
567
+ { gen_helper_sve_ld1ss_be_r_mte,
568
+ gen_helper_sve_ld2ss_be_r_mte,
569
+ gen_helper_sve_ld3ss_be_r_mte,
570
+ gen_helper_sve_ld4ss_be_r_mte },
571
+ { gen_helper_sve_ld1sdu_be_r_mte, NULL, NULL, NULL },
572
+
573
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
574
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
575
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
576
+ { gen_helper_sve_ld1dd_be_r_mte,
577
+ gen_helper_sve_ld2dd_be_r_mte,
578
+ gen_helper_sve_ld3dd_be_r_mte,
579
+ gen_helper_sve_ld4dd_be_r_mte } } },
580
};
581
- gen_helper_gvec_mem *fn = fns[s->be_data == MO_BE][dtype][nreg];
582
+ gen_helper_gvec_mem *fn
583
+ = fns[s->mte_active[0]][s->be_data == MO_BE][dtype][nreg];
584
585
- /* While there are holes in the table, they are not
586
+ /*
587
+ * While there are holes in the table, they are not
588
* accessible via the instruction encoding.
589
*/
590
assert(fn != NULL);
591
- do_mem_zpa(s, zt, pg, addr, dtype, fn);
592
+ do_mem_zpa(s, zt, pg, addr, dtype, nreg, false, fn);
593
}
594
595
static bool trans_LD_zprr(DisasContext *s, arg_rprr_load *a)
596
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
597
TCGv_i64 addr = new_tmp_a64(s);
598
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
599
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
600
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
601
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
602
fns[s->be_data == MO_BE][a->dtype]);
603
}
604
return true;
605
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
606
TCGv_i64 addr = new_tmp_a64(s);
607
608
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
609
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
610
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
611
fns[s->be_data == MO_BE][a->dtype]);
612
}
613
return true;
614
@@ -XXX,XX +XXX,XX @@ static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
615
fn = fn_multiple[be][nreg - 1][msz];
616
}
617
assert(fn != NULL);
618
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), fn);
619
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
620
}
621
622
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
150
--
623
--
151
2.20.1
624
2.20.1
152
625
153
626
diff view generated by jsdifflib
1
From: Eric Auger <eric.auger@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
The SMMUNotifierNode struct is not necessary and brings extra
3
Because the elements are sequential, we can eliminate many tests all
4
complexity so let's remove it. We now directly track the SMMUDevices
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
which have registered IOMMU MR notifiers.
5
6
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
This is inspired from the same transformation on intel-iommu
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
done in commit b4a4ba0d68f50f218ee3957b6638dbee32a5eeef
8
Message-id: 20200626033144.790098-35-richard.henderson@linaro.org
9
("intel-iommu: remove IntelIOMMUNotifierNode")
10
11
Signed-off-by: Eric Auger <eric.auger@redhat.com>
12
Reviewed-by: Peter Xu <peterx@redhat.com>
13
Message-id: 20190409160219.19026-1-eric.auger@redhat.com
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
---
10
---
16
include/hw/arm/smmu-common.h | 8 ++------
11
target/arm/helper-sve.h | 47 +++++++++++
17
hw/arm/smmu-common.c | 6 +++---
12
target/arm/sve_helper.c | 95 ++++++++++++++++------
18
hw/arm/smmuv3.c | 28 +++++++---------------------
13
target/arm/translate-sve.c | 162 ++++++++++++++++++++++++-------------
19
3 files changed, 12 insertions(+), 30 deletions(-)
14
3 files changed, 226 insertions(+), 78 deletions(-)
20
15
21
diff --git a/include/hw/arm/smmu-common.h b/include/hw/arm/smmu-common.h
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
22
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
23
--- a/include/hw/arm/smmu-common.h
18
--- a/target/arm/helper-sve.h
24
+++ b/include/hw/arm/smmu-common.h
19
+++ b/target/arm/helper-sve.h
25
@@ -XXX,XX +XXX,XX @@ typedef struct SMMUDevice {
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_st1hd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
AddressSpace as;
21
DEF_HELPER_FLAGS_4(sve_st1sd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
uint32_t cfg_cache_hits;
22
DEF_HELPER_FLAGS_4(sve_st1sd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
uint32_t cfg_cache_misses;
23
29
+ QLIST_ENTRY(SMMUDevice) next;
24
+DEF_HELPER_FLAGS_4(sve_st1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
} SMMUDevice;
25
+DEF_HELPER_FLAGS_4(sve_st2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
26
+DEF_HELPER_FLAGS_4(sve_st3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
32
-typedef struct SMMUNotifierNode {
27
+DEF_HELPER_FLAGS_4(sve_st4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
33
- SMMUDevice *sdev;
28
+
34
- QLIST_ENTRY(SMMUNotifierNode) next;
29
+DEF_HELPER_FLAGS_4(sve_st1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
-} SMMUNotifierNode;
30
+DEF_HELPER_FLAGS_4(sve_st2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
-
31
+DEF_HELPER_FLAGS_4(sve_st3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
typedef struct SMMUPciBus {
32
+DEF_HELPER_FLAGS_4(sve_st4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
PCIBus *bus;
33
+
39
SMMUDevice *pbdev[0]; /* Parent array is sparse, so dynamically alloc */
34
+DEF_HELPER_FLAGS_4(sve_st1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
@@ -XXX,XX +XXX,XX @@ typedef struct SMMUState {
35
+DEF_HELPER_FLAGS_4(sve_st2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
GHashTable *iotlb;
36
+DEF_HELPER_FLAGS_4(sve_st3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
SMMUPciBus *smmu_pcibus_by_bus_num[SMMU_PCI_BUS_MAX];
37
+DEF_HELPER_FLAGS_4(sve_st4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
PCIBus *pci_bus;
38
+
44
- QLIST_HEAD(, SMMUNotifierNode) notifiers_list;
39
+DEF_HELPER_FLAGS_4(sve_st1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+ QLIST_HEAD(, SMMUDevice) devices_with_notifiers;
40
+DEF_HELPER_FLAGS_4(sve_st2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
uint8_t bus_num;
41
+DEF_HELPER_FLAGS_4(sve_st3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
PCIBus *primary_bus;
42
+DEF_HELPER_FLAGS_4(sve_st4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
} SMMUState;
43
+
49
diff --git a/hw/arm/smmu-common.c b/hw/arm/smmu-common.c
44
+DEF_HELPER_FLAGS_4(sve_st1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_st2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_st3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_st4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
+
49
+DEF_HELPER_FLAGS_4(sve_st1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_st2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_st3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_4(sve_st4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_st1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_st2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_st3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_4(sve_st4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_4(sve_st1bh_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_st1bs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_st1bd_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_st1hs_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_st1hd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_st1hs_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_st1hd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_st1sd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_st1sd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+
71
DEF_HELPER_FLAGS_6(sve_ldbsu_zsu, TCG_CALL_NO_WG,
72
void, env, ptr, ptr, ptr, tl, i32)
73
DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu, TCG_CALL_NO_WG,
74
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
50
index XXXXXXX..XXXXXXX 100644
75
index XXXXXXX..XXXXXXX 100644
51
--- a/hw/arm/smmu-common.c
76
--- a/target/arm/sve_helper.c
52
+++ b/hw/arm/smmu-common.c
77
+++ b/target/arm/sve_helper.c
53
@@ -XXX,XX +XXX,XX @@ inline void smmu_inv_notifiers_mr(IOMMUMemoryRegion *mr)
78
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_LDNF1_2(dd, MO_64, MO_64)
54
/* Unmap all notifiers of all mr's */
79
*/
55
void smmu_inv_notifiers_all(SMMUState *s)
80
81
static inline QEMU_ALWAYS_INLINE
82
-void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
83
- const uintptr_t retaddr, const int esz,
84
- const int msz, const int N,
85
+void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr,
86
+ uint32_t desc, const uintptr_t retaddr,
87
+ const int esz, const int msz, const int N, uint32_t mtedesc,
88
sve_ldst1_host_fn *host_fn,
89
- sve_ldst1_tlb_fn *tlb_fn)
90
+ sve_ldst1_tlb_fn *tlb_fn,
91
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
56
{
92
{
57
- SMMUNotifierNode *node;
93
const unsigned rd = simd_data(desc);
58
+ SMMUDevice *sdev;
94
const intptr_t reg_max = simd_oprsz(desc);
59
95
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
60
- QLIST_FOREACH(node, &s->notifiers_list, next) {
96
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
61
- smmu_inv_notifiers_mr(&node->sdev->iommu);
97
BP_MEM_WRITE, retaddr);
62
+ QLIST_FOREACH(sdev, &s->devices_with_notifiers, next) {
98
63
+ smmu_inv_notifiers_mr(&sdev->iommu);
99
- /* TODO: MTE check. */
100
+ /*
101
+ * Handle mte checks for all active elements.
102
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
103
+ */
104
+ if (mte_check_fn && mtedesc) {
105
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
106
+ mtedesc, retaddr);
107
+ }
108
109
flags = info.page[0].flags | info.page[1].flags;
110
if (unlikely(flags != 0)) {
111
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
64
}
112
}
65
}
113
}
66
114
67
diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c
115
-#define DO_STN_1(N, NAME, ESZ) \
116
-void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
117
- target_ulong addr, uint32_t desc) \
118
-{ \
119
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
120
- sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
121
+static inline QEMU_ALWAYS_INLINE
122
+void sve_stN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
123
+ uint32_t desc, const uintptr_t ra,
124
+ const int esz, const int msz, const int N,
125
+ sve_ldst1_host_fn *host_fn,
126
+ sve_ldst1_tlb_fn *tlb_fn)
127
+{
128
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
129
+ int bit55 = extract64(addr, 55, 1);
130
+
131
+ /* Remove mtedesc from the normal sve descriptor. */
132
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
133
+
134
+ /* Perform gross MTE suppression early. */
135
+ if (!tbi_check(desc, bit55) ||
136
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
137
+ mtedesc = 0;
138
+ }
139
+
140
+ sve_stN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
141
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
142
}
143
144
-#define DO_STN_2(N, NAME, ESZ, MSZ) \
145
-void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
146
- target_ulong addr, uint32_t desc) \
147
-{ \
148
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
149
- sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
150
-} \
151
-void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
152
- target_ulong addr, uint32_t desc) \
153
-{ \
154
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
155
- sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
156
+#define DO_STN_1(N, NAME, ESZ) \
157
+void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
158
+ target_ulong addr, uint32_t desc) \
159
+{ \
160
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, 0, \
161
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb, NULL); \
162
+} \
163
+void HELPER(sve_st##N##NAME##_r_mte)(CPUARMState *env, void *vg, \
164
+ target_ulong addr, uint32_t desc) \
165
+{ \
166
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
167
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
168
+}
169
+
170
+#define DO_STN_2(N, NAME, ESZ, MSZ) \
171
+void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
172
+ target_ulong addr, uint32_t desc) \
173
+{ \
174
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
175
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb, NULL); \
176
+} \
177
+void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
178
+ target_ulong addr, uint32_t desc) \
179
+{ \
180
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
181
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb, NULL); \
182
+} \
183
+void HELPER(sve_st##N##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
184
+ target_ulong addr, uint32_t desc) \
185
+{ \
186
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
187
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
188
+} \
189
+void HELPER(sve_st##N##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
190
+ target_ulong addr, uint32_t desc) \
191
+{ \
192
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
193
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
194
}
195
196
DO_STN_1(1, bb, MO_8)
197
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
68
index XXXXXXX..XXXXXXX 100644
198
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/smmuv3.c
199
--- a/target/arm/translate-sve.c
70
+++ b/hw/arm/smmuv3.c
200
+++ b/target/arm/translate-sve.c
71
@@ -XXX,XX +XXX,XX @@ static void smmuv3_notify_iova(IOMMUMemoryRegion *mr,
201
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
72
/* invalidate an asid/iova tuple in all mr's */
202
static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
73
static void smmuv3_inv_notifiers_iova(SMMUState *s, int asid, dma_addr_t iova)
203
int msz, int esz, int nreg)
74
{
204
{
75
- SMMUNotifierNode *node;
205
- static gen_helper_gvec_mem * const fn_single[2][4][4] = {
76
+ SMMUDevice *sdev;
206
- { { gen_helper_sve_st1bb_r,
77
207
- gen_helper_sve_st1bh_r,
78
- QLIST_FOREACH(node, &s->notifiers_list, next) {
208
- gen_helper_sve_st1bs_r,
79
- IOMMUMemoryRegion *mr = &node->sdev->iommu;
209
- gen_helper_sve_st1bd_r },
80
+ QLIST_FOREACH(sdev, &s->devices_with_notifiers, next) {
210
- { NULL,
81
+ IOMMUMemoryRegion *mr = &sdev->iommu;
211
- gen_helper_sve_st1hh_le_r,
82
IOMMUNotifier *n;
212
- gen_helper_sve_st1hs_le_r,
83
213
- gen_helper_sve_st1hd_le_r },
84
trace_smmuv3_inv_notifiers_iova(mr->parent_obj.name, asid, iova);
214
- { NULL, NULL,
85
@@ -XXX,XX +XXX,XX @@ static void smmuv3_notify_flag_changed(IOMMUMemoryRegion *iommu,
215
- gen_helper_sve_st1ss_le_r,
86
SMMUDevice *sdev = container_of(iommu, SMMUDevice, iommu);
216
- gen_helper_sve_st1sd_le_r },
87
SMMUv3State *s3 = sdev->smmu;
217
- { NULL, NULL, NULL,
88
SMMUState *s = &(s3->smmu_state);
218
- gen_helper_sve_st1dd_le_r } },
89
- SMMUNotifierNode *node = NULL;
219
- { { gen_helper_sve_st1bb_r,
90
- SMMUNotifierNode *next_node = NULL;
220
- gen_helper_sve_st1bh_r,
91
221
- gen_helper_sve_st1bs_r,
92
if (new & IOMMU_NOTIFIER_MAP) {
222
- gen_helper_sve_st1bd_r },
93
int bus_num = pci_bus_num(sdev->bus);
223
- { NULL,
94
@@ -XXX,XX +XXX,XX @@ static void smmuv3_notify_flag_changed(IOMMUMemoryRegion *iommu,
224
- gen_helper_sve_st1hh_be_r,
95
225
- gen_helper_sve_st1hs_be_r,
96
if (old == IOMMU_NOTIFIER_NONE) {
226
- gen_helper_sve_st1hd_be_r },
97
trace_smmuv3_notify_flag_add(iommu->parent_obj.name);
227
- { NULL, NULL,
98
- node = g_malloc0(sizeof(*node));
228
- gen_helper_sve_st1ss_be_r,
99
- node->sdev = sdev;
229
- gen_helper_sve_st1sd_be_r },
100
- QLIST_INSERT_HEAD(&s->notifiers_list, node, next);
230
- { NULL, NULL, NULL,
101
- return;
231
- gen_helper_sve_st1dd_be_r } },
102
- }
232
+ static gen_helper_gvec_mem * const fn_single[2][2][4][4] = {
103
-
233
+ { { { gen_helper_sve_st1bb_r,
104
- /* update notifier node with new flags */
234
+ gen_helper_sve_st1bh_r,
105
- QLIST_FOREACH_SAFE(node, &s->notifiers_list, next, next_node) {
235
+ gen_helper_sve_st1bs_r,
106
- if (node->sdev == sdev) {
236
+ gen_helper_sve_st1bd_r },
107
- if (new == IOMMU_NOTIFIER_NONE) {
237
+ { NULL,
108
- trace_smmuv3_notify_flag_del(iommu->parent_obj.name);
238
+ gen_helper_sve_st1hh_le_r,
109
- QLIST_REMOVE(node, next);
239
+ gen_helper_sve_st1hs_le_r,
110
- g_free(node);
240
+ gen_helper_sve_st1hd_le_r },
111
- }
241
+ { NULL, NULL,
112
- return;
242
+ gen_helper_sve_st1ss_le_r,
113
- }
243
+ gen_helper_sve_st1sd_le_r },
114
+ QLIST_INSERT_HEAD(&s->devices_with_notifiers, sdev, next);
244
+ { NULL, NULL, NULL,
115
+ } else if (new == IOMMU_NOTIFIER_NONE) {
245
+ gen_helper_sve_st1dd_le_r } },
116
+ trace_smmuv3_notify_flag_del(iommu->parent_obj.name);
246
+ { { gen_helper_sve_st1bb_r,
117
+ QLIST_REMOVE(sdev, next);
247
+ gen_helper_sve_st1bh_r,
248
+ gen_helper_sve_st1bs_r,
249
+ gen_helper_sve_st1bd_r },
250
+ { NULL,
251
+ gen_helper_sve_st1hh_be_r,
252
+ gen_helper_sve_st1hs_be_r,
253
+ gen_helper_sve_st1hd_be_r },
254
+ { NULL, NULL,
255
+ gen_helper_sve_st1ss_be_r,
256
+ gen_helper_sve_st1sd_be_r },
257
+ { NULL, NULL, NULL,
258
+ gen_helper_sve_st1dd_be_r } } },
259
+
260
+ { { { gen_helper_sve_st1bb_r_mte,
261
+ gen_helper_sve_st1bh_r_mte,
262
+ gen_helper_sve_st1bs_r_mte,
263
+ gen_helper_sve_st1bd_r_mte },
264
+ { NULL,
265
+ gen_helper_sve_st1hh_le_r_mte,
266
+ gen_helper_sve_st1hs_le_r_mte,
267
+ gen_helper_sve_st1hd_le_r_mte },
268
+ { NULL, NULL,
269
+ gen_helper_sve_st1ss_le_r_mte,
270
+ gen_helper_sve_st1sd_le_r_mte },
271
+ { NULL, NULL, NULL,
272
+ gen_helper_sve_st1dd_le_r_mte } },
273
+ { { gen_helper_sve_st1bb_r_mte,
274
+ gen_helper_sve_st1bh_r_mte,
275
+ gen_helper_sve_st1bs_r_mte,
276
+ gen_helper_sve_st1bd_r_mte },
277
+ { NULL,
278
+ gen_helper_sve_st1hh_be_r_mte,
279
+ gen_helper_sve_st1hs_be_r_mte,
280
+ gen_helper_sve_st1hd_be_r_mte },
281
+ { NULL, NULL,
282
+ gen_helper_sve_st1ss_be_r_mte,
283
+ gen_helper_sve_st1sd_be_r_mte },
284
+ { NULL, NULL, NULL,
285
+ gen_helper_sve_st1dd_be_r_mte } } },
286
};
287
- static gen_helper_gvec_mem * const fn_multiple[2][3][4] = {
288
- { { gen_helper_sve_st2bb_r,
289
- gen_helper_sve_st2hh_le_r,
290
- gen_helper_sve_st2ss_le_r,
291
- gen_helper_sve_st2dd_le_r },
292
- { gen_helper_sve_st3bb_r,
293
- gen_helper_sve_st3hh_le_r,
294
- gen_helper_sve_st3ss_le_r,
295
- gen_helper_sve_st3dd_le_r },
296
- { gen_helper_sve_st4bb_r,
297
- gen_helper_sve_st4hh_le_r,
298
- gen_helper_sve_st4ss_le_r,
299
- gen_helper_sve_st4dd_le_r } },
300
- { { gen_helper_sve_st2bb_r,
301
- gen_helper_sve_st2hh_be_r,
302
- gen_helper_sve_st2ss_be_r,
303
- gen_helper_sve_st2dd_be_r },
304
- { gen_helper_sve_st3bb_r,
305
- gen_helper_sve_st3hh_be_r,
306
- gen_helper_sve_st3ss_be_r,
307
- gen_helper_sve_st3dd_be_r },
308
- { gen_helper_sve_st4bb_r,
309
- gen_helper_sve_st4hh_be_r,
310
- gen_helper_sve_st4ss_be_r,
311
- gen_helper_sve_st4dd_be_r } },
312
+ static gen_helper_gvec_mem * const fn_multiple[2][2][3][4] = {
313
+ { { { gen_helper_sve_st2bb_r,
314
+ gen_helper_sve_st2hh_le_r,
315
+ gen_helper_sve_st2ss_le_r,
316
+ gen_helper_sve_st2dd_le_r },
317
+ { gen_helper_sve_st3bb_r,
318
+ gen_helper_sve_st3hh_le_r,
319
+ gen_helper_sve_st3ss_le_r,
320
+ gen_helper_sve_st3dd_le_r },
321
+ { gen_helper_sve_st4bb_r,
322
+ gen_helper_sve_st4hh_le_r,
323
+ gen_helper_sve_st4ss_le_r,
324
+ gen_helper_sve_st4dd_le_r } },
325
+ { { gen_helper_sve_st2bb_r,
326
+ gen_helper_sve_st2hh_be_r,
327
+ gen_helper_sve_st2ss_be_r,
328
+ gen_helper_sve_st2dd_be_r },
329
+ { gen_helper_sve_st3bb_r,
330
+ gen_helper_sve_st3hh_be_r,
331
+ gen_helper_sve_st3ss_be_r,
332
+ gen_helper_sve_st3dd_be_r },
333
+ { gen_helper_sve_st4bb_r,
334
+ gen_helper_sve_st4hh_be_r,
335
+ gen_helper_sve_st4ss_be_r,
336
+ gen_helper_sve_st4dd_be_r } } },
337
+ { { { gen_helper_sve_st2bb_r_mte,
338
+ gen_helper_sve_st2hh_le_r_mte,
339
+ gen_helper_sve_st2ss_le_r_mte,
340
+ gen_helper_sve_st2dd_le_r_mte },
341
+ { gen_helper_sve_st3bb_r_mte,
342
+ gen_helper_sve_st3hh_le_r_mte,
343
+ gen_helper_sve_st3ss_le_r_mte,
344
+ gen_helper_sve_st3dd_le_r_mte },
345
+ { gen_helper_sve_st4bb_r_mte,
346
+ gen_helper_sve_st4hh_le_r_mte,
347
+ gen_helper_sve_st4ss_le_r_mte,
348
+ gen_helper_sve_st4dd_le_r_mte } },
349
+ { { gen_helper_sve_st2bb_r_mte,
350
+ gen_helper_sve_st2hh_be_r_mte,
351
+ gen_helper_sve_st2ss_be_r_mte,
352
+ gen_helper_sve_st2dd_be_r_mte },
353
+ { gen_helper_sve_st3bb_r_mte,
354
+ gen_helper_sve_st3hh_be_r_mte,
355
+ gen_helper_sve_st3ss_be_r_mte,
356
+ gen_helper_sve_st3dd_be_r_mte },
357
+ { gen_helper_sve_st4bb_r_mte,
358
+ gen_helper_sve_st4hh_be_r_mte,
359
+ gen_helper_sve_st4ss_be_r_mte,
360
+ gen_helper_sve_st4dd_be_r_mte } } },
361
};
362
gen_helper_gvec_mem *fn;
363
int be = s->be_data == MO_BE;
364
365
if (nreg == 0) {
366
/* ST1 */
367
- fn = fn_single[be][msz][esz];
368
+ fn = fn_single[s->mte_active[0]][be][msz][esz];
369
+ nreg = 1;
370
} else {
371
/* ST2, ST3, ST4 -- msz == esz, enforced by encoding */
372
assert(msz == esz);
373
- fn = fn_multiple[be][nreg - 1][msz];
374
+ fn = fn_multiple[s->mte_active[0]][be][nreg - 1][msz];
118
}
375
}
376
assert(fn != NULL);
377
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
378
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), nreg, true, fn);
119
}
379
}
120
380
381
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
121
--
382
--
122
2.20.1
383
2.20.1
123
384
124
385
diff view generated by jsdifflib
1
The magic value pushed onto the callee stack as an integrity
1
From: Richard Henderson <richard.henderson@linaro.org>
2
check is different if floating point is present.
3
2
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-36-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20190416125744.27770-15-peter.maydell@linaro.org
7
---
10
---
8
target/arm/helper.c | 22 +++++++++++++++++++---
11
target/arm/helper-sve.h | 98 ++++++++++++++++
9
1 file changed, 19 insertions(+), 3 deletions(-)
12
target/arm/sve_helper.c | 99 ++++++++++++++--
13
target/arm/translate-sve.c | 232 +++++++++++++++++++++++++------------
14
3 files changed, 343 insertions(+), 86 deletions(-)
10
15
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
12
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/helper.c
18
--- a/target/arm/helper-sve.h
14
+++ b/target/arm/helper.c
19
+++ b/target/arm/helper-sve.h
15
@@ -XXX,XX +XXX,XX @@ load_fail:
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
16
return false;
21
DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_ldff1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_ldff1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_ldff1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_ldff1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+DEF_HELPER_FLAGS_4(sve_ldff1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_4(sve_ldff1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_ldff1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+
32
+DEF_HELPER_FLAGS_4(sve_ldff1hh_le_r_mte, TCG_CALL_NO_WG,
33
+ void, env, ptr, tl, i32)
34
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_le_r_mte, TCG_CALL_NO_WG,
35
+ void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_le_r_mte, TCG_CALL_NO_WG,
37
+ void, env, ptr, tl, i32)
38
+DEF_HELPER_FLAGS_4(sve_ldff1hss_le_r_mte, TCG_CALL_NO_WG,
39
+ void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ldff1hds_le_r_mte, TCG_CALL_NO_WG,
41
+ void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ldff1hh_be_r_mte, TCG_CALL_NO_WG,
44
+ void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_be_r_mte, TCG_CALL_NO_WG,
46
+ void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_be_r_mte, TCG_CALL_NO_WG,
48
+ void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ldff1hss_be_r_mte, TCG_CALL_NO_WG,
50
+ void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ldff1hds_be_r_mte, TCG_CALL_NO_WG,
52
+ void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_ldff1ss_le_r_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_le_r_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, tl, i32)
58
+DEF_HELPER_FLAGS_4(sve_ldff1sds_le_r_mte, TCG_CALL_NO_WG,
59
+ void, env, ptr, tl, i32)
60
+
61
+DEF_HELPER_FLAGS_4(sve_ldff1ss_be_r_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_be_r_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r_mte, TCG_CALL_NO_WG,
69
+ void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r_mte, TCG_CALL_NO_WG,
71
+ void, env, ptr, tl, i32)
72
+
73
DEF_HELPER_FLAGS_4(sve_ldnf1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
80
+DEF_HELPER_FLAGS_4(sve_ldnf1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ldnf1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_4(sve_ldnf1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_4(sve_ldnf1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ldnf1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+
88
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_le_r_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_le_r_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_le_r_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_le_r_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_le_r_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, tl, i32)
98
+
99
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_be_r_mte, TCG_CALL_NO_WG,
100
+ void, env, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_be_r_mte, TCG_CALL_NO_WG,
102
+ void, env, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_be_r_mte, TCG_CALL_NO_WG,
104
+ void, env, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_be_r_mte, TCG_CALL_NO_WG,
106
+ void, env, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_be_r_mte, TCG_CALL_NO_WG,
108
+ void, env, ptr, tl, i32)
109
+
110
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_le_r_mte, TCG_CALL_NO_WG,
111
+ void, env, ptr, tl, i32)
112
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_le_r_mte, TCG_CALL_NO_WG,
113
+ void, env, ptr, tl, i32)
114
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_le_r_mte, TCG_CALL_NO_WG,
115
+ void, env, ptr, tl, i32)
116
+
117
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_be_r_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_be_r_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, tl, i32)
123
+
124
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r_mte, TCG_CALL_NO_WG,
125
+ void, env, ptr, tl, i32)
126
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r_mte, TCG_CALL_NO_WG,
127
+ void, env, ptr, tl, i32)
128
+
129
DEF_HELPER_FLAGS_4(sve_st1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
130
DEF_HELPER_FLAGS_4(sve_st2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
131
DEF_HELPER_FLAGS_4(sve_st3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
132
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/target/arm/sve_helper.c
135
+++ b/target/arm/sve_helper.c
136
@@ -XXX,XX +XXX,XX @@ static void record_fault(CPUARMState *env, uintptr_t i, uintptr_t oprsz)
137
*/
138
static inline QEMU_ALWAYS_INLINE
139
void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
140
- uint32_t desc, const uintptr_t retaddr,
141
+ uint32_t desc, const uintptr_t retaddr, uint32_t mtedesc,
142
const int esz, const int msz, const SVEContFault fault,
143
sve_ldst1_host_fn *host_fn,
144
sve_ldst1_tlb_fn *tlb_fn)
145
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
146
mem_off = info.mem_off_first[0];
147
flags = info.page[0].flags;
148
149
+ /*
150
+ * Disable MTE checking if the Tagged bit is not set. Since TBI must
151
+ * be set within MTEDESC for MTE, !mtedesc => !mte_active.
152
+ */
153
+ if (arm_tlb_mte_tagged(&info.page[0].attrs)) {
154
+ mtedesc = 0;
155
+ }
156
+
157
if (fault == FAULT_FIRST) {
158
+ /* Trapping mte check for the first-fault element. */
159
+ if (mtedesc) {
160
+ mte_check1(env, mtedesc, addr + mem_off, retaddr);
161
+ }
162
+
163
/*
164
* Special handling of the first active element,
165
* if it crosses a page boundary or is MMIO.
166
*/
167
bool is_split = mem_off == info.mem_off_split;
168
- /* TODO: MTE check. */
169
if (unlikely(flags != 0) || unlikely(is_split)) {
170
/*
171
* Use the slow path for cross-page handling.
172
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
173
/* Watchpoint hit, see below. */
174
goto do_fault;
175
}
176
- /* TODO: MTE check. */
177
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
178
+ goto do_fault;
179
+ }
180
/*
181
* Use the slow path for cross-page handling.
182
* This is RAM, without a watchpoint, and will not trap.
183
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
184
& BP_MEM_READ)) {
185
goto do_fault;
186
}
187
- /* TODO: MTE check. */
188
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
189
+ goto do_fault;
190
+ }
191
host_fn(vd, reg_off, host + mem_off);
192
}
193
reg_off += 1 << esz;
194
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
195
record_fault(env, reg_off, reg_max);
17
}
196
}
18
197
19
+static uint32_t v7m_integrity_sig(CPUARMState *env, uint32_t lr)
198
-#define DO_LDFF1_LDNF1_1(PART, ESZ) \
199
+static inline QEMU_ALWAYS_INLINE
200
+void sve_ldnfff1_r_mte(CPUARMState *env, void *vg, target_ulong addr,
201
+ uint32_t desc, const uintptr_t retaddr,
202
+ const int esz, const int msz, const SVEContFault fault,
203
+ sve_ldst1_host_fn *host_fn,
204
+ sve_ldst1_tlb_fn *tlb_fn)
20
+{
205
+{
21
+ /*
206
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
22
+ * Return the integrity signature value for the callee-saves
207
+ int bit55 = extract64(addr, 55, 1);
23
+ * stack frame section. @lr is the exception return payload/LR value
208
+
24
+ * whose FType bit forms bit 0 of the signature if FP is present.
209
+ /* Remove mtedesc from the normal sve descriptor. */
25
+ */
210
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
26
+ uint32_t sig = 0xfefa125a;
211
+
27
+
212
+ /* Perform gross MTE suppression early. */
28
+ if (!arm_feature(env, ARM_FEATURE_VFP) || (lr & R_V7M_EXCRET_FTYPE_MASK)) {
213
+ if (!tbi_check(desc, bit55) ||
29
+ sig |= 1;
214
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
215
+ mtedesc = 0;
30
+ }
216
+ }
31
+ return sig;
217
+
218
+ sve_ldnfff1_r(env, vg, addr, desc, retaddr, mtedesc,
219
+ esz, msz, fault, host_fn, tlb_fn);
32
+}
220
+}
33
+
221
+
34
static bool v7m_push_callee_stack(ARMCPU *cpu, uint32_t lr, bool dotailchain,
222
+#define DO_LDFF1_LDNF1_1(PART, ESZ) \
35
bool ignore_faults)
223
void HELPER(sve_ldff1##PART##_r)(CPUARMState *env, void *vg, \
224
target_ulong addr, uint32_t desc) \
225
{ \
226
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
227
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_FIRST, \
228
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
229
} \
230
void HELPER(sve_ldnf1##PART##_r)(CPUARMState *env, void *vg, \
231
target_ulong addr, uint32_t desc) \
232
{ \
233
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
234
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_NO, \
235
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
236
+} \
237
+void HELPER(sve_ldff1##PART##_r_mte)(CPUARMState *env, void *vg, \
238
+ target_ulong addr, uint32_t desc) \
239
+{ \
240
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
241
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
242
+} \
243
+void HELPER(sve_ldnf1##PART##_r_mte)(CPUARMState *env, void *vg, \
244
+ target_ulong addr, uint32_t desc) \
245
+{ \
246
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
247
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
248
}
249
250
-#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
251
+#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
252
void HELPER(sve_ldff1##PART##_le_r)(CPUARMState *env, void *vg, \
253
target_ulong addr, uint32_t desc) \
254
{ \
255
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
256
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
257
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
258
} \
259
void HELPER(sve_ldnf1##PART##_le_r)(CPUARMState *env, void *vg, \
260
target_ulong addr, uint32_t desc) \
261
{ \
262
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
263
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
264
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
265
} \
266
void HELPER(sve_ldff1##PART##_be_r)(CPUARMState *env, void *vg, \
267
target_ulong addr, uint32_t desc) \
268
{ \
269
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
270
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
271
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
272
} \
273
void HELPER(sve_ldnf1##PART##_be_r)(CPUARMState *env, void *vg, \
274
target_ulong addr, uint32_t desc) \
275
{ \
276
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
277
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
278
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
279
+} \
280
+void HELPER(sve_ldff1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
284
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
285
+} \
286
+void HELPER(sve_ldnf1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
287
+ target_ulong addr, uint32_t desc) \
288
+{ \
289
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
290
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
291
+} \
292
+void HELPER(sve_ldff1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
293
+ target_ulong addr, uint32_t desc) \
294
+{ \
295
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
296
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
297
+} \
298
+void HELPER(sve_ldnf1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
299
+ target_ulong addr, uint32_t desc) \
300
+{ \
301
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
302
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
303
}
304
305
DO_LDFF1_LDNF1_1(bb, MO_8)
306
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
307
index XXXXXXX..XXXXXXX 100644
308
--- a/target/arm/translate-sve.c
309
+++ b/target/arm/translate-sve.c
310
@@ -XXX,XX +XXX,XX @@ static bool trans_LD_zpri(DisasContext *s, arg_rpri_load *a)
311
312
static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
36
{
313
{
37
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_callee_stack(ARMCPU *cpu, uint32_t lr, bool dotailchain,
314
- static gen_helper_gvec_mem * const fns[2][16] = {
38
bool stacked_ok;
315
- /* Little-endian */
39
uint32_t limit;
316
- { gen_helper_sve_ldff1bb_r,
40
bool want_psp;
317
- gen_helper_sve_ldff1bhu_r,
41
+ uint32_t sig;
318
- gen_helper_sve_ldff1bsu_r,
42
319
- gen_helper_sve_ldff1bdu_r,
43
if (dotailchain) {
320
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
44
bool mode = lr & R_V7M_EXCRET_MODE_MASK;
321
+ { /* mte inactive, little-endian */
45
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_callee_stack(ARMCPU *cpu, uint32_t lr, bool dotailchain,
322
+ { gen_helper_sve_ldff1bb_r,
46
/* Write as much of the stack frame as we can. A write failure may
323
+ gen_helper_sve_ldff1bhu_r,
47
* cause us to pend a derived exception.
324
+ gen_helper_sve_ldff1bsu_r,
48
*/
325
+ gen_helper_sve_ldff1bdu_r,
49
+ sig = v7m_integrity_sig(env, lr);
326
50
stacked_ok =
327
- gen_helper_sve_ldff1sds_le_r,
51
- v7m_stack_write(cpu, frameptr, 0xfefa125b, mmu_idx, ignore_faults) &&
328
- gen_helper_sve_ldff1hh_le_r,
52
+ v7m_stack_write(cpu, frameptr, sig, mmu_idx, ignore_faults) &&
329
- gen_helper_sve_ldff1hsu_le_r,
53
v7m_stack_write(cpu, frameptr + 0x8, env->regs[4], mmu_idx,
330
- gen_helper_sve_ldff1hdu_le_r,
54
ignore_faults) &&
331
+ gen_helper_sve_ldff1sds_le_r,
55
v7m_stack_write(cpu, frameptr + 0xc, env->regs[5], mmu_idx,
332
+ gen_helper_sve_ldff1hh_le_r,
56
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
333
+ gen_helper_sve_ldff1hsu_le_r,
57
if (return_to_secure &&
334
+ gen_helper_sve_ldff1hdu_le_r,
58
((excret & R_V7M_EXCRET_ES_MASK) == 0 ||
335
59
(excret & R_V7M_EXCRET_DCRS_MASK) == 0)) {
336
- gen_helper_sve_ldff1hds_le_r,
60
- uint32_t expected_sig = 0xfefa125b;
337
- gen_helper_sve_ldff1hss_le_r,
61
uint32_t actual_sig;
338
- gen_helper_sve_ldff1ss_le_r,
62
339
- gen_helper_sve_ldff1sdu_le_r,
63
pop_ok = v7m_stack_read(cpu, &actual_sig, frameptr, mmu_idx);
340
+ gen_helper_sve_ldff1hds_le_r,
64
341
+ gen_helper_sve_ldff1hss_le_r,
65
- if (pop_ok && expected_sig != actual_sig) {
342
+ gen_helper_sve_ldff1ss_le_r,
66
+ if (pop_ok && v7m_integrity_sig(env, excret) != actual_sig) {
343
+ gen_helper_sve_ldff1sdu_le_r,
67
/* Take a SecureFault on the current stack */
344
68
env->v7m.sfsr |= R_V7M_SFSR_INVIS_MASK;
345
- gen_helper_sve_ldff1bds_r,
69
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
346
- gen_helper_sve_ldff1bss_r,
347
- gen_helper_sve_ldff1bhs_r,
348
- gen_helper_sve_ldff1dd_le_r },
349
+ gen_helper_sve_ldff1bds_r,
350
+ gen_helper_sve_ldff1bss_r,
351
+ gen_helper_sve_ldff1bhs_r,
352
+ gen_helper_sve_ldff1dd_le_r },
353
354
- /* Big-endian */
355
- { gen_helper_sve_ldff1bb_r,
356
- gen_helper_sve_ldff1bhu_r,
357
- gen_helper_sve_ldff1bsu_r,
358
- gen_helper_sve_ldff1bdu_r,
359
+ /* mte inactive, big-endian */
360
+ { gen_helper_sve_ldff1bb_r,
361
+ gen_helper_sve_ldff1bhu_r,
362
+ gen_helper_sve_ldff1bsu_r,
363
+ gen_helper_sve_ldff1bdu_r,
364
365
- gen_helper_sve_ldff1sds_be_r,
366
- gen_helper_sve_ldff1hh_be_r,
367
- gen_helper_sve_ldff1hsu_be_r,
368
- gen_helper_sve_ldff1hdu_be_r,
369
+ gen_helper_sve_ldff1sds_be_r,
370
+ gen_helper_sve_ldff1hh_be_r,
371
+ gen_helper_sve_ldff1hsu_be_r,
372
+ gen_helper_sve_ldff1hdu_be_r,
373
374
- gen_helper_sve_ldff1hds_be_r,
375
- gen_helper_sve_ldff1hss_be_r,
376
- gen_helper_sve_ldff1ss_be_r,
377
- gen_helper_sve_ldff1sdu_be_r,
378
+ gen_helper_sve_ldff1hds_be_r,
379
+ gen_helper_sve_ldff1hss_be_r,
380
+ gen_helper_sve_ldff1ss_be_r,
381
+ gen_helper_sve_ldff1sdu_be_r,
382
383
- gen_helper_sve_ldff1bds_r,
384
- gen_helper_sve_ldff1bss_r,
385
- gen_helper_sve_ldff1bhs_r,
386
- gen_helper_sve_ldff1dd_be_r },
387
+ gen_helper_sve_ldff1bds_r,
388
+ gen_helper_sve_ldff1bss_r,
389
+ gen_helper_sve_ldff1bhs_r,
390
+ gen_helper_sve_ldff1dd_be_r } },
391
+
392
+ { /* mte active, little-endian */
393
+ { gen_helper_sve_ldff1bb_r_mte,
394
+ gen_helper_sve_ldff1bhu_r_mte,
395
+ gen_helper_sve_ldff1bsu_r_mte,
396
+ gen_helper_sve_ldff1bdu_r_mte,
397
+
398
+ gen_helper_sve_ldff1sds_le_r_mte,
399
+ gen_helper_sve_ldff1hh_le_r_mte,
400
+ gen_helper_sve_ldff1hsu_le_r_mte,
401
+ gen_helper_sve_ldff1hdu_le_r_mte,
402
+
403
+ gen_helper_sve_ldff1hds_le_r_mte,
404
+ gen_helper_sve_ldff1hss_le_r_mte,
405
+ gen_helper_sve_ldff1ss_le_r_mte,
406
+ gen_helper_sve_ldff1sdu_le_r_mte,
407
+
408
+ gen_helper_sve_ldff1bds_r_mte,
409
+ gen_helper_sve_ldff1bss_r_mte,
410
+ gen_helper_sve_ldff1bhs_r_mte,
411
+ gen_helper_sve_ldff1dd_le_r_mte },
412
+
413
+ /* mte active, big-endian */
414
+ { gen_helper_sve_ldff1bb_r_mte,
415
+ gen_helper_sve_ldff1bhu_r_mte,
416
+ gen_helper_sve_ldff1bsu_r_mte,
417
+ gen_helper_sve_ldff1bdu_r_mte,
418
+
419
+ gen_helper_sve_ldff1sds_be_r_mte,
420
+ gen_helper_sve_ldff1hh_be_r_mte,
421
+ gen_helper_sve_ldff1hsu_be_r_mte,
422
+ gen_helper_sve_ldff1hdu_be_r_mte,
423
+
424
+ gen_helper_sve_ldff1hds_be_r_mte,
425
+ gen_helper_sve_ldff1hss_be_r_mte,
426
+ gen_helper_sve_ldff1ss_be_r_mte,
427
+ gen_helper_sve_ldff1sdu_be_r_mte,
428
+
429
+ gen_helper_sve_ldff1bds_r_mte,
430
+ gen_helper_sve_ldff1bss_r_mte,
431
+ gen_helper_sve_ldff1bhs_r_mte,
432
+ gen_helper_sve_ldff1dd_be_r_mte } },
433
};
434
435
if (sve_access_check(s)) {
436
TCGv_i64 addr = new_tmp_a64(s);
437
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
438
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
439
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
440
- fns[s->be_data == MO_BE][a->dtype]);
441
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
442
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
443
}
444
return true;
445
}
446
447
static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
448
{
449
- static gen_helper_gvec_mem * const fns[2][16] = {
450
- /* Little-endian */
451
- { gen_helper_sve_ldnf1bb_r,
452
- gen_helper_sve_ldnf1bhu_r,
453
- gen_helper_sve_ldnf1bsu_r,
454
- gen_helper_sve_ldnf1bdu_r,
455
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
456
+ { /* mte inactive, little-endian */
457
+ { gen_helper_sve_ldnf1bb_r,
458
+ gen_helper_sve_ldnf1bhu_r,
459
+ gen_helper_sve_ldnf1bsu_r,
460
+ gen_helper_sve_ldnf1bdu_r,
461
462
- gen_helper_sve_ldnf1sds_le_r,
463
- gen_helper_sve_ldnf1hh_le_r,
464
- gen_helper_sve_ldnf1hsu_le_r,
465
- gen_helper_sve_ldnf1hdu_le_r,
466
+ gen_helper_sve_ldnf1sds_le_r,
467
+ gen_helper_sve_ldnf1hh_le_r,
468
+ gen_helper_sve_ldnf1hsu_le_r,
469
+ gen_helper_sve_ldnf1hdu_le_r,
470
471
- gen_helper_sve_ldnf1hds_le_r,
472
- gen_helper_sve_ldnf1hss_le_r,
473
- gen_helper_sve_ldnf1ss_le_r,
474
- gen_helper_sve_ldnf1sdu_le_r,
475
+ gen_helper_sve_ldnf1hds_le_r,
476
+ gen_helper_sve_ldnf1hss_le_r,
477
+ gen_helper_sve_ldnf1ss_le_r,
478
+ gen_helper_sve_ldnf1sdu_le_r,
479
480
- gen_helper_sve_ldnf1bds_r,
481
- gen_helper_sve_ldnf1bss_r,
482
- gen_helper_sve_ldnf1bhs_r,
483
- gen_helper_sve_ldnf1dd_le_r },
484
+ gen_helper_sve_ldnf1bds_r,
485
+ gen_helper_sve_ldnf1bss_r,
486
+ gen_helper_sve_ldnf1bhs_r,
487
+ gen_helper_sve_ldnf1dd_le_r },
488
489
- /* Big-endian */
490
- { gen_helper_sve_ldnf1bb_r,
491
- gen_helper_sve_ldnf1bhu_r,
492
- gen_helper_sve_ldnf1bsu_r,
493
- gen_helper_sve_ldnf1bdu_r,
494
+ /* mte inactive, big-endian */
495
+ { gen_helper_sve_ldnf1bb_r,
496
+ gen_helper_sve_ldnf1bhu_r,
497
+ gen_helper_sve_ldnf1bsu_r,
498
+ gen_helper_sve_ldnf1bdu_r,
499
500
- gen_helper_sve_ldnf1sds_be_r,
501
- gen_helper_sve_ldnf1hh_be_r,
502
- gen_helper_sve_ldnf1hsu_be_r,
503
- gen_helper_sve_ldnf1hdu_be_r,
504
+ gen_helper_sve_ldnf1sds_be_r,
505
+ gen_helper_sve_ldnf1hh_be_r,
506
+ gen_helper_sve_ldnf1hsu_be_r,
507
+ gen_helper_sve_ldnf1hdu_be_r,
508
509
- gen_helper_sve_ldnf1hds_be_r,
510
- gen_helper_sve_ldnf1hss_be_r,
511
- gen_helper_sve_ldnf1ss_be_r,
512
- gen_helper_sve_ldnf1sdu_be_r,
513
+ gen_helper_sve_ldnf1hds_be_r,
514
+ gen_helper_sve_ldnf1hss_be_r,
515
+ gen_helper_sve_ldnf1ss_be_r,
516
+ gen_helper_sve_ldnf1sdu_be_r,
517
518
- gen_helper_sve_ldnf1bds_r,
519
- gen_helper_sve_ldnf1bss_r,
520
- gen_helper_sve_ldnf1bhs_r,
521
- gen_helper_sve_ldnf1dd_be_r },
522
+ gen_helper_sve_ldnf1bds_r,
523
+ gen_helper_sve_ldnf1bss_r,
524
+ gen_helper_sve_ldnf1bhs_r,
525
+ gen_helper_sve_ldnf1dd_be_r } },
526
+
527
+ { /* mte inactive, little-endian */
528
+ { gen_helper_sve_ldnf1bb_r_mte,
529
+ gen_helper_sve_ldnf1bhu_r_mte,
530
+ gen_helper_sve_ldnf1bsu_r_mte,
531
+ gen_helper_sve_ldnf1bdu_r_mte,
532
+
533
+ gen_helper_sve_ldnf1sds_le_r_mte,
534
+ gen_helper_sve_ldnf1hh_le_r_mte,
535
+ gen_helper_sve_ldnf1hsu_le_r_mte,
536
+ gen_helper_sve_ldnf1hdu_le_r_mte,
537
+
538
+ gen_helper_sve_ldnf1hds_le_r_mte,
539
+ gen_helper_sve_ldnf1hss_le_r_mte,
540
+ gen_helper_sve_ldnf1ss_le_r_mte,
541
+ gen_helper_sve_ldnf1sdu_le_r_mte,
542
+
543
+ gen_helper_sve_ldnf1bds_r_mte,
544
+ gen_helper_sve_ldnf1bss_r_mte,
545
+ gen_helper_sve_ldnf1bhs_r_mte,
546
+ gen_helper_sve_ldnf1dd_le_r_mte },
547
+
548
+ /* mte inactive, big-endian */
549
+ { gen_helper_sve_ldnf1bb_r_mte,
550
+ gen_helper_sve_ldnf1bhu_r_mte,
551
+ gen_helper_sve_ldnf1bsu_r_mte,
552
+ gen_helper_sve_ldnf1bdu_r_mte,
553
+
554
+ gen_helper_sve_ldnf1sds_be_r_mte,
555
+ gen_helper_sve_ldnf1hh_be_r_mte,
556
+ gen_helper_sve_ldnf1hsu_be_r_mte,
557
+ gen_helper_sve_ldnf1hdu_be_r_mte,
558
+
559
+ gen_helper_sve_ldnf1hds_be_r_mte,
560
+ gen_helper_sve_ldnf1hss_be_r_mte,
561
+ gen_helper_sve_ldnf1ss_be_r_mte,
562
+ gen_helper_sve_ldnf1sdu_be_r_mte,
563
+
564
+ gen_helper_sve_ldnf1bds_r_mte,
565
+ gen_helper_sve_ldnf1bss_r_mte,
566
+ gen_helper_sve_ldnf1bhs_r_mte,
567
+ gen_helper_sve_ldnf1dd_be_r_mte } },
568
};
569
570
if (sve_access_check(s)) {
571
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
572
TCGv_i64 addr = new_tmp_a64(s);
573
574
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
575
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
576
- fns[s->be_data == MO_BE][a->dtype]);
577
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
578
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
579
}
580
return true;
581
}
70
--
582
--
71
2.20.1
583
2.20.1
72
584
73
585
diff view generated by jsdifflib
1
For M-profile the MVFR* ID registers are memory mapped, in the
1
From: Richard Henderson <richard.henderson@linaro.org>
2
range we implement via the NVIC. Allow them to be read.
3
(If the CPU has no FPU, these registers are defined to be RAZ.)
4
2
3
We still need to handle tbi for user-only when mte is inactive.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-37-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20190416125744.27770-3-peter.maydell@linaro.org
8
---
9
---
9
hw/intc/armv7m_nvic.c | 6 ++++++
10
target/arm/translate-a64.h | 1 +
10
1 file changed, 6 insertions(+)
11
target/arm/translate-a64.c | 2 +-
12
target/arm/translate-sve.c | 6 ++++--
13
3 files changed, 6 insertions(+), 3 deletions(-)
11
14
12
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
13
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/intc/armv7m_nvic.c
17
--- a/target/arm/translate-a64.h
15
+++ b/hw/intc/armv7m_nvic.c
18
+++ b/target/arm/translate-a64.h
16
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
19
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
17
return 0;
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
18
}
21
unsigned int imms, unsigned int immr);
19
return cpu->env.v7m.sfar;
22
bool sve_access_check(DisasContext *s);
20
+ case 0xf40: /* MVFR0 */
23
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
21
+ return cpu->isar.mvfr0;
24
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
22
+ case 0xf44: /* MVFR1 */
25
bool tag_checked, int log2_size);
23
+ return cpu->isar.mvfr1;
26
TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
24
+ case 0xf48: /* MVFR2 */
27
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
25
+ return cpu->isar.mvfr2;
28
index XXXXXXX..XXXXXXX 100644
26
default:
29
--- a/target/arm/translate-a64.c
27
bad_offset:
30
+++ b/target/arm/translate-a64.c
28
qemu_log_mask(LOG_GUEST_ERROR, "NVIC: Bad read offset 0x%x\n", offset);
31
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
32
* of the write-back address.
33
*/
34
35
-static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
36
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
37
{
38
TCGv_i64 clean = new_tmp_a64(s);
39
#ifdef CONFIG_USER_ONLY
40
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-sve.c
43
+++ b/target/arm/translate-sve.c
44
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
45
* For e.g. LD4, there are not enough arguments to pass all 4
46
* registers as pointers, so encode the regno into the data field.
47
* For consistency, do this even for LD1.
48
- * TODO: mte_n check here while callers are updated.
49
*/
50
- if (mte_n && s->mte_active[0]) {
51
+ if (s->mte_active[0]) {
52
int msz = dtype_msz(dtype);
53
54
desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
55
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
56
desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
57
desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
58
desc <<= SVE_MTEDESC_SHIFT;
59
+ } else {
60
+ addr = clean_data_tbi(s, addr);
61
}
62
+
63
desc = simd_desc(vsz, vsz, zt | desc);
64
t_desc = tcg_const_i32(desc);
65
t_pg = tcg_temp_new_ptr();
29
--
66
--
30
2.20.1
67
2.20.1
31
68
32
69
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
No code used the tc6393xb_gpio_in_get() and tc6393xb_gpio_out_set()
3
Because the elements are non-sequential, we cannot eliminate many
4
functions since their introduction in commit 88d2c950b002. Time to
4
tests straight away like we can for sequential operations. But
5
remove them.
5
we often have the PTE details handy, so we can test for Tagged.
6
6
7
Suggested-by: Markus Armbruster <armbru@redhat.com>
8
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
9
Message-id: 20190412165416.7977-4-philmd@redhat.com
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-38-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
11
---
13
include/hw/devices.h | 3 ---
12
target/arm/helper-sve.h | 285 ++++++++++++++++
14
hw/display/tc6393xb.c | 16 ----------------
13
target/arm/sve_helper.c | 185 +++++++++--
15
2 files changed, 19 deletions(-)
14
target/arm/translate-sve.c | 650 +++++++++++++++++++++++++------------
15
3 files changed, 872 insertions(+), 248 deletions(-)
16
16
17
diff --git a/include/hw/devices.h b/include/hw/devices.h
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
18
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/devices.h
19
--- a/target/arm/helper-sve.h
20
+++ b/include/hw/devices.h
20
+++ b/target/arm/helper-sve.h
21
@@ -XXX,XX +XXX,XX @@ void retu_key_event(void *retu, int state);
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldsds_le_zd, TCG_CALL_NO_WG,
22
typedef struct TC6393xbState TC6393xbState;
22
DEF_HELPER_FLAGS_6(sve_ldsds_be_zd, TCG_CALL_NO_WG,
23
TC6393xbState *tc6393xb_init(struct MemoryRegion *sysmem,
23
void, env, ptr, ptr, ptr, tl, i32)
24
uint32_t base, qemu_irq irq);
24
25
-void tc6393xb_gpio_out_set(TC6393xbState *s, int line,
25
+DEF_HELPER_FLAGS_6(sve_ldbsu_zsu_mte, TCG_CALL_NO_WG,
26
- qemu_irq handler);
26
+ void, env, ptr, ptr, ptr, tl, i32)
27
-qemu_irq *tc6393xb_gpio_in_get(TC6393xbState *s);
27
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu_mte, TCG_CALL_NO_WG,
28
qemu_irq tc6393xb_l3v_get(TC6393xbState *s);
28
+ void, env, ptr, ptr, ptr, tl, i32)
29
29
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zsu_mte, TCG_CALL_NO_WG,
30
#endif
30
+ void, env, ptr, ptr, ptr, tl, i32)
31
diff --git a/hw/display/tc6393xb.c b/hw/display/tc6393xb.c
31
+DEF_HELPER_FLAGS_6(sve_ldss_le_zsu_mte, TCG_CALL_NO_WG,
32
+ void, env, ptr, ptr, ptr, tl, i32)
33
+DEF_HELPER_FLAGS_6(sve_ldss_be_zsu_mte, TCG_CALL_NO_WG,
34
+ void, env, ptr, ptr, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_6(sve_ldbss_zsu_mte, TCG_CALL_NO_WG,
36
+ void, env, ptr, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zsu_mte, TCG_CALL_NO_WG,
38
+ void, env, ptr, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zsu_mte, TCG_CALL_NO_WG,
40
+ void, env, ptr, ptr, ptr, tl, i32)
41
+
42
+DEF_HELPER_FLAGS_6(sve_ldbsu_zss_mte, TCG_CALL_NO_WG,
43
+ void, env, ptr, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zss_mte, TCG_CALL_NO_WG,
45
+ void, env, ptr, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zss_mte, TCG_CALL_NO_WG,
47
+ void, env, ptr, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_6(sve_ldss_le_zss_mte, TCG_CALL_NO_WG,
49
+ void, env, ptr, ptr, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_6(sve_ldss_be_zss_mte, TCG_CALL_NO_WG,
51
+ void, env, ptr, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_6(sve_ldbss_zss_mte, TCG_CALL_NO_WG,
53
+ void, env, ptr, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zss_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zss_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_6(sve_ldbdu_zsu_mte, TCG_CALL_NO_WG,
60
+ void, env, ptr, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zsu_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zsu_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zsu_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, ptr, ptr, tl, i32)
67
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zsu_mte, TCG_CALL_NO_WG,
68
+ void, env, ptr, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_6(sve_lddd_le_zsu_mte, TCG_CALL_NO_WG,
70
+ void, env, ptr, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_6(sve_lddd_be_zsu_mte, TCG_CALL_NO_WG,
72
+ void, env, ptr, ptr, ptr, tl, i32)
73
+DEF_HELPER_FLAGS_6(sve_ldbds_zsu_mte, TCG_CALL_NO_WG,
74
+ void, env, ptr, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zsu_mte, TCG_CALL_NO_WG,
76
+ void, env, ptr, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zsu_mte, TCG_CALL_NO_WG,
78
+ void, env, ptr, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zsu_mte, TCG_CALL_NO_WG,
80
+ void, env, ptr, ptr, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zsu_mte, TCG_CALL_NO_WG,
82
+ void, env, ptr, ptr, ptr, tl, i32)
83
+
84
+DEF_HELPER_FLAGS_6(sve_ldbdu_zss_mte, TCG_CALL_NO_WG,
85
+ void, env, ptr, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zss_mte, TCG_CALL_NO_WG,
87
+ void, env, ptr, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zss_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, ptr, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zss_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zss_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_6(sve_lddd_le_zss_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_6(sve_lddd_be_zss_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_6(sve_ldbds_zss_mte, TCG_CALL_NO_WG,
99
+ void, env, ptr, ptr, ptr, tl, i32)
100
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zss_mte, TCG_CALL_NO_WG,
101
+ void, env, ptr, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zss_mte, TCG_CALL_NO_WG,
103
+ void, env, ptr, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zss_mte, TCG_CALL_NO_WG,
105
+ void, env, ptr, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zss_mte, TCG_CALL_NO_WG,
107
+ void, env, ptr, ptr, ptr, tl, i32)
108
+
109
+DEF_HELPER_FLAGS_6(sve_ldbdu_zd_mte, TCG_CALL_NO_WG,
110
+ void, env, ptr, ptr, ptr, tl, i32)
111
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zd_mte, TCG_CALL_NO_WG,
112
+ void, env, ptr, ptr, ptr, tl, i32)
113
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zd_mte, TCG_CALL_NO_WG,
114
+ void, env, ptr, ptr, ptr, tl, i32)
115
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zd_mte, TCG_CALL_NO_WG,
116
+ void, env, ptr, ptr, ptr, tl, i32)
117
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zd_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, ptr, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_6(sve_lddd_le_zd_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, ptr, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_6(sve_lddd_be_zd_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, ptr, ptr, tl, i32)
123
+DEF_HELPER_FLAGS_6(sve_ldbds_zd_mte, TCG_CALL_NO_WG,
124
+ void, env, ptr, ptr, ptr, tl, i32)
125
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zd_mte, TCG_CALL_NO_WG,
126
+ void, env, ptr, ptr, ptr, tl, i32)
127
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zd_mte, TCG_CALL_NO_WG,
128
+ void, env, ptr, ptr, ptr, tl, i32)
129
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zd_mte, TCG_CALL_NO_WG,
130
+ void, env, ptr, ptr, ptr, tl, i32)
131
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zd_mte, TCG_CALL_NO_WG,
132
+ void, env, ptr, ptr, ptr, tl, i32)
133
+
134
DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu, TCG_CALL_NO_WG,
135
void, env, ptr, ptr, ptr, tl, i32)
136
DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu, TCG_CALL_NO_WG,
137
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd, TCG_CALL_NO_WG,
138
DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd, TCG_CALL_NO_WG,
139
void, env, ptr, ptr, ptr, tl, i32)
140
141
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu_mte, TCG_CALL_NO_WG,
142
+ void, env, ptr, ptr, ptr, tl, i32)
143
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu_mte, TCG_CALL_NO_WG,
144
+ void, env, ptr, ptr, ptr, tl, i32)
145
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zsu_mte, TCG_CALL_NO_WG,
146
+ void, env, ptr, ptr, ptr, tl, i32)
147
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zsu_mte, TCG_CALL_NO_WG,
148
+ void, env, ptr, ptr, ptr, tl, i32)
149
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zsu_mte, TCG_CALL_NO_WG,
150
+ void, env, ptr, ptr, ptr, tl, i32)
151
+DEF_HELPER_FLAGS_6(sve_ldffbss_zsu_mte, TCG_CALL_NO_WG,
152
+ void, env, ptr, ptr, ptr, tl, i32)
153
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zsu_mte, TCG_CALL_NO_WG,
154
+ void, env, ptr, ptr, ptr, tl, i32)
155
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zsu_mte, TCG_CALL_NO_WG,
156
+ void, env, ptr, ptr, ptr, tl, i32)
157
+
158
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zss_mte, TCG_CALL_NO_WG,
159
+ void, env, ptr, ptr, ptr, tl, i32)
160
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zss_mte, TCG_CALL_NO_WG,
161
+ void, env, ptr, ptr, ptr, tl, i32)
162
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zss_mte, TCG_CALL_NO_WG,
163
+ void, env, ptr, ptr, ptr, tl, i32)
164
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zss_mte, TCG_CALL_NO_WG,
165
+ void, env, ptr, ptr, ptr, tl, i32)
166
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zss_mte, TCG_CALL_NO_WG,
167
+ void, env, ptr, ptr, ptr, tl, i32)
168
+DEF_HELPER_FLAGS_6(sve_ldffbss_zss_mte, TCG_CALL_NO_WG,
169
+ void, env, ptr, ptr, ptr, tl, i32)
170
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zss_mte, TCG_CALL_NO_WG,
171
+ void, env, ptr, ptr, ptr, tl, i32)
172
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zss_mte, TCG_CALL_NO_WG,
173
+ void, env, ptr, ptr, ptr, tl, i32)
174
+
175
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zsu_mte, TCG_CALL_NO_WG,
176
+ void, env, ptr, ptr, ptr, tl, i32)
177
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zsu_mte, TCG_CALL_NO_WG,
178
+ void, env, ptr, ptr, ptr, tl, i32)
179
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zsu_mte, TCG_CALL_NO_WG,
180
+ void, env, ptr, ptr, ptr, tl, i32)
181
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zsu_mte, TCG_CALL_NO_WG,
182
+ void, env, ptr, ptr, ptr, tl, i32)
183
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zsu_mte, TCG_CALL_NO_WG,
184
+ void, env, ptr, ptr, ptr, tl, i32)
185
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zsu_mte, TCG_CALL_NO_WG,
186
+ void, env, ptr, ptr, ptr, tl, i32)
187
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zsu_mte, TCG_CALL_NO_WG,
188
+ void, env, ptr, ptr, ptr, tl, i32)
189
+DEF_HELPER_FLAGS_6(sve_ldffbds_zsu_mte, TCG_CALL_NO_WG,
190
+ void, env, ptr, ptr, ptr, tl, i32)
191
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zsu_mte, TCG_CALL_NO_WG,
192
+ void, env, ptr, ptr, ptr, tl, i32)
193
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zsu_mte, TCG_CALL_NO_WG,
194
+ void, env, ptr, ptr, ptr, tl, i32)
195
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zsu_mte, TCG_CALL_NO_WG,
196
+ void, env, ptr, ptr, ptr, tl, i32)
197
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zsu_mte, TCG_CALL_NO_WG,
198
+ void, env, ptr, ptr, ptr, tl, i32)
199
+
200
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zss_mte, TCG_CALL_NO_WG,
201
+ void, env, ptr, ptr, ptr, tl, i32)
202
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zss_mte, TCG_CALL_NO_WG,
203
+ void, env, ptr, ptr, ptr, tl, i32)
204
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zss_mte, TCG_CALL_NO_WG,
205
+ void, env, ptr, ptr, ptr, tl, i32)
206
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zss_mte, TCG_CALL_NO_WG,
207
+ void, env, ptr, ptr, ptr, tl, i32)
208
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zss_mte, TCG_CALL_NO_WG,
209
+ void, env, ptr, ptr, ptr, tl, i32)
210
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zss_mte, TCG_CALL_NO_WG,
211
+ void, env, ptr, ptr, ptr, tl, i32)
212
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zss_mte, TCG_CALL_NO_WG,
213
+ void, env, ptr, ptr, ptr, tl, i32)
214
+DEF_HELPER_FLAGS_6(sve_ldffbds_zss_mte, TCG_CALL_NO_WG,
215
+ void, env, ptr, ptr, ptr, tl, i32)
216
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zss_mte, TCG_CALL_NO_WG,
217
+ void, env, ptr, ptr, ptr, tl, i32)
218
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zss_mte, TCG_CALL_NO_WG,
219
+ void, env, ptr, ptr, ptr, tl, i32)
220
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zss_mte, TCG_CALL_NO_WG,
221
+ void, env, ptr, ptr, ptr, tl, i32)
222
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zss_mte, TCG_CALL_NO_WG,
223
+ void, env, ptr, ptr, ptr, tl, i32)
224
+
225
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zd_mte, TCG_CALL_NO_WG,
226
+ void, env, ptr, ptr, ptr, tl, i32)
227
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zd_mte, TCG_CALL_NO_WG,
228
+ void, env, ptr, ptr, ptr, tl, i32)
229
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zd_mte, TCG_CALL_NO_WG,
230
+ void, env, ptr, ptr, ptr, tl, i32)
231
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zd_mte, TCG_CALL_NO_WG,
232
+ void, env, ptr, ptr, ptr, tl, i32)
233
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zd_mte, TCG_CALL_NO_WG,
234
+ void, env, ptr, ptr, ptr, tl, i32)
235
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zd_mte, TCG_CALL_NO_WG,
236
+ void, env, ptr, ptr, ptr, tl, i32)
237
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zd_mte, TCG_CALL_NO_WG,
238
+ void, env, ptr, ptr, ptr, tl, i32)
239
+DEF_HELPER_FLAGS_6(sve_ldffbds_zd_mte, TCG_CALL_NO_WG,
240
+ void, env, ptr, ptr, ptr, tl, i32)
241
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zd_mte, TCG_CALL_NO_WG,
242
+ void, env, ptr, ptr, ptr, tl, i32)
243
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zd_mte, TCG_CALL_NO_WG,
244
+ void, env, ptr, ptr, ptr, tl, i32)
245
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd_mte, TCG_CALL_NO_WG,
246
+ void, env, ptr, ptr, ptr, tl, i32)
247
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd_mte, TCG_CALL_NO_WG,
248
+ void, env, ptr, ptr, ptr, tl, i32)
249
+
250
DEF_HELPER_FLAGS_6(sve_stbs_zsu, TCG_CALL_NO_WG,
251
void, env, ptr, ptr, ptr, tl, i32)
252
DEF_HELPER_FLAGS_6(sve_sths_le_zsu, TCG_CALL_NO_WG,
253
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd, TCG_CALL_NO_WG,
254
DEF_HELPER_FLAGS_6(sve_stdd_be_zd, TCG_CALL_NO_WG,
255
void, env, ptr, ptr, ptr, tl, i32)
256
257
+DEF_HELPER_FLAGS_6(sve_stbs_zsu_mte, TCG_CALL_NO_WG,
258
+ void, env, ptr, ptr, ptr, tl, i32)
259
+DEF_HELPER_FLAGS_6(sve_sths_le_zsu_mte, TCG_CALL_NO_WG,
260
+ void, env, ptr, ptr, ptr, tl, i32)
261
+DEF_HELPER_FLAGS_6(sve_sths_be_zsu_mte, TCG_CALL_NO_WG,
262
+ void, env, ptr, ptr, ptr, tl, i32)
263
+DEF_HELPER_FLAGS_6(sve_stss_le_zsu_mte, TCG_CALL_NO_WG,
264
+ void, env, ptr, ptr, ptr, tl, i32)
265
+DEF_HELPER_FLAGS_6(sve_stss_be_zsu_mte, TCG_CALL_NO_WG,
266
+ void, env, ptr, ptr, ptr, tl, i32)
267
+
268
+DEF_HELPER_FLAGS_6(sve_stbs_zss_mte, TCG_CALL_NO_WG,
269
+ void, env, ptr, ptr, ptr, tl, i32)
270
+DEF_HELPER_FLAGS_6(sve_sths_le_zss_mte, TCG_CALL_NO_WG,
271
+ void, env, ptr, ptr, ptr, tl, i32)
272
+DEF_HELPER_FLAGS_6(sve_sths_be_zss_mte, TCG_CALL_NO_WG,
273
+ void, env, ptr, ptr, ptr, tl, i32)
274
+DEF_HELPER_FLAGS_6(sve_stss_le_zss_mte, TCG_CALL_NO_WG,
275
+ void, env, ptr, ptr, ptr, tl, i32)
276
+DEF_HELPER_FLAGS_6(sve_stss_be_zss_mte, TCG_CALL_NO_WG,
277
+ void, env, ptr, ptr, ptr, tl, i32)
278
+
279
+DEF_HELPER_FLAGS_6(sve_stbd_zsu_mte, TCG_CALL_NO_WG,
280
+ void, env, ptr, ptr, ptr, tl, i32)
281
+DEF_HELPER_FLAGS_6(sve_sthd_le_zsu_mte, TCG_CALL_NO_WG,
282
+ void, env, ptr, ptr, ptr, tl, i32)
283
+DEF_HELPER_FLAGS_6(sve_sthd_be_zsu_mte, TCG_CALL_NO_WG,
284
+ void, env, ptr, ptr, ptr, tl, i32)
285
+DEF_HELPER_FLAGS_6(sve_stsd_le_zsu_mte, TCG_CALL_NO_WG,
286
+ void, env, ptr, ptr, ptr, tl, i32)
287
+DEF_HELPER_FLAGS_6(sve_stsd_be_zsu_mte, TCG_CALL_NO_WG,
288
+ void, env, ptr, ptr, ptr, tl, i32)
289
+DEF_HELPER_FLAGS_6(sve_stdd_le_zsu_mte, TCG_CALL_NO_WG,
290
+ void, env, ptr, ptr, ptr, tl, i32)
291
+DEF_HELPER_FLAGS_6(sve_stdd_be_zsu_mte, TCG_CALL_NO_WG,
292
+ void, env, ptr, ptr, ptr, tl, i32)
293
+
294
+DEF_HELPER_FLAGS_6(sve_stbd_zss_mte, TCG_CALL_NO_WG,
295
+ void, env, ptr, ptr, ptr, tl, i32)
296
+DEF_HELPER_FLAGS_6(sve_sthd_le_zss_mte, TCG_CALL_NO_WG,
297
+ void, env, ptr, ptr, ptr, tl, i32)
298
+DEF_HELPER_FLAGS_6(sve_sthd_be_zss_mte, TCG_CALL_NO_WG,
299
+ void, env, ptr, ptr, ptr, tl, i32)
300
+DEF_HELPER_FLAGS_6(sve_stsd_le_zss_mte, TCG_CALL_NO_WG,
301
+ void, env, ptr, ptr, ptr, tl, i32)
302
+DEF_HELPER_FLAGS_6(sve_stsd_be_zss_mte, TCG_CALL_NO_WG,
303
+ void, env, ptr, ptr, ptr, tl, i32)
304
+DEF_HELPER_FLAGS_6(sve_stdd_le_zss_mte, TCG_CALL_NO_WG,
305
+ void, env, ptr, ptr, ptr, tl, i32)
306
+DEF_HELPER_FLAGS_6(sve_stdd_be_zss_mte, TCG_CALL_NO_WG,
307
+ void, env, ptr, ptr, ptr, tl, i32)
308
+
309
+DEF_HELPER_FLAGS_6(sve_stbd_zd_mte, TCG_CALL_NO_WG,
310
+ void, env, ptr, ptr, ptr, tl, i32)
311
+DEF_HELPER_FLAGS_6(sve_sthd_le_zd_mte, TCG_CALL_NO_WG,
312
+ void, env, ptr, ptr, ptr, tl, i32)
313
+DEF_HELPER_FLAGS_6(sve_sthd_be_zd_mte, TCG_CALL_NO_WG,
314
+ void, env, ptr, ptr, ptr, tl, i32)
315
+DEF_HELPER_FLAGS_6(sve_stsd_le_zd_mte, TCG_CALL_NO_WG,
316
+ void, env, ptr, ptr, ptr, tl, i32)
317
+DEF_HELPER_FLAGS_6(sve_stsd_be_zd_mte, TCG_CALL_NO_WG,
318
+ void, env, ptr, ptr, ptr, tl, i32)
319
+DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG,
320
+ void, env, ptr, ptr, ptr, tl, i32)
321
+DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG,
322
+ void, env, ptr, ptr, ptr, tl, i32)
323
+
324
DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
325
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
32
index XXXXXXX..XXXXXXX 100644
326
index XXXXXXX..XXXXXXX 100644
33
--- a/hw/display/tc6393xb.c
327
--- a/target/arm/sve_helper.c
34
+++ b/hw/display/tc6393xb.c
328
+++ b/target/arm/sve_helper.c
35
@@ -XXX,XX +XXX,XX @@ struct TC6393xbState {
329
@@ -XXX,XX +XXX,XX @@ static target_ulong off_zd_d(void *reg, intptr_t reg_ofs)
36
blanked : 1;
330
static inline QEMU_ALWAYS_INLINE
331
void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
332
target_ulong base, uint32_t desc, uintptr_t retaddr,
333
- int esize, int msize, zreg_off_fn *off_fn,
334
+ uint32_t mtedesc, int esize, int msize,
335
+ zreg_off_fn *off_fn,
336
sve_ldst1_host_fn *host_fn,
337
sve_ldst1_tlb_fn *tlb_fn)
338
{
339
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
340
cpu_check_watchpoint(env_cpu(env), addr, msize,
341
info.attrs, BP_MEM_READ, retaddr);
342
}
343
- /* TODO: MTE check */
344
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
345
+ mte_check1(env, mtedesc, addr, retaddr);
346
+ }
347
host_fn(&scratch, reg_off, info.host);
348
} else {
349
/* Element crosses the page boundary. */
350
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
351
msize, info.attrs,
352
BP_MEM_READ, retaddr);
353
}
354
- /* TODO: MTE check */
355
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
356
+ mte_check1(env, mtedesc, addr, retaddr);
357
+ }
358
tlb_fn(env, &scratch, reg_off, addr, retaddr);
359
}
360
}
361
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
362
memcpy(vd, &scratch, reg_max);
363
}
364
365
+static inline QEMU_ALWAYS_INLINE
366
+void sve_ld1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
367
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
368
+ int esize, int msize, zreg_off_fn *off_fn,
369
+ sve_ldst1_host_fn *host_fn,
370
+ sve_ldst1_tlb_fn *tlb_fn)
371
+{
372
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
373
+ /* Remove mtedesc from the normal sve descriptor. */
374
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
375
+
376
+ /*
377
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
378
+ * offset base entirely over the address space hole to change the
379
+ * pointer tag, or change the bit55 selector. So we could here
380
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
381
+ */
382
+ sve_ld1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
383
+ esize, msize, off_fn, host_fn, tlb_fn);
384
+}
385
+
386
#define DO_LD1_ZPZ_S(MEM, OFS, MSZ) \
387
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
388
void *vm, target_ulong base, uint32_t desc) \
389
{ \
390
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
391
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
392
off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
393
+} \
394
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
395
+ void *vm, target_ulong base, uint32_t desc) \
396
+{ \
397
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
398
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
399
}
400
401
#define DO_LD1_ZPZ_D(MEM, OFS, MSZ) \
402
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
403
void *vm, target_ulong base, uint32_t desc) \
404
{ \
405
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
406
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
407
off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
408
+} \
409
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
410
+ void *vm, target_ulong base, uint32_t desc) \
411
+{ \
412
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
413
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
414
}
415
416
DO_LD1_ZPZ_S(bsu, zsu, MO_8)
417
@@ -XXX,XX +XXX,XX @@ DO_LD1_ZPZ_D(dd_be, zd, MO_64)
418
static inline QEMU_ALWAYS_INLINE
419
void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
420
target_ulong base, uint32_t desc, uintptr_t retaddr,
421
- const int esz, const int msz, zreg_off_fn *off_fn,
422
+ uint32_t mtedesc, const int esz, const int msz,
423
+ zreg_off_fn *off_fn,
424
sve_ldst1_host_fn *host_fn,
425
sve_ldst1_tlb_fn *tlb_fn)
426
{
427
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
428
* Probe the first element, allowing faults.
429
*/
430
addr = base + (off_fn(vm, reg_off) << scale);
431
+ if (mtedesc) {
432
+ mte_check1(env, mtedesc, addr, retaddr);
433
+ }
434
tlb_fn(env, vd, reg_off, addr, retaddr);
435
436
/* After any fault, zero the other elements. */
437
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
438
(env_cpu(env), addr, msize) & BP_MEM_READ)) {
439
goto fault;
440
}
441
- /* TODO: MTE check. */
442
+ if (mtedesc &&
443
+ arm_tlb_mte_tagged(&info.attrs) &&
444
+ !mte_probe1(env, mtedesc, addr)) {
445
+ goto fault;
446
+ }
447
448
host_fn(vd, reg_off, info.host);
449
}
450
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
451
record_fault(env, reg_off, reg_max);
452
}
453
454
-#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
455
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
456
- void *vm, target_ulong base, uint32_t desc) \
457
-{ \
458
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
459
- off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
460
+static inline QEMU_ALWAYS_INLINE
461
+void sve_ldff1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
462
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
463
+ const int esz, const int msz,
464
+ zreg_off_fn *off_fn,
465
+ sve_ldst1_host_fn *host_fn,
466
+ sve_ldst1_tlb_fn *tlb_fn)
467
+{
468
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
469
+ /* Remove mtedesc from the normal sve descriptor. */
470
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
471
+
472
+ /*
473
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
474
+ * offset base entirely over the address space hole to change the
475
+ * pointer tag, or change the bit55 selector. So we could here
476
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
477
+ */
478
+ sve_ldff1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
479
+ esz, msz, off_fn, host_fn, tlb_fn);
480
}
481
482
-#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
483
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
484
- void *vm, target_ulong base, uint32_t desc) \
485
-{ \
486
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
487
- off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
488
+#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
489
+void HELPER(sve_ldff##MEM##_##OFS) \
490
+ (CPUARMState *env, void *vd, void *vg, \
491
+ void *vm, target_ulong base, uint32_t desc) \
492
+{ \
493
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_32, MSZ, \
494
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
495
+} \
496
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
497
+ (CPUARMState *env, void *vd, void *vg, \
498
+ void *vm, target_ulong base, uint32_t desc) \
499
+{ \
500
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
501
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
502
+}
503
+
504
+#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
505
+void HELPER(sve_ldff##MEM##_##OFS) \
506
+ (CPUARMState *env, void *vd, void *vg, \
507
+ void *vm, target_ulong base, uint32_t desc) \
508
+{ \
509
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_64, MSZ, \
510
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
511
+} \
512
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
513
+ (CPUARMState *env, void *vd, void *vg, \
514
+ void *vm, target_ulong base, uint32_t desc) \
515
+{ \
516
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
517
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
518
}
519
520
DO_LDFF1_ZPZ_S(bsu, zsu, MO_8)
521
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_ZPZ_D(dd_be, zd, MO_64)
522
static inline QEMU_ALWAYS_INLINE
523
void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
524
target_ulong base, uint32_t desc, uintptr_t retaddr,
525
- int esize, int msize, zreg_off_fn *off_fn,
526
+ uint32_t mtedesc, int esize, int msize,
527
+ zreg_off_fn *off_fn,
528
sve_ldst1_host_fn *host_fn,
529
sve_ldst1_tlb_fn *tlb_fn)
530
{
531
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
532
cpu_check_watchpoint(env_cpu(env), addr, msize,
533
info.attrs, BP_MEM_WRITE, retaddr);
534
}
535
- /* TODO: MTE check. */
536
+
537
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
538
+ mte_check1(env, mtedesc, addr, retaddr);
539
+ }
540
}
541
i += 1;
542
reg_off += esize;
543
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
544
} while (reg_off < reg_max);
545
}
546
547
-#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
548
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
549
- void *vm, target_ulong base, uint32_t desc) \
550
-{ \
551
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
552
- off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
553
+static inline QEMU_ALWAYS_INLINE
554
+void sve_st1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
555
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
556
+ int esize, int msize, zreg_off_fn *off_fn,
557
+ sve_ldst1_host_fn *host_fn,
558
+ sve_ldst1_tlb_fn *tlb_fn)
559
+{
560
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
561
+ /* Remove mtedesc from the normal sve descriptor. */
562
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
563
+
564
+ /*
565
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
566
+ * offset base entirely over the address space hole to change the
567
+ * pointer tag, or change the bit55 selector. So we could here
568
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
569
+ */
570
+ sve_st1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
571
+ esize, msize, off_fn, host_fn, tlb_fn);
572
}
573
574
-#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
575
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
576
+#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
577
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
578
void *vm, target_ulong base, uint32_t desc) \
579
-{ \
580
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
581
- off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
582
+{ \
583
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
584
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
585
+} \
586
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
587
+ void *vm, target_ulong base, uint32_t desc) \
588
+{ \
589
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
590
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
591
+}
592
+
593
+#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
594
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
595
+ void *vm, target_ulong base, uint32_t desc) \
596
+{ \
597
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
598
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
599
+} \
600
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
601
+ void *vm, target_ulong base, uint32_t desc) \
602
+{ \
603
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
604
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
605
}
606
607
DO_ST1_ZPZ_S(bs, zsu, MO_8)
608
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
609
index XXXXXXX..XXXXXXX 100644
610
--- a/target/arm/translate-sve.c
611
+++ b/target/arm/translate-sve.c
612
@@ -XXX,XX +XXX,XX @@ static bool trans_ST_zpri(DisasContext *s, arg_rpri_store *a)
613
*/
614
615
static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
616
- int scale, TCGv_i64 scalar, int msz,
617
+ int scale, TCGv_i64 scalar, int msz, bool is_write,
618
gen_helper_gvec_mem_scatter *fn)
619
{
620
unsigned vsz = vec_full_reg_size(s);
621
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
622
TCGv_ptr t_pg = tcg_temp_new_ptr();
623
TCGv_ptr t_zt = tcg_temp_new_ptr();
624
TCGv_i32 t_desc;
625
- int desc;
626
+ int desc = 0;
627
628
+ if (s->mte_active[0]) {
629
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
630
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
631
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
632
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
633
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
634
+ desc <<= SVE_MTEDESC_SHIFT;
635
+ }
636
desc = simd_desc(vsz, vsz, scale);
637
t_desc = tcg_const_i32(desc);
638
639
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
640
tcg_temp_free_i32(t_desc);
641
}
642
643
-/* Indexed by [be][ff][xs][u][msz]. */
644
-static gen_helper_gvec_mem_scatter * const gather_load_fn32[2][2][2][2][3] = {
645
- /* Little-endian */
646
- { { { { gen_helper_sve_ldbss_zsu,
647
- gen_helper_sve_ldhss_le_zsu,
648
- NULL, },
649
- { gen_helper_sve_ldbsu_zsu,
650
- gen_helper_sve_ldhsu_le_zsu,
651
- gen_helper_sve_ldss_le_zsu, } },
652
- { { gen_helper_sve_ldbss_zss,
653
- gen_helper_sve_ldhss_le_zss,
654
- NULL, },
655
- { gen_helper_sve_ldbsu_zss,
656
- gen_helper_sve_ldhsu_le_zss,
657
- gen_helper_sve_ldss_le_zss, } } },
658
+/* Indexed by [mte][be][ff][xs][u][msz]. */
659
+static gen_helper_gvec_mem_scatter * const
660
+gather_load_fn32[2][2][2][2][2][3] = {
661
+ { /* MTE Inactive */
662
+ { /* Little-endian */
663
+ { { { gen_helper_sve_ldbss_zsu,
664
+ gen_helper_sve_ldhss_le_zsu,
665
+ NULL, },
666
+ { gen_helper_sve_ldbsu_zsu,
667
+ gen_helper_sve_ldhsu_le_zsu,
668
+ gen_helper_sve_ldss_le_zsu, } },
669
+ { { gen_helper_sve_ldbss_zss,
670
+ gen_helper_sve_ldhss_le_zss,
671
+ NULL, },
672
+ { gen_helper_sve_ldbsu_zss,
673
+ gen_helper_sve_ldhsu_le_zss,
674
+ gen_helper_sve_ldss_le_zss, } } },
675
676
- /* First-fault */
677
- { { { gen_helper_sve_ldffbss_zsu,
678
- gen_helper_sve_ldffhss_le_zsu,
679
- NULL, },
680
- { gen_helper_sve_ldffbsu_zsu,
681
- gen_helper_sve_ldffhsu_le_zsu,
682
- gen_helper_sve_ldffss_le_zsu, } },
683
- { { gen_helper_sve_ldffbss_zss,
684
- gen_helper_sve_ldffhss_le_zss,
685
- NULL, },
686
- { gen_helper_sve_ldffbsu_zss,
687
- gen_helper_sve_ldffhsu_le_zss,
688
- gen_helper_sve_ldffss_le_zss, } } } },
689
+ /* First-fault */
690
+ { { { gen_helper_sve_ldffbss_zsu,
691
+ gen_helper_sve_ldffhss_le_zsu,
692
+ NULL, },
693
+ { gen_helper_sve_ldffbsu_zsu,
694
+ gen_helper_sve_ldffhsu_le_zsu,
695
+ gen_helper_sve_ldffss_le_zsu, } },
696
+ { { gen_helper_sve_ldffbss_zss,
697
+ gen_helper_sve_ldffhss_le_zss,
698
+ NULL, },
699
+ { gen_helper_sve_ldffbsu_zss,
700
+ gen_helper_sve_ldffhsu_le_zss,
701
+ gen_helper_sve_ldffss_le_zss, } } } },
702
703
- /* Big-endian */
704
- { { { { gen_helper_sve_ldbss_zsu,
705
- gen_helper_sve_ldhss_be_zsu,
706
- NULL, },
707
- { gen_helper_sve_ldbsu_zsu,
708
- gen_helper_sve_ldhsu_be_zsu,
709
- gen_helper_sve_ldss_be_zsu, } },
710
- { { gen_helper_sve_ldbss_zss,
711
- gen_helper_sve_ldhss_be_zss,
712
- NULL, },
713
- { gen_helper_sve_ldbsu_zss,
714
- gen_helper_sve_ldhsu_be_zss,
715
- gen_helper_sve_ldss_be_zss, } } },
716
+ { /* Big-endian */
717
+ { { { gen_helper_sve_ldbss_zsu,
718
+ gen_helper_sve_ldhss_be_zsu,
719
+ NULL, },
720
+ { gen_helper_sve_ldbsu_zsu,
721
+ gen_helper_sve_ldhsu_be_zsu,
722
+ gen_helper_sve_ldss_be_zsu, } },
723
+ { { gen_helper_sve_ldbss_zss,
724
+ gen_helper_sve_ldhss_be_zss,
725
+ NULL, },
726
+ { gen_helper_sve_ldbsu_zss,
727
+ gen_helper_sve_ldhsu_be_zss,
728
+ gen_helper_sve_ldss_be_zss, } } },
729
730
- /* First-fault */
731
- { { { gen_helper_sve_ldffbss_zsu,
732
- gen_helper_sve_ldffhss_be_zsu,
733
- NULL, },
734
- { gen_helper_sve_ldffbsu_zsu,
735
- gen_helper_sve_ldffhsu_be_zsu,
736
- gen_helper_sve_ldffss_be_zsu, } },
737
- { { gen_helper_sve_ldffbss_zss,
738
- gen_helper_sve_ldffhss_be_zss,
739
- NULL, },
740
- { gen_helper_sve_ldffbsu_zss,
741
- gen_helper_sve_ldffhsu_be_zss,
742
- gen_helper_sve_ldffss_be_zss, } } } },
743
+ /* First-fault */
744
+ { { { gen_helper_sve_ldffbss_zsu,
745
+ gen_helper_sve_ldffhss_be_zsu,
746
+ NULL, },
747
+ { gen_helper_sve_ldffbsu_zsu,
748
+ gen_helper_sve_ldffhsu_be_zsu,
749
+ gen_helper_sve_ldffss_be_zsu, } },
750
+ { { gen_helper_sve_ldffbss_zss,
751
+ gen_helper_sve_ldffhss_be_zss,
752
+ NULL, },
753
+ { gen_helper_sve_ldffbsu_zss,
754
+ gen_helper_sve_ldffhsu_be_zss,
755
+ gen_helper_sve_ldffss_be_zss, } } } } },
756
+ { /* MTE Active */
757
+ { /* Little-endian */
758
+ { { { gen_helper_sve_ldbss_zsu_mte,
759
+ gen_helper_sve_ldhss_le_zsu_mte,
760
+ NULL, },
761
+ { gen_helper_sve_ldbsu_zsu_mte,
762
+ gen_helper_sve_ldhsu_le_zsu_mte,
763
+ gen_helper_sve_ldss_le_zsu_mte, } },
764
+ { { gen_helper_sve_ldbss_zss_mte,
765
+ gen_helper_sve_ldhss_le_zss_mte,
766
+ NULL, },
767
+ { gen_helper_sve_ldbsu_zss_mte,
768
+ gen_helper_sve_ldhsu_le_zss_mte,
769
+ gen_helper_sve_ldss_le_zss_mte, } } },
770
+
771
+ /* First-fault */
772
+ { { { gen_helper_sve_ldffbss_zsu_mte,
773
+ gen_helper_sve_ldffhss_le_zsu_mte,
774
+ NULL, },
775
+ { gen_helper_sve_ldffbsu_zsu_mte,
776
+ gen_helper_sve_ldffhsu_le_zsu_mte,
777
+ gen_helper_sve_ldffss_le_zsu_mte, } },
778
+ { { gen_helper_sve_ldffbss_zss_mte,
779
+ gen_helper_sve_ldffhss_le_zss_mte,
780
+ NULL, },
781
+ { gen_helper_sve_ldffbsu_zss_mte,
782
+ gen_helper_sve_ldffhsu_le_zss_mte,
783
+ gen_helper_sve_ldffss_le_zss_mte, } } } },
784
+
785
+ { /* Big-endian */
786
+ { { { gen_helper_sve_ldbss_zsu_mte,
787
+ gen_helper_sve_ldhss_be_zsu_mte,
788
+ NULL, },
789
+ { gen_helper_sve_ldbsu_zsu_mte,
790
+ gen_helper_sve_ldhsu_be_zsu_mte,
791
+ gen_helper_sve_ldss_be_zsu_mte, } },
792
+ { { gen_helper_sve_ldbss_zss_mte,
793
+ gen_helper_sve_ldhss_be_zss_mte,
794
+ NULL, },
795
+ { gen_helper_sve_ldbsu_zss_mte,
796
+ gen_helper_sve_ldhsu_be_zss_mte,
797
+ gen_helper_sve_ldss_be_zss_mte, } } },
798
+
799
+ /* First-fault */
800
+ { { { gen_helper_sve_ldffbss_zsu_mte,
801
+ gen_helper_sve_ldffhss_be_zsu_mte,
802
+ NULL, },
803
+ { gen_helper_sve_ldffbsu_zsu_mte,
804
+ gen_helper_sve_ldffhsu_be_zsu_mte,
805
+ gen_helper_sve_ldffss_be_zsu_mte, } },
806
+ { { gen_helper_sve_ldffbss_zss_mte,
807
+ gen_helper_sve_ldffhss_be_zss_mte,
808
+ NULL, },
809
+ { gen_helper_sve_ldffbsu_zss_mte,
810
+ gen_helper_sve_ldffhsu_be_zss_mte,
811
+ gen_helper_sve_ldffss_be_zss_mte, } } } } },
37
};
812
};
38
813
39
-qemu_irq *tc6393xb_gpio_in_get(TC6393xbState *s)
814
/* Note that we overload xs=2 to indicate 64-bit offset. */
40
-{
815
-static gen_helper_gvec_mem_scatter * const gather_load_fn64[2][2][3][2][4] = {
41
- return s->gpio_in;
816
- /* Little-endian */
42
-}
817
- { { { { gen_helper_sve_ldbds_zsu,
43
-
818
- gen_helper_sve_ldhds_le_zsu,
44
static void tc6393xb_gpio_set(void *opaque, int line, int level)
819
- gen_helper_sve_ldsds_le_zsu,
820
- NULL, },
821
- { gen_helper_sve_ldbdu_zsu,
822
- gen_helper_sve_ldhdu_le_zsu,
823
- gen_helper_sve_ldsdu_le_zsu,
824
- gen_helper_sve_lddd_le_zsu, } },
825
- { { gen_helper_sve_ldbds_zss,
826
- gen_helper_sve_ldhds_le_zss,
827
- gen_helper_sve_ldsds_le_zss,
828
- NULL, },
829
- { gen_helper_sve_ldbdu_zss,
830
- gen_helper_sve_ldhdu_le_zss,
831
- gen_helper_sve_ldsdu_le_zss,
832
- gen_helper_sve_lddd_le_zss, } },
833
- { { gen_helper_sve_ldbds_zd,
834
- gen_helper_sve_ldhds_le_zd,
835
- gen_helper_sve_ldsds_le_zd,
836
- NULL, },
837
- { gen_helper_sve_ldbdu_zd,
838
- gen_helper_sve_ldhdu_le_zd,
839
- gen_helper_sve_ldsdu_le_zd,
840
- gen_helper_sve_lddd_le_zd, } } },
841
+static gen_helper_gvec_mem_scatter * const
842
+gather_load_fn64[2][2][2][3][2][4] = {
843
+ { /* MTE Inactive */
844
+ { /* Little-endian */
845
+ { { { gen_helper_sve_ldbds_zsu,
846
+ gen_helper_sve_ldhds_le_zsu,
847
+ gen_helper_sve_ldsds_le_zsu,
848
+ NULL, },
849
+ { gen_helper_sve_ldbdu_zsu,
850
+ gen_helper_sve_ldhdu_le_zsu,
851
+ gen_helper_sve_ldsdu_le_zsu,
852
+ gen_helper_sve_lddd_le_zsu, } },
853
+ { { gen_helper_sve_ldbds_zss,
854
+ gen_helper_sve_ldhds_le_zss,
855
+ gen_helper_sve_ldsds_le_zss,
856
+ NULL, },
857
+ { gen_helper_sve_ldbdu_zss,
858
+ gen_helper_sve_ldhdu_le_zss,
859
+ gen_helper_sve_ldsdu_le_zss,
860
+ gen_helper_sve_lddd_le_zss, } },
861
+ { { gen_helper_sve_ldbds_zd,
862
+ gen_helper_sve_ldhds_le_zd,
863
+ gen_helper_sve_ldsds_le_zd,
864
+ NULL, },
865
+ { gen_helper_sve_ldbdu_zd,
866
+ gen_helper_sve_ldhdu_le_zd,
867
+ gen_helper_sve_ldsdu_le_zd,
868
+ gen_helper_sve_lddd_le_zd, } } },
869
870
- /* First-fault */
871
- { { { gen_helper_sve_ldffbds_zsu,
872
- gen_helper_sve_ldffhds_le_zsu,
873
- gen_helper_sve_ldffsds_le_zsu,
874
- NULL, },
875
- { gen_helper_sve_ldffbdu_zsu,
876
- gen_helper_sve_ldffhdu_le_zsu,
877
- gen_helper_sve_ldffsdu_le_zsu,
878
- gen_helper_sve_ldffdd_le_zsu, } },
879
- { { gen_helper_sve_ldffbds_zss,
880
- gen_helper_sve_ldffhds_le_zss,
881
- gen_helper_sve_ldffsds_le_zss,
882
- NULL, },
883
- { gen_helper_sve_ldffbdu_zss,
884
- gen_helper_sve_ldffhdu_le_zss,
885
- gen_helper_sve_ldffsdu_le_zss,
886
- gen_helper_sve_ldffdd_le_zss, } },
887
- { { gen_helper_sve_ldffbds_zd,
888
- gen_helper_sve_ldffhds_le_zd,
889
- gen_helper_sve_ldffsds_le_zd,
890
- NULL, },
891
- { gen_helper_sve_ldffbdu_zd,
892
- gen_helper_sve_ldffhdu_le_zd,
893
- gen_helper_sve_ldffsdu_le_zd,
894
- gen_helper_sve_ldffdd_le_zd, } } } },
895
+ /* First-fault */
896
+ { { { gen_helper_sve_ldffbds_zsu,
897
+ gen_helper_sve_ldffhds_le_zsu,
898
+ gen_helper_sve_ldffsds_le_zsu,
899
+ NULL, },
900
+ { gen_helper_sve_ldffbdu_zsu,
901
+ gen_helper_sve_ldffhdu_le_zsu,
902
+ gen_helper_sve_ldffsdu_le_zsu,
903
+ gen_helper_sve_ldffdd_le_zsu, } },
904
+ { { gen_helper_sve_ldffbds_zss,
905
+ gen_helper_sve_ldffhds_le_zss,
906
+ gen_helper_sve_ldffsds_le_zss,
907
+ NULL, },
908
+ { gen_helper_sve_ldffbdu_zss,
909
+ gen_helper_sve_ldffhdu_le_zss,
910
+ gen_helper_sve_ldffsdu_le_zss,
911
+ gen_helper_sve_ldffdd_le_zss, } },
912
+ { { gen_helper_sve_ldffbds_zd,
913
+ gen_helper_sve_ldffhds_le_zd,
914
+ gen_helper_sve_ldffsds_le_zd,
915
+ NULL, },
916
+ { gen_helper_sve_ldffbdu_zd,
917
+ gen_helper_sve_ldffhdu_le_zd,
918
+ gen_helper_sve_ldffsdu_le_zd,
919
+ gen_helper_sve_ldffdd_le_zd, } } } },
920
+ { /* Big-endian */
921
+ { { { gen_helper_sve_ldbds_zsu,
922
+ gen_helper_sve_ldhds_be_zsu,
923
+ gen_helper_sve_ldsds_be_zsu,
924
+ NULL, },
925
+ { gen_helper_sve_ldbdu_zsu,
926
+ gen_helper_sve_ldhdu_be_zsu,
927
+ gen_helper_sve_ldsdu_be_zsu,
928
+ gen_helper_sve_lddd_be_zsu, } },
929
+ { { gen_helper_sve_ldbds_zss,
930
+ gen_helper_sve_ldhds_be_zss,
931
+ gen_helper_sve_ldsds_be_zss,
932
+ NULL, },
933
+ { gen_helper_sve_ldbdu_zss,
934
+ gen_helper_sve_ldhdu_be_zss,
935
+ gen_helper_sve_ldsdu_be_zss,
936
+ gen_helper_sve_lddd_be_zss, } },
937
+ { { gen_helper_sve_ldbds_zd,
938
+ gen_helper_sve_ldhds_be_zd,
939
+ gen_helper_sve_ldsds_be_zd,
940
+ NULL, },
941
+ { gen_helper_sve_ldbdu_zd,
942
+ gen_helper_sve_ldhdu_be_zd,
943
+ gen_helper_sve_ldsdu_be_zd,
944
+ gen_helper_sve_lddd_be_zd, } } },
945
946
- /* Big-endian */
947
- { { { { gen_helper_sve_ldbds_zsu,
948
- gen_helper_sve_ldhds_be_zsu,
949
- gen_helper_sve_ldsds_be_zsu,
950
- NULL, },
951
- { gen_helper_sve_ldbdu_zsu,
952
- gen_helper_sve_ldhdu_be_zsu,
953
- gen_helper_sve_ldsdu_be_zsu,
954
- gen_helper_sve_lddd_be_zsu, } },
955
- { { gen_helper_sve_ldbds_zss,
956
- gen_helper_sve_ldhds_be_zss,
957
- gen_helper_sve_ldsds_be_zss,
958
- NULL, },
959
- { gen_helper_sve_ldbdu_zss,
960
- gen_helper_sve_ldhdu_be_zss,
961
- gen_helper_sve_ldsdu_be_zss,
962
- gen_helper_sve_lddd_be_zss, } },
963
- { { gen_helper_sve_ldbds_zd,
964
- gen_helper_sve_ldhds_be_zd,
965
- gen_helper_sve_ldsds_be_zd,
966
- NULL, },
967
- { gen_helper_sve_ldbdu_zd,
968
- gen_helper_sve_ldhdu_be_zd,
969
- gen_helper_sve_ldsdu_be_zd,
970
- gen_helper_sve_lddd_be_zd, } } },
971
+ /* First-fault */
972
+ { { { gen_helper_sve_ldffbds_zsu,
973
+ gen_helper_sve_ldffhds_be_zsu,
974
+ gen_helper_sve_ldffsds_be_zsu,
975
+ NULL, },
976
+ { gen_helper_sve_ldffbdu_zsu,
977
+ gen_helper_sve_ldffhdu_be_zsu,
978
+ gen_helper_sve_ldffsdu_be_zsu,
979
+ gen_helper_sve_ldffdd_be_zsu, } },
980
+ { { gen_helper_sve_ldffbds_zss,
981
+ gen_helper_sve_ldffhds_be_zss,
982
+ gen_helper_sve_ldffsds_be_zss,
983
+ NULL, },
984
+ { gen_helper_sve_ldffbdu_zss,
985
+ gen_helper_sve_ldffhdu_be_zss,
986
+ gen_helper_sve_ldffsdu_be_zss,
987
+ gen_helper_sve_ldffdd_be_zss, } },
988
+ { { gen_helper_sve_ldffbds_zd,
989
+ gen_helper_sve_ldffhds_be_zd,
990
+ gen_helper_sve_ldffsds_be_zd,
991
+ NULL, },
992
+ { gen_helper_sve_ldffbdu_zd,
993
+ gen_helper_sve_ldffhdu_be_zd,
994
+ gen_helper_sve_ldffsdu_be_zd,
995
+ gen_helper_sve_ldffdd_be_zd, } } } } },
996
+ { /* MTE Active */
997
+ { /* Little-endian */
998
+ { { { gen_helper_sve_ldbds_zsu_mte,
999
+ gen_helper_sve_ldhds_le_zsu_mte,
1000
+ gen_helper_sve_ldsds_le_zsu_mte,
1001
+ NULL, },
1002
+ { gen_helper_sve_ldbdu_zsu_mte,
1003
+ gen_helper_sve_ldhdu_le_zsu_mte,
1004
+ gen_helper_sve_ldsdu_le_zsu_mte,
1005
+ gen_helper_sve_lddd_le_zsu_mte, } },
1006
+ { { gen_helper_sve_ldbds_zss_mte,
1007
+ gen_helper_sve_ldhds_le_zss_mte,
1008
+ gen_helper_sve_ldsds_le_zss_mte,
1009
+ NULL, },
1010
+ { gen_helper_sve_ldbdu_zss_mte,
1011
+ gen_helper_sve_ldhdu_le_zss_mte,
1012
+ gen_helper_sve_ldsdu_le_zss_mte,
1013
+ gen_helper_sve_lddd_le_zss_mte, } },
1014
+ { { gen_helper_sve_ldbds_zd_mte,
1015
+ gen_helper_sve_ldhds_le_zd_mte,
1016
+ gen_helper_sve_ldsds_le_zd_mte,
1017
+ NULL, },
1018
+ { gen_helper_sve_ldbdu_zd_mte,
1019
+ gen_helper_sve_ldhdu_le_zd_mte,
1020
+ gen_helper_sve_ldsdu_le_zd_mte,
1021
+ gen_helper_sve_lddd_le_zd_mte, } } },
1022
1023
- /* First-fault */
1024
- { { { gen_helper_sve_ldffbds_zsu,
1025
- gen_helper_sve_ldffhds_be_zsu,
1026
- gen_helper_sve_ldffsds_be_zsu,
1027
- NULL, },
1028
- { gen_helper_sve_ldffbdu_zsu,
1029
- gen_helper_sve_ldffhdu_be_zsu,
1030
- gen_helper_sve_ldffsdu_be_zsu,
1031
- gen_helper_sve_ldffdd_be_zsu, } },
1032
- { { gen_helper_sve_ldffbds_zss,
1033
- gen_helper_sve_ldffhds_be_zss,
1034
- gen_helper_sve_ldffsds_be_zss,
1035
- NULL, },
1036
- { gen_helper_sve_ldffbdu_zss,
1037
- gen_helper_sve_ldffhdu_be_zss,
1038
- gen_helper_sve_ldffsdu_be_zss,
1039
- gen_helper_sve_ldffdd_be_zss, } },
1040
- { { gen_helper_sve_ldffbds_zd,
1041
- gen_helper_sve_ldffhds_be_zd,
1042
- gen_helper_sve_ldffsds_be_zd,
1043
- NULL, },
1044
- { gen_helper_sve_ldffbdu_zd,
1045
- gen_helper_sve_ldffhdu_be_zd,
1046
- gen_helper_sve_ldffsdu_be_zd,
1047
- gen_helper_sve_ldffdd_be_zd, } } } },
1048
+ /* First-fault */
1049
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1050
+ gen_helper_sve_ldffhds_le_zsu_mte,
1051
+ gen_helper_sve_ldffsds_le_zsu_mte,
1052
+ NULL, },
1053
+ { gen_helper_sve_ldffbdu_zsu_mte,
1054
+ gen_helper_sve_ldffhdu_le_zsu_mte,
1055
+ gen_helper_sve_ldffsdu_le_zsu_mte,
1056
+ gen_helper_sve_ldffdd_le_zsu_mte, } },
1057
+ { { gen_helper_sve_ldffbds_zss_mte,
1058
+ gen_helper_sve_ldffhds_le_zss_mte,
1059
+ gen_helper_sve_ldffsds_le_zss_mte,
1060
+ NULL, },
1061
+ { gen_helper_sve_ldffbdu_zss_mte,
1062
+ gen_helper_sve_ldffhdu_le_zss_mte,
1063
+ gen_helper_sve_ldffsdu_le_zss_mte,
1064
+ gen_helper_sve_ldffdd_le_zss_mte, } },
1065
+ { { gen_helper_sve_ldffbds_zd_mte,
1066
+ gen_helper_sve_ldffhds_le_zd_mte,
1067
+ gen_helper_sve_ldffsds_le_zd_mte,
1068
+ NULL, },
1069
+ { gen_helper_sve_ldffbdu_zd_mte,
1070
+ gen_helper_sve_ldffhdu_le_zd_mte,
1071
+ gen_helper_sve_ldffsdu_le_zd_mte,
1072
+ gen_helper_sve_ldffdd_le_zd_mte, } } } },
1073
+ { /* Big-endian */
1074
+ { { { gen_helper_sve_ldbds_zsu_mte,
1075
+ gen_helper_sve_ldhds_be_zsu_mte,
1076
+ gen_helper_sve_ldsds_be_zsu_mte,
1077
+ NULL, },
1078
+ { gen_helper_sve_ldbdu_zsu_mte,
1079
+ gen_helper_sve_ldhdu_be_zsu_mte,
1080
+ gen_helper_sve_ldsdu_be_zsu_mte,
1081
+ gen_helper_sve_lddd_be_zsu_mte, } },
1082
+ { { gen_helper_sve_ldbds_zss_mte,
1083
+ gen_helper_sve_ldhds_be_zss_mte,
1084
+ gen_helper_sve_ldsds_be_zss_mte,
1085
+ NULL, },
1086
+ { gen_helper_sve_ldbdu_zss_mte,
1087
+ gen_helper_sve_ldhdu_be_zss_mte,
1088
+ gen_helper_sve_ldsdu_be_zss_mte,
1089
+ gen_helper_sve_lddd_be_zss_mte, } },
1090
+ { { gen_helper_sve_ldbds_zd_mte,
1091
+ gen_helper_sve_ldhds_be_zd_mte,
1092
+ gen_helper_sve_ldsds_be_zd_mte,
1093
+ NULL, },
1094
+ { gen_helper_sve_ldbdu_zd_mte,
1095
+ gen_helper_sve_ldhdu_be_zd_mte,
1096
+ gen_helper_sve_ldsdu_be_zd_mte,
1097
+ gen_helper_sve_lddd_be_zd_mte, } } },
1098
+
1099
+ /* First-fault */
1100
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1101
+ gen_helper_sve_ldffhds_be_zsu_mte,
1102
+ gen_helper_sve_ldffsds_be_zsu_mte,
1103
+ NULL, },
1104
+ { gen_helper_sve_ldffbdu_zsu_mte,
1105
+ gen_helper_sve_ldffhdu_be_zsu_mte,
1106
+ gen_helper_sve_ldffsdu_be_zsu_mte,
1107
+ gen_helper_sve_ldffdd_be_zsu_mte, } },
1108
+ { { gen_helper_sve_ldffbds_zss_mte,
1109
+ gen_helper_sve_ldffhds_be_zss_mte,
1110
+ gen_helper_sve_ldffsds_be_zss_mte,
1111
+ NULL, },
1112
+ { gen_helper_sve_ldffbdu_zss_mte,
1113
+ gen_helper_sve_ldffhdu_be_zss_mte,
1114
+ gen_helper_sve_ldffsdu_be_zss_mte,
1115
+ gen_helper_sve_ldffdd_be_zss_mte, } },
1116
+ { { gen_helper_sve_ldffbds_zd_mte,
1117
+ gen_helper_sve_ldffhds_be_zd_mte,
1118
+ gen_helper_sve_ldffsds_be_zd_mte,
1119
+ NULL, },
1120
+ { gen_helper_sve_ldffbdu_zd_mte,
1121
+ gen_helper_sve_ldffhdu_be_zd_mte,
1122
+ gen_helper_sve_ldffsdu_be_zd_mte,
1123
+ gen_helper_sve_ldffdd_be_zd_mte, } } } } },
1124
};
1125
1126
static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
45
{
1127
{
46
// TC6393xbState *s = opaque;
1128
gen_helper_gvec_mem_scatter *fn = NULL;
47
@@ -XXX,XX +XXX,XX @@ static void tc6393xb_gpio_set(void *opaque, int line, int level)
1129
- int be = s->be_data == MO_BE;
48
// FIXME: how does the chip reflect the GPIO input level change?
1130
+ bool be = s->be_data == MO_BE;
1131
+ bool mte = s->mte_active[0];
1132
1133
if (!sve_access_check(s)) {
1134
return true;
1135
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1136
1137
switch (a->esz) {
1138
case MO_32:
1139
- fn = gather_load_fn32[be][a->ff][a->xs][a->u][a->msz];
1140
+ fn = gather_load_fn32[mte][be][a->ff][a->xs][a->u][a->msz];
1141
break;
1142
case MO_64:
1143
- fn = gather_load_fn64[be][a->ff][a->xs][a->u][a->msz];
1144
+ fn = gather_load_fn64[mte][be][a->ff][a->xs][a->u][a->msz];
1145
break;
1146
}
1147
assert(fn != NULL);
1148
1149
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1150
- cpu_reg_sp(s, a->rn), a->msz, fn);
1151
+ cpu_reg_sp(s, a->rn), a->msz, false, fn);
1152
return true;
49
}
1153
}
50
1154
51
-void tc6393xb_gpio_out_set(TC6393xbState *s, int line,
1155
static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
52
- qemu_irq handler)
53
-{
54
- if (line >= TC6393XB_GPIOS) {
55
- fprintf(stderr, "TC6393xb: no GPIO pin %d\n", line);
56
- return;
57
- }
58
-
59
- s->handler[line] = handler;
60
-}
61
-
62
static void tc6393xb_gpio_handler_update(TC6393xbState *s)
63
{
1156
{
64
uint32_t level, diff;
1157
gen_helper_gvec_mem_scatter *fn = NULL;
1158
- int be = s->be_data == MO_BE;
1159
+ bool be = s->be_data == MO_BE;
1160
+ bool mte = s->mte_active[0];
1161
TCGv_i64 imm;
1162
1163
if (a->esz < a->msz || (a->esz == a->msz && !a->u)) {
1164
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1165
1166
switch (a->esz) {
1167
case MO_32:
1168
- fn = gather_load_fn32[be][a->ff][0][a->u][a->msz];
1169
+ fn = gather_load_fn32[mte][be][a->ff][0][a->u][a->msz];
1170
break;
1171
case MO_64:
1172
- fn = gather_load_fn64[be][a->ff][2][a->u][a->msz];
1173
+ fn = gather_load_fn64[mte][be][a->ff][2][a->u][a->msz];
1174
break;
1175
}
1176
assert(fn != NULL);
1177
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1178
* by loading the immediate into the scalar parameter.
1179
*/
1180
imm = tcg_const_i64(a->imm << a->msz);
1181
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1182
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, false, fn);
1183
tcg_temp_free_i64(imm);
1184
return true;
1185
}
1186
1187
-/* Indexed by [be][xs][msz]. */
1188
-static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][3] = {
1189
- /* Little-endian */
1190
- { { gen_helper_sve_stbs_zsu,
1191
- gen_helper_sve_sths_le_zsu,
1192
- gen_helper_sve_stss_le_zsu, },
1193
- { gen_helper_sve_stbs_zss,
1194
- gen_helper_sve_sths_le_zss,
1195
- gen_helper_sve_stss_le_zss, } },
1196
- /* Big-endian */
1197
- { { gen_helper_sve_stbs_zsu,
1198
- gen_helper_sve_sths_be_zsu,
1199
- gen_helper_sve_stss_be_zsu, },
1200
- { gen_helper_sve_stbs_zss,
1201
- gen_helper_sve_sths_be_zss,
1202
- gen_helper_sve_stss_be_zss, } },
1203
+/* Indexed by [mte][be][xs][msz]. */
1204
+static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = {
1205
+ { /* MTE Inactive */
1206
+ { /* Little-endian */
1207
+ { gen_helper_sve_stbs_zsu,
1208
+ gen_helper_sve_sths_le_zsu,
1209
+ gen_helper_sve_stss_le_zsu, },
1210
+ { gen_helper_sve_stbs_zss,
1211
+ gen_helper_sve_sths_le_zss,
1212
+ gen_helper_sve_stss_le_zss, } },
1213
+ { /* Big-endian */
1214
+ { gen_helper_sve_stbs_zsu,
1215
+ gen_helper_sve_sths_be_zsu,
1216
+ gen_helper_sve_stss_be_zsu, },
1217
+ { gen_helper_sve_stbs_zss,
1218
+ gen_helper_sve_sths_be_zss,
1219
+ gen_helper_sve_stss_be_zss, } } },
1220
+ { /* MTE Active */
1221
+ { /* Little-endian */
1222
+ { gen_helper_sve_stbs_zsu_mte,
1223
+ gen_helper_sve_sths_le_zsu_mte,
1224
+ gen_helper_sve_stss_le_zsu_mte, },
1225
+ { gen_helper_sve_stbs_zss_mte,
1226
+ gen_helper_sve_sths_le_zss_mte,
1227
+ gen_helper_sve_stss_le_zss_mte, } },
1228
+ { /* Big-endian */
1229
+ { gen_helper_sve_stbs_zsu_mte,
1230
+ gen_helper_sve_sths_be_zsu_mte,
1231
+ gen_helper_sve_stss_be_zsu_mte, },
1232
+ { gen_helper_sve_stbs_zss_mte,
1233
+ gen_helper_sve_sths_be_zss_mte,
1234
+ gen_helper_sve_stss_be_zss_mte, } } },
1235
};
1236
1237
/* Note that we overload xs=2 to indicate 64-bit offset. */
1238
-static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][3][4] = {
1239
- /* Little-endian */
1240
- { { gen_helper_sve_stbd_zsu,
1241
- gen_helper_sve_sthd_le_zsu,
1242
- gen_helper_sve_stsd_le_zsu,
1243
- gen_helper_sve_stdd_le_zsu, },
1244
- { gen_helper_sve_stbd_zss,
1245
- gen_helper_sve_sthd_le_zss,
1246
- gen_helper_sve_stsd_le_zss,
1247
- gen_helper_sve_stdd_le_zss, },
1248
- { gen_helper_sve_stbd_zd,
1249
- gen_helper_sve_sthd_le_zd,
1250
- gen_helper_sve_stsd_le_zd,
1251
- gen_helper_sve_stdd_le_zd, } },
1252
- /* Big-endian */
1253
- { { gen_helper_sve_stbd_zsu,
1254
- gen_helper_sve_sthd_be_zsu,
1255
- gen_helper_sve_stsd_be_zsu,
1256
- gen_helper_sve_stdd_be_zsu, },
1257
- { gen_helper_sve_stbd_zss,
1258
- gen_helper_sve_sthd_be_zss,
1259
- gen_helper_sve_stsd_be_zss,
1260
- gen_helper_sve_stdd_be_zss, },
1261
- { gen_helper_sve_stbd_zd,
1262
- gen_helper_sve_sthd_be_zd,
1263
- gen_helper_sve_stsd_be_zd,
1264
- gen_helper_sve_stdd_be_zd, } },
1265
+static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][2][3][4] = {
1266
+ { /* MTE Inactive */
1267
+ { /* Little-endian */
1268
+ { gen_helper_sve_stbd_zsu,
1269
+ gen_helper_sve_sthd_le_zsu,
1270
+ gen_helper_sve_stsd_le_zsu,
1271
+ gen_helper_sve_stdd_le_zsu, },
1272
+ { gen_helper_sve_stbd_zss,
1273
+ gen_helper_sve_sthd_le_zss,
1274
+ gen_helper_sve_stsd_le_zss,
1275
+ gen_helper_sve_stdd_le_zss, },
1276
+ { gen_helper_sve_stbd_zd,
1277
+ gen_helper_sve_sthd_le_zd,
1278
+ gen_helper_sve_stsd_le_zd,
1279
+ gen_helper_sve_stdd_le_zd, } },
1280
+ { /* Big-endian */
1281
+ { gen_helper_sve_stbd_zsu,
1282
+ gen_helper_sve_sthd_be_zsu,
1283
+ gen_helper_sve_stsd_be_zsu,
1284
+ gen_helper_sve_stdd_be_zsu, },
1285
+ { gen_helper_sve_stbd_zss,
1286
+ gen_helper_sve_sthd_be_zss,
1287
+ gen_helper_sve_stsd_be_zss,
1288
+ gen_helper_sve_stdd_be_zss, },
1289
+ { gen_helper_sve_stbd_zd,
1290
+ gen_helper_sve_sthd_be_zd,
1291
+ gen_helper_sve_stsd_be_zd,
1292
+ gen_helper_sve_stdd_be_zd, } } },
1293
+ { /* MTE Inactive */
1294
+ { /* Little-endian */
1295
+ { gen_helper_sve_stbd_zsu_mte,
1296
+ gen_helper_sve_sthd_le_zsu_mte,
1297
+ gen_helper_sve_stsd_le_zsu_mte,
1298
+ gen_helper_sve_stdd_le_zsu_mte, },
1299
+ { gen_helper_sve_stbd_zss_mte,
1300
+ gen_helper_sve_sthd_le_zss_mte,
1301
+ gen_helper_sve_stsd_le_zss_mte,
1302
+ gen_helper_sve_stdd_le_zss_mte, },
1303
+ { gen_helper_sve_stbd_zd_mte,
1304
+ gen_helper_sve_sthd_le_zd_mte,
1305
+ gen_helper_sve_stsd_le_zd_mte,
1306
+ gen_helper_sve_stdd_le_zd_mte, } },
1307
+ { /* Big-endian */
1308
+ { gen_helper_sve_stbd_zsu_mte,
1309
+ gen_helper_sve_sthd_be_zsu_mte,
1310
+ gen_helper_sve_stsd_be_zsu_mte,
1311
+ gen_helper_sve_stdd_be_zsu_mte, },
1312
+ { gen_helper_sve_stbd_zss_mte,
1313
+ gen_helper_sve_sthd_be_zss_mte,
1314
+ gen_helper_sve_stsd_be_zss_mte,
1315
+ gen_helper_sve_stdd_be_zss_mte, },
1316
+ { gen_helper_sve_stbd_zd_mte,
1317
+ gen_helper_sve_sthd_be_zd_mte,
1318
+ gen_helper_sve_stsd_be_zd_mte,
1319
+ gen_helper_sve_stdd_be_zd_mte, } } },
1320
};
1321
1322
static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1323
{
1324
gen_helper_gvec_mem_scatter *fn;
1325
- int be = s->be_data == MO_BE;
1326
+ bool be = s->be_data == MO_BE;
1327
+ bool mte = s->mte_active[0];
1328
1329
if (a->esz < a->msz || (a->msz == 0 && a->scale)) {
1330
return false;
1331
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1332
}
1333
switch (a->esz) {
1334
case MO_32:
1335
- fn = scatter_store_fn32[be][a->xs][a->msz];
1336
+ fn = scatter_store_fn32[mte][be][a->xs][a->msz];
1337
break;
1338
case MO_64:
1339
- fn = scatter_store_fn64[be][a->xs][a->msz];
1340
+ fn = scatter_store_fn64[mte][be][a->xs][a->msz];
1341
break;
1342
default:
1343
g_assert_not_reached();
1344
}
1345
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1346
- cpu_reg_sp(s, a->rn), a->msz, fn);
1347
+ cpu_reg_sp(s, a->rn), a->msz, true, fn);
1348
return true;
1349
}
1350
1351
static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1352
{
1353
gen_helper_gvec_mem_scatter *fn = NULL;
1354
- int be = s->be_data == MO_BE;
1355
+ bool be = s->be_data == MO_BE;
1356
+ bool mte = s->mte_active[0];
1357
TCGv_i64 imm;
1358
1359
if (a->esz < a->msz) {
1360
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1361
1362
switch (a->esz) {
1363
case MO_32:
1364
- fn = scatter_store_fn32[be][0][a->msz];
1365
+ fn = scatter_store_fn32[mte][be][0][a->msz];
1366
break;
1367
case MO_64:
1368
- fn = scatter_store_fn64[be][2][a->msz];
1369
+ fn = scatter_store_fn64[mte][be][2][a->msz];
1370
break;
1371
}
1372
assert(fn != NULL);
1373
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1374
* by loading the immediate into the scalar parameter.
1375
*/
1376
imm = tcg_const_i64(a->imm << a->msz);
1377
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1378
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, true, fn);
1379
tcg_temp_free_i64(imm);
1380
return true;
1381
}
65
--
1382
--
66
2.20.1
1383
2.20.1
67
1384
68
1385
diff view generated by jsdifflib
1
Enable the FPU by default for the Cortex-M4 and Cortex-M33.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
There are a number of paths by which the TBI is still intact
4
for user-only in the SVE helpers.
5
6
Because we currently always set TBI for user-only, we do not
7
need to pass down the actual TBI setting from above, and we
8
can remove the top byte in the inner-most primitives, so that
9
none are forgotten. Moreover, this keeps the "dirty" pointer
10
around at the higher levels, where we need it for any MTE checking.
11
12
Since the normal case, especially for user-only, goes through
13
RAM, this clearing merely adds two insns per page lookup, which
14
will be completely in the noise.
15
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20200626033144.790098-39-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20190416125744.27770-27-peter.maydell@linaro.org
6
---
20
---
7
target/arm/cpu.c | 8 ++++++++
21
target/arm/cpu.c | 3 +++
8
1 file changed, 8 insertions(+)
22
target/arm/sve_helper.c | 19 +++++++++++++++++--
23
target/arm/translate-a64.c | 5 +++++
24
3 files changed, 25 insertions(+), 2 deletions(-)
9
25
10
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
26
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
11
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
12
--- a/target/arm/cpu.c
28
--- a/target/arm/cpu.c
13
+++ b/target/arm/cpu.c
29
+++ b/target/arm/cpu.c
14
@@ -XXX,XX +XXX,XX @@ static void cortex_m4_initfn(Object *obj)
30
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
15
set_feature(&cpu->env, ARM_FEATURE_M);
31
* Enable TBI0 and TBI1. While the real kernel only enables TBI0,
16
set_feature(&cpu->env, ARM_FEATURE_M_MAIN);
32
* turning on both here will produce smaller code and otherwise
17
set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP);
33
* make no difference to the user-level emulation.
18
+ set_feature(&cpu->env, ARM_FEATURE_VFP4);
34
+ *
19
cpu->midr = 0x410fc240; /* r0p0 */
35
+ * In sve_probe_page, we assume that this is set.
20
cpu->pmsav7_dregion = 8;
36
+ * Do not modify this without other changes.
21
+ cpu->isar.mvfr0 = 0x10110021;
37
*/
22
+ cpu->isar.mvfr1 = 0x11000011;
38
env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
23
+ cpu->isar.mvfr2 = 0x00000000;
39
#else
24
cpu->id_pfr0 = 0x00000030;
40
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
25
cpu->id_pfr1 = 0x00000200;
41
index XXXXXXX..XXXXXXX 100644
26
cpu->id_dfr0 = 0x00100000;
42
--- a/target/arm/sve_helper.c
27
@@ -XXX,XX +XXX,XX @@ static void cortex_m33_initfn(Object *obj)
43
+++ b/target/arm/sve_helper.c
28
set_feature(&cpu->env, ARM_FEATURE_M_MAIN);
44
@@ -XXX,XX +XXX,XX @@ static void sve_##NAME##_host(void *vd, intptr_t reg_off, void *host) \
29
set_feature(&cpu->env, ARM_FEATURE_M_SECURITY);
45
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
30
set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP);
46
target_ulong addr, uintptr_t ra) \
31
+ set_feature(&cpu->env, ARM_FEATURE_VFP4);
47
{ \
32
cpu->midr = 0x410fd213; /* r0p3 */
48
- *(TYPEE *)(vd + H(reg_off)) = (TYPEM)TLB(env, addr, ra); \
33
cpu->pmsav7_dregion = 16;
49
+ *(TYPEE *)(vd + H(reg_off)) = \
34
cpu->sau_sregion = 8;
50
+ (TYPEM)TLB(env, useronly_clean_ptr(addr), ra); \
35
+ cpu->isar.mvfr0 = 0x10110021;
51
}
36
+ cpu->isar.mvfr1 = 0x11000011;
52
37
+ cpu->isar.mvfr2 = 0x00000040;
53
#define DO_ST_TLB(NAME, H, TYPEE, TYPEM, TLB) \
38
cpu->id_pfr0 = 0x00000030;
54
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
39
cpu->id_pfr1 = 0x00000210;
55
target_ulong addr, uintptr_t ra) \
40
cpu->id_dfr0 = 0x00200000;
56
{ \
57
- TLB(env, addr, (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
58
+ TLB(env, useronly_clean_ptr(addr), \
59
+ (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
60
}
61
62
#define DO_LD_PRIM_1(NAME, H, TE, TM) \
63
@@ -XXX,XX +XXX,XX @@ static bool sve_probe_page(SVEHostPage *info, bool nofault,
64
int flags;
65
66
addr += mem_off;
67
+
68
+ /*
69
+ * User-only currently always issues with TBI. See the comment
70
+ * above useronly_clean_ptr. Usually we clean this top byte away
71
+ * during translation, but we can't do that for e.g. vector + imm
72
+ * addressing modes.
73
+ *
74
+ * We currently always enable TBI for user-only, and do not provide
75
+ * a way to turn it off. So clean the pointer unconditionally here,
76
+ * rather than look it up here, or pass it down from above.
77
+ */
78
+ addr = useronly_clean_ptr(addr);
79
+
80
flags = probe_access_flags(env, addr, access_type, mmu_idx, nofault,
81
&info->host, retaddr);
82
info->flags = flags;
83
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/target/arm/translate-a64.c
86
+++ b/target/arm/translate-a64.c
87
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
88
dc->features = env->features;
89
dc->dcz_blocksize = arm_cpu->dcz_blocksize;
90
91
+#ifdef CONFIG_USER_ONLY
92
+ /* In sve_probe_page, we assume TBI is enabled. */
93
+ tcg_debug_assert(dc->tbid & 1);
94
+#endif
95
+
96
/* Single step state. The code-generation logic here is:
97
* SS_ACTIVE == 0:
98
* generate code with no special handling for single-stepping (except
41
--
99
--
42
2.20.1
100
2.20.1
43
101
44
102
diff view generated by jsdifflib
1
We are close to running out of TB flags for AArch32; we could
1
From: Richard Henderson <richard.henderson@linaro.org>
2
start using the cs_base word, but before we do that we can
3
economise on our usage by sharing the same bits for the VFP
4
VECSTRIDE field and the XScale XSCALE_CPAR field. This
5
works because no XScale CPU ever had VFP.
6
2
3
This is DC GVA and DC GZVA, and the tag check for DC ZVA.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-40-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20190416125744.27770-18-peter.maydell@linaro.org
10
---
9
---
11
target/arm/cpu.h | 10 ++++++----
10
target/arm/cpu.h | 4 +++-
12
target/arm/cpu.c | 7 +++++++
11
target/arm/helper.c | 16 ++++++++++++++++
13
target/arm/helper.c | 6 +++++-
12
target/arm/translate-a64.c | 39 ++++++++++++++++++++++++++++++++++++++
14
target/arm/translate.c | 9 +++++++--
13
3 files changed, 58 insertions(+), 1 deletion(-)
15
4 files changed, 25 insertions(+), 7 deletions(-)
16
14
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_ANY, BE_DATA, 23, 1)
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
22
FIELD(TBFLAG_A32, THUMB, 0, 1)
20
#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
23
FIELD(TBFLAG_A32, VECLEN, 1, 3)
21
#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
24
FIELD(TBFLAG_A32, VECSTRIDE, 4, 2)
22
#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
25
+/*
23
-#define ARM_LAST_SPECIAL ARM_CP_DC_ZVA
26
+ * We store the bottom two bits of the CPAR as TB flags and handle
24
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
27
+ * checks on the other bits at runtime. This shares the same bits as
25
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
28
+ * VECSTRIDE, which is OK as no XScale CPU has VFP.
26
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
29
+ */
27
#define ARM_CP_FPU 0x1000
30
+FIELD(TBFLAG_A32, XSCALE_CPAR, 4, 2)
28
#define ARM_CP_SVE 0x2000
31
/*
29
#define ARM_CP_NO_GDB 0x4000
32
* Indicates whether cp register reads and writes by guest code should access
33
* the secure or nonsecure bank of banked registers; note that this is not
34
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, NS, 6, 1)
35
FIELD(TBFLAG_A32, VFPEN, 7, 1)
36
FIELD(TBFLAG_A32, CONDEXEC, 8, 8)
37
FIELD(TBFLAG_A32, SCTLR_B, 16, 1)
38
-/* We store the bottom two bits of the CPAR as TB flags and handle
39
- * checks on the other bits at runtime
40
- */
41
-FIELD(TBFLAG_A32, XSCALE_CPAR, 17, 2)
42
/* For M profile only, Handler (ie not Thread) mode */
43
FIELD(TBFLAG_A32, HANDLER, 21, 1)
44
/* For M profile only, whether we should generate stack-limit checks */
45
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
46
index XXXXXXX..XXXXXXX 100644
47
--- a/target/arm/cpu.c
48
+++ b/target/arm/cpu.c
49
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
50
set_feature(env, ARM_FEATURE_THUMB_DSP);
51
}
52
53
+ /*
54
+ * We rely on no XScale CPU having VFP so we can use the same bits in the
55
+ * TB flags field for VECSTRIDE and XSCALE_CPAR.
56
+ */
57
+ assert(!(arm_feature(env, ARM_FEATURE_VFP) &&
58
+ arm_feature(env, ARM_FEATURE_XSCALE)));
59
+
60
if (arm_feature(env, ARM_FEATURE_V7) &&
61
!arm_feature(env, ARM_FEATURE_M) &&
62
!arm_feature(env, ARM_FEATURE_PMSA)) {
63
diff --git a/target/arm/helper.c b/target/arm/helper.c
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
64
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/helper.c
32
--- a/target/arm/helper.c
66
+++ b/target/arm/helper.c
33
+++ b/target/arm/helper.c
67
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
34
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
68
|| arm_el_is_aa64(env, 1) || arm_feature(env, ARM_FEATURE_M)) {
35
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
69
flags = FIELD_DP32(flags, TBFLAG_A32, VFPEN, 1);
36
.type = ARM_CP_NOP, .access = PL0_W,
37
.accessfn = aa64_cacheop_poc_access },
38
+ { .name = "DC_GVA", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 3,
40
+ .access = PL0_W, .type = ARM_CP_DC_GVA,
41
+#ifndef CONFIG_USER_ONLY
42
+ /* Avoid overhead of an access check that always passes in user-mode */
43
+ .accessfn = aa64_zva_access,
44
+#endif
45
+ },
46
+ { .name = "DC_GZVA", .state = ARM_CP_STATE_AA64,
47
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 4,
48
+ .access = PL0_W, .type = ARM_CP_DC_GZVA,
49
+#ifndef CONFIG_USER_ONLY
50
+ /* Avoid overhead of an access check that always passes in user-mode */
51
+ .accessfn = aa64_zva_access,
52
+#endif
53
+ },
54
REGINFO_SENTINEL
55
};
56
57
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
58
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/translate-a64.c
60
+++ b/target/arm/translate-a64.c
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
70
}
62
}
71
- flags = FIELD_DP32(flags, TBFLAG_A32, XSCALE_CPAR, env->cp15.c15_cpar);
63
gen_helper_dc_zva(cpu_env, tcg_rt);
72
+ /* Note that XSCALE_CPAR shares bits with VECSTRIDE */
64
return;
73
+ if (arm_feature(env, ARM_FEATURE_XSCALE)) {
65
+ case ARM_CP_DC_GVA:
74
+ flags = FIELD_DP32(flags, TBFLAG_A32,
66
+ {
75
+ XSCALE_CPAR, env->cp15.c15_cpar);
67
+ TCGv_i64 clean_addr, tag;
68
+
69
+ /*
70
+ * DC_GVA, like DC_ZVA, requires that we supply the original
71
+ * pointer for an invalid page. Probe that address first.
72
+ */
73
+ tcg_rt = cpu_reg(s, rt);
74
+ clean_addr = clean_data_tbi(s, tcg_rt);
75
+ gen_probe_access(s, clean_addr, MMU_DATA_STORE, MO_8);
76
+
77
+ if (s->ata) {
78
+ /* Extract the tag from the register to match STZGM. */
79
+ tag = tcg_temp_new_i64();
80
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
81
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
82
+ tcg_temp_free_i64(tag);
83
+ }
76
+ }
84
+ }
85
+ return;
86
+ case ARM_CP_DC_GZVA:
87
+ {
88
+ TCGv_i64 clean_addr, tag;
89
+
90
+ /* For DC_GZVA, we can rely on DC_ZVA for the proper fault. */
91
+ tcg_rt = cpu_reg(s, rt);
92
+ clean_addr = clean_data_tbi(s, tcg_rt);
93
+ gen_helper_dc_zva(cpu_env, clean_addr);
94
+
95
+ if (s->ata) {
96
+ /* Extract the tag from the register to match STZGM. */
97
+ tag = tcg_temp_new_i64();
98
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
99
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
100
+ tcg_temp_free_i64(tag);
101
+ }
102
+ }
103
+ return;
104
default:
105
break;
77
}
106
}
78
79
flags = FIELD_DP32(flags, TBFLAG_ANY, MMUIDX, arm_to_core_mmu_idx(mmu_idx));
80
diff --git a/target/arm/translate.c b/target/arm/translate.c
81
index XXXXXXX..XXXXXXX 100644
82
--- a/target/arm/translate.c
83
+++ b/target/arm/translate.c
84
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
85
dc->fp_excp_el = FIELD_EX32(tb_flags, TBFLAG_ANY, FPEXC_EL);
86
dc->vfp_enabled = FIELD_EX32(tb_flags, TBFLAG_A32, VFPEN);
87
dc->vec_len = FIELD_EX32(tb_flags, TBFLAG_A32, VECLEN);
88
- dc->vec_stride = FIELD_EX32(tb_flags, TBFLAG_A32, VECSTRIDE);
89
- dc->c15_cpar = FIELD_EX32(tb_flags, TBFLAG_A32, XSCALE_CPAR);
90
+ if (arm_feature(env, ARM_FEATURE_XSCALE)) {
91
+ dc->c15_cpar = FIELD_EX32(tb_flags, TBFLAG_A32, XSCALE_CPAR);
92
+ dc->vec_stride = 0;
93
+ } else {
94
+ dc->vec_stride = FIELD_EX32(tb_flags, TBFLAG_A32, VECSTRIDE);
95
+ dc->c15_cpar = 0;
96
+ }
97
dc->v7m_handler_mode = FIELD_EX32(tb_flags, TBFLAG_A32, HANDLER);
98
dc->v8m_secure = arm_feature(env, ARM_FEATURE_M_SECURITY) &&
99
regime_is_secure(env, dc->mmu_idx);
100
--
107
--
101
2.20.1
108
2.20.1
102
109
103
110
diff view generated by jsdifflib
1
Currently the code in v7m_push_stack() which detects a violation
1
From: Richard Henderson <richard.henderson@linaro.org>
2
of the v8M stack limit simply returns early if it does so. This
3
is OK for the current integer-only code, but won't work for the
4
floating point handling we're about to add. We need to continue
5
executing the rest of the function so that we check for other
6
exceptions like not having permission to use the FPU and so
7
that we correctly set the FPCCR state if we are doing lazy
8
stacking. Refactor to avoid the early return.
9
2
3
D1.10 specifies that exception handlers begin with tag checks overridden.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-41-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20190416125744.27770-10-peter.maydell@linaro.org
13
---
9
---
14
target/arm/helper.c | 23 ++++++++++++++++++-----
10
target/arm/helper.c | 3 +++
15
1 file changed, 18 insertions(+), 5 deletions(-)
11
1 file changed, 3 insertions(+)
16
12
17
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
18
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
20
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
21
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
22
* should ignore further stack faults trying to process
18
break;
23
* that derived exception.)
24
*/
25
- bool stacked_ok;
26
+ bool stacked_ok = true, limitviol = false;
27
CPUARMState *env = &cpu->env;
28
uint32_t xpsr = xpsr_read(env);
29
uint32_t frameptr = env->regs[13];
30
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
31
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE,
32
env->v7m.secure);
33
env->regs[13] = limit;
34
- return true;
35
+ /*
36
+ * We won't try to perform any further memory accesses but
37
+ * we must continue through the following code to check for
38
+ * permission faults during FPU state preservation, and we
39
+ * must update FPCCR if lazy stacking is enabled.
40
+ */
41
+ limitviol = true;
42
+ stacked_ok = false;
43
}
19
}
44
}
20
}
45
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
46
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
22
+ new_mode |= PSTATE_TCO;
47
* (which may be taken in preference to the one we started with
48
* if it has higher priority).
49
*/
50
- stacked_ok =
51
+ stacked_ok = stacked_ok &&
52
v7m_stack_write(cpu, frameptr, env->regs[0], mmu_idx, false) &&
53
v7m_stack_write(cpu, frameptr + 4, env->regs[1], mmu_idx, false) &&
54
v7m_stack_write(cpu, frameptr + 8, env->regs[2], mmu_idx, false) &&
55
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
56
v7m_stack_write(cpu, frameptr + 24, env->regs[15], mmu_idx, false) &&
57
v7m_stack_write(cpu, frameptr + 28, xpsr, mmu_idx, false);
58
59
- /* Update SP regardless of whether any of the stack accesses failed. */
60
- env->regs[13] = frameptr;
61
+ /*
62
+ * If we broke a stack limit then SP was already updated earlier;
63
+ * otherwise we update SP regardless of whether any of the stack
64
+ * accesses failed or we took some other kind of fault.
65
+ */
66
+ if (!limitviol) {
67
+ env->regs[13] = frameptr;
68
+ }
23
+ }
69
24
70
return !stacked_ok;
25
pstate_write(env, PSTATE_DAIF | new_mode);
71
}
26
env->aarch64 = 1;
72
--
27
--
73
2.20.1
28
2.20.1
74
29
75
30
diff view generated by jsdifflib
1
Pushing registers to the stack for v7M needs to handle three cases:
1
From: Richard Henderson <richard.henderson@linaro.org>
2
* the "normal" case where we pend exceptions
2
3
* an "ignore faults" case where we set FSR bits but
3
We need to check the memattr of a page in order to determine
4
do not pend exceptions (this is used when we are
4
whether it is Tagged for MTE. Between Stage1 and Stage2,
5
handling some kinds of derived exception on exception entry)
5
this becomes simpler if we always collect this data, instead
6
* a "lazy FP stacking" case, where different FSR bits
6
of occasionally being presented with NULL.
7
are set and the exception is pended differently
7
8
8
Use the nonnull attribute to allow the compiler to check that
9
Implement this by changing the existing flag argument that
9
all pointer arguments are non-null.
10
tells us whether to ignore faults or not into an enum that
10
11
specifies which of the 3 modes we should handle.
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200626033144.790098-42-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 20190416125744.27770-23-peter.maydell@linaro.org
16
---
15
---
17
target/arm/helper.c | 118 +++++++++++++++++++++++++++++---------------
16
target/arm/internals.h | 3 ++-
18
1 file changed, 79 insertions(+), 39 deletions(-)
17
target/arm/helper.c | 60 ++++++++++++++++++++---------------------
19
18
target/arm/m_helper.c | 11 +++++---
19
target/arm/tlb_helper.c | 4 ++-
20
4 files changed, 42 insertions(+), 36 deletions(-)
21
22
diff --git a/target/arm/internals.h b/target/arm/internals.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/internals.h
25
+++ b/target/arm/internals.h
26
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
27
MMUAccessType access_type, ARMMMUIdx mmu_idx,
28
hwaddr *phys_ptr, MemTxAttrs *attrs, int *prot,
29
target_ulong *page_size,
30
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
31
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
32
+ __attribute__((nonnull));
33
34
void arm_log_exception(int idx);
35
20
diff --git a/target/arm/helper.c b/target/arm/helper.c
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
21
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/helper.c
38
--- a/target/arm/helper.c
23
+++ b/target/arm/helper.c
39
+++ b/target/arm/helper.c
24
@@ -XXX,XX +XXX,XX @@ static bool v7m_cpacr_pass(CPUARMState *env, bool is_secure, bool is_priv)
40
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
41
bool s1_is_el0,
42
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
43
target_ulong *page_size_ptr,
44
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
45
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
46
+ __attribute__((nonnull));
47
#endif
48
49
static void switch_mode(CPUARMState *env, int mode);
50
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
51
arm_tlb_bti_gp(txattrs) = true;
25
}
52
}
26
}
53
27
54
- if (cacheattrs != NULL) {
28
+/*
55
- if (mmu_idx == ARMMMUIdx_Stage2) {
29
+ * What kind of stack write are we doing? This affects how exceptions
56
- cacheattrs->attrs = convert_stage2_attrs(env,
30
+ * generated during the stacking are treated.
57
- extract32(attrs, 0, 4));
31
+ */
58
- } else {
32
+typedef enum StackingMode {
59
- /* Index into MAIR registers for cache attributes */
33
+ STACK_NORMAL,
60
- uint8_t attrindx = extract32(attrs, 0, 3);
34
+ STACK_IGNFAULTS,
61
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
35
+ STACK_LAZYFP,
62
- assert(attrindx <= 7);
36
+} StackingMode;
63
- cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
37
+
64
- }
38
static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
65
- cacheattrs->shareability = extract32(attrs, 6, 2);
39
- ARMMMUIdx mmu_idx, bool ignfault)
66
+ if (mmu_idx == ARMMMUIdx_Stage2) {
40
+ ARMMMUIdx mmu_idx, StackingMode mode)
67
+ cacheattrs->attrs = convert_stage2_attrs(env, extract32(attrs, 0, 4));
41
{
68
+ } else {
42
CPUState *cs = CPU(cpu);
69
+ /* Index into MAIR registers for cache attributes */
43
CPUARMState *env = &cpu->env;
70
+ uint8_t attrindx = extract32(attrs, 0, 3);
71
+ uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
72
+ assert(attrindx <= 7);
73
+ cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
74
}
75
+ cacheattrs->shareability = extract32(attrs, 6, 2);
76
77
*phys_ptr = descaddr;
78
*page_size_ptr = page_size;
79
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
80
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
81
mmu_idx == ARMMMUIdx_E10_0,
82
phys_ptr, attrs, &s2_prot,
83
- page_size, fi,
84
- cacheattrs != NULL ? &cacheattrs2 : NULL);
85
+ page_size, fi, &cacheattrs2);
86
fi->s2addr = ipa;
87
/* Combine the S1 and S2 perms. */
88
*prot &= s2_prot;
89
90
- /* Combine the S1 and S2 cache attributes, if needed */
91
- if (!ret && cacheattrs != NULL) {
92
- if (env->cp15.hcr_el2 & HCR_DC) {
93
- /*
94
- * HCR.DC forces the first stage attributes to
95
- * Normal Non-Shareable,
96
- * Inner Write-Back Read-Allocate Write-Allocate,
97
- * Outer Write-Back Read-Allocate Write-Allocate.
98
- */
99
- cacheattrs->attrs = 0xff;
100
- cacheattrs->shareability = 0;
101
- }
102
- *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
103
+ /* If S2 fails, return early. */
104
+ if (ret) {
105
+ return ret;
106
}
107
108
- return ret;
109
+ /* Combine the S1 and S2 cache attributes. */
110
+ if (env->cp15.hcr_el2 & HCR_DC) {
111
+ /*
112
+ * HCR.DC forces the first stage attributes to
113
+ * Normal Non-Shareable,
114
+ * Inner Write-Back Read-Allocate Write-Allocate,
115
+ * Outer Write-Back Read-Allocate Write-Allocate.
116
+ */
117
+ cacheattrs->attrs = 0xff;
118
+ cacheattrs->shareability = 0;
119
+ }
120
+ *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
121
+ return 0;
122
} else {
123
/*
124
* For non-EL2 CPUs a stage1+stage2 translation is just stage 1.
125
@@ -XXX,XX +XXX,XX @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
126
bool ret;
127
ARMMMUFaultInfo fi = {};
128
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
129
+ ARMCacheAttrs cacheattrs = {};
130
131
*attrs = (MemTxAttrs) {};
132
133
ret = get_phys_addr(env, addr, 0, mmu_idx, &phys_addr,
134
- attrs, &prot, &page_size, &fi, NULL);
135
+ attrs, &prot, &page_size, &fi, &cacheattrs);
136
137
if (ret) {
138
return -1;
139
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
140
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/m_helper.c
142
+++ b/target/arm/m_helper.c
44
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
143
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
45
&attrs, &prot, &page_size, &fi, NULL)) {
144
hwaddr physaddr;
145
int prot;
146
ARMMMUFaultInfo fi = {};
147
+ ARMCacheAttrs cacheattrs = {};
148
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
149
int exc;
150
bool exc_secure;
151
152
if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &physaddr,
153
- &attrs, &prot, &page_size, &fi, NULL)) {
154
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
46
/* MPU/SAU lookup failed */
155
/* MPU/SAU lookup failed */
47
if (fi.type == ARMFault_QEMU_SFault) {
156
if (fi.type == ARMFault_QEMU_SFault) {
48
- qemu_log_mask(CPU_LOG_INT,
157
if (mode == STACK_LAZYFP) {
49
- "...SecureFault with SFSR.AUVIOL during stacking\n");
158
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
50
- env->v7m.sfsr |= R_V7M_SFSR_AUVIOL_MASK | R_V7M_SFSR_SFARVALID_MASK;
159
hwaddr physaddr;
51
+ if (mode == STACK_LAZYFP) {
160
int prot;
52
+ qemu_log_mask(CPU_LOG_INT,
161
ARMMMUFaultInfo fi = {};
53
+ "...SecureFault with SFSR.LSPERR "
162
+ ARMCacheAttrs cacheattrs = {};
54
+ "during lazy stacking\n");
163
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
55
+ env->v7m.sfsr |= R_V7M_SFSR_LSPERR_MASK;
164
int exc;
56
+ } else {
165
bool exc_secure;
57
+ qemu_log_mask(CPU_LOG_INT,
166
uint32_t value;
58
+ "...SecureFault with SFSR.AUVIOL "
167
59
+ "during stacking\n");
168
if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
60
+ env->v7m.sfsr |= R_V7M_SFSR_AUVIOL_MASK;
169
- &attrs, &prot, &page_size, &fi, NULL)) {
61
+ }
170
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
62
+ env->v7m.sfsr |= R_V7M_SFSR_SFARVALID_MASK;
171
/* MPU/SAU lookup failed */
63
env->v7m.sfar = addr;
172
if (fi.type == ARMFault_QEMU_SFault) {
64
exc = ARMV7M_EXCP_SECURE;
173
qemu_log_mask(CPU_LOG_INT,
65
exc_secure = false;
174
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
66
} else {
175
V8M_SAttributes sattrs = {};
67
- qemu_log_mask(CPU_LOG_INT, "...MemManageFault with CFSR.MSTKERR\n");
176
MemTxAttrs attrs = {};
68
- env->v7m.cfsr[secure] |= R_V7M_CFSR_MSTKERR_MASK;
177
ARMMMUFaultInfo fi = {};
69
+ if (mode == STACK_LAZYFP) {
178
+ ARMCacheAttrs cacheattrs = {};
70
+ qemu_log_mask(CPU_LOG_INT,
179
MemTxResult txres;
71
+ "...MemManageFault with CFSR.MLSPERR\n");
180
target_ulong page_size;
72
+ env->v7m.cfsr[secure] |= R_V7M_CFSR_MLSPERR_MASK;
181
hwaddr physaddr;
73
+ } else {
182
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
74
+ qemu_log_mask(CPU_LOG_INT,
183
"...really SecureFault with SFSR.INVEP\n");
75
+ "...MemManageFault with CFSR.MSTKERR\n");
184
return false;
76
+ env->v7m.cfsr[secure] |= R_V7M_CFSR_MSTKERR_MASK;
185
}
77
+ }
186
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx,
78
exc = ARMV7M_EXCP_MEM;
187
- &physaddr, &attrs, &prot, &page_size, &fi, NULL)) {
79
exc_secure = secure;
188
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &physaddr,
80
}
189
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
81
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
190
/* the MPU lookup failed */
82
attrs, &txres);
191
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
83
if (txres != MEMTX_OK) {
192
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
84
/* BusFault trying to write the data */
193
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
85
- qemu_log_mask(CPU_LOG_INT, "...BusFault with BFSR.STKERR\n");
194
index XXXXXXX..XXXXXXX 100644
86
- env->v7m.cfsr[M_REG_NS] |= R_V7M_CFSR_STKERR_MASK;
195
--- a/target/arm/tlb_helper.c
87
+ if (mode == STACK_LAZYFP) {
196
+++ b/target/arm/tlb_helper.c
88
+ qemu_log_mask(CPU_LOG_INT, "...BusFault with BFSR.LSPERR\n");
197
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
89
+ env->v7m.cfsr[M_REG_NS] |= R_V7M_CFSR_LSPERR_MASK;
198
int prot, ret;
90
+ } else {
199
MemTxAttrs attrs = {};
91
+ qemu_log_mask(CPU_LOG_INT, "...BusFault with BFSR.STKERR\n");
200
ARMMMUFaultInfo fi = {};
92
+ env->v7m.cfsr[M_REG_NS] |= R_V7M_CFSR_STKERR_MASK;
201
+ ARMCacheAttrs cacheattrs = {};
93
+ }
202
94
exc = ARMV7M_EXCP_BUS;
203
/*
95
exc_secure = false;
204
* Walk the page table and (if the mapping exists) add the page
96
goto pend_fault;
205
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
97
@@ -XXX,XX +XXX,XX @@ pend_fault:
98
* later if we have two derived exceptions.
99
* The only case when we must not pend the exception but instead
100
* throw it away is if we are doing the push of the callee registers
101
- * and we've already generated a derived exception. Even in this
102
- * case we will still update the fault status registers.
103
+ * and we've already generated a derived exception (this is indicated
104
+ * by the caller passing STACK_IGNFAULTS). Even in this case we will
105
+ * still update the fault status registers.
106
*/
206
*/
107
- if (!ignfault) {
207
ret = get_phys_addr(&cpu->env, address, access_type,
108
+ switch (mode) {
208
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
109
+ case STACK_NORMAL:
209
- &phys_addr, &attrs, &prot, &page_size, &fi, NULL);
110
armv7m_nvic_set_pending_derived(env->nvic, exc, exc_secure);
210
+ &phys_addr, &attrs, &prot, &page_size,
111
+ break;
211
+ &fi, &cacheattrs);
112
+ case STACK_LAZYFP:
212
if (likely(!ret)) {
113
+ armv7m_nvic_set_pending_lazyfp(env->nvic, exc, exc_secure);
213
/*
114
+ break;
214
* Map a single [sub]page. Regions smaller than our declared
115
+ case STACK_IGNFAULTS:
116
+ break;
117
}
118
return false;
119
}
120
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_callee_stack(ARMCPU *cpu, uint32_t lr, bool dotailchain,
121
uint32_t limit;
122
bool want_psp;
123
uint32_t sig;
124
+ StackingMode smode = ignore_faults ? STACK_IGNFAULTS : STACK_NORMAL;
125
126
if (dotailchain) {
127
bool mode = lr & R_V7M_EXCRET_MODE_MASK;
128
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_callee_stack(ARMCPU *cpu, uint32_t lr, bool dotailchain,
129
*/
130
sig = v7m_integrity_sig(env, lr);
131
stacked_ok =
132
- v7m_stack_write(cpu, frameptr, sig, mmu_idx, ignore_faults) &&
133
- v7m_stack_write(cpu, frameptr + 0x8, env->regs[4], mmu_idx,
134
- ignore_faults) &&
135
- v7m_stack_write(cpu, frameptr + 0xc, env->regs[5], mmu_idx,
136
- ignore_faults) &&
137
- v7m_stack_write(cpu, frameptr + 0x10, env->regs[6], mmu_idx,
138
- ignore_faults) &&
139
- v7m_stack_write(cpu, frameptr + 0x14, env->regs[7], mmu_idx,
140
- ignore_faults) &&
141
- v7m_stack_write(cpu, frameptr + 0x18, env->regs[8], mmu_idx,
142
- ignore_faults) &&
143
- v7m_stack_write(cpu, frameptr + 0x1c, env->regs[9], mmu_idx,
144
- ignore_faults) &&
145
- v7m_stack_write(cpu, frameptr + 0x20, env->regs[10], mmu_idx,
146
- ignore_faults) &&
147
- v7m_stack_write(cpu, frameptr + 0x24, env->regs[11], mmu_idx,
148
- ignore_faults);
149
+ v7m_stack_write(cpu, frameptr, sig, mmu_idx, smode) &&
150
+ v7m_stack_write(cpu, frameptr + 0x8, env->regs[4], mmu_idx, smode) &&
151
+ v7m_stack_write(cpu, frameptr + 0xc, env->regs[5], mmu_idx, smode) &&
152
+ v7m_stack_write(cpu, frameptr + 0x10, env->regs[6], mmu_idx, smode) &&
153
+ v7m_stack_write(cpu, frameptr + 0x14, env->regs[7], mmu_idx, smode) &&
154
+ v7m_stack_write(cpu, frameptr + 0x18, env->regs[8], mmu_idx, smode) &&
155
+ v7m_stack_write(cpu, frameptr + 0x1c, env->regs[9], mmu_idx, smode) &&
156
+ v7m_stack_write(cpu, frameptr + 0x20, env->regs[10], mmu_idx, smode) &&
157
+ v7m_stack_write(cpu, frameptr + 0x24, env->regs[11], mmu_idx, smode);
158
159
/* Update SP regardless of whether any of the stack accesses failed. */
160
*frame_sp_p = frameptr;
161
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
162
* if it has higher priority).
163
*/
164
stacked_ok = stacked_ok &&
165
- v7m_stack_write(cpu, frameptr, env->regs[0], mmu_idx, false) &&
166
- v7m_stack_write(cpu, frameptr + 4, env->regs[1], mmu_idx, false) &&
167
- v7m_stack_write(cpu, frameptr + 8, env->regs[2], mmu_idx, false) &&
168
- v7m_stack_write(cpu, frameptr + 12, env->regs[3], mmu_idx, false) &&
169
- v7m_stack_write(cpu, frameptr + 16, env->regs[12], mmu_idx, false) &&
170
- v7m_stack_write(cpu, frameptr + 20, env->regs[14], mmu_idx, false) &&
171
- v7m_stack_write(cpu, frameptr + 24, env->regs[15], mmu_idx, false) &&
172
- v7m_stack_write(cpu, frameptr + 28, xpsr, mmu_idx, false);
173
+ v7m_stack_write(cpu, frameptr, env->regs[0], mmu_idx, STACK_NORMAL) &&
174
+ v7m_stack_write(cpu, frameptr + 4, env->regs[1],
175
+ mmu_idx, STACK_NORMAL) &&
176
+ v7m_stack_write(cpu, frameptr + 8, env->regs[2],
177
+ mmu_idx, STACK_NORMAL) &&
178
+ v7m_stack_write(cpu, frameptr + 12, env->regs[3],
179
+ mmu_idx, STACK_NORMAL) &&
180
+ v7m_stack_write(cpu, frameptr + 16, env->regs[12],
181
+ mmu_idx, STACK_NORMAL) &&
182
+ v7m_stack_write(cpu, frameptr + 20, env->regs[14],
183
+ mmu_idx, STACK_NORMAL) &&
184
+ v7m_stack_write(cpu, frameptr + 24, env->regs[15],
185
+ mmu_idx, STACK_NORMAL) &&
186
+ v7m_stack_write(cpu, frameptr + 28, xpsr, mmu_idx, STACK_NORMAL);
187
188
if (env->v7m.control[M_REG_S] & R_V7M_CONTROL_FPCA_MASK) {
189
/* FPU is active, try to save its registers */
190
@@ -XXX,XX +XXX,XX @@ static bool v7m_push_stack(ARMCPU *cpu)
191
faddr += 8; /* skip the slot for the FPSCR */
192
}
193
stacked_ok = stacked_ok &&
194
- v7m_stack_write(cpu, faddr, slo, mmu_idx, false) &&
195
- v7m_stack_write(cpu, faddr + 4, shi, mmu_idx, false);
196
+ v7m_stack_write(cpu, faddr, slo,
197
+ mmu_idx, STACK_NORMAL) &&
198
+ v7m_stack_write(cpu, faddr + 4, shi,
199
+ mmu_idx, STACK_NORMAL);
200
}
201
stacked_ok = stacked_ok &&
202
v7m_stack_write(cpu, frameptr + 0x60,
203
- vfp_get_fpscr(env), mmu_idx, false);
204
+ vfp_get_fpscr(env), mmu_idx, STACK_NORMAL);
205
if (cpacr_pass) {
206
for (i = 0; i < ((framesize == 0xa8) ? 32 : 16); i += 2) {
207
*aa32_vfp_dreg(env, i / 2) = 0;
208
--
215
--
209
2.20.1
216
2.20.1
210
217
211
218
diff view generated by jsdifflib
1
Like AArch64, M-profile floating point has no FPEXC enable
1
From: Richard Henderson <richard.henderson@linaro.org>
2
bit to gate floating point; so always set the VFPEN TB flag.
3
2
4
M-profile also has CPACR and NSACR similar to A-profile;
3
This "bit" is a particular value of the page's MemAttr.
5
they behave slightly differently:
6
* the CPACR is banked between Secure and Non-Secure
7
* if the NSACR forces a trap then this is taken to
8
the Secure state, not the Non-Secure state
9
4
10
Honour the CPACR and NSACR settings. The NSACR handling
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
requires us to borrow the exception.target_el field
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
(usually meaningless for M profile) to distinguish the
7
Message-id: 20200626033144.790098-43-richard.henderson@linaro.org
13
NOCP UsageFault taken to Secure state from the more
14
usual fault taken to the current security state.
15
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20190416125744.27770-6-peter.maydell@linaro.org
19
---
9
---
20
target/arm/helper.c | 55 +++++++++++++++++++++++++++++++++++++++---
10
target/arm/helper.c | 48 ++++++++++++++++++++++++++++++++++++++---
21
target/arm/translate.c | 10 ++++++--
11
target/arm/tlb_helper.c | 5 +++++
22
2 files changed, 60 insertions(+), 5 deletions(-)
12
2 files changed, 50 insertions(+), 3 deletions(-)
23
13
24
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
25
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
27
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
28
@@ -XXX,XX +XXX,XX @@ uint32_t arm_phys_excp_target_el(CPUState *cs, uint32_t excp_idx,
18
@@ -XXX,XX +XXX,XX @@ static uint8_t combine_cacheattr_nibble(uint8_t s1, uint8_t s2)
29
return target_el;
19
*/
20
static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
21
{
22
- uint8_t s1lo = extract32(s1.attrs, 0, 4), s2lo = extract32(s2.attrs, 0, 4);
23
- uint8_t s1hi = extract32(s1.attrs, 4, 4), s2hi = extract32(s2.attrs, 4, 4);
24
+ uint8_t s1lo, s2lo, s1hi, s2hi;
25
ARMCacheAttrs ret;
26
+ bool tagged = false;
27
+
28
+ if (s1.attrs == 0xf0) {
29
+ tagged = true;
30
+ s1.attrs = 0xff;
31
+ }
32
+
33
+ s1lo = extract32(s1.attrs, 0, 4);
34
+ s2lo = extract32(s2.attrs, 0, 4);
35
+ s1hi = extract32(s1.attrs, 4, 4);
36
+ s2hi = extract32(s2.attrs, 4, 4);
37
38
/* Combine shareability attributes (table D4-43) */
39
if (s1.shareability == 2 || s2.shareability == 2) {
40
@@ -XXX,XX +XXX,XX @@ static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
41
}
42
}
43
44
+ /* TODO: CombineS1S2Desc does not consider transient, only WB, RWA. */
45
+ if (tagged && ret.attrs == 0xff) {
46
+ ret.attrs = 0xf0;
47
+ }
48
+
49
return ret;
30
}
50
}
31
51
32
+/*
52
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
33
+ * Return true if the v7M CPACR permits access to the FPU for the specified
53
* Normal Non-Shareable,
34
+ * security state and privilege level.
54
* Inner Write-Back Read-Allocate Write-Allocate,
35
+ */
55
* Outer Write-Back Read-Allocate Write-Allocate.
36
+static bool v7m_cpacr_pass(CPUARMState *env, bool is_secure, bool is_priv)
56
+ * Do not overwrite Tagged within attrs.
37
+{
57
*/
38
+ switch (extract32(env->v7m.cpacr[is_secure], 20, 2)) {
58
- cacheattrs->attrs = 0xff;
39
+ case 0:
59
+ if (cacheattrs->attrs != 0xf0) {
40
+ case 2: /* UNPREDICTABLE: we treat like 0 */
60
+ cacheattrs->attrs = 0xff;
41
+ return false;
61
+ }
42
+ case 1:
62
cacheattrs->shareability = 0;
43
+ return is_priv;
63
}
44
+ case 3:
64
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
45
+ return true;
65
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
46
+ default:
66
/* Definitely a real MMU, not an MPU */
47
+ g_assert_not_reached();
67
48
+ }
68
if (regime_translation_disabled(env, mmu_idx)) {
49
+}
69
+ uint64_t hcr;
70
+ uint8_t memattr;
50
+
71
+
51
static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
72
/*
52
ARMMMUIdx mmu_idx, bool ignfault)
73
* MMU disabled. S1 addresses within aa64 translation regimes are
53
{
74
* still checked for bounds -- see AArch64.TranslateAddressS1Off.
54
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
75
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
55
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_UNDEFINSTR_MASK;
76
*phys_ptr = address;
56
break;
77
*prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
57
case EXCP_NOCP:
78
*page_size = TARGET_PAGE_SIZE;
58
- armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, env->v7m.secure);
59
- env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_NOCP_MASK;
60
+ {
61
+ /*
62
+ * NOCP might be directed to something other than the current
63
+ * security state if this fault is because of NSACR; we indicate
64
+ * the target security state using exception.target_el.
65
+ */
66
+ int target_secstate;
67
+
79
+
68
+ if (env->exception.target_el == 3) {
80
+ /* Fill in cacheattr a-la AArch64.TranslateAddressS1Off. */
69
+ target_secstate = M_REG_S;
81
+ hcr = arm_hcr_el2_eff(env);
82
+ cacheattrs->shareability = 0;
83
+ if (hcr & HCR_DC) {
84
+ if (hcr & HCR_DCT) {
85
+ memattr = 0xf0; /* Tagged, Normal, WB, RWA */
86
+ } else {
87
+ memattr = 0xff; /* Normal, WB, RWA */
88
+ }
89
+ } else if (access_type == MMU_INST_FETCH) {
90
+ if (regime_sctlr(env, mmu_idx) & SCTLR_I) {
91
+ memattr = 0xee; /* Normal, WT, RA, NT */
92
+ } else {
93
+ memattr = 0x44; /* Normal, NC, No */
94
+ }
95
+ cacheattrs->shareability = 2; /* outer sharable */
70
+ } else {
96
+ } else {
71
+ target_secstate = env->v7m.secure;
97
+ memattr = 0x00; /* Device, nGnRnE */
72
+ }
98
+ }
73
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, target_secstate);
99
+ cacheattrs->attrs = memattr;
74
+ env->v7m.cfsr[target_secstate] |= R_V7M_CFSR_NOCP_MASK;
75
break;
76
+ }
77
case EXCP_INVSTATE:
78
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, env->v7m.secure);
79
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_INVSTATE_MASK;
80
@@ -XXX,XX +XXX,XX @@ int fp_exception_el(CPUARMState *env, int cur_el)
81
return 0;
100
return 0;
82
}
101
}
83
102
84
+ if (arm_feature(env, ARM_FEATURE_M)) {
103
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
85
+ /* CPACR can cause a NOCP UsageFault taken to current security state */
104
index XXXXXXX..XXXXXXX 100644
86
+ if (!v7m_cpacr_pass(env, env->v7m.secure, cur_el != 0)) {
105
--- a/target/arm/tlb_helper.c
87
+ return 1;
106
+++ b/target/arm/tlb_helper.c
107
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
108
phys_addr &= TARGET_PAGE_MASK;
109
address &= TARGET_PAGE_MASK;
110
}
111
+ /* Notice and record tagged memory. */
112
+ if (cpu_isar_feature(aa64_mte, cpu) && cacheattrs.attrs == 0xf0) {
113
+ arm_tlb_mte_tagged(&attrs) = true;
88
+ }
114
+ }
89
+
115
+
90
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY) && !env->v7m.secure) {
116
tlb_set_page_with_attrs(cs, address, phys_addr, attrs,
91
+ if (!extract32(env->v7m.nsacr, 10, 1)) {
117
prot, mmu_idx, page_size);
92
+ /* FP insns cause a NOCP UsageFault taken to Secure */
118
return true;
93
+ return 3;
94
+ }
95
+ }
96
+
97
+ return 0;
98
+ }
99
+
100
/* The CPACR controls traps to EL1, or PL1 if we're 32 bit:
101
* 0, 2 : trap EL0 and EL1/PL1 accesses
102
* 1 : trap only EL0 accesses
103
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
104
flags = FIELD_DP32(flags, TBFLAG_A32, SCTLR_B, arm_sctlr_b(env));
105
flags = FIELD_DP32(flags, TBFLAG_A32, NS, !access_secure_reg(env));
106
if (env->vfp.xregs[ARM_VFP_FPEXC] & (1 << 30)
107
- || arm_el_is_aa64(env, 1)) {
108
+ || arm_el_is_aa64(env, 1) || arm_feature(env, ARM_FEATURE_M)) {
109
flags = FIELD_DP32(flags, TBFLAG_A32, VFPEN, 1);
110
}
111
flags = FIELD_DP32(flags, TBFLAG_A32, XSCALE_CPAR, env->cp15.c15_cpar);
112
diff --git a/target/arm/translate.c b/target/arm/translate.c
113
index XXXXXXX..XXXXXXX 100644
114
--- a/target/arm/translate.c
115
+++ b/target/arm/translate.c
116
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
117
* for attempts to execute invalid vfp/neon encodings with FP disabled.
118
*/
119
if (s->fp_excp_el) {
120
- gen_exception_insn(s, 4, EXCP_UDEF,
121
- syn_fp_access_trap(1, 0xe, false), s->fp_excp_el);
122
+ if (arm_dc_feature(s, ARM_FEATURE_M)) {
123
+ gen_exception_insn(s, 4, EXCP_NOCP, syn_uncategorized(),
124
+ s->fp_excp_el);
125
+ } else {
126
+ gen_exception_insn(s, 4, EXCP_UDEF,
127
+ syn_fp_access_trap(1, 0xe, false),
128
+ s->fp_excp_el);
129
+ }
130
return 0;
131
}
132
133
--
119
--
134
2.20.1
120
2.20.1
135
121
136
122
diff view generated by jsdifflib
1
The M-profile floating point support has three associated config
1
From: Richard Henderson <richard.henderson@linaro.org>
2
registers: FPCAR, FPCCR and FPDSCR. It also makes the registers
2
3
CPACR and NSACR have behaviour other than reads-as-zero.
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Add support for all of these as simple reads-as-written registers.
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
We will hook up actual functionality later.
5
Message-id: 20200626033144.790098-44-richard.henderson@linaro.org
6
7
The main complexity here is handling the FPCCR register, which
8
has a mix of banked and unbanked bits.
9
10
Note that we don't share storage with the A-profile
11
cpu->cp15.nsacr and cpu->cp15.cpacr_el1, though the behaviour
12
is quite similar, for two reasons:
13
* the M profile CPACR is banked between security states
14
* it preserves the invariant that M profile uses no state
15
inside the cp15 substruct
16
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
19
Message-id: 20190416125744.27770-4-peter.maydell@linaro.org
20
---
7
---
21
target/arm/cpu.h | 34 ++++++++++++
8
target/arm/cpu.h | 6 ++++++
22
hw/intc/armv7m_nvic.c | 125 ++++++++++++++++++++++++++++++++++++++++++
9
hw/arm/virt.c | 55 ++++++++++++++++++++++++++++++++++++++++++++++--
23
target/arm/cpu.c | 5 ++
10
target/arm/cpu.c | 52 +++++++++++++++++++++++++++++++++++++++++----
24
target/arm/machine.c | 16 ++++++
11
3 files changed, 107 insertions(+), 6 deletions(-)
25
4 files changed, 180 insertions(+)
26
12
27
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
28
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/cpu.h
15
--- a/target/arm/cpu.h
30
+++ b/target/arm/cpu.h
16
+++ b/target/arm/cpu.h
31
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
17
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
32
uint32_t scr[M_REG_NUM_BANKS];
18
/* MemoryRegion to use for secure physical accesses */
33
uint32_t msplim[M_REG_NUM_BANKS];
19
MemoryRegion *secure_memory;
34
uint32_t psplim[M_REG_NUM_BANKS];
20
35
+ uint32_t fpcar[M_REG_NUM_BANKS];
21
+ /* MemoryRegion to use for allocation tag accesses */
36
+ uint32_t fpccr[M_REG_NUM_BANKS];
22
+ MemoryRegion *tag_memory;
37
+ uint32_t fpdscr[M_REG_NUM_BANKS];
23
+ MemoryRegion *secure_tag_memory;
38
+ uint32_t cpacr[M_REG_NUM_BANKS];
24
+
39
+ uint32_t nsacr;
25
/* For v8M, pointer to the IDAU interface provided by board/SoC */
40
} v7m;
26
Object *idau;
41
27
42
/* Information associated with an exception about to be taken:
28
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
43
@@ -XXX,XX +XXX,XX @@ FIELD(V7M_CSSELR, LEVEL, 1, 3)
29
typedef enum ARMASIdx {
44
*/
30
ARMASIdx_NS = 0,
45
FIELD(V7M_CSSELR, INDEX, 0, 4)
31
ARMASIdx_S = 1,
46
32
+ ARMASIdx_TagNS = 2,
47
+/* v7M FPCCR bits */
33
+ ARMASIdx_TagS = 3,
48
+FIELD(V7M_FPCCR, LSPACT, 0, 1)
34
} ARMASIdx;
49
+FIELD(V7M_FPCCR, USER, 1, 1)
35
50
+FIELD(V7M_FPCCR, S, 2, 1)
36
/* Return the Exception Level targeted by debug exceptions. */
51
+FIELD(V7M_FPCCR, THREAD, 3, 1)
37
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
52
+FIELD(V7M_FPCCR, HFRDY, 4, 1)
53
+FIELD(V7M_FPCCR, MMRDY, 5, 1)
54
+FIELD(V7M_FPCCR, BFRDY, 6, 1)
55
+FIELD(V7M_FPCCR, SFRDY, 7, 1)
56
+FIELD(V7M_FPCCR, MONRDY, 8, 1)
57
+FIELD(V7M_FPCCR, SPLIMVIOL, 9, 1)
58
+FIELD(V7M_FPCCR, UFRDY, 10, 1)
59
+FIELD(V7M_FPCCR, RES0, 11, 15)
60
+FIELD(V7M_FPCCR, TS, 26, 1)
61
+FIELD(V7M_FPCCR, CLRONRETS, 27, 1)
62
+FIELD(V7M_FPCCR, CLRONRET, 28, 1)
63
+FIELD(V7M_FPCCR, LSPENS, 29, 1)
64
+FIELD(V7M_FPCCR, LSPEN, 30, 1)
65
+FIELD(V7M_FPCCR, ASPEN, 31, 1)
66
+/* These bits are banked. Others are non-banked and live in the M_REG_S bank */
67
+#define R_V7M_FPCCR_BANKED_MASK \
68
+ (R_V7M_FPCCR_LSPACT_MASK | \
69
+ R_V7M_FPCCR_USER_MASK | \
70
+ R_V7M_FPCCR_THREAD_MASK | \
71
+ R_V7M_FPCCR_MMRDY_MASK | \
72
+ R_V7M_FPCCR_SPLIMVIOL_MASK | \
73
+ R_V7M_FPCCR_UFRDY_MASK | \
74
+ R_V7M_FPCCR_ASPEN_MASK)
75
+
76
/*
77
* System register ID fields.
78
*/
79
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
80
index XXXXXXX..XXXXXXX 100644
38
index XXXXXXX..XXXXXXX 100644
81
--- a/hw/intc/armv7m_nvic.c
39
--- a/hw/arm/virt.c
82
+++ b/hw/intc/armv7m_nvic.c
40
+++ b/hw/arm/virt.c
83
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
41
@@ -XXX,XX +XXX,XX @@ static void create_platform_bus(VirtMachineState *vms)
84
}
42
sysbus_mmio_get_region(s, 0));
85
case 0xd84: /* CSSELR */
43
}
86
return cpu->env.v7m.csselr[attrs.secure];
44
87
+ case 0xd88: /* CPACR */
45
+static void create_tag_ram(MemoryRegion *tag_sysmem,
88
+ if (!arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
46
+ hwaddr base, hwaddr size,
89
+ return 0;
47
+ const char *name)
48
+{
49
+ MemoryRegion *tagram = g_new(MemoryRegion, 1);
50
+
51
+ memory_region_init_ram(tagram, NULL, name, size / 32, &error_fatal);
52
+ memory_region_add_subregion(tag_sysmem, base / 32, tagram);
53
+}
54
+
55
static void create_secure_ram(VirtMachineState *vms,
56
- MemoryRegion *secure_sysmem)
57
+ MemoryRegion *secure_sysmem,
58
+ MemoryRegion *secure_tag_sysmem)
59
{
60
MemoryRegion *secram = g_new(MemoryRegion, 1);
61
char *nodename;
62
@@ -XXX,XX +XXX,XX @@ static void create_secure_ram(VirtMachineState *vms,
63
qemu_fdt_setprop_string(vms->fdt, nodename, "status", "disabled");
64
qemu_fdt_setprop_string(vms->fdt, nodename, "secure-status", "okay");
65
66
+ if (secure_tag_sysmem) {
67
+ create_tag_ram(secure_tag_sysmem, base, size, "mach-virt.secure-tag");
68
+ }
69
+
70
g_free(nodename);
71
}
72
73
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
74
const CPUArchIdList *possible_cpus;
75
MemoryRegion *sysmem = get_system_memory();
76
MemoryRegion *secure_sysmem = NULL;
77
+ MemoryRegion *tag_sysmem = NULL;
78
+ MemoryRegion *secure_tag_sysmem = NULL;
79
int n, virt_max_cpus;
80
bool firmware_loaded;
81
bool aarch64 = true;
82
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
83
"secure-memory", &error_abort);
84
}
85
86
+ /*
87
+ * The cpu adds the property if and only if MemTag is supported.
88
+ * If it is, we must allocate the ram to back that up.
89
+ */
90
+ if (object_property_find(cpuobj, "tag-memory", NULL)) {
91
+ if (!tag_sysmem) {
92
+ tag_sysmem = g_new(MemoryRegion, 1);
93
+ memory_region_init(tag_sysmem, OBJECT(machine),
94
+ "tag-memory", UINT64_MAX / 32);
95
+
96
+ if (vms->secure) {
97
+ secure_tag_sysmem = g_new(MemoryRegion, 1);
98
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
99
+ "secure-tag-memory", UINT64_MAX / 32);
100
+
101
+ /* As with ram, secure-tag takes precedence over tag. */
102
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
103
+ tag_sysmem, -1);
104
+ }
105
+ }
106
+
107
+ object_property_set_link(cpuobj, OBJECT(tag_sysmem),
108
+ "tag-memory", &error_abort);
109
+ if (vms->secure) {
110
+ object_property_set_link(cpuobj, OBJECT(secure_tag_sysmem),
111
+ "secure-tag-memory", &error_abort);
112
+ }
90
+ }
113
+ }
91
+ return cpu->env.v7m.cpacr[attrs.secure];
114
+
92
+ case 0xd8c: /* NSACR */
115
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
93
+ if (!attrs.secure || !arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
116
object_unref(cpuobj);
94
+ return 0;
117
}
95
+ }
118
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
96
+ return cpu->env.v7m.nsacr;
119
create_uart(vms, VIRT_UART, sysmem, serial_hd(0));
97
/* TODO: Implement debug registers. */
120
98
case 0xd90: /* MPU_TYPE */
121
if (vms->secure) {
99
/* Unified MPU; if the MPU is not present this value is zero */
122
- create_secure_ram(vms, secure_sysmem);
100
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
123
+ create_secure_ram(vms, secure_sysmem, secure_tag_sysmem);
101
return 0;
124
create_uart(vms, VIRT_SECURE_UART, secure_sysmem, serial_hd(1));
102
}
125
}
103
return cpu->env.v7m.sfar;
126
104
+ case 0xf34: /* FPCCR */
127
+ if (tag_sysmem) {
105
+ if (!arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
128
+ create_tag_ram(tag_sysmem, vms->memmap[VIRT_MEM].base,
106
+ return 0;
129
+ machine->ram_size, "mach-virt.tag");
107
+ }
130
+ }
108
+ if (attrs.secure) {
131
+
109
+ return cpu->env.v7m.fpccr[M_REG_S];
132
vms->highmem_ecam &= vms->highmem && (!firmware_loaded || aarch64);
110
+ } else {
133
111
+ /*
134
create_rtc(vms);
112
+ * NS can read LSPEN, CLRONRET and MONRDY. It can read
113
+ * BFRDY and HFRDY if AIRCR.BFHFNMINS != 0;
114
+ * other non-banked bits RAZ.
115
+ * TODO: MONRDY should RAZ/WI if DEMCR.SDME is set.
116
+ */
117
+ uint32_t value = cpu->env.v7m.fpccr[M_REG_S];
118
+ uint32_t mask = R_V7M_FPCCR_LSPEN_MASK |
119
+ R_V7M_FPCCR_CLRONRET_MASK |
120
+ R_V7M_FPCCR_MONRDY_MASK;
121
+
122
+ if (s->cpu->env.v7m.aircr & R_V7M_AIRCR_BFHFNMINS_MASK) {
123
+ mask |= R_V7M_FPCCR_BFRDY_MASK | R_V7M_FPCCR_HFRDY_MASK;
124
+ }
125
+
126
+ value &= mask;
127
+
128
+ value |= cpu->env.v7m.fpccr[M_REG_NS];
129
+ return value;
130
+ }
131
+ case 0xf38: /* FPCAR */
132
+ if (!arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
133
+ return 0;
134
+ }
135
+ return cpu->env.v7m.fpcar[attrs.secure];
136
+ case 0xf3c: /* FPDSCR */
137
+ if (!arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
138
+ return 0;
139
+ }
140
+ return cpu->env.v7m.fpdscr[attrs.secure];
141
case 0xf40: /* MVFR0 */
142
return cpu->isar.mvfr0;
143
case 0xf44: /* MVFR1 */
144
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
145
cpu->env.v7m.csselr[attrs.secure] = value & R_V7M_CSSELR_INDEX_MASK;
146
}
147
break;
148
+ case 0xd88: /* CPACR */
149
+ if (arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
150
+ /* We implement only the Floating Point extension's CP10/CP11 */
151
+ cpu->env.v7m.cpacr[attrs.secure] = value & (0xf << 20);
152
+ }
153
+ break;
154
+ case 0xd8c: /* NSACR */
155
+ if (attrs.secure && arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
156
+ /* We implement only the Floating Point extension's CP10/CP11 */
157
+ cpu->env.v7m.nsacr = value & (3 << 10);
158
+ }
159
+ break;
160
case 0xd90: /* MPU_TYPE */
161
return; /* RO */
162
case 0xd94: /* MPU_CTRL */
163
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
164
}
165
break;
166
}
167
+ case 0xf34: /* FPCCR */
168
+ if (arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
169
+ /* Not all bits here are banked. */
170
+ uint32_t fpccr_s;
171
+
172
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
173
+ /* Don't allow setting of bits not present in v7M */
174
+ value &= (R_V7M_FPCCR_LSPACT_MASK |
175
+ R_V7M_FPCCR_USER_MASK |
176
+ R_V7M_FPCCR_THREAD_MASK |
177
+ R_V7M_FPCCR_HFRDY_MASK |
178
+ R_V7M_FPCCR_MMRDY_MASK |
179
+ R_V7M_FPCCR_BFRDY_MASK |
180
+ R_V7M_FPCCR_MONRDY_MASK |
181
+ R_V7M_FPCCR_LSPEN_MASK |
182
+ R_V7M_FPCCR_ASPEN_MASK);
183
+ }
184
+ value &= ~R_V7M_FPCCR_RES0_MASK;
185
+
186
+ if (!attrs.secure) {
187
+ /* Some non-banked bits are configurably writable by NS */
188
+ fpccr_s = cpu->env.v7m.fpccr[M_REG_S];
189
+ if (!(fpccr_s & R_V7M_FPCCR_LSPENS_MASK)) {
190
+ uint32_t lspen = FIELD_EX32(value, V7M_FPCCR, LSPEN);
191
+ fpccr_s = FIELD_DP32(fpccr_s, V7M_FPCCR, LSPEN, lspen);
192
+ }
193
+ if (!(fpccr_s & R_V7M_FPCCR_CLRONRETS_MASK)) {
194
+ uint32_t cor = FIELD_EX32(value, V7M_FPCCR, CLRONRET);
195
+ fpccr_s = FIELD_DP32(fpccr_s, V7M_FPCCR, CLRONRET, cor);
196
+ }
197
+ if ((s->cpu->env.v7m.aircr & R_V7M_AIRCR_BFHFNMINS_MASK)) {
198
+ uint32_t hfrdy = FIELD_EX32(value, V7M_FPCCR, HFRDY);
199
+ uint32_t bfrdy = FIELD_EX32(value, V7M_FPCCR, BFRDY);
200
+ fpccr_s = FIELD_DP32(fpccr_s, V7M_FPCCR, HFRDY, hfrdy);
201
+ fpccr_s = FIELD_DP32(fpccr_s, V7M_FPCCR, BFRDY, bfrdy);
202
+ }
203
+ /* TODO MONRDY should RAZ/WI if DEMCR.SDME is set */
204
+ {
205
+ uint32_t monrdy = FIELD_EX32(value, V7M_FPCCR, MONRDY);
206
+ fpccr_s = FIELD_DP32(fpccr_s, V7M_FPCCR, MONRDY, monrdy);
207
+ }
208
+
209
+ /*
210
+ * All other non-banked bits are RAZ/WI from NS; write
211
+ * just the banked bits to fpccr[M_REG_NS].
212
+ */
213
+ value &= R_V7M_FPCCR_BANKED_MASK;
214
+ cpu->env.v7m.fpccr[M_REG_NS] = value;
215
+ } else {
216
+ fpccr_s = value;
217
+ }
218
+ cpu->env.v7m.fpccr[M_REG_S] = fpccr_s;
219
+ }
220
+ break;
221
+ case 0xf38: /* FPCAR */
222
+ if (arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
223
+ value &= ~7;
224
+ cpu->env.v7m.fpcar[attrs.secure] = value;
225
+ }
226
+ break;
227
+ case 0xf3c: /* FPDSCR */
228
+ if (arm_feature(&cpu->env, ARM_FEATURE_VFP)) {
229
+ value &= 0x07c00000;
230
+ cpu->env.v7m.fpdscr[attrs.secure] = value;
231
+ }
232
+ break;
233
case 0xf50: /* ICIALLU */
234
case 0xf58: /* ICIMVAU */
235
case 0xf5c: /* DCIMVAC */
236
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
135
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
237
index XXXXXXX..XXXXXXX 100644
136
index XXXXXXX..XXXXXXX 100644
238
--- a/target/arm/cpu.c
137
--- a/target/arm/cpu.c
239
+++ b/target/arm/cpu.c
138
+++ b/target/arm/cpu.c
240
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
139
@@ -XXX,XX +XXX,XX @@ void arm_cpu_post_init(Object *obj)
241
env->v7m.ccr[M_REG_S] |= R_V7M_CCR_UNALIGN_TRP_MASK;
140
if (kvm_enabled()) {
141
kvm_arm_add_vcpu_properties(obj);
142
}
143
+
144
+#ifndef CONFIG_USER_ONLY
145
+ if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) &&
146
+ cpu_isar_feature(aa64_mte, cpu)) {
147
+ object_property_add_link(obj, "tag-memory",
148
+ TYPE_MEMORY_REGION,
149
+ (Object **)&cpu->tag_memory,
150
+ qdev_prop_allow_set_link_before_realize,
151
+ OBJ_PROP_LINK_STRONG);
152
+
153
+ if (arm_feature(&cpu->env, ARM_FEATURE_EL3)) {
154
+ object_property_add_link(obj, "secure-tag-memory",
155
+ TYPE_MEMORY_REGION,
156
+ (Object **)&cpu->secure_tag_memory,
157
+ qdev_prop_allow_set_link_before_realize,
158
+ OBJ_PROP_LINK_STRONG);
159
+ }
160
+ }
161
+#endif
162
}
163
164
static void arm_cpu_finalizefn(Object *obj)
165
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
166
#ifndef CONFIG_USER_ONLY
167
MachineState *ms = MACHINE(qdev_get_machine());
168
unsigned int smp_cpus = ms->smp.cpus;
169
+ bool has_secure = cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY);
170
171
- if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
172
- cs->num_ases = 2;
173
+ /*
174
+ * We must set cs->num_ases to the final value before
175
+ * the first call to cpu_address_space_init.
176
+ */
177
+ if (cpu->tag_memory != NULL) {
178
+ cs->num_ases = 3 + has_secure;
179
+ } else {
180
+ cs->num_ases = 1 + has_secure;
181
+ }
182
183
+ if (has_secure) {
184
if (!cpu->secure_memory) {
185
cpu->secure_memory = cs->memory;
242
}
186
}
243
187
cpu_address_space_init(cs, ARMASIdx_S, "cpu-secure-memory",
244
+ if (arm_feature(env, ARM_FEATURE_VFP)) {
188
cpu->secure_memory);
245
+ env->v7m.fpccr[M_REG_NS] = R_V7M_FPCCR_ASPEN_MASK;
189
- } else {
246
+ env->v7m.fpccr[M_REG_S] = R_V7M_FPCCR_ASPEN_MASK |
190
- cs->num_ases = 1;
247
+ R_V7M_FPCCR_LSPEN_MASK | R_V7M_FPCCR_S_MASK;
191
}
192
+
193
+ if (cpu->tag_memory != NULL) {
194
+ cpu_address_space_init(cs, ARMASIdx_TagNS, "cpu-tag-memory",
195
+ cpu->tag_memory);
196
+ if (has_secure) {
197
+ cpu_address_space_init(cs, ARMASIdx_TagS, "cpu-tag-memory",
198
+ cpu->secure_tag_memory);
248
+ }
199
+ }
249
/* Unlike A/R profile, M profile defines the reset LR value */
200
+ } else if (cpu_isar_feature(aa64_mte, cpu)) {
250
env->regs[14] = 0xffffffff;
201
+ /*
251
202
+ * Since there is no tag memory, we can't meaningfully support MTE
252
diff --git a/target/arm/machine.c b/target/arm/machine.c
203
+ * to its fullest. To avoid problems later, when we would come to
253
index XXXXXXX..XXXXXXX 100644
204
+ * use the tag memory, downgrade support to insns only.
254
--- a/target/arm/machine.c
205
+ */
255
+++ b/target/arm/machine.c
206
+ cpu->isar.id_aa64pfr1 =
256
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_v8m = {
207
+ FIELD_DP64(cpu->isar.id_aa64pfr1, ID_AA64PFR1, MTE, 1);
257
}
208
+ }
258
};
209
+
259
210
cpu_address_space_init(cs, ARMASIdx_NS, "cpu-memory", cs->memory);
260
+static const VMStateDescription vmstate_m_fp = {
211
261
+ .name = "cpu/m/fp",
212
/* No core_count specified, default to smp_cpus. */
262
+ .version_id = 1,
263
+ .minimum_version_id = 1,
264
+ .needed = vfp_needed,
265
+ .fields = (VMStateField[]) {
266
+ VMSTATE_UINT32_ARRAY(env.v7m.fpcar, ARMCPU, M_REG_NUM_BANKS),
267
+ VMSTATE_UINT32_ARRAY(env.v7m.fpccr, ARMCPU, M_REG_NUM_BANKS),
268
+ VMSTATE_UINT32_ARRAY(env.v7m.fpdscr, ARMCPU, M_REG_NUM_BANKS),
269
+ VMSTATE_UINT32_ARRAY(env.v7m.cpacr, ARMCPU, M_REG_NUM_BANKS),
270
+ VMSTATE_UINT32(env.v7m.nsacr, ARMCPU),
271
+ VMSTATE_END_OF_LIST()
272
+ }
273
+};
274
+
275
static const VMStateDescription vmstate_m = {
276
.name = "cpu/m",
277
.version_id = 4,
278
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
279
&vmstate_m_scr,
280
&vmstate_m_other_sp,
281
&vmstate_m_v8m,
282
+ &vmstate_m_fp,
283
NULL
284
}
285
};
286
--
213
--
287
2.20.1
214
2.20.1
288
215
289
216
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This device is used by both ARM (BCM2836, for raspi2) and AArch64
3
Look up the physical address for the given virtual address,
4
(BCM2837, for raspi3) targets, and is not CPU-specific.
4
convert that to a tag physical address, and finally return
5
Move it to common object, so we build it once for all targets.
5
the host address that backs it.
6
6
7
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20190427133028.12874-1-philmd@redhat.com
8
Message-id: 20200626033144.790098-45-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
11
---
12
hw/dma/Makefile.objs | 2 +-
12
target/arm/mte_helper.c | 131 ++++++++++++++++++++++++++++++++++++++++
13
1 file changed, 1 insertion(+), 1 deletion(-)
13
1 file changed, 131 insertions(+)
14
14
15
diff --git a/hw/dma/Makefile.objs b/hw/dma/Makefile.objs
15
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/dma/Makefile.objs
17
--- a/target/arm/mte_helper.c
18
+++ b/hw/dma/Makefile.objs
18
+++ b/target/arm/mte_helper.c
19
@@ -XXX,XX +XXX,XX @@ common-obj-$(CONFIG_XLNX_ZYNQMP_ARM) += xlnx-zdma.o
19
@@ -XXX,XX +XXX,XX @@
20
20
#include "cpu.h"
21
obj-$(CONFIG_OMAP) += omap_dma.o soc_dma.o
21
#include "internals.h"
22
obj-$(CONFIG_PXA2XX) += pxa2xx_dma.o
22
#include "exec/exec-all.h"
23
-obj-$(CONFIG_RASPI) += bcm2835_dma.o
23
+#include "exec/ram_addr.h"
24
+common-obj-$(CONFIG_RASPI) += bcm2835_dma.o
24
#include "exec/cpu_ldst.h"
25
#include "exec/helper-proto.h"
26
27
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
28
int ptr_size, MMUAccessType tag_access,
29
int tag_size, uintptr_t ra)
30
{
31
+#ifdef CONFIG_USER_ONLY
32
/* Tag storage not implemented. */
33
return NULL;
34
+#else
35
+ uintptr_t index;
36
+ CPUIOTLBEntry *iotlbentry;
37
+ int in_page, flags;
38
+ ram_addr_t ptr_ra;
39
+ hwaddr ptr_paddr, tag_paddr, xlat;
40
+ MemoryRegion *mr;
41
+ ARMASIdx tag_asi;
42
+ AddressSpace *tag_as;
43
+ void *host;
44
+
45
+ /*
46
+ * Probe the first byte of the virtual address. This raises an
47
+ * exception for inaccessible pages, and resolves the virtual address
48
+ * into the softmmu tlb.
49
+ *
50
+ * When RA == 0, this is for mte_probe1. The page is expected to be
51
+ * valid. Indicate to probe_access_flags no-fault, then assert that
52
+ * we received a valid page.
53
+ */
54
+ flags = probe_access_flags(env, ptr, ptr_access, ptr_mmu_idx,
55
+ ra == 0, &host, ra);
56
+ assert(!(flags & TLB_INVALID_MASK));
57
+
58
+ /*
59
+ * Find the iotlbentry for ptr. This *must* be present in the TLB
60
+ * because we just found the mapping.
61
+ * TODO: Perhaps there should be a cputlb helper that returns a
62
+ * matching tlb entry + iotlb entry.
63
+ */
64
+ index = tlb_index(env, ptr_mmu_idx, ptr);
65
+# ifdef CONFIG_DEBUG_TCG
66
+ {
67
+ CPUTLBEntry *entry = tlb_entry(env, ptr_mmu_idx, ptr);
68
+ target_ulong comparator = (ptr_access == MMU_DATA_LOAD
69
+ ? entry->addr_read
70
+ : tlb_addr_write(entry));
71
+ g_assert(tlb_hit(comparator, ptr));
72
+ }
73
+# endif
74
+ iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
75
+
76
+ /* If the virtual page MemAttr != Tagged, access unchecked. */
77
+ if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
78
+ return NULL;
79
+ }
80
+
81
+ /*
82
+ * If not backed by host ram, there is no tag storage: access unchecked.
83
+ * This is probably a guest os bug though, so log it.
84
+ */
85
+ if (unlikely(flags & TLB_MMIO)) {
86
+ qemu_log_mask(LOG_GUEST_ERROR,
87
+ "Page @ 0x%" PRIx64 " indicates Tagged Normal memory "
88
+ "but is not backed by host ram\n", ptr);
89
+ return NULL;
90
+ }
91
+
92
+ /*
93
+ * The Normal memory access can extend to the next page. E.g. a single
94
+ * 8-byte access to the last byte of a page will check only the last
95
+ * tag on the first page.
96
+ * Any page access exception has priority over tag check exception.
97
+ */
98
+ in_page = -(ptr | TARGET_PAGE_MASK);
99
+ if (unlikely(ptr_size > in_page)) {
100
+ void *ignore;
101
+ flags |= probe_access_flags(env, ptr + in_page, ptr_access,
102
+ ptr_mmu_idx, ra == 0, &ignore, ra);
103
+ assert(!(flags & TLB_INVALID_MASK));
104
+ }
105
+
106
+ /* Any debug exception has priority over a tag check exception. */
107
+ if (unlikely(flags & TLB_WATCHPOINT)) {
108
+ int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
109
+ assert(ra != 0);
110
+ cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
111
+ iotlbentry->attrs, wp, ra);
112
+ }
113
+
114
+ /*
115
+ * Find the physical address within the normal mem space.
116
+ * The memory region lookup must succeed because TLB_MMIO was
117
+ * not set in the cputlb lookup above.
118
+ */
119
+ mr = memory_region_from_host(host, &ptr_ra);
120
+ tcg_debug_assert(mr != NULL);
121
+ tcg_debug_assert(memory_region_is_ram(mr));
122
+ ptr_paddr = ptr_ra;
123
+ do {
124
+ ptr_paddr += mr->addr;
125
+ mr = mr->container;
126
+ } while (mr);
127
+
128
+ /* Convert to the physical address in tag space. */
129
+ tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
130
+
131
+ /* Look up the address in tag space. */
132
+ tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
133
+ tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
134
+ mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
135
+ tag_access == MMU_DATA_STORE,
136
+ iotlbentry->attrs);
137
+
138
+ /*
139
+ * Note that @mr will never be NULL. If there is nothing in the address
140
+ * space at @tag_paddr, the translation will return the unallocated memory
141
+ * region. For our purposes, the result must be ram.
142
+ */
143
+ if (unlikely(!memory_region_is_ram(mr))) {
144
+ /* ??? Failure is a board configuration error. */
145
+ qemu_log_mask(LOG_UNIMP,
146
+ "Tag Memory @ 0x%" HWADDR_PRIx " not found for "
147
+ "Normal Memory @ 0x%" HWADDR_PRIx "\n",
148
+ tag_paddr, ptr_paddr);
149
+ return NULL;
150
+ }
151
+
152
+ /*
153
+ * Ensure the tag memory is dirty on write, for migration.
154
+ * Tag memory can never contain code or display memory (vga).
155
+ */
156
+ if (tag_access == MMU_DATA_STORE) {
157
+ ram_addr_t tag_ra = memory_region_get_ram_addr(mr) + xlat;
158
+ cpu_physical_memory_set_dirty_flag(tag_ra, DIRTY_MEMORY_MIGRATION);
159
+ }
160
+
161
+ return memory_region_get_ram_ptr(mr) + xlat;
162
+#endif
163
}
164
165
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
25
--
166
--
26
2.20.1
167
2.20.1
27
168
28
169
diff view generated by jsdifflib
1
Normally configure identifies the source path by looking
1
From: Richard Henderson <richard.henderson@linaro.org>
2
at the location where the configure script itself exists.
3
We also provide a --source-path option which lets the user
4
manually override this.
5
2
6
There isn't really an obvious use case for the --source-path
3
We now implement all of the components of MTE, without actually
7
option, and in commit 927128222b0a91f56c13a in 2017 we
4
supporting any tagged memory. All MTE instructions will work,
8
accidentally added some logic that looks at $source_path
5
trivially, so we can enable support.
9
before the command line option that overrides it has been
10
processed.
11
6
12
The fact that nobody complained suggests that there isn't
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
any use of this option and we aren't testing it either;
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
14
remove it. This allows us to move the "make $source_path
9
Message-id: 20200626033144.790098-46-richard.henderson@linaro.org
15
absolute" logic up so that there is no window in the script
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
where $source_path is set but not yet absolute.
11
---
12
target/arm/cpu64.c | 5 +++++
13
1 file changed, 5 insertions(+)
17
14
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
19
Reviewed-by: Daniel P. Berrangé <berrange@redhat.com>
16
index XXXXXXX..XXXXXXX 100644
20
Message-id: 20190318134019.23729-1-peter.maydell@linaro.org
17
--- a/target/arm/cpu64.c
21
---
18
+++ b/target/arm/cpu64.c
22
configure | 10 ++--------
19
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
23
1 file changed, 2 insertions(+), 8 deletions(-)
20
24
21
t = cpu->isar.id_aa64pfr1;
25
diff --git a/configure b/configure
22
t = FIELD_DP64(t, ID_AA64PFR1, BT, 1);
26
index XXXXXXX..XXXXXXX 100755
23
+ /*
27
--- a/configure
24
+ * Begin with full support for MTE; will be downgraded to MTE=1
28
+++ b/configure
25
+ * during realize if the board provides no tag memory.
29
@@ -XXX,XX +XXX,XX @@ ld_has() {
26
+ */
30
27
+ t = FIELD_DP64(t, ID_AA64PFR1, MTE, 2);
31
# default parameters
28
cpu->isar.id_aa64pfr1 = t;
32
source_path=$(dirname "$0")
29
33
+# make source path absolute
30
t = cpu->isar.id_aa64mmfr1;
34
+source_path=$(cd "$source_path"; pwd)
35
cpu=""
36
iasl="iasl"
37
interp_prefix="/usr/gnemul/qemu-%M"
38
@@ -XXX,XX +XXX,XX @@ for opt do
39
;;
40
--cxx=*) CXX="$optarg"
41
;;
42
- --source-path=*) source_path="$optarg"
43
- ;;
44
--cpu=*) cpu="$optarg"
45
;;
46
--extra-cflags=*) QEMU_CFLAGS="$QEMU_CFLAGS $optarg"
47
@@ -XXX,XX +XXX,XX @@ if test "$debug_info" = "yes"; then
48
LDFLAGS="-g $LDFLAGS"
49
fi
50
51
-# make source path absolute
52
-source_path=$(cd "$source_path"; pwd)
53
-
54
# running configure in the source tree?
55
# we know that's the case if configure is there.
56
if test -f "./configure"; then
57
@@ -XXX,XX +XXX,XX @@ for opt do
58
;;
59
--interp-prefix=*) interp_prefix="$optarg"
60
;;
61
- --source-path=*)
62
- ;;
63
--cross-prefix=*)
64
;;
65
--cc=*)
66
@@ -XXX,XX +XXX,XX @@ $(echo Available targets: $default_target_list | \
67
--target-list-exclude=LIST exclude a set of targets from the default target-list
68
69
Advanced options (experts only):
70
- --source-path=PATH path of source code [$source_path]
71
--cross-prefix=PREFIX use PREFIX for compile tools [$cross_prefix]
72
--cc=CC use C compiler CC [$cc]
73
--iasl=IASL use ACPI compiler IASL [$iasl]
74
--
31
--
75
2.20.1
32
2.20.1
76
33
77
34
diff view generated by jsdifflib