1
Mostly this is RTH's memtag series, but there are also some cleanups
1
Two small bugfixes, plus most of RTH's refactoring of cpregs
2
from Philippe.
2
handling.
3
3
4
thanks
5
-- PMM
4
-- PMM
6
5
7
The following changes since commit 10f7ffabf9c507fc02382b89912003b1c43c3231:
6
The following changes since commit 1fba9dc71a170b3a05b9d3272dd8ecfe7f26e215:
8
7
9
Merge remote-tracking branch 'remotes/mcayland/tags/qemu-macppc-20200626' into staging (2020-06-26 12:14:18 +0100)
8
Merge tag 'pull-request-2022-05-04' of https://gitlab.com/thuth/qemu into staging (2022-05-04 08:07:02 -0700)
10
9
11
are available in the Git repository at:
10
are available in the Git repository at:
12
11
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200626
12
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20220505
14
13
15
for you to fetch changes up to c7459633baa71d1781fde4a245d6ec9ce2f008cf:
14
for you to fetch changes up to 99a50d1a67c602126fc2b3a4812d3000eba9bf34:
16
15
17
target/arm: Enable MTE (2020-06-26 14:32:24 +0100)
16
target/arm: read access to performance counters from EL0 (2022-05-05 09:36:22 +0100)
18
17
19
----------------------------------------------------------------
18
----------------------------------------------------------------
20
target-arm queue:
19
target-arm queue:
21
* hw/arm/aspeed: improve QOM usage
20
* Enable read access to performance counters from EL0
22
* hw/misc/pca9552: trace GPIO change events
21
* Enable SCTLR_EL1.BT0 for aarch64-linux-user
23
* target/arm: Implement ARMv8.5-MemTag for system emulation
22
* Refactoring of cpreg handling
24
23
25
----------------------------------------------------------------
24
----------------------------------------------------------------
26
Philippe Mathieu-Daudé (12):
25
Alex Zuepke (1):
27
hw/arm/aspeed: Remove extraneous MemoryRegion object owner
26
target/arm: read access to performance counters from EL0
28
hw/arm/aspeed: Rename AspeedBoardState as AspeedMachineState
29
hw/arm/aspeed: QOM'ify AspeedMachineState
30
hw/i2c/core: Add i2c_try_create_slave() and i2c_realize_and_unref()
31
hw/misc/pca9552: Rename 'nr_leds' as 'pin_count'
32
hw/misc/pca9552: Rename generic code as pca955x
33
hw/misc/pca9552: Add generic PCA955xClass, parent of TYPE_PCA9552
34
hw/misc/pca9552: Add a 'description' property for debugging purpose
35
hw/misc/pca9552: Trace GPIO High/Low events
36
hw/arm/aspeed: Describe each PCA9552 device
37
hw/misc/pca9552: Trace GPIO change events
38
hw/misc/pca9552: Model qdev output GPIOs
39
27
40
Richard Henderson (45):
28
Richard Henderson (22):
41
target/arm: Add isar tests for mte
29
target/arm: Enable SCTLR_EL1.BT0 for aarch64-linux-user
42
target/arm: Improve masking of SCR RES0 bits
30
target/arm: Split out cpregs.h
43
target/arm: Add support for MTE to SCTLR_ELx
31
target/arm: Reorg CPAccessResult and access_check_cp_reg
44
target/arm: Add support for MTE to HCR_EL2 and SCR_EL3
32
target/arm: Replace sentinels with ARRAY_SIZE in cpregs.h
45
target/arm: Rename DISAS_UPDATE to DISAS_UPDATE_EXIT
33
target/arm: Make some more cpreg data static const
46
target/arm: Add DISAS_UPDATE_NOCHAIN
34
target/arm: Reorg ARMCPRegInfo type field bits
47
target/arm: Add MTE system registers
35
target/arm: Avoid bare abort() or assert(0)
48
target/arm: Add MTE bits to tb_flags
36
target/arm: Change cpreg access permissions to enum
49
target/arm: Implement the IRG instruction
37
target/arm: Name CPState type
50
target/arm: Revise decoding for disas_add_sub_imm
38
target/arm: Name CPSecureState type
51
target/arm: Implement the ADDG, SUBG instructions
39
target/arm: Drop always-true test in define_arm_vh_e2h_redirects_aliases
52
target/arm: Implement the GMI instruction
40
target/arm: Store cpregs key in the hash table directly
53
target/arm: Implement the SUBP instruction
41
target/arm: Merge allocation of the cpreg and its name
54
target/arm: Define arm_cpu_do_unaligned_access for user-only
42
target/arm: Hoist computation of key in add_cpreg_to_hashtable
55
target/arm: Implement LDG, STG, ST2G instructions
43
target/arm: Consolidate cpreg updates in add_cpreg_to_hashtable
56
target/arm: Implement the STGP instruction
44
target/arm: Use bool for is64 and ns in add_cpreg_to_hashtable
57
target/arm: Restrict the values of DCZID.BS under TCG
45
target/arm: Hoist isbanked computation in add_cpreg_to_hashtable
58
target/arm: Simplify DC_ZVA
46
target/arm: Perform override check early in add_cpreg_to_hashtable
59
target/arm: Implement the LDGM, STGM, STZGM instructions
47
target/arm: Reformat comments in add_cpreg_to_hashtable
60
target/arm: Implement the access tag cache flushes
48
target/arm: Remove HOST_BIG_ENDIAN ifdef in add_cpreg_to_hashtable
61
target/arm: Move regime_el to internals.h
49
target/arm: Add isar predicates for FEAT_Debugv8p2
62
target/arm: Move regime_tcr to internals.h
50
target/arm: Add isar_feature_{aa64,any}_ras
63
target/arm: Add gen_mte_check1
64
target/arm: Add gen_mte_checkN
65
target/arm: Implement helper_mte_check1
66
target/arm: Implement helper_mte_checkN
67
target/arm: Add helper_mte_check_zva
68
target/arm: Use mte_checkN for sve unpredicated loads
69
target/arm: Use mte_checkN for sve unpredicated stores
70
target/arm: Use mte_check1 for sve LD1R
71
target/arm: Tidy trans_LD1R_zpri
72
target/arm: Add arm_tlb_bti_gp
73
target/arm: Add mte helpers for sve scalar + int loads
74
target/arm: Add mte helpers for sve scalar + int stores
75
target/arm: Add mte helpers for sve scalar + int ff/nf loads
76
target/arm: Handle TBI for sve scalar + int memory ops
77
target/arm: Add mte helpers for sve scatter/gather memory ops
78
target/arm: Complete TBI clearing for user-only for SVE
79
target/arm: Implement data cache set allocation tags
80
target/arm: Set PSTATE.TCO on exception entry
81
target/arm: Always pass cacheattr to get_phys_addr
82
target/arm: Cache the Tagged bit for a page in MemTxAttrs
83
target/arm: Create tagged ram when MTE is enabled
84
target/arm: Add allocation tag storage for system mode
85
target/arm: Enable MTE
86
51
87
include/hw/arm/aspeed.h | 12 +-
52
target/arm/cpregs.h | 453 ++++++++++++++++++++++++++++++++++++++
88
include/hw/i2c/i2c.h | 2 +
53
target/arm/cpu.h | 393 +++------------------------------
89
include/hw/misc/pca9552.h | 16 +-
54
hw/arm/pxa2xx.c | 2 +-
90
target/arm/cpu.h | 50 +-
55
hw/arm/pxa2xx_pic.c | 2 +-
91
target/arm/helper-a64.h | 16 +
56
hw/intc/arm_gicv3_cpuif.c | 6 +-
92
target/arm/helper-sve.h | 488 ++++++++++++++
57
hw/intc/arm_gicv3_kvm.c | 3 +-
93
target/arm/helper.h | 2 +
58
target/arm/cpu.c | 25 +--
94
target/arm/internals.h | 153 ++++-
59
target/arm/cpu64.c | 2 +-
95
target/arm/translate-a64.h | 5 +
60
target/arm/cpu_tcg.c | 5 +-
96
target/arm/translate.h | 23 +-
61
target/arm/gdbstub.c | 5 +-
97
hw/arm/aspeed.c | 46 +-
62
target/arm/helper.c | 358 +++++++++++++-----------------
98
hw/arm/virt.c | 55 +-
63
target/arm/hvf/hvf.c | 2 +-
99
hw/i2c/core.c | 18 +-
64
target/arm/kvm-stub.c | 4 +-
100
hw/misc/pca9552.c | 216 +++++--
65
target/arm/kvm.c | 4 +-
101
target/arm/cpu.c | 81 ++-
66
target/arm/machine.c | 4 +-
102
target/arm/cpu64.c | 5 +
67
target/arm/op_helper.c | 57 ++---
103
target/arm/helper-a64.c | 94 +--
68
target/arm/translate-a64.c | 14 +-
104
target/arm/helper.c | 423 ++++++++++---
69
target/arm/translate-neon.c | 2 +-
105
target/arm/m_helper.c | 11 +-
70
target/arm/translate.c | 13 +-
106
target/arm/mte_helper.c | 906 ++++++++++++++++++++++++++
71
tests/tcg/aarch64/bti-3.c | 42 ++++
107
target/arm/op_helper.c | 16 +
72
tests/tcg/aarch64/Makefile.target | 6 +-
108
target/arm/sve_helper.c | 616 ++++++++++++++----
73
21 files changed, 738 insertions(+), 664 deletions(-)
109
target/arm/tlb_helper.c | 13 +-
74
create mode 100644 target/arm/cpregs.h
110
target/arm/translate-a64.c | 657 ++++++++++++++++---
75
create mode 100644 tests/tcg/aarch64/bti-3.c
111
target/arm/translate-sve.c | 1366 ++++++++++++++++++++++++++--------------
112
target/arm/translate-vfp.inc.c | 4 +-
113
target/arm/translate.c | 16 +-
114
hw/misc/trace-events | 4 +
115
target/arm/Makefile.objs | 1 +
116
29 files changed, 4391 insertions(+), 924 deletions(-)
117
create mode 100644 target/arm/mte_helper.c
118
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
I'm confused by this code, 'bmc' is created as:
4
5
bmc = g_new0(AspeedBoardState, 1);
6
7
Then we use it as QOM owner for different MemoryRegion objects.
8
But looking at memory_region_init_ram (similarly for ROM):
9
10
void memory_region_init_ram(MemoryRegion *mr,
11
struct Object *owner,
12
const char *name,
13
uint64_t size,
14
Error **errp)
15
{
16
DeviceState *owner_dev;
17
Error *err = NULL;
18
19
memory_region_init_ram_nomigrate(mr, owner, name, size, &err);
20
if (err) {
21
error_propagate(errp, err);
22
return;
23
}
24
/* This will assert if owner is neither NULL nor a DeviceState.
25
* We only want the owner here for the purposes of defining a
26
* unique name for migration. TODO: Ideally we should implement
27
* a naming scheme for Objects which are not DeviceStates, in
28
* which case we can relax this restriction.
29
*/
30
owner_dev = DEVICE(owner);
31
vmstate_register_ram(mr, owner_dev);
32
}
33
34
The expected assertion is not triggered ('bmc' is not NULL neither
35
a DeviceState).
36
37
'bmc' structure is defined as:
38
39
struct AspeedBoardState {
40
AspeedSoCState soc;
41
MemoryRegion ram_container;
42
MemoryRegion max_ram;
43
};
44
45
What happens is when using 'OBJECT(bmc)', the QOM macros cast the
46
memory pointed by bmc, which first member is 'soc', which is
47
initialized ...:
48
49
object_initialize_child(OBJECT(machine), "soc",
50
&bmc->soc, amc->soc_name);
51
52
The 'soc' object is indeed a DeviceState, so the assertion passes.
53
54
Since this is fragile and only happens to work by luck, remove the
55
dangerous OBJECT(bmc) owner argument.
56
57
Note, this probably breaks migration for this machine.
58
59
Reviewed-by: Cédric Le Goater <clg@kaod.org>
60
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
61
Message-id: 20200623072132.2868-2-f4bug@amsat.org
62
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
63
---
64
hw/arm/aspeed.c | 6 +++---
65
1 file changed, 3 insertions(+), 3 deletions(-)
66
67
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/aspeed.c
70
+++ b/hw/arm/aspeed.c
71
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
72
* needed by the flash modules of the Aspeed machines.
73
*/
74
if (ASPEED_MACHINE(machine)->mmio_exec) {
75
- memory_region_init_alias(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
76
+ memory_region_init_alias(boot_rom, NULL, "aspeed.boot_rom",
77
&fl->mmio, 0, fl->size);
78
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
79
boot_rom);
80
} else {
81
- memory_region_init_rom(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
82
+ memory_region_init_rom(boot_rom, NULL, "aspeed.boot_rom",
83
fl->size, &error_abort);
84
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
85
boot_rom);
86
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
87
if (machine->kernel_filename && sc->num_cpus > 1) {
88
/* With no u-boot we must set up a boot stub for the secondary CPU */
89
MemoryRegion *smpboot = g_new(MemoryRegion, 1);
90
- memory_region_init_ram(smpboot, OBJECT(bmc), "aspeed.smpboot",
91
+ memory_region_init_ram(smpboot, NULL, "aspeed.smpboot",
92
0x80, &error_abort);
93
memory_region_add_subregion(get_system_memory(),
94
AST_SMP_MAILBOX_BASE, smpboot);
95
--
96
2.20.1
97
98
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
To have a more consistent naming, rename AspeedBoardState
4
as AspeedMachineState.
5
6
Suggested-by: Cédric Le Goater <clg@kaod.org>
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Message-id: 20200623072132.2868-3-f4bug@amsat.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
include/hw/arm/aspeed.h | 4 ++--
13
hw/arm/aspeed.c | 20 ++++++++++----------
14
2 files changed, 12 insertions(+), 12 deletions(-)
15
16
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/aspeed.h
19
+++ b/include/hw/arm/aspeed.h
20
@@ -XXX,XX +XXX,XX @@
21
22
#include "hw/boards.h"
23
24
-typedef struct AspeedBoardState AspeedBoardState;
25
+typedef struct AspeedMachineState AspeedMachineState;
26
27
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
28
#define ASPEED_MACHINE(obj) \
29
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineClass {
30
const char *spi_model;
31
uint32_t num_cs;
32
uint32_t macs_mask;
33
- void (*i2c_init)(AspeedBoardState *bmc);
34
+ void (*i2c_init)(AspeedMachineState *bmc);
35
} AspeedMachineClass;
36
37
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
.board_id = -1, /* device-tree-only board */
44
};
45
46
-struct AspeedBoardState {
47
+struct AspeedMachineState {
48
AspeedSoCState soc;
49
MemoryRegion ram_container;
50
MemoryRegion max_ram;
51
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
52
53
static void aspeed_machine_init(MachineState *machine)
54
{
55
- AspeedBoardState *bmc;
56
+ AspeedMachineState *bmc;
57
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
58
AspeedSoCClass *sc;
59
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
60
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
61
int i;
62
NICInfo *nd = &nd_table[0];
63
64
- bmc = g_new0(AspeedBoardState, 1);
65
+ bmc = g_new0(AspeedMachineState, 1);
66
67
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
68
4 * GiB);
69
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
70
arm_load_kernel(ARM_CPU(first_cpu), machine, &aspeed_board_binfo);
71
}
72
73
-static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
74
+static void palmetto_bmc_i2c_init(AspeedMachineState *bmc)
75
{
76
AspeedSoCState *soc = &bmc->soc;
77
DeviceState *dev;
78
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
79
object_property_set_int(OBJECT(dev), 110000, "temperature3", &error_abort);
80
}
81
82
-static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
83
+static void ast2500_evb_i2c_init(AspeedMachineState *bmc)
84
{
85
AspeedSoCState *soc = &bmc->soc;
86
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
87
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
88
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
89
}
90
91
-static void ast2600_evb_i2c_init(AspeedBoardState *bmc)
92
+static void ast2600_evb_i2c_init(AspeedMachineState *bmc)
93
{
94
/* Start with some devices on our I2C busses */
95
ast2500_evb_i2c_init(bmc);
96
}
97
98
-static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
99
+static void romulus_bmc_i2c_init(AspeedMachineState *bmc)
100
{
101
AspeedSoCState *soc = &bmc->soc;
102
103
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
104
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
105
}
106
107
-static void swift_bmc_i2c_init(AspeedBoardState *bmc)
108
+static void swift_bmc_i2c_init(AspeedMachineState *bmc)
109
{
110
AspeedSoCState *soc = &bmc->soc;
111
112
@@ -XXX,XX +XXX,XX @@ static void swift_bmc_i2c_init(AspeedBoardState *bmc)
113
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 12), "tmp105", 0x4a);
114
}
115
116
-static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
117
+static void sonorapass_bmc_i2c_init(AspeedMachineState *bmc)
118
{
119
AspeedSoCState *soc = &bmc->soc;
120
121
@@ -XXX,XX +XXX,XX @@ static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
122
123
}
124
125
-static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
126
+static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
127
{
128
AspeedSoCState *soc = &bmc->soc;
129
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
AspeedMachineState seems crippled. We use incorrectly 2
4
different structures to do the same thing. Merge them
5
altogether:
6
- Move AspeedMachine fields to AspeedMachineState
7
- AspeedMachineState is now QOM
8
- Remove unused AspeedMachine structure
9
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Message-id: 20200623072132.2868-4-f4bug@amsat.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
include/hw/arm/aspeed.h | 8 +-------
16
hw/arm/aspeed.c | 11 +++++++----
17
2 files changed, 8 insertions(+), 11 deletions(-)
18
19
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/include/hw/arm/aspeed.h
22
+++ b/include/hw/arm/aspeed.h
23
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineState AspeedMachineState;
24
25
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
26
#define ASPEED_MACHINE(obj) \
27
- OBJECT_CHECK(AspeedMachine, (obj), TYPE_ASPEED_MACHINE)
28
-
29
-typedef struct AspeedMachine {
30
- MachineState parent_obj;
31
-
32
- bool mmio_exec;
33
-} AspeedMachine;
34
+ OBJECT_CHECK(AspeedMachineState, (obj), TYPE_ASPEED_MACHINE)
35
36
#define ASPEED_MAC0_ON (1 << 0)
37
#define ASPEED_MAC1_ON (1 << 1)
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
};
44
45
struct AspeedMachineState {
46
+ /* Private */
47
+ MachineState parent_obj;
48
+ /* Public */
49
+
50
AspeedSoCState soc;
51
MemoryRegion ram_container;
52
MemoryRegion max_ram;
53
+ bool mmio_exec;
54
};
55
56
/* Palmetto hardware value: 0x120CE416 */
57
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
58
59
static void aspeed_machine_init(MachineState *machine)
60
{
61
- AspeedMachineState *bmc;
62
+ AspeedMachineState *bmc = ASPEED_MACHINE(machine);
63
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
64
AspeedSoCClass *sc;
65
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
66
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
67
int i;
68
NICInfo *nd = &nd_table[0];
69
70
- bmc = g_new0(AspeedMachineState, 1);
71
-
72
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
73
4 * GiB);
74
memory_region_add_subregion(&bmc->ram_container, 0, machine->ram);
75
@@ -XXX,XX +XXX,XX @@ static const TypeInfo aspeed_machine_types[] = {
76
}, {
77
.name = TYPE_ASPEED_MACHINE,
78
.parent = TYPE_MACHINE,
79
- .instance_size = sizeof(AspeedMachine),
80
+ .instance_size = sizeof(AspeedMachineState),
81
.instance_init = aspeed_machine_instance_init,
82
.class_size = sizeof(AspeedMachineClass),
83
.class_init = aspeed_machine_class_init,
84
--
85
2.20.1
86
87
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Extract i2c_try_create_slave() and i2c_realize_and_unref()
4
from i2c_create_slave().
5
We can now set properties on a I2CSlave before it is realized.
6
7
This is in line with the recent qdev/QOM changes merged
8
in commit 6675a653d2e.
9
10
Reviewed-by: Corey Minyard <cminyard@mvista.com>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Reviewed-by: Markus Armbruster <armbru@redhat.com>
14
Tested-by: Cédric Le Goater <clg@kaod.org>
15
Message-id: 20200623072723.6324-2-f4bug@amsat.org
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
---
18
include/hw/i2c/i2c.h | 2 ++
19
hw/i2c/core.c | 18 ++++++++++++++++--
20
2 files changed, 18 insertions(+), 2 deletions(-)
21
22
diff --git a/include/hw/i2c/i2c.h b/include/hw/i2c/i2c.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/include/hw/i2c/i2c.h
25
+++ b/include/hw/i2c/i2c.h
26
@@ -XXX,XX +XXX,XX @@ int i2c_send(I2CBus *bus, uint8_t data);
27
uint8_t i2c_recv(I2CBus *bus);
28
29
DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr);
30
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr);
31
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp);
32
33
/* lm832x.c */
34
void lm832x_key_event(DeviceState *dev, int key, int state);
35
diff --git a/hw/i2c/core.c b/hw/i2c/core.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/hw/i2c/core.c
38
+++ b/hw/i2c/core.c
39
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_i2c_slave = {
40
}
41
};
42
43
-DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
44
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr)
45
{
46
DeviceState *dev;
47
48
dev = qdev_new(name);
49
qdev_prop_set_uint8(dev, "address", addr);
50
- qdev_realize_and_unref(dev, &bus->qbus, &error_fatal);
51
+ return dev;
52
+}
53
+
54
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp)
55
+{
56
+ return qdev_realize_and_unref(dev, &bus->qbus, errp);
57
+}
58
+
59
+DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
60
+{
61
+ DeviceState *dev;
62
+
63
+ dev = i2c_try_create_slave(name, addr);
64
+ i2c_realize_and_unref(dev, bus, &error_fatal);
65
+
66
return dev;
67
}
68
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
The PCA9552 device does not expose LEDs, but simple pins
4
to connnect LEDs to. To be clearer with the device model,
5
rename 'nr_leds' as 'pin_count'.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-3-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 2 +-
14
hw/misc/pca9552.c | 10 +++++-----
15
2 files changed, 6 insertions(+), 6 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
22
23
uint8_t regs[PCA9552_NR_REGS];
24
uint8_t max_reg;
25
- uint8_t nr_leds;
26
+ uint8_t pin_count;
27
} PCA9552State;
28
29
#endif
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
35
{
36
int i;
37
38
- for (i = 0; i < s->nr_leds; i++) {
39
+ for (i = 0; i < s->pin_count; i++) {
40
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
41
uint8_t input_shift = (i % 8);
42
uint8_t config = pca9552_pin_get_config(s, i);
43
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
44
error_setg(errp, "%s: error reading %s", __func__, name);
45
return;
46
}
47
- if (led < 0 || led > s->nr_leds) {
48
+ if (led < 0 || led > s->pin_count) {
49
error_setg(errp, "%s invalid led %s", __func__, name);
50
return;
51
}
52
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
53
error_setg(errp, "%s: error reading %s", __func__, name);
54
return;
55
}
56
- if (led < 0 || led > s->nr_leds) {
57
+ if (led < 0 || led > s->pin_count) {
58
error_setg(errp, "%s invalid led %s", __func__, name);
59
return;
60
}
61
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
62
* PCA955X device
63
*/
64
s->max_reg = PCA9552_LS3;
65
- s->nr_leds = 16;
66
+ s->pin_count = 16;
67
68
- for (led = 0; led < s->nr_leds; led++) {
69
+ for (led = 0; led < s->pin_count; led++) {
70
char *name;
71
72
name = g_strdup_printf("led%d", led);
73
--
74
2.20.1
75
76
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Various code from the PCA9552 device model is generic to the
4
PCA955X family. We'll split the generic code in a base class
5
in the next commit. To ease review, first do a dumb renaming.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-4-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 10 ++---
14
hw/misc/pca9552.c | 80 +++++++++++++++++++--------------------
15
2 files changed, 45 insertions(+), 45 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@
22
#include "hw/i2c/i2c.h"
23
24
#define TYPE_PCA9552 "pca9552"
25
-#define PCA9552(obj) OBJECT_CHECK(PCA9552State, (obj), TYPE_PCA9552)
26
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
28
-#define PCA9552_NR_REGS 10
29
+#define PCA955X_NR_REGS 10
30
31
-typedef struct PCA9552State {
32
+typedef struct PCA955xState {
33
/*< private >*/
34
I2CSlave i2c;
35
/*< public >*/
36
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
37
uint8_t len;
38
uint8_t pointer;
39
40
- uint8_t regs[PCA9552_NR_REGS];
41
+ uint8_t regs[PCA955X_NR_REGS];
42
uint8_t max_reg;
43
uint8_t pin_count;
44
-} PCA9552State;
45
+} PCA955xState;
46
47
#endif
48
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/hw/misc/pca9552.c
51
+++ b/hw/misc/pca9552.c
52
@@ -XXX,XX +XXX,XX @@
53
54
static const char *led_state[] = {"on", "off", "pwm0", "pwm1"};
55
56
-static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
57
+static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
58
{
59
uint8_t reg = PCA9552_LS0 + (pin / 4);
60
uint8_t shift = (pin % 4) << 1;
61
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
62
return extract32(s->regs[reg], shift, 2);
63
}
64
65
-static void pca9552_update_pin_input(PCA9552State *s)
66
+static void pca955x_update_pin_input(PCA955xState *s)
67
{
68
int i;
69
70
for (i = 0; i < s->pin_count; i++) {
71
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
72
uint8_t input_shift = (i % 8);
73
- uint8_t config = pca9552_pin_get_config(s, i);
74
+ uint8_t config = pca955x_pin_get_config(s, i);
75
76
switch (config) {
77
case PCA9552_LED_ON:
78
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
79
}
80
}
81
82
-static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
83
+static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
84
{
85
switch (reg) {
86
case PCA9552_INPUT0:
87
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
88
}
89
}
90
91
-static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
92
+static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
93
{
94
switch (reg) {
95
case PCA9552_PSC0:
96
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
97
case PCA9552_LS2:
98
case PCA9552_LS3:
99
s->regs[reg] = data;
100
- pca9552_update_pin_input(s);
101
+ pca955x_update_pin_input(s);
102
break;
103
104
case PCA9552_INPUT0:
105
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
106
* after each byte is sent to or received by the device. The index
107
* rollovers to 0 when the maximum register address is reached.
108
*/
109
-static void pca9552_autoinc(PCA9552State *s)
110
+static void pca955x_autoinc(PCA955xState *s)
111
{
112
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
113
uint8_t reg = s->pointer & 0xf;
114
@@ -XXX,XX +XXX,XX @@ static void pca9552_autoinc(PCA9552State *s)
115
}
116
}
117
118
-static uint8_t pca9552_recv(I2CSlave *i2c)
119
+static uint8_t pca955x_recv(I2CSlave *i2c)
120
{
121
- PCA9552State *s = PCA9552(i2c);
122
+ PCA955xState *s = PCA955X(i2c);
123
uint8_t ret;
124
125
- ret = pca9552_read(s, s->pointer & 0xf);
126
+ ret = pca955x_read(s, s->pointer & 0xf);
127
128
/*
129
* From the Specs:
130
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_recv(I2CSlave *i2c)
131
__func__);
132
}
133
134
- pca9552_autoinc(s);
135
+ pca955x_autoinc(s);
136
137
return ret;
138
}
139
140
-static int pca9552_send(I2CSlave *i2c, uint8_t data)
141
+static int pca955x_send(I2CSlave *i2c, uint8_t data)
142
{
143
- PCA9552State *s = PCA9552(i2c);
144
+ PCA955xState *s = PCA955X(i2c);
145
146
/* First byte sent by is the register address */
147
if (s->len == 0) {
148
s->pointer = data;
149
s->len++;
150
} else {
151
- pca9552_write(s, s->pointer & 0xf, data);
152
+ pca955x_write(s, s->pointer & 0xf, data);
153
154
- pca9552_autoinc(s);
155
+ pca955x_autoinc(s);
156
}
157
158
return 0;
159
}
160
161
-static int pca9552_event(I2CSlave *i2c, enum i2c_event event)
162
+static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
163
{
164
- PCA9552State *s = PCA9552(i2c);
165
+ PCA955xState *s = PCA955X(i2c);
166
167
s->len = 0;
168
return 0;
169
}
170
171
-static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
172
+static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
173
void *opaque, Error **errp)
174
{
175
- PCA9552State *s = PCA9552(obj);
176
+ PCA955xState *s = PCA955X(obj);
177
int led, rc, reg;
178
uint8_t state;
179
180
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
181
* reading the INPUTx reg
182
*/
183
reg = PCA9552_LS0 + led / 4;
184
- state = (pca9552_read(s, reg) >> (led % 8)) & 0x3;
185
+ state = (pca955x_read(s, reg) >> (led % 8)) & 0x3;
186
visit_type_str(v, name, (char **)&led_state[state], errp);
187
}
188
189
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
190
((state & 0x3) << (led_num << 1));
191
}
192
193
-static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
194
+static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
195
void *opaque, Error **errp)
196
{
197
- PCA9552State *s = PCA9552(obj);
198
+ PCA955xState *s = PCA955X(obj);
199
Error *local_err = NULL;
200
int led, rc, reg, val;
201
uint8_t state;
202
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
203
}
204
205
reg = PCA9552_LS0 + led / 4;
206
- val = pca9552_read(s, reg);
207
+ val = pca955x_read(s, reg);
208
val = pca955x_ledsel(val, led % 4, state);
209
- pca9552_write(s, reg, val);
210
+ pca955x_write(s, reg, val);
211
}
212
213
static const VMStateDescription pca9552_vmstate = {
214
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription pca9552_vmstate = {
215
.version_id = 0,
216
.minimum_version_id = 0,
217
.fields = (VMStateField[]) {
218
- VMSTATE_UINT8(len, PCA9552State),
219
- VMSTATE_UINT8(pointer, PCA9552State),
220
- VMSTATE_UINT8_ARRAY(regs, PCA9552State, PCA9552_NR_REGS),
221
- VMSTATE_I2C_SLAVE(i2c, PCA9552State),
222
+ VMSTATE_UINT8(len, PCA955xState),
223
+ VMSTATE_UINT8(pointer, PCA955xState),
224
+ VMSTATE_UINT8_ARRAY(regs, PCA955xState, PCA955X_NR_REGS),
225
+ VMSTATE_I2C_SLAVE(i2c, PCA955xState),
226
VMSTATE_END_OF_LIST()
227
}
228
};
229
230
static void pca9552_reset(DeviceState *dev)
231
{
232
- PCA9552State *s = PCA9552(dev);
233
+ PCA955xState *s = PCA955X(dev);
234
235
s->regs[PCA9552_PSC0] = 0xFF;
236
s->regs[PCA9552_PWM0] = 0x80;
237
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
238
s->regs[PCA9552_LS2] = 0x55;
239
s->regs[PCA9552_LS3] = 0x55;
240
241
- pca9552_update_pin_input(s);
242
+ pca955x_update_pin_input(s);
243
244
s->pointer = 0xFF;
245
s->len = 0;
246
}
247
248
-static void pca9552_initfn(Object *obj)
249
+static void pca955x_initfn(Object *obj)
250
{
251
- PCA9552State *s = PCA9552(obj);
252
+ PCA955xState *s = PCA955X(obj);
253
int led;
254
255
/* If support for the other PCA955X devices are implemented, these
256
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
257
char *name;
258
259
name = g_strdup_printf("led%d", led);
260
- object_property_add(obj, name, "bool", pca9552_get_led, pca9552_set_led,
261
+ object_property_add(obj, name, "bool", pca955x_get_led, pca955x_set_led,
262
NULL, NULL);
263
g_free(name);
264
}
265
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
266
DeviceClass *dc = DEVICE_CLASS(klass);
267
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
268
269
- k->event = pca9552_event;
270
- k->recv = pca9552_recv;
271
- k->send = pca9552_send;
272
+ k->event = pca955x_event;
273
+ k->recv = pca955x_recv;
274
+ k->send = pca955x_send;
275
dc->reset = pca9552_reset;
276
dc->vmsd = &pca9552_vmstate;
277
}
278
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
279
static const TypeInfo pca9552_info = {
280
.name = TYPE_PCA9552,
281
.parent = TYPE_I2C_SLAVE,
282
- .instance_init = pca9552_initfn,
283
- .instance_size = sizeof(PCA9552State),
284
+ .instance_init = pca955x_initfn,
285
+ .instance_size = sizeof(PCA955xState),
286
.class_init = pca9552_class_init,
287
};
288
289
-static void pca9552_register_types(void)
290
+static void pca955x_register_types(void)
291
{
292
type_register_static(&pca9552_info);
293
}
294
295
-type_init(pca9552_register_types)
296
+type_init(pca955x_register_types)
297
--
298
2.20.1
299
300
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Extract the code common to the PCA955x family in PCA955xClass,
4
keeping the PCA9552 specific parts into pca9552_class_init().
5
Remove the 'TODO' comment added in commit 5141d4158cf.
6
7
Suggested-by: Cédric Le Goater <clg@kaod.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-5-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 6 ++--
15
hw/misc/pca9552.c | 66 ++++++++++++++++++++++++++++-----------
16
2 files changed, 51 insertions(+), 21 deletions(-)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@
23
#include "hw/i2c/i2c.h"
24
25
#define TYPE_PCA9552 "pca9552"
26
-#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
+#define TYPE_PCA955X "pca955x"
28
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA955X)
29
30
#define PCA955X_NR_REGS 10
31
+#define PCA955X_PIN_COUNT_MAX 16
32
33
typedef struct PCA955xState {
34
/*< private >*/
35
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
36
uint8_t pointer;
37
38
uint8_t regs[PCA955X_NR_REGS];
39
- uint8_t max_reg;
40
- uint8_t pin_count;
41
} PCA955xState;
42
43
#endif
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
* https://www.nxp.com/docs/en/application-note/AN264.pdf
50
*
51
* Copyright (c) 2017-2018, IBM Corporation.
52
+ * Copyright (c) 2020 Philippe Mathieu-Daudé
53
*
54
* This work is licensed under the terms of the GNU GPL, version 2 or
55
* later. See the COPYING file in the top-level directory.
56
@@ -XXX,XX +XXX,XX @@
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
60
+typedef struct PCA955xClass {
61
+ /*< private >*/
62
+ I2CSlaveClass parent_class;
63
+ /*< public >*/
64
+
65
+ uint8_t pin_count;
66
+ uint8_t max_reg;
67
+} PCA955xClass;
68
+
69
+#define PCA955X_CLASS(klass) \
70
+ OBJECT_CLASS_CHECK(PCA955xClass, (klass), TYPE_PCA955X)
71
+#define PCA955X_GET_CLASS(obj) \
72
+ OBJECT_GET_CLASS(PCA955xClass, (obj), TYPE_PCA955X)
73
+
74
#define PCA9552_LED_ON 0x0
75
#define PCA9552_LED_OFF 0x1
76
#define PCA9552_LED_PWM0 0x2
77
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
78
79
static void pca955x_update_pin_input(PCA955xState *s)
80
{
81
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
82
int i;
83
84
- for (i = 0; i < s->pin_count; i++) {
85
+ for (i = 0; i < k->pin_count; i++) {
86
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
87
uint8_t input_shift = (i % 8);
88
uint8_t config = pca955x_pin_get_config(s, i);
89
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
90
*/
91
static void pca955x_autoinc(PCA955xState *s)
92
{
93
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
94
+
95
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
96
uint8_t reg = s->pointer & 0xf;
97
98
- reg = (reg + 1) % (s->max_reg + 1);
99
+ reg = (reg + 1) % (k->max_reg + 1);
100
s->pointer = reg | PCA9552_AUTOINC;
101
}
102
}
103
@@ -XXX,XX +XXX,XX @@ static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
104
static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
105
void *opaque, Error **errp)
106
{
107
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
108
PCA955xState *s = PCA955X(obj);
109
int led, rc, reg;
110
uint8_t state;
111
@@ -XXX,XX +XXX,XX @@ static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
112
error_setg(errp, "%s: error reading %s", __func__, name);
113
return;
114
}
115
- if (led < 0 || led > s->pin_count) {
116
+ if (led < 0 || led > k->pin_count) {
117
error_setg(errp, "%s invalid led %s", __func__, name);
118
return;
119
}
120
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
121
static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
122
void *opaque, Error **errp)
123
{
124
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
125
PCA955xState *s = PCA955X(obj);
126
Error *local_err = NULL;
127
int led, rc, reg, val;
128
@@ -XXX,XX +XXX,XX @@ static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
129
error_setg(errp, "%s: error reading %s", __func__, name);
130
return;
131
}
132
- if (led < 0 || led > s->pin_count) {
133
+ if (led < 0 || led > k->pin_count) {
134
error_setg(errp, "%s invalid led %s", __func__, name);
135
return;
136
}
137
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
138
139
static void pca955x_initfn(Object *obj)
140
{
141
- PCA955xState *s = PCA955X(obj);
142
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
143
int led;
144
145
- /* If support for the other PCA955X devices are implemented, these
146
- * constant values might be part of class structure describing the
147
- * PCA955X device
148
- */
149
- s->max_reg = PCA9552_LS3;
150
- s->pin_count = 16;
151
-
152
- for (led = 0; led < s->pin_count; led++) {
153
+ assert(k->pin_count <= PCA955X_PIN_COUNT_MAX);
154
+ for (led = 0; led < k->pin_count; led++) {
155
char *name;
156
157
name = g_strdup_printf("led%d", led);
158
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
159
}
160
}
161
162
-static void pca9552_class_init(ObjectClass *klass, void *data)
163
+static void pca955x_class_init(ObjectClass *klass, void *data)
164
{
165
- DeviceClass *dc = DEVICE_CLASS(klass);
166
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
167
168
k->event = pca955x_event;
169
k->recv = pca955x_recv;
170
k->send = pca955x_send;
171
+}
172
+
173
+static const TypeInfo pca955x_info = {
174
+ .name = TYPE_PCA955X,
175
+ .parent = TYPE_I2C_SLAVE,
176
+ .instance_init = pca955x_initfn,
177
+ .instance_size = sizeof(PCA955xState),
178
+ .class_init = pca955x_class_init,
179
+ .abstract = true,
180
+};
181
+
182
+static void pca9552_class_init(ObjectClass *oc, void *data)
183
+{
184
+ DeviceClass *dc = DEVICE_CLASS(oc);
185
+ PCA955xClass *pc = PCA955X_CLASS(oc);
186
+
187
dc->reset = pca9552_reset;
188
dc->vmsd = &pca9552_vmstate;
189
+ pc->max_reg = PCA9552_LS3;
190
+ pc->pin_count = 16;
191
}
192
193
static const TypeInfo pca9552_info = {
194
.name = TYPE_PCA9552,
195
- .parent = TYPE_I2C_SLAVE,
196
- .instance_init = pca955x_initfn,
197
- .instance_size = sizeof(PCA955xState),
198
+ .parent = TYPE_PCA955X,
199
.class_init = pca9552_class_init,
200
};
201
202
static void pca955x_register_types(void)
203
{
204
+ type_register_static(&pca955x_info);
205
type_register_static(&pca9552_info);
206
}
207
208
--
209
2.20.1
210
211
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Add a description field to distinguish between multiple devices.
4
5
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20200623072723.6324-6-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
include/hw/misc/pca9552.h | 1 +
12
hw/misc/pca9552.c | 18 ++++++++++++++++++
13
2 files changed, 19 insertions(+)
14
15
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
16
index XXXXXXX..XXXXXXX 100644
17
--- a/include/hw/misc/pca9552.h
18
+++ b/include/hw/misc/pca9552.h
19
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
20
uint8_t pointer;
21
22
uint8_t regs[PCA955X_NR_REGS];
23
+ char *description; /* For debugging purpose only */
24
} PCA955xState;
25
26
#endif
27
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/misc/pca9552.c
30
+++ b/hw/misc/pca9552.c
31
@@ -XXX,XX +XXX,XX @@
32
#include "qemu/osdep.h"
33
#include "qemu/log.h"
34
#include "qemu/module.h"
35
+#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
#include "migration/vmstate.h"
39
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
40
}
41
}
42
43
+static void pca955x_realize(DeviceState *dev, Error **errp)
44
+{
45
+ PCA955xState *s = PCA955X(dev);
46
+
47
+ if (!s->description) {
48
+ s->description = g_strdup("pca-unspecified");
49
+ }
50
+}
51
+
52
+static Property pca955x_properties[] = {
53
+ DEFINE_PROP_STRING("description", PCA955xState, description),
54
+ DEFINE_PROP_END_OF_LIST(),
55
+};
56
+
57
static void pca955x_class_init(ObjectClass *klass, void *data)
58
{
59
+ DeviceClass *dc = DEVICE_CLASS(klass);
60
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
61
62
k->event = pca955x_event;
63
k->recv = pca955x_recv;
64
k->send = pca955x_send;
65
+ dc->realize = pca955x_realize;
66
+ device_class_set_props(dc, pca955x_properties);
67
}
68
69
static const TypeInfo pca955x_info = {
70
--
71
2.20.1
72
73
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Add a trivial representation of the PCA9552 GPIOs.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_status
8
1592689902.327837:pca955x_gpio_status pca-unspecified GPIOs 0-15 [*...............]
9
1592689902.329934:pca955x_gpio_status pca-unspecified GPIOs 0-15 [**..............]
10
1592689902.330717:pca955x_gpio_status pca-unspecified GPIOs 0-15 [***.............]
11
1592689902.331431:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****............]
12
1592689902.332163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*..]
13
1592689902.332888:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........**.]
14
1592689902.333629:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
15
1592690032.793289:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
16
1592690033.303163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
17
1592690033.812962:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
18
1592690034.323234:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
19
1592690034.832922:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
20
21
We notice the GPIO #14 (front-power LED) starts to blink.
22
23
This LED is described in the witherspoon device-tree [*]:
24
25
front-power {
26
retain-state-shutdown;
27
default-state = "keep";
28
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
29
};
30
31
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
32
33
Suggested-by: Cédric Le Goater <clg@kaod.org>
34
Reviewed-by: Cédric Le Goater <clg@kaod.org>
35
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
36
Tested-by: Cédric Le Goater <clg@kaod.org>
37
Message-id: 20200623072723.6324-7-f4bug@amsat.org
38
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
39
---
40
hw/misc/pca9552.c | 39 +++++++++++++++++++++++++++++++++++++++
41
hw/misc/trace-events | 3 +++
42
2 files changed, 42 insertions(+)
43
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
#include "qemu/osdep.h"
50
#include "qemu/log.h"
51
#include "qemu/module.h"
52
+#include "qemu/bitops.h"
53
#include "hw/qdev-properties.h"
54
#include "hw/misc/pca9552.h"
55
#include "hw/misc/pca9552_regs.h"
56
#include "migration/vmstate.h"
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
+#include "trace.h"
60
61
typedef struct PCA955xClass {
62
/*< private >*/
63
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
64
return extract32(s->regs[reg], shift, 2);
65
}
66
67
+/* Return INPUT status (bit #N belongs to GPIO #N) */
68
+static uint16_t pca955x_pins_get_status(PCA955xState *s)
69
+{
70
+ return (s->regs[PCA9552_INPUT1] << 8) | s->regs[PCA9552_INPUT0];
71
+}
72
+
73
+static void pca955x_display_pins_status(PCA955xState *s,
74
+ uint16_t previous_pins_status)
75
+{
76
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
77
+ uint16_t pins_status, pins_changed;
78
+ int i;
79
+
80
+ pins_status = pca955x_pins_get_status(s);
81
+ pins_changed = previous_pins_status ^ pins_status;
82
+ if (!pins_changed) {
83
+ return;
84
+ }
85
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_STATUS)) {
86
+ char *buf = g_newa(char, k->pin_count + 1);
87
+
88
+ for (i = 0; i < k->pin_count; i++) {
89
+ if (extract32(pins_status, i, 1)) {
90
+ buf[i] = '*';
91
+ } else {
92
+ buf[i] = '.';
93
+ }
94
+ }
95
+ buf[i] = '\0';
96
+ trace_pca955x_gpio_status(s->description, buf);
97
+ }
98
+}
99
+
100
static void pca955x_update_pin_input(PCA955xState *s)
101
{
102
PCA955xClass *k = PCA955X_GET_CLASS(s);
103
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
104
105
static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
106
{
107
+ uint16_t pins_status;
108
+
109
switch (reg) {
110
case PCA9552_PSC0:
111
case PCA9552_PWM0:
112
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
113
case PCA9552_LS1:
114
case PCA9552_LS2:
115
case PCA9552_LS3:
116
+ pins_status = pca955x_pins_get_status(s);
117
s->regs[reg] = data;
118
pca955x_update_pin_input(s);
119
+ pca955x_display_pins_status(s, pins_status);
120
break;
121
122
case PCA9552_INPUT0:
123
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
124
index XXXXXXX..XXXXXXX 100644
125
--- a/hw/misc/trace-events
126
+++ b/hw/misc/trace-events
127
@@ -XXX,XX +XXX,XX @@ via1_adb_poll(uint8_t data, const char *vadbint, int status, int index, int size
128
# grlib_ahb_apb_pnp.c
129
grlib_ahb_pnp_read(uint64_t addr, uint32_t value) "AHB PnP read addr:0x%03"PRIx64" data:0x%08x"
130
grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx64" data:0x%08x"
131
+
132
+# pca9552.c
133
+pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
134
--
135
2.20.1
136
137
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
We have 2 distinct PCA9552 devices. Set their description
4
to distinguish them when looking at the trace events.
5
6
Description name taken from:
7
https://github.com/open-power/witherspoon-xml/blob/master/witherspoon.xml
8
9
Reviewed-by: Cédric Le Goater <clg@kaod.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Corey Minyard <cminyard@mvista.com>
12
Reviewed-by: Markus Armbruster <armbru@redhat.com>
13
Tested-by: Cédric Le Goater <clg@kaod.org>
14
Message-id: 20200623072723.6324-8-f4bug@amsat.org
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
---
17
hw/arm/aspeed.c | 13 +++++++++----
18
1 file changed, 9 insertions(+), 4 deletions(-)
19
20
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/hw/arm/aspeed.c
23
+++ b/hw/arm/aspeed.c
24
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
25
{
26
AspeedSoCState *soc = &bmc->soc;
27
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
28
+ DeviceState *dev;
29
30
/* Bus 3: TODO bmp280@77 */
31
/* Bus 3: TODO max31785@52 */
32
/* Bus 3: TODO dps310@76 */
33
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
34
- 0x60);
35
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
36
+ qdev_prop_set_string(dev, "description", "pca1");
37
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3),
38
+ &error_fatal);
39
40
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
41
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
42
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
43
44
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
45
eeprom_buf);
46
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
47
- 0x60);
48
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
49
+ qdev_prop_set_string(dev, "description", "pca0");
50
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11),
51
+ &error_fatal);
52
/* Bus 11: TODO ucd90160@64 */
53
}
54
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Emit a trace event when a GPIO change its state.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_change
8
1592690552.687372:pca955x_gpio_change pca1 GPIO id:0 status: 0 -> 1
9
1592690552.690169:pca955x_gpio_change pca1 GPIO id:1 status: 0 -> 1
10
1592690552.691673:pca955x_gpio_change pca1 GPIO id:2 status: 0 -> 1
11
1592690552.696886:pca955x_gpio_change pca1 GPIO id:3 status: 0 -> 1
12
1592690552.698614:pca955x_gpio_change pca1 GPIO id:13 status: 0 -> 1
13
1592690552.699833:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
14
1592690552.700842:pca955x_gpio_change pca1 GPIO id:15 status: 0 -> 1
15
1592690683.841921:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
16
1592690683.861660:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
17
1592690684.371460:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
18
1592690684.882115:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
19
1592690685.391411:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
20
1592690685.901391:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
21
1592690686.411678:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
22
1592690686.921279:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
23
24
We notice the GPIO #14 (front-power LED) starts to blink.
25
26
This LED is described in the witherspoon device-tree [*]:
27
28
front-power {
29
retain-state-shutdown;
30
default-state = "keep";
31
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
32
};
33
34
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
35
36
Reviewed-by: Cédric Le Goater <clg@kaod.org>
37
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
38
Tested-by: Cédric Le Goater <clg@kaod.org>
39
Message-id: 20200623072723.6324-9-f4bug@amsat.org
40
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
41
---
42
hw/misc/pca9552.c | 15 +++++++++++++++
43
hw/misc/trace-events | 1 +
44
2 files changed, 16 insertions(+)
45
46
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/hw/misc/pca9552.c
49
+++ b/hw/misc/pca9552.c
50
@@ -XXX,XX +XXX,XX @@ static void pca955x_display_pins_status(PCA955xState *s,
51
buf[i] = '\0';
52
trace_pca955x_gpio_status(s->description, buf);
53
}
54
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_CHANGE)) {
55
+ for (i = 0; i < k->pin_count; i++) {
56
+ if (extract32(pins_changed, i, 1)) {
57
+ unsigned new_state = extract32(pins_status, i, 1);
58
+
59
+ /*
60
+ * We display the state using the PCA logic ("active-high").
61
+ * This is not the state of the LED, which signal might be
62
+ * wired "active-low" on the board.
63
+ */
64
+ trace_pca955x_gpio_change(s->description, i,
65
+ !new_state, new_state);
66
+ }
67
+ }
68
+ }
69
}
70
71
static void pca955x_update_pin_input(PCA955xState *s)
72
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/misc/trace-events
75
+++ b/hw/misc/trace-events
76
@@ -XXX,XX +XXX,XX @@ grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx6
77
78
# pca9552.c
79
pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
80
+pca955x_gpio_change(const char *description, unsigned id, unsigned prev_state, unsigned current_state) "%s GPIO id:%u status: %u -> %u"
81
--
82
2.20.1
83
84
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
The PCA9552 has 16 GPIOs which can be used as input,
4
output or PWM mode. QEMU models the output GPIO with
5
the qemu_irq type. Let the device expose the 16 GPIOs
6
to allow us to later connect LEDs to these outputs.
7
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-10-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 1 +
15
hw/misc/pca9552.c | 6 ++++++
16
2 files changed, 7 insertions(+)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
23
uint8_t pointer;
24
25
uint8_t regs[PCA955X_NR_REGS];
26
+ qemu_irq gpio[PCA955X_PIN_COUNT_MAX];
27
char *description; /* For debugging purpose only */
28
} PCA955xState;
29
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@
35
#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
+#include "hw/irq.h"
39
#include "migration/vmstate.h"
40
#include "qapi/error.h"
41
#include "qapi/visitor.h"
42
@@ -XXX,XX +XXX,XX @@ static void pca955x_update_pin_input(PCA955xState *s)
43
44
switch (config) {
45
case PCA9552_LED_ON:
46
+ qemu_set_irq(s->gpio[i], 1);
47
s->regs[input_reg] |= 1 << input_shift;
48
break;
49
case PCA9552_LED_OFF:
50
+ qemu_set_irq(s->gpio[i], 0);
51
s->regs[input_reg] &= ~(1 << input_shift);
52
break;
53
case PCA9552_LED_PWM0:
54
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
55
56
static void pca955x_realize(DeviceState *dev, Error **errp)
57
{
58
+ PCA955xClass *k = PCA955X_GET_CLASS(dev);
59
PCA955xState *s = PCA955X(dev);
60
61
if (!s->description) {
62
s->description = g_strdup("pca-unspecified");
63
}
64
+
65
+ qdev_init_gpio_out(dev, s->gpio, k->pin_count);
66
}
67
68
static Property pca955x_properties[] = {
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This controls whether the PACI{A,B}SP instructions trap with BTYPE=3
4
(indirect branch from register other than x16/x17). The linux kernel
5
sets this in bti_enable().
6
7
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/998
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20220427042312.294300-1-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-10-richard.henderson@linaro.org
11
[PMM: remove stray change to makefile comment]
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
13
---
8
target/arm/helper-a64.h | 2 ++
14
target/arm/cpu.c | 2 ++
9
target/arm/internals.h | 5 +++
15
tests/tcg/aarch64/bti-3.c | 42 +++++++++++++++++++++++++++++++
10
target/arm/mte_helper.c | 72 ++++++++++++++++++++++++++++++++++++++
16
tests/tcg/aarch64/Makefile.target | 6 ++---
11
target/arm/translate-a64.c | 18 ++++++++++
17
3 files changed, 47 insertions(+), 3 deletions(-)
12
target/arm/Makefile.objs | 1 +
18
create mode 100644 tests/tcg/aarch64/bti-3.c
13
5 files changed, 98 insertions(+)
14
create mode 100644 target/arm/mte_helper.c
15
19
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
17
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-a64.h
22
--- a/target/arm/cpu.c
19
+++ b/target/arm/helper-a64.h
23
+++ b/target/arm/cpu.c
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autda, TCG_CALL_NO_WG, i64, env, i64, i64)
24
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
21
DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
25
/* Enable all PAC keys. */
22
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
env->cp15.sctlr_el[1] |= (SCTLR_EnIA | SCTLR_EnIB |
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
27
SCTLR_EnDA | SCTLR_EnDB);
24
+
28
+ /* Trap on btype=3 for PACIxSP. */
25
+DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
29
+ env->cp15.sctlr_el[1] |= SCTLR_BT0;
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
30
/* and to the FP/Neon instructions */
27
index XXXXXXX..XXXXXXX 100644
31
env->cp15.cpacr_el1 = deposit64(env->cp15.cpacr_el1, 20, 2, 3);
28
--- a/target/arm/internals.h
32
/* and to the SVE instructions */
29
+++ b/target/arm/internals.h
33
diff --git a/tests/tcg/aarch64/bti-3.c b/tests/tcg/aarch64/bti-3.c
30
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
31
*/
32
#define GMID_EL1_BS 6
33
34
+static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
35
+{
36
+ return deposit64(ptr, 56, 4, rtag);
37
+}
38
+
39
#endif
40
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
41
new file mode 100644
34
new file mode 100644
42
index XXXXXXX..XXXXXXX
35
index XXXXXXX..XXXXXXX
43
--- /dev/null
36
--- /dev/null
44
+++ b/target/arm/mte_helper.c
37
+++ b/tests/tcg/aarch64/bti-3.c
45
@@ -XXX,XX +XXX,XX @@
38
@@ -XXX,XX +XXX,XX @@
46
+/*
39
+/*
47
+ * ARM v8.5-MemTag Operations
40
+ * BTI vs PACIASP
48
+ *
49
+ * Copyright (c) 2020 Linaro, Ltd.
50
+ *
51
+ * This library is free software; you can redistribute it and/or
52
+ * modify it under the terms of the GNU Lesser General Public
53
+ * License as published by the Free Software Foundation; either
54
+ * version 2.1 of the License, or (at your option) any later version.
55
+ *
56
+ * This library is distributed in the hope that it will be useful,
57
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
58
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
59
+ * Lesser General Public License for more details.
60
+ *
61
+ * You should have received a copy of the GNU Lesser General Public
62
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
63
+ */
41
+ */
64
+
42
+
65
+#include "qemu/osdep.h"
43
+#include "bti-crt.inc.c"
66
+#include "cpu.h"
67
+#include "internals.h"
68
+#include "exec/exec-all.h"
69
+#include "exec/cpu_ldst.h"
70
+#include "exec/helper-proto.h"
71
+
44
+
72
+
45
+static void skip2_sigill(int sig, siginfo_t *info, ucontext_t *uc)
73
+static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
74
+{
46
+{
75
+ if (exclude == 0xffff) {
47
+ uc->uc_mcontext.pc += 8;
76
+ return 0;
48
+ uc->uc_mcontext.pstate = 1;
77
+ }
78
+ if (offset == 0) {
79
+ while (exclude & (1 << tag)) {
80
+ tag = (tag + 1) & 15;
81
+ }
82
+ } else {
83
+ do {
84
+ do {
85
+ tag = (tag + 1) & 15;
86
+ } while (exclude & (1 << tag));
87
+ } while (--offset > 0);
88
+ }
89
+ return tag;
90
+}
49
+}
91
+
50
+
92
+uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
51
+#define BTYPE_1() \
52
+ asm("mov %0,#1; adr x16, 1f; br x16; 1: hint #25; mov %0,#0" \
53
+ : "=r"(skipped) : : "x16", "x30")
54
+
55
+#define BTYPE_2() \
56
+ asm("mov %0,#1; adr x16, 1f; blr x16; 1: hint #25; mov %0,#0" \
57
+ : "=r"(skipped) : : "x16", "x30")
58
+
59
+#define BTYPE_3() \
60
+ asm("mov %0,#1; adr x15, 1f; br x15; 1: hint #25; mov %0,#0" \
61
+ : "=r"(skipped) : : "x15", "x30")
62
+
63
+#define TEST(WHICH, EXPECT) \
64
+ do { WHICH(); fail += skipped ^ EXPECT; } while (0)
65
+
66
+int main()
93
+{
67
+{
94
+ int rtag;
68
+ int fail = 0;
69
+ int skipped;
95
+
70
+
96
+ /*
71
+ /* Signal-like with SA_SIGINFO. */
97
+ * Our IMPDEF choice for GCR_EL1.RRND==1 is to behave as if
72
+ signal_info(SIGILL, skip2_sigill);
98
+ * GCR_EL1.RRND==0, always producing deterministic results.
99
+ */
100
+ uint16_t exclude = extract32(rm | env->cp15.gcr_el1, 0, 16);
101
+ int start = extract32(env->cp15.rgsr_el1, 0, 4);
102
+ int seed = extract32(env->cp15.rgsr_el1, 8, 16);
103
+ int offset, i;
104
+
73
+
105
+ /* RandomTag */
74
+ /* With SCTLR_EL1.BT0 set, PACIASP is not compatible with type=3. */
106
+ for (i = offset = 0; i < 4; ++i) {
75
+ TEST(BTYPE_1, 0);
107
+ /* NextRandomTagBit */
76
+ TEST(BTYPE_2, 0);
108
+ int top = (extract32(seed, 5, 1) ^ extract32(seed, 3, 1) ^
77
+ TEST(BTYPE_3, 1);
109
+ extract32(seed, 2, 1) ^ extract32(seed, 0, 1));
110
+ seed = (top << 15) | (seed >> 1);
111
+ offset |= top << i;
112
+ }
113
+ rtag = choose_nonexcluded_tag(start, offset, exclude);
114
+ env->cp15.rgsr_el1 = rtag | (seed << 8);
115
+
78
+
116
+ return address_with_allocation_tag(rn, rtag);
79
+ return fail;
117
+}
80
+}
118
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
81
diff --git a/tests/tcg/aarch64/Makefile.target b/tests/tcg/aarch64/Makefile.target
119
index XXXXXXX..XXXXXXX 100644
82
index XXXXXXX..XXXXXXX 100644
120
--- a/target/arm/translate-a64.c
83
--- a/tests/tcg/aarch64/Makefile.target
121
+++ b/target/arm/translate-a64.c
84
+++ b/tests/tcg/aarch64/Makefile.target
122
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
85
@@ -XXX,XX +XXX,XX @@ endif
123
return clean;
86
# BTI Tests
124
}
87
# bti-1 tests the elf notes, so we require special compiler support.
125
88
ifneq ($(CROSS_CC_HAS_ARMV8_BTI),)
126
+/* Insert a zero tag into src, with the result at dst. */
89
-AARCH64_TESTS += bti-1
127
+static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
90
-bti-1: CFLAGS += -mbranch-protection=standard
128
+{
91
-bti-1: LDFLAGS += -nostdlib
129
+ tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
92
+AARCH64_TESTS += bti-1 bti-3
130
+}
93
+bti-1 bti-3: CFLAGS += -mbranch-protection=standard
131
+
94
+bti-1 bti-3: LDFLAGS += -nostdlib
132
typedef struct DisasCompare64 {
95
endif
133
TCGCond cond;
96
# bti-2 tests PROT_BTI, so no special compiler support required.
134
TCGv_i64 value;
97
AARCH64_TESTS += bti-2
135
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
136
case 3: /* SDIV */
137
handle_div(s, true, sf, rm, rn, rd);
138
break;
139
+ case 4: /* IRG */
140
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
141
+ goto do_unallocated;
142
+ }
143
+ if (s->ata) {
144
+ gen_helper_irg(cpu_reg_sp(s, rd), cpu_env,
145
+ cpu_reg_sp(s, rn), cpu_reg(s, rm));
146
+ } else {
147
+ gen_address_with_allocation_tag0(cpu_reg_sp(s, rd),
148
+ cpu_reg_sp(s, rn));
149
+ }
150
+ break;
151
case 8: /* LSLV */
152
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
153
break;
154
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
155
index XXXXXXX..XXXXXXX 100644
156
--- a/target/arm/Makefile.objs
157
+++ b/target/arm/Makefile.objs
158
@@ -XXX,XX +XXX,XX @@ obj-$(CONFIG_SOFTMMU) += psci.o
159
obj-$(TARGET_AARCH64) += translate-a64.o helper-a64.o
160
obj-$(TARGET_AARCH64) += translate-sve.o sve_helper.o
161
obj-$(TARGET_AARCH64) += pauth_helper.o
162
+obj-$(TARGET_AARCH64) += mte_helper.o
163
--
98
--
164
2.20.1
99
2.25.1
165
166
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is DC GVA and DC GZVA, and the tag check for DC ZVA.
3
Move ARMCPRegInfo and all related declarations to a new
4
internal header, out of the public cpu.h.
4
5
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-40-richard.henderson@linaro.org
9
Message-id: 20220501055028.646596-2-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/cpu.h | 4 +++-
12
target/arm/cpregs.h | 413 +++++++++++++++++++++++++++++++++++++
11
target/arm/helper.c | 16 ++++++++++++++++
13
target/arm/cpu.h | 368 ---------------------------------
12
target/arm/translate-a64.c | 39 ++++++++++++++++++++++++++++++++++++++
14
hw/arm/pxa2xx.c | 1 +
13
3 files changed, 58 insertions(+), 1 deletion(-)
15
hw/arm/pxa2xx_pic.c | 1 +
16
hw/intc/arm_gicv3_cpuif.c | 1 +
17
hw/intc/arm_gicv3_kvm.c | 2 +
18
target/arm/cpu.c | 1 +
19
target/arm/cpu64.c | 1 +
20
target/arm/cpu_tcg.c | 1 +
21
target/arm/gdbstub.c | 3 +-
22
target/arm/helper.c | 1 +
23
target/arm/op_helper.c | 1 +
24
target/arm/translate-a64.c | 4 +-
25
target/arm/translate.c | 3 +-
26
14 files changed, 427 insertions(+), 374 deletions(-)
27
create mode 100644 target/arm/cpregs.h
14
28
29
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
30
new file mode 100644
31
index XXXXXXX..XXXXXXX
32
--- /dev/null
33
+++ b/target/arm/cpregs.h
34
@@ -XXX,XX +XXX,XX @@
35
+/*
36
+ * QEMU ARM CP Register access and descriptions
37
+ *
38
+ * Copyright (c) 2022 Linaro Ltd
39
+ *
40
+ * This program is free software; you can redistribute it and/or
41
+ * modify it under the terms of the GNU General Public License
42
+ * as published by the Free Software Foundation; either version 2
43
+ * of the License, or (at your option) any later version.
44
+ *
45
+ * This program is distributed in the hope that it will be useful,
46
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
47
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
48
+ * GNU General Public License for more details.
49
+ *
50
+ * You should have received a copy of the GNU General Public License
51
+ * along with this program; if not, see
52
+ * <http://www.gnu.org/licenses/gpl-2.0.html>
53
+ */
54
+
55
+#ifndef TARGET_ARM_CPREGS_H
56
+#define TARGET_ARM_CPREGS_H
57
+
58
+/*
59
+ * ARMCPRegInfo type field bits. If the SPECIAL bit is set this is a
60
+ * special-behaviour cp reg and bits [11..8] indicate what behaviour
61
+ * it has. Otherwise it is a simple cp reg, where CONST indicates that
62
+ * TCG can assume the value to be constant (ie load at translate time)
63
+ * and 64BIT indicates a 64 bit wide coprocessor register. SUPPRESS_TB_END
64
+ * indicates that the TB should not be ended after a write to this register
65
+ * (the default is that the TB ends after cp writes). OVERRIDE permits
66
+ * a register definition to override a previous definition for the
67
+ * same (cp, is64, crn, crm, opc1, opc2) tuple: either the new or the
68
+ * old must have the OVERRIDE bit set.
69
+ * ALIAS indicates that this register is an alias view of some underlying
70
+ * state which is also visible via another register, and that the other
71
+ * register is handling migration and reset; registers marked ALIAS will not be
72
+ * migrated but may have their state set by syncing of register state from KVM.
73
+ * NO_RAW indicates that this register has no underlying state and does not
74
+ * support raw access for state saving/loading; it will not be used for either
75
+ * migration or KVM state synchronization. (Typically this is for "registers"
76
+ * which are actually used as instructions for cache maintenance and so on.)
77
+ * IO indicates that this register does I/O and therefore its accesses
78
+ * need to be marked with gen_io_start() and also end the TB. In particular,
79
+ * registers which implement clocks or timers require this.
80
+ * RAISES_EXC is for when the read or write hook might raise an exception;
81
+ * the generated code will synchronize the CPU state before calling the hook
82
+ * so that it is safe for the hook to call raise_exception().
83
+ * NEWEL is for writes to registers that might change the exception
84
+ * level - typically on older ARM chips. For those cases we need to
85
+ * re-read the new el when recomputing the translation flags.
86
+ */
87
+#define ARM_CP_SPECIAL 0x0001
88
+#define ARM_CP_CONST 0x0002
89
+#define ARM_CP_64BIT 0x0004
90
+#define ARM_CP_SUPPRESS_TB_END 0x0008
91
+#define ARM_CP_OVERRIDE 0x0010
92
+#define ARM_CP_ALIAS 0x0020
93
+#define ARM_CP_IO 0x0040
94
+#define ARM_CP_NO_RAW 0x0080
95
+#define ARM_CP_NOP (ARM_CP_SPECIAL | 0x0100)
96
+#define ARM_CP_WFI (ARM_CP_SPECIAL | 0x0200)
97
+#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
98
+#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
99
+#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
100
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
101
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
102
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
103
+#define ARM_CP_FPU 0x1000
104
+#define ARM_CP_SVE 0x2000
105
+#define ARM_CP_NO_GDB 0x4000
106
+#define ARM_CP_RAISES_EXC 0x8000
107
+#define ARM_CP_NEWEL 0x10000
108
+/* Used only as a terminator for ARMCPRegInfo lists */
109
+#define ARM_CP_SENTINEL 0xfffff
110
+/* Mask of only the flag bits in a type field */
111
+#define ARM_CP_FLAG_MASK 0x1f0ff
112
+
113
+/*
114
+ * Valid values for ARMCPRegInfo state field, indicating which of
115
+ * the AArch32 and AArch64 execution states this register is visible in.
116
+ * If the reginfo doesn't explicitly specify then it is AArch32 only.
117
+ * If the reginfo is declared to be visible in both states then a second
118
+ * reginfo is synthesised for the AArch32 view of the AArch64 register,
119
+ * such that the AArch32 view is the lower 32 bits of the AArch64 one.
120
+ * Note that we rely on the values of these enums as we iterate through
121
+ * the various states in some places.
122
+ */
123
+enum {
124
+ ARM_CP_STATE_AA32 = 0,
125
+ ARM_CP_STATE_AA64 = 1,
126
+ ARM_CP_STATE_BOTH = 2,
127
+};
128
+
129
+/*
130
+ * ARM CP register secure state flags. These flags identify security state
131
+ * attributes for a given CP register entry.
132
+ * The existence of both or neither secure and non-secure flags indicates that
133
+ * the register has both a secure and non-secure hash entry. A single one of
134
+ * these flags causes the register to only be hashed for the specified
135
+ * security state.
136
+ * Although definitions may have any combination of the S/NS bits, each
137
+ * registered entry will only have one to identify whether the entry is secure
138
+ * or non-secure.
139
+ */
140
+enum {
141
+ ARM_CP_SECSTATE_S = (1 << 0), /* bit[0]: Secure state register */
142
+ ARM_CP_SECSTATE_NS = (1 << 1), /* bit[1]: Non-secure state register */
143
+};
144
+
145
+/*
146
+ * Return true if cptype is a valid type field. This is used to try to
147
+ * catch errors where the sentinel has been accidentally left off the end
148
+ * of a list of registers.
149
+ */
150
+static inline bool cptype_valid(int cptype)
151
+{
152
+ return ((cptype & ~ARM_CP_FLAG_MASK) == 0)
153
+ || ((cptype & ARM_CP_SPECIAL) &&
154
+ ((cptype & ~ARM_CP_FLAG_MASK) <= ARM_LAST_SPECIAL));
155
+}
156
+
157
+/*
158
+ * Access rights:
159
+ * We define bits for Read and Write access for what rev C of the v7-AR ARM ARM
160
+ * defines as PL0 (user), PL1 (fiq/irq/svc/abt/und/sys, ie privileged), and
161
+ * PL2 (hyp). The other level which has Read and Write bits is Secure PL1
162
+ * (ie any of the privileged modes in Secure state, or Monitor mode).
163
+ * If a register is accessible in one privilege level it's always accessible
164
+ * in higher privilege levels too. Since "Secure PL1" also follows this rule
165
+ * (ie anything visible in PL2 is visible in S-PL1, some things are only
166
+ * visible in S-PL1) but "Secure PL1" is a bit of a mouthful, we bend the
167
+ * terminology a little and call this PL3.
168
+ * In AArch64 things are somewhat simpler as the PLx bits line up exactly
169
+ * with the ELx exception levels.
170
+ *
171
+ * If access permissions for a register are more complex than can be
172
+ * described with these bits, then use a laxer set of restrictions, and
173
+ * do the more restrictive/complex check inside a helper function.
174
+ */
175
+#define PL3_R 0x80
176
+#define PL3_W 0x40
177
+#define PL2_R (0x20 | PL3_R)
178
+#define PL2_W (0x10 | PL3_W)
179
+#define PL1_R (0x08 | PL2_R)
180
+#define PL1_W (0x04 | PL2_W)
181
+#define PL0_R (0x02 | PL1_R)
182
+#define PL0_W (0x01 | PL1_W)
183
+
184
+/*
185
+ * For user-mode some registers are accessible to EL0 via a kernel
186
+ * trap-and-emulate ABI. In this case we define the read permissions
187
+ * as actually being PL0_R. However some bits of any given register
188
+ * may still be masked.
189
+ */
190
+#ifdef CONFIG_USER_ONLY
191
+#define PL0U_R PL0_R
192
+#else
193
+#define PL0U_R PL1_R
194
+#endif
195
+
196
+#define PL3_RW (PL3_R | PL3_W)
197
+#define PL2_RW (PL2_R | PL2_W)
198
+#define PL1_RW (PL1_R | PL1_W)
199
+#define PL0_RW (PL0_R | PL0_W)
200
+
201
+typedef enum CPAccessResult {
202
+ /* Access is permitted */
203
+ CP_ACCESS_OK = 0,
204
+ /*
205
+ * Access fails due to a configurable trap or enable which would
206
+ * result in a categorized exception syndrome giving information about
207
+ * the failing instruction (ie syndrome category 0x3, 0x4, 0x5, 0x6,
208
+ * 0xc or 0x18). The exception is taken to the usual target EL (EL1 or
209
+ * PL1 if in EL0, otherwise to the current EL).
210
+ */
211
+ CP_ACCESS_TRAP = 1,
212
+ /*
213
+ * Access fails and results in an exception syndrome 0x0 ("uncategorized").
214
+ * Note that this is not a catch-all case -- the set of cases which may
215
+ * result in this failure is specifically defined by the architecture.
216
+ */
217
+ CP_ACCESS_TRAP_UNCATEGORIZED = 2,
218
+ /* As CP_ACCESS_TRAP, but for traps directly to EL2 or EL3 */
219
+ CP_ACCESS_TRAP_EL2 = 3,
220
+ CP_ACCESS_TRAP_EL3 = 4,
221
+ /* As CP_ACCESS_UNCATEGORIZED, but for traps directly to EL2 or EL3 */
222
+ CP_ACCESS_TRAP_UNCATEGORIZED_EL2 = 5,
223
+ CP_ACCESS_TRAP_UNCATEGORIZED_EL3 = 6,
224
+} CPAccessResult;
225
+
226
+typedef struct ARMCPRegInfo ARMCPRegInfo;
227
+
228
+/*
229
+ * Access functions for coprocessor registers. These cannot fail and
230
+ * may not raise exceptions.
231
+ */
232
+typedef uint64_t CPReadFn(CPUARMState *env, const ARMCPRegInfo *opaque);
233
+typedef void CPWriteFn(CPUARMState *env, const ARMCPRegInfo *opaque,
234
+ uint64_t value);
235
+/* Access permission check functions for coprocessor registers. */
236
+typedef CPAccessResult CPAccessFn(CPUARMState *env,
237
+ const ARMCPRegInfo *opaque,
238
+ bool isread);
239
+/* Hook function for register reset */
240
+typedef void CPResetFn(CPUARMState *env, const ARMCPRegInfo *opaque);
241
+
242
+#define CP_ANY 0xff
243
+
244
+/* Definition of an ARM coprocessor register */
245
+struct ARMCPRegInfo {
246
+ /* Name of register (useful mainly for debugging, need not be unique) */
247
+ const char *name;
248
+ /*
249
+ * Location of register: coprocessor number and (crn,crm,opc1,opc2)
250
+ * tuple. Any of crm, opc1 and opc2 may be CP_ANY to indicate a
251
+ * 'wildcard' field -- any value of that field in the MRC/MCR insn
252
+ * will be decoded to this register. The register read and write
253
+ * callbacks will be passed an ARMCPRegInfo with the crn/crm/opc1/opc2
254
+ * used by the program, so it is possible to register a wildcard and
255
+ * then behave differently on read/write if necessary.
256
+ * For 64 bit registers, only crm and opc1 are relevant; crn and opc2
257
+ * must both be zero.
258
+ * For AArch64-visible registers, opc0 is also used.
259
+ * Since there are no "coprocessors" in AArch64, cp is purely used as a
260
+ * way to distinguish (for KVM's benefit) guest-visible system registers
261
+ * from demuxed ones provided to preserve the "no side effects on
262
+ * KVM register read/write from QEMU" semantics. cp==0x13 is guest
263
+ * visible (to match KVM's encoding); cp==0 will be converted to
264
+ * cp==0x13 when the ARMCPRegInfo is registered, for convenience.
265
+ */
266
+ uint8_t cp;
267
+ uint8_t crn;
268
+ uint8_t crm;
269
+ uint8_t opc0;
270
+ uint8_t opc1;
271
+ uint8_t opc2;
272
+ /* Execution state in which this register is visible: ARM_CP_STATE_* */
273
+ int state;
274
+ /* Register type: ARM_CP_* bits/values */
275
+ int type;
276
+ /* Access rights: PL*_[RW] */
277
+ int access;
278
+ /* Security state: ARM_CP_SECSTATE_* bits/values */
279
+ int secure;
280
+ /*
281
+ * The opaque pointer passed to define_arm_cp_regs_with_opaque() when
282
+ * this register was defined: can be used to hand data through to the
283
+ * register read/write functions, since they are passed the ARMCPRegInfo*.
284
+ */
285
+ void *opaque;
286
+ /*
287
+ * Value of this register, if it is ARM_CP_CONST. Otherwise, if
288
+ * fieldoffset is non-zero, the reset value of the register.
289
+ */
290
+ uint64_t resetvalue;
291
+ /*
292
+ * Offset of the field in CPUARMState for this register.
293
+ * This is not needed if either:
294
+ * 1. type is ARM_CP_CONST or one of the ARM_CP_SPECIALs
295
+ * 2. both readfn and writefn are specified
296
+ */
297
+ ptrdiff_t fieldoffset; /* offsetof(CPUARMState, field) */
298
+
299
+ /*
300
+ * Offsets of the secure and non-secure fields in CPUARMState for the
301
+ * register if it is banked. These fields are only used during the static
302
+ * registration of a register. During hashing the bank associated
303
+ * with a given security state is copied to fieldoffset which is used from
304
+ * there on out.
305
+ *
306
+ * It is expected that register definitions use either fieldoffset or
307
+ * bank_fieldoffsets in the definition but not both. It is also expected
308
+ * that both bank offsets are set when defining a banked register. This
309
+ * use indicates that a register is banked.
310
+ */
311
+ ptrdiff_t bank_fieldoffsets[2];
312
+
313
+ /*
314
+ * Function for making any access checks for this register in addition to
315
+ * those specified by the 'access' permissions bits. If NULL, no extra
316
+ * checks required. The access check is performed at runtime, not at
317
+ * translate time.
318
+ */
319
+ CPAccessFn *accessfn;
320
+ /*
321
+ * Function for handling reads of this register. If NULL, then reads
322
+ * will be done by loading from the offset into CPUARMState specified
323
+ * by fieldoffset.
324
+ */
325
+ CPReadFn *readfn;
326
+ /*
327
+ * Function for handling writes of this register. If NULL, then writes
328
+ * will be done by writing to the offset into CPUARMState specified
329
+ * by fieldoffset.
330
+ */
331
+ CPWriteFn *writefn;
332
+ /*
333
+ * Function for doing a "raw" read; used when we need to copy
334
+ * coprocessor state to the kernel for KVM or out for
335
+ * migration. This only needs to be provided if there is also a
336
+ * readfn and it has side effects (for instance clear-on-read bits).
337
+ */
338
+ CPReadFn *raw_readfn;
339
+ /*
340
+ * Function for doing a "raw" write; used when we need to copy KVM
341
+ * kernel coprocessor state into userspace, or for inbound
342
+ * migration. This only needs to be provided if there is also a
343
+ * writefn and it masks out "unwritable" bits or has write-one-to-clear
344
+ * or similar behaviour.
345
+ */
346
+ CPWriteFn *raw_writefn;
347
+ /*
348
+ * Function for resetting the register. If NULL, then reset will be done
349
+ * by writing resetvalue to the field specified in fieldoffset. If
350
+ * fieldoffset is 0 then no reset will be done.
351
+ */
352
+ CPResetFn *resetfn;
353
+
354
+ /*
355
+ * "Original" writefn and readfn.
356
+ * For ARMv8.1-VHE register aliases, we overwrite the read/write
357
+ * accessor functions of various EL1/EL0 to perform the runtime
358
+ * check for which sysreg should actually be modified, and then
359
+ * forwards the operation. Before overwriting the accessors,
360
+ * the original function is copied here, so that accesses that
361
+ * really do go to the EL1/EL0 version proceed normally.
362
+ * (The corresponding EL2 register is linked via opaque.)
363
+ */
364
+ CPReadFn *orig_readfn;
365
+ CPWriteFn *orig_writefn;
366
+};
367
+
368
+/*
369
+ * Macros which are lvalues for the field in CPUARMState for the
370
+ * ARMCPRegInfo *ri.
371
+ */
372
+#define CPREG_FIELD32(env, ri) \
373
+ (*(uint32_t *)((char *)(env) + (ri)->fieldoffset))
374
+#define CPREG_FIELD64(env, ri) \
375
+ (*(uint64_t *)((char *)(env) + (ri)->fieldoffset))
376
+
377
+#define REGINFO_SENTINEL { .type = ARM_CP_SENTINEL }
378
+
379
+void define_arm_cp_regs_with_opaque(ARMCPU *cpu,
380
+ const ARMCPRegInfo *regs, void *opaque);
381
+void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
382
+ const ARMCPRegInfo *regs, void *opaque);
383
+static inline void define_arm_cp_regs(ARMCPU *cpu, const ARMCPRegInfo *regs)
384
+{
385
+ define_arm_cp_regs_with_opaque(cpu, regs, 0);
386
+}
387
+static inline void define_one_arm_cp_reg(ARMCPU *cpu, const ARMCPRegInfo *regs)
388
+{
389
+ define_one_arm_cp_reg_with_opaque(cpu, regs, 0);
390
+}
391
+const ARMCPRegInfo *get_arm_cp_reginfo(GHashTable *cpregs, uint32_t encoded_cp);
392
+
393
+/*
394
+ * Definition of an ARM co-processor register as viewed from
395
+ * userspace. This is used for presenting sanitised versions of
396
+ * registers to userspace when emulating the Linux AArch64 CPU
397
+ * ID/feature ABI (advertised as HWCAP_CPUID).
398
+ */
399
+typedef struct ARMCPRegUserSpaceInfo {
400
+ /* Name of register */
401
+ const char *name;
402
+
403
+ /* Is the name actually a glob pattern */
404
+ bool is_glob;
405
+
406
+ /* Only some bits are exported to user space */
407
+ uint64_t exported_bits;
408
+
409
+ /* Fixed bits are applied after the mask */
410
+ uint64_t fixed_bits;
411
+} ARMCPRegUserSpaceInfo;
412
+
413
+#define REGUSERINFO_SENTINEL { .name = NULL }
414
+
415
+void modify_arm_cp_regs(ARMCPRegInfo *regs, const ARMCPRegUserSpaceInfo *mods);
416
+
417
+/* CPWriteFn that can be used to implement writes-ignored behaviour */
418
+void arm_cp_write_ignore(CPUARMState *env, const ARMCPRegInfo *ri,
419
+ uint64_t value);
420
+/* CPReadFn that can be used for read-as-zero behaviour */
421
+uint64_t arm_cp_read_zero(CPUARMState *env, const ARMCPRegInfo *ri);
422
+
423
+/*
424
+ * CPResetFn that does nothing, for use if no reset is required even
425
+ * if fieldoffset is non zero.
426
+ */
427
+void arm_cp_reset_ignore(CPUARMState *env, const ARMCPRegInfo *opaque);
428
+
429
+/*
430
+ * Return true if this reginfo struct's field in the cpu state struct
431
+ * is 64 bits wide.
432
+ */
433
+static inline bool cpreg_field_is_64bit(const ARMCPRegInfo *ri)
434
+{
435
+ return (ri->state == ARM_CP_STATE_AA64) || (ri->type & ARM_CP_64BIT);
436
+}
437
+
438
+static inline bool cp_access_ok(int current_el,
439
+ const ARMCPRegInfo *ri, int isread)
440
+{
441
+ return (ri->access >> ((current_el * 2) + isread)) & 1;
442
+}
443
+
444
+/* Raw read of a coprocessor register (as needed for migration, etc) */
445
+uint64_t read_raw_cp_reg(CPUARMState *env, const ARMCPRegInfo *ri);
446
+
447
+#endif /* TARGET_ARM_CPREGS_H */
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
448
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
16
index XXXXXXX..XXXXXXX 100644
449
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.h
450
--- a/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
451
+++ b/target/arm/cpu.h
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
452
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
20
#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
453
return kvmid;
21
#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
454
}
22
#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
455
23
-#define ARM_LAST_SPECIAL ARM_CP_DC_ZVA
456
-/* ARMCPRegInfo type field bits. If the SPECIAL bit is set this is a
24
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
457
- * special-behaviour cp reg and bits [11..8] indicate what behaviour
25
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
458
- * it has. Otherwise it is a simple cp reg, where CONST indicates that
26
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
459
- * TCG can assume the value to be constant (ie load at translate time)
27
#define ARM_CP_FPU 0x1000
460
- * and 64BIT indicates a 64 bit wide coprocessor register. SUPPRESS_TB_END
28
#define ARM_CP_SVE 0x2000
461
- * indicates that the TB should not be ended after a write to this register
29
#define ARM_CP_NO_GDB 0x4000
462
- * (the default is that the TB ends after cp writes). OVERRIDE permits
463
- * a register definition to override a previous definition for the
464
- * same (cp, is64, crn, crm, opc1, opc2) tuple: either the new or the
465
- * old must have the OVERRIDE bit set.
466
- * ALIAS indicates that this register is an alias view of some underlying
467
- * state which is also visible via another register, and that the other
468
- * register is handling migration and reset; registers marked ALIAS will not be
469
- * migrated but may have their state set by syncing of register state from KVM.
470
- * NO_RAW indicates that this register has no underlying state and does not
471
- * support raw access for state saving/loading; it will not be used for either
472
- * migration or KVM state synchronization. (Typically this is for "registers"
473
- * which are actually used as instructions for cache maintenance and so on.)
474
- * IO indicates that this register does I/O and therefore its accesses
475
- * need to be marked with gen_io_start() and also end the TB. In particular,
476
- * registers which implement clocks or timers require this.
477
- * RAISES_EXC is for when the read or write hook might raise an exception;
478
- * the generated code will synchronize the CPU state before calling the hook
479
- * so that it is safe for the hook to call raise_exception().
480
- * NEWEL is for writes to registers that might change the exception
481
- * level - typically on older ARM chips. For those cases we need to
482
- * re-read the new el when recomputing the translation flags.
483
- */
484
-#define ARM_CP_SPECIAL 0x0001
485
-#define ARM_CP_CONST 0x0002
486
-#define ARM_CP_64BIT 0x0004
487
-#define ARM_CP_SUPPRESS_TB_END 0x0008
488
-#define ARM_CP_OVERRIDE 0x0010
489
-#define ARM_CP_ALIAS 0x0020
490
-#define ARM_CP_IO 0x0040
491
-#define ARM_CP_NO_RAW 0x0080
492
-#define ARM_CP_NOP (ARM_CP_SPECIAL | 0x0100)
493
-#define ARM_CP_WFI (ARM_CP_SPECIAL | 0x0200)
494
-#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
495
-#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
496
-#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
497
-#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
498
-#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
499
-#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
500
-#define ARM_CP_FPU 0x1000
501
-#define ARM_CP_SVE 0x2000
502
-#define ARM_CP_NO_GDB 0x4000
503
-#define ARM_CP_RAISES_EXC 0x8000
504
-#define ARM_CP_NEWEL 0x10000
505
-/* Used only as a terminator for ARMCPRegInfo lists */
506
-#define ARM_CP_SENTINEL 0xfffff
507
-/* Mask of only the flag bits in a type field */
508
-#define ARM_CP_FLAG_MASK 0x1f0ff
509
-
510
-/* Valid values for ARMCPRegInfo state field, indicating which of
511
- * the AArch32 and AArch64 execution states this register is visible in.
512
- * If the reginfo doesn't explicitly specify then it is AArch32 only.
513
- * If the reginfo is declared to be visible in both states then a second
514
- * reginfo is synthesised for the AArch32 view of the AArch64 register,
515
- * such that the AArch32 view is the lower 32 bits of the AArch64 one.
516
- * Note that we rely on the values of these enums as we iterate through
517
- * the various states in some places.
518
- */
519
-enum {
520
- ARM_CP_STATE_AA32 = 0,
521
- ARM_CP_STATE_AA64 = 1,
522
- ARM_CP_STATE_BOTH = 2,
523
-};
524
-
525
-/* ARM CP register secure state flags. These flags identify security state
526
- * attributes for a given CP register entry.
527
- * The existence of both or neither secure and non-secure flags indicates that
528
- * the register has both a secure and non-secure hash entry. A single one of
529
- * these flags causes the register to only be hashed for the specified
530
- * security state.
531
- * Although definitions may have any combination of the S/NS bits, each
532
- * registered entry will only have one to identify whether the entry is secure
533
- * or non-secure.
534
- */
535
-enum {
536
- ARM_CP_SECSTATE_S = (1 << 0), /* bit[0]: Secure state register */
537
- ARM_CP_SECSTATE_NS = (1 << 1), /* bit[1]: Non-secure state register */
538
-};
539
-
540
-/* Return true if cptype is a valid type field. This is used to try to
541
- * catch errors where the sentinel has been accidentally left off the end
542
- * of a list of registers.
543
- */
544
-static inline bool cptype_valid(int cptype)
545
-{
546
- return ((cptype & ~ARM_CP_FLAG_MASK) == 0)
547
- || ((cptype & ARM_CP_SPECIAL) &&
548
- ((cptype & ~ARM_CP_FLAG_MASK) <= ARM_LAST_SPECIAL));
549
-}
550
-
551
-/* Access rights:
552
- * We define bits for Read and Write access for what rev C of the v7-AR ARM ARM
553
- * defines as PL0 (user), PL1 (fiq/irq/svc/abt/und/sys, ie privileged), and
554
- * PL2 (hyp). The other level which has Read and Write bits is Secure PL1
555
- * (ie any of the privileged modes in Secure state, or Monitor mode).
556
- * If a register is accessible in one privilege level it's always accessible
557
- * in higher privilege levels too. Since "Secure PL1" also follows this rule
558
- * (ie anything visible in PL2 is visible in S-PL1, some things are only
559
- * visible in S-PL1) but "Secure PL1" is a bit of a mouthful, we bend the
560
- * terminology a little and call this PL3.
561
- * In AArch64 things are somewhat simpler as the PLx bits line up exactly
562
- * with the ELx exception levels.
563
- *
564
- * If access permissions for a register are more complex than can be
565
- * described with these bits, then use a laxer set of restrictions, and
566
- * do the more restrictive/complex check inside a helper function.
567
- */
568
-#define PL3_R 0x80
569
-#define PL3_W 0x40
570
-#define PL2_R (0x20 | PL3_R)
571
-#define PL2_W (0x10 | PL3_W)
572
-#define PL1_R (0x08 | PL2_R)
573
-#define PL1_W (0x04 | PL2_W)
574
-#define PL0_R (0x02 | PL1_R)
575
-#define PL0_W (0x01 | PL1_W)
576
-
577
-/*
578
- * For user-mode some registers are accessible to EL0 via a kernel
579
- * trap-and-emulate ABI. In this case we define the read permissions
580
- * as actually being PL0_R. However some bits of any given register
581
- * may still be masked.
582
- */
583
-#ifdef CONFIG_USER_ONLY
584
-#define PL0U_R PL0_R
585
-#else
586
-#define PL0U_R PL1_R
587
-#endif
588
-
589
-#define PL3_RW (PL3_R | PL3_W)
590
-#define PL2_RW (PL2_R | PL2_W)
591
-#define PL1_RW (PL1_R | PL1_W)
592
-#define PL0_RW (PL0_R | PL0_W)
593
-
594
/* Return the highest implemented Exception Level */
595
static inline int arm_highest_el(CPUARMState *env)
596
{
597
@@ -XXX,XX +XXX,XX @@ static inline int arm_current_el(CPUARMState *env)
598
}
599
}
600
601
-typedef struct ARMCPRegInfo ARMCPRegInfo;
602
-
603
-typedef enum CPAccessResult {
604
- /* Access is permitted */
605
- CP_ACCESS_OK = 0,
606
- /* Access fails due to a configurable trap or enable which would
607
- * result in a categorized exception syndrome giving information about
608
- * the failing instruction (ie syndrome category 0x3, 0x4, 0x5, 0x6,
609
- * 0xc or 0x18). The exception is taken to the usual target EL (EL1 or
610
- * PL1 if in EL0, otherwise to the current EL).
611
- */
612
- CP_ACCESS_TRAP = 1,
613
- /* Access fails and results in an exception syndrome 0x0 ("uncategorized").
614
- * Note that this is not a catch-all case -- the set of cases which may
615
- * result in this failure is specifically defined by the architecture.
616
- */
617
- CP_ACCESS_TRAP_UNCATEGORIZED = 2,
618
- /* As CP_ACCESS_TRAP, but for traps directly to EL2 or EL3 */
619
- CP_ACCESS_TRAP_EL2 = 3,
620
- CP_ACCESS_TRAP_EL3 = 4,
621
- /* As CP_ACCESS_UNCATEGORIZED, but for traps directly to EL2 or EL3 */
622
- CP_ACCESS_TRAP_UNCATEGORIZED_EL2 = 5,
623
- CP_ACCESS_TRAP_UNCATEGORIZED_EL3 = 6,
624
-} CPAccessResult;
625
-
626
-/* Access functions for coprocessor registers. These cannot fail and
627
- * may not raise exceptions.
628
- */
629
-typedef uint64_t CPReadFn(CPUARMState *env, const ARMCPRegInfo *opaque);
630
-typedef void CPWriteFn(CPUARMState *env, const ARMCPRegInfo *opaque,
631
- uint64_t value);
632
-/* Access permission check functions for coprocessor registers. */
633
-typedef CPAccessResult CPAccessFn(CPUARMState *env,
634
- const ARMCPRegInfo *opaque,
635
- bool isread);
636
-/* Hook function for register reset */
637
-typedef void CPResetFn(CPUARMState *env, const ARMCPRegInfo *opaque);
638
-
639
-#define CP_ANY 0xff
640
-
641
-/* Definition of an ARM coprocessor register */
642
-struct ARMCPRegInfo {
643
- /* Name of register (useful mainly for debugging, need not be unique) */
644
- const char *name;
645
- /* Location of register: coprocessor number and (crn,crm,opc1,opc2)
646
- * tuple. Any of crm, opc1 and opc2 may be CP_ANY to indicate a
647
- * 'wildcard' field -- any value of that field in the MRC/MCR insn
648
- * will be decoded to this register. The register read and write
649
- * callbacks will be passed an ARMCPRegInfo with the crn/crm/opc1/opc2
650
- * used by the program, so it is possible to register a wildcard and
651
- * then behave differently on read/write if necessary.
652
- * For 64 bit registers, only crm and opc1 are relevant; crn and opc2
653
- * must both be zero.
654
- * For AArch64-visible registers, opc0 is also used.
655
- * Since there are no "coprocessors" in AArch64, cp is purely used as a
656
- * way to distinguish (for KVM's benefit) guest-visible system registers
657
- * from demuxed ones provided to preserve the "no side effects on
658
- * KVM register read/write from QEMU" semantics. cp==0x13 is guest
659
- * visible (to match KVM's encoding); cp==0 will be converted to
660
- * cp==0x13 when the ARMCPRegInfo is registered, for convenience.
661
- */
662
- uint8_t cp;
663
- uint8_t crn;
664
- uint8_t crm;
665
- uint8_t opc0;
666
- uint8_t opc1;
667
- uint8_t opc2;
668
- /* Execution state in which this register is visible: ARM_CP_STATE_* */
669
- int state;
670
- /* Register type: ARM_CP_* bits/values */
671
- int type;
672
- /* Access rights: PL*_[RW] */
673
- int access;
674
- /* Security state: ARM_CP_SECSTATE_* bits/values */
675
- int secure;
676
- /* The opaque pointer passed to define_arm_cp_regs_with_opaque() when
677
- * this register was defined: can be used to hand data through to the
678
- * register read/write functions, since they are passed the ARMCPRegInfo*.
679
- */
680
- void *opaque;
681
- /* Value of this register, if it is ARM_CP_CONST. Otherwise, if
682
- * fieldoffset is non-zero, the reset value of the register.
683
- */
684
- uint64_t resetvalue;
685
- /* Offset of the field in CPUARMState for this register.
686
- *
687
- * This is not needed if either:
688
- * 1. type is ARM_CP_CONST or one of the ARM_CP_SPECIALs
689
- * 2. both readfn and writefn are specified
690
- */
691
- ptrdiff_t fieldoffset; /* offsetof(CPUARMState, field) */
692
-
693
- /* Offsets of the secure and non-secure fields in CPUARMState for the
694
- * register if it is banked. These fields are only used during the static
695
- * registration of a register. During hashing the bank associated
696
- * with a given security state is copied to fieldoffset which is used from
697
- * there on out.
698
- *
699
- * It is expected that register definitions use either fieldoffset or
700
- * bank_fieldoffsets in the definition but not both. It is also expected
701
- * that both bank offsets are set when defining a banked register. This
702
- * use indicates that a register is banked.
703
- */
704
- ptrdiff_t bank_fieldoffsets[2];
705
-
706
- /* Function for making any access checks for this register in addition to
707
- * those specified by the 'access' permissions bits. If NULL, no extra
708
- * checks required. The access check is performed at runtime, not at
709
- * translate time.
710
- */
711
- CPAccessFn *accessfn;
712
- /* Function for handling reads of this register. If NULL, then reads
713
- * will be done by loading from the offset into CPUARMState specified
714
- * by fieldoffset.
715
- */
716
- CPReadFn *readfn;
717
- /* Function for handling writes of this register. If NULL, then writes
718
- * will be done by writing to the offset into CPUARMState specified
719
- * by fieldoffset.
720
- */
721
- CPWriteFn *writefn;
722
- /* Function for doing a "raw" read; used when we need to copy
723
- * coprocessor state to the kernel for KVM or out for
724
- * migration. This only needs to be provided if there is also a
725
- * readfn and it has side effects (for instance clear-on-read bits).
726
- */
727
- CPReadFn *raw_readfn;
728
- /* Function for doing a "raw" write; used when we need to copy KVM
729
- * kernel coprocessor state into userspace, or for inbound
730
- * migration. This only needs to be provided if there is also a
731
- * writefn and it masks out "unwritable" bits or has write-one-to-clear
732
- * or similar behaviour.
733
- */
734
- CPWriteFn *raw_writefn;
735
- /* Function for resetting the register. If NULL, then reset will be done
736
- * by writing resetvalue to the field specified in fieldoffset. If
737
- * fieldoffset is 0 then no reset will be done.
738
- */
739
- CPResetFn *resetfn;
740
-
741
- /*
742
- * "Original" writefn and readfn.
743
- * For ARMv8.1-VHE register aliases, we overwrite the read/write
744
- * accessor functions of various EL1/EL0 to perform the runtime
745
- * check for which sysreg should actually be modified, and then
746
- * forwards the operation. Before overwriting the accessors,
747
- * the original function is copied here, so that accesses that
748
- * really do go to the EL1/EL0 version proceed normally.
749
- * (The corresponding EL2 register is linked via opaque.)
750
- */
751
- CPReadFn *orig_readfn;
752
- CPWriteFn *orig_writefn;
753
-};
754
-
755
-/* Macros which are lvalues for the field in CPUARMState for the
756
- * ARMCPRegInfo *ri.
757
- */
758
-#define CPREG_FIELD32(env, ri) \
759
- (*(uint32_t *)((char *)(env) + (ri)->fieldoffset))
760
-#define CPREG_FIELD64(env, ri) \
761
- (*(uint64_t *)((char *)(env) + (ri)->fieldoffset))
762
-
763
-#define REGINFO_SENTINEL { .type = ARM_CP_SENTINEL }
764
-
765
-void define_arm_cp_regs_with_opaque(ARMCPU *cpu,
766
- const ARMCPRegInfo *regs, void *opaque);
767
-void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
768
- const ARMCPRegInfo *regs, void *opaque);
769
-static inline void define_arm_cp_regs(ARMCPU *cpu, const ARMCPRegInfo *regs)
770
-{
771
- define_arm_cp_regs_with_opaque(cpu, regs, 0);
772
-}
773
-static inline void define_one_arm_cp_reg(ARMCPU *cpu, const ARMCPRegInfo *regs)
774
-{
775
- define_one_arm_cp_reg_with_opaque(cpu, regs, 0);
776
-}
777
-const ARMCPRegInfo *get_arm_cp_reginfo(GHashTable *cpregs, uint32_t encoded_cp);
778
-
779
-/*
780
- * Definition of an ARM co-processor register as viewed from
781
- * userspace. This is used for presenting sanitised versions of
782
- * registers to userspace when emulating the Linux AArch64 CPU
783
- * ID/feature ABI (advertised as HWCAP_CPUID).
784
- */
785
-typedef struct ARMCPRegUserSpaceInfo {
786
- /* Name of register */
787
- const char *name;
788
-
789
- /* Is the name actually a glob pattern */
790
- bool is_glob;
791
-
792
- /* Only some bits are exported to user space */
793
- uint64_t exported_bits;
794
-
795
- /* Fixed bits are applied after the mask */
796
- uint64_t fixed_bits;
797
-} ARMCPRegUserSpaceInfo;
798
-
799
-#define REGUSERINFO_SENTINEL { .name = NULL }
800
-
801
-void modify_arm_cp_regs(ARMCPRegInfo *regs, const ARMCPRegUserSpaceInfo *mods);
802
-
803
-/* CPWriteFn that can be used to implement writes-ignored behaviour */
804
-void arm_cp_write_ignore(CPUARMState *env, const ARMCPRegInfo *ri,
805
- uint64_t value);
806
-/* CPReadFn that can be used for read-as-zero behaviour */
807
-uint64_t arm_cp_read_zero(CPUARMState *env, const ARMCPRegInfo *ri);
808
-
809
-/* CPResetFn that does nothing, for use if no reset is required even
810
- * if fieldoffset is non zero.
811
- */
812
-void arm_cp_reset_ignore(CPUARMState *env, const ARMCPRegInfo *opaque);
813
-
814
-/* Return true if this reginfo struct's field in the cpu state struct
815
- * is 64 bits wide.
816
- */
817
-static inline bool cpreg_field_is_64bit(const ARMCPRegInfo *ri)
818
-{
819
- return (ri->state == ARM_CP_STATE_AA64) || (ri->type & ARM_CP_64BIT);
820
-}
821
-
822
-static inline bool cp_access_ok(int current_el,
823
- const ARMCPRegInfo *ri, int isread)
824
-{
825
- return (ri->access >> ((current_el * 2) + isread)) & 1;
826
-}
827
-
828
-/* Raw read of a coprocessor register (as needed for migration, etc) */
829
-uint64_t read_raw_cp_reg(CPUARMState *env, const ARMCPRegInfo *ri);
830
-
831
/**
832
* write_list_to_cpustate
833
* @cpu: ARMCPU
834
diff --git a/hw/arm/pxa2xx.c b/hw/arm/pxa2xx.c
835
index XXXXXXX..XXXXXXX 100644
836
--- a/hw/arm/pxa2xx.c
837
+++ b/hw/arm/pxa2xx.c
838
@@ -XXX,XX +XXX,XX @@
839
#include "qemu/cutils.h"
840
#include "qemu/log.h"
841
#include "qom/object.h"
842
+#include "target/arm/cpregs.h"
843
844
static struct {
845
hwaddr io_base;
846
diff --git a/hw/arm/pxa2xx_pic.c b/hw/arm/pxa2xx_pic.c
847
index XXXXXXX..XXXXXXX 100644
848
--- a/hw/arm/pxa2xx_pic.c
849
+++ b/hw/arm/pxa2xx_pic.c
850
@@ -XXX,XX +XXX,XX @@
851
#include "hw/sysbus.h"
852
#include "migration/vmstate.h"
853
#include "qom/object.h"
854
+#include "target/arm/cpregs.h"
855
856
#define ICIP    0x00    /* Interrupt Controller IRQ Pending register */
857
#define ICMR    0x04    /* Interrupt Controller Mask register */
858
diff --git a/hw/intc/arm_gicv3_cpuif.c b/hw/intc/arm_gicv3_cpuif.c
859
index XXXXXXX..XXXXXXX 100644
860
--- a/hw/intc/arm_gicv3_cpuif.c
861
+++ b/hw/intc/arm_gicv3_cpuif.c
862
@@ -XXX,XX +XXX,XX @@
863
#include "gicv3_internal.h"
864
#include "hw/irq.h"
865
#include "cpu.h"
866
+#include "target/arm/cpregs.h"
867
868
/*
869
* Special case return value from hppvi_index(); must be larger than
870
diff --git a/hw/intc/arm_gicv3_kvm.c b/hw/intc/arm_gicv3_kvm.c
871
index XXXXXXX..XXXXXXX 100644
872
--- a/hw/intc/arm_gicv3_kvm.c
873
+++ b/hw/intc/arm_gicv3_kvm.c
874
@@ -XXX,XX +XXX,XX @@
875
#include "vgic_common.h"
876
#include "migration/blocker.h"
877
#include "qom/object.h"
878
+#include "target/arm/cpregs.h"
879
+
880
881
#ifdef DEBUG_GICV3_KVM
882
#define DPRINTF(fmt, ...) \
883
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
884
index XXXXXXX..XXXXXXX 100644
885
--- a/target/arm/cpu.c
886
+++ b/target/arm/cpu.c
887
@@ -XXX,XX +XXX,XX @@
888
#include "kvm_arm.h"
889
#include "disas/capstone.h"
890
#include "fpu/softfloat.h"
891
+#include "cpregs.h"
892
893
static void arm_cpu_set_pc(CPUState *cs, vaddr value)
894
{
895
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
896
index XXXXXXX..XXXXXXX 100644
897
--- a/target/arm/cpu64.c
898
+++ b/target/arm/cpu64.c
899
@@ -XXX,XX +XXX,XX @@
900
#include "hvf_arm.h"
901
#include "qapi/visitor.h"
902
#include "hw/qdev-properties.h"
903
+#include "cpregs.h"
904
905
906
#ifndef CONFIG_USER_ONLY
907
diff --git a/target/arm/cpu_tcg.c b/target/arm/cpu_tcg.c
908
index XXXXXXX..XXXXXXX 100644
909
--- a/target/arm/cpu_tcg.c
910
+++ b/target/arm/cpu_tcg.c
911
@@ -XXX,XX +XXX,XX @@
912
#if !defined(CONFIG_USER_ONLY)
913
#include "hw/boards.h"
914
#endif
915
+#include "cpregs.h"
916
917
/* CPU models. These are not needed for the AArch64 linux-user build. */
918
#if !defined(CONFIG_USER_ONLY) || !defined(TARGET_AARCH64)
919
diff --git a/target/arm/gdbstub.c b/target/arm/gdbstub.c
920
index XXXXXXX..XXXXXXX 100644
921
--- a/target/arm/gdbstub.c
922
+++ b/target/arm/gdbstub.c
923
@@ -XXX,XX +XXX,XX @@
924
*/
925
#include "qemu/osdep.h"
926
#include "cpu.h"
927
-#include "internals.h"
928
#include "exec/gdbstub.h"
929
+#include "internals.h"
930
+#include "cpregs.h"
931
932
typedef struct RegisterSysregXmlParam {
933
CPUState *cs;
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
934
diff --git a/target/arm/helper.c b/target/arm/helper.c
31
index XXXXXXX..XXXXXXX 100644
935
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.c
936
--- a/target/arm/helper.c
33
+++ b/target/arm/helper.c
937
+++ b/target/arm/helper.c
34
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
938
@@ -XXX,XX +XXX,XX @@
35
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
939
#include "exec/cpu_ldst.h"
36
.type = ARM_CP_NOP, .access = PL0_W,
940
#include "semihosting/common-semi.h"
37
.accessfn = aa64_cacheop_poc_access },
941
#endif
38
+ { .name = "DC_GVA", .state = ARM_CP_STATE_AA64,
942
+#include "cpregs.h"
39
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 3,
943
40
+ .access = PL0_W, .type = ARM_CP_DC_GVA,
944
#define ARM_CPU_FREQ 1000000000 /* FIXME: 1 GHz, should be configurable */
41
+#ifndef CONFIG_USER_ONLY
945
#define PMCR_NUM_COUNTERS 4 /* QEMU IMPDEF choice */
42
+ /* Avoid overhead of an access check that always passes in user-mode */
946
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
43
+ .accessfn = aa64_zva_access,
947
index XXXXXXX..XXXXXXX 100644
44
+#endif
948
--- a/target/arm/op_helper.c
45
+ },
949
+++ b/target/arm/op_helper.c
46
+ { .name = "DC_GZVA", .state = ARM_CP_STATE_AA64,
950
@@ -XXX,XX +XXX,XX @@
47
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 4,
951
#include "internals.h"
48
+ .access = PL0_W, .type = ARM_CP_DC_GZVA,
952
#include "exec/exec-all.h"
49
+#ifndef CONFIG_USER_ONLY
953
#include "exec/cpu_ldst.h"
50
+ /* Avoid overhead of an access check that always passes in user-mode */
954
+#include "cpregs.h"
51
+ .accessfn = aa64_zva_access,
955
52
+#endif
956
#define SIGNBIT (uint32_t)0x80000000
53
+ },
957
#define SIGNBIT64 ((uint64_t)1 << 63)
54
REGINFO_SENTINEL
55
};
56
57
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
958
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
58
index XXXXXXX..XXXXXXX 100644
959
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/translate-a64.c
960
--- a/target/arm/translate-a64.c
60
+++ b/target/arm/translate-a64.c
961
+++ b/target/arm/translate-a64.c
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
962
@@ -XXX,XX +XXX,XX @@
62
}
963
#include "translate.h"
63
gen_helper_dc_zva(cpu_env, tcg_rt);
964
#include "internals.h"
64
return;
965
#include "qemu/host-utils.h"
65
+ case ARM_CP_DC_GVA:
966
-
66
+ {
967
#include "semihosting/semihost.h"
67
+ TCGv_i64 clean_addr, tag;
968
#include "exec/gen-icount.h"
68
+
969
-
69
+ /*
970
#include "exec/helper-proto.h"
70
+ * DC_GVA, like DC_ZVA, requires that we supply the original
971
#include "exec/helper-gen.h"
71
+ * pointer for an invalid page. Probe that address first.
972
#include "exec/log.h"
72
+ */
973
-
73
+ tcg_rt = cpu_reg(s, rt);
974
+#include "cpregs.h"
74
+ clean_addr = clean_data_tbi(s, tcg_rt);
975
#include "translate-a64.h"
75
+ gen_probe_access(s, clean_addr, MMU_DATA_STORE, MO_8);
976
#include "qemu/atomic128.h"
76
+
977
77
+ if (s->ata) {
978
diff --git a/target/arm/translate.c b/target/arm/translate.c
78
+ /* Extract the tag from the register to match STZGM. */
979
index XXXXXXX..XXXXXXX 100644
79
+ tag = tcg_temp_new_i64();
980
--- a/target/arm/translate.c
80
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
981
+++ b/target/arm/translate.c
81
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
982
@@ -XXX,XX +XXX,XX @@
82
+ tcg_temp_free_i64(tag);
983
#include "qemu/bitops.h"
83
+ }
984
#include "arm_ldst.h"
84
+ }
985
#include "semihosting/semihost.h"
85
+ return;
986
-
86
+ case ARM_CP_DC_GZVA:
987
#include "exec/helper-proto.h"
87
+ {
988
#include "exec/helper-gen.h"
88
+ TCGv_i64 clean_addr, tag;
989
-
89
+
990
#include "exec/log.h"
90
+ /* For DC_GZVA, we can rely on DC_ZVA for the proper fault. */
991
+#include "cpregs.h"
91
+ tcg_rt = cpu_reg(s, rt);
992
92
+ clean_addr = clean_data_tbi(s, tcg_rt);
993
93
+ gen_helper_dc_zva(cpu_env, clean_addr);
994
#define ENABLE_ARCH_4T arm_dc_feature(s, ARM_FEATURE_V4T)
94
+
95
+ if (s->ata) {
96
+ /* Extract the tag from the register to match STZGM. */
97
+ tag = tcg_temp_new_i64();
98
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
99
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
100
+ tcg_temp_free_i64(tag);
101
+ }
102
+ }
103
+ return;
104
default:
105
break;
106
}
107
--
995
--
108
2.20.1
996
2.25.1
109
997
110
998
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Because the elements are non-sequential, we cannot eliminate many
3
Rearrange the values of the enumerators of CPAccessResult
4
tests straight away like we can for sequential operations. But
4
so that we may directly extract the target el. For the two
5
we often have the PTE details handy, so we can test for Tagged.
5
special cases in access_check_cp_reg, use CPAccessResult.
6
6
7
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-38-richard.henderson@linaro.org
10
Message-id: 20220501055028.646596-3-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
---
12
target/arm/helper-sve.h | 285 ++++++++++++++++
13
target/arm/cpregs.h | 26 ++++++++++++--------
13
target/arm/sve_helper.c | 185 +++++++++--
14
target/arm/op_helper.c | 56 +++++++++++++++++++++---------------------
14
target/arm/translate-sve.c | 650 +++++++++++++++++++++++++------------
15
2 files changed, 44 insertions(+), 38 deletions(-)
15
3 files changed, 872 insertions(+), 248 deletions(-)
16
16
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
18
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/helper-sve.h
19
--- a/target/arm/cpregs.h
20
+++ b/target/arm/helper-sve.h
20
+++ b/target/arm/cpregs.h
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldsds_le_zd, TCG_CALL_NO_WG,
21
@@ -XXX,XX +XXX,XX @@ static inline bool cptype_valid(int cptype)
22
DEF_HELPER_FLAGS_6(sve_ldsds_be_zd, TCG_CALL_NO_WG,
22
typedef enum CPAccessResult {
23
void, env, ptr, ptr, ptr, tl, i32)
23
/* Access is permitted */
24
24
CP_ACCESS_OK = 0,
25
+DEF_HELPER_FLAGS_6(sve_ldbsu_zsu_mte, TCG_CALL_NO_WG,
26
+ void, env, ptr, ptr, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu_mte, TCG_CALL_NO_WG,
28
+ void, env, ptr, ptr, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zsu_mte, TCG_CALL_NO_WG,
30
+ void, env, ptr, ptr, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_6(sve_ldss_le_zsu_mte, TCG_CALL_NO_WG,
32
+ void, env, ptr, ptr, ptr, tl, i32)
33
+DEF_HELPER_FLAGS_6(sve_ldss_be_zsu_mte, TCG_CALL_NO_WG,
34
+ void, env, ptr, ptr, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_6(sve_ldbss_zsu_mte, TCG_CALL_NO_WG,
36
+ void, env, ptr, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zsu_mte, TCG_CALL_NO_WG,
38
+ void, env, ptr, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zsu_mte, TCG_CALL_NO_WG,
40
+ void, env, ptr, ptr, ptr, tl, i32)
41
+
42
+DEF_HELPER_FLAGS_6(sve_ldbsu_zss_mte, TCG_CALL_NO_WG,
43
+ void, env, ptr, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zss_mte, TCG_CALL_NO_WG,
45
+ void, env, ptr, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zss_mte, TCG_CALL_NO_WG,
47
+ void, env, ptr, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_6(sve_ldss_le_zss_mte, TCG_CALL_NO_WG,
49
+ void, env, ptr, ptr, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_6(sve_ldss_be_zss_mte, TCG_CALL_NO_WG,
51
+ void, env, ptr, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_6(sve_ldbss_zss_mte, TCG_CALL_NO_WG,
53
+ void, env, ptr, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zss_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zss_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_6(sve_ldbdu_zsu_mte, TCG_CALL_NO_WG,
60
+ void, env, ptr, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zsu_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zsu_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zsu_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, ptr, ptr, tl, i32)
67
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zsu_mte, TCG_CALL_NO_WG,
68
+ void, env, ptr, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_6(sve_lddd_le_zsu_mte, TCG_CALL_NO_WG,
70
+ void, env, ptr, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_6(sve_lddd_be_zsu_mte, TCG_CALL_NO_WG,
72
+ void, env, ptr, ptr, ptr, tl, i32)
73
+DEF_HELPER_FLAGS_6(sve_ldbds_zsu_mte, TCG_CALL_NO_WG,
74
+ void, env, ptr, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zsu_mte, TCG_CALL_NO_WG,
76
+ void, env, ptr, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zsu_mte, TCG_CALL_NO_WG,
78
+ void, env, ptr, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zsu_mte, TCG_CALL_NO_WG,
80
+ void, env, ptr, ptr, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zsu_mte, TCG_CALL_NO_WG,
82
+ void, env, ptr, ptr, ptr, tl, i32)
83
+
84
+DEF_HELPER_FLAGS_6(sve_ldbdu_zss_mte, TCG_CALL_NO_WG,
85
+ void, env, ptr, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zss_mte, TCG_CALL_NO_WG,
87
+ void, env, ptr, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zss_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, ptr, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zss_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zss_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_6(sve_lddd_le_zss_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_6(sve_lddd_be_zss_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_6(sve_ldbds_zss_mte, TCG_CALL_NO_WG,
99
+ void, env, ptr, ptr, ptr, tl, i32)
100
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zss_mte, TCG_CALL_NO_WG,
101
+ void, env, ptr, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zss_mte, TCG_CALL_NO_WG,
103
+ void, env, ptr, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zss_mte, TCG_CALL_NO_WG,
105
+ void, env, ptr, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zss_mte, TCG_CALL_NO_WG,
107
+ void, env, ptr, ptr, ptr, tl, i32)
108
+
109
+DEF_HELPER_FLAGS_6(sve_ldbdu_zd_mte, TCG_CALL_NO_WG,
110
+ void, env, ptr, ptr, ptr, tl, i32)
111
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zd_mte, TCG_CALL_NO_WG,
112
+ void, env, ptr, ptr, ptr, tl, i32)
113
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zd_mte, TCG_CALL_NO_WG,
114
+ void, env, ptr, ptr, ptr, tl, i32)
115
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zd_mte, TCG_CALL_NO_WG,
116
+ void, env, ptr, ptr, ptr, tl, i32)
117
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zd_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, ptr, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_6(sve_lddd_le_zd_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, ptr, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_6(sve_lddd_be_zd_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, ptr, ptr, tl, i32)
123
+DEF_HELPER_FLAGS_6(sve_ldbds_zd_mte, TCG_CALL_NO_WG,
124
+ void, env, ptr, ptr, ptr, tl, i32)
125
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zd_mte, TCG_CALL_NO_WG,
126
+ void, env, ptr, ptr, ptr, tl, i32)
127
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zd_mte, TCG_CALL_NO_WG,
128
+ void, env, ptr, ptr, ptr, tl, i32)
129
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zd_mte, TCG_CALL_NO_WG,
130
+ void, env, ptr, ptr, ptr, tl, i32)
131
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zd_mte, TCG_CALL_NO_WG,
132
+ void, env, ptr, ptr, ptr, tl, i32)
133
+
134
DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu, TCG_CALL_NO_WG,
135
void, env, ptr, ptr, ptr, tl, i32)
136
DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu, TCG_CALL_NO_WG,
137
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd, TCG_CALL_NO_WG,
138
DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd, TCG_CALL_NO_WG,
139
void, env, ptr, ptr, ptr, tl, i32)
140
141
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu_mte, TCG_CALL_NO_WG,
142
+ void, env, ptr, ptr, ptr, tl, i32)
143
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu_mte, TCG_CALL_NO_WG,
144
+ void, env, ptr, ptr, ptr, tl, i32)
145
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zsu_mte, TCG_CALL_NO_WG,
146
+ void, env, ptr, ptr, ptr, tl, i32)
147
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zsu_mte, TCG_CALL_NO_WG,
148
+ void, env, ptr, ptr, ptr, tl, i32)
149
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zsu_mte, TCG_CALL_NO_WG,
150
+ void, env, ptr, ptr, ptr, tl, i32)
151
+DEF_HELPER_FLAGS_6(sve_ldffbss_zsu_mte, TCG_CALL_NO_WG,
152
+ void, env, ptr, ptr, ptr, tl, i32)
153
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zsu_mte, TCG_CALL_NO_WG,
154
+ void, env, ptr, ptr, ptr, tl, i32)
155
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zsu_mte, TCG_CALL_NO_WG,
156
+ void, env, ptr, ptr, ptr, tl, i32)
157
+
158
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zss_mte, TCG_CALL_NO_WG,
159
+ void, env, ptr, ptr, ptr, tl, i32)
160
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zss_mte, TCG_CALL_NO_WG,
161
+ void, env, ptr, ptr, ptr, tl, i32)
162
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zss_mte, TCG_CALL_NO_WG,
163
+ void, env, ptr, ptr, ptr, tl, i32)
164
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zss_mte, TCG_CALL_NO_WG,
165
+ void, env, ptr, ptr, ptr, tl, i32)
166
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zss_mte, TCG_CALL_NO_WG,
167
+ void, env, ptr, ptr, ptr, tl, i32)
168
+DEF_HELPER_FLAGS_6(sve_ldffbss_zss_mte, TCG_CALL_NO_WG,
169
+ void, env, ptr, ptr, ptr, tl, i32)
170
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zss_mte, TCG_CALL_NO_WG,
171
+ void, env, ptr, ptr, ptr, tl, i32)
172
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zss_mte, TCG_CALL_NO_WG,
173
+ void, env, ptr, ptr, ptr, tl, i32)
174
+
175
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zsu_mte, TCG_CALL_NO_WG,
176
+ void, env, ptr, ptr, ptr, tl, i32)
177
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zsu_mte, TCG_CALL_NO_WG,
178
+ void, env, ptr, ptr, ptr, tl, i32)
179
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zsu_mte, TCG_CALL_NO_WG,
180
+ void, env, ptr, ptr, ptr, tl, i32)
181
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zsu_mte, TCG_CALL_NO_WG,
182
+ void, env, ptr, ptr, ptr, tl, i32)
183
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zsu_mte, TCG_CALL_NO_WG,
184
+ void, env, ptr, ptr, ptr, tl, i32)
185
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zsu_mte, TCG_CALL_NO_WG,
186
+ void, env, ptr, ptr, ptr, tl, i32)
187
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zsu_mte, TCG_CALL_NO_WG,
188
+ void, env, ptr, ptr, ptr, tl, i32)
189
+DEF_HELPER_FLAGS_6(sve_ldffbds_zsu_mte, TCG_CALL_NO_WG,
190
+ void, env, ptr, ptr, ptr, tl, i32)
191
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zsu_mte, TCG_CALL_NO_WG,
192
+ void, env, ptr, ptr, ptr, tl, i32)
193
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zsu_mte, TCG_CALL_NO_WG,
194
+ void, env, ptr, ptr, ptr, tl, i32)
195
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zsu_mte, TCG_CALL_NO_WG,
196
+ void, env, ptr, ptr, ptr, tl, i32)
197
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zsu_mte, TCG_CALL_NO_WG,
198
+ void, env, ptr, ptr, ptr, tl, i32)
199
+
200
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zss_mte, TCG_CALL_NO_WG,
201
+ void, env, ptr, ptr, ptr, tl, i32)
202
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zss_mte, TCG_CALL_NO_WG,
203
+ void, env, ptr, ptr, ptr, tl, i32)
204
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zss_mte, TCG_CALL_NO_WG,
205
+ void, env, ptr, ptr, ptr, tl, i32)
206
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zss_mte, TCG_CALL_NO_WG,
207
+ void, env, ptr, ptr, ptr, tl, i32)
208
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zss_mte, TCG_CALL_NO_WG,
209
+ void, env, ptr, ptr, ptr, tl, i32)
210
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zss_mte, TCG_CALL_NO_WG,
211
+ void, env, ptr, ptr, ptr, tl, i32)
212
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zss_mte, TCG_CALL_NO_WG,
213
+ void, env, ptr, ptr, ptr, tl, i32)
214
+DEF_HELPER_FLAGS_6(sve_ldffbds_zss_mte, TCG_CALL_NO_WG,
215
+ void, env, ptr, ptr, ptr, tl, i32)
216
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zss_mte, TCG_CALL_NO_WG,
217
+ void, env, ptr, ptr, ptr, tl, i32)
218
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zss_mte, TCG_CALL_NO_WG,
219
+ void, env, ptr, ptr, ptr, tl, i32)
220
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zss_mte, TCG_CALL_NO_WG,
221
+ void, env, ptr, ptr, ptr, tl, i32)
222
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zss_mte, TCG_CALL_NO_WG,
223
+ void, env, ptr, ptr, ptr, tl, i32)
224
+
225
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zd_mte, TCG_CALL_NO_WG,
226
+ void, env, ptr, ptr, ptr, tl, i32)
227
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zd_mte, TCG_CALL_NO_WG,
228
+ void, env, ptr, ptr, ptr, tl, i32)
229
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zd_mte, TCG_CALL_NO_WG,
230
+ void, env, ptr, ptr, ptr, tl, i32)
231
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zd_mte, TCG_CALL_NO_WG,
232
+ void, env, ptr, ptr, ptr, tl, i32)
233
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zd_mte, TCG_CALL_NO_WG,
234
+ void, env, ptr, ptr, ptr, tl, i32)
235
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zd_mte, TCG_CALL_NO_WG,
236
+ void, env, ptr, ptr, ptr, tl, i32)
237
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zd_mte, TCG_CALL_NO_WG,
238
+ void, env, ptr, ptr, ptr, tl, i32)
239
+DEF_HELPER_FLAGS_6(sve_ldffbds_zd_mte, TCG_CALL_NO_WG,
240
+ void, env, ptr, ptr, ptr, tl, i32)
241
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zd_mte, TCG_CALL_NO_WG,
242
+ void, env, ptr, ptr, ptr, tl, i32)
243
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zd_mte, TCG_CALL_NO_WG,
244
+ void, env, ptr, ptr, ptr, tl, i32)
245
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd_mte, TCG_CALL_NO_WG,
246
+ void, env, ptr, ptr, ptr, tl, i32)
247
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd_mte, TCG_CALL_NO_WG,
248
+ void, env, ptr, ptr, ptr, tl, i32)
249
+
250
DEF_HELPER_FLAGS_6(sve_stbs_zsu, TCG_CALL_NO_WG,
251
void, env, ptr, ptr, ptr, tl, i32)
252
DEF_HELPER_FLAGS_6(sve_sths_le_zsu, TCG_CALL_NO_WG,
253
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd, TCG_CALL_NO_WG,
254
DEF_HELPER_FLAGS_6(sve_stdd_be_zd, TCG_CALL_NO_WG,
255
void, env, ptr, ptr, ptr, tl, i32)
256
257
+DEF_HELPER_FLAGS_6(sve_stbs_zsu_mte, TCG_CALL_NO_WG,
258
+ void, env, ptr, ptr, ptr, tl, i32)
259
+DEF_HELPER_FLAGS_6(sve_sths_le_zsu_mte, TCG_CALL_NO_WG,
260
+ void, env, ptr, ptr, ptr, tl, i32)
261
+DEF_HELPER_FLAGS_6(sve_sths_be_zsu_mte, TCG_CALL_NO_WG,
262
+ void, env, ptr, ptr, ptr, tl, i32)
263
+DEF_HELPER_FLAGS_6(sve_stss_le_zsu_mte, TCG_CALL_NO_WG,
264
+ void, env, ptr, ptr, ptr, tl, i32)
265
+DEF_HELPER_FLAGS_6(sve_stss_be_zsu_mte, TCG_CALL_NO_WG,
266
+ void, env, ptr, ptr, ptr, tl, i32)
267
+
268
+DEF_HELPER_FLAGS_6(sve_stbs_zss_mte, TCG_CALL_NO_WG,
269
+ void, env, ptr, ptr, ptr, tl, i32)
270
+DEF_HELPER_FLAGS_6(sve_sths_le_zss_mte, TCG_CALL_NO_WG,
271
+ void, env, ptr, ptr, ptr, tl, i32)
272
+DEF_HELPER_FLAGS_6(sve_sths_be_zss_mte, TCG_CALL_NO_WG,
273
+ void, env, ptr, ptr, ptr, tl, i32)
274
+DEF_HELPER_FLAGS_6(sve_stss_le_zss_mte, TCG_CALL_NO_WG,
275
+ void, env, ptr, ptr, ptr, tl, i32)
276
+DEF_HELPER_FLAGS_6(sve_stss_be_zss_mte, TCG_CALL_NO_WG,
277
+ void, env, ptr, ptr, ptr, tl, i32)
278
+
279
+DEF_HELPER_FLAGS_6(sve_stbd_zsu_mte, TCG_CALL_NO_WG,
280
+ void, env, ptr, ptr, ptr, tl, i32)
281
+DEF_HELPER_FLAGS_6(sve_sthd_le_zsu_mte, TCG_CALL_NO_WG,
282
+ void, env, ptr, ptr, ptr, tl, i32)
283
+DEF_HELPER_FLAGS_6(sve_sthd_be_zsu_mte, TCG_CALL_NO_WG,
284
+ void, env, ptr, ptr, ptr, tl, i32)
285
+DEF_HELPER_FLAGS_6(sve_stsd_le_zsu_mte, TCG_CALL_NO_WG,
286
+ void, env, ptr, ptr, ptr, tl, i32)
287
+DEF_HELPER_FLAGS_6(sve_stsd_be_zsu_mte, TCG_CALL_NO_WG,
288
+ void, env, ptr, ptr, ptr, tl, i32)
289
+DEF_HELPER_FLAGS_6(sve_stdd_le_zsu_mte, TCG_CALL_NO_WG,
290
+ void, env, ptr, ptr, ptr, tl, i32)
291
+DEF_HELPER_FLAGS_6(sve_stdd_be_zsu_mte, TCG_CALL_NO_WG,
292
+ void, env, ptr, ptr, ptr, tl, i32)
293
+
294
+DEF_HELPER_FLAGS_6(sve_stbd_zss_mte, TCG_CALL_NO_WG,
295
+ void, env, ptr, ptr, ptr, tl, i32)
296
+DEF_HELPER_FLAGS_6(sve_sthd_le_zss_mte, TCG_CALL_NO_WG,
297
+ void, env, ptr, ptr, ptr, tl, i32)
298
+DEF_HELPER_FLAGS_6(sve_sthd_be_zss_mte, TCG_CALL_NO_WG,
299
+ void, env, ptr, ptr, ptr, tl, i32)
300
+DEF_HELPER_FLAGS_6(sve_stsd_le_zss_mte, TCG_CALL_NO_WG,
301
+ void, env, ptr, ptr, ptr, tl, i32)
302
+DEF_HELPER_FLAGS_6(sve_stsd_be_zss_mte, TCG_CALL_NO_WG,
303
+ void, env, ptr, ptr, ptr, tl, i32)
304
+DEF_HELPER_FLAGS_6(sve_stdd_le_zss_mte, TCG_CALL_NO_WG,
305
+ void, env, ptr, ptr, ptr, tl, i32)
306
+DEF_HELPER_FLAGS_6(sve_stdd_be_zss_mte, TCG_CALL_NO_WG,
307
+ void, env, ptr, ptr, ptr, tl, i32)
308
+
309
+DEF_HELPER_FLAGS_6(sve_stbd_zd_mte, TCG_CALL_NO_WG,
310
+ void, env, ptr, ptr, ptr, tl, i32)
311
+DEF_HELPER_FLAGS_6(sve_sthd_le_zd_mte, TCG_CALL_NO_WG,
312
+ void, env, ptr, ptr, ptr, tl, i32)
313
+DEF_HELPER_FLAGS_6(sve_sthd_be_zd_mte, TCG_CALL_NO_WG,
314
+ void, env, ptr, ptr, ptr, tl, i32)
315
+DEF_HELPER_FLAGS_6(sve_stsd_le_zd_mte, TCG_CALL_NO_WG,
316
+ void, env, ptr, ptr, ptr, tl, i32)
317
+DEF_HELPER_FLAGS_6(sve_stsd_be_zd_mte, TCG_CALL_NO_WG,
318
+ void, env, ptr, ptr, ptr, tl, i32)
319
+DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG,
320
+ void, env, ptr, ptr, ptr, tl, i32)
321
+DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG,
322
+ void, env, ptr, ptr, ptr, tl, i32)
323
+
324
DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
325
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
326
index XXXXXXX..XXXXXXX 100644
327
--- a/target/arm/sve_helper.c
328
+++ b/target/arm/sve_helper.c
329
@@ -XXX,XX +XXX,XX @@ static target_ulong off_zd_d(void *reg, intptr_t reg_ofs)
330
static inline QEMU_ALWAYS_INLINE
331
void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
332
target_ulong base, uint32_t desc, uintptr_t retaddr,
333
- int esize, int msize, zreg_off_fn *off_fn,
334
+ uint32_t mtedesc, int esize, int msize,
335
+ zreg_off_fn *off_fn,
336
sve_ldst1_host_fn *host_fn,
337
sve_ldst1_tlb_fn *tlb_fn)
338
{
339
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
340
cpu_check_watchpoint(env_cpu(env), addr, msize,
341
info.attrs, BP_MEM_READ, retaddr);
342
}
343
- /* TODO: MTE check */
344
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
345
+ mte_check1(env, mtedesc, addr, retaddr);
346
+ }
347
host_fn(&scratch, reg_off, info.host);
348
} else {
349
/* Element crosses the page boundary. */
350
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
351
msize, info.attrs,
352
BP_MEM_READ, retaddr);
353
}
354
- /* TODO: MTE check */
355
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
356
+ mte_check1(env, mtedesc, addr, retaddr);
357
+ }
358
tlb_fn(env, &scratch, reg_off, addr, retaddr);
359
}
360
}
361
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
362
memcpy(vd, &scratch, reg_max);
363
}
364
365
+static inline QEMU_ALWAYS_INLINE
366
+void sve_ld1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
367
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
368
+ int esize, int msize, zreg_off_fn *off_fn,
369
+ sve_ldst1_host_fn *host_fn,
370
+ sve_ldst1_tlb_fn *tlb_fn)
371
+{
372
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
373
+ /* Remove mtedesc from the normal sve descriptor. */
374
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
375
+
25
+
376
+ /*
26
+ /*
377
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
27
+ * Combined with one of the following, the low 2 bits indicate the
378
+ * offset base entirely over the address space hole to change the
28
+ * target exception level. If 0, the exception is taken to the usual
379
+ * pointer tag, or change the bit55 selector. So we could here
29
+ * target EL (EL1 or PL1 if in EL0, otherwise to the current EL).
380
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
381
+ */
30
+ */
382
+ sve_ld1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
31
+ CP_ACCESS_EL_MASK = 3,
383
+ esize, msize, off_fn, host_fn, tlb_fn);
384
+}
385
+
32
+
386
#define DO_LD1_ZPZ_S(MEM, OFS, MSZ) \
33
/*
387
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
34
* Access fails due to a configurable trap or enable which would
388
void *vm, target_ulong base, uint32_t desc) \
35
* result in a categorized exception syndrome giving information about
389
{ \
36
* the failing instruction (ie syndrome category 0x3, 0x4, 0x5, 0x6,
390
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
37
- * 0xc or 0x18). The exception is taken to the usual target EL (EL1 or
391
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
38
- * PL1 if in EL0, otherwise to the current EL).
392
off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
39
+ * 0xc or 0x18).
393
+} \
40
*/
394
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
41
- CP_ACCESS_TRAP = 1,
395
+ void *vm, target_ulong base, uint32_t desc) \
42
+ CP_ACCESS_TRAP = (1 << 2),
396
+{ \
43
+ CP_ACCESS_TRAP_EL2 = CP_ACCESS_TRAP | 2,
397
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
44
+ CP_ACCESS_TRAP_EL3 = CP_ACCESS_TRAP | 3,
398
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
45
+
399
}
46
/*
400
47
* Access fails and results in an exception syndrome 0x0 ("uncategorized").
401
#define DO_LD1_ZPZ_D(MEM, OFS, MSZ) \
48
* Note that this is not a catch-all case -- the set of cases which may
402
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
49
* result in this failure is specifically defined by the architecture.
403
void *vm, target_ulong base, uint32_t desc) \
50
*/
404
{ \
51
- CP_ACCESS_TRAP_UNCATEGORIZED = 2,
405
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
52
- /* As CP_ACCESS_TRAP, but for traps directly to EL2 or EL3 */
406
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
53
- CP_ACCESS_TRAP_EL2 = 3,
407
off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
54
- CP_ACCESS_TRAP_EL3 = 4,
408
+} \
55
- /* As CP_ACCESS_UNCATEGORIZED, but for traps directly to EL2 or EL3 */
409
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
56
- CP_ACCESS_TRAP_UNCATEGORIZED_EL2 = 5,
410
+ void *vm, target_ulong base, uint32_t desc) \
57
- CP_ACCESS_TRAP_UNCATEGORIZED_EL3 = 6,
411
+{ \
58
+ CP_ACCESS_TRAP_UNCATEGORIZED = (2 << 2),
412
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
59
+ CP_ACCESS_TRAP_UNCATEGORIZED_EL2 = CP_ACCESS_TRAP_UNCATEGORIZED | 2,
413
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
60
+ CP_ACCESS_TRAP_UNCATEGORIZED_EL3 = CP_ACCESS_TRAP_UNCATEGORIZED | 3,
414
}
61
} CPAccessResult;
415
62
416
DO_LD1_ZPZ_S(bsu, zsu, MO_8)
63
typedef struct ARMCPRegInfo ARMCPRegInfo;
417
@@ -XXX,XX +XXX,XX @@ DO_LD1_ZPZ_D(dd_be, zd, MO_64)
64
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
418
static inline QEMU_ALWAYS_INLINE
65
index XXXXXXX..XXXXXXX 100644
419
void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
66
--- a/target/arm/op_helper.c
420
target_ulong base, uint32_t desc, uintptr_t retaddr,
67
+++ b/target/arm/op_helper.c
421
- const int esz, const int msz, zreg_off_fn *off_fn,
68
@@ -XXX,XX +XXX,XX @@ void HELPER(access_check_cp_reg)(CPUARMState *env, void *rip, uint32_t syndrome,
422
+ uint32_t mtedesc, const int esz, const int msz,
69
uint32_t isread)
423
+ zreg_off_fn *off_fn,
424
sve_ldst1_host_fn *host_fn,
425
sve_ldst1_tlb_fn *tlb_fn)
426
{
70
{
427
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
71
const ARMCPRegInfo *ri = rip;
428
* Probe the first element, allowing faults.
72
+ CPAccessResult res = CP_ACCESS_OK;
429
*/
73
int target_el;
430
addr = base + (off_fn(vm, reg_off) << scale);
74
431
+ if (mtedesc) {
75
if (arm_feature(env, ARM_FEATURE_XSCALE) && ri->cp < 14
432
+ mte_check1(env, mtedesc, addr, retaddr);
76
&& extract32(env->cp15.c15_cpar, ri->cp, 1) == 0) {
77
- raise_exception(env, EXCP_UDEF, syndrome, exception_target_el(env));
78
+ res = CP_ACCESS_TRAP;
79
+ goto fail;
80
}
81
82
/*
83
@@ -XXX,XX +XXX,XX @@ void HELPER(access_check_cp_reg)(CPUARMState *env, void *rip, uint32_t syndrome,
84
mask &= ~((1 << 4) | (1 << 14));
85
86
if (env->cp15.hstr_el2 & mask) {
87
- target_el = 2;
88
- goto exept;
89
+ res = CP_ACCESS_TRAP_EL2;
90
+ goto fail;
91
}
92
}
93
94
- if (!ri->accessfn) {
95
+ if (ri->accessfn) {
96
+ res = ri->accessfn(env, ri, isread);
433
+ }
97
+ }
434
tlb_fn(env, vd, reg_off, addr, retaddr);
98
+ if (likely(res == CP_ACCESS_OK)) {
435
99
return;
436
/* After any fault, zero the other elements. */
100
}
437
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
101
438
(env_cpu(env), addr, msize) & BP_MEM_READ)) {
102
- switch (ri->accessfn(env, ri, isread)) {
439
goto fault;
103
- case CP_ACCESS_OK:
440
}
104
- return;
441
- /* TODO: MTE check. */
105
+ fail:
442
+ if (mtedesc &&
106
+ switch (res & ~CP_ACCESS_EL_MASK) {
443
+ arm_tlb_mte_tagged(&info.attrs) &&
107
case CP_ACCESS_TRAP:
444
+ !mte_probe1(env, mtedesc, addr)) {
108
- target_el = exception_target_el(env);
445
+ goto fault;
109
- break;
446
+ }
110
- case CP_ACCESS_TRAP_EL2:
447
111
- /* Requesting a trap to EL2 when we're in EL3 is
448
host_fn(vd, reg_off, info.host);
112
- * a bug in the access function.
449
}
113
- */
450
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
114
- assert(arm_current_el(env) != 3);
451
record_fault(env, reg_off, reg_max);
115
- target_el = 2;
452
}
116
- break;
453
117
- case CP_ACCESS_TRAP_EL3:
454
-#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
118
- target_el = 3;
455
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
456
- void *vm, target_ulong base, uint32_t desc) \
457
-{ \
458
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
459
- off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
460
+static inline QEMU_ALWAYS_INLINE
461
+void sve_ldff1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
462
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
463
+ const int esz, const int msz,
464
+ zreg_off_fn *off_fn,
465
+ sve_ldst1_host_fn *host_fn,
466
+ sve_ldst1_tlb_fn *tlb_fn)
467
+{
468
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
469
+ /* Remove mtedesc from the normal sve descriptor. */
470
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
471
+
472
+ /*
473
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
474
+ * offset base entirely over the address space hole to change the
475
+ * pointer tag, or change the bit55 selector. So we could here
476
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
477
+ */
478
+ sve_ldff1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
479
+ esz, msz, off_fn, host_fn, tlb_fn);
480
}
481
482
-#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
483
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
484
- void *vm, target_ulong base, uint32_t desc) \
485
-{ \
486
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
487
- off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
488
+#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
489
+void HELPER(sve_ldff##MEM##_##OFS) \
490
+ (CPUARMState *env, void *vd, void *vg, \
491
+ void *vm, target_ulong base, uint32_t desc) \
492
+{ \
493
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_32, MSZ, \
494
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
495
+} \
496
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
497
+ (CPUARMState *env, void *vd, void *vg, \
498
+ void *vm, target_ulong base, uint32_t desc) \
499
+{ \
500
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
501
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
502
+}
503
+
504
+#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
505
+void HELPER(sve_ldff##MEM##_##OFS) \
506
+ (CPUARMState *env, void *vd, void *vg, \
507
+ void *vm, target_ulong base, uint32_t desc) \
508
+{ \
509
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_64, MSZ, \
510
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
511
+} \
512
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
513
+ (CPUARMState *env, void *vd, void *vg, \
514
+ void *vm, target_ulong base, uint32_t desc) \
515
+{ \
516
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
517
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
518
}
519
520
DO_LDFF1_ZPZ_S(bsu, zsu, MO_8)
521
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_ZPZ_D(dd_be, zd, MO_64)
522
static inline QEMU_ALWAYS_INLINE
523
void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
524
target_ulong base, uint32_t desc, uintptr_t retaddr,
525
- int esize, int msize, zreg_off_fn *off_fn,
526
+ uint32_t mtedesc, int esize, int msize,
527
+ zreg_off_fn *off_fn,
528
sve_ldst1_host_fn *host_fn,
529
sve_ldst1_tlb_fn *tlb_fn)
530
{
531
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
532
cpu_check_watchpoint(env_cpu(env), addr, msize,
533
info.attrs, BP_MEM_WRITE, retaddr);
534
}
535
- /* TODO: MTE check. */
536
+
537
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
538
+ mte_check1(env, mtedesc, addr, retaddr);
539
+ }
540
}
541
i += 1;
542
reg_off += esize;
543
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
544
} while (reg_off < reg_max);
545
}
546
547
-#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
548
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
549
- void *vm, target_ulong base, uint32_t desc) \
550
-{ \
551
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
552
- off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
553
+static inline QEMU_ALWAYS_INLINE
554
+void sve_st1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
555
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
556
+ int esize, int msize, zreg_off_fn *off_fn,
557
+ sve_ldst1_host_fn *host_fn,
558
+ sve_ldst1_tlb_fn *tlb_fn)
559
+{
560
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
561
+ /* Remove mtedesc from the normal sve descriptor. */
562
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
563
+
564
+ /*
565
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
566
+ * offset base entirely over the address space hole to change the
567
+ * pointer tag, or change the bit55 selector. So we could here
568
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
569
+ */
570
+ sve_st1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
571
+ esize, msize, off_fn, host_fn, tlb_fn);
572
}
573
574
-#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
575
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
576
+#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
577
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
578
void *vm, target_ulong base, uint32_t desc) \
579
-{ \
580
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
581
- off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
582
+{ \
583
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
584
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
585
+} \
586
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
587
+ void *vm, target_ulong base, uint32_t desc) \
588
+{ \
589
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
590
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
591
+}
592
+
593
+#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
594
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
595
+ void *vm, target_ulong base, uint32_t desc) \
596
+{ \
597
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
598
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
599
+} \
600
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
601
+ void *vm, target_ulong base, uint32_t desc) \
602
+{ \
603
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
604
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
605
}
606
607
DO_ST1_ZPZ_S(bs, zsu, MO_8)
608
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
609
index XXXXXXX..XXXXXXX 100644
610
--- a/target/arm/translate-sve.c
611
+++ b/target/arm/translate-sve.c
612
@@ -XXX,XX +XXX,XX @@ static bool trans_ST_zpri(DisasContext *s, arg_rpri_store *a)
613
*/
614
615
static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
616
- int scale, TCGv_i64 scalar, int msz,
617
+ int scale, TCGv_i64 scalar, int msz, bool is_write,
618
gen_helper_gvec_mem_scatter *fn)
619
{
620
unsigned vsz = vec_full_reg_size(s);
621
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
622
TCGv_ptr t_pg = tcg_temp_new_ptr();
623
TCGv_ptr t_zt = tcg_temp_new_ptr();
624
TCGv_i32 t_desc;
625
- int desc;
626
+ int desc = 0;
627
628
+ if (s->mte_active[0]) {
629
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
630
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
631
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
632
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
633
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
634
+ desc <<= SVE_MTEDESC_SHIFT;
635
+ }
636
desc = simd_desc(vsz, vsz, scale);
637
t_desc = tcg_const_i32(desc);
638
639
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
640
tcg_temp_free_i32(t_desc);
641
}
642
643
-/* Indexed by [be][ff][xs][u][msz]. */
644
-static gen_helper_gvec_mem_scatter * const gather_load_fn32[2][2][2][2][3] = {
645
- /* Little-endian */
646
- { { { { gen_helper_sve_ldbss_zsu,
647
- gen_helper_sve_ldhss_le_zsu,
648
- NULL, },
649
- { gen_helper_sve_ldbsu_zsu,
650
- gen_helper_sve_ldhsu_le_zsu,
651
- gen_helper_sve_ldss_le_zsu, } },
652
- { { gen_helper_sve_ldbss_zss,
653
- gen_helper_sve_ldhss_le_zss,
654
- NULL, },
655
- { gen_helper_sve_ldbsu_zss,
656
- gen_helper_sve_ldhsu_le_zss,
657
- gen_helper_sve_ldss_le_zss, } } },
658
+/* Indexed by [mte][be][ff][xs][u][msz]. */
659
+static gen_helper_gvec_mem_scatter * const
660
+gather_load_fn32[2][2][2][2][2][3] = {
661
+ { /* MTE Inactive */
662
+ { /* Little-endian */
663
+ { { { gen_helper_sve_ldbss_zsu,
664
+ gen_helper_sve_ldhss_le_zsu,
665
+ NULL, },
666
+ { gen_helper_sve_ldbsu_zsu,
667
+ gen_helper_sve_ldhsu_le_zsu,
668
+ gen_helper_sve_ldss_le_zsu, } },
669
+ { { gen_helper_sve_ldbss_zss,
670
+ gen_helper_sve_ldhss_le_zss,
671
+ NULL, },
672
+ { gen_helper_sve_ldbsu_zss,
673
+ gen_helper_sve_ldhsu_le_zss,
674
+ gen_helper_sve_ldss_le_zss, } } },
675
676
- /* First-fault */
677
- { { { gen_helper_sve_ldffbss_zsu,
678
- gen_helper_sve_ldffhss_le_zsu,
679
- NULL, },
680
- { gen_helper_sve_ldffbsu_zsu,
681
- gen_helper_sve_ldffhsu_le_zsu,
682
- gen_helper_sve_ldffss_le_zsu, } },
683
- { { gen_helper_sve_ldffbss_zss,
684
- gen_helper_sve_ldffhss_le_zss,
685
- NULL, },
686
- { gen_helper_sve_ldffbsu_zss,
687
- gen_helper_sve_ldffhsu_le_zss,
688
- gen_helper_sve_ldffss_le_zss, } } } },
689
+ /* First-fault */
690
+ { { { gen_helper_sve_ldffbss_zsu,
691
+ gen_helper_sve_ldffhss_le_zsu,
692
+ NULL, },
693
+ { gen_helper_sve_ldffbsu_zsu,
694
+ gen_helper_sve_ldffhsu_le_zsu,
695
+ gen_helper_sve_ldffss_le_zsu, } },
696
+ { { gen_helper_sve_ldffbss_zss,
697
+ gen_helper_sve_ldffhss_le_zss,
698
+ NULL, },
699
+ { gen_helper_sve_ldffbsu_zss,
700
+ gen_helper_sve_ldffhsu_le_zss,
701
+ gen_helper_sve_ldffss_le_zss, } } } },
702
703
- /* Big-endian */
704
- { { { { gen_helper_sve_ldbss_zsu,
705
- gen_helper_sve_ldhss_be_zsu,
706
- NULL, },
707
- { gen_helper_sve_ldbsu_zsu,
708
- gen_helper_sve_ldhsu_be_zsu,
709
- gen_helper_sve_ldss_be_zsu, } },
710
- { { gen_helper_sve_ldbss_zss,
711
- gen_helper_sve_ldhss_be_zss,
712
- NULL, },
713
- { gen_helper_sve_ldbsu_zss,
714
- gen_helper_sve_ldhsu_be_zss,
715
- gen_helper_sve_ldss_be_zss, } } },
716
+ { /* Big-endian */
717
+ { { { gen_helper_sve_ldbss_zsu,
718
+ gen_helper_sve_ldhss_be_zsu,
719
+ NULL, },
720
+ { gen_helper_sve_ldbsu_zsu,
721
+ gen_helper_sve_ldhsu_be_zsu,
722
+ gen_helper_sve_ldss_be_zsu, } },
723
+ { { gen_helper_sve_ldbss_zss,
724
+ gen_helper_sve_ldhss_be_zss,
725
+ NULL, },
726
+ { gen_helper_sve_ldbsu_zss,
727
+ gen_helper_sve_ldhsu_be_zss,
728
+ gen_helper_sve_ldss_be_zss, } } },
729
730
- /* First-fault */
731
- { { { gen_helper_sve_ldffbss_zsu,
732
- gen_helper_sve_ldffhss_be_zsu,
733
- NULL, },
734
- { gen_helper_sve_ldffbsu_zsu,
735
- gen_helper_sve_ldffhsu_be_zsu,
736
- gen_helper_sve_ldffss_be_zsu, } },
737
- { { gen_helper_sve_ldffbss_zss,
738
- gen_helper_sve_ldffhss_be_zss,
739
- NULL, },
740
- { gen_helper_sve_ldffbsu_zss,
741
- gen_helper_sve_ldffhsu_be_zss,
742
- gen_helper_sve_ldffss_be_zss, } } } },
743
+ /* First-fault */
744
+ { { { gen_helper_sve_ldffbss_zsu,
745
+ gen_helper_sve_ldffhss_be_zsu,
746
+ NULL, },
747
+ { gen_helper_sve_ldffbsu_zsu,
748
+ gen_helper_sve_ldffhsu_be_zsu,
749
+ gen_helper_sve_ldffss_be_zsu, } },
750
+ { { gen_helper_sve_ldffbss_zss,
751
+ gen_helper_sve_ldffhss_be_zss,
752
+ NULL, },
753
+ { gen_helper_sve_ldffbsu_zss,
754
+ gen_helper_sve_ldffhsu_be_zss,
755
+ gen_helper_sve_ldffss_be_zss, } } } } },
756
+ { /* MTE Active */
757
+ { /* Little-endian */
758
+ { { { gen_helper_sve_ldbss_zsu_mte,
759
+ gen_helper_sve_ldhss_le_zsu_mte,
760
+ NULL, },
761
+ { gen_helper_sve_ldbsu_zsu_mte,
762
+ gen_helper_sve_ldhsu_le_zsu_mte,
763
+ gen_helper_sve_ldss_le_zsu_mte, } },
764
+ { { gen_helper_sve_ldbss_zss_mte,
765
+ gen_helper_sve_ldhss_le_zss_mte,
766
+ NULL, },
767
+ { gen_helper_sve_ldbsu_zss_mte,
768
+ gen_helper_sve_ldhsu_le_zss_mte,
769
+ gen_helper_sve_ldss_le_zss_mte, } } },
770
+
771
+ /* First-fault */
772
+ { { { gen_helper_sve_ldffbss_zsu_mte,
773
+ gen_helper_sve_ldffhss_le_zsu_mte,
774
+ NULL, },
775
+ { gen_helper_sve_ldffbsu_zsu_mte,
776
+ gen_helper_sve_ldffhsu_le_zsu_mte,
777
+ gen_helper_sve_ldffss_le_zsu_mte, } },
778
+ { { gen_helper_sve_ldffbss_zss_mte,
779
+ gen_helper_sve_ldffhss_le_zss_mte,
780
+ NULL, },
781
+ { gen_helper_sve_ldffbsu_zss_mte,
782
+ gen_helper_sve_ldffhsu_le_zss_mte,
783
+ gen_helper_sve_ldffss_le_zss_mte, } } } },
784
+
785
+ { /* Big-endian */
786
+ { { { gen_helper_sve_ldbss_zsu_mte,
787
+ gen_helper_sve_ldhss_be_zsu_mte,
788
+ NULL, },
789
+ { gen_helper_sve_ldbsu_zsu_mte,
790
+ gen_helper_sve_ldhsu_be_zsu_mte,
791
+ gen_helper_sve_ldss_be_zsu_mte, } },
792
+ { { gen_helper_sve_ldbss_zss_mte,
793
+ gen_helper_sve_ldhss_be_zss_mte,
794
+ NULL, },
795
+ { gen_helper_sve_ldbsu_zss_mte,
796
+ gen_helper_sve_ldhsu_be_zss_mte,
797
+ gen_helper_sve_ldss_be_zss_mte, } } },
798
+
799
+ /* First-fault */
800
+ { { { gen_helper_sve_ldffbss_zsu_mte,
801
+ gen_helper_sve_ldffhss_be_zsu_mte,
802
+ NULL, },
803
+ { gen_helper_sve_ldffbsu_zsu_mte,
804
+ gen_helper_sve_ldffhsu_be_zsu_mte,
805
+ gen_helper_sve_ldffss_be_zsu_mte, } },
806
+ { { gen_helper_sve_ldffbss_zss_mte,
807
+ gen_helper_sve_ldffhss_be_zss_mte,
808
+ NULL, },
809
+ { gen_helper_sve_ldffbsu_zss_mte,
810
+ gen_helper_sve_ldffhsu_be_zss_mte,
811
+ gen_helper_sve_ldffss_be_zss_mte, } } } } },
812
};
813
814
/* Note that we overload xs=2 to indicate 64-bit offset. */
815
-static gen_helper_gvec_mem_scatter * const gather_load_fn64[2][2][3][2][4] = {
816
- /* Little-endian */
817
- { { { { gen_helper_sve_ldbds_zsu,
818
- gen_helper_sve_ldhds_le_zsu,
819
- gen_helper_sve_ldsds_le_zsu,
820
- NULL, },
821
- { gen_helper_sve_ldbdu_zsu,
822
- gen_helper_sve_ldhdu_le_zsu,
823
- gen_helper_sve_ldsdu_le_zsu,
824
- gen_helper_sve_lddd_le_zsu, } },
825
- { { gen_helper_sve_ldbds_zss,
826
- gen_helper_sve_ldhds_le_zss,
827
- gen_helper_sve_ldsds_le_zss,
828
- NULL, },
829
- { gen_helper_sve_ldbdu_zss,
830
- gen_helper_sve_ldhdu_le_zss,
831
- gen_helper_sve_ldsdu_le_zss,
832
- gen_helper_sve_lddd_le_zss, } },
833
- { { gen_helper_sve_ldbds_zd,
834
- gen_helper_sve_ldhds_le_zd,
835
- gen_helper_sve_ldsds_le_zd,
836
- NULL, },
837
- { gen_helper_sve_ldbdu_zd,
838
- gen_helper_sve_ldhdu_le_zd,
839
- gen_helper_sve_ldsdu_le_zd,
840
- gen_helper_sve_lddd_le_zd, } } },
841
+static gen_helper_gvec_mem_scatter * const
842
+gather_load_fn64[2][2][2][3][2][4] = {
843
+ { /* MTE Inactive */
844
+ { /* Little-endian */
845
+ { { { gen_helper_sve_ldbds_zsu,
846
+ gen_helper_sve_ldhds_le_zsu,
847
+ gen_helper_sve_ldsds_le_zsu,
848
+ NULL, },
849
+ { gen_helper_sve_ldbdu_zsu,
850
+ gen_helper_sve_ldhdu_le_zsu,
851
+ gen_helper_sve_ldsdu_le_zsu,
852
+ gen_helper_sve_lddd_le_zsu, } },
853
+ { { gen_helper_sve_ldbds_zss,
854
+ gen_helper_sve_ldhds_le_zss,
855
+ gen_helper_sve_ldsds_le_zss,
856
+ NULL, },
857
+ { gen_helper_sve_ldbdu_zss,
858
+ gen_helper_sve_ldhdu_le_zss,
859
+ gen_helper_sve_ldsdu_le_zss,
860
+ gen_helper_sve_lddd_le_zss, } },
861
+ { { gen_helper_sve_ldbds_zd,
862
+ gen_helper_sve_ldhds_le_zd,
863
+ gen_helper_sve_ldsds_le_zd,
864
+ NULL, },
865
+ { gen_helper_sve_ldbdu_zd,
866
+ gen_helper_sve_ldhdu_le_zd,
867
+ gen_helper_sve_ldsdu_le_zd,
868
+ gen_helper_sve_lddd_le_zd, } } },
869
870
- /* First-fault */
871
- { { { gen_helper_sve_ldffbds_zsu,
872
- gen_helper_sve_ldffhds_le_zsu,
873
- gen_helper_sve_ldffsds_le_zsu,
874
- NULL, },
875
- { gen_helper_sve_ldffbdu_zsu,
876
- gen_helper_sve_ldffhdu_le_zsu,
877
- gen_helper_sve_ldffsdu_le_zsu,
878
- gen_helper_sve_ldffdd_le_zsu, } },
879
- { { gen_helper_sve_ldffbds_zss,
880
- gen_helper_sve_ldffhds_le_zss,
881
- gen_helper_sve_ldffsds_le_zss,
882
- NULL, },
883
- { gen_helper_sve_ldffbdu_zss,
884
- gen_helper_sve_ldffhdu_le_zss,
885
- gen_helper_sve_ldffsdu_le_zss,
886
- gen_helper_sve_ldffdd_le_zss, } },
887
- { { gen_helper_sve_ldffbds_zd,
888
- gen_helper_sve_ldffhds_le_zd,
889
- gen_helper_sve_ldffsds_le_zd,
890
- NULL, },
891
- { gen_helper_sve_ldffbdu_zd,
892
- gen_helper_sve_ldffhdu_le_zd,
893
- gen_helper_sve_ldffsdu_le_zd,
894
- gen_helper_sve_ldffdd_le_zd, } } } },
895
+ /* First-fault */
896
+ { { { gen_helper_sve_ldffbds_zsu,
897
+ gen_helper_sve_ldffhds_le_zsu,
898
+ gen_helper_sve_ldffsds_le_zsu,
899
+ NULL, },
900
+ { gen_helper_sve_ldffbdu_zsu,
901
+ gen_helper_sve_ldffhdu_le_zsu,
902
+ gen_helper_sve_ldffsdu_le_zsu,
903
+ gen_helper_sve_ldffdd_le_zsu, } },
904
+ { { gen_helper_sve_ldffbds_zss,
905
+ gen_helper_sve_ldffhds_le_zss,
906
+ gen_helper_sve_ldffsds_le_zss,
907
+ NULL, },
908
+ { gen_helper_sve_ldffbdu_zss,
909
+ gen_helper_sve_ldffhdu_le_zss,
910
+ gen_helper_sve_ldffsdu_le_zss,
911
+ gen_helper_sve_ldffdd_le_zss, } },
912
+ { { gen_helper_sve_ldffbds_zd,
913
+ gen_helper_sve_ldffhds_le_zd,
914
+ gen_helper_sve_ldffsds_le_zd,
915
+ NULL, },
916
+ { gen_helper_sve_ldffbdu_zd,
917
+ gen_helper_sve_ldffhdu_le_zd,
918
+ gen_helper_sve_ldffsdu_le_zd,
919
+ gen_helper_sve_ldffdd_le_zd, } } } },
920
+ { /* Big-endian */
921
+ { { { gen_helper_sve_ldbds_zsu,
922
+ gen_helper_sve_ldhds_be_zsu,
923
+ gen_helper_sve_ldsds_be_zsu,
924
+ NULL, },
925
+ { gen_helper_sve_ldbdu_zsu,
926
+ gen_helper_sve_ldhdu_be_zsu,
927
+ gen_helper_sve_ldsdu_be_zsu,
928
+ gen_helper_sve_lddd_be_zsu, } },
929
+ { { gen_helper_sve_ldbds_zss,
930
+ gen_helper_sve_ldhds_be_zss,
931
+ gen_helper_sve_ldsds_be_zss,
932
+ NULL, },
933
+ { gen_helper_sve_ldbdu_zss,
934
+ gen_helper_sve_ldhdu_be_zss,
935
+ gen_helper_sve_ldsdu_be_zss,
936
+ gen_helper_sve_lddd_be_zss, } },
937
+ { { gen_helper_sve_ldbds_zd,
938
+ gen_helper_sve_ldhds_be_zd,
939
+ gen_helper_sve_ldsds_be_zd,
940
+ NULL, },
941
+ { gen_helper_sve_ldbdu_zd,
942
+ gen_helper_sve_ldhdu_be_zd,
943
+ gen_helper_sve_ldsdu_be_zd,
944
+ gen_helper_sve_lddd_be_zd, } } },
945
946
- /* Big-endian */
947
- { { { { gen_helper_sve_ldbds_zsu,
948
- gen_helper_sve_ldhds_be_zsu,
949
- gen_helper_sve_ldsds_be_zsu,
950
- NULL, },
951
- { gen_helper_sve_ldbdu_zsu,
952
- gen_helper_sve_ldhdu_be_zsu,
953
- gen_helper_sve_ldsdu_be_zsu,
954
- gen_helper_sve_lddd_be_zsu, } },
955
- { { gen_helper_sve_ldbds_zss,
956
- gen_helper_sve_ldhds_be_zss,
957
- gen_helper_sve_ldsds_be_zss,
958
- NULL, },
959
- { gen_helper_sve_ldbdu_zss,
960
- gen_helper_sve_ldhdu_be_zss,
961
- gen_helper_sve_ldsdu_be_zss,
962
- gen_helper_sve_lddd_be_zss, } },
963
- { { gen_helper_sve_ldbds_zd,
964
- gen_helper_sve_ldhds_be_zd,
965
- gen_helper_sve_ldsds_be_zd,
966
- NULL, },
967
- { gen_helper_sve_ldbdu_zd,
968
- gen_helper_sve_ldhdu_be_zd,
969
- gen_helper_sve_ldsdu_be_zd,
970
- gen_helper_sve_lddd_be_zd, } } },
971
+ /* First-fault */
972
+ { { { gen_helper_sve_ldffbds_zsu,
973
+ gen_helper_sve_ldffhds_be_zsu,
974
+ gen_helper_sve_ldffsds_be_zsu,
975
+ NULL, },
976
+ { gen_helper_sve_ldffbdu_zsu,
977
+ gen_helper_sve_ldffhdu_be_zsu,
978
+ gen_helper_sve_ldffsdu_be_zsu,
979
+ gen_helper_sve_ldffdd_be_zsu, } },
980
+ { { gen_helper_sve_ldffbds_zss,
981
+ gen_helper_sve_ldffhds_be_zss,
982
+ gen_helper_sve_ldffsds_be_zss,
983
+ NULL, },
984
+ { gen_helper_sve_ldffbdu_zss,
985
+ gen_helper_sve_ldffhdu_be_zss,
986
+ gen_helper_sve_ldffsdu_be_zss,
987
+ gen_helper_sve_ldffdd_be_zss, } },
988
+ { { gen_helper_sve_ldffbds_zd,
989
+ gen_helper_sve_ldffhds_be_zd,
990
+ gen_helper_sve_ldffsds_be_zd,
991
+ NULL, },
992
+ { gen_helper_sve_ldffbdu_zd,
993
+ gen_helper_sve_ldffhdu_be_zd,
994
+ gen_helper_sve_ldffsdu_be_zd,
995
+ gen_helper_sve_ldffdd_be_zd, } } } } },
996
+ { /* MTE Active */
997
+ { /* Little-endian */
998
+ { { { gen_helper_sve_ldbds_zsu_mte,
999
+ gen_helper_sve_ldhds_le_zsu_mte,
1000
+ gen_helper_sve_ldsds_le_zsu_mte,
1001
+ NULL, },
1002
+ { gen_helper_sve_ldbdu_zsu_mte,
1003
+ gen_helper_sve_ldhdu_le_zsu_mte,
1004
+ gen_helper_sve_ldsdu_le_zsu_mte,
1005
+ gen_helper_sve_lddd_le_zsu_mte, } },
1006
+ { { gen_helper_sve_ldbds_zss_mte,
1007
+ gen_helper_sve_ldhds_le_zss_mte,
1008
+ gen_helper_sve_ldsds_le_zss_mte,
1009
+ NULL, },
1010
+ { gen_helper_sve_ldbdu_zss_mte,
1011
+ gen_helper_sve_ldhdu_le_zss_mte,
1012
+ gen_helper_sve_ldsdu_le_zss_mte,
1013
+ gen_helper_sve_lddd_le_zss_mte, } },
1014
+ { { gen_helper_sve_ldbds_zd_mte,
1015
+ gen_helper_sve_ldhds_le_zd_mte,
1016
+ gen_helper_sve_ldsds_le_zd_mte,
1017
+ NULL, },
1018
+ { gen_helper_sve_ldbdu_zd_mte,
1019
+ gen_helper_sve_ldhdu_le_zd_mte,
1020
+ gen_helper_sve_ldsdu_le_zd_mte,
1021
+ gen_helper_sve_lddd_le_zd_mte, } } },
1022
1023
- /* First-fault */
1024
- { { { gen_helper_sve_ldffbds_zsu,
1025
- gen_helper_sve_ldffhds_be_zsu,
1026
- gen_helper_sve_ldffsds_be_zsu,
1027
- NULL, },
1028
- { gen_helper_sve_ldffbdu_zsu,
1029
- gen_helper_sve_ldffhdu_be_zsu,
1030
- gen_helper_sve_ldffsdu_be_zsu,
1031
- gen_helper_sve_ldffdd_be_zsu, } },
1032
- { { gen_helper_sve_ldffbds_zss,
1033
- gen_helper_sve_ldffhds_be_zss,
1034
- gen_helper_sve_ldffsds_be_zss,
1035
- NULL, },
1036
- { gen_helper_sve_ldffbdu_zss,
1037
- gen_helper_sve_ldffhdu_be_zss,
1038
- gen_helper_sve_ldffsdu_be_zss,
1039
- gen_helper_sve_ldffdd_be_zss, } },
1040
- { { gen_helper_sve_ldffbds_zd,
1041
- gen_helper_sve_ldffhds_be_zd,
1042
- gen_helper_sve_ldffsds_be_zd,
1043
- NULL, },
1044
- { gen_helper_sve_ldffbdu_zd,
1045
- gen_helper_sve_ldffhdu_be_zd,
1046
- gen_helper_sve_ldffsdu_be_zd,
1047
- gen_helper_sve_ldffdd_be_zd, } } } },
1048
+ /* First-fault */
1049
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1050
+ gen_helper_sve_ldffhds_le_zsu_mte,
1051
+ gen_helper_sve_ldffsds_le_zsu_mte,
1052
+ NULL, },
1053
+ { gen_helper_sve_ldffbdu_zsu_mte,
1054
+ gen_helper_sve_ldffhdu_le_zsu_mte,
1055
+ gen_helper_sve_ldffsdu_le_zsu_mte,
1056
+ gen_helper_sve_ldffdd_le_zsu_mte, } },
1057
+ { { gen_helper_sve_ldffbds_zss_mte,
1058
+ gen_helper_sve_ldffhds_le_zss_mte,
1059
+ gen_helper_sve_ldffsds_le_zss_mte,
1060
+ NULL, },
1061
+ { gen_helper_sve_ldffbdu_zss_mte,
1062
+ gen_helper_sve_ldffhdu_le_zss_mte,
1063
+ gen_helper_sve_ldffsdu_le_zss_mte,
1064
+ gen_helper_sve_ldffdd_le_zss_mte, } },
1065
+ { { gen_helper_sve_ldffbds_zd_mte,
1066
+ gen_helper_sve_ldffhds_le_zd_mte,
1067
+ gen_helper_sve_ldffsds_le_zd_mte,
1068
+ NULL, },
1069
+ { gen_helper_sve_ldffbdu_zd_mte,
1070
+ gen_helper_sve_ldffhdu_le_zd_mte,
1071
+ gen_helper_sve_ldffsdu_le_zd_mte,
1072
+ gen_helper_sve_ldffdd_le_zd_mte, } } } },
1073
+ { /* Big-endian */
1074
+ { { { gen_helper_sve_ldbds_zsu_mte,
1075
+ gen_helper_sve_ldhds_be_zsu_mte,
1076
+ gen_helper_sve_ldsds_be_zsu_mte,
1077
+ NULL, },
1078
+ { gen_helper_sve_ldbdu_zsu_mte,
1079
+ gen_helper_sve_ldhdu_be_zsu_mte,
1080
+ gen_helper_sve_ldsdu_be_zsu_mte,
1081
+ gen_helper_sve_lddd_be_zsu_mte, } },
1082
+ { { gen_helper_sve_ldbds_zss_mte,
1083
+ gen_helper_sve_ldhds_be_zss_mte,
1084
+ gen_helper_sve_ldsds_be_zss_mte,
1085
+ NULL, },
1086
+ { gen_helper_sve_ldbdu_zss_mte,
1087
+ gen_helper_sve_ldhdu_be_zss_mte,
1088
+ gen_helper_sve_ldsdu_be_zss_mte,
1089
+ gen_helper_sve_lddd_be_zss_mte, } },
1090
+ { { gen_helper_sve_ldbds_zd_mte,
1091
+ gen_helper_sve_ldhds_be_zd_mte,
1092
+ gen_helper_sve_ldsds_be_zd_mte,
1093
+ NULL, },
1094
+ { gen_helper_sve_ldbdu_zd_mte,
1095
+ gen_helper_sve_ldhdu_be_zd_mte,
1096
+ gen_helper_sve_ldsdu_be_zd_mte,
1097
+ gen_helper_sve_lddd_be_zd_mte, } } },
1098
+
1099
+ /* First-fault */
1100
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1101
+ gen_helper_sve_ldffhds_be_zsu_mte,
1102
+ gen_helper_sve_ldffsds_be_zsu_mte,
1103
+ NULL, },
1104
+ { gen_helper_sve_ldffbdu_zsu_mte,
1105
+ gen_helper_sve_ldffhdu_be_zsu_mte,
1106
+ gen_helper_sve_ldffsdu_be_zsu_mte,
1107
+ gen_helper_sve_ldffdd_be_zsu_mte, } },
1108
+ { { gen_helper_sve_ldffbds_zss_mte,
1109
+ gen_helper_sve_ldffhds_be_zss_mte,
1110
+ gen_helper_sve_ldffsds_be_zss_mte,
1111
+ NULL, },
1112
+ { gen_helper_sve_ldffbdu_zss_mte,
1113
+ gen_helper_sve_ldffhdu_be_zss_mte,
1114
+ gen_helper_sve_ldffsdu_be_zss_mte,
1115
+ gen_helper_sve_ldffdd_be_zss_mte, } },
1116
+ { { gen_helper_sve_ldffbds_zd_mte,
1117
+ gen_helper_sve_ldffhds_be_zd_mte,
1118
+ gen_helper_sve_ldffsds_be_zd_mte,
1119
+ NULL, },
1120
+ { gen_helper_sve_ldffbdu_zd_mte,
1121
+ gen_helper_sve_ldffhdu_be_zd_mte,
1122
+ gen_helper_sve_ldffsdu_be_zd_mte,
1123
+ gen_helper_sve_ldffdd_be_zd_mte, } } } } },
1124
};
1125
1126
static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1127
{
1128
gen_helper_gvec_mem_scatter *fn = NULL;
1129
- int be = s->be_data == MO_BE;
1130
+ bool be = s->be_data == MO_BE;
1131
+ bool mte = s->mte_active[0];
1132
1133
if (!sve_access_check(s)) {
1134
return true;
1135
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1136
1137
switch (a->esz) {
1138
case MO_32:
1139
- fn = gather_load_fn32[be][a->ff][a->xs][a->u][a->msz];
1140
+ fn = gather_load_fn32[mte][be][a->ff][a->xs][a->u][a->msz];
1141
break;
119
break;
1142
case MO_64:
120
case CP_ACCESS_TRAP_UNCATEGORIZED:
1143
- fn = gather_load_fn64[be][a->ff][a->xs][a->u][a->msz];
121
- target_el = exception_target_el(env);
1144
+ fn = gather_load_fn64[mte][be][a->ff][a->xs][a->u][a->msz];
122
- syndrome = syn_uncategorized();
1145
break;
123
- break;
1146
}
124
- case CP_ACCESS_TRAP_UNCATEGORIZED_EL2:
1147
assert(fn != NULL);
125
- target_el = 2;
1148
126
- syndrome = syn_uncategorized();
1149
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
127
- break;
1150
- cpu_reg_sp(s, a->rn), a->msz, fn);
128
- case CP_ACCESS_TRAP_UNCATEGORIZED_EL3:
1151
+ cpu_reg_sp(s, a->rn), a->msz, false, fn);
129
- target_el = 3;
1152
return true;
130
syndrome = syn_uncategorized();
1153
}
1154
1155
static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1156
{
1157
gen_helper_gvec_mem_scatter *fn = NULL;
1158
- int be = s->be_data == MO_BE;
1159
+ bool be = s->be_data == MO_BE;
1160
+ bool mte = s->mte_active[0];
1161
TCGv_i64 imm;
1162
1163
if (a->esz < a->msz || (a->esz == a->msz && !a->u)) {
1164
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1165
1166
switch (a->esz) {
1167
case MO_32:
1168
- fn = gather_load_fn32[be][a->ff][0][a->u][a->msz];
1169
+ fn = gather_load_fn32[mte][be][a->ff][0][a->u][a->msz];
1170
break;
1171
case MO_64:
1172
- fn = gather_load_fn64[be][a->ff][2][a->u][a->msz];
1173
+ fn = gather_load_fn64[mte][be][a->ff][2][a->u][a->msz];
1174
break;
1175
}
1176
assert(fn != NULL);
1177
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1178
* by loading the immediate into the scalar parameter.
1179
*/
1180
imm = tcg_const_i64(a->imm << a->msz);
1181
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1182
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, false, fn);
1183
tcg_temp_free_i64(imm);
1184
return true;
1185
}
1186
1187
-/* Indexed by [be][xs][msz]. */
1188
-static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][3] = {
1189
- /* Little-endian */
1190
- { { gen_helper_sve_stbs_zsu,
1191
- gen_helper_sve_sths_le_zsu,
1192
- gen_helper_sve_stss_le_zsu, },
1193
- { gen_helper_sve_stbs_zss,
1194
- gen_helper_sve_sths_le_zss,
1195
- gen_helper_sve_stss_le_zss, } },
1196
- /* Big-endian */
1197
- { { gen_helper_sve_stbs_zsu,
1198
- gen_helper_sve_sths_be_zsu,
1199
- gen_helper_sve_stss_be_zsu, },
1200
- { gen_helper_sve_stbs_zss,
1201
- gen_helper_sve_sths_be_zss,
1202
- gen_helper_sve_stss_be_zss, } },
1203
+/* Indexed by [mte][be][xs][msz]. */
1204
+static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = {
1205
+ { /* MTE Inactive */
1206
+ { /* Little-endian */
1207
+ { gen_helper_sve_stbs_zsu,
1208
+ gen_helper_sve_sths_le_zsu,
1209
+ gen_helper_sve_stss_le_zsu, },
1210
+ { gen_helper_sve_stbs_zss,
1211
+ gen_helper_sve_sths_le_zss,
1212
+ gen_helper_sve_stss_le_zss, } },
1213
+ { /* Big-endian */
1214
+ { gen_helper_sve_stbs_zsu,
1215
+ gen_helper_sve_sths_be_zsu,
1216
+ gen_helper_sve_stss_be_zsu, },
1217
+ { gen_helper_sve_stbs_zss,
1218
+ gen_helper_sve_sths_be_zss,
1219
+ gen_helper_sve_stss_be_zss, } } },
1220
+ { /* MTE Active */
1221
+ { /* Little-endian */
1222
+ { gen_helper_sve_stbs_zsu_mte,
1223
+ gen_helper_sve_sths_le_zsu_mte,
1224
+ gen_helper_sve_stss_le_zsu_mte, },
1225
+ { gen_helper_sve_stbs_zss_mte,
1226
+ gen_helper_sve_sths_le_zss_mte,
1227
+ gen_helper_sve_stss_le_zss_mte, } },
1228
+ { /* Big-endian */
1229
+ { gen_helper_sve_stbs_zsu_mte,
1230
+ gen_helper_sve_sths_be_zsu_mte,
1231
+ gen_helper_sve_stss_be_zsu_mte, },
1232
+ { gen_helper_sve_stbs_zss_mte,
1233
+ gen_helper_sve_sths_be_zss_mte,
1234
+ gen_helper_sve_stss_be_zss_mte, } } },
1235
};
1236
1237
/* Note that we overload xs=2 to indicate 64-bit offset. */
1238
-static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][3][4] = {
1239
- /* Little-endian */
1240
- { { gen_helper_sve_stbd_zsu,
1241
- gen_helper_sve_sthd_le_zsu,
1242
- gen_helper_sve_stsd_le_zsu,
1243
- gen_helper_sve_stdd_le_zsu, },
1244
- { gen_helper_sve_stbd_zss,
1245
- gen_helper_sve_sthd_le_zss,
1246
- gen_helper_sve_stsd_le_zss,
1247
- gen_helper_sve_stdd_le_zss, },
1248
- { gen_helper_sve_stbd_zd,
1249
- gen_helper_sve_sthd_le_zd,
1250
- gen_helper_sve_stsd_le_zd,
1251
- gen_helper_sve_stdd_le_zd, } },
1252
- /* Big-endian */
1253
- { { gen_helper_sve_stbd_zsu,
1254
- gen_helper_sve_sthd_be_zsu,
1255
- gen_helper_sve_stsd_be_zsu,
1256
- gen_helper_sve_stdd_be_zsu, },
1257
- { gen_helper_sve_stbd_zss,
1258
- gen_helper_sve_sthd_be_zss,
1259
- gen_helper_sve_stsd_be_zss,
1260
- gen_helper_sve_stdd_be_zss, },
1261
- { gen_helper_sve_stbd_zd,
1262
- gen_helper_sve_sthd_be_zd,
1263
- gen_helper_sve_stsd_be_zd,
1264
- gen_helper_sve_stdd_be_zd, } },
1265
+static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][2][3][4] = {
1266
+ { /* MTE Inactive */
1267
+ { /* Little-endian */
1268
+ { gen_helper_sve_stbd_zsu,
1269
+ gen_helper_sve_sthd_le_zsu,
1270
+ gen_helper_sve_stsd_le_zsu,
1271
+ gen_helper_sve_stdd_le_zsu, },
1272
+ { gen_helper_sve_stbd_zss,
1273
+ gen_helper_sve_sthd_le_zss,
1274
+ gen_helper_sve_stsd_le_zss,
1275
+ gen_helper_sve_stdd_le_zss, },
1276
+ { gen_helper_sve_stbd_zd,
1277
+ gen_helper_sve_sthd_le_zd,
1278
+ gen_helper_sve_stsd_le_zd,
1279
+ gen_helper_sve_stdd_le_zd, } },
1280
+ { /* Big-endian */
1281
+ { gen_helper_sve_stbd_zsu,
1282
+ gen_helper_sve_sthd_be_zsu,
1283
+ gen_helper_sve_stsd_be_zsu,
1284
+ gen_helper_sve_stdd_be_zsu, },
1285
+ { gen_helper_sve_stbd_zss,
1286
+ gen_helper_sve_sthd_be_zss,
1287
+ gen_helper_sve_stsd_be_zss,
1288
+ gen_helper_sve_stdd_be_zss, },
1289
+ { gen_helper_sve_stbd_zd,
1290
+ gen_helper_sve_sthd_be_zd,
1291
+ gen_helper_sve_stsd_be_zd,
1292
+ gen_helper_sve_stdd_be_zd, } } },
1293
+ { /* MTE Inactive */
1294
+ { /* Little-endian */
1295
+ { gen_helper_sve_stbd_zsu_mte,
1296
+ gen_helper_sve_sthd_le_zsu_mte,
1297
+ gen_helper_sve_stsd_le_zsu_mte,
1298
+ gen_helper_sve_stdd_le_zsu_mte, },
1299
+ { gen_helper_sve_stbd_zss_mte,
1300
+ gen_helper_sve_sthd_le_zss_mte,
1301
+ gen_helper_sve_stsd_le_zss_mte,
1302
+ gen_helper_sve_stdd_le_zss_mte, },
1303
+ { gen_helper_sve_stbd_zd_mte,
1304
+ gen_helper_sve_sthd_le_zd_mte,
1305
+ gen_helper_sve_stsd_le_zd_mte,
1306
+ gen_helper_sve_stdd_le_zd_mte, } },
1307
+ { /* Big-endian */
1308
+ { gen_helper_sve_stbd_zsu_mte,
1309
+ gen_helper_sve_sthd_be_zsu_mte,
1310
+ gen_helper_sve_stsd_be_zsu_mte,
1311
+ gen_helper_sve_stdd_be_zsu_mte, },
1312
+ { gen_helper_sve_stbd_zss_mte,
1313
+ gen_helper_sve_sthd_be_zss_mte,
1314
+ gen_helper_sve_stsd_be_zss_mte,
1315
+ gen_helper_sve_stdd_be_zss_mte, },
1316
+ { gen_helper_sve_stbd_zd_mte,
1317
+ gen_helper_sve_sthd_be_zd_mte,
1318
+ gen_helper_sve_stsd_be_zd_mte,
1319
+ gen_helper_sve_stdd_be_zd_mte, } } },
1320
};
1321
1322
static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1323
{
1324
gen_helper_gvec_mem_scatter *fn;
1325
- int be = s->be_data == MO_BE;
1326
+ bool be = s->be_data == MO_BE;
1327
+ bool mte = s->mte_active[0];
1328
1329
if (a->esz < a->msz || (a->msz == 0 && a->scale)) {
1330
return false;
1331
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1332
}
1333
switch (a->esz) {
1334
case MO_32:
1335
- fn = scatter_store_fn32[be][a->xs][a->msz];
1336
+ fn = scatter_store_fn32[mte][be][a->xs][a->msz];
1337
break;
1338
case MO_64:
1339
- fn = scatter_store_fn64[be][a->xs][a->msz];
1340
+ fn = scatter_store_fn64[mte][be][a->xs][a->msz];
1341
break;
131
break;
1342
default:
132
default:
1343
g_assert_not_reached();
133
g_assert_not_reached();
1344
}
134
}
1345
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
135
1346
- cpu_reg_sp(s, a->rn), a->msz, fn);
136
-exept:
1347
+ cpu_reg_sp(s, a->rn), a->msz, true, fn);
137
+ target_el = res & CP_ACCESS_EL_MASK;
1348
return true;
138
+ switch (target_el) {
139
+ case 0:
140
+ target_el = exception_target_el(env);
141
+ break;
142
+ case 2:
143
+ assert(arm_current_el(env) != 3);
144
+ assert(arm_is_el2_enabled(env));
145
+ break;
146
+ case 3:
147
+ assert(arm_feature(env, ARM_FEATURE_EL3));
148
+ break;
149
+ default:
150
+ /* No "direct" traps to EL1 */
151
+ g_assert_not_reached();
152
+ }
153
+
154
raise_exception(env, EXCP_UDEF, syndrome, target_el);
1349
}
155
}
1350
156
1351
static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1352
{
1353
gen_helper_gvec_mem_scatter *fn = NULL;
1354
- int be = s->be_data == MO_BE;
1355
+ bool be = s->be_data == MO_BE;
1356
+ bool mte = s->mte_active[0];
1357
TCGv_i64 imm;
1358
1359
if (a->esz < a->msz) {
1360
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1361
1362
switch (a->esz) {
1363
case MO_32:
1364
- fn = scatter_store_fn32[be][0][a->msz];
1365
+ fn = scatter_store_fn32[mte][be][0][a->msz];
1366
break;
1367
case MO_64:
1368
- fn = scatter_store_fn64[be][2][a->msz];
1369
+ fn = scatter_store_fn64[mte][be][2][a->msz];
1370
break;
1371
}
1372
assert(fn != NULL);
1373
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1374
* by loading the immediate into the scalar parameter.
1375
*/
1376
imm = tcg_const_i64(a->imm << a->msz);
1377
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1378
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, true, fn);
1379
tcg_temp_free_i64(imm);
1380
return true;
1381
}
1382
--
157
--
1383
2.20.1
158
2.25.1
1384
159
1385
160
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Remove a possible source of error by removing REGINFO_SENTINEL
4
and using ARRAY_SIZE (convinently hidden inside a macro) to
5
find the end of the set of regs being registered or modified.
6
7
The space saved by not having the extra array element reduces
8
the executable's .data.rel.ro section by about 9k.
9
10
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-16-richard.henderson@linaro.org
13
Message-id: 20220501055028.646596-4-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
15
---
8
target/arm/helper-a64.h | 7 ++
16
target/arm/cpregs.h | 53 +++++++++---------
9
target/arm/helper.h | 2 +
17
hw/arm/pxa2xx.c | 1 -
10
target/arm/mte_helper.c | 194 +++++++++++++++++++++++++++++++++++++
18
hw/arm/pxa2xx_pic.c | 1 -
11
target/arm/op_helper.c | 16 +++
19
hw/intc/arm_gicv3_cpuif.c | 5 --
12
target/arm/translate-a64.c | 172 +++++++++++++++++++++++++++++++-
20
hw/intc/arm_gicv3_kvm.c | 1 -
13
5 files changed, 386 insertions(+), 5 deletions(-)
21
target/arm/cpu64.c | 1 -
22
target/arm/cpu_tcg.c | 4 --
23
target/arm/helper.c | 111 ++++++++------------------------------
24
8 files changed, 48 insertions(+), 129 deletions(-)
14
25
15
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
26
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
16
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper-a64.h
28
--- a/target/arm/cpregs.h
18
+++ b/target/arm/helper-a64.h
29
+++ b/target/arm/cpregs.h
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
30
@@ -XXX,XX +XXX,XX @@
20
31
#define ARM_CP_NO_GDB 0x4000
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
32
#define ARM_CP_RAISES_EXC 0x8000
22
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
33
#define ARM_CP_NEWEL 0x10000
23
+DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
34
-/* Used only as a terminator for ARMCPRegInfo lists */
24
+DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
35
-#define ARM_CP_SENTINEL 0xfffff
25
+DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
36
/* Mask of only the flag bits in a type field */
26
+DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
37
#define ARM_CP_FLAG_MASK 0x1f0ff
27
+DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
38
28
+DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
39
@@ -XXX,XX +XXX,XX @@ enum {
29
+DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
40
ARM_CP_SECSTATE_NS = (1 << 1), /* bit[1]: Non-secure state register */
30
diff --git a/target/arm/helper.h b/target/arm/helper.h
41
};
31
index XXXXXXX..XXXXXXX 100644
42
32
--- a/target/arm/helper.h
43
-/*
33
+++ b/target/arm/helper.h
44
- * Return true if cptype is a valid type field. This is used to try to
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
45
- * catch errors where the sentinel has been accidentally left off the end
35
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
46
- * of a list of registers.
36
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
47
- */
37
48
-static inline bool cptype_valid(int cptype)
38
+DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
49
-{
39
+
50
- return ((cptype & ~ARM_CP_FLAG_MASK) == 0)
40
DEF_HELPER_1(vfp_get_fpscr, i32, env)
51
- || ((cptype & ARM_CP_SPECIAL) &&
41
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
52
- ((cptype & ~ARM_CP_FLAG_MASK) <= ARM_LAST_SPECIAL));
42
53
-}
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
54
-
44
index XXXXXXX..XXXXXXX 100644
55
/*
45
--- a/target/arm/mte_helper.c
56
* Access rights:
46
+++ b/target/arm/mte_helper.c
57
* We define bits for Read and Write access for what rev C of the v7-AR ARM ARM
47
@@ -XXX,XX +XXX,XX @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
58
@@ -XXX,XX +XXX,XX @@ struct ARMCPRegInfo {
48
return tag;
59
#define CPREG_FIELD64(env, ri) \
49
}
60
(*(uint64_t *)((char *)(env) + (ri)->fieldoffset))
50
61
51
+/**
62
-#define REGINFO_SENTINEL { .type = ARM_CP_SENTINEL }
52
+ * allocation_tag_mem:
63
+void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu, const ARMCPRegInfo *reg,
53
+ * @env: the cpu environment
64
+ void *opaque);
54
+ * @ptr_mmu_idx: the addressing regime to use for the virtual address
65
55
+ * @ptr: the virtual address for which to look up tag memory
66
-void define_arm_cp_regs_with_opaque(ARMCPU *cpu,
56
+ * @ptr_access: the access to use for the virtual address
67
- const ARMCPRegInfo *regs, void *opaque);
57
+ * @ptr_size: the number of bytes in the normal memory access
68
-void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
58
+ * @tag_access: the access to use for the tag memory
69
- const ARMCPRegInfo *regs, void *opaque);
59
+ * @tag_size: the number of bytes in the tag memory access
70
-static inline void define_arm_cp_regs(ARMCPU *cpu, const ARMCPRegInfo *regs)
60
+ * @ra: the return address for exception handling
71
-{
61
+ *
72
- define_arm_cp_regs_with_opaque(cpu, regs, 0);
62
+ * Our tag memory is formatted as a sequence of little-endian nibbles.
73
-}
63
+ * That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
74
static inline void define_one_arm_cp_reg(ARMCPU *cpu, const ARMCPRegInfo *regs)
64
+ * tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
65
+ * for the higher addr.
66
+ *
67
+ * Here, resolve the physical address from the virtual address, and return
68
+ * a pointer to the corresponding tag byte. Exit with exception if the
69
+ * virtual address is not accessible for @ptr_access.
70
+ *
71
+ * The @ptr_size and @tag_size values may not have an obvious relation
72
+ * due to the alignment of @ptr, and the number of tag checks required.
73
+ *
74
+ * If there is no tag storage corresponding to @ptr, return NULL.
75
+ */
76
+static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
77
+ uint64_t ptr, MMUAccessType ptr_access,
78
+ int ptr_size, MMUAccessType tag_access,
79
+ int tag_size, uintptr_t ra)
80
+{
81
+ /* Tag storage not implemented. */
82
+ return NULL;
83
+}
84
+
85
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
86
{
75
{
87
int rtag;
76
- define_one_arm_cp_reg_with_opaque(cpu, regs, 0);
88
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
77
+ define_one_arm_cp_reg_with_opaque(cpu, regs, NULL);
89
90
return address_with_allocation_tag(ptr + offset, rtag);
91
}
78
}
92
+
79
+
93
+static int load_tag1(uint64_t ptr, uint8_t *mem)
80
+void define_arm_cp_regs_with_opaque_len(ARMCPU *cpu, const ARMCPRegInfo *regs,
94
+{
81
+ void *opaque, size_t len);
95
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
96
+ return extract32(*mem, ofs, 4);
97
+}
98
+
82
+
99
+uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
83
+#define define_arm_cp_regs_with_opaque(CPU, REGS, OPAQUE) \
100
+{
84
+ do { \
101
+ int mmu_idx = cpu_mmu_index(env, false);
85
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(REGS) == 0); \
102
+ uint8_t *mem;
86
+ define_arm_cp_regs_with_opaque_len(CPU, REGS, OPAQUE, \
103
+ int rtag = 0;
87
+ ARRAY_SIZE(REGS)); \
88
+ } while (0)
104
+
89
+
105
+ /* Trap if accessing an invalid page. */
90
+#define define_arm_cp_regs(CPU, REGS) \
106
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
91
+ define_arm_cp_regs_with_opaque(CPU, REGS, NULL)
107
+ MMU_DATA_LOAD, 1, GETPC());
108
+
92
+
109
+ /* Load if page supports tags. */
93
const ARMCPRegInfo *get_arm_cp_reginfo(GHashTable *cpregs, uint32_t encoded_cp);
110
+ if (mem) {
94
111
+ rtag = load_tag1(ptr, mem);
95
/*
112
+ }
96
@@ -XXX,XX +XXX,XX @@ typedef struct ARMCPRegUserSpaceInfo {
97
uint64_t fixed_bits;
98
} ARMCPRegUserSpaceInfo;
99
100
-#define REGUSERINFO_SENTINEL { .name = NULL }
101
+void modify_arm_cp_regs_with_len(ARMCPRegInfo *regs, size_t regs_len,
102
+ const ARMCPRegUserSpaceInfo *mods,
103
+ size_t mods_len);
104
105
-void modify_arm_cp_regs(ARMCPRegInfo *regs, const ARMCPRegUserSpaceInfo *mods);
106
+#define modify_arm_cp_regs(REGS, MODS) \
107
+ do { \
108
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(REGS) == 0); \
109
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(MODS) == 0); \
110
+ modify_arm_cp_regs_with_len(REGS, ARRAY_SIZE(REGS), \
111
+ MODS, ARRAY_SIZE(MODS)); \
112
+ } while (0)
113
114
/* CPWriteFn that can be used to implement writes-ignored behaviour */
115
void arm_cp_write_ignore(CPUARMState *env, const ARMCPRegInfo *ri,
116
diff --git a/hw/arm/pxa2xx.c b/hw/arm/pxa2xx.c
117
index XXXXXXX..XXXXXXX 100644
118
--- a/hw/arm/pxa2xx.c
119
+++ b/hw/arm/pxa2xx.c
120
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pxa_cp_reginfo[] = {
121
{ .name = "PWRMODE", .cp = 14, .crn = 7, .crm = 0, .opc1 = 0, .opc2 = 0,
122
.access = PL1_RW, .type = ARM_CP_IO,
123
.readfn = arm_cp_read_zero, .writefn = pxa2xx_pwrmode_write },
124
- REGINFO_SENTINEL
125
};
126
127
static void pxa2xx_setup_cp14(PXA2xxState *s)
128
diff --git a/hw/arm/pxa2xx_pic.c b/hw/arm/pxa2xx_pic.c
129
index XXXXXXX..XXXXXXX 100644
130
--- a/hw/arm/pxa2xx_pic.c
131
+++ b/hw/arm/pxa2xx_pic.c
132
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pxa_pic_cp_reginfo[] = {
133
REGINFO_FOR_PIC_CP("ICLR2", 8),
134
REGINFO_FOR_PIC_CP("ICFP2", 9),
135
REGINFO_FOR_PIC_CP("ICPR2", 0xa),
136
- REGINFO_SENTINEL
137
};
138
139
static const MemoryRegionOps pxa2xx_pic_ops = {
140
diff --git a/hw/intc/arm_gicv3_cpuif.c b/hw/intc/arm_gicv3_cpuif.c
141
index XXXXXXX..XXXXXXX 100644
142
--- a/hw/intc/arm_gicv3_cpuif.c
143
+++ b/hw/intc/arm_gicv3_cpuif.c
144
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gicv3_cpuif_reginfo[] = {
145
.readfn = icc_igrpen1_el3_read,
146
.writefn = icc_igrpen1_el3_write,
147
},
148
- REGINFO_SENTINEL
149
};
150
151
static uint64_t ich_ap_read(CPUARMState *env, const ARMCPRegInfo *ri)
152
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gicv3_cpuif_hcr_reginfo[] = {
153
.readfn = ich_vmcr_read,
154
.writefn = ich_vmcr_write,
155
},
156
- REGINFO_SENTINEL
157
};
158
159
static const ARMCPRegInfo gicv3_cpuif_ich_apxr1_reginfo[] = {
160
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gicv3_cpuif_ich_apxr1_reginfo[] = {
161
.readfn = ich_ap_read,
162
.writefn = ich_ap_write,
163
},
164
- REGINFO_SENTINEL
165
};
166
167
static const ARMCPRegInfo gicv3_cpuif_ich_apxr23_reginfo[] = {
168
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gicv3_cpuif_ich_apxr23_reginfo[] = {
169
.readfn = ich_ap_read,
170
.writefn = ich_ap_write,
171
},
172
- REGINFO_SENTINEL
173
};
174
175
static void gicv3_cpuif_el_change_hook(ARMCPU *cpu, void *opaque)
176
@@ -XXX,XX +XXX,XX @@ void gicv3_init_cpuif(GICv3State *s)
177
.readfn = ich_lr_read,
178
.writefn = ich_lr_write,
179
},
180
- REGINFO_SENTINEL
181
};
182
define_arm_cp_regs(cpu, lr_regset);
183
}
184
diff --git a/hw/intc/arm_gicv3_kvm.c b/hw/intc/arm_gicv3_kvm.c
185
index XXXXXXX..XXXXXXX 100644
186
--- a/hw/intc/arm_gicv3_kvm.c
187
+++ b/hw/intc/arm_gicv3_kvm.c
188
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gicv3_cpuif_reginfo[] = {
189
*/
190
.resetfn = arm_gicv3_icc_reset,
191
},
192
- REGINFO_SENTINEL
193
};
194
195
/**
196
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
197
index XXXXXXX..XXXXXXX 100644
198
--- a/target/arm/cpu64.c
199
+++ b/target/arm/cpu64.c
200
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cortex_a72_a57_a53_cp_reginfo[] = {
201
{ .name = "L2MERRSR",
202
.cp = 15, .opc1 = 3, .crm = 15,
203
.access = PL1_RW, .type = ARM_CP_CONST | ARM_CP_64BIT, .resetvalue = 0 },
204
- REGINFO_SENTINEL
205
};
206
207
static void aarch64_a57_initfn(Object *obj)
208
diff --git a/target/arm/cpu_tcg.c b/target/arm/cpu_tcg.c
209
index XXXXXXX..XXXXXXX 100644
210
--- a/target/arm/cpu_tcg.c
211
+++ b/target/arm/cpu_tcg.c
212
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cortexa8_cp_reginfo[] = {
213
.access = PL1_RW, .type = ARM_CP_CONST, .resetvalue = 0 },
214
{ .name = "L2AUXCR", .cp = 15, .crn = 9, .crm = 0, .opc1 = 1, .opc2 = 2,
215
.access = PL1_RW, .type = ARM_CP_CONST, .resetvalue = 0 },
216
- REGINFO_SENTINEL
217
};
218
219
static void cortex_a8_initfn(Object *obj)
220
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cortexa9_cp_reginfo[] = {
221
.access = PL1_RW, .resetvalue = 0, .type = ARM_CP_CONST },
222
{ .name = "TLB_ATTR", .cp = 15, .crn = 15, .crm = 7, .opc1 = 5, .opc2 = 2,
223
.access = PL1_RW, .resetvalue = 0, .type = ARM_CP_CONST },
224
- REGINFO_SENTINEL
225
};
226
227
static void cortex_a9_initfn(Object *obj)
228
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cortexa15_cp_reginfo[] = {
229
#endif
230
{ .name = "L2ECTLR", .cp = 15, .crn = 9, .crm = 0, .opc1 = 1, .opc2 = 3,
231
.access = PL1_RW, .type = ARM_CP_CONST, .resetvalue = 0 },
232
- REGINFO_SENTINEL
233
};
234
235
static void cortex_a7_initfn(Object *obj)
236
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cortexr5_cp_reginfo[] = {
237
.access = PL1_RW, .type = ARM_CP_CONST },
238
{ .name = "DCACHE_INVAL", .cp = 15, .opc1 = 0, .crn = 15, .crm = 5,
239
.opc2 = 0, .access = PL1_W, .type = ARM_CP_NOP },
240
- REGINFO_SENTINEL
241
};
242
243
static void cortex_r5_initfn(Object *obj)
244
diff --git a/target/arm/helper.c b/target/arm/helper.c
245
index XXXXXXX..XXXXXXX 100644
246
--- a/target/arm/helper.c
247
+++ b/target/arm/helper.c
248
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cp_reginfo[] = {
249
.secure = ARM_CP_SECSTATE_S,
250
.fieldoffset = offsetof(CPUARMState, cp15.contextidr_s),
251
.resetvalue = 0, .writefn = contextidr_write, .raw_writefn = raw_write, },
252
- REGINFO_SENTINEL
253
};
254
255
static const ARMCPRegInfo not_v8_cp_reginfo[] = {
256
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo not_v8_cp_reginfo[] = {
257
{ .name = "CACHEMAINT", .cp = 15, .crn = 7, .crm = CP_ANY,
258
.opc1 = 0, .opc2 = CP_ANY, .access = PL1_W,
259
.type = ARM_CP_NOP | ARM_CP_OVERRIDE },
260
- REGINFO_SENTINEL
261
};
262
263
static const ARMCPRegInfo not_v6_cp_reginfo[] = {
264
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo not_v6_cp_reginfo[] = {
265
*/
266
{ .name = "WFI_v5", .cp = 15, .crn = 7, .crm = 8, .opc1 = 0, .opc2 = 2,
267
.access = PL1_W, .type = ARM_CP_WFI },
268
- REGINFO_SENTINEL
269
};
270
271
static const ARMCPRegInfo not_v7_cp_reginfo[] = {
272
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo not_v7_cp_reginfo[] = {
273
.opc1 = 0, .opc2 = 0, .access = PL1_RW, .type = ARM_CP_NOP },
274
{ .name = "NMRR", .cp = 15, .crn = 10, .crm = 2,
275
.opc1 = 0, .opc2 = 1, .access = PL1_RW, .type = ARM_CP_NOP },
276
- REGINFO_SENTINEL
277
};
278
279
static void cpacr_write(CPUARMState *env, const ARMCPRegInfo *ri,
280
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v6_cp_reginfo[] = {
281
.crn = 1, .crm = 0, .opc1 = 0, .opc2 = 2, .accessfn = cpacr_access,
282
.access = PL1_RW, .fieldoffset = offsetof(CPUARMState, cp15.cpacr_el1),
283
.resetfn = cpacr_reset, .writefn = cpacr_write, .readfn = cpacr_read },
284
- REGINFO_SENTINEL
285
};
286
287
typedef struct pm_event {
288
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v7_cp_reginfo[] = {
289
{ .name = "TLBIMVAA", .cp = 15, .opc1 = 0, .crn = 8, .crm = 7, .opc2 = 3,
290
.type = ARM_CP_NO_RAW, .access = PL1_W, .accessfn = access_ttlb,
291
.writefn = tlbimvaa_write },
292
- REGINFO_SENTINEL
293
};
294
295
static const ARMCPRegInfo v7mp_cp_reginfo[] = {
296
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v7mp_cp_reginfo[] = {
297
{ .name = "TLBIMVAAIS", .cp = 15, .opc1 = 0, .crn = 8, .crm = 3, .opc2 = 3,
298
.type = ARM_CP_NO_RAW, .access = PL1_W, .accessfn = access_ttlb,
299
.writefn = tlbimvaa_is_write },
300
- REGINFO_SENTINEL
301
};
302
303
static const ARMCPRegInfo pmovsset_cp_reginfo[] = {
304
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pmovsset_cp_reginfo[] = {
305
.fieldoffset = offsetof(CPUARMState, cp15.c9_pmovsr),
306
.writefn = pmovsset_write,
307
.raw_writefn = raw_write },
308
- REGINFO_SENTINEL
309
};
310
311
static void teecr_write(CPUARMState *env, const ARMCPRegInfo *ri,
312
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo t2ee_cp_reginfo[] = {
313
{ .name = "TEEHBR", .cp = 14, .crn = 1, .crm = 0, .opc1 = 6, .opc2 = 0,
314
.access = PL0_RW, .fieldoffset = offsetof(CPUARMState, teehbr),
315
.accessfn = teehbr_access, .resetvalue = 0 },
316
- REGINFO_SENTINEL
317
};
318
319
static const ARMCPRegInfo v6k_cp_reginfo[] = {
320
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v6k_cp_reginfo[] = {
321
.bank_fieldoffsets = { offsetoflow32(CPUARMState, cp15.tpidrprw_s),
322
offsetoflow32(CPUARMState, cp15.tpidrprw_ns) },
323
.resetvalue = 0 },
324
- REGINFO_SENTINEL
325
};
326
327
#ifndef CONFIG_USER_ONLY
328
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo generic_timer_cp_reginfo[] = {
329
.fieldoffset = offsetof(CPUARMState, cp15.c14_timer[GTIMER_SEC].cval),
330
.writefn = gt_sec_cval_write, .raw_writefn = raw_write,
331
},
332
- REGINFO_SENTINEL
333
};
334
335
static CPAccessResult e2h_access(CPUARMState *env, const ARMCPRegInfo *ri,
336
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo generic_timer_cp_reginfo[] = {
337
.access = PL0_R, .type = ARM_CP_NO_RAW | ARM_CP_IO,
338
.readfn = gt_virt_cnt_read,
339
},
340
- REGINFO_SENTINEL
341
};
342
343
#endif
344
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vapa_cp_reginfo[] = {
345
.access = PL1_W, .accessfn = ats_access,
346
.writefn = ats_write, .type = ARM_CP_NO_RAW | ARM_CP_RAISES_EXC },
347
#endif
348
- REGINFO_SENTINEL
349
};
350
351
/* Return basic MPU access permission bits. */
352
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pmsav7_cp_reginfo[] = {
353
.fieldoffset = offsetof(CPUARMState, pmsav7.rnr[M_REG_NS]),
354
.writefn = pmsav7_rgnr_write,
355
.resetfn = arm_cp_reset_ignore },
356
- REGINFO_SENTINEL
357
};
358
359
static const ARMCPRegInfo pmsav5_cp_reginfo[] = {
360
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pmsav5_cp_reginfo[] = {
361
{ .name = "946_PRBS7", .cp = 15, .crn = 6, .crm = 7, .opc1 = 0,
362
.opc2 = CP_ANY, .access = PL1_RW, .resetvalue = 0,
363
.fieldoffset = offsetof(CPUARMState, cp15.c6_region[7]) },
364
- REGINFO_SENTINEL
365
};
366
367
static void vmsa_ttbcr_raw_write(CPUARMState *env, const ARMCPRegInfo *ri,
368
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vmsa_pmsa_cp_reginfo[] = {
369
.access = PL1_RW, .accessfn = access_tvm_trvm,
370
.fieldoffset = offsetof(CPUARMState, cp15.far_el[1]),
371
.resetvalue = 0, },
372
- REGINFO_SENTINEL
373
};
374
375
static const ARMCPRegInfo vmsa_cp_reginfo[] = {
376
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vmsa_cp_reginfo[] = {
377
/* No offsetoflow32 -- pass the entire TCR to writefn/raw_writefn. */
378
.bank_fieldoffsets = { offsetof(CPUARMState, cp15.tcr_el[3]),
379
offsetof(CPUARMState, cp15.tcr_el[1])} },
380
- REGINFO_SENTINEL
381
};
382
383
/* Note that unlike TTBCR, writing to TTBCR2 does not require flushing
384
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo omap_cp_reginfo[] = {
385
{ .name = "C9", .cp = 15, .crn = 9,
386
.crm = CP_ANY, .opc1 = CP_ANY, .opc2 = CP_ANY, .access = PL1_RW,
387
.type = ARM_CP_CONST | ARM_CP_OVERRIDE, .resetvalue = 0 },
388
- REGINFO_SENTINEL
389
};
390
391
static void xscale_cpar_write(CPUARMState *env, const ARMCPRegInfo *ri,
392
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo xscale_cp_reginfo[] = {
393
{ .name = "XSCALE_UNLOCK_DCACHE",
394
.cp = 15, .opc1 = 0, .crn = 9, .crm = 2, .opc2 = 1,
395
.access = PL1_W, .type = ARM_CP_NOP },
396
- REGINFO_SENTINEL
397
};
398
399
static const ARMCPRegInfo dummy_c15_cp_reginfo[] = {
400
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dummy_c15_cp_reginfo[] = {
401
.access = PL1_RW,
402
.type = ARM_CP_CONST | ARM_CP_NO_RAW | ARM_CP_OVERRIDE,
403
.resetvalue = 0 },
404
- REGINFO_SENTINEL
405
};
406
407
static const ARMCPRegInfo cache_dirty_status_cp_reginfo[] = {
408
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cache_dirty_status_cp_reginfo[] = {
409
{ .name = "CDSR", .cp = 15, .crn = 7, .crm = 10, .opc1 = 0, .opc2 = 6,
410
.access = PL1_R, .type = ARM_CP_CONST | ARM_CP_NO_RAW,
411
.resetvalue = 0 },
412
- REGINFO_SENTINEL
413
};
414
415
static const ARMCPRegInfo cache_block_ops_cp_reginfo[] = {
416
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cache_block_ops_cp_reginfo[] = {
417
.access = PL0_W, .type = ARM_CP_NOP|ARM_CP_64BIT },
418
{ .name = "CIDCR", .cp = 15, .crm = 14, .opc1 = 0,
419
.access = PL1_W, .type = ARM_CP_NOP|ARM_CP_64BIT },
420
- REGINFO_SENTINEL
421
};
422
423
static const ARMCPRegInfo cache_test_clean_cp_reginfo[] = {
424
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo cache_test_clean_cp_reginfo[] = {
425
{ .name = "TCI_DCACHE", .cp = 15, .crn = 7, .crm = 14, .opc1 = 0, .opc2 = 3,
426
.access = PL0_R, .type = ARM_CP_CONST | ARM_CP_NO_RAW,
427
.resetvalue = (1 << 30) },
428
- REGINFO_SENTINEL
429
};
430
431
static const ARMCPRegInfo strongarm_cp_reginfo[] = {
432
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo strongarm_cp_reginfo[] = {
433
.crm = CP_ANY, .opc1 = CP_ANY, .opc2 = CP_ANY,
434
.access = PL1_RW, .resetvalue = 0,
435
.type = ARM_CP_CONST | ARM_CP_OVERRIDE | ARM_CP_NO_RAW },
436
- REGINFO_SENTINEL
437
};
438
439
static uint64_t midr_read(CPUARMState *env, const ARMCPRegInfo *ri)
440
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo lpae_cp_reginfo[] = {
441
.bank_fieldoffsets = { offsetof(CPUARMState, cp15.ttbr1_s),
442
offsetof(CPUARMState, cp15.ttbr1_ns) },
443
.writefn = vmsa_ttbr_write, },
444
- REGINFO_SENTINEL
445
};
446
447
static uint64_t aa64_fpcr_read(CPUARMState *env, const ARMCPRegInfo *ri)
448
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v8_cp_reginfo[] = {
449
.access = PL1_RW, .accessfn = access_trap_aa32s_el1,
450
.writefn = sdcr_write,
451
.fieldoffset = offsetoflow32(CPUARMState, cp15.mdcr_el3) },
452
- REGINFO_SENTINEL
453
};
454
455
/* Used to describe the behaviour of EL2 regs when EL2 does not exist. */
456
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el3_no_el2_cp_reginfo[] = {
457
.type = ARM_CP_CONST,
458
.cp = 15, .opc1 = 4, .crn = 6, .crm = 0, .opc2 = 2,
459
.access = PL2_RW, .resetvalue = 0 },
460
- REGINFO_SENTINEL
461
};
462
463
/* Ditto, but for registers which exist in ARMv8 but not v7 */
464
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el3_no_el2_v8_cp_reginfo[] = {
465
.cp = 15, .opc1 = 4, .crn = 1, .crm = 1, .opc2 = 4,
466
.access = PL2_RW,
467
.type = ARM_CP_CONST, .resetvalue = 0 },
468
- REGINFO_SENTINEL
469
};
470
471
static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
472
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el2_cp_reginfo[] = {
473
.cp = 15, .opc0 = 3, .opc1 = 4, .crn = 1, .crm = 1, .opc2 = 3,
474
.access = PL2_RW,
475
.fieldoffset = offsetof(CPUARMState, cp15.hstr_el2) },
476
- REGINFO_SENTINEL
477
};
478
479
static const ARMCPRegInfo el2_v8_cp_reginfo[] = {
480
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el2_v8_cp_reginfo[] = {
481
.access = PL2_RW,
482
.fieldoffset = offsetofhigh32(CPUARMState, cp15.hcr_el2),
483
.writefn = hcr_writehigh },
484
- REGINFO_SENTINEL
485
};
486
487
static CPAccessResult sel2_access(CPUARMState *env, const ARMCPRegInfo *ri,
488
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el2_sec_cp_reginfo[] = {
489
.opc0 = 3, .opc1 = 4, .crn = 2, .crm = 6, .opc2 = 2,
490
.access = PL2_RW, .accessfn = sel2_access,
491
.fieldoffset = offsetof(CPUARMState, cp15.vstcr_el2) },
492
- REGINFO_SENTINEL
493
};
494
495
static CPAccessResult nsacr_access(CPUARMState *env, const ARMCPRegInfo *ri,
496
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el3_cp_reginfo[] = {
497
.opc0 = 1, .opc1 = 6, .crn = 8, .crm = 7, .opc2 = 5,
498
.access = PL3_W, .type = ARM_CP_NO_RAW,
499
.writefn = tlbi_aa64_vae3_write },
500
- REGINFO_SENTINEL
501
};
502
503
#ifndef CONFIG_USER_ONLY
504
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo debug_cp_reginfo[] = {
505
.cp = 14, .opc0 = 2, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 0,
506
.access = PL1_RW, .accessfn = access_tda,
507
.type = ARM_CP_NOP },
508
- REGINFO_SENTINEL
509
};
510
511
static const ARMCPRegInfo debug_lpae_cp_reginfo[] = {
512
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo debug_lpae_cp_reginfo[] = {
513
.access = PL0_R, .type = ARM_CP_CONST|ARM_CP_64BIT, .resetvalue = 0 },
514
{ .name = "DBGDSAR", .cp = 14, .crm = 2, .opc1 = 0,
515
.access = PL0_R, .type = ARM_CP_CONST|ARM_CP_64BIT, .resetvalue = 0 },
516
- REGINFO_SENTINEL
517
};
518
519
/* Return the exception level to which exceptions should be taken
520
@@ -XXX,XX +XXX,XX @@ static void define_debug_regs(ARMCPU *cpu)
521
.fieldoffset = offsetof(CPUARMState, cp15.dbgbcr[i]),
522
.writefn = dbgbcr_write, .raw_writefn = raw_write
523
},
524
- REGINFO_SENTINEL
525
};
526
define_arm_cp_regs(cpu, dbgregs);
527
}
528
@@ -XXX,XX +XXX,XX @@ static void define_debug_regs(ARMCPU *cpu)
529
.fieldoffset = offsetof(CPUARMState, cp15.dbgwcr[i]),
530
.writefn = dbgwcr_write, .raw_writefn = raw_write
531
},
532
- REGINFO_SENTINEL
533
};
534
define_arm_cp_regs(cpu, dbgregs);
535
}
536
@@ -XXX,XX +XXX,XX @@ static void define_pmu_regs(ARMCPU *cpu)
537
.type = ARM_CP_IO,
538
.readfn = pmevtyper_readfn, .writefn = pmevtyper_writefn,
539
.raw_writefn = pmevtyper_rawwrite },
540
- REGINFO_SENTINEL
541
};
542
define_arm_cp_regs(cpu, pmev_regs);
543
g_free(pmevcntr_name);
544
@@ -XXX,XX +XXX,XX @@ static void define_pmu_regs(ARMCPU *cpu)
545
.cp = 15, .opc1 = 0, .crn = 9, .crm = 14, .opc2 = 5,
546
.access = PL0_R, .accessfn = pmreg_access, .type = ARM_CP_CONST,
547
.resetvalue = extract64(cpu->pmceid1, 32, 32) },
548
- REGINFO_SENTINEL
549
};
550
define_arm_cp_regs(cpu, v81_pmu_regs);
551
}
552
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo lor_reginfo[] = {
553
.opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7,
554
.access = PL1_R, .accessfn = access_lor_ns,
555
.type = ARM_CP_CONST, .resetvalue = 0 },
556
- REGINFO_SENTINEL
557
};
558
559
#ifdef TARGET_AARCH64
560
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pauth_reginfo[] = {
561
.opc0 = 3, .opc1 = 0, .crn = 2, .crm = 1, .opc2 = 3,
562
.access = PL1_RW, .accessfn = access_pauth,
563
.fieldoffset = offsetof(CPUARMState, keys.apib.hi) },
564
- REGINFO_SENTINEL
565
};
566
567
static const ARMCPRegInfo tlbirange_reginfo[] = {
568
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo tlbirange_reginfo[] = {
569
.opc0 = 1, .opc1 = 6, .crn = 8, .crm = 6, .opc2 = 5,
570
.access = PL3_W, .type = ARM_CP_NO_RAW,
571
.writefn = tlbi_aa64_rvae3_write },
572
- REGINFO_SENTINEL
573
};
574
575
static const ARMCPRegInfo tlbios_reginfo[] = {
576
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo tlbios_reginfo[] = {
577
.opc0 = 1, .opc1 = 6, .crn = 8, .crm = 1, .opc2 = 5,
578
.access = PL3_W, .type = ARM_CP_NO_RAW,
579
.writefn = tlbi_aa64_vae3is_write },
580
- REGINFO_SENTINEL
581
};
582
583
static uint64_t rndr_readfn(CPUARMState *env, const ARMCPRegInfo *ri)
584
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo rndr_reginfo[] = {
585
.type = ARM_CP_NO_RAW | ARM_CP_SUPPRESS_TB_END | ARM_CP_IO,
586
.opc0 = 3, .opc1 = 3, .crn = 2, .crm = 4, .opc2 = 1,
587
.access = PL0_R, .readfn = rndr_readfn },
588
- REGINFO_SENTINEL
589
};
590
591
#ifndef CONFIG_USER_ONLY
592
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpop_reg[] = {
593
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 1,
594
.access = PL0_W, .type = ARM_CP_NO_RAW | ARM_CP_SUPPRESS_TB_END,
595
.accessfn = aa64_cacheop_poc_access, .writefn = dccvap_writefn },
596
- REGINFO_SENTINEL
597
};
598
599
static const ARMCPRegInfo dcpodp_reg[] = {
600
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
601
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 1,
602
.access = PL0_W, .type = ARM_CP_NO_RAW | ARM_CP_SUPPRESS_TB_END,
603
.accessfn = aa64_cacheop_poc_access, .writefn = dccvap_writefn },
604
- REGINFO_SENTINEL
605
};
606
#endif /*CONFIG_USER_ONLY*/
607
608
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
609
{ .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
610
.opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
611
.type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
612
- REGINFO_SENTINEL
613
};
614
615
static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
616
{ .name = "TCO", .state = ARM_CP_STATE_AA64,
617
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
618
.type = ARM_CP_CONST, .access = PL0_RW, },
619
- REGINFO_SENTINEL
620
};
621
622
static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
623
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
624
.accessfn = aa64_zva_access,
625
#endif
626
},
627
- REGINFO_SENTINEL
628
};
629
630
#endif
631
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo predinv_reginfo[] = {
632
{ .name = "CPPRCTX", .state = ARM_CP_STATE_AA32,
633
.cp = 15, .opc1 = 0, .crn = 7, .crm = 3, .opc2 = 7,
634
.type = ARM_CP_NOP, .access = PL0_W, .accessfn = access_predinv },
635
- REGINFO_SENTINEL
636
};
637
638
static uint64_t ccsidr2_read(CPUARMState *env, const ARMCPRegInfo *ri)
639
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo ccsidr2_reginfo[] = {
640
.access = PL1_R,
641
.accessfn = access_aa64_tid2,
642
.readfn = ccsidr2_read, .type = ARM_CP_NO_RAW },
643
- REGINFO_SENTINEL
644
};
645
646
static CPAccessResult access_aa64_tid3(CPUARMState *env, const ARMCPRegInfo *ri,
647
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo jazelle_regs[] = {
648
.cp = 14, .crn = 2, .crm = 0, .opc1 = 7, .opc2 = 0,
649
.accessfn = access_joscr_jmcr,
650
.access = PL1_RW, .type = ARM_CP_CONST, .resetvalue = 0 },
651
- REGINFO_SENTINEL
652
};
653
654
static const ARMCPRegInfo vhe_reginfo[] = {
655
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vhe_reginfo[] = {
656
.access = PL2_RW, .accessfn = e2h_access,
657
.writefn = gt_virt_cval_write, .raw_writefn = raw_write },
658
#endif
659
- REGINFO_SENTINEL
660
};
661
662
#ifndef CONFIG_USER_ONLY
663
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo ats1e1_reginfo[] = {
664
.opc0 = 1, .opc1 = 0, .crn = 7, .crm = 9, .opc2 = 1,
665
.access = PL1_W, .type = ARM_CP_NO_RAW | ARM_CP_RAISES_EXC,
666
.writefn = ats_write64 },
667
- REGINFO_SENTINEL
668
};
669
670
static const ARMCPRegInfo ats1cp_reginfo[] = {
671
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo ats1cp_reginfo[] = {
672
.cp = 15, .opc1 = 0, .crn = 7, .crm = 9, .opc2 = 1,
673
.access = PL1_W, .type = ARM_CP_NO_RAW | ARM_CP_RAISES_EXC,
674
.writefn = ats_write },
675
- REGINFO_SENTINEL
676
};
677
#endif
678
679
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo actlr2_hactlr2_reginfo[] = {
680
.cp = 15, .opc1 = 4, .crn = 1, .crm = 0, .opc2 = 3,
681
.access = PL2_RW, .type = ARM_CP_CONST,
682
.resetvalue = 0 },
683
- REGINFO_SENTINEL
684
};
685
686
void register_cp_regs_for_features(ARMCPU *cpu)
687
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
688
.access = PL1_R, .type = ARM_CP_CONST,
689
.accessfn = access_aa32_tid3,
690
.resetvalue = cpu->isar.id_isar6 },
691
- REGINFO_SENTINEL
692
};
693
define_arm_cp_regs(cpu, v6_idregs);
694
define_arm_cp_regs(cpu, v6_cp_reginfo);
695
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
696
.opc0 = 3, .opc1 = 3, .crn = 9, .crm = 12, .opc2 = 7,
697
.access = PL0_R, .accessfn = pmreg_access, .type = ARM_CP_CONST,
698
.resetvalue = cpu->pmceid1 },
699
- REGINFO_SENTINEL
700
};
701
#ifdef CONFIG_USER_ONLY
702
ARMCPRegUserSpaceInfo v8_user_idregs[] = {
703
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
704
.exported_bits = 0x000000f0ffffffff },
705
{ .name = "ID_AA64ISAR*_EL1_RESERVED",
706
.is_glob = true },
707
- REGUSERINFO_SENTINEL
708
};
709
modify_arm_cp_regs(v8_idregs, v8_user_idregs);
710
#endif
711
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
712
.access = PL2_RW,
713
.resetvalue = vmpidr_def,
714
.fieldoffset = offsetof(CPUARMState, cp15.vmpidr_el2) },
715
- REGINFO_SENTINEL
716
};
717
define_arm_cp_regs(cpu, vpidr_regs);
718
define_arm_cp_regs(cpu, el2_cp_reginfo);
719
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
720
.access = PL2_RW, .accessfn = access_el3_aa32ns,
721
.type = ARM_CP_NO_RAW,
722
.writefn = arm_cp_write_ignore, .readfn = mpidr_read },
723
- REGINFO_SENTINEL
724
};
725
define_arm_cp_regs(cpu, vpidr_regs);
726
define_arm_cp_regs(cpu, el3_no_el2_cp_reginfo);
727
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
728
.raw_writefn = raw_write, .writefn = sctlr_write,
729
.fieldoffset = offsetof(CPUARMState, cp15.sctlr_el[3]),
730
.resetvalue = cpu->reset_sctlr },
731
- REGINFO_SENTINEL
732
};
733
734
define_arm_cp_regs(cpu, el3_regs);
735
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
736
{ .name = "DUMMY",
737
.cp = 15, .crn = 0, .crm = 7, .opc1 = 0, .opc2 = CP_ANY,
738
.access = PL1_R, .type = ARM_CP_CONST, .resetvalue = 0 },
739
- REGINFO_SENTINEL
740
};
741
ARMCPRegInfo id_v8_midr_cp_reginfo[] = {
742
{ .name = "MIDR_EL1", .state = ARM_CP_STATE_BOTH,
743
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
744
.access = PL1_R,
745
.accessfn = access_aa64_tid1,
746
.type = ARM_CP_CONST, .resetvalue = cpu->revidr },
747
- REGINFO_SENTINEL
748
};
749
ARMCPRegInfo id_cp_reginfo[] = {
750
/* These are common to v8 and pre-v8 */
751
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
752
.access = PL1_R,
753
.accessfn = access_aa32_tid1,
754
.type = ARM_CP_CONST, .resetvalue = 0 },
755
- REGINFO_SENTINEL
756
};
757
/* TLBTR is specific to VMSA */
758
ARMCPRegInfo id_tlbtr_reginfo = {
759
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
760
{ .name = "MIDR_EL1",
761
.exported_bits = 0x00000000ffffffff },
762
{ .name = "REVIDR_EL1" },
763
- REGUSERINFO_SENTINEL
764
};
765
modify_arm_cp_regs(id_v8_midr_cp_reginfo, id_v8_user_midr_cp_reginfo);
766
#endif
767
if (arm_feature(env, ARM_FEATURE_OMAPCP) ||
768
arm_feature(env, ARM_FEATURE_STRONGARM)) {
769
- ARMCPRegInfo *r;
770
+ size_t i;
771
/* Register the blanket "writes ignored" value first to cover the
772
* whole space. Then update the specific ID registers to allow write
773
* access, so that they ignore writes rather than causing them to
774
* UNDEF.
775
*/
776
define_one_arm_cp_reg(cpu, &crn0_wi_reginfo);
777
- for (r = id_pre_v8_midr_cp_reginfo;
778
- r->type != ARM_CP_SENTINEL; r++) {
779
- r->access = PL1_RW;
780
+ for (i = 0; i < ARRAY_SIZE(id_pre_v8_midr_cp_reginfo); ++i) {
781
+ id_pre_v8_midr_cp_reginfo[i].access = PL1_RW;
782
}
783
- for (r = id_cp_reginfo; r->type != ARM_CP_SENTINEL; r++) {
784
- r->access = PL1_RW;
785
+ for (i = 0; i < ARRAY_SIZE(id_cp_reginfo); ++i) {
786
+ id_cp_reginfo[i].access = PL1_RW;
787
}
788
id_mpuir_reginfo.access = PL1_RW;
789
id_tlbtr_reginfo.access = PL1_RW;
790
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
791
{ .name = "MPIDR_EL1", .state = ARM_CP_STATE_BOTH,
792
.opc0 = 3, .crn = 0, .crm = 0, .opc1 = 0, .opc2 = 5,
793
.access = PL1_R, .readfn = mpidr_read, .type = ARM_CP_NO_RAW },
794
- REGINFO_SENTINEL
795
};
796
#ifdef CONFIG_USER_ONLY
797
ARMCPRegUserSpaceInfo mpidr_user_cp_reginfo[] = {
798
{ .name = "MPIDR_EL1",
799
.fixed_bits = 0x0000000080000000 },
800
- REGUSERINFO_SENTINEL
801
};
802
modify_arm_cp_regs(mpidr_cp_reginfo, mpidr_user_cp_reginfo);
803
#endif
804
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
805
.opc0 = 3, .opc1 = 6, .crn = 1, .crm = 0, .opc2 = 1,
806
.access = PL3_RW, .type = ARM_CP_CONST,
807
.resetvalue = 0 },
808
- REGINFO_SENTINEL
809
};
810
define_arm_cp_regs(cpu, auxcr_reginfo);
811
if (cpu_isar_feature(aa32_ac2, cpu)) {
812
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
813
.type = ARM_CP_CONST,
814
.opc0 = 3, .opc1 = 1, .crn = 15, .crm = 3, .opc2 = 0,
815
.access = PL1_R, .resetvalue = cpu->reset_cbar },
816
- REGINFO_SENTINEL
817
};
818
/* We don't implement a r/w 64 bit CBAR currently */
819
assert(arm_feature(env, ARM_FEATURE_CBAR_RO));
820
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
821
.bank_fieldoffsets = { offsetof(CPUARMState, cp15.vbar_s),
822
offsetof(CPUARMState, cp15.vbar_ns) },
823
.resetvalue = 0 },
824
- REGINFO_SENTINEL
825
};
826
define_arm_cp_regs(cpu, vbar_cp_reginfo);
827
}
828
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
829
r->writefn);
830
}
831
}
832
- /* Bad type field probably means missing sentinel at end of reg list */
833
- assert(cptype_valid(r->type));
113
+
834
+
114
+ return address_with_allocation_tag(xt, rtag);
835
for (crm = crmmin; crm <= crmmax; crm++) {
115
+}
836
for (opc1 = opc1min; opc1 <= opc1max; opc1++) {
116
+
837
for (opc2 = opc2min; opc2 <= opc2max; opc2++) {
117
+static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
838
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
118
+{
119
+ if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
120
+ arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
121
+ cpu_mmu_index(env, false), ra);
122
+ g_assert_not_reached();
123
+ }
124
+}
125
+
126
+/* For use in a non-parallel context, store to the given nibble. */
127
+static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
128
+{
129
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
130
+ *mem = deposit32(*mem, ofs, 4, tag);
131
+}
132
+
133
+/* For use in a parallel context, atomically store to the given nibble. */
134
+static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
135
+{
136
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
137
+ uint8_t old = atomic_read(mem);
138
+
139
+ while (1) {
140
+ uint8_t new = deposit32(old, ofs, 4, tag);
141
+ uint8_t cmp = atomic_cmpxchg(mem, old, new);
142
+ if (likely(cmp == old)) {
143
+ return;
144
+ }
145
+ old = cmp;
146
+ }
147
+}
148
+
149
+typedef void stg_store1(uint64_t, uint8_t *, int);
150
+
151
+static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
152
+ uintptr_t ra, stg_store1 store1)
153
+{
154
+ int mmu_idx = cpu_mmu_index(env, false);
155
+ uint8_t *mem;
156
+
157
+ check_tag_aligned(env, ptr, ra);
158
+
159
+ /* Trap if accessing an invalid page. */
160
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
161
+ MMU_DATA_STORE, 1, ra);
162
+
163
+ /* Store if page supports tags. */
164
+ if (mem) {
165
+ store1(ptr, mem, allocation_tag_from_addr(xt));
166
+ }
167
+}
168
+
169
+void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
170
+{
171
+ do_stg(env, ptr, xt, GETPC(), store_tag1);
172
+}
173
+
174
+void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
175
+{
176
+ do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
177
+}
178
+
179
+void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
180
+{
181
+ int mmu_idx = cpu_mmu_index(env, false);
182
+ uintptr_t ra = GETPC();
183
+
184
+ check_tag_aligned(env, ptr, ra);
185
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
186
+}
187
+
188
+static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
189
+ uintptr_t ra, stg_store1 store1)
190
+{
191
+ int mmu_idx = cpu_mmu_index(env, false);
192
+ int tag = allocation_tag_from_addr(xt);
193
+ uint8_t *mem1, *mem2;
194
+
195
+ check_tag_aligned(env, ptr, ra);
196
+
197
+ /*
198
+ * Trap if accessing an invalid page(s).
199
+ * This takes priority over !allocation_tag_access_enabled.
200
+ */
201
+ if (ptr & TAG_GRANULE) {
202
+ /* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
203
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
204
+ TAG_GRANULE, MMU_DATA_STORE, 1, ra);
205
+ mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
206
+ MMU_DATA_STORE, TAG_GRANULE,
207
+ MMU_DATA_STORE, 1, ra);
208
+
209
+ /* Store if page(s) support tags. */
210
+ if (mem1) {
211
+ store1(TAG_GRANULE, mem1, tag);
212
+ }
213
+ if (mem2) {
214
+ store1(0, mem2, tag);
215
+ }
216
+ } else {
217
+ /* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
218
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
219
+ 2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
220
+ if (mem1) {
221
+ tag |= tag << 4;
222
+ atomic_set(mem1, tag);
223
+ }
224
+ }
225
+}
226
+
227
+void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
228
+{
229
+ do_st2g(env, ptr, xt, GETPC(), store_tag1);
230
+}
231
+
232
+void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
233
+{
234
+ do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
235
+}
236
+
237
+void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
238
+{
239
+ int mmu_idx = cpu_mmu_index(env, false);
240
+ uintptr_t ra = GETPC();
241
+ int in_page = -(ptr | TARGET_PAGE_MASK);
242
+
243
+ check_tag_aligned(env, ptr, ra);
244
+
245
+ if (likely(in_page >= 2 * TAG_GRANULE)) {
246
+ probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
247
+ } else {
248
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
249
+ probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
250
+ }
251
+}
252
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
253
index XXXXXXX..XXXXXXX 100644
254
--- a/target/arm/op_helper.c
255
+++ b/target/arm/op_helper.c
256
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
257
return ((uint32_t)x >> shift) | (x << (32 - shift));
258
}
839
}
259
}
840
}
260
+
841
261
+void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
842
-void define_arm_cp_regs_with_opaque(ARMCPU *cpu,
262
+ uint32_t access_type, uint32_t mmu_idx,
843
- const ARMCPRegInfo *regs, void *opaque)
263
+ uint32_t size)
844
+/* Define a whole list of registers */
264
+{
845
+void define_arm_cp_regs_with_opaque_len(ARMCPU *cpu, const ARMCPRegInfo *regs,
265
+ uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
846
+ void *opaque, size_t len)
266
+ uintptr_t ra = GETPC();
847
{
267
+
848
- /* Define a whole list of registers */
268
+ if (likely(size <= in_page)) {
849
- const ARMCPRegInfo *r;
269
+ probe_access(env, ptr, size, access_type, mmu_idx, ra);
850
- for (r = regs; r->type != ARM_CP_SENTINEL; r++) {
270
+ } else {
851
- define_one_arm_cp_reg_with_opaque(cpu, r, opaque);
271
+ probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
852
+ size_t i;
272
+ probe_access(env, ptr + in_page, size - in_page,
853
+ for (i = 0; i < len; ++i) {
273
+ access_type, mmu_idx, ra);
854
+ define_one_arm_cp_reg_with_opaque(cpu, regs + i, opaque);
274
+ }
275
+}
276
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
277
index XXXXXXX..XXXXXXX 100644
278
--- a/target/arm/translate-a64.c
279
+++ b/target/arm/translate-a64.c
280
@@ -XXX,XX +XXX,XX @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
281
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
282
}
283
284
+static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
285
+ MMUAccessType acc, int log2_size)
286
+{
287
+ TCGv_i32 t_acc = tcg_const_i32(acc);
288
+ TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
289
+ TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
290
+
291
+ gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
292
+ tcg_temp_free_i32(t_acc);
293
+ tcg_temp_free_i32(t_idx);
294
+ tcg_temp_free_i32(t_size);
295
+}
296
+
297
typedef struct DisasCompare64 {
298
TCGCond cond;
299
TCGv_i64 value;
300
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
301
}
855
}
302
}
856
}
303
857
304
+/*
858
@@ -XXX,XX +XXX,XX @@ void define_arm_cp_regs_with_opaque(ARMCPU *cpu,
305
+ * Load/Store memory tags
859
* user-space cannot alter any values and dynamic values pertaining to
306
+ *
860
* execution state are hidden from user space view anyway.
307
+ * 31 30 29 24 22 21 12 10 5 0
861
*/
308
+ * +-----+-------------+-----+---+------+-----+------+------+
862
-void modify_arm_cp_regs(ARMCPRegInfo *regs, const ARMCPRegUserSpaceInfo *mods)
309
+ * | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
863
+void modify_arm_cp_regs_with_len(ARMCPRegInfo *regs, size_t regs_len,
310
+ * +-----+-------------+-----+---+------+-----+------+------+
864
+ const ARMCPRegUserSpaceInfo *mods,
311
+ */
865
+ size_t mods_len)
312
+static void disas_ldst_tag(DisasContext *s, uint32_t insn)
866
{
313
+{
867
- const ARMCPRegUserSpaceInfo *m;
314
+ int rt = extract32(insn, 0, 5);
868
- ARMCPRegInfo *r;
315
+ int rn = extract32(insn, 5, 5);
869
-
316
+ uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
870
- for (m = mods; m->name; m++) {
317
+ int op2 = extract32(insn, 10, 2);
871
+ for (size_t mi = 0; mi < mods_len; ++mi) {
318
+ int op1 = extract32(insn, 22, 2);
872
+ const ARMCPRegUserSpaceInfo *m = mods + mi;
319
+ bool is_load = false, is_pair = false, is_zero = false;
873
GPatternSpec *pat = NULL;
320
+ int index = 0;
321
+ TCGv_i64 addr, clean_addr, tcg_rt;
322
+
874
+
323
+ /* We checked insn bits [29:24,21] in the caller. */
875
if (m->is_glob) {
324
+ if (extract32(insn, 30, 2) != 3) {
876
pat = g_pattern_spec_new(m->name);
325
+ goto do_unallocated;
877
}
326
+ }
878
- for (r = regs; r->type != ARM_CP_SENTINEL; r++) {
879
+ for (size_t ri = 0; ri < regs_len; ++ri) {
880
+ ARMCPRegInfo *r = regs + ri;
327
+
881
+
328
+ /*
882
if (pat && g_pattern_match_string(pat, r->name)) {
329
+ * @index is a tri-state variable which has 3 states:
883
r->type = ARM_CP_CONST;
330
+ * < 0 : post-index, writeback
884
r->access = PL0U_R;
331
+ * = 0 : signed offset
332
+ * > 0 : pre-index, writeback
333
+ */
334
+ switch (op1) {
335
+ case 0:
336
+ if (op2 != 0) {
337
+ /* STG */
338
+ index = op2 - 2;
339
+ break;
340
+ }
341
+ goto do_unallocated;
342
+ case 1:
343
+ if (op2 != 0) {
344
+ /* STZG */
345
+ is_zero = true;
346
+ index = op2 - 2;
347
+ } else {
348
+ /* LDG */
349
+ is_load = true;
350
+ }
351
+ break;
352
+ case 2:
353
+ if (op2 != 0) {
354
+ /* ST2G */
355
+ is_pair = true;
356
+ index = op2 - 2;
357
+ break;
358
+ }
359
+ goto do_unallocated;
360
+ case 3:
361
+ if (op2 != 0) {
362
+ /* STZ2G */
363
+ is_pair = is_zero = true;
364
+ index = op2 - 2;
365
+ break;
366
+ }
367
+ goto do_unallocated;
368
+
369
+ default:
370
+ do_unallocated:
371
+ unallocated_encoding(s);
372
+ return;
373
+ }
374
+
375
+ if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
376
+ goto do_unallocated;
377
+ }
378
+
379
+ if (rn == 31) {
380
+ gen_check_sp_alignment(s);
381
+ }
382
+
383
+ addr = read_cpu_reg_sp(s, rn, true);
384
+ if (index >= 0) {
385
+ /* pre-index or signed offset */
386
+ tcg_gen_addi_i64(addr, addr, offset);
387
+ }
388
+
389
+ if (is_load) {
390
+ tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
391
+ tcg_rt = cpu_reg(s, rt);
392
+ if (s->ata) {
393
+ gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
394
+ } else {
395
+ clean_addr = clean_data_tbi(s, addr);
396
+ gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
397
+ gen_address_with_allocation_tag0(tcg_rt, addr);
398
+ }
399
+ } else {
400
+ tcg_rt = cpu_reg_sp(s, rt);
401
+ if (!s->ata) {
402
+ /*
403
+ * For STG and ST2G, we need to check alignment and probe memory.
404
+ * TODO: For STZG and STZ2G, we could rely on the stores below,
405
+ * at least for system mode; user-only won't enforce alignment.
406
+ */
407
+ if (is_pair) {
408
+ gen_helper_st2g_stub(cpu_env, addr);
409
+ } else {
410
+ gen_helper_stg_stub(cpu_env, addr);
411
+ }
412
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
413
+ if (is_pair) {
414
+ gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
415
+ } else {
416
+ gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
417
+ }
418
+ } else {
419
+ if (is_pair) {
420
+ gen_helper_st2g(cpu_env, addr, tcg_rt);
421
+ } else {
422
+ gen_helper_stg(cpu_env, addr, tcg_rt);
423
+ }
424
+ }
425
+ }
426
+
427
+ if (is_zero) {
428
+ TCGv_i64 clean_addr = clean_data_tbi(s, addr);
429
+ TCGv_i64 tcg_zero = tcg_const_i64(0);
430
+ int mem_index = get_mem_index(s);
431
+ int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
432
+
433
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
434
+ MO_Q | MO_ALIGN_16);
435
+ for (i = 8; i < n; i += 8) {
436
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
437
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
438
+ }
439
+ tcg_temp_free_i64(tcg_zero);
440
+ }
441
+
442
+ if (index != 0) {
443
+ /* pre-index or post-index */
444
+ if (index < 0) {
445
+ /* post-index */
446
+ tcg_gen_addi_i64(addr, addr, offset);
447
+ }
448
+ tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
449
+ }
450
+}
451
+
452
/* Loads and stores */
453
static void disas_ldst(DisasContext *s, uint32_t insn)
454
{
455
@@ -XXX,XX +XXX,XX @@ static void disas_ldst(DisasContext *s, uint32_t insn)
456
case 0x0d: /* AdvSIMD load/store single structure */
457
disas_ldst_single_struct(s, insn);
458
break;
459
- case 0x19: /* LDAPR/STLR (unscaled immediate) */
460
- if (extract32(insn, 10, 2) != 0 ||
461
- extract32(insn, 21, 1) != 0) {
462
+ case 0x19:
463
+ if (extract32(insn, 21, 1) != 0) {
464
+ disas_ldst_tag(s, insn);
465
+ } else if (extract32(insn, 10, 2) == 0) {
466
+ disas_ldst_ldapr_stlr(s, insn);
467
+ } else {
468
unallocated_encoding(s);
469
- break;
470
}
471
- disas_ldst_ldapr_stlr(s, insn);
472
break;
473
default:
474
unallocated_encoding(s);
475
--
885
--
476
2.20.1
886
2.25.1
477
887
478
888
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We need to check the memattr of a page in order to determine
3
These particular data structures are not modified at runtime.
4
whether it is Tagged for MTE. Between Stage1 and Stage2,
5
this becomes simpler if we always collect this data, instead
6
of occasionally being presented with NULL.
7
4
8
Use the nonnull attribute to allow the compiler to check that
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
9
all pointer arguments are non-null.
10
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200626033144.790098-42-richard.henderson@linaro.org
8
Message-id: 20220501055028.646596-5-richard.henderson@linaro.org
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
---
10
---
16
target/arm/internals.h | 3 ++-
11
target/arm/helper.c | 16 ++++++++--------
17
target/arm/helper.c | 60 ++++++++++++++++++++---------------------
12
1 file changed, 8 insertions(+), 8 deletions(-)
18
target/arm/m_helper.c | 11 +++++---
19
target/arm/tlb_helper.c | 4 ++-
20
4 files changed, 42 insertions(+), 36 deletions(-)
21
13
22
diff --git a/target/arm/internals.h b/target/arm/internals.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/internals.h
25
+++ b/target/arm/internals.h
26
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
27
MMUAccessType access_type, ARMMMUIdx mmu_idx,
28
hwaddr *phys_ptr, MemTxAttrs *attrs, int *prot,
29
target_ulong *page_size,
30
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
31
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
32
+ __attribute__((nonnull));
33
34
void arm_log_exception(int idx);
35
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
37
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
39
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
40
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
18
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
41
bool s1_is_el0,
19
.resetvalue = cpu->pmceid1 },
42
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
20
};
43
target_ulong *page_size_ptr,
21
#ifdef CONFIG_USER_ONLY
44
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
22
- ARMCPRegUserSpaceInfo v8_user_idregs[] = {
45
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
23
+ static const ARMCPRegUserSpaceInfo v8_user_idregs[] = {
46
+ __attribute__((nonnull));
24
{ .name = "ID_AA64PFR0_EL1",
47
#endif
25
.exported_bits = 0x000f000f00ff0000,
48
26
.fixed_bits = 0x0000000000000011 },
49
static void switch_mode(CPUARMState *env, int mode);
27
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
50
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
28
*/
51
arm_tlb_bti_gp(txattrs) = true;
29
if (arm_feature(env, ARM_FEATURE_EL3)) {
30
if (arm_feature(env, ARM_FEATURE_AARCH64)) {
31
- ARMCPRegInfo nsacr = {
32
+ static const ARMCPRegInfo nsacr = {
33
.name = "NSACR", .type = ARM_CP_CONST,
34
.cp = 15, .opc1 = 0, .crn = 1, .crm = 1, .opc2 = 2,
35
.access = PL1_RW, .accessfn = nsacr_access,
36
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
37
};
38
define_one_arm_cp_reg(cpu, &nsacr);
39
} else {
40
- ARMCPRegInfo nsacr = {
41
+ static const ARMCPRegInfo nsacr = {
42
.name = "NSACR",
43
.cp = 15, .opc1 = 0, .crn = 1, .crm = 1, .opc2 = 2,
44
.access = PL3_RW | PL1_R,
45
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
46
}
47
} else {
48
if (arm_feature(env, ARM_FEATURE_V8)) {
49
- ARMCPRegInfo nsacr = {
50
+ static const ARMCPRegInfo nsacr = {
51
.name = "NSACR", .type = ARM_CP_CONST,
52
.cp = 15, .opc1 = 0, .crn = 1, .crm = 1, .opc2 = 2,
53
.access = PL1_R,
54
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
55
.access = PL1_R, .type = ARM_CP_CONST,
56
.resetvalue = cpu->pmsav7_dregion << 8
57
};
58
- ARMCPRegInfo crn0_wi_reginfo = {
59
+ static const ARMCPRegInfo crn0_wi_reginfo = {
60
.name = "CRN0_WI", .cp = 15, .crn = 0, .crm = CP_ANY,
61
.opc1 = CP_ANY, .opc2 = CP_ANY, .access = PL1_W,
62
.type = ARM_CP_NOP | ARM_CP_OVERRIDE
63
};
64
#ifdef CONFIG_USER_ONLY
65
- ARMCPRegUserSpaceInfo id_v8_user_midr_cp_reginfo[] = {
66
+ static const ARMCPRegUserSpaceInfo id_v8_user_midr_cp_reginfo[] = {
67
{ .name = "MIDR_EL1",
68
.exported_bits = 0x00000000ffffffff },
69
{ .name = "REVIDR_EL1" },
70
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
71
.access = PL1_R, .readfn = mpidr_read, .type = ARM_CP_NO_RAW },
72
};
73
#ifdef CONFIG_USER_ONLY
74
- ARMCPRegUserSpaceInfo mpidr_user_cp_reginfo[] = {
75
+ static const ARMCPRegUserSpaceInfo mpidr_user_cp_reginfo[] = {
76
{ .name = "MPIDR_EL1",
77
.fixed_bits = 0x0000000080000000 },
78
};
79
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
52
}
80
}
53
81
54
- if (cacheattrs != NULL) {
82
if (arm_feature(env, ARM_FEATURE_VBAR)) {
55
- if (mmu_idx == ARMMMUIdx_Stage2) {
83
- ARMCPRegInfo vbar_cp_reginfo[] = {
56
- cacheattrs->attrs = convert_stage2_attrs(env,
84
+ static const ARMCPRegInfo vbar_cp_reginfo[] = {
57
- extract32(attrs, 0, 4));
85
{ .name = "VBAR", .state = ARM_CP_STATE_BOTH,
58
- } else {
86
.opc0 = 3, .crn = 12, .crm = 0, .opc1 = 0, .opc2 = 0,
59
- /* Index into MAIR registers for cache attributes */
87
.access = PL1_RW, .writefn = vbar_write,
60
- uint8_t attrindx = extract32(attrs, 0, 3);
61
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
62
- assert(attrindx <= 7);
63
- cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
64
- }
65
- cacheattrs->shareability = extract32(attrs, 6, 2);
66
+ if (mmu_idx == ARMMMUIdx_Stage2) {
67
+ cacheattrs->attrs = convert_stage2_attrs(env, extract32(attrs, 0, 4));
68
+ } else {
69
+ /* Index into MAIR registers for cache attributes */
70
+ uint8_t attrindx = extract32(attrs, 0, 3);
71
+ uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
72
+ assert(attrindx <= 7);
73
+ cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
74
}
75
+ cacheattrs->shareability = extract32(attrs, 6, 2);
76
77
*phys_ptr = descaddr;
78
*page_size_ptr = page_size;
79
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
80
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
81
mmu_idx == ARMMMUIdx_E10_0,
82
phys_ptr, attrs, &s2_prot,
83
- page_size, fi,
84
- cacheattrs != NULL ? &cacheattrs2 : NULL);
85
+ page_size, fi, &cacheattrs2);
86
fi->s2addr = ipa;
87
/* Combine the S1 and S2 perms. */
88
*prot &= s2_prot;
89
90
- /* Combine the S1 and S2 cache attributes, if needed */
91
- if (!ret && cacheattrs != NULL) {
92
- if (env->cp15.hcr_el2 & HCR_DC) {
93
- /*
94
- * HCR.DC forces the first stage attributes to
95
- * Normal Non-Shareable,
96
- * Inner Write-Back Read-Allocate Write-Allocate,
97
- * Outer Write-Back Read-Allocate Write-Allocate.
98
- */
99
- cacheattrs->attrs = 0xff;
100
- cacheattrs->shareability = 0;
101
- }
102
- *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
103
+ /* If S2 fails, return early. */
104
+ if (ret) {
105
+ return ret;
106
}
107
108
- return ret;
109
+ /* Combine the S1 and S2 cache attributes. */
110
+ if (env->cp15.hcr_el2 & HCR_DC) {
111
+ /*
112
+ * HCR.DC forces the first stage attributes to
113
+ * Normal Non-Shareable,
114
+ * Inner Write-Back Read-Allocate Write-Allocate,
115
+ * Outer Write-Back Read-Allocate Write-Allocate.
116
+ */
117
+ cacheattrs->attrs = 0xff;
118
+ cacheattrs->shareability = 0;
119
+ }
120
+ *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
121
+ return 0;
122
} else {
123
/*
124
* For non-EL2 CPUs a stage1+stage2 translation is just stage 1.
125
@@ -XXX,XX +XXX,XX @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
126
bool ret;
127
ARMMMUFaultInfo fi = {};
128
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
129
+ ARMCacheAttrs cacheattrs = {};
130
131
*attrs = (MemTxAttrs) {};
132
133
ret = get_phys_addr(env, addr, 0, mmu_idx, &phys_addr,
134
- attrs, &prot, &page_size, &fi, NULL);
135
+ attrs, &prot, &page_size, &fi, &cacheattrs);
136
137
if (ret) {
138
return -1;
139
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
140
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/m_helper.c
142
+++ b/target/arm/m_helper.c
143
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
144
hwaddr physaddr;
145
int prot;
146
ARMMMUFaultInfo fi = {};
147
+ ARMCacheAttrs cacheattrs = {};
148
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
149
int exc;
150
bool exc_secure;
151
152
if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &physaddr,
153
- &attrs, &prot, &page_size, &fi, NULL)) {
154
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
155
/* MPU/SAU lookup failed */
156
if (fi.type == ARMFault_QEMU_SFault) {
157
if (mode == STACK_LAZYFP) {
158
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
159
hwaddr physaddr;
160
int prot;
161
ARMMMUFaultInfo fi = {};
162
+ ARMCacheAttrs cacheattrs = {};
163
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
164
int exc;
165
bool exc_secure;
166
uint32_t value;
167
168
if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
169
- &attrs, &prot, &page_size, &fi, NULL)) {
170
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
171
/* MPU/SAU lookup failed */
172
if (fi.type == ARMFault_QEMU_SFault) {
173
qemu_log_mask(CPU_LOG_INT,
174
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
175
V8M_SAttributes sattrs = {};
176
MemTxAttrs attrs = {};
177
ARMMMUFaultInfo fi = {};
178
+ ARMCacheAttrs cacheattrs = {};
179
MemTxResult txres;
180
target_ulong page_size;
181
hwaddr physaddr;
182
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
183
"...really SecureFault with SFSR.INVEP\n");
184
return false;
185
}
186
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx,
187
- &physaddr, &attrs, &prot, &page_size, &fi, NULL)) {
188
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &physaddr,
189
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
190
/* the MPU lookup failed */
191
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
192
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
193
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
194
index XXXXXXX..XXXXXXX 100644
195
--- a/target/arm/tlb_helper.c
196
+++ b/target/arm/tlb_helper.c
197
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
198
int prot, ret;
199
MemTxAttrs attrs = {};
200
ARMMMUFaultInfo fi = {};
201
+ ARMCacheAttrs cacheattrs = {};
202
203
/*
204
* Walk the page table and (if the mapping exists) add the page
205
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
206
*/
207
ret = get_phys_addr(&cpu->env, address, access_type,
208
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
209
- &phys_addr, &attrs, &prot, &page_size, &fi, NULL);
210
+ &phys_addr, &attrs, &prot, &page_size,
211
+ &fi, &cacheattrs);
212
if (likely(!ret)) {
213
/*
214
* Map a single [sub]page. Regions smaller than our declared
215
--
88
--
216
2.20.1
89
2.25.1
217
90
218
91
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Add an option that writes back the PC, like DISAS_UPDATE_EXIT,
3
Instead of defining ARM_CP_FLAG_MASK to remove flags,
4
but does not exit back to the main loop.
4
define ARM_CP_SPECIAL_MASK to isolate special cases.
5
5
Sort the specials to the low bits. Use an enum.
6
7
Split the large comment block so as to document each
8
value separately.
9
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20220501055028.646596-6-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-7-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
14
---
11
target/arm/translate.h | 2 ++
15
target/arm/cpregs.h | 130 +++++++++++++++++++++++--------------
12
target/arm/translate-a64.c | 3 +++
16
target/arm/cpu.c | 4 +-
13
target/arm/translate.c | 4 ++++
17
target/arm/helper.c | 4 +-
14
3 files changed, 9 insertions(+)
18
target/arm/translate-a64.c | 6 +-
15
19
target/arm/translate.c | 6 +-
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
20
5 files changed, 92 insertions(+), 58 deletions(-)
17
index XXXXXXX..XXXXXXX 100644
21
18
--- a/target/arm/translate.h
22
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
19
+++ b/target/arm/translate.h
23
index XXXXXXX..XXXXXXX 100644
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
24
--- a/target/arm/cpregs.h
21
* return from cpu_tb_exec.
25
+++ b/target/arm/cpregs.h
26
@@ -XXX,XX +XXX,XX @@
27
#define TARGET_ARM_CPREGS_H
28
29
/*
30
- * ARMCPRegInfo type field bits. If the SPECIAL bit is set this is a
31
- * special-behaviour cp reg and bits [11..8] indicate what behaviour
32
- * it has. Otherwise it is a simple cp reg, where CONST indicates that
33
- * TCG can assume the value to be constant (ie load at translate time)
34
- * and 64BIT indicates a 64 bit wide coprocessor register. SUPPRESS_TB_END
35
- * indicates that the TB should not be ended after a write to this register
36
- * (the default is that the TB ends after cp writes). OVERRIDE permits
37
- * a register definition to override a previous definition for the
38
- * same (cp, is64, crn, crm, opc1, opc2) tuple: either the new or the
39
- * old must have the OVERRIDE bit set.
40
- * ALIAS indicates that this register is an alias view of some underlying
41
- * state which is also visible via another register, and that the other
42
- * register is handling migration and reset; registers marked ALIAS will not be
43
- * migrated but may have their state set by syncing of register state from KVM.
44
- * NO_RAW indicates that this register has no underlying state and does not
45
- * support raw access for state saving/loading; it will not be used for either
46
- * migration or KVM state synchronization. (Typically this is for "registers"
47
- * which are actually used as instructions for cache maintenance and so on.)
48
- * IO indicates that this register does I/O and therefore its accesses
49
- * need to be marked with gen_io_start() and also end the TB. In particular,
50
- * registers which implement clocks or timers require this.
51
- * RAISES_EXC is for when the read or write hook might raise an exception;
52
- * the generated code will synchronize the CPU state before calling the hook
53
- * so that it is safe for the hook to call raise_exception().
54
- * NEWEL is for writes to registers that might change the exception
55
- * level - typically on older ARM chips. For those cases we need to
56
- * re-read the new el when recomputing the translation flags.
57
+ * ARMCPRegInfo type field bits:
22
*/
58
*/
23
#define DISAS_EXIT DISAS_TARGET_9
59
-#define ARM_CP_SPECIAL 0x0001
24
+/* CPU state was modified dynamically; no need to exit, but do not chain. */
60
-#define ARM_CP_CONST 0x0002
25
+#define DISAS_UPDATE_NOCHAIN DISAS_TARGET_10
61
-#define ARM_CP_64BIT 0x0004
26
62
-#define ARM_CP_SUPPRESS_TB_END 0x0008
27
#ifdef TARGET_AARCH64
63
-#define ARM_CP_OVERRIDE 0x0010
28
void a64_translate_init(void);
64
-#define ARM_CP_ALIAS 0x0020
65
-#define ARM_CP_IO 0x0040
66
-#define ARM_CP_NO_RAW 0x0080
67
-#define ARM_CP_NOP (ARM_CP_SPECIAL | 0x0100)
68
-#define ARM_CP_WFI (ARM_CP_SPECIAL | 0x0200)
69
-#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
70
-#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
71
-#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
72
-#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
73
-#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
74
-#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
75
-#define ARM_CP_FPU 0x1000
76
-#define ARM_CP_SVE 0x2000
77
-#define ARM_CP_NO_GDB 0x4000
78
-#define ARM_CP_RAISES_EXC 0x8000
79
-#define ARM_CP_NEWEL 0x10000
80
-/* Mask of only the flag bits in a type field */
81
-#define ARM_CP_FLAG_MASK 0x1f0ff
82
+enum {
83
+ /*
84
+ * Register must be handled specially during translation.
85
+ * The method is one of the values below:
86
+ */
87
+ ARM_CP_SPECIAL_MASK = 0x000f,
88
+ /* Special: no change to PE state: writes ignored, reads ignored. */
89
+ ARM_CP_NOP = 0x0001,
90
+ /* Special: sysreg is WFI, for v5 and v6. */
91
+ ARM_CP_WFI = 0x0002,
92
+ /* Special: sysreg is NZCV. */
93
+ ARM_CP_NZCV = 0x0003,
94
+ /* Special: sysreg is CURRENTEL. */
95
+ ARM_CP_CURRENTEL = 0x0004,
96
+ /* Special: sysreg is DC ZVA or similar. */
97
+ ARM_CP_DC_ZVA = 0x0005,
98
+ ARM_CP_DC_GVA = 0x0006,
99
+ ARM_CP_DC_GZVA = 0x0007,
100
+
101
+ /* Flag: reads produce resetvalue; writes ignored. */
102
+ ARM_CP_CONST = 1 << 4,
103
+ /* Flag: For ARM_CP_STATE_AA32, sysreg is 64-bit. */
104
+ ARM_CP_64BIT = 1 << 5,
105
+ /*
106
+ * Flag: TB should not be ended after a write to this register
107
+ * (the default is that the TB ends after cp writes).
108
+ */
109
+ ARM_CP_SUPPRESS_TB_END = 1 << 6,
110
+ /*
111
+ * Flag: Permit a register definition to override a previous definition
112
+ * for the same (cp, is64, crn, crm, opc1, opc2) tuple: either the new
113
+ * or the old must have the ARM_CP_OVERRIDE bit set.
114
+ */
115
+ ARM_CP_OVERRIDE = 1 << 7,
116
+ /*
117
+ * Flag: Register is an alias view of some underlying state which is also
118
+ * visible via another register, and that the other register is handling
119
+ * migration and reset; registers marked ARM_CP_ALIAS will not be migrated
120
+ * but may have their state set by syncing of register state from KVM.
121
+ */
122
+ ARM_CP_ALIAS = 1 << 8,
123
+ /*
124
+ * Flag: Register does I/O and therefore its accesses need to be marked
125
+ * with gen_io_start() and also end the TB. In particular, registers which
126
+ * implement clocks or timers require this.
127
+ */
128
+ ARM_CP_IO = 1 << 9,
129
+ /*
130
+ * Flag: Register has no underlying state and does not support raw access
131
+ * for state saving/loading; it will not be used for either migration or
132
+ * KVM state synchronization. Typically this is for "registers" which are
133
+ * actually used as instructions for cache maintenance and so on.
134
+ */
135
+ ARM_CP_NO_RAW = 1 << 10,
136
+ /*
137
+ * Flag: The read or write hook might raise an exception; the generated
138
+ * code will synchronize the CPU state before calling the hook so that it
139
+ * is safe for the hook to call raise_exception().
140
+ */
141
+ ARM_CP_RAISES_EXC = 1 << 11,
142
+ /*
143
+ * Flag: Writes to the sysreg might change the exception level - typically
144
+ * on older ARM chips. For those cases we need to re-read the new el when
145
+ * recomputing the translation flags.
146
+ */
147
+ ARM_CP_NEWEL = 1 << 12,
148
+ /*
149
+ * Flag: Access check for this sysreg is identical to accessing FPU state
150
+ * from an instruction: use translation fp_access_check().
151
+ */
152
+ ARM_CP_FPU = 1 << 13,
153
+ /*
154
+ * Flag: Access check for this sysreg is identical to accessing SVE state
155
+ * from an instruction: use translation sve_access_check().
156
+ */
157
+ ARM_CP_SVE = 1 << 14,
158
+ /* Flag: Do not expose in gdb sysreg xml. */
159
+ ARM_CP_NO_GDB = 1 << 15,
160
+};
161
162
/*
163
* Valid values for ARMCPRegInfo state field, indicating which of
164
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
165
index XXXXXXX..XXXXXXX 100644
166
--- a/target/arm/cpu.c
167
+++ b/target/arm/cpu.c
168
@@ -XXX,XX +XXX,XX @@ static void cp_reg_reset(gpointer key, gpointer value, gpointer opaque)
169
ARMCPRegInfo *ri = value;
170
ARMCPU *cpu = opaque;
171
172
- if (ri->type & (ARM_CP_SPECIAL | ARM_CP_ALIAS)) {
173
+ if (ri->type & (ARM_CP_SPECIAL_MASK | ARM_CP_ALIAS)) {
174
return;
175
}
176
177
@@ -XXX,XX +XXX,XX @@ static void cp_reg_check_reset(gpointer key, gpointer value, gpointer opaque)
178
ARMCPU *cpu = opaque;
179
uint64_t oldvalue, newvalue;
180
181
- if (ri->type & (ARM_CP_SPECIAL | ARM_CP_ALIAS | ARM_CP_NO_RAW)) {
182
+ if (ri->type & (ARM_CP_SPECIAL_MASK | ARM_CP_ALIAS | ARM_CP_NO_RAW)) {
183
return;
184
}
185
186
diff --git a/target/arm/helper.c b/target/arm/helper.c
187
index XXXXXXX..XXXXXXX 100644
188
--- a/target/arm/helper.c
189
+++ b/target/arm/helper.c
190
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
191
* multiple times. Special registers (ie NOP/WFI) are
192
* never migratable and not even raw-accessible.
193
*/
194
- if ((r->type & ARM_CP_SPECIAL)) {
195
+ if (r->type & ARM_CP_SPECIAL_MASK) {
196
r2->type |= ARM_CP_NO_RAW;
197
}
198
if (((r->crm == CP_ANY) && crm != 0) ||
199
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
200
/* Check that the register definition has enough info to handle
201
* reads and writes if they are permitted.
202
*/
203
- if (!(r->type & (ARM_CP_SPECIAL|ARM_CP_CONST))) {
204
+ if (!(r->type & (ARM_CP_SPECIAL_MASK | ARM_CP_CONST))) {
205
if (r->access & PL3_R) {
206
assert((r->fieldoffset ||
207
(r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1])) ||
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
208
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
30
index XXXXXXX..XXXXXXX 100644
209
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/translate-a64.c
210
--- a/target/arm/translate-a64.c
32
+++ b/target/arm/translate-a64.c
211
+++ b/target/arm/translate-a64.c
33
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
212
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
34
case DISAS_EXIT:
213
}
35
tcg_gen_exit_tb(NULL, 0);
214
36
break;
215
/* Handle special cases first */
37
+ case DISAS_UPDATE_NOCHAIN:
216
- switch (ri->type & ~(ARM_CP_FLAG_MASK & ~ARM_CP_SPECIAL)) {
38
+ gen_a64_set_pc_im(dc->base.pc_next);
217
+ switch (ri->type & ARM_CP_SPECIAL_MASK) {
39
+ /* fall through */
218
+ case 0:
40
case DISAS_JUMP:
219
+ break;
41
tcg_gen_lookup_and_goto_ptr();
220
case ARM_CP_NOP:
42
break;
221
return;
222
case ARM_CP_NZCV:
223
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
224
}
225
return;
226
default:
227
- break;
228
+ g_assert_not_reached();
229
}
230
if ((ri->type & ARM_CP_FPU) && !fp_access_check(s)) {
231
return;
43
diff --git a/target/arm/translate.c b/target/arm/translate.c
232
diff --git a/target/arm/translate.c b/target/arm/translate.c
44
index XXXXXXX..XXXXXXX 100644
233
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/translate.c
234
--- a/target/arm/translate.c
46
+++ b/target/arm/translate.c
235
+++ b/target/arm/translate.c
47
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
236
@@ -XXX,XX +XXX,XX @@ static void do_coproc_insn(DisasContext *s, int cpnum, int is64,
48
case DISAS_NEXT:
237
}
49
case DISAS_TOO_MANY:
238
50
case DISAS_UPDATE_EXIT:
239
/* Handle special cases first */
51
+ case DISAS_UPDATE_NOCHAIN:
240
- switch (ri->type & ~(ARM_CP_FLAG_MASK & ~ARM_CP_SPECIAL)) {
52
gen_set_pc_im(dc, dc->base.pc_next);
241
+ switch (ri->type & ARM_CP_SPECIAL_MASK) {
53
/* fall through */
242
+ case 0:
243
+ break;
244
case ARM_CP_NOP:
245
return;
246
case ARM_CP_WFI:
247
@@ -XXX,XX +XXX,XX @@ static void do_coproc_insn(DisasContext *s, int cpnum, int is64,
248
s->base.is_jmp = DISAS_WFI;
249
return;
54
default:
250
default:
55
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
251
- break;
56
case DISAS_TOO_MANY:
252
+ g_assert_not_reached();
57
gen_goto_tb(dc, 1, dc->base.pc_next);
253
}
58
break;
254
59
+ case DISAS_UPDATE_NOCHAIN:
255
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
60
+ gen_set_pc_im(dc, dc->base.pc_next);
61
+ /* fall through */
62
case DISAS_JUMP:
63
gen_goto_ptr();
64
break;
65
--
256
--
66
2.20.1
257
2.25.1
67
68
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Emphasize that the is_jmp option exits to the main loop.
3
Standardize on g_assert_not_reached() for "should not happen".
4
Retain abort() when preceeded by fprintf or error_report.
4
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220501055028.646596-7-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-6-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
---
10
target/arm/translate.h | 14 ++++++++------
11
target/arm/helper.c | 7 +++----
11
target/arm/translate-a64.c | 8 ++++----
12
target/arm/hvf/hvf.c | 2 +-
12
target/arm/translate-vfp.inc.c | 4 ++--
13
target/arm/kvm-stub.c | 4 ++--
13
target/arm/translate.c | 12 ++++++------
14
target/arm/kvm.c | 4 ++--
14
4 files changed, 20 insertions(+), 18 deletions(-)
15
target/arm/machine.c | 4 ++--
16
target/arm/translate-a64.c | 4 ++--
17
target/arm/translate-neon.c | 2 +-
18
target/arm/translate.c | 4 ++--
19
8 files changed, 15 insertions(+), 16 deletions(-)
15
20
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
21
diff --git a/target/arm/helper.c b/target/arm/helper.c
17
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.h
23
--- a/target/arm/helper.c
19
+++ b/target/arm/translate.h
24
+++ b/target/arm/helper.c
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
25
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
21
26
break;
22
/* is_jmp field values */
27
default:
23
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
28
/* broken reginfo with out-of-range opc1 */
24
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
29
- assert(false);
25
+/* CPU state was modified dynamically; exit to main loop for interrupts. */
30
- break;
26
+#define DISAS_UPDATE_EXIT DISAS_TARGET_1
31
+ g_assert_not_reached();
27
/* These instructions trap after executing, so the A32/T32 decoder must
32
}
28
* defer them until after the conditional execution state has been updated.
33
/* assert our permissions are not too lax (stricter is fine) */
29
* WFI also needs special handling when single-stepping.
34
assert((r->access & ~mask) == 0);
30
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
35
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_v5(CPUARMState *env, uint32_t address,
31
* custom end-of-TB code)
36
break;
32
*/
37
default:
33
#define DISAS_BX_EXCRET DISAS_TARGET_8
38
/* Never happens, but compiler isn't smart enough to tell. */
34
-/* For instructions which want an immediate exit to the main loop,
39
- abort();
35
- * as opposed to attempting to use lookup_and_goto_ptr. Unlike
40
+ g_assert_not_reached();
36
- * DISAS_UPDATE this doesn't write the PC on exiting the translation
41
}
37
- * loop so you need to ensure something (gen_a64_set_pc_im or runtime
42
}
38
- * helper) has done so before we reach return from cpu_tb_exec.
43
*prot = ap_to_rw_prot(env, mmu_idx, ap, domain_prot);
39
+/*
44
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_v6(CPUARMState *env, uint32_t address,
40
+ * For instructions which want an immediate exit to the main loop, as opposed
45
break;
41
+ * to attempting to use lookup_and_goto_ptr. Unlike DISAS_UPDATE_EXIT, this
46
default:
42
+ * doesn't write the PC on exiting the translation loop so you need to ensure
47
/* Never happens, but compiler isn't smart enough to tell. */
43
+ * something (gen_a64_set_pc_im or runtime helper) has done so before we reach
48
- abort();
44
+ * return from cpu_tb_exec.
49
+ g_assert_not_reached();
45
*/
50
}
46
#define DISAS_EXIT DISAS_TARGET_9
51
}
52
if (domain_prot == 3) {
53
diff --git a/target/arm/hvf/hvf.c b/target/arm/hvf/hvf.c
54
index XXXXXXX..XXXXXXX 100644
55
--- a/target/arm/hvf/hvf.c
56
+++ b/target/arm/hvf/hvf.c
57
@@ -XXX,XX +XXX,XX @@ int hvf_vcpu_exec(CPUState *cpu)
58
/* we got kicked, no exit to process */
59
return 0;
60
default:
61
- assert(0);
62
+ g_assert_not_reached();
63
}
64
65
hvf_sync_vtimer(cpu);
66
diff --git a/target/arm/kvm-stub.c b/target/arm/kvm-stub.c
67
index XXXXXXX..XXXXXXX 100644
68
--- a/target/arm/kvm-stub.c
69
+++ b/target/arm/kvm-stub.c
70
@@ -XXX,XX +XXX,XX @@
71
72
bool write_kvmstate_to_list(ARMCPU *cpu)
73
{
74
- abort();
75
+ g_assert_not_reached();
76
}
77
78
bool write_list_to_kvmstate(ARMCPU *cpu, int level)
79
{
80
- abort();
81
+ g_assert_not_reached();
82
}
83
diff --git a/target/arm/kvm.c b/target/arm/kvm.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/target/arm/kvm.c
86
+++ b/target/arm/kvm.c
87
@@ -XXX,XX +XXX,XX @@ bool write_kvmstate_to_list(ARMCPU *cpu)
88
ret = kvm_vcpu_ioctl(cs, KVM_GET_ONE_REG, &r);
89
break;
90
default:
91
- abort();
92
+ g_assert_not_reached();
93
}
94
if (ret) {
95
ok = false;
96
@@ -XXX,XX +XXX,XX @@ bool write_list_to_kvmstate(ARMCPU *cpu, int level)
97
r.addr = (uintptr_t)(cpu->cpreg_values + i);
98
break;
99
default:
100
- abort();
101
+ g_assert_not_reached();
102
}
103
ret = kvm_vcpu_ioctl(cs, KVM_SET_ONE_REG, &r);
104
if (ret) {
105
diff --git a/target/arm/machine.c b/target/arm/machine.c
106
index XXXXXXX..XXXXXXX 100644
107
--- a/target/arm/machine.c
108
+++ b/target/arm/machine.c
109
@@ -XXX,XX +XXX,XX @@ static int cpu_pre_save(void *opaque)
110
if (kvm_enabled()) {
111
if (!write_kvmstate_to_list(cpu)) {
112
/* This should never fail */
113
- abort();
114
+ g_assert_not_reached();
115
}
116
117
/*
118
@@ -XXX,XX +XXX,XX @@ static int cpu_pre_save(void *opaque)
119
} else {
120
if (!write_cpustate_to_list(cpu, false)) {
121
/* This should never fail. */
122
- abort();
123
+ g_assert_not_reached();
124
}
125
}
47
126
48
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
127
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
49
index XXXXXXX..XXXXXXX 100644
128
index XXXXXXX..XXXXXXX 100644
50
--- a/target/arm/translate-a64.c
129
--- a/target/arm/translate-a64.c
51
+++ b/target/arm/translate-a64.c
130
+++ b/target/arm/translate-a64.c
52
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
131
@@ -XXX,XX +XXX,XX @@ static void handle_fp_1src_half(DisasContext *s, int opcode, int rd, int rn)
53
gen_helper_msr_i_daifclear(cpu_env, t1);
132
gen_helper_advsimd_rinth(tcg_res, tcg_op, fpst);
54
tcg_temp_free_i32(t1);
55
/* For DAIFClear, exit the cpu loop to re-evaluate pending IRQs. */
56
- s->base.is_jmp = DISAS_UPDATE;
57
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
58
break;
133
break;
59
60
default:
134
default:
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
135
- abort();
62
136
+ g_assert_not_reached();
63
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
64
/* I/O operations must end the TB here (whether read or write) */
65
- s->base.is_jmp = DISAS_UPDATE;
66
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
67
}
137
}
68
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
138
69
/*
139
write_fp_sreg(s, rd, tcg_res);
70
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
140
@@ -XXX,XX +XXX,XX @@ static void handle_fp_fcvt(DisasContext *s, int opcode,
71
* but allow this to be suppressed by the register definition
141
break;
72
* (usually only necessary to work around guest bugs).
142
}
73
*/
143
default:
74
- s->base.is_jmp = DISAS_UPDATE;
144
- abort();
75
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
145
+ g_assert_not_reached();
76
}
146
}
77
}
147
}
78
148
79
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
149
diff --git a/target/arm/translate-neon.c b/target/arm/translate-neon.c
80
gen_goto_tb(dc, 1, dc->base.pc_next);
81
break;
82
default:
83
- case DISAS_UPDATE:
84
+ case DISAS_UPDATE_EXIT:
85
gen_a64_set_pc_im(dc->base.pc_next);
86
/* fall through */
87
case DISAS_EXIT:
88
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
89
index XXXXXXX..XXXXXXX 100644
150
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate-vfp.inc.c
151
--- a/target/arm/translate-neon.c
91
+++ b/target/arm/translate-vfp.inc.c
152
+++ b/target/arm/translate-neon.c
92
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
153
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDST_single(DisasContext *s, arg_VLDST_single *a)
93
* this to be the last insn in the TB).
154
}
94
*/
155
break;
95
if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
156
default:
96
- s->base.is_jmp = DISAS_UPDATE;
157
- abort();
97
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
158
+ g_assert_not_reached();
98
gen_io_start();
159
}
99
}
160
if ((vd + a->stride * (nregs - 1)) > 31) {
100
gen_helper_v7m_preserve_fp_state(cpu_env);
161
/*
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
102
tcg_temp_free_i32(fptr);
103
104
/* End the TB, because we have updated FP control bits */
105
- s->base.is_jmp = DISAS_UPDATE;
106
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
107
return true;
108
}
109
diff --git a/target/arm/translate.c b/target/arm/translate.c
162
diff --git a/target/arm/translate.c b/target/arm/translate.c
110
index XXXXXXX..XXXXXXX 100644
163
index XXXXXXX..XXXXXXX 100644
111
--- a/target/arm/translate.c
164
--- a/target/arm/translate.c
112
+++ b/target/arm/translate.c
165
+++ b/target/arm/translate.c
113
@@ -XXX,XX +XXX,XX @@ static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
114
tcg_temp_free_i32(tcg_tgtmode);
115
tcg_temp_free_i32(tcg_regno);
116
tcg_temp_free_i32(tcg_reg);
117
- s->base.is_jmp = DISAS_UPDATE;
118
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
119
}
120
121
static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
122
@@ -XXX,XX +XXX,XX @@ static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
123
tcg_temp_free_i32(tcg_tgtmode);
124
tcg_temp_free_i32(tcg_regno);
125
store_reg(s, rn, tcg_reg);
126
- s->base.is_jmp = DISAS_UPDATE;
127
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
128
}
129
130
/* Store value to PC as for an exception return (ie don't
131
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
166
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
132
tcg_temp_free_i32(tmp);
167
offset = 4;
168
break;
169
default:
170
- abort();
171
+ g_assert_not_reached();
133
}
172
}
134
tcg_temp_free_i32(addr);
173
tcg_gen_addi_i32(addr, addr, offset);
135
- s->base.is_jmp = DISAS_UPDATE;
174
tmp = load_reg(s, 14);
136
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
175
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
137
}
176
offset = 0;
138
139
/* Generate a label used for skipping this instruction */
140
@@ -XXX,XX +XXX,XX @@ static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
141
}
142
if (a->E != (s->be_data == MO_BE)) {
143
gen_helper_setend(cpu_env);
144
- s->base.is_jmp = DISAS_UPDATE;
145
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
146
}
147
return true;
148
}
149
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
150
break;
177
break;
151
case DISAS_NEXT:
152
case DISAS_TOO_MANY:
153
- case DISAS_UPDATE:
154
+ case DISAS_UPDATE_EXIT:
155
gen_set_pc_im(dc, dc->base.pc_next);
156
/* fall through */
157
default:
178
default:
158
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
179
- abort();
159
case DISAS_JUMP:
180
+ g_assert_not_reached();
160
gen_goto_ptr();
181
}
161
break;
182
tcg_gen_addi_i32(addr, addr, offset);
162
- case DISAS_UPDATE:
183
gen_helper_set_r13_banked(cpu_env, tcg_constant_i32(mode), addr);
163
+ case DISAS_UPDATE_EXIT:
164
gen_set_pc_im(dc, dc->base.pc_next);
165
/* fall through */
166
default:
167
--
184
--
168
2.20.1
185
2.25.1
169
170
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We will shortly need this in mte_helper.c as well.
3
Create a typedef as well, and use it in ARMCPRegInfo.
4
This won't be perfect for debugging, but it'll nicely
5
display the most common cases.
4
6
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-22-richard.henderson@linaro.org
9
Message-id: 20220501055028.646596-8-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/internals.h | 36 ++++++++++++++++++++++++++++++++++++
12
target/arm/cpregs.h | 44 +++++++++++++++++++++++---------------------
11
target/arm/helper.c | 36 ------------------------------------
13
target/arm/helper.c | 2 +-
12
2 files changed, 36 insertions(+), 36 deletions(-)
14
2 files changed, 24 insertions(+), 22 deletions(-)
13
15
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
16
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
18
--- a/target/arm/cpregs.h
17
+++ b/target/arm/internals.h
19
+++ b/target/arm/cpregs.h
18
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_pan(CPUARMState *env, ARMMMUIdx mmu_idx)
20
@@ -XXX,XX +XXX,XX @@ enum {
19
}
21
* described with these bits, then use a laxer set of restrictions, and
20
}
22
* do the more restrictive/complex check inside a helper function.
21
22
+/* Return the exception level which controls this address translation regime */
23
+static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
24
+{
25
+ switch (mmu_idx) {
26
+ case ARMMMUIdx_E20_0:
27
+ case ARMMMUIdx_E20_2:
28
+ case ARMMMUIdx_E20_2_PAN:
29
+ case ARMMMUIdx_Stage2:
30
+ case ARMMMUIdx_E2:
31
+ return 2;
32
+ case ARMMMUIdx_SE3:
33
+ return 3;
34
+ case ARMMMUIdx_SE10_0:
35
+ return arm_el_is_aa64(env, 3) ? 1 : 3;
36
+ case ARMMMUIdx_SE10_1:
37
+ case ARMMMUIdx_SE10_1_PAN:
38
+ case ARMMMUIdx_Stage1_E0:
39
+ case ARMMMUIdx_Stage1_E1:
40
+ case ARMMMUIdx_Stage1_E1_PAN:
41
+ case ARMMMUIdx_E10_0:
42
+ case ARMMMUIdx_E10_1:
43
+ case ARMMMUIdx_E10_1_PAN:
44
+ case ARMMMUIdx_MPrivNegPri:
45
+ case ARMMMUIdx_MUserNegPri:
46
+ case ARMMMUIdx_MPriv:
47
+ case ARMMMUIdx_MUser:
48
+ case ARMMMUIdx_MSPrivNegPri:
49
+ case ARMMMUIdx_MSUserNegPri:
50
+ case ARMMMUIdx_MSPriv:
51
+ case ARMMMUIdx_MSUser:
52
+ return 1;
53
+ default:
54
+ g_assert_not_reached();
55
+ }
56
+}
57
+
58
/* Return the FSR value for a debug exception (watchpoint, hardware
59
* breakpoint or BKPT insn) targeting the specified exception level.
60
*/
23
*/
24
-#define PL3_R 0x80
25
-#define PL3_W 0x40
26
-#define PL2_R (0x20 | PL3_R)
27
-#define PL2_W (0x10 | PL3_W)
28
-#define PL1_R (0x08 | PL2_R)
29
-#define PL1_W (0x04 | PL2_W)
30
-#define PL0_R (0x02 | PL1_R)
31
-#define PL0_W (0x01 | PL1_W)
32
+typedef enum {
33
+ PL3_R = 0x80,
34
+ PL3_W = 0x40,
35
+ PL2_R = 0x20 | PL3_R,
36
+ PL2_W = 0x10 | PL3_W,
37
+ PL1_R = 0x08 | PL2_R,
38
+ PL1_W = 0x04 | PL2_W,
39
+ PL0_R = 0x02 | PL1_R,
40
+ PL0_W = 0x01 | PL1_W,
41
42
-/*
43
- * For user-mode some registers are accessible to EL0 via a kernel
44
- * trap-and-emulate ABI. In this case we define the read permissions
45
- * as actually being PL0_R. However some bits of any given register
46
- * may still be masked.
47
- */
48
+ /*
49
+ * For user-mode some registers are accessible to EL0 via a kernel
50
+ * trap-and-emulate ABI. In this case we define the read permissions
51
+ * as actually being PL0_R. However some bits of any given register
52
+ * may still be masked.
53
+ */
54
#ifdef CONFIG_USER_ONLY
55
-#define PL0U_R PL0_R
56
+ PL0U_R = PL0_R,
57
#else
58
-#define PL0U_R PL1_R
59
+ PL0U_R = PL1_R,
60
#endif
61
62
-#define PL3_RW (PL3_R | PL3_W)
63
-#define PL2_RW (PL2_R | PL2_W)
64
-#define PL1_RW (PL1_R | PL1_W)
65
-#define PL0_RW (PL0_R | PL0_W)
66
+ PL3_RW = PL3_R | PL3_W,
67
+ PL2_RW = PL2_R | PL2_W,
68
+ PL1_RW = PL1_R | PL1_W,
69
+ PL0_RW = PL0_R | PL0_W,
70
+} CPAccessRights;
71
72
typedef enum CPAccessResult {
73
/* Access is permitted */
74
@@ -XXX,XX +XXX,XX @@ struct ARMCPRegInfo {
75
/* Register type: ARM_CP_* bits/values */
76
int type;
77
/* Access rights: PL*_[RW] */
78
- int access;
79
+ CPAccessRights access;
80
/* Security state: ARM_CP_SECSTATE_* bits/values */
81
int secure;
82
/*
61
diff --git a/target/arm/helper.c b/target/arm/helper.c
83
diff --git a/target/arm/helper.c b/target/arm/helper.c
62
index XXXXXXX..XXXXXXX 100644
84
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/helper.c
85
--- a/target/arm/helper.c
64
+++ b/target/arm/helper.c
86
+++ b/target/arm/helper.c
65
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
87
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
66
}
88
* to encompass the generic architectural permission check.
67
#endif /* !CONFIG_USER_ONLY */
89
*/
68
90
if (r->state != ARM_CP_STATE_AA32) {
69
-/* Return the exception level which controls this address translation regime */
91
- int mask = 0;
70
-static uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
92
+ CPAccessRights mask;
71
-{
93
switch (r->opc1) {
72
- switch (mmu_idx) {
94
case 0:
73
- case ARMMMUIdx_E20_0:
95
/* min_EL EL1, but some accessible to EL0 via kernel ABI */
74
- case ARMMMUIdx_E20_2:
75
- case ARMMMUIdx_E20_2_PAN:
76
- case ARMMMUIdx_Stage2:
77
- case ARMMMUIdx_E2:
78
- return 2;
79
- case ARMMMUIdx_SE3:
80
- return 3;
81
- case ARMMMUIdx_SE10_0:
82
- return arm_el_is_aa64(env, 3) ? 1 : 3;
83
- case ARMMMUIdx_SE10_1:
84
- case ARMMMUIdx_SE10_1_PAN:
85
- case ARMMMUIdx_Stage1_E0:
86
- case ARMMMUIdx_Stage1_E1:
87
- case ARMMMUIdx_Stage1_E1_PAN:
88
- case ARMMMUIdx_E10_0:
89
- case ARMMMUIdx_E10_1:
90
- case ARMMMUIdx_E10_1_PAN:
91
- case ARMMMUIdx_MPrivNegPri:
92
- case ARMMMUIdx_MUserNegPri:
93
- case ARMMMUIdx_MPriv:
94
- case ARMMMUIdx_MUser:
95
- case ARMMMUIdx_MSPrivNegPri:
96
- case ARMMMUIdx_MSUserNegPri:
97
- case ARMMMUIdx_MSPriv:
98
- case ARMMMUIdx_MSUser:
99
- return 1;
100
- default:
101
- g_assert_not_reached();
102
- }
103
-}
104
-
105
uint64_t arm_sctlr(CPUARMState *env, int el)
106
{
107
/* Only EL0 needs to be adjusted for EL1&0 or EL2&0. */
108
--
96
--
109
2.20.1
97
2.25.1
110
111
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We will shortly need this in mte_helper.c as well.
3
Give this enum a name and use in ARMCPRegInfo,
4
add_cpreg_to_hashtable and define_one_arm_cp_reg_with_opaque.
4
5
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-23-richard.henderson@linaro.org
9
Message-id: 20220501055028.646596-9-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/internals.h | 9 +++++++++
12
target/arm/cpregs.h | 6 +++---
11
target/arm/helper.c | 9 ---------
13
target/arm/helper.c | 6 ++++--
12
2 files changed, 9 insertions(+), 9 deletions(-)
14
2 files changed, 7 insertions(+), 5 deletions(-)
13
15
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
16
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
18
--- a/target/arm/cpregs.h
17
+++ b/target/arm/internals.h
19
+++ b/target/arm/cpregs.h
18
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
20
@@ -XXX,XX +XXX,XX @@ enum {
19
}
21
* Note that we rely on the values of these enums as we iterate through
20
}
22
* the various states in some places.
21
22
+/* Return the TCR controlling this translation regime */
23
+static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
24
+{
25
+ if (mmu_idx == ARMMMUIdx_Stage2) {
26
+ return &env->cp15.vtcr_el2;
27
+ }
28
+ return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
29
+}
30
+
31
/* Return the FSR value for a debug exception (watchpoint, hardware
32
* breakpoint or BKPT insn) targeting the specified exception level.
33
*/
23
*/
24
-enum {
25
+typedef enum {
26
ARM_CP_STATE_AA32 = 0,
27
ARM_CP_STATE_AA64 = 1,
28
ARM_CP_STATE_BOTH = 2,
29
-};
30
+} CPState;
31
32
/*
33
* ARM CP register secure state flags. These flags identify security state
34
@@ -XXX,XX +XXX,XX @@ struct ARMCPRegInfo {
35
uint8_t opc1;
36
uint8_t opc2;
37
/* Execution state in which this register is visible: ARM_CP_STATE_* */
38
- int state;
39
+ CPState state;
40
/* Register type: ARM_CP_* bits/values */
41
int type;
42
/* Access rights: PL*_[RW] */
34
diff --git a/target/arm/helper.c b/target/arm/helper.c
43
diff --git a/target/arm/helper.c b/target/arm/helper.c
35
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/helper.c
45
--- a/target/arm/helper.c
37
+++ b/target/arm/helper.c
46
+++ b/target/arm/helper.c
38
@@ -XXX,XX +XXX,XX @@ static inline uint64_t regime_ttbr(CPUARMState *env, ARMMMUIdx mmu_idx,
47
@@ -XXX,XX +XXX,XX @@ CpuDefinitionInfoList *qmp_query_cpu_definitions(Error **errp)
39
48
}
40
#endif /* !CONFIG_USER_ONLY */
49
41
50
static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
42
-/* Return the TCR controlling this translation regime */
51
- void *opaque, int state, int secstate,
43
-static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
52
+ void *opaque, CPState state, int secstate,
44
-{
53
int crm, int opc1, int opc2,
45
- if (mmu_idx == ARMMMUIdx_Stage2) {
54
const char *name)
46
- return &env->cp15.vtcr_el2;
55
{
47
- }
56
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
48
- return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
57
* bits; the ARM_CP_64BIT* flag applies only to the AArch32 view of
49
-}
58
* the register, if any.
50
-
59
*/
51
/* Convert a possible stage1+2 MMU index into the appropriate
60
- int crm, opc1, opc2, state;
52
* stage 1 MMU index
61
+ int crm, opc1, opc2;
53
*/
62
int crmmin = (r->crm == CP_ANY) ? 0 : r->crm;
63
int crmmax = (r->crm == CP_ANY) ? 15 : r->crm;
64
int opc1min = (r->opc1 == CP_ANY) ? 0 : r->opc1;
65
int opc1max = (r->opc1 == CP_ANY) ? 7 : r->opc1;
66
int opc2min = (r->opc2 == CP_ANY) ? 0 : r->opc2;
67
int opc2max = (r->opc2 == CP_ANY) ? 7 : r->opc2;
68
+ CPState state;
69
+
70
/* 64 bit registers have only CRm and Opc1 fields */
71
assert(!((r->type & ARM_CP_64BIT) && (r->opc2 || r->crn)));
72
/* op0 only exists in the AArch64 encodings */
54
--
73
--
55
2.20.1
74
2.25.1
56
75
57
76
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Introduce an lvalue macro to wrap target_tlb_bit0.
3
Give this enum a name and use in ARMCPRegInfo and add_cpreg_to_hashtable.
4
Add the enumerator ARM_CP_SECSTATE_BOTH to clarify how 0
5
is handled in define_one_arm_cp_reg_with_opaque.
4
6
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-33-richard.henderson@linaro.org
9
Message-id: 20220501055028.646596-10-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/cpu.h | 13 +++++++++++++
12
target/arm/cpregs.h | 7 ++++---
11
target/arm/helper.c | 2 +-
13
target/arm/helper.c | 7 +++++--
12
target/arm/translate-a64.c | 2 +-
14
2 files changed, 9 insertions(+), 5 deletions(-)
13
3 files changed, 15 insertions(+), 2 deletions(-)
14
15
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
16
diff --git a/target/arm/cpregs.h b/target/arm/cpregs.h
16
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.h
18
--- a/target/arm/cpregs.h
18
+++ b/target/arm/cpu.h
19
+++ b/target/arm/cpregs.h
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
20
@@ -XXX,XX +XXX,XX @@ typedef enum {
20
/* Shared between translate-sve.c and sve_helper.c. */
21
* registered entry will only have one to identify whether the entry is secure
21
extern const uint64_t pred_esz_masks[4];
22
* or non-secure.
22
23
*/
23
+/* Helper for the macros below, validating the argument type. */
24
-enum {
24
+static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
25
+typedef enum {
25
+{
26
+ ARM_CP_SECSTATE_BOTH = 0, /* define one cpreg for each secstate */
26
+ return x;
27
ARM_CP_SECSTATE_S = (1 << 0), /* bit[0]: Secure state register */
27
+}
28
ARM_CP_SECSTATE_NS = (1 << 1), /* bit[1]: Non-secure state register */
28
+
29
-};
29
+/*
30
+} CPSecureState;
30
+ * Lvalue macros for ARM TLB bits that we must cache in the TCG TLB.
31
31
+ * Using these should be a bit more self-documenting than using the
32
+ * generic target bits directly.
33
+ */
34
+#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
35
+
36
/*
32
/*
37
* Naming convention for isar_feature functions:
33
* Access rights:
38
* Functions which test 32-bit ID registers should have _aa32_ in
34
@@ -XXX,XX +XXX,XX @@ struct ARMCPRegInfo {
35
/* Access rights: PL*_[RW] */
36
CPAccessRights access;
37
/* Security state: ARM_CP_SECSTATE_* bits/values */
38
- int secure;
39
+ CPSecureState secure;
40
/*
41
* The opaque pointer passed to define_arm_cp_regs_with_opaque() when
42
* this register was defined: can be used to hand data through to the
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
43
diff --git a/target/arm/helper.c b/target/arm/helper.c
40
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/helper.c
45
--- a/target/arm/helper.c
42
+++ b/target/arm/helper.c
46
+++ b/target/arm/helper.c
43
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
47
@@ -XXX,XX +XXX,XX @@ CpuDefinitionInfoList *qmp_query_cpu_definitions(Error **errp)
44
}
45
/* When in aarch64 mode, and BTI is enabled, remember GP in the IOTLB. */
46
if (aarch64 && guarded && cpu_isar_feature(aa64_bti, cpu)) {
47
- txattrs->target_tlb_bit0 = true;
48
+ arm_tlb_bti_gp(txattrs) = true;
49
}
50
51
if (cacheattrs != NULL) {
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
53
index XXXXXXX..XXXXXXX 100644
54
--- a/target/arm/translate-a64.c
55
+++ b/target/arm/translate-a64.c
56
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
57
* table entry even for that case.
58
*/
59
return (tlb_hit(entry->addr_code, addr) &&
60
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs.target_tlb_bit0);
61
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
62
#endif
63
}
48
}
64
49
50
static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
51
- void *opaque, CPState state, int secstate,
52
+ void *opaque, CPState state,
53
+ CPSecureState secstate,
54
int crm, int opc1, int opc2,
55
const char *name)
56
{
57
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
58
r->secure, crm, opc1, opc2,
59
r->name);
60
break;
61
- default:
62
+ case ARM_CP_SECSTATE_BOTH:
63
name = g_strdup_printf("%s_S", r->name);
64
add_cpreg_to_hashtable(cpu, r, opaque, state,
65
ARM_CP_SECSTATE_S,
66
@@ -XXX,XX +XXX,XX @@ void define_one_arm_cp_reg_with_opaque(ARMCPU *cpu,
67
ARM_CP_SECSTATE_NS,
68
crm, opc1, opc2, r->name);
69
break;
70
+ default:
71
+ g_assert_not_reached();
72
}
73
} else {
74
/* AArch64 registers get mapped to non-secure instance
65
--
75
--
66
2.20.1
76
2.25.1
67
68
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is TFSRE0_EL1, TFSR_EL1, TFSR_EL2, TFSR_EL3,
3
The new_key field is always non-zero -- drop the if.
4
RGSR_EL1, GCR_EL1, GMID_EL1, and PSTATE.TCO.
5
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20220501055028.646596-11-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-8-richard.henderson@linaro.org
8
[PMM: reinstated dropped PL3_RW mask]
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
target/arm/cpu.h | 4 ++
11
target/arm/helper.c | 23 +++++++++++------------
12
target/arm/internals.h | 9 ++++
12
1 file changed, 11 insertions(+), 12 deletions(-)
13
target/arm/helper.c | 94 ++++++++++++++++++++++++++++++++++++++
14
target/arm/translate-a64.c | 21 +++++++++
15
4 files changed, 128 insertions(+)
16
13
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
uint64_t pmccfiltr_el0; /* Performance Monitor Filter Register */
23
uint64_t vpidr_el2; /* Virtualization Processor ID Register */
24
uint64_t vmpidr_el2; /* Virtualization Multiprocessor ID Register */
25
+ uint64_t tfsr_el[4]; /* tfsre0_el1 is index 0. */
26
+ uint64_t gcr_el1;
27
+ uint64_t rgsr_el1;
28
} cp15;
29
30
struct {
31
@@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu);
32
#define PSTATE_SS (1U << 21)
33
#define PSTATE_PAN (1U << 22)
34
#define PSTATE_UAO (1U << 23)
35
+#define PSTATE_TCO (1U << 25)
36
#define PSTATE_V (1U << 28)
37
#define PSTATE_C (1U << 29)
38
#define PSTATE_Z (1U << 30)
39
diff --git a/target/arm/internals.h b/target/arm/internals.h
40
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/internals.h
42
+++ b/target/arm/internals.h
43
@@ -XXX,XX +XXX,XX @@ static inline uint32_t aarch64_pstate_valid_mask(const ARMISARegisters *id)
44
if (isar_feature_aa64_uao(id)) {
45
valid |= PSTATE_UAO;
46
}
47
+ if (isar_feature_aa64_mte(id)) {
48
+ valid |= PSTATE_TCO;
49
+ }
50
51
return valid;
52
}
53
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
54
55
#endif /* !CONFIG_USER_ONLY */
56
57
+/*
58
+ * The log2 of the words in the tag block, for GMID_EL1.BS.
59
+ * The is the maximum, 256 bytes, which manipulates 64-bits of tags.
60
+ */
61
+#define GMID_EL1_BS 6
62
+
63
#endif
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
65
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
67
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
68
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
18
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
69
{ K(3, 0, 1, 2, 0), K(3, 4, 1, 2, 0), K(3, 5, 1, 2, 0),
19
70
"ZCR_EL1", "ZCR_EL2", "ZCR_EL12", isar_feature_aa64_sve },
20
for (i = 0; i < ARRAY_SIZE(aliases); i++) {
71
21
const struct E2HAlias *a = &aliases[i];
72
+ { K(3, 0, 5, 6, 0), K(3, 4, 5, 6, 0), K(3, 5, 5, 6, 0),
22
- ARMCPRegInfo *src_reg, *dst_reg;
73
+ "TFSR_EL1", "TFSR_EL2", "TFSR_EL12", isar_feature_aa64_mte },
23
+ ARMCPRegInfo *src_reg, *dst_reg, *new_reg;
74
+
24
+ uint32_t *new_key;
75
/* TODO: ARMv8.2-SPE -- PMSCR_EL2 */
25
+ bool ok;
76
/* TODO: ARMv8.4-Trace -- TRFCR_EL2 */
26
77
};
27
if (a->feature && !a->feature(&cpu->isar)) {
78
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
28
continue;
79
};
29
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
80
#endif /*CONFIG_USER_ONLY*/
30
g_assert(src_reg->opaque == NULL);
81
31
82
+static CPAccessResult access_aa64_tid5(CPUARMState *env, const ARMCPRegInfo *ri,
32
/* Create alias before redirection so we dup the right data. */
83
+ bool isread)
33
- if (a->new_key) {
84
+{
34
- ARMCPRegInfo *new_reg = g_memdup(src_reg, sizeof(ARMCPRegInfo));
85
+ if ((arm_current_el(env) < 2) && (arm_hcr_el2_eff(env) & HCR_TID5)) {
35
- uint32_t *new_key = g_memdup(&a->new_key, sizeof(uint32_t));
86
+ return CP_ACCESS_TRAP_EL2;
36
- bool ok;
87
+ }
37
+ new_reg = g_memdup(src_reg, sizeof(ARMCPRegInfo));
88
+
38
+ new_key = g_memdup(&a->new_key, sizeof(uint32_t));
89
+ return CP_ACCESS_OK;
39
90
+}
40
- new_reg->name = a->new_name;
91
+
41
- new_reg->type |= ARM_CP_ALIAS;
92
+static CPAccessResult access_mte(CPUARMState *env, const ARMCPRegInfo *ri,
42
- /* Remove PL1/PL0 access, leaving PL2/PL3 R/W in place. */
93
+ bool isread)
43
- new_reg->access &= PL2_RW | PL3_RW;
94
+{
44
+ new_reg->name = a->new_name;
95
+ int el = arm_current_el(env);
45
+ new_reg->type |= ARM_CP_ALIAS;
96
+
46
+ /* Remove PL1/PL0 access, leaving PL2/PL3 R/W in place. */
97
+ if (el < 2 &&
47
+ new_reg->access &= PL2_RW | PL3_RW;
98
+ arm_feature(env, ARM_FEATURE_EL2) &&
48
99
+ !(arm_hcr_el2_eff(env) & HCR_ATA)) {
49
- ok = g_hash_table_insert(cpu->cp_regs, new_key, new_reg);
100
+ return CP_ACCESS_TRAP_EL2;
50
- g_assert(ok);
101
+ }
51
- }
102
+ if (el < 3 &&
52
+ ok = g_hash_table_insert(cpu->cp_regs, new_key, new_reg);
103
+ arm_feature(env, ARM_FEATURE_EL3) &&
53
+ g_assert(ok);
104
+ !(env->cp15.scr_el3 & SCR_ATA)) {
54
105
+ return CP_ACCESS_TRAP_EL3;
55
src_reg->opaque = dst_reg;
106
+ }
56
src_reg->orig_readfn = src_reg->readfn ?: raw_read;
107
+ return CP_ACCESS_OK;
108
+}
109
+
110
+static uint64_t tco_read(CPUARMState *env, const ARMCPRegInfo *ri)
111
+{
112
+ return env->pstate & PSTATE_TCO;
113
+}
114
+
115
+static void tco_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t val)
116
+{
117
+ env->pstate = (env->pstate & ~PSTATE_TCO) | (val & PSTATE_TCO);
118
+}
119
+
120
+static const ARMCPRegInfo mte_reginfo[] = {
121
+ { .name = "TFSRE0_EL1", .state = ARM_CP_STATE_AA64,
122
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 1,
123
+ .access = PL1_RW, .accessfn = access_mte,
124
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[0]) },
125
+ { .name = "TFSR_EL1", .state = ARM_CP_STATE_AA64,
126
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 0,
127
+ .access = PL1_RW, .accessfn = access_mte,
128
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[1]) },
129
+ { .name = "TFSR_EL2", .state = ARM_CP_STATE_AA64,
130
+ .opc0 = 3, .opc1 = 4, .crn = 5, .crm = 6, .opc2 = 0,
131
+ .access = PL2_RW, .accessfn = access_mte,
132
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[2]) },
133
+ { .name = "TFSR_EL3", .state = ARM_CP_STATE_AA64,
134
+ .opc0 = 3, .opc1 = 6, .crn = 5, .crm = 6, .opc2 = 0,
135
+ .access = PL3_RW,
136
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[3]) },
137
+ { .name = "RGSR_EL1", .state = ARM_CP_STATE_AA64,
138
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 5,
139
+ .access = PL1_RW, .accessfn = access_mte,
140
+ .fieldoffset = offsetof(CPUARMState, cp15.rgsr_el1) },
141
+ { .name = "GCR_EL1", .state = ARM_CP_STATE_AA64,
142
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 6,
143
+ .access = PL1_RW, .accessfn = access_mte,
144
+ .fieldoffset = offsetof(CPUARMState, cp15.gcr_el1) },
145
+ { .name = "GMID_EL1", .state = ARM_CP_STATE_AA64,
146
+ .opc0 = 3, .opc1 = 1, .crn = 0, .crm = 0, .opc2 = 4,
147
+ .access = PL1_R, .accessfn = access_aa64_tid5,
148
+ .type = ARM_CP_CONST, .resetvalue = GMID_EL1_BS },
149
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
150
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
151
+ .type = ARM_CP_NO_RAW,
152
+ .access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
153
+ REGINFO_SENTINEL
154
+};
155
+
156
+static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
157
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
158
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
159
+ .type = ARM_CP_CONST, .access = PL0_RW, },
160
+ REGINFO_SENTINEL
161
+};
162
#endif
163
164
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
165
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
166
}
167
}
168
#endif /*CONFIG_USER_ONLY*/
169
+
170
+ /*
171
+ * If full MTE is enabled, add all of the system registers.
172
+ * If only "instructions available at EL0" are enabled,
173
+ * then define only a RAZ/WI version of PSTATE.TCO.
174
+ */
175
+ if (cpu_isar_feature(aa64_mte, cpu)) {
176
+ define_arm_cp_regs(cpu, mte_reginfo);
177
+ } else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
178
+ define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
179
+ }
180
#endif
181
182
if (cpu_isar_feature(any_predinv, cpu)) {
183
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
184
index XXXXXXX..XXXXXXX 100644
185
--- a/target/arm/translate-a64.c
186
+++ b/target/arm/translate-a64.c
187
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
188
s->base.is_jmp = DISAS_UPDATE_EXIT;
189
break;
190
191
+ case 0x1c: /* TCO */
192
+ if (dc_isar_feature(aa64_mte, s)) {
193
+ /* Full MTE is enabled -- set the TCO bit as directed. */
194
+ if (crm & 1) {
195
+ set_pstate_bits(PSTATE_TCO);
196
+ } else {
197
+ clear_pstate_bits(PSTATE_TCO);
198
+ }
199
+ t1 = tcg_const_i32(s->current_el);
200
+ gen_helper_rebuild_hflags_a64(cpu_env, t1);
201
+ tcg_temp_free_i32(t1);
202
+ /* Many factors, including TCO, go into MTE_ACTIVE. */
203
+ s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
204
+ } else if (dc_isar_feature(aa64_mte_insn_reg, s)) {
205
+ /* Only "instructions accessible at EL0" -- PSTATE.TCO is WI. */
206
+ s->base.is_jmp = DISAS_NEXT;
207
+ } else {
208
+ goto do_unallocated;
209
+ }
210
+ break;
211
+
212
default:
213
do_unallocated:
214
unallocated_encoding(s);
215
--
57
--
216
2.20.1
58
2.25.1
217
218
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
There are a number of paths by which the TBI is still intact
3
Cast the uint32_t key into a gpointer directly, which
4
for user-only in the SVE helpers.
4
allows us to avoid allocating storage for each key.
5
5
6
Because we currently always set TBI for user-only, we do not
6
Use g_hash_table_lookup when we already have a gpointer
7
need to pass down the actual TBI setting from above, and we
7
(e.g. for callbacks like count_cpreg), or when using
8
can remove the top byte in the inner-most primitives, so that
8
get_arm_cp_reginfo would require casting away const.
9
none are forgotten. Moreover, this keeps the "dirty" pointer
10
around at the higher levels, where we need it for any MTE checking.
11
9
12
Since the normal case, especially for user-only, goes through
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
RAM, this clearing merely adds two insns per page lookup, which
14
will be completely in the noise.
15
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20220501055028.646596-12-richard.henderson@linaro.org
18
Message-id: 20200626033144.790098-39-richard.henderson@linaro.org
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
---
14
---
21
target/arm/cpu.c | 3 +++
15
target/arm/cpu.c | 4 ++--
22
target/arm/sve_helper.c | 19 +++++++++++++++++--
16
target/arm/gdbstub.c | 2 +-
23
target/arm/translate-a64.c | 5 +++++
17
target/arm/helper.c | 41 ++++++++++++++++++-----------------------
24
3 files changed, 25 insertions(+), 2 deletions(-)
18
3 files changed, 21 insertions(+), 26 deletions(-)
25
19
26
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
20
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
27
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/cpu.c
22
--- a/target/arm/cpu.c
29
+++ b/target/arm/cpu.c
23
+++ b/target/arm/cpu.c
30
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
24
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_initfn(Object *obj)
31
* Enable TBI0 and TBI1. While the real kernel only enables TBI0,
25
ARMCPU *cpu = ARM_CPU(obj);
32
* turning on both here will produce smaller code and otherwise
26
33
* make no difference to the user-level emulation.
27
cpu_set_cpustate_pointers(cpu);
34
+ *
28
- cpu->cp_regs = g_hash_table_new_full(g_int_hash, g_int_equal,
35
+ * In sve_probe_page, we assume that this is set.
29
- g_free, cpreg_hashtable_data_destroy);
36
+ * Do not modify this without other changes.
30
+ cpu->cp_regs = g_hash_table_new_full(g_direct_hash, g_direct_equal,
37
*/
31
+ NULL, cpreg_hashtable_data_destroy);
38
env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
32
39
#else
33
QLIST_INIT(&cpu->pre_el_change_hooks);
40
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
34
QLIST_INIT(&cpu->el_change_hooks);
35
diff --git a/target/arm/gdbstub.c b/target/arm/gdbstub.c
41
index XXXXXXX..XXXXXXX 100644
36
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/sve_helper.c
37
--- a/target/arm/gdbstub.c
43
+++ b/target/arm/sve_helper.c
38
+++ b/target/arm/gdbstub.c
44
@@ -XXX,XX +XXX,XX @@ static void sve_##NAME##_host(void *vd, intptr_t reg_off, void *host) \
39
@@ -XXX,XX +XXX,XX @@ static void arm_gen_one_xml_sysreg_tag(GString *s, DynamicGDBXMLInfo *dyn_xml,
45
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
40
static void arm_register_sysreg_for_xml(gpointer key, gpointer value,
46
target_ulong addr, uintptr_t ra) \
41
gpointer p)
47
{ \
42
{
48
- *(TYPEE *)(vd + H(reg_off)) = (TYPEM)TLB(env, addr, ra); \
43
- uint32_t ri_key = *(uint32_t *)key;
49
+ *(TYPEE *)(vd + H(reg_off)) = \
44
+ uint32_t ri_key = (uintptr_t)key;
50
+ (TYPEM)TLB(env, useronly_clean_ptr(addr), ra); \
45
ARMCPRegInfo *ri = value;
46
RegisterSysregXmlParam *param = (RegisterSysregXmlParam *)p;
47
GString *s = param->s;
48
diff --git a/target/arm/helper.c b/target/arm/helper.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/target/arm/helper.c
51
+++ b/target/arm/helper.c
52
@@ -XXX,XX +XXX,XX @@ bool write_list_to_cpustate(ARMCPU *cpu)
53
static void add_cpreg_to_list(gpointer key, gpointer opaque)
54
{
55
ARMCPU *cpu = opaque;
56
- uint64_t regidx;
57
- const ARMCPRegInfo *ri;
58
-
59
- regidx = *(uint32_t *)key;
60
- ri = get_arm_cp_reginfo(cpu->cp_regs, regidx);
61
+ uint32_t regidx = (uintptr_t)key;
62
+ const ARMCPRegInfo *ri = get_arm_cp_reginfo(cpu->cp_regs, regidx);
63
64
if (!(ri->type & (ARM_CP_NO_RAW|ARM_CP_ALIAS))) {
65
cpu->cpreg_indexes[cpu->cpreg_array_len] = cpreg_to_kvm_id(regidx);
66
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_list(gpointer key, gpointer opaque)
67
static void count_cpreg(gpointer key, gpointer opaque)
68
{
69
ARMCPU *cpu = opaque;
70
- uint64_t regidx;
71
const ARMCPRegInfo *ri;
72
73
- regidx = *(uint32_t *)key;
74
- ri = get_arm_cp_reginfo(cpu->cp_regs, regidx);
75
+ ri = g_hash_table_lookup(cpu->cp_regs, key);
76
77
if (!(ri->type & (ARM_CP_NO_RAW|ARM_CP_ALIAS))) {
78
cpu->cpreg_array_len++;
79
@@ -XXX,XX +XXX,XX @@ static void count_cpreg(gpointer key, gpointer opaque)
80
81
static gint cpreg_key_compare(gconstpointer a, gconstpointer b)
82
{
83
- uint64_t aidx = cpreg_to_kvm_id(*(uint32_t *)a);
84
- uint64_t bidx = cpreg_to_kvm_id(*(uint32_t *)b);
85
+ uint64_t aidx = cpreg_to_kvm_id((uintptr_t)a);
86
+ uint64_t bidx = cpreg_to_kvm_id((uintptr_t)b);
87
88
if (aidx > bidx) {
89
return 1;
90
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
91
for (i = 0; i < ARRAY_SIZE(aliases); i++) {
92
const struct E2HAlias *a = &aliases[i];
93
ARMCPRegInfo *src_reg, *dst_reg, *new_reg;
94
- uint32_t *new_key;
95
bool ok;
96
97
if (a->feature && !a->feature(&cpu->isar)) {
98
continue;
99
}
100
101
- src_reg = g_hash_table_lookup(cpu->cp_regs, &a->src_key);
102
- dst_reg = g_hash_table_lookup(cpu->cp_regs, &a->dst_key);
103
+ src_reg = g_hash_table_lookup(cpu->cp_regs,
104
+ (gpointer)(uintptr_t)a->src_key);
105
+ dst_reg = g_hash_table_lookup(cpu->cp_regs,
106
+ (gpointer)(uintptr_t)a->dst_key);
107
g_assert(src_reg != NULL);
108
g_assert(dst_reg != NULL);
109
110
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
111
112
/* Create alias before redirection so we dup the right data. */
113
new_reg = g_memdup(src_reg, sizeof(ARMCPRegInfo));
114
- new_key = g_memdup(&a->new_key, sizeof(uint32_t));
115
116
new_reg->name = a->new_name;
117
new_reg->type |= ARM_CP_ALIAS;
118
/* Remove PL1/PL0 access, leaving PL2/PL3 R/W in place. */
119
new_reg->access &= PL2_RW | PL3_RW;
120
121
- ok = g_hash_table_insert(cpu->cp_regs, new_key, new_reg);
122
+ ok = g_hash_table_insert(cpu->cp_regs,
123
+ (gpointer)(uintptr_t)a->new_key, new_reg);
124
g_assert(ok);
125
126
src_reg->opaque = dst_reg;
127
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
128
/* Private utility function for define_one_arm_cp_reg_with_opaque():
129
* add a single reginfo struct to the hash table.
130
*/
131
- uint32_t *key = g_new(uint32_t, 1);
132
+ uint32_t key;
133
ARMCPRegInfo *r2 = g_memdup(r, sizeof(ARMCPRegInfo));
134
int is64 = (r->type & ARM_CP_64BIT) ? 1 : 0;
135
int ns = (secstate & ARM_CP_SECSTATE_NS) ? 1 : 0;
136
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
137
if (r->cp == 0 || r->state == ARM_CP_STATE_BOTH) {
138
r2->cp = CP_REG_ARM64_SYSREG_CP;
139
}
140
- *key = ENCODE_AA64_CP_REG(r2->cp, r2->crn, crm,
141
- r2->opc0, opc1, opc2);
142
+ key = ENCODE_AA64_CP_REG(r2->cp, r2->crn, crm,
143
+ r2->opc0, opc1, opc2);
144
} else {
145
- *key = ENCODE_CP_REG(r2->cp, is64, ns, r2->crn, crm, opc1, opc2);
146
+ key = ENCODE_CP_REG(r2->cp, is64, ns, r2->crn, crm, opc1, opc2);
147
}
148
if (opaque) {
149
r2->opaque = opaque;
150
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
151
* requested.
152
*/
153
if (!(r->type & ARM_CP_OVERRIDE)) {
154
- ARMCPRegInfo *oldreg;
155
- oldreg = g_hash_table_lookup(cpu->cp_regs, key);
156
+ const ARMCPRegInfo *oldreg = get_arm_cp_reginfo(cpu->cp_regs, key);
157
if (oldreg && !(oldreg->type & ARM_CP_OVERRIDE)) {
158
fprintf(stderr, "Register redefined: cp=%d %d bit "
159
"crn=%d crm=%d opc1=%d opc2=%d, "
160
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
161
g_assert_not_reached();
162
}
163
}
164
- g_hash_table_insert(cpu->cp_regs, key, r2);
165
+ g_hash_table_insert(cpu->cp_regs, (gpointer)(uintptr_t)key, r2);
51
}
166
}
52
167
53
#define DO_ST_TLB(NAME, H, TYPEE, TYPEM, TLB) \
168
54
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
169
@@ -XXX,XX +XXX,XX @@ void modify_arm_cp_regs_with_len(ARMCPRegInfo *regs, size_t regs_len,
55
target_ulong addr, uintptr_t ra) \
170
56
{ \
171
const ARMCPRegInfo *get_arm_cp_reginfo(GHashTable *cpregs, uint32_t encoded_cp)
57
- TLB(env, addr, (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
172
{
58
+ TLB(env, useronly_clean_ptr(addr), \
173
- return g_hash_table_lookup(cpregs, &encoded_cp);
59
+ (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
174
+ return g_hash_table_lookup(cpregs, (gpointer)(uintptr_t)encoded_cp);
60
}
175
}
61
176
62
#define DO_LD_PRIM_1(NAME, H, TE, TM) \
177
void arm_cp_write_ignore(CPUARMState *env, const ARMCPRegInfo *ri,
63
@@ -XXX,XX +XXX,XX @@ static bool sve_probe_page(SVEHostPage *info, bool nofault,
64
int flags;
65
66
addr += mem_off;
67
+
68
+ /*
69
+ * User-only currently always issues with TBI. See the comment
70
+ * above useronly_clean_ptr. Usually we clean this top byte away
71
+ * during translation, but we can't do that for e.g. vector + imm
72
+ * addressing modes.
73
+ *
74
+ * We currently always enable TBI for user-only, and do not provide
75
+ * a way to turn it off. So clean the pointer unconditionally here,
76
+ * rather than look it up here, or pass it down from above.
77
+ */
78
+ addr = useronly_clean_ptr(addr);
79
+
80
flags = probe_access_flags(env, addr, access_type, mmu_idx, nofault,
81
&info->host, retaddr);
82
info->flags = flags;
83
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/target/arm/translate-a64.c
86
+++ b/target/arm/translate-a64.c
87
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
88
dc->features = env->features;
89
dc->dcz_blocksize = arm_cpu->dcz_blocksize;
90
91
+#ifdef CONFIG_USER_ONLY
92
+ /* In sve_probe_page, we assume TBI is enabled. */
93
+ tcg_debug_assert(dc->tbid & 1);
94
+#endif
95
+
96
/* Single step state. The code-generation logic here is:
97
* SS_ACTIVE == 0:
98
* generate code with no special handling for single-stepping (except
99
--
178
--
100
2.20.1
179
2.25.1
101
102
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Use the same code as system mode, so that we generate the same
3
Simplify freeing cp_regs hash table entries by using a single
4
exception + syndrome for the unaligned access.
4
allocation for the entire value.
5
5
6
For the moment, if MTE is enabled so that this path is reachable,
6
This fixes a theoretical bug if we were to ever free the entire
7
this would generate a SIGSEGV in the user-only cpu_loop. Decoding
7
hash table, because we've been installing string literal constants
8
the syndrome to produce the proper SIGBUS will be done later.
8
into the cpreg structure in define_arm_vh_e2h_redirects_aliases.
9
However, at present we only free entries created for AArch32
10
wildcard cpregs which get overwritten by more specific cpregs,
11
so this bug is never exposed.
9
12
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 20220501055028.646596-13-richard.henderson@linaro.org
12
Message-id: 20200626033144.790098-15-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
17
---
15
target/arm/cpu.c | 2 +-
18
target/arm/cpu.c | 16 +---------------
16
target/arm/tlb_helper.c | 4 ++--
19
target/arm/helper.c | 10 ++++++++--
17
2 files changed, 3 insertions(+), 3 deletions(-)
20
2 files changed, 9 insertions(+), 17 deletions(-)
18
21
19
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
22
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
20
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.c
24
--- a/target/arm/cpu.c
22
+++ b/target/arm/cpu.c
25
+++ b/target/arm/cpu.c
23
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
26
@@ -XXX,XX +XXX,XX @@ uint64_t arm_cpu_mp_affinity(int idx, uint8_t clustersz)
24
cc->tlb_fill = arm_cpu_tlb_fill;
27
return (Aff1 << ARM_AFF1_SHIFT) | Aff0;
25
cc->debug_excp_handler = arm_debug_excp_handler;
28
}
26
cc->debug_check_watchpoint = arm_debug_check_watchpoint;
29
27
-#if !defined(CONFIG_USER_ONLY)
30
-static void cpreg_hashtable_data_destroy(gpointer data)
28
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
31
-{
29
+#if !defined(CONFIG_USER_ONLY)
32
- /*
30
cc->do_transaction_failed = arm_cpu_do_transaction_failed;
33
- * Destroy function for cpu->cp_regs hashtable data entries.
31
cc->adjust_watchpoint_address = arm_adjust_watchpoint_address;
34
- * We must free the name string because it was g_strdup()ed in
32
#endif /* CONFIG_TCG && !CONFIG_USER_ONLY */
35
- * add_cpreg_to_hashtable(). It's OK to cast away the 'const'
33
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
36
- * from r->name because we know we definitely allocated it.
37
- */
38
- ARMCPRegInfo *r = data;
39
-
40
- g_free((void *)r->name);
41
- g_free(r);
42
-}
43
-
44
static void arm_cpu_initfn(Object *obj)
45
{
46
ARMCPU *cpu = ARM_CPU(obj);
47
48
cpu_set_cpustate_pointers(cpu);
49
cpu->cp_regs = g_hash_table_new_full(g_direct_hash, g_direct_equal,
50
- NULL, cpreg_hashtable_data_destroy);
51
+ NULL, g_free);
52
53
QLIST_INIT(&cpu->pre_el_change_hooks);
54
QLIST_INIT(&cpu->el_change_hooks);
55
diff --git a/target/arm/helper.c b/target/arm/helper.c
34
index XXXXXXX..XXXXXXX 100644
56
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/tlb_helper.c
57
--- a/target/arm/helper.c
36
+++ b/target/arm/tlb_helper.c
58
+++ b/target/arm/helper.c
37
@@ -XXX,XX +XXX,XX @@
59
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
38
#include "internals.h"
60
* add a single reginfo struct to the hash table.
39
#include "exec/exec-all.h"
61
*/
40
62
uint32_t key;
41
-#if !defined(CONFIG_USER_ONLY)
63
- ARMCPRegInfo *r2 = g_memdup(r, sizeof(ARMCPRegInfo));
42
-
64
+ ARMCPRegInfo *r2;
43
static inline uint32_t merge_syn_data_abort(uint32_t template_syn,
65
int is64 = (r->type & ARM_CP_64BIT) ? 1 : 0;
44
unsigned int target_el,
66
int ns = (secstate & ARM_CP_SECSTATE_NS) ? 1 : 0;
45
bool same_el, bool ea,
67
+ size_t name_len;
46
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
47
arm_deliver_fault(cpu, vaddr, access_type, mmu_idx, &fi);
48
}
49
50
+#if !defined(CONFIG_USER_ONLY)
51
+
68
+
52
/*
69
+ /* Combine cpreg and name into one allocation. */
53
* arm_cpu_do_transaction_failed: handle a memory system error response
70
+ name_len = strlen(name) + 1;
54
* (eg "no device/memory present at address") by raising an external abort
71
+ r2 = g_malloc(sizeof(*r2) + name_len);
72
+ *r2 = *r;
73
+ r2->name = memcpy(r2 + 1, name, name_len);
74
75
- r2->name = g_strdup(name);
76
/* Reset the secure state to the specific incoming state. This is
77
* necessary as the register may have been defined with both states.
78
*/
55
--
79
--
56
2.20.1
80
2.25.1
57
58
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Move the computation of key to the top of the function.
4
Hoist the resolution of cp as well, as an input to the
5
computation of key.
6
7
This will be required by a subsequent patch.
2
8
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-44-richard.henderson@linaro.org
11
Message-id: 20220501055028.646596-14-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
13
---
8
target/arm/cpu.h | 6 ++++++
14
target/arm/helper.c | 49 +++++++++++++++++++++++++--------------------
9
hw/arm/virt.c | 55 ++++++++++++++++++++++++++++++++++++++++++++++--
15
1 file changed, 27 insertions(+), 22 deletions(-)
10
target/arm/cpu.c | 52 +++++++++++++++++++++++++++++++++++++++++----
11
3 files changed, 107 insertions(+), 6 deletions(-)
12
16
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
17
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.h
19
--- a/target/arm/helper.c
16
+++ b/target/arm/cpu.h
20
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
21
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
18
/* MemoryRegion to use for secure physical accesses */
22
ARMCPRegInfo *r2;
19
MemoryRegion *secure_memory;
23
int is64 = (r->type & ARM_CP_64BIT) ? 1 : 0;
20
24
int ns = (secstate & ARM_CP_SECSTATE_NS) ? 1 : 0;
21
+ /* MemoryRegion to use for allocation tag accesses */
25
+ int cp = r->cp;
22
+ MemoryRegion *tag_memory;
26
size_t name_len;
23
+ MemoryRegion *secure_tag_memory;
27
24
+
28
+ switch (state) {
25
/* For v8M, pointer to the IDAU interface provided by board/SoC */
29
+ case ARM_CP_STATE_AA32:
26
Object *idau;
30
+ /* We assume it is a cp15 register if the .cp field is left unset. */
27
31
+ if (cp == 0 && r->state == ARM_CP_STATE_BOTH) {
28
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
32
+ cp = 15;
29
typedef enum ARMASIdx {
33
+ }
30
ARMASIdx_NS = 0,
34
+ key = ENCODE_CP_REG(cp, is64, ns, r->crn, crm, opc1, opc2);
31
ARMASIdx_S = 1,
35
+ break;
32
+ ARMASIdx_TagNS = 2,
36
+ case ARM_CP_STATE_AA64:
33
+ ARMASIdx_TagS = 3,
37
+ /*
34
} ARMASIdx;
38
+ * To allow abbreviation of ARMCPRegInfo definitions, we treat
35
39
+ * cp == 0 as equivalent to the value for "standard guest-visible
36
/* Return the Exception Level targeted by debug exceptions. */
40
+ * sysreg". STATE_BOTH definitions are also always "standard sysreg"
37
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
41
+ * in their AArch64 view (the .cp value may be non-zero for the
38
index XXXXXXX..XXXXXXX 100644
42
+ * benefit of the AArch32 view).
39
--- a/hw/arm/virt.c
43
+ */
40
+++ b/hw/arm/virt.c
44
+ if (cp == 0 || r->state == ARM_CP_STATE_BOTH) {
41
@@ -XXX,XX +XXX,XX @@ static void create_platform_bus(VirtMachineState *vms)
45
+ cp = CP_REG_ARM64_SYSREG_CP;
42
sysbus_mmio_get_region(s, 0));
46
+ }
43
}
47
+ key = ENCODE_AA64_CP_REG(cp, r->crn, crm, r->opc0, opc1, opc2);
44
48
+ break;
45
+static void create_tag_ram(MemoryRegion *tag_sysmem,
49
+ default:
46
+ hwaddr base, hwaddr size,
50
+ g_assert_not_reached();
47
+ const char *name)
48
+{
49
+ MemoryRegion *tagram = g_new(MemoryRegion, 1);
50
+
51
+ memory_region_init_ram(tagram, NULL, name, size / 32, &error_fatal);
52
+ memory_region_add_subregion(tag_sysmem, base / 32, tagram);
53
+}
54
+
55
static void create_secure_ram(VirtMachineState *vms,
56
- MemoryRegion *secure_sysmem)
57
+ MemoryRegion *secure_sysmem,
58
+ MemoryRegion *secure_tag_sysmem)
59
{
60
MemoryRegion *secram = g_new(MemoryRegion, 1);
61
char *nodename;
62
@@ -XXX,XX +XXX,XX @@ static void create_secure_ram(VirtMachineState *vms,
63
qemu_fdt_setprop_string(vms->fdt, nodename, "status", "disabled");
64
qemu_fdt_setprop_string(vms->fdt, nodename, "secure-status", "okay");
65
66
+ if (secure_tag_sysmem) {
67
+ create_tag_ram(secure_tag_sysmem, base, size, "mach-virt.secure-tag");
68
+ }
51
+ }
69
+
52
+
70
g_free(nodename);
53
/* Combine cpreg and name into one allocation. */
71
}
54
name_len = strlen(name) + 1;
72
55
r2 = g_malloc(sizeof(*r2) + name_len);
73
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
56
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
74
const CPUArchIdList *possible_cpus;
75
MemoryRegion *sysmem = get_system_memory();
76
MemoryRegion *secure_sysmem = NULL;
77
+ MemoryRegion *tag_sysmem = NULL;
78
+ MemoryRegion *secure_tag_sysmem = NULL;
79
int n, virt_max_cpus;
80
bool firmware_loaded;
81
bool aarch64 = true;
82
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
83
"secure-memory", &error_abort);
84
}
57
}
85
58
86
+ /*
59
if (r->state == ARM_CP_STATE_BOTH) {
87
+ * The cpu adds the property if and only if MemTag is supported.
60
- /* We assume it is a cp15 register if the .cp field is left unset.
88
+ * If it is, we must allocate the ram to back that up.
61
- */
89
+ */
62
- if (r2->cp == 0) {
90
+ if (object_property_find(cpuobj, "tag-memory", NULL)) {
63
- r2->cp = 15;
91
+ if (!tag_sysmem) {
64
- }
92
+ tag_sysmem = g_new(MemoryRegion, 1);
65
-
93
+ memory_region_init(tag_sysmem, OBJECT(machine),
66
#if HOST_BIG_ENDIAN
94
+ "tag-memory", UINT64_MAX / 32);
67
if (r2->fieldoffset) {
95
+
68
r2->fieldoffset += sizeof(uint32_t);
96
+ if (vms->secure) {
69
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
97
+ secure_tag_sysmem = g_new(MemoryRegion, 1);
70
#endif
98
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
71
}
99
+ "secure-tag-memory", UINT64_MAX / 32);
100
+
101
+ /* As with ram, secure-tag takes precedence over tag. */
102
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
103
+ tag_sysmem, -1);
104
+ }
105
+ }
106
+
107
+ object_property_set_link(cpuobj, OBJECT(tag_sysmem),
108
+ "tag-memory", &error_abort);
109
+ if (vms->secure) {
110
+ object_property_set_link(cpuobj, OBJECT(secure_tag_sysmem),
111
+ "secure-tag-memory", &error_abort);
112
+ }
113
+ }
114
+
115
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
116
object_unref(cpuobj);
117
}
72
}
118
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
73
- if (state == ARM_CP_STATE_AA64) {
119
create_uart(vms, VIRT_UART, sysmem, serial_hd(0));
74
- /* To allow abbreviation of ARMCPRegInfo
120
75
- * definitions, we treat cp == 0 as equivalent to
121
if (vms->secure) {
76
- * the value for "standard guest-visible sysreg".
122
- create_secure_ram(vms, secure_sysmem);
77
- * STATE_BOTH definitions are also always "standard
123
+ create_secure_ram(vms, secure_sysmem, secure_tag_sysmem);
78
- * sysreg" in their AArch64 view (the .cp value may
124
create_uart(vms, VIRT_SECURE_UART, secure_sysmem, serial_hd(1));
79
- * be non-zero for the benefit of the AArch32 view).
80
- */
81
- if (r->cp == 0 || r->state == ARM_CP_STATE_BOTH) {
82
- r2->cp = CP_REG_ARM64_SYSREG_CP;
83
- }
84
- key = ENCODE_AA64_CP_REG(r2->cp, r2->crn, crm,
85
- r2->opc0, opc1, opc2);
86
- } else {
87
- key = ENCODE_CP_REG(r2->cp, is64, ns, r2->crn, crm, opc1, opc2);
88
- }
89
if (opaque) {
90
r2->opaque = opaque;
125
}
91
}
126
92
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
127
+ if (tag_sysmem) {
93
/* Make sure reginfo passed to helpers for wildcarded regs
128
+ create_tag_ram(tag_sysmem, vms->memmap[VIRT_MEM].base,
94
* has the correct crm/opc1/opc2 for this reg, not CP_ANY:
129
+ machine->ram_size, "mach-virt.tag");
95
*/
130
+ }
96
+ r2->cp = cp;
131
+
97
r2->crm = crm;
132
vms->highmem_ecam &= vms->highmem && (!firmware_loaded || aarch64);
98
r2->opc1 = opc1;
133
99
r2->opc2 = opc2;
134
create_rtc(vms);
135
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
136
index XXXXXXX..XXXXXXX 100644
137
--- a/target/arm/cpu.c
138
+++ b/target/arm/cpu.c
139
@@ -XXX,XX +XXX,XX @@ void arm_cpu_post_init(Object *obj)
140
if (kvm_enabled()) {
141
kvm_arm_add_vcpu_properties(obj);
142
}
143
+
144
+#ifndef CONFIG_USER_ONLY
145
+ if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) &&
146
+ cpu_isar_feature(aa64_mte, cpu)) {
147
+ object_property_add_link(obj, "tag-memory",
148
+ TYPE_MEMORY_REGION,
149
+ (Object **)&cpu->tag_memory,
150
+ qdev_prop_allow_set_link_before_realize,
151
+ OBJ_PROP_LINK_STRONG);
152
+
153
+ if (arm_feature(&cpu->env, ARM_FEATURE_EL3)) {
154
+ object_property_add_link(obj, "secure-tag-memory",
155
+ TYPE_MEMORY_REGION,
156
+ (Object **)&cpu->secure_tag_memory,
157
+ qdev_prop_allow_set_link_before_realize,
158
+ OBJ_PROP_LINK_STRONG);
159
+ }
160
+ }
161
+#endif
162
}
163
164
static void arm_cpu_finalizefn(Object *obj)
165
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
166
#ifndef CONFIG_USER_ONLY
167
MachineState *ms = MACHINE(qdev_get_machine());
168
unsigned int smp_cpus = ms->smp.cpus;
169
+ bool has_secure = cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY);
170
171
- if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
172
- cs->num_ases = 2;
173
+ /*
174
+ * We must set cs->num_ases to the final value before
175
+ * the first call to cpu_address_space_init.
176
+ */
177
+ if (cpu->tag_memory != NULL) {
178
+ cs->num_ases = 3 + has_secure;
179
+ } else {
180
+ cs->num_ases = 1 + has_secure;
181
+ }
182
183
+ if (has_secure) {
184
if (!cpu->secure_memory) {
185
cpu->secure_memory = cs->memory;
186
}
187
cpu_address_space_init(cs, ARMASIdx_S, "cpu-secure-memory",
188
cpu->secure_memory);
189
- } else {
190
- cs->num_ases = 1;
191
}
192
+
193
+ if (cpu->tag_memory != NULL) {
194
+ cpu_address_space_init(cs, ARMASIdx_TagNS, "cpu-tag-memory",
195
+ cpu->tag_memory);
196
+ if (has_secure) {
197
+ cpu_address_space_init(cs, ARMASIdx_TagS, "cpu-tag-memory",
198
+ cpu->secure_tag_memory);
199
+ }
200
+ } else if (cpu_isar_feature(aa64_mte, cpu)) {
201
+ /*
202
+ * Since there is no tag memory, we can't meaningfully support MTE
203
+ * to its fullest. To avoid problems later, when we would come to
204
+ * use the tag memory, downgrade support to insns only.
205
+ */
206
+ cpu->isar.id_aa64pfr1 =
207
+ FIELD_DP64(cpu->isar.id_aa64pfr1, ID_AA64PFR1, MTE, 1);
208
+ }
209
+
210
cpu_address_space_init(cs, ARMASIdx_NS, "cpu-memory", cs->memory);
211
212
/* No core_count specified, default to smp_cpus. */
213
--
100
--
214
2.20.1
101
2.25.1
215
216
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
D1.10 specifies that exception handlers begin with tag checks overridden.
3
Put most of the value writeback to the same place,
4
and improve the comment that goes with them.
4
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220501055028.646596-15-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-41-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
---
10
target/arm/helper.c | 3 +++
11
target/arm/helper.c | 28 ++++++++++++----------------
11
1 file changed, 3 insertions(+)
12
1 file changed, 12 insertions(+), 16 deletions(-)
12
13
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
16
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
18
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
18
break;
19
*r2 = *r;
20
r2->name = memcpy(r2 + 1, name, name_len);
21
22
- /* Reset the secure state to the specific incoming state. This is
23
- * necessary as the register may have been defined with both states.
24
+ /*
25
+ * Update fields to match the instantiation, overwiting wildcards
26
+ * such as CP_ANY, ARM_CP_STATE_BOTH, or ARM_CP_SECSTATE_BOTH.
27
*/
28
+ r2->cp = cp;
29
+ r2->crm = crm;
30
+ r2->opc1 = opc1;
31
+ r2->opc2 = opc2;
32
+ r2->state = state;
33
r2->secure = secstate;
34
+ if (opaque) {
35
+ r2->opaque = opaque;
36
+ }
37
38
if (r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1]) {
39
/* Register is banked (using both entries in array).
40
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
41
#endif
19
}
42
}
20
}
43
}
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
44
- if (opaque) {
22
+ new_mode |= PSTATE_TCO;
45
- r2->opaque = opaque;
23
+ }
46
- }
24
47
- /* reginfo passed to helpers is correct for the actual access,
25
pstate_write(env, PSTATE_DAIF | new_mode);
48
- * and is never ARM_CP_STATE_BOTH:
26
env->aarch64 = 1;
49
- */
50
- r2->state = state;
51
- /* Make sure reginfo passed to helpers for wildcarded regs
52
- * has the correct crm/opc1/opc2 for this reg, not CP_ANY:
53
- */
54
- r2->cp = cp;
55
- r2->crm = crm;
56
- r2->opc1 = opc1;
57
- r2->opc2 = opc2;
58
+
59
/* By convention, for wildcarded registers only the first
60
* entry is used for migration; the others are marked as
61
* ALIAS so we don't try to transfer the register
27
--
62
--
28
2.20.1
63
2.25.1
29
30
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Like the regular data cache flushes, these are nops within qemu.
3
Bool is a more appropriate type for these variables.
4
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20220501055028.646596-16-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-21-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
9
---
10
target/arm/helper.c | 65 +++++++++++++++++++++++++++++++++++++++++++++
10
target/arm/helper.c | 4 ++--
11
1 file changed, 65 insertions(+)
11
1 file changed, 2 insertions(+), 2 deletions(-)
12
12
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
16
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
17
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
18
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
19
.type = ARM_CP_NO_RAW,
20
.access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
21
+ { .name = "DC_IGVAC", .state = ARM_CP_STATE_AA64,
22
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 3,
23
+ .type = ARM_CP_NOP, .access = PL1_W,
24
+ .accessfn = aa64_cacheop_poc_access },
25
+ { .name = "DC_IGSW", .state = ARM_CP_STATE_AA64,
26
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 4,
27
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
28
+ { .name = "DC_IGDVAC", .state = ARM_CP_STATE_AA64,
29
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 5,
30
+ .type = ARM_CP_NOP, .access = PL1_W,
31
+ .accessfn = aa64_cacheop_poc_access },
32
+ { .name = "DC_IGDSW", .state = ARM_CP_STATE_AA64,
33
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 6,
34
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
35
+ { .name = "DC_CGSW", .state = ARM_CP_STATE_AA64,
36
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 4,
37
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
38
+ { .name = "DC_CGDSW", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 6,
40
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
41
+ { .name = "DC_CIGSW", .state = ARM_CP_STATE_AA64,
42
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 4,
43
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
44
+ { .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
45
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
46
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
47
REGINFO_SENTINEL
48
};
49
50
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
51
.type = ARM_CP_CONST, .access = PL0_RW, },
52
REGINFO_SENTINEL
53
};
54
+
55
+static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
56
+ { .name = "DC_CGVAC", .state = ARM_CP_STATE_AA64,
57
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 3,
58
+ .type = ARM_CP_NOP, .access = PL0_W,
59
+ .accessfn = aa64_cacheop_poc_access },
60
+ { .name = "DC_CGDVAC", .state = ARM_CP_STATE_AA64,
61
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 5,
62
+ .type = ARM_CP_NOP, .access = PL0_W,
63
+ .accessfn = aa64_cacheop_poc_access },
64
+ { .name = "DC_CGVAP", .state = ARM_CP_STATE_AA64,
65
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 3,
66
+ .type = ARM_CP_NOP, .access = PL0_W,
67
+ .accessfn = aa64_cacheop_poc_access },
68
+ { .name = "DC_CGDVAP", .state = ARM_CP_STATE_AA64,
69
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 5,
70
+ .type = ARM_CP_NOP, .access = PL0_W,
71
+ .accessfn = aa64_cacheop_poc_access },
72
+ { .name = "DC_CGVADP", .state = ARM_CP_STATE_AA64,
73
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 3,
74
+ .type = ARM_CP_NOP, .access = PL0_W,
75
+ .accessfn = aa64_cacheop_poc_access },
76
+ { .name = "DC_CGDVADP", .state = ARM_CP_STATE_AA64,
77
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 5,
78
+ .type = ARM_CP_NOP, .access = PL0_W,
79
+ .accessfn = aa64_cacheop_poc_access },
80
+ { .name = "DC_CIGVAC", .state = ARM_CP_STATE_AA64,
81
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 3,
82
+ .type = ARM_CP_NOP, .access = PL0_W,
83
+ .accessfn = aa64_cacheop_poc_access },
84
+ { .name = "DC_CIGDVAC", .state = ARM_CP_STATE_AA64,
85
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
86
+ .type = ARM_CP_NOP, .access = PL0_W,
87
+ .accessfn = aa64_cacheop_poc_access },
88
+ REGINFO_SENTINEL
89
+};
90
+
91
#endif
92
93
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
94
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
95
*/
18
*/
96
if (cpu_isar_feature(aa64_mte, cpu)) {
19
uint32_t key;
97
define_arm_cp_regs(cpu, mte_reginfo);
20
ARMCPRegInfo *r2;
98
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
21
- int is64 = (r->type & ARM_CP_64BIT) ? 1 : 0;
99
} else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
22
- int ns = (secstate & ARM_CP_SECSTATE_NS) ? 1 : 0;
100
define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
23
+ bool is64 = r->type & ARM_CP_64BIT;
101
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
24
+ bool ns = secstate & ARM_CP_SECSTATE_NS;
102
}
25
int cp = r->cp;
103
#endif
26
size_t name_len;
104
27
105
--
28
--
106
2.20.1
29
2.25.1
107
108
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Computing isbanked only once makes the code
4
a bit easier to read.
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220501055028.646596-17-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-5-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
target/arm/helper.c | 14 +++++++++++---
11
target/arm/helper.c | 6 ++++--
9
1 file changed, 11 insertions(+), 3 deletions(-)
12
1 file changed, 4 insertions(+), 2 deletions(-)
10
13
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
14
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
15
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
18
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
16
if (cpu_isar_feature(aa64_pauth, cpu)) {
19
bool is64 = r->type & ARM_CP_64BIT;
17
valid_mask |= SCR_API | SCR_APK;
20
bool ns = secstate & ARM_CP_SECSTATE_NS;
18
}
21
int cp = r->cp;
19
+ if (cpu_isar_feature(aa64_mte, cpu)) {
22
+ bool isbanked;
20
+ valid_mask |= SCR_ATA;
23
size_t name_len;
21
+ }
24
22
} else {
25
switch (state) {
23
valid_mask &= ~(SCR_RW | SCR_ST);
26
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
27
r2->opaque = opaque;
24
}
28
}
25
@@ -XXX,XX +XXX,XX @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
29
26
if (cpu_isar_feature(aa64_pauth, cpu)) {
30
- if (r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1]) {
27
valid_mask |= HCR_API | HCR_APK;
31
+ isbanked = r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1];
28
}
32
+ if (isbanked) {
29
+ if (cpu_isar_feature(aa64_mte, cpu)) {
33
/* Register is banked (using both entries in array).
30
+ valid_mask |= HCR_ATA | HCR_DCT | HCR_TID5;
34
* Overwriting fieldoffset as the array is only used to define
31
+ }
35
* banked registers but later only fieldoffset is used.
36
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
32
}
37
}
33
38
34
/* Clear RES0 bits. */
39
if (state == ARM_CP_STATE_AA32) {
35
value &= valid_mask;
40
- if (r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1]) {
36
41
+ if (isbanked) {
37
- /* These bits change the MMU setup:
42
/* If the register is banked then we don't need to migrate or
38
+ /*
43
* reset the 32-bit instance in certain cases:
39
+ * These bits change the MMU setup:
44
*
40
* HCR_VM enables stage 2 translation
41
* HCR_PTW forbids certain page-table setups
42
- * HCR_DC Disables stage1 and enables stage2 translation
43
+ * HCR_DC disables stage1 and enables stage2 translation
44
+ * HCR_DCT enables tagging on (disabled) stage1 translation
45
*/
46
- if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC)) {
47
+ if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC | HCR_DCT)) {
48
tlb_flush(CPU(cpu));
49
}
50
env->cp15.hcr_el2 = value;
51
--
45
--
52
2.20.1
46
2.25.1
53
54
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This does not attempt to rectify all of the res0 bits, but does
3
Perform the override check early, so that it is still done
4
clear the mte bits when not enabled. Since there is no high-part
4
even when we decide to discard an unreachable cpreg.
5
mapping of SCTLR, aa32 mode cannot write to these bits.
6
5
6
Use assert not printf+abort.
7
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20220501055028.646596-18-richard.henderson@linaro.org
9
Message-id: 20200626033144.790098-4-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
---
12
target/arm/helper.c | 23 +++++++++++++++++------
13
target/arm/helper.c | 22 ++++++++--------------
13
1 file changed, 17 insertions(+), 6 deletions(-)
14
1 file changed, 8 insertions(+), 14 deletions(-)
14
15
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
18
--- a/target/arm/helper.c
18
+++ b/target/arm/helper.c
19
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
20
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
20
{
21
g_assert_not_reached();
21
ARMCPU *cpu = env_archcpu(env);
22
}
22
23
23
+ if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
24
+ /* Overriding of an existing definition must be explicitly requested. */
24
+ /* M bit is RAZ/WI for PMSA with no MPU implemented */
25
+ if (!(r->type & ARM_CP_OVERRIDE)) {
25
+ value &= ~SCTLR_M;
26
+ const ARMCPRegInfo *oldreg = get_arm_cp_reginfo(cpu->cp_regs, key);
26
+ }
27
+ if (oldreg) {
27
+
28
+ assert(oldreg->type & ARM_CP_OVERRIDE);
28
+ /* ??? Lots of these bits are not implemented. */
29
+
30
+ if (ri->state == ARM_CP_STATE_AA64 && !cpu_isar_feature(aa64_mte, cpu)) {
31
+ if (ri->opc1 == 6) { /* SCTLR_EL3 */
32
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF | SCTLR_ATA);
33
+ } else {
34
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF0 | SCTLR_TCF |
35
+ SCTLR_ATA0 | SCTLR_ATA);
36
+ }
29
+ }
37
+ }
30
+ }
38
+
31
+
39
if (raw_read(env, ri) == value) {
32
/* Combine cpreg and name into one allocation. */
40
/* Skip the TLB flush if nothing actually changed; Linux likes
33
name_len = strlen(name) + 1;
41
* to do a lot of pointless SCTLR writes.
34
r2 = g_malloc(sizeof(*r2) + name_len);
42
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
35
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
43
return;
36
assert(!raw_accessors_invalid(r2));
44
}
37
}
45
38
46
- if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
39
- /* Overriding of an existing definition must be explicitly
47
- /* M bit is RAZ/WI for PMSA with no MPU implemented */
40
- * requested.
48
- value &= ~SCTLR_M;
41
- */
42
- if (!(r->type & ARM_CP_OVERRIDE)) {
43
- const ARMCPRegInfo *oldreg = get_arm_cp_reginfo(cpu->cp_regs, key);
44
- if (oldreg && !(oldreg->type & ARM_CP_OVERRIDE)) {
45
- fprintf(stderr, "Register redefined: cp=%d %d bit "
46
- "crn=%d crm=%d opc1=%d opc2=%d, "
47
- "was %s, now %s\n", r2->cp, 32 + 32 * is64,
48
- r2->crn, r2->crm, r2->opc1, r2->opc2,
49
- oldreg->name, r2->name);
50
- g_assert_not_reached();
51
- }
49
- }
52
- }
50
-
53
g_hash_table_insert(cpu->cp_regs, (gpointer)(uintptr_t)key, r2);
51
raw_write(env, ri, value);
54
}
52
- /* ??? Lots of these bits are not implemented. */
53
+
54
/* This may enable/disable the MMU, so do a TLB flush. */
55
tlb_flush(CPU(cpu));
56
55
57
--
56
--
58
2.20.1
57
2.25.1
59
60
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Cache the composite ATA setting.
3
Put the block comments into the current coding style.
4
4
5
Cache when MTE is fully enabled, i.e. access to tags are enabled
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
and tag checks affect the PE. Do this for both the normal context
7
and the UNPRIV context.
8
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20220501055028.646596-19-richard.henderson@linaro.org
11
Message-id: 20200626033144.790098-9-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
9
---
14
target/arm/cpu.h | 12 ++++++++----
10
target/arm/helper.c | 24 +++++++++++++++---------
15
target/arm/internals.h | 18 +++++++++++++++++
11
1 file changed, 15 insertions(+), 9 deletions(-)
16
target/arm/translate.h | 5 +++++
17
target/arm/helper.c | 40 ++++++++++++++++++++++++++++++++++++++
18
target/arm/translate-a64.c | 4 ++++
19
5 files changed, 75 insertions(+), 4 deletions(-)
20
12
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
22
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
25
@@ -XXX,XX +XXX,XX @@ typedef ARMCPU ArchCPU;
26
* | | | TBFLAG_A32 | |
27
* | | +-----+----------+ TBFLAG_AM32 |
28
* | TBFLAG_ANY | |TBFLAG_M32| |
29
- * | | +-+----------+--------------|
30
- * | | | TBFLAG_A64 |
31
- * +--------------+---------+---------------------------+
32
- * 31 20 15 0
33
+ * | +-----------+----------+--------------|
34
+ * | | TBFLAG_A64 |
35
+ * +--------------+-------------------------------------+
36
+ * 31 20 0
37
*
38
* Unless otherwise noted, these bits are cached in env->hflags.
39
*/
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, BT, 9, 1)
41
FIELD(TBFLAG_A64, BTYPE, 10, 2) /* Not cached. */
42
FIELD(TBFLAG_A64, TBID, 12, 2)
43
FIELD(TBFLAG_A64, UNPRIV, 14, 1)
44
+FIELD(TBFLAG_A64, ATA, 15, 1)
45
+FIELD(TBFLAG_A64, TCMA, 16, 2)
46
+FIELD(TBFLAG_A64, MTE_ACTIVE, 18, 1)
47
+FIELD(TBFLAG_A64, MTE0_ACTIVE, 19, 1)
48
49
/**
50
* cpu_mmu_index:
51
diff --git a/target/arm/internals.h b/target/arm/internals.h
52
index XXXXXXX..XXXXXXX 100644
53
--- a/target/arm/internals.h
54
+++ b/target/arm/internals.h
55
@@ -XXX,XX +XXX,XX @@ static inline int exception_target_el(CPUARMState *env)
56
return target_el;
57
}
58
59
+/* Determine if allocation tags are available. */
60
+static inline bool allocation_tag_access_enabled(CPUARMState *env, int el,
61
+ uint64_t sctlr)
62
+{
63
+ if (el < 3
64
+ && arm_feature(env, ARM_FEATURE_EL3)
65
+ && !(env->cp15.scr_el3 & SCR_ATA)) {
66
+ return false;
67
+ }
68
+ if (el < 2
69
+ && arm_feature(env, ARM_FEATURE_EL2)
70
+ && !(arm_hcr_el2_eff(env) & HCR_ATA)) {
71
+ return false;
72
+ }
73
+ sctlr &= (el == 0 ? SCTLR_ATA0 : SCTLR_ATA);
74
+ return sctlr != 0;
75
+}
76
+
77
#ifndef CONFIG_USER_ONLY
78
79
/* Security attributes for an address, as returned by v8m_security_lookup. */
80
diff --git a/target/arm/translate.h b/target/arm/translate.h
81
index XXXXXXX..XXXXXXX 100644
82
--- a/target/arm/translate.h
83
+++ b/target/arm/translate.h
84
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
85
ARMMMUIdx mmu_idx; /* MMU index to use for normal loads/stores */
86
uint8_t tbii; /* TBI1|TBI0 for insns */
87
uint8_t tbid; /* TBI1|TBI0 for data */
88
+ uint8_t tcma; /* TCMA1|TCMA0 for MTE */
89
bool ns; /* Use non-secure CPREG bank on access */
90
int fp_excp_el; /* FP exception EL or 0 if enabled */
91
int sve_excp_el; /* SVE exception EL or 0 if enabled */
92
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
93
bool unpriv;
94
/* True if v8.3-PAuth is active. */
95
bool pauth_active;
96
+ /* True if v8.5-MTE access to tags is enabled. */
97
+ bool ata;
98
+ /* True if v8.5-MTE tag checks affect the PE; index with is_unpriv. */
99
+ bool mte_active[2];
100
/* True with v8.5-BTI and SCTLR_ELx.BT* set. */
101
bool bt;
102
/* True if any CP15 access is trapped by HSTR_EL2 */
103
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
104
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
106
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
107
@@ -XXX,XX +XXX,XX @@ static int aa64_va_parameter_tbid(uint64_t tcr, ARMMMUIdx mmu_idx)
17
@@ -XXX,XX +XXX,XX @@ CpuDefinitionInfoList *qmp_query_cpu_definitions(Error **errp)
108
}
18
return cpu_list;
109
}
19
}
110
20
111
+static int aa64_va_parameter_tcma(uint64_t tcr, ARMMMUIdx mmu_idx)
21
+/*
112
+{
22
+ * Private utility function for define_one_arm_cp_reg_with_opaque():
113
+ if (regime_has_2_ranges(mmu_idx)) {
23
+ * add a single reginfo struct to the hash table.
114
+ return extract64(tcr, 57, 2);
24
+ */
115
+ } else {
25
static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
116
+ /* Replicate the single TCMA bit so we always have 2 bits. */
26
void *opaque, CPState state,
117
+ return extract32(tcr, 30, 1) * 3;
27
CPSecureState secstate,
118
+ }
28
int crm, int opc1, int opc2,
119
+}
29
const char *name)
120
+
121
ARMVAParameters aa64_va_parameters(CPUARMState *env, uint64_t va,
122
ARMMMUIdx mmu_idx, bool data)
123
{
30
{
124
@@ -XXX,XX +XXX,XX @@ static uint32_t rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
31
- /* Private utility function for define_one_arm_cp_reg_with_opaque():
32
- * add a single reginfo struct to the hash table.
33
- */
34
uint32_t key;
35
ARMCPRegInfo *r2;
36
bool is64 = r->type & ARM_CP_64BIT;
37
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
38
39
isbanked = r->bank_fieldoffsets[0] && r->bank_fieldoffsets[1];
40
if (isbanked) {
41
- /* Register is banked (using both entries in array).
42
+ /*
43
+ * Register is banked (using both entries in array).
44
* Overwriting fieldoffset as the array is only used to define
45
* banked registers but later only fieldoffset is used.
46
*/
47
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
48
49
if (state == ARM_CP_STATE_AA32) {
50
if (isbanked) {
51
- /* If the register is banked then we don't need to migrate or
52
+ /*
53
+ * If the register is banked then we don't need to migrate or
54
* reset the 32-bit instance in certain cases:
55
*
56
* 1) If the register has both 32-bit and 64-bit instances then we
57
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
58
r2->type |= ARM_CP_ALIAS;
59
}
60
} else if ((secstate != r->secure) && !ns) {
61
- /* The register is not banked so we only want to allow migration of
62
- * the non-secure instance.
63
+ /*
64
+ * The register is not banked so we only want to allow migration
65
+ * of the non-secure instance.
66
*/
67
r2->type |= ARM_CP_ALIAS;
68
}
69
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
125
}
70
}
126
}
71
}
127
72
128
+ if (cpu_isar_feature(aa64_mte, env_archcpu(env))) {
73
- /* By convention, for wildcarded registers only the first
129
+ /*
74
+ /*
130
+ * Set MTE_ACTIVE if any access may be Checked, and leave clear
75
+ * By convention, for wildcarded registers only the first
131
+ * if all accesses must be Unchecked:
76
* entry is used for migration; the others are marked as
132
+ * 1) If no TBI, then there are no tags in the address to check,
77
* ALIAS so we don't try to transfer the register
133
+ * 2) If Tag Check Override, then all accesses are Unchecked,
78
* multiple times. Special registers (ie NOP/WFI) are
134
+ * 3) If Tag Check Fail == 0, then Checked access have no effect,
79
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
135
+ * 4) If no Allocation Tag Access, then all accesses are Unchecked.
80
r2->type |= ARM_CP_ALIAS | ARM_CP_NO_GDB;
136
+ */
81
}
137
+ if (allocation_tag_access_enabled(env, el, sctlr)) {
82
138
+ flags = FIELD_DP32(flags, TBFLAG_A64, ATA, 1);
83
- /* Check that raw accesses are either forbidden or handled. Note that
139
+ if (tbid
84
+ /*
140
+ && !(env->pstate & PSTATE_TCO)
85
+ * Check that raw accesses are either forbidden or handled. Note that
141
+ && (sctlr & (el == 0 ? SCTLR_TCF0 : SCTLR_TCF))) {
86
* we can't assert this earlier because the setup of fieldoffset for
142
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE_ACTIVE, 1);
87
* banked registers has to be done first.
143
+ }
88
*/
144
+ }
145
+ /* And again for unprivileged accesses, if required. */
146
+ if (FIELD_EX32(flags, TBFLAG_A64, UNPRIV)
147
+ && tbid
148
+ && !(env->pstate & PSTATE_TCO)
149
+ && (sctlr & SCTLR_TCF0)
150
+ && allocation_tag_access_enabled(env, 0, sctlr)) {
151
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE0_ACTIVE, 1);
152
+ }
153
+ /* Cache TCMA as well as TBI. */
154
+ flags = FIELD_DP32(flags, TBFLAG_A64, TCMA,
155
+ aa64_va_parameter_tcma(tcr, mmu_idx));
156
+ }
157
+
158
return rebuild_hflags_common(env, fp_el, mmu_idx, flags);
159
}
160
161
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
162
index XXXXXXX..XXXXXXX 100644
163
--- a/target/arm/translate-a64.c
164
+++ b/target/arm/translate-a64.c
165
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
166
dc->mmu_idx = core_to_aa64_mmu_idx(core_mmu_idx);
167
dc->tbii = FIELD_EX32(tb_flags, TBFLAG_A64, TBII);
168
dc->tbid = FIELD_EX32(tb_flags, TBFLAG_A64, TBID);
169
+ dc->tcma = FIELD_EX32(tb_flags, TBFLAG_A64, TCMA);
170
dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
171
#if !defined(CONFIG_USER_ONLY)
172
dc->user = (dc->current_el == 0);
173
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
174
dc->bt = FIELD_EX32(tb_flags, TBFLAG_A64, BT);
175
dc->btype = FIELD_EX32(tb_flags, TBFLAG_A64, BTYPE);
176
dc->unpriv = FIELD_EX32(tb_flags, TBFLAG_A64, UNPRIV);
177
+ dc->ata = FIELD_EX32(tb_flags, TBFLAG_A64, ATA);
178
+ dc->mte_active[0] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE_ACTIVE);
179
+ dc->mte_active[1] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE0_ACTIVE);
180
dc->vec_len = 0;
181
dc->vec_stride = 0;
182
dc->cp_regs = arm_cpu->cp_regs;
183
--
89
--
184
2.20.1
90
2.25.1
185
186
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This "bit" is a particular value of the page's MemAttr.
3
Since e03b56863d2bc, our host endian indicator is unconditionally
4
set, which means that we can use a normal C condition.
4
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220501055028.646596-20-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-43-richard.henderson@linaro.org
9
[PMM: quote correct git hash in commit message]
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/helper.c | 48 ++++++++++++++++++++++++++++++++++++++---
12
target/arm/helper.c | 9 +++------
11
target/arm/tlb_helper.c | 5 +++++
13
1 file changed, 3 insertions(+), 6 deletions(-)
12
2 files changed, 50 insertions(+), 3 deletions(-)
13
14
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
17
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
18
@@ -XXX,XX +XXX,XX @@ static uint8_t combine_cacheattr_nibble(uint8_t s1, uint8_t s2)
19
@@ -XXX,XX +XXX,XX @@ static void add_cpreg_to_hashtable(ARMCPU *cpu, const ARMCPRegInfo *r,
19
*/
20
r2->type |= ARM_CP_ALIAS;
20
static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
21
}
21
{
22
22
- uint8_t s1lo = extract32(s1.attrs, 0, 4), s2lo = extract32(s2.attrs, 0, 4);
23
- if (r->state == ARM_CP_STATE_BOTH) {
23
- uint8_t s1hi = extract32(s1.attrs, 4, 4), s2hi = extract32(s2.attrs, 4, 4);
24
-#if HOST_BIG_ENDIAN
24
+ uint8_t s1lo, s2lo, s1hi, s2hi;
25
- if (r2->fieldoffset) {
25
ARMCacheAttrs ret;
26
- r2->fieldoffset += sizeof(uint32_t);
26
+ bool tagged = false;
27
- }
27
+
28
-#endif
28
+ if (s1.attrs == 0xf0) {
29
+ if (HOST_BIG_ENDIAN &&
29
+ tagged = true;
30
+ r->state == ARM_CP_STATE_BOTH && r2->fieldoffset) {
30
+ s1.attrs = 0xff;
31
+ r2->fieldoffset += sizeof(uint32_t);
31
+ }
32
+
33
+ s1lo = extract32(s1.attrs, 0, 4);
34
+ s2lo = extract32(s2.attrs, 0, 4);
35
+ s1hi = extract32(s1.attrs, 4, 4);
36
+ s2hi = extract32(s2.attrs, 4, 4);
37
38
/* Combine shareability attributes (table D4-43) */
39
if (s1.shareability == 2 || s2.shareability == 2) {
40
@@ -XXX,XX +XXX,XX @@ static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
41
}
32
}
42
}
33
}
43
34
44
+ /* TODO: CombineS1S2Desc does not consider transient, only WB, RWA. */
45
+ if (tagged && ret.attrs == 0xff) {
46
+ ret.attrs = 0xf0;
47
+ }
48
+
49
return ret;
50
}
51
52
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
53
* Normal Non-Shareable,
54
* Inner Write-Back Read-Allocate Write-Allocate,
55
* Outer Write-Back Read-Allocate Write-Allocate.
56
+ * Do not overwrite Tagged within attrs.
57
*/
58
- cacheattrs->attrs = 0xff;
59
+ if (cacheattrs->attrs != 0xf0) {
60
+ cacheattrs->attrs = 0xff;
61
+ }
62
cacheattrs->shareability = 0;
63
}
64
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
65
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
66
/* Definitely a real MMU, not an MPU */
67
68
if (regime_translation_disabled(env, mmu_idx)) {
69
+ uint64_t hcr;
70
+ uint8_t memattr;
71
+
72
/*
73
* MMU disabled. S1 addresses within aa64 translation regimes are
74
* still checked for bounds -- see AArch64.TranslateAddressS1Off.
75
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
76
*phys_ptr = address;
77
*prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
78
*page_size = TARGET_PAGE_SIZE;
79
+
80
+ /* Fill in cacheattr a-la AArch64.TranslateAddressS1Off. */
81
+ hcr = arm_hcr_el2_eff(env);
82
+ cacheattrs->shareability = 0;
83
+ if (hcr & HCR_DC) {
84
+ if (hcr & HCR_DCT) {
85
+ memattr = 0xf0; /* Tagged, Normal, WB, RWA */
86
+ } else {
87
+ memattr = 0xff; /* Normal, WB, RWA */
88
+ }
89
+ } else if (access_type == MMU_INST_FETCH) {
90
+ if (regime_sctlr(env, mmu_idx) & SCTLR_I) {
91
+ memattr = 0xee; /* Normal, WT, RA, NT */
92
+ } else {
93
+ memattr = 0x44; /* Normal, NC, No */
94
+ }
95
+ cacheattrs->shareability = 2; /* outer sharable */
96
+ } else {
97
+ memattr = 0x00; /* Device, nGnRnE */
98
+ }
99
+ cacheattrs->attrs = memattr;
100
return 0;
101
}
102
103
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
104
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/tlb_helper.c
106
+++ b/target/arm/tlb_helper.c
107
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
108
phys_addr &= TARGET_PAGE_MASK;
109
address &= TARGET_PAGE_MASK;
110
}
111
+ /* Notice and record tagged memory. */
112
+ if (cpu_isar_feature(aa64_mte, cpu) && cacheattrs.attrs == 0xf0) {
113
+ arm_tlb_mte_tagged(&attrs) = true;
114
+ }
115
+
116
tlb_set_page_with_attrs(cs, address, phys_addr, attrs,
117
prot, mmu_idx, page_size);
118
return true;
119
--
35
--
120
2.20.1
36
2.25.1
121
122
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
2
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-34-richard.henderson@linaro.org
5
Message-id: 20220501055028.646596-24-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
7
---
11
target/arm/cpu.h | 1 +
8
target/arm/cpu.h | 15 +++++++++++++++
12
target/arm/helper-sve.h | 58 ++++++++++
9
1 file changed, 15 insertions(+)
13
target/arm/internals.h | 6 +
14
target/arm/sve_helper.c | 218 ++++++++++++++++++++++++++++++-------
15
target/arm/translate-sve.c | 186 ++++++++++++++++++++++---------
16
5 files changed, 378 insertions(+), 91 deletions(-)
17
10
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
19
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.h
13
--- a/target/arm/cpu.h
21
+++ b/target/arm/cpu.h
14
+++ b/target/arm/cpu.h
22
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa32_ssbs(const ARMISARegisters *id)
23
* generic target bits directly.
16
return FIELD_EX32(id->id_pfr2, ID_PFR2, SSBS) != 0;
24
*/
25
#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
26
+#define arm_tlb_mte_tagged(x) (typecheck_memtxattrs(x)->target_tlb_bit1)
27
28
/*
29
* Naming convention for isar_feature functions:
30
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper-sve.h
33
+++ b/target/arm/helper-sve.h
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ld1sds_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
DEF_HELPER_FLAGS_4(sve_ld1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
38
+DEF_HELPER_FLAGS_4(sve_ld1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_4(sve_ld2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ld3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_ld4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ld1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_4(sve_ld2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ld3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_ld4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+
48
+DEF_HELPER_FLAGS_4(sve_ld1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ld2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_ld3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ld4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+
53
+DEF_HELPER_FLAGS_4(sve_ld1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_4(sve_ld2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_ld3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ld4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+
58
+DEF_HELPER_FLAGS_4(sve_ld1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
59
+DEF_HELPER_FLAGS_4(sve_ld2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_ld3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_ld4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_ld1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_ld2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ld3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_ld4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ld1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_ld2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ld3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_4(sve_ld4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_4(sve_ld1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_4(sve_ld1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_4(sve_ld1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_4(sve_ld1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_4(sve_ld1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_4(sve_ld1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
+
80
+DEF_HELPER_FLAGS_4(sve_ld1hsu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ld1hdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ld1hss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ld1hds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+
85
+DEF_HELPER_FLAGS_4(sve_ld1hsu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ld1hdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_4(sve_ld1hss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_4(sve_ld1hds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
89
+
90
+DEF_HELPER_FLAGS_4(sve_ld1sdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
91
+DEF_HELPER_FLAGS_4(sve_ld1sds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
92
+
93
+DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ld1sds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
95
+
96
DEF_HELPER_FLAGS_4(sve_ldff1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
97
DEF_HELPER_FLAGS_4(sve_ldff1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
98
DEF_HELPER_FLAGS_4(sve_ldff1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
99
diff --git a/target/arm/internals.h b/target/arm/internals.h
100
index XXXXXXX..XXXXXXX 100644
101
--- a/target/arm/internals.h
102
+++ b/target/arm/internals.h
103
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
104
#define LOG2_TAG_GRANULE 4
105
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
106
107
+/*
108
+ * The SVE simd_data field, for memory ops, contains either
109
+ * rd (5 bits) or a shift count (2 bits).
110
+ */
111
+#define SVE_MTEDESC_SHIFT 5
112
+
113
/* Bits within a descriptor passed to the helper_mte_check* functions. */
114
FIELD(MTEDESC, MIDX, 0, 4)
115
FIELD(MTEDESC, TBI, 4, 2)
116
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
117
index XXXXXXX..XXXXXXX 100644
118
--- a/target/arm/sve_helper.c
119
+++ b/target/arm/sve_helper.c
120
@@ -XXX,XX +XXX,XX @@ static void sve_cont_ldst_watchpoints(SVEContLdSt *info, CPUARMState *env,
121
#endif
122
}
17
}
123
18
124
+typedef uint64_t mte_check_fn(CPUARMState *, uint32_t, uint64_t, uintptr_t);
19
+static inline bool isar_feature_aa32_debugv8p2(const ARMISARegisters *id)
125
+
126
+static inline QEMU_ALWAYS_INLINE
127
+void sve_cont_ldst_mte_check_int(SVEContLdSt *info, CPUARMState *env,
128
+ uint64_t *vg, target_ulong addr, int esize,
129
+ int msize, uint32_t mtedesc, uintptr_t ra,
130
+ mte_check_fn *check)
131
+{
20
+{
132
+ intptr_t mem_off, reg_off, reg_last;
21
+ return FIELD_EX32(id->id_dfr0, ID_DFR0, COPDBG) >= 8;
133
+
134
+ /* Process the page only if MemAttr == Tagged. */
135
+ if (arm_tlb_mte_tagged(&info->page[0].attrs)) {
136
+ mem_off = info->mem_off_first[0];
137
+ reg_off = info->reg_off_first[0];
138
+ reg_last = info->reg_off_split;
139
+ if (reg_last < 0) {
140
+ reg_last = info->reg_off_last[0];
141
+ }
142
+
143
+ do {
144
+ uint64_t pg = vg[reg_off >> 6];
145
+ do {
146
+ if ((pg >> (reg_off & 63)) & 1) {
147
+ check(env, mtedesc, addr, ra);
148
+ }
149
+ reg_off += esize;
150
+ mem_off += msize;
151
+ } while (reg_off <= reg_last && (reg_off & 63));
152
+ } while (reg_off <= reg_last);
153
+ }
154
+
155
+ mem_off = info->mem_off_first[1];
156
+ if (mem_off >= 0 && arm_tlb_mte_tagged(&info->page[1].attrs)) {
157
+ reg_off = info->reg_off_first[1];
158
+ reg_last = info->reg_off_last[1];
159
+
160
+ do {
161
+ uint64_t pg = vg[reg_off >> 6];
162
+ do {
163
+ if ((pg >> (reg_off & 63)) & 1) {
164
+ check(env, mtedesc, addr, ra);
165
+ }
166
+ reg_off += esize;
167
+ mem_off += msize;
168
+ } while (reg_off & 63);
169
+ } while (reg_off <= reg_last);
170
+ }
171
+}
22
+}
172
+
23
+
173
+typedef void sve_cont_ldst_mte_check_fn(SVEContLdSt *info, CPUARMState *env,
24
/*
174
+ uint64_t *vg, target_ulong addr,
25
* 64-bit feature tests via id registers.
175
+ int esize, int msize, uint32_t mtedesc,
26
*/
176
+ uintptr_t ra);
27
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_ssbs(const ARMISARegisters *id)
177
+
28
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, SSBS) != 0;
178
+static void sve_cont_ldst_mte_check1(SVEContLdSt *info, CPUARMState *env,
29
}
179
+ uint64_t *vg, target_ulong addr,
30
180
+ int esize, int msize, uint32_t mtedesc,
31
+static inline bool isar_feature_aa64_debugv8p2(const ARMISARegisters *id)
181
+ uintptr_t ra)
182
+{
32
+{
183
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
33
+ return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, DEBUGVER) >= 8;
184
+ mtedesc, ra, mte_check1);
185
+}
34
+}
186
+
35
+
187
+static void sve_cont_ldst_mte_checkN(SVEContLdSt *info, CPUARMState *env,
36
static inline bool isar_feature_aa64_sve2(const ARMISARegisters *id)
188
+ uint64_t *vg, target_ulong addr,
37
{
189
+ int esize, int msize, uint32_t mtedesc,
38
return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SVEVER) != 0;
190
+ uintptr_t ra)
39
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_any_tts2uxn(const ARMISARegisters *id)
40
return isar_feature_aa64_tts2uxn(id) || isar_feature_aa32_tts2uxn(id);
41
}
42
43
+static inline bool isar_feature_any_debugv8p2(const ARMISARegisters *id)
191
+{
44
+{
192
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
45
+ return isar_feature_aa64_debugv8p2(id) || isar_feature_aa32_debugv8p2(id);
193
+ mtedesc, ra, mte_checkN);
194
+}
46
+}
195
+
47
+
196
+
197
/*
48
/*
198
* Common helper for all contiguous 1,2,3,4-register predicated stores.
49
* Forward to the above feature tests given an ARMCPU pointer.
199
*/
50
*/
200
static inline QEMU_ALWAYS_INLINE
201
void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
202
uint32_t desc, const uintptr_t retaddr,
203
- const int esz, const int msz, const int N,
204
+ const int esz, const int msz, const int N, uint32_t mtedesc,
205
sve_ldst1_host_fn *host_fn,
206
- sve_ldst1_tlb_fn *tlb_fn)
207
+ sve_ldst1_tlb_fn *tlb_fn,
208
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
209
{
210
const unsigned rd = simd_data(desc);
211
const intptr_t reg_max = simd_oprsz(desc);
212
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
213
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
214
BP_MEM_READ, retaddr);
215
216
- /* TODO: MTE check. */
217
+ /*
218
+ * Handle mte checks for all active elements.
219
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
220
+ */
221
+ if (mte_check_fn && mtedesc) {
222
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
223
+ mtedesc, retaddr);
224
+ }
225
226
flags = info.page[0].flags | info.page[1].flags;
227
if (unlikely(flags != 0)) {
228
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
229
}
230
}
231
232
-#define DO_LD1_1(NAME, ESZ) \
233
-void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
234
- target_ulong addr, uint32_t desc) \
235
-{ \
236
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
237
- sve_##NAME##_host, sve_##NAME##_tlb); \
238
+static inline QEMU_ALWAYS_INLINE
239
+void sve_ldN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
240
+ uint32_t desc, const uintptr_t ra,
241
+ const int esz, const int msz, const int N,
242
+ sve_ldst1_host_fn *host_fn,
243
+ sve_ldst1_tlb_fn *tlb_fn)
244
+{
245
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
246
+ int bit55 = extract64(addr, 55, 1);
247
+
248
+ /* Remove mtedesc from the normal sve descriptor. */
249
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
250
+
251
+ /* Perform gross MTE suppression early. */
252
+ if (!tbi_check(desc, bit55) ||
253
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
254
+ mtedesc = 0;
255
+ }
256
+
257
+ sve_ldN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
258
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
259
}
260
261
-#define DO_LD1_2(NAME, ESZ, MSZ) \
262
-void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
263
- target_ulong addr, uint32_t desc) \
264
-{ \
265
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
266
- sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
267
-} \
268
-void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
269
- target_ulong addr, uint32_t desc) \
270
-{ \
271
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
272
- sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
273
+#define DO_LD1_1(NAME, ESZ) \
274
+void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
275
+ target_ulong addr, uint32_t desc) \
276
+{ \
277
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, 0, \
278
+ sve_##NAME##_host, sve_##NAME##_tlb, NULL); \
279
+} \
280
+void HELPER(sve_##NAME##_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
284
+ sve_##NAME##_host, sve_##NAME##_tlb); \
285
+}
286
+
287
+#define DO_LD1_2(NAME, ESZ, MSZ) \
288
+void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
289
+ target_ulong addr, uint32_t desc) \
290
+{ \
291
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
292
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb, NULL); \
293
+} \
294
+void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
295
+ target_ulong addr, uint32_t desc) \
296
+{ \
297
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
298
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb, NULL); \
299
+} \
300
+void HELPER(sve_##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
301
+ target_ulong addr, uint32_t desc) \
302
+{ \
303
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
304
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
305
+} \
306
+void HELPER(sve_##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
307
+ target_ulong addr, uint32_t desc) \
308
+{ \
309
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
310
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
311
}
312
313
DO_LD1_1(ld1bb, MO_8)
314
@@ -XXX,XX +XXX,XX @@ DO_LD1_2(ld1dd, MO_64, MO_64)
315
#undef DO_LD1_1
316
#undef DO_LD1_2
317
318
-#define DO_LDN_1(N) \
319
-void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
320
- target_ulong addr, uint32_t desc) \
321
-{ \
322
- sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
323
- sve_ld1bb_host, sve_ld1bb_tlb); \
324
+#define DO_LDN_1(N) \
325
+void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
326
+ target_ulong addr, uint32_t desc) \
327
+{ \
328
+ sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, 0, \
329
+ sve_ld1bb_host, sve_ld1bb_tlb, NULL); \
330
+} \
331
+void HELPER(sve_ld##N##bb_r_mte)(CPUARMState *env, void *vg, \
332
+ target_ulong addr, uint32_t desc) \
333
+{ \
334
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
335
+ sve_ld1bb_host, sve_ld1bb_tlb); \
336
}
337
338
-#define DO_LDN_2(N, SUFF, ESZ) \
339
-void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
340
- target_ulong addr, uint32_t desc) \
341
-{ \
342
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
343
- sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
344
-} \
345
-void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
346
- target_ulong addr, uint32_t desc) \
347
-{ \
348
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
349
- sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
350
+#define DO_LDN_2(N, SUFF, ESZ) \
351
+void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
352
+ target_ulong addr, uint32_t desc) \
353
+{ \
354
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
355
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb, NULL); \
356
+} \
357
+void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
358
+ target_ulong addr, uint32_t desc) \
359
+{ \
360
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
361
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb, NULL); \
362
+} \
363
+void HELPER(sve_ld##N##SUFF##_le_r_mte)(CPUARMState *env, void *vg, \
364
+ target_ulong addr, uint32_t desc) \
365
+{ \
366
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
367
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
368
+} \
369
+void HELPER(sve_ld##N##SUFF##_be_r_mte)(CPUARMState *env, void *vg, \
370
+ target_ulong addr, uint32_t desc) \
371
+{ \
372
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
373
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
374
}
375
376
DO_LDN_1(2)
377
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
378
index XXXXXXX..XXXXXXX 100644
379
--- a/target/arm/translate-sve.c
380
+++ b/target/arm/translate-sve.c
381
@@ -XXX,XX +XXX,XX @@ static const uint8_t dtype_esz[16] = {
382
};
383
384
static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
385
- int dtype, gen_helper_gvec_mem *fn)
386
+ int dtype, uint32_t mte_n, bool is_write,
387
+ gen_helper_gvec_mem *fn)
388
{
389
unsigned vsz = vec_full_reg_size(s);
390
TCGv_ptr t_pg;
391
TCGv_i32 t_desc;
392
- int desc;
393
+ int desc = 0;
394
395
- /* For e.g. LD4, there are not enough arguments to pass all 4
396
+ /*
397
+ * For e.g. LD4, there are not enough arguments to pass all 4
398
* registers as pointers, so encode the regno into the data field.
399
* For consistency, do this even for LD1.
400
+ * TODO: mte_n check here while callers are updated.
401
*/
402
- desc = simd_desc(vsz, vsz, zt);
403
+ if (mte_n && s->mte_active[0]) {
404
+ int msz = dtype_msz(dtype);
405
+
406
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
407
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
408
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
409
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
410
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
411
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
412
+ desc <<= SVE_MTEDESC_SHIFT;
413
+ }
414
+ desc = simd_desc(vsz, vsz, zt | desc);
415
t_desc = tcg_const_i32(desc);
416
t_pg = tcg_temp_new_ptr();
417
418
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
419
static void do_ld_zpa(DisasContext *s, int zt, int pg,
420
TCGv_i64 addr, int dtype, int nreg)
421
{
422
- static gen_helper_gvec_mem * const fns[2][16][4] = {
423
- /* Little-endian */
424
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
425
+ static gen_helper_gvec_mem * const fns[2][2][16][4] = {
426
+ { /* mte inactive, little-endian */
427
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
428
gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
429
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
430
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
431
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
432
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
433
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
434
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
435
436
- { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
437
- { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
438
- gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
439
- { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
440
- { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
441
+ { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
442
+ { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
443
+ gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
444
+ { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
445
+ { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
446
447
- { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
448
- { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
449
- { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
450
- gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
451
- { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
452
+ { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
453
+ { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
454
+ { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
455
+ gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
456
+ { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
457
458
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
459
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
460
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
461
- { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
462
- gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
463
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
464
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
465
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
466
+ { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
467
+ gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
468
469
- /* Big-endian */
470
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
471
- gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
472
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
473
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
474
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
475
+ /* mte inactive, big-endian */
476
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
477
+ gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
478
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
479
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
480
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
481
482
- { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
483
- { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
484
- gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
485
- { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
486
- { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
487
+ { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
488
+ { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
489
+ gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
490
+ { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
491
+ { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
492
493
- { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
494
- { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
495
- { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
496
- gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
497
- { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
498
+ { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
499
+ { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
500
+ { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
501
+ gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
502
+ { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
503
504
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
505
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
506
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
507
- { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
508
- gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } }
509
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
510
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
511
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
512
+ { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
513
+ gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } } },
514
+
515
+ { /* mte active, little-endian */
516
+ { { gen_helper_sve_ld1bb_r_mte,
517
+ gen_helper_sve_ld2bb_r_mte,
518
+ gen_helper_sve_ld3bb_r_mte,
519
+ gen_helper_sve_ld4bb_r_mte },
520
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
521
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
522
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
523
+
524
+ { gen_helper_sve_ld1sds_le_r_mte, NULL, NULL, NULL },
525
+ { gen_helper_sve_ld1hh_le_r_mte,
526
+ gen_helper_sve_ld2hh_le_r_mte,
527
+ gen_helper_sve_ld3hh_le_r_mte,
528
+ gen_helper_sve_ld4hh_le_r_mte },
529
+ { gen_helper_sve_ld1hsu_le_r_mte, NULL, NULL, NULL },
530
+ { gen_helper_sve_ld1hdu_le_r_mte, NULL, NULL, NULL },
531
+
532
+ { gen_helper_sve_ld1hds_le_r_mte, NULL, NULL, NULL },
533
+ { gen_helper_sve_ld1hss_le_r_mte, NULL, NULL, NULL },
534
+ { gen_helper_sve_ld1ss_le_r_mte,
535
+ gen_helper_sve_ld2ss_le_r_mte,
536
+ gen_helper_sve_ld3ss_le_r_mte,
537
+ gen_helper_sve_ld4ss_le_r_mte },
538
+ { gen_helper_sve_ld1sdu_le_r_mte, NULL, NULL, NULL },
539
+
540
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
541
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
542
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
543
+ { gen_helper_sve_ld1dd_le_r_mte,
544
+ gen_helper_sve_ld2dd_le_r_mte,
545
+ gen_helper_sve_ld3dd_le_r_mte,
546
+ gen_helper_sve_ld4dd_le_r_mte } },
547
+
548
+ /* mte active, big-endian */
549
+ { { gen_helper_sve_ld1bb_r_mte,
550
+ gen_helper_sve_ld2bb_r_mte,
551
+ gen_helper_sve_ld3bb_r_mte,
552
+ gen_helper_sve_ld4bb_r_mte },
553
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
554
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
555
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
556
+
557
+ { gen_helper_sve_ld1sds_be_r_mte, NULL, NULL, NULL },
558
+ { gen_helper_sve_ld1hh_be_r_mte,
559
+ gen_helper_sve_ld2hh_be_r_mte,
560
+ gen_helper_sve_ld3hh_be_r_mte,
561
+ gen_helper_sve_ld4hh_be_r_mte },
562
+ { gen_helper_sve_ld1hsu_be_r_mte, NULL, NULL, NULL },
563
+ { gen_helper_sve_ld1hdu_be_r_mte, NULL, NULL, NULL },
564
+
565
+ { gen_helper_sve_ld1hds_be_r_mte, NULL, NULL, NULL },
566
+ { gen_helper_sve_ld1hss_be_r_mte, NULL, NULL, NULL },
567
+ { gen_helper_sve_ld1ss_be_r_mte,
568
+ gen_helper_sve_ld2ss_be_r_mte,
569
+ gen_helper_sve_ld3ss_be_r_mte,
570
+ gen_helper_sve_ld4ss_be_r_mte },
571
+ { gen_helper_sve_ld1sdu_be_r_mte, NULL, NULL, NULL },
572
+
573
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
574
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
575
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
576
+ { gen_helper_sve_ld1dd_be_r_mte,
577
+ gen_helper_sve_ld2dd_be_r_mte,
578
+ gen_helper_sve_ld3dd_be_r_mte,
579
+ gen_helper_sve_ld4dd_be_r_mte } } },
580
};
581
- gen_helper_gvec_mem *fn = fns[s->be_data == MO_BE][dtype][nreg];
582
+ gen_helper_gvec_mem *fn
583
+ = fns[s->mte_active[0]][s->be_data == MO_BE][dtype][nreg];
584
585
- /* While there are holes in the table, they are not
586
+ /*
587
+ * While there are holes in the table, they are not
588
* accessible via the instruction encoding.
589
*/
590
assert(fn != NULL);
591
- do_mem_zpa(s, zt, pg, addr, dtype, fn);
592
+ do_mem_zpa(s, zt, pg, addr, dtype, nreg, false, fn);
593
}
594
595
static bool trans_LD_zprr(DisasContext *s, arg_rprr_load *a)
596
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
597
TCGv_i64 addr = new_tmp_a64(s);
598
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
599
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
600
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
601
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
602
fns[s->be_data == MO_BE][a->dtype]);
603
}
604
return true;
605
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
606
TCGv_i64 addr = new_tmp_a64(s);
607
608
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
609
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
610
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
611
fns[s->be_data == MO_BE][a->dtype]);
612
}
613
return true;
614
@@ -XXX,XX +XXX,XX @@ static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
615
fn = fn_multiple[be][nreg - 1][msz];
616
}
617
assert(fn != NULL);
618
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), fn);
619
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
620
}
621
622
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
623
--
51
--
624
2.20.1
52
2.25.1
625
626
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Add the aa64 predicate for detecting RAS support from id registers.
4
We already have the aa32 version from the M-profile work.
5
Add the 'any' predicate for testing both aa64 and aa32.
2
6
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-2-richard.henderson@linaro.org
9
Message-id: 20220501055028.646596-34-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
11
---
8
target/arm/cpu.h | 10 ++++++++++
12
target/arm/cpu.h | 10 ++++++++++
9
1 file changed, 10 insertions(+)
13
1 file changed, 10 insertions(+)
10
14
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
12
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
14
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_bti(const ARMISARegisters *id)
19
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_aa32_el1(const ARMISARegisters *id)
16
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, BT) != 0;
20
return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, EL1) >= 2;
17
}
21
}
18
22
19
+static inline bool isar_feature_aa64_mte_insn_reg(const ARMISARegisters *id)
23
+static inline bool isar_feature_aa64_ras(const ARMISARegisters *id)
20
+{
24
+{
21
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) != 0;
25
+ return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, RAS) != 0;
22
+}
26
+}
23
+
27
+
24
+static inline bool isar_feature_aa64_mte(const ARMISARegisters *id)
28
static inline bool isar_feature_aa64_sve(const ARMISARegisters *id)
29
{
30
return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, SVE) != 0;
31
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_any_debugv8p2(const ARMISARegisters *id)
32
return isar_feature_aa64_debugv8p2(id) || isar_feature_aa32_debugv8p2(id);
33
}
34
35
+static inline bool isar_feature_any_ras(const ARMISARegisters *id)
25
+{
36
+{
26
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) >= 2;
37
+ return isar_feature_aa64_ras(id) || isar_feature_aa32_ras(id);
27
+}
38
+}
28
+
39
+
29
static inline bool isar_feature_aa64_pmu_8_1(const ARMISARegisters *id)
40
/*
30
{
41
* Forward to the above feature tests given an ARMCPU pointer.
31
return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, PMUVER) >= 4 &&
42
*/
32
--
43
--
33
2.20.1
44
2.25.1
34
35
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Alex Zuepke <alex.zuepke@tum.de>
2
2
3
Protect reads of aa64 id registers with ARM_CP_STATE_AA64.
3
The ARMv8 manual defines that PMUSERENR_EL0.ER enables read-access
4
Use this as a simpler test than arm_el_is_aa64, since EL3
4
to both PMXEVCNTR_EL0 and PMEVCNTR<n>_EL0 registers, however,
5
cannot change mode.
5
we only use it for PMXEVCNTR_EL0. Extend to PMEVCNTR<n>_EL0 as well.
6
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Alex Zuepke <alex.zuepke@tum.de>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-3-richard.henderson@linaro.org
9
Message-id: 20220428132717.84190-1-alex.zuepke@tum.de
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
11
---
12
target/arm/helper.c | 15 ++++++++-------
12
target/arm/helper.c | 4 ++--
13
1 file changed, 8 insertions(+), 7 deletions(-)
13
1 file changed, 2 insertions(+), 2 deletions(-)
14
14
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
18
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
19
@@ -XXX,XX +XXX,XX @@ static void define_pmu_regs(ARMCPU *cpu)
20
uint32_t valid_mask = 0x3fff;
20
.crm = 8 | (3 & (i >> 3)), .opc1 = 0, .opc2 = i & 7,
21
ARMCPU *cpu = env_archcpu(env);
21
.access = PL0_RW, .type = ARM_CP_IO | ARM_CP_ALIAS,
22
22
.readfn = pmevcntr_readfn, .writefn = pmevcntr_writefn,
23
- if (arm_el_is_aa64(env, 3)) {
23
- .accessfn = pmreg_access },
24
+ if (ri->state == ARM_CP_STATE_AA64) {
24
+ .accessfn = pmreg_access_xevcntr },
25
value |= SCR_FW | SCR_AW; /* these two bits are RES1. */
25
{ .name = pmevcntr_el0_name, .state = ARM_CP_STATE_AA64,
26
valid_mask &= ~SCR_NET;
26
.opc0 = 3, .opc1 = 3, .crn = 14, .crm = 8 | (3 & (i >> 3)),
27
+
27
- .opc2 = i & 7, .access = PL0_RW, .accessfn = pmreg_access,
28
+ if (cpu_isar_feature(aa64_lor, cpu)) {
28
+ .opc2 = i & 7, .access = PL0_RW, .accessfn = pmreg_access_xevcntr,
29
+ valid_mask |= SCR_TLOR;
29
.type = ARM_CP_IO,
30
+ }
30
.readfn = pmevcntr_readfn, .writefn = pmevcntr_writefn,
31
+ if (cpu_isar_feature(aa64_pauth, cpu)) {
31
.raw_readfn = pmevcntr_rawread,
32
+ valid_mask |= SCR_API | SCR_APK;
33
+ }
34
} else {
35
valid_mask &= ~(SCR_RW | SCR_ST);
36
}
37
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
38
valid_mask &= ~SCR_SMD;
39
}
40
}
41
- if (cpu_isar_feature(aa64_lor, cpu)) {
42
- valid_mask |= SCR_TLOR;
43
- }
44
- if (cpu_isar_feature(aa64_pauth, cpu)) {
45
- valid_mask |= SCR_API | SCR_APK;
46
- }
47
48
/* Clear all-context RES0 bits. */
49
value &= valid_mask;
50
--
32
--
51
2.20.1
33
2.25.1
52
53
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
The current Arm ARM has adjusted the official decode of
4
"Add/subtract (immediate)" so that the shift field is only bit 22,
5
and bit 23 is part of the op1 field of the parent category
6
"Data processing - immediate".
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20200626033144.790098-11-richard.henderson@linaro.org
11
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/translate-a64.c | 23 ++++++++---------------
16
1 file changed, 8 insertions(+), 15 deletions(-)
17
18
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/translate-a64.c
21
+++ b/target/arm/translate-a64.c
22
@@ -XXX,XX +XXX,XX @@ static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
23
/*
24
* Add/subtract (immediate)
25
*
26
- * 31 30 29 28 24 23 22 21 10 9 5 4 0
27
- * +--+--+--+-----------+-----+-------------+-----+-----+
28
- * |sf|op| S| 1 0 0 0 1 |shift| imm12 | Rn | Rd |
29
- * +--+--+--+-----------+-----+-------------+-----+-----+
30
+ * 31 30 29 28 23 22 21 10 9 5 4 0
31
+ * +--+--+--+-------------+--+-------------+-----+-----+
32
+ * |sf|op| S| 1 0 0 0 1 0 |sh| imm12 | Rn | Rd |
33
+ * +--+--+--+-------------+--+-------------+-----+-----+
34
*
35
* sf: 0 -> 32bit, 1 -> 64bit
36
* op: 0 -> add , 1 -> sub
37
* S: 1 -> set flags
38
- * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
39
+ * sh: 1 -> LSL imm by 12
40
*/
41
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
42
{
43
int rd = extract32(insn, 0, 5);
44
int rn = extract32(insn, 5, 5);
45
uint64_t imm = extract32(insn, 10, 12);
46
- int shift = extract32(insn, 22, 2);
47
+ bool shift = extract32(insn, 22, 1);
48
bool setflags = extract32(insn, 29, 1);
49
bool sub_op = extract32(insn, 30, 1);
50
bool is_64bit = extract32(insn, 31, 1);
51
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
53
TCGv_i64 tcg_result;
54
55
- switch (shift) {
56
- case 0x0:
57
- break;
58
- case 0x1:
59
+ if (shift) {
60
imm <<= 12;
61
- break;
62
- default:
63
- unallocated_encoding(s);
64
- return;
65
}
66
67
tcg_result = tcg_temp_new_i64();
68
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
69
case 0x20: case 0x21: /* PC-rel. addressing */
70
disas_pc_rel_adr(s, insn);
71
break;
72
- case 0x22: case 0x23: /* Add/subtract (immediate) */
73
+ case 0x22: /* Add/subtract (immediate) */
74
disas_add_sub_imm(s, insn);
75
break;
76
case 0x24: /* Logical (immediate) */
77
--
78
2.20.1
79
80
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-12-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-a64.h | 1 +
9
target/arm/internals.h | 9 +++++++
10
target/arm/mte_helper.c | 10 ++++++++
11
target/arm/translate-a64.c | 51 ++++++++++++++++++++++++++++++++++++++
12
4 files changed, 71 insertions(+)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
19
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
20
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
diff --git a/target/arm/internals.h b/target/arm/internals.h
24
index XXXXXXX..XXXXXXX 100644
25
--- a/target/arm/internals.h
26
+++ b/target/arm/internals.h
27
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
28
*/
29
#define GMID_EL1_BS 6
30
31
+/* We associate one allocation tag per 16 bytes, the minimum. */
32
+#define LOG2_TAG_GRANULE 4
33
+#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
34
+
35
+static inline int allocation_tag_from_addr(uint64_t ptr)
36
+{
37
+ return extract64(ptr, 56, 4);
38
+}
39
+
40
static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
41
{
42
return deposit64(ptr, 56, 4, rtag);
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
48
49
return address_with_allocation_tag(rn, rtag);
50
}
51
+
52
+uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
53
+ int32_t offset, uint32_t tag_offset)
54
+{
55
+ int start_tag = allocation_tag_from_addr(ptr);
56
+ uint16_t exclude = extract32(env->cp15.gcr_el1, 0, 16);
57
+ int rtag = choose_nonexcluded_tag(start_tag, tag_offset, exclude);
58
+
59
+ return address_with_allocation_tag(ptr + offset, rtag);
60
+}
61
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
62
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate-a64.c
64
+++ b/target/arm/translate-a64.c
65
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
66
tcg_temp_free_i64(tcg_result);
67
}
68
69
+/*
70
+ * Add/subtract (immediate, with tags)
71
+ *
72
+ * 31 30 29 28 23 22 21 16 14 10 9 5 4 0
73
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
74
+ * |sf|op| S| 1 0 0 0 1 1 |o2| uimm6 |o3| uimm4 | Rn | Rd |
75
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
76
+ *
77
+ * op: 0 -> add, 1 -> sub
78
+ */
79
+static void disas_add_sub_imm_with_tags(DisasContext *s, uint32_t insn)
80
+{
81
+ int rd = extract32(insn, 0, 5);
82
+ int rn = extract32(insn, 5, 5);
83
+ int uimm4 = extract32(insn, 10, 4);
84
+ int uimm6 = extract32(insn, 16, 6);
85
+ bool sub_op = extract32(insn, 30, 1);
86
+ TCGv_i64 tcg_rn, tcg_rd;
87
+ int imm;
88
+
89
+ /* Test all of sf=1, S=0, o2=0, o3=0. */
90
+ if ((insn & 0xa040c000u) != 0x80000000u ||
91
+ !dc_isar_feature(aa64_mte_insn_reg, s)) {
92
+ unallocated_encoding(s);
93
+ return;
94
+ }
95
+
96
+ imm = uimm6 << LOG2_TAG_GRANULE;
97
+ if (sub_op) {
98
+ imm = -imm;
99
+ }
100
+
101
+ tcg_rn = cpu_reg_sp(s, rn);
102
+ tcg_rd = cpu_reg_sp(s, rd);
103
+
104
+ if (s->ata) {
105
+ TCGv_i32 offset = tcg_const_i32(imm);
106
+ TCGv_i32 tag_offset = tcg_const_i32(uimm4);
107
+
108
+ gen_helper_addsubg(tcg_rd, cpu_env, tcg_rn, offset, tag_offset);
109
+ tcg_temp_free_i32(tag_offset);
110
+ tcg_temp_free_i32(offset);
111
+ } else {
112
+ tcg_gen_addi_i64(tcg_rd, tcg_rn, imm);
113
+ gen_address_with_allocation_tag0(tcg_rd, tcg_rd);
114
+ }
115
+}
116
+
117
/* The input should be a value in the bottom e bits (with higher
118
* bits zero); returns that value replicated into every element
119
* of size e in a 64 bit integer.
120
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
121
case 0x22: /* Add/subtract (immediate) */
122
disas_add_sub_imm(s, insn);
123
break;
124
+ case 0x23: /* Add/subtract (immediate, with tags) */
125
+ disas_add_sub_imm_with_tags(s, insn);
126
+ break;
127
case 0x24: /* Logical (immediate) */
128
disas_logic_imm(s, insn);
129
break;
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-13-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 15 +++++++++++++++
9
1 file changed, 15 insertions(+)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
16
cpu_reg_sp(s, rn));
17
}
18
break;
19
+ case 5: /* GMI */
20
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
21
+ goto do_unallocated;
22
+ } else {
23
+ TCGv_i64 t1 = tcg_const_i64(1);
24
+ TCGv_i64 t2 = tcg_temp_new_i64();
25
+
26
+ tcg_gen_extract_i64(t2, cpu_reg_sp(s, rn), 56, 4);
27
+ tcg_gen_shl_i64(t1, t1, t2);
28
+ tcg_gen_or_i64(cpu_reg(s, rd), cpu_reg(s, rm), t1);
29
+
30
+ tcg_temp_free_i64(t1);
31
+ tcg_temp_free_i64(t2);
32
+ }
33
+ break;
34
case 8: /* LSLV */
35
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
36
break;
37
--
38
2.20.1
39
40
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-14-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 24 ++++++++++++++++++++++--
9
1 file changed, 22 insertions(+), 2 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
16
*/
17
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
18
{
19
- unsigned int sf, rm, opcode, rn, rd;
20
+ unsigned int sf, rm, opcode, rn, rd, setflag;
21
sf = extract32(insn, 31, 1);
22
+ setflag = extract32(insn, 29, 1);
23
rm = extract32(insn, 16, 5);
24
opcode = extract32(insn, 10, 6);
25
rn = extract32(insn, 5, 5);
26
rd = extract32(insn, 0, 5);
27
28
- if (extract32(insn, 29, 1)) {
29
+ if (setflag && opcode != 0) {
30
unallocated_encoding(s);
31
return;
32
}
33
34
switch (opcode) {
35
+ case 0: /* SUBP(S) */
36
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
37
+ goto do_unallocated;
38
+ } else {
39
+ TCGv_i64 tcg_n, tcg_m, tcg_d;
40
+
41
+ tcg_n = read_cpu_reg_sp(s, rn, true);
42
+ tcg_m = read_cpu_reg_sp(s, rm, true);
43
+ tcg_gen_sextract_i64(tcg_n, tcg_n, 0, 56);
44
+ tcg_gen_sextract_i64(tcg_m, tcg_m, 0, 56);
45
+ tcg_d = cpu_reg(s, rd);
46
+
47
+ if (setflag) {
48
+ gen_sub_CC(true, tcg_d, tcg_n, tcg_m);
49
+ } else {
50
+ tcg_gen_sub_i64(tcg_d, tcg_n, tcg_m);
51
+ }
52
+ }
53
+ break;
54
case 2: /* UDIV */
55
handle_div(s, false, sf, rm, rn, rd);
56
break;
57
--
58
2.20.1
59
60
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-17-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 29 ++++++++++++++++++++++++++---
9
1 file changed, 26 insertions(+), 3 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_ld_lit(DisasContext *s, uint32_t insn)
16
* +-----+-------+---+---+-------+---+-------+-------+------+------+
17
*
18
* opc: LDP/STP/LDNP/STNP 00 -> 32 bit, 10 -> 64 bit
19
- * LDPSW 01
20
+ * LDPSW/STGP 01
21
* LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
22
* V: 0 -> GPR, 1 -> Vector
23
* idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
24
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
25
bool is_signed = false;
26
bool postindex = false;
27
bool wback = false;
28
+ bool set_tag = false;
29
30
TCGv_i64 clean_addr, dirty_addr;
31
32
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
33
34
if (is_vector) {
35
size = 2 + opc;
36
+ } else if (opc == 1 && !is_load) {
37
+ /* STGP */
38
+ if (!dc_isar_feature(aa64_mte_insn_reg, s) || index == 0) {
39
+ unallocated_encoding(s);
40
+ return;
41
+ }
42
+ size = 3;
43
+ set_tag = true;
44
} else {
45
size = 2 + extract32(opc, 1, 1);
46
is_signed = extract32(opc, 0, 1);
47
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
48
return;
49
}
50
51
- offset <<= size;
52
+ offset <<= (set_tag ? LOG2_TAG_GRANULE : size);
53
54
if (rn == 31) {
55
gen_check_sp_alignment(s);
56
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
57
if (!postindex) {
58
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
59
}
60
- clean_addr = clean_data_tbi(s, dirty_addr);
61
62
+ if (set_tag) {
63
+ if (!s->ata) {
64
+ /*
65
+ * TODO: We could rely on the stores below, at least for
66
+ * system mode, if we arrange to add MO_ALIGN_16.
67
+ */
68
+ gen_helper_stg_stub(cpu_env, dirty_addr);
69
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
70
+ gen_helper_stg_parallel(cpu_env, dirty_addr, dirty_addr);
71
+ } else {
72
+ gen_helper_stg(cpu_env, dirty_addr, dirty_addr);
73
+ }
74
+ }
75
+
76
+ clean_addr = clean_data_tbi(s, dirty_addr);
77
if (is_vector) {
78
if (is_load) {
79
do_fp_ld(s, rt, clean_addr, size);
80
--
81
2.20.1
82
83
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
We can simplify our DC_ZVA if we recognize that the largest BS
4
that we actually use in system mode is 64. Let us just assert
5
that it fits within TARGET_PAGE_SIZE.
6
7
For DC_GVA and STZGM, we want to be able to write whole bytes
8
of tag memory, so assert that BS is >= 2 * TAG_GRANULE, or 32.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-18-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/cpu.c | 24 ++++++++++++++++++++++++
16
1 file changed, 24 insertions(+)
17
18
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.c
21
+++ b/target/arm/cpu.c
22
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
23
}
24
#endif
25
26
+ if (tcg_enabled()) {
27
+ int dcz_blocklen = 4 << cpu->dcz_blocksize;
28
+
29
+ /*
30
+ * We only support DCZ blocklen that fits on one page.
31
+ *
32
+ * Architectually this is always true. However TARGET_PAGE_SIZE
33
+ * is variable and, for compatibility with -machine virt-2.7,
34
+ * is only 1KiB, as an artifact of legacy ARMv5 subpage support.
35
+ * But even then, while the largest architectural DCZ blocklen
36
+ * is 2KiB, no cpu actually uses such a large blocklen.
37
+ */
38
+ assert(dcz_blocklen <= TARGET_PAGE_SIZE);
39
+
40
+ /*
41
+ * We only support DCZ blocksize >= 2*TAG_GRANULE, which is to say
42
+ * both nibbles of each byte storing tag data may be written at once.
43
+ * Since TAG_GRANULE is 16, this means that blocklen must be >= 32.
44
+ */
45
+ if (cpu_isar_feature(aa64_mte, cpu)) {
46
+ assert(dcz_blocklen >= 2 * TAG_GRANULE);
47
+ }
48
+ }
49
+
50
qemu_init_vcpu(cs);
51
cpu_reset(cs);
52
53
--
54
2.20.1
55
56
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Now that we know that the operation is on a single page,
4
we need not loop over pages while probing.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-19-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-a64.c | 94 +++++++++++------------------------------
12
1 file changed, 25 insertions(+), 69 deletions(-)
13
14
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.c
17
+++ b/target/arm/helper-a64.c
18
@@ -XXX,XX +XXX,XX @@ void HELPER(dc_zva)(CPUARMState *env, uint64_t vaddr_in)
19
* (which matches the usual QEMU behaviour of not implementing either
20
* alignment faults or any memory attribute handling).
21
*/
22
-
23
- ARMCPU *cpu = env_archcpu(env);
24
- uint64_t blocklen = 4 << cpu->dcz_blocksize;
25
+ int blocklen = 4 << env_archcpu(env)->dcz_blocksize;
26
uint64_t vaddr = vaddr_in & ~(blocklen - 1);
27
+ int mmu_idx = cpu_mmu_index(env, false);
28
+ void *mem;
29
+
30
+ /*
31
+ * Trapless lookup. In addition to actual invalid page, may
32
+ * return NULL for I/O, watchpoints, clean pages, etc.
33
+ */
34
+ mem = tlb_vaddr_to_host(env, vaddr, MMU_DATA_STORE, mmu_idx);
35
36
#ifndef CONFIG_USER_ONLY
37
- {
38
+ if (unlikely(!mem)) {
39
+ uintptr_t ra = GETPC();
40
+
41
/*
42
- * Slightly awkwardly, QEMU's TARGET_PAGE_SIZE may be less than
43
- * the block size so we might have to do more than one TLB lookup.
44
- * We know that in fact for any v8 CPU the page size is at least 4K
45
- * and the block size must be 2K or less, but TARGET_PAGE_SIZE is only
46
- * 1K as an artefact of legacy v5 subpage support being present in the
47
- * same QEMU executable. So in practice the hostaddr[] array has
48
- * two entries, given the current setting of TARGET_PAGE_BITS_MIN.
49
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
50
+ * the original pointer for an invalid page. But watchpoints require
51
+ * that we probe the actual space. So do both.
52
*/
53
- int maxidx = DIV_ROUND_UP(blocklen, TARGET_PAGE_SIZE);
54
- void *hostaddr[DIV_ROUND_UP(2 * KiB, 1 << TARGET_PAGE_BITS_MIN)];
55
- int try, i;
56
- unsigned mmu_idx = cpu_mmu_index(env, false);
57
- TCGMemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
58
+ (void) probe_write(env, vaddr_in, 1, mmu_idx, ra);
59
+ mem = probe_write(env, vaddr, blocklen, mmu_idx, ra);
60
61
- assert(maxidx <= ARRAY_SIZE(hostaddr));
62
-
63
- for (try = 0; try < 2; try++) {
64
-
65
- for (i = 0; i < maxidx; i++) {
66
- hostaddr[i] = tlb_vaddr_to_host(env,
67
- vaddr + TARGET_PAGE_SIZE * i,
68
- 1, mmu_idx);
69
- if (!hostaddr[i]) {
70
- break;
71
- }
72
- }
73
- if (i == maxidx) {
74
- /*
75
- * If it's all in the TLB it's fair game for just writing to;
76
- * we know we don't need to update dirty status, etc.
77
- */
78
- for (i = 0; i < maxidx - 1; i++) {
79
- memset(hostaddr[i], 0, TARGET_PAGE_SIZE);
80
- }
81
- memset(hostaddr[i], 0, blocklen - (i * TARGET_PAGE_SIZE));
82
- return;
83
- }
84
+ if (unlikely(!mem)) {
85
/*
86
- * OK, try a store and see if we can populate the tlb. This
87
- * might cause an exception if the memory isn't writable,
88
- * in which case we will longjmp out of here. We must for
89
- * this purpose use the actual register value passed to us
90
- * so that we get the fault address right.
91
+ * The only remaining reason for mem == NULL is I/O.
92
+ * Just do a series of byte writes as the architecture demands.
93
*/
94
- helper_ret_stb_mmu(env, vaddr_in, 0, oi, GETPC());
95
- /* Now we can populate the other TLB entries, if any */
96
- for (i = 0; i < maxidx; i++) {
97
- uint64_t va = vaddr + TARGET_PAGE_SIZE * i;
98
- if (va != (vaddr_in & TARGET_PAGE_MASK)) {
99
- helper_ret_stb_mmu(env, va, 0, oi, GETPC());
100
- }
101
+ for (int i = 0; i < blocklen; i++) {
102
+ cpu_stb_mmuidx_ra(env, vaddr + i, 0, mmu_idx, ra);
103
}
104
- }
105
-
106
- /*
107
- * Slow path (probably attempt to do this to an I/O device or
108
- * similar, or clearing of a block of code we have translations
109
- * cached for). Just do a series of byte writes as the architecture
110
- * demands. It's not worth trying to use a cpu_physical_memory_map(),
111
- * memset(), unmap() sequence here because:
112
- * + we'd need to account for the blocksize being larger than a page
113
- * + the direct-RAM access case is almost always going to be dealt
114
- * with in the fastpath code above, so there's no speed benefit
115
- * + we would have to deal with the map returning NULL because the
116
- * bounce buffer was in use
117
- */
118
- for (i = 0; i < blocklen; i++) {
119
- helper_ret_stb_mmu(env, vaddr + i, 0, oi, GETPC());
120
+ return;
121
}
122
}
123
-#else
124
- memset(g2h(vaddr), 0, blocklen);
125
#endif
126
+
127
+ memset(mem, 0, blocklen);
128
}
129
--
130
2.20.1
131
132
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-20-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-a64.h | 3 ++
9
target/arm/translate.h | 2 +
10
target/arm/mte_helper.c | 84 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-a64.c | 72 ++++++++++++++++++++++++++++----
12
4 files changed, 153 insertions(+), 8 deletions(-)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
19
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
20
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
21
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
22
+DEF_HELPER_FLAGS_2(ldgm, TCG_CALL_NO_WG, i64, env, i64)
23
+DEF_HELPER_FLAGS_3(stgm, TCG_CALL_NO_WG, void, env, i64, i64)
24
+DEF_HELPER_FLAGS_3(stzgm_tags, TCG_CALL_NO_WG, void, env, i64, i64)
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/translate.h
28
+++ b/target/arm/translate.h
29
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
30
* < 0, set by the current instruction.
31
*/
32
int8_t btype;
33
+ /* A copy of cpu->dcz_blocksize. */
34
+ uint8_t dcz_blocksize;
35
/* True if this page is guarded. */
36
bool guarded_page;
37
/* Bottom two bits of XScale c15_cpar coprocessor access control reg */
38
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/target/arm/mte_helper.c
41
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@ void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
43
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
44
}
45
}
46
+
47
+#define LDGM_STGM_SIZE (4 << GMID_EL1_BS)
48
+
49
+uint64_t HELPER(ldgm)(CPUARMState *env, uint64_t ptr)
50
+{
51
+ int mmu_idx = cpu_mmu_index(env, false);
52
+ uintptr_t ra = GETPC();
53
+ void *tag_mem;
54
+
55
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
56
+
57
+ /* Trap if accessing an invalid page. */
58
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD,
59
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
60
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
61
+
62
+ /* The tag is squashed to zero if the page does not support tags. */
63
+ if (!tag_mem) {
64
+ return 0;
65
+ }
66
+
67
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
68
+ /*
69
+ * We are loading 64-bits worth of tags. The ordering of elements
70
+ * within the word corresponds to a 64-bit little-endian operation.
71
+ */
72
+ return ldq_le_p(tag_mem);
73
+}
74
+
75
+void HELPER(stgm)(CPUARMState *env, uint64_t ptr, uint64_t val)
76
+{
77
+ int mmu_idx = cpu_mmu_index(env, false);
78
+ uintptr_t ra = GETPC();
79
+ void *tag_mem;
80
+
81
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
82
+
83
+ /* Trap if accessing an invalid page. */
84
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
85
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
86
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
87
+
88
+ /*
89
+ * Tag store only happens if the page support tags,
90
+ * and if the OS has enabled access to the tags.
91
+ */
92
+ if (!tag_mem) {
93
+ return;
94
+ }
95
+
96
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
97
+ /*
98
+ * We are storing 64-bits worth of tags. The ordering of elements
99
+ * within the word corresponds to a 64-bit little-endian operation.
100
+ */
101
+ stq_le_p(tag_mem, val);
102
+}
103
+
104
+void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
105
+{
106
+ uintptr_t ra = GETPC();
107
+ int mmu_idx = cpu_mmu_index(env, false);
108
+ int log2_dcz_bytes, log2_tag_bytes;
109
+ intptr_t dcz_bytes, tag_bytes;
110
+ uint8_t *mem;
111
+
112
+ /*
113
+ * In arm_cpu_realizefn, we assert that dcz > LOG2_TAG_GRANULE+1,
114
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway,
115
+ * to make sure that we can access one complete tag byte here.
116
+ */
117
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
118
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
119
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
120
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
121
+ ptr &= -dcz_bytes;
122
+
123
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, dcz_bytes,
124
+ MMU_DATA_STORE, tag_bytes, ra);
125
+ if (mem) {
126
+ int tag_pair = (val & 0xf) * 0x11;
127
+ memset(mem, tag_pair, tag_bytes);
128
+ }
129
+}
130
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
131
index XXXXXXX..XXXXXXX 100644
132
--- a/target/arm/translate-a64.c
133
+++ b/target/arm/translate-a64.c
134
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
135
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
136
int op2 = extract32(insn, 10, 2);
137
int op1 = extract32(insn, 22, 2);
138
- bool is_load = false, is_pair = false, is_zero = false;
139
+ bool is_load = false, is_pair = false, is_zero = false, is_mult = false;
140
int index = 0;
141
TCGv_i64 addr, clean_addr, tcg_rt;
142
143
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
144
if (op2 != 0) {
145
/* STG */
146
index = op2 - 2;
147
- break;
148
+ } else {
149
+ /* STZGM */
150
+ if (s->current_el == 0 || offset != 0) {
151
+ goto do_unallocated;
152
+ }
153
+ is_mult = is_zero = true;
154
}
155
- goto do_unallocated;
156
+ break;
157
case 1:
158
if (op2 != 0) {
159
/* STZG */
160
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
161
/* ST2G */
162
is_pair = true;
163
index = op2 - 2;
164
- break;
165
+ } else {
166
+ /* STGM */
167
+ if (s->current_el == 0 || offset != 0) {
168
+ goto do_unallocated;
169
+ }
170
+ is_mult = true;
171
}
172
- goto do_unallocated;
173
+ break;
174
case 3:
175
if (op2 != 0) {
176
/* STZ2G */
177
is_pair = is_zero = true;
178
index = op2 - 2;
179
- break;
180
+ } else {
181
+ /* LDGM */
182
+ if (s->current_el == 0 || offset != 0) {
183
+ goto do_unallocated;
184
+ }
185
+ is_mult = is_load = true;
186
}
187
- goto do_unallocated;
188
+ break;
189
190
default:
191
do_unallocated:
192
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
193
return;
194
}
195
196
- if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
197
+ if (is_mult
198
+ ? !dc_isar_feature(aa64_mte, s)
199
+ : !dc_isar_feature(aa64_mte_insn_reg, s)) {
200
goto do_unallocated;
201
}
202
203
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
204
tcg_gen_addi_i64(addr, addr, offset);
205
}
206
207
+ if (is_mult) {
208
+ tcg_rt = cpu_reg(s, rt);
209
+
210
+ if (is_zero) {
211
+ int size = 4 << s->dcz_blocksize;
212
+
213
+ if (s->ata) {
214
+ gen_helper_stzgm_tags(cpu_env, addr, tcg_rt);
215
+ }
216
+ /*
217
+ * The non-tags portion of STZGM is mostly like DC_ZVA,
218
+ * except the alignment happens before the access.
219
+ */
220
+ clean_addr = clean_data_tbi(s, addr);
221
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
222
+ gen_helper_dc_zva(cpu_env, clean_addr);
223
+ } else if (s->ata) {
224
+ if (is_load) {
225
+ gen_helper_ldgm(tcg_rt, cpu_env, addr);
226
+ } else {
227
+ gen_helper_stgm(cpu_env, addr, tcg_rt);
228
+ }
229
+ } else {
230
+ MMUAccessType acc = is_load ? MMU_DATA_LOAD : MMU_DATA_STORE;
231
+ int size = 4 << GMID_EL1_BS;
232
+
233
+ clean_addr = clean_data_tbi(s, addr);
234
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
235
+ gen_probe_access(s, clean_addr, acc, size);
236
+
237
+ if (is_load) {
238
+ /* The result tags are zeros. */
239
+ tcg_gen_movi_i64(tcg_rt, 0);
240
+ }
241
+ }
242
+ return;
243
+ }
244
+
245
if (is_load) {
246
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
247
tcg_rt = cpu_reg(s, rt);
248
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
249
dc->vec_stride = 0;
250
dc->cp_regs = arm_cpu->cp_regs;
251
dc->features = env->features;
252
+ dc->dcz_blocksize = arm_cpu->dcz_blocksize;
253
254
/* Single step state. The code-generation logic here is:
255
* SS_ACTIVE == 0:
256
--
257
2.20.1
258
259
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform a single logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-24-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/internals.h | 8 +++
14
target/arm/translate-a64.h | 2 +
15
target/arm/mte_helper.c | 8 +++
16
target/arm/translate-a64.c | 100 ++++++++++++++++++++++++++++---------
17
5 files changed, 95 insertions(+), 24 deletions(-)
18
19
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper-a64.h
22
+++ b/target/arm/helper-a64.h
23
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
24
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
25
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
27
+DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
28
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
29
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
30
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
31
diff --git a/target/arm/internals.h b/target/arm/internals.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/internals.h
34
+++ b/target/arm/internals.h
35
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
36
#define LOG2_TAG_GRANULE 4
37
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
38
39
+/* Bits within a descriptor passed to the helper_mte_check* functions. */
40
+FIELD(MTEDESC, MIDX, 0, 4)
41
+FIELD(MTEDESC, TBI, 4, 2)
42
+FIELD(MTEDESC, TCMA, 6, 2)
43
+FIELD(MTEDESC, WRITE, 8, 1)
44
+FIELD(MTEDESC, ESIZE, 9, 5)
45
+FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
46
+
47
static inline int allocation_tag_from_addr(uint64_t ptr)
48
{
49
return extract64(ptr, 56, 4);
50
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/translate-a64.h
53
+++ b/target/arm/translate-a64.h
54
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
55
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
56
unsigned int imms, unsigned int immr);
57
bool sve_access_check(DisasContext *s);
58
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
59
+ bool tag_checked, int log2_size);
60
61
/* We should have at some point before trying to access an FP register
62
* done the necessary access check, so assert that
63
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/mte_helper.c
66
+++ b/target/arm/mte_helper.c
67
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
68
memset(mem, tag_pair, tag_bytes);
69
}
70
}
71
+
72
+/*
73
+ * Perform an MTE checked access for a single logical or atomic access.
74
+ */
75
+uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
76
+{
77
+ return ptr;
78
+}
79
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
80
index XXXXXXX..XXXXXXX 100644
81
--- a/target/arm/translate-a64.c
82
+++ b/target/arm/translate-a64.c
83
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
84
}
85
86
/*
87
- * Return a "clean" address for ADDR according to TBID.
88
- * This is always a fresh temporary, as we need to be able to
89
- * increment this independently of a dirty write-back address.
90
+ * Handle MTE and/or TBI.
91
+ *
92
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
93
+ * for the tag to be present in the FAR_ELx register. But for user-only
94
+ * mode we do not have a TLB with which to implement this, so we must
95
+ * remove the top byte now.
96
+ *
97
+ * Always return a fresh temporary that we can increment independently
98
+ * of the write-back address.
99
*/
100
+
101
static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
102
{
103
TCGv_i64 clean = new_tmp_a64(s);
104
- /*
105
- * In order to get the correct value in the FAR_ELx register,
106
- * we must present the memory subsystem with the "dirty" address
107
- * including the TBI. In system mode we can make this work via
108
- * the TLB, dropping the TBI during translation. But for user-only
109
- * mode we don't have that option, and must remove the top byte now.
110
- */
111
#ifdef CONFIG_USER_ONLY
112
gen_top_byte_ignore(s, clean, addr, s->tbid);
113
#else
114
@@ -XXX,XX +XXX,XX @@ static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
115
tcg_temp_free_i32(t_size);
116
}
117
118
+/*
119
+ * For MTE, check a single logical or atomic access. This probes a single
120
+ * address, the exact one specified. The size and alignment of the access
121
+ * is not relevant to MTE, per se, but watchpoints do require the size,
122
+ * and we want to recognize those before making any other changes to state.
123
+ */
124
+static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
125
+ bool is_write, bool tag_checked,
126
+ int log2_size, bool is_unpriv,
127
+ int core_idx)
128
+{
129
+ if (tag_checked && s->mte_active[is_unpriv]) {
130
+ TCGv_i32 tcg_desc;
131
+ TCGv_i64 ret;
132
+ int desc = 0;
133
+
134
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, core_idx);
135
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
136
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
137
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
138
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_size);
139
+ tcg_desc = tcg_const_i32(desc);
140
+
141
+ ret = new_tmp_a64(s);
142
+ gen_helper_mte_check1(ret, cpu_env, tcg_desc, addr);
143
+ tcg_temp_free_i32(tcg_desc);
144
+
145
+ return ret;
146
+ }
147
+ return clean_data_tbi(s, addr);
148
+}
149
+
150
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
151
+ bool tag_checked, int log2_size)
152
+{
153
+ return gen_mte_check1_mmuidx(s, addr, is_write, tag_checked, log2_size,
154
+ false, get_mem_index(s));
155
+}
156
+
157
typedef struct DisasCompare64 {
158
TCGCond cond;
159
TCGv_i64 value;
160
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap(DisasContext *s, int rs, int rt,
161
if (rn == 31) {
162
gen_check_sp_alignment(s);
163
}
164
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
165
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size);
166
tcg_gen_atomic_cmpxchg_i64(tcg_rs, clean_addr, tcg_rs, tcg_rt, memidx,
167
size | MO_ALIGN | s->be_data);
168
}
169
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt,
170
if (rn == 31) {
171
gen_check_sp_alignment(s);
172
}
173
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
174
+
175
+ /* This is a single atomic access, despite the "pair". */
176
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size + 1);
177
178
if (size == 2) {
179
TCGv_i64 cmp = tcg_temp_new_i64();
180
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
181
if (is_lasr) {
182
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
183
}
184
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
185
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
186
+ true, rn != 31, size);
187
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, false);
188
return;
189
190
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
191
if (rn == 31) {
192
gen_check_sp_alignment(s);
193
}
194
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
195
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
196
+ false, rn != 31, size);
197
s->is_ldex = true;
198
gen_load_exclusive(s, rt, rt2, clean_addr, size, false);
199
if (is_lasr) {
200
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
201
gen_check_sp_alignment(s);
202
}
203
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
204
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
205
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
206
+ true, rn != 31, size);
207
do_gpr_st(s, cpu_reg(s, rt), clean_addr, size, true, rt,
208
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
209
return;
210
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
211
if (rn == 31) {
212
gen_check_sp_alignment(s);
213
}
214
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
215
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
216
+ false, rn != 31, size);
217
do_gpr_ld(s, cpu_reg(s, rt), clean_addr, size, false, false, true, rt,
218
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
219
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
220
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
221
if (is_lasr) {
222
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
223
}
224
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
225
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
226
+ true, rn != 31, size);
227
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, true);
228
return;
229
}
230
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
231
if (rn == 31) {
232
gen_check_sp_alignment(s);
233
}
234
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
235
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
236
+ false, rn != 31, size);
237
s->is_ldex = true;
238
gen_load_exclusive(s, rt, rt2, clean_addr, size, true);
239
if (is_lasr) {
240
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
241
bool iss_valid = !is_vector;
242
bool post_index;
243
bool writeback;
244
+ int memidx;
245
246
TCGv_i64 clean_addr, dirty_addr;
247
248
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
249
if (!post_index) {
250
tcg_gen_addi_i64(dirty_addr, dirty_addr, imm9);
251
}
252
- clean_addr = clean_data_tbi(s, dirty_addr);
253
+
254
+ memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
255
+ clean_addr = gen_mte_check1_mmuidx(s, dirty_addr, is_store,
256
+ writeback || rn != 31,
257
+ size, is_unpriv, memidx);
258
259
if (is_vector) {
260
if (is_store) {
261
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
262
}
263
} else {
264
TCGv_i64 tcg_rt = cpu_reg(s, rt);
265
- int memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
266
bool iss_sf = disas_ldst_compute_iss_sf(size, is_signed, opc);
267
268
if (is_store) {
269
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn,
270
ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
271
272
tcg_gen_add_i64(dirty_addr, dirty_addr, tcg_rm);
273
- clean_addr = clean_data_tbi(s, dirty_addr);
274
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, true, size);
275
276
if (is_vector) {
277
if (is_store) {
278
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn,
279
dirty_addr = read_cpu_reg_sp(s, rn, 1);
280
offset = imm12 << size;
281
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
282
- clean_addr = clean_data_tbi(s, dirty_addr);
283
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, rn != 31, size);
284
285
if (is_vector) {
286
if (is_store) {
287
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
288
if (rn == 31) {
289
gen_check_sp_alignment(s);
290
}
291
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
292
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), false, rn != 31, size);
293
294
if (o3_opc == 014) {
295
/*
296
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pac(DisasContext *s, uint32_t insn,
297
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
298
299
/* Note that "clean" and "dirty" here refer to TBI not PAC. */
300
- clean_addr = clean_data_tbi(s, dirty_addr);
301
+ clean_addr = gen_mte_check1(s, dirty_addr, false,
302
+ is_wback || rn != 31, size);
303
304
tcg_rt = cpu_reg(s, rt);
305
do_gpr_ld(s, tcg_rt, clean_addr, size, /* is_signed */ false,
306
--
307
2.20.1
308
309
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform multiple logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-25-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/translate-a64.h | 2 ++
14
target/arm/mte_helper.c | 8 +++++
15
target/arm/translate-a64.c | 71 +++++++++++++++++++++++++++++---------
16
4 files changed, 66 insertions(+), 16 deletions(-)
17
18
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/helper-a64.h
21
+++ b/target/arm/helper-a64.h
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
25
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
26
+DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
27
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
28
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
29
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
30
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/translate-a64.h
33
+++ b/target/arm/translate-a64.h
34
@@ -XXX,XX +XXX,XX @@ bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
35
bool sve_access_check(DisasContext *s);
36
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
37
bool tag_checked, int log2_size);
38
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
39
+ bool tag_checked, int count, int log2_esize);
40
41
/* We should have at some point before trying to access an FP register
42
* done the necessary access check, so assert that
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
48
{
49
return ptr;
50
}
51
+
52
+/*
53
+ * Perform an MTE checked access for multiple logical accesses.
54
+ */
55
+uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
56
+{
57
+ return ptr;
58
+}
59
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/target/arm/translate-a64.c
62
+++ b/target/arm/translate-a64.c
63
@@ -XXX,XX +XXX,XX @@ TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
64
false, get_mem_index(s));
65
}
66
67
+/*
68
+ * For MTE, check multiple logical sequential accesses.
69
+ */
70
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
71
+ bool tag_checked, int log2_esize, int total_size)
72
+{
73
+ if (tag_checked && s->mte_active[0] && total_size != (1 << log2_esize)) {
74
+ TCGv_i32 tcg_desc;
75
+ TCGv_i64 ret;
76
+ int desc = 0;
77
+
78
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
79
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
80
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
81
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
82
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_esize);
83
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, total_size);
84
+ tcg_desc = tcg_const_i32(desc);
85
+
86
+ ret = new_tmp_a64(s);
87
+ gen_helper_mte_checkN(ret, cpu_env, tcg_desc, addr);
88
+ tcg_temp_free_i32(tcg_desc);
89
+
90
+ return ret;
91
+ }
92
+ return gen_mte_check1(s, addr, is_write, tag_checked, log2_esize);
93
+}
94
+
95
typedef struct DisasCompare64 {
96
TCGCond cond;
97
TCGv_i64 value;
98
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
99
}
100
}
101
102
- clean_addr = clean_data_tbi(s, dirty_addr);
103
+ clean_addr = gen_mte_checkN(s, dirty_addr, !is_load,
104
+ (wback || rn != 31) && !set_tag,
105
+ size, 2 << size);
106
+
107
if (is_vector) {
108
if (is_load) {
109
do_fp_ld(s, rt, clean_addr, size);
110
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
111
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
112
MemOp endian = s->be_data;
113
114
- int ebytes; /* bytes per element */
115
+ int total; /* total bytes */
116
int elements; /* elements per vector */
117
int rpt; /* num iterations */
118
int selem; /* structure elements */
119
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
120
endian = MO_LE;
121
}
122
123
- /* Consecutive little-endian elements from a single register
124
+ total = rpt * selem * (is_q ? 16 : 8);
125
+ tcg_rn = cpu_reg_sp(s, rn);
126
+
127
+ /*
128
+ * Issue the MTE check vs the logical repeat count, before we
129
+ * promote consecutive little-endian elements below.
130
+ */
131
+ clean_addr = gen_mte_checkN(s, tcg_rn, is_store, is_postidx || rn != 31,
132
+ size, total);
133
+
134
+ /*
135
+ * Consecutive little-endian elements from a single register
136
* can be promoted to a larger little-endian operation.
137
*/
138
if (selem == 1 && endian == MO_LE) {
139
size = 3;
140
}
141
- ebytes = 1 << size;
142
- elements = (is_q ? 16 : 8) / ebytes;
143
-
144
- tcg_rn = cpu_reg_sp(s, rn);
145
- clean_addr = clean_data_tbi(s, tcg_rn);
146
- tcg_ebytes = tcg_const_i64(ebytes);
147
+ elements = (is_q ? 16 : 8) >> size;
148
149
+ tcg_ebytes = tcg_const_i64(1 << size);
150
for (r = 0; r < rpt; r++) {
151
int e;
152
for (e = 0; e < elements; e++) {
153
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
154
155
if (is_postidx) {
156
if (rm == 31) {
157
- tcg_gen_addi_i64(tcg_rn, tcg_rn, rpt * elements * selem * ebytes);
158
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
159
} else {
160
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
161
}
162
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
163
int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
164
bool replicate = false;
165
int index = is_q << 3 | S << 2 | size;
166
- int ebytes, xs;
167
+ int xs, total;
168
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
169
170
if (extract32(insn, 31, 1)) {
171
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
172
return;
173
}
174
175
- ebytes = 1 << scale;
176
-
177
if (rn == 31) {
178
gen_check_sp_alignment(s);
179
}
180
181
+ total = selem << scale;
182
tcg_rn = cpu_reg_sp(s, rn);
183
- clean_addr = clean_data_tbi(s, tcg_rn);
184
- tcg_ebytes = tcg_const_i64(ebytes);
185
186
+ clean_addr = gen_mte_checkN(s, tcg_rn, !is_load, is_postidx || rn != 31,
187
+ scale, total);
188
+
189
+ tcg_ebytes = tcg_const_i64(1 << scale);
190
for (xs = 0; xs < selem; xs++) {
191
if (replicate) {
192
/* Load and replicate to all elements */
193
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
194
195
if (is_postidx) {
196
if (rm == 31) {
197
- tcg_gen_addi_i64(tcg_rn, tcg_rn, selem * ebytes);
198
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
199
} else {
200
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
201
}
202
--
203
2.20.1
204
205
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-26-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 48 +++++++++++++++
11
target/arm/mte_helper.c | 132 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 179 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, WRITE, 8, 1)
19
FIELD(MTEDESC, ESIZE, 9, 5)
20
FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
21
22
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
23
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
24
+ uint64_t ptr, uintptr_t ra);
25
+
26
static inline int allocation_tag_from_addr(uint64_t ptr)
27
{
28
return extract64(ptr, 56, 4);
29
@@ -XXX,XX +XXX,XX @@ static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
30
return deposit64(ptr, 56, 4, rtag);
31
}
32
33
+/* Return true if tbi bits mean that the access is checked. */
34
+static inline bool tbi_check(uint32_t desc, int bit55)
35
+{
36
+ return (desc >> (R_MTEDESC_TBI_SHIFT + bit55)) & 1;
37
+}
38
+
39
+/* Return true if tcma bits mean that the access is unchecked. */
40
+static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
41
+{
42
+ /*
43
+ * We had extracted bit55 and ptr_tag for other reasons, so fold
44
+ * (ptr<59:55> == 00000 || ptr<59:55> == 11111) into a single test.
45
+ */
46
+ bool match = ((ptr_tag + bit55) & 0xf) == 0;
47
+ bool tcma = (desc >> (R_MTEDESC_TCMA_SHIFT + bit55)) & 1;
48
+ return tcma && match;
49
+}
50
+
51
+/*
52
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
53
+ * for the tag to be present in the FAR_ELx register. But for user-only
54
+ * mode, we do not have a TLB with which to implement this, so we must
55
+ * remove the top byte.
56
+ */
57
+static inline uint64_t useronly_clean_ptr(uint64_t ptr)
58
+{
59
+ /* TBI is known to be enabled. */
60
+#ifdef CONFIG_USER_ONLY
61
+ ptr = sextract64(ptr, 0, 56);
62
+#endif
63
+ return ptr;
64
+}
65
+
66
+static inline uint64_t useronly_maybe_clean_ptr(uint32_t desc, uint64_t ptr)
67
+{
68
+#ifdef CONFIG_USER_ONLY
69
+ int64_t clean_ptr = sextract64(ptr, 0, 56);
70
+ if (tbi_check(desc, clean_ptr < 0)) {
71
+ ptr = clean_ptr;
72
+ }
73
+#endif
74
+ return ptr;
75
+}
76
+
77
#endif
78
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/target/arm/mte_helper.c
81
+++ b/target/arm/mte_helper.c
82
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
83
}
84
}
85
86
+/* Record a tag check failure. */
87
+static void mte_check_fail(CPUARMState *env, int mmu_idx,
88
+ uint64_t dirty_ptr, uintptr_t ra)
89
+{
90
+ ARMMMUIdx arm_mmu_idx = core_to_aa64_mmu_idx(mmu_idx);
91
+ int el, reg_el, tcf, select;
92
+ uint64_t sctlr;
93
+
94
+ reg_el = regime_el(env, arm_mmu_idx);
95
+ sctlr = env->cp15.sctlr_el[reg_el];
96
+
97
+ switch (arm_mmu_idx) {
98
+ case ARMMMUIdx_E10_0:
99
+ case ARMMMUIdx_E20_0:
100
+ el = 0;
101
+ tcf = extract64(sctlr, 38, 2);
102
+ break;
103
+ default:
104
+ el = reg_el;
105
+ tcf = extract64(sctlr, 40, 2);
106
+ }
107
+
108
+ switch (tcf) {
109
+ case 1:
110
+ /*
111
+ * Tag check fail causes a synchronous exception.
112
+ *
113
+ * In restore_state_to_opc, we set the exception syndrome
114
+ * for the load or store operation. Unwind first so we
115
+ * may overwrite that with the syndrome for the tag check.
116
+ */
117
+ cpu_restore_state(env_cpu(env), ra, true);
118
+ env->exception.vaddress = dirty_ptr;
119
+ raise_exception(env, EXCP_DATA_ABORT,
120
+ syn_data_abort_no_iss(el != 0, 0, 0, 0, 0, 0, 0x11),
121
+ exception_target_el(env));
122
+ /* noreturn, but fall through to the assert anyway */
123
+
124
+ case 0:
125
+ /*
126
+ * Tag check fail does not affect the PE.
127
+ * We eliminate this case by not setting MTE_ACTIVE
128
+ * in tb_flags, so that we never make this runtime call.
129
+ */
130
+ g_assert_not_reached();
131
+
132
+ case 2:
133
+ /* Tag check fail causes asynchronous flag set. */
134
+ mmu_idx = arm_mmu_idx_el(env, el);
135
+ if (regime_has_2_ranges(mmu_idx)) {
136
+ select = extract64(dirty_ptr, 55, 1);
137
+ } else {
138
+ select = 0;
139
+ }
140
+ env->cp15.tfsr_el[el] |= 1 << select;
141
+ break;
142
+
143
+ default:
144
+ /* Case 3: Reserved. */
145
+ qemu_log_mask(LOG_GUEST_ERROR,
146
+ "Tag check failure with SCTLR_EL%d.TCF%s "
147
+ "set to reserved value %d\n",
148
+ reg_el, el ? "" : "0", tcf);
149
+ break;
150
+ }
151
+}
152
+
153
/*
154
* Perform an MTE checked access for a single logical or atomic access.
155
*/
156
+static bool mte_probe1_int(CPUARMState *env, uint32_t desc, uint64_t ptr,
157
+ uintptr_t ra, int bit55)
158
+{
159
+ int mem_tag, mmu_idx, ptr_tag, size;
160
+ MMUAccessType type;
161
+ uint8_t *mem;
162
+
163
+ ptr_tag = allocation_tag_from_addr(ptr);
164
+
165
+ if (tcma_check(desc, bit55, ptr_tag)) {
166
+ return true;
167
+ }
168
+
169
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
170
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
171
+ size = FIELD_EX32(desc, MTEDESC, ESIZE);
172
+
173
+ mem = allocation_tag_mem(env, mmu_idx, ptr, type, size,
174
+ MMU_DATA_LOAD, 1, ra);
175
+ if (!mem) {
176
+ return true;
177
+ }
178
+
179
+ mem_tag = load_tag1(ptr, mem);
180
+ return ptr_tag == mem_tag;
181
+}
182
+
183
+/*
184
+ * No-fault version of mte_check1, to be used by SVE for MemSingleNF.
185
+ * Returns false if the access is Checked and the check failed. This
186
+ * is only intended to probe the tag -- the validity of the page must
187
+ * be checked beforehand.
188
+ */
189
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr)
190
+{
191
+ int bit55 = extract64(ptr, 55, 1);
192
+
193
+ /* If TBI is disabled, the access is unchecked. */
194
+ if (unlikely(!tbi_check(desc, bit55))) {
195
+ return true;
196
+ }
197
+
198
+ return mte_probe1_int(env, desc, ptr, 0, bit55);
199
+}
200
+
201
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
202
+ uint64_t ptr, uintptr_t ra)
203
+{
204
+ int bit55 = extract64(ptr, 55, 1);
205
+
206
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
207
+ if (unlikely(!tbi_check(desc, bit55))) {
208
+ return ptr;
209
+ }
210
+
211
+ if (unlikely(!mte_probe1_int(env, desc, ptr, ra, bit55))) {
212
+ int mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
213
+ mte_check_fail(env, mmu_idx, ptr, ra);
214
+ }
215
+
216
+ return useronly_clean_ptr(ptr);
217
+}
218
+
219
uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
220
{
221
- return ptr;
222
+ return mte_check1(env, desc, ptr, GETPC());
223
}
224
225
/*
226
--
227
2.20.1
228
229
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-27-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 2 +
11
target/arm/mte_helper.c | 165 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 166 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
19
bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
20
uint64_t mte_check1(CPUARMState *env, uint32_t desc,
21
uint64_t ptr, uintptr_t ra);
22
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
23
+ uint64_t ptr, uintptr_t ra);
24
25
static inline int allocation_tag_from_addr(uint64_t ptr)
26
{
27
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/mte_helper.c
30
+++ b/target/arm/mte_helper.c
31
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
32
/*
33
* Perform an MTE checked access for multiple logical accesses.
34
*/
35
+
36
+/**
37
+ * checkN:
38
+ * @tag: tag memory to test
39
+ * @odd: true to begin testing at tags at odd nibble
40
+ * @cmp: the tag to compare against
41
+ * @count: number of tags to test
42
+ *
43
+ * Return the number of successful tests.
44
+ * Thus a return value < @count indicates a failure.
45
+ *
46
+ * A note about sizes: count is expected to be small.
47
+ *
48
+ * The most common use will be LDP/STP of two integer registers,
49
+ * which means 16 bytes of memory touching at most 2 tags, but
50
+ * often the access is aligned and thus just 1 tag.
51
+ *
52
+ * Using AdvSIMD LD/ST (multiple), one can access 64 bytes of memory,
53
+ * touching at most 5 tags. SVE LDR/STR (vector) with the default
54
+ * vector length is also 64 bytes; the maximum architectural length
55
+ * is 256 bytes touching at most 9 tags.
56
+ *
57
+ * The loop below uses 7 logical operations and 1 memory operation
58
+ * per tag pair. An implementation that loads an aligned word and
59
+ * uses masking to ignore adjacent tags requires 18 logical operations
60
+ * and thus does not begin to pay off until 6 tags.
61
+ * Which, according to the survey above, is unlikely to be common.
62
+ */
63
+static int checkN(uint8_t *mem, int odd, int cmp, int count)
64
+{
65
+ int n = 0, diff;
66
+
67
+ /* Replicate the test tag and compare. */
68
+ cmp *= 0x11;
69
+ diff = *mem++ ^ cmp;
70
+
71
+ if (odd) {
72
+ goto start_odd;
73
+ }
74
+
75
+ while (1) {
76
+ /* Test even tag. */
77
+ if (unlikely((diff) & 0x0f)) {
78
+ break;
79
+ }
80
+ if (++n == count) {
81
+ break;
82
+ }
83
+
84
+ start_odd:
85
+ /* Test odd tag. */
86
+ if (unlikely((diff) & 0xf0)) {
87
+ break;
88
+ }
89
+ if (++n == count) {
90
+ break;
91
+ }
92
+
93
+ diff = *mem++ ^ cmp;
94
+ }
95
+ return n;
96
+}
97
+
98
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
99
+ uint64_t ptr, uintptr_t ra)
100
+{
101
+ int mmu_idx, ptr_tag, bit55;
102
+ uint64_t ptr_last, ptr_end, prev_page, next_page;
103
+ uint64_t tag_first, tag_end;
104
+ uint64_t tag_byte_first, tag_byte_end;
105
+ uint32_t esize, total, tag_count, tag_size, n, c;
106
+ uint8_t *mem1, *mem2;
107
+ MMUAccessType type;
108
+
109
+ bit55 = extract64(ptr, 55, 1);
110
+
111
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
112
+ if (unlikely(!tbi_check(desc, bit55))) {
113
+ return ptr;
114
+ }
115
+
116
+ ptr_tag = allocation_tag_from_addr(ptr);
117
+
118
+ if (tcma_check(desc, bit55, ptr_tag)) {
119
+ goto done;
120
+ }
121
+
122
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
123
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
124
+ esize = FIELD_EX32(desc, MTEDESC, ESIZE);
125
+ total = FIELD_EX32(desc, MTEDESC, TSIZE);
126
+
127
+ /* Find the addr of the end of the access, and of the last element. */
128
+ ptr_end = ptr + total;
129
+ ptr_last = ptr_end - esize;
130
+
131
+ /* Round the bounds to the tag granule, and compute the number of tags. */
132
+ tag_first = QEMU_ALIGN_DOWN(ptr, TAG_GRANULE);
133
+ tag_end = QEMU_ALIGN_UP(ptr_last, TAG_GRANULE);
134
+ tag_count = (tag_end - tag_first) / TAG_GRANULE;
135
+
136
+ /* Round the bounds to twice the tag granule, and compute the bytes. */
137
+ tag_byte_first = QEMU_ALIGN_DOWN(ptr, 2 * TAG_GRANULE);
138
+ tag_byte_end = QEMU_ALIGN_UP(ptr_last, 2 * TAG_GRANULE);
139
+
140
+ /* Locate the page boundaries. */
141
+ prev_page = ptr & TARGET_PAGE_MASK;
142
+ next_page = prev_page + TARGET_PAGE_SIZE;
143
+
144
+ if (likely(tag_end - prev_page <= TARGET_PAGE_SIZE)) {
145
+ /* Memory access stays on one page. */
146
+ tag_size = (tag_byte_end - tag_byte_first) / (2 * TAG_GRANULE);
147
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, total,
148
+ MMU_DATA_LOAD, tag_size, ra);
149
+ if (!mem1) {
150
+ goto done;
151
+ }
152
+ /* Perform all of the comparisons. */
153
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, tag_count);
154
+ } else {
155
+ /* Memory access crosses to next page. */
156
+ tag_size = (next_page - tag_byte_first) / (2 * TAG_GRANULE);
157
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, next_page - ptr,
158
+ MMU_DATA_LOAD, tag_size, ra);
159
+
160
+ tag_size = (tag_byte_end - next_page) / (2 * TAG_GRANULE);
161
+ mem2 = allocation_tag_mem(env, mmu_idx, next_page, type,
162
+ ptr_end - next_page,
163
+ MMU_DATA_LOAD, tag_size, ra);
164
+
165
+ /*
166
+ * Perform all of the comparisons.
167
+ * Note the possible but unlikely case of the operation spanning
168
+ * two pages that do not both have tagging enabled.
169
+ */
170
+ n = c = (next_page - tag_first) / TAG_GRANULE;
171
+ if (mem1) {
172
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, c);
173
+ }
174
+ if (n == c) {
175
+ if (!mem2) {
176
+ goto done;
177
+ }
178
+ n += checkN(mem2, 0, ptr_tag, tag_count - c);
179
+ }
180
+ }
181
+
182
+ /*
183
+ * If we failed, we know which granule. Compute the element that
184
+ * is first in that granule, and signal failure on that element.
185
+ */
186
+ if (unlikely(n < tag_count)) {
187
+ uint64_t fail_ofs;
188
+
189
+ fail_ofs = tag_first + n * TAG_GRANULE - ptr;
190
+ fail_ofs = ROUND_UP(fail_ofs, esize);
191
+ mte_check_fail(env, mmu_idx, ptr + fail_ofs, ra);
192
+ }
193
+
194
+ done:
195
+ return useronly_clean_ptr(ptr);
196
+}
197
+
198
uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
199
{
200
- return ptr;
201
+ return mte_checkN(env, desc, ptr, GETPC());
202
}
203
--
204
2.20.1
205
206
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Use a special helper for DC_ZVA, rather than the more
4
general mte_checkN.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-28-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-a64.h | 1 +
12
target/arm/mte_helper.c | 106 +++++++++++++++++++++++++++++++++++++
13
target/arm/translate-a64.c | 16 +++++-
14
3 files changed, 122 insertions(+), 1 deletion(-)
15
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-a64.h
19
+++ b/target/arm/helper-a64.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
21
22
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
23
DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
24
+DEF_HELPER_FLAGS_3(mte_check_zva, TCG_CALL_NO_WG, i64, env, i32, i64)
25
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
26
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
27
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
28
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
29
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/mte_helper.c
31
+++ b/target/arm/mte_helper.c
32
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
33
{
34
return mte_checkN(env, desc, ptr, GETPC());
35
}
36
+
37
+/*
38
+ * Perform an MTE checked access for DC_ZVA.
39
+ */
40
+uint64_t HELPER(mte_check_zva)(CPUARMState *env, uint32_t desc, uint64_t ptr)
41
+{
42
+ uintptr_t ra = GETPC();
43
+ int log2_dcz_bytes, log2_tag_bytes;
44
+ int mmu_idx, bit55;
45
+ intptr_t dcz_bytes, tag_bytes, i;
46
+ void *mem;
47
+ uint64_t ptr_tag, mem_tag, align_ptr;
48
+
49
+ bit55 = extract64(ptr, 55, 1);
50
+
51
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
52
+ if (unlikely(!tbi_check(desc, bit55))) {
53
+ return ptr;
54
+ }
55
+
56
+ ptr_tag = allocation_tag_from_addr(ptr);
57
+
58
+ if (tcma_check(desc, bit55, ptr_tag)) {
59
+ goto done;
60
+ }
61
+
62
+ /*
63
+ * In arm_cpu_realizefn, we asserted that dcz > LOG2_TAG_GRANULE+1,
64
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway, to make
65
+ * sure that we can access one complete tag byte here.
66
+ */
67
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
68
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
69
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
70
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
71
+ align_ptr = ptr & -dcz_bytes;
72
+
73
+ /*
74
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
75
+ * the original pointer for an invalid page. But watchpoints require
76
+ * that we probe the actual space. So do both.
77
+ */
78
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
79
+ (void) probe_write(env, ptr, 1, mmu_idx, ra);
80
+ mem = allocation_tag_mem(env, mmu_idx, align_ptr, MMU_DATA_STORE,
81
+ dcz_bytes, MMU_DATA_LOAD, tag_bytes, ra);
82
+ if (!mem) {
83
+ goto done;
84
+ }
85
+
86
+ /*
87
+ * Unlike the reasoning for checkN, DC_ZVA is always aligned, and thus
88
+ * it is quite easy to perform all of the comparisons at once without
89
+ * any extra masking.
90
+ *
91
+ * The most common zva block size is 64; some of the thunderx cpus use
92
+ * a block size of 128. For user-only, aarch64_max_initfn will set the
93
+ * block size to 512. Fill out the other cases for future-proofing.
94
+ *
95
+ * In order to be able to find the first miscompare later, we want the
96
+ * tag bytes to be in little-endian order.
97
+ */
98
+ switch (log2_tag_bytes) {
99
+ case 0: /* zva_blocksize 32 */
100
+ mem_tag = *(uint8_t *)mem;
101
+ ptr_tag *= 0x11u;
102
+ break;
103
+ case 1: /* zva_blocksize 64 */
104
+ mem_tag = cpu_to_le16(*(uint16_t *)mem);
105
+ ptr_tag *= 0x1111u;
106
+ break;
107
+ case 2: /* zva_blocksize 128 */
108
+ mem_tag = cpu_to_le32(*(uint32_t *)mem);
109
+ ptr_tag *= 0x11111111u;
110
+ break;
111
+ case 3: /* zva_blocksize 256 */
112
+ mem_tag = cpu_to_le64(*(uint64_t *)mem);
113
+ ptr_tag *= 0x1111111111111111ull;
114
+ break;
115
+
116
+ default: /* zva_blocksize 512, 1024, 2048 */
117
+ ptr_tag *= 0x1111111111111111ull;
118
+ i = 0;
119
+ do {
120
+ mem_tag = cpu_to_le64(*(uint64_t *)(mem + i));
121
+ if (unlikely(mem_tag != ptr_tag)) {
122
+ goto fail;
123
+ }
124
+ i += 8;
125
+ align_ptr += 16 * TAG_GRANULE;
126
+ } while (i < tag_bytes);
127
+ goto done;
128
+ }
129
+
130
+ if (likely(mem_tag == ptr_tag)) {
131
+ goto done;
132
+ }
133
+
134
+ fail:
135
+ /* Locate the first nibble that differs. */
136
+ i = ctz64(mem_tag ^ ptr_tag) >> 4;
137
+ mte_check_fail(env, mmu_idx, align_ptr + i * TAG_GRANULE, ra);
138
+
139
+ done:
140
+ return useronly_clean_ptr(ptr);
141
+}
142
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
143
index XXXXXXX..XXXXXXX 100644
144
--- a/target/arm/translate-a64.c
145
+++ b/target/arm/translate-a64.c
146
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
147
return;
148
case ARM_CP_DC_ZVA:
149
/* Writes clear the aligned block of memory which rt points into. */
150
- tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
151
+ if (s->mte_active[0]) {
152
+ TCGv_i32 t_desc;
153
+ int desc = 0;
154
+
155
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
156
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
157
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
158
+ t_desc = tcg_const_i32(desc);
159
+
160
+ tcg_rt = new_tmp_a64(s);
161
+ gen_helper_mte_check_zva(tcg_rt, cpu_env, t_desc, cpu_reg(s, rt));
162
+ tcg_temp_free_i32(t_desc);
163
+ } else {
164
+ tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
165
+ }
166
gen_helper_dc_zva(cpu_env, tcg_rt);
167
return;
168
default:
169
--
170
2.20.1
171
172
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-29-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
9
1 file changed, 33 insertions(+), 28 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
16
int len_remain = len % 8;
17
int nparts = len / 8 + ctpop8(len_remain);
18
int midx = get_mem_index(s);
19
- TCGv_i64 addr, t0, t1;
20
+ TCGv_i64 dirty_addr, clean_addr, t0, t1;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
- /* Note that unpredicated load/store of vector/predicate registers
30
+ /*
31
+ * Note that unpredicated load/store of vector/predicate registers
32
* are defined as a stream of bytes, which equates to little-endian
33
- * operations on larger quantities. There is no nice way to force
34
- * a little-endian load for aarch64_be-linux-user out of line.
35
- *
36
+ * operations on larger quantities.
37
* Attempt to keep code expansion to a minimum by limiting the
38
* amount of unrolling done.
39
*/
40
if (nparts <= 4) {
41
int i;
42
43
+ t0 = tcg_temp_new_i64();
44
for (i = 0; i < len_align; i += 8) {
45
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
46
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
47
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
48
tcg_gen_st_i64(t0, cpu_env, vofs + i);
49
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
50
}
51
+ tcg_temp_free_i64(t0);
52
} else {
53
TCGLabel *loop = gen_new_label();
54
TCGv_ptr tp, i = tcg_const_local_ptr(0);
55
56
+ /* Copy the clean address into a local temp, live across the loop. */
57
+ t0 = clean_addr;
58
+ clean_addr = tcg_temp_local_new_i64();
59
+ tcg_gen_mov_i64(clean_addr, t0);
60
+ tcg_temp_free_i64(t0);
61
+
62
gen_set_label(loop);
63
64
- /* Minimize the number of local temps that must be re-read from
65
- * the stack each iteration. Instead, re-compute values other
66
- * than the loop counter.
67
- */
68
+ t0 = tcg_temp_new_i64();
69
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
70
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
71
+
72
tp = tcg_temp_new_ptr();
73
- tcg_gen_addi_ptr(tp, i, imm);
74
- tcg_gen_extu_ptr_i64(addr, tp);
75
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
76
-
77
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
78
-
79
tcg_gen_add_ptr(tp, cpu_env, i);
80
tcg_gen_addi_ptr(i, i, 8);
81
tcg_gen_st_i64(t0, tp, vofs);
82
tcg_temp_free_ptr(tp);
83
+ tcg_temp_free_i64(t0);
84
85
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
86
tcg_temp_free_ptr(i);
87
}
88
89
- /* Predicate register loads can be any multiple of 2.
90
+ /*
91
+ * Predicate register loads can be any multiple of 2.
92
* Note that we still store the entire 64-bit unit into cpu_env.
93
*/
94
if (len_remain) {
95
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
96
-
97
+ t0 = tcg_temp_new_i64();
98
switch (len_remain) {
99
case 2:
100
case 4:
101
case 8:
102
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
103
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx,
104
+ MO_LE | ctz32(len_remain));
105
break;
106
107
case 6:
108
t1 = tcg_temp_new_i64();
109
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL);
110
- tcg_gen_addi_i64(addr, addr, 4);
111
- tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW);
112
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUL);
113
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
114
+ tcg_gen_qemu_ld_i64(t1, clean_addr, midx, MO_LEUW);
115
tcg_gen_deposit_i64(t0, t0, t1, 32, 32);
116
tcg_temp_free_i64(t1);
117
break;
118
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
119
g_assert_not_reached();
120
}
121
tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
122
+ tcg_temp_free_i64(t0);
123
}
124
- tcg_temp_free_i64(addr);
125
- tcg_temp_free_i64(t0);
126
+ tcg_temp_free_i64(clean_addr);
127
}
128
129
/* Similarly for stores. */
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-30-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
9
1 file changed, 33 insertions(+), 28 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
16
int len_remain = len % 8;
17
int nparts = len / 8 + ctpop8(len_remain);
18
int midx = get_mem_index(s);
19
- TCGv_i64 addr, t0;
20
+ TCGv_i64 dirty_addr, clean_addr, t0;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
/* Note that unpredicated load/store of vector/predicate registers
30
* are defined as a stream of bytes, which equates to little-endian
31
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
32
if (nparts <= 4) {
33
int i;
34
35
+ t0 = tcg_temp_new_i64();
36
for (i = 0; i < len_align; i += 8) {
37
tcg_gen_ld_i64(t0, cpu_env, vofs + i);
38
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
39
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
40
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
41
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
42
}
43
+ tcg_temp_free_i64(t0);
44
} else {
45
TCGLabel *loop = gen_new_label();
46
- TCGv_ptr t2, i = tcg_const_local_ptr(0);
47
+ TCGv_ptr tp, i = tcg_const_local_ptr(0);
48
+
49
+ /* Copy the clean address into a local temp, live across the loop. */
50
+ t0 = clean_addr;
51
+ clean_addr = tcg_temp_local_new_i64();
52
+ tcg_gen_mov_i64(clean_addr, t0);
53
+ tcg_temp_free_i64(t0);
54
55
gen_set_label(loop);
56
57
- t2 = tcg_temp_new_ptr();
58
- tcg_gen_add_ptr(t2, cpu_env, i);
59
- tcg_gen_ld_i64(t0, t2, vofs);
60
-
61
- /* Minimize the number of local temps that must be re-read from
62
- * the stack each iteration. Instead, re-compute values other
63
- * than the loop counter.
64
- */
65
- tcg_gen_addi_ptr(t2, i, imm);
66
- tcg_gen_extu_ptr_i64(addr, t2);
67
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
68
- tcg_temp_free_ptr(t2);
69
-
70
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
71
-
72
+ t0 = tcg_temp_new_i64();
73
+ tp = tcg_temp_new_ptr();
74
+ tcg_gen_add_ptr(tp, cpu_env, i);
75
+ tcg_gen_ld_i64(t0, tp, vofs);
76
tcg_gen_addi_ptr(i, i, 8);
77
+ tcg_temp_free_ptr(tp);
78
+
79
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
80
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
81
+ tcg_temp_free_i64(t0);
82
83
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
84
tcg_temp_free_ptr(i);
85
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
86
87
/* Predicate register stores can be any multiple of 2. */
88
if (len_remain) {
89
+ t0 = tcg_temp_new_i64();
90
tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
91
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
92
93
switch (len_remain) {
94
case 2:
95
case 4:
96
case 8:
97
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
98
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx,
99
+ MO_LE | ctz32(len_remain));
100
break;
101
102
case 6:
103
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUL);
104
- tcg_gen_addi_i64(addr, addr, 4);
105
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUL);
106
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
107
tcg_gen_shri_i64(t0, t0, 32);
108
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUW);
109
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUW);
110
break;
111
112
default:
113
g_assert_not_reached();
114
}
115
+ tcg_temp_free_i64(t0);
116
}
117
- tcg_temp_free_i64(addr);
118
- tcg_temp_free_i64(t0);
119
+ tcg_temp_free_i64(clean_addr);
120
}
121
122
static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
123
--
124
2.20.1
125
126
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-31-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 6 ++++--
9
1 file changed, 4 insertions(+), 2 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
16
unsigned esz = dtype_esz[a->dtype];
17
unsigned msz = dtype_msz(a->dtype);
18
TCGLabel *over = gen_new_label();
19
- TCGv_i64 temp;
20
+ TCGv_i64 temp, clean_addr;
21
22
/* If the guarding predicate has no bits set, no load occurs. */
23
if (psz <= 8) {
24
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
25
/* Load the data. */
26
temp = tcg_temp_new_i64();
27
tcg_gen_addi_i64(temp, cpu_reg_sp(s, a->rn), a->imm << msz);
28
- tcg_gen_qemu_ld_i64(temp, temp, get_mem_index(s),
29
+ clean_addr = gen_mte_check1(s, temp, false, true, msz);
30
+
31
+ tcg_gen_qemu_ld_i64(temp, clean_addr, get_mem_index(s),
32
s->be_data | dtype_mop[a->dtype]);
33
34
/* Broadcast to *all* elements. */
35
--
36
2.20.1
37
38
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Move the variable declarations to the top of the function,
4
but do not create a new label before sve_access_check.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-32-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/translate-sve.c | 12 +++++++-----
12
1 file changed, 7 insertions(+), 5 deletions(-)
13
14
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate-sve.c
17
+++ b/target/arm/translate-sve.c
18
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RQ_zpri(DisasContext *s, arg_rpri_load *a)
19
/* Load and broadcast element. */
20
static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
21
{
22
- if (!sve_access_check(s)) {
23
- return true;
24
- }
25
-
26
unsigned vsz = vec_full_reg_size(s);
27
unsigned psz = pred_full_reg_size(s);
28
unsigned esz = dtype_esz[a->dtype];
29
unsigned msz = dtype_msz(a->dtype);
30
- TCGLabel *over = gen_new_label();
31
+ TCGLabel *over;
32
TCGv_i64 temp, clean_addr;
33
34
+ if (!sve_access_check(s)) {
35
+ return true;
36
+ }
37
+
38
+ over = gen_new_label();
39
+
40
/* If the guarding predicate has no bits set, no load occurs. */
41
if (psz <= 8) {
42
/* Reduce the pred_esz_masks value simply to reduce the
43
--
44
2.20.1
45
46
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-35-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-sve.h | 47 +++++++++++
12
target/arm/sve_helper.c | 95 ++++++++++++++++------
13
target/arm/translate-sve.c | 162 ++++++++++++++++++++++++-------------
14
3 files changed, 226 insertions(+), 78 deletions(-)
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_st1hd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
DEF_HELPER_FLAGS_4(sve_st1sd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_st1sd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_st1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_st2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_st3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_st4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+
29
+DEF_HELPER_FLAGS_4(sve_st1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_st2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_4(sve_st3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
32
+DEF_HELPER_FLAGS_4(sve_st4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
33
+
34
+DEF_HELPER_FLAGS_4(sve_st1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_4(sve_st2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_st3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_4(sve_st4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
+
39
+DEF_HELPER_FLAGS_4(sve_st1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_st2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_st3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+DEF_HELPER_FLAGS_4(sve_st4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
+
44
+DEF_HELPER_FLAGS_4(sve_st1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_st2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_st3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_st4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
+
49
+DEF_HELPER_FLAGS_4(sve_st1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_st2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_st3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_4(sve_st4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_st1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_st2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_st3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_4(sve_st4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_4(sve_st1bh_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_st1bs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_st1bd_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_st1hs_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_st1hd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_st1hs_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_st1hd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_st1sd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_st1sd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+
71
DEF_HELPER_FLAGS_6(sve_ldbsu_zsu, TCG_CALL_NO_WG,
72
void, env, ptr, ptr, ptr, tl, i32)
73
DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu, TCG_CALL_NO_WG,
74
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
75
index XXXXXXX..XXXXXXX 100644
76
--- a/target/arm/sve_helper.c
77
+++ b/target/arm/sve_helper.c
78
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_LDNF1_2(dd, MO_64, MO_64)
79
*/
80
81
static inline QEMU_ALWAYS_INLINE
82
-void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
83
- const uintptr_t retaddr, const int esz,
84
- const int msz, const int N,
85
+void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr,
86
+ uint32_t desc, const uintptr_t retaddr,
87
+ const int esz, const int msz, const int N, uint32_t mtedesc,
88
sve_ldst1_host_fn *host_fn,
89
- sve_ldst1_tlb_fn *tlb_fn)
90
+ sve_ldst1_tlb_fn *tlb_fn,
91
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
92
{
93
const unsigned rd = simd_data(desc);
94
const intptr_t reg_max = simd_oprsz(desc);
95
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
96
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
97
BP_MEM_WRITE, retaddr);
98
99
- /* TODO: MTE check. */
100
+ /*
101
+ * Handle mte checks for all active elements.
102
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
103
+ */
104
+ if (mte_check_fn && mtedesc) {
105
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
106
+ mtedesc, retaddr);
107
+ }
108
109
flags = info.page[0].flags | info.page[1].flags;
110
if (unlikely(flags != 0)) {
111
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
112
}
113
}
114
115
-#define DO_STN_1(N, NAME, ESZ) \
116
-void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
117
- target_ulong addr, uint32_t desc) \
118
-{ \
119
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
120
- sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
121
+static inline QEMU_ALWAYS_INLINE
122
+void sve_stN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
123
+ uint32_t desc, const uintptr_t ra,
124
+ const int esz, const int msz, const int N,
125
+ sve_ldst1_host_fn *host_fn,
126
+ sve_ldst1_tlb_fn *tlb_fn)
127
+{
128
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
129
+ int bit55 = extract64(addr, 55, 1);
130
+
131
+ /* Remove mtedesc from the normal sve descriptor. */
132
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
133
+
134
+ /* Perform gross MTE suppression early. */
135
+ if (!tbi_check(desc, bit55) ||
136
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
137
+ mtedesc = 0;
138
+ }
139
+
140
+ sve_stN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
141
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
142
}
143
144
-#define DO_STN_2(N, NAME, ESZ, MSZ) \
145
-void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
146
- target_ulong addr, uint32_t desc) \
147
-{ \
148
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
149
- sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
150
-} \
151
-void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
152
- target_ulong addr, uint32_t desc) \
153
-{ \
154
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
155
- sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
156
+#define DO_STN_1(N, NAME, ESZ) \
157
+void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
158
+ target_ulong addr, uint32_t desc) \
159
+{ \
160
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, 0, \
161
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb, NULL); \
162
+} \
163
+void HELPER(sve_st##N##NAME##_r_mte)(CPUARMState *env, void *vg, \
164
+ target_ulong addr, uint32_t desc) \
165
+{ \
166
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
167
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
168
+}
169
+
170
+#define DO_STN_2(N, NAME, ESZ, MSZ) \
171
+void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
172
+ target_ulong addr, uint32_t desc) \
173
+{ \
174
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
175
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb, NULL); \
176
+} \
177
+void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
178
+ target_ulong addr, uint32_t desc) \
179
+{ \
180
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
181
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb, NULL); \
182
+} \
183
+void HELPER(sve_st##N##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
184
+ target_ulong addr, uint32_t desc) \
185
+{ \
186
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
187
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
188
+} \
189
+void HELPER(sve_st##N##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
190
+ target_ulong addr, uint32_t desc) \
191
+{ \
192
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
193
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
194
}
195
196
DO_STN_1(1, bb, MO_8)
197
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
198
index XXXXXXX..XXXXXXX 100644
199
--- a/target/arm/translate-sve.c
200
+++ b/target/arm/translate-sve.c
201
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
202
static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
203
int msz, int esz, int nreg)
204
{
205
- static gen_helper_gvec_mem * const fn_single[2][4][4] = {
206
- { { gen_helper_sve_st1bb_r,
207
- gen_helper_sve_st1bh_r,
208
- gen_helper_sve_st1bs_r,
209
- gen_helper_sve_st1bd_r },
210
- { NULL,
211
- gen_helper_sve_st1hh_le_r,
212
- gen_helper_sve_st1hs_le_r,
213
- gen_helper_sve_st1hd_le_r },
214
- { NULL, NULL,
215
- gen_helper_sve_st1ss_le_r,
216
- gen_helper_sve_st1sd_le_r },
217
- { NULL, NULL, NULL,
218
- gen_helper_sve_st1dd_le_r } },
219
- { { gen_helper_sve_st1bb_r,
220
- gen_helper_sve_st1bh_r,
221
- gen_helper_sve_st1bs_r,
222
- gen_helper_sve_st1bd_r },
223
- { NULL,
224
- gen_helper_sve_st1hh_be_r,
225
- gen_helper_sve_st1hs_be_r,
226
- gen_helper_sve_st1hd_be_r },
227
- { NULL, NULL,
228
- gen_helper_sve_st1ss_be_r,
229
- gen_helper_sve_st1sd_be_r },
230
- { NULL, NULL, NULL,
231
- gen_helper_sve_st1dd_be_r } },
232
+ static gen_helper_gvec_mem * const fn_single[2][2][4][4] = {
233
+ { { { gen_helper_sve_st1bb_r,
234
+ gen_helper_sve_st1bh_r,
235
+ gen_helper_sve_st1bs_r,
236
+ gen_helper_sve_st1bd_r },
237
+ { NULL,
238
+ gen_helper_sve_st1hh_le_r,
239
+ gen_helper_sve_st1hs_le_r,
240
+ gen_helper_sve_st1hd_le_r },
241
+ { NULL, NULL,
242
+ gen_helper_sve_st1ss_le_r,
243
+ gen_helper_sve_st1sd_le_r },
244
+ { NULL, NULL, NULL,
245
+ gen_helper_sve_st1dd_le_r } },
246
+ { { gen_helper_sve_st1bb_r,
247
+ gen_helper_sve_st1bh_r,
248
+ gen_helper_sve_st1bs_r,
249
+ gen_helper_sve_st1bd_r },
250
+ { NULL,
251
+ gen_helper_sve_st1hh_be_r,
252
+ gen_helper_sve_st1hs_be_r,
253
+ gen_helper_sve_st1hd_be_r },
254
+ { NULL, NULL,
255
+ gen_helper_sve_st1ss_be_r,
256
+ gen_helper_sve_st1sd_be_r },
257
+ { NULL, NULL, NULL,
258
+ gen_helper_sve_st1dd_be_r } } },
259
+
260
+ { { { gen_helper_sve_st1bb_r_mte,
261
+ gen_helper_sve_st1bh_r_mte,
262
+ gen_helper_sve_st1bs_r_mte,
263
+ gen_helper_sve_st1bd_r_mte },
264
+ { NULL,
265
+ gen_helper_sve_st1hh_le_r_mte,
266
+ gen_helper_sve_st1hs_le_r_mte,
267
+ gen_helper_sve_st1hd_le_r_mte },
268
+ { NULL, NULL,
269
+ gen_helper_sve_st1ss_le_r_mte,
270
+ gen_helper_sve_st1sd_le_r_mte },
271
+ { NULL, NULL, NULL,
272
+ gen_helper_sve_st1dd_le_r_mte } },
273
+ { { gen_helper_sve_st1bb_r_mte,
274
+ gen_helper_sve_st1bh_r_mte,
275
+ gen_helper_sve_st1bs_r_mte,
276
+ gen_helper_sve_st1bd_r_mte },
277
+ { NULL,
278
+ gen_helper_sve_st1hh_be_r_mte,
279
+ gen_helper_sve_st1hs_be_r_mte,
280
+ gen_helper_sve_st1hd_be_r_mte },
281
+ { NULL, NULL,
282
+ gen_helper_sve_st1ss_be_r_mte,
283
+ gen_helper_sve_st1sd_be_r_mte },
284
+ { NULL, NULL, NULL,
285
+ gen_helper_sve_st1dd_be_r_mte } } },
286
};
287
- static gen_helper_gvec_mem * const fn_multiple[2][3][4] = {
288
- { { gen_helper_sve_st2bb_r,
289
- gen_helper_sve_st2hh_le_r,
290
- gen_helper_sve_st2ss_le_r,
291
- gen_helper_sve_st2dd_le_r },
292
- { gen_helper_sve_st3bb_r,
293
- gen_helper_sve_st3hh_le_r,
294
- gen_helper_sve_st3ss_le_r,
295
- gen_helper_sve_st3dd_le_r },
296
- { gen_helper_sve_st4bb_r,
297
- gen_helper_sve_st4hh_le_r,
298
- gen_helper_sve_st4ss_le_r,
299
- gen_helper_sve_st4dd_le_r } },
300
- { { gen_helper_sve_st2bb_r,
301
- gen_helper_sve_st2hh_be_r,
302
- gen_helper_sve_st2ss_be_r,
303
- gen_helper_sve_st2dd_be_r },
304
- { gen_helper_sve_st3bb_r,
305
- gen_helper_sve_st3hh_be_r,
306
- gen_helper_sve_st3ss_be_r,
307
- gen_helper_sve_st3dd_be_r },
308
- { gen_helper_sve_st4bb_r,
309
- gen_helper_sve_st4hh_be_r,
310
- gen_helper_sve_st4ss_be_r,
311
- gen_helper_sve_st4dd_be_r } },
312
+ static gen_helper_gvec_mem * const fn_multiple[2][2][3][4] = {
313
+ { { { gen_helper_sve_st2bb_r,
314
+ gen_helper_sve_st2hh_le_r,
315
+ gen_helper_sve_st2ss_le_r,
316
+ gen_helper_sve_st2dd_le_r },
317
+ { gen_helper_sve_st3bb_r,
318
+ gen_helper_sve_st3hh_le_r,
319
+ gen_helper_sve_st3ss_le_r,
320
+ gen_helper_sve_st3dd_le_r },
321
+ { gen_helper_sve_st4bb_r,
322
+ gen_helper_sve_st4hh_le_r,
323
+ gen_helper_sve_st4ss_le_r,
324
+ gen_helper_sve_st4dd_le_r } },
325
+ { { gen_helper_sve_st2bb_r,
326
+ gen_helper_sve_st2hh_be_r,
327
+ gen_helper_sve_st2ss_be_r,
328
+ gen_helper_sve_st2dd_be_r },
329
+ { gen_helper_sve_st3bb_r,
330
+ gen_helper_sve_st3hh_be_r,
331
+ gen_helper_sve_st3ss_be_r,
332
+ gen_helper_sve_st3dd_be_r },
333
+ { gen_helper_sve_st4bb_r,
334
+ gen_helper_sve_st4hh_be_r,
335
+ gen_helper_sve_st4ss_be_r,
336
+ gen_helper_sve_st4dd_be_r } } },
337
+ { { { gen_helper_sve_st2bb_r_mte,
338
+ gen_helper_sve_st2hh_le_r_mte,
339
+ gen_helper_sve_st2ss_le_r_mte,
340
+ gen_helper_sve_st2dd_le_r_mte },
341
+ { gen_helper_sve_st3bb_r_mte,
342
+ gen_helper_sve_st3hh_le_r_mte,
343
+ gen_helper_sve_st3ss_le_r_mte,
344
+ gen_helper_sve_st3dd_le_r_mte },
345
+ { gen_helper_sve_st4bb_r_mte,
346
+ gen_helper_sve_st4hh_le_r_mte,
347
+ gen_helper_sve_st4ss_le_r_mte,
348
+ gen_helper_sve_st4dd_le_r_mte } },
349
+ { { gen_helper_sve_st2bb_r_mte,
350
+ gen_helper_sve_st2hh_be_r_mte,
351
+ gen_helper_sve_st2ss_be_r_mte,
352
+ gen_helper_sve_st2dd_be_r_mte },
353
+ { gen_helper_sve_st3bb_r_mte,
354
+ gen_helper_sve_st3hh_be_r_mte,
355
+ gen_helper_sve_st3ss_be_r_mte,
356
+ gen_helper_sve_st3dd_be_r_mte },
357
+ { gen_helper_sve_st4bb_r_mte,
358
+ gen_helper_sve_st4hh_be_r_mte,
359
+ gen_helper_sve_st4ss_be_r_mte,
360
+ gen_helper_sve_st4dd_be_r_mte } } },
361
};
362
gen_helper_gvec_mem *fn;
363
int be = s->be_data == MO_BE;
364
365
if (nreg == 0) {
366
/* ST1 */
367
- fn = fn_single[be][msz][esz];
368
+ fn = fn_single[s->mte_active[0]][be][msz][esz];
369
+ nreg = 1;
370
} else {
371
/* ST2, ST3, ST4 -- msz == esz, enforced by encoding */
372
assert(msz == esz);
373
- fn = fn_multiple[be][nreg - 1][msz];
374
+ fn = fn_multiple[s->mte_active[0]][be][nreg - 1][msz];
375
}
376
assert(fn != NULL);
377
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
378
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), nreg, true, fn);
379
}
380
381
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
382
--
383
2.20.1
384
385
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-36-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-sve.h | 98 ++++++++++++++++
12
target/arm/sve_helper.c | 99 ++++++++++++++--
13
target/arm/translate-sve.c | 232 +++++++++++++++++++++++++------------
14
3 files changed, 343 insertions(+), 86 deletions(-)
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_ldff1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_ldff1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_ldff1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_ldff1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+DEF_HELPER_FLAGS_4(sve_ldff1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_4(sve_ldff1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_ldff1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+
32
+DEF_HELPER_FLAGS_4(sve_ldff1hh_le_r_mte, TCG_CALL_NO_WG,
33
+ void, env, ptr, tl, i32)
34
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_le_r_mte, TCG_CALL_NO_WG,
35
+ void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_le_r_mte, TCG_CALL_NO_WG,
37
+ void, env, ptr, tl, i32)
38
+DEF_HELPER_FLAGS_4(sve_ldff1hss_le_r_mte, TCG_CALL_NO_WG,
39
+ void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ldff1hds_le_r_mte, TCG_CALL_NO_WG,
41
+ void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ldff1hh_be_r_mte, TCG_CALL_NO_WG,
44
+ void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_be_r_mte, TCG_CALL_NO_WG,
46
+ void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_be_r_mte, TCG_CALL_NO_WG,
48
+ void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ldff1hss_be_r_mte, TCG_CALL_NO_WG,
50
+ void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ldff1hds_be_r_mte, TCG_CALL_NO_WG,
52
+ void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_ldff1ss_le_r_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_le_r_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, tl, i32)
58
+DEF_HELPER_FLAGS_4(sve_ldff1sds_le_r_mte, TCG_CALL_NO_WG,
59
+ void, env, ptr, tl, i32)
60
+
61
+DEF_HELPER_FLAGS_4(sve_ldff1ss_be_r_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_be_r_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r_mte, TCG_CALL_NO_WG,
69
+ void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r_mte, TCG_CALL_NO_WG,
71
+ void, env, ptr, tl, i32)
72
+
73
DEF_HELPER_FLAGS_4(sve_ldnf1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
80
+DEF_HELPER_FLAGS_4(sve_ldnf1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ldnf1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_4(sve_ldnf1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_4(sve_ldnf1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ldnf1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+
88
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_le_r_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_le_r_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_le_r_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_le_r_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_le_r_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, tl, i32)
98
+
99
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_be_r_mte, TCG_CALL_NO_WG,
100
+ void, env, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_be_r_mte, TCG_CALL_NO_WG,
102
+ void, env, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_be_r_mte, TCG_CALL_NO_WG,
104
+ void, env, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_be_r_mte, TCG_CALL_NO_WG,
106
+ void, env, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_be_r_mte, TCG_CALL_NO_WG,
108
+ void, env, ptr, tl, i32)
109
+
110
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_le_r_mte, TCG_CALL_NO_WG,
111
+ void, env, ptr, tl, i32)
112
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_le_r_mte, TCG_CALL_NO_WG,
113
+ void, env, ptr, tl, i32)
114
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_le_r_mte, TCG_CALL_NO_WG,
115
+ void, env, ptr, tl, i32)
116
+
117
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_be_r_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_be_r_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, tl, i32)
123
+
124
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r_mte, TCG_CALL_NO_WG,
125
+ void, env, ptr, tl, i32)
126
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r_mte, TCG_CALL_NO_WG,
127
+ void, env, ptr, tl, i32)
128
+
129
DEF_HELPER_FLAGS_4(sve_st1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
130
DEF_HELPER_FLAGS_4(sve_st2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
131
DEF_HELPER_FLAGS_4(sve_st3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
132
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/target/arm/sve_helper.c
135
+++ b/target/arm/sve_helper.c
136
@@ -XXX,XX +XXX,XX @@ static void record_fault(CPUARMState *env, uintptr_t i, uintptr_t oprsz)
137
*/
138
static inline QEMU_ALWAYS_INLINE
139
void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
140
- uint32_t desc, const uintptr_t retaddr,
141
+ uint32_t desc, const uintptr_t retaddr, uint32_t mtedesc,
142
const int esz, const int msz, const SVEContFault fault,
143
sve_ldst1_host_fn *host_fn,
144
sve_ldst1_tlb_fn *tlb_fn)
145
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
146
mem_off = info.mem_off_first[0];
147
flags = info.page[0].flags;
148
149
+ /*
150
+ * Disable MTE checking if the Tagged bit is not set. Since TBI must
151
+ * be set within MTEDESC for MTE, !mtedesc => !mte_active.
152
+ */
153
+ if (arm_tlb_mte_tagged(&info.page[0].attrs)) {
154
+ mtedesc = 0;
155
+ }
156
+
157
if (fault == FAULT_FIRST) {
158
+ /* Trapping mte check for the first-fault element. */
159
+ if (mtedesc) {
160
+ mte_check1(env, mtedesc, addr + mem_off, retaddr);
161
+ }
162
+
163
/*
164
* Special handling of the first active element,
165
* if it crosses a page boundary or is MMIO.
166
*/
167
bool is_split = mem_off == info.mem_off_split;
168
- /* TODO: MTE check. */
169
if (unlikely(flags != 0) || unlikely(is_split)) {
170
/*
171
* Use the slow path for cross-page handling.
172
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
173
/* Watchpoint hit, see below. */
174
goto do_fault;
175
}
176
- /* TODO: MTE check. */
177
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
178
+ goto do_fault;
179
+ }
180
/*
181
* Use the slow path for cross-page handling.
182
* This is RAM, without a watchpoint, and will not trap.
183
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
184
& BP_MEM_READ)) {
185
goto do_fault;
186
}
187
- /* TODO: MTE check. */
188
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
189
+ goto do_fault;
190
+ }
191
host_fn(vd, reg_off, host + mem_off);
192
}
193
reg_off += 1 << esz;
194
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
195
record_fault(env, reg_off, reg_max);
196
}
197
198
-#define DO_LDFF1_LDNF1_1(PART, ESZ) \
199
+static inline QEMU_ALWAYS_INLINE
200
+void sve_ldnfff1_r_mte(CPUARMState *env, void *vg, target_ulong addr,
201
+ uint32_t desc, const uintptr_t retaddr,
202
+ const int esz, const int msz, const SVEContFault fault,
203
+ sve_ldst1_host_fn *host_fn,
204
+ sve_ldst1_tlb_fn *tlb_fn)
205
+{
206
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
207
+ int bit55 = extract64(addr, 55, 1);
208
+
209
+ /* Remove mtedesc from the normal sve descriptor. */
210
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
211
+
212
+ /* Perform gross MTE suppression early. */
213
+ if (!tbi_check(desc, bit55) ||
214
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
215
+ mtedesc = 0;
216
+ }
217
+
218
+ sve_ldnfff1_r(env, vg, addr, desc, retaddr, mtedesc,
219
+ esz, msz, fault, host_fn, tlb_fn);
220
+}
221
+
222
+#define DO_LDFF1_LDNF1_1(PART, ESZ) \
223
void HELPER(sve_ldff1##PART##_r)(CPUARMState *env, void *vg, \
224
target_ulong addr, uint32_t desc) \
225
{ \
226
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
227
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_FIRST, \
228
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
229
} \
230
void HELPER(sve_ldnf1##PART##_r)(CPUARMState *env, void *vg, \
231
target_ulong addr, uint32_t desc) \
232
{ \
233
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
234
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_NO, \
235
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
236
+} \
237
+void HELPER(sve_ldff1##PART##_r_mte)(CPUARMState *env, void *vg, \
238
+ target_ulong addr, uint32_t desc) \
239
+{ \
240
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
241
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
242
+} \
243
+void HELPER(sve_ldnf1##PART##_r_mte)(CPUARMState *env, void *vg, \
244
+ target_ulong addr, uint32_t desc) \
245
+{ \
246
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
247
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
248
}
249
250
-#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
251
+#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
252
void HELPER(sve_ldff1##PART##_le_r)(CPUARMState *env, void *vg, \
253
target_ulong addr, uint32_t desc) \
254
{ \
255
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
256
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
257
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
258
} \
259
void HELPER(sve_ldnf1##PART##_le_r)(CPUARMState *env, void *vg, \
260
target_ulong addr, uint32_t desc) \
261
{ \
262
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
263
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
264
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
265
} \
266
void HELPER(sve_ldff1##PART##_be_r)(CPUARMState *env, void *vg, \
267
target_ulong addr, uint32_t desc) \
268
{ \
269
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
270
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
271
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
272
} \
273
void HELPER(sve_ldnf1##PART##_be_r)(CPUARMState *env, void *vg, \
274
target_ulong addr, uint32_t desc) \
275
{ \
276
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
277
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
278
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
279
+} \
280
+void HELPER(sve_ldff1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
284
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
285
+} \
286
+void HELPER(sve_ldnf1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
287
+ target_ulong addr, uint32_t desc) \
288
+{ \
289
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
290
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
291
+} \
292
+void HELPER(sve_ldff1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
293
+ target_ulong addr, uint32_t desc) \
294
+{ \
295
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
296
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
297
+} \
298
+void HELPER(sve_ldnf1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
299
+ target_ulong addr, uint32_t desc) \
300
+{ \
301
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
302
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
303
}
304
305
DO_LDFF1_LDNF1_1(bb, MO_8)
306
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
307
index XXXXXXX..XXXXXXX 100644
308
--- a/target/arm/translate-sve.c
309
+++ b/target/arm/translate-sve.c
310
@@ -XXX,XX +XXX,XX @@ static bool trans_LD_zpri(DisasContext *s, arg_rpri_load *a)
311
312
static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
313
{
314
- static gen_helper_gvec_mem * const fns[2][16] = {
315
- /* Little-endian */
316
- { gen_helper_sve_ldff1bb_r,
317
- gen_helper_sve_ldff1bhu_r,
318
- gen_helper_sve_ldff1bsu_r,
319
- gen_helper_sve_ldff1bdu_r,
320
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
321
+ { /* mte inactive, little-endian */
322
+ { gen_helper_sve_ldff1bb_r,
323
+ gen_helper_sve_ldff1bhu_r,
324
+ gen_helper_sve_ldff1bsu_r,
325
+ gen_helper_sve_ldff1bdu_r,
326
327
- gen_helper_sve_ldff1sds_le_r,
328
- gen_helper_sve_ldff1hh_le_r,
329
- gen_helper_sve_ldff1hsu_le_r,
330
- gen_helper_sve_ldff1hdu_le_r,
331
+ gen_helper_sve_ldff1sds_le_r,
332
+ gen_helper_sve_ldff1hh_le_r,
333
+ gen_helper_sve_ldff1hsu_le_r,
334
+ gen_helper_sve_ldff1hdu_le_r,
335
336
- gen_helper_sve_ldff1hds_le_r,
337
- gen_helper_sve_ldff1hss_le_r,
338
- gen_helper_sve_ldff1ss_le_r,
339
- gen_helper_sve_ldff1sdu_le_r,
340
+ gen_helper_sve_ldff1hds_le_r,
341
+ gen_helper_sve_ldff1hss_le_r,
342
+ gen_helper_sve_ldff1ss_le_r,
343
+ gen_helper_sve_ldff1sdu_le_r,
344
345
- gen_helper_sve_ldff1bds_r,
346
- gen_helper_sve_ldff1bss_r,
347
- gen_helper_sve_ldff1bhs_r,
348
- gen_helper_sve_ldff1dd_le_r },
349
+ gen_helper_sve_ldff1bds_r,
350
+ gen_helper_sve_ldff1bss_r,
351
+ gen_helper_sve_ldff1bhs_r,
352
+ gen_helper_sve_ldff1dd_le_r },
353
354
- /* Big-endian */
355
- { gen_helper_sve_ldff1bb_r,
356
- gen_helper_sve_ldff1bhu_r,
357
- gen_helper_sve_ldff1bsu_r,
358
- gen_helper_sve_ldff1bdu_r,
359
+ /* mte inactive, big-endian */
360
+ { gen_helper_sve_ldff1bb_r,
361
+ gen_helper_sve_ldff1bhu_r,
362
+ gen_helper_sve_ldff1bsu_r,
363
+ gen_helper_sve_ldff1bdu_r,
364
365
- gen_helper_sve_ldff1sds_be_r,
366
- gen_helper_sve_ldff1hh_be_r,
367
- gen_helper_sve_ldff1hsu_be_r,
368
- gen_helper_sve_ldff1hdu_be_r,
369
+ gen_helper_sve_ldff1sds_be_r,
370
+ gen_helper_sve_ldff1hh_be_r,
371
+ gen_helper_sve_ldff1hsu_be_r,
372
+ gen_helper_sve_ldff1hdu_be_r,
373
374
- gen_helper_sve_ldff1hds_be_r,
375
- gen_helper_sve_ldff1hss_be_r,
376
- gen_helper_sve_ldff1ss_be_r,
377
- gen_helper_sve_ldff1sdu_be_r,
378
+ gen_helper_sve_ldff1hds_be_r,
379
+ gen_helper_sve_ldff1hss_be_r,
380
+ gen_helper_sve_ldff1ss_be_r,
381
+ gen_helper_sve_ldff1sdu_be_r,
382
383
- gen_helper_sve_ldff1bds_r,
384
- gen_helper_sve_ldff1bss_r,
385
- gen_helper_sve_ldff1bhs_r,
386
- gen_helper_sve_ldff1dd_be_r },
387
+ gen_helper_sve_ldff1bds_r,
388
+ gen_helper_sve_ldff1bss_r,
389
+ gen_helper_sve_ldff1bhs_r,
390
+ gen_helper_sve_ldff1dd_be_r } },
391
+
392
+ { /* mte active, little-endian */
393
+ { gen_helper_sve_ldff1bb_r_mte,
394
+ gen_helper_sve_ldff1bhu_r_mte,
395
+ gen_helper_sve_ldff1bsu_r_mte,
396
+ gen_helper_sve_ldff1bdu_r_mte,
397
+
398
+ gen_helper_sve_ldff1sds_le_r_mte,
399
+ gen_helper_sve_ldff1hh_le_r_mte,
400
+ gen_helper_sve_ldff1hsu_le_r_mte,
401
+ gen_helper_sve_ldff1hdu_le_r_mte,
402
+
403
+ gen_helper_sve_ldff1hds_le_r_mte,
404
+ gen_helper_sve_ldff1hss_le_r_mte,
405
+ gen_helper_sve_ldff1ss_le_r_mte,
406
+ gen_helper_sve_ldff1sdu_le_r_mte,
407
+
408
+ gen_helper_sve_ldff1bds_r_mte,
409
+ gen_helper_sve_ldff1bss_r_mte,
410
+ gen_helper_sve_ldff1bhs_r_mte,
411
+ gen_helper_sve_ldff1dd_le_r_mte },
412
+
413
+ /* mte active, big-endian */
414
+ { gen_helper_sve_ldff1bb_r_mte,
415
+ gen_helper_sve_ldff1bhu_r_mte,
416
+ gen_helper_sve_ldff1bsu_r_mte,
417
+ gen_helper_sve_ldff1bdu_r_mte,
418
+
419
+ gen_helper_sve_ldff1sds_be_r_mte,
420
+ gen_helper_sve_ldff1hh_be_r_mte,
421
+ gen_helper_sve_ldff1hsu_be_r_mte,
422
+ gen_helper_sve_ldff1hdu_be_r_mte,
423
+
424
+ gen_helper_sve_ldff1hds_be_r_mte,
425
+ gen_helper_sve_ldff1hss_be_r_mte,
426
+ gen_helper_sve_ldff1ss_be_r_mte,
427
+ gen_helper_sve_ldff1sdu_be_r_mte,
428
+
429
+ gen_helper_sve_ldff1bds_r_mte,
430
+ gen_helper_sve_ldff1bss_r_mte,
431
+ gen_helper_sve_ldff1bhs_r_mte,
432
+ gen_helper_sve_ldff1dd_be_r_mte } },
433
};
434
435
if (sve_access_check(s)) {
436
TCGv_i64 addr = new_tmp_a64(s);
437
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
438
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
439
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
440
- fns[s->be_data == MO_BE][a->dtype]);
441
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
442
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
443
}
444
return true;
445
}
446
447
static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
448
{
449
- static gen_helper_gvec_mem * const fns[2][16] = {
450
- /* Little-endian */
451
- { gen_helper_sve_ldnf1bb_r,
452
- gen_helper_sve_ldnf1bhu_r,
453
- gen_helper_sve_ldnf1bsu_r,
454
- gen_helper_sve_ldnf1bdu_r,
455
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
456
+ { /* mte inactive, little-endian */
457
+ { gen_helper_sve_ldnf1bb_r,
458
+ gen_helper_sve_ldnf1bhu_r,
459
+ gen_helper_sve_ldnf1bsu_r,
460
+ gen_helper_sve_ldnf1bdu_r,
461
462
- gen_helper_sve_ldnf1sds_le_r,
463
- gen_helper_sve_ldnf1hh_le_r,
464
- gen_helper_sve_ldnf1hsu_le_r,
465
- gen_helper_sve_ldnf1hdu_le_r,
466
+ gen_helper_sve_ldnf1sds_le_r,
467
+ gen_helper_sve_ldnf1hh_le_r,
468
+ gen_helper_sve_ldnf1hsu_le_r,
469
+ gen_helper_sve_ldnf1hdu_le_r,
470
471
- gen_helper_sve_ldnf1hds_le_r,
472
- gen_helper_sve_ldnf1hss_le_r,
473
- gen_helper_sve_ldnf1ss_le_r,
474
- gen_helper_sve_ldnf1sdu_le_r,
475
+ gen_helper_sve_ldnf1hds_le_r,
476
+ gen_helper_sve_ldnf1hss_le_r,
477
+ gen_helper_sve_ldnf1ss_le_r,
478
+ gen_helper_sve_ldnf1sdu_le_r,
479
480
- gen_helper_sve_ldnf1bds_r,
481
- gen_helper_sve_ldnf1bss_r,
482
- gen_helper_sve_ldnf1bhs_r,
483
- gen_helper_sve_ldnf1dd_le_r },
484
+ gen_helper_sve_ldnf1bds_r,
485
+ gen_helper_sve_ldnf1bss_r,
486
+ gen_helper_sve_ldnf1bhs_r,
487
+ gen_helper_sve_ldnf1dd_le_r },
488
489
- /* Big-endian */
490
- { gen_helper_sve_ldnf1bb_r,
491
- gen_helper_sve_ldnf1bhu_r,
492
- gen_helper_sve_ldnf1bsu_r,
493
- gen_helper_sve_ldnf1bdu_r,
494
+ /* mte inactive, big-endian */
495
+ { gen_helper_sve_ldnf1bb_r,
496
+ gen_helper_sve_ldnf1bhu_r,
497
+ gen_helper_sve_ldnf1bsu_r,
498
+ gen_helper_sve_ldnf1bdu_r,
499
500
- gen_helper_sve_ldnf1sds_be_r,
501
- gen_helper_sve_ldnf1hh_be_r,
502
- gen_helper_sve_ldnf1hsu_be_r,
503
- gen_helper_sve_ldnf1hdu_be_r,
504
+ gen_helper_sve_ldnf1sds_be_r,
505
+ gen_helper_sve_ldnf1hh_be_r,
506
+ gen_helper_sve_ldnf1hsu_be_r,
507
+ gen_helper_sve_ldnf1hdu_be_r,
508
509
- gen_helper_sve_ldnf1hds_be_r,
510
- gen_helper_sve_ldnf1hss_be_r,
511
- gen_helper_sve_ldnf1ss_be_r,
512
- gen_helper_sve_ldnf1sdu_be_r,
513
+ gen_helper_sve_ldnf1hds_be_r,
514
+ gen_helper_sve_ldnf1hss_be_r,
515
+ gen_helper_sve_ldnf1ss_be_r,
516
+ gen_helper_sve_ldnf1sdu_be_r,
517
518
- gen_helper_sve_ldnf1bds_r,
519
- gen_helper_sve_ldnf1bss_r,
520
- gen_helper_sve_ldnf1bhs_r,
521
- gen_helper_sve_ldnf1dd_be_r },
522
+ gen_helper_sve_ldnf1bds_r,
523
+ gen_helper_sve_ldnf1bss_r,
524
+ gen_helper_sve_ldnf1bhs_r,
525
+ gen_helper_sve_ldnf1dd_be_r } },
526
+
527
+ { /* mte inactive, little-endian */
528
+ { gen_helper_sve_ldnf1bb_r_mte,
529
+ gen_helper_sve_ldnf1bhu_r_mte,
530
+ gen_helper_sve_ldnf1bsu_r_mte,
531
+ gen_helper_sve_ldnf1bdu_r_mte,
532
+
533
+ gen_helper_sve_ldnf1sds_le_r_mte,
534
+ gen_helper_sve_ldnf1hh_le_r_mte,
535
+ gen_helper_sve_ldnf1hsu_le_r_mte,
536
+ gen_helper_sve_ldnf1hdu_le_r_mte,
537
+
538
+ gen_helper_sve_ldnf1hds_le_r_mte,
539
+ gen_helper_sve_ldnf1hss_le_r_mte,
540
+ gen_helper_sve_ldnf1ss_le_r_mte,
541
+ gen_helper_sve_ldnf1sdu_le_r_mte,
542
+
543
+ gen_helper_sve_ldnf1bds_r_mte,
544
+ gen_helper_sve_ldnf1bss_r_mte,
545
+ gen_helper_sve_ldnf1bhs_r_mte,
546
+ gen_helper_sve_ldnf1dd_le_r_mte },
547
+
548
+ /* mte inactive, big-endian */
549
+ { gen_helper_sve_ldnf1bb_r_mte,
550
+ gen_helper_sve_ldnf1bhu_r_mte,
551
+ gen_helper_sve_ldnf1bsu_r_mte,
552
+ gen_helper_sve_ldnf1bdu_r_mte,
553
+
554
+ gen_helper_sve_ldnf1sds_be_r_mte,
555
+ gen_helper_sve_ldnf1hh_be_r_mte,
556
+ gen_helper_sve_ldnf1hsu_be_r_mte,
557
+ gen_helper_sve_ldnf1hdu_be_r_mte,
558
+
559
+ gen_helper_sve_ldnf1hds_be_r_mte,
560
+ gen_helper_sve_ldnf1hss_be_r_mte,
561
+ gen_helper_sve_ldnf1ss_be_r_mte,
562
+ gen_helper_sve_ldnf1sdu_be_r_mte,
563
+
564
+ gen_helper_sve_ldnf1bds_r_mte,
565
+ gen_helper_sve_ldnf1bss_r_mte,
566
+ gen_helper_sve_ldnf1bhs_r_mte,
567
+ gen_helper_sve_ldnf1dd_be_r_mte } },
568
};
569
570
if (sve_access_check(s)) {
571
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
572
TCGv_i64 addr = new_tmp_a64(s);
573
574
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
575
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
576
- fns[s->be_data == MO_BE][a->dtype]);
577
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
578
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
579
}
580
return true;
581
}
582
--
583
2.20.1
584
585
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
We still need to handle tbi for user-only when mte is inactive.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-37-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/translate-a64.h | 1 +
11
target/arm/translate-a64.c | 2 +-
12
target/arm/translate-sve.c | 6 ++++--
13
3 files changed, 6 insertions(+), 3 deletions(-)
14
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate-a64.h
18
+++ b/target/arm/translate-a64.h
19
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
21
unsigned int imms, unsigned int immr);
22
bool sve_access_check(DisasContext *s);
23
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
24
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
25
bool tag_checked, int log2_size);
26
TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
27
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-a64.c
30
+++ b/target/arm/translate-a64.c
31
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
32
* of the write-back address.
33
*/
34
35
-static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
36
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
37
{
38
TCGv_i64 clean = new_tmp_a64(s);
39
#ifdef CONFIG_USER_ONLY
40
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-sve.c
43
+++ b/target/arm/translate-sve.c
44
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
45
* For e.g. LD4, there are not enough arguments to pass all 4
46
* registers as pointers, so encode the regno into the data field.
47
* For consistency, do this even for LD1.
48
- * TODO: mte_n check here while callers are updated.
49
*/
50
- if (mte_n && s->mte_active[0]) {
51
+ if (s->mte_active[0]) {
52
int msz = dtype_msz(dtype);
53
54
desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
55
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
56
desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
57
desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
58
desc <<= SVE_MTEDESC_SHIFT;
59
+ } else {
60
+ addr = clean_data_tbi(s, addr);
61
}
62
+
63
desc = simd_desc(vsz, vsz, zt | desc);
64
t_desc = tcg_const_i32(desc);
65
t_pg = tcg_temp_new_ptr();
66
--
67
2.20.1
68
69
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Look up the physical address for the given virtual address,
4
convert that to a tag physical address, and finally return
5
the host address that backs it.
6
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-45-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/mte_helper.c | 131 ++++++++++++++++++++++++++++++++++++++++
13
1 file changed, 131 insertions(+)
14
15
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/mte_helper.c
18
+++ b/target/arm/mte_helper.c
19
@@ -XXX,XX +XXX,XX @@
20
#include "cpu.h"
21
#include "internals.h"
22
#include "exec/exec-all.h"
23
+#include "exec/ram_addr.h"
24
#include "exec/cpu_ldst.h"
25
#include "exec/helper-proto.h"
26
27
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
28
int ptr_size, MMUAccessType tag_access,
29
int tag_size, uintptr_t ra)
30
{
31
+#ifdef CONFIG_USER_ONLY
32
/* Tag storage not implemented. */
33
return NULL;
34
+#else
35
+ uintptr_t index;
36
+ CPUIOTLBEntry *iotlbentry;
37
+ int in_page, flags;
38
+ ram_addr_t ptr_ra;
39
+ hwaddr ptr_paddr, tag_paddr, xlat;
40
+ MemoryRegion *mr;
41
+ ARMASIdx tag_asi;
42
+ AddressSpace *tag_as;
43
+ void *host;
44
+
45
+ /*
46
+ * Probe the first byte of the virtual address. This raises an
47
+ * exception for inaccessible pages, and resolves the virtual address
48
+ * into the softmmu tlb.
49
+ *
50
+ * When RA == 0, this is for mte_probe1. The page is expected to be
51
+ * valid. Indicate to probe_access_flags no-fault, then assert that
52
+ * we received a valid page.
53
+ */
54
+ flags = probe_access_flags(env, ptr, ptr_access, ptr_mmu_idx,
55
+ ra == 0, &host, ra);
56
+ assert(!(flags & TLB_INVALID_MASK));
57
+
58
+ /*
59
+ * Find the iotlbentry for ptr. This *must* be present in the TLB
60
+ * because we just found the mapping.
61
+ * TODO: Perhaps there should be a cputlb helper that returns a
62
+ * matching tlb entry + iotlb entry.
63
+ */
64
+ index = tlb_index(env, ptr_mmu_idx, ptr);
65
+# ifdef CONFIG_DEBUG_TCG
66
+ {
67
+ CPUTLBEntry *entry = tlb_entry(env, ptr_mmu_idx, ptr);
68
+ target_ulong comparator = (ptr_access == MMU_DATA_LOAD
69
+ ? entry->addr_read
70
+ : tlb_addr_write(entry));
71
+ g_assert(tlb_hit(comparator, ptr));
72
+ }
73
+# endif
74
+ iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
75
+
76
+ /* If the virtual page MemAttr != Tagged, access unchecked. */
77
+ if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
78
+ return NULL;
79
+ }
80
+
81
+ /*
82
+ * If not backed by host ram, there is no tag storage: access unchecked.
83
+ * This is probably a guest os bug though, so log it.
84
+ */
85
+ if (unlikely(flags & TLB_MMIO)) {
86
+ qemu_log_mask(LOG_GUEST_ERROR,
87
+ "Page @ 0x%" PRIx64 " indicates Tagged Normal memory "
88
+ "but is not backed by host ram\n", ptr);
89
+ return NULL;
90
+ }
91
+
92
+ /*
93
+ * The Normal memory access can extend to the next page. E.g. a single
94
+ * 8-byte access to the last byte of a page will check only the last
95
+ * tag on the first page.
96
+ * Any page access exception has priority over tag check exception.
97
+ */
98
+ in_page = -(ptr | TARGET_PAGE_MASK);
99
+ if (unlikely(ptr_size > in_page)) {
100
+ void *ignore;
101
+ flags |= probe_access_flags(env, ptr + in_page, ptr_access,
102
+ ptr_mmu_idx, ra == 0, &ignore, ra);
103
+ assert(!(flags & TLB_INVALID_MASK));
104
+ }
105
+
106
+ /* Any debug exception has priority over a tag check exception. */
107
+ if (unlikely(flags & TLB_WATCHPOINT)) {
108
+ int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
109
+ assert(ra != 0);
110
+ cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
111
+ iotlbentry->attrs, wp, ra);
112
+ }
113
+
114
+ /*
115
+ * Find the physical address within the normal mem space.
116
+ * The memory region lookup must succeed because TLB_MMIO was
117
+ * not set in the cputlb lookup above.
118
+ */
119
+ mr = memory_region_from_host(host, &ptr_ra);
120
+ tcg_debug_assert(mr != NULL);
121
+ tcg_debug_assert(memory_region_is_ram(mr));
122
+ ptr_paddr = ptr_ra;
123
+ do {
124
+ ptr_paddr += mr->addr;
125
+ mr = mr->container;
126
+ } while (mr);
127
+
128
+ /* Convert to the physical address in tag space. */
129
+ tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
130
+
131
+ /* Look up the address in tag space. */
132
+ tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
133
+ tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
134
+ mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
135
+ tag_access == MMU_DATA_STORE,
136
+ iotlbentry->attrs);
137
+
138
+ /*
139
+ * Note that @mr will never be NULL. If there is nothing in the address
140
+ * space at @tag_paddr, the translation will return the unallocated memory
141
+ * region. For our purposes, the result must be ram.
142
+ */
143
+ if (unlikely(!memory_region_is_ram(mr))) {
144
+ /* ??? Failure is a board configuration error. */
145
+ qemu_log_mask(LOG_UNIMP,
146
+ "Tag Memory @ 0x%" HWADDR_PRIx " not found for "
147
+ "Normal Memory @ 0x%" HWADDR_PRIx "\n",
148
+ tag_paddr, ptr_paddr);
149
+ return NULL;
150
+ }
151
+
152
+ /*
153
+ * Ensure the tag memory is dirty on write, for migration.
154
+ * Tag memory can never contain code or display memory (vga).
155
+ */
156
+ if (tag_access == MMU_DATA_STORE) {
157
+ ram_addr_t tag_ra = memory_region_get_ram_addr(mr) + xlat;
158
+ cpu_physical_memory_set_dirty_flag(tag_ra, DIRTY_MEMORY_MIGRATION);
159
+ }
160
+
161
+ return memory_region_get_ram_ptr(mr) + xlat;
162
+#endif
163
}
164
165
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
166
--
167
2.20.1
168
169
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
We now implement all of the components of MTE, without actually
4
supporting any tagged memory. All MTE instructions will work,
5
trivially, so we can enable support.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-46-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/cpu64.c | 5 +++++
13
1 file changed, 5 insertions(+)
14
15
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu64.c
18
+++ b/target/arm/cpu64.c
19
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
20
21
t = cpu->isar.id_aa64pfr1;
22
t = FIELD_DP64(t, ID_AA64PFR1, BT, 1);
23
+ /*
24
+ * Begin with full support for MTE; will be downgraded to MTE=1
25
+ * during realize if the board provides no tag memory.
26
+ */
27
+ t = FIELD_DP64(t, ID_AA64PFR1, MTE, 2);
28
cpu->isar.id_aa64pfr1 = t;
29
30
t = cpu->isar.id_aa64mmfr1;
31
--
32
2.20.1
33
34
diff view generated by jsdifflib