1
Most of this is the Neon decodetree patches, followed by Edgar's versal cleanups.
1
Mostly this is RTH's memtag series, but there are also some cleanups
2
from Philippe.
2
3
3
thanks
4
thanks
4
-- PMM
5
-- PMM
5
6
7
The following changes since commit 10f7ffabf9c507fc02382b89912003b1c43c3231:
6
8
7
The following changes since commit 2ef486e76d64436be90f7359a3071fb2a56ce835:
9
Merge remote-tracking branch 'remotes/mcayland/tags/qemu-macppc-20200626' into staging (2020-06-26 12:14:18 +0100)
8
9
Merge remote-tracking branch 'remotes/marcel/tags/rdma-pull-request' into staging (2020-05-03 14:12:56 +0100)
10
10
11
are available in the Git repository at:
11
are available in the Git repository at:
12
12
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200504
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200626
14
14
15
for you to fetch changes up to 9aefc6cf9b73f66062d2f914a0136756e7a28211:
15
for you to fetch changes up to c7459633baa71d1781fde4a245d6ec9ce2f008cf:
16
16
17
target/arm: Move gen_ function typedefs to translate.h (2020-05-04 12:59:26 +0100)
17
target/arm: Enable MTE (2020-06-26 14:32:24 +0100)
18
18
19
----------------------------------------------------------------
19
----------------------------------------------------------------
20
target-arm queue:
20
target-arm queue:
21
* Start of conversion of Neon insns to decodetree
21
* hw/arm/aspeed: improve QOM usage
22
* versal board: support SD and RTC
22
* hw/misc/pca9552: trace GPIO change events
23
* Implement ARMv8.2-TTS2UXN
23
* target/arm: Implement ARMv8.5-MemTag for system emulation
24
* Make VQDMULL undefined when U=1
25
* Some minor code cleanups
26
24
27
----------------------------------------------------------------
25
----------------------------------------------------------------
28
Edgar E. Iglesias (11):
26
Philippe Mathieu-Daudé (12):
29
hw/arm: versal: Remove inclusion of arm_gicv3_common.h
27
hw/arm/aspeed: Remove extraneous MemoryRegion object owner
30
hw/arm: versal: Move misplaced comment
28
hw/arm/aspeed: Rename AspeedBoardState as AspeedMachineState
31
hw/arm: versal-virt: Fix typo xlnx-ve -> xlnx-versal
29
hw/arm/aspeed: QOM'ify AspeedMachineState
32
hw/arm: versal: Embed the UARTs into the SoC type
30
hw/i2c/core: Add i2c_try_create_slave() and i2c_realize_and_unref()
33
hw/arm: versal: Embed the GEMs into the SoC type
31
hw/misc/pca9552: Rename 'nr_leds' as 'pin_count'
34
hw/arm: versal: Embed the ADMAs into the SoC type
32
hw/misc/pca9552: Rename generic code as pca955x
35
hw/arm: versal: Embed the APUs into the SoC type
33
hw/misc/pca9552: Add generic PCA955xClass, parent of TYPE_PCA9552
36
hw/arm: versal: Add support for SD
34
hw/misc/pca9552: Add a 'description' property for debugging purpose
37
hw/arm: versal: Add support for the RTC
35
hw/misc/pca9552: Trace GPIO High/Low events
38
hw/arm: versal-virt: Add support for SD
36
hw/arm/aspeed: Describe each PCA9552 device
39
hw/arm: versal-virt: Add support for the RTC
37
hw/misc/pca9552: Trace GPIO change events
38
hw/misc/pca9552: Model qdev output GPIOs
40
39
41
Fredrik Strupe (1):
40
Richard Henderson (45):
42
target/arm: Make VQDMULL undefined when U=1
41
target/arm: Add isar tests for mte
42
target/arm: Improve masking of SCR RES0 bits
43
target/arm: Add support for MTE to SCTLR_ELx
44
target/arm: Add support for MTE to HCR_EL2 and SCR_EL3
45
target/arm: Rename DISAS_UPDATE to DISAS_UPDATE_EXIT
46
target/arm: Add DISAS_UPDATE_NOCHAIN
47
target/arm: Add MTE system registers
48
target/arm: Add MTE bits to tb_flags
49
target/arm: Implement the IRG instruction
50
target/arm: Revise decoding for disas_add_sub_imm
51
target/arm: Implement the ADDG, SUBG instructions
52
target/arm: Implement the GMI instruction
53
target/arm: Implement the SUBP instruction
54
target/arm: Define arm_cpu_do_unaligned_access for user-only
55
target/arm: Implement LDG, STG, ST2G instructions
56
target/arm: Implement the STGP instruction
57
target/arm: Restrict the values of DCZID.BS under TCG
58
target/arm: Simplify DC_ZVA
59
target/arm: Implement the LDGM, STGM, STZGM instructions
60
target/arm: Implement the access tag cache flushes
61
target/arm: Move regime_el to internals.h
62
target/arm: Move regime_tcr to internals.h
63
target/arm: Add gen_mte_check1
64
target/arm: Add gen_mte_checkN
65
target/arm: Implement helper_mte_check1
66
target/arm: Implement helper_mte_checkN
67
target/arm: Add helper_mte_check_zva
68
target/arm: Use mte_checkN for sve unpredicated loads
69
target/arm: Use mte_checkN for sve unpredicated stores
70
target/arm: Use mte_check1 for sve LD1R
71
target/arm: Tidy trans_LD1R_zpri
72
target/arm: Add arm_tlb_bti_gp
73
target/arm: Add mte helpers for sve scalar + int loads
74
target/arm: Add mte helpers for sve scalar + int stores
75
target/arm: Add mte helpers for sve scalar + int ff/nf loads
76
target/arm: Handle TBI for sve scalar + int memory ops
77
target/arm: Add mte helpers for sve scatter/gather memory ops
78
target/arm: Complete TBI clearing for user-only for SVE
79
target/arm: Implement data cache set allocation tags
80
target/arm: Set PSTATE.TCO on exception entry
81
target/arm: Always pass cacheattr to get_phys_addr
82
target/arm: Cache the Tagged bit for a page in MemTxAttrs
83
target/arm: Create tagged ram when MTE is enabled
84
target/arm: Add allocation tag storage for system mode
85
target/arm: Enable MTE
43
86
44
Peter Maydell (25):
87
include/hw/arm/aspeed.h | 12 +-
45
target/arm: Don't use a TLB for ARMMMUIdx_Stage2
88
include/hw/i2c/i2c.h | 2 +
46
target/arm: Use enum constant in get_phys_addr_lpae() call
89
include/hw/misc/pca9552.h | 16 +-
47
target/arm: Add new 's1_is_el0' argument to get_phys_addr_lpae()
90
target/arm/cpu.h | 50 +-
48
target/arm: Implement ARMv8.2-TTS2UXN
91
target/arm/helper-a64.h | 16 +
49
target/arm: Use correct variable for setting 'max' cpu's ID_AA64DFR0
92
target/arm/helper-sve.h | 488 ++++++++++++++
50
target/arm/translate-vfp.inc.c: Remove duplicate simd_r32 check
93
target/arm/helper.h | 2 +
51
target/arm: Don't allow Thumb Neon insns without FEATURE_NEON
94
target/arm/internals.h | 153 ++++-
52
target/arm: Add stubs for AArch32 Neon decodetree
95
target/arm/translate-a64.h | 5 +
53
target/arm: Convert VCMLA (vector) to decodetree
96
target/arm/translate.h | 23 +-
54
target/arm: Convert VCADD (vector) to decodetree
97
hw/arm/aspeed.c | 46 +-
55
target/arm: Convert V[US]DOT (vector) to decodetree
98
hw/arm/virt.c | 55 +-
56
target/arm: Convert VFM[AS]L (vector) to decodetree
99
hw/i2c/core.c | 18 +-
57
target/arm: Convert VCMLA (scalar) to decodetree
100
hw/misc/pca9552.c | 216 +++++--
58
target/arm: Convert V[US]DOT (scalar) to decodetree
101
target/arm/cpu.c | 81 ++-
59
target/arm: Convert VFM[AS]L (scalar) to decodetree
102
target/arm/cpu64.c | 5 +
60
target/arm: Convert Neon load/store multiple structures to decodetree
103
target/arm/helper-a64.c | 94 +--
61
target/arm: Convert Neon 'load single structure to all lanes' to decodetree
104
target/arm/helper.c | 423 ++++++++++---
62
target/arm: Convert Neon 'load/store single structure' to decodetree
105
target/arm/m_helper.c | 11 +-
63
target/arm: Convert Neon 3-reg-same VADD/VSUB to decodetree
106
target/arm/mte_helper.c | 906 ++++++++++++++++++++++++++
64
target/arm: Convert Neon 3-reg-same logic ops to decodetree
107
target/arm/op_helper.c | 16 +
65
target/arm: Convert Neon 3-reg-same VMAX/VMIN to decodetree
108
target/arm/sve_helper.c | 616 ++++++++++++++----
66
target/arm: Convert Neon 3-reg-same comparisons to decodetree
109
target/arm/tlb_helper.c | 13 +-
67
target/arm: Convert Neon 3-reg-same VQADD/VQSUB to decodetree
110
target/arm/translate-a64.c | 657 ++++++++++++++++---
68
target/arm: Convert Neon 3-reg-same VMUL, VMLA, VMLS, VSHL to decodetree
111
target/arm/translate-sve.c | 1366 ++++++++++++++++++++++++++--------------
69
target/arm: Move gen_ function typedefs to translate.h
112
target/arm/translate-vfp.inc.c | 4 +-
113
target/arm/translate.c | 16 +-
114
hw/misc/trace-events | 4 +
115
target/arm/Makefile.objs | 1 +
116
29 files changed, 4391 insertions(+), 924 deletions(-)
117
create mode 100644 target/arm/mte_helper.c
70
118
71
Philippe Mathieu-Daudé (2):
72
hw/arm/mps2-tz: Use TYPE_IOTKIT instead of hardcoded string
73
target/arm: Use uint64_t for midr field in CPU state struct
74
75
include/hw/arm/xlnx-versal.h | 31 +-
76
target/arm/cpu-param.h | 2 +-
77
target/arm/cpu.h | 38 ++-
78
target/arm/translate-a64.h | 9 -
79
target/arm/translate.h | 26 ++
80
target/arm/neon-dp.decode | 86 +++++
81
target/arm/neon-ls.decode | 52 +++
82
target/arm/neon-shared.decode | 66 ++++
83
hw/arm/mps2-tz.c | 2 +-
84
hw/arm/xlnx-versal-virt.c | 74 ++++-
85
hw/arm/xlnx-versal.c | 115 +++++--
86
target/arm/cpu.c | 3 +-
87
target/arm/cpu64.c | 8 +-
88
target/arm/helper.c | 183 ++++------
89
target/arm/translate-a64.c | 17 -
90
target/arm/translate-neon.inc.c | 714 +++++++++++++++++++++++++++++++++++++++
91
target/arm/translate-vfp.inc.c | 6 -
92
target/arm/translate.c | 716 +++-------------------------------------
93
target/arm/Makefile.objs | 18 +
94
19 files changed, 1302 insertions(+), 864 deletions(-)
95
create mode 100644 target/arm/neon-dp.decode
96
create mode 100644 target/arm/neon-ls.decode
97
create mode 100644 target/arm/neon-shared.decode
98
create mode 100644 target/arm/translate-neon.inc.c
99
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
I'm confused by this code, 'bmc' is created as:
4
5
bmc = g_new0(AspeedBoardState, 1);
6
7
Then we use it as QOM owner for different MemoryRegion objects.
8
But looking at memory_region_init_ram (similarly for ROM):
9
10
void memory_region_init_ram(MemoryRegion *mr,
11
struct Object *owner,
12
const char *name,
13
uint64_t size,
14
Error **errp)
15
{
16
DeviceState *owner_dev;
17
Error *err = NULL;
18
19
memory_region_init_ram_nomigrate(mr, owner, name, size, &err);
20
if (err) {
21
error_propagate(errp, err);
22
return;
23
}
24
/* This will assert if owner is neither NULL nor a DeviceState.
25
* We only want the owner here for the purposes of defining a
26
* unique name for migration. TODO: Ideally we should implement
27
* a naming scheme for Objects which are not DeviceStates, in
28
* which case we can relax this restriction.
29
*/
30
owner_dev = DEVICE(owner);
31
vmstate_register_ram(mr, owner_dev);
32
}
33
34
The expected assertion is not triggered ('bmc' is not NULL neither
35
a DeviceState).
36
37
'bmc' structure is defined as:
38
39
struct AspeedBoardState {
40
AspeedSoCState soc;
41
MemoryRegion ram_container;
42
MemoryRegion max_ram;
43
};
44
45
What happens is when using 'OBJECT(bmc)', the QOM macros cast the
46
memory pointed by bmc, which first member is 'soc', which is
47
initialized ...:
48
49
object_initialize_child(OBJECT(machine), "soc",
50
&bmc->soc, amc->soc_name);
51
52
The 'soc' object is indeed a DeviceState, so the assertion passes.
53
54
Since this is fragile and only happens to work by luck, remove the
55
dangerous OBJECT(bmc) owner argument.
56
57
Note, this probably breaks migration for this machine.
58
59
Reviewed-by: Cédric Le Goater <clg@kaod.org>
60
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
61
Message-id: 20200623072132.2868-2-f4bug@amsat.org
62
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
63
---
64
hw/arm/aspeed.c | 6 +++---
65
1 file changed, 3 insertions(+), 3 deletions(-)
66
67
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/aspeed.c
70
+++ b/hw/arm/aspeed.c
71
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
72
* needed by the flash modules of the Aspeed machines.
73
*/
74
if (ASPEED_MACHINE(machine)->mmio_exec) {
75
- memory_region_init_alias(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
76
+ memory_region_init_alias(boot_rom, NULL, "aspeed.boot_rom",
77
&fl->mmio, 0, fl->size);
78
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
79
boot_rom);
80
} else {
81
- memory_region_init_rom(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
82
+ memory_region_init_rom(boot_rom, NULL, "aspeed.boot_rom",
83
fl->size, &error_abort);
84
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
85
boot_rom);
86
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
87
if (machine->kernel_filename && sc->num_cpus > 1) {
88
/* With no u-boot we must set up a boot stub for the secondary CPU */
89
MemoryRegion *smpboot = g_new(MemoryRegion, 1);
90
- memory_region_init_ram(smpboot, OBJECT(bmc), "aspeed.smpboot",
91
+ memory_region_init_ram(smpboot, NULL, "aspeed.smpboot",
92
0x80, &error_abort);
93
memory_region_add_subregion(get_system_memory(),
94
AST_SMP_MAILBOX_BASE, smpboot);
95
--
96
2.20.1
97
98
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
To have a more consistent naming, rename AspeedBoardState
4
as AspeedMachineState.
5
6
Suggested-by: Cédric Le Goater <clg@kaod.org>
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Message-id: 20200623072132.2868-3-f4bug@amsat.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
include/hw/arm/aspeed.h | 4 ++--
13
hw/arm/aspeed.c | 20 ++++++++++----------
14
2 files changed, 12 insertions(+), 12 deletions(-)
15
16
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/aspeed.h
19
+++ b/include/hw/arm/aspeed.h
20
@@ -XXX,XX +XXX,XX @@
21
22
#include "hw/boards.h"
23
24
-typedef struct AspeedBoardState AspeedBoardState;
25
+typedef struct AspeedMachineState AspeedMachineState;
26
27
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
28
#define ASPEED_MACHINE(obj) \
29
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineClass {
30
const char *spi_model;
31
uint32_t num_cs;
32
uint32_t macs_mask;
33
- void (*i2c_init)(AspeedBoardState *bmc);
34
+ void (*i2c_init)(AspeedMachineState *bmc);
35
} AspeedMachineClass;
36
37
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
.board_id = -1, /* device-tree-only board */
44
};
45
46
-struct AspeedBoardState {
47
+struct AspeedMachineState {
48
AspeedSoCState soc;
49
MemoryRegion ram_container;
50
MemoryRegion max_ram;
51
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
52
53
static void aspeed_machine_init(MachineState *machine)
54
{
55
- AspeedBoardState *bmc;
56
+ AspeedMachineState *bmc;
57
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
58
AspeedSoCClass *sc;
59
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
60
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
61
int i;
62
NICInfo *nd = &nd_table[0];
63
64
- bmc = g_new0(AspeedBoardState, 1);
65
+ bmc = g_new0(AspeedMachineState, 1);
66
67
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
68
4 * GiB);
69
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
70
arm_load_kernel(ARM_CPU(first_cpu), machine, &aspeed_board_binfo);
71
}
72
73
-static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
74
+static void palmetto_bmc_i2c_init(AspeedMachineState *bmc)
75
{
76
AspeedSoCState *soc = &bmc->soc;
77
DeviceState *dev;
78
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
79
object_property_set_int(OBJECT(dev), 110000, "temperature3", &error_abort);
80
}
81
82
-static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
83
+static void ast2500_evb_i2c_init(AspeedMachineState *bmc)
84
{
85
AspeedSoCState *soc = &bmc->soc;
86
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
87
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
88
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
89
}
90
91
-static void ast2600_evb_i2c_init(AspeedBoardState *bmc)
92
+static void ast2600_evb_i2c_init(AspeedMachineState *bmc)
93
{
94
/* Start with some devices on our I2C busses */
95
ast2500_evb_i2c_init(bmc);
96
}
97
98
-static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
99
+static void romulus_bmc_i2c_init(AspeedMachineState *bmc)
100
{
101
AspeedSoCState *soc = &bmc->soc;
102
103
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
104
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
105
}
106
107
-static void swift_bmc_i2c_init(AspeedBoardState *bmc)
108
+static void swift_bmc_i2c_init(AspeedMachineState *bmc)
109
{
110
AspeedSoCState *soc = &bmc->soc;
111
112
@@ -XXX,XX +XXX,XX @@ static void swift_bmc_i2c_init(AspeedBoardState *bmc)
113
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 12), "tmp105", 0x4a);
114
}
115
116
-static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
117
+static void sonorapass_bmc_i2c_init(AspeedMachineState *bmc)
118
{
119
AspeedSoCState *soc = &bmc->soc;
120
121
@@ -XXX,XX +XXX,XX @@ static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
122
123
}
124
125
-static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
126
+static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
127
{
128
AspeedSoCState *soc = &bmc->soc;
129
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
By using the TYPE_* definitions for devices, we can:
3
AspeedMachineState seems crippled. We use incorrectly 2
4
- quickly find where devices are used with 'git-grep'
4
different structures to do the same thing. Merge them
5
- easily rename a device (one-line change).
5
altogether:
6
- Move AspeedMachine fields to AspeedMachineState
7
- AspeedMachineState is now QOM
8
- Remove unused AspeedMachine structure
6
9
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20200428154650.21991-1-f4bug@amsat.org
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Message-id: 20200623072132.2868-4-f4bug@amsat.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
14
---
12
hw/arm/mps2-tz.c | 2 +-
15
include/hw/arm/aspeed.h | 8 +-------
13
1 file changed, 1 insertion(+), 1 deletion(-)
16
hw/arm/aspeed.c | 11 +++++++----
17
2 files changed, 8 insertions(+), 11 deletions(-)
14
18
15
diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c
19
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
16
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/arm/mps2-tz.c
21
--- a/include/hw/arm/aspeed.h
18
+++ b/hw/arm/mps2-tz.c
22
+++ b/include/hw/arm/aspeed.h
19
@@ -XXX,XX +XXX,XX @@ static void mps2tz_common_init(MachineState *machine)
23
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineState AspeedMachineState;
20
exit(EXIT_FAILURE);
24
21
}
25
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
22
26
#define ASPEED_MACHINE(obj) \
23
- sysbus_init_child_obj(OBJECT(machine), "iotkit", &mms->iotkit,
27
- OBJECT_CHECK(AspeedMachine, (obj), TYPE_ASPEED_MACHINE)
24
+ sysbus_init_child_obj(OBJECT(machine), TYPE_IOTKIT, &mms->iotkit,
28
-
25
sizeof(mms->iotkit), mmc->armsse_type);
29
-typedef struct AspeedMachine {
26
iotkitdev = DEVICE(&mms->iotkit);
30
- MachineState parent_obj;
27
object_property_set_link(OBJECT(&mms->iotkit), OBJECT(system_memory),
31
-
32
- bool mmio_exec;
33
-} AspeedMachine;
34
+ OBJECT_CHECK(AspeedMachineState, (obj), TYPE_ASPEED_MACHINE)
35
36
#define ASPEED_MAC0_ON (1 << 0)
37
#define ASPEED_MAC1_ON (1 << 1)
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
};
44
45
struct AspeedMachineState {
46
+ /* Private */
47
+ MachineState parent_obj;
48
+ /* Public */
49
+
50
AspeedSoCState soc;
51
MemoryRegion ram_container;
52
MemoryRegion max_ram;
53
+ bool mmio_exec;
54
};
55
56
/* Palmetto hardware value: 0x120CE416 */
57
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
58
59
static void aspeed_machine_init(MachineState *machine)
60
{
61
- AspeedMachineState *bmc;
62
+ AspeedMachineState *bmc = ASPEED_MACHINE(machine);
63
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
64
AspeedSoCClass *sc;
65
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
66
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
67
int i;
68
NICInfo *nd = &nd_table[0];
69
70
- bmc = g_new0(AspeedMachineState, 1);
71
-
72
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
73
4 * GiB);
74
memory_region_add_subregion(&bmc->ram_container, 0, machine->ram);
75
@@ -XXX,XX +XXX,XX @@ static const TypeInfo aspeed_machine_types[] = {
76
}, {
77
.name = TYPE_ASPEED_MACHINE,
78
.parent = TYPE_MACHINE,
79
- .instance_size = sizeof(AspeedMachine),
80
+ .instance_size = sizeof(AspeedMachineState),
81
.instance_init = aspeed_machine_instance_init,
82
.class_size = sizeof(AspeedMachineClass),
83
.class_init = aspeed_machine_class_init,
28
--
84
--
29
2.20.1
85
2.20.1
30
86
31
87
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Extract i2c_try_create_slave() and i2c_realize_and_unref()
4
from i2c_create_slave().
5
We can now set properties on a I2CSlave before it is realized.
6
7
This is in line with the recent qdev/QOM changes merged
8
in commit 6675a653d2e.
9
10
Reviewed-by: Corey Minyard <cminyard@mvista.com>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Reviewed-by: Markus Armbruster <armbru@redhat.com>
14
Tested-by: Cédric Le Goater <clg@kaod.org>
15
Message-id: 20200623072723.6324-2-f4bug@amsat.org
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
---
18
include/hw/i2c/i2c.h | 2 ++
19
hw/i2c/core.c | 18 ++++++++++++++++--
20
2 files changed, 18 insertions(+), 2 deletions(-)
21
22
diff --git a/include/hw/i2c/i2c.h b/include/hw/i2c/i2c.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/include/hw/i2c/i2c.h
25
+++ b/include/hw/i2c/i2c.h
26
@@ -XXX,XX +XXX,XX @@ int i2c_send(I2CBus *bus, uint8_t data);
27
uint8_t i2c_recv(I2CBus *bus);
28
29
DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr);
30
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr);
31
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp);
32
33
/* lm832x.c */
34
void lm832x_key_event(DeviceState *dev, int key, int state);
35
diff --git a/hw/i2c/core.c b/hw/i2c/core.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/hw/i2c/core.c
38
+++ b/hw/i2c/core.c
39
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_i2c_slave = {
40
}
41
};
42
43
-DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
44
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr)
45
{
46
DeviceState *dev;
47
48
dev = qdev_new(name);
49
qdev_prop_set_uint8(dev, "address", addr);
50
- qdev_realize_and_unref(dev, &bus->qbus, &error_fatal);
51
+ return dev;
52
+}
53
+
54
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp)
55
+{
56
+ return qdev_realize_and_unref(dev, &bus->qbus, errp);
57
+}
58
+
59
+DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
60
+{
61
+ DeviceState *dev;
62
+
63
+ dev = i2c_try_create_slave(name, addr);
64
+ i2c_realize_and_unref(dev, bus, &error_fatal);
65
+
66
return dev;
67
}
68
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
The PCA9552 device does not expose LEDs, but simple pins
4
to connnect LEDs to. To be clearer with the device model,
5
rename 'nr_leds' as 'pin_count'.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-3-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 2 +-
14
hw/misc/pca9552.c | 10 +++++-----
15
2 files changed, 6 insertions(+), 6 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
22
23
uint8_t regs[PCA9552_NR_REGS];
24
uint8_t max_reg;
25
- uint8_t nr_leds;
26
+ uint8_t pin_count;
27
} PCA9552State;
28
29
#endif
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
35
{
36
int i;
37
38
- for (i = 0; i < s->nr_leds; i++) {
39
+ for (i = 0; i < s->pin_count; i++) {
40
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
41
uint8_t input_shift = (i % 8);
42
uint8_t config = pca9552_pin_get_config(s, i);
43
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
44
error_setg(errp, "%s: error reading %s", __func__, name);
45
return;
46
}
47
- if (led < 0 || led > s->nr_leds) {
48
+ if (led < 0 || led > s->pin_count) {
49
error_setg(errp, "%s invalid led %s", __func__, name);
50
return;
51
}
52
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
53
error_setg(errp, "%s: error reading %s", __func__, name);
54
return;
55
}
56
- if (led < 0 || led > s->nr_leds) {
57
+ if (led < 0 || led > s->pin_count) {
58
error_setg(errp, "%s invalid led %s", __func__, name);
59
return;
60
}
61
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
62
* PCA955X device
63
*/
64
s->max_reg = PCA9552_LS3;
65
- s->nr_leds = 16;
66
+ s->pin_count = 16;
67
68
- for (led = 0; led < s->nr_leds; led++) {
69
+ for (led = 0; led < s->pin_count; led++) {
70
char *name;
71
72
name = g_strdup_printf("led%d", led);
73
--
74
2.20.1
75
76
diff view generated by jsdifflib
New patch
1
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
3
Various code from the PCA9552 device model is generic to the
4
PCA955X family. We'll split the generic code in a base class
5
in the next commit. To ease review, first do a dumb renaming.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-4-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 10 ++---
14
hw/misc/pca9552.c | 80 +++++++++++++++++++--------------------
15
2 files changed, 45 insertions(+), 45 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@
22
#include "hw/i2c/i2c.h"
23
24
#define TYPE_PCA9552 "pca9552"
25
-#define PCA9552(obj) OBJECT_CHECK(PCA9552State, (obj), TYPE_PCA9552)
26
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
28
-#define PCA9552_NR_REGS 10
29
+#define PCA955X_NR_REGS 10
30
31
-typedef struct PCA9552State {
32
+typedef struct PCA955xState {
33
/*< private >*/
34
I2CSlave i2c;
35
/*< public >*/
36
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
37
uint8_t len;
38
uint8_t pointer;
39
40
- uint8_t regs[PCA9552_NR_REGS];
41
+ uint8_t regs[PCA955X_NR_REGS];
42
uint8_t max_reg;
43
uint8_t pin_count;
44
-} PCA9552State;
45
+} PCA955xState;
46
47
#endif
48
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/hw/misc/pca9552.c
51
+++ b/hw/misc/pca9552.c
52
@@ -XXX,XX +XXX,XX @@
53
54
static const char *led_state[] = {"on", "off", "pwm0", "pwm1"};
55
56
-static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
57
+static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
58
{
59
uint8_t reg = PCA9552_LS0 + (pin / 4);
60
uint8_t shift = (pin % 4) << 1;
61
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
62
return extract32(s->regs[reg], shift, 2);
63
}
64
65
-static void pca9552_update_pin_input(PCA9552State *s)
66
+static void pca955x_update_pin_input(PCA955xState *s)
67
{
68
int i;
69
70
for (i = 0; i < s->pin_count; i++) {
71
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
72
uint8_t input_shift = (i % 8);
73
- uint8_t config = pca9552_pin_get_config(s, i);
74
+ uint8_t config = pca955x_pin_get_config(s, i);
75
76
switch (config) {
77
case PCA9552_LED_ON:
78
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
79
}
80
}
81
82
-static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
83
+static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
84
{
85
switch (reg) {
86
case PCA9552_INPUT0:
87
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
88
}
89
}
90
91
-static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
92
+static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
93
{
94
switch (reg) {
95
case PCA9552_PSC0:
96
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
97
case PCA9552_LS2:
98
case PCA9552_LS3:
99
s->regs[reg] = data;
100
- pca9552_update_pin_input(s);
101
+ pca955x_update_pin_input(s);
102
break;
103
104
case PCA9552_INPUT0:
105
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
106
* after each byte is sent to or received by the device. The index
107
* rollovers to 0 when the maximum register address is reached.
108
*/
109
-static void pca9552_autoinc(PCA9552State *s)
110
+static void pca955x_autoinc(PCA955xState *s)
111
{
112
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
113
uint8_t reg = s->pointer & 0xf;
114
@@ -XXX,XX +XXX,XX @@ static void pca9552_autoinc(PCA9552State *s)
115
}
116
}
117
118
-static uint8_t pca9552_recv(I2CSlave *i2c)
119
+static uint8_t pca955x_recv(I2CSlave *i2c)
120
{
121
- PCA9552State *s = PCA9552(i2c);
122
+ PCA955xState *s = PCA955X(i2c);
123
uint8_t ret;
124
125
- ret = pca9552_read(s, s->pointer & 0xf);
126
+ ret = pca955x_read(s, s->pointer & 0xf);
127
128
/*
129
* From the Specs:
130
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_recv(I2CSlave *i2c)
131
__func__);
132
}
133
134
- pca9552_autoinc(s);
135
+ pca955x_autoinc(s);
136
137
return ret;
138
}
139
140
-static int pca9552_send(I2CSlave *i2c, uint8_t data)
141
+static int pca955x_send(I2CSlave *i2c, uint8_t data)
142
{
143
- PCA9552State *s = PCA9552(i2c);
144
+ PCA955xState *s = PCA955X(i2c);
145
146
/* First byte sent by is the register address */
147
if (s->len == 0) {
148
s->pointer = data;
149
s->len++;
150
} else {
151
- pca9552_write(s, s->pointer & 0xf, data);
152
+ pca955x_write(s, s->pointer & 0xf, data);
153
154
- pca9552_autoinc(s);
155
+ pca955x_autoinc(s);
156
}
157
158
return 0;
159
}
160
161
-static int pca9552_event(I2CSlave *i2c, enum i2c_event event)
162
+static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
163
{
164
- PCA9552State *s = PCA9552(i2c);
165
+ PCA955xState *s = PCA955X(i2c);
166
167
s->len = 0;
168
return 0;
169
}
170
171
-static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
172
+static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
173
void *opaque, Error **errp)
174
{
175
- PCA9552State *s = PCA9552(obj);
176
+ PCA955xState *s = PCA955X(obj);
177
int led, rc, reg;
178
uint8_t state;
179
180
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
181
* reading the INPUTx reg
182
*/
183
reg = PCA9552_LS0 + led / 4;
184
- state = (pca9552_read(s, reg) >> (led % 8)) & 0x3;
185
+ state = (pca955x_read(s, reg) >> (led % 8)) & 0x3;
186
visit_type_str(v, name, (char **)&led_state[state], errp);
187
}
188
189
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
190
((state & 0x3) << (led_num << 1));
191
}
192
193
-static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
194
+static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
195
void *opaque, Error **errp)
196
{
197
- PCA9552State *s = PCA9552(obj);
198
+ PCA955xState *s = PCA955X(obj);
199
Error *local_err = NULL;
200
int led, rc, reg, val;
201
uint8_t state;
202
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
203
}
204
205
reg = PCA9552_LS0 + led / 4;
206
- val = pca9552_read(s, reg);
207
+ val = pca955x_read(s, reg);
208
val = pca955x_ledsel(val, led % 4, state);
209
- pca9552_write(s, reg, val);
210
+ pca955x_write(s, reg, val);
211
}
212
213
static const VMStateDescription pca9552_vmstate = {
214
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription pca9552_vmstate = {
215
.version_id = 0,
216
.minimum_version_id = 0,
217
.fields = (VMStateField[]) {
218
- VMSTATE_UINT8(len, PCA9552State),
219
- VMSTATE_UINT8(pointer, PCA9552State),
220
- VMSTATE_UINT8_ARRAY(regs, PCA9552State, PCA9552_NR_REGS),
221
- VMSTATE_I2C_SLAVE(i2c, PCA9552State),
222
+ VMSTATE_UINT8(len, PCA955xState),
223
+ VMSTATE_UINT8(pointer, PCA955xState),
224
+ VMSTATE_UINT8_ARRAY(regs, PCA955xState, PCA955X_NR_REGS),
225
+ VMSTATE_I2C_SLAVE(i2c, PCA955xState),
226
VMSTATE_END_OF_LIST()
227
}
228
};
229
230
static void pca9552_reset(DeviceState *dev)
231
{
232
- PCA9552State *s = PCA9552(dev);
233
+ PCA955xState *s = PCA955X(dev);
234
235
s->regs[PCA9552_PSC0] = 0xFF;
236
s->regs[PCA9552_PWM0] = 0x80;
237
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
238
s->regs[PCA9552_LS2] = 0x55;
239
s->regs[PCA9552_LS3] = 0x55;
240
241
- pca9552_update_pin_input(s);
242
+ pca955x_update_pin_input(s);
243
244
s->pointer = 0xFF;
245
s->len = 0;
246
}
247
248
-static void pca9552_initfn(Object *obj)
249
+static void pca955x_initfn(Object *obj)
250
{
251
- PCA9552State *s = PCA9552(obj);
252
+ PCA955xState *s = PCA955X(obj);
253
int led;
254
255
/* If support for the other PCA955X devices are implemented, these
256
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
257
char *name;
258
259
name = g_strdup_printf("led%d", led);
260
- object_property_add(obj, name, "bool", pca9552_get_led, pca9552_set_led,
261
+ object_property_add(obj, name, "bool", pca955x_get_led, pca955x_set_led,
262
NULL, NULL);
263
g_free(name);
264
}
265
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
266
DeviceClass *dc = DEVICE_CLASS(klass);
267
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
268
269
- k->event = pca9552_event;
270
- k->recv = pca9552_recv;
271
- k->send = pca9552_send;
272
+ k->event = pca955x_event;
273
+ k->recv = pca955x_recv;
274
+ k->send = pca955x_send;
275
dc->reset = pca9552_reset;
276
dc->vmsd = &pca9552_vmstate;
277
}
278
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
279
static const TypeInfo pca9552_info = {
280
.name = TYPE_PCA9552,
281
.parent = TYPE_I2C_SLAVE,
282
- .instance_init = pca9552_initfn,
283
- .instance_size = sizeof(PCA9552State),
284
+ .instance_init = pca955x_initfn,
285
+ .instance_size = sizeof(PCA955xState),
286
.class_init = pca9552_class_init,
287
};
288
289
-static void pca9552_register_types(void)
290
+static void pca955x_register_types(void)
291
{
292
type_register_static(&pca9552_info);
293
}
294
295
-type_init(pca9552_register_types)
296
+type_init(pca955x_register_types)
297
--
298
2.20.1
299
300
diff view generated by jsdifflib
New patch
1
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
3
Extract the code common to the PCA955x family in PCA955xClass,
4
keeping the PCA9552 specific parts into pca9552_class_init().
5
Remove the 'TODO' comment added in commit 5141d4158cf.
6
7
Suggested-by: Cédric Le Goater <clg@kaod.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-5-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 6 ++--
15
hw/misc/pca9552.c | 66 ++++++++++++++++++++++++++++-----------
16
2 files changed, 51 insertions(+), 21 deletions(-)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@
23
#include "hw/i2c/i2c.h"
24
25
#define TYPE_PCA9552 "pca9552"
26
-#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
+#define TYPE_PCA955X "pca955x"
28
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA955X)
29
30
#define PCA955X_NR_REGS 10
31
+#define PCA955X_PIN_COUNT_MAX 16
32
33
typedef struct PCA955xState {
34
/*< private >*/
35
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
36
uint8_t pointer;
37
38
uint8_t regs[PCA955X_NR_REGS];
39
- uint8_t max_reg;
40
- uint8_t pin_count;
41
} PCA955xState;
42
43
#endif
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
* https://www.nxp.com/docs/en/application-note/AN264.pdf
50
*
51
* Copyright (c) 2017-2018, IBM Corporation.
52
+ * Copyright (c) 2020 Philippe Mathieu-Daudé
53
*
54
* This work is licensed under the terms of the GNU GPL, version 2 or
55
* later. See the COPYING file in the top-level directory.
56
@@ -XXX,XX +XXX,XX @@
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
60
+typedef struct PCA955xClass {
61
+ /*< private >*/
62
+ I2CSlaveClass parent_class;
63
+ /*< public >*/
64
+
65
+ uint8_t pin_count;
66
+ uint8_t max_reg;
67
+} PCA955xClass;
68
+
69
+#define PCA955X_CLASS(klass) \
70
+ OBJECT_CLASS_CHECK(PCA955xClass, (klass), TYPE_PCA955X)
71
+#define PCA955X_GET_CLASS(obj) \
72
+ OBJECT_GET_CLASS(PCA955xClass, (obj), TYPE_PCA955X)
73
+
74
#define PCA9552_LED_ON 0x0
75
#define PCA9552_LED_OFF 0x1
76
#define PCA9552_LED_PWM0 0x2
77
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
78
79
static void pca955x_update_pin_input(PCA955xState *s)
80
{
81
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
82
int i;
83
84
- for (i = 0; i < s->pin_count; i++) {
85
+ for (i = 0; i < k->pin_count; i++) {
86
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
87
uint8_t input_shift = (i % 8);
88
uint8_t config = pca955x_pin_get_config(s, i);
89
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
90
*/
91
static void pca955x_autoinc(PCA955xState *s)
92
{
93
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
94
+
95
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
96
uint8_t reg = s->pointer & 0xf;
97
98
- reg = (reg + 1) % (s->max_reg + 1);
99
+ reg = (reg + 1) % (k->max_reg + 1);
100
s->pointer = reg | PCA9552_AUTOINC;
101
}
102
}
103
@@ -XXX,XX +XXX,XX @@ static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
104
static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
105
void *opaque, Error **errp)
106
{
107
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
108
PCA955xState *s = PCA955X(obj);
109
int led, rc, reg;
110
uint8_t state;
111
@@ -XXX,XX +XXX,XX @@ static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
112
error_setg(errp, "%s: error reading %s", __func__, name);
113
return;
114
}
115
- if (led < 0 || led > s->pin_count) {
116
+ if (led < 0 || led > k->pin_count) {
117
error_setg(errp, "%s invalid led %s", __func__, name);
118
return;
119
}
120
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
121
static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
122
void *opaque, Error **errp)
123
{
124
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
125
PCA955xState *s = PCA955X(obj);
126
Error *local_err = NULL;
127
int led, rc, reg, val;
128
@@ -XXX,XX +XXX,XX @@ static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
129
error_setg(errp, "%s: error reading %s", __func__, name);
130
return;
131
}
132
- if (led < 0 || led > s->pin_count) {
133
+ if (led < 0 || led > k->pin_count) {
134
error_setg(errp, "%s invalid led %s", __func__, name);
135
return;
136
}
137
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
138
139
static void pca955x_initfn(Object *obj)
140
{
141
- PCA955xState *s = PCA955X(obj);
142
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
143
int led;
144
145
- /* If support for the other PCA955X devices are implemented, these
146
- * constant values might be part of class structure describing the
147
- * PCA955X device
148
- */
149
- s->max_reg = PCA9552_LS3;
150
- s->pin_count = 16;
151
-
152
- for (led = 0; led < s->pin_count; led++) {
153
+ assert(k->pin_count <= PCA955X_PIN_COUNT_MAX);
154
+ for (led = 0; led < k->pin_count; led++) {
155
char *name;
156
157
name = g_strdup_printf("led%d", led);
158
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
159
}
160
}
161
162
-static void pca9552_class_init(ObjectClass *klass, void *data)
163
+static void pca955x_class_init(ObjectClass *klass, void *data)
164
{
165
- DeviceClass *dc = DEVICE_CLASS(klass);
166
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
167
168
k->event = pca955x_event;
169
k->recv = pca955x_recv;
170
k->send = pca955x_send;
171
+}
172
+
173
+static const TypeInfo pca955x_info = {
174
+ .name = TYPE_PCA955X,
175
+ .parent = TYPE_I2C_SLAVE,
176
+ .instance_init = pca955x_initfn,
177
+ .instance_size = sizeof(PCA955xState),
178
+ .class_init = pca955x_class_init,
179
+ .abstract = true,
180
+};
181
+
182
+static void pca9552_class_init(ObjectClass *oc, void *data)
183
+{
184
+ DeviceClass *dc = DEVICE_CLASS(oc);
185
+ PCA955xClass *pc = PCA955X_CLASS(oc);
186
+
187
dc->reset = pca9552_reset;
188
dc->vmsd = &pca9552_vmstate;
189
+ pc->max_reg = PCA9552_LS3;
190
+ pc->pin_count = 16;
191
}
192
193
static const TypeInfo pca9552_info = {
194
.name = TYPE_PCA9552,
195
- .parent = TYPE_I2C_SLAVE,
196
- .instance_init = pca955x_initfn,
197
- .instance_size = sizeof(PCA955xState),
198
+ .parent = TYPE_PCA955X,
199
.class_init = pca9552_class_init,
200
};
201
202
static void pca955x_register_types(void)
203
{
204
+ type_register_static(&pca955x_info);
205
type_register_static(&pca9552_info);
206
}
207
208
--
209
2.20.1
210
211
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Add support for SD.
3
Add a description field to distinguish between multiple devices.
4
4
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
7
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20200427181649.26851-11-edgar.iglesias@gmail.com
8
Message-id: 20200623072723.6324-6-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
hw/arm/xlnx-versal-virt.c | 46 +++++++++++++++++++++++++++++++++++++++
11
include/hw/misc/pca9552.h | 1 +
12
1 file changed, 46 insertions(+)
12
hw/misc/pca9552.c | 18 ++++++++++++++++++
13
2 files changed, 19 insertions(+)
13
14
14
diff --git a/hw/arm/xlnx-versal-virt.c b/hw/arm/xlnx-versal-virt.c
15
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/hw/arm/xlnx-versal-virt.c
17
--- a/include/hw/misc/pca9552.h
17
+++ b/hw/arm/xlnx-versal-virt.c
18
+++ b/include/hw/misc/pca9552.h
19
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
20
uint8_t pointer;
21
22
uint8_t regs[PCA955X_NR_REGS];
23
+ char *description; /* For debugging purpose only */
24
} PCA955xState;
25
26
#endif
27
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/misc/pca9552.c
30
+++ b/hw/misc/pca9552.c
18
@@ -XXX,XX +XXX,XX @@
31
@@ -XXX,XX +XXX,XX @@
19
#include "hw/arm/sysbus-fdt.h"
32
#include "qemu/osdep.h"
20
#include "hw/arm/fdt.h"
33
#include "qemu/log.h"
21
#include "cpu.h"
34
#include "qemu/module.h"
22
+#include "hw/qdev-properties.h"
35
+#include "hw/qdev-properties.h"
23
#include "hw/arm/xlnx-versal.h"
36
#include "hw/misc/pca9552.h"
24
37
#include "hw/misc/pca9552_regs.h"
25
#define TYPE_XLNX_VERSAL_VIRT_MACHINE MACHINE_TYPE_NAME("xlnx-versal-virt")
38
#include "migration/vmstate.h"
26
@@ -XXX,XX +XXX,XX @@ static void fdt_add_zdma_nodes(VersalVirt *s)
39
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
27
}
40
}
28
}
41
}
29
42
30
+static void fdt_add_sd_nodes(VersalVirt *s)
43
+static void pca955x_realize(DeviceState *dev, Error **errp)
31
+{
44
+{
32
+ const char clocknames[] = "clk_xin\0clk_ahb";
45
+ PCA955xState *s = PCA955X(dev);
33
+ const char compat[] = "arasan,sdhci-8.9a";
34
+ int i;
35
+
46
+
36
+ for (i = ARRAY_SIZE(s->soc.pmc.iou.sd) - 1; i >= 0; i--) {
47
+ if (!s->description) {
37
+ uint64_t addr = MM_PMC_SD0 + MM_PMC_SD0_SIZE * i;
48
+ s->description = g_strdup("pca-unspecified");
38
+ char *name = g_strdup_printf("/sdhci@%" PRIx64, addr);
39
+
40
+ qemu_fdt_add_subnode(s->fdt, name);
41
+
42
+ qemu_fdt_setprop_cells(s->fdt, name, "clocks",
43
+ s->phandle.clk_25Mhz, s->phandle.clk_25Mhz);
44
+ qemu_fdt_setprop(s->fdt, name, "clock-names",
45
+ clocknames, sizeof(clocknames));
46
+ qemu_fdt_setprop_cells(s->fdt, name, "interrupts",
47
+ GIC_FDT_IRQ_TYPE_SPI, VERSAL_SD0_IRQ_0 + i * 2,
48
+ GIC_FDT_IRQ_FLAGS_LEVEL_HI);
49
+ qemu_fdt_setprop_sized_cells(s->fdt, name, "reg",
50
+ 2, addr, 2, MM_PMC_SD0_SIZE);
51
+ qemu_fdt_setprop(s->fdt, name, "compatible", compat, sizeof(compat));
52
+ g_free(name);
53
+ }
49
+ }
54
+}
50
+}
55
+
51
+
56
static void fdt_nop_memory_nodes(void *fdt, Error **errp)
52
+static Property pca955x_properties[] = {
53
+ DEFINE_PROP_STRING("description", PCA955xState, description),
54
+ DEFINE_PROP_END_OF_LIST(),
55
+};
56
+
57
static void pca955x_class_init(ObjectClass *klass, void *data)
57
{
58
{
58
Error *err = NULL;
59
+ DeviceClass *dc = DEVICE_CLASS(klass);
59
@@ -XXX,XX +XXX,XX @@ static void create_virtio_regions(VersalVirt *s)
60
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
60
}
61
62
k->event = pca955x_event;
63
k->recv = pca955x_recv;
64
k->send = pca955x_send;
65
+ dc->realize = pca955x_realize;
66
+ device_class_set_props(dc, pca955x_properties);
61
}
67
}
62
68
63
+static void sd_plugin_card(SDHCIState *sd, DriveInfo *di)
69
static const TypeInfo pca955x_info = {
64
+{
65
+ BlockBackend *blk = di ? blk_by_legacy_dinfo(di) : NULL;
66
+ DeviceState *card;
67
+
68
+ card = qdev_create(qdev_get_child_bus(DEVICE(sd), "sd-bus"), TYPE_SD_CARD);
69
+ object_property_add_child(OBJECT(sd), "card[*]", OBJECT(card),
70
+ &error_fatal);
71
+ qdev_prop_set_drive(card, "drive", blk, &error_fatal);
72
+ object_property_set_bool(OBJECT(card), true, "realized", &error_fatal);
73
+}
74
+
75
static void versal_virt_init(MachineState *machine)
76
{
77
VersalVirt *s = XLNX_VERSAL_VIRT_MACHINE(machine);
78
int psci_conduit = QEMU_PSCI_CONDUIT_DISABLED;
79
+ int i;
80
81
/*
82
* If the user provides an Operating System to be loaded, we expect them
83
@@ -XXX,XX +XXX,XX @@ static void versal_virt_init(MachineState *machine)
84
fdt_add_gic_nodes(s);
85
fdt_add_timer_nodes(s);
86
fdt_add_zdma_nodes(s);
87
+ fdt_add_sd_nodes(s);
88
fdt_add_cpu_nodes(s, psci_conduit);
89
fdt_add_clk_node(s, "/clk125", 125000000, s->phandle.clk_125Mhz);
90
fdt_add_clk_node(s, "/clk25", 25000000, s->phandle.clk_25Mhz);
91
@@ -XXX,XX +XXX,XX @@ static void versal_virt_init(MachineState *machine)
92
memory_region_add_subregion_overlap(get_system_memory(),
93
0, &s->soc.fpd.apu.mr, 0);
94
95
+ /* Plugin SD cards. */
96
+ for (i = 0; i < ARRAY_SIZE(s->soc.pmc.iou.sd); i++) {
97
+ sd_plugin_card(&s->soc.pmc.iou.sd[i], drive_get_next(IF_SD));
98
+ }
99
+
100
s->binfo.ram_size = machine->ram_size;
101
s->binfo.loader_start = 0x0;
102
s->binfo.get_dtb = versal_virt_get_dtb;
103
--
70
--
104
2.20.1
71
2.20.1
105
72
106
73
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Add a trivial representation of the PCA9552 GPIOs.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_status
8
1592689902.327837:pca955x_gpio_status pca-unspecified GPIOs 0-15 [*...............]
9
1592689902.329934:pca955x_gpio_status pca-unspecified GPIOs 0-15 [**..............]
10
1592689902.330717:pca955x_gpio_status pca-unspecified GPIOs 0-15 [***.............]
11
1592689902.331431:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****............]
12
1592689902.332163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*..]
13
1592689902.332888:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........**.]
14
1592689902.333629:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
15
1592690032.793289:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
16
1592690033.303163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
17
1592690033.812962:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
18
1592690034.323234:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
19
1592690034.832922:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
20
21
We notice the GPIO #14 (front-power LED) starts to blink.
22
23
This LED is described in the witherspoon device-tree [*]:
24
25
front-power {
26
retain-state-shutdown;
27
default-state = "keep";
28
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
29
};
30
31
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
32
33
Suggested-by: Cédric Le Goater <clg@kaod.org>
34
Reviewed-by: Cédric Le Goater <clg@kaod.org>
35
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
36
Tested-by: Cédric Le Goater <clg@kaod.org>
37
Message-id: 20200623072723.6324-7-f4bug@amsat.org
38
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
39
---
40
hw/misc/pca9552.c | 39 +++++++++++++++++++++++++++++++++++++++
41
hw/misc/trace-events | 3 +++
42
2 files changed, 42 insertions(+)
43
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
#include "qemu/osdep.h"
50
#include "qemu/log.h"
51
#include "qemu/module.h"
52
+#include "qemu/bitops.h"
53
#include "hw/qdev-properties.h"
54
#include "hw/misc/pca9552.h"
55
#include "hw/misc/pca9552_regs.h"
56
#include "migration/vmstate.h"
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
+#include "trace.h"
60
61
typedef struct PCA955xClass {
62
/*< private >*/
63
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
64
return extract32(s->regs[reg], shift, 2);
65
}
66
67
+/* Return INPUT status (bit #N belongs to GPIO #N) */
68
+static uint16_t pca955x_pins_get_status(PCA955xState *s)
69
+{
70
+ return (s->regs[PCA9552_INPUT1] << 8) | s->regs[PCA9552_INPUT0];
71
+}
72
+
73
+static void pca955x_display_pins_status(PCA955xState *s,
74
+ uint16_t previous_pins_status)
75
+{
76
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
77
+ uint16_t pins_status, pins_changed;
78
+ int i;
79
+
80
+ pins_status = pca955x_pins_get_status(s);
81
+ pins_changed = previous_pins_status ^ pins_status;
82
+ if (!pins_changed) {
83
+ return;
84
+ }
85
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_STATUS)) {
86
+ char *buf = g_newa(char, k->pin_count + 1);
87
+
88
+ for (i = 0; i < k->pin_count; i++) {
89
+ if (extract32(pins_status, i, 1)) {
90
+ buf[i] = '*';
91
+ } else {
92
+ buf[i] = '.';
93
+ }
94
+ }
95
+ buf[i] = '\0';
96
+ trace_pca955x_gpio_status(s->description, buf);
97
+ }
98
+}
99
+
100
static void pca955x_update_pin_input(PCA955xState *s)
101
{
102
PCA955xClass *k = PCA955X_GET_CLASS(s);
103
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
104
105
static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
106
{
107
+ uint16_t pins_status;
108
+
109
switch (reg) {
110
case PCA9552_PSC0:
111
case PCA9552_PWM0:
112
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
113
case PCA9552_LS1:
114
case PCA9552_LS2:
115
case PCA9552_LS3:
116
+ pins_status = pca955x_pins_get_status(s);
117
s->regs[reg] = data;
118
pca955x_update_pin_input(s);
119
+ pca955x_display_pins_status(s, pins_status);
120
break;
121
122
case PCA9552_INPUT0:
123
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
124
index XXXXXXX..XXXXXXX 100644
125
--- a/hw/misc/trace-events
126
+++ b/hw/misc/trace-events
127
@@ -XXX,XX +XXX,XX @@ via1_adb_poll(uint8_t data, const char *vadbint, int status, int index, int size
128
# grlib_ahb_apb_pnp.c
129
grlib_ahb_pnp_read(uint64_t addr, uint32_t value) "AHB PnP read addr:0x%03"PRIx64" data:0x%08x"
130
grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx64" data:0x%08x"
131
+
132
+# pca9552.c
133
+pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
134
--
135
2.20.1
136
137
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
We have 2 distinct PCA9552 devices. Set their description
4
to distinguish them when looking at the trace events.
5
6
Description name taken from:
7
https://github.com/open-power/witherspoon-xml/blob/master/witherspoon.xml
8
9
Reviewed-by: Cédric Le Goater <clg@kaod.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Corey Minyard <cminyard@mvista.com>
12
Reviewed-by: Markus Armbruster <armbru@redhat.com>
13
Tested-by: Cédric Le Goater <clg@kaod.org>
14
Message-id: 20200623072723.6324-8-f4bug@amsat.org
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
---
17
hw/arm/aspeed.c | 13 +++++++++----
18
1 file changed, 9 insertions(+), 4 deletions(-)
19
20
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/hw/arm/aspeed.c
23
+++ b/hw/arm/aspeed.c
24
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
25
{
26
AspeedSoCState *soc = &bmc->soc;
27
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
28
+ DeviceState *dev;
29
30
/* Bus 3: TODO bmp280@77 */
31
/* Bus 3: TODO max31785@52 */
32
/* Bus 3: TODO dps310@76 */
33
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
34
- 0x60);
35
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
36
+ qdev_prop_set_string(dev, "description", "pca1");
37
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3),
38
+ &error_fatal);
39
40
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
41
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
42
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
43
44
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
45
eeprom_buf);
46
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
47
- 0x60);
48
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
49
+ qdev_prop_set_string(dev, "description", "pca0");
50
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11),
51
+ &error_fatal);
52
/* Bus 11: TODO ucd90160@64 */
53
}
54
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Emit a trace event when a GPIO change its state.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_change
8
1592690552.687372:pca955x_gpio_change pca1 GPIO id:0 status: 0 -> 1
9
1592690552.690169:pca955x_gpio_change pca1 GPIO id:1 status: 0 -> 1
10
1592690552.691673:pca955x_gpio_change pca1 GPIO id:2 status: 0 -> 1
11
1592690552.696886:pca955x_gpio_change pca1 GPIO id:3 status: 0 -> 1
12
1592690552.698614:pca955x_gpio_change pca1 GPIO id:13 status: 0 -> 1
13
1592690552.699833:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
14
1592690552.700842:pca955x_gpio_change pca1 GPIO id:15 status: 0 -> 1
15
1592690683.841921:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
16
1592690683.861660:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
17
1592690684.371460:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
18
1592690684.882115:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
19
1592690685.391411:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
20
1592690685.901391:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
21
1592690686.411678:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
22
1592690686.921279:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
23
24
We notice the GPIO #14 (front-power LED) starts to blink.
25
26
This LED is described in the witherspoon device-tree [*]:
27
28
front-power {
29
retain-state-shutdown;
30
default-state = "keep";
31
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
32
};
33
34
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
35
36
Reviewed-by: Cédric Le Goater <clg@kaod.org>
37
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
38
Tested-by: Cédric Le Goater <clg@kaod.org>
39
Message-id: 20200623072723.6324-9-f4bug@amsat.org
40
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
41
---
42
hw/misc/pca9552.c | 15 +++++++++++++++
43
hw/misc/trace-events | 1 +
44
2 files changed, 16 insertions(+)
45
46
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/hw/misc/pca9552.c
49
+++ b/hw/misc/pca9552.c
50
@@ -XXX,XX +XXX,XX @@ static void pca955x_display_pins_status(PCA955xState *s,
51
buf[i] = '\0';
52
trace_pca955x_gpio_status(s->description, buf);
53
}
54
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_CHANGE)) {
55
+ for (i = 0; i < k->pin_count; i++) {
56
+ if (extract32(pins_changed, i, 1)) {
57
+ unsigned new_state = extract32(pins_status, i, 1);
58
+
59
+ /*
60
+ * We display the state using the PCA logic ("active-high").
61
+ * This is not the state of the LED, which signal might be
62
+ * wired "active-low" on the board.
63
+ */
64
+ trace_pca955x_gpio_change(s->description, i,
65
+ !new_state, new_state);
66
+ }
67
+ }
68
+ }
69
}
70
71
static void pca955x_update_pin_input(PCA955xState *s)
72
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/misc/trace-events
75
+++ b/hw/misc/trace-events
76
@@ -XXX,XX +XXX,XX @@ grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx6
77
78
# pca9552.c
79
pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
80
+pca955x_gpio_change(const char *description, unsigned id, unsigned prev_state, unsigned current_state) "%s GPIO id:%u status: %u -> %u"
81
--
82
2.20.1
83
84
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
The PCA9552 has 16 GPIOs which can be used as input,
4
output or PWM mode. QEMU models the output GPIO with
5
the qemu_irq type. Let the device expose the 16 GPIOs
6
to allow us to later connect LEDs to these outputs.
7
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-10-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 1 +
15
hw/misc/pca9552.c | 6 ++++++
16
2 files changed, 7 insertions(+)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
23
uint8_t pointer;
24
25
uint8_t regs[PCA955X_NR_REGS];
26
+ qemu_irq gpio[PCA955X_PIN_COUNT_MAX];
27
char *description; /* For debugging purpose only */
28
} PCA955xState;
29
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@
35
#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
+#include "hw/irq.h"
39
#include "migration/vmstate.h"
40
#include "qapi/error.h"
41
#include "qapi/visitor.h"
42
@@ -XXX,XX +XXX,XX @@ static void pca955x_update_pin_input(PCA955xState *s)
43
44
switch (config) {
45
case PCA9552_LED_ON:
46
+ qemu_set_irq(s->gpio[i], 1);
47
s->regs[input_reg] |= 1 << input_shift;
48
break;
49
case PCA9552_LED_OFF:
50
+ qemu_set_irq(s->gpio[i], 0);
51
s->regs[input_reg] &= ~(1 << input_shift);
52
break;
53
case PCA9552_LED_PWM0:
54
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
55
56
static void pca955x_realize(DeviceState *dev, Error **errp)
57
{
58
+ PCA955xClass *k = PCA955X_GET_CLASS(dev);
59
PCA955xState *s = PCA955X(dev);
60
61
if (!s->description) {
62
s->description = g_strdup("pca-unspecified");
63
}
64
+
65
+ qdev_init_gpio_out(dev, s->gpio, k->pin_count);
66
}
67
68
static Property pca955x_properties[] = {
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
1
Convert the Neon VMUL, VMLA, VMLS and VSHL insns in the
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3-reg-same grouping to decodetree.
3
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-2-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-20-peter.maydell@linaro.org
7
---
7
---
8
target/arm/neon-dp.decode | 9 +++++++
8
target/arm/cpu.h | 10 ++++++++++
9
target/arm/translate-neon.inc.c | 44 +++++++++++++++++++++++++++++++++
9
1 file changed, 10 insertions(+)
10
target/arm/translate.c | 28 +++------------------
11
3 files changed, 56 insertions(+), 25 deletions(-)
12
10
13
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/neon-dp.decode
13
--- a/target/arm/cpu.h
16
+++ b/target/arm/neon-dp.decode
14
+++ b/target/arm/cpu.h
17
@@ -XXX,XX +XXX,XX @@ VCGT_U_3s 1111 001 1 0 . .. .... .... 0011 . . . 0 .... @3same
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_bti(const ARMISARegisters *id)
18
VCGE_S_3s 1111 001 0 0 . .. .... .... 0011 . . . 1 .... @3same
16
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, BT) != 0;
19
VCGE_U_3s 1111 001 1 0 . .. .... .... 0011 . . . 1 .... @3same
17
}
20
18
21
+VSHL_S_3s 1111 001 0 0 . .. .... .... 0100 . . . 0 .... @3same
19
+static inline bool isar_feature_aa64_mte_insn_reg(const ARMISARegisters *id)
22
+VSHL_U_3s 1111 001 1 0 . .. .... .... 0100 . . . 0 .... @3same
23
+
24
VMAX_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 0 .... @3same
25
VMAX_U_3s 1111 001 1 0 . .. .... .... 0110 . . . 0 .... @3same
26
VMIN_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 1 .... @3same
27
@@ -XXX,XX +XXX,XX @@ VSUB_3s 1111 001 1 0 . .. .... .... 1000 . . . 0 .... @3same
28
29
VTST_3s 1111 001 0 0 . .. .... .... 1000 . . . 1 .... @3same
30
VCEQ_3s 1111 001 1 0 . .. .... .... 1000 . . . 1 .... @3same
31
+
32
+VMLA_3s 1111 001 0 0 . .. .... .... 1001 . . . 0 .... @3same
33
+VMLS_3s 1111 001 1 0 . .. .... .... 1001 . . . 0 .... @3same
34
+
35
+VMUL_3s 1111 001 0 0 . .. .... .... 1001 . . . 1 .... @3same
36
+VMUL_p_3s 1111 001 1 0 . .. .... .... 1001 . . . 1 .... @3same
37
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
38
index XXXXXXX..XXXXXXX 100644
39
--- a/target/arm/translate-neon.inc.c
40
+++ b/target/arm/translate-neon.inc.c
41
@@ -XXX,XX +XXX,XX @@ DO_3SAME_NO_SZ_3(VMAX_S, tcg_gen_gvec_smax)
42
DO_3SAME_NO_SZ_3(VMAX_U, tcg_gen_gvec_umax)
43
DO_3SAME_NO_SZ_3(VMIN_S, tcg_gen_gvec_smin)
44
DO_3SAME_NO_SZ_3(VMIN_U, tcg_gen_gvec_umin)
45
+DO_3SAME_NO_SZ_3(VMUL, tcg_gen_gvec_mul)
46
47
#define DO_3SAME_CMP(INSN, COND) \
48
static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
49
@@ -XXX,XX +XXX,XX @@ DO_3SAME_GVEC4(VQADD_S, sqadd_op)
50
DO_3SAME_GVEC4(VQADD_U, uqadd_op)
51
DO_3SAME_GVEC4(VQSUB_S, sqsub_op)
52
DO_3SAME_GVEC4(VQSUB_U, uqsub_op)
53
+
54
+static void gen_VMUL_p_3s(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
55
+ uint32_t rm_ofs, uint32_t oprsz, uint32_t maxsz)
56
+{
20
+{
57
+ tcg_gen_gvec_3_ool(rd_ofs, rn_ofs, rm_ofs, oprsz, maxsz,
21
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) != 0;
58
+ 0, gen_helper_gvec_pmul_b);
59
+}
22
+}
60
+
23
+
61
+static bool trans_VMUL_p_3s(DisasContext *s, arg_3same *a)
24
+static inline bool isar_feature_aa64_mte(const ARMISARegisters *id)
62
+{
25
+{
63
+ if (a->size != 0) {
26
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) >= 2;
64
+ return false;
65
+ }
66
+ return do_3same(s, a, gen_VMUL_p_3s);
67
+}
27
+}
68
+
28
+
69
+#define DO_3SAME_GVEC3_NO_SZ_3(INSN, OPARRAY) \
29
static inline bool isar_feature_aa64_pmu_8_1(const ARMISARegisters *id)
70
+ static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
30
{
71
+ uint32_t rn_ofs, uint32_t rm_ofs, \
31
return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, PMUVER) >= 4 &&
72
+ uint32_t oprsz, uint32_t maxsz) \
73
+ { \
74
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, \
75
+ oprsz, maxsz, &OPARRAY[vece]); \
76
+ } \
77
+ DO_3SAME_NO_SZ_3(INSN, gen_##INSN##_3s)
78
+
79
+
80
+DO_3SAME_GVEC3_NO_SZ_3(VMLA, mla_op)
81
+DO_3SAME_GVEC3_NO_SZ_3(VMLS, mls_op)
82
+
83
+#define DO_3SAME_GVEC3_SHIFT(INSN, OPARRAY) \
84
+ static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
85
+ uint32_t rn_ofs, uint32_t rm_ofs, \
86
+ uint32_t oprsz, uint32_t maxsz) \
87
+ { \
88
+ /* Note the operation is vshl vd,vm,vn */ \
89
+ tcg_gen_gvec_3(rd_ofs, rm_ofs, rn_ofs, \
90
+ oprsz, maxsz, &OPARRAY[vece]); \
91
+ } \
92
+ DO_3SAME(INSN, gen_##INSN##_3s)
93
+
94
+DO_3SAME_GVEC3_SHIFT(VSHL_S, sshl_op)
95
+DO_3SAME_GVEC3_SHIFT(VSHL_U, ushl_op)
96
diff --git a/target/arm/translate.c b/target/arm/translate.c
97
index XXXXXXX..XXXXXXX 100644
98
--- a/target/arm/translate.c
99
+++ b/target/arm/translate.c
100
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
101
}
102
return 1;
103
104
- case NEON_3R_VMUL: /* VMUL */
105
- if (u) {
106
- /* Polynomial case allows only P8. */
107
- if (size != 0) {
108
- return 1;
109
- }
110
- tcg_gen_gvec_3_ool(rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size,
111
- 0, gen_helper_gvec_pmul_b);
112
- } else {
113
- tcg_gen_gvec_mul(size, rd_ofs, rn_ofs, rm_ofs,
114
- vec_size, vec_size);
115
- }
116
- return 0;
117
-
118
- case NEON_3R_VML: /* VMLA, VMLS */
119
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size,
120
- u ? &mls_op[size] : &mla_op[size]);
121
- return 0;
122
-
123
- case NEON_3R_VSHL:
124
- /* Note the operation is vshl vd,vm,vn */
125
- tcg_gen_gvec_3(rd_ofs, rm_ofs, rn_ofs, vec_size, vec_size,
126
- u ? &ushl_op[size] : &sshl_op[size]);
127
- return 0;
128
-
129
case NEON_3R_VADD_VSUB:
130
case NEON_3R_LOGIC:
131
case NEON_3R_VMAX:
132
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
133
case NEON_3R_VCGE:
134
case NEON_3R_VQADD:
135
case NEON_3R_VQSUB:
136
+ case NEON_3R_VMUL:
137
+ case NEON_3R_VML:
138
+ case NEON_3R_VSHL:
139
/* Already handled by decodetree */
140
return 1;
141
}
142
--
32
--
143
2.20.1
33
2.20.1
144
34
145
35
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Protect reads of aa64 id registers with ARM_CP_STATE_AA64.
4
Use this as a simpler test than arm_el_is_aa64, since EL3
5
cannot change mode.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-3-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper.c | 15 ++++++++-------
13
1 file changed, 8 insertions(+), 7 deletions(-)
14
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
18
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
20
uint32_t valid_mask = 0x3fff;
21
ARMCPU *cpu = env_archcpu(env);
22
23
- if (arm_el_is_aa64(env, 3)) {
24
+ if (ri->state == ARM_CP_STATE_AA64) {
25
value |= SCR_FW | SCR_AW; /* these two bits are RES1. */
26
valid_mask &= ~SCR_NET;
27
+
28
+ if (cpu_isar_feature(aa64_lor, cpu)) {
29
+ valid_mask |= SCR_TLOR;
30
+ }
31
+ if (cpu_isar_feature(aa64_pauth, cpu)) {
32
+ valid_mask |= SCR_API | SCR_APK;
33
+ }
34
} else {
35
valid_mask &= ~(SCR_RW | SCR_ST);
36
}
37
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
38
valid_mask &= ~SCR_SMD;
39
}
40
}
41
- if (cpu_isar_feature(aa64_lor, cpu)) {
42
- valid_mask |= SCR_TLOR;
43
- }
44
- if (cpu_isar_feature(aa64_pauth, cpu)) {
45
- valid_mask |= SCR_API | SCR_APK;
46
- }
47
48
/* Clear all-context RES0 bits. */
49
value &= valid_mask;
50
--
51
2.20.1
52
53
diff view generated by jsdifflib
1
Convert VCMLA (scalar) in the 2reg-scalar-ext group to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This does not attempt to rectify all of the res0 bits, but does
4
clear the mte bits when not enabled. Since there is no high-part
5
mapping of SCTLR, aa32 mode cannot write to these bits.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-4-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200430181003.21682-9-peter.maydell@linaro.org
6
---
11
---
7
target/arm/neon-shared.decode | 5 +++++
12
target/arm/helper.c | 23 +++++++++++++++++------
8
target/arm/translate-neon.inc.c | 40 +++++++++++++++++++++++++++++++++
13
1 file changed, 17 insertions(+), 6 deletions(-)
9
target/arm/translate.c | 26 +--------------------
10
3 files changed, 46 insertions(+), 25 deletions(-)
11
14
12
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/neon-shared.decode
17
--- a/target/arm/helper.c
15
+++ b/target/arm/neon-shared.decode
18
+++ b/target/arm/helper.c
16
@@ -XXX,XX +XXX,XX @@ VFML 1111 110 0 s:1 . 10 .... .... 1000 . 0 . 1 .... \
19
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
17
vm=%vm_sp vn=%vn_sp vd=%vd_dp q=0
20
{
18
VFML 1111 110 0 s:1 . 10 .... .... 1000 . 1 . 1 .... \
21
ARMCPU *cpu = env_archcpu(env);
19
vm=%vm_dp vn=%vn_dp vd=%vd_dp q=1
22
20
+
23
+ if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
21
+VCMLA_scalar 1111 1110 0 . rot:2 .... .... 1000 . q:1 index:1 0 vm:4 \
24
+ /* M bit is RAZ/WI for PMSA with no MPU implemented */
22
+ vn=%vn_dp vd=%vd_dp size=0
25
+ value &= ~SCTLR_M;
23
+VCMLA_scalar 1111 1110 1 . rot:2 .... .... 1000 . q:1 . 0 .... \
24
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp size=1 index=0
25
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/translate-neon.inc.c
28
+++ b/target/arm/translate-neon.inc.c
29
@@ -XXX,XX +XXX,XX @@ static bool trans_VFML(DisasContext *s, arg_VFML *a)
30
gen_helper_gvec_fmlal_a32);
31
return true;
32
}
33
+
34
+static bool trans_VCMLA_scalar(DisasContext *s, arg_VCMLA_scalar *a)
35
+{
36
+ gen_helper_gvec_3_ptr *fn_gvec_ptr;
37
+ int opr_sz;
38
+ TCGv_ptr fpst;
39
+
40
+ if (!dc_isar_feature(aa32_vcma, s)) {
41
+ return false;
42
+ }
43
+ if (a->size == 0 && !dc_isar_feature(aa32_fp16_arith, s)) {
44
+ return false;
45
+ }
26
+ }
46
+
27
+
47
+ /* UNDEF accesses to D16-D31 if they don't exist. */
28
+ /* ??? Lots of these bits are not implemented. */
48
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
29
+
49
+ ((a->vd | a->vn | a->vm) & 0x10)) {
30
+ if (ri->state == ARM_CP_STATE_AA64 && !cpu_isar_feature(aa64_mte, cpu)) {
50
+ return false;
31
+ if (ri->opc1 == 6) { /* SCTLR_EL3 */
32
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF | SCTLR_ATA);
33
+ } else {
34
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF0 | SCTLR_TCF |
35
+ SCTLR_ATA0 | SCTLR_ATA);
36
+ }
51
+ }
37
+ }
52
+
38
+
53
+ if ((a->vd | a->vn) & a->q) {
39
if (raw_read(env, ri) == value) {
54
+ return false;
40
/* Skip the TLB flush if nothing actually changed; Linux likes
55
+ }
41
* to do a lot of pointless SCTLR writes.
42
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
43
return;
44
}
45
46
- if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
47
- /* M bit is RAZ/WI for PMSA with no MPU implemented */
48
- value &= ~SCTLR_M;
49
- }
50
-
51
raw_write(env, ri, value);
52
- /* ??? Lots of these bits are not implemented. */
56
+
53
+
57
+ if (!vfp_access_check(s)) {
54
/* This may enable/disable the MMU, so do a TLB flush. */
58
+ return true;
55
tlb_flush(CPU(cpu));
59
+ }
60
+
61
+ fn_gvec_ptr = (a->size ? gen_helper_gvec_fcmlas_idx
62
+ : gen_helper_gvec_fcmlah_idx);
63
+ opr_sz = (1 + a->q) * 8;
64
+ fpst = get_fpstatus_ptr(1);
65
+ tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd),
66
+ vfp_reg_offset(1, a->vn),
67
+ vfp_reg_offset(1, a->vm),
68
+ fpst, opr_sz, opr_sz,
69
+ (a->index << 2) | a->rot, fn_gvec_ptr);
70
+ tcg_temp_free_ptr(fpst);
71
+ return true;
72
+}
73
diff --git a/target/arm/translate.c b/target/arm/translate.c
74
index XXXXXXX..XXXXXXX 100644
75
--- a/target/arm/translate.c
76
+++ b/target/arm/translate.c
77
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
78
bool is_long = false, q = extract32(insn, 6, 1);
79
bool ptr_is_env = false;
80
81
- if ((insn & 0xff000f10) == 0xfe000800) {
82
- /* VCMLA (indexed) -- 1111 1110 S.RR .... .... 1000 ...0 .... */
83
- int rot = extract32(insn, 20, 2);
84
- int size = extract32(insn, 23, 1);
85
- int index;
86
-
87
- if (!dc_isar_feature(aa32_vcma, s)) {
88
- return 1;
89
- }
90
- if (size == 0) {
91
- if (!dc_isar_feature(aa32_fp16_arith, s)) {
92
- return 1;
93
- }
94
- /* For fp16, rm is just Vm, and index is M. */
95
- rm = extract32(insn, 0, 4);
96
- index = extract32(insn, 5, 1);
97
- } else {
98
- /* For fp32, rm is the usual M:Vm, and index is 0. */
99
- VFP_DREG_M(rm, insn);
100
- index = 0;
101
- }
102
- data = (index << 2) | rot;
103
- fn_gvec_ptr = (size ? gen_helper_gvec_fcmlas_idx
104
- : gen_helper_gvec_fcmlah_idx);
105
- } else if ((insn & 0xffb00f00) == 0xfe200d00) {
106
+ if ((insn & 0xffb00f00) == 0xfe200d00) {
107
/* V[US]DOT -- 1111 1110 0.10 .... .... 1101 .Q.U .... */
108
int u = extract32(insn, 4, 1);
109
56
110
--
57
--
111
2.20.1
58
2.20.1
112
59
113
60
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-5-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper.c | 14 +++++++++++---
9
1 file changed, 11 insertions(+), 3 deletions(-)
10
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/helper.c
14
+++ b/target/arm/helper.c
15
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
16
if (cpu_isar_feature(aa64_pauth, cpu)) {
17
valid_mask |= SCR_API | SCR_APK;
18
}
19
+ if (cpu_isar_feature(aa64_mte, cpu)) {
20
+ valid_mask |= SCR_ATA;
21
+ }
22
} else {
23
valid_mask &= ~(SCR_RW | SCR_ST);
24
}
25
@@ -XXX,XX +XXX,XX @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
26
if (cpu_isar_feature(aa64_pauth, cpu)) {
27
valid_mask |= HCR_API | HCR_APK;
28
}
29
+ if (cpu_isar_feature(aa64_mte, cpu)) {
30
+ valid_mask |= HCR_ATA | HCR_DCT | HCR_TID5;
31
+ }
32
}
33
34
/* Clear RES0 bits. */
35
value &= valid_mask;
36
37
- /* These bits change the MMU setup:
38
+ /*
39
+ * These bits change the MMU setup:
40
* HCR_VM enables stage 2 translation
41
* HCR_PTW forbids certain page-table setups
42
- * HCR_DC Disables stage1 and enables stage2 translation
43
+ * HCR_DC disables stage1 and enables stage2 translation
44
+ * HCR_DCT enables tagging on (disabled) stage1 translation
45
*/
46
- if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC)) {
47
+ if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC | HCR_DCT)) {
48
tlb_flush(CPU(cpu));
49
}
50
env->cp15.hcr_el2 = value;
51
--
52
2.20.1
53
54
diff view generated by jsdifflib
1
Convert the VFM[AS]L (scalar) insns in the 2reg-scalar-ext group
1
From: Richard Henderson <richard.henderson@linaro.org>
2
to decodetree. These are the last ones in the group so we can remove
3
all the legacy decode for the group.
4
2
5
Note that in disas_thumb2_insn() the parts of this encoding space
3
Emphasize that the is_jmp option exits to the main loop.
6
where the decodetree decoder returns false will correctly be directed
7
to illegal_op by the "(insn & (1 << 28))" check so they won't fall
8
into disas_coproc_insn() by mistake.
9
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-6-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200430181003.21682-11-peter.maydell@linaro.org
13
---
9
---
14
target/arm/neon-shared.decode | 7 +++
10
target/arm/translate.h | 14 ++++++++------
15
target/arm/translate-neon.inc.c | 32 ++++++++++
11
target/arm/translate-a64.c | 8 ++++----
16
target/arm/translate.c | 107 +-------------------------------
12
target/arm/translate-vfp.inc.c | 4 ++--
17
3 files changed, 40 insertions(+), 106 deletions(-)
13
target/arm/translate.c | 12 ++++++------
14
4 files changed, 20 insertions(+), 18 deletions(-)
18
15
19
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
20
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/neon-shared.decode
18
--- a/target/arm/translate.h
22
+++ b/target/arm/neon-shared.decode
19
+++ b/target/arm/translate.h
23
@@ -XXX,XX +XXX,XX @@ VCMLA_scalar 1111 1110 1 . rot:2 .... .... 1000 . q:1 . 0 .... \
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
24
21
25
VDOT_scalar 1111 1110 0 . 10 .... .... 1101 . q:1 index:1 u:1 rm:4 \
22
/* is_jmp field values */
26
vm=%vm_dp vn=%vn_dp vd=%vd_dp
23
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
27
+
24
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
28
+%vfml_scalar_q0_rm 0:3 5:1
25
+/* CPU state was modified dynamically; exit to main loop for interrupts. */
29
+%vfml_scalar_q1_index 5:1 3:1
26
+#define DISAS_UPDATE_EXIT DISAS_TARGET_1
30
+VFML_scalar 1111 1110 0 . 0 s:1 .... .... 1000 . 0 . 1 index:1 ... \
27
/* These instructions trap after executing, so the A32/T32 decoder must
31
+ rm=%vfml_scalar_q0_rm vn=%vn_sp vd=%vd_dp q=0
28
* defer them until after the conditional execution state has been updated.
32
+VFML_scalar 1111 1110 0 . 0 s:1 .... .... 1000 . 1 . 1 . rm:3 \
29
* WFI also needs special handling when single-stepping.
33
+ index=%vfml_scalar_q1_index vn=%vn_dp vd=%vd_dp q=1
30
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
34
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
31
* custom end-of-TB code)
32
*/
33
#define DISAS_BX_EXCRET DISAS_TARGET_8
34
-/* For instructions which want an immediate exit to the main loop,
35
- * as opposed to attempting to use lookup_and_goto_ptr. Unlike
36
- * DISAS_UPDATE this doesn't write the PC on exiting the translation
37
- * loop so you need to ensure something (gen_a64_set_pc_im or runtime
38
- * helper) has done so before we reach return from cpu_tb_exec.
39
+/*
40
+ * For instructions which want an immediate exit to the main loop, as opposed
41
+ * to attempting to use lookup_and_goto_ptr. Unlike DISAS_UPDATE_EXIT, this
42
+ * doesn't write the PC on exiting the translation loop so you need to ensure
43
+ * something (gen_a64_set_pc_im or runtime helper) has done so before we reach
44
+ * return from cpu_tb_exec.
45
*/
46
#define DISAS_EXIT DISAS_TARGET_9
47
48
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
35
index XXXXXXX..XXXXXXX 100644
49
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/translate-neon.inc.c
50
--- a/target/arm/translate-a64.c
37
+++ b/target/arm/translate-neon.inc.c
51
+++ b/target/arm/translate-a64.c
38
@@ -XXX,XX +XXX,XX @@ static bool trans_VDOT_scalar(DisasContext *s, arg_VDOT_scalar *a)
52
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
39
tcg_temp_free_ptr(fpst);
53
gen_helper_msr_i_daifclear(cpu_env, t1);
54
tcg_temp_free_i32(t1);
55
/* For DAIFClear, exit the cpu loop to re-evaluate pending IRQs. */
56
- s->base.is_jmp = DISAS_UPDATE;
57
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
58
break;
59
60
default:
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
63
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
64
/* I/O operations must end the TB here (whether read or write) */
65
- s->base.is_jmp = DISAS_UPDATE;
66
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
67
}
68
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
69
/*
70
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
71
* but allow this to be suppressed by the register definition
72
* (usually only necessary to work around guest bugs).
73
*/
74
- s->base.is_jmp = DISAS_UPDATE;
75
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
76
}
77
}
78
79
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
80
gen_goto_tb(dc, 1, dc->base.pc_next);
81
break;
82
default:
83
- case DISAS_UPDATE:
84
+ case DISAS_UPDATE_EXIT:
85
gen_a64_set_pc_im(dc->base.pc_next);
86
/* fall through */
87
case DISAS_EXIT:
88
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
89
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate-vfp.inc.c
91
+++ b/target/arm/translate-vfp.inc.c
92
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
93
* this to be the last insn in the TB).
94
*/
95
if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
96
- s->base.is_jmp = DISAS_UPDATE;
97
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
98
gen_io_start();
99
}
100
gen_helper_v7m_preserve_fp_state(cpu_env);
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
102
tcg_temp_free_i32(fptr);
103
104
/* End the TB, because we have updated FP control bits */
105
- s->base.is_jmp = DISAS_UPDATE;
106
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
40
return true;
107
return true;
41
}
108
}
42
+
43
+static bool trans_VFML_scalar(DisasContext *s, arg_VFML_scalar *a)
44
+{
45
+ int opr_sz;
46
+
47
+ if (!dc_isar_feature(aa32_fhm, s)) {
48
+ return false;
49
+ }
50
+
51
+ /* UNDEF accesses to D16-D31 if they don't exist. */
52
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
53
+ ((a->vd & 0x10) || (a->q && (a->vn & 0x10)))) {
54
+ return false;
55
+ }
56
+
57
+ if (a->vd & a->q) {
58
+ return false;
59
+ }
60
+
61
+ if (!vfp_access_check(s)) {
62
+ return true;
63
+ }
64
+
65
+ opr_sz = (1 + a->q) * 8;
66
+ tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd),
67
+ vfp_reg_offset(a->q, a->vn),
68
+ vfp_reg_offset(a->q, a->rm),
69
+ cpu_env, opr_sz, opr_sz,
70
+ (a->index << 2) | a->s, /* is_2 == 0 */
71
+ gen_helper_gvec_fmlal_idx_a32);
72
+ return true;
73
+}
74
diff --git a/target/arm/translate.c b/target/arm/translate.c
109
diff --git a/target/arm/translate.c b/target/arm/translate.c
75
index XXXXXXX..XXXXXXX 100644
110
index XXXXXXX..XXXXXXX 100644
76
--- a/target/arm/translate.c
111
--- a/target/arm/translate.c
77
+++ b/target/arm/translate.c
112
+++ b/target/arm/translate.c
78
@@ -XXX,XX +XXX,XX @@ static int disas_dsp_insn(DisasContext *s, uint32_t insn)
113
@@ -XXX,XX +XXX,XX @@ static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
114
tcg_temp_free_i32(tcg_tgtmode);
115
tcg_temp_free_i32(tcg_regno);
116
tcg_temp_free_i32(tcg_reg);
117
- s->base.is_jmp = DISAS_UPDATE;
118
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
79
}
119
}
80
120
81
#define VFP_REG_SHR(x, n) (((n) > 0) ? (x) >> (n) : (x) << -(n))
121
static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
82
-#define VFP_SREG(insn, bigbit, smallbit) \
122
@@ -XXX,XX +XXX,XX @@ static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
83
- ((VFP_REG_SHR(insn, bigbit - 1) & 0x1e) | (((insn) >> (smallbit)) & 1))
123
tcg_temp_free_i32(tcg_tgtmode);
84
#define VFP_DREG(reg, insn, bigbit, smallbit) do { \
124
tcg_temp_free_i32(tcg_regno);
85
if (dc_isar_feature(aa32_simd_r32, s)) { \
125
store_reg(s, rn, tcg_reg);
86
reg = (((insn) >> (bigbit)) & 0x0f) \
126
- s->base.is_jmp = DISAS_UPDATE;
87
@@ -XXX,XX +XXX,XX @@ static int disas_dsp_insn(DisasContext *s, uint32_t insn)
127
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
88
reg = ((insn) >> (bigbit)) & 0x0f; \
89
}} while (0)
90
91
-#define VFP_SREG_D(insn) VFP_SREG(insn, 12, 22)
92
#define VFP_DREG_D(reg, insn) VFP_DREG(reg, insn, 12, 22)
93
-#define VFP_SREG_N(insn) VFP_SREG(insn, 16, 7)
94
#define VFP_DREG_N(reg, insn) VFP_DREG(reg, insn, 16, 7)
95
-#define VFP_SREG_M(insn) VFP_SREG(insn, 0, 5)
96
#define VFP_DREG_M(reg, insn) VFP_DREG(reg, insn, 0, 5)
97
98
static void gen_neon_dup_low16(TCGv_i32 var)
99
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
100
return 0;
101
}
128
}
102
129
103
-/* Advanced SIMD two registers and a scalar extension.
130
/* Store value to PC as for an exception return (ie don't
104
- * 31 24 23 22 20 16 12 11 10 9 8 3 0
131
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
105
- * +-----------------+----+---+----+----+----+---+----+---+----+---------+----+
132
tcg_temp_free_i32(tmp);
106
- * | 1 1 1 1 1 1 1 0 | o1 | D | o2 | Vn | Vd | 1 | o3 | 0 | o4 | N Q M U | Vm |
107
- * +-----------------+----+---+----+----+----+---+----+---+----+---------+----+
108
- *
109
- */
110
-
111
-static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
112
-{
113
- gen_helper_gvec_3 *fn_gvec = NULL;
114
- gen_helper_gvec_3_ptr *fn_gvec_ptr = NULL;
115
- int rd, rn, rm, opr_sz, data;
116
- int off_rn, off_rm;
117
- bool is_long = false, q = extract32(insn, 6, 1);
118
- bool ptr_is_env = false;
119
-
120
- if ((insn & 0xffa00f10) == 0xfe000810) {
121
- /* VFM[AS]L -- 1111 1110 0.0S .... .... 1000 .Q.1 .... */
122
- int is_s = extract32(insn, 20, 1);
123
- int vm20 = extract32(insn, 0, 3);
124
- int vm3 = extract32(insn, 3, 1);
125
- int m = extract32(insn, 5, 1);
126
- int index;
127
-
128
- if (!dc_isar_feature(aa32_fhm, s)) {
129
- return 1;
130
- }
131
- if (q) {
132
- rm = vm20;
133
- index = m * 2 + vm3;
134
- } else {
135
- rm = vm20 * 2 + m;
136
- index = vm3;
137
- }
138
- is_long = true;
139
- data = (index << 2) | is_s; /* is_2 == 0 */
140
- fn_gvec_ptr = gen_helper_gvec_fmlal_idx_a32;
141
- ptr_is_env = true;
142
- } else {
143
- return 1;
144
- }
145
-
146
- VFP_DREG_D(rd, insn);
147
- if (rd & q) {
148
- return 1;
149
- }
150
- if (q || !is_long) {
151
- VFP_DREG_N(rn, insn);
152
- if (rn & q & !is_long) {
153
- return 1;
154
- }
155
- off_rn = vfp_reg_offset(1, rn);
156
- off_rm = vfp_reg_offset(1, rm);
157
- } else {
158
- rn = VFP_SREG_N(insn);
159
- off_rn = vfp_reg_offset(0, rn);
160
- off_rm = vfp_reg_offset(0, rm);
161
- }
162
- if (s->fp_excp_el) {
163
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
164
- syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
165
- return 0;
166
- }
167
- if (!s->vfp_enabled) {
168
- return 1;
169
- }
170
-
171
- opr_sz = (1 + q) * 8;
172
- if (fn_gvec_ptr) {
173
- TCGv_ptr ptr;
174
- if (ptr_is_env) {
175
- ptr = cpu_env;
176
- } else {
177
- ptr = get_fpstatus_ptr(1);
178
- }
179
- tcg_gen_gvec_3_ptr(vfp_reg_offset(1, rd), off_rn, off_rm, ptr,
180
- opr_sz, opr_sz, data, fn_gvec_ptr);
181
- if (!ptr_is_env) {
182
- tcg_temp_free_ptr(ptr);
183
- }
184
- } else {
185
- tcg_gen_gvec_3_ool(vfp_reg_offset(1, rd), off_rn, off_rm,
186
- opr_sz, opr_sz, data, fn_gvec);
187
- }
188
- return 0;
189
-}
190
-
191
static int disas_coproc_insn(DisasContext *s, uint32_t insn)
192
{
193
int cpnum, is64, crn, crm, opc1, opc2, isread, rt, rt2;
194
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
195
}
196
}
197
}
198
- } else if ((insn & 0x0f000a00) == 0x0e000800
199
- && arm_dc_feature(s, ARM_FEATURE_V8)) {
200
- if (disas_neon_insn_2reg_scalar_ext(s, insn)) {
201
- goto illegal_op;
202
- }
203
- return;
204
}
205
goto illegal_op;
206
}
133
}
207
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
134
tcg_temp_free_i32(addr);
208
}
135
- s->base.is_jmp = DISAS_UPDATE;
136
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
137
}
138
139
/* Generate a label used for skipping this instruction */
140
@@ -XXX,XX +XXX,XX @@ static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
141
}
142
if (a->E != (s->be_data == MO_BE)) {
143
gen_helper_setend(cpu_env);
144
- s->base.is_jmp = DISAS_UPDATE;
145
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
146
}
147
return true;
148
}
149
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
209
break;
150
break;
210
}
151
case DISAS_NEXT:
211
- if ((insn & 0xff000a00) == 0xfe000800
152
case DISAS_TOO_MANY:
212
- && arm_dc_feature(s, ARM_FEATURE_V8)) {
153
- case DISAS_UPDATE:
213
- /* The Thumb2 and ARM encodings are identical. */
154
+ case DISAS_UPDATE_EXIT:
214
- if (disas_neon_insn_2reg_scalar_ext(s, insn)) {
155
gen_set_pc_im(dc, dc->base.pc_next);
215
- goto illegal_op;
156
/* fall through */
216
- }
157
default:
217
- } else if (((insn >> 24) & 3) == 3) {
158
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
218
+ if (((insn >> 24) & 3) == 3) {
159
case DISAS_JUMP:
219
/* Translate into the equivalent ARM encoding. */
160
gen_goto_ptr();
220
insn = (insn & 0xe2ffffff) | ((insn & (1 << 28)) >> 4) | (1 << 28);
161
break;
221
if (disas_neon_data_insn(s, insn)) {
162
- case DISAS_UPDATE:
163
+ case DISAS_UPDATE_EXIT:
164
gen_set_pc_im(dc, dc->base.pc_next);
165
/* fall through */
166
default:
222
--
167
--
223
2.20.1
168
2.20.1
224
169
225
170
diff view generated by jsdifflib
1
Convert the Neon VQADD/VQSUB insns in the 3-reg-same grouping
1
From: Richard Henderson <richard.henderson@linaro.org>
2
to decodetree.
3
2
3
Add an option that writes back the PC, like DISAS_UPDATE_EXIT,
4
but does not exit back to the main loop.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-7-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-19-peter.maydell@linaro.org
7
---
10
---
8
target/arm/neon-dp.decode | 6 ++++++
11
target/arm/translate.h | 2 ++
9
target/arm/translate-neon.inc.c | 15 +++++++++++++++
12
target/arm/translate-a64.c | 3 +++
10
target/arm/translate.c | 14 ++------------
13
target/arm/translate.c | 4 ++++
11
3 files changed, 23 insertions(+), 12 deletions(-)
14
3 files changed, 9 insertions(+)
12
15
13
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
14
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/neon-dp.decode
18
--- a/target/arm/translate.h
16
+++ b/target/arm/neon-dp.decode
19
+++ b/target/arm/translate.h
17
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
18
@3same .... ... . . . size:2 .... .... .... . q:1 . . .... \
21
* return from cpu_tb_exec.
19
&3same vm=%vm_dp vn=%vn_dp vd=%vd_dp
22
*/
20
23
#define DISAS_EXIT DISAS_TARGET_9
21
+VQADD_S_3s 1111 001 0 0 . .. .... .... 0000 . . . 1 .... @3same
24
+/* CPU state was modified dynamically; no need to exit, but do not chain. */
22
+VQADD_U_3s 1111 001 1 0 . .. .... .... 0000 . . . 1 .... @3same
25
+#define DISAS_UPDATE_NOCHAIN DISAS_TARGET_10
23
+
26
24
@3same_logic .... ... . . . .. .... .... .... . q:1 .. .... \
27
#ifdef TARGET_AARCH64
25
&3same vm=%vm_dp vn=%vn_dp vd=%vd_dp size=0
28
void a64_translate_init(void);
26
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
27
@@ -XXX,XX +XXX,XX @@ VBSL_3s 1111 001 1 0 . 01 .... .... 0001 ... 1 .... @3same_logic
28
VBIT_3s 1111 001 1 0 . 10 .... .... 0001 ... 1 .... @3same_logic
29
VBIF_3s 1111 001 1 0 . 11 .... .... 0001 ... 1 .... @3same_logic
30
31
+VQSUB_S_3s 1111 001 0 0 . .. .... .... 0010 . . . 1 .... @3same
32
+VQSUB_U_3s 1111 001 1 0 . .. .... .... 0010 . . . 1 .... @3same
33
+
34
VCGT_S_3s 1111 001 0 0 . .. .... .... 0011 . . . 0 .... @3same
35
VCGT_U_3s 1111 001 1 0 . .. .... .... 0011 . . . 0 .... @3same
36
VCGE_S_3s 1111 001 0 0 . .. .... .... 0011 . . . 1 .... @3same
37
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
38
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
39
--- a/target/arm/translate-neon.inc.c
31
--- a/target/arm/translate-a64.c
40
+++ b/target/arm/translate-neon.inc.c
32
+++ b/target/arm/translate-a64.c
41
@@ -XXX,XX +XXX,XX @@ static void gen_VTST_3s(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
33
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
42
tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, oprsz, maxsz, &cmtst_op[vece]);
34
case DISAS_EXIT:
43
}
35
tcg_gen_exit_tb(NULL, 0);
44
DO_3SAME_NO_SZ_3(VTST, gen_VTST_3s)
36
break;
45
+
37
+ case DISAS_UPDATE_NOCHAIN:
46
+#define DO_3SAME_GVEC4(INSN, OPARRAY) \
38
+ gen_a64_set_pc_im(dc->base.pc_next);
47
+ static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
39
+ /* fall through */
48
+ uint32_t rn_ofs, uint32_t rm_ofs, \
40
case DISAS_JUMP:
49
+ uint32_t oprsz, uint32_t maxsz) \
41
tcg_gen_lookup_and_goto_ptr();
50
+ { \
42
break;
51
+ tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc), \
52
+ rn_ofs, rm_ofs, oprsz, maxsz, &OPARRAY[vece]); \
53
+ } \
54
+ DO_3SAME(INSN, gen_##INSN##_3s)
55
+
56
+DO_3SAME_GVEC4(VQADD_S, sqadd_op)
57
+DO_3SAME_GVEC4(VQADD_U, uqadd_op)
58
+DO_3SAME_GVEC4(VQSUB_S, sqsub_op)
59
+DO_3SAME_GVEC4(VQSUB_U, uqsub_op)
60
diff --git a/target/arm/translate.c b/target/arm/translate.c
43
diff --git a/target/arm/translate.c b/target/arm/translate.c
61
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
62
--- a/target/arm/translate.c
45
--- a/target/arm/translate.c
63
+++ b/target/arm/translate.c
46
+++ b/target/arm/translate.c
64
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
47
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
65
}
48
case DISAS_NEXT:
66
return 1;
49
case DISAS_TOO_MANY:
67
50
case DISAS_UPDATE_EXIT:
68
- case NEON_3R_VQADD:
51
+ case DISAS_UPDATE_NOCHAIN:
69
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
52
gen_set_pc_im(dc, dc->base.pc_next);
70
- rn_ofs, rm_ofs, vec_size, vec_size,
53
/* fall through */
71
- (u ? uqadd_op : sqadd_op) + size);
54
default:
72
- return 0;
55
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
73
-
56
case DISAS_TOO_MANY:
74
- case NEON_3R_VQSUB:
57
gen_goto_tb(dc, 1, dc->base.pc_next);
75
- tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
58
break;
76
- rn_ofs, rm_ofs, vec_size, vec_size,
59
+ case DISAS_UPDATE_NOCHAIN:
77
- (u ? uqsub_op : sqsub_op) + size);
60
+ gen_set_pc_im(dc, dc->base.pc_next);
78
- return 0;
61
+ /* fall through */
79
-
62
case DISAS_JUMP:
80
case NEON_3R_VMUL: /* VMUL */
63
gen_goto_ptr();
81
if (u) {
64
break;
82
/* Polynomial case allows only P8. */
83
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
84
case NEON_3R_VTST_VCEQ:
85
case NEON_3R_VCGT:
86
case NEON_3R_VCGE:
87
+ case NEON_3R_VQADD:
88
+ case NEON_3R_VQSUB:
89
/* Already handled by decodetree */
90
return 1;
91
}
92
--
65
--
93
2.20.1
66
2.20.1
94
67
95
68
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
This is TFSRE0_EL1, TFSR_EL1, TFSR_EL2, TFSR_EL3,
4
RGSR_EL1, GCR_EL1, GMID_EL1, and PSTATE.TCO.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-8-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/cpu.h | 4 ++
12
target/arm/internals.h | 9 ++++
13
target/arm/helper.c | 94 ++++++++++++++++++++++++++++++++++++++
14
target/arm/translate-a64.c | 21 +++++++++
15
4 files changed, 128 insertions(+)
16
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
uint64_t pmccfiltr_el0; /* Performance Monitor Filter Register */
23
uint64_t vpidr_el2; /* Virtualization Processor ID Register */
24
uint64_t vmpidr_el2; /* Virtualization Multiprocessor ID Register */
25
+ uint64_t tfsr_el[4]; /* tfsre0_el1 is index 0. */
26
+ uint64_t gcr_el1;
27
+ uint64_t rgsr_el1;
28
} cp15;
29
30
struct {
31
@@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu);
32
#define PSTATE_SS (1U << 21)
33
#define PSTATE_PAN (1U << 22)
34
#define PSTATE_UAO (1U << 23)
35
+#define PSTATE_TCO (1U << 25)
36
#define PSTATE_V (1U << 28)
37
#define PSTATE_C (1U << 29)
38
#define PSTATE_Z (1U << 30)
39
diff --git a/target/arm/internals.h b/target/arm/internals.h
40
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/internals.h
42
+++ b/target/arm/internals.h
43
@@ -XXX,XX +XXX,XX @@ static inline uint32_t aarch64_pstate_valid_mask(const ARMISARegisters *id)
44
if (isar_feature_aa64_uao(id)) {
45
valid |= PSTATE_UAO;
46
}
47
+ if (isar_feature_aa64_mte(id)) {
48
+ valid |= PSTATE_TCO;
49
+ }
50
51
return valid;
52
}
53
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
54
55
#endif /* !CONFIG_USER_ONLY */
56
57
+/*
58
+ * The log2 of the words in the tag block, for GMID_EL1.BS.
59
+ * The is the maximum, 256 bytes, which manipulates 64-bits of tags.
60
+ */
61
+#define GMID_EL1_BS 6
62
+
63
#endif
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/helper.c
67
+++ b/target/arm/helper.c
68
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
69
{ K(3, 0, 1, 2, 0), K(3, 4, 1, 2, 0), K(3, 5, 1, 2, 0),
70
"ZCR_EL1", "ZCR_EL2", "ZCR_EL12", isar_feature_aa64_sve },
71
72
+ { K(3, 0, 5, 6, 0), K(3, 4, 5, 6, 0), K(3, 5, 5, 6, 0),
73
+ "TFSR_EL1", "TFSR_EL2", "TFSR_EL12", isar_feature_aa64_mte },
74
+
75
/* TODO: ARMv8.2-SPE -- PMSCR_EL2 */
76
/* TODO: ARMv8.4-Trace -- TRFCR_EL2 */
77
};
78
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
79
};
80
#endif /*CONFIG_USER_ONLY*/
81
82
+static CPAccessResult access_aa64_tid5(CPUARMState *env, const ARMCPRegInfo *ri,
83
+ bool isread)
84
+{
85
+ if ((arm_current_el(env) < 2) && (arm_hcr_el2_eff(env) & HCR_TID5)) {
86
+ return CP_ACCESS_TRAP_EL2;
87
+ }
88
+
89
+ return CP_ACCESS_OK;
90
+}
91
+
92
+static CPAccessResult access_mte(CPUARMState *env, const ARMCPRegInfo *ri,
93
+ bool isread)
94
+{
95
+ int el = arm_current_el(env);
96
+
97
+ if (el < 2 &&
98
+ arm_feature(env, ARM_FEATURE_EL2) &&
99
+ !(arm_hcr_el2_eff(env) & HCR_ATA)) {
100
+ return CP_ACCESS_TRAP_EL2;
101
+ }
102
+ if (el < 3 &&
103
+ arm_feature(env, ARM_FEATURE_EL3) &&
104
+ !(env->cp15.scr_el3 & SCR_ATA)) {
105
+ return CP_ACCESS_TRAP_EL3;
106
+ }
107
+ return CP_ACCESS_OK;
108
+}
109
+
110
+static uint64_t tco_read(CPUARMState *env, const ARMCPRegInfo *ri)
111
+{
112
+ return env->pstate & PSTATE_TCO;
113
+}
114
+
115
+static void tco_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t val)
116
+{
117
+ env->pstate = (env->pstate & ~PSTATE_TCO) | (val & PSTATE_TCO);
118
+}
119
+
120
+static const ARMCPRegInfo mte_reginfo[] = {
121
+ { .name = "TFSRE0_EL1", .state = ARM_CP_STATE_AA64,
122
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 1,
123
+ .access = PL1_RW, .accessfn = access_mte,
124
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[0]) },
125
+ { .name = "TFSR_EL1", .state = ARM_CP_STATE_AA64,
126
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 0,
127
+ .access = PL1_RW, .accessfn = access_mte,
128
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[1]) },
129
+ { .name = "TFSR_EL2", .state = ARM_CP_STATE_AA64,
130
+ .opc0 = 3, .opc1 = 4, .crn = 5, .crm = 6, .opc2 = 0,
131
+ .access = PL2_RW, .accessfn = access_mte,
132
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[2]) },
133
+ { .name = "TFSR_EL3", .state = ARM_CP_STATE_AA64,
134
+ .opc0 = 3, .opc1 = 6, .crn = 5, .crm = 6, .opc2 = 0,
135
+ .access = PL3_RW,
136
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[3]) },
137
+ { .name = "RGSR_EL1", .state = ARM_CP_STATE_AA64,
138
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 5,
139
+ .access = PL1_RW, .accessfn = access_mte,
140
+ .fieldoffset = offsetof(CPUARMState, cp15.rgsr_el1) },
141
+ { .name = "GCR_EL1", .state = ARM_CP_STATE_AA64,
142
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 6,
143
+ .access = PL1_RW, .accessfn = access_mte,
144
+ .fieldoffset = offsetof(CPUARMState, cp15.gcr_el1) },
145
+ { .name = "GMID_EL1", .state = ARM_CP_STATE_AA64,
146
+ .opc0 = 3, .opc1 = 1, .crn = 0, .crm = 0, .opc2 = 4,
147
+ .access = PL1_R, .accessfn = access_aa64_tid5,
148
+ .type = ARM_CP_CONST, .resetvalue = GMID_EL1_BS },
149
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
150
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
151
+ .type = ARM_CP_NO_RAW,
152
+ .access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
153
+ REGINFO_SENTINEL
154
+};
155
+
156
+static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
157
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
158
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
159
+ .type = ARM_CP_CONST, .access = PL0_RW, },
160
+ REGINFO_SENTINEL
161
+};
162
#endif
163
164
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
165
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
166
}
167
}
168
#endif /*CONFIG_USER_ONLY*/
169
+
170
+ /*
171
+ * If full MTE is enabled, add all of the system registers.
172
+ * If only "instructions available at EL0" are enabled,
173
+ * then define only a RAZ/WI version of PSTATE.TCO.
174
+ */
175
+ if (cpu_isar_feature(aa64_mte, cpu)) {
176
+ define_arm_cp_regs(cpu, mte_reginfo);
177
+ } else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
178
+ define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
179
+ }
180
#endif
181
182
if (cpu_isar_feature(any_predinv, cpu)) {
183
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
184
index XXXXXXX..XXXXXXX 100644
185
--- a/target/arm/translate-a64.c
186
+++ b/target/arm/translate-a64.c
187
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
188
s->base.is_jmp = DISAS_UPDATE_EXIT;
189
break;
190
191
+ case 0x1c: /* TCO */
192
+ if (dc_isar_feature(aa64_mte, s)) {
193
+ /* Full MTE is enabled -- set the TCO bit as directed. */
194
+ if (crm & 1) {
195
+ set_pstate_bits(PSTATE_TCO);
196
+ } else {
197
+ clear_pstate_bits(PSTATE_TCO);
198
+ }
199
+ t1 = tcg_const_i32(s->current_el);
200
+ gen_helper_rebuild_hflags_a64(cpu_env, t1);
201
+ tcg_temp_free_i32(t1);
202
+ /* Many factors, including TCO, go into MTE_ACTIVE. */
203
+ s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
204
+ } else if (dc_isar_feature(aa64_mte_insn_reg, s)) {
205
+ /* Only "instructions accessible at EL0" -- PSTATE.TCO is WI. */
206
+ s->base.is_jmp = DISAS_NEXT;
207
+ } else {
208
+ goto do_unallocated;
209
+ }
210
+ break;
211
+
212
default:
213
do_unallocated:
214
unallocated_encoding(s);
215
--
216
2.20.1
217
218
diff view generated by jsdifflib
1
We define ARMMMUIdx_Stage2 as being an MMU index which uses a QEMU
1
From: Richard Henderson <richard.henderson@linaro.org>
2
TLB. However we never actually use the TLB -- all stage 2 lookups
3
are done by direct calls to get_phys_addr_lpae() followed by a
4
physical address load via address_space_ld*().
5
2
6
Remove Stage2 from the list of ARM MMU indexes which correspond to
3
Cache the composite ATA setting.
7
real core MMU indexes, and instead put it in the set of "NOTLB" ARM
8
MMU indexes.
9
4
10
This allows us to drop NB_MMU_MODES to 11. It also means we can
5
Cache when MTE is fully enabled, i.e. access to tags are enabled
11
safely add support for the ARMv8.3-TTS2UXN extension, which adds
6
and tag checks affect the PE. Do this for both the normal context
12
permission bits to the stage 2 descriptors which define execute
7
and the UNPRIV context.
13
permission separatel for EL0 and EL1; supporting that while keeping
14
Stage2 in a QEMU TLB would require us to use separate TLBs for
15
"Stage2 for an EL0 access" and "Stage2 for an EL1 access", which is a
16
lot of extra complication given we aren't even using the QEMU TLB.
17
8
18
In the process of updating the comment on our MMU index use,
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
19
fix a couple of other minor errors:
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
20
* NS EL2 EL2&0 was missing from the list in the comment
11
Message-id: 20200626033144.790098-9-richard.henderson@linaro.org
21
* some text hadn't been updated from when we bumped NB_MMU_MODES
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
22
above 8
13
---
14
target/arm/cpu.h | 12 ++++++++----
15
target/arm/internals.h | 18 +++++++++++++++++
16
target/arm/translate.h | 5 +++++
17
target/arm/helper.c | 40 ++++++++++++++++++++++++++++++++++++++
18
target/arm/translate-a64.c | 4 ++++
19
5 files changed, 75 insertions(+), 4 deletions(-)
23
20
24
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
25
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
26
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
27
Message-id: 20200330210400.11724-2-peter.maydell@linaro.org
28
---
29
target/arm/cpu-param.h | 2 +-
30
target/arm/cpu.h | 21 +++++---
31
target/arm/helper.c | 112 ++++-------------------------------------
32
3 files changed, 27 insertions(+), 108 deletions(-)
33
34
diff --git a/target/arm/cpu-param.h b/target/arm/cpu-param.h
35
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/cpu-param.h
37
+++ b/target/arm/cpu-param.h
38
@@ -XXX,XX +XXX,XX @@
39
# define TARGET_PAGE_BITS_MIN 10
40
#endif
41
42
-#define NB_MMU_MODES 12
43
+#define NB_MMU_MODES 11
44
45
#endif
46
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
47
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
48
--- a/target/arm/cpu.h
23
--- a/target/arm/cpu.h
49
+++ b/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
50
@@ -XXX,XX +XXX,XX @@ bool write_cpustate_to_list(ARMCPU *cpu, bool kvm_sync);
25
@@ -XXX,XX +XXX,XX @@ typedef ARMCPU ArchCPU;
51
* handling via the TLB. The only way to do a stage 1 translation without
26
* | | | TBFLAG_A32 | |
52
* the immediate stage 2 translation is via the ATS or AT system insns,
27
* | | +-----+----------+ TBFLAG_AM32 |
53
* which can be slow-pathed and always do a page table walk.
28
* | TBFLAG_ANY | |TBFLAG_M32| |
54
+ * The only use of stage 2 translations is either as part of an s1+2
29
- * | | +-+----------+--------------|
55
+ * lookup or when loading the descriptors during a stage 1 page table walk,
30
- * | | | TBFLAG_A64 |
56
+ * and in both those cases we don't use the TLB.
31
- * +--------------+---------+---------------------------+
57
* 4. we can also safely fold together the "32 bit EL3" and "64 bit EL3"
32
- * 31 20 15 0
58
* translation regimes, because they map reasonably well to each other
33
+ * | +-----------+----------+--------------|
59
* and they can't both be active at the same time.
34
+ * | | TBFLAG_A64 |
60
@@ -XXX,XX +XXX,XX @@ bool write_cpustate_to_list(ARMCPU *cpu, bool kvm_sync);
35
+ * +--------------+-------------------------------------+
61
* NS EL1 EL1&0 stage 1+2 (aka NS PL1)
36
+ * 31 20 0
62
* NS EL1 EL1&0 stage 1+2 +PAN
63
* NS EL0 EL2&0
64
+ * NS EL2 EL2&0
65
* NS EL2 EL2&0 +PAN
66
* NS EL2 (aka NS PL2)
67
* S EL0 EL1&0 (aka S PL0)
68
* S EL1 EL1&0 (not used if EL3 is 32 bit)
69
* S EL1 EL1&0 +PAN
70
* S EL3 (aka S PL1)
71
- * NS EL1&0 stage 2
72
*
37
*
73
- * for a total of 12 different mmu_idx.
38
* Unless otherwise noted, these bits are cached in env->hflags.
74
+ * for a total of 11 different mmu_idx.
39
*/
75
*
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, BT, 9, 1)
76
* R profile CPUs have an MPU, but can use the same set of MMU indexes
41
FIELD(TBFLAG_A64, BTYPE, 10, 2) /* Not cached. */
77
* as A profile. They only need to distinguish NS EL0 and NS EL1 (and
42
FIELD(TBFLAG_A64, TBID, 12, 2)
78
@@ -XXX,XX +XXX,XX @@ bool write_cpustate_to_list(ARMCPU *cpu, bool kvm_sync);
43
FIELD(TBFLAG_A64, UNPRIV, 14, 1)
79
* are not quite the same -- different CPU types (most notably M profile
44
+FIELD(TBFLAG_A64, ATA, 15, 1)
80
* vs A/R profile) would like to use MMU indexes with different semantics,
45
+FIELD(TBFLAG_A64, TCMA, 16, 2)
81
* but since we don't ever need to use all of those in a single CPU we
46
+FIELD(TBFLAG_A64, MTE_ACTIVE, 18, 1)
82
- * can avoid setting NB_MMU_MODES to more than 8. The lower bits of
47
+FIELD(TBFLAG_A64, MTE0_ACTIVE, 19, 1)
83
+ * can avoid having to set NB_MMU_MODES to "total number of A profile MMU
48
84
+ * modes + total number of M profile MMU modes". The lower bits of
49
/**
85
* ARMMMUIdx are the core TLB mmu index, and the higher bits are always
50
* cpu_mmu_index:
86
* the same for any particular CPU.
51
diff --git a/target/arm/internals.h b/target/arm/internals.h
87
* Variables of type ARMMUIdx are always full values, and the core
52
index XXXXXXX..XXXXXXX 100644
88
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdx {
53
--- a/target/arm/internals.h
89
ARMMMUIdx_SE10_1_PAN = 9 | ARM_MMU_IDX_A,
54
+++ b/target/arm/internals.h
90
ARMMMUIdx_SE3 = 10 | ARM_MMU_IDX_A,
55
@@ -XXX,XX +XXX,XX @@ static inline int exception_target_el(CPUARMState *env)
91
56
return target_el;
92
- ARMMMUIdx_Stage2 = 11 | ARM_MMU_IDX_A,
57
}
93
-
58
94
/*
59
+/* Determine if allocation tags are available. */
95
* These are not allocated TLBs and are used only for AT system
60
+static inline bool allocation_tag_access_enabled(CPUARMState *env, int el,
96
* instructions or for the first stage of an S12 page table walk.
61
+ uint64_t sctlr)
97
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdx {
62
+{
98
ARMMMUIdx_Stage1_E0 = 0 | ARM_MMU_IDX_NOTLB,
63
+ if (el < 3
99
ARMMMUIdx_Stage1_E1 = 1 | ARM_MMU_IDX_NOTLB,
64
+ && arm_feature(env, ARM_FEATURE_EL3)
100
ARMMMUIdx_Stage1_E1_PAN = 2 | ARM_MMU_IDX_NOTLB,
65
+ && !(env->cp15.scr_el3 & SCR_ATA)) {
101
+ /*
66
+ return false;
102
+ * Not allocated a TLB: used only for second stage of an S12 page
67
+ }
103
+ * table walk, or for descriptor loads during first stage of an S1
68
+ if (el < 2
104
+ * page table walk. Note that if we ever want to have a TLB for this
69
+ && arm_feature(env, ARM_FEATURE_EL2)
105
+ * then various TLB flush insns which currently are no-ops or flush
70
+ && !(arm_hcr_el2_eff(env) & HCR_ATA)) {
106
+ * only stage 1 MMU indexes will need to change to flush stage 2.
71
+ return false;
107
+ */
72
+ }
108
+ ARMMMUIdx_Stage2 = 3 | ARM_MMU_IDX_NOTLB,
73
+ sctlr &= (el == 0 ? SCTLR_ATA0 : SCTLR_ATA);
109
74
+ return sctlr != 0;
110
/*
75
+}
111
* M-profile.
76
+
112
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
77
#ifndef CONFIG_USER_ONLY
113
TO_CORE_BIT(SE10_1),
78
114
TO_CORE_BIT(SE10_1_PAN),
79
/* Security attributes for an address, as returned by v8m_security_lookup. */
115
TO_CORE_BIT(SE3),
80
diff --git a/target/arm/translate.h b/target/arm/translate.h
116
- TO_CORE_BIT(Stage2),
81
index XXXXXXX..XXXXXXX 100644
117
82
--- a/target/arm/translate.h
118
TO_CORE_BIT(MUser),
83
+++ b/target/arm/translate.h
119
TO_CORE_BIT(MPriv),
84
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
85
ARMMMUIdx mmu_idx; /* MMU index to use for normal loads/stores */
86
uint8_t tbii; /* TBI1|TBI0 for insns */
87
uint8_t tbid; /* TBI1|TBI0 for data */
88
+ uint8_t tcma; /* TCMA1|TCMA0 for MTE */
89
bool ns; /* Use non-secure CPREG bank on access */
90
int fp_excp_el; /* FP exception EL or 0 if enabled */
91
int sve_excp_el; /* SVE exception EL or 0 if enabled */
92
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
93
bool unpriv;
94
/* True if v8.3-PAuth is active. */
95
bool pauth_active;
96
+ /* True if v8.5-MTE access to tags is enabled. */
97
+ bool ata;
98
+ /* True if v8.5-MTE tag checks affect the PE; index with is_unpriv. */
99
+ bool mte_active[2];
100
/* True with v8.5-BTI and SCTLR_ELx.BT* set. */
101
bool bt;
102
/* True if any CP15 access is trapped by HSTR_EL2 */
120
diff --git a/target/arm/helper.c b/target/arm/helper.c
103
diff --git a/target/arm/helper.c b/target/arm/helper.c
121
index XXXXXXX..XXXXXXX 100644
104
index XXXXXXX..XXXXXXX 100644
122
--- a/target/arm/helper.c
105
--- a/target/arm/helper.c
123
+++ b/target/arm/helper.c
106
+++ b/target/arm/helper.c
124
@@ -XXX,XX +XXX,XX @@ static void tlbiall_nsnh_write(CPUARMState *env, const ARMCPRegInfo *ri,
107
@@ -XXX,XX +XXX,XX @@ static int aa64_va_parameter_tbid(uint64_t tcr, ARMMMUIdx mmu_idx)
125
tlb_flush_by_mmuidx(cs,
126
ARMMMUIdxBit_E10_1 |
127
ARMMMUIdxBit_E10_1_PAN |
128
- ARMMMUIdxBit_E10_0 |
129
- ARMMMUIdxBit_Stage2);
130
+ ARMMMUIdxBit_E10_0);
131
}
132
133
static void tlbiall_nsnh_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
134
@@ -XXX,XX +XXX,XX @@ static void tlbiall_nsnh_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
135
tlb_flush_by_mmuidx_all_cpus_synced(cs,
136
ARMMMUIdxBit_E10_1 |
137
ARMMMUIdxBit_E10_1_PAN |
138
- ARMMMUIdxBit_E10_0 |
139
- ARMMMUIdxBit_Stage2);
140
+ ARMMMUIdxBit_E10_0);
141
}
142
143
-static void tlbiipas2_write(CPUARMState *env, const ARMCPRegInfo *ri,
144
- uint64_t value)
145
-{
146
- /* Invalidate by IPA. This has to invalidate any structures that
147
- * contain only stage 2 translation information, but does not need
148
- * to apply to structures that contain combined stage 1 and stage 2
149
- * translation information.
150
- * This must NOP if EL2 isn't implemented or SCR_EL3.NS is zero.
151
- */
152
- CPUState *cs = env_cpu(env);
153
- uint64_t pageaddr;
154
-
155
- if (!arm_feature(env, ARM_FEATURE_EL2) || !(env->cp15.scr_el3 & SCR_NS)) {
156
- return;
157
- }
158
-
159
- pageaddr = sextract64(value << 12, 0, 40);
160
-
161
- tlb_flush_page_by_mmuidx(cs, pageaddr, ARMMMUIdxBit_Stage2);
162
-}
163
-
164
-static void tlbiipas2_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
165
- uint64_t value)
166
-{
167
- CPUState *cs = env_cpu(env);
168
- uint64_t pageaddr;
169
-
170
- if (!arm_feature(env, ARM_FEATURE_EL2) || !(env->cp15.scr_el3 & SCR_NS)) {
171
- return;
172
- }
173
-
174
- pageaddr = sextract64(value << 12, 0, 40);
175
-
176
- tlb_flush_page_by_mmuidx_all_cpus_synced(cs, pageaddr,
177
- ARMMMUIdxBit_Stage2);
178
-}
179
180
static void tlbiall_hyp_write(CPUARMState *env, const ARMCPRegInfo *ri,
181
uint64_t value)
182
@@ -XXX,XX +XXX,XX @@ static void vttbr_write(CPUARMState *env, const ARMCPRegInfo *ri,
183
tlb_flush_by_mmuidx(cs,
184
ARMMMUIdxBit_E10_1 |
185
ARMMMUIdxBit_E10_1_PAN |
186
- ARMMMUIdxBit_E10_0 |
187
- ARMMMUIdxBit_Stage2);
188
+ ARMMMUIdxBit_E10_0);
189
raw_write(env, ri, value);
190
}
108
}
191
}
109
}
192
@@ -XXX,XX +XXX,XX @@ static int alle1_tlbmask(CPUARMState *env)
110
193
return ARMMMUIdxBit_SE10_1 |
111
+static int aa64_va_parameter_tcma(uint64_t tcr, ARMMMUIdx mmu_idx)
194
ARMMMUIdxBit_SE10_1_PAN |
112
+{
195
ARMMMUIdxBit_SE10_0;
113
+ if (regime_has_2_ranges(mmu_idx)) {
196
- } else if (arm_feature(env, ARM_FEATURE_EL2)) {
114
+ return extract64(tcr, 57, 2);
197
- return ARMMMUIdxBit_E10_1 |
115
+ } else {
198
- ARMMMUIdxBit_E10_1_PAN |
116
+ /* Replicate the single TCMA bit so we always have 2 bits. */
199
- ARMMMUIdxBit_E10_0 |
117
+ return extract32(tcr, 30, 1) * 3;
200
- ARMMMUIdxBit_Stage2;
118
+ }
201
} else {
119
+}
202
return ARMMMUIdxBit_E10_1 |
120
+
203
ARMMMUIdxBit_E10_1_PAN |
121
ARMVAParameters aa64_va_parameters(CPUARMState *env, uint64_t va,
204
@@ -XXX,XX +XXX,XX @@ static void tlbi_aa64_vae3is_write(CPUARMState *env, const ARMCPRegInfo *ri,
122
ARMMMUIdx mmu_idx, bool data)
205
ARMMMUIdxBit_SE3);
123
{
124
@@ -XXX,XX +XXX,XX @@ static uint32_t rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
125
}
126
}
127
128
+ if (cpu_isar_feature(aa64_mte, env_archcpu(env))) {
129
+ /*
130
+ * Set MTE_ACTIVE if any access may be Checked, and leave clear
131
+ * if all accesses must be Unchecked:
132
+ * 1) If no TBI, then there are no tags in the address to check,
133
+ * 2) If Tag Check Override, then all accesses are Unchecked,
134
+ * 3) If Tag Check Fail == 0, then Checked access have no effect,
135
+ * 4) If no Allocation Tag Access, then all accesses are Unchecked.
136
+ */
137
+ if (allocation_tag_access_enabled(env, el, sctlr)) {
138
+ flags = FIELD_DP32(flags, TBFLAG_A64, ATA, 1);
139
+ if (tbid
140
+ && !(env->pstate & PSTATE_TCO)
141
+ && (sctlr & (el == 0 ? SCTLR_TCF0 : SCTLR_TCF))) {
142
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE_ACTIVE, 1);
143
+ }
144
+ }
145
+ /* And again for unprivileged accesses, if required. */
146
+ if (FIELD_EX32(flags, TBFLAG_A64, UNPRIV)
147
+ && tbid
148
+ && !(env->pstate & PSTATE_TCO)
149
+ && (sctlr & SCTLR_TCF0)
150
+ && allocation_tag_access_enabled(env, 0, sctlr)) {
151
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE0_ACTIVE, 1);
152
+ }
153
+ /* Cache TCMA as well as TBI. */
154
+ flags = FIELD_DP32(flags, TBFLAG_A64, TCMA,
155
+ aa64_va_parameter_tcma(tcr, mmu_idx));
156
+ }
157
+
158
return rebuild_hflags_common(env, fp_el, mmu_idx, flags);
206
}
159
}
207
160
208
-static void tlbi_aa64_ipas2e1_write(CPUARMState *env, const ARMCPRegInfo *ri,
161
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
209
- uint64_t value)
162
index XXXXXXX..XXXXXXX 100644
210
-{
163
--- a/target/arm/translate-a64.c
211
- /* Invalidate by IPA. This has to invalidate any structures that
164
+++ b/target/arm/translate-a64.c
212
- * contain only stage 2 translation information, but does not need
165
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
213
- * to apply to structures that contain combined stage 1 and stage 2
166
dc->mmu_idx = core_to_aa64_mmu_idx(core_mmu_idx);
214
- * translation information.
167
dc->tbii = FIELD_EX32(tb_flags, TBFLAG_A64, TBII);
215
- * This must NOP if EL2 isn't implemented or SCR_EL3.NS is zero.
168
dc->tbid = FIELD_EX32(tb_flags, TBFLAG_A64, TBID);
216
- */
169
+ dc->tcma = FIELD_EX32(tb_flags, TBFLAG_A64, TCMA);
217
- ARMCPU *cpu = env_archcpu(env);
170
dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
218
- CPUState *cs = CPU(cpu);
171
#if !defined(CONFIG_USER_ONLY)
219
- uint64_t pageaddr;
172
dc->user = (dc->current_el == 0);
220
-
173
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
221
- if (!arm_feature(env, ARM_FEATURE_EL2) || !(env->cp15.scr_el3 & SCR_NS)) {
174
dc->bt = FIELD_EX32(tb_flags, TBFLAG_A64, BT);
222
- return;
175
dc->btype = FIELD_EX32(tb_flags, TBFLAG_A64, BTYPE);
223
- }
176
dc->unpriv = FIELD_EX32(tb_flags, TBFLAG_A64, UNPRIV);
224
-
177
+ dc->ata = FIELD_EX32(tb_flags, TBFLAG_A64, ATA);
225
- pageaddr = sextract64(value << 12, 0, 48);
178
+ dc->mte_active[0] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE_ACTIVE);
226
-
179
+ dc->mte_active[1] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE0_ACTIVE);
227
- tlb_flush_page_by_mmuidx(cs, pageaddr, ARMMMUIdxBit_Stage2);
180
dc->vec_len = 0;
228
-}
181
dc->vec_stride = 0;
229
-
182
dc->cp_regs = arm_cpu->cp_regs;
230
-static void tlbi_aa64_ipas2e1is_write(CPUARMState *env, const ARMCPRegInfo *ri,
231
- uint64_t value)
232
-{
233
- CPUState *cs = env_cpu(env);
234
- uint64_t pageaddr;
235
-
236
- if (!arm_feature(env, ARM_FEATURE_EL2) || !(env->cp15.scr_el3 & SCR_NS)) {
237
- return;
238
- }
239
-
240
- pageaddr = sextract64(value << 12, 0, 48);
241
-
242
- tlb_flush_page_by_mmuidx_all_cpus_synced(cs, pageaddr,
243
- ARMMMUIdxBit_Stage2);
244
-}
245
-
246
static CPAccessResult aa64_zva_access(CPUARMState *env, const ARMCPRegInfo *ri,
247
bool isread)
248
{
249
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v8_cp_reginfo[] = {
250
.writefn = tlbi_aa64_vae1_write },
251
{ .name = "TLBI_IPAS2E1IS", .state = ARM_CP_STATE_AA64,
252
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 1,
253
- .access = PL2_W, .type = ARM_CP_NO_RAW,
254
- .writefn = tlbi_aa64_ipas2e1is_write },
255
+ .access = PL2_W, .type = ARM_CP_NOP },
256
{ .name = "TLBI_IPAS2LE1IS", .state = ARM_CP_STATE_AA64,
257
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 5,
258
- .access = PL2_W, .type = ARM_CP_NO_RAW,
259
- .writefn = tlbi_aa64_ipas2e1is_write },
260
+ .access = PL2_W, .type = ARM_CP_NOP },
261
{ .name = "TLBI_ALLE1IS", .state = ARM_CP_STATE_AA64,
262
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 3, .opc2 = 4,
263
.access = PL2_W, .type = ARM_CP_NO_RAW,
264
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v8_cp_reginfo[] = {
265
.writefn = tlbi_aa64_alle1is_write },
266
{ .name = "TLBI_IPAS2E1", .state = ARM_CP_STATE_AA64,
267
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 1,
268
- .access = PL2_W, .type = ARM_CP_NO_RAW,
269
- .writefn = tlbi_aa64_ipas2e1_write },
270
+ .access = PL2_W, .type = ARM_CP_NOP },
271
{ .name = "TLBI_IPAS2LE1", .state = ARM_CP_STATE_AA64,
272
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 5,
273
- .access = PL2_W, .type = ARM_CP_NO_RAW,
274
- .writefn = tlbi_aa64_ipas2e1_write },
275
+ .access = PL2_W, .type = ARM_CP_NOP },
276
{ .name = "TLBI_ALLE1", .state = ARM_CP_STATE_AA64,
277
.opc0 = 1, .opc1 = 4, .crn = 8, .crm = 7, .opc2 = 4,
278
.access = PL2_W, .type = ARM_CP_NO_RAW,
279
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo v8_cp_reginfo[] = {
280
.writefn = tlbimva_hyp_is_write },
281
{ .name = "TLBIIPAS2",
282
.cp = 15, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 1,
283
- .type = ARM_CP_NO_RAW, .access = PL2_W,
284
- .writefn = tlbiipas2_write },
285
+ .type = ARM_CP_NOP, .access = PL2_W },
286
{ .name = "TLBIIPAS2IS",
287
.cp = 15, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 1,
288
- .type = ARM_CP_NO_RAW, .access = PL2_W,
289
- .writefn = tlbiipas2_is_write },
290
+ .type = ARM_CP_NOP, .access = PL2_W },
291
{ .name = "TLBIIPAS2L",
292
.cp = 15, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 5,
293
- .type = ARM_CP_NO_RAW, .access = PL2_W,
294
- .writefn = tlbiipas2_write },
295
+ .type = ARM_CP_NOP, .access = PL2_W },
296
{ .name = "TLBIIPAS2LIS",
297
.cp = 15, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 5,
298
- .type = ARM_CP_NO_RAW, .access = PL2_W,
299
- .writefn = tlbiipas2_is_write },
300
+ .type = ARM_CP_NOP, .access = PL2_W },
301
/* 32 bit cache operations */
302
{ .name = "ICIALLUIS", .cp = 15, .opc1 = 0, .crn = 7, .crm = 1, .opc2 = 0,
303
.type = ARM_CP_NOP, .access = PL1_W, .accessfn = aa64_cacheop_pou_access },
304
--
183
--
305
2.20.1
184
2.20.1
306
185
307
186
diff view generated by jsdifflib
1
Add the infrastructure for building and invoking a decodetree decoder
1
From: Richard Henderson <richard.henderson@linaro.org>
2
for the AArch32 Neon encodings. At the moment the new decoder covers
3
nothing, so we always fall back to the existing hand-written decode.
4
2
5
We follow the same pattern we did for the VFP decodetree conversion
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
(commit 78e138bc1f672c145ef6ace74617d and following): code that deals
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
with Neon will be moving gradually out to translate-neon.vfp.inc,
5
Message-id: 20200626033144.790098-10-richard.henderson@linaro.org
8
which we #include into translate.c.
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-a64.h | 2 ++
9
target/arm/internals.h | 5 +++
10
target/arm/mte_helper.c | 72 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-a64.c | 18 ++++++++++
12
target/arm/Makefile.objs | 1 +
13
5 files changed, 98 insertions(+)
14
create mode 100644 target/arm/mte_helper.c
9
15
10
In order to share the decode files between A32 and T32, we
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
11
split Neon into 3 parts:
17
index XXXXXXX..XXXXXXX 100644
12
* data-processing
18
--- a/target/arm/helper-a64.h
13
* load-store
19
+++ b/target/arm/helper-a64.h
14
* 'shared' encodings
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autda, TCG_CALL_NO_WG, i64, env, i64, i64)
15
21
DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
16
The first two groups of instructions have similar but not identical
22
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
17
A32 and T32 encodings, so we need to manually transform the T32
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
18
encoding into the A32 one before calling the decoder; the third group
24
+
19
covers the Neon instructions which are identical in A32 and T32.
25
+DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
20
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
21
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
27
index XXXXXXX..XXXXXXX 100644
22
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
28
--- a/target/arm/internals.h
23
Message-id: 20200430181003.21682-4-peter.maydell@linaro.org
29
+++ b/target/arm/internals.h
24
---
30
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
25
target/arm/neon-dp.decode | 29 ++++++++++++++++++++++++++
31
*/
26
target/arm/neon-ls.decode | 29 ++++++++++++++++++++++++++
32
#define GMID_EL1_BS 6
27
target/arm/neon-shared.decode | 27 +++++++++++++++++++++++++
33
28
target/arm/translate-neon.inc.c | 32 +++++++++++++++++++++++++++++
34
+static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
29
target/arm/translate.c | 36 +++++++++++++++++++++++++++++++--
35
+{
30
target/arm/Makefile.objs | 18 +++++++++++++++++
36
+ return deposit64(ptr, 56, 4, rtag);
31
6 files changed, 169 insertions(+), 2 deletions(-)
37
+}
32
create mode 100644 target/arm/neon-dp.decode
38
+
33
create mode 100644 target/arm/neon-ls.decode
39
#endif
34
create mode 100644 target/arm/neon-shared.decode
40
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
35
create mode 100644 target/arm/translate-neon.inc.c
36
37
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
38
new file mode 100644
41
new file mode 100644
39
index XXXXXXX..XXXXXXX
42
index XXXXXXX..XXXXXXX
40
--- /dev/null
43
--- /dev/null
41
+++ b/target/arm/neon-dp.decode
44
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@
43
+# AArch32 Neon data-processing instruction descriptions
44
+#
45
+# Copyright (c) 2020 Linaro, Ltd
46
+#
47
+# This library is free software; you can redistribute it and/or
48
+# modify it under the terms of the GNU Lesser General Public
49
+# License as published by the Free Software Foundation; either
50
+# version 2 of the License, or (at your option) any later version.
51
+#
52
+# This library is distributed in the hope that it will be useful,
53
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
54
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
55
+# Lesser General Public License for more details.
56
+#
57
+# You should have received a copy of the GNU Lesser General Public
58
+# License along with this library; if not, see <http://www.gnu.org/licenses/>.
59
+
60
+#
61
+# This file is processed by scripts/decodetree.py
62
+#
63
+
64
+# Encodings for Neon data processing instructions where the T32 encoding
65
+# is a simple transformation of the A32 encoding.
66
+# More specifically, this file covers instructions where the A32 encoding is
67
+# 0b1111_001p_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
68
+# and the T32 encoding is
69
+# 0b111p_1111_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
70
+# This file works on the A32 encoding only; calling code for T32 has to
71
+# transform the insn into the A32 version first.
72
diff --git a/target/arm/neon-ls.decode b/target/arm/neon-ls.decode
73
new file mode 100644
74
index XXXXXXX..XXXXXXX
75
--- /dev/null
76
+++ b/target/arm/neon-ls.decode
77
@@ -XXX,XX +XXX,XX @@
78
+# AArch32 Neon load/store instruction descriptions
79
+#
80
+# Copyright (c) 2020 Linaro, Ltd
81
+#
82
+# This library is free software; you can redistribute it and/or
83
+# modify it under the terms of the GNU Lesser General Public
84
+# License as published by the Free Software Foundation; either
85
+# version 2 of the License, or (at your option) any later version.
86
+#
87
+# This library is distributed in the hope that it will be useful,
88
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
89
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
90
+# Lesser General Public License for more details.
91
+#
92
+# You should have received a copy of the GNU Lesser General Public
93
+# License along with this library; if not, see <http://www.gnu.org/licenses/>.
94
+
95
+#
96
+# This file is processed by scripts/decodetree.py
97
+#
98
+
99
+# Encodings for Neon load/store instructions where the T32 encoding
100
+# is a simple transformation of the A32 encoding.
101
+# More specifically, this file covers instructions where the A32 encoding is
102
+# 0b1111_0100_xxx0_xxxx_xxxx_xxxx_xxxx_xxxx
103
+# and the T32 encoding is
104
+# 0b1111_1001_xxx0_xxxx_xxxx_xxxx_xxxx_xxxx
105
+# This file works on the A32 encoding only; calling code for T32 has to
106
+# transform the insn into the A32 version first.
107
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
108
new file mode 100644
109
index XXXXXXX..XXXXXXX
110
--- /dev/null
111
+++ b/target/arm/neon-shared.decode
112
@@ -XXX,XX +XXX,XX @@
113
+# AArch32 Neon instruction descriptions
114
+#
115
+# Copyright (c) 2020 Linaro, Ltd
116
+#
117
+# This library is free software; you can redistribute it and/or
118
+# modify it under the terms of the GNU Lesser General Public
119
+# License as published by the Free Software Foundation; either
120
+# version 2 of the License, or (at your option) any later version.
121
+#
122
+# This library is distributed in the hope that it will be useful,
123
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
124
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
125
+# Lesser General Public License for more details.
126
+#
127
+# You should have received a copy of the GNU Lesser General Public
128
+# License along with this library; if not, see <http://www.gnu.org/licenses/>.
129
+
130
+#
131
+# This file is processed by scripts/decodetree.py
132
+#
133
+
134
+# Encodings for Neon instructions whose encoding is the same for
135
+# both A32 and T32.
136
+
137
+# More specifically, this covers:
138
+# 2reg scalar ext: 0b1111_1110_xxxx_xxxx_xxxx_1x0x_xxxx_xxxx
139
+# 3same ext: 0b1111_110x_xxxx_xxxx_xxxx_1x0x_xxxx_xxxx
140
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
141
new file mode 100644
142
index XXXXXXX..XXXXXXX
143
--- /dev/null
144
+++ b/target/arm/translate-neon.inc.c
145
@@ -XXX,XX +XXX,XX @@
45
@@ -XXX,XX +XXX,XX @@
146
+/*
46
+/*
147
+ * ARM translation: AArch32 Neon instructions
47
+ * ARM v8.5-MemTag Operations
148
+ *
48
+ *
149
+ * Copyright (c) 2003 Fabrice Bellard
49
+ * Copyright (c) 2020 Linaro, Ltd.
150
+ * Copyright (c) 2005-2007 CodeSourcery
151
+ * Copyright (c) 2007 OpenedHand, Ltd.
152
+ * Copyright (c) 2020 Linaro, Ltd.
153
+ *
50
+ *
154
+ * This library is free software; you can redistribute it and/or
51
+ * This library is free software; you can redistribute it and/or
155
+ * modify it under the terms of the GNU Lesser General Public
52
+ * modify it under the terms of the GNU Lesser General Public
156
+ * License as published by the Free Software Foundation; either
53
+ * License as published by the Free Software Foundation; either
157
+ * version 2 of the License, or (at your option) any later version.
54
+ * version 2.1 of the License, or (at your option) any later version.
158
+ *
55
+ *
159
+ * This library is distributed in the hope that it will be useful,
56
+ * This library is distributed in the hope that it will be useful,
160
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
57
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
161
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
58
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
162
+ * Lesser General Public License for more details.
59
+ * Lesser General Public License for more details.
163
+ *
60
+ *
164
+ * You should have received a copy of the GNU Lesser General Public
61
+ * You should have received a copy of the GNU Lesser General Public
165
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
62
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
166
+ */
63
+ */
167
+
64
+
168
+/*
65
+#include "qemu/osdep.h"
169
+ * This file is intended to be included from translate.c; it uses
66
+#include "cpu.h"
170
+ * some macros and definitions provided by that file.
67
+#include "internals.h"
171
+ * It might be possible to convert it to a standalone .c file eventually.
68
+#include "exec/exec-all.h"
172
+ */
69
+#include "exec/cpu_ldst.h"
70
+#include "exec/helper-proto.h"
173
+
71
+
174
+/* Include the generated Neon decoder */
72
+
175
+#include "decode-neon-dp.inc.c"
73
+static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
176
+#include "decode-neon-ls.inc.c"
74
+{
177
+#include "decode-neon-shared.inc.c"
75
+ if (exclude == 0xffff) {
178
diff --git a/target/arm/translate.c b/target/arm/translate.c
76
+ return 0;
77
+ }
78
+ if (offset == 0) {
79
+ while (exclude & (1 << tag)) {
80
+ tag = (tag + 1) & 15;
81
+ }
82
+ } else {
83
+ do {
84
+ do {
85
+ tag = (tag + 1) & 15;
86
+ } while (exclude & (1 << tag));
87
+ } while (--offset > 0);
88
+ }
89
+ return tag;
90
+}
91
+
92
+uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
93
+{
94
+ int rtag;
95
+
96
+ /*
97
+ * Our IMPDEF choice for GCR_EL1.RRND==1 is to behave as if
98
+ * GCR_EL1.RRND==0, always producing deterministic results.
99
+ */
100
+ uint16_t exclude = extract32(rm | env->cp15.gcr_el1, 0, 16);
101
+ int start = extract32(env->cp15.rgsr_el1, 0, 4);
102
+ int seed = extract32(env->cp15.rgsr_el1, 8, 16);
103
+ int offset, i;
104
+
105
+ /* RandomTag */
106
+ for (i = offset = 0; i < 4; ++i) {
107
+ /* NextRandomTagBit */
108
+ int top = (extract32(seed, 5, 1) ^ extract32(seed, 3, 1) ^
109
+ extract32(seed, 2, 1) ^ extract32(seed, 0, 1));
110
+ seed = (top << 15) | (seed >> 1);
111
+ offset |= top << i;
112
+ }
113
+ rtag = choose_nonexcluded_tag(start, offset, exclude);
114
+ env->cp15.rgsr_el1 = rtag | (seed << 8);
115
+
116
+ return address_with_allocation_tag(rn, rtag);
117
+}
118
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
179
index XXXXXXX..XXXXXXX 100644
119
index XXXXXXX..XXXXXXX 100644
180
--- a/target/arm/translate.c
120
--- a/target/arm/translate-a64.c
181
+++ b/target/arm/translate.c
121
+++ b/target/arm/translate-a64.c
182
@@ -XXX,XX +XXX,XX @@ static TCGv_ptr vfp_reg_ptr(bool dp, int reg)
122
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
183
123
return clean;
184
#define ARM_CP_RW_BIT (1 << 20)
124
}
185
125
186
-/* Include the VFP decoder */
126
+/* Insert a zero tag into src, with the result at dst. */
187
+/* Include the VFP and Neon decoders */
127
+static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
188
#include "translate-vfp.inc.c"
128
+{
189
+#include "translate-neon.inc.c"
129
+ tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
190
130
+}
191
static inline void iwmmxt_load_reg(TCGv_i64 var, int reg)
192
{
193
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
194
/* Unconditional instructions. */
195
/* TODO: Perhaps merge these into one decodetree output file. */
196
if (disas_a32_uncond(s, insn) ||
197
- disas_vfp_uncond(s, insn)) {
198
+ disas_vfp_uncond(s, insn) ||
199
+ disas_neon_dp(s, insn) ||
200
+ disas_neon_ls(s, insn) ||
201
+ disas_neon_shared(s, insn)) {
202
return;
203
}
204
/* fall back to legacy decoder */
205
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
206
ARCH(6T2);
207
}
208
209
+ if ((insn & 0xef000000) == 0xef000000) {
210
+ /*
211
+ * T32 encodings 0b111p_1111_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
212
+ * transform into
213
+ * A32 encodings 0b1111_001p_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
214
+ */
215
+ uint32_t a32_insn = (insn & 0xe2ffffff) |
216
+ ((insn & (1 << 28)) >> 4) | (1 << 28);
217
+
131
+
218
+ if (disas_neon_dp(s, a32_insn)) {
132
typedef struct DisasCompare64 {
219
+ return;
133
TCGCond cond;
134
TCGv_i64 value;
135
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
136
case 3: /* SDIV */
137
handle_div(s, true, sf, rm, rn, rd);
138
break;
139
+ case 4: /* IRG */
140
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
141
+ goto do_unallocated;
220
+ }
142
+ }
221
+ }
143
+ if (s->ata) {
222
+
144
+ gen_helper_irg(cpu_reg_sp(s, rd), cpu_env,
223
+ if ((insn & 0xff100000) == 0xf9000000) {
145
+ cpu_reg_sp(s, rn), cpu_reg(s, rm));
224
+ /*
146
+ } else {
225
+ * T32 encodings 0b1111_1001_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
147
+ gen_address_with_allocation_tag0(cpu_reg_sp(s, rd),
226
+ * transform into
148
+ cpu_reg_sp(s, rn));
227
+ * A32 encodings 0b1111_0100_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
228
+ */
229
+ uint32_t a32_insn = (insn & 0x00ffffff) | 0xf4000000;
230
+
231
+ if (disas_neon_ls(s, a32_insn)) {
232
+ return;
233
+ }
149
+ }
234
+ }
150
+ break;
235
+
151
case 8: /* LSLV */
236
/*
152
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
237
* TODO: Perhaps merge these into one decodetree output file.
153
break;
238
* Note disas_vfp is written for a32 with cond field in the
239
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
240
*/
241
if (disas_t32(s, insn) ||
242
disas_vfp_uncond(s, insn) ||
243
+ disas_neon_shared(s, insn) ||
244
((insn >> 28) == 0xe && disas_vfp(s, insn))) {
245
return;
246
}
247
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
154
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
248
index XXXXXXX..XXXXXXX 100644
155
index XXXXXXX..XXXXXXX 100644
249
--- a/target/arm/Makefile.objs
156
--- a/target/arm/Makefile.objs
250
+++ b/target/arm/Makefile.objs
157
+++ b/target/arm/Makefile.objs
251
@@ -XXX,XX +XXX,XX @@ target/arm/decode-sve.inc.c: $(SRC_PATH)/target/arm/sve.decode $(DECODETREE)
158
@@ -XXX,XX +XXX,XX @@ obj-$(CONFIG_SOFTMMU) += psci.o
252
     $(PYTHON) $(DECODETREE) --decode disas_sve -o $@ $<,\
159
obj-$(TARGET_AARCH64) += translate-a64.o helper-a64.o
253
     "GEN", $(TARGET_DIR)$@)
160
obj-$(TARGET_AARCH64) += translate-sve.o sve_helper.o
254
161
obj-$(TARGET_AARCH64) += pauth_helper.o
255
+target/arm/decode-neon-shared.inc.c: $(SRC_PATH)/target/arm/neon-shared.decode $(DECODETREE)
162
+obj-$(TARGET_AARCH64) += mte_helper.o
256
+    $(call quiet-command,\
257
+     $(PYTHON) $(DECODETREE) --static-decode disas_neon_shared -o $@ $<,\
258
+     "GEN", $(TARGET_DIR)$@)
259
+
260
+target/arm/decode-neon-dp.inc.c: $(SRC_PATH)/target/arm/neon-dp.decode $(DECODETREE)
261
+    $(call quiet-command,\
262
+     $(PYTHON) $(DECODETREE) --static-decode disas_neon_dp -o $@ $<,\
263
+     "GEN", $(TARGET_DIR)$@)
264
+
265
+target/arm/decode-neon-ls.inc.c: $(SRC_PATH)/target/arm/neon-ls.decode $(DECODETREE)
266
+    $(call quiet-command,\
267
+     $(PYTHON) $(DECODETREE) --static-decode disas_neon_ls -o $@ $<,\
268
+     "GEN", $(TARGET_DIR)$@)
269
+
270
target/arm/decode-vfp.inc.c: $(SRC_PATH)/target/arm/vfp.decode $(DECODETREE)
271
    $(call quiet-command,\
272
     $(PYTHON) $(DECODETREE) --static-decode disas_vfp -o $@ $<,\
273
@@ -XXX,XX +XXX,XX @@ target/arm/decode-t16.inc.c: $(SRC_PATH)/target/arm/t16.decode $(DECODETREE)
274
     "GEN", $(TARGET_DIR)$@)
275
276
target/arm/translate-sve.o: target/arm/decode-sve.inc.c
277
+target/arm/translate.o: target/arm/decode-neon-shared.inc.c
278
+target/arm/translate.o: target/arm/decode-neon-dp.inc.c
279
+target/arm/translate.o: target/arm/decode-neon-ls.inc.c
280
target/arm/translate.o: target/arm/decode-vfp.inc.c
281
target/arm/translate.o: target/arm/decode-vfp-uncond.inc.c
282
target/arm/translate.o: target/arm/decode-a32.inc.c
283
--
163
--
284
2.20.1
164
2.20.1
285
165
286
166
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
The current Arm ARM has adjusted the official decode of
4
"Add/subtract (immediate)" so that the shift field is only bit 22,
5
and bit 23 is part of the op1 field of the parent category
6
"Data processing - immediate".
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20200626033144.790098-11-richard.henderson@linaro.org
11
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/translate-a64.c | 23 ++++++++---------------
16
1 file changed, 8 insertions(+), 15 deletions(-)
17
18
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/translate-a64.c
21
+++ b/target/arm/translate-a64.c
22
@@ -XXX,XX +XXX,XX @@ static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
23
/*
24
* Add/subtract (immediate)
25
*
26
- * 31 30 29 28 24 23 22 21 10 9 5 4 0
27
- * +--+--+--+-----------+-----+-------------+-----+-----+
28
- * |sf|op| S| 1 0 0 0 1 |shift| imm12 | Rn | Rd |
29
- * +--+--+--+-----------+-----+-------------+-----+-----+
30
+ * 31 30 29 28 23 22 21 10 9 5 4 0
31
+ * +--+--+--+-------------+--+-------------+-----+-----+
32
+ * |sf|op| S| 1 0 0 0 1 0 |sh| imm12 | Rn | Rd |
33
+ * +--+--+--+-------------+--+-------------+-----+-----+
34
*
35
* sf: 0 -> 32bit, 1 -> 64bit
36
* op: 0 -> add , 1 -> sub
37
* S: 1 -> set flags
38
- * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
39
+ * sh: 1 -> LSL imm by 12
40
*/
41
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
42
{
43
int rd = extract32(insn, 0, 5);
44
int rn = extract32(insn, 5, 5);
45
uint64_t imm = extract32(insn, 10, 12);
46
- int shift = extract32(insn, 22, 2);
47
+ bool shift = extract32(insn, 22, 1);
48
bool setflags = extract32(insn, 29, 1);
49
bool sub_op = extract32(insn, 30, 1);
50
bool is_64bit = extract32(insn, 31, 1);
51
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
53
TCGv_i64 tcg_result;
54
55
- switch (shift) {
56
- case 0x0:
57
- break;
58
- case 0x1:
59
+ if (shift) {
60
imm <<= 12;
61
- break;
62
- default:
63
- unallocated_encoding(s);
64
- return;
65
}
66
67
tcg_result = tcg_temp_new_i64();
68
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
69
case 0x20: case 0x21: /* PC-rel. addressing */
70
disas_pc_rel_adr(s, insn);
71
break;
72
- case 0x22: case 0x23: /* Add/subtract (immediate) */
73
+ case 0x22: /* Add/subtract (immediate) */
74
disas_add_sub_imm(s, insn);
75
break;
76
case 0x24: /* Logical (immediate) */
77
--
78
2.20.1
79
80
diff view generated by jsdifflib
1
Convert the V[US]DOT (scalar) insns in the 2reg-scalar-ext group
1
From: Richard Henderson <richard.henderson@linaro.org>
2
to decodetree.
3
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-12-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-10-peter.maydell@linaro.org
7
---
7
---
8
target/arm/neon-shared.decode | 3 +++
8
target/arm/helper-a64.h | 1 +
9
target/arm/translate-neon.inc.c | 35 +++++++++++++++++++++++++++++++++
9
target/arm/internals.h | 9 +++++++
10
target/arm/translate.c | 13 +-----------
10
target/arm/mte_helper.c | 10 ++++++++
11
3 files changed, 39 insertions(+), 12 deletions(-)
11
target/arm/translate-a64.c | 51 ++++++++++++++++++++++++++++++++++++++
12
4 files changed, 71 insertions(+)
12
13
13
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
14
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/neon-shared.decode
16
--- a/target/arm/helper-a64.h
16
+++ b/target/arm/neon-shared.decode
17
+++ b/target/arm/helper-a64.h
17
@@ -XXX,XX +XXX,XX @@ VCMLA_scalar 1111 1110 0 . rot:2 .... .... 1000 . q:1 index:1 0 vm:4 \
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
18
vn=%vn_dp vd=%vd_dp size=0
19
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
19
VCMLA_scalar 1111 1110 1 . rot:2 .... .... 1000 . q:1 . 0 .... \
20
20
vm=%vm_dp vn=%vn_dp vd=%vd_dp size=1 index=0
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
diff --git a/target/arm/internals.h b/target/arm/internals.h
24
index XXXXXXX..XXXXXXX 100644
25
--- a/target/arm/internals.h
26
+++ b/target/arm/internals.h
27
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
28
*/
29
#define GMID_EL1_BS 6
30
31
+/* We associate one allocation tag per 16 bytes, the minimum. */
32
+#define LOG2_TAG_GRANULE 4
33
+#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
21
+
34
+
22
+VDOT_scalar 1111 1110 0 . 10 .... .... 1101 . q:1 index:1 u:1 rm:4 \
35
+static inline int allocation_tag_from_addr(uint64_t ptr)
23
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp
36
+{
24
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
37
+ return extract64(ptr, 56, 4);
38
+}
39
+
40
static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
41
{
42
return deposit64(ptr, 56, 4, rtag);
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
25
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/translate-neon.inc.c
45
--- a/target/arm/mte_helper.c
27
+++ b/target/arm/translate-neon.inc.c
46
+++ b/target/arm/mte_helper.c
28
@@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA_scalar(DisasContext *s, arg_VCMLA_scalar *a)
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
29
tcg_temp_free_ptr(fpst);
48
30
return true;
49
return address_with_allocation_tag(rn, rtag);
31
}
50
}
32
+
51
+
33
+static bool trans_VDOT_scalar(DisasContext *s, arg_VDOT_scalar *a)
52
+uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
53
+ int32_t offset, uint32_t tag_offset)
34
+{
54
+{
35
+ gen_helper_gvec_3 *fn_gvec;
55
+ int start_tag = allocation_tag_from_addr(ptr);
36
+ int opr_sz;
56
+ uint16_t exclude = extract32(env->cp15.gcr_el1, 0, 16);
37
+ TCGv_ptr fpst;
57
+ int rtag = choose_nonexcluded_tag(start_tag, tag_offset, exclude);
38
+
58
+
39
+ if (!dc_isar_feature(aa32_dp, s)) {
59
+ return address_with_allocation_tag(ptr + offset, rtag);
40
+ return false;
60
+}
61
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
62
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate-a64.c
64
+++ b/target/arm/translate-a64.c
65
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
66
tcg_temp_free_i64(tcg_result);
67
}
68
69
+/*
70
+ * Add/subtract (immediate, with tags)
71
+ *
72
+ * 31 30 29 28 23 22 21 16 14 10 9 5 4 0
73
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
74
+ * |sf|op| S| 1 0 0 0 1 1 |o2| uimm6 |o3| uimm4 | Rn | Rd |
75
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
76
+ *
77
+ * op: 0 -> add, 1 -> sub
78
+ */
79
+static void disas_add_sub_imm_with_tags(DisasContext *s, uint32_t insn)
80
+{
81
+ int rd = extract32(insn, 0, 5);
82
+ int rn = extract32(insn, 5, 5);
83
+ int uimm4 = extract32(insn, 10, 4);
84
+ int uimm6 = extract32(insn, 16, 6);
85
+ bool sub_op = extract32(insn, 30, 1);
86
+ TCGv_i64 tcg_rn, tcg_rd;
87
+ int imm;
88
+
89
+ /* Test all of sf=1, S=0, o2=0, o3=0. */
90
+ if ((insn & 0xa040c000u) != 0x80000000u ||
91
+ !dc_isar_feature(aa64_mte_insn_reg, s)) {
92
+ unallocated_encoding(s);
93
+ return;
41
+ }
94
+ }
42
+
95
+
43
+ /* UNDEF accesses to D16-D31 if they don't exist. */
96
+ imm = uimm6 << LOG2_TAG_GRANULE;
44
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
97
+ if (sub_op) {
45
+ ((a->vd | a->vn) & 0x10)) {
98
+ imm = -imm;
46
+ return false;
47
+ }
99
+ }
48
+
100
+
49
+ if ((a->vd | a->vn) & a->q) {
101
+ tcg_rn = cpu_reg_sp(s, rn);
50
+ return false;
102
+ tcg_rd = cpu_reg_sp(s, rd);
103
+
104
+ if (s->ata) {
105
+ TCGv_i32 offset = tcg_const_i32(imm);
106
+ TCGv_i32 tag_offset = tcg_const_i32(uimm4);
107
+
108
+ gen_helper_addsubg(tcg_rd, cpu_env, tcg_rn, offset, tag_offset);
109
+ tcg_temp_free_i32(tag_offset);
110
+ tcg_temp_free_i32(offset);
111
+ } else {
112
+ tcg_gen_addi_i64(tcg_rd, tcg_rn, imm);
113
+ gen_address_with_allocation_tag0(tcg_rd, tcg_rd);
51
+ }
114
+ }
115
+}
52
+
116
+
53
+ if (!vfp_access_check(s)) {
117
/* The input should be a value in the bottom e bits (with higher
54
+ return true;
118
* bits zero); returns that value replicated into every element
55
+ }
119
* of size e in a 64 bit integer.
56
+
120
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
57
+ fn_gvec = a->u ? gen_helper_gvec_udot_idx_b : gen_helper_gvec_sdot_idx_b;
121
case 0x22: /* Add/subtract (immediate) */
58
+ opr_sz = (1 + a->q) * 8;
122
disas_add_sub_imm(s, insn);
59
+ fpst = get_fpstatus_ptr(1);
123
break;
60
+ tcg_gen_gvec_3_ool(vfp_reg_offset(1, a->vd),
124
+ case 0x23: /* Add/subtract (immediate, with tags) */
61
+ vfp_reg_offset(1, a->vn),
125
+ disas_add_sub_imm_with_tags(s, insn);
62
+ vfp_reg_offset(1, a->rm),
126
+ break;
63
+ opr_sz, opr_sz, a->index, fn_gvec);
127
case 0x24: /* Logical (immediate) */
64
+ tcg_temp_free_ptr(fpst);
128
disas_logic_imm(s, insn);
65
+ return true;
129
break;
66
+}
67
diff --git a/target/arm/translate.c b/target/arm/translate.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/target/arm/translate.c
70
+++ b/target/arm/translate.c
71
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
72
bool is_long = false, q = extract32(insn, 6, 1);
73
bool ptr_is_env = false;
74
75
- if ((insn & 0xffb00f00) == 0xfe200d00) {
76
- /* V[US]DOT -- 1111 1110 0.10 .... .... 1101 .Q.U .... */
77
- int u = extract32(insn, 4, 1);
78
-
79
- if (!dc_isar_feature(aa32_dp, s)) {
80
- return 1;
81
- }
82
- fn_gvec = u ? gen_helper_gvec_udot_idx_b : gen_helper_gvec_sdot_idx_b;
83
- /* rm is just Vm, and index is M. */
84
- data = extract32(insn, 5, 1); /* index */
85
- rm = extract32(insn, 0, 4);
86
- } else if ((insn & 0xffa00f10) == 0xfe000810) {
87
+ if ((insn & 0xffa00f10) == 0xfe000810) {
88
/* VFM[AS]L -- 1111 1110 0.0S .... .... 1000 .Q.1 .... */
89
int is_s = extract32(insn, 20, 1);
90
int vm20 = extract32(insn, 0, 3);
91
--
130
--
92
2.20.1
131
2.20.1
93
132
94
133
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-13-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 15 +++++++++++++++
9
1 file changed, 15 insertions(+)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
16
cpu_reg_sp(s, rn));
17
}
18
break;
19
+ case 5: /* GMI */
20
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
21
+ goto do_unallocated;
22
+ } else {
23
+ TCGv_i64 t1 = tcg_const_i64(1);
24
+ TCGv_i64 t2 = tcg_temp_new_i64();
25
+
26
+ tcg_gen_extract_i64(t2, cpu_reg_sp(s, rn), 56, 4);
27
+ tcg_gen_shl_i64(t1, t1, t2);
28
+ tcg_gen_or_i64(cpu_reg(s, rd), cpu_reg(s, rm), t1);
29
+
30
+ tcg_temp_free_i64(t1);
31
+ tcg_temp_free_i64(t2);
32
+ }
33
+ break;
34
case 8: /* LSLV */
35
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
36
break;
37
--
38
2.20.1
39
40
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-14-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 24 ++++++++++++++++++++++--
9
1 file changed, 22 insertions(+), 2 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
16
*/
17
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
18
{
19
- unsigned int sf, rm, opcode, rn, rd;
20
+ unsigned int sf, rm, opcode, rn, rd, setflag;
21
sf = extract32(insn, 31, 1);
22
+ setflag = extract32(insn, 29, 1);
23
rm = extract32(insn, 16, 5);
24
opcode = extract32(insn, 10, 6);
25
rn = extract32(insn, 5, 5);
26
rd = extract32(insn, 0, 5);
27
28
- if (extract32(insn, 29, 1)) {
29
+ if (setflag && opcode != 0) {
30
unallocated_encoding(s);
31
return;
32
}
33
34
switch (opcode) {
35
+ case 0: /* SUBP(S) */
36
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
37
+ goto do_unallocated;
38
+ } else {
39
+ TCGv_i64 tcg_n, tcg_m, tcg_d;
40
+
41
+ tcg_n = read_cpu_reg_sp(s, rn, true);
42
+ tcg_m = read_cpu_reg_sp(s, rm, true);
43
+ tcg_gen_sextract_i64(tcg_n, tcg_n, 0, 56);
44
+ tcg_gen_sextract_i64(tcg_m, tcg_m, 0, 56);
45
+ tcg_d = cpu_reg(s, rd);
46
+
47
+ if (setflag) {
48
+ gen_sub_CC(true, tcg_d, tcg_n, tcg_m);
49
+ } else {
50
+ tcg_gen_sub_i64(tcg_d, tcg_n, tcg_m);
51
+ }
52
+ }
53
+ break;
54
case 2: /* UDIV */
55
handle_div(s, false, sf, rm, rn, rd);
56
break;
57
--
58
2.20.1
59
60
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Use the same code as system mode, so that we generate the same
4
exception + syndrome for the unaligned access.
5
6
For the moment, if MTE is enabled so that this path is reachable,
7
this would generate a SIGSEGV in the user-only cpu_loop. Decoding
8
the syndrome to produce the proper SIGBUS will be done later.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-15-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/cpu.c | 2 +-
16
target/arm/tlb_helper.c | 4 ++--
17
2 files changed, 3 insertions(+), 3 deletions(-)
18
19
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.c
22
+++ b/target/arm/cpu.c
23
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
24
cc->tlb_fill = arm_cpu_tlb_fill;
25
cc->debug_excp_handler = arm_debug_excp_handler;
26
cc->debug_check_watchpoint = arm_debug_check_watchpoint;
27
-#if !defined(CONFIG_USER_ONLY)
28
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
29
+#if !defined(CONFIG_USER_ONLY)
30
cc->do_transaction_failed = arm_cpu_do_transaction_failed;
31
cc->adjust_watchpoint_address = arm_adjust_watchpoint_address;
32
#endif /* CONFIG_TCG && !CONFIG_USER_ONLY */
33
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
34
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/tlb_helper.c
36
+++ b/target/arm/tlb_helper.c
37
@@ -XXX,XX +XXX,XX @@
38
#include "internals.h"
39
#include "exec/exec-all.h"
40
41
-#if !defined(CONFIG_USER_ONLY)
42
-
43
static inline uint32_t merge_syn_data_abort(uint32_t template_syn,
44
unsigned int target_el,
45
bool same_el, bool ea,
46
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
47
arm_deliver_fault(cpu, vaddr, access_type, mmu_idx, &fi);
48
}
49
50
+#if !defined(CONFIG_USER_ONLY)
51
+
52
/*
53
* arm_cpu_do_transaction_failed: handle a memory system error response
54
* (eg "no device/memory present at address") by raising an external abort
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Embed the UARTs into the SoC type.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-16-richard.henderson@linaro.org
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
10
Message-id: 20200427181649.26851-5-edgar.iglesias@gmail.com
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
7
---
13
include/hw/arm/xlnx-versal.h | 3 ++-
8
target/arm/helper-a64.h | 7 ++
14
hw/arm/xlnx-versal.c | 12 ++++++------
9
target/arm/helper.h | 2 +
15
2 files changed, 8 insertions(+), 7 deletions(-)
10
target/arm/mte_helper.c | 194 +++++++++++++++++++++++++++++++++++++
11
target/arm/op_helper.c | 16 +++
12
target/arm/translate-a64.c | 172 +++++++++++++++++++++++++++++++-
13
5 files changed, 386 insertions(+), 5 deletions(-)
16
14
17
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
15
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
18
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/arm/xlnx-versal.h
17
--- a/target/arm/helper-a64.h
20
+++ b/include/hw/arm/xlnx-versal.h
18
+++ b/target/arm/helper-a64.h
21
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
22
#include "hw/sysbus.h"
20
23
#include "hw/arm/boot.h"
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
24
#include "hw/intc/arm_gicv3.h"
22
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
25
+#include "hw/char/pl011.h"
23
+DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
26
24
+DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
27
#define TYPE_XLNX_VERSAL "xlnx-versal"
25
+DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
28
#define XLNX_VERSAL(obj) OBJECT_CHECK(Versal, (obj), TYPE_XLNX_VERSAL)
26
+DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
29
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
27
+DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
30
MemoryRegion mr_ocm;
28
+DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
31
29
+DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
32
struct {
30
diff --git a/target/arm/helper.h b/target/arm/helper.h
33
- SysBusDevice *uart[XLNX_VERSAL_NR_UARTS];
34
+ PL011State uart[XLNX_VERSAL_NR_UARTS];
35
SysBusDevice *gem[XLNX_VERSAL_NR_GEMS];
36
SysBusDevice *adma[XLNX_VERSAL_NR_ADMAS];
37
} iou;
38
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
39
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/xlnx-versal.c
32
--- a/target/arm/helper.h
41
+++ b/hw/arm/xlnx-versal.c
33
+++ b/target/arm/helper.h
42
@@ -XXX,XX +XXX,XX @@
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
43
#include "kvm_arm.h"
35
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
44
#include "hw/misc/unimp.h"
36
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
45
#include "hw/arm/xlnx-versal.h"
37
46
-#include "hw/char/pl011.h"
38
+DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
47
39
+
48
#define XLNX_VERSAL_ACPU_TYPE ARM_CPU_TYPE_NAME("cortex-a72")
40
DEF_HELPER_1(vfp_get_fpscr, i32, env)
49
#define GEM_REVISION 0x40070106
41
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
50
@@ -XXX,XX +XXX,XX @@ static void versal_create_uarts(Versal *s, qemu_irq *pic)
42
51
DeviceState *dev;
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
52
MemoryRegion *mr;
44
index XXXXXXX..XXXXXXX 100644
53
45
--- a/target/arm/mte_helper.c
54
- dev = qdev_create(NULL, TYPE_PL011);
46
+++ b/target/arm/mte_helper.c
55
- s->lpd.iou.uart[i] = SYS_BUS_DEVICE(dev);
47
@@ -XXX,XX +XXX,XX @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
56
+ sysbus_init_child_obj(OBJECT(s), name,
48
return tag;
57
+ &s->lpd.iou.uart[i], sizeof(s->lpd.iou.uart[i]),
49
}
58
+ TYPE_PL011);
50
59
+ dev = DEVICE(&s->lpd.iou.uart[i]);
51
+/**
60
qdev_prop_set_chr(dev, "chardev", serial_hd(i));
52
+ * allocation_tag_mem:
61
- object_property_add_child(OBJECT(s), name, OBJECT(dev), &error_fatal);
53
+ * @env: the cpu environment
62
qdev_init_nofail(dev);
54
+ * @ptr_mmu_idx: the addressing regime to use for the virtual address
63
55
+ * @ptr: the virtual address for which to look up tag memory
64
- mr = sysbus_mmio_get_region(s->lpd.iou.uart[i], 0);
56
+ * @ptr_access: the access to use for the virtual address
65
+ mr = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
57
+ * @ptr_size: the number of bytes in the normal memory access
66
memory_region_add_subregion(&s->mr_ps, addrs[i], mr);
58
+ * @tag_access: the access to use for the tag memory
67
59
+ * @tag_size: the number of bytes in the tag memory access
68
- sysbus_connect_irq(s->lpd.iou.uart[i], 0, pic[irqs[i]]);
60
+ * @ra: the return address for exception handling
69
+ sysbus_connect_irq(SYS_BUS_DEVICE(dev), 0, pic[irqs[i]]);
61
+ *
70
g_free(name);
62
+ * Our tag memory is formatted as a sequence of little-endian nibbles.
63
+ * That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
64
+ * tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
65
+ * for the higher addr.
66
+ *
67
+ * Here, resolve the physical address from the virtual address, and return
68
+ * a pointer to the corresponding tag byte. Exit with exception if the
69
+ * virtual address is not accessible for @ptr_access.
70
+ *
71
+ * The @ptr_size and @tag_size values may not have an obvious relation
72
+ * due to the alignment of @ptr, and the number of tag checks required.
73
+ *
74
+ * If there is no tag storage corresponding to @ptr, return NULL.
75
+ */
76
+static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
77
+ uint64_t ptr, MMUAccessType ptr_access,
78
+ int ptr_size, MMUAccessType tag_access,
79
+ int tag_size, uintptr_t ra)
80
+{
81
+ /* Tag storage not implemented. */
82
+ return NULL;
83
+}
84
+
85
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
86
{
87
int rtag;
88
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
89
90
return address_with_allocation_tag(ptr + offset, rtag);
91
}
92
+
93
+static int load_tag1(uint64_t ptr, uint8_t *mem)
94
+{
95
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
96
+ return extract32(*mem, ofs, 4);
97
+}
98
+
99
+uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
100
+{
101
+ int mmu_idx = cpu_mmu_index(env, false);
102
+ uint8_t *mem;
103
+ int rtag = 0;
104
+
105
+ /* Trap if accessing an invalid page. */
106
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
107
+ MMU_DATA_LOAD, 1, GETPC());
108
+
109
+ /* Load if page supports tags. */
110
+ if (mem) {
111
+ rtag = load_tag1(ptr, mem);
112
+ }
113
+
114
+ return address_with_allocation_tag(xt, rtag);
115
+}
116
+
117
+static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
118
+{
119
+ if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
120
+ arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
121
+ cpu_mmu_index(env, false), ra);
122
+ g_assert_not_reached();
123
+ }
124
+}
125
+
126
+/* For use in a non-parallel context, store to the given nibble. */
127
+static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
128
+{
129
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
130
+ *mem = deposit32(*mem, ofs, 4, tag);
131
+}
132
+
133
+/* For use in a parallel context, atomically store to the given nibble. */
134
+static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
135
+{
136
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
137
+ uint8_t old = atomic_read(mem);
138
+
139
+ while (1) {
140
+ uint8_t new = deposit32(old, ofs, 4, tag);
141
+ uint8_t cmp = atomic_cmpxchg(mem, old, new);
142
+ if (likely(cmp == old)) {
143
+ return;
144
+ }
145
+ old = cmp;
146
+ }
147
+}
148
+
149
+typedef void stg_store1(uint64_t, uint8_t *, int);
150
+
151
+static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
152
+ uintptr_t ra, stg_store1 store1)
153
+{
154
+ int mmu_idx = cpu_mmu_index(env, false);
155
+ uint8_t *mem;
156
+
157
+ check_tag_aligned(env, ptr, ra);
158
+
159
+ /* Trap if accessing an invalid page. */
160
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
161
+ MMU_DATA_STORE, 1, ra);
162
+
163
+ /* Store if page supports tags. */
164
+ if (mem) {
165
+ store1(ptr, mem, allocation_tag_from_addr(xt));
166
+ }
167
+}
168
+
169
+void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
170
+{
171
+ do_stg(env, ptr, xt, GETPC(), store_tag1);
172
+}
173
+
174
+void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
175
+{
176
+ do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
177
+}
178
+
179
+void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
180
+{
181
+ int mmu_idx = cpu_mmu_index(env, false);
182
+ uintptr_t ra = GETPC();
183
+
184
+ check_tag_aligned(env, ptr, ra);
185
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
186
+}
187
+
188
+static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
189
+ uintptr_t ra, stg_store1 store1)
190
+{
191
+ int mmu_idx = cpu_mmu_index(env, false);
192
+ int tag = allocation_tag_from_addr(xt);
193
+ uint8_t *mem1, *mem2;
194
+
195
+ check_tag_aligned(env, ptr, ra);
196
+
197
+ /*
198
+ * Trap if accessing an invalid page(s).
199
+ * This takes priority over !allocation_tag_access_enabled.
200
+ */
201
+ if (ptr & TAG_GRANULE) {
202
+ /* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
203
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
204
+ TAG_GRANULE, MMU_DATA_STORE, 1, ra);
205
+ mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
206
+ MMU_DATA_STORE, TAG_GRANULE,
207
+ MMU_DATA_STORE, 1, ra);
208
+
209
+ /* Store if page(s) support tags. */
210
+ if (mem1) {
211
+ store1(TAG_GRANULE, mem1, tag);
212
+ }
213
+ if (mem2) {
214
+ store1(0, mem2, tag);
215
+ }
216
+ } else {
217
+ /* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
218
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
219
+ 2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
220
+ if (mem1) {
221
+ tag |= tag << 4;
222
+ atomic_set(mem1, tag);
223
+ }
224
+ }
225
+}
226
+
227
+void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
228
+{
229
+ do_st2g(env, ptr, xt, GETPC(), store_tag1);
230
+}
231
+
232
+void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
233
+{
234
+ do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
235
+}
236
+
237
+void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
238
+{
239
+ int mmu_idx = cpu_mmu_index(env, false);
240
+ uintptr_t ra = GETPC();
241
+ int in_page = -(ptr | TARGET_PAGE_MASK);
242
+
243
+ check_tag_aligned(env, ptr, ra);
244
+
245
+ if (likely(in_page >= 2 * TAG_GRANULE)) {
246
+ probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
247
+ } else {
248
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
249
+ probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
250
+ }
251
+}
252
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
253
index XXXXXXX..XXXXXXX 100644
254
--- a/target/arm/op_helper.c
255
+++ b/target/arm/op_helper.c
256
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
257
return ((uint32_t)x >> shift) | (x << (32 - shift));
71
}
258
}
72
}
259
}
260
+
261
+void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
262
+ uint32_t access_type, uint32_t mmu_idx,
263
+ uint32_t size)
264
+{
265
+ uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
266
+ uintptr_t ra = GETPC();
267
+
268
+ if (likely(size <= in_page)) {
269
+ probe_access(env, ptr, size, access_type, mmu_idx, ra);
270
+ } else {
271
+ probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
272
+ probe_access(env, ptr + in_page, size - in_page,
273
+ access_type, mmu_idx, ra);
274
+ }
275
+}
276
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
277
index XXXXXXX..XXXXXXX 100644
278
--- a/target/arm/translate-a64.c
279
+++ b/target/arm/translate-a64.c
280
@@ -XXX,XX +XXX,XX @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
281
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
282
}
283
284
+static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
285
+ MMUAccessType acc, int log2_size)
286
+{
287
+ TCGv_i32 t_acc = tcg_const_i32(acc);
288
+ TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
289
+ TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
290
+
291
+ gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
292
+ tcg_temp_free_i32(t_acc);
293
+ tcg_temp_free_i32(t_idx);
294
+ tcg_temp_free_i32(t_size);
295
+}
296
+
297
typedef struct DisasCompare64 {
298
TCGCond cond;
299
TCGv_i64 value;
300
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
301
}
302
}
303
304
+/*
305
+ * Load/Store memory tags
306
+ *
307
+ * 31 30 29 24 22 21 12 10 5 0
308
+ * +-----+-------------+-----+---+------+-----+------+------+
309
+ * | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
310
+ * +-----+-------------+-----+---+------+-----+------+------+
311
+ */
312
+static void disas_ldst_tag(DisasContext *s, uint32_t insn)
313
+{
314
+ int rt = extract32(insn, 0, 5);
315
+ int rn = extract32(insn, 5, 5);
316
+ uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
317
+ int op2 = extract32(insn, 10, 2);
318
+ int op1 = extract32(insn, 22, 2);
319
+ bool is_load = false, is_pair = false, is_zero = false;
320
+ int index = 0;
321
+ TCGv_i64 addr, clean_addr, tcg_rt;
322
+
323
+ /* We checked insn bits [29:24,21] in the caller. */
324
+ if (extract32(insn, 30, 2) != 3) {
325
+ goto do_unallocated;
326
+ }
327
+
328
+ /*
329
+ * @index is a tri-state variable which has 3 states:
330
+ * < 0 : post-index, writeback
331
+ * = 0 : signed offset
332
+ * > 0 : pre-index, writeback
333
+ */
334
+ switch (op1) {
335
+ case 0:
336
+ if (op2 != 0) {
337
+ /* STG */
338
+ index = op2 - 2;
339
+ break;
340
+ }
341
+ goto do_unallocated;
342
+ case 1:
343
+ if (op2 != 0) {
344
+ /* STZG */
345
+ is_zero = true;
346
+ index = op2 - 2;
347
+ } else {
348
+ /* LDG */
349
+ is_load = true;
350
+ }
351
+ break;
352
+ case 2:
353
+ if (op2 != 0) {
354
+ /* ST2G */
355
+ is_pair = true;
356
+ index = op2 - 2;
357
+ break;
358
+ }
359
+ goto do_unallocated;
360
+ case 3:
361
+ if (op2 != 0) {
362
+ /* STZ2G */
363
+ is_pair = is_zero = true;
364
+ index = op2 - 2;
365
+ break;
366
+ }
367
+ goto do_unallocated;
368
+
369
+ default:
370
+ do_unallocated:
371
+ unallocated_encoding(s);
372
+ return;
373
+ }
374
+
375
+ if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
376
+ goto do_unallocated;
377
+ }
378
+
379
+ if (rn == 31) {
380
+ gen_check_sp_alignment(s);
381
+ }
382
+
383
+ addr = read_cpu_reg_sp(s, rn, true);
384
+ if (index >= 0) {
385
+ /* pre-index or signed offset */
386
+ tcg_gen_addi_i64(addr, addr, offset);
387
+ }
388
+
389
+ if (is_load) {
390
+ tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
391
+ tcg_rt = cpu_reg(s, rt);
392
+ if (s->ata) {
393
+ gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
394
+ } else {
395
+ clean_addr = clean_data_tbi(s, addr);
396
+ gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
397
+ gen_address_with_allocation_tag0(tcg_rt, addr);
398
+ }
399
+ } else {
400
+ tcg_rt = cpu_reg_sp(s, rt);
401
+ if (!s->ata) {
402
+ /*
403
+ * For STG and ST2G, we need to check alignment and probe memory.
404
+ * TODO: For STZG and STZ2G, we could rely on the stores below,
405
+ * at least for system mode; user-only won't enforce alignment.
406
+ */
407
+ if (is_pair) {
408
+ gen_helper_st2g_stub(cpu_env, addr);
409
+ } else {
410
+ gen_helper_stg_stub(cpu_env, addr);
411
+ }
412
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
413
+ if (is_pair) {
414
+ gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
415
+ } else {
416
+ gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
417
+ }
418
+ } else {
419
+ if (is_pair) {
420
+ gen_helper_st2g(cpu_env, addr, tcg_rt);
421
+ } else {
422
+ gen_helper_stg(cpu_env, addr, tcg_rt);
423
+ }
424
+ }
425
+ }
426
+
427
+ if (is_zero) {
428
+ TCGv_i64 clean_addr = clean_data_tbi(s, addr);
429
+ TCGv_i64 tcg_zero = tcg_const_i64(0);
430
+ int mem_index = get_mem_index(s);
431
+ int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
432
+
433
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
434
+ MO_Q | MO_ALIGN_16);
435
+ for (i = 8; i < n; i += 8) {
436
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
437
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
438
+ }
439
+ tcg_temp_free_i64(tcg_zero);
440
+ }
441
+
442
+ if (index != 0) {
443
+ /* pre-index or post-index */
444
+ if (index < 0) {
445
+ /* post-index */
446
+ tcg_gen_addi_i64(addr, addr, offset);
447
+ }
448
+ tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
449
+ }
450
+}
451
+
452
/* Loads and stores */
453
static void disas_ldst(DisasContext *s, uint32_t insn)
454
{
455
@@ -XXX,XX +XXX,XX @@ static void disas_ldst(DisasContext *s, uint32_t insn)
456
case 0x0d: /* AdvSIMD load/store single structure */
457
disas_ldst_single_struct(s, insn);
458
break;
459
- case 0x19: /* LDAPR/STLR (unscaled immediate) */
460
- if (extract32(insn, 10, 2) != 0 ||
461
- extract32(insn, 21, 1) != 0) {
462
+ case 0x19:
463
+ if (extract32(insn, 21, 1) != 0) {
464
+ disas_ldst_tag(s, insn);
465
+ } else if (extract32(insn, 10, 2) == 0) {
466
+ disas_ldst_ldapr_stlr(s, insn);
467
+ } else {
468
unallocated_encoding(s);
469
- break;
470
}
471
- disas_ldst_ldapr_stlr(s, insn);
472
break;
473
default:
474
unallocated_encoding(s);
73
--
475
--
74
2.20.1
476
2.20.1
75
477
76
478
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-17-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 29 ++++++++++++++++++++++++++---
9
1 file changed, 26 insertions(+), 3 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_ld_lit(DisasContext *s, uint32_t insn)
16
* +-----+-------+---+---+-------+---+-------+-------+------+------+
17
*
18
* opc: LDP/STP/LDNP/STNP 00 -> 32 bit, 10 -> 64 bit
19
- * LDPSW 01
20
+ * LDPSW/STGP 01
21
* LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
22
* V: 0 -> GPR, 1 -> Vector
23
* idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
24
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
25
bool is_signed = false;
26
bool postindex = false;
27
bool wback = false;
28
+ bool set_tag = false;
29
30
TCGv_i64 clean_addr, dirty_addr;
31
32
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
33
34
if (is_vector) {
35
size = 2 + opc;
36
+ } else if (opc == 1 && !is_load) {
37
+ /* STGP */
38
+ if (!dc_isar_feature(aa64_mte_insn_reg, s) || index == 0) {
39
+ unallocated_encoding(s);
40
+ return;
41
+ }
42
+ size = 3;
43
+ set_tag = true;
44
} else {
45
size = 2 + extract32(opc, 1, 1);
46
is_signed = extract32(opc, 0, 1);
47
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
48
return;
49
}
50
51
- offset <<= size;
52
+ offset <<= (set_tag ? LOG2_TAG_GRANULE : size);
53
54
if (rn == 31) {
55
gen_check_sp_alignment(s);
56
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
57
if (!postindex) {
58
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
59
}
60
- clean_addr = clean_data_tbi(s, dirty_addr);
61
62
+ if (set_tag) {
63
+ if (!s->ata) {
64
+ /*
65
+ * TODO: We could rely on the stores below, at least for
66
+ * system mode, if we arrange to add MO_ALIGN_16.
67
+ */
68
+ gen_helper_stg_stub(cpu_env, dirty_addr);
69
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
70
+ gen_helper_stg_parallel(cpu_env, dirty_addr, dirty_addr);
71
+ } else {
72
+ gen_helper_stg(cpu_env, dirty_addr, dirty_addr);
73
+ }
74
+ }
75
+
76
+ clean_addr = clean_data_tbi(s, dirty_addr);
77
if (is_vector) {
78
if (is_load) {
79
do_fp_ld(s, rt, clean_addr, size);
80
--
81
2.20.1
82
83
diff view generated by jsdifflib
1
Convert the Neon 3-reg-same VMAX and VMIN insns to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We can simplify our DC_ZVA if we recognize that the largest BS
4
that we actually use in system mode is 64. Let us just assert
5
that it fits within TARGET_PAGE_SIZE.
6
7
For DC_GVA and STZGM, we want to be able to write whole bytes
8
of tag memory, so assert that BS is >= 2 * TAG_GRANULE, or 32.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-18-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200430181003.21682-17-peter.maydell@linaro.org
6
---
14
---
7
target/arm/neon-dp.decode | 5 +++++
15
target/arm/cpu.c | 24 ++++++++++++++++++++++++
8
target/arm/translate-neon.inc.c | 14 ++++++++++++++
16
1 file changed, 24 insertions(+)
9
target/arm/translate.c | 21 ++-------------------
10
3 files changed, 21 insertions(+), 19 deletions(-)
11
17
12
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
18
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
13
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/neon-dp.decode
20
--- a/target/arm/cpu.c
15
+++ b/target/arm/neon-dp.decode
21
+++ b/target/arm/cpu.c
16
@@ -XXX,XX +XXX,XX @@ VBSL_3s 1111 001 1 0 . 01 .... .... 0001 ... 1 .... @3same_logic
22
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
17
VBIT_3s 1111 001 1 0 . 10 .... .... 0001 ... 1 .... @3same_logic
23
}
18
VBIF_3s 1111 001 1 0 . 11 .... .... 0001 ... 1 .... @3same_logic
24
#endif
19
25
20
+VMAX_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 0 .... @3same
26
+ if (tcg_enabled()) {
21
+VMAX_U_3s 1111 001 1 0 . .. .... .... 0110 . . . 0 .... @3same
27
+ int dcz_blocklen = 4 << cpu->dcz_blocksize;
22
+VMIN_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 1 .... @3same
23
+VMIN_U_3s 1111 001 1 0 . .. .... .... 0110 . . . 1 .... @3same
24
+
28
+
25
VADD_3s 1111 001 0 0 . .. .... .... 1000 . . . 0 .... @3same
29
+ /*
26
VSUB_3s 1111 001 1 0 . .. .... .... 1000 . . . 0 .... @3same
30
+ * We only support DCZ blocklen that fits on one page.
27
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
31
+ *
28
index XXXXXXX..XXXXXXX 100644
32
+ * Architectually this is always true. However TARGET_PAGE_SIZE
29
--- a/target/arm/translate-neon.inc.c
33
+ * is variable and, for compatibility with -machine virt-2.7,
30
+++ b/target/arm/translate-neon.inc.c
34
+ * is only 1KiB, as an artifact of legacy ARMv5 subpage support.
31
@@ -XXX,XX +XXX,XX @@ DO_3SAME(VEOR, tcg_gen_gvec_xor)
35
+ * But even then, while the largest architectural DCZ blocklen
32
DO_3SAME_BITSEL(VBSL, rd_ofs, rn_ofs, rm_ofs)
36
+ * is 2KiB, no cpu actually uses such a large blocklen.
33
DO_3SAME_BITSEL(VBIT, rm_ofs, rn_ofs, rd_ofs)
37
+ */
34
DO_3SAME_BITSEL(VBIF, rm_ofs, rd_ofs, rn_ofs)
38
+ assert(dcz_blocklen <= TARGET_PAGE_SIZE);
35
+
39
+
36
+#define DO_3SAME_NO_SZ_3(INSN, FUNC) \
40
+ /*
37
+ static bool trans_##INSN##_3s(DisasContext *s, arg_3same *a) \
41
+ * We only support DCZ blocksize >= 2*TAG_GRANULE, which is to say
38
+ { \
42
+ * both nibbles of each byte storing tag data may be written at once.
39
+ if (a->size == 3) { \
43
+ * Since TAG_GRANULE is 16, this means that blocklen must be >= 32.
40
+ return false; \
44
+ */
41
+ } \
45
+ if (cpu_isar_feature(aa64_mte, cpu)) {
42
+ return do_3same(s, a, FUNC); \
46
+ assert(dcz_blocklen >= 2 * TAG_GRANULE);
47
+ }
43
+ }
48
+ }
44
+
49
+
45
+DO_3SAME_NO_SZ_3(VMAX_S, tcg_gen_gvec_smax)
50
qemu_init_vcpu(cs);
46
+DO_3SAME_NO_SZ_3(VMAX_U, tcg_gen_gvec_umax)
51
cpu_reset(cs);
47
+DO_3SAME_NO_SZ_3(VMIN_S, tcg_gen_gvec_smin)
52
48
+DO_3SAME_NO_SZ_3(VMIN_U, tcg_gen_gvec_umin)
49
diff --git a/target/arm/translate.c b/target/arm/translate.c
50
index XXXXXXX..XXXXXXX 100644
51
--- a/target/arm/translate.c
52
+++ b/target/arm/translate.c
53
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
54
rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
55
return 0;
56
57
- case NEON_3R_VMAX:
58
- if (u) {
59
- tcg_gen_gvec_umax(size, rd_ofs, rn_ofs, rm_ofs,
60
- vec_size, vec_size);
61
- } else {
62
- tcg_gen_gvec_smax(size, rd_ofs, rn_ofs, rm_ofs,
63
- vec_size, vec_size);
64
- }
65
- return 0;
66
- case NEON_3R_VMIN:
67
- if (u) {
68
- tcg_gen_gvec_umin(size, rd_ofs, rn_ofs, rm_ofs,
69
- vec_size, vec_size);
70
- } else {
71
- tcg_gen_gvec_smin(size, rd_ofs, rn_ofs, rm_ofs,
72
- vec_size, vec_size);
73
- }
74
- return 0;
75
-
76
case NEON_3R_VSHL:
77
/* Note the operation is vshl vd,vm,vn */
78
tcg_gen_gvec_3(rd_ofs, rm_ofs, rn_ofs, vec_size, vec_size,
79
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
80
81
case NEON_3R_VADD_VSUB:
82
case NEON_3R_LOGIC:
83
+ case NEON_3R_VMAX:
84
+ case NEON_3R_VMIN:
85
/* Already handled by decodetree */
86
return 1;
87
}
88
--
53
--
89
2.20.1
54
2.20.1
90
55
91
56
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Embed the ADMAs into the SoC type.
3
Now that we know that the operation is on a single page,
4
we need not loop over pages while probing.
4
5
5
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Message-id: 20200626033144.790098-19-richard.henderson@linaro.org
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
10
Message-id: 20200427181649.26851-7-edgar.iglesias@gmail.com
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
10
---
13
include/hw/arm/xlnx-versal.h | 3 ++-
11
target/arm/helper-a64.c | 94 +++++++++++------------------------------
14
hw/arm/xlnx-versal.c | 14 +++++++-------
12
1 file changed, 25 insertions(+), 69 deletions(-)
15
2 files changed, 9 insertions(+), 8 deletions(-)
16
13
17
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
14
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
18
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/arm/xlnx-versal.h
16
--- a/target/arm/helper-a64.c
20
+++ b/include/hw/arm/xlnx-versal.h
17
+++ b/target/arm/helper-a64.c
21
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ void HELPER(dc_zva)(CPUARMState *env, uint64_t vaddr_in)
22
#include "hw/arm/boot.h"
19
* (which matches the usual QEMU behaviour of not implementing either
23
#include "hw/intc/arm_gicv3.h"
20
* alignment faults or any memory attribute handling).
24
#include "hw/char/pl011.h"
21
*/
25
+#include "hw/dma/xlnx-zdma.h"
22
-
26
#include "hw/net/cadence_gem.h"
23
- ARMCPU *cpu = env_archcpu(env);
27
24
- uint64_t blocklen = 4 << cpu->dcz_blocksize;
28
#define TYPE_XLNX_VERSAL "xlnx-versal"
25
+ int blocklen = 4 << env_archcpu(env)->dcz_blocksize;
29
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
26
uint64_t vaddr = vaddr_in & ~(blocklen - 1);
30
struct {
27
+ int mmu_idx = cpu_mmu_index(env, false);
31
PL011State uart[XLNX_VERSAL_NR_UARTS];
28
+ void *mem;
32
CadenceGEMState gem[XLNX_VERSAL_NR_GEMS];
29
+
33
- SysBusDevice *adma[XLNX_VERSAL_NR_ADMAS];
30
+ /*
34
+ XlnxZDMA adma[XLNX_VERSAL_NR_ADMAS];
31
+ * Trapless lookup. In addition to actual invalid page, may
35
} iou;
32
+ * return NULL for I/O, watchpoints, clean pages, etc.
36
} lpd;
33
+ */
37
34
+ mem = tlb_vaddr_to_host(env, vaddr, MMU_DATA_STORE, mmu_idx);
38
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
35
39
index XXXXXXX..XXXXXXX 100644
36
#ifndef CONFIG_USER_ONLY
40
--- a/hw/arm/xlnx-versal.c
37
- {
41
+++ b/hw/arm/xlnx-versal.c
38
+ if (unlikely(!mem)) {
42
@@ -XXX,XX +XXX,XX @@ static void versal_create_admas(Versal *s, qemu_irq *pic)
39
+ uintptr_t ra = GETPC();
43
DeviceState *dev;
40
+
44
MemoryRegion *mr;
41
/*
45
42
- * Slightly awkwardly, QEMU's TARGET_PAGE_SIZE may be less than
46
- dev = qdev_create(NULL, "xlnx.zdma");
43
- * the block size so we might have to do more than one TLB lookup.
47
- s->lpd.iou.adma[i] = SYS_BUS_DEVICE(dev);
44
- * We know that in fact for any v8 CPU the page size is at least 4K
48
- object_property_set_int(OBJECT(s->lpd.iou.adma[i]), 128, "bus-width",
45
- * and the block size must be 2K or less, but TARGET_PAGE_SIZE is only
49
- &error_abort);
46
- * 1K as an artefact of legacy v5 subpage support being present in the
50
- object_property_add_child(OBJECT(s), name, OBJECT(dev), &error_fatal);
47
- * same QEMU executable. So in practice the hostaddr[] array has
51
+ sysbus_init_child_obj(OBJECT(s), name,
48
- * two entries, given the current setting of TARGET_PAGE_BITS_MIN.
52
+ &s->lpd.iou.adma[i], sizeof(s->lpd.iou.adma[i]),
49
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
53
+ TYPE_XLNX_ZDMA);
50
+ * the original pointer for an invalid page. But watchpoints require
54
+ dev = DEVICE(&s->lpd.iou.adma[i]);
51
+ * that we probe the actual space. So do both.
55
+ object_property_set_int(OBJECT(dev), 128, "bus-width", &error_abort);
52
*/
56
qdev_init_nofail(dev);
53
- int maxidx = DIV_ROUND_UP(blocklen, TARGET_PAGE_SIZE);
57
54
- void *hostaddr[DIV_ROUND_UP(2 * KiB, 1 << TARGET_PAGE_BITS_MIN)];
58
- mr = sysbus_mmio_get_region(s->lpd.iou.adma[i], 0);
55
- int try, i;
59
+ mr = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
56
- unsigned mmu_idx = cpu_mmu_index(env, false);
60
memory_region_add_subregion(&s->mr_ps,
57
- TCGMemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
61
MM_ADMA_CH0 + i * MM_ADMA_CH0_SIZE, mr);
58
+ (void) probe_write(env, vaddr_in, 1, mmu_idx, ra);
62
59
+ mem = probe_write(env, vaddr, blocklen, mmu_idx, ra);
63
- sysbus_connect_irq(s->lpd.iou.adma[i], 0, pic[VERSAL_ADMA_IRQ_0 + i]);
60
64
+ sysbus_connect_irq(SYS_BUS_DEVICE(dev), 0, pic[VERSAL_ADMA_IRQ_0 + i]);
61
- assert(maxidx <= ARRAY_SIZE(hostaddr));
65
g_free(name);
62
-
63
- for (try = 0; try < 2; try++) {
64
-
65
- for (i = 0; i < maxidx; i++) {
66
- hostaddr[i] = tlb_vaddr_to_host(env,
67
- vaddr + TARGET_PAGE_SIZE * i,
68
- 1, mmu_idx);
69
- if (!hostaddr[i]) {
70
- break;
71
- }
72
- }
73
- if (i == maxidx) {
74
- /*
75
- * If it's all in the TLB it's fair game for just writing to;
76
- * we know we don't need to update dirty status, etc.
77
- */
78
- for (i = 0; i < maxidx - 1; i++) {
79
- memset(hostaddr[i], 0, TARGET_PAGE_SIZE);
80
- }
81
- memset(hostaddr[i], 0, blocklen - (i * TARGET_PAGE_SIZE));
82
- return;
83
- }
84
+ if (unlikely(!mem)) {
85
/*
86
- * OK, try a store and see if we can populate the tlb. This
87
- * might cause an exception if the memory isn't writable,
88
- * in which case we will longjmp out of here. We must for
89
- * this purpose use the actual register value passed to us
90
- * so that we get the fault address right.
91
+ * The only remaining reason for mem == NULL is I/O.
92
+ * Just do a series of byte writes as the architecture demands.
93
*/
94
- helper_ret_stb_mmu(env, vaddr_in, 0, oi, GETPC());
95
- /* Now we can populate the other TLB entries, if any */
96
- for (i = 0; i < maxidx; i++) {
97
- uint64_t va = vaddr + TARGET_PAGE_SIZE * i;
98
- if (va != (vaddr_in & TARGET_PAGE_MASK)) {
99
- helper_ret_stb_mmu(env, va, 0, oi, GETPC());
100
- }
101
+ for (int i = 0; i < blocklen; i++) {
102
+ cpu_stb_mmuidx_ra(env, vaddr + i, 0, mmu_idx, ra);
103
}
104
- }
105
-
106
- /*
107
- * Slow path (probably attempt to do this to an I/O device or
108
- * similar, or clearing of a block of code we have translations
109
- * cached for). Just do a series of byte writes as the architecture
110
- * demands. It's not worth trying to use a cpu_physical_memory_map(),
111
- * memset(), unmap() sequence here because:
112
- * + we'd need to account for the blocksize being larger than a page
113
- * + the direct-RAM access case is almost always going to be dealt
114
- * with in the fastpath code above, so there's no speed benefit
115
- * + we would have to deal with the map returning NULL because the
116
- * bounce buffer was in use
117
- */
118
- for (i = 0; i < blocklen; i++) {
119
- helper_ret_stb_mmu(env, vaddr + i, 0, oi, GETPC());
120
+ return;
121
}
66
}
122
}
123
-#else
124
- memset(g2h(vaddr), 0, blocklen);
125
#endif
126
+
127
+ memset(mem, 0, blocklen);
67
}
128
}
68
--
129
--
69
2.20.1
130
2.20.1
70
131
71
132
diff view generated by jsdifflib
1
We're going to want at least some of the NeonGen* typedefs
1
From: Richard Henderson <richard.henderson@linaro.org>
2
for the refactored 32-bit Neon decoder, so move them all
2
3
to translate.h since it makes more sense to keep them in
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
one group.
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
5
Message-id: 20200626033144.790098-20-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200430181003.21682-23-peter.maydell@linaro.org
9
---
7
---
10
target/arm/translate.h | 17 +++++++++++++++++
8
target/arm/helper-a64.h | 3 ++
11
target/arm/translate-a64.c | 17 -----------------
9
target/arm/translate.h | 2 +
12
2 files changed, 17 insertions(+), 17 deletions(-)
10
target/arm/mte_helper.c | 84 ++++++++++++++++++++++++++++++++++++++
13
11
target/arm/translate-a64.c | 72 ++++++++++++++++++++++++++++----
12
4 files changed, 153 insertions(+), 8 deletions(-)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
19
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
20
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
21
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
22
+DEF_HELPER_FLAGS_2(ldgm, TCG_CALL_NO_WG, i64, env, i64)
23
+DEF_HELPER_FLAGS_3(stgm, TCG_CALL_NO_WG, void, env, i64, i64)
24
+DEF_HELPER_FLAGS_3(stzgm_tags, TCG_CALL_NO_WG, void, env, i64, i64)
14
diff --git a/target/arm/translate.h b/target/arm/translate.h
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
15
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate.h
27
--- a/target/arm/translate.h
17
+++ b/target/arm/translate.h
28
+++ b/target/arm/translate.h
18
@@ -XXX,XX +XXX,XX @@ typedef void GVecGen3Fn(unsigned, uint32_t, uint32_t,
29
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
19
typedef void GVecGen4Fn(unsigned, uint32_t, uint32_t, uint32_t,
30
* < 0, set by the current instruction.
20
uint32_t, uint32_t, uint32_t);
31
*/
21
32
int8_t btype;
22
+/* Function prototype for gen_ functions for calling Neon helpers */
33
+ /* A copy of cpu->dcz_blocksize. */
23
+typedef void NeonGenOneOpEnvFn(TCGv_i32, TCGv_ptr, TCGv_i32);
34
+ uint8_t dcz_blocksize;
24
+typedef void NeonGenTwoOpFn(TCGv_i32, TCGv_i32, TCGv_i32);
35
/* True if this page is guarded. */
25
+typedef void NeonGenTwoOpEnvFn(TCGv_i32, TCGv_ptr, TCGv_i32, TCGv_i32);
36
bool guarded_page;
26
+typedef void NeonGenTwo64OpFn(TCGv_i64, TCGv_i64, TCGv_i64);
37
/* Bottom two bits of XScale c15_cpar coprocessor access control reg */
27
+typedef void NeonGenTwo64OpEnvFn(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i64);
38
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
28
+typedef void NeonGenNarrowFn(TCGv_i32, TCGv_i64);
39
index XXXXXXX..XXXXXXX 100644
29
+typedef void NeonGenNarrowEnvFn(TCGv_i32, TCGv_ptr, TCGv_i64);
40
--- a/target/arm/mte_helper.c
30
+typedef void NeonGenWidenFn(TCGv_i64, TCGv_i32);
41
+++ b/target/arm/mte_helper.c
31
+typedef void NeonGenTwoSingleOPFn(TCGv_i32, TCGv_i32, TCGv_i32, TCGv_ptr);
42
@@ -XXX,XX +XXX,XX @@ void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
32
+typedef void NeonGenTwoDoubleOPFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_ptr);
43
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
33
+typedef void NeonGenOneOpFn(TCGv_i64, TCGv_i64);
44
}
34
+typedef void CryptoTwoOpFn(TCGv_ptr, TCGv_ptr);
45
}
35
+typedef void CryptoThreeOpIntFn(TCGv_ptr, TCGv_ptr, TCGv_i32);
46
+
36
+typedef void CryptoThreeOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
47
+#define LDGM_STGM_SIZE (4 << GMID_EL1_BS)
37
+typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
48
+
38
+
49
+uint64_t HELPER(ldgm)(CPUARMState *env, uint64_t ptr)
39
#endif /* TARGET_ARM_TRANSLATE_H */
50
+{
51
+ int mmu_idx = cpu_mmu_index(env, false);
52
+ uintptr_t ra = GETPC();
53
+ void *tag_mem;
54
+
55
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
56
+
57
+ /* Trap if accessing an invalid page. */
58
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD,
59
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
60
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
61
+
62
+ /* The tag is squashed to zero if the page does not support tags. */
63
+ if (!tag_mem) {
64
+ return 0;
65
+ }
66
+
67
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
68
+ /*
69
+ * We are loading 64-bits worth of tags. The ordering of elements
70
+ * within the word corresponds to a 64-bit little-endian operation.
71
+ */
72
+ return ldq_le_p(tag_mem);
73
+}
74
+
75
+void HELPER(stgm)(CPUARMState *env, uint64_t ptr, uint64_t val)
76
+{
77
+ int mmu_idx = cpu_mmu_index(env, false);
78
+ uintptr_t ra = GETPC();
79
+ void *tag_mem;
80
+
81
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
82
+
83
+ /* Trap if accessing an invalid page. */
84
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
85
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
86
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
87
+
88
+ /*
89
+ * Tag store only happens if the page support tags,
90
+ * and if the OS has enabled access to the tags.
91
+ */
92
+ if (!tag_mem) {
93
+ return;
94
+ }
95
+
96
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
97
+ /*
98
+ * We are storing 64-bits worth of tags. The ordering of elements
99
+ * within the word corresponds to a 64-bit little-endian operation.
100
+ */
101
+ stq_le_p(tag_mem, val);
102
+}
103
+
104
+void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
105
+{
106
+ uintptr_t ra = GETPC();
107
+ int mmu_idx = cpu_mmu_index(env, false);
108
+ int log2_dcz_bytes, log2_tag_bytes;
109
+ intptr_t dcz_bytes, tag_bytes;
110
+ uint8_t *mem;
111
+
112
+ /*
113
+ * In arm_cpu_realizefn, we assert that dcz > LOG2_TAG_GRANULE+1,
114
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway,
115
+ * to make sure that we can access one complete tag byte here.
116
+ */
117
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
118
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
119
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
120
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
121
+ ptr &= -dcz_bytes;
122
+
123
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, dcz_bytes,
124
+ MMU_DATA_STORE, tag_bytes, ra);
125
+ if (mem) {
126
+ int tag_pair = (val & 0xf) * 0x11;
127
+ memset(mem, tag_pair, tag_bytes);
128
+ }
129
+}
40
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
130
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
41
index XXXXXXX..XXXXXXX 100644
131
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-a64.c
132
--- a/target/arm/translate-a64.c
43
+++ b/target/arm/translate-a64.c
133
+++ b/target/arm/translate-a64.c
44
@@ -XXX,XX +XXX,XX @@ typedef struct AArch64DecodeTable {
134
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
45
AArch64DecodeFn *disas_fn;
135
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
46
} AArch64DecodeTable;
136
int op2 = extract32(insn, 10, 2);
47
137
int op1 = extract32(insn, 22, 2);
48
-/* Function prototype for gen_ functions for calling Neon helpers */
138
- bool is_load = false, is_pair = false, is_zero = false;
49
-typedef void NeonGenOneOpEnvFn(TCGv_i32, TCGv_ptr, TCGv_i32);
139
+ bool is_load = false, is_pair = false, is_zero = false, is_mult = false;
50
-typedef void NeonGenTwoOpFn(TCGv_i32, TCGv_i32, TCGv_i32);
140
int index = 0;
51
-typedef void NeonGenTwoOpEnvFn(TCGv_i32, TCGv_ptr, TCGv_i32, TCGv_i32);
141
TCGv_i64 addr, clean_addr, tcg_rt;
52
-typedef void NeonGenTwo64OpFn(TCGv_i64, TCGv_i64, TCGv_i64);
142
53
-typedef void NeonGenTwo64OpEnvFn(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i64);
143
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
54
-typedef void NeonGenNarrowFn(TCGv_i32, TCGv_i64);
144
if (op2 != 0) {
55
-typedef void NeonGenNarrowEnvFn(TCGv_i32, TCGv_ptr, TCGv_i64);
145
/* STG */
56
-typedef void NeonGenWidenFn(TCGv_i64, TCGv_i32);
146
index = op2 - 2;
57
-typedef void NeonGenTwoSingleOPFn(TCGv_i32, TCGv_i32, TCGv_i32, TCGv_ptr);
147
- break;
58
-typedef void NeonGenTwoDoubleOPFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_ptr);
148
+ } else {
59
-typedef void NeonGenOneOpFn(TCGv_i64, TCGv_i64);
149
+ /* STZGM */
60
-typedef void CryptoTwoOpFn(TCGv_ptr, TCGv_ptr);
150
+ if (s->current_el == 0 || offset != 0) {
61
-typedef void CryptoThreeOpIntFn(TCGv_ptr, TCGv_ptr, TCGv_i32);
151
+ goto do_unallocated;
62
-typedef void CryptoThreeOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
152
+ }
63
-typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
153
+ is_mult = is_zero = true;
64
-
154
}
65
/* initialize TCG globals. */
155
- goto do_unallocated;
66
void a64_translate_init(void)
156
+ break;
67
{
157
case 1:
158
if (op2 != 0) {
159
/* STZG */
160
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
161
/* ST2G */
162
is_pair = true;
163
index = op2 - 2;
164
- break;
165
+ } else {
166
+ /* STGM */
167
+ if (s->current_el == 0 || offset != 0) {
168
+ goto do_unallocated;
169
+ }
170
+ is_mult = true;
171
}
172
- goto do_unallocated;
173
+ break;
174
case 3:
175
if (op2 != 0) {
176
/* STZ2G */
177
is_pair = is_zero = true;
178
index = op2 - 2;
179
- break;
180
+ } else {
181
+ /* LDGM */
182
+ if (s->current_el == 0 || offset != 0) {
183
+ goto do_unallocated;
184
+ }
185
+ is_mult = is_load = true;
186
}
187
- goto do_unallocated;
188
+ break;
189
190
default:
191
do_unallocated:
192
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
193
return;
194
}
195
196
- if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
197
+ if (is_mult
198
+ ? !dc_isar_feature(aa64_mte, s)
199
+ : !dc_isar_feature(aa64_mte_insn_reg, s)) {
200
goto do_unallocated;
201
}
202
203
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
204
tcg_gen_addi_i64(addr, addr, offset);
205
}
206
207
+ if (is_mult) {
208
+ tcg_rt = cpu_reg(s, rt);
209
+
210
+ if (is_zero) {
211
+ int size = 4 << s->dcz_blocksize;
212
+
213
+ if (s->ata) {
214
+ gen_helper_stzgm_tags(cpu_env, addr, tcg_rt);
215
+ }
216
+ /*
217
+ * The non-tags portion of STZGM is mostly like DC_ZVA,
218
+ * except the alignment happens before the access.
219
+ */
220
+ clean_addr = clean_data_tbi(s, addr);
221
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
222
+ gen_helper_dc_zva(cpu_env, clean_addr);
223
+ } else if (s->ata) {
224
+ if (is_load) {
225
+ gen_helper_ldgm(tcg_rt, cpu_env, addr);
226
+ } else {
227
+ gen_helper_stgm(cpu_env, addr, tcg_rt);
228
+ }
229
+ } else {
230
+ MMUAccessType acc = is_load ? MMU_DATA_LOAD : MMU_DATA_STORE;
231
+ int size = 4 << GMID_EL1_BS;
232
+
233
+ clean_addr = clean_data_tbi(s, addr);
234
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
235
+ gen_probe_access(s, clean_addr, acc, size);
236
+
237
+ if (is_load) {
238
+ /* The result tags are zeros. */
239
+ tcg_gen_movi_i64(tcg_rt, 0);
240
+ }
241
+ }
242
+ return;
243
+ }
244
+
245
if (is_load) {
246
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
247
tcg_rt = cpu_reg(s, rt);
248
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
249
dc->vec_stride = 0;
250
dc->cp_regs = arm_cpu->cp_regs;
251
dc->features = env->features;
252
+ dc->dcz_blocksize = arm_cpu->dcz_blocksize;
253
254
/* Single step state. The code-generation logic here is:
255
* SS_ACTIVE == 0:
68
--
256
--
69
2.20.1
257
2.20.1
70
258
71
259
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Remove inclusion of arm_gicv3_common.h, this already gets
3
Like the regular data cache flushes, these are nops within qemu.
4
included via xlnx-versal.h.
5
4
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
7
Message-id: 20200626033144.790098-21-richard.henderson@linaro.org
9
Message-id: 20200427181649.26851-2-edgar.iglesias@gmail.com
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
9
---
12
hw/arm/xlnx-versal.c | 1 -
10
target/arm/helper.c | 65 +++++++++++++++++++++++++++++++++++++++++++++
13
1 file changed, 1 deletion(-)
11
1 file changed, 65 insertions(+)
14
12
15
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/arm/xlnx-versal.c
15
--- a/target/arm/helper.c
18
+++ b/hw/arm/xlnx-versal.c
16
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@
17
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
20
#include "hw/arm/boot.h"
18
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
21
#include "kvm_arm.h"
19
.type = ARM_CP_NO_RAW,
22
#include "hw/misc/unimp.h"
20
.access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
23
-#include "hw/intc/arm_gicv3_common.h"
21
+ { .name = "DC_IGVAC", .state = ARM_CP_STATE_AA64,
24
#include "hw/arm/xlnx-versal.h"
22
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 3,
25
#include "hw/char/pl011.h"
23
+ .type = ARM_CP_NOP, .access = PL1_W,
24
+ .accessfn = aa64_cacheop_poc_access },
25
+ { .name = "DC_IGSW", .state = ARM_CP_STATE_AA64,
26
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 4,
27
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
28
+ { .name = "DC_IGDVAC", .state = ARM_CP_STATE_AA64,
29
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 5,
30
+ .type = ARM_CP_NOP, .access = PL1_W,
31
+ .accessfn = aa64_cacheop_poc_access },
32
+ { .name = "DC_IGDSW", .state = ARM_CP_STATE_AA64,
33
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 6,
34
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
35
+ { .name = "DC_CGSW", .state = ARM_CP_STATE_AA64,
36
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 4,
37
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
38
+ { .name = "DC_CGDSW", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 6,
40
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
41
+ { .name = "DC_CIGSW", .state = ARM_CP_STATE_AA64,
42
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 4,
43
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
44
+ { .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
45
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
46
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
47
REGINFO_SENTINEL
48
};
49
50
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
51
.type = ARM_CP_CONST, .access = PL0_RW, },
52
REGINFO_SENTINEL
53
};
54
+
55
+static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
56
+ { .name = "DC_CGVAC", .state = ARM_CP_STATE_AA64,
57
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 3,
58
+ .type = ARM_CP_NOP, .access = PL0_W,
59
+ .accessfn = aa64_cacheop_poc_access },
60
+ { .name = "DC_CGDVAC", .state = ARM_CP_STATE_AA64,
61
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 5,
62
+ .type = ARM_CP_NOP, .access = PL0_W,
63
+ .accessfn = aa64_cacheop_poc_access },
64
+ { .name = "DC_CGVAP", .state = ARM_CP_STATE_AA64,
65
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 3,
66
+ .type = ARM_CP_NOP, .access = PL0_W,
67
+ .accessfn = aa64_cacheop_poc_access },
68
+ { .name = "DC_CGDVAP", .state = ARM_CP_STATE_AA64,
69
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 5,
70
+ .type = ARM_CP_NOP, .access = PL0_W,
71
+ .accessfn = aa64_cacheop_poc_access },
72
+ { .name = "DC_CGVADP", .state = ARM_CP_STATE_AA64,
73
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 3,
74
+ .type = ARM_CP_NOP, .access = PL0_W,
75
+ .accessfn = aa64_cacheop_poc_access },
76
+ { .name = "DC_CGDVADP", .state = ARM_CP_STATE_AA64,
77
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 5,
78
+ .type = ARM_CP_NOP, .access = PL0_W,
79
+ .accessfn = aa64_cacheop_poc_access },
80
+ { .name = "DC_CIGVAC", .state = ARM_CP_STATE_AA64,
81
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 3,
82
+ .type = ARM_CP_NOP, .access = PL0_W,
83
+ .accessfn = aa64_cacheop_poc_access },
84
+ { .name = "DC_CIGDVAC", .state = ARM_CP_STATE_AA64,
85
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
86
+ .type = ARM_CP_NOP, .access = PL0_W,
87
+ .accessfn = aa64_cacheop_poc_access },
88
+ REGINFO_SENTINEL
89
+};
90
+
91
#endif
92
93
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
94
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
95
*/
96
if (cpu_isar_feature(aa64_mte, cpu)) {
97
define_arm_cp_regs(cpu, mte_reginfo);
98
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
99
} else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
100
define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
101
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
102
}
103
#endif
26
104
27
--
105
--
28
2.20.1
106
2.20.1
29
107
30
108
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Add support for SD.
3
We will shortly need this in mte_helper.c as well.
4
4
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Message-id: 20200626033144.790098-22-richard.henderson@linaro.org
8
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
9
Message-id: 20200427181649.26851-9-edgar.iglesias@gmail.com
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
9
---
12
include/hw/arm/xlnx-versal.h | 12 ++++++++++++
10
target/arm/internals.h | 36 ++++++++++++++++++++++++++++++++++++
13
hw/arm/xlnx-versal.c | 31 +++++++++++++++++++++++++++++++
11
target/arm/helper.c | 36 ------------------------------------
14
2 files changed, 43 insertions(+)
12
2 files changed, 36 insertions(+), 36 deletions(-)
15
13
16
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
17
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/xlnx-versal.h
16
--- a/target/arm/internals.h
19
+++ b/include/hw/arm/xlnx-versal.h
17
+++ b/target/arm/internals.h
20
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_pan(CPUARMState *env, ARMMMUIdx mmu_idx)
21
22
#include "hw/sysbus.h"
23
#include "hw/arm/boot.h"
24
+#include "hw/sd/sdhci.h"
25
#include "hw/intc/arm_gicv3.h"
26
#include "hw/char/pl011.h"
27
#include "hw/dma/xlnx-zdma.h"
28
@@ -XXX,XX +XXX,XX @@
29
#define XLNX_VERSAL_NR_UARTS 2
30
#define XLNX_VERSAL_NR_GEMS 2
31
#define XLNX_VERSAL_NR_ADMAS 8
32
+#define XLNX_VERSAL_NR_SDS 2
33
#define XLNX_VERSAL_NR_IRQS 192
34
35
typedef struct Versal {
36
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
37
} iou;
38
} lpd;
39
40
+ /* The Platform Management Controller subsystem. */
41
+ struct {
42
+ struct {
43
+ SDHCIState sd[XLNX_VERSAL_NR_SDS];
44
+ } iou;
45
+ } pmc;
46
+
47
struct {
48
MemoryRegion *mr_ddr;
49
uint32_t psci_conduit;
50
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
51
#define VERSAL_GEM1_IRQ_0 58
52
#define VERSAL_GEM1_WAKE_IRQ_0 59
53
#define VERSAL_ADMA_IRQ_0 60
54
+#define VERSAL_SD0_IRQ_0 126
55
56
/* Architecturally reserved IRQs suitable for virtualization. */
57
#define VERSAL_RSVD_IRQ_FIRST 111
58
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
59
#define MM_FPD_CRF 0xfd1a0000U
60
#define MM_FPD_CRF_SIZE 0x140000
61
62
+#define MM_PMC_SD0 0xf1040000U
63
+#define MM_PMC_SD0_SIZE 0x10000
64
#define MM_PMC_CRP 0xf1260000U
65
#define MM_PMC_CRP_SIZE 0x10000
66
#endif
67
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/xlnx-versal.c
70
+++ b/hw/arm/xlnx-versal.c
71
@@ -XXX,XX +XXX,XX @@ static void versal_create_admas(Versal *s, qemu_irq *pic)
72
}
19
}
73
}
20
}
74
21
75
+#define SDHCI_CAPABILITIES 0x280737ec6481 /* Same as on ZynqMP. */
22
+/* Return the exception level which controls this address translation regime */
76
+static void versal_create_sds(Versal *s, qemu_irq *pic)
23
+static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
77
+{
24
+{
78
+ int i;
25
+ switch (mmu_idx) {
79
+
26
+ case ARMMMUIdx_E20_0:
80
+ for (i = 0; i < ARRAY_SIZE(s->pmc.iou.sd); i++) {
27
+ case ARMMMUIdx_E20_2:
81
+ DeviceState *dev;
28
+ case ARMMMUIdx_E20_2_PAN:
82
+ MemoryRegion *mr;
29
+ case ARMMMUIdx_Stage2:
83
+
30
+ case ARMMMUIdx_E2:
84
+ sysbus_init_child_obj(OBJECT(s), "sd[*]",
31
+ return 2;
85
+ &s->pmc.iou.sd[i], sizeof(s->pmc.iou.sd[i]),
32
+ case ARMMMUIdx_SE3:
86
+ TYPE_SYSBUS_SDHCI);
33
+ return 3;
87
+ dev = DEVICE(&s->pmc.iou.sd[i]);
34
+ case ARMMMUIdx_SE10_0:
88
+
35
+ return arm_el_is_aa64(env, 3) ? 1 : 3;
89
+ object_property_set_uint(OBJECT(dev),
36
+ case ARMMMUIdx_SE10_1:
90
+ 3, "sd-spec-version", &error_fatal);
37
+ case ARMMMUIdx_SE10_1_PAN:
91
+ object_property_set_uint(OBJECT(dev), SDHCI_CAPABILITIES, "capareg",
38
+ case ARMMMUIdx_Stage1_E0:
92
+ &error_fatal);
39
+ case ARMMMUIdx_Stage1_E1:
93
+ object_property_set_uint(OBJECT(dev), UHS_I, "uhs", &error_fatal);
40
+ case ARMMMUIdx_Stage1_E1_PAN:
94
+ qdev_init_nofail(dev);
41
+ case ARMMMUIdx_E10_0:
95
+
42
+ case ARMMMUIdx_E10_1:
96
+ mr = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
43
+ case ARMMMUIdx_E10_1_PAN:
97
+ memory_region_add_subregion(&s->mr_ps,
44
+ case ARMMMUIdx_MPrivNegPri:
98
+ MM_PMC_SD0 + i * MM_PMC_SD0_SIZE, mr);
45
+ case ARMMMUIdx_MUserNegPri:
99
+
46
+ case ARMMMUIdx_MPriv:
100
+ sysbus_connect_irq(SYS_BUS_DEVICE(dev), 0,
47
+ case ARMMMUIdx_MUser:
101
+ pic[VERSAL_SD0_IRQ_0 + i * 2]);
48
+ case ARMMMUIdx_MSPrivNegPri:
49
+ case ARMMMUIdx_MSUserNegPri:
50
+ case ARMMMUIdx_MSPriv:
51
+ case ARMMMUIdx_MSUser:
52
+ return 1;
53
+ default:
54
+ g_assert_not_reached();
102
+ }
55
+ }
103
+}
56
+}
104
+
57
+
105
/* This takes the board allocated linear DDR memory and creates aliases
58
/* Return the FSR value for a debug exception (watchpoint, hardware
106
* for each split DDR range/aperture on the Versal address map.
59
* breakpoint or BKPT insn) targeting the specified exception level.
107
*/
60
*/
108
@@ -XXX,XX +XXX,XX @@ static void versal_realize(DeviceState *dev, Error **errp)
61
diff --git a/target/arm/helper.c b/target/arm/helper.c
109
versal_create_uarts(s, pic);
62
index XXXXXXX..XXXXXXX 100644
110
versal_create_gems(s, pic);
63
--- a/target/arm/helper.c
111
versal_create_admas(s, pic);
64
+++ b/target/arm/helper.c
112
+ versal_create_sds(s, pic);
65
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
113
versal_map_ddr(s);
66
}
114
versal_unimp(s);
67
#endif /* !CONFIG_USER_ONLY */
115
68
69
-/* Return the exception level which controls this address translation regime */
70
-static uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
71
-{
72
- switch (mmu_idx) {
73
- case ARMMMUIdx_E20_0:
74
- case ARMMMUIdx_E20_2:
75
- case ARMMMUIdx_E20_2_PAN:
76
- case ARMMMUIdx_Stage2:
77
- case ARMMMUIdx_E2:
78
- return 2;
79
- case ARMMMUIdx_SE3:
80
- return 3;
81
- case ARMMMUIdx_SE10_0:
82
- return arm_el_is_aa64(env, 3) ? 1 : 3;
83
- case ARMMMUIdx_SE10_1:
84
- case ARMMMUIdx_SE10_1_PAN:
85
- case ARMMMUIdx_Stage1_E0:
86
- case ARMMMUIdx_Stage1_E1:
87
- case ARMMMUIdx_Stage1_E1_PAN:
88
- case ARMMMUIdx_E10_0:
89
- case ARMMMUIdx_E10_1:
90
- case ARMMMUIdx_E10_1_PAN:
91
- case ARMMMUIdx_MPrivNegPri:
92
- case ARMMMUIdx_MUserNegPri:
93
- case ARMMMUIdx_MPriv:
94
- case ARMMMUIdx_MUser:
95
- case ARMMMUIdx_MSPrivNegPri:
96
- case ARMMMUIdx_MSUserNegPri:
97
- case ARMMMUIdx_MSPriv:
98
- case ARMMMUIdx_MSUser:
99
- return 1;
100
- default:
101
- g_assert_not_reached();
102
- }
103
-}
104
-
105
uint64_t arm_sctlr(CPUARMState *env, int el)
106
{
107
/* Only EL0 needs to be adjusted for EL1&0 or EL2&0. */
116
--
108
--
117
2.20.1
109
2.20.1
118
110
119
111
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Add support for the RTC.
3
We will shortly need this in mte_helper.c as well.
4
4
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
7
Message-id: 20200626033144.790098-23-richard.henderson@linaro.org
8
Message-id: 20200427181649.26851-12-edgar.iglesias@gmail.com
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
9
---
11
hw/arm/xlnx-versal-virt.c | 22 ++++++++++++++++++++++
10
target/arm/internals.h | 9 +++++++++
12
1 file changed, 22 insertions(+)
11
target/arm/helper.c | 9 ---------
12
2 files changed, 9 insertions(+), 9 deletions(-)
13
13
14
diff --git a/hw/arm/xlnx-versal-virt.c b/hw/arm/xlnx-versal-virt.c
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/hw/arm/xlnx-versal-virt.c
16
--- a/target/arm/internals.h
17
+++ b/hw/arm/xlnx-versal-virt.c
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ static void fdt_add_sd_nodes(VersalVirt *s)
18
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
19
}
19
}
20
}
20
}
21
21
22
+static void fdt_add_rtc_node(VersalVirt *s)
22
+/* Return the TCR controlling this translation regime */
23
+static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
23
+{
24
+{
24
+ const char compat[] = "xlnx,zynqmp-rtc";
25
+ if (mmu_idx == ARMMMUIdx_Stage2) {
25
+ const char interrupt_names[] = "alarm\0sec";
26
+ return &env->cp15.vtcr_el2;
26
+ char *name = g_strdup_printf("/rtc@%x", MM_PMC_RTC);
27
+ }
27
+
28
+ return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
28
+ qemu_fdt_add_subnode(s->fdt, name);
29
+
30
+ qemu_fdt_setprop_cells(s->fdt, name, "interrupts",
31
+ GIC_FDT_IRQ_TYPE_SPI, VERSAL_RTC_ALARM_IRQ,
32
+ GIC_FDT_IRQ_FLAGS_LEVEL_HI,
33
+ GIC_FDT_IRQ_TYPE_SPI, VERSAL_RTC_SECONDS_IRQ,
34
+ GIC_FDT_IRQ_FLAGS_LEVEL_HI);
35
+ qemu_fdt_setprop(s->fdt, name, "interrupt-names",
36
+ interrupt_names, sizeof(interrupt_names));
37
+ qemu_fdt_setprop_sized_cells(s->fdt, name, "reg",
38
+ 2, MM_PMC_RTC, 2, MM_PMC_RTC_SIZE);
39
+ qemu_fdt_setprop(s->fdt, name, "compatible", compat, sizeof(compat));
40
+ g_free(name);
41
+}
29
+}
42
+
30
+
43
static void fdt_nop_memory_nodes(void *fdt, Error **errp)
31
/* Return the FSR value for a debug exception (watchpoint, hardware
44
{
32
* breakpoint or BKPT insn) targeting the specified exception level.
45
Error *err = NULL;
33
*/
46
@@ -XXX,XX +XXX,XX @@ static void versal_virt_init(MachineState *machine)
34
diff --git a/target/arm/helper.c b/target/arm/helper.c
47
fdt_add_timer_nodes(s);
35
index XXXXXXX..XXXXXXX 100644
48
fdt_add_zdma_nodes(s);
36
--- a/target/arm/helper.c
49
fdt_add_sd_nodes(s);
37
+++ b/target/arm/helper.c
50
+ fdt_add_rtc_node(s);
38
@@ -XXX,XX +XXX,XX @@ static inline uint64_t regime_ttbr(CPUARMState *env, ARMMMUIdx mmu_idx,
51
fdt_add_cpu_nodes(s, psci_conduit);
39
52
fdt_add_clk_node(s, "/clk125", 125000000, s->phandle.clk_125Mhz);
40
#endif /* !CONFIG_USER_ONLY */
53
fdt_add_clk_node(s, "/clk25", 25000000, s->phandle.clk_25Mhz);
41
42
-/* Return the TCR controlling this translation regime */
43
-static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
44
-{
45
- if (mmu_idx == ARMMMUIdx_Stage2) {
46
- return &env->cp15.vtcr_el2;
47
- }
48
- return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
49
-}
50
-
51
/* Convert a possible stage1+2 MMU index into the appropriate
52
* stage 1 MMU index
53
*/
54
--
54
--
55
2.20.1
55
2.20.1
56
56
57
57
diff view generated by jsdifflib
1
Convert the Neon 3-reg-same VADD and VSUB insns to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Note that we don't need the neon_3r_sizes[op] check here because all
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
size values are OK for VADD and VSUB; we'll add this when we convert
4
perform a single logical memory access. Leave the helper blank
5
the first insn that has size restrictions.
5
for now to reduce the patch size.
6
6
7
For this we need one of the GVecGen*Fn typedefs currently in
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
translate-a64.h; move them all to translate.h as a block so they
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
are visible to the 32-bit decoder.
9
Message-id: 20200626033144.790098-24-richard.henderson@linaro.org
10
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200430181003.21682-15-peter.maydell@linaro.org
14
---
11
---
15
target/arm/translate-a64.h | 9 --------
12
target/arm/helper-a64.h | 1 +
16
target/arm/translate.h | 9 ++++++++
13
target/arm/internals.h | 8 +++
17
target/arm/neon-dp.decode | 17 +++++++++++++++
14
target/arm/translate-a64.h | 2 +
18
target/arm/translate-neon.inc.c | 38 +++++++++++++++++++++++++++++++++
15
target/arm/mte_helper.c | 8 +++
19
target/arm/translate.c | 14 ++++--------
16
target/arm/translate-a64.c | 100 ++++++++++++++++++++++++++++---------
20
5 files changed, 68 insertions(+), 19 deletions(-)
17
5 files changed, 95 insertions(+), 24 deletions(-)
21
18
19
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper-a64.h
22
+++ b/target/arm/helper-a64.h
23
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
24
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
25
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
27
+DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
28
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
29
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
30
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
31
diff --git a/target/arm/internals.h b/target/arm/internals.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/internals.h
34
+++ b/target/arm/internals.h
35
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
36
#define LOG2_TAG_GRANULE 4
37
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
38
39
+/* Bits within a descriptor passed to the helper_mte_check* functions. */
40
+FIELD(MTEDESC, MIDX, 0, 4)
41
+FIELD(MTEDESC, TBI, 4, 2)
42
+FIELD(MTEDESC, TCMA, 6, 2)
43
+FIELD(MTEDESC, WRITE, 8, 1)
44
+FIELD(MTEDESC, ESIZE, 9, 5)
45
+FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
46
+
47
static inline int allocation_tag_from_addr(uint64_t ptr)
48
{
49
return extract64(ptr, 56, 4);
22
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
50
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
23
index XXXXXXX..XXXXXXX 100644
51
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/translate-a64.h
52
--- a/target/arm/translate-a64.h
25
+++ b/target/arm/translate-a64.h
53
+++ b/target/arm/translate-a64.h
26
@@ -XXX,XX +XXX,XX @@ static inline int vec_full_reg_size(DisasContext *s)
54
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
27
55
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
28
bool disas_sve(DisasContext *, uint32_t);
56
unsigned int imms, unsigned int immr);
29
57
bool sve_access_check(DisasContext *s);
30
-/* Note that the gvec expanders operate on offsets + sizes. */
58
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
31
-typedef void GVecGen2Fn(unsigned, uint32_t, uint32_t, uint32_t, uint32_t);
59
+ bool tag_checked, int log2_size);
32
-typedef void GVecGen2iFn(unsigned, uint32_t, uint32_t, int64_t,
60
33
- uint32_t, uint32_t);
61
/* We should have at some point before trying to access an FP register
34
-typedef void GVecGen3Fn(unsigned, uint32_t, uint32_t,
62
* done the necessary access check, so assert that
35
- uint32_t, uint32_t, uint32_t);
63
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
36
-typedef void GVecGen4Fn(unsigned, uint32_t, uint32_t, uint32_t,
64
index XXXXXXX..XXXXXXX 100644
37
- uint32_t, uint32_t, uint32_t);
65
--- a/target/arm/mte_helper.c
38
-
66
+++ b/target/arm/mte_helper.c
39
#endif /* TARGET_ARM_TRANSLATE_A64_H */
67
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
40
diff --git a/target/arm/translate.h b/target/arm/translate.h
68
memset(mem, tag_pair, tag_bytes);
41
index XXXXXXX..XXXXXXX 100644
69
}
42
--- a/target/arm/translate.h
43
+++ b/target/arm/translate.h
44
@@ -XXX,XX +XXX,XX @@ void gen_sshl_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b);
45
#define dc_isar_feature(name, ctx) \
46
({ DisasContext *ctx_ = (ctx); isar_feature_##name(ctx_->isar); })
47
48
+/* Note that the gvec expanders operate on offsets + sizes. */
49
+typedef void GVecGen2Fn(unsigned, uint32_t, uint32_t, uint32_t, uint32_t);
50
+typedef void GVecGen2iFn(unsigned, uint32_t, uint32_t, int64_t,
51
+ uint32_t, uint32_t);
52
+typedef void GVecGen3Fn(unsigned, uint32_t, uint32_t,
53
+ uint32_t, uint32_t, uint32_t);
54
+typedef void GVecGen4Fn(unsigned, uint32_t, uint32_t, uint32_t,
55
+ uint32_t, uint32_t, uint32_t);
56
+
57
#endif /* TARGET_ARM_TRANSLATE_H */
58
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
59
index XXXXXXX..XXXXXXX 100644
60
--- a/target/arm/neon-dp.decode
61
+++ b/target/arm/neon-dp.decode
62
@@ -XXX,XX +XXX,XX @@
63
#
64
# This file is processed by scripts/decodetree.py
65
#
66
+# VFP/Neon register fields; same as vfp.decode
67
+%vm_dp 5:1 0:4
68
+%vn_dp 7:1 16:4
69
+%vd_dp 22:1 12:4
70
71
# Encodings for Neon data processing instructions where the T32 encoding
72
# is a simple transformation of the A32 encoding.
73
@@ -XXX,XX +XXX,XX @@
74
# 0b111p_1111_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
75
# This file works on the A32 encoding only; calling code for T32 has to
76
# transform the insn into the A32 version first.
77
+
78
+######################################################################
79
+# 3-reg-same grouping:
80
+# 1111 001 U 0 D sz:2 Vn:4 Vd:4 opc:4 N Q M op Vm:4
81
+######################################################################
82
+
83
+&3same vm vn vd q size
84
+
85
+@3same .... ... . . . size:2 .... .... .... . q:1 . . .... \
86
+ &3same vm=%vm_dp vn=%vn_dp vd=%vd_dp
87
+
88
+VADD_3s 1111 001 0 0 . .. .... .... 1000 . . . 0 .... @3same
89
+VSUB_3s 1111 001 1 0 . .. .... .... 1000 . . . 0 .... @3same
90
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
91
index XXXXXXX..XXXXXXX 100644
92
--- a/target/arm/translate-neon.inc.c
93
+++ b/target/arm/translate-neon.inc.c
94
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDST_single(DisasContext *s, arg_VLDST_single *a)
95
96
return true;
97
}
70
}
98
+
71
+
99
+static bool do_3same(DisasContext *s, arg_3same *a, GVecGen3Fn fn)
72
+/*
73
+ * Perform an MTE checked access for a single logical or atomic access.
74
+ */
75
+uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
100
+{
76
+{
101
+ int vec_size = a->q ? 16 : 8;
77
+ return ptr;
102
+ int rd_ofs = neon_reg_offset(a->vd, 0);
78
+}
103
+ int rn_ofs = neon_reg_offset(a->vn, 0);
79
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
104
+ int rm_ofs = neon_reg_offset(a->vm, 0);
80
index XXXXXXX..XXXXXXX 100644
105
+
81
--- a/target/arm/translate-a64.c
106
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
82
+++ b/target/arm/translate-a64.c
107
+ return false;
83
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
84
}
85
86
/*
87
- * Return a "clean" address for ADDR according to TBID.
88
- * This is always a fresh temporary, as we need to be able to
89
- * increment this independently of a dirty write-back address.
90
+ * Handle MTE and/or TBI.
91
+ *
92
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
93
+ * for the tag to be present in the FAR_ELx register. But for user-only
94
+ * mode we do not have a TLB with which to implement this, so we must
95
+ * remove the top byte now.
96
+ *
97
+ * Always return a fresh temporary that we can increment independently
98
+ * of the write-back address.
99
*/
100
+
101
static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
102
{
103
TCGv_i64 clean = new_tmp_a64(s);
104
- /*
105
- * In order to get the correct value in the FAR_ELx register,
106
- * we must present the memory subsystem with the "dirty" address
107
- * including the TBI. In system mode we can make this work via
108
- * the TLB, dropping the TBI during translation. But for user-only
109
- * mode we don't have that option, and must remove the top byte now.
110
- */
111
#ifdef CONFIG_USER_ONLY
112
gen_top_byte_ignore(s, clean, addr, s->tbid);
113
#else
114
@@ -XXX,XX +XXX,XX @@ static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
115
tcg_temp_free_i32(t_size);
116
}
117
118
+/*
119
+ * For MTE, check a single logical or atomic access. This probes a single
120
+ * address, the exact one specified. The size and alignment of the access
121
+ * is not relevant to MTE, per se, but watchpoints do require the size,
122
+ * and we want to recognize those before making any other changes to state.
123
+ */
124
+static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
125
+ bool is_write, bool tag_checked,
126
+ int log2_size, bool is_unpriv,
127
+ int core_idx)
128
+{
129
+ if (tag_checked && s->mte_active[is_unpriv]) {
130
+ TCGv_i32 tcg_desc;
131
+ TCGv_i64 ret;
132
+ int desc = 0;
133
+
134
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, core_idx);
135
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
136
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
137
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
138
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_size);
139
+ tcg_desc = tcg_const_i32(desc);
140
+
141
+ ret = new_tmp_a64(s);
142
+ gen_helper_mte_check1(ret, cpu_env, tcg_desc, addr);
143
+ tcg_temp_free_i32(tcg_desc);
144
+
145
+ return ret;
108
+ }
146
+ }
109
+
147
+ return clean_data_tbi(s, addr);
110
+ /* UNDEF accesses to D16-D31 if they don't exist. */
111
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
112
+ ((a->vd | a->vn | a->vm) & 0x10)) {
113
+ return false;
114
+ }
115
+
116
+ if ((a->vn | a->vm | a->vd) & a->q) {
117
+ return false;
118
+ }
119
+
120
+ if (!vfp_access_check(s)) {
121
+ return true;
122
+ }
123
+
124
+ fn(a->size, rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
125
+ return true;
126
+}
148
+}
127
+
149
+
128
+#define DO_3SAME(INSN, FUNC) \
150
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
129
+ static bool trans_##INSN##_3s(DisasContext *s, arg_3same *a) \
151
+ bool tag_checked, int log2_size)
130
+ { \
152
+{
131
+ return do_3same(s, a, FUNC); \
153
+ return gen_mte_check1_mmuidx(s, addr, is_write, tag_checked, log2_size,
132
+ }
154
+ false, get_mem_index(s));
133
+
155
+}
134
+DO_3SAME(VADD, tcg_gen_gvec_add)
156
+
135
+DO_3SAME(VSUB, tcg_gen_gvec_sub)
157
typedef struct DisasCompare64 {
136
diff --git a/target/arm/translate.c b/target/arm/translate.c
158
TCGCond cond;
137
index XXXXXXX..XXXXXXX 100644
159
TCGv_i64 value;
138
--- a/target/arm/translate.c
160
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap(DisasContext *s, int rs, int rt,
139
+++ b/target/arm/translate.c
161
if (rn == 31) {
140
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
162
gen_check_sp_alignment(s);
163
}
164
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
165
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size);
166
tcg_gen_atomic_cmpxchg_i64(tcg_rs, clean_addr, tcg_rs, tcg_rt, memidx,
167
size | MO_ALIGN | s->be_data);
168
}
169
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt,
170
if (rn == 31) {
171
gen_check_sp_alignment(s);
172
}
173
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
174
+
175
+ /* This is a single atomic access, despite the "pair". */
176
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size + 1);
177
178
if (size == 2) {
179
TCGv_i64 cmp = tcg_temp_new_i64();
180
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
181
if (is_lasr) {
182
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
183
}
184
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
185
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
186
+ true, rn != 31, size);
187
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, false);
188
return;
189
190
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
191
if (rn == 31) {
192
gen_check_sp_alignment(s);
193
}
194
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
195
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
196
+ false, rn != 31, size);
197
s->is_ldex = true;
198
gen_load_exclusive(s, rt, rt2, clean_addr, size, false);
199
if (is_lasr) {
200
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
201
gen_check_sp_alignment(s);
202
}
203
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
204
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
205
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
206
+ true, rn != 31, size);
207
do_gpr_st(s, cpu_reg(s, rt), clean_addr, size, true, rt,
208
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
209
return;
210
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
211
if (rn == 31) {
212
gen_check_sp_alignment(s);
213
}
214
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
215
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
216
+ false, rn != 31, size);
217
do_gpr_ld(s, cpu_reg(s, rt), clean_addr, size, false, false, true, rt,
218
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
219
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
220
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
221
if (is_lasr) {
222
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
141
}
223
}
142
return 0;
224
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
143
225
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
144
- case NEON_3R_VADD_VSUB:
226
+ true, rn != 31, size);
145
- if (u) {
227
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, true);
146
- tcg_gen_gvec_sub(size, rd_ofs, rn_ofs, rm_ofs,
228
return;
147
- vec_size, vec_size);
229
}
148
- } else {
230
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
149
- tcg_gen_gvec_add(size, rd_ofs, rn_ofs, rm_ofs,
231
if (rn == 31) {
150
- vec_size, vec_size);
232
gen_check_sp_alignment(s);
151
- }
233
}
152
- return 0;
234
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
153
-
235
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
154
case NEON_3R_VQADD:
236
+ false, rn != 31, size);
155
tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
237
s->is_ldex = true;
156
rn_ofs, rm_ofs, vec_size, vec_size,
238
gen_load_exclusive(s, rt, rt2, clean_addr, size, true);
157
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
239
if (is_lasr) {
158
tcg_gen_gvec_3(rd_ofs, rm_ofs, rn_ofs, vec_size, vec_size,
240
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
159
u ? &ushl_op[size] : &sshl_op[size]);
241
bool iss_valid = !is_vector;
160
return 0;
242
bool post_index;
161
+
243
bool writeback;
162
+ case NEON_3R_VADD_VSUB:
244
+ int memidx;
163
+ /* Already handled by decodetree */
245
164
+ return 1;
246
TCGv_i64 clean_addr, dirty_addr;
165
}
247
166
248
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
167
if (size == 3) {
249
if (!post_index) {
250
tcg_gen_addi_i64(dirty_addr, dirty_addr, imm9);
251
}
252
- clean_addr = clean_data_tbi(s, dirty_addr);
253
+
254
+ memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
255
+ clean_addr = gen_mte_check1_mmuidx(s, dirty_addr, is_store,
256
+ writeback || rn != 31,
257
+ size, is_unpriv, memidx);
258
259
if (is_vector) {
260
if (is_store) {
261
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
262
}
263
} else {
264
TCGv_i64 tcg_rt = cpu_reg(s, rt);
265
- int memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
266
bool iss_sf = disas_ldst_compute_iss_sf(size, is_signed, opc);
267
268
if (is_store) {
269
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn,
270
ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
271
272
tcg_gen_add_i64(dirty_addr, dirty_addr, tcg_rm);
273
- clean_addr = clean_data_tbi(s, dirty_addr);
274
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, true, size);
275
276
if (is_vector) {
277
if (is_store) {
278
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn,
279
dirty_addr = read_cpu_reg_sp(s, rn, 1);
280
offset = imm12 << size;
281
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
282
- clean_addr = clean_data_tbi(s, dirty_addr);
283
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, rn != 31, size);
284
285
if (is_vector) {
286
if (is_store) {
287
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
288
if (rn == 31) {
289
gen_check_sp_alignment(s);
290
}
291
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
292
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), false, rn != 31, size);
293
294
if (o3_opc == 014) {
295
/*
296
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pac(DisasContext *s, uint32_t insn,
297
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
298
299
/* Note that "clean" and "dirty" here refer to TBI not PAC. */
300
- clean_addr = clean_data_tbi(s, dirty_addr);
301
+ clean_addr = gen_mte_check1(s, dirty_addr, false,
302
+ is_wback || rn != 31, size);
303
304
tcg_rt = cpu_reg(s, rt);
305
do_gpr_ld(s, tcg_rt, clean_addr, size, /* is_signed */ false,
168
--
306
--
169
2.20.1
307
2.20.1
170
308
171
309
diff view generated by jsdifflib
1
Convert the Neon comparison ops in the 3-reg-same grouping
1
From: Richard Henderson <richard.henderson@linaro.org>
2
to decodetree.
2
3
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform multiple logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-25-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-18-peter.maydell@linaro.org
7
---
11
---
8
target/arm/neon-dp.decode | 8 ++++++++
12
target/arm/helper-a64.h | 1 +
9
target/arm/translate-neon.inc.c | 22 ++++++++++++++++++++++
13
target/arm/translate-a64.h | 2 ++
10
target/arm/translate.c | 23 +++--------------------
14
target/arm/mte_helper.c | 8 +++++
11
3 files changed, 33 insertions(+), 20 deletions(-)
15
target/arm/translate-a64.c | 71 +++++++++++++++++++++++++++++---------
12
16
4 files changed, 66 insertions(+), 16 deletions(-)
13
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
17
14
index XXXXXXX..XXXXXXX 100644
18
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
--- a/target/arm/neon-dp.decode
19
index XXXXXXX..XXXXXXX 100644
16
+++ b/target/arm/neon-dp.decode
20
--- a/target/arm/helper-a64.h
17
@@ -XXX,XX +XXX,XX @@ VBSL_3s 1111 001 1 0 . 01 .... .... 0001 ... 1 .... @3same_logic
21
+++ b/target/arm/helper-a64.h
18
VBIT_3s 1111 001 1 0 . 10 .... .... 0001 ... 1 .... @3same_logic
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
19
VBIF_3s 1111 001 1 0 . 11 .... .... 0001 ... 1 .... @3same_logic
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
20
24
21
+VCGT_S_3s 1111 001 0 0 . .. .... .... 0011 . . . 0 .... @3same
25
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
22
+VCGT_U_3s 1111 001 1 0 . .. .... .... 0011 . . . 0 .... @3same
26
+DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
23
+VCGE_S_3s 1111 001 0 0 . .. .... .... 0011 . . . 1 .... @3same
27
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
24
+VCGE_U_3s 1111 001 1 0 . .. .... .... 0011 . . . 1 .... @3same
28
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
25
+
29
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
26
VMAX_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 0 .... @3same
30
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
27
VMAX_U_3s 1111 001 1 0 . .. .... .... 0110 . . . 0 .... @3same
31
index XXXXXXX..XXXXXXX 100644
28
VMIN_S_3s 1111 001 0 0 . .. .... .... 0110 . . . 1 .... @3same
32
--- a/target/arm/translate-a64.h
29
@@ -XXX,XX +XXX,XX @@ VMIN_U_3s 1111 001 1 0 . .. .... .... 0110 . . . 1 .... @3same
33
+++ b/target/arm/translate-a64.h
30
34
@@ -XXX,XX +XXX,XX @@ bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
31
VADD_3s 1111 001 0 0 . .. .... .... 1000 . . . 0 .... @3same
35
bool sve_access_check(DisasContext *s);
32
VSUB_3s 1111 001 1 0 . .. .... .... 1000 . . . 0 .... @3same
36
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
33
+
37
bool tag_checked, int log2_size);
34
+VTST_3s 1111 001 0 0 . .. .... .... 1000 . . . 1 .... @3same
38
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
35
+VCEQ_3s 1111 001 1 0 . .. .... .... 1000 . . . 1 .... @3same
39
+ bool tag_checked, int count, int log2_esize);
36
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
40
37
index XXXXXXX..XXXXXXX 100644
41
/* We should have at some point before trying to access an FP register
38
--- a/target/arm/translate-neon.inc.c
42
* done the necessary access check, so assert that
39
+++ b/target/arm/translate-neon.inc.c
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
40
@@ -XXX,XX +XXX,XX @@ DO_3SAME_NO_SZ_3(VMAX_S, tcg_gen_gvec_smax)
44
index XXXXXXX..XXXXXXX 100644
41
DO_3SAME_NO_SZ_3(VMAX_U, tcg_gen_gvec_umax)
45
--- a/target/arm/mte_helper.c
42
DO_3SAME_NO_SZ_3(VMIN_S, tcg_gen_gvec_smin)
46
+++ b/target/arm/mte_helper.c
43
DO_3SAME_NO_SZ_3(VMIN_U, tcg_gen_gvec_umin)
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
44
+
48
{
45
+#define DO_3SAME_CMP(INSN, COND) \
49
return ptr;
46
+ static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
50
}
47
+ uint32_t rn_ofs, uint32_t rm_ofs, \
51
+
48
+ uint32_t oprsz, uint32_t maxsz) \
52
+/*
49
+ { \
53
+ * Perform an MTE checked access for multiple logical accesses.
50
+ tcg_gen_gvec_cmp(COND, vece, rd_ofs, rn_ofs, rm_ofs, oprsz, maxsz); \
54
+ */
51
+ } \
55
+uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
52
+ DO_3SAME_NO_SZ_3(INSN, gen_##INSN##_3s)
53
+
54
+DO_3SAME_CMP(VCGT_S, TCG_COND_GT)
55
+DO_3SAME_CMP(VCGT_U, TCG_COND_GTU)
56
+DO_3SAME_CMP(VCGE_S, TCG_COND_GE)
57
+DO_3SAME_CMP(VCGE_U, TCG_COND_GEU)
58
+DO_3SAME_CMP(VCEQ, TCG_COND_EQ)
59
+
60
+static void gen_VTST_3s(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
61
+ uint32_t rm_ofs, uint32_t oprsz, uint32_t maxsz)
62
+{
56
+{
63
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, oprsz, maxsz, &cmtst_op[vece]);
57
+ return ptr;
64
+}
58
+}
65
+DO_3SAME_NO_SZ_3(VTST, gen_VTST_3s)
59
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
66
diff --git a/target/arm/translate.c b/target/arm/translate.c
60
index XXXXXXX..XXXXXXX 100644
67
index XXXXXXX..XXXXXXX 100644
61
--- a/target/arm/translate-a64.c
68
--- a/target/arm/translate.c
62
+++ b/target/arm/translate-a64.c
69
+++ b/target/arm/translate.c
63
@@ -XXX,XX +XXX,XX @@ TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
70
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
64
false, get_mem_index(s));
71
u ? &mls_op[size] : &mla_op[size]);
65
}
72
return 0;
66
73
67
+/*
74
- case NEON_3R_VTST_VCEQ:
68
+ * For MTE, check multiple logical sequential accesses.
75
- if (u) { /* VCEQ */
69
+ */
76
- tcg_gen_gvec_cmp(TCG_COND_EQ, size, rd_ofs, rn_ofs, rm_ofs,
70
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
77
- vec_size, vec_size);
71
+ bool tag_checked, int log2_esize, int total_size)
78
- } else { /* VTST */
72
+{
79
- tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
73
+ if (tag_checked && s->mte_active[0] && total_size != (1 << log2_esize)) {
80
- vec_size, vec_size, &cmtst_op[size]);
74
+ TCGv_i32 tcg_desc;
81
- }
75
+ TCGv_i64 ret;
82
- return 0;
76
+ int desc = 0;
77
+
78
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
79
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
80
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
81
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
82
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_esize);
83
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, total_size);
84
+ tcg_desc = tcg_const_i32(desc);
85
+
86
+ ret = new_tmp_a64(s);
87
+ gen_helper_mte_checkN(ret, cpu_env, tcg_desc, addr);
88
+ tcg_temp_free_i32(tcg_desc);
89
+
90
+ return ret;
91
+ }
92
+ return gen_mte_check1(s, addr, is_write, tag_checked, log2_esize);
93
+}
94
+
95
typedef struct DisasCompare64 {
96
TCGCond cond;
97
TCGv_i64 value;
98
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
99
}
100
}
101
102
- clean_addr = clean_data_tbi(s, dirty_addr);
103
+ clean_addr = gen_mte_checkN(s, dirty_addr, !is_load,
104
+ (wback || rn != 31) && !set_tag,
105
+ size, 2 << size);
106
+
107
if (is_vector) {
108
if (is_load) {
109
do_fp_ld(s, rt, clean_addr, size);
110
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
111
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
112
MemOp endian = s->be_data;
113
114
- int ebytes; /* bytes per element */
115
+ int total; /* total bytes */
116
int elements; /* elements per vector */
117
int rpt; /* num iterations */
118
int selem; /* structure elements */
119
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
120
endian = MO_LE;
121
}
122
123
- /* Consecutive little-endian elements from a single register
124
+ total = rpt * selem * (is_q ? 16 : 8);
125
+ tcg_rn = cpu_reg_sp(s, rn);
126
+
127
+ /*
128
+ * Issue the MTE check vs the logical repeat count, before we
129
+ * promote consecutive little-endian elements below.
130
+ */
131
+ clean_addr = gen_mte_checkN(s, tcg_rn, is_store, is_postidx || rn != 31,
132
+ size, total);
133
+
134
+ /*
135
+ * Consecutive little-endian elements from a single register
136
* can be promoted to a larger little-endian operation.
137
*/
138
if (selem == 1 && endian == MO_LE) {
139
size = 3;
140
}
141
- ebytes = 1 << size;
142
- elements = (is_q ? 16 : 8) / ebytes;
83
-
143
-
84
- case NEON_3R_VCGT:
144
- tcg_rn = cpu_reg_sp(s, rn);
85
- tcg_gen_gvec_cmp(u ? TCG_COND_GTU : TCG_COND_GT, size,
145
- clean_addr = clean_data_tbi(s, tcg_rn);
86
- rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
146
- tcg_ebytes = tcg_const_i64(ebytes);
87
- return 0;
147
+ elements = (is_q ? 16 : 8) >> size;
148
149
+ tcg_ebytes = tcg_const_i64(1 << size);
150
for (r = 0; r < rpt; r++) {
151
int e;
152
for (e = 0; e < elements; e++) {
153
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
154
155
if (is_postidx) {
156
if (rm == 31) {
157
- tcg_gen_addi_i64(tcg_rn, tcg_rn, rpt * elements * selem * ebytes);
158
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
159
} else {
160
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
161
}
162
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
163
int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
164
bool replicate = false;
165
int index = is_q << 3 | S << 2 | size;
166
- int ebytes, xs;
167
+ int xs, total;
168
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
169
170
if (extract32(insn, 31, 1)) {
171
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
172
return;
173
}
174
175
- ebytes = 1 << scale;
88
-
176
-
89
- case NEON_3R_VCGE:
177
if (rn == 31) {
90
- tcg_gen_gvec_cmp(u ? TCG_COND_GEU : TCG_COND_GE, size,
178
gen_check_sp_alignment(s);
91
- rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
179
}
92
- return 0;
180
93
-
181
+ total = selem << scale;
94
case NEON_3R_VSHL:
182
tcg_rn = cpu_reg_sp(s, rn);
95
/* Note the operation is vshl vd,vm,vn */
183
- clean_addr = clean_data_tbi(s, tcg_rn);
96
tcg_gen_gvec_3(rd_ofs, rm_ofs, rn_ofs, vec_size, vec_size,
184
- tcg_ebytes = tcg_const_i64(ebytes);
97
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
185
98
case NEON_3R_LOGIC:
186
+ clean_addr = gen_mte_checkN(s, tcg_rn, !is_load, is_postidx || rn != 31,
99
case NEON_3R_VMAX:
187
+ scale, total);
100
case NEON_3R_VMIN:
188
+
101
+ case NEON_3R_VTST_VCEQ:
189
+ tcg_ebytes = tcg_const_i64(1 << scale);
102
+ case NEON_3R_VCGT:
190
for (xs = 0; xs < selem; xs++) {
103
+ case NEON_3R_VCGE:
191
if (replicate) {
104
/* Already handled by decodetree */
192
/* Load and replicate to all elements */
105
return 1;
193
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
194
195
if (is_postidx) {
196
if (rm == 31) {
197
- tcg_gen_addi_i64(tcg_rn, tcg_rn, selem * ebytes);
198
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
199
} else {
200
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
106
}
201
}
107
--
202
--
108
2.20.1
203
2.20.1
109
204
110
205
diff view generated by jsdifflib
1
Convert the Neon "load/store multiple structures" insns to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-26-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200430181003.21682-12-peter.maydell@linaro.org
6
---
9
---
7
target/arm/neon-ls.decode | 7 ++
10
target/arm/internals.h | 48 +++++++++++++++
8
target/arm/translate-neon.inc.c | 124 ++++++++++++++++++++++++++++++++
11
target/arm/mte_helper.c | 132 +++++++++++++++++++++++++++++++++++++++-
9
target/arm/translate.c | 91 +----------------------
12
2 files changed, 179 insertions(+), 1 deletion(-)
10
3 files changed, 133 insertions(+), 89 deletions(-)
13
11
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
12
diff --git a/target/arm/neon-ls.decode b/target/arm/neon-ls.decode
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/neon-ls.decode
16
--- a/target/arm/internals.h
15
+++ b/target/arm/neon-ls.decode
17
+++ b/target/arm/internals.h
16
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, WRITE, 8, 1)
17
# 0b1111_1001_xxx0_xxxx_xxxx_xxxx_xxxx_xxxx
19
FIELD(MTEDESC, ESIZE, 9, 5)
18
# This file works on the A32 encoding only; calling code for T32 has to
20
FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
19
# transform the insn into the A32 version first.
21
20
+
22
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
21
+%vd_dp 22:1 12:4
23
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
22
+
24
+ uint64_t ptr, uintptr_t ra);
23
+# Neon load/store multiple structures
25
+
24
+
26
static inline int allocation_tag_from_addr(uint64_t ptr)
25
+VLDST_multiple 1111 0100 0 . l:1 0 rn:4 .... itype:4 size:2 align:2 rm:4 \
27
{
26
+ vd=%vd_dp
28
return extract64(ptr, 56, 4);
27
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
29
@@ -XXX,XX +XXX,XX @@ static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
30
return deposit64(ptr, 56, 4, rtag);
31
}
32
33
+/* Return true if tbi bits mean that the access is checked. */
34
+static inline bool tbi_check(uint32_t desc, int bit55)
35
+{
36
+ return (desc >> (R_MTEDESC_TBI_SHIFT + bit55)) & 1;
37
+}
38
+
39
+/* Return true if tcma bits mean that the access is unchecked. */
40
+static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
41
+{
42
+ /*
43
+ * We had extracted bit55 and ptr_tag for other reasons, so fold
44
+ * (ptr<59:55> == 00000 || ptr<59:55> == 11111) into a single test.
45
+ */
46
+ bool match = ((ptr_tag + bit55) & 0xf) == 0;
47
+ bool tcma = (desc >> (R_MTEDESC_TCMA_SHIFT + bit55)) & 1;
48
+ return tcma && match;
49
+}
50
+
51
+/*
52
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
53
+ * for the tag to be present in the FAR_ELx register. But for user-only
54
+ * mode, we do not have a TLB with which to implement this, so we must
55
+ * remove the top byte.
56
+ */
57
+static inline uint64_t useronly_clean_ptr(uint64_t ptr)
58
+{
59
+ /* TBI is known to be enabled. */
60
+#ifdef CONFIG_USER_ONLY
61
+ ptr = sextract64(ptr, 0, 56);
62
+#endif
63
+ return ptr;
64
+}
65
+
66
+static inline uint64_t useronly_maybe_clean_ptr(uint32_t desc, uint64_t ptr)
67
+{
68
+#ifdef CONFIG_USER_ONLY
69
+ int64_t clean_ptr = sextract64(ptr, 0, 56);
70
+ if (tbi_check(desc, clean_ptr < 0)) {
71
+ ptr = clean_ptr;
72
+ }
73
+#endif
74
+ return ptr;
75
+}
76
+
77
#endif
78
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
28
index XXXXXXX..XXXXXXX 100644
79
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-neon.inc.c
80
--- a/target/arm/mte_helper.c
30
+++ b/target/arm/translate-neon.inc.c
81
+++ b/target/arm/mte_helper.c
31
@@ -XXX,XX +XXX,XX @@ static bool trans_VFML_scalar(DisasContext *s, arg_VFML_scalar *a)
82
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
32
gen_helper_gvec_fmlal_idx_a32);
83
}
33
return true;
34
}
84
}
35
+
85
36
+static struct {
86
+/* Record a tag check failure. */
37
+ int nregs;
87
+static void mte_check_fail(CPUARMState *env, int mmu_idx,
38
+ int interleave;
88
+ uint64_t dirty_ptr, uintptr_t ra)
39
+ int spacing;
89
+{
40
+} const neon_ls_element_type[11] = {
90
+ ARMMMUIdx arm_mmu_idx = core_to_aa64_mmu_idx(mmu_idx);
41
+ {1, 4, 1},
91
+ int el, reg_el, tcf, select;
42
+ {1, 4, 2},
92
+ uint64_t sctlr;
43
+ {4, 1, 1},
93
+
44
+ {2, 2, 2},
94
+ reg_el = regime_el(env, arm_mmu_idx);
45
+ {1, 3, 1},
95
+ sctlr = env->cp15.sctlr_el[reg_el];
46
+ {1, 3, 2},
96
+
47
+ {3, 1, 1},
97
+ switch (arm_mmu_idx) {
48
+ {1, 1, 1},
98
+ case ARMMMUIdx_E10_0:
49
+ {1, 2, 1},
99
+ case ARMMMUIdx_E20_0:
50
+ {1, 2, 2},
100
+ el = 0;
51
+ {2, 1, 1}
101
+ tcf = extract64(sctlr, 38, 2);
52
+};
53
+
54
+static void gen_neon_ldst_base_update(DisasContext *s, int rm, int rn,
55
+ int stride)
56
+{
57
+ if (rm != 15) {
58
+ TCGv_i32 base;
59
+
60
+ base = load_reg(s, rn);
61
+ if (rm == 13) {
62
+ tcg_gen_addi_i32(base, base, stride);
63
+ } else {
64
+ TCGv_i32 index;
65
+ index = load_reg(s, rm);
66
+ tcg_gen_add_i32(base, base, index);
67
+ tcg_temp_free_i32(index);
68
+ }
69
+ store_reg(s, rn, base);
70
+ }
71
+}
72
+
73
+static bool trans_VLDST_multiple(DisasContext *s, arg_VLDST_multiple *a)
74
+{
75
+ /* Neon load/store multiple structures */
76
+ int nregs, interleave, spacing, reg, n;
77
+ MemOp endian = s->be_data;
78
+ int mmu_idx = get_mem_index(s);
79
+ int size = a->size;
80
+ TCGv_i64 tmp64;
81
+ TCGv_i32 addr, tmp;
82
+
83
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
84
+ return false;
85
+ }
86
+
87
+ /* UNDEF accesses to D16-D31 if they don't exist */
88
+ if (!dc_isar_feature(aa32_simd_r32, s) && (a->vd & 0x10)) {
89
+ return false;
90
+ }
91
+ if (a->itype > 10) {
92
+ return false;
93
+ }
94
+ /* Catch UNDEF cases for bad values of align field */
95
+ switch (a->itype & 0xc) {
96
+ case 4:
97
+ if (a->align >= 2) {
98
+ return false;
99
+ }
100
+ break;
101
+ case 8:
102
+ if (a->align == 3) {
103
+ return false;
104
+ }
105
+ break;
102
+ break;
106
+ default:
103
+ default:
104
+ el = reg_el;
105
+ tcf = extract64(sctlr, 40, 2);
106
+ }
107
+
108
+ switch (tcf) {
109
+ case 1:
110
+ /*
111
+ * Tag check fail causes a synchronous exception.
112
+ *
113
+ * In restore_state_to_opc, we set the exception syndrome
114
+ * for the load or store operation. Unwind first so we
115
+ * may overwrite that with the syndrome for the tag check.
116
+ */
117
+ cpu_restore_state(env_cpu(env), ra, true);
118
+ env->exception.vaddress = dirty_ptr;
119
+ raise_exception(env, EXCP_DATA_ABORT,
120
+ syn_data_abort_no_iss(el != 0, 0, 0, 0, 0, 0, 0x11),
121
+ exception_target_el(env));
122
+ /* noreturn, but fall through to the assert anyway */
123
+
124
+ case 0:
125
+ /*
126
+ * Tag check fail does not affect the PE.
127
+ * We eliminate this case by not setting MTE_ACTIVE
128
+ * in tb_flags, so that we never make this runtime call.
129
+ */
130
+ g_assert_not_reached();
131
+
132
+ case 2:
133
+ /* Tag check fail causes asynchronous flag set. */
134
+ mmu_idx = arm_mmu_idx_el(env, el);
135
+ if (regime_has_2_ranges(mmu_idx)) {
136
+ select = extract64(dirty_ptr, 55, 1);
137
+ } else {
138
+ select = 0;
139
+ }
140
+ env->cp15.tfsr_el[el] |= 1 << select;
107
+ break;
141
+ break;
108
+ }
142
+
109
+ nregs = neon_ls_element_type[a->itype].nregs;
143
+ default:
110
+ interleave = neon_ls_element_type[a->itype].interleave;
144
+ /* Case 3: Reserved. */
111
+ spacing = neon_ls_element_type[a->itype].spacing;
145
+ qemu_log_mask(LOG_GUEST_ERROR,
112
+ if (size == 3 && (interleave | spacing) != 1) {
146
+ "Tag check failure with SCTLR_EL%d.TCF%s "
113
+ return false;
147
+ "set to reserved value %d\n",
114
+ }
148
+ reg_el, el ? "" : "0", tcf);
115
+
149
+ break;
116
+ if (!vfp_access_check(s)) {
150
+ }
151
+}
152
+
153
/*
154
* Perform an MTE checked access for a single logical or atomic access.
155
*/
156
+static bool mte_probe1_int(CPUARMState *env, uint32_t desc, uint64_t ptr,
157
+ uintptr_t ra, int bit55)
158
+{
159
+ int mem_tag, mmu_idx, ptr_tag, size;
160
+ MMUAccessType type;
161
+ uint8_t *mem;
162
+
163
+ ptr_tag = allocation_tag_from_addr(ptr);
164
+
165
+ if (tcma_check(desc, bit55, ptr_tag)) {
117
+ return true;
166
+ return true;
118
+ }
167
+ }
119
+
168
+
120
+ /* For our purposes, bytes are always little-endian. */
169
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
121
+ if (size == 0) {
170
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
122
+ endian = MO_LE;
171
+ size = FIELD_EX32(desc, MTEDESC, ESIZE);
123
+ }
172
+
124
+ /*
173
+ mem = allocation_tag_mem(env, mmu_idx, ptr, type, size,
125
+ * Consecutive little-endian elements from a single register
174
+ MMU_DATA_LOAD, 1, ra);
126
+ * can be promoted to a larger little-endian operation.
175
+ if (!mem) {
127
+ */
176
+ return true;
128
+ if (interleave == 1 && endian == MO_LE) {
177
+ }
129
+ size = 3;
178
+
130
+ }
179
+ mem_tag = load_tag1(ptr, mem);
131
+ tmp64 = tcg_temp_new_i64();
180
+ return ptr_tag == mem_tag;
132
+ addr = tcg_temp_new_i32();
181
+}
133
+ tmp = tcg_const_i32(1 << size);
182
+
134
+ load_reg_var(s, addr, a->rn);
183
+/*
135
+ for (reg = 0; reg < nregs; reg++) {
184
+ * No-fault version of mte_check1, to be used by SVE for MemSingleNF.
136
+ for (n = 0; n < 8 >> size; n++) {
185
+ * Returns false if the access is Checked and the check failed. This
137
+ int xs;
186
+ * is only intended to probe the tag -- the validity of the page must
138
+ for (xs = 0; xs < interleave; xs++) {
187
+ * be checked beforehand.
139
+ int tt = a->vd + reg + spacing * xs;
188
+ */
140
+
189
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr)
141
+ if (a->l) {
190
+{
142
+ gen_aa32_ld_i64(s, tmp64, addr, mmu_idx, endian | size);
191
+ int bit55 = extract64(ptr, 55, 1);
143
+ neon_store_element64(tt, n, size, tmp64);
192
+
144
+ } else {
193
+ /* If TBI is disabled, the access is unchecked. */
145
+ neon_load_element64(tmp64, tt, n, size);
194
+ if (unlikely(!tbi_check(desc, bit55))) {
146
+ gen_aa32_st_i64(s, tmp64, addr, mmu_idx, endian | size);
195
+ return true;
147
+ }
196
+ }
148
+ tcg_gen_add_i32(addr, addr, tmp);
197
+
149
+ }
198
+ return mte_probe1_int(env, desc, ptr, 0, bit55);
150
+ }
199
+}
151
+ }
200
+
152
+ tcg_temp_free_i32(addr);
201
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
153
+ tcg_temp_free_i32(tmp);
202
+ uint64_t ptr, uintptr_t ra)
154
+ tcg_temp_free_i64(tmp64);
203
+{
155
+
204
+ int bit55 = extract64(ptr, 55, 1);
156
+ gen_neon_ldst_base_update(s, a->rm, a->rn, nregs * interleave * 8);
205
+
157
+ return true;
206
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
158
+}
207
+ if (unlikely(!tbi_check(desc, bit55))) {
159
diff --git a/target/arm/translate.c b/target/arm/translate.c
208
+ return ptr;
160
index XXXXXXX..XXXXXXX 100644
209
+ }
161
--- a/target/arm/translate.c
210
+
162
+++ b/target/arm/translate.c
211
+ if (unlikely(!mte_probe1_int(env, desc, ptr, ra, bit55))) {
163
@@ -XXX,XX +XXX,XX @@ static void gen_neon_trn_u16(TCGv_i32 t0, TCGv_i32 t1)
212
+ int mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
213
+ mte_check_fail(env, mmu_idx, ptr, ra);
214
+ }
215
+
216
+ return useronly_clean_ptr(ptr);
217
+}
218
+
219
uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
220
{
221
- return ptr;
222
+ return mte_check1(env, desc, ptr, GETPC());
164
}
223
}
165
224
166
225
/*
167
-static struct {
168
- int nregs;
169
- int interleave;
170
- int spacing;
171
-} const neon_ls_element_type[11] = {
172
- {1, 4, 1},
173
- {1, 4, 2},
174
- {4, 1, 1},
175
- {2, 2, 2},
176
- {1, 3, 1},
177
- {1, 3, 2},
178
- {3, 1, 1},
179
- {1, 1, 1},
180
- {1, 2, 1},
181
- {1, 2, 2},
182
- {2, 1, 1}
183
-};
184
-
185
/* Translate a NEON load/store element instruction. Return nonzero if the
186
instruction is invalid. */
187
static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
188
{
189
int rd, rn, rm;
190
- int op;
191
int nregs;
192
- int interleave;
193
- int spacing;
194
int stride;
195
int size;
196
int reg;
197
int load;
198
- int n;
199
int vec_size;
200
- int mmu_idx;
201
- MemOp endian;
202
TCGv_i32 addr;
203
TCGv_i32 tmp;
204
- TCGv_i32 tmp2;
205
- TCGv_i64 tmp64;
206
207
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
208
return 1;
209
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
210
rn = (insn >> 16) & 0xf;
211
rm = insn & 0xf;
212
load = (insn & (1 << 21)) != 0;
213
- endian = s->be_data;
214
- mmu_idx = get_mem_index(s);
215
if ((insn & (1 << 23)) == 0) {
216
- /* Load store all elements. */
217
- op = (insn >> 8) & 0xf;
218
- size = (insn >> 6) & 3;
219
- if (op > 10)
220
- return 1;
221
- /* Catch UNDEF cases for bad values of align field */
222
- switch (op & 0xc) {
223
- case 4:
224
- if (((insn >> 5) & 1) == 1) {
225
- return 1;
226
- }
227
- break;
228
- case 8:
229
- if (((insn >> 4) & 3) == 3) {
230
- return 1;
231
- }
232
- break;
233
- default:
234
- break;
235
- }
236
- nregs = neon_ls_element_type[op].nregs;
237
- interleave = neon_ls_element_type[op].interleave;
238
- spacing = neon_ls_element_type[op].spacing;
239
- if (size == 3 && (interleave | spacing) != 1) {
240
- return 1;
241
- }
242
- /* For our purposes, bytes are always little-endian. */
243
- if (size == 0) {
244
- endian = MO_LE;
245
- }
246
- /* Consecutive little-endian elements from a single register
247
- * can be promoted to a larger little-endian operation.
248
- */
249
- if (interleave == 1 && endian == MO_LE) {
250
- size = 3;
251
- }
252
- tmp64 = tcg_temp_new_i64();
253
- addr = tcg_temp_new_i32();
254
- tmp2 = tcg_const_i32(1 << size);
255
- load_reg_var(s, addr, rn);
256
- for (reg = 0; reg < nregs; reg++) {
257
- for (n = 0; n < 8 >> size; n++) {
258
- int xs;
259
- for (xs = 0; xs < interleave; xs++) {
260
- int tt = rd + reg + spacing * xs;
261
-
262
- if (load) {
263
- gen_aa32_ld_i64(s, tmp64, addr, mmu_idx, endian | size);
264
- neon_store_element64(tt, n, size, tmp64);
265
- } else {
266
- neon_load_element64(tmp64, tt, n, size);
267
- gen_aa32_st_i64(s, tmp64, addr, mmu_idx, endian | size);
268
- }
269
- tcg_gen_add_i32(addr, addr, tmp2);
270
- }
271
- }
272
- }
273
- tcg_temp_free_i32(addr);
274
- tcg_temp_free_i32(tmp2);
275
- tcg_temp_free_i64(tmp64);
276
- stride = nregs * interleave * 8;
277
+ /* Load store all elements -- handled already by decodetree */
278
+ return 1;
279
} else {
280
size = (insn >> 10) & 3;
281
if (size == 3) {
282
--
226
--
283
2.20.1
227
2.20.1
284
228
285
229
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Embed the GEMs into the SoC type.
3
Fill out the stub that was added earlier.
4
4
5
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Message-id: 20200626033144.790098-27-richard.henderson@linaro.org
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
10
Message-id: 20200427181649.26851-6-edgar.iglesias@gmail.com
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
9
---
13
include/hw/arm/xlnx-versal.h | 3 ++-
10
target/arm/internals.h | 2 +
14
hw/arm/xlnx-versal.c | 15 ++++++++-------
11
target/arm/mte_helper.c | 165 +++++++++++++++++++++++++++++++++++++++-
15
2 files changed, 10 insertions(+), 8 deletions(-)
12
2 files changed, 166 insertions(+), 1 deletion(-)
16
13
17
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
18
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/arm/xlnx-versal.h
16
--- a/target/arm/internals.h
20
+++ b/include/hw/arm/xlnx-versal.h
17
+++ b/target/arm/internals.h
21
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
22
#include "hw/arm/boot.h"
19
bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
23
#include "hw/intc/arm_gicv3.h"
20
uint64_t mte_check1(CPUARMState *env, uint32_t desc,
24
#include "hw/char/pl011.h"
21
uint64_t ptr, uintptr_t ra);
25
+#include "hw/net/cadence_gem.h"
22
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
26
23
+ uint64_t ptr, uintptr_t ra);
27
#define TYPE_XLNX_VERSAL "xlnx-versal"
24
28
#define XLNX_VERSAL(obj) OBJECT_CHECK(Versal, (obj), TYPE_XLNX_VERSAL)
25
static inline int allocation_tag_from_addr(uint64_t ptr)
29
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
26
{
30
27
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
31
struct {
32
PL011State uart[XLNX_VERSAL_NR_UARTS];
33
- SysBusDevice *gem[XLNX_VERSAL_NR_GEMS];
34
+ CadenceGEMState gem[XLNX_VERSAL_NR_GEMS];
35
SysBusDevice *adma[XLNX_VERSAL_NR_ADMAS];
36
} iou;
37
} lpd;
38
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
39
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/xlnx-versal.c
29
--- a/target/arm/mte_helper.c
41
+++ b/hw/arm/xlnx-versal.c
30
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@ static void versal_create_gems(Versal *s, qemu_irq *pic)
31
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
43
DeviceState *dev;
32
/*
44
MemoryRegion *mr;
33
* Perform an MTE checked access for multiple logical accesses.
45
34
*/
46
- dev = qdev_create(NULL, "cadence_gem");
35
+
47
- s->lpd.iou.gem[i] = SYS_BUS_DEVICE(dev);
36
+/**
48
- object_property_add_child(OBJECT(s), name, OBJECT(dev), &error_fatal);
37
+ * checkN:
49
+ sysbus_init_child_obj(OBJECT(s), name,
38
+ * @tag: tag memory to test
50
+ &s->lpd.iou.gem[i], sizeof(s->lpd.iou.gem[i]),
39
+ * @odd: true to begin testing at tags at odd nibble
51
+ TYPE_CADENCE_GEM);
40
+ * @cmp: the tag to compare against
52
+ dev = DEVICE(&s->lpd.iou.gem[i]);
41
+ * @count: number of tags to test
53
if (nd->used) {
42
+ *
54
qemu_check_nic_model(nd, "cadence_gem");
43
+ * Return the number of successful tests.
55
qdev_set_nic_properties(dev, nd);
44
+ * Thus a return value < @count indicates a failure.
56
}
45
+ *
57
- object_property_set_int(OBJECT(s->lpd.iou.gem[i]),
46
+ * A note about sizes: count is expected to be small.
58
+ object_property_set_int(OBJECT(dev),
47
+ *
59
2, "num-priority-queues",
48
+ * The most common use will be LDP/STP of two integer registers,
60
&error_abort);
49
+ * which means 16 bytes of memory touching at most 2 tags, but
61
- object_property_set_link(OBJECT(s->lpd.iou.gem[i]),
50
+ * often the access is aligned and thus just 1 tag.
62
+ object_property_set_link(OBJECT(dev),
51
+ *
63
OBJECT(&s->mr_ps), "dma",
52
+ * Using AdvSIMD LD/ST (multiple), one can access 64 bytes of memory,
64
&error_abort);
53
+ * touching at most 5 tags. SVE LDR/STR (vector) with the default
65
qdev_init_nofail(dev);
54
+ * vector length is also 64 bytes; the maximum architectural length
66
55
+ * is 256 bytes touching at most 9 tags.
67
- mr = sysbus_mmio_get_region(s->lpd.iou.gem[i], 0);
56
+ *
68
+ mr = sysbus_mmio_get_region(SYS_BUS_DEVICE(dev), 0);
57
+ * The loop below uses 7 logical operations and 1 memory operation
69
memory_region_add_subregion(&s->mr_ps, addrs[i], mr);
58
+ * per tag pair. An implementation that loads an aligned word and
70
59
+ * uses masking to ignore adjacent tags requires 18 logical operations
71
- sysbus_connect_irq(s->lpd.iou.gem[i], 0, pic[irqs[i]]);
60
+ * and thus does not begin to pay off until 6 tags.
72
+ sysbus_connect_irq(SYS_BUS_DEVICE(dev), 0, pic[irqs[i]]);
61
+ * Which, according to the survey above, is unlikely to be common.
73
g_free(name);
62
+ */
74
}
63
+static int checkN(uint8_t *mem, int odd, int cmp, int count)
64
+{
65
+ int n = 0, diff;
66
+
67
+ /* Replicate the test tag and compare. */
68
+ cmp *= 0x11;
69
+ diff = *mem++ ^ cmp;
70
+
71
+ if (odd) {
72
+ goto start_odd;
73
+ }
74
+
75
+ while (1) {
76
+ /* Test even tag. */
77
+ if (unlikely((diff) & 0x0f)) {
78
+ break;
79
+ }
80
+ if (++n == count) {
81
+ break;
82
+ }
83
+
84
+ start_odd:
85
+ /* Test odd tag. */
86
+ if (unlikely((diff) & 0xf0)) {
87
+ break;
88
+ }
89
+ if (++n == count) {
90
+ break;
91
+ }
92
+
93
+ diff = *mem++ ^ cmp;
94
+ }
95
+ return n;
96
+}
97
+
98
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
99
+ uint64_t ptr, uintptr_t ra)
100
+{
101
+ int mmu_idx, ptr_tag, bit55;
102
+ uint64_t ptr_last, ptr_end, prev_page, next_page;
103
+ uint64_t tag_first, tag_end;
104
+ uint64_t tag_byte_first, tag_byte_end;
105
+ uint32_t esize, total, tag_count, tag_size, n, c;
106
+ uint8_t *mem1, *mem2;
107
+ MMUAccessType type;
108
+
109
+ bit55 = extract64(ptr, 55, 1);
110
+
111
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
112
+ if (unlikely(!tbi_check(desc, bit55))) {
113
+ return ptr;
114
+ }
115
+
116
+ ptr_tag = allocation_tag_from_addr(ptr);
117
+
118
+ if (tcma_check(desc, bit55, ptr_tag)) {
119
+ goto done;
120
+ }
121
+
122
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
123
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
124
+ esize = FIELD_EX32(desc, MTEDESC, ESIZE);
125
+ total = FIELD_EX32(desc, MTEDESC, TSIZE);
126
+
127
+ /* Find the addr of the end of the access, and of the last element. */
128
+ ptr_end = ptr + total;
129
+ ptr_last = ptr_end - esize;
130
+
131
+ /* Round the bounds to the tag granule, and compute the number of tags. */
132
+ tag_first = QEMU_ALIGN_DOWN(ptr, TAG_GRANULE);
133
+ tag_end = QEMU_ALIGN_UP(ptr_last, TAG_GRANULE);
134
+ tag_count = (tag_end - tag_first) / TAG_GRANULE;
135
+
136
+ /* Round the bounds to twice the tag granule, and compute the bytes. */
137
+ tag_byte_first = QEMU_ALIGN_DOWN(ptr, 2 * TAG_GRANULE);
138
+ tag_byte_end = QEMU_ALIGN_UP(ptr_last, 2 * TAG_GRANULE);
139
+
140
+ /* Locate the page boundaries. */
141
+ prev_page = ptr & TARGET_PAGE_MASK;
142
+ next_page = prev_page + TARGET_PAGE_SIZE;
143
+
144
+ if (likely(tag_end - prev_page <= TARGET_PAGE_SIZE)) {
145
+ /* Memory access stays on one page. */
146
+ tag_size = (tag_byte_end - tag_byte_first) / (2 * TAG_GRANULE);
147
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, total,
148
+ MMU_DATA_LOAD, tag_size, ra);
149
+ if (!mem1) {
150
+ goto done;
151
+ }
152
+ /* Perform all of the comparisons. */
153
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, tag_count);
154
+ } else {
155
+ /* Memory access crosses to next page. */
156
+ tag_size = (next_page - tag_byte_first) / (2 * TAG_GRANULE);
157
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, next_page - ptr,
158
+ MMU_DATA_LOAD, tag_size, ra);
159
+
160
+ tag_size = (tag_byte_end - next_page) / (2 * TAG_GRANULE);
161
+ mem2 = allocation_tag_mem(env, mmu_idx, next_page, type,
162
+ ptr_end - next_page,
163
+ MMU_DATA_LOAD, tag_size, ra);
164
+
165
+ /*
166
+ * Perform all of the comparisons.
167
+ * Note the possible but unlikely case of the operation spanning
168
+ * two pages that do not both have tagging enabled.
169
+ */
170
+ n = c = (next_page - tag_first) / TAG_GRANULE;
171
+ if (mem1) {
172
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, c);
173
+ }
174
+ if (n == c) {
175
+ if (!mem2) {
176
+ goto done;
177
+ }
178
+ n += checkN(mem2, 0, ptr_tag, tag_count - c);
179
+ }
180
+ }
181
+
182
+ /*
183
+ * If we failed, we know which granule. Compute the element that
184
+ * is first in that granule, and signal failure on that element.
185
+ */
186
+ if (unlikely(n < tag_count)) {
187
+ uint64_t fail_ofs;
188
+
189
+ fail_ofs = tag_first + n * TAG_GRANULE - ptr;
190
+ fail_ofs = ROUND_UP(fail_ofs, esize);
191
+ mte_check_fail(env, mmu_idx, ptr + fail_ofs, ra);
192
+ }
193
+
194
+ done:
195
+ return useronly_clean_ptr(ptr);
196
+}
197
+
198
uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
199
{
200
- return ptr;
201
+ return mte_checkN(env, desc, ptr, GETPC());
75
}
202
}
76
--
203
--
77
2.20.1
204
2.20.1
78
205
79
206
diff view generated by jsdifflib
1
Convert the V[US]DOT (vector) insns to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Use a special helper for DC_ZVA, rather than the more
4
general mte_checkN.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-28-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200430181003.21682-7-peter.maydell@linaro.org
6
---
10
---
7
target/arm/neon-shared.decode | 4 ++++
11
target/arm/helper-a64.h | 1 +
8
target/arm/translate-neon.inc.c | 32 ++++++++++++++++++++++++++++++++
12
target/arm/mte_helper.c | 106 +++++++++++++++++++++++++++++++++++++
9
target/arm/translate.c | 9 +--------
13
target/arm/translate-a64.c | 16 +++++-
10
3 files changed, 37 insertions(+), 8 deletions(-)
14
3 files changed, 122 insertions(+), 1 deletion(-)
11
15
12
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
13
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/neon-shared.decode
18
--- a/target/arm/helper-a64.h
15
+++ b/target/arm/neon-shared.decode
19
+++ b/target/arm/helper-a64.h
16
@@ -XXX,XX +XXX,XX @@ VCMLA 1111 110 rot:2 . 1 size:1 .... .... 1000 . q:1 . 0 .... \
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
17
21
18
VCADD 1111 110 rot:1 1 . 0 size:1 .... .... 1000 . q:1 . 0 .... \
22
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
19
vm=%vm_dp vn=%vn_dp vd=%vd_dp
23
DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
20
+
24
+DEF_HELPER_FLAGS_3(mte_check_zva, TCG_CALL_NO_WG, i64, env, i32, i64)
21
+# VUDOT and VSDOT
25
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+VDOT 1111 110 00 . 10 .... .... 1101 . q:1 . u:1 .... \
26
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp
27
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
24
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
28
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
25
index XXXXXXX..XXXXXXX 100644
29
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/translate-neon.inc.c
30
--- a/target/arm/mte_helper.c
27
+++ b/target/arm/translate-neon.inc.c
31
+++ b/target/arm/mte_helper.c
28
@@ -XXX,XX +XXX,XX @@ static bool trans_VCADD(DisasContext *s, arg_VCADD *a)
32
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
29
tcg_temp_free_ptr(fpst);
33
{
30
return true;
34
return mte_checkN(env, desc, ptr, GETPC());
31
}
35
}
32
+
36
+
33
+static bool trans_VDOT(DisasContext *s, arg_VDOT *a)
37
+/*
38
+ * Perform an MTE checked access for DC_ZVA.
39
+ */
40
+uint64_t HELPER(mte_check_zva)(CPUARMState *env, uint32_t desc, uint64_t ptr)
34
+{
41
+{
35
+ int opr_sz;
42
+ uintptr_t ra = GETPC();
36
+ gen_helper_gvec_3 *fn_gvec;
43
+ int log2_dcz_bytes, log2_tag_bytes;
44
+ int mmu_idx, bit55;
45
+ intptr_t dcz_bytes, tag_bytes, i;
46
+ void *mem;
47
+ uint64_t ptr_tag, mem_tag, align_ptr;
37
+
48
+
38
+ if (!dc_isar_feature(aa32_dp, s)) {
49
+ bit55 = extract64(ptr, 55, 1);
39
+ return false;
50
+
51
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
52
+ if (unlikely(!tbi_check(desc, bit55))) {
53
+ return ptr;
40
+ }
54
+ }
41
+
55
+
42
+ /* UNDEF accesses to D16-D31 if they don't exist. */
56
+ ptr_tag = allocation_tag_from_addr(ptr);
43
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
57
+
44
+ ((a->vd | a->vn | a->vm) & 0x10)) {
58
+ if (tcma_check(desc, bit55, ptr_tag)) {
45
+ return false;
59
+ goto done;
46
+ }
60
+ }
47
+
61
+
48
+ if ((a->vn | a->vm | a->vd) & a->q) {
62
+ /*
49
+ return false;
63
+ * In arm_cpu_realizefn, we asserted that dcz > LOG2_TAG_GRANULE+1,
64
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway, to make
65
+ * sure that we can access one complete tag byte here.
66
+ */
67
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
68
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
69
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
70
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
71
+ align_ptr = ptr & -dcz_bytes;
72
+
73
+ /*
74
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
75
+ * the original pointer for an invalid page. But watchpoints require
76
+ * that we probe the actual space. So do both.
77
+ */
78
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
79
+ (void) probe_write(env, ptr, 1, mmu_idx, ra);
80
+ mem = allocation_tag_mem(env, mmu_idx, align_ptr, MMU_DATA_STORE,
81
+ dcz_bytes, MMU_DATA_LOAD, tag_bytes, ra);
82
+ if (!mem) {
83
+ goto done;
50
+ }
84
+ }
51
+
85
+
52
+ if (!vfp_access_check(s)) {
86
+ /*
53
+ return true;
87
+ * Unlike the reasoning for checkN, DC_ZVA is always aligned, and thus
88
+ * it is quite easy to perform all of the comparisons at once without
89
+ * any extra masking.
90
+ *
91
+ * The most common zva block size is 64; some of the thunderx cpus use
92
+ * a block size of 128. For user-only, aarch64_max_initfn will set the
93
+ * block size to 512. Fill out the other cases for future-proofing.
94
+ *
95
+ * In order to be able to find the first miscompare later, we want the
96
+ * tag bytes to be in little-endian order.
97
+ */
98
+ switch (log2_tag_bytes) {
99
+ case 0: /* zva_blocksize 32 */
100
+ mem_tag = *(uint8_t *)mem;
101
+ ptr_tag *= 0x11u;
102
+ break;
103
+ case 1: /* zva_blocksize 64 */
104
+ mem_tag = cpu_to_le16(*(uint16_t *)mem);
105
+ ptr_tag *= 0x1111u;
106
+ break;
107
+ case 2: /* zva_blocksize 128 */
108
+ mem_tag = cpu_to_le32(*(uint32_t *)mem);
109
+ ptr_tag *= 0x11111111u;
110
+ break;
111
+ case 3: /* zva_blocksize 256 */
112
+ mem_tag = cpu_to_le64(*(uint64_t *)mem);
113
+ ptr_tag *= 0x1111111111111111ull;
114
+ break;
115
+
116
+ default: /* zva_blocksize 512, 1024, 2048 */
117
+ ptr_tag *= 0x1111111111111111ull;
118
+ i = 0;
119
+ do {
120
+ mem_tag = cpu_to_le64(*(uint64_t *)(mem + i));
121
+ if (unlikely(mem_tag != ptr_tag)) {
122
+ goto fail;
123
+ }
124
+ i += 8;
125
+ align_ptr += 16 * TAG_GRANULE;
126
+ } while (i < tag_bytes);
127
+ goto done;
54
+ }
128
+ }
55
+
129
+
56
+ opr_sz = (1 + a->q) * 8;
130
+ if (likely(mem_tag == ptr_tag)) {
57
+ fn_gvec = a->u ? gen_helper_gvec_udot_b : gen_helper_gvec_sdot_b;
131
+ goto done;
58
+ tcg_gen_gvec_3_ool(vfp_reg_offset(1, a->vd),
132
+ }
59
+ vfp_reg_offset(1, a->vn),
133
+
60
+ vfp_reg_offset(1, a->vm),
134
+ fail:
61
+ opr_sz, opr_sz, 0, fn_gvec);
135
+ /* Locate the first nibble that differs. */
62
+ return true;
136
+ i = ctz64(mem_tag ^ ptr_tag) >> 4;
137
+ mte_check_fail(env, mmu_idx, align_ptr + i * TAG_GRANULE, ra);
138
+
139
+ done:
140
+ return useronly_clean_ptr(ptr);
63
+}
141
+}
64
diff --git a/target/arm/translate.c b/target/arm/translate.c
142
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
65
index XXXXXXX..XXXXXXX 100644
143
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/translate.c
144
--- a/target/arm/translate-a64.c
67
+++ b/target/arm/translate.c
145
+++ b/target/arm/translate-a64.c
68
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
146
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
69
bool is_long = false, q = extract32(insn, 6, 1);
147
return;
70
bool ptr_is_env = false;
148
case ARM_CP_DC_ZVA:
71
149
/* Writes clear the aligned block of memory which rt points into. */
72
- if ((insn & 0xfeb00f00) == 0xfc200d00) {
150
- tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
73
- /* V[US]DOT -- 1111 1100 0.10 .... .... 1101 .Q.U .... */
151
+ if (s->mte_active[0]) {
74
- bool u = extract32(insn, 4, 1);
152
+ TCGv_i32 t_desc;
75
- if (!dc_isar_feature(aa32_dp, s)) {
153
+ int desc = 0;
76
- return 1;
154
+
77
- }
155
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
78
- fn_gvec = u ? gen_helper_gvec_udot_b : gen_helper_gvec_sdot_b;
156
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
79
- } else if ((insn & 0xff300f10) == 0xfc200810) {
157
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
80
+ if ((insn & 0xff300f10) == 0xfc200810) {
158
+ t_desc = tcg_const_i32(desc);
81
/* VFM[AS]L -- 1111 1100 S.10 .... .... 1000 .Q.1 .... */
159
+
82
int is_s = extract32(insn, 23, 1);
160
+ tcg_rt = new_tmp_a64(s);
83
if (!dc_isar_feature(aa32_fhm, s)) {
161
+ gen_helper_mte_check_zva(tcg_rt, cpu_env, t_desc, cpu_reg(s, rt));
162
+ tcg_temp_free_i32(t_desc);
163
+ } else {
164
+ tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
165
+ }
166
gen_helper_dc_zva(cpu_env, tcg_rt);
167
return;
168
default:
84
--
169
--
85
2.20.1
170
2.20.1
86
171
87
172
diff view generated by jsdifflib
1
Convert the Neon logic ops in the 3-reg-same grouping to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
Note that for the logic ops the 'size' field forms part of their
3
decode and the actual operations are always bitwise.
4
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-29-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200430181003.21682-16-peter.maydell@linaro.org
8
---
7
---
9
target/arm/neon-dp.decode | 12 +++++++++++
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
10
target/arm/translate-neon.inc.c | 19 +++++++++++++++++
9
1 file changed, 33 insertions(+), 28 deletions(-)
11
target/arm/translate.c | 38 +--------------------------------
12
3 files changed, 32 insertions(+), 37 deletions(-)
13
10
14
diff --git a/target/arm/neon-dp.decode b/target/arm/neon-dp.decode
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
15
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/neon-dp.decode
13
--- a/target/arm/translate-sve.c
17
+++ b/target/arm/neon-dp.decode
14
+++ b/target/arm/translate-sve.c
18
@@ -XXX,XX +XXX,XX @@
15
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
19
@3same .... ... . . . size:2 .... .... .... . q:1 . . .... \
16
int len_remain = len % 8;
20
&3same vm=%vm_dp vn=%vn_dp vd=%vd_dp
17
int nparts = len / 8 + ctpop8(len_remain);
21
18
int midx = get_mem_index(s);
22
+@3same_logic .... ... . . . .. .... .... .... . q:1 .. .... \
19
- TCGv_i64 addr, t0, t1;
23
+ &3same vm=%vm_dp vn=%vn_dp vd=%vd_dp size=0
20
+ TCGv_i64 dirty_addr, clean_addr, t0, t1;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
- /* Note that unpredicated load/store of vector/predicate registers
30
+ /*
31
+ * Note that unpredicated load/store of vector/predicate registers
32
* are defined as a stream of bytes, which equates to little-endian
33
- * operations on larger quantities. There is no nice way to force
34
- * a little-endian load for aarch64_be-linux-user out of line.
35
- *
36
+ * operations on larger quantities.
37
* Attempt to keep code expansion to a minimum by limiting the
38
* amount of unrolling done.
39
*/
40
if (nparts <= 4) {
41
int i;
42
43
+ t0 = tcg_temp_new_i64();
44
for (i = 0; i < len_align; i += 8) {
45
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
46
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
47
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
48
tcg_gen_st_i64(t0, cpu_env, vofs + i);
49
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
50
}
51
+ tcg_temp_free_i64(t0);
52
} else {
53
TCGLabel *loop = gen_new_label();
54
TCGv_ptr tp, i = tcg_const_local_ptr(0);
55
56
+ /* Copy the clean address into a local temp, live across the loop. */
57
+ t0 = clean_addr;
58
+ clean_addr = tcg_temp_local_new_i64();
59
+ tcg_gen_mov_i64(clean_addr, t0);
60
+ tcg_temp_free_i64(t0);
24
+
61
+
25
+VAND_3s 1111 001 0 0 . 00 .... .... 0001 ... 1 .... @3same_logic
62
gen_set_label(loop);
26
+VBIC_3s 1111 001 0 0 . 01 .... .... 0001 ... 1 .... @3same_logic
63
27
+VORR_3s 1111 001 0 0 . 10 .... .... 0001 ... 1 .... @3same_logic
64
- /* Minimize the number of local temps that must be re-read from
28
+VORN_3s 1111 001 0 0 . 11 .... .... 0001 ... 1 .... @3same_logic
65
- * the stack each iteration. Instead, re-compute values other
29
+VEOR_3s 1111 001 1 0 . 00 .... .... 0001 ... 1 .... @3same_logic
66
- * than the loop counter.
30
+VBSL_3s 1111 001 1 0 . 01 .... .... 0001 ... 1 .... @3same_logic
67
- */
31
+VBIT_3s 1111 001 1 0 . 10 .... .... 0001 ... 1 .... @3same_logic
68
+ t0 = tcg_temp_new_i64();
32
+VBIF_3s 1111 001 1 0 . 11 .... .... 0001 ... 1 .... @3same_logic
69
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
70
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
33
+
71
+
34
VADD_3s 1111 001 0 0 . .. .... .... 1000 . . . 0 .... @3same
72
tp = tcg_temp_new_ptr();
35
VSUB_3s 1111 001 1 0 . .. .... .... 1000 . . . 0 .... @3same
73
- tcg_gen_addi_ptr(tp, i, imm);
36
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
74
- tcg_gen_extu_ptr_i64(addr, tp);
37
index XXXXXXX..XXXXXXX 100644
75
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
38
--- a/target/arm/translate-neon.inc.c
39
+++ b/target/arm/translate-neon.inc.c
40
@@ -XXX,XX +XXX,XX @@ static bool do_3same(DisasContext *s, arg_3same *a, GVecGen3Fn fn)
41
42
DO_3SAME(VADD, tcg_gen_gvec_add)
43
DO_3SAME(VSUB, tcg_gen_gvec_sub)
44
+DO_3SAME(VAND, tcg_gen_gvec_and)
45
+DO_3SAME(VBIC, tcg_gen_gvec_andc)
46
+DO_3SAME(VORR, tcg_gen_gvec_or)
47
+DO_3SAME(VORN, tcg_gen_gvec_orc)
48
+DO_3SAME(VEOR, tcg_gen_gvec_xor)
49
+
50
+/* These insns are all gvec_bitsel but with the inputs in various orders. */
51
+#define DO_3SAME_BITSEL(INSN, O1, O2, O3) \
52
+ static void gen_##INSN##_3s(unsigned vece, uint32_t rd_ofs, \
53
+ uint32_t rn_ofs, uint32_t rm_ofs, \
54
+ uint32_t oprsz, uint32_t maxsz) \
55
+ { \
56
+ tcg_gen_gvec_bitsel(vece, rd_ofs, O1, O2, O3, oprsz, maxsz); \
57
+ } \
58
+ DO_3SAME(INSN, gen_##INSN##_3s)
59
+
60
+DO_3SAME_BITSEL(VBSL, rd_ofs, rn_ofs, rm_ofs)
61
+DO_3SAME_BITSEL(VBIT, rm_ofs, rn_ofs, rd_ofs)
62
+DO_3SAME_BITSEL(VBIF, rm_ofs, rd_ofs, rn_ofs)
63
diff --git a/target/arm/translate.c b/target/arm/translate.c
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/translate.c
66
+++ b/target/arm/translate.c
67
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
68
}
69
return 1;
70
71
- case NEON_3R_LOGIC: /* Logic ops. */
72
- switch ((u << 2) | size) {
73
- case 0: /* VAND */
74
- tcg_gen_gvec_and(0, rd_ofs, rn_ofs, rm_ofs,
75
- vec_size, vec_size);
76
- break;
77
- case 1: /* VBIC */
78
- tcg_gen_gvec_andc(0, rd_ofs, rn_ofs, rm_ofs,
79
- vec_size, vec_size);
80
- break;
81
- case 2: /* VORR */
82
- tcg_gen_gvec_or(0, rd_ofs, rn_ofs, rm_ofs,
83
- vec_size, vec_size);
84
- break;
85
- case 3: /* VORN */
86
- tcg_gen_gvec_orc(0, rd_ofs, rn_ofs, rm_ofs,
87
- vec_size, vec_size);
88
- break;
89
- case 4: /* VEOR */
90
- tcg_gen_gvec_xor(0, rd_ofs, rn_ofs, rm_ofs,
91
- vec_size, vec_size);
92
- break;
93
- case 5: /* VBSL */
94
- tcg_gen_gvec_bitsel(MO_8, rd_ofs, rd_ofs, rn_ofs, rm_ofs,
95
- vec_size, vec_size);
96
- break;
97
- case 6: /* VBIT */
98
- tcg_gen_gvec_bitsel(MO_8, rd_ofs, rm_ofs, rn_ofs, rd_ofs,
99
- vec_size, vec_size);
100
- break;
101
- case 7: /* VBIF */
102
- tcg_gen_gvec_bitsel(MO_8, rd_ofs, rm_ofs, rd_ofs, rn_ofs,
103
- vec_size, vec_size);
104
- break;
105
- }
106
- return 0;
107
-
76
-
108
case NEON_3R_VQADD:
77
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
109
tcg_gen_gvec_4(rd_ofs, offsetof(CPUARMState, vfp.qc),
78
-
110
rn_ofs, rm_ofs, vec_size, vec_size,
79
tcg_gen_add_ptr(tp, cpu_env, i);
111
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
80
tcg_gen_addi_ptr(i, i, 8);
112
return 0;
81
tcg_gen_st_i64(t0, tp, vofs);
113
82
tcg_temp_free_ptr(tp);
114
case NEON_3R_VADD_VSUB:
83
+ tcg_temp_free_i64(t0);
115
+ case NEON_3R_LOGIC:
84
116
/* Already handled by decodetree */
85
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
117
return 1;
86
tcg_temp_free_ptr(i);
87
}
88
89
- /* Predicate register loads can be any multiple of 2.
90
+ /*
91
+ * Predicate register loads can be any multiple of 2.
92
* Note that we still store the entire 64-bit unit into cpu_env.
93
*/
94
if (len_remain) {
95
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
96
-
97
+ t0 = tcg_temp_new_i64();
98
switch (len_remain) {
99
case 2:
100
case 4:
101
case 8:
102
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
103
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx,
104
+ MO_LE | ctz32(len_remain));
105
break;
106
107
case 6:
108
t1 = tcg_temp_new_i64();
109
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL);
110
- tcg_gen_addi_i64(addr, addr, 4);
111
- tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW);
112
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUL);
113
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
114
+ tcg_gen_qemu_ld_i64(t1, clean_addr, midx, MO_LEUW);
115
tcg_gen_deposit_i64(t0, t0, t1, 32, 32);
116
tcg_temp_free_i64(t1);
117
break;
118
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
119
g_assert_not_reached();
118
}
120
}
121
tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
122
+ tcg_temp_free_i64(t0);
123
}
124
- tcg_temp_free_i64(addr);
125
- tcg_temp_free_i64(t0);
126
+ tcg_temp_free_i64(clean_addr);
127
}
128
129
/* Similarly for stores. */
119
--
130
--
120
2.20.1
131
2.20.1
121
132
122
133
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Embed the APUs into the SoC type.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-30-richard.henderson@linaro.org
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
10
Message-id: 20200427181649.26851-8-edgar.iglesias@gmail.com
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
7
---
13
include/hw/arm/xlnx-versal.h | 2 +-
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
14
hw/arm/xlnx-versal-virt.c | 4 ++--
9
1 file changed, 33 insertions(+), 28 deletions(-)
15
hw/arm/xlnx-versal.c | 19 +++++--------------
16
3 files changed, 8 insertions(+), 17 deletions(-)
17
10
18
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
19
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/arm/xlnx-versal.h
13
--- a/target/arm/translate-sve.c
21
+++ b/include/hw/arm/xlnx-versal.h
14
+++ b/target/arm/translate-sve.c
22
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
15
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
23
struct {
16
int len_remain = len % 8;
24
struct {
17
int nparts = len / 8 + ctpop8(len_remain);
25
MemoryRegion mr;
18
int midx = get_mem_index(s);
26
- ARMCPU *cpu[XLNX_VERSAL_NR_ACPUS];
19
- TCGv_i64 addr, t0;
27
+ ARMCPU cpu[XLNX_VERSAL_NR_ACPUS];
20
+ TCGv_i64 dirty_addr, clean_addr, t0;
28
GICv3State gic;
21
29
} apu;
22
- addr = tcg_temp_new_i64();
30
} fpd;
23
- t0 = tcg_temp_new_i64();
31
diff --git a/hw/arm/xlnx-versal-virt.c b/hw/arm/xlnx-versal-virt.c
24
+ dirty_addr = tcg_temp_new_i64();
32
index XXXXXXX..XXXXXXX 100644
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
33
--- a/hw/arm/xlnx-versal-virt.c
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
34
+++ b/hw/arm/xlnx-versal-virt.c
27
+ tcg_temp_free_i64(dirty_addr);
35
@@ -XXX,XX +XXX,XX @@ static void versal_virt_init(MachineState *machine)
28
36
s->binfo.get_dtb = versal_virt_get_dtb;
29
/* Note that unpredicated load/store of vector/predicate registers
37
s->binfo.modify_dtb = versal_virt_modify_dtb;
30
* are defined as a stream of bytes, which equates to little-endian
38
if (machine->kernel_filename) {
31
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
39
- arm_load_kernel(s->soc.fpd.apu.cpu[0], machine, &s->binfo);
32
if (nparts <= 4) {
40
+ arm_load_kernel(&s->soc.fpd.apu.cpu[0], machine, &s->binfo);
33
int i;
34
35
+ t0 = tcg_temp_new_i64();
36
for (i = 0; i < len_align; i += 8) {
37
tcg_gen_ld_i64(t0, cpu_env, vofs + i);
38
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
39
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
40
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
41
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
42
}
43
+ tcg_temp_free_i64(t0);
41
} else {
44
} else {
42
- AddressSpace *as = arm_boot_address_space(s->soc.fpd.apu.cpu[0],
45
TCGLabel *loop = gen_new_label();
43
+ AddressSpace *as = arm_boot_address_space(&s->soc.fpd.apu.cpu[0],
46
- TCGv_ptr t2, i = tcg_const_local_ptr(0);
44
&s->binfo);
47
+ TCGv_ptr tp, i = tcg_const_local_ptr(0);
45
/* Some boot-loaders (e.g u-boot) don't like blobs at address 0 (NULL).
48
+
46
* Offset things by 4K. */
49
+ /* Copy the clean address into a local temp, live across the loop. */
47
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
50
+ t0 = clean_addr;
48
index XXXXXXX..XXXXXXX 100644
51
+ clean_addr = tcg_temp_local_new_i64();
49
--- a/hw/arm/xlnx-versal.c
52
+ tcg_gen_mov_i64(clean_addr, t0);
50
+++ b/hw/arm/xlnx-versal.c
53
+ tcg_temp_free_i64(t0);
51
@@ -XXX,XX +XXX,XX @@ static void versal_create_apu_cpus(Versal *s)
54
52
55
gen_set_label(loop);
53
for (i = 0; i < ARRAY_SIZE(s->fpd.apu.cpu); i++) {
56
54
Object *obj;
57
- t2 = tcg_temp_new_ptr();
55
- char *name;
58
- tcg_gen_add_ptr(t2, cpu_env, i);
59
- tcg_gen_ld_i64(t0, t2, vofs);
56
-
60
-
57
- obj = object_new(XLNX_VERSAL_ACPU_TYPE);
61
- /* Minimize the number of local temps that must be re-read from
58
- if (!obj) {
62
- * the stack each iteration. Instead, re-compute values other
59
- error_report("Unable to create apu.cpu[%d] of type %s",
63
- * than the loop counter.
60
- i, XLNX_VERSAL_ACPU_TYPE);
64
- */
61
- exit(EXIT_FAILURE);
65
- tcg_gen_addi_ptr(t2, i, imm);
62
- }
66
- tcg_gen_extu_ptr_i64(addr, t2);
67
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
68
- tcg_temp_free_ptr(t2);
63
-
69
-
64
- name = g_strdup_printf("apu-cpu[%d]", i);
70
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
65
- object_property_add_child(OBJECT(s), name, obj, &error_fatal);
71
-
66
- g_free(name);
72
+ t0 = tcg_temp_new_i64();
67
73
+ tp = tcg_temp_new_ptr();
68
+ object_initialize_child(OBJECT(s), "apu-cpu[*]",
74
+ tcg_gen_add_ptr(tp, cpu_env, i);
69
+ &s->fpd.apu.cpu[i], sizeof(s->fpd.apu.cpu[i]),
75
+ tcg_gen_ld_i64(t0, tp, vofs);
70
+ XLNX_VERSAL_ACPU_TYPE, &error_abort, NULL);
76
tcg_gen_addi_ptr(i, i, 8);
71
+ obj = OBJECT(&s->fpd.apu.cpu[i]);
77
+ tcg_temp_free_ptr(tp);
72
object_property_set_int(obj, s->cfg.psci_conduit,
78
+
73
"psci-conduit", &error_abort);
79
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
74
if (i) {
80
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
75
@@ -XXX,XX +XXX,XX @@ static void versal_create_apu_cpus(Versal *s)
81
+ tcg_temp_free_i64(t0);
76
object_property_set_link(obj, OBJECT(&s->fpd.apu.mr), "memory",
82
77
&error_abort);
83
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
78
object_property_set_bool(obj, true, "realized", &error_fatal);
84
tcg_temp_free_ptr(i);
79
- s->fpd.apu.cpu[i] = ARM_CPU(obj);
85
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
86
87
/* Predicate register stores can be any multiple of 2. */
88
if (len_remain) {
89
+ t0 = tcg_temp_new_i64();
90
tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
91
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
92
93
switch (len_remain) {
94
case 2:
95
case 4:
96
case 8:
97
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
98
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx,
99
+ MO_LE | ctz32(len_remain));
100
break;
101
102
case 6:
103
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUL);
104
- tcg_gen_addi_i64(addr, addr, 4);
105
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUL);
106
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
107
tcg_gen_shri_i64(t0, t0, 32);
108
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUW);
109
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUW);
110
break;
111
112
default:
113
g_assert_not_reached();
114
}
115
+ tcg_temp_free_i64(t0);
80
}
116
}
117
- tcg_temp_free_i64(addr);
118
- tcg_temp_free_i64(t0);
119
+ tcg_temp_free_i64(clean_addr);
81
}
120
}
82
121
83
@@ -XXX,XX +XXX,XX @@ static void versal_create_apu_gic(Versal *s, qemu_irq *pic)
122
static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
84
}
85
86
for (i = 0; i < nr_apu_cpus; i++) {
87
- DeviceState *cpudev = DEVICE(s->fpd.apu.cpu[i]);
88
+ DeviceState *cpudev = DEVICE(&s->fpd.apu.cpu[i]);
89
int ppibase = XLNX_VERSAL_NR_IRQS + i * GIC_INTERNAL + GIC_NR_SGIS;
90
qemu_irq maint_irq;
91
int ti;
92
--
123
--
93
2.20.1
124
2.20.1
94
125
95
126
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move misplaced comment.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Message-id: 20200626033144.790098-31-richard.henderson@linaro.org
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
9
Message-id: 20200427181649.26851-3-edgar.iglesias@gmail.com
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
7
---
12
hw/arm/xlnx-versal.c | 2 +-
8
target/arm/translate-sve.c | 6 ++++--
13
1 file changed, 1 insertion(+), 1 deletion(-)
9
1 file changed, 4 insertions(+), 2 deletions(-)
14
10
15
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
16
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/arm/xlnx-versal.c
13
--- a/target/arm/translate-sve.c
18
+++ b/hw/arm/xlnx-versal.c
14
+++ b/target/arm/translate-sve.c
19
@@ -XXX,XX +XXX,XX @@ static void versal_create_apu_cpus(Versal *s)
15
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
20
16
unsigned esz = dtype_esz[a->dtype];
21
obj = object_new(XLNX_VERSAL_ACPU_TYPE);
17
unsigned msz = dtype_msz(a->dtype);
22
if (!obj) {
18
TCGLabel *over = gen_new_label();
23
- /* Secondary CPUs start in PSCI powered-down state */
19
- TCGv_i64 temp;
24
error_report("Unable to create apu.cpu[%d] of type %s",
20
+ TCGv_i64 temp, clean_addr;
25
i, XLNX_VERSAL_ACPU_TYPE);
21
26
exit(EXIT_FAILURE);
22
/* If the guarding predicate has no bits set, no load occurs. */
27
@@ -XXX,XX +XXX,XX @@ static void versal_create_apu_cpus(Versal *s)
23
if (psz <= 8) {
28
object_property_set_int(obj, s->cfg.psci_conduit,
24
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
29
"psci-conduit", &error_abort);
25
/* Load the data. */
30
if (i) {
26
temp = tcg_temp_new_i64();
31
+ /* Secondary CPUs start in PSCI powered-down state */
27
tcg_gen_addi_i64(temp, cpu_reg_sp(s, a->rn), a->imm << msz);
32
object_property_set_bool(obj, true,
28
- tcg_gen_qemu_ld_i64(temp, temp, get_mem_index(s),
33
"start-powered-off", &error_abort);
29
+ clean_addr = gen_mte_check1(s, temp, false, true, msz);
34
}
30
+
31
+ tcg_gen_qemu_ld_i64(temp, clean_addr, get_mem_index(s),
32
s->be_data | dtype_mop[a->dtype]);
33
34
/* Broadcast to *all* elements. */
35
--
35
--
36
2.20.1
36
2.20.1
37
37
38
38
diff view generated by jsdifflib
1
Convert the VCADD (vector) insns to decodetree.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move the variable declarations to the top of the function,
4
but do not create a new label before sve_access_check.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-32-richard.henderson@linaro.org
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200430181003.21682-6-peter.maydell@linaro.org
6
---
10
---
7
target/arm/neon-shared.decode | 3 +++
11
target/arm/translate-sve.c | 12 +++++++-----
8
target/arm/translate-neon.inc.c | 37 +++++++++++++++++++++++++++++++++
12
1 file changed, 7 insertions(+), 5 deletions(-)
9
target/arm/translate.c | 11 +---------
10
3 files changed, 41 insertions(+), 10 deletions(-)
11
13
12
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
14
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/neon-shared.decode
16
--- a/target/arm/translate-sve.c
15
+++ b/target/arm/neon-shared.decode
17
+++ b/target/arm/translate-sve.c
16
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RQ_zpri(DisasContext *s, arg_rpri_load *a)
17
19
/* Load and broadcast element. */
18
VCMLA 1111 110 rot:2 . 1 size:1 .... .... 1000 . q:1 . 0 .... \
20
static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
19
vm=%vm_dp vn=%vn_dp vd=%vd_dp
21
{
20
+
22
- if (!sve_access_check(s)) {
21
+VCADD 1111 110 rot:1 1 . 0 size:1 .... .... 1000 . q:1 . 0 .... \
23
- return true;
22
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp
24
- }
23
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
25
-
24
index XXXXXXX..XXXXXXX 100644
26
unsigned vsz = vec_full_reg_size(s);
25
--- a/target/arm/translate-neon.inc.c
27
unsigned psz = pred_full_reg_size(s);
26
+++ b/target/arm/translate-neon.inc.c
28
unsigned esz = dtype_esz[a->dtype];
27
@@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA(DisasContext *s, arg_VCMLA *a)
29
unsigned msz = dtype_msz(a->dtype);
28
tcg_temp_free_ptr(fpst);
30
- TCGLabel *over = gen_new_label();
29
return true;
31
+ TCGLabel *over;
30
}
32
TCGv_i64 temp, clean_addr;
31
+
33
32
+static bool trans_VCADD(DisasContext *s, arg_VCADD *a)
34
+ if (!sve_access_check(s)) {
33
+{
34
+ int opr_sz;
35
+ TCGv_ptr fpst;
36
+ gen_helper_gvec_3_ptr *fn_gvec_ptr;
37
+
38
+ if (!dc_isar_feature(aa32_vcma, s)
39
+ || (!a->size && !dc_isar_feature(aa32_fp16_arith, s))) {
40
+ return false;
41
+ }
42
+
43
+ /* UNDEF accesses to D16-D31 if they don't exist. */
44
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
45
+ ((a->vd | a->vn | a->vm) & 0x10)) {
46
+ return false;
47
+ }
48
+
49
+ if ((a->vn | a->vm | a->vd) & a->q) {
50
+ return false;
51
+ }
52
+
53
+ if (!vfp_access_check(s)) {
54
+ return true;
35
+ return true;
55
+ }
36
+ }
56
+
37
+
57
+ opr_sz = (1 + a->q) * 8;
38
+ over = gen_new_label();
58
+ fpst = get_fpstatus_ptr(1);
39
+
59
+ fn_gvec_ptr = a->size ? gen_helper_gvec_fcadds : gen_helper_gvec_fcaddh;
40
/* If the guarding predicate has no bits set, no load occurs. */
60
+ tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd),
41
if (psz <= 8) {
61
+ vfp_reg_offset(1, a->vn),
42
/* Reduce the pred_esz_masks value simply to reduce the
62
+ vfp_reg_offset(1, a->vm),
63
+ fpst, opr_sz, opr_sz, a->rot,
64
+ fn_gvec_ptr);
65
+ tcg_temp_free_ptr(fpst);
66
+ return true;
67
+}
68
diff --git a/target/arm/translate.c b/target/arm/translate.c
69
index XXXXXXX..XXXXXXX 100644
70
--- a/target/arm/translate.c
71
+++ b/target/arm/translate.c
72
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
73
bool is_long = false, q = extract32(insn, 6, 1);
74
bool ptr_is_env = false;
75
76
- if ((insn & 0xfea00f10) == 0xfc800800) {
77
- /* VCADD -- 1111 110R 1.0S .... .... 1000 ...0 .... */
78
- int size = extract32(insn, 20, 1);
79
- data = extract32(insn, 24, 1); /* rot */
80
- if (!dc_isar_feature(aa32_vcma, s)
81
- || (!size && !dc_isar_feature(aa32_fp16_arith, s))) {
82
- return 1;
83
- }
84
- fn_gvec_ptr = size ? gen_helper_gvec_fcadds : gen_helper_gvec_fcaddh;
85
- } else if ((insn & 0xfeb00f00) == 0xfc200d00) {
86
+ if ((insn & 0xfeb00f00) == 0xfc200d00) {
87
/* V[US]DOT -- 1111 1100 0.10 .... .... 1101 .Q.U .... */
88
bool u = extract32(insn, 4, 1);
89
if (!dc_isar_feature(aa32_dp, s)) {
90
--
43
--
91
2.20.1
44
2.20.1
92
45
93
46
diff view generated by jsdifflib
1
The ARMv8.2-TTS2UXN feature extends the XN field in stage 2
1
From: Richard Henderson <richard.henderson@linaro.org>
2
translation table descriptors from just bit [54] to bits [54:53],
3
allowing stage 2 to control execution permissions separately for EL0
4
and EL1. Implement the new semantics of the XN field and enable
5
the feature for our 'max' CPU.
6
2
3
Introduce an lvalue macro to wrap target_tlb_bit0.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-33-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20200330210400.11724-5-peter.maydell@linaro.org
11
---
9
---
12
target/arm/cpu.h | 15 +++++++++++++++
10
target/arm/cpu.h | 13 +++++++++++++
13
target/arm/cpu.c | 1 +
11
target/arm/helper.c | 2 +-
14
target/arm/cpu64.c | 2 ++
12
target/arm/translate-a64.c | 2 +-
15
target/arm/helper.c | 37 +++++++++++++++++++++++++++++++------
13
3 files changed, 15 insertions(+), 2 deletions(-)
16
4 files changed, 49 insertions(+), 6 deletions(-)
17
14
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
19
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
21
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
22
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa32_ccidx(const ARMISARegisters *id)
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
23
return FIELD_EX32(id->id_mmfr4, ID_MMFR4, CCIDX) != 0;
20
/* Shared between translate-sve.c and sve_helper.c. */
24
}
21
extern const uint64_t pred_esz_masks[4];
25
22
26
+static inline bool isar_feature_aa32_tts2uxn(const ARMISARegisters *id)
23
+/* Helper for the macros below, validating the argument type. */
24
+static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
27
+{
25
+{
28
+ return FIELD_EX32(id->id_mmfr4, ID_MMFR4, XNX) != 0;
26
+ return x;
29
+}
27
+}
30
+
28
+
31
/*
29
+/*
32
* 64-bit feature tests via id registers.
30
+ * Lvalue macros for ARM TLB bits that we must cache in the TCG TLB.
33
*/
31
+ * Using these should be a bit more self-documenting than using the
34
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_ccidx(const ARMISARegisters *id)
32
+ * generic target bits directly.
35
return FIELD_EX64(id->id_aa64mmfr2, ID_AA64MMFR2, CCIDX) != 0;
33
+ */
36
}
34
+#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
37
38
+static inline bool isar_feature_aa64_tts2uxn(const ARMISARegisters *id)
39
+{
40
+ return FIELD_EX64(id->id_aa64mmfr1, ID_AA64MMFR1, XNX) != 0;
41
+}
42
+
35
+
43
/*
36
/*
44
* Feature tests for "does this exist in either 32-bit or 64-bit?"
37
* Naming convention for isar_feature functions:
45
*/
38
* Functions which test 32-bit ID registers should have _aa32_ in
46
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_any_ccidx(const ARMISARegisters *id)
47
return isar_feature_aa64_ccidx(id) || isar_feature_aa32_ccidx(id);
48
}
49
50
+static inline bool isar_feature_any_tts2uxn(const ARMISARegisters *id)
51
+{
52
+ return isar_feature_aa64_tts2uxn(id) || isar_feature_aa32_tts2uxn(id);
53
+}
54
+
55
/*
56
* Forward to the above feature tests given an ARMCPU pointer.
57
*/
58
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
59
index XXXXXXX..XXXXXXX 100644
60
--- a/target/arm/cpu.c
61
+++ b/target/arm/cpu.c
62
@@ -XXX,XX +XXX,XX @@ static void arm_max_initfn(Object *obj)
63
t = FIELD_DP32(t, ID_MMFR4, HPDS, 1); /* AA32HPD */
64
t = FIELD_DP32(t, ID_MMFR4, AC2, 1); /* ACTLR2, HACTLR2 */
65
t = FIELD_DP32(t, ID_MMFR4, CNP, 1); /* TTCNP */
66
+ t = FIELD_DP32(t, ID_MMFR4, XNX, 1); /* TTS2UXN */
67
cpu->isar.id_mmfr4 = t;
68
}
69
#endif
70
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
71
index XXXXXXX..XXXXXXX 100644
72
--- a/target/arm/cpu64.c
73
+++ b/target/arm/cpu64.c
74
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
75
t = FIELD_DP64(t, ID_AA64MMFR1, VH, 1);
76
t = FIELD_DP64(t, ID_AA64MMFR1, PAN, 2); /* ATS1E1 */
77
t = FIELD_DP64(t, ID_AA64MMFR1, VMIDBITS, 2); /* VMID16 */
78
+ t = FIELD_DP64(t, ID_AA64MMFR1, XNX, 1); /* TTS2UXN */
79
cpu->isar.id_aa64mmfr1 = t;
80
81
t = cpu->isar.id_aa64mmfr2;
82
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
83
u = FIELD_DP32(u, ID_MMFR4, HPDS, 1); /* AA32HPD */
84
u = FIELD_DP32(u, ID_MMFR4, AC2, 1); /* ACTLR2, HACTLR2 */
85
u = FIELD_DP32(u, ID_MMFR4, CNP, 1); /* TTCNP */
86
+ u = FIELD_DP32(u, ID_MMFR4, XNX, 1); /* TTS2UXN */
87
cpu->isar.id_mmfr4 = u;
88
89
u = cpu->isar.id_aa64dfr0;
90
diff --git a/target/arm/helper.c b/target/arm/helper.c
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
91
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
92
--- a/target/arm/helper.c
41
--- a/target/arm/helper.c
93
+++ b/target/arm/helper.c
42
+++ b/target/arm/helper.c
94
@@ -XXX,XX +XXX,XX @@ simple_ap_to_rw_prot(CPUARMState *env, ARMMMUIdx mmu_idx, int ap)
95
*
96
* @env: CPUARMState
97
* @s2ap: The 2-bit stage2 access permissions (S2AP)
98
- * @xn: XN (execute-never) bit
99
+ * @xn: XN (execute-never) bits
100
+ * @s1_is_el0: true if this is S2 of an S1+2 walk for EL0
101
*/
102
-static int get_S2prot(CPUARMState *env, int s2ap, int xn)
103
+static int get_S2prot(CPUARMState *env, int s2ap, int xn, bool s1_is_el0)
104
{
105
int prot = 0;
106
107
@@ -XXX,XX +XXX,XX @@ static int get_S2prot(CPUARMState *env, int s2ap, int xn)
108
if (s2ap & 2) {
109
prot |= PAGE_WRITE;
110
}
111
- if (!xn) {
112
- if (arm_el_is_aa64(env, 2) || prot & PAGE_READ) {
113
+
114
+ if (cpu_isar_feature(any_tts2uxn, env_archcpu(env))) {
115
+ switch (xn) {
116
+ case 0:
117
prot |= PAGE_EXEC;
118
+ break;
119
+ case 1:
120
+ if (s1_is_el0) {
121
+ prot |= PAGE_EXEC;
122
+ }
123
+ break;
124
+ case 2:
125
+ break;
126
+ case 3:
127
+ if (!s1_is_el0) {
128
+ prot |= PAGE_EXEC;
129
+ }
130
+ break;
131
+ default:
132
+ g_assert_not_reached();
133
+ }
134
+ } else {
135
+ if (!extract32(xn, 1, 1)) {
136
+ if (arm_el_is_aa64(env, 2) || prot & PAGE_READ) {
137
+ prot |= PAGE_EXEC;
138
+ }
139
}
140
}
141
return prot;
142
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
43
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
143
}
44
}
144
45
/* When in aarch64 mode, and BTI is enabled, remember GP in the IOTLB. */
145
ap = extract32(attrs, 4, 2);
46
if (aarch64 && guarded && cpu_isar_feature(aa64_bti, cpu)) {
146
- xn = extract32(attrs, 12, 1);
47
- txattrs->target_tlb_bit0 = true;
147
48
+ arm_tlb_bti_gp(txattrs) = true;
148
if (mmu_idx == ARMMMUIdx_Stage2) {
149
ns = true;
150
- *prot = get_S2prot(env, ap, xn);
151
+ xn = extract32(attrs, 11, 2);
152
+ *prot = get_S2prot(env, ap, xn, s1_is_el0);
153
} else {
154
ns = extract32(attrs, 3, 1);
155
+ xn = extract32(attrs, 12, 1);
156
pxn = extract32(attrs, 11, 1);
157
*prot = get_S1prot(env, mmu_idx, aarch64, ap, ns, xn, pxn);
158
}
49
}
50
51
if (cacheattrs != NULL) {
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
53
index XXXXXXX..XXXXXXX 100644
54
--- a/target/arm/translate-a64.c
55
+++ b/target/arm/translate-a64.c
56
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
57
* table entry even for that case.
58
*/
59
return (tlb_hit(entry->addr_code, addr) &&
60
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs.target_tlb_bit0);
61
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
62
#endif
63
}
64
159
--
65
--
160
2.20.1
66
2.20.1
161
67
162
68
diff view generated by jsdifflib
1
Convert the VCMLA (vector) insns in the 3same extension group to
1
From: Richard Henderson <richard.henderson@linaro.org>
2
decodetree.
3
2
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-34-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-5-peter.maydell@linaro.org
7
---
10
---
8
target/arm/neon-shared.decode | 11 ++++++++++
11
target/arm/cpu.h | 1 +
9
target/arm/translate-neon.inc.c | 37 +++++++++++++++++++++++++++++++++
12
target/arm/helper-sve.h | 58 ++++++++++
10
target/arm/translate.c | 11 +---------
13
target/arm/internals.h | 6 +
11
3 files changed, 49 insertions(+), 10 deletions(-)
14
target/arm/sve_helper.c | 218 ++++++++++++++++++++++++++++++-------
15
target/arm/translate-sve.c | 186 ++++++++++++++++++++++---------
16
5 files changed, 378 insertions(+), 91 deletions(-)
12
17
13
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/neon-shared.decode
20
--- a/target/arm/cpu.h
16
+++ b/target/arm/neon-shared.decode
21
+++ b/target/arm/cpu.h
17
@@ -XXX,XX +XXX,XX @@
22
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
18
# More specifically, this covers:
23
* generic target bits directly.
19
# 2reg scalar ext: 0b1111_1110_xxxx_xxxx_xxxx_1x0x_xxxx_xxxx
24
*/
20
# 3same ext: 0b1111_110x_xxxx_xxxx_xxxx_1x0x_xxxx_xxxx
25
#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
21
+
26
+#define arm_tlb_mte_tagged(x) (typecheck_memtxattrs(x)->target_tlb_bit1)
22
+# VFP/Neon register fields; same as vfp.decode
27
23
+%vm_dp 5:1 0:4
28
/*
24
+%vm_sp 0:4 5:1
29
* Naming convention for isar_feature functions:
25
+%vn_dp 7:1 16:4
30
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
26
+%vn_sp 16:4 7:1
27
+%vd_dp 22:1 12:4
28
+%vd_sp 12:4 22:1
29
+
30
+VCMLA 1111 110 rot:2 . 1 size:1 .... .... 1000 . q:1 . 0 .... \
31
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp
32
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
33
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
34
--- a/target/arm/translate-neon.inc.c
32
--- a/target/arm/helper-sve.h
35
+++ b/target/arm/translate-neon.inc.c
33
+++ b/target/arm/helper-sve.h
36
@@ -XXX,XX +XXX,XX @@
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ld1sds_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
#include "decode-neon-dp.inc.c"
35
DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
#include "decode-neon-ls.inc.c"
36
DEF_HELPER_FLAGS_4(sve_ld1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
#include "decode-neon-shared.inc.c"
37
40
+
38
+DEF_HELPER_FLAGS_4(sve_ld1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+static bool trans_VCMLA(DisasContext *s, arg_VCMLA *a)
39
+DEF_HELPER_FLAGS_4(sve_ld2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ld3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_ld4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ld1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_4(sve_ld2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ld3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_ld4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+
48
+DEF_HELPER_FLAGS_4(sve_ld1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ld2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_ld3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ld4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+
53
+DEF_HELPER_FLAGS_4(sve_ld1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_4(sve_ld2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_ld3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ld4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+
58
+DEF_HELPER_FLAGS_4(sve_ld1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
59
+DEF_HELPER_FLAGS_4(sve_ld2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_ld3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_ld4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_ld1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_ld2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ld3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_ld4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ld1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_ld2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ld3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_4(sve_ld4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_4(sve_ld1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_4(sve_ld1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_4(sve_ld1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_4(sve_ld1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_4(sve_ld1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_4(sve_ld1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
+
80
+DEF_HELPER_FLAGS_4(sve_ld1hsu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ld1hdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ld1hss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ld1hds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+
85
+DEF_HELPER_FLAGS_4(sve_ld1hsu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ld1hdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_4(sve_ld1hss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_4(sve_ld1hds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
89
+
90
+DEF_HELPER_FLAGS_4(sve_ld1sdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
91
+DEF_HELPER_FLAGS_4(sve_ld1sds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
92
+
93
+DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ld1sds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
95
+
96
DEF_HELPER_FLAGS_4(sve_ldff1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
97
DEF_HELPER_FLAGS_4(sve_ldff1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
98
DEF_HELPER_FLAGS_4(sve_ldff1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
99
diff --git a/target/arm/internals.h b/target/arm/internals.h
100
index XXXXXXX..XXXXXXX 100644
101
--- a/target/arm/internals.h
102
+++ b/target/arm/internals.h
103
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
104
#define LOG2_TAG_GRANULE 4
105
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
106
107
+/*
108
+ * The SVE simd_data field, for memory ops, contains either
109
+ * rd (5 bits) or a shift count (2 bits).
110
+ */
111
+#define SVE_MTEDESC_SHIFT 5
112
+
113
/* Bits within a descriptor passed to the helper_mte_check* functions. */
114
FIELD(MTEDESC, MIDX, 0, 4)
115
FIELD(MTEDESC, TBI, 4, 2)
116
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
117
index XXXXXXX..XXXXXXX 100644
118
--- a/target/arm/sve_helper.c
119
+++ b/target/arm/sve_helper.c
120
@@ -XXX,XX +XXX,XX @@ static void sve_cont_ldst_watchpoints(SVEContLdSt *info, CPUARMState *env,
121
#endif
122
}
123
124
+typedef uint64_t mte_check_fn(CPUARMState *, uint32_t, uint64_t, uintptr_t);
125
+
126
+static inline QEMU_ALWAYS_INLINE
127
+void sve_cont_ldst_mte_check_int(SVEContLdSt *info, CPUARMState *env,
128
+ uint64_t *vg, target_ulong addr, int esize,
129
+ int msize, uint32_t mtedesc, uintptr_t ra,
130
+ mte_check_fn *check)
42
+{
131
+{
43
+ int opr_sz;
132
+ intptr_t mem_off, reg_off, reg_last;
44
+ TCGv_ptr fpst;
133
+
45
+ gen_helper_gvec_3_ptr *fn_gvec_ptr;
134
+ /* Process the page only if MemAttr == Tagged. */
46
+
135
+ if (arm_tlb_mte_tagged(&info->page[0].attrs)) {
47
+ if (!dc_isar_feature(aa32_vcma, s)
136
+ mem_off = info->mem_off_first[0];
48
+ || (!a->size && !dc_isar_feature(aa32_fp16_arith, s))) {
137
+ reg_off = info->reg_off_first[0];
49
+ return false;
138
+ reg_last = info->reg_off_split;
139
+ if (reg_last < 0) {
140
+ reg_last = info->reg_off_last[0];
141
+ }
142
+
143
+ do {
144
+ uint64_t pg = vg[reg_off >> 6];
145
+ do {
146
+ if ((pg >> (reg_off & 63)) & 1) {
147
+ check(env, mtedesc, addr, ra);
148
+ }
149
+ reg_off += esize;
150
+ mem_off += msize;
151
+ } while (reg_off <= reg_last && (reg_off & 63));
152
+ } while (reg_off <= reg_last);
50
+ }
153
+ }
51
+
154
+
52
+ /* UNDEF accesses to D16-D31 if they don't exist. */
155
+ mem_off = info->mem_off_first[1];
53
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
156
+ if (mem_off >= 0 && arm_tlb_mte_tagged(&info->page[1].attrs)) {
54
+ ((a->vd | a->vn | a->vm) & 0x10)) {
157
+ reg_off = info->reg_off_first[1];
55
+ return false;
158
+ reg_last = info->reg_off_last[1];
159
+
160
+ do {
161
+ uint64_t pg = vg[reg_off >> 6];
162
+ do {
163
+ if ((pg >> (reg_off & 63)) & 1) {
164
+ check(env, mtedesc, addr, ra);
165
+ }
166
+ reg_off += esize;
167
+ mem_off += msize;
168
+ } while (reg_off & 63);
169
+ } while (reg_off <= reg_last);
56
+ }
170
+ }
57
+
171
+}
58
+ if ((a->vn | a->vm | a->vd) & a->q) {
172
+
59
+ return false;
173
+typedef void sve_cont_ldst_mte_check_fn(SVEContLdSt *info, CPUARMState *env,
174
+ uint64_t *vg, target_ulong addr,
175
+ int esize, int msize, uint32_t mtedesc,
176
+ uintptr_t ra);
177
+
178
+static void sve_cont_ldst_mte_check1(SVEContLdSt *info, CPUARMState *env,
179
+ uint64_t *vg, target_ulong addr,
180
+ int esize, int msize, uint32_t mtedesc,
181
+ uintptr_t ra)
182
+{
183
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
184
+ mtedesc, ra, mte_check1);
185
+}
186
+
187
+static void sve_cont_ldst_mte_checkN(SVEContLdSt *info, CPUARMState *env,
188
+ uint64_t *vg, target_ulong addr,
189
+ int esize, int msize, uint32_t mtedesc,
190
+ uintptr_t ra)
191
+{
192
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
193
+ mtedesc, ra, mte_checkN);
194
+}
195
+
196
+
197
/*
198
* Common helper for all contiguous 1,2,3,4-register predicated stores.
199
*/
200
static inline QEMU_ALWAYS_INLINE
201
void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
202
uint32_t desc, const uintptr_t retaddr,
203
- const int esz, const int msz, const int N,
204
+ const int esz, const int msz, const int N, uint32_t mtedesc,
205
sve_ldst1_host_fn *host_fn,
206
- sve_ldst1_tlb_fn *tlb_fn)
207
+ sve_ldst1_tlb_fn *tlb_fn,
208
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
209
{
210
const unsigned rd = simd_data(desc);
211
const intptr_t reg_max = simd_oprsz(desc);
212
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
213
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
214
BP_MEM_READ, retaddr);
215
216
- /* TODO: MTE check. */
217
+ /*
218
+ * Handle mte checks for all active elements.
219
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
220
+ */
221
+ if (mte_check_fn && mtedesc) {
222
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
223
+ mtedesc, retaddr);
60
+ }
224
+ }
61
+
225
62
+ if (!vfp_access_check(s)) {
226
flags = info.page[0].flags | info.page[1].flags;
63
+ return true;
227
if (unlikely(flags != 0)) {
228
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
229
}
230
}
231
232
-#define DO_LD1_1(NAME, ESZ) \
233
-void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
234
- target_ulong addr, uint32_t desc) \
235
-{ \
236
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
237
- sve_##NAME##_host, sve_##NAME##_tlb); \
238
+static inline QEMU_ALWAYS_INLINE
239
+void sve_ldN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
240
+ uint32_t desc, const uintptr_t ra,
241
+ const int esz, const int msz, const int N,
242
+ sve_ldst1_host_fn *host_fn,
243
+ sve_ldst1_tlb_fn *tlb_fn)
244
+{
245
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
246
+ int bit55 = extract64(addr, 55, 1);
247
+
248
+ /* Remove mtedesc from the normal sve descriptor. */
249
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
250
+
251
+ /* Perform gross MTE suppression early. */
252
+ if (!tbi_check(desc, bit55) ||
253
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
254
+ mtedesc = 0;
64
+ }
255
+ }
65
+
256
+
66
+ opr_sz = (1 + a->q) * 8;
257
+ sve_ldN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
67
+ fpst = get_fpstatus_ptr(1);
258
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
68
+ fn_gvec_ptr = a->size ? gen_helper_gvec_fcmlas : gen_helper_gvec_fcmlah;
259
}
69
+ tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd),
260
70
+ vfp_reg_offset(1, a->vn),
261
-#define DO_LD1_2(NAME, ESZ, MSZ) \
71
+ vfp_reg_offset(1, a->vm),
262
-void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
72
+ fpst, opr_sz, opr_sz, a->rot,
263
- target_ulong addr, uint32_t desc) \
73
+ fn_gvec_ptr);
264
-{ \
74
+ tcg_temp_free_ptr(fpst);
265
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
75
+ return true;
266
- sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
267
-} \
268
-void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
269
- target_ulong addr, uint32_t desc) \
270
-{ \
271
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
272
- sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
273
+#define DO_LD1_1(NAME, ESZ) \
274
+void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
275
+ target_ulong addr, uint32_t desc) \
276
+{ \
277
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, 0, \
278
+ sve_##NAME##_host, sve_##NAME##_tlb, NULL); \
279
+} \
280
+void HELPER(sve_##NAME##_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
284
+ sve_##NAME##_host, sve_##NAME##_tlb); \
76
+}
285
+}
77
diff --git a/target/arm/translate.c b/target/arm/translate.c
286
+
287
+#define DO_LD1_2(NAME, ESZ, MSZ) \
288
+void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
289
+ target_ulong addr, uint32_t desc) \
290
+{ \
291
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
292
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb, NULL); \
293
+} \
294
+void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
295
+ target_ulong addr, uint32_t desc) \
296
+{ \
297
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
298
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb, NULL); \
299
+} \
300
+void HELPER(sve_##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
301
+ target_ulong addr, uint32_t desc) \
302
+{ \
303
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
304
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
305
+} \
306
+void HELPER(sve_##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
307
+ target_ulong addr, uint32_t desc) \
308
+{ \
309
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
310
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
311
}
312
313
DO_LD1_1(ld1bb, MO_8)
314
@@ -XXX,XX +XXX,XX @@ DO_LD1_2(ld1dd, MO_64, MO_64)
315
#undef DO_LD1_1
316
#undef DO_LD1_2
317
318
-#define DO_LDN_1(N) \
319
-void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
320
- target_ulong addr, uint32_t desc) \
321
-{ \
322
- sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
323
- sve_ld1bb_host, sve_ld1bb_tlb); \
324
+#define DO_LDN_1(N) \
325
+void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
326
+ target_ulong addr, uint32_t desc) \
327
+{ \
328
+ sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, 0, \
329
+ sve_ld1bb_host, sve_ld1bb_tlb, NULL); \
330
+} \
331
+void HELPER(sve_ld##N##bb_r_mte)(CPUARMState *env, void *vg, \
332
+ target_ulong addr, uint32_t desc) \
333
+{ \
334
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
335
+ sve_ld1bb_host, sve_ld1bb_tlb); \
336
}
337
338
-#define DO_LDN_2(N, SUFF, ESZ) \
339
-void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
340
- target_ulong addr, uint32_t desc) \
341
-{ \
342
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
343
- sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
344
-} \
345
-void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
346
- target_ulong addr, uint32_t desc) \
347
-{ \
348
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
349
- sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
350
+#define DO_LDN_2(N, SUFF, ESZ) \
351
+void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
352
+ target_ulong addr, uint32_t desc) \
353
+{ \
354
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
355
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb, NULL); \
356
+} \
357
+void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
358
+ target_ulong addr, uint32_t desc) \
359
+{ \
360
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
361
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb, NULL); \
362
+} \
363
+void HELPER(sve_ld##N##SUFF##_le_r_mte)(CPUARMState *env, void *vg, \
364
+ target_ulong addr, uint32_t desc) \
365
+{ \
366
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
367
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
368
+} \
369
+void HELPER(sve_ld##N##SUFF##_be_r_mte)(CPUARMState *env, void *vg, \
370
+ target_ulong addr, uint32_t desc) \
371
+{ \
372
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
373
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
374
}
375
376
DO_LDN_1(2)
377
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
78
index XXXXXXX..XXXXXXX 100644
378
index XXXXXXX..XXXXXXX 100644
79
--- a/target/arm/translate.c
379
--- a/target/arm/translate-sve.c
80
+++ b/target/arm/translate.c
380
+++ b/target/arm/translate-sve.c
81
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
381
@@ -XXX,XX +XXX,XX @@ static const uint8_t dtype_esz[16] = {
82
bool is_long = false, q = extract32(insn, 6, 1);
382
};
83
bool ptr_is_env = false;
383
84
384
static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
85
- if ((insn & 0xfe200f10) == 0xfc200800) {
385
- int dtype, gen_helper_gvec_mem *fn)
86
- /* VCMLA -- 1111 110R R.1S .... .... 1000 ...0 .... */
386
+ int dtype, uint32_t mte_n, bool is_write,
87
- int size = extract32(insn, 20, 1);
387
+ gen_helper_gvec_mem *fn)
88
- data = extract32(insn, 23, 2); /* rot */
388
{
89
- if (!dc_isar_feature(aa32_vcma, s)
389
unsigned vsz = vec_full_reg_size(s);
90
- || (!size && !dc_isar_feature(aa32_fp16_arith, s))) {
390
TCGv_ptr t_pg;
91
- return 1;
391
TCGv_i32 t_desc;
92
- }
392
- int desc;
93
- fn_gvec_ptr = size ? gen_helper_gvec_fcmlas : gen_helper_gvec_fcmlah;
393
+ int desc = 0;
94
- } else if ((insn & 0xfea00f10) == 0xfc800800) {
394
95
+ if ((insn & 0xfea00f10) == 0xfc800800) {
395
- /* For e.g. LD4, there are not enough arguments to pass all 4
96
/* VCADD -- 1111 110R 1.0S .... .... 1000 ...0 .... */
396
+ /*
97
int size = extract32(insn, 20, 1);
397
+ * For e.g. LD4, there are not enough arguments to pass all 4
98
data = extract32(insn, 24, 1); /* rot */
398
* registers as pointers, so encode the regno into the data field.
399
* For consistency, do this even for LD1.
400
+ * TODO: mte_n check here while callers are updated.
401
*/
402
- desc = simd_desc(vsz, vsz, zt);
403
+ if (mte_n && s->mte_active[0]) {
404
+ int msz = dtype_msz(dtype);
405
+
406
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
407
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
408
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
409
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
410
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
411
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
412
+ desc <<= SVE_MTEDESC_SHIFT;
413
+ }
414
+ desc = simd_desc(vsz, vsz, zt | desc);
415
t_desc = tcg_const_i32(desc);
416
t_pg = tcg_temp_new_ptr();
417
418
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
419
static void do_ld_zpa(DisasContext *s, int zt, int pg,
420
TCGv_i64 addr, int dtype, int nreg)
421
{
422
- static gen_helper_gvec_mem * const fns[2][16][4] = {
423
- /* Little-endian */
424
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
425
+ static gen_helper_gvec_mem * const fns[2][2][16][4] = {
426
+ { /* mte inactive, little-endian */
427
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
428
gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
429
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
430
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
431
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
432
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
433
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
434
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
435
436
- { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
437
- { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
438
- gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
439
- { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
440
- { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
441
+ { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
442
+ { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
443
+ gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
444
+ { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
445
+ { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
446
447
- { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
448
- { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
449
- { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
450
- gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
451
- { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
452
+ { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
453
+ { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
454
+ { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
455
+ gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
456
+ { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
457
458
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
459
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
460
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
461
- { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
462
- gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
463
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
464
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
465
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
466
+ { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
467
+ gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
468
469
- /* Big-endian */
470
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
471
- gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
472
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
473
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
474
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
475
+ /* mte inactive, big-endian */
476
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
477
+ gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
478
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
479
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
480
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
481
482
- { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
483
- { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
484
- gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
485
- { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
486
- { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
487
+ { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
488
+ { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
489
+ gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
490
+ { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
491
+ { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
492
493
- { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
494
- { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
495
- { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
496
- gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
497
- { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
498
+ { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
499
+ { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
500
+ { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
501
+ gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
502
+ { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
503
504
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
505
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
506
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
507
- { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
508
- gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } }
509
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
510
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
511
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
512
+ { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
513
+ gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } } },
514
+
515
+ { /* mte active, little-endian */
516
+ { { gen_helper_sve_ld1bb_r_mte,
517
+ gen_helper_sve_ld2bb_r_mte,
518
+ gen_helper_sve_ld3bb_r_mte,
519
+ gen_helper_sve_ld4bb_r_mte },
520
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
521
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
522
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
523
+
524
+ { gen_helper_sve_ld1sds_le_r_mte, NULL, NULL, NULL },
525
+ { gen_helper_sve_ld1hh_le_r_mte,
526
+ gen_helper_sve_ld2hh_le_r_mte,
527
+ gen_helper_sve_ld3hh_le_r_mte,
528
+ gen_helper_sve_ld4hh_le_r_mte },
529
+ { gen_helper_sve_ld1hsu_le_r_mte, NULL, NULL, NULL },
530
+ { gen_helper_sve_ld1hdu_le_r_mte, NULL, NULL, NULL },
531
+
532
+ { gen_helper_sve_ld1hds_le_r_mte, NULL, NULL, NULL },
533
+ { gen_helper_sve_ld1hss_le_r_mte, NULL, NULL, NULL },
534
+ { gen_helper_sve_ld1ss_le_r_mte,
535
+ gen_helper_sve_ld2ss_le_r_mte,
536
+ gen_helper_sve_ld3ss_le_r_mte,
537
+ gen_helper_sve_ld4ss_le_r_mte },
538
+ { gen_helper_sve_ld1sdu_le_r_mte, NULL, NULL, NULL },
539
+
540
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
541
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
542
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
543
+ { gen_helper_sve_ld1dd_le_r_mte,
544
+ gen_helper_sve_ld2dd_le_r_mte,
545
+ gen_helper_sve_ld3dd_le_r_mte,
546
+ gen_helper_sve_ld4dd_le_r_mte } },
547
+
548
+ /* mte active, big-endian */
549
+ { { gen_helper_sve_ld1bb_r_mte,
550
+ gen_helper_sve_ld2bb_r_mte,
551
+ gen_helper_sve_ld3bb_r_mte,
552
+ gen_helper_sve_ld4bb_r_mte },
553
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
554
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
555
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
556
+
557
+ { gen_helper_sve_ld1sds_be_r_mte, NULL, NULL, NULL },
558
+ { gen_helper_sve_ld1hh_be_r_mte,
559
+ gen_helper_sve_ld2hh_be_r_mte,
560
+ gen_helper_sve_ld3hh_be_r_mte,
561
+ gen_helper_sve_ld4hh_be_r_mte },
562
+ { gen_helper_sve_ld1hsu_be_r_mte, NULL, NULL, NULL },
563
+ { gen_helper_sve_ld1hdu_be_r_mte, NULL, NULL, NULL },
564
+
565
+ { gen_helper_sve_ld1hds_be_r_mte, NULL, NULL, NULL },
566
+ { gen_helper_sve_ld1hss_be_r_mte, NULL, NULL, NULL },
567
+ { gen_helper_sve_ld1ss_be_r_mte,
568
+ gen_helper_sve_ld2ss_be_r_mte,
569
+ gen_helper_sve_ld3ss_be_r_mte,
570
+ gen_helper_sve_ld4ss_be_r_mte },
571
+ { gen_helper_sve_ld1sdu_be_r_mte, NULL, NULL, NULL },
572
+
573
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
574
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
575
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
576
+ { gen_helper_sve_ld1dd_be_r_mte,
577
+ gen_helper_sve_ld2dd_be_r_mte,
578
+ gen_helper_sve_ld3dd_be_r_mte,
579
+ gen_helper_sve_ld4dd_be_r_mte } } },
580
};
581
- gen_helper_gvec_mem *fn = fns[s->be_data == MO_BE][dtype][nreg];
582
+ gen_helper_gvec_mem *fn
583
+ = fns[s->mte_active[0]][s->be_data == MO_BE][dtype][nreg];
584
585
- /* While there are holes in the table, they are not
586
+ /*
587
+ * While there are holes in the table, they are not
588
* accessible via the instruction encoding.
589
*/
590
assert(fn != NULL);
591
- do_mem_zpa(s, zt, pg, addr, dtype, fn);
592
+ do_mem_zpa(s, zt, pg, addr, dtype, nreg, false, fn);
593
}
594
595
static bool trans_LD_zprr(DisasContext *s, arg_rprr_load *a)
596
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
597
TCGv_i64 addr = new_tmp_a64(s);
598
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
599
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
600
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
601
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
602
fns[s->be_data == MO_BE][a->dtype]);
603
}
604
return true;
605
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
606
TCGv_i64 addr = new_tmp_a64(s);
607
608
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
609
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
610
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
611
fns[s->be_data == MO_BE][a->dtype]);
612
}
613
return true;
614
@@ -XXX,XX +XXX,XX @@ static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
615
fn = fn_multiple[be][nreg - 1][msz];
616
}
617
assert(fn != NULL);
618
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), fn);
619
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
620
}
621
622
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
99
--
623
--
100
2.20.1
624
2.20.1
101
625
102
626
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
hw/arm: versal: Add support for the RTC.
3
Because the elements are sequential, we can eliminate many tests all
4
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
8
Message-id: 20200626033144.790098-35-richard.henderson@linaro.org
9
Message-id: 20200427181649.26851-10-edgar.iglesias@gmail.com
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
include/hw/arm/xlnx-versal.h | 8 ++++++++
11
target/arm/helper-sve.h | 47 +++++++++++
13
hw/arm/xlnx-versal.c | 21 +++++++++++++++++++++
12
target/arm/sve_helper.c | 95 ++++++++++++++++------
14
2 files changed, 29 insertions(+)
13
target/arm/translate-sve.c | 162 ++++++++++++++++++++++++-------------
15
14
3 files changed, 226 insertions(+), 78 deletions(-)
16
diff --git a/include/hw/arm/xlnx-versal.h b/include/hw/arm/xlnx-versal.h
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/xlnx-versal.h
18
--- a/target/arm/helper-sve.h
19
+++ b/include/hw/arm/xlnx-versal.h
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_st1hd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
#include "hw/char/pl011.h"
21
DEF_HELPER_FLAGS_4(sve_st1sd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
#include "hw/dma/xlnx-zdma.h"
22
DEF_HELPER_FLAGS_4(sve_st1sd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
#include "hw/net/cadence_gem.h"
23
24
+#include "hw/rtc/xlnx-zynqmp-rtc.h"
24
+DEF_HELPER_FLAGS_4(sve_st1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
25
+DEF_HELPER_FLAGS_4(sve_st2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
#define TYPE_XLNX_VERSAL "xlnx-versal"
26
+DEF_HELPER_FLAGS_4(sve_st3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
#define XLNX_VERSAL(obj) OBJECT_CHECK(Versal, (obj), TYPE_XLNX_VERSAL)
27
+DEF_HELPER_FLAGS_4(sve_st4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
28
+
29
struct {
29
+DEF_HELPER_FLAGS_4(sve_st1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
SDHCIState sd[XLNX_VERSAL_NR_SDS];
30
+DEF_HELPER_FLAGS_4(sve_st2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
} iou;
31
+DEF_HELPER_FLAGS_4(sve_st3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
32
+
32
+DEF_HELPER_FLAGS_4(sve_st4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
33
+ XlnxZynqMPRTC rtc;
33
+
34
} pmc;
34
+DEF_HELPER_FLAGS_4(sve_st1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
35
+DEF_HELPER_FLAGS_4(sve_st2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
struct {
36
+DEF_HELPER_FLAGS_4(sve_st3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
37
+DEF_HELPER_FLAGS_4(sve_st4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
#define VERSAL_GEM1_IRQ_0 58
38
+
39
#define VERSAL_GEM1_WAKE_IRQ_0 59
39
+DEF_HELPER_FLAGS_4(sve_st1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
#define VERSAL_ADMA_IRQ_0 60
40
+DEF_HELPER_FLAGS_4(sve_st2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+#define VERSAL_RTC_APB_ERR_IRQ 121
41
+DEF_HELPER_FLAGS_4(sve_st3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
#define VERSAL_SD0_IRQ_0 126
42
+DEF_HELPER_FLAGS_4(sve_st4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
+#define VERSAL_RTC_ALARM_IRQ 142
43
+
44
+#define VERSAL_RTC_SECONDS_IRQ 143
44
+DEF_HELPER_FLAGS_4(sve_st1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
45
+DEF_HELPER_FLAGS_4(sve_st2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
/* Architecturally reserved IRQs suitable for virtualization. */
46
+DEF_HELPER_FLAGS_4(sve_st3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
#define VERSAL_RSVD_IRQ_FIRST 111
47
+DEF_HELPER_FLAGS_4(sve_st4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
@@ -XXX,XX +XXX,XX @@ typedef struct Versal {
48
+
49
#define MM_PMC_SD0_SIZE 0x10000
49
+DEF_HELPER_FLAGS_4(sve_st1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
#define MM_PMC_CRP 0xf1260000U
50
+DEF_HELPER_FLAGS_4(sve_st2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
#define MM_PMC_CRP_SIZE 0x10000
51
+DEF_HELPER_FLAGS_4(sve_st3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+#define MM_PMC_RTC 0xf12a0000
52
+DEF_HELPER_FLAGS_4(sve_st4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
53
+#define MM_PMC_RTC_SIZE 0x10000
53
+
54
#endif
54
+DEF_HELPER_FLAGS_4(sve_st1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
diff --git a/hw/arm/xlnx-versal.c b/hw/arm/xlnx-versal.c
55
+DEF_HELPER_FLAGS_4(sve_st2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_st3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_4(sve_st4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_4(sve_st1bh_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_st1bs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_st1bd_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_st1hs_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_st1hd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_st1hs_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_st1hd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_st1sd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_st1sd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+
71
DEF_HELPER_FLAGS_6(sve_ldbsu_zsu, TCG_CALL_NO_WG,
72
void, env, ptr, ptr, ptr, tl, i32)
73
DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu, TCG_CALL_NO_WG,
74
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
56
index XXXXXXX..XXXXXXX 100644
75
index XXXXXXX..XXXXXXX 100644
57
--- a/hw/arm/xlnx-versal.c
76
--- a/target/arm/sve_helper.c
58
+++ b/hw/arm/xlnx-versal.c
77
+++ b/target/arm/sve_helper.c
59
@@ -XXX,XX +XXX,XX @@ static void versal_create_sds(Versal *s, qemu_irq *pic)
78
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_LDNF1_2(dd, MO_64, MO_64)
79
*/
80
81
static inline QEMU_ALWAYS_INLINE
82
-void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
83
- const uintptr_t retaddr, const int esz,
84
- const int msz, const int N,
85
+void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr,
86
+ uint32_t desc, const uintptr_t retaddr,
87
+ const int esz, const int msz, const int N, uint32_t mtedesc,
88
sve_ldst1_host_fn *host_fn,
89
- sve_ldst1_tlb_fn *tlb_fn)
90
+ sve_ldst1_tlb_fn *tlb_fn,
91
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
92
{
93
const unsigned rd = simd_data(desc);
94
const intptr_t reg_max = simd_oprsz(desc);
95
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
96
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
97
BP_MEM_WRITE, retaddr);
98
99
- /* TODO: MTE check. */
100
+ /*
101
+ * Handle mte checks for all active elements.
102
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
103
+ */
104
+ if (mte_check_fn && mtedesc) {
105
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
106
+ mtedesc, retaddr);
107
+ }
108
109
flags = info.page[0].flags | info.page[1].flags;
110
if (unlikely(flags != 0)) {
111
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
60
}
112
}
61
}
113
}
62
114
63
+static void versal_create_rtc(Versal *s, qemu_irq *pic)
115
-#define DO_STN_1(N, NAME, ESZ) \
116
-void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
117
- target_ulong addr, uint32_t desc) \
118
-{ \
119
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
120
- sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
121
+static inline QEMU_ALWAYS_INLINE
122
+void sve_stN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
123
+ uint32_t desc, const uintptr_t ra,
124
+ const int esz, const int msz, const int N,
125
+ sve_ldst1_host_fn *host_fn,
126
+ sve_ldst1_tlb_fn *tlb_fn)
64
+{
127
+{
65
+ SysBusDevice *sbd;
128
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
66
+ MemoryRegion *mr;
129
+ int bit55 = extract64(addr, 55, 1);
67
+
130
+
68
+ sysbus_init_child_obj(OBJECT(s), "rtc", &s->pmc.rtc, sizeof(s->pmc.rtc),
131
+ /* Remove mtedesc from the normal sve descriptor. */
69
+ TYPE_XLNX_ZYNQMP_RTC);
132
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
70
+ sbd = SYS_BUS_DEVICE(&s->pmc.rtc);
133
+
71
+ qdev_init_nofail(DEVICE(sbd));
134
+ /* Perform gross MTE suppression early. */
72
+
135
+ if (!tbi_check(desc, bit55) ||
73
+ mr = sysbus_mmio_get_region(sbd, 0);
136
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
74
+ memory_region_add_subregion(&s->mr_ps, MM_PMC_RTC, mr);
137
+ mtedesc = 0;
75
+
138
+ }
76
+ /*
139
+
77
+ * TODO: Connect the ALARM and SECONDS interrupts once our RTC model
140
+ sve_stN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
78
+ * supports them.
141
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
79
+ */
142
}
80
+ sysbus_connect_irq(sbd, 1, pic[VERSAL_RTC_APB_ERR_IRQ]);
143
144
-#define DO_STN_2(N, NAME, ESZ, MSZ) \
145
-void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
146
- target_ulong addr, uint32_t desc) \
147
-{ \
148
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
149
- sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
150
-} \
151
-void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
152
- target_ulong addr, uint32_t desc) \
153
-{ \
154
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
155
- sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
156
+#define DO_STN_1(N, NAME, ESZ) \
157
+void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
158
+ target_ulong addr, uint32_t desc) \
159
+{ \
160
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, 0, \
161
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb, NULL); \
162
+} \
163
+void HELPER(sve_st##N##NAME##_r_mte)(CPUARMState *env, void *vg, \
164
+ target_ulong addr, uint32_t desc) \
165
+{ \
166
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
167
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
81
+}
168
+}
82
+
169
+
83
/* This takes the board allocated linear DDR memory and creates aliases
170
+#define DO_STN_2(N, NAME, ESZ, MSZ) \
84
* for each split DDR range/aperture on the Versal address map.
171
+void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
85
*/
172
+ target_ulong addr, uint32_t desc) \
86
@@ -XXX,XX +XXX,XX @@ static void versal_realize(DeviceState *dev, Error **errp)
173
+{ \
87
versal_create_gems(s, pic);
174
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
88
versal_create_admas(s, pic);
175
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb, NULL); \
89
versal_create_sds(s, pic);
176
+} \
90
+ versal_create_rtc(s, pic);
177
+void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
91
versal_map_ddr(s);
178
+ target_ulong addr, uint32_t desc) \
92
versal_unimp(s);
179
+{ \
93
180
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
181
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb, NULL); \
182
+} \
183
+void HELPER(sve_st##N##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
184
+ target_ulong addr, uint32_t desc) \
185
+{ \
186
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
187
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
188
+} \
189
+void HELPER(sve_st##N##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
190
+ target_ulong addr, uint32_t desc) \
191
+{ \
192
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
193
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
194
}
195
196
DO_STN_1(1, bb, MO_8)
197
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
198
index XXXXXXX..XXXXXXX 100644
199
--- a/target/arm/translate-sve.c
200
+++ b/target/arm/translate-sve.c
201
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
202
static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
203
int msz, int esz, int nreg)
204
{
205
- static gen_helper_gvec_mem * const fn_single[2][4][4] = {
206
- { { gen_helper_sve_st1bb_r,
207
- gen_helper_sve_st1bh_r,
208
- gen_helper_sve_st1bs_r,
209
- gen_helper_sve_st1bd_r },
210
- { NULL,
211
- gen_helper_sve_st1hh_le_r,
212
- gen_helper_sve_st1hs_le_r,
213
- gen_helper_sve_st1hd_le_r },
214
- { NULL, NULL,
215
- gen_helper_sve_st1ss_le_r,
216
- gen_helper_sve_st1sd_le_r },
217
- { NULL, NULL, NULL,
218
- gen_helper_sve_st1dd_le_r } },
219
- { { gen_helper_sve_st1bb_r,
220
- gen_helper_sve_st1bh_r,
221
- gen_helper_sve_st1bs_r,
222
- gen_helper_sve_st1bd_r },
223
- { NULL,
224
- gen_helper_sve_st1hh_be_r,
225
- gen_helper_sve_st1hs_be_r,
226
- gen_helper_sve_st1hd_be_r },
227
- { NULL, NULL,
228
- gen_helper_sve_st1ss_be_r,
229
- gen_helper_sve_st1sd_be_r },
230
- { NULL, NULL, NULL,
231
- gen_helper_sve_st1dd_be_r } },
232
+ static gen_helper_gvec_mem * const fn_single[2][2][4][4] = {
233
+ { { { gen_helper_sve_st1bb_r,
234
+ gen_helper_sve_st1bh_r,
235
+ gen_helper_sve_st1bs_r,
236
+ gen_helper_sve_st1bd_r },
237
+ { NULL,
238
+ gen_helper_sve_st1hh_le_r,
239
+ gen_helper_sve_st1hs_le_r,
240
+ gen_helper_sve_st1hd_le_r },
241
+ { NULL, NULL,
242
+ gen_helper_sve_st1ss_le_r,
243
+ gen_helper_sve_st1sd_le_r },
244
+ { NULL, NULL, NULL,
245
+ gen_helper_sve_st1dd_le_r } },
246
+ { { gen_helper_sve_st1bb_r,
247
+ gen_helper_sve_st1bh_r,
248
+ gen_helper_sve_st1bs_r,
249
+ gen_helper_sve_st1bd_r },
250
+ { NULL,
251
+ gen_helper_sve_st1hh_be_r,
252
+ gen_helper_sve_st1hs_be_r,
253
+ gen_helper_sve_st1hd_be_r },
254
+ { NULL, NULL,
255
+ gen_helper_sve_st1ss_be_r,
256
+ gen_helper_sve_st1sd_be_r },
257
+ { NULL, NULL, NULL,
258
+ gen_helper_sve_st1dd_be_r } } },
259
+
260
+ { { { gen_helper_sve_st1bb_r_mte,
261
+ gen_helper_sve_st1bh_r_mte,
262
+ gen_helper_sve_st1bs_r_mte,
263
+ gen_helper_sve_st1bd_r_mte },
264
+ { NULL,
265
+ gen_helper_sve_st1hh_le_r_mte,
266
+ gen_helper_sve_st1hs_le_r_mte,
267
+ gen_helper_sve_st1hd_le_r_mte },
268
+ { NULL, NULL,
269
+ gen_helper_sve_st1ss_le_r_mte,
270
+ gen_helper_sve_st1sd_le_r_mte },
271
+ { NULL, NULL, NULL,
272
+ gen_helper_sve_st1dd_le_r_mte } },
273
+ { { gen_helper_sve_st1bb_r_mte,
274
+ gen_helper_sve_st1bh_r_mte,
275
+ gen_helper_sve_st1bs_r_mte,
276
+ gen_helper_sve_st1bd_r_mte },
277
+ { NULL,
278
+ gen_helper_sve_st1hh_be_r_mte,
279
+ gen_helper_sve_st1hs_be_r_mte,
280
+ gen_helper_sve_st1hd_be_r_mte },
281
+ { NULL, NULL,
282
+ gen_helper_sve_st1ss_be_r_mte,
283
+ gen_helper_sve_st1sd_be_r_mte },
284
+ { NULL, NULL, NULL,
285
+ gen_helper_sve_st1dd_be_r_mte } } },
286
};
287
- static gen_helper_gvec_mem * const fn_multiple[2][3][4] = {
288
- { { gen_helper_sve_st2bb_r,
289
- gen_helper_sve_st2hh_le_r,
290
- gen_helper_sve_st2ss_le_r,
291
- gen_helper_sve_st2dd_le_r },
292
- { gen_helper_sve_st3bb_r,
293
- gen_helper_sve_st3hh_le_r,
294
- gen_helper_sve_st3ss_le_r,
295
- gen_helper_sve_st3dd_le_r },
296
- { gen_helper_sve_st4bb_r,
297
- gen_helper_sve_st4hh_le_r,
298
- gen_helper_sve_st4ss_le_r,
299
- gen_helper_sve_st4dd_le_r } },
300
- { { gen_helper_sve_st2bb_r,
301
- gen_helper_sve_st2hh_be_r,
302
- gen_helper_sve_st2ss_be_r,
303
- gen_helper_sve_st2dd_be_r },
304
- { gen_helper_sve_st3bb_r,
305
- gen_helper_sve_st3hh_be_r,
306
- gen_helper_sve_st3ss_be_r,
307
- gen_helper_sve_st3dd_be_r },
308
- { gen_helper_sve_st4bb_r,
309
- gen_helper_sve_st4hh_be_r,
310
- gen_helper_sve_st4ss_be_r,
311
- gen_helper_sve_st4dd_be_r } },
312
+ static gen_helper_gvec_mem * const fn_multiple[2][2][3][4] = {
313
+ { { { gen_helper_sve_st2bb_r,
314
+ gen_helper_sve_st2hh_le_r,
315
+ gen_helper_sve_st2ss_le_r,
316
+ gen_helper_sve_st2dd_le_r },
317
+ { gen_helper_sve_st3bb_r,
318
+ gen_helper_sve_st3hh_le_r,
319
+ gen_helper_sve_st3ss_le_r,
320
+ gen_helper_sve_st3dd_le_r },
321
+ { gen_helper_sve_st4bb_r,
322
+ gen_helper_sve_st4hh_le_r,
323
+ gen_helper_sve_st4ss_le_r,
324
+ gen_helper_sve_st4dd_le_r } },
325
+ { { gen_helper_sve_st2bb_r,
326
+ gen_helper_sve_st2hh_be_r,
327
+ gen_helper_sve_st2ss_be_r,
328
+ gen_helper_sve_st2dd_be_r },
329
+ { gen_helper_sve_st3bb_r,
330
+ gen_helper_sve_st3hh_be_r,
331
+ gen_helper_sve_st3ss_be_r,
332
+ gen_helper_sve_st3dd_be_r },
333
+ { gen_helper_sve_st4bb_r,
334
+ gen_helper_sve_st4hh_be_r,
335
+ gen_helper_sve_st4ss_be_r,
336
+ gen_helper_sve_st4dd_be_r } } },
337
+ { { { gen_helper_sve_st2bb_r_mte,
338
+ gen_helper_sve_st2hh_le_r_mte,
339
+ gen_helper_sve_st2ss_le_r_mte,
340
+ gen_helper_sve_st2dd_le_r_mte },
341
+ { gen_helper_sve_st3bb_r_mte,
342
+ gen_helper_sve_st3hh_le_r_mte,
343
+ gen_helper_sve_st3ss_le_r_mte,
344
+ gen_helper_sve_st3dd_le_r_mte },
345
+ { gen_helper_sve_st4bb_r_mte,
346
+ gen_helper_sve_st4hh_le_r_mte,
347
+ gen_helper_sve_st4ss_le_r_mte,
348
+ gen_helper_sve_st4dd_le_r_mte } },
349
+ { { gen_helper_sve_st2bb_r_mte,
350
+ gen_helper_sve_st2hh_be_r_mte,
351
+ gen_helper_sve_st2ss_be_r_mte,
352
+ gen_helper_sve_st2dd_be_r_mte },
353
+ { gen_helper_sve_st3bb_r_mte,
354
+ gen_helper_sve_st3hh_be_r_mte,
355
+ gen_helper_sve_st3ss_be_r_mte,
356
+ gen_helper_sve_st3dd_be_r_mte },
357
+ { gen_helper_sve_st4bb_r_mte,
358
+ gen_helper_sve_st4hh_be_r_mte,
359
+ gen_helper_sve_st4ss_be_r_mte,
360
+ gen_helper_sve_st4dd_be_r_mte } } },
361
};
362
gen_helper_gvec_mem *fn;
363
int be = s->be_data == MO_BE;
364
365
if (nreg == 0) {
366
/* ST1 */
367
- fn = fn_single[be][msz][esz];
368
+ fn = fn_single[s->mte_active[0]][be][msz][esz];
369
+ nreg = 1;
370
} else {
371
/* ST2, ST3, ST4 -- msz == esz, enforced by encoding */
372
assert(msz == esz);
373
- fn = fn_multiple[be][nreg - 1][msz];
374
+ fn = fn_multiple[s->mte_active[0]][be][nreg - 1][msz];
375
}
376
assert(fn != NULL);
377
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
378
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), nreg, true, fn);
379
}
380
381
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
94
--
382
--
95
2.20.1
383
2.20.1
96
384
97
385
diff view generated by jsdifflib
1
We were accidentally permitting decode of Thumb Neon insns even if
1
From: Richard Henderson <richard.henderson@linaro.org>
2
the CPU didn't have the FEATURE_NEON bit set, because the feature
3
check was being done before the call to disas_neon_data_insn() and
4
disas_neon_ls_insn() in the Arm decoder but was omitted from the
5
Thumb decoder. Push the feature bit check down into the called
6
functions so it is done for both Arm and Thumb encodings.
7
2
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-36-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Message-id: 20200430181003.21682-3-peter.maydell@linaro.org
12
---
10
---
13
target/arm/translate.c | 16 ++++++++--------
11
target/arm/helper-sve.h | 98 ++++++++++++++++
14
1 file changed, 8 insertions(+), 8 deletions(-)
12
target/arm/sve_helper.c | 99 ++++++++++++++--
13
target/arm/translate-sve.c | 232 +++++++++++++++++++++++++------------
14
3 files changed, 343 insertions(+), 86 deletions(-)
15
15
16
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.c
18
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/translate.c
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
TCGv_i32 tmp2;
21
DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
TCGv_i64 tmp64;
22
DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
23
24
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
24
+DEF_HELPER_FLAGS_4(sve_ldff1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+ return 1;
25
+DEF_HELPER_FLAGS_4(sve_ldff1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_ldff1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_ldff1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+DEF_HELPER_FLAGS_4(sve_ldff1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_4(sve_ldff1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_ldff1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+
32
+DEF_HELPER_FLAGS_4(sve_ldff1hh_le_r_mte, TCG_CALL_NO_WG,
33
+ void, env, ptr, tl, i32)
34
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_le_r_mte, TCG_CALL_NO_WG,
35
+ void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_le_r_mte, TCG_CALL_NO_WG,
37
+ void, env, ptr, tl, i32)
38
+DEF_HELPER_FLAGS_4(sve_ldff1hss_le_r_mte, TCG_CALL_NO_WG,
39
+ void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ldff1hds_le_r_mte, TCG_CALL_NO_WG,
41
+ void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ldff1hh_be_r_mte, TCG_CALL_NO_WG,
44
+ void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_be_r_mte, TCG_CALL_NO_WG,
46
+ void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_be_r_mte, TCG_CALL_NO_WG,
48
+ void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ldff1hss_be_r_mte, TCG_CALL_NO_WG,
50
+ void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ldff1hds_be_r_mte, TCG_CALL_NO_WG,
52
+ void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_ldff1ss_le_r_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_le_r_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, tl, i32)
58
+DEF_HELPER_FLAGS_4(sve_ldff1sds_le_r_mte, TCG_CALL_NO_WG,
59
+ void, env, ptr, tl, i32)
60
+
61
+DEF_HELPER_FLAGS_4(sve_ldff1ss_be_r_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_be_r_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r_mte, TCG_CALL_NO_WG,
69
+ void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r_mte, TCG_CALL_NO_WG,
71
+ void, env, ptr, tl, i32)
72
+
73
DEF_HELPER_FLAGS_4(sve_ldnf1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
80
+DEF_HELPER_FLAGS_4(sve_ldnf1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ldnf1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_4(sve_ldnf1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_4(sve_ldnf1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ldnf1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+
88
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_le_r_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_le_r_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_le_r_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_le_r_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_le_r_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, tl, i32)
98
+
99
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_be_r_mte, TCG_CALL_NO_WG,
100
+ void, env, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_be_r_mte, TCG_CALL_NO_WG,
102
+ void, env, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_be_r_mte, TCG_CALL_NO_WG,
104
+ void, env, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_be_r_mte, TCG_CALL_NO_WG,
106
+ void, env, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_be_r_mte, TCG_CALL_NO_WG,
108
+ void, env, ptr, tl, i32)
109
+
110
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_le_r_mte, TCG_CALL_NO_WG,
111
+ void, env, ptr, tl, i32)
112
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_le_r_mte, TCG_CALL_NO_WG,
113
+ void, env, ptr, tl, i32)
114
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_le_r_mte, TCG_CALL_NO_WG,
115
+ void, env, ptr, tl, i32)
116
+
117
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_be_r_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_be_r_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, tl, i32)
123
+
124
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r_mte, TCG_CALL_NO_WG,
125
+ void, env, ptr, tl, i32)
126
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r_mte, TCG_CALL_NO_WG,
127
+ void, env, ptr, tl, i32)
128
+
129
DEF_HELPER_FLAGS_4(sve_st1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
130
DEF_HELPER_FLAGS_4(sve_st2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
131
DEF_HELPER_FLAGS_4(sve_st3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
132
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/target/arm/sve_helper.c
135
+++ b/target/arm/sve_helper.c
136
@@ -XXX,XX +XXX,XX @@ static void record_fault(CPUARMState *env, uintptr_t i, uintptr_t oprsz)
137
*/
138
static inline QEMU_ALWAYS_INLINE
139
void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
140
- uint32_t desc, const uintptr_t retaddr,
141
+ uint32_t desc, const uintptr_t retaddr, uint32_t mtedesc,
142
const int esz, const int msz, const SVEContFault fault,
143
sve_ldst1_host_fn *host_fn,
144
sve_ldst1_tlb_fn *tlb_fn)
145
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
146
mem_off = info.mem_off_first[0];
147
flags = info.page[0].flags;
148
149
+ /*
150
+ * Disable MTE checking if the Tagged bit is not set. Since TBI must
151
+ * be set within MTEDESC for MTE, !mtedesc => !mte_active.
152
+ */
153
+ if (arm_tlb_mte_tagged(&info.page[0].attrs)) {
154
+ mtedesc = 0;
26
+ }
155
+ }
27
+
156
+
28
/* FIXME: this access check should not take precedence over UNDEF
157
if (fault == FAULT_FIRST) {
29
* for invalid encodings; we will generate incorrect syndrome information
158
+ /* Trapping mte check for the first-fault element. */
30
* for attempts to execute invalid vfp/neon encodings with FP disabled.
159
+ if (mtedesc) {
31
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
160
+ mte_check1(env, mtedesc, addr + mem_off, retaddr);
32
TCGv_ptr ptr1, ptr2, ptr3;
161
+ }
33
TCGv_i64 tmp64;
162
+
34
163
/*
35
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
164
* Special handling of the first active element,
36
+ return 1;
165
* if it crosses a page boundary or is MMIO.
166
*/
167
bool is_split = mem_off == info.mem_off_split;
168
- /* TODO: MTE check. */
169
if (unlikely(flags != 0) || unlikely(is_split)) {
170
/*
171
* Use the slow path for cross-page handling.
172
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
173
/* Watchpoint hit, see below. */
174
goto do_fault;
175
}
176
- /* TODO: MTE check. */
177
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
178
+ goto do_fault;
179
+ }
180
/*
181
* Use the slow path for cross-page handling.
182
* This is RAM, without a watchpoint, and will not trap.
183
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
184
& BP_MEM_READ)) {
185
goto do_fault;
186
}
187
- /* TODO: MTE check. */
188
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
189
+ goto do_fault;
190
+ }
191
host_fn(vd, reg_off, host + mem_off);
192
}
193
reg_off += 1 << esz;
194
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
195
record_fault(env, reg_off, reg_max);
196
}
197
198
-#define DO_LDFF1_LDNF1_1(PART, ESZ) \
199
+static inline QEMU_ALWAYS_INLINE
200
+void sve_ldnfff1_r_mte(CPUARMState *env, void *vg, target_ulong addr,
201
+ uint32_t desc, const uintptr_t retaddr,
202
+ const int esz, const int msz, const SVEContFault fault,
203
+ sve_ldst1_host_fn *host_fn,
204
+ sve_ldst1_tlb_fn *tlb_fn)
205
+{
206
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
207
+ int bit55 = extract64(addr, 55, 1);
208
+
209
+ /* Remove mtedesc from the normal sve descriptor. */
210
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
211
+
212
+ /* Perform gross MTE suppression early. */
213
+ if (!tbi_check(desc, bit55) ||
214
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
215
+ mtedesc = 0;
37
+ }
216
+ }
38
+
217
+
39
/* FIXME: this access check should not take precedence over UNDEF
218
+ sve_ldnfff1_r(env, vg, addr, desc, retaddr, mtedesc,
40
* for invalid encodings; we will generate incorrect syndrome information
219
+ esz, msz, fault, host_fn, tlb_fn);
41
* for attempts to execute invalid vfp/neon encodings with FP disabled.
220
+}
42
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
221
+
43
222
+#define DO_LDFF1_LDNF1_1(PART, ESZ) \
44
if (((insn >> 25) & 7) == 1) {
223
void HELPER(sve_ldff1##PART##_r)(CPUARMState *env, void *vg, \
45
/* NEON Data processing. */
224
target_ulong addr, uint32_t desc) \
46
- if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
225
{ \
47
- goto illegal_op;
226
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
48
- }
227
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_FIRST, \
49
-
228
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
50
if (disas_neon_data_insn(s, insn)) {
229
} \
51
goto illegal_op;
230
void HELPER(sve_ldnf1##PART##_r)(CPUARMState *env, void *vg, \
52
}
231
target_ulong addr, uint32_t desc) \
53
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
232
{ \
54
}
233
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
55
if ((insn & 0x0f100000) == 0x04000000) {
234
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_NO, \
56
/* NEON load/store. */
235
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
57
- if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
236
+} \
58
- goto illegal_op;
237
+void HELPER(sve_ldff1##PART##_r_mte)(CPUARMState *env, void *vg, \
59
- }
238
+ target_ulong addr, uint32_t desc) \
60
-
239
+{ \
61
if (disas_neon_ls_insn(s, insn)) {
240
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
62
goto illegal_op;
241
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
63
}
242
+} \
243
+void HELPER(sve_ldnf1##PART##_r_mte)(CPUARMState *env, void *vg, \
244
+ target_ulong addr, uint32_t desc) \
245
+{ \
246
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
247
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
248
}
249
250
-#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
251
+#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
252
void HELPER(sve_ldff1##PART##_le_r)(CPUARMState *env, void *vg, \
253
target_ulong addr, uint32_t desc) \
254
{ \
255
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
256
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
257
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
258
} \
259
void HELPER(sve_ldnf1##PART##_le_r)(CPUARMState *env, void *vg, \
260
target_ulong addr, uint32_t desc) \
261
{ \
262
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
263
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
264
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
265
} \
266
void HELPER(sve_ldff1##PART##_be_r)(CPUARMState *env, void *vg, \
267
target_ulong addr, uint32_t desc) \
268
{ \
269
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
270
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
271
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
272
} \
273
void HELPER(sve_ldnf1##PART##_be_r)(CPUARMState *env, void *vg, \
274
target_ulong addr, uint32_t desc) \
275
{ \
276
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
277
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
278
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
279
+} \
280
+void HELPER(sve_ldff1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
284
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
285
+} \
286
+void HELPER(sve_ldnf1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
287
+ target_ulong addr, uint32_t desc) \
288
+{ \
289
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
290
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
291
+} \
292
+void HELPER(sve_ldff1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
293
+ target_ulong addr, uint32_t desc) \
294
+{ \
295
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
296
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
297
+} \
298
+void HELPER(sve_ldnf1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
299
+ target_ulong addr, uint32_t desc) \
300
+{ \
301
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
302
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
303
}
304
305
DO_LDFF1_LDNF1_1(bb, MO_8)
306
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
307
index XXXXXXX..XXXXXXX 100644
308
--- a/target/arm/translate-sve.c
309
+++ b/target/arm/translate-sve.c
310
@@ -XXX,XX +XXX,XX @@ static bool trans_LD_zpri(DisasContext *s, arg_rpri_load *a)
311
312
static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
313
{
314
- static gen_helper_gvec_mem * const fns[2][16] = {
315
- /* Little-endian */
316
- { gen_helper_sve_ldff1bb_r,
317
- gen_helper_sve_ldff1bhu_r,
318
- gen_helper_sve_ldff1bsu_r,
319
- gen_helper_sve_ldff1bdu_r,
320
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
321
+ { /* mte inactive, little-endian */
322
+ { gen_helper_sve_ldff1bb_r,
323
+ gen_helper_sve_ldff1bhu_r,
324
+ gen_helper_sve_ldff1bsu_r,
325
+ gen_helper_sve_ldff1bdu_r,
326
327
- gen_helper_sve_ldff1sds_le_r,
328
- gen_helper_sve_ldff1hh_le_r,
329
- gen_helper_sve_ldff1hsu_le_r,
330
- gen_helper_sve_ldff1hdu_le_r,
331
+ gen_helper_sve_ldff1sds_le_r,
332
+ gen_helper_sve_ldff1hh_le_r,
333
+ gen_helper_sve_ldff1hsu_le_r,
334
+ gen_helper_sve_ldff1hdu_le_r,
335
336
- gen_helper_sve_ldff1hds_le_r,
337
- gen_helper_sve_ldff1hss_le_r,
338
- gen_helper_sve_ldff1ss_le_r,
339
- gen_helper_sve_ldff1sdu_le_r,
340
+ gen_helper_sve_ldff1hds_le_r,
341
+ gen_helper_sve_ldff1hss_le_r,
342
+ gen_helper_sve_ldff1ss_le_r,
343
+ gen_helper_sve_ldff1sdu_le_r,
344
345
- gen_helper_sve_ldff1bds_r,
346
- gen_helper_sve_ldff1bss_r,
347
- gen_helper_sve_ldff1bhs_r,
348
- gen_helper_sve_ldff1dd_le_r },
349
+ gen_helper_sve_ldff1bds_r,
350
+ gen_helper_sve_ldff1bss_r,
351
+ gen_helper_sve_ldff1bhs_r,
352
+ gen_helper_sve_ldff1dd_le_r },
353
354
- /* Big-endian */
355
- { gen_helper_sve_ldff1bb_r,
356
- gen_helper_sve_ldff1bhu_r,
357
- gen_helper_sve_ldff1bsu_r,
358
- gen_helper_sve_ldff1bdu_r,
359
+ /* mte inactive, big-endian */
360
+ { gen_helper_sve_ldff1bb_r,
361
+ gen_helper_sve_ldff1bhu_r,
362
+ gen_helper_sve_ldff1bsu_r,
363
+ gen_helper_sve_ldff1bdu_r,
364
365
- gen_helper_sve_ldff1sds_be_r,
366
- gen_helper_sve_ldff1hh_be_r,
367
- gen_helper_sve_ldff1hsu_be_r,
368
- gen_helper_sve_ldff1hdu_be_r,
369
+ gen_helper_sve_ldff1sds_be_r,
370
+ gen_helper_sve_ldff1hh_be_r,
371
+ gen_helper_sve_ldff1hsu_be_r,
372
+ gen_helper_sve_ldff1hdu_be_r,
373
374
- gen_helper_sve_ldff1hds_be_r,
375
- gen_helper_sve_ldff1hss_be_r,
376
- gen_helper_sve_ldff1ss_be_r,
377
- gen_helper_sve_ldff1sdu_be_r,
378
+ gen_helper_sve_ldff1hds_be_r,
379
+ gen_helper_sve_ldff1hss_be_r,
380
+ gen_helper_sve_ldff1ss_be_r,
381
+ gen_helper_sve_ldff1sdu_be_r,
382
383
- gen_helper_sve_ldff1bds_r,
384
- gen_helper_sve_ldff1bss_r,
385
- gen_helper_sve_ldff1bhs_r,
386
- gen_helper_sve_ldff1dd_be_r },
387
+ gen_helper_sve_ldff1bds_r,
388
+ gen_helper_sve_ldff1bss_r,
389
+ gen_helper_sve_ldff1bhs_r,
390
+ gen_helper_sve_ldff1dd_be_r } },
391
+
392
+ { /* mte active, little-endian */
393
+ { gen_helper_sve_ldff1bb_r_mte,
394
+ gen_helper_sve_ldff1bhu_r_mte,
395
+ gen_helper_sve_ldff1bsu_r_mte,
396
+ gen_helper_sve_ldff1bdu_r_mte,
397
+
398
+ gen_helper_sve_ldff1sds_le_r_mte,
399
+ gen_helper_sve_ldff1hh_le_r_mte,
400
+ gen_helper_sve_ldff1hsu_le_r_mte,
401
+ gen_helper_sve_ldff1hdu_le_r_mte,
402
+
403
+ gen_helper_sve_ldff1hds_le_r_mte,
404
+ gen_helper_sve_ldff1hss_le_r_mte,
405
+ gen_helper_sve_ldff1ss_le_r_mte,
406
+ gen_helper_sve_ldff1sdu_le_r_mte,
407
+
408
+ gen_helper_sve_ldff1bds_r_mte,
409
+ gen_helper_sve_ldff1bss_r_mte,
410
+ gen_helper_sve_ldff1bhs_r_mte,
411
+ gen_helper_sve_ldff1dd_le_r_mte },
412
+
413
+ /* mte active, big-endian */
414
+ { gen_helper_sve_ldff1bb_r_mte,
415
+ gen_helper_sve_ldff1bhu_r_mte,
416
+ gen_helper_sve_ldff1bsu_r_mte,
417
+ gen_helper_sve_ldff1bdu_r_mte,
418
+
419
+ gen_helper_sve_ldff1sds_be_r_mte,
420
+ gen_helper_sve_ldff1hh_be_r_mte,
421
+ gen_helper_sve_ldff1hsu_be_r_mte,
422
+ gen_helper_sve_ldff1hdu_be_r_mte,
423
+
424
+ gen_helper_sve_ldff1hds_be_r_mte,
425
+ gen_helper_sve_ldff1hss_be_r_mte,
426
+ gen_helper_sve_ldff1ss_be_r_mte,
427
+ gen_helper_sve_ldff1sdu_be_r_mte,
428
+
429
+ gen_helper_sve_ldff1bds_r_mte,
430
+ gen_helper_sve_ldff1bss_r_mte,
431
+ gen_helper_sve_ldff1bhs_r_mte,
432
+ gen_helper_sve_ldff1dd_be_r_mte } },
433
};
434
435
if (sve_access_check(s)) {
436
TCGv_i64 addr = new_tmp_a64(s);
437
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
438
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
439
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
440
- fns[s->be_data == MO_BE][a->dtype]);
441
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
442
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
443
}
444
return true;
445
}
446
447
static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
448
{
449
- static gen_helper_gvec_mem * const fns[2][16] = {
450
- /* Little-endian */
451
- { gen_helper_sve_ldnf1bb_r,
452
- gen_helper_sve_ldnf1bhu_r,
453
- gen_helper_sve_ldnf1bsu_r,
454
- gen_helper_sve_ldnf1bdu_r,
455
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
456
+ { /* mte inactive, little-endian */
457
+ { gen_helper_sve_ldnf1bb_r,
458
+ gen_helper_sve_ldnf1bhu_r,
459
+ gen_helper_sve_ldnf1bsu_r,
460
+ gen_helper_sve_ldnf1bdu_r,
461
462
- gen_helper_sve_ldnf1sds_le_r,
463
- gen_helper_sve_ldnf1hh_le_r,
464
- gen_helper_sve_ldnf1hsu_le_r,
465
- gen_helper_sve_ldnf1hdu_le_r,
466
+ gen_helper_sve_ldnf1sds_le_r,
467
+ gen_helper_sve_ldnf1hh_le_r,
468
+ gen_helper_sve_ldnf1hsu_le_r,
469
+ gen_helper_sve_ldnf1hdu_le_r,
470
471
- gen_helper_sve_ldnf1hds_le_r,
472
- gen_helper_sve_ldnf1hss_le_r,
473
- gen_helper_sve_ldnf1ss_le_r,
474
- gen_helper_sve_ldnf1sdu_le_r,
475
+ gen_helper_sve_ldnf1hds_le_r,
476
+ gen_helper_sve_ldnf1hss_le_r,
477
+ gen_helper_sve_ldnf1ss_le_r,
478
+ gen_helper_sve_ldnf1sdu_le_r,
479
480
- gen_helper_sve_ldnf1bds_r,
481
- gen_helper_sve_ldnf1bss_r,
482
- gen_helper_sve_ldnf1bhs_r,
483
- gen_helper_sve_ldnf1dd_le_r },
484
+ gen_helper_sve_ldnf1bds_r,
485
+ gen_helper_sve_ldnf1bss_r,
486
+ gen_helper_sve_ldnf1bhs_r,
487
+ gen_helper_sve_ldnf1dd_le_r },
488
489
- /* Big-endian */
490
- { gen_helper_sve_ldnf1bb_r,
491
- gen_helper_sve_ldnf1bhu_r,
492
- gen_helper_sve_ldnf1bsu_r,
493
- gen_helper_sve_ldnf1bdu_r,
494
+ /* mte inactive, big-endian */
495
+ { gen_helper_sve_ldnf1bb_r,
496
+ gen_helper_sve_ldnf1bhu_r,
497
+ gen_helper_sve_ldnf1bsu_r,
498
+ gen_helper_sve_ldnf1bdu_r,
499
500
- gen_helper_sve_ldnf1sds_be_r,
501
- gen_helper_sve_ldnf1hh_be_r,
502
- gen_helper_sve_ldnf1hsu_be_r,
503
- gen_helper_sve_ldnf1hdu_be_r,
504
+ gen_helper_sve_ldnf1sds_be_r,
505
+ gen_helper_sve_ldnf1hh_be_r,
506
+ gen_helper_sve_ldnf1hsu_be_r,
507
+ gen_helper_sve_ldnf1hdu_be_r,
508
509
- gen_helper_sve_ldnf1hds_be_r,
510
- gen_helper_sve_ldnf1hss_be_r,
511
- gen_helper_sve_ldnf1ss_be_r,
512
- gen_helper_sve_ldnf1sdu_be_r,
513
+ gen_helper_sve_ldnf1hds_be_r,
514
+ gen_helper_sve_ldnf1hss_be_r,
515
+ gen_helper_sve_ldnf1ss_be_r,
516
+ gen_helper_sve_ldnf1sdu_be_r,
517
518
- gen_helper_sve_ldnf1bds_r,
519
- gen_helper_sve_ldnf1bss_r,
520
- gen_helper_sve_ldnf1bhs_r,
521
- gen_helper_sve_ldnf1dd_be_r },
522
+ gen_helper_sve_ldnf1bds_r,
523
+ gen_helper_sve_ldnf1bss_r,
524
+ gen_helper_sve_ldnf1bhs_r,
525
+ gen_helper_sve_ldnf1dd_be_r } },
526
+
527
+ { /* mte inactive, little-endian */
528
+ { gen_helper_sve_ldnf1bb_r_mte,
529
+ gen_helper_sve_ldnf1bhu_r_mte,
530
+ gen_helper_sve_ldnf1bsu_r_mte,
531
+ gen_helper_sve_ldnf1bdu_r_mte,
532
+
533
+ gen_helper_sve_ldnf1sds_le_r_mte,
534
+ gen_helper_sve_ldnf1hh_le_r_mte,
535
+ gen_helper_sve_ldnf1hsu_le_r_mte,
536
+ gen_helper_sve_ldnf1hdu_le_r_mte,
537
+
538
+ gen_helper_sve_ldnf1hds_le_r_mte,
539
+ gen_helper_sve_ldnf1hss_le_r_mte,
540
+ gen_helper_sve_ldnf1ss_le_r_mte,
541
+ gen_helper_sve_ldnf1sdu_le_r_mte,
542
+
543
+ gen_helper_sve_ldnf1bds_r_mte,
544
+ gen_helper_sve_ldnf1bss_r_mte,
545
+ gen_helper_sve_ldnf1bhs_r_mte,
546
+ gen_helper_sve_ldnf1dd_le_r_mte },
547
+
548
+ /* mte inactive, big-endian */
549
+ { gen_helper_sve_ldnf1bb_r_mte,
550
+ gen_helper_sve_ldnf1bhu_r_mte,
551
+ gen_helper_sve_ldnf1bsu_r_mte,
552
+ gen_helper_sve_ldnf1bdu_r_mte,
553
+
554
+ gen_helper_sve_ldnf1sds_be_r_mte,
555
+ gen_helper_sve_ldnf1hh_be_r_mte,
556
+ gen_helper_sve_ldnf1hsu_be_r_mte,
557
+ gen_helper_sve_ldnf1hdu_be_r_mte,
558
+
559
+ gen_helper_sve_ldnf1hds_be_r_mte,
560
+ gen_helper_sve_ldnf1hss_be_r_mte,
561
+ gen_helper_sve_ldnf1ss_be_r_mte,
562
+ gen_helper_sve_ldnf1sdu_be_r_mte,
563
+
564
+ gen_helper_sve_ldnf1bds_r_mte,
565
+ gen_helper_sve_ldnf1bss_r_mte,
566
+ gen_helper_sve_ldnf1bhs_r_mte,
567
+ gen_helper_sve_ldnf1dd_be_r_mte } },
568
};
569
570
if (sve_access_check(s)) {
571
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
572
TCGv_i64 addr = new_tmp_a64(s);
573
574
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
575
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
576
- fns[s->be_data == MO_BE][a->dtype]);
577
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
578
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
579
}
580
return true;
581
}
64
--
582
--
65
2.20.1
583
2.20.1
66
584
67
585
diff view generated by jsdifflib
1
From: Fredrik Strupe <fredrik@strupe.net>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
According to Arm ARM, VQDMULL is only valid when U=0, while having
3
We still need to handle tbi for user-only when mte is inactive.
4
U=1 is unallocated.
5
4
6
Signed-off-by: Fredrik Strupe <fredrik@strupe.net>
7
Fixes: 695272dcb976 ("target-arm: Handle UNDEF cases for Neon 3-regs-different-widths")
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-37-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
9
---
11
target/arm/translate.c | 2 +-
10
target/arm/translate-a64.h | 1 +
12
1 file changed, 1 insertion(+), 1 deletion(-)
11
target/arm/translate-a64.c | 2 +-
12
target/arm/translate-sve.c | 6 ++++--
13
3 files changed, 6 insertions(+), 3 deletions(-)
13
14
14
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate.c
17
--- a/target/arm/translate-a64.h
17
+++ b/target/arm/translate.c
18
+++ b/target/arm/translate-a64.h
18
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
19
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
19
{0, 0, 0, 0}, /* VMLSL */
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
20
{0, 0, 0, 9}, /* VQDMLSL */
21
unsigned int imms, unsigned int immr);
21
{0, 0, 0, 0}, /* Integer VMULL */
22
bool sve_access_check(DisasContext *s);
22
- {0, 0, 0, 1}, /* VQDMULL */
23
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
23
+ {0, 0, 0, 9}, /* VQDMULL */
24
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
24
{0, 0, 0, 0xa}, /* Polynomial VMULL */
25
bool tag_checked, int log2_size);
25
{0, 0, 0, 7}, /* Reserved: always UNDEF */
26
TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
26
};
27
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-a64.c
30
+++ b/target/arm/translate-a64.c
31
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
32
* of the write-back address.
33
*/
34
35
-static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
36
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
37
{
38
TCGv_i64 clean = new_tmp_a64(s);
39
#ifdef CONFIG_USER_ONLY
40
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-sve.c
43
+++ b/target/arm/translate-sve.c
44
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
45
* For e.g. LD4, there are not enough arguments to pass all 4
46
* registers as pointers, so encode the regno into the data field.
47
* For consistency, do this even for LD1.
48
- * TODO: mte_n check here while callers are updated.
49
*/
50
- if (mte_n && s->mte_active[0]) {
51
+ if (s->mte_active[0]) {
52
int msz = dtype_msz(dtype);
53
54
desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
55
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
56
desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
57
desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
58
desc <<= SVE_MTEDESC_SHIFT;
59
+ } else {
60
+ addr = clean_data_tbi(s, addr);
61
}
62
+
63
desc = simd_desc(vsz, vsz, zt | desc);
64
t_desc = tcg_const_i32(desc);
65
t_pg = tcg_temp_new_ptr();
27
--
66
--
28
2.20.1
67
2.20.1
29
68
30
69
diff view generated by jsdifflib
1
Convert the Neon "load/store single structure to one lane" insns to
1
From: Richard Henderson <richard.henderson@linaro.org>
2
decodetree.
3
2
4
As this is the last set of insns in the neon load/store group,
3
Because the elements are non-sequential, we cannot eliminate many
5
we can remove the whole disas_neon_ls_insn() function.
4
tests straight away like we can for sequential operations. But
5
we often have the PTE details handy, so we can test for Tagged.
6
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-38-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200430181003.21682-14-peter.maydell@linaro.org
10
---
11
---
11
target/arm/neon-ls.decode | 11 +++
12
target/arm/helper-sve.h | 285 ++++++++++++++++
12
target/arm/translate-neon.inc.c | 89 +++++++++++++++++++
13
target/arm/sve_helper.c | 185 +++++++++--
13
target/arm/translate.c | 147 --------------------------------
14
target/arm/translate-sve.c | 650 +++++++++++++++++++++++++------------
14
3 files changed, 100 insertions(+), 147 deletions(-)
15
3 files changed, 872 insertions(+), 248 deletions(-)
15
16
16
diff --git a/target/arm/neon-ls.decode b/target/arm/neon-ls.decode
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/neon-ls.decode
19
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/neon-ls.decode
20
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ VLDST_multiple 1111 0100 0 . l:1 0 rn:4 .... itype:4 size:2 align:2 rm:4 \
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldsds_le_zd, TCG_CALL_NO_WG,
21
22
DEF_HELPER_FLAGS_6(sve_ldsds_be_zd, TCG_CALL_NO_WG,
22
VLD_all_lanes 1111 0100 1 . 1 0 rn:4 .... 11 n:2 size:2 t:1 a:1 rm:4 \
23
void, env, ptr, ptr, ptr, tl, i32)
23
vd=%vd_dp
24
24
+
25
+DEF_HELPER_FLAGS_6(sve_ldbsu_zsu_mte, TCG_CALL_NO_WG,
25
+# Neon load/store single structure to one lane
26
+ void, env, ptr, ptr, ptr, tl, i32)
26
+%imm1_5_p1 5:1 !function=plus1
27
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu_mte, TCG_CALL_NO_WG,
27
+%imm1_6_p1 6:1 !function=plus1
28
+ void, env, ptr, ptr, ptr, tl, i32)
28
+
29
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zsu_mte, TCG_CALL_NO_WG,
29
+VLDST_single 1111 0100 1 . l:1 0 rn:4 .... 00 n:2 reg_idx:3 align:1 rm:4 \
30
+ void, env, ptr, ptr, ptr, tl, i32)
30
+ vd=%vd_dp size=0 stride=1
31
+DEF_HELPER_FLAGS_6(sve_ldss_le_zsu_mte, TCG_CALL_NO_WG,
31
+VLDST_single 1111 0100 1 . l:1 0 rn:4 .... 01 n:2 reg_idx:2 align:2 rm:4 \
32
+ void, env, ptr, ptr, ptr, tl, i32)
32
+ vd=%vd_dp size=1 stride=%imm1_5_p1
33
+DEF_HELPER_FLAGS_6(sve_ldss_be_zsu_mte, TCG_CALL_NO_WG,
33
+VLDST_single 1111 0100 1 . l:1 0 rn:4 .... 10 n:2 reg_idx:1 align:3 rm:4 \
34
+ void, env, ptr, ptr, ptr, tl, i32)
34
+ vd=%vd_dp size=2 stride=%imm1_6_p1
35
+DEF_HELPER_FLAGS_6(sve_ldbss_zsu_mte, TCG_CALL_NO_WG,
35
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
36
+ void, env, ptr, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zsu_mte, TCG_CALL_NO_WG,
38
+ void, env, ptr, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zsu_mte, TCG_CALL_NO_WG,
40
+ void, env, ptr, ptr, ptr, tl, i32)
41
+
42
+DEF_HELPER_FLAGS_6(sve_ldbsu_zss_mte, TCG_CALL_NO_WG,
43
+ void, env, ptr, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zss_mte, TCG_CALL_NO_WG,
45
+ void, env, ptr, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zss_mte, TCG_CALL_NO_WG,
47
+ void, env, ptr, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_6(sve_ldss_le_zss_mte, TCG_CALL_NO_WG,
49
+ void, env, ptr, ptr, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_6(sve_ldss_be_zss_mte, TCG_CALL_NO_WG,
51
+ void, env, ptr, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_6(sve_ldbss_zss_mte, TCG_CALL_NO_WG,
53
+ void, env, ptr, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zss_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zss_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_6(sve_ldbdu_zsu_mte, TCG_CALL_NO_WG,
60
+ void, env, ptr, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zsu_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zsu_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zsu_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, ptr, ptr, tl, i32)
67
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zsu_mte, TCG_CALL_NO_WG,
68
+ void, env, ptr, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_6(sve_lddd_le_zsu_mte, TCG_CALL_NO_WG,
70
+ void, env, ptr, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_6(sve_lddd_be_zsu_mte, TCG_CALL_NO_WG,
72
+ void, env, ptr, ptr, ptr, tl, i32)
73
+DEF_HELPER_FLAGS_6(sve_ldbds_zsu_mte, TCG_CALL_NO_WG,
74
+ void, env, ptr, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zsu_mte, TCG_CALL_NO_WG,
76
+ void, env, ptr, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zsu_mte, TCG_CALL_NO_WG,
78
+ void, env, ptr, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zsu_mte, TCG_CALL_NO_WG,
80
+ void, env, ptr, ptr, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zsu_mte, TCG_CALL_NO_WG,
82
+ void, env, ptr, ptr, ptr, tl, i32)
83
+
84
+DEF_HELPER_FLAGS_6(sve_ldbdu_zss_mte, TCG_CALL_NO_WG,
85
+ void, env, ptr, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zss_mte, TCG_CALL_NO_WG,
87
+ void, env, ptr, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zss_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, ptr, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zss_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zss_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_6(sve_lddd_le_zss_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_6(sve_lddd_be_zss_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_6(sve_ldbds_zss_mte, TCG_CALL_NO_WG,
99
+ void, env, ptr, ptr, ptr, tl, i32)
100
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zss_mte, TCG_CALL_NO_WG,
101
+ void, env, ptr, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zss_mte, TCG_CALL_NO_WG,
103
+ void, env, ptr, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zss_mte, TCG_CALL_NO_WG,
105
+ void, env, ptr, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zss_mte, TCG_CALL_NO_WG,
107
+ void, env, ptr, ptr, ptr, tl, i32)
108
+
109
+DEF_HELPER_FLAGS_6(sve_ldbdu_zd_mte, TCG_CALL_NO_WG,
110
+ void, env, ptr, ptr, ptr, tl, i32)
111
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zd_mte, TCG_CALL_NO_WG,
112
+ void, env, ptr, ptr, ptr, tl, i32)
113
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zd_mte, TCG_CALL_NO_WG,
114
+ void, env, ptr, ptr, ptr, tl, i32)
115
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zd_mte, TCG_CALL_NO_WG,
116
+ void, env, ptr, ptr, ptr, tl, i32)
117
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zd_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, ptr, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_6(sve_lddd_le_zd_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, ptr, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_6(sve_lddd_be_zd_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, ptr, ptr, tl, i32)
123
+DEF_HELPER_FLAGS_6(sve_ldbds_zd_mte, TCG_CALL_NO_WG,
124
+ void, env, ptr, ptr, ptr, tl, i32)
125
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zd_mte, TCG_CALL_NO_WG,
126
+ void, env, ptr, ptr, ptr, tl, i32)
127
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zd_mte, TCG_CALL_NO_WG,
128
+ void, env, ptr, ptr, ptr, tl, i32)
129
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zd_mte, TCG_CALL_NO_WG,
130
+ void, env, ptr, ptr, ptr, tl, i32)
131
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zd_mte, TCG_CALL_NO_WG,
132
+ void, env, ptr, ptr, ptr, tl, i32)
133
+
134
DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu, TCG_CALL_NO_WG,
135
void, env, ptr, ptr, ptr, tl, i32)
136
DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu, TCG_CALL_NO_WG,
137
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd, TCG_CALL_NO_WG,
138
DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd, TCG_CALL_NO_WG,
139
void, env, ptr, ptr, ptr, tl, i32)
140
141
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu_mte, TCG_CALL_NO_WG,
142
+ void, env, ptr, ptr, ptr, tl, i32)
143
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu_mte, TCG_CALL_NO_WG,
144
+ void, env, ptr, ptr, ptr, tl, i32)
145
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zsu_mte, TCG_CALL_NO_WG,
146
+ void, env, ptr, ptr, ptr, tl, i32)
147
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zsu_mte, TCG_CALL_NO_WG,
148
+ void, env, ptr, ptr, ptr, tl, i32)
149
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zsu_mte, TCG_CALL_NO_WG,
150
+ void, env, ptr, ptr, ptr, tl, i32)
151
+DEF_HELPER_FLAGS_6(sve_ldffbss_zsu_mte, TCG_CALL_NO_WG,
152
+ void, env, ptr, ptr, ptr, tl, i32)
153
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zsu_mte, TCG_CALL_NO_WG,
154
+ void, env, ptr, ptr, ptr, tl, i32)
155
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zsu_mte, TCG_CALL_NO_WG,
156
+ void, env, ptr, ptr, ptr, tl, i32)
157
+
158
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zss_mte, TCG_CALL_NO_WG,
159
+ void, env, ptr, ptr, ptr, tl, i32)
160
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zss_mte, TCG_CALL_NO_WG,
161
+ void, env, ptr, ptr, ptr, tl, i32)
162
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zss_mte, TCG_CALL_NO_WG,
163
+ void, env, ptr, ptr, ptr, tl, i32)
164
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zss_mte, TCG_CALL_NO_WG,
165
+ void, env, ptr, ptr, ptr, tl, i32)
166
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zss_mte, TCG_CALL_NO_WG,
167
+ void, env, ptr, ptr, ptr, tl, i32)
168
+DEF_HELPER_FLAGS_6(sve_ldffbss_zss_mte, TCG_CALL_NO_WG,
169
+ void, env, ptr, ptr, ptr, tl, i32)
170
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zss_mte, TCG_CALL_NO_WG,
171
+ void, env, ptr, ptr, ptr, tl, i32)
172
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zss_mte, TCG_CALL_NO_WG,
173
+ void, env, ptr, ptr, ptr, tl, i32)
174
+
175
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zsu_mte, TCG_CALL_NO_WG,
176
+ void, env, ptr, ptr, ptr, tl, i32)
177
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zsu_mte, TCG_CALL_NO_WG,
178
+ void, env, ptr, ptr, ptr, tl, i32)
179
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zsu_mte, TCG_CALL_NO_WG,
180
+ void, env, ptr, ptr, ptr, tl, i32)
181
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zsu_mte, TCG_CALL_NO_WG,
182
+ void, env, ptr, ptr, ptr, tl, i32)
183
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zsu_mte, TCG_CALL_NO_WG,
184
+ void, env, ptr, ptr, ptr, tl, i32)
185
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zsu_mte, TCG_CALL_NO_WG,
186
+ void, env, ptr, ptr, ptr, tl, i32)
187
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zsu_mte, TCG_CALL_NO_WG,
188
+ void, env, ptr, ptr, ptr, tl, i32)
189
+DEF_HELPER_FLAGS_6(sve_ldffbds_zsu_mte, TCG_CALL_NO_WG,
190
+ void, env, ptr, ptr, ptr, tl, i32)
191
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zsu_mte, TCG_CALL_NO_WG,
192
+ void, env, ptr, ptr, ptr, tl, i32)
193
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zsu_mte, TCG_CALL_NO_WG,
194
+ void, env, ptr, ptr, ptr, tl, i32)
195
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zsu_mte, TCG_CALL_NO_WG,
196
+ void, env, ptr, ptr, ptr, tl, i32)
197
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zsu_mte, TCG_CALL_NO_WG,
198
+ void, env, ptr, ptr, ptr, tl, i32)
199
+
200
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zss_mte, TCG_CALL_NO_WG,
201
+ void, env, ptr, ptr, ptr, tl, i32)
202
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zss_mte, TCG_CALL_NO_WG,
203
+ void, env, ptr, ptr, ptr, tl, i32)
204
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zss_mte, TCG_CALL_NO_WG,
205
+ void, env, ptr, ptr, ptr, tl, i32)
206
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zss_mte, TCG_CALL_NO_WG,
207
+ void, env, ptr, ptr, ptr, tl, i32)
208
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zss_mte, TCG_CALL_NO_WG,
209
+ void, env, ptr, ptr, ptr, tl, i32)
210
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zss_mte, TCG_CALL_NO_WG,
211
+ void, env, ptr, ptr, ptr, tl, i32)
212
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zss_mte, TCG_CALL_NO_WG,
213
+ void, env, ptr, ptr, ptr, tl, i32)
214
+DEF_HELPER_FLAGS_6(sve_ldffbds_zss_mte, TCG_CALL_NO_WG,
215
+ void, env, ptr, ptr, ptr, tl, i32)
216
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zss_mte, TCG_CALL_NO_WG,
217
+ void, env, ptr, ptr, ptr, tl, i32)
218
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zss_mte, TCG_CALL_NO_WG,
219
+ void, env, ptr, ptr, ptr, tl, i32)
220
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zss_mte, TCG_CALL_NO_WG,
221
+ void, env, ptr, ptr, ptr, tl, i32)
222
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zss_mte, TCG_CALL_NO_WG,
223
+ void, env, ptr, ptr, ptr, tl, i32)
224
+
225
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zd_mte, TCG_CALL_NO_WG,
226
+ void, env, ptr, ptr, ptr, tl, i32)
227
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zd_mte, TCG_CALL_NO_WG,
228
+ void, env, ptr, ptr, ptr, tl, i32)
229
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zd_mte, TCG_CALL_NO_WG,
230
+ void, env, ptr, ptr, ptr, tl, i32)
231
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zd_mte, TCG_CALL_NO_WG,
232
+ void, env, ptr, ptr, ptr, tl, i32)
233
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zd_mte, TCG_CALL_NO_WG,
234
+ void, env, ptr, ptr, ptr, tl, i32)
235
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zd_mte, TCG_CALL_NO_WG,
236
+ void, env, ptr, ptr, ptr, tl, i32)
237
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zd_mte, TCG_CALL_NO_WG,
238
+ void, env, ptr, ptr, ptr, tl, i32)
239
+DEF_HELPER_FLAGS_6(sve_ldffbds_zd_mte, TCG_CALL_NO_WG,
240
+ void, env, ptr, ptr, ptr, tl, i32)
241
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zd_mte, TCG_CALL_NO_WG,
242
+ void, env, ptr, ptr, ptr, tl, i32)
243
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zd_mte, TCG_CALL_NO_WG,
244
+ void, env, ptr, ptr, ptr, tl, i32)
245
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd_mte, TCG_CALL_NO_WG,
246
+ void, env, ptr, ptr, ptr, tl, i32)
247
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd_mte, TCG_CALL_NO_WG,
248
+ void, env, ptr, ptr, ptr, tl, i32)
249
+
250
DEF_HELPER_FLAGS_6(sve_stbs_zsu, TCG_CALL_NO_WG,
251
void, env, ptr, ptr, ptr, tl, i32)
252
DEF_HELPER_FLAGS_6(sve_sths_le_zsu, TCG_CALL_NO_WG,
253
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd, TCG_CALL_NO_WG,
254
DEF_HELPER_FLAGS_6(sve_stdd_be_zd, TCG_CALL_NO_WG,
255
void, env, ptr, ptr, ptr, tl, i32)
256
257
+DEF_HELPER_FLAGS_6(sve_stbs_zsu_mte, TCG_CALL_NO_WG,
258
+ void, env, ptr, ptr, ptr, tl, i32)
259
+DEF_HELPER_FLAGS_6(sve_sths_le_zsu_mte, TCG_CALL_NO_WG,
260
+ void, env, ptr, ptr, ptr, tl, i32)
261
+DEF_HELPER_FLAGS_6(sve_sths_be_zsu_mte, TCG_CALL_NO_WG,
262
+ void, env, ptr, ptr, ptr, tl, i32)
263
+DEF_HELPER_FLAGS_6(sve_stss_le_zsu_mte, TCG_CALL_NO_WG,
264
+ void, env, ptr, ptr, ptr, tl, i32)
265
+DEF_HELPER_FLAGS_6(sve_stss_be_zsu_mte, TCG_CALL_NO_WG,
266
+ void, env, ptr, ptr, ptr, tl, i32)
267
+
268
+DEF_HELPER_FLAGS_6(sve_stbs_zss_mte, TCG_CALL_NO_WG,
269
+ void, env, ptr, ptr, ptr, tl, i32)
270
+DEF_HELPER_FLAGS_6(sve_sths_le_zss_mte, TCG_CALL_NO_WG,
271
+ void, env, ptr, ptr, ptr, tl, i32)
272
+DEF_HELPER_FLAGS_6(sve_sths_be_zss_mte, TCG_CALL_NO_WG,
273
+ void, env, ptr, ptr, ptr, tl, i32)
274
+DEF_HELPER_FLAGS_6(sve_stss_le_zss_mte, TCG_CALL_NO_WG,
275
+ void, env, ptr, ptr, ptr, tl, i32)
276
+DEF_HELPER_FLAGS_6(sve_stss_be_zss_mte, TCG_CALL_NO_WG,
277
+ void, env, ptr, ptr, ptr, tl, i32)
278
+
279
+DEF_HELPER_FLAGS_6(sve_stbd_zsu_mte, TCG_CALL_NO_WG,
280
+ void, env, ptr, ptr, ptr, tl, i32)
281
+DEF_HELPER_FLAGS_6(sve_sthd_le_zsu_mte, TCG_CALL_NO_WG,
282
+ void, env, ptr, ptr, ptr, tl, i32)
283
+DEF_HELPER_FLAGS_6(sve_sthd_be_zsu_mte, TCG_CALL_NO_WG,
284
+ void, env, ptr, ptr, ptr, tl, i32)
285
+DEF_HELPER_FLAGS_6(sve_stsd_le_zsu_mte, TCG_CALL_NO_WG,
286
+ void, env, ptr, ptr, ptr, tl, i32)
287
+DEF_HELPER_FLAGS_6(sve_stsd_be_zsu_mte, TCG_CALL_NO_WG,
288
+ void, env, ptr, ptr, ptr, tl, i32)
289
+DEF_HELPER_FLAGS_6(sve_stdd_le_zsu_mte, TCG_CALL_NO_WG,
290
+ void, env, ptr, ptr, ptr, tl, i32)
291
+DEF_HELPER_FLAGS_6(sve_stdd_be_zsu_mte, TCG_CALL_NO_WG,
292
+ void, env, ptr, ptr, ptr, tl, i32)
293
+
294
+DEF_HELPER_FLAGS_6(sve_stbd_zss_mte, TCG_CALL_NO_WG,
295
+ void, env, ptr, ptr, ptr, tl, i32)
296
+DEF_HELPER_FLAGS_6(sve_sthd_le_zss_mte, TCG_CALL_NO_WG,
297
+ void, env, ptr, ptr, ptr, tl, i32)
298
+DEF_HELPER_FLAGS_6(sve_sthd_be_zss_mte, TCG_CALL_NO_WG,
299
+ void, env, ptr, ptr, ptr, tl, i32)
300
+DEF_HELPER_FLAGS_6(sve_stsd_le_zss_mte, TCG_CALL_NO_WG,
301
+ void, env, ptr, ptr, ptr, tl, i32)
302
+DEF_HELPER_FLAGS_6(sve_stsd_be_zss_mte, TCG_CALL_NO_WG,
303
+ void, env, ptr, ptr, ptr, tl, i32)
304
+DEF_HELPER_FLAGS_6(sve_stdd_le_zss_mte, TCG_CALL_NO_WG,
305
+ void, env, ptr, ptr, ptr, tl, i32)
306
+DEF_HELPER_FLAGS_6(sve_stdd_be_zss_mte, TCG_CALL_NO_WG,
307
+ void, env, ptr, ptr, ptr, tl, i32)
308
+
309
+DEF_HELPER_FLAGS_6(sve_stbd_zd_mte, TCG_CALL_NO_WG,
310
+ void, env, ptr, ptr, ptr, tl, i32)
311
+DEF_HELPER_FLAGS_6(sve_sthd_le_zd_mte, TCG_CALL_NO_WG,
312
+ void, env, ptr, ptr, ptr, tl, i32)
313
+DEF_HELPER_FLAGS_6(sve_sthd_be_zd_mte, TCG_CALL_NO_WG,
314
+ void, env, ptr, ptr, ptr, tl, i32)
315
+DEF_HELPER_FLAGS_6(sve_stsd_le_zd_mte, TCG_CALL_NO_WG,
316
+ void, env, ptr, ptr, ptr, tl, i32)
317
+DEF_HELPER_FLAGS_6(sve_stsd_be_zd_mte, TCG_CALL_NO_WG,
318
+ void, env, ptr, ptr, ptr, tl, i32)
319
+DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG,
320
+ void, env, ptr, ptr, ptr, tl, i32)
321
+DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG,
322
+ void, env, ptr, ptr, ptr, tl, i32)
323
+
324
DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
325
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
36
index XXXXXXX..XXXXXXX 100644
326
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/translate-neon.inc.c
327
--- a/target/arm/sve_helper.c
38
+++ b/target/arm/translate-neon.inc.c
328
+++ b/target/arm/sve_helper.c
39
@@ -XXX,XX +XXX,XX @@
329
@@ -XXX,XX +XXX,XX @@ static target_ulong off_zd_d(void *reg, intptr_t reg_ofs)
40
* It might be possible to convert it to a standalone .c file eventually.
330
static inline QEMU_ALWAYS_INLINE
331
void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
332
target_ulong base, uint32_t desc, uintptr_t retaddr,
333
- int esize, int msize, zreg_off_fn *off_fn,
334
+ uint32_t mtedesc, int esize, int msize,
335
+ zreg_off_fn *off_fn,
336
sve_ldst1_host_fn *host_fn,
337
sve_ldst1_tlb_fn *tlb_fn)
338
{
339
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
340
cpu_check_watchpoint(env_cpu(env), addr, msize,
341
info.attrs, BP_MEM_READ, retaddr);
342
}
343
- /* TODO: MTE check */
344
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
345
+ mte_check1(env, mtedesc, addr, retaddr);
346
+ }
347
host_fn(&scratch, reg_off, info.host);
348
} else {
349
/* Element crosses the page boundary. */
350
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
351
msize, info.attrs,
352
BP_MEM_READ, retaddr);
353
}
354
- /* TODO: MTE check */
355
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
356
+ mte_check1(env, mtedesc, addr, retaddr);
357
+ }
358
tlb_fn(env, &scratch, reg_off, addr, retaddr);
359
}
360
}
361
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
362
memcpy(vd, &scratch, reg_max);
363
}
364
365
+static inline QEMU_ALWAYS_INLINE
366
+void sve_ld1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
367
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
368
+ int esize, int msize, zreg_off_fn *off_fn,
369
+ sve_ldst1_host_fn *host_fn,
370
+ sve_ldst1_tlb_fn *tlb_fn)
371
+{
372
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
373
+ /* Remove mtedesc from the normal sve descriptor. */
374
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
375
+
376
+ /*
377
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
378
+ * offset base entirely over the address space hole to change the
379
+ * pointer tag, or change the bit55 selector. So we could here
380
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
381
+ */
382
+ sve_ld1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
383
+ esize, msize, off_fn, host_fn, tlb_fn);
384
+}
385
+
386
#define DO_LD1_ZPZ_S(MEM, OFS, MSZ) \
387
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
388
void *vm, target_ulong base, uint32_t desc) \
389
{ \
390
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
391
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
392
off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
393
+} \
394
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
395
+ void *vm, target_ulong base, uint32_t desc) \
396
+{ \
397
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
398
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
399
}
400
401
#define DO_LD1_ZPZ_D(MEM, OFS, MSZ) \
402
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
403
void *vm, target_ulong base, uint32_t desc) \
404
{ \
405
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
406
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
407
off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
408
+} \
409
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
410
+ void *vm, target_ulong base, uint32_t desc) \
411
+{ \
412
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
413
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
414
}
415
416
DO_LD1_ZPZ_S(bsu, zsu, MO_8)
417
@@ -XXX,XX +XXX,XX @@ DO_LD1_ZPZ_D(dd_be, zd, MO_64)
418
static inline QEMU_ALWAYS_INLINE
419
void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
420
target_ulong base, uint32_t desc, uintptr_t retaddr,
421
- const int esz, const int msz, zreg_off_fn *off_fn,
422
+ uint32_t mtedesc, const int esz, const int msz,
423
+ zreg_off_fn *off_fn,
424
sve_ldst1_host_fn *host_fn,
425
sve_ldst1_tlb_fn *tlb_fn)
426
{
427
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
428
* Probe the first element, allowing faults.
429
*/
430
addr = base + (off_fn(vm, reg_off) << scale);
431
+ if (mtedesc) {
432
+ mte_check1(env, mtedesc, addr, retaddr);
433
+ }
434
tlb_fn(env, vd, reg_off, addr, retaddr);
435
436
/* After any fault, zero the other elements. */
437
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
438
(env_cpu(env), addr, msize) & BP_MEM_READ)) {
439
goto fault;
440
}
441
- /* TODO: MTE check. */
442
+ if (mtedesc &&
443
+ arm_tlb_mte_tagged(&info.attrs) &&
444
+ !mte_probe1(env, mtedesc, addr)) {
445
+ goto fault;
446
+ }
447
448
host_fn(vd, reg_off, info.host);
449
}
450
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
451
record_fault(env, reg_off, reg_max);
452
}
453
454
-#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
455
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
456
- void *vm, target_ulong base, uint32_t desc) \
457
-{ \
458
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
459
- off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
460
+static inline QEMU_ALWAYS_INLINE
461
+void sve_ldff1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
462
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
463
+ const int esz, const int msz,
464
+ zreg_off_fn *off_fn,
465
+ sve_ldst1_host_fn *host_fn,
466
+ sve_ldst1_tlb_fn *tlb_fn)
467
+{
468
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
469
+ /* Remove mtedesc from the normal sve descriptor. */
470
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
471
+
472
+ /*
473
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
474
+ * offset base entirely over the address space hole to change the
475
+ * pointer tag, or change the bit55 selector. So we could here
476
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
477
+ */
478
+ sve_ldff1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
479
+ esz, msz, off_fn, host_fn, tlb_fn);
480
}
481
482
-#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
483
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
484
- void *vm, target_ulong base, uint32_t desc) \
485
-{ \
486
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
487
- off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
488
+#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
489
+void HELPER(sve_ldff##MEM##_##OFS) \
490
+ (CPUARMState *env, void *vd, void *vg, \
491
+ void *vm, target_ulong base, uint32_t desc) \
492
+{ \
493
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_32, MSZ, \
494
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
495
+} \
496
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
497
+ (CPUARMState *env, void *vd, void *vg, \
498
+ void *vm, target_ulong base, uint32_t desc) \
499
+{ \
500
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
501
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
502
+}
503
+
504
+#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
505
+void HELPER(sve_ldff##MEM##_##OFS) \
506
+ (CPUARMState *env, void *vd, void *vg, \
507
+ void *vm, target_ulong base, uint32_t desc) \
508
+{ \
509
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_64, MSZ, \
510
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
511
+} \
512
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
513
+ (CPUARMState *env, void *vd, void *vg, \
514
+ void *vm, target_ulong base, uint32_t desc) \
515
+{ \
516
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
517
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
518
}
519
520
DO_LDFF1_ZPZ_S(bsu, zsu, MO_8)
521
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_ZPZ_D(dd_be, zd, MO_64)
522
static inline QEMU_ALWAYS_INLINE
523
void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
524
target_ulong base, uint32_t desc, uintptr_t retaddr,
525
- int esize, int msize, zreg_off_fn *off_fn,
526
+ uint32_t mtedesc, int esize, int msize,
527
+ zreg_off_fn *off_fn,
528
sve_ldst1_host_fn *host_fn,
529
sve_ldst1_tlb_fn *tlb_fn)
530
{
531
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
532
cpu_check_watchpoint(env_cpu(env), addr, msize,
533
info.attrs, BP_MEM_WRITE, retaddr);
534
}
535
- /* TODO: MTE check. */
536
+
537
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
538
+ mte_check1(env, mtedesc, addr, retaddr);
539
+ }
540
}
541
i += 1;
542
reg_off += esize;
543
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
544
} while (reg_off < reg_max);
545
}
546
547
-#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
548
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
549
- void *vm, target_ulong base, uint32_t desc) \
550
-{ \
551
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
552
- off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
553
+static inline QEMU_ALWAYS_INLINE
554
+void sve_st1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
555
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
556
+ int esize, int msize, zreg_off_fn *off_fn,
557
+ sve_ldst1_host_fn *host_fn,
558
+ sve_ldst1_tlb_fn *tlb_fn)
559
+{
560
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
561
+ /* Remove mtedesc from the normal sve descriptor. */
562
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
563
+
564
+ /*
565
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
566
+ * offset base entirely over the address space hole to change the
567
+ * pointer tag, or change the bit55 selector. So we could here
568
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
569
+ */
570
+ sve_st1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
571
+ esize, msize, off_fn, host_fn, tlb_fn);
572
}
573
574
-#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
575
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
576
+#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
577
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
578
void *vm, target_ulong base, uint32_t desc) \
579
-{ \
580
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
581
- off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
582
+{ \
583
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
584
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
585
+} \
586
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
587
+ void *vm, target_ulong base, uint32_t desc) \
588
+{ \
589
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
590
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
591
+}
592
+
593
+#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
594
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
595
+ void *vm, target_ulong base, uint32_t desc) \
596
+{ \
597
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
598
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
599
+} \
600
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
601
+ void *vm, target_ulong base, uint32_t desc) \
602
+{ \
603
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
604
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
605
}
606
607
DO_ST1_ZPZ_S(bs, zsu, MO_8)
608
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
609
index XXXXXXX..XXXXXXX 100644
610
--- a/target/arm/translate-sve.c
611
+++ b/target/arm/translate-sve.c
612
@@ -XXX,XX +XXX,XX @@ static bool trans_ST_zpri(DisasContext *s, arg_rpri_store *a)
41
*/
613
*/
42
614
43
+static inline int plus1(DisasContext *s, int x)
615
static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
44
+{
616
- int scale, TCGv_i64 scalar, int msz,
45
+ return x + 1;
617
+ int scale, TCGv_i64 scalar, int msz, bool is_write,
46
+}
618
gen_helper_gvec_mem_scatter *fn)
47
+
619
{
48
/* Include the generated Neon decoder */
620
unsigned vsz = vec_full_reg_size(s);
49
#include "decode-neon-dp.inc.c"
621
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
50
#include "decode-neon-ls.inc.c"
622
TCGv_ptr t_pg = tcg_temp_new_ptr();
51
@@ -XXX,XX +XXX,XX @@ static bool trans_VLD_all_lanes(DisasContext *s, arg_VLD_all_lanes *a)
623
TCGv_ptr t_zt = tcg_temp_new_ptr();
52
624
TCGv_i32 t_desc;
625
- int desc;
626
+ int desc = 0;
627
628
+ if (s->mte_active[0]) {
629
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
630
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
631
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
632
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
633
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
634
+ desc <<= SVE_MTEDESC_SHIFT;
635
+ }
636
desc = simd_desc(vsz, vsz, scale);
637
t_desc = tcg_const_i32(desc);
638
639
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
640
tcg_temp_free_i32(t_desc);
641
}
642
643
-/* Indexed by [be][ff][xs][u][msz]. */
644
-static gen_helper_gvec_mem_scatter * const gather_load_fn32[2][2][2][2][3] = {
645
- /* Little-endian */
646
- { { { { gen_helper_sve_ldbss_zsu,
647
- gen_helper_sve_ldhss_le_zsu,
648
- NULL, },
649
- { gen_helper_sve_ldbsu_zsu,
650
- gen_helper_sve_ldhsu_le_zsu,
651
- gen_helper_sve_ldss_le_zsu, } },
652
- { { gen_helper_sve_ldbss_zss,
653
- gen_helper_sve_ldhss_le_zss,
654
- NULL, },
655
- { gen_helper_sve_ldbsu_zss,
656
- gen_helper_sve_ldhsu_le_zss,
657
- gen_helper_sve_ldss_le_zss, } } },
658
+/* Indexed by [mte][be][ff][xs][u][msz]. */
659
+static gen_helper_gvec_mem_scatter * const
660
+gather_load_fn32[2][2][2][2][2][3] = {
661
+ { /* MTE Inactive */
662
+ { /* Little-endian */
663
+ { { { gen_helper_sve_ldbss_zsu,
664
+ gen_helper_sve_ldhss_le_zsu,
665
+ NULL, },
666
+ { gen_helper_sve_ldbsu_zsu,
667
+ gen_helper_sve_ldhsu_le_zsu,
668
+ gen_helper_sve_ldss_le_zsu, } },
669
+ { { gen_helper_sve_ldbss_zss,
670
+ gen_helper_sve_ldhss_le_zss,
671
+ NULL, },
672
+ { gen_helper_sve_ldbsu_zss,
673
+ gen_helper_sve_ldhsu_le_zss,
674
+ gen_helper_sve_ldss_le_zss, } } },
675
676
- /* First-fault */
677
- { { { gen_helper_sve_ldffbss_zsu,
678
- gen_helper_sve_ldffhss_le_zsu,
679
- NULL, },
680
- { gen_helper_sve_ldffbsu_zsu,
681
- gen_helper_sve_ldffhsu_le_zsu,
682
- gen_helper_sve_ldffss_le_zsu, } },
683
- { { gen_helper_sve_ldffbss_zss,
684
- gen_helper_sve_ldffhss_le_zss,
685
- NULL, },
686
- { gen_helper_sve_ldffbsu_zss,
687
- gen_helper_sve_ldffhsu_le_zss,
688
- gen_helper_sve_ldffss_le_zss, } } } },
689
+ /* First-fault */
690
+ { { { gen_helper_sve_ldffbss_zsu,
691
+ gen_helper_sve_ldffhss_le_zsu,
692
+ NULL, },
693
+ { gen_helper_sve_ldffbsu_zsu,
694
+ gen_helper_sve_ldffhsu_le_zsu,
695
+ gen_helper_sve_ldffss_le_zsu, } },
696
+ { { gen_helper_sve_ldffbss_zss,
697
+ gen_helper_sve_ldffhss_le_zss,
698
+ NULL, },
699
+ { gen_helper_sve_ldffbsu_zss,
700
+ gen_helper_sve_ldffhsu_le_zss,
701
+ gen_helper_sve_ldffss_le_zss, } } } },
702
703
- /* Big-endian */
704
- { { { { gen_helper_sve_ldbss_zsu,
705
- gen_helper_sve_ldhss_be_zsu,
706
- NULL, },
707
- { gen_helper_sve_ldbsu_zsu,
708
- gen_helper_sve_ldhsu_be_zsu,
709
- gen_helper_sve_ldss_be_zsu, } },
710
- { { gen_helper_sve_ldbss_zss,
711
- gen_helper_sve_ldhss_be_zss,
712
- NULL, },
713
- { gen_helper_sve_ldbsu_zss,
714
- gen_helper_sve_ldhsu_be_zss,
715
- gen_helper_sve_ldss_be_zss, } } },
716
+ { /* Big-endian */
717
+ { { { gen_helper_sve_ldbss_zsu,
718
+ gen_helper_sve_ldhss_be_zsu,
719
+ NULL, },
720
+ { gen_helper_sve_ldbsu_zsu,
721
+ gen_helper_sve_ldhsu_be_zsu,
722
+ gen_helper_sve_ldss_be_zsu, } },
723
+ { { gen_helper_sve_ldbss_zss,
724
+ gen_helper_sve_ldhss_be_zss,
725
+ NULL, },
726
+ { gen_helper_sve_ldbsu_zss,
727
+ gen_helper_sve_ldhsu_be_zss,
728
+ gen_helper_sve_ldss_be_zss, } } },
729
730
- /* First-fault */
731
- { { { gen_helper_sve_ldffbss_zsu,
732
- gen_helper_sve_ldffhss_be_zsu,
733
- NULL, },
734
- { gen_helper_sve_ldffbsu_zsu,
735
- gen_helper_sve_ldffhsu_be_zsu,
736
- gen_helper_sve_ldffss_be_zsu, } },
737
- { { gen_helper_sve_ldffbss_zss,
738
- gen_helper_sve_ldffhss_be_zss,
739
- NULL, },
740
- { gen_helper_sve_ldffbsu_zss,
741
- gen_helper_sve_ldffhsu_be_zss,
742
- gen_helper_sve_ldffss_be_zss, } } } },
743
+ /* First-fault */
744
+ { { { gen_helper_sve_ldffbss_zsu,
745
+ gen_helper_sve_ldffhss_be_zsu,
746
+ NULL, },
747
+ { gen_helper_sve_ldffbsu_zsu,
748
+ gen_helper_sve_ldffhsu_be_zsu,
749
+ gen_helper_sve_ldffss_be_zsu, } },
750
+ { { gen_helper_sve_ldffbss_zss,
751
+ gen_helper_sve_ldffhss_be_zss,
752
+ NULL, },
753
+ { gen_helper_sve_ldffbsu_zss,
754
+ gen_helper_sve_ldffhsu_be_zss,
755
+ gen_helper_sve_ldffss_be_zss, } } } } },
756
+ { /* MTE Active */
757
+ { /* Little-endian */
758
+ { { { gen_helper_sve_ldbss_zsu_mte,
759
+ gen_helper_sve_ldhss_le_zsu_mte,
760
+ NULL, },
761
+ { gen_helper_sve_ldbsu_zsu_mte,
762
+ gen_helper_sve_ldhsu_le_zsu_mte,
763
+ gen_helper_sve_ldss_le_zsu_mte, } },
764
+ { { gen_helper_sve_ldbss_zss_mte,
765
+ gen_helper_sve_ldhss_le_zss_mte,
766
+ NULL, },
767
+ { gen_helper_sve_ldbsu_zss_mte,
768
+ gen_helper_sve_ldhsu_le_zss_mte,
769
+ gen_helper_sve_ldss_le_zss_mte, } } },
770
+
771
+ /* First-fault */
772
+ { { { gen_helper_sve_ldffbss_zsu_mte,
773
+ gen_helper_sve_ldffhss_le_zsu_mte,
774
+ NULL, },
775
+ { gen_helper_sve_ldffbsu_zsu_mte,
776
+ gen_helper_sve_ldffhsu_le_zsu_mte,
777
+ gen_helper_sve_ldffss_le_zsu_mte, } },
778
+ { { gen_helper_sve_ldffbss_zss_mte,
779
+ gen_helper_sve_ldffhss_le_zss_mte,
780
+ NULL, },
781
+ { gen_helper_sve_ldffbsu_zss_mte,
782
+ gen_helper_sve_ldffhsu_le_zss_mte,
783
+ gen_helper_sve_ldffss_le_zss_mte, } } } },
784
+
785
+ { /* Big-endian */
786
+ { { { gen_helper_sve_ldbss_zsu_mte,
787
+ gen_helper_sve_ldhss_be_zsu_mte,
788
+ NULL, },
789
+ { gen_helper_sve_ldbsu_zsu_mte,
790
+ gen_helper_sve_ldhsu_be_zsu_mte,
791
+ gen_helper_sve_ldss_be_zsu_mte, } },
792
+ { { gen_helper_sve_ldbss_zss_mte,
793
+ gen_helper_sve_ldhss_be_zss_mte,
794
+ NULL, },
795
+ { gen_helper_sve_ldbsu_zss_mte,
796
+ gen_helper_sve_ldhsu_be_zss_mte,
797
+ gen_helper_sve_ldss_be_zss_mte, } } },
798
+
799
+ /* First-fault */
800
+ { { { gen_helper_sve_ldffbss_zsu_mte,
801
+ gen_helper_sve_ldffhss_be_zsu_mte,
802
+ NULL, },
803
+ { gen_helper_sve_ldffbsu_zsu_mte,
804
+ gen_helper_sve_ldffhsu_be_zsu_mte,
805
+ gen_helper_sve_ldffss_be_zsu_mte, } },
806
+ { { gen_helper_sve_ldffbss_zss_mte,
807
+ gen_helper_sve_ldffhss_be_zss_mte,
808
+ NULL, },
809
+ { gen_helper_sve_ldffbsu_zss_mte,
810
+ gen_helper_sve_ldffhsu_be_zss_mte,
811
+ gen_helper_sve_ldffss_be_zss_mte, } } } } },
812
};
813
814
/* Note that we overload xs=2 to indicate 64-bit offset. */
815
-static gen_helper_gvec_mem_scatter * const gather_load_fn64[2][2][3][2][4] = {
816
- /* Little-endian */
817
- { { { { gen_helper_sve_ldbds_zsu,
818
- gen_helper_sve_ldhds_le_zsu,
819
- gen_helper_sve_ldsds_le_zsu,
820
- NULL, },
821
- { gen_helper_sve_ldbdu_zsu,
822
- gen_helper_sve_ldhdu_le_zsu,
823
- gen_helper_sve_ldsdu_le_zsu,
824
- gen_helper_sve_lddd_le_zsu, } },
825
- { { gen_helper_sve_ldbds_zss,
826
- gen_helper_sve_ldhds_le_zss,
827
- gen_helper_sve_ldsds_le_zss,
828
- NULL, },
829
- { gen_helper_sve_ldbdu_zss,
830
- gen_helper_sve_ldhdu_le_zss,
831
- gen_helper_sve_ldsdu_le_zss,
832
- gen_helper_sve_lddd_le_zss, } },
833
- { { gen_helper_sve_ldbds_zd,
834
- gen_helper_sve_ldhds_le_zd,
835
- gen_helper_sve_ldsds_le_zd,
836
- NULL, },
837
- { gen_helper_sve_ldbdu_zd,
838
- gen_helper_sve_ldhdu_le_zd,
839
- gen_helper_sve_ldsdu_le_zd,
840
- gen_helper_sve_lddd_le_zd, } } },
841
+static gen_helper_gvec_mem_scatter * const
842
+gather_load_fn64[2][2][2][3][2][4] = {
843
+ { /* MTE Inactive */
844
+ { /* Little-endian */
845
+ { { { gen_helper_sve_ldbds_zsu,
846
+ gen_helper_sve_ldhds_le_zsu,
847
+ gen_helper_sve_ldsds_le_zsu,
848
+ NULL, },
849
+ { gen_helper_sve_ldbdu_zsu,
850
+ gen_helper_sve_ldhdu_le_zsu,
851
+ gen_helper_sve_ldsdu_le_zsu,
852
+ gen_helper_sve_lddd_le_zsu, } },
853
+ { { gen_helper_sve_ldbds_zss,
854
+ gen_helper_sve_ldhds_le_zss,
855
+ gen_helper_sve_ldsds_le_zss,
856
+ NULL, },
857
+ { gen_helper_sve_ldbdu_zss,
858
+ gen_helper_sve_ldhdu_le_zss,
859
+ gen_helper_sve_ldsdu_le_zss,
860
+ gen_helper_sve_lddd_le_zss, } },
861
+ { { gen_helper_sve_ldbds_zd,
862
+ gen_helper_sve_ldhds_le_zd,
863
+ gen_helper_sve_ldsds_le_zd,
864
+ NULL, },
865
+ { gen_helper_sve_ldbdu_zd,
866
+ gen_helper_sve_ldhdu_le_zd,
867
+ gen_helper_sve_ldsdu_le_zd,
868
+ gen_helper_sve_lddd_le_zd, } } },
869
870
- /* First-fault */
871
- { { { gen_helper_sve_ldffbds_zsu,
872
- gen_helper_sve_ldffhds_le_zsu,
873
- gen_helper_sve_ldffsds_le_zsu,
874
- NULL, },
875
- { gen_helper_sve_ldffbdu_zsu,
876
- gen_helper_sve_ldffhdu_le_zsu,
877
- gen_helper_sve_ldffsdu_le_zsu,
878
- gen_helper_sve_ldffdd_le_zsu, } },
879
- { { gen_helper_sve_ldffbds_zss,
880
- gen_helper_sve_ldffhds_le_zss,
881
- gen_helper_sve_ldffsds_le_zss,
882
- NULL, },
883
- { gen_helper_sve_ldffbdu_zss,
884
- gen_helper_sve_ldffhdu_le_zss,
885
- gen_helper_sve_ldffsdu_le_zss,
886
- gen_helper_sve_ldffdd_le_zss, } },
887
- { { gen_helper_sve_ldffbds_zd,
888
- gen_helper_sve_ldffhds_le_zd,
889
- gen_helper_sve_ldffsds_le_zd,
890
- NULL, },
891
- { gen_helper_sve_ldffbdu_zd,
892
- gen_helper_sve_ldffhdu_le_zd,
893
- gen_helper_sve_ldffsdu_le_zd,
894
- gen_helper_sve_ldffdd_le_zd, } } } },
895
+ /* First-fault */
896
+ { { { gen_helper_sve_ldffbds_zsu,
897
+ gen_helper_sve_ldffhds_le_zsu,
898
+ gen_helper_sve_ldffsds_le_zsu,
899
+ NULL, },
900
+ { gen_helper_sve_ldffbdu_zsu,
901
+ gen_helper_sve_ldffhdu_le_zsu,
902
+ gen_helper_sve_ldffsdu_le_zsu,
903
+ gen_helper_sve_ldffdd_le_zsu, } },
904
+ { { gen_helper_sve_ldffbds_zss,
905
+ gen_helper_sve_ldffhds_le_zss,
906
+ gen_helper_sve_ldffsds_le_zss,
907
+ NULL, },
908
+ { gen_helper_sve_ldffbdu_zss,
909
+ gen_helper_sve_ldffhdu_le_zss,
910
+ gen_helper_sve_ldffsdu_le_zss,
911
+ gen_helper_sve_ldffdd_le_zss, } },
912
+ { { gen_helper_sve_ldffbds_zd,
913
+ gen_helper_sve_ldffhds_le_zd,
914
+ gen_helper_sve_ldffsds_le_zd,
915
+ NULL, },
916
+ { gen_helper_sve_ldffbdu_zd,
917
+ gen_helper_sve_ldffhdu_le_zd,
918
+ gen_helper_sve_ldffsdu_le_zd,
919
+ gen_helper_sve_ldffdd_le_zd, } } } },
920
+ { /* Big-endian */
921
+ { { { gen_helper_sve_ldbds_zsu,
922
+ gen_helper_sve_ldhds_be_zsu,
923
+ gen_helper_sve_ldsds_be_zsu,
924
+ NULL, },
925
+ { gen_helper_sve_ldbdu_zsu,
926
+ gen_helper_sve_ldhdu_be_zsu,
927
+ gen_helper_sve_ldsdu_be_zsu,
928
+ gen_helper_sve_lddd_be_zsu, } },
929
+ { { gen_helper_sve_ldbds_zss,
930
+ gen_helper_sve_ldhds_be_zss,
931
+ gen_helper_sve_ldsds_be_zss,
932
+ NULL, },
933
+ { gen_helper_sve_ldbdu_zss,
934
+ gen_helper_sve_ldhdu_be_zss,
935
+ gen_helper_sve_ldsdu_be_zss,
936
+ gen_helper_sve_lddd_be_zss, } },
937
+ { { gen_helper_sve_ldbds_zd,
938
+ gen_helper_sve_ldhds_be_zd,
939
+ gen_helper_sve_ldsds_be_zd,
940
+ NULL, },
941
+ { gen_helper_sve_ldbdu_zd,
942
+ gen_helper_sve_ldhdu_be_zd,
943
+ gen_helper_sve_ldsdu_be_zd,
944
+ gen_helper_sve_lddd_be_zd, } } },
945
946
- /* Big-endian */
947
- { { { { gen_helper_sve_ldbds_zsu,
948
- gen_helper_sve_ldhds_be_zsu,
949
- gen_helper_sve_ldsds_be_zsu,
950
- NULL, },
951
- { gen_helper_sve_ldbdu_zsu,
952
- gen_helper_sve_ldhdu_be_zsu,
953
- gen_helper_sve_ldsdu_be_zsu,
954
- gen_helper_sve_lddd_be_zsu, } },
955
- { { gen_helper_sve_ldbds_zss,
956
- gen_helper_sve_ldhds_be_zss,
957
- gen_helper_sve_ldsds_be_zss,
958
- NULL, },
959
- { gen_helper_sve_ldbdu_zss,
960
- gen_helper_sve_ldhdu_be_zss,
961
- gen_helper_sve_ldsdu_be_zss,
962
- gen_helper_sve_lddd_be_zss, } },
963
- { { gen_helper_sve_ldbds_zd,
964
- gen_helper_sve_ldhds_be_zd,
965
- gen_helper_sve_ldsds_be_zd,
966
- NULL, },
967
- { gen_helper_sve_ldbdu_zd,
968
- gen_helper_sve_ldhdu_be_zd,
969
- gen_helper_sve_ldsdu_be_zd,
970
- gen_helper_sve_lddd_be_zd, } } },
971
+ /* First-fault */
972
+ { { { gen_helper_sve_ldffbds_zsu,
973
+ gen_helper_sve_ldffhds_be_zsu,
974
+ gen_helper_sve_ldffsds_be_zsu,
975
+ NULL, },
976
+ { gen_helper_sve_ldffbdu_zsu,
977
+ gen_helper_sve_ldffhdu_be_zsu,
978
+ gen_helper_sve_ldffsdu_be_zsu,
979
+ gen_helper_sve_ldffdd_be_zsu, } },
980
+ { { gen_helper_sve_ldffbds_zss,
981
+ gen_helper_sve_ldffhds_be_zss,
982
+ gen_helper_sve_ldffsds_be_zss,
983
+ NULL, },
984
+ { gen_helper_sve_ldffbdu_zss,
985
+ gen_helper_sve_ldffhdu_be_zss,
986
+ gen_helper_sve_ldffsdu_be_zss,
987
+ gen_helper_sve_ldffdd_be_zss, } },
988
+ { { gen_helper_sve_ldffbds_zd,
989
+ gen_helper_sve_ldffhds_be_zd,
990
+ gen_helper_sve_ldffsds_be_zd,
991
+ NULL, },
992
+ { gen_helper_sve_ldffbdu_zd,
993
+ gen_helper_sve_ldffhdu_be_zd,
994
+ gen_helper_sve_ldffsdu_be_zd,
995
+ gen_helper_sve_ldffdd_be_zd, } } } } },
996
+ { /* MTE Active */
997
+ { /* Little-endian */
998
+ { { { gen_helper_sve_ldbds_zsu_mte,
999
+ gen_helper_sve_ldhds_le_zsu_mte,
1000
+ gen_helper_sve_ldsds_le_zsu_mte,
1001
+ NULL, },
1002
+ { gen_helper_sve_ldbdu_zsu_mte,
1003
+ gen_helper_sve_ldhdu_le_zsu_mte,
1004
+ gen_helper_sve_ldsdu_le_zsu_mte,
1005
+ gen_helper_sve_lddd_le_zsu_mte, } },
1006
+ { { gen_helper_sve_ldbds_zss_mte,
1007
+ gen_helper_sve_ldhds_le_zss_mte,
1008
+ gen_helper_sve_ldsds_le_zss_mte,
1009
+ NULL, },
1010
+ { gen_helper_sve_ldbdu_zss_mte,
1011
+ gen_helper_sve_ldhdu_le_zss_mte,
1012
+ gen_helper_sve_ldsdu_le_zss_mte,
1013
+ gen_helper_sve_lddd_le_zss_mte, } },
1014
+ { { gen_helper_sve_ldbds_zd_mte,
1015
+ gen_helper_sve_ldhds_le_zd_mte,
1016
+ gen_helper_sve_ldsds_le_zd_mte,
1017
+ NULL, },
1018
+ { gen_helper_sve_ldbdu_zd_mte,
1019
+ gen_helper_sve_ldhdu_le_zd_mte,
1020
+ gen_helper_sve_ldsdu_le_zd_mte,
1021
+ gen_helper_sve_lddd_le_zd_mte, } } },
1022
1023
- /* First-fault */
1024
- { { { gen_helper_sve_ldffbds_zsu,
1025
- gen_helper_sve_ldffhds_be_zsu,
1026
- gen_helper_sve_ldffsds_be_zsu,
1027
- NULL, },
1028
- { gen_helper_sve_ldffbdu_zsu,
1029
- gen_helper_sve_ldffhdu_be_zsu,
1030
- gen_helper_sve_ldffsdu_be_zsu,
1031
- gen_helper_sve_ldffdd_be_zsu, } },
1032
- { { gen_helper_sve_ldffbds_zss,
1033
- gen_helper_sve_ldffhds_be_zss,
1034
- gen_helper_sve_ldffsds_be_zss,
1035
- NULL, },
1036
- { gen_helper_sve_ldffbdu_zss,
1037
- gen_helper_sve_ldffhdu_be_zss,
1038
- gen_helper_sve_ldffsdu_be_zss,
1039
- gen_helper_sve_ldffdd_be_zss, } },
1040
- { { gen_helper_sve_ldffbds_zd,
1041
- gen_helper_sve_ldffhds_be_zd,
1042
- gen_helper_sve_ldffsds_be_zd,
1043
- NULL, },
1044
- { gen_helper_sve_ldffbdu_zd,
1045
- gen_helper_sve_ldffhdu_be_zd,
1046
- gen_helper_sve_ldffsdu_be_zd,
1047
- gen_helper_sve_ldffdd_be_zd, } } } },
1048
+ /* First-fault */
1049
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1050
+ gen_helper_sve_ldffhds_le_zsu_mte,
1051
+ gen_helper_sve_ldffsds_le_zsu_mte,
1052
+ NULL, },
1053
+ { gen_helper_sve_ldffbdu_zsu_mte,
1054
+ gen_helper_sve_ldffhdu_le_zsu_mte,
1055
+ gen_helper_sve_ldffsdu_le_zsu_mte,
1056
+ gen_helper_sve_ldffdd_le_zsu_mte, } },
1057
+ { { gen_helper_sve_ldffbds_zss_mte,
1058
+ gen_helper_sve_ldffhds_le_zss_mte,
1059
+ gen_helper_sve_ldffsds_le_zss_mte,
1060
+ NULL, },
1061
+ { gen_helper_sve_ldffbdu_zss_mte,
1062
+ gen_helper_sve_ldffhdu_le_zss_mte,
1063
+ gen_helper_sve_ldffsdu_le_zss_mte,
1064
+ gen_helper_sve_ldffdd_le_zss_mte, } },
1065
+ { { gen_helper_sve_ldffbds_zd_mte,
1066
+ gen_helper_sve_ldffhds_le_zd_mte,
1067
+ gen_helper_sve_ldffsds_le_zd_mte,
1068
+ NULL, },
1069
+ { gen_helper_sve_ldffbdu_zd_mte,
1070
+ gen_helper_sve_ldffhdu_le_zd_mte,
1071
+ gen_helper_sve_ldffsdu_le_zd_mte,
1072
+ gen_helper_sve_ldffdd_le_zd_mte, } } } },
1073
+ { /* Big-endian */
1074
+ { { { gen_helper_sve_ldbds_zsu_mte,
1075
+ gen_helper_sve_ldhds_be_zsu_mte,
1076
+ gen_helper_sve_ldsds_be_zsu_mte,
1077
+ NULL, },
1078
+ { gen_helper_sve_ldbdu_zsu_mte,
1079
+ gen_helper_sve_ldhdu_be_zsu_mte,
1080
+ gen_helper_sve_ldsdu_be_zsu_mte,
1081
+ gen_helper_sve_lddd_be_zsu_mte, } },
1082
+ { { gen_helper_sve_ldbds_zss_mte,
1083
+ gen_helper_sve_ldhds_be_zss_mte,
1084
+ gen_helper_sve_ldsds_be_zss_mte,
1085
+ NULL, },
1086
+ { gen_helper_sve_ldbdu_zss_mte,
1087
+ gen_helper_sve_ldhdu_be_zss_mte,
1088
+ gen_helper_sve_ldsdu_be_zss_mte,
1089
+ gen_helper_sve_lddd_be_zss_mte, } },
1090
+ { { gen_helper_sve_ldbds_zd_mte,
1091
+ gen_helper_sve_ldhds_be_zd_mte,
1092
+ gen_helper_sve_ldsds_be_zd_mte,
1093
+ NULL, },
1094
+ { gen_helper_sve_ldbdu_zd_mte,
1095
+ gen_helper_sve_ldhdu_be_zd_mte,
1096
+ gen_helper_sve_ldsdu_be_zd_mte,
1097
+ gen_helper_sve_lddd_be_zd_mte, } } },
1098
+
1099
+ /* First-fault */
1100
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1101
+ gen_helper_sve_ldffhds_be_zsu_mte,
1102
+ gen_helper_sve_ldffsds_be_zsu_mte,
1103
+ NULL, },
1104
+ { gen_helper_sve_ldffbdu_zsu_mte,
1105
+ gen_helper_sve_ldffhdu_be_zsu_mte,
1106
+ gen_helper_sve_ldffsdu_be_zsu_mte,
1107
+ gen_helper_sve_ldffdd_be_zsu_mte, } },
1108
+ { { gen_helper_sve_ldffbds_zss_mte,
1109
+ gen_helper_sve_ldffhds_be_zss_mte,
1110
+ gen_helper_sve_ldffsds_be_zss_mte,
1111
+ NULL, },
1112
+ { gen_helper_sve_ldffbdu_zss_mte,
1113
+ gen_helper_sve_ldffhdu_be_zss_mte,
1114
+ gen_helper_sve_ldffsdu_be_zss_mte,
1115
+ gen_helper_sve_ldffdd_be_zss_mte, } },
1116
+ { { gen_helper_sve_ldffbds_zd_mte,
1117
+ gen_helper_sve_ldffhds_be_zd_mte,
1118
+ gen_helper_sve_ldffsds_be_zd_mte,
1119
+ NULL, },
1120
+ { gen_helper_sve_ldffbdu_zd_mte,
1121
+ gen_helper_sve_ldffhdu_be_zd_mte,
1122
+ gen_helper_sve_ldffsdu_be_zd_mte,
1123
+ gen_helper_sve_ldffdd_be_zd_mte, } } } } },
1124
};
1125
1126
static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1127
{
1128
gen_helper_gvec_mem_scatter *fn = NULL;
1129
- int be = s->be_data == MO_BE;
1130
+ bool be = s->be_data == MO_BE;
1131
+ bool mte = s->mte_active[0];
1132
1133
if (!sve_access_check(s)) {
1134
return true;
1135
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1136
1137
switch (a->esz) {
1138
case MO_32:
1139
- fn = gather_load_fn32[be][a->ff][a->xs][a->u][a->msz];
1140
+ fn = gather_load_fn32[mte][be][a->ff][a->xs][a->u][a->msz];
1141
break;
1142
case MO_64:
1143
- fn = gather_load_fn64[be][a->ff][a->xs][a->u][a->msz];
1144
+ fn = gather_load_fn64[mte][be][a->ff][a->xs][a->u][a->msz];
1145
break;
1146
}
1147
assert(fn != NULL);
1148
1149
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1150
- cpu_reg_sp(s, a->rn), a->msz, fn);
1151
+ cpu_reg_sp(s, a->rn), a->msz, false, fn);
53
return true;
1152
return true;
54
}
1153
}
55
+
1154
56
+static bool trans_VLDST_single(DisasContext *s, arg_VLDST_single *a)
1155
static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
57
+{
1156
{
58
+ /* Neon load/store single structure to one lane */
1157
gen_helper_gvec_mem_scatter *fn = NULL;
59
+ int reg;
1158
- int be = s->be_data == MO_BE;
60
+ int nregs = a->n + 1;
1159
+ bool be = s->be_data == MO_BE;
61
+ int vd = a->vd;
1160
+ bool mte = s->mte_active[0];
62
+ TCGv_i32 addr, tmp;
1161
TCGv_i64 imm;
63
+
1162
64
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
1163
if (a->esz < a->msz || (a->esz == a->msz && !a->u)) {
65
+ return false;
1164
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
66
+ }
1165
67
+
1166
switch (a->esz) {
68
+ /* UNDEF accesses to D16-D31 if they don't exist */
1167
case MO_32:
69
+ if (!dc_isar_feature(aa32_simd_r32, s) && (a->vd & 0x10)) {
1168
- fn = gather_load_fn32[be][a->ff][0][a->u][a->msz];
70
+ return false;
1169
+ fn = gather_load_fn32[mte][be][a->ff][0][a->u][a->msz];
71
+ }
1170
break;
72
+
1171
case MO_64:
73
+ /* Catch the UNDEF cases. This is unavoidably a bit messy. */
1172
- fn = gather_load_fn64[be][a->ff][2][a->u][a->msz];
74
+ switch (nregs) {
1173
+ fn = gather_load_fn64[mte][be][a->ff][2][a->u][a->msz];
75
+ case 1:
1174
break;
76
+ if (((a->align & (1 << a->size)) != 0) ||
1175
}
77
+ (a->size == 2 && ((a->align & 3) == 1 || (a->align & 3) == 2))) {
1176
assert(fn != NULL);
78
+ return false;
1177
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
79
+ }
1178
* by loading the immediate into the scalar parameter.
80
+ break;
1179
*/
81
+ case 3:
1180
imm = tcg_const_i64(a->imm << a->msz);
82
+ if ((a->align & 1) != 0) {
1181
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
83
+ return false;
1182
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, false, fn);
84
+ }
1183
tcg_temp_free_i64(imm);
85
+ /* fall through */
1184
return true;
86
+ case 2:
87
+ if (a->size == 2 && (a->align & 2) != 0) {
88
+ return false;
89
+ }
90
+ break;
91
+ case 4:
92
+ if ((a->size == 2) && ((a->align & 3) == 3)) {
93
+ return false;
94
+ }
95
+ break;
96
+ default:
97
+ abort();
98
+ }
99
+ if ((vd + a->stride * (nregs - 1)) > 31) {
100
+ /*
101
+ * Attempts to write off the end of the register file are
102
+ * UNPREDICTABLE; we choose to UNDEF because otherwise we would
103
+ * access off the end of the array that holds the register data.
104
+ */
105
+ return false;
106
+ }
107
+
108
+ if (!vfp_access_check(s)) {
109
+ return true;
110
+ }
111
+
112
+ tmp = tcg_temp_new_i32();
113
+ addr = tcg_temp_new_i32();
114
+ load_reg_var(s, addr, a->rn);
115
+ /*
116
+ * TODO: if we implemented alignment exceptions, we should check
117
+ * addr against the alignment encoded in a->align here.
118
+ */
119
+ for (reg = 0; reg < nregs; reg++) {
120
+ if (a->l) {
121
+ gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
122
+ s->be_data | a->size);
123
+ neon_store_element(vd, a->reg_idx, a->size, tmp);
124
+ } else { /* Store */
125
+ neon_load_element(tmp, vd, a->reg_idx, a->size);
126
+ gen_aa32_st_i32(s, tmp, addr, get_mem_index(s),
127
+ s->be_data | a->size);
128
+ }
129
+ vd += a->stride;
130
+ tcg_gen_addi_i32(addr, addr, 1 << a->size);
131
+ }
132
+ tcg_temp_free_i32(addr);
133
+ tcg_temp_free_i32(tmp);
134
+
135
+ gen_neon_ldst_base_update(s, a->rm, a->rn, (1 << a->size) * nregs);
136
+
137
+ return true;
138
+}
139
diff --git a/target/arm/translate.c b/target/arm/translate.c
140
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/translate.c
142
+++ b/target/arm/translate.c
143
@@ -XXX,XX +XXX,XX @@ static void gen_neon_trn_u16(TCGv_i32 t0, TCGv_i32 t1)
144
tcg_temp_free_i32(rd);
145
}
1185
}
146
1186
147
-
1187
-/* Indexed by [be][xs][msz]. */
148
-/* Translate a NEON load/store element instruction. Return nonzero if the
1188
-static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][3] = {
149
- instruction is invalid. */
1189
- /* Little-endian */
150
-static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
1190
- { { gen_helper_sve_stbs_zsu,
151
-{
1191
- gen_helper_sve_sths_le_zsu,
152
- int rd, rn, rm;
1192
- gen_helper_sve_stss_le_zsu, },
153
- int nregs;
1193
- { gen_helper_sve_stbs_zss,
154
- int stride;
1194
- gen_helper_sve_sths_le_zss,
155
- int size;
1195
- gen_helper_sve_stss_le_zss, } },
156
- int reg;
1196
- /* Big-endian */
157
- int load;
1197
- { { gen_helper_sve_stbs_zsu,
158
- TCGv_i32 addr;
1198
- gen_helper_sve_sths_be_zsu,
159
- TCGv_i32 tmp;
1199
- gen_helper_sve_stss_be_zsu, },
160
-
1200
- { gen_helper_sve_stbs_zss,
161
- if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
1201
- gen_helper_sve_sths_be_zss,
162
- return 1;
1202
- gen_helper_sve_stss_be_zss, } },
163
- }
1203
+/* Indexed by [mte][be][xs][msz]. */
164
-
1204
+static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = {
165
- /* FIXME: this access check should not take precedence over UNDEF
1205
+ { /* MTE Inactive */
166
- * for invalid encodings; we will generate incorrect syndrome information
1206
+ { /* Little-endian */
167
- * for attempts to execute invalid vfp/neon encodings with FP disabled.
1207
+ { gen_helper_sve_stbs_zsu,
168
- */
1208
+ gen_helper_sve_sths_le_zsu,
169
- if (s->fp_excp_el) {
1209
+ gen_helper_sve_stss_le_zsu, },
170
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
1210
+ { gen_helper_sve_stbs_zss,
171
- syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
1211
+ gen_helper_sve_sths_le_zss,
172
- return 0;
1212
+ gen_helper_sve_stss_le_zss, } },
173
- }
1213
+ { /* Big-endian */
174
-
1214
+ { gen_helper_sve_stbs_zsu,
175
- if (!s->vfp_enabled)
1215
+ gen_helper_sve_sths_be_zsu,
176
- return 1;
1216
+ gen_helper_sve_stss_be_zsu, },
177
- VFP_DREG_D(rd, insn);
1217
+ { gen_helper_sve_stbs_zss,
178
- rn = (insn >> 16) & 0xf;
1218
+ gen_helper_sve_sths_be_zss,
179
- rm = insn & 0xf;
1219
+ gen_helper_sve_stss_be_zss, } } },
180
- load = (insn & (1 << 21)) != 0;
1220
+ { /* MTE Active */
181
- if ((insn & (1 << 23)) == 0) {
1221
+ { /* Little-endian */
182
- /* Load store all elements -- handled already by decodetree */
1222
+ { gen_helper_sve_stbs_zsu_mte,
183
- return 1;
1223
+ gen_helper_sve_sths_le_zsu_mte,
184
- } else {
1224
+ gen_helper_sve_stss_le_zsu_mte, },
185
- size = (insn >> 10) & 3;
1225
+ { gen_helper_sve_stbs_zss_mte,
186
- if (size == 3) {
1226
+ gen_helper_sve_sths_le_zss_mte,
187
- /* Load single element to all lanes -- handled by decodetree */
1227
+ gen_helper_sve_stss_le_zss_mte, } },
188
- return 1;
1228
+ { /* Big-endian */
189
- } else {
1229
+ { gen_helper_sve_stbs_zsu_mte,
190
- /* Single element. */
1230
+ gen_helper_sve_sths_be_zsu_mte,
191
- int idx = (insn >> 4) & 0xf;
1231
+ gen_helper_sve_stss_be_zsu_mte, },
192
- int reg_idx;
1232
+ { gen_helper_sve_stbs_zss_mte,
193
- switch (size) {
1233
+ gen_helper_sve_sths_be_zss_mte,
194
- case 0:
1234
+ gen_helper_sve_stss_be_zss_mte, } } },
195
- reg_idx = (insn >> 5) & 7;
1235
};
196
- stride = 1;
1236
197
- break;
1237
/* Note that we overload xs=2 to indicate 64-bit offset. */
198
- case 1:
1238
-static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][3][4] = {
199
- reg_idx = (insn >> 6) & 3;
1239
- /* Little-endian */
200
- stride = (insn & (1 << 5)) ? 2 : 1;
1240
- { { gen_helper_sve_stbd_zsu,
201
- break;
1241
- gen_helper_sve_sthd_le_zsu,
202
- case 2:
1242
- gen_helper_sve_stsd_le_zsu,
203
- reg_idx = (insn >> 7) & 1;
1243
- gen_helper_sve_stdd_le_zsu, },
204
- stride = (insn & (1 << 6)) ? 2 : 1;
1244
- { gen_helper_sve_stbd_zss,
205
- break;
1245
- gen_helper_sve_sthd_le_zss,
206
- default:
1246
- gen_helper_sve_stsd_le_zss,
207
- abort();
1247
- gen_helper_sve_stdd_le_zss, },
208
- }
1248
- { gen_helper_sve_stbd_zd,
209
- nregs = ((insn >> 8) & 3) + 1;
1249
- gen_helper_sve_sthd_le_zd,
210
- /* Catch the UNDEF cases. This is unavoidably a bit messy. */
1250
- gen_helper_sve_stsd_le_zd,
211
- switch (nregs) {
1251
- gen_helper_sve_stdd_le_zd, } },
212
- case 1:
1252
- /* Big-endian */
213
- if (((idx & (1 << size)) != 0) ||
1253
- { { gen_helper_sve_stbd_zsu,
214
- (size == 2 && ((idx & 3) == 1 || (idx & 3) == 2))) {
1254
- gen_helper_sve_sthd_be_zsu,
215
- return 1;
1255
- gen_helper_sve_stsd_be_zsu,
216
- }
1256
- gen_helper_sve_stdd_be_zsu, },
217
- break;
1257
- { gen_helper_sve_stbd_zss,
218
- case 3:
1258
- gen_helper_sve_sthd_be_zss,
219
- if ((idx & 1) != 0) {
1259
- gen_helper_sve_stsd_be_zss,
220
- return 1;
1260
- gen_helper_sve_stdd_be_zss, },
221
- }
1261
- { gen_helper_sve_stbd_zd,
222
- /* fall through */
1262
- gen_helper_sve_sthd_be_zd,
223
- case 2:
1263
- gen_helper_sve_stsd_be_zd,
224
- if (size == 2 && (idx & 2) != 0) {
1264
- gen_helper_sve_stdd_be_zd, } },
225
- return 1;
1265
+static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][2][3][4] = {
226
- }
1266
+ { /* MTE Inactive */
227
- break;
1267
+ { /* Little-endian */
228
- case 4:
1268
+ { gen_helper_sve_stbd_zsu,
229
- if ((size == 2) && ((idx & 3) == 3)) {
1269
+ gen_helper_sve_sthd_le_zsu,
230
- return 1;
1270
+ gen_helper_sve_stsd_le_zsu,
231
- }
1271
+ gen_helper_sve_stdd_le_zsu, },
232
- break;
1272
+ { gen_helper_sve_stbd_zss,
233
- default:
1273
+ gen_helper_sve_sthd_le_zss,
234
- abort();
1274
+ gen_helper_sve_stsd_le_zss,
235
- }
1275
+ gen_helper_sve_stdd_le_zss, },
236
- if ((rd + stride * (nregs - 1)) > 31) {
1276
+ { gen_helper_sve_stbd_zd,
237
- /* Attempts to write off the end of the register file
1277
+ gen_helper_sve_sthd_le_zd,
238
- * are UNPREDICTABLE; we choose to UNDEF because otherwise
1278
+ gen_helper_sve_stsd_le_zd,
239
- * the neon_load_reg() would write off the end of the array.
1279
+ gen_helper_sve_stdd_le_zd, } },
240
- */
1280
+ { /* Big-endian */
241
- return 1;
1281
+ { gen_helper_sve_stbd_zsu,
242
- }
1282
+ gen_helper_sve_sthd_be_zsu,
243
- tmp = tcg_temp_new_i32();
1283
+ gen_helper_sve_stsd_be_zsu,
244
- addr = tcg_temp_new_i32();
1284
+ gen_helper_sve_stdd_be_zsu, },
245
- load_reg_var(s, addr, rn);
1285
+ { gen_helper_sve_stbd_zss,
246
- for (reg = 0; reg < nregs; reg++) {
1286
+ gen_helper_sve_sthd_be_zss,
247
- if (load) {
1287
+ gen_helper_sve_stsd_be_zss,
248
- gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
1288
+ gen_helper_sve_stdd_be_zss, },
249
- s->be_data | size);
1289
+ { gen_helper_sve_stbd_zd,
250
- neon_store_element(rd, reg_idx, size, tmp);
1290
+ gen_helper_sve_sthd_be_zd,
251
- } else { /* Store */
1291
+ gen_helper_sve_stsd_be_zd,
252
- neon_load_element(tmp, rd, reg_idx, size);
1292
+ gen_helper_sve_stdd_be_zd, } } },
253
- gen_aa32_st_i32(s, tmp, addr, get_mem_index(s),
1293
+ { /* MTE Inactive */
254
- s->be_data | size);
1294
+ { /* Little-endian */
255
- }
1295
+ { gen_helper_sve_stbd_zsu_mte,
256
- rd += stride;
1296
+ gen_helper_sve_sthd_le_zsu_mte,
257
- tcg_gen_addi_i32(addr, addr, 1 << size);
1297
+ gen_helper_sve_stsd_le_zsu_mte,
258
- }
1298
+ gen_helper_sve_stdd_le_zsu_mte, },
259
- tcg_temp_free_i32(addr);
1299
+ { gen_helper_sve_stbd_zss_mte,
260
- tcg_temp_free_i32(tmp);
1300
+ gen_helper_sve_sthd_le_zss_mte,
261
- stride = nregs * (1 << size);
1301
+ gen_helper_sve_stsd_le_zss_mte,
262
- }
1302
+ gen_helper_sve_stdd_le_zss_mte, },
263
- }
1303
+ { gen_helper_sve_stbd_zd_mte,
264
- if (rm != 15) {
1304
+ gen_helper_sve_sthd_le_zd_mte,
265
- TCGv_i32 base;
1305
+ gen_helper_sve_stsd_le_zd_mte,
266
-
1306
+ gen_helper_sve_stdd_le_zd_mte, } },
267
- base = load_reg(s, rn);
1307
+ { /* Big-endian */
268
- if (rm == 13) {
1308
+ { gen_helper_sve_stbd_zsu_mte,
269
- tcg_gen_addi_i32(base, base, stride);
1309
+ gen_helper_sve_sthd_be_zsu_mte,
270
- } else {
1310
+ gen_helper_sve_stsd_be_zsu_mte,
271
- TCGv_i32 index;
1311
+ gen_helper_sve_stdd_be_zsu_mte, },
272
- index = load_reg(s, rm);
1312
+ { gen_helper_sve_stbd_zss_mte,
273
- tcg_gen_add_i32(base, base, index);
1313
+ gen_helper_sve_sthd_be_zss_mte,
274
- tcg_temp_free_i32(index);
1314
+ gen_helper_sve_stsd_be_zss_mte,
275
- }
1315
+ gen_helper_sve_stdd_be_zss_mte, },
276
- store_reg(s, rn, base);
1316
+ { gen_helper_sve_stbd_zd_mte,
277
- }
1317
+ gen_helper_sve_sthd_be_zd_mte,
278
- return 0;
1318
+ gen_helper_sve_stsd_be_zd_mte,
279
-}
1319
+ gen_helper_sve_stdd_be_zd_mte, } } },
280
-
1320
};
281
static inline void gen_neon_narrow(int size, TCGv_i32 dest, TCGv_i64 src)
1321
1322
static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
282
{
1323
{
283
switch (size) {
1324
gen_helper_gvec_mem_scatter *fn;
284
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
1325
- int be = s->be_data == MO_BE;
285
}
1326
+ bool be = s->be_data == MO_BE;
286
return;
1327
+ bool mte = s->mte_active[0];
287
}
1328
288
- if ((insn & 0x0f100000) == 0x04000000) {
1329
if (a->esz < a->msz || (a->msz == 0 && a->scale)) {
289
- /* NEON load/store. */
1330
return false;
290
- if (disas_neon_ls_insn(s, insn)) {
1331
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
291
- goto illegal_op;
1332
}
292
- }
1333
switch (a->esz) {
293
- return;
1334
case MO_32:
294
- }
1335
- fn = scatter_store_fn32[be][a->xs][a->msz];
295
if ((insn & 0x0e000f00) == 0x0c000100) {
1336
+ fn = scatter_store_fn32[mte][be][a->xs][a->msz];
296
if (arm_dc_feature(s, ARM_FEATURE_IWMMXT)) {
297
/* iWMMXt register transfer. */
298
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
299
}
300
break;
1337
break;
301
case 12:
1338
case MO_64:
302
- if ((insn & 0x01100000) == 0x01000000) {
1339
- fn = scatter_store_fn64[be][a->xs][a->msz];
303
- if (disas_neon_ls_insn(s, insn)) {
1340
+ fn = scatter_store_fn64[mte][be][a->xs][a->msz];
304
- goto illegal_op;
1341
break;
305
- }
306
- break;
307
- }
308
goto illegal_op;
309
default:
1342
default:
310
illegal_op:
1343
g_assert_not_reached();
1344
}
1345
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1346
- cpu_reg_sp(s, a->rn), a->msz, fn);
1347
+ cpu_reg_sp(s, a->rn), a->msz, true, fn);
1348
return true;
1349
}
1350
1351
static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1352
{
1353
gen_helper_gvec_mem_scatter *fn = NULL;
1354
- int be = s->be_data == MO_BE;
1355
+ bool be = s->be_data == MO_BE;
1356
+ bool mte = s->mte_active[0];
1357
TCGv_i64 imm;
1358
1359
if (a->esz < a->msz) {
1360
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1361
1362
switch (a->esz) {
1363
case MO_32:
1364
- fn = scatter_store_fn32[be][0][a->msz];
1365
+ fn = scatter_store_fn32[mte][be][0][a->msz];
1366
break;
1367
case MO_64:
1368
- fn = scatter_store_fn64[be][2][a->msz];
1369
+ fn = scatter_store_fn64[mte][be][2][a->msz];
1370
break;
1371
}
1372
assert(fn != NULL);
1373
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1374
* by loading the immediate into the scalar parameter.
1375
*/
1376
imm = tcg_const_i64(a->imm << a->msz);
1377
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1378
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, true, fn);
1379
tcg_temp_free_i64(imm);
1380
return true;
1381
}
311
--
1382
--
312
2.20.1
1383
2.20.1
313
1384
314
1385
diff view generated by jsdifflib
1
Convert the Neon "load single structure to all lanes" insns to
1
From: Richard Henderson <richard.henderson@linaro.org>
2
decodetree.
3
2
3
There are a number of paths by which the TBI is still intact
4
for user-only in the SVE helpers.
5
6
Because we currently always set TBI for user-only, we do not
7
need to pass down the actual TBI setting from above, and we
8
can remove the top byte in the inner-most primitives, so that
9
none are forgotten. Moreover, this keeps the "dirty" pointer
10
around at the higher levels, where we need it for any MTE checking.
11
12
Since the normal case, especially for user-only, goes through
13
RAM, this clearing merely adds two insns per page lookup, which
14
will be completely in the noise.
15
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20200626033144.790098-39-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20200430181003.21682-13-peter.maydell@linaro.org
7
---
20
---
8
target/arm/neon-ls.decode | 5 +++
21
target/arm/cpu.c | 3 +++
9
target/arm/translate-neon.inc.c | 73 +++++++++++++++++++++++++++++++++
22
target/arm/sve_helper.c | 19 +++++++++++++++++--
10
target/arm/translate.c | 55 +------------------------
23
target/arm/translate-a64.c | 5 +++++
11
3 files changed, 80 insertions(+), 53 deletions(-)
24
3 files changed, 25 insertions(+), 2 deletions(-)
12
25
13
diff --git a/target/arm/neon-ls.decode b/target/arm/neon-ls.decode
26
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
14
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/neon-ls.decode
28
--- a/target/arm/cpu.c
16
+++ b/target/arm/neon-ls.decode
29
+++ b/target/arm/cpu.c
17
@@ -XXX,XX +XXX,XX @@
30
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
18
31
* Enable TBI0 and TBI1. While the real kernel only enables TBI0,
19
VLDST_multiple 1111 0100 0 . l:1 0 rn:4 .... itype:4 size:2 align:2 rm:4 \
32
* turning on both here will produce smaller code and otherwise
20
vd=%vd_dp
33
* make no difference to the user-level emulation.
21
+
34
+ *
22
+# Neon load single element to all lanes
35
+ * In sve_probe_page, we assume that this is set.
23
+
36
+ * Do not modify this without other changes.
24
+VLD_all_lanes 1111 0100 1 . 1 0 rn:4 .... 11 n:2 size:2 t:1 a:1 rm:4 \
37
*/
25
+ vd=%vd_dp
38
env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
26
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
39
#else
40
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
27
index XXXXXXX..XXXXXXX 100644
41
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/translate-neon.inc.c
42
--- a/target/arm/sve_helper.c
29
+++ b/target/arm/translate-neon.inc.c
43
+++ b/target/arm/sve_helper.c
30
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDST_multiple(DisasContext *s, arg_VLDST_multiple *a)
44
@@ -XXX,XX +XXX,XX @@ static void sve_##NAME##_host(void *vd, intptr_t reg_off, void *host) \
31
gen_neon_ldst_base_update(s, a->rm, a->rn, nregs * interleave * 8);
45
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
32
return true;
46
target_ulong addr, uintptr_t ra) \
47
{ \
48
- *(TYPEE *)(vd + H(reg_off)) = (TYPEM)TLB(env, addr, ra); \
49
+ *(TYPEE *)(vd + H(reg_off)) = \
50
+ (TYPEM)TLB(env, useronly_clean_ptr(addr), ra); \
33
}
51
}
34
+
52
35
+static bool trans_VLD_all_lanes(DisasContext *s, arg_VLD_all_lanes *a)
53
#define DO_ST_TLB(NAME, H, TYPEE, TYPEM, TLB) \
36
+{
54
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
37
+ /* Neon load single structure to all lanes */
55
target_ulong addr, uintptr_t ra) \
38
+ int reg, stride, vec_size;
56
{ \
39
+ int vd = a->vd;
57
- TLB(env, addr, (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
40
+ int size = a->size;
58
+ TLB(env, useronly_clean_ptr(addr), \
41
+ int nregs = a->n + 1;
59
+ (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
42
+ TCGv_i32 addr, tmp;
60
}
43
+
61
44
+ if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
62
#define DO_LD_PRIM_1(NAME, H, TE, TM) \
45
+ return false;
63
@@ -XXX,XX +XXX,XX @@ static bool sve_probe_page(SVEHostPage *info, bool nofault,
46
+ }
64
int flags;
47
+
65
48
+ /* UNDEF accesses to D16-D31 if they don't exist */
66
addr += mem_off;
49
+ if (!dc_isar_feature(aa32_simd_r32, s) && (a->vd & 0x10)) {
50
+ return false;
51
+ }
52
+
53
+ if (size == 3) {
54
+ if (nregs != 4 || a->a == 0) {
55
+ return false;
56
+ }
57
+ /* For VLD4 size == 3 a == 1 means 32 bits at 16 byte alignment */
58
+ size = 2;
59
+ }
60
+ if (nregs == 1 && a->a == 1 && size == 0) {
61
+ return false;
62
+ }
63
+ if (nregs == 3 && a->a == 1) {
64
+ return false;
65
+ }
66
+
67
+ if (!vfp_access_check(s)) {
68
+ return true;
69
+ }
70
+
67
+
71
+ /*
68
+ /*
72
+ * VLD1 to all lanes: T bit indicates how many Dregs to write.
69
+ * User-only currently always issues with TBI. See the comment
73
+ * VLD2/3/4 to all lanes: T bit indicates register stride.
70
+ * above useronly_clean_ptr. Usually we clean this top byte away
71
+ * during translation, but we can't do that for e.g. vector + imm
72
+ * addressing modes.
73
+ *
74
+ * We currently always enable TBI for user-only, and do not provide
75
+ * a way to turn it off. So clean the pointer unconditionally here,
76
+ * rather than look it up here, or pass it down from above.
74
+ */
77
+ */
75
+ stride = a->t ? 2 : 1;
78
+ addr = useronly_clean_ptr(addr);
76
+ vec_size = nregs == 1 ? stride * 8 : 8;
77
+
79
+
78
+ tmp = tcg_temp_new_i32();
80
flags = probe_access_flags(env, addr, access_type, mmu_idx, nofault,
79
+ addr = tcg_temp_new_i32();
81
&info->host, retaddr);
80
+ load_reg_var(s, addr, a->rn);
82
info->flags = flags;
81
+ for (reg = 0; reg < nregs; reg++) {
83
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
82
+ gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
84
index XXXXXXX..XXXXXXX 100644
83
+ s->be_data | size);
85
--- a/target/arm/translate-a64.c
84
+ if ((vd & 1) && vec_size == 16) {
86
+++ b/target/arm/translate-a64.c
85
+ /*
87
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
86
+ * We cannot write 16 bytes at once because the
88
dc->features = env->features;
87
+ * destination is unaligned.
89
dc->dcz_blocksize = arm_cpu->dcz_blocksize;
88
+ */
90
89
+ tcg_gen_gvec_dup_i32(size, neon_reg_offset(vd, 0),
91
+#ifdef CONFIG_USER_ONLY
90
+ 8, 8, tmp);
92
+ /* In sve_probe_page, we assume TBI is enabled. */
91
+ tcg_gen_gvec_mov(0, neon_reg_offset(vd + 1, 0),
93
+ tcg_debug_assert(dc->tbid & 1);
92
+ neon_reg_offset(vd, 0), 8, 8);
94
+#endif
93
+ } else {
94
+ tcg_gen_gvec_dup_i32(size, neon_reg_offset(vd, 0),
95
+ vec_size, vec_size, tmp);
96
+ }
97
+ tcg_gen_addi_i32(addr, addr, 1 << size);
98
+ vd += stride;
99
+ }
100
+ tcg_temp_free_i32(tmp);
101
+ tcg_temp_free_i32(addr);
102
+
95
+
103
+ gen_neon_ldst_base_update(s, a->rm, a->rn, (1 << size) * nregs);
96
/* Single step state. The code-generation logic here is:
104
+
97
* SS_ACTIVE == 0:
105
+ return true;
98
* generate code with no special handling for single-stepping (except
106
+}
107
diff --git a/target/arm/translate.c b/target/arm/translate.c
108
index XXXXXXX..XXXXXXX 100644
109
--- a/target/arm/translate.c
110
+++ b/target/arm/translate.c
111
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
112
int size;
113
int reg;
114
int load;
115
- int vec_size;
116
TCGv_i32 addr;
117
TCGv_i32 tmp;
118
119
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
120
} else {
121
size = (insn >> 10) & 3;
122
if (size == 3) {
123
- /* Load single element to all lanes. */
124
- int a = (insn >> 4) & 1;
125
- if (!load) {
126
- return 1;
127
- }
128
- size = (insn >> 6) & 3;
129
- nregs = ((insn >> 8) & 3) + 1;
130
-
131
- if (size == 3) {
132
- if (nregs != 4 || a == 0) {
133
- return 1;
134
- }
135
- /* For VLD4 size==3 a == 1 means 32 bits at 16 byte alignment */
136
- size = 2;
137
- }
138
- if (nregs == 1 && a == 1 && size == 0) {
139
- return 1;
140
- }
141
- if (nregs == 3 && a == 1) {
142
- return 1;
143
- }
144
- addr = tcg_temp_new_i32();
145
- load_reg_var(s, addr, rn);
146
-
147
- /* VLD1 to all lanes: bit 5 indicates how many Dregs to write.
148
- * VLD2/3/4 to all lanes: bit 5 indicates register stride.
149
- */
150
- stride = (insn & (1 << 5)) ? 2 : 1;
151
- vec_size = nregs == 1 ? stride * 8 : 8;
152
-
153
- tmp = tcg_temp_new_i32();
154
- for (reg = 0; reg < nregs; reg++) {
155
- gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
156
- s->be_data | size);
157
- if ((rd & 1) && vec_size == 16) {
158
- /* We cannot write 16 bytes at once because the
159
- * destination is unaligned.
160
- */
161
- tcg_gen_gvec_dup_i32(size, neon_reg_offset(rd, 0),
162
- 8, 8, tmp);
163
- tcg_gen_gvec_mov(0, neon_reg_offset(rd + 1, 0),
164
- neon_reg_offset(rd, 0), 8, 8);
165
- } else {
166
- tcg_gen_gvec_dup_i32(size, neon_reg_offset(rd, 0),
167
- vec_size, vec_size, tmp);
168
- }
169
- tcg_gen_addi_i32(addr, addr, 1 << size);
170
- rd += stride;
171
- }
172
- tcg_temp_free_i32(tmp);
173
- tcg_temp_free_i32(addr);
174
- stride = (1 << size) * nregs;
175
+ /* Load single element to all lanes -- handled by decodetree */
176
+ return 1;
177
} else {
178
/* Single element. */
179
int idx = (insn >> 4) & 0xf;
180
--
99
--
181
2.20.1
100
2.20.1
182
101
183
102
diff view generated by jsdifflib
1
Somewhere along theline we accidentally added a duplicate
1
From: Richard Henderson <richard.henderson@linaro.org>
2
"using D16-D31 when they don't exist" check to do_vfm_dp()
3
(probably an artifact of a patchseries rebase). Remove it.
4
2
3
This is DC GVA and DC GZVA, and the tag check for DC ZVA.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-40-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20200430181003.21682-2-peter.maydell@linaro.org
9
---
9
---
10
target/arm/translate-vfp.inc.c | 6 ------
10
target/arm/cpu.h | 4 +++-
11
1 file changed, 6 deletions(-)
11
target/arm/helper.c | 16 ++++++++++++++++
12
target/arm/translate-a64.c | 39 ++++++++++++++++++++++++++++++++++++++
13
3 files changed, 58 insertions(+), 1 deletion(-)
12
14
13
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/translate-vfp.inc.c
17
--- a/target/arm/cpu.h
16
+++ b/target/arm/translate-vfp.inc.c
18
+++ b/target/arm/cpu.h
17
@@ -XXX,XX +XXX,XX @@ static bool do_vfm_dp(DisasContext *s, arg_VFMA_dp *a, bool neg_n, bool neg_d)
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
18
return false;
20
#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
19
}
21
#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
20
22
#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
21
- /* UNDEF accesses to D16-D31 if they don't exist. */
23
-#define ARM_LAST_SPECIAL ARM_CP_DC_ZVA
22
- if (!dc_isar_feature(aa32_simd_r32, s) &&
24
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
23
- ((a->vd | a->vn | a->vm) & 0x10)) {
25
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
24
- return false;
26
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
25
- }
27
#define ARM_CP_FPU 0x1000
26
-
28
#define ARM_CP_SVE 0x2000
27
if (!vfp_access_check(s)) {
29
#define ARM_CP_NO_GDB 0x4000
28
return true;
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.c
33
+++ b/target/arm/helper.c
34
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
35
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
36
.type = ARM_CP_NOP, .access = PL0_W,
37
.accessfn = aa64_cacheop_poc_access },
38
+ { .name = "DC_GVA", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 3,
40
+ .access = PL0_W, .type = ARM_CP_DC_GVA,
41
+#ifndef CONFIG_USER_ONLY
42
+ /* Avoid overhead of an access check that always passes in user-mode */
43
+ .accessfn = aa64_zva_access,
44
+#endif
45
+ },
46
+ { .name = "DC_GZVA", .state = ARM_CP_STATE_AA64,
47
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 4,
48
+ .access = PL0_W, .type = ARM_CP_DC_GZVA,
49
+#ifndef CONFIG_USER_ONLY
50
+ /* Avoid overhead of an access check that always passes in user-mode */
51
+ .accessfn = aa64_zva_access,
52
+#endif
53
+ },
54
REGINFO_SENTINEL
55
};
56
57
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
58
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/translate-a64.c
60
+++ b/target/arm/translate-a64.c
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
}
63
gen_helper_dc_zva(cpu_env, tcg_rt);
64
return;
65
+ case ARM_CP_DC_GVA:
66
+ {
67
+ TCGv_i64 clean_addr, tag;
68
+
69
+ /*
70
+ * DC_GVA, like DC_ZVA, requires that we supply the original
71
+ * pointer for an invalid page. Probe that address first.
72
+ */
73
+ tcg_rt = cpu_reg(s, rt);
74
+ clean_addr = clean_data_tbi(s, tcg_rt);
75
+ gen_probe_access(s, clean_addr, MMU_DATA_STORE, MO_8);
76
+
77
+ if (s->ata) {
78
+ /* Extract the tag from the register to match STZGM. */
79
+ tag = tcg_temp_new_i64();
80
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
81
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
82
+ tcg_temp_free_i64(tag);
83
+ }
84
+ }
85
+ return;
86
+ case ARM_CP_DC_GZVA:
87
+ {
88
+ TCGv_i64 clean_addr, tag;
89
+
90
+ /* For DC_GZVA, we can rely on DC_ZVA for the proper fault. */
91
+ tcg_rt = cpu_reg(s, rt);
92
+ clean_addr = clean_data_tbi(s, tcg_rt);
93
+ gen_helper_dc_zva(cpu_env, clean_addr);
94
+
95
+ if (s->ata) {
96
+ /* Extract the tag from the register to match STZGM. */
97
+ tag = tcg_temp_new_i64();
98
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
99
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
100
+ tcg_temp_free_i64(tag);
101
+ }
102
+ }
103
+ return;
104
default:
105
break;
29
}
106
}
30
--
107
--
31
2.20.1
108
2.20.1
32
109
33
110
diff view generated by jsdifflib
1
The access_type argument to get_phys_addr_lpae() is an MMUAccessType;
1
From: Richard Henderson <richard.henderson@linaro.org>
2
use the enum constant MMU_DATA_LOAD rather than a literal 0 when we
3
call it in S1_ptw_translate().
4
2
3
D1.10 specifies that exception handlers begin with tag checks overridden.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-41-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200330210400.11724-3-peter.maydell@linaro.org
9
---
9
---
10
target/arm/helper.c | 5 +++--
10
target/arm/helper.c | 3 +++
11
1 file changed, 3 insertions(+), 2 deletions(-)
11
1 file changed, 3 insertions(+)
12
12
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
16
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
17
@@ -XXX,XX +XXX,XX @@ static hwaddr S1_ptw_translate(CPUARMState *env, ARMMMUIdx mmu_idx,
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
18
pcacheattrs = &cacheattrs;
18
break;
19
}
19
}
20
20
}
21
- ret = get_phys_addr_lpae(env, addr, 0, ARMMMUIdx_Stage2, &s2pa,
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
22
- &txattrs, &s2prot, &s2size, fi, pcacheattrs);
22
+ new_mode |= PSTATE_TCO;
23
+ ret = get_phys_addr_lpae(env, addr, MMU_DATA_LOAD, ARMMMUIdx_Stage2,
23
+ }
24
+ &s2pa, &txattrs, &s2prot, &s2size, fi,
24
25
+ pcacheattrs);
25
pstate_write(env, PSTATE_DAIF | new_mode);
26
if (ret) {
26
env->aarch64 = 1;
27
assert(fi->type != ARMFault_None);
28
fi->s2addr = addr;
29
--
27
--
30
2.20.1
28
2.20.1
31
29
32
30
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Fix typo xlnx-ve -> xlnx-versal.
3
We need to check the memattr of a page in order to determine
4
4
whether it is Tagged for MTE. Between Stage1 and Stage2,
5
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
this becomes simpler if we always collect this data, instead
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
of occasionally being presented with NULL.
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
8
Reviewed-by: Luc Michel <luc.michel@greensocs.com>
8
Use the nonnull attribute to allow the compiler to check that
9
Message-id: 20200427181649.26851-4-edgar.iglesias@gmail.com
9
all pointer arguments are non-null.
10
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200626033144.790098-42-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
15
---
12
hw/arm/xlnx-versal-virt.c | 2 +-
16
target/arm/internals.h | 3 ++-
13
1 file changed, 1 insertion(+), 1 deletion(-)
17
target/arm/helper.c | 60 ++++++++++++++++++++---------------------
14
18
target/arm/m_helper.c | 11 +++++---
15
diff --git a/hw/arm/xlnx-versal-virt.c b/hw/arm/xlnx-versal-virt.c
19
target/arm/tlb_helper.c | 4 ++-
16
index XXXXXXX..XXXXXXX 100644
20
4 files changed, 42 insertions(+), 36 deletions(-)
17
--- a/hw/arm/xlnx-versal-virt.c
21
18
+++ b/hw/arm/xlnx-versal-virt.c
22
diff --git a/target/arm/internals.h b/target/arm/internals.h
19
@@ -XXX,XX +XXX,XX @@ static void versal_virt_init(MachineState *machine)
23
index XXXXXXX..XXXXXXX 100644
20
psci_conduit = QEMU_PSCI_CONDUIT_SMC;
24
--- a/target/arm/internals.h
25
+++ b/target/arm/internals.h
26
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
27
MMUAccessType access_type, ARMMMUIdx mmu_idx,
28
hwaddr *phys_ptr, MemTxAttrs *attrs, int *prot,
29
target_ulong *page_size,
30
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
31
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
32
+ __attribute__((nonnull));
33
34
void arm_log_exception(int idx);
35
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/helper.c
39
+++ b/target/arm/helper.c
40
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
41
bool s1_is_el0,
42
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
43
target_ulong *page_size_ptr,
44
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
45
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
46
+ __attribute__((nonnull));
47
#endif
48
49
static void switch_mode(CPUARMState *env, int mode);
50
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
51
arm_tlb_bti_gp(txattrs) = true;
21
}
52
}
22
53
23
- sysbus_init_child_obj(OBJECT(machine), "xlnx-ve", &s->soc,
54
- if (cacheattrs != NULL) {
24
+ sysbus_init_child_obj(OBJECT(machine), "xlnx-versal", &s->soc,
55
- if (mmu_idx == ARMMMUIdx_Stage2) {
25
sizeof(s->soc), TYPE_XLNX_VERSAL);
56
- cacheattrs->attrs = convert_stage2_attrs(env,
26
object_property_set_link(OBJECT(&s->soc), OBJECT(machine->ram),
57
- extract32(attrs, 0, 4));
27
"ddr", &error_abort);
58
- } else {
59
- /* Index into MAIR registers for cache attributes */
60
- uint8_t attrindx = extract32(attrs, 0, 3);
61
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
62
- assert(attrindx <= 7);
63
- cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
64
- }
65
- cacheattrs->shareability = extract32(attrs, 6, 2);
66
+ if (mmu_idx == ARMMMUIdx_Stage2) {
67
+ cacheattrs->attrs = convert_stage2_attrs(env, extract32(attrs, 0, 4));
68
+ } else {
69
+ /* Index into MAIR registers for cache attributes */
70
+ uint8_t attrindx = extract32(attrs, 0, 3);
71
+ uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
72
+ assert(attrindx <= 7);
73
+ cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
74
}
75
+ cacheattrs->shareability = extract32(attrs, 6, 2);
76
77
*phys_ptr = descaddr;
78
*page_size_ptr = page_size;
79
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
80
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
81
mmu_idx == ARMMMUIdx_E10_0,
82
phys_ptr, attrs, &s2_prot,
83
- page_size, fi,
84
- cacheattrs != NULL ? &cacheattrs2 : NULL);
85
+ page_size, fi, &cacheattrs2);
86
fi->s2addr = ipa;
87
/* Combine the S1 and S2 perms. */
88
*prot &= s2_prot;
89
90
- /* Combine the S1 and S2 cache attributes, if needed */
91
- if (!ret && cacheattrs != NULL) {
92
- if (env->cp15.hcr_el2 & HCR_DC) {
93
- /*
94
- * HCR.DC forces the first stage attributes to
95
- * Normal Non-Shareable,
96
- * Inner Write-Back Read-Allocate Write-Allocate,
97
- * Outer Write-Back Read-Allocate Write-Allocate.
98
- */
99
- cacheattrs->attrs = 0xff;
100
- cacheattrs->shareability = 0;
101
- }
102
- *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
103
+ /* If S2 fails, return early. */
104
+ if (ret) {
105
+ return ret;
106
}
107
108
- return ret;
109
+ /* Combine the S1 and S2 cache attributes. */
110
+ if (env->cp15.hcr_el2 & HCR_DC) {
111
+ /*
112
+ * HCR.DC forces the first stage attributes to
113
+ * Normal Non-Shareable,
114
+ * Inner Write-Back Read-Allocate Write-Allocate,
115
+ * Outer Write-Back Read-Allocate Write-Allocate.
116
+ */
117
+ cacheattrs->attrs = 0xff;
118
+ cacheattrs->shareability = 0;
119
+ }
120
+ *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
121
+ return 0;
122
} else {
123
/*
124
* For non-EL2 CPUs a stage1+stage2 translation is just stage 1.
125
@@ -XXX,XX +XXX,XX @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
126
bool ret;
127
ARMMMUFaultInfo fi = {};
128
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
129
+ ARMCacheAttrs cacheattrs = {};
130
131
*attrs = (MemTxAttrs) {};
132
133
ret = get_phys_addr(env, addr, 0, mmu_idx, &phys_addr,
134
- attrs, &prot, &page_size, &fi, NULL);
135
+ attrs, &prot, &page_size, &fi, &cacheattrs);
136
137
if (ret) {
138
return -1;
139
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
140
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/m_helper.c
142
+++ b/target/arm/m_helper.c
143
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
144
hwaddr physaddr;
145
int prot;
146
ARMMMUFaultInfo fi = {};
147
+ ARMCacheAttrs cacheattrs = {};
148
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
149
int exc;
150
bool exc_secure;
151
152
if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &physaddr,
153
- &attrs, &prot, &page_size, &fi, NULL)) {
154
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
155
/* MPU/SAU lookup failed */
156
if (fi.type == ARMFault_QEMU_SFault) {
157
if (mode == STACK_LAZYFP) {
158
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
159
hwaddr physaddr;
160
int prot;
161
ARMMMUFaultInfo fi = {};
162
+ ARMCacheAttrs cacheattrs = {};
163
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
164
int exc;
165
bool exc_secure;
166
uint32_t value;
167
168
if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
169
- &attrs, &prot, &page_size, &fi, NULL)) {
170
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
171
/* MPU/SAU lookup failed */
172
if (fi.type == ARMFault_QEMU_SFault) {
173
qemu_log_mask(CPU_LOG_INT,
174
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
175
V8M_SAttributes sattrs = {};
176
MemTxAttrs attrs = {};
177
ARMMMUFaultInfo fi = {};
178
+ ARMCacheAttrs cacheattrs = {};
179
MemTxResult txres;
180
target_ulong page_size;
181
hwaddr physaddr;
182
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
183
"...really SecureFault with SFSR.INVEP\n");
184
return false;
185
}
186
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx,
187
- &physaddr, &attrs, &prot, &page_size, &fi, NULL)) {
188
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &physaddr,
189
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
190
/* the MPU lookup failed */
191
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
192
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
193
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
194
index XXXXXXX..XXXXXXX 100644
195
--- a/target/arm/tlb_helper.c
196
+++ b/target/arm/tlb_helper.c
197
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
198
int prot, ret;
199
MemTxAttrs attrs = {};
200
ARMMMUFaultInfo fi = {};
201
+ ARMCacheAttrs cacheattrs = {};
202
203
/*
204
* Walk the page table and (if the mapping exists) add the page
205
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
206
*/
207
ret = get_phys_addr(&cpu->env, address, access_type,
208
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
209
- &phys_addr, &attrs, &prot, &page_size, &fi, NULL);
210
+ &phys_addr, &attrs, &prot, &page_size,
211
+ &fi, &cacheattrs);
212
if (likely(!ret)) {
213
/*
214
* Map a single [sub]page. Regions smaller than our declared
28
--
215
--
29
2.20.1
216
2.20.1
30
217
31
218
diff view generated by jsdifflib
1
For ARMv8.2-TTS2UXN, the stage 2 page table walk wants to know
1
From: Richard Henderson <richard.henderson@linaro.org>
2
whether the stage 1 access is for EL0 or not, because whether
3
exec permission is given can depend on whether this is an EL0
4
or EL1 access. Add a new argument to get_phys_addr_lpae() so
5
the call sites can pass this information in.
6
2
7
Since get_phys_addr_lpae() doesn't already have a doc comment,
3
This "bit" is a particular value of the page's MemAttr.
8
add one so we have a place to put the documentation of the
9
semantics of the new s1_is_el0 argument.
10
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-43-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20200330210400.11724-4-peter.maydell@linaro.org
15
---
9
---
16
target/arm/helper.c | 29 ++++++++++++++++++++++++++++-
10
target/arm/helper.c | 48 ++++++++++++++++++++++++++++++++++++++---
17
1 file changed, 28 insertions(+), 1 deletion(-)
11
target/arm/tlb_helper.c | 5 +++++
12
2 files changed, 50 insertions(+), 3 deletions(-)
18
13
19
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
20
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
22
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
23
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ static uint8_t combine_cacheattr_nibble(uint8_t s1, uint8_t s2)
24
19
*/
25
static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
20
static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
26
MMUAccessType access_type, ARMMMUIdx mmu_idx,
21
{
27
+ bool s1_is_el0,
22
- uint8_t s1lo = extract32(s1.attrs, 0, 4), s2lo = extract32(s2.attrs, 0, 4);
28
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
23
- uint8_t s1hi = extract32(s1.attrs, 4, 4), s2hi = extract32(s2.attrs, 4, 4);
29
target_ulong *page_size_ptr,
24
+ uint8_t s1lo, s2lo, s1hi, s2hi;
30
ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
25
ARMCacheAttrs ret;
31
@@ -XXX,XX +XXX,XX @@ static hwaddr S1_ptw_translate(CPUARMState *env, ARMMMUIdx mmu_idx,
26
+ bool tagged = false;
27
+
28
+ if (s1.attrs == 0xf0) {
29
+ tagged = true;
30
+ s1.attrs = 0xff;
31
+ }
32
+
33
+ s1lo = extract32(s1.attrs, 0, 4);
34
+ s2lo = extract32(s2.attrs, 0, 4);
35
+ s1hi = extract32(s1.attrs, 4, 4);
36
+ s2hi = extract32(s2.attrs, 4, 4);
37
38
/* Combine shareability attributes (table D4-43) */
39
if (s1.shareability == 2 || s2.shareability == 2) {
40
@@ -XXX,XX +XXX,XX @@ static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
32
}
41
}
33
42
}
34
ret = get_phys_addr_lpae(env, addr, MMU_DATA_LOAD, ARMMMUIdx_Stage2,
43
35
+ false,
44
+ /* TODO: CombineS1S2Desc does not consider transient, only WB, RWA. */
36
&s2pa, &txattrs, &s2prot, &s2size, fi,
45
+ if (tagged && ret.attrs == 0xff) {
37
pcacheattrs);
46
+ ret.attrs = 0xf0;
38
if (ret) {
47
+ }
39
@@ -XXX,XX +XXX,XX @@ static ARMVAParameters aa32_va_parameters(CPUARMState *env, uint32_t va,
48
+
40
};
49
return ret;
41
}
50
}
42
51
43
+/**
44
+ * get_phys_addr_lpae: perform one stage of page table walk, LPAE format
45
+ *
46
+ * Returns false if the translation was successful. Otherwise, phys_ptr, attrs,
47
+ * prot and page_size may not be filled in, and the populated fsr value provides
48
+ * information on why the translation aborted, in the format of a long-format
49
+ * DFSR/IFSR fault register, with the following caveats:
50
+ * * the WnR bit is never set (the caller must do this).
51
+ *
52
+ * @env: CPUARMState
53
+ * @address: virtual address to get physical address for
54
+ * @access_type: MMU_DATA_LOAD, MMU_DATA_STORE or MMU_INST_FETCH
55
+ * @mmu_idx: MMU index indicating required translation regime
56
+ * @s1_is_el0: if @mmu_idx is ARMMMUIdx_Stage2 (so this is a stage 2 page table
57
+ * walk), must be true if this is stage 2 of a stage 1+2 walk for an
58
+ * EL0 access). If @mmu_idx is anything else, @s1_is_el0 is ignored.
59
+ * @phys_ptr: set to the physical address corresponding to the virtual address
60
+ * @attrs: set to the memory transaction attributes to use
61
+ * @prot: set to the permissions for the page containing phys_ptr
62
+ * @page_size_ptr: set to the size of the page containing phys_ptr
63
+ * @fi: set to fault info if the translation fails
64
+ * @cacheattrs: (if non-NULL) set to the cacheability/shareability attributes
65
+ */
66
static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
67
MMUAccessType access_type, ARMMMUIdx mmu_idx,
68
+ bool s1_is_el0,
69
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
70
target_ulong *page_size_ptr,
71
ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
72
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
52
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
73
53
* Normal Non-Shareable,
74
/* S1 is done. Now do S2 translation. */
54
* Inner Write-Back Read-Allocate Write-Allocate,
75
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
55
* Outer Write-Back Read-Allocate Write-Allocate.
76
+ mmu_idx == ARMMMUIdx_E10_0,
56
+ * Do not overwrite Tagged within attrs.
77
phys_ptr, attrs, &s2_prot,
57
*/
78
page_size, fi,
58
- cacheattrs->attrs = 0xff;
79
cacheattrs != NULL ? &cacheattrs2 : NULL);
59
+ if (cacheattrs->attrs != 0xf0) {
60
+ cacheattrs->attrs = 0xff;
61
+ }
62
cacheattrs->shareability = 0;
63
}
64
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
80
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
65
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
66
/* Definitely a real MMU, not an MPU */
67
68
if (regime_translation_disabled(env, mmu_idx)) {
69
+ uint64_t hcr;
70
+ uint8_t memattr;
71
+
72
/*
73
* MMU disabled. S1 addresses within aa64 translation regimes are
74
* still checked for bounds -- see AArch64.TranslateAddressS1Off.
75
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
76
*phys_ptr = address;
77
*prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
78
*page_size = TARGET_PAGE_SIZE;
79
+
80
+ /* Fill in cacheattr a-la AArch64.TranslateAddressS1Off. */
81
+ hcr = arm_hcr_el2_eff(env);
82
+ cacheattrs->shareability = 0;
83
+ if (hcr & HCR_DC) {
84
+ if (hcr & HCR_DCT) {
85
+ memattr = 0xf0; /* Tagged, Normal, WB, RWA */
86
+ } else {
87
+ memattr = 0xff; /* Normal, WB, RWA */
88
+ }
89
+ } else if (access_type == MMU_INST_FETCH) {
90
+ if (regime_sctlr(env, mmu_idx) & SCTLR_I) {
91
+ memattr = 0xee; /* Normal, WT, RA, NT */
92
+ } else {
93
+ memattr = 0x44; /* Normal, NC, No */
94
+ }
95
+ cacheattrs->shareability = 2; /* outer sharable */
96
+ } else {
97
+ memattr = 0x00; /* Device, nGnRnE */
98
+ }
99
+ cacheattrs->attrs = memattr;
100
return 0;
81
}
101
}
82
102
83
if (regime_using_lpae_format(env, mmu_idx)) {
103
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
84
- return get_phys_addr_lpae(env, address, access_type, mmu_idx,
104
index XXXXXXX..XXXXXXX 100644
85
+ return get_phys_addr_lpae(env, address, access_type, mmu_idx, false,
105
--- a/target/arm/tlb_helper.c
86
phys_ptr, attrs, prot, page_size,
106
+++ b/target/arm/tlb_helper.c
87
fi, cacheattrs);
107
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
88
} else if (regime_sctlr(env, mmu_idx) & SCTLR_XP) {
108
phys_addr &= TARGET_PAGE_MASK;
109
address &= TARGET_PAGE_MASK;
110
}
111
+ /* Notice and record tagged memory. */
112
+ if (cpu_isar_feature(aa64_mte, cpu) && cacheattrs.attrs == 0xf0) {
113
+ arm_tlb_mte_tagged(&attrs) = true;
114
+ }
115
+
116
tlb_set_page_with_attrs(cs, address, phys_addr, attrs,
117
prot, mmu_idx, page_size);
118
return true;
89
--
119
--
90
2.20.1
120
2.20.1
91
121
92
122
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
MIDR_EL1 is a 64-bit system register with the top 32-bit being RES0.
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Represent it in QEMU's ARMCPU struct with a uint64_t, not a
5
uint32_t.
6
7
This fixes an error when compiling with -Werror=conversion
8
because we were manipulating the register value using a
9
local uint64_t variable:
10
11
target/arm/cpu64.c: In function ‘aarch64_max_initfn’:
12
target/arm/cpu64.c:628:21: error: conversion from ‘uint64_t’ {aka ‘long unsigned int’} to ‘uint32_t’ {aka ‘unsigned int’} may change value [-Werror=conversion]
13
628 | cpu->midr = t;
14
| ^
15
16
and future-proofs us against a possible future architecture
17
change using some of the top 32 bits.
18
19
Suggested-by: Laurent Desnogues <laurent.desnogues@gmail.com>
20
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
21
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
22
Reviewed-by: Laurent Desnogues <laurent.desnogues@gmail.com>
23
Message-id: 20200428172634.29707-1-f4bug@amsat.org
24
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-44-richard.henderson@linaro.org
25
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
26
---
7
---
27
target/arm/cpu.h | 2 +-
8
target/arm/cpu.h | 6 ++++++
28
target/arm/cpu.c | 2 +-
9
hw/arm/virt.c | 55 ++++++++++++++++++++++++++++++++++++++++++++++--
29
2 files changed, 2 insertions(+), 2 deletions(-)
10
target/arm/cpu.c | 52 +++++++++++++++++++++++++++++++++++++++++----
11
3 files changed, 107 insertions(+), 6 deletions(-)
30
12
31
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
32
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/cpu.h
15
--- a/target/arm/cpu.h
34
+++ b/target/arm/cpu.h
16
+++ b/target/arm/cpu.h
35
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
17
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
36
uint64_t id_aa64dfr0;
18
/* MemoryRegion to use for secure physical accesses */
37
uint64_t id_aa64dfr1;
19
MemoryRegion *secure_memory;
38
} isar;
20
39
- uint32_t midr;
21
+ /* MemoryRegion to use for allocation tag accesses */
40
+ uint64_t midr;
22
+ MemoryRegion *tag_memory;
41
uint32_t revidr;
23
+ MemoryRegion *secure_tag_memory;
42
uint32_t reset_fpsid;
24
+
43
uint32_t ctr;
25
/* For v8M, pointer to the IDAU interface provided by board/SoC */
26
Object *idau;
27
28
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
29
typedef enum ARMASIdx {
30
ARMASIdx_NS = 0,
31
ARMASIdx_S = 1,
32
+ ARMASIdx_TagNS = 2,
33
+ ARMASIdx_TagS = 3,
34
} ARMASIdx;
35
36
/* Return the Exception Level targeted by debug exceptions. */
37
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
38
index XXXXXXX..XXXXXXX 100644
39
--- a/hw/arm/virt.c
40
+++ b/hw/arm/virt.c
41
@@ -XXX,XX +XXX,XX @@ static void create_platform_bus(VirtMachineState *vms)
42
sysbus_mmio_get_region(s, 0));
43
}
44
45
+static void create_tag_ram(MemoryRegion *tag_sysmem,
46
+ hwaddr base, hwaddr size,
47
+ const char *name)
48
+{
49
+ MemoryRegion *tagram = g_new(MemoryRegion, 1);
50
+
51
+ memory_region_init_ram(tagram, NULL, name, size / 32, &error_fatal);
52
+ memory_region_add_subregion(tag_sysmem, base / 32, tagram);
53
+}
54
+
55
static void create_secure_ram(VirtMachineState *vms,
56
- MemoryRegion *secure_sysmem)
57
+ MemoryRegion *secure_sysmem,
58
+ MemoryRegion *secure_tag_sysmem)
59
{
60
MemoryRegion *secram = g_new(MemoryRegion, 1);
61
char *nodename;
62
@@ -XXX,XX +XXX,XX @@ static void create_secure_ram(VirtMachineState *vms,
63
qemu_fdt_setprop_string(vms->fdt, nodename, "status", "disabled");
64
qemu_fdt_setprop_string(vms->fdt, nodename, "secure-status", "okay");
65
66
+ if (secure_tag_sysmem) {
67
+ create_tag_ram(secure_tag_sysmem, base, size, "mach-virt.secure-tag");
68
+ }
69
+
70
g_free(nodename);
71
}
72
73
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
74
const CPUArchIdList *possible_cpus;
75
MemoryRegion *sysmem = get_system_memory();
76
MemoryRegion *secure_sysmem = NULL;
77
+ MemoryRegion *tag_sysmem = NULL;
78
+ MemoryRegion *secure_tag_sysmem = NULL;
79
int n, virt_max_cpus;
80
bool firmware_loaded;
81
bool aarch64 = true;
82
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
83
"secure-memory", &error_abort);
84
}
85
86
+ /*
87
+ * The cpu adds the property if and only if MemTag is supported.
88
+ * If it is, we must allocate the ram to back that up.
89
+ */
90
+ if (object_property_find(cpuobj, "tag-memory", NULL)) {
91
+ if (!tag_sysmem) {
92
+ tag_sysmem = g_new(MemoryRegion, 1);
93
+ memory_region_init(tag_sysmem, OBJECT(machine),
94
+ "tag-memory", UINT64_MAX / 32);
95
+
96
+ if (vms->secure) {
97
+ secure_tag_sysmem = g_new(MemoryRegion, 1);
98
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
99
+ "secure-tag-memory", UINT64_MAX / 32);
100
+
101
+ /* As with ram, secure-tag takes precedence over tag. */
102
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
103
+ tag_sysmem, -1);
104
+ }
105
+ }
106
+
107
+ object_property_set_link(cpuobj, OBJECT(tag_sysmem),
108
+ "tag-memory", &error_abort);
109
+ if (vms->secure) {
110
+ object_property_set_link(cpuobj, OBJECT(secure_tag_sysmem),
111
+ "secure-tag-memory", &error_abort);
112
+ }
113
+ }
114
+
115
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
116
object_unref(cpuobj);
117
}
118
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
119
create_uart(vms, VIRT_UART, sysmem, serial_hd(0));
120
121
if (vms->secure) {
122
- create_secure_ram(vms, secure_sysmem);
123
+ create_secure_ram(vms, secure_sysmem, secure_tag_sysmem);
124
create_uart(vms, VIRT_SECURE_UART, secure_sysmem, serial_hd(1));
125
}
126
127
+ if (tag_sysmem) {
128
+ create_tag_ram(tag_sysmem, vms->memmap[VIRT_MEM].base,
129
+ machine->ram_size, "mach-virt.tag");
130
+ }
131
+
132
vms->highmem_ecam &= vms->highmem && (!firmware_loaded || aarch64);
133
134
create_rtc(vms);
44
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
135
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
45
index XXXXXXX..XXXXXXX 100644
136
index XXXXXXX..XXXXXXX 100644
46
--- a/target/arm/cpu.c
137
--- a/target/arm/cpu.c
47
+++ b/target/arm/cpu.c
138
+++ b/target/arm/cpu.c
48
@@ -XXX,XX +XXX,XX @@ static const ARMCPUInfo arm_cpus[] = {
139
@@ -XXX,XX +XXX,XX @@ void arm_cpu_post_init(Object *obj)
49
static Property arm_cpu_properties[] = {
140
if (kvm_enabled()) {
50
DEFINE_PROP_BOOL("start-powered-off", ARMCPU, start_powered_off, false),
141
kvm_arm_add_vcpu_properties(obj);
51
DEFINE_PROP_UINT32("psci-conduit", ARMCPU, psci_conduit, 0),
142
}
52
- DEFINE_PROP_UINT32("midr", ARMCPU, midr, 0),
143
+
53
+ DEFINE_PROP_UINT64("midr", ARMCPU, midr, 0),
144
+#ifndef CONFIG_USER_ONLY
54
DEFINE_PROP_UINT64("mp-affinity", ARMCPU,
145
+ if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) &&
55
mp_affinity, ARM64_AFFINITY_INVALID),
146
+ cpu_isar_feature(aa64_mte, cpu)) {
56
DEFINE_PROP_INT32("node-id", ARMCPU, node_id, CPU_UNSET_NUMA_NODE_ID),
147
+ object_property_add_link(obj, "tag-memory",
148
+ TYPE_MEMORY_REGION,
149
+ (Object **)&cpu->tag_memory,
150
+ qdev_prop_allow_set_link_before_realize,
151
+ OBJ_PROP_LINK_STRONG);
152
+
153
+ if (arm_feature(&cpu->env, ARM_FEATURE_EL3)) {
154
+ object_property_add_link(obj, "secure-tag-memory",
155
+ TYPE_MEMORY_REGION,
156
+ (Object **)&cpu->secure_tag_memory,
157
+ qdev_prop_allow_set_link_before_realize,
158
+ OBJ_PROP_LINK_STRONG);
159
+ }
160
+ }
161
+#endif
162
}
163
164
static void arm_cpu_finalizefn(Object *obj)
165
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
166
#ifndef CONFIG_USER_ONLY
167
MachineState *ms = MACHINE(qdev_get_machine());
168
unsigned int smp_cpus = ms->smp.cpus;
169
+ bool has_secure = cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY);
170
171
- if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
172
- cs->num_ases = 2;
173
+ /*
174
+ * We must set cs->num_ases to the final value before
175
+ * the first call to cpu_address_space_init.
176
+ */
177
+ if (cpu->tag_memory != NULL) {
178
+ cs->num_ases = 3 + has_secure;
179
+ } else {
180
+ cs->num_ases = 1 + has_secure;
181
+ }
182
183
+ if (has_secure) {
184
if (!cpu->secure_memory) {
185
cpu->secure_memory = cs->memory;
186
}
187
cpu_address_space_init(cs, ARMASIdx_S, "cpu-secure-memory",
188
cpu->secure_memory);
189
- } else {
190
- cs->num_ases = 1;
191
}
192
+
193
+ if (cpu->tag_memory != NULL) {
194
+ cpu_address_space_init(cs, ARMASIdx_TagNS, "cpu-tag-memory",
195
+ cpu->tag_memory);
196
+ if (has_secure) {
197
+ cpu_address_space_init(cs, ARMASIdx_TagS, "cpu-tag-memory",
198
+ cpu->secure_tag_memory);
199
+ }
200
+ } else if (cpu_isar_feature(aa64_mte, cpu)) {
201
+ /*
202
+ * Since there is no tag memory, we can't meaningfully support MTE
203
+ * to its fullest. To avoid problems later, when we would come to
204
+ * use the tag memory, downgrade support to insns only.
205
+ */
206
+ cpu->isar.id_aa64pfr1 =
207
+ FIELD_DP64(cpu->isar.id_aa64pfr1, ID_AA64PFR1, MTE, 1);
208
+ }
209
+
210
cpu_address_space_init(cs, ARMASIdx_NS, "cpu-memory", cs->memory);
211
212
/* No core_count specified, default to smp_cpus. */
57
--
213
--
58
2.20.1
214
2.20.1
59
215
60
216
diff view generated by jsdifflib
1
Convert the VFM[AS]L (vector) insns to decodetree. This is the last
1
From: Richard Henderson <richard.henderson@linaro.org>
2
insn in the legacy decoder for the 3same_ext group, so we can
3
delete the legacy decoder function for the group entirely.
4
2
5
Note that in disas_thumb2_insn() the parts of this encoding space
3
Look up the physical address for the given virtual address,
6
where the decodetree decoder returns false will correctly be directed
4
convert that to a tag physical address, and finally return
7
to illegal_op by the "(insn & (1 << 28))" check so they won't fall
5
the host address that backs it.
8
into disas_coproc_insn() by mistake.
9
6
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-45-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200430181003.21682-8-peter.maydell@linaro.org
13
---
11
---
14
target/arm/neon-shared.decode | 6 +++
12
target/arm/mte_helper.c | 131 ++++++++++++++++++++++++++++++++++++++++
15
target/arm/translate-neon.inc.c | 31 +++++++++++
13
1 file changed, 131 insertions(+)
16
target/arm/translate.c | 92 +--------------------------------
17
3 files changed, 38 insertions(+), 91 deletions(-)
18
14
19
diff --git a/target/arm/neon-shared.decode b/target/arm/neon-shared.decode
15
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
20
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/neon-shared.decode
17
--- a/target/arm/mte_helper.c
22
+++ b/target/arm/neon-shared.decode
18
+++ b/target/arm/mte_helper.c
23
@@ -XXX,XX +XXX,XX @@ VCADD 1111 110 rot:1 1 . 0 size:1 .... .... 1000 . q:1 . 0 .... \
19
@@ -XXX,XX +XXX,XX @@
24
# VUDOT and VSDOT
20
#include "cpu.h"
25
VDOT 1111 110 00 . 10 .... .... 1101 . q:1 . u:1 .... \
21
#include "internals.h"
26
vm=%vm_dp vn=%vn_dp vd=%vd_dp
22
#include "exec/exec-all.h"
23
+#include "exec/ram_addr.h"
24
#include "exec/cpu_ldst.h"
25
#include "exec/helper-proto.h"
26
27
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
28
int ptr_size, MMUAccessType tag_access,
29
int tag_size, uintptr_t ra)
30
{
31
+#ifdef CONFIG_USER_ONLY
32
/* Tag storage not implemented. */
33
return NULL;
34
+#else
35
+ uintptr_t index;
36
+ CPUIOTLBEntry *iotlbentry;
37
+ int in_page, flags;
38
+ ram_addr_t ptr_ra;
39
+ hwaddr ptr_paddr, tag_paddr, xlat;
40
+ MemoryRegion *mr;
41
+ ARMASIdx tag_asi;
42
+ AddressSpace *tag_as;
43
+ void *host;
27
+
44
+
28
+# VFM[AS]L
45
+ /*
29
+VFML 1111 110 0 s:1 . 10 .... .... 1000 . 0 . 1 .... \
46
+ * Probe the first byte of the virtual address. This raises an
30
+ vm=%vm_sp vn=%vn_sp vd=%vd_dp q=0
47
+ * exception for inaccessible pages, and resolves the virtual address
31
+VFML 1111 110 0 s:1 . 10 .... .... 1000 . 1 . 1 .... \
48
+ * into the softmmu tlb.
32
+ vm=%vm_dp vn=%vn_dp vd=%vd_dp q=1
49
+ *
33
diff --git a/target/arm/translate-neon.inc.c b/target/arm/translate-neon.inc.c
50
+ * When RA == 0, this is for mte_probe1. The page is expected to be
34
index XXXXXXX..XXXXXXX 100644
51
+ * valid. Indicate to probe_access_flags no-fault, then assert that
35
--- a/target/arm/translate-neon.inc.c
52
+ * we received a valid page.
36
+++ b/target/arm/translate-neon.inc.c
53
+ */
37
@@ -XXX,XX +XXX,XX @@ static bool trans_VDOT(DisasContext *s, arg_VDOT *a)
54
+ flags = probe_access_flags(env, ptr, ptr_access, ptr_mmu_idx,
38
opr_sz, opr_sz, 0, fn_gvec);
55
+ ra == 0, &host, ra);
39
return true;
56
+ assert(!(flags & TLB_INVALID_MASK));
40
}
41
+
57
+
42
+static bool trans_VFML(DisasContext *s, arg_VFML *a)
58
+ /*
43
+{
59
+ * Find the iotlbentry for ptr. This *must* be present in the TLB
44
+ int opr_sz;
60
+ * because we just found the mapping.
61
+ * TODO: Perhaps there should be a cputlb helper that returns a
62
+ * matching tlb entry + iotlb entry.
63
+ */
64
+ index = tlb_index(env, ptr_mmu_idx, ptr);
65
+# ifdef CONFIG_DEBUG_TCG
66
+ {
67
+ CPUTLBEntry *entry = tlb_entry(env, ptr_mmu_idx, ptr);
68
+ target_ulong comparator = (ptr_access == MMU_DATA_LOAD
69
+ ? entry->addr_read
70
+ : tlb_addr_write(entry));
71
+ g_assert(tlb_hit(comparator, ptr));
72
+ }
73
+# endif
74
+ iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
45
+
75
+
46
+ if (!dc_isar_feature(aa32_fhm, s)) {
76
+ /* If the virtual page MemAttr != Tagged, access unchecked. */
47
+ return false;
77
+ if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
78
+ return NULL;
48
+ }
79
+ }
49
+
80
+
50
+ /* UNDEF accesses to D16-D31 if they don't exist. */
81
+ /*
51
+ if (!dc_isar_feature(aa32_simd_r32, s) &&
82
+ * If not backed by host ram, there is no tag storage: access unchecked.
52
+ (a->vd & 0x10)) {
83
+ * This is probably a guest os bug though, so log it.
53
+ return false;
84
+ */
85
+ if (unlikely(flags & TLB_MMIO)) {
86
+ qemu_log_mask(LOG_GUEST_ERROR,
87
+ "Page @ 0x%" PRIx64 " indicates Tagged Normal memory "
88
+ "but is not backed by host ram\n", ptr);
89
+ return NULL;
54
+ }
90
+ }
55
+
91
+
56
+ if (a->vd & a->q) {
92
+ /*
57
+ return false;
93
+ * The Normal memory access can extend to the next page. E.g. a single
94
+ * 8-byte access to the last byte of a page will check only the last
95
+ * tag on the first page.
96
+ * Any page access exception has priority over tag check exception.
97
+ */
98
+ in_page = -(ptr | TARGET_PAGE_MASK);
99
+ if (unlikely(ptr_size > in_page)) {
100
+ void *ignore;
101
+ flags |= probe_access_flags(env, ptr + in_page, ptr_access,
102
+ ptr_mmu_idx, ra == 0, &ignore, ra);
103
+ assert(!(flags & TLB_INVALID_MASK));
58
+ }
104
+ }
59
+
105
+
60
+ if (!vfp_access_check(s)) {
106
+ /* Any debug exception has priority over a tag check exception. */
61
+ return true;
107
+ if (unlikely(flags & TLB_WATCHPOINT)) {
108
+ int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
109
+ assert(ra != 0);
110
+ cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
111
+ iotlbentry->attrs, wp, ra);
62
+ }
112
+ }
63
+
113
+
64
+ opr_sz = (1 + a->q) * 8;
114
+ /*
65
+ tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd),
115
+ * Find the physical address within the normal mem space.
66
+ vfp_reg_offset(a->q, a->vn),
116
+ * The memory region lookup must succeed because TLB_MMIO was
67
+ vfp_reg_offset(a->q, a->vm),
117
+ * not set in the cputlb lookup above.
68
+ cpu_env, opr_sz, opr_sz, a->s, /* is_2 == 0 */
118
+ */
69
+ gen_helper_gvec_fmlal_a32);
119
+ mr = memory_region_from_host(host, &ptr_ra);
70
+ return true;
120
+ tcg_debug_assert(mr != NULL);
71
+}
121
+ tcg_debug_assert(memory_region_is_ram(mr));
72
diff --git a/target/arm/translate.c b/target/arm/translate.c
122
+ ptr_paddr = ptr_ra;
73
index XXXXXXX..XXXXXXX 100644
123
+ do {
74
--- a/target/arm/translate.c
124
+ ptr_paddr += mr->addr;
75
+++ b/target/arm/translate.c
125
+ mr = mr->container;
76
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
126
+ } while (mr);
77
return 0;
127
+
128
+ /* Convert to the physical address in tag space. */
129
+ tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
130
+
131
+ /* Look up the address in tag space. */
132
+ tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
133
+ tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
134
+ mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
135
+ tag_access == MMU_DATA_STORE,
136
+ iotlbentry->attrs);
137
+
138
+ /*
139
+ * Note that @mr will never be NULL. If there is nothing in the address
140
+ * space at @tag_paddr, the translation will return the unallocated memory
141
+ * region. For our purposes, the result must be ram.
142
+ */
143
+ if (unlikely(!memory_region_is_ram(mr))) {
144
+ /* ??? Failure is a board configuration error. */
145
+ qemu_log_mask(LOG_UNIMP,
146
+ "Tag Memory @ 0x%" HWADDR_PRIx " not found for "
147
+ "Normal Memory @ 0x%" HWADDR_PRIx "\n",
148
+ tag_paddr, ptr_paddr);
149
+ return NULL;
150
+ }
151
+
152
+ /*
153
+ * Ensure the tag memory is dirty on write, for migration.
154
+ * Tag memory can never contain code or display memory (vga).
155
+ */
156
+ if (tag_access == MMU_DATA_STORE) {
157
+ ram_addr_t tag_ra = memory_region_get_ram_addr(mr) + xlat;
158
+ cpu_physical_memory_set_dirty_flag(tag_ra, DIRTY_MEMORY_MIGRATION);
159
+ }
160
+
161
+ return memory_region_get_ram_ptr(mr) + xlat;
162
+#endif
78
}
163
}
79
164
80
-/* Advanced SIMD three registers of the same length extension.
165
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
81
- * 31 25 23 22 20 16 12 11 10 9 8 3 0
82
- * +---------------+-----+---+-----+----+----+---+----+---+----+---------+----+
83
- * | 1 1 1 1 1 1 0 | op1 | D | op2 | Vn | Vd | 1 | o3 | 0 | o4 | N Q M U | Vm |
84
- * +---------------+-----+---+-----+----+----+---+----+---+----+---------+----+
85
- */
86
-static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
87
-{
88
- gen_helper_gvec_3 *fn_gvec = NULL;
89
- gen_helper_gvec_3_ptr *fn_gvec_ptr = NULL;
90
- int rd, rn, rm, opr_sz;
91
- int data = 0;
92
- int off_rn, off_rm;
93
- bool is_long = false, q = extract32(insn, 6, 1);
94
- bool ptr_is_env = false;
95
-
96
- if ((insn & 0xff300f10) == 0xfc200810) {
97
- /* VFM[AS]L -- 1111 1100 S.10 .... .... 1000 .Q.1 .... */
98
- int is_s = extract32(insn, 23, 1);
99
- if (!dc_isar_feature(aa32_fhm, s)) {
100
- return 1;
101
- }
102
- is_long = true;
103
- data = is_s; /* is_2 == 0 */
104
- fn_gvec_ptr = gen_helper_gvec_fmlal_a32;
105
- ptr_is_env = true;
106
- } else {
107
- return 1;
108
- }
109
-
110
- VFP_DREG_D(rd, insn);
111
- if (rd & q) {
112
- return 1;
113
- }
114
- if (q || !is_long) {
115
- VFP_DREG_N(rn, insn);
116
- VFP_DREG_M(rm, insn);
117
- if ((rn | rm) & q & !is_long) {
118
- return 1;
119
- }
120
- off_rn = vfp_reg_offset(1, rn);
121
- off_rm = vfp_reg_offset(1, rm);
122
- } else {
123
- rn = VFP_SREG_N(insn);
124
- rm = VFP_SREG_M(insn);
125
- off_rn = vfp_reg_offset(0, rn);
126
- off_rm = vfp_reg_offset(0, rm);
127
- }
128
-
129
- if (s->fp_excp_el) {
130
- gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
131
- syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
132
- return 0;
133
- }
134
- if (!s->vfp_enabled) {
135
- return 1;
136
- }
137
-
138
- opr_sz = (1 + q) * 8;
139
- if (fn_gvec_ptr) {
140
- TCGv_ptr ptr;
141
- if (ptr_is_env) {
142
- ptr = cpu_env;
143
- } else {
144
- ptr = get_fpstatus_ptr(1);
145
- }
146
- tcg_gen_gvec_3_ptr(vfp_reg_offset(1, rd), off_rn, off_rm, ptr,
147
- opr_sz, opr_sz, data, fn_gvec_ptr);
148
- if (!ptr_is_env) {
149
- tcg_temp_free_ptr(ptr);
150
- }
151
- } else {
152
- tcg_gen_gvec_3_ool(vfp_reg_offset(1, rd), off_rn, off_rm,
153
- opr_sz, opr_sz, data, fn_gvec);
154
- }
155
- return 0;
156
-}
157
-
158
/* Advanced SIMD two registers and a scalar extension.
159
* 31 24 23 22 20 16 12 11 10 9 8 3 0
160
* +-----------------+----+---+----+----+----+---+----+---+----+---------+----+
161
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
162
}
163
}
164
}
165
- } else if ((insn & 0x0e000a00) == 0x0c000800
166
- && arm_dc_feature(s, ARM_FEATURE_V8)) {
167
- if (disas_neon_insn_3same_ext(s, insn)) {
168
- goto illegal_op;
169
- }
170
- return;
171
} else if ((insn & 0x0f000a00) == 0x0e000800
172
&& arm_dc_feature(s, ARM_FEATURE_V8)) {
173
if (disas_neon_insn_2reg_scalar_ext(s, insn)) {
174
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
175
}
176
break;
177
}
178
- if ((insn & 0xfe000a00) == 0xfc000800
179
+ if ((insn & 0xff000a00) == 0xfe000800
180
&& arm_dc_feature(s, ARM_FEATURE_V8)) {
181
/* The Thumb2 and ARM encodings are identical. */
182
- if (disas_neon_insn_3same_ext(s, insn)) {
183
- goto illegal_op;
184
- }
185
- } else if ((insn & 0xff000a00) == 0xfe000800
186
- && arm_dc_feature(s, ARM_FEATURE_V8)) {
187
- /* The Thumb2 and ARM encodings are identical. */
188
if (disas_neon_insn_2reg_scalar_ext(s, insn)) {
189
goto illegal_op;
190
}
191
--
166
--
192
2.20.1
167
2.20.1
193
168
194
169
diff view generated by jsdifflib
1
In aarch64_max_initfn() we update both 32-bit and 64-bit ID
1
From: Richard Henderson <richard.henderson@linaro.org>
2
registers. The intended pattern is that for 64-bit ID registers we
3
use FIELD_DP64 and the uint64_t 't' register, while 32-bit ID
4
registers use FIELD_DP32 and the uint32_t 'u' register. For
5
ID_AA64DFR0 we accidentally used 'u', meaning that the top 32 bits of
6
this 64-bit ID register would end up always zero. Luckily at the
7
moment that's what they should be anyway, so this bug has no visible
8
effects.
9
2
10
Use the right-sized variable.
3
We now implement all of the components of MTE, without actually
4
supporting any tagged memory. All MTE instructions will work,
5
trivially, so we can enable support.
11
6
12
Fixes: 3bec78447a958d481991
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-46-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Laurent Desnogues <laurent.desnogues@gmail.com>
15
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
16
Message-id: 20200423110915.10527-1-peter.maydell@linaro.org
17
---
11
---
18
target/arm/cpu64.c | 6 +++---
12
target/arm/cpu64.c | 5 +++++
19
1 file changed, 3 insertions(+), 3 deletions(-)
13
1 file changed, 5 insertions(+)
20
14
21
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
15
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
22
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/cpu64.c
17
--- a/target/arm/cpu64.c
24
+++ b/target/arm/cpu64.c
18
+++ b/target/arm/cpu64.c
25
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
19
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
26
u = FIELD_DP32(u, ID_MMFR4, XNX, 1); /* TTS2UXN */
20
27
cpu->isar.id_mmfr4 = u;
21
t = cpu->isar.id_aa64pfr1;
28
22
t = FIELD_DP64(t, ID_AA64PFR1, BT, 1);
29
- u = cpu->isar.id_aa64dfr0;
23
+ /*
30
- u = FIELD_DP64(u, ID_AA64DFR0, PMUVER, 5); /* v8.4-PMU */
24
+ * Begin with full support for MTE; will be downgraded to MTE=1
31
- cpu->isar.id_aa64dfr0 = u;
25
+ * during realize if the board provides no tag memory.
32
+ t = cpu->isar.id_aa64dfr0;
26
+ */
33
+ t = FIELD_DP64(t, ID_AA64DFR0, PMUVER, 5); /* v8.4-PMU */
27
+ t = FIELD_DP64(t, ID_AA64PFR1, MTE, 2);
34
+ cpu->isar.id_aa64dfr0 = t;
28
cpu->isar.id_aa64pfr1 = t;
35
29
36
u = cpu->isar.id_dfr0;
30
t = cpu->isar.id_aa64mmfr1;
37
u = FIELD_DP32(u, ID_DFR0, PERFMON, 5); /* v8.4-PMU */
38
--
31
--
39
2.20.1
32
2.20.1
40
33
41
34
diff view generated by jsdifflib