1
As promised, another pullreq... This one's mostly RTH's patches.
1
Mostly this is RTH's memtag series, but there are also some cleanups
2
from Philippe.
2
3
3
thanks
4
thanks
4
-- PMM
5
-- PMM
5
6
6
The following changes since commit 784c2e4f232adf5ef47a84a262ec72a07d068d6a:
7
The following changes since commit 10f7ffabf9c507fc02382b89912003b1c43c3231:
7
8
8
Merge remote-tracking branch 'remotes/jasowang/tags/net-pull-request' into staging (2018-10-19 15:30:40 +0100)
9
Merge remote-tracking branch 'remotes/mcayland/tags/qemu-macppc-20200626' into staging (2020-06-26 12:14:18 +0100)
9
10
10
are available in the Git repository at:
11
are available in the Git repository at:
11
12
12
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20181019
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200626
13
14
14
for you to fetch changes up to 88c9add25e7120e8622796c81ad3f3fb7f8d40e7:
15
for you to fetch changes up to c7459633baa71d1781fde4a245d6ec9ce2f008cf:
15
16
16
target/arm: Only flush tlb if ASID changes (2018-10-19 17:38:48 +0100)
17
target/arm: Enable MTE (2020-06-26 14:32:24 +0100)
17
18
18
----------------------------------------------------------------
19
----------------------------------------------------------------
19
target-arm queue:
20
target-arm queue:
20
* ssi-sd: Make devices picking up backends unavailable with -device
21
* hw/arm/aspeed: improve QOM usage
21
* Add support for VCPU event states
22
* hw/misc/pca9552: trace GPIO change events
22
* Move towards making ID registers the source of truth for
23
* target/arm: Implement ARMv8.5-MemTag for system emulation
23
whether a guest CPU implements a feature, rather than having
24
parallel ID registers and feature bit flags
25
* Implement various HCR hypervisor trap/config bits
26
* Get IL bit correct for v7 syndrome values
27
* Report correct syndrome for FP/SIMD traps to Hyp mode
28
* hw/arm/boot: Increase compliance with kernel arm64 boot protocol
29
* Refactor A32 Neon to use generic vector infrastructure
30
* Fix a bug in A32 VLD2 "(multiple 2-element structures)" insn
31
* net: cadence_gem: Report features correctly in ID register
32
* Avoid some unnecessary TLB flushes on TTBR register writes
33
24
34
----------------------------------------------------------------
25
----------------------------------------------------------------
35
Dongjiu Geng (1):
26
Philippe Mathieu-Daudé (12):
36
target/arm: Add support for VCPU event states
27
hw/arm/aspeed: Remove extraneous MemoryRegion object owner
28
hw/arm/aspeed: Rename AspeedBoardState as AspeedMachineState
29
hw/arm/aspeed: QOM'ify AspeedMachineState
30
hw/i2c/core: Add i2c_try_create_slave() and i2c_realize_and_unref()
31
hw/misc/pca9552: Rename 'nr_leds' as 'pin_count'
32
hw/misc/pca9552: Rename generic code as pca955x
33
hw/misc/pca9552: Add generic PCA955xClass, parent of TYPE_PCA9552
34
hw/misc/pca9552: Add a 'description' property for debugging purpose
35
hw/misc/pca9552: Trace GPIO High/Low events
36
hw/arm/aspeed: Describe each PCA9552 device
37
hw/misc/pca9552: Trace GPIO change events
38
hw/misc/pca9552: Model qdev output GPIOs
37
39
38
Edgar E. Iglesias (2):
40
Richard Henderson (45):
39
net: cadence_gem: Announce availability of priority queues
41
target/arm: Add isar tests for mte
40
net: cadence_gem: Announce 64bit addressing support
42
target/arm: Improve masking of SCR RES0 bits
43
target/arm: Add support for MTE to SCTLR_ELx
44
target/arm: Add support for MTE to HCR_EL2 and SCR_EL3
45
target/arm: Rename DISAS_UPDATE to DISAS_UPDATE_EXIT
46
target/arm: Add DISAS_UPDATE_NOCHAIN
47
target/arm: Add MTE system registers
48
target/arm: Add MTE bits to tb_flags
49
target/arm: Implement the IRG instruction
50
target/arm: Revise decoding for disas_add_sub_imm
51
target/arm: Implement the ADDG, SUBG instructions
52
target/arm: Implement the GMI instruction
53
target/arm: Implement the SUBP instruction
54
target/arm: Define arm_cpu_do_unaligned_access for user-only
55
target/arm: Implement LDG, STG, ST2G instructions
56
target/arm: Implement the STGP instruction
57
target/arm: Restrict the values of DCZID.BS under TCG
58
target/arm: Simplify DC_ZVA
59
target/arm: Implement the LDGM, STGM, STZGM instructions
60
target/arm: Implement the access tag cache flushes
61
target/arm: Move regime_el to internals.h
62
target/arm: Move regime_tcr to internals.h
63
target/arm: Add gen_mte_check1
64
target/arm: Add gen_mte_checkN
65
target/arm: Implement helper_mte_check1
66
target/arm: Implement helper_mte_checkN
67
target/arm: Add helper_mte_check_zva
68
target/arm: Use mte_checkN for sve unpredicated loads
69
target/arm: Use mte_checkN for sve unpredicated stores
70
target/arm: Use mte_check1 for sve LD1R
71
target/arm: Tidy trans_LD1R_zpri
72
target/arm: Add arm_tlb_bti_gp
73
target/arm: Add mte helpers for sve scalar + int loads
74
target/arm: Add mte helpers for sve scalar + int stores
75
target/arm: Add mte helpers for sve scalar + int ff/nf loads
76
target/arm: Handle TBI for sve scalar + int memory ops
77
target/arm: Add mte helpers for sve scatter/gather memory ops
78
target/arm: Complete TBI clearing for user-only for SVE
79
target/arm: Implement data cache set allocation tags
80
target/arm: Set PSTATE.TCO on exception entry
81
target/arm: Always pass cacheattr to get_phys_addr
82
target/arm: Cache the Tagged bit for a page in MemTxAttrs
83
target/arm: Create tagged ram when MTE is enabled
84
target/arm: Add allocation tag storage for system mode
85
target/arm: Enable MTE
41
86
42
Markus Armbruster (1):
87
include/hw/arm/aspeed.h | 12 +-
43
ssi-sd: Make devices picking up backends unavailable with -device
88
include/hw/i2c/i2c.h | 2 +
89
include/hw/misc/pca9552.h | 16 +-
90
target/arm/cpu.h | 50 +-
91
target/arm/helper-a64.h | 16 +
92
target/arm/helper-sve.h | 488 ++++++++++++++
93
target/arm/helper.h | 2 +
94
target/arm/internals.h | 153 ++++-
95
target/arm/translate-a64.h | 5 +
96
target/arm/translate.h | 23 +-
97
hw/arm/aspeed.c | 46 +-
98
hw/arm/virt.c | 55 +-
99
hw/i2c/core.c | 18 +-
100
hw/misc/pca9552.c | 216 +++++--
101
target/arm/cpu.c | 81 ++-
102
target/arm/cpu64.c | 5 +
103
target/arm/helper-a64.c | 94 +--
104
target/arm/helper.c | 423 ++++++++++---
105
target/arm/m_helper.c | 11 +-
106
target/arm/mte_helper.c | 906 ++++++++++++++++++++++++++
107
target/arm/op_helper.c | 16 +
108
target/arm/sve_helper.c | 616 ++++++++++++++----
109
target/arm/tlb_helper.c | 13 +-
110
target/arm/translate-a64.c | 657 ++++++++++++++++---
111
target/arm/translate-sve.c | 1366 ++++++++++++++++++++++++++--------------
112
target/arm/translate-vfp.inc.c | 4 +-
113
target/arm/translate.c | 16 +-
114
hw/misc/trace-events | 4 +
115
target/arm/Makefile.objs | 1 +
116
29 files changed, 4391 insertions(+), 924 deletions(-)
117
create mode 100644 target/arm/mte_helper.c
44
118
45
Peter Maydell (10):
46
target/arm: Improve debug logging of AArch32 exception return
47
target/arm: Make switch_mode() file-local
48
target/arm: Implement HCR.FB
49
target/arm: Implement HCR.DC
50
target/arm: ISR_EL1 bits track virtual interrupts if IMO/FMO set
51
target/arm: Implement HCR.VI and VF
52
target/arm: Implement HCR.PTW
53
target/arm: New utility function to extract EC from syndrome
54
target/arm: Get IL bit correct for v7 syndrome values
55
target/arm: Report correct syndrome for FP/SIMD traps to Hyp mode
56
57
Richard Henderson (30):
58
target/arm: Move some system registers into a substructure
59
target/arm: V8M should not imply V7VE
60
target/arm: Convert v8 extensions from feature bits to isar tests
61
target/arm: Convert division from feature bits to isar0 tests
62
target/arm: Convert jazelle from feature bit to isar1 test
63
target/arm: Convert t32ee from feature bit to isar3 test
64
target/arm: Convert sve from feature bit to aa64pfr0 test
65
target/arm: Convert v8.2-fp16 from feature bit to aa64pfr0 test
66
target/arm: Hoist address increment for vector memory ops
67
target/arm: Don't call tcg_clear_temp_count
68
target/arm: Use tcg_gen_gvec_dup_i64 for LD[1-4]R
69
target/arm: Promote consecutive memory ops for aa64
70
target/arm: Mark some arrays const
71
target/arm: Use gvec for NEON VDUP
72
target/arm: Use gvec for NEON VMOV, VMVN, VBIC & VORR (immediate)
73
target/arm: Use gvec for NEON_3R_LOGIC insns
74
target/arm: Use gvec for NEON_3R_VADD_VSUB insns
75
target/arm: Use gvec for NEON_2RM_VMN, NEON_2RM_VNEG
76
target/arm: Use gvec for NEON_3R_VMUL
77
target/arm: Use gvec for VSHR, VSHL
78
target/arm: Use gvec for VSRA
79
target/arm: Use gvec for VSRI, VSLI
80
target/arm: Use gvec for NEON_3R_VML
81
target/arm: Use gvec for NEON_3R_VTST_VCEQ, NEON_3R_VCGT, NEON_3R_VCGE
82
target/arm: Use gvec for NEON VLD all lanes
83
target/arm: Reorg NEON VLD/VST all elements
84
target/arm: Promote consecutive memory ops for aa32
85
target/arm: Reorg NEON VLD/VST single element to one lane
86
target/arm: Remove writefn from TTBR0_EL3
87
target/arm: Only flush tlb if ASID changes
88
89
Stewart Hildebrand (1):
90
hw/arm/boot: Increase compliance with kernel arm64 boot protocol
91
92
target/arm/cpu.h | 227 ++++++-
93
target/arm/internals.h | 45 +-
94
target/arm/kvm_arm.h | 24 +
95
target/arm/translate.h | 21 +
96
hw/arm/boot.c | 18 +
97
hw/intc/armv7m_nvic.c | 12 +-
98
hw/net/cadence_gem.c | 9 +-
99
hw/sd/ssi-sd.c | 2 +
100
linux-user/aarch64/signal.c | 4 +-
101
linux-user/elfload.c | 60 +-
102
linux-user/syscall.c | 10 +-
103
target/arm/cpu.c | 242 ++++----
104
target/arm/cpu64.c | 148 +++--
105
target/arm/helper.c | 397 ++++++++----
106
target/arm/kvm.c | 60 ++
107
target/arm/kvm32.c | 13 +
108
target/arm/kvm64.c | 15 +-
109
target/arm/machine.c | 28 +-
110
target/arm/op_helper.c | 2 +-
111
target/arm/translate-a64.c | 715 ++++-----------------
112
target/arm/translate.c | 1451 ++++++++++++++++++++++++++++---------------
113
21 files changed, 2021 insertions(+), 1482 deletions(-)
114
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Announce the availability of the various priority queues.
3
I'm confused by this code, 'bmc' is created as:
4
This fixes an issue where guest kernels would miss to
5
configure secondary queues due to inproper feature bits.
6
4
7
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
bmc = g_new0(AspeedBoardState, 1);
8
Message-id: 20181017213932.19973-2-edgar.iglesias@gmail.com
6
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Then we use it as QOM owner for different MemoryRegion objects.
8
But looking at memory_region_init_ram (similarly for ROM):
9
10
void memory_region_init_ram(MemoryRegion *mr,
11
struct Object *owner,
12
const char *name,
13
uint64_t size,
14
Error **errp)
15
{
16
DeviceState *owner_dev;
17
Error *err = NULL;
18
19
memory_region_init_ram_nomigrate(mr, owner, name, size, &err);
20
if (err) {
21
error_propagate(errp, err);
22
return;
23
}
24
/* This will assert if owner is neither NULL nor a DeviceState.
25
* We only want the owner here for the purposes of defining a
26
* unique name for migration. TODO: Ideally we should implement
27
* a naming scheme for Objects which are not DeviceStates, in
28
* which case we can relax this restriction.
29
*/
30
owner_dev = DEVICE(owner);
31
vmstate_register_ram(mr, owner_dev);
32
}
33
34
The expected assertion is not triggered ('bmc' is not NULL neither
35
a DeviceState).
36
37
'bmc' structure is defined as:
38
39
struct AspeedBoardState {
40
AspeedSoCState soc;
41
MemoryRegion ram_container;
42
MemoryRegion max_ram;
43
};
44
45
What happens is when using 'OBJECT(bmc)', the QOM macros cast the
46
memory pointed by bmc, which first member is 'soc', which is
47
initialized ...:
48
49
object_initialize_child(OBJECT(machine), "soc",
50
&bmc->soc, amc->soc_name);
51
52
The 'soc' object is indeed a DeviceState, so the assertion passes.
53
54
Since this is fragile and only happens to work by luck, remove the
55
dangerous OBJECT(bmc) owner argument.
56
57
Note, this probably breaks migration for this machine.
58
59
Reviewed-by: Cédric Le Goater <clg@kaod.org>
60
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
61
Message-id: 20200623072132.2868-2-f4bug@amsat.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
62
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
63
---
12
hw/net/cadence_gem.c | 8 +++++++-
64
hw/arm/aspeed.c | 6 +++---
13
1 file changed, 7 insertions(+), 1 deletion(-)
65
1 file changed, 3 insertions(+), 3 deletions(-)
14
66
15
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
67
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
16
index XXXXXXX..XXXXXXX 100644
68
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/net/cadence_gem.c
69
--- a/hw/arm/aspeed.c
18
+++ b/hw/net/cadence_gem.c
70
+++ b/hw/arm/aspeed.c
19
@@ -XXX,XX +XXX,XX @@ static void gem_reset(DeviceState *d)
71
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
20
int i;
72
* needed by the flash modules of the Aspeed machines.
21
CadenceGEMState *s = CADENCE_GEM(d);
73
*/
22
const uint8_t *a;
74
if (ASPEED_MACHINE(machine)->mmio_exec) {
23
+ uint32_t queues_mask = 0;
75
- memory_region_init_alias(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
24
76
+ memory_region_init_alias(boot_rom, NULL, "aspeed.boot_rom",
25
DB_PRINT("\n");
77
&fl->mmio, 0, fl->size);
26
78
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
27
@@ -XXX,XX +XXX,XX @@ static void gem_reset(DeviceState *d)
79
boot_rom);
28
s->regs[GEM_DESCONF] = 0x02500111;
80
} else {
29
s->regs[GEM_DESCONF2] = 0x2ab13fff;
81
- memory_region_init_rom(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
30
s->regs[GEM_DESCONF5] = 0x002f2045;
82
+ memory_region_init_rom(boot_rom, NULL, "aspeed.boot_rom",
31
- s->regs[GEM_DESCONF6] = 0x00000200;
83
fl->size, &error_abort);
32
+ s->regs[GEM_DESCONF6] = 0x0;
84
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
33
+
85
boot_rom);
34
+ if (s->num_priority_queues > 1) {
86
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
35
+ queues_mask = MAKE_64BIT_MASK(1, s->num_priority_queues - 1);
87
if (machine->kernel_filename && sc->num_cpus > 1) {
36
+ s->regs[GEM_DESCONF6] |= queues_mask;
88
/* With no u-boot we must set up a boot stub for the secondary CPU */
37
+ }
89
MemoryRegion *smpboot = g_new(MemoryRegion, 1);
38
90
- memory_region_init_ram(smpboot, OBJECT(bmc), "aspeed.smpboot",
39
/* Set MAC address */
91
+ memory_region_init_ram(smpboot, NULL, "aspeed.smpboot",
40
a = &s->conf.macaddr.a[0];
92
0x80, &error_abort);
93
memory_region_add_subregion(get_system_memory(),
94
AST_SMP_MAILBOX_BASE, smpboot);
41
--
95
--
42
2.19.1
96
2.20.1
43
97
44
98
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
To have a more consistent naming, rename AspeedBoardState
4
as AspeedMachineState.
5
6
Suggested-by: Cédric Le Goater <clg@kaod.org>
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Message-id: 20200623072132.2868-3-f4bug@amsat.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
include/hw/arm/aspeed.h | 4 ++--
13
hw/arm/aspeed.c | 20 ++++++++++----------
14
2 files changed, 12 insertions(+), 12 deletions(-)
15
16
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/aspeed.h
19
+++ b/include/hw/arm/aspeed.h
20
@@ -XXX,XX +XXX,XX @@
21
22
#include "hw/boards.h"
23
24
-typedef struct AspeedBoardState AspeedBoardState;
25
+typedef struct AspeedMachineState AspeedMachineState;
26
27
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
28
#define ASPEED_MACHINE(obj) \
29
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineClass {
30
const char *spi_model;
31
uint32_t num_cs;
32
uint32_t macs_mask;
33
- void (*i2c_init)(AspeedBoardState *bmc);
34
+ void (*i2c_init)(AspeedMachineState *bmc);
35
} AspeedMachineClass;
36
37
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
.board_id = -1, /* device-tree-only board */
44
};
45
46
-struct AspeedBoardState {
47
+struct AspeedMachineState {
48
AspeedSoCState soc;
49
MemoryRegion ram_container;
50
MemoryRegion max_ram;
51
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
52
53
static void aspeed_machine_init(MachineState *machine)
54
{
55
- AspeedBoardState *bmc;
56
+ AspeedMachineState *bmc;
57
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
58
AspeedSoCClass *sc;
59
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
60
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
61
int i;
62
NICInfo *nd = &nd_table[0];
63
64
- bmc = g_new0(AspeedBoardState, 1);
65
+ bmc = g_new0(AspeedMachineState, 1);
66
67
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
68
4 * GiB);
69
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
70
arm_load_kernel(ARM_CPU(first_cpu), machine, &aspeed_board_binfo);
71
}
72
73
-static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
74
+static void palmetto_bmc_i2c_init(AspeedMachineState *bmc)
75
{
76
AspeedSoCState *soc = &bmc->soc;
77
DeviceState *dev;
78
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
79
object_property_set_int(OBJECT(dev), 110000, "temperature3", &error_abort);
80
}
81
82
-static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
83
+static void ast2500_evb_i2c_init(AspeedMachineState *bmc)
84
{
85
AspeedSoCState *soc = &bmc->soc;
86
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
87
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
88
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
89
}
90
91
-static void ast2600_evb_i2c_init(AspeedBoardState *bmc)
92
+static void ast2600_evb_i2c_init(AspeedMachineState *bmc)
93
{
94
/* Start with some devices on our I2C busses */
95
ast2500_evb_i2c_init(bmc);
96
}
97
98
-static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
99
+static void romulus_bmc_i2c_init(AspeedMachineState *bmc)
100
{
101
AspeedSoCState *soc = &bmc->soc;
102
103
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
104
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
105
}
106
107
-static void swift_bmc_i2c_init(AspeedBoardState *bmc)
108
+static void swift_bmc_i2c_init(AspeedMachineState *bmc)
109
{
110
AspeedSoCState *soc = &bmc->soc;
111
112
@@ -XXX,XX +XXX,XX @@ static void swift_bmc_i2c_init(AspeedBoardState *bmc)
113
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 12), "tmp105", 0x4a);
114
}
115
116
-static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
117
+static void sonorapass_bmc_i2c_init(AspeedMachineState *bmc)
118
{
119
AspeedSoCState *soc = &bmc->soc;
120
121
@@ -XXX,XX +XXX,XX @@ static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
122
123
}
124
125
-static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
126
+static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
127
{
128
AspeedSoCState *soc = &bmc->soc;
129
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
AspeedMachineState seems crippled. We use incorrectly 2
4
different structures to do the same thing. Merge them
5
altogether:
6
- Move AspeedMachine fields to AspeedMachineState
7
- AspeedMachineState is now QOM
8
- Remove unused AspeedMachine structure
9
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Message-id: 20200623072132.2868-4-f4bug@amsat.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
include/hw/arm/aspeed.h | 8 +-------
16
hw/arm/aspeed.c | 11 +++++++----
17
2 files changed, 8 insertions(+), 11 deletions(-)
18
19
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/include/hw/arm/aspeed.h
22
+++ b/include/hw/arm/aspeed.h
23
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineState AspeedMachineState;
24
25
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
26
#define ASPEED_MACHINE(obj) \
27
- OBJECT_CHECK(AspeedMachine, (obj), TYPE_ASPEED_MACHINE)
28
-
29
-typedef struct AspeedMachine {
30
- MachineState parent_obj;
31
-
32
- bool mmio_exec;
33
-} AspeedMachine;
34
+ OBJECT_CHECK(AspeedMachineState, (obj), TYPE_ASPEED_MACHINE)
35
36
#define ASPEED_MAC0_ON (1 << 0)
37
#define ASPEED_MAC1_ON (1 << 1)
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
};
44
45
struct AspeedMachineState {
46
+ /* Private */
47
+ MachineState parent_obj;
48
+ /* Public */
49
+
50
AspeedSoCState soc;
51
MemoryRegion ram_container;
52
MemoryRegion max_ram;
53
+ bool mmio_exec;
54
};
55
56
/* Palmetto hardware value: 0x120CE416 */
57
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
58
59
static void aspeed_machine_init(MachineState *machine)
60
{
61
- AspeedMachineState *bmc;
62
+ AspeedMachineState *bmc = ASPEED_MACHINE(machine);
63
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
64
AspeedSoCClass *sc;
65
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
66
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
67
int i;
68
NICInfo *nd = &nd_table[0];
69
70
- bmc = g_new0(AspeedMachineState, 1);
71
-
72
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
73
4 * GiB);
74
memory_region_add_subregion(&bmc->ram_container, 0, machine->ram);
75
@@ -XXX,XX +XXX,XX @@ static const TypeInfo aspeed_machine_types[] = {
76
}, {
77
.name = TYPE_ASPEED_MACHINE,
78
.parent = TYPE_MACHINE,
79
- .instance_size = sizeof(AspeedMachine),
80
+ .instance_size = sizeof(AspeedMachineState),
81
.instance_init = aspeed_machine_instance_init,
82
.class_size = sizeof(AspeedMachineClass),
83
.class_init = aspeed_machine_class_init,
84
--
85
2.20.1
86
87
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Extract i2c_try_create_slave() and i2c_realize_and_unref()
4
from i2c_create_slave().
5
We can now set properties on a I2CSlave before it is realized.
6
7
This is in line with the recent qdev/QOM changes merged
8
in commit 6675a653d2e.
9
10
Reviewed-by: Corey Minyard <cminyard@mvista.com>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Reviewed-by: Markus Armbruster <armbru@redhat.com>
14
Tested-by: Cédric Le Goater <clg@kaod.org>
15
Message-id: 20200623072723.6324-2-f4bug@amsat.org
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
---
18
include/hw/i2c/i2c.h | 2 ++
19
hw/i2c/core.c | 18 ++++++++++++++++--
20
2 files changed, 18 insertions(+), 2 deletions(-)
21
22
diff --git a/include/hw/i2c/i2c.h b/include/hw/i2c/i2c.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/include/hw/i2c/i2c.h
25
+++ b/include/hw/i2c/i2c.h
26
@@ -XXX,XX +XXX,XX @@ int i2c_send(I2CBus *bus, uint8_t data);
27
uint8_t i2c_recv(I2CBus *bus);
28
29
DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr);
30
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr);
31
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp);
32
33
/* lm832x.c */
34
void lm832x_key_event(DeviceState *dev, int key, int state);
35
diff --git a/hw/i2c/core.c b/hw/i2c/core.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/hw/i2c/core.c
38
+++ b/hw/i2c/core.c
39
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_i2c_slave = {
40
}
41
};
42
43
-DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
44
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr)
45
{
46
DeviceState *dev;
47
48
dev = qdev_new(name);
49
qdev_prop_set_uint8(dev, "address", addr);
50
- qdev_realize_and_unref(dev, &bus->qbus, &error_fatal);
51
+ return dev;
52
+}
53
+
54
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp)
55
+{
56
+ return qdev_realize_and_unref(dev, &bus->qbus, errp);
57
+}
58
+
59
+DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
60
+{
61
+ DeviceState *dev;
62
+
63
+ dev = i2c_try_create_slave(name, addr);
64
+ i2c_realize_and_unref(dev, bus, &error_fatal);
65
+
66
return dev;
67
}
68
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
The PCA9552 device does not expose LEDs, but simple pins
4
to connnect LEDs to. To be clearer with the device model,
5
rename 'nr_leds' as 'pin_count'.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-3-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 2 +-
14
hw/misc/pca9552.c | 10 +++++-----
15
2 files changed, 6 insertions(+), 6 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
22
23
uint8_t regs[PCA9552_NR_REGS];
24
uint8_t max_reg;
25
- uint8_t nr_leds;
26
+ uint8_t pin_count;
27
} PCA9552State;
28
29
#endif
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
35
{
36
int i;
37
38
- for (i = 0; i < s->nr_leds; i++) {
39
+ for (i = 0; i < s->pin_count; i++) {
40
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
41
uint8_t input_shift = (i % 8);
42
uint8_t config = pca9552_pin_get_config(s, i);
43
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
44
error_setg(errp, "%s: error reading %s", __func__, name);
45
return;
46
}
47
- if (led < 0 || led > s->nr_leds) {
48
+ if (led < 0 || led > s->pin_count) {
49
error_setg(errp, "%s invalid led %s", __func__, name);
50
return;
51
}
52
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
53
error_setg(errp, "%s: error reading %s", __func__, name);
54
return;
55
}
56
- if (led < 0 || led > s->nr_leds) {
57
+ if (led < 0 || led > s->pin_count) {
58
error_setg(errp, "%s invalid led %s", __func__, name);
59
return;
60
}
61
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
62
* PCA955X device
63
*/
64
s->max_reg = PCA9552_LS3;
65
- s->nr_leds = 16;
66
+ s->pin_count = 16;
67
68
- for (led = 0; led < s->nr_leds; led++) {
69
+ for (led = 0; led < s->pin_count; led++) {
70
char *name;
71
72
name = g_strdup_printf("led%d", led);
73
--
74
2.20.1
75
76
diff view generated by jsdifflib
New patch
1
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
3
Various code from the PCA9552 device model is generic to the
4
PCA955X family. We'll split the generic code in a base class
5
in the next commit. To ease review, first do a dumb renaming.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-4-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 10 ++---
14
hw/misc/pca9552.c | 80 +++++++++++++++++++--------------------
15
2 files changed, 45 insertions(+), 45 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@
22
#include "hw/i2c/i2c.h"
23
24
#define TYPE_PCA9552 "pca9552"
25
-#define PCA9552(obj) OBJECT_CHECK(PCA9552State, (obj), TYPE_PCA9552)
26
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
28
-#define PCA9552_NR_REGS 10
29
+#define PCA955X_NR_REGS 10
30
31
-typedef struct PCA9552State {
32
+typedef struct PCA955xState {
33
/*< private >*/
34
I2CSlave i2c;
35
/*< public >*/
36
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
37
uint8_t len;
38
uint8_t pointer;
39
40
- uint8_t regs[PCA9552_NR_REGS];
41
+ uint8_t regs[PCA955X_NR_REGS];
42
uint8_t max_reg;
43
uint8_t pin_count;
44
-} PCA9552State;
45
+} PCA955xState;
46
47
#endif
48
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/hw/misc/pca9552.c
51
+++ b/hw/misc/pca9552.c
52
@@ -XXX,XX +XXX,XX @@
53
54
static const char *led_state[] = {"on", "off", "pwm0", "pwm1"};
55
56
-static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
57
+static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
58
{
59
uint8_t reg = PCA9552_LS0 + (pin / 4);
60
uint8_t shift = (pin % 4) << 1;
61
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
62
return extract32(s->regs[reg], shift, 2);
63
}
64
65
-static void pca9552_update_pin_input(PCA9552State *s)
66
+static void pca955x_update_pin_input(PCA955xState *s)
67
{
68
int i;
69
70
for (i = 0; i < s->pin_count; i++) {
71
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
72
uint8_t input_shift = (i % 8);
73
- uint8_t config = pca9552_pin_get_config(s, i);
74
+ uint8_t config = pca955x_pin_get_config(s, i);
75
76
switch (config) {
77
case PCA9552_LED_ON:
78
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
79
}
80
}
81
82
-static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
83
+static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
84
{
85
switch (reg) {
86
case PCA9552_INPUT0:
87
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
88
}
89
}
90
91
-static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
92
+static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
93
{
94
switch (reg) {
95
case PCA9552_PSC0:
96
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
97
case PCA9552_LS2:
98
case PCA9552_LS3:
99
s->regs[reg] = data;
100
- pca9552_update_pin_input(s);
101
+ pca955x_update_pin_input(s);
102
break;
103
104
case PCA9552_INPUT0:
105
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
106
* after each byte is sent to or received by the device. The index
107
* rollovers to 0 when the maximum register address is reached.
108
*/
109
-static void pca9552_autoinc(PCA9552State *s)
110
+static void pca955x_autoinc(PCA955xState *s)
111
{
112
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
113
uint8_t reg = s->pointer & 0xf;
114
@@ -XXX,XX +XXX,XX @@ static void pca9552_autoinc(PCA9552State *s)
115
}
116
}
117
118
-static uint8_t pca9552_recv(I2CSlave *i2c)
119
+static uint8_t pca955x_recv(I2CSlave *i2c)
120
{
121
- PCA9552State *s = PCA9552(i2c);
122
+ PCA955xState *s = PCA955X(i2c);
123
uint8_t ret;
124
125
- ret = pca9552_read(s, s->pointer & 0xf);
126
+ ret = pca955x_read(s, s->pointer & 0xf);
127
128
/*
129
* From the Specs:
130
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_recv(I2CSlave *i2c)
131
__func__);
132
}
133
134
- pca9552_autoinc(s);
135
+ pca955x_autoinc(s);
136
137
return ret;
138
}
139
140
-static int pca9552_send(I2CSlave *i2c, uint8_t data)
141
+static int pca955x_send(I2CSlave *i2c, uint8_t data)
142
{
143
- PCA9552State *s = PCA9552(i2c);
144
+ PCA955xState *s = PCA955X(i2c);
145
146
/* First byte sent by is the register address */
147
if (s->len == 0) {
148
s->pointer = data;
149
s->len++;
150
} else {
151
- pca9552_write(s, s->pointer & 0xf, data);
152
+ pca955x_write(s, s->pointer & 0xf, data);
153
154
- pca9552_autoinc(s);
155
+ pca955x_autoinc(s);
156
}
157
158
return 0;
159
}
160
161
-static int pca9552_event(I2CSlave *i2c, enum i2c_event event)
162
+static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
163
{
164
- PCA9552State *s = PCA9552(i2c);
165
+ PCA955xState *s = PCA955X(i2c);
166
167
s->len = 0;
168
return 0;
169
}
170
171
-static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
172
+static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
173
void *opaque, Error **errp)
174
{
175
- PCA9552State *s = PCA9552(obj);
176
+ PCA955xState *s = PCA955X(obj);
177
int led, rc, reg;
178
uint8_t state;
179
180
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
181
* reading the INPUTx reg
182
*/
183
reg = PCA9552_LS0 + led / 4;
184
- state = (pca9552_read(s, reg) >> (led % 8)) & 0x3;
185
+ state = (pca955x_read(s, reg) >> (led % 8)) & 0x3;
186
visit_type_str(v, name, (char **)&led_state[state], errp);
187
}
188
189
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
190
((state & 0x3) << (led_num << 1));
191
}
192
193
-static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
194
+static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
195
void *opaque, Error **errp)
196
{
197
- PCA9552State *s = PCA9552(obj);
198
+ PCA955xState *s = PCA955X(obj);
199
Error *local_err = NULL;
200
int led, rc, reg, val;
201
uint8_t state;
202
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
203
}
204
205
reg = PCA9552_LS0 + led / 4;
206
- val = pca9552_read(s, reg);
207
+ val = pca955x_read(s, reg);
208
val = pca955x_ledsel(val, led % 4, state);
209
- pca9552_write(s, reg, val);
210
+ pca955x_write(s, reg, val);
211
}
212
213
static const VMStateDescription pca9552_vmstate = {
214
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription pca9552_vmstate = {
215
.version_id = 0,
216
.minimum_version_id = 0,
217
.fields = (VMStateField[]) {
218
- VMSTATE_UINT8(len, PCA9552State),
219
- VMSTATE_UINT8(pointer, PCA9552State),
220
- VMSTATE_UINT8_ARRAY(regs, PCA9552State, PCA9552_NR_REGS),
221
- VMSTATE_I2C_SLAVE(i2c, PCA9552State),
222
+ VMSTATE_UINT8(len, PCA955xState),
223
+ VMSTATE_UINT8(pointer, PCA955xState),
224
+ VMSTATE_UINT8_ARRAY(regs, PCA955xState, PCA955X_NR_REGS),
225
+ VMSTATE_I2C_SLAVE(i2c, PCA955xState),
226
VMSTATE_END_OF_LIST()
227
}
228
};
229
230
static void pca9552_reset(DeviceState *dev)
231
{
232
- PCA9552State *s = PCA9552(dev);
233
+ PCA955xState *s = PCA955X(dev);
234
235
s->regs[PCA9552_PSC0] = 0xFF;
236
s->regs[PCA9552_PWM0] = 0x80;
237
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
238
s->regs[PCA9552_LS2] = 0x55;
239
s->regs[PCA9552_LS3] = 0x55;
240
241
- pca9552_update_pin_input(s);
242
+ pca955x_update_pin_input(s);
243
244
s->pointer = 0xFF;
245
s->len = 0;
246
}
247
248
-static void pca9552_initfn(Object *obj)
249
+static void pca955x_initfn(Object *obj)
250
{
251
- PCA9552State *s = PCA9552(obj);
252
+ PCA955xState *s = PCA955X(obj);
253
int led;
254
255
/* If support for the other PCA955X devices are implemented, these
256
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
257
char *name;
258
259
name = g_strdup_printf("led%d", led);
260
- object_property_add(obj, name, "bool", pca9552_get_led, pca9552_set_led,
261
+ object_property_add(obj, name, "bool", pca955x_get_led, pca955x_set_led,
262
NULL, NULL);
263
g_free(name);
264
}
265
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
266
DeviceClass *dc = DEVICE_CLASS(klass);
267
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
268
269
- k->event = pca9552_event;
270
- k->recv = pca9552_recv;
271
- k->send = pca9552_send;
272
+ k->event = pca955x_event;
273
+ k->recv = pca955x_recv;
274
+ k->send = pca955x_send;
275
dc->reset = pca9552_reset;
276
dc->vmsd = &pca9552_vmstate;
277
}
278
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
279
static const TypeInfo pca9552_info = {
280
.name = TYPE_PCA9552,
281
.parent = TYPE_I2C_SLAVE,
282
- .instance_init = pca9552_initfn,
283
- .instance_size = sizeof(PCA9552State),
284
+ .instance_init = pca955x_initfn,
285
+ .instance_size = sizeof(PCA955xState),
286
.class_init = pca9552_class_init,
287
};
288
289
-static void pca9552_register_types(void)
290
+static void pca955x_register_types(void)
291
{
292
type_register_static(&pca9552_info);
293
}
294
295
-type_init(pca9552_register_types)
296
+type_init(pca955x_register_types)
297
--
298
2.20.1
299
300
diff view generated by jsdifflib
New patch
1
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
3
Extract the code common to the PCA955x family in PCA955xClass,
4
keeping the PCA9552 specific parts into pca9552_class_init().
5
Remove the 'TODO' comment added in commit 5141d4158cf.
6
7
Suggested-by: Cédric Le Goater <clg@kaod.org>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-5-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 6 ++--
15
hw/misc/pca9552.c | 66 ++++++++++++++++++++++++++++-----------
16
2 files changed, 51 insertions(+), 21 deletions(-)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@
23
#include "hw/i2c/i2c.h"
24
25
#define TYPE_PCA9552 "pca9552"
26
-#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
27
+#define TYPE_PCA955X "pca955x"
28
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA955X)
29
30
#define PCA955X_NR_REGS 10
31
+#define PCA955X_PIN_COUNT_MAX 16
32
33
typedef struct PCA955xState {
34
/*< private >*/
35
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
36
uint8_t pointer;
37
38
uint8_t regs[PCA955X_NR_REGS];
39
- uint8_t max_reg;
40
- uint8_t pin_count;
41
} PCA955xState;
42
43
#endif
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
* https://www.nxp.com/docs/en/application-note/AN264.pdf
50
*
51
* Copyright (c) 2017-2018, IBM Corporation.
52
+ * Copyright (c) 2020 Philippe Mathieu-Daudé
53
*
54
* This work is licensed under the terms of the GNU GPL, version 2 or
55
* later. See the COPYING file in the top-level directory.
56
@@ -XXX,XX +XXX,XX @@
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
60
+typedef struct PCA955xClass {
61
+ /*< private >*/
62
+ I2CSlaveClass parent_class;
63
+ /*< public >*/
64
+
65
+ uint8_t pin_count;
66
+ uint8_t max_reg;
67
+} PCA955xClass;
68
+
69
+#define PCA955X_CLASS(klass) \
70
+ OBJECT_CLASS_CHECK(PCA955xClass, (klass), TYPE_PCA955X)
71
+#define PCA955X_GET_CLASS(obj) \
72
+ OBJECT_GET_CLASS(PCA955xClass, (obj), TYPE_PCA955X)
73
+
74
#define PCA9552_LED_ON 0x0
75
#define PCA9552_LED_OFF 0x1
76
#define PCA9552_LED_PWM0 0x2
77
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
78
79
static void pca955x_update_pin_input(PCA955xState *s)
80
{
81
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
82
int i;
83
84
- for (i = 0; i < s->pin_count; i++) {
85
+ for (i = 0; i < k->pin_count; i++) {
86
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
87
uint8_t input_shift = (i % 8);
88
uint8_t config = pca955x_pin_get_config(s, i);
89
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
90
*/
91
static void pca955x_autoinc(PCA955xState *s)
92
{
93
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
94
+
95
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
96
uint8_t reg = s->pointer & 0xf;
97
98
- reg = (reg + 1) % (s->max_reg + 1);
99
+ reg = (reg + 1) % (k->max_reg + 1);
100
s->pointer = reg | PCA9552_AUTOINC;
101
}
102
}
103
@@ -XXX,XX +XXX,XX @@ static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
104
static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
105
void *opaque, Error **errp)
106
{
107
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
108
PCA955xState *s = PCA955X(obj);
109
int led, rc, reg;
110
uint8_t state;
111
@@ -XXX,XX +XXX,XX @@ static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
112
error_setg(errp, "%s: error reading %s", __func__, name);
113
return;
114
}
115
- if (led < 0 || led > s->pin_count) {
116
+ if (led < 0 || led > k->pin_count) {
117
error_setg(errp, "%s invalid led %s", __func__, name);
118
return;
119
}
120
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
121
static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
122
void *opaque, Error **errp)
123
{
124
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
125
PCA955xState *s = PCA955X(obj);
126
Error *local_err = NULL;
127
int led, rc, reg, val;
128
@@ -XXX,XX +XXX,XX @@ static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
129
error_setg(errp, "%s: error reading %s", __func__, name);
130
return;
131
}
132
- if (led < 0 || led > s->pin_count) {
133
+ if (led < 0 || led > k->pin_count) {
134
error_setg(errp, "%s invalid led %s", __func__, name);
135
return;
136
}
137
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
138
139
static void pca955x_initfn(Object *obj)
140
{
141
- PCA955xState *s = PCA955X(obj);
142
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
143
int led;
144
145
- /* If support for the other PCA955X devices are implemented, these
146
- * constant values might be part of class structure describing the
147
- * PCA955X device
148
- */
149
- s->max_reg = PCA9552_LS3;
150
- s->pin_count = 16;
151
-
152
- for (led = 0; led < s->pin_count; led++) {
153
+ assert(k->pin_count <= PCA955X_PIN_COUNT_MAX);
154
+ for (led = 0; led < k->pin_count; led++) {
155
char *name;
156
157
name = g_strdup_printf("led%d", led);
158
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
159
}
160
}
161
162
-static void pca9552_class_init(ObjectClass *klass, void *data)
163
+static void pca955x_class_init(ObjectClass *klass, void *data)
164
{
165
- DeviceClass *dc = DEVICE_CLASS(klass);
166
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
167
168
k->event = pca955x_event;
169
k->recv = pca955x_recv;
170
k->send = pca955x_send;
171
+}
172
+
173
+static const TypeInfo pca955x_info = {
174
+ .name = TYPE_PCA955X,
175
+ .parent = TYPE_I2C_SLAVE,
176
+ .instance_init = pca955x_initfn,
177
+ .instance_size = sizeof(PCA955xState),
178
+ .class_init = pca955x_class_init,
179
+ .abstract = true,
180
+};
181
+
182
+static void pca9552_class_init(ObjectClass *oc, void *data)
183
+{
184
+ DeviceClass *dc = DEVICE_CLASS(oc);
185
+ PCA955xClass *pc = PCA955X_CLASS(oc);
186
+
187
dc->reset = pca9552_reset;
188
dc->vmsd = &pca9552_vmstate;
189
+ pc->max_reg = PCA9552_LS3;
190
+ pc->pin_count = 16;
191
}
192
193
static const TypeInfo pca9552_info = {
194
.name = TYPE_PCA9552,
195
- .parent = TYPE_I2C_SLAVE,
196
- .instance_init = pca955x_initfn,
197
- .instance_size = sizeof(PCA955xState),
198
+ .parent = TYPE_PCA955X,
199
.class_init = pca9552_class_init,
200
};
201
202
static void pca955x_register_types(void)
203
{
204
+ type_register_static(&pca955x_info);
205
type_register_static(&pca9552_info);
206
}
207
208
--
209
2.20.1
210
211
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Instead of shifts and masks, use direct loads and stores from
3
Add a description field to distinguish between multiple devices.
4
the neon register file.
5
4
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Reviewed-by: Cédric Le Goater <clg@kaod.org>
7
Message-id: 20181011205206.3552-21-richard.henderson@linaro.org
6
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20200623072723.6324-6-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
target/arm/translate.c | 92 +++++++++++++++++++++++-------------------
11
include/hw/misc/pca9552.h | 1 +
12
1 file changed, 50 insertions(+), 42 deletions(-)
12
hw/misc/pca9552.c | 18 ++++++++++++++++++
13
2 files changed, 19 insertions(+)
13
14
14
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate.c
17
--- a/include/hw/misc/pca9552.h
17
+++ b/target/arm/translate.c
18
+++ b/include/hw/misc/pca9552.h
18
@@ -XXX,XX +XXX,XX @@ static TCGv_i32 neon_load_reg(int reg, int pass)
19
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
19
return tmp;
20
uint8_t pointer;
21
22
uint8_t regs[PCA955X_NR_REGS];
23
+ char *description; /* For debugging purpose only */
24
} PCA955xState;
25
26
#endif
27
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/misc/pca9552.c
30
+++ b/hw/misc/pca9552.c
31
@@ -XXX,XX +XXX,XX @@
32
#include "qemu/osdep.h"
33
#include "qemu/log.h"
34
#include "qemu/module.h"
35
+#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
#include "migration/vmstate.h"
39
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
40
}
20
}
41
}
21
42
22
+static void neon_load_element(TCGv_i32 var, int reg, int ele, TCGMemOp mop)
43
+static void pca955x_realize(DeviceState *dev, Error **errp)
23
+{
44
+{
24
+ long offset = neon_element_offset(reg, ele, mop & MO_SIZE);
45
+ PCA955xState *s = PCA955X(dev);
25
+
46
+
26
+ switch (mop) {
47
+ if (!s->description) {
27
+ case MO_UB:
48
+ s->description = g_strdup("pca-unspecified");
28
+ tcg_gen_ld8u_i32(var, cpu_env, offset);
29
+ break;
30
+ case MO_UW:
31
+ tcg_gen_ld16u_i32(var, cpu_env, offset);
32
+ break;
33
+ case MO_UL:
34
+ tcg_gen_ld_i32(var, cpu_env, offset);
35
+ break;
36
+ default:
37
+ g_assert_not_reached();
38
+ }
49
+ }
39
+}
50
+}
40
+
51
+
41
static void neon_load_element64(TCGv_i64 var, int reg, int ele, TCGMemOp mop)
52
+static Property pca955x_properties[] = {
53
+ DEFINE_PROP_STRING("description", PCA955xState, description),
54
+ DEFINE_PROP_END_OF_LIST(),
55
+};
56
+
57
static void pca955x_class_init(ObjectClass *klass, void *data)
42
{
58
{
43
long offset = neon_element_offset(reg, ele, mop & MO_SIZE);
59
+ DeviceClass *dc = DEVICE_CLASS(klass);
44
@@ -XXX,XX +XXX,XX @@ static void neon_store_reg(int reg, int pass, TCGv_i32 var)
60
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
45
tcg_temp_free_i32(var);
61
62
k->event = pca955x_event;
63
k->recv = pca955x_recv;
64
k->send = pca955x_send;
65
+ dc->realize = pca955x_realize;
66
+ device_class_set_props(dc, pca955x_properties);
46
}
67
}
47
68
48
+static void neon_store_element(int reg, int ele, TCGMemOp size, TCGv_i32 var)
69
static const TypeInfo pca955x_info = {
49
+{
50
+ long offset = neon_element_offset(reg, ele, size);
51
+
52
+ switch (size) {
53
+ case MO_8:
54
+ tcg_gen_st8_i32(var, cpu_env, offset);
55
+ break;
56
+ case MO_16:
57
+ tcg_gen_st16_i32(var, cpu_env, offset);
58
+ break;
59
+ case MO_32:
60
+ tcg_gen_st_i32(var, cpu_env, offset);
61
+ break;
62
+ default:
63
+ g_assert_not_reached();
64
+ }
65
+}
66
+
67
static void neon_store_element64(int reg, int ele, TCGMemOp size, TCGv_i64 var)
68
{
69
long offset = neon_element_offset(reg, ele, size);
70
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
71
int stride;
72
int size;
73
int reg;
74
- int pass;
75
int load;
76
- int shift;
77
int n;
78
int vec_size;
79
int mmu_idx;
80
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
81
} else {
82
/* Single element. */
83
int idx = (insn >> 4) & 0xf;
84
- pass = (insn >> 7) & 1;
85
+ int reg_idx;
86
switch (size) {
87
case 0:
88
- shift = ((insn >> 5) & 3) * 8;
89
+ reg_idx = (insn >> 5) & 7;
90
stride = 1;
91
break;
92
case 1:
93
- shift = ((insn >> 6) & 1) * 16;
94
+ reg_idx = (insn >> 6) & 3;
95
stride = (insn & (1 << 5)) ? 2 : 1;
96
break;
97
case 2:
98
- shift = 0;
99
+ reg_idx = (insn >> 7) & 1;
100
stride = (insn & (1 << 6)) ? 2 : 1;
101
break;
102
default:
103
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
104
*/
105
return 1;
106
}
107
+ tmp = tcg_temp_new_i32();
108
addr = tcg_temp_new_i32();
109
load_reg_var(s, addr, rn);
110
for (reg = 0; reg < nregs; reg++) {
111
if (load) {
112
- tmp = tcg_temp_new_i32();
113
- switch (size) {
114
- case 0:
115
- gen_aa32_ld8u(s, tmp, addr, get_mem_index(s));
116
- break;
117
- case 1:
118
- gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
119
- break;
120
- case 2:
121
- gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
122
- break;
123
- default: /* Avoid compiler warnings. */
124
- abort();
125
- }
126
- if (size != 2) {
127
- tmp2 = neon_load_reg(rd, pass);
128
- tcg_gen_deposit_i32(tmp, tmp2, tmp,
129
- shift, size ? 16 : 8);
130
- tcg_temp_free_i32(tmp2);
131
- }
132
- neon_store_reg(rd, pass, tmp);
133
+ gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
134
+ s->be_data | size);
135
+ neon_store_element(rd, reg_idx, size, tmp);
136
} else { /* Store */
137
- tmp = neon_load_reg(rd, pass);
138
- if (shift)
139
- tcg_gen_shri_i32(tmp, tmp, shift);
140
- switch (size) {
141
- case 0:
142
- gen_aa32_st8(s, tmp, addr, get_mem_index(s));
143
- break;
144
- case 1:
145
- gen_aa32_st16(s, tmp, addr, get_mem_index(s));
146
- break;
147
- case 2:
148
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
149
- break;
150
- }
151
- tcg_temp_free_i32(tmp);
152
+ neon_load_element(tmp, rd, reg_idx, size);
153
+ gen_aa32_st_i32(s, tmp, addr, get_mem_index(s),
154
+ s->be_data | size);
155
}
156
rd += stride;
157
tcg_gen_addi_i32(addr, addr, 1 << size);
158
}
159
tcg_temp_free_i32(addr);
160
+ tcg_temp_free_i32(tmp);
161
stride = nregs * (1 << size);
162
}
163
}
164
--
70
--
165
2.19.1
71
2.20.1
166
72
167
73
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Add a trivial representation of the PCA9552 GPIOs.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_status
8
1592689902.327837:pca955x_gpio_status pca-unspecified GPIOs 0-15 [*...............]
9
1592689902.329934:pca955x_gpio_status pca-unspecified GPIOs 0-15 [**..............]
10
1592689902.330717:pca955x_gpio_status pca-unspecified GPIOs 0-15 [***.............]
11
1592689902.331431:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****............]
12
1592689902.332163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*..]
13
1592689902.332888:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........**.]
14
1592689902.333629:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
15
1592690032.793289:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
16
1592690033.303163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
17
1592690033.812962:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
18
1592690034.323234:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
19
1592690034.832922:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
20
21
We notice the GPIO #14 (front-power LED) starts to blink.
22
23
This LED is described in the witherspoon device-tree [*]:
24
25
front-power {
26
retain-state-shutdown;
27
default-state = "keep";
28
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
29
};
30
31
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
32
33
Suggested-by: Cédric Le Goater <clg@kaod.org>
34
Reviewed-by: Cédric Le Goater <clg@kaod.org>
35
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
36
Tested-by: Cédric Le Goater <clg@kaod.org>
37
Message-id: 20200623072723.6324-7-f4bug@amsat.org
38
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
39
---
40
hw/misc/pca9552.c | 39 +++++++++++++++++++++++++++++++++++++++
41
hw/misc/trace-events | 3 +++
42
2 files changed, 42 insertions(+)
43
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
#include "qemu/osdep.h"
50
#include "qemu/log.h"
51
#include "qemu/module.h"
52
+#include "qemu/bitops.h"
53
#include "hw/qdev-properties.h"
54
#include "hw/misc/pca9552.h"
55
#include "hw/misc/pca9552_regs.h"
56
#include "migration/vmstate.h"
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
+#include "trace.h"
60
61
typedef struct PCA955xClass {
62
/*< private >*/
63
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
64
return extract32(s->regs[reg], shift, 2);
65
}
66
67
+/* Return INPUT status (bit #N belongs to GPIO #N) */
68
+static uint16_t pca955x_pins_get_status(PCA955xState *s)
69
+{
70
+ return (s->regs[PCA9552_INPUT1] << 8) | s->regs[PCA9552_INPUT0];
71
+}
72
+
73
+static void pca955x_display_pins_status(PCA955xState *s,
74
+ uint16_t previous_pins_status)
75
+{
76
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
77
+ uint16_t pins_status, pins_changed;
78
+ int i;
79
+
80
+ pins_status = pca955x_pins_get_status(s);
81
+ pins_changed = previous_pins_status ^ pins_status;
82
+ if (!pins_changed) {
83
+ return;
84
+ }
85
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_STATUS)) {
86
+ char *buf = g_newa(char, k->pin_count + 1);
87
+
88
+ for (i = 0; i < k->pin_count; i++) {
89
+ if (extract32(pins_status, i, 1)) {
90
+ buf[i] = '*';
91
+ } else {
92
+ buf[i] = '.';
93
+ }
94
+ }
95
+ buf[i] = '\0';
96
+ trace_pca955x_gpio_status(s->description, buf);
97
+ }
98
+}
99
+
100
static void pca955x_update_pin_input(PCA955xState *s)
101
{
102
PCA955xClass *k = PCA955X_GET_CLASS(s);
103
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
104
105
static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
106
{
107
+ uint16_t pins_status;
108
+
109
switch (reg) {
110
case PCA9552_PSC0:
111
case PCA9552_PWM0:
112
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
113
case PCA9552_LS1:
114
case PCA9552_LS2:
115
case PCA9552_LS3:
116
+ pins_status = pca955x_pins_get_status(s);
117
s->regs[reg] = data;
118
pca955x_update_pin_input(s);
119
+ pca955x_display_pins_status(s, pins_status);
120
break;
121
122
case PCA9552_INPUT0:
123
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
124
index XXXXXXX..XXXXXXX 100644
125
--- a/hw/misc/trace-events
126
+++ b/hw/misc/trace-events
127
@@ -XXX,XX +XXX,XX @@ via1_adb_poll(uint8_t data, const char *vadbint, int status, int index, int size
128
# grlib_ahb_apb_pnp.c
129
grlib_ahb_pnp_read(uint64_t addr, uint32_t value) "AHB PnP read addr:0x%03"PRIx64" data:0x%08x"
130
grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx64" data:0x%08x"
131
+
132
+# pca9552.c
133
+pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
134
--
135
2.20.1
136
137
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
We have 2 distinct PCA9552 devices. Set their description
4
to distinguish them when looking at the trace events.
5
6
Description name taken from:
7
https://github.com/open-power/witherspoon-xml/blob/master/witherspoon.xml
8
9
Reviewed-by: Cédric Le Goater <clg@kaod.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Corey Minyard <cminyard@mvista.com>
12
Reviewed-by: Markus Armbruster <armbru@redhat.com>
13
Tested-by: Cédric Le Goater <clg@kaod.org>
14
Message-id: 20200623072723.6324-8-f4bug@amsat.org
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
---
17
hw/arm/aspeed.c | 13 +++++++++----
18
1 file changed, 9 insertions(+), 4 deletions(-)
19
20
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/hw/arm/aspeed.c
23
+++ b/hw/arm/aspeed.c
24
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
25
{
26
AspeedSoCState *soc = &bmc->soc;
27
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
28
+ DeviceState *dev;
29
30
/* Bus 3: TODO bmp280@77 */
31
/* Bus 3: TODO max31785@52 */
32
/* Bus 3: TODO dps310@76 */
33
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
34
- 0x60);
35
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
36
+ qdev_prop_set_string(dev, "description", "pca1");
37
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3),
38
+ &error_fatal);
39
40
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
41
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
42
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
43
44
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
45
eeprom_buf);
46
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
47
- 0x60);
48
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
49
+ qdev_prop_set_string(dev, "description", "pca0");
50
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11),
51
+ &error_fatal);
52
/* Bus 11: TODO ucd90160@64 */
53
}
54
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
New patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
2
3
Emit a trace event when a GPIO change its state.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_change
8
1592690552.687372:pca955x_gpio_change pca1 GPIO id:0 status: 0 -> 1
9
1592690552.690169:pca955x_gpio_change pca1 GPIO id:1 status: 0 -> 1
10
1592690552.691673:pca955x_gpio_change pca1 GPIO id:2 status: 0 -> 1
11
1592690552.696886:pca955x_gpio_change pca1 GPIO id:3 status: 0 -> 1
12
1592690552.698614:pca955x_gpio_change pca1 GPIO id:13 status: 0 -> 1
13
1592690552.699833:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
14
1592690552.700842:pca955x_gpio_change pca1 GPIO id:15 status: 0 -> 1
15
1592690683.841921:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
16
1592690683.861660:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
17
1592690684.371460:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
18
1592690684.882115:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
19
1592690685.391411:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
20
1592690685.901391:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
21
1592690686.411678:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
22
1592690686.921279:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
23
24
We notice the GPIO #14 (front-power LED) starts to blink.
25
26
This LED is described in the witherspoon device-tree [*]:
27
28
front-power {
29
retain-state-shutdown;
30
default-state = "keep";
31
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
32
};
33
34
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
35
36
Reviewed-by: Cédric Le Goater <clg@kaod.org>
37
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
38
Tested-by: Cédric Le Goater <clg@kaod.org>
39
Message-id: 20200623072723.6324-9-f4bug@amsat.org
40
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
41
---
42
hw/misc/pca9552.c | 15 +++++++++++++++
43
hw/misc/trace-events | 1 +
44
2 files changed, 16 insertions(+)
45
46
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/hw/misc/pca9552.c
49
+++ b/hw/misc/pca9552.c
50
@@ -XXX,XX +XXX,XX @@ static void pca955x_display_pins_status(PCA955xState *s,
51
buf[i] = '\0';
52
trace_pca955x_gpio_status(s->description, buf);
53
}
54
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_CHANGE)) {
55
+ for (i = 0; i < k->pin_count; i++) {
56
+ if (extract32(pins_changed, i, 1)) {
57
+ unsigned new_state = extract32(pins_status, i, 1);
58
+
59
+ /*
60
+ * We display the state using the PCA logic ("active-high").
61
+ * This is not the state of the LED, which signal might be
62
+ * wired "active-low" on the board.
63
+ */
64
+ trace_pca955x_gpio_change(s->description, i,
65
+ !new_state, new_state);
66
+ }
67
+ }
68
+ }
69
}
70
71
static void pca955x_update_pin_input(PCA955xState *s)
72
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/misc/trace-events
75
+++ b/hw/misc/trace-events
76
@@ -XXX,XX +XXX,XX @@ grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx6
77
78
# pca9552.c
79
pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
80
+pca955x_gpio_change(const char *description, unsigned id, unsigned prev_state, unsigned current_state) "%s GPIO id:%u status: %u -> %u"
81
--
82
2.20.1
83
84
diff view generated by jsdifflib
1
From: "Edgar E. Iglesias" <edgar.iglesias@xilinx.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Announce 64bit addressing support.
3
The PCA9552 has 16 GPIOs which can be used as input,
4
output or PWM mode. QEMU models the output GPIO with
5
the qemu_irq type. Let the device expose the 16 GPIOs
6
to allow us to later connect LEDs to these outputs.
4
7
5
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Message-id: 20181017213932.19973-3-edgar.iglesias@gmail.com
10
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Message-id: 20200623072723.6324-10-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
13
---
11
hw/net/cadence_gem.c | 3 ++-
14
include/hw/misc/pca9552.h | 1 +
12
1 file changed, 2 insertions(+), 1 deletion(-)
15
hw/misc/pca9552.c | 6 ++++++
16
2 files changed, 7 insertions(+)
13
17
14
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
15
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
16
--- a/hw/net/cadence_gem.c
20
--- a/include/hw/misc/pca9552.h
17
+++ b/hw/net/cadence_gem.c
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
23
uint8_t pointer;
24
25
uint8_t regs[PCA955X_NR_REGS];
26
+ qemu_irq gpio[PCA955X_PIN_COUNT_MAX];
27
char *description; /* For debugging purpose only */
28
} PCA955xState;
29
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
18
@@ -XXX,XX +XXX,XX @@
34
@@ -XXX,XX +XXX,XX @@
19
#define GEM_DESCONF4 (0x0000028C/4)
35
#include "hw/qdev-properties.h"
20
#define GEM_DESCONF5 (0x00000290/4)
36
#include "hw/misc/pca9552.h"
21
#define GEM_DESCONF6 (0x00000294/4)
37
#include "hw/misc/pca9552_regs.h"
22
+#define GEM_DESCONF6_64B_MASK (1U << 23)
38
+#include "hw/irq.h"
23
#define GEM_DESCONF7 (0x00000298/4)
39
#include "migration/vmstate.h"
24
40
#include "qapi/error.h"
25
#define GEM_INT_Q1_STATUS (0x00000400 / 4)
41
#include "qapi/visitor.h"
26
@@ -XXX,XX +XXX,XX @@ static void gem_reset(DeviceState *d)
42
@@ -XXX,XX +XXX,XX @@ static void pca955x_update_pin_input(PCA955xState *s)
27
s->regs[GEM_DESCONF] = 0x02500111;
43
28
s->regs[GEM_DESCONF2] = 0x2ab13fff;
44
switch (config) {
29
s->regs[GEM_DESCONF5] = 0x002f2045;
45
case PCA9552_LED_ON:
30
- s->regs[GEM_DESCONF6] = 0x0;
46
+ qemu_set_irq(s->gpio[i], 1);
31
+ s->regs[GEM_DESCONF6] = GEM_DESCONF6_64B_MASK;
47
s->regs[input_reg] |= 1 << input_shift;
32
48
break;
33
if (s->num_priority_queues > 1) {
49
case PCA9552_LED_OFF:
34
queues_mask = MAKE_64BIT_MASK(1, s->num_priority_queues - 1);
50
+ qemu_set_irq(s->gpio[i], 0);
51
s->regs[input_reg] &= ~(1 << input_shift);
52
break;
53
case PCA9552_LED_PWM0:
54
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
55
56
static void pca955x_realize(DeviceState *dev, Error **errp)
57
{
58
+ PCA955xClass *k = PCA955X_GET_CLASS(dev);
59
PCA955xState *s = PCA955X(dev);
60
61
if (!s->description) {
62
s->description = g_strdup("pca-unspecified");
63
}
64
+
65
+ qdev_init_gpio_out(dev, s->gpio, k->pin_count);
66
}
67
68
static Property pca955x_properties[] = {
35
--
69
--
36
2.19.1
70
2.20.1
37
71
38
72
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Having V6 alone imply jazelle was wrong for cortex-m0.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Change to an assertion for V6 & !M.
5
6
This was harmless, because the only place we tested ARM_FEATURE_JAZELLE
7
was for 'bxj' in disas_arm(), which is unreachable for M-profile cores.
8
9
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20181016223115.24100-6-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-2-richard.henderson@linaro.org
12
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
7
---
15
target/arm/cpu.h | 6 +++++-
8
target/arm/cpu.h | 10 ++++++++++
16
target/arm/cpu.c | 17 ++++++++++++++---
9
1 file changed, 10 insertions(+)
17
target/arm/translate.c | 2 +-
18
3 files changed, 20 insertions(+), 5 deletions(-)
19
10
20
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/cpu.h
13
--- a/target/arm/cpu.h
23
+++ b/target/arm/cpu.h
14
+++ b/target/arm/cpu.h
24
@@ -XXX,XX +XXX,XX @@ enum arm_features {
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_bti(const ARMISARegisters *id)
25
ARM_FEATURE_PMU, /* has PMU support */
16
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, BT) != 0;
26
ARM_FEATURE_VBAR, /* has cp15 VBAR */
27
ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
28
- ARM_FEATURE_JAZELLE, /* has (trivial) Jazelle implementation */
29
ARM_FEATURE_SVE, /* has Scalable Vector Extension */
30
ARM_FEATURE_V8_FP16, /* implements v8.2 half-precision float */
31
ARM_FEATURE_M_MAIN, /* M profile Main Extension */
32
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_arm_div(const ARMISARegisters *id)
33
return FIELD_EX32(id->id_isar0, ID_ISAR0, DIVIDE) > 1;
34
}
17
}
35
18
36
+static inline bool isar_feature_jazelle(const ARMISARegisters *id)
19
+static inline bool isar_feature_aa64_mte_insn_reg(const ARMISARegisters *id)
37
+{
20
+{
38
+ return FIELD_EX32(id->id_isar1, ID_ISAR1, JAZELLE) != 0;
21
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) != 0;
39
+}
22
+}
40
+
23
+
41
static inline bool isar_feature_aa32_aes(const ARMISARegisters *id)
24
+static inline bool isar_feature_aa64_mte(const ARMISARegisters *id)
25
+{
26
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) >= 2;
27
+}
28
+
29
static inline bool isar_feature_aa64_pmu_8_1(const ARMISARegisters *id)
42
{
30
{
43
return FIELD_EX32(id->id_isar5, ID_ISAR5, AES) != 0;
31
return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, PMUVER) >= 4 &&
44
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/target/arm/cpu.c
47
+++ b/target/arm/cpu.c
48
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
49
}
50
if (arm_feature(env, ARM_FEATURE_V6)) {
51
set_feature(env, ARM_FEATURE_V5);
52
- set_feature(env, ARM_FEATURE_JAZELLE);
53
if (!arm_feature(env, ARM_FEATURE_M)) {
54
+ assert(cpu_isar_feature(jazelle, cpu));
55
set_feature(env, ARM_FEATURE_AUXCR);
56
}
57
}
58
@@ -XXX,XX +XXX,XX @@ static void arm926_initfn(Object *obj)
59
set_feature(&cpu->env, ARM_FEATURE_VFP);
60
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
61
set_feature(&cpu->env, ARM_FEATURE_CACHE_TEST_CLEAN);
62
- set_feature(&cpu->env, ARM_FEATURE_JAZELLE);
63
cpu->midr = 0x41069265;
64
cpu->reset_fpsid = 0x41011090;
65
cpu->ctr = 0x1dd20d2;
66
cpu->reset_sctlr = 0x00090078;
67
+
68
+ /*
69
+ * ARMv5 does not have the ID_ISAR registers, but we can still
70
+ * set the field to indicate Jazelle support within QEMU.
71
+ */
72
+ cpu->isar.id_isar1 = FIELD_DP32(cpu->isar.id_isar1, ID_ISAR1, JAZELLE, 1);
73
}
74
75
static void arm946_initfn(Object *obj)
76
@@ -XXX,XX +XXX,XX @@ static void arm1026_initfn(Object *obj)
77
set_feature(&cpu->env, ARM_FEATURE_AUXCR);
78
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
79
set_feature(&cpu->env, ARM_FEATURE_CACHE_TEST_CLEAN);
80
- set_feature(&cpu->env, ARM_FEATURE_JAZELLE);
81
cpu->midr = 0x4106a262;
82
cpu->reset_fpsid = 0x410110a0;
83
cpu->ctr = 0x1dd20d2;
84
cpu->reset_sctlr = 0x00090078;
85
cpu->reset_auxcr = 1;
86
+
87
+ /*
88
+ * ARMv5 does not have the ID_ISAR registers, but we can still
89
+ * set the field to indicate Jazelle support within QEMU.
90
+ */
91
+ cpu->isar.id_isar1 = FIELD_DP32(cpu->isar.id_isar1, ID_ISAR1, JAZELLE, 1);
92
+
93
{
94
/* The 1026 had an IFAR at c6,c0,0,1 rather than the ARMv6 c6,c0,0,2 */
95
ARMCPRegInfo ifar = {
96
diff --git a/target/arm/translate.c b/target/arm/translate.c
97
index XXXXXXX..XXXXXXX 100644
98
--- a/target/arm/translate.c
99
+++ b/target/arm/translate.c
100
@@ -XXX,XX +XXX,XX @@
101
#define ENABLE_ARCH_5 arm_dc_feature(s, ARM_FEATURE_V5)
102
/* currently all emulated v5 cores are also v5TE, so don't bother */
103
#define ENABLE_ARCH_5TE arm_dc_feature(s, ARM_FEATURE_V5)
104
-#define ENABLE_ARCH_5J arm_dc_feature(s, ARM_FEATURE_JAZELLE)
105
+#define ENABLE_ARCH_5J dc_isar_feature(jazelle, s)
106
#define ENABLE_ARCH_6 arm_dc_feature(s, ARM_FEATURE_V6)
107
#define ENABLE_ARCH_6K arm_dc_feature(s, ARM_FEATURE_V6K)
108
#define ENABLE_ARCH_6T2 arm_dc_feature(s, ARM_FEATURE_THUMB2)
109
--
32
--
110
2.19.1
33
2.20.1
111
34
112
35
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Since QEMU does not implement ASIDs, changes to the ASID must flush the
3
Protect reads of aa64 id registers with ARM_CP_STATE_AA64.
4
tlb. However, if the ASID does not change there is no reason to flush.
4
Use this as a simpler test than arm_el_is_aa64, since EL3
5
cannot change mode.
5
6
6
In testing a boot of the Ubuntu installer to the first menu, this reduces
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
the number of flushes by 30%, or nearly 600k instances.
8
9
Reviewed-by: Aaron Lindsay <aaron@os.amperecomputing.com>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Message-id: 20200626033144.790098-3-richard.henderson@linaro.org
12
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
13
Message-id: 20181019015617.22583-3-richard.henderson@linaro.org
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
---
11
---
16
target/arm/helper.c | 8 +++-----
12
target/arm/helper.c | 15 ++++++++-------
17
1 file changed, 3 insertions(+), 5 deletions(-)
13
1 file changed, 8 insertions(+), 7 deletions(-)
18
14
19
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
20
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
22
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
23
@@ -XXX,XX +XXX,XX @@ static void vmsa_tcr_el1_write(CPUARMState *env, const ARMCPRegInfo *ri,
19
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
24
static void vmsa_ttbr_write(CPUARMState *env, const ARMCPRegInfo *ri,
20
uint32_t valid_mask = 0x3fff;
25
uint64_t value)
21
ARMCPU *cpu = env_archcpu(env);
26
{
22
27
- /* 64 bit accesses to the TTBRs can change the ASID and so we
23
- if (arm_el_is_aa64(env, 3)) {
28
- * must flush the TLB.
24
+ if (ri->state == ARM_CP_STATE_AA64) {
29
- */
25
value |= SCR_FW | SCR_AW; /* these two bits are RES1. */
30
- if (cpreg_field_is_64bit(ri)) {
26
valid_mask &= ~SCR_NET;
31
+ /* If the ASID changes (with a 64-bit write), we must flush the TLB. */
27
+
32
+ if (cpreg_field_is_64bit(ri) &&
28
+ if (cpu_isar_feature(aa64_lor, cpu)) {
33
+ extract64(raw_read(env, ri) ^ value, 48, 16) != 0) {
29
+ valid_mask |= SCR_TLOR;
34
ARMCPU *cpu = arm_env_get_cpu(env);
30
+ }
35
-
31
+ if (cpu_isar_feature(aa64_pauth, cpu)) {
36
tlb_flush(CPU(cpu));
32
+ valid_mask |= SCR_API | SCR_APK;
33
+ }
34
} else {
35
valid_mask &= ~(SCR_RW | SCR_ST);
37
}
36
}
38
raw_write(env, ri, value);
37
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
38
valid_mask &= ~SCR_SMD;
39
}
40
}
41
- if (cpu_isar_feature(aa64_lor, cpu)) {
42
- valid_mask |= SCR_TLOR;
43
- }
44
- if (cpu_isar_feature(aa64_pauth, cpu)) {
45
- valid_mask |= SCR_API | SCR_APK;
46
- }
47
48
/* Clear all-context RES0 bits. */
49
value &= valid_mask;
39
--
50
--
40
2.19.1
51
2.20.1
41
52
42
53
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
The EL3 version of this register does not include an ASID,
3
This does not attempt to rectify all of the res0 bits, but does
4
and so the tlb_flush performed by vmsa_ttbr_write is not needed.
4
clear the mte bits when not enabled. Since there is no high-part
5
mapping of SCTLR, aa32 mode cannot write to these bits.
5
6
6
Reviewed-by: Aaron Lindsay <aaron@os.amperecomputing.com>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Message-id: 20200626033144.790098-4-richard.henderson@linaro.org
9
Message-id: 20181019015617.22583-2-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
11
---
12
target/arm/helper.c | 2 +-
12
target/arm/helper.c | 23 +++++++++++++++++------
13
1 file changed, 1 insertion(+), 1 deletion(-)
13
1 file changed, 17 insertions(+), 6 deletions(-)
14
14
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
18
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el3_cp_reginfo[] = {
19
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
20
.fieldoffset = offsetof(CPUARMState, cp15.mvbar) },
20
{
21
{ .name = "TTBR0_EL3", .state = ARM_CP_STATE_AA64,
21
ARMCPU *cpu = env_archcpu(env);
22
.opc0 = 3, .opc1 = 6, .crn = 2, .crm = 0, .opc2 = 0,
22
23
- .access = PL3_RW, .writefn = vmsa_ttbr_write, .resetvalue = 0,
23
+ if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
24
+ .access = PL3_RW, .resetvalue = 0,
24
+ /* M bit is RAZ/WI for PMSA with no MPU implemented */
25
.fieldoffset = offsetof(CPUARMState, cp15.ttbr0_el[3]) },
25
+ value &= ~SCTLR_M;
26
{ .name = "TCR_EL3", .state = ARM_CP_STATE_AA64,
26
+ }
27
.opc0 = 3, .opc1 = 6, .crn = 2, .crm = 0, .opc2 = 2,
27
+
28
+ /* ??? Lots of these bits are not implemented. */
29
+
30
+ if (ri->state == ARM_CP_STATE_AA64 && !cpu_isar_feature(aa64_mte, cpu)) {
31
+ if (ri->opc1 == 6) { /* SCTLR_EL3 */
32
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF | SCTLR_ATA);
33
+ } else {
34
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF0 | SCTLR_TCF |
35
+ SCTLR_ATA0 | SCTLR_ATA);
36
+ }
37
+ }
38
+
39
if (raw_read(env, ri) == value) {
40
/* Skip the TLB flush if nothing actually changed; Linux likes
41
* to do a lot of pointless SCTLR writes.
42
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
43
return;
44
}
45
46
- if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
47
- /* M bit is RAZ/WI for PMSA with no MPU implemented */
48
- value &= ~SCTLR_M;
49
- }
50
-
51
raw_write(env, ri, value);
52
- /* ??? Lots of these bits are not implemented. */
53
+
54
/* This may enable/disable the MMU, so do a TLB flush. */
55
tlb_flush(CPU(cpu));
56
28
--
57
--
29
2.19.1
58
2.20.1
30
59
31
60
diff view generated by jsdifflib
1
The HCR_EL2 VI and VF bits are supposed to track whether there is
1
From: Richard Henderson <richard.henderson@linaro.org>
2
a pending virtual IRQ or virtual FIQ. For QEMU we store the
3
pending VIRQ/VFIQ status in cs->interrupt_request, so this means:
4
* if the register is read we must get these bit values from
5
cs->interrupt_request
6
* if the register is written then we must write the bit
7
values back into cs->interrupt_request
8
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-5-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20181012144235.19646-7-peter.maydell@linaro.org
12
---
7
---
13
target/arm/helper.c | 47 +++++++++++++++++++++++++++++++++++++++++----
8
target/arm/helper.c | 14 +++++++++++---
14
1 file changed, 43 insertions(+), 4 deletions(-)
9
1 file changed, 11 insertions(+), 3 deletions(-)
15
10
16
diff --git a/target/arm/helper.c b/target/arm/helper.c
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
17
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper.c
13
--- a/target/arm/helper.c
19
+++ b/target/arm/helper.c
14
+++ b/target/arm/helper.c
20
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el3_no_el2_v8_cp_reginfo[] = {
15
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
21
static void hcr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
16
if (cpu_isar_feature(aa64_pauth, cpu)) {
22
{
17
valid_mask |= SCR_API | SCR_APK;
23
ARMCPU *cpu = arm_env_get_cpu(env);
18
}
24
+ CPUState *cs = ENV_GET_CPU(env);
19
+ if (cpu_isar_feature(aa64_mte, cpu)) {
25
uint64_t valid_mask = HCR_MASK;
20
+ valid_mask |= SCR_ATA;
26
21
+ }
27
if (arm_feature(env, ARM_FEATURE_EL3)) {
22
} else {
28
@@ -XXX,XX +XXX,XX @@ static void hcr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
23
valid_mask &= ~(SCR_RW | SCR_ST);
24
}
25
@@ -XXX,XX +XXX,XX @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
26
if (cpu_isar_feature(aa64_pauth, cpu)) {
27
valid_mask |= HCR_API | HCR_APK;
28
}
29
+ if (cpu_isar_feature(aa64_mte, cpu)) {
30
+ valid_mask |= HCR_ATA | HCR_DCT | HCR_TID5;
31
+ }
32
}
33
29
/* Clear RES0 bits. */
34
/* Clear RES0 bits. */
30
value &= valid_mask;
35
value &= valid_mask;
31
36
37
- /* These bits change the MMU setup:
32
+ /*
38
+ /*
33
+ * VI and VF are kept in cs->interrupt_request. Modifying that
39
+ * These bits change the MMU setup:
34
+ * requires that we have the iothread lock, which is done by
35
+ * marking the reginfo structs as ARM_CP_IO.
36
+ * Note that if a write to HCR pends a VIRQ or VFIQ it is never
37
+ * possible for it to be taken immediately, because VIRQ and
38
+ * VFIQ are masked unless running at EL0 or EL1, and HCR
39
+ * can only be written at EL2.
40
+ */
41
+ g_assert(qemu_mutex_iothread_locked());
42
+ if (value & HCR_VI) {
43
+ cs->interrupt_request |= CPU_INTERRUPT_VIRQ;
44
+ } else {
45
+ cs->interrupt_request &= ~CPU_INTERRUPT_VIRQ;
46
+ }
47
+ if (value & HCR_VF) {
48
+ cs->interrupt_request |= CPU_INTERRUPT_VFIQ;
49
+ } else {
50
+ cs->interrupt_request &= ~CPU_INTERRUPT_VFIQ;
51
+ }
52
+ value &= ~(HCR_VI | HCR_VF);
53
+
54
/* These bits change the MMU setup:
55
* HCR_VM enables stage 2 translation
40
* HCR_VM enables stage 2 translation
56
* HCR_PTW forbids certain page-table setups
41
* HCR_PTW forbids certain page-table setups
57
@@ -XXX,XX +XXX,XX @@ static void hcr_writelow(CPUARMState *env, const ARMCPRegInfo *ri,
42
- * HCR_DC Disables stage1 and enables stage2 translation
58
hcr_write(env, NULL, value);
43
+ * HCR_DC disables stage1 and enables stage2 translation
59
}
44
+ * HCR_DCT enables tagging on (disabled) stage1 translation
60
45
*/
61
+static uint64_t hcr_read(CPUARMState *env, const ARMCPRegInfo *ri)
46
- if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC)) {
62
+{
47
+ if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC | HCR_DCT)) {
63
+ /* The VI and VF bits live in cs->interrupt_request */
48
tlb_flush(CPU(cpu));
64
+ uint64_t ret = env->cp15.hcr_el2 & ~(HCR_VI | HCR_VF);
49
}
65
+ CPUState *cs = ENV_GET_CPU(env);
50
env->cp15.hcr_el2 = value;
66
+
67
+ if (cs->interrupt_request & CPU_INTERRUPT_VIRQ) {
68
+ ret |= HCR_VI;
69
+ }
70
+ if (cs->interrupt_request & CPU_INTERRUPT_VFIQ) {
71
+ ret |= HCR_VF;
72
+ }
73
+ return ret;
74
+}
75
+
76
static const ARMCPRegInfo el2_cp_reginfo[] = {
77
{ .name = "HCR_EL2", .state = ARM_CP_STATE_AA64,
78
+ .type = ARM_CP_IO,
79
.opc0 = 3, .opc1 = 4, .crn = 1, .crm = 1, .opc2 = 0,
80
.access = PL2_RW, .fieldoffset = offsetof(CPUARMState, cp15.hcr_el2),
81
- .writefn = hcr_write },
82
+ .writefn = hcr_write, .readfn = hcr_read },
83
{ .name = "HCR", .state = ARM_CP_STATE_AA32,
84
- .type = ARM_CP_ALIAS,
85
+ .type = ARM_CP_ALIAS | ARM_CP_IO,
86
.cp = 15, .opc1 = 4, .crn = 1, .crm = 1, .opc2 = 0,
87
.access = PL2_RW, .fieldoffset = offsetof(CPUARMState, cp15.hcr_el2),
88
- .writefn = hcr_writelow },
89
+ .writefn = hcr_writelow, .readfn = hcr_read },
90
{ .name = "ELR_EL2", .state = ARM_CP_STATE_AA64,
91
.type = ARM_CP_ALIAS,
92
.opc0 = 3, .opc1 = 4, .crn = 4, .crm = 0, .opc2 = 1,
93
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo el2_cp_reginfo[] = {
94
95
static const ARMCPRegInfo el2_v8_cp_reginfo[] = {
96
{ .name = "HCR2", .state = ARM_CP_STATE_AA32,
97
- .type = ARM_CP_ALIAS,
98
+ .type = ARM_CP_ALIAS | ARM_CP_IO,
99
.cp = 15, .opc1 = 4, .crn = 1, .crm = 1, .opc2 = 4,
100
.access = PL2_RW,
101
.fieldoffset = offsetofhigh32(CPUARMState, cp15.hcr_el2),
102
--
51
--
103
2.19.1
52
2.20.1
104
53
105
54
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move expanders for VBSL, VBIT, and VBIF from translate-a64.c.
3
Emphasize that the is_jmp option exits to the main loop.
4
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181011205206.3552-9-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-6-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
9
---
10
target/arm/translate.h | 6 ++
10
target/arm/translate.h | 14 ++++++++------
11
target/arm/translate-a64.c | 61 --------------
11
target/arm/translate-a64.c | 8 ++++----
12
target/arm/translate.c | 162 +++++++++++++++++++++++++++----------
12
target/arm/translate-vfp.inc.c | 4 ++--
13
3 files changed, 124 insertions(+), 105 deletions(-)
13
target/arm/translate.c | 12 ++++++------
14
4 files changed, 20 insertions(+), 18 deletions(-)
14
15
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.h
18
--- a/target/arm/translate.h
18
+++ b/target/arm/translate.h
19
+++ b/target/arm/translate.h
19
@@ -XXX,XX +XXX,XX @@ static inline TCGv_i32 get_ahp_flag(void)
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
20
return ret;
21
21
}
22
/* is_jmp field values */
22
23
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
23
+
24
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
24
+/* Vector operations shared between ARM and AArch64. */
25
+/* CPU state was modified dynamically; exit to main loop for interrupts. */
25
+extern const GVecGen3 bsl_op;
26
+#define DISAS_UPDATE_EXIT DISAS_TARGET_1
26
+extern const GVecGen3 bit_op;
27
/* These instructions trap after executing, so the A32/T32 decoder must
27
+extern const GVecGen3 bif_op;
28
* defer them until after the conditional execution state has been updated.
28
+
29
* WFI also needs special handling when single-stepping.
29
/*
30
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
30
* Forward to the isar_feature_* tests given a DisasContext pointer.
31
* custom end-of-TB code)
31
*/
32
*/
33
#define DISAS_BX_EXCRET DISAS_TARGET_8
34
-/* For instructions which want an immediate exit to the main loop,
35
- * as opposed to attempting to use lookup_and_goto_ptr. Unlike
36
- * DISAS_UPDATE this doesn't write the PC on exiting the translation
37
- * loop so you need to ensure something (gen_a64_set_pc_im or runtime
38
- * helper) has done so before we reach return from cpu_tb_exec.
39
+/*
40
+ * For instructions which want an immediate exit to the main loop, as opposed
41
+ * to attempting to use lookup_and_goto_ptr. Unlike DISAS_UPDATE_EXIT, this
42
+ * doesn't write the PC on exiting the translation loop so you need to ensure
43
+ * something (gen_a64_set_pc_im or runtime helper) has done so before we reach
44
+ * return from cpu_tb_exec.
45
*/
46
#define DISAS_EXIT DISAS_TARGET_9
47
32
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
48
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
33
index XXXXXXX..XXXXXXX 100644
49
index XXXXXXX..XXXXXXX 100644
34
--- a/target/arm/translate-a64.c
50
--- a/target/arm/translate-a64.c
35
+++ b/target/arm/translate-a64.c
51
+++ b/target/arm/translate-a64.c
36
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn)
52
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
53
gen_helper_msr_i_daifclear(cpu_env, t1);
54
tcg_temp_free_i32(t1);
55
/* For DAIFClear, exit the cpu loop to re-evaluate pending IRQs. */
56
- s->base.is_jmp = DISAS_UPDATE;
57
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
58
break;
59
60
default:
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
63
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
64
/* I/O operations must end the TB here (whether read or write) */
65
- s->base.is_jmp = DISAS_UPDATE;
66
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
67
}
68
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
69
/*
70
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
71
* but allow this to be suppressed by the register definition
72
* (usually only necessary to work around guest bugs).
73
*/
74
- s->base.is_jmp = DISAS_UPDATE;
75
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
37
}
76
}
38
}
77
}
39
78
40
-static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
79
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
41
-{
80
gen_goto_tb(dc, 1, dc->base.pc_next);
42
- tcg_gen_xor_i64(rn, rn, rm);
81
break;
43
- tcg_gen_and_i64(rn, rn, rd);
82
default:
44
- tcg_gen_xor_i64(rd, rm, rn);
83
- case DISAS_UPDATE:
45
-}
84
+ case DISAS_UPDATE_EXIT:
46
-
85
gen_a64_set_pc_im(dc->base.pc_next);
47
-static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
86
/* fall through */
48
-{
87
case DISAS_EXIT:
49
- tcg_gen_xor_i64(rn, rn, rd);
88
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
50
- tcg_gen_and_i64(rn, rn, rm);
89
index XXXXXXX..XXXXXXX 100644
51
- tcg_gen_xor_i64(rd, rd, rn);
90
--- a/target/arm/translate-vfp.inc.c
52
-}
91
+++ b/target/arm/translate-vfp.inc.c
53
-
92
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
54
-static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
93
* this to be the last insn in the TB).
55
-{
94
*/
56
- tcg_gen_xor_i64(rn, rn, rd);
95
if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
57
- tcg_gen_andc_i64(rn, rn, rm);
96
- s->base.is_jmp = DISAS_UPDATE;
58
- tcg_gen_xor_i64(rd, rd, rn);
97
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
59
-}
98
gen_io_start();
60
-
99
}
61
-static void gen_bsl_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
100
gen_helper_v7m_preserve_fp_state(cpu_env);
62
-{
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
63
- tcg_gen_xor_vec(vece, rn, rn, rm);
102
tcg_temp_free_i32(fptr);
64
- tcg_gen_and_vec(vece, rn, rn, rd);
103
65
- tcg_gen_xor_vec(vece, rd, rm, rn);
104
/* End the TB, because we have updated FP control bits */
66
-}
105
- s->base.is_jmp = DISAS_UPDATE;
67
-
106
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
68
-static void gen_bit_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
107
return true;
69
-{
108
}
70
- tcg_gen_xor_vec(vece, rn, rn, rd);
71
- tcg_gen_and_vec(vece, rn, rn, rm);
72
- tcg_gen_xor_vec(vece, rd, rd, rn);
73
-}
74
-
75
-static void gen_bif_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
76
-{
77
- tcg_gen_xor_vec(vece, rn, rn, rd);
78
- tcg_gen_andc_vec(vece, rn, rn, rm);
79
- tcg_gen_xor_vec(vece, rd, rd, rn);
80
-}
81
-
82
/* Logic op (opcode == 3) subgroup of C3.6.16. */
83
static void disas_simd_3same_logic(DisasContext *s, uint32_t insn)
84
{
85
- static const GVecGen3 bsl_op = {
86
- .fni8 = gen_bsl_i64,
87
- .fniv = gen_bsl_vec,
88
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
89
- .load_dest = true
90
- };
91
- static const GVecGen3 bit_op = {
92
- .fni8 = gen_bit_i64,
93
- .fniv = gen_bit_vec,
94
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
95
- .load_dest = true
96
- };
97
- static const GVecGen3 bif_op = {
98
- .fni8 = gen_bif_i64,
99
- .fniv = gen_bif_vec,
100
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
101
- .load_dest = true
102
- };
103
-
104
int rd = extract32(insn, 0, 5);
105
int rn = extract32(insn, 5, 5);
106
int rm = extract32(insn, 16, 5);
107
diff --git a/target/arm/translate.c b/target/arm/translate.c
109
diff --git a/target/arm/translate.c b/target/arm/translate.c
108
index XXXXXXX..XXXXXXX 100644
110
index XXXXXXX..XXXXXXX 100644
109
--- a/target/arm/translate.c
111
--- a/target/arm/translate.c
110
+++ b/target/arm/translate.c
112
+++ b/target/arm/translate.c
111
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
113
@@ -XXX,XX +XXX,XX @@ static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
112
return 0;
114
tcg_temp_free_i32(tcg_tgtmode);
115
tcg_temp_free_i32(tcg_regno);
116
tcg_temp_free_i32(tcg_reg);
117
- s->base.is_jmp = DISAS_UPDATE;
118
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
113
}
119
}
114
120
115
-/* Bitwise select. dest = c ? t : f. Clobbers T and F. */
121
static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
116
-static void gen_neon_bsl(TCGv_i32 dest, TCGv_i32 t, TCGv_i32 f, TCGv_i32 c)
122
@@ -XXX,XX +XXX,XX @@ static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
117
-{
123
tcg_temp_free_i32(tcg_tgtmode);
118
- tcg_gen_and_i32(t, t, c);
124
tcg_temp_free_i32(tcg_regno);
119
- tcg_gen_andc_i32(f, f, c);
125
store_reg(s, rn, tcg_reg);
120
- tcg_gen_or_i32(dest, t, f);
126
- s->base.is_jmp = DISAS_UPDATE;
121
-}
127
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
122
-
123
static inline void gen_neon_narrow(int size, TCGv_i32 dest, TCGv_i64 src)
124
{
125
switch (size) {
126
@@ -XXX,XX +XXX,XX @@ static int do_v81_helper(DisasContext *s, gen_helper_gvec_3_ptr *fn,
127
return 1;
128
}
128
}
129
129
130
+/*
130
/* Store value to PC as for an exception return (ie don't
131
+ * Expanders for VBitOps_VBIF, VBIT, VBSL.
131
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
132
+ */
132
tcg_temp_free_i32(tmp);
133
+static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
133
}
134
+{
134
tcg_temp_free_i32(addr);
135
+ tcg_gen_xor_i64(rn, rn, rm);
135
- s->base.is_jmp = DISAS_UPDATE;
136
+ tcg_gen_and_i64(rn, rn, rd);
136
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
137
+ tcg_gen_xor_i64(rd, rm, rn);
137
}
138
+}
138
139
+
139
/* Generate a label used for skipping this instruction */
140
+static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
140
@@ -XXX,XX +XXX,XX @@ static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
141
+{
141
}
142
+ tcg_gen_xor_i64(rn, rn, rd);
142
if (a->E != (s->be_data == MO_BE)) {
143
+ tcg_gen_and_i64(rn, rn, rm);
143
gen_helper_setend(cpu_env);
144
+ tcg_gen_xor_i64(rd, rd, rn);
144
- s->base.is_jmp = DISAS_UPDATE;
145
+}
145
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
146
+
146
}
147
+static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
147
return true;
148
+{
148
}
149
+ tcg_gen_xor_i64(rn, rn, rd);
149
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
150
+ tcg_gen_andc_i64(rn, rn, rm);
151
+ tcg_gen_xor_i64(rd, rd, rn);
152
+}
153
+
154
+static void gen_bsl_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
155
+{
156
+ tcg_gen_xor_vec(vece, rn, rn, rm);
157
+ tcg_gen_and_vec(vece, rn, rn, rd);
158
+ tcg_gen_xor_vec(vece, rd, rm, rn);
159
+}
160
+
161
+static void gen_bit_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
162
+{
163
+ tcg_gen_xor_vec(vece, rn, rn, rd);
164
+ tcg_gen_and_vec(vece, rn, rn, rm);
165
+ tcg_gen_xor_vec(vece, rd, rd, rn);
166
+}
167
+
168
+static void gen_bif_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
169
+{
170
+ tcg_gen_xor_vec(vece, rn, rn, rd);
171
+ tcg_gen_andc_vec(vece, rn, rn, rm);
172
+ tcg_gen_xor_vec(vece, rd, rd, rn);
173
+}
174
+
175
+const GVecGen3 bsl_op = {
176
+ .fni8 = gen_bsl_i64,
177
+ .fniv = gen_bsl_vec,
178
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
179
+ .load_dest = true
180
+};
181
+
182
+const GVecGen3 bit_op = {
183
+ .fni8 = gen_bit_i64,
184
+ .fniv = gen_bit_vec,
185
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
186
+ .load_dest = true
187
+};
188
+
189
+const GVecGen3 bif_op = {
190
+ .fni8 = gen_bif_i64,
191
+ .fniv = gen_bif_vec,
192
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
193
+ .load_dest = true
194
+};
195
+
196
+
197
/* Translate a NEON data processing instruction. Return nonzero if the
198
instruction is invalid.
199
We process data in a mixture of 32-bit and 64-bit chunks.
200
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
201
{
202
int op;
203
int q;
204
- int rd, rn, rm;
205
+ int rd, rn, rm, rd_ofs, rn_ofs, rm_ofs;
206
int size;
207
int shift;
208
int pass;
209
int count;
210
int pairwise;
211
int u;
212
+ int vec_size;
213
uint32_t imm, mask;
214
TCGv_i32 tmp, tmp2, tmp3, tmp4, tmp5;
215
TCGv_ptr ptr1, ptr2, ptr3;
216
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
217
VFP_DREG_N(rn, insn);
218
VFP_DREG_M(rm, insn);
219
size = (insn >> 20) & 3;
220
+ vec_size = q ? 16 : 8;
221
+ rd_ofs = neon_reg_offset(rd, 0);
222
+ rn_ofs = neon_reg_offset(rn, 0);
223
+ rm_ofs = neon_reg_offset(rm, 0);
224
+
225
if ((insn & (1 << 23)) == 0) {
226
/* Three register same length. */
227
op = ((insn >> 7) & 0x1e) | ((insn >> 4) & 1);
228
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
229
q, rd, rn, rm);
230
}
231
return 1;
232
+
233
+ case NEON_3R_LOGIC: /* Logic ops. */
234
+ switch ((u << 2) | size) {
235
+ case 0: /* VAND */
236
+ tcg_gen_gvec_and(0, rd_ofs, rn_ofs, rm_ofs,
237
+ vec_size, vec_size);
238
+ break;
239
+ case 1: /* VBIC */
240
+ tcg_gen_gvec_andc(0, rd_ofs, rn_ofs, rm_ofs,
241
+ vec_size, vec_size);
242
+ break;
243
+ case 2:
244
+ if (rn == rm) {
245
+ /* VMOV */
246
+ tcg_gen_gvec_mov(0, rd_ofs, rn_ofs, vec_size, vec_size);
247
+ } else {
248
+ /* VORR */
249
+ tcg_gen_gvec_or(0, rd_ofs, rn_ofs, rm_ofs,
250
+ vec_size, vec_size);
251
+ }
252
+ break;
253
+ case 3: /* VORN */
254
+ tcg_gen_gvec_orc(0, rd_ofs, rn_ofs, rm_ofs,
255
+ vec_size, vec_size);
256
+ break;
257
+ case 4: /* VEOR */
258
+ tcg_gen_gvec_xor(0, rd_ofs, rn_ofs, rm_ofs,
259
+ vec_size, vec_size);
260
+ break;
261
+ case 5: /* VBSL */
262
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
263
+ vec_size, vec_size, &bsl_op);
264
+ break;
265
+ case 6: /* VBIT */
266
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
267
+ vec_size, vec_size, &bit_op);
268
+ break;
269
+ case 7: /* VBIF */
270
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
271
+ vec_size, vec_size, &bif_op);
272
+ break;
273
+ }
274
+ return 0;
275
}
276
- if (size == 3 && op != NEON_3R_LOGIC) {
277
+ if (size == 3) {
278
/* 64-bit element instructions. */
279
for (pass = 0; pass < (q ? 2 : 1); pass++) {
280
neon_load_reg64(cpu_V0, rn + pass);
281
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
282
case NEON_3R_VRHADD:
283
GEN_NEON_INTEGER_OP(rhadd);
284
break;
150
break;
285
- case NEON_3R_LOGIC: /* Logic ops. */
151
case DISAS_NEXT:
286
- switch ((u << 2) | size) {
152
case DISAS_TOO_MANY:
287
- case 0: /* VAND */
153
- case DISAS_UPDATE:
288
- tcg_gen_and_i32(tmp, tmp, tmp2);
154
+ case DISAS_UPDATE_EXIT:
289
- break;
155
gen_set_pc_im(dc, dc->base.pc_next);
290
- case 1: /* BIC */
156
/* fall through */
291
- tcg_gen_andc_i32(tmp, tmp, tmp2);
157
default:
292
- break;
158
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
293
- case 2: /* VORR */
159
case DISAS_JUMP:
294
- tcg_gen_or_i32(tmp, tmp, tmp2);
160
gen_goto_ptr();
295
- break;
296
- case 3: /* VORN */
297
- tcg_gen_orc_i32(tmp, tmp, tmp2);
298
- break;
299
- case 4: /* VEOR */
300
- tcg_gen_xor_i32(tmp, tmp, tmp2);
301
- break;
302
- case 5: /* VBSL */
303
- tmp3 = neon_load_reg(rd, pass);
304
- gen_neon_bsl(tmp, tmp, tmp2, tmp3);
305
- tcg_temp_free_i32(tmp3);
306
- break;
307
- case 6: /* VBIT */
308
- tmp3 = neon_load_reg(rd, pass);
309
- gen_neon_bsl(tmp, tmp, tmp3, tmp2);
310
- tcg_temp_free_i32(tmp3);
311
- break;
312
- case 7: /* VBIF */
313
- tmp3 = neon_load_reg(rd, pass);
314
- gen_neon_bsl(tmp, tmp3, tmp, tmp2);
315
- tcg_temp_free_i32(tmp3);
316
- break;
317
- }
318
- break;
319
case NEON_3R_VHSUB:
320
GEN_NEON_INTEGER_OP(hsub);
321
break;
161
break;
162
- case DISAS_UPDATE:
163
+ case DISAS_UPDATE_EXIT:
164
gen_set_pc_im(dc, dc->base.pc_next);
165
/* fall through */
166
default:
322
--
167
--
323
2.19.1
168
2.20.1
324
169
325
170
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move cmtst_op expanders from translate-a64.c.
3
Add an option that writes back the PC, like DISAS_UPDATE_EXIT,
4
but does not exit back to the main loop.
4
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181011205206.3552-17-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-7-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
---
10
target/arm/translate.h | 2 +
11
target/arm/translate.h | 2 ++
11
target/arm/translate-a64.c | 38 ------------------
12
target/arm/translate-a64.c | 3 +++
12
target/arm/translate.c | 81 +++++++++++++++++++++++++++-----------
13
target/arm/translate.c | 4 ++++
13
3 files changed, 60 insertions(+), 61 deletions(-)
14
3 files changed, 9 insertions(+)
14
15
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.h
18
--- a/target/arm/translate.h
18
+++ b/target/arm/translate.h
19
+++ b/target/arm/translate.h
19
@@ -XXX,XX +XXX,XX @@ extern const GVecGen3 bit_op;
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
20
extern const GVecGen3 bif_op;
21
* return from cpu_tb_exec.
21
extern const GVecGen3 mla_op[4];
22
*/
22
extern const GVecGen3 mls_op[4];
23
#define DISAS_EXIT DISAS_TARGET_9
23
+extern const GVecGen3 cmtst_op[4];
24
+/* CPU state was modified dynamically; no need to exit, but do not chain. */
24
extern const GVecGen2i ssra_op[4];
25
+#define DISAS_UPDATE_NOCHAIN DISAS_TARGET_10
25
extern const GVecGen2i usra_op[4];
26
26
extern const GVecGen2i sri_op[4];
27
#ifdef TARGET_AARCH64
27
extern const GVecGen2i sli_op[4];
28
void a64_translate_init(void);
28
+void gen_cmtst_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b);
29
30
/*
31
* Forward to the isar_feature_* tests given a DisasContext pointer.
32
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
33
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
34
--- a/target/arm/translate-a64.c
31
--- a/target/arm/translate-a64.c
35
+++ b/target/arm/translate-a64.c
32
+++ b/target/arm/translate-a64.c
36
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_three_reg_diff(DisasContext *s, uint32_t insn)
33
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
37
}
34
case DISAS_EXIT:
38
}
35
tcg_gen_exit_tb(NULL, 0);
39
36
break;
40
-/* CMTST : test is "if (X & Y != 0)". */
37
+ case DISAS_UPDATE_NOCHAIN:
41
-static void gen_cmtst_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
38
+ gen_a64_set_pc_im(dc->base.pc_next);
42
-{
39
+ /* fall through */
43
- tcg_gen_and_i32(d, a, b);
40
case DISAS_JUMP:
44
- tcg_gen_setcondi_i32(TCG_COND_NE, d, d, 0);
41
tcg_gen_lookup_and_goto_ptr();
45
- tcg_gen_neg_i32(d, d);
42
break;
46
-}
47
-
48
-static void gen_cmtst_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
49
-{
50
- tcg_gen_and_i64(d, a, b);
51
- tcg_gen_setcondi_i64(TCG_COND_NE, d, d, 0);
52
- tcg_gen_neg_i64(d, d);
53
-}
54
-
55
-static void gen_cmtst_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
56
-{
57
- tcg_gen_and_vec(vece, d, a, b);
58
- tcg_gen_dupi_vec(vece, a, 0);
59
- tcg_gen_cmp_vec(TCG_COND_NE, vece, d, d, a);
60
-}
61
-
62
static void handle_3same_64(DisasContext *s, int opcode, bool u,
63
TCGv_i64 tcg_rd, TCGv_i64 tcg_rn, TCGv_i64 tcg_rm)
64
{
65
@@ -XXX,XX +XXX,XX @@ static void disas_simd_3same_float(DisasContext *s, uint32_t insn)
66
/* Integer op subgroup of C3.6.16. */
67
static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
68
{
69
- static const GVecGen3 cmtst_op[4] = {
70
- { .fni4 = gen_helper_neon_tst_u8,
71
- .fniv = gen_cmtst_vec,
72
- .vece = MO_8 },
73
- { .fni4 = gen_helper_neon_tst_u16,
74
- .fniv = gen_cmtst_vec,
75
- .vece = MO_16 },
76
- { .fni4 = gen_cmtst_i32,
77
- .fniv = gen_cmtst_vec,
78
- .vece = MO_32 },
79
- { .fni8 = gen_cmtst_i64,
80
- .fniv = gen_cmtst_vec,
81
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
82
- .vece = MO_64 },
83
- };
84
-
85
int is_q = extract32(insn, 30, 1);
86
int u = extract32(insn, 29, 1);
87
int size = extract32(insn, 22, 2);
88
diff --git a/target/arm/translate.c b/target/arm/translate.c
43
diff --git a/target/arm/translate.c b/target/arm/translate.c
89
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate.c
45
--- a/target/arm/translate.c
91
+++ b/target/arm/translate.c
46
+++ b/target/arm/translate.c
92
@@ -XXX,XX +XXX,XX @@ const GVecGen3 mls_op[4] = {
47
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
93
.vece = MO_64 },
48
case DISAS_NEXT:
94
};
49
case DISAS_TOO_MANY:
95
50
case DISAS_UPDATE_EXIT:
96
+/* CMTST : test is "if (X & Y != 0)". */
51
+ case DISAS_UPDATE_NOCHAIN:
97
+static void gen_cmtst_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
52
gen_set_pc_im(dc, dc->base.pc_next);
98
+{
53
/* fall through */
99
+ tcg_gen_and_i32(d, a, b);
54
default:
100
+ tcg_gen_setcondi_i32(TCG_COND_NE, d, d, 0);
55
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
101
+ tcg_gen_neg_i32(d, d);
56
case DISAS_TOO_MANY:
102
+}
57
gen_goto_tb(dc, 1, dc->base.pc_next);
103
+
104
+void gen_cmtst_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
105
+{
106
+ tcg_gen_and_i64(d, a, b);
107
+ tcg_gen_setcondi_i64(TCG_COND_NE, d, d, 0);
108
+ tcg_gen_neg_i64(d, d);
109
+}
110
+
111
+static void gen_cmtst_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
112
+{
113
+ tcg_gen_and_vec(vece, d, a, b);
114
+ tcg_gen_dupi_vec(vece, a, 0);
115
+ tcg_gen_cmp_vec(TCG_COND_NE, vece, d, d, a);
116
+}
117
+
118
+const GVecGen3 cmtst_op[4] = {
119
+ { .fni4 = gen_helper_neon_tst_u8,
120
+ .fniv = gen_cmtst_vec,
121
+ .vece = MO_8 },
122
+ { .fni4 = gen_helper_neon_tst_u16,
123
+ .fniv = gen_cmtst_vec,
124
+ .vece = MO_16 },
125
+ { .fni4 = gen_cmtst_i32,
126
+ .fniv = gen_cmtst_vec,
127
+ .vece = MO_32 },
128
+ { .fni8 = gen_cmtst_i64,
129
+ .fniv = gen_cmtst_vec,
130
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
131
+ .vece = MO_64 },
132
+};
133
+
134
/* Translate a NEON data processing instruction. Return nonzero if the
135
instruction is invalid.
136
We process data in a mixture of 32-bit and 64-bit chunks.
137
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
138
tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size,
139
u ? &mls_op[size] : &mla_op[size]);
140
return 0;
141
+
142
+ case NEON_3R_VTST_VCEQ:
143
+ if (u) { /* VCEQ */
144
+ tcg_gen_gvec_cmp(TCG_COND_EQ, size, rd_ofs, rn_ofs, rm_ofs,
145
+ vec_size, vec_size);
146
+ } else { /* VTST */
147
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
148
+ vec_size, vec_size, &cmtst_op[size]);
149
+ }
150
+ return 0;
151
+
152
+ case NEON_3R_VCGT:
153
+ tcg_gen_gvec_cmp(u ? TCG_COND_GTU : TCG_COND_GT, size,
154
+ rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
155
+ return 0;
156
+
157
+ case NEON_3R_VCGE:
158
+ tcg_gen_gvec_cmp(u ? TCG_COND_GEU : TCG_COND_GE, size,
159
+ rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size);
160
+ return 0;
161
}
162
163
if (size == 3) {
164
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
165
case NEON_3R_VQSUB:
166
GEN_NEON_INTEGER_OP_ENV(qsub);
167
break;
58
break;
168
- case NEON_3R_VCGT:
59
+ case DISAS_UPDATE_NOCHAIN:
169
- GEN_NEON_INTEGER_OP(cgt);
60
+ gen_set_pc_im(dc, dc->base.pc_next);
170
- break;
61
+ /* fall through */
171
- case NEON_3R_VCGE:
62
case DISAS_JUMP:
172
- GEN_NEON_INTEGER_OP(cge);
63
gen_goto_ptr();
173
- break;
174
case NEON_3R_VSHL:
175
GEN_NEON_INTEGER_OP(shl);
176
break;
64
break;
177
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
178
tmp2 = neon_load_reg(rd, pass);
179
gen_neon_add(size, tmp, tmp2);
180
break;
181
- case NEON_3R_VTST_VCEQ:
182
- if (!u) { /* VTST */
183
- switch (size) {
184
- case 0: gen_helper_neon_tst_u8(tmp, tmp, tmp2); break;
185
- case 1: gen_helper_neon_tst_u16(tmp, tmp, tmp2); break;
186
- case 2: gen_helper_neon_tst_u32(tmp, tmp, tmp2); break;
187
- default: abort();
188
- }
189
- } else { /* VCEQ */
190
- switch (size) {
191
- case 0: gen_helper_neon_ceq_u8(tmp, tmp, tmp2); break;
192
- case 1: gen_helper_neon_ceq_u16(tmp, tmp, tmp2); break;
193
- case 2: gen_helper_neon_ceq_u32(tmp, tmp, tmp2); break;
194
- default: abort();
195
- }
196
- }
197
- break;
198
case NEON_3R_VMUL:
199
/* VMUL.P8; other cases already eliminated. */
200
gen_helper_neon_mul_p8(tmp, tmp, tmp2);
201
--
65
--
202
2.19.1
66
2.20.1
203
67
204
68
diff view generated by jsdifflib
1
For the v7 version of the Arm architecture, the IL bit in
1
From: Richard Henderson <richard.henderson@linaro.org>
2
syndrome register values where the field is not valid was
2
3
defined to be UNK/SBZP. In v8 this is RES1, which is what
3
This is TFSRE0_EL1, TFSR_EL1, TFSR_EL2, TFSR_EL3,
4
QEMU currently implements. Handle the desired v7 behaviour
4
RGSR_EL1, GCR_EL1, GMID_EL1, and PSTATE.TCO.
5
by squashing the IL bit for the affected cases:
5
6
* EC == EC_UNCATEGORIZED
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
* prefetch aborts
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
* data aborts where ISV is 0
8
Message-id: 20200626033144.790098-8-richard.henderson@linaro.org
9
10
(The fourth case listed in the v8 Arm ARM DDI 0487C.a in
11
section G7.2.70, "illegal state exception", can't happen
12
on a v7 CPU.)
13
14
This deals with a corner case noted in a comment.
15
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20181012144235.19646-10-peter.maydell@linaro.org
19
---
10
---
20
target/arm/internals.h | 7 ++-----
11
target/arm/cpu.h | 4 ++
21
target/arm/helper.c | 13 +++++++++++++
12
target/arm/internals.h | 9 ++++
22
2 files changed, 15 insertions(+), 5 deletions(-)
13
target/arm/helper.c | 94 ++++++++++++++++++++++++++++++++++++++
23
14
target/arm/translate-a64.c | 21 +++++++++
15
4 files changed, 128 insertions(+)
16
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
uint64_t pmccfiltr_el0; /* Performance Monitor Filter Register */
23
uint64_t vpidr_el2; /* Virtualization Processor ID Register */
24
uint64_t vmpidr_el2; /* Virtualization Multiprocessor ID Register */
25
+ uint64_t tfsr_el[4]; /* tfsre0_el1 is index 0. */
26
+ uint64_t gcr_el1;
27
+ uint64_t rgsr_el1;
28
} cp15;
29
30
struct {
31
@@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu);
32
#define PSTATE_SS (1U << 21)
33
#define PSTATE_PAN (1U << 22)
34
#define PSTATE_UAO (1U << 23)
35
+#define PSTATE_TCO (1U << 25)
36
#define PSTATE_V (1U << 28)
37
#define PSTATE_C (1U << 29)
38
#define PSTATE_Z (1U << 30)
24
diff --git a/target/arm/internals.h b/target/arm/internals.h
39
diff --git a/target/arm/internals.h b/target/arm/internals.h
25
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/internals.h
41
--- a/target/arm/internals.h
27
+++ b/target/arm/internals.h
42
+++ b/target/arm/internals.h
28
@@ -XXX,XX +XXX,XX @@ static inline uint32_t syn_get_ec(uint32_t syn)
43
@@ -XXX,XX +XXX,XX @@ static inline uint32_t aarch64_pstate_valid_mask(const ARMISARegisters *id)
29
/* Utility functions for constructing various kinds of syndrome value.
44
if (isar_feature_aa64_uao(id)) {
30
* Note that in general we follow the AArch64 syndrome values; in a
45
valid |= PSTATE_UAO;
31
* few cases the value in HSR for exceptions taken to AArch32 Hyp
46
}
32
- * mode differs slightly, so if we ever implemented Hyp mode then the
47
+ if (isar_feature_aa64_mte(id)) {
33
- * syndrome value would need some massaging on exception entry.
48
+ valid |= PSTATE_TCO;
34
- * (One example of this is that AArch64 defaults to IL bit set for
49
+ }
35
- * exceptions which don't specifically indicate information about the
50
36
- * trapping instruction, whereas AArch32 defaults to IL bit clear.)
51
return valid;
37
+ * mode differs slightly, and we fix this up when populating HSR in
52
}
38
+ * arm_cpu_do_interrupt_aarch32_hyp().
53
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
39
*/
54
40
static inline uint32_t syn_uncategorized(void)
55
#endif /* !CONFIG_USER_ONLY */
41
{
56
57
+/*
58
+ * The log2 of the words in the tag block, for GMID_EL1.BS.
59
+ * The is the maximum, 256 bytes, which manipulates 64-bits of tags.
60
+ */
61
+#define GMID_EL1_BS 6
62
+
63
#endif
42
diff --git a/target/arm/helper.c b/target/arm/helper.c
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
43
index XXXXXXX..XXXXXXX 100644
65
index XXXXXXX..XXXXXXX 100644
44
--- a/target/arm/helper.c
66
--- a/target/arm/helper.c
45
+++ b/target/arm/helper.c
67
+++ b/target/arm/helper.c
46
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch32_hyp(CPUState *cs)
68
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
69
{ K(3, 0, 1, 2, 0), K(3, 4, 1, 2, 0), K(3, 5, 1, 2, 0),
70
"ZCR_EL1", "ZCR_EL2", "ZCR_EL12", isar_feature_aa64_sve },
71
72
+ { K(3, 0, 5, 6, 0), K(3, 4, 5, 6, 0), K(3, 5, 5, 6, 0),
73
+ "TFSR_EL1", "TFSR_EL2", "TFSR_EL12", isar_feature_aa64_mte },
74
+
75
/* TODO: ARMv8.2-SPE -- PMSCR_EL2 */
76
/* TODO: ARMv8.4-Trace -- TRFCR_EL2 */
77
};
78
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
79
};
80
#endif /*CONFIG_USER_ONLY*/
81
82
+static CPAccessResult access_aa64_tid5(CPUARMState *env, const ARMCPRegInfo *ri,
83
+ bool isread)
84
+{
85
+ if ((arm_current_el(env) < 2) && (arm_hcr_el2_eff(env) & HCR_TID5)) {
86
+ return CP_ACCESS_TRAP_EL2;
87
+ }
88
+
89
+ return CP_ACCESS_OK;
90
+}
91
+
92
+static CPAccessResult access_mte(CPUARMState *env, const ARMCPRegInfo *ri,
93
+ bool isread)
94
+{
95
+ int el = arm_current_el(env);
96
+
97
+ if (el < 2 &&
98
+ arm_feature(env, ARM_FEATURE_EL2) &&
99
+ !(arm_hcr_el2_eff(env) & HCR_ATA)) {
100
+ return CP_ACCESS_TRAP_EL2;
101
+ }
102
+ if (el < 3 &&
103
+ arm_feature(env, ARM_FEATURE_EL3) &&
104
+ !(env->cp15.scr_el3 & SCR_ATA)) {
105
+ return CP_ACCESS_TRAP_EL3;
106
+ }
107
+ return CP_ACCESS_OK;
108
+}
109
+
110
+static uint64_t tco_read(CPUARMState *env, const ARMCPRegInfo *ri)
111
+{
112
+ return env->pstate & PSTATE_TCO;
113
+}
114
+
115
+static void tco_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t val)
116
+{
117
+ env->pstate = (env->pstate & ~PSTATE_TCO) | (val & PSTATE_TCO);
118
+}
119
+
120
+static const ARMCPRegInfo mte_reginfo[] = {
121
+ { .name = "TFSRE0_EL1", .state = ARM_CP_STATE_AA64,
122
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 1,
123
+ .access = PL1_RW, .accessfn = access_mte,
124
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[0]) },
125
+ { .name = "TFSR_EL1", .state = ARM_CP_STATE_AA64,
126
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 0,
127
+ .access = PL1_RW, .accessfn = access_mte,
128
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[1]) },
129
+ { .name = "TFSR_EL2", .state = ARM_CP_STATE_AA64,
130
+ .opc0 = 3, .opc1 = 4, .crn = 5, .crm = 6, .opc2 = 0,
131
+ .access = PL2_RW, .accessfn = access_mte,
132
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[2]) },
133
+ { .name = "TFSR_EL3", .state = ARM_CP_STATE_AA64,
134
+ .opc0 = 3, .opc1 = 6, .crn = 5, .crm = 6, .opc2 = 0,
135
+ .access = PL3_RW,
136
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[3]) },
137
+ { .name = "RGSR_EL1", .state = ARM_CP_STATE_AA64,
138
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 5,
139
+ .access = PL1_RW, .accessfn = access_mte,
140
+ .fieldoffset = offsetof(CPUARMState, cp15.rgsr_el1) },
141
+ { .name = "GCR_EL1", .state = ARM_CP_STATE_AA64,
142
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 6,
143
+ .access = PL1_RW, .accessfn = access_mte,
144
+ .fieldoffset = offsetof(CPUARMState, cp15.gcr_el1) },
145
+ { .name = "GMID_EL1", .state = ARM_CP_STATE_AA64,
146
+ .opc0 = 3, .opc1 = 1, .crn = 0, .crm = 0, .opc2 = 4,
147
+ .access = PL1_R, .accessfn = access_aa64_tid5,
148
+ .type = ARM_CP_CONST, .resetvalue = GMID_EL1_BS },
149
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
150
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
151
+ .type = ARM_CP_NO_RAW,
152
+ .access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
153
+ REGINFO_SENTINEL
154
+};
155
+
156
+static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
157
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
158
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
159
+ .type = ARM_CP_CONST, .access = PL0_RW, },
160
+ REGINFO_SENTINEL
161
+};
162
#endif
163
164
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
165
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
166
}
47
}
167
}
48
168
#endif /*CONFIG_USER_ONLY*/
49
if (cs->exception_index != EXCP_IRQ && cs->exception_index != EXCP_FIQ) {
169
+
50
+ if (!arm_feature(env, ARM_FEATURE_V8)) {
170
+ /*
51
+ /*
171
+ * If full MTE is enabled, add all of the system registers.
52
+ * QEMU syndrome values are v8-style. v7 has the IL bit
172
+ * If only "instructions available at EL0" are enabled,
53
+ * UNK/SBZP for "field not valid" cases, where v8 uses RES1.
173
+ * then define only a RAZ/WI version of PSTATE.TCO.
54
+ * If this is a v7 CPU, squash the IL bit in those cases.
174
+ */
55
+ */
175
+ if (cpu_isar_feature(aa64_mte, cpu)) {
56
+ if (cs->exception_index == EXCP_PREFETCH_ABORT ||
176
+ define_arm_cp_regs(cpu, mte_reginfo);
57
+ (cs->exception_index == EXCP_DATA_ABORT &&
177
+ } else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
58
+ !(env->exception.syndrome & ARM_EL_ISV)) ||
178
+ define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
59
+ syn_get_ec(env->exception.syndrome) == EC_UNCATEGORIZED) {
179
+ }
60
+ env->exception.syndrome &= ~ARM_EL_IL;
180
#endif
181
182
if (cpu_isar_feature(any_predinv, cpu)) {
183
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
184
index XXXXXXX..XXXXXXX 100644
185
--- a/target/arm/translate-a64.c
186
+++ b/target/arm/translate-a64.c
187
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
188
s->base.is_jmp = DISAS_UPDATE_EXIT;
189
break;
190
191
+ case 0x1c: /* TCO */
192
+ if (dc_isar_feature(aa64_mte, s)) {
193
+ /* Full MTE is enabled -- set the TCO bit as directed. */
194
+ if (crm & 1) {
195
+ set_pstate_bits(PSTATE_TCO);
196
+ } else {
197
+ clear_pstate_bits(PSTATE_TCO);
61
+ }
198
+ }
199
+ t1 = tcg_const_i32(s->current_el);
200
+ gen_helper_rebuild_hflags_a64(cpu_env, t1);
201
+ tcg_temp_free_i32(t1);
202
+ /* Many factors, including TCO, go into MTE_ACTIVE. */
203
+ s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
204
+ } else if (dc_isar_feature(aa64_mte_insn_reg, s)) {
205
+ /* Only "instructions accessible at EL0" -- PSTATE.TCO is WI. */
206
+ s->base.is_jmp = DISAS_NEXT;
207
+ } else {
208
+ goto do_unallocated;
62
+ }
209
+ }
63
env->cp15.esr_el[2] = env->exception.syndrome;
210
+ break;
64
}
211
+
65
212
default:
213
do_unallocated:
214
unallocated_encoding(s);
66
--
215
--
67
2.19.1
216
2.20.1
68
217
69
218
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
3
Cache the composite ATA setting.
4
5
Cache when MTE is fully enabled, i.e. access to tags are enabled
6
and tag checks affect the PE. Do this for both the normal context
7
and the UNPRIV context.
8
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20181016223115.24100-8-richard.henderson@linaro.org
11
Message-id: 20200626033144.790098-9-richard.henderson@linaro.org
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
13
---
9
target/arm/cpu.h | 16 +++++++++++++++-
14
target/arm/cpu.h | 12 ++++++++----
10
linux-user/aarch64/signal.c | 4 ++--
15
target/arm/internals.h | 18 +++++++++++++++++
11
linux-user/elfload.c | 2 +-
16
target/arm/translate.h | 5 +++++
12
linux-user/syscall.c | 10 ++++++----
17
target/arm/helper.c | 40 ++++++++++++++++++++++++++++++++++++++
13
target/arm/cpu64.c | 5 ++++-
18
target/arm/translate-a64.c | 4 ++++
14
target/arm/helper.c | 9 ++++++---
19
5 files changed, 75 insertions(+), 4 deletions(-)
15
target/arm/machine.c | 3 +--
16
target/arm/translate-a64.c | 4 ++--
17
8 files changed, 37 insertions(+), 16 deletions(-)
18
20
19
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
20
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.h
23
--- a/target/arm/cpu.h
22
+++ b/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
23
@@ -XXX,XX +XXX,XX @@ FIELD(ID_AA64ISAR1, FRINTTS, 32, 4)
25
@@ -XXX,XX +XXX,XX @@ typedef ARMCPU ArchCPU;
24
FIELD(ID_AA64ISAR1, SB, 36, 4)
26
* | | | TBFLAG_A32 | |
25
FIELD(ID_AA64ISAR1, SPECRES, 40, 4)
27
* | | +-----+----------+ TBFLAG_AM32 |
26
28
* | TBFLAG_ANY | |TBFLAG_M32| |
27
+FIELD(ID_AA64PFR0, EL0, 0, 4)
29
- * | | +-+----------+--------------|
28
+FIELD(ID_AA64PFR0, EL1, 4, 4)
30
- * | | | TBFLAG_A64 |
29
+FIELD(ID_AA64PFR0, EL2, 8, 4)
31
- * +--------------+---------+---------------------------+
30
+FIELD(ID_AA64PFR0, EL3, 12, 4)
32
- * 31 20 15 0
31
+FIELD(ID_AA64PFR0, FP, 16, 4)
33
+ * | +-----------+----------+--------------|
32
+FIELD(ID_AA64PFR0, ADVSIMD, 20, 4)
34
+ * | | TBFLAG_A64 |
33
+FIELD(ID_AA64PFR0, GIC, 24, 4)
35
+ * +--------------+-------------------------------------+
34
+FIELD(ID_AA64PFR0, RAS, 28, 4)
36
+ * 31 20 0
35
+FIELD(ID_AA64PFR0, SVE, 32, 4)
37
*
36
+
38
* Unless otherwise noted, these bits are cached in env->hflags.
37
QEMU_BUILD_BUG_ON(ARRAY_SIZE(((ARMCPU *)0)->ccsidr) <= R_V7M_CSSELR_INDEX_MASK);
39
*/
38
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, BT, 9, 1)
39
/* If adding a feature bit which corresponds to a Linux ELF
41
FIELD(TBFLAG_A64, BTYPE, 10, 2) /* Not cached. */
40
@@ -XXX,XX +XXX,XX @@ enum arm_features {
42
FIELD(TBFLAG_A64, TBID, 12, 2)
41
ARM_FEATURE_PMU, /* has PMU support */
43
FIELD(TBFLAG_A64, UNPRIV, 14, 1)
42
ARM_FEATURE_VBAR, /* has cp15 VBAR */
44
+FIELD(TBFLAG_A64, ATA, 15, 1)
43
ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
45
+FIELD(TBFLAG_A64, TCMA, 16, 2)
44
- ARM_FEATURE_SVE, /* has Scalable Vector Extension */
46
+FIELD(TBFLAG_A64, MTE_ACTIVE, 18, 1)
45
ARM_FEATURE_V8_FP16, /* implements v8.2 half-precision float */
47
+FIELD(TBFLAG_A64, MTE0_ACTIVE, 19, 1)
46
ARM_FEATURE_M_MAIN, /* M profile Main Extension */
48
47
};
49
/**
48
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_fcma(const ARMISARegisters *id)
50
* cpu_mmu_index:
49
return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, FCMA) != 0;
51
diff --git a/target/arm/internals.h b/target/arm/internals.h
52
index XXXXXXX..XXXXXXX 100644
53
--- a/target/arm/internals.h
54
+++ b/target/arm/internals.h
55
@@ -XXX,XX +XXX,XX @@ static inline int exception_target_el(CPUARMState *env)
56
return target_el;
50
}
57
}
51
58
52
+static inline bool isar_feature_aa64_sve(const ARMISARegisters *id)
59
+/* Determine if allocation tags are available. */
60
+static inline bool allocation_tag_access_enabled(CPUARMState *env, int el,
61
+ uint64_t sctlr)
53
+{
62
+{
54
+ return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, SVE) != 0;
63
+ if (el < 3
64
+ && arm_feature(env, ARM_FEATURE_EL3)
65
+ && !(env->cp15.scr_el3 & SCR_ATA)) {
66
+ return false;
67
+ }
68
+ if (el < 2
69
+ && arm_feature(env, ARM_FEATURE_EL2)
70
+ && !(arm_hcr_el2_eff(env) & HCR_ATA)) {
71
+ return false;
72
+ }
73
+ sctlr &= (el == 0 ? SCTLR_ATA0 : SCTLR_ATA);
74
+ return sctlr != 0;
55
+}
75
+}
56
+
76
+
57
/*
77
#ifndef CONFIG_USER_ONLY
58
* Forward to the above feature tests given an ARMCPU pointer.
78
59
*/
79
/* Security attributes for an address, as returned by v8m_security_lookup. */
60
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
80
diff --git a/target/arm/translate.h b/target/arm/translate.h
61
index XXXXXXX..XXXXXXX 100644
81
index XXXXXXX..XXXXXXX 100644
62
--- a/linux-user/aarch64/signal.c
82
--- a/target/arm/translate.h
63
+++ b/linux-user/aarch64/signal.c
83
+++ b/target/arm/translate.h
64
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
84
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
65
break;
85
ARMMMUIdx mmu_idx; /* MMU index to use for normal loads/stores */
66
86
uint8_t tbii; /* TBI1|TBI0 for insns */
67
case TARGET_SVE_MAGIC:
87
uint8_t tbid; /* TBI1|TBI0 for data */
68
- if (arm_feature(env, ARM_FEATURE_SVE)) {
88
+ uint8_t tcma; /* TCMA1|TCMA0 for MTE */
69
+ if (cpu_isar_feature(aa64_sve, arm_env_get_cpu(env))) {
89
bool ns; /* Use non-secure CPREG bank on access */
70
vq = (env->vfp.zcr_el[1] & 0xf) + 1;
90
int fp_excp_el; /* FP exception EL or 0 if enabled */
71
sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(vq), 16);
91
int sve_excp_el; /* SVE exception EL or 0 if enabled */
72
if (!sve && size == sve_size) {
92
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
73
@@ -XXX,XX +XXX,XX @@ static void target_setup_frame(int usig, struct target_sigaction *ka,
93
bool unpriv;
74
&layout);
94
/* True if v8.3-PAuth is active. */
75
95
bool pauth_active;
76
/* SVE state needs saving only if it exists. */
96
+ /* True if v8.5-MTE access to tags is enabled. */
77
- if (arm_feature(env, ARM_FEATURE_SVE)) {
97
+ bool ata;
78
+ if (cpu_isar_feature(aa64_sve, arm_env_get_cpu(env))) {
98
+ /* True if v8.5-MTE tag checks affect the PE; index with is_unpriv. */
79
vq = (env->vfp.zcr_el[1] & 0xf) + 1;
99
+ bool mte_active[2];
80
sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(vq), 16);
100
/* True with v8.5-BTI and SCTLR_ELx.BT* set. */
81
sve_ofs = alloc_sigframe_space(sve_size, &layout);
101
bool bt;
82
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
102
/* True if any CP15 access is trapped by HSTR_EL2 */
83
index XXXXXXX..XXXXXXX 100644
84
--- a/linux-user/elfload.c
85
+++ b/linux-user/elfload.c
86
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
87
GET_FEATURE_ID(aa64_rdm, ARM_HWCAP_A64_ASIMDRDM);
88
GET_FEATURE_ID(aa64_dp, ARM_HWCAP_A64_ASIMDDP);
89
GET_FEATURE_ID(aa64_fcma, ARM_HWCAP_A64_FCMA);
90
- GET_FEATURE(ARM_FEATURE_SVE, ARM_HWCAP_A64_SVE);
91
+ GET_FEATURE_ID(aa64_sve, ARM_HWCAP_A64_SVE);
92
93
#undef GET_FEATURE
94
#undef GET_FEATURE_ID
95
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
96
index XXXXXXX..XXXXXXX 100644
97
--- a/linux-user/syscall.c
98
+++ b/linux-user/syscall.c
99
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
100
* even though the current architectural maximum is VQ=16.
101
*/
102
ret = -TARGET_EINVAL;
103
- if (arm_feature(cpu_env, ARM_FEATURE_SVE)
104
+ if (cpu_isar_feature(aa64_sve, arm_env_get_cpu(cpu_env))
105
&& arg2 >= 0 && arg2 <= 512 * 16 && !(arg2 & 15)) {
106
CPUARMState *env = cpu_env;
107
ARMCPU *cpu = arm_env_get_cpu(env);
108
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
109
return ret;
110
case TARGET_PR_SVE_GET_VL:
111
ret = -TARGET_EINVAL;
112
- if (arm_feature(cpu_env, ARM_FEATURE_SVE)) {
113
- CPUARMState *env = cpu_env;
114
- ret = ((env->vfp.zcr_el[1] & 0xf) + 1) * 16;
115
+ {
116
+ ARMCPU *cpu = arm_env_get_cpu(cpu_env);
117
+ if (cpu_isar_feature(aa64_sve, cpu)) {
118
+ ret = ((cpu->env.vfp.zcr_el[1] & 0xf) + 1) * 16;
119
+ }
120
}
121
return ret;
122
#endif /* AARCH64 */
123
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
124
index XXXXXXX..XXXXXXX 100644
125
--- a/target/arm/cpu64.c
126
+++ b/target/arm/cpu64.c
127
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
128
t = FIELD_DP64(t, ID_AA64ISAR1, FCMA, 1);
129
cpu->isar.id_aa64isar1 = t;
130
131
+ t = cpu->isar.id_aa64pfr0;
132
+ t = FIELD_DP64(t, ID_AA64PFR0, SVE, 1);
133
+ cpu->isar.id_aa64pfr0 = t;
134
+
135
/* Replicate the same data to the 32-bit id registers. */
136
u = cpu->isar.id_isar5;
137
u = FIELD_DP32(u, ID_ISAR5, AES, 2); /* AES + PMULL */
138
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
139
* present in either.
140
*/
141
set_feature(&cpu->env, ARM_FEATURE_V8_FP16);
142
- set_feature(&cpu->env, ARM_FEATURE_SVE);
143
/* For usermode -cpu max we can use a larger and more efficient DCZ
144
* blocksize since we don't have to follow what the hardware does.
145
*/
146
diff --git a/target/arm/helper.c b/target/arm/helper.c
103
diff --git a/target/arm/helper.c b/target/arm/helper.c
147
index XXXXXXX..XXXXXXX 100644
104
index XXXXXXX..XXXXXXX 100644
148
--- a/target/arm/helper.c
105
--- a/target/arm/helper.c
149
+++ b/target/arm/helper.c
106
+++ b/target/arm/helper.c
150
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
107
@@ -XXX,XX +XXX,XX @@ static int aa64_va_parameter_tbid(uint64_t tcr, ARMMMUIdx mmu_idx)
151
define_one_arm_cp_reg(cpu, &sctlr);
152
}
108
}
153
109
}
154
- if (arm_feature(env, ARM_FEATURE_SVE)) {
110
155
+ if (cpu_isar_feature(aa64_sve, cpu)) {
111
+static int aa64_va_parameter_tcma(uint64_t tcr, ARMMMUIdx mmu_idx)
156
define_one_arm_cp_reg(cpu, &zcr_el1_reginfo);
112
+{
157
if (arm_feature(env, ARM_FEATURE_EL2)) {
113
+ if (regime_has_2_ranges(mmu_idx)) {
158
define_one_arm_cp_reg(cpu, &zcr_el2_reginfo);
114
+ return extract64(tcr, 57, 2);
159
@@ -XXX,XX +XXX,XX @@ void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
115
+ } else {
160
uint32_t flags;
116
+ /* Replicate the single TCMA bit so we always have 2 bits. */
161
117
+ return extract32(tcr, 30, 1) * 3;
162
if (is_a64(env)) {
118
+ }
163
+ ARMCPU *cpu = arm_env_get_cpu(env);
119
+}
164
+
120
+
165
*pc = env->pc;
121
ARMVAParameters aa64_va_parameters(CPUARMState *env, uint64_t va,
166
flags = ARM_TBFLAG_AARCH64_STATE_MASK;
122
ARMMMUIdx mmu_idx, bool data)
167
/* Get control bits for tagged addresses */
168
flags |= (arm_regime_tbi0(env, mmu_idx) << ARM_TBFLAG_TBI0_SHIFT);
169
flags |= (arm_regime_tbi1(env, mmu_idx) << ARM_TBFLAG_TBI1_SHIFT);
170
171
- if (arm_feature(env, ARM_FEATURE_SVE)) {
172
+ if (cpu_isar_feature(aa64_sve, cpu)) {
173
int sve_el = sve_exception_el(env, current_el);
174
uint32_t zcr_len;
175
176
@@ -XXX,XX +XXX,XX @@ void aarch64_sve_narrow_vq(CPUARMState *env, unsigned vq)
177
void aarch64_sve_change_el(CPUARMState *env, int old_el,
178
int new_el, bool el0_a64)
179
{
123
{
180
+ ARMCPU *cpu = arm_env_get_cpu(env);
124
@@ -XXX,XX +XXX,XX @@ static uint32_t rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
181
int old_len, new_len;
125
}
182
bool old_a64, new_a64;
183
184
/* Nothing to do if no SVE. */
185
- if (!arm_feature(env, ARM_FEATURE_SVE)) {
186
+ if (!cpu_isar_feature(aa64_sve, cpu)) {
187
return;
188
}
126
}
189
127
190
diff --git a/target/arm/machine.c b/target/arm/machine.c
128
+ if (cpu_isar_feature(aa64_mte, env_archcpu(env))) {
191
index XXXXXXX..XXXXXXX 100644
129
+ /*
192
--- a/target/arm/machine.c
130
+ * Set MTE_ACTIVE if any access may be Checked, and leave clear
193
+++ b/target/arm/machine.c
131
+ * if all accesses must be Unchecked:
194
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_iwmmxt = {
132
+ * 1) If no TBI, then there are no tags in the address to check,
195
static bool sve_needed(void *opaque)
133
+ * 2) If Tag Check Override, then all accesses are Unchecked,
196
{
134
+ * 3) If Tag Check Fail == 0, then Checked access have no effect,
197
ARMCPU *cpu = opaque;
135
+ * 4) If no Allocation Tag Access, then all accesses are Unchecked.
198
- CPUARMState *env = &cpu->env;
136
+ */
199
137
+ if (allocation_tag_access_enabled(env, el, sctlr)) {
200
- return arm_feature(env, ARM_FEATURE_SVE);
138
+ flags = FIELD_DP32(flags, TBFLAG_A64, ATA, 1);
201
+ return cpu_isar_feature(aa64_sve, cpu);
139
+ if (tbid
140
+ && !(env->pstate & PSTATE_TCO)
141
+ && (sctlr & (el == 0 ? SCTLR_TCF0 : SCTLR_TCF))) {
142
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE_ACTIVE, 1);
143
+ }
144
+ }
145
+ /* And again for unprivileged accesses, if required. */
146
+ if (FIELD_EX32(flags, TBFLAG_A64, UNPRIV)
147
+ && tbid
148
+ && !(env->pstate & PSTATE_TCO)
149
+ && (sctlr & SCTLR_TCF0)
150
+ && allocation_tag_access_enabled(env, 0, sctlr)) {
151
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE0_ACTIVE, 1);
152
+ }
153
+ /* Cache TCMA as well as TBI. */
154
+ flags = FIELD_DP32(flags, TBFLAG_A64, TCMA,
155
+ aa64_va_parameter_tcma(tcr, mmu_idx));
156
+ }
157
+
158
return rebuild_hflags_common(env, fp_el, mmu_idx, flags);
202
}
159
}
203
160
204
/* The first two words of each Zreg is stored in VFP state. */
205
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
161
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
206
index XXXXXXX..XXXXXXX 100644
162
index XXXXXXX..XXXXXXX 100644
207
--- a/target/arm/translate-a64.c
163
--- a/target/arm/translate-a64.c
208
+++ b/target/arm/translate-a64.c
164
+++ b/target/arm/translate-a64.c
209
@@ -XXX,XX +XXX,XX @@ void aarch64_cpu_dump_state(CPUState *cs, FILE *f,
165
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
210
cpu_fprintf(f, " FPCR=%08x FPSR=%08x\n",
166
dc->mmu_idx = core_to_aa64_mmu_idx(core_mmu_idx);
211
vfp_get_fpcr(env), vfp_get_fpsr(env));
167
dc->tbii = FIELD_EX32(tb_flags, TBFLAG_A64, TBII);
212
168
dc->tbid = FIELD_EX32(tb_flags, TBFLAG_A64, TBID);
213
- if (arm_feature(env, ARM_FEATURE_SVE) && sve_exception_el(env, el) == 0) {
169
+ dc->tcma = FIELD_EX32(tb_flags, TBFLAG_A64, TCMA);
214
+ if (cpu_isar_feature(aa64_sve, cpu) && sve_exception_el(env, el) == 0) {
170
dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
215
int j, zcr_len = sve_zcr_len_for_el(env, el);
171
#if !defined(CONFIG_USER_ONLY)
216
172
dc->user = (dc->current_el == 0);
217
for (i = 0; i <= FFR_PRED_NUM; i++) {
173
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
218
@@ -XXX,XX +XXX,XX @@ static void disas_a64_insn(CPUARMState *env, DisasContext *s)
174
dc->bt = FIELD_EX32(tb_flags, TBFLAG_A64, BT);
219
unallocated_encoding(s);
175
dc->btype = FIELD_EX32(tb_flags, TBFLAG_A64, BTYPE);
220
break;
176
dc->unpriv = FIELD_EX32(tb_flags, TBFLAG_A64, UNPRIV);
221
case 0x2:
177
+ dc->ata = FIELD_EX32(tb_flags, TBFLAG_A64, ATA);
222
- if (!arm_dc_feature(s, ARM_FEATURE_SVE) || !disas_sve(s, insn)) {
178
+ dc->mte_active[0] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE_ACTIVE);
223
+ if (!dc_isar_feature(aa64_sve, s) || !disas_sve(s, insn)) {
179
+ dc->mte_active[1] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE0_ACTIVE);
224
unallocated_encoding(s);
180
dc->vec_len = 0;
225
}
181
dc->vec_stride = 0;
226
break;
182
dc->cp_regs = arm_cpu->cp_regs;
227
--
183
--
228
2.19.1
184
2.20.1
229
185
230
186
diff view generated by jsdifflib
1
For traps of FP/SIMD instructions to AArch32 Hyp mode, the syndrome
1
From: Richard Henderson <richard.henderson@linaro.org>
2
provided in HSR has more information than is reported to AArch64.
3
Specifically, there are extra fields TA and coproc which indicate
4
whether the trapped instruction was FP or SIMD. Add this extra
5
information to the syndromes we construct, and mask it out when
6
taking the exception to AArch64.
7
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-10-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20181012144235.19646-11-peter.maydell@linaro.org
11
---
7
---
12
target/arm/internals.h | 14 +++++++++++++-
8
target/arm/helper-a64.h | 2 ++
13
target/arm/helper.c | 9 +++++++++
9
target/arm/internals.h | 5 +++
14
target/arm/translate.c | 8 ++++----
10
target/arm/mte_helper.c | 72 ++++++++++++++++++++++++++++++++++++++
15
3 files changed, 26 insertions(+), 5 deletions(-)
11
target/arm/translate-a64.c | 18 ++++++++++
12
target/arm/Makefile.objs | 1 +
13
5 files changed, 98 insertions(+)
14
create mode 100644 target/arm/mte_helper.c
16
15
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-a64.h
19
+++ b/target/arm/helper-a64.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autda, TCG_CALL_NO_WG, i64, env, i64, i64)
21
DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
22
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
+
25
+DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
17
diff --git a/target/arm/internals.h b/target/arm/internals.h
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
18
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/internals.h
28
--- a/target/arm/internals.h
20
+++ b/target/arm/internals.h
29
+++ b/target/arm/internals.h
21
@@ -XXX,XX +XXX,XX @@ static inline uint32_t syn_get_ec(uint32_t syn)
30
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
22
* few cases the value in HSR for exceptions taken to AArch32 Hyp
23
* mode differs slightly, and we fix this up when populating HSR in
24
* arm_cpu_do_interrupt_aarch32_hyp().
25
+ * The exception is FP/SIMD access traps -- these report extra information
26
+ * when taking an exception to AArch32. For those we include the extra coproc
27
+ * and TA fields, and mask them out when taking the exception to AArch64.
28
*/
31
*/
29
static inline uint32_t syn_uncategorized(void)
32
#define GMID_EL1_BS 6
30
{
33
31
@@ -XXX,XX +XXX,XX @@ static inline uint32_t syn_cp15_rrt_trap(int cv, int cond, int opc1, int crm,
34
+static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
32
35
+{
33
static inline uint32_t syn_fp_access_trap(int cv, int cond, bool is_16bit)
36
+ return deposit64(ptr, 56, 4, rtag);
34
{
35
+ /* AArch32 FP trap or any AArch64 FP/SIMD trap: TA == 0 coproc == 0xa */
36
return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
37
| (is_16bit ? 0 : ARM_EL_IL)
38
- | (cv << 24) | (cond << 20);
39
+ | (cv << 24) | (cond << 20) | 0xa;
40
+}
37
+}
41
+
38
+
42
+static inline uint32_t syn_simd_access_trap(int cv, int cond, bool is_16bit)
39
#endif
40
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
41
new file mode 100644
42
index XXXXXXX..XXXXXXX
43
--- /dev/null
44
+++ b/target/arm/mte_helper.c
45
@@ -XXX,XX +XXX,XX @@
46
+/*
47
+ * ARM v8.5-MemTag Operations
48
+ *
49
+ * Copyright (c) 2020 Linaro, Ltd.
50
+ *
51
+ * This library is free software; you can redistribute it and/or
52
+ * modify it under the terms of the GNU Lesser General Public
53
+ * License as published by the Free Software Foundation; either
54
+ * version 2.1 of the License, or (at your option) any later version.
55
+ *
56
+ * This library is distributed in the hope that it will be useful,
57
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
58
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
59
+ * Lesser General Public License for more details.
60
+ *
61
+ * You should have received a copy of the GNU Lesser General Public
62
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
63
+ */
64
+
65
+#include "qemu/osdep.h"
66
+#include "cpu.h"
67
+#include "internals.h"
68
+#include "exec/exec-all.h"
69
+#include "exec/cpu_ldst.h"
70
+#include "exec/helper-proto.h"
71
+
72
+
73
+static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
43
+{
74
+{
44
+ /* AArch32 SIMD trap: TA == 1 coproc == 0 */
75
+ if (exclude == 0xffff) {
45
+ return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
76
+ return 0;
46
+ | (is_16bit ? 0 : ARM_EL_IL)
77
+ }
47
+ | (cv << 24) | (cond << 20) | (1 << 5);
78
+ if (offset == 0) {
79
+ while (exclude & (1 << tag)) {
80
+ tag = (tag + 1) & 15;
81
+ }
82
+ } else {
83
+ do {
84
+ do {
85
+ tag = (tag + 1) & 15;
86
+ } while (exclude & (1 << tag));
87
+ } while (--offset > 0);
88
+ }
89
+ return tag;
90
+}
91
+
92
+uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
93
+{
94
+ int rtag;
95
+
96
+ /*
97
+ * Our IMPDEF choice for GCR_EL1.RRND==1 is to behave as if
98
+ * GCR_EL1.RRND==0, always producing deterministic results.
99
+ */
100
+ uint16_t exclude = extract32(rm | env->cp15.gcr_el1, 0, 16);
101
+ int start = extract32(env->cp15.rgsr_el1, 0, 4);
102
+ int seed = extract32(env->cp15.rgsr_el1, 8, 16);
103
+ int offset, i;
104
+
105
+ /* RandomTag */
106
+ for (i = offset = 0; i < 4; ++i) {
107
+ /* NextRandomTagBit */
108
+ int top = (extract32(seed, 5, 1) ^ extract32(seed, 3, 1) ^
109
+ extract32(seed, 2, 1) ^ extract32(seed, 0, 1));
110
+ seed = (top << 15) | (seed >> 1);
111
+ offset |= top << i;
112
+ }
113
+ rtag = choose_nonexcluded_tag(start, offset, exclude);
114
+ env->cp15.rgsr_el1 = rtag | (seed << 8);
115
+
116
+ return address_with_allocation_tag(rn, rtag);
117
+}
118
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
119
index XXXXXXX..XXXXXXX 100644
120
--- a/target/arm/translate-a64.c
121
+++ b/target/arm/translate-a64.c
122
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
123
return clean;
48
}
124
}
49
125
50
static inline uint32_t syn_sve_access_trap(void)
126
+/* Insert a zero tag into src, with the result at dst. */
51
diff --git a/target/arm/helper.c b/target/arm/helper.c
127
+static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
128
+{
129
+ tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
130
+}
131
+
132
typedef struct DisasCompare64 {
133
TCGCond cond;
134
TCGv_i64 value;
135
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
136
case 3: /* SDIV */
137
handle_div(s, true, sf, rm, rn, rd);
138
break;
139
+ case 4: /* IRG */
140
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
141
+ goto do_unallocated;
142
+ }
143
+ if (s->ata) {
144
+ gen_helper_irg(cpu_reg_sp(s, rd), cpu_env,
145
+ cpu_reg_sp(s, rn), cpu_reg(s, rm));
146
+ } else {
147
+ gen_address_with_allocation_tag0(cpu_reg_sp(s, rd),
148
+ cpu_reg_sp(s, rn));
149
+ }
150
+ break;
151
case 8: /* LSLV */
152
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
153
break;
154
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
52
index XXXXXXX..XXXXXXX 100644
155
index XXXXXXX..XXXXXXX 100644
53
--- a/target/arm/helper.c
156
--- a/target/arm/Makefile.objs
54
+++ b/target/arm/helper.c
157
+++ b/target/arm/Makefile.objs
55
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
158
@@ -XXX,XX +XXX,XX @@ obj-$(CONFIG_SOFTMMU) += psci.o
56
case EXCP_HVC:
159
obj-$(TARGET_AARCH64) += translate-a64.o helper-a64.o
57
case EXCP_HYP_TRAP:
160
obj-$(TARGET_AARCH64) += translate-sve.o sve_helper.o
58
case EXCP_SMC:
161
obj-$(TARGET_AARCH64) += pauth_helper.o
59
+ if (syn_get_ec(env->exception.syndrome) == EC_ADVSIMDFPACCESSTRAP) {
162
+obj-$(TARGET_AARCH64) += mte_helper.o
60
+ /*
61
+ * QEMU internal FP/SIMD syndromes from AArch32 include the
62
+ * TA and coproc fields which are only exposed if the exception
63
+ * is taken to AArch32 Hyp mode. Mask them out to get a valid
64
+ * AArch64 format syndrome.
65
+ */
66
+ env->exception.syndrome &= ~MAKE_64BIT_MASK(0, 20);
67
+ }
68
env->cp15.esr_el[new_el] = env->exception.syndrome;
69
break;
70
case EXCP_IRQ:
71
diff --git a/target/arm/translate.c b/target/arm/translate.c
72
index XXXXXXX..XXXXXXX 100644
73
--- a/target/arm/translate.c
74
+++ b/target/arm/translate.c
75
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
76
*/
77
if (s->fp_excp_el) {
78
gen_exception_insn(s, 4, EXCP_UDEF,
79
- syn_fp_access_trap(1, 0xe, false), s->fp_excp_el);
80
+ syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
81
return 0;
82
}
83
84
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
85
*/
86
if (s->fp_excp_el) {
87
gen_exception_insn(s, 4, EXCP_UDEF,
88
- syn_fp_access_trap(1, 0xe, false), s->fp_excp_el);
89
+ syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
90
return 0;
91
}
92
93
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
94
95
if (s->fp_excp_el) {
96
gen_exception_insn(s, 4, EXCP_UDEF,
97
- syn_fp_access_trap(1, 0xe, false), s->fp_excp_el);
98
+ syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
99
return 0;
100
}
101
if (!s->vfp_enabled) {
102
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
103
104
if (s->fp_excp_el) {
105
gen_exception_insn(s, 4, EXCP_UDEF,
106
- syn_fp_access_trap(1, 0xe, false), s->fp_excp_el);
107
+ syn_simd_access_trap(1, 0xe, false), s->fp_excp_el);
108
return 0;
109
}
110
if (!s->vfp_enabled) {
111
--
163
--
112
2.19.1
164
2.20.1
113
165
114
166
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
For a sequence of loads or stores from a single register,
3
The current Arm ARM has adjusted the official decode of
4
little-endian operations can be promoted to an 8-byte op.
4
"Add/subtract (immediate)" so that the shift field is only bit 22,
5
This can reduce the number of operations by a factor of 8.
5
and bit 23 is part of the op1 field of the parent category
6
"Data processing - immediate".
6
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20181011205206.3552-20-richard.henderson@linaro.org
10
Message-id: 20200626033144.790098-11-richard.henderson@linaro.org
9
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
11
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
14
---
13
target/arm/translate.c | 10 ++++++++++
15
target/arm/translate-a64.c | 23 ++++++++---------------
14
1 file changed, 10 insertions(+)
16
1 file changed, 8 insertions(+), 15 deletions(-)
15
17
16
diff --git a/target/arm/translate.c b/target/arm/translate.c
18
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
17
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.c
20
--- a/target/arm/translate-a64.c
19
+++ b/target/arm/translate.c
21
+++ b/target/arm/translate-a64.c
20
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
22
@@ -XXX,XX +XXX,XX @@ static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
21
if (size == 3 && (interleave | spacing) != 1) {
23
/*
22
return 1;
24
* Add/subtract (immediate)
23
}
25
*
24
+ /* For our purposes, bytes are always little-endian. */
26
- * 31 30 29 28 24 23 22 21 10 9 5 4 0
25
+ if (size == 0) {
27
- * +--+--+--+-----------+-----+-------------+-----+-----+
26
+ endian = MO_LE;
28
- * |sf|op| S| 1 0 0 0 1 |shift| imm12 | Rn | Rd |
27
+ }
29
- * +--+--+--+-----------+-----+-------------+-----+-----+
28
+ /* Consecutive little-endian elements from a single register
30
+ * 31 30 29 28 23 22 21 10 9 5 4 0
29
+ * can be promoted to a larger little-endian operation.
31
+ * +--+--+--+-------------+--+-------------+-----+-----+
30
+ */
32
+ * |sf|op| S| 1 0 0 0 1 0 |sh| imm12 | Rn | Rd |
31
+ if (interleave == 1 && endian == MO_LE) {
33
+ * +--+--+--+-------------+--+-------------+-----+-----+
32
+ size = 3;
34
*
33
+ }
35
* sf: 0 -> 32bit, 1 -> 64bit
34
tmp64 = tcg_temp_new_i64();
36
* op: 0 -> add , 1 -> sub
35
addr = tcg_temp_new_i32();
37
* S: 1 -> set flags
36
tmp2 = tcg_const_i32(1 << size);
38
- * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
39
+ * sh: 1 -> LSL imm by 12
40
*/
41
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
42
{
43
int rd = extract32(insn, 0, 5);
44
int rn = extract32(insn, 5, 5);
45
uint64_t imm = extract32(insn, 10, 12);
46
- int shift = extract32(insn, 22, 2);
47
+ bool shift = extract32(insn, 22, 1);
48
bool setflags = extract32(insn, 29, 1);
49
bool sub_op = extract32(insn, 30, 1);
50
bool is_64bit = extract32(insn, 31, 1);
51
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
53
TCGv_i64 tcg_result;
54
55
- switch (shift) {
56
- case 0x0:
57
- break;
58
- case 0x1:
59
+ if (shift) {
60
imm <<= 12;
61
- break;
62
- default:
63
- unallocated_encoding(s);
64
- return;
65
}
66
67
tcg_result = tcg_temp_new_i64();
68
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
69
case 0x20: case 0x21: /* PC-rel. addressing */
70
disas_pc_rel_adr(s, insn);
71
break;
72
- case 0x22: case 0x23: /* Add/subtract (immediate) */
73
+ case 0x22: /* Add/subtract (immediate) */
74
disas_add_sub_imm(s, insn);
75
break;
76
case 0x24: /* Logical (immediate) */
37
--
77
--
38
2.19.1
78
2.20.1
39
79
40
80
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move shi_op and sli_op expanders from translate-a64.c.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181011205206.3552-15-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-12-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
7
---
10
target/arm/translate.h | 2 +
8
target/arm/helper-a64.h | 1 +
11
target/arm/translate-a64.c | 152 +----------------------
9
target/arm/internals.h | 9 +++++++
12
target/arm/translate.c | 244 ++++++++++++++++++++++++++-----------
10
target/arm/mte_helper.c | 10 ++++++++
13
3 files changed, 179 insertions(+), 219 deletions(-)
11
target/arm/translate-a64.c | 51 ++++++++++++++++++++++++++++++++++++++
12
4 files changed, 71 insertions(+)
14
13
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
16
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.h
16
--- a/target/arm/helper-a64.h
18
+++ b/target/arm/translate.h
17
+++ b/target/arm/helper-a64.h
19
@@ -XXX,XX +XXX,XX @@ extern const GVecGen3 bit_op;
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
20
extern const GVecGen3 bif_op;
19
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
21
extern const GVecGen2i ssra_op[4];
20
22
extern const GVecGen2i usra_op[4];
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
23
+extern const GVecGen2i sri_op[4];
22
+DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
24
+extern const GVecGen2i sli_op[4];
23
diff --git a/target/arm/internals.h b/target/arm/internals.h
25
24
index XXXXXXX..XXXXXXX 100644
26
/*
25
--- a/target/arm/internals.h
27
* Forward to the isar_feature_* tests given a DisasContext pointer.
26
+++ b/target/arm/internals.h
27
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
28
*/
29
#define GMID_EL1_BS 6
30
31
+/* We associate one allocation tag per 16 bytes, the minimum. */
32
+#define LOG2_TAG_GRANULE 4
33
+#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
34
+
35
+static inline int allocation_tag_from_addr(uint64_t ptr)
36
+{
37
+ return extract64(ptr, 56, 4);
38
+}
39
+
40
static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
41
{
42
return deposit64(ptr, 56, 4, rtag);
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
48
49
return address_with_allocation_tag(rn, rtag);
50
}
51
+
52
+uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
53
+ int32_t offset, uint32_t tag_offset)
54
+{
55
+ int start_tag = allocation_tag_from_addr(ptr);
56
+ uint16_t exclude = extract32(env->cp15.gcr_el1, 0, 16);
57
+ int rtag = choose_nonexcluded_tag(start_tag, tag_offset, exclude);
58
+
59
+ return address_with_allocation_tag(ptr + offset, rtag);
60
+}
28
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
61
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
29
index XXXXXXX..XXXXXXX 100644
62
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/translate-a64.c
63
--- a/target/arm/translate-a64.c
31
+++ b/target/arm/translate-a64.c
64
+++ b/target/arm/translate-a64.c
32
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_two_reg_misc(DisasContext *s, uint32_t insn)
65
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
33
}
66
tcg_temp_free_i64(tcg_result);
34
}
67
}
35
68
36
-static void gen_shr8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
69
+/*
37
-{
70
+ * Add/subtract (immediate, with tags)
38
- uint64_t mask = dup_const(MO_8, 0xff >> shift);
71
+ *
39
- TCGv_i64 t = tcg_temp_new_i64();
72
+ * 31 30 29 28 23 22 21 16 14 10 9 5 4 0
40
-
73
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
41
- tcg_gen_shri_i64(t, a, shift);
74
+ * |sf|op| S| 1 0 0 0 1 1 |o2| uimm6 |o3| uimm4 | Rn | Rd |
42
- tcg_gen_andi_i64(t, t, mask);
75
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
43
- tcg_gen_andi_i64(d, d, ~mask);
76
+ *
44
- tcg_gen_or_i64(d, d, t);
77
+ * op: 0 -> add, 1 -> sub
45
- tcg_temp_free_i64(t);
78
+ */
46
-}
79
+static void disas_add_sub_imm_with_tags(DisasContext *s, uint32_t insn)
47
-
48
-static void gen_shr16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
49
-{
50
- uint64_t mask = dup_const(MO_16, 0xffff >> shift);
51
- TCGv_i64 t = tcg_temp_new_i64();
52
-
53
- tcg_gen_shri_i64(t, a, shift);
54
- tcg_gen_andi_i64(t, t, mask);
55
- tcg_gen_andi_i64(d, d, ~mask);
56
- tcg_gen_or_i64(d, d, t);
57
- tcg_temp_free_i64(t);
58
-}
59
-
60
-static void gen_shr32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
61
-{
62
- tcg_gen_shri_i32(a, a, shift);
63
- tcg_gen_deposit_i32(d, d, a, 0, 32 - shift);
64
-}
65
-
66
-static void gen_shr64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
67
-{
68
- tcg_gen_shri_i64(a, a, shift);
69
- tcg_gen_deposit_i64(d, d, a, 0, 64 - shift);
70
-}
71
-
72
-static void gen_shr_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
73
-{
74
- uint64_t mask = (2ull << ((8 << vece) - 1)) - 1;
75
- TCGv_vec t = tcg_temp_new_vec_matching(d);
76
- TCGv_vec m = tcg_temp_new_vec_matching(d);
77
-
78
- tcg_gen_dupi_vec(vece, m, mask ^ (mask >> sh));
79
- tcg_gen_shri_vec(vece, t, a, sh);
80
- tcg_gen_and_vec(vece, d, d, m);
81
- tcg_gen_or_vec(vece, d, d, t);
82
-
83
- tcg_temp_free_vec(t);
84
- tcg_temp_free_vec(m);
85
-}
86
-
87
/* SSHR[RA]/USHR[RA] - Vector shift right (optional rounding/accumulate) */
88
static void handle_vec_simd_shri(DisasContext *s, bool is_q, bool is_u,
89
int immh, int immb, int opcode, int rn, int rd)
90
{
91
- static const GVecGen2i sri_op[4] = {
92
- { .fni8 = gen_shr8_ins_i64,
93
- .fniv = gen_shr_ins_vec,
94
- .load_dest = true,
95
- .opc = INDEX_op_shri_vec,
96
- .vece = MO_8 },
97
- { .fni8 = gen_shr16_ins_i64,
98
- .fniv = gen_shr_ins_vec,
99
- .load_dest = true,
100
- .opc = INDEX_op_shri_vec,
101
- .vece = MO_16 },
102
- { .fni4 = gen_shr32_ins_i32,
103
- .fniv = gen_shr_ins_vec,
104
- .load_dest = true,
105
- .opc = INDEX_op_shri_vec,
106
- .vece = MO_32 },
107
- { .fni8 = gen_shr64_ins_i64,
108
- .fniv = gen_shr_ins_vec,
109
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
110
- .load_dest = true,
111
- .opc = INDEX_op_shri_vec,
112
- .vece = MO_64 },
113
- };
114
-
115
int size = 32 - clz32(immh) - 1;
116
int immhb = immh << 3 | immb;
117
int shift = 2 * (8 << size) - immhb;
118
@@ -XXX,XX +XXX,XX @@ static void handle_vec_simd_shri(DisasContext *s, bool is_q, bool is_u,
119
clear_vec_high(s, is_q, rd);
120
}
121
122
-static void gen_shl8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
123
-{
124
- uint64_t mask = dup_const(MO_8, 0xff << shift);
125
- TCGv_i64 t = tcg_temp_new_i64();
126
-
127
- tcg_gen_shli_i64(t, a, shift);
128
- tcg_gen_andi_i64(t, t, mask);
129
- tcg_gen_andi_i64(d, d, ~mask);
130
- tcg_gen_or_i64(d, d, t);
131
- tcg_temp_free_i64(t);
132
-}
133
-
134
-static void gen_shl16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
135
-{
136
- uint64_t mask = dup_const(MO_16, 0xffff << shift);
137
- TCGv_i64 t = tcg_temp_new_i64();
138
-
139
- tcg_gen_shli_i64(t, a, shift);
140
- tcg_gen_andi_i64(t, t, mask);
141
- tcg_gen_andi_i64(d, d, ~mask);
142
- tcg_gen_or_i64(d, d, t);
143
- tcg_temp_free_i64(t);
144
-}
145
-
146
-static void gen_shl32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
147
-{
148
- tcg_gen_deposit_i32(d, d, a, shift, 32 - shift);
149
-}
150
-
151
-static void gen_shl64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
152
-{
153
- tcg_gen_deposit_i64(d, d, a, shift, 64 - shift);
154
-}
155
-
156
-static void gen_shl_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
157
-{
158
- uint64_t mask = (1ull << sh) - 1;
159
- TCGv_vec t = tcg_temp_new_vec_matching(d);
160
- TCGv_vec m = tcg_temp_new_vec_matching(d);
161
-
162
- tcg_gen_dupi_vec(vece, m, mask);
163
- tcg_gen_shli_vec(vece, t, a, sh);
164
- tcg_gen_and_vec(vece, d, d, m);
165
- tcg_gen_or_vec(vece, d, d, t);
166
-
167
- tcg_temp_free_vec(t);
168
- tcg_temp_free_vec(m);
169
-}
170
-
171
/* SHL/SLI - Vector shift left */
172
static void handle_vec_simd_shli(DisasContext *s, bool is_q, bool insert,
173
int immh, int immb, int opcode, int rn, int rd)
174
{
175
- static const GVecGen2i shi_op[4] = {
176
- { .fni8 = gen_shl8_ins_i64,
177
- .fniv = gen_shl_ins_vec,
178
- .opc = INDEX_op_shli_vec,
179
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
180
- .load_dest = true,
181
- .vece = MO_8 },
182
- { .fni8 = gen_shl16_ins_i64,
183
- .fniv = gen_shl_ins_vec,
184
- .opc = INDEX_op_shli_vec,
185
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
186
- .load_dest = true,
187
- .vece = MO_16 },
188
- { .fni4 = gen_shl32_ins_i32,
189
- .fniv = gen_shl_ins_vec,
190
- .opc = INDEX_op_shli_vec,
191
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
192
- .load_dest = true,
193
- .vece = MO_32 },
194
- { .fni8 = gen_shl64_ins_i64,
195
- .fniv = gen_shl_ins_vec,
196
- .opc = INDEX_op_shli_vec,
197
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
198
- .load_dest = true,
199
- .vece = MO_64 },
200
- };
201
int size = 32 - clz32(immh) - 1;
202
int immhb = immh << 3 | immb;
203
int shift = immhb - (8 << size);
204
@@ -XXX,XX +XXX,XX @@ static void handle_vec_simd_shli(DisasContext *s, bool is_q, bool insert,
205
}
206
207
if (insert) {
208
- gen_gvec_op2i(s, is_q, rd, rn, shift, &shi_op[size]);
209
+ gen_gvec_op2i(s, is_q, rd, rn, shift, &sli_op[size]);
210
} else {
211
gen_gvec_fn2i(s, is_q, rd, rn, shift, tcg_gen_gvec_shli, size);
212
}
213
diff --git a/target/arm/translate.c b/target/arm/translate.c
214
index XXXXXXX..XXXXXXX 100644
215
--- a/target/arm/translate.c
216
+++ b/target/arm/translate.c
217
@@ -XXX,XX +XXX,XX @@ const GVecGen2i usra_op[4] = {
218
.vece = MO_64, },
219
};
220
221
+static void gen_shr8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
222
+{
80
+{
223
+ uint64_t mask = dup_const(MO_8, 0xff >> shift);
81
+ int rd = extract32(insn, 0, 5);
224
+ TCGv_i64 t = tcg_temp_new_i64();
82
+ int rn = extract32(insn, 5, 5);
83
+ int uimm4 = extract32(insn, 10, 4);
84
+ int uimm6 = extract32(insn, 16, 6);
85
+ bool sub_op = extract32(insn, 30, 1);
86
+ TCGv_i64 tcg_rn, tcg_rd;
87
+ int imm;
225
+
88
+
226
+ tcg_gen_shri_i64(t, a, shift);
89
+ /* Test all of sf=1, S=0, o2=0, o3=0. */
227
+ tcg_gen_andi_i64(t, t, mask);
90
+ if ((insn & 0xa040c000u) != 0x80000000u ||
228
+ tcg_gen_andi_i64(d, d, ~mask);
91
+ !dc_isar_feature(aa64_mte_insn_reg, s)) {
229
+ tcg_gen_or_i64(d, d, t);
92
+ unallocated_encoding(s);
230
+ tcg_temp_free_i64(t);
93
+ return;
231
+}
94
+ }
232
+
95
+
233
+static void gen_shr16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
96
+ imm = uimm6 << LOG2_TAG_GRANULE;
234
+{
97
+ if (sub_op) {
235
+ uint64_t mask = dup_const(MO_16, 0xffff >> shift);
98
+ imm = -imm;
236
+ TCGv_i64 t = tcg_temp_new_i64();
99
+ }
237
+
100
+
238
+ tcg_gen_shri_i64(t, a, shift);
101
+ tcg_rn = cpu_reg_sp(s, rn);
239
+ tcg_gen_andi_i64(t, t, mask);
102
+ tcg_rd = cpu_reg_sp(s, rd);
240
+ tcg_gen_andi_i64(d, d, ~mask);
241
+ tcg_gen_or_i64(d, d, t);
242
+ tcg_temp_free_i64(t);
243
+}
244
+
103
+
245
+static void gen_shr32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
104
+ if (s->ata) {
246
+{
105
+ TCGv_i32 offset = tcg_const_i32(imm);
247
+ tcg_gen_shri_i32(a, a, shift);
106
+ TCGv_i32 tag_offset = tcg_const_i32(uimm4);
248
+ tcg_gen_deposit_i32(d, d, a, 0, 32 - shift);
249
+}
250
+
107
+
251
+static void gen_shr64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
108
+ gen_helper_addsubg(tcg_rd, cpu_env, tcg_rn, offset, tag_offset);
252
+{
109
+ tcg_temp_free_i32(tag_offset);
253
+ tcg_gen_shri_i64(a, a, shift);
110
+ tcg_temp_free_i32(offset);
254
+ tcg_gen_deposit_i64(d, d, a, 0, 64 - shift);
255
+}
256
+
257
+static void gen_shr_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
258
+{
259
+ if (sh == 0) {
260
+ tcg_gen_mov_vec(d, a);
261
+ } else {
111
+ } else {
262
+ TCGv_vec t = tcg_temp_new_vec_matching(d);
112
+ tcg_gen_addi_i64(tcg_rd, tcg_rn, imm);
263
+ TCGv_vec m = tcg_temp_new_vec_matching(d);
113
+ gen_address_with_allocation_tag0(tcg_rd, tcg_rd);
264
+
265
+ tcg_gen_dupi_vec(vece, m, MAKE_64BIT_MASK((8 << vece) - sh, sh));
266
+ tcg_gen_shri_vec(vece, t, a, sh);
267
+ tcg_gen_and_vec(vece, d, d, m);
268
+ tcg_gen_or_vec(vece, d, d, t);
269
+
270
+ tcg_temp_free_vec(t);
271
+ tcg_temp_free_vec(m);
272
+ }
114
+ }
273
+}
115
+}
274
+
116
+
275
+const GVecGen2i sri_op[4] = {
117
/* The input should be a value in the bottom e bits (with higher
276
+ { .fni8 = gen_shr8_ins_i64,
118
* bits zero); returns that value replicated into every element
277
+ .fniv = gen_shr_ins_vec,
119
* of size e in a 64 bit integer.
278
+ .load_dest = true,
120
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
279
+ .opc = INDEX_op_shri_vec,
121
case 0x22: /* Add/subtract (immediate) */
280
+ .vece = MO_8 },
122
disas_add_sub_imm(s, insn);
281
+ { .fni8 = gen_shr16_ins_i64,
123
break;
282
+ .fniv = gen_shr_ins_vec,
124
+ case 0x23: /* Add/subtract (immediate, with tags) */
283
+ .load_dest = true,
125
+ disas_add_sub_imm_with_tags(s, insn);
284
+ .opc = INDEX_op_shri_vec,
126
+ break;
285
+ .vece = MO_16 },
127
case 0x24: /* Logical (immediate) */
286
+ { .fni4 = gen_shr32_ins_i32,
128
disas_logic_imm(s, insn);
287
+ .fniv = gen_shr_ins_vec,
129
break;
288
+ .load_dest = true,
289
+ .opc = INDEX_op_shri_vec,
290
+ .vece = MO_32 },
291
+ { .fni8 = gen_shr64_ins_i64,
292
+ .fniv = gen_shr_ins_vec,
293
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
294
+ .load_dest = true,
295
+ .opc = INDEX_op_shri_vec,
296
+ .vece = MO_64 },
297
+};
298
+
299
+static void gen_shl8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
300
+{
301
+ uint64_t mask = dup_const(MO_8, 0xff << shift);
302
+ TCGv_i64 t = tcg_temp_new_i64();
303
+
304
+ tcg_gen_shli_i64(t, a, shift);
305
+ tcg_gen_andi_i64(t, t, mask);
306
+ tcg_gen_andi_i64(d, d, ~mask);
307
+ tcg_gen_or_i64(d, d, t);
308
+ tcg_temp_free_i64(t);
309
+}
310
+
311
+static void gen_shl16_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
312
+{
313
+ uint64_t mask = dup_const(MO_16, 0xffff << shift);
314
+ TCGv_i64 t = tcg_temp_new_i64();
315
+
316
+ tcg_gen_shli_i64(t, a, shift);
317
+ tcg_gen_andi_i64(t, t, mask);
318
+ tcg_gen_andi_i64(d, d, ~mask);
319
+ tcg_gen_or_i64(d, d, t);
320
+ tcg_temp_free_i64(t);
321
+}
322
+
323
+static void gen_shl32_ins_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
324
+{
325
+ tcg_gen_deposit_i32(d, d, a, shift, 32 - shift);
326
+}
327
+
328
+static void gen_shl64_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
329
+{
330
+ tcg_gen_deposit_i64(d, d, a, shift, 64 - shift);
331
+}
332
+
333
+static void gen_shl_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
334
+{
335
+ if (sh == 0) {
336
+ tcg_gen_mov_vec(d, a);
337
+ } else {
338
+ TCGv_vec t = tcg_temp_new_vec_matching(d);
339
+ TCGv_vec m = tcg_temp_new_vec_matching(d);
340
+
341
+ tcg_gen_dupi_vec(vece, m, MAKE_64BIT_MASK(0, sh));
342
+ tcg_gen_shli_vec(vece, t, a, sh);
343
+ tcg_gen_and_vec(vece, d, d, m);
344
+ tcg_gen_or_vec(vece, d, d, t);
345
+
346
+ tcg_temp_free_vec(t);
347
+ tcg_temp_free_vec(m);
348
+ }
349
+}
350
+
351
+const GVecGen2i sli_op[4] = {
352
+ { .fni8 = gen_shl8_ins_i64,
353
+ .fniv = gen_shl_ins_vec,
354
+ .load_dest = true,
355
+ .opc = INDEX_op_shli_vec,
356
+ .vece = MO_8 },
357
+ { .fni8 = gen_shl16_ins_i64,
358
+ .fniv = gen_shl_ins_vec,
359
+ .load_dest = true,
360
+ .opc = INDEX_op_shli_vec,
361
+ .vece = MO_16 },
362
+ { .fni4 = gen_shl32_ins_i32,
363
+ .fniv = gen_shl_ins_vec,
364
+ .load_dest = true,
365
+ .opc = INDEX_op_shli_vec,
366
+ .vece = MO_32 },
367
+ { .fni8 = gen_shl64_ins_i64,
368
+ .fniv = gen_shl_ins_vec,
369
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
370
+ .load_dest = true,
371
+ .opc = INDEX_op_shli_vec,
372
+ .vece = MO_64 },
373
+};
374
+
375
/* Translate a NEON data processing instruction. Return nonzero if the
376
instruction is invalid.
377
We process data in a mixture of 32-bit and 64-bit chunks.
378
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
379
int pairwise;
380
int u;
381
int vec_size;
382
- uint32_t imm, mask;
383
+ uint32_t imm;
384
TCGv_i32 tmp, tmp2, tmp3, tmp4, tmp5;
385
TCGv_ptr ptr1, ptr2, ptr3;
386
TCGv_i64 tmp64;
387
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
388
}
389
return 0;
390
391
+ case 4: /* VSRI */
392
+ if (!u) {
393
+ return 1;
394
+ }
395
+ /* Right shift comes here negative. */
396
+ shift = -shift;
397
+ /* Shift out of range leaves destination unchanged. */
398
+ if (shift < 8 << size) {
399
+ tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size, vec_size,
400
+ shift, &sri_op[size]);
401
+ }
402
+ return 0;
403
+
404
case 5: /* VSHL, VSLI */
405
- if (!u) { /* VSHL */
406
+ if (u) { /* VSLI */
407
+ /* Shift out of range leaves destination unchanged. */
408
+ if (shift < 8 << size) {
409
+ tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size,
410
+ vec_size, shift, &sli_op[size]);
411
+ }
412
+ } else { /* VSHL */
413
/* Shifts larger than the element size are
414
* architecturally valid and results in zero.
415
*/
416
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
417
tcg_gen_gvec_shli(size, rd_ofs, rm_ofs, shift,
418
vec_size, vec_size);
419
}
420
- return 0;
421
}
422
- break;
423
+ return 0;
424
}
425
426
if (size == 3) {
427
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
428
else
429
gen_helper_neon_rshl_s64(cpu_V0, cpu_V0, cpu_V1);
430
break;
431
- case 4: /* VSRI */
432
- case 5: /* VSHL, VSLI */
433
- gen_helper_neon_shl_u64(cpu_V0, cpu_V0, cpu_V1);
434
- break;
435
case 6: /* VQSHLU */
436
gen_helper_neon_qshlu_s64(cpu_V0, cpu_env,
437
cpu_V0, cpu_V1);
438
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
439
/* Accumulate. */
440
neon_load_reg64(cpu_V1, rd + pass);
441
tcg_gen_add_i64(cpu_V0, cpu_V0, cpu_V1);
442
- } else if (op == 4 || (op == 5 && u)) {
443
- /* Insert */
444
- neon_load_reg64(cpu_V1, rd + pass);
445
- uint64_t mask;
446
- if (shift < -63 || shift > 63) {
447
- mask = 0;
448
- } else {
449
- if (op == 4) {
450
- mask = 0xffffffffffffffffull >> -shift;
451
- } else {
452
- mask = 0xffffffffffffffffull << shift;
453
- }
454
- }
455
- tcg_gen_andi_i64(cpu_V1, cpu_V1, ~mask);
456
- tcg_gen_or_i64(cpu_V0, cpu_V0, cpu_V1);
457
}
458
neon_store_reg64(cpu_V0, rd + pass);
459
} else { /* size < 3 */
460
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
461
case 3: /* VRSRA */
462
GEN_NEON_INTEGER_OP(rshl);
463
break;
464
- case 4: /* VSRI */
465
- case 5: /* VSHL, VSLI */
466
- switch (size) {
467
- case 0: gen_helper_neon_shl_u8(tmp, tmp, tmp2); break;
468
- case 1: gen_helper_neon_shl_u16(tmp, tmp, tmp2); break;
469
- case 2: gen_helper_neon_shl_u32(tmp, tmp, tmp2); break;
470
- default: abort();
471
- }
472
- break;
473
case 6: /* VQSHLU */
474
switch (size) {
475
case 0:
476
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
477
tmp2 = neon_load_reg(rd, pass);
478
gen_neon_add(size, tmp, tmp2);
479
tcg_temp_free_i32(tmp2);
480
- } else if (op == 4 || (op == 5 && u)) {
481
- /* Insert */
482
- switch (size) {
483
- case 0:
484
- if (op == 4)
485
- mask = 0xff >> -shift;
486
- else
487
- mask = (uint8_t)(0xff << shift);
488
- mask |= mask << 8;
489
- mask |= mask << 16;
490
- break;
491
- case 1:
492
- if (op == 4)
493
- mask = 0xffff >> -shift;
494
- else
495
- mask = (uint16_t)(0xffff << shift);
496
- mask |= mask << 16;
497
- break;
498
- case 2:
499
- if (shift < -31 || shift > 31) {
500
- mask = 0;
501
- } else {
502
- if (op == 4)
503
- mask = 0xffffffffu >> -shift;
504
- else
505
- mask = 0xffffffffu << shift;
506
- }
507
- break;
508
- default:
509
- abort();
510
- }
511
- tmp2 = neon_load_reg(rd, pass);
512
- tcg_gen_andi_i32(tmp, tmp, mask);
513
- tcg_gen_andi_i32(tmp2, tmp2, ~mask);
514
- tcg_gen_or_i32(tmp, tmp, tmp2);
515
- tcg_temp_free_i32(tmp2);
516
}
517
neon_store_reg(rd, pass, tmp);
518
}
519
--
130
--
520
2.19.1
131
2.20.1
521
132
522
133
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-4-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-13-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
7
---
8
target/arm/translate-a64.c | 28 +++-------------------------
8
target/arm/translate-a64.c | 15 +++++++++++++++
9
1 file changed, 3 insertions(+), 25 deletions(-)
9
1 file changed, 15 insertions(+)
10
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
15
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
16
for (xs = 0; xs < selem; xs++) {
16
cpu_reg_sp(s, rn));
17
if (replicate) {
17
}
18
/* Load and replicate to all elements */
18
break;
19
- uint64_t mulconst;
19
+ case 5: /* GMI */
20
TCGv_i64 tcg_tmp = tcg_temp_new_i64();
20
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
21
21
+ goto do_unallocated;
22
tcg_gen_qemu_ld_i64(tcg_tmp, tcg_addr,
22
+ } else {
23
get_mem_index(s), s->be_data + scale);
23
+ TCGv_i64 t1 = tcg_const_i64(1);
24
- switch (scale) {
24
+ TCGv_i64 t2 = tcg_temp_new_i64();
25
- case 0:
25
+
26
- mulconst = 0x0101010101010101ULL;
26
+ tcg_gen_extract_i64(t2, cpu_reg_sp(s, rn), 56, 4);
27
- break;
27
+ tcg_gen_shl_i64(t1, t1, t2);
28
- case 1:
28
+ tcg_gen_or_i64(cpu_reg(s, rd), cpu_reg(s, rm), t1);
29
- mulconst = 0x0001000100010001ULL;
29
+
30
- break;
30
+ tcg_temp_free_i64(t1);
31
- case 2:
31
+ tcg_temp_free_i64(t2);
32
- mulconst = 0x0000000100000001ULL;
32
+ }
33
- break;
33
+ break;
34
- case 3:
34
case 8: /* LSLV */
35
- mulconst = 0;
35
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
36
- break;
36
break;
37
- default:
38
- g_assert_not_reached();
39
- }
40
- if (mulconst) {
41
- tcg_gen_muli_i64(tcg_tmp, tcg_tmp, mulconst);
42
- }
43
- write_vec_element(s, tcg_tmp, rt, 0, MO_64);
44
- if (is_q) {
45
- write_vec_element(s, tcg_tmp, rt, 1, MO_64);
46
- }
47
+ tcg_gen_gvec_dup_i64(scale, vec_full_reg_offset(s, rt),
48
+ (is_q + 1) * 8, vec_full_reg_size(s),
49
+ tcg_tmp);
50
tcg_temp_free_i64(tcg_tmp);
51
- clear_vec_high(s, is_q, rt);
52
} else {
53
/* Load/store one element per register */
54
if (is_load) {
55
--
37
--
56
2.19.1
38
2.20.1
57
39
58
40
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is done generically in translator_loop.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
5
Reported-by: Laurent Desnogues <laurent.desnogues@gmail.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
5
Message-id: 20200626033144.790098-14-richard.henderson@linaro.org
8
Message-id: 20181011205206.3552-3-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
7
---
12
target/arm/translate-a64.c | 1 -
8
target/arm/translate-a64.c | 24 ++++++++++++++++++++++--
13
target/arm/translate.c | 1 -
9
1 file changed, 22 insertions(+), 2 deletions(-)
14
2 files changed, 2 deletions(-)
15
10
16
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
17
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate-a64.c
13
--- a/target/arm/translate-a64.c
19
+++ b/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
20
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
15
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
21
16
*/
22
static void aarch64_tr_tb_start(DisasContextBase *db, CPUState *cpu)
17
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
23
{
18
{
24
- tcg_clear_temp_count();
19
- unsigned int sf, rm, opcode, rn, rd;
25
}
20
+ unsigned int sf, rm, opcode, rn, rd, setflag;
26
21
sf = extract32(insn, 31, 1);
27
static void aarch64_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
22
+ setflag = extract32(insn, 29, 1);
28
diff --git a/target/arm/translate.c b/target/arm/translate.c
23
rm = extract32(insn, 16, 5);
29
index XXXXXXX..XXXXXXX 100644
24
opcode = extract32(insn, 10, 6);
30
--- a/target/arm/translate.c
25
rn = extract32(insn, 5, 5);
31
+++ b/target/arm/translate.c
26
rd = extract32(insn, 0, 5);
32
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_start(DisasContextBase *dcbase, CPUState *cpu)
27
33
tcg_gen_movi_i32(tmp, 0);
28
- if (extract32(insn, 29, 1)) {
34
store_cpu_field(tmp, condexec_bits);
29
+ if (setflag && opcode != 0) {
30
unallocated_encoding(s);
31
return;
35
}
32
}
36
- tcg_clear_temp_count();
33
37
}
34
switch (opcode) {
38
35
+ case 0: /* SUBP(S) */
39
static void arm_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
36
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
37
+ goto do_unallocated;
38
+ } else {
39
+ TCGv_i64 tcg_n, tcg_m, tcg_d;
40
+
41
+ tcg_n = read_cpu_reg_sp(s, rn, true);
42
+ tcg_m = read_cpu_reg_sp(s, rm, true);
43
+ tcg_gen_sextract_i64(tcg_n, tcg_n, 0, 56);
44
+ tcg_gen_sextract_i64(tcg_m, tcg_m, 0, 56);
45
+ tcg_d = cpu_reg(s, rd);
46
+
47
+ if (setflag) {
48
+ gen_sub_CC(true, tcg_d, tcg_n, tcg_m);
49
+ } else {
50
+ tcg_gen_sub_i64(tcg_d, tcg_n, tcg_m);
51
+ }
52
+ }
53
+ break;
54
case 2: /* UDIV */
55
handle_div(s, false, sf, rm, rn, rd);
56
break;
40
--
57
--
41
2.19.1
58
2.20.1
42
59
43
60
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Create struct ARMISARegisters, to be accessed during translation.
3
Use the same code as system mode, so that we generate the same
4
exception + syndrome for the unaligned access.
4
5
6
For the moment, if MTE is enabled so that this path is reachable,
7
this would generate a SIGSEGV in the user-only cpu_loop. Decoding
8
the syndrome to produce the proper SIGBUS will be done later.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181016223115.24100-2-richard.henderson@linaro.org
12
Message-id: 20200626033144.790098-15-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
14
---
10
target/arm/cpu.h | 32 ++++----
15
target/arm/cpu.c | 2 +-
11
hw/intc/armv7m_nvic.c | 12 +--
16
target/arm/tlb_helper.c | 4 ++--
12
target/arm/cpu.c | 178 +++++++++++++++++++++---------------------
17
2 files changed, 3 insertions(+), 3 deletions(-)
13
target/arm/cpu64.c | 70 ++++++++---------
14
target/arm/helper.c | 28 +++----
15
5 files changed, 162 insertions(+), 158 deletions(-)
16
18
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
22
* ARMv7AR ARM Architecture Reference Manual. A reset_ prefix
23
* is used for reset values of non-constant registers; no reset_
24
* prefix means a constant register.
25
+ * Some of these registers are split out into a substructure that
26
+ * is shared with the translators to control the ISA.
27
*/
28
+ struct ARMISARegisters {
29
+ uint32_t id_isar0;
30
+ uint32_t id_isar1;
31
+ uint32_t id_isar2;
32
+ uint32_t id_isar3;
33
+ uint32_t id_isar4;
34
+ uint32_t id_isar5;
35
+ uint32_t id_isar6;
36
+ uint32_t mvfr0;
37
+ uint32_t mvfr1;
38
+ uint32_t mvfr2;
39
+ uint64_t id_aa64isar0;
40
+ uint64_t id_aa64isar1;
41
+ uint64_t id_aa64pfr0;
42
+ uint64_t id_aa64pfr1;
43
+ } isar;
44
uint32_t midr;
45
uint32_t revidr;
46
uint32_t reset_fpsid;
47
- uint32_t mvfr0;
48
- uint32_t mvfr1;
49
- uint32_t mvfr2;
50
uint32_t ctr;
51
uint32_t reset_sctlr;
52
uint32_t id_pfr0;
53
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
54
uint32_t id_mmfr2;
55
uint32_t id_mmfr3;
56
uint32_t id_mmfr4;
57
- uint32_t id_isar0;
58
- uint32_t id_isar1;
59
- uint32_t id_isar2;
60
- uint32_t id_isar3;
61
- uint32_t id_isar4;
62
- uint32_t id_isar5;
63
- uint32_t id_isar6;
64
- uint64_t id_aa64pfr0;
65
- uint64_t id_aa64pfr1;
66
uint64_t id_aa64dfr0;
67
uint64_t id_aa64dfr1;
68
uint64_t id_aa64afr0;
69
uint64_t id_aa64afr1;
70
- uint64_t id_aa64isar0;
71
- uint64_t id_aa64isar1;
72
uint64_t id_aa64mmfr0;
73
uint64_t id_aa64mmfr1;
74
uint32_t dbgdidr;
75
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
76
index XXXXXXX..XXXXXXX 100644
77
--- a/hw/intc/armv7m_nvic.c
78
+++ b/hw/intc/armv7m_nvic.c
79
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
80
case 0xd5c: /* MMFR3. */
81
return cpu->id_mmfr3;
82
case 0xd60: /* ISAR0. */
83
- return cpu->id_isar0;
84
+ return cpu->isar.id_isar0;
85
case 0xd64: /* ISAR1. */
86
- return cpu->id_isar1;
87
+ return cpu->isar.id_isar1;
88
case 0xd68: /* ISAR2. */
89
- return cpu->id_isar2;
90
+ return cpu->isar.id_isar2;
91
case 0xd6c: /* ISAR3. */
92
- return cpu->id_isar3;
93
+ return cpu->isar.id_isar3;
94
case 0xd70: /* ISAR4. */
95
- return cpu->id_isar4;
96
+ return cpu->isar.id_isar4;
97
case 0xd74: /* ISAR5. */
98
- return cpu->id_isar5;
99
+ return cpu->isar.id_isar5;
100
case 0xd78: /* CLIDR */
101
return cpu->clidr;
102
case 0xd7c: /* CTR */
103
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
19
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
104
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/cpu.c
21
--- a/target/arm/cpu.c
106
+++ b/target/arm/cpu.c
22
+++ b/target/arm/cpu.c
107
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
23
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
108
g_hash_table_foreach(cpu->cp_regs, cp_reg_check_reset, cpu);
24
cc->tlb_fill = arm_cpu_tlb_fill;
109
25
cc->debug_excp_handler = arm_debug_excp_handler;
110
env->vfp.xregs[ARM_VFP_FPSID] = cpu->reset_fpsid;
26
cc->debug_check_watchpoint = arm_debug_check_watchpoint;
111
- env->vfp.xregs[ARM_VFP_MVFR0] = cpu->mvfr0;
27
-#if !defined(CONFIG_USER_ONLY)
112
- env->vfp.xregs[ARM_VFP_MVFR1] = cpu->mvfr1;
28
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
113
- env->vfp.xregs[ARM_VFP_MVFR2] = cpu->mvfr2;
29
+#if !defined(CONFIG_USER_ONLY)
114
+ env->vfp.xregs[ARM_VFP_MVFR0] = cpu->isar.mvfr0;
30
cc->do_transaction_failed = arm_cpu_do_transaction_failed;
115
+ env->vfp.xregs[ARM_VFP_MVFR1] = cpu->isar.mvfr1;
31
cc->adjust_watchpoint_address = arm_adjust_watchpoint_address;
116
+ env->vfp.xregs[ARM_VFP_MVFR2] = cpu->isar.mvfr2;
32
#endif /* CONFIG_TCG && !CONFIG_USER_ONLY */
117
33
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
118
cpu->power_state = cpu->start_powered_off ? PSCI_OFF : PSCI_ON;
34
index XXXXXXX..XXXXXXX 100644
119
s->halted = cpu->start_powered_off;
35
--- a/target/arm/tlb_helper.c
120
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
36
+++ b/target/arm/tlb_helper.c
121
* registers as well. These are id_pfr1[7:4] and id_aa64pfr0[15:12].
37
@@ -XXX,XX +XXX,XX @@
122
*/
38
#include "internals.h"
123
cpu->id_pfr1 &= ~0xf0;
39
#include "exec/exec-all.h"
124
- cpu->id_aa64pfr0 &= ~0xf000;
40
125
+ cpu->isar.id_aa64pfr0 &= ~0xf000;
41
-#if !defined(CONFIG_USER_ONLY)
126
}
42
-
127
43
static inline uint32_t merge_syn_data_abort(uint32_t template_syn,
128
if (!cpu->has_el2) {
44
unsigned int target_el,
129
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
45
bool same_el, bool ea,
130
* registers if we don't have EL2. These are id_pfr1[15:12] and
46
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
131
* id_aa64pfr0_el1[11:8].
47
arm_deliver_fault(cpu, vaddr, access_type, mmu_idx, &fi);
132
*/
133
- cpu->id_aa64pfr0 &= ~0xf00;
134
+ cpu->isar.id_aa64pfr0 &= ~0xf00;
135
cpu->id_pfr1 &= ~0xf000;
136
}
137
138
@@ -XXX,XX +XXX,XX @@ static void arm1136_r2_initfn(Object *obj)
139
set_feature(&cpu->env, ARM_FEATURE_CACHE_BLOCK_OPS);
140
cpu->midr = 0x4107b362;
141
cpu->reset_fpsid = 0x410120b4;
142
- cpu->mvfr0 = 0x11111111;
143
- cpu->mvfr1 = 0x00000000;
144
+ cpu->isar.mvfr0 = 0x11111111;
145
+ cpu->isar.mvfr1 = 0x00000000;
146
cpu->ctr = 0x1dd20d2;
147
cpu->reset_sctlr = 0x00050078;
148
cpu->id_pfr0 = 0x111;
149
@@ -XXX,XX +XXX,XX @@ static void arm1136_r2_initfn(Object *obj)
150
cpu->id_mmfr0 = 0x01130003;
151
cpu->id_mmfr1 = 0x10030302;
152
cpu->id_mmfr2 = 0x01222110;
153
- cpu->id_isar0 = 0x00140011;
154
- cpu->id_isar1 = 0x12002111;
155
- cpu->id_isar2 = 0x11231111;
156
- cpu->id_isar3 = 0x01102131;
157
- cpu->id_isar4 = 0x141;
158
+ cpu->isar.id_isar0 = 0x00140011;
159
+ cpu->isar.id_isar1 = 0x12002111;
160
+ cpu->isar.id_isar2 = 0x11231111;
161
+ cpu->isar.id_isar3 = 0x01102131;
162
+ cpu->isar.id_isar4 = 0x141;
163
cpu->reset_auxcr = 7;
164
}
48
}
165
49
166
@@ -XXX,XX +XXX,XX @@ static void arm1136_initfn(Object *obj)
50
+#if !defined(CONFIG_USER_ONLY)
167
set_feature(&cpu->env, ARM_FEATURE_CACHE_BLOCK_OPS);
51
+
168
cpu->midr = 0x4117b363;
52
/*
169
cpu->reset_fpsid = 0x410120b4;
53
* arm_cpu_do_transaction_failed: handle a memory system error response
170
- cpu->mvfr0 = 0x11111111;
54
* (eg "no device/memory present at address") by raising an external abort
171
- cpu->mvfr1 = 0x00000000;
172
+ cpu->isar.mvfr0 = 0x11111111;
173
+ cpu->isar.mvfr1 = 0x00000000;
174
cpu->ctr = 0x1dd20d2;
175
cpu->reset_sctlr = 0x00050078;
176
cpu->id_pfr0 = 0x111;
177
@@ -XXX,XX +XXX,XX @@ static void arm1136_initfn(Object *obj)
178
cpu->id_mmfr0 = 0x01130003;
179
cpu->id_mmfr1 = 0x10030302;
180
cpu->id_mmfr2 = 0x01222110;
181
- cpu->id_isar0 = 0x00140011;
182
- cpu->id_isar1 = 0x12002111;
183
- cpu->id_isar2 = 0x11231111;
184
- cpu->id_isar3 = 0x01102131;
185
- cpu->id_isar4 = 0x141;
186
+ cpu->isar.id_isar0 = 0x00140011;
187
+ cpu->isar.id_isar1 = 0x12002111;
188
+ cpu->isar.id_isar2 = 0x11231111;
189
+ cpu->isar.id_isar3 = 0x01102131;
190
+ cpu->isar.id_isar4 = 0x141;
191
cpu->reset_auxcr = 7;
192
}
193
194
@@ -XXX,XX +XXX,XX @@ static void arm1176_initfn(Object *obj)
195
set_feature(&cpu->env, ARM_FEATURE_EL3);
196
cpu->midr = 0x410fb767;
197
cpu->reset_fpsid = 0x410120b5;
198
- cpu->mvfr0 = 0x11111111;
199
- cpu->mvfr1 = 0x00000000;
200
+ cpu->isar.mvfr0 = 0x11111111;
201
+ cpu->isar.mvfr1 = 0x00000000;
202
cpu->ctr = 0x1dd20d2;
203
cpu->reset_sctlr = 0x00050078;
204
cpu->id_pfr0 = 0x111;
205
@@ -XXX,XX +XXX,XX @@ static void arm1176_initfn(Object *obj)
206
cpu->id_mmfr0 = 0x01130003;
207
cpu->id_mmfr1 = 0x10030302;
208
cpu->id_mmfr2 = 0x01222100;
209
- cpu->id_isar0 = 0x0140011;
210
- cpu->id_isar1 = 0x12002111;
211
- cpu->id_isar2 = 0x11231121;
212
- cpu->id_isar3 = 0x01102131;
213
- cpu->id_isar4 = 0x01141;
214
+ cpu->isar.id_isar0 = 0x0140011;
215
+ cpu->isar.id_isar1 = 0x12002111;
216
+ cpu->isar.id_isar2 = 0x11231121;
217
+ cpu->isar.id_isar3 = 0x01102131;
218
+ cpu->isar.id_isar4 = 0x01141;
219
cpu->reset_auxcr = 7;
220
}
221
222
@@ -XXX,XX +XXX,XX @@ static void arm11mpcore_initfn(Object *obj)
223
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
224
cpu->midr = 0x410fb022;
225
cpu->reset_fpsid = 0x410120b4;
226
- cpu->mvfr0 = 0x11111111;
227
- cpu->mvfr1 = 0x00000000;
228
+ cpu->isar.mvfr0 = 0x11111111;
229
+ cpu->isar.mvfr1 = 0x00000000;
230
cpu->ctr = 0x1d192992; /* 32K icache 32K dcache */
231
cpu->id_pfr0 = 0x111;
232
cpu->id_pfr1 = 0x1;
233
@@ -XXX,XX +XXX,XX @@ static void arm11mpcore_initfn(Object *obj)
234
cpu->id_mmfr0 = 0x01100103;
235
cpu->id_mmfr1 = 0x10020302;
236
cpu->id_mmfr2 = 0x01222000;
237
- cpu->id_isar0 = 0x00100011;
238
- cpu->id_isar1 = 0x12002111;
239
- cpu->id_isar2 = 0x11221011;
240
- cpu->id_isar3 = 0x01102131;
241
- cpu->id_isar4 = 0x141;
242
+ cpu->isar.id_isar0 = 0x00100011;
243
+ cpu->isar.id_isar1 = 0x12002111;
244
+ cpu->isar.id_isar2 = 0x11221011;
245
+ cpu->isar.id_isar3 = 0x01102131;
246
+ cpu->isar.id_isar4 = 0x141;
247
cpu->reset_auxcr = 1;
248
}
249
250
@@ -XXX,XX +XXX,XX @@ static void cortex_m3_initfn(Object *obj)
251
cpu->id_mmfr1 = 0x00000000;
252
cpu->id_mmfr2 = 0x00000000;
253
cpu->id_mmfr3 = 0x00000000;
254
- cpu->id_isar0 = 0x01141110;
255
- cpu->id_isar1 = 0x02111000;
256
- cpu->id_isar2 = 0x21112231;
257
- cpu->id_isar3 = 0x01111110;
258
- cpu->id_isar4 = 0x01310102;
259
- cpu->id_isar5 = 0x00000000;
260
- cpu->id_isar6 = 0x00000000;
261
+ cpu->isar.id_isar0 = 0x01141110;
262
+ cpu->isar.id_isar1 = 0x02111000;
263
+ cpu->isar.id_isar2 = 0x21112231;
264
+ cpu->isar.id_isar3 = 0x01111110;
265
+ cpu->isar.id_isar4 = 0x01310102;
266
+ cpu->isar.id_isar5 = 0x00000000;
267
+ cpu->isar.id_isar6 = 0x00000000;
268
}
269
270
static void cortex_m4_initfn(Object *obj)
271
@@ -XXX,XX +XXX,XX @@ static void cortex_m4_initfn(Object *obj)
272
cpu->id_mmfr1 = 0x00000000;
273
cpu->id_mmfr2 = 0x00000000;
274
cpu->id_mmfr3 = 0x00000000;
275
- cpu->id_isar0 = 0x01141110;
276
- cpu->id_isar1 = 0x02111000;
277
- cpu->id_isar2 = 0x21112231;
278
- cpu->id_isar3 = 0x01111110;
279
- cpu->id_isar4 = 0x01310102;
280
- cpu->id_isar5 = 0x00000000;
281
- cpu->id_isar6 = 0x00000000;
282
+ cpu->isar.id_isar0 = 0x01141110;
283
+ cpu->isar.id_isar1 = 0x02111000;
284
+ cpu->isar.id_isar2 = 0x21112231;
285
+ cpu->isar.id_isar3 = 0x01111110;
286
+ cpu->isar.id_isar4 = 0x01310102;
287
+ cpu->isar.id_isar5 = 0x00000000;
288
+ cpu->isar.id_isar6 = 0x00000000;
289
}
290
291
static void cortex_m33_initfn(Object *obj)
292
@@ -XXX,XX +XXX,XX @@ static void cortex_m33_initfn(Object *obj)
293
cpu->id_mmfr1 = 0x00000000;
294
cpu->id_mmfr2 = 0x01000000;
295
cpu->id_mmfr3 = 0x00000000;
296
- cpu->id_isar0 = 0x01101110;
297
- cpu->id_isar1 = 0x02212000;
298
- cpu->id_isar2 = 0x20232232;
299
- cpu->id_isar3 = 0x01111131;
300
- cpu->id_isar4 = 0x01310132;
301
- cpu->id_isar5 = 0x00000000;
302
- cpu->id_isar6 = 0x00000000;
303
+ cpu->isar.id_isar0 = 0x01101110;
304
+ cpu->isar.id_isar1 = 0x02212000;
305
+ cpu->isar.id_isar2 = 0x20232232;
306
+ cpu->isar.id_isar3 = 0x01111131;
307
+ cpu->isar.id_isar4 = 0x01310132;
308
+ cpu->isar.id_isar5 = 0x00000000;
309
+ cpu->isar.id_isar6 = 0x00000000;
310
cpu->clidr = 0x00000000;
311
cpu->ctr = 0x8000c000;
312
}
313
@@ -XXX,XX +XXX,XX @@ static void cortex_r5_initfn(Object *obj)
314
cpu->id_mmfr1 = 0x00000000;
315
cpu->id_mmfr2 = 0x01200000;
316
cpu->id_mmfr3 = 0x0211;
317
- cpu->id_isar0 = 0x02101111;
318
- cpu->id_isar1 = 0x13112111;
319
- cpu->id_isar2 = 0x21232141;
320
- cpu->id_isar3 = 0x01112131;
321
- cpu->id_isar4 = 0x0010142;
322
- cpu->id_isar5 = 0x0;
323
- cpu->id_isar6 = 0x0;
324
+ cpu->isar.id_isar0 = 0x02101111;
325
+ cpu->isar.id_isar1 = 0x13112111;
326
+ cpu->isar.id_isar2 = 0x21232141;
327
+ cpu->isar.id_isar3 = 0x01112131;
328
+ cpu->isar.id_isar4 = 0x0010142;
329
+ cpu->isar.id_isar5 = 0x0;
330
+ cpu->isar.id_isar6 = 0x0;
331
cpu->mp_is_up = true;
332
cpu->pmsav7_dregion = 16;
333
define_arm_cp_regs(cpu, cortexr5_cp_reginfo);
334
@@ -XXX,XX +XXX,XX @@ static void cortex_a8_initfn(Object *obj)
335
set_feature(&cpu->env, ARM_FEATURE_EL3);
336
cpu->midr = 0x410fc080;
337
cpu->reset_fpsid = 0x410330c0;
338
- cpu->mvfr0 = 0x11110222;
339
- cpu->mvfr1 = 0x00011111;
340
+ cpu->isar.mvfr0 = 0x11110222;
341
+ cpu->isar.mvfr1 = 0x00011111;
342
cpu->ctr = 0x82048004;
343
cpu->reset_sctlr = 0x00c50078;
344
cpu->id_pfr0 = 0x1031;
345
@@ -XXX,XX +XXX,XX @@ static void cortex_a8_initfn(Object *obj)
346
cpu->id_mmfr1 = 0x20000000;
347
cpu->id_mmfr2 = 0x01202000;
348
cpu->id_mmfr3 = 0x11;
349
- cpu->id_isar0 = 0x00101111;
350
- cpu->id_isar1 = 0x12112111;
351
- cpu->id_isar2 = 0x21232031;
352
- cpu->id_isar3 = 0x11112131;
353
- cpu->id_isar4 = 0x00111142;
354
+ cpu->isar.id_isar0 = 0x00101111;
355
+ cpu->isar.id_isar1 = 0x12112111;
356
+ cpu->isar.id_isar2 = 0x21232031;
357
+ cpu->isar.id_isar3 = 0x11112131;
358
+ cpu->isar.id_isar4 = 0x00111142;
359
cpu->dbgdidr = 0x15141000;
360
cpu->clidr = (1 << 27) | (2 << 24) | 3;
361
cpu->ccsidr[0] = 0xe007e01a; /* 16k L1 dcache. */
362
@@ -XXX,XX +XXX,XX @@ static void cortex_a9_initfn(Object *obj)
363
set_feature(&cpu->env, ARM_FEATURE_CBAR);
364
cpu->midr = 0x410fc090;
365
cpu->reset_fpsid = 0x41033090;
366
- cpu->mvfr0 = 0x11110222;
367
- cpu->mvfr1 = 0x01111111;
368
+ cpu->isar.mvfr0 = 0x11110222;
369
+ cpu->isar.mvfr1 = 0x01111111;
370
cpu->ctr = 0x80038003;
371
cpu->reset_sctlr = 0x00c50078;
372
cpu->id_pfr0 = 0x1031;
373
@@ -XXX,XX +XXX,XX @@ static void cortex_a9_initfn(Object *obj)
374
cpu->id_mmfr1 = 0x20000000;
375
cpu->id_mmfr2 = 0x01230000;
376
cpu->id_mmfr3 = 0x00002111;
377
- cpu->id_isar0 = 0x00101111;
378
- cpu->id_isar1 = 0x13112111;
379
- cpu->id_isar2 = 0x21232041;
380
- cpu->id_isar3 = 0x11112131;
381
- cpu->id_isar4 = 0x00111142;
382
+ cpu->isar.id_isar0 = 0x00101111;
383
+ cpu->isar.id_isar1 = 0x13112111;
384
+ cpu->isar.id_isar2 = 0x21232041;
385
+ cpu->isar.id_isar3 = 0x11112131;
386
+ cpu->isar.id_isar4 = 0x00111142;
387
cpu->dbgdidr = 0x35141000;
388
cpu->clidr = (1 << 27) | (1 << 24) | 3;
389
cpu->ccsidr[0] = 0xe00fe019; /* 16k L1 dcache. */
390
@@ -XXX,XX +XXX,XX @@ static void cortex_a7_initfn(Object *obj)
391
cpu->kvm_target = QEMU_KVM_ARM_TARGET_CORTEX_A7;
392
cpu->midr = 0x410fc075;
393
cpu->reset_fpsid = 0x41023075;
394
- cpu->mvfr0 = 0x10110222;
395
- cpu->mvfr1 = 0x11111111;
396
+ cpu->isar.mvfr0 = 0x10110222;
397
+ cpu->isar.mvfr1 = 0x11111111;
398
cpu->ctr = 0x84448003;
399
cpu->reset_sctlr = 0x00c50078;
400
cpu->id_pfr0 = 0x00001131;
401
@@ -XXX,XX +XXX,XX @@ static void cortex_a7_initfn(Object *obj)
402
/* a7_mpcore_r0p5_trm, page 4-4 gives 0x01101110; but
403
* table 4-41 gives 0x02101110, which includes the arm div insns.
404
*/
405
- cpu->id_isar0 = 0x02101110;
406
- cpu->id_isar1 = 0x13112111;
407
- cpu->id_isar2 = 0x21232041;
408
- cpu->id_isar3 = 0x11112131;
409
- cpu->id_isar4 = 0x10011142;
410
+ cpu->isar.id_isar0 = 0x02101110;
411
+ cpu->isar.id_isar1 = 0x13112111;
412
+ cpu->isar.id_isar2 = 0x21232041;
413
+ cpu->isar.id_isar3 = 0x11112131;
414
+ cpu->isar.id_isar4 = 0x10011142;
415
cpu->dbgdidr = 0x3515f005;
416
cpu->clidr = 0x0a200023;
417
cpu->ccsidr[0] = 0x701fe00a; /* 32K L1 dcache */
418
@@ -XXX,XX +XXX,XX @@ static void cortex_a15_initfn(Object *obj)
419
cpu->kvm_target = QEMU_KVM_ARM_TARGET_CORTEX_A15;
420
cpu->midr = 0x412fc0f1;
421
cpu->reset_fpsid = 0x410430f0;
422
- cpu->mvfr0 = 0x10110222;
423
- cpu->mvfr1 = 0x11111111;
424
+ cpu->isar.mvfr0 = 0x10110222;
425
+ cpu->isar.mvfr1 = 0x11111111;
426
cpu->ctr = 0x8444c004;
427
cpu->reset_sctlr = 0x00c50078;
428
cpu->id_pfr0 = 0x00001131;
429
@@ -XXX,XX +XXX,XX @@ static void cortex_a15_initfn(Object *obj)
430
cpu->id_mmfr1 = 0x20000000;
431
cpu->id_mmfr2 = 0x01240000;
432
cpu->id_mmfr3 = 0x02102211;
433
- cpu->id_isar0 = 0x02101110;
434
- cpu->id_isar1 = 0x13112111;
435
- cpu->id_isar2 = 0x21232041;
436
- cpu->id_isar3 = 0x11112131;
437
- cpu->id_isar4 = 0x10011142;
438
+ cpu->isar.id_isar0 = 0x02101110;
439
+ cpu->isar.id_isar1 = 0x13112111;
440
+ cpu->isar.id_isar2 = 0x21232041;
441
+ cpu->isar.id_isar3 = 0x11112131;
442
+ cpu->isar.id_isar4 = 0x10011142;
443
cpu->dbgdidr = 0x3515f021;
444
cpu->clidr = 0x0a200023;
445
cpu->ccsidr[0] = 0x701fe00a; /* 32K L1 dcache */
446
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
447
index XXXXXXX..XXXXXXX 100644
448
--- a/target/arm/cpu64.c
449
+++ b/target/arm/cpu64.c
450
@@ -XXX,XX +XXX,XX @@ static void aarch64_a57_initfn(Object *obj)
451
cpu->midr = 0x411fd070;
452
cpu->revidr = 0x00000000;
453
cpu->reset_fpsid = 0x41034070;
454
- cpu->mvfr0 = 0x10110222;
455
- cpu->mvfr1 = 0x12111111;
456
- cpu->mvfr2 = 0x00000043;
457
+ cpu->isar.mvfr0 = 0x10110222;
458
+ cpu->isar.mvfr1 = 0x12111111;
459
+ cpu->isar.mvfr2 = 0x00000043;
460
cpu->ctr = 0x8444c004;
461
cpu->reset_sctlr = 0x00c50838;
462
cpu->id_pfr0 = 0x00000131;
463
@@ -XXX,XX +XXX,XX @@ static void aarch64_a57_initfn(Object *obj)
464
cpu->id_mmfr1 = 0x40000000;
465
cpu->id_mmfr2 = 0x01260000;
466
cpu->id_mmfr3 = 0x02102211;
467
- cpu->id_isar0 = 0x02101110;
468
- cpu->id_isar1 = 0x13112111;
469
- cpu->id_isar2 = 0x21232042;
470
- cpu->id_isar3 = 0x01112131;
471
- cpu->id_isar4 = 0x00011142;
472
- cpu->id_isar5 = 0x00011121;
473
- cpu->id_isar6 = 0;
474
- cpu->id_aa64pfr0 = 0x00002222;
475
+ cpu->isar.id_isar0 = 0x02101110;
476
+ cpu->isar.id_isar1 = 0x13112111;
477
+ cpu->isar.id_isar2 = 0x21232042;
478
+ cpu->isar.id_isar3 = 0x01112131;
479
+ cpu->isar.id_isar4 = 0x00011142;
480
+ cpu->isar.id_isar5 = 0x00011121;
481
+ cpu->isar.id_isar6 = 0;
482
+ cpu->isar.id_aa64pfr0 = 0x00002222;
483
cpu->id_aa64dfr0 = 0x10305106;
484
cpu->pmceid0 = 0x00000000;
485
cpu->pmceid1 = 0x00000000;
486
- cpu->id_aa64isar0 = 0x00011120;
487
+ cpu->isar.id_aa64isar0 = 0x00011120;
488
cpu->id_aa64mmfr0 = 0x00001124;
489
cpu->dbgdidr = 0x3516d000;
490
cpu->clidr = 0x0a200023;
491
@@ -XXX,XX +XXX,XX @@ static void aarch64_a53_initfn(Object *obj)
492
cpu->midr = 0x410fd034;
493
cpu->revidr = 0x00000000;
494
cpu->reset_fpsid = 0x41034070;
495
- cpu->mvfr0 = 0x10110222;
496
- cpu->mvfr1 = 0x12111111;
497
- cpu->mvfr2 = 0x00000043;
498
+ cpu->isar.mvfr0 = 0x10110222;
499
+ cpu->isar.mvfr1 = 0x12111111;
500
+ cpu->isar.mvfr2 = 0x00000043;
501
cpu->ctr = 0x84448004; /* L1Ip = VIPT */
502
cpu->reset_sctlr = 0x00c50838;
503
cpu->id_pfr0 = 0x00000131;
504
@@ -XXX,XX +XXX,XX @@ static void aarch64_a53_initfn(Object *obj)
505
cpu->id_mmfr1 = 0x40000000;
506
cpu->id_mmfr2 = 0x01260000;
507
cpu->id_mmfr3 = 0x02102211;
508
- cpu->id_isar0 = 0x02101110;
509
- cpu->id_isar1 = 0x13112111;
510
- cpu->id_isar2 = 0x21232042;
511
- cpu->id_isar3 = 0x01112131;
512
- cpu->id_isar4 = 0x00011142;
513
- cpu->id_isar5 = 0x00011121;
514
- cpu->id_isar6 = 0;
515
- cpu->id_aa64pfr0 = 0x00002222;
516
+ cpu->isar.id_isar0 = 0x02101110;
517
+ cpu->isar.id_isar1 = 0x13112111;
518
+ cpu->isar.id_isar2 = 0x21232042;
519
+ cpu->isar.id_isar3 = 0x01112131;
520
+ cpu->isar.id_isar4 = 0x00011142;
521
+ cpu->isar.id_isar5 = 0x00011121;
522
+ cpu->isar.id_isar6 = 0;
523
+ cpu->isar.id_aa64pfr0 = 0x00002222;
524
cpu->id_aa64dfr0 = 0x10305106;
525
- cpu->id_aa64isar0 = 0x00011120;
526
+ cpu->isar.id_aa64isar0 = 0x00011120;
527
cpu->id_aa64mmfr0 = 0x00001122; /* 40 bit physical addr */
528
cpu->dbgdidr = 0x3516d000;
529
cpu->clidr = 0x0a200023;
530
@@ -XXX,XX +XXX,XX @@ static void aarch64_a72_initfn(Object *obj)
531
cpu->midr = 0x410fd083;
532
cpu->revidr = 0x00000000;
533
cpu->reset_fpsid = 0x41034080;
534
- cpu->mvfr0 = 0x10110222;
535
- cpu->mvfr1 = 0x12111111;
536
- cpu->mvfr2 = 0x00000043;
537
+ cpu->isar.mvfr0 = 0x10110222;
538
+ cpu->isar.mvfr1 = 0x12111111;
539
+ cpu->isar.mvfr2 = 0x00000043;
540
cpu->ctr = 0x8444c004;
541
cpu->reset_sctlr = 0x00c50838;
542
cpu->id_pfr0 = 0x00000131;
543
@@ -XXX,XX +XXX,XX @@ static void aarch64_a72_initfn(Object *obj)
544
cpu->id_mmfr1 = 0x40000000;
545
cpu->id_mmfr2 = 0x01260000;
546
cpu->id_mmfr3 = 0x02102211;
547
- cpu->id_isar0 = 0x02101110;
548
- cpu->id_isar1 = 0x13112111;
549
- cpu->id_isar2 = 0x21232042;
550
- cpu->id_isar3 = 0x01112131;
551
- cpu->id_isar4 = 0x00011142;
552
- cpu->id_isar5 = 0x00011121;
553
- cpu->id_aa64pfr0 = 0x00002222;
554
+ cpu->isar.id_isar0 = 0x02101110;
555
+ cpu->isar.id_isar1 = 0x13112111;
556
+ cpu->isar.id_isar2 = 0x21232042;
557
+ cpu->isar.id_isar3 = 0x01112131;
558
+ cpu->isar.id_isar4 = 0x00011142;
559
+ cpu->isar.id_isar5 = 0x00011121;
560
+ cpu->isar.id_aa64pfr0 = 0x00002222;
561
cpu->id_aa64dfr0 = 0x10305106;
562
cpu->pmceid0 = 0x00000000;
563
cpu->pmceid1 = 0x00000000;
564
- cpu->id_aa64isar0 = 0x00011120;
565
+ cpu->isar.id_aa64isar0 = 0x00011120;
566
cpu->id_aa64mmfr0 = 0x00001124;
567
cpu->dbgdidr = 0x3516d000;
568
cpu->clidr = 0x0a200023;
569
diff --git a/target/arm/helper.c b/target/arm/helper.c
570
index XXXXXXX..XXXXXXX 100644
571
--- a/target/arm/helper.c
572
+++ b/target/arm/helper.c
573
@@ -XXX,XX +XXX,XX @@ static uint64_t id_pfr1_read(CPUARMState *env, const ARMCPRegInfo *ri)
574
static uint64_t id_aa64pfr0_read(CPUARMState *env, const ARMCPRegInfo *ri)
575
{
576
ARMCPU *cpu = arm_env_get_cpu(env);
577
- uint64_t pfr0 = cpu->id_aa64pfr0;
578
+ uint64_t pfr0 = cpu->isar.id_aa64pfr0;
579
580
if (env->gicv3state) {
581
pfr0 |= 1 << 24;
582
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
583
{ .name = "ID_ISAR0", .state = ARM_CP_STATE_BOTH,
584
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 0,
585
.access = PL1_R, .type = ARM_CP_CONST,
586
- .resetvalue = cpu->id_isar0 },
587
+ .resetvalue = cpu->isar.id_isar0 },
588
{ .name = "ID_ISAR1", .state = ARM_CP_STATE_BOTH,
589
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 1,
590
.access = PL1_R, .type = ARM_CP_CONST,
591
- .resetvalue = cpu->id_isar1 },
592
+ .resetvalue = cpu->isar.id_isar1 },
593
{ .name = "ID_ISAR2", .state = ARM_CP_STATE_BOTH,
594
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 2,
595
.access = PL1_R, .type = ARM_CP_CONST,
596
- .resetvalue = cpu->id_isar2 },
597
+ .resetvalue = cpu->isar.id_isar2 },
598
{ .name = "ID_ISAR3", .state = ARM_CP_STATE_BOTH,
599
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 3,
600
.access = PL1_R, .type = ARM_CP_CONST,
601
- .resetvalue = cpu->id_isar3 },
602
+ .resetvalue = cpu->isar.id_isar3 },
603
{ .name = "ID_ISAR4", .state = ARM_CP_STATE_BOTH,
604
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 4,
605
.access = PL1_R, .type = ARM_CP_CONST,
606
- .resetvalue = cpu->id_isar4 },
607
+ .resetvalue = cpu->isar.id_isar4 },
608
{ .name = "ID_ISAR5", .state = ARM_CP_STATE_BOTH,
609
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 5,
610
.access = PL1_R, .type = ARM_CP_CONST,
611
- .resetvalue = cpu->id_isar5 },
612
+ .resetvalue = cpu->isar.id_isar5 },
613
{ .name = "ID_MMFR4", .state = ARM_CP_STATE_BOTH,
614
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 6,
615
.access = PL1_R, .type = ARM_CP_CONST,
616
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
617
{ .name = "ID_ISAR6", .state = ARM_CP_STATE_BOTH,
618
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 2, .opc2 = 7,
619
.access = PL1_R, .type = ARM_CP_CONST,
620
- .resetvalue = cpu->id_isar6 },
621
+ .resetvalue = cpu->isar.id_isar6 },
622
REGINFO_SENTINEL
623
};
624
define_arm_cp_regs(cpu, v6_idregs);
625
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
626
{ .name = "ID_AA64PFR1_EL1", .state = ARM_CP_STATE_AA64,
627
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 4, .opc2 = 1,
628
.access = PL1_R, .type = ARM_CP_CONST,
629
- .resetvalue = cpu->id_aa64pfr1},
630
+ .resetvalue = cpu->isar.id_aa64pfr1},
631
{ .name = "ID_AA64PFR2_EL1_RESERVED", .state = ARM_CP_STATE_AA64,
632
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 4, .opc2 = 2,
633
.access = PL1_R, .type = ARM_CP_CONST,
634
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
635
{ .name = "ID_AA64ISAR0_EL1", .state = ARM_CP_STATE_AA64,
636
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 6, .opc2 = 0,
637
.access = PL1_R, .type = ARM_CP_CONST,
638
- .resetvalue = cpu->id_aa64isar0 },
639
+ .resetvalue = cpu->isar.id_aa64isar0 },
640
{ .name = "ID_AA64ISAR1_EL1", .state = ARM_CP_STATE_AA64,
641
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 6, .opc2 = 1,
642
.access = PL1_R, .type = ARM_CP_CONST,
643
- .resetvalue = cpu->id_aa64isar1 },
644
+ .resetvalue = cpu->isar.id_aa64isar1 },
645
{ .name = "ID_AA64ISAR2_EL1_RESERVED", .state = ARM_CP_STATE_AA64,
646
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 6, .opc2 = 2,
647
.access = PL1_R, .type = ARM_CP_CONST,
648
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
649
{ .name = "MVFR0_EL1", .state = ARM_CP_STATE_AA64,
650
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 3, .opc2 = 0,
651
.access = PL1_R, .type = ARM_CP_CONST,
652
- .resetvalue = cpu->mvfr0 },
653
+ .resetvalue = cpu->isar.mvfr0 },
654
{ .name = "MVFR1_EL1", .state = ARM_CP_STATE_AA64,
655
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 3, .opc2 = 1,
656
.access = PL1_R, .type = ARM_CP_CONST,
657
- .resetvalue = cpu->mvfr1 },
658
+ .resetvalue = cpu->isar.mvfr1 },
659
{ .name = "MVFR2_EL1", .state = ARM_CP_STATE_AA64,
660
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 3, .opc2 = 2,
661
.access = PL1_R, .type = ARM_CP_CONST,
662
- .resetvalue = cpu->mvfr2 },
663
+ .resetvalue = cpu->isar.mvfr2 },
664
{ .name = "MVFR3_EL1_RESERVED", .state = ARM_CP_STATE_AA64,
665
.opc0 = 3, .opc1 = 0, .crn = 0, .crm = 3, .opc2 = 3,
666
.access = PL1_R, .type = ARM_CP_CONST,
667
--
55
--
668
2.19.1
56
2.20.1
669
57
670
58
diff view generated by jsdifflib
1
The HCR.FB virtualization configuration register bit requests that
1
From: Richard Henderson <richard.henderson@linaro.org>
2
TLB maintenance, branch predictor invalidate-all and icache
3
invalidate-all operations performed in NS EL1 should be upgraded
4
from "local CPU only to "broadcast within Inner Shareable domain".
5
For QEMU we NOP the branch predictor and icache operations, so
6
we only need to upgrade the TLB invalidates:
7
AArch32 TLBIALL, TLBIMVA, TLBIASID, DTLBIALL, DTLBIMVA, DTLBIASID,
8
ITLBIALL, ITLBIMVA, ITLBIASID, TLBIMVAA, TLBIMVAL, TLBIMVAAL
9
AArch64 TLBI VMALLE1, TLBI VAE1, TLBI ASIDE1, TLBI VAAE1,
10
TLBI VALE1, TLBI VAALE1
11
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-16-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20181012144235.19646-4-peter.maydell@linaro.org
15
---
7
---
16
target/arm/helper.c | 191 +++++++++++++++++++++++++++-----------------
8
target/arm/helper-a64.h | 7 ++
17
1 file changed, 116 insertions(+), 75 deletions(-)
9
target/arm/helper.h | 2 +
10
target/arm/mte_helper.c | 194 +++++++++++++++++++++++++++++++++++++
11
target/arm/op_helper.c | 16 +++
12
target/arm/translate-a64.c | 172 +++++++++++++++++++++++++++++++-
13
5 files changed, 386 insertions(+), 5 deletions(-)
18
14
19
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper.c
17
--- a/target/arm/helper-a64.h
22
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper-a64.h
23
@@ -XXX,XX +XXX,XX @@ static void contextidr_write(CPUARMState *env, const ARMCPRegInfo *ri,
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
raw_write(env, ri, value);
20
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
+DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
24
+DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
25
+DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
26
+DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
27
+DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
28
+DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
29
+DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
30
diff --git a/target/arm/helper.h b/target/arm/helper.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.h
33
+++ b/target/arm/helper.h
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
35
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
36
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
37
38
+DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
39
+
40
DEF_HELPER_1(vfp_get_fpscr, i32, env)
41
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
42
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
48
return tag;
25
}
49
}
26
50
27
-static void tlbiall_write(CPUARMState *env, const ARMCPRegInfo *ri,
51
+/**
28
- uint64_t value)
52
+ * allocation_tag_mem:
29
-{
53
+ * @env: the cpu environment
30
- /* Invalidate all (TLBIALL) */
54
+ * @ptr_mmu_idx: the addressing regime to use for the virtual address
31
- ARMCPU *cpu = arm_env_get_cpu(env);
55
+ * @ptr: the virtual address for which to look up tag memory
32
-
56
+ * @ptr_access: the access to use for the virtual address
33
- tlb_flush(CPU(cpu));
57
+ * @ptr_size: the number of bytes in the normal memory access
34
-}
58
+ * @tag_access: the access to use for the tag memory
35
-
59
+ * @tag_size: the number of bytes in the tag memory access
36
-static void tlbimva_write(CPUARMState *env, const ARMCPRegInfo *ri,
60
+ * @ra: the return address for exception handling
37
- uint64_t value)
61
+ *
38
-{
62
+ * Our tag memory is formatted as a sequence of little-endian nibbles.
39
- /* Invalidate single TLB entry by MVA and ASID (TLBIMVA) */
63
+ * That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
40
- ARMCPU *cpu = arm_env_get_cpu(env);
64
+ * tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
41
-
65
+ * for the higher addr.
42
- tlb_flush_page(CPU(cpu), value & TARGET_PAGE_MASK);
66
+ *
43
-}
67
+ * Here, resolve the physical address from the virtual address, and return
44
-
68
+ * a pointer to the corresponding tag byte. Exit with exception if the
45
-static void tlbiasid_write(CPUARMState *env, const ARMCPRegInfo *ri,
69
+ * virtual address is not accessible for @ptr_access.
46
- uint64_t value)
70
+ *
47
-{
71
+ * The @ptr_size and @tag_size values may not have an obvious relation
48
- /* Invalidate by ASID (TLBIASID) */
72
+ * due to the alignment of @ptr, and the number of tag checks required.
49
- ARMCPU *cpu = arm_env_get_cpu(env);
73
+ *
50
-
74
+ * If there is no tag storage corresponding to @ptr, return NULL.
51
- tlb_flush(CPU(cpu));
75
+ */
52
-}
76
+static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
53
-
77
+ uint64_t ptr, MMUAccessType ptr_access,
54
-static void tlbimvaa_write(CPUARMState *env, const ARMCPRegInfo *ri,
78
+ int ptr_size, MMUAccessType tag_access,
55
- uint64_t value)
79
+ int tag_size, uintptr_t ra)
56
-{
80
+{
57
- /* Invalidate single entry by MVA, all ASIDs (TLBIMVAA) */
81
+ /* Tag storage not implemented. */
58
- ARMCPU *cpu = arm_env_get_cpu(env);
82
+ return NULL;
59
-
83
+}
60
- tlb_flush_page(CPU(cpu), value & TARGET_PAGE_MASK);
84
+
61
-}
85
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
62
-
86
{
63
/* IS variants of TLB operations must affect all cores */
87
int rtag;
64
static void tlbiall_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
88
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
65
uint64_t value)
89
66
@@ -XXX,XX +XXX,XX @@ static void tlbimvaa_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
90
return address_with_allocation_tag(ptr + offset, rtag);
67
tlb_flush_page_all_cpus_synced(cs, value & TARGET_PAGE_MASK);
68
}
91
}
69
92
+
70
+/*
93
+static int load_tag1(uint64_t ptr, uint8_t *mem)
71
+ * Non-IS variants of TLB operations are upgraded to
94
+{
72
+ * IS versions if we are at NS EL1 and HCR_EL2.FB is set to
95
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
73
+ * force broadcast of these operations.
96
+ return extract32(*mem, ofs, 4);
74
+ */
97
+}
75
+static bool tlb_force_broadcast(CPUARMState *env)
98
+
76
+{
99
+uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
77
+ return (env->cp15.hcr_el2 & HCR_FB) &&
100
+{
78
+ arm_current_el(env) == 1 && arm_is_secure_below_el3(env);
101
+ int mmu_idx = cpu_mmu_index(env, false);
79
+}
102
+ uint8_t *mem;
80
+
103
+ int rtag = 0;
81
+static void tlbiall_write(CPUARMState *env, const ARMCPRegInfo *ri,
104
+
82
+ uint64_t value)
105
+ /* Trap if accessing an invalid page. */
83
+{
106
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
84
+ /* Invalidate all (TLBIALL) */
107
+ MMU_DATA_LOAD, 1, GETPC());
85
+ ARMCPU *cpu = arm_env_get_cpu(env);
108
+
86
+
109
+ /* Load if page supports tags. */
87
+ if (tlb_force_broadcast(env)) {
110
+ if (mem) {
88
+ tlbiall_is_write(env, NULL, value);
111
+ rtag = load_tag1(ptr, mem);
89
+ return;
112
+ }
90
+ }
113
+
91
+
114
+ return address_with_allocation_tag(xt, rtag);
92
+ tlb_flush(CPU(cpu));
115
+}
93
+}
116
+
94
+
117
+static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
95
+static void tlbimva_write(CPUARMState *env, const ARMCPRegInfo *ri,
118
+{
96
+ uint64_t value)
119
+ if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
97
+{
120
+ arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
98
+ /* Invalidate single TLB entry by MVA and ASID (TLBIMVA) */
121
+ cpu_mmu_index(env, false), ra);
99
+ ARMCPU *cpu = arm_env_get_cpu(env);
122
+ g_assert_not_reached();
100
+
123
+ }
101
+ if (tlb_force_broadcast(env)) {
124
+}
102
+ tlbimva_is_write(env, NULL, value);
125
+
103
+ return;
126
+/* For use in a non-parallel context, store to the given nibble. */
104
+ }
127
+static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
105
+
128
+{
106
+ tlb_flush_page(CPU(cpu), value & TARGET_PAGE_MASK);
129
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
107
+}
130
+ *mem = deposit32(*mem, ofs, 4, tag);
108
+
131
+}
109
+static void tlbiasid_write(CPUARMState *env, const ARMCPRegInfo *ri,
132
+
110
+ uint64_t value)
133
+/* For use in a parallel context, atomically store to the given nibble. */
111
+{
134
+static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
112
+ /* Invalidate by ASID (TLBIASID) */
135
+{
113
+ ARMCPU *cpu = arm_env_get_cpu(env);
136
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
114
+
137
+ uint8_t old = atomic_read(mem);
115
+ if (tlb_force_broadcast(env)) {
138
+
116
+ tlbiasid_is_write(env, NULL, value);
139
+ while (1) {
117
+ return;
140
+ uint8_t new = deposit32(old, ofs, 4, tag);
118
+ }
141
+ uint8_t cmp = atomic_cmpxchg(mem, old, new);
119
+
142
+ if (likely(cmp == old)) {
120
+ tlb_flush(CPU(cpu));
143
+ return;
121
+}
144
+ }
122
+
145
+ old = cmp;
123
+static void tlbimvaa_write(CPUARMState *env, const ARMCPRegInfo *ri,
146
+ }
124
+ uint64_t value)
147
+}
125
+{
148
+
126
+ /* Invalidate single entry by MVA, all ASIDs (TLBIMVAA) */
149
+typedef void stg_store1(uint64_t, uint8_t *, int);
127
+ ARMCPU *cpu = arm_env_get_cpu(env);
150
+
128
+
151
+static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
129
+ if (tlb_force_broadcast(env)) {
152
+ uintptr_t ra, stg_store1 store1)
130
+ tlbimvaa_is_write(env, NULL, value);
153
+{
131
+ return;
154
+ int mmu_idx = cpu_mmu_index(env, false);
132
+ }
155
+ uint8_t *mem;
133
+
156
+
134
+ tlb_flush_page(CPU(cpu), value & TARGET_PAGE_MASK);
157
+ check_tag_aligned(env, ptr, ra);
135
+}
158
+
136
+
159
+ /* Trap if accessing an invalid page. */
137
static void tlbiall_nsnh_write(CPUARMState *env, const ARMCPRegInfo *ri,
160
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
138
uint64_t value)
161
+ MMU_DATA_STORE, 1, ra);
139
{
162
+
140
@@ -XXX,XX +XXX,XX @@ static CPAccessResult aa64_cacheop_access(CPUARMState *env,
163
+ /* Store if page supports tags. */
141
* Page D4-1736 (DDI0487A.b)
164
+ if (mem) {
142
*/
165
+ store1(ptr, mem, allocation_tag_from_addr(xt));
143
166
+ }
144
-static void tlbi_aa64_vmalle1_write(CPUARMState *env, const ARMCPRegInfo *ri,
167
+}
145
- uint64_t value)
168
+
146
-{
169
+void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
147
- CPUState *cs = ENV_GET_CPU(env);
170
+{
148
-
171
+ do_stg(env, ptr, xt, GETPC(), store_tag1);
149
- if (arm_is_secure_below_el3(env)) {
172
+}
150
- tlb_flush_by_mmuidx(cs,
173
+
151
- ARMMMUIdxBit_S1SE1 |
174
+void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
152
- ARMMMUIdxBit_S1SE0);
175
+{
153
- } else {
176
+ do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
154
- tlb_flush_by_mmuidx(cs,
177
+}
155
- ARMMMUIdxBit_S12NSE1 |
178
+
156
- ARMMMUIdxBit_S12NSE0);
179
+void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
157
- }
180
+{
158
-}
181
+ int mmu_idx = cpu_mmu_index(env, false);
159
-
182
+ uintptr_t ra = GETPC();
160
static void tlbi_aa64_vmalle1is_write(CPUARMState *env, const ARMCPRegInfo *ri,
183
+
161
uint64_t value)
184
+ check_tag_aligned(env, ptr, ra);
162
{
185
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
163
@@ -XXX,XX +XXX,XX @@ static void tlbi_aa64_vmalle1is_write(CPUARMState *env, const ARMCPRegInfo *ri,
186
+}
187
+
188
+static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
189
+ uintptr_t ra, stg_store1 store1)
190
+{
191
+ int mmu_idx = cpu_mmu_index(env, false);
192
+ int tag = allocation_tag_from_addr(xt);
193
+ uint8_t *mem1, *mem2;
194
+
195
+ check_tag_aligned(env, ptr, ra);
196
+
197
+ /*
198
+ * Trap if accessing an invalid page(s).
199
+ * This takes priority over !allocation_tag_access_enabled.
200
+ */
201
+ if (ptr & TAG_GRANULE) {
202
+ /* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
203
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
204
+ TAG_GRANULE, MMU_DATA_STORE, 1, ra);
205
+ mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
206
+ MMU_DATA_STORE, TAG_GRANULE,
207
+ MMU_DATA_STORE, 1, ra);
208
+
209
+ /* Store if page(s) support tags. */
210
+ if (mem1) {
211
+ store1(TAG_GRANULE, mem1, tag);
212
+ }
213
+ if (mem2) {
214
+ store1(0, mem2, tag);
215
+ }
216
+ } else {
217
+ /* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
218
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
219
+ 2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
220
+ if (mem1) {
221
+ tag |= tag << 4;
222
+ atomic_set(mem1, tag);
223
+ }
224
+ }
225
+}
226
+
227
+void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
228
+{
229
+ do_st2g(env, ptr, xt, GETPC(), store_tag1);
230
+}
231
+
232
+void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
233
+{
234
+ do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
235
+}
236
+
237
+void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
238
+{
239
+ int mmu_idx = cpu_mmu_index(env, false);
240
+ uintptr_t ra = GETPC();
241
+ int in_page = -(ptr | TARGET_PAGE_MASK);
242
+
243
+ check_tag_aligned(env, ptr, ra);
244
+
245
+ if (likely(in_page >= 2 * TAG_GRANULE)) {
246
+ probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
247
+ } else {
248
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
249
+ probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
250
+ }
251
+}
252
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
253
index XXXXXXX..XXXXXXX 100644
254
--- a/target/arm/op_helper.c
255
+++ b/target/arm/op_helper.c
256
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
257
return ((uint32_t)x >> shift) | (x << (32 - shift));
164
}
258
}
165
}
259
}
166
260
+
167
+static void tlbi_aa64_vmalle1_write(CPUARMState *env, const ARMCPRegInfo *ri,
261
+void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
168
+ uint64_t value)
262
+ uint32_t access_type, uint32_t mmu_idx,
169
+{
263
+ uint32_t size)
170
+ CPUState *cs = ENV_GET_CPU(env);
264
+{
171
+
265
+ uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
172
+ if (tlb_force_broadcast(env)) {
266
+ uintptr_t ra = GETPC();
173
+ tlbi_aa64_vmalle1_write(env, NULL, value);
267
+
174
+ return;
268
+ if (likely(size <= in_page)) {
175
+ }
269
+ probe_access(env, ptr, size, access_type, mmu_idx, ra);
176
+
177
+ if (arm_is_secure_below_el3(env)) {
178
+ tlb_flush_by_mmuidx(cs,
179
+ ARMMMUIdxBit_S1SE1 |
180
+ ARMMMUIdxBit_S1SE0);
181
+ } else {
270
+ } else {
182
+ tlb_flush_by_mmuidx(cs,
271
+ probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
183
+ ARMMMUIdxBit_S12NSE1 |
272
+ probe_access(env, ptr + in_page, size - in_page,
184
+ ARMMMUIdxBit_S12NSE0);
273
+ access_type, mmu_idx, ra);
185
+ }
274
+ }
186
+}
275
+}
187
+
276
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
188
static void tlbi_aa64_alle1_write(CPUARMState *env, const ARMCPRegInfo *ri,
277
index XXXXXXX..XXXXXXX 100644
189
uint64_t value)
278
--- a/target/arm/translate-a64.c
190
{
279
+++ b/target/arm/translate-a64.c
191
@@ -XXX,XX +XXX,XX @@ static void tlbi_aa64_alle3is_write(CPUARMState *env, const ARMCPRegInfo *ri,
280
@@ -XXX,XX +XXX,XX @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
192
tlb_flush_by_mmuidx_all_cpus_synced(cs, ARMMMUIdxBit_S1E3);
281
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
193
}
282
}
194
283
195
-static void tlbi_aa64_vae1_write(CPUARMState *env, const ARMCPRegInfo *ri,
284
+static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
196
- uint64_t value)
285
+ MMUAccessType acc, int log2_size)
197
-{
286
+{
198
- /* Invalidate by VA, EL1&0 (AArch64 version).
287
+ TCGv_i32 t_acc = tcg_const_i32(acc);
199
- * Currently handles all of VAE1, VAAE1, VAALE1 and VALE1,
288
+ TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
200
- * since we don't support flush-for-specific-ASID-only or
289
+ TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
201
- * flush-last-level-only.
290
+
202
- */
291
+ gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
203
- ARMCPU *cpu = arm_env_get_cpu(env);
292
+ tcg_temp_free_i32(t_acc);
204
- CPUState *cs = CPU(cpu);
293
+ tcg_temp_free_i32(t_idx);
205
- uint64_t pageaddr = sextract64(value << 12, 0, 56);
294
+ tcg_temp_free_i32(t_size);
206
-
295
+}
207
- if (arm_is_secure_below_el3(env)) {
296
+
208
- tlb_flush_page_by_mmuidx(cs, pageaddr,
297
typedef struct DisasCompare64 {
209
- ARMMMUIdxBit_S1SE1 |
298
TCGCond cond;
210
- ARMMMUIdxBit_S1SE0);
299
TCGv_i64 value;
211
- } else {
300
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
212
- tlb_flush_page_by_mmuidx(cs, pageaddr,
213
- ARMMMUIdxBit_S12NSE1 |
214
- ARMMMUIdxBit_S12NSE0);
215
- }
216
-}
217
-
218
static void tlbi_aa64_vae2_write(CPUARMState *env, const ARMCPRegInfo *ri,
219
uint64_t value)
220
{
221
@@ -XXX,XX +XXX,XX @@ static void tlbi_aa64_vae1is_write(CPUARMState *env, const ARMCPRegInfo *ri,
222
}
301
}
223
}
302
}
224
303
225
+static void tlbi_aa64_vae1_write(CPUARMState *env, const ARMCPRegInfo *ri,
304
+/*
226
+ uint64_t value)
305
+ * Load/Store memory tags
227
+{
306
+ *
228
+ /* Invalidate by VA, EL1&0 (AArch64 version).
307
+ * 31 30 29 24 22 21 12 10 5 0
229
+ * Currently handles all of VAE1, VAAE1, VAALE1 and VALE1,
308
+ * +-----+-------------+-----+---+------+-----+------+------+
230
+ * since we don't support flush-for-specific-ASID-only or
309
+ * | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
231
+ * flush-last-level-only.
310
+ * +-----+-------------+-----+---+------+-----+------+------+
311
+ */
312
+static void disas_ldst_tag(DisasContext *s, uint32_t insn)
313
+{
314
+ int rt = extract32(insn, 0, 5);
315
+ int rn = extract32(insn, 5, 5);
316
+ uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
317
+ int op2 = extract32(insn, 10, 2);
318
+ int op1 = extract32(insn, 22, 2);
319
+ bool is_load = false, is_pair = false, is_zero = false;
320
+ int index = 0;
321
+ TCGv_i64 addr, clean_addr, tcg_rt;
322
+
323
+ /* We checked insn bits [29:24,21] in the caller. */
324
+ if (extract32(insn, 30, 2) != 3) {
325
+ goto do_unallocated;
326
+ }
327
+
328
+ /*
329
+ * @index is a tri-state variable which has 3 states:
330
+ * < 0 : post-index, writeback
331
+ * = 0 : signed offset
332
+ * > 0 : pre-index, writeback
232
+ */
333
+ */
233
+ ARMCPU *cpu = arm_env_get_cpu(env);
334
+ switch (op1) {
234
+ CPUState *cs = CPU(cpu);
335
+ case 0:
235
+ uint64_t pageaddr = sextract64(value << 12, 0, 56);
336
+ if (op2 != 0) {
236
+
337
+ /* STG */
237
+ if (tlb_force_broadcast(env)) {
338
+ index = op2 - 2;
238
+ tlbi_aa64_vae1is_write(env, NULL, value);
339
+ break;
340
+ }
341
+ goto do_unallocated;
342
+ case 1:
343
+ if (op2 != 0) {
344
+ /* STZG */
345
+ is_zero = true;
346
+ index = op2 - 2;
347
+ } else {
348
+ /* LDG */
349
+ is_load = true;
350
+ }
351
+ break;
352
+ case 2:
353
+ if (op2 != 0) {
354
+ /* ST2G */
355
+ is_pair = true;
356
+ index = op2 - 2;
357
+ break;
358
+ }
359
+ goto do_unallocated;
360
+ case 3:
361
+ if (op2 != 0) {
362
+ /* STZ2G */
363
+ is_pair = is_zero = true;
364
+ index = op2 - 2;
365
+ break;
366
+ }
367
+ goto do_unallocated;
368
+
369
+ default:
370
+ do_unallocated:
371
+ unallocated_encoding(s);
239
+ return;
372
+ return;
240
+ }
373
+ }
241
+
374
+
242
+ if (arm_is_secure_below_el3(env)) {
375
+ if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
243
+ tlb_flush_page_by_mmuidx(cs, pageaddr,
376
+ goto do_unallocated;
244
+ ARMMMUIdxBit_S1SE1 |
377
+ }
245
+ ARMMMUIdxBit_S1SE0);
378
+
379
+ if (rn == 31) {
380
+ gen_check_sp_alignment(s);
381
+ }
382
+
383
+ addr = read_cpu_reg_sp(s, rn, true);
384
+ if (index >= 0) {
385
+ /* pre-index or signed offset */
386
+ tcg_gen_addi_i64(addr, addr, offset);
387
+ }
388
+
389
+ if (is_load) {
390
+ tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
391
+ tcg_rt = cpu_reg(s, rt);
392
+ if (s->ata) {
393
+ gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
394
+ } else {
395
+ clean_addr = clean_data_tbi(s, addr);
396
+ gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
397
+ gen_address_with_allocation_tag0(tcg_rt, addr);
398
+ }
246
+ } else {
399
+ } else {
247
+ tlb_flush_page_by_mmuidx(cs, pageaddr,
400
+ tcg_rt = cpu_reg_sp(s, rt);
248
+ ARMMMUIdxBit_S12NSE1 |
401
+ if (!s->ata) {
249
+ ARMMMUIdxBit_S12NSE0);
402
+ /*
250
+ }
403
+ * For STG and ST2G, we need to check alignment and probe memory.
251
+}
404
+ * TODO: For STZG and STZ2G, we could rely on the stores below,
252
+
405
+ * at least for system mode; user-only won't enforce alignment.
253
static void tlbi_aa64_vae2is_write(CPUARMState *env, const ARMCPRegInfo *ri,
406
+ */
254
uint64_t value)
407
+ if (is_pair) {
408
+ gen_helper_st2g_stub(cpu_env, addr);
409
+ } else {
410
+ gen_helper_stg_stub(cpu_env, addr);
411
+ }
412
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
413
+ if (is_pair) {
414
+ gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
415
+ } else {
416
+ gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
417
+ }
418
+ } else {
419
+ if (is_pair) {
420
+ gen_helper_st2g(cpu_env, addr, tcg_rt);
421
+ } else {
422
+ gen_helper_stg(cpu_env, addr, tcg_rt);
423
+ }
424
+ }
425
+ }
426
+
427
+ if (is_zero) {
428
+ TCGv_i64 clean_addr = clean_data_tbi(s, addr);
429
+ TCGv_i64 tcg_zero = tcg_const_i64(0);
430
+ int mem_index = get_mem_index(s);
431
+ int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
432
+
433
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
434
+ MO_Q | MO_ALIGN_16);
435
+ for (i = 8; i < n; i += 8) {
436
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
437
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
438
+ }
439
+ tcg_temp_free_i64(tcg_zero);
440
+ }
441
+
442
+ if (index != 0) {
443
+ /* pre-index or post-index */
444
+ if (index < 0) {
445
+ /* post-index */
446
+ tcg_gen_addi_i64(addr, addr, offset);
447
+ }
448
+ tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
449
+ }
450
+}
451
+
452
/* Loads and stores */
453
static void disas_ldst(DisasContext *s, uint32_t insn)
255
{
454
{
455
@@ -XXX,XX +XXX,XX @@ static void disas_ldst(DisasContext *s, uint32_t insn)
456
case 0x0d: /* AdvSIMD load/store single structure */
457
disas_ldst_single_struct(s, insn);
458
break;
459
- case 0x19: /* LDAPR/STLR (unscaled immediate) */
460
- if (extract32(insn, 10, 2) != 0 ||
461
- extract32(insn, 21, 1) != 0) {
462
+ case 0x19:
463
+ if (extract32(insn, 21, 1) != 0) {
464
+ disas_ldst_tag(s, insn);
465
+ } else if (extract32(insn, 10, 2) == 0) {
466
+ disas_ldst_ldapr_stlr(s, insn);
467
+ } else {
468
unallocated_encoding(s);
469
- break;
470
}
471
- disas_ldst_ldapr_stlr(s, insn);
472
break;
473
default:
474
unallocated_encoding(s);
256
--
475
--
257
2.19.1
476
2.20.1
258
477
259
478
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
For a sequence of loads or stores from a single register,
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
little-endian operations can be promoted to an 8-byte op.
5
This can reduce the number of operations by a factor of 8.
6
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20181011205206.3552-5-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-17-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
7
---
12
target/arm/translate-a64.c | 66 +++++++++++++++++++++++---------------
8
target/arm/translate-a64.c | 29 ++++++++++++++++++++++++++---
13
1 file changed, 40 insertions(+), 26 deletions(-)
9
1 file changed, 26 insertions(+), 3 deletions(-)
14
10
15
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
16
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate-a64.c
13
--- a/target/arm/translate-a64.c
18
+++ b/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
19
@@ -XXX,XX +XXX,XX @@ static void write_vec_element_i32(DisasContext *s, TCGv_i32 tcg_src,
15
@@ -XXX,XX +XXX,XX @@ static void disas_ld_lit(DisasContext *s, uint32_t insn)
20
16
* +-----+-------+---+---+-------+---+-------+-------+------+------+
21
/* Store from vector register to memory */
17
*
22
static void do_vec_st(DisasContext *s, int srcidx, int element,
18
* opc: LDP/STP/LDNP/STNP 00 -> 32 bit, 10 -> 64 bit
23
- TCGv_i64 tcg_addr, int size)
19
- * LDPSW 01
24
+ TCGv_i64 tcg_addr, int size, TCGMemOp endian)
20
+ * LDPSW/STGP 01
25
{
21
* LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
26
- TCGMemOp memop = s->be_data + size;
22
* V: 0 -> GPR, 1 -> Vector
27
TCGv_i64 tcg_tmp = tcg_temp_new_i64();
23
* idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
28
24
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
29
read_vec_element(s, tcg_tmp, srcidx, element, size);
25
bool is_signed = false;
30
- tcg_gen_qemu_st_i64(tcg_tmp, tcg_addr, get_mem_index(s), memop);
26
bool postindex = false;
31
+ tcg_gen_qemu_st_i64(tcg_tmp, tcg_addr, get_mem_index(s), endian | size);
27
bool wback = false;
32
28
+ bool set_tag = false;
33
tcg_temp_free_i64(tcg_tmp);
29
34
}
30
TCGv_i64 clean_addr, dirty_addr;
35
31
36
/* Load from memory to vector register */
32
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
37
static void do_vec_ld(DisasContext *s, int destidx, int element,
33
38
- TCGv_i64 tcg_addr, int size)
34
if (is_vector) {
39
+ TCGv_i64 tcg_addr, int size, TCGMemOp endian)
35
size = 2 + opc;
40
{
36
+ } else if (opc == 1 && !is_load) {
41
- TCGMemOp memop = s->be_data + size;
37
+ /* STGP */
42
TCGv_i64 tcg_tmp = tcg_temp_new_i64();
38
+ if (!dc_isar_feature(aa64_mte_insn_reg, s) || index == 0) {
43
39
+ unallocated_encoding(s);
44
- tcg_gen_qemu_ld_i64(tcg_tmp, tcg_addr, get_mem_index(s), memop);
40
+ return;
45
+ tcg_gen_qemu_ld_i64(tcg_tmp, tcg_addr, get_mem_index(s), endian | size);
41
+ }
46
write_vec_element(s, tcg_tmp, destidx, element, size);
42
+ size = 3;
47
43
+ set_tag = true;
48
tcg_temp_free_i64(tcg_tmp);
44
} else {
49
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
45
size = 2 + extract32(opc, 1, 1);
50
bool is_postidx = extract32(insn, 23, 1);
46
is_signed = extract32(opc, 0, 1);
51
bool is_q = extract32(insn, 30, 1);
47
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_addr, tcg_rn, tcg_ebytes;
48
return;
53
+ TCGMemOp endian = s->be_data;
49
}
54
50
55
- int ebytes = 1 << size;
51
- offset <<= size;
56
- int elements = (is_q ? 128 : 64) / (8 << size);
52
+ offset <<= (set_tag ? LOG2_TAG_GRANULE : size);
57
+ int ebytes; /* bytes per element */
53
58
+ int elements; /* elements per vector */
54
if (rn == 31) {
59
int rpt; /* num iterations */
60
int selem; /* structure elements */
61
int r;
62
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
63
gen_check_sp_alignment(s);
55
gen_check_sp_alignment(s);
56
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
57
if (!postindex) {
58
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
64
}
59
}
65
60
- clean_addr = clean_data_tbi(s, dirty_addr);
66
+ /* For our purposes, bytes are always little-endian. */
61
67
+ if (size == 0) {
62
+ if (set_tag) {
68
+ endian = MO_LE;
63
+ if (!s->ata) {
69
+ }
64
+ /*
70
+
65
+ * TODO: We could rely on the stores below, at least for
71
+ /* Consecutive little-endian elements from a single register
66
+ * system mode, if we arrange to add MO_ALIGN_16.
72
+ * can be promoted to a larger little-endian operation.
67
+ */
73
+ */
68
+ gen_helper_stg_stub(cpu_env, dirty_addr);
74
+ if (selem == 1 && endian == MO_LE) {
69
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
75
+ size = 3;
70
+ gen_helper_stg_parallel(cpu_env, dirty_addr, dirty_addr);
76
+ }
71
+ } else {
77
+ ebytes = 1 << size;
72
+ gen_helper_stg(cpu_env, dirty_addr, dirty_addr);
78
+ elements = (is_q ? 16 : 8) / ebytes;
79
+
80
tcg_rn = cpu_reg_sp(s, rn);
81
tcg_addr = tcg_temp_new_i64();
82
tcg_gen_mov_i64(tcg_addr, tcg_rn);
83
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
84
for (r = 0; r < rpt; r++) {
85
int e;
86
for (e = 0; e < elements; e++) {
87
- int tt = (rt + r) % 32;
88
int xs;
89
for (xs = 0; xs < selem; xs++) {
90
+ int tt = (rt + r + xs) % 32;
91
if (is_store) {
92
- do_vec_st(s, tt, e, tcg_addr, size);
93
+ do_vec_st(s, tt, e, tcg_addr, size, endian);
94
} else {
95
- do_vec_ld(s, tt, e, tcg_addr, size);
96
-
97
- /* For non-quad operations, setting a slice of the low
98
- * 64 bits of the register clears the high 64 bits (in
99
- * the ARM ARM pseudocode this is implicit in the fact
100
- * that 'rval' is a 64 bit wide variable).
101
- * For quad operations, we might still need to zero the
102
- * high bits of SVE. We optimize by noticing that we only
103
- * need to do this the first time we touch a register.
104
- */
105
- if (e == 0 && (r == 0 || xs == selem - 1)) {
106
- clear_vec_high(s, is_q, tt);
107
- }
108
+ do_vec_ld(s, tt, e, tcg_addr, size, endian);
109
}
110
tcg_gen_add_i64(tcg_addr, tcg_addr, tcg_ebytes);
111
- tt = (tt + 1) % 32;
112
}
113
}
114
}
115
116
+ if (!is_store) {
117
+ /* For non-quad operations, setting a slice of the low
118
+ * 64 bits of the register clears the high 64 bits (in
119
+ * the ARM ARM pseudocode this is implicit in the fact
120
+ * that 'rval' is a 64 bit wide variable).
121
+ * For quad operations, we might still need to zero the
122
+ * high bits of SVE.
123
+ */
124
+ for (r = 0; r < rpt * selem; r++) {
125
+ int tt = (rt + r) % 32;
126
+ clear_vec_high(s, is_q, tt);
127
+ }
73
+ }
128
+ }
74
+ }
129
+
75
+
130
if (is_postidx) {
76
+ clean_addr = clean_data_tbi(s, dirty_addr);
131
int rm = extract32(insn, 16, 5);
77
if (is_vector) {
132
if (rm == 31) {
78
if (is_load) {
133
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
79
do_fp_ld(s, rt, clean_addr, size);
134
} else {
135
/* Load/store one element per register */
136
if (is_load) {
137
- do_vec_ld(s, rt, index, tcg_addr, scale);
138
+ do_vec_ld(s, rt, index, tcg_addr, scale, s->be_data);
139
} else {
140
- do_vec_st(s, rt, index, tcg_addr, scale);
141
+ do_vec_st(s, rt, index, tcg_addr, scale, s->be_data);
142
}
143
}
144
tcg_gen_add_i64(tcg_addr, tcg_addr, tcg_ebytes);
145
--
80
--
146
2.19.1
81
2.20.1
147
82
148
83
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Instantiating mps2-an505 (cortex-m33) will fail make check when
3
We can simplify our DC_ZVA if we recognize that the largest BS
4
V7VE asserts that ID_ISAR0.Divide includes ARM division. It is
4
that we actually use in system mode is 64. Let us just assert
5
also wrong to include ARM_FEATURE_LPAE.
5
that it fits within TARGET_PAGE_SIZE.
6
6
7
For DC_GVA and STZGM, we want to be able to write whole bytes
8
of tag memory, so assert that BS is >= 2 * TAG_GRANULE, or 32.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20181016223115.24100-3-richard.henderson@linaro.org
12
Message-id: 20200626033144.790098-18-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
14
---
12
target/arm/cpu.c | 6 +++++-
15
target/arm/cpu.c | 24 ++++++++++++++++++++++++
13
1 file changed, 5 insertions(+), 1 deletion(-)
16
1 file changed, 24 insertions(+)
14
17
15
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
18
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
16
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.c
20
--- a/target/arm/cpu.c
18
+++ b/target/arm/cpu.c
21
+++ b/target/arm/cpu.c
19
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
22
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
20
23
}
21
/* Some features automatically imply others: */
24
#endif
22
if (arm_feature(env, ARM_FEATURE_V8)) {
25
23
- set_feature(env, ARM_FEATURE_V7VE);
26
+ if (tcg_enabled()) {
24
+ if (arm_feature(env, ARM_FEATURE_M)) {
27
+ int dcz_blocklen = 4 << cpu->dcz_blocksize;
25
+ set_feature(env, ARM_FEATURE_V7);
28
+
26
+ } else {
29
+ /*
27
+ set_feature(env, ARM_FEATURE_V7VE);
30
+ * We only support DCZ blocklen that fits on one page.
31
+ *
32
+ * Architectually this is always true. However TARGET_PAGE_SIZE
33
+ * is variable and, for compatibility with -machine virt-2.7,
34
+ * is only 1KiB, as an artifact of legacy ARMv5 subpage support.
35
+ * But even then, while the largest architectural DCZ blocklen
36
+ * is 2KiB, no cpu actually uses such a large blocklen.
37
+ */
38
+ assert(dcz_blocklen <= TARGET_PAGE_SIZE);
39
+
40
+ /*
41
+ * We only support DCZ blocksize >= 2*TAG_GRANULE, which is to say
42
+ * both nibbles of each byte storing tag data may be written at once.
43
+ * Since TAG_GRANULE is 16, this means that blocklen must be >= 32.
44
+ */
45
+ if (cpu_isar_feature(aa64_mte, cpu)) {
46
+ assert(dcz_blocklen >= 2 * TAG_GRANULE);
28
+ }
47
+ }
29
}
48
+ }
30
if (arm_feature(env, ARM_FEATURE_V7VE)) {
49
+
31
/* v7 Virtualization Extensions. In real hardware this implies
50
qemu_init_vcpu(cs);
51
cpu_reset(cs);
52
32
--
53
--
33
2.19.1
54
2.20.1
34
55
35
56
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Now that we know that the operation is on a single page,
4
we need not loop over pages while probing.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-12-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-19-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
target/arm/translate.c | 31 +++++++++++++++----------------
11
target/arm/helper-a64.c | 94 +++++++++++------------------------------
9
1 file changed, 15 insertions(+), 16 deletions(-)
12
1 file changed, 25 insertions(+), 69 deletions(-)
10
13
11
diff --git a/target/arm/translate.c b/target/arm/translate.c
14
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate.c
16
--- a/target/arm/helper-a64.c
14
+++ b/target/arm/translate.c
17
+++ b/target/arm/helper-a64.c
15
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
18
@@ -XXX,XX +XXX,XX @@ void HELPER(dc_zva)(CPUARMState *env, uint64_t vaddr_in)
16
vec_size, vec_size);
19
* (which matches the usual QEMU behaviour of not implementing either
17
}
20
* alignment faults or any memory attribute handling).
18
return 0;
21
*/
22
-
23
- ARMCPU *cpu = env_archcpu(env);
24
- uint64_t blocklen = 4 << cpu->dcz_blocksize;
25
+ int blocklen = 4 << env_archcpu(env)->dcz_blocksize;
26
uint64_t vaddr = vaddr_in & ~(blocklen - 1);
27
+ int mmu_idx = cpu_mmu_index(env, false);
28
+ void *mem;
19
+
29
+
20
+ case NEON_3R_VMUL: /* VMUL */
30
+ /*
21
+ if (u) {
31
+ * Trapless lookup. In addition to actual invalid page, may
22
+ /* Polynomial case allows only P8 and is handled below. */
32
+ * return NULL for I/O, watchpoints, clean pages, etc.
23
+ if (size != 0) {
33
+ */
24
+ return 1;
34
+ mem = tlb_vaddr_to_host(env, vaddr, MMU_DATA_STORE, mmu_idx);
25
+ }
35
26
+ } else {
36
#ifndef CONFIG_USER_ONLY
27
+ tcg_gen_gvec_mul(size, rd_ofs, rn_ofs, rm_ofs,
37
- {
28
+ vec_size, vec_size);
38
+ if (unlikely(!mem)) {
29
+ return 0;
39
+ uintptr_t ra = GETPC();
30
+ }
40
+
31
+ break;
41
/*
32
}
42
- * Slightly awkwardly, QEMU's TARGET_PAGE_SIZE may be less than
33
if (size == 3) {
43
- * the block size so we might have to do more than one TLB lookup.
34
/* 64-bit element instructions. */
44
- * We know that in fact for any v8 CPU the page size is at least 4K
35
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
45
- * and the block size must be 2K or less, but TARGET_PAGE_SIZE is only
36
return 1;
46
- * 1K as an artefact of legacy v5 subpage support being present in the
37
}
47
- * same QEMU executable. So in practice the hostaddr[] array has
38
break;
48
- * two entries, given the current setting of TARGET_PAGE_BITS_MIN.
39
- case NEON_3R_VMUL:
49
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
40
- if (u && (size != 0)) {
50
+ * the original pointer for an invalid page. But watchpoints require
41
- /* UNDEF on invalid size for polynomial subcase */
51
+ * that we probe the actual space. So do both.
42
- return 1;
52
*/
43
- }
53
- int maxidx = DIV_ROUND_UP(blocklen, TARGET_PAGE_SIZE);
44
- break;
54
- void *hostaddr[DIV_ROUND_UP(2 * KiB, 1 << TARGET_PAGE_BITS_MIN)];
45
case NEON_3R_VFM_VQRDMLSH:
55
- int try, i;
46
if (!arm_dc_feature(s, ARM_FEATURE_VFP4)) {
56
- unsigned mmu_idx = cpu_mmu_index(env, false);
47
return 1;
57
- TCGMemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
48
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
58
+ (void) probe_write(env, vaddr_in, 1, mmu_idx, ra);
49
}
59
+ mem = probe_write(env, vaddr, blocklen, mmu_idx, ra);
50
break;
60
51
case NEON_3R_VMUL:
61
- assert(maxidx <= ARRAY_SIZE(hostaddr));
52
- if (u) { /* polynomial */
62
-
53
- gen_helper_neon_mul_p8(tmp, tmp, tmp2);
63
- for (try = 0; try < 2; try++) {
54
- } else { /* Integer */
64
-
55
- switch (size) {
65
- for (i = 0; i < maxidx; i++) {
56
- case 0: gen_helper_neon_mul_u8(tmp, tmp, tmp2); break;
66
- hostaddr[i] = tlb_vaddr_to_host(env,
57
- case 1: gen_helper_neon_mul_u16(tmp, tmp, tmp2); break;
67
- vaddr + TARGET_PAGE_SIZE * i,
58
- case 2: tcg_gen_mul_i32(tmp, tmp, tmp2); break;
68
- 1, mmu_idx);
59
- default: abort();
69
- if (!hostaddr[i]) {
70
- break;
60
- }
71
- }
61
- }
72
- }
62
+ /* VMUL.P8; other cases already eliminated. */
73
- if (i == maxidx) {
63
+ gen_helper_neon_mul_p8(tmp, tmp, tmp2);
74
- /*
64
break;
75
- * If it's all in the TLB it's fair game for just writing to;
65
case NEON_3R_VPMAX:
76
- * we know we don't need to update dirty status, etc.
66
GEN_NEON_INTEGER_OP(pmax);
77
- */
78
- for (i = 0; i < maxidx - 1; i++) {
79
- memset(hostaddr[i], 0, TARGET_PAGE_SIZE);
80
- }
81
- memset(hostaddr[i], 0, blocklen - (i * TARGET_PAGE_SIZE));
82
- return;
83
- }
84
+ if (unlikely(!mem)) {
85
/*
86
- * OK, try a store and see if we can populate the tlb. This
87
- * might cause an exception if the memory isn't writable,
88
- * in which case we will longjmp out of here. We must for
89
- * this purpose use the actual register value passed to us
90
- * so that we get the fault address right.
91
+ * The only remaining reason for mem == NULL is I/O.
92
+ * Just do a series of byte writes as the architecture demands.
93
*/
94
- helper_ret_stb_mmu(env, vaddr_in, 0, oi, GETPC());
95
- /* Now we can populate the other TLB entries, if any */
96
- for (i = 0; i < maxidx; i++) {
97
- uint64_t va = vaddr + TARGET_PAGE_SIZE * i;
98
- if (va != (vaddr_in & TARGET_PAGE_MASK)) {
99
- helper_ret_stb_mmu(env, va, 0, oi, GETPC());
100
- }
101
+ for (int i = 0; i < blocklen; i++) {
102
+ cpu_stb_mmuidx_ra(env, vaddr + i, 0, mmu_idx, ra);
103
}
104
- }
105
-
106
- /*
107
- * Slow path (probably attempt to do this to an I/O device or
108
- * similar, or clearing of a block of code we have translations
109
- * cached for). Just do a series of byte writes as the architecture
110
- * demands. It's not worth trying to use a cpu_physical_memory_map(),
111
- * memset(), unmap() sequence here because:
112
- * + we'd need to account for the blocksize being larger than a page
113
- * + the direct-RAM access case is almost always going to be dealt
114
- * with in the fastpath code above, so there's no speed benefit
115
- * + we would have to deal with the map returning NULL because the
116
- * bounce buffer was in use
117
- */
118
- for (i = 0; i < blocklen; i++) {
119
- helper_ret_stb_mmu(env, vaddr + i, 0, oi, GETPC());
120
+ return;
121
}
122
}
123
-#else
124
- memset(g2h(vaddr), 0, blocklen);
125
#endif
126
+
127
+ memset(mem, 0, blocklen);
128
}
67
--
129
--
68
2.19.1
130
2.20.1
69
131
70
132
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move mla_op and mls_op expanders from translate-a64.c.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181011205206.3552-16-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-20-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
7
---
10
target/arm/translate.h | 2 +
8
target/arm/helper-a64.h | 3 ++
11
target/arm/translate-a64.c | 106 -----------------------------
9
target/arm/translate.h | 2 +
12
target/arm/translate.c | 134 ++++++++++++++++++++++++++++++++-----
10
target/arm/mte_helper.c | 84 ++++++++++++++++++++++++++++++++++++++
13
3 files changed, 120 insertions(+), 122 deletions(-)
11
target/arm/translate-a64.c | 72 ++++++++++++++++++++++++++++----
14
12
4 files changed, 153 insertions(+), 8 deletions(-)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
19
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
20
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
21
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
22
+DEF_HELPER_FLAGS_2(ldgm, TCG_CALL_NO_WG, i64, env, i64)
23
+DEF_HELPER_FLAGS_3(stgm, TCG_CALL_NO_WG, void, env, i64, i64)
24
+DEF_HELPER_FLAGS_3(stzgm_tags, TCG_CALL_NO_WG, void, env, i64, i64)
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
16
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.h
27
--- a/target/arm/translate.h
18
+++ b/target/arm/translate.h
28
+++ b/target/arm/translate.h
19
@@ -XXX,XX +XXX,XX @@ static inline TCGv_i32 get_ahp_flag(void)
29
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
20
extern const GVecGen3 bsl_op;
30
* < 0, set by the current instruction.
21
extern const GVecGen3 bit_op;
31
*/
22
extern const GVecGen3 bif_op;
32
int8_t btype;
23
+extern const GVecGen3 mla_op[4];
33
+ /* A copy of cpu->dcz_blocksize. */
24
+extern const GVecGen3 mls_op[4];
34
+ uint8_t dcz_blocksize;
25
extern const GVecGen2i ssra_op[4];
35
/* True if this page is guarded. */
26
extern const GVecGen2i usra_op[4];
36
bool guarded_page;
27
extern const GVecGen2i sri_op[4];
37
/* Bottom two bits of XScale c15_cpar coprocessor access control reg */
38
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/target/arm/mte_helper.c
41
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@ void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
43
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
44
}
45
}
46
+
47
+#define LDGM_STGM_SIZE (4 << GMID_EL1_BS)
48
+
49
+uint64_t HELPER(ldgm)(CPUARMState *env, uint64_t ptr)
50
+{
51
+ int mmu_idx = cpu_mmu_index(env, false);
52
+ uintptr_t ra = GETPC();
53
+ void *tag_mem;
54
+
55
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
56
+
57
+ /* Trap if accessing an invalid page. */
58
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD,
59
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
60
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
61
+
62
+ /* The tag is squashed to zero if the page does not support tags. */
63
+ if (!tag_mem) {
64
+ return 0;
65
+ }
66
+
67
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
68
+ /*
69
+ * We are loading 64-bits worth of tags. The ordering of elements
70
+ * within the word corresponds to a 64-bit little-endian operation.
71
+ */
72
+ return ldq_le_p(tag_mem);
73
+}
74
+
75
+void HELPER(stgm)(CPUARMState *env, uint64_t ptr, uint64_t val)
76
+{
77
+ int mmu_idx = cpu_mmu_index(env, false);
78
+ uintptr_t ra = GETPC();
79
+ void *tag_mem;
80
+
81
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
82
+
83
+ /* Trap if accessing an invalid page. */
84
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
85
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
86
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
87
+
88
+ /*
89
+ * Tag store only happens if the page support tags,
90
+ * and if the OS has enabled access to the tags.
91
+ */
92
+ if (!tag_mem) {
93
+ return;
94
+ }
95
+
96
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
97
+ /*
98
+ * We are storing 64-bits worth of tags. The ordering of elements
99
+ * within the word corresponds to a 64-bit little-endian operation.
100
+ */
101
+ stq_le_p(tag_mem, val);
102
+}
103
+
104
+void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
105
+{
106
+ uintptr_t ra = GETPC();
107
+ int mmu_idx = cpu_mmu_index(env, false);
108
+ int log2_dcz_bytes, log2_tag_bytes;
109
+ intptr_t dcz_bytes, tag_bytes;
110
+ uint8_t *mem;
111
+
112
+ /*
113
+ * In arm_cpu_realizefn, we assert that dcz > LOG2_TAG_GRANULE+1,
114
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway,
115
+ * to make sure that we can access one complete tag byte here.
116
+ */
117
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
118
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
119
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
120
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
121
+ ptr &= -dcz_bytes;
122
+
123
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, dcz_bytes,
124
+ MMU_DATA_STORE, tag_bytes, ra);
125
+ if (mem) {
126
+ int tag_pair = (val & 0xf) * 0x11;
127
+ memset(mem, tag_pair, tag_bytes);
128
+ }
129
+}
28
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
130
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
29
index XXXXXXX..XXXXXXX 100644
131
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/translate-a64.c
132
--- a/target/arm/translate-a64.c
31
+++ b/target/arm/translate-a64.c
133
+++ b/target/arm/translate-a64.c
32
@@ -XXX,XX +XXX,XX @@ static void disas_simd_3same_float(DisasContext *s, uint32_t insn)
134
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
33
}
135
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
34
}
136
int op2 = extract32(insn, 10, 2);
35
137
int op1 = extract32(insn, 22, 2);
36
-static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
138
- bool is_load = false, is_pair = false, is_zero = false;
37
-{
139
+ bool is_load = false, is_pair = false, is_zero = false, is_mult = false;
38
- gen_helper_neon_mul_u8(a, a, b);
140
int index = 0;
39
- gen_helper_neon_add_u8(d, d, a);
141
TCGv_i64 addr, clean_addr, tcg_rt;
40
-}
142
41
-
143
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
42
-static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
144
if (op2 != 0) {
43
-{
145
/* STG */
44
- gen_helper_neon_mul_u16(a, a, b);
146
index = op2 - 2;
45
- gen_helper_neon_add_u16(d, d, a);
147
- break;
46
-}
148
+ } else {
47
-
149
+ /* STZGM */
48
-static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
150
+ if (s->current_el == 0 || offset != 0) {
49
-{
151
+ goto do_unallocated;
50
- tcg_gen_mul_i32(a, a, b);
152
+ }
51
- tcg_gen_add_i32(d, d, a);
153
+ is_mult = is_zero = true;
52
-}
53
-
54
-static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
55
-{
56
- tcg_gen_mul_i64(a, a, b);
57
- tcg_gen_add_i64(d, d, a);
58
-}
59
-
60
-static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
61
-{
62
- tcg_gen_mul_vec(vece, a, a, b);
63
- tcg_gen_add_vec(vece, d, d, a);
64
-}
65
-
66
-static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
67
-{
68
- gen_helper_neon_mul_u8(a, a, b);
69
- gen_helper_neon_sub_u8(d, d, a);
70
-}
71
-
72
-static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
73
-{
74
- gen_helper_neon_mul_u16(a, a, b);
75
- gen_helper_neon_sub_u16(d, d, a);
76
-}
77
-
78
-static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
79
-{
80
- tcg_gen_mul_i32(a, a, b);
81
- tcg_gen_sub_i32(d, d, a);
82
-}
83
-
84
-static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
85
-{
86
- tcg_gen_mul_i64(a, a, b);
87
- tcg_gen_sub_i64(d, d, a);
88
-}
89
-
90
-static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
91
-{
92
- tcg_gen_mul_vec(vece, a, a, b);
93
- tcg_gen_sub_vec(vece, d, d, a);
94
-}
95
-
96
/* Integer op subgroup of C3.6.16. */
97
static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
98
{
99
@@ -XXX,XX +XXX,XX @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
100
.prefer_i64 = TCG_TARGET_REG_BITS == 64,
101
.vece = MO_64 },
102
};
103
- static const GVecGen3 mla_op[4] = {
104
- { .fni4 = gen_mla8_i32,
105
- .fniv = gen_mla_vec,
106
- .opc = INDEX_op_mul_vec,
107
- .load_dest = true,
108
- .vece = MO_8 },
109
- { .fni4 = gen_mla16_i32,
110
- .fniv = gen_mla_vec,
111
- .opc = INDEX_op_mul_vec,
112
- .load_dest = true,
113
- .vece = MO_16 },
114
- { .fni4 = gen_mla32_i32,
115
- .fniv = gen_mla_vec,
116
- .opc = INDEX_op_mul_vec,
117
- .load_dest = true,
118
- .vece = MO_32 },
119
- { .fni8 = gen_mla64_i64,
120
- .fniv = gen_mla_vec,
121
- .opc = INDEX_op_mul_vec,
122
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
123
- .load_dest = true,
124
- .vece = MO_64 },
125
- };
126
- static const GVecGen3 mls_op[4] = {
127
- { .fni4 = gen_mls8_i32,
128
- .fniv = gen_mls_vec,
129
- .opc = INDEX_op_mul_vec,
130
- .load_dest = true,
131
- .vece = MO_8 },
132
- { .fni4 = gen_mls16_i32,
133
- .fniv = gen_mls_vec,
134
- .opc = INDEX_op_mul_vec,
135
- .load_dest = true,
136
- .vece = MO_16 },
137
- { .fni4 = gen_mls32_i32,
138
- .fniv = gen_mls_vec,
139
- .opc = INDEX_op_mul_vec,
140
- .load_dest = true,
141
- .vece = MO_32 },
142
- { .fni8 = gen_mls64_i64,
143
- .fniv = gen_mls_vec,
144
- .opc = INDEX_op_mul_vec,
145
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
146
- .load_dest = true,
147
- .vece = MO_64 },
148
- };
149
150
int is_q = extract32(insn, 30, 1);
151
int u = extract32(insn, 29, 1);
152
diff --git a/target/arm/translate.c b/target/arm/translate.c
153
index XXXXXXX..XXXXXXX 100644
154
--- a/target/arm/translate.c
155
+++ b/target/arm/translate.c
156
@@ -XXX,XX +XXX,XX @@ static void gen_neon_narrow_op(int op, int u, int size,
157
#define NEON_3R_VABA 15
158
#define NEON_3R_VADD_VSUB 16
159
#define NEON_3R_VTST_VCEQ 17
160
-#define NEON_3R_VML 18 /* VMLA, VMLAL, VMLS, VMLSL */
161
+#define NEON_3R_VML 18 /* VMLA, VMLS */
162
#define NEON_3R_VMUL 19
163
#define NEON_3R_VPMAX 20
164
#define NEON_3R_VPMIN 21
165
@@ -XXX,XX +XXX,XX @@ const GVecGen2i sli_op[4] = {
166
.vece = MO_64 },
167
};
168
169
+static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
170
+{
171
+ gen_helper_neon_mul_u8(a, a, b);
172
+ gen_helper_neon_add_u8(d, d, a);
173
+}
174
+
175
+static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
176
+{
177
+ gen_helper_neon_mul_u8(a, a, b);
178
+ gen_helper_neon_sub_u8(d, d, a);
179
+}
180
+
181
+static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
182
+{
183
+ gen_helper_neon_mul_u16(a, a, b);
184
+ gen_helper_neon_add_u16(d, d, a);
185
+}
186
+
187
+static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
188
+{
189
+ gen_helper_neon_mul_u16(a, a, b);
190
+ gen_helper_neon_sub_u16(d, d, a);
191
+}
192
+
193
+static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
194
+{
195
+ tcg_gen_mul_i32(a, a, b);
196
+ tcg_gen_add_i32(d, d, a);
197
+}
198
+
199
+static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
200
+{
201
+ tcg_gen_mul_i32(a, a, b);
202
+ tcg_gen_sub_i32(d, d, a);
203
+}
204
+
205
+static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
206
+{
207
+ tcg_gen_mul_i64(a, a, b);
208
+ tcg_gen_add_i64(d, d, a);
209
+}
210
+
211
+static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
212
+{
213
+ tcg_gen_mul_i64(a, a, b);
214
+ tcg_gen_sub_i64(d, d, a);
215
+}
216
+
217
+static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
218
+{
219
+ tcg_gen_mul_vec(vece, a, a, b);
220
+ tcg_gen_add_vec(vece, d, d, a);
221
+}
222
+
223
+static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
224
+{
225
+ tcg_gen_mul_vec(vece, a, a, b);
226
+ tcg_gen_sub_vec(vece, d, d, a);
227
+}
228
+
229
+/* Note that while NEON does not support VMLA and VMLS as 64-bit ops,
230
+ * these tables are shared with AArch64 which does support them.
231
+ */
232
+const GVecGen3 mla_op[4] = {
233
+ { .fni4 = gen_mla8_i32,
234
+ .fniv = gen_mla_vec,
235
+ .opc = INDEX_op_mul_vec,
236
+ .load_dest = true,
237
+ .vece = MO_8 },
238
+ { .fni4 = gen_mla16_i32,
239
+ .fniv = gen_mla_vec,
240
+ .opc = INDEX_op_mul_vec,
241
+ .load_dest = true,
242
+ .vece = MO_16 },
243
+ { .fni4 = gen_mla32_i32,
244
+ .fniv = gen_mla_vec,
245
+ .opc = INDEX_op_mul_vec,
246
+ .load_dest = true,
247
+ .vece = MO_32 },
248
+ { .fni8 = gen_mla64_i64,
249
+ .fniv = gen_mla_vec,
250
+ .opc = INDEX_op_mul_vec,
251
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
252
+ .load_dest = true,
253
+ .vece = MO_64 },
254
+};
255
+
256
+const GVecGen3 mls_op[4] = {
257
+ { .fni4 = gen_mls8_i32,
258
+ .fniv = gen_mls_vec,
259
+ .opc = INDEX_op_mul_vec,
260
+ .load_dest = true,
261
+ .vece = MO_8 },
262
+ { .fni4 = gen_mls16_i32,
263
+ .fniv = gen_mls_vec,
264
+ .opc = INDEX_op_mul_vec,
265
+ .load_dest = true,
266
+ .vece = MO_16 },
267
+ { .fni4 = gen_mls32_i32,
268
+ .fniv = gen_mls_vec,
269
+ .opc = INDEX_op_mul_vec,
270
+ .load_dest = true,
271
+ .vece = MO_32 },
272
+ { .fni8 = gen_mls64_i64,
273
+ .fniv = gen_mls_vec,
274
+ .opc = INDEX_op_mul_vec,
275
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
276
+ .load_dest = true,
277
+ .vece = MO_64 },
278
+};
279
+
280
/* Translate a NEON data processing instruction. Return nonzero if the
281
instruction is invalid.
282
We process data in a mixture of 32-bit and 64-bit chunks.
283
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
284
return 0;
285
}
286
break;
287
+
288
+ case NEON_3R_VML: /* VMLA, VMLS */
289
+ tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs, vec_size, vec_size,
290
+ u ? &mls_op[size] : &mla_op[size]);
291
+ return 0;
292
}
154
}
293
+
155
- goto do_unallocated;
294
if (size == 3) {
156
+ break;
295
/* 64-bit element instructions. */
157
case 1:
296
for (pass = 0; pass < (q ? 2 : 1); pass++) {
158
if (op2 != 0) {
297
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
159
/* STZG */
298
}
160
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
299
}
161
/* ST2G */
300
break;
162
is_pair = true;
301
- case NEON_3R_VML: /* VMLA, VMLAL, VMLS,VMLSL */
163
index = op2 - 2;
302
- switch (size) {
303
- case 0: gen_helper_neon_mul_u8(tmp, tmp, tmp2); break;
304
- case 1: gen_helper_neon_mul_u16(tmp, tmp, tmp2); break;
305
- case 2: tcg_gen_mul_i32(tmp, tmp, tmp2); break;
306
- default: abort();
307
- }
308
- tcg_temp_free_i32(tmp2);
309
- tmp2 = neon_load_reg(rd, pass);
310
- if (u) { /* VMLS */
311
- gen_neon_rsb(size, tmp, tmp2);
312
- } else { /* VMLA */
313
- gen_neon_add(size, tmp, tmp2);
314
- }
315
- break;
164
- break;
316
case NEON_3R_VMUL:
165
+ } else {
317
/* VMUL.P8; other cases already eliminated. */
166
+ /* STGM */
318
gen_helper_neon_mul_p8(tmp, tmp, tmp2);
167
+ if (s->current_el == 0 || offset != 0) {
168
+ goto do_unallocated;
169
+ }
170
+ is_mult = true;
171
}
172
- goto do_unallocated;
173
+ break;
174
case 3:
175
if (op2 != 0) {
176
/* STZ2G */
177
is_pair = is_zero = true;
178
index = op2 - 2;
179
- break;
180
+ } else {
181
+ /* LDGM */
182
+ if (s->current_el == 0 || offset != 0) {
183
+ goto do_unallocated;
184
+ }
185
+ is_mult = is_load = true;
186
}
187
- goto do_unallocated;
188
+ break;
189
190
default:
191
do_unallocated:
192
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
193
return;
194
}
195
196
- if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
197
+ if (is_mult
198
+ ? !dc_isar_feature(aa64_mte, s)
199
+ : !dc_isar_feature(aa64_mte_insn_reg, s)) {
200
goto do_unallocated;
201
}
202
203
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
204
tcg_gen_addi_i64(addr, addr, offset);
205
}
206
207
+ if (is_mult) {
208
+ tcg_rt = cpu_reg(s, rt);
209
+
210
+ if (is_zero) {
211
+ int size = 4 << s->dcz_blocksize;
212
+
213
+ if (s->ata) {
214
+ gen_helper_stzgm_tags(cpu_env, addr, tcg_rt);
215
+ }
216
+ /*
217
+ * The non-tags portion of STZGM is mostly like DC_ZVA,
218
+ * except the alignment happens before the access.
219
+ */
220
+ clean_addr = clean_data_tbi(s, addr);
221
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
222
+ gen_helper_dc_zva(cpu_env, clean_addr);
223
+ } else if (s->ata) {
224
+ if (is_load) {
225
+ gen_helper_ldgm(tcg_rt, cpu_env, addr);
226
+ } else {
227
+ gen_helper_stgm(cpu_env, addr, tcg_rt);
228
+ }
229
+ } else {
230
+ MMUAccessType acc = is_load ? MMU_DATA_LOAD : MMU_DATA_STORE;
231
+ int size = 4 << GMID_EL1_BS;
232
+
233
+ clean_addr = clean_data_tbi(s, addr);
234
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
235
+ gen_probe_access(s, clean_addr, acc, size);
236
+
237
+ if (is_load) {
238
+ /* The result tags are zeros. */
239
+ tcg_gen_movi_i64(tcg_rt, 0);
240
+ }
241
+ }
242
+ return;
243
+ }
244
+
245
if (is_load) {
246
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
247
tcg_rt = cpu_reg(s, rt);
248
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
249
dc->vec_stride = 0;
250
dc->cp_regs = arm_cpu->cp_regs;
251
dc->features = env->features;
252
+ dc->dcz_blocksize = arm_cpu->dcz_blocksize;
253
254
/* Single step state. The code-generation logic here is:
255
* SS_ACTIVE == 0:
319
--
256
--
320
2.19.1
257
2.20.1
321
258
322
259
diff view generated by jsdifflib
1
If the HCR_EL2 PTW virtualizaiton configuration register bit
1
From: Richard Henderson <richard.henderson@linaro.org>
2
is set, then this means that a stage 2 Permission fault must
3
be generated if a stage 1 translation table access is made
4
to an address that is mapped as Device memory in stage 2.
5
Implement this.
6
2
3
Like the regular data cache flushes, these are nops within qemu.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-21-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20181012144235.19646-8-peter.maydell@linaro.org
10
---
9
---
11
target/arm/helper.c | 21 ++++++++++++++++++++-
10
target/arm/helper.c | 65 +++++++++++++++++++++++++++++++++++++++++++++
12
1 file changed, 20 insertions(+), 1 deletion(-)
11
1 file changed, 65 insertions(+)
13
12
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
17
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
18
@@ -XXX,XX +XXX,XX @@ static hwaddr S1_ptw_translate(CPUARMState *env, ARMMMUIdx mmu_idx,
17
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
19
hwaddr s2pa;
18
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
20
int s2prot;
19
.type = ARM_CP_NO_RAW,
21
int ret;
20
.access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
22
+ ARMCacheAttrs cacheattrs = {};
21
+ { .name = "DC_IGVAC", .state = ARM_CP_STATE_AA64,
23
+ ARMCacheAttrs *pcacheattrs = NULL;
22
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 3,
23
+ .type = ARM_CP_NOP, .access = PL1_W,
24
+ .accessfn = aa64_cacheop_poc_access },
25
+ { .name = "DC_IGSW", .state = ARM_CP_STATE_AA64,
26
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 4,
27
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
28
+ { .name = "DC_IGDVAC", .state = ARM_CP_STATE_AA64,
29
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 5,
30
+ .type = ARM_CP_NOP, .access = PL1_W,
31
+ .accessfn = aa64_cacheop_poc_access },
32
+ { .name = "DC_IGDSW", .state = ARM_CP_STATE_AA64,
33
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 6,
34
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
35
+ { .name = "DC_CGSW", .state = ARM_CP_STATE_AA64,
36
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 4,
37
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
38
+ { .name = "DC_CGDSW", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 6,
40
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
41
+ { .name = "DC_CIGSW", .state = ARM_CP_STATE_AA64,
42
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 4,
43
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
44
+ { .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
45
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
46
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
47
REGINFO_SENTINEL
48
};
49
50
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
51
.type = ARM_CP_CONST, .access = PL0_RW, },
52
REGINFO_SENTINEL
53
};
24
+
54
+
25
+ if (env->cp15.hcr_el2 & HCR_PTW) {
55
+static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
26
+ /*
56
+ { .name = "DC_CGVAC", .state = ARM_CP_STATE_AA64,
27
+ * PTW means we must fault if this S1 walk touches S2 Device
57
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 3,
28
+ * memory; otherwise we don't care about the attributes and can
58
+ .type = ARM_CP_NOP, .access = PL0_W,
29
+ * save the S2 translation the effort of computing them.
59
+ .accessfn = aa64_cacheop_poc_access },
30
+ */
60
+ { .name = "DC_CGDVAC", .state = ARM_CP_STATE_AA64,
31
+ pcacheattrs = &cacheattrs;
61
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 5,
32
+ }
62
+ .type = ARM_CP_NOP, .access = PL0_W,
33
63
+ .accessfn = aa64_cacheop_poc_access },
34
ret = get_phys_addr_lpae(env, addr, 0, ARMMMUIdx_S2NS, &s2pa,
64
+ { .name = "DC_CGVAP", .state = ARM_CP_STATE_AA64,
35
- &txattrs, &s2prot, &s2size, fi, NULL);
65
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 3,
36
+ &txattrs, &s2prot, &s2size, fi, pcacheattrs);
66
+ .type = ARM_CP_NOP, .access = PL0_W,
37
if (ret) {
67
+ .accessfn = aa64_cacheop_poc_access },
38
assert(fi->type != ARMFault_None);
68
+ { .name = "DC_CGDVAP", .state = ARM_CP_STATE_AA64,
39
fi->s2addr = addr;
69
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 5,
40
@@ -XXX,XX +XXX,XX @@ static hwaddr S1_ptw_translate(CPUARMState *env, ARMMMUIdx mmu_idx,
70
+ .type = ARM_CP_NOP, .access = PL0_W,
41
fi->s1ptw = true;
71
+ .accessfn = aa64_cacheop_poc_access },
42
return ~0;
72
+ { .name = "DC_CGVADP", .state = ARM_CP_STATE_AA64,
43
}
73
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 3,
44
+ if (pcacheattrs && (pcacheattrs->attrs & 0xf0) == 0) {
74
+ .type = ARM_CP_NOP, .access = PL0_W,
45
+ /* Access was to Device memory: generate Permission fault */
75
+ .accessfn = aa64_cacheop_poc_access },
46
+ fi->type = ARMFault_Permission;
76
+ { .name = "DC_CGDVADP", .state = ARM_CP_STATE_AA64,
47
+ fi->s2addr = addr;
77
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 5,
48
+ fi->stage2 = true;
78
+ .type = ARM_CP_NOP, .access = PL0_W,
49
+ fi->s1ptw = true;
79
+ .accessfn = aa64_cacheop_poc_access },
50
+ return ~0;
80
+ { .name = "DC_CIGVAC", .state = ARM_CP_STATE_AA64,
51
+ }
81
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 3,
52
addr = s2pa;
82
+ .type = ARM_CP_NOP, .access = PL0_W,
83
+ .accessfn = aa64_cacheop_poc_access },
84
+ { .name = "DC_CIGDVAC", .state = ARM_CP_STATE_AA64,
85
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
86
+ .type = ARM_CP_NOP, .access = PL0_W,
87
+ .accessfn = aa64_cacheop_poc_access },
88
+ REGINFO_SENTINEL
89
+};
90
+
91
#endif
92
93
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
94
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
95
*/
96
if (cpu_isar_feature(aa64_mte, cpu)) {
97
define_arm_cp_regs(cpu, mte_reginfo);
98
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
99
} else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
100
define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
101
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
53
}
102
}
54
return addr;
103
#endif
104
55
--
105
--
56
2.19.1
106
2.20.1
57
107
58
108
diff view generated by jsdifflib
1
Create and use a utility function to extract the EC field
1
From: Richard Henderson <richard.henderson@linaro.org>
2
from a syndrome, rather than open-coding the shift.
3
2
3
We will shortly need this in mte_helper.c as well.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-22-richard.henderson@linaro.org
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181012144235.19646-9-peter.maydell@linaro.org
7
---
9
---
8
target/arm/internals.h | 5 +++++
10
target/arm/internals.h | 36 ++++++++++++++++++++++++++++++++++++
9
target/arm/helper.c | 4 ++--
11
target/arm/helper.c | 36 ------------------------------------
10
target/arm/kvm64.c | 2 +-
12
2 files changed, 36 insertions(+), 36 deletions(-)
11
target/arm/op_helper.c | 2 +-
12
4 files changed, 9 insertions(+), 4 deletions(-)
13
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ enum arm_exception_class {
18
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_pan(CPUARMState *env, ARMMMUIdx mmu_idx)
19
#define ARM_EL_IL (1 << ARM_EL_IL_SHIFT)
19
}
20
#define ARM_EL_ISV (1 << ARM_EL_ISV_SHIFT)
20
}
21
21
22
+static inline uint32_t syn_get_ec(uint32_t syn)
22
+/* Return the exception level which controls this address translation regime */
23
+static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
23
+{
24
+{
24
+ return syn >> ARM_EL_EC_SHIFT;
25
+ switch (mmu_idx) {
26
+ case ARMMMUIdx_E20_0:
27
+ case ARMMMUIdx_E20_2:
28
+ case ARMMMUIdx_E20_2_PAN:
29
+ case ARMMMUIdx_Stage2:
30
+ case ARMMMUIdx_E2:
31
+ return 2;
32
+ case ARMMMUIdx_SE3:
33
+ return 3;
34
+ case ARMMMUIdx_SE10_0:
35
+ return arm_el_is_aa64(env, 3) ? 1 : 3;
36
+ case ARMMMUIdx_SE10_1:
37
+ case ARMMMUIdx_SE10_1_PAN:
38
+ case ARMMMUIdx_Stage1_E0:
39
+ case ARMMMUIdx_Stage1_E1:
40
+ case ARMMMUIdx_Stage1_E1_PAN:
41
+ case ARMMMUIdx_E10_0:
42
+ case ARMMMUIdx_E10_1:
43
+ case ARMMMUIdx_E10_1_PAN:
44
+ case ARMMMUIdx_MPrivNegPri:
45
+ case ARMMMUIdx_MUserNegPri:
46
+ case ARMMMUIdx_MPriv:
47
+ case ARMMMUIdx_MUser:
48
+ case ARMMMUIdx_MSPrivNegPri:
49
+ case ARMMMUIdx_MSUserNegPri:
50
+ case ARMMMUIdx_MSPriv:
51
+ case ARMMMUIdx_MSUser:
52
+ return 1;
53
+ default:
54
+ g_assert_not_reached();
55
+ }
25
+}
56
+}
26
+
57
+
27
/* Utility functions for constructing various kinds of syndrome value.
58
/* Return the FSR value for a debug exception (watchpoint, hardware
28
* Note that in general we follow the AArch64 syndrome values; in a
59
* breakpoint or BKPT insn) targeting the specified exception level.
29
* few cases the value in HSR for exceptions taken to AArch32 Hyp
60
*/
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
61
diff --git a/target/arm/helper.c b/target/arm/helper.c
31
index XXXXXXX..XXXXXXX 100644
62
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.c
63
--- a/target/arm/helper.c
33
+++ b/target/arm/helper.c
64
+++ b/target/arm/helper.c
34
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch32(CPUState *cs)
35
uint32_t moe;
36
37
/* If this is a debug exception we must update the DBGDSCR.MOE bits */
38
- switch (env->exception.syndrome >> ARM_EL_EC_SHIFT) {
39
+ switch (syn_get_ec(env->exception.syndrome)) {
40
case EC_BREAKPOINT:
41
case EC_BREAKPOINT_SAME_EL:
42
moe = 1;
43
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
65
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
44
if (qemu_loglevel_mask(CPU_LOG_INT)
66
}
45
&& !excp_is_internal(cs->exception_index)) {
67
#endif /* !CONFIG_USER_ONLY */
46
qemu_log_mask(CPU_LOG_INT, "...with ESR 0x%x/0x%" PRIx32 "\n",
68
47
- env->exception.syndrome >> ARM_EL_EC_SHIFT,
69
-/* Return the exception level which controls this address translation regime */
48
+ syn_get_ec(env->exception.syndrome),
70
-static uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
49
env->exception.syndrome);
71
-{
50
}
72
- switch (mmu_idx) {
51
73
- case ARMMMUIdx_E20_0:
52
diff --git a/target/arm/kvm64.c b/target/arm/kvm64.c
74
- case ARMMMUIdx_E20_2:
53
index XXXXXXX..XXXXXXX 100644
75
- case ARMMMUIdx_E20_2_PAN:
54
--- a/target/arm/kvm64.c
76
- case ARMMMUIdx_Stage2:
55
+++ b/target/arm/kvm64.c
77
- case ARMMMUIdx_E2:
56
@@ -XXX,XX +XXX,XX @@ int kvm_arch_remove_sw_breakpoint(CPUState *cs, struct kvm_sw_breakpoint *bp)
78
- return 2;
57
79
- case ARMMMUIdx_SE3:
58
bool kvm_arm_handle_debug(CPUState *cs, struct kvm_debug_exit_arch *debug_exit)
80
- return 3;
81
- case ARMMMUIdx_SE10_0:
82
- return arm_el_is_aa64(env, 3) ? 1 : 3;
83
- case ARMMMUIdx_SE10_1:
84
- case ARMMMUIdx_SE10_1_PAN:
85
- case ARMMMUIdx_Stage1_E0:
86
- case ARMMMUIdx_Stage1_E1:
87
- case ARMMMUIdx_Stage1_E1_PAN:
88
- case ARMMMUIdx_E10_0:
89
- case ARMMMUIdx_E10_1:
90
- case ARMMMUIdx_E10_1_PAN:
91
- case ARMMMUIdx_MPrivNegPri:
92
- case ARMMMUIdx_MUserNegPri:
93
- case ARMMMUIdx_MPriv:
94
- case ARMMMUIdx_MUser:
95
- case ARMMMUIdx_MSPrivNegPri:
96
- case ARMMMUIdx_MSUserNegPri:
97
- case ARMMMUIdx_MSPriv:
98
- case ARMMMUIdx_MSUser:
99
- return 1;
100
- default:
101
- g_assert_not_reached();
102
- }
103
-}
104
-
105
uint64_t arm_sctlr(CPUARMState *env, int el)
59
{
106
{
60
- int hsr_ec = debug_exit->hsr >> ARM_EL_EC_SHIFT;
107
/* Only EL0 needs to be adjusted for EL1&0 or EL2&0. */
61
+ int hsr_ec = syn_get_ec(debug_exit->hsr);
62
ARMCPU *cpu = ARM_CPU(cs);
63
CPUClass *cc = CPU_GET_CLASS(cs);
64
CPUARMState *env = &cpu->env;
65
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
66
index XXXXXXX..XXXXXXX 100644
67
--- a/target/arm/op_helper.c
68
+++ b/target/arm/op_helper.c
69
@@ -XXX,XX +XXX,XX @@ void raise_exception(CPUARMState *env, uint32_t excp,
70
* (see DDI0478C.a D1.10.4)
71
*/
72
target_el = 2;
73
- if (syndrome >> ARM_EL_EC_SHIFT == EC_ADVSIMDFPACCESSTRAP) {
74
+ if (syn_get_ec(syndrome) == EC_ADVSIMDFPACCESSTRAP) {
75
syndrome = syn_uncategorized();
76
}
77
}
78
--
108
--
79
2.19.1
109
2.20.1
80
110
81
111
diff view generated by jsdifflib
1
For AArch32, exception return happens through certain kinds
1
From: Richard Henderson <richard.henderson@linaro.org>
2
of CPSR write. We don't currently have any CPU_LOG_INT logging
3
of these events (unlike AArch64, where we log in the ERET
4
instruction). Add some suitable logging.
5
2
6
This will log exception returns like this:
3
We will shortly need this in mte_helper.c as well.
7
Exception return from AArch32 hyp to usr PC 0x80100374
8
4
9
paralleling the existing logging in the exception_return
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
helper for AArch64 exception returns:
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Exception return from AArch64 EL2 to AArch64 EL0 PC 0x8003045c
7
Message-id: 20200626033144.790098-23-richard.henderson@linaro.org
12
Exception return from AArch64 EL2 to AArch32 EL0 PC 0x8003045c
13
14
(Note that an AArch32 exception return can only be
15
AArch32->AArch32, never to AArch64.)
16
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
19
Message-id: 20181012144235.19646-2-peter.maydell@linaro.org
20
---
9
---
21
target/arm/internals.h | 18 ++++++++++++++++++
10
target/arm/internals.h | 9 +++++++++
22
target/arm/helper.c | 10 ++++++++++
11
target/arm/helper.c | 9 ---------
23
target/arm/translate.c | 7 +------
12
2 files changed, 9 insertions(+), 9 deletions(-)
24
3 files changed, 29 insertions(+), 6 deletions(-)
25
13
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
27
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/internals.h
16
--- a/target/arm/internals.h
29
+++ b/target/arm/internals.h
17
+++ b/target/arm/internals.h
30
@@ -XXX,XX +XXX,XX @@ static inline uint32_t v7m_sp_limit(CPUARMState *env)
18
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
31
}
19
}
32
}
20
}
33
21
34
+/**
22
+/* Return the TCR controlling this translation regime */
35
+ * aarch32_mode_name(): Return name of the AArch32 CPU mode
23
+static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
36
+ * @psr: Program Status Register indicating CPU mode
37
+ *
38
+ * Returns, for debug logging purposes, a printable representation
39
+ * of the AArch32 CPU mode ("svc", "usr", etc) as indicated by
40
+ * the low bits of the specified PSR.
41
+ */
42
+static inline const char *aarch32_mode_name(uint32_t psr)
43
+{
24
+{
44
+ static const char cpu_mode_names[16][4] = {
25
+ if (mmu_idx == ARMMMUIdx_Stage2) {
45
+ "usr", "fiq", "irq", "svc", "???", "???", "mon", "abt",
26
+ return &env->cp15.vtcr_el2;
46
+ "???", "???", "hyp", "und", "???", "???", "???", "sys"
27
+ }
47
+ };
28
+ return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
48
+
49
+ return cpu_mode_names[psr & 0xf];
50
+}
29
+}
51
+
30
+
52
#endif
31
/* Return the FSR value for a debug exception (watchpoint, hardware
32
* breakpoint or BKPT insn) targeting the specified exception level.
33
*/
53
diff --git a/target/arm/helper.c b/target/arm/helper.c
34
diff --git a/target/arm/helper.c b/target/arm/helper.c
54
index XXXXXXX..XXXXXXX 100644
35
index XXXXXXX..XXXXXXX 100644
55
--- a/target/arm/helper.c
36
--- a/target/arm/helper.c
56
+++ b/target/arm/helper.c
37
+++ b/target/arm/helper.c
57
@@ -XXX,XX +XXX,XX @@ void cpsr_write(CPUARMState *env, uint32_t val, uint32_t mask,
38
@@ -XXX,XX +XXX,XX @@ static inline uint64_t regime_ttbr(CPUARMState *env, ARMMMUIdx mmu_idx,
58
mask |= CPSR_IL;
39
59
val |= CPSR_IL;
40
#endif /* !CONFIG_USER_ONLY */
60
}
41
61
+ qemu_log_mask(LOG_GUEST_ERROR,
42
-/* Return the TCR controlling this translation regime */
62
+ "Illegal AArch32 mode switch attempt from %s to %s\n",
43
-static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
63
+ aarch32_mode_name(env->uncached_cpsr),
44
-{
64
+ aarch32_mode_name(val));
45
- if (mmu_idx == ARMMMUIdx_Stage2) {
65
} else {
46
- return &env->cp15.vtcr_el2;
66
+ qemu_log_mask(CPU_LOG_INT, "%s %s to %s PC 0x%" PRIx32 "\n",
47
- }
67
+ write_type == CPSRWriteExceptionReturn ?
48
- return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
68
+ "Exception return from AArch32" :
49
-}
69
+ "AArch32 mode switch from",
70
+ aarch32_mode_name(env->uncached_cpsr),
71
+ aarch32_mode_name(val), env->regs[15]);
72
switch_mode(env, val & CPSR_M);
73
}
74
}
75
diff --git a/target/arm/translate.c b/target/arm/translate.c
76
index XXXXXXX..XXXXXXX 100644
77
--- a/target/arm/translate.c
78
+++ b/target/arm/translate.c
79
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb)
80
translator_loop(ops, &dc.base, cpu, tb);
81
}
82
83
-static const char *cpu_mode_names[16] = {
84
- "usr", "fiq", "irq", "svc", "???", "???", "mon", "abt",
85
- "???", "???", "hyp", "und", "???", "???", "???", "sys"
86
-};
87
-
50
-
88
void arm_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
51
/* Convert a possible stage1+2 MMU index into the appropriate
89
int flags)
52
* stage 1 MMU index
90
{
53
*/
91
@@ -XXX,XX +XXX,XX @@ void arm_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
92
psr & CPSR_V ? 'V' : '-',
93
psr & CPSR_T ? 'T' : 'A',
94
ns_status,
95
- cpu_mode_names[psr & 0xf], (psr & 0x10) ? 32 : 26);
96
+ aarch32_mode_name(psr), (psr & 0x10) ? 32 : 26);
97
}
98
99
if (flags & CPU_DUMP_FPU) {
100
--
54
--
101
2.19.1
55
2.20.1
102
56
103
57
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move ssra_op and usra_op expanders from translate-a64.c.
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
4
perform a single logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20181011205206.3552-14-richard.henderson@linaro.org
9
Message-id: 20200626033144.790098-24-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/translate.h | 2 +
12
target/arm/helper-a64.h | 1 +
11
target/arm/translate-a64.c | 106 ----------------------------
13
target/arm/internals.h | 8 +++
12
target/arm/translate.c | 139 ++++++++++++++++++++++++++++++++++---
14
target/arm/translate-a64.h | 2 +
13
3 files changed, 130 insertions(+), 117 deletions(-)
15
target/arm/mte_helper.c | 8 +++
14
16
target/arm/translate-a64.c | 100 ++++++++++++++++++++++++++++---------
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
17
5 files changed, 95 insertions(+), 24 deletions(-)
16
index XXXXXXX..XXXXXXX 100644
18
17
--- a/target/arm/translate.h
19
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
18
+++ b/target/arm/translate.h
20
index XXXXXXX..XXXXXXX 100644
19
@@ -XXX,XX +XXX,XX @@ static inline TCGv_i32 get_ahp_flag(void)
21
--- a/target/arm/helper-a64.h
20
extern const GVecGen3 bsl_op;
22
+++ b/target/arm/helper-a64.h
21
extern const GVecGen3 bit_op;
23
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
22
extern const GVecGen3 bif_op;
24
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
+extern const GVecGen2i ssra_op[4];
25
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
+extern const GVecGen2i usra_op[4];
26
25
27
+DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
26
/*
28
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
27
* Forward to the isar_feature_* tests given a DisasContext pointer.
29
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
30
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
31
diff --git a/target/arm/internals.h b/target/arm/internals.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/internals.h
34
+++ b/target/arm/internals.h
35
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
36
#define LOG2_TAG_GRANULE 4
37
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
38
39
+/* Bits within a descriptor passed to the helper_mte_check* functions. */
40
+FIELD(MTEDESC, MIDX, 0, 4)
41
+FIELD(MTEDESC, TBI, 4, 2)
42
+FIELD(MTEDESC, TCMA, 6, 2)
43
+FIELD(MTEDESC, WRITE, 8, 1)
44
+FIELD(MTEDESC, ESIZE, 9, 5)
45
+FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
46
+
47
static inline int allocation_tag_from_addr(uint64_t ptr)
48
{
49
return extract64(ptr, 56, 4);
50
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/translate-a64.h
53
+++ b/target/arm/translate-a64.h
54
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
55
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
56
unsigned int imms, unsigned int immr);
57
bool sve_access_check(DisasContext *s);
58
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
59
+ bool tag_checked, int log2_size);
60
61
/* We should have at some point before trying to access an FP register
62
* done the necessary access check, so assert that
63
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/mte_helper.c
66
+++ b/target/arm/mte_helper.c
67
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
68
memset(mem, tag_pair, tag_bytes);
69
}
70
}
71
+
72
+/*
73
+ * Perform an MTE checked access for a single logical or atomic access.
74
+ */
75
+uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
76
+{
77
+ return ptr;
78
+}
28
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
79
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
29
index XXXXXXX..XXXXXXX 100644
80
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/translate-a64.c
81
--- a/target/arm/translate-a64.c
31
+++ b/target/arm/translate-a64.c
82
+++ b/target/arm/translate-a64.c
32
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_two_reg_misc(DisasContext *s, uint32_t insn)
83
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
33
}
34
}
84
}
35
85
36
-static void gen_ssra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
86
/*
37
-{
87
- * Return a "clean" address for ADDR according to TBID.
38
- tcg_gen_vec_sar8i_i64(a, a, shift);
88
- * This is always a fresh temporary, as we need to be able to
39
- tcg_gen_vec_add8_i64(d, d, a);
89
- * increment this independently of a dirty write-back address.
40
-}
90
+ * Handle MTE and/or TBI.
41
-
91
+ *
42
-static void gen_ssra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
92
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
43
-{
93
+ * for the tag to be present in the FAR_ELx register. But for user-only
44
- tcg_gen_vec_sar16i_i64(a, a, shift);
94
+ * mode we do not have a TLB with which to implement this, so we must
45
- tcg_gen_vec_add16_i64(d, d, a);
95
+ * remove the top byte now.
46
-}
96
+ *
47
-
97
+ * Always return a fresh temporary that we can increment independently
48
-static void gen_ssra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
98
+ * of the write-back address.
49
-{
99
*/
50
- tcg_gen_sari_i32(a, a, shift);
100
+
51
- tcg_gen_add_i32(d, d, a);
101
static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
52
-}
53
-
54
-static void gen_ssra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
55
-{
56
- tcg_gen_sari_i64(a, a, shift);
57
- tcg_gen_add_i64(d, d, a);
58
-}
59
-
60
-static void gen_ssra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
61
-{
62
- tcg_gen_sari_vec(vece, a, a, sh);
63
- tcg_gen_add_vec(vece, d, d, a);
64
-}
65
-
66
-static void gen_usra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
67
-{
68
- tcg_gen_vec_shr8i_i64(a, a, shift);
69
- tcg_gen_vec_add8_i64(d, d, a);
70
-}
71
-
72
-static void gen_usra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
73
-{
74
- tcg_gen_vec_shr16i_i64(a, a, shift);
75
- tcg_gen_vec_add16_i64(d, d, a);
76
-}
77
-
78
-static void gen_usra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
79
-{
80
- tcg_gen_shri_i32(a, a, shift);
81
- tcg_gen_add_i32(d, d, a);
82
-}
83
-
84
-static void gen_usra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
85
-{
86
- tcg_gen_shri_i64(a, a, shift);
87
- tcg_gen_add_i64(d, d, a);
88
-}
89
-
90
-static void gen_usra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
91
-{
92
- tcg_gen_shri_vec(vece, a, a, sh);
93
- tcg_gen_add_vec(vece, d, d, a);
94
-}
95
-
96
static void gen_shr8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
97
{
102
{
98
uint64_t mask = dup_const(MO_8, 0xff >> shift);
103
TCGv_i64 clean = new_tmp_a64(s);
99
@@ -XXX,XX +XXX,XX @@ static void gen_shr_ins_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
104
- /*
100
static void handle_vec_simd_shri(DisasContext *s, bool is_q, bool is_u,
105
- * In order to get the correct value in the FAR_ELx register,
101
int immh, int immb, int opcode, int rn, int rd)
106
- * we must present the memory subsystem with the "dirty" address
102
{
107
- * including the TBI. In system mode we can make this work via
103
- static const GVecGen2i ssra_op[4] = {
108
- * the TLB, dropping the TBI during translation. But for user-only
104
- { .fni8 = gen_ssra8_i64,
109
- * mode we don't have that option, and must remove the top byte now.
105
- .fniv = gen_ssra_vec,
110
- */
106
- .load_dest = true,
111
#ifdef CONFIG_USER_ONLY
107
- .opc = INDEX_op_sari_vec,
112
gen_top_byte_ignore(s, clean, addr, s->tbid);
108
- .vece = MO_8 },
113
#else
109
- { .fni8 = gen_ssra16_i64,
114
@@ -XXX,XX +XXX,XX @@ static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
110
- .fniv = gen_ssra_vec,
115
tcg_temp_free_i32(t_size);
111
- .load_dest = true,
116
}
112
- .opc = INDEX_op_sari_vec,
117
113
- .vece = MO_16 },
118
+/*
114
- { .fni4 = gen_ssra32_i32,
119
+ * For MTE, check a single logical or atomic access. This probes a single
115
- .fniv = gen_ssra_vec,
120
+ * address, the exact one specified. The size and alignment of the access
116
- .load_dest = true,
121
+ * is not relevant to MTE, per se, but watchpoints do require the size,
117
- .opc = INDEX_op_sari_vec,
122
+ * and we want to recognize those before making any other changes to state.
118
- .vece = MO_32 },
123
+ */
119
- { .fni8 = gen_ssra64_i64,
124
+static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
120
- .fniv = gen_ssra_vec,
125
+ bool is_write, bool tag_checked,
121
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
126
+ int log2_size, bool is_unpriv,
122
- .load_dest = true,
127
+ int core_idx)
123
- .opc = INDEX_op_sari_vec,
124
- .vece = MO_64 },
125
- };
126
- static const GVecGen2i usra_op[4] = {
127
- { .fni8 = gen_usra8_i64,
128
- .fniv = gen_usra_vec,
129
- .load_dest = true,
130
- .opc = INDEX_op_shri_vec,
131
- .vece = MO_8, },
132
- { .fni8 = gen_usra16_i64,
133
- .fniv = gen_usra_vec,
134
- .load_dest = true,
135
- .opc = INDEX_op_shri_vec,
136
- .vece = MO_16, },
137
- { .fni4 = gen_usra32_i32,
138
- .fniv = gen_usra_vec,
139
- .load_dest = true,
140
- .opc = INDEX_op_shri_vec,
141
- .vece = MO_32, },
142
- { .fni8 = gen_usra64_i64,
143
- .fniv = gen_usra_vec,
144
- .prefer_i64 = TCG_TARGET_REG_BITS == 64,
145
- .load_dest = true,
146
- .opc = INDEX_op_shri_vec,
147
- .vece = MO_64, },
148
- };
149
static const GVecGen2i sri_op[4] = {
150
{ .fni8 = gen_shr8_ins_i64,
151
.fniv = gen_shr_ins_vec,
152
diff --git a/target/arm/translate.c b/target/arm/translate.c
153
index XXXXXXX..XXXXXXX 100644
154
--- a/target/arm/translate.c
155
+++ b/target/arm/translate.c
156
@@ -XXX,XX +XXX,XX @@ const GVecGen3 bif_op = {
157
.load_dest = true
158
};
159
160
+static void gen_ssra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
161
+{
128
+{
162
+ tcg_gen_vec_sar8i_i64(a, a, shift);
129
+ if (tag_checked && s->mte_active[is_unpriv]) {
163
+ tcg_gen_vec_add8_i64(d, d, a);
130
+ TCGv_i32 tcg_desc;
131
+ TCGv_i64 ret;
132
+ int desc = 0;
133
+
134
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, core_idx);
135
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
136
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
137
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
138
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_size);
139
+ tcg_desc = tcg_const_i32(desc);
140
+
141
+ ret = new_tmp_a64(s);
142
+ gen_helper_mte_check1(ret, cpu_env, tcg_desc, addr);
143
+ tcg_temp_free_i32(tcg_desc);
144
+
145
+ return ret;
146
+ }
147
+ return clean_data_tbi(s, addr);
164
+}
148
+}
165
+
149
+
166
+static void gen_ssra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
150
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
151
+ bool tag_checked, int log2_size)
167
+{
152
+{
168
+ tcg_gen_vec_sar16i_i64(a, a, shift);
153
+ return gen_mte_check1_mmuidx(s, addr, is_write, tag_checked, log2_size,
169
+ tcg_gen_vec_add16_i64(d, d, a);
154
+ false, get_mem_index(s));
170
+}
155
+}
171
+
156
+
172
+static void gen_ssra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
157
typedef struct DisasCompare64 {
173
+{
158
TCGCond cond;
174
+ tcg_gen_sari_i32(a, a, shift);
159
TCGv_i64 value;
175
+ tcg_gen_add_i32(d, d, a);
160
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap(DisasContext *s, int rs, int rt,
176
+}
161
if (rn == 31) {
177
+
162
gen_check_sp_alignment(s);
178
+static void gen_ssra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
163
}
179
+{
164
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
180
+ tcg_gen_sari_i64(a, a, shift);
165
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size);
181
+ tcg_gen_add_i64(d, d, a);
166
tcg_gen_atomic_cmpxchg_i64(tcg_rs, clean_addr, tcg_rs, tcg_rt, memidx,
182
+}
167
size | MO_ALIGN | s->be_data);
183
+
168
}
184
+static void gen_ssra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
169
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt,
185
+{
170
if (rn == 31) {
186
+ tcg_gen_sari_vec(vece, a, a, sh);
171
gen_check_sp_alignment(s);
187
+ tcg_gen_add_vec(vece, d, d, a);
172
}
188
+}
173
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
189
+
174
+
190
+const GVecGen2i ssra_op[4] = {
175
+ /* This is a single atomic access, despite the "pair". */
191
+ { .fni8 = gen_ssra8_i64,
176
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size + 1);
192
+ .fniv = gen_ssra_vec,
177
193
+ .load_dest = true,
178
if (size == 2) {
194
+ .opc = INDEX_op_sari_vec,
179
TCGv_i64 cmp = tcg_temp_new_i64();
195
+ .vece = MO_8 },
180
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
196
+ { .fni8 = gen_ssra16_i64,
181
if (is_lasr) {
197
+ .fniv = gen_ssra_vec,
182
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
198
+ .load_dest = true,
183
}
199
+ .opc = INDEX_op_sari_vec,
184
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
200
+ .vece = MO_16 },
185
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
201
+ { .fni4 = gen_ssra32_i32,
186
+ true, rn != 31, size);
202
+ .fniv = gen_ssra_vec,
187
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, false);
203
+ .load_dest = true,
188
return;
204
+ .opc = INDEX_op_sari_vec,
189
205
+ .vece = MO_32 },
190
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
206
+ { .fni8 = gen_ssra64_i64,
191
if (rn == 31) {
207
+ .fniv = gen_ssra_vec,
192
gen_check_sp_alignment(s);
208
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
193
}
209
+ .load_dest = true,
194
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
210
+ .opc = INDEX_op_sari_vec,
195
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
211
+ .vece = MO_64 },
196
+ false, rn != 31, size);
212
+};
197
s->is_ldex = true;
213
+
198
gen_load_exclusive(s, rt, rt2, clean_addr, size, false);
214
+static void gen_usra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
199
if (is_lasr) {
215
+{
200
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
216
+ tcg_gen_vec_shr8i_i64(a, a, shift);
201
gen_check_sp_alignment(s);
217
+ tcg_gen_vec_add8_i64(d, d, a);
202
}
218
+}
203
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
219
+
204
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
220
+static void gen_usra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
205
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
221
+{
206
+ true, rn != 31, size);
222
+ tcg_gen_vec_shr16i_i64(a, a, shift);
207
do_gpr_st(s, cpu_reg(s, rt), clean_addr, size, true, rt,
223
+ tcg_gen_vec_add16_i64(d, d, a);
208
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
224
+}
209
return;
225
+
210
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
226
+static void gen_usra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
211
if (rn == 31) {
227
+{
212
gen_check_sp_alignment(s);
228
+ tcg_gen_shri_i32(a, a, shift);
213
}
229
+ tcg_gen_add_i32(d, d, a);
214
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
230
+}
215
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
231
+
216
+ false, rn != 31, size);
232
+static void gen_usra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
217
do_gpr_ld(s, cpu_reg(s, rt), clean_addr, size, false, false, true, rt,
233
+{
218
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
234
+ tcg_gen_shri_i64(a, a, shift);
219
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
235
+ tcg_gen_add_i64(d, d, a);
220
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
236
+}
221
if (is_lasr) {
237
+
222
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
238
+static void gen_usra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
223
}
239
+{
224
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
240
+ tcg_gen_shri_vec(vece, a, a, sh);
225
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
241
+ tcg_gen_add_vec(vece, d, d, a);
226
+ true, rn != 31, size);
242
+}
227
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, true);
243
+
228
return;
244
+const GVecGen2i usra_op[4] = {
229
}
245
+ { .fni8 = gen_usra8_i64,
230
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
246
+ .fniv = gen_usra_vec,
231
if (rn == 31) {
247
+ .load_dest = true,
232
gen_check_sp_alignment(s);
248
+ .opc = INDEX_op_shri_vec,
233
}
249
+ .vece = MO_8, },
234
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
250
+ { .fni8 = gen_usra16_i64,
235
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
251
+ .fniv = gen_usra_vec,
236
+ false, rn != 31, size);
252
+ .load_dest = true,
237
s->is_ldex = true;
253
+ .opc = INDEX_op_shri_vec,
238
gen_load_exclusive(s, rt, rt2, clean_addr, size, true);
254
+ .vece = MO_16, },
239
if (is_lasr) {
255
+ { .fni4 = gen_usra32_i32,
240
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
256
+ .fniv = gen_usra_vec,
241
bool iss_valid = !is_vector;
257
+ .load_dest = true,
242
bool post_index;
258
+ .opc = INDEX_op_shri_vec,
243
bool writeback;
259
+ .vece = MO_32, },
244
+ int memidx;
260
+ { .fni8 = gen_usra64_i64,
245
261
+ .fniv = gen_usra_vec,
246
TCGv_i64 clean_addr, dirty_addr;
262
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64,
247
263
+ .load_dest = true,
248
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
264
+ .opc = INDEX_op_shri_vec,
249
if (!post_index) {
265
+ .vece = MO_64, },
250
tcg_gen_addi_i64(dirty_addr, dirty_addr, imm9);
266
+};
251
}
267
252
- clean_addr = clean_data_tbi(s, dirty_addr);
268
/* Translate a NEON data processing instruction. Return nonzero if the
253
+
269
instruction is invalid.
254
+ memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
270
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
255
+ clean_addr = gen_mte_check1_mmuidx(s, dirty_addr, is_store,
271
}
256
+ writeback || rn != 31,
272
return 0;
257
+ size, is_unpriv, memidx);
273
258
274
+ case 1: /* VSRA */
259
if (is_vector) {
275
+ /* Right shift comes here negative. */
260
if (is_store) {
276
+ shift = -shift;
261
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
277
+ /* Shifts larger than the element size are architecturally
262
}
278
+ * valid. Unsigned results in all zeros; signed results
263
} else {
279
+ * in all sign bits.
264
TCGv_i64 tcg_rt = cpu_reg(s, rt);
280
+ */
265
- int memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
281
+ if (!u) {
266
bool iss_sf = disas_ldst_compute_iss_sf(size, is_signed, opc);
282
+ tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size, vec_size,
267
283
+ MIN(shift, (8 << size) - 1),
268
if (is_store) {
284
+ &ssra_op[size]);
269
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn,
285
+ } else if (shift >= 8 << size) {
270
ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
286
+ /* rd += 0 */
271
287
+ } else {
272
tcg_gen_add_i64(dirty_addr, dirty_addr, tcg_rm);
288
+ tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size, vec_size,
273
- clean_addr = clean_data_tbi(s, dirty_addr);
289
+ shift, &usra_op[size]);
274
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, true, size);
290
+ }
275
291
+ return 0;
276
if (is_vector) {
292
+
277
if (is_store) {
293
case 5: /* VSHL, VSLI */
278
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn,
294
if (!u) { /* VSHL */
279
dirty_addr = read_cpu_reg_sp(s, rn, 1);
295
/* Shifts larger than the element size are
280
offset = imm12 << size;
296
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
281
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
297
neon_load_reg64(cpu_V0, rm + pass);
282
- clean_addr = clean_data_tbi(s, dirty_addr);
298
tcg_gen_movi_i64(cpu_V1, imm);
283
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, rn != 31, size);
299
switch (op) {
284
300
- case 1: /* VSRA */
285
if (is_vector) {
301
- if (u)
286
if (is_store) {
302
- gen_helper_neon_shl_u64(cpu_V0, cpu_V0, cpu_V1);
287
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
303
- else
288
if (rn == 31) {
304
- gen_helper_neon_shl_s64(cpu_V0, cpu_V0, cpu_V1);
289
gen_check_sp_alignment(s);
305
- break;
290
}
306
case 2: /* VRSHR */
291
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
307
case 3: /* VRSRA */
292
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), false, rn != 31, size);
308
if (u)
293
309
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
294
if (o3_opc == 014) {
310
default:
295
/*
311
g_assert_not_reached();
296
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pac(DisasContext *s, uint32_t insn,
312
}
297
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
313
- if (op == 1 || op == 3) {
298
314
+ if (op == 3) {
299
/* Note that "clean" and "dirty" here refer to TBI not PAC. */
315
/* Accumulate. */
300
- clean_addr = clean_data_tbi(s, dirty_addr);
316
neon_load_reg64(cpu_V1, rd + pass);
301
+ clean_addr = gen_mte_check1(s, dirty_addr, false,
317
tcg_gen_add_i64(cpu_V0, cpu_V0, cpu_V1);
302
+ is_wback || rn != 31, size);
318
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
303
319
tmp2 = tcg_temp_new_i32();
304
tcg_rt = cpu_reg(s, rt);
320
tcg_gen_movi_i32(tmp2, imm);
305
do_gpr_ld(s, tcg_rt, clean_addr, size, /* is_signed */ false,
321
switch (op) {
322
- case 1: /* VSRA */
323
- GEN_NEON_INTEGER_OP(shl);
324
- break;
325
case 2: /* VRSHR */
326
case 3: /* VRSRA */
327
GEN_NEON_INTEGER_OP(rshl);
328
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
329
}
330
tcg_temp_free_i32(tmp2);
331
332
- if (op == 1 || op == 3) {
333
+ if (op == 3) {
334
/* Accumulate. */
335
tmp2 = neon_load_reg(rd, pass);
336
gen_neon_add(size, tmp, tmp2);
337
--
306
--
338
2.19.1
307
2.20.1
339
308
340
309
diff view generated by jsdifflib
1
From: Richard Henderson <rth@twiddle.net>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This can reduce the number of opcodes required for certain
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
complex forms of load-multiple (e.g. ld4.16b).
4
perform multiple logical memory access. Leave the helper blank
5
5
for now to reduce the patch size.
6
Signed-off-by: Richard Henderson <rth@twiddle.net>
6
7
Message-id: 20181011205206.3552-2-richard.henderson@linaro.org
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-25-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
---
11
target/arm/translate-a64.c | 12 ++++++++----
12
target/arm/helper-a64.h | 1 +
12
1 file changed, 8 insertions(+), 4 deletions(-)
13
target/arm/translate-a64.h | 2 ++
13
14
target/arm/mte_helper.c | 8 +++++
15
target/arm/translate-a64.c | 71 +++++++++++++++++++++++++++++---------
16
4 files changed, 66 insertions(+), 16 deletions(-)
17
18
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/helper-a64.h
21
+++ b/target/arm/helper-a64.h
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
25
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
26
+DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
27
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
28
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
29
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
30
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/translate-a64.h
33
+++ b/target/arm/translate-a64.h
34
@@ -XXX,XX +XXX,XX @@ bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
35
bool sve_access_check(DisasContext *s);
36
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
37
bool tag_checked, int log2_size);
38
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
39
+ bool tag_checked, int count, int log2_esize);
40
41
/* We should have at some point before trying to access an FP register
42
* done the necessary access check, so assert that
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
48
{
49
return ptr;
50
}
51
+
52
+/*
53
+ * Perform an MTE checked access for multiple logical accesses.
54
+ */
55
+uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
56
+{
57
+ return ptr;
58
+}
14
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
59
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
15
index XXXXXXX..XXXXXXX 100644
60
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate-a64.c
61
--- a/target/arm/translate-a64.c
17
+++ b/target/arm/translate-a64.c
62
+++ b/target/arm/translate-a64.c
63
@@ -XXX,XX +XXX,XX @@ TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
64
false, get_mem_index(s));
65
}
66
67
+/*
68
+ * For MTE, check multiple logical sequential accesses.
69
+ */
70
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
71
+ bool tag_checked, int log2_esize, int total_size)
72
+{
73
+ if (tag_checked && s->mte_active[0] && total_size != (1 << log2_esize)) {
74
+ TCGv_i32 tcg_desc;
75
+ TCGv_i64 ret;
76
+ int desc = 0;
77
+
78
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
79
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
80
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
81
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
82
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_esize);
83
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, total_size);
84
+ tcg_desc = tcg_const_i32(desc);
85
+
86
+ ret = new_tmp_a64(s);
87
+ gen_helper_mte_checkN(ret, cpu_env, tcg_desc, addr);
88
+ tcg_temp_free_i32(tcg_desc);
89
+
90
+ return ret;
91
+ }
92
+ return gen_mte_check1(s, addr, is_write, tag_checked, log2_esize);
93
+}
94
+
95
typedef struct DisasCompare64 {
96
TCGCond cond;
97
TCGv_i64 value;
98
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
99
}
100
}
101
102
- clean_addr = clean_data_tbi(s, dirty_addr);
103
+ clean_addr = gen_mte_checkN(s, dirty_addr, !is_load,
104
+ (wback || rn != 31) && !set_tag,
105
+ size, 2 << size);
106
+
107
if (is_vector) {
108
if (is_load) {
109
do_fp_ld(s, rt, clean_addr, size);
18
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
110
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
19
bool is_store = !extract32(insn, 22, 1);
111
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
20
bool is_postidx = extract32(insn, 23, 1);
112
MemOp endian = s->be_data;
21
bool is_q = extract32(insn, 30, 1);
113
22
- TCGv_i64 tcg_addr, tcg_rn;
114
- int ebytes; /* bytes per element */
23
+ TCGv_i64 tcg_addr, tcg_rn, tcg_ebytes;
115
+ int total; /* total bytes */
24
116
int elements; /* elements per vector */
25
int ebytes = 1 << size;
117
int rpt; /* num iterations */
26
int elements = (is_q ? 128 : 64) / (8 << size);
118
int selem; /* structure elements */
27
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
119
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
28
tcg_rn = cpu_reg_sp(s, rn);
120
endian = MO_LE;
29
tcg_addr = tcg_temp_new_i64();
121
}
30
tcg_gen_mov_i64(tcg_addr, tcg_rn);
122
31
+ tcg_ebytes = tcg_const_i64(ebytes);
123
- /* Consecutive little-endian elements from a single register
32
124
+ total = rpt * selem * (is_q ? 16 : 8);
125
+ tcg_rn = cpu_reg_sp(s, rn);
126
+
127
+ /*
128
+ * Issue the MTE check vs the logical repeat count, before we
129
+ * promote consecutive little-endian elements below.
130
+ */
131
+ clean_addr = gen_mte_checkN(s, tcg_rn, is_store, is_postidx || rn != 31,
132
+ size, total);
133
+
134
+ /*
135
+ * Consecutive little-endian elements from a single register
136
* can be promoted to a larger little-endian operation.
137
*/
138
if (selem == 1 && endian == MO_LE) {
139
size = 3;
140
}
141
- ebytes = 1 << size;
142
- elements = (is_q ? 16 : 8) / ebytes;
143
-
144
- tcg_rn = cpu_reg_sp(s, rn);
145
- clean_addr = clean_data_tbi(s, tcg_rn);
146
- tcg_ebytes = tcg_const_i64(ebytes);
147
+ elements = (is_q ? 16 : 8) >> size;
148
149
+ tcg_ebytes = tcg_const_i64(1 << size);
33
for (r = 0; r < rpt; r++) {
150
for (r = 0; r < rpt; r++) {
34
int e;
151
int e;
152
for (e = 0; e < elements; e++) {
35
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
153
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
36
clear_vec_high(s, is_q, tt);
154
37
}
155
if (is_postidx) {
38
}
156
if (rm == 31) {
39
- tcg_gen_addi_i64(tcg_addr, tcg_addr, ebytes);
157
- tcg_gen_addi_i64(tcg_rn, tcg_rn, rpt * elements * selem * ebytes);
40
+ tcg_gen_add_i64(tcg_addr, tcg_addr, tcg_ebytes);
158
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
41
tt = (tt + 1) % 32;
159
} else {
42
}
43
}
44
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
45
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
160
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
46
}
161
}
47
}
48
+ tcg_temp_free_i64(tcg_ebytes);
49
tcg_temp_free_i64(tcg_addr);
50
}
51
52
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
162
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
163
int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
53
bool replicate = false;
164
bool replicate = false;
54
int index = is_q << 3 | S << 2 | size;
165
int index = is_q << 3 | S << 2 | size;
55
int ebytes, xs;
166
- int ebytes, xs;
56
- TCGv_i64 tcg_addr, tcg_rn;
167
+ int xs, total;
57
+ TCGv_i64 tcg_addr, tcg_rn, tcg_ebytes;
168
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
58
169
59
switch (scale) {
170
if (extract32(insn, 31, 1)) {
60
case 3:
61
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
171
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
172
return;
173
}
174
175
- ebytes = 1 << scale;
176
-
177
if (rn == 31) {
178
gen_check_sp_alignment(s);
179
}
180
181
+ total = selem << scale;
62
tcg_rn = cpu_reg_sp(s, rn);
182
tcg_rn = cpu_reg_sp(s, rn);
63
tcg_addr = tcg_temp_new_i64();
183
- clean_addr = clean_data_tbi(s, tcg_rn);
64
tcg_gen_mov_i64(tcg_addr, tcg_rn);
184
- tcg_ebytes = tcg_const_i64(ebytes);
65
+ tcg_ebytes = tcg_const_i64(ebytes);
185
66
186
+ clean_addr = gen_mte_checkN(s, tcg_rn, !is_load, is_postidx || rn != 31,
187
+ scale, total);
188
+
189
+ tcg_ebytes = tcg_const_i64(1 << scale);
67
for (xs = 0; xs < selem; xs++) {
190
for (xs = 0; xs < selem; xs++) {
68
if (replicate) {
191
if (replicate) {
192
/* Load and replicate to all elements */
69
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
193
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
70
do_vec_st(s, rt, index, tcg_addr, scale);
194
71
}
195
if (is_postidx) {
72
}
196
if (rm == 31) {
73
- tcg_gen_addi_i64(tcg_addr, tcg_addr, ebytes);
197
- tcg_gen_addi_i64(tcg_rn, tcg_rn, selem * ebytes);
74
+ tcg_gen_add_i64(tcg_addr, tcg_addr, tcg_ebytes);
198
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
75
rt = (rt + 1) % 32;
199
} else {
76
}
77
78
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
79
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
200
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
80
}
201
}
81
}
82
+ tcg_temp_free_i64(tcg_ebytes);
83
tcg_temp_free_i64(tcg_addr);
84
}
85
86
--
202
--
87
2.19.1
203
2.20.1
88
204
89
205
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-26-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 48 +++++++++++++++
11
target/arm/mte_helper.c | 132 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 179 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, WRITE, 8, 1)
19
FIELD(MTEDESC, ESIZE, 9, 5)
20
FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
21
22
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
23
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
24
+ uint64_t ptr, uintptr_t ra);
25
+
26
static inline int allocation_tag_from_addr(uint64_t ptr)
27
{
28
return extract64(ptr, 56, 4);
29
@@ -XXX,XX +XXX,XX @@ static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
30
return deposit64(ptr, 56, 4, rtag);
31
}
32
33
+/* Return true if tbi bits mean that the access is checked. */
34
+static inline bool tbi_check(uint32_t desc, int bit55)
35
+{
36
+ return (desc >> (R_MTEDESC_TBI_SHIFT + bit55)) & 1;
37
+}
38
+
39
+/* Return true if tcma bits mean that the access is unchecked. */
40
+static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
41
+{
42
+ /*
43
+ * We had extracted bit55 and ptr_tag for other reasons, so fold
44
+ * (ptr<59:55> == 00000 || ptr<59:55> == 11111) into a single test.
45
+ */
46
+ bool match = ((ptr_tag + bit55) & 0xf) == 0;
47
+ bool tcma = (desc >> (R_MTEDESC_TCMA_SHIFT + bit55)) & 1;
48
+ return tcma && match;
49
+}
50
+
51
+/*
52
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
53
+ * for the tag to be present in the FAR_ELx register. But for user-only
54
+ * mode, we do not have a TLB with which to implement this, so we must
55
+ * remove the top byte.
56
+ */
57
+static inline uint64_t useronly_clean_ptr(uint64_t ptr)
58
+{
59
+ /* TBI is known to be enabled. */
60
+#ifdef CONFIG_USER_ONLY
61
+ ptr = sextract64(ptr, 0, 56);
62
+#endif
63
+ return ptr;
64
+}
65
+
66
+static inline uint64_t useronly_maybe_clean_ptr(uint32_t desc, uint64_t ptr)
67
+{
68
+#ifdef CONFIG_USER_ONLY
69
+ int64_t clean_ptr = sextract64(ptr, 0, 56);
70
+ if (tbi_check(desc, clean_ptr < 0)) {
71
+ ptr = clean_ptr;
72
+ }
73
+#endif
74
+ return ptr;
75
+}
76
+
77
#endif
78
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/target/arm/mte_helper.c
81
+++ b/target/arm/mte_helper.c
82
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
83
}
84
}
85
86
+/* Record a tag check failure. */
87
+static void mte_check_fail(CPUARMState *env, int mmu_idx,
88
+ uint64_t dirty_ptr, uintptr_t ra)
89
+{
90
+ ARMMMUIdx arm_mmu_idx = core_to_aa64_mmu_idx(mmu_idx);
91
+ int el, reg_el, tcf, select;
92
+ uint64_t sctlr;
93
+
94
+ reg_el = regime_el(env, arm_mmu_idx);
95
+ sctlr = env->cp15.sctlr_el[reg_el];
96
+
97
+ switch (arm_mmu_idx) {
98
+ case ARMMMUIdx_E10_0:
99
+ case ARMMMUIdx_E20_0:
100
+ el = 0;
101
+ tcf = extract64(sctlr, 38, 2);
102
+ break;
103
+ default:
104
+ el = reg_el;
105
+ tcf = extract64(sctlr, 40, 2);
106
+ }
107
+
108
+ switch (tcf) {
109
+ case 1:
110
+ /*
111
+ * Tag check fail causes a synchronous exception.
112
+ *
113
+ * In restore_state_to_opc, we set the exception syndrome
114
+ * for the load or store operation. Unwind first so we
115
+ * may overwrite that with the syndrome for the tag check.
116
+ */
117
+ cpu_restore_state(env_cpu(env), ra, true);
118
+ env->exception.vaddress = dirty_ptr;
119
+ raise_exception(env, EXCP_DATA_ABORT,
120
+ syn_data_abort_no_iss(el != 0, 0, 0, 0, 0, 0, 0x11),
121
+ exception_target_el(env));
122
+ /* noreturn, but fall through to the assert anyway */
123
+
124
+ case 0:
125
+ /*
126
+ * Tag check fail does not affect the PE.
127
+ * We eliminate this case by not setting MTE_ACTIVE
128
+ * in tb_flags, so that we never make this runtime call.
129
+ */
130
+ g_assert_not_reached();
131
+
132
+ case 2:
133
+ /* Tag check fail causes asynchronous flag set. */
134
+ mmu_idx = arm_mmu_idx_el(env, el);
135
+ if (regime_has_2_ranges(mmu_idx)) {
136
+ select = extract64(dirty_ptr, 55, 1);
137
+ } else {
138
+ select = 0;
139
+ }
140
+ env->cp15.tfsr_el[el] |= 1 << select;
141
+ break;
142
+
143
+ default:
144
+ /* Case 3: Reserved. */
145
+ qemu_log_mask(LOG_GUEST_ERROR,
146
+ "Tag check failure with SCTLR_EL%d.TCF%s "
147
+ "set to reserved value %d\n",
148
+ reg_el, el ? "" : "0", tcf);
149
+ break;
150
+ }
151
+}
152
+
153
/*
154
* Perform an MTE checked access for a single logical or atomic access.
155
*/
156
+static bool mte_probe1_int(CPUARMState *env, uint32_t desc, uint64_t ptr,
157
+ uintptr_t ra, int bit55)
158
+{
159
+ int mem_tag, mmu_idx, ptr_tag, size;
160
+ MMUAccessType type;
161
+ uint8_t *mem;
162
+
163
+ ptr_tag = allocation_tag_from_addr(ptr);
164
+
165
+ if (tcma_check(desc, bit55, ptr_tag)) {
166
+ return true;
167
+ }
168
+
169
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
170
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
171
+ size = FIELD_EX32(desc, MTEDESC, ESIZE);
172
+
173
+ mem = allocation_tag_mem(env, mmu_idx, ptr, type, size,
174
+ MMU_DATA_LOAD, 1, ra);
175
+ if (!mem) {
176
+ return true;
177
+ }
178
+
179
+ mem_tag = load_tag1(ptr, mem);
180
+ return ptr_tag == mem_tag;
181
+}
182
+
183
+/*
184
+ * No-fault version of mte_check1, to be used by SVE for MemSingleNF.
185
+ * Returns false if the access is Checked and the check failed. This
186
+ * is only intended to probe the tag -- the validity of the page must
187
+ * be checked beforehand.
188
+ */
189
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr)
190
+{
191
+ int bit55 = extract64(ptr, 55, 1);
192
+
193
+ /* If TBI is disabled, the access is unchecked. */
194
+ if (unlikely(!tbi_check(desc, bit55))) {
195
+ return true;
196
+ }
197
+
198
+ return mte_probe1_int(env, desc, ptr, 0, bit55);
199
+}
200
+
201
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
202
+ uint64_t ptr, uintptr_t ra)
203
+{
204
+ int bit55 = extract64(ptr, 55, 1);
205
+
206
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
207
+ if (unlikely(!tbi_check(desc, bit55))) {
208
+ return ptr;
209
+ }
210
+
211
+ if (unlikely(!mte_probe1_int(env, desc, ptr, ra, bit55))) {
212
+ int mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
213
+ mte_check_fail(env, mmu_idx, ptr, ra);
214
+ }
215
+
216
+ return useronly_clean_ptr(ptr);
217
+}
218
+
219
uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
220
{
221
- return ptr;
222
+ return mte_check1(env, desc, ptr, GETPC());
223
}
224
225
/*
226
--
227
2.20.1
228
229
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Fill out the stub that was added earlier.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-27-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 2 +
11
target/arm/mte_helper.c | 165 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 166 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
17
+++ b/target/arm/internals.h
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
19
bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
20
uint64_t mte_check1(CPUARMState *env, uint32_t desc,
21
uint64_t ptr, uintptr_t ra);
22
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
23
+ uint64_t ptr, uintptr_t ra);
24
25
static inline int allocation_tag_from_addr(uint64_t ptr)
26
{
27
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/mte_helper.c
30
+++ b/target/arm/mte_helper.c
31
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
32
/*
33
* Perform an MTE checked access for multiple logical accesses.
34
*/
35
+
36
+/**
37
+ * checkN:
38
+ * @tag: tag memory to test
39
+ * @odd: true to begin testing at tags at odd nibble
40
+ * @cmp: the tag to compare against
41
+ * @count: number of tags to test
42
+ *
43
+ * Return the number of successful tests.
44
+ * Thus a return value < @count indicates a failure.
45
+ *
46
+ * A note about sizes: count is expected to be small.
47
+ *
48
+ * The most common use will be LDP/STP of two integer registers,
49
+ * which means 16 bytes of memory touching at most 2 tags, but
50
+ * often the access is aligned and thus just 1 tag.
51
+ *
52
+ * Using AdvSIMD LD/ST (multiple), one can access 64 bytes of memory,
53
+ * touching at most 5 tags. SVE LDR/STR (vector) with the default
54
+ * vector length is also 64 bytes; the maximum architectural length
55
+ * is 256 bytes touching at most 9 tags.
56
+ *
57
+ * The loop below uses 7 logical operations and 1 memory operation
58
+ * per tag pair. An implementation that loads an aligned word and
59
+ * uses masking to ignore adjacent tags requires 18 logical operations
60
+ * and thus does not begin to pay off until 6 tags.
61
+ * Which, according to the survey above, is unlikely to be common.
62
+ */
63
+static int checkN(uint8_t *mem, int odd, int cmp, int count)
64
+{
65
+ int n = 0, diff;
66
+
67
+ /* Replicate the test tag and compare. */
68
+ cmp *= 0x11;
69
+ diff = *mem++ ^ cmp;
70
+
71
+ if (odd) {
72
+ goto start_odd;
73
+ }
74
+
75
+ while (1) {
76
+ /* Test even tag. */
77
+ if (unlikely((diff) & 0x0f)) {
78
+ break;
79
+ }
80
+ if (++n == count) {
81
+ break;
82
+ }
83
+
84
+ start_odd:
85
+ /* Test odd tag. */
86
+ if (unlikely((diff) & 0xf0)) {
87
+ break;
88
+ }
89
+ if (++n == count) {
90
+ break;
91
+ }
92
+
93
+ diff = *mem++ ^ cmp;
94
+ }
95
+ return n;
96
+}
97
+
98
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
99
+ uint64_t ptr, uintptr_t ra)
100
+{
101
+ int mmu_idx, ptr_tag, bit55;
102
+ uint64_t ptr_last, ptr_end, prev_page, next_page;
103
+ uint64_t tag_first, tag_end;
104
+ uint64_t tag_byte_first, tag_byte_end;
105
+ uint32_t esize, total, tag_count, tag_size, n, c;
106
+ uint8_t *mem1, *mem2;
107
+ MMUAccessType type;
108
+
109
+ bit55 = extract64(ptr, 55, 1);
110
+
111
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
112
+ if (unlikely(!tbi_check(desc, bit55))) {
113
+ return ptr;
114
+ }
115
+
116
+ ptr_tag = allocation_tag_from_addr(ptr);
117
+
118
+ if (tcma_check(desc, bit55, ptr_tag)) {
119
+ goto done;
120
+ }
121
+
122
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
123
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
124
+ esize = FIELD_EX32(desc, MTEDESC, ESIZE);
125
+ total = FIELD_EX32(desc, MTEDESC, TSIZE);
126
+
127
+ /* Find the addr of the end of the access, and of the last element. */
128
+ ptr_end = ptr + total;
129
+ ptr_last = ptr_end - esize;
130
+
131
+ /* Round the bounds to the tag granule, and compute the number of tags. */
132
+ tag_first = QEMU_ALIGN_DOWN(ptr, TAG_GRANULE);
133
+ tag_end = QEMU_ALIGN_UP(ptr_last, TAG_GRANULE);
134
+ tag_count = (tag_end - tag_first) / TAG_GRANULE;
135
+
136
+ /* Round the bounds to twice the tag granule, and compute the bytes. */
137
+ tag_byte_first = QEMU_ALIGN_DOWN(ptr, 2 * TAG_GRANULE);
138
+ tag_byte_end = QEMU_ALIGN_UP(ptr_last, 2 * TAG_GRANULE);
139
+
140
+ /* Locate the page boundaries. */
141
+ prev_page = ptr & TARGET_PAGE_MASK;
142
+ next_page = prev_page + TARGET_PAGE_SIZE;
143
+
144
+ if (likely(tag_end - prev_page <= TARGET_PAGE_SIZE)) {
145
+ /* Memory access stays on one page. */
146
+ tag_size = (tag_byte_end - tag_byte_first) / (2 * TAG_GRANULE);
147
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, total,
148
+ MMU_DATA_LOAD, tag_size, ra);
149
+ if (!mem1) {
150
+ goto done;
151
+ }
152
+ /* Perform all of the comparisons. */
153
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, tag_count);
154
+ } else {
155
+ /* Memory access crosses to next page. */
156
+ tag_size = (next_page - tag_byte_first) / (2 * TAG_GRANULE);
157
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, next_page - ptr,
158
+ MMU_DATA_LOAD, tag_size, ra);
159
+
160
+ tag_size = (tag_byte_end - next_page) / (2 * TAG_GRANULE);
161
+ mem2 = allocation_tag_mem(env, mmu_idx, next_page, type,
162
+ ptr_end - next_page,
163
+ MMU_DATA_LOAD, tag_size, ra);
164
+
165
+ /*
166
+ * Perform all of the comparisons.
167
+ * Note the possible but unlikely case of the operation spanning
168
+ * two pages that do not both have tagging enabled.
169
+ */
170
+ n = c = (next_page - tag_first) / TAG_GRANULE;
171
+ if (mem1) {
172
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, c);
173
+ }
174
+ if (n == c) {
175
+ if (!mem2) {
176
+ goto done;
177
+ }
178
+ n += checkN(mem2, 0, ptr_tag, tag_count - c);
179
+ }
180
+ }
181
+
182
+ /*
183
+ * If we failed, we know which granule. Compute the element that
184
+ * is first in that granule, and signal failure on that element.
185
+ */
186
+ if (unlikely(n < tag_count)) {
187
+ uint64_t fail_ofs;
188
+
189
+ fail_ofs = tag_first + n * TAG_GRANULE - ptr;
190
+ fail_ofs = ROUND_UP(fail_ofs, esize);
191
+ mte_check_fail(env, mmu_idx, ptr + fail_ofs, ra);
192
+ }
193
+
194
+ done:
195
+ return useronly_clean_ptr(ptr);
196
+}
197
+
198
uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
199
{
200
- return ptr;
201
+ return mte_checkN(env, desc, ptr, GETPC());
202
}
203
--
204
2.20.1
205
206
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Use a special helper for DC_ZVA, rather than the more
4
general mte_checkN.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-13-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-28-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
target/arm/translate.c | 70 +++++++++++++++++++++++++++++-------------
11
target/arm/helper-a64.h | 1 +
9
1 file changed, 48 insertions(+), 22 deletions(-)
12
target/arm/mte_helper.c | 106 +++++++++++++++++++++++++++++++++++++
13
target/arm/translate-a64.c | 16 +++++-
14
3 files changed, 122 insertions(+), 1 deletion(-)
10
15
11
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
12
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate.c
18
--- a/target/arm/helper-a64.h
14
+++ b/target/arm/translate.c
19
+++ b/target/arm/helper-a64.h
15
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
16
size--;
21
17
}
22
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
18
shift = (insn >> 16) & ((1 << (3 + size)) - 1);
23
DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
19
- /* To avoid excessive duplication of ops we implement shift
24
+DEF_HELPER_FLAGS_3(mte_check_zva, TCG_CALL_NO_WG, i64, env, i32, i64)
20
- by immediate using the variable shift operations. */
25
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
21
if (op < 8) {
26
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
22
/* Shift by immediate:
27
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
23
VSHR, VSRA, VRSHR, VRSRA, VSRI, VSHL, VQSHL, VQSHLU. */
28
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
24
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
29
index XXXXXXX..XXXXXXX 100644
25
}
30
--- a/target/arm/mte_helper.c
26
/* Right shifts are encoded as N - shift, where N is the
31
+++ b/target/arm/mte_helper.c
27
element size in bits. */
32
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
28
- if (op <= 4)
33
{
29
+ if (op <= 4) {
34
return mte_checkN(env, desc, ptr, GETPC());
30
shift = shift - (1 << (size + 3));
35
}
31
+ }
32
+
36
+
33
+ switch (op) {
37
+/*
34
+ case 0: /* VSHR */
38
+ * Perform an MTE checked access for DC_ZVA.
35
+ /* Right shift comes here negative. */
39
+ */
36
+ shift = -shift;
40
+uint64_t HELPER(mte_check_zva)(CPUARMState *env, uint32_t desc, uint64_t ptr)
37
+ /* Shifts larger than the element size are architecturally
41
+{
38
+ * valid. Unsigned results in all zeros; signed results
42
+ uintptr_t ra = GETPC();
39
+ * in all sign bits.
43
+ int log2_dcz_bytes, log2_tag_bytes;
40
+ */
44
+ int mmu_idx, bit55;
41
+ if (!u) {
45
+ intptr_t dcz_bytes, tag_bytes, i;
42
+ tcg_gen_gvec_sari(size, rd_ofs, rm_ofs,
46
+ void *mem;
43
+ MIN(shift, (8 << size) - 1),
47
+ uint64_t ptr_tag, mem_tag, align_ptr;
44
+ vec_size, vec_size);
45
+ } else if (shift >= 8 << size) {
46
+ tcg_gen_gvec_dup8i(rd_ofs, vec_size, vec_size, 0);
47
+ } else {
48
+ tcg_gen_gvec_shri(size, rd_ofs, rm_ofs, shift,
49
+ vec_size, vec_size);
50
+ }
51
+ return 0;
52
+
48
+
53
+ case 5: /* VSHL, VSLI */
49
+ bit55 = extract64(ptr, 55, 1);
54
+ if (!u) { /* VSHL */
55
+ /* Shifts larger than the element size are
56
+ * architecturally valid and results in zero.
57
+ */
58
+ if (shift >= 8 << size) {
59
+ tcg_gen_gvec_dup8i(rd_ofs, vec_size, vec_size, 0);
60
+ } else {
61
+ tcg_gen_gvec_shli(size, rd_ofs, rm_ofs, shift,
62
+ vec_size, vec_size);
63
+ }
64
+ return 0;
65
+ }
66
+ break;
67
+ }
68
+
50
+
69
if (size == 3) {
51
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
70
count = q + 1;
52
+ if (unlikely(!tbi_check(desc, bit55))) {
71
} else {
53
+ return ptr;
72
count = q ? 4: 2;
54
+ }
73
}
74
- switch (size) {
75
- case 0:
76
- imm = (uint8_t) shift;
77
- imm |= imm << 8;
78
- imm |= imm << 16;
79
- break;
80
- case 1:
81
- imm = (uint16_t) shift;
82
- imm |= imm << 16;
83
- break;
84
- case 2:
85
- case 3:
86
- imm = shift;
87
- break;
88
- default:
89
- abort();
90
- }
91
+
55
+
92
+ /* To avoid excessive duplication of ops we implement shift
56
+ ptr_tag = allocation_tag_from_addr(ptr);
93
+ * by immediate using the variable shift operations.
57
+
94
+ */
58
+ if (tcma_check(desc, bit55, ptr_tag)) {
95
+ imm = dup_const(size, shift);
59
+ goto done;
96
60
+ }
97
for (pass = 0; pass < count; pass++) {
61
+
98
if (size == 3) {
62
+ /*
99
neon_load_reg64(cpu_V0, rm + pass);
63
+ * In arm_cpu_realizefn, we asserted that dcz > LOG2_TAG_GRANULE+1,
100
tcg_gen_movi_i64(cpu_V1, imm);
64
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway, to make
101
switch (op) {
65
+ * sure that we can access one complete tag byte here.
102
- case 0: /* VSHR */
66
+ */
103
case 1: /* VSRA */
67
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
104
if (u)
68
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
105
gen_helper_neon_shl_u64(cpu_V0, cpu_V0, cpu_V1);
69
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
106
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
70
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
107
cpu_V0, cpu_V1);
71
+ align_ptr = ptr & -dcz_bytes;
108
}
72
+
109
break;
73
+ /*
110
+ default:
74
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
111
+ g_assert_not_reached();
75
+ * the original pointer for an invalid page. But watchpoints require
112
}
76
+ * that we probe the actual space. So do both.
113
if (op == 1 || op == 3) {
77
+ */
114
/* Accumulate. */
78
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
115
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
79
+ (void) probe_write(env, ptr, 1, mmu_idx, ra);
116
tmp2 = tcg_temp_new_i32();
80
+ mem = allocation_tag_mem(env, mmu_idx, align_ptr, MMU_DATA_STORE,
117
tcg_gen_movi_i32(tmp2, imm);
81
+ dcz_bytes, MMU_DATA_LOAD, tag_bytes, ra);
118
switch (op) {
82
+ if (!mem) {
119
- case 0: /* VSHR */
83
+ goto done;
120
case 1: /* VSRA */
84
+ }
121
GEN_NEON_INTEGER_OP(shl);
85
+
122
break;
86
+ /*
123
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
87
+ * Unlike the reasoning for checkN, DC_ZVA is always aligned, and thus
124
case 7: /* VQSHL */
88
+ * it is quite easy to perform all of the comparisons at once without
125
GEN_NEON_INTEGER_OP_ENV(qshl);
89
+ * any extra masking.
126
break;
90
+ *
127
+ default:
91
+ * The most common zva block size is 64; some of the thunderx cpus use
128
+ g_assert_not_reached();
92
+ * a block size of 128. For user-only, aarch64_max_initfn will set the
129
}
93
+ * block size to 512. Fill out the other cases for future-proofing.
130
tcg_temp_free_i32(tmp2);
94
+ *
131
95
+ * In order to be able to find the first miscompare later, we want the
96
+ * tag bytes to be in little-endian order.
97
+ */
98
+ switch (log2_tag_bytes) {
99
+ case 0: /* zva_blocksize 32 */
100
+ mem_tag = *(uint8_t *)mem;
101
+ ptr_tag *= 0x11u;
102
+ break;
103
+ case 1: /* zva_blocksize 64 */
104
+ mem_tag = cpu_to_le16(*(uint16_t *)mem);
105
+ ptr_tag *= 0x1111u;
106
+ break;
107
+ case 2: /* zva_blocksize 128 */
108
+ mem_tag = cpu_to_le32(*(uint32_t *)mem);
109
+ ptr_tag *= 0x11111111u;
110
+ break;
111
+ case 3: /* zva_blocksize 256 */
112
+ mem_tag = cpu_to_le64(*(uint64_t *)mem);
113
+ ptr_tag *= 0x1111111111111111ull;
114
+ break;
115
+
116
+ default: /* zva_blocksize 512, 1024, 2048 */
117
+ ptr_tag *= 0x1111111111111111ull;
118
+ i = 0;
119
+ do {
120
+ mem_tag = cpu_to_le64(*(uint64_t *)(mem + i));
121
+ if (unlikely(mem_tag != ptr_tag)) {
122
+ goto fail;
123
+ }
124
+ i += 8;
125
+ align_ptr += 16 * TAG_GRANULE;
126
+ } while (i < tag_bytes);
127
+ goto done;
128
+ }
129
+
130
+ if (likely(mem_tag == ptr_tag)) {
131
+ goto done;
132
+ }
133
+
134
+ fail:
135
+ /* Locate the first nibble that differs. */
136
+ i = ctz64(mem_tag ^ ptr_tag) >> 4;
137
+ mte_check_fail(env, mmu_idx, align_ptr + i * TAG_GRANULE, ra);
138
+
139
+ done:
140
+ return useronly_clean_ptr(ptr);
141
+}
142
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
143
index XXXXXXX..XXXXXXX 100644
144
--- a/target/arm/translate-a64.c
145
+++ b/target/arm/translate-a64.c
146
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
147
return;
148
case ARM_CP_DC_ZVA:
149
/* Writes clear the aligned block of memory which rt points into. */
150
- tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
151
+ if (s->mte_active[0]) {
152
+ TCGv_i32 t_desc;
153
+ int desc = 0;
154
+
155
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
156
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
157
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
158
+ t_desc = tcg_const_i32(desc);
159
+
160
+ tcg_rt = new_tmp_a64(s);
161
+ gen_helper_mte_check_zva(tcg_rt, cpu_env, t_desc, cpu_reg(s, rt));
162
+ tcg_temp_free_i32(t_desc);
163
+ } else {
164
+ tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
165
+ }
166
gen_helper_dc_zva(cpu_env, tcg_rt);
167
return;
168
default:
132
--
169
--
133
2.19.1
170
2.20.1
134
171
135
172
diff view generated by jsdifflib
1
From: Stewart Hildebrand <Stewart.Hildebrand@dornerworks.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
"The Image must be placed text_offset bytes from a 2MB aligned base
4
address anywhere in usable system RAM and called there."
5
6
For the virt board, we write our startup bootloader at the very
7
bottom of RAM, so that bit can't be used for the image. To avoid
8
overlap in case the image requests to be loaded at an offset
9
smaller than our bootloader, we increment the load offset to the
10
next 2MB.
11
12
This fixes a boot failure for Xen AArch64.
13
14
Signed-off-by: Stewart Hildebrand <stewart.hildebrand@dornerworks.com>
15
Tested-by: Andre Przywara <andre.przywara@arm.com>
16
Message-id: b8a89518794b4436af0c151ed10de4fa@dornerworks.com
17
[PMM: Rephrased a comment a bit]
18
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-29-richard.henderson@linaro.org
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
---
7
---
21
hw/arm/boot.c | 18 ++++++++++++++++++
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
22
1 file changed, 18 insertions(+)
9
1 file changed, 33 insertions(+), 28 deletions(-)
23
10
24
diff --git a/hw/arm/boot.c b/hw/arm/boot.c
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
25
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
26
--- a/hw/arm/boot.c
13
--- a/target/arm/translate-sve.c
27
+++ b/hw/arm/boot.c
14
+++ b/target/arm/translate-sve.c
28
@@ -XXX,XX +XXX,XX @@
15
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
29
#include "qemu/config-file.h"
16
int len_remain = len % 8;
30
#include "qemu/option.h"
17
int nparts = len / 8 + ctpop8(len_remain);
31
#include "exec/address-spaces.h"
18
int midx = get_mem_index(s);
32
+#include "qemu/units.h"
19
- TCGv_i64 addr, t0, t1;
33
20
+ TCGv_i64 dirty_addr, clean_addr, t0, t1;
34
/* Kernel boot protocol is specified in the kernel docs
21
35
* Documentation/arm/Booting and Documentation/arm64/booting.txt
22
- addr = tcg_temp_new_i64();
36
@@ -XXX,XX +XXX,XX @@
23
- t0 = tcg_temp_new_i64();
37
#define ARM64_TEXT_OFFSET_OFFSET 8
24
+ dirty_addr = tcg_temp_new_i64();
38
#define ARM64_MAGIC_OFFSET 56
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
39
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
40
+#define BOOTLOADER_MAX_SIZE (4 * KiB)
27
+ tcg_temp_free_i64(dirty_addr);
28
29
- /* Note that unpredicated load/store of vector/predicate registers
30
+ /*
31
+ * Note that unpredicated load/store of vector/predicate registers
32
* are defined as a stream of bytes, which equates to little-endian
33
- * operations on larger quantities. There is no nice way to force
34
- * a little-endian load for aarch64_be-linux-user out of line.
35
- *
36
+ * operations on larger quantities.
37
* Attempt to keep code expansion to a minimum by limiting the
38
* amount of unrolling done.
39
*/
40
if (nparts <= 4) {
41
int i;
42
43
+ t0 = tcg_temp_new_i64();
44
for (i = 0; i < len_align; i += 8) {
45
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
46
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
47
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
48
tcg_gen_st_i64(t0, cpu_env, vofs + i);
49
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
50
}
51
+ tcg_temp_free_i64(t0);
52
} else {
53
TCGLabel *loop = gen_new_label();
54
TCGv_ptr tp, i = tcg_const_local_ptr(0);
55
56
+ /* Copy the clean address into a local temp, live across the loop. */
57
+ t0 = clean_addr;
58
+ clean_addr = tcg_temp_local_new_i64();
59
+ tcg_gen_mov_i64(clean_addr, t0);
60
+ tcg_temp_free_i64(t0);
41
+
61
+
42
AddressSpace *arm_boot_address_space(ARMCPU *cpu,
62
gen_set_label(loop);
43
const struct arm_boot_info *info)
63
44
{
64
- /* Minimize the number of local temps that must be re-read from
45
@@ -XXX,XX +XXX,XX @@ static void write_bootloader(const char *name, hwaddr addr,
65
- * the stack each iteration. Instead, re-compute values other
46
code[i] = tswap32(insn);
66
- * than the loop counter.
67
- */
68
+ t0 = tcg_temp_new_i64();
69
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
70
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
71
+
72
tp = tcg_temp_new_ptr();
73
- tcg_gen_addi_ptr(tp, i, imm);
74
- tcg_gen_extu_ptr_i64(addr, tp);
75
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
76
-
77
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
78
-
79
tcg_gen_add_ptr(tp, cpu_env, i);
80
tcg_gen_addi_ptr(i, i, 8);
81
tcg_gen_st_i64(t0, tp, vofs);
82
tcg_temp_free_ptr(tp);
83
+ tcg_temp_free_i64(t0);
84
85
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
86
tcg_temp_free_ptr(i);
47
}
87
}
48
88
49
+ assert((len * sizeof(uint32_t)) < BOOTLOADER_MAX_SIZE);
89
- /* Predicate register loads can be any multiple of 2.
50
+
90
+ /*
51
rom_add_blob_fixed_as(name, code, len * sizeof(uint32_t), addr, as);
91
+ * Predicate register loads can be any multiple of 2.
52
92
* Note that we still store the entire 64-bit unit into cpu_env.
53
g_free(code);
93
*/
54
@@ -XXX,XX +XXX,XX @@ static uint64_t load_aarch64_image(const char *filename, hwaddr mem_base,
94
if (len_remain) {
55
memcpy(&hdrvals, buffer + ARM64_TEXT_OFFSET_OFFSET, sizeof(hdrvals));
95
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
56
if (hdrvals[1] != 0) {
96
-
57
kernel_load_offset = le64_to_cpu(hdrvals[0]);
97
+ t0 = tcg_temp_new_i64();
58
+
98
switch (len_remain) {
59
+ /*
99
case 2:
60
+ * We write our startup "bootloader" at the very bottom of RAM,
100
case 4:
61
+ * so that bit can't be used for the image. Luckily the Image
101
case 8:
62
+ * format specification is that the image requests only an offset
102
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
63
+ * from a 2MB boundary, not an absolute load address. So if the
103
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx,
64
+ * image requests an offset that might mean it overlaps with the
104
+ MO_LE | ctz32(len_remain));
65
+ * bootloader, we can just load it starting at 2MB+offset rather
105
break;
66
+ * than 0MB + offset.
106
67
+ */
107
case 6:
68
+ if (kernel_load_offset < BOOTLOADER_MAX_SIZE) {
108
t1 = tcg_temp_new_i64();
69
+ kernel_load_offset += 2 * MiB;
109
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL);
70
+ }
110
- tcg_gen_addi_i64(addr, addr, 4);
111
- tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW);
112
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUL);
113
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
114
+ tcg_gen_qemu_ld_i64(t1, clean_addr, midx, MO_LEUW);
115
tcg_gen_deposit_i64(t0, t0, t1, 32, 32);
116
tcg_temp_free_i64(t1);
117
break;
118
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
119
g_assert_not_reached();
71
}
120
}
121
tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
122
+ tcg_temp_free_i64(t0);
72
}
123
}
73
124
- tcg_temp_free_i64(addr);
125
- tcg_temp_free_i64(t0);
126
+ tcg_temp_free_i64(clean_addr);
127
}
128
129
/* Similarly for stores. */
74
--
130
--
75
2.19.1
131
2.20.1
76
132
77
133
diff view generated by jsdifflib
1
From: Markus Armbruster <armbru@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Device models aren't supposed to go on fishing expeditions for
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
backends. They should expose suitable properties for the user to set.
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
For onboard devices, board code sets them.
5
Message-id: 20200626033144.790098-30-richard.henderson@linaro.org
6
7
Device ssi-sd picks up its block backend in its init() method with
8
drive_get_next() instead. This mistake is already marked FIXME since
9
commit af9e40a.
10
11
Unset user_creatable to remove the mistake from our external
12
interface. Since the SSI bus doesn't support hotplug, only -device
13
can be affected. Only certain ARM machines have ssi-sd and provide an
14
SSI bus for it; this patch breaks -device ssi-sd for these machines.
15
No actual use of -device ssi-sd is known.
16
17
Signed-off-by: Markus Armbruster <armbru@redhat.com>
18
Acked-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
19
Acked-by: Thomas Huth <thuth@redhat.com>
20
Message-id: 20181009060835.4608-1-armbru@redhat.com
21
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
22
---
7
---
23
hw/sd/ssi-sd.c | 2 ++
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
24
1 file changed, 2 insertions(+)
9
1 file changed, 33 insertions(+), 28 deletions(-)
25
10
26
diff --git a/hw/sd/ssi-sd.c b/hw/sd/ssi-sd.c
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
27
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
28
--- a/hw/sd/ssi-sd.c
13
--- a/target/arm/translate-sve.c
29
+++ b/hw/sd/ssi-sd.c
14
+++ b/target/arm/translate-sve.c
30
@@ -XXX,XX +XXX,XX @@ static void ssi_sd_class_init(ObjectClass *klass, void *data)
15
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
31
k->cs_polarity = SSI_CS_LOW;
16
int len_remain = len % 8;
32
dc->vmsd = &vmstate_ssi_sd;
17
int nparts = len / 8 + ctpop8(len_remain);
33
dc->reset = ssi_sd_reset;
18
int midx = get_mem_index(s);
34
+ /* Reason: init() method uses drive_get_next() */
19
- TCGv_i64 addr, t0;
35
+ dc->user_creatable = false;
20
+ TCGv_i64 dirty_addr, clean_addr, t0;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
/* Note that unpredicated load/store of vector/predicate registers
30
* are defined as a stream of bytes, which equates to little-endian
31
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
32
if (nparts <= 4) {
33
int i;
34
35
+ t0 = tcg_temp_new_i64();
36
for (i = 0; i < len_align; i += 8) {
37
tcg_gen_ld_i64(t0, cpu_env, vofs + i);
38
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
39
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
40
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
41
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
42
}
43
+ tcg_temp_free_i64(t0);
44
} else {
45
TCGLabel *loop = gen_new_label();
46
- TCGv_ptr t2, i = tcg_const_local_ptr(0);
47
+ TCGv_ptr tp, i = tcg_const_local_ptr(0);
48
+
49
+ /* Copy the clean address into a local temp, live across the loop. */
50
+ t0 = clean_addr;
51
+ clean_addr = tcg_temp_local_new_i64();
52
+ tcg_gen_mov_i64(clean_addr, t0);
53
+ tcg_temp_free_i64(t0);
54
55
gen_set_label(loop);
56
57
- t2 = tcg_temp_new_ptr();
58
- tcg_gen_add_ptr(t2, cpu_env, i);
59
- tcg_gen_ld_i64(t0, t2, vofs);
60
-
61
- /* Minimize the number of local temps that must be re-read from
62
- * the stack each iteration. Instead, re-compute values other
63
- * than the loop counter.
64
- */
65
- tcg_gen_addi_ptr(t2, i, imm);
66
- tcg_gen_extu_ptr_i64(addr, t2);
67
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
68
- tcg_temp_free_ptr(t2);
69
-
70
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
71
-
72
+ t0 = tcg_temp_new_i64();
73
+ tp = tcg_temp_new_ptr();
74
+ tcg_gen_add_ptr(tp, cpu_env, i);
75
+ tcg_gen_ld_i64(t0, tp, vofs);
76
tcg_gen_addi_ptr(i, i, 8);
77
+ tcg_temp_free_ptr(tp);
78
+
79
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
80
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
81
+ tcg_temp_free_i64(t0);
82
83
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
84
tcg_temp_free_ptr(i);
85
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
86
87
/* Predicate register stores can be any multiple of 2. */
88
if (len_remain) {
89
+ t0 = tcg_temp_new_i64();
90
tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
91
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
92
93
switch (len_remain) {
94
case 2:
95
case 4:
96
case 8:
97
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
98
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx,
99
+ MO_LE | ctz32(len_remain));
100
break;
101
102
case 6:
103
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUL);
104
- tcg_gen_addi_i64(addr, addr, 4);
105
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUL);
106
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
107
tcg_gen_shri_i64(t0, t0, 32);
108
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUW);
109
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUW);
110
break;
111
112
default:
113
g_assert_not_reached();
114
}
115
+ tcg_temp_free_i64(t0);
116
}
117
- tcg_temp_free_i64(addr);
118
- tcg_temp_free_i64(t0);
119
+ tcg_temp_free_i64(clean_addr);
36
}
120
}
37
121
38
static const TypeInfo ssi_sd_info = {
122
static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
39
--
123
--
40
2.19.1
124
2.20.1
41
125
42
126
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-11-richard.henderson@linaro.org
5
Message-id: 20200626033144.790098-31-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
7
---
8
target/arm/translate.c | 16 ++++++++--------
8
target/arm/translate-sve.c | 6 ++++--
9
1 file changed, 8 insertions(+), 8 deletions(-)
9
1 file changed, 4 insertions(+), 2 deletions(-)
10
10
11
diff --git a/target/arm/translate.c b/target/arm/translate.c
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate.c
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
15
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
16
tcg_temp_free_ptr(ptr1);
16
unsigned esz = dtype_esz[a->dtype];
17
tcg_temp_free_ptr(ptr2);
17
unsigned msz = dtype_msz(a->dtype);
18
break;
18
TCGLabel *over = gen_new_label();
19
- TCGv_i64 temp;
20
+ TCGv_i64 temp, clean_addr;
21
22
/* If the guarding predicate has no bits set, no load occurs. */
23
if (psz <= 8) {
24
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
25
/* Load the data. */
26
temp = tcg_temp_new_i64();
27
tcg_gen_addi_i64(temp, cpu_reg_sp(s, a->rn), a->imm << msz);
28
- tcg_gen_qemu_ld_i64(temp, temp, get_mem_index(s),
29
+ clean_addr = gen_mte_check1(s, temp, false, true, msz);
19
+
30
+
20
+ case NEON_2RM_VMVN:
31
+ tcg_gen_qemu_ld_i64(temp, clean_addr, get_mem_index(s),
21
+ tcg_gen_gvec_not(0, rd_ofs, rm_ofs, vec_size, vec_size);
32
s->be_data | dtype_mop[a->dtype]);
22
+ break;
33
23
+ case NEON_2RM_VNEG:
34
/* Broadcast to *all* elements. */
24
+ tcg_gen_gvec_neg(size, rd_ofs, rm_ofs, vec_size, vec_size);
25
+ break;
26
+
27
default:
28
elementwise:
29
for (pass = 0; pass < (q ? 4 : 2); pass++) {
30
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
31
case NEON_2RM_VCNT:
32
gen_helper_neon_cnt_u8(tmp, tmp);
33
break;
34
- case NEON_2RM_VMVN:
35
- tcg_gen_not_i32(tmp, tmp);
36
- break;
37
case NEON_2RM_VQABS:
38
switch (size) {
39
case 0:
40
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
41
default: abort();
42
}
43
break;
44
- case NEON_2RM_VNEG:
45
- tmp2 = tcg_const_i32(0);
46
- gen_neon_rsb(size, tmp, tmp2);
47
- tcg_temp_free_i32(tmp2);
48
- break;
49
case NEON_2RM_VCGT0_F:
50
{
51
TCGv_ptr fpstatus = get_fpstatus_ptr(1);
52
--
35
--
53
2.19.1
36
2.20.1
54
37
55
38
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move the variable declarations to the top of the function,
4
but do not create a new label before sve_access_check.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-10-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-32-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
target/arm/translate.c | 29 ++++++++++-------------------
11
target/arm/translate-sve.c | 12 +++++++-----
9
1 file changed, 10 insertions(+), 19 deletions(-)
12
1 file changed, 7 insertions(+), 5 deletions(-)
10
13
11
diff --git a/target/arm/translate.c b/target/arm/translate.c
14
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate.c
16
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate.c
17
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
18
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RQ_zpri(DisasContext *s, arg_rpri_load *a)
16
break;
19
/* Load and broadcast element. */
17
}
20
static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
18
return 0;
21
{
22
- if (!sve_access_check(s)) {
23
- return true;
24
- }
25
-
26
unsigned vsz = vec_full_reg_size(s);
27
unsigned psz = pred_full_reg_size(s);
28
unsigned esz = dtype_esz[a->dtype];
29
unsigned msz = dtype_msz(a->dtype);
30
- TCGLabel *over = gen_new_label();
31
+ TCGLabel *over;
32
TCGv_i64 temp, clean_addr;
33
34
+ if (!sve_access_check(s)) {
35
+ return true;
36
+ }
19
+
37
+
20
+ case NEON_3R_VADD_VSUB:
38
+ over = gen_new_label();
21
+ if (u) {
39
+
22
+ tcg_gen_gvec_sub(size, rd_ofs, rn_ofs, rm_ofs,
40
/* If the guarding predicate has no bits set, no load occurs. */
23
+ vec_size, vec_size);
41
if (psz <= 8) {
24
+ } else {
42
/* Reduce the pred_esz_masks value simply to reduce the
25
+ tcg_gen_gvec_add(size, rd_ofs, rn_ofs, rm_ofs,
26
+ vec_size, vec_size);
27
+ }
28
+ return 0;
29
}
30
if (size == 3) {
31
/* 64-bit element instructions. */
32
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
33
cpu_V1, cpu_V0);
34
}
35
break;
36
- case NEON_3R_VADD_VSUB:
37
- if (u) {
38
- tcg_gen_sub_i64(CPU_V001);
39
- } else {
40
- tcg_gen_add_i64(CPU_V001);
41
- }
42
- break;
43
default:
44
abort();
45
}
46
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
47
tmp2 = neon_load_reg(rd, pass);
48
gen_neon_add(size, tmp, tmp2);
49
break;
50
- case NEON_3R_VADD_VSUB:
51
- if (!u) { /* VADD */
52
- gen_neon_add(size, tmp, tmp2);
53
- } else { /* VSUB */
54
- switch (size) {
55
- case 0: gen_helper_neon_sub_u8(tmp, tmp, tmp2); break;
56
- case 1: gen_helper_neon_sub_u16(tmp, tmp, tmp2); break;
57
- case 2: tcg_gen_sub_i32(tmp, tmp, tmp2); break;
58
- default: abort();
59
- }
60
- }
61
- break;
62
case NEON_3R_VTST_VCEQ:
63
if (!u) { /* VTST */
64
switch (size) {
65
--
43
--
66
2.19.1
44
2.20.1
67
45
68
46
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Most of the v8 extensions are self-contained within the ISAR
3
Introduce an lvalue macro to wrap target_tlb_bit0.
4
registers and are not implied by other feature bits, which
5
makes them the easiest to convert.
6
4
7
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20181016223115.24100-4-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-33-richard.henderson@linaro.org
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
9
---
13
target/arm/cpu.h | 131 +++++++++++++++++++++++++++++++++----
10
target/arm/cpu.h | 13 +++++++++++++
14
target/arm/translate.h | 7 ++
11
target/arm/helper.c | 2 +-
15
linux-user/elfload.c | 46 ++++++++-----
12
target/arm/translate-a64.c | 2 +-
16
target/arm/cpu.c | 27 +++++---
13
3 files changed, 15 insertions(+), 2 deletions(-)
17
target/arm/cpu64.c | 57 +++++++++-------
18
target/arm/translate-a64.c | 101 ++++++++++++++--------------
19
target/arm/translate.c | 36 +++++-----
20
7 files changed, 273 insertions(+), 132 deletions(-)
21
14
22
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
23
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
25
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
26
@@ -XXX,XX +XXX,XX @@ typedef enum ARMPSCIState {
27
PSCI_ON_PENDING = 2
28
} ARMPSCIState;
29
30
+typedef struct ARMISARegisters ARMISARegisters;
31
+
32
/**
33
* ARMCPU:
34
* @env: #CPUARMState
35
@@ -XXX,XX +XXX,XX @@ enum arm_features {
36
ARM_FEATURE_LPAE, /* has Large Physical Address Extension */
37
ARM_FEATURE_V8,
38
ARM_FEATURE_AARCH64, /* supports 64 bit mode */
39
- ARM_FEATURE_V8_AES, /* implements AES part of v8 Crypto Extensions */
40
ARM_FEATURE_CBAR, /* has cp15 CBAR */
41
ARM_FEATURE_CRC, /* ARMv8 CRC instructions */
42
ARM_FEATURE_CBAR_RO, /* has cp15 CBAR and it is read-only */
43
ARM_FEATURE_EL2, /* has EL2 Virtualization support */
44
ARM_FEATURE_EL3, /* has EL3 Secure monitor support */
45
- ARM_FEATURE_V8_SHA1, /* implements SHA1 part of v8 Crypto Extensions */
46
- ARM_FEATURE_V8_SHA256, /* implements SHA256 part of v8 Crypto Extensions */
47
- ARM_FEATURE_V8_PMULL, /* implements PMULL part of v8 Crypto Extensions */
48
ARM_FEATURE_THUMB_DSP, /* DSP insns supported in the Thumb encodings */
49
ARM_FEATURE_PMU, /* has PMU support */
50
ARM_FEATURE_VBAR, /* has cp15 VBAR */
51
ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
52
ARM_FEATURE_JAZELLE, /* has (trivial) Jazelle implementation */
53
ARM_FEATURE_SVE, /* has Scalable Vector Extension */
54
- ARM_FEATURE_V8_SHA512, /* implements SHA512 part of v8 Crypto Extensions */
55
- ARM_FEATURE_V8_SHA3, /* implements SHA3 part of v8 Crypto Extensions */
56
- ARM_FEATURE_V8_SM3, /* implements SM3 part of v8 Crypto Extensions */
57
- ARM_FEATURE_V8_SM4, /* implements SM4 part of v8 Crypto Extensions */
58
- ARM_FEATURE_V8_ATOMICS, /* ARMv8.1-Atomics feature */
59
- ARM_FEATURE_V8_RDM, /* implements v8.1 simd round multiply */
60
- ARM_FEATURE_V8_DOTPROD, /* implements v8.2 simd dot product */
61
ARM_FEATURE_V8_FP16, /* implements v8.2 half-precision float */
62
- ARM_FEATURE_V8_FCMA, /* has complex number part of v8.3 extensions. */
63
ARM_FEATURE_M_MAIN, /* M profile Main Extension */
64
};
65
66
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
67
/* Shared between translate-sve.c and sve_helper.c. */
20
/* Shared between translate-sve.c and sve_helper.c. */
68
extern const uint64_t pred_esz_masks[4];
21
extern const uint64_t pred_esz_masks[4];
69
22
70
+/*
23
+/* Helper for the macros below, validating the argument type. */
71
+ * 32-bit feature tests via id registers.
24
+static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
72
+ */
73
+static inline bool isar_feature_aa32_aes(const ARMISARegisters *id)
74
+{
25
+{
75
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, AES) != 0;
26
+ return x;
76
+}
77
+
78
+static inline bool isar_feature_aa32_pmull(const ARMISARegisters *id)
79
+{
80
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, AES) > 1;
81
+}
82
+
83
+static inline bool isar_feature_aa32_sha1(const ARMISARegisters *id)
84
+{
85
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, SHA1) != 0;
86
+}
87
+
88
+static inline bool isar_feature_aa32_sha2(const ARMISARegisters *id)
89
+{
90
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, SHA2) != 0;
91
+}
92
+
93
+static inline bool isar_feature_aa32_crc32(const ARMISARegisters *id)
94
+{
95
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, CRC32) != 0;
96
+}
97
+
98
+static inline bool isar_feature_aa32_rdm(const ARMISARegisters *id)
99
+{
100
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, RDM) != 0;
101
+}
102
+
103
+static inline bool isar_feature_aa32_vcma(const ARMISARegisters *id)
104
+{
105
+ return FIELD_EX32(id->id_isar5, ID_ISAR5, VCMA) != 0;
106
+}
107
+
108
+static inline bool isar_feature_aa32_dp(const ARMISARegisters *id)
109
+{
110
+ return FIELD_EX32(id->id_isar6, ID_ISAR6, DP) != 0;
111
+}
27
+}
112
+
28
+
113
+/*
29
+/*
114
+ * 64-bit feature tests via id registers.
30
+ * Lvalue macros for ARM TLB bits that we must cache in the TCG TLB.
31
+ * Using these should be a bit more self-documenting than using the
32
+ * generic target bits directly.
115
+ */
33
+ */
116
+static inline bool isar_feature_aa64_aes(const ARMISARegisters *id)
34
+#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
117
+{
118
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, AES) != 0;
119
+}
120
+
35
+
121
+static inline bool isar_feature_aa64_pmull(const ARMISARegisters *id)
36
/*
122
+{
37
* Naming convention for isar_feature functions:
123
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, AES) > 1;
38
* Functions which test 32-bit ID registers should have _aa32_ in
124
+}
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
125
+
126
+static inline bool isar_feature_aa64_sha1(const ARMISARegisters *id)
127
+{
128
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SHA1) != 0;
129
+}
130
+
131
+static inline bool isar_feature_aa64_sha256(const ARMISARegisters *id)
132
+{
133
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SHA2) != 0;
134
+}
135
+
136
+static inline bool isar_feature_aa64_sha512(const ARMISARegisters *id)
137
+{
138
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SHA2) > 1;
139
+}
140
+
141
+static inline bool isar_feature_aa64_crc32(const ARMISARegisters *id)
142
+{
143
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, CRC32) != 0;
144
+}
145
+
146
+static inline bool isar_feature_aa64_atomics(const ARMISARegisters *id)
147
+{
148
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, ATOMIC) != 0;
149
+}
150
+
151
+static inline bool isar_feature_aa64_rdm(const ARMISARegisters *id)
152
+{
153
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, RDM) != 0;
154
+}
155
+
156
+static inline bool isar_feature_aa64_sha3(const ARMISARegisters *id)
157
+{
158
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SHA3) != 0;
159
+}
160
+
161
+static inline bool isar_feature_aa64_sm3(const ARMISARegisters *id)
162
+{
163
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SM3) != 0;
164
+}
165
+
166
+static inline bool isar_feature_aa64_sm4(const ARMISARegisters *id)
167
+{
168
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, SM4) != 0;
169
+}
170
+
171
+static inline bool isar_feature_aa64_dp(const ARMISARegisters *id)
172
+{
173
+ return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, DP) != 0;
174
+}
175
+
176
+static inline bool isar_feature_aa64_fcma(const ARMISARegisters *id)
177
+{
178
+ return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, FCMA) != 0;
179
+}
180
+
181
+/*
182
+ * Forward to the above feature tests given an ARMCPU pointer.
183
+ */
184
+#define cpu_isar_feature(name, cpu) \
185
+ ({ ARMCPU *cpu_ = (cpu); isar_feature_##name(&cpu_->isar); })
186
+
187
#endif
188
diff --git a/target/arm/translate.h b/target/arm/translate.h
189
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
190
--- a/target/arm/translate.h
41
--- a/target/arm/helper.c
191
+++ b/target/arm/translate.h
42
+++ b/target/arm/helper.c
192
@@ -XXX,XX +XXX,XX @@
43
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
193
/* internal defines */
194
typedef struct DisasContext {
195
DisasContextBase base;
196
+ const ARMISARegisters *isar;
197
198
target_ulong pc;
199
target_ulong page_start;
200
@@ -XXX,XX +XXX,XX @@ static inline TCGv_i32 get_ahp_flag(void)
201
return ret;
202
}
203
204
+/*
205
+ * Forward to the isar_feature_* tests given a DisasContext pointer.
206
+ */
207
+#define dc_isar_feature(name, ctx) \
208
+ ({ DisasContext *ctx_ = (ctx); isar_feature_##name(ctx_->isar); })
209
+
210
#endif /* TARGET_ARM_TRANSLATE_H */
211
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
212
index XXXXXXX..XXXXXXX 100644
213
--- a/linux-user/elfload.c
214
+++ b/linux-user/elfload.c
215
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
216
/* probe for the extra features */
217
#define GET_FEATURE(feat, hwcap) \
218
do { if (arm_feature(&cpu->env, feat)) { hwcaps |= hwcap; } } while (0)
219
+
220
+#define GET_FEATURE_ID(feat, hwcap) \
221
+ do { if (cpu_isar_feature(feat, cpu)) { hwcaps |= hwcap; } } while (0)
222
+
223
/* EDSP is in v5TE and above, but all our v5 CPUs are v5TE */
224
GET_FEATURE(ARM_FEATURE_V5, ARM_HWCAP_ARM_EDSP);
225
GET_FEATURE(ARM_FEATURE_VFP, ARM_HWCAP_ARM_VFP);
226
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap2(void)
227
ARMCPU *cpu = ARM_CPU(thread_cpu);
228
uint32_t hwcaps = 0;
229
230
- GET_FEATURE(ARM_FEATURE_V8_AES, ARM_HWCAP2_ARM_AES);
231
- GET_FEATURE(ARM_FEATURE_V8_PMULL, ARM_HWCAP2_ARM_PMULL);
232
- GET_FEATURE(ARM_FEATURE_V8_SHA1, ARM_HWCAP2_ARM_SHA1);
233
- GET_FEATURE(ARM_FEATURE_V8_SHA256, ARM_HWCAP2_ARM_SHA2);
234
- GET_FEATURE(ARM_FEATURE_CRC, ARM_HWCAP2_ARM_CRC32);
235
+ GET_FEATURE_ID(aa32_aes, ARM_HWCAP2_ARM_AES);
236
+ GET_FEATURE_ID(aa32_pmull, ARM_HWCAP2_ARM_PMULL);
237
+ GET_FEATURE_ID(aa32_sha1, ARM_HWCAP2_ARM_SHA1);
238
+ GET_FEATURE_ID(aa32_sha2, ARM_HWCAP2_ARM_SHA2);
239
+ GET_FEATURE_ID(aa32_crc32, ARM_HWCAP2_ARM_CRC32);
240
return hwcaps;
241
}
242
243
#undef GET_FEATURE
244
+#undef GET_FEATURE_ID
245
246
#else
247
/* 64 bit ARM definitions */
248
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
249
/* probe for the extra features */
250
#define GET_FEATURE(feat, hwcap) \
251
do { if (arm_feature(&cpu->env, feat)) { hwcaps |= hwcap; } } while (0)
252
- GET_FEATURE(ARM_FEATURE_V8_AES, ARM_HWCAP_A64_AES);
253
- GET_FEATURE(ARM_FEATURE_V8_PMULL, ARM_HWCAP_A64_PMULL);
254
- GET_FEATURE(ARM_FEATURE_V8_SHA1, ARM_HWCAP_A64_SHA1);
255
- GET_FEATURE(ARM_FEATURE_V8_SHA256, ARM_HWCAP_A64_SHA2);
256
- GET_FEATURE(ARM_FEATURE_CRC, ARM_HWCAP_A64_CRC32);
257
- GET_FEATURE(ARM_FEATURE_V8_SHA3, ARM_HWCAP_A64_SHA3);
258
- GET_FEATURE(ARM_FEATURE_V8_SM3, ARM_HWCAP_A64_SM3);
259
- GET_FEATURE(ARM_FEATURE_V8_SM4, ARM_HWCAP_A64_SM4);
260
- GET_FEATURE(ARM_FEATURE_V8_SHA512, ARM_HWCAP_A64_SHA512);
261
+#define GET_FEATURE_ID(feat, hwcap) \
262
+ do { if (cpu_isar_feature(feat, cpu)) { hwcaps |= hwcap; } } while (0)
263
+
264
+ GET_FEATURE_ID(aa64_aes, ARM_HWCAP_A64_AES);
265
+ GET_FEATURE_ID(aa64_pmull, ARM_HWCAP_A64_PMULL);
266
+ GET_FEATURE_ID(aa64_sha1, ARM_HWCAP_A64_SHA1);
267
+ GET_FEATURE_ID(aa64_sha256, ARM_HWCAP_A64_SHA2);
268
+ GET_FEATURE_ID(aa64_sha512, ARM_HWCAP_A64_SHA512);
269
+ GET_FEATURE_ID(aa64_crc32, ARM_HWCAP_A64_CRC32);
270
+ GET_FEATURE_ID(aa64_sha3, ARM_HWCAP_A64_SHA3);
271
+ GET_FEATURE_ID(aa64_sm3, ARM_HWCAP_A64_SM3);
272
+ GET_FEATURE_ID(aa64_sm4, ARM_HWCAP_A64_SM4);
273
GET_FEATURE(ARM_FEATURE_V8_FP16,
274
ARM_HWCAP_A64_FPHP | ARM_HWCAP_A64_ASIMDHP);
275
- GET_FEATURE(ARM_FEATURE_V8_ATOMICS, ARM_HWCAP_A64_ATOMICS);
276
- GET_FEATURE(ARM_FEATURE_V8_RDM, ARM_HWCAP_A64_ASIMDRDM);
277
- GET_FEATURE(ARM_FEATURE_V8_DOTPROD, ARM_HWCAP_A64_ASIMDDP);
278
- GET_FEATURE(ARM_FEATURE_V8_FCMA, ARM_HWCAP_A64_FCMA);
279
+ GET_FEATURE_ID(aa64_atomics, ARM_HWCAP_A64_ATOMICS);
280
+ GET_FEATURE_ID(aa64_rdm, ARM_HWCAP_A64_ASIMDRDM);
281
+ GET_FEATURE_ID(aa64_dp, ARM_HWCAP_A64_ASIMDDP);
282
+ GET_FEATURE_ID(aa64_fcma, ARM_HWCAP_A64_FCMA);
283
GET_FEATURE(ARM_FEATURE_SVE, ARM_HWCAP_A64_SVE);
284
+
285
#undef GET_FEATURE
286
+#undef GET_FEATURE_ID
287
288
return hwcaps;
289
}
290
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
291
index XXXXXXX..XXXXXXX 100644
292
--- a/target/arm/cpu.c
293
+++ b/target/arm/cpu.c
294
@@ -XXX,XX +XXX,XX @@ static void arm_max_initfn(Object *obj)
295
cortex_a15_initfn(obj);
296
#ifdef CONFIG_USER_ONLY
297
/* We don't set these in system emulation mode for the moment,
298
- * since we don't correctly set the ID registers to advertise them,
299
+ * since we don't correctly set (all of) the ID registers to
300
+ * advertise them.
301
*/
302
set_feature(&cpu->env, ARM_FEATURE_V8);
303
- set_feature(&cpu->env, ARM_FEATURE_V8_AES);
304
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA1);
305
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA256);
306
- set_feature(&cpu->env, ARM_FEATURE_V8_PMULL);
307
- set_feature(&cpu->env, ARM_FEATURE_CRC);
308
- set_feature(&cpu->env, ARM_FEATURE_V8_RDM);
309
- set_feature(&cpu->env, ARM_FEATURE_V8_DOTPROD);
310
- set_feature(&cpu->env, ARM_FEATURE_V8_FCMA);
311
+ {
312
+ uint32_t t;
313
+
314
+ t = cpu->isar.id_isar5;
315
+ t = FIELD_DP32(t, ID_ISAR5, AES, 2);
316
+ t = FIELD_DP32(t, ID_ISAR5, SHA1, 1);
317
+ t = FIELD_DP32(t, ID_ISAR5, SHA2, 1);
318
+ t = FIELD_DP32(t, ID_ISAR5, CRC32, 1);
319
+ t = FIELD_DP32(t, ID_ISAR5, RDM, 1);
320
+ t = FIELD_DP32(t, ID_ISAR5, VCMA, 1);
321
+ cpu->isar.id_isar5 = t;
322
+
323
+ t = cpu->isar.id_isar6;
324
+ t = FIELD_DP32(t, ID_ISAR6, DP, 1);
325
+ cpu->isar.id_isar6 = t;
326
+ }
327
#endif
328
}
44
}
329
}
45
/* When in aarch64 mode, and BTI is enabled, remember GP in the IOTLB. */
330
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
46
if (aarch64 && guarded && cpu_isar_feature(aa64_bti, cpu)) {
331
index XXXXXXX..XXXXXXX 100644
47
- txattrs->target_tlb_bit0 = true;
332
--- a/target/arm/cpu64.c
48
+ arm_tlb_bti_gp(txattrs) = true;
333
+++ b/target/arm/cpu64.c
49
}
334
@@ -XXX,XX +XXX,XX @@ static void aarch64_a57_initfn(Object *obj)
50
335
set_feature(&cpu->env, ARM_FEATURE_GENERIC_TIMER);
51
if (cacheattrs != NULL) {
336
set_feature(&cpu->env, ARM_FEATURE_AARCH64);
337
set_feature(&cpu->env, ARM_FEATURE_CBAR_RO);
338
- set_feature(&cpu->env, ARM_FEATURE_V8_AES);
339
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA1);
340
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA256);
341
- set_feature(&cpu->env, ARM_FEATURE_V8_PMULL);
342
- set_feature(&cpu->env, ARM_FEATURE_CRC);
343
set_feature(&cpu->env, ARM_FEATURE_EL2);
344
set_feature(&cpu->env, ARM_FEATURE_EL3);
345
set_feature(&cpu->env, ARM_FEATURE_PMU);
346
@@ -XXX,XX +XXX,XX @@ static void aarch64_a53_initfn(Object *obj)
347
set_feature(&cpu->env, ARM_FEATURE_GENERIC_TIMER);
348
set_feature(&cpu->env, ARM_FEATURE_AARCH64);
349
set_feature(&cpu->env, ARM_FEATURE_CBAR_RO);
350
- set_feature(&cpu->env, ARM_FEATURE_V8_AES);
351
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA1);
352
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA256);
353
- set_feature(&cpu->env, ARM_FEATURE_V8_PMULL);
354
- set_feature(&cpu->env, ARM_FEATURE_CRC);
355
set_feature(&cpu->env, ARM_FEATURE_EL2);
356
set_feature(&cpu->env, ARM_FEATURE_EL3);
357
set_feature(&cpu->env, ARM_FEATURE_PMU);
358
@@ -XXX,XX +XXX,XX @@ static void aarch64_a72_initfn(Object *obj)
359
set_feature(&cpu->env, ARM_FEATURE_GENERIC_TIMER);
360
set_feature(&cpu->env, ARM_FEATURE_AARCH64);
361
set_feature(&cpu->env, ARM_FEATURE_CBAR_RO);
362
- set_feature(&cpu->env, ARM_FEATURE_V8_AES);
363
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA1);
364
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA256);
365
- set_feature(&cpu->env, ARM_FEATURE_V8_PMULL);
366
- set_feature(&cpu->env, ARM_FEATURE_CRC);
367
set_feature(&cpu->env, ARM_FEATURE_EL2);
368
set_feature(&cpu->env, ARM_FEATURE_EL3);
369
set_feature(&cpu->env, ARM_FEATURE_PMU);
370
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
371
if (kvm_enabled()) {
372
kvm_arm_set_cpu_features_from_host(cpu);
373
} else {
374
+ uint64_t t;
375
+ uint32_t u;
376
aarch64_a57_initfn(obj);
377
+
378
+ t = cpu->isar.id_aa64isar0;
379
+ t = FIELD_DP64(t, ID_AA64ISAR0, AES, 2); /* AES + PMULL */
380
+ t = FIELD_DP64(t, ID_AA64ISAR0, SHA1, 1);
381
+ t = FIELD_DP64(t, ID_AA64ISAR0, SHA2, 2); /* SHA512 */
382
+ t = FIELD_DP64(t, ID_AA64ISAR0, CRC32, 1);
383
+ t = FIELD_DP64(t, ID_AA64ISAR0, ATOMIC, 2);
384
+ t = FIELD_DP64(t, ID_AA64ISAR0, RDM, 1);
385
+ t = FIELD_DP64(t, ID_AA64ISAR0, SHA3, 1);
386
+ t = FIELD_DP64(t, ID_AA64ISAR0, SM3, 1);
387
+ t = FIELD_DP64(t, ID_AA64ISAR0, SM4, 1);
388
+ t = FIELD_DP64(t, ID_AA64ISAR0, DP, 1);
389
+ cpu->isar.id_aa64isar0 = t;
390
+
391
+ t = cpu->isar.id_aa64isar1;
392
+ t = FIELD_DP64(t, ID_AA64ISAR1, FCMA, 1);
393
+ cpu->isar.id_aa64isar1 = t;
394
+
395
+ /* Replicate the same data to the 32-bit id registers. */
396
+ u = cpu->isar.id_isar5;
397
+ u = FIELD_DP32(u, ID_ISAR5, AES, 2); /* AES + PMULL */
398
+ u = FIELD_DP32(u, ID_ISAR5, SHA1, 1);
399
+ u = FIELD_DP32(u, ID_ISAR5, SHA2, 1);
400
+ u = FIELD_DP32(u, ID_ISAR5, CRC32, 1);
401
+ u = FIELD_DP32(u, ID_ISAR5, RDM, 1);
402
+ u = FIELD_DP32(u, ID_ISAR5, VCMA, 1);
403
+ cpu->isar.id_isar5 = u;
404
+
405
+ u = cpu->isar.id_isar6;
406
+ u = FIELD_DP32(u, ID_ISAR6, DP, 1);
407
+ cpu->isar.id_isar6 = u;
408
+
409
#ifdef CONFIG_USER_ONLY
410
/* We don't set these in system emulation mode for the moment,
411
* since we don't correctly set the ID registers to advertise them,
412
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
413
* whereas the architecture requires them to be present in both if
414
* present in either.
415
*/
416
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA512);
417
- set_feature(&cpu->env, ARM_FEATURE_V8_SHA3);
418
- set_feature(&cpu->env, ARM_FEATURE_V8_SM3);
419
- set_feature(&cpu->env, ARM_FEATURE_V8_SM4);
420
- set_feature(&cpu->env, ARM_FEATURE_V8_ATOMICS);
421
- set_feature(&cpu->env, ARM_FEATURE_V8_RDM);
422
- set_feature(&cpu->env, ARM_FEATURE_V8_DOTPROD);
423
set_feature(&cpu->env, ARM_FEATURE_V8_FP16);
424
- set_feature(&cpu->env, ARM_FEATURE_V8_FCMA);
425
set_feature(&cpu->env, ARM_FEATURE_SVE);
426
/* For usermode -cpu max we can use a larger and more efficient DCZ
427
* blocksize since we don't have to follow what the hardware does.
428
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
429
index XXXXXXX..XXXXXXX 100644
53
index XXXXXXX..XXXXXXX 100644
430
--- a/target/arm/translate-a64.c
54
--- a/target/arm/translate-a64.c
431
+++ b/target/arm/translate-a64.c
55
+++ b/target/arm/translate-a64.c
432
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
56
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
433
}
57
* table entry even for that case.
434
if (rt2 == 31
58
*/
435
&& ((rt | rs) & 1) == 0
59
return (tlb_hit(entry->addr_code, addr) &&
436
- && arm_dc_feature(s, ARM_FEATURE_V8_ATOMICS)) {
60
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs.target_tlb_bit0);
437
+ && dc_isar_feature(aa64_atomics, s)) {
61
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
438
/* CASP / CASPL */
62
#endif
439
gen_compare_and_swap_pair(s, rs, rt, rn, size | 2);
63
}
440
return;
441
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
442
}
443
if (rt2 == 31
444
&& ((rt | rs) & 1) == 0
445
- && arm_dc_feature(s, ARM_FEATURE_V8_ATOMICS)) {
446
+ && dc_isar_feature(aa64_atomics, s)) {
447
/* CASPA / CASPAL */
448
gen_compare_and_swap_pair(s, rs, rt, rn, size | 2);
449
return;
450
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
451
case 0xb: /* CASL */
452
case 0xe: /* CASA */
453
case 0xf: /* CASAL */
454
- if (rt2 == 31 && arm_dc_feature(s, ARM_FEATURE_V8_ATOMICS)) {
455
+ if (rt2 == 31 && dc_isar_feature(aa64_atomics, s)) {
456
gen_compare_and_swap(s, rs, rt, rn, size);
457
return;
458
}
459
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
460
int rs = extract32(insn, 16, 5);
461
int rn = extract32(insn, 5, 5);
462
int o3_opc = extract32(insn, 12, 4);
463
- int feature = ARM_FEATURE_V8_ATOMICS;
464
TCGv_i64 tcg_rn, tcg_rs;
465
AtomicThreeOpFn *fn;
466
467
- if (is_vector) {
468
+ if (is_vector || !dc_isar_feature(aa64_atomics, s)) {
469
unallocated_encoding(s);
470
return;
471
}
472
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
473
unallocated_encoding(s);
474
return;
475
}
476
- if (!arm_dc_feature(s, feature)) {
477
- unallocated_encoding(s);
478
- return;
479
- }
480
481
if (rn == 31) {
482
gen_check_sp_alignment(s);
483
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
484
TCGv_i64 tcg_acc, tcg_val;
485
TCGv_i32 tcg_bytes;
486
487
- if (!arm_dc_feature(s, ARM_FEATURE_CRC)
488
+ if (!dc_isar_feature(aa64_crc32, s)
489
|| (sf == 1 && sz != 3)
490
|| (sf == 0 && sz == 3)) {
491
unallocated_encoding(s);
492
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_three_reg_same_extra(DisasContext *s,
493
bool u = extract32(insn, 29, 1);
494
TCGv_i32 ele1, ele2, ele3;
495
TCGv_i64 res;
496
- int feature;
497
+ bool feature;
498
499
switch (u * 16 + opcode) {
500
case 0x10: /* SQRDMLAH (vector) */
501
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_three_reg_same_extra(DisasContext *s,
502
unallocated_encoding(s);
503
return;
504
}
505
- feature = ARM_FEATURE_V8_RDM;
506
+ feature = dc_isar_feature(aa64_rdm, s);
507
break;
508
default:
509
unallocated_encoding(s);
510
return;
511
}
512
- if (!arm_dc_feature(s, feature)) {
513
+ if (!feature) {
514
unallocated_encoding(s);
515
return;
516
}
517
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn)
518
return;
519
}
520
if (size == 3) {
521
- if (!arm_dc_feature(s, ARM_FEATURE_V8_PMULL)) {
522
+ if (!dc_isar_feature(aa64_pmull, s)) {
523
unallocated_encoding(s);
524
return;
525
}
526
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn)
527
int size = extract32(insn, 22, 2);
528
bool u = extract32(insn, 29, 1);
529
bool is_q = extract32(insn, 30, 1);
530
- int feature, rot;
531
+ bool feature;
532
+ int rot;
533
534
switch (u * 16 + opcode) {
535
case 0x10: /* SQRDMLAH (vector) */
536
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn)
537
unallocated_encoding(s);
538
return;
539
}
540
- feature = ARM_FEATURE_V8_RDM;
541
+ feature = dc_isar_feature(aa64_rdm, s);
542
break;
543
case 0x02: /* SDOT (vector) */
544
case 0x12: /* UDOT (vector) */
545
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn)
546
unallocated_encoding(s);
547
return;
548
}
549
- feature = ARM_FEATURE_V8_DOTPROD;
550
+ feature = dc_isar_feature(aa64_dp, s);
551
break;
552
case 0x18: /* FCMLA, #0 */
553
case 0x19: /* FCMLA, #90 */
554
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn)
555
unallocated_encoding(s);
556
return;
557
}
558
- feature = ARM_FEATURE_V8_FCMA;
559
+ feature = dc_isar_feature(aa64_fcma, s);
560
break;
561
default:
562
unallocated_encoding(s);
563
return;
564
}
565
- if (!arm_dc_feature(s, feature)) {
566
+ if (!feature) {
567
unallocated_encoding(s);
568
return;
569
}
570
@@ -XXX,XX +XXX,XX @@ static void disas_simd_indexed(DisasContext *s, uint32_t insn)
571
break;
572
case 0x1d: /* SQRDMLAH */
573
case 0x1f: /* SQRDMLSH */
574
- if (!arm_dc_feature(s, ARM_FEATURE_V8_RDM)) {
575
+ if (!dc_isar_feature(aa64_rdm, s)) {
576
unallocated_encoding(s);
577
return;
578
}
579
break;
580
case 0x0e: /* SDOT */
581
case 0x1e: /* UDOT */
582
- if (size != MO_32 || !arm_dc_feature(s, ARM_FEATURE_V8_DOTPROD)) {
583
+ if (size != MO_32 || !dc_isar_feature(aa64_dp, s)) {
584
unallocated_encoding(s);
585
return;
586
}
587
@@ -XXX,XX +XXX,XX @@ static void disas_simd_indexed(DisasContext *s, uint32_t insn)
588
case 0x13: /* FCMLA #90 */
589
case 0x15: /* FCMLA #180 */
590
case 0x17: /* FCMLA #270 */
591
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FCMA)) {
592
+ if (!dc_isar_feature(aa64_fcma, s)) {
593
unallocated_encoding(s);
594
return;
595
}
596
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_aes(DisasContext *s, uint32_t insn)
597
TCGv_i32 tcg_decrypt;
598
CryptoThreeOpIntFn *genfn;
599
600
- if (!arm_dc_feature(s, ARM_FEATURE_V8_AES)
601
- || size != 0) {
602
+ if (!dc_isar_feature(aa64_aes, s) || size != 0) {
603
unallocated_encoding(s);
604
return;
605
}
606
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_three_reg_sha(DisasContext *s, uint32_t insn)
607
int rd = extract32(insn, 0, 5);
608
CryptoThreeOpFn *genfn;
609
TCGv_ptr tcg_rd_ptr, tcg_rn_ptr, tcg_rm_ptr;
610
- int feature = ARM_FEATURE_V8_SHA256;
611
+ bool feature;
612
613
if (size != 0) {
614
unallocated_encoding(s);
615
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_three_reg_sha(DisasContext *s, uint32_t insn)
616
case 2: /* SHA1M */
617
case 3: /* SHA1SU0 */
618
genfn = NULL;
619
- feature = ARM_FEATURE_V8_SHA1;
620
+ feature = dc_isar_feature(aa64_sha1, s);
621
break;
622
case 4: /* SHA256H */
623
genfn = gen_helper_crypto_sha256h;
624
+ feature = dc_isar_feature(aa64_sha256, s);
625
break;
626
case 5: /* SHA256H2 */
627
genfn = gen_helper_crypto_sha256h2;
628
+ feature = dc_isar_feature(aa64_sha256, s);
629
break;
630
case 6: /* SHA256SU1 */
631
genfn = gen_helper_crypto_sha256su1;
632
+ feature = dc_isar_feature(aa64_sha256, s);
633
break;
634
default:
635
unallocated_encoding(s);
636
return;
637
}
638
639
- if (!arm_dc_feature(s, feature)) {
640
+ if (!feature) {
641
unallocated_encoding(s);
642
return;
643
}
644
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_two_reg_sha(DisasContext *s, uint32_t insn)
645
int rn = extract32(insn, 5, 5);
646
int rd = extract32(insn, 0, 5);
647
CryptoTwoOpFn *genfn;
648
- int feature;
649
+ bool feature;
650
TCGv_ptr tcg_rd_ptr, tcg_rn_ptr;
651
652
if (size != 0) {
653
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_two_reg_sha(DisasContext *s, uint32_t insn)
654
655
switch (opcode) {
656
case 0: /* SHA1H */
657
- feature = ARM_FEATURE_V8_SHA1;
658
+ feature = dc_isar_feature(aa64_sha1, s);
659
genfn = gen_helper_crypto_sha1h;
660
break;
661
case 1: /* SHA1SU1 */
662
- feature = ARM_FEATURE_V8_SHA1;
663
+ feature = dc_isar_feature(aa64_sha1, s);
664
genfn = gen_helper_crypto_sha1su1;
665
break;
666
case 2: /* SHA256SU0 */
667
- feature = ARM_FEATURE_V8_SHA256;
668
+ feature = dc_isar_feature(aa64_sha256, s);
669
genfn = gen_helper_crypto_sha256su0;
670
break;
671
default:
672
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_two_reg_sha(DisasContext *s, uint32_t insn)
673
return;
674
}
675
676
- if (!arm_dc_feature(s, feature)) {
677
+ if (!feature) {
678
unallocated_encoding(s);
679
return;
680
}
681
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_three_reg_sha512(DisasContext *s, uint32_t insn)
682
int rm = extract32(insn, 16, 5);
683
int rn = extract32(insn, 5, 5);
684
int rd = extract32(insn, 0, 5);
685
- int feature;
686
+ bool feature;
687
CryptoThreeOpFn *genfn;
688
689
if (o == 0) {
690
switch (opcode) {
691
case 0: /* SHA512H */
692
- feature = ARM_FEATURE_V8_SHA512;
693
+ feature = dc_isar_feature(aa64_sha512, s);
694
genfn = gen_helper_crypto_sha512h;
695
break;
696
case 1: /* SHA512H2 */
697
- feature = ARM_FEATURE_V8_SHA512;
698
+ feature = dc_isar_feature(aa64_sha512, s);
699
genfn = gen_helper_crypto_sha512h2;
700
break;
701
case 2: /* SHA512SU1 */
702
- feature = ARM_FEATURE_V8_SHA512;
703
+ feature = dc_isar_feature(aa64_sha512, s);
704
genfn = gen_helper_crypto_sha512su1;
705
break;
706
case 3: /* RAX1 */
707
- feature = ARM_FEATURE_V8_SHA3;
708
+ feature = dc_isar_feature(aa64_sha3, s);
709
genfn = NULL;
710
break;
711
}
712
} else {
713
switch (opcode) {
714
case 0: /* SM3PARTW1 */
715
- feature = ARM_FEATURE_V8_SM3;
716
+ feature = dc_isar_feature(aa64_sm3, s);
717
genfn = gen_helper_crypto_sm3partw1;
718
break;
719
case 1: /* SM3PARTW2 */
720
- feature = ARM_FEATURE_V8_SM3;
721
+ feature = dc_isar_feature(aa64_sm3, s);
722
genfn = gen_helper_crypto_sm3partw2;
723
break;
724
case 2: /* SM4EKEY */
725
- feature = ARM_FEATURE_V8_SM4;
726
+ feature = dc_isar_feature(aa64_sm4, s);
727
genfn = gen_helper_crypto_sm4ekey;
728
break;
729
default:
730
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_three_reg_sha512(DisasContext *s, uint32_t insn)
731
}
732
}
733
734
- if (!arm_dc_feature(s, feature)) {
735
+ if (!feature) {
736
unallocated_encoding(s);
737
return;
738
}
739
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_two_reg_sha512(DisasContext *s, uint32_t insn)
740
int rn = extract32(insn, 5, 5);
741
int rd = extract32(insn, 0, 5);
742
TCGv_ptr tcg_rd_ptr, tcg_rn_ptr;
743
- int feature;
744
+ bool feature;
745
CryptoTwoOpFn *genfn;
746
747
switch (opcode) {
748
case 0: /* SHA512SU0 */
749
- feature = ARM_FEATURE_V8_SHA512;
750
+ feature = dc_isar_feature(aa64_sha512, s);
751
genfn = gen_helper_crypto_sha512su0;
752
break;
753
case 1: /* SM4E */
754
- feature = ARM_FEATURE_V8_SM4;
755
+ feature = dc_isar_feature(aa64_sm4, s);
756
genfn = gen_helper_crypto_sm4e;
757
break;
758
default:
759
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_two_reg_sha512(DisasContext *s, uint32_t insn)
760
return;
761
}
762
763
- if (!arm_dc_feature(s, feature)) {
764
+ if (!feature) {
765
unallocated_encoding(s);
766
return;
767
}
768
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_four_reg(DisasContext *s, uint32_t insn)
769
int ra = extract32(insn, 10, 5);
770
int rn = extract32(insn, 5, 5);
771
int rd = extract32(insn, 0, 5);
772
- int feature;
773
+ bool feature;
774
775
switch (op0) {
776
case 0: /* EOR3 */
777
case 1: /* BCAX */
778
- feature = ARM_FEATURE_V8_SHA3;
779
+ feature = dc_isar_feature(aa64_sha3, s);
780
break;
781
case 2: /* SM3SS1 */
782
- feature = ARM_FEATURE_V8_SM3;
783
+ feature = dc_isar_feature(aa64_sm3, s);
784
break;
785
default:
786
unallocated_encoding(s);
787
return;
788
}
789
790
- if (!arm_dc_feature(s, feature)) {
791
+ if (!feature) {
792
unallocated_encoding(s);
793
return;
794
}
795
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_xar(DisasContext *s, uint32_t insn)
796
TCGv_i64 tcg_op1, tcg_op2, tcg_res[2];
797
int pass;
798
799
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA3)) {
800
+ if (!dc_isar_feature(aa64_sha3, s)) {
801
unallocated_encoding(s);
802
return;
803
}
804
@@ -XXX,XX +XXX,XX @@ static void disas_crypto_three_reg_imm2(DisasContext *s, uint32_t insn)
805
TCGv_ptr tcg_rd_ptr, tcg_rn_ptr, tcg_rm_ptr;
806
TCGv_i32 tcg_imm2, tcg_opcode;
807
808
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SM3)) {
809
+ if (!dc_isar_feature(aa64_sm3, s)) {
810
unallocated_encoding(s);
811
return;
812
}
813
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
814
ARMCPU *arm_cpu = arm_env_get_cpu(env);
815
int bound;
816
817
+ dc->isar = &arm_cpu->isar;
818
dc->pc = dc->base.pc_first;
819
dc->condjmp = 0;
820
821
diff --git a/target/arm/translate.c b/target/arm/translate.c
822
index XXXXXXX..XXXXXXX 100644
823
--- a/target/arm/translate.c
824
+++ b/target/arm/translate.c
825
@@ -XXX,XX +XXX,XX @@ static const uint8_t neon_2rm_sizes[] = {
826
static int do_v81_helper(DisasContext *s, gen_helper_gvec_3_ptr *fn,
827
int q, int rd, int rn, int rm)
828
{
829
- if (arm_dc_feature(s, ARM_FEATURE_V8_RDM)) {
830
+ if (dc_isar_feature(aa32_rdm, s)) {
831
int opr_sz = (1 + q) * 8;
832
tcg_gen_gvec_3_ptr(vfp_reg_offset(1, rd),
833
vfp_reg_offset(1, rn),
834
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
835
return 1;
836
}
837
if (!u) { /* SHA-1 */
838
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA1)) {
839
+ if (!dc_isar_feature(aa32_sha1, s)) {
840
return 1;
841
}
842
ptr1 = vfp_reg_ptr(true, rd);
843
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
844
gen_helper_crypto_sha1_3reg(ptr1, ptr2, ptr3, tmp4);
845
tcg_temp_free_i32(tmp4);
846
} else { /* SHA-256 */
847
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA256) || size == 3) {
848
+ if (!dc_isar_feature(aa32_sha2, s) || size == 3) {
849
return 1;
850
}
851
ptr1 = vfp_reg_ptr(true, rd);
852
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
853
if (op == 14 && size == 2) {
854
TCGv_i64 tcg_rn, tcg_rm, tcg_rd;
855
856
- if (!arm_dc_feature(s, ARM_FEATURE_V8_PMULL)) {
857
+ if (!dc_isar_feature(aa32_pmull, s)) {
858
return 1;
859
}
860
tcg_rn = tcg_temp_new_i64();
861
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
862
{
863
NeonGenThreeOpEnvFn *fn;
864
865
- if (!arm_dc_feature(s, ARM_FEATURE_V8_RDM)) {
866
+ if (!dc_isar_feature(aa32_rdm, s)) {
867
return 1;
868
}
869
if (u && ((rd | rn) & 1)) {
870
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
871
break;
872
}
873
case NEON_2RM_AESE: case NEON_2RM_AESMC:
874
- if (!arm_dc_feature(s, ARM_FEATURE_V8_AES)
875
- || ((rm | rd) & 1)) {
876
+ if (!dc_isar_feature(aa32_aes, s) || ((rm | rd) & 1)) {
877
return 1;
878
}
879
ptr1 = vfp_reg_ptr(true, rd);
880
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
881
tcg_temp_free_i32(tmp3);
882
break;
883
case NEON_2RM_SHA1H:
884
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA1)
885
- || ((rm | rd) & 1)) {
886
+ if (!dc_isar_feature(aa32_sha1, s) || ((rm | rd) & 1)) {
887
return 1;
888
}
889
ptr1 = vfp_reg_ptr(true, rd);
890
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
891
}
892
/* bit 6 (q): set -> SHA256SU0, cleared -> SHA1SU1 */
893
if (q) {
894
- if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA256)) {
895
+ if (!dc_isar_feature(aa32_sha2, s)) {
896
return 1;
897
}
898
- } else if (!arm_dc_feature(s, ARM_FEATURE_V8_SHA1)) {
899
+ } else if (!dc_isar_feature(aa32_sha1, s)) {
900
return 1;
901
}
902
ptr1 = vfp_reg_ptr(true, rd);
903
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
904
/* VCMLA -- 1111 110R R.1S .... .... 1000 ...0 .... */
905
int size = extract32(insn, 20, 1);
906
data = extract32(insn, 23, 2); /* rot */
907
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FCMA)
908
+ if (!dc_isar_feature(aa32_vcma, s)
909
|| (!size && !arm_dc_feature(s, ARM_FEATURE_V8_FP16))) {
910
return 1;
911
}
912
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
913
/* VCADD -- 1111 110R 1.0S .... .... 1000 ...0 .... */
914
int size = extract32(insn, 20, 1);
915
data = extract32(insn, 24, 1); /* rot */
916
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FCMA)
917
+ if (!dc_isar_feature(aa32_vcma, s)
918
|| (!size && !arm_dc_feature(s, ARM_FEATURE_V8_FP16))) {
919
return 1;
920
}
921
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
922
} else if ((insn & 0xfeb00f00) == 0xfc200d00) {
923
/* V[US]DOT -- 1111 1100 0.10 .... .... 1101 .Q.U .... */
924
bool u = extract32(insn, 4, 1);
925
- if (!arm_dc_feature(s, ARM_FEATURE_V8_DOTPROD)) {
926
+ if (!dc_isar_feature(aa32_dp, s)) {
927
return 1;
928
}
929
fn_gvec = u ? gen_helper_gvec_udot_b : gen_helper_gvec_sdot_b;
930
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
931
int size = extract32(insn, 23, 1);
932
int index;
933
934
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FCMA)) {
935
+ if (!dc_isar_feature(aa32_vcma, s)) {
936
return 1;
937
}
938
if (size == 0) {
939
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
940
} else if ((insn & 0xffb00f00) == 0xfe200d00) {
941
/* V[US]DOT -- 1111 1110 0.10 .... .... 1101 .Q.U .... */
942
int u = extract32(insn, 4, 1);
943
- if (!arm_dc_feature(s, ARM_FEATURE_V8_DOTPROD)) {
944
+ if (!dc_isar_feature(aa32_dp, s)) {
945
return 1;
946
}
947
fn_gvec = u ? gen_helper_gvec_udot_idx_b : gen_helper_gvec_sdot_idx_b;
948
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
949
* op1 == 3 is UNPREDICTABLE but handle as UNDEFINED.
950
* Bits 8, 10 and 11 should be zero.
951
*/
952
- if (!arm_dc_feature(s, ARM_FEATURE_CRC) || op1 == 0x3 ||
953
- (c & 0xd) != 0) {
954
+ if (!dc_isar_feature(aa32_crc32, s) || op1 == 0x3 || (c & 0xd) != 0) {
955
goto illegal_op;
956
}
957
958
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
959
case 0x28:
960
case 0x29:
961
case 0x2a:
962
- if (!arm_dc_feature(s, ARM_FEATURE_CRC)) {
963
+ if (!dc_isar_feature(aa32_crc32, s)) {
964
goto illegal_op;
965
}
966
break;
967
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
968
CPUARMState *env = cs->env_ptr;
969
ARMCPU *cpu = arm_env_get_cpu(env);
970
971
+ dc->isar = &cpu->isar;
972
dc->pc = dc->base.pc_first;
973
dc->condjmp = 0;
974
64
975
--
65
--
976
2.19.1
66
2.20.1
977
67
978
68
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Both arm and thumb2 division are controlled by the same ISAR field,
3
Because the elements are sequential, we can eliminate many tests all
4
which takes care of the arm implies thumb case. Having M imply
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
thumb2 division was wrong for cortex-m0, which is v6m and does not
6
have thumb2 at all, much less thumb2 division.
7
5
8
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20181016223115.24100-5-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-34-richard.henderson@linaro.org
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
10
---
14
target/arm/cpu.h | 12 ++++++++++--
11
target/arm/cpu.h | 1 +
15
linux-user/elfload.c | 4 ++--
12
target/arm/helper-sve.h | 58 ++++++++++
16
target/arm/cpu.c | 10 +---------
13
target/arm/internals.h | 6 +
17
target/arm/translate.c | 4 ++--
14
target/arm/sve_helper.c | 218 ++++++++++++++++++++++++++++++-------
18
4 files changed, 15 insertions(+), 15 deletions(-)
15
target/arm/translate-sve.c | 186 ++++++++++++++++++++++---------
16
5 files changed, 378 insertions(+), 91 deletions(-)
19
17
20
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/cpu.h
20
--- a/target/arm/cpu.h
23
+++ b/target/arm/cpu.h
21
+++ b/target/arm/cpu.h
24
@@ -XXX,XX +XXX,XX @@ enum arm_features {
22
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
25
ARM_FEATURE_VFP3,
23
* generic target bits directly.
26
ARM_FEATURE_VFP_FP16,
24
*/
27
ARM_FEATURE_NEON,
25
#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
28
- ARM_FEATURE_THUMB_DIV, /* divide supported in Thumb encoding */
26
+#define arm_tlb_mte_tagged(x) (typecheck_memtxattrs(x)->target_tlb_bit1)
29
ARM_FEATURE_M, /* Microcontroller profile. */
27
30
ARM_FEATURE_OMAPCP, /* OMAP specific CP15 ops handling. */
31
ARM_FEATURE_THUMB2EE,
32
@@ -XXX,XX +XXX,XX @@ enum arm_features {
33
ARM_FEATURE_V5,
34
ARM_FEATURE_STRONGARM,
35
ARM_FEATURE_VAPA, /* cp15 VA to PA lookups */
36
- ARM_FEATURE_ARM_DIV, /* divide supported in ARM encoding */
37
ARM_FEATURE_VFP4, /* VFPv4 (implies that NEON is v2) */
38
ARM_FEATURE_GENERIC_TIMER,
39
ARM_FEATURE_MVFR, /* Media and VFP Feature Registers 0 and 1 */
40
@@ -XXX,XX +XXX,XX @@ extern const uint64_t pred_esz_masks[4];
41
/*
28
/*
42
* 32-bit feature tests via id registers.
29
* Naming convention for isar_feature functions:
30
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper-sve.h
33
+++ b/target/arm/helper-sve.h
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ld1sds_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
DEF_HELPER_FLAGS_4(sve_ld1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
38
+DEF_HELPER_FLAGS_4(sve_ld1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_4(sve_ld2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ld3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_ld4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ld1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_4(sve_ld2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ld3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_ld4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+
48
+DEF_HELPER_FLAGS_4(sve_ld1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ld2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_ld3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ld4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+
53
+DEF_HELPER_FLAGS_4(sve_ld1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_4(sve_ld2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_ld3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ld4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+
58
+DEF_HELPER_FLAGS_4(sve_ld1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
59
+DEF_HELPER_FLAGS_4(sve_ld2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_ld3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_ld4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_ld1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_ld2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ld3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_ld4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ld1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_ld2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ld3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_4(sve_ld4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_4(sve_ld1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_4(sve_ld1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_4(sve_ld1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_4(sve_ld1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_4(sve_ld1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_4(sve_ld1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
+
80
+DEF_HELPER_FLAGS_4(sve_ld1hsu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ld1hdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ld1hss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ld1hds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+
85
+DEF_HELPER_FLAGS_4(sve_ld1hsu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ld1hdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_4(sve_ld1hss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_4(sve_ld1hds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
89
+
90
+DEF_HELPER_FLAGS_4(sve_ld1sdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
91
+DEF_HELPER_FLAGS_4(sve_ld1sds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
92
+
93
+DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ld1sds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
95
+
96
DEF_HELPER_FLAGS_4(sve_ldff1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
97
DEF_HELPER_FLAGS_4(sve_ldff1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
98
DEF_HELPER_FLAGS_4(sve_ldff1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
99
diff --git a/target/arm/internals.h b/target/arm/internals.h
100
index XXXXXXX..XXXXXXX 100644
101
--- a/target/arm/internals.h
102
+++ b/target/arm/internals.h
103
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
104
#define LOG2_TAG_GRANULE 4
105
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
106
107
+/*
108
+ * The SVE simd_data field, for memory ops, contains either
109
+ * rd (5 bits) or a shift count (2 bits).
110
+ */
111
+#define SVE_MTEDESC_SHIFT 5
112
+
113
/* Bits within a descriptor passed to the helper_mte_check* functions. */
114
FIELD(MTEDESC, MIDX, 0, 4)
115
FIELD(MTEDESC, TBI, 4, 2)
116
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
117
index XXXXXXX..XXXXXXX 100644
118
--- a/target/arm/sve_helper.c
119
+++ b/target/arm/sve_helper.c
120
@@ -XXX,XX +XXX,XX @@ static void sve_cont_ldst_watchpoints(SVEContLdSt *info, CPUARMState *env,
121
#endif
122
}
123
124
+typedef uint64_t mte_check_fn(CPUARMState *, uint32_t, uint64_t, uintptr_t);
125
+
126
+static inline QEMU_ALWAYS_INLINE
127
+void sve_cont_ldst_mte_check_int(SVEContLdSt *info, CPUARMState *env,
128
+ uint64_t *vg, target_ulong addr, int esize,
129
+ int msize, uint32_t mtedesc, uintptr_t ra,
130
+ mte_check_fn *check)
131
+{
132
+ intptr_t mem_off, reg_off, reg_last;
133
+
134
+ /* Process the page only if MemAttr == Tagged. */
135
+ if (arm_tlb_mte_tagged(&info->page[0].attrs)) {
136
+ mem_off = info->mem_off_first[0];
137
+ reg_off = info->reg_off_first[0];
138
+ reg_last = info->reg_off_split;
139
+ if (reg_last < 0) {
140
+ reg_last = info->reg_off_last[0];
141
+ }
142
+
143
+ do {
144
+ uint64_t pg = vg[reg_off >> 6];
145
+ do {
146
+ if ((pg >> (reg_off & 63)) & 1) {
147
+ check(env, mtedesc, addr, ra);
148
+ }
149
+ reg_off += esize;
150
+ mem_off += msize;
151
+ } while (reg_off <= reg_last && (reg_off & 63));
152
+ } while (reg_off <= reg_last);
153
+ }
154
+
155
+ mem_off = info->mem_off_first[1];
156
+ if (mem_off >= 0 && arm_tlb_mte_tagged(&info->page[1].attrs)) {
157
+ reg_off = info->reg_off_first[1];
158
+ reg_last = info->reg_off_last[1];
159
+
160
+ do {
161
+ uint64_t pg = vg[reg_off >> 6];
162
+ do {
163
+ if ((pg >> (reg_off & 63)) & 1) {
164
+ check(env, mtedesc, addr, ra);
165
+ }
166
+ reg_off += esize;
167
+ mem_off += msize;
168
+ } while (reg_off & 63);
169
+ } while (reg_off <= reg_last);
170
+ }
171
+}
172
+
173
+typedef void sve_cont_ldst_mte_check_fn(SVEContLdSt *info, CPUARMState *env,
174
+ uint64_t *vg, target_ulong addr,
175
+ int esize, int msize, uint32_t mtedesc,
176
+ uintptr_t ra);
177
+
178
+static void sve_cont_ldst_mte_check1(SVEContLdSt *info, CPUARMState *env,
179
+ uint64_t *vg, target_ulong addr,
180
+ int esize, int msize, uint32_t mtedesc,
181
+ uintptr_t ra)
182
+{
183
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
184
+ mtedesc, ra, mte_check1);
185
+}
186
+
187
+static void sve_cont_ldst_mte_checkN(SVEContLdSt *info, CPUARMState *env,
188
+ uint64_t *vg, target_ulong addr,
189
+ int esize, int msize, uint32_t mtedesc,
190
+ uintptr_t ra)
191
+{
192
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
193
+ mtedesc, ra, mte_checkN);
194
+}
195
+
196
+
197
/*
198
* Common helper for all contiguous 1,2,3,4-register predicated stores.
43
*/
199
*/
44
+static inline bool isar_feature_thumb_div(const ARMISARegisters *id)
200
static inline QEMU_ALWAYS_INLINE
201
void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
202
uint32_t desc, const uintptr_t retaddr,
203
- const int esz, const int msz, const int N,
204
+ const int esz, const int msz, const int N, uint32_t mtedesc,
205
sve_ldst1_host_fn *host_fn,
206
- sve_ldst1_tlb_fn *tlb_fn)
207
+ sve_ldst1_tlb_fn *tlb_fn,
208
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
209
{
210
const unsigned rd = simd_data(desc);
211
const intptr_t reg_max = simd_oprsz(desc);
212
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
213
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
214
BP_MEM_READ, retaddr);
215
216
- /* TODO: MTE check. */
217
+ /*
218
+ * Handle mte checks for all active elements.
219
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
220
+ */
221
+ if (mte_check_fn && mtedesc) {
222
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
223
+ mtedesc, retaddr);
224
+ }
225
226
flags = info.page[0].flags | info.page[1].flags;
227
if (unlikely(flags != 0)) {
228
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
229
}
230
}
231
232
-#define DO_LD1_1(NAME, ESZ) \
233
-void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
234
- target_ulong addr, uint32_t desc) \
235
-{ \
236
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
237
- sve_##NAME##_host, sve_##NAME##_tlb); \
238
+static inline QEMU_ALWAYS_INLINE
239
+void sve_ldN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
240
+ uint32_t desc, const uintptr_t ra,
241
+ const int esz, const int msz, const int N,
242
+ sve_ldst1_host_fn *host_fn,
243
+ sve_ldst1_tlb_fn *tlb_fn)
45
+{
244
+{
46
+ return FIELD_EX32(id->id_isar0, ID_ISAR0, DIVIDE) != 0;
245
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
246
+ int bit55 = extract64(addr, 55, 1);
247
+
248
+ /* Remove mtedesc from the normal sve descriptor. */
249
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
250
+
251
+ /* Perform gross MTE suppression early. */
252
+ if (!tbi_check(desc, bit55) ||
253
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
254
+ mtedesc = 0;
255
+ }
256
+
257
+ sve_ldN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
258
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
259
}
260
261
-#define DO_LD1_2(NAME, ESZ, MSZ) \
262
-void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
263
- target_ulong addr, uint32_t desc) \
264
-{ \
265
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
266
- sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
267
-} \
268
-void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
269
- target_ulong addr, uint32_t desc) \
270
-{ \
271
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
272
- sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
273
+#define DO_LD1_1(NAME, ESZ) \
274
+void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
275
+ target_ulong addr, uint32_t desc) \
276
+{ \
277
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, 0, \
278
+ sve_##NAME##_host, sve_##NAME##_tlb, NULL); \
279
+} \
280
+void HELPER(sve_##NAME##_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
284
+ sve_##NAME##_host, sve_##NAME##_tlb); \
47
+}
285
+}
48
+
286
+
49
+static inline bool isar_feature_arm_div(const ARMISARegisters *id)
287
+#define DO_LD1_2(NAME, ESZ, MSZ) \
50
+{
288
+void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
51
+ return FIELD_EX32(id->id_isar0, ID_ISAR0, DIVIDE) > 1;
289
+ target_ulong addr, uint32_t desc) \
52
+}
290
+{ \
53
+
291
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
54
static inline bool isar_feature_aa32_aes(const ARMISARegisters *id)
292
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb, NULL); \
293
+} \
294
+void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
295
+ target_ulong addr, uint32_t desc) \
296
+{ \
297
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
298
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb, NULL); \
299
+} \
300
+void HELPER(sve_##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
301
+ target_ulong addr, uint32_t desc) \
302
+{ \
303
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
304
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
305
+} \
306
+void HELPER(sve_##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
307
+ target_ulong addr, uint32_t desc) \
308
+{ \
309
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
310
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
311
}
312
313
DO_LD1_1(ld1bb, MO_8)
314
@@ -XXX,XX +XXX,XX @@ DO_LD1_2(ld1dd, MO_64, MO_64)
315
#undef DO_LD1_1
316
#undef DO_LD1_2
317
318
-#define DO_LDN_1(N) \
319
-void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
320
- target_ulong addr, uint32_t desc) \
321
-{ \
322
- sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
323
- sve_ld1bb_host, sve_ld1bb_tlb); \
324
+#define DO_LDN_1(N) \
325
+void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
326
+ target_ulong addr, uint32_t desc) \
327
+{ \
328
+ sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, 0, \
329
+ sve_ld1bb_host, sve_ld1bb_tlb, NULL); \
330
+} \
331
+void HELPER(sve_ld##N##bb_r_mte)(CPUARMState *env, void *vg, \
332
+ target_ulong addr, uint32_t desc) \
333
+{ \
334
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
335
+ sve_ld1bb_host, sve_ld1bb_tlb); \
336
}
337
338
-#define DO_LDN_2(N, SUFF, ESZ) \
339
-void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
340
- target_ulong addr, uint32_t desc) \
341
-{ \
342
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
343
- sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
344
-} \
345
-void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
346
- target_ulong addr, uint32_t desc) \
347
-{ \
348
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
349
- sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
350
+#define DO_LDN_2(N, SUFF, ESZ) \
351
+void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
352
+ target_ulong addr, uint32_t desc) \
353
+{ \
354
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
355
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb, NULL); \
356
+} \
357
+void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
358
+ target_ulong addr, uint32_t desc) \
359
+{ \
360
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
361
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb, NULL); \
362
+} \
363
+void HELPER(sve_ld##N##SUFF##_le_r_mte)(CPUARMState *env, void *vg, \
364
+ target_ulong addr, uint32_t desc) \
365
+{ \
366
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
367
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
368
+} \
369
+void HELPER(sve_ld##N##SUFF##_be_r_mte)(CPUARMState *env, void *vg, \
370
+ target_ulong addr, uint32_t desc) \
371
+{ \
372
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
373
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
374
}
375
376
DO_LDN_1(2)
377
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
378
index XXXXXXX..XXXXXXX 100644
379
--- a/target/arm/translate-sve.c
380
+++ b/target/arm/translate-sve.c
381
@@ -XXX,XX +XXX,XX @@ static const uint8_t dtype_esz[16] = {
382
};
383
384
static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
385
- int dtype, gen_helper_gvec_mem *fn)
386
+ int dtype, uint32_t mte_n, bool is_write,
387
+ gen_helper_gvec_mem *fn)
55
{
388
{
56
return FIELD_EX32(id->id_isar5, ID_ISAR5, AES) != 0;
389
unsigned vsz = vec_full_reg_size(s);
57
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
390
TCGv_ptr t_pg;
58
index XXXXXXX..XXXXXXX 100644
391
TCGv_i32 t_desc;
59
--- a/linux-user/elfload.c
392
- int desc;
60
+++ b/linux-user/elfload.c
393
+ int desc = 0;
61
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
394
62
GET_FEATURE(ARM_FEATURE_VFP3, ARM_HWCAP_ARM_VFPv3);
395
- /* For e.g. LD4, there are not enough arguments to pass all 4
63
GET_FEATURE(ARM_FEATURE_V6K, ARM_HWCAP_ARM_TLS);
396
+ /*
64
GET_FEATURE(ARM_FEATURE_VFP4, ARM_HWCAP_ARM_VFPv4);
397
+ * For e.g. LD4, there are not enough arguments to pass all 4
65
- GET_FEATURE(ARM_FEATURE_ARM_DIV, ARM_HWCAP_ARM_IDIVA);
398
* registers as pointers, so encode the regno into the data field.
66
- GET_FEATURE(ARM_FEATURE_THUMB_DIV, ARM_HWCAP_ARM_IDIVT);
399
* For consistency, do this even for LD1.
67
+ GET_FEATURE_ID(arm_div, ARM_HWCAP_ARM_IDIVA);
400
+ * TODO: mte_n check here while callers are updated.
68
+ GET_FEATURE_ID(thumb_div, ARM_HWCAP_ARM_IDIVT);
401
*/
69
/* All QEMU's VFPv3 CPUs have 32 registers, see VFP_DREG in translate.c.
402
- desc = simd_desc(vsz, vsz, zt);
70
* Note that the ARM_HWCAP_ARM_VFPv3D16 bit is always the inverse of
403
+ if (mte_n && s->mte_active[0]) {
71
* ARM_HWCAP_ARM_VFPD32 (and so always clear for QEMU); it is unrelated
404
+ int msz = dtype_msz(dtype);
72
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
405
+
73
index XXXXXXX..XXXXXXX 100644
406
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
74
--- a/target/arm/cpu.c
407
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
75
+++ b/target/arm/cpu.c
408
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
76
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
409
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
77
* Presence of EL2 itself is ARM_FEATURE_EL2, and of the
410
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
78
* Security Extensions is ARM_FEATURE_EL3.
411
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
79
*/
412
+ desc <<= SVE_MTEDESC_SHIFT;
80
- set_feature(env, ARM_FEATURE_ARM_DIV);
413
+ }
81
+ assert(cpu_isar_feature(arm_div, cpu));
414
+ desc = simd_desc(vsz, vsz, zt | desc);
82
set_feature(env, ARM_FEATURE_LPAE);
415
t_desc = tcg_const_i32(desc);
83
set_feature(env, ARM_FEATURE_V7);
416
t_pg = tcg_temp_new_ptr();
417
418
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
419
static void do_ld_zpa(DisasContext *s, int zt, int pg,
420
TCGv_i64 addr, int dtype, int nreg)
421
{
422
- static gen_helper_gvec_mem * const fns[2][16][4] = {
423
- /* Little-endian */
424
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
425
+ static gen_helper_gvec_mem * const fns[2][2][16][4] = {
426
+ { /* mte inactive, little-endian */
427
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
428
gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
429
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
430
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
431
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
432
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
433
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
434
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
435
436
- { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
437
- { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
438
- gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
439
- { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
440
- { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
441
+ { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
442
+ { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
443
+ gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
444
+ { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
445
+ { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
446
447
- { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
448
- { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
449
- { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
450
- gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
451
- { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
452
+ { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
453
+ { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
454
+ { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
455
+ gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
456
+ { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
457
458
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
459
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
460
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
461
- { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
462
- gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
463
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
464
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
465
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
466
+ { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
467
+ gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
468
469
- /* Big-endian */
470
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
471
- gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
472
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
473
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
474
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
475
+ /* mte inactive, big-endian */
476
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
477
+ gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
478
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
479
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
480
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
481
482
- { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
483
- { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
484
- gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
485
- { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
486
- { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
487
+ { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
488
+ { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
489
+ gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
490
+ { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
491
+ { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
492
493
- { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
494
- { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
495
- { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
496
- gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
497
- { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
498
+ { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
499
+ { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
500
+ { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
501
+ gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
502
+ { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
503
504
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
505
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
506
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
507
- { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
508
- gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } }
509
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
510
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
511
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
512
+ { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
513
+ gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } } },
514
+
515
+ { /* mte active, little-endian */
516
+ { { gen_helper_sve_ld1bb_r_mte,
517
+ gen_helper_sve_ld2bb_r_mte,
518
+ gen_helper_sve_ld3bb_r_mte,
519
+ gen_helper_sve_ld4bb_r_mte },
520
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
521
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
522
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
523
+
524
+ { gen_helper_sve_ld1sds_le_r_mte, NULL, NULL, NULL },
525
+ { gen_helper_sve_ld1hh_le_r_mte,
526
+ gen_helper_sve_ld2hh_le_r_mte,
527
+ gen_helper_sve_ld3hh_le_r_mte,
528
+ gen_helper_sve_ld4hh_le_r_mte },
529
+ { gen_helper_sve_ld1hsu_le_r_mte, NULL, NULL, NULL },
530
+ { gen_helper_sve_ld1hdu_le_r_mte, NULL, NULL, NULL },
531
+
532
+ { gen_helper_sve_ld1hds_le_r_mte, NULL, NULL, NULL },
533
+ { gen_helper_sve_ld1hss_le_r_mte, NULL, NULL, NULL },
534
+ { gen_helper_sve_ld1ss_le_r_mte,
535
+ gen_helper_sve_ld2ss_le_r_mte,
536
+ gen_helper_sve_ld3ss_le_r_mte,
537
+ gen_helper_sve_ld4ss_le_r_mte },
538
+ { gen_helper_sve_ld1sdu_le_r_mte, NULL, NULL, NULL },
539
+
540
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
541
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
542
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
543
+ { gen_helper_sve_ld1dd_le_r_mte,
544
+ gen_helper_sve_ld2dd_le_r_mte,
545
+ gen_helper_sve_ld3dd_le_r_mte,
546
+ gen_helper_sve_ld4dd_le_r_mte } },
547
+
548
+ /* mte active, big-endian */
549
+ { { gen_helper_sve_ld1bb_r_mte,
550
+ gen_helper_sve_ld2bb_r_mte,
551
+ gen_helper_sve_ld3bb_r_mte,
552
+ gen_helper_sve_ld4bb_r_mte },
553
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
554
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
555
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
556
+
557
+ { gen_helper_sve_ld1sds_be_r_mte, NULL, NULL, NULL },
558
+ { gen_helper_sve_ld1hh_be_r_mte,
559
+ gen_helper_sve_ld2hh_be_r_mte,
560
+ gen_helper_sve_ld3hh_be_r_mte,
561
+ gen_helper_sve_ld4hh_be_r_mte },
562
+ { gen_helper_sve_ld1hsu_be_r_mte, NULL, NULL, NULL },
563
+ { gen_helper_sve_ld1hdu_be_r_mte, NULL, NULL, NULL },
564
+
565
+ { gen_helper_sve_ld1hds_be_r_mte, NULL, NULL, NULL },
566
+ { gen_helper_sve_ld1hss_be_r_mte, NULL, NULL, NULL },
567
+ { gen_helper_sve_ld1ss_be_r_mte,
568
+ gen_helper_sve_ld2ss_be_r_mte,
569
+ gen_helper_sve_ld3ss_be_r_mte,
570
+ gen_helper_sve_ld4ss_be_r_mte },
571
+ { gen_helper_sve_ld1sdu_be_r_mte, NULL, NULL, NULL },
572
+
573
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
574
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
575
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
576
+ { gen_helper_sve_ld1dd_be_r_mte,
577
+ gen_helper_sve_ld2dd_be_r_mte,
578
+ gen_helper_sve_ld3dd_be_r_mte,
579
+ gen_helper_sve_ld4dd_be_r_mte } } },
580
};
581
- gen_helper_gvec_mem *fn = fns[s->be_data == MO_BE][dtype][nreg];
582
+ gen_helper_gvec_mem *fn
583
+ = fns[s->mte_active[0]][s->be_data == MO_BE][dtype][nreg];
584
585
- /* While there are holes in the table, they are not
586
+ /*
587
+ * While there are holes in the table, they are not
588
* accessible via the instruction encoding.
589
*/
590
assert(fn != NULL);
591
- do_mem_zpa(s, zt, pg, addr, dtype, fn);
592
+ do_mem_zpa(s, zt, pg, addr, dtype, nreg, false, fn);
593
}
594
595
static bool trans_LD_zprr(DisasContext *s, arg_rprr_load *a)
596
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
597
TCGv_i64 addr = new_tmp_a64(s);
598
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
599
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
600
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
601
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
602
fns[s->be_data == MO_BE][a->dtype]);
84
}
603
}
85
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
604
return true;
86
if (arm_feature(env, ARM_FEATURE_V5)) {
605
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
87
set_feature(env, ARM_FEATURE_V4T);
606
TCGv_i64 addr = new_tmp_a64(s);
607
608
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
609
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
610
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
611
fns[s->be_data == MO_BE][a->dtype]);
88
}
612
}
89
- if (arm_feature(env, ARM_FEATURE_M)) {
613
return true;
90
- set_feature(env, ARM_FEATURE_THUMB_DIV);
614
@@ -XXX,XX +XXX,XX @@ static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
91
- }
615
fn = fn_multiple[be][nreg - 1][msz];
92
- if (arm_feature(env, ARM_FEATURE_ARM_DIV)) {
616
}
93
- set_feature(env, ARM_FEATURE_THUMB_DIV);
617
assert(fn != NULL);
94
- }
618
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), fn);
95
if (arm_feature(env, ARM_FEATURE_VFP4)) {
619
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
96
set_feature(env, ARM_FEATURE_VFP3);
620
}
97
set_feature(env, ARM_FEATURE_VFP_FP16);
621
98
@@ -XXX,XX +XXX,XX @@ static void cortex_r5_initfn(Object *obj)
622
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
99
ARMCPU *cpu = ARM_CPU(obj);
100
101
set_feature(&cpu->env, ARM_FEATURE_V7);
102
- set_feature(&cpu->env, ARM_FEATURE_THUMB_DIV);
103
- set_feature(&cpu->env, ARM_FEATURE_ARM_DIV);
104
set_feature(&cpu->env, ARM_FEATURE_V7MP);
105
set_feature(&cpu->env, ARM_FEATURE_PMSA);
106
cpu->midr = 0x411fc153; /* r1p3 */
107
diff --git a/target/arm/translate.c b/target/arm/translate.c
108
index XXXXXXX..XXXXXXX 100644
109
--- a/target/arm/translate.c
110
+++ b/target/arm/translate.c
111
@@ -XXX,XX +XXX,XX @@ static void disas_arm_insn(DisasContext *s, unsigned int insn)
112
case 1:
113
case 3:
114
/* SDIV, UDIV */
115
- if (!arm_dc_feature(s, ARM_FEATURE_ARM_DIV)) {
116
+ if (!dc_isar_feature(arm_div, s)) {
117
goto illegal_op;
118
}
119
if (((insn >> 5) & 7) || (rd != 15)) {
120
@@ -XXX,XX +XXX,XX @@ static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
121
tmp2 = load_reg(s, rm);
122
if ((op & 0x50) == 0x10) {
123
/* sdiv, udiv */
124
- if (!arm_dc_feature(s, ARM_FEATURE_THUMB_DIV)) {
125
+ if (!dc_isar_feature(thumb_div, s)) {
126
goto illegal_op;
127
}
128
if (op & 0x20)
129
--
623
--
130
2.19.1
624
2.20.1
131
625
132
626
diff view generated by jsdifflib
New patch
1
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-35-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-sve.h | 47 +++++++++++
12
target/arm/sve_helper.c | 95 ++++++++++++++++------
13
target/arm/translate-sve.c | 162 ++++++++++++++++++++++++-------------
14
3 files changed, 226 insertions(+), 78 deletions(-)
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_st1hd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
DEF_HELPER_FLAGS_4(sve_st1sd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_st1sd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_st1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_st2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_st3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_st4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+
29
+DEF_HELPER_FLAGS_4(sve_st1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_st2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_4(sve_st3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
32
+DEF_HELPER_FLAGS_4(sve_st4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
33
+
34
+DEF_HELPER_FLAGS_4(sve_st1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_4(sve_st2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_st3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_4(sve_st4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
+
39
+DEF_HELPER_FLAGS_4(sve_st1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_st2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_st3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+DEF_HELPER_FLAGS_4(sve_st4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
+
44
+DEF_HELPER_FLAGS_4(sve_st1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_st2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_st3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_st4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
+
49
+DEF_HELPER_FLAGS_4(sve_st1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_st2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_st3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_4(sve_st4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_st1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_st2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_st3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_4(sve_st4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_4(sve_st1bh_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_st1bs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_st1bd_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_st1hs_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_st1hd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_st1hs_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_st1hd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_st1sd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_st1sd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+
71
DEF_HELPER_FLAGS_6(sve_ldbsu_zsu, TCG_CALL_NO_WG,
72
void, env, ptr, ptr, ptr, tl, i32)
73
DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu, TCG_CALL_NO_WG,
74
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
75
index XXXXXXX..XXXXXXX 100644
76
--- a/target/arm/sve_helper.c
77
+++ b/target/arm/sve_helper.c
78
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_LDNF1_2(dd, MO_64, MO_64)
79
*/
80
81
static inline QEMU_ALWAYS_INLINE
82
-void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
83
- const uintptr_t retaddr, const int esz,
84
- const int msz, const int N,
85
+void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr,
86
+ uint32_t desc, const uintptr_t retaddr,
87
+ const int esz, const int msz, const int N, uint32_t mtedesc,
88
sve_ldst1_host_fn *host_fn,
89
- sve_ldst1_tlb_fn *tlb_fn)
90
+ sve_ldst1_tlb_fn *tlb_fn,
91
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
92
{
93
const unsigned rd = simd_data(desc);
94
const intptr_t reg_max = simd_oprsz(desc);
95
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
96
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
97
BP_MEM_WRITE, retaddr);
98
99
- /* TODO: MTE check. */
100
+ /*
101
+ * Handle mte checks for all active elements.
102
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
103
+ */
104
+ if (mte_check_fn && mtedesc) {
105
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
106
+ mtedesc, retaddr);
107
+ }
108
109
flags = info.page[0].flags | info.page[1].flags;
110
if (unlikely(flags != 0)) {
111
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
112
}
113
}
114
115
-#define DO_STN_1(N, NAME, ESZ) \
116
-void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
117
- target_ulong addr, uint32_t desc) \
118
-{ \
119
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
120
- sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
121
+static inline QEMU_ALWAYS_INLINE
122
+void sve_stN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
123
+ uint32_t desc, const uintptr_t ra,
124
+ const int esz, const int msz, const int N,
125
+ sve_ldst1_host_fn *host_fn,
126
+ sve_ldst1_tlb_fn *tlb_fn)
127
+{
128
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
129
+ int bit55 = extract64(addr, 55, 1);
130
+
131
+ /* Remove mtedesc from the normal sve descriptor. */
132
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
133
+
134
+ /* Perform gross MTE suppression early. */
135
+ if (!tbi_check(desc, bit55) ||
136
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
137
+ mtedesc = 0;
138
+ }
139
+
140
+ sve_stN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
141
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
142
}
143
144
-#define DO_STN_2(N, NAME, ESZ, MSZ) \
145
-void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
146
- target_ulong addr, uint32_t desc) \
147
-{ \
148
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
149
- sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
150
-} \
151
-void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
152
- target_ulong addr, uint32_t desc) \
153
-{ \
154
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
155
- sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
156
+#define DO_STN_1(N, NAME, ESZ) \
157
+void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
158
+ target_ulong addr, uint32_t desc) \
159
+{ \
160
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, 0, \
161
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb, NULL); \
162
+} \
163
+void HELPER(sve_st##N##NAME##_r_mte)(CPUARMState *env, void *vg, \
164
+ target_ulong addr, uint32_t desc) \
165
+{ \
166
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
167
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
168
+}
169
+
170
+#define DO_STN_2(N, NAME, ESZ, MSZ) \
171
+void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
172
+ target_ulong addr, uint32_t desc) \
173
+{ \
174
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
175
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb, NULL); \
176
+} \
177
+void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
178
+ target_ulong addr, uint32_t desc) \
179
+{ \
180
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
181
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb, NULL); \
182
+} \
183
+void HELPER(sve_st##N##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
184
+ target_ulong addr, uint32_t desc) \
185
+{ \
186
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
187
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
188
+} \
189
+void HELPER(sve_st##N##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
190
+ target_ulong addr, uint32_t desc) \
191
+{ \
192
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
193
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
194
}
195
196
DO_STN_1(1, bb, MO_8)
197
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
198
index XXXXXXX..XXXXXXX 100644
199
--- a/target/arm/translate-sve.c
200
+++ b/target/arm/translate-sve.c
201
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
202
static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
203
int msz, int esz, int nreg)
204
{
205
- static gen_helper_gvec_mem * const fn_single[2][4][4] = {
206
- { { gen_helper_sve_st1bb_r,
207
- gen_helper_sve_st1bh_r,
208
- gen_helper_sve_st1bs_r,
209
- gen_helper_sve_st1bd_r },
210
- { NULL,
211
- gen_helper_sve_st1hh_le_r,
212
- gen_helper_sve_st1hs_le_r,
213
- gen_helper_sve_st1hd_le_r },
214
- { NULL, NULL,
215
- gen_helper_sve_st1ss_le_r,
216
- gen_helper_sve_st1sd_le_r },
217
- { NULL, NULL, NULL,
218
- gen_helper_sve_st1dd_le_r } },
219
- { { gen_helper_sve_st1bb_r,
220
- gen_helper_sve_st1bh_r,
221
- gen_helper_sve_st1bs_r,
222
- gen_helper_sve_st1bd_r },
223
- { NULL,
224
- gen_helper_sve_st1hh_be_r,
225
- gen_helper_sve_st1hs_be_r,
226
- gen_helper_sve_st1hd_be_r },
227
- { NULL, NULL,
228
- gen_helper_sve_st1ss_be_r,
229
- gen_helper_sve_st1sd_be_r },
230
- { NULL, NULL, NULL,
231
- gen_helper_sve_st1dd_be_r } },
232
+ static gen_helper_gvec_mem * const fn_single[2][2][4][4] = {
233
+ { { { gen_helper_sve_st1bb_r,
234
+ gen_helper_sve_st1bh_r,
235
+ gen_helper_sve_st1bs_r,
236
+ gen_helper_sve_st1bd_r },
237
+ { NULL,
238
+ gen_helper_sve_st1hh_le_r,
239
+ gen_helper_sve_st1hs_le_r,
240
+ gen_helper_sve_st1hd_le_r },
241
+ { NULL, NULL,
242
+ gen_helper_sve_st1ss_le_r,
243
+ gen_helper_sve_st1sd_le_r },
244
+ { NULL, NULL, NULL,
245
+ gen_helper_sve_st1dd_le_r } },
246
+ { { gen_helper_sve_st1bb_r,
247
+ gen_helper_sve_st1bh_r,
248
+ gen_helper_sve_st1bs_r,
249
+ gen_helper_sve_st1bd_r },
250
+ { NULL,
251
+ gen_helper_sve_st1hh_be_r,
252
+ gen_helper_sve_st1hs_be_r,
253
+ gen_helper_sve_st1hd_be_r },
254
+ { NULL, NULL,
255
+ gen_helper_sve_st1ss_be_r,
256
+ gen_helper_sve_st1sd_be_r },
257
+ { NULL, NULL, NULL,
258
+ gen_helper_sve_st1dd_be_r } } },
259
+
260
+ { { { gen_helper_sve_st1bb_r_mte,
261
+ gen_helper_sve_st1bh_r_mte,
262
+ gen_helper_sve_st1bs_r_mte,
263
+ gen_helper_sve_st1bd_r_mte },
264
+ { NULL,
265
+ gen_helper_sve_st1hh_le_r_mte,
266
+ gen_helper_sve_st1hs_le_r_mte,
267
+ gen_helper_sve_st1hd_le_r_mte },
268
+ { NULL, NULL,
269
+ gen_helper_sve_st1ss_le_r_mte,
270
+ gen_helper_sve_st1sd_le_r_mte },
271
+ { NULL, NULL, NULL,
272
+ gen_helper_sve_st1dd_le_r_mte } },
273
+ { { gen_helper_sve_st1bb_r_mte,
274
+ gen_helper_sve_st1bh_r_mte,
275
+ gen_helper_sve_st1bs_r_mte,
276
+ gen_helper_sve_st1bd_r_mte },
277
+ { NULL,
278
+ gen_helper_sve_st1hh_be_r_mte,
279
+ gen_helper_sve_st1hs_be_r_mte,
280
+ gen_helper_sve_st1hd_be_r_mte },
281
+ { NULL, NULL,
282
+ gen_helper_sve_st1ss_be_r_mte,
283
+ gen_helper_sve_st1sd_be_r_mte },
284
+ { NULL, NULL, NULL,
285
+ gen_helper_sve_st1dd_be_r_mte } } },
286
};
287
- static gen_helper_gvec_mem * const fn_multiple[2][3][4] = {
288
- { { gen_helper_sve_st2bb_r,
289
- gen_helper_sve_st2hh_le_r,
290
- gen_helper_sve_st2ss_le_r,
291
- gen_helper_sve_st2dd_le_r },
292
- { gen_helper_sve_st3bb_r,
293
- gen_helper_sve_st3hh_le_r,
294
- gen_helper_sve_st3ss_le_r,
295
- gen_helper_sve_st3dd_le_r },
296
- { gen_helper_sve_st4bb_r,
297
- gen_helper_sve_st4hh_le_r,
298
- gen_helper_sve_st4ss_le_r,
299
- gen_helper_sve_st4dd_le_r } },
300
- { { gen_helper_sve_st2bb_r,
301
- gen_helper_sve_st2hh_be_r,
302
- gen_helper_sve_st2ss_be_r,
303
- gen_helper_sve_st2dd_be_r },
304
- { gen_helper_sve_st3bb_r,
305
- gen_helper_sve_st3hh_be_r,
306
- gen_helper_sve_st3ss_be_r,
307
- gen_helper_sve_st3dd_be_r },
308
- { gen_helper_sve_st4bb_r,
309
- gen_helper_sve_st4hh_be_r,
310
- gen_helper_sve_st4ss_be_r,
311
- gen_helper_sve_st4dd_be_r } },
312
+ static gen_helper_gvec_mem * const fn_multiple[2][2][3][4] = {
313
+ { { { gen_helper_sve_st2bb_r,
314
+ gen_helper_sve_st2hh_le_r,
315
+ gen_helper_sve_st2ss_le_r,
316
+ gen_helper_sve_st2dd_le_r },
317
+ { gen_helper_sve_st3bb_r,
318
+ gen_helper_sve_st3hh_le_r,
319
+ gen_helper_sve_st3ss_le_r,
320
+ gen_helper_sve_st3dd_le_r },
321
+ { gen_helper_sve_st4bb_r,
322
+ gen_helper_sve_st4hh_le_r,
323
+ gen_helper_sve_st4ss_le_r,
324
+ gen_helper_sve_st4dd_le_r } },
325
+ { { gen_helper_sve_st2bb_r,
326
+ gen_helper_sve_st2hh_be_r,
327
+ gen_helper_sve_st2ss_be_r,
328
+ gen_helper_sve_st2dd_be_r },
329
+ { gen_helper_sve_st3bb_r,
330
+ gen_helper_sve_st3hh_be_r,
331
+ gen_helper_sve_st3ss_be_r,
332
+ gen_helper_sve_st3dd_be_r },
333
+ { gen_helper_sve_st4bb_r,
334
+ gen_helper_sve_st4hh_be_r,
335
+ gen_helper_sve_st4ss_be_r,
336
+ gen_helper_sve_st4dd_be_r } } },
337
+ { { { gen_helper_sve_st2bb_r_mte,
338
+ gen_helper_sve_st2hh_le_r_mte,
339
+ gen_helper_sve_st2ss_le_r_mte,
340
+ gen_helper_sve_st2dd_le_r_mte },
341
+ { gen_helper_sve_st3bb_r_mte,
342
+ gen_helper_sve_st3hh_le_r_mte,
343
+ gen_helper_sve_st3ss_le_r_mte,
344
+ gen_helper_sve_st3dd_le_r_mte },
345
+ { gen_helper_sve_st4bb_r_mte,
346
+ gen_helper_sve_st4hh_le_r_mte,
347
+ gen_helper_sve_st4ss_le_r_mte,
348
+ gen_helper_sve_st4dd_le_r_mte } },
349
+ { { gen_helper_sve_st2bb_r_mte,
350
+ gen_helper_sve_st2hh_be_r_mte,
351
+ gen_helper_sve_st2ss_be_r_mte,
352
+ gen_helper_sve_st2dd_be_r_mte },
353
+ { gen_helper_sve_st3bb_r_mte,
354
+ gen_helper_sve_st3hh_be_r_mte,
355
+ gen_helper_sve_st3ss_be_r_mte,
356
+ gen_helper_sve_st3dd_be_r_mte },
357
+ { gen_helper_sve_st4bb_r_mte,
358
+ gen_helper_sve_st4hh_be_r_mte,
359
+ gen_helper_sve_st4ss_be_r_mte,
360
+ gen_helper_sve_st4dd_be_r_mte } } },
361
};
362
gen_helper_gvec_mem *fn;
363
int be = s->be_data == MO_BE;
364
365
if (nreg == 0) {
366
/* ST1 */
367
- fn = fn_single[be][msz][esz];
368
+ fn = fn_single[s->mte_active[0]][be][msz][esz];
369
+ nreg = 1;
370
} else {
371
/* ST2, ST3, ST4 -- msz == esz, enforced by encoding */
372
assert(msz == esz);
373
- fn = fn_multiple[be][nreg - 1][msz];
374
+ fn = fn_multiple[s->mte_active[0]][be][nreg - 1][msz];
375
}
376
assert(fn != NULL);
377
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
378
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), nreg, true, fn);
379
}
380
381
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
382
--
383
2.20.1
384
385
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Instead of shifts and masks, use direct loads and stores from the neon
3
Because the elements are sequential, we can eliminate many tests all
4
register file. Mirror the iteration structure of the ARM pseudocode
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
more closely. Correct the parameters of the VLD2 A2 insn.
6
5
7
Note that this includes a bugfix for handling of the insn
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
"VLD2 (multiple 2-element structures)" -- we were using an
9
incorrect stride value.
10
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20181011205206.3552-19-richard.henderson@linaro.org
8
Message-id: 20200626033144.790098-36-richard.henderson@linaro.org
13
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
---
10
---
16
target/arm/translate.c | 170 ++++++++++++++++++-----------------------
11
target/arm/helper-sve.h | 98 ++++++++++++++++
17
1 file changed, 74 insertions(+), 96 deletions(-)
12
target/arm/sve_helper.c | 99 ++++++++++++++--
13
target/arm/translate-sve.c | 232 +++++++++++++++++++++++++------------
14
3 files changed, 343 insertions(+), 86 deletions(-)
18
15
19
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
20
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/translate.c
18
--- a/target/arm/helper-sve.h
22
+++ b/target/arm/translate.c
19
+++ b/target/arm/helper-sve.h
23
@@ -XXX,XX +XXX,XX @@ static TCGv_i32 neon_load_reg(int reg, int pass)
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
24
return tmp;
21
DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_ldff1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_ldff1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_ldff1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_ldff1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+DEF_HELPER_FLAGS_4(sve_ldff1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_4(sve_ldff1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_ldff1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+
32
+DEF_HELPER_FLAGS_4(sve_ldff1hh_le_r_mte, TCG_CALL_NO_WG,
33
+ void, env, ptr, tl, i32)
34
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_le_r_mte, TCG_CALL_NO_WG,
35
+ void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_le_r_mte, TCG_CALL_NO_WG,
37
+ void, env, ptr, tl, i32)
38
+DEF_HELPER_FLAGS_4(sve_ldff1hss_le_r_mte, TCG_CALL_NO_WG,
39
+ void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ldff1hds_le_r_mte, TCG_CALL_NO_WG,
41
+ void, env, ptr, tl, i32)
42
+
43
+DEF_HELPER_FLAGS_4(sve_ldff1hh_be_r_mte, TCG_CALL_NO_WG,
44
+ void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_be_r_mte, TCG_CALL_NO_WG,
46
+ void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_be_r_mte, TCG_CALL_NO_WG,
48
+ void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ldff1hss_be_r_mte, TCG_CALL_NO_WG,
50
+ void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ldff1hds_be_r_mte, TCG_CALL_NO_WG,
52
+ void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_ldff1ss_le_r_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_le_r_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, tl, i32)
58
+DEF_HELPER_FLAGS_4(sve_ldff1sds_le_r_mte, TCG_CALL_NO_WG,
59
+ void, env, ptr, tl, i32)
60
+
61
+DEF_HELPER_FLAGS_4(sve_ldff1ss_be_r_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_be_r_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r_mte, TCG_CALL_NO_WG,
69
+ void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r_mte, TCG_CALL_NO_WG,
71
+ void, env, ptr, tl, i32)
72
+
73
DEF_HELPER_FLAGS_4(sve_ldnf1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
80
+DEF_HELPER_FLAGS_4(sve_ldnf1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ldnf1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_4(sve_ldnf1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_4(sve_ldnf1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ldnf1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+
88
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_le_r_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_le_r_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_le_r_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_le_r_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_le_r_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, tl, i32)
98
+
99
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_be_r_mte, TCG_CALL_NO_WG,
100
+ void, env, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_be_r_mte, TCG_CALL_NO_WG,
102
+ void, env, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_be_r_mte, TCG_CALL_NO_WG,
104
+ void, env, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_be_r_mte, TCG_CALL_NO_WG,
106
+ void, env, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_be_r_mte, TCG_CALL_NO_WG,
108
+ void, env, ptr, tl, i32)
109
+
110
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_le_r_mte, TCG_CALL_NO_WG,
111
+ void, env, ptr, tl, i32)
112
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_le_r_mte, TCG_CALL_NO_WG,
113
+ void, env, ptr, tl, i32)
114
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_le_r_mte, TCG_CALL_NO_WG,
115
+ void, env, ptr, tl, i32)
116
+
117
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_be_r_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_be_r_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, tl, i32)
123
+
124
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r_mte, TCG_CALL_NO_WG,
125
+ void, env, ptr, tl, i32)
126
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r_mte, TCG_CALL_NO_WG,
127
+ void, env, ptr, tl, i32)
128
+
129
DEF_HELPER_FLAGS_4(sve_st1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
130
DEF_HELPER_FLAGS_4(sve_st2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
131
DEF_HELPER_FLAGS_4(sve_st3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
132
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/target/arm/sve_helper.c
135
+++ b/target/arm/sve_helper.c
136
@@ -XXX,XX +XXX,XX @@ static void record_fault(CPUARMState *env, uintptr_t i, uintptr_t oprsz)
137
*/
138
static inline QEMU_ALWAYS_INLINE
139
void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
140
- uint32_t desc, const uintptr_t retaddr,
141
+ uint32_t desc, const uintptr_t retaddr, uint32_t mtedesc,
142
const int esz, const int msz, const SVEContFault fault,
143
sve_ldst1_host_fn *host_fn,
144
sve_ldst1_tlb_fn *tlb_fn)
145
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
146
mem_off = info.mem_off_first[0];
147
flags = info.page[0].flags;
148
149
+ /*
150
+ * Disable MTE checking if the Tagged bit is not set. Since TBI must
151
+ * be set within MTEDESC for MTE, !mtedesc => !mte_active.
152
+ */
153
+ if (arm_tlb_mte_tagged(&info.page[0].attrs)) {
154
+ mtedesc = 0;
155
+ }
156
+
157
if (fault == FAULT_FIRST) {
158
+ /* Trapping mte check for the first-fault element. */
159
+ if (mtedesc) {
160
+ mte_check1(env, mtedesc, addr + mem_off, retaddr);
161
+ }
162
+
163
/*
164
* Special handling of the first active element,
165
* if it crosses a page boundary or is MMIO.
166
*/
167
bool is_split = mem_off == info.mem_off_split;
168
- /* TODO: MTE check. */
169
if (unlikely(flags != 0) || unlikely(is_split)) {
170
/*
171
* Use the slow path for cross-page handling.
172
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
173
/* Watchpoint hit, see below. */
174
goto do_fault;
175
}
176
- /* TODO: MTE check. */
177
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
178
+ goto do_fault;
179
+ }
180
/*
181
* Use the slow path for cross-page handling.
182
* This is RAM, without a watchpoint, and will not trap.
183
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
184
& BP_MEM_READ)) {
185
goto do_fault;
186
}
187
- /* TODO: MTE check. */
188
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
189
+ goto do_fault;
190
+ }
191
host_fn(vd, reg_off, host + mem_off);
192
}
193
reg_off += 1 << esz;
194
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
195
record_fault(env, reg_off, reg_max);
25
}
196
}
26
197
27
+static void neon_load_element64(TCGv_i64 var, int reg, int ele, TCGMemOp mop)
198
-#define DO_LDFF1_LDNF1_1(PART, ESZ) \
199
+static inline QEMU_ALWAYS_INLINE
200
+void sve_ldnfff1_r_mte(CPUARMState *env, void *vg, target_ulong addr,
201
+ uint32_t desc, const uintptr_t retaddr,
202
+ const int esz, const int msz, const SVEContFault fault,
203
+ sve_ldst1_host_fn *host_fn,
204
+ sve_ldst1_tlb_fn *tlb_fn)
28
+{
205
+{
29
+ long offset = neon_element_offset(reg, ele, mop & MO_SIZE);
206
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
30
+
207
+ int bit55 = extract64(addr, 55, 1);
31
+ switch (mop) {
208
+
32
+ case MO_UB:
209
+ /* Remove mtedesc from the normal sve descriptor. */
33
+ tcg_gen_ld8u_i64(var, cpu_env, offset);
210
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
34
+ break;
211
+
35
+ case MO_UW:
212
+ /* Perform gross MTE suppression early. */
36
+ tcg_gen_ld16u_i64(var, cpu_env, offset);
213
+ if (!tbi_check(desc, bit55) ||
37
+ break;
214
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
38
+ case MO_UL:
215
+ mtedesc = 0;
39
+ tcg_gen_ld32u_i64(var, cpu_env, offset);
40
+ break;
41
+ case MO_Q:
42
+ tcg_gen_ld_i64(var, cpu_env, offset);
43
+ break;
44
+ default:
45
+ g_assert_not_reached();
46
+ }
216
+ }
217
+
218
+ sve_ldnfff1_r(env, vg, addr, desc, retaddr, mtedesc,
219
+ esz, msz, fault, host_fn, tlb_fn);
47
+}
220
+}
48
+
221
+
49
static void neon_store_reg(int reg, int pass, TCGv_i32 var)
222
+#define DO_LDFF1_LDNF1_1(PART, ESZ) \
223
void HELPER(sve_ldff1##PART##_r)(CPUARMState *env, void *vg, \
224
target_ulong addr, uint32_t desc) \
225
{ \
226
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
227
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_FIRST, \
228
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
229
} \
230
void HELPER(sve_ldnf1##PART##_r)(CPUARMState *env, void *vg, \
231
target_ulong addr, uint32_t desc) \
232
{ \
233
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
234
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_NO, \
235
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
236
+} \
237
+void HELPER(sve_ldff1##PART##_r_mte)(CPUARMState *env, void *vg, \
238
+ target_ulong addr, uint32_t desc) \
239
+{ \
240
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
241
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
242
+} \
243
+void HELPER(sve_ldnf1##PART##_r_mte)(CPUARMState *env, void *vg, \
244
+ target_ulong addr, uint32_t desc) \
245
+{ \
246
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
247
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
248
}
249
250
-#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
251
+#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
252
void HELPER(sve_ldff1##PART##_le_r)(CPUARMState *env, void *vg, \
253
target_ulong addr, uint32_t desc) \
254
{ \
255
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
256
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
257
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
258
} \
259
void HELPER(sve_ldnf1##PART##_le_r)(CPUARMState *env, void *vg, \
260
target_ulong addr, uint32_t desc) \
261
{ \
262
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
263
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
264
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
265
} \
266
void HELPER(sve_ldff1##PART##_be_r)(CPUARMState *env, void *vg, \
267
target_ulong addr, uint32_t desc) \
268
{ \
269
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
270
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
271
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
272
} \
273
void HELPER(sve_ldnf1##PART##_be_r)(CPUARMState *env, void *vg, \
274
target_ulong addr, uint32_t desc) \
275
{ \
276
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
277
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
278
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
279
+} \
280
+void HELPER(sve_ldff1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
284
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
285
+} \
286
+void HELPER(sve_ldnf1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
287
+ target_ulong addr, uint32_t desc) \
288
+{ \
289
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
290
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
291
+} \
292
+void HELPER(sve_ldff1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
293
+ target_ulong addr, uint32_t desc) \
294
+{ \
295
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
296
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
297
+} \
298
+void HELPER(sve_ldnf1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
299
+ target_ulong addr, uint32_t desc) \
300
+{ \
301
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
302
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
303
}
304
305
DO_LDFF1_LDNF1_1(bb, MO_8)
306
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
307
index XXXXXXX..XXXXXXX 100644
308
--- a/target/arm/translate-sve.c
309
+++ b/target/arm/translate-sve.c
310
@@ -XXX,XX +XXX,XX @@ static bool trans_LD_zpri(DisasContext *s, arg_rpri_load *a)
311
312
static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
50
{
313
{
51
tcg_gen_st_i32(var, cpu_env, neon_reg_offset(reg, pass));
314
- static gen_helper_gvec_mem * const fns[2][16] = {
52
tcg_temp_free_i32(var);
315
- /* Little-endian */
316
- { gen_helper_sve_ldff1bb_r,
317
- gen_helper_sve_ldff1bhu_r,
318
- gen_helper_sve_ldff1bsu_r,
319
- gen_helper_sve_ldff1bdu_r,
320
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
321
+ { /* mte inactive, little-endian */
322
+ { gen_helper_sve_ldff1bb_r,
323
+ gen_helper_sve_ldff1bhu_r,
324
+ gen_helper_sve_ldff1bsu_r,
325
+ gen_helper_sve_ldff1bdu_r,
326
327
- gen_helper_sve_ldff1sds_le_r,
328
- gen_helper_sve_ldff1hh_le_r,
329
- gen_helper_sve_ldff1hsu_le_r,
330
- gen_helper_sve_ldff1hdu_le_r,
331
+ gen_helper_sve_ldff1sds_le_r,
332
+ gen_helper_sve_ldff1hh_le_r,
333
+ gen_helper_sve_ldff1hsu_le_r,
334
+ gen_helper_sve_ldff1hdu_le_r,
335
336
- gen_helper_sve_ldff1hds_le_r,
337
- gen_helper_sve_ldff1hss_le_r,
338
- gen_helper_sve_ldff1ss_le_r,
339
- gen_helper_sve_ldff1sdu_le_r,
340
+ gen_helper_sve_ldff1hds_le_r,
341
+ gen_helper_sve_ldff1hss_le_r,
342
+ gen_helper_sve_ldff1ss_le_r,
343
+ gen_helper_sve_ldff1sdu_le_r,
344
345
- gen_helper_sve_ldff1bds_r,
346
- gen_helper_sve_ldff1bss_r,
347
- gen_helper_sve_ldff1bhs_r,
348
- gen_helper_sve_ldff1dd_le_r },
349
+ gen_helper_sve_ldff1bds_r,
350
+ gen_helper_sve_ldff1bss_r,
351
+ gen_helper_sve_ldff1bhs_r,
352
+ gen_helper_sve_ldff1dd_le_r },
353
354
- /* Big-endian */
355
- { gen_helper_sve_ldff1bb_r,
356
- gen_helper_sve_ldff1bhu_r,
357
- gen_helper_sve_ldff1bsu_r,
358
- gen_helper_sve_ldff1bdu_r,
359
+ /* mte inactive, big-endian */
360
+ { gen_helper_sve_ldff1bb_r,
361
+ gen_helper_sve_ldff1bhu_r,
362
+ gen_helper_sve_ldff1bsu_r,
363
+ gen_helper_sve_ldff1bdu_r,
364
365
- gen_helper_sve_ldff1sds_be_r,
366
- gen_helper_sve_ldff1hh_be_r,
367
- gen_helper_sve_ldff1hsu_be_r,
368
- gen_helper_sve_ldff1hdu_be_r,
369
+ gen_helper_sve_ldff1sds_be_r,
370
+ gen_helper_sve_ldff1hh_be_r,
371
+ gen_helper_sve_ldff1hsu_be_r,
372
+ gen_helper_sve_ldff1hdu_be_r,
373
374
- gen_helper_sve_ldff1hds_be_r,
375
- gen_helper_sve_ldff1hss_be_r,
376
- gen_helper_sve_ldff1ss_be_r,
377
- gen_helper_sve_ldff1sdu_be_r,
378
+ gen_helper_sve_ldff1hds_be_r,
379
+ gen_helper_sve_ldff1hss_be_r,
380
+ gen_helper_sve_ldff1ss_be_r,
381
+ gen_helper_sve_ldff1sdu_be_r,
382
383
- gen_helper_sve_ldff1bds_r,
384
- gen_helper_sve_ldff1bss_r,
385
- gen_helper_sve_ldff1bhs_r,
386
- gen_helper_sve_ldff1dd_be_r },
387
+ gen_helper_sve_ldff1bds_r,
388
+ gen_helper_sve_ldff1bss_r,
389
+ gen_helper_sve_ldff1bhs_r,
390
+ gen_helper_sve_ldff1dd_be_r } },
391
+
392
+ { /* mte active, little-endian */
393
+ { gen_helper_sve_ldff1bb_r_mte,
394
+ gen_helper_sve_ldff1bhu_r_mte,
395
+ gen_helper_sve_ldff1bsu_r_mte,
396
+ gen_helper_sve_ldff1bdu_r_mte,
397
+
398
+ gen_helper_sve_ldff1sds_le_r_mte,
399
+ gen_helper_sve_ldff1hh_le_r_mte,
400
+ gen_helper_sve_ldff1hsu_le_r_mte,
401
+ gen_helper_sve_ldff1hdu_le_r_mte,
402
+
403
+ gen_helper_sve_ldff1hds_le_r_mte,
404
+ gen_helper_sve_ldff1hss_le_r_mte,
405
+ gen_helper_sve_ldff1ss_le_r_mte,
406
+ gen_helper_sve_ldff1sdu_le_r_mte,
407
+
408
+ gen_helper_sve_ldff1bds_r_mte,
409
+ gen_helper_sve_ldff1bss_r_mte,
410
+ gen_helper_sve_ldff1bhs_r_mte,
411
+ gen_helper_sve_ldff1dd_le_r_mte },
412
+
413
+ /* mte active, big-endian */
414
+ { gen_helper_sve_ldff1bb_r_mte,
415
+ gen_helper_sve_ldff1bhu_r_mte,
416
+ gen_helper_sve_ldff1bsu_r_mte,
417
+ gen_helper_sve_ldff1bdu_r_mte,
418
+
419
+ gen_helper_sve_ldff1sds_be_r_mte,
420
+ gen_helper_sve_ldff1hh_be_r_mte,
421
+ gen_helper_sve_ldff1hsu_be_r_mte,
422
+ gen_helper_sve_ldff1hdu_be_r_mte,
423
+
424
+ gen_helper_sve_ldff1hds_be_r_mte,
425
+ gen_helper_sve_ldff1hss_be_r_mte,
426
+ gen_helper_sve_ldff1ss_be_r_mte,
427
+ gen_helper_sve_ldff1sdu_be_r_mte,
428
+
429
+ gen_helper_sve_ldff1bds_r_mte,
430
+ gen_helper_sve_ldff1bss_r_mte,
431
+ gen_helper_sve_ldff1bhs_r_mte,
432
+ gen_helper_sve_ldff1dd_be_r_mte } },
433
};
434
435
if (sve_access_check(s)) {
436
TCGv_i64 addr = new_tmp_a64(s);
437
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
438
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
439
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
440
- fns[s->be_data == MO_BE][a->dtype]);
441
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
442
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
443
}
444
return true;
53
}
445
}
54
446
55
+static void neon_store_element64(int reg, int ele, TCGMemOp size, TCGv_i64 var)
447
static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
56
+{
57
+ long offset = neon_element_offset(reg, ele, size);
58
+
59
+ switch (size) {
60
+ case MO_8:
61
+ tcg_gen_st8_i64(var, cpu_env, offset);
62
+ break;
63
+ case MO_16:
64
+ tcg_gen_st16_i64(var, cpu_env, offset);
65
+ break;
66
+ case MO_32:
67
+ tcg_gen_st32_i64(var, cpu_env, offset);
68
+ break;
69
+ case MO_64:
70
+ tcg_gen_st_i64(var, cpu_env, offset);
71
+ break;
72
+ default:
73
+ g_assert_not_reached();
74
+ }
75
+}
76
+
77
static inline void neon_load_reg64(TCGv_i64 var, int reg)
78
{
448
{
79
tcg_gen_ld_i64(var, cpu_env, vfp_reg_offset(1, reg));
449
- static gen_helper_gvec_mem * const fns[2][16] = {
80
@@ -XXX,XX +XXX,XX @@ static struct {
450
- /* Little-endian */
81
int interleave;
451
- { gen_helper_sve_ldnf1bb_r,
82
int spacing;
452
- gen_helper_sve_ldnf1bhu_r,
83
} const neon_ls_element_type[11] = {
453
- gen_helper_sve_ldnf1bsu_r,
84
- {4, 4, 1},
454
- gen_helper_sve_ldnf1bdu_r,
85
- {4, 4, 2},
455
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
86
+ {1, 4, 1},
456
+ { /* mte inactive, little-endian */
87
+ {1, 4, 2},
457
+ { gen_helper_sve_ldnf1bb_r,
88
{4, 1, 1},
458
+ gen_helper_sve_ldnf1bhu_r,
89
- {4, 2, 1},
459
+ gen_helper_sve_ldnf1bsu_r,
90
- {3, 3, 1},
460
+ gen_helper_sve_ldnf1bdu_r,
91
- {3, 3, 2},
461
92
+ {2, 2, 2},
462
- gen_helper_sve_ldnf1sds_le_r,
93
+ {1, 3, 1},
463
- gen_helper_sve_ldnf1hh_le_r,
94
+ {1, 3, 2},
464
- gen_helper_sve_ldnf1hsu_le_r,
95
{3, 1, 1},
465
- gen_helper_sve_ldnf1hdu_le_r,
96
{1, 1, 1},
466
+ gen_helper_sve_ldnf1sds_le_r,
97
- {2, 2, 1},
467
+ gen_helper_sve_ldnf1hh_le_r,
98
- {2, 2, 2},
468
+ gen_helper_sve_ldnf1hsu_le_r,
99
+ {1, 2, 1},
469
+ gen_helper_sve_ldnf1hdu_le_r,
100
+ {1, 2, 2},
470
101
{2, 1, 1}
471
- gen_helper_sve_ldnf1hds_le_r,
102
};
472
- gen_helper_sve_ldnf1hss_le_r,
103
473
- gen_helper_sve_ldnf1ss_le_r,
104
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
474
- gen_helper_sve_ldnf1sdu_le_r,
105
int shift;
475
+ gen_helper_sve_ldnf1hds_le_r,
106
int n;
476
+ gen_helper_sve_ldnf1hss_le_r,
107
int vec_size;
477
+ gen_helper_sve_ldnf1ss_le_r,
108
+ int mmu_idx;
478
+ gen_helper_sve_ldnf1sdu_le_r,
109
+ TCGMemOp endian;
479
110
TCGv_i32 addr;
480
- gen_helper_sve_ldnf1bds_r,
111
TCGv_i32 tmp;
481
- gen_helper_sve_ldnf1bss_r,
112
TCGv_i32 tmp2;
482
- gen_helper_sve_ldnf1bhs_r,
113
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
483
- gen_helper_sve_ldnf1dd_le_r },
114
rn = (insn >> 16) & 0xf;
484
+ gen_helper_sve_ldnf1bds_r,
115
rm = insn & 0xf;
485
+ gen_helper_sve_ldnf1bss_r,
116
load = (insn & (1 << 21)) != 0;
486
+ gen_helper_sve_ldnf1bhs_r,
117
+ endian = s->be_data;
487
+ gen_helper_sve_ldnf1dd_le_r },
118
+ mmu_idx = get_mem_index(s);
488
119
if ((insn & (1 << 23)) == 0) {
489
- /* Big-endian */
120
/* Load store all elements. */
490
- { gen_helper_sve_ldnf1bb_r,
121
op = (insn >> 8) & 0xf;
491
- gen_helper_sve_ldnf1bhu_r,
122
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
492
- gen_helper_sve_ldnf1bsu_r,
123
nregs = neon_ls_element_type[op].nregs;
493
- gen_helper_sve_ldnf1bdu_r,
124
interleave = neon_ls_element_type[op].interleave;
494
+ /* mte inactive, big-endian */
125
spacing = neon_ls_element_type[op].spacing;
495
+ { gen_helper_sve_ldnf1bb_r,
126
- if (size == 3 && (interleave | spacing) != 1)
496
+ gen_helper_sve_ldnf1bhu_r,
127
+ if (size == 3 && (interleave | spacing) != 1) {
497
+ gen_helper_sve_ldnf1bsu_r,
128
return 1;
498
+ gen_helper_sve_ldnf1bdu_r,
129
+ }
499
130
+ tmp64 = tcg_temp_new_i64();
500
- gen_helper_sve_ldnf1sds_be_r,
131
addr = tcg_temp_new_i32();
501
- gen_helper_sve_ldnf1hh_be_r,
132
+ tmp2 = tcg_const_i32(1 << size);
502
- gen_helper_sve_ldnf1hsu_be_r,
133
load_reg_var(s, addr, rn);
503
- gen_helper_sve_ldnf1hdu_be_r,
134
- stride = (1 << size) * interleave;
504
+ gen_helper_sve_ldnf1sds_be_r,
135
for (reg = 0; reg < nregs; reg++) {
505
+ gen_helper_sve_ldnf1hh_be_r,
136
- if (interleave > 2 || (interleave == 2 && nregs == 2)) {
506
+ gen_helper_sve_ldnf1hsu_be_r,
137
- load_reg_var(s, addr, rn);
507
+ gen_helper_sve_ldnf1hdu_be_r,
138
- tcg_gen_addi_i32(addr, addr, (1 << size) * reg);
508
139
- } else if (interleave == 2 && nregs == 4 && reg == 2) {
509
- gen_helper_sve_ldnf1hds_be_r,
140
- load_reg_var(s, addr, rn);
510
- gen_helper_sve_ldnf1hss_be_r,
141
- tcg_gen_addi_i32(addr, addr, 1 << size);
511
- gen_helper_sve_ldnf1ss_be_r,
142
- }
512
- gen_helper_sve_ldnf1sdu_be_r,
143
- if (size == 3) {
513
+ gen_helper_sve_ldnf1hds_be_r,
144
- tmp64 = tcg_temp_new_i64();
514
+ gen_helper_sve_ldnf1hss_be_r,
145
- if (load) {
515
+ gen_helper_sve_ldnf1ss_be_r,
146
- gen_aa32_ld64(s, tmp64, addr, get_mem_index(s));
516
+ gen_helper_sve_ldnf1sdu_be_r,
147
- neon_store_reg64(tmp64, rd);
517
148
- } else {
518
- gen_helper_sve_ldnf1bds_r,
149
- neon_load_reg64(tmp64, rd);
519
- gen_helper_sve_ldnf1bss_r,
150
- gen_aa32_st64(s, tmp64, addr, get_mem_index(s));
520
- gen_helper_sve_ldnf1bhs_r,
151
- }
521
- gen_helper_sve_ldnf1dd_be_r },
152
- tcg_temp_free_i64(tmp64);
522
+ gen_helper_sve_ldnf1bds_r,
153
- tcg_gen_addi_i32(addr, addr, stride);
523
+ gen_helper_sve_ldnf1bss_r,
154
- } else {
524
+ gen_helper_sve_ldnf1bhs_r,
155
- for (pass = 0; pass < 2; pass++) {
525
+ gen_helper_sve_ldnf1dd_be_r } },
156
- if (size == 2) {
526
+
157
- if (load) {
527
+ { /* mte inactive, little-endian */
158
- tmp = tcg_temp_new_i32();
528
+ { gen_helper_sve_ldnf1bb_r_mte,
159
- gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
529
+ gen_helper_sve_ldnf1bhu_r_mte,
160
- neon_store_reg(rd, pass, tmp);
530
+ gen_helper_sve_ldnf1bsu_r_mte,
161
- } else {
531
+ gen_helper_sve_ldnf1bdu_r_mte,
162
- tmp = neon_load_reg(rd, pass);
532
+
163
- gen_aa32_st32(s, tmp, addr, get_mem_index(s));
533
+ gen_helper_sve_ldnf1sds_le_r_mte,
164
- tcg_temp_free_i32(tmp);
534
+ gen_helper_sve_ldnf1hh_le_r_mte,
165
- }
535
+ gen_helper_sve_ldnf1hsu_le_r_mte,
166
- tcg_gen_addi_i32(addr, addr, stride);
536
+ gen_helper_sve_ldnf1hdu_le_r_mte,
167
- } else if (size == 1) {
537
+
168
- if (load) {
538
+ gen_helper_sve_ldnf1hds_le_r_mte,
169
- tmp = tcg_temp_new_i32();
539
+ gen_helper_sve_ldnf1hss_le_r_mte,
170
- gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
540
+ gen_helper_sve_ldnf1ss_le_r_mte,
171
- tcg_gen_addi_i32(addr, addr, stride);
541
+ gen_helper_sve_ldnf1sdu_le_r_mte,
172
- tmp2 = tcg_temp_new_i32();
542
+
173
- gen_aa32_ld16u(s, tmp2, addr, get_mem_index(s));
543
+ gen_helper_sve_ldnf1bds_r_mte,
174
- tcg_gen_addi_i32(addr, addr, stride);
544
+ gen_helper_sve_ldnf1bss_r_mte,
175
- tcg_gen_shli_i32(tmp2, tmp2, 16);
545
+ gen_helper_sve_ldnf1bhs_r_mte,
176
- tcg_gen_or_i32(tmp, tmp, tmp2);
546
+ gen_helper_sve_ldnf1dd_le_r_mte },
177
- tcg_temp_free_i32(tmp2);
547
+
178
- neon_store_reg(rd, pass, tmp);
548
+ /* mte inactive, big-endian */
179
- } else {
549
+ { gen_helper_sve_ldnf1bb_r_mte,
180
- tmp = neon_load_reg(rd, pass);
550
+ gen_helper_sve_ldnf1bhu_r_mte,
181
- tmp2 = tcg_temp_new_i32();
551
+ gen_helper_sve_ldnf1bsu_r_mte,
182
- tcg_gen_shri_i32(tmp2, tmp, 16);
552
+ gen_helper_sve_ldnf1bdu_r_mte,
183
- gen_aa32_st16(s, tmp, addr, get_mem_index(s));
553
+
184
- tcg_temp_free_i32(tmp);
554
+ gen_helper_sve_ldnf1sds_be_r_mte,
185
- tcg_gen_addi_i32(addr, addr, stride);
555
+ gen_helper_sve_ldnf1hh_be_r_mte,
186
- gen_aa32_st16(s, tmp2, addr, get_mem_index(s));
556
+ gen_helper_sve_ldnf1hsu_be_r_mte,
187
- tcg_temp_free_i32(tmp2);
557
+ gen_helper_sve_ldnf1hdu_be_r_mte,
188
- tcg_gen_addi_i32(addr, addr, stride);
558
+
189
- }
559
+ gen_helper_sve_ldnf1hds_be_r_mte,
190
- } else /* size == 0 */ {
560
+ gen_helper_sve_ldnf1hss_be_r_mte,
191
- if (load) {
561
+ gen_helper_sve_ldnf1ss_be_r_mte,
192
- tmp2 = NULL;
562
+ gen_helper_sve_ldnf1sdu_be_r_mte,
193
- for (n = 0; n < 4; n++) {
563
+
194
- tmp = tcg_temp_new_i32();
564
+ gen_helper_sve_ldnf1bds_r_mte,
195
- gen_aa32_ld8u(s, tmp, addr, get_mem_index(s));
565
+ gen_helper_sve_ldnf1bss_r_mte,
196
- tcg_gen_addi_i32(addr, addr, stride);
566
+ gen_helper_sve_ldnf1bhs_r_mte,
197
- if (n == 0) {
567
+ gen_helper_sve_ldnf1dd_be_r_mte } },
198
- tmp2 = tmp;
568
};
199
- } else {
569
200
- tcg_gen_shli_i32(tmp, tmp, n * 8);
570
if (sve_access_check(s)) {
201
- tcg_gen_or_i32(tmp2, tmp2, tmp);
571
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
202
- tcg_temp_free_i32(tmp);
572
TCGv_i64 addr = new_tmp_a64(s);
203
- }
573
204
- }
574
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
205
- neon_store_reg(rd, pass, tmp2);
575
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
206
- } else {
576
- fns[s->be_data == MO_BE][a->dtype]);
207
- tmp2 = neon_load_reg(rd, pass);
577
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
208
- for (n = 0; n < 4; n++) {
578
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
209
- tmp = tcg_temp_new_i32();
579
}
210
- if (n == 0) {
580
return true;
211
- tcg_gen_mov_i32(tmp, tmp2);
581
}
212
- } else {
213
- tcg_gen_shri_i32(tmp, tmp2, n * 8);
214
- }
215
- gen_aa32_st8(s, tmp, addr, get_mem_index(s));
216
- tcg_temp_free_i32(tmp);
217
- tcg_gen_addi_i32(addr, addr, stride);
218
- }
219
- tcg_temp_free_i32(tmp2);
220
- }
221
+ for (n = 0; n < 8 >> size; n++) {
222
+ int xs;
223
+ for (xs = 0; xs < interleave; xs++) {
224
+ int tt = rd + reg + spacing * xs;
225
+
226
+ if (load) {
227
+ gen_aa32_ld_i64(s, tmp64, addr, mmu_idx, endian | size);
228
+ neon_store_element64(tt, n, size, tmp64);
229
+ } else {
230
+ neon_load_element64(tmp64, tt, n, size);
231
+ gen_aa32_st_i64(s, tmp64, addr, mmu_idx, endian | size);
232
}
233
+ tcg_gen_add_i32(addr, addr, tmp2);
234
}
235
}
236
- rd += spacing;
237
}
238
tcg_temp_free_i32(addr);
239
- stride = nregs * 8;
240
+ tcg_temp_free_i32(tmp2);
241
+ tcg_temp_free_i64(tmp64);
242
+ stride = nregs * interleave * 8;
243
} else {
244
size = (insn >> 10) & 3;
245
if (size == 3) {
246
--
582
--
247
2.19.1
583
2.20.1
248
584
249
585
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We still need to handle tbi for user-only when mte is inactive.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-8-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-37-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
9
---
8
target/arm/translate.c | 67 ++++++++++++++++++++++++------------------
10
target/arm/translate-a64.h | 1 +
9
1 file changed, 39 insertions(+), 28 deletions(-)
11
target/arm/translate-a64.c | 2 +-
12
target/arm/translate-sve.c | 6 ++++--
13
3 files changed, 6 insertions(+), 3 deletions(-)
10
14
11
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
12
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate.c
17
--- a/target/arm/translate-a64.h
14
+++ b/target/arm/translate.c
18
+++ b/target/arm/translate-a64.h
15
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
19
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
16
return 1;
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
17
}
21
unsigned int imms, unsigned int immr);
18
} else { /* (insn & 0x00380080) == 0 */
22
bool sve_access_check(DisasContext *s);
19
- int invert;
23
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
20
+ int invert, reg_ofs, vec_size;
24
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
25
bool tag_checked, int log2_size);
26
TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
27
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-a64.c
30
+++ b/target/arm/translate-a64.c
31
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
32
* of the write-back address.
33
*/
34
35
-static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
36
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
37
{
38
TCGv_i64 clean = new_tmp_a64(s);
39
#ifdef CONFIG_USER_ONLY
40
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-sve.c
43
+++ b/target/arm/translate-sve.c
44
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
45
* For e.g. LD4, there are not enough arguments to pass all 4
46
* registers as pointers, so encode the regno into the data field.
47
* For consistency, do this even for LD1.
48
- * TODO: mte_n check here while callers are updated.
49
*/
50
- if (mte_n && s->mte_active[0]) {
51
+ if (s->mte_active[0]) {
52
int msz = dtype_msz(dtype);
53
54
desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
55
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
56
desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
57
desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
58
desc <<= SVE_MTEDESC_SHIFT;
59
+ } else {
60
+ addr = clean_data_tbi(s, addr);
61
}
21
+
62
+
22
if (q && (rd & 1)) {
63
desc = simd_desc(vsz, vsz, zt | desc);
23
return 1;
64
t_desc = tcg_const_i32(desc);
24
}
65
t_pg = tcg_temp_new_ptr();
25
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
26
break;
27
case 14:
28
imm |= (imm << 8) | (imm << 16) | (imm << 24);
29
- if (invert)
30
+ if (invert) {
31
imm = ~imm;
32
+ }
33
break;
34
case 15:
35
if (invert) {
36
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
37
| ((imm & 0x40) ? (0x1f << 25) : (1 << 30));
38
break;
39
}
40
- if (invert)
41
+ if (invert) {
42
imm = ~imm;
43
+ }
44
45
- for (pass = 0; pass < (q ? 4 : 2); pass++) {
46
- if (op & 1 && op < 12) {
47
- tmp = neon_load_reg(rd, pass);
48
- if (invert) {
49
- /* The immediate value has already been inverted, so
50
- BIC becomes AND. */
51
- tcg_gen_andi_i32(tmp, tmp, imm);
52
- } else {
53
- tcg_gen_ori_i32(tmp, tmp, imm);
54
- }
55
+ reg_ofs = neon_reg_offset(rd, 0);
56
+ vec_size = q ? 16 : 8;
57
+
58
+ if (op & 1 && op < 12) {
59
+ if (invert) {
60
+ /* The immediate value has already been inverted,
61
+ * so BIC becomes AND.
62
+ */
63
+ tcg_gen_gvec_andi(MO_32, reg_ofs, reg_ofs, imm,
64
+ vec_size, vec_size);
65
} else {
66
- /* VMOV, VMVN. */
67
- tmp = tcg_temp_new_i32();
68
- if (op == 14 && invert) {
69
- int n;
70
- uint32_t val;
71
- val = 0;
72
- for (n = 0; n < 4; n++) {
73
- if (imm & (1 << (n + (pass & 1) * 4)))
74
- val |= 0xff << (n * 8);
75
- }
76
- tcg_gen_movi_i32(tmp, val);
77
- } else {
78
- tcg_gen_movi_i32(tmp, imm);
79
- }
80
+ tcg_gen_gvec_ori(MO_32, reg_ofs, reg_ofs, imm,
81
+ vec_size, vec_size);
82
+ }
83
+ } else {
84
+ /* VMOV, VMVN. */
85
+ if (op == 14 && invert) {
86
+ TCGv_i64 t64 = tcg_temp_new_i64();
87
+
88
+ for (pass = 0; pass <= q; ++pass) {
89
+ uint64_t val = 0;
90
+ int n;
91
+
92
+ for (n = 0; n < 8; n++) {
93
+ if (imm & (1 << (n + pass * 8))) {
94
+ val |= 0xffull << (n * 8);
95
+ }
96
+ }
97
+ tcg_gen_movi_i64(t64, val);
98
+ neon_store_reg64(t64, rd + pass);
99
+ }
100
+ tcg_temp_free_i64(t64);
101
+ } else {
102
+ tcg_gen_gvec_dup32i(reg_ofs, vec_size, vec_size, imm);
103
}
104
- neon_store_reg(rd, pass, tmp);
105
}
106
}
107
} else { /* (insn & 0x00800010 == 0x00800000) */
108
--
66
--
109
2.19.1
67
2.20.1
110
68
111
69
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Also introduces neon_element_offset to find the env offset
3
Because the elements are non-sequential, we cannot eliminate many
4
of a specific element within a neon register.
4
tests straight away like we can for sequential operations. But
5
we often have the PTE details handy, so we can test for Tagged.
5
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20181011205206.3552-7-richard.henderson@linaro.org
9
Message-id: 20200626033144.790098-38-richard.henderson@linaro.org
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
---
11
target/arm/translate.c | 63 ++++++++++++++++++++++++------------------
12
target/arm/helper-sve.h | 285 ++++++++++++++++
12
1 file changed, 36 insertions(+), 27 deletions(-)
13
target/arm/sve_helper.c | 185 +++++++++--
14
target/arm/translate-sve.c | 650 +++++++++++++++++++++++++------------
15
3 files changed, 872 insertions(+), 248 deletions(-)
13
16
14
diff --git a/target/arm/translate.c b/target/arm/translate.c
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
15
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate.c
19
--- a/target/arm/helper-sve.h
17
+++ b/target/arm/translate.c
20
+++ b/target/arm/helper-sve.h
18
@@ -XXX,XX +XXX,XX @@ neon_reg_offset (int reg, int n)
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldsds_le_zd, TCG_CALL_NO_WG,
19
return vfp_reg_offset(0, sreg);
22
DEF_HELPER_FLAGS_6(sve_ldsds_be_zd, TCG_CALL_NO_WG,
23
void, env, ptr, ptr, ptr, tl, i32)
24
25
+DEF_HELPER_FLAGS_6(sve_ldbsu_zsu_mte, TCG_CALL_NO_WG,
26
+ void, env, ptr, ptr, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu_mte, TCG_CALL_NO_WG,
28
+ void, env, ptr, ptr, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zsu_mte, TCG_CALL_NO_WG,
30
+ void, env, ptr, ptr, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_6(sve_ldss_le_zsu_mte, TCG_CALL_NO_WG,
32
+ void, env, ptr, ptr, ptr, tl, i32)
33
+DEF_HELPER_FLAGS_6(sve_ldss_be_zsu_mte, TCG_CALL_NO_WG,
34
+ void, env, ptr, ptr, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_6(sve_ldbss_zsu_mte, TCG_CALL_NO_WG,
36
+ void, env, ptr, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zsu_mte, TCG_CALL_NO_WG,
38
+ void, env, ptr, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zsu_mte, TCG_CALL_NO_WG,
40
+ void, env, ptr, ptr, ptr, tl, i32)
41
+
42
+DEF_HELPER_FLAGS_6(sve_ldbsu_zss_mte, TCG_CALL_NO_WG,
43
+ void, env, ptr, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zss_mte, TCG_CALL_NO_WG,
45
+ void, env, ptr, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zss_mte, TCG_CALL_NO_WG,
47
+ void, env, ptr, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_6(sve_ldss_le_zss_mte, TCG_CALL_NO_WG,
49
+ void, env, ptr, ptr, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_6(sve_ldss_be_zss_mte, TCG_CALL_NO_WG,
51
+ void, env, ptr, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_6(sve_ldbss_zss_mte, TCG_CALL_NO_WG,
53
+ void, env, ptr, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zss_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zss_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_6(sve_ldbdu_zsu_mte, TCG_CALL_NO_WG,
60
+ void, env, ptr, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zsu_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zsu_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zsu_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, ptr, ptr, tl, i32)
67
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zsu_mte, TCG_CALL_NO_WG,
68
+ void, env, ptr, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_6(sve_lddd_le_zsu_mte, TCG_CALL_NO_WG,
70
+ void, env, ptr, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_6(sve_lddd_be_zsu_mte, TCG_CALL_NO_WG,
72
+ void, env, ptr, ptr, ptr, tl, i32)
73
+DEF_HELPER_FLAGS_6(sve_ldbds_zsu_mte, TCG_CALL_NO_WG,
74
+ void, env, ptr, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zsu_mte, TCG_CALL_NO_WG,
76
+ void, env, ptr, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zsu_mte, TCG_CALL_NO_WG,
78
+ void, env, ptr, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zsu_mte, TCG_CALL_NO_WG,
80
+ void, env, ptr, ptr, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zsu_mte, TCG_CALL_NO_WG,
82
+ void, env, ptr, ptr, ptr, tl, i32)
83
+
84
+DEF_HELPER_FLAGS_6(sve_ldbdu_zss_mte, TCG_CALL_NO_WG,
85
+ void, env, ptr, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zss_mte, TCG_CALL_NO_WG,
87
+ void, env, ptr, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zss_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, ptr, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zss_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zss_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_6(sve_lddd_le_zss_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_6(sve_lddd_be_zss_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_6(sve_ldbds_zss_mte, TCG_CALL_NO_WG,
99
+ void, env, ptr, ptr, ptr, tl, i32)
100
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zss_mte, TCG_CALL_NO_WG,
101
+ void, env, ptr, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zss_mte, TCG_CALL_NO_WG,
103
+ void, env, ptr, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zss_mte, TCG_CALL_NO_WG,
105
+ void, env, ptr, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zss_mte, TCG_CALL_NO_WG,
107
+ void, env, ptr, ptr, ptr, tl, i32)
108
+
109
+DEF_HELPER_FLAGS_6(sve_ldbdu_zd_mte, TCG_CALL_NO_WG,
110
+ void, env, ptr, ptr, ptr, tl, i32)
111
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zd_mte, TCG_CALL_NO_WG,
112
+ void, env, ptr, ptr, ptr, tl, i32)
113
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zd_mte, TCG_CALL_NO_WG,
114
+ void, env, ptr, ptr, ptr, tl, i32)
115
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zd_mte, TCG_CALL_NO_WG,
116
+ void, env, ptr, ptr, ptr, tl, i32)
117
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zd_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, ptr, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_6(sve_lddd_le_zd_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, ptr, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_6(sve_lddd_be_zd_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, ptr, ptr, tl, i32)
123
+DEF_HELPER_FLAGS_6(sve_ldbds_zd_mte, TCG_CALL_NO_WG,
124
+ void, env, ptr, ptr, ptr, tl, i32)
125
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zd_mte, TCG_CALL_NO_WG,
126
+ void, env, ptr, ptr, ptr, tl, i32)
127
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zd_mte, TCG_CALL_NO_WG,
128
+ void, env, ptr, ptr, ptr, tl, i32)
129
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zd_mte, TCG_CALL_NO_WG,
130
+ void, env, ptr, ptr, ptr, tl, i32)
131
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zd_mte, TCG_CALL_NO_WG,
132
+ void, env, ptr, ptr, ptr, tl, i32)
133
+
134
DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu, TCG_CALL_NO_WG,
135
void, env, ptr, ptr, ptr, tl, i32)
136
DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu, TCG_CALL_NO_WG,
137
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd, TCG_CALL_NO_WG,
138
DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd, TCG_CALL_NO_WG,
139
void, env, ptr, ptr, ptr, tl, i32)
140
141
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu_mte, TCG_CALL_NO_WG,
142
+ void, env, ptr, ptr, ptr, tl, i32)
143
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu_mte, TCG_CALL_NO_WG,
144
+ void, env, ptr, ptr, ptr, tl, i32)
145
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zsu_mte, TCG_CALL_NO_WG,
146
+ void, env, ptr, ptr, ptr, tl, i32)
147
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zsu_mte, TCG_CALL_NO_WG,
148
+ void, env, ptr, ptr, ptr, tl, i32)
149
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zsu_mte, TCG_CALL_NO_WG,
150
+ void, env, ptr, ptr, ptr, tl, i32)
151
+DEF_HELPER_FLAGS_6(sve_ldffbss_zsu_mte, TCG_CALL_NO_WG,
152
+ void, env, ptr, ptr, ptr, tl, i32)
153
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zsu_mte, TCG_CALL_NO_WG,
154
+ void, env, ptr, ptr, ptr, tl, i32)
155
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zsu_mte, TCG_CALL_NO_WG,
156
+ void, env, ptr, ptr, ptr, tl, i32)
157
+
158
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zss_mte, TCG_CALL_NO_WG,
159
+ void, env, ptr, ptr, ptr, tl, i32)
160
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zss_mte, TCG_CALL_NO_WG,
161
+ void, env, ptr, ptr, ptr, tl, i32)
162
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zss_mte, TCG_CALL_NO_WG,
163
+ void, env, ptr, ptr, ptr, tl, i32)
164
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zss_mte, TCG_CALL_NO_WG,
165
+ void, env, ptr, ptr, ptr, tl, i32)
166
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zss_mte, TCG_CALL_NO_WG,
167
+ void, env, ptr, ptr, ptr, tl, i32)
168
+DEF_HELPER_FLAGS_6(sve_ldffbss_zss_mte, TCG_CALL_NO_WG,
169
+ void, env, ptr, ptr, ptr, tl, i32)
170
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zss_mte, TCG_CALL_NO_WG,
171
+ void, env, ptr, ptr, ptr, tl, i32)
172
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zss_mte, TCG_CALL_NO_WG,
173
+ void, env, ptr, ptr, ptr, tl, i32)
174
+
175
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zsu_mte, TCG_CALL_NO_WG,
176
+ void, env, ptr, ptr, ptr, tl, i32)
177
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zsu_mte, TCG_CALL_NO_WG,
178
+ void, env, ptr, ptr, ptr, tl, i32)
179
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zsu_mte, TCG_CALL_NO_WG,
180
+ void, env, ptr, ptr, ptr, tl, i32)
181
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zsu_mte, TCG_CALL_NO_WG,
182
+ void, env, ptr, ptr, ptr, tl, i32)
183
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zsu_mte, TCG_CALL_NO_WG,
184
+ void, env, ptr, ptr, ptr, tl, i32)
185
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zsu_mte, TCG_CALL_NO_WG,
186
+ void, env, ptr, ptr, ptr, tl, i32)
187
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zsu_mte, TCG_CALL_NO_WG,
188
+ void, env, ptr, ptr, ptr, tl, i32)
189
+DEF_HELPER_FLAGS_6(sve_ldffbds_zsu_mte, TCG_CALL_NO_WG,
190
+ void, env, ptr, ptr, ptr, tl, i32)
191
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zsu_mte, TCG_CALL_NO_WG,
192
+ void, env, ptr, ptr, ptr, tl, i32)
193
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zsu_mte, TCG_CALL_NO_WG,
194
+ void, env, ptr, ptr, ptr, tl, i32)
195
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zsu_mte, TCG_CALL_NO_WG,
196
+ void, env, ptr, ptr, ptr, tl, i32)
197
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zsu_mte, TCG_CALL_NO_WG,
198
+ void, env, ptr, ptr, ptr, tl, i32)
199
+
200
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zss_mte, TCG_CALL_NO_WG,
201
+ void, env, ptr, ptr, ptr, tl, i32)
202
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zss_mte, TCG_CALL_NO_WG,
203
+ void, env, ptr, ptr, ptr, tl, i32)
204
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zss_mte, TCG_CALL_NO_WG,
205
+ void, env, ptr, ptr, ptr, tl, i32)
206
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zss_mte, TCG_CALL_NO_WG,
207
+ void, env, ptr, ptr, ptr, tl, i32)
208
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zss_mte, TCG_CALL_NO_WG,
209
+ void, env, ptr, ptr, ptr, tl, i32)
210
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zss_mte, TCG_CALL_NO_WG,
211
+ void, env, ptr, ptr, ptr, tl, i32)
212
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zss_mte, TCG_CALL_NO_WG,
213
+ void, env, ptr, ptr, ptr, tl, i32)
214
+DEF_HELPER_FLAGS_6(sve_ldffbds_zss_mte, TCG_CALL_NO_WG,
215
+ void, env, ptr, ptr, ptr, tl, i32)
216
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zss_mte, TCG_CALL_NO_WG,
217
+ void, env, ptr, ptr, ptr, tl, i32)
218
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zss_mte, TCG_CALL_NO_WG,
219
+ void, env, ptr, ptr, ptr, tl, i32)
220
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zss_mte, TCG_CALL_NO_WG,
221
+ void, env, ptr, ptr, ptr, tl, i32)
222
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zss_mte, TCG_CALL_NO_WG,
223
+ void, env, ptr, ptr, ptr, tl, i32)
224
+
225
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zd_mte, TCG_CALL_NO_WG,
226
+ void, env, ptr, ptr, ptr, tl, i32)
227
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zd_mte, TCG_CALL_NO_WG,
228
+ void, env, ptr, ptr, ptr, tl, i32)
229
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zd_mte, TCG_CALL_NO_WG,
230
+ void, env, ptr, ptr, ptr, tl, i32)
231
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zd_mte, TCG_CALL_NO_WG,
232
+ void, env, ptr, ptr, ptr, tl, i32)
233
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zd_mte, TCG_CALL_NO_WG,
234
+ void, env, ptr, ptr, ptr, tl, i32)
235
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zd_mte, TCG_CALL_NO_WG,
236
+ void, env, ptr, ptr, ptr, tl, i32)
237
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zd_mte, TCG_CALL_NO_WG,
238
+ void, env, ptr, ptr, ptr, tl, i32)
239
+DEF_HELPER_FLAGS_6(sve_ldffbds_zd_mte, TCG_CALL_NO_WG,
240
+ void, env, ptr, ptr, ptr, tl, i32)
241
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zd_mte, TCG_CALL_NO_WG,
242
+ void, env, ptr, ptr, ptr, tl, i32)
243
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zd_mte, TCG_CALL_NO_WG,
244
+ void, env, ptr, ptr, ptr, tl, i32)
245
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd_mte, TCG_CALL_NO_WG,
246
+ void, env, ptr, ptr, ptr, tl, i32)
247
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd_mte, TCG_CALL_NO_WG,
248
+ void, env, ptr, ptr, ptr, tl, i32)
249
+
250
DEF_HELPER_FLAGS_6(sve_stbs_zsu, TCG_CALL_NO_WG,
251
void, env, ptr, ptr, ptr, tl, i32)
252
DEF_HELPER_FLAGS_6(sve_sths_le_zsu, TCG_CALL_NO_WG,
253
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd, TCG_CALL_NO_WG,
254
DEF_HELPER_FLAGS_6(sve_stdd_be_zd, TCG_CALL_NO_WG,
255
void, env, ptr, ptr, ptr, tl, i32)
256
257
+DEF_HELPER_FLAGS_6(sve_stbs_zsu_mte, TCG_CALL_NO_WG,
258
+ void, env, ptr, ptr, ptr, tl, i32)
259
+DEF_HELPER_FLAGS_6(sve_sths_le_zsu_mte, TCG_CALL_NO_WG,
260
+ void, env, ptr, ptr, ptr, tl, i32)
261
+DEF_HELPER_FLAGS_6(sve_sths_be_zsu_mte, TCG_CALL_NO_WG,
262
+ void, env, ptr, ptr, ptr, tl, i32)
263
+DEF_HELPER_FLAGS_6(sve_stss_le_zsu_mte, TCG_CALL_NO_WG,
264
+ void, env, ptr, ptr, ptr, tl, i32)
265
+DEF_HELPER_FLAGS_6(sve_stss_be_zsu_mte, TCG_CALL_NO_WG,
266
+ void, env, ptr, ptr, ptr, tl, i32)
267
+
268
+DEF_HELPER_FLAGS_6(sve_stbs_zss_mte, TCG_CALL_NO_WG,
269
+ void, env, ptr, ptr, ptr, tl, i32)
270
+DEF_HELPER_FLAGS_6(sve_sths_le_zss_mte, TCG_CALL_NO_WG,
271
+ void, env, ptr, ptr, ptr, tl, i32)
272
+DEF_HELPER_FLAGS_6(sve_sths_be_zss_mte, TCG_CALL_NO_WG,
273
+ void, env, ptr, ptr, ptr, tl, i32)
274
+DEF_HELPER_FLAGS_6(sve_stss_le_zss_mte, TCG_CALL_NO_WG,
275
+ void, env, ptr, ptr, ptr, tl, i32)
276
+DEF_HELPER_FLAGS_6(sve_stss_be_zss_mte, TCG_CALL_NO_WG,
277
+ void, env, ptr, ptr, ptr, tl, i32)
278
+
279
+DEF_HELPER_FLAGS_6(sve_stbd_zsu_mte, TCG_CALL_NO_WG,
280
+ void, env, ptr, ptr, ptr, tl, i32)
281
+DEF_HELPER_FLAGS_6(sve_sthd_le_zsu_mte, TCG_CALL_NO_WG,
282
+ void, env, ptr, ptr, ptr, tl, i32)
283
+DEF_HELPER_FLAGS_6(sve_sthd_be_zsu_mte, TCG_CALL_NO_WG,
284
+ void, env, ptr, ptr, ptr, tl, i32)
285
+DEF_HELPER_FLAGS_6(sve_stsd_le_zsu_mte, TCG_CALL_NO_WG,
286
+ void, env, ptr, ptr, ptr, tl, i32)
287
+DEF_HELPER_FLAGS_6(sve_stsd_be_zsu_mte, TCG_CALL_NO_WG,
288
+ void, env, ptr, ptr, ptr, tl, i32)
289
+DEF_HELPER_FLAGS_6(sve_stdd_le_zsu_mte, TCG_CALL_NO_WG,
290
+ void, env, ptr, ptr, ptr, tl, i32)
291
+DEF_HELPER_FLAGS_6(sve_stdd_be_zsu_mte, TCG_CALL_NO_WG,
292
+ void, env, ptr, ptr, ptr, tl, i32)
293
+
294
+DEF_HELPER_FLAGS_6(sve_stbd_zss_mte, TCG_CALL_NO_WG,
295
+ void, env, ptr, ptr, ptr, tl, i32)
296
+DEF_HELPER_FLAGS_6(sve_sthd_le_zss_mte, TCG_CALL_NO_WG,
297
+ void, env, ptr, ptr, ptr, tl, i32)
298
+DEF_HELPER_FLAGS_6(sve_sthd_be_zss_mte, TCG_CALL_NO_WG,
299
+ void, env, ptr, ptr, ptr, tl, i32)
300
+DEF_HELPER_FLAGS_6(sve_stsd_le_zss_mte, TCG_CALL_NO_WG,
301
+ void, env, ptr, ptr, ptr, tl, i32)
302
+DEF_HELPER_FLAGS_6(sve_stsd_be_zss_mte, TCG_CALL_NO_WG,
303
+ void, env, ptr, ptr, ptr, tl, i32)
304
+DEF_HELPER_FLAGS_6(sve_stdd_le_zss_mte, TCG_CALL_NO_WG,
305
+ void, env, ptr, ptr, ptr, tl, i32)
306
+DEF_HELPER_FLAGS_6(sve_stdd_be_zss_mte, TCG_CALL_NO_WG,
307
+ void, env, ptr, ptr, ptr, tl, i32)
308
+
309
+DEF_HELPER_FLAGS_6(sve_stbd_zd_mte, TCG_CALL_NO_WG,
310
+ void, env, ptr, ptr, ptr, tl, i32)
311
+DEF_HELPER_FLAGS_6(sve_sthd_le_zd_mte, TCG_CALL_NO_WG,
312
+ void, env, ptr, ptr, ptr, tl, i32)
313
+DEF_HELPER_FLAGS_6(sve_sthd_be_zd_mte, TCG_CALL_NO_WG,
314
+ void, env, ptr, ptr, ptr, tl, i32)
315
+DEF_HELPER_FLAGS_6(sve_stsd_le_zd_mte, TCG_CALL_NO_WG,
316
+ void, env, ptr, ptr, ptr, tl, i32)
317
+DEF_HELPER_FLAGS_6(sve_stsd_be_zd_mte, TCG_CALL_NO_WG,
318
+ void, env, ptr, ptr, ptr, tl, i32)
319
+DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG,
320
+ void, env, ptr, ptr, ptr, tl, i32)
321
+DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG,
322
+ void, env, ptr, ptr, ptr, tl, i32)
323
+
324
DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
325
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
326
index XXXXXXX..XXXXXXX 100644
327
--- a/target/arm/sve_helper.c
328
+++ b/target/arm/sve_helper.c
329
@@ -XXX,XX +XXX,XX @@ static target_ulong off_zd_d(void *reg, intptr_t reg_ofs)
330
static inline QEMU_ALWAYS_INLINE
331
void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
332
target_ulong base, uint32_t desc, uintptr_t retaddr,
333
- int esize, int msize, zreg_off_fn *off_fn,
334
+ uint32_t mtedesc, int esize, int msize,
335
+ zreg_off_fn *off_fn,
336
sve_ldst1_host_fn *host_fn,
337
sve_ldst1_tlb_fn *tlb_fn)
338
{
339
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
340
cpu_check_watchpoint(env_cpu(env), addr, msize,
341
info.attrs, BP_MEM_READ, retaddr);
342
}
343
- /* TODO: MTE check */
344
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
345
+ mte_check1(env, mtedesc, addr, retaddr);
346
+ }
347
host_fn(&scratch, reg_off, info.host);
348
} else {
349
/* Element crosses the page boundary. */
350
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
351
msize, info.attrs,
352
BP_MEM_READ, retaddr);
353
}
354
- /* TODO: MTE check */
355
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
356
+ mte_check1(env, mtedesc, addr, retaddr);
357
+ }
358
tlb_fn(env, &scratch, reg_off, addr, retaddr);
359
}
360
}
361
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
362
memcpy(vd, &scratch, reg_max);
20
}
363
}
21
364
22
+/* Return the offset of a 2**SIZE piece of a NEON register, at index ELE,
365
+static inline QEMU_ALWAYS_INLINE
23
+ * where 0 is the least significant end of the register.
366
+void sve_ld1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
24
+ */
367
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
25
+static inline long
368
+ int esize, int msize, zreg_off_fn *off_fn,
26
+neon_element_offset(int reg, int element, TCGMemOp size)
369
+ sve_ldst1_host_fn *host_fn,
370
+ sve_ldst1_tlb_fn *tlb_fn)
27
+{
371
+{
28
+ int element_size = 1 << size;
372
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
29
+ int ofs = element * element_size;
373
+ /* Remove mtedesc from the normal sve descriptor. */
30
+#ifdef HOST_WORDS_BIGENDIAN
374
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
31
+ /* Calculate the offset assuming fully little-endian,
375
+
32
+ * then XOR to account for the order of the 8-byte units.
376
+ /*
377
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
378
+ * offset base entirely over the address space hole to change the
379
+ * pointer tag, or change the bit55 selector. So we could here
380
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
33
+ */
381
+ */
34
+ if (element_size < 8) {
382
+ sve_ld1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
35
+ ofs ^= 8 - element_size;
383
+ esize, msize, off_fn, host_fn, tlb_fn);
384
+}
385
+
386
#define DO_LD1_ZPZ_S(MEM, OFS, MSZ) \
387
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
388
void *vm, target_ulong base, uint32_t desc) \
389
{ \
390
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
391
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
392
off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
393
+} \
394
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
395
+ void *vm, target_ulong base, uint32_t desc) \
396
+{ \
397
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
398
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
399
}
400
401
#define DO_LD1_ZPZ_D(MEM, OFS, MSZ) \
402
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
403
void *vm, target_ulong base, uint32_t desc) \
404
{ \
405
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
406
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
407
off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
408
+} \
409
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
410
+ void *vm, target_ulong base, uint32_t desc) \
411
+{ \
412
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
413
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
414
}
415
416
DO_LD1_ZPZ_S(bsu, zsu, MO_8)
417
@@ -XXX,XX +XXX,XX @@ DO_LD1_ZPZ_D(dd_be, zd, MO_64)
418
static inline QEMU_ALWAYS_INLINE
419
void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
420
target_ulong base, uint32_t desc, uintptr_t retaddr,
421
- const int esz, const int msz, zreg_off_fn *off_fn,
422
+ uint32_t mtedesc, const int esz, const int msz,
423
+ zreg_off_fn *off_fn,
424
sve_ldst1_host_fn *host_fn,
425
sve_ldst1_tlb_fn *tlb_fn)
426
{
427
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
428
* Probe the first element, allowing faults.
429
*/
430
addr = base + (off_fn(vm, reg_off) << scale);
431
+ if (mtedesc) {
432
+ mte_check1(env, mtedesc, addr, retaddr);
36
+ }
433
+ }
37
+#endif
434
tlb_fn(env, vd, reg_off, addr, retaddr);
38
+ return neon_reg_offset(reg, 0) + ofs;
435
436
/* After any fault, zero the other elements. */
437
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
438
(env_cpu(env), addr, msize) & BP_MEM_READ)) {
439
goto fault;
440
}
441
- /* TODO: MTE check. */
442
+ if (mtedesc &&
443
+ arm_tlb_mte_tagged(&info.attrs) &&
444
+ !mte_probe1(env, mtedesc, addr)) {
445
+ goto fault;
446
+ }
447
448
host_fn(vd, reg_off, info.host);
449
}
450
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
451
record_fault(env, reg_off, reg_max);
452
}
453
454
-#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
455
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
456
- void *vm, target_ulong base, uint32_t desc) \
457
-{ \
458
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
459
- off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
460
+static inline QEMU_ALWAYS_INLINE
461
+void sve_ldff1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
462
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
463
+ const int esz, const int msz,
464
+ zreg_off_fn *off_fn,
465
+ sve_ldst1_host_fn *host_fn,
466
+ sve_ldst1_tlb_fn *tlb_fn)
467
+{
468
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
469
+ /* Remove mtedesc from the normal sve descriptor. */
470
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
471
+
472
+ /*
473
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
474
+ * offset base entirely over the address space hole to change the
475
+ * pointer tag, or change the bit55 selector. So we could here
476
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
477
+ */
478
+ sve_ldff1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
479
+ esz, msz, off_fn, host_fn, tlb_fn);
480
}
481
482
-#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
483
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
484
- void *vm, target_ulong base, uint32_t desc) \
485
-{ \
486
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
487
- off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
488
+#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
489
+void HELPER(sve_ldff##MEM##_##OFS) \
490
+ (CPUARMState *env, void *vd, void *vg, \
491
+ void *vm, target_ulong base, uint32_t desc) \
492
+{ \
493
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_32, MSZ, \
494
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
495
+} \
496
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
497
+ (CPUARMState *env, void *vd, void *vg, \
498
+ void *vm, target_ulong base, uint32_t desc) \
499
+{ \
500
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
501
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
39
+}
502
+}
40
+
503
+
41
static TCGv_i32 neon_load_reg(int reg, int pass)
504
+#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
505
+void HELPER(sve_ldff##MEM##_##OFS) \
506
+ (CPUARMState *env, void *vd, void *vg, \
507
+ void *vm, target_ulong base, uint32_t desc) \
508
+{ \
509
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_64, MSZ, \
510
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
511
+} \
512
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
513
+ (CPUARMState *env, void *vd, void *vg, \
514
+ void *vm, target_ulong base, uint32_t desc) \
515
+{ \
516
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
517
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
518
}
519
520
DO_LDFF1_ZPZ_S(bsu, zsu, MO_8)
521
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_ZPZ_D(dd_be, zd, MO_64)
522
static inline QEMU_ALWAYS_INLINE
523
void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
524
target_ulong base, uint32_t desc, uintptr_t retaddr,
525
- int esize, int msize, zreg_off_fn *off_fn,
526
+ uint32_t mtedesc, int esize, int msize,
527
+ zreg_off_fn *off_fn,
528
sve_ldst1_host_fn *host_fn,
529
sve_ldst1_tlb_fn *tlb_fn)
42
{
530
{
43
TCGv_i32 tmp = tcg_temp_new_i32();
531
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
44
@@ -XXX,XX +XXX,XX @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn)
532
cpu_check_watchpoint(env_cpu(env), addr, msize,
45
tmp = load_reg(s, rd);
533
info.attrs, BP_MEM_WRITE, retaddr);
46
if (insn & (1 << 23)) {
47
/* VDUP */
48
- if (size == 0) {
49
- gen_neon_dup_u8(tmp, 0);
50
- } else if (size == 1) {
51
- gen_neon_dup_low16(tmp);
52
- }
53
- for (n = 0; n <= pass * 2; n++) {
54
- tmp2 = tcg_temp_new_i32();
55
- tcg_gen_mov_i32(tmp2, tmp);
56
- neon_store_reg(rn, n, tmp2);
57
- }
58
- neon_store_reg(rn, n, tmp);
59
+ int vec_size = pass ? 16 : 8;
60
+ tcg_gen_gvec_dup_i32(size, neon_reg_offset(rn, 0),
61
+ vec_size, vec_size, tmp);
62
+ tcg_temp_free_i32(tmp);
63
} else {
64
/* VMOV */
65
switch (size) {
66
@@ -XXX,XX +XXX,XX @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
67
tcg_temp_free_i32(tmp);
68
} else if ((insn & 0x380) == 0) {
69
/* VDUP */
70
+ int element;
71
+ TCGMemOp size;
72
+
73
if ((insn & (7 << 16)) == 0 || (q && (rd & 1))) {
74
return 1;
75
}
534
}
76
- if (insn & (1 << 19)) {
535
- /* TODO: MTE check. */
77
- tmp = neon_load_reg(rm, 1);
536
+
78
- } else {
537
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
79
- tmp = neon_load_reg(rm, 0);
538
+ mte_check1(env, mtedesc, addr, retaddr);
80
- }
539
+ }
81
if (insn & (1 << 16)) {
82
- gen_neon_dup_u8(tmp, ((insn >> 17) & 3) * 8);
83
+ size = MO_8;
84
+ element = (insn >> 17) & 7;
85
} else if (insn & (1 << 17)) {
86
- if ((insn >> 18) & 1)
87
- gen_neon_dup_high16(tmp);
88
- else
89
- gen_neon_dup_low16(tmp);
90
+ size = MO_16;
91
+ element = (insn >> 18) & 3;
92
+ } else {
93
+ size = MO_32;
94
+ element = (insn >> 19) & 1;
95
}
96
- for (pass = 0; pass < (q ? 4 : 2); pass++) {
97
- tmp2 = tcg_temp_new_i32();
98
- tcg_gen_mov_i32(tmp2, tmp);
99
- neon_store_reg(rd, pass, tmp2);
100
- }
101
- tcg_temp_free_i32(tmp);
102
+ tcg_gen_gvec_dup_mem(size, neon_reg_offset(rd, 0),
103
+ neon_element_offset(rm, element, size),
104
+ q ? 16 : 8, q ? 16 : 8);
105
} else {
106
return 1;
107
}
540
}
541
i += 1;
542
reg_off += esize;
543
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
544
} while (reg_off < reg_max);
545
}
546
547
-#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
548
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
549
- void *vm, target_ulong base, uint32_t desc) \
550
-{ \
551
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
552
- off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
553
+static inline QEMU_ALWAYS_INLINE
554
+void sve_st1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
555
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
556
+ int esize, int msize, zreg_off_fn *off_fn,
557
+ sve_ldst1_host_fn *host_fn,
558
+ sve_ldst1_tlb_fn *tlb_fn)
559
+{
560
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
561
+ /* Remove mtedesc from the normal sve descriptor. */
562
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
563
+
564
+ /*
565
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
566
+ * offset base entirely over the address space hole to change the
567
+ * pointer tag, or change the bit55 selector. So we could here
568
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
569
+ */
570
+ sve_st1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
571
+ esize, msize, off_fn, host_fn, tlb_fn);
572
}
573
574
-#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
575
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
576
+#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
577
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
578
void *vm, target_ulong base, uint32_t desc) \
579
-{ \
580
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
581
- off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
582
+{ \
583
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
584
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
585
+} \
586
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
587
+ void *vm, target_ulong base, uint32_t desc) \
588
+{ \
589
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
590
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
591
+}
592
+
593
+#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
594
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
595
+ void *vm, target_ulong base, uint32_t desc) \
596
+{ \
597
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
598
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
599
+} \
600
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
601
+ void *vm, target_ulong base, uint32_t desc) \
602
+{ \
603
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
604
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
605
}
606
607
DO_ST1_ZPZ_S(bs, zsu, MO_8)
608
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
609
index XXXXXXX..XXXXXXX 100644
610
--- a/target/arm/translate-sve.c
611
+++ b/target/arm/translate-sve.c
612
@@ -XXX,XX +XXX,XX @@ static bool trans_ST_zpri(DisasContext *s, arg_rpri_store *a)
613
*/
614
615
static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
616
- int scale, TCGv_i64 scalar, int msz,
617
+ int scale, TCGv_i64 scalar, int msz, bool is_write,
618
gen_helper_gvec_mem_scatter *fn)
619
{
620
unsigned vsz = vec_full_reg_size(s);
621
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
622
TCGv_ptr t_pg = tcg_temp_new_ptr();
623
TCGv_ptr t_zt = tcg_temp_new_ptr();
624
TCGv_i32 t_desc;
625
- int desc;
626
+ int desc = 0;
627
628
+ if (s->mte_active[0]) {
629
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
630
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
631
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
632
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
633
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
634
+ desc <<= SVE_MTEDESC_SHIFT;
635
+ }
636
desc = simd_desc(vsz, vsz, scale);
637
t_desc = tcg_const_i32(desc);
638
639
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
640
tcg_temp_free_i32(t_desc);
641
}
642
643
-/* Indexed by [be][ff][xs][u][msz]. */
644
-static gen_helper_gvec_mem_scatter * const gather_load_fn32[2][2][2][2][3] = {
645
- /* Little-endian */
646
- { { { { gen_helper_sve_ldbss_zsu,
647
- gen_helper_sve_ldhss_le_zsu,
648
- NULL, },
649
- { gen_helper_sve_ldbsu_zsu,
650
- gen_helper_sve_ldhsu_le_zsu,
651
- gen_helper_sve_ldss_le_zsu, } },
652
- { { gen_helper_sve_ldbss_zss,
653
- gen_helper_sve_ldhss_le_zss,
654
- NULL, },
655
- { gen_helper_sve_ldbsu_zss,
656
- gen_helper_sve_ldhsu_le_zss,
657
- gen_helper_sve_ldss_le_zss, } } },
658
+/* Indexed by [mte][be][ff][xs][u][msz]. */
659
+static gen_helper_gvec_mem_scatter * const
660
+gather_load_fn32[2][2][2][2][2][3] = {
661
+ { /* MTE Inactive */
662
+ { /* Little-endian */
663
+ { { { gen_helper_sve_ldbss_zsu,
664
+ gen_helper_sve_ldhss_le_zsu,
665
+ NULL, },
666
+ { gen_helper_sve_ldbsu_zsu,
667
+ gen_helper_sve_ldhsu_le_zsu,
668
+ gen_helper_sve_ldss_le_zsu, } },
669
+ { { gen_helper_sve_ldbss_zss,
670
+ gen_helper_sve_ldhss_le_zss,
671
+ NULL, },
672
+ { gen_helper_sve_ldbsu_zss,
673
+ gen_helper_sve_ldhsu_le_zss,
674
+ gen_helper_sve_ldss_le_zss, } } },
675
676
- /* First-fault */
677
- { { { gen_helper_sve_ldffbss_zsu,
678
- gen_helper_sve_ldffhss_le_zsu,
679
- NULL, },
680
- { gen_helper_sve_ldffbsu_zsu,
681
- gen_helper_sve_ldffhsu_le_zsu,
682
- gen_helper_sve_ldffss_le_zsu, } },
683
- { { gen_helper_sve_ldffbss_zss,
684
- gen_helper_sve_ldffhss_le_zss,
685
- NULL, },
686
- { gen_helper_sve_ldffbsu_zss,
687
- gen_helper_sve_ldffhsu_le_zss,
688
- gen_helper_sve_ldffss_le_zss, } } } },
689
+ /* First-fault */
690
+ { { { gen_helper_sve_ldffbss_zsu,
691
+ gen_helper_sve_ldffhss_le_zsu,
692
+ NULL, },
693
+ { gen_helper_sve_ldffbsu_zsu,
694
+ gen_helper_sve_ldffhsu_le_zsu,
695
+ gen_helper_sve_ldffss_le_zsu, } },
696
+ { { gen_helper_sve_ldffbss_zss,
697
+ gen_helper_sve_ldffhss_le_zss,
698
+ NULL, },
699
+ { gen_helper_sve_ldffbsu_zss,
700
+ gen_helper_sve_ldffhsu_le_zss,
701
+ gen_helper_sve_ldffss_le_zss, } } } },
702
703
- /* Big-endian */
704
- { { { { gen_helper_sve_ldbss_zsu,
705
- gen_helper_sve_ldhss_be_zsu,
706
- NULL, },
707
- { gen_helper_sve_ldbsu_zsu,
708
- gen_helper_sve_ldhsu_be_zsu,
709
- gen_helper_sve_ldss_be_zsu, } },
710
- { { gen_helper_sve_ldbss_zss,
711
- gen_helper_sve_ldhss_be_zss,
712
- NULL, },
713
- { gen_helper_sve_ldbsu_zss,
714
- gen_helper_sve_ldhsu_be_zss,
715
- gen_helper_sve_ldss_be_zss, } } },
716
+ { /* Big-endian */
717
+ { { { gen_helper_sve_ldbss_zsu,
718
+ gen_helper_sve_ldhss_be_zsu,
719
+ NULL, },
720
+ { gen_helper_sve_ldbsu_zsu,
721
+ gen_helper_sve_ldhsu_be_zsu,
722
+ gen_helper_sve_ldss_be_zsu, } },
723
+ { { gen_helper_sve_ldbss_zss,
724
+ gen_helper_sve_ldhss_be_zss,
725
+ NULL, },
726
+ { gen_helper_sve_ldbsu_zss,
727
+ gen_helper_sve_ldhsu_be_zss,
728
+ gen_helper_sve_ldss_be_zss, } } },
729
730
- /* First-fault */
731
- { { { gen_helper_sve_ldffbss_zsu,
732
- gen_helper_sve_ldffhss_be_zsu,
733
- NULL, },
734
- { gen_helper_sve_ldffbsu_zsu,
735
- gen_helper_sve_ldffhsu_be_zsu,
736
- gen_helper_sve_ldffss_be_zsu, } },
737
- { { gen_helper_sve_ldffbss_zss,
738
- gen_helper_sve_ldffhss_be_zss,
739
- NULL, },
740
- { gen_helper_sve_ldffbsu_zss,
741
- gen_helper_sve_ldffhsu_be_zss,
742
- gen_helper_sve_ldffss_be_zss, } } } },
743
+ /* First-fault */
744
+ { { { gen_helper_sve_ldffbss_zsu,
745
+ gen_helper_sve_ldffhss_be_zsu,
746
+ NULL, },
747
+ { gen_helper_sve_ldffbsu_zsu,
748
+ gen_helper_sve_ldffhsu_be_zsu,
749
+ gen_helper_sve_ldffss_be_zsu, } },
750
+ { { gen_helper_sve_ldffbss_zss,
751
+ gen_helper_sve_ldffhss_be_zss,
752
+ NULL, },
753
+ { gen_helper_sve_ldffbsu_zss,
754
+ gen_helper_sve_ldffhsu_be_zss,
755
+ gen_helper_sve_ldffss_be_zss, } } } } },
756
+ { /* MTE Active */
757
+ { /* Little-endian */
758
+ { { { gen_helper_sve_ldbss_zsu_mte,
759
+ gen_helper_sve_ldhss_le_zsu_mte,
760
+ NULL, },
761
+ { gen_helper_sve_ldbsu_zsu_mte,
762
+ gen_helper_sve_ldhsu_le_zsu_mte,
763
+ gen_helper_sve_ldss_le_zsu_mte, } },
764
+ { { gen_helper_sve_ldbss_zss_mte,
765
+ gen_helper_sve_ldhss_le_zss_mte,
766
+ NULL, },
767
+ { gen_helper_sve_ldbsu_zss_mte,
768
+ gen_helper_sve_ldhsu_le_zss_mte,
769
+ gen_helper_sve_ldss_le_zss_mte, } } },
770
+
771
+ /* First-fault */
772
+ { { { gen_helper_sve_ldffbss_zsu_mte,
773
+ gen_helper_sve_ldffhss_le_zsu_mte,
774
+ NULL, },
775
+ { gen_helper_sve_ldffbsu_zsu_mte,
776
+ gen_helper_sve_ldffhsu_le_zsu_mte,
777
+ gen_helper_sve_ldffss_le_zsu_mte, } },
778
+ { { gen_helper_sve_ldffbss_zss_mte,
779
+ gen_helper_sve_ldffhss_le_zss_mte,
780
+ NULL, },
781
+ { gen_helper_sve_ldffbsu_zss_mte,
782
+ gen_helper_sve_ldffhsu_le_zss_mte,
783
+ gen_helper_sve_ldffss_le_zss_mte, } } } },
784
+
785
+ { /* Big-endian */
786
+ { { { gen_helper_sve_ldbss_zsu_mte,
787
+ gen_helper_sve_ldhss_be_zsu_mte,
788
+ NULL, },
789
+ { gen_helper_sve_ldbsu_zsu_mte,
790
+ gen_helper_sve_ldhsu_be_zsu_mte,
791
+ gen_helper_sve_ldss_be_zsu_mte, } },
792
+ { { gen_helper_sve_ldbss_zss_mte,
793
+ gen_helper_sve_ldhss_be_zss_mte,
794
+ NULL, },
795
+ { gen_helper_sve_ldbsu_zss_mte,
796
+ gen_helper_sve_ldhsu_be_zss_mte,
797
+ gen_helper_sve_ldss_be_zss_mte, } } },
798
+
799
+ /* First-fault */
800
+ { { { gen_helper_sve_ldffbss_zsu_mte,
801
+ gen_helper_sve_ldffhss_be_zsu_mte,
802
+ NULL, },
803
+ { gen_helper_sve_ldffbsu_zsu_mte,
804
+ gen_helper_sve_ldffhsu_be_zsu_mte,
805
+ gen_helper_sve_ldffss_be_zsu_mte, } },
806
+ { { gen_helper_sve_ldffbss_zss_mte,
807
+ gen_helper_sve_ldffhss_be_zss_mte,
808
+ NULL, },
809
+ { gen_helper_sve_ldffbsu_zss_mte,
810
+ gen_helper_sve_ldffhsu_be_zss_mte,
811
+ gen_helper_sve_ldffss_be_zss_mte, } } } } },
812
};
813
814
/* Note that we overload xs=2 to indicate 64-bit offset. */
815
-static gen_helper_gvec_mem_scatter * const gather_load_fn64[2][2][3][2][4] = {
816
- /* Little-endian */
817
- { { { { gen_helper_sve_ldbds_zsu,
818
- gen_helper_sve_ldhds_le_zsu,
819
- gen_helper_sve_ldsds_le_zsu,
820
- NULL, },
821
- { gen_helper_sve_ldbdu_zsu,
822
- gen_helper_sve_ldhdu_le_zsu,
823
- gen_helper_sve_ldsdu_le_zsu,
824
- gen_helper_sve_lddd_le_zsu, } },
825
- { { gen_helper_sve_ldbds_zss,
826
- gen_helper_sve_ldhds_le_zss,
827
- gen_helper_sve_ldsds_le_zss,
828
- NULL, },
829
- { gen_helper_sve_ldbdu_zss,
830
- gen_helper_sve_ldhdu_le_zss,
831
- gen_helper_sve_ldsdu_le_zss,
832
- gen_helper_sve_lddd_le_zss, } },
833
- { { gen_helper_sve_ldbds_zd,
834
- gen_helper_sve_ldhds_le_zd,
835
- gen_helper_sve_ldsds_le_zd,
836
- NULL, },
837
- { gen_helper_sve_ldbdu_zd,
838
- gen_helper_sve_ldhdu_le_zd,
839
- gen_helper_sve_ldsdu_le_zd,
840
- gen_helper_sve_lddd_le_zd, } } },
841
+static gen_helper_gvec_mem_scatter * const
842
+gather_load_fn64[2][2][2][3][2][4] = {
843
+ { /* MTE Inactive */
844
+ { /* Little-endian */
845
+ { { { gen_helper_sve_ldbds_zsu,
846
+ gen_helper_sve_ldhds_le_zsu,
847
+ gen_helper_sve_ldsds_le_zsu,
848
+ NULL, },
849
+ { gen_helper_sve_ldbdu_zsu,
850
+ gen_helper_sve_ldhdu_le_zsu,
851
+ gen_helper_sve_ldsdu_le_zsu,
852
+ gen_helper_sve_lddd_le_zsu, } },
853
+ { { gen_helper_sve_ldbds_zss,
854
+ gen_helper_sve_ldhds_le_zss,
855
+ gen_helper_sve_ldsds_le_zss,
856
+ NULL, },
857
+ { gen_helper_sve_ldbdu_zss,
858
+ gen_helper_sve_ldhdu_le_zss,
859
+ gen_helper_sve_ldsdu_le_zss,
860
+ gen_helper_sve_lddd_le_zss, } },
861
+ { { gen_helper_sve_ldbds_zd,
862
+ gen_helper_sve_ldhds_le_zd,
863
+ gen_helper_sve_ldsds_le_zd,
864
+ NULL, },
865
+ { gen_helper_sve_ldbdu_zd,
866
+ gen_helper_sve_ldhdu_le_zd,
867
+ gen_helper_sve_ldsdu_le_zd,
868
+ gen_helper_sve_lddd_le_zd, } } },
869
870
- /* First-fault */
871
- { { { gen_helper_sve_ldffbds_zsu,
872
- gen_helper_sve_ldffhds_le_zsu,
873
- gen_helper_sve_ldffsds_le_zsu,
874
- NULL, },
875
- { gen_helper_sve_ldffbdu_zsu,
876
- gen_helper_sve_ldffhdu_le_zsu,
877
- gen_helper_sve_ldffsdu_le_zsu,
878
- gen_helper_sve_ldffdd_le_zsu, } },
879
- { { gen_helper_sve_ldffbds_zss,
880
- gen_helper_sve_ldffhds_le_zss,
881
- gen_helper_sve_ldffsds_le_zss,
882
- NULL, },
883
- { gen_helper_sve_ldffbdu_zss,
884
- gen_helper_sve_ldffhdu_le_zss,
885
- gen_helper_sve_ldffsdu_le_zss,
886
- gen_helper_sve_ldffdd_le_zss, } },
887
- { { gen_helper_sve_ldffbds_zd,
888
- gen_helper_sve_ldffhds_le_zd,
889
- gen_helper_sve_ldffsds_le_zd,
890
- NULL, },
891
- { gen_helper_sve_ldffbdu_zd,
892
- gen_helper_sve_ldffhdu_le_zd,
893
- gen_helper_sve_ldffsdu_le_zd,
894
- gen_helper_sve_ldffdd_le_zd, } } } },
895
+ /* First-fault */
896
+ { { { gen_helper_sve_ldffbds_zsu,
897
+ gen_helper_sve_ldffhds_le_zsu,
898
+ gen_helper_sve_ldffsds_le_zsu,
899
+ NULL, },
900
+ { gen_helper_sve_ldffbdu_zsu,
901
+ gen_helper_sve_ldffhdu_le_zsu,
902
+ gen_helper_sve_ldffsdu_le_zsu,
903
+ gen_helper_sve_ldffdd_le_zsu, } },
904
+ { { gen_helper_sve_ldffbds_zss,
905
+ gen_helper_sve_ldffhds_le_zss,
906
+ gen_helper_sve_ldffsds_le_zss,
907
+ NULL, },
908
+ { gen_helper_sve_ldffbdu_zss,
909
+ gen_helper_sve_ldffhdu_le_zss,
910
+ gen_helper_sve_ldffsdu_le_zss,
911
+ gen_helper_sve_ldffdd_le_zss, } },
912
+ { { gen_helper_sve_ldffbds_zd,
913
+ gen_helper_sve_ldffhds_le_zd,
914
+ gen_helper_sve_ldffsds_le_zd,
915
+ NULL, },
916
+ { gen_helper_sve_ldffbdu_zd,
917
+ gen_helper_sve_ldffhdu_le_zd,
918
+ gen_helper_sve_ldffsdu_le_zd,
919
+ gen_helper_sve_ldffdd_le_zd, } } } },
920
+ { /* Big-endian */
921
+ { { { gen_helper_sve_ldbds_zsu,
922
+ gen_helper_sve_ldhds_be_zsu,
923
+ gen_helper_sve_ldsds_be_zsu,
924
+ NULL, },
925
+ { gen_helper_sve_ldbdu_zsu,
926
+ gen_helper_sve_ldhdu_be_zsu,
927
+ gen_helper_sve_ldsdu_be_zsu,
928
+ gen_helper_sve_lddd_be_zsu, } },
929
+ { { gen_helper_sve_ldbds_zss,
930
+ gen_helper_sve_ldhds_be_zss,
931
+ gen_helper_sve_ldsds_be_zss,
932
+ NULL, },
933
+ { gen_helper_sve_ldbdu_zss,
934
+ gen_helper_sve_ldhdu_be_zss,
935
+ gen_helper_sve_ldsdu_be_zss,
936
+ gen_helper_sve_lddd_be_zss, } },
937
+ { { gen_helper_sve_ldbds_zd,
938
+ gen_helper_sve_ldhds_be_zd,
939
+ gen_helper_sve_ldsds_be_zd,
940
+ NULL, },
941
+ { gen_helper_sve_ldbdu_zd,
942
+ gen_helper_sve_ldhdu_be_zd,
943
+ gen_helper_sve_ldsdu_be_zd,
944
+ gen_helper_sve_lddd_be_zd, } } },
945
946
- /* Big-endian */
947
- { { { { gen_helper_sve_ldbds_zsu,
948
- gen_helper_sve_ldhds_be_zsu,
949
- gen_helper_sve_ldsds_be_zsu,
950
- NULL, },
951
- { gen_helper_sve_ldbdu_zsu,
952
- gen_helper_sve_ldhdu_be_zsu,
953
- gen_helper_sve_ldsdu_be_zsu,
954
- gen_helper_sve_lddd_be_zsu, } },
955
- { { gen_helper_sve_ldbds_zss,
956
- gen_helper_sve_ldhds_be_zss,
957
- gen_helper_sve_ldsds_be_zss,
958
- NULL, },
959
- { gen_helper_sve_ldbdu_zss,
960
- gen_helper_sve_ldhdu_be_zss,
961
- gen_helper_sve_ldsdu_be_zss,
962
- gen_helper_sve_lddd_be_zss, } },
963
- { { gen_helper_sve_ldbds_zd,
964
- gen_helper_sve_ldhds_be_zd,
965
- gen_helper_sve_ldsds_be_zd,
966
- NULL, },
967
- { gen_helper_sve_ldbdu_zd,
968
- gen_helper_sve_ldhdu_be_zd,
969
- gen_helper_sve_ldsdu_be_zd,
970
- gen_helper_sve_lddd_be_zd, } } },
971
+ /* First-fault */
972
+ { { { gen_helper_sve_ldffbds_zsu,
973
+ gen_helper_sve_ldffhds_be_zsu,
974
+ gen_helper_sve_ldffsds_be_zsu,
975
+ NULL, },
976
+ { gen_helper_sve_ldffbdu_zsu,
977
+ gen_helper_sve_ldffhdu_be_zsu,
978
+ gen_helper_sve_ldffsdu_be_zsu,
979
+ gen_helper_sve_ldffdd_be_zsu, } },
980
+ { { gen_helper_sve_ldffbds_zss,
981
+ gen_helper_sve_ldffhds_be_zss,
982
+ gen_helper_sve_ldffsds_be_zss,
983
+ NULL, },
984
+ { gen_helper_sve_ldffbdu_zss,
985
+ gen_helper_sve_ldffhdu_be_zss,
986
+ gen_helper_sve_ldffsdu_be_zss,
987
+ gen_helper_sve_ldffdd_be_zss, } },
988
+ { { gen_helper_sve_ldffbds_zd,
989
+ gen_helper_sve_ldffhds_be_zd,
990
+ gen_helper_sve_ldffsds_be_zd,
991
+ NULL, },
992
+ { gen_helper_sve_ldffbdu_zd,
993
+ gen_helper_sve_ldffhdu_be_zd,
994
+ gen_helper_sve_ldffsdu_be_zd,
995
+ gen_helper_sve_ldffdd_be_zd, } } } } },
996
+ { /* MTE Active */
997
+ { /* Little-endian */
998
+ { { { gen_helper_sve_ldbds_zsu_mte,
999
+ gen_helper_sve_ldhds_le_zsu_mte,
1000
+ gen_helper_sve_ldsds_le_zsu_mte,
1001
+ NULL, },
1002
+ { gen_helper_sve_ldbdu_zsu_mte,
1003
+ gen_helper_sve_ldhdu_le_zsu_mte,
1004
+ gen_helper_sve_ldsdu_le_zsu_mte,
1005
+ gen_helper_sve_lddd_le_zsu_mte, } },
1006
+ { { gen_helper_sve_ldbds_zss_mte,
1007
+ gen_helper_sve_ldhds_le_zss_mte,
1008
+ gen_helper_sve_ldsds_le_zss_mte,
1009
+ NULL, },
1010
+ { gen_helper_sve_ldbdu_zss_mte,
1011
+ gen_helper_sve_ldhdu_le_zss_mte,
1012
+ gen_helper_sve_ldsdu_le_zss_mte,
1013
+ gen_helper_sve_lddd_le_zss_mte, } },
1014
+ { { gen_helper_sve_ldbds_zd_mte,
1015
+ gen_helper_sve_ldhds_le_zd_mte,
1016
+ gen_helper_sve_ldsds_le_zd_mte,
1017
+ NULL, },
1018
+ { gen_helper_sve_ldbdu_zd_mte,
1019
+ gen_helper_sve_ldhdu_le_zd_mte,
1020
+ gen_helper_sve_ldsdu_le_zd_mte,
1021
+ gen_helper_sve_lddd_le_zd_mte, } } },
1022
1023
- /* First-fault */
1024
- { { { gen_helper_sve_ldffbds_zsu,
1025
- gen_helper_sve_ldffhds_be_zsu,
1026
- gen_helper_sve_ldffsds_be_zsu,
1027
- NULL, },
1028
- { gen_helper_sve_ldffbdu_zsu,
1029
- gen_helper_sve_ldffhdu_be_zsu,
1030
- gen_helper_sve_ldffsdu_be_zsu,
1031
- gen_helper_sve_ldffdd_be_zsu, } },
1032
- { { gen_helper_sve_ldffbds_zss,
1033
- gen_helper_sve_ldffhds_be_zss,
1034
- gen_helper_sve_ldffsds_be_zss,
1035
- NULL, },
1036
- { gen_helper_sve_ldffbdu_zss,
1037
- gen_helper_sve_ldffhdu_be_zss,
1038
- gen_helper_sve_ldffsdu_be_zss,
1039
- gen_helper_sve_ldffdd_be_zss, } },
1040
- { { gen_helper_sve_ldffbds_zd,
1041
- gen_helper_sve_ldffhds_be_zd,
1042
- gen_helper_sve_ldffsds_be_zd,
1043
- NULL, },
1044
- { gen_helper_sve_ldffbdu_zd,
1045
- gen_helper_sve_ldffhdu_be_zd,
1046
- gen_helper_sve_ldffsdu_be_zd,
1047
- gen_helper_sve_ldffdd_be_zd, } } } },
1048
+ /* First-fault */
1049
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1050
+ gen_helper_sve_ldffhds_le_zsu_mte,
1051
+ gen_helper_sve_ldffsds_le_zsu_mte,
1052
+ NULL, },
1053
+ { gen_helper_sve_ldffbdu_zsu_mte,
1054
+ gen_helper_sve_ldffhdu_le_zsu_mte,
1055
+ gen_helper_sve_ldffsdu_le_zsu_mte,
1056
+ gen_helper_sve_ldffdd_le_zsu_mte, } },
1057
+ { { gen_helper_sve_ldffbds_zss_mte,
1058
+ gen_helper_sve_ldffhds_le_zss_mte,
1059
+ gen_helper_sve_ldffsds_le_zss_mte,
1060
+ NULL, },
1061
+ { gen_helper_sve_ldffbdu_zss_mte,
1062
+ gen_helper_sve_ldffhdu_le_zss_mte,
1063
+ gen_helper_sve_ldffsdu_le_zss_mte,
1064
+ gen_helper_sve_ldffdd_le_zss_mte, } },
1065
+ { { gen_helper_sve_ldffbds_zd_mte,
1066
+ gen_helper_sve_ldffhds_le_zd_mte,
1067
+ gen_helper_sve_ldffsds_le_zd_mte,
1068
+ NULL, },
1069
+ { gen_helper_sve_ldffbdu_zd_mte,
1070
+ gen_helper_sve_ldffhdu_le_zd_mte,
1071
+ gen_helper_sve_ldffsdu_le_zd_mte,
1072
+ gen_helper_sve_ldffdd_le_zd_mte, } } } },
1073
+ { /* Big-endian */
1074
+ { { { gen_helper_sve_ldbds_zsu_mte,
1075
+ gen_helper_sve_ldhds_be_zsu_mte,
1076
+ gen_helper_sve_ldsds_be_zsu_mte,
1077
+ NULL, },
1078
+ { gen_helper_sve_ldbdu_zsu_mte,
1079
+ gen_helper_sve_ldhdu_be_zsu_mte,
1080
+ gen_helper_sve_ldsdu_be_zsu_mte,
1081
+ gen_helper_sve_lddd_be_zsu_mte, } },
1082
+ { { gen_helper_sve_ldbds_zss_mte,
1083
+ gen_helper_sve_ldhds_be_zss_mte,
1084
+ gen_helper_sve_ldsds_be_zss_mte,
1085
+ NULL, },
1086
+ { gen_helper_sve_ldbdu_zss_mte,
1087
+ gen_helper_sve_ldhdu_be_zss_mte,
1088
+ gen_helper_sve_ldsdu_be_zss_mte,
1089
+ gen_helper_sve_lddd_be_zss_mte, } },
1090
+ { { gen_helper_sve_ldbds_zd_mte,
1091
+ gen_helper_sve_ldhds_be_zd_mte,
1092
+ gen_helper_sve_ldsds_be_zd_mte,
1093
+ NULL, },
1094
+ { gen_helper_sve_ldbdu_zd_mte,
1095
+ gen_helper_sve_ldhdu_be_zd_mte,
1096
+ gen_helper_sve_ldsdu_be_zd_mte,
1097
+ gen_helper_sve_lddd_be_zd_mte, } } },
1098
+
1099
+ /* First-fault */
1100
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1101
+ gen_helper_sve_ldffhds_be_zsu_mte,
1102
+ gen_helper_sve_ldffsds_be_zsu_mte,
1103
+ NULL, },
1104
+ { gen_helper_sve_ldffbdu_zsu_mte,
1105
+ gen_helper_sve_ldffhdu_be_zsu_mte,
1106
+ gen_helper_sve_ldffsdu_be_zsu_mte,
1107
+ gen_helper_sve_ldffdd_be_zsu_mte, } },
1108
+ { { gen_helper_sve_ldffbds_zss_mte,
1109
+ gen_helper_sve_ldffhds_be_zss_mte,
1110
+ gen_helper_sve_ldffsds_be_zss_mte,
1111
+ NULL, },
1112
+ { gen_helper_sve_ldffbdu_zss_mte,
1113
+ gen_helper_sve_ldffhdu_be_zss_mte,
1114
+ gen_helper_sve_ldffsdu_be_zss_mte,
1115
+ gen_helper_sve_ldffdd_be_zss_mte, } },
1116
+ { { gen_helper_sve_ldffbds_zd_mte,
1117
+ gen_helper_sve_ldffhds_be_zd_mte,
1118
+ gen_helper_sve_ldffsds_be_zd_mte,
1119
+ NULL, },
1120
+ { gen_helper_sve_ldffbdu_zd_mte,
1121
+ gen_helper_sve_ldffhdu_be_zd_mte,
1122
+ gen_helper_sve_ldffsdu_be_zd_mte,
1123
+ gen_helper_sve_ldffdd_be_zd_mte, } } } } },
1124
};
1125
1126
static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1127
{
1128
gen_helper_gvec_mem_scatter *fn = NULL;
1129
- int be = s->be_data == MO_BE;
1130
+ bool be = s->be_data == MO_BE;
1131
+ bool mte = s->mte_active[0];
1132
1133
if (!sve_access_check(s)) {
1134
return true;
1135
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1136
1137
switch (a->esz) {
1138
case MO_32:
1139
- fn = gather_load_fn32[be][a->ff][a->xs][a->u][a->msz];
1140
+ fn = gather_load_fn32[mte][be][a->ff][a->xs][a->u][a->msz];
1141
break;
1142
case MO_64:
1143
- fn = gather_load_fn64[be][a->ff][a->xs][a->u][a->msz];
1144
+ fn = gather_load_fn64[mte][be][a->ff][a->xs][a->u][a->msz];
1145
break;
1146
}
1147
assert(fn != NULL);
1148
1149
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1150
- cpu_reg_sp(s, a->rn), a->msz, fn);
1151
+ cpu_reg_sp(s, a->rn), a->msz, false, fn);
1152
return true;
1153
}
1154
1155
static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1156
{
1157
gen_helper_gvec_mem_scatter *fn = NULL;
1158
- int be = s->be_data == MO_BE;
1159
+ bool be = s->be_data == MO_BE;
1160
+ bool mte = s->mte_active[0];
1161
TCGv_i64 imm;
1162
1163
if (a->esz < a->msz || (a->esz == a->msz && !a->u)) {
1164
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1165
1166
switch (a->esz) {
1167
case MO_32:
1168
- fn = gather_load_fn32[be][a->ff][0][a->u][a->msz];
1169
+ fn = gather_load_fn32[mte][be][a->ff][0][a->u][a->msz];
1170
break;
1171
case MO_64:
1172
- fn = gather_load_fn64[be][a->ff][2][a->u][a->msz];
1173
+ fn = gather_load_fn64[mte][be][a->ff][2][a->u][a->msz];
1174
break;
1175
}
1176
assert(fn != NULL);
1177
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1178
* by loading the immediate into the scalar parameter.
1179
*/
1180
imm = tcg_const_i64(a->imm << a->msz);
1181
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1182
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, false, fn);
1183
tcg_temp_free_i64(imm);
1184
return true;
1185
}
1186
1187
-/* Indexed by [be][xs][msz]. */
1188
-static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][3] = {
1189
- /* Little-endian */
1190
- { { gen_helper_sve_stbs_zsu,
1191
- gen_helper_sve_sths_le_zsu,
1192
- gen_helper_sve_stss_le_zsu, },
1193
- { gen_helper_sve_stbs_zss,
1194
- gen_helper_sve_sths_le_zss,
1195
- gen_helper_sve_stss_le_zss, } },
1196
- /* Big-endian */
1197
- { { gen_helper_sve_stbs_zsu,
1198
- gen_helper_sve_sths_be_zsu,
1199
- gen_helper_sve_stss_be_zsu, },
1200
- { gen_helper_sve_stbs_zss,
1201
- gen_helper_sve_sths_be_zss,
1202
- gen_helper_sve_stss_be_zss, } },
1203
+/* Indexed by [mte][be][xs][msz]. */
1204
+static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = {
1205
+ { /* MTE Inactive */
1206
+ { /* Little-endian */
1207
+ { gen_helper_sve_stbs_zsu,
1208
+ gen_helper_sve_sths_le_zsu,
1209
+ gen_helper_sve_stss_le_zsu, },
1210
+ { gen_helper_sve_stbs_zss,
1211
+ gen_helper_sve_sths_le_zss,
1212
+ gen_helper_sve_stss_le_zss, } },
1213
+ { /* Big-endian */
1214
+ { gen_helper_sve_stbs_zsu,
1215
+ gen_helper_sve_sths_be_zsu,
1216
+ gen_helper_sve_stss_be_zsu, },
1217
+ { gen_helper_sve_stbs_zss,
1218
+ gen_helper_sve_sths_be_zss,
1219
+ gen_helper_sve_stss_be_zss, } } },
1220
+ { /* MTE Active */
1221
+ { /* Little-endian */
1222
+ { gen_helper_sve_stbs_zsu_mte,
1223
+ gen_helper_sve_sths_le_zsu_mte,
1224
+ gen_helper_sve_stss_le_zsu_mte, },
1225
+ { gen_helper_sve_stbs_zss_mte,
1226
+ gen_helper_sve_sths_le_zss_mte,
1227
+ gen_helper_sve_stss_le_zss_mte, } },
1228
+ { /* Big-endian */
1229
+ { gen_helper_sve_stbs_zsu_mte,
1230
+ gen_helper_sve_sths_be_zsu_mte,
1231
+ gen_helper_sve_stss_be_zsu_mte, },
1232
+ { gen_helper_sve_stbs_zss_mte,
1233
+ gen_helper_sve_sths_be_zss_mte,
1234
+ gen_helper_sve_stss_be_zss_mte, } } },
1235
};
1236
1237
/* Note that we overload xs=2 to indicate 64-bit offset. */
1238
-static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][3][4] = {
1239
- /* Little-endian */
1240
- { { gen_helper_sve_stbd_zsu,
1241
- gen_helper_sve_sthd_le_zsu,
1242
- gen_helper_sve_stsd_le_zsu,
1243
- gen_helper_sve_stdd_le_zsu, },
1244
- { gen_helper_sve_stbd_zss,
1245
- gen_helper_sve_sthd_le_zss,
1246
- gen_helper_sve_stsd_le_zss,
1247
- gen_helper_sve_stdd_le_zss, },
1248
- { gen_helper_sve_stbd_zd,
1249
- gen_helper_sve_sthd_le_zd,
1250
- gen_helper_sve_stsd_le_zd,
1251
- gen_helper_sve_stdd_le_zd, } },
1252
- /* Big-endian */
1253
- { { gen_helper_sve_stbd_zsu,
1254
- gen_helper_sve_sthd_be_zsu,
1255
- gen_helper_sve_stsd_be_zsu,
1256
- gen_helper_sve_stdd_be_zsu, },
1257
- { gen_helper_sve_stbd_zss,
1258
- gen_helper_sve_sthd_be_zss,
1259
- gen_helper_sve_stsd_be_zss,
1260
- gen_helper_sve_stdd_be_zss, },
1261
- { gen_helper_sve_stbd_zd,
1262
- gen_helper_sve_sthd_be_zd,
1263
- gen_helper_sve_stsd_be_zd,
1264
- gen_helper_sve_stdd_be_zd, } },
1265
+static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][2][3][4] = {
1266
+ { /* MTE Inactive */
1267
+ { /* Little-endian */
1268
+ { gen_helper_sve_stbd_zsu,
1269
+ gen_helper_sve_sthd_le_zsu,
1270
+ gen_helper_sve_stsd_le_zsu,
1271
+ gen_helper_sve_stdd_le_zsu, },
1272
+ { gen_helper_sve_stbd_zss,
1273
+ gen_helper_sve_sthd_le_zss,
1274
+ gen_helper_sve_stsd_le_zss,
1275
+ gen_helper_sve_stdd_le_zss, },
1276
+ { gen_helper_sve_stbd_zd,
1277
+ gen_helper_sve_sthd_le_zd,
1278
+ gen_helper_sve_stsd_le_zd,
1279
+ gen_helper_sve_stdd_le_zd, } },
1280
+ { /* Big-endian */
1281
+ { gen_helper_sve_stbd_zsu,
1282
+ gen_helper_sve_sthd_be_zsu,
1283
+ gen_helper_sve_stsd_be_zsu,
1284
+ gen_helper_sve_stdd_be_zsu, },
1285
+ { gen_helper_sve_stbd_zss,
1286
+ gen_helper_sve_sthd_be_zss,
1287
+ gen_helper_sve_stsd_be_zss,
1288
+ gen_helper_sve_stdd_be_zss, },
1289
+ { gen_helper_sve_stbd_zd,
1290
+ gen_helper_sve_sthd_be_zd,
1291
+ gen_helper_sve_stsd_be_zd,
1292
+ gen_helper_sve_stdd_be_zd, } } },
1293
+ { /* MTE Inactive */
1294
+ { /* Little-endian */
1295
+ { gen_helper_sve_stbd_zsu_mte,
1296
+ gen_helper_sve_sthd_le_zsu_mte,
1297
+ gen_helper_sve_stsd_le_zsu_mte,
1298
+ gen_helper_sve_stdd_le_zsu_mte, },
1299
+ { gen_helper_sve_stbd_zss_mte,
1300
+ gen_helper_sve_sthd_le_zss_mte,
1301
+ gen_helper_sve_stsd_le_zss_mte,
1302
+ gen_helper_sve_stdd_le_zss_mte, },
1303
+ { gen_helper_sve_stbd_zd_mte,
1304
+ gen_helper_sve_sthd_le_zd_mte,
1305
+ gen_helper_sve_stsd_le_zd_mte,
1306
+ gen_helper_sve_stdd_le_zd_mte, } },
1307
+ { /* Big-endian */
1308
+ { gen_helper_sve_stbd_zsu_mte,
1309
+ gen_helper_sve_sthd_be_zsu_mte,
1310
+ gen_helper_sve_stsd_be_zsu_mte,
1311
+ gen_helper_sve_stdd_be_zsu_mte, },
1312
+ { gen_helper_sve_stbd_zss_mte,
1313
+ gen_helper_sve_sthd_be_zss_mte,
1314
+ gen_helper_sve_stsd_be_zss_mte,
1315
+ gen_helper_sve_stdd_be_zss_mte, },
1316
+ { gen_helper_sve_stbd_zd_mte,
1317
+ gen_helper_sve_sthd_be_zd_mte,
1318
+ gen_helper_sve_stsd_be_zd_mte,
1319
+ gen_helper_sve_stdd_be_zd_mte, } } },
1320
};
1321
1322
static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1323
{
1324
gen_helper_gvec_mem_scatter *fn;
1325
- int be = s->be_data == MO_BE;
1326
+ bool be = s->be_data == MO_BE;
1327
+ bool mte = s->mte_active[0];
1328
1329
if (a->esz < a->msz || (a->msz == 0 && a->scale)) {
1330
return false;
1331
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1332
}
1333
switch (a->esz) {
1334
case MO_32:
1335
- fn = scatter_store_fn32[be][a->xs][a->msz];
1336
+ fn = scatter_store_fn32[mte][be][a->xs][a->msz];
1337
break;
1338
case MO_64:
1339
- fn = scatter_store_fn64[be][a->xs][a->msz];
1340
+ fn = scatter_store_fn64[mte][be][a->xs][a->msz];
1341
break;
1342
default:
1343
g_assert_not_reached();
1344
}
1345
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
1346
- cpu_reg_sp(s, a->rn), a->msz, fn);
1347
+ cpu_reg_sp(s, a->rn), a->msz, true, fn);
1348
return true;
1349
}
1350
1351
static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1352
{
1353
gen_helper_gvec_mem_scatter *fn = NULL;
1354
- int be = s->be_data == MO_BE;
1355
+ bool be = s->be_data == MO_BE;
1356
+ bool mte = s->mte_active[0];
1357
TCGv_i64 imm;
1358
1359
if (a->esz < a->msz) {
1360
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1361
1362
switch (a->esz) {
1363
case MO_32:
1364
- fn = scatter_store_fn32[be][0][a->msz];
1365
+ fn = scatter_store_fn32[mte][be][0][a->msz];
1366
break;
1367
case MO_64:
1368
- fn = scatter_store_fn64[be][2][a->msz];
1369
+ fn = scatter_store_fn64[mte][be][2][a->msz];
1370
break;
1371
}
1372
assert(fn != NULL);
1373
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1374
* by loading the immediate into the scalar parameter.
1375
*/
1376
imm = tcg_const_i64(a->imm << a->msz);
1377
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1378
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, true, fn);
1379
tcg_temp_free_i64(imm);
1380
return true;
1381
}
108
--
1382
--
109
2.19.1
1383
2.20.1
110
1384
111
1385
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
There are a number of paths by which the TBI is still intact
4
for user-only in the SVE helpers.
5
6
Because we currently always set TBI for user-only, we do not
7
need to pass down the actual TBI setting from above, and we
8
can remove the top byte in the inner-most primitives, so that
9
none are forgotten. Moreover, this keeps the "dirty" pointer
10
around at the higher levels, where we need it for any MTE checking.
11
12
Since the normal case, especially for user-only, goes through
13
RAM, this clearing merely adds two insns per page lookup, which
14
will be completely in the noise.
15
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20181011205206.3552-18-richard.henderson@linaro.org
18
Message-id: 20200626033144.790098-39-richard.henderson@linaro.org
5
[PMM: added parens in ?: expression]
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
20
---
9
target/arm/translate.c | 81 ++++++++++++++----------------------------
21
target/arm/cpu.c | 3 +++
10
1 file changed, 26 insertions(+), 55 deletions(-)
22
target/arm/sve_helper.c | 19 +++++++++++++++++--
23
target/arm/translate-a64.c | 5 +++++
24
3 files changed, 25 insertions(+), 2 deletions(-)
11
25
12
diff --git a/target/arm/translate.c b/target/arm/translate.c
26
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
13
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/translate.c
28
--- a/target/arm/cpu.c
15
+++ b/target/arm/translate.c
29
+++ b/target/arm/cpu.c
16
@@ -XXX,XX +XXX,XX @@ static void gen_vfp_msr(TCGv_i32 tmp)
30
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
17
tcg_temp_free_i32(tmp);
31
* Enable TBI0 and TBI1. While the real kernel only enables TBI0,
32
* turning on both here will produce smaller code and otherwise
33
* make no difference to the user-level emulation.
34
+ *
35
+ * In sve_probe_page, we assume that this is set.
36
+ * Do not modify this without other changes.
37
*/
38
env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
39
#else
40
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/sve_helper.c
43
+++ b/target/arm/sve_helper.c
44
@@ -XXX,XX +XXX,XX @@ static void sve_##NAME##_host(void *vd, intptr_t reg_off, void *host) \
45
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
46
target_ulong addr, uintptr_t ra) \
47
{ \
48
- *(TYPEE *)(vd + H(reg_off)) = (TYPEM)TLB(env, addr, ra); \
49
+ *(TYPEE *)(vd + H(reg_off)) = \
50
+ (TYPEM)TLB(env, useronly_clean_ptr(addr), ra); \
18
}
51
}
19
52
20
-static void gen_neon_dup_u8(TCGv_i32 var, int shift)
53
#define DO_ST_TLB(NAME, H, TYPEE, TYPEM, TLB) \
21
-{
54
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
22
- TCGv_i32 tmp = tcg_temp_new_i32();
55
target_ulong addr, uintptr_t ra) \
23
- if (shift)
56
{ \
24
- tcg_gen_shri_i32(var, var, shift);
57
- TLB(env, addr, (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
25
- tcg_gen_ext8u_i32(var, var);
58
+ TLB(env, useronly_clean_ptr(addr), \
26
- tcg_gen_shli_i32(tmp, var, 8);
59
+ (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
27
- tcg_gen_or_i32(var, var, tmp);
28
- tcg_gen_shli_i32(tmp, var, 16);
29
- tcg_gen_or_i32(var, var, tmp);
30
- tcg_temp_free_i32(tmp);
31
-}
32
-
33
static void gen_neon_dup_low16(TCGv_i32 var)
34
{
35
TCGv_i32 tmp = tcg_temp_new_i32();
36
@@ -XXX,XX +XXX,XX @@ static void gen_neon_dup_high16(TCGv_i32 var)
37
tcg_temp_free_i32(tmp);
38
}
60
}
39
61
40
-static TCGv_i32 gen_load_and_replicate(DisasContext *s, TCGv_i32 addr, int size)
62
#define DO_LD_PRIM_1(NAME, H, TE, TM) \
41
-{
63
@@ -XXX,XX +XXX,XX @@ static bool sve_probe_page(SVEHostPage *info, bool nofault,
42
- /* Load a single Neon element and replicate into a 32 bit TCG reg */
64
int flags;
43
- TCGv_i32 tmp = tcg_temp_new_i32();
65
44
- switch (size) {
66
addr += mem_off;
45
- case 0:
46
- gen_aa32_ld8u(s, tmp, addr, get_mem_index(s));
47
- gen_neon_dup_u8(tmp, 0);
48
- break;
49
- case 1:
50
- gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
51
- gen_neon_dup_low16(tmp);
52
- break;
53
- case 2:
54
- gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
55
- break;
56
- default: /* Avoid compiler warnings. */
57
- abort();
58
- }
59
- return tmp;
60
-}
61
-
62
static int handle_vsel(uint32_t insn, uint32_t rd, uint32_t rn, uint32_t rm,
63
uint32_t dp)
64
{
65
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
66
int load;
67
int shift;
68
int n;
69
+ int vec_size;
70
TCGv_i32 addr;
71
TCGv_i32 tmp;
72
TCGv_i32 tmp2;
73
@@ -XXX,XX +XXX,XX @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
74
}
75
addr = tcg_temp_new_i32();
76
load_reg_var(s, addr, rn);
77
- if (nregs == 1) {
78
- /* VLD1 to all lanes: bit 5 indicates how many Dregs to write */
79
- tmp = gen_load_and_replicate(s, addr, size);
80
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd, 0));
81
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd, 1));
82
- if (insn & (1 << 5)) {
83
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd + 1, 0));
84
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd + 1, 1));
85
- }
86
- tcg_temp_free_i32(tmp);
87
- } else {
88
- /* VLD2/3/4 to all lanes: bit 5 indicates register stride */
89
- stride = (insn & (1 << 5)) ? 2 : 1;
90
- for (reg = 0; reg < nregs; reg++) {
91
- tmp = gen_load_and_replicate(s, addr, size);
92
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd, 0));
93
- tcg_gen_st_i32(tmp, cpu_env, neon_reg_offset(rd, 1));
94
- tcg_temp_free_i32(tmp);
95
- tcg_gen_addi_i32(addr, addr, 1 << size);
96
- rd += stride;
97
+
67
+
98
+ /* VLD1 to all lanes: bit 5 indicates how many Dregs to write.
68
+ /*
99
+ * VLD2/3/4 to all lanes: bit 5 indicates register stride.
69
+ * User-only currently always issues with TBI. See the comment
100
+ */
70
+ * above useronly_clean_ptr. Usually we clean this top byte away
101
+ stride = (insn & (1 << 5)) ? 2 : 1;
71
+ * during translation, but we can't do that for e.g. vector + imm
102
+ vec_size = nregs == 1 ? stride * 8 : 8;
72
+ * addressing modes.
73
+ *
74
+ * We currently always enable TBI for user-only, and do not provide
75
+ * a way to turn it off. So clean the pointer unconditionally here,
76
+ * rather than look it up here, or pass it down from above.
77
+ */
78
+ addr = useronly_clean_ptr(addr);
103
+
79
+
104
+ tmp = tcg_temp_new_i32();
80
flags = probe_access_flags(env, addr, access_type, mmu_idx, nofault,
105
+ for (reg = 0; reg < nregs; reg++) {
81
&info->host, retaddr);
106
+ gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s),
82
info->flags = flags;
107
+ s->be_data | size);
83
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
108
+ if ((rd & 1) && vec_size == 16) {
84
index XXXXXXX..XXXXXXX 100644
109
+ /* We cannot write 16 bytes at once because the
85
--- a/target/arm/translate-a64.c
110
+ * destination is unaligned.
86
+++ b/target/arm/translate-a64.c
111
+ */
87
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
112
+ tcg_gen_gvec_dup_i32(size, neon_reg_offset(rd, 0),
88
dc->features = env->features;
113
+ 8, 8, tmp);
89
dc->dcz_blocksize = arm_cpu->dcz_blocksize;
114
+ tcg_gen_gvec_mov(0, neon_reg_offset(rd + 1, 0),
90
115
+ neon_reg_offset(rd, 0), 8, 8);
91
+#ifdef CONFIG_USER_ONLY
116
+ } else {
92
+ /* In sve_probe_page, we assume TBI is enabled. */
117
+ tcg_gen_gvec_dup_i32(size, neon_reg_offset(rd, 0),
93
+ tcg_debug_assert(dc->tbid & 1);
118
+ vec_size, vec_size, tmp);
94
+#endif
119
}
95
+
120
+ tcg_gen_addi_i32(addr, addr, 1 << size);
96
/* Single step state. The code-generation logic here is:
121
+ rd += stride;
97
* SS_ACTIVE == 0:
122
}
98
* generate code with no special handling for single-stepping (except
123
+ tcg_temp_free_i32(tmp);
124
tcg_temp_free_i32(addr);
125
stride = (1 << size) * nregs;
126
} else {
127
--
99
--
128
2.19.1
100
2.20.1
129
101
130
102
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
3
This is DC GVA and DC GZVA, and the tag check for DC ZVA.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20181016223115.24100-9-richard.henderson@linaro.org
7
Message-id: 20200626033144.790098-40-richard.henderson@linaro.org
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
9
---
9
target/arm/cpu.h | 17 +++++++++++++++-
10
target/arm/cpu.h | 4 +++-
10
linux-user/elfload.c | 6 +-----
11
target/arm/helper.c | 16 ++++++++++++++++
11
target/arm/cpu64.c | 16 ++++++++-------
12
target/arm/translate-a64.c | 39 ++++++++++++++++++++++++++++++++++++++
12
target/arm/helper.c | 2 +-
13
3 files changed, 58 insertions(+), 1 deletion(-)
13
target/arm/translate-a64.c | 40 +++++++++++++++++++-------------------
14
target/arm/translate.c | 6 +++---
15
6 files changed, 50 insertions(+), 37 deletions(-)
16
14
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ enum arm_features {
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
22
ARM_FEATURE_PMU, /* has PMU support */
20
#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
23
ARM_FEATURE_VBAR, /* has cp15 VBAR */
21
#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
24
ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
22
#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
25
- ARM_FEATURE_V8_FP16, /* implements v8.2 half-precision float */
23
-#define ARM_LAST_SPECIAL ARM_CP_DC_ZVA
26
ARM_FEATURE_M_MAIN, /* M profile Main Extension */
24
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
27
};
25
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
28
26
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
29
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa32_dp(const ARMISARegisters *id)
27
#define ARM_CP_FPU 0x1000
30
return FIELD_EX32(id->id_isar6, ID_ISAR6, DP) != 0;
28
#define ARM_CP_SVE 0x2000
31
}
29
#define ARM_CP_NO_GDB 0x4000
32
33
+static inline bool isar_feature_aa32_fp16_arith(const ARMISARegisters *id)
34
+{
35
+ /*
36
+ * This is a placeholder for use by VCMA until the rest of
37
+ * the ARMv8.2-FP16 extension is implemented for aa32 mode.
38
+ * At which point we can properly set and check MVFR1.FPHP.
39
+ */
40
+ return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, FP) == 1;
41
+}
42
+
43
/*
44
* 64-bit feature tests via id registers.
45
*/
46
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_fcma(const ARMISARegisters *id)
47
return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, FCMA) != 0;
48
}
49
50
+static inline bool isar_feature_aa64_fp16(const ARMISARegisters *id)
51
+{
52
+ /* We always set the AdvSIMD and FP fields identically wrt FP16. */
53
+ return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, FP) == 1;
54
+}
55
+
56
static inline bool isar_feature_aa64_sve(const ARMISARegisters *id)
57
{
58
return FIELD_EX64(id->id_aa64pfr0, ID_AA64PFR0, SVE) != 0;
59
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/linux-user/elfload.c
62
+++ b/linux-user/elfload.c
63
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
64
hwcaps |= ARM_HWCAP_A64_ASIMD;
65
66
/* probe for the extra features */
67
-#define GET_FEATURE(feat, hwcap) \
68
- do { if (arm_feature(&cpu->env, feat)) { hwcaps |= hwcap; } } while (0)
69
#define GET_FEATURE_ID(feat, hwcap) \
70
do { if (cpu_isar_feature(feat, cpu)) { hwcaps |= hwcap; } } while (0)
71
72
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
73
GET_FEATURE_ID(aa64_sha3, ARM_HWCAP_A64_SHA3);
74
GET_FEATURE_ID(aa64_sm3, ARM_HWCAP_A64_SM3);
75
GET_FEATURE_ID(aa64_sm4, ARM_HWCAP_A64_SM4);
76
- GET_FEATURE(ARM_FEATURE_V8_FP16,
77
- ARM_HWCAP_A64_FPHP | ARM_HWCAP_A64_ASIMDHP);
78
+ GET_FEATURE_ID(aa64_fp16, ARM_HWCAP_A64_FPHP | ARM_HWCAP_A64_ASIMDHP);
79
GET_FEATURE_ID(aa64_atomics, ARM_HWCAP_A64_ATOMICS);
80
GET_FEATURE_ID(aa64_rdm, ARM_HWCAP_A64_ASIMDRDM);
81
GET_FEATURE_ID(aa64_dp, ARM_HWCAP_A64_ASIMDDP);
82
GET_FEATURE_ID(aa64_fcma, ARM_HWCAP_A64_FCMA);
83
GET_FEATURE_ID(aa64_sve, ARM_HWCAP_A64_SVE);
84
85
-#undef GET_FEATURE
86
#undef GET_FEATURE_ID
87
88
return hwcaps;
89
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
90
index XXXXXXX..XXXXXXX 100644
91
--- a/target/arm/cpu64.c
92
+++ b/target/arm/cpu64.c
93
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
94
95
t = cpu->isar.id_aa64pfr0;
96
t = FIELD_DP64(t, ID_AA64PFR0, SVE, 1);
97
+ t = FIELD_DP64(t, ID_AA64PFR0, FP, 1);
98
+ t = FIELD_DP64(t, ID_AA64PFR0, ADVSIMD, 1);
99
cpu->isar.id_aa64pfr0 = t;
100
101
/* Replicate the same data to the 32-bit id registers. */
102
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
103
u = FIELD_DP32(u, ID_ISAR6, DP, 1);
104
cpu->isar.id_isar6 = u;
105
106
-#ifdef CONFIG_USER_ONLY
107
- /* We don't set these in system emulation mode for the moment,
108
- * since we don't correctly set the ID registers to advertise them,
109
- * and in some cases they're only available in AArch64 and not AArch32,
110
- * whereas the architecture requires them to be present in both if
111
- * present in either.
112
+ /*
113
+ * FIXME: We do not yet support ARMv8.2-fp16 for AArch32 yet,
114
+ * so do not set MVFR1.FPHP. Strictly speaking this is not legal,
115
+ * but it is also not legal to enable SVE without support for FP16,
116
+ * and enabling SVE in system mode is more useful in the short term.
117
*/
118
- set_feature(&cpu->env, ARM_FEATURE_V8_FP16);
119
+
120
+#ifdef CONFIG_USER_ONLY
121
/* For usermode -cpu max we can use a larger and more efficient DCZ
122
* blocksize since we don't have to follow what the hardware does.
123
*/
124
diff --git a/target/arm/helper.c b/target/arm/helper.c
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
125
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
126
--- a/target/arm/helper.c
32
--- a/target/arm/helper.c
127
+++ b/target/arm/helper.c
33
+++ b/target/arm/helper.c
128
@@ -XXX,XX +XXX,XX @@ void HELPER(vfp_set_fpscr)(CPUARMState *env, uint32_t val)
34
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
129
uint32_t changed;
35
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
130
36
.type = ARM_CP_NOP, .access = PL0_W,
131
/* When ARMv8.2-FP16 is not supported, FZ16 is RES0. */
37
.accessfn = aa64_cacheop_poc_access },
132
- if (!arm_feature(env, ARM_FEATURE_V8_FP16)) {
38
+ { .name = "DC_GVA", .state = ARM_CP_STATE_AA64,
133
+ if (!cpu_isar_feature(aa64_fp16, arm_env_get_cpu(env))) {
39
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 3,
134
val &= ~FPCR_FZ16;
40
+ .access = PL0_W, .type = ARM_CP_DC_GVA,
135
}
41
+#ifndef CONFIG_USER_ONLY
42
+ /* Avoid overhead of an access check that always passes in user-mode */
43
+ .accessfn = aa64_zva_access,
44
+#endif
45
+ },
46
+ { .name = "DC_GZVA", .state = ARM_CP_STATE_AA64,
47
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 4,
48
+ .access = PL0_W, .type = ARM_CP_DC_GZVA,
49
+#ifndef CONFIG_USER_ONLY
50
+ /* Avoid overhead of an access check that always passes in user-mode */
51
+ .accessfn = aa64_zva_access,
52
+#endif
53
+ },
54
REGINFO_SENTINEL
55
};
136
56
137
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
57
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
138
index XXXXXXX..XXXXXXX 100644
58
index XXXXXXX..XXXXXXX 100644
139
--- a/target/arm/translate-a64.c
59
--- a/target/arm/translate-a64.c
140
+++ b/target/arm/translate-a64.c
60
+++ b/target/arm/translate-a64.c
141
@@ -XXX,XX +XXX,XX @@ static void disas_fp_compare(DisasContext *s, uint32_t insn)
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
142
break;
143
case 3:
144
size = MO_16;
145
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
146
+ if (dc_isar_feature(aa64_fp16, s)) {
147
break;
148
}
62
}
149
/* fallthru */
63
gen_helper_dc_zva(cpu_env, tcg_rt);
150
@@ -XXX,XX +XXX,XX @@ static void disas_fp_ccomp(DisasContext *s, uint32_t insn)
151
break;
152
case 3:
153
size = MO_16;
154
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
155
+ if (dc_isar_feature(aa64_fp16, s)) {
156
break;
157
}
158
/* fallthru */
159
@@ -XXX,XX +XXX,XX @@ static void disas_fp_csel(DisasContext *s, uint32_t insn)
160
break;
161
case 3:
162
sz = MO_16;
163
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
164
+ if (dc_isar_feature(aa64_fp16, s)) {
165
break;
166
}
167
/* fallthru */
168
@@ -XXX,XX +XXX,XX @@ static void disas_fp_1src(DisasContext *s, uint32_t insn)
169
handle_fp_1src_double(s, opcode, rd, rn);
170
break;
171
case 3:
172
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
173
+ if (!dc_isar_feature(aa64_fp16, s)) {
174
unallocated_encoding(s);
175
return;
176
}
177
@@ -XXX,XX +XXX,XX @@ static void disas_fp_2src(DisasContext *s, uint32_t insn)
178
handle_fp_2src_double(s, opcode, rd, rn, rm);
179
break;
180
case 3:
181
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
182
+ if (!dc_isar_feature(aa64_fp16, s)) {
183
unallocated_encoding(s);
184
return;
185
}
186
@@ -XXX,XX +XXX,XX @@ static void disas_fp_3src(DisasContext *s, uint32_t insn)
187
handle_fp_3src_double(s, o0, o1, rd, rn, rm, ra);
188
break;
189
case 3:
190
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
191
+ if (!dc_isar_feature(aa64_fp16, s)) {
192
unallocated_encoding(s);
193
return;
194
}
195
@@ -XXX,XX +XXX,XX @@ static void disas_fp_imm(DisasContext *s, uint32_t insn)
196
break;
197
case 3:
198
sz = MO_16;
199
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
200
+ if (dc_isar_feature(aa64_fp16, s)) {
201
break;
202
}
203
/* fallthru */
204
@@ -XXX,XX +XXX,XX @@ static void disas_fp_fixed_conv(DisasContext *s, uint32_t insn)
205
case 1: /* float64 */
206
break;
207
case 3: /* float16 */
208
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
209
+ if (dc_isar_feature(aa64_fp16, s)) {
210
break;
211
}
212
/* fallthru */
213
@@ -XXX,XX +XXX,XX @@ static void disas_fp_int_conv(DisasContext *s, uint32_t insn)
214
break;
215
case 0x6: /* 16-bit float, 32-bit int */
216
case 0xe: /* 16-bit float, 64-bit int */
217
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
218
+ if (dc_isar_feature(aa64_fp16, s)) {
219
break;
220
}
221
/* fallthru */
222
@@ -XXX,XX +XXX,XX @@ static void disas_fp_int_conv(DisasContext *s, uint32_t insn)
223
case 1: /* float64 */
224
break;
225
case 3: /* float16 */
226
- if (arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
227
+ if (dc_isar_feature(aa64_fp16, s)) {
228
break;
229
}
230
/* fallthru */
231
@@ -XXX,XX +XXX,XX @@ static void disas_simd_across_lanes(DisasContext *s, uint32_t insn)
232
*/
233
is_min = extract32(size, 1, 1);
234
is_fp = true;
235
- if (!is_u && arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
236
+ if (!is_u && dc_isar_feature(aa64_fp16, s)) {
237
size = 1;
238
} else if (!is_u || !is_q || extract32(size, 0, 1)) {
239
unallocated_encoding(s);
240
@@ -XXX,XX +XXX,XX @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
241
242
if (o2 != 0 || ((cmode == 0xf) && is_neg && !is_q)) {
243
/* Check for FMOV (vector, immediate) - half-precision */
244
- if (!(arm_dc_feature(s, ARM_FEATURE_V8_FP16) && o2 && cmode == 0xf)) {
245
+ if (!(dc_isar_feature(aa64_fp16, s) && o2 && cmode == 0xf)) {
246
unallocated_encoding(s);
247
return;
248
}
249
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_pairwise(DisasContext *s, uint32_t insn)
250
case 0x2f: /* FMINP */
251
/* FP op, size[0] is 32 or 64 bit*/
252
if (!u) {
253
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
254
+ if (!dc_isar_feature(aa64_fp16, s)) {
255
unallocated_encoding(s);
256
return;
257
} else {
258
@@ -XXX,XX +XXX,XX @@ static void handle_simd_shift_intfp_conv(DisasContext *s, bool is_scalar,
259
size = MO_32;
260
} else if (immh & 2) {
261
size = MO_16;
262
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
263
+ if (!dc_isar_feature(aa64_fp16, s)) {
264
unallocated_encoding(s);
265
return;
266
}
267
@@ -XXX,XX +XXX,XX @@ static void handle_simd_shift_fpint_conv(DisasContext *s, bool is_scalar,
268
size = MO_32;
269
} else if (immh & 0x2) {
270
size = MO_16;
271
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
272
+ if (!dc_isar_feature(aa64_fp16, s)) {
273
unallocated_encoding(s);
274
return;
275
}
276
@@ -XXX,XX +XXX,XX @@ static void disas_simd_scalar_three_reg_same_fp16(DisasContext *s,
277
return;
64
return;
278
}
65
+ case ARM_CP_DC_GVA:
279
66
+ {
280
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
67
+ TCGv_i64 clean_addr, tag;
281
+ if (!dc_isar_feature(aa64_fp16, s)) {
68
+
282
unallocated_encoding(s);
69
+ /*
283
}
70
+ * DC_GVA, like DC_ZVA, requires that we supply the original
284
71
+ * pointer for an invalid page. Probe that address first.
285
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_fp16(DisasContext *s, uint32_t insn)
72
+ */
286
TCGv_ptr fpst;
73
+ tcg_rt = cpu_reg(s, rt);
287
bool pairwise = false;
74
+ clean_addr = clean_data_tbi(s, tcg_rt);
288
75
+ gen_probe_access(s, clean_addr, MMU_DATA_STORE, MO_8);
289
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
76
+
290
+ if (!dc_isar_feature(aa64_fp16, s)) {
77
+ if (s->ata) {
291
unallocated_encoding(s);
78
+ /* Extract the tag from the register to match STZGM. */
292
return;
79
+ tag = tcg_temp_new_i64();
293
}
80
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
294
@@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn)
81
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
295
case 0x1c: /* FCADD, #90 */
82
+ tcg_temp_free_i64(tag);
296
case 0x1e: /* FCADD, #270 */
83
+ }
297
if (size == 0
84
+ }
298
- || (size == 1 && !arm_dc_feature(s, ARM_FEATURE_V8_FP16))
85
+ return;
299
+ || (size == 1 && !dc_isar_feature(aa64_fp16, s))
86
+ case ARM_CP_DC_GZVA:
300
|| (size == 3 && !is_q)) {
87
+ {
301
unallocated_encoding(s);
88
+ TCGv_i64 clean_addr, tag;
302
return;
89
+
303
@@ -XXX,XX +XXX,XX @@ static void disas_simd_two_reg_misc_fp16(DisasContext *s, uint32_t insn)
90
+ /* For DC_GZVA, we can rely on DC_ZVA for the proper fault. */
304
bool need_fpst = true;
91
+ tcg_rt = cpu_reg(s, rt);
305
int rmode;
92
+ clean_addr = clean_data_tbi(s, tcg_rt);
306
93
+ gen_helper_dc_zva(cpu_env, clean_addr);
307
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
94
+
308
+ if (!dc_isar_feature(aa64_fp16, s)) {
95
+ if (s->ata) {
309
unallocated_encoding(s);
96
+ /* Extract the tag from the register to match STZGM. */
310
return;
97
+ tag = tcg_temp_new_i64();
311
}
98
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
312
@@ -XXX,XX +XXX,XX @@ static void disas_simd_indexed(DisasContext *s, uint32_t insn)
99
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
313
}
100
+ tcg_temp_free_i64(tag);
101
+ }
102
+ }
103
+ return;
104
default:
314
break;
105
break;
315
}
106
}
316
- if (is_fp16 && !arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
317
+ if (is_fp16 && !dc_isar_feature(aa64_fp16, s)) {
318
unallocated_encoding(s);
319
return;
320
}
321
diff --git a/target/arm/translate.c b/target/arm/translate.c
322
index XXXXXXX..XXXXXXX 100644
323
--- a/target/arm/translate.c
324
+++ b/target/arm/translate.c
325
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
326
int size = extract32(insn, 20, 1);
327
data = extract32(insn, 23, 2); /* rot */
328
if (!dc_isar_feature(aa32_vcma, s)
329
- || (!size && !arm_dc_feature(s, ARM_FEATURE_V8_FP16))) {
330
+ || (!size && !dc_isar_feature(aa32_fp16_arith, s))) {
331
return 1;
332
}
333
fn_gvec_ptr = size ? gen_helper_gvec_fcmlas : gen_helper_gvec_fcmlah;
334
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_3same_ext(DisasContext *s, uint32_t insn)
335
int size = extract32(insn, 20, 1);
336
data = extract32(insn, 24, 1); /* rot */
337
if (!dc_isar_feature(aa32_vcma, s)
338
- || (!size && !arm_dc_feature(s, ARM_FEATURE_V8_FP16))) {
339
+ || (!size && !dc_isar_feature(aa32_fp16_arith, s))) {
340
return 1;
341
}
342
fn_gvec_ptr = size ? gen_helper_gvec_fcadds : gen_helper_gvec_fcaddh;
343
@@ -XXX,XX +XXX,XX @@ static int disas_neon_insn_2reg_scalar_ext(DisasContext *s, uint32_t insn)
344
return 1;
345
}
346
if (size == 0) {
347
- if (!arm_dc_feature(s, ARM_FEATURE_V8_FP16)) {
348
+ if (!dc_isar_feature(aa32_fp16_arith, s)) {
349
return 1;
350
}
351
/* For fp16, rm is just Vm, and index is M. */
352
--
107
--
353
2.19.1
108
2.20.1
354
109
355
110
diff view generated by jsdifflib
1
The A/I/F bits in ISR_EL1 should track the virtual interrupt
1
From: Richard Henderson <richard.henderson@linaro.org>
2
status, not the physical interrupt status, if the associated
3
HCR_EL2.AMO/IMO/FMO bit is set. Implement this, rather than
4
always showing the physical interrupt status.
5
2
6
We don't currently implement anything to do with external
3
D1.10 specifies that exception handlers begin with tag checks overridden.
7
aborts, so this applies only to the I and F bits (though it
8
ought to be possible for the outer guest to present a virtual
9
external abort to the inner guest, even if QEMU doesn't
10
emulate physical external aborts, so there is missing
11
functionality in this area).
12
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-41-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 20181012144235.19646-6-peter.maydell@linaro.org
16
---
9
---
17
target/arm/helper.c | 22 ++++++++++++++++++----
10
target/arm/helper.c | 3 +++
18
1 file changed, 18 insertions(+), 4 deletions(-)
11
1 file changed, 3 insertions(+)
19
12
20
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
21
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
23
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
24
@@ -XXX,XX +XXX,XX @@ static uint64_t isr_read(CPUARMState *env, const ARMCPRegInfo *ri)
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
25
CPUState *cs = ENV_GET_CPU(env);
18
break;
26
uint64_t ret = 0;
19
}
27
28
- if (cs->interrupt_request & CPU_INTERRUPT_HARD) {
29
- ret |= CPSR_I;
30
+ if (arm_hcr_el2_imo(env)) {
31
+ if (cs->interrupt_request & CPU_INTERRUPT_VIRQ) {
32
+ ret |= CPSR_I;
33
+ }
34
+ } else {
35
+ if (cs->interrupt_request & CPU_INTERRUPT_HARD) {
36
+ ret |= CPSR_I;
37
+ }
38
}
20
}
39
- if (cs->interrupt_request & CPU_INTERRUPT_FIQ) {
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
40
- ret |= CPSR_F;
22
+ new_mode |= PSTATE_TCO;
41
+
23
+ }
42
+ if (arm_hcr_el2_fmo(env)) {
24
43
+ if (cs->interrupt_request & CPU_INTERRUPT_VFIQ) {
25
pstate_write(env, PSTATE_DAIF | new_mode);
44
+ ret |= CPSR_F;
26
env->aarch64 = 1;
45
+ }
46
+ } else {
47
+ if (cs->interrupt_request & CPU_INTERRUPT_FIQ) {
48
+ ret |= CPSR_F;
49
+ }
50
}
51
+
52
/* External aborts are not possible in QEMU so A bit is always clear */
53
return ret;
54
}
55
--
27
--
56
2.19.1
28
2.20.1
57
29
58
30
diff view generated by jsdifflib
1
The switch_mode() function is defined in target/arm/helper.c and used
1
From: Richard Henderson <richard.henderson@linaro.org>
2
only in that file and nowhere else, so we can make it file-local
2
3
rather than global.
3
We need to check the memattr of a page in order to determine
4
4
whether it is Tagged for MTE. Between Stage1 and Stage2,
5
this becomes simpler if we always collect this data, instead
6
of occasionally being presented with NULL.
7
8
Use the nonnull attribute to allow the compiler to check that
9
all pointer arguments are non-null.
10
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200626033144.790098-42-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20181012144235.19646-3-peter.maydell@linaro.org
8
---
15
---
9
target/arm/internals.h | 1 -
16
target/arm/internals.h | 3 ++-
10
target/arm/helper.c | 6 ++++--
17
target/arm/helper.c | 60 ++++++++++++++++++++---------------------
11
2 files changed, 4 insertions(+), 3 deletions(-)
18
target/arm/m_helper.c | 11 +++++---
19
target/arm/tlb_helper.c | 4 ++-
20
4 files changed, 42 insertions(+), 36 deletions(-)
12
21
13
diff --git a/target/arm/internals.h b/target/arm/internals.h
22
diff --git a/target/arm/internals.h b/target/arm/internals.h
14
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/internals.h
24
--- a/target/arm/internals.h
16
+++ b/target/arm/internals.h
25
+++ b/target/arm/internals.h
17
@@ -XXX,XX +XXX,XX @@ static inline int bank_number(int mode)
26
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
18
g_assert_not_reached();
27
MMUAccessType access_type, ARMMMUIdx mmu_idx,
19
}
28
hwaddr *phys_ptr, MemTxAttrs *attrs, int *prot,
20
29
target_ulong *page_size,
21
-void switch_mode(CPUARMState *, int);
30
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
22
void arm_cpu_register_gdb_regs_for_features(ARMCPU *cpu);
31
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
23
void arm_translate_init(void);
32
+ __attribute__((nonnull));
33
34
void arm_log_exception(int idx);
24
35
25
diff --git a/target/arm/helper.c b/target/arm/helper.c
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
26
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/helper.c
38
--- a/target/arm/helper.c
28
+++ b/target/arm/helper.c
39
+++ b/target/arm/helper.c
29
@@ -XXX,XX +XXX,XX @@ static void v8m_security_lookup(CPUARMState *env, uint32_t address,
40
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
30
V8M_SAttributes *sattrs);
41
bool s1_is_el0,
42
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
43
target_ulong *page_size_ptr,
44
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
45
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
46
+ __attribute__((nonnull));
31
#endif
47
#endif
32
48
33
+static void switch_mode(CPUARMState *env, int mode);
49
static void switch_mode(CPUARMState *env, int mode);
34
+
50
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
35
static int vfp_gdb_get_reg(CPUARMState *env, uint8_t *buf, int reg)
51
arm_tlb_bti_gp(txattrs) = true;
36
{
52
}
37
int nregs;
53
38
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_tt)(CPUARMState *env, uint32_t addr, uint32_t op)
54
- if (cacheattrs != NULL) {
39
return 0;
55
- if (mmu_idx == ARMMMUIdx_Stage2) {
40
}
56
- cacheattrs->attrs = convert_stage2_attrs(env,
41
57
- extract32(attrs, 0, 4));
42
-void switch_mode(CPUARMState *env, int mode)
58
- } else {
43
+static void switch_mode(CPUARMState *env, int mode)
59
- /* Index into MAIR registers for cache attributes */
44
{
60
- uint8_t attrindx = extract32(attrs, 0, 3);
45
ARMCPU *cpu = arm_env_get_cpu(env);
61
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
46
62
- assert(attrindx <= 7);
47
@@ -XXX,XX +XXX,XX @@ void aarch64_sync_64_to_32(CPUARMState *env)
63
- cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
48
64
- }
49
#else
65
- cacheattrs->shareability = extract32(attrs, 6, 2);
50
66
+ if (mmu_idx == ARMMMUIdx_Stage2) {
51
-void switch_mode(CPUARMState *env, int mode)
67
+ cacheattrs->attrs = convert_stage2_attrs(env, extract32(attrs, 0, 4));
52
+static void switch_mode(CPUARMState *env, int mode)
68
+ } else {
53
{
69
+ /* Index into MAIR registers for cache attributes */
54
int old_mode;
70
+ uint8_t attrindx = extract32(attrs, 0, 3);
55
int i;
71
+ uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
72
+ assert(attrindx <= 7);
73
+ cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
74
}
75
+ cacheattrs->shareability = extract32(attrs, 6, 2);
76
77
*phys_ptr = descaddr;
78
*page_size_ptr = page_size;
79
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
80
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
81
mmu_idx == ARMMMUIdx_E10_0,
82
phys_ptr, attrs, &s2_prot,
83
- page_size, fi,
84
- cacheattrs != NULL ? &cacheattrs2 : NULL);
85
+ page_size, fi, &cacheattrs2);
86
fi->s2addr = ipa;
87
/* Combine the S1 and S2 perms. */
88
*prot &= s2_prot;
89
90
- /* Combine the S1 and S2 cache attributes, if needed */
91
- if (!ret && cacheattrs != NULL) {
92
- if (env->cp15.hcr_el2 & HCR_DC) {
93
- /*
94
- * HCR.DC forces the first stage attributes to
95
- * Normal Non-Shareable,
96
- * Inner Write-Back Read-Allocate Write-Allocate,
97
- * Outer Write-Back Read-Allocate Write-Allocate.
98
- */
99
- cacheattrs->attrs = 0xff;
100
- cacheattrs->shareability = 0;
101
- }
102
- *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
103
+ /* If S2 fails, return early. */
104
+ if (ret) {
105
+ return ret;
106
}
107
108
- return ret;
109
+ /* Combine the S1 and S2 cache attributes. */
110
+ if (env->cp15.hcr_el2 & HCR_DC) {
111
+ /*
112
+ * HCR.DC forces the first stage attributes to
113
+ * Normal Non-Shareable,
114
+ * Inner Write-Back Read-Allocate Write-Allocate,
115
+ * Outer Write-Back Read-Allocate Write-Allocate.
116
+ */
117
+ cacheattrs->attrs = 0xff;
118
+ cacheattrs->shareability = 0;
119
+ }
120
+ *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
121
+ return 0;
122
} else {
123
/*
124
* For non-EL2 CPUs a stage1+stage2 translation is just stage 1.
125
@@ -XXX,XX +XXX,XX @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
126
bool ret;
127
ARMMMUFaultInfo fi = {};
128
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
129
+ ARMCacheAttrs cacheattrs = {};
130
131
*attrs = (MemTxAttrs) {};
132
133
ret = get_phys_addr(env, addr, 0, mmu_idx, &phys_addr,
134
- attrs, &prot, &page_size, &fi, NULL);
135
+ attrs, &prot, &page_size, &fi, &cacheattrs);
136
137
if (ret) {
138
return -1;
139
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
140
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/m_helper.c
142
+++ b/target/arm/m_helper.c
143
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
144
hwaddr physaddr;
145
int prot;
146
ARMMMUFaultInfo fi = {};
147
+ ARMCacheAttrs cacheattrs = {};
148
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
149
int exc;
150
bool exc_secure;
151
152
if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &physaddr,
153
- &attrs, &prot, &page_size, &fi, NULL)) {
154
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
155
/* MPU/SAU lookup failed */
156
if (fi.type == ARMFault_QEMU_SFault) {
157
if (mode == STACK_LAZYFP) {
158
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
159
hwaddr physaddr;
160
int prot;
161
ARMMMUFaultInfo fi = {};
162
+ ARMCacheAttrs cacheattrs = {};
163
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
164
int exc;
165
bool exc_secure;
166
uint32_t value;
167
168
if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
169
- &attrs, &prot, &page_size, &fi, NULL)) {
170
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
171
/* MPU/SAU lookup failed */
172
if (fi.type == ARMFault_QEMU_SFault) {
173
qemu_log_mask(CPU_LOG_INT,
174
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
175
V8M_SAttributes sattrs = {};
176
MemTxAttrs attrs = {};
177
ARMMMUFaultInfo fi = {};
178
+ ARMCacheAttrs cacheattrs = {};
179
MemTxResult txres;
180
target_ulong page_size;
181
hwaddr physaddr;
182
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
183
"...really SecureFault with SFSR.INVEP\n");
184
return false;
185
}
186
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx,
187
- &physaddr, &attrs, &prot, &page_size, &fi, NULL)) {
188
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &physaddr,
189
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
190
/* the MPU lookup failed */
191
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
192
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
193
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
194
index XXXXXXX..XXXXXXX 100644
195
--- a/target/arm/tlb_helper.c
196
+++ b/target/arm/tlb_helper.c
197
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
198
int prot, ret;
199
MemTxAttrs attrs = {};
200
ARMMMUFaultInfo fi = {};
201
+ ARMCacheAttrs cacheattrs = {};
202
203
/*
204
* Walk the page table and (if the mapping exists) add the page
205
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
206
*/
207
ret = get_phys_addr(&cpu->env, address, access_type,
208
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
209
- &phys_addr, &attrs, &prot, &page_size, &fi, NULL);
210
+ &phys_addr, &attrs, &prot, &page_size,
211
+ &fi, &cacheattrs);
212
if (likely(!ret)) {
213
/*
214
* Map a single [sub]page. Regions smaller than our declared
56
--
215
--
57
2.19.1
216
2.20.1
58
217
59
218
diff view generated by jsdifflib
1
The HCR.DC virtualization configuration register bit has the
1
From: Richard Henderson <richard.henderson@linaro.org>
2
following effects:
3
* SCTLR.M behaves as if it is 0 for all purposes except
4
direct reads of the bit
5
* HCR.VM behaves as if it is 1 for all purposes except
6
direct reads of the bit
7
* the memory type produced by the first stage of the EL1&EL0
8
translation regime is Normal Non-Shareable,
9
Inner Write-Back Read-Allocate Write-Allocate,
10
Outer Write-Back Read-Allocate Write-Allocate.
11
2
12
Implement this behaviour.
3
This "bit" is a particular value of the page's MemAttr.
13
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-43-richard.henderson@linaro.org
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
16
Message-id: 20181012144235.19646-5-peter.maydell@linaro.org
17
---
9
---
18
target/arm/helper.c | 23 +++++++++++++++++++++--
10
target/arm/helper.c | 48 ++++++++++++++++++++++++++++++++++++++---
19
1 file changed, 21 insertions(+), 2 deletions(-)
11
target/arm/tlb_helper.c | 5 +++++
12
2 files changed, 50 insertions(+), 3 deletions(-)
20
13
21
diff --git a/target/arm/helper.c b/target/arm/helper.c
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
22
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/helper.c
16
--- a/target/arm/helper.c
24
+++ b/target/arm/helper.c
17
+++ b/target/arm/helper.c
25
@@ -XXX,XX +XXX,XX @@ static uint64_t do_ats_write(CPUARMState *env, uint64_t value,
18
@@ -XXX,XX +XXX,XX @@ static uint8_t combine_cacheattr_nibble(uint8_t s1, uint8_t s2)
26
* * The Non-secure TTBCR.EAE bit is set to 1
19
*/
27
* * The implementation includes EL2, and the value of HCR.VM is 1
20
static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
28
*
21
{
29
+ * (Note that HCR.DC makes HCR.VM behave as if it is 1.)
22
- uint8_t s1lo = extract32(s1.attrs, 0, 4), s2lo = extract32(s2.attrs, 0, 4);
30
+ *
23
- uint8_t s1hi = extract32(s1.attrs, 4, 4), s2hi = extract32(s2.attrs, 4, 4);
31
* ATS1Hx always uses the 64bit format (not supported yet).
24
+ uint8_t s1lo, s2lo, s1hi, s2hi;
32
*/
25
ARMCacheAttrs ret;
33
format64 = arm_s1_regime_using_lpae_format(env, mmu_idx);
26
+ bool tagged = false;
34
27
+
35
if (arm_feature(env, ARM_FEATURE_EL2)) {
28
+ if (s1.attrs == 0xf0) {
36
if (mmu_idx == ARMMMUIdx_S12NSE0 || mmu_idx == ARMMMUIdx_S12NSE1) {
29
+ tagged = true;
37
- format64 |= env->cp15.hcr_el2 & HCR_VM;
30
+ s1.attrs = 0xff;
38
+ format64 |= env->cp15.hcr_el2 & (HCR_VM | HCR_DC);
31
+ }
39
} else {
32
+
40
format64 |= arm_current_el(env) == 2;
33
+ s1lo = extract32(s1.attrs, 0, 4);
41
}
34
+ s2lo = extract32(s2.attrs, 0, 4);
42
@@ -XXX,XX +XXX,XX @@ static inline bool regime_translation_disabled(CPUARMState *env,
35
+ s1hi = extract32(s1.attrs, 4, 4);
43
}
36
+ s2hi = extract32(s2.attrs, 4, 4);
44
37
45
if (mmu_idx == ARMMMUIdx_S2NS) {
38
/* Combine shareability attributes (table D4-43) */
46
- return (env->cp15.hcr_el2 & HCR_VM) == 0;
39
if (s1.shareability == 2 || s2.shareability == 2) {
47
+ /* HCR.DC means HCR.VM behaves as 1 */
40
@@ -XXX,XX +XXX,XX @@ static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
48
+ return (env->cp15.hcr_el2 & (HCR_DC | HCR_VM)) == 0;
49
}
50
51
if (env->cp15.hcr_el2 & HCR_TGE) {
52
@@ -XXX,XX +XXX,XX @@ static inline bool regime_translation_disabled(CPUARMState *env,
53
}
41
}
54
}
42
}
55
43
56
+ if ((env->cp15.hcr_el2 & HCR_DC) &&
44
+ /* TODO: CombineS1S2Desc does not consider transient, only WB, RWA. */
57
+ (mmu_idx == ARMMMUIdx_S1NSE0 || mmu_idx == ARMMMUIdx_S1NSE1)) {
45
+ if (tagged && ret.attrs == 0xff) {
58
+ /* HCR.DC means SCTLR_EL1.M behaves as 0 */
46
+ ret.attrs = 0xf0;
59
+ return true;
60
+ }
47
+ }
61
+
48
+
62
return (regime_sctlr(env, mmu_idx) & SCTLR_M) == 0;
49
return ret;
63
}
50
}
64
51
65
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr(CPUARMState *env, target_ulong address,
52
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
66
53
* Normal Non-Shareable,
67
/* Combine the S1 and S2 cache attributes, if needed */
54
* Inner Write-Back Read-Allocate Write-Allocate,
68
if (!ret && cacheattrs != NULL) {
55
* Outer Write-Back Read-Allocate Write-Allocate.
69
+ if (env->cp15.hcr_el2 & HCR_DC) {
56
+ * Do not overwrite Tagged within attrs.
70
+ /*
57
*/
71
+ * HCR.DC forces the first stage attributes to
58
- cacheattrs->attrs = 0xff;
72
+ * Normal Non-Shareable,
59
+ if (cacheattrs->attrs != 0xf0) {
73
+ * Inner Write-Back Read-Allocate Write-Allocate,
74
+ * Outer Write-Back Read-Allocate Write-Allocate.
75
+ */
76
+ cacheattrs->attrs = 0xff;
60
+ cacheattrs->attrs = 0xff;
77
+ cacheattrs->shareability = 0;
78
+ }
61
+ }
79
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
62
cacheattrs->shareability = 0;
80
}
63
}
81
64
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
65
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
66
/* Definitely a real MMU, not an MPU */
67
68
if (regime_translation_disabled(env, mmu_idx)) {
69
+ uint64_t hcr;
70
+ uint8_t memattr;
71
+
72
/*
73
* MMU disabled. S1 addresses within aa64 translation regimes are
74
* still checked for bounds -- see AArch64.TranslateAddressS1Off.
75
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
76
*phys_ptr = address;
77
*prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
78
*page_size = TARGET_PAGE_SIZE;
79
+
80
+ /* Fill in cacheattr a-la AArch64.TranslateAddressS1Off. */
81
+ hcr = arm_hcr_el2_eff(env);
82
+ cacheattrs->shareability = 0;
83
+ if (hcr & HCR_DC) {
84
+ if (hcr & HCR_DCT) {
85
+ memattr = 0xf0; /* Tagged, Normal, WB, RWA */
86
+ } else {
87
+ memattr = 0xff; /* Normal, WB, RWA */
88
+ }
89
+ } else if (access_type == MMU_INST_FETCH) {
90
+ if (regime_sctlr(env, mmu_idx) & SCTLR_I) {
91
+ memattr = 0xee; /* Normal, WT, RA, NT */
92
+ } else {
93
+ memattr = 0x44; /* Normal, NC, No */
94
+ }
95
+ cacheattrs->shareability = 2; /* outer sharable */
96
+ } else {
97
+ memattr = 0x00; /* Device, nGnRnE */
98
+ }
99
+ cacheattrs->attrs = memattr;
100
return 0;
101
}
102
103
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
104
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/tlb_helper.c
106
+++ b/target/arm/tlb_helper.c
107
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
108
phys_addr &= TARGET_PAGE_MASK;
109
address &= TARGET_PAGE_MASK;
110
}
111
+ /* Notice and record tagged memory. */
112
+ if (cpu_isar_feature(aa64_mte, cpu) && cacheattrs.attrs == 0xf0) {
113
+ arm_tlb_mte_tagged(&attrs) = true;
114
+ }
115
+
116
tlb_set_page_with_attrs(cs, address, phys_addr, attrs,
117
prot, mmu_idx, page_size);
118
return true;
82
--
119
--
83
2.19.1
120
2.20.1
84
121
85
122
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20181016223115.24100-7-richard.henderson@linaro.org
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-44-richard.henderson@linaro.org
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
7
---
9
target/arm/cpu.h | 6 +++++-
8
target/arm/cpu.h | 6 ++++++
10
linux-user/elfload.c | 2 +-
9
hw/arm/virt.c | 55 ++++++++++++++++++++++++++++++++++++++++++++++--
11
target/arm/cpu.c | 4 ----
10
target/arm/cpu.c | 52 +++++++++++++++++++++++++++++++++++++++++----
12
target/arm/helper.c | 2 +-
11
3 files changed, 107 insertions(+), 6 deletions(-)
13
target/arm/machine.c | 3 +--
14
5 files changed, 8 insertions(+), 9 deletions(-)
15
12
16
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
17
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/cpu.h
15
--- a/target/arm/cpu.h
19
+++ b/target/arm/cpu.h
16
+++ b/target/arm/cpu.h
20
@@ -XXX,XX +XXX,XX @@ enum arm_features {
17
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
21
ARM_FEATURE_NEON,
18
/* MemoryRegion to use for secure physical accesses */
22
ARM_FEATURE_M, /* Microcontroller profile. */
19
MemoryRegion *secure_memory;
23
ARM_FEATURE_OMAPCP, /* OMAP specific CP15 ops handling. */
20
24
- ARM_FEATURE_THUMB2EE,
21
+ /* MemoryRegion to use for allocation tag accesses */
25
ARM_FEATURE_V7MP, /* v7 Multiprocessing Extensions */
22
+ MemoryRegion *tag_memory;
26
ARM_FEATURE_V7VE, /* v7 Virtualization Extensions (non-EL2 parts) */
23
+ MemoryRegion *secure_tag_memory;
27
ARM_FEATURE_V4T,
24
+
28
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_jazelle(const ARMISARegisters *id)
25
/* For v8M, pointer to the IDAU interface provided by board/SoC */
29
return FIELD_EX32(id->id_isar1, ID_ISAR1, JAZELLE) != 0;
26
Object *idau;
27
28
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
29
typedef enum ARMASIdx {
30
ARMASIdx_NS = 0,
31
ARMASIdx_S = 1,
32
+ ARMASIdx_TagNS = 2,
33
+ ARMASIdx_TagS = 3,
34
} ARMASIdx;
35
36
/* Return the Exception Level targeted by debug exceptions. */
37
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
38
index XXXXXXX..XXXXXXX 100644
39
--- a/hw/arm/virt.c
40
+++ b/hw/arm/virt.c
41
@@ -XXX,XX +XXX,XX @@ static void create_platform_bus(VirtMachineState *vms)
42
sysbus_mmio_get_region(s, 0));
30
}
43
}
31
44
32
+static inline bool isar_feature_t32ee(const ARMISARegisters *id)
45
+static void create_tag_ram(MemoryRegion *tag_sysmem,
46
+ hwaddr base, hwaddr size,
47
+ const char *name)
33
+{
48
+{
34
+ return FIELD_EX32(id->id_isar3, ID_ISAR3, T32EE) != 0;
49
+ MemoryRegion *tagram = g_new(MemoryRegion, 1);
50
+
51
+ memory_region_init_ram(tagram, NULL, name, size / 32, &error_fatal);
52
+ memory_region_add_subregion(tag_sysmem, base / 32, tagram);
35
+}
53
+}
36
+
54
+
37
static inline bool isar_feature_aa32_aes(const ARMISARegisters *id)
55
static void create_secure_ram(VirtMachineState *vms,
56
- MemoryRegion *secure_sysmem)
57
+ MemoryRegion *secure_sysmem,
58
+ MemoryRegion *secure_tag_sysmem)
38
{
59
{
39
return FIELD_EX32(id->id_isar5, ID_ISAR5, AES) != 0;
60
MemoryRegion *secram = g_new(MemoryRegion, 1);
40
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
61
char *nodename;
41
index XXXXXXX..XXXXXXX 100644
62
@@ -XXX,XX +XXX,XX @@ static void create_secure_ram(VirtMachineState *vms,
42
--- a/linux-user/elfload.c
63
qemu_fdt_setprop_string(vms->fdt, nodename, "status", "disabled");
43
+++ b/linux-user/elfload.c
64
qemu_fdt_setprop_string(vms->fdt, nodename, "secure-status", "okay");
44
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
65
45
GET_FEATURE(ARM_FEATURE_V5, ARM_HWCAP_ARM_EDSP);
66
+ if (secure_tag_sysmem) {
46
GET_FEATURE(ARM_FEATURE_VFP, ARM_HWCAP_ARM_VFP);
67
+ create_tag_ram(secure_tag_sysmem, base, size, "mach-virt.secure-tag");
47
GET_FEATURE(ARM_FEATURE_IWMMXT, ARM_HWCAP_ARM_IWMMXT);
68
+ }
48
- GET_FEATURE(ARM_FEATURE_THUMB2EE, ARM_HWCAP_ARM_THUMBEE);
69
+
49
+ GET_FEATURE_ID(t32ee, ARM_HWCAP_ARM_THUMBEE);
70
g_free(nodename);
50
GET_FEATURE(ARM_FEATURE_NEON, ARM_HWCAP_ARM_NEON);
71
}
51
GET_FEATURE(ARM_FEATURE_VFP3, ARM_HWCAP_ARM_VFPv3);
72
52
GET_FEATURE(ARM_FEATURE_V6K, ARM_HWCAP_ARM_TLS);
73
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
74
const CPUArchIdList *possible_cpus;
75
MemoryRegion *sysmem = get_system_memory();
76
MemoryRegion *secure_sysmem = NULL;
77
+ MemoryRegion *tag_sysmem = NULL;
78
+ MemoryRegion *secure_tag_sysmem = NULL;
79
int n, virt_max_cpus;
80
bool firmware_loaded;
81
bool aarch64 = true;
82
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
83
"secure-memory", &error_abort);
84
}
85
86
+ /*
87
+ * The cpu adds the property if and only if MemTag is supported.
88
+ * If it is, we must allocate the ram to back that up.
89
+ */
90
+ if (object_property_find(cpuobj, "tag-memory", NULL)) {
91
+ if (!tag_sysmem) {
92
+ tag_sysmem = g_new(MemoryRegion, 1);
93
+ memory_region_init(tag_sysmem, OBJECT(machine),
94
+ "tag-memory", UINT64_MAX / 32);
95
+
96
+ if (vms->secure) {
97
+ secure_tag_sysmem = g_new(MemoryRegion, 1);
98
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
99
+ "secure-tag-memory", UINT64_MAX / 32);
100
+
101
+ /* As with ram, secure-tag takes precedence over tag. */
102
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
103
+ tag_sysmem, -1);
104
+ }
105
+ }
106
+
107
+ object_property_set_link(cpuobj, OBJECT(tag_sysmem),
108
+ "tag-memory", &error_abort);
109
+ if (vms->secure) {
110
+ object_property_set_link(cpuobj, OBJECT(secure_tag_sysmem),
111
+ "secure-tag-memory", &error_abort);
112
+ }
113
+ }
114
+
115
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
116
object_unref(cpuobj);
117
}
118
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
119
create_uart(vms, VIRT_UART, sysmem, serial_hd(0));
120
121
if (vms->secure) {
122
- create_secure_ram(vms, secure_sysmem);
123
+ create_secure_ram(vms, secure_sysmem, secure_tag_sysmem);
124
create_uart(vms, VIRT_SECURE_UART, secure_sysmem, serial_hd(1));
125
}
126
127
+ if (tag_sysmem) {
128
+ create_tag_ram(tag_sysmem, vms->memmap[VIRT_MEM].base,
129
+ machine->ram_size, "mach-virt.tag");
130
+ }
131
+
132
vms->highmem_ecam &= vms->highmem && (!firmware_loaded || aarch64);
133
134
create_rtc(vms);
53
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
135
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
54
index XXXXXXX..XXXXXXX 100644
136
index XXXXXXX..XXXXXXX 100644
55
--- a/target/arm/cpu.c
137
--- a/target/arm/cpu.c
56
+++ b/target/arm/cpu.c
138
+++ b/target/arm/cpu.c
57
@@ -XXX,XX +XXX,XX @@ static void cortex_a8_initfn(Object *obj)
139
@@ -XXX,XX +XXX,XX @@ void arm_cpu_post_init(Object *obj)
58
set_feature(&cpu->env, ARM_FEATURE_V7);
140
if (kvm_enabled()) {
59
set_feature(&cpu->env, ARM_FEATURE_VFP3);
141
kvm_arm_add_vcpu_properties(obj);
60
set_feature(&cpu->env, ARM_FEATURE_NEON);
142
}
61
- set_feature(&cpu->env, ARM_FEATURE_THUMB2EE);
143
+
62
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
144
+#ifndef CONFIG_USER_ONLY
63
set_feature(&cpu->env, ARM_FEATURE_EL3);
145
+ if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) &&
64
cpu->midr = 0x410fc080;
146
+ cpu_isar_feature(aa64_mte, cpu)) {
65
@@ -XXX,XX +XXX,XX @@ static void cortex_a9_initfn(Object *obj)
147
+ object_property_add_link(obj, "tag-memory",
66
set_feature(&cpu->env, ARM_FEATURE_VFP3);
148
+ TYPE_MEMORY_REGION,
67
set_feature(&cpu->env, ARM_FEATURE_VFP_FP16);
149
+ (Object **)&cpu->tag_memory,
68
set_feature(&cpu->env, ARM_FEATURE_NEON);
150
+ qdev_prop_allow_set_link_before_realize,
69
- set_feature(&cpu->env, ARM_FEATURE_THUMB2EE);
151
+ OBJ_PROP_LINK_STRONG);
70
set_feature(&cpu->env, ARM_FEATURE_EL3);
152
+
71
/* Note that A9 supports the MP extensions even for
153
+ if (arm_feature(&cpu->env, ARM_FEATURE_EL3)) {
72
* A9UP and single-core A9MP (which are both different
154
+ object_property_add_link(obj, "secure-tag-memory",
73
@@ -XXX,XX +XXX,XX @@ static void cortex_a7_initfn(Object *obj)
155
+ TYPE_MEMORY_REGION,
74
set_feature(&cpu->env, ARM_FEATURE_V7VE);
156
+ (Object **)&cpu->secure_tag_memory,
75
set_feature(&cpu->env, ARM_FEATURE_VFP4);
157
+ qdev_prop_allow_set_link_before_realize,
76
set_feature(&cpu->env, ARM_FEATURE_NEON);
158
+ OBJ_PROP_LINK_STRONG);
77
- set_feature(&cpu->env, ARM_FEATURE_THUMB2EE);
159
+ }
78
set_feature(&cpu->env, ARM_FEATURE_GENERIC_TIMER);
160
+ }
79
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
161
+#endif
80
set_feature(&cpu->env, ARM_FEATURE_CBAR_RO);
81
@@ -XXX,XX +XXX,XX @@ static void cortex_a15_initfn(Object *obj)
82
set_feature(&cpu->env, ARM_FEATURE_V7VE);
83
set_feature(&cpu->env, ARM_FEATURE_VFP4);
84
set_feature(&cpu->env, ARM_FEATURE_NEON);
85
- set_feature(&cpu->env, ARM_FEATURE_THUMB2EE);
86
set_feature(&cpu->env, ARM_FEATURE_GENERIC_TIMER);
87
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
88
set_feature(&cpu->env, ARM_FEATURE_CBAR_RO);
89
diff --git a/target/arm/helper.c b/target/arm/helper.c
90
index XXXXXXX..XXXXXXX 100644
91
--- a/target/arm/helper.c
92
+++ b/target/arm/helper.c
93
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
94
define_arm_cp_regs(cpu, vmsa_pmsa_cp_reginfo);
95
define_arm_cp_regs(cpu, vmsa_cp_reginfo);
96
}
97
- if (arm_feature(env, ARM_FEATURE_THUMB2EE)) {
98
+ if (cpu_isar_feature(t32ee, cpu)) {
99
define_arm_cp_regs(cpu, t2ee_cp_reginfo);
100
}
101
if (arm_feature(env, ARM_FEATURE_GENERIC_TIMER)) {
102
diff --git a/target/arm/machine.c b/target/arm/machine.c
103
index XXXXXXX..XXXXXXX 100644
104
--- a/target/arm/machine.c
105
+++ b/target/arm/machine.c
106
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
107
static bool thumb2ee_needed(void *opaque)
108
{
109
ARMCPU *cpu = opaque;
110
- CPUARMState *env = &cpu->env;
111
112
- return arm_feature(env, ARM_FEATURE_THUMB2EE);
113
+ return cpu_isar_feature(t32ee, cpu);
114
}
162
}
115
163
116
static const VMStateDescription vmstate_thumb2ee = {
164
static void arm_cpu_finalizefn(Object *obj)
165
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
166
#ifndef CONFIG_USER_ONLY
167
MachineState *ms = MACHINE(qdev_get_machine());
168
unsigned int smp_cpus = ms->smp.cpus;
169
+ bool has_secure = cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY);
170
171
- if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
172
- cs->num_ases = 2;
173
+ /*
174
+ * We must set cs->num_ases to the final value before
175
+ * the first call to cpu_address_space_init.
176
+ */
177
+ if (cpu->tag_memory != NULL) {
178
+ cs->num_ases = 3 + has_secure;
179
+ } else {
180
+ cs->num_ases = 1 + has_secure;
181
+ }
182
183
+ if (has_secure) {
184
if (!cpu->secure_memory) {
185
cpu->secure_memory = cs->memory;
186
}
187
cpu_address_space_init(cs, ARMASIdx_S, "cpu-secure-memory",
188
cpu->secure_memory);
189
- } else {
190
- cs->num_ases = 1;
191
}
192
+
193
+ if (cpu->tag_memory != NULL) {
194
+ cpu_address_space_init(cs, ARMASIdx_TagNS, "cpu-tag-memory",
195
+ cpu->tag_memory);
196
+ if (has_secure) {
197
+ cpu_address_space_init(cs, ARMASIdx_TagS, "cpu-tag-memory",
198
+ cpu->secure_tag_memory);
199
+ }
200
+ } else if (cpu_isar_feature(aa64_mte, cpu)) {
201
+ /*
202
+ * Since there is no tag memory, we can't meaningfully support MTE
203
+ * to its fullest. To avoid problems later, when we would come to
204
+ * use the tag memory, downgrade support to insns only.
205
+ */
206
+ cpu->isar.id_aa64pfr1 =
207
+ FIELD_DP64(cpu->isar.id_aa64pfr1, ID_AA64PFR1, MTE, 1);
208
+ }
209
+
210
cpu_address_space_init(cs, ARMASIdx_NS, "cpu-memory", cs->memory);
211
212
/* No core_count specified, default to smp_cpus. */
117
--
213
--
118
2.19.1
214
2.20.1
119
215
120
216
diff view generated by jsdifflib
1
From: Dongjiu Geng <gengdongjiu@huawei.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This patch extends the qemu-kvm state sync logic with support for
3
Look up the physical address for the given virtual address,
4
KVM_GET/SET_VCPU_EVENTS, giving access to yet missing SError exception.
4
convert that to a tag physical address, and finally return
5
And also it can support the exception state migration.
5
the host address that backs it.
6
6
7
The SError exception states include SError pending state and ESR value,
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
the kvm_put/get_vcpu_events() will be called when set or get system
8
Message-id: 20200626033144.790098-45-richard.henderson@linaro.org
9
registers. When do migration, if source machine has SError pending,
10
QEMU will do this migration regardless whether the target machine supports
11
to specify guest ESR value, because if target machine does not support that,
12
it can also inject the SError with zero ESR value.
13
14
Signed-off-by: Dongjiu Geng <gengdongjiu@huawei.com>
15
Reviewed-by: Andrew Jones <drjones@redhat.com>
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Message-id: 1538067351-23931-3-git-send-email-gengdongjiu@huawei.com
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
---
11
---
20
target/arm/cpu.h | 7 ++++++
12
target/arm/mte_helper.c | 131 ++++++++++++++++++++++++++++++++++++++++
21
target/arm/kvm_arm.h | 24 ++++++++++++++++++
13
1 file changed, 131 insertions(+)
22
target/arm/kvm.c | 60 ++++++++++++++++++++++++++++++++++++++++++++
23
target/arm/kvm32.c | 13 ++++++++++
24
target/arm/kvm64.c | 13 ++++++++++
25
target/arm/machine.c | 22 ++++++++++++++++
26
6 files changed, 139 insertions(+)
27
14
28
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
29
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/cpu.h
17
--- a/target/arm/mte_helper.c
31
+++ b/target/arm/cpu.h
18
+++ b/target/arm/mte_helper.c
32
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
19
@@ -XXX,XX +XXX,XX @@
33
*/
20
#include "cpu.h"
34
} exception;
21
#include "internals.h"
35
22
#include "exec/exec-all.h"
36
+ /* Information associated with an SError */
23
+#include "exec/ram_addr.h"
37
+ struct {
24
#include "exec/cpu_ldst.h"
38
+ uint8_t pending;
25
#include "exec/helper-proto.h"
39
+ uint8_t has_esr;
26
40
+ uint64_t esr;
27
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
41
+ } serror;
28
int ptr_size, MMUAccessType tag_access,
29
int tag_size, uintptr_t ra)
30
{
31
+#ifdef CONFIG_USER_ONLY
32
/* Tag storage not implemented. */
33
return NULL;
34
+#else
35
+ uintptr_t index;
36
+ CPUIOTLBEntry *iotlbentry;
37
+ int in_page, flags;
38
+ ram_addr_t ptr_ra;
39
+ hwaddr ptr_paddr, tag_paddr, xlat;
40
+ MemoryRegion *mr;
41
+ ARMASIdx tag_asi;
42
+ AddressSpace *tag_as;
43
+ void *host;
42
+
44
+
43
/* Thumb-2 EE state. */
45
+ /*
44
uint32_t teecr;
46
+ * Probe the first byte of the virtual address. This raises an
45
uint32_t teehbr;
47
+ * exception for inaccessible pages, and resolves the virtual address
46
diff --git a/target/arm/kvm_arm.h b/target/arm/kvm_arm.h
48
+ * into the softmmu tlb.
47
index XXXXXXX..XXXXXXX 100644
49
+ *
48
--- a/target/arm/kvm_arm.h
50
+ * When RA == 0, this is for mte_probe1. The page is expected to be
49
+++ b/target/arm/kvm_arm.h
51
+ * valid. Indicate to probe_access_flags no-fault, then assert that
50
@@ -XXX,XX +XXX,XX @@ bool write_kvmstate_to_list(ARMCPU *cpu);
52
+ * we received a valid page.
51
*/
53
+ */
52
void kvm_arm_reset_vcpu(ARMCPU *cpu);
54
+ flags = probe_access_flags(env, ptr, ptr_access, ptr_mmu_idx,
53
55
+ ra == 0, &host, ra);
54
+/**
56
+ assert(!(flags & TLB_INVALID_MASK));
55
+ * kvm_arm_init_serror_injection:
56
+ * @cs: CPUState
57
+ *
58
+ * Check whether KVM can set guest SError syndrome.
59
+ */
60
+void kvm_arm_init_serror_injection(CPUState *cs);
61
+
57
+
62
+/**
58
+ /*
63
+ * kvm_get_vcpu_events:
59
+ * Find the iotlbentry for ptr. This *must* be present in the TLB
64
+ * @cpu: ARMCPU
60
+ * because we just found the mapping.
65
+ *
61
+ * TODO: Perhaps there should be a cputlb helper that returns a
66
+ * Get VCPU related state from kvm.
62
+ * matching tlb entry + iotlb entry.
67
+ */
63
+ */
68
+int kvm_get_vcpu_events(ARMCPU *cpu);
64
+ index = tlb_index(env, ptr_mmu_idx, ptr);
65
+# ifdef CONFIG_DEBUG_TCG
66
+ {
67
+ CPUTLBEntry *entry = tlb_entry(env, ptr_mmu_idx, ptr);
68
+ target_ulong comparator = (ptr_access == MMU_DATA_LOAD
69
+ ? entry->addr_read
70
+ : tlb_addr_write(entry));
71
+ g_assert(tlb_hit(comparator, ptr));
72
+ }
73
+# endif
74
+ iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
69
+
75
+
70
+/**
76
+ /* If the virtual page MemAttr != Tagged, access unchecked. */
71
+ * kvm_put_vcpu_events:
77
+ if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
72
+ * @cpu: ARMCPU
78
+ return NULL;
73
+ *
74
+ * Put VCPU related state to kvm.
75
+ */
76
+int kvm_put_vcpu_events(ARMCPU *cpu);
77
+
78
#ifdef CONFIG_KVM
79
/**
80
* kvm_arm_create_scratch_host_vcpu:
81
diff --git a/target/arm/kvm.c b/target/arm/kvm.c
82
index XXXXXXX..XXXXXXX 100644
83
--- a/target/arm/kvm.c
84
+++ b/target/arm/kvm.c
85
@@ -XXX,XX +XXX,XX @@ const KVMCapabilityInfo kvm_arch_required_capabilities[] = {
86
};
87
88
static bool cap_has_mp_state;
89
+static bool cap_has_inject_serror_esr;
90
91
static ARMHostCPUFeatures arm_host_cpu_features;
92
93
@@ -XXX,XX +XXX,XX @@ int kvm_arm_vcpu_init(CPUState *cs)
94
return kvm_vcpu_ioctl(cs, KVM_ARM_VCPU_INIT, &init);
95
}
96
97
+void kvm_arm_init_serror_injection(CPUState *cs)
98
+{
99
+ cap_has_inject_serror_esr = kvm_check_extension(cs->kvm_state,
100
+ KVM_CAP_ARM_INJECT_SERROR_ESR);
101
+}
102
+
103
bool kvm_arm_create_scratch_host_vcpu(const uint32_t *cpus_to_try,
104
int *fdarray,
105
struct kvm_vcpu_init *init)
106
@@ -XXX,XX +XXX,XX @@ int kvm_arm_sync_mpstate_to_qemu(ARMCPU *cpu)
107
return 0;
108
}
109
110
+int kvm_put_vcpu_events(ARMCPU *cpu)
111
+{
112
+ CPUARMState *env = &cpu->env;
113
+ struct kvm_vcpu_events events;
114
+ int ret;
115
+
116
+ if (!kvm_has_vcpu_events()) {
117
+ return 0;
118
+ }
79
+ }
119
+
80
+
120
+ memset(&events, 0, sizeof(events));
81
+ /*
121
+ events.exception.serror_pending = env->serror.pending;
82
+ * If not backed by host ram, there is no tag storage: access unchecked.
122
+
83
+ * This is probably a guest os bug though, so log it.
123
+ /* Inject SError to guest with specified syndrome if host kernel
124
+ * supports it, otherwise inject SError without syndrome.
125
+ */
84
+ */
126
+ if (cap_has_inject_serror_esr) {
85
+ if (unlikely(flags & TLB_MMIO)) {
127
+ events.exception.serror_has_esr = env->serror.has_esr;
86
+ qemu_log_mask(LOG_GUEST_ERROR,
128
+ events.exception.serror_esr = env->serror.esr;
87
+ "Page @ 0x%" PRIx64 " indicates Tagged Normal memory "
88
+ "but is not backed by host ram\n", ptr);
89
+ return NULL;
129
+ }
90
+ }
130
+
91
+
131
+ ret = kvm_vcpu_ioctl(CPU(cpu), KVM_SET_VCPU_EVENTS, &events);
92
+ /*
132
+ if (ret) {
93
+ * The Normal memory access can extend to the next page. E.g. a single
133
+ error_report("failed to put vcpu events");
94
+ * 8-byte access to the last byte of a page will check only the last
95
+ * tag on the first page.
96
+ * Any page access exception has priority over tag check exception.
97
+ */
98
+ in_page = -(ptr | TARGET_PAGE_MASK);
99
+ if (unlikely(ptr_size > in_page)) {
100
+ void *ignore;
101
+ flags |= probe_access_flags(env, ptr + in_page, ptr_access,
102
+ ptr_mmu_idx, ra == 0, &ignore, ra);
103
+ assert(!(flags & TLB_INVALID_MASK));
134
+ }
104
+ }
135
+
105
+
136
+ return ret;
106
+ /* Any debug exception has priority over a tag check exception. */
137
+}
107
+ if (unlikely(flags & TLB_WATCHPOINT)) {
138
+
108
+ int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
139
+int kvm_get_vcpu_events(ARMCPU *cpu)
109
+ assert(ra != 0);
140
+{
110
+ cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
141
+ CPUARMState *env = &cpu->env;
111
+ iotlbentry->attrs, wp, ra);
142
+ struct kvm_vcpu_events events;
143
+ int ret;
144
+
145
+ if (!kvm_has_vcpu_events()) {
146
+ return 0;
147
+ }
112
+ }
148
+
113
+
149
+ memset(&events, 0, sizeof(events));
114
+ /*
150
+ ret = kvm_vcpu_ioctl(CPU(cpu), KVM_GET_VCPU_EVENTS, &events);
115
+ * Find the physical address within the normal mem space.
151
+ if (ret) {
116
+ * The memory region lookup must succeed because TLB_MMIO was
152
+ error_report("failed to get vcpu events");
117
+ * not set in the cputlb lookup above.
153
+ return ret;
118
+ */
119
+ mr = memory_region_from_host(host, &ptr_ra);
120
+ tcg_debug_assert(mr != NULL);
121
+ tcg_debug_assert(memory_region_is_ram(mr));
122
+ ptr_paddr = ptr_ra;
123
+ do {
124
+ ptr_paddr += mr->addr;
125
+ mr = mr->container;
126
+ } while (mr);
127
+
128
+ /* Convert to the physical address in tag space. */
129
+ tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
130
+
131
+ /* Look up the address in tag space. */
132
+ tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
133
+ tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
134
+ mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
135
+ tag_access == MMU_DATA_STORE,
136
+ iotlbentry->attrs);
137
+
138
+ /*
139
+ * Note that @mr will never be NULL. If there is nothing in the address
140
+ * space at @tag_paddr, the translation will return the unallocated memory
141
+ * region. For our purposes, the result must be ram.
142
+ */
143
+ if (unlikely(!memory_region_is_ram(mr))) {
144
+ /* ??? Failure is a board configuration error. */
145
+ qemu_log_mask(LOG_UNIMP,
146
+ "Tag Memory @ 0x%" HWADDR_PRIx " not found for "
147
+ "Normal Memory @ 0x%" HWADDR_PRIx "\n",
148
+ tag_paddr, ptr_paddr);
149
+ return NULL;
154
+ }
150
+ }
155
+
151
+
156
+ env->serror.pending = events.exception.serror_pending;
152
+ /*
157
+ env->serror.has_esr = events.exception.serror_has_esr;
153
+ * Ensure the tag memory is dirty on write, for migration.
158
+ env->serror.esr = events.exception.serror_esr;
154
+ * Tag memory can never contain code or display memory (vga).
159
+
155
+ */
160
+ return 0;
156
+ if (tag_access == MMU_DATA_STORE) {
161
+}
157
+ ram_addr_t tag_ra = memory_region_get_ram_addr(mr) + xlat;
162
+
158
+ cpu_physical_memory_set_dirty_flag(tag_ra, DIRTY_MEMORY_MIGRATION);
163
void kvm_arch_pre_run(CPUState *cs, struct kvm_run *run)
164
{
165
}
166
diff --git a/target/arm/kvm32.c b/target/arm/kvm32.c
167
index XXXXXXX..XXXXXXX 100644
168
--- a/target/arm/kvm32.c
169
+++ b/target/arm/kvm32.c
170
@@ -XXX,XX +XXX,XX @@ int kvm_arch_init_vcpu(CPUState *cs)
171
}
172
cpu->mp_affinity = mpidr & ARM32_AFFINITY_MASK;
173
174
+ /* Check whether userspace can specify guest syndrome value */
175
+ kvm_arm_init_serror_injection(cs);
176
+
177
return kvm_arm_init_cpreg_list(cpu);
178
}
179
180
@@ -XXX,XX +XXX,XX @@ int kvm_arch_put_registers(CPUState *cs, int level)
181
return ret;
182
}
183
184
+ ret = kvm_put_vcpu_events(cpu);
185
+ if (ret) {
186
+ return ret;
187
+ }
159
+ }
188
+
160
+
189
/* Note that we do not call write_cpustate_to_list()
161
+ return memory_region_get_ram_ptr(mr) + xlat;
190
* here, so we are only writing the tuple list back to
162
+#endif
191
* KVM. This is safe because nothing can change the
192
@@ -XXX,XX +XXX,XX @@ int kvm_arch_get_registers(CPUState *cs)
193
}
194
vfp_set_fpscr(env, fpscr);
195
196
+ ret = kvm_get_vcpu_events(cpu);
197
+ if (ret) {
198
+ return ret;
199
+ }
200
+
201
if (!write_kvmstate_to_list(cpu)) {
202
return EINVAL;
203
}
204
diff --git a/target/arm/kvm64.c b/target/arm/kvm64.c
205
index XXXXXXX..XXXXXXX 100644
206
--- a/target/arm/kvm64.c
207
+++ b/target/arm/kvm64.c
208
@@ -XXX,XX +XXX,XX @@ int kvm_arch_init_vcpu(CPUState *cs)
209
210
kvm_arm_init_debug(cs);
211
212
+ /* Check whether user space can specify guest syndrome value */
213
+ kvm_arm_init_serror_injection(cs);
214
+
215
return kvm_arm_init_cpreg_list(cpu);
216
}
163
}
217
164
218
@@ -XXX,XX +XXX,XX @@ int kvm_arch_put_registers(CPUState *cs, int level)
165
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
219
return ret;
220
}
221
222
+ ret = kvm_put_vcpu_events(cpu);
223
+ if (ret) {
224
+ return ret;
225
+ }
226
+
227
if (!write_list_to_kvmstate(cpu, level)) {
228
return EINVAL;
229
}
230
@@ -XXX,XX +XXX,XX @@ int kvm_arch_get_registers(CPUState *cs)
231
}
232
vfp_set_fpcr(env, fpr);
233
234
+ ret = kvm_get_vcpu_events(cpu);
235
+ if (ret) {
236
+ return ret;
237
+ }
238
+
239
if (!write_kvmstate_to_list(cpu)) {
240
return EINVAL;
241
}
242
diff --git a/target/arm/machine.c b/target/arm/machine.c
243
index XXXXXXX..XXXXXXX 100644
244
--- a/target/arm/machine.c
245
+++ b/target/arm/machine.c
246
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_sve = {
247
};
248
#endif /* AARCH64 */
249
250
+static bool serror_needed(void *opaque)
251
+{
252
+ ARMCPU *cpu = opaque;
253
+ CPUARMState *env = &cpu->env;
254
+
255
+ return env->serror.pending != 0;
256
+}
257
+
258
+static const VMStateDescription vmstate_serror = {
259
+ .name = "cpu/serror",
260
+ .version_id = 1,
261
+ .minimum_version_id = 1,
262
+ .needed = serror_needed,
263
+ .fields = (VMStateField[]) {
264
+ VMSTATE_UINT8(env.serror.pending, ARMCPU),
265
+ VMSTATE_UINT8(env.serror.has_esr, ARMCPU),
266
+ VMSTATE_UINT64(env.serror.esr, ARMCPU),
267
+ VMSTATE_END_OF_LIST()
268
+ }
269
+};
270
+
271
static bool m_needed(void *opaque)
272
{
273
ARMCPU *cpu = opaque;
274
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_arm_cpu = {
275
#ifdef TARGET_AARCH64
276
&vmstate_sve,
277
#endif
278
+ &vmstate_serror,
279
NULL
280
}
281
};
282
--
166
--
283
2.19.1
167
2.20.1
284
168
285
169
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We now implement all of the components of MTE, without actually
4
supporting any tagged memory. All MTE instructions will work,
5
trivially, so we can enable support.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
9
Message-id: 20200626033144.790098-46-richard.henderson@linaro.org
5
Message-id: 20181011205206.3552-6-richard.henderson@linaro.org
6
[PMM: drop change to now-deleted cpu_mode_names array]
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/translate.c | 4 ++--
12
target/arm/cpu64.c | 5 +++++
11
1 file changed, 2 insertions(+), 2 deletions(-)
13
1 file changed, 5 insertions(+)
12
14
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
14
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/translate.c
17
--- a/target/arm/cpu64.c
16
+++ b/target/arm/translate.c
18
+++ b/target/arm/cpu64.c
17
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 cpu_F0d, cpu_F1d;
19
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
18
20
19
#include "exec/gen-icount.h"
21
t = cpu->isar.id_aa64pfr1;
20
22
t = FIELD_DP64(t, ID_AA64PFR1, BT, 1);
21
-static const char *regnames[] =
23
+ /*
22
+static const char * const regnames[] =
24
+ * Begin with full support for MTE; will be downgraded to MTE=1
23
{ "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
25
+ * during realize if the board provides no tag memory.
24
"r8", "r9", "r10", "r11", "r12", "r13", "r14", "pc" };
26
+ */
25
27
+ t = FIELD_DP64(t, ID_AA64PFR1, MTE, 2);
26
@@ -XXX,XX +XXX,XX @@ static struct {
28
cpu->isar.id_aa64pfr1 = t;
27
int nregs;
29
28
int interleave;
30
t = cpu->isar.id_aa64mmfr1;
29
int spacing;
30
-} neon_ls_element_type[11] = {
31
+} const neon_ls_element_type[11] = {
32
{4, 4, 1},
33
{4, 4, 2},
34
{4, 1, 1},
35
--
31
--
36
2.19.1
32
2.20.1
37
33
38
34
diff view generated by jsdifflib