1
First ARM pullreq of the 2.10 cycle...
1
The following changes since commit 5a67d7735d4162630769ef495cf813244fc850df:
2
2
3
thanks
3
Merge remote-tracking branch 'remotes/berrange-gitlab/tags/tls-deps-pull-request' into staging (2021-07-02 08:22:39 +0100)
4
-- PMM
5
4
6
The following changes since commit 64c8ed97cceabac4fafe17fca8d88ef08183f439:
5
are available in the Git repository at:
7
6
8
Open 2.10 development tree (2017-04-20 15:42:31 +0100)
7
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20210702
9
8
10
are available in the git repository at:
9
for you to fetch changes up to 04ea4d3cfd0a21b248ece8eb7a9436a3d9898dd8:
11
10
12
git://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20170420
11
target/arm: Implement MVE shifts by register (2021-07-02 11:48:38 +0100)
13
14
for you to fetch changes up to f4e8e4edda875cab9df91dc4ae9767f7cb1f50aa:
15
16
arm: Remove workarounds for old M-profile exception return implementation (2017-04-20 17:39:17 +0100)
17
12
18
----------------------------------------------------------------
13
----------------------------------------------------------------
19
target-arm queue:
14
target-arm queue:
20
* implement M profile exception return properly
15
* more MVE instructions
21
* cadence GEM: fix multiqueue handling bugs
16
* hw/gpio/gpio_pwr: use shutdown function for reboot
22
* pxa2xx.c: QOMify a device
17
* target/arm: Check NaN mode before silencing NaN
23
* arm/kvm: Remove trailing newlines from error_report()
18
* tests: Boot and halt a Linux guest on the Raspberry Pi 2 machine
24
* stellaris: Don't hw_error() on bad register accesses
19
* hw/arm: Add basic power management to raspi.
25
* Add assertion about FSC format for syndrome registers
20
* docs/system/arm: Add quanta-gbs-bmc, quanta-q7l1-bmc
26
* Move excnames[] array into arm_log_exceptions()
27
* exynos: minor code cleanups
28
* hw/arm/boot: take Linux/arm64 TEXT_OFFSET header field into account
29
* Fix APSR writes via M profile MSR
30
21
31
----------------------------------------------------------------
22
----------------------------------------------------------------
32
Alistair Francis (5):
23
Joe Komlodi (1):
33
cadence_gem: Read the correct queue descriptor
24
target/arm: Check NaN mode before silencing NaN
34
cadence_gem: Correct the multi-queue can rx logic
35
cadence_gem: Correct the interupt logic
36
cadence_gem: Make the revision a property
37
xlnx-zynqmp: Set the Cadence GEM revision
38
25
39
Ard Biesheuvel (1):
26
Maxim Uvarov (1):
40
hw/arm/boot: take Linux/arm64 TEXT_OFFSET header field into account
27
hw/gpio/gpio_pwr: use shutdown function for reboot
41
28
42
Ishani Chugh (1):
29
Nolan Leake (1):
43
arm/kvm: Remove trailing newlines from error_report()
30
hw/arm: Add basic power management to raspi.
44
31
45
Krzysztof Kozlowski (3):
32
Patrick Venture (2):
46
hw/arm/exynos: Convert fprintf to qemu_log_mask/error_report
33
docs/system/arm: Add quanta-q7l1-bmc reference
47
hw/char/exynos4210_uart: Constify static array and few arguments
34
docs/system/arm: Add quanta-gbs-bmc reference
48
hw/misc/exynos4210_pmu: Reorder local variables for readability
49
35
50
Peter Maydell (13):
36
Peter Maydell (18):
51
target/arm: Add missing entries to excnames[] for log strings
37
target/arm: Fix MVE widening/narrowing VLDR/VSTR offset calculation
52
arm: Move excnames[] array into arm_log_exceptions()
38
target/arm: Fix bugs in MVE VRMLALDAVH, VRMLSLDAVH
53
target/arm: Add assertion about FSC format for syndrome registers
39
target/arm: Make asimd_imm_const() public
54
stellaris: Don't hw_error() on bad register accesses
40
target/arm: Use asimd_imm_const for A64 decode
55
arm: Don't implement BXJ on M-profile CPUs
41
target/arm: Use dup_const() instead of bitfield_replicate()
56
arm: Thumb shift operations should not permit interworking branches
42
target/arm: Implement MVE logical immediate insns
57
arm: Factor out "generate right kind of step exception"
43
target/arm: Implement MVE vector shift left by immediate insns
58
arm: Move gen_set_condexec() and gen_set_pc_im() up in the file
44
target/arm: Implement MVE vector shift right by immediate insns
59
arm: Move condition-failed codepath generation out of if()
45
target/arm: Implement MVE VSHLL
60
arm: Abstract out "are we singlestepping" test to utility function
46
target/arm: Implement MVE VSRI, VSLI
61
arm: Track M profile handler mode state in TB flags
47
target/arm: Implement MVE VSHRN, VRSHRN
62
arm: Implement M profile exception return properly
48
target/arm: Implement MVE saturating narrowing shifts
63
arm: Remove workarounds for old M-profile exception return implementation
49
target/arm: Implement MVE VSHLC
50
target/arm: Implement MVE VADDLV
51
target/arm: Implement MVE long shifts by immediate
52
target/arm: Implement MVE long shifts by register
53
target/arm: Implement MVE shifts by immediate
54
target/arm: Implement MVE shifts by register
64
55
65
Suramya Shah (1):
56
Philippe Mathieu-Daudé (1):
66
hw/arm: Qomify pxa2xx.c
57
tests: Boot and halt a Linux guest on the Raspberry Pi 2 machine
67
58
68
include/hw/net/cadence_gem.h | 1 +
59
docs/system/arm/aspeed.rst | 1 +
69
target/arm/cpu.h | 10 +++
60
docs/system/arm/nuvoton.rst | 5 +-
70
target/arm/internals.h | 21 -----
61
include/hw/arm/bcm2835_peripherals.h | 3 +-
71
target/arm/translate.h | 5 ++
62
include/hw/misc/bcm2835_powermgt.h | 29 ++
72
hw/arm/boot.c | 64 ++++++++++++---
63
target/arm/helper-mve.h | 108 +++++++
73
hw/arm/exynos4_boards.c | 7 +-
64
target/arm/translate.h | 41 +++
74
hw/arm/pxa2xx.c | 14 ++--
65
target/arm/mve.decode | 177 ++++++++++-
75
hw/arm/stellaris.c | 60 ++++++++------
66
target/arm/t32.decode | 71 ++++-
76
hw/arm/xlnx-zynqmp.c | 6 +-
67
hw/arm/bcm2835_peripherals.c | 13 +-
77
hw/char/exynos4210_uart.c | 8 +-
68
hw/gpio/gpio_pwr.c | 2 +-
78
hw/misc/exynos4210_pmu.c | 4 +-
69
hw/misc/bcm2835_powermgt.c | 160 ++++++++++
79
hw/net/cadence_gem.c | 45 +++++++----
70
target/arm/helper-a64.c | 12 +-
80
hw/timer/exynos4210_mct.c | 6 +-
71
target/arm/mve_helper.c | 524 +++++++++++++++++++++++++++++++--
81
hw/timer/exynos4210_pwm.c | 13 ++--
72
target/arm/translate-a64.c | 86 +-----
82
hw/timer/exynos4210_rtc.c | 19 ++---
73
target/arm/translate-mve.c | 261 +++++++++++++++-
83
target/arm/cpu.c | 43 +---------
74
target/arm/translate-neon.c | 81 -----
84
target/arm/helper.c | 19 +++++
75
target/arm/translate.c | 327 +++++++++++++++++++-
85
target/arm/kvm64.c | 4 +-
76
target/arm/vfp_helper.c | 24 +-
86
target/arm/op_helper.c | 23 ++++--
77
hw/misc/meson.build | 1 +
87
target/arm/translate.c | 181 +++++++++++++++++++++++++++++--------------
78
tests/acceptance/boot_linux_console.py | 43 +++
88
20 files changed, 341 insertions(+), 212 deletions(-)
79
20 files changed, 1760 insertions(+), 209 deletions(-)
80
create mode 100644 include/hw/misc/bcm2835_powermgt.h
81
create mode 100644 hw/misc/bcm2835_powermgt.c
89
82
diff view generated by jsdifflib
1
From: Alistair Francis <alistair.francis@xilinx.com>
1
From: Patrick Venture <venture@google.com>
2
2
3
Signed-off-by: Alistair Francis <alistair.francis@xilinx.com>
3
Adds a line-item reference to the supported quanta-q71l-bmc aspeed
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
entry.
5
Message-id: 026dbe01a1d42619eee30ce3f2079741bf04bc73.1491947224.git.alistair.francis@xilinx.com
5
6
Signed-off-by: Patrick Venture <venture@google.com>
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20210615192848.1065297-2-venture@google.com
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
hw/arm/xlnx-zynqmp.c | 6 +++++-
11
docs/system/arm/aspeed.rst | 1 +
9
1 file changed, 5 insertions(+), 1 deletion(-)
12
1 file changed, 1 insertion(+)
10
13
11
diff --git a/hw/arm/xlnx-zynqmp.c b/hw/arm/xlnx-zynqmp.c
14
diff --git a/docs/system/arm/aspeed.rst b/docs/system/arm/aspeed.rst
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/hw/arm/xlnx-zynqmp.c
16
--- a/docs/system/arm/aspeed.rst
14
+++ b/hw/arm/xlnx-zynqmp.c
17
+++ b/docs/system/arm/aspeed.rst
15
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ etc.
16
#define ARM_PHYS_TIMER_PPI 30
19
AST2400 SoC based machines :
17
#define ARM_VIRT_TIMER_PPI 27
20
18
21
- ``palmetto-bmc`` OpenPOWER Palmetto POWER8 BMC
19
+#define GEM_REVISION 0x40070106
22
+- ``quanta-q71l-bmc`` OpenBMC Quanta BMC
20
+
23
21
#define GIC_BASE_ADDR 0xf9000000
24
AST2500 SoC based machines :
22
#define GIC_DIST_ADDR 0xf9010000
25
23
#define GIC_CPU_ADDR 0xf9020000
24
@@ -XXX,XX +XXX,XX @@ static void xlnx_zynqmp_realize(DeviceState *dev, Error **errp)
25
qemu_check_nic_model(nd, TYPE_CADENCE_GEM);
26
qdev_set_nic_properties(DEVICE(&s->gem[i]), nd);
27
}
28
+ object_property_set_int(OBJECT(&s->gem[i]), GEM_REVISION, "revision",
29
+ &error_abort);
30
object_property_set_int(OBJECT(&s->gem[i]), 2, "num-priority-queues",
31
- &error_abort);
32
+ &error_abort);
33
object_property_set_bool(OBJECT(&s->gem[i]), true, "realized", &err);
34
if (err) {
35
error_propagate(errp, err);
36
--
26
--
37
2.7.4
27
2.20.1
38
28
39
29
diff view generated by jsdifflib
1
From: Alistair Francis <alistair.francis@xilinx.com>
1
From: Patrick Venture <venture@google.com>
2
2
3
Expose the Cadence GEM revision as a property.
3
Add line item reference to quanta-gbs-bmc machine.
4
4
5
Signed-off-by: Alistair Francis <alistair.francis@xilinx.com>
5
Signed-off-by: Patrick Venture <venture@google.com>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Reviewed-by: Cédric Le Goater <clg@kaod.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Message-id: 20210615192848.1065297-3-venture@google.com
8
Message-id: 541324373cf87b50f8be0439a0cb89f5028b016f.1491947224.git.alistair.francis@xilinx.com
8
[PMM: fixed underline Sphinx warning]
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
include/hw/net/cadence_gem.h | 1 +
11
docs/system/arm/nuvoton.rst | 5 +++--
12
hw/net/cadence_gem.c | 6 +++++-
12
1 file changed, 3 insertions(+), 2 deletions(-)
13
2 files changed, 6 insertions(+), 1 deletion(-)
14
13
15
diff --git a/include/hw/net/cadence_gem.h b/include/hw/net/cadence_gem.h
14
diff --git a/docs/system/arm/nuvoton.rst b/docs/system/arm/nuvoton.rst
16
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
17
--- a/include/hw/net/cadence_gem.h
16
--- a/docs/system/arm/nuvoton.rst
18
+++ b/include/hw/net/cadence_gem.h
17
+++ b/docs/system/arm/nuvoton.rst
19
@@ -XXX,XX +XXX,XX @@ typedef struct CadenceGEMState {
20
uint8_t num_priority_queues;
21
uint8_t num_type1_screeners;
22
uint8_t num_type2_screeners;
23
+ uint32_t revision;
24
25
/* GEM registers backing store */
26
uint32_t regs[CADENCE_GEM_MAXREG];
27
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/net/cadence_gem.c
30
+++ b/hw/net/cadence_gem.c
31
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@
32
#define DESC_1_RX_SOF 0x00004000
19
-Nuvoton iBMC boards (``npcm750-evb``, ``quanta-gsj``)
33
#define DESC_1_RX_EOF 0x00008000
20
-=====================================================
34
21
+Nuvoton iBMC boards (``*-bmc``, ``npcm750-evb``, ``quanta-gsj``)
35
+#define GEM_MODID_VALUE 0x00020118
22
+================================================================
36
+
23
37
static inline unsigned tx_desc_get_buffer(unsigned *desc)
24
The `Nuvoton iBMC`_ chips (NPCM7xx) are a family of ARM-based SoCs that are
38
{
25
designed to be used as Baseboard Management Controllers (BMCs) in various
39
return desc[0];
26
@@ -XXX,XX +XXX,XX @@ segment. The following machines are based on this chip :
40
@@ -XXX,XX +XXX,XX @@ static void gem_reset(DeviceState *d)
27
The NPCM730 SoC has two Cortex-A9 cores and is targeted for Data Center and
41
s->regs[GEM_TXPAUSE] = 0x0000ffff;
28
Hyperscale applications. The following machines are based on this chip :
42
s->regs[GEM_TXPARTIALSF] = 0x000003ff;
29
43
s->regs[GEM_RXPARTIALSF] = 0x000003ff;
30
+- ``quanta-gbs-bmc`` Quanta GBS server BMC
44
- s->regs[GEM_MODID] = 0x00020118;
31
- ``quanta-gsj`` Quanta GSJ server BMC
45
+ s->regs[GEM_MODID] = s->revision;
32
46
s->regs[GEM_DESCONF] = 0x02500111;
33
There are also two more SoCs, NPCM710 and NPCM705, which are single-core
47
s->regs[GEM_DESCONF2] = 0x2ab13fff;
48
s->regs[GEM_DESCONF5] = 0x002f2145;
49
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_cadence_gem = {
50
51
static Property gem_properties[] = {
52
DEFINE_NIC_PROPERTIES(CadenceGEMState, conf),
53
+ DEFINE_PROP_UINT32("revision", CadenceGEMState, revision,
54
+ GEM_MODID_VALUE),
55
DEFINE_PROP_UINT8("num-priority-queues", CadenceGEMState,
56
num_priority_queues, 1),
57
DEFINE_PROP_UINT8("num-type1-screeners", CadenceGEMState,
58
--
34
--
59
2.7.4
35
2.20.1
60
36
61
37
diff view generated by jsdifflib
1
From: Suramya Shah <shah.suramya@gmail.com>
1
From: Nolan Leake <nolan@sigbus.net>
2
2
3
Signed-off-by: Suramya Shah <shah.suramya@gmail.com>
3
This is just enough to make reboot and poweroff work. Works for
4
Message-id: 20170415180316.2694-1-shah.suramya@gmail.com
4
linux, u-boot, and the arm trusted firmware. Not tested, but should
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
work for plan9, and bare-metal/hobby OSes, since they seem to generally
6
do what linux does for reset.
7
8
The watchdog timer functionality is not yet implemented.
9
10
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/64
11
Signed-off-by: Nolan Leake <nolan@sigbus.net>
12
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Tested-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
14
Message-id: 20210625210209.1870217-1-nolan@sigbus.net
15
[PMM: tweaked commit title; fixed region size to 0x200;
16
moved header file to include/]
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
18
---
8
hw/arm/pxa2xx.c | 14 ++++++--------
19
include/hw/arm/bcm2835_peripherals.h | 3 +-
9
1 file changed, 6 insertions(+), 8 deletions(-)
20
include/hw/misc/bcm2835_powermgt.h | 29 +++++
10
21
hw/arm/bcm2835_peripherals.c | 13 ++-
11
diff --git a/hw/arm/pxa2xx.c b/hw/arm/pxa2xx.c
22
hw/misc/bcm2835_powermgt.c | 160 +++++++++++++++++++++++++++
23
hw/misc/meson.build | 1 +
24
5 files changed, 204 insertions(+), 2 deletions(-)
25
create mode 100644 include/hw/misc/bcm2835_powermgt.h
26
create mode 100644 hw/misc/bcm2835_powermgt.c
27
28
diff --git a/include/hw/arm/bcm2835_peripherals.h b/include/hw/arm/bcm2835_peripherals.h
12
index XXXXXXX..XXXXXXX 100644
29
index XXXXXXX..XXXXXXX 100644
13
--- a/hw/arm/pxa2xx.c
30
--- a/include/hw/arm/bcm2835_peripherals.h
14
+++ b/hw/arm/pxa2xx.c
31
+++ b/include/hw/arm/bcm2835_peripherals.h
15
@@ -XXX,XX +XXX,XX @@ static void pxa2xx_ssp_reset(DeviceState *d)
32
@@ -XXX,XX +XXX,XX @@
16
s->rx_start = s->rx_level = 0;
33
#include "hw/misc/bcm2835_mphi.h"
34
#include "hw/misc/bcm2835_thermal.h"
35
#include "hw/misc/bcm2835_cprman.h"
36
+#include "hw/misc/bcm2835_powermgt.h"
37
#include "hw/sd/sdhci.h"
38
#include "hw/sd/bcm2835_sdhost.h"
39
#include "hw/gpio/bcm2835_gpio.h"
40
@@ -XXX,XX +XXX,XX @@ struct BCM2835PeripheralState {
41
BCM2835MphiState mphi;
42
UnimplementedDeviceState txp;
43
UnimplementedDeviceState armtmr;
44
- UnimplementedDeviceState powermgt;
45
+ BCM2835PowerMgtState powermgt;
46
BCM2835CprmanState cprman;
47
PL011State uart0;
48
BCM2835AuxState aux;
49
diff --git a/include/hw/misc/bcm2835_powermgt.h b/include/hw/misc/bcm2835_powermgt.h
50
new file mode 100644
51
index XXXXXXX..XXXXXXX
52
--- /dev/null
53
+++ b/include/hw/misc/bcm2835_powermgt.h
54
@@ -XXX,XX +XXX,XX @@
55
+/*
56
+ * BCM2835 Power Management emulation
57
+ *
58
+ * Copyright (C) 2017 Marcin Chojnacki <marcinch7@gmail.com>
59
+ * Copyright (C) 2021 Nolan Leake <nolan@sigbus.net>
60
+ *
61
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
62
+ * See the COPYING file in the top-level directory.
63
+ */
64
+
65
+#ifndef BCM2835_POWERMGT_H
66
+#define BCM2835_POWERMGT_H
67
+
68
+#include "hw/sysbus.h"
69
+#include "qom/object.h"
70
+
71
+#define TYPE_BCM2835_POWERMGT "bcm2835-powermgt"
72
+OBJECT_DECLARE_SIMPLE_TYPE(BCM2835PowerMgtState, BCM2835_POWERMGT)
73
+
74
+struct BCM2835PowerMgtState {
75
+ SysBusDevice busdev;
76
+ MemoryRegion iomem;
77
+
78
+ uint32_t rstc;
79
+ uint32_t rsts;
80
+ uint32_t wdog;
81
+};
82
+
83
+#endif
84
diff --git a/hw/arm/bcm2835_peripherals.c b/hw/arm/bcm2835_peripherals.c
85
index XXXXXXX..XXXXXXX 100644
86
--- a/hw/arm/bcm2835_peripherals.c
87
+++ b/hw/arm/bcm2835_peripherals.c
88
@@ -XXX,XX +XXX,XX @@ static void bcm2835_peripherals_init(Object *obj)
89
90
object_property_add_const_link(OBJECT(&s->dwc2), "dma-mr",
91
OBJECT(&s->gpu_bus_mr));
92
+
93
+ /* Power Management */
94
+ object_initialize_child(obj, "powermgt", &s->powermgt,
95
+ TYPE_BCM2835_POWERMGT);
17
}
96
}
18
97
19
-static int pxa2xx_ssp_init(SysBusDevice *sbd)
98
static void bcm2835_peripherals_realize(DeviceState *dev, Error **errp)
20
+static void pxa2xx_ssp_init(Object *obj)
99
@@ -XXX,XX +XXX,XX @@ static void bcm2835_peripherals_realize(DeviceState *dev, Error **errp)
21
{
100
qdev_get_gpio_in_named(DEVICE(&s->ic), BCM2835_IC_GPU_IRQ,
22
- DeviceState *dev = DEVICE(sbd);
101
INTERRUPT_USB));
23
- PXA2xxSSPState *s = PXA2XX_SSP(dev);
102
24
-
103
+ /* Power Management */
25
+ DeviceState *dev = DEVICE(obj);
104
+ if (!sysbus_realize(SYS_BUS_DEVICE(&s->powermgt), errp)) {
26
+ PXA2xxSSPState *s = PXA2XX_SSP(obj);
105
+ return;
27
+ SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
106
+ }
28
sysbus_init_irq(sbd, &s->irq);
107
+
29
108
+ memory_region_add_subregion(&s->peri_mr, PM_OFFSET,
30
- memory_region_init_io(&s->iomem, OBJECT(s), &pxa2xx_ssp_ops, s,
109
+ sysbus_mmio_get_region(SYS_BUS_DEVICE(&s->powermgt), 0));
31
+ memory_region_init_io(&s->iomem, obj, &pxa2xx_ssp_ops, s,
110
+
32
"pxa2xx-ssp", 0x1000);
111
create_unimp(s, &s->txp, "bcm2835-txp", TXP_OFFSET, 0x1000);
33
sysbus_init_mmio(sbd, &s->iomem);
112
create_unimp(s, &s->armtmr, "bcm2835-sp804", ARMCTRL_TIMER0_1_OFFSET, 0x40);
34
113
- create_unimp(s, &s->powermgt, "bcm2835-powermgt", PM_OFFSET, 0x114);
35
s->bus = ssi_create_bus(dev, "ssi");
114
create_unimp(s, &s->i2s, "bcm2835-i2s", I2S_OFFSET, 0x100);
36
- return 0;
115
create_unimp(s, &s->smi, "bcm2835-smi", SMI_OFFSET, 0x100);
37
}
116
create_unimp(s, &s->spi[0], "bcm2835-spi0", SPI0_OFFSET, 0x20);
38
117
diff --git a/hw/misc/bcm2835_powermgt.c b/hw/misc/bcm2835_powermgt.c
39
/* Real-Time Clock */
118
new file mode 100644
40
@@ -XXX,XX +XXX,XX @@ PXA2xxState *pxa255_init(MemoryRegion *address_space, unsigned int sdram_size)
119
index XXXXXXX..XXXXXXX
41
120
--- /dev/null
42
static void pxa2xx_ssp_class_init(ObjectClass *klass, void *data)
121
+++ b/hw/misc/bcm2835_powermgt.c
43
{
122
@@ -XXX,XX +XXX,XX @@
44
- SysBusDeviceClass *sdc = SYS_BUS_DEVICE_CLASS(klass);
123
+/*
45
DeviceClass *dc = DEVICE_CLASS(klass);
124
+ * BCM2835 Power Management emulation
46
125
+ *
47
- sdc->init = pxa2xx_ssp_init;
126
+ * Copyright (C) 2017 Marcin Chojnacki <marcinch7@gmail.com>
48
dc->reset = pxa2xx_ssp_reset;
127
+ * Copyright (C) 2021 Nolan Leake <nolan@sigbus.net>
49
dc->vmsd = &vmstate_pxa2xx_ssp;
128
+ *
50
}
129
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
51
@@ -XXX,XX +XXX,XX @@ static const TypeInfo pxa2xx_ssp_info = {
130
+ * See the COPYING file in the top-level directory.
52
.name = TYPE_PXA2XX_SSP,
131
+ */
53
.parent = TYPE_SYS_BUS_DEVICE,
132
+
54
.instance_size = sizeof(PXA2xxSSPState),
133
+#include "qemu/osdep.h"
55
+ .instance_init = pxa2xx_ssp_init,
134
+#include "qemu/log.h"
56
.class_init = pxa2xx_ssp_class_init,
135
+#include "qemu/module.h"
57
};
136
+#include "hw/misc/bcm2835_powermgt.h"
58
137
+#include "migration/vmstate.h"
138
+#include "sysemu/runstate.h"
139
+
140
+#define PASSWORD 0x5a000000
141
+#define PASSWORD_MASK 0xff000000
142
+
143
+#define R_RSTC 0x1c
144
+#define V_RSTC_RESET 0x20
145
+#define R_RSTS 0x20
146
+#define V_RSTS_POWEROFF 0x555 /* Linux uses partition 63 to indicate halt. */
147
+#define R_WDOG 0x24
148
+
149
+static uint64_t bcm2835_powermgt_read(void *opaque, hwaddr offset,
150
+ unsigned size)
151
+{
152
+ BCM2835PowerMgtState *s = (BCM2835PowerMgtState *)opaque;
153
+ uint32_t res = 0;
154
+
155
+ switch (offset) {
156
+ case R_RSTC:
157
+ res = s->rstc;
158
+ break;
159
+ case R_RSTS:
160
+ res = s->rsts;
161
+ break;
162
+ case R_WDOG:
163
+ res = s->wdog;
164
+ break;
165
+
166
+ default:
167
+ qemu_log_mask(LOG_UNIMP,
168
+ "bcm2835_powermgt_read: Unknown offset 0x%08"HWADDR_PRIx
169
+ "\n", offset);
170
+ res = 0;
171
+ break;
172
+ }
173
+
174
+ return res;
175
+}
176
+
177
+static void bcm2835_powermgt_write(void *opaque, hwaddr offset,
178
+ uint64_t value, unsigned size)
179
+{
180
+ BCM2835PowerMgtState *s = (BCM2835PowerMgtState *)opaque;
181
+
182
+ if ((value & PASSWORD_MASK) != PASSWORD) {
183
+ qemu_log_mask(LOG_GUEST_ERROR,
184
+ "bcm2835_powermgt_write: Bad password 0x%"PRIx64
185
+ " at offset 0x%08"HWADDR_PRIx"\n",
186
+ value, offset);
187
+ return;
188
+ }
189
+
190
+ value = value & ~PASSWORD_MASK;
191
+
192
+ switch (offset) {
193
+ case R_RSTC:
194
+ s->rstc = value;
195
+ if (value & V_RSTC_RESET) {
196
+ if ((s->rsts & 0xfff) == V_RSTS_POWEROFF) {
197
+ qemu_system_shutdown_request(SHUTDOWN_CAUSE_GUEST_SHUTDOWN);
198
+ } else {
199
+ qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_RESET);
200
+ }
201
+ }
202
+ break;
203
+ case R_RSTS:
204
+ qemu_log_mask(LOG_UNIMP,
205
+ "bcm2835_powermgt_write: RSTS\n");
206
+ s->rsts = value;
207
+ break;
208
+ case R_WDOG:
209
+ qemu_log_mask(LOG_UNIMP,
210
+ "bcm2835_powermgt_write: WDOG\n");
211
+ s->wdog = value;
212
+ break;
213
+
214
+ default:
215
+ qemu_log_mask(LOG_UNIMP,
216
+ "bcm2835_powermgt_write: Unknown offset 0x%08"HWADDR_PRIx
217
+ "\n", offset);
218
+ break;
219
+ }
220
+}
221
+
222
+static const MemoryRegionOps bcm2835_powermgt_ops = {
223
+ .read = bcm2835_powermgt_read,
224
+ .write = bcm2835_powermgt_write,
225
+ .endianness = DEVICE_NATIVE_ENDIAN,
226
+ .impl.min_access_size = 4,
227
+ .impl.max_access_size = 4,
228
+};
229
+
230
+static const VMStateDescription vmstate_bcm2835_powermgt = {
231
+ .name = TYPE_BCM2835_POWERMGT,
232
+ .version_id = 1,
233
+ .minimum_version_id = 1,
234
+ .fields = (VMStateField[]) {
235
+ VMSTATE_UINT32(rstc, BCM2835PowerMgtState),
236
+ VMSTATE_UINT32(rsts, BCM2835PowerMgtState),
237
+ VMSTATE_UINT32(wdog, BCM2835PowerMgtState),
238
+ VMSTATE_END_OF_LIST()
239
+ }
240
+};
241
+
242
+static void bcm2835_powermgt_init(Object *obj)
243
+{
244
+ BCM2835PowerMgtState *s = BCM2835_POWERMGT(obj);
245
+
246
+ memory_region_init_io(&s->iomem, obj, &bcm2835_powermgt_ops, s,
247
+ TYPE_BCM2835_POWERMGT, 0x200);
248
+ sysbus_init_mmio(SYS_BUS_DEVICE(s), &s->iomem);
249
+}
250
+
251
+static void bcm2835_powermgt_reset(DeviceState *dev)
252
+{
253
+ BCM2835PowerMgtState *s = BCM2835_POWERMGT(dev);
254
+
255
+ /* https://elinux.org/BCM2835_registers#PM */
256
+ s->rstc = 0x00000102;
257
+ s->rsts = 0x00001000;
258
+ s->wdog = 0x00000000;
259
+}
260
+
261
+static void bcm2835_powermgt_class_init(ObjectClass *klass, void *data)
262
+{
263
+ DeviceClass *dc = DEVICE_CLASS(klass);
264
+
265
+ dc->reset = bcm2835_powermgt_reset;
266
+ dc->vmsd = &vmstate_bcm2835_powermgt;
267
+}
268
+
269
+static TypeInfo bcm2835_powermgt_info = {
270
+ .name = TYPE_BCM2835_POWERMGT,
271
+ .parent = TYPE_SYS_BUS_DEVICE,
272
+ .instance_size = sizeof(BCM2835PowerMgtState),
273
+ .class_init = bcm2835_powermgt_class_init,
274
+ .instance_init = bcm2835_powermgt_init,
275
+};
276
+
277
+static void bcm2835_powermgt_register_types(void)
278
+{
279
+ type_register_static(&bcm2835_powermgt_info);
280
+}
281
+
282
+type_init(bcm2835_powermgt_register_types)
283
diff --git a/hw/misc/meson.build b/hw/misc/meson.build
284
index XXXXXXX..XXXXXXX 100644
285
--- a/hw/misc/meson.build
286
+++ b/hw/misc/meson.build
287
@@ -XXX,XX +XXX,XX @@ softmmu_ss.add(when: 'CONFIG_RASPI', if_true: files(
288
'bcm2835_rng.c',
289
'bcm2835_thermal.c',
290
'bcm2835_cprman.c',
291
+ 'bcm2835_powermgt.c',
292
))
293
softmmu_ss.add(when: 'CONFIG_SLAVIO', if_true: files('slavio_misc.c'))
294
softmmu_ss.add(when: 'CONFIG_ZYNQ', if_true: files('zynq_slcr.c', 'zynq-xadc.c'))
59
--
295
--
60
2.7.4
296
2.20.1
61
297
62
298
diff view generated by jsdifflib
1
From: Alistair Francis <alistair.francis@xilinx.com>
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
2
3
Read the correct descriptor instead of hardcoding the first (q=0).
3
Add a test booting and quickly shutdown a raspi2 machine,
4
to test the power management model:
4
5
5
Signed-off-by: Alistair Francis <alistair.francis@xilinx.com>
6
(1/1) tests/acceptance/boot_linux_console.py:BootLinuxConsole.test_arm_raspi2_initrd:
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
console: [ 0.000000] Booting Linux on physical CPU 0xf00
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
console: [ 0.000000] Linux version 4.14.98-v7+ (dom@dom-XPS-13-9370) (gcc version 4.9.3 (crosstool-NG crosstool-ng-1.22.0-88-g8460611)) #1200 SMP Tue Feb 12 20:27:48 GMT 2019
8
Message-id: 988b183dcf951856d8b3379f7e911ec95233bbf4.1491947224.git.alistair.francis@xilinx.com
9
console: [ 0.000000] CPU: ARMv7 Processor [410fc075] revision 5 (ARMv7), cr=10c5387d
10
console: [ 0.000000] CPU: div instructions available: patching division code
11
console: [ 0.000000] CPU: PIPT / VIPT nonaliasing data cache, VIPT aliasing instruction cache
12
console: [ 0.000000] OF: fdt: Machine model: Raspberry Pi 2 Model B
13
...
14
console: Boot successful.
15
console: cat /proc/cpuinfo
16
console: / # cat /proc/cpuinfo
17
...
18
console: processor : 3
19
console: model name : ARMv7 Processor rev 5 (v7l)
20
console: BogoMIPS : 125.00
21
console: Features : half thumb fastmult vfp edsp neon vfpv3 tls vfpv4 idiva idivt vfpd32 lpae evtstrm
22
console: CPU implementer : 0x41
23
console: CPU architecture: 7
24
console: CPU variant : 0x0
25
console: CPU part : 0xc07
26
console: CPU revision : 5
27
console: Hardware : BCM2835
28
console: Revision : 0000
29
console: Serial : 0000000000000000
30
console: cat /proc/iomem
31
console: / # cat /proc/iomem
32
console: 00000000-3bffffff : System RAM
33
console: 00008000-00afffff : Kernel code
34
console: 00c00000-00d468ef : Kernel data
35
console: 3f006000-3f006fff : dwc_otg
36
console: 3f007000-3f007eff : /soc/dma@7e007000
37
console: 3f00b880-3f00b8bf : /soc/mailbox@7e00b880
38
console: 3f100000-3f100027 : /soc/watchdog@7e100000
39
console: 3f101000-3f102fff : /soc/cprman@7e101000
40
console: 3f200000-3f2000b3 : /soc/gpio@7e200000
41
PASS (24.59 s)
42
RESULTS : PASS 1 | ERROR 0 | FAIL 0 | SKIP 0 | WARN 0 | INTERRUPT 0 | CANCEL 0
43
JOB TIME : 25.02 s
44
45
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
46
Reviewed-by: Wainer dos Santos Moschetta <wainersm@redhat.com>
47
Message-id: 20210531113837.1689775-1-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
48
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
49
---
11
hw/net/cadence_gem.c | 4 ++--
50
tests/acceptance/boot_linux_console.py | 43 ++++++++++++++++++++++++++
12
1 file changed, 2 insertions(+), 2 deletions(-)
51
1 file changed, 43 insertions(+)
13
52
14
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
53
diff --git a/tests/acceptance/boot_linux_console.py b/tests/acceptance/boot_linux_console.py
15
index XXXXXXX..XXXXXXX 100644
54
index XXXXXXX..XXXXXXX 100644
16
--- a/hw/net/cadence_gem.c
55
--- a/tests/acceptance/boot_linux_console.py
17
+++ b/hw/net/cadence_gem.c
56
+++ b/tests/acceptance/boot_linux_console.py
18
@@ -XXX,XX +XXX,XX @@ static void gem_get_rx_desc(CadenceGEMState *s, int q)
57
@@ -XXX,XX +XXX,XX @@
19
{
58
from avocado import skip
20
DB_PRINT("read descriptor 0x%x\n", (unsigned)s->rx_desc_addr[q]);
59
from avocado import skipUnless
21
/* read current descriptor */
60
from avocado_qemu import Test
22
- cpu_physical_memory_read(s->rx_desc_addr[0],
61
+from avocado_qemu import exec_command
23
- (uint8_t *)s->rx_desc[0], sizeof(s->rx_desc[0]));
62
from avocado_qemu import exec_command_and_wait_for_pattern
24
+ cpu_physical_memory_read(s->rx_desc_addr[q],
63
from avocado_qemu import interrupt_interactive_console_until_pattern
25
+ (uint8_t *)s->rx_desc[q], sizeof(s->rx_desc[q]));
64
from avocado_qemu import wait_for_console_pattern
26
65
@@ -XXX,XX +XXX,XX @@ def test_arm_raspi2_uart0(self):
27
/* Descriptor owned by software ? */
66
"""
28
if (rx_desc_get_ownership(s->rx_desc[q]) == 1) {
67
self.do_test_arm_raspi2(0)
68
69
+ def test_arm_raspi2_initrd(self):
70
+ """
71
+ :avocado: tags=arch:arm
72
+ :avocado: tags=machine:raspi2
73
+ """
74
+ deb_url = ('http://archive.raspberrypi.org/debian/'
75
+ 'pool/main/r/raspberrypi-firmware/'
76
+ 'raspberrypi-kernel_1.20190215-1_armhf.deb')
77
+ deb_hash = 'cd284220b32128c5084037553db3c482426f3972'
78
+ deb_path = self.fetch_asset(deb_url, asset_hash=deb_hash)
79
+ kernel_path = self.extract_from_deb(deb_path, '/boot/kernel7.img')
80
+ dtb_path = self.extract_from_deb(deb_path, '/boot/bcm2709-rpi-2-b.dtb')
81
+
82
+ initrd_url = ('https://github.com/groeck/linux-build-test/raw/'
83
+ '2eb0a73b5d5a28df3170c546ddaaa9757e1e0848/rootfs/'
84
+ 'arm/rootfs-armv7a.cpio.gz')
85
+ initrd_hash = '604b2e45cdf35045846b8bbfbf2129b1891bdc9c'
86
+ initrd_path_gz = self.fetch_asset(initrd_url, asset_hash=initrd_hash)
87
+ initrd_path = os.path.join(self.workdir, 'rootfs.cpio')
88
+ archive.gzip_uncompress(initrd_path_gz, initrd_path)
89
+
90
+ self.vm.set_console()
91
+ kernel_command_line = (self.KERNEL_COMMON_COMMAND_LINE +
92
+ 'earlycon=pl011,0x3f201000 console=ttyAMA0 '
93
+ 'panic=-1 noreboot ' +
94
+ 'dwc_otg.fiq_fsm_enable=0')
95
+ self.vm.add_args('-kernel', kernel_path,
96
+ '-dtb', dtb_path,
97
+ '-initrd', initrd_path,
98
+ '-append', kernel_command_line,
99
+ '-no-reboot')
100
+ self.vm.launch()
101
+ self.wait_for_console_pattern('Boot successful.')
102
+
103
+ exec_command_and_wait_for_pattern(self, 'cat /proc/cpuinfo',
104
+ 'BCM2835')
105
+ exec_command_and_wait_for_pattern(self, 'cat /proc/iomem',
106
+ '/soc/cprman@7e101000')
107
+ exec_command(self, 'halt')
108
+ # Wait for VM to shut down gracefully
109
+ self.vm.wait()
110
+
111
def test_arm_exynos4210_initrd(self):
112
"""
113
:avocado: tags=arch:arm
29
--
114
--
30
2.7.4
115
2.20.1
31
116
32
117
diff view generated by jsdifflib
1
From: Krzysztof Kozlowski <krzk@kernel.org>
1
From: Joe Komlodi <joe.komlodi@xilinx.com>
2
2
3
Short declaration of 'i' was in the middle of declarations with
3
If the CPU is running in default NaN mode (FPCR.DN == 1) and we execute
4
assignments. Make it a little bit more readable. Additionally switch
4
FRSQRTE, FRECPE, or FRECPX with a signaling NaN, parts_silence_nan_frac() will
5
from "unsigned" to "unsigned int" as this pattern is more widely used.
5
assert due to fpst->default_nan_mode being set.
6
No functional change.
7
6
8
Signed-off-by: Krzysztof Kozlowski <krzk@kernel.org>
7
To avoid this, we check to see what NaN mode we're running in before we call
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
floatxx_silence_nan().
10
Message-id: 20170313184750.429-4-krzk@kernel.org
9
10
Signed-off-by: Joe Komlodi <joe.komlodi@xilinx.com>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 1624662174-175828-2-git-send-email-joe.komlodi@xilinx.com
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
15
---
14
hw/misc/exynos4210_pmu.c | 4 ++--
16
target/arm/helper-a64.c | 12 +++++++++---
15
1 file changed, 2 insertions(+), 2 deletions(-)
17
target/arm/vfp_helper.c | 24 ++++++++++++++++++------
18
2 files changed, 27 insertions(+), 9 deletions(-)
16
19
17
diff --git a/hw/misc/exynos4210_pmu.c b/hw/misc/exynos4210_pmu.c
20
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
18
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
19
--- a/hw/misc/exynos4210_pmu.c
22
--- a/target/arm/helper-a64.c
20
+++ b/hw/misc/exynos4210_pmu.c
23
+++ b/target/arm/helper-a64.c
21
@@ -XXX,XX +XXX,XX @@ static uint64_t exynos4210_pmu_read(void *opaque, hwaddr offset,
24
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(frecpx_f16)(uint32_t a, void *fpstp)
22
unsigned size)
25
float16 nan = a;
23
{
26
if (float16_is_signaling_nan(a, fpst)) {
24
Exynos4210PmuState *s = (Exynos4210PmuState *)opaque;
27
float_raise(float_flag_invalid, fpst);
25
- unsigned i;
28
- nan = float16_silence_nan(a, fpst);
26
const Exynos4210PmuReg *reg_p = exynos4210_pmu_regs;
29
+ if (!fpst->default_nan_mode) {
27
+ unsigned int i;
30
+ nan = float16_silence_nan(a, fpst);
28
31
+ }
29
for (i = 0; i < PMU_NUM_OF_REGISTERS; i++) {
32
}
30
if (reg_p->offset == offset) {
33
if (fpst->default_nan_mode) {
31
@@ -XXX,XX +XXX,XX @@ static void exynos4210_pmu_write(void *opaque, hwaddr offset,
34
nan = float16_default_nan(fpst);
32
uint64_t val, unsigned size)
35
@@ -XXX,XX +XXX,XX @@ float32 HELPER(frecpx_f32)(float32 a, void *fpstp)
33
{
36
float32 nan = a;
34
Exynos4210PmuState *s = (Exynos4210PmuState *)opaque;
37
if (float32_is_signaling_nan(a, fpst)) {
35
- unsigned i;
38
float_raise(float_flag_invalid, fpst);
36
const Exynos4210PmuReg *reg_p = exynos4210_pmu_regs;
39
- nan = float32_silence_nan(a, fpst);
37
+ unsigned int i;
40
+ if (!fpst->default_nan_mode) {
38
41
+ nan = float32_silence_nan(a, fpst);
39
for (i = 0; i < PMU_NUM_OF_REGISTERS; i++) {
42
+ }
40
if (reg_p->offset == offset) {
43
}
44
if (fpst->default_nan_mode) {
45
nan = float32_default_nan(fpst);
46
@@ -XXX,XX +XXX,XX @@ float64 HELPER(frecpx_f64)(float64 a, void *fpstp)
47
float64 nan = a;
48
if (float64_is_signaling_nan(a, fpst)) {
49
float_raise(float_flag_invalid, fpst);
50
- nan = float64_silence_nan(a, fpst);
51
+ if (!fpst->default_nan_mode) {
52
+ nan = float64_silence_nan(a, fpst);
53
+ }
54
}
55
if (fpst->default_nan_mode) {
56
nan = float64_default_nan(fpst);
57
diff --git a/target/arm/vfp_helper.c b/target/arm/vfp_helper.c
58
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/vfp_helper.c
60
+++ b/target/arm/vfp_helper.c
61
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(recpe_f16)(uint32_t input, void *fpstp)
62
float16 nan = f16;
63
if (float16_is_signaling_nan(f16, fpst)) {
64
float_raise(float_flag_invalid, fpst);
65
- nan = float16_silence_nan(f16, fpst);
66
+ if (!fpst->default_nan_mode) {
67
+ nan = float16_silence_nan(f16, fpst);
68
+ }
69
}
70
if (fpst->default_nan_mode) {
71
nan = float16_default_nan(fpst);
72
@@ -XXX,XX +XXX,XX @@ float32 HELPER(recpe_f32)(float32 input, void *fpstp)
73
float32 nan = f32;
74
if (float32_is_signaling_nan(f32, fpst)) {
75
float_raise(float_flag_invalid, fpst);
76
- nan = float32_silence_nan(f32, fpst);
77
+ if (!fpst->default_nan_mode) {
78
+ nan = float32_silence_nan(f32, fpst);
79
+ }
80
}
81
if (fpst->default_nan_mode) {
82
nan = float32_default_nan(fpst);
83
@@ -XXX,XX +XXX,XX @@ float64 HELPER(recpe_f64)(float64 input, void *fpstp)
84
float64 nan = f64;
85
if (float64_is_signaling_nan(f64, fpst)) {
86
float_raise(float_flag_invalid, fpst);
87
- nan = float64_silence_nan(f64, fpst);
88
+ if (!fpst->default_nan_mode) {
89
+ nan = float64_silence_nan(f64, fpst);
90
+ }
91
}
92
if (fpst->default_nan_mode) {
93
nan = float64_default_nan(fpst);
94
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(rsqrte_f16)(uint32_t input, void *fpstp)
95
float16 nan = f16;
96
if (float16_is_signaling_nan(f16, s)) {
97
float_raise(float_flag_invalid, s);
98
- nan = float16_silence_nan(f16, s);
99
+ if (!s->default_nan_mode) {
100
+ nan = float16_silence_nan(f16, fpstp);
101
+ }
102
}
103
if (s->default_nan_mode) {
104
nan = float16_default_nan(s);
105
@@ -XXX,XX +XXX,XX @@ float32 HELPER(rsqrte_f32)(float32 input, void *fpstp)
106
float32 nan = f32;
107
if (float32_is_signaling_nan(f32, s)) {
108
float_raise(float_flag_invalid, s);
109
- nan = float32_silence_nan(f32, s);
110
+ if (!s->default_nan_mode) {
111
+ nan = float32_silence_nan(f32, fpstp);
112
+ }
113
}
114
if (s->default_nan_mode) {
115
nan = float32_default_nan(s);
116
@@ -XXX,XX +XXX,XX @@ float64 HELPER(rsqrte_f64)(float64 input, void *fpstp)
117
float64 nan = f64;
118
if (float64_is_signaling_nan(f64, s)) {
119
float_raise(float_flag_invalid, s);
120
- nan = float64_silence_nan(f64, s);
121
+ if (!s->default_nan_mode) {
122
+ nan = float64_silence_nan(f64, fpstp);
123
+ }
124
}
125
if (s->default_nan_mode) {
126
nan = float64_default_nan(s);
41
--
127
--
42
2.7.4
128
2.20.1
43
129
44
130
diff view generated by jsdifflib
1
From: Ishani Chugh <chugh.ishani@research.iiit.ac.in>
1
From: Maxim Uvarov <maxim.uvarov@linaro.org>
2
2
3
Signed-off-by: Ishani Chugh <chugh.ishani@research.iiit.ac.in>
3
qemu has 2 type of functions: shutdown and reboot. Shutdown
4
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
4
function has to be used for machine shutdown. Otherwise we cause
5
Message-id: 1491629987-6826-1-git-send-email-chugh.ishani@research.iiit.ac.in
5
a reset with a bogus "cause" value, when we intended a shutdown.
6
7
Signed-off-by: Maxim Uvarov <maxim.uvarov@linaro.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Message-id: 20210625111842.3790-3-maxim.uvarov@linaro.org
10
[PMM: tweaked commit message]
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
12
---
8
target/arm/kvm64.c | 4 ++--
13
hw/gpio/gpio_pwr.c | 2 +-
9
1 file changed, 2 insertions(+), 2 deletions(-)
14
1 file changed, 1 insertion(+), 1 deletion(-)
10
15
11
diff --git a/target/arm/kvm64.c b/target/arm/kvm64.c
16
diff --git a/hw/gpio/gpio_pwr.c b/hw/gpio/gpio_pwr.c
12
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/kvm64.c
18
--- a/hw/gpio/gpio_pwr.c
14
+++ b/target/arm/kvm64.c
19
+++ b/hw/gpio/gpio_pwr.c
15
@@ -XXX,XX +XXX,XX @@ bool kvm_arm_handle_debug(CPUState *cs, struct kvm_debug_exit_arch *debug_exit)
20
@@ -XXX,XX +XXX,XX @@ static void gpio_pwr_reset(void *opaque, int n, int level)
16
* single step at this point so something has gone wrong.
21
static void gpio_pwr_shutdown(void *opaque, int n, int level)
17
*/
22
{
18
error_report("%s: guest single-step while debugging unsupported"
23
if (level) {
19
- " (%"PRIx64", %"PRIx32")\n",
24
- qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_SHUTDOWN);
20
+ " (%"PRIx64", %"PRIx32")",
25
+ qemu_system_shutdown_request(SHUTDOWN_CAUSE_GUEST_SHUTDOWN);
21
__func__, env->pc, debug_exit->hsr);
22
return false;
23
}
24
@@ -XXX,XX +XXX,XX @@ bool kvm_arm_handle_debug(CPUState *cs, struct kvm_debug_exit_arch *debug_exit)
25
break;
26
}
26
}
27
default:
27
}
28
- error_report("%s: unhandled debug exit (%"PRIx32", %"PRIx64")\n",
29
+ error_report("%s: unhandled debug exit (%"PRIx32", %"PRIx64")",
30
__func__, debug_exit->hsr, env->pc);
31
}
32
28
33
--
29
--
34
2.7.4
30
2.20.1
35
31
36
32
diff view generated by jsdifflib
1
From: Krzysztof Kozlowski <krzk@kernel.org>
1
In do_ldst(), the calculation of the offset needs to be based on the
2
size of the memory access, not the size of the elements in the
3
vector. This meant we were getting it wrong for the widening and
4
narrowing variants of the various VLDR and VSTR insns.
2
5
3
The static array exynos4210_uart_regs with register values is not
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
modified so it can be made const.
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210628135835.6690-2-peter.maydell@linaro.org
9
---
10
target/arm/translate-mve.c | 17 +++++++++--------
11
1 file changed, 9 insertions(+), 8 deletions(-)
5
12
6
Few other functions accept driver or uart state as an argument but they
13
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
7
do not change it and do not cast it so this can be made const for code
8
safeness.
9
10
Signed-off-by: Krzysztof Kozlowski <krzk@kernel.org>
11
Message-id: 20170313184750.429-3-krzk@kernel.org
12
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
hw/char/exynos4210_uart.c | 8 ++++----
16
1 file changed, 4 insertions(+), 4 deletions(-)
17
18
diff --git a/hw/char/exynos4210_uart.c b/hw/char/exynos4210_uart.c
19
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
20
--- a/hw/char/exynos4210_uart.c
15
--- a/target/arm/translate-mve.c
21
+++ b/hw/char/exynos4210_uart.c
16
+++ b/target/arm/translate-mve.c
22
@@ -XXX,XX +XXX,XX @@ typedef struct Exynos4210UartReg {
17
@@ -XXX,XX +XXX,XX @@ static bool mve_skip_first_beat(DisasContext *s)
23
uint32_t reset_value;
18
}
24
} Exynos4210UartReg;
25
26
-static Exynos4210UartReg exynos4210_uart_regs[] = {
27
+static const Exynos4210UartReg exynos4210_uart_regs[] = {
28
{"ULCON", ULCON, 0x00000000},
29
{"UCON", UCON, 0x00003000},
30
{"UFCON", UFCON, 0x00000000},
31
@@ -XXX,XX +XXX,XX @@ static uint8_t fifo_retrieve(Exynos4210UartFIFO *q)
32
return ret;
33
}
19
}
34
20
35
-static int fifo_elements_number(Exynos4210UartFIFO *q)
21
-static bool do_ldst(DisasContext *s, arg_VLDR_VSTR *a, MVEGenLdStFn *fn)
36
+static int fifo_elements_number(const Exynos4210UartFIFO *q)
22
+static bool do_ldst(DisasContext *s, arg_VLDR_VSTR *a, MVEGenLdStFn *fn,
23
+ unsigned msize)
37
{
24
{
38
if (q->sp < q->rp) {
25
TCGv_i32 addr;
39
return q->size - q->rp + q->sp;
26
uint32_t offset;
40
@@ -XXX,XX +XXX,XX @@ static int fifo_elements_number(Exynos4210UartFIFO *q)
27
@@ -XXX,XX +XXX,XX @@ static bool do_ldst(DisasContext *s, arg_VLDR_VSTR *a, MVEGenLdStFn *fn)
41
return q->sp - q->rp;
28
return true;
29
}
30
31
- offset = a->imm << a->size;
32
+ offset = a->imm << msize;
33
if (!a->a) {
34
offset = -offset;
35
}
36
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDR_VSTR(DisasContext *s, arg_VLDR_VSTR *a)
37
{ gen_helper_mve_vstrw, gen_helper_mve_vldrw },
38
{ NULL, NULL }
39
};
40
- return do_ldst(s, a, ldstfns[a->size][a->l]);
41
+ return do_ldst(s, a, ldstfns[a->size][a->l], a->size);
42
}
42
}
43
43
44
-static int fifo_empty_elements_number(Exynos4210UartFIFO *q)
44
-#define DO_VLDST_WIDE_NARROW(OP, SLD, ULD, ST) \
45
+static int fifo_empty_elements_number(const Exynos4210UartFIFO *q)
45
+#define DO_VLDST_WIDE_NARROW(OP, SLD, ULD, ST, MSIZE) \
46
static bool trans_##OP(DisasContext *s, arg_VLDR_VSTR *a) \
47
{ \
48
static MVEGenLdStFn * const ldstfns[2][2] = { \
49
{ gen_helper_mve_##ST, gen_helper_mve_##SLD }, \
50
{ NULL, gen_helper_mve_##ULD }, \
51
}; \
52
- return do_ldst(s, a, ldstfns[a->u][a->l]); \
53
+ return do_ldst(s, a, ldstfns[a->u][a->l], MSIZE); \
54
}
55
56
-DO_VLDST_WIDE_NARROW(VLDSTB_H, vldrb_sh, vldrb_uh, vstrb_h)
57
-DO_VLDST_WIDE_NARROW(VLDSTB_W, vldrb_sw, vldrb_uw, vstrb_w)
58
-DO_VLDST_WIDE_NARROW(VLDSTH_W, vldrh_sw, vldrh_uw, vstrh_w)
59
+DO_VLDST_WIDE_NARROW(VLDSTB_H, vldrb_sh, vldrb_uh, vstrb_h, MO_8)
60
+DO_VLDST_WIDE_NARROW(VLDSTB_W, vldrb_sw, vldrb_uw, vstrb_w, MO_8)
61
+DO_VLDST_WIDE_NARROW(VLDSTH_W, vldrh_sw, vldrh_uw, vstrh_w, MO_16)
62
63
static bool trans_VDUP(DisasContext *s, arg_VDUP *a)
46
{
64
{
47
return q->size - fifo_elements_number(q);
48
}
49
@@ -XXX,XX +XXX,XX @@ static void fifo_reset(Exynos4210UartFIFO *q)
50
q->rp = 0;
51
}
52
53
-static uint32_t exynos4210_uart_Tx_FIFO_trigger_level(Exynos4210UartState *s)
54
+static uint32_t exynos4210_uart_Tx_FIFO_trigger_level(const Exynos4210UartState *s)
55
{
56
uint32_t level = 0;
57
uint32_t reg;
58
--
65
--
59
2.7.4
66
2.20.1
60
67
61
68
diff view generated by jsdifflib
1
From: Krzysztof Kozlowski <krzk@kernel.org>
1
The initial implementation of the MVE VRMLALDAVH and VRMLSLDAVH
2
insns had some bugs:
3
* the 32x32 multiply of elements was being done as 32x32->32,
4
not 32x32->64
5
* we were incorrectly maintaining the accumulator in its full
6
72-bit form across all 4 beats of the insn; in the pseudocode
7
it is squashed back into the 64 bits of the RdaHi:RdaLo
8
registers after each beat
2
9
3
qemu_log_mask() and error_report() are preferred over fprintf() for
10
In particular, fixing the second of these allows us to recast
4
logging errors. Also remove square brackets [] and additional new line
11
the implementation to avoid 128-bit arithmetic entirely.
5
characters in printed messages.
6
12
7
Signed-off-by: Krzysztof Kozlowski <krzk@kernel.org>
13
Since the element size here is always 4, we can also drop the
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
14
parameterization of ESIZE to make the code a little more readable.
9
Message-id: 20170313184750.429-2-krzk@kernel.org
15
10
[PMM: wrapped long line]
16
Suggested-by: Richard Henderson <richard.henderson@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
19
Message-id: 20210628135835.6690-3-peter.maydell@linaro.org
13
---
20
---
14
hw/arm/exynos4_boards.c | 7 ++++---
21
target/arm/mve_helper.c | 38 +++++++++++++++++++++-----------------
15
hw/timer/exynos4210_mct.c | 6 ++++--
22
1 file changed, 21 insertions(+), 17 deletions(-)
16
hw/timer/exynos4210_pwm.c | 13 +++++++------
17
hw/timer/exynos4210_rtc.c | 19 ++++++++++---------
18
4 files changed, 25 insertions(+), 20 deletions(-)
19
23
20
diff --git a/hw/arm/exynos4_boards.c b/hw/arm/exynos4_boards.c
24
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
21
index XXXXXXX..XXXXXXX 100644
25
index XXXXXXX..XXXXXXX 100644
22
--- a/hw/arm/exynos4_boards.c
26
--- a/target/arm/mve_helper.c
23
+++ b/hw/arm/exynos4_boards.c
27
+++ b/target/arm/mve_helper.c
24
@@ -XXX,XX +XXX,XX @@
28
@@ -XXX,XX +XXX,XX @@
25
*/
29
*/
26
30
27
#include "qemu/osdep.h"
31
#include "qemu/osdep.h"
28
+#include "qemu/error-report.h"
32
-#include "qemu/int128.h"
29
#include "qemu-common.h"
30
#include "cpu.h"
33
#include "cpu.h"
31
#include "sysemu/sysemu.h"
34
#include "internals.h"
32
@@ -XXX,XX +XXX,XX @@ static Exynos4210State *exynos4_boards_init_common(MachineState *machine,
35
#include "vec_internal.h"
33
MachineClass *mc = MACHINE_GET_CLASS(machine);
36
@@ -XXX,XX +XXX,XX @@ DO_LDAV(vmlsldavsw, 4, int32_t, false, +=, -=)
34
37
DO_LDAV(vmlsldavxsw, 4, int32_t, true, +=, -=)
35
if (smp_cpus != EXYNOS4210_NCPUS && !qtest_enabled()) {
38
36
- fprintf(stderr, "%s board supports only %d CPU cores. Ignoring smp_cpus"
39
/*
37
- " value.\n",
40
- * Rounding multiply add long dual accumulate high: we must keep
38
- mc->name, EXYNOS4210_NCPUS);
41
- * a 72-bit internal accumulator value and return the top 64 bits.
39
+ error_report("%s board supports only %d CPU cores, ignoring smp_cpus"
42
+ * Rounding multiply add long dual accumulate high. In the pseudocode
40
+ " value",
43
+ * this is implemented with a 72-bit internal accumulator value of which
41
+ mc->name, EXYNOS4210_NCPUS);
44
+ * the top 64 bits are returned. We optimize this to avoid having to
45
+ * use 128-bit arithmetic -- we can do this because the 74-bit accumulator
46
+ * is squashed back into 64-bits after each beat.
47
*/
48
-#define DO_LDAVH(OP, ESIZE, TYPE, XCHG, EVENACC, ODDACC, TO128) \
49
+#define DO_LDAVH(OP, TYPE, LTYPE, XCHG, SUB) \
50
uint64_t HELPER(glue(mve_, OP))(CPUARMState *env, void *vn, \
51
void *vm, uint64_t a) \
52
{ \
53
uint16_t mask = mve_element_mask(env); \
54
unsigned e; \
55
TYPE *n = vn, *m = vm; \
56
- Int128 acc = int128_lshift(TO128(a), 8); \
57
- for (e = 0; e < 16 / ESIZE; e++, mask >>= ESIZE) { \
58
+ for (e = 0; e < 16 / 4; e++, mask >>= 4) { \
59
if (mask & 1) { \
60
+ LTYPE mul; \
61
if (e & 1) { \
62
- acc = ODDACC(acc, TO128(n[H##ESIZE(e - 1 * XCHG)] * \
63
- m[H##ESIZE(e)])); \
64
+ mul = (LTYPE)n[H4(e - 1 * XCHG)] * m[H4(e)]; \
65
+ if (SUB) { \
66
+ mul = -mul; \
67
+ } \
68
} else { \
69
- acc = EVENACC(acc, TO128(n[H##ESIZE(e + 1 * XCHG)] * \
70
- m[H##ESIZE(e)])); \
71
+ mul = (LTYPE)n[H4(e + 1 * XCHG)] * m[H4(e)]; \
72
} \
73
- acc = int128_add(acc, int128_make64(1 << 7)); \
74
+ mul = (mul >> 8) + ((mul >> 7) & 1); \
75
+ a += mul; \
76
} \
77
} \
78
mve_advance_vpt(env); \
79
- return int128_getlo(int128_rshift(acc, 8)); \
80
+ return a; \
42
}
81
}
43
82
44
exynos4_board_binfo.ram_size = exynos4_board_ram_size[board_type];
83
-DO_LDAVH(vrmlaldavhsw, 4, int32_t, false, int128_add, int128_add, int128_makes64)
45
diff --git a/hw/timer/exynos4210_mct.c b/hw/timer/exynos4210_mct.c
84
-DO_LDAVH(vrmlaldavhxsw, 4, int32_t, true, int128_add, int128_add, int128_makes64)
46
index XXXXXXX..XXXXXXX 100644
85
+DO_LDAVH(vrmlaldavhsw, int32_t, int64_t, false, false)
47
--- a/hw/timer/exynos4210_mct.c
86
+DO_LDAVH(vrmlaldavhxsw, int32_t, int64_t, true, false)
48
+++ b/hw/timer/exynos4210_mct.c
87
49
@@ -XXX,XX +XXX,XX @@
88
-DO_LDAVH(vrmlaldavhuw, 4, uint32_t, false, int128_add, int128_add, int128_make64)
50
*/
89
+DO_LDAVH(vrmlaldavhuw, uint32_t, uint64_t, false, false)
51
90
52
#include "qemu/osdep.h"
91
-DO_LDAVH(vrmlsldavhsw, 4, int32_t, false, int128_add, int128_sub, int128_makes64)
53
+#include "qemu/log.h"
92
-DO_LDAVH(vrmlsldavhxsw, 4, int32_t, true, int128_add, int128_sub, int128_makes64)
54
#include "hw/sysbus.h"
93
+DO_LDAVH(vrmlsldavhsw, int32_t, int64_t, false, true)
55
#include "qemu/timer.h"
94
+DO_LDAVH(vrmlsldavhxsw, int32_t, int64_t, true, true)
56
#include "qemu/main-loop.h"
95
57
@@ -XXX,XX +XXX,XX @@ break;
96
/* Vector add across vector */
58
case L0_TCNTO: case L1_TCNTO:
97
#define DO_VADDV(OP, ESIZE, TYPE) \
59
case L0_ICNTO: case L1_ICNTO:
60
case L0_FRCNTO: case L1_FRCNTO:
61
- fprintf(stderr, "\n[exynos4210.mct: write to RO register "
62
- TARGET_FMT_plx "]\n\n", offset);
63
+ qemu_log_mask(LOG_GUEST_ERROR,
64
+ "exynos4210.mct: write to RO register " TARGET_FMT_plx,
65
+ offset);
66
break;
67
68
case L0_INT_CSTAT: case L1_INT_CSTAT:
69
diff --git a/hw/timer/exynos4210_pwm.c b/hw/timer/exynos4210_pwm.c
70
index XXXXXXX..XXXXXXX 100644
71
--- a/hw/timer/exynos4210_pwm.c
72
+++ b/hw/timer/exynos4210_pwm.c
73
@@ -XXX,XX +XXX,XX @@
74
*/
75
76
#include "qemu/osdep.h"
77
+#include "qemu/log.h"
78
#include "hw/sysbus.h"
79
#include "qemu/timer.h"
80
#include "qemu-common.h"
81
@@ -XXX,XX +XXX,XX @@ static uint64_t exynos4210_pwm_read(void *opaque, hwaddr offset,
82
break;
83
84
default:
85
- fprintf(stderr,
86
- "[exynos4210.pwm: bad read offset " TARGET_FMT_plx "]\n",
87
- offset);
88
+ qemu_log_mask(LOG_GUEST_ERROR,
89
+ "exynos4210.pwm: bad read offset " TARGET_FMT_plx,
90
+ offset);
91
break;
92
}
93
return value;
94
@@ -XXX,XX +XXX,XX @@ static void exynos4210_pwm_write(void *opaque, hwaddr offset,
95
break;
96
97
default:
98
- fprintf(stderr,
99
- "[exynos4210.pwm: bad write offset " TARGET_FMT_plx "]\n",
100
- offset);
101
+ qemu_log_mask(LOG_GUEST_ERROR,
102
+ "exynos4210.pwm: bad write offset " TARGET_FMT_plx,
103
+ offset);
104
break;
105
106
}
107
diff --git a/hw/timer/exynos4210_rtc.c b/hw/timer/exynos4210_rtc.c
108
index XXXXXXX..XXXXXXX 100644
109
--- a/hw/timer/exynos4210_rtc.c
110
+++ b/hw/timer/exynos4210_rtc.c
111
@@ -XXX,XX +XXX,XX @@
112
*/
113
114
#include "qemu/osdep.h"
115
+#include "qemu/log.h"
116
#include "hw/sysbus.h"
117
#include "qemu/timer.h"
118
#include "qemu-common.h"
119
@@ -XXX,XX +XXX,XX @@ static uint64_t exynos4210_rtc_read(void *opaque, hwaddr offset,
120
break;
121
122
default:
123
- fprintf(stderr,
124
- "[exynos4210.rtc: bad read offset " TARGET_FMT_plx "]\n",
125
- offset);
126
+ qemu_log_mask(LOG_GUEST_ERROR,
127
+ "exynos4210.rtc: bad read offset " TARGET_FMT_plx,
128
+ offset);
129
break;
130
}
131
return value;
132
@@ -XXX,XX +XXX,XX @@ static void exynos4210_rtc_write(void *opaque, hwaddr offset,
133
if (value > TICNT_THRESHOLD) {
134
s->reg_ticcnt = value;
135
} else {
136
- fprintf(stderr,
137
- "[exynos4210.rtc: bad TICNT value %u ]\n",
138
- (uint32_t)value);
139
+ qemu_log_mask(LOG_GUEST_ERROR,
140
+ "exynos4210.rtc: bad TICNT value %u",
141
+ (uint32_t)value);
142
}
143
break;
144
145
@@ -XXX,XX +XXX,XX @@ static void exynos4210_rtc_write(void *opaque, hwaddr offset,
146
break;
147
148
default:
149
- fprintf(stderr,
150
- "[exynos4210.rtc: bad write offset " TARGET_FMT_plx "]\n",
151
- offset);
152
+ qemu_log_mask(LOG_GUEST_ERROR,
153
+ "exynos4210.rtc: bad write offset " TARGET_FMT_plx,
154
+ offset);
155
break;
156
157
}
158
--
98
--
159
2.7.4
99
2.20.1
160
100
161
101
diff view generated by jsdifflib
1
Now that we've rewritten M-profile exception return so that the magic
1
The function asimd_imm_const() in translate-neon.c is an
2
PC values are not visible to other parts of QEMU, we can delete the
2
implementation of the pseudocode AdvSIMDExpandImm(), which we will
3
special casing of them elsewhere.
3
also want for MVE. Move the implementation to translate.c, with a
4
prototype in translate.h.
4
5
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Richard Henderson <rth@twiddle.net>
8
Message-id: 20210628135835.6690-4-peter.maydell@linaro.org
8
Message-id: 1491844419-12485-10-git-send-email-peter.maydell@linaro.org
9
---
9
---
10
target/arm/cpu.c | 43 ++-----------------------------------------
10
target/arm/translate.h | 16 ++++++++++
11
target/arm/translate.c | 8 --------
11
target/arm/translate-neon.c | 63 -------------------------------------
12
2 files changed, 2 insertions(+), 49 deletions(-)
12
target/arm/translate.c | 57 +++++++++++++++++++++++++++++++++
13
3 files changed, 73 insertions(+), 63 deletions(-)
13
14
14
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.c
17
--- a/target/arm/translate.h
17
+++ b/target/arm/cpu.c
18
+++ b/target/arm/translate.h
18
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
19
@@ -XXX,XX +XXX,XX @@ static inline MemOp finalize_memop(DisasContext *s, MemOp opc)
20
return opc | s->be_data;
19
}
21
}
20
22
21
#if !defined(CONFIG_USER_ONLY) || !defined(TARGET_AARCH64)
23
+/**
22
-static void arm_v7m_unassigned_access(CPUState *cpu, hwaddr addr,
24
+ * asimd_imm_const: Expand an encoded SIMD constant value
23
- bool is_write, bool is_exec, int opaque,
25
+ *
24
- unsigned size)
26
+ * Expand a SIMD constant value. This is essentially the pseudocode
27
+ * AdvSIMDExpandImm, except that we also perform the boolean NOT needed for
28
+ * VMVN and VBIC (when cmode < 14 && op == 1).
29
+ *
30
+ * The combination cmode == 15 op == 1 is a reserved encoding for AArch32;
31
+ * callers must catch this.
32
+ *
33
+ * cmode = 2,3,4,5,6,7,10,11,12,13 imm=0 was UNPREDICTABLE in v7A but
34
+ * is either not unpredictable or merely CONSTRAINED UNPREDICTABLE in v8A;
35
+ * we produce an immediate constant value of 0 in these cases.
36
+ */
37
+uint64_t asimd_imm_const(uint32_t imm, int cmode, int op);
38
+
39
#endif /* TARGET_ARM_TRANSLATE_H */
40
diff --git a/target/arm/translate-neon.c b/target/arm/translate-neon.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/translate-neon.c
43
+++ b/target/arm/translate-neon.c
44
@@ -XXX,XX +XXX,XX @@ DO_FP_2SH(VCVT_UH, gen_helper_gvec_vcvt_uh)
45
DO_FP_2SH(VCVT_HS, gen_helper_gvec_vcvt_hs)
46
DO_FP_2SH(VCVT_HU, gen_helper_gvec_vcvt_hu)
47
48
-static uint64_t asimd_imm_const(uint32_t imm, int cmode, int op)
25
-{
49
-{
26
- ARMCPU *arm = ARM_CPU(cpu);
50
- /*
27
- CPUARMState *env = &arm->env;
51
- * Expand the encoded constant.
52
- * Note that cmode = 2,3,4,5,6,7,10,11,12,13 imm=0 is UNPREDICTABLE.
53
- * We choose to not special-case this and will behave as if a
54
- * valid constant encoding of 0 had been given.
55
- * cmode = 15 op = 1 must UNDEF; we assume decode has handled that.
56
- */
57
- switch (cmode) {
58
- case 0: case 1:
59
- /* no-op */
60
- break;
61
- case 2: case 3:
62
- imm <<= 8;
63
- break;
64
- case 4: case 5:
65
- imm <<= 16;
66
- break;
67
- case 6: case 7:
68
- imm <<= 24;
69
- break;
70
- case 8: case 9:
71
- imm |= imm << 16;
72
- break;
73
- case 10: case 11:
74
- imm = (imm << 8) | (imm << 24);
75
- break;
76
- case 12:
77
- imm = (imm << 8) | 0xff;
78
- break;
79
- case 13:
80
- imm = (imm << 16) | 0xffff;
81
- break;
82
- case 14:
83
- if (op) {
84
- /*
85
- * This is the only case where the top and bottom 32 bits
86
- * of the encoded constant differ.
87
- */
88
- uint64_t imm64 = 0;
89
- int n;
28
-
90
-
29
- /* ARMv7-M interrupt return works by loading a magic value into the PC.
91
- for (n = 0; n < 8; n++) {
30
- * On real hardware the load causes the return to occur. The qemu
92
- if (imm & (1 << n)) {
31
- * implementation performs the jump normally, then does the exception
93
- imm64 |= (0xffULL << (n * 8));
32
- * return by throwing a special exception when when the CPU tries to
94
- }
33
- * execute code at the magic address.
95
- }
34
- */
96
- return imm64;
35
- if (env->v7m.exception != 0 && addr >= 0xfffffff0 && is_exec) {
97
- }
36
- cpu->exception_index = EXCP_EXCEPTION_EXIT;
98
- imm |= (imm << 8) | (imm << 16) | (imm << 24);
37
- cpu_loop_exit(cpu);
99
- break;
100
- case 15:
101
- imm = ((imm & 0x80) << 24) | ((imm & 0x3f) << 19)
102
- | ((imm & 0x40) ? (0x1f << 25) : (1 << 30));
103
- break;
38
- }
104
- }
39
-
105
- if (op) {
40
- /* In real hardware an attempt to access parts of the address space
106
- imm = ~imm;
41
- * with nothing there will usually cause an external abort.
107
- }
42
- * However our QEMU board models are often missing device models where
108
- return dup_const(MO_32, imm);
43
- * the guest can boot anyway with the default read-as-zero/writes-ignored
44
- * behaviour that you get without a QEMU unassigned_access hook.
45
- * So just return here to retain that default behaviour.
46
- */
47
-}
109
-}
48
-
110
-
49
static bool arm_v7m_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
111
static bool do_1reg_imm(DisasContext *s, arg_1reg_imm *a,
112
GVecGen2iFn *fn)
50
{
113
{
51
CPUClass *cc = CPU_GET_CLASS(cs);
52
@@ -XXX,XX +XXX,XX @@ static bool arm_v7m_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
53
CPUARMState *env = &cpu->env;
54
bool ret = false;
55
56
- /* ARMv7-M interrupt return works by loading a magic value
57
- * into the PC. On real hardware the load causes the
58
- * return to occur. The qemu implementation performs the
59
- * jump normally, then does the exception return when the
60
- * CPU tries to execute code at the magic address.
61
- * This will cause the magic PC value to be pushed to
62
- * the stack if an interrupt occurred at the wrong time.
63
- * We avoid this by disabling interrupts when
64
- * pc contains a magic address.
65
- *
66
- * ARMv7-M interrupt masking works differently than -A or -R.
67
+ /* ARMv7-M interrupt masking works differently than -A or -R.
68
* There is no FIQ/IRQ distinction. Instead of I and F bits
69
* masking FIQ and IRQ interrupts, an exception is taken only
70
* if it is higher priority than the current execution priority
71
@@ -XXX,XX +XXX,XX @@ static bool arm_v7m_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
72
* currently active exception).
73
*/
74
if (interrupt_request & CPU_INTERRUPT_HARD
75
- && (armv7m_nvic_can_take_pending_exception(env->nvic))
76
- && (env->regs[15] < 0xfffffff0)) {
77
+ && (armv7m_nvic_can_take_pending_exception(env->nvic))) {
78
cs->exception_index = EXCP_IRQ;
79
cc->do_interrupt(cs);
80
ret = true;
81
@@ -XXX,XX +XXX,XX @@ static void arm_v7m_class_init(ObjectClass *oc, void *data)
82
cc->do_interrupt = arm_v7m_cpu_do_interrupt;
83
#endif
84
85
- cc->do_unassigned_access = arm_v7m_unassigned_access;
86
cc->cpu_exec_interrupt = arm_v7m_cpu_exec_interrupt;
87
}
88
89
diff --git a/target/arm/translate.c b/target/arm/translate.c
114
diff --git a/target/arm/translate.c b/target/arm/translate.c
90
index XXXXXXX..XXXXXXX 100644
115
index XXXXXXX..XXXXXXX 100644
91
--- a/target/arm/translate.c
116
--- a/target/arm/translate.c
92
+++ b/target/arm/translate.c
117
+++ b/target/arm/translate.c
93
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
118
@@ -XXX,XX +XXX,XX @@ void arm_translate_init(void)
94
dc->is_jmp = DISAS_EXC;
119
a64_translate_init();
95
break;
120
}
96
}
121
97
-#else
122
+uint64_t asimd_imm_const(uint32_t imm, int cmode, int op)
98
- if (arm_dc_feature(dc, ARM_FEATURE_M)) {
123
+{
99
- /* Branches to the magic exception-return addresses should
124
+ /* Expand the encoded constant as per AdvSIMDExpandImm pseudocode */
100
- * already have been caught via the arm_v7m_unassigned_access hook,
125
+ switch (cmode) {
101
- * and never get here.
126
+ case 0: case 1:
102
- */
127
+ /* no-op */
103
- assert(dc->pc < 0xfffffff0);
128
+ break;
104
- }
129
+ case 2: case 3:
105
#endif
130
+ imm <<= 8;
106
131
+ break;
107
if (unlikely(!QTAILQ_EMPTY(&cs->breakpoints))) {
132
+ case 4: case 5:
133
+ imm <<= 16;
134
+ break;
135
+ case 6: case 7:
136
+ imm <<= 24;
137
+ break;
138
+ case 8: case 9:
139
+ imm |= imm << 16;
140
+ break;
141
+ case 10: case 11:
142
+ imm = (imm << 8) | (imm << 24);
143
+ break;
144
+ case 12:
145
+ imm = (imm << 8) | 0xff;
146
+ break;
147
+ case 13:
148
+ imm = (imm << 16) | 0xffff;
149
+ break;
150
+ case 14:
151
+ if (op) {
152
+ /*
153
+ * This is the only case where the top and bottom 32 bits
154
+ * of the encoded constant differ.
155
+ */
156
+ uint64_t imm64 = 0;
157
+ int n;
158
+
159
+ for (n = 0; n < 8; n++) {
160
+ if (imm & (1 << n)) {
161
+ imm64 |= (0xffULL << (n * 8));
162
+ }
163
+ }
164
+ return imm64;
165
+ }
166
+ imm |= (imm << 8) | (imm << 16) | (imm << 24);
167
+ break;
168
+ case 15:
169
+ imm = ((imm & 0x80) << 24) | ((imm & 0x3f) << 19)
170
+ | ((imm & 0x40) ? (0x1f << 25) : (1 << 30));
171
+ break;
172
+ }
173
+ if (op) {
174
+ imm = ~imm;
175
+ }
176
+ return dup_const(MO_32, imm);
177
+}
178
+
179
/* Generate a label used for skipping this instruction */
180
void arm_gen_condlabel(DisasContext *s)
181
{
108
--
182
--
109
2.7.4
183
2.20.1
110
184
111
185
diff view generated by jsdifflib
1
For M profile exception-return handling we'd like to generate different
1
The A64 AdvSIMD modified-immediate grouping uses almost the same
2
code for some instructions depending on whether we are in Handler
2
constant encoding that A32 Neon does; reuse asimd_imm_const() (to
3
mode or Thread mode. This isn't the same as "are we privileged
3
which we add the AArch64-specific case for cmode 15 op 1) instead of
4
or user", so we need an extra bit in the TB flags to distinguish.
4
reimplementing it all.
5
5
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <rth@twiddle.net>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20210628135835.6690-5-peter.maydell@linaro.org
9
Message-id: 1491844419-12485-8-git-send-email-peter.maydell@linaro.org
10
---
9
---
11
target/arm/cpu.h | 9 +++++++++
10
target/arm/translate.h | 3 +-
12
target/arm/translate.h | 1 +
11
target/arm/translate-a64.c | 86 ++++----------------------------------
13
target/arm/translate.c | 1 +
12
target/arm/translate.c | 17 +++++++-
14
3 files changed, 11 insertions(+)
13
3 files changed, 24 insertions(+), 82 deletions(-)
15
14
16
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/cpu.h
19
+++ b/target/arm/cpu.h
20
@@ -XXX,XX +XXX,XX @@ static inline bool arm_cpu_data_is_big_endian(CPUARMState *env)
21
#define ARM_TBFLAG_NS_MASK (1 << ARM_TBFLAG_NS_SHIFT)
22
#define ARM_TBFLAG_BE_DATA_SHIFT 20
23
#define ARM_TBFLAG_BE_DATA_MASK (1 << ARM_TBFLAG_BE_DATA_SHIFT)
24
+/* For M profile only, Handler (ie not Thread) mode */
25
+#define ARM_TBFLAG_HANDLER_SHIFT 21
26
+#define ARM_TBFLAG_HANDLER_MASK (1 << ARM_TBFLAG_HANDLER_SHIFT)
27
28
/* Bit usage when in AArch64 state */
29
#define ARM_TBFLAG_TBI0_SHIFT 0 /* TBI0 for EL0/1 or TBI for EL2/3 */
30
@@ -XXX,XX +XXX,XX @@ static inline bool arm_cpu_data_is_big_endian(CPUARMState *env)
31
(((F) & ARM_TBFLAG_NS_MASK) >> ARM_TBFLAG_NS_SHIFT)
32
#define ARM_TBFLAG_BE_DATA(F) \
33
(((F) & ARM_TBFLAG_BE_DATA_MASK) >> ARM_TBFLAG_BE_DATA_SHIFT)
34
+#define ARM_TBFLAG_HANDLER(F) \
35
+ (((F) & ARM_TBFLAG_HANDLER_MASK) >> ARM_TBFLAG_HANDLER_SHIFT)
36
#define ARM_TBFLAG_TBI0(F) \
37
(((F) & ARM_TBFLAG_TBI0_MASK) >> ARM_TBFLAG_TBI0_SHIFT)
38
#define ARM_TBFLAG_TBI1(F) \
39
@@ -XXX,XX +XXX,XX @@ static inline void cpu_get_tb_cpu_state(CPUARMState *env, target_ulong *pc,
40
}
41
*flags |= fp_exception_el(env) << ARM_TBFLAG_FPEXC_EL_SHIFT;
42
43
+ if (env->v7m.exception != 0) {
44
+ *flags |= ARM_TBFLAG_HANDLER_MASK;
45
+ }
46
+
47
*cs_base = 0;
48
}
49
50
diff --git a/target/arm/translate.h b/target/arm/translate.h
15
diff --git a/target/arm/translate.h b/target/arm/translate.h
51
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/translate.h
17
--- a/target/arm/translate.h
53
+++ b/target/arm/translate.h
18
+++ b/target/arm/translate.h
54
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
19
@@ -XXX,XX +XXX,XX @@ static inline MemOp finalize_memop(DisasContext *s, MemOp opc)
55
bool vfp_enabled; /* FP enabled via FPSCR.EN */
20
* VMVN and VBIC (when cmode < 14 && op == 1).
56
int vec_len;
21
*
57
int vec_stride;
22
* The combination cmode == 15 op == 1 is a reserved encoding for AArch32;
58
+ bool v7m_handler_mode;
23
- * callers must catch this.
59
/* Immediate value in AArch32 SVC insn; must be set if is_jmp == DISAS_SWI
24
+ * callers must catch this; we return the 64-bit constant value defined
60
* so that top level loop can generate correct syndrome information.
25
+ * for AArch64.
61
*/
26
*
27
* cmode = 2,3,4,5,6,7,10,11,12,13 imm=0 was UNPREDICTABLE in v7A but
28
* is either not unpredictable or merely CONSTRAINED UNPREDICTABLE in v8A;
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
30
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/translate-a64.c
32
+++ b/target/arm/translate-a64.c
33
@@ -XXX,XX +XXX,XX @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
34
{
35
int rd = extract32(insn, 0, 5);
36
int cmode = extract32(insn, 12, 4);
37
- int cmode_3_1 = extract32(cmode, 1, 3);
38
- int cmode_0 = extract32(cmode, 0, 1);
39
int o2 = extract32(insn, 11, 1);
40
uint64_t abcdefgh = extract32(insn, 5, 5) | (extract32(insn, 16, 3) << 5);
41
bool is_neg = extract32(insn, 29, 1);
42
@@ -XXX,XX +XXX,XX @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
43
return;
44
}
45
46
- /* See AdvSIMDExpandImm() in ARM ARM */
47
- switch (cmode_3_1) {
48
- case 0: /* Replicate(Zeros(24):imm8, 2) */
49
- case 1: /* Replicate(Zeros(16):imm8:Zeros(8), 2) */
50
- case 2: /* Replicate(Zeros(8):imm8:Zeros(16), 2) */
51
- case 3: /* Replicate(imm8:Zeros(24), 2) */
52
- {
53
- int shift = cmode_3_1 * 8;
54
- imm = bitfield_replicate(abcdefgh << shift, 32);
55
- break;
56
- }
57
- case 4: /* Replicate(Zeros(8):imm8, 4) */
58
- case 5: /* Replicate(imm8:Zeros(8), 4) */
59
- {
60
- int shift = (cmode_3_1 & 0x1) * 8;
61
- imm = bitfield_replicate(abcdefgh << shift, 16);
62
- break;
63
- }
64
- case 6:
65
- if (cmode_0) {
66
- /* Replicate(Zeros(8):imm8:Ones(16), 2) */
67
- imm = (abcdefgh << 16) | 0xffff;
68
- } else {
69
- /* Replicate(Zeros(16):imm8:Ones(8), 2) */
70
- imm = (abcdefgh << 8) | 0xff;
71
- }
72
- imm = bitfield_replicate(imm, 32);
73
- break;
74
- case 7:
75
- if (!cmode_0 && !is_neg) {
76
- imm = bitfield_replicate(abcdefgh, 8);
77
- } else if (!cmode_0 && is_neg) {
78
- int i;
79
- imm = 0;
80
- for (i = 0; i < 8; i++) {
81
- if ((abcdefgh) & (1 << i)) {
82
- imm |= 0xffULL << (i * 8);
83
- }
84
- }
85
- } else if (cmode_0) {
86
- if (is_neg) {
87
- imm = (abcdefgh & 0x3f) << 48;
88
- if (abcdefgh & 0x80) {
89
- imm |= 0x8000000000000000ULL;
90
- }
91
- if (abcdefgh & 0x40) {
92
- imm |= 0x3fc0000000000000ULL;
93
- } else {
94
- imm |= 0x4000000000000000ULL;
95
- }
96
- } else {
97
- if (o2) {
98
- /* FMOV (vector, immediate) - half-precision */
99
- imm = vfp_expand_imm(MO_16, abcdefgh);
100
- /* now duplicate across the lanes */
101
- imm = bitfield_replicate(imm, 16);
102
- } else {
103
- imm = (abcdefgh & 0x3f) << 19;
104
- if (abcdefgh & 0x80) {
105
- imm |= 0x80000000;
106
- }
107
- if (abcdefgh & 0x40) {
108
- imm |= 0x3e000000;
109
- } else {
110
- imm |= 0x40000000;
111
- }
112
- imm |= (imm << 32);
113
- }
114
- }
115
- }
116
- break;
117
- default:
118
- g_assert_not_reached();
119
- }
120
-
121
- if (cmode_3_1 != 7 && is_neg) {
122
- imm = ~imm;
123
+ if (cmode == 15 && o2 && !is_neg) {
124
+ /* FMOV (vector, immediate) - half-precision */
125
+ imm = vfp_expand_imm(MO_16, abcdefgh);
126
+ /* now duplicate across the lanes */
127
+ imm = bitfield_replicate(imm, 16);
128
+ } else {
129
+ imm = asimd_imm_const(abcdefgh, cmode, is_neg);
130
}
131
132
if (!((cmode & 0x9) == 0x1 || (cmode & 0xd) == 0x9)) {
62
diff --git a/target/arm/translate.c b/target/arm/translate.c
133
diff --git a/target/arm/translate.c b/target/arm/translate.c
63
index XXXXXXX..XXXXXXX 100644
134
index XXXXXXX..XXXXXXX 100644
64
--- a/target/arm/translate.c
135
--- a/target/arm/translate.c
65
+++ b/target/arm/translate.c
136
+++ b/target/arm/translate.c
66
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
137
@@ -XXX,XX +XXX,XX @@ uint64_t asimd_imm_const(uint32_t imm, int cmode, int op)
67
dc->vec_len = ARM_TBFLAG_VECLEN(tb->flags);
138
case 14:
68
dc->vec_stride = ARM_TBFLAG_VECSTRIDE(tb->flags);
139
if (op) {
69
dc->c15_cpar = ARM_TBFLAG_XSCALE_CPAR(tb->flags);
140
/*
70
+ dc->v7m_handler_mode = ARM_TBFLAG_HANDLER(tb->flags);
141
- * This is the only case where the top and bottom 32 bits
71
dc->cp_regs = cpu->cp_regs;
142
- * of the encoded constant differ.
72
dc->features = env->features;
143
+ * This and cmode == 15 op == 1 are the only cases where
73
144
+ * the top and bottom 32 bits of the encoded constant differ.
145
*/
146
uint64_t imm64 = 0;
147
int n;
148
@@ -XXX,XX +XXX,XX @@ uint64_t asimd_imm_const(uint32_t imm, int cmode, int op)
149
imm |= (imm << 8) | (imm << 16) | (imm << 24);
150
break;
151
case 15:
152
+ if (op) {
153
+ /* Reserved encoding for AArch32; valid for AArch64 */
154
+ uint64_t imm64 = (uint64_t)(imm & 0x3f) << 48;
155
+ if (imm & 0x80) {
156
+ imm64 |= 0x8000000000000000ULL;
157
+ }
158
+ if (imm & 0x40) {
159
+ imm64 |= 0x3fc0000000000000ULL;
160
+ } else {
161
+ imm64 |= 0x4000000000000000ULL;
162
+ }
163
+ return imm64;
164
+ }
165
imm = ((imm & 0x80) << 24) | ((imm & 0x3f) << 19)
166
| ((imm & 0x40) ? (0x1f << 25) : (1 << 30));
167
break;
74
--
168
--
75
2.7.4
169
2.20.1
76
170
77
171
diff view generated by jsdifflib
1
From: Alistair Francis <alistair.francis@xilinx.com>
1
Use dup_const() instead of bitfield_replicate() in
2
disas_simd_mod_imm().
2
3
3
Correct the buffer descriptor busy logic to work correctly when using
4
(We can't replace the other use of bitfield_replicate() in this file,
4
multiple queues.
5
in logic_imm_decode_wmask(), because that location needs to handle 2
6
and 4 bit elements, which dup_const() cannot.)
5
7
6
Signed-off-by: Alistair Francis <alistair.francis@xilinx.com>
7
Message-id: 8a7e8059984e27d46a276a66299d035a0afd280f.1491947224.git.alistair.francis@xilinx.com
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20210628135835.6690-6-peter.maydell@linaro.org
10
---
11
---
11
hw/net/cadence_gem.c | 17 ++++++++++-------
12
target/arm/translate-a64.c | 2 +-
12
1 file changed, 10 insertions(+), 7 deletions(-)
13
1 file changed, 1 insertion(+), 1 deletion(-)
13
14
14
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
15
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/hw/net/cadence_gem.c
17
--- a/target/arm/translate-a64.c
17
+++ b/hw/net/cadence_gem.c
18
+++ b/target/arm/translate-a64.c
18
@@ -XXX,XX +XXX,XX @@ static int gem_can_receive(NetClientState *nc)
19
@@ -XXX,XX +XXX,XX @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
20
/* FMOV (vector, immediate) - half-precision */
21
imm = vfp_expand_imm(MO_16, abcdefgh);
22
/* now duplicate across the lanes */
23
- imm = bitfield_replicate(imm, 16);
24
+ imm = dup_const(MO_16, imm);
25
} else {
26
imm = asimd_imm_const(abcdefgh, cmode, is_neg);
19
}
27
}
20
21
for (i = 0; i < s->num_priority_queues; i++) {
22
- if (rx_desc_get_ownership(s->rx_desc[i]) == 1) {
23
- if (s->can_rx_state != 2) {
24
- s->can_rx_state = 2;
25
- DB_PRINT("can't receive - busy buffer descriptor (q%d) 0x%x\n",
26
- i, s->rx_desc_addr[i]);
27
- }
28
- return 0;
29
+ if (rx_desc_get_ownership(s->rx_desc[i]) != 1) {
30
+ break;
31
+ }
32
+ };
33
+
34
+ if (i == s->num_priority_queues) {
35
+ if (s->can_rx_state != 2) {
36
+ s->can_rx_state = 2;
37
+ DB_PRINT("can't receive - all the buffer descriptors are busy\n");
38
}
39
+ return 0;
40
}
41
42
if (s->can_rx_state != 0) {
43
--
28
--
44
2.7.4
29
2.20.1
45
30
46
31
diff view generated by jsdifflib
1
The excnames[] array is defined in internals.h because we used
1
Implement the MVE logical-immediate insns (VMOV, VMVN,
2
to use it from two different source files for handling logging
2
VORR and VBIC). These have essentially the same encoding
3
of AArch32 and AArch64 exception entry. Refactoring means that
3
as their Neon equivalents, and we implement the decode
4
it's now used only in arm_log_exception() in helper.c, so move
4
in the same way.
5
the array into that function.
6
5
7
Suggested-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 1491821097-5647-1-git-send-email-peter.maydell@linaro.org
8
Message-id: 20210628135835.6690-7-peter.maydell@linaro.org
11
---
9
---
12
target/arm/cpu.h | 2 +-
10
target/arm/helper-mve.h | 4 +++
13
target/arm/internals.h | 23 -----------------------
11
target/arm/mve.decode | 17 +++++++++++++
14
target/arm/helper.c | 19 +++++++++++++++++++
12
target/arm/mve_helper.c | 24 ++++++++++++++++++
15
3 files changed, 20 insertions(+), 24 deletions(-)
13
target/arm/translate-mve.c | 50 ++++++++++++++++++++++++++++++++++++++
14
4 files changed, 95 insertions(+)
16
15
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
16
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
18
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
18
--- a/target/arm/helper-mve.h
20
+++ b/target/arm/cpu.h
19
+++ b/target/arm/helper-mve.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_vaddvsh, TCG_CALL_NO_WG, i32, env, ptr, i32)
21
DEF_HELPER_FLAGS_3(mve_vaddvuh, TCG_CALL_NO_WG, i32, env, ptr, i32)
22
DEF_HELPER_FLAGS_3(mve_vaddvsw, TCG_CALL_NO_WG, i32, env, ptr, i32)
23
DEF_HELPER_FLAGS_3(mve_vaddvuw, TCG_CALL_NO_WG, i32, env, ptr, i32)
24
+
25
+DEF_HELPER_FLAGS_3(mve_vmovi, TCG_CALL_NO_WG, void, env, ptr, i64)
26
+DEF_HELPER_FLAGS_3(mve_vandi, TCG_CALL_NO_WG, void, env, ptr, i64)
27
+DEF_HELPER_FLAGS_3(mve_vorri, TCG_CALL_NO_WG, void, env, ptr, i64)
28
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
29
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/mve.decode
31
+++ b/target/arm/mve.decode
21
@@ -XXX,XX +XXX,XX @@
32
@@ -XXX,XX +XXX,XX @@
22
#define EXCP_SEMIHOST 16 /* semihosting call */
33
# VQDMULL has size in bit 28: 0 for 16 bit, 1 for 32 bit
23
#define EXCP_NOCP 17 /* v7M NOCP UsageFault */
34
%size_28 28:1 !function=plus_1
24
#define EXCP_INVSTATE 18 /* v7M INVSTATE UsageFault */
35
25
-/* NB: new EXCP_ defines should be added to the excnames[] array too */
36
+# 1imm format immediate
26
+/* NB: add new EXCP_ defines to the array in arm_log_exception() too */
37
+%imm_28_16_0 28:1 16:3 0:4
27
38
+
28
#define ARMV7M_EXCP_RESET 1
39
&vldr_vstr rn qd imm p a w size l u
29
#define ARMV7M_EXCP_NMI 2
40
&1op qd qm size
30
diff --git a/target/arm/internals.h b/target/arm/internals.h
41
&2op qd qm qn size
42
&2scalar qd qn rm size
43
+&1imm qd imm cmode op
44
45
@vldr_vstr ....... . . . . l:1 rn:4 ... ...... imm:7 &vldr_vstr qd=%qd u=0
46
# Note that both Rn and Qd are 3 bits only (no D bit)
47
@@ -XXX,XX +XXX,XX @@
48
@2op_nosz .... .... .... .... .... .... .... .... &2op qd=%qd qm=%qm qn=%qn size=0
49
@2op_sz28 .... .... .... .... .... .... .... .... &2op qd=%qd qm=%qm qn=%qn \
50
size=%size_28
51
+@1imm .... .... .... .... .... cmode:4 .. op:1 . .... &1imm qd=%qd imm=%imm_28_16_0
52
53
# The _rev suffix indicates that Vn and Vm are reversed. This is
54
# the case for shifts. In the Arm ARM these insns are documented
55
@@ -XXX,XX +XXX,XX @@ VADDV 111 u:1 1110 1111 size:2 01 ... 0 1111 0 0 a:1 0 qm:3 0 rda=%rd
56
# Predicate operations
57
%mask_22_13 22:1 13:3
58
VPST 1111 1110 0 . 11 000 1 ... 0 1111 0100 1101 mask=%mask_22_13
59
+
60
+# Logical immediate operations (1 reg and modified-immediate)
61
+
62
+# The cmode/op bits here decode VORR/VBIC/VMOV/VMVN, but
63
+# not in a way we can conveniently represent in decodetree without
64
+# a lot of repetition:
65
+# VORR: op=0, (cmode & 1) && cmode < 12
66
+# VBIC: op=1, (cmode & 1) && cmode < 12
67
+# VMOV: everything else
68
+# So we have a single decode line and check the cmode/op in the
69
+# trans function.
70
+Vimm_1r 111 . 1111 1 . 00 0 ... ... 0 .... 0 1 . 1 .... @1imm
71
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
31
index XXXXXXX..XXXXXXX 100644
72
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/internals.h
73
--- a/target/arm/mve_helper.c
33
+++ b/target/arm/internals.h
74
+++ b/target/arm/mve_helper.c
34
@@ -XXX,XX +XXX,XX @@ static inline bool excp_is_internal(int excp)
75
@@ -XXX,XX +XXX,XX @@ DO_1OP(vnegw, 4, int32_t, DO_NEG)
35
|| excp == EXCP_SEMIHOST;
76
DO_1OP(vfnegh, 8, uint64_t, DO_FNEGH)
77
DO_1OP(vfnegs, 8, uint64_t, DO_FNEGS)
78
79
+/*
80
+ * 1 operand immediates: Vda is destination and possibly also one source.
81
+ * All these insns work at 64-bit widths.
82
+ */
83
+#define DO_1OP_IMM(OP, FN) \
84
+ void HELPER(mve_##OP)(CPUARMState *env, void *vda, uint64_t imm) \
85
+ { \
86
+ uint64_t *da = vda; \
87
+ uint16_t mask = mve_element_mask(env); \
88
+ unsigned e; \
89
+ for (e = 0; e < 16 / 8; e++, mask >>= 8) { \
90
+ mergemask(&da[H8(e)], FN(da[H8(e)], imm), mask); \
91
+ } \
92
+ mve_advance_vpt(env); \
93
+ }
94
+
95
+#define DO_MOVI(N, I) (I)
96
+#define DO_ANDI(N, I) ((N) & (I))
97
+#define DO_ORRI(N, I) ((N) | (I))
98
+
99
+DO_1OP_IMM(vmovi, DO_MOVI)
100
+DO_1OP_IMM(vandi, DO_ANDI)
101
+DO_1OP_IMM(vorri, DO_ORRI)
102
+
103
#define DO_2OP(OP, ESIZE, TYPE, FN) \
104
void HELPER(glue(mve_, OP))(CPUARMState *env, \
105
void *vd, void *vn, void *vm) \
106
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
107
index XXXXXXX..XXXXXXX 100644
108
--- a/target/arm/translate-mve.c
109
+++ b/target/arm/translate-mve.c
110
@@ -XXX,XX +XXX,XX @@ typedef void MVEGenTwoOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_ptr);
111
typedef void MVEGenTwoOpScalarFn(TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_i32);
112
typedef void MVEGenDualAccOpFn(TCGv_i64, TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_i64);
113
typedef void MVEGenVADDVFn(TCGv_i32, TCGv_ptr, TCGv_ptr, TCGv_i32);
114
+typedef void MVEGenOneOpImmFn(TCGv_ptr, TCGv_ptr, TCGv_i64);
115
116
/* Return the offset of a Qn register (same semantics as aa32_vfp_qreg()) */
117
static inline long mve_qreg_offset(unsigned reg)
118
@@ -XXX,XX +XXX,XX @@ static bool trans_VADDV(DisasContext *s, arg_VADDV *a)
119
mve_update_eci(s);
120
return true;
36
}
121
}
37
122
+
38
-/* Exception names for debug logging; note that not all of these
123
+static bool do_1imm(DisasContext *s, arg_1imm *a, MVEGenOneOpImmFn *fn)
39
- * precisely correspond to architectural exceptions.
124
+{
40
- */
125
+ TCGv_ptr qd;
41
-static const char * const excnames[] = {
126
+ uint64_t imm;
42
- [EXCP_UDEF] = "Undefined Instruction",
127
+
43
- [EXCP_SWI] = "SVC",
128
+ if (!dc_isar_feature(aa32_mve, s) ||
44
- [EXCP_PREFETCH_ABORT] = "Prefetch Abort",
129
+ !mve_check_qreg_bank(s, a->qd) ||
45
- [EXCP_DATA_ABORT] = "Data Abort",
130
+ !fn) {
46
- [EXCP_IRQ] = "IRQ",
131
+ return false;
47
- [EXCP_FIQ] = "FIQ",
132
+ }
48
- [EXCP_BKPT] = "Breakpoint",
133
+ if (!mve_eci_check(s) || !vfp_access_check(s)) {
49
- [EXCP_EXCEPTION_EXIT] = "QEMU v7M exception exit",
134
+ return true;
50
- [EXCP_KERNEL_TRAP] = "QEMU intercept of kernel commpage",
135
+ }
51
- [EXCP_HVC] = "Hypervisor Call",
136
+
52
- [EXCP_HYP_TRAP] = "Hypervisor Trap",
137
+ imm = asimd_imm_const(a->imm, a->cmode, a->op);
53
- [EXCP_SMC] = "Secure Monitor Call",
138
+
54
- [EXCP_VIRQ] = "Virtual IRQ",
139
+ qd = mve_qreg_ptr(a->qd);
55
- [EXCP_VFIQ] = "Virtual FIQ",
140
+ fn(cpu_env, qd, tcg_constant_i64(imm));
56
- [EXCP_SEMIHOST] = "Semihosting call",
141
+ tcg_temp_free_ptr(qd);
57
- [EXCP_NOCP] = "v7M NOCP UsageFault",
142
+ mve_update_eci(s);
58
- [EXCP_INVSTATE] = "v7M INVSTATE UsageFault",
143
+ return true;
59
-};
144
+}
60
-
145
+
61
/* Scale factor for generic timers, ie number of ns per tick.
146
+static bool trans_Vimm_1r(DisasContext *s, arg_1imm *a)
62
* This gives a 62.5MHz timer.
147
+{
63
*/
148
+ /* Handle decode of cmode/op here between VORR/VBIC/VMOV */
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
149
+ MVEGenOneOpImmFn *fn;
65
index XXXXXXX..XXXXXXX 100644
150
+
66
--- a/target/arm/helper.c
151
+ if ((a->cmode & 1) && a->cmode < 12) {
67
+++ b/target/arm/helper.c
152
+ if (a->op) {
68
@@ -XXX,XX +XXX,XX @@ static void arm_log_exception(int idx)
153
+ /*
69
{
154
+ * For op=1, the immediate will be inverted by asimd_imm_const(),
70
if (qemu_loglevel_mask(CPU_LOG_INT)) {
155
+ * so the VBIC becomes a logical AND operation.
71
const char *exc = NULL;
156
+ */
72
+ static const char * const excnames[] = {
157
+ fn = gen_helper_mve_vandi;
73
+ [EXCP_UDEF] = "Undefined Instruction",
158
+ } else {
74
+ [EXCP_SWI] = "SVC",
159
+ fn = gen_helper_mve_vorri;
75
+ [EXCP_PREFETCH_ABORT] = "Prefetch Abort",
160
+ }
76
+ [EXCP_DATA_ABORT] = "Data Abort",
161
+ } else {
77
+ [EXCP_IRQ] = "IRQ",
162
+ /* There is one unallocated cmode/op combination in this space */
78
+ [EXCP_FIQ] = "FIQ",
163
+ if (a->cmode == 15 && a->op == 1) {
79
+ [EXCP_BKPT] = "Breakpoint",
164
+ return false;
80
+ [EXCP_EXCEPTION_EXIT] = "QEMU v7M exception exit",
165
+ }
81
+ [EXCP_KERNEL_TRAP] = "QEMU intercept of kernel commpage",
166
+ /* asimd_imm_const() sorts out VMVNI vs VMOVI for us */
82
+ [EXCP_HVC] = "Hypervisor Call",
167
+ fn = gen_helper_mve_vmovi;
83
+ [EXCP_HYP_TRAP] = "Hypervisor Trap",
168
+ }
84
+ [EXCP_SMC] = "Secure Monitor Call",
169
+ return do_1imm(s, a, fn);
85
+ [EXCP_VIRQ] = "Virtual IRQ",
170
+}
86
+ [EXCP_VFIQ] = "Virtual FIQ",
87
+ [EXCP_SEMIHOST] = "Semihosting call",
88
+ [EXCP_NOCP] = "v7M NOCP UsageFault",
89
+ [EXCP_INVSTATE] = "v7M INVSTATE UsageFault",
90
+ };
91
92
if (idx >= 0 && idx < ARRAY_SIZE(excnames)) {
93
exc = excnames[idx];
94
--
171
--
95
2.7.4
172
2.20.1
96
173
97
174
diff view generated by jsdifflib
1
In tlb_fill() we construct a syndrome register value from a
1
Implement the MVE shift-vector-left-by-immediate insns VSHL, VQSHL
2
fault status register value which is filled in by arm_tlb_fill().
2
and VQSHLU.
3
arm_tlb_fill() returns FSR values which might be in the format
3
4
used with short-format page descriptors, or the format used
4
The size-and-immediate encoding here is the same as Neon, and we
5
with long-format (LPAE) descriptors. The syndrome register
5
handle it the same way neon-dp.decode does.
6
always uses LPAE-format FSR status codes.
7
8
It isn't actually possible to end up delivering a syndrome
9
register value to the guest for a fault which is reported
10
with a short-format FSR (that kind of stage 1 fault will only
11
happen for an AArch32 translation regime which doesn't have
12
a syndrome register, and can never be redirected to an AArch64
13
or Hyp exception level). Add an assertion which checks this,
14
and adjust the code so that we construct a syndrome with
15
an invalid status code, rather than allowing set bits in
16
the FSR input to randomly corrupt other fields in the syndrome.
17
6
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
20
Message-id: 1491486152-24304-1-git-send-email-peter.maydell@linaro.org
9
Message-id: 20210628135835.6690-8-peter.maydell@linaro.org
21
---
10
---
22
target/arm/op_helper.c | 23 ++++++++++++++++++-----
11
target/arm/helper-mve.h | 16 +++++++++++
23
1 file changed, 18 insertions(+), 5 deletions(-)
12
target/arm/mve.decode | 23 +++++++++++++++
24
13
target/arm/mve_helper.c | 57 ++++++++++++++++++++++++++++++++++++++
25
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
14
target/arm/translate-mve.c | 51 ++++++++++++++++++++++++++++++++++
26
index XXXXXXX..XXXXXXX 100644
15
4 files changed, 147 insertions(+)
27
--- a/target/arm/op_helper.c
16
28
+++ b/target/arm/op_helper.c
17
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
29
@@ -XXX,XX +XXX,XX @@ void tlb_fill(CPUState *cs, target_ulong addr, MMUAccessType access_type,
18
index XXXXXXX..XXXXXXX 100644
30
if (unlikely(ret)) {
19
--- a/target/arm/helper-mve.h
31
ARMCPU *cpu = ARM_CPU(cs);
20
+++ b/target/arm/helper-mve.h
32
CPUARMState *env = &cpu->env;
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_vaddvuw, TCG_CALL_NO_WG, i32, env, ptr, i32)
33
- uint32_t syn, exc;
22
DEF_HELPER_FLAGS_3(mve_vmovi, TCG_CALL_NO_WG, void, env, ptr, i64)
34
+ uint32_t syn, exc, fsc;
23
DEF_HELPER_FLAGS_3(mve_vandi, TCG_CALL_NO_WG, void, env, ptr, i64)
35
unsigned int target_el;
24
DEF_HELPER_FLAGS_3(mve_vorri, TCG_CALL_NO_WG, void, env, ptr, i64)
36
bool same_el;
25
+
37
26
+DEF_HELPER_FLAGS_4(mve_vshli_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
38
@@ -XXX,XX +XXX,XX @@ void tlb_fill(CPUState *cs, target_ulong addr, MMUAccessType access_type,
27
+DEF_HELPER_FLAGS_4(mve_vshli_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
39
env->cp15.hpfar_el2 = extract64(fi.s2addr, 12, 47) << 4;
28
+DEF_HELPER_FLAGS_4(mve_vshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
40
}
29
+
41
same_el = arm_current_el(env) == target_el;
30
+DEF_HELPER_FLAGS_4(mve_vqshli_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
42
- /* AArch64 syndrome does not have an LPAE bit */
31
+DEF_HELPER_FLAGS_4(mve_vqshli_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
43
- syn = fsr & ~(1 << 9);
32
+DEF_HELPER_FLAGS_4(mve_vqshli_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
44
+
33
+
45
+ if (fsr & (1 << 9)) {
34
+DEF_HELPER_FLAGS_4(mve_vqshli_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
46
+ /* LPAE format fault status register : bottom 6 bits are
35
+DEF_HELPER_FLAGS_4(mve_vqshli_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
47
+ * status code in the same form as needed for syndrome
36
+DEF_HELPER_FLAGS_4(mve_vqshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
48
+ */
37
+
49
+ fsc = extract32(fsr, 0, 6);
38
+DEF_HELPER_FLAGS_4(mve_vqshlui_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
50
+ } else {
39
+DEF_HELPER_FLAGS_4(mve_vqshlui_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
51
+ /* Short format FSR : this fault will never actually be reported
40
+DEF_HELPER_FLAGS_4(mve_vqshlui_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
52
+ * to an EL that uses a syndrome register. Check that here,
41
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
53
+ * and use a (currently) reserved FSR code in case the constructed
42
index XXXXXXX..XXXXXXX 100644
54
+ * syndrome does leak into the guest somehow.
43
--- a/target/arm/mve.decode
55
+ */
44
+++ b/target/arm/mve.decode
56
+ assert(target_el != 2 && !arm_el_is_aa64(env, target_el));
45
@@ -XXX,XX +XXX,XX @@
57
+ fsc = 0x3f;
46
&2op qd qm qn size
58
+ }
47
&2scalar qd qn rm size
59
48
&1imm qd imm cmode op
60
/* For insn and data aborts we assume there is no instruction syndrome
49
+&2shift qd qm shift size
61
* information; this is always true for exceptions reported to EL1.
50
62
*/
51
@vldr_vstr ....... . . . . l:1 rn:4 ... ...... imm:7 &vldr_vstr qd=%qd u=0
63
if (access_type == MMU_INST_FETCH) {
52
# Note that both Rn and Qd are 3 bits only (no D bit)
64
- syn = syn_insn_abort(same_el, 0, fi.s1ptw, syn);
53
@@ -XXX,XX +XXX,XX @@
65
+ syn = syn_insn_abort(same_el, 0, fi.s1ptw, fsc);
54
@2scalar .... .... .. size:2 .... .... .... .... rm:4 &2scalar qd=%qd qn=%qn
66
exc = EXCP_PREFETCH_ABORT;
55
@2scalar_nosz .... .... .... .... .... .... .... rm:4 &2scalar qd=%qd qn=%qn
67
} else {
56
68
syn = merge_syn_data_abort(env->exception.syndrome, target_el,
57
+@2_shl_b .... .... .. 001 shift:3 .... .... .... .... &2shift qd=%qd qm=%qm size=0
69
same_el, fi.s1ptw,
58
+@2_shl_h .... .... .. 01 shift:4 .... .... .... .... &2shift qd=%qd qm=%qm size=1
70
- access_type == MMU_DATA_STORE, syn);
59
+@2_shl_w .... .... .. 1 shift:5 .... .... .... .... &2shift qd=%qd qm=%qm size=2
71
+ access_type == MMU_DATA_STORE, fsc);
60
+
72
if (access_type == MMU_DATA_STORE
61
# Vector loads and stores
73
&& arm_feature(env, ARM_FEATURE_V6)) {
62
74
fsr |= (1 << 11);
63
# Widening loads and narrowing stores:
64
@@ -XXX,XX +XXX,XX @@ VPST 1111 1110 0 . 11 000 1 ... 0 1111 0100 1101 mask=%mask_22_13
65
# So we have a single decode line and check the cmode/op in the
66
# trans function.
67
Vimm_1r 111 . 1111 1 . 00 0 ... ... 0 .... 0 1 . 1 .... @1imm
68
+
69
+# Shifts by immediate
70
+
71
+VSHLI 111 0 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_b
72
+VSHLI 111 0 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_h
73
+VSHLI 111 0 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_w
74
+
75
+VQSHLI_S 111 0 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_b
76
+VQSHLI_S 111 0 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_h
77
+VQSHLI_S 111 0 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_w
78
+
79
+VQSHLI_U 111 1 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_b
80
+VQSHLI_U 111 1 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_h
81
+VQSHLI_U 111 1 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_w
82
+
83
+VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_b
84
+VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_h
85
+VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_w
86
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
87
index XXXXXXX..XXXXXXX 100644
88
--- a/target/arm/mve_helper.c
89
+++ b/target/arm/mve_helper.c
90
@@ -XXX,XX +XXX,XX @@ DO_2OP_SAT(vqsubsw, 4, int32_t, DO_SQSUB_W)
91
WRAP_QRSHL_HELPER(do_sqrshl_bhs, N, M, true, satp)
92
#define DO_UQRSHL_OP(N, M, satp) \
93
WRAP_QRSHL_HELPER(do_uqrshl_bhs, N, M, true, satp)
94
+#define DO_SUQSHL_OP(N, M, satp) \
95
+ WRAP_QRSHL_HELPER(do_suqrshl_bhs, N, M, false, satp)
96
97
DO_2OP_SAT_S(vqshls, DO_SQSHL_OP)
98
DO_2OP_SAT_U(vqshlu, DO_UQSHL_OP)
99
@@ -XXX,XX +XXX,XX @@ DO_VADDV(vaddvsw, 4, uint32_t)
100
DO_VADDV(vaddvub, 1, uint8_t)
101
DO_VADDV(vaddvuh, 2, uint16_t)
102
DO_VADDV(vaddvuw, 4, uint32_t)
103
+
104
+/* Shifts by immediate */
105
+#define DO_2SHIFT(OP, ESIZE, TYPE, FN) \
106
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
107
+ void *vm, uint32_t shift) \
108
+ { \
109
+ TYPE *d = vd, *m = vm; \
110
+ uint16_t mask = mve_element_mask(env); \
111
+ unsigned e; \
112
+ for (e = 0; e < 16 / ESIZE; e++, mask >>= ESIZE) { \
113
+ mergemask(&d[H##ESIZE(e)], \
114
+ FN(m[H##ESIZE(e)], shift), mask); \
115
+ } \
116
+ mve_advance_vpt(env); \
117
+ }
118
+
119
+#define DO_2SHIFT_SAT(OP, ESIZE, TYPE, FN) \
120
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
121
+ void *vm, uint32_t shift) \
122
+ { \
123
+ TYPE *d = vd, *m = vm; \
124
+ uint16_t mask = mve_element_mask(env); \
125
+ unsigned e; \
126
+ bool qc = false; \
127
+ for (e = 0; e < 16 / ESIZE; e++, mask >>= ESIZE) { \
128
+ bool sat = false; \
129
+ mergemask(&d[H##ESIZE(e)], \
130
+ FN(m[H##ESIZE(e)], shift, &sat), mask); \
131
+ qc |= sat & mask & 1; \
132
+ } \
133
+ if (qc) { \
134
+ env->vfp.qc[0] = qc; \
135
+ } \
136
+ mve_advance_vpt(env); \
137
+ }
138
+
139
+/* provide unsigned 2-op shift helpers for all sizes */
140
+#define DO_2SHIFT_U(OP, FN) \
141
+ DO_2SHIFT(OP##b, 1, uint8_t, FN) \
142
+ DO_2SHIFT(OP##h, 2, uint16_t, FN) \
143
+ DO_2SHIFT(OP##w, 4, uint32_t, FN)
144
+
145
+#define DO_2SHIFT_SAT_U(OP, FN) \
146
+ DO_2SHIFT_SAT(OP##b, 1, uint8_t, FN) \
147
+ DO_2SHIFT_SAT(OP##h, 2, uint16_t, FN) \
148
+ DO_2SHIFT_SAT(OP##w, 4, uint32_t, FN)
149
+#define DO_2SHIFT_SAT_S(OP, FN) \
150
+ DO_2SHIFT_SAT(OP##b, 1, int8_t, FN) \
151
+ DO_2SHIFT_SAT(OP##h, 2, int16_t, FN) \
152
+ DO_2SHIFT_SAT(OP##w, 4, int32_t, FN)
153
+
154
+DO_2SHIFT_U(vshli_u, DO_VSHLU)
155
+DO_2SHIFT_SAT_U(vqshli_u, DO_UQSHL_OP)
156
+DO_2SHIFT_SAT_S(vqshli_s, DO_SQSHL_OP)
157
+DO_2SHIFT_SAT_S(vqshlui_s, DO_SUQSHL_OP)
158
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
159
index XXXXXXX..XXXXXXX 100644
160
--- a/target/arm/translate-mve.c
161
+++ b/target/arm/translate-mve.c
162
@@ -XXX,XX +XXX,XX @@ typedef void MVEGenLdStFn(TCGv_ptr, TCGv_ptr, TCGv_i32);
163
typedef void MVEGenOneOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
164
typedef void MVEGenTwoOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_ptr);
165
typedef void MVEGenTwoOpScalarFn(TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_i32);
166
+typedef void MVEGenTwoOpShiftFn(TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_i32);
167
typedef void MVEGenDualAccOpFn(TCGv_i64, TCGv_ptr, TCGv_ptr, TCGv_ptr, TCGv_i64);
168
typedef void MVEGenVADDVFn(TCGv_i32, TCGv_ptr, TCGv_ptr, TCGv_i32);
169
typedef void MVEGenOneOpImmFn(TCGv_ptr, TCGv_ptr, TCGv_i64);
170
@@ -XXX,XX +XXX,XX @@ static bool trans_Vimm_1r(DisasContext *s, arg_1imm *a)
171
}
172
return do_1imm(s, a, fn);
173
}
174
+
175
+static bool do_2shift(DisasContext *s, arg_2shift *a, MVEGenTwoOpShiftFn fn,
176
+ bool negateshift)
177
+{
178
+ TCGv_ptr qd, qm;
179
+ int shift = a->shift;
180
+
181
+ if (!dc_isar_feature(aa32_mve, s) ||
182
+ !mve_check_qreg_bank(s, a->qd | a->qm) ||
183
+ !fn) {
184
+ return false;
185
+ }
186
+ if (!mve_eci_check(s) || !vfp_access_check(s)) {
187
+ return true;
188
+ }
189
+
190
+ /*
191
+ * When we handle a right shift insn using a left-shift helper
192
+ * which permits a negative shift count to indicate a right-shift,
193
+ * we must negate the shift count.
194
+ */
195
+ if (negateshift) {
196
+ shift = -shift;
197
+ }
198
+
199
+ qd = mve_qreg_ptr(a->qd);
200
+ qm = mve_qreg_ptr(a->qm);
201
+ fn(cpu_env, qd, qm, tcg_constant_i32(shift));
202
+ tcg_temp_free_ptr(qd);
203
+ tcg_temp_free_ptr(qm);
204
+ mve_update_eci(s);
205
+ return true;
206
+}
207
+
208
+#define DO_2SHIFT(INSN, FN, NEGATESHIFT) \
209
+ static bool trans_##INSN(DisasContext *s, arg_2shift *a) \
210
+ { \
211
+ static MVEGenTwoOpShiftFn * const fns[] = { \
212
+ gen_helper_mve_##FN##b, \
213
+ gen_helper_mve_##FN##h, \
214
+ gen_helper_mve_##FN##w, \
215
+ NULL, \
216
+ }; \
217
+ return do_2shift(s, a, fns[a->size], NEGATESHIFT); \
218
+ }
219
+
220
+DO_2SHIFT(VSHLI, vshli_u, false)
221
+DO_2SHIFT(VQSHLI_S, vqshli_s, false)
222
+DO_2SHIFT(VQSHLI_U, vqshli_u, false)
223
+DO_2SHIFT(VQSHLUI, vqshlui_s, false)
75
--
224
--
76
2.7.4
225
2.20.1
77
226
78
227
diff view generated by jsdifflib
1
Move the utility routines gen_set_condexec() and gen_set_pc_im()
1
Implement the MVE vector shift right by immediate insns VSHRI and
2
up in the file, as we will want to use them from a function
2
VRSHRI. As with Neon, we implement these by using helper functions
3
placed earlier in the file than their current location.
3
which perform left shifts but allow negative shift counts to indicate
4
right shifts.
4
5
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Richard Henderson <rth@twiddle.net>
8
Message-id: 20210628135835.6690-9-peter.maydell@linaro.org
8
Message-id: 1491844419-12485-5-git-send-email-peter.maydell@linaro.org
9
---
9
---
10
target/arm/translate.c | 31 +++++++++++++++----------------
10
target/arm/helper-mve.h | 12 ++++++++++++
11
1 file changed, 15 insertions(+), 16 deletions(-)
11
target/arm/translate.h | 20 ++++++++++++++++++++
12
target/arm/mve.decode | 28 ++++++++++++++++++++++++++++
13
target/arm/mve_helper.c | 7 +++++++
14
target/arm/translate-mve.c | 5 +++++
15
target/arm/translate-neon.c | 18 ------------------
16
6 files changed, 72 insertions(+), 18 deletions(-)
12
17
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
18
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
14
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/translate.c
20
--- a/target/arm/helper-mve.h
16
+++ b/target/arm/translate.c
21
+++ b/target/arm/helper-mve.h
17
@@ -XXX,XX +XXX,XX @@ static const uint8_t table_logic_cc[16] = {
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_vmovi, TCG_CALL_NO_WG, void, env, ptr, i64)
18
1, /* mvn */
23
DEF_HELPER_FLAGS_3(mve_vandi, TCG_CALL_NO_WG, void, env, ptr, i64)
19
};
24
DEF_HELPER_FLAGS_3(mve_vorri, TCG_CALL_NO_WG, void, env, ptr, i64)
20
25
21
+static inline void gen_set_condexec(DisasContext *s)
26
+DEF_HELPER_FLAGS_4(mve_vshli_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
+DEF_HELPER_FLAGS_4(mve_vshli_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
+DEF_HELPER_FLAGS_4(mve_vshli_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
29
+
30
DEF_HELPER_FLAGS_4(mve_vshli_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
31
DEF_HELPER_FLAGS_4(mve_vshli_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
32
DEF_HELPER_FLAGS_4(mve_vshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
33
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vqshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
34
DEF_HELPER_FLAGS_4(mve_vqshlui_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
35
DEF_HELPER_FLAGS_4(mve_vqshlui_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
36
DEF_HELPER_FLAGS_4(mve_vqshlui_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
37
+
38
+DEF_HELPER_FLAGS_4(mve_vrshli_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
39
+DEF_HELPER_FLAGS_4(mve_vrshli_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
40
+DEF_HELPER_FLAGS_4(mve_vrshli_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
41
+
42
+DEF_HELPER_FLAGS_4(mve_vrshli_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
43
+DEF_HELPER_FLAGS_4(mve_vrshli_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
44
+DEF_HELPER_FLAGS_4(mve_vrshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
45
diff --git a/target/arm/translate.h b/target/arm/translate.h
46
index XXXXXXX..XXXXXXX 100644
47
--- a/target/arm/translate.h
48
+++ b/target/arm/translate.h
49
@@ -XXX,XX +XXX,XX @@ static inline int times_2_plus_1(DisasContext *s, int x)
50
return x * 2 + 1;
51
}
52
53
+static inline int rsub_64(DisasContext *s, int x)
22
+{
54
+{
23
+ if (s->condexec_mask) {
55
+ return 64 - x;
24
+ uint32_t val = (s->condexec_cond << 4) | (s->condexec_mask >> 1);
25
+ TCGv_i32 tmp = tcg_temp_new_i32();
26
+ tcg_gen_movi_i32(tmp, val);
27
+ store_cpu_field(tmp, condexec_bits);
28
+ }
29
+}
56
+}
30
+
57
+
31
+static inline void gen_set_pc_im(DisasContext *s, target_ulong val)
58
+static inline int rsub_32(DisasContext *s, int x)
32
+{
59
+{
33
+ tcg_gen_movi_i32(cpu_R[15], val);
60
+ return 32 - x;
34
+}
61
+}
35
+
62
+
36
/* Set PC and Thumb state from an immediate address. */
63
+static inline int rsub_16(DisasContext *s, int x)
37
static inline void gen_bx_im(DisasContext *s, uint32_t addr)
64
+{
65
+ return 16 - x;
66
+}
67
+
68
+static inline int rsub_8(DisasContext *s, int x)
69
+{
70
+ return 8 - x;
71
+}
72
+
73
static inline int arm_dc_feature(DisasContext *dc, int feature)
38
{
74
{
39
@@ -XXX,XX +XXX,XX @@ DO_GEN_ST(8, MO_UB)
75
return (dc->features & (1ULL << feature)) != 0;
40
DO_GEN_ST(16, MO_UW)
76
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
41
DO_GEN_ST(32, MO_UL)
77
index XXXXXXX..XXXXXXX 100644
42
78
--- a/target/arm/mve.decode
43
-static inline void gen_set_pc_im(DisasContext *s, target_ulong val)
79
+++ b/target/arm/mve.decode
80
@@ -XXX,XX +XXX,XX @@
81
@2_shl_h .... .... .. 01 shift:4 .... .... .... .... &2shift qd=%qd qm=%qm size=1
82
@2_shl_w .... .... .. 1 shift:5 .... .... .... .... &2shift qd=%qd qm=%qm size=2
83
84
+# Right shifts are encoded as N - shift, where N is the element size in bits.
85
+%rshift_i5 16:5 !function=rsub_32
86
+%rshift_i4 16:4 !function=rsub_16
87
+%rshift_i3 16:3 !function=rsub_8
88
+
89
+@2_shr_b .... .... .. 001 ... .... .... .... .... &2shift qd=%qd qm=%qm \
90
+ size=0 shift=%rshift_i3
91
+@2_shr_h .... .... .. 01 .... .... .... .... .... &2shift qd=%qd qm=%qm \
92
+ size=1 shift=%rshift_i4
93
+@2_shr_w .... .... .. 1 ..... .... .... .... .... &2shift qd=%qd qm=%qm \
94
+ size=2 shift=%rshift_i5
95
+
96
# Vector loads and stores
97
98
# Widening loads and narrowing stores:
99
@@ -XXX,XX +XXX,XX @@ VQSHLI_U 111 1 1111 1 . ... ... ... 0 0111 0 1 . 1 ... 0 @2_shl_w
100
VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_b
101
VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_h
102
VQSHLUI 111 1 1111 1 . ... ... ... 0 0110 0 1 . 1 ... 0 @2_shl_w
103
+
104
+VSHRI_S 111 0 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_b
105
+VSHRI_S 111 0 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_h
106
+VSHRI_S 111 0 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_w
107
+
108
+VSHRI_U 111 1 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_b
109
+VSHRI_U 111 1 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_h
110
+VSHRI_U 111 1 1111 1 . ... ... ... 0 0000 0 1 . 1 ... 0 @2_shr_w
111
+
112
+VRSHRI_S 111 0 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_b
113
+VRSHRI_S 111 0 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_h
114
+VRSHRI_S 111 0 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_w
115
+
116
+VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_b
117
+VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_h
118
+VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_w
119
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
120
index XXXXXXX..XXXXXXX 100644
121
--- a/target/arm/mve_helper.c
122
+++ b/target/arm/mve_helper.c
123
@@ -XXX,XX +XXX,XX @@ DO_VADDV(vaddvuw, 4, uint32_t)
124
DO_2SHIFT(OP##b, 1, uint8_t, FN) \
125
DO_2SHIFT(OP##h, 2, uint16_t, FN) \
126
DO_2SHIFT(OP##w, 4, uint32_t, FN)
127
+#define DO_2SHIFT_S(OP, FN) \
128
+ DO_2SHIFT(OP##b, 1, int8_t, FN) \
129
+ DO_2SHIFT(OP##h, 2, int16_t, FN) \
130
+ DO_2SHIFT(OP##w, 4, int32_t, FN)
131
132
#define DO_2SHIFT_SAT_U(OP, FN) \
133
DO_2SHIFT_SAT(OP##b, 1, uint8_t, FN) \
134
@@ -XXX,XX +XXX,XX @@ DO_VADDV(vaddvuw, 4, uint32_t)
135
DO_2SHIFT_SAT(OP##w, 4, int32_t, FN)
136
137
DO_2SHIFT_U(vshli_u, DO_VSHLU)
138
+DO_2SHIFT_S(vshli_s, DO_VSHLS)
139
DO_2SHIFT_SAT_U(vqshli_u, DO_UQSHL_OP)
140
DO_2SHIFT_SAT_S(vqshli_s, DO_SQSHL_OP)
141
DO_2SHIFT_SAT_S(vqshlui_s, DO_SUQSHL_OP)
142
+DO_2SHIFT_U(vrshli_u, DO_VRSHLU)
143
+DO_2SHIFT_S(vrshli_s, DO_VRSHLS)
144
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
145
index XXXXXXX..XXXXXXX 100644
146
--- a/target/arm/translate-mve.c
147
+++ b/target/arm/translate-mve.c
148
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT(VSHLI, vshli_u, false)
149
DO_2SHIFT(VQSHLI_S, vqshli_s, false)
150
DO_2SHIFT(VQSHLI_U, vqshli_u, false)
151
DO_2SHIFT(VQSHLUI, vqshlui_s, false)
152
+/* These right shifts use a left-shift helper with negated shift count */
153
+DO_2SHIFT(VSHRI_S, vshli_s, true)
154
+DO_2SHIFT(VSHRI_U, vshli_u, true)
155
+DO_2SHIFT(VRSHRI_S, vrshli_s, true)
156
+DO_2SHIFT(VRSHRI_U, vrshli_u, true)
157
diff --git a/target/arm/translate-neon.c b/target/arm/translate-neon.c
158
index XXXXXXX..XXXXXXX 100644
159
--- a/target/arm/translate-neon.c
160
+++ b/target/arm/translate-neon.c
161
@@ -XXX,XX +XXX,XX @@ static inline int plus1(DisasContext *s, int x)
162
return x + 1;
163
}
164
165
-static inline int rsub_64(DisasContext *s, int x)
44
-{
166
-{
45
- tcg_gen_movi_i32(cpu_R[15], val);
167
- return 64 - x;
46
-}
168
-}
47
-
169
-
48
static inline void gen_hvc(DisasContext *s, int imm16)
170
-static inline int rsub_32(DisasContext *s, int x)
49
{
50
/* The pre HVC helper handles cases when HVC gets trapped
51
@@ -XXX,XX +XXX,XX @@ static inline void gen_smc(DisasContext *s)
52
s->is_jmp = DISAS_SMC;
53
}
54
55
-static inline void
56
-gen_set_condexec (DisasContext *s)
57
-{
171
-{
58
- if (s->condexec_mask) {
172
- return 32 - x;
59
- uint32_t val = (s->condexec_cond << 4) | (s->condexec_mask >> 1);
173
-}
60
- TCGv_i32 tmp = tcg_temp_new_i32();
174
-static inline int rsub_16(DisasContext *s, int x)
61
- tcg_gen_movi_i32(tmp, val);
175
-{
62
- store_cpu_field(tmp, condexec_bits);
176
- return 16 - x;
63
- }
177
-}
178
-static inline int rsub_8(DisasContext *s, int x)
179
-{
180
- return 8 - x;
64
-}
181
-}
65
-
182
-
66
static void gen_exception_internal_insn(DisasContext *s, int offset, int excp)
183
static inline int neon_3same_fp_size(DisasContext *s, int x)
67
{
184
{
68
gen_set_condexec(s);
185
/* Convert 0==fp32, 1==fp16 into a MO_* value */
69
--
186
--
70
2.7.4
187
2.20.1
71
188
72
189
diff view generated by jsdifflib
1
Move the code to generate the "condition failed" instruction
1
Implement the MVE VHLL (vector shift left long) insn. This has two
2
codepath out of the if (singlestepping) {} else {}. This
2
encodings: the T1 encoding is the usual shift-by-immediate format,
3
will allow adding support for handling a new is_jmp type
3
and the T2 encoding is a special case where the shift count is always
4
which can't be neatly split into "singlestepping case"
4
equal to the element size.
5
versus "not singlestepping case".
6
5
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Reviewed-by: Richard Henderson <rth@twiddle.net>
8
Message-id: 20210628135835.6690-10-peter.maydell@linaro.org
10
Message-id: 1491844419-12485-6-git-send-email-peter.maydell@linaro.org
11
---
9
---
12
target/arm/translate.c | 24 +++++++++++-------------
10
target/arm/helper-mve.h | 9 +++++++
13
1 file changed, 11 insertions(+), 13 deletions(-)
11
target/arm/mve.decode | 53 +++++++++++++++++++++++++++++++++++---
12
target/arm/mve_helper.c | 32 +++++++++++++++++++++++
13
target/arm/translate-mve.c | 15 +++++++++++
14
4 files changed, 105 insertions(+), 4 deletions(-)
14
15
15
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
16
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.c
18
--- a/target/arm/helper-mve.h
18
+++ b/target/arm/translate.c
19
+++ b/target/arm/helper-mve.h
19
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vrshli_sw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
20
/* At this stage dc->condjmp will only be set when the skipped
21
DEF_HELPER_FLAGS_4(mve_vrshli_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
21
instruction was a conditional branch or trap, and the PC has
22
DEF_HELPER_FLAGS_4(mve_vrshli_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
22
already been written. */
23
DEF_HELPER_FLAGS_4(mve_vrshli_uw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
23
+ gen_set_condexec(dc);
24
+
24
if (unlikely(cs->singlestep_enabled || dc->ss_active)) {
25
+DEF_HELPER_FLAGS_4(mve_vshllbsb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
25
/* Unconditional and "condition passed" instruction codepath. */
26
+DEF_HELPER_FLAGS_4(mve_vshllbsh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
26
- gen_set_condexec(dc);
27
+DEF_HELPER_FLAGS_4(mve_vshllbub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
switch (dc->is_jmp) {
28
+DEF_HELPER_FLAGS_4(mve_vshllbuh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
case DISAS_SWI:
29
+DEF_HELPER_FLAGS_4(mve_vshlltsb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
29
gen_ss_advance(dc);
30
+DEF_HELPER_FLAGS_4(mve_vshlltsh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
30
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
31
+DEF_HELPER_FLAGS_4(mve_vshlltub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
31
/* FIXME: Single stepping a WFI insn will not halt the CPU. */
32
+DEF_HELPER_FLAGS_4(mve_vshlltuh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
32
gen_singlestep_exception(dc);
33
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
33
}
34
index XXXXXXX..XXXXXXX 100644
34
- if (dc->condjmp) {
35
--- a/target/arm/mve.decode
35
- /* "Condition failed" instruction codepath. */
36
+++ b/target/arm/mve.decode
36
- gen_set_label(dc->condlabel);
37
@@ -XXX,XX +XXX,XX @@
37
- gen_set_condexec(dc);
38
@2_shl_h .... .... .. 01 shift:4 .... .... .... .... &2shift qd=%qd qm=%qm size=1
38
- gen_set_pc_im(dc, dc->pc);
39
@2_shl_w .... .... .. 1 shift:5 .... .... .... .... &2shift qd=%qd qm=%qm size=2
39
- gen_singlestep_exception(dc);
40
40
- }
41
+@2_shll_b .... .... ... 01 shift:3 .... .... .... .... &2shift qd=%qd qm=%qm size=0
41
} else {
42
+@2_shll_h .... .... ... 1 shift:4 .... .... .... .... &2shift qd=%qd qm=%qm size=1
42
/* While branches must always occur at the end of an IT block,
43
+# VSHLL encoding T2 where shift == esize
43
there are a few other things that can cause us to terminate
44
+@2_shll_esize_b .... .... .... 00 .. .... .... .... .... &2shift \
44
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
45
+ qd=%qd qm=%qm size=0 shift=8
45
- Hardware watchpoints.
46
+@2_shll_esize_h .... .... .... 01 .. .... .... .... .... &2shift \
46
Hardware breakpoints have already been handled and skip this code.
47
+ qd=%qd qm=%qm size=1 shift=16
47
*/
48
+
48
- gen_set_condexec(dc);
49
# Right shifts are encoded as N - shift, where N is the element size in bits.
49
switch(dc->is_jmp) {
50
%rshift_i5 16:5 !function=rsub_32
50
case DISAS_NEXT:
51
%rshift_i4 16:4 !function=rsub_16
51
gen_goto_tb(dc, 1, dc->pc);
52
@@ -XXX,XX +XXX,XX @@ VADD 1110 1111 0 . .. ... 0 ... 0 1000 . 1 . 0 ... 0 @2op
52
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
53
VSUB 1111 1111 0 . .. ... 0 ... 0 1000 . 1 . 0 ... 0 @2op
53
gen_exception(EXCP_SMC, syn_aa32_smc(), 3);
54
VMUL 1110 1111 0 . .. ... 0 ... 0 1001 . 1 . 1 ... 0 @2op
54
break;
55
55
}
56
-VMULH_S 111 0 1110 0 . .. ...1 ... 0 1110 . 0 . 0 ... 1 @2op
56
- if (dc->condjmp) {
57
-VMULH_U 111 1 1110 0 . .. ...1 ... 0 1110 . 0 . 0 ... 1 @2op
57
- gen_set_label(dc->condlabel);
58
+# The VSHLL T2 encoding is not a @2op pattern, but is here because it
58
- gen_set_condexec(dc);
59
+# overlaps what would be size=0b11 VMULH/VRMULH
60
+{
61
+ VSHLL_BS 111 0 1110 0 . 11 .. 01 ... 0 1110 0 0 . 0 ... 1 @2_shll_esize_b
62
+ VSHLL_BS 111 0 1110 0 . 11 .. 01 ... 0 1110 0 0 . 0 ... 1 @2_shll_esize_h
63
64
-VRMULH_S 111 0 1110 0 . .. ...1 ... 1 1110 . 0 . 0 ... 1 @2op
65
-VRMULH_U 111 1 1110 0 . .. ...1 ... 1 1110 . 0 . 0 ... 1 @2op
66
+ VMULH_S 111 0 1110 0 . .. ...1 ... 0 1110 . 0 . 0 ... 1 @2op
67
+}
68
+
69
+{
70
+ VSHLL_BU 111 1 1110 0 . 11 .. 01 ... 0 1110 0 0 . 0 ... 1 @2_shll_esize_b
71
+ VSHLL_BU 111 1 1110 0 . 11 .. 01 ... 0 1110 0 0 . 0 ... 1 @2_shll_esize_h
72
+
73
+ VMULH_U 111 1 1110 0 . .. ...1 ... 0 1110 . 0 . 0 ... 1 @2op
74
+}
75
+
76
+{
77
+ VSHLL_TS 111 0 1110 0 . 11 .. 01 ... 1 1110 0 0 . 0 ... 1 @2_shll_esize_b
78
+ VSHLL_TS 111 0 1110 0 . 11 .. 01 ... 1 1110 0 0 . 0 ... 1 @2_shll_esize_h
79
+
80
+ VRMULH_S 111 0 1110 0 . .. ...1 ... 1 1110 . 0 . 0 ... 1 @2op
81
+}
82
+
83
+{
84
+ VSHLL_TU 111 1 1110 0 . 11 .. 01 ... 1 1110 0 0 . 0 ... 1 @2_shll_esize_b
85
+ VSHLL_TU 111 1 1110 0 . 11 .. 01 ... 1 1110 0 0 . 0 ... 1 @2_shll_esize_h
86
+
87
+ VRMULH_U 111 1 1110 0 . .. ...1 ... 1 1110 . 0 . 0 ... 1 @2op
88
+}
89
90
VMAX_S 111 0 1111 0 . .. ... 0 ... 0 0110 . 1 . 0 ... 0 @2op
91
VMAX_U 111 1 1111 0 . .. ... 0 ... 0 0110 . 1 . 0 ... 0 @2op
92
@@ -XXX,XX +XXX,XX @@ VRSHRI_S 111 0 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_w
93
VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_b
94
VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_h
95
VRSHRI_U 111 1 1111 1 . ... ... ... 0 0010 0 1 . 1 ... 0 @2_shr_w
96
+
97
+# VSHLL T1 encoding; the T2 VSHLL encoding is elsewhere in this file
98
+VSHLL_BS 111 0 1110 1 . 1 .. ... ... 0 1111 0 1 . 0 ... 0 @2_shll_b
99
+VSHLL_BS 111 0 1110 1 . 1 .. ... ... 0 1111 0 1 . 0 ... 0 @2_shll_h
100
+
101
+VSHLL_BU 111 1 1110 1 . 1 .. ... ... 0 1111 0 1 . 0 ... 0 @2_shll_b
102
+VSHLL_BU 111 1 1110 1 . 1 .. ... ... 0 1111 0 1 . 0 ... 0 @2_shll_h
103
+
104
+VSHLL_TS 111 0 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_b
105
+VSHLL_TS 111 0 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_h
106
+
107
+VSHLL_TU 111 1 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_b
108
+VSHLL_TU 111 1 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_h
109
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
110
index XXXXXXX..XXXXXXX 100644
111
--- a/target/arm/mve_helper.c
112
+++ b/target/arm/mve_helper.c
113
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT_SAT_S(vqshli_s, DO_SQSHL_OP)
114
DO_2SHIFT_SAT_S(vqshlui_s, DO_SUQSHL_OP)
115
DO_2SHIFT_U(vrshli_u, DO_VRSHLU)
116
DO_2SHIFT_S(vrshli_s, DO_VRSHLS)
117
+
118
+/*
119
+ * Long shifts taking half-sized inputs from top or bottom of the input
120
+ * vector and producing a double-width result. ESIZE, TYPE are for
121
+ * the input, and LESIZE, LTYPE for the output.
122
+ * Unlike the normal shift helpers, we do not handle negative shift counts,
123
+ * because the long shift is strictly left-only.
124
+ */
125
+#define DO_VSHLL(OP, TOP, ESIZE, TYPE, LESIZE, LTYPE) \
126
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
127
+ void *vm, uint32_t shift) \
128
+ { \
129
+ LTYPE *d = vd; \
130
+ TYPE *m = vm; \
131
+ uint16_t mask = mve_element_mask(env); \
132
+ unsigned le; \
133
+ assert(shift <= 16); \
134
+ for (le = 0; le < 16 / LESIZE; le++, mask >>= LESIZE) { \
135
+ LTYPE r = (LTYPE)m[H##ESIZE(le * 2 + TOP)] << shift; \
136
+ mergemask(&d[H##LESIZE(le)], r, mask); \
137
+ } \
138
+ mve_advance_vpt(env); \
59
+ }
139
+ }
60
+
140
+
61
+ if (dc->condjmp) {
141
+#define DO_VSHLL_ALL(OP, TOP) \
62
+ /* "Condition failed" instruction codepath for the branch/trap insn */
142
+ DO_VSHLL(OP##sb, TOP, 1, int8_t, 2, int16_t) \
63
+ gen_set_label(dc->condlabel);
143
+ DO_VSHLL(OP##ub, TOP, 1, uint8_t, 2, uint16_t) \
64
+ gen_set_condexec(dc);
144
+ DO_VSHLL(OP##sh, TOP, 2, int16_t, 4, int32_t) \
65
+ if (unlikely(cs->singlestep_enabled || dc->ss_active)) {
145
+ DO_VSHLL(OP##uh, TOP, 2, uint16_t, 4, uint32_t) \
66
+ gen_set_pc_im(dc, dc->pc);
146
+
67
+ gen_singlestep_exception(dc);
147
+DO_VSHLL_ALL(vshllb, false)
68
+ } else {
148
+DO_VSHLL_ALL(vshllt, true)
69
gen_goto_tb(dc, 1, dc->pc);
149
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
70
- dc->condjmp = 0;
150
index XXXXXXX..XXXXXXX 100644
71
}
151
--- a/target/arm/translate-mve.c
72
}
152
+++ b/target/arm/translate-mve.c
73
153
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT(VSHRI_S, vshli_s, true)
154
DO_2SHIFT(VSHRI_U, vshli_u, true)
155
DO_2SHIFT(VRSHRI_S, vrshli_s, true)
156
DO_2SHIFT(VRSHRI_U, vrshli_u, true)
157
+
158
+#define DO_VSHLL(INSN, FN) \
159
+ static bool trans_##INSN(DisasContext *s, arg_2shift *a) \
160
+ { \
161
+ static MVEGenTwoOpShiftFn * const fns[] = { \
162
+ gen_helper_mve_##FN##b, \
163
+ gen_helper_mve_##FN##h, \
164
+ }; \
165
+ return do_2shift(s, a, fns[a->size], false); \
166
+ }
167
+
168
+DO_VSHLL(VSHLL_BS, vshllbs)
169
+DO_VSHLL(VSHLL_BU, vshllbu)
170
+DO_VSHLL(VSHLL_TS, vshllts)
171
+DO_VSHLL(VSHLL_TU, vshlltu)
74
--
172
--
75
2.7.4
173
2.20.1
76
174
77
175
diff view generated by jsdifflib
1
From: Alistair Francis <alistair.francis@xilinx.com>
1
Implement the MVE VSRI and VSLI insns, which perform a
2
shift-and-insert operation.
2
3
3
This patch fixes two mistakes in the interrupt logic.
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20210628135835.6690-11-peter.maydell@linaro.org
7
---
8
target/arm/helper-mve.h | 8 ++++++++
9
target/arm/mve.decode | 9 ++++++++
10
target/arm/mve_helper.c | 42 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-mve.c | 3 +++
12
4 files changed, 62 insertions(+)
4
13
5
First we only trigger single-queue or multi-queue interrupts if the status
14
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
6
register is set. This logic was already used for non multi-queue interrupts
7
but it also applies to multi-queue interrupts.
8
9
Secondly we need to lower the interrupts if the ISR isn't set. As part
10
of this we can remove the other interrupt lowering logic and consolidate
11
it inside gem_update_int_status().
12
13
Signed-off-by: Alistair Francis <alistair.francis@xilinx.com>
14
Message-id: 438bcc014f8f8a2f8f68f322cb6a53f4c04688c2.1491947224.git.alistair.francis@xilinx.com
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
---
18
hw/net/cadence_gem.c | 18 +++++++++++++-----
19
1 file changed, 13 insertions(+), 5 deletions(-)
20
21
diff --git a/hw/net/cadence_gem.c b/hw/net/cadence_gem.c
22
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
23
--- a/hw/net/cadence_gem.c
16
--- a/target/arm/helper-mve.h
24
+++ b/hw/net/cadence_gem.c
17
+++ b/target/arm/helper-mve.h
25
@@ -XXX,XX +XXX,XX @@ static void gem_update_int_status(CadenceGEMState *s)
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vshlltsb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
26
{
19
DEF_HELPER_FLAGS_4(mve_vshlltsh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
int i;
20
DEF_HELPER_FLAGS_4(mve_vshlltub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
21
DEF_HELPER_FLAGS_4(mve_vshlltuh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
29
- if ((s->num_priority_queues == 1) && s->regs[GEM_ISR]) {
22
+
30
+ if (!s->regs[GEM_ISR]) {
23
+DEF_HELPER_FLAGS_4(mve_vsrib, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
31
+ /* ISR isn't set, clear all the interrupts */
24
+DEF_HELPER_FLAGS_4(mve_vsrih, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
32
+ for (i = 0; i < s->num_priority_queues; ++i) {
25
+DEF_HELPER_FLAGS_4(mve_vsriw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
33
+ qemu_set_irq(s->irq[i], 0);
26
+
34
+ }
27
+DEF_HELPER_FLAGS_4(mve_vslib, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
35
+ return;
28
+DEF_HELPER_FLAGS_4(mve_vslih, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
29
+DEF_HELPER_FLAGS_4(mve_vsliw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
30
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
31
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/mve.decode
33
+++ b/target/arm/mve.decode
34
@@ -XXX,XX +XXX,XX @@ VSHLL_TS 111 0 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_h
35
36
VSHLL_TU 111 1 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_b
37
VSHLL_TU 111 1 1110 1 . 1 .. ... ... 1 1111 0 1 . 0 ... 0 @2_shll_h
38
+
39
+# Shift-and-insert
40
+VSRI 111 1 1111 1 . ... ... ... 0 0100 0 1 . 1 ... 0 @2_shr_b
41
+VSRI 111 1 1111 1 . ... ... ... 0 0100 0 1 . 1 ... 0 @2_shr_h
42
+VSRI 111 1 1111 1 . ... ... ... 0 0100 0 1 . 1 ... 0 @2_shr_w
43
+
44
+VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_b
45
+VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_h
46
+VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_w
47
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/mve_helper.c
50
+++ b/target/arm/mve_helper.c
51
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT_SAT_S(vqshlui_s, DO_SUQSHL_OP)
52
DO_2SHIFT_U(vrshli_u, DO_VRSHLU)
53
DO_2SHIFT_S(vrshli_s, DO_VRSHLS)
54
55
+/* Shift-and-insert; we always work with 64 bits at a time */
56
+#define DO_2SHIFT_INSERT(OP, ESIZE, SHIFTFN, MASKFN) \
57
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
58
+ void *vm, uint32_t shift) \
59
+ { \
60
+ uint64_t *d = vd, *m = vm; \
61
+ uint16_t mask; \
62
+ uint64_t shiftmask; \
63
+ unsigned e; \
64
+ if (shift == 0 || shift == ESIZE * 8) { \
65
+ /* \
66
+ * Only VSLI can shift by 0; only VSRI can shift by <dt>. \
67
+ * The generic logic would give the right answer for 0 but \
68
+ * fails for <dt>. \
69
+ */ \
70
+ goto done; \
71
+ } \
72
+ assert(shift < ESIZE * 8); \
73
+ mask = mve_element_mask(env); \
74
+ /* ESIZE / 2 gives the MO_* value if ESIZE is in [1,2,4] */ \
75
+ shiftmask = dup_const(ESIZE / 2, MASKFN(ESIZE * 8, shift)); \
76
+ for (e = 0; e < 16 / 8; e++, mask >>= 8) { \
77
+ uint64_t r = (SHIFTFN(m[H8(e)], shift) & shiftmask) | \
78
+ (d[H8(e)] & ~shiftmask); \
79
+ mergemask(&d[H8(e)], r, mask); \
80
+ } \
81
+done: \
82
+ mve_advance_vpt(env); \
36
+ }
83
+ }
37
+
84
+
38
+ /* If we get here we know s->regs[GEM_ISR] is set, so we don't need to
85
+#define DO_SHL(N, SHIFT) ((N) << (SHIFT))
39
+ * check it again.
86
+#define DO_SHR(N, SHIFT) ((N) >> (SHIFT))
40
+ */
87
+#define SHL_MASK(EBITS, SHIFT) MAKE_64BIT_MASK((SHIFT), (EBITS) - (SHIFT))
41
+ if (s->num_priority_queues == 1) {
88
+#define SHR_MASK(EBITS, SHIFT) MAKE_64BIT_MASK(0, (EBITS) - (SHIFT))
42
/* No priority queues, just trigger the interrupt */
89
+
43
DB_PRINT("asserting int.\n");
90
+DO_2SHIFT_INSERT(vsrib, 1, DO_SHR, SHR_MASK)
44
qemu_set_irq(s->irq[0], 1);
91
+DO_2SHIFT_INSERT(vsrih, 2, DO_SHR, SHR_MASK)
45
@@ -XXX,XX +XXX,XX @@ static uint64_t gem_read(void *opaque, hwaddr offset, unsigned size)
92
+DO_2SHIFT_INSERT(vsriw, 4, DO_SHR, SHR_MASK)
46
{
93
+DO_2SHIFT_INSERT(vslib, 1, DO_SHL, SHL_MASK)
47
CadenceGEMState *s;
94
+DO_2SHIFT_INSERT(vslih, 2, DO_SHL, SHL_MASK)
48
uint32_t retval;
95
+DO_2SHIFT_INSERT(vsliw, 4, DO_SHL, SHL_MASK)
49
- int i;
96
+
50
s = (CadenceGEMState *)opaque;
97
/*
51
98
* Long shifts taking half-sized inputs from top or bottom of the input
52
offset >>= 2;
99
* vector and producing a double-width result. ESIZE, TYPE are for
53
@@ -XXX,XX +XXX,XX @@ static uint64_t gem_read(void *opaque, hwaddr offset, unsigned size)
100
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
54
switch (offset) {
101
index XXXXXXX..XXXXXXX 100644
55
case GEM_ISR:
102
--- a/target/arm/translate-mve.c
56
DB_PRINT("lowering irqs on ISR read\n");
103
+++ b/target/arm/translate-mve.c
57
- for (i = 0; i < s->num_priority_queues; ++i) {
104
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT(VSHRI_U, vshli_u, true)
58
- qemu_set_irq(s->irq[i], 0);
105
DO_2SHIFT(VRSHRI_S, vrshli_s, true)
59
- }
106
DO_2SHIFT(VRSHRI_U, vrshli_u, true)
60
+ /* The interrupts get updated at the end of the function. */
107
61
break;
108
+DO_2SHIFT(VSRI, vsri, false)
62
case GEM_PHYMNTNC:
109
+DO_2SHIFT(VSLI, vsli, false)
63
if (retval & GEM_PHYMNTNC_OP_R) {
110
+
111
#define DO_VSHLL(INSN, FN) \
112
static bool trans_##INSN(DisasContext *s, arg_2shift *a) \
113
{ \
64
--
114
--
65
2.7.4
115
2.20.1
66
116
67
117
diff view generated by jsdifflib
1
We currently have two places that do:
1
Implement the MVE shift-right-and-narrow insn VSHRN and VRSHRN.
2
if (dc->ss_active) {
3
gen_step_complete_exception(dc);
4
} else {
5
gen_exception_internal(EXCP_DEBUG);
6
}
7
2
8
Factor this out into its own function, as we're about to add
3
do_urshr() is borrowed from sve_helper.c.
9
a third place that needs the same logic.
10
4
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Reviewed-by: Richard Henderson <rth@twiddle.net>
7
Message-id: 20210628135835.6690-12-peter.maydell@linaro.org
14
Message-id: 1491844419-12485-4-git-send-email-peter.maydell@linaro.org
15
---
8
---
16
target/arm/translate.c | 28 ++++++++++++++++------------
9
target/arm/helper-mve.h | 10 ++++++++++
17
1 file changed, 16 insertions(+), 12 deletions(-)
10
target/arm/mve.decode | 11 +++++++++++
11
target/arm/mve_helper.c | 40 ++++++++++++++++++++++++++++++++++++++
12
target/arm/translate-mve.c | 15 ++++++++++++++
13
4 files changed, 76 insertions(+)
18
14
19
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
20
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/translate.c
17
--- a/target/arm/helper-mve.h
22
+++ b/target/arm/translate.c
18
+++ b/target/arm/helper-mve.h
23
@@ -XXX,XX +XXX,XX @@ static void gen_step_complete_exception(DisasContext *s)
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vsriw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
24
s->is_jmp = DISAS_EXC;
20
DEF_HELPER_FLAGS_4(mve_vslib, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
25
}
21
DEF_HELPER_FLAGS_4(mve_vslih, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
26
22
DEF_HELPER_FLAGS_4(mve_vsliw, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
+static void gen_singlestep_exception(DisasContext *s)
23
+
24
+DEF_HELPER_FLAGS_4(mve_vshrnbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
25
+DEF_HELPER_FLAGS_4(mve_vshrnbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
26
+DEF_HELPER_FLAGS_4(mve_vshrntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
+DEF_HELPER_FLAGS_4(mve_vshrnth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
+
29
+DEF_HELPER_FLAGS_4(mve_vrshrnbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
30
+DEF_HELPER_FLAGS_4(mve_vrshrnbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
31
+DEF_HELPER_FLAGS_4(mve_vrshrntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
32
+DEF_HELPER_FLAGS_4(mve_vrshrnth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
33
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
34
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/mve.decode
36
+++ b/target/arm/mve.decode
37
@@ -XXX,XX +XXX,XX @@ VSRI 111 1 1111 1 . ... ... ... 0 0100 0 1 . 1 ... 0 @2_shr_w
38
VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_b
39
VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_h
40
VSLI 111 1 1111 1 . ... ... ... 0 0101 0 1 . 1 ... 0 @2_shl_w
41
+
42
+# Narrowing shifts (which only support b and h sizes)
43
+VSHRNB 111 0 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_b
44
+VSHRNB 111 0 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_h
45
+VSHRNT 111 0 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_b
46
+VSHRNT 111 0 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_h
47
+
48
+VRSHRNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_b
49
+VRSHRNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_h
50
+VRSHRNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_b
51
+VRSHRNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_h
52
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
53
index XXXXXXX..XXXXXXX 100644
54
--- a/target/arm/mve_helper.c
55
+++ b/target/arm/mve_helper.c
56
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT_INSERT(vsliw, 4, DO_SHL, SHL_MASK)
57
58
DO_VSHLL_ALL(vshllb, false)
59
DO_VSHLL_ALL(vshllt, true)
60
+
61
+/*
62
+ * Narrowing right shifts, taking a double sized input, shifting it
63
+ * and putting the result in either the top or bottom half of the output.
64
+ * ESIZE, TYPE are the output, and LESIZE, LTYPE the input.
65
+ */
66
+#define DO_VSHRN(OP, TOP, ESIZE, TYPE, LESIZE, LTYPE, FN) \
67
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
68
+ void *vm, uint32_t shift) \
69
+ { \
70
+ LTYPE *m = vm; \
71
+ TYPE *d = vd; \
72
+ uint16_t mask = mve_element_mask(env); \
73
+ unsigned le; \
74
+ for (le = 0; le < 16 / LESIZE; le++, mask >>= LESIZE) { \
75
+ TYPE r = FN(m[H##LESIZE(le)], shift); \
76
+ mergemask(&d[H##ESIZE(le * 2 + TOP)], r, mask); \
77
+ } \
78
+ mve_advance_vpt(env); \
79
+ }
80
+
81
+#define DO_VSHRN_ALL(OP, FN) \
82
+ DO_VSHRN(OP##bb, false, 1, uint8_t, 2, uint16_t, FN) \
83
+ DO_VSHRN(OP##bh, false, 2, uint16_t, 4, uint32_t, FN) \
84
+ DO_VSHRN(OP##tb, true, 1, uint8_t, 2, uint16_t, FN) \
85
+ DO_VSHRN(OP##th, true, 2, uint16_t, 4, uint32_t, FN)
86
+
87
+static inline uint64_t do_urshr(uint64_t x, unsigned sh)
28
+{
88
+{
29
+ /* Generate the right kind of exception for singlestep, which is
89
+ if (likely(sh < 64)) {
30
+ * either the architectural singlestep or EXCP_DEBUG for QEMU's
90
+ return (x >> sh) + ((x >> (sh - 1)) & 1);
31
+ * gdb singlestepping.
91
+ } else if (sh == 64) {
32
+ */
92
+ return x >> 63;
33
+ if (s->ss_active) {
34
+ gen_step_complete_exception(s);
35
+ } else {
93
+ } else {
36
+ gen_exception_internal(EXCP_DEBUG);
94
+ return 0;
37
+ }
95
+ }
38
+}
96
+}
39
+
97
+
40
static void gen_smul_dual(TCGv_i32 a, TCGv_i32 b)
98
+DO_VSHRN_ALL(vshrn, DO_SHR)
41
{
99
+DO_VSHRN_ALL(vrshrn, do_urshr)
42
TCGv_i32 tmp1 = tcg_temp_new_i32();
100
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
43
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
101
index XXXXXXX..XXXXXXX 100644
44
gen_set_pc_im(dc, dc->pc);
102
--- a/target/arm/translate-mve.c
45
/* fall through */
103
+++ b/target/arm/translate-mve.c
46
default:
104
@@ -XXX,XX +XXX,XX @@ DO_VSHLL(VSHLL_BS, vshllbs)
47
- if (dc->ss_active) {
105
DO_VSHLL(VSHLL_BU, vshllbu)
48
- gen_step_complete_exception(dc);
106
DO_VSHLL(VSHLL_TS, vshllts)
49
- } else {
107
DO_VSHLL(VSHLL_TU, vshlltu)
50
- /* FIXME: Single stepping a WFI insn will not halt
108
+
51
- the CPU. */
109
+#define DO_2SHIFT_N(INSN, FN) \
52
- gen_exception_internal(EXCP_DEBUG);
110
+ static bool trans_##INSN(DisasContext *s, arg_2shift *a) \
53
- }
111
+ { \
54
+ /* FIXME: Single stepping a WFI insn will not halt the CPU. */
112
+ static MVEGenTwoOpShiftFn * const fns[] = { \
55
+ gen_singlestep_exception(dc);
113
+ gen_helper_mve_##FN##b, \
56
}
114
+ gen_helper_mve_##FN##h, \
57
if (dc->condjmp) {
115
+ }; \
58
/* "Condition failed" instruction codepath. */
116
+ return do_2shift(s, a, fns[a->size], false); \
59
gen_set_label(dc->condlabel);
117
+ }
60
gen_set_condexec(dc);
118
+
61
gen_set_pc_im(dc, dc->pc);
119
+DO_2SHIFT_N(VSHRNB, vshrnb)
62
- if (dc->ss_active) {
120
+DO_2SHIFT_N(VSHRNT, vshrnt)
63
- gen_step_complete_exception(dc);
121
+DO_2SHIFT_N(VRSHRNB, vrshrnb)
64
- } else {
122
+DO_2SHIFT_N(VRSHRNT, vrshrnt)
65
- gen_exception_internal(EXCP_DEBUG);
66
- }
67
+ gen_singlestep_exception(dc);
68
}
69
} else {
70
/* While branches must always occur at the end of an IT block,
71
--
123
--
72
2.7.4
124
2.20.1
73
125
74
126
diff view generated by jsdifflib
1
Current recommended style is to log a guest error on bad register
1
Implement the MVE saturating shift-right-and-narrow insns
2
accesses, not kill the whole system with hw_error(). Change the
2
VQSHRN, VQSHRUN, VQRSHRN and VQRSHRUN.
3
hw_error() calls to log as LOG_GUEST_ERROR or LOG_UNIMP or use
3
4
g_assert_not_reached() as appropriate.
4
do_srshr() is borrowed from sve_helper.c.
5
5
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 1491486314-25823-1-git-send-email-peter.maydell@linaro.org
8
Message-id: 20210628135835.6690-13-peter.maydell@linaro.org
9
---
9
---
10
hw/arm/stellaris.c | 60 +++++++++++++++++++++++++++++++++---------------------
10
target/arm/helper-mve.h | 30 +++++++++++
11
1 file changed, 37 insertions(+), 23 deletions(-)
11
target/arm/mve.decode | 28 ++++++++++
12
12
target/arm/mve_helper.c | 104 +++++++++++++++++++++++++++++++++++++
13
diff --git a/hw/arm/stellaris.c b/hw/arm/stellaris.c
13
target/arm/translate-mve.c | 12 +++++
14
index XXXXXXX..XXXXXXX 100644
14
4 files changed, 174 insertions(+)
15
--- a/hw/arm/stellaris.c
15
16
+++ b/hw/arm/stellaris.c
16
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
17
@@ -XXX,XX +XXX,XX @@ static void gptm_reload(gptm_state *s, int n, int reset)
17
index XXXXXXX..XXXXXXX 100644
18
} else if (s->mode[n] == 0xa) {
18
--- a/target/arm/helper-mve.h
19
/* PWM mode. Not implemented. */
19
+++ b/target/arm/helper-mve.h
20
} else {
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vrshrnbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
21
- hw_error("TODO: 16-bit timer mode 0x%x\n", s->mode[n]);
21
DEF_HELPER_FLAGS_4(mve_vrshrnbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
22
+ qemu_log_mask(LOG_UNIMP,
22
DEF_HELPER_FLAGS_4(mve_vrshrntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
23
+ "GPTM: 16-bit timer mode unimplemented: 0x%x\n",
23
DEF_HELPER_FLAGS_4(mve_vrshrnth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
24
+ s->mode[n]);
24
+
25
+ return;
25
+DEF_HELPER_FLAGS_4(mve_vqshrnb_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
26
}
26
+DEF_HELPER_FLAGS_4(mve_vqshrnb_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
27
s->tick[n] = tick;
27
+DEF_HELPER_FLAGS_4(mve_vqshrnt_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
timer_mod(s->timer[n], tick);
28
+DEF_HELPER_FLAGS_4(mve_vqshrnt_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
29
@@ -XXX,XX +XXX,XX @@ static void gptm_tick(void *opaque)
29
+
30
} else if (s->mode[n] == 0xa) {
30
+DEF_HELPER_FLAGS_4(mve_vqshrnb_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
31
/* PWM mode. Not implemented. */
31
+DEF_HELPER_FLAGS_4(mve_vqshrnb_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
32
} else {
32
+DEF_HELPER_FLAGS_4(mve_vqshrnt_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
33
- hw_error("TODO: 16-bit timer mode 0x%x\n", s->mode[n]);
33
+DEF_HELPER_FLAGS_4(mve_vqshrnt_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
34
+ qemu_log_mask(LOG_UNIMP,
34
+
35
+ "GPTM: 16-bit timer mode unimplemented: 0x%x\n",
35
+DEF_HELPER_FLAGS_4(mve_vqshrunbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
36
+ s->mode[n]);
36
+DEF_HELPER_FLAGS_4(mve_vqshrunbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
37
}
37
+DEF_HELPER_FLAGS_4(mve_vqshruntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
38
gptm_update_irq(s);
38
+DEF_HELPER_FLAGS_4(mve_vqshrunth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
39
}
39
+
40
@@ -XXX,XX +XXX,XX @@ static void gptm_write(void *opaque, hwaddr offset,
40
+DEF_HELPER_FLAGS_4(mve_vqrshrnb_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
41
s->match_prescale[0] = value;
41
+DEF_HELPER_FLAGS_4(mve_vqrshrnb_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
42
break;
42
+DEF_HELPER_FLAGS_4(mve_vqrshrnt_sb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
43
default:
43
+DEF_HELPER_FLAGS_4(mve_vqrshrnt_sh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
44
- hw_error("gptm_write: Bad offset 0x%x\n", (int)offset);
44
+
45
+ qemu_log_mask(LOG_GUEST_ERROR,
45
+DEF_HELPER_FLAGS_4(mve_vqrshrnb_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
46
+ "GPTM: read at bad offset 0x%x\n", (int)offset);
46
+DEF_HELPER_FLAGS_4(mve_vqrshrnb_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
47
}
47
+DEF_HELPER_FLAGS_4(mve_vqrshrnt_ub, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
48
gptm_update_irq(s);
48
+DEF_HELPER_FLAGS_4(mve_vqrshrnt_uh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
49
}
49
+
50
@@ -XXX,XX +XXX,XX @@ static int ssys_board_class(const ssys_state *s)
50
+DEF_HELPER_FLAGS_4(mve_vqrshrunbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
51
}
51
+DEF_HELPER_FLAGS_4(mve_vqrshrunbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
52
/* for unknown classes, fall through */
52
+DEF_HELPER_FLAGS_4(mve_vqrshruntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
53
default:
53
+DEF_HELPER_FLAGS_4(mve_vqrshrunth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
54
- hw_error("ssys_board_class: Unknown class 0x%08x\n", did0);
54
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
55
+ /* This can only happen if the hardwired constant did0 value
55
index XXXXXXX..XXXXXXX 100644
56
+ * in this board's stellaris_board_info struct is wrong.
56
--- a/target/arm/mve.decode
57
+ */
57
+++ b/target/arm/mve.decode
58
+ g_assert_not_reached();
58
@@ -XXX,XX +XXX,XX @@ VRSHRNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_b
59
VRSHRNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 1 @2_shr_h
60
VRSHRNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_b
61
VRSHRNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 1 @2_shr_h
62
+
63
+VQSHRNB_S 111 0 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 0 @2_shr_b
64
+VQSHRNB_S 111 0 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 0 @2_shr_h
65
+VQSHRNT_S 111 0 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 0 @2_shr_b
66
+VQSHRNT_S 111 0 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 0 @2_shr_h
67
+VQSHRNB_U 111 1 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 0 @2_shr_b
68
+VQSHRNB_U 111 1 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 0 @2_shr_h
69
+VQSHRNT_U 111 1 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 0 @2_shr_b
70
+VQSHRNT_U 111 1 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 0 @2_shr_h
71
+
72
+VQSHRUNB 111 0 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_b
73
+VQSHRUNB 111 0 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_h
74
+VQSHRUNT 111 0 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_b
75
+VQSHRUNT 111 0 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_h
76
+
77
+VQRSHRNB_S 111 0 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 1 @2_shr_b
78
+VQRSHRNB_S 111 0 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 1 @2_shr_h
79
+VQRSHRNT_S 111 0 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 1 @2_shr_b
80
+VQRSHRNT_S 111 0 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 1 @2_shr_h
81
+VQRSHRNB_U 111 1 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 1 @2_shr_b
82
+VQRSHRNB_U 111 1 1110 1 . ... ... ... 0 1111 0 1 . 0 ... 1 @2_shr_h
83
+VQRSHRNT_U 111 1 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 1 @2_shr_b
84
+VQRSHRNT_U 111 1 1110 1 . ... ... ... 1 1111 0 1 . 0 ... 1 @2_shr_h
85
+
86
+VQRSHRUNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_b
87
+VQRSHRUNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_h
88
+VQRSHRUNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_b
89
+VQRSHRUNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_h
90
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
91
index XXXXXXX..XXXXXXX 100644
92
--- a/target/arm/mve_helper.c
93
+++ b/target/arm/mve_helper.c
94
@@ -XXX,XX +XXX,XX @@ static inline uint64_t do_urshr(uint64_t x, unsigned sh)
59
}
95
}
60
}
96
}
61
97
62
@@ -XXX,XX +XXX,XX @@ static uint64_t ssys_read(void *opaque, hwaddr offset,
98
+static inline int64_t do_srshr(int64_t x, unsigned sh)
63
case DID0_CLASS_SANDSTORM:
99
+{
64
return pllcfg_sandstorm[xtal];
100
+ if (likely(sh < 64)) {
65
default:
101
+ return (x >> sh) + ((x >> (sh - 1)) & 1);
66
- hw_error("ssys_read: Unhandled class for PLLCFG read.\n");
102
+ } else {
67
- return 0;
103
+ /* Rounding the sign bit always produces 0. */
68
+ g_assert_not_reached();
104
+ return 0;
69
}
105
+ }
70
}
106
+}
71
case 0x070: /* RCC2 */
107
+
72
@@ -XXX,XX +XXX,XX @@ static uint64_t ssys_read(void *opaque, hwaddr offset,
108
DO_VSHRN_ALL(vshrn, DO_SHR)
73
case 0x1e4: /* USER1 */
109
DO_VSHRN_ALL(vrshrn, do_urshr)
74
return s->user1;
110
+
75
default:
111
+static inline int32_t do_sat_bhs(int64_t val, int64_t min, int64_t max,
76
- hw_error("ssys_read: Bad offset 0x%x\n", (int)offset);
112
+ bool *satp)
77
+ qemu_log_mask(LOG_GUEST_ERROR,
113
+{
78
+ "SSYS: read at bad offset 0x%x\n", (int)offset);
114
+ if (val > max) {
79
return 0;
115
+ *satp = true;
80
}
116
+ return max;
81
}
117
+ } else if (val < min) {
82
@@ -XXX,XX +XXX,XX @@ static void ssys_write(void *opaque, hwaddr offset,
118
+ *satp = true;
83
s->ldoarst = value;
119
+ return min;
84
break;
120
+ } else {
85
default:
121
+ return val;
86
- hw_error("ssys_write: Bad offset 0x%x\n", (int)offset);
122
+ }
87
+ qemu_log_mask(LOG_GUEST_ERROR,
123
+}
88
+ "SSYS: write at bad offset 0x%x\n", (int)offset);
124
+
89
}
125
+/* Saturating narrowing right shifts */
90
ssys_update(s);
126
+#define DO_VSHRN_SAT(OP, TOP, ESIZE, TYPE, LESIZE, LTYPE, FN) \
91
}
127
+ void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
92
@@ -XXX,XX +XXX,XX @@ static uint64_t stellaris_i2c_read(void *opaque, hwaddr offset,
128
+ void *vm, uint32_t shift) \
93
case 0x20: /* MCR */
129
+ { \
94
return s->mcr;
130
+ LTYPE *m = vm; \
95
default:
131
+ TYPE *d = vd; \
96
- hw_error("strllaris_i2c_read: Bad offset 0x%x\n", (int)offset);
132
+ uint16_t mask = mve_element_mask(env); \
97
+ qemu_log_mask(LOG_GUEST_ERROR,
133
+ bool qc = false; \
98
+ "stellaris_i2c: read at bad offset 0x%x\n", (int)offset);
134
+ unsigned le; \
99
return 0;
135
+ for (le = 0; le < 16 / LESIZE; le++, mask >>= LESIZE) { \
100
}
136
+ bool sat = false; \
101
}
137
+ TYPE r = FN(m[H##LESIZE(le)], shift, &sat); \
102
@@ -XXX,XX +XXX,XX @@ static void stellaris_i2c_write(void *opaque, hwaddr offset,
138
+ mergemask(&d[H##ESIZE(le * 2 + TOP)], r, mask); \
103
s->mris &= ~value;
139
+ qc |= sat && (mask & 1 << (TOP * ESIZE)); \
104
break;
140
+ } \
105
case 0x20: /* MCR */
141
+ if (qc) { \
106
- if (value & 1)
142
+ env->vfp.qc[0] = qc; \
107
- hw_error(
143
+ } \
108
- "stellaris_i2c_write: Loopback not implemented\n");
144
+ mve_advance_vpt(env); \
109
- if (value & 0x20)
145
+ }
110
- hw_error(
146
+
111
- "stellaris_i2c_write: Slave mode not implemented\n");
147
+#define DO_VSHRN_SAT_UB(BOP, TOP, FN) \
112
+ if (value & 1) {
148
+ DO_VSHRN_SAT(BOP, false, 1, uint8_t, 2, uint16_t, FN) \
113
+ qemu_log_mask(LOG_UNIMP, "stellaris_i2c: Loopback not implemented");
149
+ DO_VSHRN_SAT(TOP, true, 1, uint8_t, 2, uint16_t, FN)
114
+ }
150
+
115
+ if (value & 0x20) {
151
+#define DO_VSHRN_SAT_UH(BOP, TOP, FN) \
116
+ qemu_log_mask(LOG_UNIMP,
152
+ DO_VSHRN_SAT(BOP, false, 2, uint16_t, 4, uint32_t, FN) \
117
+ "stellaris_i2c: Slave mode not implemented");
153
+ DO_VSHRN_SAT(TOP, true, 2, uint16_t, 4, uint32_t, FN)
118
+ }
154
+
119
s->mcr = value & 0x31;
155
+#define DO_VSHRN_SAT_SB(BOP, TOP, FN) \
120
break;
156
+ DO_VSHRN_SAT(BOP, false, 1, int8_t, 2, int16_t, FN) \
121
default:
157
+ DO_VSHRN_SAT(TOP, true, 1, int8_t, 2, int16_t, FN)
122
- hw_error("stellaris_i2c_write: Bad offset 0x%x\n",
158
+
123
- (int)offset);
159
+#define DO_VSHRN_SAT_SH(BOP, TOP, FN) \
124
+ qemu_log_mask(LOG_GUEST_ERROR,
160
+ DO_VSHRN_SAT(BOP, false, 2, int16_t, 4, int32_t, FN) \
125
+ "stellaris_i2c: write at bad offset 0x%x\n", (int)offset);
161
+ DO_VSHRN_SAT(TOP, true, 2, int16_t, 4, int32_t, FN)
126
}
162
+
127
stellaris_i2c_update(s);
163
+#define DO_SHRN_SB(N, M, SATP) \
128
}
164
+ do_sat_bhs((int64_t)(N) >> (M), INT8_MIN, INT8_MAX, SATP)
129
@@ -XXX,XX +XXX,XX @@ static uint64_t stellaris_adc_read(void *opaque, hwaddr offset,
165
+#define DO_SHRN_UB(N, M, SATP) \
130
case 0x30: /* SAC */
166
+ do_sat_bhs((uint64_t)(N) >> (M), 0, UINT8_MAX, SATP)
131
return s->sac;
167
+#define DO_SHRUN_B(N, M, SATP) \
132
default:
168
+ do_sat_bhs((int64_t)(N) >> (M), 0, UINT8_MAX, SATP)
133
- hw_error("strllaris_adc_read: Bad offset 0x%x\n",
169
+
134
- (int)offset);
170
+#define DO_SHRN_SH(N, M, SATP) \
135
+ qemu_log_mask(LOG_GUEST_ERROR,
171
+ do_sat_bhs((int64_t)(N) >> (M), INT16_MIN, INT16_MAX, SATP)
136
+ "stellaris_adc: read at bad offset 0x%x\n", (int)offset);
172
+#define DO_SHRN_UH(N, M, SATP) \
137
return 0;
173
+ do_sat_bhs((uint64_t)(N) >> (M), 0, UINT16_MAX, SATP)
138
}
174
+#define DO_SHRUN_H(N, M, SATP) \
139
}
175
+ do_sat_bhs((int64_t)(N) >> (M), 0, UINT16_MAX, SATP)
140
@@ -XXX,XX +XXX,XX @@ static void stellaris_adc_write(void *opaque, hwaddr offset,
176
+
141
return;
177
+#define DO_RSHRN_SB(N, M, SATP) \
142
case 0x04: /* SSCTL */
178
+ do_sat_bhs(do_srshr(N, M), INT8_MIN, INT8_MAX, SATP)
143
if (value != 6) {
179
+#define DO_RSHRN_UB(N, M, SATP) \
144
- hw_error("ADC: Unimplemented sequence %" PRIx64 "\n",
180
+ do_sat_bhs(do_urshr(N, M), 0, UINT8_MAX, SATP)
145
- value);
181
+#define DO_RSHRUN_B(N, M, SATP) \
146
+ qemu_log_mask(LOG_UNIMP,
182
+ do_sat_bhs(do_srshr(N, M), 0, UINT8_MAX, SATP)
147
+ "ADC: Unimplemented sequence %" PRIx64 "\n",
183
+
148
+ value);
184
+#define DO_RSHRN_SH(N, M, SATP) \
149
}
185
+ do_sat_bhs(do_srshr(N, M), INT16_MIN, INT16_MAX, SATP)
150
s->ssctl[n] = value;
186
+#define DO_RSHRN_UH(N, M, SATP) \
151
return;
187
+ do_sat_bhs(do_urshr(N, M), 0, UINT16_MAX, SATP)
152
@@ -XXX,XX +XXX,XX @@ static void stellaris_adc_write(void *opaque, hwaddr offset,
188
+#define DO_RSHRUN_H(N, M, SATP) \
153
s->sspri = value;
189
+ do_sat_bhs(do_srshr(N, M), 0, UINT16_MAX, SATP)
154
break;
190
+
155
case 0x28: /* PSSI */
191
+DO_VSHRN_SAT_SB(vqshrnb_sb, vqshrnt_sb, DO_SHRN_SB)
156
- hw_error("Not implemented: ADC sample initiate\n");
192
+DO_VSHRN_SAT_SH(vqshrnb_sh, vqshrnt_sh, DO_SHRN_SH)
157
+ qemu_log_mask(LOG_UNIMP, "ADC: sample initiate unimplemented");
193
+DO_VSHRN_SAT_UB(vqshrnb_ub, vqshrnt_ub, DO_SHRN_UB)
158
break;
194
+DO_VSHRN_SAT_UH(vqshrnb_uh, vqshrnt_uh, DO_SHRN_UH)
159
case 0x30: /* SAC */
195
+DO_VSHRN_SAT_SB(vqshrunbb, vqshruntb, DO_SHRUN_B)
160
s->sac = value;
196
+DO_VSHRN_SAT_SH(vqshrunbh, vqshrunth, DO_SHRUN_H)
161
break;
197
+
162
default:
198
+DO_VSHRN_SAT_SB(vqrshrnb_sb, vqrshrnt_sb, DO_RSHRN_SB)
163
- hw_error("stellaris_adc_write: Bad offset 0x%x\n", (int)offset);
199
+DO_VSHRN_SAT_SH(vqrshrnb_sh, vqrshrnt_sh, DO_RSHRN_SH)
164
+ qemu_log_mask(LOG_GUEST_ERROR,
200
+DO_VSHRN_SAT_UB(vqrshrnb_ub, vqrshrnt_ub, DO_RSHRN_UB)
165
+ "stellaris_adc: write at bad offset 0x%x\n", (int)offset);
201
+DO_VSHRN_SAT_UH(vqrshrnb_uh, vqrshrnt_uh, DO_RSHRN_UH)
166
}
202
+DO_VSHRN_SAT_SB(vqrshrunbb, vqrshruntb, DO_RSHRUN_B)
167
stellaris_adc_update(s);
203
+DO_VSHRN_SAT_SH(vqrshrunbh, vqrshrunth, DO_RSHRUN_H)
168
}
204
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
205
index XXXXXXX..XXXXXXX 100644
206
--- a/target/arm/translate-mve.c
207
+++ b/target/arm/translate-mve.c
208
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT_N(VSHRNB, vshrnb)
209
DO_2SHIFT_N(VSHRNT, vshrnt)
210
DO_2SHIFT_N(VRSHRNB, vrshrnb)
211
DO_2SHIFT_N(VRSHRNT, vrshrnt)
212
+DO_2SHIFT_N(VQSHRNB_S, vqshrnb_s)
213
+DO_2SHIFT_N(VQSHRNT_S, vqshrnt_s)
214
+DO_2SHIFT_N(VQSHRNB_U, vqshrnb_u)
215
+DO_2SHIFT_N(VQSHRNT_U, vqshrnt_u)
216
+DO_2SHIFT_N(VQSHRUNB, vqshrunb)
217
+DO_2SHIFT_N(VQSHRUNT, vqshrunt)
218
+DO_2SHIFT_N(VQRSHRNB_S, vqrshrnb_s)
219
+DO_2SHIFT_N(VQRSHRNT_S, vqrshrnt_s)
220
+DO_2SHIFT_N(VQRSHRNB_U, vqrshrnb_u)
221
+DO_2SHIFT_N(VQRSHRNT_U, vqrshrnt_u)
222
+DO_2SHIFT_N(VQRSHRUNB, vqrshrunb)
223
+DO_2SHIFT_N(VQRSHRUNT, vqrshrunt)
169
--
224
--
170
2.7.4
225
2.20.1
171
226
172
227
diff view generated by jsdifflib
1
Recent changes have added new EXCP_ values to ARM but forgot
1
Implement the MVE VSHLC insn, which performs a shift left of the
2
to update the excnames[] array which is used to provide
2
entire vector with carry in bits provided from a general purpose
3
human-readable strings when printing information about the
3
register and carry out bits written back to that register.
4
exception for debug logging. Add the missing entries, and
5
add a comment to the list of #defines to help avoid the mistake
6
being repeated in future.
7
4
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Message-id: 20210628135835.6690-14-peter.maydell@linaro.org
11
Message-id: 1491486340-25988-1-git-send-email-peter.maydell@linaro.org
12
---
8
---
13
target/arm/cpu.h | 1 +
9
target/arm/helper-mve.h | 2 ++
14
target/arm/internals.h | 2 ++
10
target/arm/mve.decode | 2 ++
15
2 files changed, 3 insertions(+)
11
target/arm/mve_helper.c | 38 ++++++++++++++++++++++++++++++++++++++
12
target/arm/translate-mve.c | 30 ++++++++++++++++++++++++++++++
13
4 files changed, 72 insertions(+)
16
14
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
18
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
17
--- a/target/arm/helper-mve.h
20
+++ b/target/arm/cpu.h
18
+++ b/target/arm/helper-mve.h
21
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vqrshrunbb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
22
#define EXCP_SEMIHOST 16 /* semihosting call */
20
DEF_HELPER_FLAGS_4(mve_vqrshrunbh, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
23
#define EXCP_NOCP 17 /* v7M NOCP UsageFault */
21
DEF_HELPER_FLAGS_4(mve_vqrshruntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
24
#define EXCP_INVSTATE 18 /* v7M INVSTATE UsageFault */
22
DEF_HELPER_FLAGS_4(mve_vqrshrunth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
25
+/* NB: new EXCP_ defines should be added to the excnames[] array too */
23
+
26
24
+DEF_HELPER_FLAGS_4(mve_vshlc, TCG_CALL_NO_WG, i32, env, ptr, i32, i32)
27
#define ARMV7M_EXCP_RESET 1
25
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
28
#define ARMV7M_EXCP_NMI 2
29
diff --git a/target/arm/internals.h b/target/arm/internals.h
30
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/internals.h
27
--- a/target/arm/mve.decode
32
+++ b/target/arm/internals.h
28
+++ b/target/arm/mve.decode
33
@@ -XXX,XX +XXX,XX @@ static const char * const excnames[] = {
29
@@ -XXX,XX +XXX,XX @@ VQRSHRUNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_b
34
[EXCP_VIRQ] = "Virtual IRQ",
30
VQRSHRUNB 111 1 1110 1 . ... ... ... 0 1111 1 1 . 0 ... 0 @2_shr_h
35
[EXCP_VFIQ] = "Virtual FIQ",
31
VQRSHRUNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_b
36
[EXCP_SEMIHOST] = "Semihosting call",
32
VQRSHRUNT 111 1 1110 1 . ... ... ... 1 1111 1 1 . 0 ... 0 @2_shr_h
37
+ [EXCP_NOCP] = "v7M NOCP UsageFault",
33
+
38
+ [EXCP_INVSTATE] = "v7M INVSTATE UsageFault",
34
+VSHLC 111 0 1110 1 . 1 imm:5 ... 0 1111 1100 rdm:4 qd=%qd
39
};
35
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
40
36
index XXXXXXX..XXXXXXX 100644
41
/* Scale factor for generic timers, ie number of ns per tick.
37
--- a/target/arm/mve_helper.c
38
+++ b/target/arm/mve_helper.c
39
@@ -XXX,XX +XXX,XX @@ DO_VSHRN_SAT_UB(vqrshrnb_ub, vqrshrnt_ub, DO_RSHRN_UB)
40
DO_VSHRN_SAT_UH(vqrshrnb_uh, vqrshrnt_uh, DO_RSHRN_UH)
41
DO_VSHRN_SAT_SB(vqrshrunbb, vqrshruntb, DO_RSHRUN_B)
42
DO_VSHRN_SAT_SH(vqrshrunbh, vqrshrunth, DO_RSHRUN_H)
43
+
44
+uint32_t HELPER(mve_vshlc)(CPUARMState *env, void *vd, uint32_t rdm,
45
+ uint32_t shift)
46
+{
47
+ uint32_t *d = vd;
48
+ uint16_t mask = mve_element_mask(env);
49
+ unsigned e;
50
+ uint32_t r;
51
+
52
+ /*
53
+ * For each 32-bit element, we shift it left, bringing in the
54
+ * low 'shift' bits of rdm at the bottom. Bits shifted out at
55
+ * the top become the new rdm, if the predicate mask permits.
56
+ * The final rdm value is returned to update the register.
57
+ * shift == 0 here means "shift by 32 bits".
58
+ */
59
+ if (shift == 0) {
60
+ for (e = 0; e < 16 / 4; e++, mask >>= 4) {
61
+ r = rdm;
62
+ if (mask & 1) {
63
+ rdm = d[H4(e)];
64
+ }
65
+ mergemask(&d[H4(e)], r, mask);
66
+ }
67
+ } else {
68
+ uint32_t shiftmask = MAKE_64BIT_MASK(0, shift);
69
+
70
+ for (e = 0; e < 16 / 4; e++, mask >>= 4) {
71
+ r = (d[H4(e)] << shift) | (rdm & shiftmask);
72
+ if (mask & 1) {
73
+ rdm = d[H4(e)] >> (32 - shift);
74
+ }
75
+ mergemask(&d[H4(e)], r, mask);
76
+ }
77
+ }
78
+ mve_advance_vpt(env);
79
+ return rdm;
80
+}
81
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
82
index XXXXXXX..XXXXXXX 100644
83
--- a/target/arm/translate-mve.c
84
+++ b/target/arm/translate-mve.c
85
@@ -XXX,XX +XXX,XX @@ DO_2SHIFT_N(VQRSHRNB_U, vqrshrnb_u)
86
DO_2SHIFT_N(VQRSHRNT_U, vqrshrnt_u)
87
DO_2SHIFT_N(VQRSHRUNB, vqrshrunb)
88
DO_2SHIFT_N(VQRSHRUNT, vqrshrunt)
89
+
90
+static bool trans_VSHLC(DisasContext *s, arg_VSHLC *a)
91
+{
92
+ /*
93
+ * Whole Vector Left Shift with Carry. The carry is taken
94
+ * from a general purpose register and written back there.
95
+ * An imm of 0 means "shift by 32".
96
+ */
97
+ TCGv_ptr qd;
98
+ TCGv_i32 rdm;
99
+
100
+ if (!dc_isar_feature(aa32_mve, s) || !mve_check_qreg_bank(s, a->qd)) {
101
+ return false;
102
+ }
103
+ if (a->rdm == 13 || a->rdm == 15) {
104
+ /* CONSTRAINED UNPREDICTABLE: we UNDEF */
105
+ return false;
106
+ }
107
+ if (!mve_eci_check(s) || !vfp_access_check(s)) {
108
+ return true;
109
+ }
110
+
111
+ qd = mve_qreg_ptr(a->qd);
112
+ rdm = load_reg(s, a->rdm);
113
+ gen_helper_mve_vshlc(rdm, cpu_env, qd, rdm, tcg_constant_i32(a->imm));
114
+ store_reg(s, a->rdm, rdm);
115
+ tcg_temp_free_ptr(qd);
116
+ mve_update_eci(s);
117
+ return true;
118
+}
42
--
119
--
43
2.7.4
120
2.20.1
44
121
45
122
diff view generated by jsdifflib
1
From: Ard Biesheuvel <ard.biesheuvel@linaro.org>
1
Implement the MVE VADDLV insn; this is similar to VADDV, except
2
that it accumulates 32-bit elements into a 64-bit accumulator
3
stored in a pair of general-purpose registers.
2
4
3
The arm64 boot protocol stipulates that the kernel must be loaded
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
TEXT_OFFSET bytes beyond a 2 MB aligned base address, where TEXT_OFFSET
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
could be any 4 KB multiple between 0 and 2 MB, and whose value can be
7
Message-id: 20210628135835.6690-15-peter.maydell@linaro.org
6
found in the header of the Image file.
8
---
9
target/arm/helper-mve.h | 3 ++
10
target/arm/mve.decode | 6 +++-
11
target/arm/mve_helper.c | 19 ++++++++++++
12
target/arm/translate-mve.c | 63 ++++++++++++++++++++++++++++++++++++++
13
4 files changed, 90 insertions(+), 1 deletion(-)
7
14
8
So after attempts to load the arm64 kernel image as an ELF file or as a
15
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
9
U-Boot image have failed (both of which have their own way of specifying
10
the load offset), try to determine the TEXT_OFFSET from the image after
11
loading it but before mapping it as a ROM mapping into the guest address
12
space.
13
14
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
15
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
16
Message-id: 1489414630-21609-1-git-send-email-ard.biesheuvel@linaro.org
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
---
19
hw/arm/boot.c | 64 +++++++++++++++++++++++++++++++++++++++++++++++++----------
20
1 file changed, 53 insertions(+), 11 deletions(-)
21
22
diff --git a/hw/arm/boot.c b/hw/arm/boot.c
23
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
24
--- a/hw/arm/boot.c
17
--- a/target/arm/helper-mve.h
25
+++ b/hw/arm/boot.c
18
+++ b/target/arm/helper-mve.h
26
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_vaddvuh, TCG_CALL_NO_WG, i32, env, ptr, i32)
27
#define KERNEL_LOAD_ADDR 0x00010000
20
DEF_HELPER_FLAGS_3(mve_vaddvsw, TCG_CALL_NO_WG, i32, env, ptr, i32)
28
#define KERNEL64_LOAD_ADDR 0x00080000
21
DEF_HELPER_FLAGS_3(mve_vaddvuw, TCG_CALL_NO_WG, i32, env, ptr, i32)
29
22
30
+#define ARM64_TEXT_OFFSET_OFFSET 8
23
+DEF_HELPER_FLAGS_3(mve_vaddlv_s, TCG_CALL_NO_WG, i64, env, ptr, i64)
31
+#define ARM64_MAGIC_OFFSET 56
24
+DEF_HELPER_FLAGS_3(mve_vaddlv_u, TCG_CALL_NO_WG, i64, env, ptr, i64)
32
+
25
+
33
typedef enum {
26
DEF_HELPER_FLAGS_3(mve_vmovi, TCG_CALL_NO_WG, void, env, ptr, i64)
34
FIXUP_NONE = 0, /* do nothing */
27
DEF_HELPER_FLAGS_3(mve_vandi, TCG_CALL_NO_WG, void, env, ptr, i64)
35
FIXUP_TERMINATOR, /* end of insns */
28
DEF_HELPER_FLAGS_3(mve_vorri, TCG_CALL_NO_WG, void, env, ptr, i64)
36
@@ -XXX,XX +XXX,XX @@ static uint64_t arm_load_elf(struct arm_boot_info *info, uint64_t *pentry,
29
diff --git a/target/arm/mve.decode b/target/arm/mve.decode
37
return ret;
30
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/mve.decode
32
+++ b/target/arm/mve.decode
33
@@ -XXX,XX +XXX,XX @@ VQDMULH_scalar 1110 1110 0 . .. ... 1 ... 0 1110 . 110 .... @2scalar
34
VQRDMULH_scalar 1111 1110 0 . .. ... 1 ... 0 1110 . 110 .... @2scalar
35
36
# Vector add across vector
37
-VADDV 111 u:1 1110 1111 size:2 01 ... 0 1111 0 0 a:1 0 qm:3 0 rda=%rdalo
38
+{
39
+ VADDV 111 u:1 1110 1111 size:2 01 ... 0 1111 0 0 a:1 0 qm:3 0 rda=%rdalo
40
+ VADDLV 111 u:1 1110 1 ... 1001 ... 0 1111 00 a:1 0 qm:3 0 \
41
+ rdahi=%rdahi rdalo=%rdalo
42
+}
43
44
# Predicate operations
45
%mask_22_13 22:1 13:3
46
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/target/arm/mve_helper.c
49
+++ b/target/arm/mve_helper.c
50
@@ -XXX,XX +XXX,XX @@ DO_VADDV(vaddvub, 1, uint8_t)
51
DO_VADDV(vaddvuh, 2, uint16_t)
52
DO_VADDV(vaddvuw, 4, uint32_t)
53
54
+#define DO_VADDLV(OP, TYPE, LTYPE) \
55
+ uint64_t HELPER(glue(mve_, OP))(CPUARMState *env, void *vm, \
56
+ uint64_t ra) \
57
+ { \
58
+ uint16_t mask = mve_element_mask(env); \
59
+ unsigned e; \
60
+ TYPE *m = vm; \
61
+ for (e = 0; e < 16 / 4; e++, mask >>= 4) { \
62
+ if (mask & 1) { \
63
+ ra += (LTYPE)m[H4(e)]; \
64
+ } \
65
+ } \
66
+ mve_advance_vpt(env); \
67
+ return ra; \
68
+ } \
69
+
70
+DO_VADDLV(vaddlv_s, int32_t, int64_t)
71
+DO_VADDLV(vaddlv_u, uint32_t, uint64_t)
72
+
73
/* Shifts by immediate */
74
#define DO_2SHIFT(OP, ESIZE, TYPE, FN) \
75
void HELPER(glue(mve_, OP))(CPUARMState *env, void *vd, \
76
diff --git a/target/arm/translate-mve.c b/target/arm/translate-mve.c
77
index XXXXXXX..XXXXXXX 100644
78
--- a/target/arm/translate-mve.c
79
+++ b/target/arm/translate-mve.c
80
@@ -XXX,XX +XXX,XX @@ static bool trans_VADDV(DisasContext *s, arg_VADDV *a)
81
return true;
38
}
82
}
39
83
40
+static uint64_t load_aarch64_image(const char *filename, hwaddr mem_base,
84
+static bool trans_VADDLV(DisasContext *s, arg_VADDLV *a)
41
+ hwaddr *entry)
42
+{
85
+{
43
+ hwaddr kernel_load_offset = KERNEL64_LOAD_ADDR;
86
+ /*
44
+ uint8_t *buffer;
87
+ * Vector Add Long Across Vector: accumulate the 32-bit
45
+ int size;
88
+ * elements of the vector into a 64-bit result stored in
89
+ * a pair of general-purpose registers.
90
+ * No need to check Qm's bank: it is only 3 bits in decode.
91
+ */
92
+ TCGv_ptr qm;
93
+ TCGv_i64 rda;
94
+ TCGv_i32 rdalo, rdahi;
46
+
95
+
47
+ /* On aarch64, it's the bootloader's job to uncompress the kernel. */
96
+ if (!dc_isar_feature(aa32_mve, s)) {
48
+ size = load_image_gzipped_buffer(filename, LOAD_IMAGE_MAX_GUNZIP_BYTES,
97
+ return false;
49
+ &buffer);
98
+ }
50
+
99
+ /*
51
+ if (size < 0) {
100
+ * rdahi == 13 is UNPREDICTABLE; rdahi == 15 is a related
52
+ gsize len;
101
+ * encoding; rdalo always has bit 0 clear so cannot be 13 or 15.
53
+
102
+ */
54
+ /* Load as raw file otherwise */
103
+ if (a->rdahi == 13 || a->rdahi == 15) {
55
+ if (!g_file_get_contents(filename, (char **)&buffer, &len, NULL)) {
104
+ return false;
56
+ return -1;
105
+ }
57
+ }
106
+ if (!mve_eci_check(s) || !vfp_access_check(s)) {
58
+ size = len;
107
+ return true;
59
+ }
108
+ }
60
+
109
+
61
+ /* check the arm64 magic header value -- very old kernels may not have it */
110
+ /*
62
+ if (memcmp(buffer + ARM64_MAGIC_OFFSET, "ARM\x64", 4) == 0) {
111
+ * This insn is subject to beat-wise execution. Partial execution
63
+ uint64_t hdrvals[2];
112
+ * of an A=0 (no-accumulate) insn which does not execute the first
64
+
113
+ * beat must start with the current value of RdaHi:RdaLo, not zero.
65
+ /* The arm64 Image header has text_offset and image_size fields at 8 and
114
+ */
66
+ * 16 bytes into the Image header, respectively. The text_offset field
115
+ if (a->a || mve_skip_first_beat(s)) {
67
+ * is only valid if the image_size is non-zero.
116
+ /* Accumulate input from RdaHi:RdaLo */
68
+ */
117
+ rda = tcg_temp_new_i64();
69
+ memcpy(&hdrvals, buffer + ARM64_TEXT_OFFSET_OFFSET, sizeof(hdrvals));
118
+ rdalo = load_reg(s, a->rdalo);
70
+ if (hdrvals[1] != 0) {
119
+ rdahi = load_reg(s, a->rdahi);
71
+ kernel_load_offset = le64_to_cpu(hdrvals[0]);
120
+ tcg_gen_concat_i32_i64(rda, rdalo, rdahi);
72
+ }
121
+ tcg_temp_free_i32(rdalo);
122
+ tcg_temp_free_i32(rdahi);
123
+ } else {
124
+ /* Accumulate starting at zero */
125
+ rda = tcg_const_i64(0);
73
+ }
126
+ }
74
+
127
+
75
+ *entry = mem_base + kernel_load_offset;
128
+ qm = mve_qreg_ptr(a->qm);
76
+ rom_add_blob_fixed(filename, buffer, size, *entry);
129
+ if (a->u) {
130
+ gen_helper_mve_vaddlv_u(rda, cpu_env, qm, rda);
131
+ } else {
132
+ gen_helper_mve_vaddlv_s(rda, cpu_env, qm, rda);
133
+ }
134
+ tcg_temp_free_ptr(qm);
77
+
135
+
78
+ g_free(buffer);
136
+ rdalo = tcg_temp_new_i32();
79
+
137
+ rdahi = tcg_temp_new_i32();
80
+ return size;
138
+ tcg_gen_extrl_i64_i32(rdalo, rda);
139
+ tcg_gen_extrh_i64_i32(rdahi, rda);
140
+ store_reg(s, a->rdalo, rdalo);
141
+ store_reg(s, a->rdahi, rdahi);
142
+ tcg_temp_free_i64(rda);
143
+ mve_update_eci(s);
144
+ return true;
81
+}
145
+}
82
+
146
+
83
static void arm_load_kernel_notify(Notifier *notifier, void *data)
147
static bool do_1imm(DisasContext *s, arg_1imm *a, MVEGenOneOpImmFn *fn)
84
{
148
{
85
CPUState *cs;
149
TCGv_ptr qd;
86
@@ -XXX,XX +XXX,XX @@ static void arm_load_kernel_notify(Notifier *notifier, void *data)
87
int is_linux = 0;
88
uint64_t elf_entry, elf_low_addr, elf_high_addr;
89
int elf_machine;
90
- hwaddr entry, kernel_load_offset;
91
+ hwaddr entry;
92
static const ARMInsnFixup *primary_loader;
93
ArmLoadKernelNotifier *n = DO_UPCAST(ArmLoadKernelNotifier,
94
notifier, notifier);
95
@@ -XXX,XX +XXX,XX @@ static void arm_load_kernel_notify(Notifier *notifier, void *data)
96
97
if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64)) {
98
primary_loader = bootloader_aarch64;
99
- kernel_load_offset = KERNEL64_LOAD_ADDR;
100
elf_machine = EM_AARCH64;
101
} else {
102
primary_loader = bootloader;
103
if (!info->write_board_setup) {
104
primary_loader += BOOTLOADER_NO_BOARD_SETUP_OFFSET;
105
}
106
- kernel_load_offset = KERNEL_LOAD_ADDR;
107
elf_machine = EM_ARM;
108
}
109
110
@@ -XXX,XX +XXX,XX @@ static void arm_load_kernel_notify(Notifier *notifier, void *data)
111
kernel_size = load_uimage(info->kernel_filename, &entry, NULL,
112
&is_linux, NULL, NULL);
113
}
114
- /* On aarch64, it's the bootloader's job to uncompress the kernel. */
115
if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) && kernel_size < 0) {
116
- entry = info->loader_start + kernel_load_offset;
117
- kernel_size = load_image_gzipped(info->kernel_filename, entry,
118
- info->ram_size - kernel_load_offset);
119
+ kernel_size = load_aarch64_image(info->kernel_filename,
120
+ info->loader_start, &entry);
121
is_linux = 1;
122
- }
123
- if (kernel_size < 0) {
124
- entry = info->loader_start + kernel_load_offset;
125
+ } else if (kernel_size < 0) {
126
+ /* 32-bit ARM */
127
+ entry = info->loader_start + KERNEL_LOAD_ADDR;
128
kernel_size = load_image_targphys(info->kernel_filename, entry,
129
- info->ram_size - kernel_load_offset);
130
+ info->ram_size - KERNEL_LOAD_ADDR);
131
is_linux = 1;
132
}
133
if (kernel_size < 0) {
134
--
150
--
135
2.7.4
151
2.20.1
136
152
137
153
diff view generated by jsdifflib
1
In Thumb mode, the only instructions which can cause an interworking
1
The MVE extension to v8.1M includes some new shift instructions which
2
branch by writing the PC are BLX, BX, BXJ, LDR, POP and LDM. Unlike
2
sit entirely within the non-coprocessor part of the encoding space
3
ARM mode, data processing instructions which target the PC do not
3
and which operate only on general-purpose registers. They take up
4
cause interworking branches.
4
the space which was previously UNPREDICTABLE MOVS and ORRS encodings
5
5
with Rm == 13 or 15.
6
When we added support for doing interworking branches on writes to
6
7
PC from data processing instructions in commit 21aeb3430ce7ba, we
7
Implement the long shifts by immediate, which perform shifts on a
8
accidentally changed a Thumb instruction to have interworking
8
pair of general-purpose registers treated as a 64-bit quantity, with
9
branch behaviour for writes to PC. (MOV, MOVS register-shifted
9
an immediate shift count between 1 and 32.
10
register, encoding T2; this is the standard encoding for
10
11
LSL/LSR/ASR/ROR (register).)
11
Awkwardly, because the MOVS and ORRS trans functions do not UNDEF for
12
12
the Rm==13,15 case, we need to explicitly emit code to UNDEF for the
13
For this encoding, behaviour with Rd == R15 is specified as
13
cases where v8.1M now requires that. (Trying to change MOVS and ORRS
14
UNPREDICTABLE, so allowing an interworking branch is within
14
is too difficult, because the functions that generate the code are
15
spec, but it's confusing and differs from our handling of this
15
shared between a dozen different kinds of arithmetic or logical
16
class of UNPREDICTABLE for other Thumb ALU operations. Make
16
instruction for all A32, T16 and T32 encodings, and for some insns
17
it perform a simple (non-interworking) branch like the others.
17
and some encodings Rm==13,15 are valid.)
18
19
We make the helper functions we need for UQSHLL and SQSHLL take
20
a 32-bit value which the helper casts to int8_t because we'll need
21
these helpers also for the shift-by-register insns, where the shift
22
count might be < 0 or > 32.
18
23
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
24
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
Reviewed-by: Richard Henderson <rth@twiddle.net>
25
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
21
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
26
Message-id: 20210628135835.6690-16-peter.maydell@linaro.org
22
Message-id: 1491844419-12485-3-git-send-email-peter.maydell@linaro.org
23
---
27
---
24
target/arm/translate.c | 2 +-
28
target/arm/helper-mve.h | 3 ++
25
1 file changed, 1 insertion(+), 1 deletion(-)
29
target/arm/translate.h | 1 +
26
30
target/arm/t32.decode | 28 +++++++++++++
31
target/arm/mve_helper.c | 10 +++++
32
target/arm/translate.c | 90 +++++++++++++++++++++++++++++++++++++++++
33
5 files changed, 132 insertions(+)
34
35
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
36
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/helper-mve.h
38
+++ b/target/arm/helper-mve.h
39
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vqrshruntb, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
40
DEF_HELPER_FLAGS_4(mve_vqrshrunth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
41
42
DEF_HELPER_FLAGS_4(mve_vshlc, TCG_CALL_NO_WG, i32, env, ptr, i32, i32)
43
+
44
+DEF_HELPER_FLAGS_3(mve_sqshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
45
+DEF_HELPER_FLAGS_3(mve_uqshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
46
diff --git a/target/arm/translate.h b/target/arm/translate.h
47
index XXXXXXX..XXXXXXX 100644
48
--- a/target/arm/translate.h
49
+++ b/target/arm/translate.h
50
@@ -XXX,XX +XXX,XX @@ typedef void CryptoTwoOpFn(TCGv_ptr, TCGv_ptr);
51
typedef void CryptoThreeOpIntFn(TCGv_ptr, TCGv_ptr, TCGv_i32);
52
typedef void CryptoThreeOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
53
typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
54
+typedef void WideShiftImmFn(TCGv_i64, TCGv_i64, int64_t shift);
55
56
/**
57
* arm_tbflags_from_tb:
58
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
59
index XXXXXXX..XXXXXXX 100644
60
--- a/target/arm/t32.decode
61
+++ b/target/arm/t32.decode
62
@@ -XXX,XX +XXX,XX @@
63
&mcr !extern cp opc1 crn crm opc2 rt
64
&mcrr !extern cp opc1 crm rt rt2
65
66
+&mve_shl_ri rdalo rdahi shim
67
+
68
+# rdahi: bits [3:1] from insn, bit 0 is 1
69
+# rdalo: bits [3:1] from insn, bit 0 is 0
70
+%rdahi_9 9:3 !function=times_2_plus_1
71
+%rdalo_17 17:3 !function=times_2
72
+
73
# Data-processing (register)
74
75
%imm5_12_6 12:3 6:2
76
@@ -XXX,XX +XXX,XX @@
77
@S_xrr_shi ....... .... . rn:4 .... .... .. shty:2 rm:4 \
78
&s_rrr_shi shim=%imm5_12_6 s=1 rd=0
79
80
+@mve_shl_ri ....... .... . ... . . ... ... . .. .. .... \
81
+ &mve_shl_ri shim=%imm5_12_6 rdalo=%rdalo_17 rdahi=%rdahi_9
82
+
83
{
84
TST_xrri 1110101 0000 1 .... 0 ... 1111 .... .... @S_xrr_shi
85
AND_rrri 1110101 0000 . .... 0 ... .... .... .... @s_rrr_shi
86
}
87
BIC_rrri 1110101 0001 . .... 0 ... .... .... .... @s_rrr_shi
88
{
89
+ # The v8.1M MVE shift insns overlap in encoding with MOVS/ORRS
90
+ # and are distinguished by having Rm==13 or 15. Those are UNPREDICTABLE
91
+ # cases for MOVS/ORRS. We decode the MVE cases first, ensuring that
92
+ # they explicitly call unallocated_encoding() for cases that must UNDEF
93
+ # (eg "using a new shift insn on a v8.1M CPU without MVE"), and letting
94
+ # the rest fall through (where ORR_rrri and MOV_rxri will end up
95
+ # handling them as r13 and r15 accesses with the same semantics as A32).
96
+ [
97
+ LSLL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 00 1111 @mve_shl_ri
98
+ LSRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 01 1111 @mve_shl_ri
99
+ ASRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 10 1111 @mve_shl_ri
100
+
101
+ UQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 00 1111 @mve_shl_ri
102
+ URSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 01 1111 @mve_shl_ri
103
+ SRSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 10 1111 @mve_shl_ri
104
+ SQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 11 1111 @mve_shl_ri
105
+ ]
106
+
107
MOV_rxri 1110101 0010 . 1111 0 ... .... .... .... @s_rxr_shi
108
ORR_rrri 1110101 0010 . .... 0 ... .... .... .... @s_rrr_shi
109
}
110
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
111
index XXXXXXX..XXXXXXX 100644
112
--- a/target/arm/mve_helper.c
113
+++ b/target/arm/mve_helper.c
114
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(mve_vshlc)(CPUARMState *env, void *vd, uint32_t rdm,
115
mve_advance_vpt(env);
116
return rdm;
117
}
118
+
119
+uint64_t HELPER(mve_sqshll)(CPUARMState *env, uint64_t n, uint32_t shift)
120
+{
121
+ return do_sqrshl_d(n, (int8_t)shift, false, &env->QF);
122
+}
123
+
124
+uint64_t HELPER(mve_uqshll)(CPUARMState *env, uint64_t n, uint32_t shift)
125
+{
126
+ return do_uqrshl_d(n, (int8_t)shift, false, &env->QF);
127
+}
27
diff --git a/target/arm/translate.c b/target/arm/translate.c
128
diff --git a/target/arm/translate.c b/target/arm/translate.c
28
index XXXXXXX..XXXXXXX 100644
129
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate.c
130
--- a/target/arm/translate.c
30
+++ b/target/arm/translate.c
131
+++ b/target/arm/translate.c
31
@@ -XXX,XX +XXX,XX @@ static int disas_thumb2_insn(CPUARMState *env, DisasContext *s, uint16_t insn_hw
132
@@ -XXX,XX +XXX,XX @@ static bool trans_MOVT(DisasContext *s, arg_MOVW *a)
32
gen_arm_shift_reg(tmp, op, tmp2, logic_cc);
133
return true;
33
if (logic_cc)
134
}
34
gen_logic_CC(tmp);
135
35
- store_reg_bx(s, rd, tmp);
136
+/*
36
+ store_reg(s, rd, tmp);
137
+ * v8.1M MVE wide-shifts
37
break;
138
+ */
38
case 1: /* Sign/zero extend. */
139
+static bool do_mve_shl_ri(DisasContext *s, arg_mve_shl_ri *a,
39
op = (insn >> 20) & 7;
140
+ WideShiftImmFn *fn)
141
+{
142
+ TCGv_i64 rda;
143
+ TCGv_i32 rdalo, rdahi;
144
+
145
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
146
+ /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
147
+ return false;
148
+ }
149
+ if (a->rdahi == 15) {
150
+ /* These are a different encoding (SQSHL/SRSHR/UQSHL/URSHR) */
151
+ return false;
152
+ }
153
+ if (!dc_isar_feature(aa32_mve, s) ||
154
+ !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
155
+ a->rdahi == 13) {
156
+ /* RdaHi == 13 is UNPREDICTABLE; we choose to UNDEF */
157
+ unallocated_encoding(s);
158
+ return true;
159
+ }
160
+
161
+ if (a->shim == 0) {
162
+ a->shim = 32;
163
+ }
164
+
165
+ rda = tcg_temp_new_i64();
166
+ rdalo = load_reg(s, a->rdalo);
167
+ rdahi = load_reg(s, a->rdahi);
168
+ tcg_gen_concat_i32_i64(rda, rdalo, rdahi);
169
+
170
+ fn(rda, rda, a->shim);
171
+
172
+ tcg_gen_extrl_i64_i32(rdalo, rda);
173
+ tcg_gen_extrh_i64_i32(rdahi, rda);
174
+ store_reg(s, a->rdalo, rdalo);
175
+ store_reg(s, a->rdahi, rdahi);
176
+ tcg_temp_free_i64(rda);
177
+
178
+ return true;
179
+}
180
+
181
+static bool trans_ASRL_ri(DisasContext *s, arg_mve_shl_ri *a)
182
+{
183
+ return do_mve_shl_ri(s, a, tcg_gen_sari_i64);
184
+}
185
+
186
+static bool trans_LSLL_ri(DisasContext *s, arg_mve_shl_ri *a)
187
+{
188
+ return do_mve_shl_ri(s, a, tcg_gen_shli_i64);
189
+}
190
+
191
+static bool trans_LSRL_ri(DisasContext *s, arg_mve_shl_ri *a)
192
+{
193
+ return do_mve_shl_ri(s, a, tcg_gen_shri_i64);
194
+}
195
+
196
+static void gen_mve_sqshll(TCGv_i64 r, TCGv_i64 n, int64_t shift)
197
+{
198
+ gen_helper_mve_sqshll(r, cpu_env, n, tcg_constant_i32(shift));
199
+}
200
+
201
+static bool trans_SQSHLL_ri(DisasContext *s, arg_mve_shl_ri *a)
202
+{
203
+ return do_mve_shl_ri(s, a, gen_mve_sqshll);
204
+}
205
+
206
+static void gen_mve_uqshll(TCGv_i64 r, TCGv_i64 n, int64_t shift)
207
+{
208
+ gen_helper_mve_uqshll(r, cpu_env, n, tcg_constant_i32(shift));
209
+}
210
+
211
+static bool trans_UQSHLL_ri(DisasContext *s, arg_mve_shl_ri *a)
212
+{
213
+ return do_mve_shl_ri(s, a, gen_mve_uqshll);
214
+}
215
+
216
+static bool trans_SRSHRL_ri(DisasContext *s, arg_mve_shl_ri *a)
217
+{
218
+ return do_mve_shl_ri(s, a, gen_srshr64_i64);
219
+}
220
+
221
+static bool trans_URSHRL_ri(DisasContext *s, arg_mve_shl_ri *a)
222
+{
223
+ return do_mve_shl_ri(s, a, gen_urshr64_i64);
224
+}
225
+
226
/*
227
* Multiply and multiply accumulate
228
*/
40
--
229
--
41
2.7.4
230
2.20.1
42
231
43
232
diff view generated by jsdifflib
1
For M-profile CPUs, the BXJ instruction does not exist at all, and
1
Implement the MVE long shifts by register, which perform shifts on a
2
the encoding should always UNDEF. We were accidentally implementing
2
pair of general-purpose registers treated as a 64-bit quantity, with
3
it to behave like A-profile BXJ; correct the error.
3
the shift count in another general-purpose register, which might be
4
either positive or negative.
5
6
Like the long-shifts-by-immediate, these encodings sit in the space
7
that was previously the UNPREDICTABLE MOVS/ORRS with Rm==13,15.
8
Because LSLL_rr and ASRL_rr overlap with both MOV_rxri/ORR_rrri and
9
also with CSEL (as one of the previously-UNPREDICTABLE Rm==13 cases),
10
we have to move the CSEL pattern into the same decodetree group.
4
11
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Reviewed-by: Richard Henderson <rth@twiddle.net>
14
Message-id: 20210628135835.6690-17-peter.maydell@linaro.org
8
Message-id: 1491844419-12485-2-git-send-email-peter.maydell@linaro.org
9
---
15
---
10
target/arm/translate.c | 7 ++++++-
16
target/arm/helper-mve.h | 6 +++
11
1 file changed, 6 insertions(+), 1 deletion(-)
17
target/arm/translate.h | 1 +
12
18
target/arm/t32.decode | 16 +++++--
19
target/arm/mve_helper.c | 93 +++++++++++++++++++++++++++++++++++++++++
20
target/arm/translate.c | 69 ++++++++++++++++++++++++++++++
21
5 files changed, 182 insertions(+), 3 deletions(-)
22
23
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
24
index XXXXXXX..XXXXXXX 100644
25
--- a/target/arm/helper-mve.h
26
+++ b/target/arm/helper-mve.h
27
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(mve_vqrshrunth, TCG_CALL_NO_WG, void, env, ptr, ptr, i32)
28
29
DEF_HELPER_FLAGS_4(mve_vshlc, TCG_CALL_NO_WG, i32, env, ptr, i32, i32)
30
31
+DEF_HELPER_FLAGS_3(mve_sshrl, TCG_CALL_NO_RWG, i64, env, i64, i32)
32
+DEF_HELPER_FLAGS_3(mve_ushll, TCG_CALL_NO_RWG, i64, env, i64, i32)
33
DEF_HELPER_FLAGS_3(mve_sqshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
34
DEF_HELPER_FLAGS_3(mve_uqshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
35
+DEF_HELPER_FLAGS_3(mve_sqrshrl, TCG_CALL_NO_RWG, i64, env, i64, i32)
36
+DEF_HELPER_FLAGS_3(mve_uqrshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
37
+DEF_HELPER_FLAGS_3(mve_sqrshrl48, TCG_CALL_NO_RWG, i64, env, i64, i32)
38
+DEF_HELPER_FLAGS_3(mve_uqrshll48, TCG_CALL_NO_RWG, i64, env, i64, i32)
39
diff --git a/target/arm/translate.h b/target/arm/translate.h
40
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/translate.h
42
+++ b/target/arm/translate.h
43
@@ -XXX,XX +XXX,XX @@ typedef void CryptoThreeOpIntFn(TCGv_ptr, TCGv_ptr, TCGv_i32);
44
typedef void CryptoThreeOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
45
typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
46
typedef void WideShiftImmFn(TCGv_i64, TCGv_i64, int64_t shift);
47
+typedef void WideShiftFn(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i32);
48
49
/**
50
* arm_tbflags_from_tb:
51
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
52
index XXXXXXX..XXXXXXX 100644
53
--- a/target/arm/t32.decode
54
+++ b/target/arm/t32.decode
55
@@ -XXX,XX +XXX,XX @@
56
&mcrr !extern cp opc1 crm rt rt2
57
58
&mve_shl_ri rdalo rdahi shim
59
+&mve_shl_rr rdalo rdahi rm
60
61
# rdahi: bits [3:1] from insn, bit 0 is 1
62
# rdalo: bits [3:1] from insn, bit 0 is 0
63
@@ -XXX,XX +XXX,XX @@
64
65
@mve_shl_ri ....... .... . ... . . ... ... . .. .. .... \
66
&mve_shl_ri shim=%imm5_12_6 rdalo=%rdalo_17 rdahi=%rdahi_9
67
+@mve_shl_rr ....... .... . ... . rm:4 ... . .. .. .... \
68
+ &mve_shl_rr rdalo=%rdalo_17 rdahi=%rdahi_9
69
70
{
71
TST_xrri 1110101 0000 1 .... 0 ... 1111 .... .... @S_xrr_shi
72
@@ -XXX,XX +XXX,XX @@ BIC_rrri 1110101 0001 . .... 0 ... .... .... .... @s_rrr_shi
73
URSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 01 1111 @mve_shl_ri
74
SRSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 10 1111 @mve_shl_ri
75
SQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 11 1111 @mve_shl_ri
76
+
77
+ LSLL_rr 1110101 0010 1 ... 0 .... ... 1 0000 1101 @mve_shl_rr
78
+ ASRL_rr 1110101 0010 1 ... 0 .... ... 1 0010 1101 @mve_shl_rr
79
+ UQRSHLL64_rr 1110101 0010 1 ... 1 .... ... 1 0000 1101 @mve_shl_rr
80
+ SQRSHRL64_rr 1110101 0010 1 ... 1 .... ... 1 0010 1101 @mve_shl_rr
81
+ UQRSHLL48_rr 1110101 0010 1 ... 1 .... ... 1 1000 1101 @mve_shl_rr
82
+ SQRSHRL48_rr 1110101 0010 1 ... 1 .... ... 1 1010 1101 @mve_shl_rr
83
]
84
85
MOV_rxri 1110101 0010 . 1111 0 ... .... .... .... @s_rxr_shi
86
ORR_rrri 1110101 0010 . .... 0 ... .... .... .... @s_rrr_shi
87
+
88
+ # v8.1M CSEL and friends
89
+ CSEL 1110101 0010 1 rn:4 10 op:2 rd:4 fcond:4 rm:4
90
}
91
{
92
MVN_rxri 1110101 0011 . 1111 0 ... .... .... .... @s_rxr_shi
93
@@ -XXX,XX +XXX,XX @@ SBC_rrri 1110101 1011 . .... 0 ... .... .... .... @s_rrr_shi
94
}
95
RSB_rrri 1110101 1110 . .... 0 ... .... .... .... @s_rrr_shi
96
97
-# v8.1M CSEL and friends
98
-CSEL 1110101 0010 1 rn:4 10 op:2 rd:4 fcond:4 rm:4
99
-
100
# Data-processing (register-shifted register)
101
102
MOV_rxrr 1111 1010 0 shty:2 s:1 rm:4 1111 rd:4 0000 rs:4 \
103
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
104
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/mve_helper.c
106
+++ b/target/arm/mve_helper.c
107
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(mve_vshlc)(CPUARMState *env, void *vd, uint32_t rdm,
108
return rdm;
109
}
110
111
+uint64_t HELPER(mve_sshrl)(CPUARMState *env, uint64_t n, uint32_t shift)
112
+{
113
+ return do_sqrshl_d(n, -(int8_t)shift, false, NULL);
114
+}
115
+
116
+uint64_t HELPER(mve_ushll)(CPUARMState *env, uint64_t n, uint32_t shift)
117
+{
118
+ return do_uqrshl_d(n, (int8_t)shift, false, NULL);
119
+}
120
+
121
uint64_t HELPER(mve_sqshll)(CPUARMState *env, uint64_t n, uint32_t shift)
122
{
123
return do_sqrshl_d(n, (int8_t)shift, false, &env->QF);
124
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mve_uqshll)(CPUARMState *env, uint64_t n, uint32_t shift)
125
{
126
return do_uqrshl_d(n, (int8_t)shift, false, &env->QF);
127
}
128
+
129
+uint64_t HELPER(mve_sqrshrl)(CPUARMState *env, uint64_t n, uint32_t shift)
130
+{
131
+ return do_sqrshl_d(n, -(int8_t)shift, true, &env->QF);
132
+}
133
+
134
+uint64_t HELPER(mve_uqrshll)(CPUARMState *env, uint64_t n, uint32_t shift)
135
+{
136
+ return do_uqrshl_d(n, (int8_t)shift, true, &env->QF);
137
+}
138
+
139
+/* Operate on 64-bit values, but saturate at 48 bits */
140
+static inline int64_t do_sqrshl48_d(int64_t src, int64_t shift,
141
+ bool round, uint32_t *sat)
142
+{
143
+ if (shift <= -48) {
144
+ /* Rounding the sign bit always produces 0. */
145
+ if (round) {
146
+ return 0;
147
+ }
148
+ return src >> 63;
149
+ } else if (shift < 0) {
150
+ if (round) {
151
+ src >>= -shift - 1;
152
+ return (src >> 1) + (src & 1);
153
+ }
154
+ return src >> -shift;
155
+ } else if (shift < 48) {
156
+ int64_t val = src << shift;
157
+ int64_t extval = sextract64(val, 0, 48);
158
+ if (!sat || val == extval) {
159
+ return extval;
160
+ }
161
+ } else if (!sat || src == 0) {
162
+ return 0;
163
+ }
164
+
165
+ *sat = 1;
166
+ return (1ULL << 47) - (src >= 0);
167
+}
168
+
169
+/* Operate on 64-bit values, but saturate at 48 bits */
170
+static inline uint64_t do_uqrshl48_d(uint64_t src, int64_t shift,
171
+ bool round, uint32_t *sat)
172
+{
173
+ uint64_t val, extval;
174
+
175
+ if (shift <= -(48 + round)) {
176
+ return 0;
177
+ } else if (shift < 0) {
178
+ if (round) {
179
+ val = src >> (-shift - 1);
180
+ val = (val >> 1) + (val & 1);
181
+ } else {
182
+ val = src >> -shift;
183
+ }
184
+ extval = extract64(val, 0, 48);
185
+ if (!sat || val == extval) {
186
+ return extval;
187
+ }
188
+ } else if (shift < 48) {
189
+ uint64_t val = src << shift;
190
+ uint64_t extval = extract64(val, 0, 48);
191
+ if (!sat || val == extval) {
192
+ return extval;
193
+ }
194
+ } else if (!sat || src == 0) {
195
+ return 0;
196
+ }
197
+
198
+ *sat = 1;
199
+ return MAKE_64BIT_MASK(0, 48);
200
+}
201
+
202
+uint64_t HELPER(mve_sqrshrl48)(CPUARMState *env, uint64_t n, uint32_t shift)
203
+{
204
+ return do_sqrshl48_d(n, -(int8_t)shift, true, &env->QF);
205
+}
206
+
207
+uint64_t HELPER(mve_uqrshll48)(CPUARMState *env, uint64_t n, uint32_t shift)
208
+{
209
+ return do_uqrshl48_d(n, (int8_t)shift, true, &env->QF);
210
+}
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
211
diff --git a/target/arm/translate.c b/target/arm/translate.c
14
index XXXXXXX..XXXXXXX 100644
212
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/translate.c
213
--- a/target/arm/translate.c
16
+++ b/target/arm/translate.c
214
+++ b/target/arm/translate.c
17
@@ -XXX,XX +XXX,XX @@ static int disas_thumb2_insn(CPUARMState *env, DisasContext *s, uint16_t insn_hw
215
@@ -XXX,XX +XXX,XX @@ static bool trans_URSHRL_ri(DisasContext *s, arg_mve_shl_ri *a)
18
}
216
return do_mve_shl_ri(s, a, gen_urshr64_i64);
19
break;
217
}
20
case 4: /* bxj */
218
21
- /* Trivial implementation equivalent to bx. */
219
+static bool do_mve_shl_rr(DisasContext *s, arg_mve_shl_rr *a, WideShiftFn *fn)
22
+ /* Trivial implementation equivalent to bx.
220
+{
23
+ * This instruction doesn't exist at all for M-profile.
221
+ TCGv_i64 rda;
24
+ */
222
+ TCGv_i32 rdalo, rdahi;
25
+ if (arm_dc_feature(s, ARM_FEATURE_M)) {
223
+
26
+ goto illegal_op;
224
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
27
+ }
225
+ /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
28
tmp = load_reg(s, rn);
226
+ return false;
29
gen_bx(s, tmp);
227
+ }
30
break;
228
+ if (a->rdahi == 15) {
229
+ /* These are a different encoding (SQSHL/SRSHR/UQSHL/URSHR) */
230
+ return false;
231
+ }
232
+ if (!dc_isar_feature(aa32_mve, s) ||
233
+ !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
234
+ a->rdahi == 13 || a->rm == 13 || a->rm == 15 ||
235
+ a->rm == a->rdahi || a->rm == a->rdalo) {
236
+ /* These rdahi/rdalo/rm cases are UNPREDICTABLE; we choose to UNDEF */
237
+ unallocated_encoding(s);
238
+ return true;
239
+ }
240
+
241
+ rda = tcg_temp_new_i64();
242
+ rdalo = load_reg(s, a->rdalo);
243
+ rdahi = load_reg(s, a->rdahi);
244
+ tcg_gen_concat_i32_i64(rda, rdalo, rdahi);
245
+
246
+ /* The helper takes care of the sign-extension of the low 8 bits of Rm */
247
+ fn(rda, cpu_env, rda, cpu_R[a->rm]);
248
+
249
+ tcg_gen_extrl_i64_i32(rdalo, rda);
250
+ tcg_gen_extrh_i64_i32(rdahi, rda);
251
+ store_reg(s, a->rdalo, rdalo);
252
+ store_reg(s, a->rdahi, rdahi);
253
+ tcg_temp_free_i64(rda);
254
+
255
+ return true;
256
+}
257
+
258
+static bool trans_LSLL_rr(DisasContext *s, arg_mve_shl_rr *a)
259
+{
260
+ return do_mve_shl_rr(s, a, gen_helper_mve_ushll);
261
+}
262
+
263
+static bool trans_ASRL_rr(DisasContext *s, arg_mve_shl_rr *a)
264
+{
265
+ return do_mve_shl_rr(s, a, gen_helper_mve_sshrl);
266
+}
267
+
268
+static bool trans_UQRSHLL64_rr(DisasContext *s, arg_mve_shl_rr *a)
269
+{
270
+ return do_mve_shl_rr(s, a, gen_helper_mve_uqrshll);
271
+}
272
+
273
+static bool trans_SQRSHRL64_rr(DisasContext *s, arg_mve_shl_rr *a)
274
+{
275
+ return do_mve_shl_rr(s, a, gen_helper_mve_sqrshrl);
276
+}
277
+
278
+static bool trans_UQRSHLL48_rr(DisasContext *s, arg_mve_shl_rr *a)
279
+{
280
+ return do_mve_shl_rr(s, a, gen_helper_mve_uqrshll48);
281
+}
282
+
283
+static bool trans_SQRSHRL48_rr(DisasContext *s, arg_mve_shl_rr *a)
284
+{
285
+ return do_mve_shl_rr(s, a, gen_helper_mve_sqrshrl48);
286
+}
287
+
288
/*
289
* Multiply and multiply accumulate
290
*/
31
--
291
--
32
2.7.4
292
2.20.1
33
293
34
294
diff view generated by jsdifflib
1
We now test for "are we singlestepping" in several places and
1
Implement the MVE shifts by immediate, which perform shifts
2
it's not a trivial check because we need to care about both
2
on a single general-purpose register.
3
architectural singlestep and QEMU gdbstub singlestep. We're
3
4
also about to add another place that needs to make this check,
4
These patterns overlap with the long-shift-by-immediates,
5
so pull the condition out into a function.
5
so we have to rearrange the grouping a little here.
6
6
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <rth@twiddle.net>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Message-id: 20210628135835.6690-18-peter.maydell@linaro.org
10
Message-id: 1491844419-12485-7-git-send-email-peter.maydell@linaro.org
11
---
10
---
12
target/arm/translate.c | 20 +++++++++++++++-----
11
target/arm/helper-mve.h | 3 ++
13
1 file changed, 15 insertions(+), 5 deletions(-)
12
target/arm/translate.h | 1 +
14
13
target/arm/t32.decode | 31 ++++++++++++++-----
14
target/arm/mve_helper.c | 10 ++++++
15
target/arm/translate.c | 68 +++++++++++++++++++++++++++++++++++++++--
16
5 files changed, 104 insertions(+), 9 deletions(-)
17
18
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/helper-mve.h
21
+++ b/target/arm/helper-mve.h
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_sqrshrl, TCG_CALL_NO_RWG, i64, env, i64, i32)
23
DEF_HELPER_FLAGS_3(mve_uqrshll, TCG_CALL_NO_RWG, i64, env, i64, i32)
24
DEF_HELPER_FLAGS_3(mve_sqrshrl48, TCG_CALL_NO_RWG, i64, env, i64, i32)
25
DEF_HELPER_FLAGS_3(mve_uqrshll48, TCG_CALL_NO_RWG, i64, env, i64, i32)
26
+
27
+DEF_HELPER_FLAGS_3(mve_uqshl, TCG_CALL_NO_RWG, i32, env, i32, i32)
28
+DEF_HELPER_FLAGS_3(mve_sqshl, TCG_CALL_NO_RWG, i32, env, i32, i32)
29
diff --git a/target/arm/translate.h b/target/arm/translate.h
30
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/translate.h
32
+++ b/target/arm/translate.h
33
@@ -XXX,XX +XXX,XX @@ typedef void CryptoThreeOpFn(TCGv_ptr, TCGv_ptr, TCGv_ptr);
34
typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
35
typedef void WideShiftImmFn(TCGv_i64, TCGv_i64, int64_t shift);
36
typedef void WideShiftFn(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i32);
37
+typedef void ShiftImmFn(TCGv_i32, TCGv_i32, int32_t shift);
38
39
/**
40
* arm_tbflags_from_tb:
41
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
42
index XXXXXXX..XXXXXXX 100644
43
--- a/target/arm/t32.decode
44
+++ b/target/arm/t32.decode
45
@@ -XXX,XX +XXX,XX @@
46
47
&mve_shl_ri rdalo rdahi shim
48
&mve_shl_rr rdalo rdahi rm
49
+&mve_sh_ri rda shim
50
51
# rdahi: bits [3:1] from insn, bit 0 is 1
52
# rdalo: bits [3:1] from insn, bit 0 is 0
53
@@ -XXX,XX +XXX,XX @@
54
&mve_shl_ri shim=%imm5_12_6 rdalo=%rdalo_17 rdahi=%rdahi_9
55
@mve_shl_rr ....... .... . ... . rm:4 ... . .. .. .... \
56
&mve_shl_rr rdalo=%rdalo_17 rdahi=%rdahi_9
57
+@mve_sh_ri ....... .... . rda:4 . ... ... . .. .. .... \
58
+ &mve_sh_ri shim=%imm5_12_6
59
60
{
61
TST_xrri 1110101 0000 1 .... 0 ... 1111 .... .... @S_xrr_shi
62
@@ -XXX,XX +XXX,XX @@ BIC_rrri 1110101 0001 . .... 0 ... .... .... .... @s_rrr_shi
63
# the rest fall through (where ORR_rrri and MOV_rxri will end up
64
# handling them as r13 and r15 accesses with the same semantics as A32).
65
[
66
- LSLL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 00 1111 @mve_shl_ri
67
- LSRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 01 1111 @mve_shl_ri
68
- ASRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 10 1111 @mve_shl_ri
69
+ {
70
+ UQSHL_ri 1110101 0010 1 .... 0 ... 1111 .. 00 1111 @mve_sh_ri
71
+ LSLL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 00 1111 @mve_shl_ri
72
+ UQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 00 1111 @mve_shl_ri
73
+ }
74
75
- UQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 00 1111 @mve_shl_ri
76
- URSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 01 1111 @mve_shl_ri
77
- SRSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 10 1111 @mve_shl_ri
78
- SQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 11 1111 @mve_shl_ri
79
+ {
80
+ URSHR_ri 1110101 0010 1 .... 0 ... 1111 .. 01 1111 @mve_sh_ri
81
+ LSRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 01 1111 @mve_shl_ri
82
+ URSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 01 1111 @mve_shl_ri
83
+ }
84
+
85
+ {
86
+ SRSHR_ri 1110101 0010 1 .... 0 ... 1111 .. 10 1111 @mve_sh_ri
87
+ ASRL_ri 1110101 0010 1 ... 0 0 ... ... 1 .. 10 1111 @mve_shl_ri
88
+ SRSHRL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 10 1111 @mve_shl_ri
89
+ }
90
+
91
+ {
92
+ SQSHL_ri 1110101 0010 1 .... 0 ... 1111 .. 11 1111 @mve_sh_ri
93
+ SQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 11 1111 @mve_shl_ri
94
+ }
95
96
LSLL_rr 1110101 0010 1 ... 0 .... ... 1 0000 1101 @mve_shl_rr
97
ASRL_rr 1110101 0010 1 ... 0 .... ... 1 0010 1101 @mve_shl_rr
98
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
99
index XXXXXXX..XXXXXXX 100644
100
--- a/target/arm/mve_helper.c
101
+++ b/target/arm/mve_helper.c
102
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mve_uqrshll48)(CPUARMState *env, uint64_t n, uint32_t shift)
103
{
104
return do_uqrshl48_d(n, (int8_t)shift, true, &env->QF);
105
}
106
+
107
+uint32_t HELPER(mve_uqshl)(CPUARMState *env, uint32_t n, uint32_t shift)
108
+{
109
+ return do_uqrshl_bhs(n, (int8_t)shift, 32, false, &env->QF);
110
+}
111
+
112
+uint32_t HELPER(mve_sqshl)(CPUARMState *env, uint32_t n, uint32_t shift)
113
+{
114
+ return do_sqrshl_bhs(n, (int8_t)shift, 32, false, &env->QF);
115
+}
15
diff --git a/target/arm/translate.c b/target/arm/translate.c
116
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
index XXXXXXX..XXXXXXX 100644
117
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate.c
118
--- a/target/arm/translate.c
18
+++ b/target/arm/translate.c
119
+++ b/target/arm/translate.c
19
@@ -XXX,XX +XXX,XX @@ static void gen_singlestep_exception(DisasContext *s)
120
@@ -XXX,XX +XXX,XX @@ static void gen_srshr16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
20
}
121
122
static void gen_srshr32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
123
{
124
- TCGv_i32 t = tcg_temp_new_i32();
125
+ TCGv_i32 t;
126
127
+ /* Handle shift by the input size for the benefit of trans_SRSHR_ri */
128
+ if (sh == 32) {
129
+ tcg_gen_movi_i32(d, 0);
130
+ return;
131
+ }
132
+ t = tcg_temp_new_i32();
133
tcg_gen_extract_i32(t, a, sh - 1, 1);
134
tcg_gen_sari_i32(d, a, sh);
135
tcg_gen_add_i32(d, d, t);
136
@@ -XXX,XX +XXX,XX @@ static void gen_urshr16_i64(TCGv_i64 d, TCGv_i64 a, int64_t sh)
137
138
static void gen_urshr32_i32(TCGv_i32 d, TCGv_i32 a, int32_t sh)
139
{
140
- TCGv_i32 t = tcg_temp_new_i32();
141
+ TCGv_i32 t;
142
143
+ /* Handle shift by the input size for the benefit of trans_URSHR_ri */
144
+ if (sh == 32) {
145
+ tcg_gen_extract_i32(d, a, sh - 1, 1);
146
+ return;
147
+ }
148
+ t = tcg_temp_new_i32();
149
tcg_gen_extract_i32(t, a, sh - 1, 1);
150
tcg_gen_shri_i32(d, a, sh);
151
tcg_gen_add_i32(d, d, t);
152
@@ -XXX,XX +XXX,XX @@ static bool trans_SQRSHRL48_rr(DisasContext *s, arg_mve_shl_rr *a)
153
return do_mve_shl_rr(s, a, gen_helper_mve_sqrshrl48);
21
}
154
}
22
155
23
+static inline bool is_singlestepping(DisasContext *s)
156
+static bool do_mve_sh_ri(DisasContext *s, arg_mve_sh_ri *a, ShiftImmFn *fn)
24
+{
157
+{
25
+ /* Return true if we are singlestepping either because of
158
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
26
+ * architectural singlestep or QEMU gdbstub singlestep. This does
159
+ /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
27
+ * not include the command line '-singlestep' mode which is rather
160
+ return false;
28
+ * misnamed as it only means "one instruction per TB" and doesn't
161
+ }
29
+ * affect the code we generate.
162
+ if (!dc_isar_feature(aa32_mve, s) ||
30
+ */
163
+ !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
31
+ return s->singlestep_enabled || s->ss_active;
164
+ a->rda == 13 || a->rda == 15) {
32
+}
165
+ /* These rda cases are UNPREDICTABLE; we choose to UNDEF */
33
+
166
+ unallocated_encoding(s);
34
static void gen_smul_dual(TCGv_i32 a, TCGv_i32 b)
167
+ return true;
35
{
168
+ }
36
TCGv_i32 tmp1 = tcg_temp_new_i32();
169
+
37
@@ -XXX,XX +XXX,XX @@ static inline void gen_goto_tb(DisasContext *s, int n, target_ulong dest)
170
+ if (a->shim == 0) {
38
171
+ a->shim = 32;
39
static inline void gen_jmp (DisasContext *s, uint32_t dest)
172
+ }
40
{
173
+ fn(cpu_R[a->rda], cpu_R[a->rda], a->shim);
41
- if (unlikely(s->singlestep_enabled || s->ss_active)) {
174
+
42
+ if (unlikely(is_singlestepping(s))) {
175
+ return true;
43
/* An indirect jump so that we still trigger the debug exception. */
176
+}
44
if (s->thumb)
177
+
45
dest |= 1;
178
+static bool trans_URSHR_ri(DisasContext *s, arg_mve_sh_ri *a)
46
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
179
+{
47
((dc->pc >= next_page_start - 3) && insn_crosses_page(env, dc));
180
+ return do_mve_sh_ri(s, a, gen_urshr32_i32);
48
181
+}
49
} while (!dc->is_jmp && !tcg_op_buf_full() &&
182
+
50
- !cs->singlestep_enabled &&
183
+static bool trans_SRSHR_ri(DisasContext *s, arg_mve_sh_ri *a)
51
+ !is_singlestepping(dc) &&
184
+{
52
!singlestep &&
185
+ return do_mve_sh_ri(s, a, gen_srshr32_i32);
53
- !dc->ss_active &&
186
+}
54
!end_of_page &&
187
+
55
num_insns < max_insns);
188
+static void gen_mve_sqshl(TCGv_i32 r, TCGv_i32 n, int32_t shift)
56
189
+{
57
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
190
+ gen_helper_mve_sqshl(r, cpu_env, n, tcg_constant_i32(shift));
58
instruction was a conditional branch or trap, and the PC has
191
+}
59
already been written. */
192
+
60
gen_set_condexec(dc);
193
+static bool trans_SQSHL_ri(DisasContext *s, arg_mve_sh_ri *a)
61
- if (unlikely(cs->singlestep_enabled || dc->ss_active)) {
194
+{
62
+ if (unlikely(is_singlestepping(dc))) {
195
+ return do_mve_sh_ri(s, a, gen_mve_sqshl);
63
/* Unconditional and "condition passed" instruction codepath. */
196
+}
64
switch (dc->is_jmp) {
197
+
65
case DISAS_SWI:
198
+static void gen_mve_uqshl(TCGv_i32 r, TCGv_i32 n, int32_t shift)
66
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
199
+{
67
/* "Condition failed" instruction codepath for the branch/trap insn */
200
+ gen_helper_mve_uqshl(r, cpu_env, n, tcg_constant_i32(shift));
68
gen_set_label(dc->condlabel);
201
+}
69
gen_set_condexec(dc);
202
+
70
- if (unlikely(cs->singlestep_enabled || dc->ss_active)) {
203
+static bool trans_UQSHL_ri(DisasContext *s, arg_mve_sh_ri *a)
71
+ if (unlikely(is_singlestepping(dc))) {
204
+{
72
gen_set_pc_im(dc, dc->pc);
205
+ return do_mve_sh_ri(s, a, gen_mve_uqshl);
73
gen_singlestep_exception(dc);
206
+}
74
} else {
207
+
208
/*
209
* Multiply and multiply accumulate
210
*/
75
--
211
--
76
2.7.4
212
2.20.1
77
213
78
214
diff view generated by jsdifflib
1
On M profile, return from exceptions happen when code in Handler mode
1
Implement the MVE shifts by register, which perform
2
executes one of the following function call return instructions:
2
shifts on a single general-purpose register.
3
* POP or LDM which loads the PC
4
* LDR to PC
5
* BX register
6
and the new PC value is 0xFFxxxxxx.
7
8
QEMU tries to implement this by not treating the instruction
9
specially but then catching the attempt to execute from the magic
10
address value. This is not ideal, because:
11
* there are guest visible differences from the architecturally
12
specified behaviour (for instance jumping to 0xFFxxxxxx via a
13
different instruction should not cause an exception return but it
14
will in the QEMU implementation)
15
* we have to account for it in various places (like refusing to take
16
an interrupt if the PC is at a magic value, and making sure that
17
the MPU doesn't deny execution at the magic value addresses)
18
19
Drop these hacks, and instead implement exception return the way the
20
architecture specifies -- by having the relevant instructions check
21
for the magic value and raise the 'do an exception return' QEMU
22
internal exception immediately.
23
24
The effect on the generated code is minor:
25
26
bx lr, old code (and new code for Thread mode):
27
TCG:
28
mov_i32 tmp5,r14
29
movi_i32 tmp6,$0xfffffffffffffffe
30
and_i32 pc,tmp5,tmp6
31
movi_i32 tmp6,$0x1
32
and_i32 tmp5,tmp5,tmp6
33
st_i32 tmp5,env,$0x218
34
exit_tb $0x0
35
set_label $L0
36
exit_tb $0x7f2aabd61993
37
x86_64 generated code:
38
0x7f2aabe87019: mov %ebx,%ebp
39
0x7f2aabe8701b: and $0xfffffffffffffffe,%ebp
40
0x7f2aabe8701e: mov %ebp,0x3c(%r14)
41
0x7f2aabe87022: and $0x1,%ebx
42
0x7f2aabe87025: mov %ebx,0x218(%r14)
43
0x7f2aabe8702c: xor %eax,%eax
44
0x7f2aabe8702e: jmpq 0x7f2aabe7c016
45
46
bx lr, new code when in Handler mode:
47
TCG:
48
mov_i32 tmp5,r14
49
movi_i32 tmp6,$0xfffffffffffffffe
50
and_i32 pc,tmp5,tmp6
51
movi_i32 tmp6,$0x1
52
and_i32 tmp5,tmp5,tmp6
53
st_i32 tmp5,env,$0x218
54
movi_i32 tmp5,$0xffffffffff000000
55
brcond_i32 pc,tmp5,geu,$L1
56
exit_tb $0x0
57
set_label $L1
58
movi_i32 tmp5,$0x8
59
call exception_internal,$0x0,$0,env,tmp5
60
x86_64 generated code:
61
0x7fe8fa1264e3: mov %ebp,%ebx
62
0x7fe8fa1264e5: and $0xfffffffffffffffe,%ebx
63
0x7fe8fa1264e8: mov %ebx,0x3c(%r14)
64
0x7fe8fa1264ec: and $0x1,%ebp
65
0x7fe8fa1264ef: mov %ebp,0x218(%r14)
66
0x7fe8fa1264f6: cmp $0xff000000,%ebx
67
0x7fe8fa1264fc: jae 0x7fe8fa126509
68
0x7fe8fa126502: xor %eax,%eax
69
0x7fe8fa126504: jmpq 0x7fe8fa122016
70
0x7fe8fa126509: mov %r14,%rdi
71
0x7fe8fa12650c: mov $0x8,%esi
72
0x7fe8fa126511: mov $0x56095dbeccf5,%r10
73
0x7fe8fa12651b: callq *%r10
74
75
which is a difference of one cmp/branch-not-taken. This will
76
be lost in the noise of having to exit generated code and
77
look up the next TB anyway.
78
3
79
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
80
Reviewed-by: Richard Henderson <rth@twiddle.net>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
81
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Message-id: 20210628135835.6690-19-peter.maydell@linaro.org
82
Message-id: 1491844419-12485-9-git-send-email-peter.maydell@linaro.org
83
---
7
---
84
target/arm/translate.h | 4 +++
8
target/arm/helper-mve.h | 2 ++
85
target/arm/translate.c | 66 +++++++++++++++++++++++++++++++++++++++++++++-----
9
target/arm/translate.h | 1 +
86
2 files changed, 64 insertions(+), 6 deletions(-)
10
target/arm/t32.decode | 18 ++++++++++++++----
11
target/arm/mve_helper.c | 10 ++++++++++
12
target/arm/translate.c | 30 ++++++++++++++++++++++++++++++
13
5 files changed, 57 insertions(+), 4 deletions(-)
87
14
15
diff --git a/target/arm/helper-mve.h b/target/arm/helper-mve.h
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper-mve.h
18
+++ b/target/arm/helper-mve.h
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(mve_uqrshll48, TCG_CALL_NO_RWG, i64, env, i64, i32)
20
21
DEF_HELPER_FLAGS_3(mve_uqshl, TCG_CALL_NO_RWG, i32, env, i32, i32)
22
DEF_HELPER_FLAGS_3(mve_sqshl, TCG_CALL_NO_RWG, i32, env, i32, i32)
23
+DEF_HELPER_FLAGS_3(mve_uqrshl, TCG_CALL_NO_RWG, i32, env, i32, i32)
24
+DEF_HELPER_FLAGS_3(mve_sqrshr, TCG_CALL_NO_RWG, i32, env, i32, i32)
88
diff --git a/target/arm/translate.h b/target/arm/translate.h
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
89
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate.h
27
--- a/target/arm/translate.h
91
+++ b/target/arm/translate.h
28
+++ b/target/arm/translate.h
92
@@ -XXX,XX +XXX,XX @@ static void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
29
@@ -XXX,XX +XXX,XX @@ typedef void AtomicThreeOpFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGArg, MemOp);
93
#define DISAS_HVC 8
30
typedef void WideShiftImmFn(TCGv_i64, TCGv_i64, int64_t shift);
94
#define DISAS_SMC 9
31
typedef void WideShiftFn(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i32);
95
#define DISAS_YIELD 10
32
typedef void ShiftImmFn(TCGv_i32, TCGv_i32, int32_t shift);
96
+/* M profile branch which might be an exception return (and so needs
33
+typedef void ShiftFn(TCGv_i32, TCGv_ptr, TCGv_i32, TCGv_i32);
97
+ * custom end-of-TB code)
34
98
+ */
35
/**
99
+#define DISAS_BX_EXCRET 11
36
* arm_tbflags_from_tb:
100
37
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
101
#ifdef TARGET_AARCH64
38
index XXXXXXX..XXXXXXX 100644
102
void a64_translate_init(void);
39
--- a/target/arm/t32.decode
40
+++ b/target/arm/t32.decode
41
@@ -XXX,XX +XXX,XX @@
42
&mve_shl_ri rdalo rdahi shim
43
&mve_shl_rr rdalo rdahi rm
44
&mve_sh_ri rda shim
45
+&mve_sh_rr rda rm
46
47
# rdahi: bits [3:1] from insn, bit 0 is 1
48
# rdalo: bits [3:1] from insn, bit 0 is 0
49
@@ -XXX,XX +XXX,XX @@
50
&mve_shl_rr rdalo=%rdalo_17 rdahi=%rdahi_9
51
@mve_sh_ri ....... .... . rda:4 . ... ... . .. .. .... \
52
&mve_sh_ri shim=%imm5_12_6
53
+@mve_sh_rr ....... .... . rda:4 rm:4 .... .... .... &mve_sh_rr
54
55
{
56
TST_xrri 1110101 0000 1 .... 0 ... 1111 .... .... @S_xrr_shi
57
@@ -XXX,XX +XXX,XX @@ BIC_rrri 1110101 0001 . .... 0 ... .... .... .... @s_rrr_shi
58
SQSHLL_ri 1110101 0010 1 ... 1 0 ... ... 1 .. 11 1111 @mve_shl_ri
59
}
60
61
- LSLL_rr 1110101 0010 1 ... 0 .... ... 1 0000 1101 @mve_shl_rr
62
- ASRL_rr 1110101 0010 1 ... 0 .... ... 1 0010 1101 @mve_shl_rr
63
- UQRSHLL64_rr 1110101 0010 1 ... 1 .... ... 1 0000 1101 @mve_shl_rr
64
- SQRSHRL64_rr 1110101 0010 1 ... 1 .... ... 1 0010 1101 @mve_shl_rr
65
+ {
66
+ UQRSHL_rr 1110101 0010 1 .... .... 1111 0000 1101 @mve_sh_rr
67
+ LSLL_rr 1110101 0010 1 ... 0 .... ... 1 0000 1101 @mve_shl_rr
68
+ UQRSHLL64_rr 1110101 0010 1 ... 1 .... ... 1 0000 1101 @mve_shl_rr
69
+ }
70
+
71
+ {
72
+ SQRSHR_rr 1110101 0010 1 .... .... 1111 0010 1101 @mve_sh_rr
73
+ ASRL_rr 1110101 0010 1 ... 0 .... ... 1 0010 1101 @mve_shl_rr
74
+ SQRSHRL64_rr 1110101 0010 1 ... 1 .... ... 1 0010 1101 @mve_shl_rr
75
+ }
76
+
77
UQRSHLL48_rr 1110101 0010 1 ... 1 .... ... 1 1000 1101 @mve_shl_rr
78
SQRSHRL48_rr 1110101 0010 1 ... 1 .... ... 1 1010 1101 @mve_shl_rr
79
]
80
diff --git a/target/arm/mve_helper.c b/target/arm/mve_helper.c
81
index XXXXXXX..XXXXXXX 100644
82
--- a/target/arm/mve_helper.c
83
+++ b/target/arm/mve_helper.c
84
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(mve_sqshl)(CPUARMState *env, uint32_t n, uint32_t shift)
85
{
86
return do_sqrshl_bhs(n, (int8_t)shift, 32, false, &env->QF);
87
}
88
+
89
+uint32_t HELPER(mve_uqrshl)(CPUARMState *env, uint32_t n, uint32_t shift)
90
+{
91
+ return do_uqrshl_bhs(n, (int8_t)shift, 32, true, &env->QF);
92
+}
93
+
94
+uint32_t HELPER(mve_sqrshr)(CPUARMState *env, uint32_t n, uint32_t shift)
95
+{
96
+ return do_sqrshl_bhs(n, -(int8_t)shift, 32, true, &env->QF);
97
+}
103
diff --git a/target/arm/translate.c b/target/arm/translate.c
98
diff --git a/target/arm/translate.c b/target/arm/translate.c
104
index XXXXXXX..XXXXXXX 100644
99
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/translate.c
100
--- a/target/arm/translate.c
106
+++ b/target/arm/translate.c
101
+++ b/target/arm/translate.c
107
@@ -XXX,XX +XXX,XX @@ static inline void gen_bx(DisasContext *s, TCGv_i32 var)
102
@@ -XXX,XX +XXX,XX @@ static bool trans_UQSHL_ri(DisasContext *s, arg_mve_sh_ri *a)
108
store_cpu_field(var, thumb);
103
return do_mve_sh_ri(s, a, gen_mve_uqshl);
109
}
104
}
110
105
111
+/* Set PC and Thumb state from var. var is marked as dead.
106
+static bool do_mve_sh_rr(DisasContext *s, arg_mve_sh_rr *a, ShiftFn *fn)
112
+ * For M-profile CPUs, include logic to detect exception-return
113
+ * branches and handle them. This is needed for Thumb POP/LDM to PC, LDR to PC,
114
+ * and BX reg, and no others, and happens only for code in Handler mode.
115
+ */
116
+static inline void gen_bx_excret(DisasContext *s, TCGv_i32 var)
117
+{
107
+{
118
+ /* Generate the same code here as for a simple bx, but flag via
108
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
119
+ * s->is_jmp that we need to do the rest of the work later.
109
+ /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
120
+ */
110
+ return false;
121
+ gen_bx(s, var);
122
+ if (s->v7m_handler_mode && arm_dc_feature(s, ARM_FEATURE_M)) {
123
+ s->is_jmp = DISAS_BX_EXCRET;
124
+ }
111
+ }
112
+ if (!dc_isar_feature(aa32_mve, s) ||
113
+ !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
114
+ a->rda == 13 || a->rda == 15 || a->rm == 13 || a->rm == 15 ||
115
+ a->rm == a->rda) {
116
+ /* These rda/rm cases are UNPREDICTABLE; we choose to UNDEF */
117
+ unallocated_encoding(s);
118
+ return true;
119
+ }
120
+
121
+ /* The helper takes care of the sign-extension of the low 8 bits of Rm */
122
+ fn(cpu_R[a->rda], cpu_env, cpu_R[a->rda], cpu_R[a->rm]);
123
+ return true;
125
+}
124
+}
126
+
125
+
127
+static inline void gen_bx_excret_final_code(DisasContext *s)
126
+static bool trans_SQRSHR_rr(DisasContext *s, arg_mve_sh_rr *a)
128
+{
127
+{
129
+ /* Generate the code to finish possible exception return and end the TB */
128
+ return do_mve_sh_rr(s, a, gen_helper_mve_sqrshr);
130
+ TCGLabel *excret_label = gen_new_label();
131
+
132
+ /* Is the new PC value in the magic range indicating exception return? */
133
+ tcg_gen_brcondi_i32(TCG_COND_GEU, cpu_R[15], 0xff000000, excret_label);
134
+ /* No: end the TB as we would for a DISAS_JMP */
135
+ if (is_singlestepping(s)) {
136
+ gen_singlestep_exception(s);
137
+ } else {
138
+ tcg_gen_exit_tb(0);
139
+ }
140
+ gen_set_label(excret_label);
141
+ /* Yes: this is an exception return.
142
+ * At this point in runtime env->regs[15] and env->thumb will hold
143
+ * the exception-return magic number, which do_v7m_exception_exit()
144
+ * will read. Nothing else will be able to see those values because
145
+ * the cpu-exec main loop guarantees that we will always go straight
146
+ * from raising the exception to the exception-handling code.
147
+ *
148
+ * gen_ss_advance(s) does nothing on M profile currently but
149
+ * calling it is conceptually the right thing as we have executed
150
+ * this instruction (compare SWI, HVC, SMC handling).
151
+ */
152
+ gen_ss_advance(s);
153
+ gen_exception_internal(EXCP_EXCEPTION_EXIT);
154
+}
129
+}
155
+
130
+
156
/* Variant of store_reg which uses branch&exchange logic when storing
131
+static bool trans_UQRSHL_rr(DisasContext *s, arg_mve_sh_rr *a)
157
to r15 in ARM architecture v7 and above. The source must be a temporary
132
+{
158
and will be marked as dead. */
133
+ return do_mve_sh_rr(s, a, gen_helper_mve_uqrshl);
159
@@ -XXX,XX +XXX,XX @@ static inline void store_reg_bx(DisasContext *s, int reg, TCGv_i32 var)
134
+}
160
static inline void store_reg_from_load(DisasContext *s, int reg, TCGv_i32 var)
135
+
161
{
136
/*
162
if (reg == 15 && ENABLE_ARCH_5) {
137
* Multiply and multiply accumulate
163
- gen_bx(s, var);
138
*/
164
+ gen_bx_excret(s, var);
165
} else {
166
store_reg(s, reg, var);
167
}
168
@@ -XXX,XX +XXX,XX @@ static int disas_thumb2_insn(CPUARMState *env, DisasContext *s, uint16_t insn_hw
169
tmp = tcg_temp_new_i32();
170
gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
171
if (i == 15) {
172
- gen_bx(s, tmp);
173
+ gen_bx_excret(s, tmp);
174
} else if (i == rn) {
175
loaded_var = tmp;
176
loaded_base = 1;
177
@@ -XXX,XX +XXX,XX @@ static int disas_thumb2_insn(CPUARMState *env, DisasContext *s, uint16_t insn_hw
178
goto illegal_op;
179
}
180
if (rs == 15) {
181
- gen_bx(s, tmp);
182
+ gen_bx_excret(s, tmp);
183
} else {
184
store_reg(s, rs, tmp);
185
}
186
@@ -XXX,XX +XXX,XX @@ static void disas_thumb_insn(CPUARMState *env, DisasContext *s)
187
tmp2 = tcg_temp_new_i32();
188
tcg_gen_movi_i32(tmp2, val);
189
store_reg(s, 14, tmp2);
190
+ gen_bx(s, tmp);
191
+ } else {
192
+ /* Only BX works as exception-return, not BLX */
193
+ gen_bx_excret(s, tmp);
194
}
195
- /* already thumb, no need to check */
196
- gen_bx(s, tmp);
197
break;
198
}
199
break;
200
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUARMState *env, TranslationBlock *tb)
201
instruction was a conditional branch or trap, and the PC has
202
already been written. */
203
gen_set_condexec(dc);
204
- if (unlikely(is_singlestepping(dc))) {
205
+ if (dc->is_jmp == DISAS_BX_EXCRET) {
206
+ /* Exception return branches need some special case code at the
207
+ * end of the TB, which is complex enough that it has to
208
+ * handle the single-step vs not and the condition-failed
209
+ * insn codepath itself.
210
+ */
211
+ gen_bx_excret_final_code(dc);
212
+ } else if (unlikely(is_singlestepping(dc))) {
213
/* Unconditional and "condition passed" instruction codepath. */
214
switch (dc->is_jmp) {
215
case DISAS_SWI:
216
--
139
--
217
2.7.4
140
2.20.1
218
141
219
142
diff view generated by jsdifflib