1 | Big fat pullreq this time around, because it has all of RTH's | 1 | The following changes since commit 8f6330a807f2642dc2a3cdf33347aa28a4c00a87: |
---|---|---|---|
2 | SVE2 emulation patchset in it. | ||
3 | 2 | ||
4 | -- PMM | 3 | Merge tag 'pull-maintainer-updates-060324-1' of https://gitlab.com/stsquad/qemu into staging (2024-03-06 16:56:20 +0000) |
5 | |||
6 | The following changes since commit 0dab1d36f55c3ed649bb8e4c74b9269ef3a63049: | ||
7 | |||
8 | Merge remote-tracking branch 'remotes/stefanha-gitlab/tags/block-pull-request' into staging (2021-05-24 15:48:08 +0100) | ||
9 | 4 | ||
10 | are available in the Git repository at: | 5 | are available in the Git repository at: |
11 | 6 | ||
12 | https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20210525 | 7 | https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20240308 |
13 | 8 | ||
14 | for you to fetch changes up to f8680aaa6e5bfc6022b75157c23db7d2ea98ab11: | 9 | for you to fetch changes up to bbf6c6dbead82292a20951eb1204442a6b838de9: |
15 | 10 | ||
16 | target/arm: Enable SVE2 and related extensions (2021-05-25 16:01:44 +0100) | 11 | target/arm: Move v7m-related code from cpu32.c into a separate file (2024-03-08 14:45:03 +0000) |
17 | 12 | ||
18 | ---------------------------------------------------------------- | 13 | ---------------------------------------------------------------- |
19 | target-arm queue: | 14 | target-arm queue: |
20 | * Implement SVE2 emulation | 15 | * Implement FEAT_ECV |
21 | * Implement integer matrix multiply accumulate | 16 | * STM32L4x5: Implement GPIO device |
22 | * Implement FEAT_TLBIOS | 17 | * Fix 32-bit SMOPA |
23 | * Implement FEAT_TLBRANGE | 18 | * Refactor v7m related code from cpu32.c into its own file |
24 | * disas/libvixl: Protect C system header for C++ compiler | 19 | * hw/rtc/sun4v-rtc: Relicense to GPLv2-or-later |
25 | * Use correct SP in M-profile exception return | ||
26 | * AN524, AN547: Correct modelling of internal SRAMs | ||
27 | * hw/intc/arm_gicv3_cpuif: Fix EOIR write access check logic | ||
28 | * hw/arm/smmuv3: Another range invalidation fix | ||
29 | 20 | ||
30 | ---------------------------------------------------------------- | 21 | ---------------------------------------------------------------- |
31 | Eric Auger (1): | 22 | Inès Varhol (3): |
32 | hw/arm/smmuv3: Another range invalidation fix | 23 | hw/gpio: Implement STM32L4x5 GPIO |
24 | hw/arm: Connect STM32L4x5 GPIO to STM32L4x5 SoC | ||
25 | tests/qtest: Add STM32L4x5 GPIO QTest testcase | ||
33 | 26 | ||
34 | Peter Maydell (8): | 27 | Peter Maydell (9): |
35 | hw/intc/arm_gicv3_cpuif: Fix EOIR write access check logic | 28 | target/arm: Move some register related defines to internals.h |
36 | hw/arm/mps2-tz: Don't duplicate modelling of SRAM in AN524 | 29 | target/arm: Timer _EL02 registers UNDEF for E2H == 0 |
37 | hw/arm/mps2-tz: Make SRAM_ADDR_WIDTH board-specific | 30 | target/arm: use FIELD macro for CNTHCTL bit definitions |
38 | hw/arm/armsse.c: Correct modelling of SSE-300 internal SRAMs | 31 | target/arm: Don't allow RES0 CNTHCTL_EL2 bits to be written |
39 | hw/arm/armsse: Convert armsse_realize() to use ERRP_GUARD | 32 | target/arm: Implement new FEAT_ECV trap bits |
40 | hw/arm/mps2-tz: Allow board to specify a boot RAM size | 33 | target/arm: Define CNTPCTSS_EL0 and CNTVCTSS_EL0 |
41 | hw/arm: Model TCMs in the SSE-300, not the AN547 | 34 | target/arm: Implement FEAT_ECV CNTPOFF_EL2 handling |
42 | target/arm: Use correct SP in M-profile exception return | 35 | target/arm: Enable FEAT_ECV for 'max' CPU |
36 | hw/rtc/sun4v-rtc: Relicense to GPLv2-or-later | ||
43 | 37 | ||
44 | Philippe Mathieu-Daudé (1): | 38 | Richard Henderson (1): |
45 | disas/libvixl: Protect C system header for C++ compiler | 39 | target/arm: Fix 32-bit SMOPA |
46 | 40 | ||
47 | Rebecca Cran (3): | 41 | Thomas Huth (1): |
48 | target/arm: Add support for FEAT_TLBIRANGE | 42 | target/arm: Move v7m-related code from cpu32.c into a separate file |
49 | target/arm: Add support for FEAT_TLBIOS | ||
50 | target/arm: set ID_AA64ISAR0.TLB to 2 for max AARCH64 CPU type | ||
51 | 43 | ||
52 | Richard Henderson (84): | 44 | MAINTAINERS | 1 + |
53 | accel/tcg: Replace g_new() + memcpy() by g_memdup() | 45 | docs/system/arm/b-l475e-iot01a.rst | 2 +- |
54 | accel/tcg: Pass length argument to tlb_flush_range_locked() | 46 | docs/system/arm/emulation.rst | 1 + |
55 | accel/tlb: Rename TLBFlushPageBitsByMMUIdxData -> TLBFlushRangeData | 47 | include/hw/arm/stm32l4x5_soc.h | 2 + |
56 | accel/tcg: Remove {encode,decode}_pbm_to_runon | 48 | include/hw/gpio/stm32l4x5_gpio.h | 71 +++++ |
57 | accel/tcg: Add tlb_flush_range_by_mmuidx() | 49 | include/hw/misc/stm32l4x5_syscfg.h | 3 +- |
58 | accel/tcg: Add tlb_flush_range_by_mmuidx_all_cpus() | 50 | include/hw/rtc/sun4v-rtc.h | 2 +- |
59 | accel/tlb: Add tlb_flush_range_by_mmuidx_all_cpus_synced() | 51 | target/arm/cpu-features.h | 10 + |
60 | accel/tcg: Rename tlb_flush_page_bits -> range]_by_mmuidx_async_0 | 52 | target/arm/cpu.h | 129 +-------- |
61 | accel/tlb: Rename tlb_flush_[page_bits > range]_by_mmuidx_async_[2 > 1] | 53 | target/arm/internals.h | 151 ++++++++++ |
62 | target/arm: Add ID_AA64ZFR0 fields and isar_feature_aa64_sve2 | 54 | hw/arm/stm32l4x5_soc.c | 71 ++++- |
63 | target/arm: Implement SVE2 Integer Multiply - Unpredicated | 55 | hw/gpio/stm32l4x5_gpio.c | 477 ++++++++++++++++++++++++++++++++ |
64 | target/arm: Implement SVE2 integer pairwise add and accumulate long | 56 | hw/misc/stm32l4x5_syscfg.c | 1 + |
65 | target/arm: Implement SVE2 integer unary operations (predicated) | 57 | hw/rtc/sun4v-rtc.c | 2 +- |
66 | target/arm: Split out saturating/rounding shifts from neon | 58 | target/arm/helper.c | 189 ++++++++++++- |
67 | target/arm: Implement SVE2 saturating/rounding bitwise shift left (predicated) | 59 | target/arm/tcg/cpu-v7m.c | 290 +++++++++++++++++++ |
68 | target/arm: Implement SVE2 integer halving add/subtract (predicated) | 60 | target/arm/tcg/cpu32.c | 261 ------------------ |
69 | target/arm: Implement SVE2 integer pairwise arithmetic | 61 | target/arm/tcg/cpu64.c | 1 + |
70 | target/arm: Implement SVE2 saturating add/subtract (predicated) | 62 | target/arm/tcg/sme_helper.c | 77 +++--- |
71 | target/arm: Implement SVE2 integer add/subtract long | 63 | tests/qtest/stm32l4x5_gpio-test.c | 551 +++++++++++++++++++++++++++++++++++++ |
72 | target/arm: Implement SVE2 integer add/subtract interleaved long | 64 | tests/tcg/aarch64/sme-smopa-1.c | 47 ++++ |
73 | target/arm: Implement SVE2 integer add/subtract wide | 65 | tests/tcg/aarch64/sme-smopa-2.c | 54 ++++ |
74 | target/arm: Implement SVE2 integer multiply long | 66 | hw/arm/Kconfig | 3 +- |
75 | target/arm: Implement SVE2 PMULLB, PMULLT | 67 | hw/gpio/Kconfig | 3 + |
76 | target/arm: Implement SVE2 bitwise shift left long | 68 | hw/gpio/meson.build | 1 + |
77 | target/arm: Implement SVE2 bitwise exclusive-or interleaved | 69 | hw/gpio/trace-events | 6 + |
78 | target/arm: Implement SVE2 bitwise permute | 70 | target/arm/meson.build | 3 + |
79 | target/arm: Implement SVE2 complex integer add | 71 | target/arm/tcg/meson.build | 3 + |
80 | target/arm: Implement SVE2 integer absolute difference and accumulate long | 72 | target/arm/trace-events | 1 + |
81 | target/arm: Implement SVE2 integer add/subtract long with carry | 73 | tests/qtest/meson.build | 3 +- |
82 | target/arm: Implement SVE2 bitwise shift right and accumulate | 74 | tests/tcg/aarch64/Makefile.target | 2 +- |
83 | target/arm: Implement SVE2 bitwise shift and insert | 75 | 31 files changed, 1962 insertions(+), 456 deletions(-) |
84 | target/arm: Implement SVE2 integer absolute difference and accumulate | 76 | create mode 100644 include/hw/gpio/stm32l4x5_gpio.h |
85 | target/arm: Implement SVE2 saturating extract narrow | 77 | create mode 100644 hw/gpio/stm32l4x5_gpio.c |
86 | target/arm: Implement SVE2 SHRN, RSHRN | 78 | create mode 100644 target/arm/tcg/cpu-v7m.c |
87 | target/arm: Implement SVE2 SQSHRUN, SQRSHRUN | 79 | create mode 100644 tests/qtest/stm32l4x5_gpio-test.c |
88 | target/arm: Implement SVE2 UQSHRN, UQRSHRN | 80 | create mode 100644 tests/tcg/aarch64/sme-smopa-1.c |
89 | target/arm: Implement SVE2 SQSHRN, SQRSHRN | 81 | create mode 100644 tests/tcg/aarch64/sme-smopa-2.c |
90 | target/arm: Implement SVE2 WHILEGT, WHILEGE, WHILEHI, WHILEHS | ||
91 | target/arm: Implement SVE2 WHILERW, WHILEWR | ||
92 | target/arm: Implement SVE2 bitwise ternary operations | ||
93 | target/arm: Implement SVE2 saturating multiply-add long | ||
94 | target/arm: Implement SVE2 saturating multiply-add high | ||
95 | target/arm: Implement SVE2 integer multiply-add long | ||
96 | target/arm: Implement SVE2 complex integer multiply-add | ||
97 | target/arm: Implement SVE2 XAR | ||
98 | target/arm: Use correct output type for gvec_sdot_*_b | ||
99 | target/arm: Pass separate addend to {U, S}DOT helpers | ||
100 | target/arm: Pass separate addend to FCMLA helpers | ||
101 | target/arm: Split out formats for 2 vectors + 1 index | ||
102 | target/arm: Split out formats for 3 vectors + 1 index | ||
103 | target/arm: Implement SVE2 integer multiply (indexed) | ||
104 | target/arm: Implement SVE2 integer multiply-add (indexed) | ||
105 | target/arm: Implement SVE2 saturating multiply-add high (indexed) | ||
106 | target/arm: Implement SVE2 saturating multiply-add (indexed) | ||
107 | target/arm: Implement SVE2 saturating multiply (indexed) | ||
108 | target/arm: Implement SVE2 signed saturating doubling multiply high | ||
109 | target/arm: Implement SVE2 saturating multiply high (indexed) | ||
110 | target/arm: Implement SVE2 multiply-add long (indexed) | ||
111 | target/arm: Implement SVE2 integer multiply long (indexed) | ||
112 | target/arm: Implement SVE2 complex integer multiply-add (indexed) | ||
113 | target/arm: Implement SVE2 complex integer dot product | ||
114 | target/arm: Macroize helper_gvec_{s,u}dot_{b,h} | ||
115 | target/arm: Macroize helper_gvec_{s,u}dot_idx_{b,h} | ||
116 | target/arm: Implement SVE mixed sign dot product (indexed) | ||
117 | target/arm: Implement SVE mixed sign dot product | ||
118 | target/arm: Implement SVE2 crypto unary operations | ||
119 | target/arm: Implement SVE2 crypto destructive binary operations | ||
120 | target/arm: Implement SVE2 crypto constructive binary operations | ||
121 | target/arm: Implement SVE2 FCVTNT | ||
122 | target/arm: Share table of sve load functions | ||
123 | target/arm: Tidy do_ldrq | ||
124 | target/arm: Implement SVE2 LD1RO | ||
125 | target/arm: Implement 128-bit ZIP, UZP, TRN | ||
126 | target/arm: Move endian adjustment macros to vec_internal.h | ||
127 | target/arm: Implement aarch64 SUDOT, USDOT | ||
128 | target/arm: Split out do_neon_ddda_fpst | ||
129 | target/arm: Remove unused fpst from VDOT_scalar | ||
130 | target/arm: Fix decode for VDOT (indexed) | ||
131 | target/arm: Split out do_neon_ddda | ||
132 | target/arm: Split decode of VSDOT and VUDOT | ||
133 | target/arm: Implement aarch32 VSUDOT, VUSDOT | ||
134 | target/arm: Implement integer matrix multiply accumulate | ||
135 | linux-user/aarch64: Enable hwcap bits for sve2 and related extensions | ||
136 | target/arm: Enable SVE2 and related extensions | ||
137 | 82 | ||
138 | Stephen Long (17): | ||
139 | target/arm: Implement SVE2 floating-point pairwise | ||
140 | target/arm: Implement SVE2 MATCH, NMATCH | ||
141 | target/arm: Implement SVE2 ADDHNB, ADDHNT | ||
142 | target/arm: Implement SVE2 RADDHNB, RADDHNT | ||
143 | target/arm: Implement SVE2 SUBHNB, SUBHNT | ||
144 | target/arm: Implement SVE2 RSUBHNB, RSUBHNT | ||
145 | target/arm: Implement SVE2 HISTCNT, HISTSEG | ||
146 | target/arm: Implement SVE2 scatter store insns | ||
147 | target/arm: Implement SVE2 gather load insns | ||
148 | target/arm: Implement SVE2 FMMLA | ||
149 | target/arm: Implement SVE2 SPLICE, EXT | ||
150 | target/arm: Implement SVE2 TBL, TBX | ||
151 | target/arm: Implement SVE2 FCVTLT | ||
152 | target/arm: Implement SVE2 FCVTXNT, FCVTX | ||
153 | target/arm: Implement SVE2 FLOGB | ||
154 | target/arm: Implement SVE2 bitwise shift immediate | ||
155 | target/arm: Implement SVE2 fp multiply-add long | ||
156 | |||
157 | disas/libvixl/vixl/code-buffer.h | 2 +- | ||
158 | disas/libvixl/vixl/globals.h | 16 +- | ||
159 | disas/libvixl/vixl/invalset.h | 2 +- | ||
160 | disas/libvixl/vixl/platform.h | 2 + | ||
161 | disas/libvixl/vixl/utils.h | 2 +- | ||
162 | include/exec/exec-all.h | 44 + | ||
163 | include/hw/arm/armsse.h | 2 + | ||
164 | target/arm/cpu.h | 76 + | ||
165 | target/arm/helper-sve.h | 722 ++++++++- | ||
166 | target/arm/helper.h | 110 +- | ||
167 | target/arm/translate-a64.h | 3 + | ||
168 | target/arm/vec_internal.h | 167 ++ | ||
169 | target/arm/neon-shared.decode | 24 +- | ||
170 | target/arm/sve.decode | 574 ++++++- | ||
171 | accel/tcg/cputlb.c | 231 ++- | ||
172 | hw/arm/armsse.c | 35 +- | ||
173 | hw/arm/mps2-tz.c | 39 +- | ||
174 | hw/arm/smmuv3.c | 50 +- | ||
175 | hw/intc/arm_gicv3_cpuif.c | 48 +- | ||
176 | linux-user/elfload.c | 10 + | ||
177 | target/arm/cpu.c | 2 + | ||
178 | target/arm/cpu64.c | 14 + | ||
179 | target/arm/cpu_tcg.c | 1 + | ||
180 | target/arm/helper.c | 327 +++- | ||
181 | target/arm/kvm64.c | 21 +- | ||
182 | target/arm/m_helper.c | 3 +- | ||
183 | target/arm/neon_helper.c | 507 +----- | ||
184 | target/arm/sve_helper.c | 2110 +++++++++++++++++++++++-- | ||
185 | target/arm/translate-a64.c | 111 +- | ||
186 | target/arm/translate-neon.c | 231 +-- | ||
187 | target/arm/translate-sve.c | 3200 +++++++++++++++++++++++++++++++++++--- | ||
188 | target/arm/vec_helper.c | 887 ++++++++--- | ||
189 | disas/libvixl/vixl/utils.cc | 2 +- | ||
190 | 33 files changed, 8275 insertions(+), 1300 deletions(-) | ||
191 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Eric Auger <eric.auger@redhat.com> | ||
2 | 1 | ||
3 | 6d9cd115b9 ("hw/arm/smmuv3: Enforce invalidation on a power of two range") | ||
4 | failed to completely fix misalignment issues with range | ||
5 | invalidation. For instance invalidations patterns like "invalidate 32 | ||
6 | 4kB pages starting from 0xff395000 are not correctly handled" due | ||
7 | to the fact the previous fix only made sure the number of invalidated | ||
8 | pages were a power of 2 but did not properly handle the start | ||
9 | address was not aligned with the range. This can be noticed when | ||
10 | boothing a fedora 33 with protected virtio-blk-pci. | ||
11 | |||
12 | Signed-off-by: Eric Auger <eric.auger@redhat.com> | ||
13 | Fixes: 6d9cd115b9 ("hw/arm/smmuv3: Enforce invalidation on a power of two range") | ||
14 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
15 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
16 | --- | ||
17 | hw/arm/smmuv3.c | 50 +++++++++++++++++++++++++------------------------ | ||
18 | 1 file changed, 26 insertions(+), 24 deletions(-) | ||
19 | |||
20 | diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c | ||
21 | index XXXXXXX..XXXXXXX 100644 | ||
22 | --- a/hw/arm/smmuv3.c | ||
23 | +++ b/hw/arm/smmuv3.c | ||
24 | @@ -XXX,XX +XXX,XX @@ static void smmuv3_inv_notifiers_iova(SMMUState *s, int asid, dma_addr_t iova, | ||
25 | |||
26 | static void smmuv3_s1_range_inval(SMMUState *s, Cmd *cmd) | ||
27 | { | ||
28 | - uint8_t scale = 0, num = 0, ttl = 0; | ||
29 | - dma_addr_t addr = CMD_ADDR(cmd); | ||
30 | + dma_addr_t end, addr = CMD_ADDR(cmd); | ||
31 | uint8_t type = CMD_TYPE(cmd); | ||
32 | uint16_t vmid = CMD_VMID(cmd); | ||
33 | + uint8_t scale = CMD_SCALE(cmd); | ||
34 | + uint8_t num = CMD_NUM(cmd); | ||
35 | + uint8_t ttl = CMD_TTL(cmd); | ||
36 | bool leaf = CMD_LEAF(cmd); | ||
37 | uint8_t tg = CMD_TG(cmd); | ||
38 | - uint64_t first_page = 0, last_page; | ||
39 | - uint64_t num_pages = 1; | ||
40 | + uint64_t num_pages; | ||
41 | + uint8_t granule; | ||
42 | int asid = -1; | ||
43 | |||
44 | - if (tg) { | ||
45 | - scale = CMD_SCALE(cmd); | ||
46 | - num = CMD_NUM(cmd); | ||
47 | - ttl = CMD_TTL(cmd); | ||
48 | - num_pages = (num + 1) * BIT_ULL(scale); | ||
49 | - } | ||
50 | - | ||
51 | if (type == SMMU_CMD_TLBI_NH_VA) { | ||
52 | asid = CMD_ASID(cmd); | ||
53 | } | ||
54 | |||
55 | + if (!tg) { | ||
56 | + trace_smmuv3_s1_range_inval(vmid, asid, addr, tg, 1, ttl, leaf); | ||
57 | + smmuv3_inv_notifiers_iova(s, asid, addr, tg, 1); | ||
58 | + smmu_iotlb_inv_iova(s, asid, addr, tg, 1, ttl); | ||
59 | + return; | ||
60 | + } | ||
61 | + | ||
62 | + /* RIL in use */ | ||
63 | + | ||
64 | + num_pages = (num + 1) * BIT_ULL(scale); | ||
65 | + granule = tg * 2 + 10; | ||
66 | + | ||
67 | /* Split invalidations into ^2 range invalidations */ | ||
68 | - last_page = num_pages - 1; | ||
69 | - while (num_pages) { | ||
70 | - uint8_t granule = tg * 2 + 10; | ||
71 | - uint64_t mask, count; | ||
72 | + end = addr + (num_pages << granule) - 1; | ||
73 | |||
74 | - mask = dma_aligned_pow2_mask(first_page, last_page, 64 - granule); | ||
75 | - count = mask + 1; | ||
76 | + while (addr != end + 1) { | ||
77 | + uint64_t mask = dma_aligned_pow2_mask(addr, end, 64); | ||
78 | |||
79 | - trace_smmuv3_s1_range_inval(vmid, asid, addr, tg, count, ttl, leaf); | ||
80 | - smmuv3_inv_notifiers_iova(s, asid, addr, tg, count); | ||
81 | - smmu_iotlb_inv_iova(s, asid, addr, tg, count, ttl); | ||
82 | - | ||
83 | - num_pages -= count; | ||
84 | - first_page += count; | ||
85 | - addr += count * BIT_ULL(granule); | ||
86 | + num_pages = (mask + 1) >> granule; | ||
87 | + trace_smmuv3_s1_range_inval(vmid, asid, addr, tg, num_pages, ttl, leaf); | ||
88 | + smmuv3_inv_notifiers_iova(s, asid, addr, tg, num_pages); | ||
89 | + smmu_iotlb_inv_iova(s, asid, addr, tg, num_pages, ttl); | ||
90 | + addr += mask + 1; | ||
91 | } | ||
92 | } | ||
93 | |||
94 | -- | ||
95 | 2.20.1 | ||
96 | |||
97 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | In icc_eoir_write() we assume that we can identify the group of the | ||
2 | IRQ being completed based purely on which register is being written | ||
3 | to and the current CPU state, and that "CPU state matches group | ||
4 | indicated by register" is the only necessary access check. | ||
5 | 1 | ||
6 | This isn't correct: if the CPU is not in Secure state then EOIR1 will | ||
7 | only complete Group 1 NS IRQs, but if the CPU is in EL3 it can | ||
8 | complete both Group 1 S and Group 1 NS IRQs. (The pseudocode | ||
9 | ICC_EOIR1_EL1 makes this clear.) We were also missing the logic to | ||
10 | prevent EOIR0 writes completing G0 IRQs when they should not. | ||
11 | |||
12 | Rearrange the logic to first identify the group of the current | ||
13 | highest priority interrupt and then look at whether we should | ||
14 | complete it or ignore the access based on which register was accessed | ||
15 | and the state of the CPU. The resulting behavioural change is: | ||
16 | * EL3 can now complete G1NS interrupts | ||
17 | * G0 interrupt completion is now ignored if the GIC | ||
18 | and the CPU have the security extension enabled and | ||
19 | the CPU is not secure | ||
20 | |||
21 | Reported-by: Chan Kim <ckim@etri.re.kr> | ||
22 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
23 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
24 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
25 | Message-id: 20210510150016.24910-1-peter.maydell@linaro.org | ||
26 | --- | ||
27 | hw/intc/arm_gicv3_cpuif.c | 48 ++++++++++++++++++++++++++------------- | ||
28 | 1 file changed, 32 insertions(+), 16 deletions(-) | ||
29 | |||
30 | diff --git a/hw/intc/arm_gicv3_cpuif.c b/hw/intc/arm_gicv3_cpuif.c | ||
31 | index XXXXXXX..XXXXXXX 100644 | ||
32 | --- a/hw/intc/arm_gicv3_cpuif.c | ||
33 | +++ b/hw/intc/arm_gicv3_cpuif.c | ||
34 | @@ -XXX,XX +XXX,XX @@ static void icc_eoir_write(CPUARMState *env, const ARMCPRegInfo *ri, | ||
35 | GICv3CPUState *cs = icc_cs_from_env(env); | ||
36 | int irq = value & 0xffffff; | ||
37 | int grp; | ||
38 | + bool is_eoir0 = ri->crm == 8; | ||
39 | |||
40 | - if (icv_access(env, ri->crm == 8 ? HCR_FMO : HCR_IMO)) { | ||
41 | + if (icv_access(env, is_eoir0 ? HCR_FMO : HCR_IMO)) { | ||
42 | icv_eoir_write(env, ri, value); | ||
43 | return; | ||
44 | } | ||
45 | |||
46 | - trace_gicv3_icc_eoir_write(ri->crm == 8 ? 0 : 1, | ||
47 | + trace_gicv3_icc_eoir_write(is_eoir0 ? 0 : 1, | ||
48 | gicv3_redist_affid(cs), value); | ||
49 | |||
50 | - if (ri->crm == 8) { | ||
51 | - /* EOIR0 */ | ||
52 | - grp = GICV3_G0; | ||
53 | - } else { | ||
54 | - /* EOIR1 */ | ||
55 | - if (arm_is_secure(env)) { | ||
56 | - grp = GICV3_G1; | ||
57 | - } else { | ||
58 | - grp = GICV3_G1NS; | ||
59 | - } | ||
60 | - } | ||
61 | - | ||
62 | if (irq >= cs->gic->num_irq) { | ||
63 | /* This handles two cases: | ||
64 | * 1. If software writes the ID of a spurious interrupt [ie 1020-1023] | ||
65 | @@ -XXX,XX +XXX,XX @@ static void icc_eoir_write(CPUARMState *env, const ARMCPRegInfo *ri, | ||
66 | return; | ||
67 | } | ||
68 | |||
69 | - if (icc_highest_active_group(cs) != grp) { | ||
70 | - return; | ||
71 | + grp = icc_highest_active_group(cs); | ||
72 | + switch (grp) { | ||
73 | + case GICV3_G0: | ||
74 | + if (!is_eoir0) { | ||
75 | + return; | ||
76 | + } | ||
77 | + if (!(cs->gic->gicd_ctlr & GICD_CTLR_DS) | ||
78 | + && arm_feature(env, ARM_FEATURE_EL3) && !arm_is_secure(env)) { | ||
79 | + return; | ||
80 | + } | ||
81 | + break; | ||
82 | + case GICV3_G1: | ||
83 | + if (is_eoir0) { | ||
84 | + return; | ||
85 | + } | ||
86 | + if (!arm_is_secure(env)) { | ||
87 | + return; | ||
88 | + } | ||
89 | + break; | ||
90 | + case GICV3_G1NS: | ||
91 | + if (is_eoir0) { | ||
92 | + return; | ||
93 | + } | ||
94 | + if (!arm_is_el3_or_mon(env) && arm_is_secure(env)) { | ||
95 | + return; | ||
96 | + } | ||
97 | + break; | ||
98 | + default: | ||
99 | + g_assert_not_reached(); | ||
100 | } | ||
101 | |||
102 | icc_drop_prio(cs, grp); | ||
103 | -- | ||
104 | 2.20.1 | ||
105 | |||
106 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | The SRAM at 0x2000_0000 is part of the SSE-200 itself, and we model | ||
2 | it that way in hw/arm/armsse.c (along with the associated MPCs). We | ||
3 | incorrectly also added an entry to the RAMInfo array for the AN524 in | ||
4 | hw/arm/mps2-tz.c, which was pointless because the CPU would never see | ||
5 | it. Delete it. | ||
6 | 1 | ||
7 | The bug had no guest-visible effect because devices in the SSE-200 | ||
8 | take priority over those in the board model (armsse.c maps | ||
9 | s->board_memory at priority -2). | ||
10 | |||
11 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
13 | Message-id: 20210510190844.17799-2-peter.maydell@linaro.org | ||
14 | --- | ||
15 | hw/arm/mps2-tz.c | 8 +------- | ||
16 | 1 file changed, 1 insertion(+), 7 deletions(-) | ||
17 | |||
18 | diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/hw/arm/mps2-tz.c | ||
21 | +++ b/hw/arm/mps2-tz.c | ||
22 | @@ -XXX,XX +XXX,XX @@ static const RAMInfo an524_raminfo[] = { { | ||
23 | .size = 512 * KiB, | ||
24 | .mpc = 0, | ||
25 | .mrindex = 0, | ||
26 | - }, { | ||
27 | - .name = "sram", | ||
28 | - .base = 0x20000000, | ||
29 | - .size = 32 * 4 * KiB, | ||
30 | - .mpc = -1, | ||
31 | - .mrindex = 1, | ||
32 | }, { | ||
33 | /* We don't model QSPI flash yet; for now expose it as simple ROM */ | ||
34 | .name = "QSPI", | ||
35 | .base = 0x28000000, | ||
36 | .size = 8 * MiB, | ||
37 | .mpc = 1, | ||
38 | - .mrindex = 2, | ||
39 | + .mrindex = 1, | ||
40 | .flags = IS_ROM, | ||
41 | }, { | ||
42 | .name = "DDR", | ||
43 | -- | ||
44 | 2.20.1 | ||
45 | |||
46 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | The AN547 sets the SRAM_ADDR_WIDTH for the SSE-300 to 21; | ||
2 | since this is not the default value for the SSE-300, model this | ||
3 | in mps2-tz.c as a per-board value. | ||
4 | 1 | ||
5 | Reported-by: Devaraj Ranganna <devaraj.ranganna@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
8 | Message-id: 20210510190844.17799-3-peter.maydell@linaro.org | ||
9 | --- | ||
10 | hw/arm/mps2-tz.c | 6 ++++++ | ||
11 | 1 file changed, 6 insertions(+) | ||
12 | |||
13 | diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c | ||
14 | index XXXXXXX..XXXXXXX 100644 | ||
15 | --- a/hw/arm/mps2-tz.c | ||
16 | +++ b/hw/arm/mps2-tz.c | ||
17 | @@ -XXX,XX +XXX,XX @@ struct MPS2TZMachineClass { | ||
18 | int numirq; /* Number of external interrupts */ | ||
19 | int uart_overflow_irq; /* number of the combined UART overflow IRQ */ | ||
20 | uint32_t init_svtor; /* init-svtor setting for SSE */ | ||
21 | + uint32_t sram_addr_width; /* SRAM_ADDR_WIDTH setting for SSE */ | ||
22 | const RAMInfo *raminfo; | ||
23 | const char *armsse_type; | ||
24 | }; | ||
25 | @@ -XXX,XX +XXX,XX @@ static void mps2tz_common_init(MachineState *machine) | ||
26 | OBJECT(system_memory), &error_abort); | ||
27 | qdev_prop_set_uint32(iotkitdev, "EXP_NUMIRQ", mmc->numirq); | ||
28 | qdev_prop_set_uint32(iotkitdev, "init-svtor", mmc->init_svtor); | ||
29 | + qdev_prop_set_uint32(iotkitdev, "SRAM_ADDR_WIDTH", mmc->sram_addr_width); | ||
30 | qdev_connect_clock_in(iotkitdev, "MAINCLK", mms->sysclk); | ||
31 | qdev_connect_clock_in(iotkitdev, "S32KCLK", mms->s32kclk); | ||
32 | sysbus_realize(SYS_BUS_DEVICE(&mms->iotkit), &error_fatal); | ||
33 | @@ -XXX,XX +XXX,XX @@ static void mps2tz_an505_class_init(ObjectClass *oc, void *data) | ||
34 | mmc->numirq = 92; | ||
35 | mmc->uart_overflow_irq = 47; | ||
36 | mmc->init_svtor = 0x10000000; | ||
37 | + mmc->sram_addr_width = 15; | ||
38 | mmc->raminfo = an505_raminfo; | ||
39 | mmc->armsse_type = TYPE_IOTKIT; | ||
40 | mps2tz_set_default_ram_info(mmc); | ||
41 | @@ -XXX,XX +XXX,XX @@ static void mps2tz_an521_class_init(ObjectClass *oc, void *data) | ||
42 | mmc->numirq = 92; | ||
43 | mmc->uart_overflow_irq = 47; | ||
44 | mmc->init_svtor = 0x10000000; | ||
45 | + mmc->sram_addr_width = 15; | ||
46 | mmc->raminfo = an505_raminfo; /* AN521 is the same as AN505 here */ | ||
47 | mmc->armsse_type = TYPE_SSE200; | ||
48 | mps2tz_set_default_ram_info(mmc); | ||
49 | @@ -XXX,XX +XXX,XX @@ static void mps3tz_an524_class_init(ObjectClass *oc, void *data) | ||
50 | mmc->numirq = 95; | ||
51 | mmc->uart_overflow_irq = 47; | ||
52 | mmc->init_svtor = 0x10000000; | ||
53 | + mmc->sram_addr_width = 15; | ||
54 | mmc->raminfo = an524_raminfo; | ||
55 | mmc->armsse_type = TYPE_SSE200; | ||
56 | mps2tz_set_default_ram_info(mmc); | ||
57 | @@ -XXX,XX +XXX,XX @@ static void mps3tz_an547_class_init(ObjectClass *oc, void *data) | ||
58 | mmc->numirq = 96; | ||
59 | mmc->uart_overflow_irq = 48; | ||
60 | mmc->init_svtor = 0x00000000; | ||
61 | + mmc->sram_addr_width = 21; | ||
62 | mmc->raminfo = an547_raminfo; | ||
63 | mmc->armsse_type = TYPE_SSE300; | ||
64 | mps2tz_set_default_ram_info(mmc); | ||
65 | -- | ||
66 | 2.20.1 | ||
67 | |||
68 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | The SSE-300 was not correctly modelling its internal SRAMs: | ||
2 | * the SRAM address width default is 18 | ||
3 | * the SRAM is mapped at 0x2100_0000, not 0x2000_0000 like | ||
4 | the SSE-200 and IoTKit | ||
5 | 1 | ||
6 | The default address width is no longer guest-visible since | ||
7 | our only SSE-300 board sets it explicitly to a non-default | ||
8 | value, but following the hardware's default will help for | ||
9 | any future boards we need to model. | ||
10 | |||
11 | Reported-by: Devaraj Ranganna <devaraj.ranganna@linaro.org> | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
14 | Message-id: 20210510190844.17799-4-peter.maydell@linaro.org | ||
15 | --- | ||
16 | hw/arm/armsse.c | 8 ++++++-- | ||
17 | 1 file changed, 6 insertions(+), 2 deletions(-) | ||
18 | |||
19 | diff --git a/hw/arm/armsse.c b/hw/arm/armsse.c | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/hw/arm/armsse.c | ||
22 | +++ b/hw/arm/armsse.c | ||
23 | @@ -XXX,XX +XXX,XX @@ struct ARMSSEInfo { | ||
24 | const char *cpu_type; | ||
25 | uint32_t sse_version; | ||
26 | int sram_banks; | ||
27 | + uint32_t sram_bank_base; | ||
28 | int num_cpus; | ||
29 | uint32_t sys_version; | ||
30 | uint32_t iidr; | ||
31 | @@ -XXX,XX +XXX,XX @@ static Property sse300_properties[] = { | ||
32 | DEFINE_PROP_LINK("memory", ARMSSE, board_memory, TYPE_MEMORY_REGION, | ||
33 | MemoryRegion *), | ||
34 | DEFINE_PROP_UINT32("EXP_NUMIRQ", ARMSSE, exp_numirq, 64), | ||
35 | - DEFINE_PROP_UINT32("SRAM_ADDR_WIDTH", ARMSSE, sram_addr_width, 15), | ||
36 | + DEFINE_PROP_UINT32("SRAM_ADDR_WIDTH", ARMSSE, sram_addr_width, 18), | ||
37 | DEFINE_PROP_UINT32("init-svtor", ARMSSE, init_svtor, 0x10000000), | ||
38 | DEFINE_PROP_BOOL("CPU0_FPU", ARMSSE, cpu_fpu[0], true), | ||
39 | DEFINE_PROP_BOOL("CPU0_DSP", ARMSSE, cpu_dsp[0], true), | ||
40 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
41 | .sse_version = ARMSSE_IOTKIT, | ||
42 | .cpu_type = ARM_CPU_TYPE_NAME("cortex-m33"), | ||
43 | .sram_banks = 1, | ||
44 | + .sram_bank_base = 0x20000000, | ||
45 | .num_cpus = 1, | ||
46 | .sys_version = 0x41743, | ||
47 | .iidr = 0, | ||
48 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
49 | .sse_version = ARMSSE_SSE200, | ||
50 | .cpu_type = ARM_CPU_TYPE_NAME("cortex-m33"), | ||
51 | .sram_banks = 4, | ||
52 | + .sram_bank_base = 0x20000000, | ||
53 | .num_cpus = 2, | ||
54 | .sys_version = 0x22041743, | ||
55 | .iidr = 0, | ||
56 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
57 | .sse_version = ARMSSE_SSE300, | ||
58 | .cpu_type = ARM_CPU_TYPE_NAME("cortex-m55"), | ||
59 | .sram_banks = 2, | ||
60 | + .sram_bank_base = 0x21000000, | ||
61 | .num_cpus = 1, | ||
62 | .sys_version = 0x7e00043b, | ||
63 | .iidr = 0x74a0043b, | ||
64 | @@ -XXX,XX +XXX,XX @@ static void armsse_realize(DeviceState *dev, Error **errp) | ||
65 | /* Map the upstream end of the MPC into the right place... */ | ||
66 | sbd_mpc = SYS_BUS_DEVICE(&s->mpc[i]); | ||
67 | memory_region_add_subregion(&s->container, | ||
68 | - 0x20000000 + i * sram_bank_size, | ||
69 | + info->sram_bank_base + i * sram_bank_size, | ||
70 | sysbus_mmio_get_region(sbd_mpc, 1)); | ||
71 | /* ...and its register interface */ | ||
72 | memory_region_add_subregion(&s->container, 0x50083000 + i * 0x1000, | ||
73 | -- | ||
74 | 2.20.1 | ||
75 | |||
76 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | Convert armsse_realize() to use ERRP_GUARD(), following | ||
2 | the rules in include/qapi/error.h. | ||
3 | 1 | ||
4 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
5 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210510190844.17799-5-peter.maydell@linaro.org | ||
7 | --- | ||
8 | hw/arm/armsse.c | 8 ++++---- | ||
9 | 1 file changed, 4 insertions(+), 4 deletions(-) | ||
10 | |||
11 | diff --git a/hw/arm/armsse.c b/hw/arm/armsse.c | ||
12 | index XXXXXXX..XXXXXXX 100644 | ||
13 | --- a/hw/arm/armsse.c | ||
14 | +++ b/hw/arm/armsse.c | ||
15 | @@ -XXX,XX +XXX,XX @@ static void armsse_realize(DeviceState *dev, Error **errp) | ||
16 | const ARMSSEDeviceInfo *devinfo; | ||
17 | int i; | ||
18 | MemoryRegion *mr; | ||
19 | - Error *err = NULL; | ||
20 | SysBusDevice *sbd_apb_ppc0; | ||
21 | SysBusDevice *sbd_secctl; | ||
22 | DeviceState *dev_apb_ppc0; | ||
23 | @@ -XXX,XX +XXX,XX @@ static void armsse_realize(DeviceState *dev, Error **errp) | ||
24 | DeviceState *dev_splitter; | ||
25 | uint32_t addr_width_max; | ||
26 | |||
27 | + ERRP_GUARD(); | ||
28 | + | ||
29 | if (!s->board_memory) { | ||
30 | error_setg(errp, "memory property was not set"); | ||
31 | return; | ||
32 | @@ -XXX,XX +XXX,XX @@ static void armsse_realize(DeviceState *dev, Error **errp) | ||
33 | uint32_t sram_bank_size = 1 << s->sram_addr_width; | ||
34 | |||
35 | memory_region_init_ram(&s->sram[i], NULL, ramname, | ||
36 | - sram_bank_size, &err); | ||
37 | + sram_bank_size, errp); | ||
38 | g_free(ramname); | ||
39 | - if (err) { | ||
40 | - error_propagate(errp, err); | ||
41 | + if (*errp) { | ||
42 | return; | ||
43 | } | ||
44 | object_property_set_link(OBJECT(&s->mpc[i]), "downstream", | ||
45 | -- | ||
46 | 2.20.1 | ||
47 | |||
48 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | cpu.h has a lot of #defines relating to CPU register fields. |
---|---|---|---|
2 | Most of these aren't actually used outside target/arm code, | ||
3 | so there's no point in cluttering up the cpu.h file with them. | ||
4 | Move some easy ones to internals.h. | ||
2 | 5 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-71-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
7 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
8 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20240301183219.2424889-2-peter.maydell@linaro.org | ||
7 | --- | 10 | --- |
8 | target/arm/cpu.h | 5 +++++ | 11 | target/arm/cpu.h | 128 ----------------------------------------- |
9 | target/arm/sve.decode | 4 ++++ | 12 | target/arm/internals.h | 128 +++++++++++++++++++++++++++++++++++++++++ |
10 | target/arm/translate-sve.c | 16 ++++++++++++++++ | 13 | 2 files changed, 128 insertions(+), 128 deletions(-) |
11 | 3 files changed, 25 insertions(+) | ||
12 | 14 | ||
13 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | 15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h |
14 | index XXXXXXX..XXXXXXX 100644 | 16 | index XXXXXXX..XXXXXXX 100644 |
15 | --- a/target/arm/cpu.h | 17 | --- a/target/arm/cpu.h |
16 | +++ b/target/arm/cpu.h | 18 | +++ b/target/arm/cpu.h |
17 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2_bitperm(const ARMISARegisters *id) | 19 | @@ -XXX,XX +XXX,XX @@ typedef struct ARMGenericTimer { |
18 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, BITPERM) != 0; | 20 | uint64_t ctl; /* Timer Control register */ |
19 | } | 21 | } ARMGenericTimer; |
20 | 22 | ||
21 | +static inline bool isar_feature_aa64_sve2_sha3(const ARMISARegisters *id) | 23 | -#define VTCR_NSW (1u << 29) |
22 | +{ | 24 | -#define VTCR_NSA (1u << 30) |
23 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SHA3) != 0; | 25 | -#define VSTCR_SW VTCR_NSW |
24 | +} | 26 | -#define VSTCR_SA VTCR_NSA |
25 | + | 27 | - |
26 | static inline bool isar_feature_aa64_sve2_sm4(const ARMISARegisters *id) | 28 | /* Define a maximum sized vector register. |
27 | { | 29 | * For 32-bit, this is a 128-bit NEON/AdvSIMD register. |
28 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SM4) != 0; | 30 | * For 64-bit, this is a 2048-bit SVE register. |
29 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | 31 | @@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu); |
32 | #define SCTLR_SPINTMASK (1ULL << 62) /* FEAT_NMI */ | ||
33 | #define SCTLR_TIDCP (1ULL << 63) /* FEAT_TIDCP1 */ | ||
34 | |||
35 | -/* Bit definitions for CPACR (AArch32 only) */ | ||
36 | -FIELD(CPACR, CP10, 20, 2) | ||
37 | -FIELD(CPACR, CP11, 22, 2) | ||
38 | -FIELD(CPACR, TRCDIS, 28, 1) /* matches CPACR_EL1.TTA */ | ||
39 | -FIELD(CPACR, D32DIS, 30, 1) /* up to v7; RAZ in v8 */ | ||
40 | -FIELD(CPACR, ASEDIS, 31, 1) | ||
41 | - | ||
42 | -/* Bit definitions for CPACR_EL1 (AArch64 only) */ | ||
43 | -FIELD(CPACR_EL1, ZEN, 16, 2) | ||
44 | -FIELD(CPACR_EL1, FPEN, 20, 2) | ||
45 | -FIELD(CPACR_EL1, SMEN, 24, 2) | ||
46 | -FIELD(CPACR_EL1, TTA, 28, 1) /* matches CPACR.TRCDIS */ | ||
47 | - | ||
48 | -/* Bit definitions for HCPTR (AArch32 only) */ | ||
49 | -FIELD(HCPTR, TCP10, 10, 1) | ||
50 | -FIELD(HCPTR, TCP11, 11, 1) | ||
51 | -FIELD(HCPTR, TASE, 15, 1) | ||
52 | -FIELD(HCPTR, TTA, 20, 1) | ||
53 | -FIELD(HCPTR, TAM, 30, 1) /* matches CPTR_EL2.TAM */ | ||
54 | -FIELD(HCPTR, TCPAC, 31, 1) /* matches CPTR_EL2.TCPAC */ | ||
55 | - | ||
56 | -/* Bit definitions for CPTR_EL2 (AArch64 only) */ | ||
57 | -FIELD(CPTR_EL2, TZ, 8, 1) /* !E2H */ | ||
58 | -FIELD(CPTR_EL2, TFP, 10, 1) /* !E2H, matches HCPTR.TCP10 */ | ||
59 | -FIELD(CPTR_EL2, TSM, 12, 1) /* !E2H */ | ||
60 | -FIELD(CPTR_EL2, ZEN, 16, 2) /* E2H */ | ||
61 | -FIELD(CPTR_EL2, FPEN, 20, 2) /* E2H */ | ||
62 | -FIELD(CPTR_EL2, SMEN, 24, 2) /* E2H */ | ||
63 | -FIELD(CPTR_EL2, TTA, 28, 1) | ||
64 | -FIELD(CPTR_EL2, TAM, 30, 1) /* matches HCPTR.TAM */ | ||
65 | -FIELD(CPTR_EL2, TCPAC, 31, 1) /* matches HCPTR.TCPAC */ | ||
66 | - | ||
67 | -/* Bit definitions for CPTR_EL3 (AArch64 only) */ | ||
68 | -FIELD(CPTR_EL3, EZ, 8, 1) | ||
69 | -FIELD(CPTR_EL3, TFP, 10, 1) | ||
70 | -FIELD(CPTR_EL3, ESM, 12, 1) | ||
71 | -FIELD(CPTR_EL3, TTA, 20, 1) | ||
72 | -FIELD(CPTR_EL3, TAM, 30, 1) | ||
73 | -FIELD(CPTR_EL3, TCPAC, 31, 1) | ||
74 | - | ||
75 | -#define MDCR_MTPME (1U << 28) | ||
76 | -#define MDCR_TDCC (1U << 27) | ||
77 | -#define MDCR_HLP (1U << 26) /* MDCR_EL2 */ | ||
78 | -#define MDCR_SCCD (1U << 23) /* MDCR_EL3 */ | ||
79 | -#define MDCR_HCCD (1U << 23) /* MDCR_EL2 */ | ||
80 | -#define MDCR_EPMAD (1U << 21) | ||
81 | -#define MDCR_EDAD (1U << 20) | ||
82 | -#define MDCR_TTRF (1U << 19) | ||
83 | -#define MDCR_STE (1U << 18) /* MDCR_EL3 */ | ||
84 | -#define MDCR_SPME (1U << 17) /* MDCR_EL3 */ | ||
85 | -#define MDCR_HPMD (1U << 17) /* MDCR_EL2 */ | ||
86 | -#define MDCR_SDD (1U << 16) | ||
87 | -#define MDCR_SPD (3U << 14) | ||
88 | -#define MDCR_TDRA (1U << 11) | ||
89 | -#define MDCR_TDOSA (1U << 10) | ||
90 | -#define MDCR_TDA (1U << 9) | ||
91 | -#define MDCR_TDE (1U << 8) | ||
92 | -#define MDCR_HPME (1U << 7) | ||
93 | -#define MDCR_TPM (1U << 6) | ||
94 | -#define MDCR_TPMCR (1U << 5) | ||
95 | -#define MDCR_HPMN (0x1fU) | ||
96 | - | ||
97 | -/* Not all of the MDCR_EL3 bits are present in the 32-bit SDCR */ | ||
98 | -#define SDCR_VALID_MASK (MDCR_MTPME | MDCR_TDCC | MDCR_SCCD | \ | ||
99 | - MDCR_EPMAD | MDCR_EDAD | MDCR_TTRF | \ | ||
100 | - MDCR_STE | MDCR_SPME | MDCR_SPD) | ||
101 | - | ||
102 | #define CPSR_M (0x1fU) | ||
103 | #define CPSR_T (1U << 5) | ||
104 | #define CPSR_F (1U << 6) | ||
105 | @@ -XXX,XX +XXX,XX @@ FIELD(CPTR_EL3, TCPAC, 31, 1) | ||
106 | #define XPSR_NZCV CPSR_NZCV | ||
107 | #define XPSR_IT CPSR_IT | ||
108 | |||
109 | -#define TTBCR_N (7U << 0) /* TTBCR.EAE==0 */ | ||
110 | -#define TTBCR_T0SZ (7U << 0) /* TTBCR.EAE==1 */ | ||
111 | -#define TTBCR_PD0 (1U << 4) | ||
112 | -#define TTBCR_PD1 (1U << 5) | ||
113 | -#define TTBCR_EPD0 (1U << 7) | ||
114 | -#define TTBCR_IRGN0 (3U << 8) | ||
115 | -#define TTBCR_ORGN0 (3U << 10) | ||
116 | -#define TTBCR_SH0 (3U << 12) | ||
117 | -#define TTBCR_T1SZ (3U << 16) | ||
118 | -#define TTBCR_A1 (1U << 22) | ||
119 | -#define TTBCR_EPD1 (1U << 23) | ||
120 | -#define TTBCR_IRGN1 (3U << 24) | ||
121 | -#define TTBCR_ORGN1 (3U << 26) | ||
122 | -#define TTBCR_SH1 (1U << 28) | ||
123 | -#define TTBCR_EAE (1U << 31) | ||
124 | - | ||
125 | -FIELD(VTCR, T0SZ, 0, 6) | ||
126 | -FIELD(VTCR, SL0, 6, 2) | ||
127 | -FIELD(VTCR, IRGN0, 8, 2) | ||
128 | -FIELD(VTCR, ORGN0, 10, 2) | ||
129 | -FIELD(VTCR, SH0, 12, 2) | ||
130 | -FIELD(VTCR, TG0, 14, 2) | ||
131 | -FIELD(VTCR, PS, 16, 3) | ||
132 | -FIELD(VTCR, VS, 19, 1) | ||
133 | -FIELD(VTCR, HA, 21, 1) | ||
134 | -FIELD(VTCR, HD, 22, 1) | ||
135 | -FIELD(VTCR, HWU59, 25, 1) | ||
136 | -FIELD(VTCR, HWU60, 26, 1) | ||
137 | -FIELD(VTCR, HWU61, 27, 1) | ||
138 | -FIELD(VTCR, HWU62, 28, 1) | ||
139 | -FIELD(VTCR, NSW, 29, 1) | ||
140 | -FIELD(VTCR, NSA, 30, 1) | ||
141 | -FIELD(VTCR, DS, 32, 1) | ||
142 | -FIELD(VTCR, SL2, 33, 1) | ||
143 | - | ||
144 | /* Bit definitions for ARMv8 SPSR (PSTATE) format. | ||
145 | * Only these are valid when in AArch64 mode; in | ||
146 | * AArch32 mode SPSRs are basically CPSR-format. | ||
147 | @@ -XXX,XX +XXX,XX @@ static inline void xpsr_write(CPUARMState *env, uint32_t val, uint32_t mask) | ||
148 | #define HCR_TWEDEN (1ULL << 59) | ||
149 | #define HCR_TWEDEL MAKE_64BIT_MASK(60, 4) | ||
150 | |||
151 | -#define HCRX_ENAS0 (1ULL << 0) | ||
152 | -#define HCRX_ENALS (1ULL << 1) | ||
153 | -#define HCRX_ENASR (1ULL << 2) | ||
154 | -#define HCRX_FNXS (1ULL << 3) | ||
155 | -#define HCRX_FGTNXS (1ULL << 4) | ||
156 | -#define HCRX_SMPME (1ULL << 5) | ||
157 | -#define HCRX_TALLINT (1ULL << 6) | ||
158 | -#define HCRX_VINMI (1ULL << 7) | ||
159 | -#define HCRX_VFNMI (1ULL << 8) | ||
160 | -#define HCRX_CMOW (1ULL << 9) | ||
161 | -#define HCRX_MCE2 (1ULL << 10) | ||
162 | -#define HCRX_MSCEN (1ULL << 11) | ||
163 | - | ||
164 | -#define HPFAR_NS (1ULL << 63) | ||
165 | - | ||
166 | #define SCR_NS (1ULL << 0) | ||
167 | #define SCR_IRQ (1ULL << 1) | ||
168 | #define SCR_FIQ (1ULL << 2) | ||
169 | @@ -XXX,XX +XXX,XX @@ static inline void xpsr_write(CPUARMState *env, uint32_t val, uint32_t mask) | ||
170 | #define SCR_GPF (1ULL << 48) | ||
171 | #define SCR_NSE (1ULL << 62) | ||
172 | |||
173 | -#define HSTR_TTEE (1 << 16) | ||
174 | -#define HSTR_TJDBX (1 << 17) | ||
175 | - | ||
176 | -#define CNTHCTL_CNTVMASK (1 << 18) | ||
177 | -#define CNTHCTL_CNTPMASK (1 << 19) | ||
178 | - | ||
179 | /* Return the current FPSCR value. */ | ||
180 | uint32_t vfp_get_fpscr(CPUARMState *env); | ||
181 | void vfp_set_fpscr(CPUARMState *env, uint32_t val); | ||
182 | diff --git a/target/arm/internals.h b/target/arm/internals.h | ||
30 | index XXXXXXX..XXXXXXX 100644 | 183 | index XXXXXXX..XXXXXXX 100644 |
31 | --- a/target/arm/sve.decode | 184 | --- a/target/arm/internals.h |
32 | +++ b/target/arm/sve.decode | 185 | +++ b/target/arm/internals.h |
33 | @@ -XXX,XX +XXX,XX @@ AESMC 01000101 00 10000011100 decrypt:1 00000 rd:5 | 186 | @@ -XXX,XX +XXX,XX @@ FIELD(DBGWCR, WT, 20, 1) |
34 | AESE 01000101 00 10001 0 11100 0 ..... ..... @rdn_rm_e0 | 187 | FIELD(DBGWCR, MASK, 24, 5) |
35 | AESD 01000101 00 10001 0 11100 1 ..... ..... @rdn_rm_e0 | 188 | FIELD(DBGWCR, SSCE, 29, 1) |
36 | SM4E 01000101 00 10001 1 11100 0 ..... ..... @rdn_rm_e0 | 189 | |
37 | + | 190 | +#define VTCR_NSW (1u << 29) |
38 | +# SVE2 crypto constructive binary operations | 191 | +#define VTCR_NSA (1u << 30) |
39 | +SM4EKEY 01000101 00 1 ..... 11110 0 ..... ..... @rd_rn_rm_e0 | 192 | +#define VSTCR_SW VTCR_NSW |
40 | +RAX1 01000101 00 1 ..... 11110 1 ..... ..... @rd_rn_rm_e0 | 193 | +#define VSTCR_SA VTCR_NSA |
41 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | 194 | + |
42 | index XXXXXXX..XXXXXXX 100644 | 195 | +/* Bit definitions for CPACR (AArch32 only) */ |
43 | --- a/target/arm/translate-sve.c | 196 | +FIELD(CPACR, CP10, 20, 2) |
44 | +++ b/target/arm/translate-sve.c | 197 | +FIELD(CPACR, CP11, 22, 2) |
45 | @@ -XXX,XX +XXX,XX @@ static bool trans_SM4E(DisasContext *s, arg_rrr_esz *a) | 198 | +FIELD(CPACR, TRCDIS, 28, 1) /* matches CPACR_EL1.TTA */ |
46 | { | 199 | +FIELD(CPACR, D32DIS, 30, 1) /* up to v7; RAZ in v8 */ |
47 | return do_sm4(s, a, gen_helper_crypto_sm4e); | 200 | +FIELD(CPACR, ASEDIS, 31, 1) |
48 | } | 201 | + |
49 | + | 202 | +/* Bit definitions for CPACR_EL1 (AArch64 only) */ |
50 | +static bool trans_SM4EKEY(DisasContext *s, arg_rrr_esz *a) | 203 | +FIELD(CPACR_EL1, ZEN, 16, 2) |
51 | +{ | 204 | +FIELD(CPACR_EL1, FPEN, 20, 2) |
52 | + return do_sm4(s, a, gen_helper_crypto_sm4ekey); | 205 | +FIELD(CPACR_EL1, SMEN, 24, 2) |
53 | +} | 206 | +FIELD(CPACR_EL1, TTA, 28, 1) /* matches CPACR.TRCDIS */ |
54 | + | 207 | + |
55 | +static bool trans_RAX1(DisasContext *s, arg_rrr_esz *a) | 208 | +/* Bit definitions for HCPTR (AArch32 only) */ |
56 | +{ | 209 | +FIELD(HCPTR, TCP10, 10, 1) |
57 | + if (!dc_isar_feature(aa64_sve2_sha3, s)) { | 210 | +FIELD(HCPTR, TCP11, 11, 1) |
58 | + return false; | 211 | +FIELD(HCPTR, TASE, 15, 1) |
59 | + } | 212 | +FIELD(HCPTR, TTA, 20, 1) |
60 | + if (sve_access_check(s)) { | 213 | +FIELD(HCPTR, TAM, 30, 1) /* matches CPTR_EL2.TAM */ |
61 | + gen_gvec_fn_zzz(s, gen_gvec_rax1, MO_64, a->rd, a->rn, a->rm); | 214 | +FIELD(HCPTR, TCPAC, 31, 1) /* matches CPTR_EL2.TCPAC */ |
62 | + } | 215 | + |
63 | + return true; | 216 | +/* Bit definitions for CPTR_EL2 (AArch64 only) */ |
64 | +} | 217 | +FIELD(CPTR_EL2, TZ, 8, 1) /* !E2H */ |
218 | +FIELD(CPTR_EL2, TFP, 10, 1) /* !E2H, matches HCPTR.TCP10 */ | ||
219 | +FIELD(CPTR_EL2, TSM, 12, 1) /* !E2H */ | ||
220 | +FIELD(CPTR_EL2, ZEN, 16, 2) /* E2H */ | ||
221 | +FIELD(CPTR_EL2, FPEN, 20, 2) /* E2H */ | ||
222 | +FIELD(CPTR_EL2, SMEN, 24, 2) /* E2H */ | ||
223 | +FIELD(CPTR_EL2, TTA, 28, 1) | ||
224 | +FIELD(CPTR_EL2, TAM, 30, 1) /* matches HCPTR.TAM */ | ||
225 | +FIELD(CPTR_EL2, TCPAC, 31, 1) /* matches HCPTR.TCPAC */ | ||
226 | + | ||
227 | +/* Bit definitions for CPTR_EL3 (AArch64 only) */ | ||
228 | +FIELD(CPTR_EL3, EZ, 8, 1) | ||
229 | +FIELD(CPTR_EL3, TFP, 10, 1) | ||
230 | +FIELD(CPTR_EL3, ESM, 12, 1) | ||
231 | +FIELD(CPTR_EL3, TTA, 20, 1) | ||
232 | +FIELD(CPTR_EL3, TAM, 30, 1) | ||
233 | +FIELD(CPTR_EL3, TCPAC, 31, 1) | ||
234 | + | ||
235 | +#define MDCR_MTPME (1U << 28) | ||
236 | +#define MDCR_TDCC (1U << 27) | ||
237 | +#define MDCR_HLP (1U << 26) /* MDCR_EL2 */ | ||
238 | +#define MDCR_SCCD (1U << 23) /* MDCR_EL3 */ | ||
239 | +#define MDCR_HCCD (1U << 23) /* MDCR_EL2 */ | ||
240 | +#define MDCR_EPMAD (1U << 21) | ||
241 | +#define MDCR_EDAD (1U << 20) | ||
242 | +#define MDCR_TTRF (1U << 19) | ||
243 | +#define MDCR_STE (1U << 18) /* MDCR_EL3 */ | ||
244 | +#define MDCR_SPME (1U << 17) /* MDCR_EL3 */ | ||
245 | +#define MDCR_HPMD (1U << 17) /* MDCR_EL2 */ | ||
246 | +#define MDCR_SDD (1U << 16) | ||
247 | +#define MDCR_SPD (3U << 14) | ||
248 | +#define MDCR_TDRA (1U << 11) | ||
249 | +#define MDCR_TDOSA (1U << 10) | ||
250 | +#define MDCR_TDA (1U << 9) | ||
251 | +#define MDCR_TDE (1U << 8) | ||
252 | +#define MDCR_HPME (1U << 7) | ||
253 | +#define MDCR_TPM (1U << 6) | ||
254 | +#define MDCR_TPMCR (1U << 5) | ||
255 | +#define MDCR_HPMN (0x1fU) | ||
256 | + | ||
257 | +/* Not all of the MDCR_EL3 bits are present in the 32-bit SDCR */ | ||
258 | +#define SDCR_VALID_MASK (MDCR_MTPME | MDCR_TDCC | MDCR_SCCD | \ | ||
259 | + MDCR_EPMAD | MDCR_EDAD | MDCR_TTRF | \ | ||
260 | + MDCR_STE | MDCR_SPME | MDCR_SPD) | ||
261 | + | ||
262 | +#define TTBCR_N (7U << 0) /* TTBCR.EAE==0 */ | ||
263 | +#define TTBCR_T0SZ (7U << 0) /* TTBCR.EAE==1 */ | ||
264 | +#define TTBCR_PD0 (1U << 4) | ||
265 | +#define TTBCR_PD1 (1U << 5) | ||
266 | +#define TTBCR_EPD0 (1U << 7) | ||
267 | +#define TTBCR_IRGN0 (3U << 8) | ||
268 | +#define TTBCR_ORGN0 (3U << 10) | ||
269 | +#define TTBCR_SH0 (3U << 12) | ||
270 | +#define TTBCR_T1SZ (3U << 16) | ||
271 | +#define TTBCR_A1 (1U << 22) | ||
272 | +#define TTBCR_EPD1 (1U << 23) | ||
273 | +#define TTBCR_IRGN1 (3U << 24) | ||
274 | +#define TTBCR_ORGN1 (3U << 26) | ||
275 | +#define TTBCR_SH1 (1U << 28) | ||
276 | +#define TTBCR_EAE (1U << 31) | ||
277 | + | ||
278 | +FIELD(VTCR, T0SZ, 0, 6) | ||
279 | +FIELD(VTCR, SL0, 6, 2) | ||
280 | +FIELD(VTCR, IRGN0, 8, 2) | ||
281 | +FIELD(VTCR, ORGN0, 10, 2) | ||
282 | +FIELD(VTCR, SH0, 12, 2) | ||
283 | +FIELD(VTCR, TG0, 14, 2) | ||
284 | +FIELD(VTCR, PS, 16, 3) | ||
285 | +FIELD(VTCR, VS, 19, 1) | ||
286 | +FIELD(VTCR, HA, 21, 1) | ||
287 | +FIELD(VTCR, HD, 22, 1) | ||
288 | +FIELD(VTCR, HWU59, 25, 1) | ||
289 | +FIELD(VTCR, HWU60, 26, 1) | ||
290 | +FIELD(VTCR, HWU61, 27, 1) | ||
291 | +FIELD(VTCR, HWU62, 28, 1) | ||
292 | +FIELD(VTCR, NSW, 29, 1) | ||
293 | +FIELD(VTCR, NSA, 30, 1) | ||
294 | +FIELD(VTCR, DS, 32, 1) | ||
295 | +FIELD(VTCR, SL2, 33, 1) | ||
296 | + | ||
297 | +#define HCRX_ENAS0 (1ULL << 0) | ||
298 | +#define HCRX_ENALS (1ULL << 1) | ||
299 | +#define HCRX_ENASR (1ULL << 2) | ||
300 | +#define HCRX_FNXS (1ULL << 3) | ||
301 | +#define HCRX_FGTNXS (1ULL << 4) | ||
302 | +#define HCRX_SMPME (1ULL << 5) | ||
303 | +#define HCRX_TALLINT (1ULL << 6) | ||
304 | +#define HCRX_VINMI (1ULL << 7) | ||
305 | +#define HCRX_VFNMI (1ULL << 8) | ||
306 | +#define HCRX_CMOW (1ULL << 9) | ||
307 | +#define HCRX_MCE2 (1ULL << 10) | ||
308 | +#define HCRX_MSCEN (1ULL << 11) | ||
309 | + | ||
310 | +#define HPFAR_NS (1ULL << 63) | ||
311 | + | ||
312 | +#define HSTR_TTEE (1 << 16) | ||
313 | +#define HSTR_TJDBX (1 << 17) | ||
314 | + | ||
315 | +#define CNTHCTL_CNTVMASK (1 << 18) | ||
316 | +#define CNTHCTL_CNTPMASK (1 << 19) | ||
317 | + | ||
318 | /* We use a few fake FSR values for internal purposes in M profile. | ||
319 | * M profile cores don't have A/R format FSRs, but currently our | ||
320 | * get_phys_addr() code assumes A/R profile and reports failures via | ||
65 | -- | 321 | -- |
66 | 2.20.1 | 322 | 2.34.1 |
67 | 323 | ||
68 | 324 | diff view generated by jsdifflib |
1 | When an M-profile CPU is restoring registers from the stack on | 1 | The timer _EL02 registers should UNDEF for invalid accesses from EL2 |
---|---|---|---|
2 | exception return, the stack pointer to use is determined based on | 2 | or EL3 when HCR_EL2.E2H == 0, not take a cp access trap. We were |
3 | bits in the magic exception return type value. We were not getting | 3 | delivering the exception to EL2 with the wrong syndrome. |
4 | this logic entirely correct. | ||
5 | |||
6 | Whether we use one of the Secure stack pointers or one of the | ||
7 | Non-Secure stack pointers depends on the EXCRET.S bit. However, | ||
8 | whether we use the MSP or the PSP then depends on the SPSEL bit in | ||
9 | either the CONTROL_S or CONTROL_NS register. We were incorrectly | ||
10 | selecting MSP vs PSP based on the EXCRET.SPSEL bit. | ||
11 | |||
12 | (In the pseudocode this is in the PopStack() function, which calls | ||
13 | LookUpSp_with_security_mode() which in turn looks at the relevant | ||
14 | CONTROL.SPSEL bit.) | ||
15 | |||
16 | The buggy behaviour wasn't noticeable in most cases, because we write | ||
17 | EXCRET.SPSEL to the CONTROL.SPSEL bit for the S/NS register selected | ||
18 | by EXCRET.ES, so we only do the wrong thing when EXCRET.S and | ||
19 | EXCRET.ES are different. This will happen when secure code takes a | ||
20 | secure exception, which then tail-chains to a non-secure exception | ||
21 | which finally returns to the original secure code. | ||
22 | 4 | ||
23 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 5 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
24 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | 6 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> |
25 | Message-id: 20210520130905.2049-1-peter.maydell@linaro.org | 7 | Message-id: 20240301183219.2424889-3-peter.maydell@linaro.org |
26 | --- | 8 | --- |
27 | target/arm/m_helper.c | 3 ++- | 9 | target/arm/helper.c | 2 +- |
28 | 1 file changed, 2 insertions(+), 1 deletion(-) | 10 | 1 file changed, 1 insertion(+), 1 deletion(-) |
29 | 11 | ||
30 | diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c | 12 | diff --git a/target/arm/helper.c b/target/arm/helper.c |
31 | index XXXXXXX..XXXXXXX 100644 | 13 | index XXXXXXX..XXXXXXX 100644 |
32 | --- a/target/arm/m_helper.c | 14 | --- a/target/arm/helper.c |
33 | +++ b/target/arm/m_helper.c | 15 | +++ b/target/arm/helper.c |
34 | @@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu) | 16 | @@ -XXX,XX +XXX,XX @@ static CPAccessResult e2h_access(CPUARMState *env, const ARMCPRegInfo *ri, |
35 | * We use this limited C variable scope so we don't accidentally | 17 | return CP_ACCESS_OK; |
36 | * use 'frame_sp_p' after we do something that makes it invalid. | 18 | } |
37 | */ | 19 | if (!(arm_hcr_el2_eff(env) & HCR_E2H)) { |
38 | + bool spsel = env->v7m.control[return_to_secure] & R_V7M_CONTROL_SPSEL_MASK; | 20 | - return CP_ACCESS_TRAP; |
39 | uint32_t *frame_sp_p = get_v7m_sp_ptr(env, | 21 | + return CP_ACCESS_TRAP_UNCATEGORIZED; |
40 | return_to_secure, | 22 | } |
41 | !return_to_handler, | 23 | return CP_ACCESS_OK; |
42 | - return_to_sp_process); | 24 | } |
43 | + spsel); | ||
44 | uint32_t frameptr = *frame_sp_p; | ||
45 | bool pop_ok = true; | ||
46 | ARMMMUIdx mmu_idx; | ||
47 | -- | 25 | -- |
48 | 2.20.1 | 26 | 2.34.1 |
49 | |||
50 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | We prefer the FIELD macro over ad-hoc #defines for register bits; |
---|---|---|---|
2 | switch CNTHCTL to that style before we add any more bits. | ||
2 | 3 | ||
3 | Will be used for SVE2 isa subset enablement. | 4 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
5 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
6 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20240301183219.2424889-4-peter.maydell@linaro.org | ||
8 | --- | ||
9 | target/arm/internals.h | 27 +++++++++++++++++++++++++-- | ||
10 | target/arm/helper.c | 9 ++++----- | ||
11 | 2 files changed, 29 insertions(+), 7 deletions(-) | ||
4 | 12 | ||
5 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | 13 | diff --git a/target/arm/internals.h b/target/arm/internals.h |
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-2-richard.henderson@linaro.org | ||
8 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/cpu.h | 16 ++++++++++++++++ | ||
12 | target/arm/helper.c | 3 +-- | ||
13 | target/arm/kvm64.c | 21 +++++++++++++++------ | ||
14 | 3 files changed, 32 insertions(+), 8 deletions(-) | ||
15 | |||
16 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
17 | index XXXXXXX..XXXXXXX 100644 | 14 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/target/arm/cpu.h | 15 | --- a/target/arm/internals.h |
19 | +++ b/target/arm/cpu.h | 16 | +++ b/target/arm/internals.h |
20 | @@ -XXX,XX +XXX,XX @@ struct ARMCPU { | 17 | @@ -XXX,XX +XXX,XX @@ FIELD(VTCR, SL2, 33, 1) |
21 | uint64_t id_aa64mmfr2; | 18 | #define HSTR_TTEE (1 << 16) |
22 | uint64_t id_aa64dfr0; | 19 | #define HSTR_TJDBX (1 << 17) |
23 | uint64_t id_aa64dfr1; | 20 | |
24 | + uint64_t id_aa64zfr0; | 21 | -#define CNTHCTL_CNTVMASK (1 << 18) |
25 | } isar; | 22 | -#define CNTHCTL_CNTPMASK (1 << 19) |
26 | uint64_t midr; | 23 | +/* |
27 | uint32_t revidr; | 24 | + * Depending on the value of HCR_EL2.E2H, bits 0 and 1 |
28 | @@ -XXX,XX +XXX,XX @@ FIELD(ID_AA64DFR0, DOUBLELOCK, 36, 4) | 25 | + * have different bit definitions, and EL1PCTEN might be |
29 | FIELD(ID_AA64DFR0, TRACEFILT, 40, 4) | 26 | + * bit 0 or bit 10. We use _E2H1 and _E2H0 suffixes to |
30 | FIELD(ID_AA64DFR0, MTPMU, 48, 4) | 27 | + * disambiguate if necessary. |
31 | 28 | + */ | |
32 | +FIELD(ID_AA64ZFR0, SVEVER, 0, 4) | 29 | +FIELD(CNTHCTL, EL0PCTEN_E2H1, 0, 1) |
33 | +FIELD(ID_AA64ZFR0, AES, 4, 4) | 30 | +FIELD(CNTHCTL, EL0VCTEN_E2H1, 1, 1) |
34 | +FIELD(ID_AA64ZFR0, BITPERM, 16, 4) | 31 | +FIELD(CNTHCTL, EL1PCTEN_E2H0, 0, 1) |
35 | +FIELD(ID_AA64ZFR0, BFLOAT16, 20, 4) | 32 | +FIELD(CNTHCTL, EL1PCEN_E2H0, 1, 1) |
36 | +FIELD(ID_AA64ZFR0, SHA3, 32, 4) | 33 | +FIELD(CNTHCTL, EVNTEN, 2, 1) |
37 | +FIELD(ID_AA64ZFR0, SM4, 40, 4) | 34 | +FIELD(CNTHCTL, EVNTDIR, 3, 1) |
38 | +FIELD(ID_AA64ZFR0, I8MM, 44, 4) | 35 | +FIELD(CNTHCTL, EVNTI, 4, 4) |
39 | +FIELD(ID_AA64ZFR0, F32MM, 52, 4) | 36 | +FIELD(CNTHCTL, EL0VTEN, 8, 1) |
40 | +FIELD(ID_AA64ZFR0, F64MM, 56, 4) | 37 | +FIELD(CNTHCTL, EL0PTEN, 9, 1) |
41 | + | 38 | +FIELD(CNTHCTL, EL1PCTEN_E2H1, 10, 1) |
42 | FIELD(ID_DFR0, COPDBG, 0, 4) | 39 | +FIELD(CNTHCTL, EL1PTEN, 11, 1) |
43 | FIELD(ID_DFR0, COPSDBG, 4, 4) | 40 | +FIELD(CNTHCTL, ECV, 12, 1) |
44 | FIELD(ID_DFR0, MMAPDBG, 8, 4) | 41 | +FIELD(CNTHCTL, EL1TVT, 13, 1) |
45 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_ssbs(const ARMISARegisters *id) | 42 | +FIELD(CNTHCTL, EL1TVCT, 14, 1) |
46 | return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, SSBS) != 0; | 43 | +FIELD(CNTHCTL, EL1NVPCT, 15, 1) |
47 | } | 44 | +FIELD(CNTHCTL, EL1NVVCT, 16, 1) |
48 | 45 | +FIELD(CNTHCTL, EVNTIS, 17, 1) | |
49 | +static inline bool isar_feature_aa64_sve2(const ARMISARegisters *id) | 46 | +FIELD(CNTHCTL, CNTVMASK, 18, 1) |
50 | +{ | 47 | +FIELD(CNTHCTL, CNTPMASK, 19, 1) |
51 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SVEVER) != 0; | 48 | |
52 | +} | 49 | /* We use a few fake FSR values for internal purposes in M profile. |
53 | + | 50 | * M profile cores don't have A/R format FSRs, but currently our |
54 | /* | ||
55 | * Feature tests for "does this exist in either 32-bit or 64-bit?" | ||
56 | */ | ||
57 | diff --git a/target/arm/helper.c b/target/arm/helper.c | 51 | diff --git a/target/arm/helper.c b/target/arm/helper.c |
58 | index XXXXXXX..XXXXXXX 100644 | 52 | index XXXXXXX..XXXXXXX 100644 |
59 | --- a/target/arm/helper.c | 53 | --- a/target/arm/helper.c |
60 | +++ b/target/arm/helper.c | 54 | +++ b/target/arm/helper.c |
61 | @@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu) | 55 | @@ -XXX,XX +XXX,XX @@ static void gt_update_irq(ARMCPU *cpu, int timeridx) |
62 | .opc0 = 3, .opc1 = 0, .crn = 0, .crm = 4, .opc2 = 4, | 56 | * It is RES0 in Secure and NonSecure state. |
63 | .access = PL1_R, .type = ARM_CP_CONST, | 57 | */ |
64 | .accessfn = access_aa64_tid3, | 58 | if ((ss == ARMSS_Root || ss == ARMSS_Realm) && |
65 | - /* At present, only SVEver == 0 is defined anyway. */ | 59 | - ((timeridx == GTIMER_VIRT && (cnthctl & CNTHCTL_CNTVMASK)) || |
66 | - .resetvalue = 0 }, | 60 | - (timeridx == GTIMER_PHYS && (cnthctl & CNTHCTL_CNTPMASK)))) { |
67 | + .resetvalue = cpu->isar.id_aa64zfr0 }, | 61 | + ((timeridx == GTIMER_VIRT && (cnthctl & R_CNTHCTL_CNTVMASK_MASK)) || |
68 | { .name = "ID_AA64PFR5_EL1_RESERVED", .state = ARM_CP_STATE_AA64, | 62 | + (timeridx == GTIMER_PHYS && (cnthctl & R_CNTHCTL_CNTPMASK_MASK)))) { |
69 | .opc0 = 3, .opc1 = 0, .crn = 0, .crm = 4, .opc2 = 5, | 63 | irqstate = 0; |
70 | .access = PL1_R, .type = ARM_CP_CONST, | 64 | } |
71 | diff --git a/target/arm/kvm64.c b/target/arm/kvm64.c | 65 | |
72 | index XXXXXXX..XXXXXXX 100644 | 66 | @@ -XXX,XX +XXX,XX @@ static void gt_cnthctl_write(CPUARMState *env, const ARMCPRegInfo *ri, |
73 | --- a/target/arm/kvm64.c | 67 | { |
74 | +++ b/target/arm/kvm64.c | 68 | ARMCPU *cpu = env_archcpu(env); |
75 | @@ -XXX,XX +XXX,XX @@ bool kvm_arm_get_host_cpu_features(ARMHostCPUFeatures *ahcf) | 69 | uint32_t oldval = env->cp15.cnthctl_el2; |
76 | |||
77 | sve_supported = ioctl(fdarray[0], KVM_CHECK_EXTENSION, KVM_CAP_ARM_SVE) > 0; | ||
78 | |||
79 | - kvm_arm_destroy_scratch_host_vcpu(fdarray); | ||
80 | - | 70 | - |
81 | - if (err < 0) { | 71 | raw_write(env, ri, value); |
82 | - return false; | 72 | |
83 | - } | 73 | - if ((oldval ^ value) & CNTHCTL_CNTVMASK) { |
84 | - | 74 | + if ((oldval ^ value) & R_CNTHCTL_CNTVMASK_MASK) { |
85 | /* Add feature bits that can't appear until after VCPU init. */ | 75 | gt_update_irq(cpu, GTIMER_VIRT); |
86 | if (sve_supported) { | 76 | - } else if ((oldval ^ value) & CNTHCTL_CNTPMASK) { |
87 | t = ahcf->isar.id_aa64pfr0; | 77 | + } else if ((oldval ^ value) & R_CNTHCTL_CNTPMASK_MASK) { |
88 | t = FIELD_DP64(t, ID_AA64PFR0, SVE, 1); | 78 | gt_update_irq(cpu, GTIMER_PHYS); |
89 | ahcf->isar.id_aa64pfr0 = t; | ||
90 | + | ||
91 | + /* | ||
92 | + * Before v5.1, KVM did not support SVE and did not expose | ||
93 | + * ID_AA64ZFR0_EL1 even as RAZ. After v5.1, KVM still does | ||
94 | + * not expose the register to "user" requests like this | ||
95 | + * unless the host supports SVE. | ||
96 | + */ | ||
97 | + err |= read_sys_reg64(fdarray[2], &ahcf->isar.id_aa64zfr0, | ||
98 | + ARM64_SYS_REG(3, 0, 0, 4, 4)); | ||
99 | + } | ||
100 | + | ||
101 | + kvm_arm_destroy_scratch_host_vcpu(fdarray); | ||
102 | + | ||
103 | + if (err < 0) { | ||
104 | + return false; | ||
105 | } | 79 | } |
106 | 80 | } | |
107 | /* | ||
108 | -- | 81 | -- |
109 | 2.20.1 | 82 | 2.34.1 |
110 | 83 | ||
111 | 84 | diff view generated by jsdifflib |
1 | The SSE-300 has an ITCM at 0x0000_0000 and a DTCM at 0x2000_0000. | 1 | Don't allow the guest to write CNTHCTL_EL2 bits which don't exist. |
---|---|---|---|
2 | Currently we model these in the AN547 board, but this is conceptually | 2 | This is not strictly architecturally required, but it is how we've |
3 | wrong, because they are a part of the SSE-300 itself. Move the | 3 | tended to implement registers more recently. |
4 | modelling of the TCMs out of mps2-tz.c into sse300.c. | ||
5 | 4 | ||
6 | This has no guest-visible effects. | 5 | In particular, bits [19:18] are only present with FEAT_RME, |
6 | and bits [17:12] will only be present with FEAT_ECV. | ||
7 | 7 | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
9 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | 9 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> |
10 | Message-id: 20210510190844.17799-7-peter.maydell@linaro.org | 10 | Message-id: 20240301183219.2424889-5-peter.maydell@linaro.org |
11 | --- | 11 | --- |
12 | include/hw/arm/armsse.h | 2 ++ | 12 | target/arm/helper.c | 18 ++++++++++++++++++ |
13 | hw/arm/armsse.c | 19 +++++++++++++++++++ | 13 | 1 file changed, 18 insertions(+) |
14 | hw/arm/mps2-tz.c | 12 ------------ | ||
15 | 3 files changed, 21 insertions(+), 12 deletions(-) | ||
16 | 14 | ||
17 | diff --git a/include/hw/arm/armsse.h b/include/hw/arm/armsse.h | 15 | diff --git a/target/arm/helper.c b/target/arm/helper.c |
18 | index XXXXXXX..XXXXXXX 100644 | 16 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/include/hw/arm/armsse.h | 17 | --- a/target/arm/helper.c |
20 | +++ b/include/hw/arm/armsse.h | 18 | +++ b/target/arm/helper.c |
21 | @@ -XXX,XX +XXX,XX @@ struct ARMSSE { | 19 | @@ -XXX,XX +XXX,XX @@ static void gt_cnthctl_write(CPUARMState *env, const ARMCPRegInfo *ri, |
22 | MemoryRegion alias2; | 20 | { |
23 | MemoryRegion alias3[SSE_MAX_CPUS]; | 21 | ARMCPU *cpu = env_archcpu(env); |
24 | MemoryRegion sram[MAX_SRAM_BANKS]; | 22 | uint32_t oldval = env->cp15.cnthctl_el2; |
25 | + MemoryRegion itcm; | 23 | + uint32_t valid_mask = |
26 | + MemoryRegion dtcm; | 24 | + R_CNTHCTL_EL0PCTEN_E2H1_MASK | |
27 | 25 | + R_CNTHCTL_EL0VCTEN_E2H1_MASK | | |
28 | qemu_irq *exp_irqs[SSE_MAX_CPUS]; | 26 | + R_CNTHCTL_EVNTEN_MASK | |
29 | qemu_irq ppc0_irq; | 27 | + R_CNTHCTL_EVNTDIR_MASK | |
30 | diff --git a/hw/arm/armsse.c b/hw/arm/armsse.c | 28 | + R_CNTHCTL_EVNTI_MASK | |
31 | index XXXXXXX..XXXXXXX 100644 | 29 | + R_CNTHCTL_EL0VTEN_MASK | |
32 | --- a/hw/arm/armsse.c | 30 | + R_CNTHCTL_EL0PTEN_MASK | |
33 | +++ b/hw/arm/armsse.c | 31 | + R_CNTHCTL_EL1PCTEN_E2H1_MASK | |
34 | @@ -XXX,XX +XXX,XX @@ | 32 | + R_CNTHCTL_EL1PTEN_MASK; |
35 | #include "qemu/log.h" | 33 | + |
36 | #include "qemu/module.h" | 34 | + if (cpu_isar_feature(aa64_rme, cpu)) { |
37 | #include "qemu/bitops.h" | 35 | + valid_mask |= R_CNTHCTL_CNTVMASK_MASK | R_CNTHCTL_CNTPMASK_MASK; |
38 | +#include "qemu/units.h" | ||
39 | #include "qapi/error.h" | ||
40 | #include "trace.h" | ||
41 | #include "hw/sysbus.h" | ||
42 | @@ -XXX,XX +XXX,XX @@ struct ARMSSEInfo { | ||
43 | bool has_cpuid; | ||
44 | bool has_cpu_pwrctrl; | ||
45 | bool has_sse_counter; | ||
46 | + bool has_tcms; | ||
47 | Property *props; | ||
48 | const ARMSSEDeviceInfo *devinfo; | ||
49 | const bool *irq_is_common; | ||
50 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
51 | .has_cpuid = false, | ||
52 | .has_cpu_pwrctrl = false, | ||
53 | .has_sse_counter = false, | ||
54 | + .has_tcms = false, | ||
55 | .props = iotkit_properties, | ||
56 | .devinfo = iotkit_devices, | ||
57 | .irq_is_common = sse200_irq_is_common, | ||
58 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
59 | .has_cpuid = true, | ||
60 | .has_cpu_pwrctrl = false, | ||
61 | .has_sse_counter = false, | ||
62 | + .has_tcms = false, | ||
63 | .props = sse200_properties, | ||
64 | .devinfo = sse200_devices, | ||
65 | .irq_is_common = sse200_irq_is_common, | ||
66 | @@ -XXX,XX +XXX,XX @@ static const ARMSSEInfo armsse_variants[] = { | ||
67 | .has_cpuid = true, | ||
68 | .has_cpu_pwrctrl = true, | ||
69 | .has_sse_counter = true, | ||
70 | + .has_tcms = true, | ||
71 | .props = sse300_properties, | ||
72 | .devinfo = sse300_devices, | ||
73 | .irq_is_common = sse300_irq_is_common, | ||
74 | @@ -XXX,XX +XXX,XX @@ static void armsse_realize(DeviceState *dev, Error **errp) | ||
75 | sysbus_mmio_get_region(sbd, 1)); | ||
76 | } | ||
77 | |||
78 | + if (info->has_tcms) { | ||
79 | + /* The SSE-300 has an ITCM at 0x0000_0000 and a DTCM at 0x2000_0000 */ | ||
80 | + memory_region_init_ram(&s->itcm, NULL, "sse300-itcm", 512 * KiB, errp); | ||
81 | + if (*errp) { | ||
82 | + return; | ||
83 | + } | ||
84 | + memory_region_init_ram(&s->dtcm, NULL, "sse300-dtcm", 512 * KiB, errp); | ||
85 | + if (*errp) { | ||
86 | + return; | ||
87 | + } | ||
88 | + memory_region_add_subregion(&s->container, 0x00000000, &s->itcm); | ||
89 | + memory_region_add_subregion(&s->container, 0x20000000, &s->dtcm); | ||
90 | + } | 36 | + } |
91 | + | 37 | + |
92 | /* Devices behind APB PPC0: | 38 | + /* Clear RES0 bits */ |
93 | * 0x40000000: timer0 | 39 | + value &= valid_mask; |
94 | * 0x40001000: timer1 | 40 | + |
95 | diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c | 41 | raw_write(env, ri, value); |
96 | index XXXXXXX..XXXXXXX 100644 | 42 | |
97 | --- a/hw/arm/mps2-tz.c | 43 | if ((oldval ^ value) & R_CNTHCTL_CNTVMASK_MASK) { |
98 | +++ b/hw/arm/mps2-tz.c | ||
99 | @@ -XXX,XX +XXX,XX @@ static const RAMInfo an524_raminfo[] = { { | ||
100 | }; | ||
101 | |||
102 | static const RAMInfo an547_raminfo[] = { { | ||
103 | - .name = "itcm", | ||
104 | - .base = 0x00000000, | ||
105 | - .size = 512 * KiB, | ||
106 | - .mpc = -1, | ||
107 | - .mrindex = 0, | ||
108 | - }, { | ||
109 | .name = "sram", | ||
110 | .base = 0x01000000, | ||
111 | .size = 2 * MiB, | ||
112 | .mpc = 0, | ||
113 | .mrindex = 1, | ||
114 | - }, { | ||
115 | - .name = "dtcm", | ||
116 | - .base = 0x20000000, | ||
117 | - .size = 4 * 128 * KiB, | ||
118 | - .mpc = -1, | ||
119 | - .mrindex = 2, | ||
120 | }, { | ||
121 | .name = "sram 2", | ||
122 | .base = 0x21000000, | ||
123 | -- | 44 | -- |
124 | 2.20.1 | 45 | 2.34.1 |
125 | |||
126 | diff view generated by jsdifflib |
1 | From: Rebecca Cran <rebecca@nuviainc.com> | 1 | The functionality defined by ID_AA64MMFR0_EL1.ECV == 1 is: |
---|---|---|---|
2 | * four new trap bits for various counter and timer registers | ||
3 | * the CNTHCTL_EL2.EVNTIS and CNTKCTL_EL1.EVNTIS bits which control | ||
4 | scaling of the event stream. This is a no-op for us, because we don't | ||
5 | implement the event stream (our WFE is a NOP): all we need to do is | ||
6 | allow CNTHCTL_EL2.ENVTIS to be read and written. | ||
7 | * extensions to PMSCR_EL1.PCT, PMSCR_EL2.PCT, TRFCR_EL1.TS and | ||
8 | TRFCR_EL2.TS: these are all no-ops for us, because we don't implement | ||
9 | FEAT_SPE or FEAT_TRF. | ||
10 | * new registers CNTPCTSS_EL0 and NCTVCTSS_EL0 which are | ||
11 | "self-sychronizing" views of the CNTPCT_EL0 and CNTVCT_EL0, meaning | ||
12 | that no barriers are needed around their accesses. For us these | ||
13 | are just the same as the normal views, because all our sysregs are | ||
14 | inherently self-sychronizing. | ||
2 | 15 | ||
3 | ARMv8.4 adds the mandatory FEAT_TLBIRANGE. It provides TLBI | 16 | In this commit we implement the trap handling and permit the new |
4 | maintenance instructions that apply to a range of input addresses. | 17 | CNTHCTL_EL2 bits to be written. |
5 | 18 | ||
6 | Signed-off-by: Rebecca Cran <rebecca@nuviainc.com> | 19 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
7 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | 20 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> |
8 | Message-id: 20210512182337.18563-2-rebecca@nuviainc.com | 21 | Message-id: 20240301183219.2424889-6-peter.maydell@linaro.org |
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | 22 | --- |
11 | target/arm/cpu.h | 5 + | 23 | target/arm/cpu-features.h | 5 ++++ |
12 | target/arm/helper.c | 281 ++++++++++++++++++++++++++++++++++++++++++++ | 24 | target/arm/helper.c | 51 +++++++++++++++++++++++++++++++++++---- |
13 | 2 files changed, 286 insertions(+) | 25 | 2 files changed, 51 insertions(+), 5 deletions(-) |
14 | 26 | ||
15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | 27 | diff --git a/target/arm/cpu-features.h b/target/arm/cpu-features.h |
16 | index XXXXXXX..XXXXXXX 100644 | 28 | index XXXXXXX..XXXXXXX 100644 |
17 | --- a/target/arm/cpu.h | 29 | --- a/target/arm/cpu-features.h |
18 | +++ b/target/arm/cpu.h | 30 | +++ b/target/arm/cpu-features.h |
19 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_pauth_arch(const ARMISARegisters *id) | 31 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_fgt(const ARMISARegisters *id) |
20 | return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, APA) != 0; | 32 | return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR0, FGT) != 0; |
21 | } | 33 | } |
22 | 34 | ||
23 | +static inline bool isar_feature_aa64_tlbirange(const ARMISARegisters *id) | 35 | +static inline bool isar_feature_aa64_ecv_traps(const ARMISARegisters *id) |
24 | +{ | 36 | +{ |
25 | + return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, TLB) == 2; | 37 | + return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR0, ECV) > 0; |
26 | +} | 38 | +} |
27 | + | 39 | + |
28 | static inline bool isar_feature_aa64_sb(const ARMISARegisters *id) | 40 | static inline bool isar_feature_aa64_vh(const ARMISARegisters *id) |
29 | { | 41 | { |
30 | return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, SB) != 0; | 42 | return FIELD_EX64(id->id_aa64mmfr1, ID_AA64MMFR1, VH) != 0; |
31 | diff --git a/target/arm/helper.c b/target/arm/helper.c | 43 | diff --git a/target/arm/helper.c b/target/arm/helper.c |
32 | index XXXXXXX..XXXXXXX 100644 | 44 | index XXXXXXX..XXXXXXX 100644 |
33 | --- a/target/arm/helper.c | 45 | --- a/target/arm/helper.c |
34 | +++ b/target/arm/helper.c | 46 | +++ b/target/arm/helper.c |
35 | @@ -XXX,XX +XXX,XX @@ static void tlbi_aa64_vae3is_write(CPUARMState *env, const ARMCPRegInfo *ri, | 47 | @@ -XXX,XX +XXX,XX @@ static CPAccessResult gt_counter_access(CPUARMState *env, int timeridx, |
36 | ARMMMUIdxBit_SE3, bits); | 48 | : !extract32(env->cp15.cnthctl_el2, 0, 1))) { |
49 | return CP_ACCESS_TRAP_EL2; | ||
50 | } | ||
51 | + if (has_el2 && timeridx == GTIMER_VIRT) { | ||
52 | + if (FIELD_EX64(env->cp15.cnthctl_el2, CNTHCTL, EL1TVCT)) { | ||
53 | + return CP_ACCESS_TRAP_EL2; | ||
54 | + } | ||
55 | + } | ||
56 | break; | ||
57 | } | ||
58 | return CP_ACCESS_OK; | ||
59 | @@ -XXX,XX +XXX,XX @@ static CPAccessResult gt_timer_access(CPUARMState *env, int timeridx, | ||
60 | } | ||
61 | } | ||
62 | } | ||
63 | + if (has_el2 && timeridx == GTIMER_VIRT) { | ||
64 | + if (FIELD_EX64(env->cp15.cnthctl_el2, CNTHCTL, EL1TVT)) { | ||
65 | + return CP_ACCESS_TRAP_EL2; | ||
66 | + } | ||
67 | + } | ||
68 | break; | ||
69 | } | ||
70 | return CP_ACCESS_OK; | ||
71 | @@ -XXX,XX +XXX,XX @@ static void gt_cnthctl_write(CPUARMState *env, const ARMCPRegInfo *ri, | ||
72 | if (cpu_isar_feature(aa64_rme, cpu)) { | ||
73 | valid_mask |= R_CNTHCTL_CNTVMASK_MASK | R_CNTHCTL_CNTPMASK_MASK; | ||
74 | } | ||
75 | + if (cpu_isar_feature(aa64_ecv_traps, cpu)) { | ||
76 | + valid_mask |= | ||
77 | + R_CNTHCTL_EL1TVT_MASK | | ||
78 | + R_CNTHCTL_EL1TVCT_MASK | | ||
79 | + R_CNTHCTL_EL1NVPCT_MASK | | ||
80 | + R_CNTHCTL_EL1NVVCT_MASK | | ||
81 | + R_CNTHCTL_EVNTIS_MASK; | ||
82 | + } | ||
83 | |||
84 | /* Clear RES0 bits */ | ||
85 | value &= valid_mask; | ||
86 | @@ -XXX,XX +XXX,XX @@ static CPAccessResult e2h_access(CPUARMState *env, const ARMCPRegInfo *ri, | ||
87 | { | ||
88 | if (arm_current_el(env) == 1) { | ||
89 | /* This must be a FEAT_NV access */ | ||
90 | - /* TODO: FEAT_ECV will need to check CNTHCTL_EL2 here */ | ||
91 | return CP_ACCESS_OK; | ||
92 | } | ||
93 | if (!(arm_hcr_el2_eff(env) & HCR_E2H)) { | ||
94 | @@ -XXX,XX +XXX,XX @@ static CPAccessResult e2h_access(CPUARMState *env, const ARMCPRegInfo *ri, | ||
95 | return CP_ACCESS_OK; | ||
37 | } | 96 | } |
38 | 97 | ||
39 | +#ifdef TARGET_AARCH64 | 98 | +static CPAccessResult access_el1nvpct(CPUARMState *env, const ARMCPRegInfo *ri, |
40 | +static uint64_t tlbi_aa64_range_get_length(CPUARMState *env, | 99 | + bool isread) |
41 | + uint64_t value) | ||
42 | +{ | 100 | +{ |
43 | + unsigned int page_shift; | 101 | + if (arm_current_el(env) == 1) { |
44 | + unsigned int page_size_granule; | 102 | + /* This must be a FEAT_NV access with NVx == 101 */ |
45 | + uint64_t num; | 103 | + if (FIELD_EX64(env->cp15.cnthctl_el2, CNTHCTL, EL1NVPCT)) { |
46 | + uint64_t scale; | 104 | + return CP_ACCESS_TRAP_EL2; |
47 | + uint64_t exponent; | 105 | + } |
48 | + uint64_t length; | ||
49 | + | ||
50 | + num = extract64(value, 39, 4); | ||
51 | + scale = extract64(value, 44, 2); | ||
52 | + page_size_granule = extract64(value, 46, 2); | ||
53 | + | ||
54 | + page_shift = page_size_granule * 2 + 12; | ||
55 | + | ||
56 | + if (page_size_granule == 0) { | ||
57 | + qemu_log_mask(LOG_GUEST_ERROR, "Invalid page size granule %d\n", | ||
58 | + page_size_granule); | ||
59 | + return 0; | ||
60 | + } | 106 | + } |
61 | + | 107 | + return e2h_access(env, ri, isread); |
62 | + exponent = (5 * scale) + 1; | ||
63 | + length = (num + 1) << (exponent + page_shift); | ||
64 | + | ||
65 | + return length; | ||
66 | +} | 108 | +} |
67 | + | 109 | + |
68 | +static uint64_t tlbi_aa64_range_get_base(CPUARMState *env, uint64_t value, | 110 | +static CPAccessResult access_el1nvvct(CPUARMState *env, const ARMCPRegInfo *ri, |
69 | + bool two_ranges) | 111 | + bool isread) |
70 | +{ | 112 | +{ |
71 | + /* TODO: ARMv8.7 FEAT_LPA2 */ | 113 | + if (arm_current_el(env) == 1) { |
72 | + uint64_t pageaddr; | 114 | + /* This must be a FEAT_NV access with NVx == 101 */ |
73 | + | 115 | + if (FIELD_EX64(env->cp15.cnthctl_el2, CNTHCTL, EL1NVVCT)) { |
74 | + if (two_ranges) { | 116 | + return CP_ACCESS_TRAP_EL2; |
75 | + pageaddr = sextract64(value, 0, 37) << TARGET_PAGE_BITS; | 117 | + } |
76 | + } else { | ||
77 | + pageaddr = extract64(value, 0, 37) << TARGET_PAGE_BITS; | ||
78 | + } | 118 | + } |
79 | + | 119 | + return e2h_access(env, ri, isread); |
80 | + return pageaddr; | ||
81 | +} | 120 | +} |
82 | + | 121 | + |
83 | +static void do_rvae_write(CPUARMState *env, uint64_t value, | 122 | /* Test if system register redirection is to occur in the current state. */ |
84 | + int idxmap, bool synced) | 123 | static bool redirect_for_e2h(CPUARMState *env) |
85 | +{ | ||
86 | + ARMMMUIdx one_idx = ARM_MMU_IDX_A | ctz32(idxmap); | ||
87 | + bool two_ranges = regime_has_2_ranges(one_idx); | ||
88 | + uint64_t baseaddr, length; | ||
89 | + int bits; | ||
90 | + | ||
91 | + baseaddr = tlbi_aa64_range_get_base(env, value, two_ranges); | ||
92 | + length = tlbi_aa64_range_get_length(env, value); | ||
93 | + bits = tlbbits_for_regime(env, one_idx, baseaddr); | ||
94 | + | ||
95 | + if (synced) { | ||
96 | + tlb_flush_range_by_mmuidx_all_cpus_synced(env_cpu(env), | ||
97 | + baseaddr, | ||
98 | + length, | ||
99 | + idxmap, | ||
100 | + bits); | ||
101 | + } else { | ||
102 | + tlb_flush_range_by_mmuidx(env_cpu(env), baseaddr, | ||
103 | + length, idxmap, bits); | ||
104 | + } | ||
105 | +} | ||
106 | + | ||
107 | +static void tlbi_aa64_rvae1_write(CPUARMState *env, | ||
108 | + const ARMCPRegInfo *ri, | ||
109 | + uint64_t value) | ||
110 | +{ | ||
111 | + /* | ||
112 | + * Invalidate by VA range, EL1&0. | ||
113 | + * Currently handles all of RVAE1, RVAAE1, RVAALE1 and RVALE1, | ||
114 | + * since we don't support flush-for-specific-ASID-only or | ||
115 | + * flush-last-level-only. | ||
116 | + */ | ||
117 | + | ||
118 | + do_rvae_write(env, value, vae1_tlbmask(env), | ||
119 | + tlb_force_broadcast(env)); | ||
120 | +} | ||
121 | + | ||
122 | +static void tlbi_aa64_rvae1is_write(CPUARMState *env, | ||
123 | + const ARMCPRegInfo *ri, | ||
124 | + uint64_t value) | ||
125 | +{ | ||
126 | + /* | ||
127 | + * Invalidate by VA range, Inner/Outer Shareable EL1&0. | ||
128 | + * Currently handles all of RVAE1IS, RVAE1OS, RVAAE1IS, RVAAE1OS, | ||
129 | + * RVAALE1IS, RVAALE1OS, RVALE1IS and RVALE1OS, since we don't support | ||
130 | + * flush-for-specific-ASID-only, flush-last-level-only or inner/outer | ||
131 | + * shareable specific flushes. | ||
132 | + */ | ||
133 | + | ||
134 | + do_rvae_write(env, value, vae1_tlbmask(env), true); | ||
135 | +} | ||
136 | + | ||
137 | +static int vae2_tlbmask(CPUARMState *env) | ||
138 | +{ | ||
139 | + return (arm_is_secure_below_el3(env) | ||
140 | + ? ARMMMUIdxBit_SE2 : ARMMMUIdxBit_E2); | ||
141 | +} | ||
142 | + | ||
143 | +static void tlbi_aa64_rvae2_write(CPUARMState *env, | ||
144 | + const ARMCPRegInfo *ri, | ||
145 | + uint64_t value) | ||
146 | +{ | ||
147 | + /* | ||
148 | + * Invalidate by VA range, EL2. | ||
149 | + * Currently handles all of RVAE2 and RVALE2, | ||
150 | + * since we don't support flush-for-specific-ASID-only or | ||
151 | + * flush-last-level-only. | ||
152 | + */ | ||
153 | + | ||
154 | + do_rvae_write(env, value, vae2_tlbmask(env), | ||
155 | + tlb_force_broadcast(env)); | ||
156 | + | ||
157 | + | ||
158 | +} | ||
159 | + | ||
160 | +static void tlbi_aa64_rvae2is_write(CPUARMState *env, | ||
161 | + const ARMCPRegInfo *ri, | ||
162 | + uint64_t value) | ||
163 | +{ | ||
164 | + /* | ||
165 | + * Invalidate by VA range, Inner/Outer Shareable, EL2. | ||
166 | + * Currently handles all of RVAE2IS, RVAE2OS, RVALE2IS and RVALE2OS, | ||
167 | + * since we don't support flush-for-specific-ASID-only, | ||
168 | + * flush-last-level-only or inner/outer shareable specific flushes. | ||
169 | + */ | ||
170 | + | ||
171 | + do_rvae_write(env, value, vae2_tlbmask(env), true); | ||
172 | + | ||
173 | +} | ||
174 | + | ||
175 | +static void tlbi_aa64_rvae3_write(CPUARMState *env, | ||
176 | + const ARMCPRegInfo *ri, | ||
177 | + uint64_t value) | ||
178 | +{ | ||
179 | + /* | ||
180 | + * Invalidate by VA range, EL3. | ||
181 | + * Currently handles all of RVAE3 and RVALE3, | ||
182 | + * since we don't support flush-for-specific-ASID-only or | ||
183 | + * flush-last-level-only. | ||
184 | + */ | ||
185 | + | ||
186 | + do_rvae_write(env, value, ARMMMUIdxBit_SE3, | ||
187 | + tlb_force_broadcast(env)); | ||
188 | +} | ||
189 | + | ||
190 | +static void tlbi_aa64_rvae3is_write(CPUARMState *env, | ||
191 | + const ARMCPRegInfo *ri, | ||
192 | + uint64_t value) | ||
193 | +{ | ||
194 | + /* | ||
195 | + * Invalidate by VA range, EL3, Inner/Outer Shareable. | ||
196 | + * Currently handles all of RVAE3IS, RVAE3OS, RVALE3IS and RVALE3OS, | ||
197 | + * since we don't support flush-for-specific-ASID-only, | ||
198 | + * flush-last-level-only or inner/outer specific flushes. | ||
199 | + */ | ||
200 | + | ||
201 | + do_rvae_write(env, value, ARMMMUIdxBit_SE3, true); | ||
202 | +} | ||
203 | +#endif | ||
204 | + | ||
205 | static CPAccessResult aa64_zva_access(CPUARMState *env, const ARMCPRegInfo *ri, | ||
206 | bool isread) | ||
207 | { | 124 | { |
208 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pauth_reginfo[] = { | 125 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vhe_reginfo[] = { |
209 | REGINFO_SENTINEL | 126 | { .name = "CNTP_CTL_EL02", .state = ARM_CP_STATE_AA64, |
127 | .opc0 = 3, .opc1 = 5, .crn = 14, .crm = 2, .opc2 = 1, | ||
128 | .type = ARM_CP_IO | ARM_CP_ALIAS, | ||
129 | - .access = PL2_RW, .accessfn = e2h_access, | ||
130 | + .access = PL2_RW, .accessfn = access_el1nvpct, | ||
131 | .nv2_redirect_offset = 0x180 | NV2_REDIR_NO_NV1, | ||
132 | .fieldoffset = offsetof(CPUARMState, cp15.c14_timer[GTIMER_PHYS].ctl), | ||
133 | .writefn = gt_phys_ctl_write, .raw_writefn = raw_write }, | ||
134 | { .name = "CNTV_CTL_EL02", .state = ARM_CP_STATE_AA64, | ||
135 | .opc0 = 3, .opc1 = 5, .crn = 14, .crm = 3, .opc2 = 1, | ||
136 | .type = ARM_CP_IO | ARM_CP_ALIAS, | ||
137 | - .access = PL2_RW, .accessfn = e2h_access, | ||
138 | + .access = PL2_RW, .accessfn = access_el1nvvct, | ||
139 | .nv2_redirect_offset = 0x170 | NV2_REDIR_NO_NV1, | ||
140 | .fieldoffset = offsetof(CPUARMState, cp15.c14_timer[GTIMER_VIRT].ctl), | ||
141 | .writefn = gt_virt_ctl_write, .raw_writefn = raw_write }, | ||
142 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo vhe_reginfo[] = { | ||
143 | .type = ARM_CP_IO | ARM_CP_ALIAS, | ||
144 | .fieldoffset = offsetof(CPUARMState, cp15.c14_timer[GTIMER_PHYS].cval), | ||
145 | .nv2_redirect_offset = 0x178 | NV2_REDIR_NO_NV1, | ||
146 | - .access = PL2_RW, .accessfn = e2h_access, | ||
147 | + .access = PL2_RW, .accessfn = access_el1nvpct, | ||
148 | .writefn = gt_phys_cval_write, .raw_writefn = raw_write }, | ||
149 | { .name = "CNTV_CVAL_EL02", .state = ARM_CP_STATE_AA64, | ||
150 | .opc0 = 3, .opc1 = 5, .crn = 14, .crm = 3, .opc2 = 2, | ||
151 | .type = ARM_CP_IO | ARM_CP_ALIAS, | ||
152 | .nv2_redirect_offset = 0x168 | NV2_REDIR_NO_NV1, | ||
153 | .fieldoffset = offsetof(CPUARMState, cp15.c14_timer[GTIMER_VIRT].cval), | ||
154 | - .access = PL2_RW, .accessfn = e2h_access, | ||
155 | + .access = PL2_RW, .accessfn = access_el1nvvct, | ||
156 | .writefn = gt_virt_cval_write, .raw_writefn = raw_write }, | ||
157 | #endif | ||
210 | }; | 158 | }; |
211 | |||
212 | +static const ARMCPRegInfo tlbirange_reginfo[] = { | ||
213 | + { .name = "TLBI_RVAE1IS", .state = ARM_CP_STATE_AA64, | ||
214 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 2, .opc2 = 1, | ||
215 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
216 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
217 | + { .name = "TLBI_RVAAE1IS", .state = ARM_CP_STATE_AA64, | ||
218 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 2, .opc2 = 3, | ||
219 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
220 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
221 | + { .name = "TLBI_RVALE1IS", .state = ARM_CP_STATE_AA64, | ||
222 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 2, .opc2 = 5, | ||
223 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
224 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
225 | + { .name = "TLBI_RVAALE1IS", .state = ARM_CP_STATE_AA64, | ||
226 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 2, .opc2 = 7, | ||
227 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
228 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
229 | + { .name = "TLBI_RVAE1OS", .state = ARM_CP_STATE_AA64, | ||
230 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 5, .opc2 = 1, | ||
231 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
232 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
233 | + { .name = "TLBI_RVAAE1OS", .state = ARM_CP_STATE_AA64, | ||
234 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 5, .opc2 = 3, | ||
235 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
236 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
237 | + { .name = "TLBI_RVALE1OS", .state = ARM_CP_STATE_AA64, | ||
238 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 5, .opc2 = 5, | ||
239 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
240 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
241 | + { .name = "TLBI_RVAALE1OS", .state = ARM_CP_STATE_AA64, | ||
242 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 5, .opc2 = 7, | ||
243 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
244 | + .writefn = tlbi_aa64_rvae1is_write }, | ||
245 | + { .name = "TLBI_RVAE1", .state = ARM_CP_STATE_AA64, | ||
246 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 6, .opc2 = 1, | ||
247 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
248 | + .writefn = tlbi_aa64_rvae1_write }, | ||
249 | + { .name = "TLBI_RVAAE1", .state = ARM_CP_STATE_AA64, | ||
250 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 6, .opc2 = 3, | ||
251 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
252 | + .writefn = tlbi_aa64_rvae1_write }, | ||
253 | + { .name = "TLBI_RVALE1", .state = ARM_CP_STATE_AA64, | ||
254 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 6, .opc2 = 5, | ||
255 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
256 | + .writefn = tlbi_aa64_rvae1_write }, | ||
257 | + { .name = "TLBI_RVAALE1", .state = ARM_CP_STATE_AA64, | ||
258 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 6, .opc2 = 7, | ||
259 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | ||
260 | + .writefn = tlbi_aa64_rvae1_write }, | ||
261 | + { .name = "TLBI_RIPAS2E1IS", .state = ARM_CP_STATE_AA64, | ||
262 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 2, | ||
263 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
264 | + { .name = "TLBI_RIPAS2LE1IS", .state = ARM_CP_STATE_AA64, | ||
265 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 0, .opc2 = 6, | ||
266 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
267 | + { .name = "TLBI_RVAE2IS", .state = ARM_CP_STATE_AA64, | ||
268 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 2, .opc2 = 1, | ||
269 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
270 | + .writefn = tlbi_aa64_rvae2is_write }, | ||
271 | + { .name = "TLBI_RVALE2IS", .state = ARM_CP_STATE_AA64, | ||
272 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 2, .opc2 = 5, | ||
273 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
274 | + .writefn = tlbi_aa64_rvae2is_write }, | ||
275 | + { .name = "TLBI_RIPAS2E1", .state = ARM_CP_STATE_AA64, | ||
276 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 2, | ||
277 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
278 | + { .name = "TLBI_RIPAS2LE1", .state = ARM_CP_STATE_AA64, | ||
279 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 6, | ||
280 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
281 | + { .name = "TLBI_RVAE2OS", .state = ARM_CP_STATE_AA64, | ||
282 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 5, .opc2 = 1, | ||
283 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
284 | + .writefn = tlbi_aa64_rvae2is_write }, | ||
285 | + { .name = "TLBI_RVALE2OS", .state = ARM_CP_STATE_AA64, | ||
286 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 5, .opc2 = 5, | ||
287 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
288 | + .writefn = tlbi_aa64_rvae2is_write }, | ||
289 | + { .name = "TLBI_RVAE2", .state = ARM_CP_STATE_AA64, | ||
290 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 6, .opc2 = 1, | ||
291 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
292 | + .writefn = tlbi_aa64_rvae2_write }, | ||
293 | + { .name = "TLBI_RVALE2", .state = ARM_CP_STATE_AA64, | ||
294 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 6, .opc2 = 5, | ||
295 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | ||
296 | + .writefn = tlbi_aa64_rvae2_write }, | ||
297 | + { .name = "TLBI_RVAE3IS", .state = ARM_CP_STATE_AA64, | ||
298 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 2, .opc2 = 1, | ||
299 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
300 | + .writefn = tlbi_aa64_rvae3is_write }, | ||
301 | + { .name = "TLBI_RVALE3IS", .state = ARM_CP_STATE_AA64, | ||
302 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 2, .opc2 = 5, | ||
303 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
304 | + .writefn = tlbi_aa64_rvae3is_write }, | ||
305 | + { .name = "TLBI_RVAE3OS", .state = ARM_CP_STATE_AA64, | ||
306 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 5, .opc2 = 1, | ||
307 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
308 | + .writefn = tlbi_aa64_rvae3is_write }, | ||
309 | + { .name = "TLBI_RVALE3OS", .state = ARM_CP_STATE_AA64, | ||
310 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 5, .opc2 = 5, | ||
311 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
312 | + .writefn = tlbi_aa64_rvae3is_write }, | ||
313 | + { .name = "TLBI_RVAE3", .state = ARM_CP_STATE_AA64, | ||
314 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 6, .opc2 = 1, | ||
315 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
316 | + .writefn = tlbi_aa64_rvae3_write }, | ||
317 | + { .name = "TLBI_RVALE3", .state = ARM_CP_STATE_AA64, | ||
318 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 6, .opc2 = 5, | ||
319 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
320 | + .writefn = tlbi_aa64_rvae3_write }, | ||
321 | + REGINFO_SENTINEL | ||
322 | +}; | ||
323 | + | ||
324 | static uint64_t rndr_readfn(CPUARMState *env, const ARMCPRegInfo *ri) | ||
325 | { | ||
326 | Error *err = NULL; | ||
327 | @@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu) | ||
328 | if (cpu_isar_feature(aa64_rndr, cpu)) { | ||
329 | define_arm_cp_regs(cpu, rndr_reginfo); | ||
330 | } | ||
331 | + if (cpu_isar_feature(aa64_tlbirange, cpu)) { | ||
332 | + define_arm_cp_regs(cpu, tlbirange_reginfo); | ||
333 | + } | ||
334 | #ifndef CONFIG_USER_ONLY | ||
335 | /* Data Cache clean instructions up to PoP */ | ||
336 | if (cpu_isar_feature(aa64_dcpop, cpu)) { | ||
337 | -- | 159 | -- |
338 | 2.20.1 | 160 | 2.34.1 |
339 | |||
340 | diff view generated by jsdifflib |
1 | From: Rebecca Cran <rebecca@nuviainc.com> | 1 | For FEAT_ECV, new registers CNTPCTSS_EL0 and CNTVCTSS_EL0 are |
---|---|---|---|
2 | defined, which are "self-synchronized" views of the physical and | ||
3 | virtual counts as seen in the CNTPCT_EL0 and CNTVCT_EL0 registers | ||
4 | (meaning that no barriers are needed around accesses to them to | ||
5 | ensure that reads of them do not occur speculatively and out-of-order | ||
6 | with other instructions). | ||
2 | 7 | ||
3 | ARMv8.4 adds the mandatory FEAT_TLBIOS. It provides TLBI | 8 | For QEMU, all our system registers are self-synchronized, so we can |
4 | maintenance instructions that extend to the Outer Shareable domain. | 9 | simply copy the existing implementation of CNTPCT_EL0 and CNTVCT_EL0 |
10 | to the new register encodings. | ||
5 | 11 | ||
6 | Signed-off-by: Rebecca Cran <rebecca@nuviainc.com> | 12 | This means we now implement all the functionality required for |
13 | ID_AA64MMFR0_EL1.ECV == 0b0001. | ||
14 | |||
15 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | 16 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> |
8 | Message-id: 20210512182337.18563-3-rebecca@nuviainc.com | 17 | Message-id: 20240301183219.2424889-7-peter.maydell@linaro.org |
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | 18 | --- |
11 | target/arm/cpu.h | 5 +++++ | ||
12 | target/arm/helper.c | 43 +++++++++++++++++++++++++++++++++++++++++++ | 19 | target/arm/helper.c | 43 +++++++++++++++++++++++++++++++++++++++++++ |
13 | 2 files changed, 48 insertions(+) | 20 | 1 file changed, 43 insertions(+) |
14 | 21 | ||
15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/target/arm/cpu.h | ||
18 | +++ b/target/arm/cpu.h | ||
19 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_tlbirange(const ARMISARegisters *id) | ||
20 | return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, TLB) == 2; | ||
21 | } | ||
22 | |||
23 | +static inline bool isar_feature_aa64_tlbios(const ARMISARegisters *id) | ||
24 | +{ | ||
25 | + return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, TLB) != 0; | ||
26 | +} | ||
27 | + | ||
28 | static inline bool isar_feature_aa64_sb(const ARMISARegisters *id) | ||
29 | { | ||
30 | return FIELD_EX64(id->id_aa64isar1, ID_AA64ISAR1, SB) != 0; | ||
31 | diff --git a/target/arm/helper.c b/target/arm/helper.c | 22 | diff --git a/target/arm/helper.c b/target/arm/helper.c |
32 | index XXXXXXX..XXXXXXX 100644 | 23 | index XXXXXXX..XXXXXXX 100644 |
33 | --- a/target/arm/helper.c | 24 | --- a/target/arm/helper.c |
34 | +++ b/target/arm/helper.c | 25 | +++ b/target/arm/helper.c |
35 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo tlbirange_reginfo[] = { | 26 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo generic_timer_cp_reginfo[] = { |
36 | REGINFO_SENTINEL | 27 | }, |
37 | }; | 28 | }; |
38 | 29 | ||
39 | +static const ARMCPRegInfo tlbios_reginfo[] = { | 30 | +/* |
40 | + { .name = "TLBI_VMALLE1OS", .state = ARM_CP_STATE_AA64, | 31 | + * FEAT_ECV adds extra views of CNTVCT_EL0 and CNTPCT_EL0 which |
41 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 1, .opc2 = 0, | 32 | + * are "self-synchronizing". For QEMU all sysregs are self-synchronizing, |
42 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | 33 | + * so our implementations here are identical to the normal registers. |
43 | + .writefn = tlbi_aa64_vmalle1is_write }, | 34 | + */ |
44 | + { .name = "TLBI_ASIDE1OS", .state = ARM_CP_STATE_AA64, | 35 | +static const ARMCPRegInfo gen_timer_ecv_cp_reginfo[] = { |
45 | + .opc0 = 1, .opc1 = 0, .crn = 8, .crm = 1, .opc2 = 2, | 36 | + { .name = "CNTVCTSS", .cp = 15, .crm = 14, .opc1 = 9, |
46 | + .access = PL1_W, .type = ARM_CP_NO_RAW, | 37 | + .access = PL0_R, .type = ARM_CP_64BIT | ARM_CP_NO_RAW | ARM_CP_IO, |
47 | + .writefn = tlbi_aa64_vmalle1is_write }, | 38 | + .accessfn = gt_vct_access, |
48 | + { .name = "TLBI_ALLE2OS", .state = ARM_CP_STATE_AA64, | 39 | + .readfn = gt_virt_cnt_read, .resetfn = arm_cp_reset_ignore, |
49 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 1, .opc2 = 0, | 40 | + }, |
50 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | 41 | + { .name = "CNTVCTSS_EL0", .state = ARM_CP_STATE_AA64, |
51 | + .writefn = tlbi_aa64_alle2is_write }, | 42 | + .opc0 = 3, .opc1 = 3, .crn = 14, .crm = 0, .opc2 = 6, |
52 | + { .name = "TLBI_ALLE1OS", .state = ARM_CP_STATE_AA64, | 43 | + .access = PL0_R, .type = ARM_CP_NO_RAW | ARM_CP_IO, |
53 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 1, .opc2 = 4, | 44 | + .accessfn = gt_vct_access, .readfn = gt_virt_cnt_read, |
54 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | 45 | + }, |
55 | + .writefn = tlbi_aa64_alle1is_write }, | 46 | + { .name = "CNTPCTSS", .cp = 15, .crm = 14, .opc1 = 8, |
56 | + { .name = "TLBI_VMALLS12E1OS", .state = ARM_CP_STATE_AA64, | 47 | + .access = PL0_R, .type = ARM_CP_64BIT | ARM_CP_NO_RAW | ARM_CP_IO, |
57 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 1, .opc2 = 6, | 48 | + .accessfn = gt_pct_access, |
58 | + .access = PL2_W, .type = ARM_CP_NO_RAW, | 49 | + .readfn = gt_cnt_read, .resetfn = arm_cp_reset_ignore, |
59 | + .writefn = tlbi_aa64_alle1is_write }, | 50 | + }, |
60 | + { .name = "TLBI_IPAS2E1OS", .state = ARM_CP_STATE_AA64, | 51 | + { .name = "CNTPCTSS_EL0", .state = ARM_CP_STATE_AA64, |
61 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 0, | 52 | + .opc0 = 3, .opc1 = 3, .crn = 14, .crm = 0, .opc2 = 5, |
62 | + .access = PL2_W, .type = ARM_CP_NOP }, | 53 | + .access = PL0_R, .type = ARM_CP_NO_RAW | ARM_CP_IO, |
63 | + { .name = "TLBI_RIPAS2E1OS", .state = ARM_CP_STATE_AA64, | 54 | + .accessfn = gt_pct_access, .readfn = gt_cnt_read, |
64 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 3, | 55 | + }, |
65 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
66 | + { .name = "TLBI_IPAS2LE1OS", .state = ARM_CP_STATE_AA64, | ||
67 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 4, | ||
68 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
69 | + { .name = "TLBI_RIPAS2LE1OS", .state = ARM_CP_STATE_AA64, | ||
70 | + .opc0 = 1, .opc1 = 4, .crn = 8, .crm = 4, .opc2 = 7, | ||
71 | + .access = PL2_W, .type = ARM_CP_NOP }, | ||
72 | + { .name = "TLBI_ALLE3OS", .state = ARM_CP_STATE_AA64, | ||
73 | + .opc0 = 1, .opc1 = 6, .crn = 8, .crm = 1, .opc2 = 0, | ||
74 | + .access = PL3_W, .type = ARM_CP_NO_RAW, | ||
75 | + .writefn = tlbi_aa64_alle3is_write }, | ||
76 | + REGINFO_SENTINEL | ||
77 | +}; | 56 | +}; |
78 | + | 57 | + |
79 | static uint64_t rndr_readfn(CPUARMState *env, const ARMCPRegInfo *ri) | 58 | #else |
80 | { | 59 | |
81 | Error *err = NULL; | 60 | /* |
61 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo generic_timer_cp_reginfo[] = { | ||
62 | }, | ||
63 | }; | ||
64 | |||
65 | +/* | ||
66 | + * CNTVCTSS_EL0 has the same trap conditions as CNTVCT_EL0, so it also | ||
67 | + * is exposed to userspace by Linux. | ||
68 | + */ | ||
69 | +static const ARMCPRegInfo gen_timer_ecv_cp_reginfo[] = { | ||
70 | + { .name = "CNTVCTSS_EL0", .state = ARM_CP_STATE_AA64, | ||
71 | + .opc0 = 3, .opc1 = 3, .crn = 14, .crm = 0, .opc2 = 6, | ||
72 | + .access = PL0_R, .type = ARM_CP_NO_RAW | ARM_CP_IO, | ||
73 | + .readfn = gt_virt_cnt_read, | ||
74 | + }, | ||
75 | +}; | ||
76 | + | ||
77 | #endif | ||
78 | |||
79 | static void par_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) | ||
82 | @@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu) | 80 | @@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu) |
83 | if (cpu_isar_feature(aa64_tlbirange, cpu)) { | 81 | if (arm_feature(env, ARM_FEATURE_GENERIC_TIMER)) { |
84 | define_arm_cp_regs(cpu, tlbirange_reginfo); | 82 | define_arm_cp_regs(cpu, generic_timer_cp_reginfo); |
85 | } | 83 | } |
86 | + if (cpu_isar_feature(aa64_tlbios, cpu)) { | 84 | + if (cpu_isar_feature(aa64_ecv_traps, cpu)) { |
87 | + define_arm_cp_regs(cpu, tlbios_reginfo); | 85 | + define_arm_cp_regs(cpu, gen_timer_ecv_cp_reginfo); |
88 | + } | 86 | + } |
89 | #ifndef CONFIG_USER_ONLY | 87 | if (arm_feature(env, ARM_FEATURE_VAPA)) { |
90 | /* Data Cache clean instructions up to PoP */ | 88 | ARMCPRegInfo vapa_cp_reginfo[] = { |
91 | if (cpu_isar_feature(aa64_dcpop, cpu)) { | 89 | { .name = "PAR", .cp = 15, .crn = 7, .crm = 4, .opc1 = 0, .opc2 = 0, |
92 | -- | 90 | -- |
93 | 2.20.1 | 91 | 2.34.1 |
94 | |||
95 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | When ID_AA64MMFR0_EL1.ECV is 0b0010, a new register CNTPOFF_EL2 is |
---|---|---|---|
2 | implemented. This is similar to the existing CNTVOFF_EL2, except | ||
3 | that it controls a hypervisor-adjustable offset made to the physical | ||
4 | counter and timer. | ||
2 | 5 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 6 | Implement the handling for this register, which includes control/trap |
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 7 | bits in SCR_EL3 and CNTHCTL_EL2. |
5 | Message-id: 20210525010358.152808-70-richard.henderson@linaro.org | 8 | |
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
10 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
11 | Message-id: 20240301183219.2424889-8-peter.maydell@linaro.org | ||
7 | --- | 12 | --- |
8 | target/arm/cpu.h | 5 +++++ | 13 | target/arm/cpu-features.h | 5 +++ |
9 | target/arm/sve.decode | 7 +++++++ | 14 | target/arm/cpu.h | 1 + |
10 | target/arm/translate-sve.c | 38 ++++++++++++++++++++++++++++++++++++++ | 15 | target/arm/helper.c | 68 +++++++++++++++++++++++++++++++++++++-- |
11 | 3 files changed, 50 insertions(+) | 16 | target/arm/trace-events | 1 + |
17 | 4 files changed, 73 insertions(+), 2 deletions(-) | ||
12 | 18 | ||
19 | diff --git a/target/arm/cpu-features.h b/target/arm/cpu-features.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/target/arm/cpu-features.h | ||
22 | +++ b/target/arm/cpu-features.h | ||
23 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_ecv_traps(const ARMISARegisters *id) | ||
24 | return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR0, ECV) > 0; | ||
25 | } | ||
26 | |||
27 | +static inline bool isar_feature_aa64_ecv(const ARMISARegisters *id) | ||
28 | +{ | ||
29 | + return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR0, ECV) > 1; | ||
30 | +} | ||
31 | + | ||
32 | static inline bool isar_feature_aa64_vh(const ARMISARegisters *id) | ||
33 | { | ||
34 | return FIELD_EX64(id->id_aa64mmfr1, ID_AA64MMFR1, VH) != 0; | ||
13 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | 35 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h |
14 | index XXXXXXX..XXXXXXX 100644 | 36 | index XXXXXXX..XXXXXXX 100644 |
15 | --- a/target/arm/cpu.h | 37 | --- a/target/arm/cpu.h |
16 | +++ b/target/arm/cpu.h | 38 | +++ b/target/arm/cpu.h |
17 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2_bitperm(const ARMISARegisters *id) | 39 | @@ -XXX,XX +XXX,XX @@ typedef struct CPUArchState { |
18 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, BITPERM) != 0; | 40 | uint64_t c14_cntkctl; /* Timer Control register */ |
41 | uint64_t cnthctl_el2; /* Counter/Timer Hyp Control register */ | ||
42 | uint64_t cntvoff_el2; /* Counter Virtual Offset register */ | ||
43 | + uint64_t cntpoff_el2; /* Counter Physical Offset register */ | ||
44 | ARMGenericTimer c14_timer[NUM_GTIMERS]; | ||
45 | uint32_t c15_cpar; /* XScale Coprocessor Access Register */ | ||
46 | uint32_t c15_ticonfig; /* TI925T configuration byte. */ | ||
47 | diff --git a/target/arm/helper.c b/target/arm/helper.c | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/target/arm/helper.c | ||
50 | +++ b/target/arm/helper.c | ||
51 | @@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) | ||
52 | if (cpu_isar_feature(aa64_rme, cpu)) { | ||
53 | valid_mask |= SCR_NSE | SCR_GPF; | ||
54 | } | ||
55 | + if (cpu_isar_feature(aa64_ecv, cpu)) { | ||
56 | + valid_mask |= SCR_ECVEN; | ||
57 | + } | ||
58 | } else { | ||
59 | valid_mask &= ~(SCR_RW | SCR_ST); | ||
60 | if (cpu_isar_feature(aa32_ras, cpu)) { | ||
61 | @@ -XXX,XX +XXX,XX @@ void gt_rme_post_el_change(ARMCPU *cpu, void *ignored) | ||
62 | gt_update_irq(cpu, GTIMER_PHYS); | ||
19 | } | 63 | } |
20 | 64 | ||
21 | +static inline bool isar_feature_aa64_sve2_sm4(const ARMISARegisters *id) | 65 | +static uint64_t gt_phys_raw_cnt_offset(CPUARMState *env) |
22 | +{ | 66 | +{ |
23 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SM4) != 0; | 67 | + if ((env->cp15.scr_el3 & SCR_ECVEN) && |
68 | + FIELD_EX64(env->cp15.cnthctl_el2, CNTHCTL, ECV) && | ||
69 | + arm_is_el2_enabled(env) && | ||
70 | + (arm_hcr_el2_eff(env) & (HCR_E2H | HCR_TGE)) != (HCR_E2H | HCR_TGE)) { | ||
71 | + return env->cp15.cntpoff_el2; | ||
72 | + } | ||
73 | + return 0; | ||
24 | +} | 74 | +} |
25 | + | 75 | + |
26 | static inline bool isar_feature_aa64_sve_i8mm(const ARMISARegisters *id) | 76 | +static uint64_t gt_phys_cnt_offset(CPUARMState *env) |
27 | { | ||
28 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, I8MM) != 0; | ||
29 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
30 | index XXXXXXX..XXXXXXX 100644 | ||
31 | --- a/target/arm/sve.decode | ||
32 | +++ b/target/arm/sve.decode | ||
33 | @@ -XXX,XX +XXX,XX @@ | ||
34 | @pd_pn_pm ........ esz:2 .. rm:4 ....... rn:4 . rd:4 &rrr_esz | ||
35 | @rdn_rm ........ esz:2 ...... ...... rm:5 rd:5 \ | ||
36 | &rrr_esz rn=%reg_movprfx | ||
37 | +@rdn_rm_e0 ........ .. ...... ...... rm:5 rd:5 \ | ||
38 | + &rrr_esz rn=%reg_movprfx esz=0 | ||
39 | @rdn_sh_i8u ........ esz:2 ...... ...... ..... rd:5 \ | ||
40 | &rri_esz rn=%reg_movprfx imm=%sh8_i8u | ||
41 | @rdn_i8u ........ esz:2 ...... ... imm:8 rd:5 \ | ||
42 | @@ -XXX,XX +XXX,XX @@ STNT1_zprz 1110010 .. 10 ..... 001 ... ..... ..... \ | ||
43 | # SVE2 crypto unary operations | ||
44 | # AESMC and AESIMC | ||
45 | AESMC 01000101 00 10000011100 decrypt:1 00000 rd:5 | ||
46 | + | ||
47 | +# SVE2 crypto destructive binary operations | ||
48 | +AESE 01000101 00 10001 0 11100 0 ..... ..... @rdn_rm_e0 | ||
49 | +AESD 01000101 00 10001 0 11100 1 ..... ..... @rdn_rm_e0 | ||
50 | +SM4E 01000101 00 10001 1 11100 0 ..... ..... @rdn_rm_e0 | ||
51 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
52 | index XXXXXXX..XXXXXXX 100644 | ||
53 | --- a/target/arm/translate-sve.c | ||
54 | +++ b/target/arm/translate-sve.c | ||
55 | @@ -XXX,XX +XXX,XX @@ static bool trans_AESMC(DisasContext *s, arg_AESMC *a) | ||
56 | } | ||
57 | return true; | ||
58 | } | ||
59 | + | ||
60 | +static bool do_aese(DisasContext *s, arg_rrr_esz *a, bool decrypt) | ||
61 | +{ | 77 | +{ |
62 | + if (!dc_isar_feature(aa64_sve2_aes, s)) { | 78 | + if (arm_current_el(env) >= 2) { |
63 | + return false; | 79 | + return 0; |
64 | + } | 80 | + } |
65 | + if (sve_access_check(s)) { | 81 | + return gt_phys_raw_cnt_offset(env); |
66 | + gen_gvec_ool_zzz(s, gen_helper_crypto_aese, | ||
67 | + a->rd, a->rn, a->rm, decrypt); | ||
68 | + } | ||
69 | + return true; | ||
70 | +} | 82 | +} |
71 | + | 83 | + |
72 | +static bool trans_AESE(DisasContext *s, arg_rrr_esz *a) | 84 | static void gt_recalc_timer(ARMCPU *cpu, int timeridx) |
85 | { | ||
86 | ARMGenericTimer *gt = &cpu->env.cp15.c14_timer[timeridx]; | ||
87 | @@ -XXX,XX +XXX,XX @@ static void gt_recalc_timer(ARMCPU *cpu, int timeridx) | ||
88 | * reset timer to when ISTATUS next has to change | ||
89 | */ | ||
90 | uint64_t offset = timeridx == GTIMER_VIRT ? | ||
91 | - cpu->env.cp15.cntvoff_el2 : 0; | ||
92 | + cpu->env.cp15.cntvoff_el2 : gt_phys_raw_cnt_offset(&cpu->env); | ||
93 | uint64_t count = gt_get_countervalue(&cpu->env); | ||
94 | /* Note that this must be unsigned 64 bit arithmetic: */ | ||
95 | int istatus = count - offset >= gt->cval; | ||
96 | @@ -XXX,XX +XXX,XX @@ static void gt_timer_reset(CPUARMState *env, const ARMCPRegInfo *ri, | ||
97 | |||
98 | static uint64_t gt_cnt_read(CPUARMState *env, const ARMCPRegInfo *ri) | ||
99 | { | ||
100 | - return gt_get_countervalue(env); | ||
101 | + return gt_get_countervalue(env) - gt_phys_cnt_offset(env); | ||
102 | } | ||
103 | |||
104 | static uint64_t gt_virt_cnt_offset(CPUARMState *env) | ||
105 | @@ -XXX,XX +XXX,XX @@ static uint64_t gt_tval_read(CPUARMState *env, const ARMCPRegInfo *ri, | ||
106 | case GTIMER_HYPVIRT: | ||
107 | offset = gt_virt_cnt_offset(env); | ||
108 | break; | ||
109 | + case GTIMER_PHYS: | ||
110 | + offset = gt_phys_cnt_offset(env); | ||
111 | + break; | ||
112 | } | ||
113 | |||
114 | return (uint32_t)(env->cp15.c14_timer[timeridx].cval - | ||
115 | @@ -XXX,XX +XXX,XX @@ static void gt_tval_write(CPUARMState *env, const ARMCPRegInfo *ri, | ||
116 | case GTIMER_HYPVIRT: | ||
117 | offset = gt_virt_cnt_offset(env); | ||
118 | break; | ||
119 | + case GTIMER_PHYS: | ||
120 | + offset = gt_phys_cnt_offset(env); | ||
121 | + break; | ||
122 | } | ||
123 | |||
124 | trace_arm_gt_tval_write(timeridx, value); | ||
125 | @@ -XXX,XX +XXX,XX @@ static void gt_cnthctl_write(CPUARMState *env, const ARMCPRegInfo *ri, | ||
126 | R_CNTHCTL_EL1NVVCT_MASK | | ||
127 | R_CNTHCTL_EVNTIS_MASK; | ||
128 | } | ||
129 | + if (cpu_isar_feature(aa64_ecv, cpu)) { | ||
130 | + valid_mask |= R_CNTHCTL_ECV_MASK; | ||
131 | + } | ||
132 | |||
133 | /* Clear RES0 bits */ | ||
134 | value &= valid_mask; | ||
135 | @@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo gen_timer_ecv_cp_reginfo[] = { | ||
136 | }, | ||
137 | }; | ||
138 | |||
139 | +static CPAccessResult gt_cntpoff_access(CPUARMState *env, | ||
140 | + const ARMCPRegInfo *ri, | ||
141 | + bool isread) | ||
73 | +{ | 142 | +{ |
74 | + return do_aese(s, a, false); | 143 | + if (arm_current_el(env) == 2 && !(env->cp15.scr_el3 & SCR_ECVEN)) { |
144 | + return CP_ACCESS_TRAP_EL3; | ||
145 | + } | ||
146 | + return CP_ACCESS_OK; | ||
75 | +} | 147 | +} |
76 | + | 148 | + |
77 | +static bool trans_AESD(DisasContext *s, arg_rrr_esz *a) | 149 | +static void gt_cntpoff_write(CPUARMState *env, const ARMCPRegInfo *ri, |
150 | + uint64_t value) | ||
78 | +{ | 151 | +{ |
79 | + return do_aese(s, a, true); | 152 | + ARMCPU *cpu = env_archcpu(env); |
153 | + | ||
154 | + trace_arm_gt_cntpoff_write(value); | ||
155 | + raw_write(env, ri, value); | ||
156 | + gt_recalc_timer(cpu, GTIMER_PHYS); | ||
80 | +} | 157 | +} |
81 | + | 158 | + |
82 | +static bool do_sm4(DisasContext *s, arg_rrr_esz *a, gen_helper_gvec_3 *fn) | 159 | +static const ARMCPRegInfo gen_timer_cntpoff_reginfo = { |
83 | +{ | 160 | + .name = "CNTPOFF_EL2", .state = ARM_CP_STATE_AA64, |
84 | + if (!dc_isar_feature(aa64_sve2_sm4, s)) { | 161 | + .opc0 = 3, .opc1 = 4, .crn = 14, .crm = 0, .opc2 = 6, |
85 | + return false; | 162 | + .access = PL2_RW, .type = ARM_CP_IO, .resetvalue = 0, |
163 | + .accessfn = gt_cntpoff_access, .writefn = gt_cntpoff_write, | ||
164 | + .nv2_redirect_offset = 0x1a8, | ||
165 | + .fieldoffset = offsetof(CPUARMState, cp15.cntpoff_el2), | ||
166 | +}; | ||
167 | #else | ||
168 | |||
169 | /* | ||
170 | @@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu) | ||
171 | if (cpu_isar_feature(aa64_ecv_traps, cpu)) { | ||
172 | define_arm_cp_regs(cpu, gen_timer_ecv_cp_reginfo); | ||
173 | } | ||
174 | +#ifndef CONFIG_USER_ONLY | ||
175 | + if (cpu_isar_feature(aa64_ecv, cpu)) { | ||
176 | + define_one_arm_cp_reg(cpu, &gen_timer_cntpoff_reginfo); | ||
86 | + } | 177 | + } |
87 | + if (sve_access_check(s)) { | 178 | +#endif |
88 | + gen_gvec_ool_zzz(s, fn, a->rd, a->rn, a->rm, 0); | 179 | if (arm_feature(env, ARM_FEATURE_VAPA)) { |
89 | + } | 180 | ARMCPRegInfo vapa_cp_reginfo[] = { |
90 | + return true; | 181 | { .name = "PAR", .cp = 15, .crn = 7, .crm = 4, .opc1 = 0, .opc2 = 0, |
91 | +} | 182 | diff --git a/target/arm/trace-events b/target/arm/trace-events |
92 | + | 183 | index XXXXXXX..XXXXXXX 100644 |
93 | +static bool trans_SM4E(DisasContext *s, arg_rrr_esz *a) | 184 | --- a/target/arm/trace-events |
94 | +{ | 185 | +++ b/target/arm/trace-events |
95 | + return do_sm4(s, a, gen_helper_crypto_sm4e); | 186 | @@ -XXX,XX +XXX,XX @@ arm_gt_tval_write(int timer, uint64_t value) "gt_tval_write: timer %d value 0x%" |
96 | +} | 187 | arm_gt_ctl_write(int timer, uint64_t value) "gt_ctl_write: timer %d value 0x%" PRIx64 |
188 | arm_gt_imask_toggle(int timer) "gt_ctl_write: timer %d IMASK toggle" | ||
189 | arm_gt_cntvoff_write(uint64_t value) "gt_cntvoff_write: value 0x%" PRIx64 | ||
190 | +arm_gt_cntpoff_write(uint64_t value) "gt_cntpoff_write: value 0x%" PRIx64 | ||
191 | arm_gt_update_irq(int timer, int irqstate) "gt_update_irq: timer %d irqstate %d" | ||
192 | |||
193 | # kvm.c | ||
97 | -- | 194 | -- |
98 | 2.20.1 | 195 | 2.34.1 |
99 | |||
100 | diff view generated by jsdifflib |
1 | Currently we model the ITCM in the AN547's RAMInfo list. This is incorrect | 1 | Enable all FEAT_ECV features on the 'max' CPU. |
---|---|---|---|
2 | because this RAM is really a part of the SSE-300. We can't just delete | ||
3 | it from the RAMInfo list, though, because this would make boot_ram_size() | ||
4 | assert because it wouldn't be able to find an entry in the list covering | ||
5 | guest address 0. | ||
6 | |||
7 | Allow a board to specify a boot RAM size manually if it doesn't have | ||
8 | any RAM itself at address 0 and is relying on the SSE for that, and | ||
9 | set the correct value for the AN547. The other boards can continue | ||
10 | to use the "look it up from the RAMInfo list" logic. | ||
11 | 2 | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 3 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
4 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
13 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | 5 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> |
14 | Message-id: 20210510190844.17799-6-peter.maydell@linaro.org | 6 | Message-id: 20240301183219.2424889-9-peter.maydell@linaro.org |
15 | --- | 7 | --- |
16 | hw/arm/mps2-tz.c | 13 +++++++++++++ | 8 | docs/system/arm/emulation.rst | 1 + |
17 | 1 file changed, 13 insertions(+) | 9 | target/arm/tcg/cpu64.c | 1 + |
10 | 2 files changed, 2 insertions(+) | ||
18 | 11 | ||
19 | diff --git a/hw/arm/mps2-tz.c b/hw/arm/mps2-tz.c | 12 | diff --git a/docs/system/arm/emulation.rst b/docs/system/arm/emulation.rst |
20 | index XXXXXXX..XXXXXXX 100644 | 13 | index XXXXXXX..XXXXXXX 100644 |
21 | --- a/hw/arm/mps2-tz.c | 14 | --- a/docs/system/arm/emulation.rst |
22 | +++ b/hw/arm/mps2-tz.c | 15 | +++ b/docs/system/arm/emulation.rst |
23 | @@ -XXX,XX +XXX,XX @@ struct MPS2TZMachineClass { | 16 | @@ -XXX,XX +XXX,XX @@ the following architecture extensions: |
24 | uint32_t sram_addr_width; /* SRAM_ADDR_WIDTH setting for SSE */ | 17 | - FEAT_DotProd (Advanced SIMD dot product instructions) |
25 | const RAMInfo *raminfo; | 18 | - FEAT_DoubleFault (Double Fault Extension) |
26 | const char *armsse_type; | 19 | - FEAT_E0PD (Preventing EL0 access to halves of address maps) |
27 | + uint32_t boot_ram_size; /* size of ram at address 0; 0 == find in raminfo */ | 20 | +- FEAT_ECV (Enhanced Counter Virtualization) |
28 | }; | 21 | - FEAT_EPAC (Enhanced pointer authentication) |
29 | 22 | - FEAT_ETS (Enhanced Translation Synchronization) | |
30 | struct MPS2TZMachineState { | 23 | - FEAT_EVT (Enhanced Virtualization Traps) |
31 | @@ -XXX,XX +XXX,XX @@ static uint32_t boot_ram_size(MPS2TZMachineState *mms) | 24 | diff --git a/target/arm/tcg/cpu64.c b/target/arm/tcg/cpu64.c |
32 | const RAMInfo *p; | 25 | index XXXXXXX..XXXXXXX 100644 |
33 | MPS2TZMachineClass *mmc = MPS2TZ_MACHINE_GET_CLASS(mms); | 26 | --- a/target/arm/tcg/cpu64.c |
34 | 27 | +++ b/target/arm/tcg/cpu64.c | |
35 | + /* | 28 | @@ -XXX,XX +XXX,XX @@ void aarch64_max_tcg_initfn(Object *obj) |
36 | + * Use a per-board specification (for when the boot RAM is in | 29 | t = FIELD_DP64(t, ID_AA64MMFR0, TGRAN64_2, 2); /* 64k stage2 supported */ |
37 | + * the SSE and so doesn't have a RAMInfo list entry) | 30 | t = FIELD_DP64(t, ID_AA64MMFR0, TGRAN4_2, 2); /* 4k stage2 supported */ |
38 | + */ | 31 | t = FIELD_DP64(t, ID_AA64MMFR0, FGT, 1); /* FEAT_FGT */ |
39 | + if (mmc->boot_ram_size) { | 32 | + t = FIELD_DP64(t, ID_AA64MMFR0, ECV, 2); /* FEAT_ECV */ |
40 | + return mmc->boot_ram_size; | 33 | cpu->isar.id_aa64mmfr0 = t; |
41 | + } | 34 | |
42 | + | 35 | t = cpu->isar.id_aa64mmfr1; |
43 | for (p = mmc->raminfo; p->name; p++) { | ||
44 | if (p->base == boot_mem_base(mms)) { | ||
45 | return p->size; | ||
46 | @@ -XXX,XX +XXX,XX @@ static void mps2tz_an505_class_init(ObjectClass *oc, void *data) | ||
47 | mmc->sram_addr_width = 15; | ||
48 | mmc->raminfo = an505_raminfo; | ||
49 | mmc->armsse_type = TYPE_IOTKIT; | ||
50 | + mmc->boot_ram_size = 0; | ||
51 | mps2tz_set_default_ram_info(mmc); | ||
52 | } | ||
53 | |||
54 | @@ -XXX,XX +XXX,XX @@ static void mps2tz_an521_class_init(ObjectClass *oc, void *data) | ||
55 | mmc->sram_addr_width = 15; | ||
56 | mmc->raminfo = an505_raminfo; /* AN521 is the same as AN505 here */ | ||
57 | mmc->armsse_type = TYPE_SSE200; | ||
58 | + mmc->boot_ram_size = 0; | ||
59 | mps2tz_set_default_ram_info(mmc); | ||
60 | } | ||
61 | |||
62 | @@ -XXX,XX +XXX,XX @@ static void mps3tz_an524_class_init(ObjectClass *oc, void *data) | ||
63 | mmc->sram_addr_width = 15; | ||
64 | mmc->raminfo = an524_raminfo; | ||
65 | mmc->armsse_type = TYPE_SSE200; | ||
66 | + mmc->boot_ram_size = 0; | ||
67 | mps2tz_set_default_ram_info(mmc); | ||
68 | |||
69 | object_class_property_add_str(oc, "remap", mps2_get_remap, mps2_set_remap); | ||
70 | @@ -XXX,XX +XXX,XX @@ static void mps3tz_an547_class_init(ObjectClass *oc, void *data) | ||
71 | mmc->sram_addr_width = 21; | ||
72 | mmc->raminfo = an547_raminfo; | ||
73 | mmc->armsse_type = TYPE_SSE300; | ||
74 | + mmc->boot_ram_size = 512 * KiB; | ||
75 | mps2tz_set_default_ram_info(mmc); | ||
76 | } | ||
77 | |||
78 | -- | 36 | -- |
79 | 2.20.1 | 37 | 2.34.1 |
80 | 38 | ||
81 | 39 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Using g_memdup is a bit more compact than g_new + memcpy. | ||
4 | |||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
7 | Message-id: 20210509151618.2331764-2-f4bug@amsat.org | ||
8 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
9 | [PMD: Split from bigger patch] | ||
10 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
11 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | --- | ||
14 | accel/tcg/cputlb.c | 15 ++++----------- | ||
15 | 1 file changed, 4 insertions(+), 11 deletions(-) | ||
16 | |||
17 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/accel/tcg/cputlb.c | ||
20 | +++ b/accel/tcg/cputlb.c | ||
21 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
22 | } else if (encode_pbm_to_runon(&runon, d)) { | ||
23 | async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
24 | } else { | ||
25 | - TLBFlushPageBitsByMMUIdxData *p | ||
26 | - = g_new(TLBFlushPageBitsByMMUIdxData, 1); | ||
27 | - | ||
28 | /* Otherwise allocate a structure, freed by the worker. */ | ||
29 | - *p = d; | ||
30 | + TLBFlushPageBitsByMMUIdxData *p = g_memdup(&d, sizeof(d)); | ||
31 | async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
32 | RUN_ON_CPU_HOST_PTR(p)); | ||
33 | } | ||
34 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
35 | flush_all_helper(src_cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
36 | } else { | ||
37 | CPUState *dst_cpu; | ||
38 | - TLBFlushPageBitsByMMUIdxData *p; | ||
39 | |||
40 | /* Allocate a separate data block for each destination cpu. */ | ||
41 | CPU_FOREACH(dst_cpu) { | ||
42 | if (dst_cpu != src_cpu) { | ||
43 | - p = g_new(TLBFlushPageBitsByMMUIdxData, 1); | ||
44 | - *p = d; | ||
45 | + TLBFlushPageBitsByMMUIdxData *p = g_memdup(&d, sizeof(d)); | ||
46 | async_run_on_cpu(dst_cpu, | ||
47 | tlb_flush_page_bits_by_mmuidx_async_2, | ||
48 | RUN_ON_CPU_HOST_PTR(p)); | ||
49 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
50 | /* Allocate a separate data block for each destination cpu. */ | ||
51 | CPU_FOREACH(dst_cpu) { | ||
52 | if (dst_cpu != src_cpu) { | ||
53 | - p = g_new(TLBFlushPageBitsByMMUIdxData, 1); | ||
54 | - *p = d; | ||
55 | + p = g_memdup(&d, sizeof(d)); | ||
56 | async_run_on_cpu(dst_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
57 | RUN_ON_CPU_HOST_PTR(p)); | ||
58 | } | ||
59 | } | ||
60 | |||
61 | - p = g_new(TLBFlushPageBitsByMMUIdxData, 1); | ||
62 | - *p = d; | ||
63 | + p = g_memdup(&d, sizeof(d)); | ||
64 | async_safe_run_on_cpu(src_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
65 | RUN_ON_CPU_HOST_PTR(p)); | ||
66 | } | ||
67 | -- | ||
68 | 2.20.1 | ||
69 | |||
70 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Rename tlb_flush_page_bits_locked() -> tlb_flush_range_locked(), and | ||
4 | have callers pass a length argument (currently TARGET_PAGE_SIZE) via | ||
5 | the TLBFlushPageBitsByMMUIdxData structure. | ||
6 | |||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
8 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
9 | Message-id: 20210509151618.2331764-3-f4bug@amsat.org | ||
10 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
11 | [PMD: Split from bigger patch] | ||
12 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
13 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
14 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
15 | --- | ||
16 | accel/tcg/cputlb.c | 48 +++++++++++++++++++++++++++++++--------------- | ||
17 | 1 file changed, 33 insertions(+), 15 deletions(-) | ||
18 | |||
19 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/accel/tcg/cputlb.c | ||
22 | +++ b/accel/tcg/cputlb.c | ||
23 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_all_cpus_synced(CPUState *src, target_ulong addr) | ||
24 | tlb_flush_page_by_mmuidx_all_cpus_synced(src, addr, ALL_MMUIDX_BITS); | ||
25 | } | ||
26 | |||
27 | -static void tlb_flush_page_bits_locked(CPUArchState *env, int midx, | ||
28 | - target_ulong page, unsigned bits) | ||
29 | +static void tlb_flush_range_locked(CPUArchState *env, int midx, | ||
30 | + target_ulong addr, target_ulong len, | ||
31 | + unsigned bits) | ||
32 | { | ||
33 | CPUTLBDesc *d = &env_tlb(env)->d[midx]; | ||
34 | CPUTLBDescFast *f = &env_tlb(env)->f[midx]; | ||
35 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_locked(CPUArchState *env, int midx, | ||
36 | * If @bits is smaller than the tlb size, there may be multiple entries | ||
37 | * within the TLB; otherwise all addresses that match under @mask hit | ||
38 | * the same TLB entry. | ||
39 | - * | ||
40 | * TODO: Perhaps allow bits to be a few bits less than the size. | ||
41 | * For now, just flush the entire TLB. | ||
42 | + * | ||
43 | + * If @len is larger than the tlb size, then it will take longer to | ||
44 | + * test all of the entries in the TLB than it will to flush it all. | ||
45 | */ | ||
46 | - if (mask < f->mask) { | ||
47 | + if (mask < f->mask || len > f->mask) { | ||
48 | tlb_debug("forcing full flush midx %d (" | ||
49 | - TARGET_FMT_lx "/" TARGET_FMT_lx ")\n", | ||
50 | - midx, page, mask); | ||
51 | + TARGET_FMT_lx "/" TARGET_FMT_lx "+" TARGET_FMT_lx ")\n", | ||
52 | + midx, addr, mask, len); | ||
53 | tlb_flush_one_mmuidx_locked(env, midx, get_clock_realtime()); | ||
54 | return; | ||
55 | } | ||
56 | |||
57 | - /* Check if we need to flush due to large pages. */ | ||
58 | - if ((page & d->large_page_mask) == d->large_page_addr) { | ||
59 | + /* | ||
60 | + * Check if we need to flush due to large pages. | ||
61 | + * Because large_page_mask contains all 1's from the msb, | ||
62 | + * we only need to test the end of the range. | ||
63 | + */ | ||
64 | + if (((addr + len - 1) & d->large_page_mask) == d->large_page_addr) { | ||
65 | tlb_debug("forcing full flush midx %d (" | ||
66 | TARGET_FMT_lx "/" TARGET_FMT_lx ")\n", | ||
67 | midx, d->large_page_addr, d->large_page_mask); | ||
68 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_locked(CPUArchState *env, int midx, | ||
69 | return; | ||
70 | } | ||
71 | |||
72 | - if (tlb_flush_entry_mask_locked(tlb_entry(env, midx, page), page, mask)) { | ||
73 | - tlb_n_used_entries_dec(env, midx); | ||
74 | + for (target_ulong i = 0; i < len; i += TARGET_PAGE_SIZE) { | ||
75 | + target_ulong page = addr + i; | ||
76 | + CPUTLBEntry *entry = tlb_entry(env, midx, page); | ||
77 | + | ||
78 | + if (tlb_flush_entry_mask_locked(entry, page, mask)) { | ||
79 | + tlb_n_used_entries_dec(env, midx); | ||
80 | + } | ||
81 | + tlb_flush_vtlb_page_mask_locked(env, midx, page, mask); | ||
82 | } | ||
83 | - tlb_flush_vtlb_page_mask_locked(env, midx, page, mask); | ||
84 | } | ||
85 | |||
86 | typedef struct { | ||
87 | target_ulong addr; | ||
88 | + target_ulong len; | ||
89 | uint16_t idxmap; | ||
90 | uint16_t bits; | ||
91 | } TLBFlushPageBitsByMMUIdxData; | ||
92 | @@ -XXX,XX +XXX,XX @@ tlb_flush_page_bits_by_mmuidx_async_0(CPUState *cpu, | ||
93 | |||
94 | assert_cpu_is_self(cpu); | ||
95 | |||
96 | - tlb_debug("page addr:" TARGET_FMT_lx "/%u mmu_map:0x%x\n", | ||
97 | - d.addr, d.bits, d.idxmap); | ||
98 | + tlb_debug("range:" TARGET_FMT_lx "/%u+" TARGET_FMT_lx " mmu_map:0x%x\n", | ||
99 | + d.addr, d.bits, d.len, d.idxmap); | ||
100 | |||
101 | qemu_spin_lock(&env_tlb(env)->c.lock); | ||
102 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
103 | if ((d.idxmap >> mmu_idx) & 1) { | ||
104 | - tlb_flush_page_bits_locked(env, mmu_idx, d.addr, d.bits); | ||
105 | + tlb_flush_range_locked(env, mmu_idx, d.addr, d.len, d.bits); | ||
106 | } | ||
107 | } | ||
108 | qemu_spin_unlock(&env_tlb(env)->c.lock); | ||
109 | |||
110 | - tb_flush_jmp_cache(cpu, d.addr); | ||
111 | + for (target_ulong i = 0; i < d.len; i += TARGET_PAGE_SIZE) { | ||
112 | + tb_flush_jmp_cache(cpu, d.addr + i); | ||
113 | + } | ||
114 | } | ||
115 | |||
116 | static bool encode_pbm_to_runon(run_on_cpu_data *out, | ||
117 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
118 | |||
119 | /* This should already be page aligned */ | ||
120 | d.addr = addr & TARGET_PAGE_MASK; | ||
121 | + d.len = TARGET_PAGE_SIZE; | ||
122 | d.idxmap = idxmap; | ||
123 | d.bits = bits; | ||
124 | |||
125 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
126 | |||
127 | /* This should already be page aligned */ | ||
128 | d.addr = addr & TARGET_PAGE_MASK; | ||
129 | + d.len = TARGET_PAGE_SIZE; | ||
130 | d.idxmap = idxmap; | ||
131 | d.bits = bits; | ||
132 | |||
133 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
134 | |||
135 | /* This should already be page aligned */ | ||
136 | d.addr = addr & TARGET_PAGE_MASK; | ||
137 | + d.len = TARGET_PAGE_SIZE; | ||
138 | d.idxmap = idxmap; | ||
139 | d.bits = bits; | ||
140 | |||
141 | -- | ||
142 | 2.20.1 | ||
143 | |||
144 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Rename the structure to match the rename of tlb_flush_range_locked. | ||
4 | |||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
7 | Message-id: 20210509151618.2331764-4-f4bug@amsat.org | ||
8 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
9 | [PMD: Split from bigger patch] | ||
10 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
11 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | --- | ||
14 | accel/tcg/cputlb.c | 24 ++++++++++++------------ | ||
15 | 1 file changed, 12 insertions(+), 12 deletions(-) | ||
16 | |||
17 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/accel/tcg/cputlb.c | ||
20 | +++ b/accel/tcg/cputlb.c | ||
21 | @@ -XXX,XX +XXX,XX @@ typedef struct { | ||
22 | target_ulong len; | ||
23 | uint16_t idxmap; | ||
24 | uint16_t bits; | ||
25 | -} TLBFlushPageBitsByMMUIdxData; | ||
26 | +} TLBFlushRangeData; | ||
27 | |||
28 | static void | ||
29 | tlb_flush_page_bits_by_mmuidx_async_0(CPUState *cpu, | ||
30 | - TLBFlushPageBitsByMMUIdxData d) | ||
31 | + TLBFlushRangeData d) | ||
32 | { | ||
33 | CPUArchState *env = cpu->env_ptr; | ||
34 | int mmu_idx; | ||
35 | @@ -XXX,XX +XXX,XX @@ tlb_flush_page_bits_by_mmuidx_async_0(CPUState *cpu, | ||
36 | } | ||
37 | |||
38 | static bool encode_pbm_to_runon(run_on_cpu_data *out, | ||
39 | - TLBFlushPageBitsByMMUIdxData d) | ||
40 | + TLBFlushRangeData d) | ||
41 | { | ||
42 | /* We need 6 bits to hold to hold @bits up to 63. */ | ||
43 | if (d.idxmap <= MAKE_64BIT_MASK(0, TARGET_PAGE_BITS - 6)) { | ||
44 | @@ -XXX,XX +XXX,XX @@ static bool encode_pbm_to_runon(run_on_cpu_data *out, | ||
45 | return false; | ||
46 | } | ||
47 | |||
48 | -static TLBFlushPageBitsByMMUIdxData | ||
49 | +static TLBFlushRangeData | ||
50 | decode_runon_to_pbm(run_on_cpu_data data) | ||
51 | { | ||
52 | target_ulong addr_map_bits = (target_ulong) data.target_ptr; | ||
53 | - return (TLBFlushPageBitsByMMUIdxData){ | ||
54 | + return (TLBFlushRangeData){ | ||
55 | .addr = addr_map_bits & TARGET_PAGE_MASK, | ||
56 | .idxmap = (addr_map_bits & ~TARGET_PAGE_MASK) >> 6, | ||
57 | .bits = addr_map_bits & 0x3f | ||
58 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_by_mmuidx_async_1(CPUState *cpu, | ||
59 | static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
60 | run_on_cpu_data data) | ||
61 | { | ||
62 | - TLBFlushPageBitsByMMUIdxData *d = data.host_ptr; | ||
63 | + TLBFlushRangeData *d = data.host_ptr; | ||
64 | tlb_flush_page_bits_by_mmuidx_async_0(cpu, *d); | ||
65 | g_free(d); | ||
66 | } | ||
67 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
68 | void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
69 | uint16_t idxmap, unsigned bits) | ||
70 | { | ||
71 | - TLBFlushPageBitsByMMUIdxData d; | ||
72 | + TLBFlushRangeData d; | ||
73 | run_on_cpu_data runon; | ||
74 | |||
75 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
76 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
77 | async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
78 | } else { | ||
79 | /* Otherwise allocate a structure, freed by the worker. */ | ||
80 | - TLBFlushPageBitsByMMUIdxData *p = g_memdup(&d, sizeof(d)); | ||
81 | + TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
82 | async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
83 | RUN_ON_CPU_HOST_PTR(p)); | ||
84 | } | ||
85 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
86 | uint16_t idxmap, | ||
87 | unsigned bits) | ||
88 | { | ||
89 | - TLBFlushPageBitsByMMUIdxData d; | ||
90 | + TLBFlushRangeData d; | ||
91 | run_on_cpu_data runon; | ||
92 | |||
93 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
94 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
95 | /* Allocate a separate data block for each destination cpu. */ | ||
96 | CPU_FOREACH(dst_cpu) { | ||
97 | if (dst_cpu != src_cpu) { | ||
98 | - TLBFlushPageBitsByMMUIdxData *p = g_memdup(&d, sizeof(d)); | ||
99 | + TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
100 | async_run_on_cpu(dst_cpu, | ||
101 | tlb_flush_page_bits_by_mmuidx_async_2, | ||
102 | RUN_ON_CPU_HOST_PTR(p)); | ||
103 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
104 | uint16_t idxmap, | ||
105 | unsigned bits) | ||
106 | { | ||
107 | - TLBFlushPageBitsByMMUIdxData d; | ||
108 | + TLBFlushRangeData d; | ||
109 | run_on_cpu_data runon; | ||
110 | |||
111 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
112 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
113 | runon); | ||
114 | } else { | ||
115 | CPUState *dst_cpu; | ||
116 | - TLBFlushPageBitsByMMUIdxData *p; | ||
117 | + TLBFlushRangeData *p; | ||
118 | |||
119 | /* Allocate a separate data block for each destination cpu. */ | ||
120 | CPU_FOREACH(dst_cpu) { | ||
121 | -- | ||
122 | 2.20.1 | ||
123 | |||
124 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | We will not be able to fit address + length into a 64-bit packet. | ||
4 | Drop this optimization before re-organizing this code. | ||
5 | |||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
8 | Message-id: 20210509151618.2331764-10-f4bug@amsat.org | ||
9 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
10 | [PMD: Split from bigger patch] | ||
11 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
12 | [PMM: Moved patch earlier in the series] | ||
13 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
14 | --- | ||
15 | accel/tcg/cputlb.c | 86 +++++++++++----------------------------------- | ||
16 | 1 file changed, 20 insertions(+), 66 deletions(-) | ||
17 | |||
18 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/accel/tcg/cputlb.c | ||
21 | +++ b/accel/tcg/cputlb.c | ||
22 | @@ -XXX,XX +XXX,XX @@ tlb_flush_page_bits_by_mmuidx_async_0(CPUState *cpu, | ||
23 | } | ||
24 | } | ||
25 | |||
26 | -static bool encode_pbm_to_runon(run_on_cpu_data *out, | ||
27 | - TLBFlushRangeData d) | ||
28 | -{ | ||
29 | - /* We need 6 bits to hold to hold @bits up to 63. */ | ||
30 | - if (d.idxmap <= MAKE_64BIT_MASK(0, TARGET_PAGE_BITS - 6)) { | ||
31 | - *out = RUN_ON_CPU_TARGET_PTR(d.addr | (d.idxmap << 6) | d.bits); | ||
32 | - return true; | ||
33 | - } | ||
34 | - return false; | ||
35 | -} | ||
36 | - | ||
37 | -static TLBFlushRangeData | ||
38 | -decode_runon_to_pbm(run_on_cpu_data data) | ||
39 | -{ | ||
40 | - target_ulong addr_map_bits = (target_ulong) data.target_ptr; | ||
41 | - return (TLBFlushRangeData){ | ||
42 | - .addr = addr_map_bits & TARGET_PAGE_MASK, | ||
43 | - .idxmap = (addr_map_bits & ~TARGET_PAGE_MASK) >> 6, | ||
44 | - .bits = addr_map_bits & 0x3f | ||
45 | - }; | ||
46 | -} | ||
47 | - | ||
48 | -static void tlb_flush_page_bits_by_mmuidx_async_1(CPUState *cpu, | ||
49 | - run_on_cpu_data runon) | ||
50 | -{ | ||
51 | - tlb_flush_page_bits_by_mmuidx_async_0(cpu, decode_runon_to_pbm(runon)); | ||
52 | -} | ||
53 | - | ||
54 | static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
55 | run_on_cpu_data data) | ||
56 | { | ||
57 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
58 | uint16_t idxmap, unsigned bits) | ||
59 | { | ||
60 | TLBFlushRangeData d; | ||
61 | - run_on_cpu_data runon; | ||
62 | |||
63 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
64 | if (bits >= TARGET_LONG_BITS) { | ||
65 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
66 | |||
67 | if (qemu_cpu_is_self(cpu)) { | ||
68 | tlb_flush_page_bits_by_mmuidx_async_0(cpu, d); | ||
69 | - } else if (encode_pbm_to_runon(&runon, d)) { | ||
70 | - async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
71 | } else { | ||
72 | /* Otherwise allocate a structure, freed by the worker. */ | ||
73 | TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
74 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
75 | unsigned bits) | ||
76 | { | ||
77 | TLBFlushRangeData d; | ||
78 | - run_on_cpu_data runon; | ||
79 | + CPUState *dst_cpu; | ||
80 | |||
81 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
82 | if (bits >= TARGET_LONG_BITS) { | ||
83 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
84 | d.idxmap = idxmap; | ||
85 | d.bits = bits; | ||
86 | |||
87 | - if (encode_pbm_to_runon(&runon, d)) { | ||
88 | - flush_all_helper(src_cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
89 | - } else { | ||
90 | - CPUState *dst_cpu; | ||
91 | - | ||
92 | - /* Allocate a separate data block for each destination cpu. */ | ||
93 | - CPU_FOREACH(dst_cpu) { | ||
94 | - if (dst_cpu != src_cpu) { | ||
95 | - TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
96 | - async_run_on_cpu(dst_cpu, | ||
97 | - tlb_flush_page_bits_by_mmuidx_async_2, | ||
98 | - RUN_ON_CPU_HOST_PTR(p)); | ||
99 | - } | ||
100 | + /* Allocate a separate data block for each destination cpu. */ | ||
101 | + CPU_FOREACH(dst_cpu) { | ||
102 | + if (dst_cpu != src_cpu) { | ||
103 | + TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
104 | + async_run_on_cpu(dst_cpu, | ||
105 | + tlb_flush_page_bits_by_mmuidx_async_2, | ||
106 | + RUN_ON_CPU_HOST_PTR(p)); | ||
107 | } | ||
108 | } | ||
109 | |||
110 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
111 | uint16_t idxmap, | ||
112 | unsigned bits) | ||
113 | { | ||
114 | - TLBFlushRangeData d; | ||
115 | - run_on_cpu_data runon; | ||
116 | + TLBFlushRangeData d, *p; | ||
117 | + CPUState *dst_cpu; | ||
118 | |||
119 | /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
120 | if (bits >= TARGET_LONG_BITS) { | ||
121 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
122 | d.idxmap = idxmap; | ||
123 | d.bits = bits; | ||
124 | |||
125 | - if (encode_pbm_to_runon(&runon, d)) { | ||
126 | - flush_all_helper(src_cpu, tlb_flush_page_bits_by_mmuidx_async_1, runon); | ||
127 | - async_safe_run_on_cpu(src_cpu, tlb_flush_page_bits_by_mmuidx_async_1, | ||
128 | - runon); | ||
129 | - } else { | ||
130 | - CPUState *dst_cpu; | ||
131 | - TLBFlushRangeData *p; | ||
132 | - | ||
133 | - /* Allocate a separate data block for each destination cpu. */ | ||
134 | - CPU_FOREACH(dst_cpu) { | ||
135 | - if (dst_cpu != src_cpu) { | ||
136 | - p = g_memdup(&d, sizeof(d)); | ||
137 | - async_run_on_cpu(dst_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
138 | - RUN_ON_CPU_HOST_PTR(p)); | ||
139 | - } | ||
140 | + /* Allocate a separate data block for each destination cpu. */ | ||
141 | + CPU_FOREACH(dst_cpu) { | ||
142 | + if (dst_cpu != src_cpu) { | ||
143 | + p = g_memdup(&d, sizeof(d)); | ||
144 | + async_run_on_cpu(dst_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
145 | + RUN_ON_CPU_HOST_PTR(p)); | ||
146 | } | ||
147 | - | ||
148 | - p = g_memdup(&d, sizeof(d)); | ||
149 | - async_safe_run_on_cpu(src_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
150 | - RUN_ON_CPU_HOST_PTR(p)); | ||
151 | } | ||
152 | + | ||
153 | + p = g_memdup(&d, sizeof(d)); | ||
154 | + async_safe_run_on_cpu(src_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
155 | + RUN_ON_CPU_HOST_PTR(p)); | ||
156 | } | ||
157 | |||
158 | /* update the TLBs so that writes to code in the virtual page 'addr' | ||
159 | -- | ||
160 | 2.20.1 | ||
161 | |||
162 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Forward tlb_flush_page_bits_by_mmuidx to tlb_flush_range_by_mmuidx | ||
4 | passing TARGET_PAGE_SIZE. | ||
5 | |||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
8 | Message-id: 20210509151618.2331764-5-f4bug@amsat.org | ||
9 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
10 | [PMD: Split from bigger patch] | ||
11 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
12 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
14 | --- | ||
15 | include/exec/exec-all.h | 19 +++++++++++++++++++ | ||
16 | accel/tcg/cputlb.c | 20 +++++++++++++++----- | ||
17 | 2 files changed, 34 insertions(+), 5 deletions(-) | ||
18 | |||
19 | diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/include/exec/exec-all.h | ||
22 | +++ b/include/exec/exec-all.h | ||
23 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *cpu, target_ulong addr, | ||
24 | void tlb_flush_page_bits_by_mmuidx_all_cpus_synced | ||
25 | (CPUState *cpu, target_ulong addr, uint16_t idxmap, unsigned bits); | ||
26 | |||
27 | +/** | ||
28 | + * tlb_flush_range_by_mmuidx | ||
29 | + * @cpu: CPU whose TLB should be flushed | ||
30 | + * @addr: virtual address of the start of the range to be flushed | ||
31 | + * @len: length of range to be flushed | ||
32 | + * @idxmap: bitmap of mmu indexes to flush | ||
33 | + * @bits: number of significant bits in address | ||
34 | + * | ||
35 | + * For each mmuidx in @idxmap, flush all pages within [@addr,@addr+@len), | ||
36 | + * comparing only the low @bits worth of each virtual page. | ||
37 | + */ | ||
38 | +void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
39 | + target_ulong len, uint16_t idxmap, | ||
40 | + unsigned bits); | ||
41 | /** | ||
42 | * tlb_set_page_with_attrs: | ||
43 | * @cpu: CPU to add this TLB entry for | ||
44 | @@ -XXX,XX +XXX,XX @@ tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *cpu, target_ulong addr, | ||
45 | uint16_t idxmap, unsigned bits) | ||
46 | { | ||
47 | } | ||
48 | +static inline void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
49 | + target_ulong len, uint16_t idxmap, | ||
50 | + unsigned bits) | ||
51 | +{ | ||
52 | +} | ||
53 | #endif | ||
54 | /** | ||
55 | * probe_access: | ||
56 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
57 | index XXXXXXX..XXXXXXX 100644 | ||
58 | --- a/accel/tcg/cputlb.c | ||
59 | +++ b/accel/tcg/cputlb.c | ||
60 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
61 | g_free(d); | ||
62 | } | ||
63 | |||
64 | -void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
65 | - uint16_t idxmap, unsigned bits) | ||
66 | +void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
67 | + target_ulong len, uint16_t idxmap, | ||
68 | + unsigned bits) | ||
69 | { | ||
70 | TLBFlushRangeData d; | ||
71 | |||
72 | - /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
73 | - if (bits >= TARGET_LONG_BITS) { | ||
74 | + /* | ||
75 | + * If all bits are significant, and len is small, | ||
76 | + * this devolves to tlb_flush_page. | ||
77 | + */ | ||
78 | + if (bits >= TARGET_LONG_BITS && len <= TARGET_PAGE_SIZE) { | ||
79 | tlb_flush_page_by_mmuidx(cpu, addr, idxmap); | ||
80 | return; | ||
81 | } | ||
82 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
83 | |||
84 | /* This should already be page aligned */ | ||
85 | d.addr = addr & TARGET_PAGE_MASK; | ||
86 | - d.len = TARGET_PAGE_SIZE; | ||
87 | + d.len = len; | ||
88 | d.idxmap = idxmap; | ||
89 | d.bits = bits; | ||
90 | |||
91 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
92 | } | ||
93 | } | ||
94 | |||
95 | +void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
96 | + uint16_t idxmap, unsigned bits) | ||
97 | +{ | ||
98 | + tlb_flush_range_by_mmuidx(cpu, addr, TARGET_PAGE_SIZE, idxmap, bits); | ||
99 | +} | ||
100 | + | ||
101 | void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
102 | target_ulong addr, | ||
103 | uint16_t idxmap, | ||
104 | -- | ||
105 | 2.20.1 | ||
106 | |||
107 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Forward tlb_flush_page_bits_by_mmuidx_all_cpus to | ||
4 | tlb_flush_range_by_mmuidx_all_cpus passing TARGET_PAGE_SIZE. | ||
5 | |||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
8 | Message-id: 20210509151618.2331764-6-f4bug@amsat.org | ||
9 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
10 | [PMD: Split from bigger patch] | ||
11 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
12 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
14 | --- | ||
15 | include/exec/exec-all.h | 13 +++++++++++++ | ||
16 | accel/tcg/cputlb.c | 24 +++++++++++++++++------- | ||
17 | 2 files changed, 30 insertions(+), 7 deletions(-) | ||
18 | |||
19 | diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/include/exec/exec-all.h | ||
22 | +++ b/include/exec/exec-all.h | ||
23 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced | ||
24 | void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
25 | target_ulong len, uint16_t idxmap, | ||
26 | unsigned bits); | ||
27 | + | ||
28 | +/* Similarly, with broadcast and syncing. */ | ||
29 | +void tlb_flush_range_by_mmuidx_all_cpus(CPUState *cpu, target_ulong addr, | ||
30 | + target_ulong len, uint16_t idxmap, | ||
31 | + unsigned bits); | ||
32 | + | ||
33 | /** | ||
34 | * tlb_set_page_with_attrs: | ||
35 | * @cpu: CPU to add this TLB entry for | ||
36 | @@ -XXX,XX +XXX,XX @@ static inline void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
37 | unsigned bits) | ||
38 | { | ||
39 | } | ||
40 | +static inline void tlb_flush_range_by_mmuidx_all_cpus(CPUState *cpu, | ||
41 | + target_ulong addr, | ||
42 | + target_ulong len, | ||
43 | + uint16_t idxmap, | ||
44 | + unsigned bits) | ||
45 | +{ | ||
46 | +} | ||
47 | #endif | ||
48 | /** | ||
49 | * probe_access: | ||
50 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/accel/tcg/cputlb.c | ||
53 | +++ b/accel/tcg/cputlb.c | ||
54 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
55 | tlb_flush_range_by_mmuidx(cpu, addr, TARGET_PAGE_SIZE, idxmap, bits); | ||
56 | } | ||
57 | |||
58 | -void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
59 | - target_ulong addr, | ||
60 | - uint16_t idxmap, | ||
61 | - unsigned bits) | ||
62 | +void tlb_flush_range_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
63 | + target_ulong addr, target_ulong len, | ||
64 | + uint16_t idxmap, unsigned bits) | ||
65 | { | ||
66 | TLBFlushRangeData d; | ||
67 | CPUState *dst_cpu; | ||
68 | |||
69 | - /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
70 | - if (bits >= TARGET_LONG_BITS) { | ||
71 | + /* | ||
72 | + * If all bits are significant, and len is small, | ||
73 | + * this devolves to tlb_flush_page. | ||
74 | + */ | ||
75 | + if (bits >= TARGET_LONG_BITS && len <= TARGET_PAGE_SIZE) { | ||
76 | tlb_flush_page_by_mmuidx_all_cpus(src_cpu, addr, idxmap); | ||
77 | return; | ||
78 | } | ||
79 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
80 | |||
81 | /* This should already be page aligned */ | ||
82 | d.addr = addr & TARGET_PAGE_MASK; | ||
83 | - d.len = TARGET_PAGE_SIZE; | ||
84 | + d.len = len; | ||
85 | d.idxmap = idxmap; | ||
86 | d.bits = bits; | ||
87 | |||
88 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
89 | tlb_flush_page_bits_by_mmuidx_async_0(src_cpu, d); | ||
90 | } | ||
91 | |||
92 | +void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
93 | + target_ulong addr, | ||
94 | + uint16_t idxmap, unsigned bits) | ||
95 | +{ | ||
96 | + tlb_flush_range_by_mmuidx_all_cpus(src_cpu, addr, TARGET_PAGE_SIZE, | ||
97 | + idxmap, bits); | ||
98 | +} | ||
99 | + | ||
100 | void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
101 | target_ulong addr, | ||
102 | uint16_t idxmap, | ||
103 | -- | ||
104 | 2.20.1 | ||
105 | |||
106 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Forward tlb_flush_page_bits_by_mmuidx_all_cpus_synced to | ||
4 | tlb_flush_range_by_mmuidx_all_cpus_synced passing TARGET_PAGE_SIZE. | ||
5 | |||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
8 | Message-id: 20210509151618.2331764-7-f4bug@amsat.org | ||
9 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
10 | [PMD: Split from bigger patch] | ||
11 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
12 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
14 | --- | ||
15 | include/exec/exec-all.h | 12 ++++++++++++ | ||
16 | accel/tcg/cputlb.c | 27 ++++++++++++++++++++------- | ||
17 | 2 files changed, 32 insertions(+), 7 deletions(-) | ||
18 | |||
19 | diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/include/exec/exec-all.h | ||
22 | +++ b/include/exec/exec-all.h | ||
23 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
24 | void tlb_flush_range_by_mmuidx_all_cpus(CPUState *cpu, target_ulong addr, | ||
25 | target_ulong len, uint16_t idxmap, | ||
26 | unsigned bits); | ||
27 | +void tlb_flush_range_by_mmuidx_all_cpus_synced(CPUState *cpu, | ||
28 | + target_ulong addr, | ||
29 | + target_ulong len, | ||
30 | + uint16_t idxmap, | ||
31 | + unsigned bits); | ||
32 | |||
33 | /** | ||
34 | * tlb_set_page_with_attrs: | ||
35 | @@ -XXX,XX +XXX,XX @@ static inline void tlb_flush_range_by_mmuidx_all_cpus(CPUState *cpu, | ||
36 | unsigned bits) | ||
37 | { | ||
38 | } | ||
39 | +static inline void tlb_flush_range_by_mmuidx_all_cpus_synced(CPUState *cpu, | ||
40 | + target_ulong addr, | ||
41 | + target_long len, | ||
42 | + uint16_t idxmap, | ||
43 | + unsigned bits) | ||
44 | +{ | ||
45 | +} | ||
46 | #endif | ||
47 | /** | ||
48 | * probe_access: | ||
49 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/accel/tcg/cputlb.c | ||
52 | +++ b/accel/tcg/cputlb.c | ||
53 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
54 | idxmap, bits); | ||
55 | } | ||
56 | |||
57 | -void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
58 | - target_ulong addr, | ||
59 | - uint16_t idxmap, | ||
60 | - unsigned bits) | ||
61 | +void tlb_flush_range_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
62 | + target_ulong addr, | ||
63 | + target_ulong len, | ||
64 | + uint16_t idxmap, | ||
65 | + unsigned bits) | ||
66 | { | ||
67 | TLBFlushRangeData d, *p; | ||
68 | CPUState *dst_cpu; | ||
69 | |||
70 | - /* If all bits are significant, this devolves to tlb_flush_page. */ | ||
71 | - if (bits >= TARGET_LONG_BITS) { | ||
72 | + /* | ||
73 | + * If all bits are significant, and len is small, | ||
74 | + * this devolves to tlb_flush_page. | ||
75 | + */ | ||
76 | + if (bits >= TARGET_LONG_BITS && len <= TARGET_PAGE_SIZE) { | ||
77 | tlb_flush_page_by_mmuidx_all_cpus_synced(src_cpu, addr, idxmap); | ||
78 | return; | ||
79 | } | ||
80 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
81 | |||
82 | /* This should already be page aligned */ | ||
83 | d.addr = addr & TARGET_PAGE_MASK; | ||
84 | - d.len = TARGET_PAGE_SIZE; | ||
85 | + d.len = len; | ||
86 | d.idxmap = idxmap; | ||
87 | d.bits = bits; | ||
88 | |||
89 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
90 | RUN_ON_CPU_HOST_PTR(p)); | ||
91 | } | ||
92 | |||
93 | +void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
94 | + target_ulong addr, | ||
95 | + uint16_t idxmap, | ||
96 | + unsigned bits) | ||
97 | +{ | ||
98 | + tlb_flush_range_by_mmuidx_all_cpus_synced(src_cpu, addr, TARGET_PAGE_SIZE, | ||
99 | + idxmap, bits); | ||
100 | +} | ||
101 | + | ||
102 | /* update the TLBs so that writes to code in the virtual page 'addr' | ||
103 | can be detected */ | ||
104 | void tlb_protect_code(ram_addr_t ram_addr) | ||
105 | -- | ||
106 | 2.20.1 | ||
107 | |||
108 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Rename to match tlb_flush_range_locked. | ||
4 | |||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
7 | Message-id: 20210509151618.2331764-8-f4bug@amsat.org | ||
8 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
9 | [PMD: Split from bigger patch] | ||
10 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
11 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | --- | ||
14 | accel/tcg/cputlb.c | 11 +++++------ | ||
15 | 1 file changed, 5 insertions(+), 6 deletions(-) | ||
16 | |||
17 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/accel/tcg/cputlb.c | ||
20 | +++ b/accel/tcg/cputlb.c | ||
21 | @@ -XXX,XX +XXX,XX @@ typedef struct { | ||
22 | uint16_t bits; | ||
23 | } TLBFlushRangeData; | ||
24 | |||
25 | -static void | ||
26 | -tlb_flush_page_bits_by_mmuidx_async_0(CPUState *cpu, | ||
27 | - TLBFlushRangeData d) | ||
28 | +static void tlb_flush_range_by_mmuidx_async_0(CPUState *cpu, | ||
29 | + TLBFlushRangeData d) | ||
30 | { | ||
31 | CPUArchState *env = cpu->env_ptr; | ||
32 | int mmu_idx; | ||
33 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
34 | run_on_cpu_data data) | ||
35 | { | ||
36 | TLBFlushRangeData *d = data.host_ptr; | ||
37 | - tlb_flush_page_bits_by_mmuidx_async_0(cpu, *d); | ||
38 | + tlb_flush_range_by_mmuidx_async_0(cpu, *d); | ||
39 | g_free(d); | ||
40 | } | ||
41 | |||
42 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
43 | d.bits = bits; | ||
44 | |||
45 | if (qemu_cpu_is_self(cpu)) { | ||
46 | - tlb_flush_page_bits_by_mmuidx_async_0(cpu, d); | ||
47 | + tlb_flush_range_by_mmuidx_async_0(cpu, d); | ||
48 | } else { | ||
49 | /* Otherwise allocate a structure, freed by the worker. */ | ||
50 | TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
51 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
52 | } | ||
53 | } | ||
54 | |||
55 | - tlb_flush_page_bits_by_mmuidx_async_0(src_cpu, d); | ||
56 | + tlb_flush_range_by_mmuidx_async_0(src_cpu, d); | ||
57 | } | ||
58 | |||
59 | void tlb_flush_page_bits_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
60 | -- | ||
61 | 2.20.1 | ||
62 | |||
63 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Rename to match tlb_flush_range_locked. | ||
4 | |||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
7 | Message-id: 20210509151618.2331764-9-f4bug@amsat.org | ||
8 | Message-Id: <20210508201640.1045808-1-richard.henderson@linaro.org> | ||
9 | [PMD: Split from bigger patch] | ||
10 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
11 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
13 | --- | ||
14 | accel/tcg/cputlb.c | 12 ++++++------ | ||
15 | 1 file changed, 6 insertions(+), 6 deletions(-) | ||
16 | |||
17 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/accel/tcg/cputlb.c | ||
20 | +++ b/accel/tcg/cputlb.c | ||
21 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_range_by_mmuidx_async_0(CPUState *cpu, | ||
22 | } | ||
23 | } | ||
24 | |||
25 | -static void tlb_flush_page_bits_by_mmuidx_async_2(CPUState *cpu, | ||
26 | - run_on_cpu_data data) | ||
27 | +static void tlb_flush_range_by_mmuidx_async_1(CPUState *cpu, | ||
28 | + run_on_cpu_data data) | ||
29 | { | ||
30 | TLBFlushRangeData *d = data.host_ptr; | ||
31 | tlb_flush_range_by_mmuidx_async_0(cpu, *d); | ||
32 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx(CPUState *cpu, target_ulong addr, | ||
33 | } else { | ||
34 | /* Otherwise allocate a structure, freed by the worker. */ | ||
35 | TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
36 | - async_run_on_cpu(cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
37 | + async_run_on_cpu(cpu, tlb_flush_range_by_mmuidx_async_1, | ||
38 | RUN_ON_CPU_HOST_PTR(p)); | ||
39 | } | ||
40 | } | ||
41 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx_all_cpus(CPUState *src_cpu, | ||
42 | if (dst_cpu != src_cpu) { | ||
43 | TLBFlushRangeData *p = g_memdup(&d, sizeof(d)); | ||
44 | async_run_on_cpu(dst_cpu, | ||
45 | - tlb_flush_page_bits_by_mmuidx_async_2, | ||
46 | + tlb_flush_range_by_mmuidx_async_1, | ||
47 | RUN_ON_CPU_HOST_PTR(p)); | ||
48 | } | ||
49 | } | ||
50 | @@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx_all_cpus_synced(CPUState *src_cpu, | ||
51 | CPU_FOREACH(dst_cpu) { | ||
52 | if (dst_cpu != src_cpu) { | ||
53 | p = g_memdup(&d, sizeof(d)); | ||
54 | - async_run_on_cpu(dst_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
55 | + async_run_on_cpu(dst_cpu, tlb_flush_range_by_mmuidx_async_1, | ||
56 | RUN_ON_CPU_HOST_PTR(p)); | ||
57 | } | ||
58 | } | ||
59 | |||
60 | p = g_memdup(&d, sizeof(d)); | ||
61 | - async_safe_run_on_cpu(src_cpu, tlb_flush_page_bits_by_mmuidx_async_2, | ||
62 | + async_safe_run_on_cpu(src_cpu, tlb_flush_range_by_mmuidx_async_1, | ||
63 | RUN_ON_CPU_HOST_PTR(p)); | ||
64 | } | ||
65 | |||
66 | -- | ||
67 | 2.20.1 | ||
68 | |||
69 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Rebecca Cran <rebecca@nuviainc.com> | ||
2 | 1 | ||
3 | Indicate support for FEAT_TLBIOS and FEAT_TLBIRANGE by setting | ||
4 | ID_AA64ISAR0.TLB to 2 for the max AARCH64 CPU type. | ||
5 | |||
6 | Signed-off-by: Rebecca Cran <rebecca@nuviainc.com> | ||
7 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
8 | Message-id: 20210512182337.18563-4-rebecca@nuviainc.com | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/cpu64.c | 1 + | ||
12 | 1 file changed, 1 insertion(+) | ||
13 | |||
14 | diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/cpu64.c | ||
17 | +++ b/target/arm/cpu64.c | ||
18 | @@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj) | ||
19 | t = FIELD_DP64(t, ID_AA64ISAR0, DP, 1); | ||
20 | t = FIELD_DP64(t, ID_AA64ISAR0, FHM, 1); | ||
21 | t = FIELD_DP64(t, ID_AA64ISAR0, TS, 2); /* v8.5-CondM */ | ||
22 | + t = FIELD_DP64(t, ID_AA64ISAR0, TLB, 2); /* FEAT_TLBIRANGE */ | ||
23 | t = FIELD_DP64(t, ID_AA64ISAR0, RNDR, 1); | ||
24 | cpu->isar.id_aa64isar0 = t; | ||
25 | |||
26 | -- | ||
27 | 2.20.1 | ||
28 | |||
29 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
2 | 1 | ||
3 | When selecting an ARM target on Debian unstable, we get: | ||
4 | |||
5 | Compiling C++ object libcommon.fa.p/disas_libvixl_vixl_utils.cc.o | ||
6 | FAILED: libcommon.fa.p/disas_libvixl_vixl_utils.cc.o | ||
7 | c++ -Ilibcommon.fa.p -I. -I.. [...] -o libcommon.fa.p/disas_libvixl_vixl_utils.cc.o -c ../disas/libvixl/vixl/utils.cc | ||
8 | In file included from /home/philmd/qemu/disas/libvixl/vixl/utils.h:30, | ||
9 | from ../disas/libvixl/vixl/utils.cc:27: | ||
10 | /usr/include/string.h:36:43: error: missing binary operator before token "(" | ||
11 | 36 | #if defined __cplusplus && (__GNUC_PREREQ (4, 4) \ | ||
12 | | ^ | ||
13 | /usr/include/string.h:53:62: error: missing binary operator before token "(" | ||
14 | 53 | #if defined __USE_MISC || defined __USE_XOPEN || __GLIBC_USE (ISOC2X) | ||
15 | | ^ | ||
16 | /usr/include/string.h:165:21: error: missing binary operator before token "(" | ||
17 | 165 | || __GLIBC_USE (LIB_EXT2) || __GLIBC_USE (ISOC2X)) | ||
18 | | ^ | ||
19 | /usr/include/string.h:174:43: error: missing binary operator before token "(" | ||
20 | 174 | #if defined __USE_XOPEN2K8 || __GLIBC_USE (LIB_EXT2) || __GLIBC_USE (ISOC2X) | ||
21 | | ^ | ||
22 | /usr/include/string.h:492:19: error: missing binary operator before token "(" | ||
23 | 492 | #if __GNUC_PREREQ (3,4) | ||
24 | | ^ | ||
25 | |||
26 | Relevant information from the host: | ||
27 | |||
28 | $ lsb_release -d | ||
29 | Description: Debian GNU/Linux 11 (bullseye) | ||
30 | $ gcc --version | ||
31 | gcc (Debian 10.2.1-6) 10.2.1 20210110 | ||
32 | $ dpkg -S /usr/include/string.h | ||
33 | libc6-dev: /usr/include/string.h | ||
34 | $ apt-cache show libc6-dev | ||
35 | Package: libc6-dev | ||
36 | Version: 2.31-11 | ||
37 | |||
38 | Partially cherry-pick vixl commit 78973f258039f6e96 [*]: | ||
39 | |||
40 | Refactor VIXL to use `extern` block when including C header | ||
41 | that do not have a C++ counterpart. | ||
42 | |||
43 | which is similar to commit 875df03b221 ('osdep: protect qemu/osdep.h | ||
44 | with extern "C"'). | ||
45 | |||
46 | [*] https://git.linaro.org/arm/vixl.git/commit/?id=78973f258039f6e96 | ||
47 | |||
48 | Buglink: https://bugs.launchpad.net/qemu/+bug/1914870 | ||
49 | Suggested-by: Thomas Huth <thuth@redhat.com> | ||
50 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
51 | Reviewed-by: Thomas Huth <thuth@redhat.com> | ||
52 | Message-id: 20210516171023.510778-1-f4bug@amsat.org | ||
53 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
54 | --- | ||
55 | disas/libvixl/vixl/code-buffer.h | 2 +- | ||
56 | disas/libvixl/vixl/globals.h | 16 +++++++++------- | ||
57 | disas/libvixl/vixl/invalset.h | 2 +- | ||
58 | disas/libvixl/vixl/platform.h | 2 ++ | ||
59 | disas/libvixl/vixl/utils.h | 2 +- | ||
60 | disas/libvixl/vixl/utils.cc | 2 +- | ||
61 | 6 files changed, 15 insertions(+), 11 deletions(-) | ||
62 | |||
63 | diff --git a/disas/libvixl/vixl/code-buffer.h b/disas/libvixl/vixl/code-buffer.h | ||
64 | index XXXXXXX..XXXXXXX 100644 | ||
65 | --- a/disas/libvixl/vixl/code-buffer.h | ||
66 | +++ b/disas/libvixl/vixl/code-buffer.h | ||
67 | @@ -XXX,XX +XXX,XX @@ | ||
68 | #ifndef VIXL_CODE_BUFFER_H | ||
69 | #define VIXL_CODE_BUFFER_H | ||
70 | |||
71 | -#include <string.h> | ||
72 | +#include <cstring> | ||
73 | #include "vixl/globals.h" | ||
74 | |||
75 | namespace vixl { | ||
76 | diff --git a/disas/libvixl/vixl/globals.h b/disas/libvixl/vixl/globals.h | ||
77 | index XXXXXXX..XXXXXXX 100644 | ||
78 | --- a/disas/libvixl/vixl/globals.h | ||
79 | +++ b/disas/libvixl/vixl/globals.h | ||
80 | @@ -XXX,XX +XXX,XX @@ | ||
81 | #define __STDC_FORMAT_MACROS | ||
82 | #endif | ||
83 | |||
84 | -#include <stdint.h> | ||
85 | +extern "C" { | ||
86 | #include <inttypes.h> | ||
87 | - | ||
88 | -#include <assert.h> | ||
89 | -#include <stdarg.h> | ||
90 | -#include <stdio.h> | ||
91 | #include <stdint.h> | ||
92 | -#include <stdlib.h> | ||
93 | -#include <stddef.h> | ||
94 | +} | ||
95 | + | ||
96 | +#include <cassert> | ||
97 | +#include <cstdarg> | ||
98 | +#include <cstddef> | ||
99 | +#include <cstdio> | ||
100 | +#include <cstdlib> | ||
101 | + | ||
102 | #include "vixl/platform.h" | ||
103 | |||
104 | |||
105 | diff --git a/disas/libvixl/vixl/invalset.h b/disas/libvixl/vixl/invalset.h | ||
106 | index XXXXXXX..XXXXXXX 100644 | ||
107 | --- a/disas/libvixl/vixl/invalset.h | ||
108 | +++ b/disas/libvixl/vixl/invalset.h | ||
109 | @@ -XXX,XX +XXX,XX @@ | ||
110 | #ifndef VIXL_INVALSET_H_ | ||
111 | #define VIXL_INVALSET_H_ | ||
112 | |||
113 | -#include <string.h> | ||
114 | +#include <cstring> | ||
115 | |||
116 | #include <algorithm> | ||
117 | #include <vector> | ||
118 | diff --git a/disas/libvixl/vixl/platform.h b/disas/libvixl/vixl/platform.h | ||
119 | index XXXXXXX..XXXXXXX 100644 | ||
120 | --- a/disas/libvixl/vixl/platform.h | ||
121 | +++ b/disas/libvixl/vixl/platform.h | ||
122 | @@ -XXX,XX +XXX,XX @@ | ||
123 | #define PLATFORM_H | ||
124 | |||
125 | // Define platform specific functionalities. | ||
126 | +extern "C" { | ||
127 | #include <signal.h> | ||
128 | +} | ||
129 | |||
130 | namespace vixl { | ||
131 | inline void HostBreakpoint() { raise(SIGINT); } | ||
132 | diff --git a/disas/libvixl/vixl/utils.h b/disas/libvixl/vixl/utils.h | ||
133 | index XXXXXXX..XXXXXXX 100644 | ||
134 | --- a/disas/libvixl/vixl/utils.h | ||
135 | +++ b/disas/libvixl/vixl/utils.h | ||
136 | @@ -XXX,XX +XXX,XX @@ | ||
137 | #ifndef VIXL_UTILS_H | ||
138 | #define VIXL_UTILS_H | ||
139 | |||
140 | -#include <string.h> | ||
141 | #include <cmath> | ||
142 | +#include <cstring> | ||
143 | #include "vixl/globals.h" | ||
144 | #include "vixl/compiler-intrinsics.h" | ||
145 | |||
146 | diff --git a/disas/libvixl/vixl/utils.cc b/disas/libvixl/vixl/utils.cc | ||
147 | index XXXXXXX..XXXXXXX 100644 | ||
148 | --- a/disas/libvixl/vixl/utils.cc | ||
149 | +++ b/disas/libvixl/vixl/utils.cc | ||
150 | @@ -XXX,XX +XXX,XX @@ | ||
151 | // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | ||
152 | |||
153 | #include "vixl/utils.h" | ||
154 | -#include <stdio.h> | ||
155 | +#include <cstdio> | ||
156 | |||
157 | namespace vixl { | ||
158 | |||
159 | -- | ||
160 | 2.20.1 | ||
161 | |||
162 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | For MUL, we can rely on generic support. For SMULH and UMULH, | ||
4 | create some trivial helpers. For PMUL, back in a21bb78e5817, | ||
5 | we organized helper_gvec_pmul_b in preparation for this use. | ||
6 | |||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-3-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper.h | 10 ++++ | ||
13 | target/arm/sve.decode | 10 ++++ | ||
14 | target/arm/translate-sve.c | 50 ++++++++++++++++++++ | ||
15 | target/arm/vec_helper.c | 96 ++++++++++++++++++++++++++++++++++++++ | ||
16 | 4 files changed, 166 insertions(+) | ||
17 | |||
18 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/target/arm/helper.h | ||
21 | +++ b/target/arm/helper.h | ||
22 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(gvec_cgt0_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_3(gvec_cge0_b, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | DEF_HELPER_FLAGS_3(gvec_cge0_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | |||
26 | +DEF_HELPER_FLAGS_4(gvec_smulh_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(gvec_smulh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(gvec_smulh_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_4(gvec_smulh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_4(gvec_umulh_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(gvec_umulh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(gvec_umulh_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_4(gvec_umulh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
35 | + | ||
36 | DEF_HELPER_FLAGS_4(gvec_sshl_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | DEF_HELPER_FLAGS_4(gvec_sshl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
38 | DEF_HELPER_FLAGS_4(gvec_ushl_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ ST1_zprz 1110010 .. 00 ..... 100 ... ..... ..... \ | ||
44 | @rprr_scatter_store xs=0 esz=3 scale=0 | ||
45 | ST1_zprz 1110010 .. 00 ..... 110 ... ..... ..... \ | ||
46 | @rprr_scatter_store xs=1 esz=3 scale=0 | ||
47 | + | ||
48 | +#### SVE2 Support | ||
49 | + | ||
50 | +### SVE2 Integer Multiply - Unpredicated | ||
51 | + | ||
52 | +# SVE2 integer multiply vectors (unpredicated) | ||
53 | +MUL_zzz 00000100 .. 1 ..... 0110 00 ..... ..... @rd_rn_rm | ||
54 | +SMULH_zzz 00000100 .. 1 ..... 0110 10 ..... ..... @rd_rn_rm | ||
55 | +UMULH_zzz 00000100 .. 1 ..... 0110 11 ..... ..... @rd_rn_rm | ||
56 | +PMUL_zzz 00000100 00 1 ..... 0110 01 ..... ..... @rd_rn_rm_e0 | ||
57 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | ||
59 | --- a/target/arm/translate-sve.c | ||
60 | +++ b/target/arm/translate-sve.c | ||
61 | @@ -XXX,XX +XXX,XX @@ static bool trans_MOVPRFX_z(DisasContext *s, arg_rpr_esz *a) | ||
62 | { | ||
63 | return do_movz_zpz(s, a->rd, a->rn, a->pg, a->esz, false); | ||
64 | } | ||
65 | + | ||
66 | +/* | ||
67 | + * SVE2 Integer Multiply - Unpredicated | ||
68 | + */ | ||
69 | + | ||
70 | +static bool trans_MUL_zzz(DisasContext *s, arg_rrr_esz *a) | ||
71 | +{ | ||
72 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
73 | + return false; | ||
74 | + } | ||
75 | + if (sve_access_check(s)) { | ||
76 | + gen_gvec_fn_zzz(s, tcg_gen_gvec_mul, a->esz, a->rd, a->rn, a->rm); | ||
77 | + } | ||
78 | + return true; | ||
79 | +} | ||
80 | + | ||
81 | +static bool do_sve2_zzz_ool(DisasContext *s, arg_rrr_esz *a, | ||
82 | + gen_helper_gvec_3 *fn) | ||
83 | +{ | ||
84 | + if (fn == NULL || !dc_isar_feature(aa64_sve2, s)) { | ||
85 | + return false; | ||
86 | + } | ||
87 | + if (sve_access_check(s)) { | ||
88 | + gen_gvec_ool_zzz(s, fn, a->rd, a->rn, a->rm, 0); | ||
89 | + } | ||
90 | + return true; | ||
91 | +} | ||
92 | + | ||
93 | +static bool trans_SMULH_zzz(DisasContext *s, arg_rrr_esz *a) | ||
94 | +{ | ||
95 | + static gen_helper_gvec_3 * const fns[4] = { | ||
96 | + gen_helper_gvec_smulh_b, gen_helper_gvec_smulh_h, | ||
97 | + gen_helper_gvec_smulh_s, gen_helper_gvec_smulh_d, | ||
98 | + }; | ||
99 | + return do_sve2_zzz_ool(s, a, fns[a->esz]); | ||
100 | +} | ||
101 | + | ||
102 | +static bool trans_UMULH_zzz(DisasContext *s, arg_rrr_esz *a) | ||
103 | +{ | ||
104 | + static gen_helper_gvec_3 * const fns[4] = { | ||
105 | + gen_helper_gvec_umulh_b, gen_helper_gvec_umulh_h, | ||
106 | + gen_helper_gvec_umulh_s, gen_helper_gvec_umulh_d, | ||
107 | + }; | ||
108 | + return do_sve2_zzz_ool(s, a, fns[a->esz]); | ||
109 | +} | ||
110 | + | ||
111 | +static bool trans_PMUL_zzz(DisasContext *s, arg_rrr_esz *a) | ||
112 | +{ | ||
113 | + return do_sve2_zzz_ool(s, a, gen_helper_gvec_pmul_b); | ||
114 | +} | ||
115 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
116 | index XXXXXXX..XXXXXXX 100644 | ||
117 | --- a/target/arm/vec_helper.c | ||
118 | +++ b/target/arm/vec_helper.c | ||
119 | @@ -XXX,XX +XXX,XX @@ void HELPER(simd_tblx)(void *vd, void *vm, void *venv, uint32_t desc) | ||
120 | clear_tail(vd, oprsz, simd_maxsz(desc)); | ||
121 | } | ||
122 | #endif | ||
123 | + | ||
124 | +/* | ||
125 | + * NxN -> N highpart multiply | ||
126 | + * | ||
127 | + * TODO: expose this as a generic vector operation. | ||
128 | + */ | ||
129 | + | ||
130 | +void HELPER(gvec_smulh_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
131 | +{ | ||
132 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
133 | + int8_t *d = vd, *n = vn, *m = vm; | ||
134 | + | ||
135 | + for (i = 0; i < opr_sz; ++i) { | ||
136 | + d[i] = ((int32_t)n[i] * m[i]) >> 8; | ||
137 | + } | ||
138 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
139 | +} | ||
140 | + | ||
141 | +void HELPER(gvec_smulh_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
142 | +{ | ||
143 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
144 | + int16_t *d = vd, *n = vn, *m = vm; | ||
145 | + | ||
146 | + for (i = 0; i < opr_sz / 2; ++i) { | ||
147 | + d[i] = ((int32_t)n[i] * m[i]) >> 16; | ||
148 | + } | ||
149 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
150 | +} | ||
151 | + | ||
152 | +void HELPER(gvec_smulh_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
153 | +{ | ||
154 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
155 | + int32_t *d = vd, *n = vn, *m = vm; | ||
156 | + | ||
157 | + for (i = 0; i < opr_sz / 4; ++i) { | ||
158 | + d[i] = ((int64_t)n[i] * m[i]) >> 32; | ||
159 | + } | ||
160 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
161 | +} | ||
162 | + | ||
163 | +void HELPER(gvec_smulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
164 | +{ | ||
165 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
166 | + uint64_t *d = vd, *n = vn, *m = vm; | ||
167 | + uint64_t discard; | ||
168 | + | ||
169 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
170 | + muls64(&discard, &d[i], n[i], m[i]); | ||
171 | + } | ||
172 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
173 | +} | ||
174 | + | ||
175 | +void HELPER(gvec_umulh_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
176 | +{ | ||
177 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
178 | + uint8_t *d = vd, *n = vn, *m = vm; | ||
179 | + | ||
180 | + for (i = 0; i < opr_sz; ++i) { | ||
181 | + d[i] = ((uint32_t)n[i] * m[i]) >> 8; | ||
182 | + } | ||
183 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
184 | +} | ||
185 | + | ||
186 | +void HELPER(gvec_umulh_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
187 | +{ | ||
188 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
189 | + uint16_t *d = vd, *n = vn, *m = vm; | ||
190 | + | ||
191 | + for (i = 0; i < opr_sz / 2; ++i) { | ||
192 | + d[i] = ((uint32_t)n[i] * m[i]) >> 16; | ||
193 | + } | ||
194 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
195 | +} | ||
196 | + | ||
197 | +void HELPER(gvec_umulh_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
198 | +{ | ||
199 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
200 | + uint32_t *d = vd, *n = vn, *m = vm; | ||
201 | + | ||
202 | + for (i = 0; i < opr_sz / 4; ++i) { | ||
203 | + d[i] = ((uint64_t)n[i] * m[i]) >> 32; | ||
204 | + } | ||
205 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
206 | +} | ||
207 | + | ||
208 | +void HELPER(gvec_umulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
209 | +{ | ||
210 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
211 | + uint64_t *d = vd, *n = vn, *m = vm; | ||
212 | + uint64_t discard; | ||
213 | + | ||
214 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
215 | + mulu64(&discard, &d[i], n[i], m[i]); | ||
216 | + } | ||
217 | + clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
218 | +} | ||
219 | -- | ||
220 | 2.20.1 | ||
221 | |||
222 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-4-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 14 ++++++++++++ | ||
9 | target/arm/sve.decode | 5 +++++ | ||
10 | target/arm/sve_helper.c | 44 ++++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 39 +++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 102 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve_umulh_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(sve_umulh_zpzz_d, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_5(sve2_sadalp_zpzz_h, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_sadalp_zpzz_s, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_sadalp_zpzz_d, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | + | ||
29 | +DEF_HELPER_FLAGS_5(sve2_uadalp_zpzz_h, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_5(sve2_uadalp_zpzz_s, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_uadalp_zpzz_d, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | + | ||
36 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_s, TCG_CALL_NO_RWG, | ||
37 | void, ptr, ptr, ptr, ptr, i32) | ||
38 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_d, TCG_CALL_NO_RWG, | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ MUL_zzz 00000100 .. 1 ..... 0110 00 ..... ..... @rd_rn_rm | ||
44 | SMULH_zzz 00000100 .. 1 ..... 0110 10 ..... ..... @rd_rn_rm | ||
45 | UMULH_zzz 00000100 .. 1 ..... 0110 11 ..... ..... @rd_rn_rm | ||
46 | PMUL_zzz 00000100 00 1 ..... 0110 01 ..... ..... @rd_rn_rm_e0 | ||
47 | + | ||
48 | +### SVE2 Integer - Predicated | ||
49 | + | ||
50 | +SADALP_zpzz 01000100 .. 000 100 101 ... ..... ..... @rdm_pg_rn | ||
51 | +UADALP_zpzz 01000100 .. 000 101 101 ... ..... ..... @rdm_pg_rn | ||
52 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/target/arm/sve_helper.c | ||
55 | +++ b/target/arm/sve_helper.c | ||
56 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ_D(sve_asr_zpzz_d, int64_t, DO_ASR) | ||
57 | DO_ZPZZ_D(sve_lsr_zpzz_d, uint64_t, DO_LSR) | ||
58 | DO_ZPZZ_D(sve_lsl_zpzz_d, uint64_t, DO_LSL) | ||
59 | |||
60 | +static inline uint16_t do_sadalp_h(int16_t n, int16_t m) | ||
61 | +{ | ||
62 | + int8_t n1 = n, n2 = n >> 8; | ||
63 | + return m + n1 + n2; | ||
64 | +} | ||
65 | + | ||
66 | +static inline uint32_t do_sadalp_s(int32_t n, int32_t m) | ||
67 | +{ | ||
68 | + int16_t n1 = n, n2 = n >> 16; | ||
69 | + return m + n1 + n2; | ||
70 | +} | ||
71 | + | ||
72 | +static inline uint64_t do_sadalp_d(int64_t n, int64_t m) | ||
73 | +{ | ||
74 | + int32_t n1 = n, n2 = n >> 32; | ||
75 | + return m + n1 + n2; | ||
76 | +} | ||
77 | + | ||
78 | +DO_ZPZZ(sve2_sadalp_zpzz_h, int16_t, H1_2, do_sadalp_h) | ||
79 | +DO_ZPZZ(sve2_sadalp_zpzz_s, int32_t, H1_4, do_sadalp_s) | ||
80 | +DO_ZPZZ_D(sve2_sadalp_zpzz_d, int64_t, do_sadalp_d) | ||
81 | + | ||
82 | +static inline uint16_t do_uadalp_h(uint16_t n, uint16_t m) | ||
83 | +{ | ||
84 | + uint8_t n1 = n, n2 = n >> 8; | ||
85 | + return m + n1 + n2; | ||
86 | +} | ||
87 | + | ||
88 | +static inline uint32_t do_uadalp_s(uint32_t n, uint32_t m) | ||
89 | +{ | ||
90 | + uint16_t n1 = n, n2 = n >> 16; | ||
91 | + return m + n1 + n2; | ||
92 | +} | ||
93 | + | ||
94 | +static inline uint64_t do_uadalp_d(uint64_t n, uint64_t m) | ||
95 | +{ | ||
96 | + uint32_t n1 = n, n2 = n >> 32; | ||
97 | + return m + n1 + n2; | ||
98 | +} | ||
99 | + | ||
100 | +DO_ZPZZ(sve2_uadalp_zpzz_h, uint16_t, H1_2, do_uadalp_h) | ||
101 | +DO_ZPZZ(sve2_uadalp_zpzz_s, uint32_t, H1_4, do_uadalp_s) | ||
102 | +DO_ZPZZ_D(sve2_uadalp_zpzz_d, uint64_t, do_uadalp_d) | ||
103 | + | ||
104 | #undef DO_ZPZZ | ||
105 | #undef DO_ZPZZ_D | ||
106 | |||
107 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
108 | index XXXXXXX..XXXXXXX 100644 | ||
109 | --- a/target/arm/translate-sve.c | ||
110 | +++ b/target/arm/translate-sve.c | ||
111 | @@ -XXX,XX +XXX,XX @@ static bool trans_PMUL_zzz(DisasContext *s, arg_rrr_esz *a) | ||
112 | { | ||
113 | return do_sve2_zzz_ool(s, a, gen_helper_gvec_pmul_b); | ||
114 | } | ||
115 | + | ||
116 | +/* | ||
117 | + * SVE2 Integer - Predicated | ||
118 | + */ | ||
119 | + | ||
120 | +static bool do_sve2_zpzz_ool(DisasContext *s, arg_rprr_esz *a, | ||
121 | + gen_helper_gvec_4 *fn) | ||
122 | +{ | ||
123 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
124 | + return false; | ||
125 | + } | ||
126 | + return do_zpzz_ool(s, a, fn); | ||
127 | +} | ||
128 | + | ||
129 | +static bool trans_SADALP_zpzz(DisasContext *s, arg_rprr_esz *a) | ||
130 | +{ | ||
131 | + static gen_helper_gvec_4 * const fns[3] = { | ||
132 | + gen_helper_sve2_sadalp_zpzz_h, | ||
133 | + gen_helper_sve2_sadalp_zpzz_s, | ||
134 | + gen_helper_sve2_sadalp_zpzz_d, | ||
135 | + }; | ||
136 | + if (a->esz == 0) { | ||
137 | + return false; | ||
138 | + } | ||
139 | + return do_sve2_zpzz_ool(s, a, fns[a->esz - 1]); | ||
140 | +} | ||
141 | + | ||
142 | +static bool trans_UADALP_zpzz(DisasContext *s, arg_rprr_esz *a) | ||
143 | +{ | ||
144 | + static gen_helper_gvec_4 * const fns[3] = { | ||
145 | + gen_helper_sve2_uadalp_zpzz_h, | ||
146 | + gen_helper_sve2_uadalp_zpzz_s, | ||
147 | + gen_helper_sve2_uadalp_zpzz_d, | ||
148 | + }; | ||
149 | + if (a->esz == 0) { | ||
150 | + return false; | ||
151 | + } | ||
152 | + return do_sve2_zpzz_ool(s, a, fns[a->esz - 1]); | ||
153 | +} | ||
154 | -- | ||
155 | 2.20.1 | ||
156 | |||
157 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-5-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 13 +++++++++++ | ||
9 | target/arm/sve.decode | 7 ++++++ | ||
10 | target/arm/sve_helper.c | 21 +++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 47 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 88 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_rbit_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_4(sve_rbit_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve_rbit_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_4(sve2_sqabs_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_4(sve2_sqabs_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_sqabs_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_sqabs_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | + | ||
27 | +DEF_HELPER_FLAGS_4(sve2_sqneg_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_sqneg_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_4(sve2_sqneg_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_sqneg_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | + | ||
32 | +DEF_HELPER_FLAGS_4(sve2_urecpe_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(sve2_ursqrte_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
34 | + | ||
35 | DEF_HELPER_FLAGS_5(sve_splice, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
36 | |||
37 | DEF_HELPER_FLAGS_5(sve_cmpeq_ppzz_b, TCG_CALL_NO_RWG, | ||
38 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
39 | index XXXXXXX..XXXXXXX 100644 | ||
40 | --- a/target/arm/sve.decode | ||
41 | +++ b/target/arm/sve.decode | ||
42 | @@ -XXX,XX +XXX,XX @@ PMUL_zzz 00000100 00 1 ..... 0110 01 ..... ..... @rd_rn_rm_e0 | ||
43 | |||
44 | SADALP_zpzz 01000100 .. 000 100 101 ... ..... ..... @rdm_pg_rn | ||
45 | UADALP_zpzz 01000100 .. 000 101 101 ... ..... ..... @rdm_pg_rn | ||
46 | + | ||
47 | +### SVE2 integer unary operations (predicated) | ||
48 | + | ||
49 | +URECPE 01000100 .. 000 000 101 ... ..... ..... @rd_pg_rn | ||
50 | +URSQRTE 01000100 .. 000 001 101 ... ..... ..... @rd_pg_rn | ||
51 | +SQABS 01000100 .. 001 000 101 ... ..... ..... @rd_pg_rn | ||
52 | +SQNEG 01000100 .. 001 001 101 ... ..... ..... @rd_pg_rn | ||
53 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
54 | index XXXXXXX..XXXXXXX 100644 | ||
55 | --- a/target/arm/sve_helper.c | ||
56 | +++ b/target/arm/sve_helper.c | ||
57 | @@ -XXX,XX +XXX,XX @@ DO_ZPZ(sve_rbit_h, uint16_t, H1_2, revbit16) | ||
58 | DO_ZPZ(sve_rbit_s, uint32_t, H1_4, revbit32) | ||
59 | DO_ZPZ_D(sve_rbit_d, uint64_t, revbit64) | ||
60 | |||
61 | +#define DO_SQABS(X) \ | ||
62 | + ({ __typeof(X) x_ = (X), min_ = 1ull << (sizeof(X) * 8 - 1); \ | ||
63 | + x_ >= 0 ? x_ : x_ == min_ ? -min_ - 1 : -x_; }) | ||
64 | + | ||
65 | +DO_ZPZ(sve2_sqabs_b, int8_t, H1, DO_SQABS) | ||
66 | +DO_ZPZ(sve2_sqabs_h, int16_t, H1_2, DO_SQABS) | ||
67 | +DO_ZPZ(sve2_sqabs_s, int32_t, H1_4, DO_SQABS) | ||
68 | +DO_ZPZ_D(sve2_sqabs_d, int64_t, DO_SQABS) | ||
69 | + | ||
70 | +#define DO_SQNEG(X) \ | ||
71 | + ({ __typeof(X) x_ = (X), min_ = 1ull << (sizeof(X) * 8 - 1); \ | ||
72 | + x_ == min_ ? -min_ - 1 : -x_; }) | ||
73 | + | ||
74 | +DO_ZPZ(sve2_sqneg_b, uint8_t, H1, DO_SQNEG) | ||
75 | +DO_ZPZ(sve2_sqneg_h, uint16_t, H1_2, DO_SQNEG) | ||
76 | +DO_ZPZ(sve2_sqneg_s, uint32_t, H1_4, DO_SQNEG) | ||
77 | +DO_ZPZ_D(sve2_sqneg_d, uint64_t, DO_SQNEG) | ||
78 | + | ||
79 | +DO_ZPZ(sve2_urecpe_s, uint32_t, H1_4, helper_recpe_u32) | ||
80 | +DO_ZPZ(sve2_ursqrte_s, uint32_t, H1_4, helper_rsqrte_u32) | ||
81 | + | ||
82 | /* Three-operand expander, unpredicated, in which the third operand is "wide". | ||
83 | */ | ||
84 | #define DO_ZZW(NAME, TYPE, TYPEW, H, OP) \ | ||
85 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
86 | index XXXXXXX..XXXXXXX 100644 | ||
87 | --- a/target/arm/translate-sve.c | ||
88 | +++ b/target/arm/translate-sve.c | ||
89 | @@ -XXX,XX +XXX,XX @@ static bool trans_UADALP_zpzz(DisasContext *s, arg_rprr_esz *a) | ||
90 | } | ||
91 | return do_sve2_zpzz_ool(s, a, fns[a->esz - 1]); | ||
92 | } | ||
93 | + | ||
94 | +/* | ||
95 | + * SVE2 integer unary operations (predicated) | ||
96 | + */ | ||
97 | + | ||
98 | +static bool do_sve2_zpz_ool(DisasContext *s, arg_rpr_esz *a, | ||
99 | + gen_helper_gvec_3 *fn) | ||
100 | +{ | ||
101 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
102 | + return false; | ||
103 | + } | ||
104 | + return do_zpz_ool(s, a, fn); | ||
105 | +} | ||
106 | + | ||
107 | +static bool trans_URECPE(DisasContext *s, arg_rpr_esz *a) | ||
108 | +{ | ||
109 | + if (a->esz != 2) { | ||
110 | + return false; | ||
111 | + } | ||
112 | + return do_sve2_zpz_ool(s, a, gen_helper_sve2_urecpe_s); | ||
113 | +} | ||
114 | + | ||
115 | +static bool trans_URSQRTE(DisasContext *s, arg_rpr_esz *a) | ||
116 | +{ | ||
117 | + if (a->esz != 2) { | ||
118 | + return false; | ||
119 | + } | ||
120 | + return do_sve2_zpz_ool(s, a, gen_helper_sve2_ursqrte_s); | ||
121 | +} | ||
122 | + | ||
123 | +static bool trans_SQABS(DisasContext *s, arg_rpr_esz *a) | ||
124 | +{ | ||
125 | + static gen_helper_gvec_3 * const fns[4] = { | ||
126 | + gen_helper_sve2_sqabs_b, gen_helper_sve2_sqabs_h, | ||
127 | + gen_helper_sve2_sqabs_s, gen_helper_sve2_sqabs_d, | ||
128 | + }; | ||
129 | + return do_sve2_zpz_ool(s, a, fns[a->esz]); | ||
130 | +} | ||
131 | + | ||
132 | +static bool trans_SQNEG(DisasContext *s, arg_rpr_esz *a) | ||
133 | +{ | ||
134 | + static gen_helper_gvec_3 * const fns[4] = { | ||
135 | + gen_helper_sve2_sqneg_b, gen_helper_sve2_sqneg_h, | ||
136 | + gen_helper_sve2_sqneg_s, gen_helper_sve2_sqneg_d, | ||
137 | + }; | ||
138 | + return do_sve2_zpz_ool(s, a, fns[a->esz]); | ||
139 | +} | ||
140 | -- | ||
141 | 2.20.1 | ||
142 | |||
143 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Split these operations out into a header that can be shared | ||
4 | between neon and sve. The "sat" pointer acts both as a boolean | ||
5 | for control of saturating behavior and controls the difference | ||
6 | in behavior between neon and sve -- QC bit or no QC bit. | ||
7 | |||
8 | Widen the shift operand in the new helpers, as the SVE2 insns treat | ||
9 | the whole input element as significant. For the neon uses, truncate | ||
10 | the shift to int8_t while passing the parameter. | ||
11 | |||
12 | Implement right-shift rounding as | ||
13 | |||
14 | tmp = src >> (shift - 1); | ||
15 | dst = (tmp >> 1) + (tmp & 1); | ||
16 | |||
17 | This is the same number of instructions as the current | ||
18 | |||
19 | tmp = 1 << (shift - 1); | ||
20 | dst = (src + tmp) >> shift; | ||
21 | |||
22 | without any possibility of intermediate overflow. | ||
23 | |||
24 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
25 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
26 | Message-id: 20210525010358.152808-6-richard.henderson@linaro.org | ||
27 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
28 | --- | ||
29 | target/arm/vec_internal.h | 138 +++++++++++ | ||
30 | target/arm/neon_helper.c | 507 +++++++------------------------------- | ||
31 | 2 files changed, 221 insertions(+), 424 deletions(-) | ||
32 | |||
33 | diff --git a/target/arm/vec_internal.h b/target/arm/vec_internal.h | ||
34 | index XXXXXXX..XXXXXXX 100644 | ||
35 | --- a/target/arm/vec_internal.h | ||
36 | +++ b/target/arm/vec_internal.h | ||
37 | @@ -XXX,XX +XXX,XX @@ static inline void clear_tail(void *vd, uintptr_t opr_sz, uintptr_t max_sz) | ||
38 | } | ||
39 | } | ||
40 | |||
41 | +static inline int32_t do_sqrshl_bhs(int32_t src, int32_t shift, int bits, | ||
42 | + bool round, uint32_t *sat) | ||
43 | +{ | ||
44 | + if (shift <= -bits) { | ||
45 | + /* Rounding the sign bit always produces 0. */ | ||
46 | + if (round) { | ||
47 | + return 0; | ||
48 | + } | ||
49 | + return src >> 31; | ||
50 | + } else if (shift < 0) { | ||
51 | + if (round) { | ||
52 | + src >>= -shift - 1; | ||
53 | + return (src >> 1) + (src & 1); | ||
54 | + } | ||
55 | + return src >> -shift; | ||
56 | + } else if (shift < bits) { | ||
57 | + int32_t val = src << shift; | ||
58 | + if (bits == 32) { | ||
59 | + if (!sat || val >> shift == src) { | ||
60 | + return val; | ||
61 | + } | ||
62 | + } else { | ||
63 | + int32_t extval = sextract32(val, 0, bits); | ||
64 | + if (!sat || val == extval) { | ||
65 | + return extval; | ||
66 | + } | ||
67 | + } | ||
68 | + } else if (!sat || src == 0) { | ||
69 | + return 0; | ||
70 | + } | ||
71 | + | ||
72 | + *sat = 1; | ||
73 | + return (1u << (bits - 1)) - (src >= 0); | ||
74 | +} | ||
75 | + | ||
76 | +static inline uint32_t do_uqrshl_bhs(uint32_t src, int32_t shift, int bits, | ||
77 | + bool round, uint32_t *sat) | ||
78 | +{ | ||
79 | + if (shift <= -(bits + round)) { | ||
80 | + return 0; | ||
81 | + } else if (shift < 0) { | ||
82 | + if (round) { | ||
83 | + src >>= -shift - 1; | ||
84 | + return (src >> 1) + (src & 1); | ||
85 | + } | ||
86 | + return src >> -shift; | ||
87 | + } else if (shift < bits) { | ||
88 | + uint32_t val = src << shift; | ||
89 | + if (bits == 32) { | ||
90 | + if (!sat || val >> shift == src) { | ||
91 | + return val; | ||
92 | + } | ||
93 | + } else { | ||
94 | + uint32_t extval = extract32(val, 0, bits); | ||
95 | + if (!sat || val == extval) { | ||
96 | + return extval; | ||
97 | + } | ||
98 | + } | ||
99 | + } else if (!sat || src == 0) { | ||
100 | + return 0; | ||
101 | + } | ||
102 | + | ||
103 | + *sat = 1; | ||
104 | + return MAKE_64BIT_MASK(0, bits); | ||
105 | +} | ||
106 | + | ||
107 | +static inline int32_t do_suqrshl_bhs(int32_t src, int32_t shift, int bits, | ||
108 | + bool round, uint32_t *sat) | ||
109 | +{ | ||
110 | + if (sat && src < 0) { | ||
111 | + *sat = 1; | ||
112 | + return 0; | ||
113 | + } | ||
114 | + return do_uqrshl_bhs(src, shift, bits, round, sat); | ||
115 | +} | ||
116 | + | ||
117 | +static inline int64_t do_sqrshl_d(int64_t src, int64_t shift, | ||
118 | + bool round, uint32_t *sat) | ||
119 | +{ | ||
120 | + if (shift <= -64) { | ||
121 | + /* Rounding the sign bit always produces 0. */ | ||
122 | + if (round) { | ||
123 | + return 0; | ||
124 | + } | ||
125 | + return src >> 63; | ||
126 | + } else if (shift < 0) { | ||
127 | + if (round) { | ||
128 | + src >>= -shift - 1; | ||
129 | + return (src >> 1) + (src & 1); | ||
130 | + } | ||
131 | + return src >> -shift; | ||
132 | + } else if (shift < 64) { | ||
133 | + int64_t val = src << shift; | ||
134 | + if (!sat || val >> shift == src) { | ||
135 | + return val; | ||
136 | + } | ||
137 | + } else if (!sat || src == 0) { | ||
138 | + return 0; | ||
139 | + } | ||
140 | + | ||
141 | + *sat = 1; | ||
142 | + return src < 0 ? INT64_MIN : INT64_MAX; | ||
143 | +} | ||
144 | + | ||
145 | +static inline uint64_t do_uqrshl_d(uint64_t src, int64_t shift, | ||
146 | + bool round, uint32_t *sat) | ||
147 | +{ | ||
148 | + if (shift <= -(64 + round)) { | ||
149 | + return 0; | ||
150 | + } else if (shift < 0) { | ||
151 | + if (round) { | ||
152 | + src >>= -shift - 1; | ||
153 | + return (src >> 1) + (src & 1); | ||
154 | + } | ||
155 | + return src >> -shift; | ||
156 | + } else if (shift < 64) { | ||
157 | + uint64_t val = src << shift; | ||
158 | + if (!sat || val >> shift == src) { | ||
159 | + return val; | ||
160 | + } | ||
161 | + } else if (!sat || src == 0) { | ||
162 | + return 0; | ||
163 | + } | ||
164 | + | ||
165 | + *sat = 1; | ||
166 | + return UINT64_MAX; | ||
167 | +} | ||
168 | + | ||
169 | +static inline int64_t do_suqrshl_d(int64_t src, int64_t shift, | ||
170 | + bool round, uint32_t *sat) | ||
171 | +{ | ||
172 | + if (sat && src < 0) { | ||
173 | + *sat = 1; | ||
174 | + return 0; | ||
175 | + } | ||
176 | + return do_uqrshl_d(src, shift, round, sat); | ||
177 | +} | ||
178 | + | ||
179 | #endif /* TARGET_ARM_VEC_INTERNALS_H */ | ||
180 | diff --git a/target/arm/neon_helper.c b/target/arm/neon_helper.c | ||
181 | index XXXXXXX..XXXXXXX 100644 | ||
182 | --- a/target/arm/neon_helper.c | ||
183 | +++ b/target/arm/neon_helper.c | ||
184 | @@ -XXX,XX +XXX,XX @@ | ||
185 | #include "cpu.h" | ||
186 | #include "exec/helper-proto.h" | ||
187 | #include "fpu/softfloat.h" | ||
188 | +#include "vec_internal.h" | ||
189 | |||
190 | #define SIGNBIT (uint32_t)0x80000000 | ||
191 | #define SIGNBIT64 ((uint64_t)1 << 63) | ||
192 | @@ -XXX,XX +XXX,XX @@ NEON_POP(pmax_s16, neon_s16, 2) | ||
193 | NEON_POP(pmax_u16, neon_u16, 2) | ||
194 | #undef NEON_FN | ||
195 | |||
196 | -#define NEON_FN(dest, src1, src2) do { \ | ||
197 | - int8_t tmp; \ | ||
198 | - tmp = (int8_t)src2; \ | ||
199 | - if (tmp >= (ssize_t)sizeof(src1) * 8 || \ | ||
200 | - tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
201 | - dest = 0; \ | ||
202 | - } else if (tmp < 0) { \ | ||
203 | - dest = src1 >> -tmp; \ | ||
204 | - } else { \ | ||
205 | - dest = src1 << tmp; \ | ||
206 | - }} while (0) | ||
207 | +#define NEON_FN(dest, src1, src2) \ | ||
208 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 16, false, NULL)) | ||
209 | NEON_VOP(shl_u16, neon_u16, 2) | ||
210 | #undef NEON_FN | ||
211 | |||
212 | -#define NEON_FN(dest, src1, src2) do { \ | ||
213 | - int8_t tmp; \ | ||
214 | - tmp = (int8_t)src2; \ | ||
215 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
216 | - dest = 0; \ | ||
217 | - } else if (tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
218 | - dest = src1 >> (sizeof(src1) * 8 - 1); \ | ||
219 | - } else if (tmp < 0) { \ | ||
220 | - dest = src1 >> -tmp; \ | ||
221 | - } else { \ | ||
222 | - dest = src1 << tmp; \ | ||
223 | - }} while (0) | ||
224 | +#define NEON_FN(dest, src1, src2) \ | ||
225 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 16, false, NULL)) | ||
226 | NEON_VOP(shl_s16, neon_s16, 2) | ||
227 | #undef NEON_FN | ||
228 | |||
229 | -#define NEON_FN(dest, src1, src2) do { \ | ||
230 | - int8_t tmp; \ | ||
231 | - tmp = (int8_t)src2; \ | ||
232 | - if ((tmp >= (ssize_t)sizeof(src1) * 8) \ | ||
233 | - || (tmp <= -(ssize_t)sizeof(src1) * 8)) { \ | ||
234 | - dest = 0; \ | ||
235 | - } else if (tmp < 0) { \ | ||
236 | - dest = (src1 + (1 << (-1 - tmp))) >> -tmp; \ | ||
237 | - } else { \ | ||
238 | - dest = src1 << tmp; \ | ||
239 | - }} while (0) | ||
240 | +#define NEON_FN(dest, src1, src2) \ | ||
241 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 8, true, NULL)) | ||
242 | NEON_VOP(rshl_s8, neon_s8, 4) | ||
243 | +#undef NEON_FN | ||
244 | + | ||
245 | +#define NEON_FN(dest, src1, src2) \ | ||
246 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 16, true, NULL)) | ||
247 | NEON_VOP(rshl_s16, neon_s16, 2) | ||
248 | #undef NEON_FN | ||
249 | |||
250 | -/* The addition of the rounding constant may overflow, so we use an | ||
251 | - * intermediate 64 bit accumulator. */ | ||
252 | -uint32_t HELPER(neon_rshl_s32)(uint32_t valop, uint32_t shiftop) | ||
253 | +uint32_t HELPER(neon_rshl_s32)(uint32_t val, uint32_t shift) | ||
254 | { | ||
255 | - int32_t dest; | ||
256 | - int32_t val = (int32_t)valop; | ||
257 | - int8_t shift = (int8_t)shiftop; | ||
258 | - if ((shift >= 32) || (shift <= -32)) { | ||
259 | - dest = 0; | ||
260 | - } else if (shift < 0) { | ||
261 | - int64_t big_dest = ((int64_t)val + (1 << (-1 - shift))); | ||
262 | - dest = big_dest >> -shift; | ||
263 | - } else { | ||
264 | - dest = val << shift; | ||
265 | - } | ||
266 | - return dest; | ||
267 | + return do_sqrshl_bhs(val, (int8_t)shift, 32, true, NULL); | ||
268 | } | ||
269 | |||
270 | -/* Handling addition overflow with 64 bit input values is more | ||
271 | - * tricky than with 32 bit values. */ | ||
272 | -uint64_t HELPER(neon_rshl_s64)(uint64_t valop, uint64_t shiftop) | ||
273 | +uint64_t HELPER(neon_rshl_s64)(uint64_t val, uint64_t shift) | ||
274 | { | ||
275 | - int8_t shift = (int8_t)shiftop; | ||
276 | - int64_t val = valop; | ||
277 | - if ((shift >= 64) || (shift <= -64)) { | ||
278 | - val = 0; | ||
279 | - } else if (shift < 0) { | ||
280 | - val >>= (-shift - 1); | ||
281 | - if (val == INT64_MAX) { | ||
282 | - /* In this case, it means that the rounding constant is 1, | ||
283 | - * and the addition would overflow. Return the actual | ||
284 | - * result directly. */ | ||
285 | - val = 0x4000000000000000LL; | ||
286 | - } else { | ||
287 | - val++; | ||
288 | - val >>= 1; | ||
289 | - } | ||
290 | - } else { | ||
291 | - val <<= shift; | ||
292 | - } | ||
293 | - return val; | ||
294 | + return do_sqrshl_d(val, (int8_t)shift, true, NULL); | ||
295 | } | ||
296 | |||
297 | -#define NEON_FN(dest, src1, src2) do { \ | ||
298 | - int8_t tmp; \ | ||
299 | - tmp = (int8_t)src2; \ | ||
300 | - if (tmp >= (ssize_t)sizeof(src1) * 8 || \ | ||
301 | - tmp < -(ssize_t)sizeof(src1) * 8) { \ | ||
302 | - dest = 0; \ | ||
303 | - } else if (tmp == -(ssize_t)sizeof(src1) * 8) { \ | ||
304 | - dest = src1 >> (-tmp - 1); \ | ||
305 | - } else if (tmp < 0) { \ | ||
306 | - dest = (src1 + (1 << (-1 - tmp))) >> -tmp; \ | ||
307 | - } else { \ | ||
308 | - dest = src1 << tmp; \ | ||
309 | - }} while (0) | ||
310 | +#define NEON_FN(dest, src1, src2) \ | ||
311 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 8, true, NULL)) | ||
312 | NEON_VOP(rshl_u8, neon_u8, 4) | ||
313 | +#undef NEON_FN | ||
314 | + | ||
315 | +#define NEON_FN(dest, src1, src2) \ | ||
316 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 16, true, NULL)) | ||
317 | NEON_VOP(rshl_u16, neon_u16, 2) | ||
318 | #undef NEON_FN | ||
319 | |||
320 | -/* The addition of the rounding constant may overflow, so we use an | ||
321 | - * intermediate 64 bit accumulator. */ | ||
322 | -uint32_t HELPER(neon_rshl_u32)(uint32_t val, uint32_t shiftop) | ||
323 | +uint32_t HELPER(neon_rshl_u32)(uint32_t val, uint32_t shift) | ||
324 | { | ||
325 | - uint32_t dest; | ||
326 | - int8_t shift = (int8_t)shiftop; | ||
327 | - if (shift >= 32 || shift < -32) { | ||
328 | - dest = 0; | ||
329 | - } else if (shift == -32) { | ||
330 | - dest = val >> 31; | ||
331 | - } else if (shift < 0) { | ||
332 | - uint64_t big_dest = ((uint64_t)val + (1 << (-1 - shift))); | ||
333 | - dest = big_dest >> -shift; | ||
334 | - } else { | ||
335 | - dest = val << shift; | ||
336 | - } | ||
337 | - return dest; | ||
338 | + return do_uqrshl_bhs(val, (int8_t)shift, 32, true, NULL); | ||
339 | } | ||
340 | |||
341 | -/* Handling addition overflow with 64 bit input values is more | ||
342 | - * tricky than with 32 bit values. */ | ||
343 | -uint64_t HELPER(neon_rshl_u64)(uint64_t val, uint64_t shiftop) | ||
344 | +uint64_t HELPER(neon_rshl_u64)(uint64_t val, uint64_t shift) | ||
345 | { | ||
346 | - int8_t shift = (uint8_t)shiftop; | ||
347 | - if (shift >= 64 || shift < -64) { | ||
348 | - val = 0; | ||
349 | - } else if (shift == -64) { | ||
350 | - /* Rounding a 1-bit result just preserves that bit. */ | ||
351 | - val >>= 63; | ||
352 | - } else if (shift < 0) { | ||
353 | - val >>= (-shift - 1); | ||
354 | - if (val == UINT64_MAX) { | ||
355 | - /* In this case, it means that the rounding constant is 1, | ||
356 | - * and the addition would overflow. Return the actual | ||
357 | - * result directly. */ | ||
358 | - val = 0x8000000000000000ULL; | ||
359 | - } else { | ||
360 | - val++; | ||
361 | - val >>= 1; | ||
362 | - } | ||
363 | - } else { | ||
364 | - val <<= shift; | ||
365 | - } | ||
366 | - return val; | ||
367 | + return do_uqrshl_d(val, (int8_t)shift, true, NULL); | ||
368 | } | ||
369 | |||
370 | -#define NEON_FN(dest, src1, src2) do { \ | ||
371 | - int8_t tmp; \ | ||
372 | - tmp = (int8_t)src2; \ | ||
373 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
374 | - if (src1) { \ | ||
375 | - SET_QC(); \ | ||
376 | - dest = ~0; \ | ||
377 | - } else { \ | ||
378 | - dest = 0; \ | ||
379 | - } \ | ||
380 | - } else if (tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
381 | - dest = 0; \ | ||
382 | - } else if (tmp < 0) { \ | ||
383 | - dest = src1 >> -tmp; \ | ||
384 | - } else { \ | ||
385 | - dest = src1 << tmp; \ | ||
386 | - if ((dest >> tmp) != src1) { \ | ||
387 | - SET_QC(); \ | ||
388 | - dest = ~0; \ | ||
389 | - } \ | ||
390 | - }} while (0) | ||
391 | +#define NEON_FN(dest, src1, src2) \ | ||
392 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 8, false, env->vfp.qc)) | ||
393 | NEON_VOP_ENV(qshl_u8, neon_u8, 4) | ||
394 | +#undef NEON_FN | ||
395 | + | ||
396 | +#define NEON_FN(dest, src1, src2) \ | ||
397 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 16, false, env->vfp.qc)) | ||
398 | NEON_VOP_ENV(qshl_u16, neon_u16, 2) | ||
399 | -NEON_VOP_ENV(qshl_u32, neon_u32, 1) | ||
400 | #undef NEON_FN | ||
401 | |||
402 | -uint64_t HELPER(neon_qshl_u64)(CPUARMState *env, uint64_t val, uint64_t shiftop) | ||
403 | +uint32_t HELPER(neon_qshl_u32)(CPUARMState *env, uint32_t val, uint32_t shift) | ||
404 | { | ||
405 | - int8_t shift = (int8_t)shiftop; | ||
406 | - if (shift >= 64) { | ||
407 | - if (val) { | ||
408 | - val = ~(uint64_t)0; | ||
409 | - SET_QC(); | ||
410 | - } | ||
411 | - } else if (shift <= -64) { | ||
412 | - val = 0; | ||
413 | - } else if (shift < 0) { | ||
414 | - val >>= -shift; | ||
415 | - } else { | ||
416 | - uint64_t tmp = val; | ||
417 | - val <<= shift; | ||
418 | - if ((val >> shift) != tmp) { | ||
419 | - SET_QC(); | ||
420 | - val = ~(uint64_t)0; | ||
421 | - } | ||
422 | - } | ||
423 | - return val; | ||
424 | + return do_uqrshl_bhs(val, (int8_t)shift, 32, false, env->vfp.qc); | ||
425 | } | ||
426 | |||
427 | -#define NEON_FN(dest, src1, src2) do { \ | ||
428 | - int8_t tmp; \ | ||
429 | - tmp = (int8_t)src2; \ | ||
430 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
431 | - if (src1) { \ | ||
432 | - SET_QC(); \ | ||
433 | - dest = (uint32_t)(1 << (sizeof(src1) * 8 - 1)); \ | ||
434 | - if (src1 > 0) { \ | ||
435 | - dest--; \ | ||
436 | - } \ | ||
437 | - } else { \ | ||
438 | - dest = src1; \ | ||
439 | - } \ | ||
440 | - } else if (tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
441 | - dest = src1 >> 31; \ | ||
442 | - } else if (tmp < 0) { \ | ||
443 | - dest = src1 >> -tmp; \ | ||
444 | - } else { \ | ||
445 | - dest = src1 << tmp; \ | ||
446 | - if ((dest >> tmp) != src1) { \ | ||
447 | - SET_QC(); \ | ||
448 | - dest = (uint32_t)(1 << (sizeof(src1) * 8 - 1)); \ | ||
449 | - if (src1 > 0) { \ | ||
450 | - dest--; \ | ||
451 | - } \ | ||
452 | - } \ | ||
453 | - }} while (0) | ||
454 | +uint64_t HELPER(neon_qshl_u64)(CPUARMState *env, uint64_t val, uint64_t shift) | ||
455 | +{ | ||
456 | + return do_uqrshl_d(val, (int8_t)shift, false, env->vfp.qc); | ||
457 | +} | ||
458 | + | ||
459 | +#define NEON_FN(dest, src1, src2) \ | ||
460 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 8, false, env->vfp.qc)) | ||
461 | NEON_VOP_ENV(qshl_s8, neon_s8, 4) | ||
462 | +#undef NEON_FN | ||
463 | + | ||
464 | +#define NEON_FN(dest, src1, src2) \ | ||
465 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 16, false, env->vfp.qc)) | ||
466 | NEON_VOP_ENV(qshl_s16, neon_s16, 2) | ||
467 | -NEON_VOP_ENV(qshl_s32, neon_s32, 1) | ||
468 | #undef NEON_FN | ||
469 | |||
470 | -uint64_t HELPER(neon_qshl_s64)(CPUARMState *env, uint64_t valop, uint64_t shiftop) | ||
471 | +uint32_t HELPER(neon_qshl_s32)(CPUARMState *env, uint32_t val, uint32_t shift) | ||
472 | { | ||
473 | - int8_t shift = (uint8_t)shiftop; | ||
474 | - int64_t val = valop; | ||
475 | - if (shift >= 64) { | ||
476 | - if (val) { | ||
477 | - SET_QC(); | ||
478 | - val = (val >> 63) ^ ~SIGNBIT64; | ||
479 | - } | ||
480 | - } else if (shift <= -64) { | ||
481 | - val >>= 63; | ||
482 | - } else if (shift < 0) { | ||
483 | - val >>= -shift; | ||
484 | - } else { | ||
485 | - int64_t tmp = val; | ||
486 | - val <<= shift; | ||
487 | - if ((val >> shift) != tmp) { | ||
488 | - SET_QC(); | ||
489 | - val = (tmp >> 63) ^ ~SIGNBIT64; | ||
490 | - } | ||
491 | - } | ||
492 | - return val; | ||
493 | + return do_sqrshl_bhs(val, (int8_t)shift, 32, false, env->vfp.qc); | ||
494 | } | ||
495 | |||
496 | -#define NEON_FN(dest, src1, src2) do { \ | ||
497 | - if (src1 & (1 << (sizeof(src1) * 8 - 1))) { \ | ||
498 | - SET_QC(); \ | ||
499 | - dest = 0; \ | ||
500 | - } else { \ | ||
501 | - int8_t tmp; \ | ||
502 | - tmp = (int8_t)src2; \ | ||
503 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
504 | - if (src1) { \ | ||
505 | - SET_QC(); \ | ||
506 | - dest = ~0; \ | ||
507 | - } else { \ | ||
508 | - dest = 0; \ | ||
509 | - } \ | ||
510 | - } else if (tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
511 | - dest = 0; \ | ||
512 | - } else if (tmp < 0) { \ | ||
513 | - dest = src1 >> -tmp; \ | ||
514 | - } else { \ | ||
515 | - dest = src1 << tmp; \ | ||
516 | - if ((dest >> tmp) != src1) { \ | ||
517 | - SET_QC(); \ | ||
518 | - dest = ~0; \ | ||
519 | - } \ | ||
520 | - } \ | ||
521 | - }} while (0) | ||
522 | -NEON_VOP_ENV(qshlu_s8, neon_u8, 4) | ||
523 | -NEON_VOP_ENV(qshlu_s16, neon_u16, 2) | ||
524 | +uint64_t HELPER(neon_qshl_s64)(CPUARMState *env, uint64_t val, uint64_t shift) | ||
525 | +{ | ||
526 | + return do_sqrshl_d(val, (int8_t)shift, false, env->vfp.qc); | ||
527 | +} | ||
528 | + | ||
529 | +#define NEON_FN(dest, src1, src2) \ | ||
530 | + (dest = do_suqrshl_bhs(src1, (int8_t)src2, 8, false, env->vfp.qc)) | ||
531 | +NEON_VOP_ENV(qshlu_s8, neon_s8, 4) | ||
532 | #undef NEON_FN | ||
533 | |||
534 | -uint32_t HELPER(neon_qshlu_s32)(CPUARMState *env, uint32_t valop, uint32_t shiftop) | ||
535 | +#define NEON_FN(dest, src1, src2) \ | ||
536 | + (dest = do_suqrshl_bhs(src1, (int8_t)src2, 16, false, env->vfp.qc)) | ||
537 | +NEON_VOP_ENV(qshlu_s16, neon_s16, 2) | ||
538 | +#undef NEON_FN | ||
539 | + | ||
540 | +uint32_t HELPER(neon_qshlu_s32)(CPUARMState *env, uint32_t val, uint32_t shift) | ||
541 | { | ||
542 | - if ((int32_t)valop < 0) { | ||
543 | - SET_QC(); | ||
544 | - return 0; | ||
545 | - } | ||
546 | - return helper_neon_qshl_u32(env, valop, shiftop); | ||
547 | + return do_suqrshl_bhs(val, (int8_t)shift, 32, false, env->vfp.qc); | ||
548 | } | ||
549 | |||
550 | -uint64_t HELPER(neon_qshlu_s64)(CPUARMState *env, uint64_t valop, uint64_t shiftop) | ||
551 | +uint64_t HELPER(neon_qshlu_s64)(CPUARMState *env, uint64_t val, uint64_t shift) | ||
552 | { | ||
553 | - if ((int64_t)valop < 0) { | ||
554 | - SET_QC(); | ||
555 | - return 0; | ||
556 | - } | ||
557 | - return helper_neon_qshl_u64(env, valop, shiftop); | ||
558 | + return do_suqrshl_d(val, (int8_t)shift, false, env->vfp.qc); | ||
559 | } | ||
560 | |||
561 | -#define NEON_FN(dest, src1, src2) do { \ | ||
562 | - int8_t tmp; \ | ||
563 | - tmp = (int8_t)src2; \ | ||
564 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
565 | - if (src1) { \ | ||
566 | - SET_QC(); \ | ||
567 | - dest = ~0; \ | ||
568 | - } else { \ | ||
569 | - dest = 0; \ | ||
570 | - } \ | ||
571 | - } else if (tmp < -(ssize_t)sizeof(src1) * 8) { \ | ||
572 | - dest = 0; \ | ||
573 | - } else if (tmp == -(ssize_t)sizeof(src1) * 8) { \ | ||
574 | - dest = src1 >> (sizeof(src1) * 8 - 1); \ | ||
575 | - } else if (tmp < 0) { \ | ||
576 | - dest = (src1 + (1 << (-1 - tmp))) >> -tmp; \ | ||
577 | - } else { \ | ||
578 | - dest = src1 << tmp; \ | ||
579 | - if ((dest >> tmp) != src1) { \ | ||
580 | - SET_QC(); \ | ||
581 | - dest = ~0; \ | ||
582 | - } \ | ||
583 | - }} while (0) | ||
584 | +#define NEON_FN(dest, src1, src2) \ | ||
585 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 8, true, env->vfp.qc)) | ||
586 | NEON_VOP_ENV(qrshl_u8, neon_u8, 4) | ||
587 | +#undef NEON_FN | ||
588 | + | ||
589 | +#define NEON_FN(dest, src1, src2) \ | ||
590 | + (dest = do_uqrshl_bhs(src1, (int8_t)src2, 16, true, env->vfp.qc)) | ||
591 | NEON_VOP_ENV(qrshl_u16, neon_u16, 2) | ||
592 | #undef NEON_FN | ||
593 | |||
594 | -/* The addition of the rounding constant may overflow, so we use an | ||
595 | - * intermediate 64 bit accumulator. */ | ||
596 | -uint32_t HELPER(neon_qrshl_u32)(CPUARMState *env, uint32_t val, uint32_t shiftop) | ||
597 | +uint32_t HELPER(neon_qrshl_u32)(CPUARMState *env, uint32_t val, uint32_t shift) | ||
598 | { | ||
599 | - uint32_t dest; | ||
600 | - int8_t shift = (int8_t)shiftop; | ||
601 | - if (shift >= 32) { | ||
602 | - if (val) { | ||
603 | - SET_QC(); | ||
604 | - dest = ~0; | ||
605 | - } else { | ||
606 | - dest = 0; | ||
607 | - } | ||
608 | - } else if (shift < -32) { | ||
609 | - dest = 0; | ||
610 | - } else if (shift == -32) { | ||
611 | - dest = val >> 31; | ||
612 | - } else if (shift < 0) { | ||
613 | - uint64_t big_dest = ((uint64_t)val + (1 << (-1 - shift))); | ||
614 | - dest = big_dest >> -shift; | ||
615 | - } else { | ||
616 | - dest = val << shift; | ||
617 | - if ((dest >> shift) != val) { | ||
618 | - SET_QC(); | ||
619 | - dest = ~0; | ||
620 | - } | ||
621 | - } | ||
622 | - return dest; | ||
623 | + return do_uqrshl_bhs(val, (int8_t)shift, 32, true, env->vfp.qc); | ||
624 | } | ||
625 | |||
626 | -/* Handling addition overflow with 64 bit input values is more | ||
627 | - * tricky than with 32 bit values. */ | ||
628 | -uint64_t HELPER(neon_qrshl_u64)(CPUARMState *env, uint64_t val, uint64_t shiftop) | ||
629 | +uint64_t HELPER(neon_qrshl_u64)(CPUARMState *env, uint64_t val, uint64_t shift) | ||
630 | { | ||
631 | - int8_t shift = (int8_t)shiftop; | ||
632 | - if (shift >= 64) { | ||
633 | - if (val) { | ||
634 | - SET_QC(); | ||
635 | - val = ~0; | ||
636 | - } | ||
637 | - } else if (shift < -64) { | ||
638 | - val = 0; | ||
639 | - } else if (shift == -64) { | ||
640 | - val >>= 63; | ||
641 | - } else if (shift < 0) { | ||
642 | - val >>= (-shift - 1); | ||
643 | - if (val == UINT64_MAX) { | ||
644 | - /* In this case, it means that the rounding constant is 1, | ||
645 | - * and the addition would overflow. Return the actual | ||
646 | - * result directly. */ | ||
647 | - val = 0x8000000000000000ULL; | ||
648 | - } else { | ||
649 | - val++; | ||
650 | - val >>= 1; | ||
651 | - } | ||
652 | - } else { \ | ||
653 | - uint64_t tmp = val; | ||
654 | - val <<= shift; | ||
655 | - if ((val >> shift) != tmp) { | ||
656 | - SET_QC(); | ||
657 | - val = ~0; | ||
658 | - } | ||
659 | - } | ||
660 | - return val; | ||
661 | + return do_uqrshl_d(val, (int8_t)shift, true, env->vfp.qc); | ||
662 | } | ||
663 | |||
664 | -#define NEON_FN(dest, src1, src2) do { \ | ||
665 | - int8_t tmp; \ | ||
666 | - tmp = (int8_t)src2; \ | ||
667 | - if (tmp >= (ssize_t)sizeof(src1) * 8) { \ | ||
668 | - if (src1) { \ | ||
669 | - SET_QC(); \ | ||
670 | - dest = (typeof(dest))(1 << (sizeof(src1) * 8 - 1)); \ | ||
671 | - if (src1 > 0) { \ | ||
672 | - dest--; \ | ||
673 | - } \ | ||
674 | - } else { \ | ||
675 | - dest = 0; \ | ||
676 | - } \ | ||
677 | - } else if (tmp <= -(ssize_t)sizeof(src1) * 8) { \ | ||
678 | - dest = 0; \ | ||
679 | - } else if (tmp < 0) { \ | ||
680 | - dest = (src1 + (1 << (-1 - tmp))) >> -tmp; \ | ||
681 | - } else { \ | ||
682 | - dest = src1 << tmp; \ | ||
683 | - if ((dest >> tmp) != src1) { \ | ||
684 | - SET_QC(); \ | ||
685 | - dest = (uint32_t)(1 << (sizeof(src1) * 8 - 1)); \ | ||
686 | - if (src1 > 0) { \ | ||
687 | - dest--; \ | ||
688 | - } \ | ||
689 | - } \ | ||
690 | - }} while (0) | ||
691 | +#define NEON_FN(dest, src1, src2) \ | ||
692 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 8, true, env->vfp.qc)) | ||
693 | NEON_VOP_ENV(qrshl_s8, neon_s8, 4) | ||
694 | +#undef NEON_FN | ||
695 | + | ||
696 | +#define NEON_FN(dest, src1, src2) \ | ||
697 | + (dest = do_sqrshl_bhs(src1, (int8_t)src2, 16, true, env->vfp.qc)) | ||
698 | NEON_VOP_ENV(qrshl_s16, neon_s16, 2) | ||
699 | #undef NEON_FN | ||
700 | |||
701 | -/* The addition of the rounding constant may overflow, so we use an | ||
702 | - * intermediate 64 bit accumulator. */ | ||
703 | -uint32_t HELPER(neon_qrshl_s32)(CPUARMState *env, uint32_t valop, uint32_t shiftop) | ||
704 | +uint32_t HELPER(neon_qrshl_s32)(CPUARMState *env, uint32_t val, uint32_t shift) | ||
705 | { | ||
706 | - int32_t dest; | ||
707 | - int32_t val = (int32_t)valop; | ||
708 | - int8_t shift = (int8_t)shiftop; | ||
709 | - if (shift >= 32) { | ||
710 | - if (val) { | ||
711 | - SET_QC(); | ||
712 | - dest = (val >> 31) ^ ~SIGNBIT; | ||
713 | - } else { | ||
714 | - dest = 0; | ||
715 | - } | ||
716 | - } else if (shift <= -32) { | ||
717 | - dest = 0; | ||
718 | - } else if (shift < 0) { | ||
719 | - int64_t big_dest = ((int64_t)val + (1 << (-1 - shift))); | ||
720 | - dest = big_dest >> -shift; | ||
721 | - } else { | ||
722 | - dest = val << shift; | ||
723 | - if ((dest >> shift) != val) { | ||
724 | - SET_QC(); | ||
725 | - dest = (val >> 31) ^ ~SIGNBIT; | ||
726 | - } | ||
727 | - } | ||
728 | - return dest; | ||
729 | + return do_sqrshl_bhs(val, (int8_t)shift, 32, true, env->vfp.qc); | ||
730 | } | ||
731 | |||
732 | -/* Handling addition overflow with 64 bit input values is more | ||
733 | - * tricky than with 32 bit values. */ | ||
734 | -uint64_t HELPER(neon_qrshl_s64)(CPUARMState *env, uint64_t valop, uint64_t shiftop) | ||
735 | +uint64_t HELPER(neon_qrshl_s64)(CPUARMState *env, uint64_t val, uint64_t shift) | ||
736 | { | ||
737 | - int8_t shift = (uint8_t)shiftop; | ||
738 | - int64_t val = valop; | ||
739 | - | ||
740 | - if (shift >= 64) { | ||
741 | - if (val) { | ||
742 | - SET_QC(); | ||
743 | - val = (val >> 63) ^ ~SIGNBIT64; | ||
744 | - } | ||
745 | - } else if (shift <= -64) { | ||
746 | - val = 0; | ||
747 | - } else if (shift < 0) { | ||
748 | - val >>= (-shift - 1); | ||
749 | - if (val == INT64_MAX) { | ||
750 | - /* In this case, it means that the rounding constant is 1, | ||
751 | - * and the addition would overflow. Return the actual | ||
752 | - * result directly. */ | ||
753 | - val = 0x4000000000000000ULL; | ||
754 | - } else { | ||
755 | - val++; | ||
756 | - val >>= 1; | ||
757 | - } | ||
758 | - } else { | ||
759 | - int64_t tmp = val; | ||
760 | - val <<= shift; | ||
761 | - if ((val >> shift) != tmp) { | ||
762 | - SET_QC(); | ||
763 | - val = (tmp >> 63) ^ ~SIGNBIT64; | ||
764 | - } | ||
765 | - } | ||
766 | - return val; | ||
767 | + return do_sqrshl_d(val, (int8_t)shift, true, env->vfp.qc); | ||
768 | } | ||
769 | |||
770 | uint32_t HELPER(neon_add_u8)(uint32_t a, uint32_t b) | ||
771 | -- | ||
772 | 2.20.1 | ||
773 | |||
774 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Message-id: 20210525010358.152808-7-richard.henderson@linaro.org | ||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 54 +++++++++++++++++++++++ | ||
9 | target/arm/sve.decode | 17 ++++++++ | ||
10 | target/arm/sve_helper.c | 87 ++++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 18 ++++++++ | ||
12 | 4 files changed, 176 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_uadalp_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(sve2_uadalp_zpzz_d, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_5(sve2_srshl_zpzz_b, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_srshl_zpzz_h, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_srshl_zpzz_s, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_srshl_zpzz_d, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_5(sve2_urshl_zpzz_b, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_urshl_zpzz_h, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(sve2_urshl_zpzz_s, TCG_CALL_NO_RWG, | ||
36 | + void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_5(sve2_urshl_zpzz_d, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | + | ||
40 | +DEF_HELPER_FLAGS_5(sve2_sqshl_zpzz_b, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | +DEF_HELPER_FLAGS_5(sve2_sqshl_zpzz_h, TCG_CALL_NO_RWG, | ||
43 | + void, ptr, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_5(sve2_sqshl_zpzz_s, TCG_CALL_NO_RWG, | ||
45 | + void, ptr, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_5(sve2_sqshl_zpzz_d, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, i32) | ||
48 | + | ||
49 | +DEF_HELPER_FLAGS_5(sve2_uqshl_zpzz_b, TCG_CALL_NO_RWG, | ||
50 | + void, ptr, ptr, ptr, ptr, i32) | ||
51 | +DEF_HELPER_FLAGS_5(sve2_uqshl_zpzz_h, TCG_CALL_NO_RWG, | ||
52 | + void, ptr, ptr, ptr, ptr, i32) | ||
53 | +DEF_HELPER_FLAGS_5(sve2_uqshl_zpzz_s, TCG_CALL_NO_RWG, | ||
54 | + void, ptr, ptr, ptr, ptr, i32) | ||
55 | +DEF_HELPER_FLAGS_5(sve2_uqshl_zpzz_d, TCG_CALL_NO_RWG, | ||
56 | + void, ptr, ptr, ptr, ptr, i32) | ||
57 | + | ||
58 | +DEF_HELPER_FLAGS_5(sve2_sqrshl_zpzz_b, TCG_CALL_NO_RWG, | ||
59 | + void, ptr, ptr, ptr, ptr, i32) | ||
60 | +DEF_HELPER_FLAGS_5(sve2_sqrshl_zpzz_h, TCG_CALL_NO_RWG, | ||
61 | + void, ptr, ptr, ptr, ptr, i32) | ||
62 | +DEF_HELPER_FLAGS_5(sve2_sqrshl_zpzz_s, TCG_CALL_NO_RWG, | ||
63 | + void, ptr, ptr, ptr, ptr, i32) | ||
64 | +DEF_HELPER_FLAGS_5(sve2_sqrshl_zpzz_d, TCG_CALL_NO_RWG, | ||
65 | + void, ptr, ptr, ptr, ptr, i32) | ||
66 | + | ||
67 | +DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_b, TCG_CALL_NO_RWG, | ||
68 | + void, ptr, ptr, ptr, ptr, i32) | ||
69 | +DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_h, TCG_CALL_NO_RWG, | ||
70 | + void, ptr, ptr, ptr, ptr, i32) | ||
71 | +DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_s, TCG_CALL_NO_RWG, | ||
72 | + void, ptr, ptr, ptr, ptr, i32) | ||
73 | +DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_d, TCG_CALL_NO_RWG, | ||
74 | + void, ptr, ptr, ptr, ptr, i32) | ||
75 | + | ||
76 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_s, TCG_CALL_NO_RWG, | ||
77 | void, ptr, ptr, ptr, ptr, i32) | ||
78 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_d, TCG_CALL_NO_RWG, | ||
79 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/sve.decode | ||
82 | +++ b/target/arm/sve.decode | ||
83 | @@ -XXX,XX +XXX,XX @@ URECPE 01000100 .. 000 000 101 ... ..... ..... @rd_pg_rn | ||
84 | URSQRTE 01000100 .. 000 001 101 ... ..... ..... @rd_pg_rn | ||
85 | SQABS 01000100 .. 001 000 101 ... ..... ..... @rd_pg_rn | ||
86 | SQNEG 01000100 .. 001 001 101 ... ..... ..... @rd_pg_rn | ||
87 | + | ||
88 | +### SVE2 saturating/rounding bitwise shift left (predicated) | ||
89 | + | ||
90 | +SRSHL 01000100 .. 000 010 100 ... ..... ..... @rdn_pg_rm | ||
91 | +URSHL 01000100 .. 000 011 100 ... ..... ..... @rdn_pg_rm | ||
92 | +SRSHL 01000100 .. 000 110 100 ... ..... ..... @rdm_pg_rn # SRSHLR | ||
93 | +URSHL 01000100 .. 000 111 100 ... ..... ..... @rdm_pg_rn # URSHLR | ||
94 | + | ||
95 | +SQSHL 01000100 .. 001 000 100 ... ..... ..... @rdn_pg_rm | ||
96 | +UQSHL 01000100 .. 001 001 100 ... ..... ..... @rdn_pg_rm | ||
97 | +SQSHL 01000100 .. 001 100 100 ... ..... ..... @rdm_pg_rn # SQSHLR | ||
98 | +UQSHL 01000100 .. 001 101 100 ... ..... ..... @rdm_pg_rn # UQSHLR | ||
99 | + | ||
100 | +SQRSHL 01000100 .. 001 010 100 ... ..... ..... @rdn_pg_rm | ||
101 | +UQRSHL 01000100 .. 001 011 100 ... ..... ..... @rdn_pg_rm | ||
102 | +SQRSHL 01000100 .. 001 110 100 ... ..... ..... @rdm_pg_rn # SQRSHLR | ||
103 | +UQRSHL 01000100 .. 001 111 100 ... ..... ..... @rdm_pg_rn # UQRSHLR | ||
104 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
105 | index XXXXXXX..XXXXXXX 100644 | ||
106 | --- a/target/arm/sve_helper.c | ||
107 | +++ b/target/arm/sve_helper.c | ||
108 | @@ -XXX,XX +XXX,XX @@ | ||
109 | #include "tcg/tcg-gvec-desc.h" | ||
110 | #include "fpu/softfloat.h" | ||
111 | #include "tcg/tcg.h" | ||
112 | +#include "vec_internal.h" | ||
113 | |||
114 | |||
115 | /* Note that vector data is stored in host-endian 64-bit chunks, | ||
116 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ(sve2_uadalp_zpzz_h, uint16_t, H1_2, do_uadalp_h) | ||
117 | DO_ZPZZ(sve2_uadalp_zpzz_s, uint32_t, H1_4, do_uadalp_s) | ||
118 | DO_ZPZZ_D(sve2_uadalp_zpzz_d, uint64_t, do_uadalp_d) | ||
119 | |||
120 | +#define do_srshl_b(n, m) do_sqrshl_bhs(n, m, 8, true, NULL) | ||
121 | +#define do_srshl_h(n, m) do_sqrshl_bhs(n, m, 16, true, NULL) | ||
122 | +#define do_srshl_s(n, m) do_sqrshl_bhs(n, m, 32, true, NULL) | ||
123 | +#define do_srshl_d(n, m) do_sqrshl_d(n, m, true, NULL) | ||
124 | + | ||
125 | +DO_ZPZZ(sve2_srshl_zpzz_b, int8_t, H1, do_srshl_b) | ||
126 | +DO_ZPZZ(sve2_srshl_zpzz_h, int16_t, H1_2, do_srshl_h) | ||
127 | +DO_ZPZZ(sve2_srshl_zpzz_s, int32_t, H1_4, do_srshl_s) | ||
128 | +DO_ZPZZ_D(sve2_srshl_zpzz_d, int64_t, do_srshl_d) | ||
129 | + | ||
130 | +#define do_urshl_b(n, m) do_uqrshl_bhs(n, (int8_t)m, 8, true, NULL) | ||
131 | +#define do_urshl_h(n, m) do_uqrshl_bhs(n, (int16_t)m, 16, true, NULL) | ||
132 | +#define do_urshl_s(n, m) do_uqrshl_bhs(n, m, 32, true, NULL) | ||
133 | +#define do_urshl_d(n, m) do_uqrshl_d(n, m, true, NULL) | ||
134 | + | ||
135 | +DO_ZPZZ(sve2_urshl_zpzz_b, uint8_t, H1, do_urshl_b) | ||
136 | +DO_ZPZZ(sve2_urshl_zpzz_h, uint16_t, H1_2, do_urshl_h) | ||
137 | +DO_ZPZZ(sve2_urshl_zpzz_s, uint32_t, H1_4, do_urshl_s) | ||
138 | +DO_ZPZZ_D(sve2_urshl_zpzz_d, uint64_t, do_urshl_d) | ||
139 | + | ||
140 | +/* | ||
141 | + * Unlike the NEON and AdvSIMD versions, there is no QC bit to set. | ||
142 | + * We pass in a pointer to a dummy saturation field to trigger | ||
143 | + * the saturating arithmetic but discard the information about | ||
144 | + * whether it has occurred. | ||
145 | + */ | ||
146 | +#define do_sqshl_b(n, m) \ | ||
147 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 8, false, &discard); }) | ||
148 | +#define do_sqshl_h(n, m) \ | ||
149 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 16, false, &discard); }) | ||
150 | +#define do_sqshl_s(n, m) \ | ||
151 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 32, false, &discard); }) | ||
152 | +#define do_sqshl_d(n, m) \ | ||
153 | + ({ uint32_t discard; do_sqrshl_d(n, m, false, &discard); }) | ||
154 | + | ||
155 | +DO_ZPZZ(sve2_sqshl_zpzz_b, int8_t, H1_2, do_sqshl_b) | ||
156 | +DO_ZPZZ(sve2_sqshl_zpzz_h, int16_t, H1_2, do_sqshl_h) | ||
157 | +DO_ZPZZ(sve2_sqshl_zpzz_s, int32_t, H1_4, do_sqshl_s) | ||
158 | +DO_ZPZZ_D(sve2_sqshl_zpzz_d, int64_t, do_sqshl_d) | ||
159 | + | ||
160 | +#define do_uqshl_b(n, m) \ | ||
161 | + ({ uint32_t discard; do_uqrshl_bhs(n, (int8_t)m, 8, false, &discard); }) | ||
162 | +#define do_uqshl_h(n, m) \ | ||
163 | + ({ uint32_t discard; do_uqrshl_bhs(n, (int16_t)m, 16, false, &discard); }) | ||
164 | +#define do_uqshl_s(n, m) \ | ||
165 | + ({ uint32_t discard; do_uqrshl_bhs(n, m, 32, false, &discard); }) | ||
166 | +#define do_uqshl_d(n, m) \ | ||
167 | + ({ uint32_t discard; do_uqrshl_d(n, m, false, &discard); }) | ||
168 | + | ||
169 | +DO_ZPZZ(sve2_uqshl_zpzz_b, uint8_t, H1_2, do_uqshl_b) | ||
170 | +DO_ZPZZ(sve2_uqshl_zpzz_h, uint16_t, H1_2, do_uqshl_h) | ||
171 | +DO_ZPZZ(sve2_uqshl_zpzz_s, uint32_t, H1_4, do_uqshl_s) | ||
172 | +DO_ZPZZ_D(sve2_uqshl_zpzz_d, uint64_t, do_uqshl_d) | ||
173 | + | ||
174 | +#define do_sqrshl_b(n, m) \ | ||
175 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 8, true, &discard); }) | ||
176 | +#define do_sqrshl_h(n, m) \ | ||
177 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 16, true, &discard); }) | ||
178 | +#define do_sqrshl_s(n, m) \ | ||
179 | + ({ uint32_t discard; do_sqrshl_bhs(n, m, 32, true, &discard); }) | ||
180 | +#define do_sqrshl_d(n, m) \ | ||
181 | + ({ uint32_t discard; do_sqrshl_d(n, m, true, &discard); }) | ||
182 | + | ||
183 | +DO_ZPZZ(sve2_sqrshl_zpzz_b, int8_t, H1_2, do_sqrshl_b) | ||
184 | +DO_ZPZZ(sve2_sqrshl_zpzz_h, int16_t, H1_2, do_sqrshl_h) | ||
185 | +DO_ZPZZ(sve2_sqrshl_zpzz_s, int32_t, H1_4, do_sqrshl_s) | ||
186 | +DO_ZPZZ_D(sve2_sqrshl_zpzz_d, int64_t, do_sqrshl_d) | ||
187 | + | ||
188 | +#undef do_sqrshl_d | ||
189 | + | ||
190 | +#define do_uqrshl_b(n, m) \ | ||
191 | + ({ uint32_t discard; do_uqrshl_bhs(n, (int8_t)m, 8, true, &discard); }) | ||
192 | +#define do_uqrshl_h(n, m) \ | ||
193 | + ({ uint32_t discard; do_uqrshl_bhs(n, (int16_t)m, 16, true, &discard); }) | ||
194 | +#define do_uqrshl_s(n, m) \ | ||
195 | + ({ uint32_t discard; do_uqrshl_bhs(n, m, 32, true, &discard); }) | ||
196 | +#define do_uqrshl_d(n, m) \ | ||
197 | + ({ uint32_t discard; do_uqrshl_d(n, m, true, &discard); }) | ||
198 | + | ||
199 | +DO_ZPZZ(sve2_uqrshl_zpzz_b, uint8_t, H1_2, do_uqrshl_b) | ||
200 | +DO_ZPZZ(sve2_uqrshl_zpzz_h, uint16_t, H1_2, do_uqrshl_h) | ||
201 | +DO_ZPZZ(sve2_uqrshl_zpzz_s, uint32_t, H1_4, do_uqrshl_s) | ||
202 | +DO_ZPZZ_D(sve2_uqrshl_zpzz_d, uint64_t, do_uqrshl_d) | ||
203 | + | ||
204 | +#undef do_uqrshl_d | ||
205 | + | ||
206 | #undef DO_ZPZZ | ||
207 | #undef DO_ZPZZ_D | ||
208 | |||
209 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
210 | index XXXXXXX..XXXXXXX 100644 | ||
211 | --- a/target/arm/translate-sve.c | ||
212 | +++ b/target/arm/translate-sve.c | ||
213 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQNEG(DisasContext *s, arg_rpr_esz *a) | ||
214 | }; | ||
215 | return do_sve2_zpz_ool(s, a, fns[a->esz]); | ||
216 | } | ||
217 | + | ||
218 | +#define DO_SVE2_ZPZZ(NAME, name) \ | ||
219 | +static bool trans_##NAME(DisasContext *s, arg_rprr_esz *a) \ | ||
220 | +{ \ | ||
221 | + static gen_helper_gvec_4 * const fns[4] = { \ | ||
222 | + gen_helper_sve2_##name##_zpzz_b, gen_helper_sve2_##name##_zpzz_h, \ | ||
223 | + gen_helper_sve2_##name##_zpzz_s, gen_helper_sve2_##name##_zpzz_d, \ | ||
224 | + }; \ | ||
225 | + return do_sve2_zpzz_ool(s, a, fns[a->esz]); \ | ||
226 | +} | ||
227 | + | ||
228 | +DO_SVE2_ZPZZ(SQSHL, sqshl) | ||
229 | +DO_SVE2_ZPZZ(SQRSHL, sqrshl) | ||
230 | +DO_SVE2_ZPZZ(SRSHL, srshl) | ||
231 | + | ||
232 | +DO_SVE2_ZPZZ(UQSHL, uqshl) | ||
233 | +DO_SVE2_ZPZZ(UQRSHL, uqrshl) | ||
234 | +DO_SVE2_ZPZZ(URSHL, urshl) | ||
235 | -- | ||
236 | 2.20.1 | ||
237 | |||
238 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-8-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 54 ++++++++++++++++++++++++++++++++++++++ | ||
9 | target/arm/sve.decode | 11 ++++++++ | ||
10 | target/arm/sve_helper.c | 39 +++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 8 ++++++ | ||
12 | 4 files changed, 112 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(sve2_uqrshl_zpzz_d, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_5(sve2_shadd_zpzz_b, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_shadd_zpzz_h, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_shadd_zpzz_s, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_shadd_zpzz_d, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_5(sve2_uhadd_zpzz_b, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_uhadd_zpzz_h, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(sve2_uhadd_zpzz_s, TCG_CALL_NO_RWG, | ||
36 | + void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_5(sve2_uhadd_zpzz_d, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | + | ||
40 | +DEF_HELPER_FLAGS_5(sve2_srhadd_zpzz_b, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | +DEF_HELPER_FLAGS_5(sve2_srhadd_zpzz_h, TCG_CALL_NO_RWG, | ||
43 | + void, ptr, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_5(sve2_srhadd_zpzz_s, TCG_CALL_NO_RWG, | ||
45 | + void, ptr, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_5(sve2_srhadd_zpzz_d, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, i32) | ||
48 | + | ||
49 | +DEF_HELPER_FLAGS_5(sve2_urhadd_zpzz_b, TCG_CALL_NO_RWG, | ||
50 | + void, ptr, ptr, ptr, ptr, i32) | ||
51 | +DEF_HELPER_FLAGS_5(sve2_urhadd_zpzz_h, TCG_CALL_NO_RWG, | ||
52 | + void, ptr, ptr, ptr, ptr, i32) | ||
53 | +DEF_HELPER_FLAGS_5(sve2_urhadd_zpzz_s, TCG_CALL_NO_RWG, | ||
54 | + void, ptr, ptr, ptr, ptr, i32) | ||
55 | +DEF_HELPER_FLAGS_5(sve2_urhadd_zpzz_d, TCG_CALL_NO_RWG, | ||
56 | + void, ptr, ptr, ptr, ptr, i32) | ||
57 | + | ||
58 | +DEF_HELPER_FLAGS_5(sve2_shsub_zpzz_b, TCG_CALL_NO_RWG, | ||
59 | + void, ptr, ptr, ptr, ptr, i32) | ||
60 | +DEF_HELPER_FLAGS_5(sve2_shsub_zpzz_h, TCG_CALL_NO_RWG, | ||
61 | + void, ptr, ptr, ptr, ptr, i32) | ||
62 | +DEF_HELPER_FLAGS_5(sve2_shsub_zpzz_s, TCG_CALL_NO_RWG, | ||
63 | + void, ptr, ptr, ptr, ptr, i32) | ||
64 | +DEF_HELPER_FLAGS_5(sve2_shsub_zpzz_d, TCG_CALL_NO_RWG, | ||
65 | + void, ptr, ptr, ptr, ptr, i32) | ||
66 | + | ||
67 | +DEF_HELPER_FLAGS_5(sve2_uhsub_zpzz_b, TCG_CALL_NO_RWG, | ||
68 | + void, ptr, ptr, ptr, ptr, i32) | ||
69 | +DEF_HELPER_FLAGS_5(sve2_uhsub_zpzz_h, TCG_CALL_NO_RWG, | ||
70 | + void, ptr, ptr, ptr, ptr, i32) | ||
71 | +DEF_HELPER_FLAGS_5(sve2_uhsub_zpzz_s, TCG_CALL_NO_RWG, | ||
72 | + void, ptr, ptr, ptr, ptr, i32) | ||
73 | +DEF_HELPER_FLAGS_5(sve2_uhsub_zpzz_d, TCG_CALL_NO_RWG, | ||
74 | + void, ptr, ptr, ptr, ptr, i32) | ||
75 | + | ||
76 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_s, TCG_CALL_NO_RWG, | ||
77 | void, ptr, ptr, ptr, ptr, i32) | ||
78 | DEF_HELPER_FLAGS_5(sve_sdiv_zpzz_d, TCG_CALL_NO_RWG, | ||
79 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/sve.decode | ||
82 | +++ b/target/arm/sve.decode | ||
83 | @@ -XXX,XX +XXX,XX @@ SQRSHL 01000100 .. 001 010 100 ... ..... ..... @rdn_pg_rm | ||
84 | UQRSHL 01000100 .. 001 011 100 ... ..... ..... @rdn_pg_rm | ||
85 | SQRSHL 01000100 .. 001 110 100 ... ..... ..... @rdm_pg_rn # SQRSHLR | ||
86 | UQRSHL 01000100 .. 001 111 100 ... ..... ..... @rdm_pg_rn # UQRSHLR | ||
87 | + | ||
88 | +### SVE2 integer halving add/subtract (predicated) | ||
89 | + | ||
90 | +SHADD 01000100 .. 010 000 100 ... ..... ..... @rdn_pg_rm | ||
91 | +UHADD 01000100 .. 010 001 100 ... ..... ..... @rdn_pg_rm | ||
92 | +SHSUB 01000100 .. 010 010 100 ... ..... ..... @rdn_pg_rm | ||
93 | +UHSUB 01000100 .. 010 011 100 ... ..... ..... @rdn_pg_rm | ||
94 | +SRHADD 01000100 .. 010 100 100 ... ..... ..... @rdn_pg_rm | ||
95 | +URHADD 01000100 .. 010 101 100 ... ..... ..... @rdn_pg_rm | ||
96 | +SHSUB 01000100 .. 010 110 100 ... ..... ..... @rdm_pg_rn # SHSUBR | ||
97 | +UHSUB 01000100 .. 010 111 100 ... ..... ..... @rdm_pg_rn # UHSUBR | ||
98 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
99 | index XXXXXXX..XXXXXXX 100644 | ||
100 | --- a/target/arm/sve_helper.c | ||
101 | +++ b/target/arm/sve_helper.c | ||
102 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ_D(sve2_uqrshl_zpzz_d, uint64_t, do_uqrshl_d) | ||
103 | |||
104 | #undef do_uqrshl_d | ||
105 | |||
106 | +#define DO_HADD_BHS(n, m) (((int64_t)n + m) >> 1) | ||
107 | +#define DO_HADD_D(n, m) ((n >> 1) + (m >> 1) + (n & m & 1)) | ||
108 | + | ||
109 | +DO_ZPZZ(sve2_shadd_zpzz_b, int8_t, H1, DO_HADD_BHS) | ||
110 | +DO_ZPZZ(sve2_shadd_zpzz_h, int16_t, H1_2, DO_HADD_BHS) | ||
111 | +DO_ZPZZ(sve2_shadd_zpzz_s, int32_t, H1_4, DO_HADD_BHS) | ||
112 | +DO_ZPZZ_D(sve2_shadd_zpzz_d, int64_t, DO_HADD_D) | ||
113 | + | ||
114 | +DO_ZPZZ(sve2_uhadd_zpzz_b, uint8_t, H1, DO_HADD_BHS) | ||
115 | +DO_ZPZZ(sve2_uhadd_zpzz_h, uint16_t, H1_2, DO_HADD_BHS) | ||
116 | +DO_ZPZZ(sve2_uhadd_zpzz_s, uint32_t, H1_4, DO_HADD_BHS) | ||
117 | +DO_ZPZZ_D(sve2_uhadd_zpzz_d, uint64_t, DO_HADD_D) | ||
118 | + | ||
119 | +#define DO_RHADD_BHS(n, m) (((int64_t)n + m + 1) >> 1) | ||
120 | +#define DO_RHADD_D(n, m) ((n >> 1) + (m >> 1) + ((n | m) & 1)) | ||
121 | + | ||
122 | +DO_ZPZZ(sve2_srhadd_zpzz_b, int8_t, H1, DO_RHADD_BHS) | ||
123 | +DO_ZPZZ(sve2_srhadd_zpzz_h, int16_t, H1_2, DO_RHADD_BHS) | ||
124 | +DO_ZPZZ(sve2_srhadd_zpzz_s, int32_t, H1_4, DO_RHADD_BHS) | ||
125 | +DO_ZPZZ_D(sve2_srhadd_zpzz_d, int64_t, DO_RHADD_D) | ||
126 | + | ||
127 | +DO_ZPZZ(sve2_urhadd_zpzz_b, uint8_t, H1, DO_RHADD_BHS) | ||
128 | +DO_ZPZZ(sve2_urhadd_zpzz_h, uint16_t, H1_2, DO_RHADD_BHS) | ||
129 | +DO_ZPZZ(sve2_urhadd_zpzz_s, uint32_t, H1_4, DO_RHADD_BHS) | ||
130 | +DO_ZPZZ_D(sve2_urhadd_zpzz_d, uint64_t, DO_RHADD_D) | ||
131 | + | ||
132 | +#define DO_HSUB_BHS(n, m) (((int64_t)n - m) >> 1) | ||
133 | +#define DO_HSUB_D(n, m) ((n >> 1) - (m >> 1) - (~n & m & 1)) | ||
134 | + | ||
135 | +DO_ZPZZ(sve2_shsub_zpzz_b, int8_t, H1, DO_HSUB_BHS) | ||
136 | +DO_ZPZZ(sve2_shsub_zpzz_h, int16_t, H1_2, DO_HSUB_BHS) | ||
137 | +DO_ZPZZ(sve2_shsub_zpzz_s, int32_t, H1_4, DO_HSUB_BHS) | ||
138 | +DO_ZPZZ_D(sve2_shsub_zpzz_d, int64_t, DO_HSUB_D) | ||
139 | + | ||
140 | +DO_ZPZZ(sve2_uhsub_zpzz_b, uint8_t, H1, DO_HSUB_BHS) | ||
141 | +DO_ZPZZ(sve2_uhsub_zpzz_h, uint16_t, H1_2, DO_HSUB_BHS) | ||
142 | +DO_ZPZZ(sve2_uhsub_zpzz_s, uint32_t, H1_4, DO_HSUB_BHS) | ||
143 | +DO_ZPZZ_D(sve2_uhsub_zpzz_d, uint64_t, DO_HSUB_D) | ||
144 | + | ||
145 | #undef DO_ZPZZ | ||
146 | #undef DO_ZPZZ_D | ||
147 | |||
148 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
149 | index XXXXXXX..XXXXXXX 100644 | ||
150 | --- a/target/arm/translate-sve.c | ||
151 | +++ b/target/arm/translate-sve.c | ||
152 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ(SRSHL, srshl) | ||
153 | DO_SVE2_ZPZZ(UQSHL, uqshl) | ||
154 | DO_SVE2_ZPZZ(UQRSHL, uqrshl) | ||
155 | DO_SVE2_ZPZZ(URSHL, urshl) | ||
156 | + | ||
157 | +DO_SVE2_ZPZZ(SHADD, shadd) | ||
158 | +DO_SVE2_ZPZZ(SRHADD, srhadd) | ||
159 | +DO_SVE2_ZPZZ(SHSUB, shsub) | ||
160 | + | ||
161 | +DO_SVE2_ZPZZ(UHADD, uhadd) | ||
162 | +DO_SVE2_ZPZZ(URHADD, urhadd) | ||
163 | +DO_SVE2_ZPZZ(UHSUB, uhsub) | ||
164 | -- | ||
165 | 2.20.1 | ||
166 | |||
167 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-9-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 45 ++++++++++++++++++++++ | ||
9 | target/arm/sve.decode | 8 ++++ | ||
10 | target/arm/sve_helper.c | 76 ++++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 6 +++ | ||
12 | 4 files changed, 135 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve_sel_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(sve_sel_zpzz_d, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_5(sve2_addp_zpzz_b, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_addp_zpzz_h, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_addp_zpzz_s, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_addp_zpzz_d, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_5(sve2_smaxp_zpzz_b, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_smaxp_zpzz_h, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(sve2_smaxp_zpzz_s, TCG_CALL_NO_RWG, | ||
36 | + void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_5(sve2_smaxp_zpzz_d, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | + | ||
40 | +DEF_HELPER_FLAGS_5(sve2_umaxp_zpzz_b, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | +DEF_HELPER_FLAGS_5(sve2_umaxp_zpzz_h, TCG_CALL_NO_RWG, | ||
43 | + void, ptr, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_5(sve2_umaxp_zpzz_s, TCG_CALL_NO_RWG, | ||
45 | + void, ptr, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_5(sve2_umaxp_zpzz_d, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, i32) | ||
48 | + | ||
49 | +DEF_HELPER_FLAGS_5(sve2_sminp_zpzz_b, TCG_CALL_NO_RWG, | ||
50 | + void, ptr, ptr, ptr, ptr, i32) | ||
51 | +DEF_HELPER_FLAGS_5(sve2_sminp_zpzz_h, TCG_CALL_NO_RWG, | ||
52 | + void, ptr, ptr, ptr, ptr, i32) | ||
53 | +DEF_HELPER_FLAGS_5(sve2_sminp_zpzz_s, TCG_CALL_NO_RWG, | ||
54 | + void, ptr, ptr, ptr, ptr, i32) | ||
55 | +DEF_HELPER_FLAGS_5(sve2_sminp_zpzz_d, TCG_CALL_NO_RWG, | ||
56 | + void, ptr, ptr, ptr, ptr, i32) | ||
57 | + | ||
58 | +DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_b, TCG_CALL_NO_RWG, | ||
59 | + void, ptr, ptr, ptr, ptr, i32) | ||
60 | +DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_h, TCG_CALL_NO_RWG, | ||
61 | + void, ptr, ptr, ptr, ptr, i32) | ||
62 | +DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_s, TCG_CALL_NO_RWG, | ||
63 | + void, ptr, ptr, ptr, ptr, i32) | ||
64 | +DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_d, TCG_CALL_NO_RWG, | ||
65 | + void, ptr, ptr, ptr, ptr, i32) | ||
66 | + | ||
67 | DEF_HELPER_FLAGS_5(sve_asr_zpzw_b, TCG_CALL_NO_RWG, | ||
68 | void, ptr, ptr, ptr, ptr, i32) | ||
69 | DEF_HELPER_FLAGS_5(sve_asr_zpzw_h, TCG_CALL_NO_RWG, | ||
70 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
71 | index XXXXXXX..XXXXXXX 100644 | ||
72 | --- a/target/arm/sve.decode | ||
73 | +++ b/target/arm/sve.decode | ||
74 | @@ -XXX,XX +XXX,XX @@ SRHADD 01000100 .. 010 100 100 ... ..... ..... @rdn_pg_rm | ||
75 | URHADD 01000100 .. 010 101 100 ... ..... ..... @rdn_pg_rm | ||
76 | SHSUB 01000100 .. 010 110 100 ... ..... ..... @rdm_pg_rn # SHSUBR | ||
77 | UHSUB 01000100 .. 010 111 100 ... ..... ..... @rdm_pg_rn # UHSUBR | ||
78 | + | ||
79 | +### SVE2 integer pairwise arithmetic | ||
80 | + | ||
81 | +ADDP 01000100 .. 010 001 101 ... ..... ..... @rdn_pg_rm | ||
82 | +SMAXP 01000100 .. 010 100 101 ... ..... ..... @rdn_pg_rm | ||
83 | +UMAXP 01000100 .. 010 101 101 ... ..... ..... @rdn_pg_rm | ||
84 | +SMINP 01000100 .. 010 110 101 ... ..... ..... @rdn_pg_rm | ||
85 | +UMINP 01000100 .. 010 111 101 ... ..... ..... @rdn_pg_rm | ||
86 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
87 | index XXXXXXX..XXXXXXX 100644 | ||
88 | --- a/target/arm/sve_helper.c | ||
89 | +++ b/target/arm/sve_helper.c | ||
90 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ_D(sve2_uhsub_zpzz_d, uint64_t, DO_HSUB_D) | ||
91 | #undef DO_ZPZZ | ||
92 | #undef DO_ZPZZ_D | ||
93 | |||
94 | +/* | ||
95 | + * Three operand expander, operating on element pairs. | ||
96 | + * If the slot I is even, the elements from from VN {I, I+1}. | ||
97 | + * If the slot I is odd, the elements from from VM {I-1, I}. | ||
98 | + * Load all of the input elements in each pair before overwriting output. | ||
99 | + */ | ||
100 | +#define DO_ZPZZ_PAIR(NAME, TYPE, H, OP) \ | ||
101 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *vg, uint32_t desc) \ | ||
102 | +{ \ | ||
103 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
104 | + for (i = 0; i < opr_sz; ) { \ | ||
105 | + uint16_t pg = *(uint16_t *)(vg + H1_2(i >> 3)); \ | ||
106 | + do { \ | ||
107 | + TYPE n0 = *(TYPE *)(vn + H(i)); \ | ||
108 | + TYPE m0 = *(TYPE *)(vm + H(i)); \ | ||
109 | + TYPE n1 = *(TYPE *)(vn + H(i + sizeof(TYPE))); \ | ||
110 | + TYPE m1 = *(TYPE *)(vm + H(i + sizeof(TYPE))); \ | ||
111 | + if (pg & 1) { \ | ||
112 | + *(TYPE *)(vd + H(i)) = OP(n0, n1); \ | ||
113 | + } \ | ||
114 | + i += sizeof(TYPE), pg >>= sizeof(TYPE); \ | ||
115 | + if (pg & 1) { \ | ||
116 | + *(TYPE *)(vd + H(i)) = OP(m0, m1); \ | ||
117 | + } \ | ||
118 | + i += sizeof(TYPE), pg >>= sizeof(TYPE); \ | ||
119 | + } while (i & 15); \ | ||
120 | + } \ | ||
121 | +} | ||
122 | + | ||
123 | +/* Similarly, specialized for 64-bit operands. */ | ||
124 | +#define DO_ZPZZ_PAIR_D(NAME, TYPE, OP) \ | ||
125 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *vg, uint32_t desc) \ | ||
126 | +{ \ | ||
127 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; \ | ||
128 | + TYPE *d = vd, *n = vn, *m = vm; \ | ||
129 | + uint8_t *pg = vg; \ | ||
130 | + for (i = 0; i < opr_sz; i += 2) { \ | ||
131 | + TYPE n0 = n[i], n1 = n[i + 1]; \ | ||
132 | + TYPE m0 = m[i], m1 = m[i + 1]; \ | ||
133 | + if (pg[H1(i)] & 1) { \ | ||
134 | + d[i] = OP(n0, n1); \ | ||
135 | + } \ | ||
136 | + if (pg[H1(i + 1)] & 1) { \ | ||
137 | + d[i + 1] = OP(m0, m1); \ | ||
138 | + } \ | ||
139 | + } \ | ||
140 | +} | ||
141 | + | ||
142 | +DO_ZPZZ_PAIR(sve2_addp_zpzz_b, uint8_t, H1, DO_ADD) | ||
143 | +DO_ZPZZ_PAIR(sve2_addp_zpzz_h, uint16_t, H1_2, DO_ADD) | ||
144 | +DO_ZPZZ_PAIR(sve2_addp_zpzz_s, uint32_t, H1_4, DO_ADD) | ||
145 | +DO_ZPZZ_PAIR_D(sve2_addp_zpzz_d, uint64_t, DO_ADD) | ||
146 | + | ||
147 | +DO_ZPZZ_PAIR(sve2_umaxp_zpzz_b, uint8_t, H1, DO_MAX) | ||
148 | +DO_ZPZZ_PAIR(sve2_umaxp_zpzz_h, uint16_t, H1_2, DO_MAX) | ||
149 | +DO_ZPZZ_PAIR(sve2_umaxp_zpzz_s, uint32_t, H1_4, DO_MAX) | ||
150 | +DO_ZPZZ_PAIR_D(sve2_umaxp_zpzz_d, uint64_t, DO_MAX) | ||
151 | + | ||
152 | +DO_ZPZZ_PAIR(sve2_uminp_zpzz_b, uint8_t, H1, DO_MIN) | ||
153 | +DO_ZPZZ_PAIR(sve2_uminp_zpzz_h, uint16_t, H1_2, DO_MIN) | ||
154 | +DO_ZPZZ_PAIR(sve2_uminp_zpzz_s, uint32_t, H1_4, DO_MIN) | ||
155 | +DO_ZPZZ_PAIR_D(sve2_uminp_zpzz_d, uint64_t, DO_MIN) | ||
156 | + | ||
157 | +DO_ZPZZ_PAIR(sve2_smaxp_zpzz_b, int8_t, H1, DO_MAX) | ||
158 | +DO_ZPZZ_PAIR(sve2_smaxp_zpzz_h, int16_t, H1_2, DO_MAX) | ||
159 | +DO_ZPZZ_PAIR(sve2_smaxp_zpzz_s, int32_t, H1_4, DO_MAX) | ||
160 | +DO_ZPZZ_PAIR_D(sve2_smaxp_zpzz_d, int64_t, DO_MAX) | ||
161 | + | ||
162 | +DO_ZPZZ_PAIR(sve2_sminp_zpzz_b, int8_t, H1, DO_MIN) | ||
163 | +DO_ZPZZ_PAIR(sve2_sminp_zpzz_h, int16_t, H1_2, DO_MIN) | ||
164 | +DO_ZPZZ_PAIR(sve2_sminp_zpzz_s, int32_t, H1_4, DO_MIN) | ||
165 | +DO_ZPZZ_PAIR_D(sve2_sminp_zpzz_d, int64_t, DO_MIN) | ||
166 | + | ||
167 | +#undef DO_ZPZZ_PAIR | ||
168 | +#undef DO_ZPZZ_PAIR_D | ||
169 | + | ||
170 | /* Three-operand expander, controlled by a predicate, in which the | ||
171 | * third operand is "wide". That is, for D = N op M, the same 64-bit | ||
172 | * value of M is used with all of the narrower values of N. | ||
173 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
174 | index XXXXXXX..XXXXXXX 100644 | ||
175 | --- a/target/arm/translate-sve.c | ||
176 | +++ b/target/arm/translate-sve.c | ||
177 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ(SHSUB, shsub) | ||
178 | DO_SVE2_ZPZZ(UHADD, uhadd) | ||
179 | DO_SVE2_ZPZZ(URHADD, urhadd) | ||
180 | DO_SVE2_ZPZZ(UHSUB, uhsub) | ||
181 | + | ||
182 | +DO_SVE2_ZPZZ(ADDP, addp) | ||
183 | +DO_SVE2_ZPZZ(SMAXP, smaxp) | ||
184 | +DO_SVE2_ZPZZ(UMAXP, umaxp) | ||
185 | +DO_SVE2_ZPZZ(SMINP, sminp) | ||
186 | +DO_SVE2_ZPZZ(UMINP, uminp) | ||
187 | -- | ||
188 | 2.20.1 | ||
189 | |||
190 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-10-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 54 +++++++++++ | ||
9 | target/arm/sve.decode | 11 +++ | ||
10 | target/arm/sve_helper.c | 194 ++++++++++++++++++++++++++----------- | ||
11 | target/arm/translate-sve.c | 7 ++ | ||
12 | 4 files changed, 210 insertions(+), 56 deletions(-) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(sve2_uminp_zpzz_d, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_5(sve2_sqadd_zpzz_b, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_sqadd_zpzz_h, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_sqadd_zpzz_s, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_sqadd_zpzz_d, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_5(sve2_uqadd_zpzz_b, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_uqadd_zpzz_h, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(sve2_uqadd_zpzz_s, TCG_CALL_NO_RWG, | ||
36 | + void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_5(sve2_uqadd_zpzz_d, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | + | ||
40 | +DEF_HELPER_FLAGS_5(sve2_sqsub_zpzz_b, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | +DEF_HELPER_FLAGS_5(sve2_sqsub_zpzz_h, TCG_CALL_NO_RWG, | ||
43 | + void, ptr, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_5(sve2_sqsub_zpzz_s, TCG_CALL_NO_RWG, | ||
45 | + void, ptr, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_5(sve2_sqsub_zpzz_d, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, i32) | ||
48 | + | ||
49 | +DEF_HELPER_FLAGS_5(sve2_uqsub_zpzz_b, TCG_CALL_NO_RWG, | ||
50 | + void, ptr, ptr, ptr, ptr, i32) | ||
51 | +DEF_HELPER_FLAGS_5(sve2_uqsub_zpzz_h, TCG_CALL_NO_RWG, | ||
52 | + void, ptr, ptr, ptr, ptr, i32) | ||
53 | +DEF_HELPER_FLAGS_5(sve2_uqsub_zpzz_s, TCG_CALL_NO_RWG, | ||
54 | + void, ptr, ptr, ptr, ptr, i32) | ||
55 | +DEF_HELPER_FLAGS_5(sve2_uqsub_zpzz_d, TCG_CALL_NO_RWG, | ||
56 | + void, ptr, ptr, ptr, ptr, i32) | ||
57 | + | ||
58 | +DEF_HELPER_FLAGS_5(sve2_suqadd_zpzz_b, TCG_CALL_NO_RWG, | ||
59 | + void, ptr, ptr, ptr, ptr, i32) | ||
60 | +DEF_HELPER_FLAGS_5(sve2_suqadd_zpzz_h, TCG_CALL_NO_RWG, | ||
61 | + void, ptr, ptr, ptr, ptr, i32) | ||
62 | +DEF_HELPER_FLAGS_5(sve2_suqadd_zpzz_s, TCG_CALL_NO_RWG, | ||
63 | + void, ptr, ptr, ptr, ptr, i32) | ||
64 | +DEF_HELPER_FLAGS_5(sve2_suqadd_zpzz_d, TCG_CALL_NO_RWG, | ||
65 | + void, ptr, ptr, ptr, ptr, i32) | ||
66 | + | ||
67 | +DEF_HELPER_FLAGS_5(sve2_usqadd_zpzz_b, TCG_CALL_NO_RWG, | ||
68 | + void, ptr, ptr, ptr, ptr, i32) | ||
69 | +DEF_HELPER_FLAGS_5(sve2_usqadd_zpzz_h, TCG_CALL_NO_RWG, | ||
70 | + void, ptr, ptr, ptr, ptr, i32) | ||
71 | +DEF_HELPER_FLAGS_5(sve2_usqadd_zpzz_s, TCG_CALL_NO_RWG, | ||
72 | + void, ptr, ptr, ptr, ptr, i32) | ||
73 | +DEF_HELPER_FLAGS_5(sve2_usqadd_zpzz_d, TCG_CALL_NO_RWG, | ||
74 | + void, ptr, ptr, ptr, ptr, i32) | ||
75 | + | ||
76 | DEF_HELPER_FLAGS_5(sve_asr_zpzw_b, TCG_CALL_NO_RWG, | ||
77 | void, ptr, ptr, ptr, ptr, i32) | ||
78 | DEF_HELPER_FLAGS_5(sve_asr_zpzw_h, TCG_CALL_NO_RWG, | ||
79 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/sve.decode | ||
82 | +++ b/target/arm/sve.decode | ||
83 | @@ -XXX,XX +XXX,XX @@ SMAXP 01000100 .. 010 100 101 ... ..... ..... @rdn_pg_rm | ||
84 | UMAXP 01000100 .. 010 101 101 ... ..... ..... @rdn_pg_rm | ||
85 | SMINP 01000100 .. 010 110 101 ... ..... ..... @rdn_pg_rm | ||
86 | UMINP 01000100 .. 010 111 101 ... ..... ..... @rdn_pg_rm | ||
87 | + | ||
88 | +### SVE2 saturating add/subtract (predicated) | ||
89 | + | ||
90 | +SQADD_zpzz 01000100 .. 011 000 100 ... ..... ..... @rdn_pg_rm | ||
91 | +UQADD_zpzz 01000100 .. 011 001 100 ... ..... ..... @rdn_pg_rm | ||
92 | +SQSUB_zpzz 01000100 .. 011 010 100 ... ..... ..... @rdn_pg_rm | ||
93 | +UQSUB_zpzz 01000100 .. 011 011 100 ... ..... ..... @rdn_pg_rm | ||
94 | +SUQADD 01000100 .. 011 100 100 ... ..... ..... @rdn_pg_rm | ||
95 | +USQADD 01000100 .. 011 101 100 ... ..... ..... @rdn_pg_rm | ||
96 | +SQSUB_zpzz 01000100 .. 011 110 100 ... ..... ..... @rdm_pg_rn # SQSUBR | ||
97 | +UQSUB_zpzz 01000100 .. 011 111 100 ... ..... ..... @rdm_pg_rn # UQSUBR | ||
98 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
99 | index XXXXXXX..XXXXXXX 100644 | ||
100 | --- a/target/arm/sve_helper.c | ||
101 | +++ b/target/arm/sve_helper.c | ||
102 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ(sve2_uhsub_zpzz_h, uint16_t, H1_2, DO_HSUB_BHS) | ||
103 | DO_ZPZZ(sve2_uhsub_zpzz_s, uint32_t, H1_4, DO_HSUB_BHS) | ||
104 | DO_ZPZZ_D(sve2_uhsub_zpzz_d, uint64_t, DO_HSUB_D) | ||
105 | |||
106 | +static inline int32_t do_sat_bhs(int64_t val, int64_t min, int64_t max) | ||
107 | +{ | ||
108 | + return val >= max ? max : val <= min ? min : val; | ||
109 | +} | ||
110 | + | ||
111 | +#define DO_SQADD_B(n, m) do_sat_bhs((int64_t)n + m, INT8_MIN, INT8_MAX) | ||
112 | +#define DO_SQADD_H(n, m) do_sat_bhs((int64_t)n + m, INT16_MIN, INT16_MAX) | ||
113 | +#define DO_SQADD_S(n, m) do_sat_bhs((int64_t)n + m, INT32_MIN, INT32_MAX) | ||
114 | + | ||
115 | +static inline int64_t do_sqadd_d(int64_t n, int64_t m) | ||
116 | +{ | ||
117 | + int64_t r = n + m; | ||
118 | + if (((r ^ n) & ~(n ^ m)) < 0) { | ||
119 | + /* Signed overflow. */ | ||
120 | + return r < 0 ? INT64_MAX : INT64_MIN; | ||
121 | + } | ||
122 | + return r; | ||
123 | +} | ||
124 | + | ||
125 | +DO_ZPZZ(sve2_sqadd_zpzz_b, int8_t, H1, DO_SQADD_B) | ||
126 | +DO_ZPZZ(sve2_sqadd_zpzz_h, int16_t, H1_2, DO_SQADD_H) | ||
127 | +DO_ZPZZ(sve2_sqadd_zpzz_s, int32_t, H1_4, DO_SQADD_S) | ||
128 | +DO_ZPZZ_D(sve2_sqadd_zpzz_d, int64_t, do_sqadd_d) | ||
129 | + | ||
130 | +#define DO_UQADD_B(n, m) do_sat_bhs((int64_t)n + m, 0, UINT8_MAX) | ||
131 | +#define DO_UQADD_H(n, m) do_sat_bhs((int64_t)n + m, 0, UINT16_MAX) | ||
132 | +#define DO_UQADD_S(n, m) do_sat_bhs((int64_t)n + m, 0, UINT32_MAX) | ||
133 | + | ||
134 | +static inline uint64_t do_uqadd_d(uint64_t n, uint64_t m) | ||
135 | +{ | ||
136 | + uint64_t r = n + m; | ||
137 | + return r < n ? UINT64_MAX : r; | ||
138 | +} | ||
139 | + | ||
140 | +DO_ZPZZ(sve2_uqadd_zpzz_b, uint8_t, H1, DO_UQADD_B) | ||
141 | +DO_ZPZZ(sve2_uqadd_zpzz_h, uint16_t, H1_2, DO_UQADD_H) | ||
142 | +DO_ZPZZ(sve2_uqadd_zpzz_s, uint32_t, H1_4, DO_UQADD_S) | ||
143 | +DO_ZPZZ_D(sve2_uqadd_zpzz_d, uint64_t, do_uqadd_d) | ||
144 | + | ||
145 | +#define DO_SQSUB_B(n, m) do_sat_bhs((int64_t)n - m, INT8_MIN, INT8_MAX) | ||
146 | +#define DO_SQSUB_H(n, m) do_sat_bhs((int64_t)n - m, INT16_MIN, INT16_MAX) | ||
147 | +#define DO_SQSUB_S(n, m) do_sat_bhs((int64_t)n - m, INT32_MIN, INT32_MAX) | ||
148 | + | ||
149 | +static inline int64_t do_sqsub_d(int64_t n, int64_t m) | ||
150 | +{ | ||
151 | + int64_t r = n - m; | ||
152 | + if (((r ^ n) & (n ^ m)) < 0) { | ||
153 | + /* Signed overflow. */ | ||
154 | + return r < 0 ? INT64_MAX : INT64_MIN; | ||
155 | + } | ||
156 | + return r; | ||
157 | +} | ||
158 | + | ||
159 | +DO_ZPZZ(sve2_sqsub_zpzz_b, int8_t, H1, DO_SQSUB_B) | ||
160 | +DO_ZPZZ(sve2_sqsub_zpzz_h, int16_t, H1_2, DO_SQSUB_H) | ||
161 | +DO_ZPZZ(sve2_sqsub_zpzz_s, int32_t, H1_4, DO_SQSUB_S) | ||
162 | +DO_ZPZZ_D(sve2_sqsub_zpzz_d, int64_t, do_sqsub_d) | ||
163 | + | ||
164 | +#define DO_UQSUB_B(n, m) do_sat_bhs((int64_t)n - m, 0, UINT8_MAX) | ||
165 | +#define DO_UQSUB_H(n, m) do_sat_bhs((int64_t)n - m, 0, UINT16_MAX) | ||
166 | +#define DO_UQSUB_S(n, m) do_sat_bhs((int64_t)n - m, 0, UINT32_MAX) | ||
167 | + | ||
168 | +static inline uint64_t do_uqsub_d(uint64_t n, uint64_t m) | ||
169 | +{ | ||
170 | + return n > m ? n - m : 0; | ||
171 | +} | ||
172 | + | ||
173 | +DO_ZPZZ(sve2_uqsub_zpzz_b, uint8_t, H1, DO_UQSUB_B) | ||
174 | +DO_ZPZZ(sve2_uqsub_zpzz_h, uint16_t, H1_2, DO_UQSUB_H) | ||
175 | +DO_ZPZZ(sve2_uqsub_zpzz_s, uint32_t, H1_4, DO_UQSUB_S) | ||
176 | +DO_ZPZZ_D(sve2_uqsub_zpzz_d, uint64_t, do_uqsub_d) | ||
177 | + | ||
178 | +#define DO_SUQADD_B(n, m) \ | ||
179 | + do_sat_bhs((int64_t)(int8_t)n + m, INT8_MIN, INT8_MAX) | ||
180 | +#define DO_SUQADD_H(n, m) \ | ||
181 | + do_sat_bhs((int64_t)(int16_t)n + m, INT16_MIN, INT16_MAX) | ||
182 | +#define DO_SUQADD_S(n, m) \ | ||
183 | + do_sat_bhs((int64_t)(int32_t)n + m, INT32_MIN, INT32_MAX) | ||
184 | + | ||
185 | +static inline int64_t do_suqadd_d(int64_t n, uint64_t m) | ||
186 | +{ | ||
187 | + uint64_t r = n + m; | ||
188 | + | ||
189 | + if (n < 0) { | ||
190 | + /* Note that m - abs(n) cannot underflow. */ | ||
191 | + if (r > INT64_MAX) { | ||
192 | + /* Result is either very large positive or negative. */ | ||
193 | + if (m > -n) { | ||
194 | + /* m > abs(n), so r is a very large positive. */ | ||
195 | + return INT64_MAX; | ||
196 | + } | ||
197 | + /* Result is negative. */ | ||
198 | + } | ||
199 | + } else { | ||
200 | + /* Both inputs are positive: check for overflow. */ | ||
201 | + if (r < m || r > INT64_MAX) { | ||
202 | + return INT64_MAX; | ||
203 | + } | ||
204 | + } | ||
205 | + return r; | ||
206 | +} | ||
207 | + | ||
208 | +DO_ZPZZ(sve2_suqadd_zpzz_b, uint8_t, H1, DO_SUQADD_B) | ||
209 | +DO_ZPZZ(sve2_suqadd_zpzz_h, uint16_t, H1_2, DO_SUQADD_H) | ||
210 | +DO_ZPZZ(sve2_suqadd_zpzz_s, uint32_t, H1_4, DO_SUQADD_S) | ||
211 | +DO_ZPZZ_D(sve2_suqadd_zpzz_d, uint64_t, do_suqadd_d) | ||
212 | + | ||
213 | +#define DO_USQADD_B(n, m) \ | ||
214 | + do_sat_bhs((int64_t)n + (int8_t)m, 0, UINT8_MAX) | ||
215 | +#define DO_USQADD_H(n, m) \ | ||
216 | + do_sat_bhs((int64_t)n + (int16_t)m, 0, UINT16_MAX) | ||
217 | +#define DO_USQADD_S(n, m) \ | ||
218 | + do_sat_bhs((int64_t)n + (int32_t)m, 0, UINT32_MAX) | ||
219 | + | ||
220 | +static inline uint64_t do_usqadd_d(uint64_t n, int64_t m) | ||
221 | +{ | ||
222 | + uint64_t r = n + m; | ||
223 | + | ||
224 | + if (m < 0) { | ||
225 | + return n < -m ? 0 : r; | ||
226 | + } | ||
227 | + return r < n ? UINT64_MAX : r; | ||
228 | +} | ||
229 | + | ||
230 | +DO_ZPZZ(sve2_usqadd_zpzz_b, uint8_t, H1, DO_USQADD_B) | ||
231 | +DO_ZPZZ(sve2_usqadd_zpzz_h, uint16_t, H1_2, DO_USQADD_H) | ||
232 | +DO_ZPZZ(sve2_usqadd_zpzz_s, uint32_t, H1_4, DO_USQADD_S) | ||
233 | +DO_ZPZZ_D(sve2_usqadd_zpzz_d, uint64_t, do_usqadd_d) | ||
234 | + | ||
235 | #undef DO_ZPZZ | ||
236 | #undef DO_ZPZZ_D | ||
237 | |||
238 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_sqaddi_b)(void *d, void *a, int32_t b, uint32_t desc) | ||
239 | intptr_t i, oprsz = simd_oprsz(desc); | ||
240 | |||
241 | for (i = 0; i < oprsz; i += sizeof(int8_t)) { | ||
242 | - int r = *(int8_t *)(a + i) + b; | ||
243 | - if (r > INT8_MAX) { | ||
244 | - r = INT8_MAX; | ||
245 | - } else if (r < INT8_MIN) { | ||
246 | - r = INT8_MIN; | ||
247 | - } | ||
248 | - *(int8_t *)(d + i) = r; | ||
249 | + *(int8_t *)(d + i) = DO_SQADD_B(b, *(int8_t *)(a + i)); | ||
250 | } | ||
251 | } | ||
252 | |||
253 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_sqaddi_h)(void *d, void *a, int32_t b, uint32_t desc) | ||
254 | intptr_t i, oprsz = simd_oprsz(desc); | ||
255 | |||
256 | for (i = 0; i < oprsz; i += sizeof(int16_t)) { | ||
257 | - int r = *(int16_t *)(a + i) + b; | ||
258 | - if (r > INT16_MAX) { | ||
259 | - r = INT16_MAX; | ||
260 | - } else if (r < INT16_MIN) { | ||
261 | - r = INT16_MIN; | ||
262 | - } | ||
263 | - *(int16_t *)(d + i) = r; | ||
264 | + *(int16_t *)(d + i) = DO_SQADD_H(b, *(int16_t *)(a + i)); | ||
265 | } | ||
266 | } | ||
267 | |||
268 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_sqaddi_s)(void *d, void *a, int64_t b, uint32_t desc) | ||
269 | intptr_t i, oprsz = simd_oprsz(desc); | ||
270 | |||
271 | for (i = 0; i < oprsz; i += sizeof(int32_t)) { | ||
272 | - int64_t r = *(int32_t *)(a + i) + b; | ||
273 | - if (r > INT32_MAX) { | ||
274 | - r = INT32_MAX; | ||
275 | - } else if (r < INT32_MIN) { | ||
276 | - r = INT32_MIN; | ||
277 | - } | ||
278 | - *(int32_t *)(d + i) = r; | ||
279 | + *(int32_t *)(d + i) = DO_SQADD_S(b, *(int32_t *)(a + i)); | ||
280 | } | ||
281 | } | ||
282 | |||
283 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_sqaddi_d)(void *d, void *a, int64_t b, uint32_t desc) | ||
284 | intptr_t i, oprsz = simd_oprsz(desc); | ||
285 | |||
286 | for (i = 0; i < oprsz; i += sizeof(int64_t)) { | ||
287 | - int64_t ai = *(int64_t *)(a + i); | ||
288 | - int64_t r = ai + b; | ||
289 | - if (((r ^ ai) & ~(ai ^ b)) < 0) { | ||
290 | - /* Signed overflow. */ | ||
291 | - r = (r < 0 ? INT64_MAX : INT64_MIN); | ||
292 | - } | ||
293 | - *(int64_t *)(d + i) = r; | ||
294 | + *(int64_t *)(d + i) = do_sqadd_d(b, *(int64_t *)(a + i)); | ||
295 | } | ||
296 | } | ||
297 | |||
298 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_uqaddi_b)(void *d, void *a, int32_t b, uint32_t desc) | ||
299 | intptr_t i, oprsz = simd_oprsz(desc); | ||
300 | |||
301 | for (i = 0; i < oprsz; i += sizeof(uint8_t)) { | ||
302 | - int r = *(uint8_t *)(a + i) + b; | ||
303 | - if (r > UINT8_MAX) { | ||
304 | - r = UINT8_MAX; | ||
305 | - } else if (r < 0) { | ||
306 | - r = 0; | ||
307 | - } | ||
308 | - *(uint8_t *)(d + i) = r; | ||
309 | + *(uint8_t *)(d + i) = DO_UQADD_B(b, *(uint8_t *)(a + i)); | ||
310 | } | ||
311 | } | ||
312 | |||
313 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_uqaddi_h)(void *d, void *a, int32_t b, uint32_t desc) | ||
314 | intptr_t i, oprsz = simd_oprsz(desc); | ||
315 | |||
316 | for (i = 0; i < oprsz; i += sizeof(uint16_t)) { | ||
317 | - int r = *(uint16_t *)(a + i) + b; | ||
318 | - if (r > UINT16_MAX) { | ||
319 | - r = UINT16_MAX; | ||
320 | - } else if (r < 0) { | ||
321 | - r = 0; | ||
322 | - } | ||
323 | - *(uint16_t *)(d + i) = r; | ||
324 | + *(uint16_t *)(d + i) = DO_UQADD_H(b, *(uint16_t *)(a + i)); | ||
325 | } | ||
326 | } | ||
327 | |||
328 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_uqaddi_s)(void *d, void *a, int64_t b, uint32_t desc) | ||
329 | intptr_t i, oprsz = simd_oprsz(desc); | ||
330 | |||
331 | for (i = 0; i < oprsz; i += sizeof(uint32_t)) { | ||
332 | - int64_t r = *(uint32_t *)(a + i) + b; | ||
333 | - if (r > UINT32_MAX) { | ||
334 | - r = UINT32_MAX; | ||
335 | - } else if (r < 0) { | ||
336 | - r = 0; | ||
337 | - } | ||
338 | - *(uint32_t *)(d + i) = r; | ||
339 | + *(uint32_t *)(d + i) = DO_UQADD_S(b, *(uint32_t *)(a + i)); | ||
340 | } | ||
341 | } | ||
342 | |||
343 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_uqaddi_d)(void *d, void *a, uint64_t b, uint32_t desc) | ||
344 | intptr_t i, oprsz = simd_oprsz(desc); | ||
345 | |||
346 | for (i = 0; i < oprsz; i += sizeof(uint64_t)) { | ||
347 | - uint64_t r = *(uint64_t *)(a + i) + b; | ||
348 | - if (r < b) { | ||
349 | - r = UINT64_MAX; | ||
350 | - } | ||
351 | - *(uint64_t *)(d + i) = r; | ||
352 | + *(uint64_t *)(d + i) = do_uqadd_d(b, *(uint64_t *)(a + i)); | ||
353 | } | ||
354 | } | ||
355 | |||
356 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_uqsubi_d)(void *d, void *a, uint64_t b, uint32_t desc) | ||
357 | intptr_t i, oprsz = simd_oprsz(desc); | ||
358 | |||
359 | for (i = 0; i < oprsz; i += sizeof(uint64_t)) { | ||
360 | - uint64_t ai = *(uint64_t *)(a + i); | ||
361 | - *(uint64_t *)(d + i) = (ai < b ? 0 : ai - b); | ||
362 | + *(uint64_t *)(d + i) = do_uqsub_d(*(uint64_t *)(a + i), b); | ||
363 | } | ||
364 | } | ||
365 | |||
366 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
367 | index XXXXXXX..XXXXXXX 100644 | ||
368 | --- a/target/arm/translate-sve.c | ||
369 | +++ b/target/arm/translate-sve.c | ||
370 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ(SMAXP, smaxp) | ||
371 | DO_SVE2_ZPZZ(UMAXP, umaxp) | ||
372 | DO_SVE2_ZPZZ(SMINP, sminp) | ||
373 | DO_SVE2_ZPZZ(UMINP, uminp) | ||
374 | + | ||
375 | +DO_SVE2_ZPZZ(SQADD_zpzz, sqadd) | ||
376 | +DO_SVE2_ZPZZ(UQADD_zpzz, uqadd) | ||
377 | +DO_SVE2_ZPZZ(SQSUB_zpzz, sqsub) | ||
378 | +DO_SVE2_ZPZZ(UQSUB_zpzz, uqsub) | ||
379 | +DO_SVE2_ZPZZ(SUQADD, suqadd) | ||
380 | +DO_SVE2_ZPZZ(USQADD, usqadd) | ||
381 | -- | ||
382 | 2.20.1 | ||
383 | |||
384 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-11-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 24 ++++++++++++++++++++ | ||
9 | target/arm/sve.decode | 19 ++++++++++++++++ | ||
10 | target/arm/sve_helper.c | 43 +++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 46 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 132 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve_ftmad_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_5(sve_ftmad_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve_ftmad_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_4(sve2_saddl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_4(sve2_saddl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_saddl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | + | ||
26 | +DEF_HELPER_FLAGS_4(sve2_ssubl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(sve2_ssubl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_ssubl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_4(sve2_sabdl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_sabdl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_sabdl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_4(sve2_uaddl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_4(sve2_uaddl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_4(sve2_uaddl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | + | ||
38 | +DEF_HELPER_FLAGS_4(sve2_usubl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
39 | +DEF_HELPER_FLAGS_4(sve2_usubl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
40 | +DEF_HELPER_FLAGS_4(sve2_usubl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
41 | + | ||
42 | +DEF_HELPER_FLAGS_4(sve2_uabdl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
43 | +DEF_HELPER_FLAGS_4(sve2_uabdl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_4(sve2_uabdl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
45 | + | ||
46 | DEF_HELPER_FLAGS_4(sve_ld1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
47 | DEF_HELPER_FLAGS_4(sve_ld2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
48 | DEF_HELPER_FLAGS_4(sve_ld3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
49 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/target/arm/sve.decode | ||
52 | +++ b/target/arm/sve.decode | ||
53 | @@ -XXX,XX +XXX,XX @@ SUQADD 01000100 .. 011 100 100 ... ..... ..... @rdn_pg_rm | ||
54 | USQADD 01000100 .. 011 101 100 ... ..... ..... @rdn_pg_rm | ||
55 | SQSUB_zpzz 01000100 .. 011 110 100 ... ..... ..... @rdm_pg_rn # SQSUBR | ||
56 | UQSUB_zpzz 01000100 .. 011 111 100 ... ..... ..... @rdm_pg_rn # UQSUBR | ||
57 | + | ||
58 | +#### SVE2 Widening Integer Arithmetic | ||
59 | + | ||
60 | +## SVE2 integer add/subtract long | ||
61 | + | ||
62 | +SADDLB 01000101 .. 0 ..... 00 0000 ..... ..... @rd_rn_rm | ||
63 | +SADDLT 01000101 .. 0 ..... 00 0001 ..... ..... @rd_rn_rm | ||
64 | +UADDLB 01000101 .. 0 ..... 00 0010 ..... ..... @rd_rn_rm | ||
65 | +UADDLT 01000101 .. 0 ..... 00 0011 ..... ..... @rd_rn_rm | ||
66 | + | ||
67 | +SSUBLB 01000101 .. 0 ..... 00 0100 ..... ..... @rd_rn_rm | ||
68 | +SSUBLT 01000101 .. 0 ..... 00 0101 ..... ..... @rd_rn_rm | ||
69 | +USUBLB 01000101 .. 0 ..... 00 0110 ..... ..... @rd_rn_rm | ||
70 | +USUBLT 01000101 .. 0 ..... 00 0111 ..... ..... @rd_rn_rm | ||
71 | + | ||
72 | +SABDLB 01000101 .. 0 ..... 00 1100 ..... ..... @rd_rn_rm | ||
73 | +SABDLT 01000101 .. 0 ..... 00 1101 ..... ..... @rd_rn_rm | ||
74 | +UABDLB 01000101 .. 0 ..... 00 1110 ..... ..... @rd_rn_rm | ||
75 | +UABDLT 01000101 .. 0 ..... 00 1111 ..... ..... @rd_rn_rm | ||
76 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
77 | index XXXXXXX..XXXXXXX 100644 | ||
78 | --- a/target/arm/sve_helper.c | ||
79 | +++ b/target/arm/sve_helper.c | ||
80 | @@ -XXX,XX +XXX,XX @@ DO_ZZW(sve_lsl_zzw_s, uint32_t, uint64_t, H1_4, DO_LSL) | ||
81 | #undef DO_ZPZ | ||
82 | #undef DO_ZPZ_D | ||
83 | |||
84 | +/* | ||
85 | + * Three-operand expander, unpredicated, in which the two inputs are | ||
86 | + * selected from the top or bottom half of the wide column. | ||
87 | + */ | ||
88 | +#define DO_ZZZ_TB(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
89 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
90 | +{ \ | ||
91 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
92 | + int sel1 = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPEN); \ | ||
93 | + int sel2 = extract32(desc, SIMD_DATA_SHIFT + 1, 1) * sizeof(TYPEN); \ | ||
94 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
95 | + TYPEW nn = *(TYPEN *)(vn + HN(i + sel1)); \ | ||
96 | + TYPEW mm = *(TYPEN *)(vm + HN(i + sel2)); \ | ||
97 | + *(TYPEW *)(vd + HW(i)) = OP(nn, mm); \ | ||
98 | + } \ | ||
99 | +} | ||
100 | + | ||
101 | +DO_ZZZ_TB(sve2_saddl_h, int16_t, int8_t, H1_2, H1, DO_ADD) | ||
102 | +DO_ZZZ_TB(sve2_saddl_s, int32_t, int16_t, H1_4, H1_2, DO_ADD) | ||
103 | +DO_ZZZ_TB(sve2_saddl_d, int64_t, int32_t, , H1_4, DO_ADD) | ||
104 | + | ||
105 | +DO_ZZZ_TB(sve2_ssubl_h, int16_t, int8_t, H1_2, H1, DO_SUB) | ||
106 | +DO_ZZZ_TB(sve2_ssubl_s, int32_t, int16_t, H1_4, H1_2, DO_SUB) | ||
107 | +DO_ZZZ_TB(sve2_ssubl_d, int64_t, int32_t, , H1_4, DO_SUB) | ||
108 | + | ||
109 | +DO_ZZZ_TB(sve2_sabdl_h, int16_t, int8_t, H1_2, H1, DO_ABD) | ||
110 | +DO_ZZZ_TB(sve2_sabdl_s, int32_t, int16_t, H1_4, H1_2, DO_ABD) | ||
111 | +DO_ZZZ_TB(sve2_sabdl_d, int64_t, int32_t, , H1_4, DO_ABD) | ||
112 | + | ||
113 | +DO_ZZZ_TB(sve2_uaddl_h, uint16_t, uint8_t, H1_2, H1, DO_ADD) | ||
114 | +DO_ZZZ_TB(sve2_uaddl_s, uint32_t, uint16_t, H1_4, H1_2, DO_ADD) | ||
115 | +DO_ZZZ_TB(sve2_uaddl_d, uint64_t, uint32_t, , H1_4, DO_ADD) | ||
116 | + | ||
117 | +DO_ZZZ_TB(sve2_usubl_h, uint16_t, uint8_t, H1_2, H1, DO_SUB) | ||
118 | +DO_ZZZ_TB(sve2_usubl_s, uint32_t, uint16_t, H1_4, H1_2, DO_SUB) | ||
119 | +DO_ZZZ_TB(sve2_usubl_d, uint64_t, uint32_t, , H1_4, DO_SUB) | ||
120 | + | ||
121 | +DO_ZZZ_TB(sve2_uabdl_h, uint16_t, uint8_t, H1_2, H1, DO_ABD) | ||
122 | +DO_ZZZ_TB(sve2_uabdl_s, uint32_t, uint16_t, H1_4, H1_2, DO_ABD) | ||
123 | +DO_ZZZ_TB(sve2_uabdl_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
124 | + | ||
125 | +#undef DO_ZZZ_TB | ||
126 | + | ||
127 | /* Two-operand reduction expander, controlled by a predicate. | ||
128 | * The difference between TYPERED and TYPERET has to do with | ||
129 | * sign-extension. E.g. for SMAX, TYPERED must be signed, | ||
130 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
131 | index XXXXXXX..XXXXXXX 100644 | ||
132 | --- a/target/arm/translate-sve.c | ||
133 | +++ b/target/arm/translate-sve.c | ||
134 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ(SQSUB_zpzz, sqsub) | ||
135 | DO_SVE2_ZPZZ(UQSUB_zpzz, uqsub) | ||
136 | DO_SVE2_ZPZZ(SUQADD, suqadd) | ||
137 | DO_SVE2_ZPZZ(USQADD, usqadd) | ||
138 | + | ||
139 | +/* | ||
140 | + * SVE2 Widening Integer Arithmetic | ||
141 | + */ | ||
142 | + | ||
143 | +static bool do_sve2_zzw_ool(DisasContext *s, arg_rrr_esz *a, | ||
144 | + gen_helper_gvec_3 *fn, int data) | ||
145 | +{ | ||
146 | + if (fn == NULL || !dc_isar_feature(aa64_sve2, s)) { | ||
147 | + return false; | ||
148 | + } | ||
149 | + if (sve_access_check(s)) { | ||
150 | + unsigned vsz = vec_full_reg_size(s); | ||
151 | + tcg_gen_gvec_3_ool(vec_full_reg_offset(s, a->rd), | ||
152 | + vec_full_reg_offset(s, a->rn), | ||
153 | + vec_full_reg_offset(s, a->rm), | ||
154 | + vsz, vsz, data, fn); | ||
155 | + } | ||
156 | + return true; | ||
157 | +} | ||
158 | + | ||
159 | +#define DO_SVE2_ZZZ_TB(NAME, name, SEL1, SEL2) \ | ||
160 | +static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
161 | +{ \ | ||
162 | + static gen_helper_gvec_3 * const fns[4] = { \ | ||
163 | + NULL, gen_helper_sve2_##name##_h, \ | ||
164 | + gen_helper_sve2_##name##_s, gen_helper_sve2_##name##_d, \ | ||
165 | + }; \ | ||
166 | + return do_sve2_zzw_ool(s, a, fns[a->esz], (SEL2 << 1) | SEL1); \ | ||
167 | +} | ||
168 | + | ||
169 | +DO_SVE2_ZZZ_TB(SADDLB, saddl, false, false) | ||
170 | +DO_SVE2_ZZZ_TB(SSUBLB, ssubl, false, false) | ||
171 | +DO_SVE2_ZZZ_TB(SABDLB, sabdl, false, false) | ||
172 | + | ||
173 | +DO_SVE2_ZZZ_TB(UADDLB, uaddl, false, false) | ||
174 | +DO_SVE2_ZZZ_TB(USUBLB, usubl, false, false) | ||
175 | +DO_SVE2_ZZZ_TB(UABDLB, uabdl, false, false) | ||
176 | + | ||
177 | +DO_SVE2_ZZZ_TB(SADDLT, saddl, true, true) | ||
178 | +DO_SVE2_ZZZ_TB(SSUBLT, ssubl, true, true) | ||
179 | +DO_SVE2_ZZZ_TB(SABDLT, sabdl, true, true) | ||
180 | + | ||
181 | +DO_SVE2_ZZZ_TB(UADDLT, uaddl, true, true) | ||
182 | +DO_SVE2_ZZZ_TB(USUBLT, usubl, true, true) | ||
183 | +DO_SVE2_ZZZ_TB(UABDLT, uabdl, true, true) | ||
184 | -- | ||
185 | 2.20.1 | ||
186 | |||
187 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-12-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 6 ++++++ | ||
9 | target/arm/translate-sve.c | 4 ++++ | ||
10 | 2 files changed, 10 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ SABDLB 01000101 .. 0 ..... 00 1100 ..... ..... @rd_rn_rm | ||
17 | SABDLT 01000101 .. 0 ..... 00 1101 ..... ..... @rd_rn_rm | ||
18 | UABDLB 01000101 .. 0 ..... 00 1110 ..... ..... @rd_rn_rm | ||
19 | UABDLT 01000101 .. 0 ..... 00 1111 ..... ..... @rd_rn_rm | ||
20 | + | ||
21 | +## SVE2 integer add/subtract interleaved long | ||
22 | + | ||
23 | +SADDLBT 01000101 .. 0 ..... 1000 00 ..... ..... @rd_rn_rm | ||
24 | +SSUBLBT 01000101 .. 0 ..... 1000 10 ..... ..... @rd_rn_rm | ||
25 | +SSUBLTB 01000101 .. 0 ..... 1000 11 ..... ..... @rd_rn_rm | ||
26 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
27 | index XXXXXXX..XXXXXXX 100644 | ||
28 | --- a/target/arm/translate-sve.c | ||
29 | +++ b/target/arm/translate-sve.c | ||
30 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_TB(SABDLT, sabdl, true, true) | ||
31 | DO_SVE2_ZZZ_TB(UADDLT, uaddl, true, true) | ||
32 | DO_SVE2_ZZZ_TB(USUBLT, usubl, true, true) | ||
33 | DO_SVE2_ZZZ_TB(UABDLT, uabdl, true, true) | ||
34 | + | ||
35 | +DO_SVE2_ZZZ_TB(SADDLBT, saddl, false, true) | ||
36 | +DO_SVE2_ZZZ_TB(SSUBLBT, ssubl, false, true) | ||
37 | +DO_SVE2_ZZZ_TB(SSUBLTB, ssubl, true, false) | ||
38 | -- | ||
39 | 2.20.1 | ||
40 | |||
41 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-13-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 16 ++++++++++++++++ | ||
9 | target/arm/sve.decode | 12 ++++++++++++ | ||
10 | target/arm/sve_helper.c | 30 ++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 20 ++++++++++++++++++++ | ||
12 | 4 files changed, 78 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_uabdl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_4(sve2_uabdl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_uabdl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_4(sve2_saddw_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_4(sve2_saddw_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_saddw_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | + | ||
26 | +DEF_HELPER_FLAGS_4(sve2_ssubw_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(sve2_ssubw_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_ssubw_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_4(sve2_uaddw_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_uaddw_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_uaddw_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_4(sve2_usubw_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_4(sve2_usubw_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_4(sve2_usubw_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | + | ||
38 | DEF_HELPER_FLAGS_4(sve_ld1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
39 | DEF_HELPER_FLAGS_4(sve_ld2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
40 | DEF_HELPER_FLAGS_4(sve_ld3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32) | ||
41 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/target/arm/sve.decode | ||
44 | +++ b/target/arm/sve.decode | ||
45 | @@ -XXX,XX +XXX,XX @@ UABDLT 01000101 .. 0 ..... 00 1111 ..... ..... @rd_rn_rm | ||
46 | SADDLBT 01000101 .. 0 ..... 1000 00 ..... ..... @rd_rn_rm | ||
47 | SSUBLBT 01000101 .. 0 ..... 1000 10 ..... ..... @rd_rn_rm | ||
48 | SSUBLTB 01000101 .. 0 ..... 1000 11 ..... ..... @rd_rn_rm | ||
49 | + | ||
50 | +## SVE2 integer add/subtract wide | ||
51 | + | ||
52 | +SADDWB 01000101 .. 0 ..... 010 000 ..... ..... @rd_rn_rm | ||
53 | +SADDWT 01000101 .. 0 ..... 010 001 ..... ..... @rd_rn_rm | ||
54 | +UADDWB 01000101 .. 0 ..... 010 010 ..... ..... @rd_rn_rm | ||
55 | +UADDWT 01000101 .. 0 ..... 010 011 ..... ..... @rd_rn_rm | ||
56 | + | ||
57 | +SSUBWB 01000101 .. 0 ..... 010 100 ..... ..... @rd_rn_rm | ||
58 | +SSUBWT 01000101 .. 0 ..... 010 101 ..... ..... @rd_rn_rm | ||
59 | +USUBWB 01000101 .. 0 ..... 010 110 ..... ..... @rd_rn_rm | ||
60 | +USUBWT 01000101 .. 0 ..... 010 111 ..... ..... @rd_rn_rm | ||
61 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
62 | index XXXXXXX..XXXXXXX 100644 | ||
63 | --- a/target/arm/sve_helper.c | ||
64 | +++ b/target/arm/sve_helper.c | ||
65 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_TB(sve2_uabdl_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
66 | |||
67 | #undef DO_ZZZ_TB | ||
68 | |||
69 | +#define DO_ZZZ_WTB(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
70 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
71 | +{ \ | ||
72 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
73 | + int sel2 = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPEN); \ | ||
74 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
75 | + TYPEW nn = *(TYPEW *)(vn + HW(i)); \ | ||
76 | + TYPEW mm = *(TYPEN *)(vm + HN(i + sel2)); \ | ||
77 | + *(TYPEW *)(vd + HW(i)) = OP(nn, mm); \ | ||
78 | + } \ | ||
79 | +} | ||
80 | + | ||
81 | +DO_ZZZ_WTB(sve2_saddw_h, int16_t, int8_t, H1_2, H1, DO_ADD) | ||
82 | +DO_ZZZ_WTB(sve2_saddw_s, int32_t, int16_t, H1_4, H1_2, DO_ADD) | ||
83 | +DO_ZZZ_WTB(sve2_saddw_d, int64_t, int32_t, , H1_4, DO_ADD) | ||
84 | + | ||
85 | +DO_ZZZ_WTB(sve2_ssubw_h, int16_t, int8_t, H1_2, H1, DO_SUB) | ||
86 | +DO_ZZZ_WTB(sve2_ssubw_s, int32_t, int16_t, H1_4, H1_2, DO_SUB) | ||
87 | +DO_ZZZ_WTB(sve2_ssubw_d, int64_t, int32_t, , H1_4, DO_SUB) | ||
88 | + | ||
89 | +DO_ZZZ_WTB(sve2_uaddw_h, uint16_t, uint8_t, H1_2, H1, DO_ADD) | ||
90 | +DO_ZZZ_WTB(sve2_uaddw_s, uint32_t, uint16_t, H1_4, H1_2, DO_ADD) | ||
91 | +DO_ZZZ_WTB(sve2_uaddw_d, uint64_t, uint32_t, , H1_4, DO_ADD) | ||
92 | + | ||
93 | +DO_ZZZ_WTB(sve2_usubw_h, uint16_t, uint8_t, H1_2, H1, DO_SUB) | ||
94 | +DO_ZZZ_WTB(sve2_usubw_s, uint32_t, uint16_t, H1_4, H1_2, DO_SUB) | ||
95 | +DO_ZZZ_WTB(sve2_usubw_d, uint64_t, uint32_t, , H1_4, DO_SUB) | ||
96 | + | ||
97 | +#undef DO_ZZZ_WTB | ||
98 | + | ||
99 | /* Two-operand reduction expander, controlled by a predicate. | ||
100 | * The difference between TYPERED and TYPERET has to do with | ||
101 | * sign-extension. E.g. for SMAX, TYPERED must be signed, | ||
102 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
103 | index XXXXXXX..XXXXXXX 100644 | ||
104 | --- a/target/arm/translate-sve.c | ||
105 | +++ b/target/arm/translate-sve.c | ||
106 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_TB(UABDLT, uabdl, true, true) | ||
107 | DO_SVE2_ZZZ_TB(SADDLBT, saddl, false, true) | ||
108 | DO_SVE2_ZZZ_TB(SSUBLBT, ssubl, false, true) | ||
109 | DO_SVE2_ZZZ_TB(SSUBLTB, ssubl, true, false) | ||
110 | + | ||
111 | +#define DO_SVE2_ZZZ_WTB(NAME, name, SEL2) \ | ||
112 | +static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
113 | +{ \ | ||
114 | + static gen_helper_gvec_3 * const fns[4] = { \ | ||
115 | + NULL, gen_helper_sve2_##name##_h, \ | ||
116 | + gen_helper_sve2_##name##_s, gen_helper_sve2_##name##_d, \ | ||
117 | + }; \ | ||
118 | + return do_sve2_zzw_ool(s, a, fns[a->esz], SEL2); \ | ||
119 | +} | ||
120 | + | ||
121 | +DO_SVE2_ZZZ_WTB(SADDWB, saddw, false) | ||
122 | +DO_SVE2_ZZZ_WTB(SADDWT, saddw, true) | ||
123 | +DO_SVE2_ZZZ_WTB(SSUBWB, ssubw, false) | ||
124 | +DO_SVE2_ZZZ_WTB(SSUBWT, ssubw, true) | ||
125 | + | ||
126 | +DO_SVE2_ZZZ_WTB(UADDWB, uaddw, false) | ||
127 | +DO_SVE2_ZZZ_WTB(UADDWT, uaddw, true) | ||
128 | +DO_SVE2_ZZZ_WTB(USUBWB, usubw, false) | ||
129 | +DO_SVE2_ZZZ_WTB(USUBWT, usubw, true) | ||
130 | -- | ||
131 | 2.20.1 | ||
132 | |||
133 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Exclude PMULL from this category for the moment. | ||
4 | |||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-14-richard.henderson@linaro.org | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | --- | ||
10 | target/arm/helper-sve.h | 15 +++++++++++++++ | ||
11 | target/arm/sve.decode | 9 +++++++++ | ||
12 | target/arm/sve_helper.c | 31 +++++++++++++++++++++++++++++++ | ||
13 | target/arm/translate-sve.c | 9 +++++++++ | ||
14 | 4 files changed, 64 insertions(+) | ||
15 | |||
16 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/helper-sve.h | ||
19 | +++ b/target/arm/helper-sve.h | ||
20 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG, | ||
21 | DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG, | ||
22 | void, env, ptr, ptr, ptr, tl, i32) | ||
23 | |||
24 | +DEF_HELPER_FLAGS_4(sve2_sqdmull_zzz_h, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_sqdmull_zzz_s, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_sqdmull_zzz_d, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_4(sve2_smull_zzz_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_smull_zzz_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(sve2_smull_zzz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
34 | + | ||
35 | +DEF_HELPER_FLAGS_4(sve2_umull_zzz_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_4(sve2_umull_zzz_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_4(sve2_umull_zzz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
38 | + | ||
39 | DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
40 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
41 | index XXXXXXX..XXXXXXX 100644 | ||
42 | --- a/target/arm/sve.decode | ||
43 | +++ b/target/arm/sve.decode | ||
44 | @@ -XXX,XX +XXX,XX @@ SSUBWB 01000101 .. 0 ..... 010 100 ..... ..... @rd_rn_rm | ||
45 | SSUBWT 01000101 .. 0 ..... 010 101 ..... ..... @rd_rn_rm | ||
46 | USUBWB 01000101 .. 0 ..... 010 110 ..... ..... @rd_rn_rm | ||
47 | USUBWT 01000101 .. 0 ..... 010 111 ..... ..... @rd_rn_rm | ||
48 | + | ||
49 | +## SVE2 integer multiply long | ||
50 | + | ||
51 | +SQDMULLB_zzz 01000101 .. 0 ..... 011 000 ..... ..... @rd_rn_rm | ||
52 | +SQDMULLT_zzz 01000101 .. 0 ..... 011 001 ..... ..... @rd_rn_rm | ||
53 | +SMULLB_zzz 01000101 .. 0 ..... 011 100 ..... ..... @rd_rn_rm | ||
54 | +SMULLT_zzz 01000101 .. 0 ..... 011 101 ..... ..... @rd_rn_rm | ||
55 | +UMULLB_zzz 01000101 .. 0 ..... 011 110 ..... ..... @rd_rn_rm | ||
56 | +UMULLT_zzz 01000101 .. 0 ..... 011 111 ..... ..... @rd_rn_rm | ||
57 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | ||
59 | --- a/target/arm/sve_helper.c | ||
60 | +++ b/target/arm/sve_helper.c | ||
61 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_TB(sve2_uabdl_h, uint16_t, uint8_t, H1_2, H1, DO_ABD) | ||
62 | DO_ZZZ_TB(sve2_uabdl_s, uint32_t, uint16_t, H1_4, H1_2, DO_ABD) | ||
63 | DO_ZZZ_TB(sve2_uabdl_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
64 | |||
65 | +DO_ZZZ_TB(sve2_smull_zzz_h, int16_t, int8_t, H1_2, H1, DO_MUL) | ||
66 | +DO_ZZZ_TB(sve2_smull_zzz_s, int32_t, int16_t, H1_4, H1_2, DO_MUL) | ||
67 | +DO_ZZZ_TB(sve2_smull_zzz_d, int64_t, int32_t, , H1_4, DO_MUL) | ||
68 | + | ||
69 | +DO_ZZZ_TB(sve2_umull_zzz_h, uint16_t, uint8_t, H1_2, H1, DO_MUL) | ||
70 | +DO_ZZZ_TB(sve2_umull_zzz_s, uint32_t, uint16_t, H1_4, H1_2, DO_MUL) | ||
71 | +DO_ZZZ_TB(sve2_umull_zzz_d, uint64_t, uint32_t, , H1_4, DO_MUL) | ||
72 | + | ||
73 | +/* Note that the multiply cannot overflow, but the doubling can. */ | ||
74 | +static inline int16_t do_sqdmull_h(int16_t n, int16_t m) | ||
75 | +{ | ||
76 | + int16_t val = n * m; | ||
77 | + return DO_SQADD_H(val, val); | ||
78 | +} | ||
79 | + | ||
80 | +static inline int32_t do_sqdmull_s(int32_t n, int32_t m) | ||
81 | +{ | ||
82 | + int32_t val = n * m; | ||
83 | + return DO_SQADD_S(val, val); | ||
84 | +} | ||
85 | + | ||
86 | +static inline int64_t do_sqdmull_d(int64_t n, int64_t m) | ||
87 | +{ | ||
88 | + int64_t val = n * m; | ||
89 | + return do_sqadd_d(val, val); | ||
90 | +} | ||
91 | + | ||
92 | +DO_ZZZ_TB(sve2_sqdmull_zzz_h, int16_t, int8_t, H1_2, H1, do_sqdmull_h) | ||
93 | +DO_ZZZ_TB(sve2_sqdmull_zzz_s, int32_t, int16_t, H1_4, H1_2, do_sqdmull_s) | ||
94 | +DO_ZZZ_TB(sve2_sqdmull_zzz_d, int64_t, int32_t, , H1_4, do_sqdmull_d) | ||
95 | + | ||
96 | #undef DO_ZZZ_TB | ||
97 | |||
98 | #define DO_ZZZ_WTB(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
99 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
100 | index XXXXXXX..XXXXXXX 100644 | ||
101 | --- a/target/arm/translate-sve.c | ||
102 | +++ b/target/arm/translate-sve.c | ||
103 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_TB(SADDLBT, saddl, false, true) | ||
104 | DO_SVE2_ZZZ_TB(SSUBLBT, ssubl, false, true) | ||
105 | DO_SVE2_ZZZ_TB(SSUBLTB, ssubl, true, false) | ||
106 | |||
107 | +DO_SVE2_ZZZ_TB(SQDMULLB_zzz, sqdmull_zzz, false, false) | ||
108 | +DO_SVE2_ZZZ_TB(SQDMULLT_zzz, sqdmull_zzz, true, true) | ||
109 | + | ||
110 | +DO_SVE2_ZZZ_TB(SMULLB_zzz, smull_zzz, false, false) | ||
111 | +DO_SVE2_ZZZ_TB(SMULLT_zzz, smull_zzz, true, true) | ||
112 | + | ||
113 | +DO_SVE2_ZZZ_TB(UMULLB_zzz, umull_zzz, false, false) | ||
114 | +DO_SVE2_ZZZ_TB(UMULLT_zzz, umull_zzz, true, true) | ||
115 | + | ||
116 | #define DO_SVE2_ZZZ_WTB(NAME, name, SEL2) \ | ||
117 | static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
118 | { \ | ||
119 | -- | ||
120 | 2.20.1 | ||
121 | |||
122 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-15-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/cpu.h | 10 ++++++++++ | ||
9 | target/arm/helper-sve.h | 1 + | ||
10 | target/arm/sve.decode | 2 ++ | ||
11 | target/arm/translate-sve.c | 22 ++++++++++++++++++++++ | ||
12 | target/arm/vec_helper.c | 24 ++++++++++++++++++++++++ | ||
13 | 5 files changed, 59 insertions(+) | ||
14 | |||
15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/target/arm/cpu.h | ||
18 | +++ b/target/arm/cpu.h | ||
19 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2(const ARMISARegisters *id) | ||
20 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, SVEVER) != 0; | ||
21 | } | ||
22 | |||
23 | +static inline bool isar_feature_aa64_sve2_aes(const ARMISARegisters *id) | ||
24 | +{ | ||
25 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, AES) != 0; | ||
26 | +} | ||
27 | + | ||
28 | +static inline bool isar_feature_aa64_sve2_pmull128(const ARMISARegisters *id) | ||
29 | +{ | ||
30 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, AES) >= 2; | ||
31 | +} | ||
32 | + | ||
33 | /* | ||
34 | * Feature tests for "does this exist in either 32-bit or 64-bit?" | ||
35 | */ | ||
36 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/helper-sve.h | ||
39 | +++ b/target/arm/helper-sve.h | ||
40 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_umull_zzz_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
41 | DEF_HELPER_FLAGS_4(sve2_umull_zzz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
42 | |||
43 | DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_4(sve2_pmull_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
45 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
46 | index XXXXXXX..XXXXXXX 100644 | ||
47 | --- a/target/arm/sve.decode | ||
48 | +++ b/target/arm/sve.decode | ||
49 | @@ -XXX,XX +XXX,XX @@ USUBWT 01000101 .. 0 ..... 010 111 ..... ..... @rd_rn_rm | ||
50 | |||
51 | SQDMULLB_zzz 01000101 .. 0 ..... 011 000 ..... ..... @rd_rn_rm | ||
52 | SQDMULLT_zzz 01000101 .. 0 ..... 011 001 ..... ..... @rd_rn_rm | ||
53 | +PMULLB 01000101 .. 0 ..... 011 010 ..... ..... @rd_rn_rm | ||
54 | +PMULLT 01000101 .. 0 ..... 011 011 ..... ..... @rd_rn_rm | ||
55 | SMULLB_zzz 01000101 .. 0 ..... 011 100 ..... ..... @rd_rn_rm | ||
56 | SMULLT_zzz 01000101 .. 0 ..... 011 101 ..... ..... @rd_rn_rm | ||
57 | UMULLB_zzz 01000101 .. 0 ..... 011 110 ..... ..... @rd_rn_rm | ||
58 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/target/arm/translate-sve.c | ||
61 | +++ b/target/arm/translate-sve.c | ||
62 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_TB(SMULLT_zzz, smull_zzz, true, true) | ||
63 | DO_SVE2_ZZZ_TB(UMULLB_zzz, umull_zzz, false, false) | ||
64 | DO_SVE2_ZZZ_TB(UMULLT_zzz, umull_zzz, true, true) | ||
65 | |||
66 | +static bool do_trans_pmull(DisasContext *s, arg_rrr_esz *a, bool sel) | ||
67 | +{ | ||
68 | + static gen_helper_gvec_3 * const fns[4] = { | ||
69 | + gen_helper_gvec_pmull_q, gen_helper_sve2_pmull_h, | ||
70 | + NULL, gen_helper_sve2_pmull_d, | ||
71 | + }; | ||
72 | + if (a->esz == 0 && !dc_isar_feature(aa64_sve2_pmull128, s)) { | ||
73 | + return false; | ||
74 | + } | ||
75 | + return do_sve2_zzw_ool(s, a, fns[a->esz], sel); | ||
76 | +} | ||
77 | + | ||
78 | +static bool trans_PMULLB(DisasContext *s, arg_rrr_esz *a) | ||
79 | +{ | ||
80 | + return do_trans_pmull(s, a, false); | ||
81 | +} | ||
82 | + | ||
83 | +static bool trans_PMULLT(DisasContext *s, arg_rrr_esz *a) | ||
84 | +{ | ||
85 | + return do_trans_pmull(s, a, true); | ||
86 | +} | ||
87 | + | ||
88 | #define DO_SVE2_ZZZ_WTB(NAME, name, SEL2) \ | ||
89 | static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
90 | { \ | ||
91 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
92 | index XXXXXXX..XXXXXXX 100644 | ||
93 | --- a/target/arm/vec_helper.c | ||
94 | +++ b/target/arm/vec_helper.c | ||
95 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_pmull_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
96 | d[i] = pmull_h(nn, mm); | ||
97 | } | ||
98 | } | ||
99 | + | ||
100 | +static uint64_t pmull_d(uint64_t op1, uint64_t op2) | ||
101 | +{ | ||
102 | + uint64_t result = 0; | ||
103 | + int i; | ||
104 | + | ||
105 | + for (i = 0; i < 32; ++i) { | ||
106 | + uint64_t mask = -((op1 >> i) & 1); | ||
107 | + result ^= (op2 << i) & mask; | ||
108 | + } | ||
109 | + return result; | ||
110 | +} | ||
111 | + | ||
112 | +void HELPER(sve2_pmull_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
113 | +{ | ||
114 | + intptr_t sel = H4(simd_data(desc)); | ||
115 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
116 | + uint32_t *n = vn, *m = vm; | ||
117 | + uint64_t *d = vd; | ||
118 | + | ||
119 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
120 | + d[i] = pmull_d(n[2 * i + sel], m[2 * i + sel]); | ||
121 | + } | ||
122 | +} | ||
123 | #endif | ||
124 | |||
125 | #define DO_CMP0(NAME, TYPE, OP) \ | ||
126 | -- | ||
127 | 2.20.1 | ||
128 | |||
129 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-16-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 8 ++ | ||
9 | target/arm/sve.decode | 8 ++ | ||
10 | target/arm/sve_helper.c | 22 +++++ | ||
11 | target/arm/translate-sve.c | 159 +++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 197 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_umull_zzz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | |||
20 | DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_4(sve2_pmull_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_3(sve2_sshll_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_3(sve2_sshll_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_3(sve2_sshll_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
26 | + | ||
27 | +DEF_HELPER_FLAGS_3(sve2_ushll_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_3(sve2_ushll_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_3(sve2_ushll_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
30 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
31 | index XXXXXXX..XXXXXXX 100644 | ||
32 | --- a/target/arm/sve.decode | ||
33 | +++ b/target/arm/sve.decode | ||
34 | @@ -XXX,XX +XXX,XX @@ SMULLB_zzz 01000101 .. 0 ..... 011 100 ..... ..... @rd_rn_rm | ||
35 | SMULLT_zzz 01000101 .. 0 ..... 011 101 ..... ..... @rd_rn_rm | ||
36 | UMULLB_zzz 01000101 .. 0 ..... 011 110 ..... ..... @rd_rn_rm | ||
37 | UMULLT_zzz 01000101 .. 0 ..... 011 111 ..... ..... @rd_rn_rm | ||
38 | + | ||
39 | +## SVE2 bitwise shift left long | ||
40 | + | ||
41 | +# Note bit23 == 0 is handled by esz > 0 in do_sve2_shll_tb. | ||
42 | +SSHLLB 01000101 .. 0 ..... 1010 00 ..... ..... @rd_rn_tszimm_shl | ||
43 | +SSHLLT 01000101 .. 0 ..... 1010 01 ..... ..... @rd_rn_tszimm_shl | ||
44 | +USHLLB 01000101 .. 0 ..... 1010 10 ..... ..... @rd_rn_tszimm_shl | ||
45 | +USHLLT 01000101 .. 0 ..... 1010 11 ..... ..... @rd_rn_tszimm_shl | ||
46 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
47 | index XXXXXXX..XXXXXXX 100644 | ||
48 | --- a/target/arm/sve_helper.c | ||
49 | +++ b/target/arm/sve_helper.c | ||
50 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_WTB(sve2_usubw_d, uint64_t, uint32_t, , H1_4, DO_SUB) | ||
51 | |||
52 | #undef DO_ZZZ_WTB | ||
53 | |||
54 | +#define DO_ZZI_SHLL(NAME, TYPEW, TYPEN, HW, HN) \ | ||
55 | +void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
56 | +{ \ | ||
57 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
58 | + intptr_t sel = (simd_data(desc) & 1) * sizeof(TYPEN); \ | ||
59 | + int shift = simd_data(desc) >> 1; \ | ||
60 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
61 | + TYPEW nn = *(TYPEN *)(vn + HN(i + sel)); \ | ||
62 | + *(TYPEW *)(vd + HW(i)) = nn << shift; \ | ||
63 | + } \ | ||
64 | +} | ||
65 | + | ||
66 | +DO_ZZI_SHLL(sve2_sshll_h, int16_t, int8_t, H1_2, H1) | ||
67 | +DO_ZZI_SHLL(sve2_sshll_s, int32_t, int16_t, H1_4, H1_2) | ||
68 | +DO_ZZI_SHLL(sve2_sshll_d, int64_t, int32_t, , H1_4) | ||
69 | + | ||
70 | +DO_ZZI_SHLL(sve2_ushll_h, uint16_t, uint8_t, H1_2, H1) | ||
71 | +DO_ZZI_SHLL(sve2_ushll_s, uint32_t, uint16_t, H1_4, H1_2) | ||
72 | +DO_ZZI_SHLL(sve2_ushll_d, uint64_t, uint32_t, , H1_4) | ||
73 | + | ||
74 | +#undef DO_ZZI_SHLL | ||
75 | + | ||
76 | /* Two-operand reduction expander, controlled by a predicate. | ||
77 | * The difference between TYPERED and TYPERET has to do with | ||
78 | * sign-extension. E.g. for SMAX, TYPERED must be signed, | ||
79 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/translate-sve.c | ||
82 | +++ b/target/arm/translate-sve.c | ||
83 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_WTB(UADDWB, uaddw, false) | ||
84 | DO_SVE2_ZZZ_WTB(UADDWT, uaddw, true) | ||
85 | DO_SVE2_ZZZ_WTB(USUBWB, usubw, false) | ||
86 | DO_SVE2_ZZZ_WTB(USUBWT, usubw, true) | ||
87 | + | ||
88 | +static void gen_sshll_vec(unsigned vece, TCGv_vec d, TCGv_vec n, int64_t imm) | ||
89 | +{ | ||
90 | + int top = imm & 1; | ||
91 | + int shl = imm >> 1; | ||
92 | + int halfbits = 4 << vece; | ||
93 | + | ||
94 | + if (top) { | ||
95 | + if (shl == halfbits) { | ||
96 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
97 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(halfbits, halfbits)); | ||
98 | + tcg_gen_and_vec(vece, d, n, t); | ||
99 | + tcg_temp_free_vec(t); | ||
100 | + } else { | ||
101 | + tcg_gen_sari_vec(vece, d, n, halfbits); | ||
102 | + tcg_gen_shli_vec(vece, d, d, shl); | ||
103 | + } | ||
104 | + } else { | ||
105 | + tcg_gen_shli_vec(vece, d, n, halfbits); | ||
106 | + tcg_gen_sari_vec(vece, d, d, halfbits - shl); | ||
107 | + } | ||
108 | +} | ||
109 | + | ||
110 | +static void gen_ushll_i64(unsigned vece, TCGv_i64 d, TCGv_i64 n, int imm) | ||
111 | +{ | ||
112 | + int halfbits = 4 << vece; | ||
113 | + int top = imm & 1; | ||
114 | + int shl = (imm >> 1); | ||
115 | + int shift; | ||
116 | + uint64_t mask; | ||
117 | + | ||
118 | + mask = MAKE_64BIT_MASK(0, halfbits); | ||
119 | + mask <<= shl; | ||
120 | + mask = dup_const(vece, mask); | ||
121 | + | ||
122 | + shift = shl - top * halfbits; | ||
123 | + if (shift < 0) { | ||
124 | + tcg_gen_shri_i64(d, n, -shift); | ||
125 | + } else { | ||
126 | + tcg_gen_shli_i64(d, n, shift); | ||
127 | + } | ||
128 | + tcg_gen_andi_i64(d, d, mask); | ||
129 | +} | ||
130 | + | ||
131 | +static void gen_ushll16_i64(TCGv_i64 d, TCGv_i64 n, int64_t imm) | ||
132 | +{ | ||
133 | + gen_ushll_i64(MO_16, d, n, imm); | ||
134 | +} | ||
135 | + | ||
136 | +static void gen_ushll32_i64(TCGv_i64 d, TCGv_i64 n, int64_t imm) | ||
137 | +{ | ||
138 | + gen_ushll_i64(MO_32, d, n, imm); | ||
139 | +} | ||
140 | + | ||
141 | +static void gen_ushll64_i64(TCGv_i64 d, TCGv_i64 n, int64_t imm) | ||
142 | +{ | ||
143 | + gen_ushll_i64(MO_64, d, n, imm); | ||
144 | +} | ||
145 | + | ||
146 | +static void gen_ushll_vec(unsigned vece, TCGv_vec d, TCGv_vec n, int64_t imm) | ||
147 | +{ | ||
148 | + int halfbits = 4 << vece; | ||
149 | + int top = imm & 1; | ||
150 | + int shl = imm >> 1; | ||
151 | + | ||
152 | + if (top) { | ||
153 | + if (shl == halfbits) { | ||
154 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
155 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(halfbits, halfbits)); | ||
156 | + tcg_gen_and_vec(vece, d, n, t); | ||
157 | + tcg_temp_free_vec(t); | ||
158 | + } else { | ||
159 | + tcg_gen_shri_vec(vece, d, n, halfbits); | ||
160 | + tcg_gen_shli_vec(vece, d, d, shl); | ||
161 | + } | ||
162 | + } else { | ||
163 | + if (shl == 0) { | ||
164 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
165 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
166 | + tcg_gen_and_vec(vece, d, n, t); | ||
167 | + tcg_temp_free_vec(t); | ||
168 | + } else { | ||
169 | + tcg_gen_shli_vec(vece, d, n, halfbits); | ||
170 | + tcg_gen_shri_vec(vece, d, d, halfbits - shl); | ||
171 | + } | ||
172 | + } | ||
173 | +} | ||
174 | + | ||
175 | +static bool do_sve2_shll_tb(DisasContext *s, arg_rri_esz *a, | ||
176 | + bool sel, bool uns) | ||
177 | +{ | ||
178 | + static const TCGOpcode sshll_list[] = { | ||
179 | + INDEX_op_shli_vec, INDEX_op_sari_vec, 0 | ||
180 | + }; | ||
181 | + static const TCGOpcode ushll_list[] = { | ||
182 | + INDEX_op_shli_vec, INDEX_op_shri_vec, 0 | ||
183 | + }; | ||
184 | + static const GVecGen2i ops[2][3] = { | ||
185 | + { { .fniv = gen_sshll_vec, | ||
186 | + .opt_opc = sshll_list, | ||
187 | + .fno = gen_helper_sve2_sshll_h, | ||
188 | + .vece = MO_16 }, | ||
189 | + { .fniv = gen_sshll_vec, | ||
190 | + .opt_opc = sshll_list, | ||
191 | + .fno = gen_helper_sve2_sshll_s, | ||
192 | + .vece = MO_32 }, | ||
193 | + { .fniv = gen_sshll_vec, | ||
194 | + .opt_opc = sshll_list, | ||
195 | + .fno = gen_helper_sve2_sshll_d, | ||
196 | + .vece = MO_64 } }, | ||
197 | + { { .fni8 = gen_ushll16_i64, | ||
198 | + .fniv = gen_ushll_vec, | ||
199 | + .opt_opc = ushll_list, | ||
200 | + .fno = gen_helper_sve2_ushll_h, | ||
201 | + .vece = MO_16 }, | ||
202 | + { .fni8 = gen_ushll32_i64, | ||
203 | + .fniv = gen_ushll_vec, | ||
204 | + .opt_opc = ushll_list, | ||
205 | + .fno = gen_helper_sve2_ushll_s, | ||
206 | + .vece = MO_32 }, | ||
207 | + { .fni8 = gen_ushll64_i64, | ||
208 | + .fniv = gen_ushll_vec, | ||
209 | + .opt_opc = ushll_list, | ||
210 | + .fno = gen_helper_sve2_ushll_d, | ||
211 | + .vece = MO_64 } }, | ||
212 | + }; | ||
213 | + | ||
214 | + if (a->esz < 0 || a->esz > 2 || !dc_isar_feature(aa64_sve2, s)) { | ||
215 | + return false; | ||
216 | + } | ||
217 | + if (sve_access_check(s)) { | ||
218 | + unsigned vsz = vec_full_reg_size(s); | ||
219 | + tcg_gen_gvec_2i(vec_full_reg_offset(s, a->rd), | ||
220 | + vec_full_reg_offset(s, a->rn), | ||
221 | + vsz, vsz, (a->imm << 1) | sel, | ||
222 | + &ops[uns][a->esz]); | ||
223 | + } | ||
224 | + return true; | ||
225 | +} | ||
226 | + | ||
227 | +static bool trans_SSHLLB(DisasContext *s, arg_rri_esz *a) | ||
228 | +{ | ||
229 | + return do_sve2_shll_tb(s, a, false, false); | ||
230 | +} | ||
231 | + | ||
232 | +static bool trans_SSHLLT(DisasContext *s, arg_rri_esz *a) | ||
233 | +{ | ||
234 | + return do_sve2_shll_tb(s, a, true, false); | ||
235 | +} | ||
236 | + | ||
237 | +static bool trans_USHLLB(DisasContext *s, arg_rri_esz *a) | ||
238 | +{ | ||
239 | + return do_sve2_shll_tb(s, a, false, true); | ||
240 | +} | ||
241 | + | ||
242 | +static bool trans_USHLLT(DisasContext *s, arg_rri_esz *a) | ||
243 | +{ | ||
244 | + return do_sve2_shll_tb(s, a, true, true); | ||
245 | +} | ||
246 | -- | ||
247 | 2.20.1 | ||
248 | |||
249 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-17-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 5 +++++ | ||
9 | target/arm/sve.decode | 5 +++++ | ||
10 | target/arm/sve_helper.c | 20 ++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 19 +++++++++++++++++++ | ||
12 | 4 files changed, 49 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_sshll_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_3(sve2_ushll_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_3(sve2_ushll_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_3(sve2_ushll_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_4(sve2_eoril_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_eoril_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_eoril_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_eoril_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
28 | index XXXXXXX..XXXXXXX 100644 | ||
29 | --- a/target/arm/sve.decode | ||
30 | +++ b/target/arm/sve.decode | ||
31 | @@ -XXX,XX +XXX,XX @@ SSHLLB 01000101 .. 0 ..... 1010 00 ..... ..... @rd_rn_tszimm_shl | ||
32 | SSHLLT 01000101 .. 0 ..... 1010 01 ..... ..... @rd_rn_tszimm_shl | ||
33 | USHLLB 01000101 .. 0 ..... 1010 10 ..... ..... @rd_rn_tszimm_shl | ||
34 | USHLLT 01000101 .. 0 ..... 1010 11 ..... ..... @rd_rn_tszimm_shl | ||
35 | + | ||
36 | +## SVE2 bitwise exclusive-or interleaved | ||
37 | + | ||
38 | +EORBT 01000101 .. 0 ..... 10010 0 ..... ..... @rd_rn_rm | ||
39 | +EORTB 01000101 .. 0 ..... 10010 1 ..... ..... @rd_rn_rm | ||
40 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
41 | index XXXXXXX..XXXXXXX 100644 | ||
42 | --- a/target/arm/sve_helper.c | ||
43 | +++ b/target/arm/sve_helper.c | ||
44 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_WTB(sve2_usubw_d, uint64_t, uint32_t, , H1_4, DO_SUB) | ||
45 | |||
46 | #undef DO_ZZZ_WTB | ||
47 | |||
48 | +#define DO_ZZZ_NTB(NAME, TYPE, H, OP) \ | ||
49 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
50 | +{ \ | ||
51 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
52 | + intptr_t sel1 = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPE); \ | ||
53 | + intptr_t sel2 = extract32(desc, SIMD_DATA_SHIFT + 1, 1) * sizeof(TYPE); \ | ||
54 | + for (i = 0; i < opr_sz; i += 2 * sizeof(TYPE)) { \ | ||
55 | + TYPE nn = *(TYPE *)(vn + H(i + sel1)); \ | ||
56 | + TYPE mm = *(TYPE *)(vm + H(i + sel2)); \ | ||
57 | + *(TYPE *)(vd + H(i + sel1)) = OP(nn, mm); \ | ||
58 | + } \ | ||
59 | +} | ||
60 | + | ||
61 | +DO_ZZZ_NTB(sve2_eoril_b, uint8_t, H1, DO_EOR) | ||
62 | +DO_ZZZ_NTB(sve2_eoril_h, uint16_t, H1_2, DO_EOR) | ||
63 | +DO_ZZZ_NTB(sve2_eoril_s, uint32_t, H1_4, DO_EOR) | ||
64 | +DO_ZZZ_NTB(sve2_eoril_d, uint64_t, , DO_EOR) | ||
65 | + | ||
66 | +#undef DO_ZZZ_NTB | ||
67 | + | ||
68 | #define DO_ZZI_SHLL(NAME, TYPEW, TYPEN, HW, HN) \ | ||
69 | void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
70 | { \ | ||
71 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
72 | index XXXXXXX..XXXXXXX 100644 | ||
73 | --- a/target/arm/translate-sve.c | ||
74 | +++ b/target/arm/translate-sve.c | ||
75 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_TB(SMULLT_zzz, smull_zzz, true, true) | ||
76 | DO_SVE2_ZZZ_TB(UMULLB_zzz, umull_zzz, false, false) | ||
77 | DO_SVE2_ZZZ_TB(UMULLT_zzz, umull_zzz, true, true) | ||
78 | |||
79 | +static bool do_eor_tb(DisasContext *s, arg_rrr_esz *a, bool sel1) | ||
80 | +{ | ||
81 | + static gen_helper_gvec_3 * const fns[4] = { | ||
82 | + gen_helper_sve2_eoril_b, gen_helper_sve2_eoril_h, | ||
83 | + gen_helper_sve2_eoril_s, gen_helper_sve2_eoril_d, | ||
84 | + }; | ||
85 | + return do_sve2_zzw_ool(s, a, fns[a->esz], (!sel1 << 1) | sel1); | ||
86 | +} | ||
87 | + | ||
88 | +static bool trans_EORBT(DisasContext *s, arg_rrr_esz *a) | ||
89 | +{ | ||
90 | + return do_eor_tb(s, a, false); | ||
91 | +} | ||
92 | + | ||
93 | +static bool trans_EORTB(DisasContext *s, arg_rrr_esz *a) | ||
94 | +{ | ||
95 | + return do_eor_tb(s, a, true); | ||
96 | +} | ||
97 | + | ||
98 | static bool do_trans_pmull(DisasContext *s, arg_rrr_esz *a, bool sel) | ||
99 | { | ||
100 | static gen_helper_gvec_3 * const fns[4] = { | ||
101 | -- | ||
102 | 2.20.1 | ||
103 | |||
104 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-18-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/cpu.h | 5 +++ | ||
9 | target/arm/helper-sve.h | 15 ++++++++ | ||
10 | target/arm/sve.decode | 6 ++++ | ||
11 | target/arm/sve_helper.c | 73 ++++++++++++++++++++++++++++++++++++++ | ||
12 | target/arm/translate-sve.c | 36 +++++++++++++++++++ | ||
13 | 5 files changed, 135 insertions(+) | ||
14 | |||
15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/target/arm/cpu.h | ||
18 | +++ b/target/arm/cpu.h | ||
19 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2_pmull128(const ARMISARegisters *id) | ||
20 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, AES) >= 2; | ||
21 | } | ||
22 | |||
23 | +static inline bool isar_feature_aa64_sve2_bitperm(const ARMISARegisters *id) | ||
24 | +{ | ||
25 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, BITPERM) != 0; | ||
26 | +} | ||
27 | + | ||
28 | /* | ||
29 | * Feature tests for "does this exist in either 32-bit or 64-bit?" | ||
30 | */ | ||
31 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
32 | index XXXXXXX..XXXXXXX 100644 | ||
33 | --- a/target/arm/helper-sve.h | ||
34 | +++ b/target/arm/helper-sve.h | ||
35 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_eoril_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
36 | DEF_HELPER_FLAGS_4(sve2_eoril_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | DEF_HELPER_FLAGS_4(sve2_eoril_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
38 | DEF_HELPER_FLAGS_4(sve2_eoril_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
39 | + | ||
40 | +DEF_HELPER_FLAGS_4(sve2_bext_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_4(sve2_bext_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
42 | +DEF_HELPER_FLAGS_4(sve2_bext_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
43 | +DEF_HELPER_FLAGS_4(sve2_bext_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
44 | + | ||
45 | +DEF_HELPER_FLAGS_4(sve2_bdep_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_4(sve2_bdep_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
47 | +DEF_HELPER_FLAGS_4(sve2_bdep_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
48 | +DEF_HELPER_FLAGS_4(sve2_bdep_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
49 | + | ||
50 | +DEF_HELPER_FLAGS_4(sve2_bgrp_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
51 | +DEF_HELPER_FLAGS_4(sve2_bgrp_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
52 | +DEF_HELPER_FLAGS_4(sve2_bgrp_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
53 | +DEF_HELPER_FLAGS_4(sve2_bgrp_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
54 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
55 | index XXXXXXX..XXXXXXX 100644 | ||
56 | --- a/target/arm/sve.decode | ||
57 | +++ b/target/arm/sve.decode | ||
58 | @@ -XXX,XX +XXX,XX @@ USHLLT 01000101 .. 0 ..... 1010 11 ..... ..... @rd_rn_tszimm_shl | ||
59 | |||
60 | EORBT 01000101 .. 0 ..... 10010 0 ..... ..... @rd_rn_rm | ||
61 | EORTB 01000101 .. 0 ..... 10010 1 ..... ..... @rd_rn_rm | ||
62 | + | ||
63 | +## SVE2 bitwise permute | ||
64 | + | ||
65 | +BEXT 01000101 .. 0 ..... 1011 00 ..... ..... @rd_rn_rm | ||
66 | +BDEP 01000101 .. 0 ..... 1011 01 ..... ..... @rd_rn_rm | ||
67 | +BGRP 01000101 .. 0 ..... 1011 10 ..... ..... @rd_rn_rm | ||
68 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
69 | index XXXXXXX..XXXXXXX 100644 | ||
70 | --- a/target/arm/sve_helper.c | ||
71 | +++ b/target/arm/sve_helper.c | ||
72 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_NTB(sve2_eoril_d, uint64_t, , DO_EOR) | ||
73 | |||
74 | #undef DO_ZZZ_NTB | ||
75 | |||
76 | +#define DO_BITPERM(NAME, TYPE, OP) \ | ||
77 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
78 | +{ \ | ||
79 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
80 | + for (i = 0; i < opr_sz; i += sizeof(TYPE)) { \ | ||
81 | + TYPE nn = *(TYPE *)(vn + i); \ | ||
82 | + TYPE mm = *(TYPE *)(vm + i); \ | ||
83 | + *(TYPE *)(vd + i) = OP(nn, mm, sizeof(TYPE) * 8); \ | ||
84 | + } \ | ||
85 | +} | ||
86 | + | ||
87 | +static uint64_t bitextract(uint64_t data, uint64_t mask, int n) | ||
88 | +{ | ||
89 | + uint64_t res = 0; | ||
90 | + int db, rb = 0; | ||
91 | + | ||
92 | + for (db = 0; db < n; ++db) { | ||
93 | + if ((mask >> db) & 1) { | ||
94 | + res |= ((data >> db) & 1) << rb; | ||
95 | + ++rb; | ||
96 | + } | ||
97 | + } | ||
98 | + return res; | ||
99 | +} | ||
100 | + | ||
101 | +DO_BITPERM(sve2_bext_b, uint8_t, bitextract) | ||
102 | +DO_BITPERM(sve2_bext_h, uint16_t, bitextract) | ||
103 | +DO_BITPERM(sve2_bext_s, uint32_t, bitextract) | ||
104 | +DO_BITPERM(sve2_bext_d, uint64_t, bitextract) | ||
105 | + | ||
106 | +static uint64_t bitdeposit(uint64_t data, uint64_t mask, int n) | ||
107 | +{ | ||
108 | + uint64_t res = 0; | ||
109 | + int rb, db = 0; | ||
110 | + | ||
111 | + for (rb = 0; rb < n; ++rb) { | ||
112 | + if ((mask >> rb) & 1) { | ||
113 | + res |= ((data >> db) & 1) << rb; | ||
114 | + ++db; | ||
115 | + } | ||
116 | + } | ||
117 | + return res; | ||
118 | +} | ||
119 | + | ||
120 | +DO_BITPERM(sve2_bdep_b, uint8_t, bitdeposit) | ||
121 | +DO_BITPERM(sve2_bdep_h, uint16_t, bitdeposit) | ||
122 | +DO_BITPERM(sve2_bdep_s, uint32_t, bitdeposit) | ||
123 | +DO_BITPERM(sve2_bdep_d, uint64_t, bitdeposit) | ||
124 | + | ||
125 | +static uint64_t bitgroup(uint64_t data, uint64_t mask, int n) | ||
126 | +{ | ||
127 | + uint64_t resm = 0, resu = 0; | ||
128 | + int db, rbm = 0, rbu = 0; | ||
129 | + | ||
130 | + for (db = 0; db < n; ++db) { | ||
131 | + uint64_t val = (data >> db) & 1; | ||
132 | + if ((mask >> db) & 1) { | ||
133 | + resm |= val << rbm++; | ||
134 | + } else { | ||
135 | + resu |= val << rbu++; | ||
136 | + } | ||
137 | + } | ||
138 | + | ||
139 | + return resm | (resu << rbm); | ||
140 | +} | ||
141 | + | ||
142 | +DO_BITPERM(sve2_bgrp_b, uint8_t, bitgroup) | ||
143 | +DO_BITPERM(sve2_bgrp_h, uint16_t, bitgroup) | ||
144 | +DO_BITPERM(sve2_bgrp_s, uint32_t, bitgroup) | ||
145 | +DO_BITPERM(sve2_bgrp_d, uint64_t, bitgroup) | ||
146 | + | ||
147 | +#undef DO_BITPERM | ||
148 | + | ||
149 | #define DO_ZZI_SHLL(NAME, TYPEW, TYPEN, HW, HN) \ | ||
150 | void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
151 | { \ | ||
152 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
153 | index XXXXXXX..XXXXXXX 100644 | ||
154 | --- a/target/arm/translate-sve.c | ||
155 | +++ b/target/arm/translate-sve.c | ||
156 | @@ -XXX,XX +XXX,XX @@ static bool trans_USHLLT(DisasContext *s, arg_rri_esz *a) | ||
157 | { | ||
158 | return do_sve2_shll_tb(s, a, true, true); | ||
159 | } | ||
160 | + | ||
161 | +static bool trans_BEXT(DisasContext *s, arg_rrr_esz *a) | ||
162 | +{ | ||
163 | + static gen_helper_gvec_3 * const fns[4] = { | ||
164 | + gen_helper_sve2_bext_b, gen_helper_sve2_bext_h, | ||
165 | + gen_helper_sve2_bext_s, gen_helper_sve2_bext_d, | ||
166 | + }; | ||
167 | + if (!dc_isar_feature(aa64_sve2_bitperm, s)) { | ||
168 | + return false; | ||
169 | + } | ||
170 | + return do_sve2_zzw_ool(s, a, fns[a->esz], 0); | ||
171 | +} | ||
172 | + | ||
173 | +static bool trans_BDEP(DisasContext *s, arg_rrr_esz *a) | ||
174 | +{ | ||
175 | + static gen_helper_gvec_3 * const fns[4] = { | ||
176 | + gen_helper_sve2_bdep_b, gen_helper_sve2_bdep_h, | ||
177 | + gen_helper_sve2_bdep_s, gen_helper_sve2_bdep_d, | ||
178 | + }; | ||
179 | + if (!dc_isar_feature(aa64_sve2_bitperm, s)) { | ||
180 | + return false; | ||
181 | + } | ||
182 | + return do_sve2_zzw_ool(s, a, fns[a->esz], 0); | ||
183 | +} | ||
184 | + | ||
185 | +static bool trans_BGRP(DisasContext *s, arg_rrr_esz *a) | ||
186 | +{ | ||
187 | + static gen_helper_gvec_3 * const fns[4] = { | ||
188 | + gen_helper_sve2_bgrp_b, gen_helper_sve2_bgrp_h, | ||
189 | + gen_helper_sve2_bgrp_s, gen_helper_sve2_bgrp_d, | ||
190 | + }; | ||
191 | + if (!dc_isar_feature(aa64_sve2_bitperm, s)) { | ||
192 | + return false; | ||
193 | + } | ||
194 | + return do_sve2_zzw_ool(s, a, fns[a->esz], 0); | ||
195 | +} | ||
196 | -- | ||
197 | 2.20.1 | ||
198 | |||
199 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-19-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 10 +++++++++ | ||
9 | target/arm/sve.decode | 9 ++++++++ | ||
10 | target/arm/sve_helper.c | 42 ++++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 31 ++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 92 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_bgrp_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_4(sve2_bgrp_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_bgrp_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_4(sve2_bgrp_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_4(sve2_cadd_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_cadd_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_cadd_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_cadd_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | + | ||
28 | +DEF_HELPER_FLAGS_4(sve2_sqcadd_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_4(sve2_sqcadd_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_sqcadd_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_sqcadd_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
33 | index XXXXXXX..XXXXXXX 100644 | ||
34 | --- a/target/arm/sve.decode | ||
35 | +++ b/target/arm/sve.decode | ||
36 | @@ -XXX,XX +XXX,XX @@ EORTB 01000101 .. 0 ..... 10010 1 ..... ..... @rd_rn_rm | ||
37 | BEXT 01000101 .. 0 ..... 1011 00 ..... ..... @rd_rn_rm | ||
38 | BDEP 01000101 .. 0 ..... 1011 01 ..... ..... @rd_rn_rm | ||
39 | BGRP 01000101 .. 0 ..... 1011 10 ..... ..... @rd_rn_rm | ||
40 | + | ||
41 | +#### SVE2 Accumulate | ||
42 | + | ||
43 | +## SVE2 complex integer add | ||
44 | + | ||
45 | +CADD_rot90 01000101 .. 00000 0 11011 0 ..... ..... @rdn_rm | ||
46 | +CADD_rot270 01000101 .. 00000 0 11011 1 ..... ..... @rdn_rm | ||
47 | +SQCADD_rot90 01000101 .. 00000 1 11011 0 ..... ..... @rdn_rm | ||
48 | +SQCADD_rot270 01000101 .. 00000 1 11011 1 ..... ..... @rdn_rm | ||
49 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/target/arm/sve_helper.c | ||
52 | +++ b/target/arm/sve_helper.c | ||
53 | @@ -XXX,XX +XXX,XX @@ DO_BITPERM(sve2_bgrp_d, uint64_t, bitgroup) | ||
54 | |||
55 | #undef DO_BITPERM | ||
56 | |||
57 | +#define DO_CADD(NAME, TYPE, H, ADD_OP, SUB_OP) \ | ||
58 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
59 | +{ \ | ||
60 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
61 | + int sub_r = simd_data(desc); \ | ||
62 | + if (sub_r) { \ | ||
63 | + for (i = 0; i < opr_sz; i += 2 * sizeof(TYPE)) { \ | ||
64 | + TYPE acc_r = *(TYPE *)(vn + H(i)); \ | ||
65 | + TYPE acc_i = *(TYPE *)(vn + H(i + sizeof(TYPE))); \ | ||
66 | + TYPE el2_r = *(TYPE *)(vm + H(i)); \ | ||
67 | + TYPE el2_i = *(TYPE *)(vm + H(i + sizeof(TYPE))); \ | ||
68 | + acc_r = ADD_OP(acc_r, el2_i); \ | ||
69 | + acc_i = SUB_OP(acc_i, el2_r); \ | ||
70 | + *(TYPE *)(vd + H(i)) = acc_r; \ | ||
71 | + *(TYPE *)(vd + H(i + sizeof(TYPE))) = acc_i; \ | ||
72 | + } \ | ||
73 | + } else { \ | ||
74 | + for (i = 0; i < opr_sz; i += 2 * sizeof(TYPE)) { \ | ||
75 | + TYPE acc_r = *(TYPE *)(vn + H(i)); \ | ||
76 | + TYPE acc_i = *(TYPE *)(vn + H(i + sizeof(TYPE))); \ | ||
77 | + TYPE el2_r = *(TYPE *)(vm + H(i)); \ | ||
78 | + TYPE el2_i = *(TYPE *)(vm + H(i + sizeof(TYPE))); \ | ||
79 | + acc_r = SUB_OP(acc_r, el2_i); \ | ||
80 | + acc_i = ADD_OP(acc_i, el2_r); \ | ||
81 | + *(TYPE *)(vd + H(i)) = acc_r; \ | ||
82 | + *(TYPE *)(vd + H(i + sizeof(TYPE))) = acc_i; \ | ||
83 | + } \ | ||
84 | + } \ | ||
85 | +} | ||
86 | + | ||
87 | +DO_CADD(sve2_cadd_b, int8_t, H1, DO_ADD, DO_SUB) | ||
88 | +DO_CADD(sve2_cadd_h, int16_t, H1_2, DO_ADD, DO_SUB) | ||
89 | +DO_CADD(sve2_cadd_s, int32_t, H1_4, DO_ADD, DO_SUB) | ||
90 | +DO_CADD(sve2_cadd_d, int64_t, , DO_ADD, DO_SUB) | ||
91 | + | ||
92 | +DO_CADD(sve2_sqcadd_b, int8_t, H1, DO_SQADD_B, DO_SQSUB_B) | ||
93 | +DO_CADD(sve2_sqcadd_h, int16_t, H1_2, DO_SQADD_H, DO_SQSUB_H) | ||
94 | +DO_CADD(sve2_sqcadd_s, int32_t, H1_4, DO_SQADD_S, DO_SQSUB_S) | ||
95 | +DO_CADD(sve2_sqcadd_d, int64_t, , do_sqadd_d, do_sqsub_d) | ||
96 | + | ||
97 | +#undef DO_CADD | ||
98 | + | ||
99 | #define DO_ZZI_SHLL(NAME, TYPEW, TYPEN, HW, HN) \ | ||
100 | void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
101 | { \ | ||
102 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
103 | index XXXXXXX..XXXXXXX 100644 | ||
104 | --- a/target/arm/translate-sve.c | ||
105 | +++ b/target/arm/translate-sve.c | ||
106 | @@ -XXX,XX +XXX,XX @@ static bool trans_BGRP(DisasContext *s, arg_rrr_esz *a) | ||
107 | } | ||
108 | return do_sve2_zzw_ool(s, a, fns[a->esz], 0); | ||
109 | } | ||
110 | + | ||
111 | +static bool do_cadd(DisasContext *s, arg_rrr_esz *a, bool sq, bool rot) | ||
112 | +{ | ||
113 | + static gen_helper_gvec_3 * const fns[2][4] = { | ||
114 | + { gen_helper_sve2_cadd_b, gen_helper_sve2_cadd_h, | ||
115 | + gen_helper_sve2_cadd_s, gen_helper_sve2_cadd_d }, | ||
116 | + { gen_helper_sve2_sqcadd_b, gen_helper_sve2_sqcadd_h, | ||
117 | + gen_helper_sve2_sqcadd_s, gen_helper_sve2_sqcadd_d }, | ||
118 | + }; | ||
119 | + return do_sve2_zzw_ool(s, a, fns[sq][a->esz], rot); | ||
120 | +} | ||
121 | + | ||
122 | +static bool trans_CADD_rot90(DisasContext *s, arg_rrr_esz *a) | ||
123 | +{ | ||
124 | + return do_cadd(s, a, false, false); | ||
125 | +} | ||
126 | + | ||
127 | +static bool trans_CADD_rot270(DisasContext *s, arg_rrr_esz *a) | ||
128 | +{ | ||
129 | + return do_cadd(s, a, false, true); | ||
130 | +} | ||
131 | + | ||
132 | +static bool trans_SQCADD_rot90(DisasContext *s, arg_rrr_esz *a) | ||
133 | +{ | ||
134 | + return do_cadd(s, a, true, false); | ||
135 | +} | ||
136 | + | ||
137 | +static bool trans_SQCADD_rot270(DisasContext *s, arg_rrr_esz *a) | ||
138 | +{ | ||
139 | + return do_cadd(s, a, true, true); | ||
140 | +} | ||
141 | -- | ||
142 | 2.20.1 | ||
143 | |||
144 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-20-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 14 ++++++++++ | ||
9 | target/arm/sve.decode | 12 +++++++++ | ||
10 | target/arm/sve_helper.c | 23 ++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 55 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 104 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_sqcadd_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_4(sve2_sqcadd_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_sqcadd_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_4(sve2_sqcadd_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_sabal_h, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_sabal_s, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_sabal_d, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_5(sve2_uabal_h, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(sve2_uabal_s, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(sve2_uabal_d, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ | ||
41 | &rpr_s rd pg rn s | ||
42 | &rprr_s rd pg rn rm s | ||
43 | &rprr_esz rd pg rn rm esz | ||
44 | +&rrrr_esz rd ra rn rm esz | ||
45 | &rprrr_esz rd pg rn rm ra esz | ||
46 | &rpri_esz rd pg rn imm esz | ||
47 | &ptrue rd esz pat s | ||
48 | @@ -XXX,XX +XXX,XX @@ | ||
49 | @rdn_i8s ........ esz:2 ...... ... imm:s8 rd:5 \ | ||
50 | &rri_esz rn=%reg_movprfx | ||
51 | |||
52 | +# Four operand, vector element size | ||
53 | +@rda_rn_rm ........ esz:2 . rm:5 ... ... rn:5 rd:5 \ | ||
54 | + &rrrr_esz ra=%reg_movprfx | ||
55 | + | ||
56 | # Three operand with "memory" size, aka immediate left shift | ||
57 | @rd_rn_msz_rm ........ ... rm:5 .... imm:2 rn:5 rd:5 &rrri | ||
58 | |||
59 | @@ -XXX,XX +XXX,XX @@ CADD_rot90 01000101 .. 00000 0 11011 0 ..... ..... @rdn_rm | ||
60 | CADD_rot270 01000101 .. 00000 0 11011 1 ..... ..... @rdn_rm | ||
61 | SQCADD_rot90 01000101 .. 00000 1 11011 0 ..... ..... @rdn_rm | ||
62 | SQCADD_rot270 01000101 .. 00000 1 11011 1 ..... ..... @rdn_rm | ||
63 | + | ||
64 | +## SVE2 integer absolute difference and accumulate long | ||
65 | + | ||
66 | +SABALB 01000101 .. 0 ..... 1100 00 ..... ..... @rda_rn_rm | ||
67 | +SABALT 01000101 .. 0 ..... 1100 01 ..... ..... @rda_rn_rm | ||
68 | +UABALB 01000101 .. 0 ..... 1100 10 ..... ..... @rda_rn_rm | ||
69 | +UABALT 01000101 .. 0 ..... 1100 11 ..... ..... @rda_rn_rm | ||
70 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
71 | index XXXXXXX..XXXXXXX 100644 | ||
72 | --- a/target/arm/sve_helper.c | ||
73 | +++ b/target/arm/sve_helper.c | ||
74 | @@ -XXX,XX +XXX,XX @@ DO_ZZZ_NTB(sve2_eoril_d, uint64_t, , DO_EOR) | ||
75 | |||
76 | #undef DO_ZZZ_NTB | ||
77 | |||
78 | +#define DO_ZZZW_ACC(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
79 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
80 | +{ \ | ||
81 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
82 | + intptr_t sel1 = simd_data(desc) * sizeof(TYPEN); \ | ||
83 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
84 | + TYPEW nn = *(TYPEN *)(vn + HN(i + sel1)); \ | ||
85 | + TYPEW mm = *(TYPEN *)(vm + HN(i + sel1)); \ | ||
86 | + TYPEW aa = *(TYPEW *)(va + HW(i)); \ | ||
87 | + *(TYPEW *)(vd + HW(i)) = OP(nn, mm) + aa; \ | ||
88 | + } \ | ||
89 | +} | ||
90 | + | ||
91 | +DO_ZZZW_ACC(sve2_sabal_h, int16_t, int8_t, H1_2, H1, DO_ABD) | ||
92 | +DO_ZZZW_ACC(sve2_sabal_s, int32_t, int16_t, H1_4, H1_2, DO_ABD) | ||
93 | +DO_ZZZW_ACC(sve2_sabal_d, int64_t, int32_t, , H1_4, DO_ABD) | ||
94 | + | ||
95 | +DO_ZZZW_ACC(sve2_uabal_h, uint16_t, uint8_t, H1_2, H1, DO_ABD) | ||
96 | +DO_ZZZW_ACC(sve2_uabal_s, uint32_t, uint16_t, H1_4, H1_2, DO_ABD) | ||
97 | +DO_ZZZW_ACC(sve2_uabal_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
98 | + | ||
99 | +#undef DO_ZZZW_ACC | ||
100 | + | ||
101 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
102 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
103 | { \ | ||
104 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
105 | index XXXXXXX..XXXXXXX 100644 | ||
106 | --- a/target/arm/translate-sve.c | ||
107 | +++ b/target/arm/translate-sve.c | ||
108 | @@ -XXX,XX +XXX,XX @@ static void gen_gvec_ool_zzz(DisasContext *s, gen_helper_gvec_3 *fn, | ||
109 | vsz, vsz, data, fn); | ||
110 | } | ||
111 | |||
112 | +/* Invoke an out-of-line helper on 4 Zregs. */ | ||
113 | +static void gen_gvec_ool_zzzz(DisasContext *s, gen_helper_gvec_4 *fn, | ||
114 | + int rd, int rn, int rm, int ra, int data) | ||
115 | +{ | ||
116 | + unsigned vsz = vec_full_reg_size(s); | ||
117 | + tcg_gen_gvec_4_ool(vec_full_reg_offset(s, rd), | ||
118 | + vec_full_reg_offset(s, rn), | ||
119 | + vec_full_reg_offset(s, rm), | ||
120 | + vec_full_reg_offset(s, ra), | ||
121 | + vsz, vsz, data, fn); | ||
122 | +} | ||
123 | + | ||
124 | /* Invoke an out-of-line helper on 2 Zregs and a predicate. */ | ||
125 | static void gen_gvec_ool_zzp(DisasContext *s, gen_helper_gvec_3 *fn, | ||
126 | int rd, int rn, int pg, int data) | ||
127 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQCADD_rot270(DisasContext *s, arg_rrr_esz *a) | ||
128 | { | ||
129 | return do_cadd(s, a, true, true); | ||
130 | } | ||
131 | + | ||
132 | +static bool do_sve2_zzzz_ool(DisasContext *s, arg_rrrr_esz *a, | ||
133 | + gen_helper_gvec_4 *fn, int data) | ||
134 | +{ | ||
135 | + if (fn == NULL || !dc_isar_feature(aa64_sve2, s)) { | ||
136 | + return false; | ||
137 | + } | ||
138 | + if (sve_access_check(s)) { | ||
139 | + gen_gvec_ool_zzzz(s, fn, a->rd, a->rn, a->rm, a->ra, data); | ||
140 | + } | ||
141 | + return true; | ||
142 | +} | ||
143 | + | ||
144 | +static bool do_abal(DisasContext *s, arg_rrrr_esz *a, bool uns, bool sel) | ||
145 | +{ | ||
146 | + static gen_helper_gvec_4 * const fns[2][4] = { | ||
147 | + { NULL, gen_helper_sve2_sabal_h, | ||
148 | + gen_helper_sve2_sabal_s, gen_helper_sve2_sabal_d }, | ||
149 | + { NULL, gen_helper_sve2_uabal_h, | ||
150 | + gen_helper_sve2_uabal_s, gen_helper_sve2_uabal_d }, | ||
151 | + }; | ||
152 | + return do_sve2_zzzz_ool(s, a, fns[uns][a->esz], sel); | ||
153 | +} | ||
154 | + | ||
155 | +static bool trans_SABALB(DisasContext *s, arg_rrrr_esz *a) | ||
156 | +{ | ||
157 | + return do_abal(s, a, false, false); | ||
158 | +} | ||
159 | + | ||
160 | +static bool trans_SABALT(DisasContext *s, arg_rrrr_esz *a) | ||
161 | +{ | ||
162 | + return do_abal(s, a, false, true); | ||
163 | +} | ||
164 | + | ||
165 | +static bool trans_UABALB(DisasContext *s, arg_rrrr_esz *a) | ||
166 | +{ | ||
167 | + return do_abal(s, a, true, false); | ||
168 | +} | ||
169 | + | ||
170 | +static bool trans_UABALT(DisasContext *s, arg_rrrr_esz *a) | ||
171 | +{ | ||
172 | + return do_abal(s, a, true, true); | ||
173 | +} | ||
174 | -- | ||
175 | 2.20.1 | ||
176 | |||
177 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-21-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 3 +++ | ||
9 | target/arm/sve.decode | 6 ++++++ | ||
10 | target/arm/sve_helper.c | 34 ++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 23 +++++++++++++++++++++++ | ||
12 | 4 files changed, 66 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_uabal_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_uabal_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_adcl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_adcl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
25 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
26 | index XXXXXXX..XXXXXXX 100644 | ||
27 | --- a/target/arm/sve.decode | ||
28 | +++ b/target/arm/sve.decode | ||
29 | @@ -XXX,XX +XXX,XX @@ SABALB 01000101 .. 0 ..... 1100 00 ..... ..... @rda_rn_rm | ||
30 | SABALT 01000101 .. 0 ..... 1100 01 ..... ..... @rda_rn_rm | ||
31 | UABALB 01000101 .. 0 ..... 1100 10 ..... ..... @rda_rn_rm | ||
32 | UABALT 01000101 .. 0 ..... 1100 11 ..... ..... @rda_rn_rm | ||
33 | + | ||
34 | +## SVE2 integer add/subtract long with carry | ||
35 | + | ||
36 | +# ADC and SBC decoded via size in helper dispatch. | ||
37 | +ADCLB 01000101 .. 0 ..... 11010 0 ..... ..... @rda_rn_rm | ||
38 | +ADCLT 01000101 .. 0 ..... 11010 1 ..... ..... @rda_rn_rm | ||
39 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve_helper.c | ||
42 | +++ b/target/arm/sve_helper.c | ||
43 | @@ -XXX,XX +XXX,XX @@ DO_ZZZW_ACC(sve2_uabal_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
44 | |||
45 | #undef DO_ZZZW_ACC | ||
46 | |||
47 | +void HELPER(sve2_adcl_s)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
48 | +{ | ||
49 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
50 | + int sel = H4(extract32(desc, SIMD_DATA_SHIFT, 1)); | ||
51 | + uint32_t inv = -extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
52 | + uint32_t *a = va, *n = vn; | ||
53 | + uint64_t *d = vd, *m = vm; | ||
54 | + | ||
55 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
56 | + uint32_t e1 = a[2 * i + H4(0)]; | ||
57 | + uint32_t e2 = n[2 * i + sel] ^ inv; | ||
58 | + uint64_t c = extract64(m[i], 32, 1); | ||
59 | + /* Compute and store the entire 33-bit result at once. */ | ||
60 | + d[i] = c + e1 + e2; | ||
61 | + } | ||
62 | +} | ||
63 | + | ||
64 | +void HELPER(sve2_adcl_d)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
65 | +{ | ||
66 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
67 | + int sel = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
68 | + uint64_t inv = -(uint64_t)extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
69 | + uint64_t *d = vd, *a = va, *n = vn, *m = vm; | ||
70 | + | ||
71 | + for (i = 0; i < opr_sz / 8; i += 2) { | ||
72 | + Int128 e1 = int128_make64(a[i]); | ||
73 | + Int128 e2 = int128_make64(n[i + sel] ^ inv); | ||
74 | + Int128 c = int128_make64(m[i + 1] & 1); | ||
75 | + Int128 r = int128_add(int128_add(e1, e2), c); | ||
76 | + d[i + 0] = int128_getlo(r); | ||
77 | + d[i + 1] = int128_gethi(r); | ||
78 | + } | ||
79 | +} | ||
80 | + | ||
81 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
82 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
83 | { \ | ||
84 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
85 | index XXXXXXX..XXXXXXX 100644 | ||
86 | --- a/target/arm/translate-sve.c | ||
87 | +++ b/target/arm/translate-sve.c | ||
88 | @@ -XXX,XX +XXX,XX @@ static bool trans_UABALT(DisasContext *s, arg_rrrr_esz *a) | ||
89 | { | ||
90 | return do_abal(s, a, true, true); | ||
91 | } | ||
92 | + | ||
93 | +static bool do_adcl(DisasContext *s, arg_rrrr_esz *a, bool sel) | ||
94 | +{ | ||
95 | + static gen_helper_gvec_4 * const fns[2] = { | ||
96 | + gen_helper_sve2_adcl_s, | ||
97 | + gen_helper_sve2_adcl_d, | ||
98 | + }; | ||
99 | + /* | ||
100 | + * Note that in this case the ESZ field encodes both size and sign. | ||
101 | + * Split out 'subtract' into bit 1 of the data field for the helper. | ||
102 | + */ | ||
103 | + return do_sve2_zzzz_ool(s, a, fns[a->esz & 1], (a->esz & 2) | sel); | ||
104 | +} | ||
105 | + | ||
106 | +static bool trans_ADCLB(DisasContext *s, arg_rrrr_esz *a) | ||
107 | +{ | ||
108 | + return do_adcl(s, a, false); | ||
109 | +} | ||
110 | + | ||
111 | +static bool trans_ADCLT(DisasContext *s, arg_rrrr_esz *a) | ||
112 | +{ | ||
113 | + return do_adcl(s, a, true); | ||
114 | +} | ||
115 | -- | ||
116 | 2.20.1 | ||
117 | |||
118 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-22-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 8 ++++++++ | ||
9 | target/arm/translate-sve.c | 34 ++++++++++++++++++++++++++++++++++ | ||
10 | 2 files changed, 42 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ UABALT 01000101 .. 0 ..... 1100 11 ..... ..... @rda_rn_rm | ||
17 | # ADC and SBC decoded via size in helper dispatch. | ||
18 | ADCLB 01000101 .. 0 ..... 11010 0 ..... ..... @rda_rn_rm | ||
19 | ADCLT 01000101 .. 0 ..... 11010 1 ..... ..... @rda_rn_rm | ||
20 | + | ||
21 | +## SVE2 bitwise shift right and accumulate | ||
22 | + | ||
23 | +# TODO: Use @rda and %reg_movprfx here. | ||
24 | +SSRA 01000101 .. 0 ..... 1110 00 ..... ..... @rd_rn_tszimm_shr | ||
25 | +USRA 01000101 .. 0 ..... 1110 01 ..... ..... @rd_rn_tszimm_shr | ||
26 | +SRSRA 01000101 .. 0 ..... 1110 10 ..... ..... @rd_rn_tszimm_shr | ||
27 | +URSRA 01000101 .. 0 ..... 1110 11 ..... ..... @rd_rn_tszimm_shr | ||
28 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
29 | index XXXXXXX..XXXXXXX 100644 | ||
30 | --- a/target/arm/translate-sve.c | ||
31 | +++ b/target/arm/translate-sve.c | ||
32 | @@ -XXX,XX +XXX,XX @@ static bool trans_ADCLT(DisasContext *s, arg_rrrr_esz *a) | ||
33 | { | ||
34 | return do_adcl(s, a, true); | ||
35 | } | ||
36 | + | ||
37 | +static bool do_sve2_fn2i(DisasContext *s, arg_rri_esz *a, GVecGen2iFn *fn) | ||
38 | +{ | ||
39 | + if (a->esz < 0 || !dc_isar_feature(aa64_sve2, s)) { | ||
40 | + return false; | ||
41 | + } | ||
42 | + if (sve_access_check(s)) { | ||
43 | + unsigned vsz = vec_full_reg_size(s); | ||
44 | + unsigned rd_ofs = vec_full_reg_offset(s, a->rd); | ||
45 | + unsigned rn_ofs = vec_full_reg_offset(s, a->rn); | ||
46 | + fn(a->esz, rd_ofs, rn_ofs, a->imm, vsz, vsz); | ||
47 | + } | ||
48 | + return true; | ||
49 | +} | ||
50 | + | ||
51 | +static bool trans_SSRA(DisasContext *s, arg_rri_esz *a) | ||
52 | +{ | ||
53 | + return do_sve2_fn2i(s, a, gen_gvec_ssra); | ||
54 | +} | ||
55 | + | ||
56 | +static bool trans_USRA(DisasContext *s, arg_rri_esz *a) | ||
57 | +{ | ||
58 | + return do_sve2_fn2i(s, a, gen_gvec_usra); | ||
59 | +} | ||
60 | + | ||
61 | +static bool trans_SRSRA(DisasContext *s, arg_rri_esz *a) | ||
62 | +{ | ||
63 | + return do_sve2_fn2i(s, a, gen_gvec_srsra); | ||
64 | +} | ||
65 | + | ||
66 | +static bool trans_URSRA(DisasContext *s, arg_rri_esz *a) | ||
67 | +{ | ||
68 | + return do_sve2_fn2i(s, a, gen_gvec_ursra); | ||
69 | +} | ||
70 | -- | ||
71 | 2.20.1 | ||
72 | |||
73 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-23-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 5 +++++ | ||
9 | target/arm/translate-sve.c | 10 ++++++++++ | ||
10 | 2 files changed, 15 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ SSRA 01000101 .. 0 ..... 1110 00 ..... ..... @rd_rn_tszimm_shr | ||
17 | USRA 01000101 .. 0 ..... 1110 01 ..... ..... @rd_rn_tszimm_shr | ||
18 | SRSRA 01000101 .. 0 ..... 1110 10 ..... ..... @rd_rn_tszimm_shr | ||
19 | URSRA 01000101 .. 0 ..... 1110 11 ..... ..... @rd_rn_tszimm_shr | ||
20 | + | ||
21 | +## SVE2 bitwise shift and insert | ||
22 | + | ||
23 | +SRI 01000101 .. 0 ..... 11110 0 ..... ..... @rd_rn_tszimm_shr | ||
24 | +SLI 01000101 .. 0 ..... 11110 1 ..... ..... @rd_rn_tszimm_shl | ||
25 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
26 | index XXXXXXX..XXXXXXX 100644 | ||
27 | --- a/target/arm/translate-sve.c | ||
28 | +++ b/target/arm/translate-sve.c | ||
29 | @@ -XXX,XX +XXX,XX @@ static bool trans_URSRA(DisasContext *s, arg_rri_esz *a) | ||
30 | { | ||
31 | return do_sve2_fn2i(s, a, gen_gvec_ursra); | ||
32 | } | ||
33 | + | ||
34 | +static bool trans_SRI(DisasContext *s, arg_rri_esz *a) | ||
35 | +{ | ||
36 | + return do_sve2_fn2i(s, a, gen_gvec_sri); | ||
37 | +} | ||
38 | + | ||
39 | +static bool trans_SLI(DisasContext *s, arg_rri_esz *a) | ||
40 | +{ | ||
41 | + return do_sve2_fn2i(s, a, gen_gvec_sli); | ||
42 | +} | ||
43 | -- | ||
44 | 2.20.1 | ||
45 | |||
46 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-24-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 6 ++++++ | ||
9 | target/arm/translate-sve.c | 21 +++++++++++++++++++++ | ||
10 | 2 files changed, 27 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ URSRA 01000101 .. 0 ..... 1110 11 ..... ..... @rd_rn_tszimm_shr | ||
17 | |||
18 | SRI 01000101 .. 0 ..... 11110 0 ..... ..... @rd_rn_tszimm_shr | ||
19 | SLI 01000101 .. 0 ..... 11110 1 ..... ..... @rd_rn_tszimm_shl | ||
20 | + | ||
21 | +## SVE2 integer absolute difference and accumulate | ||
22 | + | ||
23 | +# TODO: Use @rda and %reg_movprfx here. | ||
24 | +SABA 01000101 .. 0 ..... 11111 0 ..... ..... @rd_rn_rm | ||
25 | +UABA 01000101 .. 0 ..... 11111 1 ..... ..... @rd_rn_rm | ||
26 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
27 | index XXXXXXX..XXXXXXX 100644 | ||
28 | --- a/target/arm/translate-sve.c | ||
29 | +++ b/target/arm/translate-sve.c | ||
30 | @@ -XXX,XX +XXX,XX @@ static bool trans_SLI(DisasContext *s, arg_rri_esz *a) | ||
31 | { | ||
32 | return do_sve2_fn2i(s, a, gen_gvec_sli); | ||
33 | } | ||
34 | + | ||
35 | +static bool do_sve2_fn_zzz(DisasContext *s, arg_rrr_esz *a, GVecGen3Fn *fn) | ||
36 | +{ | ||
37 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
38 | + return false; | ||
39 | + } | ||
40 | + if (sve_access_check(s)) { | ||
41 | + gen_gvec_fn_zzz(s, fn, a->esz, a->rd, a->rn, a->rm); | ||
42 | + } | ||
43 | + return true; | ||
44 | +} | ||
45 | + | ||
46 | +static bool trans_SABA(DisasContext *s, arg_rrr_esz *a) | ||
47 | +{ | ||
48 | + return do_sve2_fn_zzz(s, a, gen_gvec_saba); | ||
49 | +} | ||
50 | + | ||
51 | +static bool trans_UABA(DisasContext *s, arg_rrr_esz *a) | ||
52 | +{ | ||
53 | + return do_sve2_fn_zzz(s, a, gen_gvec_uaba); | ||
54 | +} | ||
55 | -- | ||
56 | 2.20.1 | ||
57 | |||
58 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-25-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 24 ++++ | ||
9 | target/arm/sve.decode | 12 ++ | ||
10 | target/arm/sve_helper.c | 56 +++++++++ | ||
11 | target/arm/translate-sve.c | 238 +++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 330 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_uabal_d, TCG_CALL_NO_RWG, | ||
19 | |||
20 | DEF_HELPER_FLAGS_5(sve2_adcl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_5(sve2_adcl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_3(sve2_sqxtnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_3(sve2_sqxtnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_3(sve2_sqxtnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
26 | + | ||
27 | +DEF_HELPER_FLAGS_3(sve2_uqxtnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_3(sve2_uqxtnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_3(sve2_uqxtnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_3(sve2_sqxtunb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_3(sve2_sqxtunb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_3(sve2_sqxtunb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
34 | + | ||
35 | +DEF_HELPER_FLAGS_3(sve2_sqxtnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_3(sve2_sqxtnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_3(sve2_sqxtnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
38 | + | ||
39 | +DEF_HELPER_FLAGS_3(sve2_uqxtnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
40 | +DEF_HELPER_FLAGS_3(sve2_uqxtnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_3(sve2_uqxtnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
42 | + | ||
43 | +DEF_HELPER_FLAGS_3(sve2_sqxtunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
44 | +DEF_HELPER_FLAGS_3(sve2_sqxtunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
45 | +DEF_HELPER_FLAGS_3(sve2_sqxtunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
46 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
47 | index XXXXXXX..XXXXXXX 100644 | ||
48 | --- a/target/arm/sve.decode | ||
49 | +++ b/target/arm/sve.decode | ||
50 | @@ -XXX,XX +XXX,XX @@ SLI 01000101 .. 0 ..... 11110 1 ..... ..... @rd_rn_tszimm_shl | ||
51 | # TODO: Use @rda and %reg_movprfx here. | ||
52 | SABA 01000101 .. 0 ..... 11111 0 ..... ..... @rd_rn_rm | ||
53 | UABA 01000101 .. 0 ..... 11111 1 ..... ..... @rd_rn_rm | ||
54 | + | ||
55 | +#### SVE2 Narrowing | ||
56 | + | ||
57 | +## SVE2 saturating extract narrow | ||
58 | + | ||
59 | +# Bits 23, 18-16 are zero, limited in the translator via esz < 3 & imm == 0. | ||
60 | +SQXTNB 01000101 .. 1 ..... 010 000 ..... ..... @rd_rn_tszimm_shl | ||
61 | +SQXTNT 01000101 .. 1 ..... 010 001 ..... ..... @rd_rn_tszimm_shl | ||
62 | +UQXTNB 01000101 .. 1 ..... 010 010 ..... ..... @rd_rn_tszimm_shl | ||
63 | +UQXTNT 01000101 .. 1 ..... 010 011 ..... ..... @rd_rn_tszimm_shl | ||
64 | +SQXTUNB 01000101 .. 1 ..... 010 100 ..... ..... @rd_rn_tszimm_shl | ||
65 | +SQXTUNT 01000101 .. 1 ..... 010 101 ..... ..... @rd_rn_tszimm_shl | ||
66 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
67 | index XXXXXXX..XXXXXXX 100644 | ||
68 | --- a/target/arm/sve_helper.c | ||
69 | +++ b/target/arm/sve_helper.c | ||
70 | @@ -XXX,XX +XXX,XX @@ DO_ZZZW_ACC(sve2_uabal_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
71 | |||
72 | #undef DO_ZZZW_ACC | ||
73 | |||
74 | +#define DO_XTNB(NAME, TYPE, OP) \ | ||
75 | +void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
76 | +{ \ | ||
77 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
78 | + for (i = 0; i < opr_sz; i += sizeof(TYPE)) { \ | ||
79 | + TYPE nn = *(TYPE *)(vn + i); \ | ||
80 | + nn = OP(nn) & MAKE_64BIT_MASK(0, sizeof(TYPE) * 4); \ | ||
81 | + *(TYPE *)(vd + i) = nn; \ | ||
82 | + } \ | ||
83 | +} | ||
84 | + | ||
85 | +#define DO_XTNT(NAME, TYPE, TYPEN, H, OP) \ | ||
86 | +void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
87 | +{ \ | ||
88 | + intptr_t i, opr_sz = simd_oprsz(desc), odd = H(sizeof(TYPEN)); \ | ||
89 | + for (i = 0; i < opr_sz; i += sizeof(TYPE)) { \ | ||
90 | + TYPE nn = *(TYPE *)(vn + i); \ | ||
91 | + *(TYPEN *)(vd + i + odd) = OP(nn); \ | ||
92 | + } \ | ||
93 | +} | ||
94 | + | ||
95 | +#define DO_SQXTN_H(n) do_sat_bhs(n, INT8_MIN, INT8_MAX) | ||
96 | +#define DO_SQXTN_S(n) do_sat_bhs(n, INT16_MIN, INT16_MAX) | ||
97 | +#define DO_SQXTN_D(n) do_sat_bhs(n, INT32_MIN, INT32_MAX) | ||
98 | + | ||
99 | +DO_XTNB(sve2_sqxtnb_h, int16_t, DO_SQXTN_H) | ||
100 | +DO_XTNB(sve2_sqxtnb_s, int32_t, DO_SQXTN_S) | ||
101 | +DO_XTNB(sve2_sqxtnb_d, int64_t, DO_SQXTN_D) | ||
102 | + | ||
103 | +DO_XTNT(sve2_sqxtnt_h, int16_t, int8_t, H1, DO_SQXTN_H) | ||
104 | +DO_XTNT(sve2_sqxtnt_s, int32_t, int16_t, H1_2, DO_SQXTN_S) | ||
105 | +DO_XTNT(sve2_sqxtnt_d, int64_t, int32_t, H1_4, DO_SQXTN_D) | ||
106 | + | ||
107 | +#define DO_UQXTN_H(n) do_sat_bhs(n, 0, UINT8_MAX) | ||
108 | +#define DO_UQXTN_S(n) do_sat_bhs(n, 0, UINT16_MAX) | ||
109 | +#define DO_UQXTN_D(n) do_sat_bhs(n, 0, UINT32_MAX) | ||
110 | + | ||
111 | +DO_XTNB(sve2_uqxtnb_h, uint16_t, DO_UQXTN_H) | ||
112 | +DO_XTNB(sve2_uqxtnb_s, uint32_t, DO_UQXTN_S) | ||
113 | +DO_XTNB(sve2_uqxtnb_d, uint64_t, DO_UQXTN_D) | ||
114 | + | ||
115 | +DO_XTNT(sve2_uqxtnt_h, uint16_t, uint8_t, H1, DO_UQXTN_H) | ||
116 | +DO_XTNT(sve2_uqxtnt_s, uint32_t, uint16_t, H1_2, DO_UQXTN_S) | ||
117 | +DO_XTNT(sve2_uqxtnt_d, uint64_t, uint32_t, H1_4, DO_UQXTN_D) | ||
118 | + | ||
119 | +DO_XTNB(sve2_sqxtunb_h, int16_t, DO_UQXTN_H) | ||
120 | +DO_XTNB(sve2_sqxtunb_s, int32_t, DO_UQXTN_S) | ||
121 | +DO_XTNB(sve2_sqxtunb_d, int64_t, DO_UQXTN_D) | ||
122 | + | ||
123 | +DO_XTNT(sve2_sqxtunt_h, int16_t, int8_t, H1, DO_UQXTN_H) | ||
124 | +DO_XTNT(sve2_sqxtunt_s, int32_t, int16_t, H1_2, DO_UQXTN_S) | ||
125 | +DO_XTNT(sve2_sqxtunt_d, int64_t, int32_t, H1_4, DO_UQXTN_D) | ||
126 | + | ||
127 | +#undef DO_XTNB | ||
128 | +#undef DO_XTNT | ||
129 | + | ||
130 | void HELPER(sve2_adcl_s)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
131 | { | ||
132 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
133 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
134 | index XXXXXXX..XXXXXXX 100644 | ||
135 | --- a/target/arm/translate-sve.c | ||
136 | +++ b/target/arm/translate-sve.c | ||
137 | @@ -XXX,XX +XXX,XX @@ static bool trans_UABA(DisasContext *s, arg_rrr_esz *a) | ||
138 | { | ||
139 | return do_sve2_fn_zzz(s, a, gen_gvec_uaba); | ||
140 | } | ||
141 | + | ||
142 | +static bool do_sve2_narrow_extract(DisasContext *s, arg_rri_esz *a, | ||
143 | + const GVecGen2 ops[3]) | ||
144 | +{ | ||
145 | + if (a->esz < 0 || a->esz > MO_32 || a->imm != 0 || | ||
146 | + !dc_isar_feature(aa64_sve2, s)) { | ||
147 | + return false; | ||
148 | + } | ||
149 | + if (sve_access_check(s)) { | ||
150 | + unsigned vsz = vec_full_reg_size(s); | ||
151 | + tcg_gen_gvec_2(vec_full_reg_offset(s, a->rd), | ||
152 | + vec_full_reg_offset(s, a->rn), | ||
153 | + vsz, vsz, &ops[a->esz]); | ||
154 | + } | ||
155 | + return true; | ||
156 | +} | ||
157 | + | ||
158 | +static const TCGOpcode sqxtn_list[] = { | ||
159 | + INDEX_op_shli_vec, INDEX_op_smin_vec, INDEX_op_smax_vec, 0 | ||
160 | +}; | ||
161 | + | ||
162 | +static void gen_sqxtnb_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
163 | +{ | ||
164 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
165 | + int halfbits = 4 << vece; | ||
166 | + int64_t mask = (1ull << halfbits) - 1; | ||
167 | + int64_t min = -1ull << (halfbits - 1); | ||
168 | + int64_t max = -min - 1; | ||
169 | + | ||
170 | + tcg_gen_dupi_vec(vece, t, min); | ||
171 | + tcg_gen_smax_vec(vece, d, n, t); | ||
172 | + tcg_gen_dupi_vec(vece, t, max); | ||
173 | + tcg_gen_smin_vec(vece, d, d, t); | ||
174 | + tcg_gen_dupi_vec(vece, t, mask); | ||
175 | + tcg_gen_and_vec(vece, d, d, t); | ||
176 | + tcg_temp_free_vec(t); | ||
177 | +} | ||
178 | + | ||
179 | +static bool trans_SQXTNB(DisasContext *s, arg_rri_esz *a) | ||
180 | +{ | ||
181 | + static const GVecGen2 ops[3] = { | ||
182 | + { .fniv = gen_sqxtnb_vec, | ||
183 | + .opt_opc = sqxtn_list, | ||
184 | + .fno = gen_helper_sve2_sqxtnb_h, | ||
185 | + .vece = MO_16 }, | ||
186 | + { .fniv = gen_sqxtnb_vec, | ||
187 | + .opt_opc = sqxtn_list, | ||
188 | + .fno = gen_helper_sve2_sqxtnb_s, | ||
189 | + .vece = MO_32 }, | ||
190 | + { .fniv = gen_sqxtnb_vec, | ||
191 | + .opt_opc = sqxtn_list, | ||
192 | + .fno = gen_helper_sve2_sqxtnb_d, | ||
193 | + .vece = MO_64 }, | ||
194 | + }; | ||
195 | + return do_sve2_narrow_extract(s, a, ops); | ||
196 | +} | ||
197 | + | ||
198 | +static void gen_sqxtnt_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
199 | +{ | ||
200 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
201 | + int halfbits = 4 << vece; | ||
202 | + int64_t mask = (1ull << halfbits) - 1; | ||
203 | + int64_t min = -1ull << (halfbits - 1); | ||
204 | + int64_t max = -min - 1; | ||
205 | + | ||
206 | + tcg_gen_dupi_vec(vece, t, min); | ||
207 | + tcg_gen_smax_vec(vece, n, n, t); | ||
208 | + tcg_gen_dupi_vec(vece, t, max); | ||
209 | + tcg_gen_smin_vec(vece, n, n, t); | ||
210 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
211 | + tcg_gen_dupi_vec(vece, t, mask); | ||
212 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
213 | + tcg_temp_free_vec(t); | ||
214 | +} | ||
215 | + | ||
216 | +static bool trans_SQXTNT(DisasContext *s, arg_rri_esz *a) | ||
217 | +{ | ||
218 | + static const GVecGen2 ops[3] = { | ||
219 | + { .fniv = gen_sqxtnt_vec, | ||
220 | + .opt_opc = sqxtn_list, | ||
221 | + .load_dest = true, | ||
222 | + .fno = gen_helper_sve2_sqxtnt_h, | ||
223 | + .vece = MO_16 }, | ||
224 | + { .fniv = gen_sqxtnt_vec, | ||
225 | + .opt_opc = sqxtn_list, | ||
226 | + .load_dest = true, | ||
227 | + .fno = gen_helper_sve2_sqxtnt_s, | ||
228 | + .vece = MO_32 }, | ||
229 | + { .fniv = gen_sqxtnt_vec, | ||
230 | + .opt_opc = sqxtn_list, | ||
231 | + .load_dest = true, | ||
232 | + .fno = gen_helper_sve2_sqxtnt_d, | ||
233 | + .vece = MO_64 }, | ||
234 | + }; | ||
235 | + return do_sve2_narrow_extract(s, a, ops); | ||
236 | +} | ||
237 | + | ||
238 | +static const TCGOpcode uqxtn_list[] = { | ||
239 | + INDEX_op_shli_vec, INDEX_op_umin_vec, 0 | ||
240 | +}; | ||
241 | + | ||
242 | +static void gen_uqxtnb_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
243 | +{ | ||
244 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
245 | + int halfbits = 4 << vece; | ||
246 | + int64_t max = (1ull << halfbits) - 1; | ||
247 | + | ||
248 | + tcg_gen_dupi_vec(vece, t, max); | ||
249 | + tcg_gen_umin_vec(vece, d, n, t); | ||
250 | + tcg_temp_free_vec(t); | ||
251 | +} | ||
252 | + | ||
253 | +static bool trans_UQXTNB(DisasContext *s, arg_rri_esz *a) | ||
254 | +{ | ||
255 | + static const GVecGen2 ops[3] = { | ||
256 | + { .fniv = gen_uqxtnb_vec, | ||
257 | + .opt_opc = uqxtn_list, | ||
258 | + .fno = gen_helper_sve2_uqxtnb_h, | ||
259 | + .vece = MO_16 }, | ||
260 | + { .fniv = gen_uqxtnb_vec, | ||
261 | + .opt_opc = uqxtn_list, | ||
262 | + .fno = gen_helper_sve2_uqxtnb_s, | ||
263 | + .vece = MO_32 }, | ||
264 | + { .fniv = gen_uqxtnb_vec, | ||
265 | + .opt_opc = uqxtn_list, | ||
266 | + .fno = gen_helper_sve2_uqxtnb_d, | ||
267 | + .vece = MO_64 }, | ||
268 | + }; | ||
269 | + return do_sve2_narrow_extract(s, a, ops); | ||
270 | +} | ||
271 | + | ||
272 | +static void gen_uqxtnt_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
273 | +{ | ||
274 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
275 | + int halfbits = 4 << vece; | ||
276 | + int64_t max = (1ull << halfbits) - 1; | ||
277 | + | ||
278 | + tcg_gen_dupi_vec(vece, t, max); | ||
279 | + tcg_gen_umin_vec(vece, n, n, t); | ||
280 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
281 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
282 | + tcg_temp_free_vec(t); | ||
283 | +} | ||
284 | + | ||
285 | +static bool trans_UQXTNT(DisasContext *s, arg_rri_esz *a) | ||
286 | +{ | ||
287 | + static const GVecGen2 ops[3] = { | ||
288 | + { .fniv = gen_uqxtnt_vec, | ||
289 | + .opt_opc = uqxtn_list, | ||
290 | + .load_dest = true, | ||
291 | + .fno = gen_helper_sve2_uqxtnt_h, | ||
292 | + .vece = MO_16 }, | ||
293 | + { .fniv = gen_uqxtnt_vec, | ||
294 | + .opt_opc = uqxtn_list, | ||
295 | + .load_dest = true, | ||
296 | + .fno = gen_helper_sve2_uqxtnt_s, | ||
297 | + .vece = MO_32 }, | ||
298 | + { .fniv = gen_uqxtnt_vec, | ||
299 | + .opt_opc = uqxtn_list, | ||
300 | + .load_dest = true, | ||
301 | + .fno = gen_helper_sve2_uqxtnt_d, | ||
302 | + .vece = MO_64 }, | ||
303 | + }; | ||
304 | + return do_sve2_narrow_extract(s, a, ops); | ||
305 | +} | ||
306 | + | ||
307 | +static const TCGOpcode sqxtun_list[] = { | ||
308 | + INDEX_op_shli_vec, INDEX_op_umin_vec, INDEX_op_smax_vec, 0 | ||
309 | +}; | ||
310 | + | ||
311 | +static void gen_sqxtunb_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
312 | +{ | ||
313 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
314 | + int halfbits = 4 << vece; | ||
315 | + int64_t max = (1ull << halfbits) - 1; | ||
316 | + | ||
317 | + tcg_gen_dupi_vec(vece, t, 0); | ||
318 | + tcg_gen_smax_vec(vece, d, n, t); | ||
319 | + tcg_gen_dupi_vec(vece, t, max); | ||
320 | + tcg_gen_umin_vec(vece, d, d, t); | ||
321 | + tcg_temp_free_vec(t); | ||
322 | +} | ||
323 | + | ||
324 | +static bool trans_SQXTUNB(DisasContext *s, arg_rri_esz *a) | ||
325 | +{ | ||
326 | + static const GVecGen2 ops[3] = { | ||
327 | + { .fniv = gen_sqxtunb_vec, | ||
328 | + .opt_opc = sqxtun_list, | ||
329 | + .fno = gen_helper_sve2_sqxtunb_h, | ||
330 | + .vece = MO_16 }, | ||
331 | + { .fniv = gen_sqxtunb_vec, | ||
332 | + .opt_opc = sqxtun_list, | ||
333 | + .fno = gen_helper_sve2_sqxtunb_s, | ||
334 | + .vece = MO_32 }, | ||
335 | + { .fniv = gen_sqxtunb_vec, | ||
336 | + .opt_opc = sqxtun_list, | ||
337 | + .fno = gen_helper_sve2_sqxtunb_d, | ||
338 | + .vece = MO_64 }, | ||
339 | + }; | ||
340 | + return do_sve2_narrow_extract(s, a, ops); | ||
341 | +} | ||
342 | + | ||
343 | +static void gen_sqxtunt_vec(unsigned vece, TCGv_vec d, TCGv_vec n) | ||
344 | +{ | ||
345 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
346 | + int halfbits = 4 << vece; | ||
347 | + int64_t max = (1ull << halfbits) - 1; | ||
348 | + | ||
349 | + tcg_gen_dupi_vec(vece, t, 0); | ||
350 | + tcg_gen_smax_vec(vece, n, n, t); | ||
351 | + tcg_gen_dupi_vec(vece, t, max); | ||
352 | + tcg_gen_umin_vec(vece, n, n, t); | ||
353 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
354 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
355 | + tcg_temp_free_vec(t); | ||
356 | +} | ||
357 | + | ||
358 | +static bool trans_SQXTUNT(DisasContext *s, arg_rri_esz *a) | ||
359 | +{ | ||
360 | + static const GVecGen2 ops[3] = { | ||
361 | + { .fniv = gen_sqxtunt_vec, | ||
362 | + .opt_opc = sqxtun_list, | ||
363 | + .load_dest = true, | ||
364 | + .fno = gen_helper_sve2_sqxtunt_h, | ||
365 | + .vece = MO_16 }, | ||
366 | + { .fniv = gen_sqxtunt_vec, | ||
367 | + .opt_opc = sqxtun_list, | ||
368 | + .load_dest = true, | ||
369 | + .fno = gen_helper_sve2_sqxtunt_s, | ||
370 | + .vece = MO_32 }, | ||
371 | + { .fniv = gen_sqxtunt_vec, | ||
372 | + .opt_opc = sqxtun_list, | ||
373 | + .load_dest = true, | ||
374 | + .fno = gen_helper_sve2_sqxtunt_d, | ||
375 | + .vece = MO_64 }, | ||
376 | + }; | ||
377 | + return do_sve2_narrow_extract(s, a, ops); | ||
378 | +} | ||
379 | -- | ||
380 | 2.20.1 | ||
381 | |||
382 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-26-richard.henderson@linaro.org | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | --- | ||
10 | target/arm/helper-sve.h | 35 +++++++++++++++++++++++++++++ | ||
11 | target/arm/sve.decode | 8 +++++++ | ||
12 | target/arm/sve_helper.c | 46 ++++++++++++++++++++++++++++++++++++++ | ||
13 | target/arm/translate-sve.c | 25 +++++++++++++++++++++ | ||
14 | 4 files changed, 114 insertions(+) | ||
15 | |||
16 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/helper-sve.h | ||
19 | +++ b/target/arm/helper-sve.h | ||
20 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_uqxtnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_3(sve2_sqxtunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_3(sve2_sqxtunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_3(sve2_sqxtunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | + | ||
25 | +DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_d, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
31 | + | ||
32 | +DEF_HELPER_FLAGS_6(sve2_fmaxnmp_zpzz_h, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_6(sve2_fmaxnmp_zpzz_s, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_6(sve2_fmaxnmp_zpzz_d, TCG_CALL_NO_RWG, | ||
37 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
38 | + | ||
39 | +DEF_HELPER_FLAGS_6(sve2_fminnmp_zpzz_h, TCG_CALL_NO_RWG, | ||
40 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_6(sve2_fminnmp_zpzz_s, TCG_CALL_NO_RWG, | ||
42 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
43 | +DEF_HELPER_FLAGS_6(sve2_fminnmp_zpzz_d, TCG_CALL_NO_RWG, | ||
44 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
45 | + | ||
46 | +DEF_HELPER_FLAGS_6(sve2_fmaxp_zpzz_h, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
48 | +DEF_HELPER_FLAGS_6(sve2_fmaxp_zpzz_s, TCG_CALL_NO_RWG, | ||
49 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
50 | +DEF_HELPER_FLAGS_6(sve2_fmaxp_zpzz_d, TCG_CALL_NO_RWG, | ||
51 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
52 | + | ||
53 | +DEF_HELPER_FLAGS_6(sve2_fminp_zpzz_h, TCG_CALL_NO_RWG, | ||
54 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
55 | +DEF_HELPER_FLAGS_6(sve2_fminp_zpzz_s, TCG_CALL_NO_RWG, | ||
56 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
57 | +DEF_HELPER_FLAGS_6(sve2_fminp_zpzz_d, TCG_CALL_NO_RWG, | ||
58 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
59 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
60 | index XXXXXXX..XXXXXXX 100644 | ||
61 | --- a/target/arm/sve.decode | ||
62 | +++ b/target/arm/sve.decode | ||
63 | @@ -XXX,XX +XXX,XX @@ UQXTNB 01000101 .. 1 ..... 010 010 ..... ..... @rd_rn_tszimm_shl | ||
64 | UQXTNT 01000101 .. 1 ..... 010 011 ..... ..... @rd_rn_tszimm_shl | ||
65 | SQXTUNB 01000101 .. 1 ..... 010 100 ..... ..... @rd_rn_tszimm_shl | ||
66 | SQXTUNT 01000101 .. 1 ..... 010 101 ..... ..... @rd_rn_tszimm_shl | ||
67 | + | ||
68 | +## SVE2 floating-point pairwise operations | ||
69 | + | ||
70 | +FADDP 01100100 .. 010 00 0 100 ... ..... ..... @rdn_pg_rm | ||
71 | +FMAXNMP 01100100 .. 010 10 0 100 ... ..... ..... @rdn_pg_rm | ||
72 | +FMINNMP 01100100 .. 010 10 1 100 ... ..... ..... @rdn_pg_rm | ||
73 | +FMAXP 01100100 .. 010 11 0 100 ... ..... ..... @rdn_pg_rm | ||
74 | +FMINP 01100100 .. 010 11 1 100 ... ..... ..... @rdn_pg_rm | ||
75 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
76 | index XXXXXXX..XXXXXXX 100644 | ||
77 | --- a/target/arm/sve_helper.c | ||
78 | +++ b/target/arm/sve_helper.c | ||
79 | @@ -XXX,XX +XXX,XX @@ DO_ZPZZ_PAIR_D(sve2_sminp_zpzz_d, int64_t, DO_MIN) | ||
80 | #undef DO_ZPZZ_PAIR | ||
81 | #undef DO_ZPZZ_PAIR_D | ||
82 | |||
83 | +#define DO_ZPZZ_PAIR_FP(NAME, TYPE, H, OP) \ | ||
84 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *vg, \ | ||
85 | + void *status, uint32_t desc) \ | ||
86 | +{ \ | ||
87 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
88 | + for (i = 0; i < opr_sz; ) { \ | ||
89 | + uint16_t pg = *(uint16_t *)(vg + H1_2(i >> 3)); \ | ||
90 | + do { \ | ||
91 | + TYPE n0 = *(TYPE *)(vn + H(i)); \ | ||
92 | + TYPE m0 = *(TYPE *)(vm + H(i)); \ | ||
93 | + TYPE n1 = *(TYPE *)(vn + H(i + sizeof(TYPE))); \ | ||
94 | + TYPE m1 = *(TYPE *)(vm + H(i + sizeof(TYPE))); \ | ||
95 | + if (pg & 1) { \ | ||
96 | + *(TYPE *)(vd + H(i)) = OP(n0, n1, status); \ | ||
97 | + } \ | ||
98 | + i += sizeof(TYPE), pg >>= sizeof(TYPE); \ | ||
99 | + if (pg & 1) { \ | ||
100 | + *(TYPE *)(vd + H(i)) = OP(m0, m1, status); \ | ||
101 | + } \ | ||
102 | + i += sizeof(TYPE), pg >>= sizeof(TYPE); \ | ||
103 | + } while (i & 15); \ | ||
104 | + } \ | ||
105 | +} | ||
106 | + | ||
107 | +DO_ZPZZ_PAIR_FP(sve2_faddp_zpzz_h, float16, H1_2, float16_add) | ||
108 | +DO_ZPZZ_PAIR_FP(sve2_faddp_zpzz_s, float32, H1_4, float32_add) | ||
109 | +DO_ZPZZ_PAIR_FP(sve2_faddp_zpzz_d, float64, , float64_add) | ||
110 | + | ||
111 | +DO_ZPZZ_PAIR_FP(sve2_fmaxnmp_zpzz_h, float16, H1_2, float16_maxnum) | ||
112 | +DO_ZPZZ_PAIR_FP(sve2_fmaxnmp_zpzz_s, float32, H1_4, float32_maxnum) | ||
113 | +DO_ZPZZ_PAIR_FP(sve2_fmaxnmp_zpzz_d, float64, , float64_maxnum) | ||
114 | + | ||
115 | +DO_ZPZZ_PAIR_FP(sve2_fminnmp_zpzz_h, float16, H1_2, float16_minnum) | ||
116 | +DO_ZPZZ_PAIR_FP(sve2_fminnmp_zpzz_s, float32, H1_4, float32_minnum) | ||
117 | +DO_ZPZZ_PAIR_FP(sve2_fminnmp_zpzz_d, float64, , float64_minnum) | ||
118 | + | ||
119 | +DO_ZPZZ_PAIR_FP(sve2_fmaxp_zpzz_h, float16, H1_2, float16_max) | ||
120 | +DO_ZPZZ_PAIR_FP(sve2_fmaxp_zpzz_s, float32, H1_4, float32_max) | ||
121 | +DO_ZPZZ_PAIR_FP(sve2_fmaxp_zpzz_d, float64, , float64_max) | ||
122 | + | ||
123 | +DO_ZPZZ_PAIR_FP(sve2_fminp_zpzz_h, float16, H1_2, float16_min) | ||
124 | +DO_ZPZZ_PAIR_FP(sve2_fminp_zpzz_s, float32, H1_4, float32_min) | ||
125 | +DO_ZPZZ_PAIR_FP(sve2_fminp_zpzz_d, float64, , float64_min) | ||
126 | + | ||
127 | +#undef DO_ZPZZ_PAIR_FP | ||
128 | + | ||
129 | /* Three-operand expander, controlled by a predicate, in which the | ||
130 | * third operand is "wide". That is, for D = N op M, the same 64-bit | ||
131 | * value of M is used with all of the narrower values of N. | ||
132 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
133 | index XXXXXXX..XXXXXXX 100644 | ||
134 | --- a/target/arm/translate-sve.c | ||
135 | +++ b/target/arm/translate-sve.c | ||
136 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQXTUNT(DisasContext *s, arg_rri_esz *a) | ||
137 | }; | ||
138 | return do_sve2_narrow_extract(s, a, ops); | ||
139 | } | ||
140 | + | ||
141 | +static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | ||
142 | + gen_helper_gvec_4_ptr *fn) | ||
143 | +{ | ||
144 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
145 | + return false; | ||
146 | + } | ||
147 | + return do_zpzz_fp(s, a, fn); | ||
148 | +} | ||
149 | + | ||
150 | +#define DO_SVE2_ZPZZ_FP(NAME, name) \ | ||
151 | +static bool trans_##NAME(DisasContext *s, arg_rprr_esz *a) \ | ||
152 | +{ \ | ||
153 | + static gen_helper_gvec_4_ptr * const fns[4] = { \ | ||
154 | + NULL, gen_helper_sve2_##name##_zpzz_h, \ | ||
155 | + gen_helper_sve2_##name##_zpzz_s, gen_helper_sve2_##name##_zpzz_d \ | ||
156 | + }; \ | ||
157 | + return do_sve2_zpzz_fp(s, a, fns[a->esz]); \ | ||
158 | +} | ||
159 | + | ||
160 | +DO_SVE2_ZPZZ_FP(FADDP, faddp) | ||
161 | +DO_SVE2_ZPZZ_FP(FMAXNMP, fmaxnmp) | ||
162 | +DO_SVE2_ZPZZ_FP(FMINNMP, fminnmp) | ||
163 | +DO_SVE2_ZPZZ_FP(FMAXP, fmaxp) | ||
164 | +DO_SVE2_ZPZZ_FP(FMINP, fminp) | ||
165 | -- | ||
166 | 2.20.1 | ||
167 | |||
168 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-27-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 16 ++++ | ||
9 | target/arm/sve.decode | 8 ++ | ||
10 | target/arm/sve_helper.c | 54 ++++++++++++- | ||
11 | target/arm/translate-sve.c | 160 +++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 236 insertions(+), 2 deletions(-) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_sqxtunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_3(sve2_sqxtunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_3(sve2_sqxtunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_3(sve2_shrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_3(sve2_shrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_3(sve2_shrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | + | ||
26 | +DEF_HELPER_FLAGS_3(sve2_shrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_3(sve2_shrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_3(sve2_shrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_3(sve2_rshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_3(sve2_rshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_3(sve2_rshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_3(sve2_rshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_3(sve2_rshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_3(sve2_rshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | + | ||
38 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
39 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
40 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
41 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/target/arm/sve.decode | ||
44 | +++ b/target/arm/sve.decode | ||
45 | @@ -XXX,XX +XXX,XX @@ UQXTNT 01000101 .. 1 ..... 010 011 ..... ..... @rd_rn_tszimm_shl | ||
46 | SQXTUNB 01000101 .. 1 ..... 010 100 ..... ..... @rd_rn_tszimm_shl | ||
47 | SQXTUNT 01000101 .. 1 ..... 010 101 ..... ..... @rd_rn_tszimm_shl | ||
48 | |||
49 | +## SVE2 bitwise shift right narrow | ||
50 | + | ||
51 | +# Bit 23 == 0 is handled by esz > 0 in the translator. | ||
52 | +SHRNB 01000101 .. 1 ..... 00 0100 ..... ..... @rd_rn_tszimm_shr | ||
53 | +SHRNT 01000101 .. 1 ..... 00 0101 ..... ..... @rd_rn_tszimm_shr | ||
54 | +RSHRNB 01000101 .. 1 ..... 00 0110 ..... ..... @rd_rn_tszimm_shr | ||
55 | +RSHRNT 01000101 .. 1 ..... 00 0111 ..... ..... @rd_rn_tszimm_shr | ||
56 | + | ||
57 | ## SVE2 floating-point pairwise operations | ||
58 | |||
59 | FADDP 01100100 .. 010 00 0 100 ... ..... ..... @rdn_pg_rm | ||
60 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
61 | index XXXXXXX..XXXXXXX 100644 | ||
62 | --- a/target/arm/sve_helper.c | ||
63 | +++ b/target/arm/sve_helper.c | ||
64 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vg, uint32_t desc) \ | ||
65 | when N is negative, add 2**M-1. */ | ||
66 | #define DO_ASRD(N, M) ((N + (N < 0 ? ((__typeof(N))1 << M) - 1 : 0)) >> M) | ||
67 | |||
68 | +static inline uint64_t do_urshr(uint64_t x, unsigned sh) | ||
69 | +{ | ||
70 | + if (likely(sh < 64)) { | ||
71 | + return (x >> sh) + ((x >> (sh - 1)) & 1); | ||
72 | + } else if (sh == 64) { | ||
73 | + return x >> 63; | ||
74 | + } else { | ||
75 | + return 0; | ||
76 | + } | ||
77 | +} | ||
78 | + | ||
79 | DO_ZPZI(sve_asr_zpzi_b, int8_t, H1, DO_SHR) | ||
80 | DO_ZPZI(sve_asr_zpzi_h, int16_t, H1_2, DO_SHR) | ||
81 | DO_ZPZI(sve_asr_zpzi_s, int32_t, H1_4, DO_SHR) | ||
82 | @@ -XXX,XX +XXX,XX @@ DO_ZPZI(sve_asrd_h, int16_t, H1_2, DO_ASRD) | ||
83 | DO_ZPZI(sve_asrd_s, int32_t, H1_4, DO_ASRD) | ||
84 | DO_ZPZI_D(sve_asrd_d, int64_t, DO_ASRD) | ||
85 | |||
86 | -#undef DO_SHR | ||
87 | -#undef DO_SHL | ||
88 | #undef DO_ASRD | ||
89 | #undef DO_ZPZI | ||
90 | #undef DO_ZPZI_D | ||
91 | |||
92 | +#define DO_SHRNB(NAME, TYPEW, TYPEN, OP) \ | ||
93 | +void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
94 | +{ \ | ||
95 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
96 | + int shift = simd_data(desc); \ | ||
97 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
98 | + TYPEW nn = *(TYPEW *)(vn + i); \ | ||
99 | + *(TYPEW *)(vd + i) = (TYPEN)OP(nn, shift); \ | ||
100 | + } \ | ||
101 | +} | ||
102 | + | ||
103 | +#define DO_SHRNT(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
104 | +void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
105 | +{ \ | ||
106 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
107 | + int shift = simd_data(desc); \ | ||
108 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
109 | + TYPEW nn = *(TYPEW *)(vn + HW(i)); \ | ||
110 | + *(TYPEN *)(vd + HN(i + sizeof(TYPEN))) = OP(nn, shift); \ | ||
111 | + } \ | ||
112 | +} | ||
113 | + | ||
114 | +DO_SHRNB(sve2_shrnb_h, uint16_t, uint8_t, DO_SHR) | ||
115 | +DO_SHRNB(sve2_shrnb_s, uint32_t, uint16_t, DO_SHR) | ||
116 | +DO_SHRNB(sve2_shrnb_d, uint64_t, uint32_t, DO_SHR) | ||
117 | + | ||
118 | +DO_SHRNT(sve2_shrnt_h, uint16_t, uint8_t, H1_2, H1, DO_SHR) | ||
119 | +DO_SHRNT(sve2_shrnt_s, uint32_t, uint16_t, H1_4, H1_2, DO_SHR) | ||
120 | +DO_SHRNT(sve2_shrnt_d, uint64_t, uint32_t, , H1_4, DO_SHR) | ||
121 | + | ||
122 | +DO_SHRNB(sve2_rshrnb_h, uint16_t, uint8_t, do_urshr) | ||
123 | +DO_SHRNB(sve2_rshrnb_s, uint32_t, uint16_t, do_urshr) | ||
124 | +DO_SHRNB(sve2_rshrnb_d, uint64_t, uint32_t, do_urshr) | ||
125 | + | ||
126 | +DO_SHRNT(sve2_rshrnt_h, uint16_t, uint8_t, H1_2, H1, do_urshr) | ||
127 | +DO_SHRNT(sve2_rshrnt_s, uint32_t, uint16_t, H1_4, H1_2, do_urshr) | ||
128 | +DO_SHRNT(sve2_rshrnt_d, uint64_t, uint32_t, , H1_4, do_urshr) | ||
129 | + | ||
130 | +#undef DO_SHRNB | ||
131 | +#undef DO_SHRNT | ||
132 | + | ||
133 | /* Fully general four-operand expander, controlled by a predicate. | ||
134 | */ | ||
135 | #define DO_ZPZZZ(NAME, TYPE, H, OP) \ | ||
136 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
137 | index XXXXXXX..XXXXXXX 100644 | ||
138 | --- a/target/arm/translate-sve.c | ||
139 | +++ b/target/arm/translate-sve.c | ||
140 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQXTUNT(DisasContext *s, arg_rri_esz *a) | ||
141 | return do_sve2_narrow_extract(s, a, ops); | ||
142 | } | ||
143 | |||
144 | +static bool do_sve2_shr_narrow(DisasContext *s, arg_rri_esz *a, | ||
145 | + const GVecGen2i ops[3]) | ||
146 | +{ | ||
147 | + if (a->esz < 0 || a->esz > MO_32 || !dc_isar_feature(aa64_sve2, s)) { | ||
148 | + return false; | ||
149 | + } | ||
150 | + assert(a->imm > 0 && a->imm <= (8 << a->esz)); | ||
151 | + if (sve_access_check(s)) { | ||
152 | + unsigned vsz = vec_full_reg_size(s); | ||
153 | + tcg_gen_gvec_2i(vec_full_reg_offset(s, a->rd), | ||
154 | + vec_full_reg_offset(s, a->rn), | ||
155 | + vsz, vsz, a->imm, &ops[a->esz]); | ||
156 | + } | ||
157 | + return true; | ||
158 | +} | ||
159 | + | ||
160 | +static void gen_shrnb_i64(unsigned vece, TCGv_i64 d, TCGv_i64 n, int shr) | ||
161 | +{ | ||
162 | + int halfbits = 4 << vece; | ||
163 | + uint64_t mask = dup_const(vece, MAKE_64BIT_MASK(0, halfbits)); | ||
164 | + | ||
165 | + tcg_gen_shri_i64(d, n, shr); | ||
166 | + tcg_gen_andi_i64(d, d, mask); | ||
167 | +} | ||
168 | + | ||
169 | +static void gen_shrnb16_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
170 | +{ | ||
171 | + gen_shrnb_i64(MO_16, d, n, shr); | ||
172 | +} | ||
173 | + | ||
174 | +static void gen_shrnb32_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
175 | +{ | ||
176 | + gen_shrnb_i64(MO_32, d, n, shr); | ||
177 | +} | ||
178 | + | ||
179 | +static void gen_shrnb64_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
180 | +{ | ||
181 | + gen_shrnb_i64(MO_64, d, n, shr); | ||
182 | +} | ||
183 | + | ||
184 | +static void gen_shrnb_vec(unsigned vece, TCGv_vec d, TCGv_vec n, int64_t shr) | ||
185 | +{ | ||
186 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
187 | + int halfbits = 4 << vece; | ||
188 | + uint64_t mask = MAKE_64BIT_MASK(0, halfbits); | ||
189 | + | ||
190 | + tcg_gen_shri_vec(vece, n, n, shr); | ||
191 | + tcg_gen_dupi_vec(vece, t, mask); | ||
192 | + tcg_gen_and_vec(vece, d, n, t); | ||
193 | + tcg_temp_free_vec(t); | ||
194 | +} | ||
195 | + | ||
196 | +static bool trans_SHRNB(DisasContext *s, arg_rri_esz *a) | ||
197 | +{ | ||
198 | + static const TCGOpcode vec_list[] = { INDEX_op_shri_vec, 0 }; | ||
199 | + static const GVecGen2i ops[3] = { | ||
200 | + { .fni8 = gen_shrnb16_i64, | ||
201 | + .fniv = gen_shrnb_vec, | ||
202 | + .opt_opc = vec_list, | ||
203 | + .fno = gen_helper_sve2_shrnb_h, | ||
204 | + .vece = MO_16 }, | ||
205 | + { .fni8 = gen_shrnb32_i64, | ||
206 | + .fniv = gen_shrnb_vec, | ||
207 | + .opt_opc = vec_list, | ||
208 | + .fno = gen_helper_sve2_shrnb_s, | ||
209 | + .vece = MO_32 }, | ||
210 | + { .fni8 = gen_shrnb64_i64, | ||
211 | + .fniv = gen_shrnb_vec, | ||
212 | + .opt_opc = vec_list, | ||
213 | + .fno = gen_helper_sve2_shrnb_d, | ||
214 | + .vece = MO_64 }, | ||
215 | + }; | ||
216 | + return do_sve2_shr_narrow(s, a, ops); | ||
217 | +} | ||
218 | + | ||
219 | +static void gen_shrnt_i64(unsigned vece, TCGv_i64 d, TCGv_i64 n, int shr) | ||
220 | +{ | ||
221 | + int halfbits = 4 << vece; | ||
222 | + uint64_t mask = dup_const(vece, MAKE_64BIT_MASK(0, halfbits)); | ||
223 | + | ||
224 | + tcg_gen_shli_i64(n, n, halfbits - shr); | ||
225 | + tcg_gen_andi_i64(n, n, ~mask); | ||
226 | + tcg_gen_andi_i64(d, d, mask); | ||
227 | + tcg_gen_or_i64(d, d, n); | ||
228 | +} | ||
229 | + | ||
230 | +static void gen_shrnt16_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
231 | +{ | ||
232 | + gen_shrnt_i64(MO_16, d, n, shr); | ||
233 | +} | ||
234 | + | ||
235 | +static void gen_shrnt32_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
236 | +{ | ||
237 | + gen_shrnt_i64(MO_32, d, n, shr); | ||
238 | +} | ||
239 | + | ||
240 | +static void gen_shrnt64_i64(TCGv_i64 d, TCGv_i64 n, int64_t shr) | ||
241 | +{ | ||
242 | + tcg_gen_shri_i64(n, n, shr); | ||
243 | + tcg_gen_deposit_i64(d, d, n, 32, 32); | ||
244 | +} | ||
245 | + | ||
246 | +static void gen_shrnt_vec(unsigned vece, TCGv_vec d, TCGv_vec n, int64_t shr) | ||
247 | +{ | ||
248 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
249 | + int halfbits = 4 << vece; | ||
250 | + uint64_t mask = MAKE_64BIT_MASK(0, halfbits); | ||
251 | + | ||
252 | + tcg_gen_shli_vec(vece, n, n, halfbits - shr); | ||
253 | + tcg_gen_dupi_vec(vece, t, mask); | ||
254 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
255 | + tcg_temp_free_vec(t); | ||
256 | +} | ||
257 | + | ||
258 | +static bool trans_SHRNT(DisasContext *s, arg_rri_esz *a) | ||
259 | +{ | ||
260 | + static const TCGOpcode vec_list[] = { INDEX_op_shli_vec, 0 }; | ||
261 | + static const GVecGen2i ops[3] = { | ||
262 | + { .fni8 = gen_shrnt16_i64, | ||
263 | + .fniv = gen_shrnt_vec, | ||
264 | + .opt_opc = vec_list, | ||
265 | + .load_dest = true, | ||
266 | + .fno = gen_helper_sve2_shrnt_h, | ||
267 | + .vece = MO_16 }, | ||
268 | + { .fni8 = gen_shrnt32_i64, | ||
269 | + .fniv = gen_shrnt_vec, | ||
270 | + .opt_opc = vec_list, | ||
271 | + .load_dest = true, | ||
272 | + .fno = gen_helper_sve2_shrnt_s, | ||
273 | + .vece = MO_32 }, | ||
274 | + { .fni8 = gen_shrnt64_i64, | ||
275 | + .fniv = gen_shrnt_vec, | ||
276 | + .opt_opc = vec_list, | ||
277 | + .load_dest = true, | ||
278 | + .fno = gen_helper_sve2_shrnt_d, | ||
279 | + .vece = MO_64 }, | ||
280 | + }; | ||
281 | + return do_sve2_shr_narrow(s, a, ops); | ||
282 | +} | ||
283 | + | ||
284 | +static bool trans_RSHRNB(DisasContext *s, arg_rri_esz *a) | ||
285 | +{ | ||
286 | + static const GVecGen2i ops[3] = { | ||
287 | + { .fno = gen_helper_sve2_rshrnb_h }, | ||
288 | + { .fno = gen_helper_sve2_rshrnb_s }, | ||
289 | + { .fno = gen_helper_sve2_rshrnb_d }, | ||
290 | + }; | ||
291 | + return do_sve2_shr_narrow(s, a, ops); | ||
292 | +} | ||
293 | + | ||
294 | +static bool trans_RSHRNT(DisasContext *s, arg_rri_esz *a) | ||
295 | +{ | ||
296 | + static const GVecGen2i ops[3] = { | ||
297 | + { .fno = gen_helper_sve2_rshrnt_h }, | ||
298 | + { .fno = gen_helper_sve2_rshrnt_s }, | ||
299 | + { .fno = gen_helper_sve2_rshrnt_d }, | ||
300 | + }; | ||
301 | + return do_sve2_shr_narrow(s, a, ops); | ||
302 | +} | ||
303 | + | ||
304 | static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | ||
305 | gen_helper_gvec_4_ptr *fn) | ||
306 | { | ||
307 | -- | ||
308 | 2.20.1 | ||
309 | |||
310 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-28-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 16 +++++++ | ||
9 | target/arm/sve.decode | 4 ++ | ||
10 | target/arm/sve_helper.c | 35 ++++++++++++++ | ||
11 | target/arm/translate-sve.c | 98 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 153 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_rshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_3(sve2_rshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_3(sve2_rshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_3(sve2_sqshrunb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_3(sve2_sqshrunb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_3(sve2_sqshrunb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | + | ||
26 | +DEF_HELPER_FLAGS_3(sve2_sqshrunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_3(sve2_sqshrunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_3(sve2_sqshrunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_3(sve2_sqrshrunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | + | ||
38 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
39 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
40 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
41 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/target/arm/sve.decode | ||
44 | +++ b/target/arm/sve.decode | ||
45 | @@ -XXX,XX +XXX,XX @@ SQXTUNT 01000101 .. 1 ..... 010 101 ..... ..... @rd_rn_tszimm_shl | ||
46 | ## SVE2 bitwise shift right narrow | ||
47 | |||
48 | # Bit 23 == 0 is handled by esz > 0 in the translator. | ||
49 | +SQSHRUNB 01000101 .. 1 ..... 00 0000 ..... ..... @rd_rn_tszimm_shr | ||
50 | +SQSHRUNT 01000101 .. 1 ..... 00 0001 ..... ..... @rd_rn_tszimm_shr | ||
51 | +SQRSHRUNB 01000101 .. 1 ..... 00 0010 ..... ..... @rd_rn_tszimm_shr | ||
52 | +SQRSHRUNT 01000101 .. 1 ..... 00 0011 ..... ..... @rd_rn_tszimm_shr | ||
53 | SHRNB 01000101 .. 1 ..... 00 0100 ..... ..... @rd_rn_tszimm_shr | ||
54 | SHRNT 01000101 .. 1 ..... 00 0101 ..... ..... @rd_rn_tszimm_shr | ||
55 | RSHRNB 01000101 .. 1 ..... 00 0110 ..... ..... @rd_rn_tszimm_shr | ||
56 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
57 | index XXXXXXX..XXXXXXX 100644 | ||
58 | --- a/target/arm/sve_helper.c | ||
59 | +++ b/target/arm/sve_helper.c | ||
60 | @@ -XXX,XX +XXX,XX @@ static inline uint64_t do_urshr(uint64_t x, unsigned sh) | ||
61 | } | ||
62 | } | ||
63 | |||
64 | +static inline int64_t do_srshr(int64_t x, unsigned sh) | ||
65 | +{ | ||
66 | + if (likely(sh < 64)) { | ||
67 | + return (x >> sh) + ((x >> (sh - 1)) & 1); | ||
68 | + } else { | ||
69 | + /* Rounding the sign bit always produces 0. */ | ||
70 | + return 0; | ||
71 | + } | ||
72 | +} | ||
73 | + | ||
74 | DO_ZPZI(sve_asr_zpzi_b, int8_t, H1, DO_SHR) | ||
75 | DO_ZPZI(sve_asr_zpzi_h, int16_t, H1_2, DO_SHR) | ||
76 | DO_ZPZI(sve_asr_zpzi_s, int32_t, H1_4, DO_SHR) | ||
77 | @@ -XXX,XX +XXX,XX @@ DO_SHRNT(sve2_rshrnt_h, uint16_t, uint8_t, H1_2, H1, do_urshr) | ||
78 | DO_SHRNT(sve2_rshrnt_s, uint32_t, uint16_t, H1_4, H1_2, do_urshr) | ||
79 | DO_SHRNT(sve2_rshrnt_d, uint64_t, uint32_t, , H1_4, do_urshr) | ||
80 | |||
81 | +#define DO_SQSHRUN_H(x, sh) do_sat_bhs((int64_t)(x) >> sh, 0, UINT8_MAX) | ||
82 | +#define DO_SQSHRUN_S(x, sh) do_sat_bhs((int64_t)(x) >> sh, 0, UINT16_MAX) | ||
83 | +#define DO_SQSHRUN_D(x, sh) \ | ||
84 | + do_sat_bhs((int64_t)(x) >> (sh < 64 ? sh : 63), 0, UINT32_MAX) | ||
85 | + | ||
86 | +DO_SHRNB(sve2_sqshrunb_h, int16_t, uint8_t, DO_SQSHRUN_H) | ||
87 | +DO_SHRNB(sve2_sqshrunb_s, int32_t, uint16_t, DO_SQSHRUN_S) | ||
88 | +DO_SHRNB(sve2_sqshrunb_d, int64_t, uint32_t, DO_SQSHRUN_D) | ||
89 | + | ||
90 | +DO_SHRNT(sve2_sqshrunt_h, int16_t, uint8_t, H1_2, H1, DO_SQSHRUN_H) | ||
91 | +DO_SHRNT(sve2_sqshrunt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQSHRUN_S) | ||
92 | +DO_SHRNT(sve2_sqshrunt_d, int64_t, uint32_t, , H1_4, DO_SQSHRUN_D) | ||
93 | + | ||
94 | +#define DO_SQRSHRUN_H(x, sh) do_sat_bhs(do_srshr(x, sh), 0, UINT8_MAX) | ||
95 | +#define DO_SQRSHRUN_S(x, sh) do_sat_bhs(do_srshr(x, sh), 0, UINT16_MAX) | ||
96 | +#define DO_SQRSHRUN_D(x, sh) do_sat_bhs(do_srshr(x, sh), 0, UINT32_MAX) | ||
97 | + | ||
98 | +DO_SHRNB(sve2_sqrshrunb_h, int16_t, uint8_t, DO_SQRSHRUN_H) | ||
99 | +DO_SHRNB(sve2_sqrshrunb_s, int32_t, uint16_t, DO_SQRSHRUN_S) | ||
100 | +DO_SHRNB(sve2_sqrshrunb_d, int64_t, uint32_t, DO_SQRSHRUN_D) | ||
101 | + | ||
102 | +DO_SHRNT(sve2_sqrshrunt_h, int16_t, uint8_t, H1_2, H1, DO_SQRSHRUN_H) | ||
103 | +DO_SHRNT(sve2_sqrshrunt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQRSHRUN_S) | ||
104 | +DO_SHRNT(sve2_sqrshrunt_d, int64_t, uint32_t, , H1_4, DO_SQRSHRUN_D) | ||
105 | + | ||
106 | #undef DO_SHRNB | ||
107 | #undef DO_SHRNT | ||
108 | |||
109 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
110 | index XXXXXXX..XXXXXXX 100644 | ||
111 | --- a/target/arm/translate-sve.c | ||
112 | +++ b/target/arm/translate-sve.c | ||
113 | @@ -XXX,XX +XXX,XX @@ static bool trans_RSHRNT(DisasContext *s, arg_rri_esz *a) | ||
114 | return do_sve2_shr_narrow(s, a, ops); | ||
115 | } | ||
116 | |||
117 | +static void gen_sqshrunb_vec(unsigned vece, TCGv_vec d, | ||
118 | + TCGv_vec n, int64_t shr) | ||
119 | +{ | ||
120 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
121 | + int halfbits = 4 << vece; | ||
122 | + | ||
123 | + tcg_gen_sari_vec(vece, n, n, shr); | ||
124 | + tcg_gen_dupi_vec(vece, t, 0); | ||
125 | + tcg_gen_smax_vec(vece, n, n, t); | ||
126 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
127 | + tcg_gen_umin_vec(vece, d, n, t); | ||
128 | + tcg_temp_free_vec(t); | ||
129 | +} | ||
130 | + | ||
131 | +static bool trans_SQSHRUNB(DisasContext *s, arg_rri_esz *a) | ||
132 | +{ | ||
133 | + static const TCGOpcode vec_list[] = { | ||
134 | + INDEX_op_sari_vec, INDEX_op_smax_vec, INDEX_op_umin_vec, 0 | ||
135 | + }; | ||
136 | + static const GVecGen2i ops[3] = { | ||
137 | + { .fniv = gen_sqshrunb_vec, | ||
138 | + .opt_opc = vec_list, | ||
139 | + .fno = gen_helper_sve2_sqshrunb_h, | ||
140 | + .vece = MO_16 }, | ||
141 | + { .fniv = gen_sqshrunb_vec, | ||
142 | + .opt_opc = vec_list, | ||
143 | + .fno = gen_helper_sve2_sqshrunb_s, | ||
144 | + .vece = MO_32 }, | ||
145 | + { .fniv = gen_sqshrunb_vec, | ||
146 | + .opt_opc = vec_list, | ||
147 | + .fno = gen_helper_sve2_sqshrunb_d, | ||
148 | + .vece = MO_64 }, | ||
149 | + }; | ||
150 | + return do_sve2_shr_narrow(s, a, ops); | ||
151 | +} | ||
152 | + | ||
153 | +static void gen_sqshrunt_vec(unsigned vece, TCGv_vec d, | ||
154 | + TCGv_vec n, int64_t shr) | ||
155 | +{ | ||
156 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
157 | + int halfbits = 4 << vece; | ||
158 | + | ||
159 | + tcg_gen_sari_vec(vece, n, n, shr); | ||
160 | + tcg_gen_dupi_vec(vece, t, 0); | ||
161 | + tcg_gen_smax_vec(vece, n, n, t); | ||
162 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
163 | + tcg_gen_umin_vec(vece, n, n, t); | ||
164 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
165 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
166 | + tcg_temp_free_vec(t); | ||
167 | +} | ||
168 | + | ||
169 | +static bool trans_SQSHRUNT(DisasContext *s, arg_rri_esz *a) | ||
170 | +{ | ||
171 | + static const TCGOpcode vec_list[] = { | ||
172 | + INDEX_op_shli_vec, INDEX_op_sari_vec, | ||
173 | + INDEX_op_smax_vec, INDEX_op_umin_vec, 0 | ||
174 | + }; | ||
175 | + static const GVecGen2i ops[3] = { | ||
176 | + { .fniv = gen_sqshrunt_vec, | ||
177 | + .opt_opc = vec_list, | ||
178 | + .load_dest = true, | ||
179 | + .fno = gen_helper_sve2_sqshrunt_h, | ||
180 | + .vece = MO_16 }, | ||
181 | + { .fniv = gen_sqshrunt_vec, | ||
182 | + .opt_opc = vec_list, | ||
183 | + .load_dest = true, | ||
184 | + .fno = gen_helper_sve2_sqshrunt_s, | ||
185 | + .vece = MO_32 }, | ||
186 | + { .fniv = gen_sqshrunt_vec, | ||
187 | + .opt_opc = vec_list, | ||
188 | + .load_dest = true, | ||
189 | + .fno = gen_helper_sve2_sqshrunt_d, | ||
190 | + .vece = MO_64 }, | ||
191 | + }; | ||
192 | + return do_sve2_shr_narrow(s, a, ops); | ||
193 | +} | ||
194 | + | ||
195 | +static bool trans_SQRSHRUNB(DisasContext *s, arg_rri_esz *a) | ||
196 | +{ | ||
197 | + static const GVecGen2i ops[3] = { | ||
198 | + { .fno = gen_helper_sve2_sqrshrunb_h }, | ||
199 | + { .fno = gen_helper_sve2_sqrshrunb_s }, | ||
200 | + { .fno = gen_helper_sve2_sqrshrunb_d }, | ||
201 | + }; | ||
202 | + return do_sve2_shr_narrow(s, a, ops); | ||
203 | +} | ||
204 | + | ||
205 | +static bool trans_SQRSHRUNT(DisasContext *s, arg_rri_esz *a) | ||
206 | +{ | ||
207 | + static const GVecGen2i ops[3] = { | ||
208 | + { .fno = gen_helper_sve2_sqrshrunt_h }, | ||
209 | + { .fno = gen_helper_sve2_sqrshrunt_s }, | ||
210 | + { .fno = gen_helper_sve2_sqrshrunt_d }, | ||
211 | + }; | ||
212 | + return do_sve2_shr_narrow(s, a, ops); | ||
213 | +} | ||
214 | + | ||
215 | static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | ||
216 | gen_helper_gvec_4_ptr *fn) | ||
217 | { | ||
218 | -- | ||
219 | 2.20.1 | ||
220 | |||
221 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-29-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 16 +++++++ | ||
9 | target/arm/sve.decode | 4 ++ | ||
10 | target/arm/sve_helper.c | 24 ++++++++++ | ||
11 | target/arm/translate-sve.c | 93 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 137 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_sqrshrunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_3(sve2_sqrshrunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_3(sve2_sqrshrunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_3(sve2_uqshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_3(sve2_uqshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_3(sve2_uqshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | + | ||
26 | +DEF_HELPER_FLAGS_3(sve2_uqshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_3(sve2_uqshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_3(sve2_uqshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_3(sve2_uqrshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | + | ||
38 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
39 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
40 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
41 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/target/arm/sve.decode | ||
44 | +++ b/target/arm/sve.decode | ||
45 | @@ -XXX,XX +XXX,XX @@ SHRNB 01000101 .. 1 ..... 00 0100 ..... ..... @rd_rn_tszimm_shr | ||
46 | SHRNT 01000101 .. 1 ..... 00 0101 ..... ..... @rd_rn_tszimm_shr | ||
47 | RSHRNB 01000101 .. 1 ..... 00 0110 ..... ..... @rd_rn_tszimm_shr | ||
48 | RSHRNT 01000101 .. 1 ..... 00 0111 ..... ..... @rd_rn_tszimm_shr | ||
49 | +UQSHRNB 01000101 .. 1 ..... 00 1100 ..... ..... @rd_rn_tszimm_shr | ||
50 | +UQSHRNT 01000101 .. 1 ..... 00 1101 ..... ..... @rd_rn_tszimm_shr | ||
51 | +UQRSHRNB 01000101 .. 1 ..... 00 1110 ..... ..... @rd_rn_tszimm_shr | ||
52 | +UQRSHRNT 01000101 .. 1 ..... 00 1111 ..... ..... @rd_rn_tszimm_shr | ||
53 | |||
54 | ## SVE2 floating-point pairwise operations | ||
55 | |||
56 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
57 | index XXXXXXX..XXXXXXX 100644 | ||
58 | --- a/target/arm/sve_helper.c | ||
59 | +++ b/target/arm/sve_helper.c | ||
60 | @@ -XXX,XX +XXX,XX @@ DO_SHRNT(sve2_sqrshrunt_h, int16_t, uint8_t, H1_2, H1, DO_SQRSHRUN_H) | ||
61 | DO_SHRNT(sve2_sqrshrunt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQRSHRUN_S) | ||
62 | DO_SHRNT(sve2_sqrshrunt_d, int64_t, uint32_t, , H1_4, DO_SQRSHRUN_D) | ||
63 | |||
64 | +#define DO_UQSHRN_H(x, sh) MIN(x >> sh, UINT8_MAX) | ||
65 | +#define DO_UQSHRN_S(x, sh) MIN(x >> sh, UINT16_MAX) | ||
66 | +#define DO_UQSHRN_D(x, sh) MIN(x >> sh, UINT32_MAX) | ||
67 | + | ||
68 | +DO_SHRNB(sve2_uqshrnb_h, uint16_t, uint8_t, DO_UQSHRN_H) | ||
69 | +DO_SHRNB(sve2_uqshrnb_s, uint32_t, uint16_t, DO_UQSHRN_S) | ||
70 | +DO_SHRNB(sve2_uqshrnb_d, uint64_t, uint32_t, DO_UQSHRN_D) | ||
71 | + | ||
72 | +DO_SHRNT(sve2_uqshrnt_h, uint16_t, uint8_t, H1_2, H1, DO_UQSHRN_H) | ||
73 | +DO_SHRNT(sve2_uqshrnt_s, uint32_t, uint16_t, H1_4, H1_2, DO_UQSHRN_S) | ||
74 | +DO_SHRNT(sve2_uqshrnt_d, uint64_t, uint32_t, , H1_4, DO_UQSHRN_D) | ||
75 | + | ||
76 | +#define DO_UQRSHRN_H(x, sh) MIN(do_urshr(x, sh), UINT8_MAX) | ||
77 | +#define DO_UQRSHRN_S(x, sh) MIN(do_urshr(x, sh), UINT16_MAX) | ||
78 | +#define DO_UQRSHRN_D(x, sh) MIN(do_urshr(x, sh), UINT32_MAX) | ||
79 | + | ||
80 | +DO_SHRNB(sve2_uqrshrnb_h, uint16_t, uint8_t, DO_UQRSHRN_H) | ||
81 | +DO_SHRNB(sve2_uqrshrnb_s, uint32_t, uint16_t, DO_UQRSHRN_S) | ||
82 | +DO_SHRNB(sve2_uqrshrnb_d, uint64_t, uint32_t, DO_UQRSHRN_D) | ||
83 | + | ||
84 | +DO_SHRNT(sve2_uqrshrnt_h, uint16_t, uint8_t, H1_2, H1, DO_UQRSHRN_H) | ||
85 | +DO_SHRNT(sve2_uqrshrnt_s, uint32_t, uint16_t, H1_4, H1_2, DO_UQRSHRN_S) | ||
86 | +DO_SHRNT(sve2_uqrshrnt_d, uint64_t, uint32_t, , H1_4, DO_UQRSHRN_D) | ||
87 | + | ||
88 | #undef DO_SHRNB | ||
89 | #undef DO_SHRNT | ||
90 | |||
91 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
92 | index XXXXXXX..XXXXXXX 100644 | ||
93 | --- a/target/arm/translate-sve.c | ||
94 | +++ b/target/arm/translate-sve.c | ||
95 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQRSHRUNT(DisasContext *s, arg_rri_esz *a) | ||
96 | return do_sve2_shr_narrow(s, a, ops); | ||
97 | } | ||
98 | |||
99 | +static void gen_uqshrnb_vec(unsigned vece, TCGv_vec d, | ||
100 | + TCGv_vec n, int64_t shr) | ||
101 | +{ | ||
102 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
103 | + int halfbits = 4 << vece; | ||
104 | + | ||
105 | + tcg_gen_shri_vec(vece, n, n, shr); | ||
106 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
107 | + tcg_gen_umin_vec(vece, d, n, t); | ||
108 | + tcg_temp_free_vec(t); | ||
109 | +} | ||
110 | + | ||
111 | +static bool trans_UQSHRNB(DisasContext *s, arg_rri_esz *a) | ||
112 | +{ | ||
113 | + static const TCGOpcode vec_list[] = { | ||
114 | + INDEX_op_shri_vec, INDEX_op_umin_vec, 0 | ||
115 | + }; | ||
116 | + static const GVecGen2i ops[3] = { | ||
117 | + { .fniv = gen_uqshrnb_vec, | ||
118 | + .opt_opc = vec_list, | ||
119 | + .fno = gen_helper_sve2_uqshrnb_h, | ||
120 | + .vece = MO_16 }, | ||
121 | + { .fniv = gen_uqshrnb_vec, | ||
122 | + .opt_opc = vec_list, | ||
123 | + .fno = gen_helper_sve2_uqshrnb_s, | ||
124 | + .vece = MO_32 }, | ||
125 | + { .fniv = gen_uqshrnb_vec, | ||
126 | + .opt_opc = vec_list, | ||
127 | + .fno = gen_helper_sve2_uqshrnb_d, | ||
128 | + .vece = MO_64 }, | ||
129 | + }; | ||
130 | + return do_sve2_shr_narrow(s, a, ops); | ||
131 | +} | ||
132 | + | ||
133 | +static void gen_uqshrnt_vec(unsigned vece, TCGv_vec d, | ||
134 | + TCGv_vec n, int64_t shr) | ||
135 | +{ | ||
136 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
137 | + int halfbits = 4 << vece; | ||
138 | + | ||
139 | + tcg_gen_shri_vec(vece, n, n, shr); | ||
140 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
141 | + tcg_gen_umin_vec(vece, n, n, t); | ||
142 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
143 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
144 | + tcg_temp_free_vec(t); | ||
145 | +} | ||
146 | + | ||
147 | +static bool trans_UQSHRNT(DisasContext *s, arg_rri_esz *a) | ||
148 | +{ | ||
149 | + static const TCGOpcode vec_list[] = { | ||
150 | + INDEX_op_shli_vec, INDEX_op_shri_vec, INDEX_op_umin_vec, 0 | ||
151 | + }; | ||
152 | + static const GVecGen2i ops[3] = { | ||
153 | + { .fniv = gen_uqshrnt_vec, | ||
154 | + .opt_opc = vec_list, | ||
155 | + .load_dest = true, | ||
156 | + .fno = gen_helper_sve2_uqshrnt_h, | ||
157 | + .vece = MO_16 }, | ||
158 | + { .fniv = gen_uqshrnt_vec, | ||
159 | + .opt_opc = vec_list, | ||
160 | + .load_dest = true, | ||
161 | + .fno = gen_helper_sve2_uqshrnt_s, | ||
162 | + .vece = MO_32 }, | ||
163 | + { .fniv = gen_uqshrnt_vec, | ||
164 | + .opt_opc = vec_list, | ||
165 | + .load_dest = true, | ||
166 | + .fno = gen_helper_sve2_uqshrnt_d, | ||
167 | + .vece = MO_64 }, | ||
168 | + }; | ||
169 | + return do_sve2_shr_narrow(s, a, ops); | ||
170 | +} | ||
171 | + | ||
172 | +static bool trans_UQRSHRNB(DisasContext *s, arg_rri_esz *a) | ||
173 | +{ | ||
174 | + static const GVecGen2i ops[3] = { | ||
175 | + { .fno = gen_helper_sve2_uqrshrnb_h }, | ||
176 | + { .fno = gen_helper_sve2_uqrshrnb_s }, | ||
177 | + { .fno = gen_helper_sve2_uqrshrnb_d }, | ||
178 | + }; | ||
179 | + return do_sve2_shr_narrow(s, a, ops); | ||
180 | +} | ||
181 | + | ||
182 | +static bool trans_UQRSHRNT(DisasContext *s, arg_rri_esz *a) | ||
183 | +{ | ||
184 | + static const GVecGen2i ops[3] = { | ||
185 | + { .fno = gen_helper_sve2_uqrshrnt_h }, | ||
186 | + { .fno = gen_helper_sve2_uqrshrnt_s }, | ||
187 | + { .fno = gen_helper_sve2_uqrshrnt_d }, | ||
188 | + }; | ||
189 | + return do_sve2_shr_narrow(s, a, ops); | ||
190 | +} | ||
191 | + | ||
192 | static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | ||
193 | gen_helper_gvec_4_ptr *fn) | ||
194 | { | ||
195 | -- | ||
196 | 2.20.1 | ||
197 | |||
198 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | This completes the section "SVE2 bitwise shift right narrow". | ||
4 | |||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-30-richard.henderson@linaro.org | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | --- | ||
10 | target/arm/helper-sve.h | 16 ++++++ | ||
11 | target/arm/sve.decode | 4 ++ | ||
12 | target/arm/sve_helper.c | 24 +++++++++ | ||
13 | target/arm/translate-sve.c | 105 +++++++++++++++++++++++++++++++++++++ | ||
14 | 4 files changed, 149 insertions(+) | ||
15 | |||
16 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/helper-sve.h | ||
19 | +++ b/target/arm/helper-sve.h | ||
20 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_sqrshrunt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_3(sve2_sqrshrunt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_3(sve2_sqrshrunt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | |||
24 | +DEF_HELPER_FLAGS_3(sve2_sqshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_3(sve2_sqshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_3(sve2_sqshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
27 | + | ||
28 | +DEF_HELPER_FLAGS_3(sve2_sqshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_3(sve2_sqshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_3(sve2_sqshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
31 | + | ||
32 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
35 | + | ||
36 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
38 | +DEF_HELPER_FLAGS_3(sve2_sqrshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
39 | + | ||
40 | DEF_HELPER_FLAGS_3(sve2_uqshrnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
41 | DEF_HELPER_FLAGS_3(sve2_uqshrnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
42 | DEF_HELPER_FLAGS_3(sve2_uqshrnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
43 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
44 | index XXXXXXX..XXXXXXX 100644 | ||
45 | --- a/target/arm/sve.decode | ||
46 | +++ b/target/arm/sve.decode | ||
47 | @@ -XXX,XX +XXX,XX @@ SHRNB 01000101 .. 1 ..... 00 0100 ..... ..... @rd_rn_tszimm_shr | ||
48 | SHRNT 01000101 .. 1 ..... 00 0101 ..... ..... @rd_rn_tszimm_shr | ||
49 | RSHRNB 01000101 .. 1 ..... 00 0110 ..... ..... @rd_rn_tszimm_shr | ||
50 | RSHRNT 01000101 .. 1 ..... 00 0111 ..... ..... @rd_rn_tszimm_shr | ||
51 | +SQSHRNB 01000101 .. 1 ..... 00 1000 ..... ..... @rd_rn_tszimm_shr | ||
52 | +SQSHRNT 01000101 .. 1 ..... 00 1001 ..... ..... @rd_rn_tszimm_shr | ||
53 | +SQRSHRNB 01000101 .. 1 ..... 00 1010 ..... ..... @rd_rn_tszimm_shr | ||
54 | +SQRSHRNT 01000101 .. 1 ..... 00 1011 ..... ..... @rd_rn_tszimm_shr | ||
55 | UQSHRNB 01000101 .. 1 ..... 00 1100 ..... ..... @rd_rn_tszimm_shr | ||
56 | UQSHRNT 01000101 .. 1 ..... 00 1101 ..... ..... @rd_rn_tszimm_shr | ||
57 | UQRSHRNB 01000101 .. 1 ..... 00 1110 ..... ..... @rd_rn_tszimm_shr | ||
58 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/target/arm/sve_helper.c | ||
61 | +++ b/target/arm/sve_helper.c | ||
62 | @@ -XXX,XX +XXX,XX @@ DO_SHRNT(sve2_sqrshrunt_h, int16_t, uint8_t, H1_2, H1, DO_SQRSHRUN_H) | ||
63 | DO_SHRNT(sve2_sqrshrunt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQRSHRUN_S) | ||
64 | DO_SHRNT(sve2_sqrshrunt_d, int64_t, uint32_t, , H1_4, DO_SQRSHRUN_D) | ||
65 | |||
66 | +#define DO_SQSHRN_H(x, sh) do_sat_bhs(x >> sh, INT8_MIN, INT8_MAX) | ||
67 | +#define DO_SQSHRN_S(x, sh) do_sat_bhs(x >> sh, INT16_MIN, INT16_MAX) | ||
68 | +#define DO_SQSHRN_D(x, sh) do_sat_bhs(x >> sh, INT32_MIN, INT32_MAX) | ||
69 | + | ||
70 | +DO_SHRNB(sve2_sqshrnb_h, int16_t, uint8_t, DO_SQSHRN_H) | ||
71 | +DO_SHRNB(sve2_sqshrnb_s, int32_t, uint16_t, DO_SQSHRN_S) | ||
72 | +DO_SHRNB(sve2_sqshrnb_d, int64_t, uint32_t, DO_SQSHRN_D) | ||
73 | + | ||
74 | +DO_SHRNT(sve2_sqshrnt_h, int16_t, uint8_t, H1_2, H1, DO_SQSHRN_H) | ||
75 | +DO_SHRNT(sve2_sqshrnt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQSHRN_S) | ||
76 | +DO_SHRNT(sve2_sqshrnt_d, int64_t, uint32_t, , H1_4, DO_SQSHRN_D) | ||
77 | + | ||
78 | +#define DO_SQRSHRN_H(x, sh) do_sat_bhs(do_srshr(x, sh), INT8_MIN, INT8_MAX) | ||
79 | +#define DO_SQRSHRN_S(x, sh) do_sat_bhs(do_srshr(x, sh), INT16_MIN, INT16_MAX) | ||
80 | +#define DO_SQRSHRN_D(x, sh) do_sat_bhs(do_srshr(x, sh), INT32_MIN, INT32_MAX) | ||
81 | + | ||
82 | +DO_SHRNB(sve2_sqrshrnb_h, int16_t, uint8_t, DO_SQRSHRN_H) | ||
83 | +DO_SHRNB(sve2_sqrshrnb_s, int32_t, uint16_t, DO_SQRSHRN_S) | ||
84 | +DO_SHRNB(sve2_sqrshrnb_d, int64_t, uint32_t, DO_SQRSHRN_D) | ||
85 | + | ||
86 | +DO_SHRNT(sve2_sqrshrnt_h, int16_t, uint8_t, H1_2, H1, DO_SQRSHRN_H) | ||
87 | +DO_SHRNT(sve2_sqrshrnt_s, int32_t, uint16_t, H1_4, H1_2, DO_SQRSHRN_S) | ||
88 | +DO_SHRNT(sve2_sqrshrnt_d, int64_t, uint32_t, , H1_4, DO_SQRSHRN_D) | ||
89 | + | ||
90 | #define DO_UQSHRN_H(x, sh) MIN(x >> sh, UINT8_MAX) | ||
91 | #define DO_UQSHRN_S(x, sh) MIN(x >> sh, UINT16_MAX) | ||
92 | #define DO_UQSHRN_D(x, sh) MIN(x >> sh, UINT32_MAX) | ||
93 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
94 | index XXXXXXX..XXXXXXX 100644 | ||
95 | --- a/target/arm/translate-sve.c | ||
96 | +++ b/target/arm/translate-sve.c | ||
97 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQRSHRUNT(DisasContext *s, arg_rri_esz *a) | ||
98 | return do_sve2_shr_narrow(s, a, ops); | ||
99 | } | ||
100 | |||
101 | +static void gen_sqshrnb_vec(unsigned vece, TCGv_vec d, | ||
102 | + TCGv_vec n, int64_t shr) | ||
103 | +{ | ||
104 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
105 | + int halfbits = 4 << vece; | ||
106 | + int64_t max = MAKE_64BIT_MASK(0, halfbits - 1); | ||
107 | + int64_t min = -max - 1; | ||
108 | + | ||
109 | + tcg_gen_sari_vec(vece, n, n, shr); | ||
110 | + tcg_gen_dupi_vec(vece, t, min); | ||
111 | + tcg_gen_smax_vec(vece, n, n, t); | ||
112 | + tcg_gen_dupi_vec(vece, t, max); | ||
113 | + tcg_gen_smin_vec(vece, n, n, t); | ||
114 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
115 | + tcg_gen_and_vec(vece, d, n, t); | ||
116 | + tcg_temp_free_vec(t); | ||
117 | +} | ||
118 | + | ||
119 | +static bool trans_SQSHRNB(DisasContext *s, arg_rri_esz *a) | ||
120 | +{ | ||
121 | + static const TCGOpcode vec_list[] = { | ||
122 | + INDEX_op_sari_vec, INDEX_op_smax_vec, INDEX_op_smin_vec, 0 | ||
123 | + }; | ||
124 | + static const GVecGen2i ops[3] = { | ||
125 | + { .fniv = gen_sqshrnb_vec, | ||
126 | + .opt_opc = vec_list, | ||
127 | + .fno = gen_helper_sve2_sqshrnb_h, | ||
128 | + .vece = MO_16 }, | ||
129 | + { .fniv = gen_sqshrnb_vec, | ||
130 | + .opt_opc = vec_list, | ||
131 | + .fno = gen_helper_sve2_sqshrnb_s, | ||
132 | + .vece = MO_32 }, | ||
133 | + { .fniv = gen_sqshrnb_vec, | ||
134 | + .opt_opc = vec_list, | ||
135 | + .fno = gen_helper_sve2_sqshrnb_d, | ||
136 | + .vece = MO_64 }, | ||
137 | + }; | ||
138 | + return do_sve2_shr_narrow(s, a, ops); | ||
139 | +} | ||
140 | + | ||
141 | +static void gen_sqshrnt_vec(unsigned vece, TCGv_vec d, | ||
142 | + TCGv_vec n, int64_t shr) | ||
143 | +{ | ||
144 | + TCGv_vec t = tcg_temp_new_vec_matching(d); | ||
145 | + int halfbits = 4 << vece; | ||
146 | + int64_t max = MAKE_64BIT_MASK(0, halfbits - 1); | ||
147 | + int64_t min = -max - 1; | ||
148 | + | ||
149 | + tcg_gen_sari_vec(vece, n, n, shr); | ||
150 | + tcg_gen_dupi_vec(vece, t, min); | ||
151 | + tcg_gen_smax_vec(vece, n, n, t); | ||
152 | + tcg_gen_dupi_vec(vece, t, max); | ||
153 | + tcg_gen_smin_vec(vece, n, n, t); | ||
154 | + tcg_gen_shli_vec(vece, n, n, halfbits); | ||
155 | + tcg_gen_dupi_vec(vece, t, MAKE_64BIT_MASK(0, halfbits)); | ||
156 | + tcg_gen_bitsel_vec(vece, d, t, d, n); | ||
157 | + tcg_temp_free_vec(t); | ||
158 | +} | ||
159 | + | ||
160 | +static bool trans_SQSHRNT(DisasContext *s, arg_rri_esz *a) | ||
161 | +{ | ||
162 | + static const TCGOpcode vec_list[] = { | ||
163 | + INDEX_op_shli_vec, INDEX_op_sari_vec, | ||
164 | + INDEX_op_smax_vec, INDEX_op_smin_vec, 0 | ||
165 | + }; | ||
166 | + static const GVecGen2i ops[3] = { | ||
167 | + { .fniv = gen_sqshrnt_vec, | ||
168 | + .opt_opc = vec_list, | ||
169 | + .load_dest = true, | ||
170 | + .fno = gen_helper_sve2_sqshrnt_h, | ||
171 | + .vece = MO_16 }, | ||
172 | + { .fniv = gen_sqshrnt_vec, | ||
173 | + .opt_opc = vec_list, | ||
174 | + .load_dest = true, | ||
175 | + .fno = gen_helper_sve2_sqshrnt_s, | ||
176 | + .vece = MO_32 }, | ||
177 | + { .fniv = gen_sqshrnt_vec, | ||
178 | + .opt_opc = vec_list, | ||
179 | + .load_dest = true, | ||
180 | + .fno = gen_helper_sve2_sqshrnt_d, | ||
181 | + .vece = MO_64 }, | ||
182 | + }; | ||
183 | + return do_sve2_shr_narrow(s, a, ops); | ||
184 | +} | ||
185 | + | ||
186 | +static bool trans_SQRSHRNB(DisasContext *s, arg_rri_esz *a) | ||
187 | +{ | ||
188 | + static const GVecGen2i ops[3] = { | ||
189 | + { .fno = gen_helper_sve2_sqrshrnb_h }, | ||
190 | + { .fno = gen_helper_sve2_sqrshrnb_s }, | ||
191 | + { .fno = gen_helper_sve2_sqrshrnb_d }, | ||
192 | + }; | ||
193 | + return do_sve2_shr_narrow(s, a, ops); | ||
194 | +} | ||
195 | + | ||
196 | +static bool trans_SQRSHRNT(DisasContext *s, arg_rri_esz *a) | ||
197 | +{ | ||
198 | + static const GVecGen2i ops[3] = { | ||
199 | + { .fno = gen_helper_sve2_sqrshrnt_h }, | ||
200 | + { .fno = gen_helper_sve2_sqrshrnt_s }, | ||
201 | + { .fno = gen_helper_sve2_sqrshrnt_d }, | ||
202 | + }; | ||
203 | + return do_sve2_shr_narrow(s, a, ops); | ||
204 | +} | ||
205 | + | ||
206 | static void gen_uqshrnb_vec(unsigned vece, TCGv_vec d, | ||
207 | TCGv_vec n, int64_t shr) | ||
208 | { | ||
209 | -- | ||
210 | 2.20.1 | ||
211 | |||
212 | diff view generated by jsdifflib |
1 | From: Stephen Long <steplong@quicinc.com> | 1 | From: Inès Varhol <ines.varhol@telecom-paris.fr> |
---|---|---|---|
2 | 2 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 3 | Features supported : |
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | 4 | - the 8 STM32L4x5 GPIOs are initialized with their reset values |
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 5 | (except IDR, see below) |
6 | Message-id: 20210525010358.152808-43-richard.henderson@linaro.org | 6 | - input mode : setting a pin in input mode "externally" (using input |
7 | Message-Id: <20200416173109.8856-1-steplong@quicinc.com> | 7 | irqs) results in an out irq (transmitted to SYSCFG) |
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 8 | - output mode : setting a bit in ODR sets the corresponding out irq |
9 | (if this line is configured in output mode) | ||
10 | - pull-up, pull-down | ||
11 | - push-pull, open-drain | ||
12 | |||
13 | Difference with the real GPIOs : | ||
14 | - Alternate Function and Analog mode aren't implemented : | ||
15 | pins in AF/Analog behave like pins in input mode | ||
16 | - floating pins stay at their last value | ||
17 | - register IDR reset values differ from the real one : | ||
18 | values are coherent with the other registers reset values | ||
19 | and the fact that AF/Analog modes aren't implemented | ||
20 | - setting I/O output speed isn't supported | ||
21 | - locking port bits isn't supported | ||
22 | - ADC function isn't supported | ||
23 | - GPIOH has 16 pins instead of 2 pins | ||
24 | - writing to registers LCKR, AFRL, AFRH and ASCR is ineffective | ||
25 | |||
26 | Signed-off-by: Arnaud Minier <arnaud.minier@telecom-paris.fr> | ||
27 | Signed-off-by: Inès Varhol <ines.varhol@telecom-paris.fr> | ||
28 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
29 | Acked-by: Alistair Francis <alistair.francis@wdc.com> | ||
30 | Message-id: 20240305210444.310665-2-ines.varhol@telecom-paris.fr | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 31 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
10 | --- | 32 | --- |
11 | target/arm/helper-sve.h | 7 ++ | 33 | MAINTAINERS | 1 + |
12 | target/arm/sve.decode | 6 ++ | 34 | docs/system/arm/b-l475e-iot01a.rst | 2 +- |
13 | target/arm/sve_helper.c | 131 +++++++++++++++++++++++++++++++++++++ | 35 | include/hw/gpio/stm32l4x5_gpio.h | 70 +++++ |
14 | target/arm/translate-sve.c | 19 ++++++ | 36 | hw/gpio/stm32l4x5_gpio.c | 477 +++++++++++++++++++++++++++++ |
15 | 4 files changed, 163 insertions(+) | 37 | hw/gpio/Kconfig | 3 + |
38 | hw/gpio/meson.build | 1 + | ||
39 | hw/gpio/trace-events | 6 + | ||
40 | 7 files changed, 559 insertions(+), 1 deletion(-) | ||
41 | create mode 100644 include/hw/gpio/stm32l4x5_gpio.h | ||
42 | create mode 100644 hw/gpio/stm32l4x5_gpio.c | ||
16 | 43 | ||
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | 44 | diff --git a/MAINTAINERS b/MAINTAINERS |
18 | index XXXXXXX..XXXXXXX 100644 | 45 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/target/arm/helper-sve.h | 46 | --- a/MAINTAINERS |
20 | +++ b/target/arm/helper-sve.h | 47 | +++ b/MAINTAINERS |
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_nmatch_ppzz_b, TCG_CALL_NO_RWG, | 48 | @@ -XXX,XX +XXX,XX @@ F: hw/arm/stm32l4x5_soc.c |
22 | DEF_HELPER_FLAGS_5(sve2_nmatch_ppzz_h, TCG_CALL_NO_RWG, | 49 | F: hw/misc/stm32l4x5_exti.c |
23 | i32, ptr, ptr, ptr, ptr, i32) | 50 | F: hw/misc/stm32l4x5_syscfg.c |
24 | 51 | F: hw/misc/stm32l4x5_rcc.c | |
25 | +DEF_HELPER_FLAGS_5(sve2_histcnt_s, TCG_CALL_NO_RWG, | 52 | +F: hw/gpio/stm32l4x5_gpio.c |
26 | + void, ptr, ptr, ptr, ptr, i32) | 53 | F: include/hw/*/stm32l4x5_*.h |
27 | +DEF_HELPER_FLAGS_5(sve2_histcnt_d, TCG_CALL_NO_RWG, | 54 | |
28 | + void, ptr, ptr, ptr, ptr, i32) | 55 | B-L475E-IOT01A IoT Node |
29 | + | 56 | diff --git a/docs/system/arm/b-l475e-iot01a.rst b/docs/system/arm/b-l475e-iot01a.rst |
30 | +DEF_HELPER_FLAGS_4(sve2_histseg, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | + | ||
32 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
33 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
34 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
35 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
36 | index XXXXXXX..XXXXXXX 100644 | 57 | index XXXXXXX..XXXXXXX 100644 |
37 | --- a/target/arm/sve.decode | 58 | --- a/docs/system/arm/b-l475e-iot01a.rst |
38 | +++ b/target/arm/sve.decode | 59 | +++ b/docs/system/arm/b-l475e-iot01a.rst |
60 | @@ -XXX,XX +XXX,XX @@ Currently B-L475E-IOT01A machine's only supports the following devices: | ||
61 | - STM32L4x5 EXTI (Extended interrupts and events controller) | ||
62 | - STM32L4x5 SYSCFG (System configuration controller) | ||
63 | - STM32L4x5 RCC (Reset and clock control) | ||
64 | +- STM32L4x5 GPIOs (General-purpose I/Os) | ||
65 | |||
66 | Missing devices | ||
67 | """"""""""""""" | ||
68 | @@ -XXX,XX +XXX,XX @@ Missing devices | ||
69 | The B-L475E-IOT01A does *not* support the following devices: | ||
70 | |||
71 | - Serial ports (UART) | ||
72 | -- General-purpose I/Os (GPIO) | ||
73 | - Analog to Digital Converter (ADC) | ||
74 | - SPI controller | ||
75 | - Timer controller (TIMER) | ||
76 | diff --git a/include/hw/gpio/stm32l4x5_gpio.h b/include/hw/gpio/stm32l4x5_gpio.h | ||
77 | new file mode 100644 | ||
78 | index XXXXXXX..XXXXXXX | ||
79 | --- /dev/null | ||
80 | +++ b/include/hw/gpio/stm32l4x5_gpio.h | ||
39 | @@ -XXX,XX +XXX,XX @@ | 81 | @@ -XXX,XX +XXX,XX @@ |
40 | &rprrr_esz rn=%reg_movprfx | 82 | +/* |
41 | @rdn_pg_rm_ra ........ esz:2 . ra:5 ... pg:3 rm:5 rd:5 \ | 83 | + * STM32L4x5 GPIO (General Purpose Input/Ouput) |
42 | &rprrr_esz rn=%reg_movprfx | 84 | + * |
43 | +@rd_pg_rn_rm ........ esz:2 . rm:5 ... pg:3 rn:5 rd:5 &rprr_esz | 85 | + * Copyright (c) 2024 Arnaud Minier <arnaud.minier@telecom-paris.fr> |
44 | 86 | + * Copyright (c) 2024 Inès Varhol <ines.varhol@telecom-paris.fr> | |
45 | # One register operand, with governing predicate, vector element size | 87 | + * |
46 | @rd_pg_rn ........ esz:2 ... ... ... pg:3 rn:5 rd:5 &rpr_esz | 88 | + * SPDX-License-Identifier: GPL-2.0-or-later |
47 | @@ -XXX,XX +XXX,XX @@ RSUBHNT 01000101 .. 1 ..... 011 111 ..... ..... @rd_rn_rm | 89 | + * |
48 | MATCH 01000101 .. 1 ..... 100 ... ..... 0 .... @pd_pg_rn_rm | 90 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. |
49 | NMATCH 01000101 .. 1 ..... 100 ... ..... 1 .... @pd_pg_rn_rm | 91 | + * See the COPYING file in the top-level directory. |
50 | 92 | + */ | |
51 | +### SVE2 Histogram Computation | 93 | + |
52 | + | 94 | +/* |
53 | +HISTCNT 01000101 .. 1 ..... 110 ... ..... ..... @rd_pg_rn_rm | 95 | + * The reference used is the STMicroElectronics RM0351 Reference manual |
54 | +HISTSEG 01000101 .. 1 ..... 101 000 ..... ..... @rd_rn_rm | 96 | + * for STM32L4x5 and STM32L4x6 advanced Arm ® -based 32-bit MCUs. |
55 | + | 97 | + * https://www.st.com/en/microcontrollers-microprocessors/stm32l4x5/documentation.html |
56 | ## SVE2 floating-point pairwise operations | 98 | + */ |
57 | 99 | + | |
58 | FADDP 01100100 .. 010 00 0 100 ... ..... ..... @rdn_pg_rm | 100 | +#ifndef HW_STM32L4X5_GPIO_H |
59 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | 101 | +#define HW_STM32L4X5_GPIO_H |
60 | index XXXXXXX..XXXXXXX 100644 | 102 | + |
61 | --- a/target/arm/sve_helper.c | 103 | +#include "hw/sysbus.h" |
62 | +++ b/target/arm/sve_helper.c | 104 | +#include "qom/object.h" |
63 | @@ -XXX,XX +XXX,XX @@ DO_PPZZ_MATCH(sve2_nmatch_ppzz_b, MO_8, true) | 105 | + |
64 | DO_PPZZ_MATCH(sve2_nmatch_ppzz_h, MO_16, true) | 106 | +#define TYPE_STM32L4X5_GPIO "stm32l4x5-gpio" |
65 | 107 | +OBJECT_DECLARE_SIMPLE_TYPE(Stm32l4x5GpioState, STM32L4X5_GPIO) | |
66 | #undef DO_PPZZ_MATCH | 108 | + |
67 | + | 109 | +#define GPIO_NUM_PINS 16 |
68 | +void HELPER(sve2_histcnt_s)(void *vd, void *vn, void *vm, void *vg, | 110 | + |
69 | + uint32_t desc) | 111 | +struct Stm32l4x5GpioState { |
70 | +{ | 112 | + SysBusDevice parent_obj; |
71 | + ARMVectorReg scratch; | 113 | + |
72 | + intptr_t i, j; | 114 | + MemoryRegion mmio; |
73 | + intptr_t opr_sz = simd_oprsz(desc); | 115 | + |
74 | + uint32_t *d = vd, *n = vn, *m = vm; | 116 | + /* GPIO registers */ |
75 | + uint8_t *pg = vg; | 117 | + uint32_t moder; |
76 | + | 118 | + uint32_t otyper; |
77 | + if (d == n) { | 119 | + uint32_t ospeedr; |
78 | + n = memcpy(&scratch, n, opr_sz); | 120 | + uint32_t pupdr; |
79 | + if (d == m) { | 121 | + uint32_t idr; |
80 | + m = n; | 122 | + uint32_t odr; |
123 | + uint32_t lckr; | ||
124 | + uint32_t afrl; | ||
125 | + uint32_t afrh; | ||
126 | + uint32_t ascr; | ||
127 | + | ||
128 | + /* GPIO registers reset values */ | ||
129 | + uint32_t moder_reset; | ||
130 | + uint32_t ospeedr_reset; | ||
131 | + uint32_t pupdr_reset; | ||
132 | + | ||
133 | + /* | ||
134 | + * External driving of pins. | ||
135 | + * The pins can be set externally through the device | ||
136 | + * anonymous input GPIOs lines under certain conditions. | ||
137 | + * The pin must not be in push-pull output mode, | ||
138 | + * and can't be set high in open-drain mode. | ||
139 | + * Pins driven externally and configured to | ||
140 | + * output mode will in general be "disconnected" | ||
141 | + * (see `get_gpio_pinmask_to_disconnect()`) | ||
142 | + */ | ||
143 | + uint16_t disconnected_pins; | ||
144 | + uint16_t pins_connected_high; | ||
145 | + | ||
146 | + char *name; | ||
147 | + Clock *clk; | ||
148 | + qemu_irq pin[GPIO_NUM_PINS]; | ||
149 | +}; | ||
150 | + | ||
151 | +#endif | ||
152 | diff --git a/hw/gpio/stm32l4x5_gpio.c b/hw/gpio/stm32l4x5_gpio.c | ||
153 | new file mode 100644 | ||
154 | index XXXXXXX..XXXXXXX | ||
155 | --- /dev/null | ||
156 | +++ b/hw/gpio/stm32l4x5_gpio.c | ||
157 | @@ -XXX,XX +XXX,XX @@ | ||
158 | +/* | ||
159 | + * STM32L4x5 GPIO (General Purpose Input/Ouput) | ||
160 | + * | ||
161 | + * Copyright (c) 2024 Arnaud Minier <arnaud.minier@telecom-paris.fr> | ||
162 | + * Copyright (c) 2024 Inès Varhol <ines.varhol@telecom-paris.fr> | ||
163 | + * | ||
164 | + * SPDX-License-Identifier: GPL-2.0-or-later | ||
165 | + * | ||
166 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. | ||
167 | + * See the COPYING file in the top-level directory. | ||
168 | + */ | ||
169 | + | ||
170 | +/* | ||
171 | + * The reference used is the STMicroElectronics RM0351 Reference manual | ||
172 | + * for STM32L4x5 and STM32L4x6 advanced Arm ® -based 32-bit MCUs. | ||
173 | + * https://www.st.com/en/microcontrollers-microprocessors/stm32l4x5/documentation.html | ||
174 | + */ | ||
175 | + | ||
176 | +#include "qemu/osdep.h" | ||
177 | +#include "qemu/log.h" | ||
178 | +#include "hw/gpio/stm32l4x5_gpio.h" | ||
179 | +#include "hw/irq.h" | ||
180 | +#include "hw/qdev-clock.h" | ||
181 | +#include "hw/qdev-properties.h" | ||
182 | +#include "qapi/visitor.h" | ||
183 | +#include "qapi/error.h" | ||
184 | +#include "migration/vmstate.h" | ||
185 | +#include "trace.h" | ||
186 | + | ||
187 | +#define GPIO_MODER 0x00 | ||
188 | +#define GPIO_OTYPER 0x04 | ||
189 | +#define GPIO_OSPEEDR 0x08 | ||
190 | +#define GPIO_PUPDR 0x0C | ||
191 | +#define GPIO_IDR 0x10 | ||
192 | +#define GPIO_ODR 0x14 | ||
193 | +#define GPIO_BSRR 0x18 | ||
194 | +#define GPIO_LCKR 0x1C | ||
195 | +#define GPIO_AFRL 0x20 | ||
196 | +#define GPIO_AFRH 0x24 | ||
197 | +#define GPIO_BRR 0x28 | ||
198 | +#define GPIO_ASCR 0x2C | ||
199 | + | ||
200 | +/* 0b11111111_11111111_00000000_00000000 */ | ||
201 | +#define RESERVED_BITS_MASK 0xFFFF0000 | ||
202 | + | ||
203 | +static void update_gpio_idr(Stm32l4x5GpioState *s); | ||
204 | + | ||
205 | +static bool is_pull_up(Stm32l4x5GpioState *s, unsigned pin) | ||
206 | +{ | ||
207 | + return extract32(s->pupdr, 2 * pin, 2) == 1; | ||
208 | +} | ||
209 | + | ||
210 | +static bool is_pull_down(Stm32l4x5GpioState *s, unsigned pin) | ||
211 | +{ | ||
212 | + return extract32(s->pupdr, 2 * pin, 2) == 2; | ||
213 | +} | ||
214 | + | ||
215 | +static bool is_output(Stm32l4x5GpioState *s, unsigned pin) | ||
216 | +{ | ||
217 | + return extract32(s->moder, 2 * pin, 2) == 1; | ||
218 | +} | ||
219 | + | ||
220 | +static bool is_open_drain(Stm32l4x5GpioState *s, unsigned pin) | ||
221 | +{ | ||
222 | + return extract32(s->otyper, pin, 1) == 1; | ||
223 | +} | ||
224 | + | ||
225 | +static bool is_push_pull(Stm32l4x5GpioState *s, unsigned pin) | ||
226 | +{ | ||
227 | + return extract32(s->otyper, pin, 1) == 0; | ||
228 | +} | ||
229 | + | ||
230 | +static void stm32l4x5_gpio_reset_hold(Object *obj) | ||
231 | +{ | ||
232 | + Stm32l4x5GpioState *s = STM32L4X5_GPIO(obj); | ||
233 | + | ||
234 | + s->moder = s->moder_reset; | ||
235 | + s->otyper = 0x00000000; | ||
236 | + s->ospeedr = s->ospeedr_reset; | ||
237 | + s->pupdr = s->pupdr_reset; | ||
238 | + s->idr = 0x00000000; | ||
239 | + s->odr = 0x00000000; | ||
240 | + s->lckr = 0x00000000; | ||
241 | + s->afrl = 0x00000000; | ||
242 | + s->afrh = 0x00000000; | ||
243 | + s->ascr = 0x00000000; | ||
244 | + | ||
245 | + s->disconnected_pins = 0xFFFF; | ||
246 | + s->pins_connected_high = 0x0000; | ||
247 | + update_gpio_idr(s); | ||
248 | +} | ||
249 | + | ||
250 | +static void stm32l4x5_gpio_set(void *opaque, int line, int level) | ||
251 | +{ | ||
252 | + Stm32l4x5GpioState *s = opaque; | ||
253 | + /* | ||
254 | + * The pin isn't set if line is configured in output mode | ||
255 | + * except if level is 0 and the output is open-drain. | ||
256 | + * This way there will be no short-circuit prone situations. | ||
257 | + */ | ||
258 | + if (is_output(s, line) && !(is_open_drain(s, line) && (level == 0))) { | ||
259 | + qemu_log_mask(LOG_GUEST_ERROR, "Line %d can't be driven externally\n", | ||
260 | + line); | ||
261 | + return; | ||
262 | + } | ||
263 | + | ||
264 | + s->disconnected_pins &= ~(1 << line); | ||
265 | + if (level) { | ||
266 | + s->pins_connected_high |= (1 << line); | ||
267 | + } else { | ||
268 | + s->pins_connected_high &= ~(1 << line); | ||
269 | + } | ||
270 | + trace_stm32l4x5_gpio_pins(s->name, s->disconnected_pins, | ||
271 | + s->pins_connected_high); | ||
272 | + update_gpio_idr(s); | ||
273 | +} | ||
274 | + | ||
275 | + | ||
276 | +static void update_gpio_idr(Stm32l4x5GpioState *s) | ||
277 | +{ | ||
278 | + uint32_t new_idr_mask = 0; | ||
279 | + uint32_t new_idr = s->odr; | ||
280 | + uint32_t old_idr = s->idr; | ||
281 | + int new_pin_state, old_pin_state; | ||
282 | + | ||
283 | + for (int i = 0; i < GPIO_NUM_PINS; i++) { | ||
284 | + if (is_output(s, i)) { | ||
285 | + if (is_push_pull(s, i)) { | ||
286 | + new_idr_mask |= (1 << i); | ||
287 | + } else if (!(s->odr & (1 << i))) { | ||
288 | + /* open-drain ODR 0 */ | ||
289 | + new_idr_mask |= (1 << i); | ||
290 | + /* open-drain ODR 1 */ | ||
291 | + } else if (!(s->disconnected_pins & (1 << i)) && | ||
292 | + !(s->pins_connected_high & (1 << i))) { | ||
293 | + /* open-drain ODR 1 with pin connected low */ | ||
294 | + new_idr_mask |= (1 << i); | ||
295 | + new_idr &= ~(1 << i); | ||
296 | + /* open-drain ODR 1 with unactive pin */ | ||
297 | + } else if (is_pull_up(s, i)) { | ||
298 | + new_idr_mask |= (1 << i); | ||
299 | + } else if (is_pull_down(s, i)) { | ||
300 | + new_idr_mask |= (1 << i); | ||
301 | + new_idr &= ~(1 << i); | ||
302 | + } | ||
303 | + /* | ||
304 | + * The only case left is for open-drain ODR 1 | ||
305 | + * with unactive pin without pull-up or pull-down : | ||
306 | + * the value is floating. | ||
307 | + */ | ||
308 | + /* input or analog mode with connected pin */ | ||
309 | + } else if (!(s->disconnected_pins & (1 << i))) { | ||
310 | + if (s->pins_connected_high & (1 << i)) { | ||
311 | + /* pin high */ | ||
312 | + new_idr_mask |= (1 << i); | ||
313 | + new_idr |= (1 << i); | ||
314 | + } else { | ||
315 | + /* pin low */ | ||
316 | + new_idr_mask |= (1 << i); | ||
317 | + new_idr &= ~(1 << i); | ||
318 | + } | ||
319 | + /* input or analog mode with disconnected pin */ | ||
320 | + } else { | ||
321 | + if (is_pull_up(s, i)) { | ||
322 | + /* pull-up */ | ||
323 | + new_idr_mask |= (1 << i); | ||
324 | + new_idr |= (1 << i); | ||
325 | + } else if (is_pull_down(s, i)) { | ||
326 | + /* pull-down */ | ||
327 | + new_idr_mask |= (1 << i); | ||
328 | + new_idr &= ~(1 << i); | ||
329 | + } | ||
330 | + /* | ||
331 | + * The only case left is for a disconnected pin | ||
332 | + * without pull-up or pull-down : | ||
333 | + * the value is floating. | ||
334 | + */ | ||
81 | + } | 335 | + } |
82 | + } else if (d == m) { | 336 | + } |
83 | + m = memcpy(&scratch, m, opr_sz); | 337 | + |
84 | + } | 338 | + s->idr = (old_idr & ~new_idr_mask) | (new_idr & new_idr_mask); |
85 | + | 339 | + trace_stm32l4x5_gpio_update_idr(s->name, old_idr, s->idr); |
86 | + for (i = 0; i < opr_sz; i += 4) { | 340 | + |
87 | + uint64_t count = 0; | 341 | + for (int i = 0; i < GPIO_NUM_PINS; i++) { |
88 | + uint8_t pred; | 342 | + if (new_idr_mask & (1 << i)) { |
89 | + | 343 | + new_pin_state = (new_idr & (1 << i)) > 0; |
90 | + pred = pg[H1(i >> 3)] >> (i & 7); | 344 | + old_pin_state = (old_idr & (1 << i)) > 0; |
91 | + if (pred & 1) { | 345 | + if (new_pin_state > old_pin_state) { |
92 | + uint32_t nn = n[H4(i >> 2)]; | 346 | + qemu_irq_raise(s->pin[i]); |
93 | + | 347 | + } else if (new_pin_state < old_pin_state) { |
94 | + for (j = 0; j <= i; j += 4) { | 348 | + qemu_irq_lower(s->pin[i]); |
95 | + pred = pg[H1(j >> 3)] >> (j & 7); | ||
96 | + if ((pred & 1) && nn == m[H4(j >> 2)]) { | ||
97 | + ++count; | ||
98 | + } | ||
99 | + } | 349 | + } |
100 | + } | 350 | + } |
101 | + d[H4(i >> 2)] = count; | 351 | + } |
102 | + } | 352 | +} |
103 | +} | 353 | + |
104 | + | 354 | +/* |
105 | +void HELPER(sve2_histcnt_d)(void *vd, void *vn, void *vm, void *vg, | 355 | + * Return mask of pins that are both configured in output |
106 | + uint32_t desc) | 356 | + * mode and externally driven (except pins in open-drain |
107 | +{ | 357 | + * mode externally set to 0). |
108 | + ARMVectorReg scratch; | 358 | + */ |
109 | + intptr_t i, j; | 359 | +static uint32_t get_gpio_pinmask_to_disconnect(Stm32l4x5GpioState *s) |
110 | + intptr_t opr_sz = simd_oprsz(desc); | 360 | +{ |
111 | + uint64_t *d = vd, *n = vn, *m = vm; | 361 | + uint32_t pins_to_disconnect = 0; |
112 | + uint8_t *pg = vg; | 362 | + for (int i = 0; i < GPIO_NUM_PINS; i++) { |
113 | + | 363 | + /* for each connected pin in output mode */ |
114 | + if (d == n) { | 364 | + if (!(s->disconnected_pins & (1 << i)) && is_output(s, i)) { |
115 | + n = memcpy(&scratch, n, opr_sz); | 365 | + /* if either push-pull or high level */ |
116 | + if (d == m) { | 366 | + if (is_push_pull(s, i) || s->pins_connected_high & (1 << i)) { |
117 | + m = n; | 367 | + pins_to_disconnect |= (1 << i); |
118 | + } | 368 | + qemu_log_mask(LOG_GUEST_ERROR, |
119 | + } else if (d == m) { | 369 | + "Line %d can't be driven externally\n", |
120 | + m = memcpy(&scratch, m, opr_sz); | 370 | + i); |
121 | + } | ||
122 | + | ||
123 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
124 | + uint64_t count = 0; | ||
125 | + if (pg[H1(i)] & 1) { | ||
126 | + uint64_t nn = n[i]; | ||
127 | + for (j = 0; j <= i; ++j) { | ||
128 | + if ((pg[H1(j)] & 1) && nn == m[j]) { | ||
129 | + ++count; | ||
130 | + } | ||
131 | + } | 371 | + } |
132 | + } | 372 | + } |
133 | + d[i] = count; | 373 | + } |
134 | + } | 374 | + return pins_to_disconnect; |
135 | +} | 375 | +} |
136 | + | 376 | + |
137 | +/* | 377 | +/* |
138 | + * Returns the number of bytes in m0 and m1 that match n. | 378 | + * Set field `disconnected_pins` and call `update_gpio_idr()` |
139 | + * Unlike do_match2 we don't just need true/false, we need an exact count. | ||
140 | + * This requires two extra logical operations. | ||
141 | + */ | 379 | + */ |
142 | +static inline uint64_t do_histseg_cnt(uint8_t n, uint64_t m0, uint64_t m1) | 380 | +static void disconnect_gpio_pins(Stm32l4x5GpioState *s, uint16_t lines) |
143 | +{ | 381 | +{ |
144 | + const uint64_t mask = dup_const(MO_8, 0x7f); | 382 | + s->disconnected_pins |= lines; |
145 | + uint64_t cmp0, cmp1; | 383 | + trace_stm32l4x5_gpio_pins(s->name, s->disconnected_pins, |
146 | + | 384 | + s->pins_connected_high); |
147 | + cmp1 = dup_const(MO_8, n); | 385 | + update_gpio_idr(s); |
148 | + cmp0 = cmp1 ^ m0; | 386 | +} |
149 | + cmp1 = cmp1 ^ m1; | 387 | + |
150 | + | 388 | +static void disconnected_pins_set(Object *obj, Visitor *v, |
151 | + /* | 389 | + const char *name, void *opaque, Error **errp) |
152 | + * 1: clear msb of each byte to avoid carry to next byte (& mask) | 390 | +{ |
153 | + * 2: carry in to msb if byte != 0 (+ mask) | 391 | + Stm32l4x5GpioState *s = STM32L4X5_GPIO(obj); |
154 | + * 3: set msb if cmp has msb set (| cmp) | 392 | + uint16_t value; |
155 | + * 4: set ~msb to ignore them (| mask) | 393 | + if (!visit_type_uint16(v, name, &value, errp)) { |
156 | + * We now have 0xff for byte != 0 or 0x7f for byte == 0. | 394 | + return; |
157 | + * 5: invert, resulting in 0x80 if and only if byte == 0. | 395 | + } |
158 | + */ | 396 | + disconnect_gpio_pins(s, value); |
159 | + cmp0 = ~(((cmp0 & mask) + mask) | cmp0 | mask); | 397 | +} |
160 | + cmp1 = ~(((cmp1 & mask) + mask) | cmp1 | mask); | 398 | + |
161 | + | 399 | +static void disconnected_pins_get(Object *obj, Visitor *v, |
162 | + /* | 400 | + const char *name, void *opaque, Error **errp) |
163 | + * Combine the two compares in a way that the bits do | 401 | +{ |
164 | + * not overlap, and so preserves the count of set bits. | 402 | + visit_type_uint16(v, name, (uint16_t *)opaque, errp); |
165 | + * If the host has an efficient instruction for ctpop, | 403 | +} |
166 | + * then ctpop(x) + ctpop(y) has the same number of | 404 | + |
167 | + * operations as ctpop(x | (y >> 1)). If the host does | 405 | +static void clock_freq_get(Object *obj, Visitor *v, |
168 | + * not have an efficient ctpop, then we only want to | 406 | + const char *name, void *opaque, Error **errp) |
169 | + * use it once. | 407 | +{ |
170 | + */ | 408 | + Stm32l4x5GpioState *s = STM32L4X5_GPIO(obj); |
171 | + return ctpop64(cmp0 | (cmp1 >> 1)); | 409 | + uint32_t clock_freq_hz = clock_get_hz(s->clk); |
172 | +} | 410 | + visit_type_uint32(v, name, &clock_freq_hz, errp); |
173 | + | 411 | +} |
174 | +void HELPER(sve2_histseg)(void *vd, void *vn, void *vm, uint32_t desc) | 412 | + |
175 | +{ | 413 | +static void stm32l4x5_gpio_write(void *opaque, hwaddr addr, |
176 | + intptr_t i, j; | 414 | + uint64_t val64, unsigned int size) |
177 | + intptr_t opr_sz = simd_oprsz(desc); | 415 | +{ |
178 | + | 416 | + Stm32l4x5GpioState *s = opaque; |
179 | + for (i = 0; i < opr_sz; i += 16) { | 417 | + |
180 | + uint64_t n0 = *(uint64_t *)(vn + i); | 418 | + uint32_t value = val64; |
181 | + uint64_t m0 = *(uint64_t *)(vm + i); | 419 | + trace_stm32l4x5_gpio_write(s->name, addr, val64); |
182 | + uint64_t n1 = *(uint64_t *)(vn + i + 8); | 420 | + |
183 | + uint64_t m1 = *(uint64_t *)(vm + i + 8); | 421 | + switch (addr) { |
184 | + uint64_t out0 = 0; | 422 | + case GPIO_MODER: |
185 | + uint64_t out1 = 0; | 423 | + s->moder = value; |
186 | + | 424 | + disconnect_gpio_pins(s, get_gpio_pinmask_to_disconnect(s)); |
187 | + for (j = 0; j < 64; j += 8) { | 425 | + qemu_log_mask(LOG_UNIMP, |
188 | + uint64_t cnt0 = do_histseg_cnt(n0 >> j, m0, m1); | 426 | + "%s: Analog and AF modes aren't supported\n\ |
189 | + uint64_t cnt1 = do_histseg_cnt(n1 >> j, m0, m1); | 427 | + Analog and AF mode behave like input mode\n", |
190 | + out0 |= cnt0 << j; | 428 | + __func__); |
191 | + out1 |= cnt1 << j; | 429 | + return; |
192 | + } | 430 | + case GPIO_OTYPER: |
193 | + | 431 | + s->otyper = value & ~RESERVED_BITS_MASK; |
194 | + *(uint64_t *)(vd + i) = out0; | 432 | + disconnect_gpio_pins(s, get_gpio_pinmask_to_disconnect(s)); |
195 | + *(uint64_t *)(vd + i + 8) = out1; | 433 | + return; |
196 | + } | 434 | + case GPIO_OSPEEDR: |
197 | +} | 435 | + qemu_log_mask(LOG_UNIMP, |
198 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | 436 | + "%s: Changing I/O output speed isn't supported\n\ |
437 | + I/O speed is already maximal\n", | ||
438 | + __func__); | ||
439 | + s->ospeedr = value; | ||
440 | + return; | ||
441 | + case GPIO_PUPDR: | ||
442 | + s->pupdr = value; | ||
443 | + update_gpio_idr(s); | ||
444 | + return; | ||
445 | + case GPIO_IDR: | ||
446 | + qemu_log_mask(LOG_UNIMP, | ||
447 | + "%s: GPIO->IDR is read-only\n", | ||
448 | + __func__); | ||
449 | + return; | ||
450 | + case GPIO_ODR: | ||
451 | + s->odr = value & ~RESERVED_BITS_MASK; | ||
452 | + update_gpio_idr(s); | ||
453 | + return; | ||
454 | + case GPIO_BSRR: { | ||
455 | + uint32_t bits_to_reset = (value & RESERVED_BITS_MASK) >> GPIO_NUM_PINS; | ||
456 | + uint32_t bits_to_set = value & ~RESERVED_BITS_MASK; | ||
457 | + /* If both BSx and BRx are set, BSx has priority.*/ | ||
458 | + s->odr &= ~bits_to_reset; | ||
459 | + s->odr |= bits_to_set; | ||
460 | + update_gpio_idr(s); | ||
461 | + return; | ||
462 | + } | ||
463 | + case GPIO_LCKR: | ||
464 | + qemu_log_mask(LOG_UNIMP, | ||
465 | + "%s: Locking port bits configuration isn't supported\n", | ||
466 | + __func__); | ||
467 | + s->lckr = value & ~RESERVED_BITS_MASK; | ||
468 | + return; | ||
469 | + case GPIO_AFRL: | ||
470 | + qemu_log_mask(LOG_UNIMP, | ||
471 | + "%s: Alternate functions aren't supported\n", | ||
472 | + __func__); | ||
473 | + s->afrl = value; | ||
474 | + return; | ||
475 | + case GPIO_AFRH: | ||
476 | + qemu_log_mask(LOG_UNIMP, | ||
477 | + "%s: Alternate functions aren't supported\n", | ||
478 | + __func__); | ||
479 | + s->afrh = value; | ||
480 | + return; | ||
481 | + case GPIO_BRR: { | ||
482 | + uint32_t bits_to_reset = value & ~RESERVED_BITS_MASK; | ||
483 | + s->odr &= ~bits_to_reset; | ||
484 | + update_gpio_idr(s); | ||
485 | + return; | ||
486 | + } | ||
487 | + case GPIO_ASCR: | ||
488 | + qemu_log_mask(LOG_UNIMP, | ||
489 | + "%s: ADC function isn't supported\n", | ||
490 | + __func__); | ||
491 | + s->ascr = value & ~RESERVED_BITS_MASK; | ||
492 | + return; | ||
493 | + default: | ||
494 | + qemu_log_mask(LOG_GUEST_ERROR, | ||
495 | + "%s: Bad offset 0x%" HWADDR_PRIx "\n", __func__, addr); | ||
496 | + } | ||
497 | +} | ||
498 | + | ||
499 | +static uint64_t stm32l4x5_gpio_read(void *opaque, hwaddr addr, | ||
500 | + unsigned int size) | ||
501 | +{ | ||
502 | + Stm32l4x5GpioState *s = opaque; | ||
503 | + | ||
504 | + trace_stm32l4x5_gpio_read(s->name, addr); | ||
505 | + | ||
506 | + switch (addr) { | ||
507 | + case GPIO_MODER: | ||
508 | + return s->moder; | ||
509 | + case GPIO_OTYPER: | ||
510 | + return s->otyper; | ||
511 | + case GPIO_OSPEEDR: | ||
512 | + return s->ospeedr; | ||
513 | + case GPIO_PUPDR: | ||
514 | + return s->pupdr; | ||
515 | + case GPIO_IDR: | ||
516 | + return s->idr; | ||
517 | + case GPIO_ODR: | ||
518 | + return s->odr; | ||
519 | + case GPIO_BSRR: | ||
520 | + return 0; | ||
521 | + case GPIO_LCKR: | ||
522 | + return s->lckr; | ||
523 | + case GPIO_AFRL: | ||
524 | + return s->afrl; | ||
525 | + case GPIO_AFRH: | ||
526 | + return s->afrh; | ||
527 | + case GPIO_BRR: | ||
528 | + return 0; | ||
529 | + case GPIO_ASCR: | ||
530 | + return s->ascr; | ||
531 | + default: | ||
532 | + qemu_log_mask(LOG_GUEST_ERROR, | ||
533 | + "%s: Bad offset 0x%" HWADDR_PRIx "\n", __func__, addr); | ||
534 | + return 0; | ||
535 | + } | ||
536 | +} | ||
537 | + | ||
538 | +static const MemoryRegionOps stm32l4x5_gpio_ops = { | ||
539 | + .read = stm32l4x5_gpio_read, | ||
540 | + .write = stm32l4x5_gpio_write, | ||
541 | + .endianness = DEVICE_NATIVE_ENDIAN, | ||
542 | + .impl = { | ||
543 | + .min_access_size = 4, | ||
544 | + .max_access_size = 4, | ||
545 | + .unaligned = false, | ||
546 | + }, | ||
547 | + .valid = { | ||
548 | + .min_access_size = 4, | ||
549 | + .max_access_size = 4, | ||
550 | + .unaligned = false, | ||
551 | + }, | ||
552 | +}; | ||
553 | + | ||
554 | +static void stm32l4x5_gpio_init(Object *obj) | ||
555 | +{ | ||
556 | + Stm32l4x5GpioState *s = STM32L4X5_GPIO(obj); | ||
557 | + | ||
558 | + memory_region_init_io(&s->mmio, obj, &stm32l4x5_gpio_ops, s, | ||
559 | + TYPE_STM32L4X5_GPIO, 0x400); | ||
560 | + | ||
561 | + sysbus_init_mmio(SYS_BUS_DEVICE(obj), &s->mmio); | ||
562 | + | ||
563 | + qdev_init_gpio_out(DEVICE(obj), s->pin, GPIO_NUM_PINS); | ||
564 | + qdev_init_gpio_in(DEVICE(obj), stm32l4x5_gpio_set, GPIO_NUM_PINS); | ||
565 | + | ||
566 | + s->clk = qdev_init_clock_in(DEVICE(s), "clk", NULL, s, 0); | ||
567 | + | ||
568 | + object_property_add(obj, "disconnected-pins", "uint16", | ||
569 | + disconnected_pins_get, disconnected_pins_set, | ||
570 | + NULL, &s->disconnected_pins); | ||
571 | + object_property_add(obj, "clock-freq-hz", "uint32", | ||
572 | + clock_freq_get, NULL, NULL, NULL); | ||
573 | +} | ||
574 | + | ||
575 | +static void stm32l4x5_gpio_realize(DeviceState *dev, Error **errp) | ||
576 | +{ | ||
577 | + Stm32l4x5GpioState *s = STM32L4X5_GPIO(dev); | ||
578 | + if (!clock_has_source(s->clk)) { | ||
579 | + error_setg(errp, "GPIO: clk input must be connected"); | ||
580 | + return; | ||
581 | + } | ||
582 | +} | ||
583 | + | ||
584 | +static const VMStateDescription vmstate_stm32l4x5_gpio = { | ||
585 | + .name = TYPE_STM32L4X5_GPIO, | ||
586 | + .version_id = 1, | ||
587 | + .minimum_version_id = 1, | ||
588 | + .fields = (VMStateField[]){ | ||
589 | + VMSTATE_UINT32(moder, Stm32l4x5GpioState), | ||
590 | + VMSTATE_UINT32(otyper, Stm32l4x5GpioState), | ||
591 | + VMSTATE_UINT32(ospeedr, Stm32l4x5GpioState), | ||
592 | + VMSTATE_UINT32(pupdr, Stm32l4x5GpioState), | ||
593 | + VMSTATE_UINT32(idr, Stm32l4x5GpioState), | ||
594 | + VMSTATE_UINT32(odr, Stm32l4x5GpioState), | ||
595 | + VMSTATE_UINT32(lckr, Stm32l4x5GpioState), | ||
596 | + VMSTATE_UINT32(afrl, Stm32l4x5GpioState), | ||
597 | + VMSTATE_UINT32(afrh, Stm32l4x5GpioState), | ||
598 | + VMSTATE_UINT32(ascr, Stm32l4x5GpioState), | ||
599 | + VMSTATE_UINT16(disconnected_pins, Stm32l4x5GpioState), | ||
600 | + VMSTATE_UINT16(pins_connected_high, Stm32l4x5GpioState), | ||
601 | + VMSTATE_END_OF_LIST() | ||
602 | + } | ||
603 | +}; | ||
604 | + | ||
605 | +static Property stm32l4x5_gpio_properties[] = { | ||
606 | + DEFINE_PROP_STRING("name", Stm32l4x5GpioState, name), | ||
607 | + DEFINE_PROP_UINT32("mode-reset", Stm32l4x5GpioState, moder_reset, 0), | ||
608 | + DEFINE_PROP_UINT32("ospeed-reset", Stm32l4x5GpioState, ospeedr_reset, 0), | ||
609 | + DEFINE_PROP_UINT32("pupd-reset", Stm32l4x5GpioState, pupdr_reset, 0), | ||
610 | + DEFINE_PROP_END_OF_LIST(), | ||
611 | +}; | ||
612 | + | ||
613 | +static void stm32l4x5_gpio_class_init(ObjectClass *klass, void *data) | ||
614 | +{ | ||
615 | + DeviceClass *dc = DEVICE_CLASS(klass); | ||
616 | + ResettableClass *rc = RESETTABLE_CLASS(klass); | ||
617 | + | ||
618 | + device_class_set_props(dc, stm32l4x5_gpio_properties); | ||
619 | + dc->vmsd = &vmstate_stm32l4x5_gpio; | ||
620 | + dc->realize = stm32l4x5_gpio_realize; | ||
621 | + rc->phases.hold = stm32l4x5_gpio_reset_hold; | ||
622 | +} | ||
623 | + | ||
624 | +static const TypeInfo stm32l4x5_gpio_types[] = { | ||
625 | + { | ||
626 | + .name = TYPE_STM32L4X5_GPIO, | ||
627 | + .parent = TYPE_SYS_BUS_DEVICE, | ||
628 | + .instance_size = sizeof(Stm32l4x5GpioState), | ||
629 | + .instance_init = stm32l4x5_gpio_init, | ||
630 | + .class_init = stm32l4x5_gpio_class_init, | ||
631 | + }, | ||
632 | +}; | ||
633 | + | ||
634 | +DEFINE_TYPES(stm32l4x5_gpio_types) | ||
635 | diff --git a/hw/gpio/Kconfig b/hw/gpio/Kconfig | ||
199 | index XXXXXXX..XXXXXXX 100644 | 636 | index XXXXXXX..XXXXXXX 100644 |
200 | --- a/target/arm/translate-sve.c | 637 | --- a/hw/gpio/Kconfig |
201 | +++ b/target/arm/translate-sve.c | 638 | +++ b/hw/gpio/Kconfig |
202 | @@ -XXX,XX +XXX,XX @@ static bool trans_##NAME(DisasContext *s, arg_rprr_esz *a) \ | 639 | @@ -XXX,XX +XXX,XX @@ config GPIO_PWR |
203 | DO_SVE2_PPZZ_MATCH(MATCH, match) | 640 | |
204 | DO_SVE2_PPZZ_MATCH(NMATCH, nmatch) | 641 | config SIFIVE_GPIO |
205 | 642 | bool | |
206 | +static bool trans_HISTCNT(DisasContext *s, arg_rprr_esz *a) | 643 | + |
207 | +{ | 644 | +config STM32L4X5_GPIO |
208 | + static gen_helper_gvec_4 * const fns[2] = { | 645 | + bool |
209 | + gen_helper_sve2_histcnt_s, gen_helper_sve2_histcnt_d | 646 | diff --git a/hw/gpio/meson.build b/hw/gpio/meson.build |
210 | + }; | 647 | index XXXXXXX..XXXXXXX 100644 |
211 | + if (a->esz < 2) { | 648 | --- a/hw/gpio/meson.build |
212 | + return false; | 649 | +++ b/hw/gpio/meson.build |
213 | + } | 650 | @@ -XXX,XX +XXX,XX @@ system_ss.add(when: 'CONFIG_RASPI', if_true: files( |
214 | + return do_sve2_zpzz_ool(s, a, fns[a->esz - 2]); | 651 | 'bcm2835_gpio.c', |
215 | +} | 652 | 'bcm2838_gpio.c' |
216 | + | 653 | )) |
217 | +static bool trans_HISTSEG(DisasContext *s, arg_rrr_esz *a) | 654 | +system_ss.add(when: 'CONFIG_STM32L4X5_SOC', if_true: files('stm32l4x5_gpio.c')) |
218 | +{ | 655 | system_ss.add(when: 'CONFIG_ASPEED_SOC', if_true: files('aspeed_gpio.c')) |
219 | + if (a->esz != 0) { | 656 | system_ss.add(when: 'CONFIG_SIFIVE_GPIO', if_true: files('sifive_gpio.c')) |
220 | + return false; | 657 | diff --git a/hw/gpio/trace-events b/hw/gpio/trace-events |
221 | + } | 658 | index XXXXXXX..XXXXXXX 100644 |
222 | + return do_sve2_zzz_ool(s, a, gen_helper_sve2_histseg); | 659 | --- a/hw/gpio/trace-events |
223 | +} | 660 | +++ b/hw/gpio/trace-events |
224 | + | 661 | @@ -XXX,XX +XXX,XX @@ sifive_gpio_update_output_irq(int64_t line, int64_t value) "line %" PRIi64 " val |
225 | static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | 662 | # aspeed_gpio.c |
226 | gen_helper_gvec_4_ptr *fn) | 663 | aspeed_gpio_read(uint64_t offset, uint64_t value) "offset: 0x%" PRIx64 " value 0x%" PRIx64 |
227 | { | 664 | aspeed_gpio_write(uint64_t offset, uint64_t value) "offset: 0x%" PRIx64 " value 0x%" PRIx64 |
665 | + | ||
666 | +# stm32l4x5_gpio.c | ||
667 | +stm32l4x5_gpio_read(char *gpio, uint64_t addr) "GPIO%s addr: 0x%" PRIx64 " " | ||
668 | +stm32l4x5_gpio_write(char *gpio, uint64_t addr, uint64_t data) "GPIO%s addr: 0x%" PRIx64 " val: 0x%" PRIx64 "" | ||
669 | +stm32l4x5_gpio_update_idr(char *gpio, uint32_t old_idr, uint32_t new_idr) "GPIO%s from: 0x%x to: 0x%x" | ||
670 | +stm32l4x5_gpio_pins(char *gpio, uint16_t disconnected, uint16_t high) "GPIO%s disconnected pins: 0x%x levels: 0x%x" | ||
228 | -- | 671 | -- |
229 | 2.20.1 | 672 | 2.34.1 |
230 | 673 | ||
231 | 674 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | From: Inès Varhol <ines.varhol@telecom-paris.fr> |
---|---|---|---|
2 | 2 | ||
3 | Rename the existing sve_while (less-than) helper to sve_whilel | 3 | Signed-off-by: Arnaud Minier <arnaud.minier@telecom-paris.fr> |
4 | to make room for a new sve_whileg helper for greater-than. | 4 | Signed-off-by: Inès Varhol <ines.varhol@telecom-paris.fr> |
5 | 5 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | |
6 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 6 | Acked-by: Alistair Francis <alistair.francis@wdc.com> |
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 7 | Message-id: 20240305210444.310665-3-ines.varhol@telecom-paris.fr |
8 | Message-id: 20210525010358.152808-31-richard.henderson@linaro.org | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
10 | --- | 9 | --- |
11 | target/arm/helper-sve.h | 3 +- | 10 | include/hw/arm/stm32l4x5_soc.h | 2 + |
12 | target/arm/sve.decode | 2 +- | 11 | include/hw/gpio/stm32l4x5_gpio.h | 1 + |
13 | target/arm/sve_helper.c | 38 +++++++++++++++++++++++++- | 12 | include/hw/misc/stm32l4x5_syscfg.h | 3 +- |
14 | target/arm/translate-sve.c | 56 ++++++++++++++++++++++++++++---------- | 13 | hw/arm/stm32l4x5_soc.c | 71 +++++++++++++++++++++++------- |
15 | 4 files changed, 82 insertions(+), 17 deletions(-) | 14 | hw/misc/stm32l4x5_syscfg.c | 1 + |
16 | 15 | hw/arm/Kconfig | 3 +- | |
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | 16 | 6 files changed, 63 insertions(+), 18 deletions(-) |
18 | index XXXXXXX..XXXXXXX 100644 | 17 | |
19 | --- a/target/arm/helper-sve.h | 18 | diff --git a/include/hw/arm/stm32l4x5_soc.h b/include/hw/arm/stm32l4x5_soc.h |
20 | +++ b/target/arm/helper-sve.h | 19 | index XXXXXXX..XXXXXXX 100644 |
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_brkns, TCG_CALL_NO_RWG, i32, ptr, ptr, ptr, i32) | 20 | --- a/include/hw/arm/stm32l4x5_soc.h |
22 | 21 | +++ b/include/hw/arm/stm32l4x5_soc.h | |
23 | DEF_HELPER_FLAGS_3(sve_cntp, TCG_CALL_NO_RWG, i64, ptr, ptr, i32) | 22 | @@ -XXX,XX +XXX,XX @@ |
24 | 23 | #include "hw/misc/stm32l4x5_syscfg.h" | |
25 | -DEF_HELPER_FLAGS_3(sve_while, TCG_CALL_NO_RWG, i32, ptr, i32, i32) | 24 | #include "hw/misc/stm32l4x5_exti.h" |
26 | +DEF_HELPER_FLAGS_3(sve_whilel, TCG_CALL_NO_RWG, i32, ptr, i32, i32) | 25 | #include "hw/misc/stm32l4x5_rcc.h" |
27 | +DEF_HELPER_FLAGS_3(sve_whileg, TCG_CALL_NO_RWG, i32, ptr, i32, i32) | 26 | +#include "hw/gpio/stm32l4x5_gpio.h" |
28 | 27 | #include "qom/object.h" | |
29 | DEF_HELPER_FLAGS_4(sve_subri_b, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32) | 28 | |
30 | DEF_HELPER_FLAGS_4(sve_subri_h, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32) | 29 | #define TYPE_STM32L4X5_SOC "stm32l4x5-soc" |
31 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | 30 | @@ -XXX,XX +XXX,XX @@ struct Stm32l4x5SocState { |
32 | index XXXXXXX..XXXXXXX 100644 | 31 | OrIRQState exti_or_gates[NUM_EXTI_OR_GATES]; |
33 | --- a/target/arm/sve.decode | 32 | Stm32l4x5SyscfgState syscfg; |
34 | +++ b/target/arm/sve.decode | 33 | Stm32l4x5RccState rcc; |
35 | @@ -XXX,XX +XXX,XX @@ SINCDECP_z 00100101 .. 1010 d:1 u:1 10000 00 .... ..... @incdec2_pred | 34 | + Stm32l4x5GpioState gpio[NUM_GPIOS]; |
36 | CTERM 00100101 1 sf:1 1 rm:5 001000 rn:5 ne:1 0000 | 35 | |
37 | 36 | MemoryRegion sram1; | |
38 | # SVE integer compare scalar count and limit | 37 | MemoryRegion sram2; |
39 | -WHILE 00100101 esz:2 1 rm:5 000 sf:1 u:1 1 rn:5 eq:1 rd:4 | 38 | diff --git a/include/hw/gpio/stm32l4x5_gpio.h b/include/hw/gpio/stm32l4x5_gpio.h |
40 | +WHILE 00100101 esz:2 1 rm:5 000 sf:1 u:1 lt:1 rn:5 eq:1 rd:4 | 39 | index XXXXXXX..XXXXXXX 100644 |
41 | 40 | --- a/include/hw/gpio/stm32l4x5_gpio.h | |
42 | ### SVE Integer Wide Immediate - Unpredicated Group | 41 | +++ b/include/hw/gpio/stm32l4x5_gpio.h |
43 | 42 | @@ -XXX,XX +XXX,XX @@ | |
44 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | 43 | #define TYPE_STM32L4X5_GPIO "stm32l4x5-gpio" |
45 | index XXXXXXX..XXXXXXX 100644 | 44 | OBJECT_DECLARE_SIMPLE_TYPE(Stm32l4x5GpioState, STM32L4X5_GPIO) |
46 | --- a/target/arm/sve_helper.c | 45 | |
47 | +++ b/target/arm/sve_helper.c | 46 | +#define NUM_GPIOS 8 |
48 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(sve_cntp)(void *vn, void *vg, uint32_t pred_desc) | 47 | #define GPIO_NUM_PINS 16 |
49 | return sum; | 48 | |
49 | struct Stm32l4x5GpioState { | ||
50 | diff --git a/include/hw/misc/stm32l4x5_syscfg.h b/include/hw/misc/stm32l4x5_syscfg.h | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/include/hw/misc/stm32l4x5_syscfg.h | ||
53 | +++ b/include/hw/misc/stm32l4x5_syscfg.h | ||
54 | @@ -XXX,XX +XXX,XX @@ | ||
55 | |||
56 | #include "hw/sysbus.h" | ||
57 | #include "qom/object.h" | ||
58 | +#include "hw/gpio/stm32l4x5_gpio.h" | ||
59 | |||
60 | #define TYPE_STM32L4X5_SYSCFG "stm32l4x5-syscfg" | ||
61 | OBJECT_DECLARE_SIMPLE_TYPE(Stm32l4x5SyscfgState, STM32L4X5_SYSCFG) | ||
62 | |||
63 | -#define NUM_GPIOS 8 | ||
64 | -#define GPIO_NUM_PINS 16 | ||
65 | #define SYSCFG_NUM_EXTICR 4 | ||
66 | |||
67 | struct Stm32l4x5SyscfgState { | ||
68 | diff --git a/hw/arm/stm32l4x5_soc.c b/hw/arm/stm32l4x5_soc.c | ||
69 | index XXXXXXX..XXXXXXX 100644 | ||
70 | --- a/hw/arm/stm32l4x5_soc.c | ||
71 | +++ b/hw/arm/stm32l4x5_soc.c | ||
72 | @@ -XXX,XX +XXX,XX @@ | ||
73 | #include "sysemu/sysemu.h" | ||
74 | #include "hw/or-irq.h" | ||
75 | #include "hw/arm/stm32l4x5_soc.h" | ||
76 | +#include "hw/gpio/stm32l4x5_gpio.h" | ||
77 | #include "hw/qdev-clock.h" | ||
78 | #include "hw/misc/unimp.h" | ||
79 | |||
80 | @@ -XXX,XX +XXX,XX @@ static const int exti_or_gate1_lines_in[EXTI_OR_GATE1_NUM_LINES_IN] = { | ||
81 | 16, 35, 36, 37, 38, | ||
82 | }; | ||
83 | |||
84 | +static const struct { | ||
85 | + uint32_t addr; | ||
86 | + uint32_t moder_reset; | ||
87 | + uint32_t ospeedr_reset; | ||
88 | + uint32_t pupdr_reset; | ||
89 | +} stm32l4x5_gpio_cfg[NUM_GPIOS] = { | ||
90 | + { 0x48000000, 0xABFFFFFF, 0x0C000000, 0x64000000 }, | ||
91 | + { 0x48000400, 0xFFFFFEBF, 0x00000000, 0x00000100 }, | ||
92 | + { 0x48000800, 0xFFFFFFFF, 0x00000000, 0x00000000 }, | ||
93 | + { 0x48000C00, 0xFFFFFFFF, 0x00000000, 0x00000000 }, | ||
94 | + { 0x48001000, 0xFFFFFFFF, 0x00000000, 0x00000000 }, | ||
95 | + { 0x48001400, 0xFFFFFFFF, 0x00000000, 0x00000000 }, | ||
96 | + { 0x48001800, 0xFFFFFFFF, 0x00000000, 0x00000000 }, | ||
97 | + { 0x48001C00, 0x0000000F, 0x00000000, 0x00000000 }, | ||
98 | +}; | ||
99 | + | ||
100 | static void stm32l4x5_soc_initfn(Object *obj) | ||
101 | { | ||
102 | Stm32l4x5SocState *s = STM32L4X5_SOC(obj); | ||
103 | @@ -XXX,XX +XXX,XX @@ static void stm32l4x5_soc_initfn(Object *obj) | ||
104 | } | ||
105 | object_initialize_child(obj, "syscfg", &s->syscfg, TYPE_STM32L4X5_SYSCFG); | ||
106 | object_initialize_child(obj, "rcc", &s->rcc, TYPE_STM32L4X5_RCC); | ||
107 | + | ||
108 | + for (unsigned i = 0; i < NUM_GPIOS; i++) { | ||
109 | + g_autofree char *name = g_strdup_printf("gpio%c", 'a' + i); | ||
110 | + object_initialize_child(obj, name, &s->gpio[i], TYPE_STM32L4X5_GPIO); | ||
111 | + } | ||
50 | } | 112 | } |
51 | 113 | ||
52 | -uint32_t HELPER(sve_while)(void *vd, uint32_t count, uint32_t pred_desc) | 114 | static void stm32l4x5_soc_realize(DeviceState *dev_soc, Error **errp) |
53 | +uint32_t HELPER(sve_whilel)(void *vd, uint32_t count, uint32_t pred_desc) | 115 | @@ -XXX,XX +XXX,XX @@ static void stm32l4x5_soc_realize(DeviceState *dev_soc, Error **errp) |
54 | { | 116 | Stm32l4x5SocState *s = STM32L4X5_SOC(dev_soc); |
55 | intptr_t oprsz = FIELD_EX32(pred_desc, PREDDESC, OPRSZ); | 117 | const Stm32l4x5SocClass *sc = STM32L4X5_SOC_GET_CLASS(dev_soc); |
56 | intptr_t esz = FIELD_EX32(pred_desc, PREDDESC, ESZ); | 118 | MemoryRegion *system_memory = get_system_memory(); |
57 | @@ -XXX,XX +XXX,XX @@ uint32_t HELPER(sve_while)(void *vd, uint32_t count, uint32_t pred_desc) | 119 | - DeviceState *armv7m; |
58 | return predtest_ones(d, oprsz, esz_mask); | 120 | + DeviceState *armv7m, *dev; |
59 | } | 121 | SysBusDevice *busdev; |
60 | 122 | + uint32_t pin_index; | |
61 | +uint32_t HELPER(sve_whileg)(void *vd, uint32_t count, uint32_t pred_desc) | 123 | |
62 | +{ | 124 | if (!memory_region_init_rom(&s->flash, OBJECT(dev_soc), "flash", |
63 | + intptr_t oprsz = FIELD_EX32(pred_desc, PREDDESC, OPRSZ); | 125 | sc->flash_size, errp)) { |
64 | + intptr_t esz = FIELD_EX32(pred_desc, PREDDESC, ESZ); | 126 | @@ -XXX,XX +XXX,XX @@ static void stm32l4x5_soc_realize(DeviceState *dev_soc, Error **errp) |
65 | + uint64_t esz_mask = pred_esz_masks[esz]; | 127 | return; |
66 | + ARMPredicateReg *d = vd; | 128 | } |
67 | + intptr_t i, invcount, oprbits; | 129 | |
68 | + uint64_t bits; | 130 | + /* GPIOs */ |
69 | + | 131 | + for (unsigned i = 0; i < NUM_GPIOS; i++) { |
70 | + if (count == 0) { | 132 | + g_autofree char *name = g_strdup_printf("%c", 'A' + i); |
71 | + return do_zero(d, oprsz); | 133 | + dev = DEVICE(&s->gpio[i]); |
134 | + qdev_prop_set_string(dev, "name", name); | ||
135 | + qdev_prop_set_uint32(dev, "mode-reset", | ||
136 | + stm32l4x5_gpio_cfg[i].moder_reset); | ||
137 | + qdev_prop_set_uint32(dev, "ospeed-reset", | ||
138 | + stm32l4x5_gpio_cfg[i].ospeedr_reset); | ||
139 | + qdev_prop_set_uint32(dev, "pupd-reset", | ||
140 | + stm32l4x5_gpio_cfg[i].pupdr_reset); | ||
141 | + busdev = SYS_BUS_DEVICE(&s->gpio[i]); | ||
142 | + g_free(name); | ||
143 | + name = g_strdup_printf("gpio%c-out", 'a' + i); | ||
144 | + qdev_connect_clock_in(DEVICE(&s->gpio[i]), "clk", | ||
145 | + qdev_get_clock_out(DEVICE(&(s->rcc)), name)); | ||
146 | + if (!sysbus_realize(busdev, errp)) { | ||
147 | + return; | ||
148 | + } | ||
149 | + sysbus_mmio_map(busdev, 0, stm32l4x5_gpio_cfg[i].addr); | ||
72 | + } | 150 | + } |
73 | + | 151 | + |
74 | + oprbits = oprsz * 8; | 152 | /* System configuration controller */ |
75 | + tcg_debug_assert(count <= oprbits); | 153 | busdev = SYS_BUS_DEVICE(&s->syscfg); |
76 | + | 154 | if (!sysbus_realize(busdev, errp)) { |
77 | + bits = esz_mask; | 155 | return; |
78 | + if (oprbits & 63) { | 156 | } |
79 | + bits &= MAKE_64BIT_MASK(0, oprbits & 63); | 157 | sysbus_mmio_map(busdev, 0, SYSCFG_ADDR); |
80 | + } | 158 | - /* |
81 | + | 159 | - * TODO: when the GPIO device is implemented, connect it |
82 | + invcount = oprbits - count; | 160 | - * to SYCFG using `qdev_connect_gpio_out`, NUM_GPIOS and |
83 | + for (i = (oprsz - 1) / 8; i > invcount / 64; --i) { | 161 | - * GPIO_NUM_PINS. |
84 | + d->p[i] = bits; | 162 | - */ |
85 | + bits = esz_mask; | 163 | + |
86 | + } | 164 | + for (unsigned i = 0; i < NUM_GPIOS; i++) { |
87 | + | 165 | + for (unsigned j = 0; j < GPIO_NUM_PINS; j++) { |
88 | + d->p[i] = bits & MAKE_64BIT_MASK(invcount & 63, 64); | 166 | + pin_index = GPIO_NUM_PINS * i + j; |
89 | + | 167 | + qdev_connect_gpio_out(DEVICE(&s->gpio[i]), j, |
90 | + while (--i >= 0) { | 168 | + qdev_get_gpio_in(DEVICE(&s->syscfg), |
91 | + d->p[i] = 0; | 169 | + pin_index)); |
92 | + } | ||
93 | + | ||
94 | + return predtest_ones(d, oprsz, esz_mask); | ||
95 | +} | ||
96 | + | ||
97 | /* Recursive reduction on a function; | ||
98 | * C.f. the ARM ARM function ReducePredicated. | ||
99 | * | ||
100 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
101 | index XXXXXXX..XXXXXXX 100644 | ||
102 | --- a/target/arm/translate-sve.c | ||
103 | +++ b/target/arm/translate-sve.c | ||
104 | @@ -XXX,XX +XXX,XX @@ static bool trans_WHILE(DisasContext *s, arg_WHILE *a) | ||
105 | unsigned vsz = vec_full_reg_size(s); | ||
106 | unsigned desc = 0; | ||
107 | TCGCond cond; | ||
108 | + uint64_t maxval; | ||
109 | + /* Note that GE/HS has a->eq == 0 and GT/HI has a->eq == 1. */ | ||
110 | + bool eq = a->eq == a->lt; | ||
111 | |||
112 | + /* The greater-than conditions are all SVE2. */ | ||
113 | + if (!a->lt && !dc_isar_feature(aa64_sve2, s)) { | ||
114 | + return false; | ||
115 | + } | ||
116 | if (!sve_access_check(s)) { | ||
117 | return true; | ||
118 | } | ||
119 | @@ -XXX,XX +XXX,XX @@ static bool trans_WHILE(DisasContext *s, arg_WHILE *a) | ||
120 | */ | ||
121 | t0 = tcg_temp_new_i64(); | ||
122 | t1 = tcg_temp_new_i64(); | ||
123 | - tcg_gen_sub_i64(t0, op1, op0); | ||
124 | + | ||
125 | + if (a->lt) { | ||
126 | + tcg_gen_sub_i64(t0, op1, op0); | ||
127 | + if (a->u) { | ||
128 | + maxval = a->sf ? UINT64_MAX : UINT32_MAX; | ||
129 | + cond = eq ? TCG_COND_LEU : TCG_COND_LTU; | ||
130 | + } else { | ||
131 | + maxval = a->sf ? INT64_MAX : INT32_MAX; | ||
132 | + cond = eq ? TCG_COND_LE : TCG_COND_LT; | ||
133 | + } | ||
134 | + } else { | ||
135 | + tcg_gen_sub_i64(t0, op0, op1); | ||
136 | + if (a->u) { | ||
137 | + maxval = 0; | ||
138 | + cond = eq ? TCG_COND_GEU : TCG_COND_GTU; | ||
139 | + } else { | ||
140 | + maxval = a->sf ? INT64_MIN : INT32_MIN; | ||
141 | + cond = eq ? TCG_COND_GE : TCG_COND_GT; | ||
142 | + } | 170 | + } |
143 | + } | 171 | + } |
144 | 172 | ||
145 | tmax = tcg_const_i64(vsz >> a->esz); | 173 | /* EXTI device */ |
146 | - if (a->eq) { | 174 | busdev = SYS_BUS_DEVICE(&s->exti); |
147 | + if (eq) { | 175 | @@ -XXX,XX +XXX,XX @@ static void stm32l4x5_soc_realize(DeviceState *dev_soc, Error **errp) |
148 | /* Equality means one more iteration. */ | 176 | } |
149 | tcg_gen_addi_i64(t0, t0, 1); | 177 | } |
150 | 178 | ||
151 | - /* If op1 is max (un)signed integer (and the only time the addition | 179 | - for (unsigned i = 0; i < 16; i++) { |
152 | - * above could overflow), then we produce an all-true predicate by | 180 | + for (unsigned i = 0; i < GPIO_NUM_PINS; i++) { |
153 | - * setting the count to the vector length. This is because the | 181 | qdev_connect_gpio_out(DEVICE(&s->syscfg), i, |
154 | - * pseudocode is described as an increment + compare loop, and the | 182 | qdev_get_gpio_in(DEVICE(&s->exti), i)); |
155 | - * max integer would always compare true. | 183 | } |
156 | + /* | 184 | @@ -XXX,XX +XXX,XX @@ static void stm32l4x5_soc_realize(DeviceState *dev_soc, Error **errp) |
157 | + * For the less-than while, if op1 is maxval (and the only time | 185 | /* RESERVED: 0x40024400, 0x7FDBC00 */ |
158 | + * the addition above could overflow), then we produce an all-true | 186 | |
159 | + * predicate by setting the count to the vector length. This is | 187 | /* AHB2 BUS */ |
160 | + * because the pseudocode is described as an increment + compare | 188 | - create_unimplemented_device("GPIOA", 0x48000000, 0x400); |
161 | + * loop, and the maximum integer would always compare true. | 189 | - create_unimplemented_device("GPIOB", 0x48000400, 0x400); |
162 | + * Similarly, the greater-than while has the same issue with the | 190 | - create_unimplemented_device("GPIOC", 0x48000800, 0x400); |
163 | + * minimum integer due to the decrement + compare loop. | 191 | - create_unimplemented_device("GPIOD", 0x48000C00, 0x400); |
164 | */ | 192 | - create_unimplemented_device("GPIOE", 0x48001000, 0x400); |
165 | - tcg_gen_movi_i64(t1, (a->sf | 193 | - create_unimplemented_device("GPIOF", 0x48001400, 0x400); |
166 | - ? (a->u ? UINT64_MAX : INT64_MAX) | 194 | - create_unimplemented_device("GPIOG", 0x48001800, 0x400); |
167 | - : (a->u ? UINT32_MAX : INT32_MAX))); | 195 | - create_unimplemented_device("GPIOH", 0x48001C00, 0x400); |
168 | + tcg_gen_movi_i64(t1, maxval); | 196 | /* RESERVED: 0x48002000, 0x7FDBC00 */ |
169 | tcg_gen_movcond_i64(TCG_COND_EQ, t0, op1, t1, tmax, t0); | 197 | create_unimplemented_device("OTG_FS", 0x50000000, 0x40000); |
170 | } | 198 | create_unimplemented_device("ADC", 0x50040000, 0x400); |
171 | 199 | diff --git a/hw/misc/stm32l4x5_syscfg.c b/hw/misc/stm32l4x5_syscfg.c | |
172 | @@ -XXX,XX +XXX,XX @@ static bool trans_WHILE(DisasContext *s, arg_WHILE *a) | 200 | index XXXXXXX..XXXXXXX 100644 |
173 | tcg_temp_free_i64(tmax); | 201 | --- a/hw/misc/stm32l4x5_syscfg.c |
174 | 202 | +++ b/hw/misc/stm32l4x5_syscfg.c | |
175 | /* Set the count to zero if the condition is false. */ | 203 | @@ -XXX,XX +XXX,XX @@ |
176 | - cond = (a->u | 204 | #include "hw/irq.h" |
177 | - ? (a->eq ? TCG_COND_LEU : TCG_COND_LTU) | 205 | #include "migration/vmstate.h" |
178 | - : (a->eq ? TCG_COND_LE : TCG_COND_LT)); | 206 | #include "hw/misc/stm32l4x5_syscfg.h" |
179 | tcg_gen_movi_i64(t1, 0); | 207 | +#include "hw/gpio/stm32l4x5_gpio.h" |
180 | tcg_gen_movcond_i64(cond, t0, op0, op1, t0, t1); | 208 | |
181 | tcg_temp_free_i64(t1); | 209 | #define SYSCFG_MEMRMP 0x00 |
182 | @@ -XXX,XX +XXX,XX @@ static bool trans_WHILE(DisasContext *s, arg_WHILE *a) | 210 | #define SYSCFG_CFGR1 0x04 |
183 | ptr = tcg_temp_new_ptr(); | 211 | diff --git a/hw/arm/Kconfig b/hw/arm/Kconfig |
184 | tcg_gen_addi_ptr(ptr, cpu_env, pred_full_reg_offset(s, a->rd)); | 212 | index XXXXXXX..XXXXXXX 100644 |
185 | 213 | --- a/hw/arm/Kconfig | |
186 | - gen_helper_sve_while(t2, ptr, t2, t3); | 214 | +++ b/hw/arm/Kconfig |
187 | + if (a->lt) { | 215 | @@ -XXX,XX +XXX,XX @@ config STM32L4X5_SOC |
188 | + gen_helper_sve_whilel(t2, ptr, t2, t3); | 216 | bool |
189 | + } else { | 217 | select ARM_V7M |
190 | + gen_helper_sve_whileg(t2, ptr, t2, t3); | 218 | select OR_IRQ |
191 | + } | 219 | - select STM32L4X5_SYSCFG |
192 | do_pred_flags(t2); | 220 | select STM32L4X5_EXTI |
193 | 221 | + select STM32L4X5_SYSCFG | |
194 | tcg_temp_free_ptr(ptr); | 222 | select STM32L4X5_RCC |
223 | + select STM32L4X5_GPIO | ||
224 | |||
225 | config XLNX_ZYNQMP_ARM | ||
226 | bool | ||
195 | -- | 227 | -- |
196 | 2.20.1 | 228 | 2.34.1 |
197 | 229 | ||
198 | 230 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-32-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 3 ++ | ||
9 | target/arm/translate-sve.c | 67 ++++++++++++++++++++++++++++++++++++++ | ||
10 | 2 files changed, 70 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ CTERM 00100101 1 sf:1 1 rm:5 001000 rn:5 ne:1 0000 | ||
17 | # SVE integer compare scalar count and limit | ||
18 | WHILE 00100101 esz:2 1 rm:5 000 sf:1 u:1 lt:1 rn:5 eq:1 rd:4 | ||
19 | |||
20 | +# SVE2 pointer conflict compare | ||
21 | +WHILE_ptr 00100101 esz:2 1 rm:5 001 100 rn:5 rw:1 rd:4 | ||
22 | + | ||
23 | ### SVE Integer Wide Immediate - Unpredicated Group | ||
24 | |||
25 | # SVE broadcast floating-point immediate (unpredicated) | ||
26 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
27 | index XXXXXXX..XXXXXXX 100644 | ||
28 | --- a/target/arm/translate-sve.c | ||
29 | +++ b/target/arm/translate-sve.c | ||
30 | @@ -XXX,XX +XXX,XX @@ static bool trans_WHILE(DisasContext *s, arg_WHILE *a) | ||
31 | return true; | ||
32 | } | ||
33 | |||
34 | +static bool trans_WHILE_ptr(DisasContext *s, arg_WHILE_ptr *a) | ||
35 | +{ | ||
36 | + TCGv_i64 op0, op1, diff, t1, tmax; | ||
37 | + TCGv_i32 t2, t3; | ||
38 | + TCGv_ptr ptr; | ||
39 | + unsigned vsz = vec_full_reg_size(s); | ||
40 | + unsigned desc = 0; | ||
41 | + | ||
42 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
43 | + return false; | ||
44 | + } | ||
45 | + if (!sve_access_check(s)) { | ||
46 | + return true; | ||
47 | + } | ||
48 | + | ||
49 | + op0 = read_cpu_reg(s, a->rn, 1); | ||
50 | + op1 = read_cpu_reg(s, a->rm, 1); | ||
51 | + | ||
52 | + tmax = tcg_const_i64(vsz); | ||
53 | + diff = tcg_temp_new_i64(); | ||
54 | + | ||
55 | + if (a->rw) { | ||
56 | + /* WHILERW */ | ||
57 | + /* diff = abs(op1 - op0), noting that op0/1 are unsigned. */ | ||
58 | + t1 = tcg_temp_new_i64(); | ||
59 | + tcg_gen_sub_i64(diff, op0, op1); | ||
60 | + tcg_gen_sub_i64(t1, op1, op0); | ||
61 | + tcg_gen_movcond_i64(TCG_COND_GEU, diff, op0, op1, diff, t1); | ||
62 | + tcg_temp_free_i64(t1); | ||
63 | + /* Round down to a multiple of ESIZE. */ | ||
64 | + tcg_gen_andi_i64(diff, diff, -1 << a->esz); | ||
65 | + /* If op1 == op0, diff == 0, and the condition is always true. */ | ||
66 | + tcg_gen_movcond_i64(TCG_COND_EQ, diff, op0, op1, tmax, diff); | ||
67 | + } else { | ||
68 | + /* WHILEWR */ | ||
69 | + tcg_gen_sub_i64(diff, op1, op0); | ||
70 | + /* Round down to a multiple of ESIZE. */ | ||
71 | + tcg_gen_andi_i64(diff, diff, -1 << a->esz); | ||
72 | + /* If op0 >= op1, diff <= 0, the condition is always true. */ | ||
73 | + tcg_gen_movcond_i64(TCG_COND_GEU, diff, op0, op1, tmax, diff); | ||
74 | + } | ||
75 | + | ||
76 | + /* Bound to the maximum. */ | ||
77 | + tcg_gen_umin_i64(diff, diff, tmax); | ||
78 | + tcg_temp_free_i64(tmax); | ||
79 | + | ||
80 | + /* Since we're bounded, pass as a 32-bit type. */ | ||
81 | + t2 = tcg_temp_new_i32(); | ||
82 | + tcg_gen_extrl_i64_i32(t2, diff); | ||
83 | + tcg_temp_free_i64(diff); | ||
84 | + | ||
85 | + desc = FIELD_DP32(desc, PREDDESC, OPRSZ, vsz / 8); | ||
86 | + desc = FIELD_DP32(desc, PREDDESC, ESZ, a->esz); | ||
87 | + t3 = tcg_const_i32(desc); | ||
88 | + | ||
89 | + ptr = tcg_temp_new_ptr(); | ||
90 | + tcg_gen_addi_ptr(ptr, cpu_env, pred_full_reg_offset(s, a->rd)); | ||
91 | + | ||
92 | + gen_helper_sve_whilel(t2, ptr, t2, t3); | ||
93 | + do_pred_flags(t2); | ||
94 | + | ||
95 | + tcg_temp_free_ptr(ptr); | ||
96 | + tcg_temp_free_i32(t2); | ||
97 | + tcg_temp_free_i32(t3); | ||
98 | + return true; | ||
99 | +} | ||
100 | + | ||
101 | /* | ||
102 | *** SVE Integer Wide Immediate - Unpredicated Group | ||
103 | */ | ||
104 | -- | ||
105 | 2.20.1 | ||
106 | |||
107 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-33-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 6 ++ | ||
9 | target/arm/sve.decode | 12 +++ | ||
10 | target/arm/sve_helper.c | 50 +++++++++ | ||
11 | target/arm/translate-sve.c | 213 +++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 281 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve2_fminp_zpzz_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_6(sve2_fminp_zpzz_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_eor3, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_5(sve2_bcax, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_bsl1n, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_5(sve2_bsl2n, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_nbsl, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
28 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
29 | index XXXXXXX..XXXXXXX 100644 | ||
30 | --- a/target/arm/sve.decode | ||
31 | +++ b/target/arm/sve.decode | ||
32 | @@ -XXX,XX +XXX,XX @@ | ||
33 | @rda_rn_rm ........ esz:2 . rm:5 ... ... rn:5 rd:5 \ | ||
34 | &rrrr_esz ra=%reg_movprfx | ||
35 | |||
36 | +# Four operand with unused vector element size | ||
37 | +@rdn_ra_rm_e0 ........ ... rm:5 ... ... ra:5 rd:5 \ | ||
38 | + &rrrr_esz esz=0 rn=%reg_movprfx | ||
39 | + | ||
40 | # Three operand with "memory" size, aka immediate left shift | ||
41 | @rd_rn_msz_rm ........ ... rm:5 .... imm:2 rn:5 rd:5 &rrri | ||
42 | |||
43 | @@ -XXX,XX +XXX,XX @@ ORR_zzz 00000100 01 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
44 | EOR_zzz 00000100 10 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
45 | BIC_zzz 00000100 11 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
46 | |||
47 | +# SVE2 bitwise ternary operations | ||
48 | +EOR3 00000100 00 1 ..... 001 110 ..... ..... @rdn_ra_rm_e0 | ||
49 | +BSL 00000100 00 1 ..... 001 111 ..... ..... @rdn_ra_rm_e0 | ||
50 | +BCAX 00000100 01 1 ..... 001 110 ..... ..... @rdn_ra_rm_e0 | ||
51 | +BSL1N 00000100 01 1 ..... 001 111 ..... ..... @rdn_ra_rm_e0 | ||
52 | +BSL2N 00000100 10 1 ..... 001 111 ..... ..... @rdn_ra_rm_e0 | ||
53 | +NBSL 00000100 11 1 ..... 001 111 ..... ..... @rdn_ra_rm_e0 | ||
54 | + | ||
55 | ### SVE Index Generation Group | ||
56 | |||
57 | # SVE index generation (immediate start, immediate increment) | ||
58 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/target/arm/sve_helper.c | ||
61 | +++ b/target/arm/sve_helper.c | ||
62 | @@ -XXX,XX +XXX,XX @@ DO_ST1_ZPZ_D(dd_be, zd, MO_64) | ||
63 | |||
64 | #undef DO_ST1_ZPZ_S | ||
65 | #undef DO_ST1_ZPZ_D | ||
66 | + | ||
67 | +void HELPER(sve2_eor3)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
68 | +{ | ||
69 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
70 | + uint64_t *d = vd, *n = vn, *m = vm, *k = vk; | ||
71 | + | ||
72 | + for (i = 0; i < opr_sz; ++i) { | ||
73 | + d[i] = n[i] ^ m[i] ^ k[i]; | ||
74 | + } | ||
75 | +} | ||
76 | + | ||
77 | +void HELPER(sve2_bcax)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
78 | +{ | ||
79 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
80 | + uint64_t *d = vd, *n = vn, *m = vm, *k = vk; | ||
81 | + | ||
82 | + for (i = 0; i < opr_sz; ++i) { | ||
83 | + d[i] = n[i] ^ (m[i] & ~k[i]); | ||
84 | + } | ||
85 | +} | ||
86 | + | ||
87 | +void HELPER(sve2_bsl1n)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
88 | +{ | ||
89 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
90 | + uint64_t *d = vd, *n = vn, *m = vm, *k = vk; | ||
91 | + | ||
92 | + for (i = 0; i < opr_sz; ++i) { | ||
93 | + d[i] = (~n[i] & k[i]) | (m[i] & ~k[i]); | ||
94 | + } | ||
95 | +} | ||
96 | + | ||
97 | +void HELPER(sve2_bsl2n)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
98 | +{ | ||
99 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
100 | + uint64_t *d = vd, *n = vn, *m = vm, *k = vk; | ||
101 | + | ||
102 | + for (i = 0; i < opr_sz; ++i) { | ||
103 | + d[i] = (n[i] & k[i]) | (~m[i] & ~k[i]); | ||
104 | + } | ||
105 | +} | ||
106 | + | ||
107 | +void HELPER(sve2_nbsl)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
108 | +{ | ||
109 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
110 | + uint64_t *d = vd, *n = vn, *m = vm, *k = vk; | ||
111 | + | ||
112 | + for (i = 0; i < opr_sz; ++i) { | ||
113 | + d[i] = ~((n[i] & k[i]) | (m[i] & ~k[i])); | ||
114 | + } | ||
115 | +} | ||
116 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
117 | index XXXXXXX..XXXXXXX 100644 | ||
118 | --- a/target/arm/translate-sve.c | ||
119 | +++ b/target/arm/translate-sve.c | ||
120 | @@ -XXX,XX +XXX,XX @@ static void gen_gvec_fn_zzz(DisasContext *s, GVecGen3Fn *gvec_fn, | ||
121 | vec_full_reg_offset(s, rm), vsz, vsz); | ||
122 | } | ||
123 | |||
124 | +/* Invoke a vector expander on four Zregs. */ | ||
125 | +static void gen_gvec_fn_zzzz(DisasContext *s, GVecGen4Fn *gvec_fn, | ||
126 | + int esz, int rd, int rn, int rm, int ra) | ||
127 | +{ | ||
128 | + unsigned vsz = vec_full_reg_size(s); | ||
129 | + gvec_fn(esz, vec_full_reg_offset(s, rd), | ||
130 | + vec_full_reg_offset(s, rn), | ||
131 | + vec_full_reg_offset(s, rm), | ||
132 | + vec_full_reg_offset(s, ra), vsz, vsz); | ||
133 | +} | ||
134 | + | ||
135 | /* Invoke a vector move on two Zregs. */ | ||
136 | static bool do_mov_z(DisasContext *s, int rd, int rn) | ||
137 | { | ||
138 | @@ -XXX,XX +XXX,XX @@ static bool trans_BIC_zzz(DisasContext *s, arg_rrr_esz *a) | ||
139 | return do_zzz_fn(s, a, tcg_gen_gvec_andc); | ||
140 | } | ||
141 | |||
142 | +static bool do_sve2_zzzz_fn(DisasContext *s, arg_rrrr_esz *a, GVecGen4Fn *fn) | ||
143 | +{ | ||
144 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
145 | + return false; | ||
146 | + } | ||
147 | + if (sve_access_check(s)) { | ||
148 | + gen_gvec_fn_zzzz(s, fn, a->esz, a->rd, a->rn, a->rm, a->ra); | ||
149 | + } | ||
150 | + return true; | ||
151 | +} | ||
152 | + | ||
153 | +static void gen_eor3_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 k) | ||
154 | +{ | ||
155 | + tcg_gen_xor_i64(d, n, m); | ||
156 | + tcg_gen_xor_i64(d, d, k); | ||
157 | +} | ||
158 | + | ||
159 | +static void gen_eor3_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
160 | + TCGv_vec m, TCGv_vec k) | ||
161 | +{ | ||
162 | + tcg_gen_xor_vec(vece, d, n, m); | ||
163 | + tcg_gen_xor_vec(vece, d, d, k); | ||
164 | +} | ||
165 | + | ||
166 | +static void gen_eor3(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
167 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
168 | +{ | ||
169 | + static const GVecGen4 op = { | ||
170 | + .fni8 = gen_eor3_i64, | ||
171 | + .fniv = gen_eor3_vec, | ||
172 | + .fno = gen_helper_sve2_eor3, | ||
173 | + .vece = MO_64, | ||
174 | + .prefer_i64 = TCG_TARGET_REG_BITS == 64, | ||
175 | + }; | ||
176 | + tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &op); | ||
177 | +} | ||
178 | + | ||
179 | +static bool trans_EOR3(DisasContext *s, arg_rrrr_esz *a) | ||
180 | +{ | ||
181 | + return do_sve2_zzzz_fn(s, a, gen_eor3); | ||
182 | +} | ||
183 | + | ||
184 | +static void gen_bcax_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 k) | ||
185 | +{ | ||
186 | + tcg_gen_andc_i64(d, m, k); | ||
187 | + tcg_gen_xor_i64(d, d, n); | ||
188 | +} | ||
189 | + | ||
190 | +static void gen_bcax_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
191 | + TCGv_vec m, TCGv_vec k) | ||
192 | +{ | ||
193 | + tcg_gen_andc_vec(vece, d, m, k); | ||
194 | + tcg_gen_xor_vec(vece, d, d, n); | ||
195 | +} | ||
196 | + | ||
197 | +static void gen_bcax(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
198 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
199 | +{ | ||
200 | + static const GVecGen4 op = { | ||
201 | + .fni8 = gen_bcax_i64, | ||
202 | + .fniv = gen_bcax_vec, | ||
203 | + .fno = gen_helper_sve2_bcax, | ||
204 | + .vece = MO_64, | ||
205 | + .prefer_i64 = TCG_TARGET_REG_BITS == 64, | ||
206 | + }; | ||
207 | + tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &op); | ||
208 | +} | ||
209 | + | ||
210 | +static bool trans_BCAX(DisasContext *s, arg_rrrr_esz *a) | ||
211 | +{ | ||
212 | + return do_sve2_zzzz_fn(s, a, gen_bcax); | ||
213 | +} | ||
214 | + | ||
215 | +static void gen_bsl(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
216 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
217 | +{ | ||
218 | + /* BSL differs from the generic bitsel in argument ordering. */ | ||
219 | + tcg_gen_gvec_bitsel(vece, d, a, n, m, oprsz, maxsz); | ||
220 | +} | ||
221 | + | ||
222 | +static bool trans_BSL(DisasContext *s, arg_rrrr_esz *a) | ||
223 | +{ | ||
224 | + return do_sve2_zzzz_fn(s, a, gen_bsl); | ||
225 | +} | ||
226 | + | ||
227 | +static void gen_bsl1n_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 k) | ||
228 | +{ | ||
229 | + tcg_gen_andc_i64(n, k, n); | ||
230 | + tcg_gen_andc_i64(m, m, k); | ||
231 | + tcg_gen_or_i64(d, n, m); | ||
232 | +} | ||
233 | + | ||
234 | +static void gen_bsl1n_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
235 | + TCGv_vec m, TCGv_vec k) | ||
236 | +{ | ||
237 | + if (TCG_TARGET_HAS_bitsel_vec) { | ||
238 | + tcg_gen_not_vec(vece, n, n); | ||
239 | + tcg_gen_bitsel_vec(vece, d, k, n, m); | ||
240 | + } else { | ||
241 | + tcg_gen_andc_vec(vece, n, k, n); | ||
242 | + tcg_gen_andc_vec(vece, m, m, k); | ||
243 | + tcg_gen_or_vec(vece, d, n, m); | ||
244 | + } | ||
245 | +} | ||
246 | + | ||
247 | +static void gen_bsl1n(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
248 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
249 | +{ | ||
250 | + static const GVecGen4 op = { | ||
251 | + .fni8 = gen_bsl1n_i64, | ||
252 | + .fniv = gen_bsl1n_vec, | ||
253 | + .fno = gen_helper_sve2_bsl1n, | ||
254 | + .vece = MO_64, | ||
255 | + .prefer_i64 = TCG_TARGET_REG_BITS == 64, | ||
256 | + }; | ||
257 | + tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &op); | ||
258 | +} | ||
259 | + | ||
260 | +static bool trans_BSL1N(DisasContext *s, arg_rrrr_esz *a) | ||
261 | +{ | ||
262 | + return do_sve2_zzzz_fn(s, a, gen_bsl1n); | ||
263 | +} | ||
264 | + | ||
265 | +static void gen_bsl2n_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 k) | ||
266 | +{ | ||
267 | + /* | ||
268 | + * Z[dn] = (n & k) | (~m & ~k) | ||
269 | + * = | ~(m | k) | ||
270 | + */ | ||
271 | + tcg_gen_and_i64(n, n, k); | ||
272 | + if (TCG_TARGET_HAS_orc_i64) { | ||
273 | + tcg_gen_or_i64(m, m, k); | ||
274 | + tcg_gen_orc_i64(d, n, m); | ||
275 | + } else { | ||
276 | + tcg_gen_nor_i64(m, m, k); | ||
277 | + tcg_gen_or_i64(d, n, m); | ||
278 | + } | ||
279 | +} | ||
280 | + | ||
281 | +static void gen_bsl2n_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
282 | + TCGv_vec m, TCGv_vec k) | ||
283 | +{ | ||
284 | + if (TCG_TARGET_HAS_bitsel_vec) { | ||
285 | + tcg_gen_not_vec(vece, m, m); | ||
286 | + tcg_gen_bitsel_vec(vece, d, k, n, m); | ||
287 | + } else { | ||
288 | + tcg_gen_and_vec(vece, n, n, k); | ||
289 | + tcg_gen_or_vec(vece, m, m, k); | ||
290 | + tcg_gen_orc_vec(vece, d, n, m); | ||
291 | + } | ||
292 | +} | ||
293 | + | ||
294 | +static void gen_bsl2n(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
295 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
296 | +{ | ||
297 | + static const GVecGen4 op = { | ||
298 | + .fni8 = gen_bsl2n_i64, | ||
299 | + .fniv = gen_bsl2n_vec, | ||
300 | + .fno = gen_helper_sve2_bsl2n, | ||
301 | + .vece = MO_64, | ||
302 | + .prefer_i64 = TCG_TARGET_REG_BITS == 64, | ||
303 | + }; | ||
304 | + tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &op); | ||
305 | +} | ||
306 | + | ||
307 | +static bool trans_BSL2N(DisasContext *s, arg_rrrr_esz *a) | ||
308 | +{ | ||
309 | + return do_sve2_zzzz_fn(s, a, gen_bsl2n); | ||
310 | +} | ||
311 | + | ||
312 | +static void gen_nbsl_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 k) | ||
313 | +{ | ||
314 | + tcg_gen_and_i64(n, n, k); | ||
315 | + tcg_gen_andc_i64(m, m, k); | ||
316 | + tcg_gen_nor_i64(d, n, m); | ||
317 | +} | ||
318 | + | ||
319 | +static void gen_nbsl_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
320 | + TCGv_vec m, TCGv_vec k) | ||
321 | +{ | ||
322 | + tcg_gen_bitsel_vec(vece, d, k, n, m); | ||
323 | + tcg_gen_not_vec(vece, d, d); | ||
324 | +} | ||
325 | + | ||
326 | +static void gen_nbsl(unsigned vece, uint32_t d, uint32_t n, uint32_t m, | ||
327 | + uint32_t a, uint32_t oprsz, uint32_t maxsz) | ||
328 | +{ | ||
329 | + static const GVecGen4 op = { | ||
330 | + .fni8 = gen_nbsl_i64, | ||
331 | + .fniv = gen_nbsl_vec, | ||
332 | + .fno = gen_helper_sve2_nbsl, | ||
333 | + .vece = MO_64, | ||
334 | + .prefer_i64 = TCG_TARGET_REG_BITS == 64, | ||
335 | + }; | ||
336 | + tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &op); | ||
337 | +} | ||
338 | + | ||
339 | +static bool trans_NBSL(DisasContext *s, arg_rrrr_esz *a) | ||
340 | +{ | ||
341 | + return do_sve2_zzzz_fn(s, a, gen_nbsl); | ||
342 | +} | ||
343 | + | ||
344 | /* | ||
345 | *** SVE Integer Arithmetic - Unpredicated Group | ||
346 | */ | ||
347 | -- | ||
348 | 2.20.1 | ||
349 | |||
350 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-34-richard.henderson@linaro.org | ||
7 | Message-Id: <20200415145915.2859-1-steplong@quicinc.com> | ||
8 | [rth: Expanded comment for do_match2] | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper-sve.h | 10 ++++++ | ||
13 | target/arm/sve.decode | 5 +++ | ||
14 | target/arm/sve_helper.c | 64 ++++++++++++++++++++++++++++++++++++++ | ||
15 | target/arm/translate-sve.c | 22 +++++++++++++ | ||
16 | 4 files changed, 101 insertions(+) | ||
17 | |||
18 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/target/arm/helper-sve.h | ||
21 | +++ b/target/arm/helper-sve.h | ||
22 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_uqrshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_3(sve2_uqrshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | DEF_HELPER_FLAGS_3(sve2_uqrshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
25 | |||
26 | +DEF_HELPER_FLAGS_5(sve2_match_ppzz_b, TCG_CALL_NO_RWG, | ||
27 | + i32, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_match_ppzz_h, TCG_CALL_NO_RWG, | ||
29 | + i32, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_5(sve2_nmatch_ppzz_b, TCG_CALL_NO_RWG, | ||
32 | + i32, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_nmatch_ppzz_h, TCG_CALL_NO_RWG, | ||
34 | + i32, ptr, ptr, ptr, ptr, i32) | ||
35 | + | ||
36 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
37 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
38 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ UQSHRNT 01000101 .. 1 ..... 00 1101 ..... ..... @rd_rn_tszimm_shr | ||
44 | UQRSHRNB 01000101 .. 1 ..... 00 1110 ..... ..... @rd_rn_tszimm_shr | ||
45 | UQRSHRNT 01000101 .. 1 ..... 00 1111 ..... ..... @rd_rn_tszimm_shr | ||
46 | |||
47 | +### SVE2 Character Match | ||
48 | + | ||
49 | +MATCH 01000101 .. 1 ..... 100 ... ..... 0 .... @pd_pg_rn_rm | ||
50 | +NMATCH 01000101 .. 1 ..... 100 ... ..... 1 .... @pd_pg_rn_rm | ||
51 | + | ||
52 | ## SVE2 floating-point pairwise operations | ||
53 | |||
54 | FADDP 01100100 .. 010 00 0 100 ... ..... ..... @rdn_pg_rm | ||
55 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
56 | index XXXXXXX..XXXXXXX 100644 | ||
57 | --- a/target/arm/sve_helper.c | ||
58 | +++ b/target/arm/sve_helper.c | ||
59 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_nbsl)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) | ||
60 | d[i] = ~((n[i] & k[i]) | (m[i] & ~k[i])); | ||
61 | } | ||
62 | } | ||
63 | + | ||
64 | +/* | ||
65 | + * Returns true if m0 or m1 contains the low uint8_t/uint16_t in n. | ||
66 | + * See hasless(v,1) from | ||
67 | + * https://graphics.stanford.edu/~seander/bithacks.html#ZeroInWord | ||
68 | + */ | ||
69 | +static inline bool do_match2(uint64_t n, uint64_t m0, uint64_t m1, int esz) | ||
70 | +{ | ||
71 | + int bits = 8 << esz; | ||
72 | + uint64_t ones = dup_const(esz, 1); | ||
73 | + uint64_t signs = ones << (bits - 1); | ||
74 | + uint64_t cmp0, cmp1; | ||
75 | + | ||
76 | + cmp1 = dup_const(esz, n); | ||
77 | + cmp0 = cmp1 ^ m0; | ||
78 | + cmp1 = cmp1 ^ m1; | ||
79 | + cmp0 = (cmp0 - ones) & ~cmp0; | ||
80 | + cmp1 = (cmp1 - ones) & ~cmp1; | ||
81 | + return (cmp0 | cmp1) & signs; | ||
82 | +} | ||
83 | + | ||
84 | +static inline uint32_t do_match(void *vd, void *vn, void *vm, void *vg, | ||
85 | + uint32_t desc, int esz, bool nmatch) | ||
86 | +{ | ||
87 | + uint16_t esz_mask = pred_esz_masks[esz]; | ||
88 | + intptr_t opr_sz = simd_oprsz(desc); | ||
89 | + uint32_t flags = PREDTEST_INIT; | ||
90 | + intptr_t i, j, k; | ||
91 | + | ||
92 | + for (i = 0; i < opr_sz; i += 16) { | ||
93 | + uint64_t m0 = *(uint64_t *)(vm + i); | ||
94 | + uint64_t m1 = *(uint64_t *)(vm + i + 8); | ||
95 | + uint16_t pg = *(uint16_t *)(vg + H1_2(i >> 3)) & esz_mask; | ||
96 | + uint16_t out = 0; | ||
97 | + | ||
98 | + for (j = 0; j < 16; j += 8) { | ||
99 | + uint64_t n = *(uint64_t *)(vn + i + j); | ||
100 | + | ||
101 | + for (k = 0; k < 8; k += 1 << esz) { | ||
102 | + if (pg & (1 << (j + k))) { | ||
103 | + bool o = do_match2(n >> (k * 8), m0, m1, esz); | ||
104 | + out |= (o ^ nmatch) << (j + k); | ||
105 | + } | ||
106 | + } | ||
107 | + } | ||
108 | + *(uint16_t *)(vd + H1_2(i >> 3)) = out; | ||
109 | + flags = iter_predtest_fwd(out, pg, flags); | ||
110 | + } | ||
111 | + return flags; | ||
112 | +} | ||
113 | + | ||
114 | +#define DO_PPZZ_MATCH(NAME, ESZ, INV) \ | ||
115 | +uint32_t HELPER(NAME)(void *vd, void *vn, void *vm, void *vg, uint32_t desc) \ | ||
116 | +{ \ | ||
117 | + return do_match(vd, vn, vm, vg, desc, ESZ, INV); \ | ||
118 | +} | ||
119 | + | ||
120 | +DO_PPZZ_MATCH(sve2_match_ppzz_b, MO_8, false) | ||
121 | +DO_PPZZ_MATCH(sve2_match_ppzz_h, MO_16, false) | ||
122 | + | ||
123 | +DO_PPZZ_MATCH(sve2_nmatch_ppzz_b, MO_8, true) | ||
124 | +DO_PPZZ_MATCH(sve2_nmatch_ppzz_h, MO_16, true) | ||
125 | + | ||
126 | +#undef DO_PPZZ_MATCH | ||
127 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
128 | index XXXXXXX..XXXXXXX 100644 | ||
129 | --- a/target/arm/translate-sve.c | ||
130 | +++ b/target/arm/translate-sve.c | ||
131 | @@ -XXX,XX +XXX,XX @@ static bool trans_UQRSHRNT(DisasContext *s, arg_rri_esz *a) | ||
132 | return do_sve2_shr_narrow(s, a, ops); | ||
133 | } | ||
134 | |||
135 | +static bool do_sve2_ppzz_flags(DisasContext *s, arg_rprr_esz *a, | ||
136 | + gen_helper_gvec_flags_4 *fn) | ||
137 | +{ | ||
138 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
139 | + return false; | ||
140 | + } | ||
141 | + return do_ppzz_flags(s, a, fn); | ||
142 | +} | ||
143 | + | ||
144 | +#define DO_SVE2_PPZZ_MATCH(NAME, name) \ | ||
145 | +static bool trans_##NAME(DisasContext *s, arg_rprr_esz *a) \ | ||
146 | +{ \ | ||
147 | + static gen_helper_gvec_flags_4 * const fns[4] = { \ | ||
148 | + gen_helper_sve2_##name##_ppzz_b, gen_helper_sve2_##name##_ppzz_h, \ | ||
149 | + NULL, NULL \ | ||
150 | + }; \ | ||
151 | + return do_sve2_ppzz_flags(s, a, fns[a->esz]); \ | ||
152 | +} | ||
153 | + | ||
154 | +DO_SVE2_PPZZ_MATCH(MATCH, match) | ||
155 | +DO_SVE2_PPZZ_MATCH(NMATCH, nmatch) | ||
156 | + | ||
157 | static bool do_sve2_zpzz_fp(DisasContext *s, arg_rprr_esz *a, | ||
158 | gen_helper_gvec_4_ptr *fn) | ||
159 | { | ||
160 | -- | ||
161 | 2.20.1 | ||
162 | |||
163 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-35-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 14 ++++++++++ | ||
9 | target/arm/sve.decode | 14 ++++++++++ | ||
10 | target/arm/sve_helper.c | 30 +++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 54 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 112 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_bcax, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_5(sve2_bsl1n, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_bsl2n, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_5(sve2_nbsl, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_sqdmlal_zzzw_h, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_sqdmlal_zzzw_s, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_sqdmlal_zzzw_d, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_5(sve2_sqdmlsl_zzzw_h, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(sve2_sqdmlsl_zzzw_s, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(sve2_sqdmlsl_zzzw_d, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ FMAXNMP 01100100 .. 010 10 0 100 ... ..... ..... @rdn_pg_rm | ||
41 | FMINNMP 01100100 .. 010 10 1 100 ... ..... ..... @rdn_pg_rm | ||
42 | FMAXP 01100100 .. 010 11 0 100 ... ..... ..... @rdn_pg_rm | ||
43 | FMINP 01100100 .. 010 11 1 100 ... ..... ..... @rdn_pg_rm | ||
44 | + | ||
45 | +#### SVE Integer Multiply-Add (unpredicated) | ||
46 | + | ||
47 | +## SVE2 saturating multiply-add long | ||
48 | + | ||
49 | +SQDMLALB_zzzw 01000100 .. 0 ..... 0110 00 ..... ..... @rda_rn_rm | ||
50 | +SQDMLALT_zzzw 01000100 .. 0 ..... 0110 01 ..... ..... @rda_rn_rm | ||
51 | +SQDMLSLB_zzzw 01000100 .. 0 ..... 0110 10 ..... ..... @rda_rn_rm | ||
52 | +SQDMLSLT_zzzw 01000100 .. 0 ..... 0110 11 ..... ..... @rda_rn_rm | ||
53 | + | ||
54 | +## SVE2 saturating multiply-add interleaved long | ||
55 | + | ||
56 | +SQDMLALBT 01000100 .. 0 ..... 00001 0 ..... ..... @rda_rn_rm | ||
57 | +SQDMLSLBT 01000100 .. 0 ..... 00001 1 ..... ..... @rda_rn_rm | ||
58 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/target/arm/sve_helper.c | ||
61 | +++ b/target/arm/sve_helper.c | ||
62 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_adcl_d)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
63 | } | ||
64 | } | ||
65 | |||
66 | +#define DO_SQDMLAL(NAME, TYPEW, TYPEN, HW, HN, DMUL_OP, SUM_OP) \ | ||
67 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
68 | +{ \ | ||
69 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
70 | + int sel1 = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPEN); \ | ||
71 | + int sel2 = extract32(desc, SIMD_DATA_SHIFT + 1, 1) * sizeof(TYPEN); \ | ||
72 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
73 | + TYPEW nn = *(TYPEN *)(vn + HN(i + sel1)); \ | ||
74 | + TYPEW mm = *(TYPEN *)(vm + HN(i + sel2)); \ | ||
75 | + TYPEW aa = *(TYPEW *)(va + HW(i)); \ | ||
76 | + *(TYPEW *)(vd + HW(i)) = SUM_OP(aa, DMUL_OP(nn, mm)); \ | ||
77 | + } \ | ||
78 | +} | ||
79 | + | ||
80 | +DO_SQDMLAL(sve2_sqdmlal_zzzw_h, int16_t, int8_t, H1_2, H1, | ||
81 | + do_sqdmull_h, DO_SQADD_H) | ||
82 | +DO_SQDMLAL(sve2_sqdmlal_zzzw_s, int32_t, int16_t, H1_4, H1_2, | ||
83 | + do_sqdmull_s, DO_SQADD_S) | ||
84 | +DO_SQDMLAL(sve2_sqdmlal_zzzw_d, int64_t, int32_t, , H1_4, | ||
85 | + do_sqdmull_d, do_sqadd_d) | ||
86 | + | ||
87 | +DO_SQDMLAL(sve2_sqdmlsl_zzzw_h, int16_t, int8_t, H1_2, H1, | ||
88 | + do_sqdmull_h, DO_SQSUB_H) | ||
89 | +DO_SQDMLAL(sve2_sqdmlsl_zzzw_s, int32_t, int16_t, H1_4, H1_2, | ||
90 | + do_sqdmull_s, DO_SQSUB_S) | ||
91 | +DO_SQDMLAL(sve2_sqdmlsl_zzzw_d, int64_t, int32_t, , H1_4, | ||
92 | + do_sqdmull_d, do_sqsub_d) | ||
93 | + | ||
94 | +#undef DO_SQDMLAL | ||
95 | + | ||
96 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
97 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
98 | { \ | ||
99 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
100 | index XXXXXXX..XXXXXXX 100644 | ||
101 | --- a/target/arm/translate-sve.c | ||
102 | +++ b/target/arm/translate-sve.c | ||
103 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ_FP(FMAXNMP, fmaxnmp) | ||
104 | DO_SVE2_ZPZZ_FP(FMINNMP, fminnmp) | ||
105 | DO_SVE2_ZPZZ_FP(FMAXP, fmaxp) | ||
106 | DO_SVE2_ZPZZ_FP(FMINP, fminp) | ||
107 | + | ||
108 | +/* | ||
109 | + * SVE Integer Multiply-Add (unpredicated) | ||
110 | + */ | ||
111 | + | ||
112 | +static bool do_sqdmlal_zzzw(DisasContext *s, arg_rrrr_esz *a, | ||
113 | + bool sel1, bool sel2) | ||
114 | +{ | ||
115 | + static gen_helper_gvec_4 * const fns[] = { | ||
116 | + NULL, gen_helper_sve2_sqdmlal_zzzw_h, | ||
117 | + gen_helper_sve2_sqdmlal_zzzw_s, gen_helper_sve2_sqdmlal_zzzw_d, | ||
118 | + }; | ||
119 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], (sel2 << 1) | sel1); | ||
120 | +} | ||
121 | + | ||
122 | +static bool do_sqdmlsl_zzzw(DisasContext *s, arg_rrrr_esz *a, | ||
123 | + bool sel1, bool sel2) | ||
124 | +{ | ||
125 | + static gen_helper_gvec_4 * const fns[] = { | ||
126 | + NULL, gen_helper_sve2_sqdmlsl_zzzw_h, | ||
127 | + gen_helper_sve2_sqdmlsl_zzzw_s, gen_helper_sve2_sqdmlsl_zzzw_d, | ||
128 | + }; | ||
129 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], (sel2 << 1) | sel1); | ||
130 | +} | ||
131 | + | ||
132 | +static bool trans_SQDMLALB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
133 | +{ | ||
134 | + return do_sqdmlal_zzzw(s, a, false, false); | ||
135 | +} | ||
136 | + | ||
137 | +static bool trans_SQDMLALT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
138 | +{ | ||
139 | + return do_sqdmlal_zzzw(s, a, true, true); | ||
140 | +} | ||
141 | + | ||
142 | +static bool trans_SQDMLALBT(DisasContext *s, arg_rrrr_esz *a) | ||
143 | +{ | ||
144 | + return do_sqdmlal_zzzw(s, a, false, true); | ||
145 | +} | ||
146 | + | ||
147 | +static bool trans_SQDMLSLB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
148 | +{ | ||
149 | + return do_sqdmlsl_zzzw(s, a, false, false); | ||
150 | +} | ||
151 | + | ||
152 | +static bool trans_SQDMLSLT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
153 | +{ | ||
154 | + return do_sqdmlsl_zzzw(s, a, true, true); | ||
155 | +} | ||
156 | + | ||
157 | +static bool trans_SQDMLSLBT(DisasContext *s, arg_rrrr_esz *a) | ||
158 | +{ | ||
159 | + return do_sqdmlsl_zzzw(s, a, false, true); | ||
160 | +} | ||
161 | -- | ||
162 | 2.20.1 | ||
163 | |||
164 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | From: Inès Varhol <ines.varhol@telecom-paris.fr> |
---|---|---|---|
2 | 2 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 3 | The testcase contains : |
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | 4 | - `test_idr_reset_value()` : |
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 5 | Checks the reset values of MODER, OTYPER, PUPDR, ODR and IDR. |
6 | Message-id: 20210525010358.152808-73-richard.henderson@linaro.org | 6 | - `test_gpio_output_mode()` : |
7 | Message-Id: <20200428174332.17162-2-steplong@quicinc.com> | 7 | Checks that writing a bit in register ODR results in the corresponding |
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 8 | pin rising or lowering, if this pin is configured in output mode. |
9 | - `test_gpio_input_mode()` : | ||
10 | Checks that a input pin set high or low externally results | ||
11 | in the pin rising and lowering. | ||
12 | - `test_pull_up_pull_down()` : | ||
13 | Checks that a floating pin in pull-up/down mode is actually high/down. | ||
14 | - `test_push_pull()` : | ||
15 | Checks that a pin set externally is disconnected when configured in | ||
16 | push-pull output mode, and can't be set externally while in this mode. | ||
17 | - `test_open_drain()` : | ||
18 | Checks that a pin set externally high is disconnected when configured | ||
19 | in open-drain output mode, and can't be set high while in this mode. | ||
20 | - `test_bsrr_brr()` : | ||
21 | Checks that writing to BSRR and BRR has the desired result in ODR. | ||
22 | - `test_clock_enable()` : | ||
23 | Checks that GPIO clock is at the right frequency after enabling it. | ||
24 | |||
25 | Acked-by: Thomas Huth <thuth@redhat.com> | ||
26 | Signed-off-by: Arnaud Minier <arnaud.minier@telecom-paris.fr> | ||
27 | Signed-off-by: Inès Varhol <ines.varhol@telecom-paris.fr> | ||
28 | Message-id: 20240305210444.310665-4-ines.varhol@telecom-paris.fr | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 29 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
10 | --- | 30 | --- |
11 | target/arm/helper-sve.h | 5 +++++ | 31 | tests/qtest/stm32l4x5_gpio-test.c | 551 ++++++++++++++++++++++++++++++ |
12 | target/arm/sve.decode | 4 ++++ | 32 | tests/qtest/meson.build | 3 +- |
13 | target/arm/sve_helper.c | 20 ++++++++++++++++++++ | 33 | 2 files changed, 553 insertions(+), 1 deletion(-) |
14 | target/arm/translate-sve.c | 16 ++++++++++++++++ | 34 | create mode 100644 tests/qtest/stm32l4x5_gpio-test.c |
15 | 4 files changed, 45 insertions(+) | ||
16 | 35 | ||
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | 36 | diff --git a/tests/qtest/stm32l4x5_gpio-test.c b/tests/qtest/stm32l4x5_gpio-test.c |
37 | new file mode 100644 | ||
38 | index XXXXXXX..XXXXXXX | ||
39 | --- /dev/null | ||
40 | +++ b/tests/qtest/stm32l4x5_gpio-test.c | ||
41 | @@ -XXX,XX +XXX,XX @@ | ||
42 | +/* | ||
43 | + * QTest testcase for STM32L4x5_GPIO | ||
44 | + * | ||
45 | + * Copyright (c) 2024 Arnaud Minier <arnaud.minier@telecom-paris.fr> | ||
46 | + * Copyright (c) 2024 Inès Varhol <ines.varhol@telecom-paris.fr> | ||
47 | + * | ||
48 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. | ||
49 | + * See the COPYING file in the top-level directory. | ||
50 | + */ | ||
51 | + | ||
52 | +#include "qemu/osdep.h" | ||
53 | +#include "libqtest-single.h" | ||
54 | + | ||
55 | +#define GPIO_BASE_ADDR 0x48000000 | ||
56 | +#define GPIO_SIZE 0x400 | ||
57 | +#define NUM_GPIOS 8 | ||
58 | +#define NUM_GPIO_PINS 16 | ||
59 | + | ||
60 | +#define GPIO_A 0x48000000 | ||
61 | +#define GPIO_B 0x48000400 | ||
62 | +#define GPIO_C 0x48000800 | ||
63 | +#define GPIO_D 0x48000C00 | ||
64 | +#define GPIO_E 0x48001000 | ||
65 | +#define GPIO_F 0x48001400 | ||
66 | +#define GPIO_G 0x48001800 | ||
67 | +#define GPIO_H 0x48001C00 | ||
68 | + | ||
69 | +#define MODER 0x00 | ||
70 | +#define OTYPER 0x04 | ||
71 | +#define PUPDR 0x0C | ||
72 | +#define IDR 0x10 | ||
73 | +#define ODR 0x14 | ||
74 | +#define BSRR 0x18 | ||
75 | +#define BRR 0x28 | ||
76 | + | ||
77 | +#define MODER_INPUT 0 | ||
78 | +#define MODER_OUTPUT 1 | ||
79 | + | ||
80 | +#define PUPDR_NONE 0 | ||
81 | +#define PUPDR_PULLUP 1 | ||
82 | +#define PUPDR_PULLDOWN 2 | ||
83 | + | ||
84 | +#define OTYPER_PUSH_PULL 0 | ||
85 | +#define OTYPER_OPEN_DRAIN 1 | ||
86 | + | ||
87 | +const uint32_t moder_reset[NUM_GPIOS] = { | ||
88 | + 0xABFFFFFF, | ||
89 | + 0xFFFFFEBF, | ||
90 | + 0xFFFFFFFF, | ||
91 | + 0xFFFFFFFF, | ||
92 | + 0xFFFFFFFF, | ||
93 | + 0xFFFFFFFF, | ||
94 | + 0xFFFFFFFF, | ||
95 | + 0x0000000F | ||
96 | +}; | ||
97 | + | ||
98 | +const uint32_t pupdr_reset[NUM_GPIOS] = { | ||
99 | + 0x64000000, | ||
100 | + 0x00000100, | ||
101 | + 0x00000000, | ||
102 | + 0x00000000, | ||
103 | + 0x00000000, | ||
104 | + 0x00000000, | ||
105 | + 0x00000000, | ||
106 | + 0x00000000 | ||
107 | +}; | ||
108 | + | ||
109 | +const uint32_t idr_reset[NUM_GPIOS] = { | ||
110 | + 0x0000A000, | ||
111 | + 0x00000010, | ||
112 | + 0x00000000, | ||
113 | + 0x00000000, | ||
114 | + 0x00000000, | ||
115 | + 0x00000000, | ||
116 | + 0x00000000, | ||
117 | + 0x00000000 | ||
118 | +}; | ||
119 | + | ||
120 | +static uint32_t gpio_readl(unsigned int gpio, unsigned int offset) | ||
121 | +{ | ||
122 | + return readl(gpio + offset); | ||
123 | +} | ||
124 | + | ||
125 | +static void gpio_writel(unsigned int gpio, unsigned int offset, uint32_t value) | ||
126 | +{ | ||
127 | + writel(gpio + offset, value); | ||
128 | +} | ||
129 | + | ||
130 | +static void gpio_set_bit(unsigned int gpio, unsigned int reg, | ||
131 | + unsigned int pin, uint32_t value) | ||
132 | +{ | ||
133 | + uint32_t mask = 0xFFFFFFFF & ~(0x1 << pin); | ||
134 | + gpio_writel(gpio, reg, (gpio_readl(gpio, reg) & mask) | value << pin); | ||
135 | +} | ||
136 | + | ||
137 | +static void gpio_set_2bits(unsigned int gpio, unsigned int reg, | ||
138 | + unsigned int pin, uint32_t value) | ||
139 | +{ | ||
140 | + uint32_t offset = 2 * pin; | ||
141 | + uint32_t mask = 0xFFFFFFFF & ~(0x3 << offset); | ||
142 | + gpio_writel(gpio, reg, (gpio_readl(gpio, reg) & mask) | value << offset); | ||
143 | +} | ||
144 | + | ||
145 | +static unsigned int get_gpio_id(uint32_t gpio_addr) | ||
146 | +{ | ||
147 | + return (gpio_addr - GPIO_BASE_ADDR) / GPIO_SIZE; | ||
148 | +} | ||
149 | + | ||
150 | +static void gpio_set_irq(unsigned int gpio, int num, int level) | ||
151 | +{ | ||
152 | + g_autofree char *name = g_strdup_printf("/machine/soc/gpio%c", | ||
153 | + get_gpio_id(gpio) + 'a'); | ||
154 | + qtest_set_irq_in(global_qtest, name, NULL, num, level); | ||
155 | +} | ||
156 | + | ||
157 | +static void disconnect_all_pins(unsigned int gpio) | ||
158 | +{ | ||
159 | + g_autofree char *path = g_strdup_printf("/machine/soc/gpio%c", | ||
160 | + get_gpio_id(gpio) + 'a'); | ||
161 | + QDict *r; | ||
162 | + | ||
163 | + r = qtest_qmp(global_qtest, "{ 'execute': 'qom-set', 'arguments': " | ||
164 | + "{ 'path': %s, 'property': 'disconnected-pins', 'value': %d } }", | ||
165 | + path, 0xFFFF); | ||
166 | + g_assert_false(qdict_haskey(r, "error")); | ||
167 | + qobject_unref(r); | ||
168 | +} | ||
169 | + | ||
170 | +static uint32_t get_disconnected_pins(unsigned int gpio) | ||
171 | +{ | ||
172 | + g_autofree char *path = g_strdup_printf("/machine/soc/gpio%c", | ||
173 | + get_gpio_id(gpio) + 'a'); | ||
174 | + uint32_t disconnected_pins = 0; | ||
175 | + QDict *r; | ||
176 | + | ||
177 | + r = qtest_qmp(global_qtest, "{ 'execute': 'qom-get', 'arguments':" | ||
178 | + " { 'path': %s, 'property': 'disconnected-pins'} }", path); | ||
179 | + g_assert_false(qdict_haskey(r, "error")); | ||
180 | + disconnected_pins = qdict_get_int(r, "return"); | ||
181 | + qobject_unref(r); | ||
182 | + return disconnected_pins; | ||
183 | +} | ||
184 | + | ||
185 | +static uint32_t reset(uint32_t gpio, unsigned int offset) | ||
186 | +{ | ||
187 | + switch (offset) { | ||
188 | + case MODER: | ||
189 | + return moder_reset[get_gpio_id(gpio)]; | ||
190 | + case PUPDR: | ||
191 | + return pupdr_reset[get_gpio_id(gpio)]; | ||
192 | + case IDR: | ||
193 | + return idr_reset[get_gpio_id(gpio)]; | ||
194 | + } | ||
195 | + return 0x0; | ||
196 | +} | ||
197 | + | ||
198 | +static void system_reset(void) | ||
199 | +{ | ||
200 | + QDict *r; | ||
201 | + r = qtest_qmp(global_qtest, "{'execute': 'system_reset'}"); | ||
202 | + g_assert_false(qdict_haskey(r, "error")); | ||
203 | + qobject_unref(r); | ||
204 | +} | ||
205 | + | ||
206 | +static void test_idr_reset_value(void) | ||
207 | +{ | ||
208 | + /* | ||
209 | + * Checks that the values in MODER, OTYPER, PUPDR and ODR | ||
210 | + * after reset are correct, and that the value in IDR is | ||
211 | + * coherent. | ||
212 | + * Since AF and analog modes aren't implemented, IDR reset | ||
213 | + * values aren't the same as with a real board. | ||
214 | + * | ||
215 | + * Register IDR contains the actual values of all GPIO pins. | ||
216 | + * Its value depends on the pins' configuration | ||
217 | + * (intput/output/analog : register MODER, push-pull/open-drain : | ||
218 | + * register OTYPER, pull-up/pull-down/none : register PUPDR) | ||
219 | + * and on the values stored in register ODR | ||
220 | + * (in case the pin is in output mode). | ||
221 | + */ | ||
222 | + | ||
223 | + gpio_writel(GPIO_A, MODER, 0xDEADBEEF); | ||
224 | + gpio_writel(GPIO_A, ODR, 0xDEADBEEF); | ||
225 | + gpio_writel(GPIO_A, OTYPER, 0xDEADBEEF); | ||
226 | + gpio_writel(GPIO_A, PUPDR, 0xDEADBEEF); | ||
227 | + | ||
228 | + gpio_writel(GPIO_B, MODER, 0xDEADBEEF); | ||
229 | + gpio_writel(GPIO_B, ODR, 0xDEADBEEF); | ||
230 | + gpio_writel(GPIO_B, OTYPER, 0xDEADBEEF); | ||
231 | + gpio_writel(GPIO_B, PUPDR, 0xDEADBEEF); | ||
232 | + | ||
233 | + gpio_writel(GPIO_C, MODER, 0xDEADBEEF); | ||
234 | + gpio_writel(GPIO_C, ODR, 0xDEADBEEF); | ||
235 | + gpio_writel(GPIO_C, OTYPER, 0xDEADBEEF); | ||
236 | + gpio_writel(GPIO_C, PUPDR, 0xDEADBEEF); | ||
237 | + | ||
238 | + gpio_writel(GPIO_H, MODER, 0xDEADBEEF); | ||
239 | + gpio_writel(GPIO_H, ODR, 0xDEADBEEF); | ||
240 | + gpio_writel(GPIO_H, OTYPER, 0xDEADBEEF); | ||
241 | + gpio_writel(GPIO_H, PUPDR, 0xDEADBEEF); | ||
242 | + | ||
243 | + system_reset(); | ||
244 | + | ||
245 | + uint32_t moder = gpio_readl(GPIO_A, MODER); | ||
246 | + uint32_t odr = gpio_readl(GPIO_A, ODR); | ||
247 | + uint32_t otyper = gpio_readl(GPIO_A, OTYPER); | ||
248 | + uint32_t pupdr = gpio_readl(GPIO_A, PUPDR); | ||
249 | + uint32_t idr = gpio_readl(GPIO_A, IDR); | ||
250 | + /* 15: AF, 14: AF, 13: AF, 12: Analog ... */ | ||
251 | + /* here AF is the same as Analog and Input mode */ | ||
252 | + g_assert_cmphex(moder, ==, reset(GPIO_A, MODER)); | ||
253 | + g_assert_cmphex(odr, ==, reset(GPIO_A, ODR)); | ||
254 | + g_assert_cmphex(otyper, ==, reset(GPIO_A, OTYPER)); | ||
255 | + /* 15: pull-up, 14: pull-down, 13: pull-up, 12: neither ... */ | ||
256 | + g_assert_cmphex(pupdr, ==, reset(GPIO_A, PUPDR)); | ||
257 | + /* 15 : 1, 14: 0, 13: 1, 12 : reset value ... */ | ||
258 | + g_assert_cmphex(idr, ==, reset(GPIO_A, IDR)); | ||
259 | + | ||
260 | + moder = gpio_readl(GPIO_B, MODER); | ||
261 | + odr = gpio_readl(GPIO_B, ODR); | ||
262 | + otyper = gpio_readl(GPIO_B, OTYPER); | ||
263 | + pupdr = gpio_readl(GPIO_B, PUPDR); | ||
264 | + idr = gpio_readl(GPIO_B, IDR); | ||
265 | + /* ... 5: Analog, 4: AF, 3: AF, 2: Analog ... */ | ||
266 | + /* here AF is the same as Analog and Input mode */ | ||
267 | + g_assert_cmphex(moder, ==, reset(GPIO_B, MODER)); | ||
268 | + g_assert_cmphex(odr, ==, reset(GPIO_B, ODR)); | ||
269 | + g_assert_cmphex(otyper, ==, reset(GPIO_B, OTYPER)); | ||
270 | + /* ... 5: neither, 4: pull-up, 3: neither ... */ | ||
271 | + g_assert_cmphex(pupdr, ==, reset(GPIO_B, PUPDR)); | ||
272 | + /* ... 5 : reset value, 4 : 1, 3 : reset value ... */ | ||
273 | + g_assert_cmphex(idr, ==, reset(GPIO_B, IDR)); | ||
274 | + | ||
275 | + moder = gpio_readl(GPIO_C, MODER); | ||
276 | + odr = gpio_readl(GPIO_C, ODR); | ||
277 | + otyper = gpio_readl(GPIO_C, OTYPER); | ||
278 | + pupdr = gpio_readl(GPIO_C, PUPDR); | ||
279 | + idr = gpio_readl(GPIO_C, IDR); | ||
280 | + /* Analog, same as Input mode*/ | ||
281 | + g_assert_cmphex(moder, ==, reset(GPIO_C, MODER)); | ||
282 | + g_assert_cmphex(odr, ==, reset(GPIO_C, ODR)); | ||
283 | + g_assert_cmphex(otyper, ==, reset(GPIO_C, OTYPER)); | ||
284 | + /* no pull-up or pull-down */ | ||
285 | + g_assert_cmphex(pupdr, ==, reset(GPIO_C, PUPDR)); | ||
286 | + /* reset value */ | ||
287 | + g_assert_cmphex(idr, ==, reset(GPIO_C, IDR)); | ||
288 | + | ||
289 | + moder = gpio_readl(GPIO_H, MODER); | ||
290 | + odr = gpio_readl(GPIO_H, ODR); | ||
291 | + otyper = gpio_readl(GPIO_H, OTYPER); | ||
292 | + pupdr = gpio_readl(GPIO_H, PUPDR); | ||
293 | + idr = gpio_readl(GPIO_H, IDR); | ||
294 | + /* Analog, same as Input mode */ | ||
295 | + g_assert_cmphex(moder, ==, reset(GPIO_H, MODER)); | ||
296 | + g_assert_cmphex(odr, ==, reset(GPIO_H, ODR)); | ||
297 | + g_assert_cmphex(otyper, ==, reset(GPIO_H, OTYPER)); | ||
298 | + /* no pull-up or pull-down */ | ||
299 | + g_assert_cmphex(pupdr, ==, reset(GPIO_H, PUPDR)); | ||
300 | + /* reset value */ | ||
301 | + g_assert_cmphex(idr, ==, reset(GPIO_H, IDR)); | ||
302 | +} | ||
303 | + | ||
304 | +static void test_gpio_output_mode(const void *data) | ||
305 | +{ | ||
306 | + /* | ||
307 | + * Checks that setting a bit in ODR sets the corresponding | ||
308 | + * GPIO line high : it should set the right bit in IDR | ||
309 | + * and send an irq to syscfg. | ||
310 | + * Additionally, it checks that values written to ODR | ||
311 | + * when not in output mode are stored and not discarded. | ||
312 | + */ | ||
313 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
314 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
315 | + unsigned int gpio_id = get_gpio_id(gpio); | ||
316 | + | ||
317 | + qtest_irq_intercept_in(global_qtest, "/machine/soc/syscfg"); | ||
318 | + | ||
319 | + /* Set a bit in ODR and check nothing happens */ | ||
320 | + gpio_set_bit(gpio, ODR, pin, 1); | ||
321 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR)); | ||
322 | + g_assert_false(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
323 | + | ||
324 | + /* Configure the relevant line as output and check the pin is high */ | ||
325 | + gpio_set_2bits(gpio, MODER, pin, MODER_OUTPUT); | ||
326 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) | (1 << pin)); | ||
327 | + g_assert_true(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
328 | + | ||
329 | + /* Reset the bit in ODR and check the pin is low */ | ||
330 | + gpio_set_bit(gpio, ODR, pin, 0); | ||
331 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
332 | + g_assert_false(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
333 | + | ||
334 | + /* Clean the test */ | ||
335 | + gpio_writel(gpio, ODR, reset(gpio, ODR)); | ||
336 | + gpio_writel(gpio, MODER, reset(gpio, MODER)); | ||
337 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR)); | ||
338 | + g_assert_false(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
339 | +} | ||
340 | + | ||
341 | +static void test_gpio_input_mode(const void *data) | ||
342 | +{ | ||
343 | + /* | ||
344 | + * Test that setting a line high/low externally sets the | ||
345 | + * corresponding GPIO line high/low : it should set the | ||
346 | + * right bit in IDR and send an irq to syscfg. | ||
347 | + */ | ||
348 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
349 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
350 | + unsigned int gpio_id = get_gpio_id(gpio); | ||
351 | + | ||
352 | + qtest_irq_intercept_in(global_qtest, "/machine/soc/syscfg"); | ||
353 | + | ||
354 | + /* Configure a line as input, raise it, and check that the pin is high */ | ||
355 | + gpio_set_2bits(gpio, MODER, pin, MODER_INPUT); | ||
356 | + gpio_set_irq(gpio, pin, 1); | ||
357 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) | (1 << pin)); | ||
358 | + g_assert_true(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
359 | + | ||
360 | + /* Lower the line and check that the pin is low */ | ||
361 | + gpio_set_irq(gpio, pin, 0); | ||
362 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
363 | + g_assert_false(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
364 | + | ||
365 | + /* Clean the test */ | ||
366 | + gpio_writel(gpio, MODER, reset(gpio, MODER)); | ||
367 | + disconnect_all_pins(gpio); | ||
368 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR)); | ||
369 | +} | ||
370 | + | ||
371 | +static void test_pull_up_pull_down(const void *data) | ||
372 | +{ | ||
373 | + /* | ||
374 | + * Test that a floating pin with pull-up sets the pin | ||
375 | + * high and vice-versa. | ||
376 | + */ | ||
377 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
378 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
379 | + unsigned int gpio_id = get_gpio_id(gpio); | ||
380 | + | ||
381 | + qtest_irq_intercept_in(global_qtest, "/machine/soc/syscfg"); | ||
382 | + | ||
383 | + /* Configure a line as input with pull-up, check the line is set high */ | ||
384 | + gpio_set_2bits(gpio, MODER, pin, MODER_INPUT); | ||
385 | + gpio_set_2bits(gpio, PUPDR, pin, PUPDR_PULLUP); | ||
386 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) | (1 << pin)); | ||
387 | + g_assert_true(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
388 | + | ||
389 | + /* Configure the line with pull-down, check the line is low */ | ||
390 | + gpio_set_2bits(gpio, PUPDR, pin, PUPDR_PULLDOWN); | ||
391 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
392 | + g_assert_false(get_irq(gpio_id * NUM_GPIO_PINS + pin)); | ||
393 | + | ||
394 | + /* Clean the test */ | ||
395 | + gpio_writel(gpio, MODER, reset(gpio, MODER)); | ||
396 | + gpio_writel(gpio, PUPDR, reset(gpio, PUPDR)); | ||
397 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR)); | ||
398 | +} | ||
399 | + | ||
400 | +static void test_push_pull(const void *data) | ||
401 | +{ | ||
402 | + /* | ||
403 | + * Test that configuring a line in push-pull output mode | ||
404 | + * disconnects the pin, that the pin can't be set or reset | ||
405 | + * externally afterwards. | ||
406 | + */ | ||
407 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
408 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
409 | + uint32_t gpio2 = GPIO_BASE_ADDR + (GPIO_H - gpio); | ||
410 | + | ||
411 | + qtest_irq_intercept_in(global_qtest, "/machine/soc/syscfg"); | ||
412 | + | ||
413 | + /* Setting a line high externally, configuring it in push-pull output */ | ||
414 | + /* And checking the pin was disconnected */ | ||
415 | + gpio_set_irq(gpio, pin, 1); | ||
416 | + gpio_set_2bits(gpio, MODER, pin, MODER_OUTPUT); | ||
417 | + g_assert_cmphex(get_disconnected_pins(gpio), ==, 0xFFFF); | ||
418 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
419 | + | ||
420 | + /* Setting a line low externally, configuring it in push-pull output */ | ||
421 | + /* And checking the pin was disconnected */ | ||
422 | + gpio_set_irq(gpio2, pin, 0); | ||
423 | + gpio_set_bit(gpio2, ODR, pin, 1); | ||
424 | + gpio_set_2bits(gpio2, MODER, pin, MODER_OUTPUT); | ||
425 | + g_assert_cmphex(get_disconnected_pins(gpio2), ==, 0xFFFF); | ||
426 | + g_assert_cmphex(gpio_readl(gpio2, IDR), ==, reset(gpio2, IDR) | (1 << pin)); | ||
427 | + | ||
428 | + /* Trying to set a push-pull output pin, checking it doesn't work */ | ||
429 | + gpio_set_irq(gpio, pin, 1); | ||
430 | + g_assert_cmphex(get_disconnected_pins(gpio), ==, 0xFFFF); | ||
431 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
432 | + | ||
433 | + /* Trying to reset a push-pull output pin, checking it doesn't work */ | ||
434 | + gpio_set_irq(gpio2, pin, 0); | ||
435 | + g_assert_cmphex(get_disconnected_pins(gpio2), ==, 0xFFFF); | ||
436 | + g_assert_cmphex(gpio_readl(gpio2, IDR), ==, reset(gpio2, IDR) | (1 << pin)); | ||
437 | + | ||
438 | + /* Clean the test */ | ||
439 | + gpio_writel(gpio, MODER, reset(gpio, MODER)); | ||
440 | + gpio_writel(gpio2, ODR, reset(gpio2, ODR)); | ||
441 | + gpio_writel(gpio2, MODER, reset(gpio2, MODER)); | ||
442 | +} | ||
443 | + | ||
444 | +static void test_open_drain(const void *data) | ||
445 | +{ | ||
446 | + /* | ||
447 | + * Test that configuring a line in open-drain output mode | ||
448 | + * disconnects a pin set high externally and that the pin | ||
449 | + * can't be set high externally while configured in open-drain. | ||
450 | + * | ||
451 | + * However a pin set low externally shouldn't be disconnected, | ||
452 | + * and it can be set low externally when in open-drain mode. | ||
453 | + */ | ||
454 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
455 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
456 | + uint32_t gpio2 = GPIO_BASE_ADDR + (GPIO_H - gpio); | ||
457 | + | ||
458 | + qtest_irq_intercept_in(global_qtest, "/machine/soc/syscfg"); | ||
459 | + | ||
460 | + /* Setting a line high externally, configuring it in open-drain output */ | ||
461 | + /* And checking the pin was disconnected */ | ||
462 | + gpio_set_irq(gpio, pin, 1); | ||
463 | + gpio_set_bit(gpio, OTYPER, pin, OTYPER_OPEN_DRAIN); | ||
464 | + gpio_set_2bits(gpio, MODER, pin, MODER_OUTPUT); | ||
465 | + g_assert_cmphex(get_disconnected_pins(gpio), ==, 0xFFFF); | ||
466 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
467 | + | ||
468 | + /* Setting a line low externally, configuring it in open-drain output */ | ||
469 | + /* And checking the pin wasn't disconnected */ | ||
470 | + gpio_set_irq(gpio2, pin, 0); | ||
471 | + gpio_set_bit(gpio2, ODR, pin, 1); | ||
472 | + gpio_set_bit(gpio2, OTYPER, pin, OTYPER_OPEN_DRAIN); | ||
473 | + gpio_set_2bits(gpio2, MODER, pin, MODER_OUTPUT); | ||
474 | + g_assert_cmphex(get_disconnected_pins(gpio2), ==, 0xFFFF & ~(1 << pin)); | ||
475 | + g_assert_cmphex(gpio_readl(gpio2, IDR), ==, | ||
476 | + reset(gpio2, IDR) & ~(1 << pin)); | ||
477 | + | ||
478 | + /* Trying to set a open-drain output pin, checking it doesn't work */ | ||
479 | + gpio_set_irq(gpio, pin, 1); | ||
480 | + g_assert_cmphex(get_disconnected_pins(gpio), ==, 0xFFFF); | ||
481 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR) & ~(1 << pin)); | ||
482 | + | ||
483 | + /* Trying to reset a open-drain output pin, checking it works */ | ||
484 | + gpio_set_bit(gpio, ODR, pin, 1); | ||
485 | + gpio_set_irq(gpio, pin, 0); | ||
486 | + g_assert_cmphex(get_disconnected_pins(gpio2), ==, 0xFFFF & ~(1 << pin)); | ||
487 | + g_assert_cmphex(gpio_readl(gpio2, IDR), ==, | ||
488 | + reset(gpio2, IDR) & ~(1 << pin)); | ||
489 | + | ||
490 | + /* Clean the test */ | ||
491 | + disconnect_all_pins(gpio2); | ||
492 | + gpio_writel(gpio2, OTYPER, reset(gpio2, OTYPER)); | ||
493 | + gpio_writel(gpio2, ODR, reset(gpio2, ODR)); | ||
494 | + gpio_writel(gpio2, MODER, reset(gpio2, MODER)); | ||
495 | + g_assert_cmphex(gpio_readl(gpio2, IDR), ==, reset(gpio2, IDR)); | ||
496 | + disconnect_all_pins(gpio); | ||
497 | + gpio_writel(gpio, OTYPER, reset(gpio, OTYPER)); | ||
498 | + gpio_writel(gpio, ODR, reset(gpio, ODR)); | ||
499 | + gpio_writel(gpio, MODER, reset(gpio, MODER)); | ||
500 | + g_assert_cmphex(gpio_readl(gpio, IDR), ==, reset(gpio, IDR)); | ||
501 | +} | ||
502 | + | ||
503 | +static void test_bsrr_brr(const void *data) | ||
504 | +{ | ||
505 | + /* | ||
506 | + * Test that writing a '1' in BSS and BSRR | ||
507 | + * has the desired effect on ODR. | ||
508 | + * In BSRR, BSx has priority over BRx. | ||
509 | + */ | ||
510 | + unsigned int pin = ((uint64_t)data) & 0xF; | ||
511 | + uint32_t gpio = ((uint64_t)data) >> 32; | ||
512 | + | ||
513 | + gpio_writel(gpio, BSRR, (1 << pin)); | ||
514 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR) | (1 << pin)); | ||
515 | + | ||
516 | + gpio_writel(gpio, BSRR, (1 << (pin + NUM_GPIO_PINS))); | ||
517 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR)); | ||
518 | + | ||
519 | + gpio_writel(gpio, BSRR, (1 << pin)); | ||
520 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR) | (1 << pin)); | ||
521 | + | ||
522 | + gpio_writel(gpio, BRR, (1 << pin)); | ||
523 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR)); | ||
524 | + | ||
525 | + /* BSx should have priority over BRx */ | ||
526 | + gpio_writel(gpio, BSRR, (1 << pin) | (1 << (pin + NUM_GPIO_PINS))); | ||
527 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR) | (1 << pin)); | ||
528 | + | ||
529 | + gpio_writel(gpio, BRR, (1 << pin)); | ||
530 | + g_assert_cmphex(gpio_readl(gpio, ODR), ==, reset(gpio, ODR)); | ||
531 | + | ||
532 | + gpio_writel(gpio, ODR, reset(gpio, ODR)); | ||
533 | +} | ||
534 | + | ||
535 | +int main(int argc, char **argv) | ||
536 | +{ | ||
537 | + int ret; | ||
538 | + | ||
539 | + g_test_init(&argc, &argv, NULL); | ||
540 | + g_test_set_nonfatal_assertions(); | ||
541 | + qtest_add_func("stm32l4x5/gpio/test_idr_reset_value", | ||
542 | + test_idr_reset_value); | ||
543 | + /* | ||
544 | + * The inputs for the tests (gpio and pin) can be changed, | ||
545 | + * but the tests don't work for pins that are high at reset | ||
546 | + * (GPIOA15, GPIO13 and GPIOB5). | ||
547 | + * Specifically, rising the pin then checking `get_irq()` | ||
548 | + * is problematic since the pin was already high. | ||
549 | + */ | ||
550 | + qtest_add_data_func("stm32l4x5/gpio/test_gpioc5_output_mode", | ||
551 | + (void *)((uint64_t)GPIO_C << 32 | 5), | ||
552 | + test_gpio_output_mode); | ||
553 | + qtest_add_data_func("stm32l4x5/gpio/test_gpioh3_output_mode", | ||
554 | + (void *)((uint64_t)GPIO_H << 32 | 3), | ||
555 | + test_gpio_output_mode); | ||
556 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_input_mode1", | ||
557 | + (void *)((uint64_t)GPIO_D << 32 | 6), | ||
558 | + test_gpio_input_mode); | ||
559 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_input_mode2", | ||
560 | + (void *)((uint64_t)GPIO_C << 32 | 10), | ||
561 | + test_gpio_input_mode); | ||
562 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_pull_up_pull_down1", | ||
563 | + (void *)((uint64_t)GPIO_B << 32 | 5), | ||
564 | + test_pull_up_pull_down); | ||
565 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_pull_up_pull_down2", | ||
566 | + (void *)((uint64_t)GPIO_F << 32 | 1), | ||
567 | + test_pull_up_pull_down); | ||
568 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_push_pull1", | ||
569 | + (void *)((uint64_t)GPIO_G << 32 | 6), | ||
570 | + test_push_pull); | ||
571 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_push_pull2", | ||
572 | + (void *)((uint64_t)GPIO_H << 32 | 3), | ||
573 | + test_push_pull); | ||
574 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_open_drain1", | ||
575 | + (void *)((uint64_t)GPIO_C << 32 | 4), | ||
576 | + test_open_drain); | ||
577 | + qtest_add_data_func("stm32l4x5/gpio/test_gpio_open_drain2", | ||
578 | + (void *)((uint64_t)GPIO_E << 32 | 11), | ||
579 | + test_open_drain); | ||
580 | + qtest_add_data_func("stm32l4x5/gpio/test_bsrr_brr1", | ||
581 | + (void *)((uint64_t)GPIO_A << 32 | 12), | ||
582 | + test_bsrr_brr); | ||
583 | + qtest_add_data_func("stm32l4x5/gpio/test_bsrr_brr2", | ||
584 | + (void *)((uint64_t)GPIO_D << 32 | 0), | ||
585 | + test_bsrr_brr); | ||
586 | + | ||
587 | + qtest_start("-machine b-l475e-iot01a"); | ||
588 | + ret = g_test_run(); | ||
589 | + qtest_end(); | ||
590 | + | ||
591 | + return ret; | ||
592 | +} | ||
593 | diff --git a/tests/qtest/meson.build b/tests/qtest/meson.build | ||
18 | index XXXXXXX..XXXXXXX 100644 | 594 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/target/arm/helper-sve.h | 595 | --- a/tests/qtest/meson.build |
20 | +++ b/target/arm/helper-sve.h | 596 | +++ b/tests/qtest/meson.build |
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_cdot_idx_s, TCG_CALL_NO_RWG, | 597 | @@ -XXX,XX +XXX,XX @@ qtests_aspeed = \ |
22 | void, ptr, ptr, ptr, ptr, i32) | 598 | qtests_stm32l4x5 = \ |
23 | DEF_HELPER_FLAGS_5(sve2_cdot_idx_d, TCG_CALL_NO_RWG, | 599 | ['stm32l4x5_exti-test', |
24 | void, ptr, ptr, ptr, ptr, i32) | 600 | 'stm32l4x5_syscfg-test', |
25 | + | 601 | - 'stm32l4x5_rcc-test'] |
26 | +DEF_HELPER_FLAGS_5(sve2_fcvtnt_sh, TCG_CALL_NO_RWG, | 602 | + 'stm32l4x5_rcc-test', |
27 | + void, ptr, ptr, ptr, ptr, i32) | 603 | + 'stm32l4x5_gpio-test'] |
28 | +DEF_HELPER_FLAGS_5(sve2_fcvtnt_ds, TCG_CALL_NO_RWG, | 604 | |
29 | + void, ptr, ptr, ptr, ptr, i32) | 605 | qtests_arm = \ |
30 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | 606 | (config_all_devices.has_key('CONFIG_MPS2') ? ['sse-timer-test'] : []) + \ |
31 | index XXXXXXX..XXXXXXX 100644 | ||
32 | --- a/target/arm/sve.decode | ||
33 | +++ b/target/arm/sve.decode | ||
34 | @@ -XXX,XX +XXX,XX @@ SM4E 01000101 00 10001 1 11100 0 ..... ..... @rdn_rm_e0 | ||
35 | # SVE2 crypto constructive binary operations | ||
36 | SM4EKEY 01000101 00 1 ..... 11110 0 ..... ..... @rd_rn_rm_e0 | ||
37 | RAX1 01000101 00 1 ..... 11110 1 ..... ..... @rd_rn_rm_e0 | ||
38 | + | ||
39 | +### SVE2 floating-point convert precision odd elements | ||
40 | +FCVTNT_sh 01100100 10 0010 00 101 ... ..... ..... @rd_pg_rn_e0 | ||
41 | +FCVTNT_ds 01100100 11 0010 10 101 ... ..... ..... @rd_pg_rn_e0 | ||
42 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
43 | index XXXXXXX..XXXXXXX 100644 | ||
44 | --- a/target/arm/sve_helper.c | ||
45 | +++ b/target/arm/sve_helper.c | ||
46 | @@ -XXX,XX +XXX,XX @@ void HELPER(fmmla_d)(void *vd, void *vn, void *vm, void *va, | ||
47 | d[3] = float64_add(a[3], float64_add(p0, p1, status), status); | ||
48 | } | ||
49 | } | ||
50 | + | ||
51 | +#define DO_FCVTNT(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
52 | +void HELPER(NAME)(void *vd, void *vn, void *vg, void *status, uint32_t desc) \ | ||
53 | +{ \ | ||
54 | + intptr_t i = simd_oprsz(desc); \ | ||
55 | + uint64_t *g = vg; \ | ||
56 | + do { \ | ||
57 | + uint64_t pg = g[(i - 1) >> 6]; \ | ||
58 | + do { \ | ||
59 | + i -= sizeof(TYPEW); \ | ||
60 | + if (likely((pg >> (i & 63)) & 1)) { \ | ||
61 | + TYPEW nn = *(TYPEW *)(vn + HW(i)); \ | ||
62 | + *(TYPEN *)(vd + HN(i + sizeof(TYPEN))) = OP(nn, status); \ | ||
63 | + } \ | ||
64 | + } while (i & 63); \ | ||
65 | + } while (i != 0); \ | ||
66 | +} | ||
67 | + | ||
68 | +DO_FCVTNT(sve2_fcvtnt_sh, uint32_t, uint16_t, H1_4, H1_2, sve_f32_to_f16) | ||
69 | +DO_FCVTNT(sve2_fcvtnt_ds, uint64_t, uint32_t, , H1_4, float64_to_float32) | ||
70 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
71 | index XXXXXXX..XXXXXXX 100644 | ||
72 | --- a/target/arm/translate-sve.c | ||
73 | +++ b/target/arm/translate-sve.c | ||
74 | @@ -XXX,XX +XXX,XX @@ static bool trans_RAX1(DisasContext *s, arg_rrr_esz *a) | ||
75 | } | ||
76 | return true; | ||
77 | } | ||
78 | + | ||
79 | +static bool trans_FCVTNT_sh(DisasContext *s, arg_rpr_esz *a) | ||
80 | +{ | ||
81 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
82 | + return false; | ||
83 | + } | ||
84 | + return do_zpz_ptr(s, a->rd, a->rn, a->pg, false, gen_helper_sve2_fcvtnt_sh); | ||
85 | +} | ||
86 | + | ||
87 | +static bool trans_FCVTNT_ds(DisasContext *s, arg_rpr_esz *a) | ||
88 | +{ | ||
89 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
90 | + return false; | ||
91 | + } | ||
92 | + return do_zpz_ptr(s, a->rd, a->rn, a->pg, false, gen_helper_sve2_fcvtnt_ds); | ||
93 | +} | ||
94 | -- | 607 | -- |
95 | 2.20.1 | 608 | 2.34.1 |
96 | 609 | ||
97 | 610 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | From: Richard Henderson <richard.henderson@linaro.org> |
---|---|---|---|
2 | 2 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 3 | While the 8-bit input elements are sequential in the input vector, |
4 | the 32-bit output elements are not sequential in the output matrix. | ||
5 | Do not attempt to compute 2 32-bit outputs at the same time. | ||
6 | |||
7 | Cc: qemu-stable@nongnu.org | ||
8 | Fixes: 23a5e3859f5 ("target/arm: Implement SME integer outer product") | ||
9 | Resolves: https://gitlab.com/qemu-project/qemu/-/issues/2083 | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 10 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
5 | Message-id: 20210525010358.152808-60-richard.henderson@linaro.org | 11 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
12 | Message-id: 20240305163931.242795-1-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 13 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
7 | --- | 14 | --- |
8 | target/arm/helper.h | 14 ++++++ | 15 | target/arm/tcg/sme_helper.c | 77 ++++++++++++++++++------------- |
9 | target/arm/sve.decode | 8 ++++ | 16 | tests/tcg/aarch64/sme-smopa-1.c | 47 +++++++++++++++++++ |
10 | target/arm/translate-sve.c | 8 ++++ | 17 | tests/tcg/aarch64/sme-smopa-2.c | 54 ++++++++++++++++++++++ |
11 | target/arm/vec_helper.c | 88 ++++++++++++++++++++++++++++++++++++++ | 18 | tests/tcg/aarch64/Makefile.target | 2 +- |
12 | 4 files changed, 118 insertions(+) | 19 | 4 files changed, 147 insertions(+), 33 deletions(-) |
13 | 20 | create mode 100644 tests/tcg/aarch64/sme-smopa-1.c | |
14 | diff --git a/target/arm/helper.h b/target/arm/helper.h | 21 | create mode 100644 tests/tcg/aarch64/sme-smopa-2.c |
22 | |||
23 | diff --git a/target/arm/tcg/sme_helper.c b/target/arm/tcg/sme_helper.c | ||
15 | index XXXXXXX..XXXXXXX 100644 | 24 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/target/arm/helper.h | 25 | --- a/target/arm/tcg/sme_helper.c |
17 | +++ b/target/arm/helper.h | 26 | +++ b/target/arm/tcg/sme_helper.c |
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_sqrdmulh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | 27 | @@ -XXX,XX +XXX,XX @@ void HELPER(sme_bfmopa)(void *vza, void *vzn, void *vzm, void *vpn, |
19 | DEF_HELPER_FLAGS_4(sve2_sqrdmulh_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_sqrdmulh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_idx_h, TCG_CALL_NO_RWG, | ||
23 | + void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_idx_s, TCG_CALL_NO_RWG, | ||
25 | + void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_idx_d, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, i32) | ||
28 | + | ||
29 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_idx_h, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_idx_s, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_idx_d, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, i32) | ||
35 | + | ||
36 | DEF_HELPER_FLAGS_4(gvec_xar_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
37 | |||
38 | #ifdef TARGET_AARCH64 | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ SQDMULLB_zzx_d 01000100 11 1 ..... 1110.0 ..... ..... @rrx_2a esz=3 | ||
44 | SQDMULLT_zzx_s 01000100 10 1 ..... 1110.1 ..... ..... @rrx_3a esz=2 | ||
45 | SQDMULLT_zzx_d 01000100 11 1 ..... 1110.1 ..... ..... @rrx_2a esz=3 | ||
46 | |||
47 | +# SVE2 saturating multiply high (indexed) | ||
48 | +SQDMULH_zzx_h 01000100 0. 1 ..... 111100 ..... ..... @rrx_3 esz=1 | ||
49 | +SQDMULH_zzx_s 01000100 10 1 ..... 111100 ..... ..... @rrx_2 esz=2 | ||
50 | +SQDMULH_zzx_d 01000100 11 1 ..... 111100 ..... ..... @rrx_1 esz=3 | ||
51 | +SQRDMULH_zzx_h 01000100 0. 1 ..... 111101 ..... ..... @rrx_3 esz=1 | ||
52 | +SQRDMULH_zzx_s 01000100 10 1 ..... 111101 ..... ..... @rrx_2 esz=2 | ||
53 | +SQRDMULH_zzx_d 01000100 11 1 ..... 111101 ..... ..... @rrx_1 esz=3 | ||
54 | + | ||
55 | # SVE2 integer multiply (indexed) | ||
56 | MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
57 | MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
58 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/target/arm/translate-sve.c | ||
61 | +++ b/target/arm/translate-sve.c | ||
62 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRX(trans_MUL_zzx_h, gen_helper_gvec_mul_idx_h) | ||
63 | DO_SVE2_RRX(trans_MUL_zzx_s, gen_helper_gvec_mul_idx_s) | ||
64 | DO_SVE2_RRX(trans_MUL_zzx_d, gen_helper_gvec_mul_idx_d) | ||
65 | |||
66 | +DO_SVE2_RRX(trans_SQDMULH_zzx_h, gen_helper_sve2_sqdmulh_idx_h) | ||
67 | +DO_SVE2_RRX(trans_SQDMULH_zzx_s, gen_helper_sve2_sqdmulh_idx_s) | ||
68 | +DO_SVE2_RRX(trans_SQDMULH_zzx_d, gen_helper_sve2_sqdmulh_idx_d) | ||
69 | + | ||
70 | +DO_SVE2_RRX(trans_SQRDMULH_zzx_h, gen_helper_sve2_sqrdmulh_idx_h) | ||
71 | +DO_SVE2_RRX(trans_SQRDMULH_zzx_s, gen_helper_sve2_sqrdmulh_idx_s) | ||
72 | +DO_SVE2_RRX(trans_SQRDMULH_zzx_d, gen_helper_sve2_sqrdmulh_idx_d) | ||
73 | + | ||
74 | #undef DO_SVE2_RRX | ||
75 | |||
76 | #define DO_SVE2_RRX_TB(NAME, FUNC, TOP) \ | ||
77 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
78 | index XXXXXXX..XXXXXXX 100644 | ||
79 | --- a/target/arm/vec_helper.c | ||
80 | +++ b/target/arm/vec_helper.c | ||
81 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmulh_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
82 | } | 28 | } |
83 | } | 29 | } |
84 | 30 | ||
85 | +void HELPER(sve2_sqdmulh_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | 31 | -typedef uint64_t IMOPFn(uint64_t, uint64_t, uint64_t, uint8_t, bool); |
32 | +typedef uint32_t IMOPFn32(uint32_t, uint32_t, uint32_t, uint8_t, bool); | ||
33 | +static inline void do_imopa_s(uint32_t *za, uint32_t *zn, uint32_t *zm, | ||
34 | + uint8_t *pn, uint8_t *pm, | ||
35 | + uint32_t desc, IMOPFn32 *fn) | ||
86 | +{ | 36 | +{ |
87 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | 37 | + intptr_t row, col, oprsz = simd_oprsz(desc) / 4; |
88 | + int idx = simd_data(desc); | 38 | + bool neg = simd_data(desc); |
89 | + int16_t *d = vd, *n = vn, *m = (int16_t *)vm + H2(idx); | 39 | |
90 | + uint32_t discard; | 40 | -static inline void do_imopa(uint64_t *za, uint64_t *zn, uint64_t *zm, |
91 | + | 41 | - uint8_t *pn, uint8_t *pm, |
92 | + for (i = 0; i < opr_sz / 2; i += 16 / 2) { | 42 | - uint32_t desc, IMOPFn *fn) |
93 | + int16_t mm = m[i]; | 43 | + for (row = 0; row < oprsz; ++row) { |
94 | + for (j = 0; j < 16 / 2; ++j) { | 44 | + uint8_t pa = (pn[H1(row >> 1)] >> ((row & 1) * 4)) & 0xf; |
95 | + d[i + j] = do_sqrdmlah_h(n[i + j], mm, 0, false, false, &discard); | 45 | + uint32_t *za_row = &za[tile_vslice_index(row)]; |
46 | + uint32_t n = zn[H4(row)]; | ||
47 | + | ||
48 | + for (col = 0; col < oprsz; ++col) { | ||
49 | + uint8_t pb = pm[H1(col >> 1)] >> ((col & 1) * 4); | ||
50 | + uint32_t *a = &za_row[H4(col)]; | ||
51 | + | ||
52 | + *a = fn(n, zm[H4(col)], *a, pa & pb, neg); | ||
96 | + } | 53 | + } |
97 | + } | 54 | + } |
98 | +} | 55 | +} |
99 | + | 56 | + |
100 | +void HELPER(sve2_sqrdmulh_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | 57 | +typedef uint64_t IMOPFn64(uint64_t, uint64_t, uint64_t, uint8_t, bool); |
58 | +static inline void do_imopa_d(uint64_t *za, uint64_t *zn, uint64_t *zm, | ||
59 | + uint8_t *pn, uint8_t *pm, | ||
60 | + uint32_t desc, IMOPFn64 *fn) | ||
61 | { | ||
62 | intptr_t row, col, oprsz = simd_oprsz(desc) / 8; | ||
63 | bool neg = simd_data(desc); | ||
64 | @@ -XXX,XX +XXX,XX @@ static inline void do_imopa(uint64_t *za, uint64_t *zn, uint64_t *zm, | ||
65 | } | ||
66 | |||
67 | #define DEF_IMOP_32(NAME, NTYPE, MTYPE) \ | ||
68 | -static uint64_t NAME(uint64_t n, uint64_t m, uint64_t a, uint8_t p, bool neg) \ | ||
69 | +static uint32_t NAME(uint32_t n, uint32_t m, uint32_t a, uint8_t p, bool neg) \ | ||
70 | { \ | ||
71 | - uint32_t sum0 = 0, sum1 = 0; \ | ||
72 | + uint32_t sum = 0; \ | ||
73 | /* Apply P to N as a mask, making the inactive elements 0. */ \ | ||
74 | n &= expand_pred_b(p); \ | ||
75 | - sum0 += (NTYPE)(n >> 0) * (MTYPE)(m >> 0); \ | ||
76 | - sum0 += (NTYPE)(n >> 8) * (MTYPE)(m >> 8); \ | ||
77 | - sum0 += (NTYPE)(n >> 16) * (MTYPE)(m >> 16); \ | ||
78 | - sum0 += (NTYPE)(n >> 24) * (MTYPE)(m >> 24); \ | ||
79 | - sum1 += (NTYPE)(n >> 32) * (MTYPE)(m >> 32); \ | ||
80 | - sum1 += (NTYPE)(n >> 40) * (MTYPE)(m >> 40); \ | ||
81 | - sum1 += (NTYPE)(n >> 48) * (MTYPE)(m >> 48); \ | ||
82 | - sum1 += (NTYPE)(n >> 56) * (MTYPE)(m >> 56); \ | ||
83 | - if (neg) { \ | ||
84 | - sum0 = (uint32_t)a - sum0, sum1 = (uint32_t)(a >> 32) - sum1; \ | ||
85 | - } else { \ | ||
86 | - sum0 = (uint32_t)a + sum0, sum1 = (uint32_t)(a >> 32) + sum1; \ | ||
87 | - } \ | ||
88 | - return ((uint64_t)sum1 << 32) | sum0; \ | ||
89 | + sum += (NTYPE)(n >> 0) * (MTYPE)(m >> 0); \ | ||
90 | + sum += (NTYPE)(n >> 8) * (MTYPE)(m >> 8); \ | ||
91 | + sum += (NTYPE)(n >> 16) * (MTYPE)(m >> 16); \ | ||
92 | + sum += (NTYPE)(n >> 24) * (MTYPE)(m >> 24); \ | ||
93 | + return neg ? a - sum : a + sum; \ | ||
94 | } | ||
95 | |||
96 | #define DEF_IMOP_64(NAME, NTYPE, MTYPE) \ | ||
97 | @@ -XXX,XX +XXX,XX @@ DEF_IMOP_64(umopa_d, uint16_t, uint16_t) | ||
98 | DEF_IMOP_64(sumopa_d, int16_t, uint16_t) | ||
99 | DEF_IMOP_64(usmopa_d, uint16_t, int16_t) | ||
100 | |||
101 | -#define DEF_IMOPH(NAME) \ | ||
102 | - void HELPER(sme_##NAME)(void *vza, void *vzn, void *vzm, void *vpn, \ | ||
103 | - void *vpm, uint32_t desc) \ | ||
104 | - { do_imopa(vza, vzn, vzm, vpn, vpm, desc, NAME); } | ||
105 | +#define DEF_IMOPH(NAME, S) \ | ||
106 | + void HELPER(sme_##NAME##_##S)(void *vza, void *vzn, void *vzm, \ | ||
107 | + void *vpn, void *vpm, uint32_t desc) \ | ||
108 | + { do_imopa_##S(vza, vzn, vzm, vpn, vpm, desc, NAME##_##S); } | ||
109 | |||
110 | -DEF_IMOPH(smopa_s) | ||
111 | -DEF_IMOPH(umopa_s) | ||
112 | -DEF_IMOPH(sumopa_s) | ||
113 | -DEF_IMOPH(usmopa_s) | ||
114 | -DEF_IMOPH(smopa_d) | ||
115 | -DEF_IMOPH(umopa_d) | ||
116 | -DEF_IMOPH(sumopa_d) | ||
117 | -DEF_IMOPH(usmopa_d) | ||
118 | +DEF_IMOPH(smopa, s) | ||
119 | +DEF_IMOPH(umopa, s) | ||
120 | +DEF_IMOPH(sumopa, s) | ||
121 | +DEF_IMOPH(usmopa, s) | ||
122 | + | ||
123 | +DEF_IMOPH(smopa, d) | ||
124 | +DEF_IMOPH(umopa, d) | ||
125 | +DEF_IMOPH(sumopa, d) | ||
126 | +DEF_IMOPH(usmopa, d) | ||
127 | diff --git a/tests/tcg/aarch64/sme-smopa-1.c b/tests/tcg/aarch64/sme-smopa-1.c | ||
128 | new file mode 100644 | ||
129 | index XXXXXXX..XXXXXXX | ||
130 | --- /dev/null | ||
131 | +++ b/tests/tcg/aarch64/sme-smopa-1.c | ||
132 | @@ -XXX,XX +XXX,XX @@ | ||
133 | +#include <stdio.h> | ||
134 | +#include <string.h> | ||
135 | + | ||
136 | +int main() | ||
101 | +{ | 137 | +{ |
102 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | 138 | + static const int cmp[4][4] = { |
103 | + int idx = simd_data(desc); | 139 | + { 110, 134, 158, 182 }, |
104 | + int16_t *d = vd, *n = vn, *m = (int16_t *)vm + H2(idx); | 140 | + { 390, 478, 566, 654 }, |
105 | + uint32_t discard; | 141 | + { 670, 822, 974, 1126 }, |
106 | + | 142 | + { 950, 1166, 1382, 1598 } |
107 | + for (i = 0; i < opr_sz / 2; i += 16 / 2) { | 143 | + }; |
108 | + int16_t mm = m[i]; | 144 | + int dst[4][4]; |
109 | + for (j = 0; j < 16 / 2; ++j) { | 145 | + int *tmp = &dst[0][0]; |
110 | + d[i + j] = do_sqrdmlah_h(n[i + j], mm, 0, false, true, &discard); | 146 | + |
147 | + asm volatile( | ||
148 | + ".arch armv8-r+sme\n\t" | ||
149 | + "smstart\n\t" | ||
150 | + "index z0.b, #0, #1\n\t" | ||
151 | + "movprfx z1, z0\n\t" | ||
152 | + "add z1.b, z1.b, #16\n\t" | ||
153 | + "ptrue p0.b\n\t" | ||
154 | + "smopa za0.s, p0/m, p0/m, z0.b, z1.b\n\t" | ||
155 | + "ptrue p0.s, vl4\n\t" | ||
156 | + "mov w12, #0\n\t" | ||
157 | + "st1w { za0h.s[w12, #0] }, p0, [%0]\n\t" | ||
158 | + "add %0, %0, #16\n\t" | ||
159 | + "st1w { za0h.s[w12, #1] }, p0, [%0]\n\t" | ||
160 | + "add %0, %0, #16\n\t" | ||
161 | + "st1w { za0h.s[w12, #2] }, p0, [%0]\n\t" | ||
162 | + "add %0, %0, #16\n\t" | ||
163 | + "st1w { za0h.s[w12, #3] }, p0, [%0]\n\t" | ||
164 | + "smstop" | ||
165 | + : "+r"(tmp) : : "memory"); | ||
166 | + | ||
167 | + if (memcmp(cmp, dst, sizeof(dst)) == 0) { | ||
168 | + return 0; | ||
169 | + } | ||
170 | + | ||
171 | + /* See above for correct results. */ | ||
172 | + for (int i = 0; i < 4; ++i) { | ||
173 | + for (int j = 0; j < 4; ++j) { | ||
174 | + printf("%6d", dst[i][j]); | ||
111 | + } | 175 | + } |
112 | + } | 176 | + printf("\n"); |
177 | + } | ||
178 | + return 1; | ||
113 | +} | 179 | +} |
114 | + | 180 | diff --git a/tests/tcg/aarch64/sme-smopa-2.c b/tests/tcg/aarch64/sme-smopa-2.c |
115 | /* Signed saturating rounding doubling multiply-accumulate high half, 32-bit */ | 181 | new file mode 100644 |
116 | int32_t do_sqrdmlah_s(int32_t src1, int32_t src2, int32_t src3, | 182 | index XXXXXXX..XXXXXXX |
117 | bool neg, bool round, uint32_t *sat) | 183 | --- /dev/null |
118 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmulh_s)(void *vd, void *vn, void *vm, uint32_t desc) | 184 | +++ b/tests/tcg/aarch64/sme-smopa-2.c |
119 | } | 185 | @@ -XXX,XX +XXX,XX @@ |
120 | } | 186 | +#include <stdio.h> |
121 | 187 | +#include <string.h> | |
122 | +void HELPER(sve2_sqdmulh_idx_s)(void *vd, void *vn, void *vm, uint32_t desc) | 188 | + |
189 | +int main() | ||
123 | +{ | 190 | +{ |
124 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | 191 | + static const long cmp[4][4] = { |
125 | + int idx = simd_data(desc); | 192 | + { 110, 134, 158, 182 }, |
126 | + int32_t *d = vd, *n = vn, *m = (int32_t *)vm + H4(idx); | 193 | + { 390, 478, 566, 654 }, |
127 | + uint32_t discard; | 194 | + { 670, 822, 974, 1126 }, |
128 | + | 195 | + { 950, 1166, 1382, 1598 } |
129 | + for (i = 0; i < opr_sz / 4; i += 16 / 4) { | 196 | + }; |
130 | + int32_t mm = m[i]; | 197 | + long dst[4][4]; |
131 | + for (j = 0; j < 16 / 4; ++j) { | 198 | + long *tmp = &dst[0][0]; |
132 | + d[i + j] = do_sqrdmlah_s(n[i + j], mm, 0, false, false, &discard); | 199 | + long svl; |
200 | + | ||
201 | + /* Validate that we have a wide enough vector for 4 elements. */ | ||
202 | + asm(".arch armv8-r+sme-i64\n\trdsvl %0, #1" : "=r"(svl)); | ||
203 | + if (svl < 32) { | ||
204 | + return 0; | ||
205 | + } | ||
206 | + | ||
207 | + asm volatile( | ||
208 | + "smstart\n\t" | ||
209 | + "index z0.h, #0, #1\n\t" | ||
210 | + "movprfx z1, z0\n\t" | ||
211 | + "add z1.h, z1.h, #16\n\t" | ||
212 | + "ptrue p0.b\n\t" | ||
213 | + "smopa za0.d, p0/m, p0/m, z0.h, z1.h\n\t" | ||
214 | + "ptrue p0.d, vl4\n\t" | ||
215 | + "mov w12, #0\n\t" | ||
216 | + "st1d { za0h.d[w12, #0] }, p0, [%0]\n\t" | ||
217 | + "add %0, %0, #32\n\t" | ||
218 | + "st1d { za0h.d[w12, #1] }, p0, [%0]\n\t" | ||
219 | + "mov w12, #2\n\t" | ||
220 | + "add %0, %0, #32\n\t" | ||
221 | + "st1d { za0h.d[w12, #0] }, p0, [%0]\n\t" | ||
222 | + "add %0, %0, #32\n\t" | ||
223 | + "st1d { za0h.d[w12, #1] }, p0, [%0]\n\t" | ||
224 | + "smstop" | ||
225 | + : "+r"(tmp) : : "memory"); | ||
226 | + | ||
227 | + if (memcmp(cmp, dst, sizeof(dst)) == 0) { | ||
228 | + return 0; | ||
229 | + } | ||
230 | + | ||
231 | + /* See above for correct results. */ | ||
232 | + for (int i = 0; i < 4; ++i) { | ||
233 | + for (int j = 0; j < 4; ++j) { | ||
234 | + printf("%6ld", dst[i][j]); | ||
133 | + } | 235 | + } |
134 | + } | 236 | + printf("\n"); |
237 | + } | ||
238 | + return 1; | ||
135 | +} | 239 | +} |
136 | + | 240 | diff --git a/tests/tcg/aarch64/Makefile.target b/tests/tcg/aarch64/Makefile.target |
137 | +void HELPER(sve2_sqrdmulh_idx_s)(void *vd, void *vn, void *vm, uint32_t desc) | 241 | index XXXXXXX..XXXXXXX 100644 |
138 | +{ | 242 | --- a/tests/tcg/aarch64/Makefile.target |
139 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | 243 | +++ b/tests/tcg/aarch64/Makefile.target |
140 | + int idx = simd_data(desc); | 244 | @@ -XXX,XX +XXX,XX @@ endif |
141 | + int32_t *d = vd, *n = vn, *m = (int32_t *)vm + H4(idx); | 245 | |
142 | + uint32_t discard; | 246 | # SME Tests |
143 | + | 247 | ifneq ($(CROSS_AS_HAS_ARMV9_SME),) |
144 | + for (i = 0; i < opr_sz / 4; i += 16 / 4) { | 248 | -AARCH64_TESTS += sme-outprod1 |
145 | + int32_t mm = m[i]; | 249 | +AARCH64_TESTS += sme-outprod1 sme-smopa-1 sme-smopa-2 |
146 | + for (j = 0; j < 16 / 4; ++j) { | 250 | endif |
147 | + d[i + j] = do_sqrdmlah_s(n[i + j], mm, 0, false, true, &discard); | 251 | |
148 | + } | 252 | # System Registers Tests |
149 | + } | ||
150 | +} | ||
151 | + | ||
152 | /* Signed saturating rounding doubling multiply-accumulate high half, 64-bit */ | ||
153 | static int64_t do_sat128_d(Int128 r) | ||
154 | { | ||
155 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
156 | } | ||
157 | } | ||
158 | |||
159 | +void HELPER(sve2_sqdmulh_idx_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
160 | +{ | ||
161 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | ||
162 | + int idx = simd_data(desc); | ||
163 | + int64_t *d = vd, *n = vn, *m = (int64_t *)vm + idx; | ||
164 | + | ||
165 | + for (i = 0; i < opr_sz / 8; i += 16 / 8) { | ||
166 | + int64_t mm = m[i]; | ||
167 | + for (j = 0; j < 16 / 8; ++j) { | ||
168 | + d[i + j] = do_sqrdmlah_d(n[i + j], mm, 0, false, false); | ||
169 | + } | ||
170 | + } | ||
171 | +} | ||
172 | + | ||
173 | +void HELPER(sve2_sqrdmulh_idx_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
174 | +{ | ||
175 | + intptr_t i, j, opr_sz = simd_oprsz(desc); | ||
176 | + int idx = simd_data(desc); | ||
177 | + int64_t *d = vd, *n = vn, *m = (int64_t *)vm + idx; | ||
178 | + | ||
179 | + for (i = 0; i < opr_sz / 8; i += 16 / 8) { | ||
180 | + int64_t mm = m[i]; | ||
181 | + for (j = 0; j < 16 / 8; ++j) { | ||
182 | + d[i + j] = do_sqrdmlah_d(n[i + j], mm, 0, false, true); | ||
183 | + } | ||
184 | + } | ||
185 | +} | ||
186 | + | ||
187 | /* Integer 8 and 16-bit dot-product. | ||
188 | * | ||
189 | * Note that for the loops herein, host endianness does not matter | ||
190 | -- | 253 | -- |
191 | 2.20.1 | 254 | 2.34.1 |
192 | 255 | ||
193 | 256 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | The sun4v RTC device model added under commit a0e893039cf2ce0 in 2016 |
---|---|---|---|
2 | was unfortunately added with a license of GPL-v3-or-later, which is | ||
3 | not compatible with other QEMU code which has a GPL-v2-only license. | ||
2 | 4 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 5 | Relicense the code in the .c and the .h file to GPL-v2-or-later, |
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 6 | to make it compatible with the rest of QEMU. |
5 | Message-id: 20210525010358.152808-68-richard.henderson@linaro.org | 7 | |
8 | Cc: qemu-stable@nongnu.org | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | Signed-off-by: Paolo Bonzini (for Red Hat) <pbonzini@redhat.com> | ||
11 | Signed-off-by: Artyom Tarasenko <atar4qemu@gmail.com> | ||
12 | Signed-off-by: Markus Armbruster <armbru@redhat.com> | ||
13 | Signed-off-by: Alex Bennée <alex.bennee@linaro.org> | ||
14 | Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
15 | Signed-off-by: Daniel P. Berrangé <berrange@redhat.com> | ||
16 | Acked-by: Alex Bennée <alex.bennee@linaro.org> | ||
17 | Message-id: 20240223161300.938542-1-peter.maydell@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 18 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
7 | --- | 19 | --- |
8 | target/arm/helper.h | 1 + | 20 | include/hw/rtc/sun4v-rtc.h | 2 +- |
9 | target/arm/sve.decode | 4 ++++ | 21 | hw/rtc/sun4v-rtc.c | 2 +- |
10 | target/arm/translate-sve.c | 16 ++++++++++++++++ | 22 | 2 files changed, 2 insertions(+), 2 deletions(-) |
11 | target/arm/vec_helper.c | 1 + | ||
12 | 4 files changed, 22 insertions(+) | ||
13 | 23 | ||
14 | diff --git a/target/arm/helper.h b/target/arm/helper.h | 24 | diff --git a/include/hw/rtc/sun4v-rtc.h b/include/hw/rtc/sun4v-rtc.h |
15 | index XXXXXXX..XXXXXXX 100644 | 25 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/target/arm/helper.h | 26 | --- a/include/hw/rtc/sun4v-rtc.h |
17 | +++ b/target/arm/helper.h | 27 | +++ b/include/hw/rtc/sun4v-rtc.h |
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(gvec_sdot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | 28 | @@ -XXX,XX +XXX,XX @@ |
19 | DEF_HELPER_FLAGS_5(gvec_udot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | 29 | * |
20 | DEF_HELPER_FLAGS_5(gvec_sdot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | 30 | * Copyright (c) 2016 Artyom Tarasenko |
21 | DEF_HELPER_FLAGS_5(gvec_udot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | 31 | * |
22 | +DEF_HELPER_FLAGS_5(gvec_usdot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | 32 | - * This code is licensed under the GNU GPL v3 or (at your option) any later |
23 | 33 | + * This code is licensed under the GNU GPL v2 or (at your option) any later | |
24 | DEF_HELPER_FLAGS_5(gvec_sdot_idx_b, TCG_CALL_NO_RWG, | 34 | * version. |
25 | void, ptr, ptr, ptr, ptr, i32) | 35 | */ |
26 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | 36 | |
37 | diff --git a/hw/rtc/sun4v-rtc.c b/hw/rtc/sun4v-rtc.c | ||
27 | index XXXXXXX..XXXXXXX 100644 | 38 | index XXXXXXX..XXXXXXX 100644 |
28 | --- a/target/arm/sve.decode | 39 | --- a/hw/rtc/sun4v-rtc.c |
29 | +++ b/target/arm/sve.decode | 40 | +++ b/hw/rtc/sun4v-rtc.c |
30 | @@ -XXX,XX +XXX,XX @@ UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | 41 | @@ -XXX,XX +XXX,XX @@ |
31 | CMLA_zzzz 01000100 esz:2 0 rm:5 0010 rot:2 rn:5 rd:5 ra=%reg_movprfx | 42 | * |
32 | SQRDCMLAH_zzzz 01000100 esz:2 0 rm:5 0011 rot:2 rn:5 rd:5 ra=%reg_movprfx | 43 | * Copyright (c) 2016 Artyom Tarasenko |
33 | 44 | * | |
34 | +## SVE mixed sign dot product | 45 | - * This code is licensed under the GNU GPL v3 or (at your option) any later |
35 | + | 46 | + * This code is licensed under the GNU GPL v2 or (at your option) any later |
36 | +USDOT_zzzz 01000100 .. 0 ..... 011 110 ..... ..... @rda_rn_rm | 47 | * version. |
37 | + | 48 | */ |
38 | ### SVE2 floating point matrix multiply accumulate | ||
39 | |||
40 | FMMLA 01100100 .. 1 ..... 111001 ..... ..... @rda_rn_rm | ||
41 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/target/arm/translate-sve.c | ||
44 | +++ b/target/arm/translate-sve.c | ||
45 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQRDCMLAH_zzzz(DisasContext *s, arg_SQRDCMLAH_zzzz *a) | ||
46 | } | ||
47 | return true; | ||
48 | } | ||
49 | + | ||
50 | +static bool trans_USDOT_zzzz(DisasContext *s, arg_USDOT_zzzz *a) | ||
51 | +{ | ||
52 | + if (a->esz != 2 || !dc_isar_feature(aa64_sve_i8mm, s)) { | ||
53 | + return false; | ||
54 | + } | ||
55 | + if (sve_access_check(s)) { | ||
56 | + unsigned vsz = vec_full_reg_size(s); | ||
57 | + tcg_gen_gvec_4_ool(vec_full_reg_offset(s, a->rd), | ||
58 | + vec_full_reg_offset(s, a->rn), | ||
59 | + vec_full_reg_offset(s, a->rm), | ||
60 | + vec_full_reg_offset(s, a->ra), | ||
61 | + vsz, vsz, 0, gen_helper_gvec_usdot_b); | ||
62 | + } | ||
63 | + return true; | ||
64 | +} | ||
65 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
66 | index XXXXXXX..XXXXXXX 100644 | ||
67 | --- a/target/arm/vec_helper.c | ||
68 | +++ b/target/arm/vec_helper.c | ||
69 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
70 | |||
71 | DO_DOT(gvec_sdot_b, int32_t, int8_t, int8_t) | ||
72 | DO_DOT(gvec_udot_b, uint32_t, uint8_t, uint8_t) | ||
73 | +DO_DOT(gvec_usdot_b, uint32_t, uint8_t, int8_t) | ||
74 | DO_DOT(gvec_sdot_h, int64_t, int16_t, int16_t) | ||
75 | DO_DOT(gvec_udot_h, uint64_t, uint16_t, uint16_t) | ||
76 | 49 | ||
77 | -- | 50 | -- |
78 | 2.20.1 | 51 | 2.34.1 |
79 | 52 | ||
80 | 53 | diff view generated by jsdifflib |
1 | From: Richard Henderson <richard.henderson@linaro.org> | 1 | From: Thomas Huth <thuth@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | SVE2 has two additional sizes of the operation and unlike NEON, | 3 | Move the code to a separate file so that we do not have to compile |
4 | there is no saturation flag. Create new entry points for SVE2 | 4 | it anymore if CONFIG_ARM_V7M is not set. |
5 | that do not set QC. | ||
6 | 5 | ||
6 | Signed-off-by: Thomas Huth <thuth@redhat.com> | ||
7 | Message-id: 20240308141051.536599-2-thuth@redhat.com | ||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | 8 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> |
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-36-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | 9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> |
11 | --- | 10 | --- |
12 | target/arm/helper.h | 17 ++++ | 11 | target/arm/tcg/cpu-v7m.c | 290 +++++++++++++++++++++++++++++++++++++ |
13 | target/arm/sve.decode | 5 ++ | 12 | target/arm/tcg/cpu32.c | 261 --------------------------------- |
14 | target/arm/translate-sve.c | 18 +++++ | 13 | target/arm/meson.build | 3 + |
15 | target/arm/vec_helper.c | 161 +++++++++++++++++++++++++++++++++++-- | 14 | target/arm/tcg/meson.build | 3 + |
16 | 4 files changed, 195 insertions(+), 6 deletions(-) | 15 | 4 files changed, 296 insertions(+), 261 deletions(-) |
16 | create mode 100644 target/arm/tcg/cpu-v7m.c | ||
17 | 17 | ||
18 | diff --git a/target/arm/helper.h b/target/arm/helper.h | 18 | diff --git a/target/arm/tcg/cpu-v7m.c b/target/arm/tcg/cpu-v7m.c |
19 | index XXXXXXX..XXXXXXX 100644 | 19 | new file mode 100644 |
20 | --- a/target/arm/helper.h | 20 | index XXXXXXX..XXXXXXX |
21 | +++ b/target/arm/helper.h | 21 | --- /dev/null |
22 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(gvec_qrdmlah_s32, TCG_CALL_NO_RWG, | 22 | +++ b/target/arm/tcg/cpu-v7m.c |
23 | DEF_HELPER_FLAGS_5(gvec_qrdmlsh_s32, TCG_CALL_NO_RWG, | ||
24 | void, ptr, ptr, ptr, ptr, i32) | ||
25 | |||
26 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_b, TCG_CALL_NO_RWG, | ||
27 | + void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_b, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_h, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_h, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_s, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_s, TCG_CALL_NO_RWG, | ||
37 | + void, ptr, ptr, ptr, ptr, i32) | ||
38 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_d, TCG_CALL_NO_RWG, | ||
39 | + void, ptr, ptr, ptr, ptr, i32) | ||
40 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_d, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | + | ||
43 | DEF_HELPER_FLAGS_4(gvec_sdot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
44 | DEF_HELPER_FLAGS_4(gvec_udot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
45 | DEF_HELPER_FLAGS_4(gvec_sdot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
46 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
47 | index XXXXXXX..XXXXXXX 100644 | ||
48 | --- a/target/arm/sve.decode | ||
49 | +++ b/target/arm/sve.decode | ||
50 | @@ -XXX,XX +XXX,XX @@ SQDMLSLT_zzzw 01000100 .. 0 ..... 0110 11 ..... ..... @rda_rn_rm | ||
51 | |||
52 | SQDMLALBT 01000100 .. 0 ..... 00001 0 ..... ..... @rda_rn_rm | ||
53 | SQDMLSLBT 01000100 .. 0 ..... 00001 1 ..... ..... @rda_rn_rm | ||
54 | + | ||
55 | +## SVE2 saturating multiply-add high | ||
56 | + | ||
57 | +SQRDMLAH_zzzz 01000100 .. 0 ..... 01110 0 ..... ..... @rda_rn_rm | ||
58 | +SQRDMLSH_zzzz 01000100 .. 0 ..... 01110 1 ..... ..... @rda_rn_rm | ||
59 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
60 | index XXXXXXX..XXXXXXX 100644 | ||
61 | --- a/target/arm/translate-sve.c | ||
62 | +++ b/target/arm/translate-sve.c | ||
63 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQDMLSLBT(DisasContext *s, arg_rrrr_esz *a) | ||
64 | { | ||
65 | return do_sqdmlsl_zzzw(s, a, false, true); | ||
66 | } | ||
67 | + | ||
68 | +static bool trans_SQRDMLAH_zzzz(DisasContext *s, arg_rrrr_esz *a) | ||
69 | +{ | ||
70 | + static gen_helper_gvec_4 * const fns[] = { | ||
71 | + gen_helper_sve2_sqrdmlah_b, gen_helper_sve2_sqrdmlah_h, | ||
72 | + gen_helper_sve2_sqrdmlah_s, gen_helper_sve2_sqrdmlah_d, | ||
73 | + }; | ||
74 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], 0); | ||
75 | +} | ||
76 | + | ||
77 | +static bool trans_SQRDMLSH_zzzz(DisasContext *s, arg_rrrr_esz *a) | ||
78 | +{ | ||
79 | + static gen_helper_gvec_4 * const fns[] = { | ||
80 | + gen_helper_sve2_sqrdmlsh_b, gen_helper_sve2_sqrdmlsh_h, | ||
81 | + gen_helper_sve2_sqrdmlsh_s, gen_helper_sve2_sqrdmlsh_d, | ||
82 | + }; | ||
83 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], 0); | ||
84 | +} | ||
85 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
86 | index XXXXXXX..XXXXXXX 100644 | ||
87 | --- a/target/arm/vec_helper.c | ||
88 | +++ b/target/arm/vec_helper.c | ||
89 | @@ -XXX,XX +XXX,XX @@ | 23 | @@ -XXX,XX +XXX,XX @@ |
90 | #include "exec/helper-proto.h" | 24 | +/* |
91 | #include "tcg/tcg-gvec-desc.h" | 25 | + * QEMU ARMv7-M TCG-only CPUs. |
92 | #include "fpu/softfloat.h" | 26 | + * |
93 | +#include "qemu/int128.h" | 27 | + * Copyright (c) 2012 SUSE LINUX Products GmbH |
94 | #include "vec_internal.h" | 28 | + * |
95 | 29 | + * This code is licensed under the GNU GPL v2 or later. | |
96 | /* Note that vector data is stored in host-endian 64-bit chunks, | 30 | + * |
97 | @@ -XXX,XX +XXX,XX @@ | 31 | + * SPDX-License-Identifier: GPL-2.0-or-later |
98 | #define H4(x) (x) | 32 | + */ |
99 | #endif | 33 | + |
100 | 34 | +#include "qemu/osdep.h" | |
101 | +/* Signed saturating rounding doubling multiply-accumulate high half, 8-bit */ | 35 | +#include "cpu.h" |
102 | +static int8_t do_sqrdmlah_b(int8_t src1, int8_t src2, int8_t src3, | 36 | +#include "hw/core/tcg-cpu-ops.h" |
103 | + bool neg, bool round) | 37 | +#include "internals.h" |
104 | +{ | 38 | + |
39 | +#if !defined(CONFIG_USER_ONLY) | ||
40 | + | ||
41 | +#include "hw/intc/armv7m_nvic.h" | ||
42 | + | ||
43 | +static bool arm_v7m_cpu_exec_interrupt(CPUState *cs, int interrupt_request) | ||
44 | +{ | ||
45 | + CPUClass *cc = CPU_GET_CLASS(cs); | ||
46 | + ARMCPU *cpu = ARM_CPU(cs); | ||
47 | + CPUARMState *env = &cpu->env; | ||
48 | + bool ret = false; | ||
49 | + | ||
105 | + /* | 50 | + /* |
106 | + * Simplify: | 51 | + * ARMv7-M interrupt masking works differently than -A or -R. |
107 | + * = ((a3 << 8) + ((e1 * e2) << 1) + (round << 7)) >> 8 | 52 | + * There is no FIQ/IRQ distinction. Instead of I and F bits |
108 | + * = ((a3 << 7) + (e1 * e2) + (round << 6)) >> 7 | 53 | + * masking FIQ and IRQ interrupts, an exception is taken only |
54 | + * if it is higher priority than the current execution priority | ||
55 | + * (which depends on state like BASEPRI, FAULTMASK and the | ||
56 | + * currently active exception). | ||
109 | + */ | 57 | + */ |
110 | + int32_t ret = (int32_t)src1 * src2; | 58 | + if (interrupt_request & CPU_INTERRUPT_HARD |
111 | + if (neg) { | 59 | + && (armv7m_nvic_can_take_pending_exception(env->nvic))) { |
112 | + ret = -ret; | 60 | + cs->exception_index = EXCP_IRQ; |
113 | + } | 61 | + cc->tcg_ops->do_interrupt(cs); |
114 | + ret += ((int32_t)src3 << 7) + (round << 6); | 62 | + ret = true; |
115 | + ret >>= 7; | ||
116 | + | ||
117 | + if (ret != (int8_t)ret) { | ||
118 | + ret = (ret < 0 ? INT8_MIN : INT8_MAX); | ||
119 | + } | 63 | + } |
120 | + return ret; | 64 | + return ret; |
121 | +} | 65 | +} |
122 | + | 66 | + |
123 | +void HELPER(sve2_sqrdmlah_b)(void *vd, void *vn, void *vm, | 67 | +#endif /* !CONFIG_USER_ONLY */ |
124 | + void *va, uint32_t desc) | 68 | + |
125 | +{ | 69 | +static void cortex_m0_initfn(Object *obj) |
126 | + intptr_t i, opr_sz = simd_oprsz(desc); | 70 | +{ |
127 | + int8_t *d = vd, *n = vn, *m = vm, *a = va; | 71 | + ARMCPU *cpu = ARM_CPU(obj); |
128 | + | 72 | + set_feature(&cpu->env, ARM_FEATURE_V6); |
129 | + for (i = 0; i < opr_sz; ++i) { | 73 | + set_feature(&cpu->env, ARM_FEATURE_M); |
130 | + d[i] = do_sqrdmlah_b(n[i], m[i], a[i], false, true); | 74 | + |
75 | + cpu->midr = 0x410cc200; | ||
76 | + | ||
77 | + /* | ||
78 | + * These ID register values are not guest visible, because | ||
79 | + * we do not implement the Main Extension. They must be set | ||
80 | + * to values corresponding to the Cortex-M0's implemented | ||
81 | + * features, because QEMU generally controls its emulation | ||
82 | + * by looking at ID register fields. We use the same values as | ||
83 | + * for the M3. | ||
84 | + */ | ||
85 | + cpu->isar.id_pfr0 = 0x00000030; | ||
86 | + cpu->isar.id_pfr1 = 0x00000200; | ||
87 | + cpu->isar.id_dfr0 = 0x00100000; | ||
88 | + cpu->id_afr0 = 0x00000000; | ||
89 | + cpu->isar.id_mmfr0 = 0x00000030; | ||
90 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
91 | + cpu->isar.id_mmfr2 = 0x00000000; | ||
92 | + cpu->isar.id_mmfr3 = 0x00000000; | ||
93 | + cpu->isar.id_isar0 = 0x01141110; | ||
94 | + cpu->isar.id_isar1 = 0x02111000; | ||
95 | + cpu->isar.id_isar2 = 0x21112231; | ||
96 | + cpu->isar.id_isar3 = 0x01111110; | ||
97 | + cpu->isar.id_isar4 = 0x01310102; | ||
98 | + cpu->isar.id_isar5 = 0x00000000; | ||
99 | + cpu->isar.id_isar6 = 0x00000000; | ||
100 | +} | ||
101 | + | ||
102 | +static void cortex_m3_initfn(Object *obj) | ||
103 | +{ | ||
104 | + ARMCPU *cpu = ARM_CPU(obj); | ||
105 | + set_feature(&cpu->env, ARM_FEATURE_V7); | ||
106 | + set_feature(&cpu->env, ARM_FEATURE_M); | ||
107 | + set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
108 | + cpu->midr = 0x410fc231; | ||
109 | + cpu->pmsav7_dregion = 8; | ||
110 | + cpu->isar.id_pfr0 = 0x00000030; | ||
111 | + cpu->isar.id_pfr1 = 0x00000200; | ||
112 | + cpu->isar.id_dfr0 = 0x00100000; | ||
113 | + cpu->id_afr0 = 0x00000000; | ||
114 | + cpu->isar.id_mmfr0 = 0x00000030; | ||
115 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
116 | + cpu->isar.id_mmfr2 = 0x00000000; | ||
117 | + cpu->isar.id_mmfr3 = 0x00000000; | ||
118 | + cpu->isar.id_isar0 = 0x01141110; | ||
119 | + cpu->isar.id_isar1 = 0x02111000; | ||
120 | + cpu->isar.id_isar2 = 0x21112231; | ||
121 | + cpu->isar.id_isar3 = 0x01111110; | ||
122 | + cpu->isar.id_isar4 = 0x01310102; | ||
123 | + cpu->isar.id_isar5 = 0x00000000; | ||
124 | + cpu->isar.id_isar6 = 0x00000000; | ||
125 | +} | ||
126 | + | ||
127 | +static void cortex_m4_initfn(Object *obj) | ||
128 | +{ | ||
129 | + ARMCPU *cpu = ARM_CPU(obj); | ||
130 | + | ||
131 | + set_feature(&cpu->env, ARM_FEATURE_V7); | ||
132 | + set_feature(&cpu->env, ARM_FEATURE_M); | ||
133 | + set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
134 | + set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
135 | + cpu->midr = 0x410fc240; /* r0p0 */ | ||
136 | + cpu->pmsav7_dregion = 8; | ||
137 | + cpu->isar.mvfr0 = 0x10110021; | ||
138 | + cpu->isar.mvfr1 = 0x11000011; | ||
139 | + cpu->isar.mvfr2 = 0x00000000; | ||
140 | + cpu->isar.id_pfr0 = 0x00000030; | ||
141 | + cpu->isar.id_pfr1 = 0x00000200; | ||
142 | + cpu->isar.id_dfr0 = 0x00100000; | ||
143 | + cpu->id_afr0 = 0x00000000; | ||
144 | + cpu->isar.id_mmfr0 = 0x00000030; | ||
145 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
146 | + cpu->isar.id_mmfr2 = 0x00000000; | ||
147 | + cpu->isar.id_mmfr3 = 0x00000000; | ||
148 | + cpu->isar.id_isar0 = 0x01141110; | ||
149 | + cpu->isar.id_isar1 = 0x02111000; | ||
150 | + cpu->isar.id_isar2 = 0x21112231; | ||
151 | + cpu->isar.id_isar3 = 0x01111110; | ||
152 | + cpu->isar.id_isar4 = 0x01310102; | ||
153 | + cpu->isar.id_isar5 = 0x00000000; | ||
154 | + cpu->isar.id_isar6 = 0x00000000; | ||
155 | +} | ||
156 | + | ||
157 | +static void cortex_m7_initfn(Object *obj) | ||
158 | +{ | ||
159 | + ARMCPU *cpu = ARM_CPU(obj); | ||
160 | + | ||
161 | + set_feature(&cpu->env, ARM_FEATURE_V7); | ||
162 | + set_feature(&cpu->env, ARM_FEATURE_M); | ||
163 | + set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
164 | + set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
165 | + cpu->midr = 0x411fc272; /* r1p2 */ | ||
166 | + cpu->pmsav7_dregion = 8; | ||
167 | + cpu->isar.mvfr0 = 0x10110221; | ||
168 | + cpu->isar.mvfr1 = 0x12000011; | ||
169 | + cpu->isar.mvfr2 = 0x00000040; | ||
170 | + cpu->isar.id_pfr0 = 0x00000030; | ||
171 | + cpu->isar.id_pfr1 = 0x00000200; | ||
172 | + cpu->isar.id_dfr0 = 0x00100000; | ||
173 | + cpu->id_afr0 = 0x00000000; | ||
174 | + cpu->isar.id_mmfr0 = 0x00100030; | ||
175 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
176 | + cpu->isar.id_mmfr2 = 0x01000000; | ||
177 | + cpu->isar.id_mmfr3 = 0x00000000; | ||
178 | + cpu->isar.id_isar0 = 0x01101110; | ||
179 | + cpu->isar.id_isar1 = 0x02112000; | ||
180 | + cpu->isar.id_isar2 = 0x20232231; | ||
181 | + cpu->isar.id_isar3 = 0x01111131; | ||
182 | + cpu->isar.id_isar4 = 0x01310132; | ||
183 | + cpu->isar.id_isar5 = 0x00000000; | ||
184 | + cpu->isar.id_isar6 = 0x00000000; | ||
185 | +} | ||
186 | + | ||
187 | +static void cortex_m33_initfn(Object *obj) | ||
188 | +{ | ||
189 | + ARMCPU *cpu = ARM_CPU(obj); | ||
190 | + | ||
191 | + set_feature(&cpu->env, ARM_FEATURE_V8); | ||
192 | + set_feature(&cpu->env, ARM_FEATURE_M); | ||
193 | + set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
194 | + set_feature(&cpu->env, ARM_FEATURE_M_SECURITY); | ||
195 | + set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
196 | + cpu->midr = 0x410fd213; /* r0p3 */ | ||
197 | + cpu->pmsav7_dregion = 16; | ||
198 | + cpu->sau_sregion = 8; | ||
199 | + cpu->isar.mvfr0 = 0x10110021; | ||
200 | + cpu->isar.mvfr1 = 0x11000011; | ||
201 | + cpu->isar.mvfr2 = 0x00000040; | ||
202 | + cpu->isar.id_pfr0 = 0x00000030; | ||
203 | + cpu->isar.id_pfr1 = 0x00000210; | ||
204 | + cpu->isar.id_dfr0 = 0x00200000; | ||
205 | + cpu->id_afr0 = 0x00000000; | ||
206 | + cpu->isar.id_mmfr0 = 0x00101F40; | ||
207 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
208 | + cpu->isar.id_mmfr2 = 0x01000000; | ||
209 | + cpu->isar.id_mmfr3 = 0x00000000; | ||
210 | + cpu->isar.id_isar0 = 0x01101110; | ||
211 | + cpu->isar.id_isar1 = 0x02212000; | ||
212 | + cpu->isar.id_isar2 = 0x20232232; | ||
213 | + cpu->isar.id_isar3 = 0x01111131; | ||
214 | + cpu->isar.id_isar4 = 0x01310132; | ||
215 | + cpu->isar.id_isar5 = 0x00000000; | ||
216 | + cpu->isar.id_isar6 = 0x00000000; | ||
217 | + cpu->clidr = 0x00000000; | ||
218 | + cpu->ctr = 0x8000c000; | ||
219 | +} | ||
220 | + | ||
221 | +static void cortex_m55_initfn(Object *obj) | ||
222 | +{ | ||
223 | + ARMCPU *cpu = ARM_CPU(obj); | ||
224 | + | ||
225 | + set_feature(&cpu->env, ARM_FEATURE_V8); | ||
226 | + set_feature(&cpu->env, ARM_FEATURE_V8_1M); | ||
227 | + set_feature(&cpu->env, ARM_FEATURE_M); | ||
228 | + set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
229 | + set_feature(&cpu->env, ARM_FEATURE_M_SECURITY); | ||
230 | + set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
231 | + cpu->midr = 0x410fd221; /* r0p1 */ | ||
232 | + cpu->revidr = 0; | ||
233 | + cpu->pmsav7_dregion = 16; | ||
234 | + cpu->sau_sregion = 8; | ||
235 | + /* These are the MVFR* values for the FPU + full MVE configuration */ | ||
236 | + cpu->isar.mvfr0 = 0x10110221; | ||
237 | + cpu->isar.mvfr1 = 0x12100211; | ||
238 | + cpu->isar.mvfr2 = 0x00000040; | ||
239 | + cpu->isar.id_pfr0 = 0x20000030; | ||
240 | + cpu->isar.id_pfr1 = 0x00000230; | ||
241 | + cpu->isar.id_dfr0 = 0x10200000; | ||
242 | + cpu->id_afr0 = 0x00000000; | ||
243 | + cpu->isar.id_mmfr0 = 0x00111040; | ||
244 | + cpu->isar.id_mmfr1 = 0x00000000; | ||
245 | + cpu->isar.id_mmfr2 = 0x01000000; | ||
246 | + cpu->isar.id_mmfr3 = 0x00000011; | ||
247 | + cpu->isar.id_isar0 = 0x01103110; | ||
248 | + cpu->isar.id_isar1 = 0x02212000; | ||
249 | + cpu->isar.id_isar2 = 0x20232232; | ||
250 | + cpu->isar.id_isar3 = 0x01111131; | ||
251 | + cpu->isar.id_isar4 = 0x01310132; | ||
252 | + cpu->isar.id_isar5 = 0x00000000; | ||
253 | + cpu->isar.id_isar6 = 0x00000000; | ||
254 | + cpu->clidr = 0x00000000; /* caches not implemented */ | ||
255 | + cpu->ctr = 0x8303c003; | ||
256 | +} | ||
257 | + | ||
258 | +static const TCGCPUOps arm_v7m_tcg_ops = { | ||
259 | + .initialize = arm_translate_init, | ||
260 | + .synchronize_from_tb = arm_cpu_synchronize_from_tb, | ||
261 | + .debug_excp_handler = arm_debug_excp_handler, | ||
262 | + .restore_state_to_opc = arm_restore_state_to_opc, | ||
263 | + | ||
264 | +#ifdef CONFIG_USER_ONLY | ||
265 | + .record_sigsegv = arm_cpu_record_sigsegv, | ||
266 | + .record_sigbus = arm_cpu_record_sigbus, | ||
267 | +#else | ||
268 | + .tlb_fill = arm_cpu_tlb_fill, | ||
269 | + .cpu_exec_interrupt = arm_v7m_cpu_exec_interrupt, | ||
270 | + .do_interrupt = arm_v7m_cpu_do_interrupt, | ||
271 | + .do_transaction_failed = arm_cpu_do_transaction_failed, | ||
272 | + .do_unaligned_access = arm_cpu_do_unaligned_access, | ||
273 | + .adjust_watchpoint_address = arm_adjust_watchpoint_address, | ||
274 | + .debug_check_watchpoint = arm_debug_check_watchpoint, | ||
275 | + .debug_check_breakpoint = arm_debug_check_breakpoint, | ||
276 | +#endif /* !CONFIG_USER_ONLY */ | ||
277 | +}; | ||
278 | + | ||
279 | +static void arm_v7m_class_init(ObjectClass *oc, void *data) | ||
280 | +{ | ||
281 | + ARMCPUClass *acc = ARM_CPU_CLASS(oc); | ||
282 | + CPUClass *cc = CPU_CLASS(oc); | ||
283 | + | ||
284 | + acc->info = data; | ||
285 | + cc->tcg_ops = &arm_v7m_tcg_ops; | ||
286 | + cc->gdb_core_xml_file = "arm-m-profile.xml"; | ||
287 | +} | ||
288 | + | ||
289 | +static const ARMCPUInfo arm_v7m_cpus[] = { | ||
290 | + { .name = "cortex-m0", .initfn = cortex_m0_initfn, | ||
291 | + .class_init = arm_v7m_class_init }, | ||
292 | + { .name = "cortex-m3", .initfn = cortex_m3_initfn, | ||
293 | + .class_init = arm_v7m_class_init }, | ||
294 | + { .name = "cortex-m4", .initfn = cortex_m4_initfn, | ||
295 | + .class_init = arm_v7m_class_init }, | ||
296 | + { .name = "cortex-m7", .initfn = cortex_m7_initfn, | ||
297 | + .class_init = arm_v7m_class_init }, | ||
298 | + { .name = "cortex-m33", .initfn = cortex_m33_initfn, | ||
299 | + .class_init = arm_v7m_class_init }, | ||
300 | + { .name = "cortex-m55", .initfn = cortex_m55_initfn, | ||
301 | + .class_init = arm_v7m_class_init }, | ||
302 | +}; | ||
303 | + | ||
304 | +static void arm_v7m_cpu_register_types(void) | ||
305 | +{ | ||
306 | + size_t i; | ||
307 | + | ||
308 | + for (i = 0; i < ARRAY_SIZE(arm_v7m_cpus); ++i) { | ||
309 | + arm_cpu_register(&arm_v7m_cpus[i]); | ||
131 | + } | 310 | + } |
132 | +} | 311 | +} |
133 | + | 312 | + |
134 | +void HELPER(sve2_sqrdmlsh_b)(void *vd, void *vn, void *vm, | 313 | +type_init(arm_v7m_cpu_register_types) |
135 | + void *va, uint32_t desc) | 314 | diff --git a/target/arm/tcg/cpu32.c b/target/arm/tcg/cpu32.c |
136 | +{ | 315 | index XXXXXXX..XXXXXXX 100644 |
137 | + intptr_t i, opr_sz = simd_oprsz(desc); | 316 | --- a/target/arm/tcg/cpu32.c |
138 | + int8_t *d = vd, *n = vn, *m = vm, *a = va; | 317 | +++ b/target/arm/tcg/cpu32.c |
139 | + | 318 | @@ -XXX,XX +XXX,XX @@ |
140 | + for (i = 0; i < opr_sz; ++i) { | 319 | #include "hw/boards.h" |
141 | + d[i] = do_sqrdmlah_b(n[i], m[i], a[i], true, true); | 320 | #endif |
142 | + } | 321 | #include "cpregs.h" |
143 | +} | 322 | -#if !defined(CONFIG_USER_ONLY) && defined(CONFIG_TCG) |
144 | + | 323 | -#include "hw/intc/armv7m_nvic.h" |
145 | /* Signed saturating rounding doubling multiply-accumulate high half, 16-bit */ | 324 | -#endif |
146 | static int16_t do_sqrdmlah_h(int16_t src1, int16_t src2, int16_t src3, | 325 | |
147 | bool neg, bool round, uint32_t *sat) | 326 | |
327 | /* Share AArch32 -cpu max features with AArch64. */ | ||
328 | @@ -XXX,XX +XXX,XX @@ void aa32_max_features(ARMCPU *cpu) | ||
329 | /* CPU models. These are not needed for the AArch64 linux-user build. */ | ||
330 | #if !defined(CONFIG_USER_ONLY) || !defined(TARGET_AARCH64) | ||
331 | |||
332 | -#if !defined(CONFIG_USER_ONLY) | ||
333 | -static bool arm_v7m_cpu_exec_interrupt(CPUState *cs, int interrupt_request) | ||
334 | -{ | ||
335 | - CPUClass *cc = CPU_GET_CLASS(cs); | ||
336 | - ARMCPU *cpu = ARM_CPU(cs); | ||
337 | - CPUARMState *env = &cpu->env; | ||
338 | - bool ret = false; | ||
339 | - | ||
340 | - /* | ||
341 | - * ARMv7-M interrupt masking works differently than -A or -R. | ||
342 | - * There is no FIQ/IRQ distinction. Instead of I and F bits | ||
343 | - * masking FIQ and IRQ interrupts, an exception is taken only | ||
344 | - * if it is higher priority than the current execution priority | ||
345 | - * (which depends on state like BASEPRI, FAULTMASK and the | ||
346 | - * currently active exception). | ||
347 | - */ | ||
348 | - if (interrupt_request & CPU_INTERRUPT_HARD | ||
349 | - && (armv7m_nvic_can_take_pending_exception(env->nvic))) { | ||
350 | - cs->exception_index = EXCP_IRQ; | ||
351 | - cc->tcg_ops->do_interrupt(cs); | ||
352 | - ret = true; | ||
353 | - } | ||
354 | - return ret; | ||
355 | -} | ||
356 | -#endif /* !CONFIG_USER_ONLY */ | ||
357 | - | ||
358 | static void arm926_initfn(Object *obj) | ||
148 | { | 359 | { |
360 | ARMCPU *cpu = ARM_CPU(obj); | ||
361 | @@ -XXX,XX +XXX,XX @@ static void cortex_a15_initfn(Object *obj) | ||
362 | define_arm_cp_regs(cpu, cortexa15_cp_reginfo); | ||
363 | } | ||
364 | |||
365 | -static void cortex_m0_initfn(Object *obj) | ||
366 | -{ | ||
367 | - ARMCPU *cpu = ARM_CPU(obj); | ||
368 | - set_feature(&cpu->env, ARM_FEATURE_V6); | ||
369 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
370 | - | ||
371 | - cpu->midr = 0x410cc200; | ||
372 | - | ||
149 | - /* | 373 | - /* |
150 | - * Simplify: | 374 | - * These ID register values are not guest visible, because |
151 | - * = ((a3 << 16) + ((e1 * e2) << 1) + (1 << 15)) >> 16 | 375 | - * we do not implement the Main Extension. They must be set |
152 | - * = ((a3 << 15) + (e1 * e2) + (1 << 14)) >> 15 | 376 | - * to values corresponding to the Cortex-M0's implemented |
377 | - * features, because QEMU generally controls its emulation | ||
378 | - * by looking at ID register fields. We use the same values as | ||
379 | - * for the M3. | ||
153 | - */ | 380 | - */ |
154 | + /* Simplify similarly to do_sqrdmlah_b above. */ | 381 | - cpu->isar.id_pfr0 = 0x00000030; |
155 | int32_t ret = (int32_t)src1 * src2; | 382 | - cpu->isar.id_pfr1 = 0x00000200; |
156 | if (neg) { | 383 | - cpu->isar.id_dfr0 = 0x00100000; |
157 | ret = -ret; | 384 | - cpu->id_afr0 = 0x00000000; |
158 | @@ -XXX,XX +XXX,XX @@ void HELPER(neon_sqrdmulh_h)(void *vd, void *vn, void *vm, | 385 | - cpu->isar.id_mmfr0 = 0x00000030; |
159 | clear_tail(d, opr_sz, simd_maxsz(desc)); | 386 | - cpu->isar.id_mmfr1 = 0x00000000; |
387 | - cpu->isar.id_mmfr2 = 0x00000000; | ||
388 | - cpu->isar.id_mmfr3 = 0x00000000; | ||
389 | - cpu->isar.id_isar0 = 0x01141110; | ||
390 | - cpu->isar.id_isar1 = 0x02111000; | ||
391 | - cpu->isar.id_isar2 = 0x21112231; | ||
392 | - cpu->isar.id_isar3 = 0x01111110; | ||
393 | - cpu->isar.id_isar4 = 0x01310102; | ||
394 | - cpu->isar.id_isar5 = 0x00000000; | ||
395 | - cpu->isar.id_isar6 = 0x00000000; | ||
396 | -} | ||
397 | - | ||
398 | -static void cortex_m3_initfn(Object *obj) | ||
399 | -{ | ||
400 | - ARMCPU *cpu = ARM_CPU(obj); | ||
401 | - set_feature(&cpu->env, ARM_FEATURE_V7); | ||
402 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
403 | - set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
404 | - cpu->midr = 0x410fc231; | ||
405 | - cpu->pmsav7_dregion = 8; | ||
406 | - cpu->isar.id_pfr0 = 0x00000030; | ||
407 | - cpu->isar.id_pfr1 = 0x00000200; | ||
408 | - cpu->isar.id_dfr0 = 0x00100000; | ||
409 | - cpu->id_afr0 = 0x00000000; | ||
410 | - cpu->isar.id_mmfr0 = 0x00000030; | ||
411 | - cpu->isar.id_mmfr1 = 0x00000000; | ||
412 | - cpu->isar.id_mmfr2 = 0x00000000; | ||
413 | - cpu->isar.id_mmfr3 = 0x00000000; | ||
414 | - cpu->isar.id_isar0 = 0x01141110; | ||
415 | - cpu->isar.id_isar1 = 0x02111000; | ||
416 | - cpu->isar.id_isar2 = 0x21112231; | ||
417 | - cpu->isar.id_isar3 = 0x01111110; | ||
418 | - cpu->isar.id_isar4 = 0x01310102; | ||
419 | - cpu->isar.id_isar5 = 0x00000000; | ||
420 | - cpu->isar.id_isar6 = 0x00000000; | ||
421 | -} | ||
422 | - | ||
423 | -static void cortex_m4_initfn(Object *obj) | ||
424 | -{ | ||
425 | - ARMCPU *cpu = ARM_CPU(obj); | ||
426 | - | ||
427 | - set_feature(&cpu->env, ARM_FEATURE_V7); | ||
428 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
429 | - set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
430 | - set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
431 | - cpu->midr = 0x410fc240; /* r0p0 */ | ||
432 | - cpu->pmsav7_dregion = 8; | ||
433 | - cpu->isar.mvfr0 = 0x10110021; | ||
434 | - cpu->isar.mvfr1 = 0x11000011; | ||
435 | - cpu->isar.mvfr2 = 0x00000000; | ||
436 | - cpu->isar.id_pfr0 = 0x00000030; | ||
437 | - cpu->isar.id_pfr1 = 0x00000200; | ||
438 | - cpu->isar.id_dfr0 = 0x00100000; | ||
439 | - cpu->id_afr0 = 0x00000000; | ||
440 | - cpu->isar.id_mmfr0 = 0x00000030; | ||
441 | - cpu->isar.id_mmfr1 = 0x00000000; | ||
442 | - cpu->isar.id_mmfr2 = 0x00000000; | ||
443 | - cpu->isar.id_mmfr3 = 0x00000000; | ||
444 | - cpu->isar.id_isar0 = 0x01141110; | ||
445 | - cpu->isar.id_isar1 = 0x02111000; | ||
446 | - cpu->isar.id_isar2 = 0x21112231; | ||
447 | - cpu->isar.id_isar3 = 0x01111110; | ||
448 | - cpu->isar.id_isar4 = 0x01310102; | ||
449 | - cpu->isar.id_isar5 = 0x00000000; | ||
450 | - cpu->isar.id_isar6 = 0x00000000; | ||
451 | -} | ||
452 | - | ||
453 | -static void cortex_m7_initfn(Object *obj) | ||
454 | -{ | ||
455 | - ARMCPU *cpu = ARM_CPU(obj); | ||
456 | - | ||
457 | - set_feature(&cpu->env, ARM_FEATURE_V7); | ||
458 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
459 | - set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
460 | - set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
461 | - cpu->midr = 0x411fc272; /* r1p2 */ | ||
462 | - cpu->pmsav7_dregion = 8; | ||
463 | - cpu->isar.mvfr0 = 0x10110221; | ||
464 | - cpu->isar.mvfr1 = 0x12000011; | ||
465 | - cpu->isar.mvfr2 = 0x00000040; | ||
466 | - cpu->isar.id_pfr0 = 0x00000030; | ||
467 | - cpu->isar.id_pfr1 = 0x00000200; | ||
468 | - cpu->isar.id_dfr0 = 0x00100000; | ||
469 | - cpu->id_afr0 = 0x00000000; | ||
470 | - cpu->isar.id_mmfr0 = 0x00100030; | ||
471 | - cpu->isar.id_mmfr1 = 0x00000000; | ||
472 | - cpu->isar.id_mmfr2 = 0x01000000; | ||
473 | - cpu->isar.id_mmfr3 = 0x00000000; | ||
474 | - cpu->isar.id_isar0 = 0x01101110; | ||
475 | - cpu->isar.id_isar1 = 0x02112000; | ||
476 | - cpu->isar.id_isar2 = 0x20232231; | ||
477 | - cpu->isar.id_isar3 = 0x01111131; | ||
478 | - cpu->isar.id_isar4 = 0x01310132; | ||
479 | - cpu->isar.id_isar5 = 0x00000000; | ||
480 | - cpu->isar.id_isar6 = 0x00000000; | ||
481 | -} | ||
482 | - | ||
483 | -static void cortex_m33_initfn(Object *obj) | ||
484 | -{ | ||
485 | - ARMCPU *cpu = ARM_CPU(obj); | ||
486 | - | ||
487 | - set_feature(&cpu->env, ARM_FEATURE_V8); | ||
488 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
489 | - set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
490 | - set_feature(&cpu->env, ARM_FEATURE_M_SECURITY); | ||
491 | - set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
492 | - cpu->midr = 0x410fd213; /* r0p3 */ | ||
493 | - cpu->pmsav7_dregion = 16; | ||
494 | - cpu->sau_sregion = 8; | ||
495 | - cpu->isar.mvfr0 = 0x10110021; | ||
496 | - cpu->isar.mvfr1 = 0x11000011; | ||
497 | - cpu->isar.mvfr2 = 0x00000040; | ||
498 | - cpu->isar.id_pfr0 = 0x00000030; | ||
499 | - cpu->isar.id_pfr1 = 0x00000210; | ||
500 | - cpu->isar.id_dfr0 = 0x00200000; | ||
501 | - cpu->id_afr0 = 0x00000000; | ||
502 | - cpu->isar.id_mmfr0 = 0x00101F40; | ||
503 | - cpu->isar.id_mmfr1 = 0x00000000; | ||
504 | - cpu->isar.id_mmfr2 = 0x01000000; | ||
505 | - cpu->isar.id_mmfr3 = 0x00000000; | ||
506 | - cpu->isar.id_isar0 = 0x01101110; | ||
507 | - cpu->isar.id_isar1 = 0x02212000; | ||
508 | - cpu->isar.id_isar2 = 0x20232232; | ||
509 | - cpu->isar.id_isar3 = 0x01111131; | ||
510 | - cpu->isar.id_isar4 = 0x01310132; | ||
511 | - cpu->isar.id_isar5 = 0x00000000; | ||
512 | - cpu->isar.id_isar6 = 0x00000000; | ||
513 | - cpu->clidr = 0x00000000; | ||
514 | - cpu->ctr = 0x8000c000; | ||
515 | -} | ||
516 | - | ||
517 | -static void cortex_m55_initfn(Object *obj) | ||
518 | -{ | ||
519 | - ARMCPU *cpu = ARM_CPU(obj); | ||
520 | - | ||
521 | - set_feature(&cpu->env, ARM_FEATURE_V8); | ||
522 | - set_feature(&cpu->env, ARM_FEATURE_V8_1M); | ||
523 | - set_feature(&cpu->env, ARM_FEATURE_M); | ||
524 | - set_feature(&cpu->env, ARM_FEATURE_M_MAIN); | ||
525 | - set_feature(&cpu->env, ARM_FEATURE_M_SECURITY); | ||
526 | - set_feature(&cpu->env, ARM_FEATURE_THUMB_DSP); | ||
527 | - cpu->midr = 0x410fd221; /* r0p1 */ | ||
528 | - cpu->revidr = 0; | ||
529 | - cpu->pmsav7_dregion = 16; | ||
530 | - cpu->sau_sregion = 8; | ||
531 | - /* These are the MVFR* values for the FPU + full MVE configuration */ | ||
532 | - cpu->isar.mvfr0 = 0x10110221; | ||
533 | - cpu->isar.mvfr1 = 0x12100211; | ||
534 | - cpu->isar.mvfr2 = 0x00000040; | ||
535 | - cpu->isar.id_pfr0 = 0x20000030; | ||
536 | - cpu->isar.id_pfr1 = 0x00000230; | ||
537 | - cpu->isar.id_dfr0 = 0x10200000; | ||
538 | - cpu->id_afr0 = 0x00000000; | ||
539 | - cpu->isar.id_mmfr0 = 0x00111040; | ||
540 | - cpu->isar.id_mmfr1 = 0x00000000; | ||
541 | - cpu->isar.id_mmfr2 = 0x01000000; | ||
542 | - cpu->isar.id_mmfr3 = 0x00000011; | ||
543 | - cpu->isar.id_isar0 = 0x01103110; | ||
544 | - cpu->isar.id_isar1 = 0x02212000; | ||
545 | - cpu->isar.id_isar2 = 0x20232232; | ||
546 | - cpu->isar.id_isar3 = 0x01111131; | ||
547 | - cpu->isar.id_isar4 = 0x01310132; | ||
548 | - cpu->isar.id_isar5 = 0x00000000; | ||
549 | - cpu->isar.id_isar6 = 0x00000000; | ||
550 | - cpu->clidr = 0x00000000; /* caches not implemented */ | ||
551 | - cpu->ctr = 0x8303c003; | ||
552 | -} | ||
553 | - | ||
554 | static const ARMCPRegInfo cortexr5_cp_reginfo[] = { | ||
555 | /* Dummy the TCM region regs for the moment */ | ||
556 | { .name = "ATCM", .cp = 15, .opc1 = 0, .crn = 9, .crm = 1, .opc2 = 0, | ||
557 | @@ -XXX,XX +XXX,XX @@ static void pxa270c5_initfn(Object *obj) | ||
558 | cpu->reset_sctlr = 0x00000078; | ||
160 | } | 559 | } |
161 | 560 | ||
162 | +void HELPER(sve2_sqrdmlah_h)(void *vd, void *vn, void *vm, | 561 | -static const TCGCPUOps arm_v7m_tcg_ops = { |
163 | + void *va, uint32_t desc) | 562 | - .initialize = arm_translate_init, |
164 | +{ | 563 | - .synchronize_from_tb = arm_cpu_synchronize_from_tb, |
165 | + intptr_t i, opr_sz = simd_oprsz(desc); | 564 | - .debug_excp_handler = arm_debug_excp_handler, |
166 | + int16_t *d = vd, *n = vn, *m = vm, *a = va; | 565 | - .restore_state_to_opc = arm_restore_state_to_opc, |
167 | + uint32_t discard; | 566 | - |
168 | + | 567 | -#ifdef CONFIG_USER_ONLY |
169 | + for (i = 0; i < opr_sz / 2; ++i) { | 568 | - .record_sigsegv = arm_cpu_record_sigsegv, |
170 | + d[i] = do_sqrdmlah_h(n[i], m[i], a[i], false, true, &discard); | 569 | - .record_sigbus = arm_cpu_record_sigbus, |
171 | + } | 570 | -#else |
172 | +} | 571 | - .tlb_fill = arm_cpu_tlb_fill, |
173 | + | 572 | - .cpu_exec_interrupt = arm_v7m_cpu_exec_interrupt, |
174 | +void HELPER(sve2_sqrdmlsh_h)(void *vd, void *vn, void *vm, | 573 | - .do_interrupt = arm_v7m_cpu_do_interrupt, |
175 | + void *va, uint32_t desc) | 574 | - .do_transaction_failed = arm_cpu_do_transaction_failed, |
176 | +{ | 575 | - .do_unaligned_access = arm_cpu_do_unaligned_access, |
177 | + intptr_t i, opr_sz = simd_oprsz(desc); | 576 | - .adjust_watchpoint_address = arm_adjust_watchpoint_address, |
178 | + int16_t *d = vd, *n = vn, *m = vm, *a = va; | 577 | - .debug_check_watchpoint = arm_debug_check_watchpoint, |
179 | + uint32_t discard; | 578 | - .debug_check_breakpoint = arm_debug_check_breakpoint, |
180 | + | 579 | -#endif /* !CONFIG_USER_ONLY */ |
181 | + for (i = 0; i < opr_sz / 2; ++i) { | 580 | -}; |
182 | + d[i] = do_sqrdmlah_h(n[i], m[i], a[i], true, true, &discard); | 581 | - |
183 | + } | 582 | -static void arm_v7m_class_init(ObjectClass *oc, void *data) |
184 | +} | 583 | -{ |
185 | + | 584 | - ARMCPUClass *acc = ARM_CPU_CLASS(oc); |
186 | /* Signed saturating rounding doubling multiply-accumulate high half, 32-bit */ | 585 | - CPUClass *cc = CPU_CLASS(oc); |
187 | static int32_t do_sqrdmlah_s(int32_t src1, int32_t src2, int32_t src3, | 586 | - |
188 | bool neg, bool round, uint32_t *sat) | 587 | - acc->info = data; |
189 | { | 588 | - cc->tcg_ops = &arm_v7m_tcg_ops; |
190 | - /* Simplify similarly to int_qrdmlah_s16 above. */ | 589 | - cc->gdb_core_xml_file = "arm-m-profile.xml"; |
191 | + /* Simplify similarly to do_sqrdmlah_b above. */ | 590 | -} |
192 | int64_t ret = (int64_t)src1 * src2; | 591 | - |
193 | if (neg) { | 592 | #ifndef TARGET_AARCH64 |
194 | ret = -ret; | 593 | /* |
195 | @@ -XXX,XX +XXX,XX @@ void HELPER(neon_sqrdmulh_s)(void *vd, void *vn, void *vm, | 594 | * -cpu max: a CPU with as many features enabled as our emulation supports. |
196 | clear_tail(d, opr_sz, simd_maxsz(desc)); | 595 | @@ -XXX,XX +XXX,XX @@ static const ARMCPUInfo arm_tcg_cpus[] = { |
197 | } | 596 | { .name = "cortex-a8", .initfn = cortex_a8_initfn }, |
198 | 597 | { .name = "cortex-a9", .initfn = cortex_a9_initfn }, | |
199 | +void HELPER(sve2_sqrdmlah_s)(void *vd, void *vn, void *vm, | 598 | { .name = "cortex-a15", .initfn = cortex_a15_initfn }, |
200 | + void *va, uint32_t desc) | 599 | - { .name = "cortex-m0", .initfn = cortex_m0_initfn, |
201 | +{ | 600 | - .class_init = arm_v7m_class_init }, |
202 | + intptr_t i, opr_sz = simd_oprsz(desc); | 601 | - { .name = "cortex-m3", .initfn = cortex_m3_initfn, |
203 | + int32_t *d = vd, *n = vn, *m = vm, *a = va; | 602 | - .class_init = arm_v7m_class_init }, |
204 | + uint32_t discard; | 603 | - { .name = "cortex-m4", .initfn = cortex_m4_initfn, |
205 | + | 604 | - .class_init = arm_v7m_class_init }, |
206 | + for (i = 0; i < opr_sz / 4; ++i) { | 605 | - { .name = "cortex-m7", .initfn = cortex_m7_initfn, |
207 | + d[i] = do_sqrdmlah_s(n[i], m[i], a[i], false, true, &discard); | 606 | - .class_init = arm_v7m_class_init }, |
208 | + } | 607 | - { .name = "cortex-m33", .initfn = cortex_m33_initfn, |
209 | +} | 608 | - .class_init = arm_v7m_class_init }, |
210 | + | 609 | - { .name = "cortex-m55", .initfn = cortex_m55_initfn, |
211 | +void HELPER(sve2_sqrdmlsh_s)(void *vd, void *vn, void *vm, | 610 | - .class_init = arm_v7m_class_init }, |
212 | + void *va, uint32_t desc) | 611 | { .name = "cortex-r5", .initfn = cortex_r5_initfn }, |
213 | +{ | 612 | { .name = "cortex-r5f", .initfn = cortex_r5f_initfn }, |
214 | + intptr_t i, opr_sz = simd_oprsz(desc); | 613 | { .name = "cortex-r52", .initfn = cortex_r52_initfn }, |
215 | + int32_t *d = vd, *n = vn, *m = vm, *a = va; | 614 | diff --git a/target/arm/meson.build b/target/arm/meson.build |
216 | + uint32_t discard; | 615 | index XXXXXXX..XXXXXXX 100644 |
217 | + | 616 | --- a/target/arm/meson.build |
218 | + for (i = 0; i < opr_sz / 4; ++i) { | 617 | +++ b/target/arm/meson.build |
219 | + d[i] = do_sqrdmlah_s(n[i], m[i], a[i], true, true, &discard); | 618 | @@ -XXX,XX +XXX,XX @@ arm_system_ss.add(files( |
220 | + } | 619 | 'ptw.c', |
221 | +} | 620 | )) |
222 | + | 621 | |
223 | +/* Signed saturating rounding doubling multiply-accumulate high half, 64-bit */ | 622 | +arm_user_ss = ss.source_set() |
224 | +static int64_t do_sat128_d(Int128 r) | 623 | + |
225 | +{ | 624 | subdir('hvf') |
226 | + int64_t ls = int128_getlo(r); | 625 | |
227 | + int64_t hs = int128_gethi(r); | 626 | if 'CONFIG_TCG' in config_all_accel |
228 | + | 627 | @@ -XXX,XX +XXX,XX @@ endif |
229 | + if (unlikely(hs != (ls >> 63))) { | 628 | |
230 | + return hs < 0 ? INT64_MIN : INT64_MAX; | 629 | target_arch += {'arm': arm_ss} |
231 | + } | 630 | target_system_arch += {'arm': arm_system_ss} |
232 | + return ls; | 631 | +target_user_arch += {'arm': arm_user_ss} |
233 | +} | 632 | diff --git a/target/arm/tcg/meson.build b/target/arm/tcg/meson.build |
234 | + | 633 | index XXXXXXX..XXXXXXX 100644 |
235 | +static int64_t do_sqrdmlah_d(int64_t n, int64_t m, int64_t a, | 634 | --- a/target/arm/tcg/meson.build |
236 | + bool neg, bool round) | 635 | +++ b/target/arm/tcg/meson.build |
237 | +{ | 636 | @@ -XXX,XX +XXX,XX @@ arm_ss.add(when: 'TARGET_AARCH64', if_true: files( |
238 | + uint64_t l, h; | 637 | arm_system_ss.add(files( |
239 | + Int128 r, t; | 638 | 'psci.c', |
240 | + | 639 | )) |
241 | + /* As in do_sqrdmlah_b, but with 128-bit arithmetic. */ | 640 | + |
242 | + muls64(&l, &h, m, n); | 641 | +arm_system_ss.add(when: 'CONFIG_ARM_V7M', if_true: files('cpu-v7m.c')) |
243 | + r = int128_make128(l, h); | 642 | +arm_user_ss.add(when: 'TARGET_AARCH64', if_false: files('cpu-v7m.c')) |
244 | + if (neg) { | ||
245 | + r = int128_neg(r); | ||
246 | + } | ||
247 | + if (a) { | ||
248 | + t = int128_exts64(a); | ||
249 | + t = int128_lshift(t, 63); | ||
250 | + r = int128_add(r, t); | ||
251 | + } | ||
252 | + if (round) { | ||
253 | + t = int128_exts64(1ll << 62); | ||
254 | + r = int128_add(r, t); | ||
255 | + } | ||
256 | + r = int128_rshift(r, 63); | ||
257 | + | ||
258 | + return do_sat128_d(r); | ||
259 | +} | ||
260 | + | ||
261 | +void HELPER(sve2_sqrdmlah_d)(void *vd, void *vn, void *vm, | ||
262 | + void *va, uint32_t desc) | ||
263 | +{ | ||
264 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
265 | + int64_t *d = vd, *n = vn, *m = vm, *a = va; | ||
266 | + | ||
267 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
268 | + d[i] = do_sqrdmlah_d(n[i], m[i], a[i], false, true); | ||
269 | + } | ||
270 | +} | ||
271 | + | ||
272 | +void HELPER(sve2_sqrdmlsh_d)(void *vd, void *vn, void *vm, | ||
273 | + void *va, uint32_t desc) | ||
274 | +{ | ||
275 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
276 | + int64_t *d = vd, *n = vn, *m = vm, *a = va; | ||
277 | + | ||
278 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
279 | + d[i] = do_sqrdmlah_d(n[i], m[i], a[i], true, true); | ||
280 | + } | ||
281 | +} | ||
282 | + | ||
283 | /* Integer 8 and 16-bit dot-product. | ||
284 | * | ||
285 | * Note that for the loops herein, host endianness does not matter | ||
286 | -- | 643 | -- |
287 | 2.20.1 | 644 | 2.34.1 |
288 | |||
289 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-37-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 28 ++++++++++++++ | ||
9 | target/arm/sve.decode | 11 ++++++ | ||
10 | target/arm/sve_helper.c | 18 +++++++++ | ||
11 | target/arm/translate-sve.c | 76 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 133 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqdmlsl_zzzw_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_sqdmlsl_zzzw_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_smlal_zzzw_h, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_smlal_zzzw_s, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_smlal_zzzw_d, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_5(sve2_umlal_zzzw_h, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(sve2_umlal_zzzw_s, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(sve2_umlal_zzzw_d, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | + | ||
37 | +DEF_HELPER_FLAGS_5(sve2_smlsl_zzzw_h, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | +DEF_HELPER_FLAGS_5(sve2_smlsl_zzzw_s, TCG_CALL_NO_RWG, | ||
40 | + void, ptr, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_5(sve2_smlsl_zzzw_d, TCG_CALL_NO_RWG, | ||
42 | + void, ptr, ptr, ptr, ptr, i32) | ||
43 | + | ||
44 | +DEF_HELPER_FLAGS_5(sve2_umlsl_zzzw_h, TCG_CALL_NO_RWG, | ||
45 | + void, ptr, ptr, ptr, ptr, i32) | ||
46 | +DEF_HELPER_FLAGS_5(sve2_umlsl_zzzw_s, TCG_CALL_NO_RWG, | ||
47 | + void, ptr, ptr, ptr, ptr, i32) | ||
48 | +DEF_HELPER_FLAGS_5(sve2_umlsl_zzzw_d, TCG_CALL_NO_RWG, | ||
49 | + void, ptr, ptr, ptr, ptr, i32) | ||
50 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/target/arm/sve.decode | ||
53 | +++ b/target/arm/sve.decode | ||
54 | @@ -XXX,XX +XXX,XX @@ SQDMLSLBT 01000100 .. 0 ..... 00001 1 ..... ..... @rda_rn_rm | ||
55 | |||
56 | SQRDMLAH_zzzz 01000100 .. 0 ..... 01110 0 ..... ..... @rda_rn_rm | ||
57 | SQRDMLSH_zzzz 01000100 .. 0 ..... 01110 1 ..... ..... @rda_rn_rm | ||
58 | + | ||
59 | +## SVE2 integer multiply-add long | ||
60 | + | ||
61 | +SMLALB_zzzw 01000100 .. 0 ..... 010 000 ..... ..... @rda_rn_rm | ||
62 | +SMLALT_zzzw 01000100 .. 0 ..... 010 001 ..... ..... @rda_rn_rm | ||
63 | +UMLALB_zzzw 01000100 .. 0 ..... 010 010 ..... ..... @rda_rn_rm | ||
64 | +UMLALT_zzzw 01000100 .. 0 ..... 010 011 ..... ..... @rda_rn_rm | ||
65 | +SMLSLB_zzzw 01000100 .. 0 ..... 010 100 ..... ..... @rda_rn_rm | ||
66 | +SMLSLT_zzzw 01000100 .. 0 ..... 010 101 ..... ..... @rda_rn_rm | ||
67 | +UMLSLB_zzzw 01000100 .. 0 ..... 010 110 ..... ..... @rda_rn_rm | ||
68 | +UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | ||
69 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
70 | index XXXXXXX..XXXXXXX 100644 | ||
71 | --- a/target/arm/sve_helper.c | ||
72 | +++ b/target/arm/sve_helper.c | ||
73 | @@ -XXX,XX +XXX,XX @@ DO_ZZZW_ACC(sve2_uabal_h, uint16_t, uint8_t, H1_2, H1, DO_ABD) | ||
74 | DO_ZZZW_ACC(sve2_uabal_s, uint32_t, uint16_t, H1_4, H1_2, DO_ABD) | ||
75 | DO_ZZZW_ACC(sve2_uabal_d, uint64_t, uint32_t, , H1_4, DO_ABD) | ||
76 | |||
77 | +DO_ZZZW_ACC(sve2_smlal_zzzw_h, int16_t, int8_t, H1_2, H1, DO_MUL) | ||
78 | +DO_ZZZW_ACC(sve2_smlal_zzzw_s, int32_t, int16_t, H1_4, H1_2, DO_MUL) | ||
79 | +DO_ZZZW_ACC(sve2_smlal_zzzw_d, int64_t, int32_t, , H1_4, DO_MUL) | ||
80 | + | ||
81 | +DO_ZZZW_ACC(sve2_umlal_zzzw_h, uint16_t, uint8_t, H1_2, H1, DO_MUL) | ||
82 | +DO_ZZZW_ACC(sve2_umlal_zzzw_s, uint32_t, uint16_t, H1_4, H1_2, DO_MUL) | ||
83 | +DO_ZZZW_ACC(sve2_umlal_zzzw_d, uint64_t, uint32_t, , H1_4, DO_MUL) | ||
84 | + | ||
85 | +#define DO_NMUL(N, M) -(N * M) | ||
86 | + | ||
87 | +DO_ZZZW_ACC(sve2_smlsl_zzzw_h, int16_t, int8_t, H1_2, H1, DO_NMUL) | ||
88 | +DO_ZZZW_ACC(sve2_smlsl_zzzw_s, int32_t, int16_t, H1_4, H1_2, DO_NMUL) | ||
89 | +DO_ZZZW_ACC(sve2_smlsl_zzzw_d, int64_t, int32_t, , H1_4, DO_NMUL) | ||
90 | + | ||
91 | +DO_ZZZW_ACC(sve2_umlsl_zzzw_h, uint16_t, uint8_t, H1_2, H1, DO_NMUL) | ||
92 | +DO_ZZZW_ACC(sve2_umlsl_zzzw_s, uint32_t, uint16_t, H1_4, H1_2, DO_NMUL) | ||
93 | +DO_ZZZW_ACC(sve2_umlsl_zzzw_d, uint64_t, uint32_t, , H1_4, DO_NMUL) | ||
94 | + | ||
95 | #undef DO_ZZZW_ACC | ||
96 | |||
97 | #define DO_XTNB(NAME, TYPE, OP) \ | ||
98 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
99 | index XXXXXXX..XXXXXXX 100644 | ||
100 | --- a/target/arm/translate-sve.c | ||
101 | +++ b/target/arm/translate-sve.c | ||
102 | @@ -XXX,XX +XXX,XX @@ static bool trans_SQRDMLSH_zzzz(DisasContext *s, arg_rrrr_esz *a) | ||
103 | }; | ||
104 | return do_sve2_zzzz_ool(s, a, fns[a->esz], 0); | ||
105 | } | ||
106 | + | ||
107 | +static bool do_smlal_zzzw(DisasContext *s, arg_rrrr_esz *a, bool sel) | ||
108 | +{ | ||
109 | + static gen_helper_gvec_4 * const fns[] = { | ||
110 | + NULL, gen_helper_sve2_smlal_zzzw_h, | ||
111 | + gen_helper_sve2_smlal_zzzw_s, gen_helper_sve2_smlal_zzzw_d, | ||
112 | + }; | ||
113 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], sel); | ||
114 | +} | ||
115 | + | ||
116 | +static bool trans_SMLALB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
117 | +{ | ||
118 | + return do_smlal_zzzw(s, a, false); | ||
119 | +} | ||
120 | + | ||
121 | +static bool trans_SMLALT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
122 | +{ | ||
123 | + return do_smlal_zzzw(s, a, true); | ||
124 | +} | ||
125 | + | ||
126 | +static bool do_umlal_zzzw(DisasContext *s, arg_rrrr_esz *a, bool sel) | ||
127 | +{ | ||
128 | + static gen_helper_gvec_4 * const fns[] = { | ||
129 | + NULL, gen_helper_sve2_umlal_zzzw_h, | ||
130 | + gen_helper_sve2_umlal_zzzw_s, gen_helper_sve2_umlal_zzzw_d, | ||
131 | + }; | ||
132 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], sel); | ||
133 | +} | ||
134 | + | ||
135 | +static bool trans_UMLALB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
136 | +{ | ||
137 | + return do_umlal_zzzw(s, a, false); | ||
138 | +} | ||
139 | + | ||
140 | +static bool trans_UMLALT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
141 | +{ | ||
142 | + return do_umlal_zzzw(s, a, true); | ||
143 | +} | ||
144 | + | ||
145 | +static bool do_smlsl_zzzw(DisasContext *s, arg_rrrr_esz *a, bool sel) | ||
146 | +{ | ||
147 | + static gen_helper_gvec_4 * const fns[] = { | ||
148 | + NULL, gen_helper_sve2_smlsl_zzzw_h, | ||
149 | + gen_helper_sve2_smlsl_zzzw_s, gen_helper_sve2_smlsl_zzzw_d, | ||
150 | + }; | ||
151 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], sel); | ||
152 | +} | ||
153 | + | ||
154 | +static bool trans_SMLSLB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
155 | +{ | ||
156 | + return do_smlsl_zzzw(s, a, false); | ||
157 | +} | ||
158 | + | ||
159 | +static bool trans_SMLSLT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
160 | +{ | ||
161 | + return do_smlsl_zzzw(s, a, true); | ||
162 | +} | ||
163 | + | ||
164 | +static bool do_umlsl_zzzw(DisasContext *s, arg_rrrr_esz *a, bool sel) | ||
165 | +{ | ||
166 | + static gen_helper_gvec_4 * const fns[] = { | ||
167 | + NULL, gen_helper_sve2_umlsl_zzzw_h, | ||
168 | + gen_helper_sve2_umlsl_zzzw_s, gen_helper_sve2_umlsl_zzzw_d, | ||
169 | + }; | ||
170 | + return do_sve2_zzzz_ool(s, a, fns[a->esz], sel); | ||
171 | +} | ||
172 | + | ||
173 | +static bool trans_UMLSLB_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
174 | +{ | ||
175 | + return do_umlsl_zzzw(s, a, false); | ||
176 | +} | ||
177 | + | ||
178 | +static bool trans_UMLSLT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
179 | +{ | ||
180 | + return do_umlsl_zzzw(s, a, true); | ||
181 | +} | ||
182 | -- | ||
183 | 2.20.1 | ||
184 | |||
185 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-38-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 18 +++++++++++++++ | ||
9 | target/arm/vec_internal.h | 5 +++++ | ||
10 | target/arm/sve.decode | 5 +++++ | ||
11 | target/arm/sve_helper.c | 46 ++++++++++++++++++++++++++++++++++++++ | ||
12 | target/arm/translate-sve.c | 32 ++++++++++++++++++++++++++ | ||
13 | target/arm/vec_helper.c | 15 ++++++------- | ||
14 | 6 files changed, 113 insertions(+), 8 deletions(-) | ||
15 | |||
16 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/helper-sve.h | ||
19 | +++ b/target/arm/helper-sve.h | ||
20 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_umlsl_zzzw_s, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_5(sve2_umlsl_zzzw_d, TCG_CALL_NO_RWG, | ||
23 | void, ptr, ptr, ptr, ptr, i32) | ||
24 | + | ||
25 | +DEF_HELPER_FLAGS_5(sve2_cmla_zzzz_b, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_cmla_zzzz_h, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_5(sve2_cmla_zzzz_s, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_5(sve2_cmla_zzzz_d, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | + | ||
34 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_b, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_h, TCG_CALL_NO_RWG, | ||
37 | + void, ptr, ptr, ptr, ptr, i32) | ||
38 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_s, TCG_CALL_NO_RWG, | ||
39 | + void, ptr, ptr, ptr, ptr, i32) | ||
40 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_d, TCG_CALL_NO_RWG, | ||
41 | + void, ptr, ptr, ptr, ptr, i32) | ||
42 | diff --git a/target/arm/vec_internal.h b/target/arm/vec_internal.h | ||
43 | index XXXXXXX..XXXXXXX 100644 | ||
44 | --- a/target/arm/vec_internal.h | ||
45 | +++ b/target/arm/vec_internal.h | ||
46 | @@ -XXX,XX +XXX,XX @@ static inline int64_t do_suqrshl_d(int64_t src, int64_t shift, | ||
47 | return do_uqrshl_d(src, shift, round, sat); | ||
48 | } | ||
49 | |||
50 | +int8_t do_sqrdmlah_b(int8_t, int8_t, int8_t, bool, bool); | ||
51 | +int16_t do_sqrdmlah_h(int16_t, int16_t, int16_t, bool, bool, uint32_t *); | ||
52 | +int32_t do_sqrdmlah_s(int32_t, int32_t, int32_t, bool, bool, uint32_t *); | ||
53 | +int64_t do_sqrdmlah_d(int64_t, int64_t, int64_t, bool, bool); | ||
54 | + | ||
55 | #endif /* TARGET_ARM_VEC_INTERNALS_H */ | ||
56 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
57 | index XXXXXXX..XXXXXXX 100644 | ||
58 | --- a/target/arm/sve.decode | ||
59 | +++ b/target/arm/sve.decode | ||
60 | @@ -XXX,XX +XXX,XX @@ SMLSLB_zzzw 01000100 .. 0 ..... 010 100 ..... ..... @rda_rn_rm | ||
61 | SMLSLT_zzzw 01000100 .. 0 ..... 010 101 ..... ..... @rda_rn_rm | ||
62 | UMLSLB_zzzw 01000100 .. 0 ..... 010 110 ..... ..... @rda_rn_rm | ||
63 | UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | ||
64 | + | ||
65 | +## SVE2 complex integer multiply-add | ||
66 | + | ||
67 | +CMLA_zzzz 01000100 esz:2 0 rm:5 0010 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
68 | +SQRDCMLAH_zzzz 01000100 esz:2 0 rm:5 0011 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
69 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
70 | index XXXXXXX..XXXXXXX 100644 | ||
71 | --- a/target/arm/sve_helper.c | ||
72 | +++ b/target/arm/sve_helper.c | ||
73 | @@ -XXX,XX +XXX,XX @@ DO_SQDMLAL(sve2_sqdmlsl_zzzw_d, int64_t, int32_t, , H1_4, | ||
74 | |||
75 | #undef DO_SQDMLAL | ||
76 | |||
77 | +#define DO_CMLA_FUNC(NAME, TYPE, H, OP) \ | ||
78 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
79 | +{ \ | ||
80 | + intptr_t i, opr_sz = simd_oprsz(desc) / sizeof(TYPE); \ | ||
81 | + int rot = simd_data(desc); \ | ||
82 | + int sel_a = rot & 1, sel_b = sel_a ^ 1; \ | ||
83 | + bool sub_r = rot == 1 || rot == 2; \ | ||
84 | + bool sub_i = rot >= 2; \ | ||
85 | + TYPE *d = vd, *n = vn, *m = vm, *a = va; \ | ||
86 | + for (i = 0; i < opr_sz; i += 2) { \ | ||
87 | + TYPE elt1_a = n[H(i + sel_a)]; \ | ||
88 | + TYPE elt2_a = m[H(i + sel_a)]; \ | ||
89 | + TYPE elt2_b = m[H(i + sel_b)]; \ | ||
90 | + d[H(i)] = OP(elt1_a, elt2_a, a[H(i)], sub_r); \ | ||
91 | + d[H(i + 1)] = OP(elt1_a, elt2_b, a[H(i + 1)], sub_i); \ | ||
92 | + } \ | ||
93 | +} | ||
94 | + | ||
95 | +#define DO_CMLA(N, M, A, S) (A + (N * M) * (S ? -1 : 1)) | ||
96 | + | ||
97 | +DO_CMLA_FUNC(sve2_cmla_zzzz_b, uint8_t, H1, DO_CMLA) | ||
98 | +DO_CMLA_FUNC(sve2_cmla_zzzz_h, uint16_t, H2, DO_CMLA) | ||
99 | +DO_CMLA_FUNC(sve2_cmla_zzzz_s, uint32_t, H4, DO_CMLA) | ||
100 | +DO_CMLA_FUNC(sve2_cmla_zzzz_d, uint64_t, , DO_CMLA) | ||
101 | + | ||
102 | +#define DO_SQRDMLAH_B(N, M, A, S) \ | ||
103 | + do_sqrdmlah_b(N, M, A, S, true) | ||
104 | +#define DO_SQRDMLAH_H(N, M, A, S) \ | ||
105 | + ({ uint32_t discard; do_sqrdmlah_h(N, M, A, S, true, &discard); }) | ||
106 | +#define DO_SQRDMLAH_S(N, M, A, S) \ | ||
107 | + ({ uint32_t discard; do_sqrdmlah_s(N, M, A, S, true, &discard); }) | ||
108 | +#define DO_SQRDMLAH_D(N, M, A, S) \ | ||
109 | + do_sqrdmlah_d(N, M, A, S, true) | ||
110 | + | ||
111 | +DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_b, int8_t, H1, DO_SQRDMLAH_B) | ||
112 | +DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_h, int16_t, H2, DO_SQRDMLAH_H) | ||
113 | +DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_s, int32_t, H4, DO_SQRDMLAH_S) | ||
114 | +DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_d, int64_t, , DO_SQRDMLAH_D) | ||
115 | + | ||
116 | +#undef DO_CMLA | ||
117 | +#undef DO_CMLA_FUNC | ||
118 | +#undef DO_SQRDMLAH_B | ||
119 | +#undef DO_SQRDMLAH_H | ||
120 | +#undef DO_SQRDMLAH_S | ||
121 | +#undef DO_SQRDMLAH_D | ||
122 | + | ||
123 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
124 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
125 | { \ | ||
126 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
127 | index XXXXXXX..XXXXXXX 100644 | ||
128 | --- a/target/arm/translate-sve.c | ||
129 | +++ b/target/arm/translate-sve.c | ||
130 | @@ -XXX,XX +XXX,XX @@ static bool trans_UMLSLT_zzzw(DisasContext *s, arg_rrrr_esz *a) | ||
131 | { | ||
132 | return do_umlsl_zzzw(s, a, true); | ||
133 | } | ||
134 | + | ||
135 | +static bool trans_CMLA_zzzz(DisasContext *s, arg_CMLA_zzzz *a) | ||
136 | +{ | ||
137 | + static gen_helper_gvec_4 * const fns[] = { | ||
138 | + gen_helper_sve2_cmla_zzzz_b, gen_helper_sve2_cmla_zzzz_h, | ||
139 | + gen_helper_sve2_cmla_zzzz_s, gen_helper_sve2_cmla_zzzz_d, | ||
140 | + }; | ||
141 | + | ||
142 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
143 | + return false; | ||
144 | + } | ||
145 | + if (sve_access_check(s)) { | ||
146 | + gen_gvec_ool_zzzz(s, fns[a->esz], a->rd, a->rn, a->rm, a->ra, a->rot); | ||
147 | + } | ||
148 | + return true; | ||
149 | +} | ||
150 | + | ||
151 | +static bool trans_SQRDCMLAH_zzzz(DisasContext *s, arg_SQRDCMLAH_zzzz *a) | ||
152 | +{ | ||
153 | + static gen_helper_gvec_4 * const fns[] = { | ||
154 | + gen_helper_sve2_sqrdcmlah_zzzz_b, gen_helper_sve2_sqrdcmlah_zzzz_h, | ||
155 | + gen_helper_sve2_sqrdcmlah_zzzz_s, gen_helper_sve2_sqrdcmlah_zzzz_d, | ||
156 | + }; | ||
157 | + | ||
158 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
159 | + return false; | ||
160 | + } | ||
161 | + if (sve_access_check(s)) { | ||
162 | + gen_gvec_ool_zzzz(s, fns[a->esz], a->rd, a->rn, a->rm, a->ra, a->rot); | ||
163 | + } | ||
164 | + return true; | ||
165 | +} | ||
166 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
167 | index XXXXXXX..XXXXXXX 100644 | ||
168 | --- a/target/arm/vec_helper.c | ||
169 | +++ b/target/arm/vec_helper.c | ||
170 | @@ -XXX,XX +XXX,XX @@ | ||
171 | #endif | ||
172 | |||
173 | /* Signed saturating rounding doubling multiply-accumulate high half, 8-bit */ | ||
174 | -static int8_t do_sqrdmlah_b(int8_t src1, int8_t src2, int8_t src3, | ||
175 | - bool neg, bool round) | ||
176 | +int8_t do_sqrdmlah_b(int8_t src1, int8_t src2, int8_t src3, | ||
177 | + bool neg, bool round) | ||
178 | { | ||
179 | /* | ||
180 | * Simplify: | ||
181 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_b)(void *vd, void *vn, void *vm, | ||
182 | } | ||
183 | |||
184 | /* Signed saturating rounding doubling multiply-accumulate high half, 16-bit */ | ||
185 | -static int16_t do_sqrdmlah_h(int16_t src1, int16_t src2, int16_t src3, | ||
186 | - bool neg, bool round, uint32_t *sat) | ||
187 | +int16_t do_sqrdmlah_h(int16_t src1, int16_t src2, int16_t src3, | ||
188 | + bool neg, bool round, uint32_t *sat) | ||
189 | { | ||
190 | /* Simplify similarly to do_sqrdmlah_b above. */ | ||
191 | int32_t ret = (int32_t)src1 * src2; | ||
192 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_h)(void *vd, void *vn, void *vm, | ||
193 | } | ||
194 | |||
195 | /* Signed saturating rounding doubling multiply-accumulate high half, 32-bit */ | ||
196 | -static int32_t do_sqrdmlah_s(int32_t src1, int32_t src2, int32_t src3, | ||
197 | - bool neg, bool round, uint32_t *sat) | ||
198 | +int32_t do_sqrdmlah_s(int32_t src1, int32_t src2, int32_t src3, | ||
199 | + bool neg, bool round, uint32_t *sat) | ||
200 | { | ||
201 | /* Simplify similarly to do_sqrdmlah_b above. */ | ||
202 | int64_t ret = (int64_t)src1 * src2; | ||
203 | @@ -XXX,XX +XXX,XX @@ static int64_t do_sat128_d(Int128 r) | ||
204 | return ls; | ||
205 | } | ||
206 | |||
207 | -static int64_t do_sqrdmlah_d(int64_t n, int64_t m, int64_t a, | ||
208 | - bool neg, bool round) | ||
209 | +int64_t do_sqrdmlah_d(int64_t n, int64_t m, int64_t a, bool neg, bool round) | ||
210 | { | ||
211 | uint64_t l, h; | ||
212 | Int128 r, t; | ||
213 | -- | ||
214 | 2.20.1 | ||
215 | |||
216 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-39-richard.henderson@linaro.org | ||
7 | Message-Id: <20200417162231.10374-2-steplong@quicinc.com> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/helper-sve.h | 8 ++++++++ | ||
12 | target/arm/sve.decode | 5 +++++ | ||
13 | target/arm/sve_helper.c | 36 ++++++++++++++++++++++++++++++++++++ | ||
14 | target/arm/translate-sve.c | 13 +++++++++++++ | ||
15 | 4 files changed, 62 insertions(+) | ||
16 | |||
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/target/arm/helper-sve.h | ||
20 | +++ b/target/arm/helper-sve.h | ||
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(sve2_uqrshrnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_3(sve2_uqrshrnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_3(sve2_uqrshrnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
24 | |||
25 | +DEF_HELPER_FLAGS_4(sve2_addhnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_addhnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(sve2_addhnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | + | ||
29 | +DEF_HELPER_FLAGS_4(sve2_addhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_addhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_addhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | + | ||
33 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_b, TCG_CALL_NO_RWG, | ||
34 | i32, ptr, ptr, ptr, ptr, i32) | ||
35 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_h, TCG_CALL_NO_RWG, | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ UQSHRNT 01000101 .. 1 ..... 00 1101 ..... ..... @rd_rn_tszimm_shr | ||
41 | UQRSHRNB 01000101 .. 1 ..... 00 1110 ..... ..... @rd_rn_tszimm_shr | ||
42 | UQRSHRNT 01000101 .. 1 ..... 00 1111 ..... ..... @rd_rn_tszimm_shr | ||
43 | |||
44 | +## SVE2 integer add/subtract narrow high part | ||
45 | + | ||
46 | +ADDHNB 01000101 .. 1 ..... 011 000 ..... ..... @rd_rn_rm | ||
47 | +ADDHNT 01000101 .. 1 ..... 011 001 ..... ..... @rd_rn_rm | ||
48 | + | ||
49 | ### SVE2 Character Match | ||
50 | |||
51 | MATCH 01000101 .. 1 ..... 100 ... ..... 0 .... @pd_pg_rn_rm | ||
52 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/target/arm/sve_helper.c | ||
55 | +++ b/target/arm/sve_helper.c | ||
56 | @@ -XXX,XX +XXX,XX @@ DO_SHRNT(sve2_uqrshrnt_d, uint64_t, uint32_t, , H1_4, DO_UQRSHRN_D) | ||
57 | #undef DO_SHRNB | ||
58 | #undef DO_SHRNT | ||
59 | |||
60 | +#define DO_BINOPNB(NAME, TYPEW, TYPEN, SHIFT, OP) \ | ||
61 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
62 | +{ \ | ||
63 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
64 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
65 | + TYPEW nn = *(TYPEW *)(vn + i); \ | ||
66 | + TYPEW mm = *(TYPEW *)(vm + i); \ | ||
67 | + *(TYPEW *)(vd + i) = (TYPEN)OP(nn, mm, SHIFT); \ | ||
68 | + } \ | ||
69 | +} | ||
70 | + | ||
71 | +#define DO_BINOPNT(NAME, TYPEW, TYPEN, SHIFT, HW, HN, OP) \ | ||
72 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
73 | +{ \ | ||
74 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
75 | + for (i = 0; i < opr_sz; i += sizeof(TYPEW)) { \ | ||
76 | + TYPEW nn = *(TYPEW *)(vn + HW(i)); \ | ||
77 | + TYPEW mm = *(TYPEW *)(vm + HW(i)); \ | ||
78 | + *(TYPEN *)(vd + HN(i + sizeof(TYPEN))) = OP(nn, mm, SHIFT); \ | ||
79 | + } \ | ||
80 | +} | ||
81 | + | ||
82 | +#define DO_ADDHN(N, M, SH) ((N + M) >> SH) | ||
83 | + | ||
84 | +DO_BINOPNB(sve2_addhnb_h, uint16_t, uint8_t, 8, DO_ADDHN) | ||
85 | +DO_BINOPNB(sve2_addhnb_s, uint32_t, uint16_t, 16, DO_ADDHN) | ||
86 | +DO_BINOPNB(sve2_addhnb_d, uint64_t, uint32_t, 32, DO_ADDHN) | ||
87 | + | ||
88 | +DO_BINOPNT(sve2_addhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_ADDHN) | ||
89 | +DO_BINOPNT(sve2_addhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_ADDHN) | ||
90 | +DO_BINOPNT(sve2_addhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_ADDHN) | ||
91 | + | ||
92 | +#undef DO_ADDHN | ||
93 | + | ||
94 | +#undef DO_BINOPNB | ||
95 | + | ||
96 | /* Fully general four-operand expander, controlled by a predicate. | ||
97 | */ | ||
98 | #define DO_ZPZZZ(NAME, TYPE, H, OP) \ | ||
99 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
100 | index XXXXXXX..XXXXXXX 100644 | ||
101 | --- a/target/arm/translate-sve.c | ||
102 | +++ b/target/arm/translate-sve.c | ||
103 | @@ -XXX,XX +XXX,XX @@ static bool trans_UQRSHRNT(DisasContext *s, arg_rri_esz *a) | ||
104 | return do_sve2_shr_narrow(s, a, ops); | ||
105 | } | ||
106 | |||
107 | +#define DO_SVE2_ZZZ_NARROW(NAME, name) \ | ||
108 | +static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
109 | +{ \ | ||
110 | + static gen_helper_gvec_3 * const fns[4] = { \ | ||
111 | + NULL, gen_helper_sve2_##name##_h, \ | ||
112 | + gen_helper_sve2_##name##_s, gen_helper_sve2_##name##_d, \ | ||
113 | + }; \ | ||
114 | + return do_sve2_zzz_ool(s, a, fns[a->esz]); \ | ||
115 | +} | ||
116 | + | ||
117 | +DO_SVE2_ZZZ_NARROW(ADDHNB, addhnb) | ||
118 | +DO_SVE2_ZZZ_NARROW(ADDHNT, addhnt) | ||
119 | + | ||
120 | static bool do_sve2_ppzz_flags(DisasContext *s, arg_rprr_esz *a, | ||
121 | gen_helper_gvec_flags_4 *fn) | ||
122 | { | ||
123 | -- | ||
124 | 2.20.1 | ||
125 | |||
126 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-40-richard.henderson@linaro.org | ||
7 | Message-Id: <20200417162231.10374-3-steplong@quicinc.com> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/helper-sve.h | 8 ++++++++ | ||
12 | target/arm/sve.decode | 2 ++ | ||
13 | target/arm/sve_helper.c | 10 ++++++++++ | ||
14 | target/arm/translate-sve.c | 2 ++ | ||
15 | 4 files changed, 22 insertions(+) | ||
16 | |||
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/target/arm/helper-sve.h | ||
20 | +++ b/target/arm/helper-sve.h | ||
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_addhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_4(sve2_addhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_4(sve2_addhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | |||
25 | +DEF_HELPER_FLAGS_4(sve2_raddhnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_raddhnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(sve2_raddhnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | + | ||
29 | +DEF_HELPER_FLAGS_4(sve2_raddhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_raddhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_raddhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | + | ||
33 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_b, TCG_CALL_NO_RWG, | ||
34 | i32, ptr, ptr, ptr, ptr, i32) | ||
35 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_h, TCG_CALL_NO_RWG, | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ UQRSHRNT 01000101 .. 1 ..... 00 1111 ..... ..... @rd_rn_tszimm_shr | ||
41 | |||
42 | ADDHNB 01000101 .. 1 ..... 011 000 ..... ..... @rd_rn_rm | ||
43 | ADDHNT 01000101 .. 1 ..... 011 001 ..... ..... @rd_rn_rm | ||
44 | +RADDHNB 01000101 .. 1 ..... 011 010 ..... ..... @rd_rn_rm | ||
45 | +RADDHNT 01000101 .. 1 ..... 011 011 ..... ..... @rd_rn_rm | ||
46 | |||
47 | ### SVE2 Character Match | ||
48 | |||
49 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/target/arm/sve_helper.c | ||
52 | +++ b/target/arm/sve_helper.c | ||
53 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
54 | } | ||
55 | |||
56 | #define DO_ADDHN(N, M, SH) ((N + M) >> SH) | ||
57 | +#define DO_RADDHN(N, M, SH) ((N + M + ((__typeof(N))1 << (SH - 1))) >> SH) | ||
58 | |||
59 | DO_BINOPNB(sve2_addhnb_h, uint16_t, uint8_t, 8, DO_ADDHN) | ||
60 | DO_BINOPNB(sve2_addhnb_s, uint32_t, uint16_t, 16, DO_ADDHN) | ||
61 | @@ -XXX,XX +XXX,XX @@ DO_BINOPNT(sve2_addhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_ADDHN) | ||
62 | DO_BINOPNT(sve2_addhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_ADDHN) | ||
63 | DO_BINOPNT(sve2_addhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_ADDHN) | ||
64 | |||
65 | +DO_BINOPNB(sve2_raddhnb_h, uint16_t, uint8_t, 8, DO_RADDHN) | ||
66 | +DO_BINOPNB(sve2_raddhnb_s, uint32_t, uint16_t, 16, DO_RADDHN) | ||
67 | +DO_BINOPNB(sve2_raddhnb_d, uint64_t, uint32_t, 32, DO_RADDHN) | ||
68 | + | ||
69 | +DO_BINOPNT(sve2_raddhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_RADDHN) | ||
70 | +DO_BINOPNT(sve2_raddhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_RADDHN) | ||
71 | +DO_BINOPNT(sve2_raddhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_RADDHN) | ||
72 | + | ||
73 | +#undef DO_RADDHN | ||
74 | #undef DO_ADDHN | ||
75 | |||
76 | #undef DO_BINOPNB | ||
77 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
78 | index XXXXXXX..XXXXXXX 100644 | ||
79 | --- a/target/arm/translate-sve.c | ||
80 | +++ b/target/arm/translate-sve.c | ||
81 | @@ -XXX,XX +XXX,XX @@ static bool trans_##NAME(DisasContext *s, arg_rrr_esz *a) \ | ||
82 | |||
83 | DO_SVE2_ZZZ_NARROW(ADDHNB, addhnb) | ||
84 | DO_SVE2_ZZZ_NARROW(ADDHNT, addhnt) | ||
85 | +DO_SVE2_ZZZ_NARROW(RADDHNB, raddhnb) | ||
86 | +DO_SVE2_ZZZ_NARROW(RADDHNT, raddhnt) | ||
87 | |||
88 | static bool do_sve2_ppzz_flags(DisasContext *s, arg_rprr_esz *a, | ||
89 | gen_helper_gvec_flags_4 *fn) | ||
90 | -- | ||
91 | 2.20.1 | ||
92 | |||
93 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-41-richard.henderson@linaro.org | ||
7 | Message-Id: <20200417162231.10374-4-steplong@quicinc.com> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/helper-sve.h | 8 ++++++++ | ||
12 | target/arm/sve.decode | 2 ++ | ||
13 | target/arm/sve_helper.c | 10 ++++++++++ | ||
14 | target/arm/translate-sve.c | 3 +++ | ||
15 | 4 files changed, 23 insertions(+) | ||
16 | |||
17 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/target/arm/helper-sve.h | ||
20 | +++ b/target/arm/helper-sve.h | ||
21 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_raddhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | DEF_HELPER_FLAGS_4(sve2_raddhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_4(sve2_raddhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | |||
25 | +DEF_HELPER_FLAGS_4(sve2_subhnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_subhnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_4(sve2_subhnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | + | ||
29 | +DEF_HELPER_FLAGS_4(sve2_subhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_subhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_subhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | + | ||
33 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_b, TCG_CALL_NO_RWG, | ||
34 | i32, ptr, ptr, ptr, ptr, i32) | ||
35 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_h, TCG_CALL_NO_RWG, | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ ADDHNB 01000101 .. 1 ..... 011 000 ..... ..... @rd_rn_rm | ||
41 | ADDHNT 01000101 .. 1 ..... 011 001 ..... ..... @rd_rn_rm | ||
42 | RADDHNB 01000101 .. 1 ..... 011 010 ..... ..... @rd_rn_rm | ||
43 | RADDHNT 01000101 .. 1 ..... 011 011 ..... ..... @rd_rn_rm | ||
44 | +SUBHNB 01000101 .. 1 ..... 011 100 ..... ..... @rd_rn_rm | ||
45 | +SUBHNT 01000101 .. 1 ..... 011 101 ..... ..... @rd_rn_rm | ||
46 | |||
47 | ### SVE2 Character Match | ||
48 | |||
49 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/target/arm/sve_helper.c | ||
52 | +++ b/target/arm/sve_helper.c | ||
53 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
54 | |||
55 | #define DO_ADDHN(N, M, SH) ((N + M) >> SH) | ||
56 | #define DO_RADDHN(N, M, SH) ((N + M + ((__typeof(N))1 << (SH - 1))) >> SH) | ||
57 | +#define DO_SUBHN(N, M, SH) ((N - M) >> SH) | ||
58 | |||
59 | DO_BINOPNB(sve2_addhnb_h, uint16_t, uint8_t, 8, DO_ADDHN) | ||
60 | DO_BINOPNB(sve2_addhnb_s, uint32_t, uint16_t, 16, DO_ADDHN) | ||
61 | @@ -XXX,XX +XXX,XX @@ DO_BINOPNT(sve2_raddhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_RADDHN) | ||
62 | DO_BINOPNT(sve2_raddhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_RADDHN) | ||
63 | DO_BINOPNT(sve2_raddhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_RADDHN) | ||
64 | |||
65 | +DO_BINOPNB(sve2_subhnb_h, uint16_t, uint8_t, 8, DO_SUBHN) | ||
66 | +DO_BINOPNB(sve2_subhnb_s, uint32_t, uint16_t, 16, DO_SUBHN) | ||
67 | +DO_BINOPNB(sve2_subhnb_d, uint64_t, uint32_t, 32, DO_SUBHN) | ||
68 | + | ||
69 | +DO_BINOPNT(sve2_subhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_SUBHN) | ||
70 | +DO_BINOPNT(sve2_subhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_SUBHN) | ||
71 | +DO_BINOPNT(sve2_subhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_SUBHN) | ||
72 | + | ||
73 | +#undef DO_SUBHN | ||
74 | #undef DO_RADDHN | ||
75 | #undef DO_ADDHN | ||
76 | |||
77 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
78 | index XXXXXXX..XXXXXXX 100644 | ||
79 | --- a/target/arm/translate-sve.c | ||
80 | +++ b/target/arm/translate-sve.c | ||
81 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_NARROW(ADDHNT, addhnt) | ||
82 | DO_SVE2_ZZZ_NARROW(RADDHNB, raddhnb) | ||
83 | DO_SVE2_ZZZ_NARROW(RADDHNT, raddhnt) | ||
84 | |||
85 | +DO_SVE2_ZZZ_NARROW(SUBHNB, subhnb) | ||
86 | +DO_SVE2_ZZZ_NARROW(SUBHNT, subhnt) | ||
87 | + | ||
88 | static bool do_sve2_ppzz_flags(DisasContext *s, arg_rprr_esz *a, | ||
89 | gen_helper_gvec_flags_4 *fn) | ||
90 | { | ||
91 | -- | ||
92 | 2.20.1 | ||
93 | |||
94 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | This completes the section 'SVE2 integer add/subtract narrow high part' | ||
4 | |||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
8 | Message-id: 20210525010358.152808-42-richard.henderson@linaro.org | ||
9 | Message-Id: <20200417162231.10374-5-steplong@quicinc.com> | ||
10 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
11 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
12 | --- | ||
13 | target/arm/helper-sve.h | 8 ++++++++ | ||
14 | target/arm/sve.decode | 2 ++ | ||
15 | target/arm/sve_helper.c | 10 ++++++++++ | ||
16 | target/arm/translate-sve.c | 2 ++ | ||
17 | 4 files changed, 22 insertions(+) | ||
18 | |||
19 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/target/arm/helper-sve.h | ||
22 | +++ b/target/arm/helper-sve.h | ||
23 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_subhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | DEF_HELPER_FLAGS_4(sve2_subhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | DEF_HELPER_FLAGS_4(sve2_subhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | |||
27 | +DEF_HELPER_FLAGS_4(sve2_rsubhnb_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_rsubhnb_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_4(sve2_rsubhnb_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_4(sve2_rsubhnt_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_rsubhnt_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(sve2_rsubhnt_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
34 | + | ||
35 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_b, TCG_CALL_NO_RWG, | ||
36 | i32, ptr, ptr, ptr, ptr, i32) | ||
37 | DEF_HELPER_FLAGS_5(sve2_match_ppzz_h, TCG_CALL_NO_RWG, | ||
38 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
39 | index XXXXXXX..XXXXXXX 100644 | ||
40 | --- a/target/arm/sve.decode | ||
41 | +++ b/target/arm/sve.decode | ||
42 | @@ -XXX,XX +XXX,XX @@ RADDHNB 01000101 .. 1 ..... 011 010 ..... ..... @rd_rn_rm | ||
43 | RADDHNT 01000101 .. 1 ..... 011 011 ..... ..... @rd_rn_rm | ||
44 | SUBHNB 01000101 .. 1 ..... 011 100 ..... ..... @rd_rn_rm | ||
45 | SUBHNT 01000101 .. 1 ..... 011 101 ..... ..... @rd_rn_rm | ||
46 | +RSUBHNB 01000101 .. 1 ..... 011 110 ..... ..... @rd_rn_rm | ||
47 | +RSUBHNT 01000101 .. 1 ..... 011 111 ..... ..... @rd_rn_rm | ||
48 | |||
49 | ### SVE2 Character Match | ||
50 | |||
51 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
52 | index XXXXXXX..XXXXXXX 100644 | ||
53 | --- a/target/arm/sve_helper.c | ||
54 | +++ b/target/arm/sve_helper.c | ||
55 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
56 | #define DO_ADDHN(N, M, SH) ((N + M) >> SH) | ||
57 | #define DO_RADDHN(N, M, SH) ((N + M + ((__typeof(N))1 << (SH - 1))) >> SH) | ||
58 | #define DO_SUBHN(N, M, SH) ((N - M) >> SH) | ||
59 | +#define DO_RSUBHN(N, M, SH) ((N - M + ((__typeof(N))1 << (SH - 1))) >> SH) | ||
60 | |||
61 | DO_BINOPNB(sve2_addhnb_h, uint16_t, uint8_t, 8, DO_ADDHN) | ||
62 | DO_BINOPNB(sve2_addhnb_s, uint32_t, uint16_t, 16, DO_ADDHN) | ||
63 | @@ -XXX,XX +XXX,XX @@ DO_BINOPNT(sve2_subhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_SUBHN) | ||
64 | DO_BINOPNT(sve2_subhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_SUBHN) | ||
65 | DO_BINOPNT(sve2_subhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_SUBHN) | ||
66 | |||
67 | +DO_BINOPNB(sve2_rsubhnb_h, uint16_t, uint8_t, 8, DO_RSUBHN) | ||
68 | +DO_BINOPNB(sve2_rsubhnb_s, uint32_t, uint16_t, 16, DO_RSUBHN) | ||
69 | +DO_BINOPNB(sve2_rsubhnb_d, uint64_t, uint32_t, 32, DO_RSUBHN) | ||
70 | + | ||
71 | +DO_BINOPNT(sve2_rsubhnt_h, uint16_t, uint8_t, 8, H1_2, H1, DO_RSUBHN) | ||
72 | +DO_BINOPNT(sve2_rsubhnt_s, uint32_t, uint16_t, 16, H1_4, H1_2, DO_RSUBHN) | ||
73 | +DO_BINOPNT(sve2_rsubhnt_d, uint64_t, uint32_t, 32, , H1_4, DO_RSUBHN) | ||
74 | + | ||
75 | +#undef DO_RSUBHN | ||
76 | #undef DO_SUBHN | ||
77 | #undef DO_RADDHN | ||
78 | #undef DO_ADDHN | ||
79 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/translate-sve.c | ||
82 | +++ b/target/arm/translate-sve.c | ||
83 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_NARROW(RADDHNT, raddhnt) | ||
84 | |||
85 | DO_SVE2_ZZZ_NARROW(SUBHNB, subhnb) | ||
86 | DO_SVE2_ZZZ_NARROW(SUBHNT, subhnt) | ||
87 | +DO_SVE2_ZZZ_NARROW(RSUBHNB, rsubhnb) | ||
88 | +DO_SVE2_ZZZ_NARROW(RSUBHNT, rsubhnt) | ||
89 | |||
90 | static bool do_sve2_ppzz_flags(DisasContext *s, arg_rprr_esz *a, | ||
91 | gen_helper_gvec_flags_4 *fn) | ||
92 | -- | ||
93 | 2.20.1 | ||
94 | |||
95 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | In addition, use the same vector generator interface for AdvSIMD. | ||
4 | This fixes a bug in which the AdvSIMD insn failed to clear the | ||
5 | high bits of the SVE register. | ||
6 | |||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-44-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper-sve.h | 4 ++ | ||
13 | target/arm/helper.h | 2 + | ||
14 | target/arm/translate-a64.h | 3 ++ | ||
15 | target/arm/sve.decode | 4 ++ | ||
16 | target/arm/sve_helper.c | 39 ++++++++++++++ | ||
17 | target/arm/translate-a64.c | 25 ++------- | ||
18 | target/arm/translate-sve.c | 104 +++++++++++++++++++++++++++++++++++++ | ||
19 | target/arm/vec_helper.c | 12 +++++ | ||
20 | 8 files changed, 172 insertions(+), 21 deletions(-) | ||
21 | |||
22 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
23 | index XXXXXXX..XXXXXXX 100644 | ||
24 | --- a/target/arm/helper-sve.h | ||
25 | +++ b/target/arm/helper-sve.h | ||
26 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_histcnt_d, TCG_CALL_NO_RWG, | ||
27 | |||
28 | DEF_HELPER_FLAGS_4(sve2_histseg, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | |||
30 | +DEF_HELPER_FLAGS_4(sve2_xar_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_4(sve2_xar_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_xar_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | + | ||
34 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_h, TCG_CALL_NO_RWG, | ||
35 | void, ptr, ptr, ptr, ptr, ptr, i32) | ||
36 | DEF_HELPER_FLAGS_6(sve2_faddp_zpzz_s, TCG_CALL_NO_RWG, | ||
37 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
38 | index XXXXXXX..XXXXXXX 100644 | ||
39 | --- a/target/arm/helper.h | ||
40 | +++ b/target/arm/helper.h | ||
41 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(neon_sqrdmulh_h, TCG_CALL_NO_RWG, | ||
42 | DEF_HELPER_FLAGS_5(neon_sqrdmulh_s, TCG_CALL_NO_RWG, | ||
43 | void, ptr, ptr, ptr, ptr, i32) | ||
44 | |||
45 | +DEF_HELPER_FLAGS_4(gvec_xar_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
46 | + | ||
47 | #ifdef TARGET_AARCH64 | ||
48 | #include "helper-a64.h" | ||
49 | #include "helper-sve.h" | ||
50 | diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/target/arm/translate-a64.h | ||
53 | +++ b/target/arm/translate-a64.h | ||
54 | @@ -XXX,XX +XXX,XX @@ bool disas_sve(DisasContext *, uint32_t); | ||
55 | |||
56 | void gen_gvec_rax1(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs, | ||
57 | uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz); | ||
58 | +void gen_gvec_xar(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs, | ||
59 | + uint32_t rm_ofs, int64_t shift, | ||
60 | + uint32_t opr_sz, uint32_t max_sz); | ||
61 | |||
62 | #endif /* TARGET_ARM_TRANSLATE_A64_H */ | ||
63 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
64 | index XXXXXXX..XXXXXXX 100644 | ||
65 | --- a/target/arm/sve.decode | ||
66 | +++ b/target/arm/sve.decode | ||
67 | @@ -XXX,XX +XXX,XX @@ | ||
68 | &rr_dbm rd rn dbm | ||
69 | &rrri rd rn rm imm | ||
70 | &rri_esz rd rn imm esz | ||
71 | +&rrri_esz rd rn rm imm esz | ||
72 | &rrr_esz rd rn rm esz | ||
73 | &rpr_esz rd pg rn esz | ||
74 | &rpr_s rd pg rn s | ||
75 | @@ -XXX,XX +XXX,XX @@ ORR_zzz 00000100 01 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
76 | EOR_zzz 00000100 10 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
77 | BIC_zzz 00000100 11 1 ..... 001 100 ..... ..... @rd_rn_rm_e0 | ||
78 | |||
79 | +XAR 00000100 .. 1 ..... 001 101 rm:5 rd:5 &rrri_esz \ | ||
80 | + rn=%reg_movprfx esz=%tszimm16_esz imm=%tszimm16_shr | ||
81 | + | ||
82 | # SVE2 bitwise ternary operations | ||
83 | EOR3 00000100 00 1 ..... 001 110 ..... ..... @rdn_ra_rm_e0 | ||
84 | BSL 00000100 00 1 ..... 001 111 ..... ..... @rdn_ra_rm_e0 | ||
85 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
86 | index XXXXXXX..XXXXXXX 100644 | ||
87 | --- a/target/arm/sve_helper.c | ||
88 | +++ b/target/arm/sve_helper.c | ||
89 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_histseg)(void *vd, void *vn, void *vm, uint32_t desc) | ||
90 | *(uint64_t *)(vd + i + 8) = out1; | ||
91 | } | ||
92 | } | ||
93 | + | ||
94 | +void HELPER(sve2_xar_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
95 | +{ | ||
96 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
97 | + int shr = simd_data(desc); | ||
98 | + int shl = 8 - shr; | ||
99 | + uint64_t mask = dup_const(MO_8, 0xff >> shr); | ||
100 | + uint64_t *d = vd, *n = vn, *m = vm; | ||
101 | + | ||
102 | + for (i = 0; i < opr_sz; ++i) { | ||
103 | + uint64_t t = n[i] ^ m[i]; | ||
104 | + d[i] = ((t >> shr) & mask) | ((t << shl) & ~mask); | ||
105 | + } | ||
106 | +} | ||
107 | + | ||
108 | +void HELPER(sve2_xar_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
109 | +{ | ||
110 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
111 | + int shr = simd_data(desc); | ||
112 | + int shl = 16 - shr; | ||
113 | + uint64_t mask = dup_const(MO_16, 0xffff >> shr); | ||
114 | + uint64_t *d = vd, *n = vn, *m = vm; | ||
115 | + | ||
116 | + for (i = 0; i < opr_sz; ++i) { | ||
117 | + uint64_t t = n[i] ^ m[i]; | ||
118 | + d[i] = ((t >> shr) & mask) | ((t << shl) & ~mask); | ||
119 | + } | ||
120 | +} | ||
121 | + | ||
122 | +void HELPER(sve2_xar_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
123 | +{ | ||
124 | + intptr_t i, opr_sz = simd_oprsz(desc) / 4; | ||
125 | + int shr = simd_data(desc); | ||
126 | + uint32_t *d = vd, *n = vn, *m = vm; | ||
127 | + | ||
128 | + for (i = 0; i < opr_sz; ++i) { | ||
129 | + d[i] = ror32(n[i] ^ m[i], shr); | ||
130 | + } | ||
131 | +} | ||
132 | diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c | ||
133 | index XXXXXXX..XXXXXXX 100644 | ||
134 | --- a/target/arm/translate-a64.c | ||
135 | +++ b/target/arm/translate-a64.c | ||
136 | @@ -XXX,XX +XXX,XX @@ static void disas_crypto_xar(DisasContext *s, uint32_t insn) | ||
137 | int imm6 = extract32(insn, 10, 6); | ||
138 | int rn = extract32(insn, 5, 5); | ||
139 | int rd = extract32(insn, 0, 5); | ||
140 | - TCGv_i64 tcg_op1, tcg_op2, tcg_res[2]; | ||
141 | - int pass; | ||
142 | |||
143 | if (!dc_isar_feature(aa64_sha3, s)) { | ||
144 | unallocated_encoding(s); | ||
145 | @@ -XXX,XX +XXX,XX @@ static void disas_crypto_xar(DisasContext *s, uint32_t insn) | ||
146 | return; | ||
147 | } | ||
148 | |||
149 | - tcg_op1 = tcg_temp_new_i64(); | ||
150 | - tcg_op2 = tcg_temp_new_i64(); | ||
151 | - tcg_res[0] = tcg_temp_new_i64(); | ||
152 | - tcg_res[1] = tcg_temp_new_i64(); | ||
153 | - | ||
154 | - for (pass = 0; pass < 2; pass++) { | ||
155 | - read_vec_element(s, tcg_op1, rn, pass, MO_64); | ||
156 | - read_vec_element(s, tcg_op2, rm, pass, MO_64); | ||
157 | - | ||
158 | - tcg_gen_xor_i64(tcg_res[pass], tcg_op1, tcg_op2); | ||
159 | - tcg_gen_rotri_i64(tcg_res[pass], tcg_res[pass], imm6); | ||
160 | - } | ||
161 | - write_vec_element(s, tcg_res[0], rd, 0, MO_64); | ||
162 | - write_vec_element(s, tcg_res[1], rd, 1, MO_64); | ||
163 | - | ||
164 | - tcg_temp_free_i64(tcg_op1); | ||
165 | - tcg_temp_free_i64(tcg_op2); | ||
166 | - tcg_temp_free_i64(tcg_res[0]); | ||
167 | - tcg_temp_free_i64(tcg_res[1]); | ||
168 | + gen_gvec_xar(MO_64, vec_full_reg_offset(s, rd), | ||
169 | + vec_full_reg_offset(s, rn), | ||
170 | + vec_full_reg_offset(s, rm), imm6, 16, | ||
171 | + vec_full_reg_size(s)); | ||
172 | } | ||
173 | |||
174 | /* Crypto three-reg imm2 | ||
175 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
176 | index XXXXXXX..XXXXXXX 100644 | ||
177 | --- a/target/arm/translate-sve.c | ||
178 | +++ b/target/arm/translate-sve.c | ||
179 | @@ -XXX,XX +XXX,XX @@ static bool trans_BIC_zzz(DisasContext *s, arg_rrr_esz *a) | ||
180 | return do_zzz_fn(s, a, tcg_gen_gvec_andc); | ||
181 | } | ||
182 | |||
183 | +static void gen_xar8_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, int64_t sh) | ||
184 | +{ | ||
185 | + TCGv_i64 t = tcg_temp_new_i64(); | ||
186 | + uint64_t mask = dup_const(MO_8, 0xff >> sh); | ||
187 | + | ||
188 | + tcg_gen_xor_i64(t, n, m); | ||
189 | + tcg_gen_shri_i64(d, t, sh); | ||
190 | + tcg_gen_shli_i64(t, t, 8 - sh); | ||
191 | + tcg_gen_andi_i64(d, d, mask); | ||
192 | + tcg_gen_andi_i64(t, t, ~mask); | ||
193 | + tcg_gen_or_i64(d, d, t); | ||
194 | + tcg_temp_free_i64(t); | ||
195 | +} | ||
196 | + | ||
197 | +static void gen_xar16_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, int64_t sh) | ||
198 | +{ | ||
199 | + TCGv_i64 t = tcg_temp_new_i64(); | ||
200 | + uint64_t mask = dup_const(MO_16, 0xffff >> sh); | ||
201 | + | ||
202 | + tcg_gen_xor_i64(t, n, m); | ||
203 | + tcg_gen_shri_i64(d, t, sh); | ||
204 | + tcg_gen_shli_i64(t, t, 16 - sh); | ||
205 | + tcg_gen_andi_i64(d, d, mask); | ||
206 | + tcg_gen_andi_i64(t, t, ~mask); | ||
207 | + tcg_gen_or_i64(d, d, t); | ||
208 | + tcg_temp_free_i64(t); | ||
209 | +} | ||
210 | + | ||
211 | +static void gen_xar_i32(TCGv_i32 d, TCGv_i32 n, TCGv_i32 m, int32_t sh) | ||
212 | +{ | ||
213 | + tcg_gen_xor_i32(d, n, m); | ||
214 | + tcg_gen_rotri_i32(d, d, sh); | ||
215 | +} | ||
216 | + | ||
217 | +static void gen_xar_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, int64_t sh) | ||
218 | +{ | ||
219 | + tcg_gen_xor_i64(d, n, m); | ||
220 | + tcg_gen_rotri_i64(d, d, sh); | ||
221 | +} | ||
222 | + | ||
223 | +static void gen_xar_vec(unsigned vece, TCGv_vec d, TCGv_vec n, | ||
224 | + TCGv_vec m, int64_t sh) | ||
225 | +{ | ||
226 | + tcg_gen_xor_vec(vece, d, n, m); | ||
227 | + tcg_gen_rotri_vec(vece, d, d, sh); | ||
228 | +} | ||
229 | + | ||
230 | +void gen_gvec_xar(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs, | ||
231 | + uint32_t rm_ofs, int64_t shift, | ||
232 | + uint32_t opr_sz, uint32_t max_sz) | ||
233 | +{ | ||
234 | + static const TCGOpcode vecop[] = { INDEX_op_rotli_vec, 0 }; | ||
235 | + static const GVecGen3i ops[4] = { | ||
236 | + { .fni8 = gen_xar8_i64, | ||
237 | + .fniv = gen_xar_vec, | ||
238 | + .fno = gen_helper_sve2_xar_b, | ||
239 | + .opt_opc = vecop, | ||
240 | + .vece = MO_8 }, | ||
241 | + { .fni8 = gen_xar16_i64, | ||
242 | + .fniv = gen_xar_vec, | ||
243 | + .fno = gen_helper_sve2_xar_h, | ||
244 | + .opt_opc = vecop, | ||
245 | + .vece = MO_16 }, | ||
246 | + { .fni4 = gen_xar_i32, | ||
247 | + .fniv = gen_xar_vec, | ||
248 | + .fno = gen_helper_sve2_xar_s, | ||
249 | + .opt_opc = vecop, | ||
250 | + .vece = MO_32 }, | ||
251 | + { .fni8 = gen_xar_i64, | ||
252 | + .fniv = gen_xar_vec, | ||
253 | + .fno = gen_helper_gvec_xar_d, | ||
254 | + .opt_opc = vecop, | ||
255 | + .vece = MO_64 } | ||
256 | + }; | ||
257 | + int esize = 8 << vece; | ||
258 | + | ||
259 | + /* The SVE2 range is 1 .. esize; the AdvSIMD range is 0 .. esize-1. */ | ||
260 | + tcg_debug_assert(shift >= 0); | ||
261 | + tcg_debug_assert(shift <= esize); | ||
262 | + shift &= esize - 1; | ||
263 | + | ||
264 | + if (shift == 0) { | ||
265 | + /* xar with no rotate devolves to xor. */ | ||
266 | + tcg_gen_gvec_xor(vece, rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz); | ||
267 | + } else { | ||
268 | + tcg_gen_gvec_3i(rd_ofs, rn_ofs, rm_ofs, opr_sz, max_sz, | ||
269 | + shift, &ops[vece]); | ||
270 | + } | ||
271 | +} | ||
272 | + | ||
273 | +static bool trans_XAR(DisasContext *s, arg_rrri_esz *a) | ||
274 | +{ | ||
275 | + if (a->esz < 0 || !dc_isar_feature(aa64_sve2, s)) { | ||
276 | + return false; | ||
277 | + } | ||
278 | + if (sve_access_check(s)) { | ||
279 | + unsigned vsz = vec_full_reg_size(s); | ||
280 | + gen_gvec_xar(a->esz, vec_full_reg_offset(s, a->rd), | ||
281 | + vec_full_reg_offset(s, a->rn), | ||
282 | + vec_full_reg_offset(s, a->rm), a->imm, vsz, vsz); | ||
283 | + } | ||
284 | + return true; | ||
285 | +} | ||
286 | + | ||
287 | static bool do_sve2_zzzz_fn(DisasContext *s, arg_rrrr_esz *a, GVecGen4Fn *fn) | ||
288 | { | ||
289 | if (!dc_isar_feature(aa64_sve2, s)) { | ||
290 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
291 | index XXXXXXX..XXXXXXX 100644 | ||
292 | --- a/target/arm/vec_helper.c | ||
293 | +++ b/target/arm/vec_helper.c | ||
294 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_umulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
295 | } | ||
296 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
297 | } | ||
298 | + | ||
299 | +void HELPER(gvec_xar_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
300 | +{ | ||
301 | + intptr_t i, opr_sz = simd_oprsz(desc) / 8; | ||
302 | + int shr = simd_data(desc); | ||
303 | + uint64_t *d = vd, *n = vn, *m = vm; | ||
304 | + | ||
305 | + for (i = 0; i < opr_sz; ++i) { | ||
306 | + d[i] = ror64(n[i] ^ m[i], shr); | ||
307 | + } | ||
308 | + clear_tail(d, opr_sz * 8, simd_maxsz(desc)); | ||
309 | +} | ||
310 | -- | ||
311 | 2.20.1 | ||
312 | |||
313 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Add decoding logic for SVE2 64-bit/32-bit scatter non-temporal | ||
4 | store insns. | ||
5 | |||
6 | 64-bit | ||
7 | * STNT1B (vector plus scalar) | ||
8 | * STNT1H (vector plus scalar) | ||
9 | * STNT1W (vector plus scalar) | ||
10 | * STNT1D (vector plus scalar) | ||
11 | |||
12 | 32-bit | ||
13 | * STNT1B (vector plus scalar) | ||
14 | * STNT1H (vector plus scalar) | ||
15 | * STNT1W (vector plus scalar) | ||
16 | |||
17 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
18 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
19 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
20 | Message-id: 20210525010358.152808-45-richard.henderson@linaro.org | ||
21 | Message-Id: <20200422141553.8037-1-steplong@quicinc.com> | ||
22 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
23 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
24 | --- | ||
25 | target/arm/sve.decode | 10 ++++++++++ | ||
26 | target/arm/translate-sve.c | 8 ++++++++ | ||
27 | 2 files changed, 18 insertions(+) | ||
28 | |||
29 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
30 | index XXXXXXX..XXXXXXX 100644 | ||
31 | --- a/target/arm/sve.decode | ||
32 | +++ b/target/arm/sve.decode | ||
33 | @@ -XXX,XX +XXX,XX @@ UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | ||
34 | |||
35 | CMLA_zzzz 01000100 esz:2 0 rm:5 0010 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
36 | SQRDCMLAH_zzzz 01000100 esz:2 0 rm:5 0011 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
37 | + | ||
38 | +### SVE2 Memory Store Group | ||
39 | + | ||
40 | +# SVE2 64-bit scatter non-temporal store (vector plus scalar) | ||
41 | +STNT1_zprz 1110010 .. 00 ..... 001 ... ..... ..... \ | ||
42 | + @rprr_scatter_store xs=2 esz=3 scale=0 | ||
43 | + | ||
44 | +# SVE2 32-bit scatter non-temporal store (vector plus scalar) | ||
45 | +STNT1_zprz 1110010 .. 10 ..... 001 ... ..... ..... \ | ||
46 | + @rprr_scatter_store xs=0 esz=2 scale=0 | ||
47 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/target/arm/translate-sve.c | ||
50 | +++ b/target/arm/translate-sve.c | ||
51 | @@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a) | ||
52 | return true; | ||
53 | } | ||
54 | |||
55 | +static bool trans_STNT1_zprz(DisasContext *s, arg_ST1_zprz *a) | ||
56 | +{ | ||
57 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
58 | + return false; | ||
59 | + } | ||
60 | + return trans_ST1_zprz(s, a); | ||
61 | +} | ||
62 | + | ||
63 | /* | ||
64 | * Prefetches | ||
65 | */ | ||
66 | -- | ||
67 | 2.20.1 | ||
68 | |||
69 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Add decoding logic for SVE2 64-bit/32-bit gather non-temporal | ||
4 | load insns. | ||
5 | |||
6 | 64-bit | ||
7 | * LDNT1SB | ||
8 | * LDNT1B (vector plus scalar) | ||
9 | * LDNT1SH | ||
10 | * LDNT1H (vector plus scalar) | ||
11 | * LDNT1SW | ||
12 | * LDNT1W (vector plus scalar) | ||
13 | * LDNT1D (vector plus scalar) | ||
14 | |||
15 | 32-bit | ||
16 | * LDNT1SB | ||
17 | * LDNT1B (vector plus scalar) | ||
18 | * LDNT1SH | ||
19 | * LDNT1H (vector plus scalar) | ||
20 | * LDNT1W (vector plus scalar) | ||
21 | |||
22 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
23 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
24 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
25 | Message-id: 20210525010358.152808-46-richard.henderson@linaro.org | ||
26 | Message-Id: <20200422152343.12493-1-steplong@quicinc.com> | ||
27 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
28 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
29 | --- | ||
30 | target/arm/sve.decode | 11 +++++++++++ | ||
31 | target/arm/translate-sve.c | 8 ++++++++ | ||
32 | 2 files changed, 19 insertions(+) | ||
33 | |||
34 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
35 | index XXXXXXX..XXXXXXX 100644 | ||
36 | --- a/target/arm/sve.decode | ||
37 | +++ b/target/arm/sve.decode | ||
38 | @@ -XXX,XX +XXX,XX @@ UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | ||
39 | CMLA_zzzz 01000100 esz:2 0 rm:5 0010 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
40 | SQRDCMLAH_zzzz 01000100 esz:2 0 rm:5 0011 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
41 | |||
42 | +### SVE2 Memory Gather Load Group | ||
43 | + | ||
44 | +# SVE2 64-bit gather non-temporal load | ||
45 | +# (scalar plus unpacked 32-bit unscaled offsets) | ||
46 | +LDNT1_zprz 1100010 msz:2 00 rm:5 1 u:1 0 pg:3 rn:5 rd:5 \ | ||
47 | + &rprr_gather_load xs=0 esz=3 scale=0 ff=0 | ||
48 | + | ||
49 | +# SVE2 32-bit gather non-temporal load (scalar plus 32-bit unscaled offsets) | ||
50 | +LDNT1_zprz 1000010 msz:2 00 rm:5 10 u:1 pg:3 rn:5 rd:5 \ | ||
51 | + &rprr_gather_load xs=0 esz=2 scale=0 ff=0 | ||
52 | + | ||
53 | ### SVE2 Memory Store Group | ||
54 | |||
55 | # SVE2 64-bit scatter non-temporal store (vector plus scalar) | ||
56 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
57 | index XXXXXXX..XXXXXXX 100644 | ||
58 | --- a/target/arm/translate-sve.c | ||
59 | +++ b/target/arm/translate-sve.c | ||
60 | @@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a) | ||
61 | return true; | ||
62 | } | ||
63 | |||
64 | +static bool trans_LDNT1_zprz(DisasContext *s, arg_LD1_zprz *a) | ||
65 | +{ | ||
66 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
67 | + return false; | ||
68 | + } | ||
69 | + return trans_LD1_zprz(s, a); | ||
70 | +} | ||
71 | + | ||
72 | /* Indexed by [mte][be][xs][msz]. */ | ||
73 | static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = { | ||
74 | { /* MTE Inactive */ | ||
75 | -- | ||
76 | 2.20.1 | ||
77 | |||
78 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-47-richard.henderson@linaro.org | ||
7 | Message-Id: <20200422165503.13511-1-steplong@quicinc.com> | ||
8 | [rth: Fix indexing in helpers, expand macro to straight functions.] | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/cpu.h | 10 ++++++ | ||
13 | target/arm/helper-sve.h | 3 ++ | ||
14 | target/arm/sve.decode | 4 +++ | ||
15 | target/arm/sve_helper.c | 74 ++++++++++++++++++++++++++++++++++++++ | ||
16 | target/arm/translate-sve.c | 34 ++++++++++++++++++ | ||
17 | 5 files changed, 125 insertions(+) | ||
18 | |||
19 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/target/arm/cpu.h | ||
22 | +++ b/target/arm/cpu.h | ||
23 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2_bitperm(const ARMISARegisters *id) | ||
24 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, BITPERM) != 0; | ||
25 | } | ||
26 | |||
27 | +static inline bool isar_feature_aa64_sve_f32mm(const ARMISARegisters *id) | ||
28 | +{ | ||
29 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, F32MM) != 0; | ||
30 | +} | ||
31 | + | ||
32 | +static inline bool isar_feature_aa64_sve_f64mm(const ARMISARegisters *id) | ||
33 | +{ | ||
34 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, F64MM) != 0; | ||
35 | +} | ||
36 | + | ||
37 | /* | ||
38 | * Feature tests for "does this exist in either 32-bit or 64-bit?" | ||
39 | */ | ||
40 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
41 | index XXXXXXX..XXXXXXX 100644 | ||
42 | --- a/target/arm/helper-sve.h | ||
43 | +++ b/target/arm/helper-sve.h | ||
44 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_s, TCG_CALL_NO_RWG, | ||
45 | void, ptr, ptr, ptr, ptr, i32) | ||
46 | DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_d, TCG_CALL_NO_RWG, | ||
47 | void, ptr, ptr, ptr, ptr, i32) | ||
48 | + | ||
49 | +DEF_HELPER_FLAGS_6(fmmla_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, ptr, i32) | ||
50 | +DEF_HELPER_FLAGS_6(fmmla_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, ptr, i32) | ||
51 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
52 | index XXXXXXX..XXXXXXX 100644 | ||
53 | --- a/target/arm/sve.decode | ||
54 | +++ b/target/arm/sve.decode | ||
55 | @@ -XXX,XX +XXX,XX @@ UMLSLT_zzzw 01000100 .. 0 ..... 010 111 ..... ..... @rda_rn_rm | ||
56 | CMLA_zzzz 01000100 esz:2 0 rm:5 0010 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
57 | SQRDCMLAH_zzzz 01000100 esz:2 0 rm:5 0011 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
58 | |||
59 | +### SVE2 floating point matrix multiply accumulate | ||
60 | + | ||
61 | +FMMLA 01100100 .. 1 ..... 111001 ..... ..... @rda_rn_rm | ||
62 | + | ||
63 | ### SVE2 Memory Gather Load Group | ||
64 | |||
65 | # SVE2 64-bit gather non-temporal load | ||
66 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
67 | index XXXXXXX..XXXXXXX 100644 | ||
68 | --- a/target/arm/sve_helper.c | ||
69 | +++ b/target/arm/sve_helper.c | ||
70 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_xar_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
71 | d[i] = ror32(n[i] ^ m[i], shr); | ||
72 | } | ||
73 | } | ||
74 | + | ||
75 | +void HELPER(fmmla_s)(void *vd, void *vn, void *vm, void *va, | ||
76 | + void *status, uint32_t desc) | ||
77 | +{ | ||
78 | + intptr_t s, opr_sz = simd_oprsz(desc) / (sizeof(float32) * 4); | ||
79 | + | ||
80 | + for (s = 0; s < opr_sz; ++s) { | ||
81 | + float32 *n = vn + s * sizeof(float32) * 4; | ||
82 | + float32 *m = vm + s * sizeof(float32) * 4; | ||
83 | + float32 *a = va + s * sizeof(float32) * 4; | ||
84 | + float32 *d = vd + s * sizeof(float32) * 4; | ||
85 | + float32 n00 = n[H4(0)], n01 = n[H4(1)]; | ||
86 | + float32 n10 = n[H4(2)], n11 = n[H4(3)]; | ||
87 | + float32 m00 = m[H4(0)], m01 = m[H4(1)]; | ||
88 | + float32 m10 = m[H4(2)], m11 = m[H4(3)]; | ||
89 | + float32 p0, p1; | ||
90 | + | ||
91 | + /* i = 0, j = 0 */ | ||
92 | + p0 = float32_mul(n00, m00, status); | ||
93 | + p1 = float32_mul(n01, m01, status); | ||
94 | + d[H4(0)] = float32_add(a[H4(0)], float32_add(p0, p1, status), status); | ||
95 | + | ||
96 | + /* i = 0, j = 1 */ | ||
97 | + p0 = float32_mul(n00, m10, status); | ||
98 | + p1 = float32_mul(n01, m11, status); | ||
99 | + d[H4(1)] = float32_add(a[H4(1)], float32_add(p0, p1, status), status); | ||
100 | + | ||
101 | + /* i = 1, j = 0 */ | ||
102 | + p0 = float32_mul(n10, m00, status); | ||
103 | + p1 = float32_mul(n11, m01, status); | ||
104 | + d[H4(2)] = float32_add(a[H4(2)], float32_add(p0, p1, status), status); | ||
105 | + | ||
106 | + /* i = 1, j = 1 */ | ||
107 | + p0 = float32_mul(n10, m10, status); | ||
108 | + p1 = float32_mul(n11, m11, status); | ||
109 | + d[H4(3)] = float32_add(a[H4(3)], float32_add(p0, p1, status), status); | ||
110 | + } | ||
111 | +} | ||
112 | + | ||
113 | +void HELPER(fmmla_d)(void *vd, void *vn, void *vm, void *va, | ||
114 | + void *status, uint32_t desc) | ||
115 | +{ | ||
116 | + intptr_t s, opr_sz = simd_oprsz(desc) / (sizeof(float64) * 4); | ||
117 | + | ||
118 | + for (s = 0; s < opr_sz; ++s) { | ||
119 | + float64 *n = vn + s * sizeof(float64) * 4; | ||
120 | + float64 *m = vm + s * sizeof(float64) * 4; | ||
121 | + float64 *a = va + s * sizeof(float64) * 4; | ||
122 | + float64 *d = vd + s * sizeof(float64) * 4; | ||
123 | + float64 n00 = n[0], n01 = n[1], n10 = n[2], n11 = n[3]; | ||
124 | + float64 m00 = m[0], m01 = m[1], m10 = m[2], m11 = m[3]; | ||
125 | + float64 p0, p1; | ||
126 | + | ||
127 | + /* i = 0, j = 0 */ | ||
128 | + p0 = float64_mul(n00, m00, status); | ||
129 | + p1 = float64_mul(n01, m01, status); | ||
130 | + d[0] = float64_add(a[0], float64_add(p0, p1, status), status); | ||
131 | + | ||
132 | + /* i = 0, j = 1 */ | ||
133 | + p0 = float64_mul(n00, m10, status); | ||
134 | + p1 = float64_mul(n01, m11, status); | ||
135 | + d[1] = float64_add(a[1], float64_add(p0, p1, status), status); | ||
136 | + | ||
137 | + /* i = 1, j = 0 */ | ||
138 | + p0 = float64_mul(n10, m00, status); | ||
139 | + p1 = float64_mul(n11, m01, status); | ||
140 | + d[2] = float64_add(a[2], float64_add(p0, p1, status), status); | ||
141 | + | ||
142 | + /* i = 1, j = 1 */ | ||
143 | + p0 = float64_mul(n10, m10, status); | ||
144 | + p1 = float64_mul(n11, m11, status); | ||
145 | + d[3] = float64_add(a[3], float64_add(p0, p1, status), status); | ||
146 | + } | ||
147 | +} | ||
148 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
149 | index XXXXXXX..XXXXXXX 100644 | ||
150 | --- a/target/arm/translate-sve.c | ||
151 | +++ b/target/arm/translate-sve.c | ||
152 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_ZPZZ_FP(FMINP, fminp) | ||
153 | * SVE Integer Multiply-Add (unpredicated) | ||
154 | */ | ||
155 | |||
156 | +static bool trans_FMMLA(DisasContext *s, arg_rrrr_esz *a) | ||
157 | +{ | ||
158 | + gen_helper_gvec_4_ptr *fn; | ||
159 | + | ||
160 | + switch (a->esz) { | ||
161 | + case MO_32: | ||
162 | + if (!dc_isar_feature(aa64_sve_f32mm, s)) { | ||
163 | + return false; | ||
164 | + } | ||
165 | + fn = gen_helper_fmmla_s; | ||
166 | + break; | ||
167 | + case MO_64: | ||
168 | + if (!dc_isar_feature(aa64_sve_f64mm, s)) { | ||
169 | + return false; | ||
170 | + } | ||
171 | + fn = gen_helper_fmmla_d; | ||
172 | + break; | ||
173 | + default: | ||
174 | + return false; | ||
175 | + } | ||
176 | + | ||
177 | + if (sve_access_check(s)) { | ||
178 | + unsigned vsz = vec_full_reg_size(s); | ||
179 | + TCGv_ptr status = fpstatus_ptr(FPST_FPCR); | ||
180 | + tcg_gen_gvec_4_ptr(vec_full_reg_offset(s, a->rd), | ||
181 | + vec_full_reg_offset(s, a->rn), | ||
182 | + vec_full_reg_offset(s, a->rm), | ||
183 | + vec_full_reg_offset(s, a->ra), | ||
184 | + status, vsz, vsz, 0, fn); | ||
185 | + tcg_temp_free_ptr(status); | ||
186 | + } | ||
187 | + return true; | ||
188 | +} | ||
189 | + | ||
190 | static bool do_sqdmlal_zzzw(DisasContext *s, arg_rrrr_esz *a, | ||
191 | bool sel1, bool sel2) | ||
192 | { | ||
193 | -- | ||
194 | 2.20.1 | ||
195 | |||
196 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-48-richard.henderson@linaro.org | ||
7 | Message-Id: <20200423180347.9403-1-steplong@quicinc.com> | ||
8 | [rth: Rename the trans_* functions to *_sve2.] | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/sve.decode | 11 +++++++++-- | ||
13 | target/arm/translate-sve.c | 35 ++++++++++++++++++++++++++++++----- | ||
14 | 2 files changed, 39 insertions(+), 7 deletions(-) | ||
15 | |||
16 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/sve.decode | ||
19 | +++ b/target/arm/sve.decode | ||
20 | @@ -XXX,XX +XXX,XX @@ CPY_z_i 00000101 .. 01 .... 00 . ........ ..... @rdn_pg4 imm=%sh8_i8s | ||
21 | |||
22 | ### SVE Permute - Extract Group | ||
23 | |||
24 | -# SVE extract vector (immediate offset) | ||
25 | +# SVE extract vector (destructive) | ||
26 | EXT 00000101 001 ..... 000 ... rm:5 rd:5 \ | ||
27 | &rrri rn=%reg_movprfx imm=%imm8_16_10 | ||
28 | |||
29 | +# SVE2 extract vector (constructive) | ||
30 | +EXT_sve2 00000101 011 ..... 000 ... rn:5 rd:5 \ | ||
31 | + &rri imm=%imm8_16_10 | ||
32 | + | ||
33 | ### SVE Permute - Unpredicated Group | ||
34 | |||
35 | # SVE broadcast general register | ||
36 | @@ -XXX,XX +XXX,XX @@ REVH 00000101 .. 1001 01 100 ... ..... ..... @rd_pg_rn | ||
37 | REVW 00000101 .. 1001 10 100 ... ..... ..... @rd_pg_rn | ||
38 | RBIT 00000101 .. 1001 11 100 ... ..... ..... @rd_pg_rn | ||
39 | |||
40 | -# SVE vector splice (predicated) | ||
41 | +# SVE vector splice (predicated, destructive) | ||
42 | SPLICE 00000101 .. 101 100 100 ... ..... ..... @rdn_pg_rm | ||
43 | |||
44 | +# SVE2 vector splice (predicated, constructive) | ||
45 | +SPLICE_sve2 00000101 .. 101 101 100 ... ..... ..... @rd_pg_rn | ||
46 | + | ||
47 | ### SVE Select Vectors Group | ||
48 | |||
49 | # SVE select vector elements (predicated) | ||
50 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/target/arm/translate-sve.c | ||
53 | +++ b/target/arm/translate-sve.c | ||
54 | @@ -XXX,XX +XXX,XX @@ static bool trans_CPY_z_i(DisasContext *s, arg_CPY_z_i *a) | ||
55 | *** SVE Permute Extract Group | ||
56 | */ | ||
57 | |||
58 | -static bool trans_EXT(DisasContext *s, arg_EXT *a) | ||
59 | +static bool do_EXT(DisasContext *s, int rd, int rn, int rm, int imm) | ||
60 | { | ||
61 | if (!sve_access_check(s)) { | ||
62 | return true; | ||
63 | } | ||
64 | |||
65 | unsigned vsz = vec_full_reg_size(s); | ||
66 | - unsigned n_ofs = a->imm >= vsz ? 0 : a->imm; | ||
67 | + unsigned n_ofs = imm >= vsz ? 0 : imm; | ||
68 | unsigned n_siz = vsz - n_ofs; | ||
69 | - unsigned d = vec_full_reg_offset(s, a->rd); | ||
70 | - unsigned n = vec_full_reg_offset(s, a->rn); | ||
71 | - unsigned m = vec_full_reg_offset(s, a->rm); | ||
72 | + unsigned d = vec_full_reg_offset(s, rd); | ||
73 | + unsigned n = vec_full_reg_offset(s, rn); | ||
74 | + unsigned m = vec_full_reg_offset(s, rm); | ||
75 | |||
76 | /* Use host vector move insns if we have appropriate sizes | ||
77 | * and no unfortunate overlap. | ||
78 | @@ -XXX,XX +XXX,XX @@ static bool trans_EXT(DisasContext *s, arg_EXT *a) | ||
79 | return true; | ||
80 | } | ||
81 | |||
82 | +static bool trans_EXT(DisasContext *s, arg_EXT *a) | ||
83 | +{ | ||
84 | + return do_EXT(s, a->rd, a->rn, a->rm, a->imm); | ||
85 | +} | ||
86 | + | ||
87 | +static bool trans_EXT_sve2(DisasContext *s, arg_rri *a) | ||
88 | +{ | ||
89 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
90 | + return false; | ||
91 | + } | ||
92 | + return do_EXT(s, a->rd, a->rn, (a->rn + 1) % 32, a->imm); | ||
93 | +} | ||
94 | + | ||
95 | /* | ||
96 | *** SVE Permute - Unpredicated Group | ||
97 | */ | ||
98 | @@ -XXX,XX +XXX,XX @@ static bool trans_SPLICE(DisasContext *s, arg_rprr_esz *a) | ||
99 | return true; | ||
100 | } | ||
101 | |||
102 | +static bool trans_SPLICE_sve2(DisasContext *s, arg_rpr_esz *a) | ||
103 | +{ | ||
104 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
105 | + return false; | ||
106 | + } | ||
107 | + if (sve_access_check(s)) { | ||
108 | + gen_gvec_ool_zzzp(s, gen_helper_sve_splice, | ||
109 | + a->rd, a->rn, (a->rn + 1) % 32, a->pg, a->esz); | ||
110 | + } | ||
111 | + return true; | ||
112 | +} | ||
113 | + | ||
114 | /* | ||
115 | *** SVE Integer Compare - Vectors Group | ||
116 | */ | ||
117 | -- | ||
118 | 2.20.1 | ||
119 | |||
120 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | The signed dot product routines produce a signed result. | ||
4 | Since we use -fwrapv, there is no functional change. | ||
5 | |||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-49-richard.henderson@linaro.org | ||
8 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | --- | ||
11 | target/arm/vec_helper.c | 8 ++++---- | ||
12 | 1 file changed, 4 insertions(+), 4 deletions(-) | ||
13 | |||
14 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/vec_helper.c | ||
17 | +++ b/target/arm/vec_helper.c | ||
18 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_d)(void *vd, void *vn, void *vm, | ||
19 | void HELPER(gvec_sdot_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
20 | { | ||
21 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
22 | - uint32_t *d = vd; | ||
23 | + int32_t *d = vd; | ||
24 | int8_t *n = vn, *m = vm; | ||
25 | |||
26 | for (i = 0; i < opr_sz / 4; ++i) { | ||
27 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_udot_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
28 | void HELPER(gvec_sdot_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
29 | { | ||
30 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
31 | - uint64_t *d = vd; | ||
32 | + int64_t *d = vd; | ||
33 | int16_t *n = vn, *m = vm; | ||
34 | |||
35 | for (i = 0; i < opr_sz / 8; ++i) { | ||
36 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
37 | { | ||
38 | intptr_t i, segend, opr_sz = simd_oprsz(desc), opr_sz_4 = opr_sz / 4; | ||
39 | intptr_t index = simd_data(desc); | ||
40 | - uint32_t *d = vd; | ||
41 | + int32_t *d = vd; | ||
42 | int8_t *n = vn; | ||
43 | int8_t *m_indexed = (int8_t *)vm + H4(index) * 4; | ||
44 | |||
45 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
46 | { | ||
47 | intptr_t i, opr_sz = simd_oprsz(desc), opr_sz_8 = opr_sz / 8; | ||
48 | intptr_t index = simd_data(desc); | ||
49 | - uint64_t *d = vd; | ||
50 | + int64_t *d = vd; | ||
51 | int16_t *n = vn; | ||
52 | int16_t *m_indexed = (int16_t *)vm + index * 4; | ||
53 | |||
54 | -- | ||
55 | 2.20.1 | ||
56 | |||
57 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | For SVE, we potentially have a 4th argument coming from the | ||
4 | movprfx instruction. Currently we do not optimize movprfx, | ||
5 | so the problem is not visible. | ||
6 | |||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-50-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper.h | 20 +++--- | ||
13 | target/arm/sve.decode | 7 ++- | ||
14 | target/arm/translate-a64.c | 15 ++++- | ||
15 | target/arm/translate-neon.c | 10 +-- | ||
16 | target/arm/translate-sve.c | 13 ++-- | ||
17 | target/arm/vec_helper.c | 120 ++++++++++++++++++++---------------- | ||
18 | 6 files changed, 109 insertions(+), 76 deletions(-) | ||
19 | |||
20 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
21 | index XXXXXXX..XXXXXXX 100644 | ||
22 | --- a/target/arm/helper.h | ||
23 | +++ b/target/arm/helper.h | ||
24 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqrdmlah_d, TCG_CALL_NO_RWG, | ||
25 | DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_d, TCG_CALL_NO_RWG, | ||
26 | void, ptr, ptr, ptr, ptr, i32) | ||
27 | |||
28 | -DEF_HELPER_FLAGS_4(gvec_sdot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | -DEF_HELPER_FLAGS_4(gvec_udot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | -DEF_HELPER_FLAGS_4(gvec_sdot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | -DEF_HELPER_FLAGS_4(gvec_udot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(gvec_sdot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(gvec_udot_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(gvec_sdot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(gvec_udot_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
36 | |||
37 | -DEF_HELPER_FLAGS_4(gvec_sdot_idx_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
38 | -DEF_HELPER_FLAGS_4(gvec_udot_idx_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
39 | -DEF_HELPER_FLAGS_4(gvec_sdot_idx_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
40 | -DEF_HELPER_FLAGS_4(gvec_udot_idx_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_5(gvec_sdot_idx_b, TCG_CALL_NO_RWG, | ||
42 | + void, ptr, ptr, ptr, ptr, i32) | ||
43 | +DEF_HELPER_FLAGS_5(gvec_udot_idx_b, TCG_CALL_NO_RWG, | ||
44 | + void, ptr, ptr, ptr, ptr, i32) | ||
45 | +DEF_HELPER_FLAGS_5(gvec_sdot_idx_h, TCG_CALL_NO_RWG, | ||
46 | + void, ptr, ptr, ptr, ptr, i32) | ||
47 | +DEF_HELPER_FLAGS_5(gvec_udot_idx_h, TCG_CALL_NO_RWG, | ||
48 | + void, ptr, ptr, ptr, ptr, i32) | ||
49 | |||
50 | DEF_HELPER_FLAGS_5(gvec_fcaddh, TCG_CALL_NO_RWG, | ||
51 | void, ptr, ptr, ptr, ptr, i32) | ||
52 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/target/arm/sve.decode | ||
55 | +++ b/target/arm/sve.decode | ||
56 | @@ -XXX,XX +XXX,XX @@ UMIN_zzi 00100101 .. 101 011 110 ........ ..... @rdn_i8u | ||
57 | MUL_zzi 00100101 .. 110 000 110 ........ ..... @rdn_i8s | ||
58 | |||
59 | # SVE integer dot product (unpredicated) | ||
60 | -DOT_zzz 01000100 1 sz:1 0 rm:5 00000 u:1 rn:5 rd:5 ra=%reg_movprfx | ||
61 | +DOT_zzzz 01000100 1 sz:1 0 rm:5 00000 u:1 rn:5 rd:5 \ | ||
62 | + ra=%reg_movprfx | ||
63 | |||
64 | # SVE integer dot product (indexed) | ||
65 | -DOT_zzx 01000100 101 index:2 rm:3 00000 u:1 rn:5 rd:5 \ | ||
66 | +DOT_zzxw 01000100 101 index:2 rm:3 00000 u:1 rn:5 rd:5 \ | ||
67 | sz=0 ra=%reg_movprfx | ||
68 | -DOT_zzx 01000100 111 index:1 rm:4 00000 u:1 rn:5 rd:5 \ | ||
69 | +DOT_zzxw 01000100 111 index:1 rm:4 00000 u:1 rn:5 rd:5 \ | ||
70 | sz=1 ra=%reg_movprfx | ||
71 | |||
72 | # SVE floating-point complex add (predicated) | ||
73 | diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c | ||
74 | index XXXXXXX..XXXXXXX 100644 | ||
75 | --- a/target/arm/translate-a64.c | ||
76 | +++ b/target/arm/translate-a64.c | ||
77 | @@ -XXX,XX +XXX,XX @@ static void gen_gvec_op3_qc(DisasContext *s, bool is_q, int rd, int rn, | ||
78 | tcg_temp_free_ptr(qc_ptr); | ||
79 | } | ||
80 | |||
81 | +/* Expand a 4-operand operation using an out-of-line helper. */ | ||
82 | +static void gen_gvec_op4_ool(DisasContext *s, bool is_q, int rd, int rn, | ||
83 | + int rm, int ra, int data, gen_helper_gvec_4 *fn) | ||
84 | +{ | ||
85 | + tcg_gen_gvec_4_ool(vec_full_reg_offset(s, rd), | ||
86 | + vec_full_reg_offset(s, rn), | ||
87 | + vec_full_reg_offset(s, rm), | ||
88 | + vec_full_reg_offset(s, ra), | ||
89 | + is_q ? 16 : 8, vec_full_reg_size(s), data, fn); | ||
90 | +} | ||
91 | + | ||
92 | /* Set ZF and NF based on a 64 bit result. This is alas fiddlier | ||
93 | * than the 32 bit equivalent. | ||
94 | */ | ||
95 | @@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn) | ||
96 | return; | ||
97 | |||
98 | case 0x2: /* SDOT / UDOT */ | ||
99 | - gen_gvec_op3_ool(s, is_q, rd, rn, rm, 0, | ||
100 | + gen_gvec_op4_ool(s, is_q, rd, rn, rm, rd, 0, | ||
101 | u ? gen_helper_gvec_udot_b : gen_helper_gvec_sdot_b); | ||
102 | return; | ||
103 | |||
104 | @@ -XXX,XX +XXX,XX @@ static void disas_simd_indexed(DisasContext *s, uint32_t insn) | ||
105 | switch (16 * u + opcode) { | ||
106 | case 0x0e: /* SDOT */ | ||
107 | case 0x1e: /* UDOT */ | ||
108 | - gen_gvec_op3_ool(s, is_q, rd, rn, rm, index, | ||
109 | + gen_gvec_op4_ool(s, is_q, rd, rn, rm, rd, index, | ||
110 | u ? gen_helper_gvec_udot_idx_b | ||
111 | : gen_helper_gvec_sdot_idx_b); | ||
112 | return; | ||
113 | diff --git a/target/arm/translate-neon.c b/target/arm/translate-neon.c | ||
114 | index XXXXXXX..XXXXXXX 100644 | ||
115 | --- a/target/arm/translate-neon.c | ||
116 | +++ b/target/arm/translate-neon.c | ||
117 | @@ -XXX,XX +XXX,XX @@ static bool trans_VCADD(DisasContext *s, arg_VCADD *a) | ||
118 | static bool trans_VDOT(DisasContext *s, arg_VDOT *a) | ||
119 | { | ||
120 | int opr_sz; | ||
121 | - gen_helper_gvec_3 *fn_gvec; | ||
122 | + gen_helper_gvec_4 *fn_gvec; | ||
123 | |||
124 | if (!dc_isar_feature(aa32_dp, s)) { | ||
125 | return false; | ||
126 | @@ -XXX,XX +XXX,XX @@ static bool trans_VDOT(DisasContext *s, arg_VDOT *a) | ||
127 | |||
128 | opr_sz = (1 + a->q) * 8; | ||
129 | fn_gvec = a->u ? gen_helper_gvec_udot_b : gen_helper_gvec_sdot_b; | ||
130 | - tcg_gen_gvec_3_ool(vfp_reg_offset(1, a->vd), | ||
131 | + tcg_gen_gvec_4_ool(vfp_reg_offset(1, a->vd), | ||
132 | vfp_reg_offset(1, a->vn), | ||
133 | vfp_reg_offset(1, a->vm), | ||
134 | + vfp_reg_offset(1, a->vd), | ||
135 | opr_sz, opr_sz, 0, fn_gvec); | ||
136 | return true; | ||
137 | } | ||
138 | @@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA_scalar(DisasContext *s, arg_VCMLA_scalar *a) | ||
139 | |||
140 | static bool trans_VDOT_scalar(DisasContext *s, arg_VDOT_scalar *a) | ||
141 | { | ||
142 | - gen_helper_gvec_3 *fn_gvec; | ||
143 | + gen_helper_gvec_4 *fn_gvec; | ||
144 | int opr_sz; | ||
145 | TCGv_ptr fpst; | ||
146 | |||
147 | @@ -XXX,XX +XXX,XX @@ static bool trans_VDOT_scalar(DisasContext *s, arg_VDOT_scalar *a) | ||
148 | fn_gvec = a->u ? gen_helper_gvec_udot_idx_b : gen_helper_gvec_sdot_idx_b; | ||
149 | opr_sz = (1 + a->q) * 8; | ||
150 | fpst = fpstatus_ptr(FPST_STD); | ||
151 | - tcg_gen_gvec_3_ool(vfp_reg_offset(1, a->vd), | ||
152 | + tcg_gen_gvec_4_ool(vfp_reg_offset(1, a->vd), | ||
153 | vfp_reg_offset(1, a->vn), | ||
154 | vfp_reg_offset(1, a->rm), | ||
155 | + vfp_reg_offset(1, a->vd), | ||
156 | opr_sz, opr_sz, a->index, fn_gvec); | ||
157 | tcg_temp_free_ptr(fpst); | ||
158 | return true; | ||
159 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
160 | index XXXXXXX..XXXXXXX 100644 | ||
161 | --- a/target/arm/translate-sve.c | ||
162 | +++ b/target/arm/translate-sve.c | ||
163 | @@ -XXX,XX +XXX,XX @@ DO_ZZI(UMIN, umin) | ||
164 | |||
165 | #undef DO_ZZI | ||
166 | |||
167 | -static bool trans_DOT_zzz(DisasContext *s, arg_DOT_zzz *a) | ||
168 | +static bool trans_DOT_zzzz(DisasContext *s, arg_DOT_zzzz *a) | ||
169 | { | ||
170 | - static gen_helper_gvec_3 * const fns[2][2] = { | ||
171 | + static gen_helper_gvec_4 * const fns[2][2] = { | ||
172 | { gen_helper_gvec_sdot_b, gen_helper_gvec_sdot_h }, | ||
173 | { gen_helper_gvec_udot_b, gen_helper_gvec_udot_h } | ||
174 | }; | ||
175 | |||
176 | if (sve_access_check(s)) { | ||
177 | - gen_gvec_ool_zzz(s, fns[a->u][a->sz], a->rd, a->rn, a->rm, 0); | ||
178 | + gen_gvec_ool_zzzz(s, fns[a->u][a->sz], a->rd, a->rn, a->rm, a->ra, 0); | ||
179 | } | ||
180 | return true; | ||
181 | } | ||
182 | |||
183 | -static bool trans_DOT_zzx(DisasContext *s, arg_DOT_zzx *a) | ||
184 | +static bool trans_DOT_zzxw(DisasContext *s, arg_DOT_zzxw *a) | ||
185 | { | ||
186 | - static gen_helper_gvec_3 * const fns[2][2] = { | ||
187 | + static gen_helper_gvec_4 * const fns[2][2] = { | ||
188 | { gen_helper_gvec_sdot_idx_b, gen_helper_gvec_sdot_idx_h }, | ||
189 | { gen_helper_gvec_udot_idx_b, gen_helper_gvec_udot_idx_h } | ||
190 | }; | ||
191 | |||
192 | if (sve_access_check(s)) { | ||
193 | - gen_gvec_ool_zzz(s, fns[a->u][a->sz], a->rd, a->rn, a->rm, a->index); | ||
194 | + gen_gvec_ool_zzzz(s, fns[a->u][a->sz], a->rd, a->rn, a->rm, | ||
195 | + a->ra, a->index); | ||
196 | } | ||
197 | return true; | ||
198 | } | ||
199 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
200 | index XXXXXXX..XXXXXXX 100644 | ||
201 | --- a/target/arm/vec_helper.c | ||
202 | +++ b/target/arm/vec_helper.c | ||
203 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_d)(void *vd, void *vn, void *vm, | ||
204 | * All elements are treated equally, no matter where they are. | ||
205 | */ | ||
206 | |||
207 | -void HELPER(gvec_sdot_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
208 | +void HELPER(gvec_sdot_b)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
209 | { | ||
210 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
211 | - int32_t *d = vd; | ||
212 | + int32_t *d = vd, *a = va; | ||
213 | int8_t *n = vn, *m = vm; | ||
214 | |||
215 | for (i = 0; i < opr_sz / 4; ++i) { | ||
216 | - d[i] += n[i * 4 + 0] * m[i * 4 + 0] | ||
217 | - + n[i * 4 + 1] * m[i * 4 + 1] | ||
218 | - + n[i * 4 + 2] * m[i * 4 + 2] | ||
219 | - + n[i * 4 + 3] * m[i * 4 + 3]; | ||
220 | + d[i] = (a[i] + | ||
221 | + n[i * 4 + 0] * m[i * 4 + 0] + | ||
222 | + n[i * 4 + 1] * m[i * 4 + 1] + | ||
223 | + n[i * 4 + 2] * m[i * 4 + 2] + | ||
224 | + n[i * 4 + 3] * m[i * 4 + 3]); | ||
225 | } | ||
226 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
227 | } | ||
228 | |||
229 | -void HELPER(gvec_udot_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
230 | +void HELPER(gvec_udot_b)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
231 | { | ||
232 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
233 | - uint32_t *d = vd; | ||
234 | + uint32_t *d = vd, *a = va; | ||
235 | uint8_t *n = vn, *m = vm; | ||
236 | |||
237 | for (i = 0; i < opr_sz / 4; ++i) { | ||
238 | - d[i] += n[i * 4 + 0] * m[i * 4 + 0] | ||
239 | - + n[i * 4 + 1] * m[i * 4 + 1] | ||
240 | - + n[i * 4 + 2] * m[i * 4 + 2] | ||
241 | - + n[i * 4 + 3] * m[i * 4 + 3]; | ||
242 | + d[i] = (a[i] + | ||
243 | + n[i * 4 + 0] * m[i * 4 + 0] + | ||
244 | + n[i * 4 + 1] * m[i * 4 + 1] + | ||
245 | + n[i * 4 + 2] * m[i * 4 + 2] + | ||
246 | + n[i * 4 + 3] * m[i * 4 + 3]); | ||
247 | } | ||
248 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
249 | } | ||
250 | |||
251 | -void HELPER(gvec_sdot_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
252 | +void HELPER(gvec_sdot_h)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
253 | { | ||
254 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
255 | - int64_t *d = vd; | ||
256 | + int64_t *d = vd, *a = va; | ||
257 | int16_t *n = vn, *m = vm; | ||
258 | |||
259 | for (i = 0; i < opr_sz / 8; ++i) { | ||
260 | - d[i] += (int64_t)n[i * 4 + 0] * m[i * 4 + 0] | ||
261 | - + (int64_t)n[i * 4 + 1] * m[i * 4 + 1] | ||
262 | - + (int64_t)n[i * 4 + 2] * m[i * 4 + 2] | ||
263 | - + (int64_t)n[i * 4 + 3] * m[i * 4 + 3]; | ||
264 | + d[i] = (a[i] + | ||
265 | + (int64_t)n[i * 4 + 0] * m[i * 4 + 0] + | ||
266 | + (int64_t)n[i * 4 + 1] * m[i * 4 + 1] + | ||
267 | + (int64_t)n[i * 4 + 2] * m[i * 4 + 2] + | ||
268 | + (int64_t)n[i * 4 + 3] * m[i * 4 + 3]); | ||
269 | } | ||
270 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
271 | } | ||
272 | |||
273 | -void HELPER(gvec_udot_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
274 | +void HELPER(gvec_udot_h)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
275 | { | ||
276 | intptr_t i, opr_sz = simd_oprsz(desc); | ||
277 | - uint64_t *d = vd; | ||
278 | + uint64_t *d = vd, *a = va; | ||
279 | uint16_t *n = vn, *m = vm; | ||
280 | |||
281 | for (i = 0; i < opr_sz / 8; ++i) { | ||
282 | - d[i] += (uint64_t)n[i * 4 + 0] * m[i * 4 + 0] | ||
283 | - + (uint64_t)n[i * 4 + 1] * m[i * 4 + 1] | ||
284 | - + (uint64_t)n[i * 4 + 2] * m[i * 4 + 2] | ||
285 | - + (uint64_t)n[i * 4 + 3] * m[i * 4 + 3]; | ||
286 | + d[i] = (a[i] + | ||
287 | + (uint64_t)n[i * 4 + 0] * m[i * 4 + 0] + | ||
288 | + (uint64_t)n[i * 4 + 1] * m[i * 4 + 1] + | ||
289 | + (uint64_t)n[i * 4 + 2] * m[i * 4 + 2] + | ||
290 | + (uint64_t)n[i * 4 + 3] * m[i * 4 + 3]); | ||
291 | } | ||
292 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
293 | } | ||
294 | |||
295 | -void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
296 | +void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, | ||
297 | + void *va, uint32_t desc) | ||
298 | { | ||
299 | intptr_t i, segend, opr_sz = simd_oprsz(desc), opr_sz_4 = opr_sz / 4; | ||
300 | intptr_t index = simd_data(desc); | ||
301 | - int32_t *d = vd; | ||
302 | + int32_t *d = vd, *a = va; | ||
303 | int8_t *n = vn; | ||
304 | int8_t *m_indexed = (int8_t *)vm + H4(index) * 4; | ||
305 | |||
306 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
307 | int8_t m3 = m_indexed[i * 4 + 3]; | ||
308 | |||
309 | do { | ||
310 | - d[i] += n[i * 4 + 0] * m0 | ||
311 | - + n[i * 4 + 1] * m1 | ||
312 | - + n[i * 4 + 2] * m2 | ||
313 | - + n[i * 4 + 3] * m3; | ||
314 | + d[i] = (a[i] + | ||
315 | + n[i * 4 + 0] * m0 + | ||
316 | + n[i * 4 + 1] * m1 + | ||
317 | + n[i * 4 + 2] * m2 + | ||
318 | + n[i * 4 + 3] * m3); | ||
319 | } while (++i < segend); | ||
320 | segend = i + 4; | ||
321 | } while (i < opr_sz_4); | ||
322 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
323 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
324 | } | ||
325 | |||
326 | -void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
327 | +void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, | ||
328 | + void *va, uint32_t desc) | ||
329 | { | ||
330 | intptr_t i, segend, opr_sz = simd_oprsz(desc), opr_sz_4 = opr_sz / 4; | ||
331 | intptr_t index = simd_data(desc); | ||
332 | - uint32_t *d = vd; | ||
333 | + uint32_t *d = vd, *a = va; | ||
334 | uint8_t *n = vn; | ||
335 | uint8_t *m_indexed = (uint8_t *)vm + H4(index) * 4; | ||
336 | |||
337 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
338 | uint8_t m3 = m_indexed[i * 4 + 3]; | ||
339 | |||
340 | do { | ||
341 | - d[i] += n[i * 4 + 0] * m0 | ||
342 | - + n[i * 4 + 1] * m1 | ||
343 | - + n[i * 4 + 2] * m2 | ||
344 | - + n[i * 4 + 3] * m3; | ||
345 | + d[i] = (a[i] + | ||
346 | + n[i * 4 + 0] * m0 + | ||
347 | + n[i * 4 + 1] * m1 + | ||
348 | + n[i * 4 + 2] * m2 + | ||
349 | + n[i * 4 + 3] * m3); | ||
350 | } while (++i < segend); | ||
351 | segend = i + 4; | ||
352 | } while (i < opr_sz_4); | ||
353 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
354 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
355 | } | ||
356 | |||
357 | -void HELPER(gvec_sdot_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
358 | +void HELPER(gvec_sdot_idx_h)(void *vd, void *vn, void *vm, | ||
359 | + void *va, uint32_t desc) | ||
360 | { | ||
361 | intptr_t i, opr_sz = simd_oprsz(desc), opr_sz_8 = opr_sz / 8; | ||
362 | intptr_t index = simd_data(desc); | ||
363 | - int64_t *d = vd; | ||
364 | + int64_t *d = vd, *a = va; | ||
365 | int16_t *n = vn; | ||
366 | int16_t *m_indexed = (int16_t *)vm + index * 4; | ||
367 | |||
368 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
369 | * Process the entire segment all at once, writing back the results | ||
370 | * only after we've consumed all of the inputs. | ||
371 | */ | ||
372 | - for (i = 0; i < opr_sz_8 ; i += 2) { | ||
373 | - uint64_t d0, d1; | ||
374 | + for (i = 0; i < opr_sz_8; i += 2) { | ||
375 | + int64_t d0, d1; | ||
376 | |||
377 | - d0 = n[i * 4 + 0] * (int64_t)m_indexed[i * 4 + 0]; | ||
378 | + d0 = a[i + 0]; | ||
379 | + d0 += n[i * 4 + 0] * (int64_t)m_indexed[i * 4 + 0]; | ||
380 | d0 += n[i * 4 + 1] * (int64_t)m_indexed[i * 4 + 1]; | ||
381 | d0 += n[i * 4 + 2] * (int64_t)m_indexed[i * 4 + 2]; | ||
382 | d0 += n[i * 4 + 3] * (int64_t)m_indexed[i * 4 + 3]; | ||
383 | - d1 = n[i * 4 + 4] * (int64_t)m_indexed[i * 4 + 0]; | ||
384 | + | ||
385 | + d1 = a[i + 1]; | ||
386 | + d1 += n[i * 4 + 4] * (int64_t)m_indexed[i * 4 + 0]; | ||
387 | d1 += n[i * 4 + 5] * (int64_t)m_indexed[i * 4 + 1]; | ||
388 | d1 += n[i * 4 + 6] * (int64_t)m_indexed[i * 4 + 2]; | ||
389 | d1 += n[i * 4 + 7] * (int64_t)m_indexed[i * 4 + 3]; | ||
390 | |||
391 | - d[i + 0] += d0; | ||
392 | - d[i + 1] += d1; | ||
393 | + d[i + 0] = d0; | ||
394 | + d[i + 1] = d1; | ||
395 | } | ||
396 | - | ||
397 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
398 | } | ||
399 | |||
400 | -void HELPER(gvec_udot_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
401 | +void HELPER(gvec_udot_idx_h)(void *vd, void *vn, void *vm, | ||
402 | + void *va, uint32_t desc) | ||
403 | { | ||
404 | intptr_t i, opr_sz = simd_oprsz(desc), opr_sz_8 = opr_sz / 8; | ||
405 | intptr_t index = simd_data(desc); | ||
406 | - uint64_t *d = vd; | ||
407 | + uint64_t *d = vd, *a = va; | ||
408 | uint16_t *n = vn; | ||
409 | uint16_t *m_indexed = (uint16_t *)vm + index * 4; | ||
410 | |||
411 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_udot_idx_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
412 | * Process the entire segment all at once, writing back the results | ||
413 | * only after we've consumed all of the inputs. | ||
414 | */ | ||
415 | - for (i = 0; i < opr_sz_8 ; i += 2) { | ||
416 | + for (i = 0; i < opr_sz_8; i += 2) { | ||
417 | uint64_t d0, d1; | ||
418 | |||
419 | - d0 = n[i * 4 + 0] * (uint64_t)m_indexed[i * 4 + 0]; | ||
420 | + d0 = a[i + 0]; | ||
421 | + d0 += n[i * 4 + 0] * (uint64_t)m_indexed[i * 4 + 0]; | ||
422 | d0 += n[i * 4 + 1] * (uint64_t)m_indexed[i * 4 + 1]; | ||
423 | d0 += n[i * 4 + 2] * (uint64_t)m_indexed[i * 4 + 2]; | ||
424 | d0 += n[i * 4 + 3] * (uint64_t)m_indexed[i * 4 + 3]; | ||
425 | - d1 = n[i * 4 + 4] * (uint64_t)m_indexed[i * 4 + 0]; | ||
426 | + | ||
427 | + d1 = a[i + 1]; | ||
428 | + d1 += n[i * 4 + 4] * (uint64_t)m_indexed[i * 4 + 0]; | ||
429 | d1 += n[i * 4 + 5] * (uint64_t)m_indexed[i * 4 + 1]; | ||
430 | d1 += n[i * 4 + 6] * (uint64_t)m_indexed[i * 4 + 2]; | ||
431 | d1 += n[i * 4 + 7] * (uint64_t)m_indexed[i * 4 + 3]; | ||
432 | |||
433 | - d[i + 0] += d0; | ||
434 | - d[i + 1] += d1; | ||
435 | + d[i + 0] = d0; | ||
436 | + d[i + 1] = d1; | ||
437 | } | ||
438 | - | ||
439 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
440 | } | ||
441 | |||
442 | -- | ||
443 | 2.20.1 | ||
444 | |||
445 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | For SVE, we potentially have a 4th argument coming from the | ||
4 | movprfx instruction. Currently we do not optimize movprfx, | ||
5 | so the problem is not visible. | ||
6 | |||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-51-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper.h | 20 +++++++-------- | ||
13 | target/arm/translate-a64.c | 28 +++++++++++++++++---- | ||
14 | target/arm/translate-neon.c | 10 +++++--- | ||
15 | target/arm/translate-sve.c | 5 ++-- | ||
16 | target/arm/vec_helper.c | 50 +++++++++++++++---------------------- | ||
17 | 5 files changed, 62 insertions(+), 51 deletions(-) | ||
18 | |||
19 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/target/arm/helper.h | ||
22 | +++ b/target/arm/helper.h | ||
23 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(gvec_fcadds, TCG_CALL_NO_RWG, | ||
24 | DEF_HELPER_FLAGS_5(gvec_fcaddd, TCG_CALL_NO_RWG, | ||
25 | void, ptr, ptr, ptr, ptr, i32) | ||
26 | |||
27 | -DEF_HELPER_FLAGS_5(gvec_fcmlah, TCG_CALL_NO_RWG, | ||
28 | - void, ptr, ptr, ptr, ptr, i32) | ||
29 | -DEF_HELPER_FLAGS_5(gvec_fcmlah_idx, TCG_CALL_NO_RWG, | ||
30 | - void, ptr, ptr, ptr, ptr, i32) | ||
31 | -DEF_HELPER_FLAGS_5(gvec_fcmlas, TCG_CALL_NO_RWG, | ||
32 | - void, ptr, ptr, ptr, ptr, i32) | ||
33 | -DEF_HELPER_FLAGS_5(gvec_fcmlas_idx, TCG_CALL_NO_RWG, | ||
34 | - void, ptr, ptr, ptr, ptr, i32) | ||
35 | -DEF_HELPER_FLAGS_5(gvec_fcmlad, TCG_CALL_NO_RWG, | ||
36 | - void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_6(gvec_fcmlah, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
39 | +DEF_HELPER_FLAGS_6(gvec_fcmlah_idx, TCG_CALL_NO_RWG, | ||
40 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_6(gvec_fcmlas, TCG_CALL_NO_RWG, | ||
42 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
43 | +DEF_HELPER_FLAGS_6(gvec_fcmlas_idx, TCG_CALL_NO_RWG, | ||
44 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
45 | +DEF_HELPER_FLAGS_6(gvec_fcmlad, TCG_CALL_NO_RWG, | ||
46 | + void, ptr, ptr, ptr, ptr, ptr, i32) | ||
47 | |||
48 | DEF_HELPER_FLAGS_5(neon_paddh, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
49 | DEF_HELPER_FLAGS_5(neon_pmaxh, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
50 | diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/target/arm/translate-a64.c | ||
53 | +++ b/target/arm/translate-a64.c | ||
54 | @@ -XXX,XX +XXX,XX @@ static void gen_gvec_op4_ool(DisasContext *s, bool is_q, int rd, int rn, | ||
55 | is_q ? 16 : 8, vec_full_reg_size(s), data, fn); | ||
56 | } | ||
57 | |||
58 | +/* | ||
59 | + * Expand a 4-operand + fpstatus pointer + simd data value operation using | ||
60 | + * an out-of-line helper. | ||
61 | + */ | ||
62 | +static void gen_gvec_op4_fpst(DisasContext *s, bool is_q, int rd, int rn, | ||
63 | + int rm, int ra, bool is_fp16, int data, | ||
64 | + gen_helper_gvec_4_ptr *fn) | ||
65 | +{ | ||
66 | + TCGv_ptr fpst = fpstatus_ptr(is_fp16 ? FPST_FPCR_F16 : FPST_FPCR); | ||
67 | + tcg_gen_gvec_4_ptr(vec_full_reg_offset(s, rd), | ||
68 | + vec_full_reg_offset(s, rn), | ||
69 | + vec_full_reg_offset(s, rm), | ||
70 | + vec_full_reg_offset(s, ra), fpst, | ||
71 | + is_q ? 16 : 8, vec_full_reg_size(s), data, fn); | ||
72 | + tcg_temp_free_ptr(fpst); | ||
73 | +} | ||
74 | + | ||
75 | /* Set ZF and NF based on a 64 bit result. This is alas fiddlier | ||
76 | * than the 32 bit equivalent. | ||
77 | */ | ||
78 | @@ -XXX,XX +XXX,XX @@ static void disas_simd_three_reg_same_extra(DisasContext *s, uint32_t insn) | ||
79 | rot = extract32(opcode, 0, 2); | ||
80 | switch (size) { | ||
81 | case 1: | ||
82 | - gen_gvec_op3_fpst(s, is_q, rd, rn, rm, true, rot, | ||
83 | + gen_gvec_op4_fpst(s, is_q, rd, rn, rm, rd, true, rot, | ||
84 | gen_helper_gvec_fcmlah); | ||
85 | break; | ||
86 | case 2: | ||
87 | - gen_gvec_op3_fpst(s, is_q, rd, rn, rm, false, rot, | ||
88 | + gen_gvec_op4_fpst(s, is_q, rd, rn, rm, rd, false, rot, | ||
89 | gen_helper_gvec_fcmlas); | ||
90 | break; | ||
91 | case 3: | ||
92 | - gen_gvec_op3_fpst(s, is_q, rd, rn, rm, false, rot, | ||
93 | + gen_gvec_op4_fpst(s, is_q, rd, rn, rm, rd, false, rot, | ||
94 | gen_helper_gvec_fcmlad); | ||
95 | break; | ||
96 | default: | ||
97 | @@ -XXX,XX +XXX,XX @@ static void disas_simd_indexed(DisasContext *s, uint32_t insn) | ||
98 | { | ||
99 | int rot = extract32(insn, 13, 2); | ||
100 | int data = (index << 2) | rot; | ||
101 | - tcg_gen_gvec_3_ptr(vec_full_reg_offset(s, rd), | ||
102 | + tcg_gen_gvec_4_ptr(vec_full_reg_offset(s, rd), | ||
103 | vec_full_reg_offset(s, rn), | ||
104 | - vec_full_reg_offset(s, rm), fpst, | ||
105 | + vec_full_reg_offset(s, rm), | ||
106 | + vec_full_reg_offset(s, rd), fpst, | ||
107 | is_q ? 16 : 8, vec_full_reg_size(s), data, | ||
108 | size == MO_64 | ||
109 | ? gen_helper_gvec_fcmlas_idx | ||
110 | diff --git a/target/arm/translate-neon.c b/target/arm/translate-neon.c | ||
111 | index XXXXXXX..XXXXXXX 100644 | ||
112 | --- a/target/arm/translate-neon.c | ||
113 | +++ b/target/arm/translate-neon.c | ||
114 | @@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA(DisasContext *s, arg_VCMLA *a) | ||
115 | { | ||
116 | int opr_sz; | ||
117 | TCGv_ptr fpst; | ||
118 | - gen_helper_gvec_3_ptr *fn_gvec_ptr; | ||
119 | + gen_helper_gvec_4_ptr *fn_gvec_ptr; | ||
120 | |||
121 | if (!dc_isar_feature(aa32_vcma, s) | ||
122 | || (a->size == MO_16 && !dc_isar_feature(aa32_fp16_arith, s))) { | ||
123 | @@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA(DisasContext *s, arg_VCMLA *a) | ||
124 | fpst = fpstatus_ptr(a->size == MO_16 ? FPST_STD_F16 : FPST_STD); | ||
125 | fn_gvec_ptr = (a->size == MO_16) ? | ||
126 | gen_helper_gvec_fcmlah : gen_helper_gvec_fcmlas; | ||
127 | - tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd), | ||
128 | + tcg_gen_gvec_4_ptr(vfp_reg_offset(1, a->vd), | ||
129 | vfp_reg_offset(1, a->vn), | ||
130 | vfp_reg_offset(1, a->vm), | ||
131 | + vfp_reg_offset(1, a->vd), | ||
132 | fpst, opr_sz, opr_sz, a->rot, | ||
133 | fn_gvec_ptr); | ||
134 | tcg_temp_free_ptr(fpst); | ||
135 | @@ -XXX,XX +XXX,XX @@ static bool trans_VFML(DisasContext *s, arg_VFML *a) | ||
136 | |||
137 | static bool trans_VCMLA_scalar(DisasContext *s, arg_VCMLA_scalar *a) | ||
138 | { | ||
139 | - gen_helper_gvec_3_ptr *fn_gvec_ptr; | ||
140 | + gen_helper_gvec_4_ptr *fn_gvec_ptr; | ||
141 | int opr_sz; | ||
142 | TCGv_ptr fpst; | ||
143 | |||
144 | @@ -XXX,XX +XXX,XX @@ static bool trans_VCMLA_scalar(DisasContext *s, arg_VCMLA_scalar *a) | ||
145 | gen_helper_gvec_fcmlah_idx : gen_helper_gvec_fcmlas_idx; | ||
146 | opr_sz = (1 + a->q) * 8; | ||
147 | fpst = fpstatus_ptr(a->size == MO_16 ? FPST_STD_F16 : FPST_STD); | ||
148 | - tcg_gen_gvec_3_ptr(vfp_reg_offset(1, a->vd), | ||
149 | + tcg_gen_gvec_4_ptr(vfp_reg_offset(1, a->vd), | ||
150 | vfp_reg_offset(1, a->vn), | ||
151 | vfp_reg_offset(1, a->vm), | ||
152 | + vfp_reg_offset(1, a->vd), | ||
153 | fpst, opr_sz, opr_sz, | ||
154 | (a->index << 2) | a->rot, fn_gvec_ptr); | ||
155 | tcg_temp_free_ptr(fpst); | ||
156 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
157 | index XXXXXXX..XXXXXXX 100644 | ||
158 | --- a/target/arm/translate-sve.c | ||
159 | +++ b/target/arm/translate-sve.c | ||
160 | @@ -XXX,XX +XXX,XX @@ static bool trans_FCMLA_zpzzz(DisasContext *s, arg_FCMLA_zpzzz *a) | ||
161 | |||
162 | static bool trans_FCMLA_zzxz(DisasContext *s, arg_FCMLA_zzxz *a) | ||
163 | { | ||
164 | - static gen_helper_gvec_3_ptr * const fns[2] = { | ||
165 | + static gen_helper_gvec_4_ptr * const fns[2] = { | ||
166 | gen_helper_gvec_fcmlah_idx, | ||
167 | gen_helper_gvec_fcmlas_idx, | ||
168 | }; | ||
169 | @@ -XXX,XX +XXX,XX @@ static bool trans_FCMLA_zzxz(DisasContext *s, arg_FCMLA_zzxz *a) | ||
170 | if (sve_access_check(s)) { | ||
171 | unsigned vsz = vec_full_reg_size(s); | ||
172 | TCGv_ptr status = fpstatus_ptr(a->esz == MO_16 ? FPST_FPCR_F16 : FPST_FPCR); | ||
173 | - tcg_gen_gvec_3_ptr(vec_full_reg_offset(s, a->rd), | ||
174 | + tcg_gen_gvec_4_ptr(vec_full_reg_offset(s, a->rd), | ||
175 | vec_full_reg_offset(s, a->rn), | ||
176 | vec_full_reg_offset(s, a->rm), | ||
177 | + vec_full_reg_offset(s, a->ra), | ||
178 | status, vsz, vsz, | ||
179 | a->index * 4 + a->rot, | ||
180 | fns[a->esz - 1]); | ||
181 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
182 | index XXXXXXX..XXXXXXX 100644 | ||
183 | --- a/target/arm/vec_helper.c | ||
184 | +++ b/target/arm/vec_helper.c | ||
185 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcaddd)(void *vd, void *vn, void *vm, | ||
186 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
187 | } | ||
188 | |||
189 | -void HELPER(gvec_fcmlah)(void *vd, void *vn, void *vm, | ||
190 | +void HELPER(gvec_fcmlah)(void *vd, void *vn, void *vm, void *va, | ||
191 | void *vfpst, uint32_t desc) | ||
192 | { | ||
193 | uintptr_t opr_sz = simd_oprsz(desc); | ||
194 | - float16 *d = vd; | ||
195 | - float16 *n = vn; | ||
196 | - float16 *m = vm; | ||
197 | + float16 *d = vd, *n = vn, *m = vm, *a = va; | ||
198 | float_status *fpst = vfpst; | ||
199 | intptr_t flip = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
200 | uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
201 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcmlah)(void *vd, void *vn, void *vm, | ||
202 | float16 e4 = e2; | ||
203 | float16 e3 = m[H2(i + 1 - flip)] ^ neg_imag; | ||
204 | |||
205 | - d[H2(i)] = float16_muladd(e2, e1, d[H2(i)], 0, fpst); | ||
206 | - d[H2(i + 1)] = float16_muladd(e4, e3, d[H2(i + 1)], 0, fpst); | ||
207 | + d[H2(i)] = float16_muladd(e2, e1, a[H2(i)], 0, fpst); | ||
208 | + d[H2(i + 1)] = float16_muladd(e4, e3, a[H2(i + 1)], 0, fpst); | ||
209 | } | ||
210 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
211 | } | ||
212 | |||
213 | -void HELPER(gvec_fcmlah_idx)(void *vd, void *vn, void *vm, | ||
214 | +void HELPER(gvec_fcmlah_idx)(void *vd, void *vn, void *vm, void *va, | ||
215 | void *vfpst, uint32_t desc) | ||
216 | { | ||
217 | uintptr_t opr_sz = simd_oprsz(desc); | ||
218 | - float16 *d = vd; | ||
219 | - float16 *n = vn; | ||
220 | - float16 *m = vm; | ||
221 | + float16 *d = vd, *n = vn, *m = vm, *a = va; | ||
222 | float_status *fpst = vfpst; | ||
223 | intptr_t flip = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
224 | uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
225 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcmlah_idx)(void *vd, void *vn, void *vm, | ||
226 | float16 e2 = n[H2(j + flip)]; | ||
227 | float16 e4 = e2; | ||
228 | |||
229 | - d[H2(j)] = float16_muladd(e2, e1, d[H2(j)], 0, fpst); | ||
230 | - d[H2(j + 1)] = float16_muladd(e4, e3, d[H2(j + 1)], 0, fpst); | ||
231 | + d[H2(j)] = float16_muladd(e2, e1, a[H2(j)], 0, fpst); | ||
232 | + d[H2(j + 1)] = float16_muladd(e4, e3, a[H2(j + 1)], 0, fpst); | ||
233 | } | ||
234 | } | ||
235 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
236 | } | ||
237 | |||
238 | -void HELPER(gvec_fcmlas)(void *vd, void *vn, void *vm, | ||
239 | +void HELPER(gvec_fcmlas)(void *vd, void *vn, void *vm, void *va, | ||
240 | void *vfpst, uint32_t desc) | ||
241 | { | ||
242 | uintptr_t opr_sz = simd_oprsz(desc); | ||
243 | - float32 *d = vd; | ||
244 | - float32 *n = vn; | ||
245 | - float32 *m = vm; | ||
246 | + float32 *d = vd, *n = vn, *m = vm, *a = va; | ||
247 | float_status *fpst = vfpst; | ||
248 | intptr_t flip = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
249 | uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
250 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcmlas)(void *vd, void *vn, void *vm, | ||
251 | float32 e4 = e2; | ||
252 | float32 e3 = m[H4(i + 1 - flip)] ^ neg_imag; | ||
253 | |||
254 | - d[H4(i)] = float32_muladd(e2, e1, d[H4(i)], 0, fpst); | ||
255 | - d[H4(i + 1)] = float32_muladd(e4, e3, d[H4(i + 1)], 0, fpst); | ||
256 | + d[H4(i)] = float32_muladd(e2, e1, a[H4(i)], 0, fpst); | ||
257 | + d[H4(i + 1)] = float32_muladd(e4, e3, a[H4(i + 1)], 0, fpst); | ||
258 | } | ||
259 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
260 | } | ||
261 | |||
262 | -void HELPER(gvec_fcmlas_idx)(void *vd, void *vn, void *vm, | ||
263 | +void HELPER(gvec_fcmlas_idx)(void *vd, void *vn, void *vm, void *va, | ||
264 | void *vfpst, uint32_t desc) | ||
265 | { | ||
266 | uintptr_t opr_sz = simd_oprsz(desc); | ||
267 | - float32 *d = vd; | ||
268 | - float32 *n = vn; | ||
269 | - float32 *m = vm; | ||
270 | + float32 *d = vd, *n = vn, *m = vm, *a = va; | ||
271 | float_status *fpst = vfpst; | ||
272 | intptr_t flip = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
273 | uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
274 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcmlas_idx)(void *vd, void *vn, void *vm, | ||
275 | float32 e2 = n[H4(j + flip)]; | ||
276 | float32 e4 = e2; | ||
277 | |||
278 | - d[H4(j)] = float32_muladd(e2, e1, d[H4(j)], 0, fpst); | ||
279 | - d[H4(j + 1)] = float32_muladd(e4, e3, d[H4(j + 1)], 0, fpst); | ||
280 | + d[H4(j)] = float32_muladd(e2, e1, a[H4(j)], 0, fpst); | ||
281 | + d[H4(j + 1)] = float32_muladd(e4, e3, a[H4(j + 1)], 0, fpst); | ||
282 | } | ||
283 | } | ||
284 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
285 | } | ||
286 | |||
287 | -void HELPER(gvec_fcmlad)(void *vd, void *vn, void *vm, | ||
288 | +void HELPER(gvec_fcmlad)(void *vd, void *vn, void *vm, void *va, | ||
289 | void *vfpst, uint32_t desc) | ||
290 | { | ||
291 | uintptr_t opr_sz = simd_oprsz(desc); | ||
292 | - float64 *d = vd; | ||
293 | - float64 *n = vn; | ||
294 | - float64 *m = vm; | ||
295 | + float64 *d = vd, *n = vn, *m = vm, *a = va; | ||
296 | float_status *fpst = vfpst; | ||
297 | intptr_t flip = extract32(desc, SIMD_DATA_SHIFT, 1); | ||
298 | uint64_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1); | ||
299 | @@ -XXX,XX +XXX,XX @@ void HELPER(gvec_fcmlad)(void *vd, void *vn, void *vm, | ||
300 | float64 e4 = e2; | ||
301 | float64 e3 = m[i + 1 - flip] ^ neg_imag; | ||
302 | |||
303 | - d[i] = float64_muladd(e2, e1, d[i], 0, fpst); | ||
304 | - d[i + 1] = float64_muladd(e4, e3, d[i + 1], 0, fpst); | ||
305 | + d[i] = float64_muladd(e2, e1, a[i], 0, fpst); | ||
306 | + d[i + 1] = float64_muladd(e4, e3, a[i + 1], 0, fpst); | ||
307 | } | ||
308 | clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
309 | } | ||
310 | -- | ||
311 | 2.20.1 | ||
312 | |||
313 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Currently only used by FMUL, but will shortly be used more. | ||
4 | |||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Message-id: 20210525010358.152808-52-richard.henderson@linaro.org | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | --- | ||
10 | target/arm/sve.decode | 14 ++++++++++---- | ||
11 | 1 file changed, 10 insertions(+), 4 deletions(-) | ||
12 | |||
13 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
14 | index XXXXXXX..XXXXXXX 100644 | ||
15 | --- a/target/arm/sve.decode | ||
16 | +++ b/target/arm/sve.decode | ||
17 | @@ -XXX,XX +XXX,XX @@ | ||
18 | &rri_esz rd rn imm esz | ||
19 | &rrri_esz rd rn rm imm esz | ||
20 | &rrr_esz rd rn rm esz | ||
21 | +&rrx_esz rd rn rm index esz | ||
22 | &rpr_esz rd pg rn esz | ||
23 | &rpr_s rd pg rn s | ||
24 | &rprr_s rd pg rn rm s | ||
25 | @@ -XXX,XX +XXX,XX @@ | ||
26 | @rpri_scatter_store ....... msz:2 .. imm:5 ... pg:3 rn:5 rd:5 \ | ||
27 | &rpri_scatter_store | ||
28 | |||
29 | +# Two registers and a scalar by N-bit index | ||
30 | +@rrx_3 ........ .. . .. rm:3 ...... rn:5 rd:5 \ | ||
31 | + &rrx_esz index=%index3_22_19 | ||
32 | +@rrx_2 ........ .. . index:2 rm:3 ...... rn:5 rd:5 &rrx_esz | ||
33 | +@rrx_1 ........ .. . index:1 rm:4 ...... rn:5 rd:5 &rrx_esz | ||
34 | + | ||
35 | ########################################################################### | ||
36 | # Instruction patterns. Grouped according to the SVE encodingindex.xhtml. | ||
37 | |||
38 | @@ -XXX,XX +XXX,XX @@ FMLA_zzxz 01100100 111 index:1 rm:4 00000 sub:1 rn:5 rd:5 \ | ||
39 | ### SVE FP Multiply Indexed Group | ||
40 | |||
41 | # SVE floating-point multiply (indexed) | ||
42 | -FMUL_zzx 01100100 0.1 .. rm:3 001000 rn:5 rd:5 \ | ||
43 | - index=%index3_22_19 esz=1 | ||
44 | -FMUL_zzx 01100100 101 index:2 rm:3 001000 rn:5 rd:5 esz=2 | ||
45 | -FMUL_zzx 01100100 111 index:1 rm:4 001000 rn:5 rd:5 esz=3 | ||
46 | +FMUL_zzx 01100100 0. 1 ..... 001000 ..... ..... @rrx_3 esz=1 | ||
47 | +FMUL_zzx 01100100 10 1 ..... 001000 ..... ..... @rrx_2 esz=2 | ||
48 | +FMUL_zzx 01100100 11 1 ..... 001000 ..... ..... @rrx_1 esz=3 | ||
49 | |||
50 | ### SVE FP Fast Reduction Group | ||
51 | |||
52 | -- | ||
53 | 2.20.1 | ||
54 | |||
55 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Used by FMLA and DOT, but will shortly be used more. | ||
4 | Split FMLA from FMLS to avoid an extra sub field; | ||
5 | similarly for SDOT from UDOT. | ||
6 | |||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-id: 20210525010358.152808-53-richard.henderson@linaro.org | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/sve.decode | 29 +++++++++++++++++++---------- | ||
13 | target/arm/translate-sve.c | 38 ++++++++++++++++++++++++++++---------- | ||
14 | 2 files changed, 47 insertions(+), 20 deletions(-) | ||
15 | |||
16 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/target/arm/sve.decode | ||
19 | +++ b/target/arm/sve.decode | ||
20 | @@ -XXX,XX +XXX,XX @@ | ||
21 | &rprr_s rd pg rn rm s | ||
22 | &rprr_esz rd pg rn rm esz | ||
23 | &rrrr_esz rd ra rn rm esz | ||
24 | +&rrxr_esz rd rn rm ra index esz | ||
25 | &rprrr_esz rd pg rn rm ra esz | ||
26 | &rpri_esz rd pg rn imm esz | ||
27 | &ptrue rd esz pat s | ||
28 | @@ -XXX,XX +XXX,XX @@ | ||
29 | @rrx_2 ........ .. . index:2 rm:3 ...... rn:5 rd:5 &rrx_esz | ||
30 | @rrx_1 ........ .. . index:1 rm:4 ...... rn:5 rd:5 &rrx_esz | ||
31 | |||
32 | +# Three registers and a scalar by N-bit index | ||
33 | +@rrxr_3 ........ .. . .. rm:3 ...... rn:5 rd:5 \ | ||
34 | + &rrxr_esz ra=%reg_movprfx index=%index3_22_19 | ||
35 | +@rrxr_2 ........ .. . index:2 rm:3 ...... rn:5 rd:5 \ | ||
36 | + &rrxr_esz ra=%reg_movprfx | ||
37 | +@rrxr_1 ........ .. . index:1 rm:4 ...... rn:5 rd:5 \ | ||
38 | + &rrxr_esz ra=%reg_movprfx | ||
39 | + | ||
40 | ########################################################################### | ||
41 | # Instruction patterns. Grouped according to the SVE encodingindex.xhtml. | ||
42 | |||
43 | @@ -XXX,XX +XXX,XX @@ DOT_zzzz 01000100 1 sz:1 0 rm:5 00000 u:1 rn:5 rd:5 \ | ||
44 | ra=%reg_movprfx | ||
45 | |||
46 | # SVE integer dot product (indexed) | ||
47 | -DOT_zzxw 01000100 101 index:2 rm:3 00000 u:1 rn:5 rd:5 \ | ||
48 | - sz=0 ra=%reg_movprfx | ||
49 | -DOT_zzxw 01000100 111 index:1 rm:4 00000 u:1 rn:5 rd:5 \ | ||
50 | - sz=1 ra=%reg_movprfx | ||
51 | +SDOT_zzxw_s 01000100 10 1 ..... 000000 ..... ..... @rrxr_2 esz=2 | ||
52 | +SDOT_zzxw_d 01000100 11 1 ..... 000000 ..... ..... @rrxr_1 esz=3 | ||
53 | +UDOT_zzxw_s 01000100 10 1 ..... 000001 ..... ..... @rrxr_2 esz=2 | ||
54 | +UDOT_zzxw_d 01000100 11 1 ..... 000001 ..... ..... @rrxr_1 esz=3 | ||
55 | |||
56 | # SVE floating-point complex add (predicated) | ||
57 | FCADD 01100100 esz:2 00000 rot:1 100 pg:3 rm:5 rd:5 \ | ||
58 | @@ -XXX,XX +XXX,XX @@ FCMLA_zzxz 01100100 11 1 index:1 rm:4 0001 rot:2 rn:5 rd:5 \ | ||
59 | ### SVE FP Multiply-Add Indexed Group | ||
60 | |||
61 | # SVE floating-point multiply-add (indexed) | ||
62 | -FMLA_zzxz 01100100 0.1 .. rm:3 00000 sub:1 rn:5 rd:5 \ | ||
63 | - ra=%reg_movprfx index=%index3_22_19 esz=1 | ||
64 | -FMLA_zzxz 01100100 101 index:2 rm:3 00000 sub:1 rn:5 rd:5 \ | ||
65 | - ra=%reg_movprfx esz=2 | ||
66 | -FMLA_zzxz 01100100 111 index:1 rm:4 00000 sub:1 rn:5 rd:5 \ | ||
67 | - ra=%reg_movprfx esz=3 | ||
68 | +FMLA_zzxz 01100100 0. 1 ..... 000000 ..... ..... @rrxr_3 esz=1 | ||
69 | +FMLA_zzxz 01100100 10 1 ..... 000000 ..... ..... @rrxr_2 esz=2 | ||
70 | +FMLA_zzxz 01100100 11 1 ..... 000000 ..... ..... @rrxr_1 esz=3 | ||
71 | +FMLS_zzxz 01100100 0. 1 ..... 000001 ..... ..... @rrxr_3 esz=1 | ||
72 | +FMLS_zzxz 01100100 10 1 ..... 000001 ..... ..... @rrxr_2 esz=2 | ||
73 | +FMLS_zzxz 01100100 11 1 ..... 000001 ..... ..... @rrxr_1 esz=3 | ||
74 | |||
75 | ### SVE FP Multiply Indexed Group | ||
76 | |||
77 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
78 | index XXXXXXX..XXXXXXX 100644 | ||
79 | --- a/target/arm/translate-sve.c | ||
80 | +++ b/target/arm/translate-sve.c | ||
81 | @@ -XXX,XX +XXX,XX @@ static bool trans_DOT_zzzz(DisasContext *s, arg_DOT_zzzz *a) | ||
82 | return true; | ||
83 | } | ||
84 | |||
85 | -static bool trans_DOT_zzxw(DisasContext *s, arg_DOT_zzxw *a) | ||
86 | +static bool do_zzxz_ool(DisasContext *s, arg_rrxr_esz *a, | ||
87 | + gen_helper_gvec_4 *fn) | ||
88 | { | ||
89 | - static gen_helper_gvec_4 * const fns[2][2] = { | ||
90 | - { gen_helper_gvec_sdot_idx_b, gen_helper_gvec_sdot_idx_h }, | ||
91 | - { gen_helper_gvec_udot_idx_b, gen_helper_gvec_udot_idx_h } | ||
92 | - }; | ||
93 | - | ||
94 | + if (fn == NULL) { | ||
95 | + return false; | ||
96 | + } | ||
97 | if (sve_access_check(s)) { | ||
98 | - gen_gvec_ool_zzzz(s, fns[a->u][a->sz], a->rd, a->rn, a->rm, | ||
99 | - a->ra, a->index); | ||
100 | + gen_gvec_ool_zzzz(s, fn, a->rd, a->rn, a->rm, a->ra, a->index); | ||
101 | } | ||
102 | return true; | ||
103 | } | ||
104 | |||
105 | +#define DO_RRXR(NAME, FUNC) \ | ||
106 | + static bool NAME(DisasContext *s, arg_rrxr_esz *a) \ | ||
107 | + { return do_zzxz_ool(s, a, FUNC); } | ||
108 | + | ||
109 | +DO_RRXR(trans_SDOT_zzxw_s, gen_helper_gvec_sdot_idx_b) | ||
110 | +DO_RRXR(trans_SDOT_zzxw_d, gen_helper_gvec_sdot_idx_h) | ||
111 | +DO_RRXR(trans_UDOT_zzxw_s, gen_helper_gvec_udot_idx_b) | ||
112 | +DO_RRXR(trans_UDOT_zzxw_d, gen_helper_gvec_udot_idx_h) | ||
113 | + | ||
114 | +#undef DO_RRXR | ||
115 | |||
116 | /* | ||
117 | *** SVE Floating Point Multiply-Add Indexed Group | ||
118 | */ | ||
119 | |||
120 | -static bool trans_FMLA_zzxz(DisasContext *s, arg_FMLA_zzxz *a) | ||
121 | +static bool do_FMLA_zzxz(DisasContext *s, arg_rrxr_esz *a, bool sub) | ||
122 | { | ||
123 | static gen_helper_gvec_4_ptr * const fns[3] = { | ||
124 | gen_helper_gvec_fmla_idx_h, | ||
125 | @@ -XXX,XX +XXX,XX @@ static bool trans_FMLA_zzxz(DisasContext *s, arg_FMLA_zzxz *a) | ||
126 | vec_full_reg_offset(s, a->rn), | ||
127 | vec_full_reg_offset(s, a->rm), | ||
128 | vec_full_reg_offset(s, a->ra), | ||
129 | - status, vsz, vsz, (a->index << 1) | a->sub, | ||
130 | + status, vsz, vsz, (a->index << 1) | sub, | ||
131 | fns[a->esz - 1]); | ||
132 | tcg_temp_free_ptr(status); | ||
133 | } | ||
134 | return true; | ||
135 | } | ||
136 | |||
137 | +static bool trans_FMLA_zzxz(DisasContext *s, arg_FMLA_zzxz *a) | ||
138 | +{ | ||
139 | + return do_FMLA_zzxz(s, a, false); | ||
140 | +} | ||
141 | + | ||
142 | +static bool trans_FMLS_zzxz(DisasContext *s, arg_FMLA_zzxz *a) | ||
143 | +{ | ||
144 | + return do_FMLA_zzxz(s, a, true); | ||
145 | +} | ||
146 | + | ||
147 | /* | ||
148 | *** SVE Floating Point Multiply Indexed Group | ||
149 | */ | ||
150 | -- | ||
151 | 2.20.1 | ||
152 | |||
153 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-54-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 7 +++++++ | ||
9 | target/arm/translate-sve.c | 30 ++++++++++++++++++++++++++++++ | ||
10 | 2 files changed, 37 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ MUL_zzi 00100101 .. 110 000 110 ........ ..... @rdn_i8s | ||
17 | DOT_zzzz 01000100 1 sz:1 0 rm:5 00000 u:1 rn:5 rd:5 \ | ||
18 | ra=%reg_movprfx | ||
19 | |||
20 | +#### SVE Multiply - Indexed | ||
21 | + | ||
22 | # SVE integer dot product (indexed) | ||
23 | SDOT_zzxw_s 01000100 10 1 ..... 000000 ..... ..... @rrxr_2 esz=2 | ||
24 | SDOT_zzxw_d 01000100 11 1 ..... 000000 ..... ..... @rrxr_1 esz=3 | ||
25 | UDOT_zzxw_s 01000100 10 1 ..... 000001 ..... ..... @rrxr_2 esz=2 | ||
26 | UDOT_zzxw_d 01000100 11 1 ..... 000001 ..... ..... @rrxr_1 esz=3 | ||
27 | |||
28 | +# SVE2 integer multiply (indexed) | ||
29 | +MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
30 | +MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
31 | +MUL_zzx_d 01000100 11 1 ..... 111110 ..... ..... @rrx_1 esz=3 | ||
32 | + | ||
33 | # SVE floating-point complex add (predicated) | ||
34 | FCADD 01100100 esz:2 00000 rot:1 100 pg:3 rm:5 rd:5 \ | ||
35 | rn=%reg_movprfx | ||
36 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/translate-sve.c | ||
39 | +++ b/target/arm/translate-sve.c | ||
40 | @@ -XXX,XX +XXX,XX @@ static bool trans_DOT_zzzz(DisasContext *s, arg_DOT_zzzz *a) | ||
41 | return true; | ||
42 | } | ||
43 | |||
44 | +/* | ||
45 | + * SVE Multiply - Indexed | ||
46 | + */ | ||
47 | + | ||
48 | static bool do_zzxz_ool(DisasContext *s, arg_rrxr_esz *a, | ||
49 | gen_helper_gvec_4 *fn) | ||
50 | { | ||
51 | @@ -XXX,XX +XXX,XX @@ DO_RRXR(trans_UDOT_zzxw_d, gen_helper_gvec_udot_idx_h) | ||
52 | |||
53 | #undef DO_RRXR | ||
54 | |||
55 | +static bool do_sve2_zzz_data(DisasContext *s, int rd, int rn, int rm, int data, | ||
56 | + gen_helper_gvec_3 *fn) | ||
57 | +{ | ||
58 | + if (fn == NULL || !dc_isar_feature(aa64_sve2, s)) { | ||
59 | + return false; | ||
60 | + } | ||
61 | + if (sve_access_check(s)) { | ||
62 | + unsigned vsz = vec_full_reg_size(s); | ||
63 | + tcg_gen_gvec_3_ool(vec_full_reg_offset(s, rd), | ||
64 | + vec_full_reg_offset(s, rn), | ||
65 | + vec_full_reg_offset(s, rm), | ||
66 | + vsz, vsz, data, fn); | ||
67 | + } | ||
68 | + return true; | ||
69 | +} | ||
70 | + | ||
71 | +#define DO_SVE2_RRX(NAME, FUNC) \ | ||
72 | + static bool NAME(DisasContext *s, arg_rrx_esz *a) \ | ||
73 | + { return do_sve2_zzz_data(s, a->rd, a->rn, a->rm, a->index, FUNC); } | ||
74 | + | ||
75 | +DO_SVE2_RRX(trans_MUL_zzx_h, gen_helper_gvec_mul_idx_h) | ||
76 | +DO_SVE2_RRX(trans_MUL_zzx_s, gen_helper_gvec_mul_idx_s) | ||
77 | +DO_SVE2_RRX(trans_MUL_zzx_d, gen_helper_gvec_mul_idx_d) | ||
78 | + | ||
79 | +#undef DO_SVE2_RRX | ||
80 | + | ||
81 | /* | ||
82 | *** SVE Floating Point Multiply-Add Indexed Group | ||
83 | */ | ||
84 | -- | ||
85 | 2.20.1 | ||
86 | |||
87 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-55-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 8 ++++++++ | ||
9 | target/arm/translate-sve.c | 31 +++++++++++++++++++++++++++++++ | ||
10 | 2 files changed, 39 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ SDOT_zzxw_d 01000100 11 1 ..... 000000 ..... ..... @rrxr_1 esz=3 | ||
17 | UDOT_zzxw_s 01000100 10 1 ..... 000001 ..... ..... @rrxr_2 esz=2 | ||
18 | UDOT_zzxw_d 01000100 11 1 ..... 000001 ..... ..... @rrxr_1 esz=3 | ||
19 | |||
20 | +# SVE2 integer multiply-add (indexed) | ||
21 | +MLA_zzxz_h 01000100 0. 1 ..... 000010 ..... ..... @rrxr_3 esz=1 | ||
22 | +MLA_zzxz_s 01000100 10 1 ..... 000010 ..... ..... @rrxr_2 esz=2 | ||
23 | +MLA_zzxz_d 01000100 11 1 ..... 000010 ..... ..... @rrxr_1 esz=3 | ||
24 | +MLS_zzxz_h 01000100 0. 1 ..... 000011 ..... ..... @rrxr_3 esz=1 | ||
25 | +MLS_zzxz_s 01000100 10 1 ..... 000011 ..... ..... @rrxr_2 esz=2 | ||
26 | +MLS_zzxz_d 01000100 11 1 ..... 000011 ..... ..... @rrxr_1 esz=3 | ||
27 | + | ||
28 | # SVE2 integer multiply (indexed) | ||
29 | MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
30 | MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
31 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
32 | index XXXXXXX..XXXXXXX 100644 | ||
33 | --- a/target/arm/translate-sve.c | ||
34 | +++ b/target/arm/translate-sve.c | ||
35 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRX(trans_MUL_zzx_d, gen_helper_gvec_mul_idx_d) | ||
36 | |||
37 | #undef DO_SVE2_RRX | ||
38 | |||
39 | +static bool do_sve2_zzzz_data(DisasContext *s, int rd, int rn, int rm, int ra, | ||
40 | + int data, gen_helper_gvec_4 *fn) | ||
41 | +{ | ||
42 | + if (fn == NULL || !dc_isar_feature(aa64_sve2, s)) { | ||
43 | + return false; | ||
44 | + } | ||
45 | + if (sve_access_check(s)) { | ||
46 | + unsigned vsz = vec_full_reg_size(s); | ||
47 | + tcg_gen_gvec_4_ool(vec_full_reg_offset(s, rd), | ||
48 | + vec_full_reg_offset(s, rn), | ||
49 | + vec_full_reg_offset(s, rm), | ||
50 | + vec_full_reg_offset(s, ra), | ||
51 | + vsz, vsz, data, fn); | ||
52 | + } | ||
53 | + return true; | ||
54 | +} | ||
55 | + | ||
56 | +#define DO_SVE2_RRXR(NAME, FUNC) \ | ||
57 | + static bool NAME(DisasContext *s, arg_rrxr_esz *a) \ | ||
58 | + { return do_sve2_zzzz_data(s, a->rd, a->rn, a->rm, a->ra, a->index, FUNC); } | ||
59 | + | ||
60 | +DO_SVE2_RRXR(trans_MLA_zzxz_h, gen_helper_gvec_mla_idx_h) | ||
61 | +DO_SVE2_RRXR(trans_MLA_zzxz_s, gen_helper_gvec_mla_idx_s) | ||
62 | +DO_SVE2_RRXR(trans_MLA_zzxz_d, gen_helper_gvec_mla_idx_d) | ||
63 | + | ||
64 | +DO_SVE2_RRXR(trans_MLS_zzxz_h, gen_helper_gvec_mls_idx_h) | ||
65 | +DO_SVE2_RRXR(trans_MLS_zzxz_s, gen_helper_gvec_mls_idx_s) | ||
66 | +DO_SVE2_RRXR(trans_MLS_zzxz_d, gen_helper_gvec_mls_idx_d) | ||
67 | + | ||
68 | +#undef DO_SVE2_RRXR | ||
69 | + | ||
70 | /* | ||
71 | *** SVE Floating Point Multiply-Add Indexed Group | ||
72 | */ | ||
73 | -- | ||
74 | 2.20.1 | ||
75 | |||
76 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-56-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 14 ++++++++++++++ | ||
9 | target/arm/sve.decode | 8 ++++++++ | ||
10 | target/arm/sve_helper.c | 36 ++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 8 ++++++++ | ||
12 | 4 files changed, 66 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_zzzz_d, TCG_CALL_NO_RWG, | ||
19 | |||
20 | DEF_HELPER_FLAGS_6(fmmla_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_6(fmmla_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_idx_h, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_idx_s, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlah_idx_d, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | + | ||
30 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_idx_h, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_idx_s, TCG_CALL_NO_RWG, | ||
33 | + void, ptr, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_idx_d, TCG_CALL_NO_RWG, | ||
35 | + void, ptr, ptr, ptr, ptr, i32) | ||
36 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/target/arm/sve.decode | ||
39 | +++ b/target/arm/sve.decode | ||
40 | @@ -XXX,XX +XXX,XX @@ MLS_zzxz_h 01000100 0. 1 ..... 000011 ..... ..... @rrxr_3 esz=1 | ||
41 | MLS_zzxz_s 01000100 10 1 ..... 000011 ..... ..... @rrxr_2 esz=2 | ||
42 | MLS_zzxz_d 01000100 11 1 ..... 000011 ..... ..... @rrxr_1 esz=3 | ||
43 | |||
44 | +# SVE2 saturating multiply-add high (indexed) | ||
45 | +SQRDMLAH_zzxz_h 01000100 0. 1 ..... 000100 ..... ..... @rrxr_3 esz=1 | ||
46 | +SQRDMLAH_zzxz_s 01000100 10 1 ..... 000100 ..... ..... @rrxr_2 esz=2 | ||
47 | +SQRDMLAH_zzxz_d 01000100 11 1 ..... 000100 ..... ..... @rrxr_1 esz=3 | ||
48 | +SQRDMLSH_zzxz_h 01000100 0. 1 ..... 000101 ..... ..... @rrxr_3 esz=1 | ||
49 | +SQRDMLSH_zzxz_s 01000100 10 1 ..... 000101 ..... ..... @rrxr_2 esz=2 | ||
50 | +SQRDMLSH_zzxz_d 01000100 11 1 ..... 000101 ..... ..... @rrxr_1 esz=3 | ||
51 | + | ||
52 | # SVE2 integer multiply (indexed) | ||
53 | MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
54 | MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
55 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
56 | index XXXXXXX..XXXXXXX 100644 | ||
57 | --- a/target/arm/sve_helper.c | ||
58 | +++ b/target/arm/sve_helper.c | ||
59 | @@ -XXX,XX +XXX,XX @@ DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_d, int64_t, , DO_SQRDMLAH_D) | ||
60 | #undef DO_SQRDMLAH_S | ||
61 | #undef DO_SQRDMLAH_D | ||
62 | |||
63 | +#define DO_ZZXZ(NAME, TYPE, H, OP) \ | ||
64 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
65 | +{ \ | ||
66 | + intptr_t oprsz = simd_oprsz(desc), segment = 16 / sizeof(TYPE); \ | ||
67 | + intptr_t i, j, idx = simd_data(desc); \ | ||
68 | + TYPE *d = vd, *a = va, *n = vn, *m = (TYPE *)vm + H(idx); \ | ||
69 | + for (i = 0; i < oprsz / sizeof(TYPE); i += segment) { \ | ||
70 | + TYPE mm = m[i]; \ | ||
71 | + for (j = 0; j < segment; j++) { \ | ||
72 | + d[i + j] = OP(n[i + j], mm, a[i + j]); \ | ||
73 | + } \ | ||
74 | + } \ | ||
75 | +} | ||
76 | + | ||
77 | +#define DO_SQRDMLAH_H(N, M, A) \ | ||
78 | + ({ uint32_t discard; do_sqrdmlah_h(N, M, A, false, true, &discard); }) | ||
79 | +#define DO_SQRDMLAH_S(N, M, A) \ | ||
80 | + ({ uint32_t discard; do_sqrdmlah_s(N, M, A, false, true, &discard); }) | ||
81 | +#define DO_SQRDMLAH_D(N, M, A) do_sqrdmlah_d(N, M, A, false, true) | ||
82 | + | ||
83 | +DO_ZZXZ(sve2_sqrdmlah_idx_h, int16_t, H2, DO_SQRDMLAH_H) | ||
84 | +DO_ZZXZ(sve2_sqrdmlah_idx_s, int32_t, H4, DO_SQRDMLAH_S) | ||
85 | +DO_ZZXZ(sve2_sqrdmlah_idx_d, int64_t, , DO_SQRDMLAH_D) | ||
86 | + | ||
87 | +#define DO_SQRDMLSH_H(N, M, A) \ | ||
88 | + ({ uint32_t discard; do_sqrdmlah_h(N, M, A, true, true, &discard); }) | ||
89 | +#define DO_SQRDMLSH_S(N, M, A) \ | ||
90 | + ({ uint32_t discard; do_sqrdmlah_s(N, M, A, true, true, &discard); }) | ||
91 | +#define DO_SQRDMLSH_D(N, M, A) do_sqrdmlah_d(N, M, A, true, true) | ||
92 | + | ||
93 | +DO_ZZXZ(sve2_sqrdmlsh_idx_h, int16_t, H2, DO_SQRDMLSH_H) | ||
94 | +DO_ZZXZ(sve2_sqrdmlsh_idx_s, int32_t, H4, DO_SQRDMLSH_S) | ||
95 | +DO_ZZXZ(sve2_sqrdmlsh_idx_d, int64_t, , DO_SQRDMLSH_D) | ||
96 | + | ||
97 | +#undef DO_ZZXZ | ||
98 | + | ||
99 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
100 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
101 | { \ | ||
102 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
103 | index XXXXXXX..XXXXXXX 100644 | ||
104 | --- a/target/arm/translate-sve.c | ||
105 | +++ b/target/arm/translate-sve.c | ||
106 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRXR(trans_MLS_zzxz_h, gen_helper_gvec_mls_idx_h) | ||
107 | DO_SVE2_RRXR(trans_MLS_zzxz_s, gen_helper_gvec_mls_idx_s) | ||
108 | DO_SVE2_RRXR(trans_MLS_zzxz_d, gen_helper_gvec_mls_idx_d) | ||
109 | |||
110 | +DO_SVE2_RRXR(trans_SQRDMLAH_zzxz_h, gen_helper_sve2_sqrdmlah_idx_h) | ||
111 | +DO_SVE2_RRXR(trans_SQRDMLAH_zzxz_s, gen_helper_sve2_sqrdmlah_idx_s) | ||
112 | +DO_SVE2_RRXR(trans_SQRDMLAH_zzxz_d, gen_helper_sve2_sqrdmlah_idx_d) | ||
113 | + | ||
114 | +DO_SVE2_RRXR(trans_SQRDMLSH_zzxz_h, gen_helper_sve2_sqrdmlsh_idx_h) | ||
115 | +DO_SVE2_RRXR(trans_SQRDMLSH_zzxz_s, gen_helper_sve2_sqrdmlsh_idx_s) | ||
116 | +DO_SVE2_RRXR(trans_SQRDMLSH_zzxz_d, gen_helper_sve2_sqrdmlsh_idx_d) | ||
117 | + | ||
118 | #undef DO_SVE2_RRXR | ||
119 | |||
120 | /* | ||
121 | -- | ||
122 | 2.20.1 | ||
123 | |||
124 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-57-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 9 +++++++++ | ||
9 | target/arm/sve.decode | 18 ++++++++++++++++++ | ||
10 | target/arm/sve_helper.c | 30 ++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 19 +++++++++++++++++++ | ||
12 | 4 files changed, 76 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_idx_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_sqrdmlsh_idx_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_sqdmlal_idx_s, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_sqdmlal_idx_d, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_sqdmlsl_idx_s, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_5(sve2_sqdmlsl_idx_d, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, i32) | ||
31 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
32 | index XXXXXXX..XXXXXXX 100644 | ||
33 | --- a/target/arm/sve.decode | ||
34 | +++ b/target/arm/sve.decode | ||
35 | @@ -XXX,XX +XXX,XX @@ | ||
36 | %size_23 23:2 | ||
37 | %dtype_23_13 23:2 13:2 | ||
38 | %index3_22_19 22:1 19:2 | ||
39 | +%index3_19_11 19:2 11:1 | ||
40 | +%index2_20_11 20:1 11:1 | ||
41 | |||
42 | # A combination of tsz:imm3 -- extract esize. | ||
43 | %tszimm_esz 22:2 5:5 !function=tszimm_esz | ||
44 | @@ -XXX,XX +XXX,XX @@ | ||
45 | @rrxr_1 ........ .. . index:1 rm:4 ...... rn:5 rd:5 \ | ||
46 | &rrxr_esz ra=%reg_movprfx | ||
47 | |||
48 | +# Three registers and a scalar by N-bit index, alternate | ||
49 | +@rrxr_3a ........ .. ... rm:3 ...... rn:5 rd:5 \ | ||
50 | + &rrxr_esz ra=%reg_movprfx index=%index3_19_11 | ||
51 | +@rrxr_2a ........ .. .. rm:4 ...... rn:5 rd:5 \ | ||
52 | + &rrxr_esz ra=%reg_movprfx index=%index2_20_11 | ||
53 | + | ||
54 | ########################################################################### | ||
55 | # Instruction patterns. Grouped according to the SVE encodingindex.xhtml. | ||
56 | |||
57 | @@ -XXX,XX +XXX,XX @@ SQRDMLSH_zzxz_h 01000100 0. 1 ..... 000101 ..... ..... @rrxr_3 esz=1 | ||
58 | SQRDMLSH_zzxz_s 01000100 10 1 ..... 000101 ..... ..... @rrxr_2 esz=2 | ||
59 | SQRDMLSH_zzxz_d 01000100 11 1 ..... 000101 ..... ..... @rrxr_1 esz=3 | ||
60 | |||
61 | +# SVE2 saturating multiply-add (indexed) | ||
62 | +SQDMLALB_zzxw_s 01000100 10 1 ..... 0010.0 ..... ..... @rrxr_3a esz=2 | ||
63 | +SQDMLALB_zzxw_d 01000100 11 1 ..... 0010.0 ..... ..... @rrxr_2a esz=3 | ||
64 | +SQDMLALT_zzxw_s 01000100 10 1 ..... 0010.1 ..... ..... @rrxr_3a esz=2 | ||
65 | +SQDMLALT_zzxw_d 01000100 11 1 ..... 0010.1 ..... ..... @rrxr_2a esz=3 | ||
66 | +SQDMLSLB_zzxw_s 01000100 10 1 ..... 0011.0 ..... ..... @rrxr_3a esz=2 | ||
67 | +SQDMLSLB_zzxw_d 01000100 11 1 ..... 0011.0 ..... ..... @rrxr_2a esz=3 | ||
68 | +SQDMLSLT_zzxw_s 01000100 10 1 ..... 0011.1 ..... ..... @rrxr_3a esz=2 | ||
69 | +SQDMLSLT_zzxw_d 01000100 11 1 ..... 0011.1 ..... ..... @rrxr_2a esz=3 | ||
70 | + | ||
71 | # SVE2 integer multiply (indexed) | ||
72 | MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
73 | MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
74 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
75 | index XXXXXXX..XXXXXXX 100644 | ||
76 | --- a/target/arm/sve_helper.c | ||
77 | +++ b/target/arm/sve_helper.c | ||
78 | @@ -XXX,XX +XXX,XX @@ DO_ZZXZ(sve2_sqrdmlsh_idx_d, int64_t, , DO_SQRDMLSH_D) | ||
79 | |||
80 | #undef DO_ZZXZ | ||
81 | |||
82 | +#define DO_ZZXW(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
83 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
84 | +{ \ | ||
85 | + intptr_t i, j, oprsz = simd_oprsz(desc); \ | ||
86 | + intptr_t sel = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPEN); \ | ||
87 | + intptr_t idx = extract32(desc, SIMD_DATA_SHIFT + 1, 3) * sizeof(TYPEN); \ | ||
88 | + for (i = 0; i < oprsz; i += 16) { \ | ||
89 | + TYPEW mm = *(TYPEN *)(vm + HN(i + idx)); \ | ||
90 | + for (j = 0; j < 16; j += sizeof(TYPEW)) { \ | ||
91 | + TYPEW nn = *(TYPEN *)(vn + HN(i + j + sel)); \ | ||
92 | + TYPEW aa = *(TYPEW *)(va + HW(i + j)); \ | ||
93 | + *(TYPEW *)(vd + HW(i + j)) = OP(nn, mm, aa); \ | ||
94 | + } \ | ||
95 | + } \ | ||
96 | +} | ||
97 | + | ||
98 | +#define DO_SQDMLAL_S(N, M, A) DO_SQADD_S(A, do_sqdmull_s(N, M)) | ||
99 | +#define DO_SQDMLAL_D(N, M, A) do_sqadd_d(A, do_sqdmull_d(N, M)) | ||
100 | + | ||
101 | +DO_ZZXW(sve2_sqdmlal_idx_s, int32_t, int16_t, H1_4, H1_2, DO_SQDMLAL_S) | ||
102 | +DO_ZZXW(sve2_sqdmlal_idx_d, int64_t, int32_t, , H1_4, DO_SQDMLAL_D) | ||
103 | + | ||
104 | +#define DO_SQDMLSL_S(N, M, A) DO_SQSUB_S(A, do_sqdmull_s(N, M)) | ||
105 | +#define DO_SQDMLSL_D(N, M, A) do_sqsub_d(A, do_sqdmull_d(N, M)) | ||
106 | + | ||
107 | +DO_ZZXW(sve2_sqdmlsl_idx_s, int32_t, int16_t, H1_4, H1_2, DO_SQDMLSL_S) | ||
108 | +DO_ZZXW(sve2_sqdmlsl_idx_d, int64_t, int32_t, , H1_4, DO_SQDMLSL_D) | ||
109 | + | ||
110 | +#undef DO_ZZXW | ||
111 | + | ||
112 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
113 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
114 | { \ | ||
115 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
116 | index XXXXXXX..XXXXXXX 100644 | ||
117 | --- a/target/arm/translate-sve.c | ||
118 | +++ b/target/arm/translate-sve.c | ||
119 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRXR(trans_SQRDMLSH_zzxz_d, gen_helper_sve2_sqrdmlsh_idx_d) | ||
120 | |||
121 | #undef DO_SVE2_RRXR | ||
122 | |||
123 | +#define DO_SVE2_RRXR_TB(NAME, FUNC, TOP) \ | ||
124 | + static bool NAME(DisasContext *s, arg_rrxr_esz *a) \ | ||
125 | + { \ | ||
126 | + return do_sve2_zzzz_data(s, a->rd, a->rn, a->rm, a->rd, \ | ||
127 | + (a->index << 1) | TOP, FUNC); \ | ||
128 | + } | ||
129 | + | ||
130 | +DO_SVE2_RRXR_TB(trans_SQDMLALB_zzxw_s, gen_helper_sve2_sqdmlal_idx_s, false) | ||
131 | +DO_SVE2_RRXR_TB(trans_SQDMLALB_zzxw_d, gen_helper_sve2_sqdmlal_idx_d, false) | ||
132 | +DO_SVE2_RRXR_TB(trans_SQDMLALT_zzxw_s, gen_helper_sve2_sqdmlal_idx_s, true) | ||
133 | +DO_SVE2_RRXR_TB(trans_SQDMLALT_zzxw_d, gen_helper_sve2_sqdmlal_idx_d, true) | ||
134 | + | ||
135 | +DO_SVE2_RRXR_TB(trans_SQDMLSLB_zzxw_s, gen_helper_sve2_sqdmlsl_idx_s, false) | ||
136 | +DO_SVE2_RRXR_TB(trans_SQDMLSLB_zzxw_d, gen_helper_sve2_sqdmlsl_idx_d, false) | ||
137 | +DO_SVE2_RRXR_TB(trans_SQDMLSLT_zzxw_s, gen_helper_sve2_sqdmlsl_idx_s, true) | ||
138 | +DO_SVE2_RRXR_TB(trans_SQDMLSLT_zzxw_d, gen_helper_sve2_sqdmlsl_idx_d, true) | ||
139 | + | ||
140 | +#undef DO_SVE2_RRXR_TB | ||
141 | + | ||
142 | /* | ||
143 | *** SVE Floating Point Multiply-Add Indexed Group | ||
144 | */ | ||
145 | -- | ||
146 | 2.20.1 | ||
147 | |||
148 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-58-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 5 +++++ | ||
9 | target/arm/sve.decode | 12 ++++++++++++ | ||
10 | target/arm/sve_helper.c | 20 ++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 14 ++++++++++++++ | ||
12 | 4 files changed, 51 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqdmlsl_idx_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_sqdmlsl_idx_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_4(sve2_sqdmull_idx_s, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_sqdmull_idx_d, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, i32) | ||
27 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
28 | index XXXXXXX..XXXXXXX 100644 | ||
29 | --- a/target/arm/sve.decode | ||
30 | +++ b/target/arm/sve.decode | ||
31 | @@ -XXX,XX +XXX,XX @@ | ||
32 | @rrx_2 ........ .. . index:2 rm:3 ...... rn:5 rd:5 &rrx_esz | ||
33 | @rrx_1 ........ .. . index:1 rm:4 ...... rn:5 rd:5 &rrx_esz | ||
34 | |||
35 | +# Two registers and a scalar by N-bit index, alternate | ||
36 | +@rrx_3a ........ .. . .. rm:3 ...... rn:5 rd:5 \ | ||
37 | + &rrx_esz index=%index3_19_11 | ||
38 | +@rrx_2a ........ .. . . rm:4 ...... rn:5 rd:5 \ | ||
39 | + &rrx_esz index=%index2_20_11 | ||
40 | + | ||
41 | # Three registers and a scalar by N-bit index | ||
42 | @rrxr_3 ........ .. . .. rm:3 ...... rn:5 rd:5 \ | ||
43 | &rrxr_esz ra=%reg_movprfx index=%index3_22_19 | ||
44 | @@ -XXX,XX +XXX,XX @@ SQDMLSLB_zzxw_d 01000100 11 1 ..... 0011.0 ..... ..... @rrxr_2a esz=3 | ||
45 | SQDMLSLT_zzxw_s 01000100 10 1 ..... 0011.1 ..... ..... @rrxr_3a esz=2 | ||
46 | SQDMLSLT_zzxw_d 01000100 11 1 ..... 0011.1 ..... ..... @rrxr_2a esz=3 | ||
47 | |||
48 | +# SVE2 saturating multiply (indexed) | ||
49 | +SQDMULLB_zzx_s 01000100 10 1 ..... 1110.0 ..... ..... @rrx_3a esz=2 | ||
50 | +SQDMULLB_zzx_d 01000100 11 1 ..... 1110.0 ..... ..... @rrx_2a esz=3 | ||
51 | +SQDMULLT_zzx_s 01000100 10 1 ..... 1110.1 ..... ..... @rrx_3a esz=2 | ||
52 | +SQDMULLT_zzx_d 01000100 11 1 ..... 1110.1 ..... ..... @rrx_2a esz=3 | ||
53 | + | ||
54 | # SVE2 integer multiply (indexed) | ||
55 | MUL_zzx_h 01000100 0. 1 ..... 111110 ..... ..... @rrx_3 esz=1 | ||
56 | MUL_zzx_s 01000100 10 1 ..... 111110 ..... ..... @rrx_2 esz=2 | ||
57 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | ||
59 | --- a/target/arm/sve_helper.c | ||
60 | +++ b/target/arm/sve_helper.c | ||
61 | @@ -XXX,XX +XXX,XX @@ DO_ZZXW(sve2_sqdmlsl_idx_d, int64_t, int32_t, , H1_4, DO_SQDMLSL_D) | ||
62 | |||
63 | #undef DO_ZZXW | ||
64 | |||
65 | +#define DO_ZZX(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
66 | +void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
67 | +{ \ | ||
68 | + intptr_t i, j, oprsz = simd_oprsz(desc); \ | ||
69 | + intptr_t sel = extract32(desc, SIMD_DATA_SHIFT, 1) * sizeof(TYPEN); \ | ||
70 | + intptr_t idx = extract32(desc, SIMD_DATA_SHIFT + 1, 3) * sizeof(TYPEN); \ | ||
71 | + for (i = 0; i < oprsz; i += 16) { \ | ||
72 | + TYPEW mm = *(TYPEN *)(vm + HN(i + idx)); \ | ||
73 | + for (j = 0; j < 16; j += sizeof(TYPEW)) { \ | ||
74 | + TYPEW nn = *(TYPEN *)(vn + HN(i + j + sel)); \ | ||
75 | + *(TYPEW *)(vd + HW(i + j)) = OP(nn, mm); \ | ||
76 | + } \ | ||
77 | + } \ | ||
78 | +} | ||
79 | + | ||
80 | +DO_ZZX(sve2_sqdmull_idx_s, int32_t, int16_t, H1_4, H1_2, do_sqdmull_s) | ||
81 | +DO_ZZX(sve2_sqdmull_idx_d, int64_t, int32_t, , H1_4, do_sqdmull_d) | ||
82 | + | ||
83 | +#undef DO_ZZX | ||
84 | + | ||
85 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
86 | void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
87 | { \ | ||
88 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
89 | index XXXXXXX..XXXXXXX 100644 | ||
90 | --- a/target/arm/translate-sve.c | ||
91 | +++ b/target/arm/translate-sve.c | ||
92 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRX(trans_MUL_zzx_d, gen_helper_gvec_mul_idx_d) | ||
93 | |||
94 | #undef DO_SVE2_RRX | ||
95 | |||
96 | +#define DO_SVE2_RRX_TB(NAME, FUNC, TOP) \ | ||
97 | + static bool NAME(DisasContext *s, arg_rrx_esz *a) \ | ||
98 | + { \ | ||
99 | + return do_sve2_zzz_data(s, a->rd, a->rn, a->rm, \ | ||
100 | + (a->index << 1) | TOP, FUNC); \ | ||
101 | + } | ||
102 | + | ||
103 | +DO_SVE2_RRX_TB(trans_SQDMULLB_zzx_s, gen_helper_sve2_sqdmull_idx_s, false) | ||
104 | +DO_SVE2_RRX_TB(trans_SQDMULLB_zzx_d, gen_helper_sve2_sqdmull_idx_d, false) | ||
105 | +DO_SVE2_RRX_TB(trans_SQDMULLT_zzx_s, gen_helper_sve2_sqdmull_idx_s, true) | ||
106 | +DO_SVE2_RRX_TB(trans_SQDMULLT_zzx_d, gen_helper_sve2_sqdmull_idx_d, true) | ||
107 | + | ||
108 | +#undef DO_SVE2_RRX_TB | ||
109 | + | ||
110 | static bool do_sve2_zzzz_data(DisasContext *s, int rd, int rn, int rm, int ra, | ||
111 | int data, gen_helper_gvec_4 *fn) | ||
112 | { | ||
113 | -- | ||
114 | 2.20.1 | ||
115 | |||
116 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-59-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper.h | 10 +++++ | ||
9 | target/arm/sve.decode | 4 ++ | ||
10 | target/arm/translate-sve.c | 18 ++++++++ | ||
11 | target/arm/vec_helper.c | 84 ++++++++++++++++++++++++++++++++++++++ | ||
12 | 4 files changed, 116 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper.h | ||
17 | +++ b/target/arm/helper.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(neon_sqrdmulh_h, TCG_CALL_NO_RWG, | ||
19 | DEF_HELPER_FLAGS_5(neon_sqrdmulh_s, TCG_CALL_NO_RWG, | ||
20 | void, ptr, ptr, ptr, ptr, i32) | ||
21 | |||
22 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_sqdmulh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | + | ||
27 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_4(sve2_sqrdmulh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
31 | + | ||
32 | DEF_HELPER_FLAGS_4(gvec_xar_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | |||
34 | #ifdef TARGET_AARCH64 | ||
35 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
36 | index XXXXXXX..XXXXXXX 100644 | ||
37 | --- a/target/arm/sve.decode | ||
38 | +++ b/target/arm/sve.decode | ||
39 | @@ -XXX,XX +XXX,XX @@ SMULH_zzz 00000100 .. 1 ..... 0110 10 ..... ..... @rd_rn_rm | ||
40 | UMULH_zzz 00000100 .. 1 ..... 0110 11 ..... ..... @rd_rn_rm | ||
41 | PMUL_zzz 00000100 00 1 ..... 0110 01 ..... ..... @rd_rn_rm_e0 | ||
42 | |||
43 | +# SVE2 signed saturating doubling multiply high (unpredicated) | ||
44 | +SQDMULH_zzz 00000100 .. 1 ..... 0111 00 ..... ..... @rd_rn_rm | ||
45 | +SQRDMULH_zzz 00000100 .. 1 ..... 0111 01 ..... ..... @rd_rn_rm | ||
46 | + | ||
47 | ### SVE2 Integer - Predicated | ||
48 | |||
49 | SADALP_zpzz 01000100 .. 000 100 101 ... ..... ..... @rdm_pg_rn | ||
50 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/target/arm/translate-sve.c | ||
53 | +++ b/target/arm/translate-sve.c | ||
54 | @@ -XXX,XX +XXX,XX @@ static bool trans_PMUL_zzz(DisasContext *s, arg_rrr_esz *a) | ||
55 | return do_sve2_zzz_ool(s, a, gen_helper_gvec_pmul_b); | ||
56 | } | ||
57 | |||
58 | +static bool trans_SQDMULH_zzz(DisasContext *s, arg_rrr_esz *a) | ||
59 | +{ | ||
60 | + static gen_helper_gvec_3 * const fns[4] = { | ||
61 | + gen_helper_sve2_sqdmulh_b, gen_helper_sve2_sqdmulh_h, | ||
62 | + gen_helper_sve2_sqdmulh_s, gen_helper_sve2_sqdmulh_d, | ||
63 | + }; | ||
64 | + return do_sve2_zzz_ool(s, a, fns[a->esz]); | ||
65 | +} | ||
66 | + | ||
67 | +static bool trans_SQRDMULH_zzz(DisasContext *s, arg_rrr_esz *a) | ||
68 | +{ | ||
69 | + static gen_helper_gvec_3 * const fns[4] = { | ||
70 | + gen_helper_sve2_sqrdmulh_b, gen_helper_sve2_sqrdmulh_h, | ||
71 | + gen_helper_sve2_sqrdmulh_s, gen_helper_sve2_sqrdmulh_d, | ||
72 | + }; | ||
73 | + return do_sve2_zzz_ool(s, a, fns[a->esz]); | ||
74 | +} | ||
75 | + | ||
76 | /* | ||
77 | * SVE2 Integer - Predicated | ||
78 | */ | ||
79 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
80 | index XXXXXXX..XXXXXXX 100644 | ||
81 | --- a/target/arm/vec_helper.c | ||
82 | +++ b/target/arm/vec_helper.c | ||
83 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_b)(void *vd, void *vn, void *vm, | ||
84 | } | ||
85 | } | ||
86 | |||
87 | +void HELPER(sve2_sqdmulh_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
88 | +{ | ||
89 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
90 | + int8_t *d = vd, *n = vn, *m = vm; | ||
91 | + | ||
92 | + for (i = 0; i < opr_sz; ++i) { | ||
93 | + d[i] = do_sqrdmlah_b(n[i], m[i], 0, false, false); | ||
94 | + } | ||
95 | +} | ||
96 | + | ||
97 | +void HELPER(sve2_sqrdmulh_b)(void *vd, void *vn, void *vm, uint32_t desc) | ||
98 | +{ | ||
99 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
100 | + int8_t *d = vd, *n = vn, *m = vm; | ||
101 | + | ||
102 | + for (i = 0; i < opr_sz; ++i) { | ||
103 | + d[i] = do_sqrdmlah_b(n[i], m[i], 0, false, true); | ||
104 | + } | ||
105 | +} | ||
106 | + | ||
107 | /* Signed saturating rounding doubling multiply-accumulate high half, 16-bit */ | ||
108 | int16_t do_sqrdmlah_h(int16_t src1, int16_t src2, int16_t src3, | ||
109 | bool neg, bool round, uint32_t *sat) | ||
110 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_h)(void *vd, void *vn, void *vm, | ||
111 | } | ||
112 | } | ||
113 | |||
114 | +void HELPER(sve2_sqdmulh_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
115 | +{ | ||
116 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
117 | + int16_t *d = vd, *n = vn, *m = vm; | ||
118 | + uint32_t discard; | ||
119 | + | ||
120 | + for (i = 0; i < opr_sz / 2; ++i) { | ||
121 | + d[i] = do_sqrdmlah_h(n[i], m[i], 0, false, false, &discard); | ||
122 | + } | ||
123 | +} | ||
124 | + | ||
125 | +void HELPER(sve2_sqrdmulh_h)(void *vd, void *vn, void *vm, uint32_t desc) | ||
126 | +{ | ||
127 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
128 | + int16_t *d = vd, *n = vn, *m = vm; | ||
129 | + uint32_t discard; | ||
130 | + | ||
131 | + for (i = 0; i < opr_sz / 2; ++i) { | ||
132 | + d[i] = do_sqrdmlah_h(n[i], m[i], 0, false, true, &discard); | ||
133 | + } | ||
134 | +} | ||
135 | + | ||
136 | /* Signed saturating rounding doubling multiply-accumulate high half, 32-bit */ | ||
137 | int32_t do_sqrdmlah_s(int32_t src1, int32_t src2, int32_t src3, | ||
138 | bool neg, bool round, uint32_t *sat) | ||
139 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_s)(void *vd, void *vn, void *vm, | ||
140 | } | ||
141 | } | ||
142 | |||
143 | +void HELPER(sve2_sqdmulh_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
144 | +{ | ||
145 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
146 | + int32_t *d = vd, *n = vn, *m = vm; | ||
147 | + uint32_t discard; | ||
148 | + | ||
149 | + for (i = 0; i < opr_sz / 4; ++i) { | ||
150 | + d[i] = do_sqrdmlah_s(n[i], m[i], 0, false, false, &discard); | ||
151 | + } | ||
152 | +} | ||
153 | + | ||
154 | +void HELPER(sve2_sqrdmulh_s)(void *vd, void *vn, void *vm, uint32_t desc) | ||
155 | +{ | ||
156 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
157 | + int32_t *d = vd, *n = vn, *m = vm; | ||
158 | + uint32_t discard; | ||
159 | + | ||
160 | + for (i = 0; i < opr_sz / 4; ++i) { | ||
161 | + d[i] = do_sqrdmlah_s(n[i], m[i], 0, false, true, &discard); | ||
162 | + } | ||
163 | +} | ||
164 | + | ||
165 | /* Signed saturating rounding doubling multiply-accumulate high half, 64-bit */ | ||
166 | static int64_t do_sat128_d(Int128 r) | ||
167 | { | ||
168 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmlsh_d)(void *vd, void *vn, void *vm, | ||
169 | } | ||
170 | } | ||
171 | |||
172 | +void HELPER(sve2_sqdmulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
173 | +{ | ||
174 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
175 | + int64_t *d = vd, *n = vn, *m = vm; | ||
176 | + | ||
177 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
178 | + d[i] = do_sqrdmlah_d(n[i], m[i], 0, false, false); | ||
179 | + } | ||
180 | +} | ||
181 | + | ||
182 | +void HELPER(sve2_sqrdmulh_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
183 | +{ | ||
184 | + intptr_t i, opr_sz = simd_oprsz(desc); | ||
185 | + int64_t *d = vd, *n = vn, *m = vm; | ||
186 | + | ||
187 | + for (i = 0; i < opr_sz / 8; ++i) { | ||
188 | + d[i] = do_sqrdmlah_d(n[i], m[i], 0, false, true); | ||
189 | + } | ||
190 | +} | ||
191 | + | ||
192 | /* Integer 8 and 16-bit dot-product. | ||
193 | * | ||
194 | * Note that for the loops herein, host endianness does not matter | ||
195 | -- | ||
196 | 2.20.1 | ||
197 | |||
198 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Message-id: 20210525010358.152808-61-richard.henderson@linaro.org | ||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 17 +++++++++++++++++ | ||
9 | target/arm/sve.decode | 18 ++++++++++++++++++ | ||
10 | target/arm/sve_helper.c | 16 ++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 20 ++++++++++++++++++++ | ||
12 | 4 files changed, 71 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_sqdmull_idx_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_sqdmull_idx_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_smlal_idx_s, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_smlal_idx_d, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_smlsl_idx_s, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_5(sve2_smlsl_idx_d, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, i32) | ||
31 | +DEF_HELPER_FLAGS_5(sve2_umlal_idx_s, TCG_CALL_NO_RWG, | ||
32 | + void, ptr, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_5(sve2_umlal_idx_d, TCG_CALL_NO_RWG, | ||
34 | + void, ptr, ptr, ptr, ptr, i32) | ||
35 | +DEF_HELPER_FLAGS_5(sve2_umlsl_idx_s, TCG_CALL_NO_RWG, | ||
36 | + void, ptr, ptr, ptr, ptr, i32) | ||
37 | +DEF_HELPER_FLAGS_5(sve2_umlsl_idx_d, TCG_CALL_NO_RWG, | ||
38 | + void, ptr, ptr, ptr, ptr, i32) | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ SQDMLSLB_zzxw_d 01000100 11 1 ..... 0011.0 ..... ..... @rrxr_2a esz=3 | ||
44 | SQDMLSLT_zzxw_s 01000100 10 1 ..... 0011.1 ..... ..... @rrxr_3a esz=2 | ||
45 | SQDMLSLT_zzxw_d 01000100 11 1 ..... 0011.1 ..... ..... @rrxr_2a esz=3 | ||
46 | |||
47 | +# SVE2 multiply-add long (indexed) | ||
48 | +SMLALB_zzxw_s 01000100 10 1 ..... 1000.0 ..... ..... @rrxr_3a esz=2 | ||
49 | +SMLALB_zzxw_d 01000100 11 1 ..... 1000.0 ..... ..... @rrxr_2a esz=3 | ||
50 | +SMLALT_zzxw_s 01000100 10 1 ..... 1000.1 ..... ..... @rrxr_3a esz=2 | ||
51 | +SMLALT_zzxw_d 01000100 11 1 ..... 1000.1 ..... ..... @rrxr_2a esz=3 | ||
52 | +UMLALB_zzxw_s 01000100 10 1 ..... 1001.0 ..... ..... @rrxr_3a esz=2 | ||
53 | +UMLALB_zzxw_d 01000100 11 1 ..... 1001.0 ..... ..... @rrxr_2a esz=3 | ||
54 | +UMLALT_zzxw_s 01000100 10 1 ..... 1001.1 ..... ..... @rrxr_3a esz=2 | ||
55 | +UMLALT_zzxw_d 01000100 11 1 ..... 1001.1 ..... ..... @rrxr_2a esz=3 | ||
56 | +SMLSLB_zzxw_s 01000100 10 1 ..... 1010.0 ..... ..... @rrxr_3a esz=2 | ||
57 | +SMLSLB_zzxw_d 01000100 11 1 ..... 1010.0 ..... ..... @rrxr_2a esz=3 | ||
58 | +SMLSLT_zzxw_s 01000100 10 1 ..... 1010.1 ..... ..... @rrxr_3a esz=2 | ||
59 | +SMLSLT_zzxw_d 01000100 11 1 ..... 1010.1 ..... ..... @rrxr_2a esz=3 | ||
60 | +UMLSLB_zzxw_s 01000100 10 1 ..... 1011.0 ..... ..... @rrxr_3a esz=2 | ||
61 | +UMLSLB_zzxw_d 01000100 11 1 ..... 1011.0 ..... ..... @rrxr_2a esz=3 | ||
62 | +UMLSLT_zzxw_s 01000100 10 1 ..... 1011.1 ..... ..... @rrxr_3a esz=2 | ||
63 | +UMLSLT_zzxw_d 01000100 11 1 ..... 1011.1 ..... ..... @rrxr_2a esz=3 | ||
64 | + | ||
65 | # SVE2 saturating multiply (indexed) | ||
66 | SQDMULLB_zzx_s 01000100 10 1 ..... 1110.0 ..... ..... @rrx_3a esz=2 | ||
67 | SQDMULLB_zzx_d 01000100 11 1 ..... 1110.0 ..... ..... @rrx_2a esz=3 | ||
68 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
69 | index XXXXXXX..XXXXXXX 100644 | ||
70 | --- a/target/arm/sve_helper.c | ||
71 | +++ b/target/arm/sve_helper.c | ||
72 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
73 | } \ | ||
74 | } | ||
75 | |||
76 | +#define DO_MLA(N, M, A) (A + N * M) | ||
77 | + | ||
78 | +DO_ZZXW(sve2_smlal_idx_s, int32_t, int16_t, H1_4, H1_2, DO_MLA) | ||
79 | +DO_ZZXW(sve2_smlal_idx_d, int64_t, int32_t, , H1_4, DO_MLA) | ||
80 | +DO_ZZXW(sve2_umlal_idx_s, uint32_t, uint16_t, H1_4, H1_2, DO_MLA) | ||
81 | +DO_ZZXW(sve2_umlal_idx_d, uint64_t, uint32_t, , H1_4, DO_MLA) | ||
82 | + | ||
83 | +#define DO_MLS(N, M, A) (A - N * M) | ||
84 | + | ||
85 | +DO_ZZXW(sve2_smlsl_idx_s, int32_t, int16_t, H1_4, H1_2, DO_MLS) | ||
86 | +DO_ZZXW(sve2_smlsl_idx_d, int64_t, int32_t, , H1_4, DO_MLS) | ||
87 | +DO_ZZXW(sve2_umlsl_idx_s, uint32_t, uint16_t, H1_4, H1_2, DO_MLS) | ||
88 | +DO_ZZXW(sve2_umlsl_idx_d, uint64_t, uint32_t, , H1_4, DO_MLS) | ||
89 | + | ||
90 | #define DO_SQDMLAL_S(N, M, A) DO_SQADD_S(A, do_sqdmull_s(N, M)) | ||
91 | #define DO_SQDMLAL_D(N, M, A) do_sqadd_d(A, do_sqdmull_d(N, M)) | ||
92 | |||
93 | @@ -XXX,XX +XXX,XX @@ DO_ZZXW(sve2_sqdmlal_idx_d, int64_t, int32_t, , H1_4, DO_SQDMLAL_D) | ||
94 | DO_ZZXW(sve2_sqdmlsl_idx_s, int32_t, int16_t, H1_4, H1_2, DO_SQDMLSL_S) | ||
95 | DO_ZZXW(sve2_sqdmlsl_idx_d, int64_t, int32_t, , H1_4, DO_SQDMLSL_D) | ||
96 | |||
97 | +#undef DO_MLA | ||
98 | +#undef DO_MLS | ||
99 | #undef DO_ZZXW | ||
100 | |||
101 | #define DO_ZZX(NAME, TYPEW, TYPEN, HW, HN, OP) \ | ||
102 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
103 | index XXXXXXX..XXXXXXX 100644 | ||
104 | --- a/target/arm/translate-sve.c | ||
105 | +++ b/target/arm/translate-sve.c | ||
106 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRXR_TB(trans_SQDMLSLB_zzxw_d, gen_helper_sve2_sqdmlsl_idx_d, false) | ||
107 | DO_SVE2_RRXR_TB(trans_SQDMLSLT_zzxw_s, gen_helper_sve2_sqdmlsl_idx_s, true) | ||
108 | DO_SVE2_RRXR_TB(trans_SQDMLSLT_zzxw_d, gen_helper_sve2_sqdmlsl_idx_d, true) | ||
109 | |||
110 | +DO_SVE2_RRXR_TB(trans_SMLALB_zzxw_s, gen_helper_sve2_smlal_idx_s, false) | ||
111 | +DO_SVE2_RRXR_TB(trans_SMLALB_zzxw_d, gen_helper_sve2_smlal_idx_d, false) | ||
112 | +DO_SVE2_RRXR_TB(trans_SMLALT_zzxw_s, gen_helper_sve2_smlal_idx_s, true) | ||
113 | +DO_SVE2_RRXR_TB(trans_SMLALT_zzxw_d, gen_helper_sve2_smlal_idx_d, true) | ||
114 | + | ||
115 | +DO_SVE2_RRXR_TB(trans_UMLALB_zzxw_s, gen_helper_sve2_umlal_idx_s, false) | ||
116 | +DO_SVE2_RRXR_TB(trans_UMLALB_zzxw_d, gen_helper_sve2_umlal_idx_d, false) | ||
117 | +DO_SVE2_RRXR_TB(trans_UMLALT_zzxw_s, gen_helper_sve2_umlal_idx_s, true) | ||
118 | +DO_SVE2_RRXR_TB(trans_UMLALT_zzxw_d, gen_helper_sve2_umlal_idx_d, true) | ||
119 | + | ||
120 | +DO_SVE2_RRXR_TB(trans_SMLSLB_zzxw_s, gen_helper_sve2_smlsl_idx_s, false) | ||
121 | +DO_SVE2_RRXR_TB(trans_SMLSLB_zzxw_d, gen_helper_sve2_smlsl_idx_d, false) | ||
122 | +DO_SVE2_RRXR_TB(trans_SMLSLT_zzxw_s, gen_helper_sve2_smlsl_idx_s, true) | ||
123 | +DO_SVE2_RRXR_TB(trans_SMLSLT_zzxw_d, gen_helper_sve2_smlsl_idx_d, true) | ||
124 | + | ||
125 | +DO_SVE2_RRXR_TB(trans_UMLSLB_zzxw_s, gen_helper_sve2_umlsl_idx_s, false) | ||
126 | +DO_SVE2_RRXR_TB(trans_UMLSLB_zzxw_d, gen_helper_sve2_umlsl_idx_d, false) | ||
127 | +DO_SVE2_RRXR_TB(trans_UMLSLT_zzxw_s, gen_helper_sve2_umlsl_idx_s, true) | ||
128 | +DO_SVE2_RRXR_TB(trans_UMLSLT_zzxw_d, gen_helper_sve2_umlsl_idx_d, true) | ||
129 | + | ||
130 | #undef DO_SVE2_RRXR_TB | ||
131 | |||
132 | /* | ||
133 | -- | ||
134 | 2.20.1 | ||
135 | |||
136 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Message-id: 20210525010358.152808-62-richard.henderson@linaro.org | ||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 5 +++++ | ||
9 | target/arm/sve.decode | 10 ++++++++++ | ||
10 | target/arm/sve_helper.c | 6 ++++++ | ||
11 | target/arm/translate-sve.c | 10 ++++++++++ | ||
12 | 4 files changed, 31 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_umlsl_idx_s, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_umlsl_idx_d, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_4(sve2_smull_idx_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | +DEF_HELPER_FLAGS_4(sve2_smull_idx_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_4(sve2_umull_idx_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
26 | +DEF_HELPER_FLAGS_4(sve2_umull_idx_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
27 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
28 | index XXXXXXX..XXXXXXX 100644 | ||
29 | --- a/target/arm/sve.decode | ||
30 | +++ b/target/arm/sve.decode | ||
31 | @@ -XXX,XX +XXX,XX @@ UMLSLB_zzxw_d 01000100 11 1 ..... 1011.0 ..... ..... @rrxr_2a esz=3 | ||
32 | UMLSLT_zzxw_s 01000100 10 1 ..... 1011.1 ..... ..... @rrxr_3a esz=2 | ||
33 | UMLSLT_zzxw_d 01000100 11 1 ..... 1011.1 ..... ..... @rrxr_2a esz=3 | ||
34 | |||
35 | +# SVE2 integer multiply long (indexed) | ||
36 | +SMULLB_zzx_s 01000100 10 1 ..... 1100.0 ..... ..... @rrx_3a esz=2 | ||
37 | +SMULLB_zzx_d 01000100 11 1 ..... 1100.0 ..... ..... @rrx_2a esz=3 | ||
38 | +SMULLT_zzx_s 01000100 10 1 ..... 1100.1 ..... ..... @rrx_3a esz=2 | ||
39 | +SMULLT_zzx_d 01000100 11 1 ..... 1100.1 ..... ..... @rrx_2a esz=3 | ||
40 | +UMULLB_zzx_s 01000100 10 1 ..... 1101.0 ..... ..... @rrx_3a esz=2 | ||
41 | +UMULLB_zzx_d 01000100 11 1 ..... 1101.0 ..... ..... @rrx_2a esz=3 | ||
42 | +UMULLT_zzx_s 01000100 10 1 ..... 1101.1 ..... ..... @rrx_3a esz=2 | ||
43 | +UMULLT_zzx_d 01000100 11 1 ..... 1101.1 ..... ..... @rrx_2a esz=3 | ||
44 | + | ||
45 | # SVE2 saturating multiply (indexed) | ||
46 | SQDMULLB_zzx_s 01000100 10 1 ..... 1110.0 ..... ..... @rrx_3a esz=2 | ||
47 | SQDMULLB_zzx_d 01000100 11 1 ..... 1110.0 ..... ..... @rrx_2a esz=3 | ||
48 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
49 | index XXXXXXX..XXXXXXX 100644 | ||
50 | --- a/target/arm/sve_helper.c | ||
51 | +++ b/target/arm/sve_helper.c | ||
52 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
53 | DO_ZZX(sve2_sqdmull_idx_s, int32_t, int16_t, H1_4, H1_2, do_sqdmull_s) | ||
54 | DO_ZZX(sve2_sqdmull_idx_d, int64_t, int32_t, , H1_4, do_sqdmull_d) | ||
55 | |||
56 | +DO_ZZX(sve2_smull_idx_s, int32_t, int16_t, H1_4, H1_2, DO_MUL) | ||
57 | +DO_ZZX(sve2_smull_idx_d, int64_t, int32_t, , H1_4, DO_MUL) | ||
58 | + | ||
59 | +DO_ZZX(sve2_umull_idx_s, uint32_t, uint16_t, H1_4, H1_2, DO_MUL) | ||
60 | +DO_ZZX(sve2_umull_idx_d, uint64_t, uint32_t, , H1_4, DO_MUL) | ||
61 | + | ||
62 | #undef DO_ZZX | ||
63 | |||
64 | #define DO_BITPERM(NAME, TYPE, OP) \ | ||
65 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
66 | index XXXXXXX..XXXXXXX 100644 | ||
67 | --- a/target/arm/translate-sve.c | ||
68 | +++ b/target/arm/translate-sve.c | ||
69 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRX_TB(trans_SQDMULLB_zzx_d, gen_helper_sve2_sqdmull_idx_d, false) | ||
70 | DO_SVE2_RRX_TB(trans_SQDMULLT_zzx_s, gen_helper_sve2_sqdmull_idx_s, true) | ||
71 | DO_SVE2_RRX_TB(trans_SQDMULLT_zzx_d, gen_helper_sve2_sqdmull_idx_d, true) | ||
72 | |||
73 | +DO_SVE2_RRX_TB(trans_SMULLB_zzx_s, gen_helper_sve2_smull_idx_s, false) | ||
74 | +DO_SVE2_RRX_TB(trans_SMULLB_zzx_d, gen_helper_sve2_smull_idx_d, false) | ||
75 | +DO_SVE2_RRX_TB(trans_SMULLT_zzx_s, gen_helper_sve2_smull_idx_s, true) | ||
76 | +DO_SVE2_RRX_TB(trans_SMULLT_zzx_d, gen_helper_sve2_smull_idx_d, true) | ||
77 | + | ||
78 | +DO_SVE2_RRX_TB(trans_UMULLB_zzx_s, gen_helper_sve2_umull_idx_s, false) | ||
79 | +DO_SVE2_RRX_TB(trans_UMULLB_zzx_d, gen_helper_sve2_umull_idx_d, false) | ||
80 | +DO_SVE2_RRX_TB(trans_UMULLT_zzx_s, gen_helper_sve2_umull_idx_s, true) | ||
81 | +DO_SVE2_RRX_TB(trans_UMULLT_zzx_d, gen_helper_sve2_umull_idx_d, true) | ||
82 | + | ||
83 | #undef DO_SVE2_RRX_TB | ||
84 | |||
85 | static bool do_sve2_zzzz_data(DisasContext *s, int rd, int rn, int rm, int ra, | ||
86 | -- | ||
87 | 2.20.1 | ||
88 | |||
89 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Message-id: 20210525010358.152808-63-richard.henderson@linaro.org | ||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 9 +++++++++ | ||
9 | target/arm/sve.decode | 12 ++++++++++++ | ||
10 | target/arm/sve_helper.c | 28 ++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 15 +++++++++++++++ | ||
12 | 4 files changed, 64 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve2_smull_idx_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
19 | DEF_HELPER_FLAGS_4(sve2_smull_idx_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_4(sve2_umull_idx_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
21 | DEF_HELPER_FLAGS_4(sve2_umull_idx_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_cmla_idx_h, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_cmla_idx_s, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_idx_h, TCG_CALL_NO_RWG, | ||
28 | + void, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_idx_s, TCG_CALL_NO_RWG, | ||
30 | + void, ptr, ptr, ptr, ptr, i32) | ||
31 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
32 | index XXXXXXX..XXXXXXX 100644 | ||
33 | --- a/target/arm/sve.decode | ||
34 | +++ b/target/arm/sve.decode | ||
35 | @@ -XXX,XX +XXX,XX @@ SQDMLSLB_zzxw_d 01000100 11 1 ..... 0011.0 ..... ..... @rrxr_2a esz=3 | ||
36 | SQDMLSLT_zzxw_s 01000100 10 1 ..... 0011.1 ..... ..... @rrxr_3a esz=2 | ||
37 | SQDMLSLT_zzxw_d 01000100 11 1 ..... 0011.1 ..... ..... @rrxr_2a esz=3 | ||
38 | |||
39 | +# SVE2 complex integer multiply-add (indexed) | ||
40 | +CMLA_zzxz_h 01000100 10 1 index:2 rm:3 0110 rot:2 rn:5 rd:5 \ | ||
41 | + ra=%reg_movprfx | ||
42 | +CMLA_zzxz_s 01000100 11 1 index:1 rm:4 0110 rot:2 rn:5 rd:5 \ | ||
43 | + ra=%reg_movprfx | ||
44 | + | ||
45 | +# SVE2 complex saturating integer multiply-add (indexed) | ||
46 | +SQRDCMLAH_zzxz_h 01000100 10 1 index:2 rm:3 0111 rot:2 rn:5 rd:5 \ | ||
47 | + ra=%reg_movprfx | ||
48 | +SQRDCMLAH_zzxz_s 01000100 11 1 index:1 rm:4 0111 rot:2 rn:5 rd:5 \ | ||
49 | + ra=%reg_movprfx | ||
50 | + | ||
51 | # SVE2 multiply-add long (indexed) | ||
52 | SMLALB_zzxw_s 01000100 10 1 ..... 1000.0 ..... ..... @rrxr_3a esz=2 | ||
53 | SMLALB_zzxw_d 01000100 11 1 ..... 1000.0 ..... ..... @rrxr_2a esz=3 | ||
54 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
55 | index XXXXXXX..XXXXXXX 100644 | ||
56 | --- a/target/arm/sve_helper.c | ||
57 | +++ b/target/arm/sve_helper.c | ||
58 | @@ -XXX,XX +XXX,XX @@ DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_h, int16_t, H2, DO_SQRDMLAH_H) | ||
59 | DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_s, int32_t, H4, DO_SQRDMLAH_S) | ||
60 | DO_CMLA_FUNC(sve2_sqrdcmlah_zzzz_d, int64_t, , DO_SQRDMLAH_D) | ||
61 | |||
62 | +#define DO_CMLA_IDX_FUNC(NAME, TYPE, H, OP) \ | ||
63 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
64 | +{ \ | ||
65 | + intptr_t i, j, oprsz = simd_oprsz(desc); \ | ||
66 | + int rot = extract32(desc, SIMD_DATA_SHIFT, 2); \ | ||
67 | + int idx = extract32(desc, SIMD_DATA_SHIFT + 2, 2) * 2; \ | ||
68 | + int sel_a = rot & 1, sel_b = sel_a ^ 1; \ | ||
69 | + bool sub_r = rot == 1 || rot == 2; \ | ||
70 | + bool sub_i = rot >= 2; \ | ||
71 | + TYPE *d = vd, *n = vn, *m = vm, *a = va; \ | ||
72 | + for (i = 0; i < oprsz / sizeof(TYPE); i += 16 / sizeof(TYPE)) { \ | ||
73 | + TYPE elt2_a = m[H(i + idx + sel_a)]; \ | ||
74 | + TYPE elt2_b = m[H(i + idx + sel_b)]; \ | ||
75 | + for (j = 0; j < 16 / sizeof(TYPE); j += 2) { \ | ||
76 | + TYPE elt1_a = n[H(i + j + sel_a)]; \ | ||
77 | + d[H2(i + j)] = OP(elt1_a, elt2_a, a[H(i + j)], sub_r); \ | ||
78 | + d[H2(i + j + 1)] = OP(elt1_a, elt2_b, a[H(i + j + 1)], sub_i); \ | ||
79 | + } \ | ||
80 | + } \ | ||
81 | +} | ||
82 | + | ||
83 | +DO_CMLA_IDX_FUNC(sve2_cmla_idx_h, int16_t, H2, DO_CMLA) | ||
84 | +DO_CMLA_IDX_FUNC(sve2_cmla_idx_s, int32_t, H4, DO_CMLA) | ||
85 | + | ||
86 | +DO_CMLA_IDX_FUNC(sve2_sqrdcmlah_idx_h, int16_t, H2, DO_SQRDMLAH_H) | ||
87 | +DO_CMLA_IDX_FUNC(sve2_sqrdcmlah_idx_s, int32_t, H4, DO_SQRDMLAH_S) | ||
88 | + | ||
89 | #undef DO_CMLA | ||
90 | #undef DO_CMLA_FUNC | ||
91 | +#undef DO_CMLA_IDX_FUNC | ||
92 | #undef DO_SQRDMLAH_B | ||
93 | #undef DO_SQRDMLAH_H | ||
94 | #undef DO_SQRDMLAH_S | ||
95 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
96 | index XXXXXXX..XXXXXXX 100644 | ||
97 | --- a/target/arm/translate-sve.c | ||
98 | +++ b/target/arm/translate-sve.c | ||
99 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRXR_TB(trans_UMLSLT_zzxw_d, gen_helper_sve2_umlsl_idx_d, true) | ||
100 | |||
101 | #undef DO_SVE2_RRXR_TB | ||
102 | |||
103 | +#define DO_SVE2_RRXR_ROT(NAME, FUNC) \ | ||
104 | + static bool trans_##NAME(DisasContext *s, arg_##NAME *a) \ | ||
105 | + { \ | ||
106 | + return do_sve2_zzzz_data(s, a->rd, a->rn, a->rm, a->ra, \ | ||
107 | + (a->index << 2) | a->rot, FUNC); \ | ||
108 | + } | ||
109 | + | ||
110 | +DO_SVE2_RRXR_ROT(CMLA_zzxz_h, gen_helper_sve2_cmla_idx_h) | ||
111 | +DO_SVE2_RRXR_ROT(CMLA_zzxz_s, gen_helper_sve2_cmla_idx_s) | ||
112 | + | ||
113 | +DO_SVE2_RRXR_ROT(SQRDCMLAH_zzxz_h, gen_helper_sve2_sqrdcmlah_idx_h) | ||
114 | +DO_SVE2_RRXR_ROT(SQRDCMLAH_zzxz_s, gen_helper_sve2_sqrdcmlah_idx_s) | ||
115 | + | ||
116 | +#undef DO_SVE2_RRXR_ROT | ||
117 | + | ||
118 | /* | ||
119 | *** SVE Floating Point Multiply-Add Indexed Group | ||
120 | */ | ||
121 | -- | ||
122 | 2.20.1 | ||
123 | |||
124 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Message-id: 20210525010358.152808-64-richard.henderson@linaro.org | ||
5 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/helper-sve.h | 10 ++++ | ||
9 | target/arm/sve.decode | 9 ++++ | ||
10 | target/arm/sve_helper.c | 99 ++++++++++++++++++++++++++++++++++++++ | ||
11 | target/arm/translate-sve.c | 17 +++++++ | ||
12 | 4 files changed, 135 insertions(+) | ||
13 | |||
14 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/target/arm/helper-sve.h | ||
17 | +++ b/target/arm/helper-sve.h | ||
18 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_idx_h, TCG_CALL_NO_RWG, | ||
19 | void, ptr, ptr, ptr, ptr, i32) | ||
20 | DEF_HELPER_FLAGS_5(sve2_sqrdcmlah_idx_s, TCG_CALL_NO_RWG, | ||
21 | void, ptr, ptr, ptr, ptr, i32) | ||
22 | + | ||
23 | +DEF_HELPER_FLAGS_5(sve2_cdot_zzzz_s, TCG_CALL_NO_RWG, | ||
24 | + void, ptr, ptr, ptr, ptr, i32) | ||
25 | +DEF_HELPER_FLAGS_5(sve2_cdot_zzzz_d, TCG_CALL_NO_RWG, | ||
26 | + void, ptr, ptr, ptr, ptr, i32) | ||
27 | + | ||
28 | +DEF_HELPER_FLAGS_5(sve2_cdot_idx_s, TCG_CALL_NO_RWG, | ||
29 | + void, ptr, ptr, ptr, ptr, i32) | ||
30 | +DEF_HELPER_FLAGS_5(sve2_cdot_idx_d, TCG_CALL_NO_RWG, | ||
31 | + void, ptr, ptr, ptr, ptr, i32) | ||
32 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
33 | index XXXXXXX..XXXXXXX 100644 | ||
34 | --- a/target/arm/sve.decode | ||
35 | +++ b/target/arm/sve.decode | ||
36 | @@ -XXX,XX +XXX,XX @@ MUL_zzi 00100101 .. 110 000 110 ........ ..... @rdn_i8s | ||
37 | DOT_zzzz 01000100 1 sz:1 0 rm:5 00000 u:1 rn:5 rd:5 \ | ||
38 | ra=%reg_movprfx | ||
39 | |||
40 | +# SVE2 complex dot product (vectors) | ||
41 | +CDOT_zzzz 01000100 esz:2 0 rm:5 0001 rot:2 rn:5 rd:5 ra=%reg_movprfx | ||
42 | + | ||
43 | #### SVE Multiply - Indexed | ||
44 | |||
45 | # SVE integer dot product (indexed) | ||
46 | @@ -XXX,XX +XXX,XX @@ SQDMLSLB_zzxw_d 01000100 11 1 ..... 0011.0 ..... ..... @rrxr_2a esz=3 | ||
47 | SQDMLSLT_zzxw_s 01000100 10 1 ..... 0011.1 ..... ..... @rrxr_3a esz=2 | ||
48 | SQDMLSLT_zzxw_d 01000100 11 1 ..... 0011.1 ..... ..... @rrxr_2a esz=3 | ||
49 | |||
50 | +# SVE2 complex integer dot product (indexed) | ||
51 | +CDOT_zzxw_s 01000100 10 1 index:2 rm:3 0100 rot:2 rn:5 rd:5 \ | ||
52 | + ra=%reg_movprfx | ||
53 | +CDOT_zzxw_d 01000100 11 1 index:1 rm:4 0100 rot:2 rn:5 rd:5 \ | ||
54 | + ra=%reg_movprfx | ||
55 | + | ||
56 | # SVE2 complex integer multiply-add (indexed) | ||
57 | CMLA_zzxz_h 01000100 10 1 index:2 rm:3 0110 rot:2 rn:5 rd:5 \ | ||
58 | ra=%reg_movprfx | ||
59 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
60 | index XXXXXXX..XXXXXXX 100644 | ||
61 | --- a/target/arm/sve_helper.c | ||
62 | +++ b/target/arm/sve_helper.c | ||
63 | @@ -XXX,XX +XXX,XX @@ DO_CMLA_IDX_FUNC(sve2_sqrdcmlah_idx_s, int32_t, H4, DO_SQRDMLAH_S) | ||
64 | #undef DO_SQRDMLAH_S | ||
65 | #undef DO_SQRDMLAH_D | ||
66 | |||
67 | +/* Note N and M are 4 elements bundled into one unit. */ | ||
68 | +static int32_t do_cdot_s(uint32_t n, uint32_t m, int32_t a, | ||
69 | + int sel_a, int sel_b, int sub_i) | ||
70 | +{ | ||
71 | + for (int i = 0; i <= 1; i++) { | ||
72 | + int32_t elt1_r = (int8_t)(n >> (16 * i)); | ||
73 | + int32_t elt1_i = (int8_t)(n >> (16 * i + 8)); | ||
74 | + int32_t elt2_a = (int8_t)(m >> (16 * i + 8 * sel_a)); | ||
75 | + int32_t elt2_b = (int8_t)(m >> (16 * i + 8 * sel_b)); | ||
76 | + | ||
77 | + a += elt1_r * elt2_a + elt1_i * elt2_b * sub_i; | ||
78 | + } | ||
79 | + return a; | ||
80 | +} | ||
81 | + | ||
82 | +static int64_t do_cdot_d(uint64_t n, uint64_t m, int64_t a, | ||
83 | + int sel_a, int sel_b, int sub_i) | ||
84 | +{ | ||
85 | + for (int i = 0; i <= 1; i++) { | ||
86 | + int64_t elt1_r = (int16_t)(n >> (32 * i + 0)); | ||
87 | + int64_t elt1_i = (int16_t)(n >> (32 * i + 16)); | ||
88 | + int64_t elt2_a = (int16_t)(m >> (32 * i + 16 * sel_a)); | ||
89 | + int64_t elt2_b = (int16_t)(m >> (32 * i + 16 * sel_b)); | ||
90 | + | ||
91 | + a += elt1_r * elt2_a + elt1_i * elt2_b * sub_i; | ||
92 | + } | ||
93 | + return a; | ||
94 | +} | ||
95 | + | ||
96 | +void HELPER(sve2_cdot_zzzz_s)(void *vd, void *vn, void *vm, | ||
97 | + void *va, uint32_t desc) | ||
98 | +{ | ||
99 | + int opr_sz = simd_oprsz(desc); | ||
100 | + int rot = simd_data(desc); | ||
101 | + int sel_a = rot & 1; | ||
102 | + int sel_b = sel_a ^ 1; | ||
103 | + int sub_i = (rot == 0 || rot == 3 ? -1 : 1); | ||
104 | + uint32_t *d = vd, *n = vn, *m = vm, *a = va; | ||
105 | + | ||
106 | + for (int e = 0; e < opr_sz / 4; e++) { | ||
107 | + d[e] = do_cdot_s(n[e], m[e], a[e], sel_a, sel_b, sub_i); | ||
108 | + } | ||
109 | +} | ||
110 | + | ||
111 | +void HELPER(sve2_cdot_zzzz_d)(void *vd, void *vn, void *vm, | ||
112 | + void *va, uint32_t desc) | ||
113 | +{ | ||
114 | + int opr_sz = simd_oprsz(desc); | ||
115 | + int rot = simd_data(desc); | ||
116 | + int sel_a = rot & 1; | ||
117 | + int sel_b = sel_a ^ 1; | ||
118 | + int sub_i = (rot == 0 || rot == 3 ? -1 : 1); | ||
119 | + uint64_t *d = vd, *n = vn, *m = vm, *a = va; | ||
120 | + | ||
121 | + for (int e = 0; e < opr_sz / 8; e++) { | ||
122 | + d[e] = do_cdot_d(n[e], m[e], a[e], sel_a, sel_b, sub_i); | ||
123 | + } | ||
124 | +} | ||
125 | + | ||
126 | +void HELPER(sve2_cdot_idx_s)(void *vd, void *vn, void *vm, | ||
127 | + void *va, uint32_t desc) | ||
128 | +{ | ||
129 | + int opr_sz = simd_oprsz(desc); | ||
130 | + int rot = extract32(desc, SIMD_DATA_SHIFT, 2); | ||
131 | + int idx = H4(extract32(desc, SIMD_DATA_SHIFT + 2, 2)); | ||
132 | + int sel_a = rot & 1; | ||
133 | + int sel_b = sel_a ^ 1; | ||
134 | + int sub_i = (rot == 0 || rot == 3 ? -1 : 1); | ||
135 | + uint32_t *d = vd, *n = vn, *m = vm, *a = va; | ||
136 | + | ||
137 | + for (int seg = 0; seg < opr_sz / 4; seg += 4) { | ||
138 | + uint32_t seg_m = m[seg + idx]; | ||
139 | + for (int e = 0; e < 4; e++) { | ||
140 | + d[seg + e] = do_cdot_s(n[seg + e], seg_m, a[seg + e], | ||
141 | + sel_a, sel_b, sub_i); | ||
142 | + } | ||
143 | + } | ||
144 | +} | ||
145 | + | ||
146 | +void HELPER(sve2_cdot_idx_d)(void *vd, void *vn, void *vm, | ||
147 | + void *va, uint32_t desc) | ||
148 | +{ | ||
149 | + int seg, opr_sz = simd_oprsz(desc); | ||
150 | + int rot = extract32(desc, SIMD_DATA_SHIFT, 2); | ||
151 | + int idx = extract32(desc, SIMD_DATA_SHIFT + 2, 2); | ||
152 | + int sel_a = rot & 1; | ||
153 | + int sel_b = sel_a ^ 1; | ||
154 | + int sub_i = (rot == 0 || rot == 3 ? -1 : 1); | ||
155 | + uint64_t *d = vd, *n = vn, *m = vm, *a = va; | ||
156 | + | ||
157 | + for (seg = 0; seg < opr_sz / 8; seg += 2) { | ||
158 | + uint64_t seg_m = m[seg + idx]; | ||
159 | + for (int e = 0; e < 2; e++) { | ||
160 | + d[seg + e] = do_cdot_d(n[seg + e], seg_m, a[seg + e], | ||
161 | + sel_a, sel_b, sub_i); | ||
162 | + } | ||
163 | + } | ||
164 | +} | ||
165 | + | ||
166 | #define DO_ZZXZ(NAME, TYPE, H, OP) \ | ||
167 | void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
168 | { \ | ||
169 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
170 | index XXXXXXX..XXXXXXX 100644 | ||
171 | --- a/target/arm/translate-sve.c | ||
172 | +++ b/target/arm/translate-sve.c | ||
173 | @@ -XXX,XX +XXX,XX @@ DO_SVE2_RRXR_ROT(CMLA_zzxz_s, gen_helper_sve2_cmla_idx_s) | ||
174 | DO_SVE2_RRXR_ROT(SQRDCMLAH_zzxz_h, gen_helper_sve2_sqrdcmlah_idx_h) | ||
175 | DO_SVE2_RRXR_ROT(SQRDCMLAH_zzxz_s, gen_helper_sve2_sqrdcmlah_idx_s) | ||
176 | |||
177 | +DO_SVE2_RRXR_ROT(CDOT_zzxw_s, gen_helper_sve2_cdot_idx_s) | ||
178 | +DO_SVE2_RRXR_ROT(CDOT_zzxw_d, gen_helper_sve2_cdot_idx_d) | ||
179 | + | ||
180 | #undef DO_SVE2_RRXR_ROT | ||
181 | |||
182 | /* | ||
183 | @@ -XXX,XX +XXX,XX @@ static bool trans_CMLA_zzzz(DisasContext *s, arg_CMLA_zzzz *a) | ||
184 | return true; | ||
185 | } | ||
186 | |||
187 | +static bool trans_CDOT_zzzz(DisasContext *s, arg_CMLA_zzzz *a) | ||
188 | +{ | ||
189 | + if (!dc_isar_feature(aa64_sve2, s) || a->esz < MO_32) { | ||
190 | + return false; | ||
191 | + } | ||
192 | + if (sve_access_check(s)) { | ||
193 | + gen_helper_gvec_4 *fn = (a->esz == MO_32 | ||
194 | + ? gen_helper_sve2_cdot_zzzz_s | ||
195 | + : gen_helper_sve2_cdot_zzzz_d); | ||
196 | + gen_gvec_ool_zzzz(s, fn, a->rd, a->rn, a->rm, a->ra, a->rot); | ||
197 | + } | ||
198 | + return true; | ||
199 | +} | ||
200 | + | ||
201 | static bool trans_SQRDCMLAH_zzzz(DisasContext *s, arg_SQRDCMLAH_zzzz *a) | ||
202 | { | ||
203 | static gen_helper_gvec_4 * const fns[] = { | ||
204 | -- | ||
205 | 2.20.1 | ||
206 | |||
207 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | We're about to add more variations on this theme. | ||
4 | |||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-65-richard.henderson@linaro.org | ||
7 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
8 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
9 | --- | ||
10 | target/arm/vec_helper.c | 82 ++++++++++------------------------------- | ||
11 | 1 file changed, 20 insertions(+), 62 deletions(-) | ||
12 | |||
13 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
14 | index XXXXXXX..XXXXXXX 100644 | ||
15 | --- a/target/arm/vec_helper.c | ||
16 | +++ b/target/arm/vec_helper.c | ||
17 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve2_sqrdmulh_idx_d)(void *vd, void *vn, void *vm, uint32_t desc) | ||
18 | /* Integer 8 and 16-bit dot-product. | ||
19 | * | ||
20 | * Note that for the loops herein, host endianness does not matter | ||
21 | - * with respect to the ordering of data within the 64-bit lanes. | ||
22 | + * with respect to the ordering of data within the quad-width lanes. | ||
23 | * All elements are treated equally, no matter where they are. | ||
24 | */ | ||
25 | |||
26 | -void HELPER(gvec_sdot_b)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
27 | -{ | ||
28 | - intptr_t i, opr_sz = simd_oprsz(desc); | ||
29 | - int32_t *d = vd, *a = va; | ||
30 | - int8_t *n = vn, *m = vm; | ||
31 | - | ||
32 | - for (i = 0; i < opr_sz / 4; ++i) { | ||
33 | - d[i] = (a[i] + | ||
34 | - n[i * 4 + 0] * m[i * 4 + 0] + | ||
35 | - n[i * 4 + 1] * m[i * 4 + 1] + | ||
36 | - n[i * 4 + 2] * m[i * 4 + 2] + | ||
37 | - n[i * 4 + 3] * m[i * 4 + 3]); | ||
38 | - } | ||
39 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
40 | +#define DO_DOT(NAME, TYPED, TYPEN, TYPEM) \ | ||
41 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
42 | +{ \ | ||
43 | + intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
44 | + TYPED *d = vd, *a = va; \ | ||
45 | + TYPEN *n = vn; \ | ||
46 | + TYPEM *m = vm; \ | ||
47 | + for (i = 0; i < opr_sz / sizeof(TYPED); ++i) { \ | ||
48 | + d[i] = (a[i] + \ | ||
49 | + (TYPED)n[i * 4 + 0] * m[i * 4 + 0] + \ | ||
50 | + (TYPED)n[i * 4 + 1] * m[i * 4 + 1] + \ | ||
51 | + (TYPED)n[i * 4 + 2] * m[i * 4 + 2] + \ | ||
52 | + (TYPED)n[i * 4 + 3] * m[i * 4 + 3]); \ | ||
53 | + } \ | ||
54 | + clear_tail(d, opr_sz, simd_maxsz(desc)); \ | ||
55 | } | ||
56 | |||
57 | -void HELPER(gvec_udot_b)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
58 | -{ | ||
59 | - intptr_t i, opr_sz = simd_oprsz(desc); | ||
60 | - uint32_t *d = vd, *a = va; | ||
61 | - uint8_t *n = vn, *m = vm; | ||
62 | - | ||
63 | - for (i = 0; i < opr_sz / 4; ++i) { | ||
64 | - d[i] = (a[i] + | ||
65 | - n[i * 4 + 0] * m[i * 4 + 0] + | ||
66 | - n[i * 4 + 1] * m[i * 4 + 1] + | ||
67 | - n[i * 4 + 2] * m[i * 4 + 2] + | ||
68 | - n[i * 4 + 3] * m[i * 4 + 3]); | ||
69 | - } | ||
70 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
71 | -} | ||
72 | - | ||
73 | -void HELPER(gvec_sdot_h)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
74 | -{ | ||
75 | - intptr_t i, opr_sz = simd_oprsz(desc); | ||
76 | - int64_t *d = vd, *a = va; | ||
77 | - int16_t *n = vn, *m = vm; | ||
78 | - | ||
79 | - for (i = 0; i < opr_sz / 8; ++i) { | ||
80 | - d[i] = (a[i] + | ||
81 | - (int64_t)n[i * 4 + 0] * m[i * 4 + 0] + | ||
82 | - (int64_t)n[i * 4 + 1] * m[i * 4 + 1] + | ||
83 | - (int64_t)n[i * 4 + 2] * m[i * 4 + 2] + | ||
84 | - (int64_t)n[i * 4 + 3] * m[i * 4 + 3]); | ||
85 | - } | ||
86 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
87 | -} | ||
88 | - | ||
89 | -void HELPER(gvec_udot_h)(void *vd, void *vn, void *vm, void *va, uint32_t desc) | ||
90 | -{ | ||
91 | - intptr_t i, opr_sz = simd_oprsz(desc); | ||
92 | - uint64_t *d = vd, *a = va; | ||
93 | - uint16_t *n = vn, *m = vm; | ||
94 | - | ||
95 | - for (i = 0; i < opr_sz / 8; ++i) { | ||
96 | - d[i] = (a[i] + | ||
97 | - (uint64_t)n[i * 4 + 0] * m[i * 4 + 0] + | ||
98 | - (uint64_t)n[i * 4 + 1] * m[i * 4 + 1] + | ||
99 | - (uint64_t)n[i * 4 + 2] * m[i * 4 + 2] + | ||
100 | - (uint64_t)n[i * 4 + 3] * m[i * 4 + 3]); | ||
101 | - } | ||
102 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
103 | -} | ||
104 | +DO_DOT(gvec_sdot_b, int32_t, int8_t, int8_t) | ||
105 | +DO_DOT(gvec_udot_b, uint32_t, uint8_t, uint8_t) | ||
106 | +DO_DOT(gvec_sdot_h, int64_t, int16_t, int16_t) | ||
107 | +DO_DOT(gvec_udot_h, uint64_t, uint16_t, uint16_t) | ||
108 | |||
109 | void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, | ||
110 | void *va, uint32_t desc) | ||
111 | -- | ||
112 | 2.20.1 | ||
113 | |||
114 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | We're about to add more variations on this theme. | ||
4 | Accept the inner loop for the _h variants, rather | ||
5 | than keep it unrolled. | ||
6 | |||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
8 | Message-id: 20210525010358.152808-66-richard.henderson@linaro.org | ||
9 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/vec_helper.c | 160 ++++++++-------------------------------- | ||
13 | 1 file changed, 29 insertions(+), 131 deletions(-) | ||
14 | |||
15 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/target/arm/vec_helper.c | ||
18 | +++ b/target/arm/vec_helper.c | ||
19 | @@ -XXX,XX +XXX,XX @@ DO_DOT(gvec_udot_b, uint32_t, uint8_t, uint8_t) | ||
20 | DO_DOT(gvec_sdot_h, int64_t, int16_t, int16_t) | ||
21 | DO_DOT(gvec_udot_h, uint64_t, uint16_t, uint16_t) | ||
22 | |||
23 | -void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, | ||
24 | - void *va, uint32_t desc) | ||
25 | -{ | ||
26 | - intptr_t i, segend, opr_sz = simd_oprsz(desc), opr_sz_4 = opr_sz / 4; | ||
27 | - intptr_t index = simd_data(desc); | ||
28 | - int32_t *d = vd, *a = va; | ||
29 | - int8_t *n = vn; | ||
30 | - int8_t *m_indexed = (int8_t *)vm + H4(index) * 4; | ||
31 | - | ||
32 | - /* Notice the special case of opr_sz == 8, from aa64/aa32 advsimd. | ||
33 | - * Otherwise opr_sz is a multiple of 16. | ||
34 | - */ | ||
35 | - segend = MIN(4, opr_sz_4); | ||
36 | - i = 0; | ||
37 | - do { | ||
38 | - int8_t m0 = m_indexed[i * 4 + 0]; | ||
39 | - int8_t m1 = m_indexed[i * 4 + 1]; | ||
40 | - int8_t m2 = m_indexed[i * 4 + 2]; | ||
41 | - int8_t m3 = m_indexed[i * 4 + 3]; | ||
42 | - | ||
43 | - do { | ||
44 | - d[i] = (a[i] + | ||
45 | - n[i * 4 + 0] * m0 + | ||
46 | - n[i * 4 + 1] * m1 + | ||
47 | - n[i * 4 + 2] * m2 + | ||
48 | - n[i * 4 + 3] * m3); | ||
49 | - } while (++i < segend); | ||
50 | - segend = i + 4; | ||
51 | - } while (i < opr_sz_4); | ||
52 | - | ||
53 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
54 | +#define DO_DOT_IDX(NAME, TYPED, TYPEN, TYPEM, HD) \ | ||
55 | +void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
56 | +{ \ | ||
57 | + intptr_t i = 0, opr_sz = simd_oprsz(desc); \ | ||
58 | + intptr_t opr_sz_n = opr_sz / sizeof(TYPED); \ | ||
59 | + intptr_t segend = MIN(16 / sizeof(TYPED), opr_sz_n); \ | ||
60 | + intptr_t index = simd_data(desc); \ | ||
61 | + TYPED *d = vd, *a = va; \ | ||
62 | + TYPEN *n = vn; \ | ||
63 | + TYPEM *m_indexed = (TYPEM *)vm + HD(index) * 4; \ | ||
64 | + do { \ | ||
65 | + TYPED m0 = m_indexed[i * 4 + 0]; \ | ||
66 | + TYPED m1 = m_indexed[i * 4 + 1]; \ | ||
67 | + TYPED m2 = m_indexed[i * 4 + 2]; \ | ||
68 | + TYPED m3 = m_indexed[i * 4 + 3]; \ | ||
69 | + do { \ | ||
70 | + d[i] = (a[i] + \ | ||
71 | + n[i * 4 + 0] * m0 + \ | ||
72 | + n[i * 4 + 1] * m1 + \ | ||
73 | + n[i * 4 + 2] * m2 + \ | ||
74 | + n[i * 4 + 3] * m3); \ | ||
75 | + } while (++i < segend); \ | ||
76 | + segend = i + 4; \ | ||
77 | + } while (i < opr_sz_n); \ | ||
78 | + clear_tail(d, opr_sz, simd_maxsz(desc)); \ | ||
79 | } | ||
80 | |||
81 | -void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, | ||
82 | - void *va, uint32_t desc) | ||
83 | -{ | ||
84 | - intptr_t i, segend, opr_sz = simd_oprsz(desc), opr_sz_4 = opr_sz / 4; | ||
85 | - intptr_t index = simd_data(desc); | ||
86 | - uint32_t *d = vd, *a = va; | ||
87 | - uint8_t *n = vn; | ||
88 | - uint8_t *m_indexed = (uint8_t *)vm + H4(index) * 4; | ||
89 | - | ||
90 | - /* Notice the special case of opr_sz == 8, from aa64/aa32 advsimd. | ||
91 | - * Otherwise opr_sz is a multiple of 16. | ||
92 | - */ | ||
93 | - segend = MIN(4, opr_sz_4); | ||
94 | - i = 0; | ||
95 | - do { | ||
96 | - uint8_t m0 = m_indexed[i * 4 + 0]; | ||
97 | - uint8_t m1 = m_indexed[i * 4 + 1]; | ||
98 | - uint8_t m2 = m_indexed[i * 4 + 2]; | ||
99 | - uint8_t m3 = m_indexed[i * 4 + 3]; | ||
100 | - | ||
101 | - do { | ||
102 | - d[i] = (a[i] + | ||
103 | - n[i * 4 + 0] * m0 + | ||
104 | - n[i * 4 + 1] * m1 + | ||
105 | - n[i * 4 + 2] * m2 + | ||
106 | - n[i * 4 + 3] * m3); | ||
107 | - } while (++i < segend); | ||
108 | - segend = i + 4; | ||
109 | - } while (i < opr_sz_4); | ||
110 | - | ||
111 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
112 | -} | ||
113 | - | ||
114 | -void HELPER(gvec_sdot_idx_h)(void *vd, void *vn, void *vm, | ||
115 | - void *va, uint32_t desc) | ||
116 | -{ | ||
117 | - intptr_t i, opr_sz = simd_oprsz(desc), opr_sz_8 = opr_sz / 8; | ||
118 | - intptr_t index = simd_data(desc); | ||
119 | - int64_t *d = vd, *a = va; | ||
120 | - int16_t *n = vn; | ||
121 | - int16_t *m_indexed = (int16_t *)vm + index * 4; | ||
122 | - | ||
123 | - /* This is supported by SVE only, so opr_sz is always a multiple of 16. | ||
124 | - * Process the entire segment all at once, writing back the results | ||
125 | - * only after we've consumed all of the inputs. | ||
126 | - */ | ||
127 | - for (i = 0; i < opr_sz_8; i += 2) { | ||
128 | - int64_t d0, d1; | ||
129 | - | ||
130 | - d0 = a[i + 0]; | ||
131 | - d0 += n[i * 4 + 0] * (int64_t)m_indexed[i * 4 + 0]; | ||
132 | - d0 += n[i * 4 + 1] * (int64_t)m_indexed[i * 4 + 1]; | ||
133 | - d0 += n[i * 4 + 2] * (int64_t)m_indexed[i * 4 + 2]; | ||
134 | - d0 += n[i * 4 + 3] * (int64_t)m_indexed[i * 4 + 3]; | ||
135 | - | ||
136 | - d1 = a[i + 1]; | ||
137 | - d1 += n[i * 4 + 4] * (int64_t)m_indexed[i * 4 + 0]; | ||
138 | - d1 += n[i * 4 + 5] * (int64_t)m_indexed[i * 4 + 1]; | ||
139 | - d1 += n[i * 4 + 6] * (int64_t)m_indexed[i * 4 + 2]; | ||
140 | - d1 += n[i * 4 + 7] * (int64_t)m_indexed[i * 4 + 3]; | ||
141 | - | ||
142 | - d[i + 0] = d0; | ||
143 | - d[i + 1] = d1; | ||
144 | - } | ||
145 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
146 | -} | ||
147 | - | ||
148 | -void HELPER(gvec_udot_idx_h)(void *vd, void *vn, void *vm, | ||
149 | - void *va, uint32_t desc) | ||
150 | -{ | ||
151 | - intptr_t i, opr_sz = simd_oprsz(desc), opr_sz_8 = opr_sz / 8; | ||
152 | - intptr_t index = simd_data(desc); | ||
153 | - uint64_t *d = vd, *a = va; | ||
154 | - uint16_t *n = vn; | ||
155 | - uint16_t *m_indexed = (uint16_t *)vm + index * 4; | ||
156 | - | ||
157 | - /* This is supported by SVE only, so opr_sz is always a multiple of 16. | ||
158 | - * Process the entire segment all at once, writing back the results | ||
159 | - * only after we've consumed all of the inputs. | ||
160 | - */ | ||
161 | - for (i = 0; i < opr_sz_8; i += 2) { | ||
162 | - uint64_t d0, d1; | ||
163 | - | ||
164 | - d0 = a[i + 0]; | ||
165 | - d0 += n[i * 4 + 0] * (uint64_t)m_indexed[i * 4 + 0]; | ||
166 | - d0 += n[i * 4 + 1] * (uint64_t)m_indexed[i * 4 + 1]; | ||
167 | - d0 += n[i * 4 + 2] * (uint64_t)m_indexed[i * 4 + 2]; | ||
168 | - d0 += n[i * 4 + 3] * (uint64_t)m_indexed[i * 4 + 3]; | ||
169 | - | ||
170 | - d1 = a[i + 1]; | ||
171 | - d1 += n[i * 4 + 4] * (uint64_t)m_indexed[i * 4 + 0]; | ||
172 | - d1 += n[i * 4 + 5] * (uint64_t)m_indexed[i * 4 + 1]; | ||
173 | - d1 += n[i * 4 + 6] * (uint64_t)m_indexed[i * 4 + 2]; | ||
174 | - d1 += n[i * 4 + 7] * (uint64_t)m_indexed[i * 4 + 3]; | ||
175 | - | ||
176 | - d[i + 0] = d0; | ||
177 | - d[i + 1] = d1; | ||
178 | - } | ||
179 | - clear_tail(d, opr_sz, simd_maxsz(desc)); | ||
180 | -} | ||
181 | +DO_DOT_IDX(gvec_sdot_idx_b, int32_t, int8_t, int8_t, H4) | ||
182 | +DO_DOT_IDX(gvec_udot_idx_b, uint32_t, uint8_t, uint8_t, H4) | ||
183 | +DO_DOT_IDX(gvec_sdot_idx_h, int64_t, int16_t, int16_t, ) | ||
184 | +DO_DOT_IDX(gvec_udot_idx_h, uint64_t, uint16_t, uint16_t, ) | ||
185 | |||
186 | void HELPER(gvec_fcaddh)(void *vd, void *vn, void *vm, | ||
187 | void *vfpst, uint32_t desc) | ||
188 | -- | ||
189 | 2.20.1 | ||
190 | |||
191 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-67-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/cpu.h | 5 +++++ | ||
9 | target/arm/helper.h | 4 ++++ | ||
10 | target/arm/sve.decode | 4 ++++ | ||
11 | target/arm/translate-sve.c | 16 ++++++++++++++++ | ||
12 | target/arm/vec_helper.c | 2 ++ | ||
13 | 5 files changed, 31 insertions(+) | ||
14 | |||
15 | diff --git a/target/arm/cpu.h b/target/arm/cpu.h | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/target/arm/cpu.h | ||
18 | +++ b/target/arm/cpu.h | ||
19 | @@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_sve2_bitperm(const ARMISARegisters *id) | ||
20 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, BITPERM) != 0; | ||
21 | } | ||
22 | |||
23 | +static inline bool isar_feature_aa64_sve_i8mm(const ARMISARegisters *id) | ||
24 | +{ | ||
25 | + return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, I8MM) != 0; | ||
26 | +} | ||
27 | + | ||
28 | static inline bool isar_feature_aa64_sve_f32mm(const ARMISARegisters *id) | ||
29 | { | ||
30 | return FIELD_EX64(id->id_aa64zfr0, ID_AA64ZFR0, F32MM) != 0; | ||
31 | diff --git a/target/arm/helper.h b/target/arm/helper.h | ||
32 | index XXXXXXX..XXXXXXX 100644 | ||
33 | --- a/target/arm/helper.h | ||
34 | +++ b/target/arm/helper.h | ||
35 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(gvec_sdot_idx_h, TCG_CALL_NO_RWG, | ||
36 | void, ptr, ptr, ptr, ptr, i32) | ||
37 | DEF_HELPER_FLAGS_5(gvec_udot_idx_h, TCG_CALL_NO_RWG, | ||
38 | void, ptr, ptr, ptr, ptr, i32) | ||
39 | +DEF_HELPER_FLAGS_5(gvec_sudot_idx_b, TCG_CALL_NO_RWG, | ||
40 | + void, ptr, ptr, ptr, ptr, i32) | ||
41 | +DEF_HELPER_FLAGS_5(gvec_usdot_idx_b, TCG_CALL_NO_RWG, | ||
42 | + void, ptr, ptr, ptr, ptr, i32) | ||
43 | |||
44 | DEF_HELPER_FLAGS_5(gvec_fcaddh, TCG_CALL_NO_RWG, | ||
45 | void, ptr, ptr, ptr, ptr, i32) | ||
46 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
47 | index XXXXXXX..XXXXXXX 100644 | ||
48 | --- a/target/arm/sve.decode | ||
49 | +++ b/target/arm/sve.decode | ||
50 | @@ -XXX,XX +XXX,XX @@ SQRDMLSH_zzxz_h 01000100 0. 1 ..... 000101 ..... ..... @rrxr_3 esz=1 | ||
51 | SQRDMLSH_zzxz_s 01000100 10 1 ..... 000101 ..... ..... @rrxr_2 esz=2 | ||
52 | SQRDMLSH_zzxz_d 01000100 11 1 ..... 000101 ..... ..... @rrxr_1 esz=3 | ||
53 | |||
54 | +# SVE mixed sign dot product (indexed) | ||
55 | +USDOT_zzxw_s 01000100 10 1 ..... 000110 ..... ..... @rrxr_2 esz=2 | ||
56 | +SUDOT_zzxw_s 01000100 10 1 ..... 000111 ..... ..... @rrxr_2 esz=2 | ||
57 | + | ||
58 | # SVE2 saturating multiply-add (indexed) | ||
59 | SQDMLALB_zzxw_s 01000100 10 1 ..... 0010.0 ..... ..... @rrxr_3a esz=2 | ||
60 | SQDMLALB_zzxw_d 01000100 11 1 ..... 0010.0 ..... ..... @rrxr_2a esz=3 | ||
61 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
62 | index XXXXXXX..XXXXXXX 100644 | ||
63 | --- a/target/arm/translate-sve.c | ||
64 | +++ b/target/arm/translate-sve.c | ||
65 | @@ -XXX,XX +XXX,XX @@ DO_RRXR(trans_SDOT_zzxw_d, gen_helper_gvec_sdot_idx_h) | ||
66 | DO_RRXR(trans_UDOT_zzxw_s, gen_helper_gvec_udot_idx_b) | ||
67 | DO_RRXR(trans_UDOT_zzxw_d, gen_helper_gvec_udot_idx_h) | ||
68 | |||
69 | +static bool trans_SUDOT_zzxw_s(DisasContext *s, arg_rrxr_esz *a) | ||
70 | +{ | ||
71 | + if (!dc_isar_feature(aa64_sve_i8mm, s)) { | ||
72 | + return false; | ||
73 | + } | ||
74 | + return do_zzxz_ool(s, a, gen_helper_gvec_sudot_idx_b); | ||
75 | +} | ||
76 | + | ||
77 | +static bool trans_USDOT_zzxw_s(DisasContext *s, arg_rrxr_esz *a) | ||
78 | +{ | ||
79 | + if (!dc_isar_feature(aa64_sve_i8mm, s)) { | ||
80 | + return false; | ||
81 | + } | ||
82 | + return do_zzxz_ool(s, a, gen_helper_gvec_usdot_idx_b); | ||
83 | +} | ||
84 | + | ||
85 | #undef DO_RRXR | ||
86 | |||
87 | static bool do_sve2_zzz_data(DisasContext *s, int rd, int rn, int rm, int data, | ||
88 | diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c | ||
89 | index XXXXXXX..XXXXXXX 100644 | ||
90 | --- a/target/arm/vec_helper.c | ||
91 | +++ b/target/arm/vec_helper.c | ||
92 | @@ -XXX,XX +XXX,XX @@ void HELPER(NAME)(void *vd, void *vn, void *vm, void *va, uint32_t desc) \ | ||
93 | |||
94 | DO_DOT_IDX(gvec_sdot_idx_b, int32_t, int8_t, int8_t, H4) | ||
95 | DO_DOT_IDX(gvec_udot_idx_b, uint32_t, uint8_t, uint8_t, H4) | ||
96 | +DO_DOT_IDX(gvec_sudot_idx_b, int32_t, int8_t, uint8_t, H4) | ||
97 | +DO_DOT_IDX(gvec_usdot_idx_b, int32_t, uint8_t, int8_t, H4) | ||
98 | DO_DOT_IDX(gvec_sdot_idx_h, int64_t, int16_t, int16_t, ) | ||
99 | DO_DOT_IDX(gvec_udot_idx_h, uint64_t, uint16_t, uint16_t, ) | ||
100 | |||
101 | -- | ||
102 | 2.20.1 | ||
103 | |||
104 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Richard Henderson <richard.henderson@linaro.org> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Message-id: 20210525010358.152808-69-richard.henderson@linaro.org | ||
6 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
7 | --- | ||
8 | target/arm/sve.decode | 6 ++++++ | ||
9 | target/arm/translate-sve.c | 11 +++++++++++ | ||
10 | 2 files changed, 17 insertions(+) | ||
11 | |||
12 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/target/arm/sve.decode | ||
15 | +++ b/target/arm/sve.decode | ||
16 | @@ -XXX,XX +XXX,XX @@ STNT1_zprz 1110010 .. 00 ..... 001 ... ..... ..... \ | ||
17 | # SVE2 32-bit scatter non-temporal store (vector plus scalar) | ||
18 | STNT1_zprz 1110010 .. 10 ..... 001 ... ..... ..... \ | ||
19 | @rprr_scatter_store xs=0 esz=2 scale=0 | ||
20 | + | ||
21 | +### SVE2 Crypto Extensions | ||
22 | + | ||
23 | +# SVE2 crypto unary operations | ||
24 | +# AESMC and AESIMC | ||
25 | +AESMC 01000101 00 10000011100 decrypt:1 00000 rd:5 | ||
26 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
27 | index XXXXXXX..XXXXXXX 100644 | ||
28 | --- a/target/arm/translate-sve.c | ||
29 | +++ b/target/arm/translate-sve.c | ||
30 | @@ -XXX,XX +XXX,XX @@ static bool trans_USDOT_zzzz(DisasContext *s, arg_USDOT_zzzz *a) | ||
31 | } | ||
32 | return true; | ||
33 | } | ||
34 | + | ||
35 | +static bool trans_AESMC(DisasContext *s, arg_AESMC *a) | ||
36 | +{ | ||
37 | + if (!dc_isar_feature(aa64_sve2_aes, s)) { | ||
38 | + return false; | ||
39 | + } | ||
40 | + if (sve_access_check(s)) { | ||
41 | + gen_gvec_ool_zz(s, gen_helper_crypto_aesmc, a->rd, a->rd, a->decrypt); | ||
42 | + } | ||
43 | + return true; | ||
44 | +} | ||
45 | -- | ||
46 | 2.20.1 | ||
47 | |||
48 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Stephen Long <steplong@quicinc.com> | ||
2 | 1 | ||
3 | Reviewed-by: Peter Maydell <peter.maydell@linaro.org> | ||
4 | Signed-off-by: Stephen Long <steplong@quicinc.com> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
6 | Message-id: 20210525010358.152808-72-richard.henderson@linaro.org | ||
7 | Message-Id: <20200428144352.9275-1-steplong@quicinc.com> | ||
8 | [rth: rearrange the macros a little and rebase] | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Signed-off-by: Peter Maydell <peter.maydell@linaro.org> | ||
11 | --- | ||
12 | target/arm/helper-sve.h | 10 +++++ | ||
13 | target/arm/sve.decode | 5 +++ | ||
14 | target/arm/sve_helper.c | 90 ++++++++++++++++++++++++++++++-------- | ||
15 | target/arm/translate-sve.c | 33 ++++++++++++++ | ||
16 | 4 files changed, 119 insertions(+), 19 deletions(-) | ||
17 | |||
18 | diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/target/arm/helper-sve.h | ||
21 | +++ b/target/arm/helper-sve.h | ||
22 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_tbl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
23 | DEF_HELPER_FLAGS_4(sve_tbl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
24 | DEF_HELPER_FLAGS_4(sve_tbl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
25 | |||
26 | +DEF_HELPER_FLAGS_5(sve2_tbl_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
27 | +DEF_HELPER_FLAGS_5(sve2_tbl_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
28 | +DEF_HELPER_FLAGS_5(sve2_tbl_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
29 | +DEF_HELPER_FLAGS_5(sve2_tbl_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32) | ||
30 | + | ||
31 | +DEF_HELPER_FLAGS_4(sve2_tbx_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
32 | +DEF_HELPER_FLAGS_4(sve2_tbx_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
33 | +DEF_HELPER_FLAGS_4(sve2_tbx_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
34 | +DEF_HELPER_FLAGS_4(sve2_tbx_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32) | ||
35 | + | ||
36 | DEF_HELPER_FLAGS_3(sve_sunpk_h, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
37 | DEF_HELPER_FLAGS_3(sve_sunpk_s, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
38 | DEF_HELPER_FLAGS_3(sve_sunpk_d, TCG_CALL_NO_RWG, void, ptr, ptr, i32) | ||
39 | diff --git a/target/arm/sve.decode b/target/arm/sve.decode | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/target/arm/sve.decode | ||
42 | +++ b/target/arm/sve.decode | ||
43 | @@ -XXX,XX +XXX,XX @@ TBL 00000101 .. 1 ..... 001100 ..... ..... @rd_rn_rm | ||
44 | # SVE unpack vector elements | ||
45 | UNPK 00000101 esz:2 1100 u:1 h:1 001110 rn:5 rd:5 | ||
46 | |||
47 | +# SVE2 Table Lookup (three sources) | ||
48 | + | ||
49 | +TBL_sve2 00000101 .. 1 ..... 001010 ..... ..... @rd_rn_rm | ||
50 | +TBX 00000101 .. 1 ..... 001011 ..... ..... @rd_rn_rm | ||
51 | + | ||
52 | ### SVE Permute - Predicates Group | ||
53 | |||
54 | # SVE permute predicate elements | ||
55 | diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c | ||
56 | index XXXXXXX..XXXXXXX 100644 | ||
57 | --- a/target/arm/sve_helper.c | ||
58 | +++ b/target/arm/sve_helper.c | ||
59 | @@ -XXX,XX +XXX,XX @@ void HELPER(sve_rev_d)(void *vd, void *vn, uint32_t desc) | ||
60 | } | ||
61 | } | ||
62 | |||
63 | -#define DO_TBL(NAME, TYPE, H) \ | ||
64 | -void HELPER(NAME)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
65 | -{ \ | ||
66 | - intptr_t i, opr_sz = simd_oprsz(desc); \ | ||
67 | - uintptr_t elem = opr_sz / sizeof(TYPE); \ | ||
68 | - TYPE *d = vd, *n = vn, *m = vm; \ | ||
69 | - ARMVectorReg tmp; \ | ||
70 | - if (unlikely(vd == vn)) { \ | ||
71 | - n = memcpy(&tmp, vn, opr_sz); \ | ||
72 | - } \ | ||
73 | - for (i = 0; i < elem; i++) { \ | ||
74 | - TYPE j = m[H(i)]; \ | ||
75 | - d[H(i)] = j < elem ? n[H(j)] : 0; \ | ||
76 | - } \ | ||
77 | +typedef void tb_impl_fn(void *, void *, void *, void *, uintptr_t, bool); | ||
78 | + | ||
79 | +static inline void do_tbl1(void *vd, void *vn, void *vm, uint32_t desc, | ||
80 | + bool is_tbx, tb_impl_fn *fn) | ||
81 | +{ | ||
82 | + ARMVectorReg scratch; | ||
83 | + uintptr_t oprsz = simd_oprsz(desc); | ||
84 | + | ||
85 | + if (unlikely(vd == vn)) { | ||
86 | + vn = memcpy(&scratch, vn, oprsz); | ||
87 | + } | ||
88 | + | ||
89 | + fn(vd, vn, NULL, vm, oprsz, is_tbx); | ||
90 | } | ||
91 | |||
92 | -DO_TBL(sve_tbl_b, uint8_t, H1) | ||
93 | -DO_TBL(sve_tbl_h, uint16_t, H2) | ||
94 | -DO_TBL(sve_tbl_s, uint32_t, H4) | ||
95 | -DO_TBL(sve_tbl_d, uint64_t, ) | ||
96 | +static inline void do_tbl2(void *vd, void *vn0, void *vn1, void *vm, | ||
97 | + uint32_t desc, bool is_tbx, tb_impl_fn *fn) | ||
98 | +{ | ||
99 | + ARMVectorReg scratch; | ||
100 | + uintptr_t oprsz = simd_oprsz(desc); | ||
101 | |||
102 | -#undef TBL | ||
103 | + if (unlikely(vd == vn0)) { | ||
104 | + vn0 = memcpy(&scratch, vn0, oprsz); | ||
105 | + if (vd == vn1) { | ||
106 | + vn1 = vn0; | ||
107 | + } | ||
108 | + } else if (unlikely(vd == vn1)) { | ||
109 | + vn1 = memcpy(&scratch, vn1, oprsz); | ||
110 | + } | ||
111 | + | ||
112 | + fn(vd, vn0, vn1, vm, oprsz, is_tbx); | ||
113 | +} | ||
114 | + | ||
115 | +#define DO_TB(SUFF, TYPE, H) \ | ||
116 | +static inline void do_tb_##SUFF(void *vd, void *vt0, void *vt1, \ | ||
117 | + void *vm, uintptr_t oprsz, bool is_tbx) \ | ||
118 | +{ \ | ||
119 | + TYPE *d = vd, *tbl0 = vt0, *tbl1 = vt1, *indexes = vm; \ | ||
120 | + uintptr_t i, nelem = oprsz / sizeof(TYPE); \ | ||
121 | + for (i = 0; i < nelem; ++i) { \ | ||
122 | + TYPE index = indexes[H1(i)], val = 0; \ | ||
123 | + if (index < nelem) { \ | ||
124 | + val = tbl0[H(index)]; \ | ||
125 | + } else { \ | ||
126 | + index -= nelem; \ | ||
127 | + if (tbl1 && index < nelem) { \ | ||
128 | + val = tbl1[H(index)]; \ | ||
129 | + } else if (is_tbx) { \ | ||
130 | + continue; \ | ||
131 | + } \ | ||
132 | + } \ | ||
133 | + d[H(i)] = val; \ | ||
134 | + } \ | ||
135 | +} \ | ||
136 | +void HELPER(sve_tbl_##SUFF)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
137 | +{ \ | ||
138 | + do_tbl1(vd, vn, vm, desc, false, do_tb_##SUFF); \ | ||
139 | +} \ | ||
140 | +void HELPER(sve2_tbl_##SUFF)(void *vd, void *vn0, void *vn1, \ | ||
141 | + void *vm, uint32_t desc) \ | ||
142 | +{ \ | ||
143 | + do_tbl2(vd, vn0, vn1, vm, desc, false, do_tb_##SUFF); \ | ||
144 | +} \ | ||
145 | +void HELPER(sve2_tbx_##SUFF)(void *vd, void *vn, void *vm, uint32_t desc) \ | ||
146 | +{ \ | ||
147 | + do_tbl1(vd, vn, vm, desc, true, do_tb_##SUFF); \ | ||
148 | +} | ||
149 | + | ||
150 | +DO_TB(b, uint8_t, H1) | ||
151 | +DO_TB(h, uint16_t, H2) | ||
152 | +DO_TB(s, uint32_t, H4) | ||
153 | +DO_TB(d, uint64_t, ) | ||
154 | + | ||
155 | +#undef DO_TB | ||
156 | |||
157 | #define DO_UNPK(NAME, TYPED, TYPES, HD, HS) \ | ||
158 | void HELPER(NAME)(void *vd, void *vn, uint32_t desc) \ | ||
159 | diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c | ||
160 | index XXXXXXX..XXXXXXX 100644 | ||
161 | --- a/target/arm/translate-sve.c | ||
162 | +++ b/target/arm/translate-sve.c | ||
163 | @@ -XXX,XX +XXX,XX @@ static bool trans_TBL(DisasContext *s, arg_rrr_esz *a) | ||
164 | return true; | ||
165 | } | ||
166 | |||
167 | +static bool trans_TBL_sve2(DisasContext *s, arg_rrr_esz *a) | ||
168 | +{ | ||
169 | + static gen_helper_gvec_4 * const fns[4] = { | ||
170 | + gen_helper_sve2_tbl_b, gen_helper_sve2_tbl_h, | ||
171 | + gen_helper_sve2_tbl_s, gen_helper_sve2_tbl_d | ||
172 | + }; | ||
173 | + | ||
174 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
175 | + return false; | ||
176 | + } | ||
177 | + if (sve_access_check(s)) { | ||
178 | + gen_gvec_ool_zzzz(s, fns[a->esz], a->rd, a->rn, | ||
179 | + (a->rn + 1) % 32, a->rm, 0); | ||
180 | + } | ||
181 | + return true; | ||
182 | +} | ||
183 | + | ||
184 | +static bool trans_TBX(DisasContext *s, arg_rrr_esz *a) | ||
185 | +{ | ||
186 | + static gen_helper_gvec_3 * const fns[4] = { | ||
187 | + gen_helper_sve2_tbx_b, gen_helper_sve2_tbx_h, | ||
188 | + gen_helper_sve2_tbx_s, gen_helper_sve2_tbx_d | ||
189 | + }; | ||
190 | + | ||
191 | + if (!dc_isar_feature(aa64_sve2, s)) { | ||
192 | + return false; | ||
193 | + } | ||
194 | + if (sve_access_check(s)) { | ||
195 | + gen_gvec_ool_zzz(s, fns[a->esz], a->rd, a->rn, a->rm, 0); | ||
196 | + } | ||
197 | + return true; | ||
198 | +} | ||
199 | + | ||
200 | static bool trans_UNPK(DisasContext *s, arg_UNPK *a) | ||
201 | { | ||
202 | static gen_helper_gvec_2 * const fns[4][2] = { | ||
203 | -- | ||
204 | 2.20.1 | ||
205 | |||
206 | diff view generated by jsdifflib |