From nobody Fri May  9 12:09:59 2025
Delivered-To: importer@patchew.org
Received-SPF: pass (zoho.com: domain of gnu.org designates 208.118.235.17 as
 permitted sender) client-ip=208.118.235.17;
 envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org;
 helo=lists.gnu.org;
Authentication-Results: mx.zohomail.com;
	spf=pass (zoho.com: domain of gnu.org designates 208.118.235.17 as permitted
 sender)  smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org;
	dmarc=fail(p=none dis=none)  header.from=linaro.org
Return-Path: <qemu-devel-bounces+importer=patchew.org@nongnu.org>
Received: from lists.gnu.org (lists.gnu.org [208.118.235.17]) by
 mx.zohomail.com
	with SMTPS id 1539969963326480.73857013419297;
 Fri, 19 Oct 2018 10:26:03 -0700 (PDT)
Received: from localhost ([::1]:51849 helo=lists.gnu.org)
	by lists.gnu.org with esmtp (Exim 4.71)
	(envelope-from <qemu-devel-bounces+importer=patchew.org@nongnu.org>)
	id 1gDYXC-0002il-6H
	for importer@patchew.org; Fri, 19 Oct 2018 13:26:02 -0400
Received: from eggs.gnu.org ([2001:4830:134:3::10]:48089)
	by lists.gnu.org with esmtp (Exim 4.71)
	(envelope-from <pm215@archaic.org.uk>) id 1gDY7A-0002b5-U3
	for qemu-devel@nongnu.org; Fri, 19 Oct 2018 12:59:10 -0400
Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71)
	(envelope-from <pm215@archaic.org.uk>) id 1gDY79-0003xV-IY
	for qemu-devel@nongnu.org; Fri, 19 Oct 2018 12:59:08 -0400
Received: from orth.archaic.org.uk ([2001:8b0:1d0::2]:51984)
	by eggs.gnu.org with esmtps (TLS1.0:RSA_AES_256_CBC_SHA1:32)
	(Exim 4.71) (envelope-from <pm215@archaic.org.uk>)
	id 1gDY79-0002bn-4p
	for qemu-devel@nongnu.org; Fri, 19 Oct 2018 12:59:07 -0400
Received: from pm215 by orth.archaic.org.uk with local (Exim 4.89)
	(envelope-from <pm215@archaic.org.uk>) id 1gDY6E-0006nJ-Ln
	for qemu-devel@nongnu.org; Fri, 19 Oct 2018 17:58:10 +0100
From: Peter Maydell <peter.maydell@linaro.org>
To: qemu-devel@nongnu.org
Date: Fri, 19 Oct 2018 17:57:24 +0100
Message-Id: <20181019165735.22511-35-peter.maydell@linaro.org>
X-Mailer: git-send-email 2.19.1
In-Reply-To: <20181019165735.22511-1-peter.maydell@linaro.org>
References: <20181019165735.22511-1-peter.maydell@linaro.org>
MIME-Version: 1.0
Content-Transfer-Encoding: quoted-printable
X-detected-operating-system: by eggs.gnu.org: Genre and OS details not
	recognized.
X-Received-From: 2001:8b0:1d0::2
Subject: [Qemu-devel] [PULL 34/45] target/arm: Use gvec for VSRA
X-BeenThere: qemu-devel@nongnu.org
X-Mailman-Version: 2.1.21
Precedence: list
List-Id: <qemu-devel.nongnu.org>
List-Unsubscribe: <https://lists.nongnu.org/mailman/options/qemu-devel>,
	<mailto:qemu-devel-request@nongnu.org?subject=unsubscribe>
List-Archive: <http://lists.nongnu.org/archive/html/qemu-devel/>
List-Post: <mailto:qemu-devel@nongnu.org>
List-Help: <mailto:qemu-devel-request@nongnu.org?subject=help>
List-Subscribe: <https://lists.nongnu.org/mailman/listinfo/qemu-devel>,
	<mailto:qemu-devel-request@nongnu.org?subject=subscribe>
Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org
Sender: "Qemu-devel" <qemu-devel-bounces+importer=patchew.org@nongnu.org>
X-ZohoMail: RDMRC_1  RSF_0  Z_629925259 SPT_0
Content-Type: text/plain; charset="utf-8"

From: Richard Henderson <richard.henderson@linaro.org>

Move ssra_op and usra_op expanders from translate-a64.c.

Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Message-id: 20181011205206.3552-14-richard.henderson@linaro.org
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
---
 target/arm/translate.h     |   2 +
 target/arm/translate-a64.c | 106 ----------------------------
 target/arm/translate.c     | 139 ++++++++++++++++++++++++++++++++++---
 3 files changed, 130 insertions(+), 117 deletions(-)

diff --git a/target/arm/translate.h b/target/arm/translate.h
index dea59c7214d..5e13571b362 100644
--- a/target/arm/translate.h
+++ b/target/arm/translate.h
@@ -196,6 +196,8 @@ static inline TCGv_i32 get_ahp_flag(void)
 extern const GVecGen3 bsl_op;
 extern const GVecGen3 bit_op;
 extern const GVecGen3 bif_op;
+extern const GVecGen2i ssra_op[4];
+extern const GVecGen2i usra_op[4];
=20
 /*
  * Forward to the isar_feature_* tests given a DisasContext pointer.
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
index 384bcbbb00c..6d11e384898 100644
--- a/target/arm/translate-a64.c
+++ b/target/arm/translate-a64.c
@@ -9392,66 +9392,6 @@ static void disas_simd_scalar_two_reg_misc(DisasCont=
ext *s, uint32_t insn)
     }
 }
=20
-static void gen_ssra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_vec_sar8i_i64(a, a, shift);
-    tcg_gen_vec_add8_i64(d, d, a);
-}
-
-static void gen_ssra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_vec_sar16i_i64(a, a, shift);
-    tcg_gen_vec_add16_i64(d, d, a);
-}
-
-static void gen_ssra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
-{
-    tcg_gen_sari_i32(a, a, shift);
-    tcg_gen_add_i32(d, d, a);
-}
-
-static void gen_ssra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_sari_i64(a, a, shift);
-    tcg_gen_add_i64(d, d, a);
-}
-
-static void gen_ssra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
-{
-    tcg_gen_sari_vec(vece, a, a, sh);
-    tcg_gen_add_vec(vece, d, d, a);
-}
-
-static void gen_usra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_vec_shr8i_i64(a, a, shift);
-    tcg_gen_vec_add8_i64(d, d, a);
-}
-
-static void gen_usra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_vec_shr16i_i64(a, a, shift);
-    tcg_gen_vec_add16_i64(d, d, a);
-}
-
-static void gen_usra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
-{
-    tcg_gen_shri_i32(a, a, shift);
-    tcg_gen_add_i32(d, d, a);
-}
-
-static void gen_usra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
-{
-    tcg_gen_shri_i64(a, a, shift);
-    tcg_gen_add_i64(d, d, a);
-}
-
-static void gen_usra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
-{
-    tcg_gen_shri_vec(vece, a, a, sh);
-    tcg_gen_add_vec(vece, d, d, a);
-}
-
 static void gen_shr8_ins_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
 {
     uint64_t mask =3D dup_const(MO_8, 0xff >> shift);
@@ -9507,52 +9447,6 @@ static void gen_shr_ins_vec(unsigned vece, TCGv_vec =
d, TCGv_vec a, int64_t sh)
 static void handle_vec_simd_shri(DisasContext *s, bool is_q, bool is_u,
                                  int immh, int immb, int opcode, int rn, i=
nt rd)
 {
-    static const GVecGen2i ssra_op[4] =3D {
-        { .fni8 =3D gen_ssra8_i64,
-          .fniv =3D gen_ssra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_sari_vec,
-          .vece =3D MO_8 },
-        { .fni8 =3D gen_ssra16_i64,
-          .fniv =3D gen_ssra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_sari_vec,
-          .vece =3D MO_16 },
-        { .fni4 =3D gen_ssra32_i32,
-          .fniv =3D gen_ssra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_sari_vec,
-          .vece =3D MO_32 },
-        { .fni8 =3D gen_ssra64_i64,
-          .fniv =3D gen_ssra_vec,
-          .prefer_i64 =3D TCG_TARGET_REG_BITS =3D=3D 64,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_sari_vec,
-          .vece =3D MO_64 },
-    };
-    static const GVecGen2i usra_op[4] =3D {
-        { .fni8 =3D gen_usra8_i64,
-          .fniv =3D gen_usra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_shri_vec,
-          .vece =3D MO_8, },
-        { .fni8 =3D gen_usra16_i64,
-          .fniv =3D gen_usra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_shri_vec,
-          .vece =3D MO_16, },
-        { .fni4 =3D gen_usra32_i32,
-          .fniv =3D gen_usra_vec,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_shri_vec,
-          .vece =3D MO_32, },
-        { .fni8 =3D gen_usra64_i64,
-          .fniv =3D gen_usra_vec,
-          .prefer_i64 =3D TCG_TARGET_REG_BITS =3D=3D 64,
-          .load_dest =3D true,
-          .opc =3D INDEX_op_shri_vec,
-          .vece =3D MO_64, },
-    };
     static const GVecGen2i sri_op[4] =3D {
         { .fni8 =3D gen_shr8_ins_i64,
           .fniv =3D gen_shr_ins_vec,
diff --git a/target/arm/translate.c b/target/arm/translate.c
index 2d715d9b47b..b3b2ef93f4d 100644
--- a/target/arm/translate.c
+++ b/target/arm/translate.c
@@ -5770,6 +5770,113 @@ const GVecGen3 bif_op =3D {
     .load_dest =3D true
 };
=20
+static void gen_ssra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_vec_sar8i_i64(a, a, shift);
+    tcg_gen_vec_add8_i64(d, d, a);
+}
+
+static void gen_ssra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_vec_sar16i_i64(a, a, shift);
+    tcg_gen_vec_add16_i64(d, d, a);
+}
+
+static void gen_ssra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
+{
+    tcg_gen_sari_i32(a, a, shift);
+    tcg_gen_add_i32(d, d, a);
+}
+
+static void gen_ssra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_sari_i64(a, a, shift);
+    tcg_gen_add_i64(d, d, a);
+}
+
+static void gen_ssra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
+{
+    tcg_gen_sari_vec(vece, a, a, sh);
+    tcg_gen_add_vec(vece, d, d, a);
+}
+
+const GVecGen2i ssra_op[4] =3D {
+    { .fni8 =3D gen_ssra8_i64,
+      .fniv =3D gen_ssra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_sari_vec,
+      .vece =3D MO_8 },
+    { .fni8 =3D gen_ssra16_i64,
+      .fniv =3D gen_ssra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_sari_vec,
+      .vece =3D MO_16 },
+    { .fni4 =3D gen_ssra32_i32,
+      .fniv =3D gen_ssra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_sari_vec,
+      .vece =3D MO_32 },
+    { .fni8 =3D gen_ssra64_i64,
+      .fniv =3D gen_ssra_vec,
+      .prefer_i64 =3D TCG_TARGET_REG_BITS =3D=3D 64,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_sari_vec,
+      .vece =3D MO_64 },
+};
+
+static void gen_usra8_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_vec_shr8i_i64(a, a, shift);
+    tcg_gen_vec_add8_i64(d, d, a);
+}
+
+static void gen_usra16_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_vec_shr16i_i64(a, a, shift);
+    tcg_gen_vec_add16_i64(d, d, a);
+}
+
+static void gen_usra32_i32(TCGv_i32 d, TCGv_i32 a, int32_t shift)
+{
+    tcg_gen_shri_i32(a, a, shift);
+    tcg_gen_add_i32(d, d, a);
+}
+
+static void gen_usra64_i64(TCGv_i64 d, TCGv_i64 a, int64_t shift)
+{
+    tcg_gen_shri_i64(a, a, shift);
+    tcg_gen_add_i64(d, d, a);
+}
+
+static void gen_usra_vec(unsigned vece, TCGv_vec d, TCGv_vec a, int64_t sh)
+{
+    tcg_gen_shri_vec(vece, a, a, sh);
+    tcg_gen_add_vec(vece, d, d, a);
+}
+
+const GVecGen2i usra_op[4] =3D {
+    { .fni8 =3D gen_usra8_i64,
+      .fniv =3D gen_usra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_shri_vec,
+      .vece =3D MO_8, },
+    { .fni8 =3D gen_usra16_i64,
+      .fniv =3D gen_usra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_shri_vec,
+      .vece =3D MO_16, },
+    { .fni4 =3D gen_usra32_i32,
+      .fniv =3D gen_usra_vec,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_shri_vec,
+      .vece =3D MO_32, },
+    { .fni8 =3D gen_usra64_i64,
+      .fniv =3D gen_usra_vec,
+      .prefer_i64 =3D TCG_TARGET_REG_BITS =3D=3D 64,
+      .load_dest =3D true,
+      .opc =3D INDEX_op_shri_vec,
+      .vece =3D MO_64, },
+};
=20
 /* Translate a NEON data processing instruction.  Return nonzero if the
    instruction is invalid.
@@ -6408,6 +6515,25 @@ static int disas_neon_data_insn(DisasContext *s, uin=
t32_t insn)
                     }
                     return 0;
=20
+                case 1:  /* VSRA */
+                    /* Right shift comes here negative.  */
+                    shift =3D -shift;
+                    /* Shifts larger than the element size are architectur=
ally
+                     * valid.  Unsigned results in all zeros; signed resul=
ts
+                     * in all sign bits.
+                     */
+                    if (!u) {
+                        tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size, vec_size,
+                                        MIN(shift, (8 << size) - 1),
+                                        &ssra_op[size]);
+                    } else if (shift >=3D 8 << size) {
+                        /* rd +=3D 0 */
+                    } else {
+                        tcg_gen_gvec_2i(rd_ofs, rm_ofs, vec_size, vec_size,
+                                        shift, &usra_op[size]);
+                    }
+                    return 0;
+
                 case 5: /* VSHL, VSLI */
                     if (!u) { /* VSHL */
                         /* Shifts larger than the element size are
@@ -6440,12 +6566,6 @@ static int disas_neon_data_insn(DisasContext *s, uin=
t32_t insn)
                         neon_load_reg64(cpu_V0, rm + pass);
                         tcg_gen_movi_i64(cpu_V1, imm);
                         switch (op) {
-                        case 1:  /* VSRA */
-                            if (u)
-                                gen_helper_neon_shl_u64(cpu_V0, cpu_V0, cp=
u_V1);
-                            else
-                                gen_helper_neon_shl_s64(cpu_V0, cpu_V0, cp=
u_V1);
-                            break;
                         case 2: /* VRSHR */
                         case 3: /* VRSRA */
                             if (u)
@@ -6473,7 +6593,7 @@ static int disas_neon_data_insn(DisasContext *s, uint=
32_t insn)
                         default:
                             g_assert_not_reached();
                         }
-                        if (op =3D=3D 1 || op =3D=3D 3) {
+                        if (op =3D=3D 3) {
                             /* Accumulate.  */
                             neon_load_reg64(cpu_V1, rd + pass);
                             tcg_gen_add_i64(cpu_V0, cpu_V0, cpu_V1);
@@ -6500,9 +6620,6 @@ static int disas_neon_data_insn(DisasContext *s, uint=
32_t insn)
                         tmp2 =3D tcg_temp_new_i32();
                         tcg_gen_movi_i32(tmp2, imm);
                         switch (op) {
-                        case 1:  /* VSRA */
-                            GEN_NEON_INTEGER_OP(shl);
-                            break;
                         case 2: /* VRSHR */
                         case 3: /* VRSRA */
                             GEN_NEON_INTEGER_OP(rshl);
@@ -6542,7 +6659,7 @@ static int disas_neon_data_insn(DisasContext *s, uint=
32_t insn)
                         }
                         tcg_temp_free_i32(tmp2);
=20
-                        if (op =3D=3D 1 || op =3D=3D 3) {
+                        if (op =3D=3D 3) {
                             /* Accumulate.  */
                             tmp2 =3D neon_load_reg(rd, pass);
                             gen_neon_add(size, tmp, tmp2);
--=20
2.19.1