[PATCH 6/8] target/riscv: Fix format for indentation

Weiwei Li posted 8 patches 2 years, 10 months ago
There is a newer version of this series
[PATCH 6/8] target/riscv: Fix format for indentation
Posted by Weiwei Li 2 years, 10 months ago
Fix identation problems, and try to use the same indentation strategy
in the same file.

Signed-off-by: Weiwei Li <liweiwei@iscas.ac.cn>
Signed-off-by: Junqiang Wang <wangjunqiang@iscas.ac.cn>
---
 target/riscv/arch_dump.c                |   4 +-
 target/riscv/cpu.c                      |   4 +-
 target/riscv/cpu_helper.c               |  15 +--
 target/riscv/insn_trans/trans_rvv.c.inc |  46 ++++----
 target/riscv/op_helper.c                |   4 +-
 target/riscv/pmp.c                      |  19 ++--
 target/riscv/pmp.h                      |   9 +-
 target/riscv/vector_helper.c            | 134 +++++++++++++-----------
 8 files changed, 122 insertions(+), 113 deletions(-)

diff --git a/target/riscv/arch_dump.c b/target/riscv/arch_dump.c
index 736a232956..573587810e 100644
--- a/target/riscv/arch_dump.c
+++ b/target/riscv/arch_dump.c
@@ -180,8 +180,8 @@ int cpu_get_dump_info(ArchDumpInfo *info,
     info->d_class = ELFCLASS32;
 #endif
 
-    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0
-                     ? ELFDATA2MSB : ELFDATA2LSB;
+    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0 ?
+                     ELFDATA2MSB : ELFDATA2LSB;
 
     return 0;
 }
diff --git a/target/riscv/cpu.c b/target/riscv/cpu.c
index 16e465a0ab..75dab70ba7 100644
--- a/target/riscv/cpu.c
+++ b/target/riscv/cpu.c
@@ -344,8 +344,8 @@ static void riscv_any_cpu_init(Object *obj)
 
 #ifndef CONFIG_USER_ONLY
     set_satp_mode_max_supported(RISCV_CPU(obj),
-            riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
-                                    VM_1_10_SV32 : VM_1_10_SV57);
+        riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
+        VM_1_10_SV32 : VM_1_10_SV57);
 #endif
 
     set_priv_version(env, PRIV_VERSION_1_12_0);
diff --git a/target/riscv/cpu_helper.c b/target/riscv/cpu_helper.c
index de2d4a8c1d..08689ee3f6 100644
--- a/target/riscv/cpu_helper.c
+++ b/target/riscv/cpu_helper.c
@@ -68,12 +68,12 @@ void cpu_get_tb_cpu_state(CPURISCVState *env, target_ulong *pc,
         flags = FIELD_DP32(flags, TB_FLAGS, VILL, env->vill);
         flags = FIELD_DP32(flags, TB_FLAGS, SEW, sew);
         flags = FIELD_DP32(flags, TB_FLAGS, LMUL,
-                    FIELD_EX64(env->vtype, VTYPE, VLMUL));
+                           FIELD_EX64(env->vtype, VTYPE, VLMUL));
         flags = FIELD_DP32(flags, TB_FLAGS, VL_EQ_VLMAX, vl_eq_vlmax);
         flags = FIELD_DP32(flags, TB_FLAGS, VTA,
-                    FIELD_EX64(env->vtype, VTYPE, VTA));
+                           FIELD_EX64(env->vtype, VTYPE, VTA));
         flags = FIELD_DP32(flags, TB_FLAGS, VMA,
-                    FIELD_EX64(env->vtype, VTYPE, VMA));
+                           FIELD_EX64(env->vtype, VTYPE, VMA));
     } else {
         flags = FIELD_DP32(flags, TB_FLAGS, VILL, 1);
     }
@@ -1056,7 +1056,7 @@ restart:
             /* add write permission on stores or if the page is already dirty,
                so that we TLB miss on later writes to update the dirty bit */
             if ((pte & PTE_W) &&
-                    (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
+                (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
                 *prot |= PAGE_WRITE;
             }
             return TRANSLATE_SUCCESS;
@@ -1285,9 +1285,10 @@ bool riscv_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
                                        false);
 
             qemu_log_mask(CPU_LOG_MMU,
-                    "%s 2nd-stage address=%" VADDR_PRIx " ret %d physical "
-                    HWADDR_FMT_plx " prot %d\n",
-                    __func__, im_address, ret, pa, prot2);
+                          "%s 2nd-stage address=%" VADDR_PRIx
+                          " ret %d physical "
+                          HWADDR_FMT_plx " prot %d\n",
+                          __func__, im_address, ret, pa, prot2);
 
             prot &= prot2;
 
diff --git a/target/riscv/insn_trans/trans_rvv.c.inc b/target/riscv/insn_trans/trans_rvv.c.inc
index f2e3d38515..2aed66934a 100644
--- a/target/riscv/insn_trans/trans_rvv.c.inc
+++ b/target/riscv/insn_trans/trans_rvv.c.inc
@@ -238,8 +238,8 @@ static bool vext_check_store(DisasContext *s, int vd, int nf, uint8_t eew)
 {
     int8_t emul = eew - s->sew + s->lmul;
     return (emul >= -3 && emul <= 3) &&
-            require_align(vd, emul) &&
-            require_nf(vd, nf, emul);
+           require_align(vd, emul) &&
+           require_nf(vd, nf, emul);
 }
 
 /*
@@ -315,7 +315,7 @@ static bool vext_check_ld_index(DisasContext *s, int vd, int vs2,
     int8_t seg_vd;
     int8_t emul = eew - s->sew + s->lmul;
     bool ret = vext_check_st_index(s, vd, vs2, nf, eew) &&
-        require_vm(vm, vd);
+               require_vm(vm, vd);
 
     /* Each segment register group has to follow overlap rules. */
     for (int i = 0; i < nf; ++i) {
@@ -345,8 +345,8 @@ static bool vext_check_ld_index(DisasContext *s, int vd, int vs2,
 static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
 {
     return require_vm(vm, vd) &&
-        require_align(vd, s->lmul) &&
-        require_align(vs, s->lmul);
+           require_align(vd, s->lmul) &&
+           require_align(vs, s->lmul);
 }
 
 /*
@@ -365,7 +365,7 @@ static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
 static bool vext_check_sss(DisasContext *s, int vd, int vs1, int vs2, int vm)
 {
     return vext_check_ss(s, vd, vs2, vm) &&
-        require_align(vs1, s->lmul);
+           require_align(vs1, s->lmul);
 }
 
 static bool vext_check_ms(DisasContext *s, int vd, int vs)
@@ -396,7 +396,7 @@ static bool vext_check_ms(DisasContext *s, int vd, int vs)
 static bool vext_check_mss(DisasContext *s, int vd, int vs1, int vs2)
 {
     bool ret = vext_check_ms(s, vd, vs2) &&
-        require_align(vs1, s->lmul);
+               require_align(vs1, s->lmul);
     if (vd != vs1) {
         ret &= require_noover(vd, 0, vs1, s->lmul);
     }
@@ -460,14 +460,14 @@ static bool vext_narrow_check_common(DisasContext *s, int vd, int vs2,
 static bool vext_check_ds(DisasContext *s, int vd, int vs, int vm)
 {
     return vext_wide_check_common(s, vd, vm) &&
-        require_align(vs, s->lmul) &&
-        require_noover(vd, s->lmul + 1, vs, s->lmul);
+           require_align(vs, s->lmul) &&
+           require_noover(vd, s->lmul + 1, vs, s->lmul);
 }
 
 static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
 {
     return vext_wide_check_common(s, vd, vm) &&
-        require_align(vs, s->lmul + 1);
+           require_align(vs, s->lmul + 1);
 }
 
 /*
@@ -485,8 +485,8 @@ static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
 static bool vext_check_dss(DisasContext *s, int vd, int vs1, int vs2, int vm)
 {
     return vext_check_ds(s, vd, vs2, vm) &&
-        require_align(vs1, s->lmul) &&
-        require_noover(vd, s->lmul + 1, vs1, s->lmul);
+           require_align(vs1, s->lmul) &&
+           require_noover(vd, s->lmul + 1, vs1, s->lmul);
 }
 
 /*
@@ -507,7 +507,7 @@ static bool vext_check_dss(DisasContext *s, int vd, int vs1, int vs2, int vm)
 static bool vext_check_dds(DisasContext *s, int vd, int vs1, int vs2, int vm)
 {
     return vext_check_ds(s, vd, vs1, vm) &&
-        require_align(vs2, s->lmul + 1);
+           require_align(vs2, s->lmul + 1);
 }
 
 static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
@@ -535,7 +535,7 @@ static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
 static bool vext_check_sds(DisasContext *s, int vd, int vs1, int vs2, int vm)
 {
     return vext_check_sd(s, vd, vs2, vm) &&
-        require_align(vs1, s->lmul);
+           require_align(vs1, s->lmul);
 }
 
 /*
@@ -1216,7 +1216,7 @@ do_opivv_gvec(DisasContext *s, arg_rmrr *a, GVecGen3Fn *gvec_fn,
 }
 
 /* OPIVV with GVEC IR */
-#define GEN_OPIVV_GVEC_TRANS(NAME, SUF) \
+#define GEN_OPIVV_GVEC_TRANS(NAME, SUF)                            \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
 {                                                                  \
     static gen_helper_gvec_4_ptr * const fns[4] = {                \
@@ -1300,7 +1300,7 @@ do_opivx_gvec(DisasContext *s, arg_rmrr *a, GVecGen2sFn *gvec_fn,
 }
 
 /* OPIVX with GVEC IR */
-#define GEN_OPIVX_GVEC_TRANS(NAME, SUF) \
+#define GEN_OPIVX_GVEC_TRANS(NAME, SUF)                            \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
 {                                                                  \
     static gen_helper_opivx * const fns[4] = {                     \
@@ -1453,7 +1453,7 @@ do_opivi_gvec(DisasContext *s, arg_rmrr *a, GVecGen2iFn *gvec_fn,
 }
 
 /* OPIVI with GVEC IR */
-#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF) \
+#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF)           \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
 {                                                                  \
     static gen_helper_opivx * const fns[4] = {                     \
@@ -1512,7 +1512,7 @@ static bool do_opivv_widen(DisasContext *s, arg_rmrr *a,
     return false;
 }
 
-#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK) \
+#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK)                   \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
 {                                                            \
     static gen_helper_gvec_4_ptr * const fns[3] = {          \
@@ -1545,7 +1545,7 @@ static bool do_opivx_widen(DisasContext *s, arg_rmrr *a,
     return false;
 }
 
-#define GEN_OPIVX_WIDEN_TRANS(NAME) \
+#define GEN_OPIVX_WIDEN_TRANS(NAME)                          \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
 {                                                            \
     static gen_helper_opivx * const fns[3] = {               \
@@ -1594,7 +1594,7 @@ static bool do_opiwv_widen(DisasContext *s, arg_rmrr *a,
     return false;
 }
 
-#define GEN_OPIWV_WIDEN_TRANS(NAME) \
+#define GEN_OPIWV_WIDEN_TRANS(NAME)                          \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
 {                                                            \
     static gen_helper_gvec_4_ptr * const fns[3] = {          \
@@ -1627,7 +1627,7 @@ static bool do_opiwx_widen(DisasContext *s, arg_rmrr *a,
     return false;
 }
 
-#define GEN_OPIWX_WIDEN_TRANS(NAME) \
+#define GEN_OPIWX_WIDEN_TRANS(NAME)                          \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
 {                                                            \
     static gen_helper_opivx * const fns[3] = {               \
@@ -1801,7 +1801,7 @@ do_opivx_gvec_shift(DisasContext *s, arg_rmrr *a, GVecGen2sFn32 *gvec_fn,
     return opivx_trans(a->rd, a->rs1, a->rs2, a->vm, fn, s);
 }
 
-#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, SUF) \
+#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, SUF)                             \
 static bool trans_##NAME(DisasContext *s, arg_rmrr *a)                    \
 {                                                                         \
     static gen_helper_opivx * const fns[4] = {                            \
@@ -3668,7 +3668,7 @@ static bool trans_vcompress_vm(DisasContext *s, arg_r *a)
  * Whole Vector Register Move Instructions ignore vtype and vl setting.
  * Thus, we don't need to check vill bit. (Section 16.6)
  */
-#define GEN_VMV_WHOLE_TRANS(NAME, LEN)                             \
+#define GEN_VMV_WHOLE_TRANS(NAME, LEN)                                  \
 static bool trans_##NAME(DisasContext *s, arg_##NAME * a)               \
 {                                                                       \
     if (require_rvv(s) &&                                               \
diff --git a/target/riscv/op_helper.c b/target/riscv/op_helper.c
index 1eecae9547..478365131d 100644
--- a/target/riscv/op_helper.c
+++ b/target/riscv/op_helper.c
@@ -367,8 +367,8 @@ void helper_wfi(CPURISCVState *env)
     if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, MSTATUS_TW)) ||
         (rvs && prv_u && !riscv_cpu_virt_enabled(env))) {
         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
-    } else if (riscv_cpu_virt_enabled(env) && (prv_u ||
-        (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
+    } else if (riscv_cpu_virt_enabled(env) &&
+               (prv_u || (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
     } else {
         cs->halted = 1;
diff --git a/target/riscv/pmp.c b/target/riscv/pmp.c
index a08cd95658..3943b0f2e3 100644
--- a/target/riscv/pmp.c
+++ b/target/riscv/pmp.c
@@ -27,7 +27,7 @@
 #include "exec/exec-all.h"
 
 static void pmp_write_cfg(CPURISCVState *env, uint32_t addr_index,
-    uint8_t val);
+                          uint8_t val);
 static uint8_t pmp_read_cfg(CPURISCVState *env, uint32_t addr_index);
 static void pmp_update_rule(CPURISCVState *env, uint32_t pmp_index);
 
@@ -220,8 +220,8 @@ static int pmp_is_in_range(CPURISCVState *env, int pmp_index, target_ulong addr)
 {
     int result = 0;
 
-    if ((addr >= env->pmp_state.addr[pmp_index].sa)
-        && (addr <= env->pmp_state.addr[pmp_index].ea)) {
+    if ((addr >= env->pmp_state.addr[pmp_index].sa) &&
+        (addr <= env->pmp_state.addr[pmp_index].ea)) {
         result = 1;
     } else {
         result = 0;
@@ -234,8 +234,9 @@ static int pmp_is_in_range(CPURISCVState *env, int pmp_index, target_ulong addr)
  * Check if the address has required RWX privs when no PMP entry is matched.
  */
 static bool pmp_hart_has_privs_default(CPURISCVState *env, target_ulong addr,
-    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
-    target_ulong mode)
+                                       target_ulong size, pmp_priv_t privs,
+                                       pmp_priv_t *allowed_privs,
+                                       target_ulong mode)
 {
     bool ret;
 
@@ -297,8 +298,8 @@ static bool pmp_hart_has_privs_default(CPURISCVState *env, target_ulong addr,
  * Return negtive value if no match
  */
 int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
-    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
-    target_ulong mode)
+                       target_ulong size, pmp_priv_t privs,
+                       pmp_priv_t *allowed_privs, target_ulong mode)
 {
     int i = 0;
     int ret = -1;
@@ -466,7 +467,7 @@ int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
  * Handle a write to a pmpcfg CSR
  */
 void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
-    target_ulong val)
+                      target_ulong val)
 {
     int i;
     uint8_t cfg_val;
@@ -508,7 +509,7 @@ target_ulong pmpcfg_csr_read(CPURISCVState *env, uint32_t reg_index)
  * Handle a write to a pmpaddr CSR
  */
 void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
-    target_ulong val)
+                       target_ulong val)
 {
     trace_pmpaddr_csr_write(env->mhartid, addr_index, val);
 
diff --git a/target/riscv/pmp.h b/target/riscv/pmp.h
index da32c61c85..b296ea1fc6 100644
--- a/target/riscv/pmp.h
+++ b/target/riscv/pmp.h
@@ -63,18 +63,19 @@ typedef struct {
 } pmp_table_t;
 
 void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
-    target_ulong val);
+                      target_ulong val);
 target_ulong pmpcfg_csr_read(CPURISCVState *env, uint32_t reg_index);
 
 void mseccfg_csr_write(CPURISCVState *env, target_ulong val);
 target_ulong mseccfg_csr_read(CPURISCVState *env);
 
 void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
-    target_ulong val);
+                       target_ulong val);
 target_ulong pmpaddr_csr_read(CPURISCVState *env, uint32_t addr_index);
 int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
-    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
-    target_ulong mode);
+                       target_ulong size, pmp_priv_t privs,
+                       pmp_priv_t *allowed_privs,
+                       target_ulong mode);
 target_ulong pmp_get_tlb_size(CPURISCVState *env, int pmp_index,
                               target_ulong tlb_sa, target_ulong tlb_ea);
 void pmp_update_rule_addr(CPURISCVState *env, uint32_t pmp_index);
diff --git a/target/riscv/vector_helper.c b/target/riscv/vector_helper.c
index 2423affe37..df6201d043 100644
--- a/target/riscv/vector_helper.c
+++ b/target/riscv/vector_helper.c
@@ -1116,7 +1116,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2,        \
                                                                          \
         *((ETYPE *)vd + H(i)) = DO_OP(s2, (ETYPE)(target_long)s1, carry);\
     }                                                                    \
-    env->vstart = 0;                                          \
+    env->vstart = 0;                                                     \
     /* set tail elements to 1s */                                        \
     vext_set_elems_1s(vd, vta, vl * esz, total_elems * esz);             \
 }
@@ -1308,7 +1308,8 @@ GEN_VEXT_SHIFT_VV(vsra_vv_d, uint64_t, int64_t, H8, H8, DO_SRL, 0x3f)
 /* generate the helpers for shift instructions with one vector and one scalar */
 #define GEN_VEXT_SHIFT_VX(NAME, TD, TS2, HD, HS2, OP, MASK) \
 void HELPER(NAME)(void *vd, void *v0, target_ulong s1,      \
-        void *vs2, CPURISCVState *env, uint32_t desc)       \
+                  void *vs2, CPURISCVState *env,            \
+                  uint32_t desc)                            \
 {                                                           \
     uint32_t vm = vext_vm(desc);                            \
     uint32_t vl = env->vl;                                  \
@@ -1735,9 +1736,9 @@ GEN_VEXT_VX(vmulhsu_vx_d, 8)
 /* Vector Integer Divide Instructions */
 #define DO_DIVU(N, M) (unlikely(M == 0) ? (__typeof(N))(-1) : N / M)
 #define DO_REMU(N, M) (unlikely(M == 0) ? N : N % M)
-#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) :\
+#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) : \
         unlikely((N == -N) && (M == (__typeof(N))(-1))) ? N : N / M)
-#define DO_REM(N, M)  (unlikely(M == 0) ? N :\
+#define DO_REM(N, M)  (unlikely(M == 0) ? N : \
         unlikely((N == -N) && (M == (__typeof(N))(-1))) ? 0 : N % M)
 
 RVVCALL(OPIVV2, vdivu_vv_b, OP_UUU_B, H1, H1, H1, DO_DIVU)
@@ -1846,7 +1847,7 @@ GEN_VEXT_VX(vwmulsu_vx_h, 4)
 GEN_VEXT_VX(vwmulsu_vx_w, 8)
 
 /* Vector Single-Width Integer Multiply-Add Instructions */
-#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)   \
+#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
 static void do_##NAME(void *vd, void *vs1, void *vs2, int i)       \
 {                                                                  \
     TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
@@ -2277,7 +2278,8 @@ vext_vx_rm_2(void *vd, void *v0, target_long s1, void *vs2,
 /* generate helpers for fixed point instructions with OPIVX format */
 #define GEN_VEXT_VX_RM(NAME, ESZ)                         \
 void HELPER(NAME)(void *vd, void *v0, target_ulong s1,    \
-        void *vs2, CPURISCVState *env, uint32_t desc)     \
+                  void *vs2, CPURISCVState *env,          \
+                  uint32_t desc)                          \
 {                                                         \
     vext_vx_rm_2(vd, v0, s1, vs2, env, desc,              \
                  do_##NAME, ESZ);                         \
@@ -3052,7 +3054,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1,        \
     uint32_t vm = vext_vm(desc);                          \
     uint32_t vl = env->vl;                                \
     uint32_t total_elems =                                \
-        vext_get_total_elems(env, desc, ESZ);              \
+        vext_get_total_elems(env, desc, ESZ);             \
     uint32_t vta = vext_vta(desc);                        \
     uint32_t vma = vext_vma(desc);                        \
     uint32_t i;                                           \
@@ -3118,13 +3120,13 @@ GEN_VEXT_VF(vfrsub_vf_d, 8)
 static uint32_t vfwadd16(uint16_t a, uint16_t b, float_status *s)
 {
     return float32_add(float16_to_float32(a, true, s),
-            float16_to_float32(b, true, s), s);
+                       float16_to_float32(b, true, s), s);
 }
 
 static uint64_t vfwadd32(uint32_t a, uint32_t b, float_status *s)
 {
     return float64_add(float32_to_float64(a, s),
-            float32_to_float64(b, s), s);
+                       float32_to_float64(b, s), s);
 
 }
 
@@ -3140,13 +3142,13 @@ GEN_VEXT_VF(vfwadd_vf_w, 8)
 static uint32_t vfwsub16(uint16_t a, uint16_t b, float_status *s)
 {
     return float32_sub(float16_to_float32(a, true, s),
-            float16_to_float32(b, true, s), s);
+                       float16_to_float32(b, true, s), s);
 }
 
 static uint64_t vfwsub32(uint32_t a, uint32_t b, float_status *s)
 {
     return float64_sub(float32_to_float64(a, s),
-            float32_to_float64(b, s), s);
+                       float32_to_float64(b, s), s);
 
 }
 
@@ -3250,13 +3252,13 @@ GEN_VEXT_VF(vfrdiv_vf_d, 8)
 static uint32_t vfwmul16(uint16_t a, uint16_t b, float_status *s)
 {
     return float32_mul(float16_to_float32(a, true, s),
-            float16_to_float32(b, true, s), s);
+                       float16_to_float32(b, true, s), s);
 }
 
 static uint64_t vfwmul32(uint32_t a, uint32_t b, float_status *s)
 {
     return float64_mul(float32_to_float64(a, s),
-            float32_to_float64(b, s), s);
+                       float32_to_float64(b, s), s);
 
 }
 RVVCALL(OPFVV2, vfwmul_vv_h, WOP_UUU_H, H4, H2, H2, vfwmul16)
@@ -3271,7 +3273,7 @@ GEN_VEXT_VF(vfwmul_vf_w, 8)
 /* Vector Single-Width Floating-Point Fused Multiply-Add Instructions */
 #define OPFVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
 static void do_##NAME(void *vd, void *vs1, void *vs2, int i,       \
-        CPURISCVState *env)                                        \
+                      CPURISCVState *env)                          \
 {                                                                  \
     TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
     TX2 s2 = *((T2 *)vs2 + HS2(i));                                \
@@ -3303,7 +3305,7 @@ GEN_VEXT_VV_ENV(vfmacc_vv_d, 8)
 
 #define OPFVF3(NAME, TD, T1, T2, TX1, TX2, HD, HS2, OP)           \
 static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i,    \
-        CPURISCVState *env)                                       \
+                      CPURISCVState *env)                         \
 {                                                                 \
     TX2 s2 = *((T2 *)vs2 + HS2(i));                               \
     TD d = *((TD *)vd + HD(i));                                   \
@@ -3319,20 +3321,20 @@ GEN_VEXT_VF(vfmacc_vf_d, 8)
 
 static uint16_t fnmacc16(uint16_t a, uint16_t b, uint16_t d, float_status *s)
 {
-    return float16_muladd(a, b, d,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float16_muladd(a, b, d, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 static uint32_t fnmacc32(uint32_t a, uint32_t b, uint32_t d, float_status *s)
 {
-    return float32_muladd(a, b, d,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float32_muladd(a, b, d, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 static uint64_t fnmacc64(uint64_t a, uint64_t b, uint64_t d, float_status *s)
 {
-    return float64_muladd(a, b, d,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float64_muladd(a, b, d, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 RVVCALL(OPFVV3, vfnmacc_vv_h, OP_UUU_H, H2, H2, H2, fnmacc16)
@@ -3434,20 +3436,20 @@ GEN_VEXT_VF(vfmadd_vf_d, 8)
 
 static uint16_t fnmadd16(uint16_t a, uint16_t b, uint16_t d, float_status *s)
 {
-    return float16_muladd(d, b, a,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float16_muladd(d, b, a, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 static uint32_t fnmadd32(uint32_t a, uint32_t b, uint32_t d, float_status *s)
 {
-    return float32_muladd(d, b, a,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float32_muladd(d, b, a, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 static uint64_t fnmadd64(uint64_t a, uint64_t b, uint64_t d, float_status *s)
 {
-    return float64_muladd(d, b, a,
-            float_muladd_negate_c | float_muladd_negate_product, s);
+    return float64_muladd(d, b, a, float_muladd_negate_c |
+                                   float_muladd_negate_product, s);
 }
 
 RVVCALL(OPFVV3, vfnmadd_vv_h, OP_UUU_H, H2, H2, H2, fnmadd16)
@@ -3523,13 +3525,13 @@ GEN_VEXT_VF(vfnmsub_vf_d, 8)
 static uint32_t fwmacc16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
 {
     return float32_muladd(float16_to_float32(a, true, s),
-                        float16_to_float32(b, true, s), d, 0, s);
+                          float16_to_float32(b, true, s), d, 0, s);
 }
 
 static uint64_t fwmacc32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
 {
     return float64_muladd(float32_to_float64(a, s),
-                        float32_to_float64(b, s), d, 0, s);
+                          float32_to_float64(b, s), d, 0, s);
 }
 
 RVVCALL(OPFVV3, vfwmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwmacc16)
@@ -3544,15 +3546,16 @@ GEN_VEXT_VF(vfwmacc_vf_w, 8)
 static uint32_t fwnmacc16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
 {
     return float32_muladd(float16_to_float32(a, true, s),
-                        float16_to_float32(b, true, s), d,
-                        float_muladd_negate_c | float_muladd_negate_product, s);
+                          float16_to_float32(b, true, s), d,
+                          float_muladd_negate_c | float_muladd_negate_product,
+                          s);
 }
 
 static uint64_t fwnmacc32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
 {
-    return float64_muladd(float32_to_float64(a, s),
-                        float32_to_float64(b, s), d,
-                        float_muladd_negate_c | float_muladd_negate_product, s);
+    return float64_muladd(float32_to_float64(a, s), float32_to_float64(b, s),
+                          d, float_muladd_negate_c |
+                             float_muladd_negate_product, s);
 }
 
 RVVCALL(OPFVV3, vfwnmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwnmacc16)
@@ -3567,15 +3570,15 @@ GEN_VEXT_VF(vfwnmacc_vf_w, 8)
 static uint32_t fwmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
 {
     return float32_muladd(float16_to_float32(a, true, s),
-                        float16_to_float32(b, true, s), d,
-                        float_muladd_negate_c, s);
+                          float16_to_float32(b, true, s), d,
+                          float_muladd_negate_c, s);
 }
 
 static uint64_t fwmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
 {
     return float64_muladd(float32_to_float64(a, s),
-                        float32_to_float64(b, s), d,
-                        float_muladd_negate_c, s);
+                          float32_to_float64(b, s), d,
+                          float_muladd_negate_c, s);
 }
 
 RVVCALL(OPFVV3, vfwmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwmsac16)
@@ -3590,15 +3593,15 @@ GEN_VEXT_VF(vfwmsac_vf_w, 8)
 static uint32_t fwnmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
 {
     return float32_muladd(float16_to_float32(a, true, s),
-                        float16_to_float32(b, true, s), d,
-                        float_muladd_negate_product, s);
+                          float16_to_float32(b, true, s), d,
+                          float_muladd_negate_product, s);
 }
 
 static uint64_t fwnmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
 {
     return float64_muladd(float32_to_float64(a, s),
-                        float32_to_float64(b, s), d,
-                        float_muladd_negate_product, s);
+                          float32_to_float64(b, s), d,
+                          float_muladd_negate_product, s);
 }
 
 RVVCALL(OPFVV3, vfwnmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwnmsac16)
@@ -3616,9 +3619,9 @@ GEN_VEXT_VF(vfwnmsac_vf_w, 8)
 #define OP_UU_W uint32_t, uint32_t, uint32_t
 #define OP_UU_D uint64_t, uint64_t, uint64_t
 
-#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)        \
+#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)         \
 static void do_##NAME(void *vd, void *vs2, int i,      \
-        CPURISCVState *env)                            \
+                      CPURISCVState *env)              \
 {                                                      \
     TX2 s2 = *((T2 *)vs2 + HS2(i));                    \
     *((TD *)vd + HD(i)) = OP(s2, &env->fp_status);     \
@@ -3626,7 +3629,7 @@ static void do_##NAME(void *vd, void *vs2, int i,      \
 
 #define GEN_VEXT_V_ENV(NAME, ESZ)                      \
 void HELPER(NAME)(void *vd, void *v0, void *vs2,       \
-        CPURISCVState *env, uint32_t desc)             \
+                  CPURISCVState *env, uint32_t desc)   \
 {                                                      \
     uint32_t vm = vext_vm(desc);                       \
     uint32_t vl = env->vl;                             \
@@ -3703,9 +3706,9 @@ static uint64_t frsqrt7(uint64_t f, int exp_size, int frac_size)
     }
 
     int idx = ((exp & 1) << (precision - 1)) |
-                (frac >> (frac_size - precision + 1));
+              (frac >> (frac_size - precision + 1));
     uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
-                            (frac_size - precision);
+                        (frac_size - precision);
     uint64_t out_exp = (3 * MAKE_64BIT_MASK(0, exp_size - 1) + ~exp) / 2;
 
     uint64_t val = 0;
@@ -3727,9 +3730,9 @@ static float16 frsqrt7_h(float16 f, float_status *s)
      * frsqrt7(-subnormal) = canonical NaN
      */
     if (float16_is_signaling_nan(f, s) ||
-            (float16_is_infinity(f) && sign) ||
-            (float16_is_normal(f) && sign) ||
-            (float16_is_zero_or_denormal(f) && !float16_is_zero(f) && sign)) {
+        (float16_is_infinity(f) && sign) ||
+        (float16_is_normal(f) && sign) ||
+        (float16_is_zero_or_denormal(f) && !float16_is_zero(f) && sign)) {
         s->float_exception_flags |= float_flag_invalid;
         return float16_default_nan(s);
     }
@@ -3767,9 +3770,9 @@ static float32 frsqrt7_s(float32 f, float_status *s)
      * frsqrt7(-subnormal) = canonical NaN
      */
     if (float32_is_signaling_nan(f, s) ||
-            (float32_is_infinity(f) && sign) ||
-            (float32_is_normal(f) && sign) ||
-            (float32_is_zero_or_denormal(f) && !float32_is_zero(f) && sign)) {
+        (float32_is_infinity(f) && sign) ||
+        (float32_is_normal(f) && sign) ||
+        (float32_is_zero_or_denormal(f) && !float32_is_zero(f) && sign)) {
         s->float_exception_flags |= float_flag_invalid;
         return float32_default_nan(s);
     }
@@ -3807,9 +3810,9 @@ static float64 frsqrt7_d(float64 f, float_status *s)
      * frsqrt7(-subnormal) = canonical NaN
      */
     if (float64_is_signaling_nan(f, s) ||
-            (float64_is_infinity(f) && sign) ||
-            (float64_is_normal(f) && sign) ||
-            (float64_is_zero_or_denormal(f) && !float64_is_zero(f) && sign)) {
+        (float64_is_infinity(f) && sign) ||
+        (float64_is_normal(f) && sign) ||
+        (float64_is_zero_or_denormal(f) && !float64_is_zero(f) && sign)) {
         s->float_exception_flags |= float_flag_invalid;
         return float64_default_nan(s);
     }
@@ -3897,18 +3900,18 @@ static uint64_t frec7(uint64_t f, int exp_size, int frac_size,
                 ((s->float_rounding_mode == float_round_up) && sign)) {
                 /* Return greatest/negative finite value. */
                 return (sign << (exp_size + frac_size)) |
-                    (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
+                       (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
             } else {
                 /* Return +-inf. */
                 return (sign << (exp_size + frac_size)) |
-                    MAKE_64BIT_MASK(frac_size, exp_size);
+                       MAKE_64BIT_MASK(frac_size, exp_size);
             }
         }
     }
 
     int idx = frac >> (frac_size - precision);
     uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
-                            (frac_size - precision);
+                        (frac_size - precision);
     uint64_t out_exp = 2 * MAKE_64BIT_MASK(0, exp_size - 1) + ~exp;
 
     if (out_exp == 0 || out_exp == UINT64_MAX) {
@@ -4422,8 +4425,8 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
                                                               \
     for (i = env->vstart; i < vl; i++) {                      \
         ETYPE s2 = *((ETYPE *)vs2 + H(i));                    \
-        *((ETYPE *)vd + H(i))                                 \
-          = (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
+        *((ETYPE *)vd + H(i)) =                               \
+            (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
     }                                                         \
     env->vstart = 0;                                          \
     /* set tail elements to 1s */                             \
@@ -4564,7 +4567,8 @@ GEN_VEXT_V_ENV(vfncvt_f_f_w_w, 4)
 /* Vector Single-Width Integer Reduction Instructions */
 #define GEN_VEXT_RED(NAME, TD, TS2, HD, HS2, OP)          \
 void HELPER(NAME)(void *vd, void *v0, void *vs1,          \
-        void *vs2, CPURISCVState *env, uint32_t desc)     \
+                  void *vs2, CPURISCVState *env,          \
+                  uint32_t desc)                          \
 {                                                         \
     uint32_t vm = vext_vm(desc);                          \
     uint32_t vl = env->vl;                                \
@@ -5013,7 +5017,8 @@ GEN_VEXT_VSLIDEDOWN_VX(vslidedown_vx_d, uint64_t, H8)
 
 #define GEN_VEXT_VSLIE1UP(BITWIDTH, H)                                      \
 static void vslide1up_##BITWIDTH(void *vd, void *v0, uint64_t s1,           \
-                     void *vs2, CPURISCVState *env, uint32_t desc)          \
+                                 void *vs2, CPURISCVState *env,             \
+                                 uint32_t desc)                             \
 {                                                                           \
     typedef uint##BITWIDTH##_t ETYPE;                                       \
     uint32_t vm = vext_vm(desc);                                            \
@@ -5061,7 +5066,8 @@ GEN_VEXT_VSLIDE1UP_VX(vslide1up_vx_d, 64)
 
 #define GEN_VEXT_VSLIDE1DOWN(BITWIDTH, H)                                     \
 static void vslide1down_##BITWIDTH(void *vd, void *v0, uint64_t s1,           \
-                       void *vs2, CPURISCVState *env, uint32_t desc)          \
+                                   void *vs2, CPURISCVState *env,             \
+                                   uint32_t desc)                             \
 {                                                                             \
     typedef uint##BITWIDTH##_t ETYPE;                                         \
     uint32_t vm = vext_vm(desc);                                              \
-- 
2.25.1
Re: [PATCH 6/8] target/riscv: Fix format for indentation
Posted by LIU Zhiwei 2 years, 10 months ago
On 2023/3/24 20:38, Weiwei Li wrote:
> Fix identation problems, and try to use the same indentation strategy
> in the same file.
>
> Signed-off-by: Weiwei Li <liweiwei@iscas.ac.cn>
> Signed-off-by: Junqiang Wang <wangjunqiang@iscas.ac.cn>
> ---
>   target/riscv/arch_dump.c                |   4 +-
>   target/riscv/cpu.c                      |   4 +-
>   target/riscv/cpu_helper.c               |  15 +--
>   target/riscv/insn_trans/trans_rvv.c.inc |  46 ++++----
>   target/riscv/op_helper.c                |   4 +-
>   target/riscv/pmp.c                      |  19 ++--
>   target/riscv/pmp.h                      |   9 +-
>   target/riscv/vector_helper.c            | 134 +++++++++++++-----------
>   8 files changed, 122 insertions(+), 113 deletions(-)
>
> diff --git a/target/riscv/arch_dump.c b/target/riscv/arch_dump.c
> index 736a232956..573587810e 100644
> --- a/target/riscv/arch_dump.c
> +++ b/target/riscv/arch_dump.c
> @@ -180,8 +180,8 @@ int cpu_get_dump_info(ArchDumpInfo *info,
>       info->d_class = ELFCLASS32;
>   #endif
>   
> -    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0
> -                     ? ELFDATA2MSB : ELFDATA2LSB;
> +    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0 ?
> +                     ELFDATA2MSB : ELFDATA2LSB;
>   
>       return 0;
>   }
> diff --git a/target/riscv/cpu.c b/target/riscv/cpu.c
> index 16e465a0ab..75dab70ba7 100644
> --- a/target/riscv/cpu.c
> +++ b/target/riscv/cpu.c
> @@ -344,8 +344,8 @@ static void riscv_any_cpu_init(Object *obj)
>   
>   #ifndef CONFIG_USER_ONLY
>       set_satp_mode_max_supported(RISCV_CPU(obj),
> -            riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
> -                                    VM_1_10_SV32 : VM_1_10_SV57);
> +        riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
> +        VM_1_10_SV32 : VM_1_10_SV57);
>   #endif
>   
>       set_priv_version(env, PRIV_VERSION_1_12_0);
> diff --git a/target/riscv/cpu_helper.c b/target/riscv/cpu_helper.c
> index de2d4a8c1d..08689ee3f6 100644
> --- a/target/riscv/cpu_helper.c
> +++ b/target/riscv/cpu_helper.c
> @@ -68,12 +68,12 @@ void cpu_get_tb_cpu_state(CPURISCVState *env, target_ulong *pc,
>           flags = FIELD_DP32(flags, TB_FLAGS, VILL, env->vill);
>           flags = FIELD_DP32(flags, TB_FLAGS, SEW, sew);
>           flags = FIELD_DP32(flags, TB_FLAGS, LMUL,
> -                    FIELD_EX64(env->vtype, VTYPE, VLMUL));
> +                           FIELD_EX64(env->vtype, VTYPE, VLMUL));
>           flags = FIELD_DP32(flags, TB_FLAGS, VL_EQ_VLMAX, vl_eq_vlmax);
>           flags = FIELD_DP32(flags, TB_FLAGS, VTA,
> -                    FIELD_EX64(env->vtype, VTYPE, VTA));
> +                           FIELD_EX64(env->vtype, VTYPE, VTA));
>           flags = FIELD_DP32(flags, TB_FLAGS, VMA,
> -                    FIELD_EX64(env->vtype, VTYPE, VMA));
> +                           FIELD_EX64(env->vtype, VTYPE, VMA));
>       } else {
>           flags = FIELD_DP32(flags, TB_FLAGS, VILL, 1);
>       }
> @@ -1056,7 +1056,7 @@ restart:
>               /* add write permission on stores or if the page is already dirty,
>                  so that we TLB miss on later writes to update the dirty bit */
>               if ((pte & PTE_W) &&
> -                    (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
> +                (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
>                   *prot |= PAGE_WRITE;
>               }
>               return TRANSLATE_SUCCESS;
> @@ -1285,9 +1285,10 @@ bool riscv_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
>                                          false);
>   
>               qemu_log_mask(CPU_LOG_MMU,
> -                    "%s 2nd-stage address=%" VADDR_PRIx " ret %d physical "
> -                    HWADDR_FMT_plx " prot %d\n",
> -                    __func__, im_address, ret, pa, prot2);
> +                          "%s 2nd-stage address=%" VADDR_PRIx
> +                          " ret %d physical "
> +                          HWADDR_FMT_plx " prot %d\n",
> +                          __func__, im_address, ret, pa, prot2);
>   
>               prot &= prot2;
>   
> diff --git a/target/riscv/insn_trans/trans_rvv.c.inc b/target/riscv/insn_trans/trans_rvv.c.inc
> index f2e3d38515..2aed66934a 100644
> --- a/target/riscv/insn_trans/trans_rvv.c.inc
> +++ b/target/riscv/insn_trans/trans_rvv.c.inc
> @@ -238,8 +238,8 @@ static bool vext_check_store(DisasContext *s, int vd, int nf, uint8_t eew)
>   {
>       int8_t emul = eew - s->sew + s->lmul;
>       return (emul >= -3 && emul <= 3) &&
> -            require_align(vd, emul) &&
> -            require_nf(vd, nf, emul);
> +           require_align(vd, emul) &&
> +           require_nf(vd, nf, emul);
>   }
>   
>   /*
> @@ -315,7 +315,7 @@ static bool vext_check_ld_index(DisasContext *s, int vd, int vs2,
>       int8_t seg_vd;
>       int8_t emul = eew - s->sew + s->lmul;
>       bool ret = vext_check_st_index(s, vd, vs2, nf, eew) &&
> -        require_vm(vm, vd);
> +               require_vm(vm, vd);
>   
>       /* Each segment register group has to follow overlap rules. */
>       for (int i = 0; i < nf; ++i) {
> @@ -345,8 +345,8 @@ static bool vext_check_ld_index(DisasContext *s, int vd, int vs2,
>   static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
>   {
>       return require_vm(vm, vd) &&
> -        require_align(vd, s->lmul) &&
> -        require_align(vs, s->lmul);
> +           require_align(vd, s->lmul) &&
> +           require_align(vs, s->lmul);
>   }
>   
>   /*
> @@ -365,7 +365,7 @@ static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
>   static bool vext_check_sss(DisasContext *s, int vd, int vs1, int vs2, int vm)
>   {
>       return vext_check_ss(s, vd, vs2, vm) &&
> -        require_align(vs1, s->lmul);
> +           require_align(vs1, s->lmul);
>   }
>   
>   static bool vext_check_ms(DisasContext *s, int vd, int vs)
> @@ -396,7 +396,7 @@ static bool vext_check_ms(DisasContext *s, int vd, int vs)
>   static bool vext_check_mss(DisasContext *s, int vd, int vs1, int vs2)
>   {
>       bool ret = vext_check_ms(s, vd, vs2) &&
> -        require_align(vs1, s->lmul);
> +               require_align(vs1, s->lmul);
>       if (vd != vs1) {
>           ret &= require_noover(vd, 0, vs1, s->lmul);
>       }
> @@ -460,14 +460,14 @@ static bool vext_narrow_check_common(DisasContext *s, int vd, int vs2,
>   static bool vext_check_ds(DisasContext *s, int vd, int vs, int vm)
>   {
>       return vext_wide_check_common(s, vd, vm) &&
> -        require_align(vs, s->lmul) &&
> -        require_noover(vd, s->lmul + 1, vs, s->lmul);
> +           require_align(vs, s->lmul) &&
> +           require_noover(vd, s->lmul + 1, vs, s->lmul);
>   }
>   
>   static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
>   {
>       return vext_wide_check_common(s, vd, vm) &&
> -        require_align(vs, s->lmul + 1);
> +           require_align(vs, s->lmul + 1);
>   }
>   
>   /*
> @@ -485,8 +485,8 @@ static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
>   static bool vext_check_dss(DisasContext *s, int vd, int vs1, int vs2, int vm)
>   {
>       return vext_check_ds(s, vd, vs2, vm) &&
> -        require_align(vs1, s->lmul) &&
> -        require_noover(vd, s->lmul + 1, vs1, s->lmul);
> +           require_align(vs1, s->lmul) &&
> +           require_noover(vd, s->lmul + 1, vs1, s->lmul);
>   }
>   
>   /*
> @@ -507,7 +507,7 @@ static bool vext_check_dss(DisasContext *s, int vd, int vs1, int vs2, int vm)
>   static bool vext_check_dds(DisasContext *s, int vd, int vs1, int vs2, int vm)
>   {
>       return vext_check_ds(s, vd, vs1, vm) &&
> -        require_align(vs2, s->lmul + 1);
> +           require_align(vs2, s->lmul + 1);
>   }
>   
>   static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
> @@ -535,7 +535,7 @@ static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
>   static bool vext_check_sds(DisasContext *s, int vd, int vs1, int vs2, int vm)
>   {
>       return vext_check_sd(s, vd, vs2, vm) &&
> -        require_align(vs1, s->lmul);
> +           require_align(vs1, s->lmul);
>   }
>   
>   /*
> @@ -1216,7 +1216,7 @@ do_opivv_gvec(DisasContext *s, arg_rmrr *a, GVecGen3Fn *gvec_fn,
>   }
>   
>   /* OPIVV with GVEC IR */
> -#define GEN_OPIVV_GVEC_TRANS(NAME, SUF) \
> +#define GEN_OPIVV_GVEC_TRANS(NAME, SUF)                            \

At least it is not bad to use a special indentation for the macro name.

Zhiwei

>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>   {                                                                  \
>       static gen_helper_gvec_4_ptr * const fns[4] = {                \
> @@ -1300,7 +1300,7 @@ do_opivx_gvec(DisasContext *s, arg_rmrr *a, GVecGen2sFn *gvec_fn,
>   }
>   
>   /* OPIVX with GVEC IR */
> -#define GEN_OPIVX_GVEC_TRANS(NAME, SUF) \
> +#define GEN_OPIVX_GVEC_TRANS(NAME, SUF)                            \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>   {                                                                  \
>       static gen_helper_opivx * const fns[4] = {                     \
> @@ -1453,7 +1453,7 @@ do_opivi_gvec(DisasContext *s, arg_rmrr *a, GVecGen2iFn *gvec_fn,
>   }
>   
>   /* OPIVI with GVEC IR */
> -#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF) \
> +#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF)           \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>   {                                                                  \
>       static gen_helper_opivx * const fns[4] = {                     \
> @@ -1512,7 +1512,7 @@ static bool do_opivv_widen(DisasContext *s, arg_rmrr *a,
>       return false;
>   }
>   
> -#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK) \
> +#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK)                   \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>   {                                                            \
>       static gen_helper_gvec_4_ptr * const fns[3] = {          \
> @@ -1545,7 +1545,7 @@ static bool do_opivx_widen(DisasContext *s, arg_rmrr *a,
>       return false;
>   }
>   
> -#define GEN_OPIVX_WIDEN_TRANS(NAME) \
> +#define GEN_OPIVX_WIDEN_TRANS(NAME)                          \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>   {                                                            \
>       static gen_helper_opivx * const fns[3] = {               \
> @@ -1594,7 +1594,7 @@ static bool do_opiwv_widen(DisasContext *s, arg_rmrr *a,
>       return false;
>   }
>   
> -#define GEN_OPIWV_WIDEN_TRANS(NAME) \
> +#define GEN_OPIWV_WIDEN_TRANS(NAME)                          \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>   {                                                            \
>       static gen_helper_gvec_4_ptr * const fns[3] = {          \
> @@ -1627,7 +1627,7 @@ static bool do_opiwx_widen(DisasContext *s, arg_rmrr *a,
>       return false;
>   }
>   
> -#define GEN_OPIWX_WIDEN_TRANS(NAME) \
> +#define GEN_OPIWX_WIDEN_TRANS(NAME)                          \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>   {                                                            \
>       static gen_helper_opivx * const fns[3] = {               \
> @@ -1801,7 +1801,7 @@ do_opivx_gvec_shift(DisasContext *s, arg_rmrr *a, GVecGen2sFn32 *gvec_fn,
>       return opivx_trans(a->rd, a->rs1, a->rs2, a->vm, fn, s);
>   }
>   
> -#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, SUF) \
> +#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, SUF)                             \
>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)                    \
>   {                                                                         \
>       static gen_helper_opivx * const fns[4] = {                            \
> @@ -3668,7 +3668,7 @@ static bool trans_vcompress_vm(DisasContext *s, arg_r *a)
>    * Whole Vector Register Move Instructions ignore vtype and vl setting.
>    * Thus, we don't need to check vill bit. (Section 16.6)
>    */
> -#define GEN_VMV_WHOLE_TRANS(NAME, LEN)                             \
> +#define GEN_VMV_WHOLE_TRANS(NAME, LEN)                                  \
>   static bool trans_##NAME(DisasContext *s, arg_##NAME * a)               \
>   {                                                                       \
>       if (require_rvv(s) &&                                               \
> diff --git a/target/riscv/op_helper.c b/target/riscv/op_helper.c
> index 1eecae9547..478365131d 100644
> --- a/target/riscv/op_helper.c
> +++ b/target/riscv/op_helper.c
> @@ -367,8 +367,8 @@ void helper_wfi(CPURISCVState *env)
>       if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, MSTATUS_TW)) ||
>           (rvs && prv_u && !riscv_cpu_virt_enabled(env))) {
>           riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
> -    } else if (riscv_cpu_virt_enabled(env) && (prv_u ||
> -        (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
> +    } else if (riscv_cpu_virt_enabled(env) &&
> +               (prv_u || (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
>           riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
>       } else {
>           cs->halted = 1;
> diff --git a/target/riscv/pmp.c b/target/riscv/pmp.c
> index a08cd95658..3943b0f2e3 100644
> --- a/target/riscv/pmp.c
> +++ b/target/riscv/pmp.c
> @@ -27,7 +27,7 @@
>   #include "exec/exec-all.h"
>   
>   static void pmp_write_cfg(CPURISCVState *env, uint32_t addr_index,
> -    uint8_t val);
> +                          uint8_t val);
>   static uint8_t pmp_read_cfg(CPURISCVState *env, uint32_t addr_index);
>   static void pmp_update_rule(CPURISCVState *env, uint32_t pmp_index);
>   
> @@ -220,8 +220,8 @@ static int pmp_is_in_range(CPURISCVState *env, int pmp_index, target_ulong addr)
>   {
>       int result = 0;
>   
> -    if ((addr >= env->pmp_state.addr[pmp_index].sa)
> -        && (addr <= env->pmp_state.addr[pmp_index].ea)) {
> +    if ((addr >= env->pmp_state.addr[pmp_index].sa) &&
> +        (addr <= env->pmp_state.addr[pmp_index].ea)) {
>           result = 1;
>       } else {
>           result = 0;
> @@ -234,8 +234,9 @@ static int pmp_is_in_range(CPURISCVState *env, int pmp_index, target_ulong addr)
>    * Check if the address has required RWX privs when no PMP entry is matched.
>    */
>   static bool pmp_hart_has_privs_default(CPURISCVState *env, target_ulong addr,
> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
> -    target_ulong mode)
> +                                       target_ulong size, pmp_priv_t privs,
> +                                       pmp_priv_t *allowed_privs,
> +                                       target_ulong mode)
>   {
>       bool ret;
>   
> @@ -297,8 +298,8 @@ static bool pmp_hart_has_privs_default(CPURISCVState *env, target_ulong addr,
>    * Return negtive value if no match
>    */
>   int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
> -    target_ulong mode)
> +                       target_ulong size, pmp_priv_t privs,
> +                       pmp_priv_t *allowed_privs, target_ulong mode)
>   {
>       int i = 0;
>       int ret = -1;
> @@ -466,7 +467,7 @@ int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
>    * Handle a write to a pmpcfg CSR
>    */
>   void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
> -    target_ulong val)
> +                      target_ulong val)
>   {
>       int i;
>       uint8_t cfg_val;
> @@ -508,7 +509,7 @@ target_ulong pmpcfg_csr_read(CPURISCVState *env, uint32_t reg_index)
>    * Handle a write to a pmpaddr CSR
>    */
>   void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
> -    target_ulong val)
> +                       target_ulong val)
>   {
>       trace_pmpaddr_csr_write(env->mhartid, addr_index, val);
>   
> diff --git a/target/riscv/pmp.h b/target/riscv/pmp.h
> index da32c61c85..b296ea1fc6 100644
> --- a/target/riscv/pmp.h
> +++ b/target/riscv/pmp.h
> @@ -63,18 +63,19 @@ typedef struct {
>   } pmp_table_t;
>   
>   void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
> -    target_ulong val);
> +                      target_ulong val);
>   target_ulong pmpcfg_csr_read(CPURISCVState *env, uint32_t reg_index);
>   
>   void mseccfg_csr_write(CPURISCVState *env, target_ulong val);
>   target_ulong mseccfg_csr_read(CPURISCVState *env);
>   
>   void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
> -    target_ulong val);
> +                       target_ulong val);
>   target_ulong pmpaddr_csr_read(CPURISCVState *env, uint32_t addr_index);
>   int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
> -    target_ulong mode);
> +                       target_ulong size, pmp_priv_t privs,
> +                       pmp_priv_t *allowed_privs,
> +                       target_ulong mode);
>   target_ulong pmp_get_tlb_size(CPURISCVState *env, int pmp_index,
>                                 target_ulong tlb_sa, target_ulong tlb_ea);
>   void pmp_update_rule_addr(CPURISCVState *env, uint32_t pmp_index);
> diff --git a/target/riscv/vector_helper.c b/target/riscv/vector_helper.c
> index 2423affe37..df6201d043 100644
> --- a/target/riscv/vector_helper.c
> +++ b/target/riscv/vector_helper.c
> @@ -1116,7 +1116,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2,        \
>                                                                            \
>           *((ETYPE *)vd + H(i)) = DO_OP(s2, (ETYPE)(target_long)s1, carry);\
>       }                                                                    \
> -    env->vstart = 0;                                          \
> +    env->vstart = 0;                                                     \
>       /* set tail elements to 1s */                                        \
>       vext_set_elems_1s(vd, vta, vl * esz, total_elems * esz);             \
>   }
> @@ -1308,7 +1308,8 @@ GEN_VEXT_SHIFT_VV(vsra_vv_d, uint64_t, int64_t, H8, H8, DO_SRL, 0x3f)
>   /* generate the helpers for shift instructions with one vector and one scalar */
>   #define GEN_VEXT_SHIFT_VX(NAME, TD, TS2, HD, HS2, OP, MASK) \
>   void HELPER(NAME)(void *vd, void *v0, target_ulong s1,      \
> -        void *vs2, CPURISCVState *env, uint32_t desc)       \
> +                  void *vs2, CPURISCVState *env,            \
> +                  uint32_t desc)                            \
>   {                                                           \
>       uint32_t vm = vext_vm(desc);                            \
>       uint32_t vl = env->vl;                                  \
> @@ -1735,9 +1736,9 @@ GEN_VEXT_VX(vmulhsu_vx_d, 8)
>   /* Vector Integer Divide Instructions */
>   #define DO_DIVU(N, M) (unlikely(M == 0) ? (__typeof(N))(-1) : N / M)
>   #define DO_REMU(N, M) (unlikely(M == 0) ? N : N % M)
> -#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) :\
> +#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) : \
>           unlikely((N == -N) && (M == (__typeof(N))(-1))) ? N : N / M)
> -#define DO_REM(N, M)  (unlikely(M == 0) ? N :\
> +#define DO_REM(N, M)  (unlikely(M == 0) ? N : \
>           unlikely((N == -N) && (M == (__typeof(N))(-1))) ? 0 : N % M)
>   
>   RVVCALL(OPIVV2, vdivu_vv_b, OP_UUU_B, H1, H1, H1, DO_DIVU)
> @@ -1846,7 +1847,7 @@ GEN_VEXT_VX(vwmulsu_vx_h, 4)
>   GEN_VEXT_VX(vwmulsu_vx_w, 8)
>   
>   /* Vector Single-Width Integer Multiply-Add Instructions */
> -#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)   \
> +#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
>   static void do_##NAME(void *vd, void *vs1, void *vs2, int i)       \
>   {                                                                  \
>       TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
> @@ -2277,7 +2278,8 @@ vext_vx_rm_2(void *vd, void *v0, target_long s1, void *vs2,
>   /* generate helpers for fixed point instructions with OPIVX format */
>   #define GEN_VEXT_VX_RM(NAME, ESZ)                         \
>   void HELPER(NAME)(void *vd, void *v0, target_ulong s1,    \
> -        void *vs2, CPURISCVState *env, uint32_t desc)     \
> +                  void *vs2, CPURISCVState *env,          \
> +                  uint32_t desc)                          \
>   {                                                         \
>       vext_vx_rm_2(vd, v0, s1, vs2, env, desc,              \
>                    do_##NAME, ESZ);                         \
> @@ -3052,7 +3054,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1,        \
>       uint32_t vm = vext_vm(desc);                          \
>       uint32_t vl = env->vl;                                \
>       uint32_t total_elems =                                \
> -        vext_get_total_elems(env, desc, ESZ);              \
> +        vext_get_total_elems(env, desc, ESZ);             \
>       uint32_t vta = vext_vta(desc);                        \
>       uint32_t vma = vext_vma(desc);                        \
>       uint32_t i;                                           \
> @@ -3118,13 +3120,13 @@ GEN_VEXT_VF(vfrsub_vf_d, 8)
>   static uint32_t vfwadd16(uint16_t a, uint16_t b, float_status *s)
>   {
>       return float32_add(float16_to_float32(a, true, s),
> -            float16_to_float32(b, true, s), s);
> +                       float16_to_float32(b, true, s), s);
>   }
>   
>   static uint64_t vfwadd32(uint32_t a, uint32_t b, float_status *s)
>   {
>       return float64_add(float32_to_float64(a, s),
> -            float32_to_float64(b, s), s);
> +                       float32_to_float64(b, s), s);
>   
>   }
>   
> @@ -3140,13 +3142,13 @@ GEN_VEXT_VF(vfwadd_vf_w, 8)
>   static uint32_t vfwsub16(uint16_t a, uint16_t b, float_status *s)
>   {
>       return float32_sub(float16_to_float32(a, true, s),
> -            float16_to_float32(b, true, s), s);
> +                       float16_to_float32(b, true, s), s);
>   }
>   
>   static uint64_t vfwsub32(uint32_t a, uint32_t b, float_status *s)
>   {
>       return float64_sub(float32_to_float64(a, s),
> -            float32_to_float64(b, s), s);
> +                       float32_to_float64(b, s), s);
>   
>   }
>   
> @@ -3250,13 +3252,13 @@ GEN_VEXT_VF(vfrdiv_vf_d, 8)
>   static uint32_t vfwmul16(uint16_t a, uint16_t b, float_status *s)
>   {
>       return float32_mul(float16_to_float32(a, true, s),
> -            float16_to_float32(b, true, s), s);
> +                       float16_to_float32(b, true, s), s);
>   }
>   
>   static uint64_t vfwmul32(uint32_t a, uint32_t b, float_status *s)
>   {
>       return float64_mul(float32_to_float64(a, s),
> -            float32_to_float64(b, s), s);
> +                       float32_to_float64(b, s), s);
>   
>   }
>   RVVCALL(OPFVV2, vfwmul_vv_h, WOP_UUU_H, H4, H2, H2, vfwmul16)
> @@ -3271,7 +3273,7 @@ GEN_VEXT_VF(vfwmul_vf_w, 8)
>   /* Vector Single-Width Floating-Point Fused Multiply-Add Instructions */
>   #define OPFVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
>   static void do_##NAME(void *vd, void *vs1, void *vs2, int i,       \
> -        CPURISCVState *env)                                        \
> +                      CPURISCVState *env)                          \
>   {                                                                  \
>       TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
>       TX2 s2 = *((T2 *)vs2 + HS2(i));                                \
> @@ -3303,7 +3305,7 @@ GEN_VEXT_VV_ENV(vfmacc_vv_d, 8)
>   
>   #define OPFVF3(NAME, TD, T1, T2, TX1, TX2, HD, HS2, OP)           \
>   static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i,    \
> -        CPURISCVState *env)                                       \
> +                      CPURISCVState *env)                         \
>   {                                                                 \
>       TX2 s2 = *((T2 *)vs2 + HS2(i));                               \
>       TD d = *((TD *)vd + HD(i));                                   \
> @@ -3319,20 +3321,20 @@ GEN_VEXT_VF(vfmacc_vf_d, 8)
>   
>   static uint16_t fnmacc16(uint16_t a, uint16_t b, uint16_t d, float_status *s)
>   {
> -    return float16_muladd(a, b, d,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float16_muladd(a, b, d, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   static uint32_t fnmacc32(uint32_t a, uint32_t b, uint32_t d, float_status *s)
>   {
> -    return float32_muladd(a, b, d,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float32_muladd(a, b, d, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   static uint64_t fnmacc64(uint64_t a, uint64_t b, uint64_t d, float_status *s)
>   {
> -    return float64_muladd(a, b, d,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float64_muladd(a, b, d, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   RVVCALL(OPFVV3, vfnmacc_vv_h, OP_UUU_H, H2, H2, H2, fnmacc16)
> @@ -3434,20 +3436,20 @@ GEN_VEXT_VF(vfmadd_vf_d, 8)
>   
>   static uint16_t fnmadd16(uint16_t a, uint16_t b, uint16_t d, float_status *s)
>   {
> -    return float16_muladd(d, b, a,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float16_muladd(d, b, a, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   static uint32_t fnmadd32(uint32_t a, uint32_t b, uint32_t d, float_status *s)
>   {
> -    return float32_muladd(d, b, a,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float32_muladd(d, b, a, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   static uint64_t fnmadd64(uint64_t a, uint64_t b, uint64_t d, float_status *s)
>   {
> -    return float64_muladd(d, b, a,
> -            float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float64_muladd(d, b, a, float_muladd_negate_c |
> +                                   float_muladd_negate_product, s);
>   }
>   
>   RVVCALL(OPFVV3, vfnmadd_vv_h, OP_UUU_H, H2, H2, H2, fnmadd16)
> @@ -3523,13 +3525,13 @@ GEN_VEXT_VF(vfnmsub_vf_d, 8)
>   static uint32_t fwmacc16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
>   {
>       return float32_muladd(float16_to_float32(a, true, s),
> -                        float16_to_float32(b, true, s), d, 0, s);
> +                          float16_to_float32(b, true, s), d, 0, s);
>   }
>   
>   static uint64_t fwmacc32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
>   {
>       return float64_muladd(float32_to_float64(a, s),
> -                        float32_to_float64(b, s), d, 0, s);
> +                          float32_to_float64(b, s), d, 0, s);
>   }
>   
>   RVVCALL(OPFVV3, vfwmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwmacc16)
> @@ -3544,15 +3546,16 @@ GEN_VEXT_VF(vfwmacc_vf_w, 8)
>   static uint32_t fwnmacc16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
>   {
>       return float32_muladd(float16_to_float32(a, true, s),
> -                        float16_to_float32(b, true, s), d,
> -                        float_muladd_negate_c | float_muladd_negate_product, s);
> +                          float16_to_float32(b, true, s), d,
> +                          float_muladd_negate_c | float_muladd_negate_product,
> +                          s);
>   }
>   
>   static uint64_t fwnmacc32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
>   {
> -    return float64_muladd(float32_to_float64(a, s),
> -                        float32_to_float64(b, s), d,
> -                        float_muladd_negate_c | float_muladd_negate_product, s);
> +    return float64_muladd(float32_to_float64(a, s), float32_to_float64(b, s),
> +                          d, float_muladd_negate_c |
> +                             float_muladd_negate_product, s);
>   }
>   
>   RVVCALL(OPFVV3, vfwnmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwnmacc16)
> @@ -3567,15 +3570,15 @@ GEN_VEXT_VF(vfwnmacc_vf_w, 8)
>   static uint32_t fwmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
>   {
>       return float32_muladd(float16_to_float32(a, true, s),
> -                        float16_to_float32(b, true, s), d,
> -                        float_muladd_negate_c, s);
> +                          float16_to_float32(b, true, s), d,
> +                          float_muladd_negate_c, s);
>   }
>   
>   static uint64_t fwmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
>   {
>       return float64_muladd(float32_to_float64(a, s),
> -                        float32_to_float64(b, s), d,
> -                        float_muladd_negate_c, s);
> +                          float32_to_float64(b, s), d,
> +                          float_muladd_negate_c, s);
>   }
>   
>   RVVCALL(OPFVV3, vfwmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwmsac16)
> @@ -3590,15 +3593,15 @@ GEN_VEXT_VF(vfwmsac_vf_w, 8)
>   static uint32_t fwnmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)
>   {
>       return float32_muladd(float16_to_float32(a, true, s),
> -                        float16_to_float32(b, true, s), d,
> -                        float_muladd_negate_product, s);
> +                          float16_to_float32(b, true, s), d,
> +                          float_muladd_negate_product, s);
>   }
>   
>   static uint64_t fwnmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)
>   {
>       return float64_muladd(float32_to_float64(a, s),
> -                        float32_to_float64(b, s), d,
> -                        float_muladd_negate_product, s);
> +                          float32_to_float64(b, s), d,
> +                          float_muladd_negate_product, s);
>   }
>   
>   RVVCALL(OPFVV3, vfwnmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwnmsac16)
> @@ -3616,9 +3619,9 @@ GEN_VEXT_VF(vfwnmsac_vf_w, 8)
>   #define OP_UU_W uint32_t, uint32_t, uint32_t
>   #define OP_UU_D uint64_t, uint64_t, uint64_t
>   
> -#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)        \
> +#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)         \
>   static void do_##NAME(void *vd, void *vs2, int i,      \
> -        CPURISCVState *env)                            \
> +                      CPURISCVState *env)              \
>   {                                                      \
>       TX2 s2 = *((T2 *)vs2 + HS2(i));                    \
>       *((TD *)vd + HD(i)) = OP(s2, &env->fp_status);     \
> @@ -3626,7 +3629,7 @@ static void do_##NAME(void *vd, void *vs2, int i,      \
>   
>   #define GEN_VEXT_V_ENV(NAME, ESZ)                      \
>   void HELPER(NAME)(void *vd, void *v0, void *vs2,       \
> -        CPURISCVState *env, uint32_t desc)             \
> +                  CPURISCVState *env, uint32_t desc)   \
>   {                                                      \
>       uint32_t vm = vext_vm(desc);                       \
>       uint32_t vl = env->vl;                             \
> @@ -3703,9 +3706,9 @@ static uint64_t frsqrt7(uint64_t f, int exp_size, int frac_size)
>       }
>   
>       int idx = ((exp & 1) << (precision - 1)) |
> -                (frac >> (frac_size - precision + 1));
> +              (frac >> (frac_size - precision + 1));
>       uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
> -                            (frac_size - precision);
> +                        (frac_size - precision);
>       uint64_t out_exp = (3 * MAKE_64BIT_MASK(0, exp_size - 1) + ~exp) / 2;
>   
>       uint64_t val = 0;
> @@ -3727,9 +3730,9 @@ static float16 frsqrt7_h(float16 f, float_status *s)
>        * frsqrt7(-subnormal) = canonical NaN
>        */
>       if (float16_is_signaling_nan(f, s) ||
> -            (float16_is_infinity(f) && sign) ||
> -            (float16_is_normal(f) && sign) ||
> -            (float16_is_zero_or_denormal(f) && !float16_is_zero(f) && sign)) {
> +        (float16_is_infinity(f) && sign) ||
> +        (float16_is_normal(f) && sign) ||
> +        (float16_is_zero_or_denormal(f) && !float16_is_zero(f) && sign)) {
>           s->float_exception_flags |= float_flag_invalid;
>           return float16_default_nan(s);
>       }
> @@ -3767,9 +3770,9 @@ static float32 frsqrt7_s(float32 f, float_status *s)
>        * frsqrt7(-subnormal) = canonical NaN
>        */
>       if (float32_is_signaling_nan(f, s) ||
> -            (float32_is_infinity(f) && sign) ||
> -            (float32_is_normal(f) && sign) ||
> -            (float32_is_zero_or_denormal(f) && !float32_is_zero(f) && sign)) {
> +        (float32_is_infinity(f) && sign) ||
> +        (float32_is_normal(f) && sign) ||
> +        (float32_is_zero_or_denormal(f) && !float32_is_zero(f) && sign)) {
>           s->float_exception_flags |= float_flag_invalid;
>           return float32_default_nan(s);
>       }
> @@ -3807,9 +3810,9 @@ static float64 frsqrt7_d(float64 f, float_status *s)
>        * frsqrt7(-subnormal) = canonical NaN
>        */
>       if (float64_is_signaling_nan(f, s) ||
> -            (float64_is_infinity(f) && sign) ||
> -            (float64_is_normal(f) && sign) ||
> -            (float64_is_zero_or_denormal(f) && !float64_is_zero(f) && sign)) {
> +        (float64_is_infinity(f) && sign) ||
> +        (float64_is_normal(f) && sign) ||
> +        (float64_is_zero_or_denormal(f) && !float64_is_zero(f) && sign)) {
>           s->float_exception_flags |= float_flag_invalid;
>           return float64_default_nan(s);
>       }
> @@ -3897,18 +3900,18 @@ static uint64_t frec7(uint64_t f, int exp_size, int frac_size,
>                   ((s->float_rounding_mode == float_round_up) && sign)) {
>                   /* Return greatest/negative finite value. */
>                   return (sign << (exp_size + frac_size)) |
> -                    (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
> +                       (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
>               } else {
>                   /* Return +-inf. */
>                   return (sign << (exp_size + frac_size)) |
> -                    MAKE_64BIT_MASK(frac_size, exp_size);
> +                       MAKE_64BIT_MASK(frac_size, exp_size);
>               }
>           }
>       }
>   
>       int idx = frac >> (frac_size - precision);
>       uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
> -                            (frac_size - precision);
> +                        (frac_size - precision);
>       uint64_t out_exp = 2 * MAKE_64BIT_MASK(0, exp_size - 1) + ~exp;
>   
>       if (out_exp == 0 || out_exp == UINT64_MAX) {
> @@ -4422,8 +4425,8 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
>                                                                 \
>       for (i = env->vstart; i < vl; i++) {                      \
>           ETYPE s2 = *((ETYPE *)vs2 + H(i));                    \
> -        *((ETYPE *)vd + H(i))                                 \
> -          = (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
> +        *((ETYPE *)vd + H(i)) =                               \
> +            (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
>       }                                                         \
>       env->vstart = 0;                                          \
>       /* set tail elements to 1s */                             \
> @@ -4564,7 +4567,8 @@ GEN_VEXT_V_ENV(vfncvt_f_f_w_w, 4)
>   /* Vector Single-Width Integer Reduction Instructions */
>   #define GEN_VEXT_RED(NAME, TD, TS2, HD, HS2, OP)          \
>   void HELPER(NAME)(void *vd, void *v0, void *vs1,          \
> -        void *vs2, CPURISCVState *env, uint32_t desc)     \
> +                  void *vs2, CPURISCVState *env,          \
> +                  uint32_t desc)                          \
>   {                                                         \
>       uint32_t vm = vext_vm(desc);                          \
>       uint32_t vl = env->vl;                                \
> @@ -5013,7 +5017,8 @@ GEN_VEXT_VSLIDEDOWN_VX(vslidedown_vx_d, uint64_t, H8)
>   
>   #define GEN_VEXT_VSLIE1UP(BITWIDTH, H)                                      \
>   static void vslide1up_##BITWIDTH(void *vd, void *v0, uint64_t s1,           \
> -                     void *vs2, CPURISCVState *env, uint32_t desc)          \
> +                                 void *vs2, CPURISCVState *env,             \
> +                                 uint32_t desc)                             \
>   {                                                                           \
>       typedef uint##BITWIDTH##_t ETYPE;                                       \
>       uint32_t vm = vext_vm(desc);                                            \
> @@ -5061,7 +5066,8 @@ GEN_VEXT_VSLIDE1UP_VX(vslide1up_vx_d, 64)
>   
>   #define GEN_VEXT_VSLIDE1DOWN(BITWIDTH, H)                                     \
>   static void vslide1down_##BITWIDTH(void *vd, void *v0, uint64_t s1,           \
> -                       void *vs2, CPURISCVState *env, uint32_t desc)          \
> +                                   void *vs2, CPURISCVState *env,             \
> +                                   uint32_t desc)                             \
>   {                                                                             \
>       typedef uint##BITWIDTH##_t ETYPE;                                         \
>       uint32_t vm = vext_vm(desc);                                              \
Re: [PATCH 6/8] target/riscv: Fix format for indentation
Posted by liweiwei 2 years, 10 months ago
On 2023/3/25 22:23, LIU Zhiwei wrote:
>
> On 2023/3/24 20:38, Weiwei Li wrote:
>> Fix identation problems, and try to use the same indentation strategy
>> in the same file.
>>
>> Signed-off-by: Weiwei Li <liweiwei@iscas.ac.cn>
>> Signed-off-by: Junqiang Wang <wangjunqiang@iscas.ac.cn>
>> ---
>>   target/riscv/arch_dump.c                |   4 +-
>>   target/riscv/cpu.c                      |   4 +-
>>   target/riscv/cpu_helper.c               |  15 +--
>>   target/riscv/insn_trans/trans_rvv.c.inc |  46 ++++----
>>   target/riscv/op_helper.c                |   4 +-
>>   target/riscv/pmp.c                      |  19 ++--
>>   target/riscv/pmp.h                      |   9 +-
>>   target/riscv/vector_helper.c            | 134 +++++++++++++-----------
>>   8 files changed, 122 insertions(+), 113 deletions(-)
>>
>> diff --git a/target/riscv/arch_dump.c b/target/riscv/arch_dump.c
>> index 736a232956..573587810e 100644
>> --- a/target/riscv/arch_dump.c
>> +++ b/target/riscv/arch_dump.c
>> @@ -180,8 +180,8 @@ int cpu_get_dump_info(ArchDumpInfo *info,
>>       info->d_class = ELFCLASS32;
>>   #endif
>>   -    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0
>> -                     ? ELFDATA2MSB : ELFDATA2LSB;
>> +    info->d_endian = (env->mstatus & MSTATUS_UBE) != 0 ?
>> +                     ELFDATA2MSB : ELFDATA2LSB;
>>         return 0;
>>   }
>> diff --git a/target/riscv/cpu.c b/target/riscv/cpu.c
>> index 16e465a0ab..75dab70ba7 100644
>> --- a/target/riscv/cpu.c
>> +++ b/target/riscv/cpu.c
>> @@ -344,8 +344,8 @@ static void riscv_any_cpu_init(Object *obj)
>>     #ifndef CONFIG_USER_ONLY
>>       set_satp_mode_max_supported(RISCV_CPU(obj),
>> -            riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
>> -                                    VM_1_10_SV32 : VM_1_10_SV57);
>> +        riscv_cpu_mxl(&RISCV_CPU(obj)->env) == MXL_RV32 ?
>> +        VM_1_10_SV32 : VM_1_10_SV57);
>>   #endif
>>         set_priv_version(env, PRIV_VERSION_1_12_0);
>> diff --git a/target/riscv/cpu_helper.c b/target/riscv/cpu_helper.c
>> index de2d4a8c1d..08689ee3f6 100644
>> --- a/target/riscv/cpu_helper.c
>> +++ b/target/riscv/cpu_helper.c
>> @@ -68,12 +68,12 @@ void cpu_get_tb_cpu_state(CPURISCVState *env, 
>> target_ulong *pc,
>>           flags = FIELD_DP32(flags, TB_FLAGS, VILL, env->vill);
>>           flags = FIELD_DP32(flags, TB_FLAGS, SEW, sew);
>>           flags = FIELD_DP32(flags, TB_FLAGS, LMUL,
>> -                    FIELD_EX64(env->vtype, VTYPE, VLMUL));
>> +                           FIELD_EX64(env->vtype, VTYPE, VLMUL));
>>           flags = FIELD_DP32(flags, TB_FLAGS, VL_EQ_VLMAX, vl_eq_vlmax);
>>           flags = FIELD_DP32(flags, TB_FLAGS, VTA,
>> -                    FIELD_EX64(env->vtype, VTYPE, VTA));
>> +                           FIELD_EX64(env->vtype, VTYPE, VTA));
>>           flags = FIELD_DP32(flags, TB_FLAGS, VMA,
>> -                    FIELD_EX64(env->vtype, VTYPE, VMA));
>> +                           FIELD_EX64(env->vtype, VTYPE, VMA));
>>       } else {
>>           flags = FIELD_DP32(flags, TB_FLAGS, VILL, 1);
>>       }
>> @@ -1056,7 +1056,7 @@ restart:
>>               /* add write permission on stores or if the page is 
>> already dirty,
>>                  so that we TLB miss on later writes to update the 
>> dirty bit */
>>               if ((pte & PTE_W) &&
>> -                    (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
>> +                (access_type == MMU_DATA_STORE || (pte & PTE_D))) {
>>                   *prot |= PAGE_WRITE;
>>               }
>>               return TRANSLATE_SUCCESS;
>> @@ -1285,9 +1285,10 @@ bool riscv_cpu_tlb_fill(CPUState *cs, vaddr 
>> address, int size,
>>                                          false);
>>                 qemu_log_mask(CPU_LOG_MMU,
>> -                    "%s 2nd-stage address=%" VADDR_PRIx " ret %d 
>> physical "
>> -                    HWADDR_FMT_plx " prot %d\n",
>> -                    __func__, im_address, ret, pa, prot2);
>> +                          "%s 2nd-stage address=%" VADDR_PRIx
>> +                          " ret %d physical "
>> +                          HWADDR_FMT_plx " prot %d\n",
>> +                          __func__, im_address, ret, pa, prot2);
>>                 prot &= prot2;
>>   diff --git a/target/riscv/insn_trans/trans_rvv.c.inc 
>> b/target/riscv/insn_trans/trans_rvv.c.inc
>> index f2e3d38515..2aed66934a 100644
>> --- a/target/riscv/insn_trans/trans_rvv.c.inc
>> +++ b/target/riscv/insn_trans/trans_rvv.c.inc
>> @@ -238,8 +238,8 @@ static bool vext_check_store(DisasContext *s, int 
>> vd, int nf, uint8_t eew)
>>   {
>>       int8_t emul = eew - s->sew + s->lmul;
>>       return (emul >= -3 && emul <= 3) &&
>> -            require_align(vd, emul) &&
>> -            require_nf(vd, nf, emul);
>> +           require_align(vd, emul) &&
>> +           require_nf(vd, nf, emul);
>>   }
>>     /*
>> @@ -315,7 +315,7 @@ static bool vext_check_ld_index(DisasContext *s, 
>> int vd, int vs2,
>>       int8_t seg_vd;
>>       int8_t emul = eew - s->sew + s->lmul;
>>       bool ret = vext_check_st_index(s, vd, vs2, nf, eew) &&
>> -        require_vm(vm, vd);
>> +               require_vm(vm, vd);
>>         /* Each segment register group has to follow overlap rules. */
>>       for (int i = 0; i < nf; ++i) {
>> @@ -345,8 +345,8 @@ static bool vext_check_ld_index(DisasContext *s, 
>> int vd, int vs2,
>>   static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
>>   {
>>       return require_vm(vm, vd) &&
>> -        require_align(vd, s->lmul) &&
>> -        require_align(vs, s->lmul);
>> +           require_align(vd, s->lmul) &&
>> +           require_align(vs, s->lmul);
>>   }
>>     /*
>> @@ -365,7 +365,7 @@ static bool vext_check_ss(DisasContext *s, int 
>> vd, int vs, int vm)
>>   static bool vext_check_sss(DisasContext *s, int vd, int vs1, int 
>> vs2, int vm)
>>   {
>>       return vext_check_ss(s, vd, vs2, vm) &&
>> -        require_align(vs1, s->lmul);
>> +           require_align(vs1, s->lmul);
>>   }
>>     static bool vext_check_ms(DisasContext *s, int vd, int vs)
>> @@ -396,7 +396,7 @@ static bool vext_check_ms(DisasContext *s, int 
>> vd, int vs)
>>   static bool vext_check_mss(DisasContext *s, int vd, int vs1, int vs2)
>>   {
>>       bool ret = vext_check_ms(s, vd, vs2) &&
>> -        require_align(vs1, s->lmul);
>> +               require_align(vs1, s->lmul);
>>       if (vd != vs1) {
>>           ret &= require_noover(vd, 0, vs1, s->lmul);
>>       }
>> @@ -460,14 +460,14 @@ static bool 
>> vext_narrow_check_common(DisasContext *s, int vd, int vs2,
>>   static bool vext_check_ds(DisasContext *s, int vd, int vs, int vm)
>>   {
>>       return vext_wide_check_common(s, vd, vm) &&
>> -        require_align(vs, s->lmul) &&
>> -        require_noover(vd, s->lmul + 1, vs, s->lmul);
>> +           require_align(vs, s->lmul) &&
>> +           require_noover(vd, s->lmul + 1, vs, s->lmul);
>>   }
>>     static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
>>   {
>>       return vext_wide_check_common(s, vd, vm) &&
>> -        require_align(vs, s->lmul + 1);
>> +           require_align(vs, s->lmul + 1);
>>   }
>>     /*
>> @@ -485,8 +485,8 @@ static bool vext_check_dd(DisasContext *s, int 
>> vd, int vs, int vm)
>>   static bool vext_check_dss(DisasContext *s, int vd, int vs1, int 
>> vs2, int vm)
>>   {
>>       return vext_check_ds(s, vd, vs2, vm) &&
>> -        require_align(vs1, s->lmul) &&
>> -        require_noover(vd, s->lmul + 1, vs1, s->lmul);
>> +           require_align(vs1, s->lmul) &&
>> +           require_noover(vd, s->lmul + 1, vs1, s->lmul);
>>   }
>>     /*
>> @@ -507,7 +507,7 @@ static bool vext_check_dss(DisasContext *s, int 
>> vd, int vs1, int vs2, int vm)
>>   static bool vext_check_dds(DisasContext *s, int vd, int vs1, int 
>> vs2, int vm)
>>   {
>>       return vext_check_ds(s, vd, vs1, vm) &&
>> -        require_align(vs2, s->lmul + 1);
>> +           require_align(vs2, s->lmul + 1);
>>   }
>>     static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
>> @@ -535,7 +535,7 @@ static bool vext_check_sd(DisasContext *s, int 
>> vd, int vs, int vm)
>>   static bool vext_check_sds(DisasContext *s, int vd, int vs1, int 
>> vs2, int vm)
>>   {
>>       return vext_check_sd(s, vd, vs2, vm) &&
>> -        require_align(vs1, s->lmul);
>> +           require_align(vs1, s->lmul);
>>   }
>>     /*
>> @@ -1216,7 +1216,7 @@ do_opivv_gvec(DisasContext *s, arg_rmrr *a, 
>> GVecGen3Fn *gvec_fn,
>>   }
>>     /* OPIVV with GVEC IR */
>> -#define GEN_OPIVV_GVEC_TRANS(NAME, SUF) \
>> +#define GEN_OPIVV_GVEC_TRANS(NAME, SUF)                            \
>
> At least it is not bad to use a special indentation for the macro name.
>
> Zhiwei

Yeah, It's acceptable to me. I modified them since most of the macro 
names in this file use the aligned indentation.

By the way, the alignment seems changed in your reply mail.

Regards,

Weiwei Li

>
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>> { \
>>       static gen_helper_gvec_4_ptr * const fns[4] = {                \
>> @@ -1300,7 +1300,7 @@ do_opivx_gvec(DisasContext *s, arg_rmrr *a, 
>> GVecGen2sFn *gvec_fn,
>>   }
>>     /* OPIVX with GVEC IR */
>> -#define GEN_OPIVX_GVEC_TRANS(NAME, SUF) \
>> +#define GEN_OPIVX_GVEC_TRANS(NAME, SUF)                            \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>> { \
>>       static gen_helper_opivx * const fns[4] = {                     \
>> @@ -1453,7 +1453,7 @@ do_opivi_gvec(DisasContext *s, arg_rmrr *a, 
>> GVecGen2iFn *gvec_fn,
>>   }
>>     /* OPIVI with GVEC IR */
>> -#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF) \
>> +#define GEN_OPIVI_GVEC_TRANS(NAME, IMM_MODE, OPIVX, SUF)           \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \
>> { \
>>       static gen_helper_opivx * const fns[4] = {                     \
>> @@ -1512,7 +1512,7 @@ static bool do_opivv_widen(DisasContext *s, 
>> arg_rmrr *a,
>>       return false;
>>   }
>>   -#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK) \
>> +#define GEN_OPIVV_WIDEN_TRANS(NAME, CHECK)                   \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>>   {                                                            \
>>       static gen_helper_gvec_4_ptr * const fns[3] = {          \
>> @@ -1545,7 +1545,7 @@ static bool do_opivx_widen(DisasContext *s, 
>> arg_rmrr *a,
>>       return false;
>>   }
>>   -#define GEN_OPIVX_WIDEN_TRANS(NAME) \
>> +#define GEN_OPIVX_WIDEN_TRANS(NAME)                          \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>>   {                                                            \
>>       static gen_helper_opivx * const fns[3] = {               \
>> @@ -1594,7 +1594,7 @@ static bool do_opiwv_widen(DisasContext *s, 
>> arg_rmrr *a,
>>       return false;
>>   }
>>   -#define GEN_OPIWV_WIDEN_TRANS(NAME) \
>> +#define GEN_OPIWV_WIDEN_TRANS(NAME)                          \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>>   {                                                            \
>>       static gen_helper_gvec_4_ptr * const fns[3] = {          \
>> @@ -1627,7 +1627,7 @@ static bool do_opiwx_widen(DisasContext *s, 
>> arg_rmrr *a,
>>       return false;
>>   }
>>   -#define GEN_OPIWX_WIDEN_TRANS(NAME) \
>> +#define GEN_OPIWX_WIDEN_TRANS(NAME)                          \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr *a)       \
>>   {                                                            \
>>       static gen_helper_opivx * const fns[3] = {               \
>> @@ -1801,7 +1801,7 @@ do_opivx_gvec_shift(DisasContext *s, arg_rmrr 
>> *a, GVecGen2sFn32 *gvec_fn,
>>       return opivx_trans(a->rd, a->rs1, a->rs2, a->vm, fn, s);
>>   }
>>   -#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, SUF) \
>> +#define GEN_OPIVX_GVEC_SHIFT_TRANS(NAME, 
>> SUF)                             \
>>   static bool trans_##NAME(DisasContext *s, arg_rmrr 
>> *a)                    \
>> { \
>>       static gen_helper_opivx * const fns[4] = 
>> {                            \
>> @@ -3668,7 +3668,7 @@ static bool trans_vcompress_vm(DisasContext *s, 
>> arg_r *a)
>>    * Whole Vector Register Move Instructions ignore vtype and vl 
>> setting.
>>    * Thus, we don't need to check vill bit. (Section 16.6)
>>    */
>> -#define GEN_VMV_WHOLE_TRANS(NAME, LEN)                             \
>> +#define GEN_VMV_WHOLE_TRANS(NAME, 
>> LEN)                                  \
>>   static bool trans_##NAME(DisasContext *s, arg_##NAME * 
>> a)               \
>> { \
>>       if (require_rvv(s) 
>> &&                                               \
>> diff --git a/target/riscv/op_helper.c b/target/riscv/op_helper.c
>> index 1eecae9547..478365131d 100644
>> --- a/target/riscv/op_helper.c
>> +++ b/target/riscv/op_helper.c
>> @@ -367,8 +367,8 @@ void helper_wfi(CPURISCVState *env)
>>       if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, 
>> MSTATUS_TW)) ||
>>           (rvs && prv_u && !riscv_cpu_virt_enabled(env))) {
>>           riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
>> -    } else if (riscv_cpu_virt_enabled(env) && (prv_u ||
>> -        (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
>> +    } else if (riscv_cpu_virt_enabled(env) &&
>> +               (prv_u || (prv_s && get_field(env->hstatus, 
>> HSTATUS_VTW)))) {
>>           riscv_raise_exception(env, 
>> RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
>>       } else {
>>           cs->halted = 1;
>> diff --git a/target/riscv/pmp.c b/target/riscv/pmp.c
>> index a08cd95658..3943b0f2e3 100644
>> --- a/target/riscv/pmp.c
>> +++ b/target/riscv/pmp.c
>> @@ -27,7 +27,7 @@
>>   #include "exec/exec-all.h"
>>     static void pmp_write_cfg(CPURISCVState *env, uint32_t addr_index,
>> -    uint8_t val);
>> +                          uint8_t val);
>>   static uint8_t pmp_read_cfg(CPURISCVState *env, uint32_t addr_index);
>>   static void pmp_update_rule(CPURISCVState *env, uint32_t pmp_index);
>>   @@ -220,8 +220,8 @@ static int pmp_is_in_range(CPURISCVState *env, 
>> int pmp_index, target_ulong addr)
>>   {
>>       int result = 0;
>>   -    if ((addr >= env->pmp_state.addr[pmp_index].sa)
>> -        && (addr <= env->pmp_state.addr[pmp_index].ea)) {
>> +    if ((addr >= env->pmp_state.addr[pmp_index].sa) &&
>> +        (addr <= env->pmp_state.addr[pmp_index].ea)) {
>>           result = 1;
>>       } else {
>>           result = 0;
>> @@ -234,8 +234,9 @@ static int pmp_is_in_range(CPURISCVState *env, 
>> int pmp_index, target_ulong addr)
>>    * Check if the address has required RWX privs when no PMP entry is 
>> matched.
>>    */
>>   static bool pmp_hart_has_privs_default(CPURISCVState *env, 
>> target_ulong addr,
>> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
>> -    target_ulong mode)
>> +                                       target_ulong size, pmp_priv_t 
>> privs,
>> +                                       pmp_priv_t *allowed_privs,
>> +                                       target_ulong mode)
>>   {
>>       bool ret;
>>   @@ -297,8 +298,8 @@ static bool 
>> pmp_hart_has_privs_default(CPURISCVState *env, target_ulong addr,
>>    * Return negtive value if no match
>>    */
>>   int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
>> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
>> -    target_ulong mode)
>> +                       target_ulong size, pmp_priv_t privs,
>> +                       pmp_priv_t *allowed_privs, target_ulong mode)
>>   {
>>       int i = 0;
>>       int ret = -1;
>> @@ -466,7 +467,7 @@ int pmp_hart_has_privs(CPURISCVState *env, 
>> target_ulong addr,
>>    * Handle a write to a pmpcfg CSR
>>    */
>>   void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
>> -    target_ulong val)
>> +                      target_ulong val)
>>   {
>>       int i;
>>       uint8_t cfg_val;
>> @@ -508,7 +509,7 @@ target_ulong pmpcfg_csr_read(CPURISCVState *env, 
>> uint32_t reg_index)
>>    * Handle a write to a pmpaddr CSR
>>    */
>>   void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
>> -    target_ulong val)
>> +                       target_ulong val)
>>   {
>>       trace_pmpaddr_csr_write(env->mhartid, addr_index, val);
>>   diff --git a/target/riscv/pmp.h b/target/riscv/pmp.h
>> index da32c61c85..b296ea1fc6 100644
>> --- a/target/riscv/pmp.h
>> +++ b/target/riscv/pmp.h
>> @@ -63,18 +63,19 @@ typedef struct {
>>   } pmp_table_t;
>>     void pmpcfg_csr_write(CPURISCVState *env, uint32_t reg_index,
>> -    target_ulong val);
>> +                      target_ulong val);
>>   target_ulong pmpcfg_csr_read(CPURISCVState *env, uint32_t reg_index);
>>     void mseccfg_csr_write(CPURISCVState *env, target_ulong val);
>>   target_ulong mseccfg_csr_read(CPURISCVState *env);
>>     void pmpaddr_csr_write(CPURISCVState *env, uint32_t addr_index,
>> -    target_ulong val);
>> +                       target_ulong val);
>>   target_ulong pmpaddr_csr_read(CPURISCVState *env, uint32_t 
>> addr_index);
>>   int pmp_hart_has_privs(CPURISCVState *env, target_ulong addr,
>> -    target_ulong size, pmp_priv_t privs, pmp_priv_t *allowed_privs,
>> -    target_ulong mode);
>> +                       target_ulong size, pmp_priv_t privs,
>> +                       pmp_priv_t *allowed_privs,
>> +                       target_ulong mode);
>>   target_ulong pmp_get_tlb_size(CPURISCVState *env, int pmp_index,
>>                                 target_ulong tlb_sa, target_ulong 
>> tlb_ea);
>>   void pmp_update_rule_addr(CPURISCVState *env, uint32_t pmp_index);
>> diff --git a/target/riscv/vector_helper.c b/target/riscv/vector_helper.c
>> index 2423affe37..df6201d043 100644
>> --- a/target/riscv/vector_helper.c
>> +++ b/target/riscv/vector_helper.c
>> @@ -1116,7 +1116,7 @@ void HELPER(NAME)(void *vd, void *v0, 
>> target_ulong s1, void *vs2,        \
>> \
>>           *((ETYPE *)vd + H(i)) = DO_OP(s2, (ETYPE)(target_long)s1, 
>> carry);\
>> } \
>> -    env->vstart = 0;                                          \
>> +    env->vstart = 
>> 0;                                                     \
>>       /* set tail elements to 1s 
>> */                                        \
>>       vext_set_elems_1s(vd, vta, vl * esz, total_elems * 
>> esz);             \
>>   }
>> @@ -1308,7 +1308,8 @@ GEN_VEXT_SHIFT_VV(vsra_vv_d, uint64_t, int64_t, 
>> H8, H8, DO_SRL, 0x3f)
>>   /* generate the helpers for shift instructions with one vector and 
>> one scalar */
>>   #define GEN_VEXT_SHIFT_VX(NAME, TD, TS2, HD, HS2, OP, MASK) \
>>   void HELPER(NAME)(void *vd, void *v0, target_ulong s1,      \
>> -        void *vs2, CPURISCVState *env, uint32_t desc)       \
>> +                  void *vs2, CPURISCVState *env,            \
>> +                  uint32_t desc)                            \
>>   {                                                           \
>>       uint32_t vm = vext_vm(desc);                            \
>>       uint32_t vl = env->vl; \
>> @@ -1735,9 +1736,9 @@ GEN_VEXT_VX(vmulhsu_vx_d, 8)
>>   /* Vector Integer Divide Instructions */
>>   #define DO_DIVU(N, M) (unlikely(M == 0) ? (__typeof(N))(-1) : N / M)
>>   #define DO_REMU(N, M) (unlikely(M == 0) ? N : N % M)
>> -#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) :\
>> +#define DO_DIV(N, M)  (unlikely(M == 0) ? (__typeof(N))(-1) : \
>>           unlikely((N == -N) && (M == (__typeof(N))(-1))) ? N : N / M)
>> -#define DO_REM(N, M)  (unlikely(M == 0) ? N :\
>> +#define DO_REM(N, M)  (unlikely(M == 0) ? N : \
>>           unlikely((N == -N) && (M == (__typeof(N))(-1))) ? 0 : N % M)
>>     RVVCALL(OPIVV2, vdivu_vv_b, OP_UUU_B, H1, H1, H1, DO_DIVU)
>> @@ -1846,7 +1847,7 @@ GEN_VEXT_VX(vwmulsu_vx_h, 4)
>>   GEN_VEXT_VX(vwmulsu_vx_w, 8)
>>     /* Vector Single-Width Integer Multiply-Add Instructions */
>> -#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP) \
>> +#define OPIVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
>>   static void do_##NAME(void *vd, void *vs1, void *vs2, int i)       \
>> { \
>>       TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
>> @@ -2277,7 +2278,8 @@ vext_vx_rm_2(void *vd, void *v0, target_long 
>> s1, void *vs2,
>>   /* generate helpers for fixed point instructions with OPIVX format */
>>   #define GEN_VEXT_VX_RM(NAME, ESZ)                         \
>>   void HELPER(NAME)(void *vd, void *v0, target_ulong s1,    \
>> -        void *vs2, CPURISCVState *env, uint32_t desc)     \
>> +                  void *vs2, CPURISCVState *env,          \
>> +                  uint32_t desc)                          \
>>   {                                                         \
>>       vext_vx_rm_2(vd, v0, s1, vs2, env, desc,              \
>>                    do_##NAME, ESZ);                         \
>> @@ -3052,7 +3054,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t 
>> s1,        \
>>       uint32_t vm = vext_vm(desc);                          \
>>       uint32_t vl = env->vl;                                \
>>       uint32_t total_elems =                                \
>> -        vext_get_total_elems(env, desc, ESZ);              \
>> +        vext_get_total_elems(env, desc, ESZ);             \
>>       uint32_t vta = vext_vta(desc);                        \
>>       uint32_t vma = vext_vma(desc);                        \
>>       uint32_t i;                                           \
>> @@ -3118,13 +3120,13 @@ GEN_VEXT_VF(vfrsub_vf_d, 8)
>>   static uint32_t vfwadd16(uint16_t a, uint16_t b, float_status *s)
>>   {
>>       return float32_add(float16_to_float32(a, true, s),
>> -            float16_to_float32(b, true, s), s);
>> +                       float16_to_float32(b, true, s), s);
>>   }
>>     static uint64_t vfwadd32(uint32_t a, uint32_t b, float_status *s)
>>   {
>>       return float64_add(float32_to_float64(a, s),
>> -            float32_to_float64(b, s), s);
>> +                       float32_to_float64(b, s), s);
>>     }
>>   @@ -3140,13 +3142,13 @@ GEN_VEXT_VF(vfwadd_vf_w, 8)
>>   static uint32_t vfwsub16(uint16_t a, uint16_t b, float_status *s)
>>   {
>>       return float32_sub(float16_to_float32(a, true, s),
>> -            float16_to_float32(b, true, s), s);
>> +                       float16_to_float32(b, true, s), s);
>>   }
>>     static uint64_t vfwsub32(uint32_t a, uint32_t b, float_status *s)
>>   {
>>       return float64_sub(float32_to_float64(a, s),
>> -            float32_to_float64(b, s), s);
>> +                       float32_to_float64(b, s), s);
>>     }
>>   @@ -3250,13 +3252,13 @@ GEN_VEXT_VF(vfrdiv_vf_d, 8)
>>   static uint32_t vfwmul16(uint16_t a, uint16_t b, float_status *s)
>>   {
>>       return float32_mul(float16_to_float32(a, true, s),
>> -            float16_to_float32(b, true, s), s);
>> +                       float16_to_float32(b, true, s), s);
>>   }
>>     static uint64_t vfwmul32(uint32_t a, uint32_t b, float_status *s)
>>   {
>>       return float64_mul(float32_to_float64(a, s),
>> -            float32_to_float64(b, s), s);
>> +                       float32_to_float64(b, s), s);
>>     }
>>   RVVCALL(OPFVV2, vfwmul_vv_h, WOP_UUU_H, H4, H2, H2, vfwmul16)
>> @@ -3271,7 +3273,7 @@ GEN_VEXT_VF(vfwmul_vf_w, 8)
>>   /* Vector Single-Width Floating-Point Fused Multiply-Add 
>> Instructions */
>>   #define OPFVV3(NAME, TD, T1, T2, TX1, TX2, HD, HS1, HS2, OP)       \
>>   static void do_##NAME(void *vd, void *vs1, void *vs2, int i,       \
>> -        CPURISCVState *env)                                        \
>> +                      CPURISCVState *env)                          \
>> { \
>>       TX1 s1 = *((T1 *)vs1 + HS1(i));                                \
>>       TX2 s2 = *((T2 *)vs2 + HS2(i));                                \
>> @@ -3303,7 +3305,7 @@ GEN_VEXT_VV_ENV(vfmacc_vv_d, 8)
>>     #define OPFVF3(NAME, TD, T1, T2, TX1, TX2, HD, HS2, OP)           \
>>   static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i,    \
>> -        CPURISCVState *env)                                       \
>> +                      CPURISCVState *env)                         \
>> { \
>>       TX2 s2 = *((T2 *)vs2 + HS2(i));                               \
>>       TD d = *((TD *)vd + HD(i));                                   \
>> @@ -3319,20 +3321,20 @@ GEN_VEXT_VF(vfmacc_vf_d, 8)
>>     static uint16_t fnmacc16(uint16_t a, uint16_t b, uint16_t d, 
>> float_status *s)
>>   {
>> -    return float16_muladd(a, b, d,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float16_muladd(a, b, d, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     static uint32_t fnmacc32(uint32_t a, uint32_t b, uint32_t d, 
>> float_status *s)
>>   {
>> -    return float32_muladd(a, b, d,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float32_muladd(a, b, d, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     static uint64_t fnmacc64(uint64_t a, uint64_t b, uint64_t d, 
>> float_status *s)
>>   {
>> -    return float64_muladd(a, b, d,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float64_muladd(a, b, d, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     RVVCALL(OPFVV3, vfnmacc_vv_h, OP_UUU_H, H2, H2, H2, fnmacc16)
>> @@ -3434,20 +3436,20 @@ GEN_VEXT_VF(vfmadd_vf_d, 8)
>>     static uint16_t fnmadd16(uint16_t a, uint16_t b, uint16_t d, 
>> float_status *s)
>>   {
>> -    return float16_muladd(d, b, a,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float16_muladd(d, b, a, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     static uint32_t fnmadd32(uint32_t a, uint32_t b, uint32_t d, 
>> float_status *s)
>>   {
>> -    return float32_muladd(d, b, a,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float32_muladd(d, b, a, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     static uint64_t fnmadd64(uint64_t a, uint64_t b, uint64_t d, 
>> float_status *s)
>>   {
>> -    return float64_muladd(d, b, a,
>> -            float_muladd_negate_c | float_muladd_negate_product, s);
>> +    return float64_muladd(d, b, a, float_muladd_negate_c |
>> +                                   float_muladd_negate_product, s);
>>   }
>>     RVVCALL(OPFVV3, vfnmadd_vv_h, OP_UUU_H, H2, H2, H2, fnmadd16)
>> @@ -3523,13 +3525,13 @@ GEN_VEXT_VF(vfnmsub_vf_d, 8)
>>   static uint32_t fwmacc16(uint16_t a, uint16_t b, uint32_t d, 
>> float_status *s)
>>   {
>>       return float32_muladd(float16_to_float32(a, true, s),
>> -                        float16_to_float32(b, true, s), d, 0, s);
>> +                          float16_to_float32(b, true, s), d, 0, s);
>>   }
>>     static uint64_t fwmacc32(uint32_t a, uint32_t b, uint64_t d, 
>> float_status *s)
>>   {
>>       return float64_muladd(float32_to_float64(a, s),
>> -                        float32_to_float64(b, s), d, 0, s);
>> +                          float32_to_float64(b, s), d, 0, s);
>>   }
>>     RVVCALL(OPFVV3, vfwmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwmacc16)
>> @@ -3544,15 +3546,16 @@ GEN_VEXT_VF(vfwmacc_vf_w, 8)
>>   static uint32_t fwnmacc16(uint16_t a, uint16_t b, uint32_t d, 
>> float_status *s)
>>   {
>>       return float32_muladd(float16_to_float32(a, true, s),
>> -                        float16_to_float32(b, true, s), d,
>> -                        float_muladd_negate_c | 
>> float_muladd_negate_product, s);
>> +                          float16_to_float32(b, true, s), d,
>> +                          float_muladd_negate_c | 
>> float_muladd_negate_product,
>> +                          s);
>>   }
>>     static uint64_t fwnmacc32(uint32_t a, uint32_t b, uint64_t d, 
>> float_status *s)
>>   {
>> -    return float64_muladd(float32_to_float64(a, s),
>> -                        float32_to_float64(b, s), d,
>> -                        float_muladd_negate_c | 
>> float_muladd_negate_product, s);
>> +    return float64_muladd(float32_to_float64(a, s), 
>> float32_to_float64(b, s),
>> +                          d, float_muladd_negate_c |
>> +                             float_muladd_negate_product, s);
>>   }
>>     RVVCALL(OPFVV3, vfwnmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwnmacc16)
>> @@ -3567,15 +3570,15 @@ GEN_VEXT_VF(vfwnmacc_vf_w, 8)
>>   static uint32_t fwmsac16(uint16_t a, uint16_t b, uint32_t d, 
>> float_status *s)
>>   {
>>       return float32_muladd(float16_to_float32(a, true, s),
>> -                        float16_to_float32(b, true, s), d,
>> -                        float_muladd_negate_c, s);
>> +                          float16_to_float32(b, true, s), d,
>> +                          float_muladd_negate_c, s);
>>   }
>>     static uint64_t fwmsac32(uint32_t a, uint32_t b, uint64_t d, 
>> float_status *s)
>>   {
>>       return float64_muladd(float32_to_float64(a, s),
>> -                        float32_to_float64(b, s), d,
>> -                        float_muladd_negate_c, s);
>> +                          float32_to_float64(b, s), d,
>> +                          float_muladd_negate_c, s);
>>   }
>>     RVVCALL(OPFVV3, vfwmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwmsac16)
>> @@ -3590,15 +3593,15 @@ GEN_VEXT_VF(vfwmsac_vf_w, 8)
>>   static uint32_t fwnmsac16(uint16_t a, uint16_t b, uint32_t d, 
>> float_status *s)
>>   {
>>       return float32_muladd(float16_to_float32(a, true, s),
>> -                        float16_to_float32(b, true, s), d,
>> -                        float_muladd_negate_product, s);
>> +                          float16_to_float32(b, true, s), d,
>> +                          float_muladd_negate_product, s);
>>   }
>>     static uint64_t fwnmsac32(uint32_t a, uint32_t b, uint64_t d, 
>> float_status *s)
>>   {
>>       return float64_muladd(float32_to_float64(a, s),
>> -                        float32_to_float64(b, s), d,
>> -                        float_muladd_negate_product, s);
>> +                          float32_to_float64(b, s), d,
>> +                          float_muladd_negate_product, s);
>>   }
>>     RVVCALL(OPFVV3, vfwnmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwnmsac16)
>> @@ -3616,9 +3619,9 @@ GEN_VEXT_VF(vfwnmsac_vf_w, 8)
>>   #define OP_UU_W uint32_t, uint32_t, uint32_t
>>   #define OP_UU_D uint64_t, uint64_t, uint64_t
>>   -#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)        \
>> +#define OPFVV1(NAME, TD, T2, TX2, HD, HS2, OP)         \
>>   static void do_##NAME(void *vd, void *vs2, int i,      \
>> -        CPURISCVState *env)                            \
>> +                      CPURISCVState *env)              \
>>   {                                                      \
>>       TX2 s2 = *((T2 *)vs2 + HS2(i));                    \
>>       *((TD *)vd + HD(i)) = OP(s2, &env->fp_status); \
>> @@ -3626,7 +3629,7 @@ static void do_##NAME(void *vd, void *vs2, int 
>> i,      \
>>     #define GEN_VEXT_V_ENV(NAME, ESZ)                      \
>>   void HELPER(NAME)(void *vd, void *v0, void *vs2,       \
>> -        CPURISCVState *env, uint32_t desc)             \
>> +                  CPURISCVState *env, uint32_t desc)   \
>>   {                                                      \
>>       uint32_t vm = vext_vm(desc);                       \
>>       uint32_t vl = env->vl;                             \
>> @@ -3703,9 +3706,9 @@ static uint64_t frsqrt7(uint64_t f, int 
>> exp_size, int frac_size)
>>       }
>>         int idx = ((exp & 1) << (precision - 1)) |
>> -                (frac >> (frac_size - precision + 1));
>> +              (frac >> (frac_size - precision + 1));
>>       uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
>> -                            (frac_size - precision);
>> +                        (frac_size - precision);
>>       uint64_t out_exp = (3 * MAKE_64BIT_MASK(0, exp_size - 1) + 
>> ~exp) / 2;
>>         uint64_t val = 0;
>> @@ -3727,9 +3730,9 @@ static float16 frsqrt7_h(float16 f, 
>> float_status *s)
>>        * frsqrt7(-subnormal) = canonical NaN
>>        */
>>       if (float16_is_signaling_nan(f, s) ||
>> -            (float16_is_infinity(f) && sign) ||
>> -            (float16_is_normal(f) && sign) ||
>> -            (float16_is_zero_or_denormal(f) && !float16_is_zero(f) 
>> && sign)) {
>> +        (float16_is_infinity(f) && sign) ||
>> +        (float16_is_normal(f) && sign) ||
>> +        (float16_is_zero_or_denormal(f) && !float16_is_zero(f) && 
>> sign)) {
>>           s->float_exception_flags |= float_flag_invalid;
>>           return float16_default_nan(s);
>>       }
>> @@ -3767,9 +3770,9 @@ static float32 frsqrt7_s(float32 f, 
>> float_status *s)
>>        * frsqrt7(-subnormal) = canonical NaN
>>        */
>>       if (float32_is_signaling_nan(f, s) ||
>> -            (float32_is_infinity(f) && sign) ||
>> -            (float32_is_normal(f) && sign) ||
>> -            (float32_is_zero_or_denormal(f) && !float32_is_zero(f) 
>> && sign)) {
>> +        (float32_is_infinity(f) && sign) ||
>> +        (float32_is_normal(f) && sign) ||
>> +        (float32_is_zero_or_denormal(f) && !float32_is_zero(f) && 
>> sign)) {
>>           s->float_exception_flags |= float_flag_invalid;
>>           return float32_default_nan(s);
>>       }
>> @@ -3807,9 +3810,9 @@ static float64 frsqrt7_d(float64 f, 
>> float_status *s)
>>        * frsqrt7(-subnormal) = canonical NaN
>>        */
>>       if (float64_is_signaling_nan(f, s) ||
>> -            (float64_is_infinity(f) && sign) ||
>> -            (float64_is_normal(f) && sign) ||
>> -            (float64_is_zero_or_denormal(f) && !float64_is_zero(f) 
>> && sign)) {
>> +        (float64_is_infinity(f) && sign) ||
>> +        (float64_is_normal(f) && sign) ||
>> +        (float64_is_zero_or_denormal(f) && !float64_is_zero(f) && 
>> sign)) {
>>           s->float_exception_flags |= float_flag_invalid;
>>           return float64_default_nan(s);
>>       }
>> @@ -3897,18 +3900,18 @@ static uint64_t frec7(uint64_t f, int 
>> exp_size, int frac_size,
>>                   ((s->float_rounding_mode == float_round_up) && 
>> sign)) {
>>                   /* Return greatest/negative finite value. */
>>                   return (sign << (exp_size + frac_size)) |
>> -                    (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
>> +                       (MAKE_64BIT_MASK(frac_size, exp_size) - 1);
>>               } else {
>>                   /* Return +-inf. */
>>                   return (sign << (exp_size + frac_size)) |
>> -                    MAKE_64BIT_MASK(frac_size, exp_size);
>> +                       MAKE_64BIT_MASK(frac_size, exp_size);
>>               }
>>           }
>>       }
>>         int idx = frac >> (frac_size - precision);
>>       uint64_t out_frac = (uint64_t)(lookup_table[idx]) <<
>> -                            (frac_size - precision);
>> +                        (frac_size - precision);
>>       uint64_t out_exp = 2 * MAKE_64BIT_MASK(0, exp_size - 1) + ~exp;
>>         if (out_exp == 0 || out_exp == UINT64_MAX) {
>> @@ -4422,8 +4425,8 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t 
>> s1, void *vs2, \
>> \
>>       for (i = env->vstart; i < vl; i++) {                      \ #define
>>           ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
>> -        *((ETYPE *)vd + H(i))                                 \
>> -          = (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
>> +        *((ETYPE *)vd + H(i)) =                               \
>> +            (!vm && !vext_elem_mask(v0, i) ? s2 : s1);        \
>>       } \
>>       env->vstart = 0;                                          \
>>       /* set tail elements to 1s */ \
>> @@ -4564,7 +4567,8 @@ GEN_VEXT_V_ENV(vfncvt_f_f_w_w, 4)
>>   /* Vector Single-Width Integer Reduction Instructions */
>>   #define GEN_VEXT_RED(NAME, TD, TS2, HD, HS2, OP)          \
>>   void HELPER(NAME)(void *vd, void *v0, void *vs1,          \
>> -        void *vs2, CPURISCVState *env, uint32_t desc)     \
>> +                  void *vs2, CPURISCVState *env,          \
>> +                  uint32_t desc)                          \
>>   {                                                         \
>>       uint32_t vm = vext_vm(desc);                          \
>>       uint32_t vl = env->vl;                                \
>> @@ -5013,7 +5017,8 @@ GEN_VEXT_VSLIDEDOWN_VX(vslidedown_vx_d, 
>> uint64_t, H8)
>>     #define GEN_VEXT_VSLIE1UP(BITWIDTH, 
>> H)                                      \
>>   static void vslide1up_##BITWIDTH(void *vd, void *v0, uint64_t 
>> s1,           \
>> -                     void *vs2, CPURISCVState *env, uint32_t 
>> desc)          \
>> +                                 void *vs2, CPURISCVState 
>> *env,             \
>> +                                 uint32_t 
>> desc)                             \
>> { \
>>       typedef uint##BITWIDTH##_t 
>> ETYPE;                                       \
>>       uint32_t vm = 
>> vext_vm(desc);                                            \
>> @@ -5061,7 +5066,8 @@ GEN_VEXT_VSLIDE1UP_VX(vslide1up_vx_d, 64)
>>     #define GEN_VEXT_VSLIDE1DOWN(BITWIDTH, 
>> H)                                     \
>>   static void vslide1down_##BITWIDTH(void *vd, void *v0, uint64_t 
>> s1,           \
>> -                       void *vs2, CPURISCVState *env, uint32_t 
>> desc)          \
>> +                                   void *vs2, CPURISCVState 
>> *env,             \
>> +                                   uint32_t 
>> desc)                             \
>> { \
>>       typedef uint##BITWIDTH##_t 
>> ETYPE;                                         \
>>       uint32_t vm = 
>> vext_vm(desc);                                              \