target/hppa/translate.c | 498 +++++++++++++++++++++------------------- 1 file changed, 263 insertions(+), 235 deletions(-)
Wrap offset and space into a single structure, with
offset represented either as a constant or as variable.
This enhances copy_iaoq_entry to copy the space, and at
the same time simplifying the code.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
Sven, I think this might solve your branch-with-space-change issue.
There are a couple of places where it does appear as if we might be
losing a space change, with a branch within a branch delay slot.
I'm trying to solve this by keeping both queue entries together.
r~
---
target/hppa/translate.c | 498 +++++++++++++++++++++-------------------
1 file changed, 263 insertions(+), 235 deletions(-)
diff --git a/target/hppa/translate.c b/target/hppa/translate.c
index be0b0494d0..e555d542c8 100644
--- a/target/hppa/translate.c
+++ b/target/hppa/translate.c
@@ -41,15 +41,31 @@ typedef struct DisasCond {
TCGv_i64 a0, a1;
} DisasCond;
+/* Describe a complete Instruction Address Queue entry. */
+typedef struct IAQEntry {
+ /*
+ * Non-constant value of ia space.
+ * When null, IASQ_Front, IASQ_Back, and IASQ_Next are all equal.
+ */
+ TCGv_i64 space;
+ /*
+ * Non-constant value of ia offset, or env destination.
+ * If non-null, will always be set on copy.
+ */
+ TCGv_i64 ofsv;
+ /*
+ * Known constant value of ia offset with no change to ia space.
+ * Unused if either spc or ofsv are non-null.
+ */
+ uint64_t ofsc;
+} IAQEntry;
+
typedef struct DisasContext {
DisasContextBase base;
CPUState *cs;
TCGOp *insn_start;
- uint64_t iaoq_f;
- uint64_t iaoq_b;
- uint64_t iaoq_n;
- TCGv_i64 iaoq_n_var;
+ IAQEntry iaq_f, iaq_b, iaq_n;
DisasCond null_cond;
TCGLabel *null_lab;
@@ -221,15 +237,13 @@ static int cmpbid_c(DisasContext *ctx, int val)
static TCGv_i64 cpu_gr[32];
static TCGv_i64 cpu_sr[4];
static TCGv_i64 cpu_srH;
-static TCGv_i64 cpu_iaoq_f;
-static TCGv_i64 cpu_iaoq_b;
-static TCGv_i64 cpu_iasq_f;
-static TCGv_i64 cpu_iasq_b;
static TCGv_i64 cpu_sar;
static TCGv_i64 cpu_psw_n;
static TCGv_i64 cpu_psw_v;
static TCGv_i64 cpu_psw_cb;
static TCGv_i64 cpu_psw_cb_msb;
+static IAQEntry cpu_iaq_f;
+static IAQEntry cpu_iaq_b;
void hppa_translate_init(void)
{
@@ -242,8 +256,6 @@ void hppa_translate_init(void)
DEF_VAR(psw_v),
DEF_VAR(psw_cb),
DEF_VAR(psw_cb_msb),
- DEF_VAR(iaoq_f),
- DEF_VAR(iaoq_b),
};
#undef DEF_VAR
@@ -282,12 +294,18 @@ void hppa_translate_init(void)
*v->var = tcg_global_mem_new(tcg_env, v->ofs, v->name);
}
- cpu_iasq_f = tcg_global_mem_new_i64(tcg_env,
- offsetof(CPUHPPAState, iasq_f),
- "iasq_f");
- cpu_iasq_b = tcg_global_mem_new_i64(tcg_env,
- offsetof(CPUHPPAState, iasq_b),
- "iasq_b");
+ cpu_iaq_f.ofsv = tcg_global_mem_new_i64(tcg_env,
+ offsetof(CPUHPPAState, iaoq_f),
+ "iaoq_f");
+ cpu_iaq_b.ofsv = tcg_global_mem_new_i64(tcg_env,
+ offsetof(CPUHPPAState, iaoq_b),
+ "iaoq_b");
+ cpu_iaq_f.space = tcg_global_mem_new_i64(tcg_env,
+ offsetof(CPUHPPAState, iasq_f),
+ "iasq_f");
+ cpu_iaq_b.space = tcg_global_mem_new_i64(tcg_env,
+ offsetof(CPUHPPAState, iasq_b),
+ "iasq_b");
}
static void set_insn_breg(DisasContext *ctx, int breg)
@@ -593,31 +611,48 @@ static uint64_t gva_offset_mask(DisasContext *ctx)
: MAKE_64BIT_MASK(0, 32));
}
-static void copy_iaoq_entry(DisasContext *ctx, TCGv_i64 dest,
- uint64_t ival, TCGv_i64 vval)
+static void copy_iaq_entry(DisasContext *ctx, IAQEntry *d, IAQEntry *s)
{
uint64_t mask = gva_offset_mask(ctx);
- if (ival != -1) {
- tcg_gen_movi_i64(dest, ival & mask);
- return;
- }
- tcg_debug_assert(vval != NULL);
-
- /*
- * We know that the IAOQ is already properly masked.
- * This optimization is primarily for "iaoq_f = iaoq_b".
- */
- if (vval == cpu_iaoq_f || vval == cpu_iaoq_b) {
- tcg_gen_mov_i64(dest, vval);
+ if (s->ofsv == NULL) {
+ /* With ofs constant, space is unchanged. */
+ assert(s->space == NULL);
+ if (d->ofsv) {
+ tcg_gen_movi_i64(d->ofsv, s->ofsc & mask);
+ } else {
+ d->ofsc = s->ofsc & mask;
+ }
} else {
- tcg_gen_andi_i64(dest, vval, mask);
+ assert(d->ofsv != NULL);
+ /*
+ * We know that the IAOQ is already properly masked.
+ * This optimization is primarily for "iaoq_f = iaoq_b".
+ */
+ if (s->ofsv == cpu_iaq_f.ofsv || s->ofsv == cpu_iaq_b.ofsv) {
+ tcg_gen_mov_i64(d->ofsv, s->ofsv);
+ } else {
+ tcg_gen_andi_i64(d->ofsv, s->ofsv, mask);
+ }
+ if (s->space) {
+ if (d->space) {
+ tcg_gen_mov_i64(d->space, s->space);
+ } else {
+ d->space = s->space;
+ }
+ }
}
}
-static inline uint64_t iaoq_dest(DisasContext *ctx, int64_t disp)
+static void copy_iaq_offset(DisasContext *ctx, TCGv_i64 dest, IAQEntry *s)
{
- return ctx->iaoq_f + disp + 8;
+ IAQEntry d = { .ofsv = dest };
+ copy_iaq_entry(ctx, &d, s);
+}
+
+static uint64_t iaoq_dest(DisasContext *ctx, int64_t disp)
+{
+ return ctx->iaq_f.ofsc + disp + 8;
}
static void gen_excp_1(int exception)
@@ -627,8 +662,9 @@ static void gen_excp_1(int exception)
static void gen_excp(DisasContext *ctx, int exception)
{
- copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_f, cpu_iaoq_f);
- copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_b, cpu_iaoq_b);
+ /* Writeback constant IAQ offsets. */
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_f);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_b);
nullify_save(ctx);
gen_excp_1(exception);
ctx->base.is_jmp = DISAS_NORETURN;
@@ -660,9 +696,10 @@ static bool gen_illegal(DisasContext *ctx)
} while (0)
#endif
-static bool use_goto_tb(DisasContext *ctx, uint64_t dest)
+static bool use_goto_tb(DisasContext *ctx, IAQEntry *b, IAQEntry *n)
{
- return translator_use_goto_tb(&ctx->base, dest);
+ return (b->ofsv == NULL && n->ofsv == NULL &&
+ translator_use_goto_tb(&ctx->base, b->ofsc));
}
/* If the next insn is to be nullified, and it's on the same page,
@@ -671,21 +708,21 @@ static bool use_goto_tb(DisasContext *ctx, uint64_t dest)
executing a TB that merely branches to the next TB. */
static bool use_nullify_skip(DisasContext *ctx)
{
- return (((ctx->iaoq_b ^ ctx->iaoq_f) & TARGET_PAGE_MASK) == 0
- && !cpu_breakpoint_test(ctx->cs, ctx->iaoq_b, BP_ANY));
+ return ctx->iaq_b.ofsv == NULL
+ && ((ctx->iaq_b.ofsc ^ ctx->iaq_f.ofsc) & TARGET_PAGE_MASK) == 0
+ && !cpu_breakpoint_test(ctx->cs, ctx->iaq_b.ofsc, BP_ANY);
}
-static void gen_goto_tb(DisasContext *ctx, int which,
- uint64_t f, uint64_t b)
+static void gen_goto_tb(DisasContext *ctx, int which, IAQEntry *b, IAQEntry *n)
{
- if (f != -1 && b != -1 && use_goto_tb(ctx, f)) {
+ if (use_goto_tb(ctx, b, n)) {
tcg_gen_goto_tb(which);
- copy_iaoq_entry(ctx, cpu_iaoq_f, f, NULL);
- copy_iaoq_entry(ctx, cpu_iaoq_b, b, NULL);
+ copy_iaq_entry(ctx, &cpu_iaq_f, b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, n);
tcg_gen_exit_tb(ctx->base.tb, which);
} else {
- copy_iaoq_entry(ctx, cpu_iaoq_f, f, cpu_iaoq_b);
- copy_iaoq_entry(ctx, cpu_iaoq_b, b, ctx->iaoq_n_var);
+ copy_iaq_entry(ctx, &cpu_iaq_f, b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, n);
tcg_gen_lookup_and_goto_ptr();
}
}
@@ -1705,35 +1742,62 @@ static bool do_fop_dedd(DisasContext *ctx, unsigned rt,
static bool do_dbranch(DisasContext *ctx, uint64_t dest,
unsigned link, bool is_n)
{
+ IAQEntry n1, n2;
+
if (ctx->null_cond.c == TCG_COND_NEVER && ctx->null_lab == NULL) {
if (link != 0) {
- copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var);
+ copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n);
}
- ctx->iaoq_n = dest;
+
+ /* Perform an unconditional branch by changing IAQ_Next. */
+ n1.space = NULL;
+ n1.ofsv = NULL;
+ n1.ofsc = dest;
+ copy_iaq_entry(ctx, &ctx->iaq_n, &n1);
+
if (is_n) {
ctx->null_cond.c = TCG_COND_ALWAYS;
}
- } else {
- nullify_over(ctx);
-
- if (link != 0) {
- copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var);
- }
-
- if (is_n && use_nullify_skip(ctx)) {
- nullify_set(ctx, 0);
- gen_goto_tb(ctx, 0, dest, dest + 4);
- } else {
- nullify_set(ctx, is_n);
- gen_goto_tb(ctx, 0, ctx->iaoq_b, dest);
- }
-
- nullify_end(ctx);
-
- nullify_set(ctx, 0);
- gen_goto_tb(ctx, 1, ctx->iaoq_b, ctx->iaoq_n);
- ctx->base.is_jmp = DISAS_NORETURN;
+ return true;
}
+
+ /* Conditional branch by virtue of nullification. */
+ nullify_over(ctx);
+
+ if (link != 0) {
+ copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n);
+ }
+
+ /*
+ * Retain any space change already in the queue.
+ * This can happen with B in delay slot of BE.
+ */
+ if (ctx->iaq_n.space) {
+ n1.space = n2.space = ctx->iaq_n.space;
+ n1.ofsv = tcg_constant_i64(dest);
+ n2.ofsv = tcg_constant_i64(dest + 4);
+ n1.ofsc = n2.ofsc = 0;
+ } else {
+ n1.space = n2.space = NULL;
+ n1.ofsv = n2.ofsv = NULL;
+ n1.ofsc = dest;
+ n2.ofsc = dest + 4;
+ }
+
+ if (is_n && use_nullify_skip(ctx)) {
+ nullify_set(ctx, 0);
+ gen_goto_tb(ctx, 0, &n1, &n2);
+ } else {
+ nullify_set(ctx, is_n);
+ gen_goto_tb(ctx, 0, &ctx->iaq_b, &n1);
+ }
+
+ nullify_end(ctx);
+
+ /* Branch not taken. */
+ nullify_set(ctx, 0);
+ gen_goto_tb(ctx, 1, &ctx->iaq_b, &ctx->iaq_n);
+ ctx->base.is_jmp = DISAS_NORETURN;
return true;
}
@@ -1745,17 +1809,15 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n,
uint64_t dest = iaoq_dest(ctx, disp);
TCGLabel *taken = NULL;
TCGCond c = cond->c;
+ IAQEntry n1, n2;
bool n;
assert(ctx->null_cond.c == TCG_COND_NEVER);
- /* Handle TRUE and NEVER as direct branches. */
+ /* Handle ALWAYS as a direct branch. */
if (c == TCG_COND_ALWAYS) {
return do_dbranch(ctx, dest, 0, is_n && disp >= 0);
}
- if (c == TCG_COND_NEVER) {
- return do_dbranch(ctx, ctx->iaoq_n, 0, is_n && disp < 0);
- }
taken = gen_new_label();
tcg_gen_brcond_i64(c, cond->a0, cond->a1, taken);
@@ -1763,33 +1825,40 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n,
/* Not taken: Condition not satisfied; nullify on backward branches. */
n = is_n && disp < 0;
- if (n && use_nullify_skip(ctx)) {
- nullify_set(ctx, 0);
- gen_goto_tb(ctx, 0, ctx->iaoq_n, ctx->iaoq_n + 4);
- } else {
- if (!n && ctx->null_lab) {
- gen_set_label(ctx->null_lab);
- ctx->null_lab = NULL;
- }
- nullify_set(ctx, n);
- if (ctx->iaoq_n == -1) {
- /* The temporary iaoq_n_var died at the branch above.
- Regenerate it here instead of saving it. */
- tcg_gen_addi_i64(ctx->iaoq_n_var, cpu_iaoq_b, 4);
- }
- gen_goto_tb(ctx, 0, ctx->iaoq_b, ctx->iaoq_n);
+ if (!n && ctx->null_lab) {
+ gen_set_label(ctx->null_lab);
+ ctx->null_lab = NULL;
}
+ nullify_set(ctx, n);
+ gen_goto_tb(ctx, 0, &ctx->iaq_b, &ctx->iaq_n);
gen_set_label(taken);
/* Taken: Condition satisfied; nullify on forward branches. */
n = is_n && disp >= 0;
+
+ /*
+ * Retain any space change already in the queue.
+ * This can happen with ADDB in delay slot of BE.
+ */
+ if (ctx->iaq_n.space) {
+ n1.space = n2.space = ctx->iaq_n.space;
+ n1.ofsv = tcg_constant_i64(dest);
+ n2.ofsv = tcg_constant_i64(dest + 4);
+ n1.ofsc = n2.ofsc = 0;
+ } else {
+ n1.space = n2.space = NULL;
+ n1.ofsv = n2.ofsv = NULL;
+ n1.ofsc = dest;
+ n2.ofsc = dest + 4;
+ }
+
if (n && use_nullify_skip(ctx)) {
nullify_set(ctx, 0);
- gen_goto_tb(ctx, 1, dest, dest + 4);
+ gen_goto_tb(ctx, 1, &n1, &n2);
} else {
nullify_set(ctx, n);
- gen_goto_tb(ctx, 1, ctx->iaoq_b, dest);
+ gen_goto_tb(ctx, 1, &ctx->iaq_b, &n1);
}
/* Not taken: the branch itself was nullified. */
@@ -1808,84 +1877,59 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n,
static bool do_ibranch(DisasContext *ctx, TCGv_i64 dest,
unsigned link, bool is_n)
{
- TCGv_i64 a0, a1, next, tmp;
- TCGCond c;
+ IAQEntry next = {
+ .space = ctx->iaq_n.space,
+ .ofsv = tcg_temp_new_i64()
+ };
assert(ctx->null_lab == NULL);
if (ctx->null_cond.c == TCG_COND_NEVER) {
+ tcg_gen_mov_i64(next.ofsv, dest);
+
if (link != 0) {
- copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var);
+ copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n);
}
- next = tcg_temp_new_i64();
- tcg_gen_mov_i64(next, dest);
if (is_n) {
if (use_nullify_skip(ctx)) {
- copy_iaoq_entry(ctx, cpu_iaoq_f, -1, next);
- tcg_gen_addi_i64(next, next, 4);
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, next);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &next);
+ tcg_gen_addi_i64(next.ofsv, next.ofsv, 4);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
nullify_set(ctx, 0);
ctx->base.is_jmp = DISAS_IAQ_N_UPDATED;
return true;
}
ctx->null_cond.c = TCG_COND_ALWAYS;
}
- ctx->iaoq_n = -1;
- ctx->iaoq_n_var = next;
- } else if (is_n && use_nullify_skip(ctx)) {
+ ctx->iaq_n = next;
+ return true;
+ }
+
+ nullify_over(ctx);
+
+ tcg_gen_mov_i64(next.ofsv, dest);
+ if (link != 0) {
+ copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n);
+ }
+
+ if (is_n && use_nullify_skip(ctx)) {
/* The (conditional) branch, B, nullifies the next insn, N,
and we're allowed to skip execution N (no single-step or
tracepoint in effect). Since the goto_ptr that we must use
for the indirect branch consumes no special resources, we
can (conditionally) skip B and continue execution. */
- /* The use_nullify_skip test implies we have a known control path. */
- tcg_debug_assert(ctx->iaoq_b != -1);
- tcg_debug_assert(ctx->iaoq_n != -1);
-
- /* We do have to handle the non-local temporary, DEST, before
- branching. Since IOAQ_F is not really live at this point, we
- can simply store DEST optimistically. Similarly with IAOQ_B. */
- copy_iaoq_entry(ctx, cpu_iaoq_f, -1, dest);
- next = tcg_temp_new_i64();
- tcg_gen_addi_i64(next, dest, 4);
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, next);
-
- nullify_over(ctx);
- if (link != 0) {
- copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var);
- }
- tcg_gen_lookup_and_goto_ptr();
- return nullify_end(ctx);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &next);
+ tcg_gen_addi_i64(next.ofsv, next.ofsv, 4);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
+ nullify_set(ctx, 0);
} else {
- c = ctx->null_cond.c;
- a0 = ctx->null_cond.a0;
- a1 = ctx->null_cond.a1;
-
- tmp = tcg_temp_new_i64();
- next = tcg_temp_new_i64();
-
- copy_iaoq_entry(ctx, tmp, ctx->iaoq_n, ctx->iaoq_n_var);
- tcg_gen_movcond_i64(c, next, a0, a1, tmp, dest);
- ctx->iaoq_n = -1;
- ctx->iaoq_n_var = next;
-
- if (link != 0) {
- tcg_gen_movcond_i64(c, cpu_gr[link], a0, a1, cpu_gr[link], tmp);
- }
-
- if (is_n) {
- /* The branch nullifies the next insn, which means the state of N
- after the branch is the inverse of the state of N that applied
- to the branch. */
- tcg_gen_setcond_i64(tcg_invert_cond(c), cpu_psw_n, a0, a1);
- cond_free(&ctx->null_cond);
- ctx->null_cond = cond_make_n();
- ctx->psw_n_nonzero = true;
- } else {
- cond_free(&ctx->null_cond);
- }
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
+ nullify_set(ctx, is_n);
}
- return true;
+ tcg_gen_lookup_and_goto_ptr();
+ ctx->base.is_jmp = DISAS_NORETURN;
+ return nullify_end(ctx);
}
/* Implement
@@ -1927,7 +1971,7 @@ static TCGv_i64 do_ibranch_priv(DisasContext *ctx, TCGv_i64 offset)
aforementioned BE. */
static void do_page_zero(DisasContext *ctx)
{
- TCGv_i64 tmp;
+ IAQEntry next = { };
/* If by some means we get here with PSW[N]=1, that implies that
the B,GATE instruction would be skipped, and we'd fault on the
@@ -1948,11 +1992,11 @@ static void do_page_zero(DisasContext *ctx)
non-sequential instruction execution. Normally the PSW[B] bit
detects this by disallowing the B,GATE instruction to execute
under such conditions. */
- if (ctx->iaoq_b != ctx->iaoq_f + 4) {
+ if (ctx->iaq_b.ofsv || ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4) {
goto do_sigill;
}
- switch (ctx->iaoq_f & -4) {
+ switch (ctx->iaq_f.ofsc & -4) {
case 0x00: /* Null pointer call */
gen_excp_1(EXCP_IMP);
ctx->base.is_jmp = DISAS_NORETURN;
@@ -1965,11 +2009,11 @@ static void do_page_zero(DisasContext *ctx)
case 0xe0: /* SET_THREAD_POINTER */
tcg_gen_st_i64(cpu_gr[26], tcg_env, offsetof(CPUHPPAState, cr[27]));
- tmp = tcg_temp_new_i64();
- tcg_gen_ori_i64(tmp, cpu_gr[31], 3);
- copy_iaoq_entry(ctx, cpu_iaoq_f, -1, tmp);
- tcg_gen_addi_i64(tmp, tmp, 4);
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp);
+ next.ofsv = tcg_temp_new_i64();
+ tcg_gen_ori_i64(next.ofsv, cpu_gr[31], 3);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &next);
+ tcg_gen_addi_i64(next.ofsv, next.ofsv, 4);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
ctx->base.is_jmp = DISAS_IAQ_N_UPDATED;
break;
@@ -2011,7 +2055,8 @@ static bool trans_mfia(DisasContext *ctx, arg_mfia *a)
{
unsigned rt = a->t;
TCGv_i64 tmp = dest_gpr(ctx, rt);
- tcg_gen_movi_i64(tmp, ctx->iaoq_f);
+
+ tcg_gen_movi_i64(tmp, ctx->iaq_f.ofsc);
save_gpr(ctx, rt, tmp);
cond_free(&ctx->null_cond);
@@ -2678,8 +2723,8 @@ static bool trans_or(DisasContext *ctx, arg_rrr_cf_d *a)
nullify_over(ctx);
/* Advance the instruction queue. */
- copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b);
- copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_n, ctx->iaoq_n_var);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_n);
nullify_set(ctx, 0);
/* Tell the qemu main loop to halt until this cpu has work. */
@@ -3805,26 +3850,21 @@ static bool trans_be(DisasContext *ctx, arg_be *a)
#ifdef CONFIG_USER_ONLY
return do_ibranch(ctx, tmp, a->l, a->n);
#else
- TCGv_i64 new_spc = tcg_temp_new_i64();
+ IAQEntry next = { .ofsv = tmp, .space = tcg_temp_new_i64() };
+ load_spr(ctx, next.space, a->sp);
- load_spr(ctx, new_spc, a->sp);
if (a->l) {
- copy_iaoq_entry(ctx, cpu_gr[31], ctx->iaoq_n, ctx->iaoq_n_var);
- tcg_gen_mov_i64(cpu_sr[0], cpu_iasq_f);
+ copy_iaq_offset(ctx, cpu_gr[31], &ctx->iaq_n);
+ tcg_gen_mov_i64(cpu_sr[0], ctx->iaq_n.space ? : cpu_iaq_b.space);
}
if (a->n && use_nullify_skip(ctx)) {
- copy_iaoq_entry(ctx, cpu_iaoq_f, -1, tmp);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &next);
tcg_gen_addi_i64(tmp, tmp, 4);
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp);
- tcg_gen_mov_i64(cpu_iasq_f, new_spc);
- tcg_gen_mov_i64(cpu_iasq_b, cpu_iasq_f);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
+ nullify_set(ctx, 0);
} else {
- copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b);
- if (ctx->iaoq_b == -1) {
- tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b);
- }
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp);
- tcg_gen_mov_i64(cpu_iasq_b, new_spc);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
nullify_set(ctx, a->n);
}
tcg_gen_lookup_and_goto_ptr();
@@ -3855,7 +3895,7 @@ static bool trans_b_gate(DisasContext *ctx, arg_b_gate *a)
* b evil
* in which instructions at evil would run with increased privs.
*/
- if (ctx->iaoq_b == -1 || ctx->iaoq_b != ctx->iaoq_f + 4) {
+ if (ctx->iaq_b.ofsv || ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4) {
return gen_illegal(ctx);
}
@@ -3897,12 +3937,12 @@ static bool trans_blr(DisasContext *ctx, arg_blr *a)
if (a->x) {
TCGv_i64 tmp = tcg_temp_new_i64();
tcg_gen_shli_i64(tmp, load_gpr(ctx, a->x), 3);
- tcg_gen_addi_i64(tmp, tmp, ctx->iaoq_f + 8);
+ tcg_gen_addi_i64(tmp, tmp, ctx->iaq_f.ofsc + 8);
/* The computation here never changes privilege level. */
return do_ibranch(ctx, tmp, a->l, a->n);
} else {
/* BLR R0,RX is a good way to load PC+8 into RX. */
- return do_dbranch(ctx, ctx->iaoq_f + 8, a->l, a->n);
+ return do_dbranch(ctx, ctx->iaq_f.ofsc + 8, a->l, a->n);
}
}
@@ -3923,23 +3963,23 @@ static bool trans_bv(DisasContext *ctx, arg_bv *a)
static bool trans_bve(DisasContext *ctx, arg_bve *a)
{
- TCGv_i64 dest;
#ifdef CONFIG_USER_ONLY
- dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b));
+ TCGv_i64 dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b));
return do_ibranch(ctx, dest, a->l, a->n);
#else
- nullify_over(ctx);
- dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b));
+ IAQEntry next;
- copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b);
- if (ctx->iaoq_b == -1) {
- tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b);
- }
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, dest);
- tcg_gen_mov_i64(cpu_iasq_b, space_select(ctx, 0, dest));
+ nullify_over(ctx);
+
+ next.ofsc = 0;
+ next.ofsv = do_ibranch_priv(ctx, load_gpr(ctx, a->b));
+ next.space = space_select(ctx, 0, next.ofsv);
+
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &next);
if (a->l) {
- copy_iaoq_entry(ctx, cpu_gr[a->l], ctx->iaoq_n, ctx->iaoq_n_var);
+ copy_iaq_offset(ctx, cpu_gr[a->l], &ctx->iaq_n);
}
nullify_set(ctx, a->n);
tcg_gen_lookup_and_goto_ptr();
@@ -4489,8 +4529,8 @@ static void hppa_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
#ifdef CONFIG_USER_ONLY
ctx->privilege = MMU_IDX_TO_PRIV(MMU_USER_IDX);
ctx->mmu_idx = MMU_USER_IDX;
- ctx->iaoq_f = ctx->base.pc_first | ctx->privilege;
- ctx->iaoq_b = ctx->base.tb->cs_base | ctx->privilege;
+ ctx->iaq_f.ofsc = ctx->base.pc_first | ctx->privilege;
+ ctx->iaq_b.ofsc = ctx->base.tb->cs_base | ctx->privilege;
ctx->unalign = (ctx->tb_flags & TB_FLAG_UNALIGN ? MO_UNALN : MO_ALIGN);
#else
ctx->privilege = (ctx->tb_flags >> TB_FLAG_PRIV_SHIFT) & 3;
@@ -4503,12 +4543,18 @@ static void hppa_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
uint64_t iasq_f = cs_base & ~0xffffffffull;
int32_t diff = cs_base;
- ctx->iaoq_f = (ctx->base.pc_first & ~iasq_f) + ctx->privilege;
- ctx->iaoq_b = (diff ? ctx->iaoq_f + diff : -1);
-#endif
- ctx->iaoq_n = -1;
- ctx->iaoq_n_var = NULL;
+ ctx->iaq_f.ofsc = (ctx->base.pc_first & ~iasq_f) + ctx->privilege;
+ ctx->iaq_f.ofsv = NULL;
+ ctx->iaq_f.space = NULL;
+ if (diff) {
+ ctx->iaq_b.ofsc = ctx->iaq_f.ofsc + diff;
+ ctx->iaq_b.ofsv = NULL;
+ ctx->iaq_b.space = NULL;
+ } else {
+ ctx->iaq_b = cpu_iaq_b;
+ }
+#endif
ctx->zero = tcg_constant_i64(0);
/* Bound the number of instructions by those left on the page. */
@@ -4533,8 +4579,13 @@ static void hppa_tr_tb_start(DisasContextBase *dcbase, CPUState *cs)
static void hppa_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
{
DisasContext *ctx = container_of(dcbase, DisasContext, base);
+ uint64_t b_ofs;
- tcg_gen_insn_start(ctx->iaoq_f, ctx->iaoq_b, 0);
+ assert(ctx->iaq_f.ofsv == NULL);
+ assert(ctx->iaq_f.space == NULL);
+ assert(ctx->iaq_f.ofsc != -1);
+ b_ofs = ctx->iaq_b.ofsv ? -1 : ctx->iaq_b.ofsc;
+ tcg_gen_insn_start(ctx->iaq_f.ofsc, b_ofs, 0);
ctx->insn_start = tcg_last_op();
}
@@ -4559,13 +4610,16 @@ static void hppa_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
/* Set up the IA queue for the next insn.
This will be overwritten by a branch. */
- if (ctx->iaoq_b == -1) {
- ctx->iaoq_n = -1;
- ctx->iaoq_n_var = tcg_temp_new_i64();
- tcg_gen_addi_i64(ctx->iaoq_n_var, cpu_iaoq_b, 4);
+ if (ctx->iaq_b.ofsv) {
+ ctx->iaq_n.ofsc = 0;
+ ctx->iaq_n.ofsv = tcg_temp_new_i64();
+ tcg_gen_addi_i64(ctx->iaq_n.ofsv, ctx->iaq_b.ofsv, 4);
+ ctx->iaq_n.space = ctx->iaq_b.space;
} else {
- ctx->iaoq_n = ctx->iaoq_b + 4;
- ctx->iaoq_n_var = NULL;
+ assert(ctx->iaq_b.space == NULL);
+ ctx->iaq_n.ofsc = ctx->iaq_b.ofsc + 4;
+ ctx->iaq_n.ofsv = NULL;
+ ctx->iaq_n.space = NULL;
}
if (unlikely(ctx->null_cond.c == TCG_COND_ALWAYS)) {
@@ -4583,48 +4637,22 @@ static void hppa_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
/* Advance the insn queue. Note that this check also detects
a priority change within the instruction queue. */
- if (ret == DISAS_NEXT && ctx->iaoq_b != ctx->iaoq_f + 4) {
- if (ctx->iaoq_b != -1 && ctx->iaoq_n != -1
- && use_goto_tb(ctx, ctx->iaoq_b)
+ if (ret == DISAS_NEXT
+ && (ctx->iaq_b.ofsv != NULL ||
+ ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4)) {
+ if (use_goto_tb(ctx, &ctx->iaq_b, &ctx->iaq_n)
&& (ctx->null_cond.c == TCG_COND_NEVER
|| ctx->null_cond.c == TCG_COND_ALWAYS)) {
nullify_set(ctx, ctx->null_cond.c == TCG_COND_ALWAYS);
- gen_goto_tb(ctx, 0, ctx->iaoq_b, ctx->iaoq_n);
+ gen_goto_tb(ctx, 0, &ctx->iaq_b, &ctx->iaq_n);
ctx->base.is_jmp = ret = DISAS_NORETURN;
} else {
ctx->base.is_jmp = ret = DISAS_IAQ_N_STALE;
}
}
- ctx->iaoq_f = ctx->iaoq_b;
- ctx->iaoq_b = ctx->iaoq_n;
+ ctx->iaq_f = ctx->iaq_b;
+ ctx->iaq_b = ctx->iaq_n;
ctx->base.pc_next += 4;
-
- switch (ret) {
- case DISAS_NORETURN:
- case DISAS_IAQ_N_UPDATED:
- break;
-
- case DISAS_NEXT:
- case DISAS_IAQ_N_STALE:
- case DISAS_IAQ_N_STALE_EXIT:
- if (ctx->iaoq_f == -1) {
- copy_iaoq_entry(ctx, cpu_iaoq_f, -1, cpu_iaoq_b);
- copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_n, ctx->iaoq_n_var);
-#ifndef CONFIG_USER_ONLY
- tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b);
-#endif
- nullify_save(ctx);
- ctx->base.is_jmp = (ret == DISAS_IAQ_N_STALE_EXIT
- ? DISAS_EXIT
- : DISAS_IAQ_N_UPDATED);
- } else if (ctx->iaoq_b == -1) {
- copy_iaoq_entry(ctx, cpu_iaoq_b, -1, ctx->iaoq_n_var);
- }
- break;
-
- default:
- g_assert_not_reached();
- }
}
static void hppa_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
@@ -4638,8 +4666,8 @@ static void hppa_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
case DISAS_TOO_MANY:
case DISAS_IAQ_N_STALE:
case DISAS_IAQ_N_STALE_EXIT:
- copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_f, cpu_iaoq_f);
- copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_b, cpu_iaoq_b);
+ copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_f);
+ copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_b);
nullify_save(ctx);
/* FALLTHRU */
case DISAS_IAQ_N_UPDATED:
@@ -4694,6 +4722,6 @@ static const TranslatorOps hppa_tr_ops = {
void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns,
vaddr pc, void *host_pc)
{
- DisasContext ctx;
+ DisasContext ctx = { };
translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base);
}
--
2.34.1
Hi Richard, Richard Henderson <richard.henderson@linaro.org> writes: > Wrap offset and space into a single structure, with > offset represented either as a constant or as variable. > This enhances copy_iaoq_entry to copy the space, and at > the same time simplifying the code. > > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> > --- > > Sven, I think this might solve your branch-with-space-change issue. > There are a couple of places where it does appear as if we might be > losing a space change, with a branch within a branch delay slot. > I'm trying to solve this by keeping both queue entries together. Thanks! I've put that on top of the other changes i have for 64bit HP-UX, but it introduces a new problem: * Checking configuration for consistency... WARNING: The disk at: 10/0/0/0.0.0 (QEMU_QEMU_HARDDISK) appears to contain a file system and boot area. Continuing the installation will destroy any existing data on this disk. Do you wish to cancel the non-interactive installation in order to respond to the warnings above? ([y]/n): n * Loading configuration utility... * Beginning installation from source: /dev/dsk/c0t2d0 ======= 03/19/24 13:59:53 EST Starting system configuration... * Configure_Disks: Begin * Will install B.11.11 onto this system. * Creating LVM physical volume "/dev/rdsk/c0t0d0" (10/0/0/0.0.0). * Creating volume group "vg00". vgcreate: Illegal path "". * Creating logical volume "vg00/lvol1" (/stand). lvcreate: Illegal path "". * Creating logical volume "vg00/lvol2" (swap_dump). lvcreate: Illegal path "". * Creating logical volume "vg00/lvol3" (/). lvcreate: Illegal path "". * Creating logical volume "vg00/lvol4" (/tmp). lvcreate: Illegal path "". * Creating logical volume "vg00/lvol5" (/home). lvcreate: Illegal path "". * Creating logical volume "vg00/lvol6" (/opt). lvcreate: Illegal path "". The 'illegal path' errors where not present before. I haven't done any debugging, as debugging hppa userspace in qemu is a really time consuming thing. But when you're add ioaq/iasq handling - The branch i was referring to is https://github.com/svenschnelle/qemu/commits/hppa_fixes64/ With the main difference in commits: https://github.com/svenschnelle/qemu/commit/a67a19ac082ab3435009c1ab3b57bfe6d9588827 https://github.com/svenschnelle/qemu/commit/ca288e26f92b61cc9148fa5cfe7552aba6241466 The second commit is really hacky, because it tries to hack around the problem the we have 2 bits Space select, 22 bits Space and 42 bits offset in PA2.0. And than there's also the problem that the iaoq_f/iaoq_b difference might get encoded into cs_base, but that works only if it's less than 32bits. I don't have a solution for that. But at least it gets a 64 bit HP-UX image from my C8000 booted in qemu. So if you know of a nice way to handle the space selector bits + 64 Bit GVA, im happy to test patches :-) > r~ > > --- > target/hppa/translate.c | 498 +++++++++++++++++++++------------------- > 1 file changed, 263 insertions(+), 235 deletions(-) > > diff --git a/target/hppa/translate.c b/target/hppa/translate.c > index be0b0494d0..e555d542c8 100644 > --- a/target/hppa/translate.c > +++ b/target/hppa/translate.c > @@ -41,15 +41,31 @@ typedef struct DisasCond { > TCGv_i64 a0, a1; > } DisasCond; > > +/* Describe a complete Instruction Address Queue entry. */ > +typedef struct IAQEntry { > + /* > + * Non-constant value of ia space. > + * When null, IASQ_Front, IASQ_Back, and IASQ_Next are all equal. > + */ > + TCGv_i64 space; > + /* > + * Non-constant value of ia offset, or env destination. > + * If non-null, will always be set on copy. > + */ > + TCGv_i64 ofsv; > + /* > + * Known constant value of ia offset with no change to ia space. > + * Unused if either spc or ofsv are non-null. > + */ > + uint64_t ofsc; > +} IAQEntry; > + > typedef struct DisasContext { > DisasContextBase base; > CPUState *cs; > TCGOp *insn_start; > > - uint64_t iaoq_f; > - uint64_t iaoq_b; > - uint64_t iaoq_n; > - TCGv_i64 iaoq_n_var; > + IAQEntry iaq_f, iaq_b, iaq_n; > > DisasCond null_cond; > TCGLabel *null_lab; > @@ -221,15 +237,13 @@ static int cmpbid_c(DisasContext *ctx, int val) > static TCGv_i64 cpu_gr[32]; > static TCGv_i64 cpu_sr[4]; > static TCGv_i64 cpu_srH; > -static TCGv_i64 cpu_iaoq_f; > -static TCGv_i64 cpu_iaoq_b; > -static TCGv_i64 cpu_iasq_f; > -static TCGv_i64 cpu_iasq_b; > static TCGv_i64 cpu_sar; > static TCGv_i64 cpu_psw_n; > static TCGv_i64 cpu_psw_v; > static TCGv_i64 cpu_psw_cb; > static TCGv_i64 cpu_psw_cb_msb; > +static IAQEntry cpu_iaq_f; > +static IAQEntry cpu_iaq_b; > > void hppa_translate_init(void) > { > @@ -242,8 +256,6 @@ void hppa_translate_init(void) > DEF_VAR(psw_v), > DEF_VAR(psw_cb), > DEF_VAR(psw_cb_msb), > - DEF_VAR(iaoq_f), > - DEF_VAR(iaoq_b), > }; > > #undef DEF_VAR > @@ -282,12 +294,18 @@ void hppa_translate_init(void) > *v->var = tcg_global_mem_new(tcg_env, v->ofs, v->name); > } > > - cpu_iasq_f = tcg_global_mem_new_i64(tcg_env, > - offsetof(CPUHPPAState, iasq_f), > - "iasq_f"); > - cpu_iasq_b = tcg_global_mem_new_i64(tcg_env, > - offsetof(CPUHPPAState, iasq_b), > - "iasq_b"); > + cpu_iaq_f.ofsv = tcg_global_mem_new_i64(tcg_env, > + offsetof(CPUHPPAState, iaoq_f), > + "iaoq_f"); > + cpu_iaq_b.ofsv = tcg_global_mem_new_i64(tcg_env, > + offsetof(CPUHPPAState, iaoq_b), > + "iaoq_b"); > + cpu_iaq_f.space = tcg_global_mem_new_i64(tcg_env, > + offsetof(CPUHPPAState, iasq_f), > + "iasq_f"); > + cpu_iaq_b.space = tcg_global_mem_new_i64(tcg_env, > + offsetof(CPUHPPAState, iasq_b), > + "iasq_b"); > } > > static void set_insn_breg(DisasContext *ctx, int breg) > @@ -593,31 +611,48 @@ static uint64_t gva_offset_mask(DisasContext *ctx) > : MAKE_64BIT_MASK(0, 32)); > } > > -static void copy_iaoq_entry(DisasContext *ctx, TCGv_i64 dest, > - uint64_t ival, TCGv_i64 vval) > +static void copy_iaq_entry(DisasContext *ctx, IAQEntry *d, IAQEntry *s) > { > uint64_t mask = gva_offset_mask(ctx); > > - if (ival != -1) { > - tcg_gen_movi_i64(dest, ival & mask); > - return; > - } > - tcg_debug_assert(vval != NULL); > - > - /* > - * We know that the IAOQ is already properly masked. > - * This optimization is primarily for "iaoq_f = iaoq_b". > - */ > - if (vval == cpu_iaoq_f || vval == cpu_iaoq_b) { > - tcg_gen_mov_i64(dest, vval); > + if (s->ofsv == NULL) { > + /* With ofs constant, space is unchanged. */ > + assert(s->space == NULL); > + if (d->ofsv) { > + tcg_gen_movi_i64(d->ofsv, s->ofsc & mask); > + } else { > + d->ofsc = s->ofsc & mask; > + } > } else { > - tcg_gen_andi_i64(dest, vval, mask); > + assert(d->ofsv != NULL); > + /* > + * We know that the IAOQ is already properly masked. > + * This optimization is primarily for "iaoq_f = iaoq_b". > + */ > + if (s->ofsv == cpu_iaq_f.ofsv || s->ofsv == cpu_iaq_b.ofsv) { > + tcg_gen_mov_i64(d->ofsv, s->ofsv); > + } else { > + tcg_gen_andi_i64(d->ofsv, s->ofsv, mask); > + } > + if (s->space) { > + if (d->space) { > + tcg_gen_mov_i64(d->space, s->space); > + } else { > + d->space = s->space; > + } > + } > } > } > > -static inline uint64_t iaoq_dest(DisasContext *ctx, int64_t disp) > +static void copy_iaq_offset(DisasContext *ctx, TCGv_i64 dest, IAQEntry *s) > { > - return ctx->iaoq_f + disp + 8; > + IAQEntry d = { .ofsv = dest }; > + copy_iaq_entry(ctx, &d, s); > +} > + > +static uint64_t iaoq_dest(DisasContext *ctx, int64_t disp) > +{ > + return ctx->iaq_f.ofsc + disp + 8; > } > > static void gen_excp_1(int exception) > @@ -627,8 +662,9 @@ static void gen_excp_1(int exception) > > static void gen_excp(DisasContext *ctx, int exception) > { > - copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_f, cpu_iaoq_f); > - copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_b, cpu_iaoq_b); > + /* Writeback constant IAQ offsets. */ > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_f); > + copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_b); > nullify_save(ctx); > gen_excp_1(exception); > ctx->base.is_jmp = DISAS_NORETURN; > @@ -660,9 +696,10 @@ static bool gen_illegal(DisasContext *ctx) > } while (0) > #endif > > -static bool use_goto_tb(DisasContext *ctx, uint64_t dest) > +static bool use_goto_tb(DisasContext *ctx, IAQEntry *b, IAQEntry *n) > { > - return translator_use_goto_tb(&ctx->base, dest); > + return (b->ofsv == NULL && n->ofsv == NULL && > + translator_use_goto_tb(&ctx->base, b->ofsc)); > } > > /* If the next insn is to be nullified, and it's on the same page, > @@ -671,21 +708,21 @@ static bool use_goto_tb(DisasContext *ctx, uint64_t dest) > executing a TB that merely branches to the next TB. */ > static bool use_nullify_skip(DisasContext *ctx) > { > - return (((ctx->iaoq_b ^ ctx->iaoq_f) & TARGET_PAGE_MASK) == 0 > - && !cpu_breakpoint_test(ctx->cs, ctx->iaoq_b, BP_ANY)); > + return ctx->iaq_b.ofsv == NULL > + && ((ctx->iaq_b.ofsc ^ ctx->iaq_f.ofsc) & TARGET_PAGE_MASK) == 0 > + && !cpu_breakpoint_test(ctx->cs, ctx->iaq_b.ofsc, BP_ANY); > } > > -static void gen_goto_tb(DisasContext *ctx, int which, > - uint64_t f, uint64_t b) > +static void gen_goto_tb(DisasContext *ctx, int which, IAQEntry *b, IAQEntry *n) > { > - if (f != -1 && b != -1 && use_goto_tb(ctx, f)) { > + if (use_goto_tb(ctx, b, n)) { > tcg_gen_goto_tb(which); > - copy_iaoq_entry(ctx, cpu_iaoq_f, f, NULL); > - copy_iaoq_entry(ctx, cpu_iaoq_b, b, NULL); > + copy_iaq_entry(ctx, &cpu_iaq_f, b); > + copy_iaq_entry(ctx, &cpu_iaq_b, n); > tcg_gen_exit_tb(ctx->base.tb, which); > } else { > - copy_iaoq_entry(ctx, cpu_iaoq_f, f, cpu_iaoq_b); > - copy_iaoq_entry(ctx, cpu_iaoq_b, b, ctx->iaoq_n_var); > + copy_iaq_entry(ctx, &cpu_iaq_f, b); > + copy_iaq_entry(ctx, &cpu_iaq_b, n); > tcg_gen_lookup_and_goto_ptr(); > } > } > @@ -1705,35 +1742,62 @@ static bool do_fop_dedd(DisasContext *ctx, unsigned rt, > static bool do_dbranch(DisasContext *ctx, uint64_t dest, > unsigned link, bool is_n) > { > + IAQEntry n1, n2; > + > if (ctx->null_cond.c == TCG_COND_NEVER && ctx->null_lab == NULL) { > if (link != 0) { > - copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var); > + copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n); > } > - ctx->iaoq_n = dest; > + > + /* Perform an unconditional branch by changing IAQ_Next. */ > + n1.space = NULL; > + n1.ofsv = NULL; > + n1.ofsc = dest; > + copy_iaq_entry(ctx, &ctx->iaq_n, &n1); > + > if (is_n) { > ctx->null_cond.c = TCG_COND_ALWAYS; > } > - } else { > - nullify_over(ctx); > - > - if (link != 0) { > - copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var); > - } > - > - if (is_n && use_nullify_skip(ctx)) { > - nullify_set(ctx, 0); > - gen_goto_tb(ctx, 0, dest, dest + 4); > - } else { > - nullify_set(ctx, is_n); > - gen_goto_tb(ctx, 0, ctx->iaoq_b, dest); > - } > - > - nullify_end(ctx); > - > - nullify_set(ctx, 0); > - gen_goto_tb(ctx, 1, ctx->iaoq_b, ctx->iaoq_n); > - ctx->base.is_jmp = DISAS_NORETURN; > + return true; > } > + > + /* Conditional branch by virtue of nullification. */ > + nullify_over(ctx); > + > + if (link != 0) { > + copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n); > + } > + > + /* > + * Retain any space change already in the queue. > + * This can happen with B in delay slot of BE. > + */ > + if (ctx->iaq_n.space) { > + n1.space = n2.space = ctx->iaq_n.space; > + n1.ofsv = tcg_constant_i64(dest); > + n2.ofsv = tcg_constant_i64(dest + 4); > + n1.ofsc = n2.ofsc = 0; > + } else { > + n1.space = n2.space = NULL; > + n1.ofsv = n2.ofsv = NULL; > + n1.ofsc = dest; > + n2.ofsc = dest + 4; > + } > + > + if (is_n && use_nullify_skip(ctx)) { > + nullify_set(ctx, 0); > + gen_goto_tb(ctx, 0, &n1, &n2); > + } else { > + nullify_set(ctx, is_n); > + gen_goto_tb(ctx, 0, &ctx->iaq_b, &n1); > + } > + > + nullify_end(ctx); > + > + /* Branch not taken. */ > + nullify_set(ctx, 0); > + gen_goto_tb(ctx, 1, &ctx->iaq_b, &ctx->iaq_n); > + ctx->base.is_jmp = DISAS_NORETURN; > return true; > } > > @@ -1745,17 +1809,15 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n, > uint64_t dest = iaoq_dest(ctx, disp); > TCGLabel *taken = NULL; > TCGCond c = cond->c; > + IAQEntry n1, n2; > bool n; > > assert(ctx->null_cond.c == TCG_COND_NEVER); > > - /* Handle TRUE and NEVER as direct branches. */ > + /* Handle ALWAYS as a direct branch. */ > if (c == TCG_COND_ALWAYS) { > return do_dbranch(ctx, dest, 0, is_n && disp >= 0); > } > - if (c == TCG_COND_NEVER) { > - return do_dbranch(ctx, ctx->iaoq_n, 0, is_n && disp < 0); > - } > > taken = gen_new_label(); > tcg_gen_brcond_i64(c, cond->a0, cond->a1, taken); > @@ -1763,33 +1825,40 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n, > > /* Not taken: Condition not satisfied; nullify on backward branches. */ > n = is_n && disp < 0; > - if (n && use_nullify_skip(ctx)) { > - nullify_set(ctx, 0); > - gen_goto_tb(ctx, 0, ctx->iaoq_n, ctx->iaoq_n + 4); > - } else { > - if (!n && ctx->null_lab) { > - gen_set_label(ctx->null_lab); > - ctx->null_lab = NULL; > - } > - nullify_set(ctx, n); > - if (ctx->iaoq_n == -1) { > - /* The temporary iaoq_n_var died at the branch above. > - Regenerate it here instead of saving it. */ > - tcg_gen_addi_i64(ctx->iaoq_n_var, cpu_iaoq_b, 4); > - } > - gen_goto_tb(ctx, 0, ctx->iaoq_b, ctx->iaoq_n); > + if (!n && ctx->null_lab) { > + gen_set_label(ctx->null_lab); > + ctx->null_lab = NULL; > } > + nullify_set(ctx, n); > + gen_goto_tb(ctx, 0, &ctx->iaq_b, &ctx->iaq_n); > > gen_set_label(taken); > > /* Taken: Condition satisfied; nullify on forward branches. */ > n = is_n && disp >= 0; > + > + /* > + * Retain any space change already in the queue. > + * This can happen with ADDB in delay slot of BE. > + */ > + if (ctx->iaq_n.space) { > + n1.space = n2.space = ctx->iaq_n.space; > + n1.ofsv = tcg_constant_i64(dest); > + n2.ofsv = tcg_constant_i64(dest + 4); > + n1.ofsc = n2.ofsc = 0; > + } else { > + n1.space = n2.space = NULL; > + n1.ofsv = n2.ofsv = NULL; > + n1.ofsc = dest; > + n2.ofsc = dest + 4; > + } > + > if (n && use_nullify_skip(ctx)) { > nullify_set(ctx, 0); > - gen_goto_tb(ctx, 1, dest, dest + 4); > + gen_goto_tb(ctx, 1, &n1, &n2); > } else { > nullify_set(ctx, n); > - gen_goto_tb(ctx, 1, ctx->iaoq_b, dest); > + gen_goto_tb(ctx, 1, &ctx->iaq_b, &n1); > } > > /* Not taken: the branch itself was nullified. */ > @@ -1808,84 +1877,59 @@ static bool do_cbranch(DisasContext *ctx, int64_t disp, bool is_n, > static bool do_ibranch(DisasContext *ctx, TCGv_i64 dest, > unsigned link, bool is_n) > { > - TCGv_i64 a0, a1, next, tmp; > - TCGCond c; > + IAQEntry next = { > + .space = ctx->iaq_n.space, > + .ofsv = tcg_temp_new_i64() > + }; > > assert(ctx->null_lab == NULL); > > if (ctx->null_cond.c == TCG_COND_NEVER) { > + tcg_gen_mov_i64(next.ofsv, dest); > + > if (link != 0) { > - copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var); > + copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n); > } > - next = tcg_temp_new_i64(); > - tcg_gen_mov_i64(next, dest); > if (is_n) { > if (use_nullify_skip(ctx)) { > - copy_iaoq_entry(ctx, cpu_iaoq_f, -1, next); > - tcg_gen_addi_i64(next, next, 4); > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, next); > + copy_iaq_entry(ctx, &cpu_iaq_f, &next); > + tcg_gen_addi_i64(next.ofsv, next.ofsv, 4); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > nullify_set(ctx, 0); > ctx->base.is_jmp = DISAS_IAQ_N_UPDATED; > return true; > } > ctx->null_cond.c = TCG_COND_ALWAYS; > } > - ctx->iaoq_n = -1; > - ctx->iaoq_n_var = next; > - } else if (is_n && use_nullify_skip(ctx)) { > + ctx->iaq_n = next; > + return true; > + } > + > + nullify_over(ctx); > + > + tcg_gen_mov_i64(next.ofsv, dest); > + if (link != 0) { > + copy_iaq_offset(ctx, cpu_gr[link], &ctx->iaq_n); > + } > + > + if (is_n && use_nullify_skip(ctx)) { > /* The (conditional) branch, B, nullifies the next insn, N, > and we're allowed to skip execution N (no single-step or > tracepoint in effect). Since the goto_ptr that we must use > for the indirect branch consumes no special resources, we > can (conditionally) skip B and continue execution. */ > - /* The use_nullify_skip test implies we have a known control path. */ > - tcg_debug_assert(ctx->iaoq_b != -1); > - tcg_debug_assert(ctx->iaoq_n != -1); > - > - /* We do have to handle the non-local temporary, DEST, before > - branching. Since IOAQ_F is not really live at this point, we > - can simply store DEST optimistically. Similarly with IAOQ_B. */ > - copy_iaoq_entry(ctx, cpu_iaoq_f, -1, dest); > - next = tcg_temp_new_i64(); > - tcg_gen_addi_i64(next, dest, 4); > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, next); > - > - nullify_over(ctx); > - if (link != 0) { > - copy_iaoq_entry(ctx, cpu_gr[link], ctx->iaoq_n, ctx->iaoq_n_var); > - } > - tcg_gen_lookup_and_goto_ptr(); > - return nullify_end(ctx); > + copy_iaq_entry(ctx, &cpu_iaq_f, &next); > + tcg_gen_addi_i64(next.ofsv, next.ofsv, 4); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > + nullify_set(ctx, 0); > } else { > - c = ctx->null_cond.c; > - a0 = ctx->null_cond.a0; > - a1 = ctx->null_cond.a1; > - > - tmp = tcg_temp_new_i64(); > - next = tcg_temp_new_i64(); > - > - copy_iaoq_entry(ctx, tmp, ctx->iaoq_n, ctx->iaoq_n_var); > - tcg_gen_movcond_i64(c, next, a0, a1, tmp, dest); > - ctx->iaoq_n = -1; > - ctx->iaoq_n_var = next; > - > - if (link != 0) { > - tcg_gen_movcond_i64(c, cpu_gr[link], a0, a1, cpu_gr[link], tmp); > - } > - > - if (is_n) { > - /* The branch nullifies the next insn, which means the state of N > - after the branch is the inverse of the state of N that applied > - to the branch. */ > - tcg_gen_setcond_i64(tcg_invert_cond(c), cpu_psw_n, a0, a1); > - cond_free(&ctx->null_cond); > - ctx->null_cond = cond_make_n(); > - ctx->psw_n_nonzero = true; > - } else { > - cond_free(&ctx->null_cond); > - } > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > + nullify_set(ctx, is_n); > } > - return true; > + tcg_gen_lookup_and_goto_ptr(); > + ctx->base.is_jmp = DISAS_NORETURN; > + return nullify_end(ctx); > } > > /* Implement > @@ -1927,7 +1971,7 @@ static TCGv_i64 do_ibranch_priv(DisasContext *ctx, TCGv_i64 offset) > aforementioned BE. */ > static void do_page_zero(DisasContext *ctx) > { > - TCGv_i64 tmp; > + IAQEntry next = { }; > > /* If by some means we get here with PSW[N]=1, that implies that > the B,GATE instruction would be skipped, and we'd fault on the > @@ -1948,11 +1992,11 @@ static void do_page_zero(DisasContext *ctx) > non-sequential instruction execution. Normally the PSW[B] bit > detects this by disallowing the B,GATE instruction to execute > under such conditions. */ > - if (ctx->iaoq_b != ctx->iaoq_f + 4) { > + if (ctx->iaq_b.ofsv || ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4) { > goto do_sigill; > } > > - switch (ctx->iaoq_f & -4) { > + switch (ctx->iaq_f.ofsc & -4) { > case 0x00: /* Null pointer call */ > gen_excp_1(EXCP_IMP); > ctx->base.is_jmp = DISAS_NORETURN; > @@ -1965,11 +2009,11 @@ static void do_page_zero(DisasContext *ctx) > > case 0xe0: /* SET_THREAD_POINTER */ > tcg_gen_st_i64(cpu_gr[26], tcg_env, offsetof(CPUHPPAState, cr[27])); > - tmp = tcg_temp_new_i64(); > - tcg_gen_ori_i64(tmp, cpu_gr[31], 3); > - copy_iaoq_entry(ctx, cpu_iaoq_f, -1, tmp); > - tcg_gen_addi_i64(tmp, tmp, 4); > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp); > + next.ofsv = tcg_temp_new_i64(); > + tcg_gen_ori_i64(next.ofsv, cpu_gr[31], 3); > + copy_iaq_entry(ctx, &cpu_iaq_f, &next); > + tcg_gen_addi_i64(next.ofsv, next.ofsv, 4); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > ctx->base.is_jmp = DISAS_IAQ_N_UPDATED; > break; > > @@ -2011,7 +2055,8 @@ static bool trans_mfia(DisasContext *ctx, arg_mfia *a) > { > unsigned rt = a->t; > TCGv_i64 tmp = dest_gpr(ctx, rt); > - tcg_gen_movi_i64(tmp, ctx->iaoq_f); > + > + tcg_gen_movi_i64(tmp, ctx->iaq_f.ofsc); > save_gpr(ctx, rt, tmp); > > cond_free(&ctx->null_cond); > @@ -2678,8 +2723,8 @@ static bool trans_or(DisasContext *ctx, arg_rrr_cf_d *a) > nullify_over(ctx); > > /* Advance the instruction queue. */ > - copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b); > - copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_n, ctx->iaoq_n_var); > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b); > + copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_n); > nullify_set(ctx, 0); > > /* Tell the qemu main loop to halt until this cpu has work. */ > @@ -3805,26 +3850,21 @@ static bool trans_be(DisasContext *ctx, arg_be *a) > #ifdef CONFIG_USER_ONLY > return do_ibranch(ctx, tmp, a->l, a->n); > #else > - TCGv_i64 new_spc = tcg_temp_new_i64(); > + IAQEntry next = { .ofsv = tmp, .space = tcg_temp_new_i64() }; > + load_spr(ctx, next.space, a->sp); > > - load_spr(ctx, new_spc, a->sp); > if (a->l) { > - copy_iaoq_entry(ctx, cpu_gr[31], ctx->iaoq_n, ctx->iaoq_n_var); > - tcg_gen_mov_i64(cpu_sr[0], cpu_iasq_f); > + copy_iaq_offset(ctx, cpu_gr[31], &ctx->iaq_n); > + tcg_gen_mov_i64(cpu_sr[0], ctx->iaq_n.space ? : cpu_iaq_b.space); > } > if (a->n && use_nullify_skip(ctx)) { > - copy_iaoq_entry(ctx, cpu_iaoq_f, -1, tmp); > + copy_iaq_entry(ctx, &cpu_iaq_f, &next); > tcg_gen_addi_i64(tmp, tmp, 4); > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp); > - tcg_gen_mov_i64(cpu_iasq_f, new_spc); > - tcg_gen_mov_i64(cpu_iasq_b, cpu_iasq_f); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > + nullify_set(ctx, 0); > } else { > - copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b); > - if (ctx->iaoq_b == -1) { > - tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b); > - } > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, tmp); > - tcg_gen_mov_i64(cpu_iasq_b, new_spc); > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > nullify_set(ctx, a->n); > } > tcg_gen_lookup_and_goto_ptr(); > @@ -3855,7 +3895,7 @@ static bool trans_b_gate(DisasContext *ctx, arg_b_gate *a) > * b evil > * in which instructions at evil would run with increased privs. > */ > - if (ctx->iaoq_b == -1 || ctx->iaoq_b != ctx->iaoq_f + 4) { > + if (ctx->iaq_b.ofsv || ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4) { > return gen_illegal(ctx); > } > > @@ -3897,12 +3937,12 @@ static bool trans_blr(DisasContext *ctx, arg_blr *a) > if (a->x) { > TCGv_i64 tmp = tcg_temp_new_i64(); > tcg_gen_shli_i64(tmp, load_gpr(ctx, a->x), 3); > - tcg_gen_addi_i64(tmp, tmp, ctx->iaoq_f + 8); > + tcg_gen_addi_i64(tmp, tmp, ctx->iaq_f.ofsc + 8); > /* The computation here never changes privilege level. */ > return do_ibranch(ctx, tmp, a->l, a->n); > } else { > /* BLR R0,RX is a good way to load PC+8 into RX. */ > - return do_dbranch(ctx, ctx->iaoq_f + 8, a->l, a->n); > + return do_dbranch(ctx, ctx->iaq_f.ofsc + 8, a->l, a->n); > } > } > > @@ -3923,23 +3963,23 @@ static bool trans_bv(DisasContext *ctx, arg_bv *a) > > static bool trans_bve(DisasContext *ctx, arg_bve *a) > { > - TCGv_i64 dest; > > #ifdef CONFIG_USER_ONLY > - dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b)); > + TCGv_i64 dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b)); > return do_ibranch(ctx, dest, a->l, a->n); > #else > - nullify_over(ctx); > - dest = do_ibranch_priv(ctx, load_gpr(ctx, a->b)); > + IAQEntry next; > > - copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_b, cpu_iaoq_b); > - if (ctx->iaoq_b == -1) { > - tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b); > - } > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, dest); > - tcg_gen_mov_i64(cpu_iasq_b, space_select(ctx, 0, dest)); > + nullify_over(ctx); > + > + next.ofsc = 0; > + next.ofsv = do_ibranch_priv(ctx, load_gpr(ctx, a->b)); > + next.space = space_select(ctx, 0, next.ofsv); > + > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_b); > + copy_iaq_entry(ctx, &cpu_iaq_b, &next); > if (a->l) { > - copy_iaoq_entry(ctx, cpu_gr[a->l], ctx->iaoq_n, ctx->iaoq_n_var); > + copy_iaq_offset(ctx, cpu_gr[a->l], &ctx->iaq_n); > } > nullify_set(ctx, a->n); > tcg_gen_lookup_and_goto_ptr(); > @@ -4489,8 +4529,8 @@ static void hppa_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) > #ifdef CONFIG_USER_ONLY > ctx->privilege = MMU_IDX_TO_PRIV(MMU_USER_IDX); > ctx->mmu_idx = MMU_USER_IDX; > - ctx->iaoq_f = ctx->base.pc_first | ctx->privilege; > - ctx->iaoq_b = ctx->base.tb->cs_base | ctx->privilege; > + ctx->iaq_f.ofsc = ctx->base.pc_first | ctx->privilege; > + ctx->iaq_b.ofsc = ctx->base.tb->cs_base | ctx->privilege; > ctx->unalign = (ctx->tb_flags & TB_FLAG_UNALIGN ? MO_UNALN : MO_ALIGN); > #else > ctx->privilege = (ctx->tb_flags >> TB_FLAG_PRIV_SHIFT) & 3; > @@ -4503,12 +4543,18 @@ static void hppa_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) > uint64_t iasq_f = cs_base & ~0xffffffffull; > int32_t diff = cs_base; > > - ctx->iaoq_f = (ctx->base.pc_first & ~iasq_f) + ctx->privilege; > - ctx->iaoq_b = (diff ? ctx->iaoq_f + diff : -1); > -#endif > - ctx->iaoq_n = -1; > - ctx->iaoq_n_var = NULL; > + ctx->iaq_f.ofsc = (ctx->base.pc_first & ~iasq_f) + ctx->privilege; > + ctx->iaq_f.ofsv = NULL; > + ctx->iaq_f.space = NULL; > > + if (diff) { > + ctx->iaq_b.ofsc = ctx->iaq_f.ofsc + diff; > + ctx->iaq_b.ofsv = NULL; > + ctx->iaq_b.space = NULL; > + } else { > + ctx->iaq_b = cpu_iaq_b; > + } > +#endif > ctx->zero = tcg_constant_i64(0); > > /* Bound the number of instructions by those left on the page. */ > @@ -4533,8 +4579,13 @@ static void hppa_tr_tb_start(DisasContextBase *dcbase, CPUState *cs) > static void hppa_tr_insn_start(DisasContextBase *dcbase, CPUState *cs) > { > DisasContext *ctx = container_of(dcbase, DisasContext, base); > + uint64_t b_ofs; > > - tcg_gen_insn_start(ctx->iaoq_f, ctx->iaoq_b, 0); > + assert(ctx->iaq_f.ofsv == NULL); > + assert(ctx->iaq_f.space == NULL); > + assert(ctx->iaq_f.ofsc != -1); > + b_ofs = ctx->iaq_b.ofsv ? -1 : ctx->iaq_b.ofsc; > + tcg_gen_insn_start(ctx->iaq_f.ofsc, b_ofs, 0); > ctx->insn_start = tcg_last_op(); > } > > @@ -4559,13 +4610,16 @@ static void hppa_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) > > /* Set up the IA queue for the next insn. > This will be overwritten by a branch. */ > - if (ctx->iaoq_b == -1) { > - ctx->iaoq_n = -1; > - ctx->iaoq_n_var = tcg_temp_new_i64(); > - tcg_gen_addi_i64(ctx->iaoq_n_var, cpu_iaoq_b, 4); > + if (ctx->iaq_b.ofsv) { > + ctx->iaq_n.ofsc = 0; > + ctx->iaq_n.ofsv = tcg_temp_new_i64(); > + tcg_gen_addi_i64(ctx->iaq_n.ofsv, ctx->iaq_b.ofsv, 4); > + ctx->iaq_n.space = ctx->iaq_b.space; > } else { > - ctx->iaoq_n = ctx->iaoq_b + 4; > - ctx->iaoq_n_var = NULL; > + assert(ctx->iaq_b.space == NULL); > + ctx->iaq_n.ofsc = ctx->iaq_b.ofsc + 4; > + ctx->iaq_n.ofsv = NULL; > + ctx->iaq_n.space = NULL; > } > > if (unlikely(ctx->null_cond.c == TCG_COND_ALWAYS)) { > @@ -4583,48 +4637,22 @@ static void hppa_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) > > /* Advance the insn queue. Note that this check also detects > a priority change within the instruction queue. */ > - if (ret == DISAS_NEXT && ctx->iaoq_b != ctx->iaoq_f + 4) { > - if (ctx->iaoq_b != -1 && ctx->iaoq_n != -1 > - && use_goto_tb(ctx, ctx->iaoq_b) > + if (ret == DISAS_NEXT > + && (ctx->iaq_b.ofsv != NULL || > + ctx->iaq_b.ofsc != ctx->iaq_f.ofsc + 4)) { > + if (use_goto_tb(ctx, &ctx->iaq_b, &ctx->iaq_n) > && (ctx->null_cond.c == TCG_COND_NEVER > || ctx->null_cond.c == TCG_COND_ALWAYS)) { > nullify_set(ctx, ctx->null_cond.c == TCG_COND_ALWAYS); > - gen_goto_tb(ctx, 0, ctx->iaoq_b, ctx->iaoq_n); > + gen_goto_tb(ctx, 0, &ctx->iaq_b, &ctx->iaq_n); > ctx->base.is_jmp = ret = DISAS_NORETURN; > } else { > ctx->base.is_jmp = ret = DISAS_IAQ_N_STALE; > } > } > - ctx->iaoq_f = ctx->iaoq_b; > - ctx->iaoq_b = ctx->iaoq_n; > + ctx->iaq_f = ctx->iaq_b; > + ctx->iaq_b = ctx->iaq_n; > ctx->base.pc_next += 4; > - > - switch (ret) { > - case DISAS_NORETURN: > - case DISAS_IAQ_N_UPDATED: > - break; > - > - case DISAS_NEXT: > - case DISAS_IAQ_N_STALE: > - case DISAS_IAQ_N_STALE_EXIT: > - if (ctx->iaoq_f == -1) { > - copy_iaoq_entry(ctx, cpu_iaoq_f, -1, cpu_iaoq_b); > - copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_n, ctx->iaoq_n_var); > -#ifndef CONFIG_USER_ONLY > - tcg_gen_mov_i64(cpu_iasq_f, cpu_iasq_b); > -#endif > - nullify_save(ctx); > - ctx->base.is_jmp = (ret == DISAS_IAQ_N_STALE_EXIT > - ? DISAS_EXIT > - : DISAS_IAQ_N_UPDATED); > - } else if (ctx->iaoq_b == -1) { > - copy_iaoq_entry(ctx, cpu_iaoq_b, -1, ctx->iaoq_n_var); > - } > - break; > - > - default: > - g_assert_not_reached(); > - } > } > > static void hppa_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) > @@ -4638,8 +4666,8 @@ static void hppa_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) > case DISAS_TOO_MANY: > case DISAS_IAQ_N_STALE: > case DISAS_IAQ_N_STALE_EXIT: > - copy_iaoq_entry(ctx, cpu_iaoq_f, ctx->iaoq_f, cpu_iaoq_f); > - copy_iaoq_entry(ctx, cpu_iaoq_b, ctx->iaoq_b, cpu_iaoq_b); > + copy_iaq_entry(ctx, &cpu_iaq_f, &ctx->iaq_f); > + copy_iaq_entry(ctx, &cpu_iaq_b, &ctx->iaq_b); > nullify_save(ctx); > /* FALLTHRU */ > case DISAS_IAQ_N_UPDATED: > @@ -4694,6 +4722,6 @@ static const TranslatorOps hppa_tr_ops = { > void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns, > vaddr pc, void *host_pc) > { > - DisasContext ctx; > + DisasContext ctx = { }; > translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base); > }
Sven Schnelle <svens@stackframe.org> writes: > Hi Richard, > > Richard Henderson <richard.henderson@linaro.org> writes: > >> Wrap offset and space into a single structure, with >> offset represented either as a constant or as variable. >> This enhances copy_iaoq_entry to copy the space, and at >> the same time simplifying the code. >> >> Signed-off-by: Richard Henderson <richard.henderson@linaro.org> >> --- >> >> Sven, I think this might solve your branch-with-space-change issue. >> There are a couple of places where it does appear as if we might be >> losing a space change, with a branch within a branch delay slot. >> I'm trying to solve this by keeping both queue entries together. > > Thanks! I've put that on top of the other changes i have for 64bit > HP-UX, but it introduces a new problem: > > * Checking configuration for consistency... > WARNING: The disk at: 10/0/0/0.0.0 (QEMU_QEMU_HARDDISK) appears to contain a > file system and boot area. Continuing the installation will destroy > any existing data on this disk. > > Do you wish to cancel the non-interactive installation in order to > respond to the warnings above? ([y]/n): n > > * Loading configuration utility... > * Beginning installation from source: /dev/dsk/c0t2d0 > ======= 03/19/24 13:59:53 EST Starting system configuration... > * Configure_Disks: Begin > * Will install B.11.11 onto this system. > * Creating LVM physical volume "/dev/rdsk/c0t0d0" (10/0/0/0.0.0). > * Creating volume group "vg00". > vgcreate: Illegal path "". > * Creating logical volume "vg00/lvol1" (/stand). > lvcreate: Illegal path "". > * Creating logical volume "vg00/lvol2" (swap_dump). > lvcreate: Illegal path "". > * Creating logical volume "vg00/lvol3" (/). > lvcreate: Illegal path "". > * Creating logical volume "vg00/lvol4" (/tmp). > lvcreate: Illegal path "". > * Creating logical volume "vg00/lvol5" (/home). > lvcreate: Illegal path "". > * Creating logical volume "vg00/lvol6" (/opt). > lvcreate: Illegal path "". > > The 'illegal path' errors where not present before. I haven't done any > debugging, as debugging hppa userspace in qemu is a really time > consuming thing. I'm seeing the same error with 32bit HP-UX 10.20 - so at least it's not caused by my 64bit hacks. Linux seems to run fine, but Linux is also utilizing address spaces not to the same extent as HP-UX.
© 2016 - 2024 Red Hat, Inc.