Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
accel/tcg/cputlb.c | 151 ++++++++++++++++++++++-----------------------
1 file changed, 74 insertions(+), 77 deletions(-)
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index a33bebf55a..8f459be5a8 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1684,10 +1684,7 @@ bool tlb_plugin_lookup(CPUState *cpu, vaddr addr, int mmu_idx,
*/
typedef struct MMULookupPageData {
- CPUTLBEntryFull *full;
- void *haddr;
vaddr addr;
- int flags;
int size;
TLBLookupOutput o;
} MMULookupPageData;
@@ -1724,10 +1721,6 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop,
};
tlb_lookup_nofail(cpu, &data->o, &i);
-
- data->full = &data->o.full;
- data->flags = data->o.flags;
- data->haddr = data->o.haddr;
}
/**
@@ -1743,24 +1736,22 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop,
static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data,
MMUAccessType access_type, uintptr_t ra)
{
- CPUTLBEntryFull *full = data->full;
- vaddr addr = data->addr;
- int flags = data->flags;
- int size = data->size;
+ int flags = data->o.flags;
/* On watchpoint hit, this will longjmp out. */
if (flags & TLB_WATCHPOINT) {
int wp = access_type == MMU_DATA_STORE ? BP_MEM_WRITE : BP_MEM_READ;
- cpu_check_watchpoint(cpu, addr, size, full->attrs, wp, ra);
+ cpu_check_watchpoint(cpu, data->addr, data->size,
+ data->o.full.attrs, wp, ra);
flags &= ~TLB_WATCHPOINT;
}
/* Note that notdirty is only set for writes. */
if (flags & TLB_NOTDIRTY) {
- notdirty_write(cpu, addr, size, full, ra);
+ notdirty_write(cpu, data->addr, data->size, &data->o.full, ra);
flags &= ~TLB_NOTDIRTY;
}
- data->flags = flags;
+ data->o.flags = flags;
}
/**
@@ -1795,7 +1786,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
if (likely(!crosspage)) {
mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
- flags = l->page[0].flags;
+ flags = l->page[0].o.flags;
if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) {
mmu_watch_or_dirty(cpu, &l->page[0], type, ra);
}
@@ -1812,7 +1803,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra);
- flags = l->page[0].flags | l->page[1].flags;
+ flags = l->page[0].o.flags | l->page[1].o.flags;
if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) {
mmu_watch_or_dirty(cpu, &l->page[0], type, ra);
mmu_watch_or_dirty(cpu, &l->page[1], type, ra);
@@ -2029,7 +2020,7 @@ static Int128 do_ld16_mmio_beN(CPUState *cpu, CPUTLBEntryFull *full,
*/
static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be)
{
- uint8_t *haddr = p->haddr;
+ uint8_t *haddr = p->o.haddr;
int i, size = p->size;
for (i = 0; i < size; i++) {
@@ -2047,7 +2038,7 @@ static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be)
*/
static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be)
{
- void *haddr = p->haddr;
+ void *haddr = p->o.haddr;
int size = p->size;
do {
@@ -2097,7 +2088,7 @@ static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be)
static uint64_t do_ld_whole_be4(MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 3;
- uint32_t x = load_atomic4(p->haddr - o);
+ uint32_t x = load_atomic4(p->o.haddr - o);
x = cpu_to_be32(x);
x <<= o * 8;
@@ -2117,7 +2108,7 @@ static uint64_t do_ld_whole_be8(CPUState *cpu, uintptr_t ra,
MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 7;
- uint64_t x = load_atomic8_or_exit(cpu, ra, p->haddr - o);
+ uint64_t x = load_atomic8_or_exit(cpu, ra, p->o.haddr - o);
x = cpu_to_be64(x);
x <<= o * 8;
@@ -2137,7 +2128,7 @@ static Int128 do_ld_whole_be16(CPUState *cpu, uintptr_t ra,
MMULookupPageData *p, uint64_t ret_be)
{
int o = p->addr & 15;
- Int128 x, y = load_atomic16_or_exit(cpu, ra, p->haddr - o);
+ Int128 x, y = load_atomic16_or_exit(cpu, ra, p->o.haddr - o);
int size = p->size;
if (!HOST_BIG_ENDIAN) {
@@ -2160,8 +2151,8 @@ static uint64_t do_ld_beN(CPUState *cpu, MMULookupPageData *p,
MemOp atom;
unsigned tmp, half_size;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld_mmio_beN(cpu, p->full, ret_be, p->addr, p->size,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld_mmio_beN(cpu, &p->o.full, ret_be, p->addr, p->size,
mmu_idx, type, ra);
}
@@ -2210,8 +2201,9 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p,
uint64_t b;
MemOp atom;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld16_mmio_beN(cpu, p->full, a, p->addr, size, mmu_idx, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld16_mmio_beN(cpu, &p->o.full, a, p->addr,
+ size, mmu_idx, ra);
}
/*
@@ -2223,7 +2215,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p,
case MO_ATOM_SUBALIGN:
p->size = size - 8;
a = do_ld_parts_beN(p, a);
- p->haddr += size - 8;
+ p->o.haddr += size - 8;
p->size = 8;
b = do_ld_parts_beN(p, 0);
break;
@@ -2242,7 +2234,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p,
case MO_ATOM_NONE:
p->size = size - 8;
a = do_ld_bytes_beN(p, a);
- b = ldq_be_p(p->haddr + size - 8);
+ b = ldq_be_p(p->o.haddr + size - 8);
break;
default:
@@ -2255,10 +2247,11 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p,
static uint8_t do_ld_1(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
MMUAccessType type, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_ld_mmio_beN(cpu, p->full, 0, p->addr, 1, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 1,
+ mmu_idx, type, ra);
} else {
- return *(uint8_t *)p->haddr;
+ return *(uint8_t *)p->o.haddr;
}
}
@@ -2267,14 +2260,15 @@ static uint16_t do_ld_2(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
{
uint16_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 2, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 2,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap16(ret);
}
} else {
/* Perform the load host endian, then swap if necessary. */
- ret = load_atom_2(cpu, ra, p->haddr, memop);
+ ret = load_atom_2(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap16(ret);
}
@@ -2287,14 +2281,15 @@ static uint32_t do_ld_4(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
{
uint32_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 4, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 4,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap32(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_4(cpu, ra, p->haddr, memop);
+ ret = load_atom_4(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap32(ret);
}
@@ -2307,14 +2302,15 @@ static uint64_t do_ld_8(CPUState *cpu, MMULookupPageData *p, int mmu_idx,
{
uint64_t ret;
- if (unlikely(p->flags & TLB_MMIO)) {
- ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 8, mmu_idx, type, ra);
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 8,
+ mmu_idx, type, ra);
if ((memop & MO_BSWAP) == MO_LE) {
ret = bswap64(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_8(cpu, ra, p->haddr, memop);
+ ret = load_atom_8(cpu, ra, p->o.haddr, memop);
if (memop & MO_BSWAP) {
ret = bswap64(ret);
}
@@ -2414,15 +2410,15 @@ static Int128 do_ld16_mmu(CPUState *cpu, vaddr addr,
cpu_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_LOAD, &l);
if (likely(!crosspage)) {
- if (unlikely(l.page[0].flags & TLB_MMIO)) {
- ret = do_ld16_mmio_beN(cpu, l.page[0].full, 0, addr, 16,
+ if (unlikely(l.page[0].o.flags & TLB_MMIO)) {
+ ret = do_ld16_mmio_beN(cpu, &l.page[0].o.full, 0, addr, 16,
l.mmu_idx, ra);
if ((l.memop & MO_BSWAP) == MO_LE) {
ret = bswap128(ret);
}
} else {
/* Perform the load host endian. */
- ret = load_atom_16(cpu, ra, l.page[0].haddr, l.memop);
+ ret = load_atom_16(cpu, ra, l.page[0].o.haddr, l.memop);
if (l.memop & MO_BSWAP) {
ret = bswap128(ret);
}
@@ -2568,10 +2564,10 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p,
MemOp atom;
unsigned tmp, half_size;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_st_mmio_leN(cpu, p->full, val_le, p->addr,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_st_mmio_leN(cpu, &p->o.full, val_le, p->addr,
p->size, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
return val_le >> (p->size * 8);
}
@@ -2582,7 +2578,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p,
atom = mop & MO_ATOM_MASK;
switch (atom) {
case MO_ATOM_SUBALIGN:
- return store_parts_leN(p->haddr, p->size, val_le);
+ return store_parts_leN(p->o.haddr, p->size, val_le);
case MO_ATOM_IFALIGN_PAIR:
case MO_ATOM_WITHIN16_PAIR:
@@ -2593,9 +2589,9 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p,
? p->size == half_size
: p->size >= half_size) {
if (!HAVE_al8_fast && p->size <= 4) {
- return store_whole_le4(p->haddr, p->size, val_le);
+ return store_whole_le4(p->o.haddr, p->size, val_le);
} else if (HAVE_al8) {
- return store_whole_le8(p->haddr, p->size, val_le);
+ return store_whole_le8(p->o.haddr, p->size, val_le);
} else {
cpu_loop_exit_atomic(cpu, ra);
}
@@ -2605,7 +2601,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p,
case MO_ATOM_IFALIGN:
case MO_ATOM_WITHIN16:
case MO_ATOM_NONE:
- return store_bytes_leN(p->haddr, p->size, val_le);
+ return store_bytes_leN(p->o.haddr, p->size, val_le);
default:
g_assert_not_reached();
@@ -2622,10 +2618,10 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
int size = p->size;
MemOp atom;
- if (unlikely(p->flags & TLB_MMIO)) {
- return do_st16_mmio_leN(cpu, p->full, val_le, p->addr,
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ return do_st16_mmio_leN(cpu, &p->o.full, val_le, p->addr,
size, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
return int128_gethi(val_le) >> ((size - 8) * 8);
}
@@ -2636,8 +2632,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
atom = mop & MO_ATOM_MASK;
switch (atom) {
case MO_ATOM_SUBALIGN:
- store_parts_leN(p->haddr, 8, int128_getlo(val_le));
- return store_parts_leN(p->haddr + 8, p->size - 8,
+ store_parts_leN(p->o.haddr, 8, int128_getlo(val_le));
+ return store_parts_leN(p->o.haddr + 8, p->size - 8,
int128_gethi(val_le));
case MO_ATOM_WITHIN16_PAIR:
@@ -2645,7 +2641,7 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
if (!HAVE_CMPXCHG128) {
cpu_loop_exit_atomic(cpu, ra);
}
- return store_whole_le16(p->haddr, p->size, val_le);
+ return store_whole_le16(p->o.haddr, p->size, val_le);
case MO_ATOM_IFALIGN_PAIR:
/*
@@ -2655,8 +2651,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
case MO_ATOM_IFALIGN:
case MO_ATOM_WITHIN16:
case MO_ATOM_NONE:
- stq_le_p(p->haddr, int128_getlo(val_le));
- return store_bytes_leN(p->haddr + 8, p->size - 8,
+ stq_le_p(p->o.haddr, int128_getlo(val_le));
+ return store_bytes_leN(p->o.haddr + 8, p->size - 8,
int128_gethi(val_le));
default:
@@ -2667,69 +2663,69 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p,
static void do_st_1(CPUState *cpu, MMULookupPageData *p, uint8_t val,
int mmu_idx, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
- do_st_mmio_leN(cpu, p->full, val, p->addr, 1, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 1, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
- *(uint8_t *)p->haddr = val;
+ *(uint8_t *)p->o.haddr = val;
}
}
static void do_st_2(CPUState *cpu, MMULookupPageData *p, uint16_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap16(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 2, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 2, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap16(val);
}
- store_atom_2(cpu, ra, p->haddr, memop, val);
+ store_atom_2(cpu, ra, p->o.haddr, memop, val);
}
}
static void do_st_4(CPUState *cpu, MMULookupPageData *p, uint32_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap32(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 4, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 4, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap32(val);
}
- store_atom_4(cpu, ra, p->haddr, memop, val);
+ store_atom_4(cpu, ra, p->o.haddr, memop, val);
}
}
static void do_st_8(CPUState *cpu, MMULookupPageData *p, uint64_t val,
int mmu_idx, MemOp memop, uintptr_t ra)
{
- if (unlikely(p->flags & TLB_MMIO)) {
+ if (unlikely(p->o.flags & TLB_MMIO)) {
if ((memop & MO_BSWAP) != MO_LE) {
val = bswap64(val);
}
- do_st_mmio_leN(cpu, p->full, val, p->addr, 8, mmu_idx, ra);
- } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) {
+ do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 8, mmu_idx, ra);
+ } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (memop & MO_BSWAP) {
val = bswap64(val);
}
- store_atom_8(cpu, ra, p->haddr, memop, val);
+ store_atom_8(cpu, ra, p->o.haddr, memop, val);
}
}
@@ -2822,19 +2818,20 @@ static void do_st16_mmu(CPUState *cpu, vaddr addr, Int128 val,
cpu_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_STORE, &l);
if (likely(!crosspage)) {
- if (unlikely(l.page[0].flags & TLB_MMIO)) {
+ if (unlikely(l.page[0].o.flags & TLB_MMIO)) {
if ((l.memop & MO_BSWAP) != MO_LE) {
val = bswap128(val);
}
- do_st16_mmio_leN(cpu, l.page[0].full, val, addr, 16, l.mmu_idx, ra);
- } else if (unlikely(l.page[0].flags & TLB_DISCARD_WRITE)) {
+ do_st16_mmio_leN(cpu, &l.page[0].o.full, val, addr,
+ 16, l.mmu_idx, ra);
+ } else if (unlikely(l.page[0].o.flags & TLB_DISCARD_WRITE)) {
/* nothing */
} else {
/* Swap to host endian if necessary, then store. */
if (l.memop & MO_BSWAP) {
val = bswap128(val);
}
- store_atom_16(cpu, ra, l.page[0].haddr, l.memop, val);
+ store_atom_16(cpu, ra, l.page[0].o.haddr, l.memop, val);
}
return;
}
--
2.43.0
On 11/14/24 08:01, Richard Henderson wrote: > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> > --- > accel/tcg/cputlb.c | 151 ++++++++++++++++++++++----------------------- > 1 file changed, 74 insertions(+), 77 deletions(-) > > diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c > index a33bebf55a..8f459be5a8 100644 > --- a/accel/tcg/cputlb.c > +++ b/accel/tcg/cputlb.c > @@ -1684,10 +1684,7 @@ bool tlb_plugin_lookup(CPUState *cpu, vaddr addr, int mmu_idx, > */ > > typedef struct MMULookupPageData { > - CPUTLBEntryFull *full; > - void *haddr; > vaddr addr; > - int flags; > int size; > TLBLookupOutput o; > } MMULookupPageData; > @@ -1724,10 +1721,6 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, > }; > > tlb_lookup_nofail(cpu, &data->o, &i); > - > - data->full = &data->o.full; > - data->flags = data->o.flags; > - data->haddr = data->o.haddr; > } > > /** > @@ -1743,24 +1736,22 @@ static void mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop, > static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data, > MMUAccessType access_type, uintptr_t ra) > { > - CPUTLBEntryFull *full = data->full; > - vaddr addr = data->addr; > - int flags = data->flags; > - int size = data->size; > + int flags = data->o.flags; > > /* On watchpoint hit, this will longjmp out. */ > if (flags & TLB_WATCHPOINT) { > int wp = access_type == MMU_DATA_STORE ? BP_MEM_WRITE : BP_MEM_READ; > - cpu_check_watchpoint(cpu, addr, size, full->attrs, wp, ra); > + cpu_check_watchpoint(cpu, data->addr, data->size, > + data->o.full.attrs, wp, ra); > flags &= ~TLB_WATCHPOINT; > } > > /* Note that notdirty is only set for writes. */ > if (flags & TLB_NOTDIRTY) { > - notdirty_write(cpu, addr, size, full, ra); > + notdirty_write(cpu, data->addr, data->size, &data->o.full, ra); > flags &= ~TLB_NOTDIRTY; > } > - data->flags = flags; > + data->o.flags = flags; > } > > /** > @@ -1795,7 +1786,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, > if (likely(!crosspage)) { > mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); > > - flags = l->page[0].flags; > + flags = l->page[0].o.flags; > if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { > mmu_watch_or_dirty(cpu, &l->page[0], type, ra); > } > @@ -1812,7 +1803,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi, > mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra); > mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra); > > - flags = l->page[0].flags | l->page[1].flags; > + flags = l->page[0].o.flags | l->page[1].o.flags; > if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) { > mmu_watch_or_dirty(cpu, &l->page[0], type, ra); > mmu_watch_or_dirty(cpu, &l->page[1], type, ra); > @@ -2029,7 +2020,7 @@ static Int128 do_ld16_mmio_beN(CPUState *cpu, CPUTLBEntryFull *full, > */ > static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) > { > - uint8_t *haddr = p->haddr; > + uint8_t *haddr = p->o.haddr; > int i, size = p->size; > > for (i = 0; i < size; i++) { > @@ -2047,7 +2038,7 @@ static uint64_t do_ld_bytes_beN(MMULookupPageData *p, uint64_t ret_be) > */ > static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) > { > - void *haddr = p->haddr; > + void *haddr = p->o.haddr; > int size = p->size; > > do { > @@ -2097,7 +2088,7 @@ static uint64_t do_ld_parts_beN(MMULookupPageData *p, uint64_t ret_be) > static uint64_t do_ld_whole_be4(MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 3; > - uint32_t x = load_atomic4(p->haddr - o); > + uint32_t x = load_atomic4(p->o.haddr - o); > > x = cpu_to_be32(x); > x <<= o * 8; > @@ -2117,7 +2108,7 @@ static uint64_t do_ld_whole_be8(CPUState *cpu, uintptr_t ra, > MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 7; > - uint64_t x = load_atomic8_or_exit(cpu, ra, p->haddr - o); > + uint64_t x = load_atomic8_or_exit(cpu, ra, p->o.haddr - o); > > x = cpu_to_be64(x); > x <<= o * 8; > @@ -2137,7 +2128,7 @@ static Int128 do_ld_whole_be16(CPUState *cpu, uintptr_t ra, > MMULookupPageData *p, uint64_t ret_be) > { > int o = p->addr & 15; > - Int128 x, y = load_atomic16_or_exit(cpu, ra, p->haddr - o); > + Int128 x, y = load_atomic16_or_exit(cpu, ra, p->o.haddr - o); > int size = p->size; > > if (!HOST_BIG_ENDIAN) { > @@ -2160,8 +2151,8 @@ static uint64_t do_ld_beN(CPUState *cpu, MMULookupPageData *p, > MemOp atom; > unsigned tmp, half_size; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld_mmio_beN(cpu, p->full, ret_be, p->addr, p->size, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld_mmio_beN(cpu, &p->o.full, ret_be, p->addr, p->size, > mmu_idx, type, ra); > } > > @@ -2210,8 +2201,9 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > uint64_t b; > MemOp atom; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld16_mmio_beN(cpu, p->full, a, p->addr, size, mmu_idx, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld16_mmio_beN(cpu, &p->o.full, a, p->addr, > + size, mmu_idx, ra); > } > > /* > @@ -2223,7 +2215,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_SUBALIGN: > p->size = size - 8; > a = do_ld_parts_beN(p, a); > - p->haddr += size - 8; > + p->o.haddr += size - 8; > p->size = 8; > b = do_ld_parts_beN(p, 0); > break; > @@ -2242,7 +2234,7 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_NONE: > p->size = size - 8; > a = do_ld_bytes_beN(p, a); > - b = ldq_be_p(p->haddr + size - 8); > + b = ldq_be_p(p->o.haddr + size - 8); > break; > > default: > @@ -2255,10 +2247,11 @@ static Int128 do_ld16_beN(CPUState *cpu, MMULookupPageData *p, > static uint8_t do_ld_1(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > MMUAccessType type, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_ld_mmio_beN(cpu, p->full, 0, p->addr, 1, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 1, > + mmu_idx, type, ra); > } else { > - return *(uint8_t *)p->haddr; > + return *(uint8_t *)p->o.haddr; > } > } > > @@ -2267,14 +2260,15 @@ static uint16_t do_ld_2(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint16_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 2, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 2, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap16(ret); > } > } else { > /* Perform the load host endian, then swap if necessary. */ > - ret = load_atom_2(cpu, ra, p->haddr, memop); > + ret = load_atom_2(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap16(ret); > } > @@ -2287,14 +2281,15 @@ static uint32_t do_ld_4(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint32_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 4, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 4, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap32(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_4(cpu, ra, p->haddr, memop); > + ret = load_atom_4(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap32(ret); > } > @@ -2307,14 +2302,15 @@ static uint64_t do_ld_8(CPUState *cpu, MMULookupPageData *p, int mmu_idx, > { > uint64_t ret; > > - if (unlikely(p->flags & TLB_MMIO)) { > - ret = do_ld_mmio_beN(cpu, p->full, 0, p->addr, 8, mmu_idx, type, ra); > + if (unlikely(p->o.flags & TLB_MMIO)) { > + ret = do_ld_mmio_beN(cpu, &p->o.full, 0, p->addr, 8, > + mmu_idx, type, ra); > if ((memop & MO_BSWAP) == MO_LE) { > ret = bswap64(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_8(cpu, ra, p->haddr, memop); > + ret = load_atom_8(cpu, ra, p->o.haddr, memop); > if (memop & MO_BSWAP) { > ret = bswap64(ret); > } > @@ -2414,15 +2410,15 @@ static Int128 do_ld16_mmu(CPUState *cpu, vaddr addr, > cpu_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD); > crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_LOAD, &l); > if (likely(!crosspage)) { > - if (unlikely(l.page[0].flags & TLB_MMIO)) { > - ret = do_ld16_mmio_beN(cpu, l.page[0].full, 0, addr, 16, > + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { > + ret = do_ld16_mmio_beN(cpu, &l.page[0].o.full, 0, addr, 16, > l.mmu_idx, ra); > if ((l.memop & MO_BSWAP) == MO_LE) { > ret = bswap128(ret); > } > } else { > /* Perform the load host endian. */ > - ret = load_atom_16(cpu, ra, l.page[0].haddr, l.memop); > + ret = load_atom_16(cpu, ra, l.page[0].o.haddr, l.memop); > if (l.memop & MO_BSWAP) { > ret = bswap128(ret); > } > @@ -2568,10 +2564,10 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > MemOp atom; > unsigned tmp, half_size; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_st_mmio_leN(cpu, p->full, val_le, p->addr, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_st_mmio_leN(cpu, &p->o.full, val_le, p->addr, > p->size, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > return val_le >> (p->size * 8); > } > > @@ -2582,7 +2578,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > atom = mop & MO_ATOM_MASK; > switch (atom) { > case MO_ATOM_SUBALIGN: > - return store_parts_leN(p->haddr, p->size, val_le); > + return store_parts_leN(p->o.haddr, p->size, val_le); > > case MO_ATOM_IFALIGN_PAIR: > case MO_ATOM_WITHIN16_PAIR: > @@ -2593,9 +2589,9 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > ? p->size == half_size > : p->size >= half_size) { > if (!HAVE_al8_fast && p->size <= 4) { > - return store_whole_le4(p->haddr, p->size, val_le); > + return store_whole_le4(p->o.haddr, p->size, val_le); > } else if (HAVE_al8) { > - return store_whole_le8(p->haddr, p->size, val_le); > + return store_whole_le8(p->o.haddr, p->size, val_le); > } else { > cpu_loop_exit_atomic(cpu, ra); > } > @@ -2605,7 +2601,7 @@ static uint64_t do_st_leN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_IFALIGN: > case MO_ATOM_WITHIN16: > case MO_ATOM_NONE: > - return store_bytes_leN(p->haddr, p->size, val_le); > + return store_bytes_leN(p->o.haddr, p->size, val_le); > > default: > g_assert_not_reached(); > @@ -2622,10 +2618,10 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > int size = p->size; > MemOp atom; > > - if (unlikely(p->flags & TLB_MMIO)) { > - return do_st16_mmio_leN(cpu, p->full, val_le, p->addr, > + if (unlikely(p->o.flags & TLB_MMIO)) { > + return do_st16_mmio_leN(cpu, &p->o.full, val_le, p->addr, > size, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > return int128_gethi(val_le) >> ((size - 8) * 8); > } > > @@ -2636,8 +2632,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > atom = mop & MO_ATOM_MASK; > switch (atom) { > case MO_ATOM_SUBALIGN: > - store_parts_leN(p->haddr, 8, int128_getlo(val_le)); > - return store_parts_leN(p->haddr + 8, p->size - 8, > + store_parts_leN(p->o.haddr, 8, int128_getlo(val_le)); > + return store_parts_leN(p->o.haddr + 8, p->size - 8, > int128_gethi(val_le)); > > case MO_ATOM_WITHIN16_PAIR: > @@ -2645,7 +2641,7 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > if (!HAVE_CMPXCHG128) { > cpu_loop_exit_atomic(cpu, ra); > } > - return store_whole_le16(p->haddr, p->size, val_le); > + return store_whole_le16(p->o.haddr, p->size, val_le); > > case MO_ATOM_IFALIGN_PAIR: > /* > @@ -2655,8 +2651,8 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > case MO_ATOM_IFALIGN: > case MO_ATOM_WITHIN16: > case MO_ATOM_NONE: > - stq_le_p(p->haddr, int128_getlo(val_le)); > - return store_bytes_leN(p->haddr + 8, p->size - 8, > + stq_le_p(p->o.haddr, int128_getlo(val_le)); > + return store_bytes_leN(p->o.haddr + 8, p->size - 8, > int128_gethi(val_le)); > > default: > @@ -2667,69 +2663,69 @@ static uint64_t do_st16_leN(CPUState *cpu, MMULookupPageData *p, > static void do_st_1(CPUState *cpu, MMULookupPageData *p, uint8_t val, > int mmu_idx, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > - do_st_mmio_leN(cpu, p->full, val, p->addr, 1, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 1, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > - *(uint8_t *)p->haddr = val; > + *(uint8_t *)p->o.haddr = val; > } > } > > static void do_st_2(CPUState *cpu, MMULookupPageData *p, uint16_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap16(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 2, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 2, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap16(val); > } > - store_atom_2(cpu, ra, p->haddr, memop, val); > + store_atom_2(cpu, ra, p->o.haddr, memop, val); > } > } > > static void do_st_4(CPUState *cpu, MMULookupPageData *p, uint32_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap32(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 4, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 4, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap32(val); > } > - store_atom_4(cpu, ra, p->haddr, memop, val); > + store_atom_4(cpu, ra, p->o.haddr, memop, val); > } > } > > static void do_st_8(CPUState *cpu, MMULookupPageData *p, uint64_t val, > int mmu_idx, MemOp memop, uintptr_t ra) > { > - if (unlikely(p->flags & TLB_MMIO)) { > + if (unlikely(p->o.flags & TLB_MMIO)) { > if ((memop & MO_BSWAP) != MO_LE) { > val = bswap64(val); > } > - do_st_mmio_leN(cpu, p->full, val, p->addr, 8, mmu_idx, ra); > - } else if (unlikely(p->flags & TLB_DISCARD_WRITE)) { > + do_st_mmio_leN(cpu, &p->o.full, val, p->addr, 8, mmu_idx, ra); > + } else if (unlikely(p->o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (memop & MO_BSWAP) { > val = bswap64(val); > } > - store_atom_8(cpu, ra, p->haddr, memop, val); > + store_atom_8(cpu, ra, p->o.haddr, memop, val); > } > } > > @@ -2822,19 +2818,20 @@ static void do_st16_mmu(CPUState *cpu, vaddr addr, Int128 val, > cpu_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST); > crosspage = mmu_lookup(cpu, addr, oi, ra, MMU_DATA_STORE, &l); > if (likely(!crosspage)) { > - if (unlikely(l.page[0].flags & TLB_MMIO)) { > + if (unlikely(l.page[0].o.flags & TLB_MMIO)) { > if ((l.memop & MO_BSWAP) != MO_LE) { > val = bswap128(val); > } > - do_st16_mmio_leN(cpu, l.page[0].full, val, addr, 16, l.mmu_idx, ra); > - } else if (unlikely(l.page[0].flags & TLB_DISCARD_WRITE)) { > + do_st16_mmio_leN(cpu, &l.page[0].o.full, val, addr, > + 16, l.mmu_idx, ra); > + } else if (unlikely(l.page[0].o.flags & TLB_DISCARD_WRITE)) { > /* nothing */ > } else { > /* Swap to host endian if necessary, then store. */ > if (l.memop & MO_BSWAP) { > val = bswap128(val); > } > - store_atom_16(cpu, ra, l.page[0].haddr, l.memop, val); > + store_atom_16(cpu, ra, l.page[0].o.haddr, l.memop, val); > } > return; > } Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
© 2016 - 2024 Red Hat, Inc.