With hardware PTW feature enabled, Present) bit and Write bit is checked
by hardware, rather Valid bit and Dirty bit. Bit P means that the page is
valid and present, and bit W means that the page is writable.
The original V bit is treated as access bit, hardware sets this bit if
there is read or write access. Bit D bit is updated by hardware if there
is write access.
Signed-off-by: Bibo Mao <maobibo@loongson.cn>
---
target/loongarch/cpu-csr.h | 2 ++
target/loongarch/cpu-mmu.h | 31 +++++++++++++++++++++++++++++++
target/loongarch/cpu_helper.c | 4 ++--
target/loongarch/tcg/tlb_helper.c | 14 +++++++-------
4 files changed, 42 insertions(+), 9 deletions(-)
diff --git a/target/loongarch/cpu-csr.h b/target/loongarch/cpu-csr.h
index 043b2f14b9..345a51d04a 100644
--- a/target/loongarch/cpu-csr.h
+++ b/target/loongarch/cpu-csr.h
@@ -68,6 +68,8 @@ FIELD(TLBENTRY, PLV, 2, 2)
FIELD(TLBENTRY, MAT, 4, 2)
FIELD(TLBENTRY, G, 6, 1)
FIELD(TLBENTRY, HUGE, 6, 1)
+FIELD(TLBENTRY, P, 7, 1)
+FIELD(TLBENTRY, W, 8, 1)
FIELD(TLBENTRY, HGLOBAL, 12, 1)
FIELD(TLBENTRY, LEVEL, 13, 2)
FIELD(TLBENTRY_32, PPN, 8, 24)
diff --git a/target/loongarch/cpu-mmu.h b/target/loongarch/cpu-mmu.h
index 311bf7b2d7..e4f3199f44 100644
--- a/target/loongarch/cpu-mmu.h
+++ b/target/loongarch/cpu-mmu.h
@@ -29,6 +29,37 @@ typedef struct MMUContext {
int mmu_index;
} MMUContext;
+static inline bool cpu_has_ptw(CPULoongArchState *env)
+{
+ return !!FIELD_EX64(env->CSR_PWCH, CSR_PWCH, HPTW_EN);
+}
+
+static inline bool pte_present(CPULoongArchState *env, uint64_t entry)
+{
+ uint8_t present;
+
+ if (cpu_has_ptw(env)) {
+ present = FIELD_EX64(entry, TLBENTRY, P);
+ } else {
+ present = FIELD_EX64(entry, TLBENTRY, V);
+ }
+
+ return !!present;
+}
+
+static inline bool pte_write(CPULoongArchState *env, uint64_t entry)
+{
+ uint8_t writable;
+
+ if (cpu_has_ptw(env)) {
+ writable = FIELD_EX64(entry, TLBENTRY, W);
+ } else {
+ writable = FIELD_EX64(entry, TLBENTRY, D);
+ }
+
+ return !!writable;
+}
+
bool check_ps(CPULoongArchState *ent, uint8_t ps);
TLBRet loongarch_check_pte(CPULoongArchState *env, MMUContext *context,
MMUAccessType access_type, int mmu_idx);
diff --git a/target/loongarch/cpu_helper.c b/target/loongarch/cpu_helper.c
index 6572d0a7cc..2e07049425 100644
--- a/target/loongarch/cpu_helper.c
+++ b/target/loongarch/cpu_helper.c
@@ -53,8 +53,8 @@ TLBRet loongarch_check_pte(CPULoongArchState *env, MMUContext *context,
tlb_entry = context->pte;
tlb_ps = context->ps;
- tlb_v = FIELD_EX64(tlb_entry, TLBENTRY, V);
- tlb_d = FIELD_EX64(tlb_entry, TLBENTRY, D);
+ tlb_v = pte_present(env, tlb_entry);
+ tlb_d = pte_write(env, tlb_entry);
tlb_plv = FIELD_EX64(tlb_entry, TLBENTRY, PLV);
if (is_la64(env)) {
tlb_ppn = FIELD_EX64(tlb_entry, TLBENTRY_64, PPN);
diff --git a/target/loongarch/tcg/tlb_helper.c b/target/loongarch/tcg/tlb_helper.c
index 7a4de7e566..10c01ead3f 100644
--- a/target/loongarch/tcg/tlb_helper.c
+++ b/target/loongarch/tcg/tlb_helper.c
@@ -119,11 +119,9 @@ static void raise_mmu_exception(CPULoongArchState *env, vaddr address,
static void invalidate_tlb_entry(CPULoongArchState *env, int index)
{
target_ulong addr, mask, pagesize;
- uint8_t tlb_ps;
+ uint8_t tlb_ps, tlb_v;
LoongArchTLB *tlb = &env->tlb[index];
int mmu_idx;
- uint8_t tlb_v0 = FIELD_EX64(tlb->tlb_entry0, TLBENTRY, V);
- uint8_t tlb_v1 = FIELD_EX64(tlb->tlb_entry1, TLBENTRY, V);
uint64_t tlb_vppn = FIELD_EX64(tlb->tlb_misc, TLB_MISC, VPPN);
uint8_t tlb_e = FIELD_EX64(tlb->tlb_misc, TLB_MISC, E);
@@ -137,7 +135,8 @@ static void invalidate_tlb_entry(CPULoongArchState *env, int index)
addr = (tlb_vppn << R_TLB_MISC_VPPN_SHIFT) & ~mask;
addr = sextract64(addr, 0, TARGET_VIRT_ADDR_SPACE_BITS);
- if (tlb_v0) {
+ tlb_v = pte_present(env, tlb->tlb_entry0);
+ if (tlb_v) {
mmu_idx = BIT(FIELD_EX64(tlb->tlb_entry0, TLBENTRY, PLV));
/* Even page is accessed in kernel mode */
if (tlb->tlb_misc & TLB_MISC_KM_PTE_LOW0) {
@@ -147,7 +146,8 @@ static void invalidate_tlb_entry(CPULoongArchState *env, int index)
mmu_idx, TARGET_LONG_BITS);
}
- if (tlb_v1) {
+ tlb_v = pte_present(env, tlb->tlb_entry1);
+ if (tlb_v) {
mmu_idx = BIT(FIELD_EX64(tlb->tlb_entry1, TLBENTRY, PLV));
/* Odd page is accessed in kernel mode */
if (tlb->tlb_misc & TLB_MISC_KM_PTE_LOW1) {
@@ -362,8 +362,8 @@ void helper_tlbwr(CPULoongArchState *env)
/* Check whether ASID/VPPN is the same */
if (old->tlb_misc == new.tlb_misc) {
/* Check whether pte is the same or invalid */
- tlb_v0 = FIELD_EX64(old->tlb_entry0, TLBENTRY, V);
- tlb_v1 = FIELD_EX64(old->tlb_entry1, TLBENTRY, V);
+ tlb_v0 = pte_present(env, old->tlb_entry0);
+ tlb_v1 = pte_present(env, old->tlb_entry1);
if ((!tlb_v0 || new.tlb_entry0 == old->tlb_entry0) &&
(!tlb_v1 || new.tlb_entry1 == old->tlb_entry1)) {
skip_inv = true;
--
2.39.3