accel/tcg/tb-jmp-cache.h | 8 +++-- accel/tcg/cpu-exec.c | 66 ++++++++++++++-------------------------- 2 files changed, 28 insertions(+), 46 deletions(-)
Unless I'm missing something egregious, the jmp cache is only every
populated with a valid entry by the same thread that reads the cache.
Therefore, the contents of any valid entry are always consistent and
there is no need for any acquire/release magic.
Indeed ->tb has to be accessed with atomics, because concurrent
invalidations would otherwise cause data races. But ->pc is only ever
accessed by one thread, and accesses to ->tb and ->pc within tb_lookup
can never race with another tb_lookup. While the TranslationBlock
(especially the flags) could be modified by a concurrent invalidation,
store-release and load-acquire operations on the cache entry would
not add any additional ordering beyond what you get from performing
the accesses within a single thread.
Because of this, there is really nothing to win in splitting the CF_PCREL
and !CF_PCREL paths. It is easier to just always use the ->pc field in
the jump cache.
I noticed this while working on splitting commit 8ed558ec0cb
("accel/tcg: Introduce TARGET_TB_PCREL", 2022-10-04) into multiple
pieces, for the sake of finding a more fine-grained bisection
result for https://gitlab.com/qemu-project/qemu/-/issues/2092.
It does not (and does not intend to) fix that issue; therefore
it may make sense to not commit it until the root cause
of issue #2092 is found.
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
---
accel/tcg/tb-jmp-cache.h | 8 +++--
accel/tcg/cpu-exec.c | 66 ++++++++++++++--------------------------
2 files changed, 28 insertions(+), 46 deletions(-)
diff --git a/accel/tcg/tb-jmp-cache.h b/accel/tcg/tb-jmp-cache.h
index bb424c8a05b..4ab8553afcc 100644
--- a/accel/tcg/tb-jmp-cache.h
+++ b/accel/tcg/tb-jmp-cache.h
@@ -13,9 +13,11 @@
#define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
/*
- * Accessed in parallel; all accesses to 'tb' must be atomic.
- * For CF_PCREL, accesses to 'pc' must be protected by a
- * load_acquire/store_release to 'tb'.
+ * Invalidated in parallel; all accesses to 'tb' must be atomic.
+ * A valid entry is read/written by a single CPU, therefore there is
+ * no need for qatomic_rcu_read() and pc is always consistent with a
+ * non-NULL value of 'tb'. Strictly speaking pc is only needed for
+ * CF_PCREL, but it's used always for simplicity.
*/
struct CPUJumpCache {
struct rcu_head rcu;
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
index 67eda9865ee..40c268bfa1f 100644
--- a/accel/tcg/cpu-exec.c
+++ b/accel/tcg/cpu-exec.c
@@ -253,43 +253,29 @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, vaddr pc,
hash = tb_jmp_cache_hash_func(pc);
jc = cpu->tb_jmp_cache;
- if (cflags & CF_PCREL) {
- /* Use acquire to ensure current load of pc from jc. */
- tb = qatomic_load_acquire(&jc->array[hash].tb);
-
- if (likely(tb &&
- jc->array[hash].pc == pc &&
- tb->cs_base == cs_base &&
- tb->flags == flags &&
- tb_cflags(tb) == cflags)) {
- return tb;
- }
- tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags);
- if (tb == NULL) {
- return NULL;
- }
- jc->array[hash].pc = pc;
- /* Ensure pc is written first. */
- qatomic_store_release(&jc->array[hash].tb, tb);
- } else {
- /* Use rcu_read to ensure current load of pc from *tb. */
- tb = qatomic_rcu_read(&jc->array[hash].tb);
-
- if (likely(tb &&
- tb->pc == pc &&
- tb->cs_base == cs_base &&
- tb->flags == flags &&
- tb_cflags(tb) == cflags)) {
- return tb;
- }
- tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags);
- if (tb == NULL) {
- return NULL;
- }
- /* Use the pc value already stored in tb->pc. */
- qatomic_set(&jc->array[hash].tb, tb);
+ tb = qatomic_read(&jc->array[hash].tb);
+ if (likely(tb &&
+ jc->array[hash].pc == pc &&
+ tb->cs_base == cs_base &&
+ tb->flags == flags &&
+ tb_cflags(tb) == cflags)) {
+ goto hit;
}
+ tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags);
+ if (tb == NULL) {
+ return NULL;
+ }
+
+ jc->array[hash].pc = pc;
+ qatomic_set(&jc->array[hash].tb, tb);
+
+hit:
+ /*
+ * As long as tb is not NULL, the contents are consistent. Therefore,
+ * the virtual PC has to match for non-CF_PCREL translations.
+ */
+ assert((tb_cflags(tb) & CF_PCREL) || tb->pc == pc);
return tb;
}
@@ -1012,14 +998,8 @@ cpu_exec_loop(CPUState *cpu, SyncClocks *sc)
*/
h = tb_jmp_cache_hash_func(pc);
jc = cpu->tb_jmp_cache;
- if (cflags & CF_PCREL) {
- jc->array[h].pc = pc;
- /* Ensure pc is written first. */
- qatomic_store_release(&jc->array[h].tb, tb);
- } else {
- /* Use the pc value already stored in tb->pc. */
- qatomic_set(&jc->array[h].tb, tb);
- }
+ jc->array[h].pc = pc;
+ qatomic_set(&jc->array[h].tb, tb);
}
#ifndef CONFIG_USER_ONLY
--
2.43.0
Paolo Bonzini <pbonzini@redhat.com> writes: > Unless I'm missing something egregious, the jmp cache is only every > populated with a valid entry by the same thread that reads the cache. > Therefore, the contents of any valid entry are always consistent and > there is no need for any acquire/release magic. > > Indeed ->tb has to be accessed with atomics, because concurrent > invalidations would otherwise cause data races. But ->pc is only ever > accessed by one thread, and accesses to ->tb and ->pc within tb_lookup > can never race with another tb_lookup. While the TranslationBlock > (especially the flags) could be modified by a concurrent invalidation, > store-release and load-acquire operations on the cache entry would > not add any additional ordering beyond what you get from performing > the accesses within a single thread. > > Because of this, there is really nothing to win in splitting the CF_PCREL > and !CF_PCREL paths. It is easier to just always use the ->pc field in > the jump cache. > > I noticed this while working on splitting commit 8ed558ec0cb > ("accel/tcg: Introduce TARGET_TB_PCREL", 2022-10-04) into multiple > pieces, for the sake of finding a more fine-grained bisection > result for https://gitlab.com/qemu-project/qemu/-/issues/2092. > It does not (and does not intend to) fix that issue; therefore > it may make sense to not commit it until the root cause > of issue #2092 is found. > > Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> > --- > accel/tcg/tb-jmp-cache.h | 8 +++-- > accel/tcg/cpu-exec.c | 66 ++++++++++++++-------------------------- > 2 files changed, 28 insertions(+), 46 deletions(-) > > diff --git a/accel/tcg/tb-jmp-cache.h b/accel/tcg/tb-jmp-cache.h > index bb424c8a05b..4ab8553afcc 100644 > --- a/accel/tcg/tb-jmp-cache.h > +++ b/accel/tcg/tb-jmp-cache.h > @@ -13,9 +13,11 @@ > #define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS) > > /* > - * Accessed in parallel; all accesses to 'tb' must be atomic. > - * For CF_PCREL, accesses to 'pc' must be protected by a > - * load_acquire/store_release to 'tb'. > + * Invalidated in parallel; all accesses to 'tb' must be atomic. > + * A valid entry is read/written by a single CPU, therefore there is > + * no need for qatomic_rcu_read() and pc is always consistent with a > + * non-NULL value of 'tb'. Strictly speaking pc is only needed for > + * CF_PCREL, but it's used always for simplicity. > */ I can follow the logic and soak testing seems to show its stable. So: Tested-by: Alex Bennée <alex.bennee@linaro.org> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> However I would note that TranslationBlock has the comment: * jmp_lock also protects the CF_INVALID cflag; a jump must not be chained * to a destination TB that has CF_INVALID set. which I don't think holds true. It might be worth summarising the lock requirements of TranslationBlock in a kdoc header. > struct CPUJumpCache { > struct rcu_head rcu; > diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c > index 67eda9865ee..40c268bfa1f 100644 > --- a/accel/tcg/cpu-exec.c > +++ b/accel/tcg/cpu-exec.c > @@ -253,43 +253,29 @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, vaddr pc, > hash = tb_jmp_cache_hash_func(pc); > jc = cpu->tb_jmp_cache; > > - if (cflags & CF_PCREL) { > - /* Use acquire to ensure current load of pc from jc. */ > - tb = qatomic_load_acquire(&jc->array[hash].tb); > - > - if (likely(tb && > - jc->array[hash].pc == pc && > - tb->cs_base == cs_base && > - tb->flags == flags && > - tb_cflags(tb) == cflags)) { > - return tb; > - } > - tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags); > - if (tb == NULL) { > - return NULL; > - } > - jc->array[hash].pc = pc; > - /* Ensure pc is written first. */ > - qatomic_store_release(&jc->array[hash].tb, tb); > - } else { > - /* Use rcu_read to ensure current load of pc from *tb. */ > - tb = qatomic_rcu_read(&jc->array[hash].tb); > - > - if (likely(tb && > - tb->pc == pc && > - tb->cs_base == cs_base && > - tb->flags == flags && > - tb_cflags(tb) == cflags)) { > - return tb; > - } > - tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags); > - if (tb == NULL) { > - return NULL; > - } > - /* Use the pc value already stored in tb->pc. */ > - qatomic_set(&jc->array[hash].tb, tb); > + tb = qatomic_read(&jc->array[hash].tb); > + if (likely(tb && > + jc->array[hash].pc == pc && > + tb->cs_base == cs_base && > + tb->flags == flags && > + tb_cflags(tb) == cflags)) { > + goto hit; > } > > + tb = tb_htable_lookup(cpu, pc, cs_base, flags, cflags); > + if (tb == NULL) { > + return NULL; > + } > + > + jc->array[hash].pc = pc; > + qatomic_set(&jc->array[hash].tb, tb); > + > +hit: > + /* > + * As long as tb is not NULL, the contents are consistent. Therefore, > + * the virtual PC has to match for non-CF_PCREL translations. > + */ > + assert((tb_cflags(tb) & CF_PCREL) || tb->pc == pc); > return tb; > } > > @@ -1012,14 +998,8 @@ cpu_exec_loop(CPUState *cpu, SyncClocks *sc) > */ > h = tb_jmp_cache_hash_func(pc); > jc = cpu->tb_jmp_cache; > - if (cflags & CF_PCREL) { > - jc->array[h].pc = pc; > - /* Ensure pc is written first. */ > - qatomic_store_release(&jc->array[h].tb, tb); > - } else { > - /* Use the pc value already stored in tb->pc. */ > - qatomic_set(&jc->array[h].tb, tb); > - } > + jc->array[h].pc = pc; > + qatomic_set(&jc->array[h].tb, tb); > } > > #ifndef CONFIG_USER_ONLY -- Alex Bennée Virtualisation Tech Lead @ Linaro
On Tue, Jan 23, 2024 at 11:02 PM Alex Bennée <alex.bennee@linaro.org> wrote: > However I would note that TranslationBlock has the comment: > > * jmp_lock also protects the CF_INVALID cflag; a jump must not be chained > * to a destination TB that has CF_INVALID set. > > which I don't think holds true. It does, both /* make sure the destination TB is valid */ if (tb_next->cflags & CF_INVALID) { goto out_unlock_next; } and qatomic_set(&tb->cflags, tb->cflags | CF_INVALID); are protected by jmp_lock. And if something is chaining to a CF_INVALID translation block before tb_phys_invalidate() sets the flag, it is cleaned up by tb_remove_from_jmp_list(tb, 0); tb_remove_from_jmp_list(tb, 1); tb_jmp_unlink(tb); before tb_phys_invalidate() returns. Paolo
On 1/23/24 01:34, Paolo Bonzini wrote: > Unless I'm missing something egregious, the jmp cache is only every > populated with a valid entry by the same thread that reads the cache. > Therefore, the contents of any valid entry are always consistent and > there is no need for any acquire/release magic. I think you're right, and I over-complicated this thinking about invalidations. > Because of this, there is really nothing to win in splitting the CF_PCREL > and !CF_PCREL paths. It is easier to just always use the ->pc field in > the jump cache. Once upon a time, PCREL was an ifdef, and the jump cache pc did not exist for !PCREL. The split has not been addressed since then. The cleanup looks good. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> r~
On 1/23/24 07:57, Richard Henderson wrote: > On 1/23/24 01:34, Paolo Bonzini wrote: >> Unless I'm missing something egregious, the jmp cache is only every >> populated with a valid entry by the same thread that reads the cache. >> Therefore, the contents of any valid entry are always consistent and >> there is no need for any acquire/release magic. > > I think you're right, and I over-complicated this thinking about invalidations. > >> Because of this, there is really nothing to win in splitting the CF_PCREL >> and !CF_PCREL paths. It is easier to just always use the ->pc field in >> the jump cache. > > Once upon a time, PCREL was an ifdef, and the jump cache pc did not exist for !PCREL. The > split has not been addressed since then. > > > The cleanup looks good. > > Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Queued, thanks. r~
© 2016 - 2024 Red Hat, Inc.