From nobody Tue Oct 28 12:39:19 2025 Delivered-To: importer@patchew.org Received-SPF: pass (zoho.com: domain of gnu.org designates 208.118.235.17 as permitted sender) client-ip=208.118.235.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Authentication-Results: mx.zohomail.com; spf=pass (zoho.com: domain of gnu.org designates 208.118.235.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org Return-Path: Received: from lists.gnu.org (lists.gnu.org [208.118.235.17]) by mx.zohomail.com with SMTPS id 1522981473048898.3525927880901; Thu, 5 Apr 2018 19:24:33 -0700 (PDT) Received: from localhost ([::1]:44271 helo=lists.gnu.org) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1f4H3I-0003Ne-6r for importer@patchew.org; Thu, 05 Apr 2018 22:24:32 -0400 Received: from eggs.gnu.org ([2001:4830:134:3::10]:58438) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1f4GsL-0002GW-UJ for qemu-devel@nongnu.org; Thu, 05 Apr 2018 22:13:16 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1f4GsJ-0003Vn-UJ for qemu-devel@nongnu.org; Thu, 05 Apr 2018 22:13:13 -0400 Received: from out5-smtp.messagingengine.com ([66.111.4.29]:35083) by eggs.gnu.org with esmtps (TLS1.0:DHE_RSA_AES_256_CBC_SHA1:32) (Exim 4.71) (envelope-from ) id 1f4GsJ-0003VI-QN for qemu-devel@nongnu.org; Thu, 05 Apr 2018 22:13:11 -0400 Received: from compute4.internal (compute4.nyi.internal [10.202.2.44]) by mailout.nyi.internal (Postfix) with ESMTP id 8684B20B5F; Thu, 5 Apr 2018 22:13:11 -0400 (EDT) Received: from mailfrontend2 ([10.202.2.163]) by compute4.internal (MEProxy); Thu, 05 Apr 2018 22:13:11 -0400 Received: from localhost (flamenco.cs.columbia.edu [128.59.20.216]) by mail.messagingengine.com (Postfix) with ESMTPA id 3EF6B1025D; Thu, 5 Apr 2018 22:13:11 -0400 (EDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=braap.org; h=cc :content-transfer-encoding:content-type:date:from:in-reply-to :message-id:mime-version:references:subject:to:x-me-sender :x-me-sender:x-sasl-enc; s=mesmtp; bh=v0vb8rkcWPaXIXnKGFutJwCF1S j7MYnFB1QSMmLOOig=; b=tBO7gWWeiX3VJTx0j8LDXKLMeISe8lTjPK4eaLRCyx KkCLY2KNe+O1wD7vHRemW0s9dXQUY+fZTaLOiKiQ82AR6sWKZHFKIJGg8dt/boqK hhNbqbwGFBYP1Kyg1ymS+CijhVJyDtotmg7v5GdCdBtBkwKiFULAbCV9mk1mTDu0 A= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d= messagingengine.com; h=cc:content-transfer-encoding:content-type :date:from:in-reply-to:message-id:mime-version:references :subject:to:x-me-sender:x-me-sender:x-sasl-enc; s=fm2; bh=v0vb8r kcWPaXIXnKGFutJwCF1Sj7MYnFB1QSMmLOOig=; b=XIuIvRlZ3H3JeCT3EePPtV oVonaCRqrfd9YPjQXI7m316sKMHRbgzHF/t3cBQOgEbDrKGoMAPKr4gtcA646fwo MLBlMJ8hBmL1UhqyF9vwuiLpdK/u2rE3mcVVhDHVutIYviMTj19iZR1qxGkrKJvD ri/sZgVyB9XAWmoXtViCONv6c3bcZ5mL+9XA0r6+ISxASDOaLBgToopzSIPVI3lf 2vVJrdEfmkDnrjYIffk1FmxqDW8mW/Nx4zKFQLZNFLpKSS+Rw46djXFPaO5NuDFu V6qUU148/vCPkpl6ow7+E77ajlnmHgwZGgRWapzUW5SU+HcQc7mCVagIHuUOiGyQ == X-ME-Sender: From: "Emilio G. Cota" To: qemu-devel@nongnu.org Date: Thu, 5 Apr 2018 22:13:02 -0400 Message-Id: <1522980788-1252-12-git-send-email-cota@braap.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1522980788-1252-1-git-send-email-cota@braap.org> References: <1522980788-1252-1-git-send-email-cota@braap.org> MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable X-detected-operating-system: by eggs.gnu.org: GNU/Linux 2.2.x-3.x [generic] [fuzzy] X-Received-From: 66.111.4.29 Subject: [Qemu-devel] [PATCH v2 11/17] translate-all: add page_locked assertions X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.21 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Richard Henderson , =?UTF-8?q?Alex=20Benn=C3=A9e?= , Paolo Bonzini Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail: RSF_0 Z_629925259 SPT_0 This is only compiled under CONFIG_DEBUG_TCG to avoid bloating the binary. In user-mode, assert_page_locked is equivalent to assert_mmap_lock. Note: There are some tb_lock assertions left that will be removed by later patches. Suggested-by: Alex Benn=C3=A9e Signed-off-by: Emilio G. Cota --- accel/tcg/translate-all.c | 90 +++++++++++++++++++++++++++++++++++++++++++= ++-- 1 file changed, 87 insertions(+), 3 deletions(-) diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c index 042378a..29bc1da 100644 --- a/accel/tcg/translate-all.c +++ b/accel/tcg/translate-all.c @@ -580,6 +580,9 @@ static inline PageDesc *page_find(tb_page_addr_t index) =20 /* In user-mode page locks aren't used; mmap_lock is enough */ #ifdef CONFIG_USER_ONLY + +#define assert_page_locked(pd) tcg_debug_assert(have_mmap_lock()) + static inline void page_lock(PageDesc *pd) { } =20 @@ -602,14 +605,91 @@ void page_collection_unlock(struct page_collection *s= et) { } #else /* !CONFIG_USER_ONLY */ =20 +#ifdef CONFIG_DEBUG_TCG + +struct page_lock_debug { + const PageDesc *pd; + QLIST_ENTRY(page_lock_debug) entry; +}; + +static __thread QLIST_HEAD(, page_lock_debug) page_lock_debug_head; + +static struct page_lock_debug *get_page_lock_debug(const PageDesc *pd) +{ + struct page_lock_debug *pld; + + QLIST_FOREACH(pld, &page_lock_debug_head, entry) { + if (pld->pd =3D=3D pd) { + return pld; + } + } + return NULL; +} + +static bool page_is_locked(const PageDesc *pd) +{ + struct page_lock_debug *pld; + + pld =3D get_page_lock_debug(pd); + return pld !=3D NULL; +} + +static void page_lock__debug(const PageDesc *pd) +{ + struct page_lock_debug *pld; + + g_assert(!page_is_locked(pd)); + pld =3D g_new(struct page_lock_debug, 1); + pld->pd =3D pd; + QLIST_INSERT_HEAD(&page_lock_debug_head, pld, entry); +} + +static void page_unlock__debug(const PageDesc *pd) +{ + struct page_lock_debug *pld; + + pld =3D get_page_lock_debug(pd); + g_assert(pld); + QLIST_REMOVE(pld, entry); + g_free(pld); +} + +static void +do_assert_page_locked(const PageDesc *pd, const char *file, int line) +{ + if (unlikely(!page_is_locked(pd))) { + error_report("assert_page_lock: PageDesc %p not locked @ %s:%d", + pd, file, line); + abort(); + } +} + +#define assert_page_locked(pd) do_assert_page_locked(pd, __FILE__, __LINE_= _) + +#else /* !CONFIG_DEBUG_TCG */ + +#define assert_page_locked(pd) + +static inline void page_lock__debug(const PageDesc *pd) +{ +} + +static inline void page_unlock__debug(const PageDesc *pd) +{ +} + +#endif /* CONFIG_DEBUG_TCG */ + static inline void page_lock(PageDesc *pd) { + page_lock__debug(pd); qemu_spin_lock(&pd->lock); } =20 static inline void page_unlock(PageDesc *pd) { qemu_spin_unlock(&pd->lock); + page_unlock__debug(pd); } =20 /* lock the page(s) of a TB in the correct acquisition order */ @@ -1091,6 +1171,7 @@ static TranslationBlock *tb_alloc(target_ulong pc) /* call with @p->lock held */ static inline void invalidate_page_bitmap(PageDesc *p) { + assert_page_locked(p); #ifdef CONFIG_SOFTMMU g_free(p->code_bitmap); p->code_bitmap =3D NULL; @@ -1247,6 +1328,7 @@ static inline void tb_page_remove(PageDesc *pd, Trans= lationBlock *tb) uintptr_t *pprev; unsigned int n1; =20 + assert_page_locked(pd); pprev =3D &pd->first_tb; PAGE_FOR_EACH_TB(pd, tb1, n1) { if (tb1 =3D=3D tb) { @@ -1395,6 +1477,7 @@ static void build_page_bitmap(PageDesc *p) int n, tb_start, tb_end; TranslationBlock *tb; =20 + assert_page_locked(p); p->code_bitmap =3D bitmap_new(TARGET_PAGE_SIZE); =20 PAGE_FOR_EACH_TB(p, tb, n) { @@ -1428,7 +1511,7 @@ static inline void tb_page_add(PageDesc *p, Translati= onBlock *tb, bool page_already_protected; #endif =20 - assert_memory_lock(); + assert_page_locked(p); =20 tb->page_addr[n] =3D page_addr; tb->page_next[n] =3D p->first_tb; @@ -1710,8 +1793,7 @@ tb_invalidate_phys_page_range__locked(struct page_col= lection *pages, uint32_t current_flags =3D 0; #endif /* TARGET_HAS_PRECISE_SMC */ =20 - assert_memory_lock(); - assert_tb_locked(); + assert_page_locked(p); =20 #if defined(TARGET_HAS_PRECISE_SMC) if (cpu !=3D NULL) { @@ -1723,6 +1805,7 @@ tb_invalidate_phys_page_range__locked(struct page_col= lection *pages, /* XXX: see if in some cases it could be faster to invalidate all the code */ PAGE_FOR_EACH_TB(p, tb, n) { + assert_page_locked(p); /* NOTE: this is subtle as a TB may span two physical pages */ if (n =3D=3D 0) { /* NOTE: tb_end may be after the end of the page, but @@ -1879,6 +1962,7 @@ void tb_invalidate_phys_page_fast(tb_page_addr_t star= t, int len) } =20 pages =3D page_collection_lock(start, start + len); + assert_page_locked(p); if (!p->code_bitmap && ++p->code_write_count >=3D SMC_BITMAP_USE_THRESHOLD) { build_page_bitmap(p); --=20 2.7.4