From nobody Tue Feb 10 13:16:37 2026 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of _spf.google.com designates 209.85.128.42 as permitted sender) client-ip=209.85.128.42; envelope-from=philippe.mathieu.daude@gmail.com; helo=mail-wm1-f42.google.com; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of _spf.google.com designates 209.85.128.42 as permitted sender) smtp.mailfrom=philippe.mathieu.daude@gmail.com ARC-Seal: i=1; a=rsa-sha256; t=1623247818; cv=none; d=zohomail.com; s=zohoarc; b=e3kefy9IFMx74KVeXzf/HB738IIriItEk1eJvUOm+0CkmlsaBmH+ygUsNf6AJt8LkTnWuy7vP9jZ7eFzhk/6oJE4FlnNE4IPfFhMz/pb3iWMRfEwDbri3QWGPaQldSRZkMjq++/qDrTIrLDBLqtXtRUyV0aadUJoQEHP2V0pD4k= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1623247818; h=Content-Type:Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:MIME-Version:Message-ID:References:Sender:Subject:To; bh=SK4Sda9oKAO+kAJdqn+QafWI9XvksejZSvmI3F9R0lo=; b=eY9ikVGd/rtujbsFYGyTCZ+KUkR1SxlXI8qQaAGx7oZyHyVguhoOPSqUGrsQ4nk3/MhFVMmm1pN7CxeEWLqpsx+T8/v6bqfOoro/U3J1PB+wxjrIZEhp8Dtmta4E08/v6es3xeWbNkJgql0sd0nMNDjo9j2m5nYh7sI/MXut06s= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of _spf.google.com designates 209.85.128.42 as permitted sender) smtp.mailfrom=philippe.mathieu.daude@gmail.com Received: from mail-wm1-f42.google.com (mail-wm1-f42.google.com [209.85.128.42]) by mx.zohomail.com with SMTPS id 1623247818936697.7067616504783; Wed, 9 Jun 2021 07:10:18 -0700 (PDT) Received: by mail-wm1-f42.google.com with SMTP id l11-20020a05600c4f0bb029017a7cd488f5so4351800wmq.0 for ; Wed, 09 Jun 2021 07:10:18 -0700 (PDT) Return-Path: Return-Path: Received: from x1w.redhat.com (235.red-83-57-168.dynamicip.rima-tde.net. [83.57.168.235]) by smtp.gmail.com with ESMTPSA id j12sm24258wrt.69.2021.06.09.07.10.16 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 09 Jun 2021 07:10:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=sender:from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=SK4Sda9oKAO+kAJdqn+QafWI9XvksejZSvmI3F9R0lo=; b=ThhnxJXqlKNpUl5qYUi6A0MSXeZgFTHQMckiI4DPHkoeqz7qS8usBhZ8su0vGTZ2iF FODzMTSK7RdUBKP+NaxESlyOILQTRZsRo5lBMyIXa6xzvse/9nK56d7A4wqjgE5ckzlN t/8paIt1g9D1UUbGh0o2A65Dj7d/zLnMetFCKhXAokFYFPG/bGbKtKhJPCa2e4tHArxe WQ421EjaxNRBhZoZV/GU0p39CrKPkDnJHVZDyIXXghvyn1L2BkAehhKcANcyac9955gy uYozoC89NtM/romJsw8tDT7du/eO1wmNo5urq4gY4BPJBoeLrujgmT6zgx/TL02vuUOl 7dhw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:sender:from:to:cc:subject:date:message-id :in-reply-to:references:mime-version:content-transfer-encoding; bh=SK4Sda9oKAO+kAJdqn+QafWI9XvksejZSvmI3F9R0lo=; b=bI9JMTt+E1yuOL1Dl49cQOjUO2NvoFO1Pwh+YQOariNwVcuVwN/x/53sZOtG1mcr0A R+CVFJWcsHuFPALVUnHaLnAU34y8kio4bDdD83IC8X3H+zQzxSx309myfTJXd15jomZr scu6/o2AEnOr24e90un4C7D5Hu728vKPojOcHxR8Kq2qMxSpqEwbR6WahcUBjLG4+Ukz Qe2f+zMTbIGSmiDKrC+3aDLyi/m2jNOCmH9xDJPlQsFdeMf83Jhh6DBjNlqO8v7gN6jB KmkhjOZqMinnql4k4EQsGvr+vy0mz7CT50dLWUEbV6vmLZI7+F62/Nk6pRXJ+jLIhEUz 0oLw== X-Gm-Message-State: AOAM531BFvJAFkr7hNeP23tZM5oSNIH6JGr9OvHZT/VdKdzW7tY/EAlk 1BHexrerqaS7TViVvXiBh88= X-Google-Smtp-Source: ABdhPJxxYSgNV6WU1qc3iTRGMNfGLj5hPfnTOD1rGGyzO0/Xjn03K5FFl2GPQtIdhU6MiAFnPyey5w== X-Received: by 2002:a05:600c:35c3:: with SMTP id r3mr10033186wmq.169.1623247816991; Wed, 09 Jun 2021 07:10:16 -0700 (PDT) Sender: =?UTF-8?Q?Philippe_Mathieu=2DDaud=C3=A9?= From: =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= To: qemu-devel@nongnu.org Cc: Peter Maydell , Richard Henderson , Mark Cave-Ayland , =?UTF-8?q?Alex=20Benn=C3=A9e?= , Paolo Bonzini , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= Subject: [PATCH v2 1/2] accel/tcg/cputlb: Extract load_helper_unaligned() from load_helper() Date: Wed, 9 Jun 2021 16:10:09 +0200 Message-Id: <20210609141010.1066750-2-f4bug@amsat.org> X-Mailer: git-send-email 2.31.1 In-Reply-To: <20210609141010.1066750-1-f4bug@amsat.org> References: <20210609141010.1066750-1-f4bug@amsat.org> MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable X-ZohoMail-DKIM: pass (identity @gmail.com) Replace a goto statement by an inlined function for easier review. No logical change intended. Inspired-by: Mark Cave-Ayland Signed-off-by: Philippe Mathieu-Daud=C3=A9 --- accel/tcg/cputlb.c | 54 ++++++++++++++++++++++++++++------------------ 1 file changed, 33 insertions(+), 21 deletions(-) diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c index f24348e9793..2b5d569412c 100644 --- a/accel/tcg/cputlb.c +++ b/accel/tcg/cputlb.c @@ -1851,6 +1851,34 @@ load_memop(const void *haddr, MemOp op) } } =20 +static inline uint64_t QEMU_ALWAYS_INLINE +load_helper_unaligned(CPUArchState *env, target_ulong addr, TCGMemOpIdx oi, + uintptr_t retaddr, MemOp op, bool code_read, + FullLoadHelper *full_load) +{ + size_t size =3D memop_size(op); + target_ulong addr1, addr2; + uint64_t res; + uint64_t r1, r2; + unsigned shift; + + addr1 =3D addr & ~((target_ulong)size - 1); + addr2 =3D addr1 + size; + r1 =3D full_load(env, addr1, oi, retaddr); + r2 =3D full_load(env, addr2, oi, retaddr); + shift =3D (addr & (size - 1)) * 8; + + if (memop_big_endian(op)) { + /* Big-endian combine. */ + res =3D (r1 << shift) | (r2 >> ((size * 8) - shift)); + } else { + /* Little-endian combine. */ + res =3D (r1 >> shift) | (r2 << ((size * 8) - shift)); + } + + return res & MAKE_64BIT_MASK(0, size * 8); +} + static inline uint64_t QEMU_ALWAYS_INLINE load_helper(CPUArchState *env, target_ulong addr, TCGMemOpIdx oi, uintptr_t retaddr, MemOp op, bool code_read, @@ -1866,7 +1894,6 @@ load_helper(CPUArchState *env, target_ulong addr, TCG= MemOpIdx oi, code_read ? MMU_INST_FETCH : MMU_DATA_LOAD; unsigned a_bits =3D get_alignment_bits(get_memop(oi)); void *haddr; - uint64_t res; size_t size =3D memop_size(op); =20 /* Handle CPU specific unaligned behaviour */ @@ -1893,9 +1920,10 @@ load_helper(CPUArchState *env, target_ulong addr, TC= GMemOpIdx oi, CPUIOTLBEntry *iotlbentry; bool need_swap; =20 - /* For anything that is unaligned, recurse through full_load. */ + /* For anything that is unaligned, recurse through byte loads. */ if ((addr & (size - 1)) !=3D 0) { - goto do_unaligned_access; + return load_helper_unaligned(env, addr, oi, retaddr, op, + code_read, full_load); } =20 iotlbentry =3D &env_tlb(env)->d[mmu_idx].iotlb[index]; @@ -1932,24 +1960,8 @@ load_helper(CPUArchState *env, target_ulong addr, TC= GMemOpIdx oi, if (size > 1 && unlikely((addr & ~TARGET_PAGE_MASK) + size - 1 >=3D TARGET_PAGE_SIZE)) { - target_ulong addr1, addr2; - uint64_t r1, r2; - unsigned shift; - do_unaligned_access: - addr1 =3D addr & ~((target_ulong)size - 1); - addr2 =3D addr1 + size; - r1 =3D full_load(env, addr1, oi, retaddr); - r2 =3D full_load(env, addr2, oi, retaddr); - shift =3D (addr & (size - 1)) * 8; - - if (memop_big_endian(op)) { - /* Big-endian combine. */ - res =3D (r1 << shift) | (r2 >> ((size * 8) - shift)); - } else { - /* Little-endian combine. */ - res =3D (r1 >> shift) | (r2 << ((size * 8) - shift)); - } - return res & MAKE_64BIT_MASK(0, size * 8); + return load_helper_unaligned(env, addr, oi, retaddr, op, + code_read, full_load); } =20 haddr =3D (void *)((uintptr_t)addr + entry->addend); --=20 2.31.1