From nobody Fri Oct 3 15:31:34 2025 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by smtp.subspace.kernel.org (Postfix) with ESMTP id DD9CB32A3C7 for ; Fri, 29 Aug 2025 15:35:27 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=217.140.110.172 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1756481730; cv=none; b=K8OHw4GmHSPlFMKjqBu5JkSW24d5ttUF79zIZXDKfwPD3GEZWmMCso4eT4jHUYa9Q+OQPqEd2FPW5MuxTU7sbMroaBelXUTwks8bS1d2i9DICaWq5+Pzgbl674O2Jvy9opmIa8kfelAdmZY0hiBRdV6A287afNhgqWRHe9LnNGw= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1756481730; c=relaxed/simple; bh=2TLw3pkBQcva92wprOdUHqezoM2m2ggtACrxJV3anSM=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=VE4DFK816D1TBgajx8gHYSeVgQ4fJDYqXjBK/SBdU4eTj/fD+RfdTboYszA2+26YwQDCCVBmlG2n/Qt2TVDA4qxSPYYNK0tCVtYRK4ykv43zRyUjG498LXX/oAj1JTvQ/s0sqfdJMZVtes8zzUVvErOw3qCov/cA/dciuf8ickc= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com; spf=pass smtp.mailfrom=arm.com; arc=none smtp.client-ip=217.140.110.172 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=arm.com Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 9D29A1C01; Fri, 29 Aug 2025 08:35:18 -0700 (PDT) Received: from e125769.cambridge.arm.com (e125769.cambridge.arm.com [10.1.196.27]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id CB8583F738; Fri, 29 Aug 2025 08:35:25 -0700 (PDT) From: Ryan Roberts To: Catalin Marinas , Will Deacon , Mark Rutland , James Morse Cc: Ryan Roberts , linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Subject: [RFC PATCH v1 2/2] arm64: tlbflush: Don't broadcast if mm was only active on local cpu Date: Fri, 29 Aug 2025 16:35:08 +0100 Message-ID: <20250829153510.2401161-3-ryan.roberts@arm.com> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20250829153510.2401161-1-ryan.roberts@arm.com> References: <20250829153510.2401161-1-ryan.roberts@arm.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset="utf-8" There are 3 variants of tlb flush that invalidate user mappings: flush_tlb_mm(), flush_tlb_page() and __flush_tlb_range(). All of these would previously unconditionally broadcast their tlbis to all cpus in the inner shareable domain. But this is a waste of effort if we can prove that the mm for which we are flushing the mappings has only ever been active on the local cpu. In that case, it is safe to avoid the broadcast and simply invalidate the current cpu. So let's track in mm_context_t::active_cpu either the mm has never been active on any cpu, has been active on more than 1 cpu, or has been active on precisely 1 cpu - and in that case, which one. We update this when switching context, being careful to ensure that it gets updated *before* installing the mm's pgtables. On the reader side, we ensure we read *after* the previous write(s) to the pgtable(s) that necessitated the tlb flush have completed. This guarrantees that if a cpu that is doing a tlb flush sees it's own id in active_cpu, then the old pgtable entry cannot have been seen by any other cpu and we can flush only the local cpu. Signed-off-by: Ryan Roberts Reviewed-by: Catalin Marinas --- arch/arm64/include/asm/mmu.h | 12 ++++ arch/arm64/include/asm/mmu_context.h | 2 + arch/arm64/include/asm/tlbflush.h | 90 +++++++++++++++++++++++++--- arch/arm64/mm/context.c | 30 +++++++++- 4 files changed, 123 insertions(+), 11 deletions(-) diff --git a/arch/arm64/include/asm/mmu.h b/arch/arm64/include/asm/mmu.h index 6e8aa8e72601..ca32fb860309 100644 --- a/arch/arm64/include/asm/mmu.h +++ b/arch/arm64/include/asm/mmu.h @@ -17,6 +17,17 @@ #include #include =20 +/* + * Sentinal values for mm_context_t::active_cpu. ACTIVE_CPU_NONE indicates= the + * mm has never been active on any CPU. ACTIVE_CPU_MULTIPLE indicates the = mm + * has been active on multiple CPUs. Any other value is the ID of the sing= le + * CPU that the mm has been active on. + */ +enum active_cpu { + ACTIVE_CPU_NONE =3D UINT_MAX, + ACTIVE_CPU_MULTIPLE =3D UINT_MAX - 1, +}; + typedef struct { atomic64_t id; #ifdef CONFIG_COMPAT @@ -26,6 +37,7 @@ typedef struct { void *vdso; unsigned long flags; u8 pkey_allocation_map; + unsigned int active_cpu; } mm_context_t; =20 /* diff --git a/arch/arm64/include/asm/mmu_context.h b/arch/arm64/include/asm/= mmu_context.h index 0dbe3b29049b..db2505a8eba8 100644 --- a/arch/arm64/include/asm/mmu_context.h +++ b/arch/arm64/include/asm/mmu_context.h @@ -180,6 +180,8 @@ init_new_context(struct task_struct *tsk, struct mm_str= uct *mm) /* pkey 0 is the default, so always reserve it. */ mm->context.pkey_allocation_map =3D BIT(0); =20 + WRITE_ONCE(mm->context.active_cpu, ACTIVE_CPU_NONE); + return 0; } =20 diff --git a/arch/arm64/include/asm/tlbflush.h b/arch/arm64/include/asm/tlb= flush.h index f66b8c4696d0..651440e0aff9 100644 --- a/arch/arm64/include/asm/tlbflush.h +++ b/arch/arm64/include/asm/tlbflush.h @@ -182,6 +182,52 @@ static inline unsigned long get_trans_granule(void) (__pages >> (5 * (scale) + 1)) - 1; \ }) =20 +/* + * Determines whether the user tlbi invalidation can be performed only on = the + * local CPU or whether it needs to be broadcast. (Returns true for local). + * Additionally issues appropriate barrier to ensure prior pgtable updates= are + * visible to the table walker. Must be paired with flush_tlb_user_post(). + */ +static inline bool flush_tlb_user_pre(struct mm_struct *mm) +{ + unsigned int self, active; + bool local; + + migrate_disable(); + + self =3D smp_processor_id(); + + /* + * The load of mm->context.active_cpu must not be reordered before the + * store to the pgtable that necessitated this flush. This ensures that + * if the value read is our cpu id, then no other cpu can have seen the + * old pgtable value and therefore does not need this old value to be + * flushed from its tlb. But we don't want to upgrade the dsb(ishst), + * needed to make the pgtable updates visible to the walker, to a + * dsb(ish) by default. So speculatively load without a barrier and if + * it indicates our cpu id, then upgrade the barrier and re-load. + */ + active =3D READ_ONCE(mm->context.active_cpu); + if (active =3D=3D self) { + dsb(ish); + active =3D READ_ONCE(mm->context.active_cpu); + } else { + dsb(ishst); + } + + local =3D active =3D=3D self; + if (!local) + migrate_enable(); + + return local; +} + +static inline void flush_tlb_user_post(bool local) +{ + if (local) + migrate_enable(); +} + /* * TLB Invalidation * =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D @@ -274,11 +320,18 @@ static inline void flush_tlb_all(void) static inline void flush_tlb_mm(struct mm_struct *mm) { unsigned long asid; + bool local; =20 - dsb(ishst); + local =3D flush_tlb_user_pre(mm); asid =3D __TLBI_VADDR(0, ASID(mm)); - __tlbi(aside1is, asid); - __tlbi_user(aside1is, asid); + if (local) { + __tlbi(aside1, asid); + __tlbi_user(aside1, asid); + } else { + __tlbi(aside1is, asid); + __tlbi_user(aside1is, asid); + } + flush_tlb_user_post(local); dsb(ish); mmu_notifier_arch_invalidate_secondary_tlbs(mm, 0, -1UL); } @@ -287,11 +340,18 @@ static inline void __flush_tlb_page_nosync(struct mm_= struct *mm, unsigned long uaddr) { unsigned long addr; + bool local; =20 - dsb(ishst); + local =3D flush_tlb_user_pre(mm); addr =3D __TLBI_VADDR(uaddr, ASID(mm)); - __tlbi(vale1is, addr); - __tlbi_user(vale1is, addr); + if (local) { + __tlbi(vale1, addr); + __tlbi_user(vale1, addr); + } else { + __tlbi(vale1is, addr); + __tlbi_user(vale1is, addr); + } + flush_tlb_user_post(local); mmu_notifier_arch_invalidate_secondary_tlbs(mm, uaddr & PAGE_MASK, (uaddr & PAGE_MASK) + PAGE_SIZE); } @@ -437,6 +497,8 @@ static inline bool __flush_tlb_range_limit_excess(unsig= ned long start, enum tlbi_op { TLBI_VALE1IS, TLBI_VAE1IS, + TLBI_VALE1, + TLBI_VAE1, }; =20 #define flush_tlb_range_op(op, ...) \ @@ -448,6 +510,12 @@ do { \ case TLBI_VAE1IS: \ __flush_tlb_range_op(vae1is, __VA_ARGS__); \ break; \ + case TLBI_VALE1: \ + __flush_tlb_range_op(vale1, __VA_ARGS__); \ + break; \ + case TLBI_VAE1: \ + __flush_tlb_range_op(vae1, __VA_ARGS__); \ + break; \ default: \ BUILD_BUG_ON_MSG(1, "Unknown TLBI op"); \ } \ @@ -460,6 +528,7 @@ static inline void __flush_tlb_range_nosync(struct mm_s= truct *mm, { unsigned long asid, pages; enum tlbi_op tlbi_op; + bool local; =20 start =3D round_down(start, stride); end =3D round_up(end, stride); @@ -470,13 +539,18 @@ static inline void __flush_tlb_range_nosync(struct mm= _struct *mm, return; } =20 - dsb(ishst); + local =3D flush_tlb_user_pre(mm); asid =3D ASID(mm); =20 - tlbi_op =3D last_level ? TLBI_VALE1IS : TLBI_VAE1IS; + if (local) + tlbi_op =3D last_level ? TLBI_VALE1 : TLBI_VAE1; + else + tlbi_op =3D last_level ? TLBI_VALE1IS : TLBI_VAE1IS; + flush_tlb_range_op(tlbi_op, start, pages, stride, asid, tlb_level, true, lpa2_is_enabled()); =20 + flush_tlb_user_post(local); mmu_notifier_arch_invalidate_secondary_tlbs(mm, start, end); } =20 diff --git a/arch/arm64/mm/context.c b/arch/arm64/mm/context.c index b2ac06246327..adf4fc26ddb6 100644 --- a/arch/arm64/mm/context.c +++ b/arch/arm64/mm/context.c @@ -214,9 +214,10 @@ static u64 new_context(struct mm_struct *mm) =20 void check_and_switch_context(struct mm_struct *mm) { - unsigned long flags; - unsigned int cpu; + unsigned int cpu =3D smp_processor_id(); u64 asid, old_active_asid; + unsigned int active; + unsigned long flags; =20 if (system_supports_cnp()) cpu_set_reserved_ttbr0(); @@ -251,7 +252,6 @@ void check_and_switch_context(struct mm_struct *mm) atomic64_set(&mm->context.id, asid); } =20 - cpu =3D smp_processor_id(); if (cpumask_test_and_clear_cpu(cpu, &tlb_flush_pending)) local_flush_tlb_all(); =20 @@ -262,6 +262,30 @@ void check_and_switch_context(struct mm_struct *mm) =20 arm64_apply_bp_hardening(); =20 + /* + * Update mm->context.active_cpu in such a manner that we avoid cmpxchg + * and dsb unless we definitely need it. If initially ACTIVE_CPU_NONE + * then we are the first cpu to run so set it to our id. If initially + * any id other than ours, we are the second cpu to run so set it to + * ACTIVE_CPU_MULTIPLE. If we update the value then we must issue + * dsb(ishst) to ensure stores to mm->context.active_cpu are ordered + * against the TTBR0 write in cpu_switch_mm()/uaccess_enable(); the + * store must be visible to another cpu before this cpu could have + * populated any TLB entries based on the pgtables that will be + * installed. + */ + active =3D READ_ONCE(mm->context.active_cpu); + if (active !=3D cpu && active !=3D ACTIVE_CPU_MULTIPLE) { + if (active =3D=3D ACTIVE_CPU_NONE) + active =3D cmpxchg_relaxed(&mm->context.active_cpu, + ACTIVE_CPU_NONE, cpu); + + if (active !=3D ACTIVE_CPU_NONE) + WRITE_ONCE(mm->context.active_cpu, ACTIVE_CPU_MULTIPLE); + + dsb(ishst); + } + /* * Defer TTBR0_EL1 setting for user threads to uaccess_enable() when * emulating PAN. --=20 2.43.0