From nobody Wed Apr 8 20:00:15 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 6F9BCC38A02 for ; Fri, 28 Oct 2022 23:40:48 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229890AbiJ1Xkq (ORCPT ); Fri, 28 Oct 2022 19:40:46 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38446 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229670AbiJ1Xkn (ORCPT ); Fri, 28 Oct 2022 19:40:43 -0400 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 0A95A71706 for ; Fri, 28 Oct 2022 16:40:41 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id B035262AD8 for ; Fri, 28 Oct 2022 23:40:40 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 5141EC433D6; Fri, 28 Oct 2022 23:40:39 +0000 (UTC) Authentication-Results: smtp.kernel.org; dkim=pass (1024-bit key) header.d=zx2c4.com header.i=@zx2c4.com header.b="PB+weiRE" DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=zx2c4.com; s=20210105; t=1667000437; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding; bh=cp1Xr1b0al/fXFCbuvHhFS7yT6OAQVKZYDYFdLMcyIc=; b=PB+weiREJSJgZVfkz58SPQiIAV1W8QU2tt8R107aoTX7yo+d9d/uWKMtINVzQoODHODEna U5yRE23sBGGxWaYcWfeqfOftRpMkOEeEMrEjsjf+yKvWMyidVak8fO4av6UM8Sd3KXr0l9 vc89vsRwN9ctWa2tSkoKQj2Nm+DoXHo= Received: by mail.zx2c4.com (ZX2C4 Mail Server) with ESMTPSA id 2b502cb8 (TLSv1.3:TLS_AES_256_GCM_SHA384:256:NO); Fri, 28 Oct 2022 23:40:36 +0000 (UTC) From: "Jason A. Donenfeld" To: catalin.marinas@arm.com, will@kernel.org, jean-philippe@linaro.org, ardb@kernel.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: "Jason A. Donenfeld" Subject: [PATCH] random: remove early archrandom abstraction Date: Sat, 29 Oct 2022 01:40:25 +0200 Message-Id: <20221028234025.82222-1-Jason@zx2c4.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" The arch_get_random*_early() abstraction is not completely useful and adds complexity, because it's not a given that there will be no calls to arch_get_random*() between random_init_early(), which uses arch_get_random*_early(), and init_cpu_features(). During that gap, crng_reseed() might be called, which uses arch_get_random*(), since it's mostly not init code. Instead we can test whether we're in the early phase in arch_get_random*() itself, and in doing so avoid all ambiguity about where we are. Fortunately, the only architecture that currently implements arch_get_random*_early() also has an alternatives-based cpu feature system, one flag of which determines whether the other flags have been initialized. This makes it possible to do the early check with zero cost once the system is initialized. Signed-off-by: Jason A. Donenfeld --- Catalin - Though this touches arm64's archrandom.h, I intend to take this through the random.git tree, if that's okay. I have other patches that will build off of this one. -Jason arch/arm64/include/asm/archrandom.h | 57 ++++++----------------------- drivers/char/random.c | 4 +- include/linux/random.h | 20 ---------- 3 files changed, 14 insertions(+), 67 deletions(-) diff --git a/arch/arm64/include/asm/archrandom.h b/arch/arm64/include/asm/a= rchrandom.h index 109e2a4454be..8a059a9033af 100644 --- a/arch/arm64/include/asm/archrandom.h +++ b/arch/arm64/include/asm/archrandom.h @@ -58,6 +58,16 @@ static inline bool __arm64_rndrrs(unsigned long *v) return ok; } =20 +static __always_inline bool __cpu_has_rng(void) +{ + if (!system_capabilities_finalized()) { + /* Open code as we run prior to the first call to cpufeature. */ + unsigned long ftr =3D read_sysreg_s(SYS_ID_AA64ISAR0_EL1); + return (ftr >> ID_AA64ISAR0_EL1_RNDR_SHIFT) & 0xf; + } + return cpus_have_const_cap(ARM64_HAS_RNG); +} + static inline size_t __must_check arch_get_random_longs(unsigned long *v, = size_t max_longs) { /* @@ -66,7 +76,7 @@ static inline size_t __must_check arch_get_random_longs(u= nsigned long *v, size_t * cpufeature code and with potential scheduling between CPUs * with and without the feature. */ - if (max_longs && cpus_have_const_cap(ARM64_HAS_RNG) && __arm64_rndr(v)) + if (max_longs && __cpu_has_rng() && __arm64_rndr(v)) return 1; return 0; } @@ -108,53 +118,10 @@ static inline size_t __must_check arch_get_random_see= d_longs(unsigned long *v, s * reseeded after each invocation. This is not a 100% fit but good * enough to implement this API if no other entropy source exists. */ - if (cpus_have_const_cap(ARM64_HAS_RNG) && __arm64_rndrrs(v)) + if (__cpu_has_rng() && __arm64_rndrrs(v)) return 1; =20 return 0; } =20 -static inline bool __init __early_cpu_has_rndr(void) -{ - /* Open code as we run prior to the first call to cpufeature. */ - unsigned long ftr =3D read_sysreg_s(SYS_ID_AA64ISAR0_EL1); - return (ftr >> ID_AA64ISAR0_EL1_RNDR_SHIFT) & 0xf; -} - -static inline size_t __init __must_check -arch_get_random_seed_longs_early(unsigned long *v, size_t max_longs) -{ - WARN_ON(system_state !=3D SYSTEM_BOOTING); - - if (!max_longs) - return 0; - - if (smccc_trng_available) { - struct arm_smccc_res res; - - max_longs =3D min_t(size_t, 3, max_longs); - arm_smccc_1_1_invoke(ARM_SMCCC_TRNG_RND64, max_longs * 64, &res); - if ((int)res.a0 >=3D 0) { - switch (max_longs) { - case 3: - *v++ =3D res.a1; - fallthrough; - case 2: - *v++ =3D res.a2; - fallthrough; - case 1: - *v++ =3D res.a3; - break; - } - return max_longs; - } - } - - if (__early_cpu_has_rndr() && __arm64_rndr(v)) - return 1; - - return 0; -} -#define arch_get_random_seed_longs_early arch_get_random_seed_longs_early - #endif /* _ASM_ARCHRANDOM_H */ diff --git a/drivers/char/random.c b/drivers/char/random.c index 6f323344d0b9..e3cf4f51ed58 100644 --- a/drivers/char/random.c +++ b/drivers/char/random.c @@ -813,13 +813,13 @@ void __init random_init_early(const char *command_lin= e) #endif =20 for (i =3D 0, arch_bits =3D sizeof(entropy) * 8; i < ARRAY_SIZE(entropy);= ) { - longs =3D arch_get_random_seed_longs_early(entropy, ARRAY_SIZE(entropy) = - i); + longs =3D arch_get_random_seed_longs(entropy, ARRAY_SIZE(entropy) - i); if (longs) { _mix_pool_bytes(entropy, sizeof(*entropy) * longs); i +=3D longs; continue; } - longs =3D arch_get_random_longs_early(entropy, ARRAY_SIZE(entropy) - i); + longs =3D arch_get_random_longs(entropy, ARRAY_SIZE(entropy) - i); if (longs) { _mix_pool_bytes(entropy, sizeof(*entropy) * longs); i +=3D longs; diff --git a/include/linux/random.h b/include/linux/random.h index 182780cafd45..2bdd3add3400 100644 --- a/include/linux/random.h +++ b/include/linux/random.h @@ -153,26 +153,6 @@ declare_get_random_var_wait(long, unsigned long) =20 #include =20 -/* - * Called from the boot CPU during startup; not valid to call once - * secondary CPUs are up and preemption is possible. - */ -#ifndef arch_get_random_seed_longs_early -static inline size_t __init arch_get_random_seed_longs_early(unsigned long= *v, size_t max_longs) -{ - WARN_ON(system_state !=3D SYSTEM_BOOTING); - return arch_get_random_seed_longs(v, max_longs); -} -#endif - -#ifndef arch_get_random_longs_early -static inline bool __init arch_get_random_longs_early(unsigned long *v, si= ze_t max_longs) -{ - WARN_ON(system_state !=3D SYSTEM_BOOTING); - return arch_get_random_longs(v, max_longs); -} -#endif - #ifdef CONFIG_SMP int random_prepare_cpu(unsigned int cpu); int random_online_cpu(unsigned int cpu); --=20 2.38.1