From nobody Sun Sep 7 12:18:18 2025 Received: from mail-wm1-f74.google.com (mail-wm1-f74.google.com [209.85.128.74]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 84989286880 for ; Wed, 16 Jul 2025 03:22:37 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.128.74 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1752636159; cv=none; b=d3IfvEl1qFJ097kpVFmNUElwOvwuQifVJCJ+fgVVBd5ZqW9n1qkL8XAYJRlnDnlz2PUK9LqBYAAbA9rNF1Vj399rHH48GW3LHwZH/eneMvy0yI3sxK6J12LmoT5G1SbwlRREyJTU5YNtJlkAY5ewmRphFWLj1qAA4ZtuKyRZtSM= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1752636159; c=relaxed/simple; bh=hyOCEHD2BP/BHB/wXZzexTCfs17VdhN8Xct7u6Z/6TQ=; h=Date:In-Reply-To:Mime-Version:References:Message-ID:Subject:From: To:Cc:Content-Type; b=SW3OTdCKO7DJi7ToT/a+JL6HQshuaUxX7ebGLTQtV7FJs1styx3lymc06VhdBvOWc8rK116Q+/wKIIgEEQINns8s6YPCogWS1WxHVPlJYDj0SAsy7E3esGov4tPGQlZuBLNCcEhEv256IIx9GR7sFeyjDKUKHf/lPgWs+UgRzuc= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=google.com; spf=pass smtp.mailfrom=flex--ardb.bounces.google.com; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b=MPNNShgM; arc=none smtp.client-ip=209.85.128.74 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=google.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=flex--ardb.bounces.google.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b="MPNNShgM" Received: by mail-wm1-f74.google.com with SMTP id 5b1f17b1804b1-456175dba68so22662555e9.2 for ; Tue, 15 Jul 2025 20:22:37 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1752636156; x=1753240956; darn=vger.kernel.org; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:from:to:cc:subject:date:message-id:reply-to; bh=Ie7ghjTAXJ1RjJATnR0+RLls6HNvil6BXsqkKY03oqo=; b=MPNNShgMVzBAwX8hvo0Vyr1Z/3QCpWPW101KdoziurF/I06Lf0CKjEUhYx3Rr8FuLq b1cu05ngGtymzeCRH55On40MJ3dgxbPS8/t1TXhXeoOnbmvjIy5nKfGFu8G+BGO3P2wb Lwh17do0CiT125c11GyMKbZaurMICS/HWYsXYBTbT3kQMtM+/WEokBz7oitcjD4HOZun EWyHbk6uGXUEamLk9Bvm/5kDnW69pZdWdF8Qb+aCxj8yPSj97Nnx4Rgae/qvLLrSdk6Y egp7RLCF1OJF/04TEtgbzX9mN8iXgxxr9TDuEk5tk/dLwk6E54wnaGXVPqeI0BoTQQPc JZRg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1752636156; x=1753240956; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=Ie7ghjTAXJ1RjJATnR0+RLls6HNvil6BXsqkKY03oqo=; b=uMxyXxiAKI0PK57YIjrWjuWsAJWHg9Y9DbddLDm104hG3EiNKyTl6GYF61VbVmvExA WHljSk4HV4dXyKvQCvcNCD6dgwPkgL9q3vv5YrXOnwZyvcu8/R7eXB7NV5V2ecVEjBJ+ XuYm6UNw6P5B8WCI8Jaq7ktkewTwGLNpq6nvwQXjLHTZ7Eqi014XWkVSHh8lMuv7SY00 XMR20Ve/iIraf5ilgfoohxRztEfmNzlVE2Os+tccHST8/7QSaPKGkoxhMP4fFDw4nNkY AZC3hldg0jEu7q2sA+FSeLaz82bwY9Al+qQ1+pGQzN2QTas1jIdGIiEOMt5w3052pZ71 qEgA== X-Gm-Message-State: AOJu0Yw0fcUx6ZksL/aA0FapDYbJDzIojskbeXw3h4XLEfGtG3278zqL czaLEim9M3P3tyJiKQAND0MzRhcPaGJSnhqAvaeZA39B7SRT7V5Oks+J+10TEX9F449M2kFBXro VIJSgmwhq+2l/J1PJuL3hEODCt2GiB8OSwO9y8JwcfvYa6wsuio9vweKyvQUtPzWkDG5XMM+nD8 sAiBeCd1GQK2Q9dQnSjKc9GWplL+Umriy1zQ== X-Google-Smtp-Source: AGHT+IFWEjTGKJC0P+l/6U1ojJjIwXduaOUN+ZHQZNlS4h1A3hUnAS0bLiY0kzHEFrxpEU9mloI1sYNT X-Received: from wmsr5.prod.google.com ([2002:a05:600c:8b05:b0:456:15bd:a297]) (user=ardb job=prod-delivery.src-stubby-dispatcher) by 2002:a05:600c:1d20:b0:43c:e70d:44f0 with SMTP id 5b1f17b1804b1-4562e36c6d3mr7961485e9.19.1752636156038; Tue, 15 Jul 2025 20:22:36 -0700 (PDT) Date: Wed, 16 Jul 2025 05:18:23 +0200 In-Reply-To: <20250716031814.2096113-24-ardb+git@google.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: Mime-Version: 1.0 References: <20250716031814.2096113-24-ardb+git@google.com> X-Developer-Key: i=ardb@kernel.org; a=openpgp; fpr=F43D03328115A198C90016883D200E9CA6329909 X-Developer-Signature: v=1; a=openpgp-sha256; l=7692; i=ardb@kernel.org; h=from:subject; bh=Dgsh787GUJSIpiPFllr2H7qbaHO4zGl5ERrfzXk+iCw=; b=owGbwMvMwCVmkMcZplerG8N4Wi2JIaNcSq26R/SC4a5fzd5rHNdKcs52YuM+4layV+nWcbeDr 31WtKt3lLIwiHExyIopsgjM/vtu5+mJUrXOs2Rh5rAygQxh4OIUgIls/sXwz/Ki78XyI6FvFkvu ULS+nP12H2PjfI0jL944fQw8+VeKW5Lhr3xO+dlShp1cwvkFkob//WZ8kNoc5xMlI2KQaFDyYNl rDgA= X-Mailer: git-send-email 2.50.0.727.gbf7dc18ff4-goog Message-ID: <20250716031814.2096113-32-ardb+git@google.com> Subject: [PATCH v5 08/22] x86/sev: Pass SVSM calling area down to early page state change API From: Ard Biesheuvel To: linux-kernel@vger.kernel.org Cc: linux-efi@vger.kernel.org, x86@kernel.org, Ard Biesheuvel , Borislav Petkov , Ingo Molnar , Kevin Loughlin , Tom Lendacky , Josh Poimboeuf , Peter Zijlstra , Nikunj A Dadhania Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset="utf-8" From: Ard Biesheuvel The early page state change API is mostly only used very early, when only the boot time SVSM calling area is in use. However, this API is also called by the kexec finishing code, which runs very late, and potentially from a different CPU (which uses a different calling area). To avoid pulling the per-CPU SVSM calling area pointers and related SEV state into the startup code, refactor the page state change API so the SVSM calling area virtual and physical addresses can be provided by the caller. No functional change intended. Signed-off-by: Ard Biesheuvel --- arch/x86/boot/compressed/sev.c | 12 +++++++++--- arch/x86/boot/startup/sev-shared.c | 17 +++++++++-------- arch/x86/boot/startup/sev-startup.c | 11 +++++++---- arch/x86/coco/sev/core.c | 3 ++- arch/x86/include/asm/sev-internal.h | 3 ++- 5 files changed, 29 insertions(+), 17 deletions(-) diff --git a/arch/x86/boot/compressed/sev.c b/arch/x86/boot/compressed/sev.c index f714235d3222..18b0ccf517eb 100644 --- a/arch/x86/boot/compressed/sev.c +++ b/arch/x86/boot/compressed/sev.c @@ -65,7 +65,9 @@ void snp_set_page_private(unsigned long paddr) if (!sev_snp_enabled()) return; =20 - __page_state_change(paddr, paddr, SNP_PAGE_STATE_PRIVATE); + __page_state_change(paddr, paddr, SNP_PAGE_STATE_PRIVATE, + (struct svsm_ca *)boot_svsm_caa_pa, + boot_svsm_caa_pa); } =20 void snp_set_page_shared(unsigned long paddr) @@ -73,7 +75,9 @@ void snp_set_page_shared(unsigned long paddr) if (!sev_snp_enabled()) return; =20 - __page_state_change(paddr, paddr, SNP_PAGE_STATE_SHARED); + __page_state_change(paddr, paddr, SNP_PAGE_STATE_SHARED, + (struct svsm_ca *)boot_svsm_caa_pa, + boot_svsm_caa_pa); } =20 bool early_setup_ghcb(void) @@ -99,7 +103,9 @@ bool early_setup_ghcb(void) void snp_accept_memory(phys_addr_t start, phys_addr_t end) { for (phys_addr_t pa =3D start; pa < end; pa +=3D PAGE_SIZE) - __page_state_change(pa, pa, SNP_PAGE_STATE_PRIVATE); + __page_state_change(pa, pa, SNP_PAGE_STATE_PRIVATE, + (struct svsm_ca *)boot_svsm_caa_pa, + boot_svsm_caa_pa); } =20 void sev_es_shutdown_ghcb(void) diff --git a/arch/x86/boot/startup/sev-shared.c b/arch/x86/boot/startup/sev= -shared.c index eb241ff1156d..83ca97df0808 100644 --- a/arch/x86/boot/startup/sev-shared.c +++ b/arch/x86/boot/startup/sev-shared.c @@ -598,7 +598,8 @@ static int __head svsm_call_msr_protocol(struct svsm_ca= ll *call) return ret; } =20 -static void __head svsm_pval_4k_page(unsigned long paddr, bool validate) +static void __head svsm_pval_4k_page(unsigned long paddr, bool validate, + struct svsm_ca *caa, u64 caa_pa) { struct svsm_pvalidate_call *pc; struct svsm_call call =3D {}; @@ -611,10 +612,10 @@ static void __head svsm_pval_4k_page(unsigned long pa= ddr, bool validate) */ flags =3D native_local_irq_save(); =20 - call.caa =3D svsm_get_caa(); + call.caa =3D caa; =20 pc =3D (struct svsm_pvalidate_call *)call.caa->svsm_buffer; - pc_pa =3D svsm_get_caa_pa() + offsetof(struct svsm_ca, svsm_buffer); + pc_pa =3D caa_pa + offsetof(struct svsm_ca, svsm_buffer); =20 pc->num_entries =3D 1; pc->cur_index =3D 0; @@ -639,12 +640,12 @@ static void __head svsm_pval_4k_page(unsigned long pa= ddr, bool validate) } =20 static void __head pvalidate_4k_page(unsigned long vaddr, unsigned long pa= ddr, - bool validate) + bool validate, struct svsm_ca *caa, u64 caa_pa) { int ret; =20 if (snp_vmpl) { - svsm_pval_4k_page(paddr, validate); + svsm_pval_4k_page(paddr, validate, caa, caa_pa); } else { ret =3D pvalidate(vaddr, RMP_PG_SIZE_4K, validate); if (ret) @@ -653,7 +654,7 @@ static void __head pvalidate_4k_page(unsigned long vadd= r, unsigned long paddr, } =20 static void __head __page_state_change(unsigned long vaddr, unsigned long = paddr, - enum psc_op op) + enum psc_op op, struct svsm_ca *caa, u64 caa_pa) { u64 val, msr; =20 @@ -662,7 +663,7 @@ static void __head __page_state_change(unsigned long va= ddr, unsigned long paddr, * state change in the RMP table. */ if (op =3D=3D SNP_PAGE_STATE_SHARED) - pvalidate_4k_page(vaddr, paddr, false); + pvalidate_4k_page(vaddr, paddr, false, caa, caa_pa); =20 /* Save the current GHCB MSR value */ msr =3D sev_es_rd_ghcb_msr(); @@ -684,7 +685,7 @@ static void __head __page_state_change(unsigned long va= ddr, unsigned long paddr, * consistent with the RMP entry. */ if (op =3D=3D SNP_PAGE_STATE_PRIVATE) - pvalidate_4k_page(vaddr, paddr, true); + pvalidate_4k_page(vaddr, paddr, true, caa, caa_pa); } =20 /* diff --git a/arch/x86/boot/startup/sev-startup.c b/arch/x86/boot/startup/se= v-startup.c index b4e2cb7bc44a..7aabda0b378e 100644 --- a/arch/x86/boot/startup/sev-startup.c +++ b/arch/x86/boot/startup/sev-startup.c @@ -132,7 +132,8 @@ noinstr void __sev_put_ghcb(struct ghcb_state *state) =20 void __head early_set_pages_state(unsigned long vaddr, unsigned long paddr, - unsigned long npages, enum psc_op op) + unsigned long npages, enum psc_op op, + struct svsm_ca *caa, u64 caa_pa) { unsigned long paddr_end; =20 @@ -142,7 +143,7 @@ early_set_pages_state(unsigned long vaddr, unsigned lon= g paddr, paddr_end =3D paddr + (npages << PAGE_SHIFT); =20 while (paddr < paddr_end) { - __page_state_change(vaddr, paddr, op); + __page_state_change(vaddr, paddr, op, caa, caa_pa); =20 vaddr +=3D PAGE_SIZE; paddr +=3D PAGE_SIZE; @@ -165,7 +166,8 @@ void __head early_snp_set_memory_private(unsigned long = vaddr, unsigned long padd * Ask the hypervisor to mark the memory pages as private in the RMP * table. */ - early_set_pages_state(vaddr, paddr, npages, SNP_PAGE_STATE_PRIVATE); + early_set_pages_state(vaddr, paddr, npages, SNP_PAGE_STATE_PRIVATE, + svsm_get_caa(), svsm_get_caa_pa()); } =20 void __head early_snp_set_memory_shared(unsigned long vaddr, unsigned long= paddr, @@ -181,7 +183,8 @@ void __head early_snp_set_memory_shared(unsigned long v= addr, unsigned long paddr return; =20 /* Ask hypervisor to mark the memory pages shared in the RMP table. */ - early_set_pages_state(vaddr, paddr, npages, SNP_PAGE_STATE_SHARED); + early_set_pages_state(vaddr, paddr, npages, SNP_PAGE_STATE_SHARED, + svsm_get_caa(), svsm_get_caa_pa()); } =20 /* diff --git a/arch/x86/coco/sev/core.c b/arch/x86/coco/sev/core.c index bea67d017bf0..7a86a2fe494d 100644 --- a/arch/x86/coco/sev/core.c +++ b/arch/x86/coco/sev/core.c @@ -585,7 +585,8 @@ static void set_pages_state(unsigned long vaddr, unsign= ed long npages, int op) =20 /* Use the MSR protocol when a GHCB is not available. */ if (!boot_ghcb) - return early_set_pages_state(vaddr, __pa(vaddr), npages, op); + return early_set_pages_state(vaddr, __pa(vaddr), npages, op, + svsm_get_caa(), svsm_get_caa_pa()); =20 vaddr =3D vaddr & PAGE_MASK; vaddr_end =3D vaddr + (npages << PAGE_SHIFT); diff --git a/arch/x86/include/asm/sev-internal.h b/arch/x86/include/asm/sev= -internal.h index ffe4755962fe..3b72d8217827 100644 --- a/arch/x86/include/asm/sev-internal.h +++ b/arch/x86/include/asm/sev-internal.h @@ -55,7 +55,8 @@ DECLARE_PER_CPU(struct sev_es_runtime_data*, runtime_data= ); DECLARE_PER_CPU(struct sev_es_save_area *, sev_vmsa); =20 void early_set_pages_state(unsigned long vaddr, unsigned long paddr, - unsigned long npages, enum psc_op op); + unsigned long npages, enum psc_op op, + struct svsm_ca *ca, u64 caa_pa); =20 DECLARE_PER_CPU(struct svsm_ca *, svsm_caa); DECLARE_PER_CPU(u64, svsm_caa_pa); --=20 2.50.0.727.gbf7dc18ff4-goog