On Wed, May 14, 2025 at 03:17:40AM +0800, Edgecombe, Rick P wrote:
> On Thu, 2025-04-24 at 11:05 +0800, Yan Zhao wrote:
> > From: Xiaoyao Li <xiaoyao.li@intel.com>
> >
> > KVM invokes tdx_clear_page() to zero pages using movdir64b().
> > Include level information to enable tdx_clear_page() to zero a huge page.
> >
> > [Yan: split out, let tdx_clear_page() accept level]
> >
> > Signed-off-by: Xiaoyao Li <xiaoyao.li@intel.com>
> > Signed-off-by: Isaku Yamahata <isaku.yamahata@intel.com>
> > Signed-off-by: Yan Zhao <yan.y.zhao@intel.com>
> > ---
> > arch/x86/kvm/vmx/tdx.c | 19 ++++++++++++++-----
> > 1 file changed, 14 insertions(+), 5 deletions(-)
> >
> > diff --git a/arch/x86/kvm/vmx/tdx.c b/arch/x86/kvm/vmx/tdx.c
> > index 03885cb2869b..1186085795ac 100644
> > --- a/arch/x86/kvm/vmx/tdx.c
> > +++ b/arch/x86/kvm/vmx/tdx.c
> > @@ -276,7 +276,7 @@ static inline void tdx_disassociate_vp(struct kvm_vcpu *vcpu)
> > vcpu->cpu = -1;
> > }
> >
> > -static void tdx_clear_page(struct page *page)
> > +static void __tdx_clear_page(struct page *page)
> > {
> > const void *zero_page = (const void *) page_to_virt(ZERO_PAGE(0));
> > void *dest = page_to_virt(page);
> > @@ -295,6 +295,15 @@ static void tdx_clear_page(struct page *page)
> > __mb();
> > }
> >
> > +static void tdx_clear_page(struct page *page, int level)
> > +{
> > + unsigned long nr = KVM_PAGES_PER_HPAGE(level);
> > + unsigned long idx = 0;
> > +
> > + while (nr--)
> > + __tdx_clear_page(nth_page(page, idx++));
>
> You shouldn't need both idx and nr.
>
> > +}
>
> Since tdx_clear_page() has a __mb(), it is probably worth checking that this
> generates efficient code, considering the loops within loops pattern.
The concern makes sense!
Will convert level to size and use "for (i = 0; i < size; i += 64)" for
movdir64b().
> > +
> > static void tdx_no_vcpus_enter_start(struct kvm *kvm)
> > {
> > struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
> > @@ -340,11 +349,10 @@ static int tdx_reclaim_page(struct page *page)
> >
> > r = __tdx_reclaim_page(page);
> > if (!r)
> > - tdx_clear_page(page);
> > + tdx_clear_page(page, PG_LEVEL_4K);
> > return r;
> > }
> >
> > -
> > /*
> > * Reclaim the TD control page(s) which are crypto-protected by TDX guest's
> > * private KeyID. Assume the cache associated with the TDX private KeyID has
> > @@ -588,7 +596,7 @@ static void tdx_reclaim_td_control_pages(struct kvm *kvm)
> > pr_tdx_error(TDH_PHYMEM_PAGE_WBINVD, err);
> > return;
> > }
> > - tdx_clear_page(kvm_tdx->td.tdr_page);
> > + tdx_clear_page(kvm_tdx->td.tdr_page, PG_LEVEL_4K);
>
> Why not the __tdx_clear_page() variant? The patch adds it, but doesn't really
> use it. Just implement it all in tdx_clear_page() then.
Ok.
> >
> > __free_page(kvm_tdx->td.tdr_page);
> > kvm_tdx->td.tdr_page = NULL;
> > @@ -1621,7 +1629,8 @@ static int tdx_sept_drop_private_spte(struct kvm *kvm, gfn_t gfn,
> > pr_tdx_error(TDH_PHYMEM_PAGE_WBINVD, err);
> > return -EIO;
> > }
> > - tdx_clear_page(page);
> > +
> > + tdx_clear_page(page, level);
> > tdx_unpin(kvm, page);
> > return 0;
> > }
>