From nobody Tue Feb 10 04:14:02 2026 Received: from mgamail.intel.com (mgamail.intel.com [192.198.163.8]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 3E635345725; Wed, 14 Jan 2026 19:21:49 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=192.198.163.8 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1768418511; cv=none; b=d/gTAvuP+lbA5OznPvJd4koXdPB1LEqsKZkT59hJ3/hDuLYTnL2McMVObXLre1gJPXPHSe6FuhHGa7zj8+MOW6H+v3X4f/DJZRy8+Mgw66Edyz0moqvawR5ks5guX4LbU2tWdevhDNi/fDabN2JFT98rM2I4tMOrNAaibsFatuc= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1768418511; c=relaxed/simple; bh=jiGWK05xuN192DYv+TPcEdDntEMizLdbArhfYH/lgMA=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version:Content-Type; b=FvKVTxl20ghz8VpQMFYp2//i3XGGFwuy12+pU0aKrDtoCtLlJM80tOEXei8DIacWQcN5KKVgsHSvbYSqYqZ76FMw6WzPB8v3LVRmsxxDEZ1YLgdv1Wxin39SolHlSJZiIOMwsSSCVmnAaPrRfOnk44trKsetgtGSXx1I6+U3UIo= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=intel.com; spf=pass smtp.mailfrom=intel.com; dkim=pass (2048-bit key) header.d=intel.com header.i=@intel.com header.b=YBrX/0SP; arc=none smtp.client-ip=192.198.163.8 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=intel.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=intel.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=intel.com header.i=@intel.com header.b="YBrX/0SP" DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1768418509; x=1799954509; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=jiGWK05xuN192DYv+TPcEdDntEMizLdbArhfYH/lgMA=; b=YBrX/0SP+QXYj4nR7EfYqJFq6vd3UkkW/UrZL26jZgqO+/0P00ASZxyj 70GBX+OQugynBvIb4OYzfcjBbjCLT96/YlxEKd37JnGuc5Tdj6GH1A/vp 8HTV1WhAFwBwkzioHpak3aeJtqafdQiKxNdV18SUe/nNiSMvhFKJ0feKB wkWVGsx6FRq5CwZWrsjVQnkLgG6gt6w4RDvSNSz7z+9INluwaL8Li3Zk2 KcD4z8fdnHWeY9K90L9TlbYkaWyiunUQed5y+nesKIx+xkiBIcTMLpnPF 6wjZt7AtZB+VRa+SgnG17AUGMw/bHYyejvrLM7145Hv5mnCwsDT83fojW A==; X-CSE-ConnectionGUID: XDQbom+IRiSgkS6NtyEaKg== X-CSE-MsgGUID: NhiLZwNPQcqa7J/2a4zAfA== X-IronPort-AV: E=McAfee;i="6800,10657,11671"; a="87305725" X-IronPort-AV: E=Sophos;i="6.21,226,1763452800"; d="scan'208";a="87305725" Received: from fmviesa001.fm.intel.com ([10.60.135.141]) by fmvoesa102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Jan 2026 11:21:48 -0800 X-CSE-ConnectionGUID: UrOFw46sSR+Opbvpw2+jpw== X-CSE-MsgGUID: /21BpcAaQ6GbBGCCIQN2Ow== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.21,226,1763452800"; d="scan'208";a="236006831" Received: from smoticic-mobl1.ger.corp.intel.com (HELO fdugast-desk.intel.com) ([10.245.244.85]) by smtpauth.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Jan 2026 11:21:42 -0800 From: Francois Dugast To: intel-xe@lists.freedesktop.org Cc: dri-devel@lists.freedesktop.org, Matthew Brost , Zi Yan , Alistair Popple , adhavan Srinivasan , Nicholas Piggin , Michael Ellerman , "Christophe Leroy (CS GROUP)" , Felix Kuehling , Alex Deucher , =?UTF-8?q?Christian=20K=C3=B6nig?= , David Airlie , Simona Vetter , Maarten Lankhorst , Maxime Ripard , Thomas Zimmermann , Lyude Paul , Danilo Krummrich , David Hildenbrand , Oscar Salvador , Andrew Morton , Jason Gunthorpe , Leon Romanovsky , Lorenzo Stoakes , "Liam R . Howlett" , Vlastimil Babka , Mike Rapoport , Suren Baghdasaryan , Michal Hocko , Balbir Singh , linuxppc-dev@lists.ozlabs.org, kvm@vger.kernel.org, linux-kernel@vger.kernel.org, amd-gfx@lists.freedesktop.org, nouveau@lists.freedesktop.org, linux-mm@kvack.org, linux-cxl@vger.kernel.org, Francois Dugast Subject: [PATCH v5 1/5] mm/zone_device: Reinitialize large zone device private folios Date: Wed, 14 Jan 2026 20:19:52 +0100 Message-ID: <20260114192111.1267147-2-francois.dugast@intel.com> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20260114192111.1267147-1-francois.dugast@intel.com> References: <20260114192111.1267147-1-francois.dugast@intel.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable From: Matthew Brost Reinitialize metadata for large zone device private folios in zone_device_page_init prior to creating a higher-order zone device private folio. This step is necessary when the folio=E2=80=99s order changes dynamically between zone_device_page_init calls to avoid building a corrupt folio. As part of the metadata reinitialization, the dev_pagemap must be passed in from the caller because the pgmap stored in the folio page may have been overwritten with a compound head. Cc: Zi Yan Cc: Alistair Popple Cc: adhavan Srinivasan Cc: Nicholas Piggin Cc: Michael Ellerman Cc: "Christophe Leroy (CS GROUP)" Cc: Felix Kuehling Cc: Alex Deucher Cc: "Christian K=C3=B6nig" Cc: David Airlie Cc: Simona Vetter Cc: Maarten Lankhorst Cc: Maxime Ripard Cc: Thomas Zimmermann Cc: Lyude Paul Cc: Danilo Krummrich Cc: David Hildenbrand Cc: Oscar Salvador Cc: Andrew Morton Cc: Jason Gunthorpe Cc: Leon Romanovsky Cc: Lorenzo Stoakes Cc: Liam R. Howlett Cc: Vlastimil Babka Cc: Mike Rapoport Cc: Suren Baghdasaryan Cc: Michal Hocko Cc: Balbir Singh Cc: linuxppc-dev@lists.ozlabs.org Cc: kvm@vger.kernel.org Cc: linux-kernel@vger.kernel.org Cc: amd-gfx@lists.freedesktop.org Cc: dri-devel@lists.freedesktop.org Cc: nouveau@lists.freedesktop.org Cc: linux-mm@kvack.org Cc: linux-cxl@vger.kernel.org Fixes: d245f9b4ab80 ("mm/zone_device: support large zone device private fol= ios") Signed-off-by: Matthew Brost Signed-off-by: Francois Dugast Acked-by: Felix Kuehling --- arch/powerpc/kvm/book3s_hv_uvmem.c | 2 +- drivers/gpu/drm/amd/amdkfd/kfd_migrate.c | 2 +- drivers/gpu/drm/drm_pagemap.c | 2 +- drivers/gpu/drm/nouveau/nouveau_dmem.c | 2 +- include/linux/memremap.h | 9 ++++++--- lib/test_hmm.c | 4 +++- mm/memremap.c | 20 +++++++++++++++++++- 7 files changed, 32 insertions(+), 9 deletions(-) diff --git a/arch/powerpc/kvm/book3s_hv_uvmem.c b/arch/powerpc/kvm/book3s_h= v_uvmem.c index e5000bef90f2..7cf9310de0ec 100644 --- a/arch/powerpc/kvm/book3s_hv_uvmem.c +++ b/arch/powerpc/kvm/book3s_hv_uvmem.c @@ -723,7 +723,7 @@ static struct page *kvmppc_uvmem_get_page(unsigned long= gpa, struct kvm *kvm) =20 dpage =3D pfn_to_page(uvmem_pfn); dpage->zone_device_data =3D pvt; - zone_device_page_init(dpage, 0); + zone_device_page_init(dpage, &kvmppc_uvmem_pgmap, 0); return dpage; out_clear: spin_lock(&kvmppc_uvmem_bitmap_lock); diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c b/drivers/gpu/drm/amd= /amdkfd/kfd_migrate.c index af53e796ea1b..6ada7b4af7c6 100644 --- a/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c +++ b/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c @@ -217,7 +217,7 @@ svm_migrate_get_vram_page(struct svm_range *prange, uns= igned long pfn) page =3D pfn_to_page(pfn); svm_range_bo_ref(prange->svm_bo); page->zone_device_data =3D prange->svm_bo; - zone_device_page_init(page, 0); + zone_device_page_init(page, page_pgmap(page), 0); } =20 static void diff --git a/drivers/gpu/drm/drm_pagemap.c b/drivers/gpu/drm/drm_pagemap.c index 03ee39a761a4..c497726b0147 100644 --- a/drivers/gpu/drm/drm_pagemap.c +++ b/drivers/gpu/drm/drm_pagemap.c @@ -201,7 +201,7 @@ static void drm_pagemap_get_devmem_page(struct page *pa= ge, struct drm_pagemap_zdd *zdd) { page->zone_device_data =3D drm_pagemap_zdd_get(zdd); - zone_device_page_init(page, 0); + zone_device_page_init(page, zdd->dpagemap->pagemap, 0); } =20 /** diff --git a/drivers/gpu/drm/nouveau/nouveau_dmem.c b/drivers/gpu/drm/nouve= au/nouveau_dmem.c index 58071652679d..3d8031296eed 100644 --- a/drivers/gpu/drm/nouveau/nouveau_dmem.c +++ b/drivers/gpu/drm/nouveau/nouveau_dmem.c @@ -425,7 +425,7 @@ nouveau_dmem_page_alloc_locked(struct nouveau_drm *drm,= bool is_large) order =3D ilog2(DMEM_CHUNK_NPAGES); } =20 - zone_device_folio_init(folio, order); + zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order); return page; } =20 diff --git a/include/linux/memremap.h b/include/linux/memremap.h index 713ec0435b48..e3c2ccf872a8 100644 --- a/include/linux/memremap.h +++ b/include/linux/memremap.h @@ -224,7 +224,8 @@ static inline bool is_fsdax_page(const struct page *pag= e) } =20 #ifdef CONFIG_ZONE_DEVICE -void zone_device_page_init(struct page *page, unsigned int order); +void zone_device_page_init(struct page *page, struct dev_pagemap *pgmap, + unsigned int order); void *memremap_pages(struct dev_pagemap *pgmap, int nid); void memunmap_pages(struct dev_pagemap *pgmap); void *devm_memremap_pages(struct device *dev, struct dev_pagemap *pgmap); @@ -234,9 +235,11 @@ bool pgmap_pfn_valid(struct dev_pagemap *pgmap, unsign= ed long pfn); =20 unsigned long memremap_compat_align(void); =20 -static inline void zone_device_folio_init(struct folio *folio, unsigned in= t order) +static inline void zone_device_folio_init(struct folio *folio, + struct dev_pagemap *pgmap, + unsigned int order) { - zone_device_page_init(&folio->page, order); + zone_device_page_init(&folio->page, pgmap, order); if (order) folio_set_large_rmappable(folio); } diff --git a/lib/test_hmm.c b/lib/test_hmm.c index 8af169d3873a..455a6862ae50 100644 --- a/lib/test_hmm.c +++ b/lib/test_hmm.c @@ -662,7 +662,9 @@ static struct page *dmirror_devmem_alloc_page(struct dm= irror *dmirror, goto error; } =20 - zone_device_folio_init(page_folio(dpage), order); + zone_device_folio_init(page_folio(dpage), + page_pgmap(folio_page(page_folio(dpage), 0)), + order); dpage->zone_device_data =3D rpage; return dpage; =20 diff --git a/mm/memremap.c b/mm/memremap.c index 63c6ab4fdf08..6f46ab14662b 100644 --- a/mm/memremap.c +++ b/mm/memremap.c @@ -477,10 +477,28 @@ void free_zone_device_folio(struct folio *folio) } } =20 -void zone_device_page_init(struct page *page, unsigned int order) +void zone_device_page_init(struct page *page, struct dev_pagemap *pgmap, + unsigned int order) { + struct page *new_page =3D page; + unsigned int i; + VM_WARN_ON_ONCE(order > MAX_ORDER_NR_PAGES); =20 + for (i =3D 0; i < (1UL << order); ++i, ++new_page) { + struct folio *new_folio =3D (struct folio *)new_page; + + new_page->flags.f &=3D ~0xffUL; /* Clear possible order, page head */ +#ifdef NR_PAGES_IN_LARGE_FOLIO + ((struct folio *)(new_page - 1))->_nr_pages =3D 0; +#endif + new_folio->mapping =3D NULL; + new_folio->pgmap =3D pgmap; /* Also clear compound head */ + new_folio->share =3D 0; /* fsdax only, unused for device private */ + VM_WARN_ON_FOLIO(folio_ref_count(new_folio), new_folio); + VM_WARN_ON_FOLIO(!folio_is_zone_device(new_folio), new_folio); + } + /* * Drivers shouldn't be allocating pages after calling * memunmap_pages(). --=20 2.43.0