From nobody Tue Feb 10 15:46:08 2026 Received: from dggsgout11.his.huawei.com (dggsgout11.his.huawei.com [45.249.212.51]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 60EB827EFE3; Sat, 20 Dec 2025 04:16:15 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=45.249.212.51 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1766204177; cv=none; b=pjTUlQp4YeUYKe+w7WUJdylk2QIvSToAa/OLDejg/CnVGd5iI0Lipji0bnAZY7je7hveHbUh7NaRLdaxFD5LN1ytH/Qv1DL9V5f6J84XGCwYpSXDMNZC1NYiWWGkCOo88LkZ+BNVxc9ZiATweIMSfN8+YFubSEBH4PwbnSlNtoc= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1766204177; c=relaxed/simple; bh=sWxn4m0VtswJJMDsL52vwPsZZi6e7XbRViKbtvda+po=; h=From:To:Cc:Subject:Date:Message-Id:In-Reply-To:References: MIME-Version; b=O5i/Mc4fROlD0xzG8EEs/7k/svmMHajURC7j/2CH+4P1WsTwpbo4G6PgH2bwgjngtFmixjyV0GgK9XZutjx88d1RMFjWWn6pHg1L8xxz+oarScq2/pzqzEuMNYLv8QV6E/MdL2f6W1w63g++//ImboAInXGgvpBqhrefbqsptDM= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=huaweicloud.com; spf=pass smtp.mailfrom=huaweicloud.com; arc=none smtp.client-ip=45.249.212.51 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=huaweicloud.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=huaweicloud.com Received: from mail.maildlp.com (unknown [172.19.163.177]) by dggsgout11.his.huawei.com (SkyGuard) with ESMTPS id 4dYB0817ZkzYQtfq; Sat, 20 Dec 2025 12:15:36 +0800 (CST) Received: from mail02.huawei.com (unknown [10.116.40.128]) by mail.maildlp.com (Postfix) with ESMTP id 7DF374058F; Sat, 20 Dec 2025 12:16:07 +0800 (CST) Received: from huaweicloud.com (unknown [10.50.87.129]) by APP4 (Coremail) with SMTP id gCh0CgD3WPn5IkZpFwpFAw--.56015S12; Sat, 20 Dec 2025 12:16:07 +0800 (CST) From: Hou Tao To: linux-kernel@vger.kernel.org Cc: linux-pci@vger.kernel.org, linux-mm@kvack.org, linux-nvme@lists.infradead.org, Bjorn Helgaas , Logan Gunthorpe , Alistair Popple , Leon Romanovsky , Greg Kroah-Hartman , Tejun Heo , "Rafael J . Wysocki" , Danilo Krummrich , Andrew Morton , David Hildenbrand , Lorenzo Stoakes , Keith Busch , Jens Axboe , Christoph Hellwig , Sagi Grimberg , houtao1@huawei.com Subject: [PATCH 08/13] mm/huge_memory: add helpers to insert huge page during mmap Date: Sat, 20 Dec 2025 12:04:41 +0800 Message-Id: <20251220040446.274991-9-houtao@huaweicloud.com> X-Mailer: git-send-email 2.29.2 In-Reply-To: <20251220040446.274991-1-houtao@huaweicloud.com> References: <20251220040446.274991-1-houtao@huaweicloud.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-CM-TRANSID: gCh0CgD3WPn5IkZpFwpFAw--.56015S12 X-Coremail-Antispam: 1UD129KBjvJXoWxZryfJFyfCF45JrW8WF47twb_yoW5uF17pF 97GFn8ZrWIqrnrurnxWFs8Ary3X3yxWayUKFW7WF1ava17t34F9a1kJw15tF15JryUCFs3 Xa17GFy5uFyUWa7anT9S1TB71UUUUU7qnTZGkaVYY2UrUUUUjbIjqfuFe4nvWSU5nxnvy2 9KBjDU0xBIdaVrnRJUUUPlb4IE77IF4wAFF20E14v26rWj6s0DM7CY07I20VC2zVCF04k2 6cxKx2IYs7xG6rWj6s0DM7CIcVAFz4kK6r1j6r18M28IrcIa0xkI8VA2jI8067AKxVWUAV Cq3wA2048vs2IY020Ec7CjxVAFwI0_Xr0E3s1l8cAvFVAK0II2c7xJM28CjxkF64kEwVA0 rcxSw2x7M28EF7xvwVC0I7IYx2IY67AKxVW7JVWDJwA2z4x0Y4vE2Ix0cI8IcVCY1x0267 AKxVW8Jr0_Cr1UM28EF7xvwVC2z280aVAFwI0_GcCE3s1l84ACjcxK6I8E87Iv6xkF7I0E 14v26rxl6s0DM2AIxVAIcxkEcVAq07x20xvEncxIr21l5I8CrVACY4xI64kE6c02F40Ex7 xfMcIj6xIIjxv20xvE14v26r126r1DMcIj6I8E87Iv67AKxVW8Jr0_Cr1UMcvjeVCFs4IE 7xkEbVWUJVW8JwACjcxG0xvY0x0EwIxGrwACI402YVCY1x02628vn2kIc2xKxwCY1x0262 kKe7AKxVW8ZVWrXwCF04k20xvY0x0EwIxGrwCFx2IqxVCFs4IE7xkEbVWUJVW8JwC20s02 6c02F40E14v26r1j6r18MI8I3I0E7480Y4vE14v26r106r1rMI8E67AF67kF1VAFwI0_GF v_WrylIxkGc2Ij64vIr41lIxAIcVC0I7IYx2IY67AKxVWUCVW8JwCI42IY6xIIjxv20xvE c7CjxVAFwI0_Gr1j6F4UJwCI42IY6xAIw20EY4v20xvaj40_Jr0_JF4lIxAIcVC2z280aV AFwI0_Jr0_Gr1lIxAIcVC2z280aVCY1x0267AKxVW8Jr0_Cr1UYxBIdaVFxhVjvjDU0xZF pf9x07j4fO7UUUUU= X-CM-SenderInfo: xkrx3t3r6k3tpzhluzxrxghudrp/ Content-Type: text/plain; charset="utf-8" From: Hou Tao vmf_insert_folio_{pmd,pud}() can be used to insert huge page during page fault. However, for simplicity, the mapping of p2pdma memory inserts all necessary pages during mmap. Therefore, add vm_insert_folio_{pmd|pud} helpers to support inserting pmd-sized and pud-sized page during mmap. Signed-off-by: Hou Tao --- include/linux/huge_mm.h | 4 +++ mm/huge_memory.c | 66 +++++++++++++++++++++++++++++++++++++++++ 2 files changed, 70 insertions(+) diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h index a4d9f964dfde..8cf8bb85be79 100644 --- a/include/linux/huge_mm.h +++ b/include/linux/huge_mm.h @@ -45,6 +45,10 @@ vm_fault_t vmf_insert_folio_pmd(struct vm_fault *vmf, st= ruct folio *folio, bool write); vm_fault_t vmf_insert_folio_pud(struct vm_fault *vmf, struct folio *folio, bool write); +int vm_insert_folio_pmd(struct vm_area_struct *vma, unsigned long addr, + struct folio *folio); +int vm_insert_folio_pud(struct vm_area_struct *vma, unsigned long addr, + struct folio *folio); =20 enum transparent_hugepage_flag { TRANSPARENT_HUGEPAGE_UNSUPPORTED, diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 40cf59301c21..11d19f8986da 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -1644,6 +1644,41 @@ vm_fault_t vmf_insert_folio_pmd(struct vm_fault *vmf= , struct folio *folio, } EXPORT_SYMBOL_GPL(vmf_insert_folio_pmd); =20 +int vm_insert_folio_pmd(struct vm_area_struct *vma, unsigned long addr, + struct folio *folio) +{ + struct mm_struct *mm =3D vma->vm_mm; + struct folio_or_pfn fop =3D { + .folio =3D folio, + .is_folio =3D true, + }; + pgd_t *pgd; + p4d_t *p4d; + pud_t *pud; + pmd_t *pmd; + vm_fault_t fault_err; + + mmap_assert_write_locked(mm); + + pgd =3D pgd_offset(mm, addr); + p4d =3D p4d_alloc(mm, pgd, addr); + if (!p4d) + return -ENOMEM; + pud =3D pud_alloc(mm, p4d, addr); + if (!pud) + return -ENOMEM; + pmd =3D pmd_alloc(mm, pud, addr); + if (!pmd) + return -ENOMEM; + + fault_err =3D insert_pmd(vma, addr, pmd, fop, vma->vm_page_prot, + vma->vm_flags & VM_WRITE); + if (fault_err !=3D VM_FAULT_NOPAGE) + return -EINVAL; + + return 0; +} + #ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD static pud_t maybe_pud_mkwrite(pud_t pud, struct vm_area_struct *vma) { @@ -1759,6 +1794,37 @@ vm_fault_t vmf_insert_folio_pud(struct vm_fault *vmf= , struct folio *folio, return insert_pud(vma, addr, vmf->pud, fop, vma->vm_page_prot, write); } EXPORT_SYMBOL_GPL(vmf_insert_folio_pud); + +int vm_insert_folio_pud(struct vm_area_struct *vma, unsigned long addr, + struct folio *folio) +{ + struct mm_struct *mm =3D vma->vm_mm; + struct folio_or_pfn fop =3D { + .folio =3D folio, + .is_folio =3D true, + }; + pgd_t *pgd; + p4d_t *p4d; + pud_t *pud; + vm_fault_t fault_err; + + mmap_assert_write_locked(mm); + + pgd =3D pgd_offset(mm, addr); + p4d =3D p4d_alloc(mm, pgd, addr); + if (!p4d) + return -ENOMEM; + pud =3D pud_alloc(mm, p4d, addr); + if (!pud) + return -ENOMEM; + + fault_err =3D insert_pud(vma, addr, pud, fop, vma->vm_page_prot, + vma->vm_flags & VM_WRITE); + if (fault_err !=3D VM_FAULT_NOPAGE) + return -EINVAL; + + return 0; +} #endif /* CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD */ =20 /** --=20 2.29.2