From nobody Mon Dec 29 20:04:55 2025 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 50770C61DF7 for ; Thu, 23 Nov 2023 13:36:43 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345531AbjKWNge (ORCPT ); Thu, 23 Nov 2023 08:36:34 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47700 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345495AbjKWNg0 (ORCPT ); Thu, 23 Nov 2023 08:36:26 -0500 Received: from out-175.mta0.migadu.com (out-175.mta0.migadu.com [IPv6:2001:41d0:1004:224b::af]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1D59A1B6 for ; Thu, 23 Nov 2023 05:36:31 -0800 (PST) X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1700746266; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=rZ0xmMUhKXlLEkuqOXd+S2Kz2QguNe+dVlDKaJye92A=; b=PvITNhCauLhoxZoDafpy0vQRZZqHvieig+ygz0sJ0KwQBDvhV9RPQ7WPfQpMHtWDAup4Yw MSno402iwprJTutoIIV7sPIPvxuHCpkEExhh6mVM441U9SnVOxB2WXeqSKZJ89qeRTXCh6 7CujJmPaqH11kwtg96cXGzPn5Hr9Caw= From: Gang Li To: Mike Kravetz , Muchun Song , Andrew Morton Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Gang Li Subject: [RFC PATCH v1 1/4] hugetlb: code clean for hugetlb_hstate_alloc_pages Date: Thu, 23 Nov 2023 21:30:33 +0800 Message-Id: <20231123133036.68540-2-gang.li@linux.dev> In-Reply-To: <20231123133036.68540-1-gang.li@linux.dev> References: <20231123133036.68540-1-gang.li@linux.dev> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" From: Gang Li This patch focus on cleaning up the code related to per node allocation and error reporting in the hugetlb alloc: - hugetlb_hstate_alloc_pages_node_specific() to handle iterates through each online node and performs allocation if necessary. - hugetlb_hstate_alloc_pages_report() report error during allocation. And the value of h->max_huge_pages is updated accordingly. This patch has no functional changes. Signed-off-by: Gang Li --- mm/hugetlb.c | 46 +++++++++++++++++++++++++++++----------------- 1 file changed, 29 insertions(+), 17 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index c466551e2fd9..7af2ee08ad1b 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -3482,6 +3482,33 @@ static void __init hugetlb_hstate_alloc_pages_onenod= e(struct hstate *h, int nid) h->max_huge_pages_node[nid] =3D i; } =20 +static bool __init hugetlb_hstate_alloc_pages_node_specific(struct hstate = *h) +{ + int i; + bool node_specific_alloc =3D false; + + for_each_online_node(i) { + if (h->max_huge_pages_node[i] > 0) { + hugetlb_hstate_alloc_pages_onenode(h, i); + node_specific_alloc =3D true; + } + } + + return node_specific_alloc; +} + +static void __init hugetlb_hstate_alloc_pages_report(unsigned long allocat= ed, struct hstate *h) +{ + if (allocated < h->max_huge_pages) { + char buf[32]; + + string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32); + pr_warn("HugeTLB: allocating %lu of page size %s failed. Only allocated= %lu hugepages.\n", + h->max_huge_pages, buf, allocated); + h->max_huge_pages =3D allocated; + } +} + /* * NOTE: this routine is called in different contexts for gigantic and * non-gigantic pages. @@ -3499,7 +3526,6 @@ static void __init hugetlb_hstate_alloc_pages(struct = hstate *h) struct folio *folio; LIST_HEAD(folio_list); nodemask_t *node_alloc_noretry; - bool node_specific_alloc =3D false; =20 /* skip gigantic hugepages allocation if hugetlb_cma enabled */ if (hstate_is_gigantic(h) && hugetlb_cma_size) { @@ -3508,14 +3534,7 @@ static void __init hugetlb_hstate_alloc_pages(struct= hstate *h) } =20 /* do node specific alloc */ - for_each_online_node(i) { - if (h->max_huge_pages_node[i] > 0) { - hugetlb_hstate_alloc_pages_onenode(h, i); - node_specific_alloc =3D true; - } - } - - if (node_specific_alloc) + if (hugetlb_hstate_alloc_pages_node_specific(h)) return; =20 /* below will do all node balanced alloc */ @@ -3558,14 +3577,7 @@ static void __init hugetlb_hstate_alloc_pages(struct= hstate *h) /* list will be empty if hstate_is_gigantic */ prep_and_add_allocated_folios(h, &folio_list); =20 - if (i < h->max_huge_pages) { - char buf[32]; - - string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32); - pr_warn("HugeTLB: allocating %lu of page size %s failed. Only allocated= %lu hugepages.\n", - h->max_huge_pages, buf, i); - h->max_huge_pages =3D i; - } + hugetlb_hstate_alloc_pages_report(i, h); kfree(node_alloc_noretry); } =20 --=20 2.20.1 From nobody Mon Dec 29 20:04:55 2025 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 26FFFC5AD4C for ; Thu, 23 Nov 2023 13:36:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345525AbjKWNga (ORCPT ); Thu, 23 Nov 2023 08:36:30 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47676 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345493AbjKWNg0 (ORCPT ); Thu, 23 Nov 2023 08:36:26 -0500 Received: from out-179.mta0.migadu.com (out-179.mta0.migadu.com [IPv6:2001:41d0:1004:224b::b3]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id DAA12BA for ; Thu, 23 Nov 2023 05:36:31 -0800 (PST) X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1700746269; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=nGnKScZww5hZXWYTDAjHhsju7kmgmcYLYNYxHMYgc28=; b=d5CHLX57cCOGxJdVkM1Rcs5gK8JkZKtdzrPAWDsIIXnn4bLS2WT8PJZozgQuz0CNFwb0nc 0T867ikyskX4+1d3xr9U6o5B04QY6IGfw21Vn6rpkQHu8CmkuXfiaUXsViDqqNzM3Lwggu BarejmtOHYkPWa7BTb5kD5s+K+Zxa3w= From: Gang Li To: Mike Kravetz , Muchun Song , Andrew Morton Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Gang Li Subject: [RFC PATCH v1 2/4] hugetlb: split hugetlb_hstate_alloc_pages Date: Thu, 23 Nov 2023 21:30:34 +0800 Message-Id: <20231123133036.68540-3-gang.li@linux.dev> In-Reply-To: <20231123133036.68540-1-gang.li@linux.dev> References: <20231123133036.68540-1-gang.li@linux.dev> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" From: Gang Li Split hugetlb_hstate_alloc_pages into gigantic and non-gigantic. These patch has no functional changes. Signed-off-by: Gang Li --- mm/hugetlb.c | 86 +++++++++++++++++++++++++++------------------------- 1 file changed, 45 insertions(+), 41 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 7af2ee08ad1b..7f9ff0855dd0 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -3509,6 +3509,47 @@ static void __init hugetlb_hstate_alloc_pages_report= (unsigned long allocated, st } } =20 +static unsigned long __init hugetlb_hstate_alloc_pages_gigantic(struct hst= ate *h) +{ + unsigned long i; + + for (i =3D 0; i < h->max_huge_pages; ++i) { + /* + * gigantic pages not added to list as they are not + * added to pools now. + */ + if (!alloc_bootmem_huge_page(h, NUMA_NO_NODE)) + break; + cond_resched(); + } + + return i; +} + +static unsigned long __init hugetlb_hstate_alloc_pages_non_gigantic(struct= hstate *h) +{ + unsigned long i; + struct folio *folio; + LIST_HEAD(folio_list); + nodemask_t node_alloc_noretry; + + /* Bit mask controlling how hard we retry per-node allocations.*/ + nodes_clear(node_alloc_noretry); + + for (i =3D 0; i < h->max_huge_pages; ++i) { + folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], + &node_alloc_noretry); + if (!folio) + break; + list_add(&folio->lru, &folio_list); + cond_resched(); + } + + prep_and_add_allocated_folios(h, &folio_list); + + return i; +} + /* * NOTE: this routine is called in different contexts for gigantic and * non-gigantic pages. @@ -3522,10 +3563,7 @@ static void __init hugetlb_hstate_alloc_pages_report= (unsigned long allocated, st */ static void __init hugetlb_hstate_alloc_pages(struct hstate *h) { - unsigned long i; - struct folio *folio; - LIST_HEAD(folio_list); - nodemask_t *node_alloc_noretry; + unsigned long allocated; =20 /* skip gigantic hugepages allocation if hugetlb_cma enabled */ if (hstate_is_gigantic(h) && hugetlb_cma_size) { @@ -3539,46 +3577,12 @@ static void __init hugetlb_hstate_alloc_pages(struc= t hstate *h) =20 /* below will do all node balanced alloc */ if (!hstate_is_gigantic(h)) { - /* - * Bit mask controlling how hard we retry per-node allocations. - * Ignore errors as lower level routines can deal with - * node_alloc_noretry =3D=3D NULL. If this kmalloc fails at boot - * time, we are likely in bigger trouble. - */ - node_alloc_noretry =3D kmalloc(sizeof(*node_alloc_noretry), - GFP_KERNEL); + allocated =3D hugetlb_hstate_alloc_pages_non_gigantic(h); } else { - /* allocations done at boot time */ - node_alloc_noretry =3D NULL; - } - - /* bit mask controlling how hard we retry per-node allocations */ - if (node_alloc_noretry) - nodes_clear(*node_alloc_noretry); - - for (i =3D 0; i < h->max_huge_pages; ++i) { - if (hstate_is_gigantic(h)) { - /* - * gigantic pages not added to list as they are not - * added to pools now. - */ - if (!alloc_bootmem_huge_page(h, NUMA_NO_NODE)) - break; - } else { - folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], - node_alloc_noretry); - if (!folio) - break; - list_add(&folio->lru, &folio_list); - } - cond_resched(); + allocated =3D hugetlb_hstate_alloc_pages_gigantic(h); } =20 - /* list will be empty if hstate_is_gigantic */ - prep_and_add_allocated_folios(h, &folio_list); - - hugetlb_hstate_alloc_pages_report(i, h); - kfree(node_alloc_noretry); + hugetlb_hstate_alloc_pages_report(allocated, h); } =20 static void __init hugetlb_init_hstates(void) --=20 2.20.1 From nobody Mon Dec 29 20:04:55 2025 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id C5E22C61DF7 for ; Thu, 23 Nov 2023 13:36:48 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345560AbjKWNgj (ORCPT ); Thu, 23 Nov 2023 08:36:39 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47684 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345504AbjKWNg0 (ORCPT ); Thu, 23 Nov 2023 08:36:26 -0500 Received: from out-175.mta0.migadu.com (out-175.mta0.migadu.com [IPv6:2001:41d0:1004:224b::af]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id DAABD1B3 for ; Thu, 23 Nov 2023 05:36:31 -0800 (PST) X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1700746272; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=ZmVKGX3xWVsiaI1L9mmzJOcNAtj7qPzayoGRggtSBII=; b=Ztzwn0HQtIgdF3nrfOjRTmuus6uH05VuvO5saXWr5t12Q51VOnt0v0Tdqe3fTgV29LfEAj ZlpS/GGfT1xwjVdXl9k5LFCEe038RDKIRwdprffhWUuffvY73PgEemYmViGVybE5iBDRMn 6grrj48RxSO/qOyc0/yPWH8AYNiUtwQ= From: Gang Li To: Mike Kravetz , Muchun Song , Andrew Morton Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Gang Li Subject: [RFC PATCH v1 3/4] hugetlb: add timing to hugetlb allocations on boot Date: Thu, 23 Nov 2023 21:30:35 +0800 Message-Id: <20231123133036.68540-4-gang.li@linux.dev> In-Reply-To: <20231123133036.68540-1-gang.li@linux.dev> References: <20231123133036.68540-1-gang.li@linux.dev> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" From: Gang Li Add timing to hugetlb allocations for further optimization. Signed-off-by: Gang Li --- mm/hugetlb.c | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 7f9ff0855dd0..ac8558724cc2 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -3563,7 +3563,7 @@ static unsigned long __init hugetlb_hstate_alloc_page= s_non_gigantic(struct hstat */ static void __init hugetlb_hstate_alloc_pages(struct hstate *h) { - unsigned long allocated; + unsigned long allocated, start; =20 /* skip gigantic hugepages allocation if hugetlb_cma enabled */ if (hstate_is_gigantic(h) && hugetlb_cma_size) { @@ -3576,11 +3576,13 @@ static void __init hugetlb_hstate_alloc_pages(struc= t hstate *h) return; =20 /* below will do all node balanced alloc */ + start =3D jiffies; if (!hstate_is_gigantic(h)) { allocated =3D hugetlb_hstate_alloc_pages_non_gigantic(h); } else { allocated =3D hugetlb_hstate_alloc_pages_gigantic(h); } + pr_info("HugeTLB: Allocation takes %u ms\n", jiffies_to_msecs(jiffies - s= tart)); =20 hugetlb_hstate_alloc_pages_report(allocated, h); } --=20 2.20.1 From nobody Mon Dec 29 20:04:55 2025 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 31803C61DF7 for ; Thu, 23 Nov 2023 13:36:36 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345514AbjKWNg2 (ORCPT ); Thu, 23 Nov 2023 08:36:28 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47672 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345468AbjKWNg0 (ORCPT ); Thu, 23 Nov 2023 08:36:26 -0500 X-Greylist: delayed 326 seconds by postgrey-1.37 at lindbergh.monkeyblade.net; Thu, 23 Nov 2023 05:36:31 PST Received: from out-186.mta0.migadu.com (out-186.mta0.migadu.com [IPv6:2001:41d0:1004:224b::ba]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A04B81A8 for ; Thu, 23 Nov 2023 05:36:31 -0800 (PST) X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1700746275; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=SVmsj9TN/aj38n19fqlW2b7zQkW3VvgWG56LLSCGces=; b=Y71JssEDzrYWYNV34cyzjaSirgHK0BrpKRWoKYtD+NlDt+B0BHOaS3CWA9Uyhz5yKbgj6g PimBt6l7h2P4kLFfkV0n8MPPjcmR9a/NNorQu2gY5FzktxNGVTBcPtxXi8OV6XSHGDx82i xNn/ddoatbelzCFAmHSUzrB3UJs23Jc= From: Gang Li To: Mike Kravetz , Muchun Song , Andrew Morton Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Gang Li Subject: [RFC PATCH v1 4/4] hugetlb: parallelize hugetlb page allocation Date: Thu, 23 Nov 2023 21:30:36 +0800 Message-Id: <20231123133036.68540-5-gang.li@linux.dev> In-Reply-To: <20231123133036.68540-1-gang.li@linux.dev> References: <20231123133036.68540-1-gang.li@linux.dev> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" From: Gang Li By distributing the allocation across threads, large hugetlb configurations can allocate pages faster, improving boot speed. Signed-off-by: Gang Li --- mm/hugetlb.c | 89 +++++++++++++++++++++++++++++++++++++++++++--------- 1 file changed, 74 insertions(+), 15 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index ac8558724cc2..df3fbe95989e 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -3509,6 +3509,55 @@ static void __init hugetlb_hstate_alloc_pages_report= (unsigned long allocated, st } } =20 +struct hugetlb_work { + struct work_struct work; + struct hstate *h; + int num; + int nid; +}; + +static atomic_t hugetlb_hstate_alloc_n_undone __initdata; +static __initdata DECLARE_COMPLETION(hugetlb_hstate_alloc_comp); + +static void __init hugetlb_alloc_node(struct work_struct *w) +{ + struct hugetlb_work *hw =3D container_of(w, struct hugetlb_work, work); + struct hstate *h =3D hw->h; + int i, num =3D hw->num; + nodemask_t node_alloc_noretry; + unsigned long flags; + + /* Bit mask controlling how hard we retry per-node allocations.*/ + nodes_clear(node_alloc_noretry); + + for (i =3D 0; i < num; ++i) { + struct folio *folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], + &node_alloc_noretry); + if (!folio) + break; + spin_lock_irqsave(&hugetlb_lock, flags); + __prep_account_new_huge_page(h, folio_nid(folio)); + enqueue_hugetlb_folio(h, folio); + spin_unlock_irqrestore(&hugetlb_lock, flags); + cond_resched(); + } + + if (atomic_dec_and_test(&hugetlb_hstate_alloc_n_undone)) + complete(&hugetlb_hstate_alloc_comp); +} + +static void __init hugetlb_vmemmap_optimize_node(struct work_struct *w) +{ + struct hugetlb_work *hw =3D container_of(w, struct hugetlb_work, work); + struct hstate *h =3D hw->h; + int nid =3D hw->nid; + + hugetlb_vmemmap_optimize_folios(h, &h->hugepage_freelists[nid]); + + if (atomic_dec_and_test(&hugetlb_hstate_alloc_n_undone)) + complete(&hugetlb_hstate_alloc_comp); +} + static unsigned long __init hugetlb_hstate_alloc_pages_gigantic(struct hst= ate *h) { unsigned long i; @@ -3528,26 +3577,36 @@ static unsigned long __init hugetlb_hstate_alloc_pa= ges_gigantic(struct hstate *h =20 static unsigned long __init hugetlb_hstate_alloc_pages_non_gigantic(struct= hstate *h) { - unsigned long i; - struct folio *folio; - LIST_HEAD(folio_list); - nodemask_t node_alloc_noretry; + int nid; + struct hugetlb_work *works; =20 - /* Bit mask controlling how hard we retry per-node allocations.*/ - nodes_clear(node_alloc_noretry); + works =3D kcalloc(num_node_state(N_MEMORY), sizeof(*works), GFP_KERNEL); + if (works =3D=3D NULL) { + pr_warn("HugeTLB: allocating struct hugetlb_work failed.\n"); + return 0; + } =20 - for (i =3D 0; i < h->max_huge_pages; ++i) { - folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], - &node_alloc_noretry); - if (!folio) - break; - list_add(&folio->lru, &folio_list); - cond_resched(); + atomic_set(&hugetlb_hstate_alloc_n_undone, num_node_state(N_MEMORY)); + for_each_node_state(nid, N_MEMORY) { + works[nid].h =3D h; + works[nid].num =3D h->max_huge_pages/num_node_state(N_MEMORY); + if (nid =3D=3D 0) + works[nid].num +=3D h->max_huge_pages % num_node_state(N_MEMORY); + INIT_WORK(&works[nid].work, hugetlb_alloc_node); + queue_work_node(nid, system_unbound_wq, &works[nid].work); } + wait_for_completion(&hugetlb_hstate_alloc_comp); =20 - prep_and_add_allocated_folios(h, &folio_list); + atomic_set(&hugetlb_hstate_alloc_n_undone, num_node_state(N_MEMORY)); + for_each_node_state(nid, N_MEMORY) { + works[nid].nid =3D nid; + INIT_WORK(&works[nid].work, hugetlb_vmemmap_optimize_node); + queue_work_node(nid, system_unbound_wq, &works[nid].work); + } + wait_for_completion(&hugetlb_hstate_alloc_comp); =20 - return i; + kfree(works); + return h->nr_huge_pages; } =20 /* --=20 2.20.1