From nobody Thu Dec 18 07:10:28 2025 Received: from out-170.mta0.migadu.com (out-170.mta0.migadu.com [91.218.175.170]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 661801CAB9 for ; Wed, 1 May 2024 17:26:34 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=91.218.175.170 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714584396; cv=none; b=AlJvB9SwTeKoH+DRsWeuKnuURJxTDI/8zDp+4UNfQtSW1zN4PKaRZA+2Srqwcck6iKj8OEqP4DRpr5C1PQ1TRshttfGxsxGCrdTFBQEYpqoVnJ8u1LratscT2lIN1srZ4YKCyiwhLHkMqt/q3ZxT8ckfMbkfeWUfyNfLVUl7DP8= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714584396; c=relaxed/simple; bh=S0/mXTQiLOtYNza4Zt4bMsm/gr/AOggi/7+c+mhm0+A=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=naSuUGQxcZs/Kqk7CXjNarS2Zx4ADMwh97igdtvBMDqyLU4vh7X05Y50Bsf+pFiJCRy6Lu+BRpKWtqViXituu1AX0ubw3axVzXDHScezN7OXratdJRCCRPdPH3bBmI4sIuCg9Oobbott32+T/v5/7lo2c6jbCR922it6DU3RhCg= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=oO8ERjmx; arc=none smtp.client-ip=91.218.175.170 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="oO8ERjmx" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714584392; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=fBg/xQS/MIq7cEhcQ96zq9LHH7YtwZWaDugtQaMbEjA=; b=oO8ERjmxEmOnQNxm4RLisy/xS7iu1hwJijzbGr5XjDAhgynlzk14K/ZUCNh7l6xdxVVh6i VmXZzO2NisPVHuroxULEK/AH/2QxNLCkvwgMItOxL8K00gWQGYiU1iY2ubooM4yA3Vat5A gsvk25slnPnudsqUwOVnqs+0FRq83OQ= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song , Yosry Ahmed , "T . J . Mercier" Cc: kernel-team@meta.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v4 2/8] memcg: dynamically allocate lruvec_stats Date: Wed, 1 May 2024 10:26:11 -0700 Message-ID: <20240501172617.678560-3-shakeel.butt@linux.dev> In-Reply-To: <20240501172617.678560-1-shakeel.butt@linux.dev> References: <20240501172617.678560-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" To decouple the dependency of lruvec_stats on NR_VM_NODE_STAT_ITEMS, we need to dynamically allocate lruvec_stats in the mem_cgroup_per_node structure. Also move the definition of lruvec_stats_percpu and lruvec_stats and related functions to the memcontrol.c to facilitate later patches. No functional changes in the patch. Signed-off-by: Shakeel Butt Reviewed-by: Yosry Ahmed Reviewed-by: T.J. Mercier Reviewed-by: Roman Gushchin --- Changes since v3: - N/A include/linux/memcontrol.h | 62 +++------------------------ mm/memcontrol.c | 87 ++++++++++++++++++++++++++++++++------ 2 files changed, 81 insertions(+), 68 deletions(-) diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index 9aba0d0462ca..ab8a6e884375 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -83,6 +83,8 @@ enum mem_cgroup_events_target { =20 struct memcg_vmstats_percpu; struct memcg_vmstats; +struct lruvec_stats_percpu; +struct lruvec_stats; =20 struct mem_cgroup_reclaim_iter { struct mem_cgroup *position; @@ -90,25 +92,6 @@ struct mem_cgroup_reclaim_iter { unsigned int generation; }; =20 -struct lruvec_stats_percpu { - /* Local (CPU and cgroup) state */ - long state[NR_VM_NODE_STAT_ITEMS]; - - /* Delta calculation for lockless upward propagation */ - long state_prev[NR_VM_NODE_STAT_ITEMS]; -}; - -struct lruvec_stats { - /* Aggregated (CPU and subtree) state */ - long state[NR_VM_NODE_STAT_ITEMS]; - - /* Non-hierarchical (CPU aggregated) state */ - long state_local[NR_VM_NODE_STAT_ITEMS]; - - /* Pending child counts during tree propagation */ - long state_pending[NR_VM_NODE_STAT_ITEMS]; -}; - /* * per-node information in memory controller. */ @@ -116,7 +99,7 @@ struct mem_cgroup_per_node { struct lruvec lruvec; =20 struct lruvec_stats_percpu __percpu *lruvec_stats_percpu; - struct lruvec_stats lruvec_stats; + struct lruvec_stats *lruvec_stats; =20 unsigned long lru_zone_size[MAX_NR_ZONES][NR_LRU_LISTS]; =20 @@ -1037,42 +1020,9 @@ static inline void mod_memcg_page_state(struct page = *page, } =20 unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx); - -static inline unsigned long lruvec_page_state(struct lruvec *lruvec, - enum node_stat_item idx) -{ - struct mem_cgroup_per_node *pn; - long x; - - if (mem_cgroup_disabled()) - return node_page_state(lruvec_pgdat(lruvec), idx); - - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats.state[idx]); -#ifdef CONFIG_SMP - if (x < 0) - x =3D 0; -#endif - return x; -} - -static inline unsigned long lruvec_page_state_local(struct lruvec *lruvec, - enum node_stat_item idx) -{ - struct mem_cgroup_per_node *pn; - long x =3D 0; - - if (mem_cgroup_disabled()) - return node_page_state(lruvec_pgdat(lruvec), idx); - - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats.state_local[idx]); -#ifdef CONFIG_SMP - if (x < 0) - x =3D 0; -#endif - return x; -} +unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item= idx); +unsigned long lruvec_page_state_local(struct lruvec *lruvec, + enum node_stat_item idx); =20 void mem_cgroup_flush_stats(struct mem_cgroup *memcg); void mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg); diff --git a/mm/memcontrol.c b/mm/memcontrol.c index c146187cda9c..7126459ec56a 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -576,6 +576,60 @@ mem_cgroup_largest_soft_limit_node(struct mem_cgroup_t= ree_per_node *mctz) return mz; } =20 +struct lruvec_stats_percpu { + /* Local (CPU and cgroup) state */ + long state[NR_VM_NODE_STAT_ITEMS]; + + /* Delta calculation for lockless upward propagation */ + long state_prev[NR_VM_NODE_STAT_ITEMS]; +}; + +struct lruvec_stats { + /* Aggregated (CPU and subtree) state */ + long state[NR_VM_NODE_STAT_ITEMS]; + + /* Non-hierarchical (CPU aggregated) state */ + long state_local[NR_VM_NODE_STAT_ITEMS]; + + /* Pending child counts during tree propagation */ + long state_pending[NR_VM_NODE_STAT_ITEMS]; +}; + +unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item= idx) +{ + struct mem_cgroup_per_node *pn; + long x; + + if (mem_cgroup_disabled()) + return node_page_state(lruvec_pgdat(lruvec), idx); + + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state[idx]); +#ifdef CONFIG_SMP + if (x < 0) + x =3D 0; +#endif + return x; +} + +unsigned long lruvec_page_state_local(struct lruvec *lruvec, + enum node_stat_item idx) +{ + struct mem_cgroup_per_node *pn; + long x =3D 0; + + if (mem_cgroup_disabled()) + return node_page_state(lruvec_pgdat(lruvec), idx); + + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state_local[idx]); +#ifdef CONFIG_SMP + if (x < 0) + x =3D 0; +#endif + return x; +} + /* Subset of vm_event_item to report for memcg event stats */ static const unsigned int memcg_vm_event_stat[] =3D { PGPGIN, @@ -5491,18 +5545,25 @@ static int alloc_mem_cgroup_per_node_info(struct me= m_cgroup *memcg, int node) if (!pn) return 1; =20 + pn->lruvec_stats =3D kzalloc_node(sizeof(struct lruvec_stats), GFP_KERNEL, + node); + if (!pn->lruvec_stats) + goto fail; + pn->lruvec_stats_percpu =3D alloc_percpu_gfp(struct lruvec_stats_percpu, GFP_KERNEL_ACCOUNT); - if (!pn->lruvec_stats_percpu) { - kfree(pn); - return 1; - } + if (!pn->lruvec_stats_percpu) + goto fail; =20 lruvec_init(&pn->lruvec); pn->memcg =3D memcg; =20 memcg->nodeinfo[node] =3D pn; return 0; +fail: + kfree(pn->lruvec_stats); + kfree(pn); + return 1; } =20 static void free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int no= de) @@ -5513,6 +5574,7 @@ static void free_mem_cgroup_per_node_info(struct mem_= cgroup *memcg, int node) return; =20 free_percpu(pn->lruvec_stats_percpu); + kfree(pn->lruvec_stats); kfree(pn); } =20 @@ -5865,18 +5927,19 @@ static void mem_cgroup_css_rstat_flush(struct cgrou= p_subsys_state *css, int cpu) =20 for_each_node_state(nid, N_MEMORY) { struct mem_cgroup_per_node *pn =3D memcg->nodeinfo[nid]; - struct mem_cgroup_per_node *ppn =3D NULL; + struct lruvec_stats *lstats =3D pn->lruvec_stats; + struct lruvec_stats *plstats =3D NULL; struct lruvec_stats_percpu *lstatc; =20 if (parent) - ppn =3D parent->nodeinfo[nid]; + plstats =3D parent->nodeinfo[nid]->lruvec_stats; =20 lstatc =3D per_cpu_ptr(pn->lruvec_stats_percpu, cpu); =20 for (i =3D 0; i < NR_VM_NODE_STAT_ITEMS; i++) { - delta =3D pn->lruvec_stats.state_pending[i]; + delta =3D lstats->state_pending[i]; if (delta) - pn->lruvec_stats.state_pending[i] =3D 0; + lstats->state_pending[i] =3D 0; =20 delta_cpu =3D 0; v =3D READ_ONCE(lstatc->state[i]); @@ -5887,12 +5950,12 @@ static void mem_cgroup_css_rstat_flush(struct cgrou= p_subsys_state *css, int cpu) } =20 if (delta_cpu) - pn->lruvec_stats.state_local[i] +=3D delta_cpu; + lstats->state_local[i] +=3D delta_cpu; =20 if (delta) { - pn->lruvec_stats.state[i] +=3D delta; - if (ppn) - ppn->lruvec_stats.state_pending[i] +=3D delta; + lstats->state[i] +=3D delta; + if (plstats) + plstats->state_pending[i] +=3D delta; } } } --=20 2.43.0