From nobody Fri Dec 19 12:29:11 2025 Received: from out-181.mta1.migadu.com (out-181.mta1.migadu.com [95.215.58.181]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 92A7D2577B for ; Sat, 27 Apr 2024 00:37:45 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.181 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178267; cv=none; b=azwWRG8Da2LcVUd5jqlbW25YFek6pzk4hhGirJuJMM2+mtLm9aYYe/zzheB8hpQH4ESgO1gQGYNVb3cPyDrKxohQ5ZNQIi1hom3B15LJW8PqQsSc0+LYVmdahF35sA8ep/iI3jaK0ejSQR9yaAZfTfsAAL02jMZR4iAMYQFg49o= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178267; c=relaxed/simple; bh=tJK4g+poJ9VBpRehNiwwG1ZwM1nTxzq5jA9+UhY0/zc=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=GgBcGdFGOHHsEYRpnhJ69VeN807mIMgJiM5VMAHve6EOflh18aImACHgChxzHjgeQR846Gi1X6FtsNitQclqN2l9L5lXwi6JVEyfdkBsOLnI/ZQJ6G9Pmh4t7V9jeiytMICmXpoXvS1QpnnwzFIqQ6SFP5tIS7V8LC9l1569MDs= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=Nm5yyRv3; arc=none smtp.client-ip=95.215.58.181 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="Nm5yyRv3" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178263; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=Gfv3iMSIWQC4QbQ/6GXaLS8Cjd7JxtChDyD5rQgDX9Q=; b=Nm5yyRv30txwaGJ50+RpR3dkkQIdU/cAkah0rz5E9NaWS6JEnqIlf8pCbX0XdNDBFnxc6N Hbz4uF6xZb+psVAA6G6MVW0R9CvKnIaxzNY6sPTfTQEuMd3bY+WiOBBBHx+n4a1jgGVc7m SpP+ZPeUku+xQYxxAw2TOWgKmw/JcxQ= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 1/7] memcg: reduce memory size of mem_cgroup_events_index Date: Fri, 26 Apr 2024 17:37:27 -0700 Message-ID: <20240427003733.3898961-2-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" mem_cgroup_events_index is a translation table to get the right index of the memcg relevant entry for the general vm_event_item. At the moment, it is defined as integer array. However on a typical system the max entry of vm_event_item (NR_VM_EVENT_ITEMS) is 113, so we don't need to use int as storage type of the array. For now just use int8_t as type and add a BUILD_BUG_ON() and will switch to short once NR_VM_EVENT_ITEMS touches 127. Signed-off-by: Shakeel Butt Reviewed-by: Roman Gushchin --- mm/memcontrol.c | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 602ad5faad4d..53769d06053f 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -607,11 +607,14 @@ static const unsigned int memcg_vm_event_stat[] =3D { }; =20 #define NR_MEMCG_EVENTS ARRAY_SIZE(memcg_vm_event_stat) -static int mem_cgroup_events_index[NR_VM_EVENT_ITEMS] __read_mostly; +static int8_t mem_cgroup_events_index[NR_VM_EVENT_ITEMS] __read_mostly; =20 static void init_memcg_events(void) { - int i; + int8_t i; + + /* Switch to short once this failure occurs. */ + BUILD_BUG_ON(NR_VM_EVENT_ITEMS >=3D 127 /* INT8_MAX */); =20 for (i =3D 0; i < NR_MEMCG_EVENTS; ++i) mem_cgroup_events_index[memcg_vm_event_stat[i]] =3D i + 1; --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-188.mta1.migadu.com (out-188.mta1.migadu.com [95.215.58.188]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 500032E852 for ; Sat, 27 Apr 2024 00:37:48 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.188 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178270; cv=none; b=myKJ8RIawiw8ZVHfb4Onk+fiSBoyMblY3Eb/y9bFkQV145q6nGAhkVuloO/vG0FGQlQ37DYJbgHRUckZkC4wUK2/34UNWzLNPLgsVsGGe0UVePE/gV0oCRsNg5sUEVmtPcm9ZndZ4vkn7l2Pe4WCPzvJ30RZC9lQ5JdOcm1ZWtA= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178270; c=relaxed/simple; bh=d1x8pyehYWPfCqWd+SbnXD6NG1swtEVTqk4fMqibs8w=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=JXvup8daAUk67dSsiV8Oh2emD0QAKDI/eL45XQNudg0S4RUeqDOSnIzpPSyOK05Mg3RDuvhlxuhwd9esdhGXfrkTkW2BG0TyMyDx0xXaQSfidSARZEw6Toqxta1DANZ9/tjemmFV4Q5ke0DC+FCxN+H9Avpko6But9TSj9ILcpc= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=TTzxVCz/; arc=none smtp.client-ip=95.215.58.188 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="TTzxVCz/" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178266; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=5HIEk0SwiGrn4sd8u1onunMHRY8Lr8R1eMIi9eNXAwg=; b=TTzxVCz/SkpxFmZQf92Tga9m8XT6tVu0QWSUn9K3qKo0E1mv96BJzNsXIz76H/EcjW0Psm aOZme5snQD4/yIdO02Ba+xutkIyUD+b/G5a8DIZz/9L/SuinYiiA5H2V5R0NuZ+SUXmxdq n/1LeeGxekFuaPZVKjK+TbCPzXMu6FA= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 2/7] memcg: dynamically allocate lruvec_stats Date: Fri, 26 Apr 2024 17:37:28 -0700 Message-ID: <20240427003733.3898961-3-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" To decouple the dependency of lruvec_stats on NR_VM_NODE_STAT_ITEMS, we need to dynamically allocate lruvec_stats in the mem_cgroup_per_node structure. Also move the definition of lruvec_stats_percpu and lruvec_stats and related functions to the memcontrol.c to facilitate later patches. No functional changes in the patch. Signed-off-by: Shakeel Butt Acked-by: Shakeel Butt Reviewed-by: Yosry Ahmed --- include/linux/memcontrol.h | 62 +++------------------------ mm/memcontrol.c | 87 ++++++++++++++++++++++++++++++++------ 2 files changed, 81 insertions(+), 68 deletions(-) diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index 9aba0d0462ca..ab8a6e884375 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -83,6 +83,8 @@ enum mem_cgroup_events_target { =20 struct memcg_vmstats_percpu; struct memcg_vmstats; +struct lruvec_stats_percpu; +struct lruvec_stats; =20 struct mem_cgroup_reclaim_iter { struct mem_cgroup *position; @@ -90,25 +92,6 @@ struct mem_cgroup_reclaim_iter { unsigned int generation; }; =20 -struct lruvec_stats_percpu { - /* Local (CPU and cgroup) state */ - long state[NR_VM_NODE_STAT_ITEMS]; - - /* Delta calculation for lockless upward propagation */ - long state_prev[NR_VM_NODE_STAT_ITEMS]; -}; - -struct lruvec_stats { - /* Aggregated (CPU and subtree) state */ - long state[NR_VM_NODE_STAT_ITEMS]; - - /* Non-hierarchical (CPU aggregated) state */ - long state_local[NR_VM_NODE_STAT_ITEMS]; - - /* Pending child counts during tree propagation */ - long state_pending[NR_VM_NODE_STAT_ITEMS]; -}; - /* * per-node information in memory controller. */ @@ -116,7 +99,7 @@ struct mem_cgroup_per_node { struct lruvec lruvec; =20 struct lruvec_stats_percpu __percpu *lruvec_stats_percpu; - struct lruvec_stats lruvec_stats; + struct lruvec_stats *lruvec_stats; =20 unsigned long lru_zone_size[MAX_NR_ZONES][NR_LRU_LISTS]; =20 @@ -1037,42 +1020,9 @@ static inline void mod_memcg_page_state(struct page = *page, } =20 unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx); - -static inline unsigned long lruvec_page_state(struct lruvec *lruvec, - enum node_stat_item idx) -{ - struct mem_cgroup_per_node *pn; - long x; - - if (mem_cgroup_disabled()) - return node_page_state(lruvec_pgdat(lruvec), idx); - - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats.state[idx]); -#ifdef CONFIG_SMP - if (x < 0) - x =3D 0; -#endif - return x; -} - -static inline unsigned long lruvec_page_state_local(struct lruvec *lruvec, - enum node_stat_item idx) -{ - struct mem_cgroup_per_node *pn; - long x =3D 0; - - if (mem_cgroup_disabled()) - return node_page_state(lruvec_pgdat(lruvec), idx); - - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats.state_local[idx]); -#ifdef CONFIG_SMP - if (x < 0) - x =3D 0; -#endif - return x; -} +unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item= idx); +unsigned long lruvec_page_state_local(struct lruvec *lruvec, + enum node_stat_item idx); =20 void mem_cgroup_flush_stats(struct mem_cgroup *memcg); void mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg); diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 53769d06053f..5e337ed6c6bf 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -576,6 +576,60 @@ mem_cgroup_largest_soft_limit_node(struct mem_cgroup_t= ree_per_node *mctz) return mz; } =20 +struct lruvec_stats_percpu { + /* Local (CPU and cgroup) state */ + long state[NR_VM_NODE_STAT_ITEMS]; + + /* Delta calculation for lockless upward propagation */ + long state_prev[NR_VM_NODE_STAT_ITEMS]; +}; + +struct lruvec_stats { + /* Aggregated (CPU and subtree) state */ + long state[NR_VM_NODE_STAT_ITEMS]; + + /* Non-hierarchical (CPU aggregated) state */ + long state_local[NR_VM_NODE_STAT_ITEMS]; + + /* Pending child counts during tree propagation */ + long state_pending[NR_VM_NODE_STAT_ITEMS]; +}; + +unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item= idx) +{ + struct mem_cgroup_per_node *pn; + long x; + + if (mem_cgroup_disabled()) + return node_page_state(lruvec_pgdat(lruvec), idx); + + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state[idx]); +#ifdef CONFIG_SMP + if (x < 0) + x =3D 0; +#endif + return x; +} + +unsigned long lruvec_page_state_local(struct lruvec *lruvec, + enum node_stat_item idx) +{ + struct mem_cgroup_per_node *pn; + long x =3D 0; + + if (mem_cgroup_disabled()) + return node_page_state(lruvec_pgdat(lruvec), idx); + + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state_local[idx]); +#ifdef CONFIG_SMP + if (x < 0) + x =3D 0; +#endif + return x; +} + /* Subset of vm_event_item to report for memcg event stats */ static const unsigned int memcg_vm_event_stat[] =3D { PGPGIN, @@ -5492,18 +5546,25 @@ static int alloc_mem_cgroup_per_node_info(struct me= m_cgroup *memcg, int node) if (!pn) return 1; =20 + pn->lruvec_stats =3D kzalloc_node(sizeof(struct lruvec_stats), GFP_KERNEL, + node); + if (!pn->lruvec_stats) + goto fail; + pn->lruvec_stats_percpu =3D alloc_percpu_gfp(struct lruvec_stats_percpu, GFP_KERNEL_ACCOUNT); - if (!pn->lruvec_stats_percpu) { - kfree(pn); - return 1; - } + if (!pn->lruvec_stats_percpu) + goto fail; =20 lruvec_init(&pn->lruvec); pn->memcg =3D memcg; =20 memcg->nodeinfo[node] =3D pn; return 0; +fail: + kfree(pn->lruvec_stats); + kfree(pn); + return 1; } =20 static void free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int no= de) @@ -5514,6 +5575,7 @@ static void free_mem_cgroup_per_node_info(struct mem_= cgroup *memcg, int node) return; =20 free_percpu(pn->lruvec_stats_percpu); + kfree(pn->lruvec_stats); kfree(pn); } =20 @@ -5866,18 +5928,19 @@ static void mem_cgroup_css_rstat_flush(struct cgrou= p_subsys_state *css, int cpu) =20 for_each_node_state(nid, N_MEMORY) { struct mem_cgroup_per_node *pn =3D memcg->nodeinfo[nid]; - struct mem_cgroup_per_node *ppn =3D NULL; + struct lruvec_stats *lstats =3D pn->lruvec_stats; + struct lruvec_stats *plstats =3D NULL; struct lruvec_stats_percpu *lstatc; =20 if (parent) - ppn =3D parent->nodeinfo[nid]; + plstats =3D parent->nodeinfo[nid]->lruvec_stats; =20 lstatc =3D per_cpu_ptr(pn->lruvec_stats_percpu, cpu); =20 for (i =3D 0; i < NR_VM_NODE_STAT_ITEMS; i++) { - delta =3D pn->lruvec_stats.state_pending[i]; + delta =3D lstats->state_pending[i]; if (delta) - pn->lruvec_stats.state_pending[i] =3D 0; + lstats->state_pending[i] =3D 0; =20 delta_cpu =3D 0; v =3D READ_ONCE(lstatc->state[i]); @@ -5888,12 +5951,12 @@ static void mem_cgroup_css_rstat_flush(struct cgrou= p_subsys_state *css, int cpu) } =20 if (delta_cpu) - pn->lruvec_stats.state_local[i] +=3D delta_cpu; + lstats->state_local[i] +=3D delta_cpu; =20 if (delta) { - pn->lruvec_stats.state[i] +=3D delta; - if (ppn) - ppn->lruvec_stats.state_pending[i] +=3D delta; + lstats->state[i] +=3D delta; + if (plstats) + plstats->state_pending[i] +=3D delta; } } } --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-174.mta1.migadu.com (out-174.mta1.migadu.com [95.215.58.174]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 81ED739856 for ; Sat, 27 Apr 2024 00:37:51 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.174 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178273; cv=none; b=eYA4vFf2Qa8LFefjCxBCZ6zzkgKvko6yj6Y1/51wNZbV12mpLHOS4poiHn9Nbfzj6fxIqJzRzg/1Q1ayh6zL89rET9t7Uvq6tYOWc+TCfVUa/CjDzlc7n0M0r0V4h2GubVrPbYLpJ2vS6JjGKywAvtcwqfqdmN0Pp/PeWmbw2Ng= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178273; c=relaxed/simple; bh=ODuHC/xW2Sztqrth+tMBMfYpvtgz/0RD1c7BzWd7jG4=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=PujVj32CYsck/xiJp0/YP1YSISJnX0N5g8J2VbqXiNHOSRpU8GbO87aqHfS3CW9QE40XHNanU2bnZM6lfeDc3o7WQF4Pj8O8q/Dn0vREzAl1E6OTmVCDR2y/zPJtDHuSpsLZvRHdxhA48Kkm51/QAmkeDqBAla4Fed4wusz8abQ= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=q3eCDp0J; arc=none smtp.client-ip=95.215.58.174 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="q3eCDp0J" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178269; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=CiP6jJ3z/Zb+7HCzkE9EgNFU2GbKmyiZu4OerwR5i68=; b=q3eCDp0J9gOqeEqaX8tgBo8ZzkToNZNxyobMgMVMJQ9HF/mBNKsNmj1YUeUiJcf9PW3Sre yx9grJUbPPPWOTn07Rw0XLZzPbR+tJ83YdB8f1HVsij6xYHqrTTowobd5UzmTg2kJ+OKbX cChmVNaWnbQODbqk5iXRD5Oq8ndtlew= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 3/7] memcg: reduce memory for the lruvec and memcg stats Date: Fri, 26 Apr 2024 17:37:29 -0700 Message-ID: <20240427003733.3898961-4-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" At the moment, the amount of memory allocated for stats related structs in the mem_cgroup corresponds to the size of enum node_stat_item. However not all fields in enum node_stat_item has corresponding memcg stats. So, let's use indirection mechanism similar to the one used for memcg vmstats management. For a given x86_64 config, the size of stats with and without patch is: structs size in bytes w/o with struct lruvec_stats 1128 648 struct lruvec_stats_percpu 752 432 struct memcg_vmstats 1832 1352 struct memcg_vmstats_percpu 1280 960 The memory savings is further compounded by the fact that these structs are allocated for each cpu and for each node. To be precise, for each memcg the memory saved would be: Memory saved =3D ((21 * 3 * NR_NODES) + (21 * 2 * NR_NODS * NR_CPUS) + (21 * 3) + (21 * 2 * NR_CPUS)) * sizeof(long) Where 21 is the number of fields eliminated. Signed-off-by: Shakeel Butt --- mm/memcontrol.c | 138 ++++++++++++++++++++++++++++++++++++++++-------- 1 file changed, 115 insertions(+), 23 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 5e337ed6c6bf..c164bc9b8ed6 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -576,35 +576,105 @@ mem_cgroup_largest_soft_limit_node(struct mem_cgroup= _tree_per_node *mctz) return mz; } =20 +/* Subset of node_stat_item for memcg stats */ +static const unsigned int memcg_node_stat_items[] =3D { + NR_INACTIVE_ANON, + NR_ACTIVE_ANON, + NR_INACTIVE_FILE, + NR_ACTIVE_FILE, + NR_UNEVICTABLE, + NR_SLAB_RECLAIMABLE_B, + NR_SLAB_UNRECLAIMABLE_B, + WORKINGSET_REFAULT_ANON, + WORKINGSET_REFAULT_FILE, + WORKINGSET_ACTIVATE_ANON, + WORKINGSET_ACTIVATE_FILE, + WORKINGSET_RESTORE_ANON, + WORKINGSET_RESTORE_FILE, + WORKINGSET_NODERECLAIM, + NR_ANON_MAPPED, + NR_FILE_MAPPED, + NR_FILE_PAGES, + NR_FILE_DIRTY, + NR_WRITEBACK, + NR_SHMEM, + NR_SHMEM_THPS, + NR_FILE_THPS, + NR_ANON_THPS, + NR_KERNEL_STACK_KB, + NR_PAGETABLE, + NR_SECONDARY_PAGETABLE, +#ifdef CONFIG_SWAP + NR_SWAPCACHE, +#endif +}; + +static const unsigned int memcg_stat_items[] =3D { + MEMCG_SWAP, + MEMCG_SOCK, + MEMCG_PERCPU_B, + MEMCG_VMALLOC, + MEMCG_KMEM, + MEMCG_ZSWAP_B, + MEMCG_ZSWAPPED, +}; + +#define NR_MEMCG_NODE_STAT_ITEMS ARRAY_SIZE(memcg_node_stat_items) +#define NR_MEMCG_STATS (NR_MEMCG_NODE_STAT_ITEMS + ARRAY_SIZE(memcg_stat_i= tems)) +static int8_t mem_cgroup_stats_index[MEMCG_NR_STAT] __read_mostly; + +static void init_memcg_stats(void) +{ + int8_t i, j =3D 0; + + /* Switch to short once this failure occurs. */ + BUILD_BUG_ON(NR_MEMCG_STATS >=3D 127 /* INT8_MAX */); + + for (i =3D 0; i < NR_MEMCG_NODE_STAT_ITEMS; ++i) + mem_cgroup_stats_index[memcg_node_stat_items[i]] =3D ++j; + + for (i =3D 0; i < ARRAY_SIZE(memcg_stat_items); ++i) + mem_cgroup_stats_index[memcg_stat_items[i]] =3D ++j; +} + +static inline int memcg_stats_index(int idx) +{ + return mem_cgroup_stats_index[idx] - 1; +} + struct lruvec_stats_percpu { /* Local (CPU and cgroup) state */ - long state[NR_VM_NODE_STAT_ITEMS]; + long state[NR_MEMCG_NODE_STAT_ITEMS]; =20 /* Delta calculation for lockless upward propagation */ - long state_prev[NR_VM_NODE_STAT_ITEMS]; + long state_prev[NR_MEMCG_NODE_STAT_ITEMS]; }; =20 struct lruvec_stats { /* Aggregated (CPU and subtree) state */ - long state[NR_VM_NODE_STAT_ITEMS]; + long state[NR_MEMCG_NODE_STAT_ITEMS]; =20 /* Non-hierarchical (CPU aggregated) state */ - long state_local[NR_VM_NODE_STAT_ITEMS]; + long state_local[NR_MEMCG_NODE_STAT_ITEMS]; =20 /* Pending child counts during tree propagation */ - long state_pending[NR_VM_NODE_STAT_ITEMS]; + long state_pending[NR_MEMCG_NODE_STAT_ITEMS]; }; =20 unsigned long lruvec_page_state(struct lruvec *lruvec, enum node_stat_item= idx) { struct mem_cgroup_per_node *pn; - long x; + long x =3D 0; + int i; =20 if (mem_cgroup_disabled()) return node_page_state(lruvec_pgdat(lruvec), idx); =20 - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats->state[idx]); + i =3D memcg_stats_index(idx); + if (i >=3D 0) { + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state[i]); + } #ifdef CONFIG_SMP if (x < 0) x =3D 0; @@ -617,12 +687,16 @@ unsigned long lruvec_page_state_local(struct lruvec *= lruvec, { struct mem_cgroup_per_node *pn; long x =3D 0; + int i; =20 if (mem_cgroup_disabled()) return node_page_state(lruvec_pgdat(lruvec), idx); =20 - pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); - x =3D READ_ONCE(pn->lruvec_stats->state_local[idx]); + i =3D memcg_stats_index(idx); + if (i >=3D 0) { + pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); + x =3D READ_ONCE(pn->lruvec_stats->state_local[i]); + } #ifdef CONFIG_SMP if (x < 0) x =3D 0; @@ -690,11 +764,11 @@ struct memcg_vmstats_percpu { /* The above should fit a single cacheline for memcg_rstat_updated() */ =20 /* Local (CPU and cgroup) page state & events */ - long state[MEMCG_NR_STAT]; + long state[NR_MEMCG_STATS]; unsigned long events[NR_MEMCG_EVENTS]; =20 /* Delta calculation for lockless upward propagation */ - long state_prev[MEMCG_NR_STAT]; + long state_prev[NR_MEMCG_STATS]; unsigned long events_prev[NR_MEMCG_EVENTS]; =20 /* Cgroup1: threshold notifications & softlimit tree updates */ @@ -704,15 +778,15 @@ struct memcg_vmstats_percpu { =20 struct memcg_vmstats { /* Aggregated (CPU and subtree) page state & events */ - long state[MEMCG_NR_STAT]; + long state[NR_MEMCG_STATS]; unsigned long events[NR_MEMCG_EVENTS]; =20 /* Non-hierarchical (CPU aggregated) page state & events */ - long state_local[MEMCG_NR_STAT]; + long state_local[NR_MEMCG_STATS]; unsigned long events_local[NR_MEMCG_EVENTS]; =20 /* Pending child counts during tree propagation */ - long state_pending[MEMCG_NR_STAT]; + long state_pending[NR_MEMCG_STATS]; unsigned long events_pending[NR_MEMCG_EVENTS]; =20 /* Stats updates since the last flush */ @@ -845,7 +919,13 @@ static void flush_memcg_stats_dwork(struct work_struct= *w) =20 unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) { - long x =3D READ_ONCE(memcg->vmstats->state[idx]); + long x; + int i =3D memcg_stats_index(idx); + + if (i < 0) + return 0; + + x =3D READ_ONCE(memcg->vmstats->state[i]); #ifdef CONFIG_SMP if (x < 0) x =3D 0; @@ -877,18 +957,25 @@ static int memcg_state_val_in_pages(int idx, int val) */ void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) { - if (mem_cgroup_disabled()) + int i =3D memcg_stats_index(idx); + + if (mem_cgroup_disabled() || i < 0) return; =20 - __this_cpu_add(memcg->vmstats_percpu->state[idx], val); + __this_cpu_add(memcg->vmstats_percpu->state[i], val); memcg_rstat_updated(memcg, memcg_state_val_in_pages(idx, val)); } =20 /* idx can be of type enum memcg_stat_item or node_stat_item. */ static unsigned long memcg_page_state_local(struct mem_cgroup *memcg, int = idx) { - long x =3D READ_ONCE(memcg->vmstats->state_local[idx]); + long x; + int i =3D memcg_stats_index(idx); + + if (i < 0) + return 0; =20 + x =3D READ_ONCE(memcg->vmstats->state_local[i]); #ifdef CONFIG_SMP if (x < 0) x =3D 0; @@ -902,6 +989,10 @@ static void __mod_memcg_lruvec_state(struct lruvec *lr= uvec, { struct mem_cgroup_per_node *pn; struct mem_cgroup *memcg; + int i =3D memcg_stats_index(idx); + + if (i < 0) + return; =20 pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); memcg =3D pn->memcg; @@ -931,10 +1022,10 @@ static void __mod_memcg_lruvec_state(struct lruvec *= lruvec, } =20 /* Update memcg */ - __this_cpu_add(memcg->vmstats_percpu->state[idx], val); + __this_cpu_add(memcg->vmstats_percpu->state[i], val); =20 /* Update lruvec */ - __this_cpu_add(pn->lruvec_stats_percpu->state[idx], val); + __this_cpu_add(pn->lruvec_stats_percpu->state[i], val); =20 memcg_rstat_updated(memcg, memcg_state_val_in_pages(idx, val)); memcg_stats_unlock(); @@ -5702,6 +5793,7 @@ mem_cgroup_css_alloc(struct cgroup_subsys_state *pare= nt_css) page_counter_init(&memcg->kmem, &parent->kmem); page_counter_init(&memcg->tcpmem, &parent->tcpmem); } else { + init_memcg_stats(); init_memcg_events(); page_counter_init(&memcg->memory, NULL); page_counter_init(&memcg->swap, NULL); @@ -5873,7 +5965,7 @@ static void mem_cgroup_css_rstat_flush(struct cgroup_= subsys_state *css, int cpu) =20 statc =3D per_cpu_ptr(memcg->vmstats_percpu, cpu); =20 - for (i =3D 0; i < MEMCG_NR_STAT; i++) { + for (i =3D 0; i < NR_MEMCG_STATS; i++) { /* * Collect the aggregated propagation counts of groups * below us. We're in a per-cpu loop here and this is @@ -5937,7 +6029,7 @@ static void mem_cgroup_css_rstat_flush(struct cgroup_= subsys_state *css, int cpu) =20 lstatc =3D per_cpu_ptr(pn->lruvec_stats_percpu, cpu); =20 - for (i =3D 0; i < NR_VM_NODE_STAT_ITEMS; i++) { + for (i =3D 0; i < NR_MEMCG_NODE_STAT_ITEMS; i++) { delta =3D lstats->state_pending[i]; if (delta) lstats->state_pending[i] =3D 0; --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-170.mta1.migadu.com (out-170.mta1.migadu.com [95.215.58.170]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id C16563B1AE for ; Sat, 27 Apr 2024 00:37:54 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.170 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178276; cv=none; b=fGf80MKbDTaul14YHiMUQB+/LurTIQt7F6b6UdBXKTIBClmocJKZTjtnF9brBLa6k4uJLBlYLuQ9czsOPqGK16m4dBICO9rIaD3AnRvQ8ioeha47CZeh/QJovFNVIXyWga30s+VN+zJcE4W5KejbursOMGYVFbMiHly1gVjijRA= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178276; c=relaxed/simple; bh=2c11/UNucjhYuvQsF9rHuh1tnmfsKj/luR52vgQMaxk=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=NIz4cRaZLwOhR0LWmmCmY6lrB2cgGi60MKMaEft0UDVOHwrE1aO8vjKnCMPBdmhRKrXKPyI/gSurV/aIxth5hFWVr7zrvTRyb6IiFx88pQ7lImlZChObpWc+UB4h9SK+IhCMn3nHuPA2qW4cTtwLdGWNKB7j5G+rDaQT+WsrQAs= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=b6gjrMhu; arc=none smtp.client-ip=95.215.58.170 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="b6gjrMhu" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178273; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=StJiZmI3BplMAQI82xrhJ2SVU48fOC0iWz8HtDgUnJo=; b=b6gjrMhuur3+QzQZZhAjDMsktwDbCY5i5Y/S3EnI6cT4PrqrIopqb7fjdGAwV1+f85fJkj oXiUuqCSxo3ONQADEVCNbYBnGKPDW1uurVXpqggrM7xrGlwYB8WMKdS6/0/lRWwBF5Hmo5 wt+amLltg08cTAW21mxIRKMXJMRUjFc= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 4/7] memcg: cleanup __mod_memcg_lruvec_state Date: Fri, 26 Apr 2024 17:37:30 -0700 Message-ID: <20240427003733.3898961-5-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" There are no memcg specific stats for NR_SHMEM_PMDMAPPED and NR_FILE_PMDMAPPED. Let's remove them. Signed-off-by: Shakeel Butt Reviewed-by: Roman Gushchin Reviewed-by: Yosry Ahmed --- mm/memcontrol.c | 2 -- 1 file changed, 2 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index c164bc9b8ed6..103e0e53e20a 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -1009,8 +1009,6 @@ static void __mod_memcg_lruvec_state(struct lruvec *l= ruvec, case NR_ANON_MAPPED: case NR_FILE_MAPPED: case NR_ANON_THPS: - case NR_SHMEM_PMDMAPPED: - case NR_FILE_PMDMAPPED: if (WARN_ON_ONCE(!in_task())) pr_warn("stat item index: %d\n", idx); break; --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-185.mta1.migadu.com (out-185.mta1.migadu.com [95.215.58.185]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 784C73C473 for ; Sat, 27 Apr 2024 00:37:57 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.185 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178279; cv=none; b=tIAY0z4MiqoB44nPKgPko0CGAyFJuYhVC/W6bXMYxFthbRGwqVPTHH5GhVZtHwNhwMkHyUTtDLQM+60jb1d8vfmdbiAgLuhSiHWaXsEq7dYcNK8ekpfw8FSZbAW4SIxQY4XFb5oJqbzD+ly5q6PYIshUL2HPYVB+tKIKnumGfss= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178279; c=relaxed/simple; bh=BJaMUZqNYD+vFZj/LHcHOqshLcOBcpf1OX6HJEFXK+Y=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=giOMp4R8sf1yDgyAPOk4/F+gem9ZsnNUQmeu93XCdif34VCtOGWR3REYSBuhu0YR8E2+FVFj4aZqWmvhKiZ+pCHoSwtBWYdaBxphBkC4VV31EQBK9t8IkbOPsMiQkfG3soVM9kJrGP7e1S24dzIydJFansn9ZGboSc8jHFZ+5YY= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=houO09vD; arc=none smtp.client-ip=95.215.58.185 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="houO09vD" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178275; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=zLHD4gb8t6zDRcF8qzenGgWmsGCHlQSRICk2vCIqKfk=; b=houO09vDhohFagXTbOfhLSJqhpc84FITSFNLRlY6ylNtuLjFRkyPyDJyqpgi6EbQHK4FZn RhE2P7AnzkgfYA/sfqlnm1T5oG8du97LuVbDILIJXqYXe1+56MCMYjkGUrV8Oga2UzpQz5 EeBGScH0Y4HpVoFLboUXkjNMFabeuPk= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 5/7] memcg: pr_warn_once for unexpected events and stats Date: Fri, 26 Apr 2024 17:37:31 -0700 Message-ID: <20240427003733.3898961-6-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" To reduce memory usage by the memcg events and stats, the kernel uses indirection table and only allocate stats and events which are being used by the memcg code. To make this more robust, let's add warnings where unexpected stats and events indexes are used. Signed-off-by: Shakeel Butt --- mm/memcontrol.c | 43 ++++++++++++++++++++++++++++++++++--------- 1 file changed, 34 insertions(+), 9 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 103e0e53e20a..36145089dcf5 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -671,9 +671,11 @@ unsigned long lruvec_page_state(struct lruvec *lruvec,= enum node_stat_item idx) return node_page_state(lruvec_pgdat(lruvec), idx); =20 i =3D memcg_stats_index(idx); - if (i >=3D 0) { + if (likely(i >=3D 0)) { pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); x =3D READ_ONCE(pn->lruvec_stats->state[i]); + } else { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); } #ifdef CONFIG_SMP if (x < 0) @@ -693,9 +695,11 @@ unsigned long lruvec_page_state_local(struct lruvec *l= ruvec, return node_page_state(lruvec_pgdat(lruvec), idx); =20 i =3D memcg_stats_index(idx); - if (i >=3D 0) { + if (likely(i >=3D 0)) { pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); x =3D READ_ONCE(pn->lruvec_stats->state_local[i]); + } else { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); } #ifdef CONFIG_SMP if (x < 0) @@ -922,8 +926,10 @@ unsigned long memcg_page_state(struct mem_cgroup *memc= g, int idx) long x; int i =3D memcg_stats_index(idx); =20 - if (i < 0) + if (unlikely(i < 0)) { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); return 0; + } =20 x =3D READ_ONCE(memcg->vmstats->state[i]); #ifdef CONFIG_SMP @@ -959,8 +965,13 @@ void __mod_memcg_state(struct mem_cgroup *memcg, int i= dx, int val) { int i =3D memcg_stats_index(idx); =20 - if (mem_cgroup_disabled() || i < 0) + if (mem_cgroup_disabled()) + return; + + if (unlikely(i < 0)) { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); return; + } =20 __this_cpu_add(memcg->vmstats_percpu->state[i], val); memcg_rstat_updated(memcg, memcg_state_val_in_pages(idx, val)); @@ -972,8 +983,10 @@ static unsigned long memcg_page_state_local(struct mem= _cgroup *memcg, int idx) long x; int i =3D memcg_stats_index(idx); =20 - if (i < 0) + if (unlikely(i < 0)) { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); return 0; + } =20 x =3D READ_ONCE(memcg->vmstats->state_local[i]); #ifdef CONFIG_SMP @@ -991,8 +1004,10 @@ static void __mod_memcg_lruvec_state(struct lruvec *l= ruvec, struct mem_cgroup *memcg; int i =3D memcg_stats_index(idx); =20 - if (i < 0) + if (unlikely(i < 0)) { + pr_warn_once("%s: stat item index: %d\n", __func__, idx); return; + } =20 pn =3D container_of(lruvec, struct mem_cgroup_per_node, lruvec); memcg =3D pn->memcg; @@ -1107,8 +1122,13 @@ void __count_memcg_events(struct mem_cgroup *memcg, = enum vm_event_item idx, { int index =3D memcg_events_index(idx); =20 - if (mem_cgroup_disabled() || index < 0) + if (mem_cgroup_disabled()) + return; + + if (unlikely(index < 0)) { + pr_warn_once("%s: event item index: %d\n", __func__, idx); return; + } =20 memcg_stats_lock(); __this_cpu_add(memcg->vmstats_percpu->events[index], count); @@ -1120,8 +1140,11 @@ static unsigned long memcg_events(struct mem_cgroup = *memcg, int event) { int index =3D memcg_events_index(event); =20 - if (index < 0) + if (unlikely(index < 0)) { + pr_warn_once("%s: event item index: %d\n", __func__, event); return 0; + } + return READ_ONCE(memcg->vmstats->events[index]); } =20 @@ -1129,8 +1152,10 @@ static unsigned long memcg_events_local(struct mem_c= group *memcg, int event) { int index =3D memcg_events_index(event); =20 - if (index < 0) + if (unlikely(index < 0)) { + pr_warn_once("%s: event item index: %d\n", __func__, event); return 0; + } =20 return READ_ONCE(memcg->vmstats->events_local[index]); } --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-179.mta0.migadu.com (out-179.mta0.migadu.com [91.218.175.179]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 281233FBA2 for ; Sat, 27 Apr 2024 00:38:02 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=91.218.175.179 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178284; cv=none; b=plcSjjEc6J2bwYj5DmdlMUDhQC2udnGVaiKauydy4FYboTuAfmZ1UoPSP/yszNfiCaIR34g9SjQfBgxHYwDH9LcpkyJ8qiEFHSuXGeuu6SAO6fuCKzp3zpaZ6DbMLEW9qvRAC5uO/83sCAAI2eULxSJHA/20dNVUm7EFR33Q8vU= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178284; c=relaxed/simple; bh=eMj3Z5zlEYlTfocBWKvpvJxSBuEHTxAA4dcpTuqRF/c=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=gQCvhxVaEv8/spAhKoP7gSPE2zBN0jxLrb7NuUPAbJTs1c63lJHD5XHmuOt+SQZjfWV2eSG82ai0iUexImhuMsTg4RMHQf5OoweokefGaY3LRwhKPxyPTgbNzfXkGebxILjg+oThl4mN/iWrrK1oafAWFkinRfYZN+dNnyT9NCc= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=sivGlrSp; arc=none smtp.client-ip=91.218.175.179 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="sivGlrSp" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178280; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=Gx3PKav0CetNdFI0jFUffF6lUGwk4q4qKzC3INQ2E8A=; b=sivGlrSpl81qjvi6K7a/ZlE+lX7x9zLqRfOxT3jTy3TpHNeCbRa/CRH5yGHNGDKvqIijXu BGuhy0LIWV81fc/WqvGITWNss9qyDo50w1qvtd+jaICAMwbaUac4gGWLUWsi6XL7uBHCmV pOaOEV7xvE4/E41E676iu4kaYa9KQZU= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 6/7] memcg: use proper type for mod_memcg_state Date: Fri, 26 Apr 2024 17:37:32 -0700 Message-ID: <20240427003733.3898961-7-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" The memcg stats update functions can take arbitrary integer but the only input which make sense is enum memcg_stat_item and we don't want these functions to be called with arbitrary integer, so replace the parameter type with enum memcg_stat_item and compiler will be able to warn if memcg stat update functions are called with incorrect index value. Signed-off-by: Shakeel Butt --- include/linux/memcontrol.h | 13 +++++++------ mm/memcontrol.c | 3 ++- 2 files changed, 9 insertions(+), 7 deletions(-) diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index ab8a6e884375..73cad69dfb5a 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -974,7 +974,8 @@ void mem_cgroup_print_oom_group(struct mem_cgroup *memc= g); void folio_memcg_lock(struct folio *folio); void folio_memcg_unlock(struct folio *folio); =20 -void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val); +void __mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, + int val); =20 /* try to stablize folio_memcg() for all the pages in a memcg */ static inline bool mem_cgroup_trylock_pages(struct mem_cgroup *memcg) @@ -995,7 +996,7 @@ static inline void mem_cgroup_unlock_pages(void) =20 /* idx can be of type enum memcg_stat_item or node_stat_item */ static inline void mod_memcg_state(struct mem_cgroup *memcg, - int idx, int val) + enum memcg_stat_item idx, int val) { unsigned long flags; =20 @@ -1005,7 +1006,7 @@ static inline void mod_memcg_state(struct mem_cgroup = *memcg, } =20 static inline void mod_memcg_page_state(struct page *page, - int idx, int val) + enum memcg_stat_item idx, int val) { struct mem_cgroup *memcg; =20 @@ -1491,19 +1492,19 @@ static inline void mem_cgroup_print_oom_group(struc= t mem_cgroup *memcg) } =20 static inline void __mod_memcg_state(struct mem_cgroup *memcg, - int idx, + enum memcg_stat_item idx, int nr) { } =20 static inline void mod_memcg_state(struct mem_cgroup *memcg, - int idx, + enum memcg_stat_item idx, int nr) { } =20 static inline void mod_memcg_page_state(struct page *page, - int idx, int val) + enum memcg_stat_item idx, int val) { } =20 diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 36145089dcf5..d11536ef59ef 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -961,7 +961,8 @@ static int memcg_state_val_in_pages(int idx, int val) * @idx: the stat item - can be enum memcg_stat_item or enum node_stat_item * @val: delta to add to the counter, can be negative */ -void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) +void __mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, + int val) { int i =3D memcg_stats_index(idx); =20 --=20 2.43.0 From nobody Fri Dec 19 12:29:11 2025 Received: from out-189.mta1.migadu.com (out-189.mta1.migadu.com [95.215.58.189]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 3DD9B40847 for ; Sat, 27 Apr 2024 00:38:06 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=95.215.58.189 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178287; cv=none; b=uyBpr//ArkeDU5KJled27LKWSpWe84TwowkM2aJiWodQ7gVWmrtleHeMLxA1zqIfH+2udu0PP2Y5Ok2MP+iIUOKAXXSKDyCk5NeC+khca3TBCTBfy10+I8NyBOx+8DbD17Tby6sjZ1/U5uKcd8SbGuvyBoRczUWy5KjdRKFU0rE= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1714178287; c=relaxed/simple; bh=dkznO5UFLYRjlvo6rhP3U2XgAo3NdmOtc60UDRFxx54=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=X9Gq1B+nVTq8zhUpLgOW90AAmKctRS+wPwRaLRfdX+y5PuWBli9+4626OaUgzIPyeTLpKUSmVw9upHbF9Hp4r3goqMoL85AEX+B37iqSomj+7VNiy3yWKv5Rip/L3Gs6SgNuyqJq8bGOadYEguMRr14VCvoiRkCtBa+2kNlN2Xc= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev; spf=pass smtp.mailfrom=linux.dev; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b=SncPF2vo; arc=none smtp.client-ip=95.215.58.189 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.dev Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.dev Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.dev header.i=@linux.dev header.b="SncPF2vo" X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1714178284; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=XaVmG+nL1Z/KgIhQvtyGGfErTWQGfN9RM5fOWCMfM60=; b=SncPF2vo7Gj5pGR2HtW7z18CEr5iYUMsR9J3GEV4lMaLOx6+PYwok7wMxTGAqdmd/XJczt pZgDrZ1qNeEJIRnvjIs1Aeyf4VLQRcEnJYOnsAE/Dy3rzsexN2xS0tWE7DQUd+u6a/yuZU qWm6PPDolOHny72+IBIloElPU+qJuFI= From: Shakeel Butt To: Andrew Morton , Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v2 7/7] mm: cleanup WORKINGSET_NODES in workingset Date: Fri, 26 Apr 2024 17:37:33 -0700 Message-ID: <20240427003733.3898961-8-shakeel.butt@linux.dev> In-Reply-To: <20240427003733.3898961-1-shakeel.butt@linux.dev> References: <20240427003733.3898961-1-shakeel.butt@linux.dev> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Migadu-Flow: FLOW_OUT Content-Type: text/plain; charset="utf-8" WORKINGSET_NODES is not exposed in the memcg stats and thus there is no need to use the memcg specific stat update functions for it. In future if we decide to expose WORKINGSET_NODES in the memcg stats, we can revert this patch. Signed-off-by: Shakeel Butt Reviewed-by: Roman Gushchin --- mm/workingset.c | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/mm/workingset.c b/mm/workingset.c index f2a0ecaf708d..c22adb93622a 100644 --- a/mm/workingset.c +++ b/mm/workingset.c @@ -618,6 +618,7 @@ struct list_lru shadow_nodes; void workingset_update_node(struct xa_node *node) { struct address_space *mapping; + struct page *page =3D virt_to_page(node); =20 /* * Track non-empty nodes that contain only shadow entries; @@ -633,12 +634,12 @@ void workingset_update_node(struct xa_node *node) if (node->count && node->count =3D=3D node->nr_values) { if (list_empty(&node->private_list)) { list_lru_add_obj(&shadow_nodes, &node->private_list); - __inc_lruvec_kmem_state(node, WORKINGSET_NODES); + __inc_node_page_state(page, WORKINGSET_NODES); } } else { if (!list_empty(&node->private_list)) { list_lru_del_obj(&shadow_nodes, &node->private_list); - __dec_lruvec_kmem_state(node, WORKINGSET_NODES); + __dec_node_page_state(page, WORKINGSET_NODES); } } } @@ -742,7 +743,7 @@ static enum lru_status shadow_lru_isolate(struct list_h= ead *item, } =20 list_lru_isolate(lru, item); - __dec_lruvec_kmem_state(node, WORKINGSET_NODES); + __dec_node_page_state(virt_to_page(node), WORKINGSET_NODES); =20 spin_unlock(lru_lock); =20 --=20 2.43.0