From nobody Thu Apr 9 14:00:16 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 28B12ECAAA1 for ; Wed, 31 Aug 2022 03:21:42 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232317AbiHaDVj (ORCPT ); Tue, 30 Aug 2022 23:21:39 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57398 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231187AbiHaDUm (ORCPT ); Tue, 30 Aug 2022 23:20:42 -0400 Received: from mail-pl1-x634.google.com (mail-pl1-x634.google.com [IPv6:2607:f8b0:4864:20::634]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 4434854CAD for ; Tue, 30 Aug 2022 20:20:33 -0700 (PDT) Received: by mail-pl1-x634.google.com with SMTP id p18so12950711plr.8 for ; Tue, 30 Aug 2022 20:20:32 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=bytedance-com.20210112.gappssmtp.com; s=20210112; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date; bh=lt2/YaYHr5rJLJwkCdxmg5YZwJpd4IJKxnznGjt9X+s=; b=0N1toz3Fx8xpZdKGsdyOMDfgo8HENQJGt/GN4Dq6KrIB0nFRHfbVwHakUBxIfB83zH OD3SQlKQwLFqQAQvNMZngYz1rHGLtqFpAZnywjpkWRfnzF2m6ZfxhBWgBjoM0Sq2jASR 2Zt850BxiEaXuKGEhE7zh9U0jpehhZmWl4WJH0TouQmFDhjCHNnnunxCD0l336wDAN/M 4Iu0fhbDHOgFEcyCDlueiskgx7uw/wLlKSFo93+IbvGxNS/eJaOB4isEYA+6uS8FQ7ZZ KguwklR7PIDPotwjIN6d+bfyAY9LiYL/hnFBIa36LXV37N/N/pRWK0daTQK+777FYSRa v+eA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date; bh=lt2/YaYHr5rJLJwkCdxmg5YZwJpd4IJKxnznGjt9X+s=; b=ixOP0rOg9tAH8CUzijFsfK/1kyI6CnaQSbqGhjheu9ulQadYrynmLj8SEACLMoc0RP PZJ1o3iK4tdXt3Qb7vTko6NyUlMmTN9D+rWUFOD7lTg9sTZIir9Ik665MavJCaorITmQ ws/KVWfrpwJO9txLkmm4DFiM23D0+XCNgZ0nVC752c0tspPaTAyaavGU/xVaZgrjcChY MYHUF0Rz0mm4/8sPDKqunli09xzG0cyu4UquhBo3N6N8LokYYqpxMh1zPtHzSyeegpf6 IrLAZKSiBsFK68hSJqs3y4uZ+yCmumsQybiNmCSCnoFVnxRirIX3MZdLrIFQfotGoCwR YxZA== X-Gm-Message-State: ACgBeo3tY/stJzbsuMIjq13D70Eb1PbUJwkaRBKwgC6NtWrAwUQQcilf G3MGtUcDEJq9RDb/FXQGQ6f2DA== X-Google-Smtp-Source: AA6agR6WVNQbj0YQ9diql3Yc2i0fOPGwhWAQ0rKwIzK2ns4up9w0TNuEUbdHfL3HCAdE6u8Kh98pqA== X-Received: by 2002:a17:90b:1812:b0:1fd:d509:93e5 with SMTP id lw18-20020a17090b181200b001fdd50993e5mr1176557pjb.25.1661916032237; Tue, 30 Aug 2022 20:20:32 -0700 (PDT) Received: from C02DW0BEMD6R.bytedance.net ([139.177.225.245]) by smtp.gmail.com with ESMTPSA id i13-20020a170902c94d00b0015e8d4eb1d5sm8633535pla.31.2022.08.30.20.20.28 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 30 Aug 2022 20:20:31 -0700 (PDT) From: Qi Zheng To: akpm@linux-foundation.org, shy828301@gmail.com, willy@infradead.org, vbabka@suse.cz, hannes@cmpxchg.org, minchan@kernel.org, rppt@kernel.org Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Qi Zheng Subject: [PATCH v2 5/7] ksm: convert ksm_mm_slot.mm_list to ksm_mm_slot.mm_node Date: Wed, 31 Aug 2022 11:19:49 +0800 Message-Id: <20220831031951.43152-6-zhengqi.arch@bytedance.com> X-Mailer: git-send-email 2.24.3 (Apple Git-128) In-Reply-To: <20220831031951.43152-1-zhengqi.arch@bytedance.com> References: <20220831031951.43152-1-zhengqi.arch@bytedance.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" In order to use common struct mm_slot, convert ksm_mm_slot.mm_list to ksm_mm_slot.mm_node in advance, no functional change. Signed-off-by: Qi Zheng --- mm/ksm.c | 40 ++++++++++++++++++++-------------------- 1 file changed, 20 insertions(+), 20 deletions(-) diff --git a/mm/ksm.c b/mm/ksm.c index 3937111f9ab8..8c52aa7e0a02 100644 --- a/mm/ksm.c +++ b/mm/ksm.c @@ -114,13 +114,13 @@ /** * struct ksm_mm_slot - ksm information per mm that is being scanned * @link: link to the mm_slots hash list - * @mm_list: link into the mm_slots list, rooted in ksm_mm_head + * @mm_node: link into the mm_slots list, rooted in ksm_mm_head * @rmap_list: head for this mm_slot's singly-linked list of rmap_items * @mm: the mm that this information is valid for */ struct ksm_mm_slot { struct hlist_node link; - struct list_head mm_list; + struct list_head mm_node; struct ksm_rmap_item *rmap_list; struct mm_struct *mm; }; @@ -231,7 +231,7 @@ static LIST_HEAD(migrate_nodes); static DEFINE_HASHTABLE(mm_slots_hash, MM_SLOTS_HASH_BITS); =20 static struct ksm_mm_slot ksm_mm_head =3D { - .mm_list =3D LIST_HEAD_INIT(ksm_mm_head.mm_list), + .mm_node =3D LIST_HEAD_INIT(ksm_mm_head.mm_node), }; static struct ksm_scan ksm_scan =3D { .mm_slot =3D &ksm_mm_head, @@ -980,8 +980,8 @@ static int unmerge_and_remove_all_rmap_items(void) int err =3D 0; =20 spin_lock(&ksm_mmlist_lock); - ksm_scan.mm_slot =3D list_entry(ksm_mm_head.mm_list.next, - struct ksm_mm_slot, mm_list); + ksm_scan.mm_slot =3D list_entry(ksm_mm_head.mm_node.next, + struct ksm_mm_slot, mm_node); spin_unlock(&ksm_mmlist_lock); =20 for (mm_slot =3D ksm_scan.mm_slot; mm_slot !=3D &ksm_mm_head; @@ -1005,11 +1005,11 @@ static int unmerge_and_remove_all_rmap_items(void) mmap_read_unlock(mm); =20 spin_lock(&ksm_mmlist_lock); - ksm_scan.mm_slot =3D list_entry(mm_slot->mm_list.next, - struct ksm_mm_slot, mm_list); + ksm_scan.mm_slot =3D list_entry(mm_slot->mm_node.next, + struct ksm_mm_slot, mm_node); if (ksm_test_exit(mm)) { hash_del(&mm_slot->link); - list_del(&mm_slot->mm_list); + list_del(&mm_slot->mm_node); spin_unlock(&ksm_mmlist_lock); =20 free_mm_slot(mm_slot); @@ -2250,7 +2250,7 @@ static struct ksm_rmap_item *scan_get_next_rmap_item(= struct page **page) struct vma_iterator vmi; int nid; =20 - if (list_empty(&ksm_mm_head.mm_list)) + if (list_empty(&ksm_mm_head.mm_node)) return NULL; =20 slot =3D ksm_scan.mm_slot; @@ -2291,7 +2291,7 @@ static struct ksm_rmap_item *scan_get_next_rmap_item(= struct page **page) root_unstable_tree[nid] =3D RB_ROOT; =20 spin_lock(&ksm_mmlist_lock); - slot =3D list_entry(slot->mm_list.next, struct ksm_mm_slot, mm_list); + slot =3D list_entry(slot->mm_node.next, struct ksm_mm_slot, mm_node); ksm_scan.mm_slot =3D slot; spin_unlock(&ksm_mmlist_lock); /* @@ -2364,8 +2364,8 @@ static struct ksm_rmap_item *scan_get_next_rmap_item(= struct page **page) remove_trailing_rmap_items(ksm_scan.rmap_list); =20 spin_lock(&ksm_mmlist_lock); - ksm_scan.mm_slot =3D list_entry(slot->mm_list.next, - struct ksm_mm_slot, mm_list); + ksm_scan.mm_slot =3D list_entry(slot->mm_node.next, + struct ksm_mm_slot, mm_node); if (ksm_scan.address =3D=3D 0) { /* * We've completed a full scan of all vmas, holding mmap_lock @@ -2377,7 +2377,7 @@ static struct ksm_rmap_item *scan_get_next_rmap_item(= struct page **page) * mmap_lock then protects against race with MADV_MERGEABLE). */ hash_del(&slot->link); - list_del(&slot->mm_list); + list_del(&slot->mm_node); spin_unlock(&ksm_mmlist_lock); =20 free_mm_slot(slot); @@ -2426,7 +2426,7 @@ static void ksm_do_scan(unsigned int scan_npages) =20 static int ksmd_should_run(void) { - return (ksm_run & KSM_RUN_MERGE) && !list_empty(&ksm_mm_head.mm_list); + return (ksm_run & KSM_RUN_MERGE) && !list_empty(&ksm_mm_head.mm_node); } =20 static int ksm_scan_thread(void *nothing) @@ -2523,7 +2523,7 @@ int __ksm_enter(struct mm_struct *mm) return -ENOMEM; =20 /* Check ksm_run too? Would need tighter locking */ - needs_wakeup =3D list_empty(&ksm_mm_head.mm_list); + needs_wakeup =3D list_empty(&ksm_mm_head.mm_node); =20 spin_lock(&ksm_mmlist_lock); insert_to_mm_slots_hash(mm, mm_slot); @@ -2538,9 +2538,9 @@ int __ksm_enter(struct mm_struct *mm) * missed: then we might as well insert at the end of the list. */ if (ksm_run & KSM_RUN_UNMERGE) - list_add_tail(&mm_slot->mm_list, &ksm_mm_head.mm_list); + list_add_tail(&mm_slot->mm_node, &ksm_mm_head.mm_node); else - list_add_tail(&mm_slot->mm_list, &ksm_scan.mm_slot->mm_list); + list_add_tail(&mm_slot->mm_node, &ksm_scan.mm_slot->mm_node); spin_unlock(&ksm_mmlist_lock); =20 set_bit(MMF_VM_MERGEABLE, &mm->flags); @@ -2571,11 +2571,11 @@ void __ksm_exit(struct mm_struct *mm) if (mm_slot && ksm_scan.mm_slot !=3D mm_slot) { if (!mm_slot->rmap_list) { hash_del(&mm_slot->link); - list_del(&mm_slot->mm_list); + list_del(&mm_slot->mm_node); easy_to_free =3D 1; } else { - list_move(&mm_slot->mm_list, - &ksm_scan.mm_slot->mm_list); + list_move(&mm_slot->mm_node, + &ksm_scan.mm_slot->mm_node); } } spin_unlock(&ksm_mmlist_lock); --=20 2.20.1