From nobody Fri Dec 19 09:29:30 2025 Received: from mx0b-00364e01.pphosted.com (mx0b-00364e01.pphosted.com [148.163.139.74]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id DDD76241698 for ; Sat, 7 Jun 2025 06:40:17 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=148.163.139.74 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1749278419; cv=none; b=qpfPVRNnA743OTJeqkkt1Jvy2sYKct9GfKXazRovBc9Vc8IUCGnZxlotPtShczR0xqPgNcMdMZ3u8nE/KXFexWUyW049s65cohjAFc/nOdd8nLuGFT8YuAdh+Tf+jzcjUZge7wy0tOQL5ArIvYlf/SOpIy/hNHSwKWy5j0rlFOs= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1749278419; c=relaxed/simple; bh=8xK9cYLtkyMChjhj6zDEoAbzPlavCj2F/8VVwgmH0dc=; h=From:Date:Subject:MIME-Version:Content-Type:Message-Id:References: In-Reply-To:To:Cc; b=T6gCCSHpe++gQQUJbmroPfdBQDqBkY086BqPszqVvR6jTrMgdjMBdYc0m/785COU56phHufCrgjS1yJPSuynmq8YU9YcVVOYGKpo7ujRxdyq6NppL/QQqC/VVz6/CEw3OCtbTkat45x5qEP5+3rYcTeffyWvZ1GMUa9C7+AdcVE= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=columbia.edu; spf=pass smtp.mailfrom=columbia.edu; dkim=pass (2048-bit key) header.d=columbia.edu header.i=@columbia.edu header.b=XlSgTddv; arc=none smtp.client-ip=148.163.139.74 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=columbia.edu Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=columbia.edu Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=columbia.edu header.i=@columbia.edu header.b="XlSgTddv" Received: from pps.filterd (m0167073.ppops.net [127.0.0.1]) by mx0b-00364e01.pphosted.com (8.18.1.2/8.18.1.2) with ESMTP id 5575d5RI006722 for ; Sat, 7 Jun 2025 02:40:11 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=columbia.edu; h=cc : content-transfer-encoding : content-type : date : from : in-reply-to : message-id : mime-version : references : subject : to; s=pps01; bh=AsCYG8CkIIPZbMRBimL2n8U8A9q+7pGknJ/irwioCr4=; b=XlSgTddvA0Q4Gv1nwjggTtrLVvK4UOBDaL6hYkxLNrNc2JfxL9xK13KJUOsfdizihaMS hvePSAfJ+DdfkwHZiXPnDvFHEQTJs+RHfSMYXtd34fh7kaUlxZ52WQAxOOfsKFGuNEcP i86YKjNQPklHiIgXAhpTC1R37IAXG7GI0BW5AtfPwlgoHcHaKjJfcbIHvdKH1kmsc07l wYzoeetg2LRvLaMTqf2b9HSay/h04s1Kct6tLAxXRyPkri+/4l/uo1ZUObxOMXkLiXuc vtPP6zvtRvdHTKQiuwiXfq5uIVaHUahA8U0WFFB02PZcdnI9AyBVPhNpTwaZiUF9hLnZ yw== Received: from mail-qk1-f199.google.com (mail-qk1-f199.google.com [209.85.222.199]) by mx0b-00364e01.pphosted.com (PPS) with ESMTPS id 473ennw0dq-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128 verify=NOT) for ; Sat, 07 Jun 2025 02:40:10 -0400 Received: by mail-qk1-f199.google.com with SMTP id af79cd13be357-7c793d573b2so521462785a.1 for ; Fri, 06 Jun 2025 23:40:10 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1749278410; x=1749883210; h=cc:to:in-reply-to:references:message-id:content-transfer-encoding :mime-version:subject:date:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=AsCYG8CkIIPZbMRBimL2n8U8A9q+7pGknJ/irwioCr4=; b=Lmkdl+kGL2ij5RG1gq9K+ro2ig9aGbIa1FceKne22Uw2MxO+wjO7UweEMMqCWUCGJ9 YlYnVLcTTi5WMZo4HK6gmazAKAsu/AI0TBZw2Z0Y6+RQ03hVNoZ7VEZCekF2eQVbBSeE FzbJO7ICy8k7tuCEvKq+F+TzcymERI/y87SOZC0NgxY80incZvYiqPYDdc1bF6iUVrX+ YUcnEZ4RPAOTKgo4P2rBRVc3XduKnP0lvEGPvH/fgW8FyhAQCsfu+ybBC1W8y37lm7hO f0o4KRzFR7JoOWb0HdC9sN9iPxisBjkc8QWuPcq2FnUvM2Ec6rgXDMLWv44ROL1raA0g fWMg== X-Forwarded-Encrypted: i=1; AJvYcCU+wdNqkRSJ4Pp8hMBo2/XugoEdBcXPcsYvT9scaAVakIifPxg7/RglA0JYd+WfblNlyWEJEayrbCFB5/Q=@vger.kernel.org X-Gm-Message-State: AOJu0YyZHnwn3EB8XMn+gJDcq3N3NMaFgH86QrE4g4OfUXvg1BtpgE6F ciQaBYLT/sxwPTqjrXgVTSMKiRIQpWPh378YoDZj2JcFh6g0pwVbCcI+m9J8z+28rvWY8CID8SM v4hYP0e5s1tlDOvQA47p8mAzw+ETuFuUFPffPAU3od2FvZfh2ggY2yya011CJxcYSPHXxug== X-Gm-Gg: ASbGncuuU3C5z76egDAZhjripkLyysn7AmOFIA8LlH9H3f3SxCviRykmTyl/RxZsoBu dRd+brv4ra/+rL4OkareQ+oOYx5+Q7V6ox4XyJ6wsdy8nekjcCIIQgfROIi9Dv2JozQhsHtsmFb TQ5k07SOw+x6IOj+lcpn2z7w1CkYD9vIieyciqF0XOjdr4veGXLSLdgKh5qC9/o6b7qptN7+V2v PcwLO5IqCbhH5MW4SNsvBBKhCRELsGwVq5qvfw1At3Alpbl1P2nvoqcqa4IQ7IwNeIHABErIKzS 5dXk89YPbnRIkIpCA2rrvr/v9Faye3JwHXcthadOvIiQ9Lh0gCgqkUqAEw== X-Received: by 2002:a05:620a:4406:b0:7c5:4711:dc56 with SMTP id af79cd13be357-7d2298fb49amr1059098785a.48.1749278409885; Fri, 06 Jun 2025 23:40:09 -0700 (PDT) X-Google-Smtp-Source: AGHT+IF2auxeh6j8sDgORgohwuvlEu1znKeOCk52yB8TwnyaKsXohV1V87jQgHvUs+w5vQE4vIOb1g== X-Received: by 2002:a05:620a:4406:b0:7c5:4711:dc56 with SMTP id af79cd13be357-7d2298fb49amr1059095785a.48.1749278409369; Fri, 06 Jun 2025 23:40:09 -0700 (PDT) Received: from [127.0.1.1] (dyn-160-39-33-242.dyn.columbia.edu. [160.39.33.242]) by smtp.gmail.com with ESMTPSA id 6a1803df08f44-6fb09ac95e5sm24461256d6.43.2025.06.06.23.40.08 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 06 Jun 2025 23:40:08 -0700 (PDT) From: Tal Zussman Date: Sat, 07 Jun 2025 02:40:01 -0400 Subject: [PATCH v2 2/4] userfaultfd: remove (VM_)BUG_ON()s Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable Message-Id: <20250607-uffd-fixes-v2-2-339dafe9a2fe@columbia.edu> References: <20250607-uffd-fixes-v2-0-339dafe9a2fe@columbia.edu> In-Reply-To: <20250607-uffd-fixes-v2-0-339dafe9a2fe@columbia.edu> To: Andrew Morton , Peter Xu , "Jason A. Donenfeld" , David Hildenbrand , Alexander Viro , Christian Brauner , Jan Kara , Andrea Arcangeli Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org, Tal Zussman X-Mailer: b4 0.14.3-dev-d7477 X-Developer-Signature: v=1; a=ed25519-sha256; t=1749278406; l=12862; i=tz2294@columbia.edu; s=20250528; h=from:subject:message-id; bh=8xK9cYLtkyMChjhj6zDEoAbzPlavCj2F/8VVwgmH0dc=; b=S+N7HEkuX1EXoPaYFwQEbrEidF/LHDXVIh0PjyFqrHyCQ7+XyLIIN+W4PRgPVwS7zx5ao/6Nq 2qTTxiB7gZCC0wWB3XOuTAoBcOGOEbzdJ6UnChfrt2PylWcXxKiDuIX X-Developer-Key: i=tz2294@columbia.edu; a=ed25519; pk=BIj5KdACscEOyAC0oIkeZqLB3L94fzBnDccEooxeM5Y= X-Proofpoint-ORIG-GUID: EZkspP3HGUpnLjU2f8bv9c0_o5OmAJgc X-Proofpoint-GUID: EZkspP3HGUpnLjU2f8bv9c0_o5OmAJgc X-Proofpoint-Spam-Details-Enc: AW1haW4tMjUwNjA3MDA0NyBTYWx0ZWRfXyLYkhe6dgNOp svP06X5WVmaOLD0OEhcOGVndAV71t53LQDVM51CEYNHgGoRyz77Bb2+2wDuR0GkdGklpLBrz4L9 gP99cyzY+/aAxXVyhJ+O4lAF6PispcrgPCzpoUbC6jXPQorYTOodkt303GT6DwcJ/lPqam3WDn1 sEqY/TY/U+Ex/nSL8o4qfQ+y3M2FY41VG1OLo/fgHcCl7MivYQ1sfGgR3smcxL01Lx+/Yve5fPc cJ3ZOoqGiWEoo96Tkw/srr6UzVMQNTLtXaHUIK1WryD3GN9cdaVzS9wIi9efkkVpR8x27waHAui Ynwb74Rg35Vdy2hcTNfGHDbsvOWvjwv0E3ILeZ8bYBhwXwCp/9lpLQTH8eRRdDVHmpA3MQHik8X XyWS+0S0 X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.293,Aquarius:18.0.1099,Hydra:6.0.736,FMLib:17.12.80.40 definitions=2025-06-07_03,2025-06-05_01,2025-03-28_01 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 malwarescore=0 suspectscore=0 bulkscore=10 phishscore=0 priorityscore=1501 mlxscore=0 impostorscore=0 mlxlogscore=999 spamscore=0 clxscore=1015 lowpriorityscore=10 adultscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2505160000 definitions=main-2506070047 BUG_ON() is deprecated [1]. Convert all the BUG_ON()s and VM_BUG_ON()s to use VM_WARN_ON_ONCE(). While at it, also convert the WARN_ON_ONCE()s in move_pages() to use VM_WARN_ON_ONCE(), as the relevant conditions are already checked in validate_range() in move_pages()'s caller. [1] https://www.kernel.org/doc/html/v6.15/process/coding-style.html#use-war= n-rather-than-bug Signed-off-by: Tal Zussman --- fs/userfaultfd.c | 59 +++++++++++++++++++++++++------------------------- mm/userfaultfd.c | 66 +++++++++++++++++++++++++++-------------------------= ---- 2 files changed, 61 insertions(+), 64 deletions(-) diff --git a/fs/userfaultfd.c b/fs/userfaultfd.c index 22f4bf956ba1..80c95c712266 100644 --- a/fs/userfaultfd.c +++ b/fs/userfaultfd.c @@ -165,14 +165,14 @@ static void userfaultfd_ctx_get(struct userfaultfd_ct= x *ctx) static void userfaultfd_ctx_put(struct userfaultfd_ctx *ctx) { if (refcount_dec_and_test(&ctx->refcount)) { - VM_BUG_ON(spin_is_locked(&ctx->fault_pending_wqh.lock)); - VM_BUG_ON(waitqueue_active(&ctx->fault_pending_wqh)); - VM_BUG_ON(spin_is_locked(&ctx->fault_wqh.lock)); - VM_BUG_ON(waitqueue_active(&ctx->fault_wqh)); - VM_BUG_ON(spin_is_locked(&ctx->event_wqh.lock)); - VM_BUG_ON(waitqueue_active(&ctx->event_wqh)); - VM_BUG_ON(spin_is_locked(&ctx->fd_wqh.lock)); - VM_BUG_ON(waitqueue_active(&ctx->fd_wqh)); + VM_WARN_ON_ONCE(spin_is_locked(&ctx->fault_pending_wqh.lock)); + VM_WARN_ON_ONCE(waitqueue_active(&ctx->fault_pending_wqh)); + VM_WARN_ON_ONCE(spin_is_locked(&ctx->fault_wqh.lock)); + VM_WARN_ON_ONCE(waitqueue_active(&ctx->fault_wqh)); + VM_WARN_ON_ONCE(spin_is_locked(&ctx->event_wqh.lock)); + VM_WARN_ON_ONCE(waitqueue_active(&ctx->event_wqh)); + VM_WARN_ON_ONCE(spin_is_locked(&ctx->fd_wqh.lock)); + VM_WARN_ON_ONCE(waitqueue_active(&ctx->fd_wqh)); mmdrop(ctx->mm); kmem_cache_free(userfaultfd_ctx_cachep, ctx); } @@ -383,12 +383,12 @@ vm_fault_t handle_userfault(struct vm_fault *vmf, uns= igned long reason) if (!ctx) goto out; =20 - BUG_ON(ctx->mm !=3D mm); + VM_WARN_ON_ONCE(ctx->mm !=3D mm); =20 /* Any unrecognized flag is a bug. */ - VM_BUG_ON(reason & ~__VM_UFFD_FLAGS); + VM_WARN_ON_ONCE(reason & ~__VM_UFFD_FLAGS); /* 0 or > 1 flags set is a bug; we expect exactly 1. */ - VM_BUG_ON(!reason || (reason & (reason - 1))); + VM_WARN_ON_ONCE(!reason || (reason & (reason - 1))); =20 if (ctx->features & UFFD_FEATURE_SIGBUS) goto out; @@ -411,12 +411,11 @@ vm_fault_t handle_userfault(struct vm_fault *vmf, uns= igned long reason) * to be sure not to return SIGBUS erroneously on * nowait invocations. */ - BUG_ON(vmf->flags & FAULT_FLAG_RETRY_NOWAIT); + VM_WARN_ON_ONCE(vmf->flags & FAULT_FLAG_RETRY_NOWAIT); #ifdef CONFIG_DEBUG_VM if (printk_ratelimit()) { - printk(KERN_WARNING - "FAULT_FLAG_ALLOW_RETRY missing %x\n", - vmf->flags); + pr_warn("FAULT_FLAG_ALLOW_RETRY missing %x\n", + vmf->flags); dump_stack(); } #endif @@ -602,7 +601,7 @@ static void userfaultfd_event_wait_completion(struct us= erfaultfd_ctx *ctx, */ out: atomic_dec(&ctx->mmap_changing); - VM_BUG_ON(atomic_read(&ctx->mmap_changing) < 0); + VM_WARN_ON_ONCE(atomic_read(&ctx->mmap_changing) < 0); userfaultfd_ctx_put(ctx); } =20 @@ -710,7 +709,7 @@ void dup_userfaultfd_fail(struct list_head *fcs) struct userfaultfd_ctx *ctx =3D fctx->new; =20 atomic_dec(&octx->mmap_changing); - VM_BUG_ON(atomic_read(&octx->mmap_changing) < 0); + VM_WARN_ON_ONCE(atomic_read(&octx->mmap_changing) < 0); userfaultfd_ctx_put(octx); userfaultfd_ctx_put(ctx); =20 @@ -1317,8 +1316,8 @@ static int userfaultfd_register(struct userfaultfd_ct= x *ctx, do { cond_resched(); =20 - BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^ - !!(cur->vm_flags & __VM_UFFD_FLAGS)); + VM_WARN_ON_ONCE(!!cur->vm_userfaultfd_ctx.ctx ^ + !!(cur->vm_flags & __VM_UFFD_FLAGS)); =20 /* check not compatible vmas */ ret =3D -EINVAL; @@ -1372,7 +1371,7 @@ static int userfaultfd_register(struct userfaultfd_ct= x *ctx, =20 found =3D true; } for_each_vma_range(vmi, cur, end); - BUG_ON(!found); + VM_WARN_ON_ONCE(!found); =20 ret =3D userfaultfd_register_range(ctx, vma, vm_flags, start, end, wp_async); @@ -1464,8 +1463,8 @@ static int userfaultfd_unregister(struct userfaultfd_= ctx *ctx, do { cond_resched(); =20 - BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^ - !!(cur->vm_flags & __VM_UFFD_FLAGS)); + VM_WARN_ON_ONCE(!!cur->vm_userfaultfd_ctx.ctx ^ + !!(cur->vm_flags & __VM_UFFD_FLAGS)); =20 /* * Check not compatible vmas, not strictly required @@ -1479,7 +1478,7 @@ static int userfaultfd_unregister(struct userfaultfd_= ctx *ctx, =20 found =3D true; } for_each_vma_range(vmi, cur, end); - BUG_ON(!found); + VM_WARN_ON_ONCE(!found); =20 vma_iter_set(&vmi, start); prev =3D vma_prev(&vmi); @@ -1490,7 +1489,7 @@ static int userfaultfd_unregister(struct userfaultfd_= ctx *ctx, for_each_vma_range(vmi, vma, end) { cond_resched(); =20 - BUG_ON(!vma_can_userfault(vma, vma->vm_flags, wp_async)); + VM_WARN_ON_ONCE(!vma_can_userfault(vma, vma->vm_flags, wp_async)); =20 /* * Nothing to do: this vma is already registered into this @@ -1564,7 +1563,7 @@ static int userfaultfd_wake(struct userfaultfd_ctx *c= tx, * len =3D=3D 0 means wake all and we don't want to wake all here, * so check it again to be sure. */ - VM_BUG_ON(!range.len); + VM_WARN_ON_ONCE(!range.len); =20 wake_userfault(ctx, &range); ret =3D 0; @@ -1621,7 +1620,7 @@ static int userfaultfd_copy(struct userfaultfd_ctx *c= tx, return -EFAULT; if (ret < 0) goto out; - BUG_ON(!ret); + VM_WARN_ON_ONCE(!ret); /* len =3D=3D 0 would wake all */ range.len =3D ret; if (!(uffdio_copy.mode & UFFDIO_COPY_MODE_DONTWAKE)) { @@ -1676,7 +1675,7 @@ static int userfaultfd_zeropage(struct userfaultfd_ct= x *ctx, if (ret < 0) goto out; /* len =3D=3D 0 would wake all */ - BUG_ON(!ret); + VM_WARN_ON_ONCE(!ret); range.len =3D ret; if (!(uffdio_zeropage.mode & UFFDIO_ZEROPAGE_MODE_DONTWAKE)) { range.start =3D uffdio_zeropage.range.start; @@ -1788,7 +1787,7 @@ static int userfaultfd_continue(struct userfaultfd_ct= x *ctx, unsigned long arg) goto out; =20 /* len =3D=3D 0 would wake all */ - BUG_ON(!ret); + VM_WARN_ON_ONCE(!ret); range.len =3D ret; if (!(uffdio_continue.mode & UFFDIO_CONTINUE_MODE_DONTWAKE)) { range.start =3D uffdio_continue.range.start; @@ -1845,7 +1844,7 @@ static inline int userfaultfd_poison(struct userfault= fd_ctx *ctx, unsigned long goto out; =20 /* len =3D=3D 0 would wake all */ - BUG_ON(!ret); + VM_WARN_ON_ONCE(!ret); range.len =3D ret; if (!(uffdio_poison.mode & UFFDIO_POISON_MODE_DONTWAKE)) { range.start =3D uffdio_poison.range.start; @@ -2106,7 +2105,7 @@ static int new_userfaultfd(int flags) struct file *file; int fd; =20 - BUG_ON(!current->mm); + VM_WARN_ON_ONCE(!current->mm); =20 /* Check the UFFD_* constants for consistency. */ BUILD_BUG_ON(UFFD_USER_MODE_ONLY & UFFD_SHARED_FCNTL_FLAGS); diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c index bc473ad21202..41e67ded5a6e 100644 --- a/mm/userfaultfd.c +++ b/mm/userfaultfd.c @@ -561,7 +561,7 @@ static __always_inline ssize_t mfill_atomic_hugetlb( } =20 while (src_addr < src_start + len) { - BUG_ON(dst_addr >=3D dst_start + len); + VM_WARN_ON_ONCE(dst_addr >=3D dst_start + len); =20 /* * Serialize via vma_lock and hugetlb_fault_mutex. @@ -602,7 +602,7 @@ static __always_inline ssize_t mfill_atomic_hugetlb( if (unlikely(err =3D=3D -ENOENT)) { up_read(&ctx->map_changing_lock); uffd_mfill_unlock(dst_vma); - BUG_ON(!folio); + VM_WARN_ON_ONCE(!folio); =20 err =3D copy_folio_from_user(folio, (const void __user *)src_addr, true); @@ -614,7 +614,7 @@ static __always_inline ssize_t mfill_atomic_hugetlb( dst_vma =3D NULL; goto retry; } else - BUG_ON(folio); + VM_WARN_ON_ONCE(folio); =20 if (!err) { dst_addr +=3D vma_hpagesize; @@ -635,9 +635,9 @@ static __always_inline ssize_t mfill_atomic_hugetlb( out: if (folio) folio_put(folio); - BUG_ON(copied < 0); - BUG_ON(err > 0); - BUG_ON(!copied && !err); + VM_WARN_ON_ONCE(copied < 0); + VM_WARN_ON_ONCE(err > 0); + VM_WARN_ON_ONCE(!copied && !err); return copied ? copied : err; } #else /* !CONFIG_HUGETLB_PAGE */ @@ -711,12 +711,12 @@ static __always_inline ssize_t mfill_atomic(struct us= erfaultfd_ctx *ctx, /* * Sanitize the command parameters: */ - BUG_ON(dst_start & ~PAGE_MASK); - BUG_ON(len & ~PAGE_MASK); + VM_WARN_ON_ONCE(dst_start & ~PAGE_MASK); + VM_WARN_ON_ONCE(len & ~PAGE_MASK); =20 /* Does the address range wrap, or is the span zero-sized? */ - BUG_ON(src_start + len <=3D src_start); - BUG_ON(dst_start + len <=3D dst_start); + VM_WARN_ON_ONCE(src_start + len <=3D src_start); + VM_WARN_ON_ONCE(dst_start + len <=3D dst_start); =20 src_addr =3D src_start; dst_addr =3D dst_start; @@ -775,7 +775,7 @@ static __always_inline ssize_t mfill_atomic(struct user= faultfd_ctx *ctx, while (src_addr < src_start + len) { pmd_t dst_pmdval; =20 - BUG_ON(dst_addr >=3D dst_start + len); + VM_WARN_ON_ONCE(dst_addr >=3D dst_start + len); =20 dst_pmd =3D mm_alloc_pmd(dst_mm, dst_addr); if (unlikely(!dst_pmd)) { @@ -818,7 +818,7 @@ static __always_inline ssize_t mfill_atomic(struct user= faultfd_ctx *ctx, =20 up_read(&ctx->map_changing_lock); uffd_mfill_unlock(dst_vma); - BUG_ON(!folio); + VM_WARN_ON_ONCE(!folio); =20 kaddr =3D kmap_local_folio(folio, 0); err =3D copy_from_user(kaddr, @@ -832,7 +832,7 @@ static __always_inline ssize_t mfill_atomic(struct user= faultfd_ctx *ctx, flush_dcache_folio(folio); goto retry; } else - BUG_ON(folio); + VM_WARN_ON_ONCE(folio); =20 if (!err) { dst_addr +=3D PAGE_SIZE; @@ -852,9 +852,9 @@ static __always_inline ssize_t mfill_atomic(struct user= faultfd_ctx *ctx, out: if (folio) folio_put(folio); - BUG_ON(copied < 0); - BUG_ON(err > 0); - BUG_ON(!copied && !err); + VM_WARN_ON_ONCE(copied < 0); + VM_WARN_ON_ONCE(err > 0); + VM_WARN_ON_ONCE(!copied && !err); return copied ? copied : err; } =20 @@ -940,11 +940,11 @@ int mwriteprotect_range(struct userfaultfd_ctx *ctx, = unsigned long start, /* * Sanitize the command parameters: */ - BUG_ON(start & ~PAGE_MASK); - BUG_ON(len & ~PAGE_MASK); + VM_WARN_ON_ONCE(start & ~PAGE_MASK); + VM_WARN_ON_ONCE(len & ~PAGE_MASK); =20 /* Does the address range wrap, or is the span zero-sized? */ - BUG_ON(start + len <=3D start); + VM_WARN_ON_ONCE(start + len <=3D start); =20 mmap_read_lock(dst_mm); =20 @@ -1709,15 +1709,13 @@ ssize_t move_pages(struct userfaultfd_ctx *ctx, uns= igned long dst_start, ssize_t moved =3D 0; =20 /* Sanitize the command parameters. */ - if (WARN_ON_ONCE(src_start & ~PAGE_MASK) || - WARN_ON_ONCE(dst_start & ~PAGE_MASK) || - WARN_ON_ONCE(len & ~PAGE_MASK)) - goto out; + VM_WARN_ON_ONCE(src_start & ~PAGE_MASK); + VM_WARN_ON_ONCE(dst_start & ~PAGE_MASK); + VM_WARN_ON_ONCE(len & ~PAGE_MASK); =20 /* Does the address range wrap, or is the span zero-sized? */ - if (WARN_ON_ONCE(src_start + len <=3D src_start) || - WARN_ON_ONCE(dst_start + len <=3D dst_start)) - goto out; + VM_WARN_ON_ONCE(src_start + len < src_start); + VM_WARN_ON_ONCE(dst_start + len < dst_start); =20 err =3D uffd_move_lock(mm, dst_start, src_start, &dst_vma, &src_vma); if (err) @@ -1867,9 +1865,9 @@ ssize_t move_pages(struct userfaultfd_ctx *ctx, unsig= ned long dst_start, up_read(&ctx->map_changing_lock); uffd_move_unlock(dst_vma, src_vma); out: - VM_WARN_ON(moved < 0); - VM_WARN_ON(err > 0); - VM_WARN_ON(!moved && !err); + VM_WARN_ON_ONCE(moved < 0); + VM_WARN_ON_ONCE(err > 0); + VM_WARN_ON_ONCE(!moved && !err); return moved ? moved : err; } =20 @@ -1956,9 +1954,9 @@ int userfaultfd_register_range(struct userfaultfd_ctx= *ctx, for_each_vma_range(vmi, vma, end) { cond_resched(); =20 - BUG_ON(!vma_can_userfault(vma, vm_flags, wp_async)); - BUG_ON(vma->vm_userfaultfd_ctx.ctx && - vma->vm_userfaultfd_ctx.ctx !=3D ctx); + VM_WARN_ON_ONCE(!vma_can_userfault(vma, vm_flags, wp_async)); + VM_WARN_ON_ONCE(vma->vm_userfaultfd_ctx.ctx && + vma->vm_userfaultfd_ctx.ctx !=3D ctx); WARN_ON(!(vma->vm_flags & VM_MAYWRITE)); =20 /* @@ -2035,8 +2033,8 @@ void userfaultfd_release_all(struct mm_struct *mm, prev =3D NULL; for_each_vma(vmi, vma) { cond_resched(); - BUG_ON(!!vma->vm_userfaultfd_ctx.ctx ^ - !!(vma->vm_flags & __VM_UFFD_FLAGS)); + VM_WARN_ON_ONCE(!!vma->vm_userfaultfd_ctx.ctx ^ + !!(vma->vm_flags & __VM_UFFD_FLAGS)); if (vma->vm_userfaultfd_ctx.ctx !=3D ctx) { prev =3D vma; continue; --=20 2.39.5