From nobody Mon Nov 25 00:29:21 2024 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 93C691CB339; Thu, 31 Oct 2024 19:25:59 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1730402759; cv=none; b=ApUF4d4I15tUdtwCBdBGXqLtb5McL12wtnbBw6DQejn43jFhtriq7uENyy7QB9VvXThg7hXJu4zeYQ7UypuExLRjqJiPxGuYMo2XCykOH6VHWoQd9oRG1KuQBrvF1wjnfO+PaOq43CTkffD3RolExCyB7+Mv1KKEU9xzVUIWhCg= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1730402759; c=relaxed/simple; bh=XmYalT4R87/DWvbsQyVmrZIIZ6DF/I/kJjYXRf7eYV4=; h=From:Date:Subject:MIME-Version:Content-Type:Message-Id:References: In-Reply-To:To:Cc; b=W7MHWS2mKCWb67rYrtUMO8TNWtRpZxxUHxu1dmQzUNH8+LOe8EeDP9tUWZkEjh6PxqgmN/tF4peheFcevtrQWY1HpYYXp0sGdGSDGwPXHFcyUcaCkzEISQdvIyF4eU9oHqQz4BtEiE8bj0frIOkhsCEvOF25TAelfyYoMier/1o= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=HMMpqSpv; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="HMMpqSpv" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 8B381C4CED1; Thu, 31 Oct 2024 19:25:53 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1730402759; bh=XmYalT4R87/DWvbsQyVmrZIIZ6DF/I/kJjYXRf7eYV4=; h=From:Date:Subject:References:In-Reply-To:To:Cc:From; b=HMMpqSpvapCFd/w8UTexjaywcFou5EJxB/4+yUX7rHyOyiD0ji6BEIscDy66WM4xE BbZmj7htwglb4XX4VjxWG9zIDHAWa5mYWRkH5GI2oaoChOEJYj7YlBFL//pSXiFw8b kMV4646w1yz7PrWygLCZ9mf7I+LehMKpqLiR8c0f9kegcyck1rcq2z214dabX3kE53 O+BT1gG2tlop4pQ7R7YEQW8OqWF9ZMxWwdk7tOuVoO6g/8WsfOIvoXtvzVypWzUM3R PTBeB/e78u+6yZjPBXtEM6MKxdLHBssFEn26aRwNmYArJ5aDvEH/LFna+z4rBlX8u3 qBb8NnfgkxgjQ== From: Mark Brown Date: Thu, 31 Oct 2024 19:25:05 +0000 Subject: [PATCH RFT v12 4/8] fork: Add shadow stack support to clone3() Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable Message-Id: <20241031-clone3-shadow-stack-v12-4-7183eb8bee17@kernel.org> References: <20241031-clone3-shadow-stack-v12-0-7183eb8bee17@kernel.org> In-Reply-To: <20241031-clone3-shadow-stack-v12-0-7183eb8bee17@kernel.org> To: "Rick P. Edgecombe" , Deepak Gupta , Szabolcs Nagy , "H.J. Lu" , Florian Weimer , Thomas Gleixner , Ingo Molnar , Borislav Petkov , Dave Hansen , x86@kernel.org, "H. Peter Anvin" , Peter Zijlstra , Juri Lelli , Vincent Guittot , Dietmar Eggemann , Steven Rostedt , Ben Segall , Mel Gorman , Valentin Schneider , Christian Brauner , Shuah Khan Cc: linux-kernel@vger.kernel.org, Catalin Marinas , Will Deacon , jannh@google.com, bsegall@google.com, Yury Khrustalev , Wilco Dijkstra , linux-kselftest@vger.kernel.org, linux-api@vger.kernel.org, Mark Brown , Kees Cook X-Mailer: b4 0.15-dev-9b746 X-Developer-Signature: v=1; a=openpgp-sha256; l=16778; i=broonie@kernel.org; h=from:subject:message-id; bh=XmYalT4R87/DWvbsQyVmrZIIZ6DF/I/kJjYXRf7eYV4=; b=owEBbQGS/pANAwAKASTWi3JdVIfQAcsmYgBnI9mkhUnofvERgRI+2bZ88hJoQscyOeKvRWUlO057 5kdOrL2JATMEAAEKAB0WIQSt5miqZ1cYtZ/in+ok1otyXVSH0AUCZyPZpAAKCRAk1otyXVSH0LV1B/ 9m1vUo5p+Ns2IeHMnXVQo6OMKuuNNIroF/30F3ByWlV1gLG7i86EJ6dKgCHv9sGvFY00bbLBc6Ik1/ UPFvQawNbTzmY1SfjHozzbLHVGyLkkJil/MEWwu4IUzEjr9Fov5i3q8lhHUG8nqS7xXxAGyCyDREkO zxm2b5uqj2kIuCRu1w+IQon002wNA/vQA/XcrnVzagQtBh1ToTIm0Lc+2ySoUQT5TxatZ0Kyu6cIna jDYfT3/A9Pnf/9zScM/dpvamfEERusFmz1Ar/GVbGwuukGbxJrlGhGZWefNfF5RL4KXW+zYf0wnXMn +xImhgFuMIAfMFxYttIo0Qg6wupPal X-Developer-Key: i=broonie@kernel.org; a=openpgp; fpr=3F2568AAC26998F9E813A1C5C3F436CA30F5D8EB Unlike with the normal stack there is no API for configuring the the shadow stack for a new thread, instead the kernel will dynamically allocate a new shadow stack with the same size as the normal stack. This appears to be due to the shadow stack series having been in development since before the more extensible clone3() was added rather than anything more deliberate. Add a paramter to clone3() specifying the shadow stack pointer to use for the new thread, this is inconsistent with the way we specify the normal stack but during review concerns were expressed about having to identify where the shadow stack pointer should be placed especially in cases where the shadow stack has been previously active. If no shadow stack is specified then the existing implicit allocation behaviour is maintained. If a shadow stack pointer is specified then it is required to have an architecture defined token placed on the stack, this will be consumed by the new task. If no valid token is present then this will be reported with -EINVAL. This token prevents new threads being created pointing at the shadow stack of an existing running thread. If the architecture does not support shadow stacks the shadow stack pointer must be not be specified, architectures that do support the feature are expected to enforce the same requirement on individual systems that lack shadow stack support. Update the existing arm64 and x86 implementations to pay attention to the newly added arguments, in order to maintain compatibility we use the existing behaviour if no shadow stack is specified. Since we are now using more fields from the kernel_clone_args we pass that into the shadow stack code rather than individual fields. Portions of the x86 architecture code were written by Rick Edgecombe. Signed-off-by: Mark Brown --- arch/arm64/mm/gcs.c | 54 +++++++++++++++++++++- arch/x86/include/asm/shstk.h | 11 +++-- arch/x86/kernel/process.c | 2 +- arch/x86/kernel/shstk.c | 57 +++++++++++++++++++++--- include/asm-generic/cacheflush.h | 11 +++++ include/linux/sched/task.h | 17 +++++++ include/uapi/linux/sched.h | 10 +++-- kernel/fork.c | 96 +++++++++++++++++++++++++++++++++++-= ---- 8 files changed, 232 insertions(+), 26 deletions(-) diff --git a/arch/arm64/mm/gcs.c b/arch/arm64/mm/gcs.c index 1f633a482558b59aac5427963d42b37fce08c8a6..c4e93b7ce05c5dfa1128923ad58= 7f9b5a7fb0051 100644 --- a/arch/arm64/mm/gcs.c +++ b/arch/arm64/mm/gcs.c @@ -43,8 +43,24 @@ int gcs_alloc_thread_stack(struct task_struct *tsk, { unsigned long addr, size; =20 - if (!system_supports_gcs()) + if (!system_supports_gcs()) { + if (args->shadow_stack_pointer) + return -EINVAL; + + return 0; + } + + /* + * If the user specified a GCS then use it, otherwise fall + * back to a default allocation strategy. Validation is done + * in arch_shstk_validate_clone(). + */ + if (args->shadow_stack_pointer) { + tsk->thread.gcs_base =3D 0; + tsk->thread.gcs_size =3D 0; + tsk->thread.gcspr_el0 =3D args->shadow_stack_pointer; return 0; + } =20 if (!task_gcs_el0_enabled(tsk)) return 0; @@ -68,6 +84,42 @@ int gcs_alloc_thread_stack(struct task_struct *tsk, return 0; } =20 +static bool gcs_consume_token(struct vm_area_struct *vma, struct page *pag= e, + unsigned long user_addr) +{ + u64 expected =3D GCS_CAP(user_addr); + u64 *token =3D page_address(page) + offset_in_page(user_addr); + + if (!cmpxchg_to_user_page(vma, page, user_addr, token, expected, 0)) + return false; + set_page_dirty_lock(page); + + return true; +} + +int arch_shstk_validate_clone(struct task_struct *tsk, + struct vm_area_struct *vma, + struct page *page, + struct kernel_clone_args *args) +{ + unsigned long gcspr_el0; + int ret =3D 0; + + /* Ensure that a token written as a result of a pivot is visible */ + gcsb_dsync(); + + gcspr_el0 =3D args->shadow_stack_pointer; + if (!gcs_consume_token(vma, page, gcspr_el0)) + return -EINVAL; + + tsk->thread.gcspr_el0 =3D gcspr_el0 + sizeof(u64); + + /* Ensure that our token consumption visible */ + gcsb_dsync(); + + return ret; +} + SYSCALL_DEFINE3(map_shadow_stack, unsigned long, addr, unsigned long, size= , unsigned int, flags) { unsigned long alloc_size; diff --git a/arch/x86/include/asm/shstk.h b/arch/x86/include/asm/shstk.h index 4cb77e004615dff003426a2eb594460ca1015f4e..252feeda69991e939942c74556e= 23e27c835e766 100644 --- a/arch/x86/include/asm/shstk.h +++ b/arch/x86/include/asm/shstk.h @@ -6,6 +6,7 @@ #include =20 struct task_struct; +struct kernel_clone_args; struct ksignal; =20 #ifdef CONFIG_X86_USER_SHADOW_STACK @@ -16,8 +17,8 @@ struct thread_shstk { =20 long shstk_prctl(struct task_struct *task, int option, unsigned long arg2); void reset_thread_features(void); -unsigned long shstk_alloc_thread_stack(struct task_struct *p, unsigned lon= g clone_flags, - unsigned long stack_size); +unsigned long shstk_alloc_thread_stack(struct task_struct *p, + const struct kernel_clone_args *args); void shstk_free(struct task_struct *p); int setup_signal_shadow_stack(struct ksignal *ksig); int restore_signal_shadow_stack(void); @@ -28,8 +29,10 @@ static inline long shstk_prctl(struct task_struct *task,= int option, unsigned long arg2) { return -EINVAL; } static inline void reset_thread_features(void) {} static inline unsigned long shstk_alloc_thread_stack(struct task_struct *p, - unsigned long clone_flags, - unsigned long stack_size) { return 0; } + const struct kernel_clone_args *args) +{ + return 0; +} static inline void shstk_free(struct task_struct *p) {} static inline int setup_signal_shadow_stack(struct ksignal *ksig) { return= 0; } static inline int restore_signal_shadow_stack(void) { return 0; } diff --git a/arch/x86/kernel/process.c b/arch/x86/kernel/process.c index f63f8fd00a91f3d1171f307b92179556ba2d716d..59456ab8d93faee29c3b223b64e= b41659df76032 100644 --- a/arch/x86/kernel/process.c +++ b/arch/x86/kernel/process.c @@ -207,7 +207,7 @@ int copy_thread(struct task_struct *p, const struct ker= nel_clone_args *args) * is disabled, new_ssp will remain 0, and fpu_clone() will know not to * update it. */ - new_ssp =3D shstk_alloc_thread_stack(p, clone_flags, args->stack_size); + new_ssp =3D shstk_alloc_thread_stack(p, args); if (IS_ERR_VALUE(new_ssp)) return PTR_ERR((void *)new_ssp); =20 diff --git a/arch/x86/kernel/shstk.c b/arch/x86/kernel/shstk.c index 059685612362d7b1865eabf400888fbfa0659c1e..056e2c9ec30531d0901297da07f= 1842b47d2fcd5 100644 --- a/arch/x86/kernel/shstk.c +++ b/arch/x86/kernel/shstk.c @@ -191,18 +191,65 @@ void reset_thread_features(void) current->thread.features_locked =3D 0; } =20 -unsigned long shstk_alloc_thread_stack(struct task_struct *tsk, unsigned l= ong clone_flags, - unsigned long stack_size) +int arch_shstk_validate_clone(struct task_struct *t, + struct vm_area_struct *vma, + struct page *page, + struct kernel_clone_args *args) +{ + /* + * SSP is aligned, so reserved bits and mode bit are a zero, just mark + * the token 64-bit. + */ + void *maddr =3D kmap_local_page(page); + int offset; + unsigned long addr, ssp; + u64 expected; + + if (!features_enabled(ARCH_SHSTK_SHSTK)) + return 0; + + ssp =3D args->shadow_stack_pointer; + addr =3D ssp - SS_FRAME_SIZE; + expected =3D ssp | BIT(0); + offset =3D offset_in_page(addr); + + if (!cmpxchg_to_user_page(vma, page, addr, (unsigned long *)(maddr + offs= et), + expected, 0)) + return -EINVAL; + set_page_dirty_lock(page); + + return 0; +} + +unsigned long shstk_alloc_thread_stack(struct task_struct *tsk, + const struct kernel_clone_args *args) { struct thread_shstk *shstk =3D &tsk->thread.shstk; + unsigned long clone_flags =3D args->flags; unsigned long addr, size; =20 /* * If shadow stack is not enabled on the new thread, skip any - * switch to a new shadow stack. + * implicit switch to a new shadow stack and reject attempts to + * explicitly specify one. */ - if (!features_enabled(ARCH_SHSTK_SHSTK)) + if (!features_enabled(ARCH_SHSTK_SHSTK)) { + if (args->shadow_stack_pointer) + return (unsigned long)ERR_PTR(-EINVAL); + return 0; + } + + /* + * If the user specified a shadow stack then use it, otherwise + * fall back to a default allocation strategy. Validation is + * done in arch_shstk_validate_clone(). + */ + if (args->shadow_stack_pointer) { + shstk->base =3D 0; + shstk->size =3D 0; + return args->shadow_stack_pointer; + } =20 /* * For CLONE_VFORK the child will share the parents shadow stack. @@ -222,7 +269,7 @@ unsigned long shstk_alloc_thread_stack(struct task_stru= ct *tsk, unsigned long cl if (!(clone_flags & CLONE_VM)) return 0; =20 - size =3D adjust_shstk_size(stack_size); + size =3D adjust_shstk_size(args->stack_size); addr =3D alloc_shstk(0, size, 0, false); if (IS_ERR_VALUE(addr)) return addr; diff --git a/include/asm-generic/cacheflush.h b/include/asm-generic/cachefl= ush.h index 7ee8a179d1036e1d8010b8b18a8f3022e41c1695..96cc0c7a5c90fd7e899d0c5fe7c= 706302265efcf 100644 --- a/include/asm-generic/cacheflush.h +++ b/include/asm-generic/cacheflush.h @@ -124,4 +124,15 @@ static inline void flush_cache_vunmap(unsigned long st= art, unsigned long end) } while (0) #endif =20 +#ifndef cmpxchg_to_user_page +#define cmpxchg_to_user_page(vma, page, vaddr, ptr, old, new) \ +({ \ + bool ret; \ + \ + ret =3D try_cmpxchg(ptr, &old, new); \ + flush_icache_user_page(vma, page, vaddr, sizeof(*ptr)); \ + ret; \ +}) +#endif + #endif /* _ASM_GENERIC_CACHEFLUSH_H */ diff --git a/include/linux/sched/task.h b/include/linux/sched/task.h index 0f2aeb37bbb047335a399326b31bc8df81b75a3a..cd36389619d5c97401f7b90e177= c6027c232783b 100644 --- a/include/linux/sched/task.h +++ b/include/linux/sched/task.h @@ -16,6 +16,7 @@ struct task_struct; struct rusage; union thread_union; struct css_set; +struct vm_area_struct; =20 /* All the bits taken by the old clone syscall. */ #define CLONE_LEGACY_FLAGS 0xffffffffULL @@ -43,6 +44,7 @@ struct kernel_clone_args { void *fn_arg; struct cgroup *cgrp; struct css_set *cset; + unsigned long shadow_stack_pointer; }; =20 /* @@ -236,4 +238,19 @@ static inline void task_unlock(struct task_struct *p) =20 DEFINE_GUARD(task_lock, struct task_struct *, task_lock(_T), task_unlock(_= T)) =20 +#ifdef CONFIG_ARCH_HAS_USER_SHADOW_STACK +int arch_shstk_validate_clone(struct task_struct *p, + struct vm_area_struct *vma, + struct page *page, + struct kernel_clone_args *args); +#else +static inline int arch_shstk_validate_clone(struct task_struct *p, + struct vm_area_struct *vma, + struct page *page, + struct kernel_clone_args *args) +{ + return 0; +} +#endif + #endif /* _LINUX_SCHED_TASK_H */ diff --git a/include/uapi/linux/sched.h b/include/uapi/linux/sched.h index 359a14cc76a4038aeacef14b2915d5ce60d0cf44..586a1c05a4e4ca05584d4d50022= 3bcf6c3add54c 100644 --- a/include/uapi/linux/sched.h +++ b/include/uapi/linux/sched.h @@ -84,6 +84,8 @@ * kernel's limit of nested PID namespaces. * @cgroup: If CLONE_INTO_CGROUP is specified set this to * a file descriptor for the cgroup. + * @shadow_stack_pointer: Value to use for shadow stack pointer in the + * child process. * * The structure is versioned by size and thus extensible. * New struct members must go at the end of the struct and @@ -101,12 +103,14 @@ struct clone_args { __aligned_u64 set_tid; __aligned_u64 set_tid_size; __aligned_u64 cgroup; + __aligned_u64 shadow_stack_pointer; }; #endif =20 -#define CLONE_ARGS_SIZE_VER0 64 /* sizeof first published struct */ -#define CLONE_ARGS_SIZE_VER1 80 /* sizeof second published struct */ -#define CLONE_ARGS_SIZE_VER2 88 /* sizeof third published struct */ +#define CLONE_ARGS_SIZE_VER0 64 /* sizeof first published struct */ +#define CLONE_ARGS_SIZE_VER1 80 /* sizeof second published struct */ +#define CLONE_ARGS_SIZE_VER2 88 /* sizeof third published struct */ +#define CLONE_ARGS_SIZE_VER3 96 /* sizeof fourth published struct */ =20 /* * Scheduling policies diff --git a/kernel/fork.c b/kernel/fork.c index 60c0b4868fd4993920f7a615a47f7e915b9a47b5..d925e0c52a5ac4394abd1d46a5e= b14386d7ba2ca 100644 --- a/kernel/fork.c +++ b/kernel/fork.c @@ -2109,6 +2109,51 @@ static void rv_task_fork(struct task_struct *p) #define rv_task_fork(p) do {} while (0) #endif =20 +static int shstk_validate_clone(struct task_struct *p, + struct kernel_clone_args *args) +{ + struct mm_struct *mm; + struct vm_area_struct *vma; + struct page *page; + unsigned long addr; + int ret; + + if (!IS_ENABLED(CONFIG_ARCH_HAS_USER_SHADOW_STACK)) + return 0; + + if (!args->shadow_stack_pointer) + return 0; + + mm =3D get_task_mm(p); + if (!mm) + return -EFAULT; + + mmap_read_lock(mm); + + addr =3D untagged_addr_remote(mm, args->shadow_stack_pointer); + page =3D get_user_page_vma_remote(mm, addr, FOLL_FORCE | FOLL_WRITE, + &vma); + if (IS_ERR(page)) { + ret =3D -EFAULT; + goto out; + } + + if (!(vma->vm_flags & VM_SHADOW_STACK) || + !(vma->vm_flags & VM_WRITE)) { + ret =3D -EFAULT; + goto out_page; + } + + ret =3D arch_shstk_validate_clone(p, vma, page, args); + +out_page: + put_page(page); +out: + mmap_read_unlock(mm); + mmput(mm); + return ret; +} + /* * This creates a new process as a copy of the old one, * but does not actually start it yet. @@ -2382,6 +2427,9 @@ __latent_entropy struct task_struct *copy_process( if (retval) goto bad_fork_cleanup_namespaces; retval =3D copy_thread(p, args); + if (retval) + goto bad_fork_cleanup_io; + retval =3D shstk_validate_clone(p, args); if (retval) goto bad_fork_cleanup_io; =20 @@ -2945,7 +2993,9 @@ noinline static int copy_clone_args_from_user(struct = kernel_clone_args *kargs, CLONE_ARGS_SIZE_VER1); BUILD_BUG_ON(offsetofend(struct clone_args, cgroup) !=3D CLONE_ARGS_SIZE_VER2); - BUILD_BUG_ON(sizeof(struct clone_args) !=3D CLONE_ARGS_SIZE_VER2); + BUILD_BUG_ON(offsetofend(struct clone_args, shadow_stack_pointer) !=3D + CLONE_ARGS_SIZE_VER3); + BUILD_BUG_ON(sizeof(struct clone_args) !=3D CLONE_ARGS_SIZE_VER3); =20 if (unlikely(usize > PAGE_SIZE)) return -E2BIG; @@ -2978,16 +3028,17 @@ noinline static int copy_clone_args_from_user(struc= t kernel_clone_args *kargs, return -EINVAL; =20 *kargs =3D (struct kernel_clone_args){ - .flags =3D args.flags, - .pidfd =3D u64_to_user_ptr(args.pidfd), - .child_tid =3D u64_to_user_ptr(args.child_tid), - .parent_tid =3D u64_to_user_ptr(args.parent_tid), - .exit_signal =3D args.exit_signal, - .stack =3D args.stack, - .stack_size =3D args.stack_size, - .tls =3D args.tls, - .set_tid_size =3D args.set_tid_size, - .cgroup =3D args.cgroup, + .flags =3D args.flags, + .pidfd =3D u64_to_user_ptr(args.pidfd), + .child_tid =3D u64_to_user_ptr(args.child_tid), + .parent_tid =3D u64_to_user_ptr(args.parent_tid), + .exit_signal =3D args.exit_signal, + .stack =3D args.stack, + .stack_size =3D args.stack_size, + .tls =3D args.tls, + .set_tid_size =3D args.set_tid_size, + .cgroup =3D args.cgroup, + .shadow_stack_pointer =3D args.shadow_stack_pointer, }; =20 if (args.set_tid && @@ -3028,6 +3079,27 @@ static inline bool clone3_stack_valid(struct kernel_= clone_args *kargs) return true; } =20 +/** + * clone3_shadow_stack_valid - check and prepare shadow stack + * @kargs: kernel clone args + * + * Verify that shadow stacks are only enabled if supported. + */ +static inline bool clone3_shadow_stack_valid(struct kernel_clone_args *kar= gs) +{ + if (!kargs->shadow_stack_pointer) + return true; + + if (!IS_ALIGNED(kargs->shadow_stack_pointer, sizeof(void *))) + return false; + + /* + * The architecture must check support on the specific + * machine. + */ + return IS_ENABLED(CONFIG_ARCH_HAS_USER_SHADOW_STACK); +} + static bool clone3_args_valid(struct kernel_clone_args *kargs) { /* Verify that no unknown flags are passed along. */ @@ -3050,7 +3122,7 @@ static bool clone3_args_valid(struct kernel_clone_arg= s *kargs) kargs->exit_signal) return false; =20 - if (!clone3_stack_valid(kargs)) + if (!clone3_stack_valid(kargs) || !clone3_shadow_stack_valid(kargs)) return false; =20 return true; --=20 2.39.2