From nobody Mon Dec 1 22:02:17 2025 Received: from mail-ej1-f45.google.com (mail-ej1-f45.google.com [209.85.218.45]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 8D4F130FC2D for ; Mon, 1 Dec 2025 12:42:18 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.218.45 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1764592941; cv=none; b=pf6Jbtad7F6wNKNU/aAO0iomjv1g1IL0aG5B29HWsrRwmVFJsK4J9GG66twnRzdaysspqQ8RfMer7zX1BzXUkCgAJZ4uotwQ/Z639Gs/+PXjAqJDvGDYo2UiX4sBOsl1fGtBd+4LzAwp6NDJXmjvFk3KH9ZrNEIoGg4ZFUjFEIw= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1764592941; c=relaxed/simple; bh=1mcgOHYRO2xHgEprO4OO6ROTvUomqwHwdE7/L7UpgqA=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=g+tcgxD8PapLDxhCqXmJopmdw/53bpVFfo/ylSQ3a2Fq/XiJ0dO2xR/Tbpe4zOS4g6IqrL9ZM8CMK/b+unx369IomARzzeI3Co9D/b3Q4+A1HpDXXcKqKtf4B042N+K5RxqWhPmfQQ26CrydhDBXU8+wZnJbJNsw6insXEWOR6s= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com; spf=pass smtp.mailfrom=gmail.com; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b=c93dizpW; arc=none smtp.client-ip=209.85.218.45 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=gmail.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="c93dizpW" Received: by mail-ej1-f45.google.com with SMTP id a640c23a62f3a-b735b89501fso510454366b.0 for ; Mon, 01 Dec 2025 04:42:18 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1764592937; x=1765197737; darn=vger.kernel.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=uj6RNmz5NRTJ9FELpWwEPrdAIia3oMPoBC/Wu0k17QA=; b=c93dizpWhR9vfsnb5tjUvV8/xeAKlHPd2tBYXEEsBMtKapzoX50srnJRhTs6o6glSd EPwKdqQQIuaG6aHKTqvUEBRgzYEda0Us4o7rgHt5s7cMxmSaYrWdQejWaay4Mr+Rmfo9 0J59YYKReK70IsuTSjP7ajGPGPzU1M9KGX+rs0IR5zTf3VmfHtAa8fBXBfFuN6phIHu8 qoU5tpRXPjPcX+bCz1+l4WLqzU1J2UX3WH1DaXjLjEXaAo/BdNVgLBceoltMugMuzZ8L SCvKNXAWUORPU9GmEgLKo5Xsxl2NAQew2WdMeqh1IWxakos1/t1jMHMqO0CwRNxnTAs0 tTuw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1764592937; x=1765197737; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-gg:x-gm-message-state:from :to:cc:subject:date:message-id:reply-to; bh=uj6RNmz5NRTJ9FELpWwEPrdAIia3oMPoBC/Wu0k17QA=; b=Vma4ACAwyDTwdiHdD4/Z3G2hwnq0dfrGMdxB8czexCUt7VgSrkhDjw32YDOg0QfoVN R4/wAzKhyOZXJkuXBhnLmJiMwvU0TAj7i2SxtVUheBala+piYAGrCHz+EqALRz+Sz4wf ZLCUmPZ08JSipxIZUOYaksLd7g9GVTCLlBYY4IBEkRbwU1iFj4CJkGASG3qDwxldhgRU ulq1GEDWHPSgZw4SKAyUKWLtKvEq2hg92zHOJDwOX6twPawNMZnxPvD2ZIa2zJFt/ioO w48MR2UPXNkbSev9gz6QMDwQH/JeofI9llxgbRXHeXJhk1yjrGpQH6SsE0lx/hgkth3O YvLA== X-Gm-Message-State: AOJu0YzI1XkxJLm0Z9yrIeSoy1GaAuJeWqDsPHj5IJO0XL2UycY8Mp4t FDhLGsz83iY2M7CnjuIvZ5BzOJDgoIyGYCqxDpH2Xc2BZ04H++bF0sdx X-Gm-Gg: ASbGncuyt3CY1yE+uMIsAxRBCsXv4P3cbkXapdChCxr9WBaGkFB/ISPDOE6GLVzjAjt cCRYoda74cEc4+ETZCD96Au8wXw/jBaqz/A3HptIkQNNpscoikx0AMRa8cnedb3zVfYQnBFYr9H hnK0w+LavNsQ2JrUL6fK7081HHivMmeF7d42FmEocBUF8N7S0NaEsrfIOD6l8P1NNhKCtrVsIna TjdY6t90pR4LQ+JLdsJd+hfSOe+4q3p1tp7Jb123vIN/oCClbpHGBxRLuQ5UKMtnhGMZawit+6f UUZ2qNTLJDr2O2NUv9NuPxKTTUSJOtlUeK9gRNSLNCDX0PG8JpiON/OKm5tyChm2seUEaz39yrw hJxihtDBsu2n2v6zNJdsZfLU7dNooPc7E8VmZ6LtTmYSaj7AphTYip8EBr1hfO3YxLFxieCvFCr ZtUiv0he3h X-Google-Smtp-Source: AGHT+IGxVGwC/sKYUwqRKG8lXLIscGWazq91V21BiqeQVaA3olRdaG4t9Av2Vc0WTgyCeiUBd+OaoA== X-Received: by 2002:a17:907:9690:b0:b76:8074:344b with SMTP id a640c23a62f3a-b76c5355c41mr2315128366b.8.1764592936640; Mon, 01 Dec 2025 04:42:16 -0800 (PST) Received: from victus-lab ([193.205.81.5]) by smtp.gmail.com with ESMTPSA id a640c23a62f3a-b76f59e8612sm1173738266b.52.2025.12.01.04.42.15 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 01 Dec 2025 04:42:16 -0800 (PST) From: Yuri Andriaccio To: Ingo Molnar , Peter Zijlstra , Juri Lelli , Vincent Guittot , Dietmar Eggemann , Steven Rostedt , Ben Segall , Mel Gorman , Valentin Schneider Cc: linux-kernel@vger.kernel.org, Luca Abeni , Yuri Andriaccio Subject: [RFC PATCH v4 10/28] sched/rt: Add {alloc/free}_rt_sched_group Date: Mon, 1 Dec 2025 13:41:43 +0100 Message-ID: <20251201124205.11169-11-yurand2000@gmail.com> X-Mailer: git-send-email 2.51.0 In-Reply-To: <20251201124205.11169-1-yurand2000@gmail.com> References: <20251201124205.11169-1-yurand2000@gmail.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset="utf-8" From: luca abeni - Add allocation and deallocation code for rt-cgroups. - Declare dl_server specific functions (only skeleton, but no implementation yet), needed by the deadline servers to be called when trying to schedule. Co-developed-by: Alessio Balsini Signed-off-by: Alessio Balsini Co-developed-by: Andrea Parri Signed-off-by: Andrea Parri Co-developed-by: Yuri Andriaccio Signed-off-by: Yuri Andriaccio Signed-off-by: luca abeni --- kernel/sched/rt.c | 125 +++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 123 insertions(+), 2 deletions(-) diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c index 6a51232de6..2301efc03f 100644 --- a/kernel/sched/rt.c +++ b/kernel/sched/rt.c @@ -92,19 +92,140 @@ void unregister_rt_sched_group(struct task_group *tg) =20 void free_rt_sched_group(struct task_group *tg) { + int i; + unsigned long flags; + if (!rt_group_sched_enabled()) return; + + if (!tg->dl_se || !tg->rt_rq) + return; + + for_each_possible_cpu(i) { + if (!tg->dl_se[i] || !tg->rt_rq[i]) + continue; + + /* + * Shutdown the dl_server and free it + * + * Since the dl timer is going to be cancelled, + * we risk to never decrease the running bw... + * Fix this issue by changing the group runtime + * to 0 immediately before freeing it. + */ + dl_init_tg(tg->dl_se[i], 0, tg->dl_se[i]->dl_period); + + raw_spin_rq_lock_irqsave(cpu_rq(i), flags); + hrtimer_cancel(&tg->dl_se[i]->dl_timer); + raw_spin_rq_unlock_irqrestore(cpu_rq(i), flags); + kfree(tg->dl_se[i]); + + /* Free the local per-cpu runqueue */ + kfree(served_rq_of_rt_rq(tg->rt_rq[i])); + } + + kfree(tg->rt_rq); + kfree(tg->dl_se); +} + +static struct task_struct *rt_server_pick(struct sched_dl_entity *dl_se) +{ + return NULL; +} + +static inline void __rt_rq_free(struct rt_rq **rt_rq) +{ + int i; + + for_each_possible_cpu(i) { + kfree(served_rq_of_rt_rq(rt_rq[i])); + } + + kfree(rt_rq); +} + +DEFINE_FREE(rt_rq_free, struct rt_rq **, if (_T) __rt_rq_free(_T)) + +static inline void __dl_se_free(struct sched_dl_entity **dl_se) +{ + int i; + + for_each_possible_cpu(i) { + kfree(dl_se[i]); + } + + kfree(dl_se); } =20 +DEFINE_FREE(dl_se_free, struct sched_dl_entity **, if (_T) __dl_se_free(_T= )) + int alloc_rt_sched_group(struct task_group *tg, struct task_group *parent) { + /* Instantiate automatic cleanup in event of kalloc fail */ + struct rt_rq **tg_rt_rq __free(rt_rq_free) =3D NULL; + struct sched_dl_entity *dl_se, **tg_dl_se __free(dl_se_free) =3D NULL; + struct rq *s_rq; + int i; + if (!rt_group_sched_enabled()) return 1; =20 + /* Allocate all necessary resources beforehand */ + tg_rt_rq =3D kcalloc(nr_cpu_ids, sizeof(struct rt_rq *), GFP_KERNEL); + if (!tg_rt_rq) + return 0; + + tg_dl_se =3D kcalloc(nr_cpu_ids, + sizeof(struct sched_dl_entity *), GFP_KERNEL); + if (!tg_dl_se) + return 0; + + for_each_possible_cpu(i) { + s_rq =3D kzalloc_node(sizeof(struct rq), + GFP_KERNEL, cpu_to_node(i)); + if (!s_rq) + return 0; + + dl_se =3D kzalloc_node(sizeof(struct sched_dl_entity), + GFP_KERNEL, cpu_to_node(i)); + if (!dl_se) { + kfree(s_rq); + return 0; + } + + tg_rt_rq[i] =3D &s_rq->rt; + tg_dl_se[i] =3D dl_se; + } + + tg->rt_rq =3D no_free_ptr(tg_rt_rq); + tg->dl_se =3D no_free_ptr(tg_dl_se); + + /* Initialize the allocated resources now. */ + init_dl_bandwidth(&tg->dl_bandwidth, 0, 0); + + for_each_possible_cpu(i) { + s_rq =3D served_rq_of_rt_rq(tg->rt_rq[i]); + dl_se =3D tg->dl_se[i]; + + init_rt_rq(&s_rq->rt); + s_rq->cpu =3D i; + s_rq->rt.tg =3D tg; + + init_dl_entity(dl_se); + dl_se->dl_runtime =3D tg->dl_bandwidth.dl_runtime; + dl_se->dl_period =3D tg->dl_bandwidth.dl_period; + dl_se->dl_deadline =3D dl_se->dl_period; + dl_se->dl_bw =3D to_ratio(dl_se->dl_period, dl_se->dl_runtime); + dl_se->dl_density =3D to_ratio(dl_se->dl_period, dl_se->dl_runtime); + dl_se->dl_server =3D 1; + + dl_server_init(dl_se, &cpu_rq(i)->dl, s_rq, rt_server_pick); + } + return 1; } =20 -#else /* !CONFIG_RT_GROUP_SCHED: */ +#else /* !CONFIG_RT_GROUP_SCHED */ =20 void unregister_rt_sched_group(struct task_group *tg) { } =20 @@ -114,7 +235,7 @@ int alloc_rt_sched_group(struct task_group *tg, struct = task_group *parent) { return 1; } -#endif /* !CONFIG_RT_GROUP_SCHED */ +#endif /* CONFIG_RT_GROUP_SCHED */ =20 static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *pr= ev) { --=20 2.51.0