From: Geliang Tang <tanggeliang@kylinos.cn>
Use the newly added bpf_for_each() helper to walk the conn_list.
Signed-off-by: Geliang Tang <tanggeliang@kylinos.cn>
---
.../selftests/bpf/progs/mptcp_bpf_burst.c | 79 ++++++++++---------
1 file changed, 40 insertions(+), 39 deletions(-)
diff --git a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c
index eb21119aa8f7..e7df5f048aa4 100644
--- a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c
+++ b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c
@@ -11,6 +11,10 @@ char _license[] SEC("license") = "GPL";
#define min(a, b) ((a) < (b) ? (a) : (b))
+#define SSK_MODE_ACTIVE 0
+#define SSK_MODE_BACKUP 1
+#define SSK_MODE_MAX 2
+
struct bpf_subflow_send_info {
__u8 subflow_id;
__u64 linger_time;
@@ -23,10 +27,6 @@ extern bool tcp_stream_memory_free(const struct sock *sk, int wake) __ksym;
extern bool bpf_mptcp_subflow_queues_empty(struct sock *sk) __ksym;
extern void mptcp_pm_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk) __ksym;
-#define SSK_MODE_ACTIVE 0
-#define SSK_MODE_BACKUP 1
-#define SSK_MODE_MAX 2
-
static __always_inline __u64 div_u64(__u64 dividend, __u32 divisor)
{
return dividend / divisor;
@@ -57,6 +57,19 @@ static __always_inline bool sk_stream_memory_free(const struct sock *sk)
return __sk_stream_memory_free(sk, 0);
}
+static struct mptcp_subflow_context *
+mptcp_lookup_subflow_by_id(struct mptcp_sock *msk, unsigned int id)
+{
+ struct mptcp_subflow_context *subflow;
+
+ bpf_for_each(mptcp_subflow, subflow, msk) {
+ if (subflow->subflow_id == id)
+ return subflow;
+ }
+
+ return NULL;
+}
+
SEC("struct_ops")
void BPF_PROG(mptcp_sched_burst_init, struct mptcp_sock *msk)
{
@@ -67,8 +80,7 @@ void BPF_PROG(mptcp_sched_burst_release, struct mptcp_sock *msk)
{
}
-static int bpf_burst_get_send(struct mptcp_sock *msk,
- struct mptcp_sched_data *data)
+static int bpf_burst_get_send(struct mptcp_sock *msk)
{
struct bpf_subflow_send_info send_info[SSK_MODE_MAX];
struct mptcp_subflow_context *subflow;
@@ -84,16 +96,10 @@ static int bpf_burst_get_send(struct mptcp_sock *msk,
send_info[i].linger_time = -1;
}
- for (i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
- bool backup;
+ bpf_for_each(mptcp_subflow, subflow, msk) {
+ bool backup = subflow->backup || subflow->request_bkup;
- subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
- if (!subflow)
- break;
-
- backup = subflow->backup || subflow->request_bkup;
-
- ssk = mptcp_subflow_tcp_sock(subflow);
+ ssk = bpf_mptcp_subflow_tcp_sock(subflow);
if (!mptcp_subflow_active(subflow))
continue;
@@ -109,7 +115,7 @@ static int bpf_burst_get_send(struct mptcp_sock *msk,
linger_time = div_u64((__u64)ssk->sk_wmem_queued << 32, pace);
if (linger_time < send_info[backup].linger_time) {
- send_info[backup].subflow_id = i;
+ send_info[backup].subflow_id = subflow->subflow_id;
send_info[backup].linger_time = linger_time;
}
}
@@ -119,10 +125,10 @@ static int bpf_burst_get_send(struct mptcp_sock *msk,
if (!nr_active)
send_info[SSK_MODE_ACTIVE].subflow_id = send_info[SSK_MODE_BACKUP].subflow_id;
- subflow = bpf_mptcp_subflow_ctx_by_pos(data, send_info[SSK_MODE_ACTIVE].subflow_id);
+ subflow = mptcp_lookup_subflow_by_id(msk, send_info[SSK_MODE_ACTIVE].subflow_id);
if (!subflow)
return -1;
- ssk = mptcp_subflow_tcp_sock(subflow);
+ ssk = bpf_mptcp_subflow_tcp_sock(subflow);
if (!ssk || !sk_stream_memory_free(ssk))
return -1;
@@ -141,23 +147,18 @@ static int bpf_burst_get_send(struct mptcp_sock *msk,
return 0;
}
-static int bpf_burst_get_retrans(struct mptcp_sock *msk,
- struct mptcp_sched_data *data)
+static int bpf_burst_get_retrans(struct mptcp_sock *msk)
{
- int backup = MPTCP_SUBFLOWS_MAX, pick = MPTCP_SUBFLOWS_MAX, subflow_id;
+ struct sock *backup = NULL, *pick = NULL;
struct mptcp_subflow_context *subflow;
int min_stale_count = INT_MAX;
- struct sock *ssk;
- for (int i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
- subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
- if (!subflow)
- break;
+ bpf_for_each(mptcp_subflow, subflow, msk) {
+ struct sock *ssk = bpf_mptcp_subflow_tcp_sock(subflow);
if (!mptcp_subflow_active(subflow))
continue;
- ssk = mptcp_subflow_tcp_sock(subflow);
/* still data outstanding at TCP level? skip this */
if (!tcp_rtx_and_write_queues_empty(ssk)) {
mptcp_pm_subflow_chk_stale(msk, ssk);
@@ -166,23 +167,23 @@ static int bpf_burst_get_retrans(struct mptcp_sock *msk,
}
if (subflow->backup || subflow->request_bkup) {
- if (backup == MPTCP_SUBFLOWS_MAX)
- backup = i;
+ if (!backup)
+ backup = ssk;
continue;
}
- if (pick == MPTCP_SUBFLOWS_MAX)
- pick = i;
+ if (!pick)
+ pick = ssk;
}
- if (pick < MPTCP_SUBFLOWS_MAX) {
- subflow_id = pick;
+ if (pick)
goto out;
- }
- subflow_id = min_stale_count > 1 ? backup : MPTCP_SUBFLOWS_MAX;
+ pick = min_stale_count > 1 ? backup : NULL;
out:
- subflow = bpf_mptcp_subflow_ctx_by_pos(data, subflow_id);
+ if (!pick)
+ return -1;
+ subflow = bpf_mptcp_subflow_ctx(pick);
if (!subflow)
return -1;
mptcp_subflow_set_scheduled(subflow, true);
@@ -194,11 +195,11 @@ int BPF_PROG(bpf_burst_get_subflow, struct mptcp_sock *msk,
struct mptcp_sched_data *data)
{
if (data->reinject)
- return bpf_burst_get_retrans(msk, data);
- return bpf_burst_get_send(msk, data);
+ return bpf_burst_get_retrans(msk);
+ return bpf_burst_get_send(msk);
}
-SEC(".struct_ops")
+SEC(".struct_ops.link")
struct mptcp_sched_ops burst = {
.init = (void *)mptcp_sched_burst_init,
.release = (void *)mptcp_sched_burst_release,
--
2.45.2
On Tue, 22 Oct 2024, Geliang Tang wrote: > From: Geliang Tang <tanggeliang@kylinos.cn> > > Use the newly added bpf_for_each() helper to walk the conn_list. > > Signed-off-by: Geliang Tang <tanggeliang@kylinos.cn> > --- > .../selftests/bpf/progs/mptcp_bpf_burst.c | 79 ++++++++++--------- > 1 file changed, 40 insertions(+), 39 deletions(-) > > diff --git a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > index eb21119aa8f7..e7df5f048aa4 100644 > --- a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > +++ b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > @@ -11,6 +11,10 @@ char _license[] SEC("license") = "GPL"; > > #define min(a, b) ((a) < (b) ? (a) : (b)) > > +#define SSK_MODE_ACTIVE 0 > +#define SSK_MODE_BACKUP 1 > +#define SSK_MODE_MAX 2 > + Hi Geliang - > struct bpf_subflow_send_info { > __u8 subflow_id; If you store a subflow pointer here instead of an index, there's no need to add the mptcp_lookup_subflow_id() helper function, an extra iteration over conn_list is eliminated, and the code is more like mptcp_subflow_get_send(). - Mat > __u64 linger_time; > @@ -23,10 +27,6 @@ extern bool tcp_stream_memory_free(const struct sock *sk, int wake) __ksym; > extern bool bpf_mptcp_subflow_queues_empty(struct sock *sk) __ksym; > extern void mptcp_pm_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk) __ksym; > > -#define SSK_MODE_ACTIVE 0 > -#define SSK_MODE_BACKUP 1 > -#define SSK_MODE_MAX 2 > - > static __always_inline __u64 div_u64(__u64 dividend, __u32 divisor) > { > return dividend / divisor; > @@ -57,6 +57,19 @@ static __always_inline bool sk_stream_memory_free(const struct sock *sk) > return __sk_stream_memory_free(sk, 0); > } > > +static struct mptcp_subflow_context * > +mptcp_lookup_subflow_by_id(struct mptcp_sock *msk, unsigned int id) > +{ > + struct mptcp_subflow_context *subflow; > + > + bpf_for_each(mptcp_subflow, subflow, msk) { > + if (subflow->subflow_id == id) > + return subflow; > + } > + > + return NULL; > +} > + > SEC("struct_ops") > void BPF_PROG(mptcp_sched_burst_init, struct mptcp_sock *msk) > { > @@ -67,8 +80,7 @@ void BPF_PROG(mptcp_sched_burst_release, struct mptcp_sock *msk) > { > } > > -static int bpf_burst_get_send(struct mptcp_sock *msk, > - struct mptcp_sched_data *data) > +static int bpf_burst_get_send(struct mptcp_sock *msk) > { > struct bpf_subflow_send_info send_info[SSK_MODE_MAX]; > struct mptcp_subflow_context *subflow; > @@ -84,16 +96,10 @@ static int bpf_burst_get_send(struct mptcp_sock *msk, > send_info[i].linger_time = -1; > } > > - for (i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) { > - bool backup; > + bpf_for_each(mptcp_subflow, subflow, msk) { > + bool backup = subflow->backup || subflow->request_bkup; > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, i); > - if (!subflow) > - break; > - > - backup = subflow->backup || subflow->request_bkup; > - > - ssk = mptcp_subflow_tcp_sock(subflow); > + ssk = bpf_mptcp_subflow_tcp_sock(subflow); > if (!mptcp_subflow_active(subflow)) > continue; > > @@ -109,7 +115,7 @@ static int bpf_burst_get_send(struct mptcp_sock *msk, > > linger_time = div_u64((__u64)ssk->sk_wmem_queued << 32, pace); > if (linger_time < send_info[backup].linger_time) { > - send_info[backup].subflow_id = i; > + send_info[backup].subflow_id = subflow->subflow_id; > send_info[backup].linger_time = linger_time; > } > } > @@ -119,10 +125,10 @@ static int bpf_burst_get_send(struct mptcp_sock *msk, > if (!nr_active) > send_info[SSK_MODE_ACTIVE].subflow_id = send_info[SSK_MODE_BACKUP].subflow_id; > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, send_info[SSK_MODE_ACTIVE].subflow_id); > + subflow = mptcp_lookup_subflow_by_id(msk, send_info[SSK_MODE_ACTIVE].subflow_id); > if (!subflow) > return -1; > - ssk = mptcp_subflow_tcp_sock(subflow); > + ssk = bpf_mptcp_subflow_tcp_sock(subflow); > if (!ssk || !sk_stream_memory_free(ssk)) > return -1; > > @@ -141,23 +147,18 @@ static int bpf_burst_get_send(struct mptcp_sock *msk, > return 0; > } > > -static int bpf_burst_get_retrans(struct mptcp_sock *msk, > - struct mptcp_sched_data *data) > +static int bpf_burst_get_retrans(struct mptcp_sock *msk) > { > - int backup = MPTCP_SUBFLOWS_MAX, pick = MPTCP_SUBFLOWS_MAX, subflow_id; > + struct sock *backup = NULL, *pick = NULL; > struct mptcp_subflow_context *subflow; > int min_stale_count = INT_MAX; > - struct sock *ssk; > > - for (int i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) { > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, i); > - if (!subflow) > - break; > + bpf_for_each(mptcp_subflow, subflow, msk) { > + struct sock *ssk = bpf_mptcp_subflow_tcp_sock(subflow); > > if (!mptcp_subflow_active(subflow)) > continue; > > - ssk = mptcp_subflow_tcp_sock(subflow); > /* still data outstanding at TCP level? skip this */ > if (!tcp_rtx_and_write_queues_empty(ssk)) { > mptcp_pm_subflow_chk_stale(msk, ssk); > @@ -166,23 +167,23 @@ static int bpf_burst_get_retrans(struct mptcp_sock *msk, > } > > if (subflow->backup || subflow->request_bkup) { > - if (backup == MPTCP_SUBFLOWS_MAX) > - backup = i; > + if (!backup) > + backup = ssk; > continue; > } > > - if (pick == MPTCP_SUBFLOWS_MAX) > - pick = i; > + if (!pick) > + pick = ssk; > } > > - if (pick < MPTCP_SUBFLOWS_MAX) { > - subflow_id = pick; > + if (pick) > goto out; > - } > - subflow_id = min_stale_count > 1 ? backup : MPTCP_SUBFLOWS_MAX; > + pick = min_stale_count > 1 ? backup : NULL; > > out: > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, subflow_id); > + if (!pick) > + return -1; > + subflow = bpf_mptcp_subflow_ctx(pick); > if (!subflow) > return -1; > mptcp_subflow_set_scheduled(subflow, true); > @@ -194,11 +195,11 @@ int BPF_PROG(bpf_burst_get_subflow, struct mptcp_sock *msk, > struct mptcp_sched_data *data) > { > if (data->reinject) > - return bpf_burst_get_retrans(msk, data); > - return bpf_burst_get_send(msk, data); > + return bpf_burst_get_retrans(msk); > + return bpf_burst_get_send(msk); > } > > -SEC(".struct_ops") > +SEC(".struct_ops.link") > struct mptcp_sched_ops burst = { > .init = (void *)mptcp_sched_burst_init, > .release = (void *)mptcp_sched_burst_release, > -- > 2.45.2 > > >
On Tue, 2024-10-22 at 17:03 -0700, Mat Martineau wrote: > On Tue, 22 Oct 2024, Geliang Tang wrote: > > > From: Geliang Tang <tanggeliang@kylinos.cn> > > > > Use the newly added bpf_for_each() helper to walk the conn_list. > > > > Signed-off-by: Geliang Tang <tanggeliang@kylinos.cn> > > --- > > .../selftests/bpf/progs/mptcp_bpf_burst.c | 79 ++++++++++------ > > --- > > 1 file changed, 40 insertions(+), 39 deletions(-) > > > > diff --git a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > > b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > > index eb21119aa8f7..e7df5f048aa4 100644 > > --- a/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > > +++ b/tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c > > @@ -11,6 +11,10 @@ char _license[] SEC("license") = "GPL"; > > > > #define min(a, b) ((a) < (b) ? (a) : (b)) > > > > +#define SSK_MODE_ACTIVE 0 > > +#define SSK_MODE_BACKUP 1 > > +#define SSK_MODE_MAX 2 > > + > > Hi Geliang - > > > struct bpf_subflow_send_info { > > __u8 subflow_id; > > If you store a subflow pointer here instead of an index, there's no > need > to add the mptcp_lookup_subflow_id() helper function, an extra > iteration over conn_list is eliminated, and the code is more like > mptcp_subflow_get_send(). I thought so too. Two changes were needed to achieve this. 1. Move sk_stream_memory_free check inside bpf_for_each() loop. 2. Implement mptcp_subflow_set_scheduled helper in BPF. > > - Mat > > > __u64 linger_time; > > @@ -23,10 +27,6 @@ extern bool tcp_stream_memory_free(const struct > > sock *sk, int wake) __ksym; > > extern bool bpf_mptcp_subflow_queues_empty(struct sock *sk) __ksym; > > extern void mptcp_pm_subflow_chk_stale(const struct mptcp_sock > > *msk, struct sock *ssk) __ksym; > > > > -#define SSK_MODE_ACTIVE 0 > > -#define SSK_MODE_BACKUP 1 > > -#define SSK_MODE_MAX 2 > > - > > static __always_inline __u64 div_u64(__u64 dividend, __u32 divisor) > > { > > return dividend / divisor; > > @@ -57,6 +57,19 @@ static __always_inline bool > > sk_stream_memory_free(const struct sock *sk) > > return __sk_stream_memory_free(sk, 0); > > } > > > > +static struct mptcp_subflow_context * > > +mptcp_lookup_subflow_by_id(struct mptcp_sock *msk, unsigned int > > id) > > +{ > > + struct mptcp_subflow_context *subflow; > > + > > + bpf_for_each(mptcp_subflow, subflow, msk) { > > + if (subflow->subflow_id == id) > > + return subflow; > > + } > > + > > + return NULL; > > +} > > + > > SEC("struct_ops") > > void BPF_PROG(mptcp_sched_burst_init, struct mptcp_sock *msk) > > { > > @@ -67,8 +80,7 @@ void BPF_PROG(mptcp_sched_burst_release, struct > > mptcp_sock *msk) > > { > > } > > > > -static int bpf_burst_get_send(struct mptcp_sock *msk, > > - struct mptcp_sched_data *data) > > +static int bpf_burst_get_send(struct mptcp_sock *msk) > > { > > struct bpf_subflow_send_info send_info[SSK_MODE_MAX]; > > struct mptcp_subflow_context *subflow; > > @@ -84,16 +96,10 @@ static int bpf_burst_get_send(struct mptcp_sock > > *msk, > > send_info[i].linger_time = -1; > > } > > > > - for (i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; > > i++) { > > - bool backup; > > + bpf_for_each(mptcp_subflow, subflow, msk) { > > + bool backup = subflow->backup || subflow- > > >request_bkup; > > > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, i); > > - if (!subflow) > > - break; > > - > > - backup = subflow->backup || subflow->request_bkup; > > - > > - ssk = mptcp_subflow_tcp_sock(subflow); > > + ssk = bpf_mptcp_subflow_tcp_sock(subflow); > > if (!mptcp_subflow_active(subflow)) > > continue; > > > > @@ -109,7 +115,7 @@ static int bpf_burst_get_send(struct mptcp_sock > > *msk, > > > > linger_time = div_u64((__u64)ssk->sk_wmem_queued > > << 32, pace); > > if (linger_time < send_info[backup].linger_time) { > > - send_info[backup].subflow_id = i; > > + send_info[backup].subflow_id = subflow- > > >subflow_id; > > send_info[backup].linger_time = > > linger_time; > > } > > } > > @@ -119,10 +125,10 @@ static int bpf_burst_get_send(struct > > mptcp_sock *msk, > > if (!nr_active) > > send_info[SSK_MODE_ACTIVE].subflow_id = > > send_info[SSK_MODE_BACKUP].subflow_id; > > > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, > > send_info[SSK_MODE_ACTIVE].subflow_id); > > + subflow = mptcp_lookup_subflow_by_id(msk, > > send_info[SSK_MODE_ACTIVE].subflow_id); BPF does not allow access to subflow in this way: subflow = send_info[SSK_MODE_ACTIVE].subflow; So I have to use bpf_core_cast() here as: subflow = bpf_core_cast(send_info[SSK_MODE_ACTIVE].subflow, struct mptcp_subflow_context); > > if (!subflow) > > return -1; > > - ssk = mptcp_subflow_tcp_sock(subflow); > > + ssk = bpf_mptcp_subflow_tcp_sock(subflow); Here BPF doesn't allow passing a cast pointer (subflow) to a kfunc (bpf_mptcp_subflow_tcp_sock). Fortunately we can use mptcp_subflow_tcp_sock instead, which is a BPF helper, not a kfunc: ssk = mptcp_subflow_tcp_sock(subflow); > > if (!ssk || !sk_stream_memory_free(ssk)) Again, BPF does not allow passing a cast pointer to a kfunc, sk_stream_memory_free(ssk) fails. It's not possible to implement a sk_stream_memory_free function in BPF, it's too complicated. So the approach I took in v8 was to move this sk_stream_memory_free check forward to the position of mptcp_subflow_active in bpf_for_each() loop. I think this doesn't change the logic of the burst scheduler, but I'd like to hear your opinion. After this, mptcp_subflow_set_scheduled(subflow, true) is not allowed too. So I have to implement a mptcp_subflow_set_scheduled helper in BPF. It's easy since WRITE_ONCE() is defined in progs/map_kptr.c. > > return -1; > > > > @@ -141,23 +147,18 @@ static int bpf_burst_get_send(struct > > mptcp_sock *msk, > > return 0; > > } > > > > -static int bpf_burst_get_retrans(struct mptcp_sock *msk, > > - struct mptcp_sched_data *data) > > +static int bpf_burst_get_retrans(struct mptcp_sock *msk) > > { > > - int backup = MPTCP_SUBFLOWS_MAX, pick = > > MPTCP_SUBFLOWS_MAX, subflow_id; > > + struct sock *backup = NULL, *pick = NULL; > > struct mptcp_subflow_context *subflow; > > int min_stale_count = INT_MAX; > > - struct sock *ssk; > > > > - for (int i = 0; i < data->subflows && i < > > MPTCP_SUBFLOWS_MAX; i++) { > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, i); > > - if (!subflow) > > - break; > > + bpf_for_each(mptcp_subflow, subflow, msk) { > > + struct sock *ssk = > > bpf_mptcp_subflow_tcp_sock(subflow); > > > > if (!mptcp_subflow_active(subflow)) > > continue; > > > > - ssk = mptcp_subflow_tcp_sock(subflow); > > /* still data outstanding at TCP level? skip this > > */ > > if (!tcp_rtx_and_write_queues_empty(ssk)) { > > mptcp_pm_subflow_chk_stale(msk, ssk); > > @@ -166,23 +167,23 @@ static int bpf_burst_get_retrans(struct > > mptcp_sock *msk, > > } > > > > if (subflow->backup || subflow->request_bkup) { > > - if (backup == MPTCP_SUBFLOWS_MAX) > > - backup = i; > > + if (!backup) > > + backup = ssk; > > continue; > > } > > > > - if (pick == MPTCP_SUBFLOWS_MAX) > > - pick = i; > > + if (!pick) > > + pick = ssk; > > } > > > > - if (pick < MPTCP_SUBFLOWS_MAX) { > > - subflow_id = pick; > > + if (pick) > > goto out; > > - } > > - subflow_id = min_stale_count > 1 ? backup : > > MPTCP_SUBFLOWS_MAX; > > + pick = min_stale_count > 1 ? backup : NULL; > > > > out: > > - subflow = bpf_mptcp_subflow_ctx_by_pos(data, subflow_id); > > + if (!pick) > > + return -1; > > + subflow = bpf_mptcp_subflow_ctx(pick); > > if (!subflow) > > return -1; > > mptcp_subflow_set_scheduled(subflow, true); > > @@ -194,11 +195,11 @@ int BPF_PROG(bpf_burst_get_subflow, struct > > mptcp_sock *msk, > > struct mptcp_sched_data *data) > > { > > if (data->reinject) > > - return bpf_burst_get_retrans(msk, data); > > - return bpf_burst_get_send(msk, data); > > + return bpf_burst_get_retrans(msk); > > + return bpf_burst_get_send(msk); > > } > > > > -SEC(".struct_ops") > > +SEC(".struct_ops.link") > > struct mptcp_sched_ops burst = { > > .init = (void *)mptcp_sched_burst_init, > > .release = (void *)mptcp_sched_burst_release, > > -- > > 2.45.2 > > > > > >
© 2016 - 2025 Red Hat, Inc.