It will only be called from bfq_bfqq_handle_idle_busy_switch() in
specific code branch, there is no need to precaculate
'bfqq_wants_to_preempt' each time bfq_bfqq_handle_idle_busy_switch()
is caleld.
Signed-off-by: Yu Kuai <yukuai3@huawei.com>
---
block/bfq-iosched.c | 32 +++++++-------------------------
1 file changed, 7 insertions(+), 25 deletions(-)
diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
index e36a16684fb4..1e57d76c8dd3 100644
--- a/block/bfq-iosched.c
+++ b/block/bfq-iosched.c
@@ -1555,10 +1555,11 @@ static int bfq_min_budget(struct bfq_data *bfqd)
* responsibility of handling the above case 2.
*/
static bool bfq_bfqq_update_budg_for_activation(struct bfq_data *bfqd,
- struct bfq_queue *bfqq,
- bool arrived_in_time)
+ struct bfq_queue *bfqq)
{
struct bfq_entity *entity = &bfqq->entity;
+ bool arrived_in_time = ktime_get_ns() <= bfqq->ttime.last_end_request +
+ bfqd->bfq_slice_idle * 3;
/*
* In the next compound condition, we check also whether there
@@ -1567,7 +1568,7 @@ static bool bfq_bfqq_update_budg_for_activation(struct bfq_data *bfqd,
* would be expired immediately after being selected for
* service. This would only cause useless overhead.
*/
- if (bfq_bfqq_non_blocking_wait_rq(bfqq) && arrived_in_time &&
+ if (arrived_in_time && bfq_bfqq_non_blocking_wait_rq(bfqq) &&
bfq_bfqq_budget_left(bfqq) > 0) {
/*
* We do not clear the flag non_blocking_wait_rq here, as
@@ -1768,17 +1769,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd,
bool *interactive)
{
bool soft_rt, in_burst, wr_or_deserves_wr,
- bfqq_wants_to_preempt,
- idle_for_long_time = bfq_bfqq_idle_for_long_time(bfqd, bfqq),
- /*
- * See the comments on
- * bfq_bfqq_update_budg_for_activation for
- * details on the usage of the next variable.
- */
- arrived_in_time = ktime_get_ns() <=
- bfqq->ttime.last_end_request +
- bfqd->bfq_slice_idle * 3;
-
+ idle_for_long_time = bfq_bfqq_idle_for_long_time(bfqd, bfqq);
/*
* bfqq deserves to be weight-raised if:
@@ -1816,14 +1807,6 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd,
(bfqq->bic || RQ_BIC(rq)->stably_merged) &&
(*interactive || soft_rt)));
- /*
- * Using the last flag, update budget and check whether bfqq
- * may want to preempt the in-service queue.
- */
- bfqq_wants_to_preempt =
- bfq_bfqq_update_budg_for_activation(bfqd, bfqq,
- arrived_in_time);
-
/*
* If bfqq happened to be activated in a burst, but has been
* idle for much more than an interactive queue, then we
@@ -1879,8 +1862,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd,
* guarantees or throughput. As for guarantees, we care
* explicitly about two cases. The first is that bfqq has to
* recover a service hole, as explained in the comments on
- * bfq_bfqq_update_budg_for_activation(), i.e., that
- * bfqq_wants_to_preempt is true. However, if bfqq does not
+ * bfq_bfqq_update_budg_for_activation(). However, if bfqq does not
* carry time-critical I/O, then bfqq's bandwidth is less
* important than that of queues that carry time-critical I/O.
* So, as a further constraint, we consider this case only if
@@ -1918,7 +1900,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd,
* (2) this switch of bfqq to busy changes the scenario.
*/
if (bfqd->in_service_queue &&
- ((bfqq_wants_to_preempt &&
+ ((bfq_bfqq_update_budg_for_activation(bfqd, bfqq) &&
bfqq->wr_coeff >= bfqd->in_service_queue->wr_coeff) ||
bfq_bfqq_higher_class_or_weight(bfqq, bfqd->in_service_queue) ||
!bfq_better_to_idle(bfqd->in_service_queue)) &&
--
2.31.1
On Sat 14-05-22 17:05:21, Yu Kuai wrote: > It will only be called from bfq_bfqq_handle_idle_busy_switch() in > specific code branch, there is no need to precaculate > 'bfqq_wants_to_preempt' each time bfq_bfqq_handle_idle_busy_switch() > is caleld. > > Signed-off-by: Yu Kuai <yukuai3@huawei.com> Please see below: > @@ -1816,14 +1807,6 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, > (bfqq->bic || RQ_BIC(rq)->stably_merged) && > (*interactive || soft_rt))); > > - /* > - * Using the last flag, update budget and check whether bfqq > - * may want to preempt the in-service queue. > - */ > - bfqq_wants_to_preempt = > - bfq_bfqq_update_budg_for_activation(bfqd, bfqq, > - arrived_in_time); > - > /* > * If bfqq happened to be activated in a burst, but has been > * idle for much more than an interactive queue, then we ... > @@ -1918,7 +1900,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, > * (2) this switch of bfqq to busy changes the scenario. > */ > if (bfqd->in_service_queue && > - ((bfqq_wants_to_preempt && > + ((bfq_bfqq_update_budg_for_activation(bfqd, bfqq) && > bfqq->wr_coeff >= bfqd->in_service_queue->wr_coeff) || > bfq_bfqq_higher_class_or_weight(bfqq, bfqd->in_service_queue) || > !bfq_better_to_idle(bfqd->in_service_queue)) && So these changes are actually wrong because bfq_bfqq_update_budg_for_activation() relies on bfq_bfqq_non_blocking_wait_rq() but bfq_add_bfqq_busy() clears that. And bfq_add_bfqq_busy() is called between the place where bfq_bfqq_update_budg_for_activation() was called previously and now so your patch breaks this logic. Honza -- Jan Kara <jack@suse.com> SUSE Labs, CR
在 2022/05/19 19:18, Jan Kara 写道: > On Sat 14-05-22 17:05:21, Yu Kuai wrote: >> It will only be called from bfq_bfqq_handle_idle_busy_switch() in >> specific code branch, there is no need to precaculate >> 'bfqq_wants_to_preempt' each time bfq_bfqq_handle_idle_busy_switch() >> is caleld. >> >> Signed-off-by: Yu Kuai <yukuai3@huawei.com> > > Please see below: > >> @@ -1816,14 +1807,6 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, >> (bfqq->bic || RQ_BIC(rq)->stably_merged) && >> (*interactive || soft_rt))); >> >> - /* >> - * Using the last flag, update budget and check whether bfqq >> - * may want to preempt the in-service queue. >> - */ >> - bfqq_wants_to_preempt = >> - bfq_bfqq_update_budg_for_activation(bfqd, bfqq, >> - arrived_in_time); >> - >> /* >> * If bfqq happened to be activated in a burst, but has been >> * idle for much more than an interactive queue, then we > ... >> @@ -1918,7 +1900,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, >> * (2) this switch of bfqq to busy changes the scenario. >> */ >> if (bfqd->in_service_queue && >> - ((bfqq_wants_to_preempt && >> + ((bfq_bfqq_update_budg_for_activation(bfqd, bfqq) && >> bfqq->wr_coeff >= bfqd->in_service_queue->wr_coeff) || >> bfq_bfqq_higher_class_or_weight(bfqq, bfqd->in_service_queue) || >> !bfq_better_to_idle(bfqd->in_service_queue)) && > > So these changes are actually wrong because > bfq_bfqq_update_budg_for_activation() relies on > bfq_bfqq_non_blocking_wait_rq() but bfq_add_bfqq_busy() clears that. And > bfq_add_bfqq_busy() is called between the place where > bfq_bfqq_update_budg_for_activation() was called previously and now so your > patch breaks this logic. Hi, You are right, thanks for the explanation, I'll remove this patch and the next patch in next version. Kuai > > Honza >
© 2016 - 2026 Red Hat, Inc.