It will only be called from bfq_bfqq_handle_idle_busy_switch() in specific code branch, there is no need to precaculate 'bfqq_wants_to_preempt' each time bfq_bfqq_handle_idle_busy_switch() is caleld. Signed-off-by: Yu Kuai <yukuai3@xxxxxxxxxx> --- block/bfq-iosched.c | 32 +++++++------------------------- 1 file changed, 7 insertions(+), 25 deletions(-) diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c index e36a16684fb4..1e57d76c8dd3 100644 --- a/block/bfq-iosched.c +++ b/block/bfq-iosched.c @@ -1555,10 +1555,11 @@ static int bfq_min_budget(struct bfq_data *bfqd) * responsibility of handling the above case 2. */ static bool bfq_bfqq_update_budg_for_activation(struct bfq_data *bfqd, - struct bfq_queue *bfqq, - bool arrived_in_time) + struct bfq_queue *bfqq) { struct bfq_entity *entity = &bfqq->entity; + bool arrived_in_time = ktime_get_ns() <= bfqq->ttime.last_end_request + + bfqd->bfq_slice_idle * 3; /* * In the next compound condition, we check also whether there @@ -1567,7 +1568,7 @@ static bool bfq_bfqq_update_budg_for_activation(struct bfq_data *bfqd, * would be expired immediately after being selected for * service. This would only cause useless overhead. */ - if (bfq_bfqq_non_blocking_wait_rq(bfqq) && arrived_in_time && + if (arrived_in_time && bfq_bfqq_non_blocking_wait_rq(bfqq) && bfq_bfqq_budget_left(bfqq) > 0) { /* * We do not clear the flag non_blocking_wait_rq here, as @@ -1768,17 +1769,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, bool *interactive) { bool soft_rt, in_burst, wr_or_deserves_wr, - bfqq_wants_to_preempt, - idle_for_long_time = bfq_bfqq_idle_for_long_time(bfqd, bfqq), - /* - * See the comments on - * bfq_bfqq_update_budg_for_activation for - * details on the usage of the next variable. - */ - arrived_in_time = ktime_get_ns() <= - bfqq->ttime.last_end_request + - bfqd->bfq_slice_idle * 3; - + idle_for_long_time = bfq_bfqq_idle_for_long_time(bfqd, bfqq); /* * bfqq deserves to be weight-raised if: @@ -1816,14 +1807,6 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, (bfqq->bic || RQ_BIC(rq)->stably_merged) && (*interactive || soft_rt))); - /* - * Using the last flag, update budget and check whether bfqq - * may want to preempt the in-service queue. - */ - bfqq_wants_to_preempt = - bfq_bfqq_update_budg_for_activation(bfqd, bfqq, - arrived_in_time); - /* * If bfqq happened to be activated in a burst, but has been * idle for much more than an interactive queue, then we @@ -1879,8 +1862,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, * guarantees or throughput. As for guarantees, we care * explicitly about two cases. The first is that bfqq has to * recover a service hole, as explained in the comments on - * bfq_bfqq_update_budg_for_activation(), i.e., that - * bfqq_wants_to_preempt is true. However, if bfqq does not + * bfq_bfqq_update_budg_for_activation(). However, if bfqq does not * carry time-critical I/O, then bfqq's bandwidth is less * important than that of queues that carry time-critical I/O. * So, as a further constraint, we consider this case only if @@ -1918,7 +1900,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, * (2) this switch of bfqq to busy changes the scenario. */ if (bfqd->in_service_queue && - ((bfqq_wants_to_preempt && + ((bfq_bfqq_update_budg_for_activation(bfqd, bfqq) && bfqq->wr_coeff >= bfqd->in_service_queue->wr_coeff) || bfq_bfqq_higher_class_or_weight(bfqq, bfqd->in_service_queue) || !bfq_better_to_idle(bfqd->in_service_queue)) && -- 2.31.1