Re: [PATCH V2] block, bfq: remove batches of confusing ifdefs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Jens,
do you think this version could be ok?

Thanks,
Paolo

> Il giorno 04 dic 2017, alle ore 11:42, Paolo Valente <paolo.valente@xxxxxxxxxx> ha scritto:
> 
> Commit a33801e8b473 ("block, bfq: move debug blkio stats behind
> CONFIG_DEBUG_BLK_CGROUP") introduced two batches of confusing ifdefs:
> one reported in [1], plus a similar one in another function. This
> commit removes both batches, in the way suggested in [1].
> 
> [1] https://www.spinics.net/lists/linux-block/msg20043.html
> 
> Fixes: a33801e8b473 ("block, bfq: move debug blkio stats behind CONFIG_DEBUG_BLK_CGROUP")
> Reported-by: Linus Torvalds <torvalds@xxxxxxxxxxxxxxxxxxxx>
> Tested-by: Luca Miccio <lucmiccio@xxxxxxxxx>
> Signed-off-by: Paolo Valente <paolo.valente@xxxxxxxxxx>
> ---
> block/bfq-iosched.c | 127 +++++++++++++++++++++++++++++-----------------------
> 1 file changed, 72 insertions(+), 55 deletions(-)
> 
> diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
> index bcb6d21..3feed64 100644
> --- a/block/bfq-iosched.c
> +++ b/block/bfq-iosched.c
> @@ -3689,35 +3689,16 @@ static struct request *__bfq_dispatch_request(struct blk_mq_hw_ctx *hctx)
> 	return rq;
> }
> 
> -static struct request *bfq_dispatch_request(struct blk_mq_hw_ctx *hctx)
> -{
> -	struct bfq_data *bfqd = hctx->queue->elevator->elevator_data;
> -	struct request *rq;
> #if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> -	struct bfq_queue *in_serv_queue, *bfqq;
> -	bool waiting_rq, idle_timer_disabled;
> -#endif
> -
> -	spin_lock_irq(&bfqd->lock);
> -
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> -	in_serv_queue = bfqd->in_service_queue;
> -	waiting_rq = in_serv_queue && bfq_bfqq_wait_request(in_serv_queue);
> -
> -	rq = __bfq_dispatch_request(hctx);
> -
> -	idle_timer_disabled =
> -		waiting_rq && !bfq_bfqq_wait_request(in_serv_queue);
> -
> -#else
> -	rq = __bfq_dispatch_request(hctx);
> -#endif
> -	spin_unlock_irq(&bfqd->lock);
> +static void bfq_update_dispatch_stats(struct request_queue *q,
> +				      struct request *rq,
> +				      struct bfq_queue *in_serv_queue,
> +				      bool idle_timer_disabled)
> +{
> +	struct bfq_queue *bfqq = rq ? RQ_BFQQ(rq) : NULL;
> 
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> -	bfqq = rq ? RQ_BFQQ(rq) : NULL;
> 	if (!idle_timer_disabled && !bfqq)
> -		return rq;
> +		return;
> 
> 	/*
> 	 * rq and bfqq are guaranteed to exist until this function
> @@ -3732,7 +3713,7 @@ static struct request *bfq_dispatch_request(struct blk_mq_hw_ctx *hctx)
> 	 * In addition, the following queue lock guarantees that
> 	 * bfqq_group(bfqq) exists as well.
> 	 */
> -	spin_lock_irq(hctx->queue->queue_lock);
> +	spin_lock_irq(q->queue_lock);
> 	if (idle_timer_disabled)
> 		/*
> 		 * Since the idle timer has been disabled,
> @@ -3751,9 +3732,37 @@ static struct request *bfq_dispatch_request(struct blk_mq_hw_ctx *hctx)
> 		bfqg_stats_set_start_empty_time(bfqg);
> 		bfqg_stats_update_io_remove(bfqg, rq->cmd_flags);
> 	}
> -	spin_unlock_irq(hctx->queue->queue_lock);
> +	spin_unlock_irq(q->queue_lock);
> +}
> +#else
> +static inline void bfq_update_dispatch_stats(struct request_queue *q,
> +					     struct request *rq,
> +					     struct bfq_queue *in_serv_queue,
> +					     bool idle_timer_disabled) {}
> #endif
> 
> +static struct request *bfq_dispatch_request(struct blk_mq_hw_ctx *hctx)
> +{
> +	struct bfq_data *bfqd = hctx->queue->elevator->elevator_data;
> +	struct request *rq;
> +	struct bfq_queue *in_serv_queue;
> +	bool waiting_rq, idle_timer_disabled;
> +
> +	spin_lock_irq(&bfqd->lock);
> +
> +	in_serv_queue = bfqd->in_service_queue;
> +	waiting_rq = in_serv_queue && bfq_bfqq_wait_request(in_serv_queue);
> +
> +	rq = __bfq_dispatch_request(hctx);
> +
> +	idle_timer_disabled =
> +		waiting_rq && !bfq_bfqq_wait_request(in_serv_queue);
> +
> +	spin_unlock_irq(&bfqd->lock);
> +
> +	bfq_update_dispatch_stats(hctx->queue, rq, in_serv_queue,
> +				  idle_timer_disabled);
> +
> 	return rq;
> }
> 
> @@ -4276,16 +4285,46 @@ static bool __bfq_insert_request(struct bfq_data *bfqd, struct request *rq)
> 	return idle_timer_disabled;
> }
> 
> +#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> +static void bfq_update_insert_stats(struct request_queue *q,
> +				    struct bfq_queue *bfqq,
> +				    bool idle_timer_disabled,
> +				    unsigned int cmd_flags)
> +{
> +	if (!bfqq)
> +		return;
> +
> +	/*
> +	 * bfqq still exists, because it can disappear only after
> +	 * either it is merged with another queue, or the process it
> +	 * is associated with exits. But both actions must be taken by
> +	 * the same process currently executing this flow of
> +	 * instructions.
> +	 *
> +	 * In addition, the following queue lock guarantees that
> +	 * bfqq_group(bfqq) exists as well.
> +	 */
> +	spin_lock_irq(q->queue_lock);
> +	bfqg_stats_update_io_add(bfqq_group(bfqq), bfqq, cmd_flags);
> +	if (idle_timer_disabled)
> +		bfqg_stats_update_idle_time(bfqq_group(bfqq));
> +	spin_unlock_irq(q->queue_lock);
> +}
> +#else
> +static inline void bfq_update_insert_stats(struct request_queue *q,
> +					   struct bfq_queue *bfqq,
> +					   bool idle_timer_disabled,
> +					   unsigned int cmd_flags) {}
> +#endif
> +
> static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
> 			       bool at_head)
> {
> 	struct request_queue *q = hctx->queue;
> 	struct bfq_data *bfqd = q->elevator->elevator_data;
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> 	struct bfq_queue *bfqq = RQ_BFQQ(rq);
> 	bool idle_timer_disabled = false;
> 	unsigned int cmd_flags;
> -#endif
> 
> 	spin_lock_irq(&bfqd->lock);
> 	if (blk_mq_sched_try_insert_merge(q, rq)) {
> @@ -4304,7 +4343,6 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
> 		else
> 			list_add_tail(&rq->queuelist, &bfqd->dispatch);
> 	} else {
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> 		idle_timer_disabled = __bfq_insert_request(bfqd, rq);
> 		/*
> 		 * Update bfqq, because, if a queue merge has occurred
> @@ -4312,9 +4350,6 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
> 		 * redirected into a new queue.
> 		 */
> 		bfqq = RQ_BFQQ(rq);
> -#else
> -		__bfq_insert_request(bfqd, rq);
> -#endif
> 
> 		if (rq_mergeable(rq)) {
> 			elv_rqhash_add(q, rq);
> @@ -4323,35 +4358,17 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
> 		}
> 	}
> 
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> 	/*
> 	 * Cache cmd_flags before releasing scheduler lock, because rq
> 	 * may disappear afterwards (for example, because of a request
> 	 * merge).
> 	 */
> 	cmd_flags = rq->cmd_flags;
> -#endif
> +
> 	spin_unlock_irq(&bfqd->lock);
> 
> -#if defined(CONFIG_BFQ_GROUP_IOSCHED) && defined(CONFIG_DEBUG_BLK_CGROUP)
> -	if (!bfqq)
> -		return;
> -	/*
> -	 * bfqq still exists, because it can disappear only after
> -	 * either it is merged with another queue, or the process it
> -	 * is associated with exits. But both actions must be taken by
> -	 * the same process currently executing this flow of
> -	 * instruction.
> -	 *
> -	 * In addition, the following queue lock guarantees that
> -	 * bfqq_group(bfqq) exists as well.
> -	 */
> -	spin_lock_irq(q->queue_lock);
> -	bfqg_stats_update_io_add(bfqq_group(bfqq), bfqq, cmd_flags);
> -	if (idle_timer_disabled)
> -		bfqg_stats_update_idle_time(bfqq_group(bfqq));
> -	spin_unlock_irq(q->queue_lock);
> -#endif
> +	bfq_update_insert_stats(q, bfqq, idle_timer_disabled,
> +				cmd_flags);
> }
> 
> static void bfq_insert_requests(struct blk_mq_hw_ctx *hctx,
> -- 
> 2.10.0
> 





[Index of Archives]     [Linux RAID]     [Linux SCSI]     [Linux ATA RAID]     [IDE]     [Linux Wireless]     [Linux Kernel]     [ATH6KL]     [Linux Bluetooth]     [Linux Netdev]     [Kernel Newbies]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Device Mapper]

  Powered by Linux