Re: [PATCH v3 07/63] block/bfq: Use the new blk_opf_t type

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu 14-07-22 11:06:33, Bart Van Assche wrote:
> Use the new blk_opf_t type for arguments and variables that represent
> request flags or a bitwise combination of a request operation and
> request flags. Rename those variables from 'op' into 'opf'.
> 
> This patch does not change any functionality.
> 
> Cc: Jan Kara <jack@xxxxxxx>
> Signed-off-by: Bart Van Assche <bvanassche@xxxxxxx>

Looks good to me. Feel free to add:

Reviewed-by: Jan Kara <jack@xxxxxxx>

								Honza

> ---
>  block/bfq-cgroup.c  | 26 +++++++++++++-------------
>  block/bfq-iosched.c | 16 ++++++++--------
>  block/bfq-iosched.h |  8 ++++----
>  3 files changed, 25 insertions(+), 25 deletions(-)
> 
> diff --git a/block/bfq-cgroup.c b/block/bfq-cgroup.c
> index 9fc605791b1e..30b15a9a47c4 100644
> --- a/block/bfq-cgroup.c
> +++ b/block/bfq-cgroup.c
> @@ -220,46 +220,46 @@ void bfqg_stats_update_avg_queue_size(struct bfq_group *bfqg)
>  }
>  
>  void bfqg_stats_update_io_add(struct bfq_group *bfqg, struct bfq_queue *bfqq,
> -			      unsigned int op)
> +			      blk_opf_t opf)
>  {
> -	blkg_rwstat_add(&bfqg->stats.queued, op, 1);
> +	blkg_rwstat_add(&bfqg->stats.queued, opf, 1);
>  	bfqg_stats_end_empty_time(&bfqg->stats);
>  	if (!(bfqq == ((struct bfq_data *)bfqg->bfqd)->in_service_queue))
>  		bfqg_stats_set_start_group_wait_time(bfqg, bfqq_group(bfqq));
>  }
>  
> -void bfqg_stats_update_io_remove(struct bfq_group *bfqg, unsigned int op)
> +void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf)
>  {
> -	blkg_rwstat_add(&bfqg->stats.queued, op, -1);
> +	blkg_rwstat_add(&bfqg->stats.queued, opf, -1);
>  }
>  
> -void bfqg_stats_update_io_merged(struct bfq_group *bfqg, unsigned int op)
> +void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf)
>  {
> -	blkg_rwstat_add(&bfqg->stats.merged, op, 1);
> +	blkg_rwstat_add(&bfqg->stats.merged, opf, 1);
>  }
>  
>  void bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns,
> -				  u64 io_start_time_ns, unsigned int op)
> +				  u64 io_start_time_ns, blk_opf_t opf)
>  {
>  	struct bfqg_stats *stats = &bfqg->stats;
>  	u64 now = ktime_get_ns();
>  
>  	if (now > io_start_time_ns)
> -		blkg_rwstat_add(&stats->service_time, op,
> +		blkg_rwstat_add(&stats->service_time, opf,
>  				now - io_start_time_ns);
>  	if (io_start_time_ns > start_time_ns)
> -		blkg_rwstat_add(&stats->wait_time, op,
> +		blkg_rwstat_add(&stats->wait_time, opf,
>  				io_start_time_ns - start_time_ns);
>  }
>  
>  #else /* CONFIG_BFQ_CGROUP_DEBUG */
>  
>  void bfqg_stats_update_io_add(struct bfq_group *bfqg, struct bfq_queue *bfqq,
> -			      unsigned int op) { }
> -void bfqg_stats_update_io_remove(struct bfq_group *bfqg, unsigned int op) { }
> -void bfqg_stats_update_io_merged(struct bfq_group *bfqg, unsigned int op) { }
> +			      blk_opf_t opf) { }
> +void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { }
> +void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) { }
>  void bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns,
> -				  u64 io_start_time_ns, unsigned int op) { }
> +				  u64 io_start_time_ns, blk_opf_t opf) { }
>  void bfqg_stats_update_dequeue(struct bfq_group *bfqg) { }
>  void bfqg_stats_set_start_empty_time(struct bfq_group *bfqg) { }
>  void bfqg_stats_update_idle_time(struct bfq_group *bfqg) { }
> diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
> index e6d7e6b01a05..c740b41fe0a4 100644
> --- a/block/bfq-iosched.c
> +++ b/block/bfq-iosched.c
> @@ -668,19 +668,19 @@ static bool bfqq_request_over_limit(struct bfq_queue *bfqq, int limit)
>   * significantly affect service guarantees coming from the BFQ scheduling
>   * algorithm.
>   */
> -static void bfq_limit_depth(unsigned int op, struct blk_mq_alloc_data *data)
> +static void bfq_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data)
>  {
>  	struct bfq_data *bfqd = data->q->elevator->elevator_data;
>  	struct bfq_io_cq *bic = bfq_bic_lookup(data->q);
> -	struct bfq_queue *bfqq = bic ? bic_to_bfqq(bic, op_is_sync(op)) : NULL;
> +	struct bfq_queue *bfqq = bic ? bic_to_bfqq(bic, op_is_sync(opf)) : NULL;
>  	int depth;
>  	unsigned limit = data->q->nr_requests;
>  
>  	/* Sync reads have full depth available */
> -	if (op_is_sync(op) && !op_is_write(op)) {
> +	if (op_is_sync(opf) && !op_is_write(opf)) {
>  		depth = 0;
>  	} else {
> -		depth = bfqd->word_depths[!!bfqd->wr_busy_queues][op_is_sync(op)];
> +		depth = bfqd->word_depths[!!bfqd->wr_busy_queues][op_is_sync(opf)];
>  		limit = (limit * depth) >> bfqd->full_depth_shift;
>  	}
>  
> @@ -693,7 +693,7 @@ static void bfq_limit_depth(unsigned int op, struct blk_mq_alloc_data *data)
>  		depth = 1;
>  
>  	bfq_log(bfqd, "[%s] wr_busy %d sync %d depth %u",
> -		__func__, bfqd->wr_busy_queues, op_is_sync(op), depth);
> +		__func__, bfqd->wr_busy_queues, op_is_sync(opf), depth);
>  	if (depth)
>  		data->shallow_depth = depth;
>  }
> @@ -6104,7 +6104,7 @@ static bool __bfq_insert_request(struct bfq_data *bfqd, struct request *rq)
>  static void bfq_update_insert_stats(struct request_queue *q,
>  				    struct bfq_queue *bfqq,
>  				    bool idle_timer_disabled,
> -				    unsigned int cmd_flags)
> +				    blk_opf_t cmd_flags)
>  {
>  	if (!bfqq)
>  		return;
> @@ -6129,7 +6129,7 @@ static void bfq_update_insert_stats(struct request_queue *q,
>  static inline void bfq_update_insert_stats(struct request_queue *q,
>  					   struct bfq_queue *bfqq,
>  					   bool idle_timer_disabled,
> -					   unsigned int cmd_flags) {}
> +					   blk_opf_t cmd_flags) {}
>  #endif /* CONFIG_BFQ_CGROUP_DEBUG */
>  
>  static struct bfq_queue *bfq_init_rq(struct request *rq);
> @@ -6141,7 +6141,7 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
>  	struct bfq_data *bfqd = q->elevator->elevator_data;
>  	struct bfq_queue *bfqq;
>  	bool idle_timer_disabled = false;
> -	unsigned int cmd_flags;
> +	blk_opf_t cmd_flags;
>  	LIST_HEAD(free);
>  
>  #ifdef CONFIG_BFQ_GROUP_IOSCHED
> diff --git a/block/bfq-iosched.h b/block/bfq-iosched.h
> index ca8177d7bf7c..ad8e513d7e87 100644
> --- a/block/bfq-iosched.h
> +++ b/block/bfq-iosched.h
> @@ -994,11 +994,11 @@ void bfq_put_async_queues(struct bfq_data *bfqd, struct bfq_group *bfqg);
>  
>  void bfqg_stats_update_legacy_io(struct request_queue *q, struct request *rq);
>  void bfqg_stats_update_io_add(struct bfq_group *bfqg, struct bfq_queue *bfqq,
> -			      unsigned int op);
> -void bfqg_stats_update_io_remove(struct bfq_group *bfqg, unsigned int op);
> -void bfqg_stats_update_io_merged(struct bfq_group *bfqg, unsigned int op);
> +			      blk_opf_t opf);
> +void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf);
> +void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf);
>  void bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns,
> -				  u64 io_start_time_ns, unsigned int op);
> +				  u64 io_start_time_ns, blk_opf_t opf);
>  void bfqg_stats_update_dequeue(struct bfq_group *bfqg);
>  void bfqg_stats_set_start_empty_time(struct bfq_group *bfqg);
>  void bfqg_stats_update_idle_time(struct bfq_group *bfqg);
-- 
Jan Kara <jack@xxxxxxxx>
SUSE Labs, CR



[Index of Archives]     [Linux RAID]     [Linux SCSI]     [Linux ATA RAID]     [IDE]     [Linux Wireless]     [Linux Kernel]     [ATH6KL]     [Linux Bluetooth]     [Linux Netdev]     [Kernel Newbies]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Device Mapper]

  Powered by Linux