Fix some random typos. Signed-off-by: Yang Hau <yuanyanghau@xxxxxxxxx> --- block/bdev.c | 2 +- block/bfq-iosched.c | 4 ++-- block/blk-cgroup-rwstat.c | 2 +- block/blk-ia-ranges.c | 2 +- block/blk-merge.c | 6 +++--- block/blk.h | 2 +- block/elevator.c | 2 +- block/kyber-iosched.c | 2 +- block/mq-deadline.c | 2 +- 9 files changed, 12 insertions(+), 12 deletions(-) diff --git a/block/bdev.c b/block/bdev.c index 1795c7d4b..6f8bff116 100644 --- a/block/bdev.c +++ b/block/bdev.c @@ -539,7 +539,7 @@ static void bd_clear_claiming(struct block_device *whole, void *holder) * @bdev: block device of interest * @holder: holder that has claimed @bdev * - * Finish exclusive open of a block device. Mark the device as exlusively + * Finish exclusive open of a block device. Mark the device as exclusively * open by the holder and wake up all waiters for exclusive open to finish. */ static void bd_finish_claiming(struct block_device *bdev, void *holder) diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c index d9ed3108c..3adb03633 100644 --- a/block/bfq-iosched.c +++ b/block/bfq-iosched.c @@ -1963,7 +1963,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd, * As for throughput, we ask bfq_better_to_idle() whether we * still need to plug I/O dispatching. If bfq_better_to_idle() * says no, then plugging is not needed any longer, either to - * boost throughput or to perserve service guarantees. Then + * boost throughput or to preserve service guarantees. Then * the best option is to stop plugging I/O, as not doing so * would certainly lower throughput. We may end up in this * case if: (1) upon a dispatch attempt, we detected that it @@ -2491,7 +2491,7 @@ static int bfq_request_merge(struct request_queue *q, struct request **req, if (__rq && elv_bio_merge_ok(__rq, bio)) { *req = __rq; - if (blk_discard_mergable(__rq)) + if (blk_discard_mergeable(__rq)) return ELEVATOR_DISCARD_MERGE; return ELEVATOR_FRONT_MERGE; } diff --git a/block/blk-cgroup-rwstat.c b/block/blk-cgroup-rwstat.c index 3304e841d..0039e4756 100644 --- a/block/blk-cgroup-rwstat.c +++ b/block/blk-cgroup-rwstat.c @@ -37,7 +37,7 @@ EXPORT_SYMBOL_GPL(blkg_rwstat_exit); * @pd: policy private data of interest * @rwstat: rwstat to print * - * Print @rwstat to @sf for the device assocaited with @pd. + * Print @rwstat to @sf for the device associated with @pd. */ u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd, const struct blkg_rwstat_sample *rwstat) diff --git a/block/blk-ia-ranges.c b/block/blk-ia-ranges.c index c9eb4241e..f65f7a2cf 100644 --- a/block/blk-ia-ranges.c +++ b/block/blk-ia-ranges.c @@ -63,7 +63,7 @@ static const struct sysfs_ops blk_ia_range_sysfs_ops = { }; /* - * Independent access range entries are not freed individually, but alltogether + * Independent access range entries are not freed individually, but all together * with struct blk_independent_access_ranges and its array of ranges. Since * kobject_add() takes a reference on the parent kobject contained in * struct blk_independent_access_ranges, the array of independent access range diff --git a/block/blk-merge.c b/block/blk-merge.c index 6460abdb2..29034bb9d 100644 --- a/block/blk-merge.c +++ b/block/blk-merge.c @@ -790,7 +790,7 @@ static void blk_account_io_merge_request(struct request *req) static enum elv_merge blk_try_req_merge(struct request *req, struct request *next) { - if (blk_discard_mergable(req)) + if (blk_discard_mergeable(req)) return ELEVATOR_DISCARD_MERGE; else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next)) return ELEVATOR_BACK_MERGE; @@ -864,7 +864,7 @@ static struct request *attempt_merge(struct request_queue *q, req->__data_len += blk_rq_bytes(next); - if (!blk_discard_mergable(req)) + if (!blk_discard_mergeable(req)) elv_merge_requests(q, req, next); /* @@ -947,7 +947,7 @@ bool blk_rq_merge_ok(struct request *rq, struct bio *bio) enum elv_merge blk_try_merge(struct request *rq, struct bio *bio) { - if (blk_discard_mergable(rq)) + if (blk_discard_mergeable(rq)) return ELEVATOR_DISCARD_MERGE; else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector) return ELEVATOR_BACK_MERGE; diff --git a/block/blk.h b/block/blk.h index cc4e8873d..eb2ae2df1 100644 --- a/block/blk.h +++ b/block/blk.h @@ -148,7 +148,7 @@ static inline bool rq_mergeable(struct request *rq) * 2) Otherwise, the request will be normal read/write requests. The ranges * need to be contiguous. */ -static inline bool blk_discard_mergable(struct request *req) +static inline bool blk_discard_mergeable(struct request *req) { if (req_op(req) == REQ_OP_DISCARD && queue_max_discard_segments(req->q) > 1) diff --git a/block/elevator.c b/block/elevator.c index 24909069f..a6ab90cff 100644 --- a/block/elevator.c +++ b/block/elevator.c @@ -310,7 +310,7 @@ enum elv_merge elv_merge(struct request_queue *q, struct request **req, if (__rq && elv_bio_merge_ok(__rq, bio)) { *req = __rq; - if (blk_discard_mergable(__rq)) + if (blk_discard_mergeable(__rq)) return ELEVATOR_DISCARD_MERGE; return ELEVATOR_BACK_MERGE; } diff --git a/block/kyber-iosched.c b/block/kyber-iosched.c index 214696923..678448cf9 100644 --- a/block/kyber-iosched.c +++ b/block/kyber-iosched.c @@ -142,7 +142,7 @@ struct kyber_cpu_latency { */ struct kyber_ctx_queue { /* - * Used to ensure operations on rq_list and kcq_map to be an atmoic one. + * Used to ensure operations on rq_list and kcq_map to be an atomic one. * Also protect the rqs on rq_list when merge. */ spinlock_t lock; diff --git a/block/mq-deadline.c b/block/mq-deadline.c index f10c2a0d1..910a23ef4 100644 --- a/block/mq-deadline.c +++ b/block/mq-deadline.c @@ -734,7 +734,7 @@ static int dd_request_merge(struct request_queue *q, struct request **rq, if (elv_bio_merge_ok(__rq, bio)) { *rq = __rq; - if (blk_discard_mergable(__rq)) + if (blk_discard_mergeable(__rq)) return ELEVATOR_DISCARD_MERGE; return ELEVATOR_FRONT_MERGE; } -- 2.39.0