[PATCH 1/2] block: fix random typos

From: Yang Hau
Date: Sun Jun 11 2023 - 21:33:36 EST


Fix some random typos.

Signed-off-by: Yang Hau <yuanyanghau@xxxxxxxxx>
---
block/bdev.c | 2 +-
block/bfq-iosched.c | 4 ++--
block/blk-cgroup-rwstat.c | 2 +-
block/blk-ia-ranges.c | 2 +-
block/blk-merge.c | 6 +++---
block/blk.h | 2 +-
block/elevator.c | 2 +-
block/kyber-iosched.c | 2 +-
block/mq-deadline.c | 2 +-
9 files changed, 12 insertions(+), 12 deletions(-)

diff --git a/block/bdev.c b/block/bdev.c
index 1795c7d4b..6f8bff116 100644
--- a/block/bdev.c
+++ b/block/bdev.c
@@ -539,7 +539,7 @@ static void bd_clear_claiming(struct block_device *whole, void *holder)
* @bdev: block device of interest
* @holder: holder that has claimed @bdev
*
- * Finish exclusive open of a block device. Mark the device as exlusively
+ * Finish exclusive open of a block device. Mark the device as exclusively
* open by the holder and wake up all waiters for exclusive open to finish.
*/
static void bd_finish_claiming(struct block_device *bdev, void *holder)
diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
index d9ed3108c..3adb03633 100644
--- a/block/bfq-iosched.c
+++ b/block/bfq-iosched.c
@@ -1963,7 +1963,7 @@ static void bfq_bfqq_handle_idle_busy_switch(struct bfq_data *bfqd,
* As for throughput, we ask bfq_better_to_idle() whether we
* still need to plug I/O dispatching. If bfq_better_to_idle()
* says no, then plugging is not needed any longer, either to
- * boost throughput or to perserve service guarantees. Then
+ * boost throughput or to preserve service guarantees. Then
* the best option is to stop plugging I/O, as not doing so
* would certainly lower throughput. We may end up in this
* case if: (1) upon a dispatch attempt, we detected that it
@@ -2491,7 +2491,7 @@ static int bfq_request_merge(struct request_queue *q, struct request **req,
if (__rq && elv_bio_merge_ok(__rq, bio)) {
*req = __rq;

- if (blk_discard_mergable(__rq))
+ if (blk_discard_mergeable(__rq))
return ELEVATOR_DISCARD_MERGE;
return ELEVATOR_FRONT_MERGE;
}
diff --git a/block/blk-cgroup-rwstat.c b/block/blk-cgroup-rwstat.c
index 3304e841d..0039e4756 100644
--- a/block/blk-cgroup-rwstat.c
+++ b/block/blk-cgroup-rwstat.c
@@ -37,7 +37,7 @@ EXPORT_SYMBOL_GPL(blkg_rwstat_exit);
* @pd: policy private data of interest
* @rwstat: rwstat to print
*
- * Print @rwstat to @sf for the device assocaited with @pd.
+ * Print @rwstat to @sf for the device associated with @pd.
*/
u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd,
const struct blkg_rwstat_sample *rwstat)
diff --git a/block/blk-ia-ranges.c b/block/blk-ia-ranges.c
index c9eb4241e..f65f7a2cf 100644
--- a/block/blk-ia-ranges.c
+++ b/block/blk-ia-ranges.c
@@ -63,7 +63,7 @@ static const struct sysfs_ops blk_ia_range_sysfs_ops = {
};

/*
- * Independent access range entries are not freed individually, but alltogether
+ * Independent access range entries are not freed individually, but all together
* with struct blk_independent_access_ranges and its array of ranges. Since
* kobject_add() takes a reference on the parent kobject contained in
* struct blk_independent_access_ranges, the array of independent access range
diff --git a/block/blk-merge.c b/block/blk-merge.c
index 6460abdb2..29034bb9d 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -790,7 +790,7 @@ static void blk_account_io_merge_request(struct request *req)
static enum elv_merge blk_try_req_merge(struct request *req,
struct request *next)
{
- if (blk_discard_mergable(req))
+ if (blk_discard_mergeable(req))
return ELEVATOR_DISCARD_MERGE;
else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next))
return ELEVATOR_BACK_MERGE;
@@ -864,7 +864,7 @@ static struct request *attempt_merge(struct request_queue *q,

req->__data_len += blk_rq_bytes(next);

- if (!blk_discard_mergable(req))
+ if (!blk_discard_mergeable(req))
elv_merge_requests(q, req, next);

/*
@@ -947,7 +947,7 @@ bool blk_rq_merge_ok(struct request *rq, struct bio *bio)

enum elv_merge blk_try_merge(struct request *rq, struct bio *bio)
{
- if (blk_discard_mergable(rq))
+ if (blk_discard_mergeable(rq))
return ELEVATOR_DISCARD_MERGE;
else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector)
return ELEVATOR_BACK_MERGE;
diff --git a/block/blk.h b/block/blk.h
index cc4e8873d..eb2ae2df1 100644
--- a/block/blk.h
+++ b/block/blk.h
@@ -148,7 +148,7 @@ static inline bool rq_mergeable(struct request *rq)
* 2) Otherwise, the request will be normal read/write requests. The ranges
* need to be contiguous.
*/
-static inline bool blk_discard_mergable(struct request *req)
+static inline bool blk_discard_mergeable(struct request *req)
{
if (req_op(req) == REQ_OP_DISCARD &&
queue_max_discard_segments(req->q) > 1)
diff --git a/block/elevator.c b/block/elevator.c
index 24909069f..a6ab90cff 100644
--- a/block/elevator.c
+++ b/block/elevator.c
@@ -310,7 +310,7 @@ enum elv_merge elv_merge(struct request_queue *q, struct request **req,
if (__rq && elv_bio_merge_ok(__rq, bio)) {
*req = __rq;

- if (blk_discard_mergable(__rq))
+ if (blk_discard_mergeable(__rq))
return ELEVATOR_DISCARD_MERGE;
return ELEVATOR_BACK_MERGE;
}
diff --git a/block/kyber-iosched.c b/block/kyber-iosched.c
index 214696923..678448cf9 100644
--- a/block/kyber-iosched.c
+++ b/block/kyber-iosched.c
@@ -142,7 +142,7 @@ struct kyber_cpu_latency {
*/
struct kyber_ctx_queue {
/*
- * Used to ensure operations on rq_list and kcq_map to be an atmoic one.
+ * Used to ensure operations on rq_list and kcq_map to be an atomic one.
* Also protect the rqs on rq_list when merge.
*/
spinlock_t lock;
diff --git a/block/mq-deadline.c b/block/mq-deadline.c
index f10c2a0d1..910a23ef4 100644
--- a/block/mq-deadline.c
+++ b/block/mq-deadline.c
@@ -734,7 +734,7 @@ static int dd_request_merge(struct request_queue *q, struct request **rq,

if (elv_bio_merge_ok(__rq, bio)) {
*rq = __rq;
- if (blk_discard_mergable(__rq))
+ if (blk_discard_mergeable(__rq))
return ELEVATOR_DISCARD_MERGE;
return ELEVATOR_FRONT_MERGE;
}
--
2.39.0