[PATCH v2 06/11] md: avoid conflicts in active_aligned_reads operations

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This patch depends on patch 03.

Change the type of active_aligned_reads from atomic_t to percpu_ref in percpu mode.

Signed-off-by: Keisuke TADA <keisuke1.tada@xxxxxxxxxx>
Signed-off-by: Toshifumi OHTAKE <toshifumi.ootake@xxxxxxxxxx>
---
 drivers/md/raid5.c | 27 +++++++++++++++++++--------
 drivers/md/raid5.h | 13 ++++---------
 2 files changed, 23 insertions(+), 17 deletions(-)

diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c
index b7056854375d..f267856efb48 100644
--- a/drivers/md/raid5.c
+++ b/drivers/md/raid5.c
@@ -5435,8 +5435,7 @@ static void raid5_align_endio(struct bio *bi)
 
 	if (!error) {
 		bio_endio(raid_bi);
-		if (active_aligned_reads_dec_and_test(conf))
-			wake_up(&conf->wait_for_quiescent);
+		active_aligned_reads_dec(conf);
 		return;
 	}
 
@@ -5507,8 +5506,8 @@ static int raid5_read_one_chunk(struct mddev *mddev, struct bio *raid_bio)
 		/* quiesce is in progress, so we need to undo io activation and wait
 		 * for it to finish
 		 */
-		if (did_inc && active_aligned_reads_dec_and_test(conf))
-			wake_up(&conf->wait_for_quiescent);
+		if (did_inc)
+			active_aligned_reads_dec(conf);
 		spin_lock_irq(&conf->device_lock);
 		wait_event_lock_irq(conf->wait_for_quiescent, conf->quiesce == 0,
 				    conf->device_lock);
@@ -6608,8 +6607,7 @@ static int  retry_aligned_read(struct r5conf *conf, struct bio *raid_bio,
 
 	bio_endio(raid_bio);
 
-	if (active_aligned_reads_dec_and_test(conf))
-		wake_up(&conf->wait_for_quiescent);
+	active_aligned_reads_dec(conf);
 	return handled;
 }
 
@@ -7315,6 +7313,7 @@ static void free_conf(struct r5conf *conf)
 {
 	int i;
 
+	percpu_ref_exit(&conf->active_aligned_reads);
 	log_exit(conf);
 
 	shrinker_free(conf->shrinker);
@@ -7396,6 +7395,12 @@ static unsigned long raid5_cache_count(struct shrinker *shrink,
 	return max_stripes - min_stripes;
 }
 
+static void percpu_wakeup_handle_req_active(struct percpu_ref *r)
+{
+	struct r5conf *conf = container_of(r, struct r5conf, active_aligned_reads);
+
+	wake_up(&conf->wait_for_quiescent);
+}
 
 static struct r5conf *setup_conf(struct mddev *mddev, bool quiesce)
 {
@@ -7483,7 +7488,11 @@ static struct r5conf *setup_conf(struct mddev *mddev, bool quiesce)
 		percpu_ref_init_flags = PERCPU_REF_ALLOW_REINIT | PERCPU_REF_INIT_DEAD;
 	else
 		percpu_ref_init_flags = PERCPU_REF_ALLOW_REINIT;
-	atomic_set(&conf->active_aligned_reads, 0);
+	ret = percpu_ref_init(&conf->active_aligned_reads, percpu_wakeup_handle_req_active,
+		percpu_ref_init_flags, GFP_KERNEL);
+	if (ret)
+		goto abort;
+	ret = -ENOMEM;
 	spin_lock_init(&conf->pending_bios_lock);
 	conf->batch_bio_dispatch = true;
 	rdev_for_each(rdev, mddev) {
@@ -7675,7 +7684,7 @@ static struct r5conf *setup_conf_for_takeover(struct mddev *mddev)
 
 	if (mddev->level == 4 || mddev->level == 5 || mddev->level == 6) {
 		conf = mddev->private;
-		quiesce = false;
+		quiesce = percpu_ref_is_dying(&conf->active_aligned_reads);
 	}
 	return setup_conf(mddev, quiesce);
 }
@@ -8607,6 +8616,7 @@ static void raid5_quiesce(struct mddev *mddev, int quiesce)
 		 * quiesce started and reverts to slow (locked) path.
 		 */
 		smp_store_release(&conf->quiesce, 2);
+		percpu_ref_kill(&conf->active_aligned_reads);
 		wait_event_cmd(conf->wait_for_quiescent,
 				    atomic_read(&conf->active_stripes) == 0 &&
 				    active_aligned_reads_is_zero(conf),
@@ -8619,6 +8629,7 @@ static void raid5_quiesce(struct mddev *mddev, int quiesce)
 	} else {
 		/* re-enable writes */
 		lock_all_device_hash_locks_irq(conf);
+		percpu_ref_reinit(&conf->active_aligned_reads);
 		conf->quiesce = 0;
 		wake_up(&conf->wait_for_quiescent);
 		wake_up(&conf->wait_for_overlap);
diff --git a/drivers/md/raid5.h b/drivers/md/raid5.h
index 5bd6bb3540c5..c4ab418e2084 100644
--- a/drivers/md/raid5.h
+++ b/drivers/md/raid5.h
@@ -621,7 +621,7 @@ struct r5conf {
 	unsigned int		retry_read_offset; /* sector offset into retry_read_aligned */
 	struct bio		*retry_read_aligned_list; /* aligned bios retry list  */
 	atomic_t		preread_active_stripes; /* stripes with scheduled io */
-	atomic_t		active_aligned_reads;
+	struct percpu_ref	active_aligned_reads;
 	atomic_t		pending_full_writes; /* full write backlog */
 	int			bypass_count; /* bypassed prereads */
 	int			bypass_threshold; /* preread nice */
@@ -698,22 +698,17 @@ struct r5conf {
 
 static inline void active_aligned_reads_inc(struct r5conf *conf)
 {
-	atomic_inc(&conf->active_aligned_reads);
+	percpu_ref_get(&conf->active_aligned_reads);
 }
 
 static inline void active_aligned_reads_dec(struct r5conf *conf)
 {
-	atomic_dec(&conf->active_aligned_reads);
+	percpu_ref_put(&conf->active_aligned_reads);
 }
 
 static inline bool active_aligned_reads_is_zero(struct r5conf *conf)
 {
-	return atomic_read(&conf->active_aligned_reads) == 0;
-}
-
-static inline bool active_aligned_reads_dec_and_test(struct r5conf *conf)
-{
-	return atomic_dec_and_test(&conf->active_aligned_reads);
+	return percpu_ref_is_zero(&conf->active_aligned_reads);
 }
 
 #if PAGE_SIZE == DEFAULT_STRIPE_SIZE
-- 
2.34.1







[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux