Re: [PATCH] raid1: prevent unnecessary call to wake_up() in fast path

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

在 2023/07/03 20:55, Jinpu Wang 写道:
Hi Kuai,

Thanks for your comment, see reply inline.

On Mon, Jul 3, 2023 at 2:35 PM Yu Kuai <yukuai3@xxxxxxxxxx> wrote:

Hi,

在 2023/07/03 16:01, Jack Wang 写道:
wake_up is called unconditionally in fast path such as make_request(),
which cause lock contention under high concurrency
      raid1_end_write_request
       wake_up
        __wake_up_common_lock
         spin_lock_irqsave

Improve performance by only call wake_up() if waitqueue is not empty

Fio test script:

[global]
name=random reads and writes
ioengine=libaio
direct=1
readwrite=randrw
rwmixread=70
iodepth=64
buffered=0
filename=/dev/md0
size=1G
runtime=30
time_based
randrepeat=0
norandommap
refill_buffers
ramp_time=10
bs=4k
numjobs=400
group_reporting=1
[job1]

Test result with ramdisk raid1 on a EPYC:

       Before this patch       With this patch
       READ    BW=4621MB/s     BW=7337MB/s
       WRITE   BW=1980MB/s     BW=1675MB/s
This was copy mistake, checked the raw output, with patch write BW is 3144MB/s.

will fix in next version.

will also adapt the subject with "md/raid1" prefix.



This is weird, I don't understand how write can be worse.

The patch is inspired by Yu Kuai's change for raid10:
https://lore.kernel.org/r/20230621105728.1268542-1-yukuai1@xxxxxxxxxxxxxxx

Cc: Yu Kuai <yukuai3@xxxxxxxxxx>
Signed-off-by: Jack Wang <jinpu.wang@xxxxxxxxx>
---
   drivers/md/raid1.c | 17 ++++++++++++-----
   1 file changed, 12 insertions(+), 5 deletions(-)

diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c
index f834d99a36f6..808c91f338e6 100644
--- a/drivers/md/raid1.c
+++ b/drivers/md/raid1.c
@@ -789,11 +789,17 @@ static int read_balance(struct r1conf *conf, struct r1bio *r1_bio, int *max_sect
       return best_disk;
   }

+static void wake_up_barrier(struct r1conf *conf)
+{
+     if (wq_has_sleeper(&conf->wait_barrier))
+             wake_up(&conf->wait_barrier);
+}
+
   static void flush_bio_list(struct r1conf *conf, struct bio *bio)
   {
       /* flush any pending bitmap writes to disk before proceeding w/ I/O */
       raid1_prepare_flush_writes(conf->mddev->bitmap);
-     wake_up(&conf->wait_barrier);
+     wake_up_barrier(conf);

       while (bio) { /* submit pending writes */
               struct bio *next = bio->bi_next;
@@ -835,6 +841,7 @@ static void flush_pending_writes(struct r1conf *conf)
               spin_unlock_irq(&conf->device_lock);
   }

+

Please remove this new line.
   /* Barriers....
    * Sometimes we need to suspend IO while we do something else,
    * either some resync/recovery, or reconfigure the array.
@@ -970,7 +977,7 @@ static bool _wait_barrier(struct r1conf *conf, int idx, bool nowait)
        * In case freeze_array() is waiting for
        * get_unqueued_pending() == extra
        */
-     wake_up(&conf->wait_barrier);
+     wake_up_barrier(conf);

This is not fast path, this is only called when array is frozen or
barrier is grabbed, and this is also called with 'resync_lock' held.

No, this one is call from
raid1_write_request->wait_barrier->_wait_barrier. and it can be seen
via perf.

I'm aware where this is called from, but fast path should be no forzen
and no barrier, and _wait_barrier() will return early. Otherwise,
spin_lock_irq(&conf->resync_lock) is called as well and current context
will wait for forzen/barrier, which is slow path.

If it can be seen via perf, which means the array is frozen of sync
thread is still in progress, and I don't think this change will be
helpful because there is still anther spinlock involved.

Thanks,
Kuai


       /* Wait for the barrier in same barrier unit bucket to drop. */

       /* Return false when nowait flag is set */
@@ -1013,7 +1020,7 @@ static bool wait_read_barrier(struct r1conf *conf, sector_t sector_nr, bool nowa
        * In case freeze_array() is waiting for
        * get_unqueued_pending() == extra
        */
-     wake_up(&conf->wait_barrier);
+     wake_up_barrier(conf);

same above.
No, this one is call from raid1_read_request-> wait_read_barrier, it
can be seen via perf results.

       /* Wait for array to be unfrozen */

       /* Return false when nowait flag is set */
@@ -1042,7 +1049,7 @@ static bool wait_barrier(struct r1conf *conf, sector_t sector_nr, bool nowait)
   static void _allow_barrier(struct r1conf *conf, int idx)
   {
       atomic_dec(&conf->nr_pending[idx]);
-     wake_up(&conf->wait_barrier);
+     wake_up_barrier(conf);
   }

   static void allow_barrier(struct r1conf *conf, sector_t sector_nr)
@@ -1171,7 +1178,7 @@ static void raid1_unplug(struct blk_plug_cb *cb, bool from_schedule)
               spin_lock_irq(&conf->device_lock);
               bio_list_merge(&conf->pending_bio_list, &plug->pending);
               spin_unlock_irq(&conf->device_lock);
-             wake_up(&conf->wait_barrier);
+             wake_up_barrier(conf);
               md_wakeup_thread(mddev->thread);
               kfree(plug);
               return;


And you missed raid1_write_request().
you meant this one:
1583         /* In case raid1d snuck in to freeze_array */
1584         wake_up(&conf->wait_barrier);
1585 }

perf result doesn't show it, I will add it too.


Thanks,
Kuai

Thx!

.




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux