Re: Stuck in md_write_start because MD_SB_CHANGE_PENDING can't be cleared

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




----- Original Message -----
> From: "NeilBrown" <neilb@xxxxxxxx>
> To: "Xiao Ni" <xni@xxxxxxxxxx>, "linux-raid" <linux-raid@xxxxxxxxxxxxxxx>
> Cc: shli@xxxxxxxxxx
> Sent: Thursday, September 7, 2017 1:37:45 PM
> Subject: Re: Stuck in md_write_start because MD_SB_CHANGE_PENDING can't be cleared
> 
> On Wed, Sep 06 2017, Xiao Ni wrote:
> 
> > ----- Original Message -----
> >> From: "Xiao Ni" <xni@xxxxxxxxxx>
> >> To: "NeilBrown" <neilb@xxxxxxxx>, "linux-raid"
> >> <linux-raid@xxxxxxxxxxxxxxx>
> >> Cc: shli@xxxxxxxxxx
> >> Sent: Tuesday, September 5, 2017 10:15:00 AM
> >> Subject: Re: Stuck in md_write_start because MD_SB_CHANGE_PENDING can't be
> >> cleared
> >> 
> >> 
> >> 
> >> On 09/05/2017 09:36 AM, NeilBrown wrote:
> >> > On Mon, Sep 04 2017, Xiao Ni wrote:
> >> >
> >> >>
> >> >> In function handle_stripe:
> >> >> 4697         if (s.handle_bad_blocks ||
> >> >> 4698             test_bit(MD_SB_CHANGE_PENDING,
> >> >> &conf->mddev->sb_flags)) {
> >> >> 4699                 set_bit(STRIPE_HANDLE, &sh->state);
> >> >> 4700                 goto finish;
> >> >> 4701         }
> >> >>
> >> >> Because MD_SB_CHANGE_PENDING is set, so the stripes can't be handled.
> >> >>
> >> > Right, of course.  I see what is happening now.
> >> >
> >> > - raid5d cannot complete stripes until the metadata is written
> >> > - the metadata cannot be written until raid5d gets the mddev_lock
> >> > - mddev_lock is held by the write to suspend_hi
> >> > - the write to suspend_hi is waiting for raid5_quiesce
> >> > - raid5_quiesce is waiting for some stripes to complete.
> >> >
> >> > We could declare that ->quiesce(, 1) cannot be called while holding the
> >> > lock.
> >> > We could possible allow it but only if md_update_sb() is called first,
> >> > though that might still be racy.
> >> >
> >> > ->quiesce(, 1) is currently called from:
> >> >   mddev_suspend
> >> >   suspend_lo_store
> >> >   suspend_hi_store
> >> >   __md_stop_writes
> >> >   mddev_detach
> >> >   set_bitmap_file
> >> >   update_array_info (when setting/removing internal bitmap)
> >> >   md_do_sync
> >> >
> >> > and most of those are call with the lock held, or take the lock.
> >> >
> >> > Maybe we should *require* that mddev_lock is held when calling
> >> > ->quiesce() and have ->quiesce() do the metadata update.
> >> >
> >> > Something like the following maybe.  Can you test it?
> >> 
> >> Hi Neil
> >> 
> >> Thanks for the analysis. I need to thing for a while :)
> >> I already added the patch and the test is running now. It usually needs
> >> more than 5
> >> hours to reproduce this problem. I'll let it run more than 24 hours.
> >> I'll update the test
> >> result later.
> >
> > Hi Neil
> >
> > The problem still exists. But it doesn't show calltrace this time. It
> > was stuck yesterday. I didn't notice that because there has no calltrace.
> >
> > echo file raid5.c +p > /sys/kernel/debug/dynamic_debug/control
> >
> > It shows that raid5d is still spinning.
> 
> Thanks for testing. I've thought some more and I think there is a better
> approach.
> The fact that we need to take the mutex to write the super block has
> caused problems several times before and is a key part of the problem
> now.
> Maybe we should relax that.  Obviously we don't want two threads
> updating the metadata at the same time, but it should be safe to
> update it in parallel with other uses of reconfix_mutex.
> 
> Holding mddev->lock while copying data from the struct mddev to the
> superblock (which we do) should ensure that the superblock is internally
> consistent, and that should be enough.
> 
> So I propose the following patch.  It certainly needs review and
> testing, but I think it should make a big improvement.
> 
> Thanks,
> NeilBrown
> 
> 
> diff --git a/drivers/md/md.c b/drivers/md/md.c
> index b01e458d31e9..414a4c1a052d 100644
> --- a/drivers/md/md.c
> +++ b/drivers/md/md.c
> @@ -2388,6 +2388,15 @@ void md_update_sb(struct mddev *mddev, int
> force_change)
>  		return;
>  	}
>  
> +	if (!force_change && !(mddev->sb_flags & ~BIT(MD_SB_UPDATE_ACTIVE)))
> +		return;
> +
> +	wait_event(mddev->sb_wait,
> +		   !test_and_set_bit(MD_SB_UPDATE_ACTIVE, &mddev->sb_flags));
> +
> +	if (!force_change && !(mddev->sb_flags & ~BIT(MD_SB_UPDATE_ACTIVE)))
> +		goto out;
> +
>  repeat:
>  	if (mddev_is_clustered(mddev)) {
>  		if (test_and_clear_bit(MD_SB_CHANGE_DEVS, &mddev->sb_flags))
> @@ -2402,7 +2411,7 @@ void md_update_sb(struct mddev *mddev, int
> force_change)
>  			bit_clear_unless(&mddev->sb_flags, BIT(MD_SB_CHANGE_PENDING),
>  							 BIT(MD_SB_CHANGE_DEVS) |
>  							 BIT(MD_SB_CHANGE_CLEAN));
> -			return;
> +			goto out;
>  		}
>  	}
>  
> @@ -2432,8 +2441,7 @@ void md_update_sb(struct mddev *mddev, int
> force_change)
>  				wake_up(&rdev->blocked_wait);
>  			}
>  		}
> -		wake_up(&mddev->sb_wait);
> -		return;
> +		goto out;
>  	}
>  
>  	spin_lock(&mddev->lock);
> @@ -2544,6 +2552,9 @@ void md_update_sb(struct mddev *mddev, int
> force_change)
>  			       BIT(MD_SB_CHANGE_DEVS) | BIT(MD_SB_CHANGE_CLEAN)))
>  		/* have to write it out again */
>  		goto repeat;
> +
> +out:
> +	clear_bit_unlock(MD_SB_UPDATE_ACTIVE, &mddev->sb_flags);
>  	wake_up(&mddev->sb_wait);
>  	if (test_bit(MD_RECOVERY_RUNNING, &mddev->recovery))
>  		sysfs_notify(&mddev->kobj, NULL, "sync_completed");
> @@ -5606,8 +5617,7 @@ int md_run(struct mddev *mddev)
>  		set_bit(MD_RECOVERY_RECOVER, &mddev->recovery);
>  	set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
>  
> -	if (mddev->sb_flags)
> -		md_update_sb(mddev, 0);
> +	md_update_sb(mddev, 0);
>  
>  	md_new_event(mddev);
>  	sysfs_notify_dirent_safe(mddev->sysfs_state);
> @@ -8643,17 +8653,14 @@ void md_check_recovery(struct mddev *mddev)
>  
>  	if (mddev->ro && !test_bit(MD_RECOVERY_NEEDED, &mddev->recovery))
>  		return;
> -	if ( ! (
> -		(mddev->sb_flags & ~ (1<<MD_SB_CHANGE_PENDING)) ||
> +	if ((
>  		test_bit(MD_RECOVERY_NEEDED, &mddev->recovery) ||
>  		test_bit(MD_RECOVERY_DONE, &mddev->recovery) ||
>  		(mddev->external == 0 && mddev->safemode == 1) ||
>  		(mddev->safemode == 2
>  		 && !mddev->in_sync && mddev->recovery_cp == MaxSector)
> -		))
> -		return;
> -
> -	if (mddev_trylock(mddev)) {
> +		     ) &&
> +	    mddev_trylock(mddev)) {
>  		int spares = 0;
>  
>  		if (!mddev->external && mddev->safemode == 1)
> @@ -8706,9 +8713,6 @@ void md_check_recovery(struct mddev *mddev)
>  			spin_unlock(&mddev->lock);
>  		}
>  
> -		if (mddev->sb_flags)
> -			md_update_sb(mddev, 0);
> -
>  		if (test_bit(MD_RECOVERY_RUNNING, &mddev->recovery) &&
>  		    !test_bit(MD_RECOVERY_DONE, &mddev->recovery)) {
>  			/* resync/recovery still happening */
> @@ -8786,6 +8790,7 @@ void md_check_recovery(struct mddev *mddev)
>  		wake_up(&mddev->sb_wait);
>  		mddev_unlock(mddev);
>  	}
> +	md_update_sb(mddev, 0);
>  }
>  EXPORT_SYMBOL(md_check_recovery);
>  
> diff --git a/drivers/md/md.h b/drivers/md/md.h
> index 09db03455801..bc8633cf7c40 100644
> --- a/drivers/md/md.h
> +++ b/drivers/md/md.h
> @@ -243,6 +243,7 @@ enum mddev_sb_flags {
>  	MD_SB_CHANGE_CLEAN,	/* transition to or from 'clean' */
>  	MD_SB_CHANGE_PENDING,	/* switch from 'clean' to 'active' in progress */
>  	MD_SB_NEED_REWRITE,	/* metadata write needs to be repeated */
> +	MD_SB_UPDATE_ACTIVE,	/* A thread is running in md_update_sb */
>  };
>  
>  struct mddev {
> diff --git a/drivers/md/raid5-cache.c b/drivers/md/raid5-cache.c
> index 2dcbafa8e66c..76169dd8ff7c 100644
> --- a/drivers/md/raid5-cache.c
> +++ b/drivers/md/raid5-cache.c
> @@ -1334,21 +1334,10 @@ static void r5l_write_super_and_discard_space(struct
> r5l_log *log,
>  	mddev = log->rdev->mddev;
>  	/*
>  	 * Discard could zero data, so before discard we must make sure
> -	 * superblock is updated to new log tail. Updating superblock (either
> -	 * directly call md_update_sb() or depend on md thread) must hold
> -	 * reconfig mutex. On the other hand, raid5_quiesce is called with
> -	 * reconfig_mutex hold. The first step of raid5_quiesce() is waitting
> -	 * for all IO finish, hence waitting for reclaim thread, while reclaim
> -	 * thread is calling this function and waitting for reconfig mutex. So
> -	 * there is a deadlock. We workaround this issue with a trylock.
> -	 * FIXME: we could miss discard if we can't take reconfig mutex
> +	 * superblock is updated to new log tail.
>  	 */
> -	set_mask_bits(&mddev->sb_flags, 0,
> -		BIT(MD_SB_CHANGE_DEVS) | BIT(MD_SB_CHANGE_PENDING));
> -	if (!mddev_trylock(mddev))
> -		return;
> +
>  	md_update_sb(mddev, 1);
> -	mddev_unlock(mddev);
>  
>  	/* discard IO error really doesn't matter, ignore it */
>  	if (log->last_checkpoint < end) {
> 

Hi Neil

The test have run for three days and the problem is fixed by this patch. 
Thanks for the help.

Could you help to look at https://www.spinics.net/lists/raid/msg58918.html.
The bug which is fixed by your patch was found when I try to reproduce that
bug. I did a simply analysis, but I'm not sure whether I'm right or not.

Regards
Xiao
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux