Re: [PATCH] xfs: use dedicated log worker wq to avoid deadlock with cil wq

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Mar 28, 2017 at 08:01:33AM -0700, Darrick J. Wong wrote:
> On Tue, Mar 28, 2017 at 08:51:58AM -0400, Brian Foster wrote:
> > On Thu, Mar 16, 2017 at 12:17:34PM -0700, Darrick J. Wong wrote:
> > > On Thu, Mar 09, 2017 at 01:18:17PM -0500, Brian Foster wrote:
> > > > The log covering background task used to be part of the xfssyncd
> > > > workqueue. That workqueue was removed as of commit 5889608df ("xfs:
> > > > syncd workqueue is no more") and the associated work item scheduled
> > > > to the xfs-log wq. The latter is used for log buffer I/O completion.
> > > > 
> > > > Since xfs_log_worker() can invoke a log flush, a deadlock is
> > > > possible between the xfs-log and xfs-cil workqueues. Consider the
> > > > following codepath from xfs_log_worker():
> > > > 
> > > > xfs_log_worker()
> > > >   xfs_log_force()
> > > >     _xfs_log_force()
> > > >       xlog_cil_force()
> > > >         xlog_cil_force_lsn()
> > > >           xlog_cil_push_now()
> > > >             flush_work()
> > > > 
> > > > The above is in xfs-log wq context and blocked waiting on the
> > > > completion of an xfs-cil work item. Concurrently, the cil push in
> > > > progress can end up blocked here:
> > > > 
> > > > xlog_cil_push_work()
> > > >   xlog_cil_push()
> > > >     xlog_write()
> > > >       xlog_state_get_iclog_space()
> > > >         xlog_wait(&log->l_flush_wait, ...)
> > > > 
> > > > The above is in xfs-cil context waiting on log buffer I/O
> > > > completion, which executes in xfs-log wq context. In this scenario
> > > > both workqueues are deadlocked waiting on eachother.
> > > > 
> > > > Add a new workqueue specifically for the high level log covering and
> > > > ail pushing worker, as was the case prior to commit 5889608df.
> > > > 
> > > > Diagnosed-by: David Jeffery <djeffery@xxxxxxxxxx>
> > > > Signed-off-by: Brian Foster <bfoster@xxxxxxxxxx>
> > > > ---
> > > > 
> > > > Note that this seems difficult to reproduce in practice because I
> > > > believe it also relies on memory pressure. Otherwise, the xfs-log wq
> > > > rescuer thread may be available to unwind the deadlock.
> > > > 
> > > > Given that, one alternative approach I considered here is to use the
> > > > rescuer thread check brought up in the other thread[1] regarding
> > > > WQ_MEM_RECLAIM. I opted for this approach because I think the former is
> > > > brittle in that it assumes context within the xfs_log_worker() code,
> > > > adds non-determinism to the overall job of the background worker, and
> > > > also doesn't address the fact that it's probably unnecessary to run this
> > > > code at high priority.
> > > > 
> > > > Brian
> > > > 
> > > > [1] http://www.spinics.net/lists/linux-xfs/msg04673.html
> > > > 
> > > >  fs/xfs/xfs_log.c   | 2 +-
> > > >  fs/xfs/xfs_mount.h | 1 +
> > > >  fs/xfs/xfs_super.c | 8 ++++++++
> > > >  3 files changed, 10 insertions(+), 1 deletion(-)
> > > > 
> > > > diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c
> > > > index b1469f0..bb58cd1 100644
> > > > --- a/fs/xfs/xfs_log.c
> > > > +++ b/fs/xfs/xfs_log.c
> > > > @@ -1293,7 +1293,7 @@ void
> > > >  xfs_log_work_queue(
> > > >  	struct xfs_mount        *mp)
> > > >  {
> > > > -	queue_delayed_work(mp->m_log_workqueue, &mp->m_log->l_work,
> > > > +	queue_delayed_work(mp->m_sync_workqueue, &mp->m_log->l_work,
> > > >  				msecs_to_jiffies(xfs_syncd_centisecs * 10));
> > > 
> > > I /think/ this looks ok?  But let me run it through xfstests before I
> > > commit to anything. :)
> > > 
> > 
> > Ping... just checking in on this one. Is this in the pipe..? Thanks.
> 
> Tested ok; do you want me to put this in for 4.11 or save it for 4.12?
> 

Thanks. No preference, though I guess that means 4.12 is probably
appropriate. Does that mean this would fall into for-next soon-ish? I'm
basically just looking to land it somewhere upstream so we have a commit
id and we can backport it downstream for customers that hit the problem.

Brian

> --D
> 
> > 
> > Brian
> > 
> > > --D
> > > 
> > > >  }
> > > >  
> > > > diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h
> > > > index 6db6fd6..22b2185 100644
> > > > --- a/fs/xfs/xfs_mount.h
> > > > +++ b/fs/xfs/xfs_mount.h
> > > > @@ -183,6 +183,7 @@ typedef struct xfs_mount {
> > > >  	struct workqueue_struct	*m_reclaim_workqueue;
> > > >  	struct workqueue_struct	*m_log_workqueue;
> > > >  	struct workqueue_struct *m_eofblocks_workqueue;
> > > > +	struct workqueue_struct	*m_sync_workqueue;
> > > >  
> > > >  	/*
> > > >  	 * Generation of the filesysyem layout.  This is incremented by each
> > > > diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c
> > > > index 890862f..4bad410 100644
> > > > --- a/fs/xfs/xfs_super.c
> > > > +++ b/fs/xfs/xfs_super.c
> > > > @@ -877,8 +877,15 @@ xfs_init_mount_workqueues(
> > > >  	if (!mp->m_eofblocks_workqueue)
> > > >  		goto out_destroy_log;
> > > >  
> > > > +	mp->m_sync_workqueue = alloc_workqueue("xfs-sync/%s", WQ_FREEZABLE, 0,
> > > > +					       mp->m_fsname);
> > > > +	if (!mp->m_sync_workqueue)
> > > > +		goto out_destroy_eofb;
> > > > +
> > > >  	return 0;
> > > >  
> > > > +out_destroy_eofb:
> > > > +	destroy_workqueue(mp->m_eofblocks_workqueue);
> > > >  out_destroy_log:
> > > >  	destroy_workqueue(mp->m_log_workqueue);
> > > >  out_destroy_reclaim:
> > > > @@ -899,6 +906,7 @@ STATIC void
> > > >  xfs_destroy_mount_workqueues(
> > > >  	struct xfs_mount	*mp)
> > > >  {
> > > > +	destroy_workqueue(mp->m_sync_workqueue);
> > > >  	destroy_workqueue(mp->m_eofblocks_workqueue);
> > > >  	destroy_workqueue(mp->m_log_workqueue);
> > > >  	destroy_workqueue(mp->m_reclaim_workqueue);
> > > > -- 
> > > > 2.7.4
> > > > 
> > > > --
> > > > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [XFS Filesystem Development (older mail)]     [Linux Filesystem Development]     [Linux Audio Users]     [Yosemite Trails]     [Linux Kernel]     [Linux RAID]     [Linux SCSI]


  Powered by Linux