On Thu, Mar 03, 2011 at 10:34:10AM -0500, Christoph Hellwig wrote: > I still don't see any point in having the ENOSPC flushing moved to a > different context. IIRC, stack usage has always been an issue, and we also call xfs_flush_inodes() with the XFS_IOLOCK held (from xfs_iomap_write_delay()) so the alternate context was used to avoid deadlocks. I don't think we have that deadlock problem now thanks to being able to combine SYNC_TRYLOCK | SYNC_WAIT flags, but I'm not sure we can ignore the stack issues. > Just add a mutex and flush inline, e.g. > > void > xfs_flush_inodes( > struct xfs_inode *ip) > { > struct xfs_mount *mp = ip->i_mount; > > if (!mutex_trylock(&xfs_syncd_lock)) > return; /* someone else is flushing right now */ > xfs_sync_data(mp, SYNC_TRYLOCK); > xfs_sync_data(mp, SYNC_TRYLOCK | SYNC_WAIT); > xfs_log_force(mp, XFS_LOG_SYNC); > mutex_unlock(&xfs_syncd_lock); > } This doesn't allow all the concurrent flushes to block on the flush in progress. i.e. if there is a flush in progress, all the others will simply return an likely get ENOSPC again because they haven't waited for any potential space to be freed up. It also realy requires a per-filesystem mutex, not a global mutex, because we don't wan't to avoid flushing filesystem X because filesystem Y is currently flushing. Yes, I could play tricks when the trylock case fails, but I'd prefer to leave it as a work item because then all the concurrent flushers all block on the same work item and it is clear from the stack traces what they are all waiting on. I've also realised the work_pending() check is unnecessary, as is the lock, because queue_work() will only queue new work if the work item isn't already pending so there's no need to check it here. Hence all this actually needs to do is: queue_work() flush_work_sync() Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx _______________________________________________ xfs mailing list xfs@xxxxxxxxxxx http://oss.sgi.com/mailman/listinfo/xfs