Re: [PATCH 2/2] repair: don't grind CPUs with large extent lists

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, May 08, 2014 at 10:01:37PM -0500, Eric Sandeen wrote:
> On 5/8/14, 8:17 PM, Dave Chinner wrote:
> > From: Dave Chinner <dchinner@xxxxxxxxxx>
> > 
> > When repairing a large filesystem with fragemented files, xfs_repair
> > can grind to a halt burning multiple CPUs in blkmap_set_ext().
> > blkmap_set_ext() inserts extents into the blockmap for the inode
> > fork and it keeps them in order, even if the inserts are not done in
> > order.
> > 
> > The ordered insert is highly inefficient - it starts at the first
> > extent, and simple walks the array to find the insertion point. i.e.
> > it is an O(n) operation. When we have a fragemented file with a
> > large number of extents, the cost of the entire mapping operation
> > is rather costly.
> > 
> > The thing is, we are doing the insertion from an *ordered btree
> > scan* which is inserting the extents in ascending offset order.
> > IOWs, we are always inserting the extent at the end of the array
> > after searching the entire array. i.e. the mapping operation cost is
> > O(N^2).
> > 
> > Fix this simply by reversing the order of the insert slot search.
> > Start at the end of the blockmap array when we do almost all
> > insertions, which brings the overhead of each insertion down to O(1)
> > complexity. This, in turn, results in the overall map building
> > operation being reduced to an O(N) operation, and so performance
> > degrades linearly with increasing extent counts rather than
> > exponentially.
> > 
> > The result is that the test filesystem (27TB, 30M inodes, at ENOSPC)
> > takes 5m10s to *fully repair* on my test system, rather that getting
> > 15 (of 60) AGs into phase three and sitting there burning 3-4 CPUs
> > making no progress for over half an hour.
> 
> Did the blkmap_grow() changes sneak in here accidentally?

Ah, forgot to mention them. Do a realloc every 4 entries when we
have hundreds of thousands of extents is just silly ;)

> > Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx>
> > ---
> >  repair/bmap.c | 36 ++++++++++++++++++++++++++++--------
> >  1 file changed, 28 insertions(+), 8 deletions(-)
> > 
> > diff --git a/repair/bmap.c b/repair/bmap.c
> > index 14161cb..1e913c5 100644
> > --- a/repair/bmap.c
> > +++ b/repair/bmap.c
> > @@ -260,7 +260,15 @@ blkmap_grow(
> >  {
> >  	pthread_key_t	key = dblkmap_key;
> >  	blkmap_t	*new_blkmap;
> > -	int		new_naexts = blkmap->naexts + 4;
> > +	int		new_naexts;
> > +
> > +	/* reduce the number of reallocations for large files */
> > +	if (blkmap->naexts < 1000)
> > +		new_naexts = blkmap->naexts + 4;
> > +	else if (blkmap->naexts < 10000)
> > +		new_naexts = blkmap->naexts + 100;
> > +	else
> > +		new_naexts = blkmap->naexts + 1000;
> >  
> >  	if (pthread_getspecific(key) != blkmap) {
> >  		key = ablkmap_key;
> > @@ -308,7 +316,7 @@ blkmap_set_ext(
> >  	xfs_dfilblks_t	c)
> >  {
> >  	blkmap_t	*blkmap = *blkmapp;
> > -	xfs_extnum_t	i;
> > +	xfs_extnum_t	i = 0;
> 
> I wonder if rather than initializing i here...
> 
> >  
> >  	if (blkmap->nexts == blkmap->naexts) {
> >  		blkmap = blkmap_grow(blkmap);
> > @@ -318,15 +326,27 @@ blkmap_set_ext(
> >  	}
> >  
> >  	ASSERT(blkmap->nexts < blkmap->naexts);
> > -	for (i = 0; i < blkmap->nexts; i++) {
> > -		if (blkmap->exts[i].startoff > o) {
> > -			memmove(blkmap->exts + i + 1,
> > -				blkmap->exts + i,
> > -				sizeof(bmap_ext_t) * (blkmap->nexts - i));
> > +
> > +	if (blkmap->nexts == 0)
> 
> ... setting i = 0 here would be a bit more explicit & clear?

*nod*. Will do.

> > +		goto insert;
> > +
> > +	/*
> > +	 * Do a reverse order search as the most common insertion during
> > +	 * a bmapbt scan is at the end of the map. Use "plus 1" indexing
> > +	 * for the loop counter so when we break out we are at the correct index
> > +	 * for insertion.
> > +	 */
> 
> perhaps say why it's most common at the end?

OK, I'll add "because we are doing an ascending offset order scan of
the bmapbt" or words to that effect.

Thanks!

Cheers,

Dave.
-- 
Dave Chinner
david@xxxxxxxxxxxxx

_______________________________________________
xfs mailing list
xfs@xxxxxxxxxxx
http://oss.sgi.com/mailman/listinfo/xfs




[Index of Archives]     [Linux XFS Devel]     [Linux Filesystem Development]     [Filesystem Testing]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux