Re: [PATCH v2] xfs: fix sb write verify for lazysbcount

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Oct 26, 2022 at 11:52:29AM -0700, Darrick J. Wong wrote:
> On Wed, Oct 26, 2022 at 05:13:44PM +0800, Long Li wrote:
> > On Tue, Oct 25, 2022 at 11:16:39AM -0700, Darrick J. Wong wrote:
> > > On Tue, Oct 25, 2022 at 05:15:27PM +0800, Long Li wrote:
> > > > When lazysbcount is enabled, fsstress and loop mount/unmount test report
> > > > the following problems:
> > > > 
> > > > XFS (loop0): SB summary counter sanity check failed
> > > > XFS (loop0): Metadata corruption detected at xfs_sb_write_verify+0x13b/0x460,
> > > > 	xfs_sb block 0x0
> > > > XFS (loop0): Unmount and run xfs_repair
> > > > XFS (loop0): First 128 bytes of corrupted metadata buffer:
> > > > 00000000: 58 46 53 42 00 00 10 00 00 00 00 00 00 28 00 00  XFSB.........(..
> > > > 00000010: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00  ................
> > > > 00000020: 69 fb 7c cd 5f dc 44 af 85 74 e0 cc d4 e3 34 5a  i.|._.D..t....4Z
> > > > 00000030: 00 00 00 00 00 20 00 06 00 00 00 00 00 00 00 80  ..... ..........
> > > > 00000040: 00 00 00 00 00 00 00 81 00 00 00 00 00 00 00 82  ................
> > > > 00000050: 00 00 00 01 00 0a 00 00 00 00 00 04 00 00 00 00  ................
> > > > 00000060: 00 00 0a 00 b4 b5 02 00 02 00 00 08 00 00 00 00  ................
> > > > 00000070: 00 00 00 00 00 00 00 00 0c 09 09 03 14 00 00 19  ................
> > > > XFS (loop0): Corruption of in-memory data (0x8) detected at _xfs_buf_ioapply
> > > > 	+0xe1e/0x10e0 (fs/xfs/xfs_buf.c:1580).  Shutting down filesystem.
> > > > XFS (loop0): Please unmount the filesystem and rectify the problem(s)
> > > > XFS (loop0): log mount/recovery failed: error -117
> > > > XFS (loop0): log mount failed
> > > > 
> > > > This will make the file system unmountable, the cause of the problem is
> > > > that during the log recovery process, incorrect count (ifree > icount)
> > > > are recovered from the log and fail to pass the boundary check in
> > > > xfs_validate_sb_write(). The following script can reproduce the problem,
> > > > but it may take a long time.
> > > > 
> > > > device=/dev/sda
> > > > testdir=/mnt/test
> > > > round=0
> > > > 
> > > > function fail()
> > > > {
> > > > 	echo "$*"
> > > > 	exit 1
> > > > }
> > > > 
> > > > mkdir -p $testdir
> > > > while [ $round -lt 10000 ]
> > > > do
> > > > 	echo "******* round $round ********"
> > > > 	mkfs.xfs -f $device
> > > > 	mount $device $testdir || fail "mount failed!"
> > > > 	fsstress -d $testdir -l 0 -n 10000 -p 4 >/dev/null &
> > > 
> > > What is the backtrace of the xfs_log_sb caller?  I speculate that it's
> > > something along the lines of adding a superblock feature?  attr2 would
> > > be my guess since this is fsstress.
> > 
> > The call trace that I reproduced:
> >  Call Trace:
> >   <TASK>
> >   dump_stack_lvl+0x4d/0x66
> >   xfs_log_sb.cold+0x2f/0x1af
> >   xfs_bmap_add_attrfork+0x687/0xb40
> >   ? get_reg+0x91/0x190
> >   ? xfs_bmap_add_attrfork+0x0/0xb40
> >   ? unwind_next_frame+0x115d/0x1b70
> >   ? xfs_attr_calc_size+0x13c/0x2e0
> >   xfs_attr_set+0xb51/0x1d50
> >   ? __kernel_text_address-0xe/0x30
> >   ? xfs_attr_set+0x0/0x1d50
> >   ? __kernel_text_address+0xe/0x30
> >   ? unwind_get_return_address+0x5f/0xa0
> >   ? stack_trace_consume_entry+0x0/0x160
> >   ? arch_stack_walk+0x98/0xf0
> >   xfs_attr_change+0x22d/0x380
> >   xfs_xattr_set+0xeb/0x160
> >   ? xfs_xattr_set+0x0/0x160
> >   ? vmemdup_user+0x27/0xa0
> >   ? setxattr_copy+0x103/0x1a0
> >   ? setxattr+0xd1/0x160
> >   ? path_setxattr+0x168/0x190
> >   ? __x64_sys_setxattr+0xc5/0x160
> >   ? xattr_resolve_name+0x23d/0x360
> >   ? xfs_xattr_set+0x0/0x160
> >   __vfs_setxattr+0x100/0x160
> >   ? __vfs_setxattr+0x0/0x160
> >   __vfs_setxattr_noperm+0x104/0x320
> >   __vfs_setxattr_locked+0x1ba/0x260
> > 
> > > 
> > > So the other racing thread would be a thread that just freed an inode
> > > cluster, committed the transaction, and now it's committing idelta and
> > > ifreedelta into the incore percpu counters via:
> > > 
> > > 	if (idelta)
> > > 		percpu_counter_add_batch(&mp->m_icount, idelta,
> > > 					 XFS_ICOUNT_BATCH);
> > > 
> > > 	if (ifreedelta)
> > > 		percpu_counter_add(&mp->m_ifree, ifreedelta);
> > > 
> > > > 	sleep 4
> > > > 	killall -w fsstress
> > > > 	umount $testdir
> > > > 	xfs_repair -e $device > /dev/null
> > > > 	if [ $? -eq 2 ];then
> > > > 		echo "ERR CODE 2: Dirty log exception during repair."
> > > > 		exit 1
> > > > 	fi
> > > > 	round=$(($round+1))
> > > > done
> > > > 
> > > > With lazysbcount is enabled, There is no additional lock protection for
> > > > reading m_ifree and m_icount in xfs_log_sb(), if other cpu modifies the
> > > > m_ifree, this will make the m_ifree greater than m_icount and written to
> > > > the log. For example consider the following sequence:
> > > > 
> > > >  CPU0				 CPU1
> > > >  xfs_log_sb			 xfs_trans_unreserve_and_mod_sb
> > > >  ----------			 ------------------------------
> > > >  percpu_counter_sum(&mp->m_icount)
> > > > 				 percpu_counter_add(&mp->m_icount, idelta)
> > > 
> > > This callsite does not exist ^^^^^^^^^^^ in the codebase, AFAICT.
> > > 
> > > > 				 percpu_counter_add_batch(&mp->m_icount,
> > > > 						idelta, XFS_ICOUNT_BATCH)
> > > >  percpu_counter_sum(&mp->m_ifree)
> > 
> > Sorry, the code I copied is wrong, as it should be:
> > 
> >  CPU0				 CPU1
> >  xfs_log_sb			 xfs_trans_unreserve_and_mod_sb
> >  ----------			 ------------------------------
> >  percpu_counter_sum(&mp->m_icount)
> > 				 percpu_counter_add_batch(&mp->m_icount,
> > 						idelta, XFS_ICOUNT_BATCH)
> > 				 percpu_counter_add(&mp->m_ifree, ifreedelta);
> >  percpu_counter_sum(&mp->m_ifree)
> > 
> > > 
> > > I think what's happening here is more like:
> > > 
> > > 1. CPU1 adds a negative idelta to m_icount.
> > > 2. CPU0 sums m_icount.
> > > 3. CPU0 sums m_ifree.
> > > 4. CPU1 adds a negative ideltafree to m_ifree.
> > 
> > I tried to reproduce the situation that you said, but it hasn't been
> > reproduced yet. Only the following sequence is reproduced:
> > 
> > 1. CPU0 sums m_icount.
> > 2. CPU1 adds a positive idelta (e.g. 32) to m_icount.
> > 3. CPU1 adds a positive ideltafree (e.g. 32) to m_ifree.
> > 4. CPU0 sums m_ifree. 
> 
> Aha, that was my second guess as to what was really going on.
> 
> Either way, we're racing with updates to two percpu counters.
> Now that the source of the bug has been clarified...
> 
> > > Now CPU0 has an ifree > icount, which it writes into the primary
> > > superblock buffer.  Eventually the AIL writes the buffer to disk, only
> > > the write verifier trips over icount < ifree and shuts down the fs.
> > > 
> > > > If we have an unclean shutdown, this will be corrected by
> > > > xfs_initialize_perag_data() rebuilding the counters from the AGF block
> > > > counts, and the correction is later than log recovery. During log recovery,
> > > > incorrect ifree/icount may be restored from the log and written sb, since
> > > > ifree and icount have not been corrected at this time, sb write check
> > > > will fail due to ifree > icount.
> > > > 
> > > > Guaranteed that ifree will never be logged as being greater than icount.
> > > > Neither icount or ifree will be accurate if we are racing with other
> > > > updates, but it will guarantee that what we write to the journal
> > > > won't trigger corruption warnings.
> > > > 
> > > > Fixes: 8756a5af1819 ("libxfs: add more bounds checking to sb sanity checks")
> > > > Signed-off-by: Long Li <leo.lilong@xxxxxxxxxx>
> > > > ---
> > > > v2:
> > > > - Add scripts that could reproduce the problem
> > > > - Guaranteed that ifree will never be logged as being greater than icount
> > > > 
> > > >  fs/xfs/libxfs/xfs_sb.c | 4 +++-
> > > >  1 file changed, 3 insertions(+), 1 deletion(-)
> > > > 
> > > > diff --git a/fs/xfs/libxfs/xfs_sb.c b/fs/xfs/libxfs/xfs_sb.c
> > > > index a20cade590e9..1eeecf2eb2a7 100644
> > > > --- a/fs/xfs/libxfs/xfs_sb.c
> > > > +++ b/fs/xfs/libxfs/xfs_sb.c
> > > > @@ -972,7 +972,9 @@ xfs_log_sb(
> > > >  	 */
> > > >  	if (xfs_has_lazysbcount(mp)) {
> > > >  		mp->m_sb.sb_icount = percpu_counter_sum(&mp->m_icount);
> > > > -		mp->m_sb.sb_ifree = percpu_counter_sum(&mp->m_ifree);
> > > > +		mp->m_sb.sb_ifree = min_t(uint64_t,
> > > > +				percpu_counter_sum(&mp->m_ifree),
> > > > +				mp->m_sb.sb_icount);
> > > 
> > > This part looks plausible, but I think xfs_unmountfs really ought to
> > > check that m_ifree < m_icount after it's quiesced the rest of the
> > > filesystem and freed the reserve block pool.  If ifree is still larger
> > > than icount, someone has corrupted the incore counters, so we should not
> > > write a clean unmount record.
> 
> ...please update the patch to include this sanity check at unmount so
> that I can get this bugfix moving towards upstream.

I have some questions about this. If we can guarantee that
m_ifree <= m_icount, why do we need add a check at umount?
On the other hand, if m_ifree > m_icount, sb write will triger
a shutdown in xfs_validate_sb_write() because the check did
not pass, therefore it will not write a clean umount record
at umount. I also haven't found a code suitable for adding 
such checks. 
> 
> --D
> 
> > > 
> > > --D
> > > 
> > > >  		mp->m_sb.sb_fdblocks = percpu_counter_sum(&mp->m_fdblocks);
> > > >  	}
> > > >  
> > > > -- 
> > > > 2.31.1
> > > > 



[Index of Archives]     [XFS Filesystem Development (older mail)]     [Linux Filesystem Development]     [Linux Audio Users]     [Yosemite Trails]     [Linux Kernel]     [Linux RAID]     [Linux SCSI]


  Powered by Linux