On Wed, Jul 18, 2018 at 08:51:51AM -0700, Darrick J. Wong wrote: > From: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > > Shorten all the metadata checking xfs_scrub_ prefixes to xchk_. After > this, the only xfs_scrub* symbols are the ones that pertain to both > scrub and repair. Whitespace damage will be fixed in a subsequent > patch. There are no functional changes. > > Signed-off-by: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > --- I have no major preference on the name, but looks like a straight search/replace to me: Reviewed-by: Brian Foster <bfoster@xxxxxxxxxx> > fs/xfs/scrub/agheader.c | 318 ++++++++++++++++++++++++----------------------- > fs/xfs/scrub/alloc.c | 56 ++++---- > fs/xfs/scrub/attr.c | 106 ++++++++-------- > fs/xfs/scrub/bmap.c | 176 +++++++++++++------------- > fs/xfs/scrub/btree.c | 166 ++++++++++++------------- > fs/xfs/scrub/btree.h | 22 ++- > fs/xfs/scrub/common.c | 172 +++++++++++++------------ > fs/xfs/scrub/common.h | 102 ++++++++------- > fs/xfs/scrub/dabtree.c | 124 +++++++++--------- > fs/xfs/scrub/dabtree.h | 14 +- > fs/xfs/scrub/dir.c | 170 +++++++++++++------------ > fs/xfs/scrub/ialloc.c | 138 ++++++++++---------- > fs/xfs/scrub/inode.c | 160 ++++++++++++------------ > fs/xfs/scrub/parent.c | 62 +++++---- > fs/xfs/scrub/quota.c | 64 +++++---- > fs/xfs/scrub/refcount.c | 114 ++++++++--------- > fs/xfs/scrub/repair.c | 4 - > fs/xfs/scrub/rmap.c | 78 ++++++------ > fs/xfs/scrub/rtbitmap.c | 30 ++-- > fs/xfs/scrub/scrub.c | 142 +++++++++++---------- > fs/xfs/scrub/scrub.h | 82 ++++++------ > fs/xfs/scrub/symlink.c | 14 +- > fs/xfs/scrub/trace.c | 2 > fs/xfs/scrub/trace.h | 70 +++++----- > 24 files changed, 1196 insertions(+), 1190 deletions(-) > > > diff --git a/fs/xfs/scrub/agheader.c b/fs/xfs/scrub/agheader.c > index 9bb0745f1ad2..c0625ec16d63 100644 > --- a/fs/xfs/scrub/agheader.c > +++ b/fs/xfs/scrub/agheader.c > @@ -28,7 +28,7 @@ > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_superblock_xref( > +xchk_superblock_xref( > struct xfs_scrub_context *sc, > struct xfs_buf *bp) > { > @@ -43,15 +43,15 @@ xfs_scrub_superblock_xref( > > agbno = XFS_SB_BLOCK(mp); > > - error = xfs_scrub_ag_init(sc, agno, &sc->sa); > - if (!xfs_scrub_xref_process_error(sc, agno, agbno, &error)) > + error = xchk_ag_init(sc, agno, &sc->sa); > + if (!xchk_xref_process_error(sc, agno, agbno, &error)) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_xref_is_not_inode_chunk(sc, agbno, 1); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > + xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); > + xchk_xref_is_not_shared(sc, agbno, 1); > > /* scrub teardown will take care of sc->sa for us */ > } > @@ -65,7 +65,7 @@ xfs_scrub_superblock_xref( > * sb 0 is ok and we can use its information to check everything else. > */ > int > -xfs_scrub_superblock( > +xchk_superblock( > struct xfs_scrub_context *sc) > { > struct xfs_mount *mp = sc->mp; > @@ -98,7 +98,7 @@ xfs_scrub_superblock( > default: > break; > } > - if (!xfs_scrub_process_error(sc, agno, XFS_SB_BLOCK(mp), &error)) > + if (!xchk_process_error(sc, agno, XFS_SB_BLOCK(mp), &error)) > return error; > > sb = XFS_BUF_TO_SBP(bp); > @@ -110,46 +110,46 @@ xfs_scrub_superblock( > * checked. > */ > if (sb->sb_blocksize != cpu_to_be32(mp->m_sb.sb_blocksize)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_dblocks != cpu_to_be64(mp->m_sb.sb_dblocks)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_rblocks != cpu_to_be64(mp->m_sb.sb_rblocks)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_rextents != cpu_to_be64(mp->m_sb.sb_rextents)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (!uuid_equal(&sb->sb_uuid, &mp->m_sb.sb_uuid)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_logstart != cpu_to_be64(mp->m_sb.sb_logstart)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_rootino != cpu_to_be64(mp->m_sb.sb_rootino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_rbmino != cpu_to_be64(mp->m_sb.sb_rbmino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_rsumino != cpu_to_be64(mp->m_sb.sb_rsumino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_rextsize != cpu_to_be32(mp->m_sb.sb_rextsize)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_agblocks != cpu_to_be32(mp->m_sb.sb_agblocks)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_agcount != cpu_to_be32(mp->m_sb.sb_agcount)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_rbmblocks != cpu_to_be32(mp->m_sb.sb_rbmblocks)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_logblocks != cpu_to_be32(mp->m_sb.sb_logblocks)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check sb_versionnum bits that are set at mkfs time. */ > vernum_mask = cpu_to_be16(~XFS_SB_VERSION_OKBITS | > @@ -163,7 +163,7 @@ xfs_scrub_superblock( > XFS_SB_VERSION_DIRV2BIT); > if ((sb->sb_versionnum & vernum_mask) != > (cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check sb_versionnum bits that can be set after mkfs time. */ > vernum_mask = cpu_to_be16(XFS_SB_VERSION_ATTRBIT | > @@ -171,40 +171,40 @@ xfs_scrub_superblock( > XFS_SB_VERSION_QUOTABIT); > if ((sb->sb_versionnum & vernum_mask) != > (cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_sectsize != cpu_to_be16(mp->m_sb.sb_sectsize)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_inodesize != cpu_to_be16(mp->m_sb.sb_inodesize)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_inopblock != cpu_to_be16(mp->m_sb.sb_inopblock)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (memcmp(sb->sb_fname, mp->m_sb.sb_fname, sizeof(sb->sb_fname))) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_blocklog != mp->m_sb.sb_blocklog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_sectlog != mp->m_sb.sb_sectlog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_inodelog != mp->m_sb.sb_inodelog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_inopblog != mp->m_sb.sb_inopblog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_agblklog != mp->m_sb.sb_agblklog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_rextslog != mp->m_sb.sb_rextslog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_imax_pct != mp->m_sb.sb_imax_pct) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > /* > * Skip the summary counters since we track them in memory anyway. > @@ -212,10 +212,10 @@ xfs_scrub_superblock( > */ > > if (sb->sb_uquotino != cpu_to_be64(mp->m_sb.sb_uquotino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_gquotino != cpu_to_be64(mp->m_sb.sb_gquotino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > /* > * Skip the quota flags since repair will force quotacheck. > @@ -223,46 +223,46 @@ xfs_scrub_superblock( > */ > > if (sb->sb_flags != mp->m_sb.sb_flags) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_shared_vn != mp->m_sb.sb_shared_vn) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_inoalignmt != cpu_to_be32(mp->m_sb.sb_inoalignmt)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_unit != cpu_to_be32(mp->m_sb.sb_unit)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_width != cpu_to_be32(mp->m_sb.sb_width)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > if (sb->sb_dirblklog != mp->m_sb.sb_dirblklog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_logsectlog != mp->m_sb.sb_logsectlog) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_logsectsize != cpu_to_be16(mp->m_sb.sb_logsectsize)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_logsunit != cpu_to_be32(mp->m_sb.sb_logsunit)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Do we see any invalid bits in sb_features2? */ > if (!xfs_sb_version_hasmorebits(&mp->m_sb)) { > if (sb->sb_features2 != 0) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > } else { > v2_ok = XFS_SB_VERSION2_OKBITS; > if (XFS_SB_VERSION_NUM(&mp->m_sb) >= XFS_SB_VERSION_5) > v2_ok |= XFS_SB_VERSION2_CRCBIT; > > if (!!(sb->sb_features2 & cpu_to_be32(~v2_ok))) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_features2 != sb->sb_bad_features2) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > } > > /* Check sb_features2 flags that are set at mkfs time. */ > @@ -272,26 +272,26 @@ xfs_scrub_superblock( > XFS_SB_VERSION2_FTYPE); > if ((sb->sb_features2 & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features2) & features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check sb_features2 flags that can be set after mkfs time. */ > features_mask = cpu_to_be32(XFS_SB_VERSION2_ATTR2BIT); > if ((sb->sb_features2 & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features2) & features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (!xfs_sb_version_hascrc(&mp->m_sb)) { > /* all v5 fields must be zero */ > if (memchr_inv(&sb->sb_features_compat, 0, > sizeof(struct xfs_dsb) - > offsetof(struct xfs_dsb, sb_features_compat))) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > } else { > /* Check compat flags; all are set at mkfs time. */ > features_mask = cpu_to_be32(XFS_SB_FEAT_COMPAT_UNKNOWN); > if ((sb->sb_features_compat & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features_compat) & features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check ro compat flags; all are set at mkfs time. */ > features_mask = cpu_to_be32(XFS_SB_FEAT_RO_COMPAT_UNKNOWN | > @@ -301,7 +301,7 @@ xfs_scrub_superblock( > if ((sb->sb_features_ro_compat & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features_ro_compat) & > features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check incompat flags; all are set at mkfs time. */ > features_mask = cpu_to_be32(XFS_SB_FEAT_INCOMPAT_UNKNOWN | > @@ -311,22 +311,22 @@ xfs_scrub_superblock( > if ((sb->sb_features_incompat & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features_incompat) & > features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Check log incompat flags; all are set at mkfs time. */ > features_mask = cpu_to_be32(XFS_SB_FEAT_INCOMPAT_LOG_UNKNOWN); > if ((sb->sb_features_log_incompat & features_mask) != > (cpu_to_be32(mp->m_sb.sb_features_log_incompat) & > features_mask)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > /* Don't care about sb_crc */ > > if (sb->sb_spino_align != cpu_to_be32(mp->m_sb.sb_spino_align)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > if (sb->sb_pquotino != cpu_to_be64(mp->m_sb.sb_pquotino)) > - xfs_scrub_block_set_preen(sc, bp); > + xchk_block_set_preen(sc, bp); > > /* Don't care about sb_lsn */ > } > @@ -334,15 +334,15 @@ xfs_scrub_superblock( > if (xfs_sb_version_hasmetauuid(&mp->m_sb)) { > /* The metadata UUID must be the same for all supers */ > if (!uuid_equal(&sb->sb_meta_uuid, &mp->m_sb.sb_meta_uuid)) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > } > > /* Everything else must be zero. */ > if (memchr_inv(sb + 1, 0, > BBTOB(bp->b_length) - sizeof(struct xfs_dsb))) > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > > - xfs_scrub_superblock_xref(sc, bp); > + xchk_superblock_xref(sc, bp); > > return error; > } > @@ -351,7 +351,7 @@ xfs_scrub_superblock( > > /* Tally freespace record lengths. */ > STATIC int > -xfs_scrub_agf_record_bno_lengths( > +xchk_agf_record_bno_lengths( > struct xfs_btree_cur *cur, > struct xfs_alloc_rec_incore *rec, > void *priv) > @@ -364,7 +364,7 @@ xfs_scrub_agf_record_bno_lengths( > > /* Check agf_freeblks */ > static inline void > -xfs_scrub_agf_xref_freeblks( > +xchk_agf_xref_freeblks( > struct xfs_scrub_context *sc) > { > struct xfs_agf *agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > @@ -375,16 +375,16 @@ xfs_scrub_agf_xref_freeblks( > return; > > error = xfs_alloc_query_all(sc->sa.bno_cur, > - xfs_scrub_agf_record_bno_lengths, &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.bno_cur)) > + xchk_agf_record_bno_lengths, &blocks); > + if (!xchk_should_check_xref(sc, &error, &sc->sa.bno_cur)) > return; > if (blocks != be32_to_cpu(agf->agf_freeblks)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > } > > /* Cross reference the AGF with the cntbt (freespace by length btree) */ > static inline void > -xfs_scrub_agf_xref_cntbt( > +xchk_agf_xref_cntbt( > struct xfs_scrub_context *sc) > { > struct xfs_agf *agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > @@ -398,25 +398,25 @@ xfs_scrub_agf_xref_cntbt( > > /* Any freespace at all? */ > error = xfs_alloc_lookup_le(sc->sa.cnt_cur, 0, -1U, &have); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > return; > if (!have) { > if (agf->agf_freeblks != be32_to_cpu(0)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > return; > } > > /* Check agf_longest */ > error = xfs_alloc_get_rec(sc->sa.cnt_cur, &agbno, &blocks, &have); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > return; > if (!have || blocks != be32_to_cpu(agf->agf_longest)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > } > > /* Check the btree block counts in the AGF against the btrees. */ > STATIC void > -xfs_scrub_agf_xref_btreeblks( > +xchk_agf_xref_btreeblks( > struct xfs_scrub_context *sc) > { > struct xfs_agf *agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > @@ -428,11 +428,11 @@ xfs_scrub_agf_xref_btreeblks( > /* Check agf_rmap_blocks; set up for agf_btreeblks check */ > if (sc->sa.rmap_cur) { > error = xfs_btree_count_blocks(sc->sa.rmap_cur, &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > btreeblks = blocks - 1; > if (blocks != be32_to_cpu(agf->agf_rmap_blocks)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > } else { > btreeblks = 0; > } > @@ -447,22 +447,22 @@ xfs_scrub_agf_xref_btreeblks( > > /* Check agf_btreeblks */ > error = xfs_btree_count_blocks(sc->sa.bno_cur, &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.bno_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.bno_cur)) > return; > btreeblks += blocks - 1; > > error = xfs_btree_count_blocks(sc->sa.cnt_cur, &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.cnt_cur)) > return; > btreeblks += blocks - 1; > > if (btreeblks != be32_to_cpu(agf->agf_btreeblks)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > } > > /* Check agf_refcount_blocks against tree size */ > static inline void > -xfs_scrub_agf_xref_refcblks( > +xchk_agf_xref_refcblks( > struct xfs_scrub_context *sc) > { > struct xfs_agf *agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > @@ -473,15 +473,15 @@ xfs_scrub_agf_xref_refcblks( > return; > > error = xfs_btree_count_blocks(sc->sa.refc_cur, &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) > return; > if (blocks != be32_to_cpu(agf->agf_refcount_blocks)) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agf_bp); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_agf_xref( > +xchk_agf_xref( > struct xfs_scrub_context *sc) > { > struct xfs_owner_info oinfo; > @@ -494,26 +494,26 @@ xfs_scrub_agf_xref( > > agbno = XFS_AGF_BLOCK(mp); > > - error = xfs_scrub_ag_btcur_init(sc, &sc->sa); > + error = xchk_ag_btcur_init(sc, &sc->sa); > if (error) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_agf_xref_freeblks(sc); > - xfs_scrub_agf_xref_cntbt(sc); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_agf_xref_freeblks(sc); > + xchk_agf_xref_cntbt(sc); > + xchk_xref_is_not_inode_chunk(sc, agbno, 1); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo); > - xfs_scrub_agf_xref_btreeblks(sc); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > - xfs_scrub_agf_xref_refcblks(sc); > + xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); > + xchk_agf_xref_btreeblks(sc); > + xchk_xref_is_not_shared(sc, agbno, 1); > + xchk_agf_xref_refcblks(sc); > > /* scrub teardown will take care of sc->sa for us */ > } > > /* Scrub the AGF. */ > int > -xfs_scrub_agf( > +xchk_agf( > struct xfs_scrub_context *sc) > { > struct xfs_mount *mp = sc->mp; > @@ -529,54 +529,54 @@ xfs_scrub_agf( > int error = 0; > > agno = sc->sa.agno = sc->sm->sm_agno; > - error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp, > + error = xchk_ag_read_headers(sc, agno, &sc->sa.agi_bp, > &sc->sa.agf_bp, &sc->sa.agfl_bp); > - if (!xfs_scrub_process_error(sc, agno, XFS_AGF_BLOCK(sc->mp), &error)) > + if (!xchk_process_error(sc, agno, XFS_AGF_BLOCK(sc->mp), &error)) > goto out; > - xfs_scrub_buffer_recheck(sc, sc->sa.agf_bp); > + xchk_buffer_recheck(sc, sc->sa.agf_bp); > > agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > > /* Check the AG length */ > eoag = be32_to_cpu(agf->agf_length); > if (eoag != xfs_ag_block_count(mp, agno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > /* Check the AGF btree roots and levels */ > agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_BNO]); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_CNT]); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > if (xfs_sb_version_hasrmapbt(&mp->m_sb)) { > agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_RMAP]); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_RMAP]); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > } > > if (xfs_sb_version_hasreflink(&mp->m_sb)) { > agbno = be32_to_cpu(agf->agf_refcount_root); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > level = be32_to_cpu(agf->agf_refcount_level); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > } > > /* Check the AGFL counters */ > @@ -588,16 +588,16 @@ xfs_scrub_agf( > else > fl_count = xfs_agfl_size(mp) - agfl_first + agfl_last + 1; > if (agfl_count != 0 && fl_count != agfl_count) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > > - xfs_scrub_agf_xref(sc); > + xchk_agf_xref(sc); > out: > return error; > } > > /* AGFL */ > > -struct xfs_scrub_agfl_info { > +struct xchk_agfl_info { > struct xfs_owner_info oinfo; > unsigned int sz_entries; > unsigned int nr_entries; > @@ -607,7 +607,7 @@ struct xfs_scrub_agfl_info { > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_agfl_block_xref( > +xchk_agfl_block_xref( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > struct xfs_owner_info *oinfo) > @@ -615,20 +615,20 @@ xfs_scrub_agfl_block_xref( > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_xref_is_not_inode_chunk(sc, agbno, 1); > + xchk_xref_is_owned_by(sc, agbno, 1, oinfo); > + xchk_xref_is_not_shared(sc, agbno, 1); > } > > /* Scrub an AGFL block. */ > STATIC int > -xfs_scrub_agfl_block( > +xchk_agfl_block( > struct xfs_mount *mp, > xfs_agblock_t agbno, > void *priv) > { > - struct xfs_scrub_agfl_info *sai = priv; > + struct xchk_agfl_info *sai = priv; > struct xfs_scrub_context *sc = sai->sc; > xfs_agnumber_t agno = sc->sa.agno; > > @@ -636,9 +636,9 @@ xfs_scrub_agfl_block( > sai->nr_entries < sai->sz_entries) > sai->entries[sai->nr_entries++] = agbno; > else > - xfs_scrub_block_set_corrupt(sc, sc->sa.agfl_bp); > + xchk_block_set_corrupt(sc, sc->sa.agfl_bp); > > - xfs_scrub_agfl_block_xref(sc, agbno, priv); > + xchk_agfl_block_xref(sc, agbno, priv); > > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return XFS_BTREE_QUERY_RANGE_ABORT; > @@ -647,7 +647,7 @@ xfs_scrub_agfl_block( > } > > static int > -xfs_scrub_agblock_cmp( > +xchk_agblock_cmp( > const void *pa, > const void *pb) > { > @@ -659,7 +659,7 @@ xfs_scrub_agblock_cmp( > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_agfl_xref( > +xchk_agfl_xref( > struct xfs_scrub_context *sc) > { > struct xfs_owner_info oinfo; > @@ -672,15 +672,15 @@ xfs_scrub_agfl_xref( > > agbno = XFS_AGFL_BLOCK(mp); > > - error = xfs_scrub_ag_btcur_init(sc, &sc->sa); > + error = xchk_ag_btcur_init(sc, &sc->sa); > if (error) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_xref_is_not_inode_chunk(sc, agbno, 1); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > + xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); > + xchk_xref_is_not_shared(sc, agbno, 1); > > /* > * Scrub teardown will take care of sc->sa for us. Leave sc->sa > @@ -690,10 +690,10 @@ xfs_scrub_agfl_xref( > > /* Scrub the AGFL. */ > int > -xfs_scrub_agfl( > +xchk_agfl( > struct xfs_scrub_context *sc) > { > - struct xfs_scrub_agfl_info sai; > + struct xchk_agfl_info sai; > struct xfs_agf *agf; > xfs_agnumber_t agno; > unsigned int agflcount; > @@ -701,15 +701,15 @@ xfs_scrub_agfl( > int error; > > agno = sc->sa.agno = sc->sm->sm_agno; > - error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp, > + error = xchk_ag_read_headers(sc, agno, &sc->sa.agi_bp, > &sc->sa.agf_bp, &sc->sa.agfl_bp); > - if (!xfs_scrub_process_error(sc, agno, XFS_AGFL_BLOCK(sc->mp), &error)) > + if (!xchk_process_error(sc, agno, XFS_AGFL_BLOCK(sc->mp), &error)) > goto out; > if (!sc->sa.agf_bp) > return -EFSCORRUPTED; > - xfs_scrub_buffer_recheck(sc, sc->sa.agfl_bp); > + xchk_buffer_recheck(sc, sc->sa.agfl_bp); > > - xfs_scrub_agfl_xref(sc); > + xchk_agfl_xref(sc); > > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > @@ -718,7 +718,7 @@ xfs_scrub_agfl( > agf = XFS_BUF_TO_AGF(sc->sa.agf_bp); > agflcount = be32_to_cpu(agf->agf_flcount); > if (agflcount > xfs_agfl_size(sc->mp)) { > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > goto out; > } > memset(&sai, 0, sizeof(sai)); > @@ -734,7 +734,7 @@ xfs_scrub_agfl( > /* Check the blocks in the AGFL. */ > xfs_rmap_ag_owner(&sai.oinfo, XFS_RMAP_OWN_AG); > error = xfs_agfl_walk(sc->mp, XFS_BUF_TO_AGF(sc->sa.agf_bp), > - sc->sa.agfl_bp, xfs_scrub_agfl_block, &sai); > + sc->sa.agfl_bp, xchk_agfl_block, &sai); > if (error == XFS_BTREE_QUERY_RANGE_ABORT) { > error = 0; > goto out_free; > @@ -743,16 +743,16 @@ xfs_scrub_agfl( > goto out_free; > > if (agflcount != sai.nr_entries) { > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > goto out_free; > } > > /* Sort entries, check for duplicates. */ > sort(sai.entries, sai.nr_entries, sizeof(sai.entries[0]), > - xfs_scrub_agblock_cmp, NULL); > + xchk_agblock_cmp, NULL); > for (i = 1; i < sai.nr_entries; i++) { > if (sai.entries[i] == sai.entries[i - 1]) { > - xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp); > + xchk_block_set_corrupt(sc, sc->sa.agf_bp); > break; > } > } > @@ -767,7 +767,7 @@ xfs_scrub_agfl( > > /* Check agi_count/agi_freecount */ > static inline void > -xfs_scrub_agi_xref_icounts( > +xchk_agi_xref_icounts( > struct xfs_scrub_context *sc) > { > struct xfs_agi *agi = XFS_BUF_TO_AGI(sc->sa.agi_bp); > @@ -779,16 +779,16 @@ xfs_scrub_agi_xref_icounts( > return; > > error = xfs_ialloc_count_inodes(sc->sa.ino_cur, &icount, &freecount); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.ino_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.ino_cur)) > return; > if (be32_to_cpu(agi->agi_count) != icount || > be32_to_cpu(agi->agi_freecount) != freecount) > - xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_xref_set_corrupt(sc, sc->sa.agi_bp); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_agi_xref( > +xchk_agi_xref( > struct xfs_scrub_context *sc) > { > struct xfs_owner_info oinfo; > @@ -801,23 +801,23 @@ xfs_scrub_agi_xref( > > agbno = XFS_AGI_BLOCK(mp); > > - error = xfs_scrub_ag_btcur_init(sc, &sc->sa); > + error = xchk_ag_btcur_init(sc, &sc->sa); > if (error) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1); > - xfs_scrub_agi_xref_icounts(sc); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_xref_is_not_inode_chunk(sc, agbno, 1); > + xchk_agi_xref_icounts(sc); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > + xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); > + xchk_xref_is_not_shared(sc, agbno, 1); > > /* scrub teardown will take care of sc->sa for us */ > } > > /* Scrub the AGI. */ > int > -xfs_scrub_agi( > +xchk_agi( > struct xfs_scrub_context *sc) > { > struct xfs_mount *mp = sc->mp; > @@ -834,36 +834,36 @@ xfs_scrub_agi( > int error = 0; > > agno = sc->sa.agno = sc->sm->sm_agno; > - error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp, > + error = xchk_ag_read_headers(sc, agno, &sc->sa.agi_bp, > &sc->sa.agf_bp, &sc->sa.agfl_bp); > - if (!xfs_scrub_process_error(sc, agno, XFS_AGI_BLOCK(sc->mp), &error)) > + if (!xchk_process_error(sc, agno, XFS_AGI_BLOCK(sc->mp), &error)) > goto out; > - xfs_scrub_buffer_recheck(sc, sc->sa.agi_bp); > + xchk_buffer_recheck(sc, sc->sa.agi_bp); > > agi = XFS_BUF_TO_AGI(sc->sa.agi_bp); > > /* Check the AG length */ > eoag = be32_to_cpu(agi->agi_length); > if (eoag != xfs_ag_block_count(mp, agno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > /* Check btree roots and levels */ > agbno = be32_to_cpu(agi->agi_root); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > level = be32_to_cpu(agi->agi_level); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > if (xfs_sb_version_hasfinobt(&mp->m_sb)) { > agbno = be32_to_cpu(agi->agi_free_root); > if (!xfs_verify_agbno(mp, agno, agbno)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > level = be32_to_cpu(agi->agi_free_level); > if (level <= 0 || level > XFS_BTREE_MAXLEVELS) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > } > > /* Check inode counters */ > @@ -871,16 +871,16 @@ xfs_scrub_agi( > icount = be32_to_cpu(agi->agi_count); > if (icount > last_agino - first_agino + 1 || > icount < be32_to_cpu(agi->agi_freecount)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > /* Check inode pointers */ > agino = be32_to_cpu(agi->agi_newino); > if (agino != NULLAGINO && !xfs_verify_agino(mp, agno, agino)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > agino = be32_to_cpu(agi->agi_dirino); > if (agino != NULLAGINO && !xfs_verify_agino(mp, agno, agino)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > /* Check unlinked inode buckets */ > for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) { > @@ -888,13 +888,13 @@ xfs_scrub_agi( > if (agino == NULLAGINO) > continue; > if (!xfs_verify_agino(mp, agno, agino)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > } > > if (agi->agi_pad32 != cpu_to_be32(0)) > - xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp); > + xchk_block_set_corrupt(sc, sc->sa.agi_bp); > > - xfs_scrub_agi_xref(sc); > + xchk_agi_xref(sc); > out: > return error; > } > diff --git a/fs/xfs/scrub/alloc.c b/fs/xfs/scrub/alloc.c > index 50e4f7fa06f0..1f6e3a6a1fdd 100644 > --- a/fs/xfs/scrub/alloc.c > +++ b/fs/xfs/scrub/alloc.c > @@ -28,11 +28,11 @@ > * Set us up to scrub free space btrees. > */ > int > -xfs_scrub_setup_ag_allocbt( > +xchk_setup_ag_allocbt( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_ag_btree(sc, ip, false); > + return xchk_setup_ag_btree(sc, ip, false); > } > > /* Free space btree scrubber. */ > @@ -41,7 +41,7 @@ xfs_scrub_setup_ag_allocbt( > * bnobt/cntbt record, respectively. > */ > STATIC void > -xfs_scrub_allocbt_xref_other( > +xchk_allocbt_xref_other( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > @@ -56,32 +56,32 @@ xfs_scrub_allocbt_xref_other( > pcur = &sc->sa.cnt_cur; > else > pcur = &sc->sa.bno_cur; > - if (!*pcur || xfs_scrub_skip_xref(sc->sm)) > + if (!*pcur || xchk_skip_xref(sc->sm)) > return; > > error = xfs_alloc_lookup_le(*pcur, agbno, len, &has_otherrec); > - if (!xfs_scrub_should_check_xref(sc, &error, pcur)) > + if (!xchk_should_check_xref(sc, &error, pcur)) > return; > if (!has_otherrec) { > - xfs_scrub_btree_xref_set_corrupt(sc, *pcur, 0); > + xchk_btree_xref_set_corrupt(sc, *pcur, 0); > return; > } > > error = xfs_alloc_get_rec(*pcur, &fbno, &flen, &has_otherrec); > - if (!xfs_scrub_should_check_xref(sc, &error, pcur)) > + if (!xchk_should_check_xref(sc, &error, pcur)) > return; > if (!has_otherrec) { > - xfs_scrub_btree_xref_set_corrupt(sc, *pcur, 0); > + xchk_btree_xref_set_corrupt(sc, *pcur, 0); > return; > } > > if (fbno != agbno || flen != len) > - xfs_scrub_btree_xref_set_corrupt(sc, *pcur, 0); > + xchk_btree_xref_set_corrupt(sc, *pcur, 0); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_allocbt_xref( > +xchk_allocbt_xref( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > @@ -89,16 +89,16 @@ xfs_scrub_allocbt_xref( > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_allocbt_xref_other(sc, agbno, len); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, len); > - xfs_scrub_xref_has_no_owner(sc, agbno, len); > - xfs_scrub_xref_is_not_shared(sc, agbno, len); > + xchk_allocbt_xref_other(sc, agbno, len); > + xchk_xref_is_not_inode_chunk(sc, agbno, len); > + xchk_xref_has_no_owner(sc, agbno, len); > + xchk_xref_is_not_shared(sc, agbno, len); > } > > /* Scrub a bnobt/cntbt record. */ > STATIC int > -xfs_scrub_allocbt_rec( > - struct xfs_scrub_btree *bs, > +xchk_allocbt_rec( > + struct xchk_btree *bs, > union xfs_btree_rec *rec) > { > struct xfs_mount *mp = bs->cur->bc_mp; > @@ -113,16 +113,16 @@ xfs_scrub_allocbt_rec( > if (bno + len <= bno || > !xfs_verify_agbno(mp, agno, bno) || > !xfs_verify_agbno(mp, agno, bno + len - 1)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > - xfs_scrub_allocbt_xref(bs->sc, bno, len); > + xchk_allocbt_xref(bs->sc, bno, len); > > return error; > } > > /* Scrub the freespace btrees for some AG. */ > STATIC int > -xfs_scrub_allocbt( > +xchk_allocbt( > struct xfs_scrub_context *sc, > xfs_btnum_t which) > { > @@ -131,26 +131,26 @@ xfs_scrub_allocbt( > > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_AG); > cur = which == XFS_BTNUM_BNO ? sc->sa.bno_cur : sc->sa.cnt_cur; > - return xfs_scrub_btree(sc, cur, xfs_scrub_allocbt_rec, &oinfo, NULL); > + return xchk_btree(sc, cur, xchk_allocbt_rec, &oinfo, NULL); > } > > int > -xfs_scrub_bnobt( > +xchk_bnobt( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_allocbt(sc, XFS_BTNUM_BNO); > + return xchk_allocbt(sc, XFS_BTNUM_BNO); > } > > int > -xfs_scrub_cntbt( > +xchk_cntbt( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_allocbt(sc, XFS_BTNUM_CNT); > + return xchk_allocbt(sc, XFS_BTNUM_CNT); > } > > /* xref check that the extent is not free */ > void > -xfs_scrub_xref_is_used_space( > +xchk_xref_is_used_space( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > @@ -158,12 +158,12 @@ xfs_scrub_xref_is_used_space( > bool is_freesp; > int error; > > - if (!sc->sa.bno_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.bno_cur || xchk_skip_xref(sc->sm)) > return; > > error = xfs_alloc_has_record(sc->sa.bno_cur, agbno, len, &is_freesp); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.bno_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.bno_cur)) > return; > if (is_freesp) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.bno_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.bno_cur, 0); > } > diff --git a/fs/xfs/scrub/attr.c b/fs/xfs/scrub/attr.c > index de51cf8a8516..0068bebddf3e 100644 > --- a/fs/xfs/scrub/attr.c > +++ b/fs/xfs/scrub/attr.c > @@ -32,7 +32,7 @@ > > /* Set us up to scrub an inode's extended attributes. */ > int > -xfs_scrub_setup_xattr( > +xchk_setup_xattr( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > @@ -50,12 +50,12 @@ xfs_scrub_setup_xattr( > if (!sc->buf) > return -ENOMEM; > > - return xfs_scrub_setup_inode_contents(sc, ip, 0); > + return xchk_setup_inode_contents(sc, ip, 0); > } > > /* Extended Attributes */ > > -struct xfs_scrub_xattr { > +struct xchk_xattr { > struct xfs_attr_list_context context; > struct xfs_scrub_context *sc; > }; > @@ -69,22 +69,22 @@ struct xfs_scrub_xattr { > * or if we get more or less data than we expected. > */ > static void > -xfs_scrub_xattr_listent( > +xchk_xattr_listent( > struct xfs_attr_list_context *context, > int flags, > unsigned char *name, > int namelen, > int valuelen) > { > - struct xfs_scrub_xattr *sx; > + struct xchk_xattr *sx; > struct xfs_da_args args = { NULL }; > int error = 0; > > - sx = container_of(context, struct xfs_scrub_xattr, context); > + sx = container_of(context, struct xchk_xattr, context); > > if (flags & XFS_ATTR_INCOMPLETE) { > /* Incomplete attr key, just mark the inode for preening. */ > - xfs_scrub_ino_set_preen(sx->sc, context->dp->i_ino); > + xchk_ino_set_preen(sx->sc, context->dp->i_ino); > return; > } > > @@ -106,11 +106,11 @@ xfs_scrub_xattr_listent( > error = xfs_attr_get_ilocked(context->dp, &args); > if (error == -EEXIST) > error = 0; > - if (!xfs_scrub_fblock_process_error(sx->sc, XFS_ATTR_FORK, args.blkno, > + if (!xchk_fblock_process_error(sx->sc, XFS_ATTR_FORK, args.blkno, > &error)) > goto fail_xref; > if (args.valuelen != valuelen) > - xfs_scrub_fblock_set_corrupt(sx->sc, XFS_ATTR_FORK, > + xchk_fblock_set_corrupt(sx->sc, XFS_ATTR_FORK, > args.blkno); > fail_xref: > if (sx->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > @@ -126,7 +126,7 @@ xfs_scrub_xattr_listent( > * the smallest address > */ > STATIC bool > -xfs_scrub_xattr_set_map( > +xchk_xattr_set_map( > struct xfs_scrub_context *sc, > unsigned long *map, > unsigned int start, > @@ -154,7 +154,7 @@ xfs_scrub_xattr_set_map( > * attr freemap has problems or points to used space. > */ > STATIC bool > -xfs_scrub_xattr_check_freemap( > +xchk_xattr_check_freemap( > struct xfs_scrub_context *sc, > unsigned long *map, > struct xfs_attr3_icleaf_hdr *leafhdr) > @@ -168,7 +168,7 @@ xfs_scrub_xattr_check_freemap( > freemap = (unsigned long *)sc->buf + BITS_TO_LONGS(mapsize); > bitmap_zero(freemap, mapsize); > for (i = 0; i < XFS_ATTR_LEAF_MAPSIZE; i++) { > - if (!xfs_scrub_xattr_set_map(sc, freemap, > + if (!xchk_xattr_set_map(sc, freemap, > leafhdr->freemap[i].base, > leafhdr->freemap[i].size)) > return false; > @@ -184,8 +184,8 @@ xfs_scrub_xattr_check_freemap( > * Returns the number of bytes used for the name/value data. > */ > STATIC void > -xfs_scrub_xattr_entry( > - struct xfs_scrub_da_btree *ds, > +xchk_xattr_entry( > + struct xchk_da_btree *ds, > int level, > char *buf_end, > struct xfs_attr_leafblock *leaf, > @@ -204,17 +204,17 @@ xfs_scrub_xattr_entry( > unsigned int namesize; > > if (ent->pad2 != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > /* Hash values in order? */ > if (be32_to_cpu(ent->hashval) < *last_hashval) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > *last_hashval = be32_to_cpu(ent->hashval); > > nameidx = be16_to_cpu(ent->nameidx); > if (nameidx < leafhdr->firstused || > nameidx >= mp->m_attr_geo->blksize) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > return; > } > > @@ -225,27 +225,27 @@ xfs_scrub_xattr_entry( > be16_to_cpu(lentry->valuelen)); > name_end = (char *)lentry + namesize; > if (lentry->namelen == 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > } else { > rentry = xfs_attr3_leaf_name_remote(leaf, idx); > namesize = xfs_attr_leaf_entsize_remote(rentry->namelen); > name_end = (char *)rentry + namesize; > if (rentry->namelen == 0 || rentry->valueblk == 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > } > if (name_end > buf_end) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > - if (!xfs_scrub_xattr_set_map(ds->sc, usedmap, nameidx, namesize)) > - xfs_scrub_da_set_corrupt(ds, level); > + if (!xchk_xattr_set_map(ds->sc, usedmap, nameidx, namesize)) > + xchk_da_set_corrupt(ds, level); > if (!(ds->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > *usedbytes += namesize; > } > > /* Scrub an attribute leaf. */ > STATIC int > -xfs_scrub_xattr_block( > - struct xfs_scrub_da_btree *ds, > +xchk_xattr_block( > + struct xchk_da_btree *ds, > int level) > { > struct xfs_attr3_icleaf_hdr leafhdr; > @@ -275,10 +275,10 @@ xfs_scrub_xattr_block( > > if (leaf->hdr.pad1 != 0 || leaf->hdr.pad2 != 0 || > leaf->hdr.info.hdr.pad != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > } else { > if (leaf->hdr.pad1 != 0 || leaf->hdr.info.pad != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > } > > /* Check the leaf header */ > @@ -286,44 +286,44 @@ xfs_scrub_xattr_block( > hdrsize = xfs_attr3_leaf_hdr_size(leaf); > > if (leafhdr.usedbytes > mp->m_attr_geo->blksize) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > if (leafhdr.firstused > mp->m_attr_geo->blksize) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > if (leafhdr.firstused < hdrsize) > - xfs_scrub_da_set_corrupt(ds, level); > - if (!xfs_scrub_xattr_set_map(ds->sc, usedmap, 0, hdrsize)) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > + if (!xchk_xattr_set_map(ds->sc, usedmap, 0, hdrsize)) > + xchk_da_set_corrupt(ds, level); > > if (ds->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > > entries = xfs_attr3_leaf_entryp(leaf); > if ((char *)&entries[leafhdr.count] > (char *)leaf + leafhdr.firstused) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > buf_end = (char *)bp->b_addr + mp->m_attr_geo->blksize; > for (i = 0, ent = entries; i < leafhdr.count; ent++, i++) { > /* Mark the leaf entry itself. */ > off = (char *)ent - (char *)leaf; > - if (!xfs_scrub_xattr_set_map(ds->sc, usedmap, off, > + if (!xchk_xattr_set_map(ds->sc, usedmap, off, > sizeof(xfs_attr_leaf_entry_t))) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out; > } > > /* Check the entry and nameval. */ > - xfs_scrub_xattr_entry(ds, level, buf_end, leaf, &leafhdr, > + xchk_xattr_entry(ds, level, buf_end, leaf, &leafhdr, > usedmap, ent, i, &usedbytes, &last_hashval); > > if (ds->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > } > > - if (!xfs_scrub_xattr_check_freemap(ds->sc, usedmap, &leafhdr)) > - xfs_scrub_da_set_corrupt(ds, level); > + if (!xchk_xattr_check_freemap(ds->sc, usedmap, &leafhdr)) > + xchk_da_set_corrupt(ds, level); > > if (leafhdr.usedbytes != usedbytes) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > out: > return 0; > @@ -331,8 +331,8 @@ xfs_scrub_xattr_block( > > /* Scrub a attribute btree record. */ > STATIC int > -xfs_scrub_xattr_rec( > - struct xfs_scrub_da_btree *ds, > +xchk_xattr_rec( > + struct xchk_da_btree *ds, > int level, > void *rec) > { > @@ -352,14 +352,14 @@ xfs_scrub_xattr_rec( > blk = &ds->state->path.blk[level]; > > /* Check the whole block, if necessary. */ > - error = xfs_scrub_xattr_block(ds, level); > + error = xchk_xattr_block(ds, level); > if (error) > goto out; > if (ds->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > > /* Check the hash of the entry. */ > - error = xfs_scrub_da_btree_hash(ds, level, &ent->hashval); > + error = xchk_da_btree_hash(ds, level, &ent->hashval); > if (error) > goto out; > > @@ -368,7 +368,7 @@ xfs_scrub_xattr_rec( > hdrsize = xfs_attr3_leaf_hdr_size(bp->b_addr); > nameidx = be16_to_cpu(ent->nameidx); > if (nameidx < hdrsize || nameidx >= mp->m_attr_geo->blksize) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out; > } > > @@ -377,12 +377,12 @@ xfs_scrub_xattr_rec( > badflags = ~(XFS_ATTR_LOCAL | XFS_ATTR_ROOT | XFS_ATTR_SECURE | > XFS_ATTR_INCOMPLETE); > if ((ent->flags & badflags) != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > if (ent->flags & XFS_ATTR_LOCAL) { > lentry = (struct xfs_attr_leaf_name_local *) > (((char *)bp->b_addr) + nameidx); > if (lentry->namelen <= 0) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out; > } > calc_hash = xfs_da_hashname(lentry->nameval, lentry->namelen); > @@ -390,13 +390,13 @@ xfs_scrub_xattr_rec( > rentry = (struct xfs_attr_leaf_name_remote *) > (((char *)bp->b_addr) + nameidx); > if (rentry->namelen <= 0) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out; > } > calc_hash = xfs_da_hashname(rentry->name, rentry->namelen); > } > if (calc_hash != hash) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > out: > return error; > @@ -404,10 +404,10 @@ xfs_scrub_xattr_rec( > > /* Scrub the extended attribute metadata. */ > int > -xfs_scrub_xattr( > +xchk_xattr( > struct xfs_scrub_context *sc) > { > - struct xfs_scrub_xattr sx; > + struct xchk_xattr sx; > struct attrlist_cursor_kern cursor = { 0 }; > xfs_dablk_t last_checked = -1U; > int error = 0; > @@ -417,7 +417,7 @@ xfs_scrub_xattr( > > memset(&sx, 0, sizeof(sx)); > /* Check attribute tree structure */ > - error = xfs_scrub_da_btree(sc, XFS_ATTR_FORK, xfs_scrub_xattr_rec, > + error = xchk_da_btree(sc, XFS_ATTR_FORK, xchk_xattr_rec, > &last_checked); > if (error) > goto out; > @@ -429,7 +429,7 @@ xfs_scrub_xattr( > sx.context.dp = sc->ip; > sx.context.cursor = &cursor; > sx.context.resynch = 1; > - sx.context.put_listent = xfs_scrub_xattr_listent; > + sx.context.put_listent = xchk_xattr_listent; > sx.context.tp = sc->tp; > sx.context.flags = ATTR_INCOMPLETE; > sx.sc = sc; > @@ -438,7 +438,7 @@ xfs_scrub_xattr( > * Look up every xattr in this file by name. > * > * Use the backend implementation of xfs_attr_list to call > - * xfs_scrub_xattr_listent on every attribute key in this inode. > + * xchk_xattr_listent on every attribute key in this inode. > * In other words, we use the same iterator/callback mechanism > * that listattr uses to scrub extended attributes, though in our > * _listent function, we check the value of the attribute. > @@ -451,7 +451,7 @@ xfs_scrub_xattr( > * locking order. > */ > error = xfs_attr_list_int_ilocked(&sx.context); > - if (!xfs_scrub_fblock_process_error(sc, XFS_ATTR_FORK, 0, &error)) > + if (!xchk_fblock_process_error(sc, XFS_ATTR_FORK, 0, &error)) > goto out; > out: > return error; > diff --git a/fs/xfs/scrub/bmap.c b/fs/xfs/scrub/bmap.c > index ebbfab173e97..19cfbd3910a2 100644 > --- a/fs/xfs/scrub/bmap.c > +++ b/fs/xfs/scrub/bmap.c > @@ -33,13 +33,13 @@ > > /* Set us up with an inode's bmap. */ > int > -xfs_scrub_setup_inode_bmap( > +xchk_setup_inode_bmap( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > int error; > > - error = xfs_scrub_get_inode(sc, ip); > + error = xchk_get_inode(sc, ip); > if (error) > goto out; > > @@ -60,7 +60,7 @@ xfs_scrub_setup_inode_bmap( > } > > /* Got the inode, lock it and we're ready to go. */ > - error = xfs_scrub_trans_alloc(sc, 0); > + error = xchk_trans_alloc(sc, 0); > if (error) > goto out; > sc->ilock_flags |= XFS_ILOCK_EXCL; > @@ -78,7 +78,7 @@ xfs_scrub_setup_inode_bmap( > * is in btree format. > */ > > -struct xfs_scrub_bmap_info { > +struct xchk_bmap_info { > struct xfs_scrub_context *sc; > xfs_fileoff_t lastoff; > bool is_rt; > @@ -88,8 +88,8 @@ struct xfs_scrub_bmap_info { > > /* Look for a corresponding rmap for this irec. */ > static inline bool > -xfs_scrub_bmap_get_rmap( > - struct xfs_scrub_bmap_info *info, > +xchk_bmap_get_rmap( > + struct xchk_bmap_info *info, > struct xfs_bmbt_irec *irec, > xfs_agblock_t agbno, > uint64_t owner, > @@ -120,7 +120,7 @@ xfs_scrub_bmap_get_rmap( > if (info->is_shared) { > error = xfs_rmap_lookup_le_range(info->sc->sa.rmap_cur, agbno, > owner, offset, rflags, rmap, &has_rmap); > - if (!xfs_scrub_should_check_xref(info->sc, &error, > + if (!xchk_should_check_xref(info->sc, &error, > &info->sc->sa.rmap_cur)) > return false; > goto out; > @@ -131,28 +131,28 @@ xfs_scrub_bmap_get_rmap( > */ > error = xfs_rmap_lookup_le(info->sc->sa.rmap_cur, agbno, 0, owner, > offset, rflags, &has_rmap); > - if (!xfs_scrub_should_check_xref(info->sc, &error, > + if (!xchk_should_check_xref(info->sc, &error, > &info->sc->sa.rmap_cur)) > return false; > if (!has_rmap) > goto out; > > error = xfs_rmap_get_rec(info->sc->sa.rmap_cur, rmap, &has_rmap); > - if (!xfs_scrub_should_check_xref(info->sc, &error, > + if (!xchk_should_check_xref(info->sc, &error, > &info->sc->sa.rmap_cur)) > return false; > > out: > if (!has_rmap) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > return has_rmap; > } > > /* Make sure that we have rmapbt records for this extent. */ > STATIC void > -xfs_scrub_bmap_xref_rmap( > - struct xfs_scrub_bmap_info *info, > +xchk_bmap_xref_rmap( > + struct xchk_bmap_info *info, > struct xfs_bmbt_irec *irec, > xfs_agblock_t agbno) > { > @@ -160,7 +160,7 @@ xfs_scrub_bmap_xref_rmap( > unsigned long long rmap_end; > uint64_t owner; > > - if (!info->sc->sa.rmap_cur || xfs_scrub_skip_xref(info->sc->sm)) > + if (!info->sc->sa.rmap_cur || xchk_skip_xref(info->sc->sm)) > return; > > if (info->whichfork == XFS_COW_FORK) > @@ -169,14 +169,14 @@ xfs_scrub_bmap_xref_rmap( > owner = info->sc->ip->i_ino; > > /* Find the rmap record for this irec. */ > - if (!xfs_scrub_bmap_get_rmap(info, irec, agbno, owner, &rmap)) > + if (!xchk_bmap_get_rmap(info, irec, agbno, owner, &rmap)) > return; > > /* Check the rmap. */ > rmap_end = (unsigned long long)rmap.rm_startblock + rmap.rm_blockcount; > if (rmap.rm_startblock > agbno || > agbno + irec->br_blockcount > rmap_end) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* > @@ -189,12 +189,12 @@ xfs_scrub_bmap_xref_rmap( > rmap.rm_blockcount; > if (rmap.rm_offset > irec->br_startoff || > irec->br_startoff + irec->br_blockcount > rmap_end) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, > + xchk_fblock_xref_set_corrupt(info->sc, > info->whichfork, irec->br_startoff); > } > > if (rmap.rm_owner != owner) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* > @@ -207,22 +207,22 @@ xfs_scrub_bmap_xref_rmap( > if (owner != XFS_RMAP_OWN_COW && > irec->br_state == XFS_EXT_UNWRITTEN && > !(rmap.rm_flags & XFS_RMAP_UNWRITTEN)) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > if (info->whichfork == XFS_ATTR_FORK && > !(rmap.rm_flags & XFS_RMAP_ATTR_FORK)) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > if (rmap.rm_flags & XFS_RMAP_BMBT_BLOCK) > - xfs_scrub_fblock_xref_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_xref_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > } > > /* Cross-reference a single rtdev extent record. */ > STATIC void > -xfs_scrub_bmap_rt_extent_xref( > - struct xfs_scrub_bmap_info *info, > +xchk_bmap_rt_extent_xref( > + struct xchk_bmap_info *info, > struct xfs_inode *ip, > struct xfs_btree_cur *cur, > struct xfs_bmbt_irec *irec) > @@ -230,14 +230,14 @@ xfs_scrub_bmap_rt_extent_xref( > if (info->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_xref_is_used_rt_space(info->sc, irec->br_startblock, > + xchk_xref_is_used_rt_space(info->sc, irec->br_startblock, > irec->br_blockcount); > } > > /* Cross-reference a single datadev extent record. */ > STATIC void > -xfs_scrub_bmap_extent_xref( > - struct xfs_scrub_bmap_info *info, > +xchk_bmap_extent_xref( > + struct xchk_bmap_info *info, > struct xfs_inode *ip, > struct xfs_btree_cur *cur, > struct xfs_bmbt_irec *irec) > @@ -255,38 +255,38 @@ xfs_scrub_bmap_extent_xref( > agbno = XFS_FSB_TO_AGBNO(mp, irec->br_startblock); > len = irec->br_blockcount; > > - error = xfs_scrub_ag_init(info->sc, agno, &info->sc->sa); > - if (!xfs_scrub_fblock_process_error(info->sc, info->whichfork, > + error = xchk_ag_init(info->sc, agno, &info->sc->sa); > + if (!xchk_fblock_process_error(info->sc, info->whichfork, > irec->br_startoff, &error)) > return; > > - xfs_scrub_xref_is_used_space(info->sc, agbno, len); > - xfs_scrub_xref_is_not_inode_chunk(info->sc, agbno, len); > - xfs_scrub_bmap_xref_rmap(info, irec, agbno); > + xchk_xref_is_used_space(info->sc, agbno, len); > + xchk_xref_is_not_inode_chunk(info->sc, agbno, len); > + xchk_bmap_xref_rmap(info, irec, agbno); > switch (info->whichfork) { > case XFS_DATA_FORK: > if (xfs_is_reflink_inode(info->sc->ip)) > break; > /* fall through */ > case XFS_ATTR_FORK: > - xfs_scrub_xref_is_not_shared(info->sc, agbno, > + xchk_xref_is_not_shared(info->sc, agbno, > irec->br_blockcount); > break; > case XFS_COW_FORK: > - xfs_scrub_xref_is_cow_staging(info->sc, agbno, > + xchk_xref_is_cow_staging(info->sc, agbno, > irec->br_blockcount); > break; > } > > - xfs_scrub_ag_free(info->sc, &info->sc->sa); > + xchk_ag_free(info->sc, &info->sc->sa); > } > > /* Scrub a single extent record. */ > STATIC int > -xfs_scrub_bmap_extent( > +xchk_bmap_extent( > struct xfs_inode *ip, > struct xfs_btree_cur *cur, > - struct xfs_scrub_bmap_info *info, > + struct xchk_bmap_info *info, > struct xfs_bmbt_irec *irec) > { > struct xfs_mount *mp = info->sc->mp; > @@ -302,12 +302,12 @@ xfs_scrub_bmap_extent( > * from the incore list, for which there is no ordering check. > */ > if (irec->br_startoff < info->lastoff) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* There should never be a "hole" extent in either extent list. */ > if (irec->br_startblock == HOLESTARTBLOCK) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* > @@ -315,40 +315,40 @@ xfs_scrub_bmap_extent( > * in-core extent scan, and we should never see these in the bmbt. > */ > if (isnullstartblock(irec->br_startblock)) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* Make sure the extent points to a valid place. */ > if (irec->br_blockcount > MAXEXTLEN) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > if (irec->br_startblock + irec->br_blockcount <= irec->br_startblock) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > end = irec->br_startblock + irec->br_blockcount - 1; > if (info->is_rt && > (!xfs_verify_rtbno(mp, irec->br_startblock) || > !xfs_verify_rtbno(mp, end))) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > if (!info->is_rt && > (!xfs_verify_fsbno(mp, irec->br_startblock) || > !xfs_verify_fsbno(mp, end) || > XFS_FSB_TO_AGNO(mp, irec->br_startblock) != > XFS_FSB_TO_AGNO(mp, end))) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > /* We don't allow unwritten extents on attr forks. */ > if (irec->br_state == XFS_EXT_UNWRITTEN && > info->whichfork == XFS_ATTR_FORK) > - xfs_scrub_fblock_set_corrupt(info->sc, info->whichfork, > + xchk_fblock_set_corrupt(info->sc, info->whichfork, > irec->br_startoff); > > if (info->is_rt) > - xfs_scrub_bmap_rt_extent_xref(info, ip, cur, irec); > + xchk_bmap_rt_extent_xref(info, ip, cur, irec); > else > - xfs_scrub_bmap_extent_xref(info, ip, cur, irec); > + xchk_bmap_extent_xref(info, ip, cur, irec); > > info->lastoff = irec->br_startoff + irec->br_blockcount; > return error; > @@ -356,12 +356,12 @@ xfs_scrub_bmap_extent( > > /* Scrub a bmbt record. */ > STATIC int > -xfs_scrub_bmapbt_rec( > - struct xfs_scrub_btree *bs, > +xchk_bmapbt_rec( > + struct xchk_btree *bs, > union xfs_btree_rec *rec) > { > struct xfs_bmbt_irec irec; > - struct xfs_scrub_bmap_info *info = bs->private; > + struct xchk_bmap_info *info = bs->private; > struct xfs_inode *ip = bs->cur->bc_private.b.ip; > struct xfs_buf *bp = NULL; > struct xfs_btree_block *block; > @@ -378,22 +378,22 @@ xfs_scrub_bmapbt_rec( > block = xfs_btree_get_block(bs->cur, i, &bp); > owner = be64_to_cpu(block->bb_u.l.bb_owner); > if (owner != ip->i_ino) > - xfs_scrub_fblock_set_corrupt(bs->sc, > + xchk_fblock_set_corrupt(bs->sc, > info->whichfork, 0); > } > } > > /* Set up the in-core record and scrub it. */ > xfs_bmbt_disk_get_all(&rec->bmbt, &irec); > - return xfs_scrub_bmap_extent(ip, bs->cur, info, &irec); > + return xchk_bmap_extent(ip, bs->cur, info, &irec); > } > > /* Scan the btree records. */ > STATIC int > -xfs_scrub_bmap_btree( > +xchk_bmap_btree( > struct xfs_scrub_context *sc, > int whichfork, > - struct xfs_scrub_bmap_info *info) > + struct xchk_bmap_info *info) > { > struct xfs_owner_info oinfo; > struct xfs_mount *mp = sc->mp; > @@ -403,12 +403,12 @@ xfs_scrub_bmap_btree( > > cur = xfs_bmbt_init_cursor(mp, sc->tp, ip, whichfork); > xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, whichfork); > - error = xfs_scrub_btree(sc, cur, xfs_scrub_bmapbt_rec, &oinfo, info); > + error = xchk_btree(sc, cur, xchk_bmapbt_rec, &oinfo, info); > xfs_btree_del_cursor(cur, error); > return error; > } > > -struct xfs_scrub_bmap_check_rmap_info { > +struct xchk_bmap_check_rmap_info { > struct xfs_scrub_context *sc; > int whichfork; > struct xfs_iext_cursor icur; > @@ -416,13 +416,13 @@ struct xfs_scrub_bmap_check_rmap_info { > > /* Can we find bmaps that fit this rmap? */ > STATIC int > -xfs_scrub_bmap_check_rmap( > +xchk_bmap_check_rmap( > struct xfs_btree_cur *cur, > struct xfs_rmap_irec *rec, > void *priv) > { > struct xfs_bmbt_irec irec; > - struct xfs_scrub_bmap_check_rmap_info *sbcri = priv; > + struct xchk_bmap_check_rmap_info *sbcri = priv; > struct xfs_ifork *ifp; > struct xfs_scrub_context *sc = sbcri->sc; > bool have_map; > @@ -439,14 +439,14 @@ xfs_scrub_bmap_check_rmap( > /* Now look up the bmbt record. */ > ifp = XFS_IFORK_PTR(sc->ip, sbcri->whichfork); > if (!ifp) { > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > goto out; > } > have_map = xfs_iext_lookup_extent(sc->ip, ifp, rec->rm_offset, > &sbcri->icur, &irec); > if (!have_map) > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > /* > * bmap extent record lengths are constrained to 2^21 blocks in length > @@ -457,14 +457,14 @@ xfs_scrub_bmap_check_rmap( > */ > while (have_map) { > if (irec.br_startoff != rec->rm_offset) > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > if (irec.br_startblock != XFS_AGB_TO_FSB(sc->mp, > cur->bc_private.a.agno, rec->rm_startblock)) > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > if (irec.br_blockcount > rec->rm_blockcount) > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > break; > @@ -475,7 +475,7 @@ xfs_scrub_bmap_check_rmap( > break; > have_map = xfs_iext_next_extent(ifp, &sbcri->icur, &irec); > if (!have_map) > - xfs_scrub_fblock_set_corrupt(sc, sbcri->whichfork, > + xchk_fblock_set_corrupt(sc, sbcri->whichfork, > rec->rm_offset); > } > > @@ -487,12 +487,12 @@ xfs_scrub_bmap_check_rmap( > > /* Make sure each rmap has a corresponding bmbt entry. */ > STATIC int > -xfs_scrub_bmap_check_ag_rmaps( > +xchk_bmap_check_ag_rmaps( > struct xfs_scrub_context *sc, > int whichfork, > xfs_agnumber_t agno) > { > - struct xfs_scrub_bmap_check_rmap_info sbcri; > + struct xchk_bmap_check_rmap_info sbcri; > struct xfs_btree_cur *cur; > struct xfs_buf *agf; > int error; > @@ -509,7 +509,7 @@ xfs_scrub_bmap_check_ag_rmaps( > > sbcri.sc = sc; > sbcri.whichfork = whichfork; > - error = xfs_rmap_query_all(cur, xfs_scrub_bmap_check_rmap, &sbcri); > + error = xfs_rmap_query_all(cur, xchk_bmap_check_rmap, &sbcri); > if (error == XFS_BTREE_QUERY_RANGE_ABORT) > error = 0; > > @@ -521,7 +521,7 @@ xfs_scrub_bmap_check_ag_rmaps( > > /* Make sure each rmap has a corresponding bmbt entry. */ > STATIC int > -xfs_scrub_bmap_check_rmaps( > +xchk_bmap_check_rmaps( > struct xfs_scrub_context *sc, > int whichfork) > { > @@ -561,7 +561,7 @@ xfs_scrub_bmap_check_rmaps( > return 0; > > for (agno = 0; agno < sc->mp->m_sb.sb_agcount; agno++) { > - error = xfs_scrub_bmap_check_ag_rmaps(sc, whichfork, agno); > + error = xchk_bmap_check_ag_rmaps(sc, whichfork, agno); > if (error) > return error; > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > @@ -578,12 +578,12 @@ xfs_scrub_bmap_check_rmaps( > * Then we unconditionally scan the incore extent cache. > */ > STATIC int > -xfs_scrub_bmap( > +xchk_bmap( > struct xfs_scrub_context *sc, > int whichfork) > { > struct xfs_bmbt_irec irec; > - struct xfs_scrub_bmap_info info = { NULL }; > + struct xchk_bmap_info info = { NULL }; > struct xfs_mount *mp = sc->mp; > struct xfs_inode *ip = sc->ip; > struct xfs_ifork *ifp; > @@ -605,7 +605,7 @@ xfs_scrub_bmap( > goto out; > /* No CoW forks on non-reflink inodes/filesystems. */ > if (!xfs_is_reflink_inode(ip)) { > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > goto out; > } > break; > @@ -614,7 +614,7 @@ xfs_scrub_bmap( > goto out_check_rmap; > if (!xfs_sb_version_hasattr(&mp->m_sb) && > !xfs_sb_version_hasattr2(&mp->m_sb)) > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > break; > default: > ASSERT(whichfork == XFS_DATA_FORK); > @@ -630,22 +630,22 @@ xfs_scrub_bmap( > goto out; > case XFS_DINODE_FMT_EXTENTS: > if (!(ifp->if_flags & XFS_IFEXTENTS)) { > - xfs_scrub_fblock_set_corrupt(sc, whichfork, 0); > + xchk_fblock_set_corrupt(sc, whichfork, 0); > goto out; > } > break; > case XFS_DINODE_FMT_BTREE: > if (whichfork == XFS_COW_FORK) { > - xfs_scrub_fblock_set_corrupt(sc, whichfork, 0); > + xchk_fblock_set_corrupt(sc, whichfork, 0); > goto out; > } > > - error = xfs_scrub_bmap_btree(sc, whichfork, &info); > + error = xchk_bmap_btree(sc, whichfork, &info); > if (error) > goto out; > break; > default: > - xfs_scrub_fblock_set_corrupt(sc, whichfork, 0); > + xchk_fblock_set_corrupt(sc, whichfork, 0); > goto out; > } > > @@ -655,37 +655,37 @@ xfs_scrub_bmap( > /* Now try to scrub the in-memory extent list. */ > if (!(ifp->if_flags & XFS_IFEXTENTS)) { > error = xfs_iread_extents(sc->tp, ip, whichfork); > - if (!xfs_scrub_fblock_process_error(sc, whichfork, 0, &error)) > + if (!xchk_fblock_process_error(sc, whichfork, 0, &error)) > goto out; > } > > /* Find the offset of the last extent in the mapping. */ > error = xfs_bmap_last_offset(ip, &endoff, whichfork); > - if (!xfs_scrub_fblock_process_error(sc, whichfork, 0, &error)) > + if (!xchk_fblock_process_error(sc, whichfork, 0, &error)) > goto out; > > /* Scrub extent records. */ > info.lastoff = 0; > ifp = XFS_IFORK_PTR(ip, whichfork); > for_each_xfs_iext(ifp, &icur, &irec) { > - if (xfs_scrub_should_terminate(sc, &error) || > + if (xchk_should_terminate(sc, &error) || > (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > break; > if (isnullstartblock(irec.br_startblock)) > continue; > if (irec.br_startoff >= endoff) { > - xfs_scrub_fblock_set_corrupt(sc, whichfork, > + xchk_fblock_set_corrupt(sc, whichfork, > irec.br_startoff); > goto out; > } > - error = xfs_scrub_bmap_extent(ip, NULL, &info, &irec); > + error = xchk_bmap_extent(ip, NULL, &info, &irec); > if (error) > goto out; > } > > out_check_rmap: > - error = xfs_scrub_bmap_check_rmaps(sc, whichfork); > - if (!xfs_scrub_fblock_xref_process_error(sc, whichfork, 0, &error)) > + error = xchk_bmap_check_rmaps(sc, whichfork); > + if (!xchk_fblock_xref_process_error(sc, whichfork, 0, &error)) > goto out; > out: > return error; > @@ -693,27 +693,27 @@ xfs_scrub_bmap( > > /* Scrub an inode's data fork. */ > int > -xfs_scrub_bmap_data( > +xchk_bmap_data( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_bmap(sc, XFS_DATA_FORK); > + return xchk_bmap(sc, XFS_DATA_FORK); > } > > /* Scrub an inode's attr fork. */ > int > -xfs_scrub_bmap_attr( > +xchk_bmap_attr( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_bmap(sc, XFS_ATTR_FORK); > + return xchk_bmap(sc, XFS_ATTR_FORK); > } > > /* Scrub an inode's CoW fork. */ > int > -xfs_scrub_bmap_cow( > +xchk_bmap_cow( > struct xfs_scrub_context *sc) > { > if (!xfs_is_reflink_inode(sc->ip)) > return -ENOENT; > > - return xfs_scrub_bmap(sc, XFS_COW_FORK); > + return xchk_bmap(sc, XFS_COW_FORK); > } > diff --git a/fs/xfs/scrub/btree.c b/fs/xfs/scrub/btree.c > index 5b472045f036..30fe9a147959 100644 > --- a/fs/xfs/scrub/btree.c > +++ b/fs/xfs/scrub/btree.c > @@ -29,7 +29,7 @@ > * operational errors in common.c. > */ > static bool > -__xfs_scrub_btree_process_error( > +__xchk_btree_process_error( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level, > @@ -43,7 +43,7 @@ __xfs_scrub_btree_process_error( > switch (*error) { > case -EDEADLOCK: > /* Used to restart an op with deadlock avoidance. */ > - trace_xfs_scrub_deadlock_retry(sc->ip, sc->sm, *error); > + trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); > break; > case -EFSBADCRC: > case -EFSCORRUPTED: > @@ -53,10 +53,10 @@ __xfs_scrub_btree_process_error( > /* fall through */ > default: > if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) > - trace_xfs_scrub_ifork_btree_op_error(sc, cur, level, > + trace_xchk_ifork_btree_op_error(sc, cur, level, > *error, ret_ip); > else > - trace_xfs_scrub_btree_op_error(sc, cur, level, > + trace_xchk_btree_op_error(sc, cur, level, > *error, ret_ip); > break; > } > @@ -64,30 +64,30 @@ __xfs_scrub_btree_process_error( > } > > bool > -xfs_scrub_btree_process_error( > +xchk_btree_process_error( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level, > int *error) > { > - return __xfs_scrub_btree_process_error(sc, cur, level, error, > + return __xchk_btree_process_error(sc, cur, level, error, > XFS_SCRUB_OFLAG_CORRUPT, __return_address); > } > > bool > -xfs_scrub_btree_xref_process_error( > +xchk_btree_xref_process_error( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level, > int *error) > { > - return __xfs_scrub_btree_process_error(sc, cur, level, error, > + return __xchk_btree_process_error(sc, cur, level, error, > XFS_SCRUB_OFLAG_XFAIL, __return_address); > } > > /* Record btree block corruption. */ > static void > -__xfs_scrub_btree_set_corrupt( > +__xchk_btree_set_corrupt( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level, > @@ -97,30 +97,30 @@ __xfs_scrub_btree_set_corrupt( > sc->sm->sm_flags |= errflag; > > if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) > - trace_xfs_scrub_ifork_btree_error(sc, cur, level, > + trace_xchk_ifork_btree_error(sc, cur, level, > ret_ip); > else > - trace_xfs_scrub_btree_error(sc, cur, level, > + trace_xchk_btree_error(sc, cur, level, > ret_ip); > } > > void > -xfs_scrub_btree_set_corrupt( > +xchk_btree_set_corrupt( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level) > { > - __xfs_scrub_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_CORRUPT, > + __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_CORRUPT, > __return_address); > } > > void > -xfs_scrub_btree_xref_set_corrupt( > +xchk_btree_xref_set_corrupt( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > int level) > { > - __xfs_scrub_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_XCORRUPT, > + __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_XCORRUPT, > __return_address); > } > > @@ -129,8 +129,8 @@ xfs_scrub_btree_xref_set_corrupt( > * keys. > */ > STATIC void > -xfs_scrub_btree_rec( > - struct xfs_scrub_btree *bs) > +xchk_btree_rec( > + struct xchk_btree *bs) > { > struct xfs_btree_cur *cur = bs->cur; > union xfs_btree_rec *rec; > @@ -144,11 +144,11 @@ xfs_scrub_btree_rec( > block = xfs_btree_get_block(cur, 0, &bp); > rec = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block); > > - trace_xfs_scrub_btree_rec(bs->sc, cur, 0); > + trace_xchk_btree_rec(bs->sc, cur, 0); > > /* If this isn't the first record, are they in order? */ > if (!bs->firstrec && !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec)) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, 0); > + xchk_btree_set_corrupt(bs->sc, cur, 0); > bs->firstrec = false; > memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len); > > @@ -160,7 +160,7 @@ xfs_scrub_btree_rec( > keyblock = xfs_btree_get_block(cur, 1, &bp); > keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[1], keyblock); > if (cur->bc_ops->diff_two_keys(cur, &key, keyp) < 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, 1); > + xchk_btree_set_corrupt(bs->sc, cur, 1); > > if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) > return; > @@ -169,7 +169,7 @@ xfs_scrub_btree_rec( > cur->bc_ops->init_high_key_from_rec(&hkey, rec); > keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[1], keyblock); > if (cur->bc_ops->diff_two_keys(cur, keyp, &hkey) < 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, 1); > + xchk_btree_set_corrupt(bs->sc, cur, 1); > } > > /* > @@ -177,8 +177,8 @@ xfs_scrub_btree_rec( > * keys. > */ > STATIC void > -xfs_scrub_btree_key( > - struct xfs_scrub_btree *bs, > +xchk_btree_key( > + struct xchk_btree *bs, > int level) > { > struct xfs_btree_cur *cur = bs->cur; > @@ -191,12 +191,12 @@ xfs_scrub_btree_key( > block = xfs_btree_get_block(cur, level, &bp); > key = xfs_btree_key_addr(cur, cur->bc_ptrs[level], block); > > - trace_xfs_scrub_btree_key(bs->sc, cur, level); > + trace_xchk_btree_key(bs->sc, cur, level); > > /* If this isn't the first key, are they in order? */ > if (!bs->firstkey[level] && > !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level], key)) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > bs->firstkey[level] = false; > memcpy(&bs->lastkey[level], key, cur->bc_ops->key_len); > > @@ -207,7 +207,7 @@ xfs_scrub_btree_key( > keyblock = xfs_btree_get_block(cur, level + 1, &bp); > keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[level + 1], keyblock); > if (cur->bc_ops->diff_two_keys(cur, key, keyp) < 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > > if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) > return; > @@ -216,7 +216,7 @@ xfs_scrub_btree_key( > key = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level], block); > keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level + 1], keyblock); > if (cur->bc_ops->diff_two_keys(cur, keyp, key) < 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > } > > /* > @@ -224,8 +224,8 @@ xfs_scrub_btree_key( > * Callers do not need to set the corrupt flag. > */ > static bool > -xfs_scrub_btree_ptr_ok( > - struct xfs_scrub_btree *bs, > +xchk_btree_ptr_ok( > + struct xchk_btree *bs, > int level, > union xfs_btree_ptr *ptr) > { > @@ -242,15 +242,15 @@ xfs_scrub_btree_ptr_ok( > else > res = xfs_btree_check_sptr(bs->cur, be32_to_cpu(ptr->s), level); > if (!res) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, level); > + xchk_btree_set_corrupt(bs->sc, bs->cur, level); > > return res; > } > > /* Check that a btree block's sibling matches what we expect it. */ > STATIC int > -xfs_scrub_btree_block_check_sibling( > - struct xfs_scrub_btree *bs, > +xchk_btree_block_check_sibling( > + struct xchk_btree *bs, > int level, > int direction, > union xfs_btree_ptr *sibling) > @@ -264,7 +264,7 @@ xfs_scrub_btree_block_check_sibling( > int error; > > error = xfs_btree_dup_cursor(cur, &ncur); > - if (!xfs_scrub_btree_process_error(bs->sc, cur, level + 1, &error) || > + if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error) || > !ncur) > return error; > > @@ -278,7 +278,7 @@ xfs_scrub_btree_block_check_sibling( > else > error = xfs_btree_decrement(ncur, level + 1, &success); > if (error == 0 && success) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > error = 0; > goto out; > } > @@ -288,23 +288,23 @@ xfs_scrub_btree_block_check_sibling( > error = xfs_btree_increment(ncur, level + 1, &success); > else > error = xfs_btree_decrement(ncur, level + 1, &success); > - if (!xfs_scrub_btree_process_error(bs->sc, cur, level + 1, &error)) > + if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error)) > goto out; > if (!success) { > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level + 1); > + xchk_btree_set_corrupt(bs->sc, cur, level + 1); > goto out; > } > > /* Compare upper level pointer to sibling pointer. */ > pblock = xfs_btree_get_block(ncur, level + 1, &pbp); > pp = xfs_btree_ptr_addr(ncur, ncur->bc_ptrs[level + 1], pblock); > - if (!xfs_scrub_btree_ptr_ok(bs, level + 1, pp)) > + if (!xchk_btree_ptr_ok(bs, level + 1, pp)) > goto out; > if (pbp) > - xfs_scrub_buffer_recheck(bs->sc, pbp); > + xchk_buffer_recheck(bs->sc, pbp); > > if (xfs_btree_diff_two_ptrs(cur, pp, sibling)) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > out: > xfs_btree_del_cursor(ncur, XFS_BTREE_ERROR); > return error; > @@ -312,8 +312,8 @@ xfs_scrub_btree_block_check_sibling( > > /* Check the siblings of a btree block. */ > STATIC int > -xfs_scrub_btree_block_check_siblings( > - struct xfs_scrub_btree *bs, > +xchk_btree_block_check_siblings( > + struct xchk_btree *bs, > struct xfs_btree_block *block) > { > struct xfs_btree_cur *cur = bs->cur; > @@ -330,7 +330,7 @@ xfs_scrub_btree_block_check_siblings( > if (level == cur->bc_nlevels - 1) { > if (!xfs_btree_ptr_is_null(cur, &leftsib) || > !xfs_btree_ptr_is_null(cur, &rightsib)) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, level); > + xchk_btree_set_corrupt(bs->sc, cur, level); > goto out; > } > > @@ -339,10 +339,10 @@ xfs_scrub_btree_block_check_siblings( > * parent level pointers? > * (These function absorbs error codes for us.) > */ > - error = xfs_scrub_btree_block_check_sibling(bs, level, -1, &leftsib); > + error = xchk_btree_block_check_sibling(bs, level, -1, &leftsib); > if (error) > return error; > - error = xfs_scrub_btree_block_check_sibling(bs, level, 1, &rightsib); > + error = xchk_btree_block_check_sibling(bs, level, 1, &rightsib); > if (error) > return error; > out: > @@ -360,8 +360,8 @@ struct check_owner { > * an rmap record for it. > */ > STATIC int > -xfs_scrub_btree_check_block_owner( > - struct xfs_scrub_btree *bs, > +xchk_btree_check_block_owner( > + struct xchk_btree *bs, > int level, > xfs_daddr_t daddr) > { > @@ -380,13 +380,13 @@ xfs_scrub_btree_check_block_owner( > > init_sa = bs->cur->bc_flags & XFS_BTREE_LONG_PTRS; > if (init_sa) { > - error = xfs_scrub_ag_init(bs->sc, agno, &bs->sc->sa); > - if (!xfs_scrub_btree_xref_process_error(bs->sc, bs->cur, > + error = xchk_ag_init(bs->sc, agno, &bs->sc->sa); > + if (!xchk_btree_xref_process_error(bs->sc, bs->cur, > level, &error)) > return error; > } > > - xfs_scrub_xref_is_used_space(bs->sc, agbno, 1); > + xchk_xref_is_used_space(bs->sc, agbno, 1); > /* > * The bnobt scrubber aliases bs->cur to bs->sc->sa.bno_cur, so we > * have to nullify it (to shut down further block owner checks) if > @@ -395,20 +395,20 @@ xfs_scrub_btree_check_block_owner( > if (!bs->sc->sa.bno_cur && btnum == XFS_BTNUM_BNO) > bs->cur = NULL; > > - xfs_scrub_xref_is_owned_by(bs->sc, agbno, 1, bs->oinfo); > + xchk_xref_is_owned_by(bs->sc, agbno, 1, bs->oinfo); > if (!bs->sc->sa.rmap_cur && btnum == XFS_BTNUM_RMAP) > bs->cur = NULL; > > if (init_sa) > - xfs_scrub_ag_free(bs->sc, &bs->sc->sa); > + xchk_ag_free(bs->sc, &bs->sc->sa); > > return error; > } > > /* Check the owner of a btree block. */ > STATIC int > -xfs_scrub_btree_check_owner( > - struct xfs_scrub_btree *bs, > +xchk_btree_check_owner( > + struct xchk_btree *bs, > int level, > struct xfs_buf *bp) > { > @@ -437,7 +437,7 @@ xfs_scrub_btree_check_owner( > return 0; > } > > - return xfs_scrub_btree_check_block_owner(bs, level, XFS_BUF_ADDR(bp)); > + return xchk_btree_check_block_owner(bs, level, XFS_BUF_ADDR(bp)); > } > > /* > @@ -445,8 +445,8 @@ xfs_scrub_btree_check_owner( > * special blocks that don't require that. > */ > STATIC void > -xfs_scrub_btree_check_minrecs( > - struct xfs_scrub_btree *bs, > +xchk_btree_check_minrecs( > + struct xchk_btree *bs, > int level, > struct xfs_btree_block *block) > { > @@ -475,7 +475,7 @@ xfs_scrub_btree_check_minrecs( > if (level >= ok_level) > return; > > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, level); > + xchk_btree_set_corrupt(bs->sc, bs->cur, level); > } > > /* > @@ -483,8 +483,8 @@ xfs_scrub_btree_check_minrecs( > * and buffer pointers (if applicable) if they're ok to use. > */ > STATIC int > -xfs_scrub_btree_get_block( > - struct xfs_scrub_btree *bs, > +xchk_btree_get_block( > + struct xchk_btree *bs, > int level, > union xfs_btree_ptr *pp, > struct xfs_btree_block **pblock, > @@ -497,7 +497,7 @@ xfs_scrub_btree_get_block( > *pbp = NULL; > > error = xfs_btree_lookup_get_block(bs->cur, level, pp, pblock); > - if (!xfs_scrub_btree_process_error(bs->sc, bs->cur, level, &error) || > + if (!xchk_btree_process_error(bs->sc, bs->cur, level, &error) || > !*pblock) > return error; > > @@ -509,19 +509,19 @@ xfs_scrub_btree_get_block( > failed_at = __xfs_btree_check_sblock(bs->cur, *pblock, > level, *pbp); > if (failed_at) { > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, level); > + xchk_btree_set_corrupt(bs->sc, bs->cur, level); > return 0; > } > if (*pbp) > - xfs_scrub_buffer_recheck(bs->sc, *pbp); > + xchk_buffer_recheck(bs->sc, *pbp); > > - xfs_scrub_btree_check_minrecs(bs, level, *pblock); > + xchk_btree_check_minrecs(bs, level, *pblock); > > /* > * Check the block's owner; this function absorbs error codes > * for us. > */ > - error = xfs_scrub_btree_check_owner(bs, level, *pbp); > + error = xchk_btree_check_owner(bs, level, *pbp); > if (error) > return error; > > @@ -529,7 +529,7 @@ xfs_scrub_btree_get_block( > * Check the block's siblings; this function absorbs error codes > * for us. > */ > - return xfs_scrub_btree_block_check_siblings(bs, *pblock); > + return xchk_btree_block_check_siblings(bs, *pblock); > } > > /* > @@ -537,8 +537,8 @@ xfs_scrub_btree_get_block( > * in the parent block. > */ > STATIC void > -xfs_scrub_btree_block_keys( > - struct xfs_scrub_btree *bs, > +xchk_btree_block_keys( > + struct xchk_btree *bs, > int level, > struct xfs_btree_block *block) > { > @@ -562,7 +562,7 @@ xfs_scrub_btree_block_keys( > parent_block); > > if (cur->bc_ops->diff_two_keys(cur, &block_keys, parent_keys) != 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, 1); > + xchk_btree_set_corrupt(bs->sc, cur, 1); > > if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) > return; > @@ -573,7 +573,7 @@ xfs_scrub_btree_block_keys( > parent_block); > > if (cur->bc_ops->diff_two_keys(cur, high_bk, high_pk) != 0) > - xfs_scrub_btree_set_corrupt(bs->sc, cur, 1); > + xchk_btree_set_corrupt(bs->sc, cur, 1); > } > > /* > @@ -582,14 +582,14 @@ xfs_scrub_btree_block_keys( > * so that the caller can verify individual records. > */ > int > -xfs_scrub_btree( > +xchk_btree( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > - xfs_scrub_btree_rec_fn scrub_fn, > + xchk_btree_rec_fn scrub_fn, > struct xfs_owner_info *oinfo, > void *private) > { > - struct xfs_scrub_btree bs = { NULL }; > + struct xchk_btree bs = { NULL }; > union xfs_btree_ptr ptr; > union xfs_btree_ptr *pp; > union xfs_btree_rec *recp; > @@ -614,7 +614,7 @@ xfs_scrub_btree( > > /* Don't try to check a tree with a height we can't handle. */ > if (cur->bc_nlevels > XFS_BTREE_MAXLEVELS) { > - xfs_scrub_btree_set_corrupt(sc, cur, 0); > + xchk_btree_set_corrupt(sc, cur, 0); > goto out; > } > > @@ -624,9 +624,9 @@ xfs_scrub_btree( > */ > level = cur->bc_nlevels - 1; > cur->bc_ops->init_ptr_from_cur(cur, &ptr); > - if (!xfs_scrub_btree_ptr_ok(&bs, cur->bc_nlevels, &ptr)) > + if (!xchk_btree_ptr_ok(&bs, cur->bc_nlevels, &ptr)) > goto out; > - error = xfs_scrub_btree_get_block(&bs, level, &ptr, &block, &bp); > + error = xchk_btree_get_block(&bs, level, &ptr, &block, &bp); > if (error || !block) > goto out; > > @@ -639,7 +639,7 @@ xfs_scrub_btree( > /* End of leaf, pop back towards the root. */ > if (cur->bc_ptrs[level] > > be16_to_cpu(block->bb_numrecs)) { > - xfs_scrub_btree_block_keys(&bs, level, block); > + xchk_btree_block_keys(&bs, level, block); > if (level < cur->bc_nlevels - 1) > cur->bc_ptrs[level + 1]++; > level++; > @@ -647,14 +647,14 @@ xfs_scrub_btree( > } > > /* Records in order for scrub? */ > - xfs_scrub_btree_rec(&bs); > + xchk_btree_rec(&bs); > > /* Call out to the record checker. */ > recp = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block); > error = bs.scrub_rec(&bs, recp); > if (error) > break; > - if (xfs_scrub_should_terminate(sc, &error) || > + if (xchk_should_terminate(sc, &error) || > (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > break; > > @@ -664,7 +664,7 @@ xfs_scrub_btree( > > /* End of node, pop back towards the root. */ > if (cur->bc_ptrs[level] > be16_to_cpu(block->bb_numrecs)) { > - xfs_scrub_btree_block_keys(&bs, level, block); > + xchk_btree_block_keys(&bs, level, block); > if (level < cur->bc_nlevels - 1) > cur->bc_ptrs[level + 1]++; > level++; > @@ -672,16 +672,16 @@ xfs_scrub_btree( > } > > /* Keys in order for scrub? */ > - xfs_scrub_btree_key(&bs, level); > + xchk_btree_key(&bs, level); > > /* Drill another level deeper. */ > pp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[level], block); > - if (!xfs_scrub_btree_ptr_ok(&bs, level, pp)) { > + if (!xchk_btree_ptr_ok(&bs, level, pp)) { > cur->bc_ptrs[level]++; > continue; > } > level--; > - error = xfs_scrub_btree_get_block(&bs, level, pp, &block, &bp); > + error = xchk_btree_get_block(&bs, level, pp, &block, &bp); > if (error || !block) > goto out; > > @@ -692,7 +692,7 @@ xfs_scrub_btree( > /* Process deferred owner checks on btree blocks. */ > list_for_each_entry_safe(co, n, &bs.to_check, list) { > if (!error && bs.cur) > - error = xfs_scrub_btree_check_block_owner(&bs, > + error = xchk_btree_check_block_owner(&bs, > co->level, co->daddr); > list_del(&co->list); > kmem_free(co); > diff --git a/fs/xfs/scrub/btree.h b/fs/xfs/scrub/btree.h > index 956627500f2c..598ac04a6c3e 100644 > --- a/fs/xfs/scrub/btree.h > +++ b/fs/xfs/scrub/btree.h > @@ -9,32 +9,32 @@ > /* btree scrub */ > > /* Check for btree operation errors. */ > -bool xfs_scrub_btree_process_error(struct xfs_scrub_context *sc, > +bool xchk_btree_process_error(struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, int level, int *error); > > /* Check for btree xref operation errors. */ > -bool xfs_scrub_btree_xref_process_error(struct xfs_scrub_context *sc, > +bool xchk_btree_xref_process_error(struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, int level, > int *error); > > /* Check for btree corruption. */ > -void xfs_scrub_btree_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_btree_set_corrupt(struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, int level); > > /* Check for btree xref discrepancies. */ > -void xfs_scrub_btree_xref_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_btree_xref_set_corrupt(struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, int level); > > -struct xfs_scrub_btree; > -typedef int (*xfs_scrub_btree_rec_fn)( > - struct xfs_scrub_btree *bs, > +struct xchk_btree; > +typedef int (*xchk_btree_rec_fn)( > + struct xchk_btree *bs, > union xfs_btree_rec *rec); > > -struct xfs_scrub_btree { > +struct xchk_btree { > /* caller-provided scrub state */ > struct xfs_scrub_context *sc; > struct xfs_btree_cur *cur; > - xfs_scrub_btree_rec_fn scrub_rec; > + xchk_btree_rec_fn scrub_rec; > struct xfs_owner_info *oinfo; > void *private; > > @@ -45,8 +45,8 @@ struct xfs_scrub_btree { > bool firstkey[XFS_BTREE_MAXLEVELS]; > struct list_head to_check; > }; > -int xfs_scrub_btree(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > - xfs_scrub_btree_rec_fn scrub_fn, > +int xchk_btree(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > + xchk_btree_rec_fn scrub_fn, > struct xfs_owner_info *oinfo, void *private); > > #endif /* __XFS_SCRUB_BTREE_H__ */ > diff --git a/fs/xfs/scrub/common.c b/fs/xfs/scrub/common.c > index 70e70c69f83f..a4004a06f482 100644 > --- a/fs/xfs/scrub/common.c > +++ b/fs/xfs/scrub/common.c > @@ -68,7 +68,7 @@ > > /* Check for operational errors. */ > static bool > -__xfs_scrub_process_error( > +__xchk_process_error( > struct xfs_scrub_context *sc, > xfs_agnumber_t agno, > xfs_agblock_t bno, > @@ -81,7 +81,7 @@ __xfs_scrub_process_error( > return true; > case -EDEADLOCK: > /* Used to restart an op with deadlock avoidance. */ > - trace_xfs_scrub_deadlock_retry(sc->ip, sc->sm, *error); > + trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); > break; > case -EFSBADCRC: > case -EFSCORRUPTED: > @@ -90,7 +90,7 @@ __xfs_scrub_process_error( > *error = 0; > /* fall through */ > default: > - trace_xfs_scrub_op_error(sc, agno, bno, *error, > + trace_xchk_op_error(sc, agno, bno, *error, > ret_ip); > break; > } > @@ -98,30 +98,30 @@ __xfs_scrub_process_error( > } > > bool > -xfs_scrub_process_error( > +xchk_process_error( > struct xfs_scrub_context *sc, > xfs_agnumber_t agno, > xfs_agblock_t bno, > int *error) > { > - return __xfs_scrub_process_error(sc, agno, bno, error, > + return __xchk_process_error(sc, agno, bno, error, > XFS_SCRUB_OFLAG_CORRUPT, __return_address); > } > > bool > -xfs_scrub_xref_process_error( > +xchk_xref_process_error( > struct xfs_scrub_context *sc, > xfs_agnumber_t agno, > xfs_agblock_t bno, > int *error) > { > - return __xfs_scrub_process_error(sc, agno, bno, error, > + return __xchk_process_error(sc, agno, bno, error, > XFS_SCRUB_OFLAG_XFAIL, __return_address); > } > > /* Check for operational errors for a file offset. */ > static bool > -__xfs_scrub_fblock_process_error( > +__xchk_fblock_process_error( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset, > @@ -134,7 +134,7 @@ __xfs_scrub_fblock_process_error( > return true; > case -EDEADLOCK: > /* Used to restart an op with deadlock avoidance. */ > - trace_xfs_scrub_deadlock_retry(sc->ip, sc->sm, *error); > + trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); > break; > case -EFSBADCRC: > case -EFSCORRUPTED: > @@ -143,7 +143,7 @@ __xfs_scrub_fblock_process_error( > *error = 0; > /* fall through */ > default: > - trace_xfs_scrub_file_op_error(sc, whichfork, offset, *error, > + trace_xchk_file_op_error(sc, whichfork, offset, *error, > ret_ip); > break; > } > @@ -151,24 +151,24 @@ __xfs_scrub_fblock_process_error( > } > > bool > -xfs_scrub_fblock_process_error( > +xchk_fblock_process_error( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset, > int *error) > { > - return __xfs_scrub_fblock_process_error(sc, whichfork, offset, error, > + return __xchk_fblock_process_error(sc, whichfork, offset, error, > XFS_SCRUB_OFLAG_CORRUPT, __return_address); > } > > bool > -xfs_scrub_fblock_xref_process_error( > +xchk_fblock_xref_process_error( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset, > int *error) > { > - return __xfs_scrub_fblock_process_error(sc, whichfork, offset, error, > + return __xchk_fblock_process_error(sc, whichfork, offset, error, > XFS_SCRUB_OFLAG_XFAIL, __return_address); > } > > @@ -186,12 +186,12 @@ xfs_scrub_fblock_xref_process_error( > > /* Record a block which could be optimized. */ > void > -xfs_scrub_block_set_preen( > +xchk_block_set_preen( > struct xfs_scrub_context *sc, > struct xfs_buf *bp) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_PREEN; > - trace_xfs_scrub_block_preen(sc, bp->b_bn, __return_address); > + trace_xchk_block_preen(sc, bp->b_bn, __return_address); > } > > /* > @@ -200,32 +200,32 @@ xfs_scrub_block_set_preen( > * the block location of the inode record itself. > */ > void > -xfs_scrub_ino_set_preen( > +xchk_ino_set_preen( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_PREEN; > - trace_xfs_scrub_ino_preen(sc, ino, __return_address); > + trace_xchk_ino_preen(sc, ino, __return_address); > } > > /* Record a corrupt block. */ > void > -xfs_scrub_block_set_corrupt( > +xchk_block_set_corrupt( > struct xfs_scrub_context *sc, > struct xfs_buf *bp) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > - trace_xfs_scrub_block_error(sc, bp->b_bn, __return_address); > + trace_xchk_block_error(sc, bp->b_bn, __return_address); > } > > /* Record a corruption while cross-referencing. */ > void > -xfs_scrub_block_xref_set_corrupt( > +xchk_block_xref_set_corrupt( > struct xfs_scrub_context *sc, > struct xfs_buf *bp) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_XCORRUPT; > - trace_xfs_scrub_block_error(sc, bp->b_bn, __return_address); > + trace_xchk_block_error(sc, bp->b_bn, __return_address); > } > > /* > @@ -234,44 +234,44 @@ xfs_scrub_block_xref_set_corrupt( > * inode record itself. > */ > void > -xfs_scrub_ino_set_corrupt( > +xchk_ino_set_corrupt( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > - trace_xfs_scrub_ino_error(sc, ino, __return_address); > + trace_xchk_ino_error(sc, ino, __return_address); > } > > /* Record a corruption while cross-referencing with an inode. */ > void > -xfs_scrub_ino_xref_set_corrupt( > +xchk_ino_xref_set_corrupt( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_XCORRUPT; > - trace_xfs_scrub_ino_error(sc, ino, __return_address); > + trace_xchk_ino_error(sc, ino, __return_address); > } > > /* Record corruption in a block indexed by a file fork. */ > void > -xfs_scrub_fblock_set_corrupt( > +xchk_fblock_set_corrupt( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > - trace_xfs_scrub_fblock_error(sc, whichfork, offset, __return_address); > + trace_xchk_fblock_error(sc, whichfork, offset, __return_address); > } > > /* Record a corruption while cross-referencing a fork block. */ > void > -xfs_scrub_fblock_xref_set_corrupt( > +xchk_fblock_xref_set_corrupt( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_XCORRUPT; > - trace_xfs_scrub_fblock_error(sc, whichfork, offset, __return_address); > + trace_xchk_fblock_error(sc, whichfork, offset, __return_address); > } > > /* > @@ -279,32 +279,32 @@ xfs_scrub_fblock_xref_set_corrupt( > * incorrect. > */ > void > -xfs_scrub_ino_set_warning( > +xchk_ino_set_warning( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_WARNING; > - trace_xfs_scrub_ino_warning(sc, ino, __return_address); > + trace_xchk_ino_warning(sc, ino, __return_address); > } > > /* Warn about a block indexed by a file fork that needs review. */ > void > -xfs_scrub_fblock_set_warning( > +xchk_fblock_set_warning( > struct xfs_scrub_context *sc, > int whichfork, > xfs_fileoff_t offset) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_WARNING; > - trace_xfs_scrub_fblock_warning(sc, whichfork, offset, __return_address); > + trace_xchk_fblock_warning(sc, whichfork, offset, __return_address); > } > > /* Signal an incomplete scrub. */ > void > -xfs_scrub_set_incomplete( > +xchk_set_incomplete( > struct xfs_scrub_context *sc) > { > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_INCOMPLETE; > - trace_xfs_scrub_incomplete(sc, __return_address); > + trace_xchk_incomplete(sc, __return_address); > } > > /* > @@ -312,18 +312,18 @@ xfs_scrub_set_incomplete( > * at least according to the reverse mapping data. > */ > > -struct xfs_scrub_rmap_ownedby_info { > +struct xchk_rmap_ownedby_info { > struct xfs_owner_info *oinfo; > xfs_filblks_t *blocks; > }; > > STATIC int > -xfs_scrub_count_rmap_ownedby_irec( > +xchk_count_rmap_ownedby_irec( > struct xfs_btree_cur *cur, > struct xfs_rmap_irec *rec, > void *priv) > { > - struct xfs_scrub_rmap_ownedby_info *sroi = priv; > + struct xchk_rmap_ownedby_info *sroi = priv; > bool irec_attr; > bool oinfo_attr; > > @@ -344,19 +344,19 @@ xfs_scrub_count_rmap_ownedby_irec( > * The caller should pass us an rmapbt cursor. > */ > int > -xfs_scrub_count_rmap_ownedby_ag( > +xchk_count_rmap_ownedby_ag( > struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > struct xfs_owner_info *oinfo, > xfs_filblks_t *blocks) > { > - struct xfs_scrub_rmap_ownedby_info sroi; > + struct xchk_rmap_ownedby_info sroi; > > sroi.oinfo = oinfo; > *blocks = 0; > sroi.blocks = blocks; > > - return xfs_rmap_query_all(cur, xfs_scrub_count_rmap_ownedby_irec, > + return xfs_rmap_query_all(cur, xchk_count_rmap_ownedby_irec, > &sroi); > } > > @@ -392,12 +392,12 @@ want_ag_read_header_failure( > /* > * Grab all the headers for an AG. > * > - * The headers should be released by xfs_scrub_ag_free, but as a fail > + * The headers should be released by xchk_ag_free, but as a fail > * safe we attach all the buffers we grab to the scrub transaction so > * they'll all be freed when we cancel it. > */ > int > -xfs_scrub_ag_read_headers( > +xchk_ag_read_headers( > struct xfs_scrub_context *sc, > xfs_agnumber_t agno, > struct xfs_buf **agi, > @@ -425,8 +425,8 @@ xfs_scrub_ag_read_headers( > > /* Release all the AG btree cursors. */ > void > -xfs_scrub_ag_btcur_free( > - struct xfs_scrub_ag *sa) > +xchk_ag_btcur_free( > + struct xchk_ag *sa) > { > if (sa->refc_cur) > xfs_btree_del_cursor(sa->refc_cur, XFS_BTREE_ERROR); > @@ -451,9 +451,9 @@ xfs_scrub_ag_btcur_free( > > /* Initialize all the btree cursors for an AG. */ > int > -xfs_scrub_ag_btcur_init( > +xchk_ag_btcur_init( > struct xfs_scrub_context *sc, > - struct xfs_scrub_ag *sa) > + struct xchk_ag *sa) > { > struct xfs_mount *mp = sc->mp; > xfs_agnumber_t agno = sa->agno; > @@ -511,11 +511,11 @@ xfs_scrub_ag_btcur_init( > > /* Release the AG header context and btree cursors. */ > void > -xfs_scrub_ag_free( > +xchk_ag_free( > struct xfs_scrub_context *sc, > - struct xfs_scrub_ag *sa) > + struct xchk_ag *sa) > { > - xfs_scrub_ag_btcur_free(sa); > + xchk_ag_btcur_free(sa); > if (sa->agfl_bp) { > xfs_trans_brelse(sc->tp, sa->agfl_bp); > sa->agfl_bp = NULL; > @@ -543,30 +543,30 @@ xfs_scrub_ag_free( > * transaction ourselves. > */ > int > -xfs_scrub_ag_init( > +xchk_ag_init( > struct xfs_scrub_context *sc, > xfs_agnumber_t agno, > - struct xfs_scrub_ag *sa) > + struct xchk_ag *sa) > { > int error; > > sa->agno = agno; > - error = xfs_scrub_ag_read_headers(sc, agno, &sa->agi_bp, > + error = xchk_ag_read_headers(sc, agno, &sa->agi_bp, > &sa->agf_bp, &sa->agfl_bp); > if (error) > return error; > > - return xfs_scrub_ag_btcur_init(sc, sa); > + return xchk_ag_btcur_init(sc, sa); > } > > /* > * Grab the per-ag structure if we haven't already gotten it. Teardown of the > - * xfs_scrub_ag will release it for us. > + * xchk_ag will release it for us. > */ > void > -xfs_scrub_perag_get( > +xchk_perag_get( > struct xfs_mount *mp, > - struct xfs_scrub_ag *sa) > + struct xchk_ag *sa) > { > if (!sa->pag) > sa->pag = xfs_perag_get(mp, sa->agno); > @@ -585,7 +585,7 @@ xfs_scrub_perag_get( > * the metadata object. > */ > int > -xfs_scrub_trans_alloc( > +xchk_trans_alloc( > struct xfs_scrub_context *sc, > uint resblks) > { > @@ -598,19 +598,19 @@ xfs_scrub_trans_alloc( > > /* Set us up with a transaction and an empty context. */ > int > -xfs_scrub_setup_fs( > +xchk_setup_fs( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > uint resblks; > > resblks = xfs_repair_calc_ag_resblks(sc); > - return xfs_scrub_trans_alloc(sc, resblks); > + return xchk_trans_alloc(sc, resblks); > } > > /* Set us up with AG headers and btree cursors. */ > int > -xfs_scrub_setup_ag_btree( > +xchk_setup_ag_btree( > struct xfs_scrub_context *sc, > struct xfs_inode *ip, > bool force_log) > @@ -625,21 +625,21 @@ xfs_scrub_setup_ag_btree( > * document why they need to do so. > */ > if (force_log) { > - error = xfs_scrub_checkpoint_log(mp); > + error = xchk_checkpoint_log(mp); > if (error) > return error; > } > > - error = xfs_scrub_setup_fs(sc, ip); > + error = xchk_setup_fs(sc, ip); > if (error) > return error; > > - return xfs_scrub_ag_init(sc, sc->sm->sm_agno, &sc->sa); > + return xchk_ag_init(sc, sc->sm->sm_agno, &sc->sa); > } > > /* Push everything out of the log onto disk. */ > int > -xfs_scrub_checkpoint_log( > +xchk_checkpoint_log( > struct xfs_mount *mp) > { > int error; > @@ -657,7 +657,7 @@ xfs_scrub_checkpoint_log( > * The inode is not locked. > */ > int > -xfs_scrub_get_inode( > +xchk_get_inode( > struct xfs_scrub_context *sc, > struct xfs_inode *ip_in) > { > @@ -704,7 +704,7 @@ xfs_scrub_get_inode( > error = -EFSCORRUPTED; > /* fall through */ > default: > - trace_xfs_scrub_op_error(sc, > + trace_xchk_op_error(sc, > XFS_INO_TO_AGNO(mp, sc->sm->sm_ino), > XFS_INO_TO_AGBNO(mp, sc->sm->sm_ino), > error, __return_address); > @@ -721,21 +721,21 @@ xfs_scrub_get_inode( > > /* Set us up to scrub a file's contents. */ > int > -xfs_scrub_setup_inode_contents( > +xchk_setup_inode_contents( > struct xfs_scrub_context *sc, > struct xfs_inode *ip, > unsigned int resblks) > { > int error; > > - error = xfs_scrub_get_inode(sc, ip); > + error = xchk_get_inode(sc, ip); > if (error) > return error; > > /* Got the inode, lock it and we're ready to go. */ > sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; > xfs_ilock(sc->ip, sc->ilock_flags); > - error = xfs_scrub_trans_alloc(sc, resblks); > + error = xchk_trans_alloc(sc, resblks); > if (error) > goto out; > sc->ilock_flags |= XFS_ILOCK_EXCL; > @@ -752,13 +752,13 @@ xfs_scrub_setup_inode_contents( > * the cursor and skip the check. > */ > bool > -xfs_scrub_should_check_xref( > +xchk_should_check_xref( > struct xfs_scrub_context *sc, > int *error, > struct xfs_btree_cur **curpp) > { > /* No point in xref if we already know we're corrupt. */ > - if (xfs_scrub_skip_xref(sc->sm)) > + if (xchk_skip_xref(sc->sm)) > return false; > > if (*error == 0) > @@ -775,7 +775,7 @@ xfs_scrub_should_check_xref( > } > > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_XFAIL; > - trace_xfs_scrub_xref_error(sc, *error, __return_address); > + trace_xchk_xref_error(sc, *error, __return_address); > > /* > * Errors encountered during cross-referencing with another > @@ -787,25 +787,25 @@ xfs_scrub_should_check_xref( > > /* Run the structure verifiers on in-memory buffers to detect bad memory. */ > void > -xfs_scrub_buffer_recheck( > +xchk_buffer_recheck( > struct xfs_scrub_context *sc, > struct xfs_buf *bp) > { > xfs_failaddr_t fa; > > if (bp->b_ops == NULL) { > - xfs_scrub_block_set_corrupt(sc, bp); > + xchk_block_set_corrupt(sc, bp); > return; > } > if (bp->b_ops->verify_struct == NULL) { > - xfs_scrub_set_incomplete(sc); > + xchk_set_incomplete(sc); > return; > } > fa = bp->b_ops->verify_struct(bp); > if (!fa) > return; > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > - trace_xfs_scrub_block_error(sc, bp->b_bn, fa); > + trace_xchk_block_error(sc, bp->b_bn, fa); > } > > /* > @@ -813,7 +813,7 @@ xfs_scrub_buffer_recheck( > * pointed to by sc->ip and the ILOCK must be held. > */ > int > -xfs_scrub_metadata_inode_forks( > +xchk_metadata_inode_forks( > struct xfs_scrub_context *sc) > { > __u32 smtype; > @@ -825,26 +825,26 @@ xfs_scrub_metadata_inode_forks( > > /* Metadata inodes don't live on the rt device. */ > if (sc->ip->i_d.di_flags & XFS_DIFLAG_REALTIME) { > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > return 0; > } > > /* They should never participate in reflink. */ > if (xfs_is_reflink_inode(sc->ip)) { > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > return 0; > } > > /* They also should never have extended attributes. */ > if (xfs_inode_hasattr(sc->ip)) { > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > return 0; > } > > /* Invoke the data fork scrubber. */ > smtype = sc->sm->sm_type; > sc->sm->sm_type = XFS_SCRUB_TYPE_BMBTD; > - error = xfs_scrub_bmap_data(sc); > + error = xchk_bmap_data(sc); > sc->sm->sm_type = smtype; > if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > return error; > @@ -853,11 +853,11 @@ xfs_scrub_metadata_inode_forks( > if (xfs_sb_version_hasreflink(&sc->mp->m_sb)) { > error = xfs_reflink_inode_has_shared_extents(sc->tp, sc->ip, > &shared); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, > &error)) > return error; > if (shared) > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > } > > return error; > @@ -871,7 +871,7 @@ xfs_scrub_metadata_inode_forks( > * we can't. > */ > int > -xfs_scrub_ilock_inverted( > +xchk_ilock_inverted( > struct xfs_inode *ip, > uint lock_mode) > { > diff --git a/fs/xfs/scrub/common.h b/fs/xfs/scrub/common.h > index 2172bd5361e2..5881cb2ecc26 100644 > --- a/fs/xfs/scrub/common.h > +++ b/fs/xfs/scrub/common.h > @@ -12,7 +12,7 @@ > * Note that we're careful not to make any judgements about *error. > */ > static inline bool > -xfs_scrub_should_terminate( > +xchk_should_terminate( > struct xfs_scrub_context *sc, > int *error) > { > @@ -24,121 +24,121 @@ xfs_scrub_should_terminate( > return false; > } > > -int xfs_scrub_trans_alloc(struct xfs_scrub_context *sc, uint resblks); > -bool xfs_scrub_process_error(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > +int xchk_trans_alloc(struct xfs_scrub_context *sc, uint resblks); > +bool xchk_process_error(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > xfs_agblock_t bno, int *error); > -bool xfs_scrub_fblock_process_error(struct xfs_scrub_context *sc, int whichfork, > +bool xchk_fblock_process_error(struct xfs_scrub_context *sc, int whichfork, > xfs_fileoff_t offset, int *error); > > -bool xfs_scrub_xref_process_error(struct xfs_scrub_context *sc, > +bool xchk_xref_process_error(struct xfs_scrub_context *sc, > xfs_agnumber_t agno, xfs_agblock_t bno, int *error); > -bool xfs_scrub_fblock_xref_process_error(struct xfs_scrub_context *sc, > +bool xchk_fblock_xref_process_error(struct xfs_scrub_context *sc, > int whichfork, xfs_fileoff_t offset, int *error); > > -void xfs_scrub_block_set_preen(struct xfs_scrub_context *sc, > +void xchk_block_set_preen(struct xfs_scrub_context *sc, > struct xfs_buf *bp); > -void xfs_scrub_ino_set_preen(struct xfs_scrub_context *sc, xfs_ino_t ino); > +void xchk_ino_set_preen(struct xfs_scrub_context *sc, xfs_ino_t ino); > > -void xfs_scrub_block_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_block_set_corrupt(struct xfs_scrub_context *sc, > struct xfs_buf *bp); > -void xfs_scrub_ino_set_corrupt(struct xfs_scrub_context *sc, xfs_ino_t ino); > -void xfs_scrub_fblock_set_corrupt(struct xfs_scrub_context *sc, int whichfork, > +void xchk_ino_set_corrupt(struct xfs_scrub_context *sc, xfs_ino_t ino); > +void xchk_fblock_set_corrupt(struct xfs_scrub_context *sc, int whichfork, > xfs_fileoff_t offset); > > -void xfs_scrub_block_xref_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_block_xref_set_corrupt(struct xfs_scrub_context *sc, > struct xfs_buf *bp); > -void xfs_scrub_ino_xref_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_ino_xref_set_corrupt(struct xfs_scrub_context *sc, > xfs_ino_t ino); > -void xfs_scrub_fblock_xref_set_corrupt(struct xfs_scrub_context *sc, > +void xchk_fblock_xref_set_corrupt(struct xfs_scrub_context *sc, > int whichfork, xfs_fileoff_t offset); > > -void xfs_scrub_ino_set_warning(struct xfs_scrub_context *sc, xfs_ino_t ino); > -void xfs_scrub_fblock_set_warning(struct xfs_scrub_context *sc, int whichfork, > +void xchk_ino_set_warning(struct xfs_scrub_context *sc, xfs_ino_t ino); > +void xchk_fblock_set_warning(struct xfs_scrub_context *sc, int whichfork, > xfs_fileoff_t offset); > > -void xfs_scrub_set_incomplete(struct xfs_scrub_context *sc); > -int xfs_scrub_checkpoint_log(struct xfs_mount *mp); > +void xchk_set_incomplete(struct xfs_scrub_context *sc); > +int xchk_checkpoint_log(struct xfs_mount *mp); > > /* Are we set up for a cross-referencing check? */ > -bool xfs_scrub_should_check_xref(struct xfs_scrub_context *sc, int *error, > +bool xchk_should_check_xref(struct xfs_scrub_context *sc, int *error, > struct xfs_btree_cur **curpp); > > /* Setup functions */ > -int xfs_scrub_setup_fs(struct xfs_scrub_context *sc, struct xfs_inode *ip); > -int xfs_scrub_setup_ag_allocbt(struct xfs_scrub_context *sc, > +int xchk_setup_fs(struct xfs_scrub_context *sc, struct xfs_inode *ip); > +int xchk_setup_ag_allocbt(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_ag_iallocbt(struct xfs_scrub_context *sc, > +int xchk_setup_ag_iallocbt(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_ag_rmapbt(struct xfs_scrub_context *sc, > +int xchk_setup_ag_rmapbt(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_ag_refcountbt(struct xfs_scrub_context *sc, > +int xchk_setup_ag_refcountbt(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_inode(struct xfs_scrub_context *sc, > +int xchk_setup_inode(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_inode_bmap(struct xfs_scrub_context *sc, > +int xchk_setup_inode_bmap(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_inode_bmap_data(struct xfs_scrub_context *sc, > +int xchk_setup_inode_bmap_data(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_directory(struct xfs_scrub_context *sc, > +int xchk_setup_directory(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_xattr(struct xfs_scrub_context *sc, > +int xchk_setup_xattr(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_symlink(struct xfs_scrub_context *sc, > +int xchk_setup_symlink(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > -int xfs_scrub_setup_parent(struct xfs_scrub_context *sc, > +int xchk_setup_parent(struct xfs_scrub_context *sc, > struct xfs_inode *ip); > #ifdef CONFIG_XFS_RT > -int xfs_scrub_setup_rt(struct xfs_scrub_context *sc, struct xfs_inode *ip); > +int xchk_setup_rt(struct xfs_scrub_context *sc, struct xfs_inode *ip); > #else > static inline int > -xfs_scrub_setup_rt(struct xfs_scrub_context *sc, struct xfs_inode *ip) > +xchk_setup_rt(struct xfs_scrub_context *sc, struct xfs_inode *ip) > { > return -ENOENT; > } > #endif > #ifdef CONFIG_XFS_QUOTA > -int xfs_scrub_setup_quota(struct xfs_scrub_context *sc, struct xfs_inode *ip); > +int xchk_setup_quota(struct xfs_scrub_context *sc, struct xfs_inode *ip); > #else > static inline int > -xfs_scrub_setup_quota(struct xfs_scrub_context *sc, struct xfs_inode *ip) > +xchk_setup_quota(struct xfs_scrub_context *sc, struct xfs_inode *ip) > { > return -ENOENT; > } > #endif > > -void xfs_scrub_ag_free(struct xfs_scrub_context *sc, struct xfs_scrub_ag *sa); > -int xfs_scrub_ag_init(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > - struct xfs_scrub_ag *sa); > -void xfs_scrub_perag_get(struct xfs_mount *mp, struct xfs_scrub_ag *sa); > -int xfs_scrub_ag_read_headers(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > +void xchk_ag_free(struct xfs_scrub_context *sc, struct xchk_ag *sa); > +int xchk_ag_init(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > + struct xchk_ag *sa); > +void xchk_perag_get(struct xfs_mount *mp, struct xchk_ag *sa); > +int xchk_ag_read_headers(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > struct xfs_buf **agi, struct xfs_buf **agf, > struct xfs_buf **agfl); > -void xfs_scrub_ag_btcur_free(struct xfs_scrub_ag *sa); > -int xfs_scrub_ag_btcur_init(struct xfs_scrub_context *sc, > - struct xfs_scrub_ag *sa); > -int xfs_scrub_count_rmap_ownedby_ag(struct xfs_scrub_context *sc, > +void xchk_ag_btcur_free(struct xchk_ag *sa); > +int xchk_ag_btcur_init(struct xfs_scrub_context *sc, > + struct xchk_ag *sa); > +int xchk_count_rmap_ownedby_ag(struct xfs_scrub_context *sc, > struct xfs_btree_cur *cur, > struct xfs_owner_info *oinfo, > xfs_filblks_t *blocks); > > -int xfs_scrub_setup_ag_btree(struct xfs_scrub_context *sc, > +int xchk_setup_ag_btree(struct xfs_scrub_context *sc, > struct xfs_inode *ip, bool force_log); > -int xfs_scrub_get_inode(struct xfs_scrub_context *sc, struct xfs_inode *ip_in); > -int xfs_scrub_setup_inode_contents(struct xfs_scrub_context *sc, > +int xchk_get_inode(struct xfs_scrub_context *sc, struct xfs_inode *ip_in); > +int xchk_setup_inode_contents(struct xfs_scrub_context *sc, > struct xfs_inode *ip, unsigned int resblks); > -void xfs_scrub_buffer_recheck(struct xfs_scrub_context *sc, struct xfs_buf *bp); > +void xchk_buffer_recheck(struct xfs_scrub_context *sc, struct xfs_buf *bp); > > /* > * Don't bother cross-referencing if we already found corruption or cross > * referencing discrepancies. > */ > -static inline bool xfs_scrub_skip_xref(struct xfs_scrub_metadata *sm) > +static inline bool xchk_skip_xref(struct xfs_scrub_metadata *sm) > { > return sm->sm_flags & (XFS_SCRUB_OFLAG_CORRUPT | > XFS_SCRUB_OFLAG_XCORRUPT); > } > > -int xfs_scrub_metadata_inode_forks(struct xfs_scrub_context *sc); > -int xfs_scrub_ilock_inverted(struct xfs_inode *ip, uint lock_mode); > +int xchk_metadata_inode_forks(struct xfs_scrub_context *sc); > +int xchk_ilock_inverted(struct xfs_inode *ip, uint lock_mode); > > #endif /* __XFS_SCRUB_COMMON_H__ */ > diff --git a/fs/xfs/scrub/dabtree.c b/fs/xfs/scrub/dabtree.c > index d700c4d4d4ef..fee80f6ddfd7 100644 > --- a/fs/xfs/scrub/dabtree.c > +++ b/fs/xfs/scrub/dabtree.c > @@ -35,8 +35,8 @@ > * operational errors in common.c. > */ > bool > -xfs_scrub_da_process_error( > - struct xfs_scrub_da_btree *ds, > +xchk_da_process_error( > + struct xchk_da_btree *ds, > int level, > int *error) > { > @@ -48,7 +48,7 @@ xfs_scrub_da_process_error( > switch (*error) { > case -EDEADLOCK: > /* Used to restart an op with deadlock avoidance. */ > - trace_xfs_scrub_deadlock_retry(sc->ip, sc->sm, *error); > + trace_xchk_deadlock_retry(sc->ip, sc->sm, *error); > break; > case -EFSBADCRC: > case -EFSCORRUPTED: > @@ -57,7 +57,7 @@ xfs_scrub_da_process_error( > *error = 0; > /* fall through */ > default: > - trace_xfs_scrub_file_op_error(sc, ds->dargs.whichfork, > + trace_xchk_file_op_error(sc, ds->dargs.whichfork, > xfs_dir2_da_to_db(ds->dargs.geo, > ds->state->path.blk[level].blkno), > *error, __return_address); > @@ -71,15 +71,15 @@ xfs_scrub_da_process_error( > * operational errors in common.c. > */ > void > -xfs_scrub_da_set_corrupt( > - struct xfs_scrub_da_btree *ds, > +xchk_da_set_corrupt( > + struct xchk_da_btree *ds, > int level) > { > struct xfs_scrub_context *sc = ds->sc; > > sc->sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > > - trace_xfs_scrub_fblock_error(sc, ds->dargs.whichfork, > + trace_xchk_fblock_error(sc, ds->dargs.whichfork, > xfs_dir2_da_to_db(ds->dargs.geo, > ds->state->path.blk[level].blkno), > __return_address); > @@ -87,8 +87,8 @@ xfs_scrub_da_set_corrupt( > > /* Find an entry at a certain level in a da btree. */ > STATIC void * > -xfs_scrub_da_btree_entry( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_entry( > + struct xchk_da_btree *ds, > int level, > int rec) > { > @@ -123,8 +123,8 @@ xfs_scrub_da_btree_entry( > > /* Scrub a da btree hash (key). */ > int > -xfs_scrub_da_btree_hash( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_hash( > + struct xchk_da_btree *ds, > int level, > __be32 *hashp) > { > @@ -136,7 +136,7 @@ xfs_scrub_da_btree_hash( > /* Is this hash in order? */ > hash = be32_to_cpu(*hashp); > if (hash < ds->hashes[level]) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > ds->hashes[level] = hash; > > if (level == 0) > @@ -144,10 +144,10 @@ xfs_scrub_da_btree_hash( > > /* Is this hash no larger than the parent hash? */ > blks = ds->state->path.blk; > - entry = xfs_scrub_da_btree_entry(ds, level - 1, blks[level - 1].index); > + entry = xchk_da_btree_entry(ds, level - 1, blks[level - 1].index); > parent_hash = be32_to_cpu(entry->hashval); > if (parent_hash < hash) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > return 0; > } > @@ -157,13 +157,13 @@ xfs_scrub_da_btree_hash( > * pointer. > */ > STATIC bool > -xfs_scrub_da_btree_ptr_ok( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_ptr_ok( > + struct xchk_da_btree *ds, > int level, > xfs_dablk_t blkno) > { > if (blkno < ds->lowest || (ds->highest != 0 && blkno >= ds->highest)) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > return false; > } > > @@ -176,7 +176,7 @@ xfs_scrub_da_btree_ptr_ok( > * leaf1, we must multiplex the verifiers. > */ > static void > -xfs_scrub_da_btree_read_verify( > +xchk_da_btree_read_verify( > struct xfs_buf *bp) > { > struct xfs_da_blkinfo *info = bp->b_addr; > @@ -198,7 +198,7 @@ xfs_scrub_da_btree_read_verify( > } > } > static void > -xfs_scrub_da_btree_write_verify( > +xchk_da_btree_write_verify( > struct xfs_buf *bp) > { > struct xfs_da_blkinfo *info = bp->b_addr; > @@ -220,7 +220,7 @@ xfs_scrub_da_btree_write_verify( > } > } > static void * > -xfs_scrub_da_btree_verify( > +xchk_da_btree_verify( > struct xfs_buf *bp) > { > struct xfs_da_blkinfo *info = bp->b_addr; > @@ -236,17 +236,17 @@ xfs_scrub_da_btree_verify( > } > } > > -static const struct xfs_buf_ops xfs_scrub_da_btree_buf_ops = { > - .name = "xfs_scrub_da_btree", > - .verify_read = xfs_scrub_da_btree_read_verify, > - .verify_write = xfs_scrub_da_btree_write_verify, > - .verify_struct = xfs_scrub_da_btree_verify, > +static const struct xfs_buf_ops xchk_da_btree_buf_ops = { > + .name = "xchk_da_btree", > + .verify_read = xchk_da_btree_read_verify, > + .verify_write = xchk_da_btree_write_verify, > + .verify_struct = xchk_da_btree_verify, > }; > > /* Check a block's sibling. */ > STATIC int > -xfs_scrub_da_btree_block_check_sibling( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_block_check_sibling( > + struct xchk_da_btree *ds, > int level, > int direction, > xfs_dablk_t sibling) > @@ -265,7 +265,7 @@ xfs_scrub_da_btree_block_check_sibling( > error = xfs_da3_path_shift(ds->state, &ds->state->altpath, > direction, false, &retval); > if (error == 0 && retval == 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > error = 0; > goto out; > } > @@ -273,19 +273,19 @@ xfs_scrub_da_btree_block_check_sibling( > /* Move the alternate cursor one block in the direction given. */ > error = xfs_da3_path_shift(ds->state, &ds->state->altpath, > direction, false, &retval); > - if (!xfs_scrub_da_process_error(ds, level, &error)) > + if (!xchk_da_process_error(ds, level, &error)) > return error; > if (retval) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > return error; > } > if (ds->state->altpath.blk[level].bp) > - xfs_scrub_buffer_recheck(ds->sc, > + xchk_buffer_recheck(ds->sc, > ds->state->altpath.blk[level].bp); > > /* Compare upper level pointer to sibling pointer. */ > if (ds->state->altpath.blk[level].blkno != sibling) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > xfs_trans_brelse(ds->dargs.trans, ds->state->altpath.blk[level].bp); > out: > return error; > @@ -293,8 +293,8 @@ xfs_scrub_da_btree_block_check_sibling( > > /* Check a block's sibling pointers. */ > STATIC int > -xfs_scrub_da_btree_block_check_siblings( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_block_check_siblings( > + struct xchk_da_btree *ds, > int level, > struct xfs_da_blkinfo *hdr) > { > @@ -308,7 +308,7 @@ xfs_scrub_da_btree_block_check_siblings( > /* Top level blocks should not have sibling pointers. */ > if (level == 0) { > if (forw != 0 || back != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > return 0; > } > > @@ -316,10 +316,10 @@ xfs_scrub_da_btree_block_check_siblings( > * Check back (left) and forw (right) pointers. These functions > * absorb error codes for us. > */ > - error = xfs_scrub_da_btree_block_check_sibling(ds, level, 0, back); > + error = xchk_da_btree_block_check_sibling(ds, level, 0, back); > if (error) > goto out; > - error = xfs_scrub_da_btree_block_check_sibling(ds, level, 1, forw); > + error = xchk_da_btree_block_check_sibling(ds, level, 1, forw); > > out: > memset(&ds->state->altpath, 0, sizeof(ds->state->altpath)); > @@ -328,8 +328,8 @@ xfs_scrub_da_btree_block_check_siblings( > > /* Load a dir/attribute block from a btree. */ > STATIC int > -xfs_scrub_da_btree_block( > - struct xfs_scrub_da_btree *ds, > +xchk_da_btree_block( > + struct xchk_da_btree *ds, > int level, > xfs_dablk_t blkno) > { > @@ -355,17 +355,17 @@ xfs_scrub_da_btree_block( > > /* Check the pointer. */ > blk->blkno = blkno; > - if (!xfs_scrub_da_btree_ptr_ok(ds, level, blkno)) > + if (!xchk_da_btree_ptr_ok(ds, level, blkno)) > goto out_nobuf; > > /* Read the buffer. */ > error = xfs_da_read_buf(dargs->trans, dargs->dp, blk->blkno, -2, > &blk->bp, dargs->whichfork, > - &xfs_scrub_da_btree_buf_ops); > - if (!xfs_scrub_da_process_error(ds, level, &error)) > + &xchk_da_btree_buf_ops); > + if (!xchk_da_process_error(ds, level, &error)) > goto out_nobuf; > if (blk->bp) > - xfs_scrub_buffer_recheck(ds->sc, blk->bp); > + xchk_buffer_recheck(ds->sc, blk->bp); > > /* > * We didn't find a dir btree root block, which means that > @@ -378,7 +378,7 @@ xfs_scrub_da_btree_block( > > /* It's /not/ ok for attr trees not to have a da btree. */ > if (blk->bp == NULL) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out_nobuf; > } > > @@ -388,17 +388,17 @@ xfs_scrub_da_btree_block( > > /* We only started zeroing the header on v5 filesystems. */ > if (xfs_sb_version_hascrc(&ds->sc->mp->m_sb) && hdr3->hdr.pad) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > > /* Check the owner. */ > if (xfs_sb_version_hascrc(&ip->i_mount->m_sb)) { > owner = be64_to_cpu(hdr3->owner); > if (owner != ip->i_ino) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > } > > /* Check the siblings. */ > - error = xfs_scrub_da_btree_block_check_siblings(ds, level, &hdr3->hdr); > + error = xchk_da_btree_block_check_siblings(ds, level, &hdr3->hdr); > if (error) > goto out; > > @@ -411,7 +411,7 @@ xfs_scrub_da_btree_block( > blk->magic = XFS_ATTR_LEAF_MAGIC; > blk->hashval = xfs_attr_leaf_lasthash(blk->bp, pmaxrecs); > if (ds->tree_level != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > break; > case XFS_DIR2_LEAFN_MAGIC: > case XFS_DIR3_LEAFN_MAGIC: > @@ -420,7 +420,7 @@ xfs_scrub_da_btree_block( > blk->magic = XFS_DIR2_LEAFN_MAGIC; > blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs); > if (ds->tree_level != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > break; > case XFS_DIR2_LEAF1_MAGIC: > case XFS_DIR3_LEAF1_MAGIC: > @@ -429,7 +429,7 @@ xfs_scrub_da_btree_block( > blk->magic = XFS_DIR2_LEAF1_MAGIC; > blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs); > if (ds->tree_level != 0) > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > break; > case XFS_DA_NODE_MAGIC: > case XFS_DA3_NODE_MAGIC: > @@ -443,13 +443,13 @@ xfs_scrub_da_btree_block( > blk->hashval = be32_to_cpu(btree[*pmaxrecs - 1].hashval); > if (level == 0) { > if (nodehdr.level >= XFS_DA_NODE_MAXDEPTH) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out_freebp; > } > ds->tree_level = nodehdr.level; > } else { > if (ds->tree_level != nodehdr.level) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out_freebp; > } > } > @@ -457,7 +457,7 @@ xfs_scrub_da_btree_block( > /* XXX: Check hdr3.pad32 once we know how to fix it. */ > break; > default: > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out_freebp; > } > > @@ -473,13 +473,13 @@ xfs_scrub_da_btree_block( > > /* Visit all nodes and leaves of a da btree. */ > int > -xfs_scrub_da_btree( > +xchk_da_btree( > struct xfs_scrub_context *sc, > int whichfork, > - xfs_scrub_da_btree_rec_fn scrub_fn, > + xchk_da_btree_rec_fn scrub_fn, > void *private) > { > - struct xfs_scrub_da_btree ds = {}; > + struct xchk_da_btree ds = {}; > struct xfs_mount *mp = sc->mp; > struct xfs_da_state_blk *blks; > struct xfs_da_node_entry *key; > @@ -517,7 +517,7 @@ xfs_scrub_da_btree( > > /* Find the root of the da tree, if present. */ > blks = ds.state->path.blk; > - error = xfs_scrub_da_btree_block(&ds, level, blkno); > + error = xchk_da_btree_block(&ds, level, blkno); > if (error) > goto out_state; > /* > @@ -542,12 +542,12 @@ xfs_scrub_da_btree( > } > > /* Dispatch record scrubbing. */ > - rec = xfs_scrub_da_btree_entry(&ds, level, > + rec = xchk_da_btree_entry(&ds, level, > blks[level].index); > error = scrub_fn(&ds, level, rec); > if (error) > break; > - if (xfs_scrub_should_terminate(sc, &error) || > + if (xchk_should_terminate(sc, &error) || > (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > break; > > @@ -566,8 +566,8 @@ xfs_scrub_da_btree( > } > > /* Hashes in order for scrub? */ > - key = xfs_scrub_da_btree_entry(&ds, level, blks[level].index); > - error = xfs_scrub_da_btree_hash(&ds, level, &key->hashval); > + key = xchk_da_btree_entry(&ds, level, blks[level].index); > + error = xchk_da_btree_hash(&ds, level, &key->hashval); > if (error) > goto out; > > @@ -575,7 +575,7 @@ xfs_scrub_da_btree( > blkno = be32_to_cpu(key->before); > level++; > ds.tree_level--; > - error = xfs_scrub_da_btree_block(&ds, level, blkno); > + error = xchk_da_btree_block(&ds, level, blkno); > if (error) > goto out; > if (blks[level].bp == NULL) > diff --git a/fs/xfs/scrub/dabtree.h b/fs/xfs/scrub/dabtree.h > index 365f9f0019e6..80e4af0e2589 100644 > --- a/fs/xfs/scrub/dabtree.h > +++ b/fs/xfs/scrub/dabtree.h > @@ -8,7 +8,7 @@ > > /* dir/attr btree */ > > -struct xfs_scrub_da_btree { > +struct xchk_da_btree { > struct xfs_da_args dargs; > xfs_dahash_t hashes[XFS_DA_NODE_MAXDEPTH]; > int maxrecs[XFS_DA_NODE_MAXDEPTH]; > @@ -28,18 +28,18 @@ struct xfs_scrub_da_btree { > int tree_level; > }; > > -typedef int (*xfs_scrub_da_btree_rec_fn)(struct xfs_scrub_da_btree *ds, > +typedef int (*xchk_da_btree_rec_fn)(struct xchk_da_btree *ds, > int level, void *rec); > > /* Check for da btree operation errors. */ > -bool xfs_scrub_da_process_error(struct xfs_scrub_da_btree *ds, int level, int *error); > +bool xchk_da_process_error(struct xchk_da_btree *ds, int level, int *error); > > /* Check for da btree corruption. */ > -void xfs_scrub_da_set_corrupt(struct xfs_scrub_da_btree *ds, int level); > +void xchk_da_set_corrupt(struct xchk_da_btree *ds, int level); > > -int xfs_scrub_da_btree_hash(struct xfs_scrub_da_btree *ds, int level, > +int xchk_da_btree_hash(struct xchk_da_btree *ds, int level, > __be32 *hashp); > -int xfs_scrub_da_btree(struct xfs_scrub_context *sc, int whichfork, > - xfs_scrub_da_btree_rec_fn scrub_fn, void *private); > +int xchk_da_btree(struct xfs_scrub_context *sc, int whichfork, > + xchk_da_btree_rec_fn scrub_fn, void *private); > > #endif /* __XFS_SCRUB_DABTREE_H__ */ > diff --git a/fs/xfs/scrub/dir.c b/fs/xfs/scrub/dir.c > index 86324775fc9b..2ac07bb73478 100644 > --- a/fs/xfs/scrub/dir.c > +++ b/fs/xfs/scrub/dir.c > @@ -31,18 +31,18 @@ > > /* Set us up to scrub directories. */ > int > -xfs_scrub_setup_directory( > +xchk_setup_directory( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_inode_contents(sc, ip, 0); > + return xchk_setup_inode_contents(sc, ip, 0); > } > > /* Directories */ > > /* Scrub a directory entry. */ > > -struct xfs_scrub_dir_ctx { > +struct xchk_dir_ctx { > /* VFS fill-directory iterator */ > struct dir_context dir_iter; > > @@ -51,8 +51,8 @@ struct xfs_scrub_dir_ctx { > > /* Check that an inode's mode matches a given DT_ type. */ > STATIC int > -xfs_scrub_dir_check_ftype( > - struct xfs_scrub_dir_ctx *sdc, > +xchk_dir_check_ftype( > + struct xchk_dir_ctx *sdc, > xfs_fileoff_t offset, > xfs_ino_t inum, > int dtype) > @@ -64,7 +64,7 @@ xfs_scrub_dir_check_ftype( > > if (!xfs_sb_version_hasftype(&mp->m_sb)) { > if (dtype != DT_UNKNOWN && dtype != DT_DIR) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > offset); > goto out; > } > @@ -78,7 +78,7 @@ xfs_scrub_dir_check_ftype( > * inodes can trigger immediate inactive cleanup of the inode. > */ > error = xfs_iget(mp, sdc->sc->tp, inum, 0, 0, &ip); > - if (!xfs_scrub_fblock_xref_process_error(sdc->sc, XFS_DATA_FORK, offset, > + if (!xchk_fblock_xref_process_error(sdc->sc, XFS_DATA_FORK, offset, > &error)) > goto out; > > @@ -86,7 +86,7 @@ xfs_scrub_dir_check_ftype( > ino_dtype = xfs_dir3_get_dtype(mp, > xfs_mode_to_ftype(VFS_I(ip)->i_mode)); > if (ino_dtype != dtype) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > iput(VFS_I(ip)); > out: > return error; > @@ -101,7 +101,7 @@ xfs_scrub_dir_check_ftype( > * we can look up this filename. Finally, we check the ftype. > */ > STATIC int > -xfs_scrub_dir_actor( > +xchk_dir_actor( > struct dir_context *dir_iter, > const char *name, > int namelen, > @@ -111,13 +111,13 @@ xfs_scrub_dir_actor( > { > struct xfs_mount *mp; > struct xfs_inode *ip; > - struct xfs_scrub_dir_ctx *sdc; > + struct xchk_dir_ctx *sdc; > struct xfs_name xname; > xfs_ino_t lookup_ino; > xfs_dablk_t offset; > int error = 0; > > - sdc = container_of(dir_iter, struct xfs_scrub_dir_ctx, dir_iter); > + sdc = container_of(dir_iter, struct xchk_dir_ctx, dir_iter); > ip = sdc->sc->ip; > mp = ip->i_mount; > offset = xfs_dir2_db_to_da(mp->m_dir_geo, > @@ -125,17 +125,17 @@ xfs_scrub_dir_actor( > > /* Does this inode number make sense? */ > if (!xfs_verify_dir_ino(mp, ino)) { > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > goto out; > } > > if (!strncmp(".", name, namelen)) { > /* If this is "." then check that the inum matches the dir. */ > if (xfs_sb_version_hasftype(&mp->m_sb) && type != DT_DIR) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > offset); > if (ino != ip->i_ino) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > offset); > } else if (!strncmp("..", name, namelen)) { > /* > @@ -143,10 +143,10 @@ xfs_scrub_dir_actor( > * matches this dir. > */ > if (xfs_sb_version_hasftype(&mp->m_sb) && type != DT_DIR) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > offset); > if (ip->i_ino == mp->m_sb.sb_rootino && ino != ip->i_ino) > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, > offset); > } > > @@ -156,23 +156,23 @@ xfs_scrub_dir_actor( > xname.type = XFS_DIR3_FT_UNKNOWN; > > error = xfs_dir_lookup(sdc->sc->tp, ip, &xname, &lookup_ino, NULL); > - if (!xfs_scrub_fblock_process_error(sdc->sc, XFS_DATA_FORK, offset, > + if (!xchk_fblock_process_error(sdc->sc, XFS_DATA_FORK, offset, > &error)) > goto out; > if (lookup_ino != ino) { > - xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset); > goto out; > } > > /* Verify the file type. This function absorbs error codes. */ > - error = xfs_scrub_dir_check_ftype(sdc, offset, lookup_ino, type); > + error = xchk_dir_check_ftype(sdc, offset, lookup_ino, type); > if (error) > goto out; > out: > /* > * A negative error code returned here is supposed to cause the > * dir_emit caller (xfs_readdir) to abort the directory iteration > - * and return zero to xfs_scrub_directory. > + * and return zero to xchk_directory. > */ > if (error == 0 && sdc->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return -EFSCORRUPTED; > @@ -181,8 +181,8 @@ xfs_scrub_dir_actor( > > /* Scrub a directory btree record. */ > STATIC int > -xfs_scrub_dir_rec( > - struct xfs_scrub_da_btree *ds, > +xchk_dir_rec( > + struct xchk_da_btree *ds, > int level, > void *rec) > { > @@ -203,7 +203,7 @@ xfs_scrub_dir_rec( > int error; > > /* Check the hash of the entry. */ > - error = xfs_scrub_da_btree_hash(ds, level, &ent->hashval); > + error = xchk_da_btree_hash(ds, level, &ent->hashval); > if (error) > goto out; > > @@ -218,18 +218,18 @@ xfs_scrub_dir_rec( > rec_bno = xfs_dir2_db_to_da(mp->m_dir_geo, db); > > if (rec_bno >= mp->m_dir_geo->leafblk) { > - xfs_scrub_da_set_corrupt(ds, level); > + xchk_da_set_corrupt(ds, level); > goto out; > } > error = xfs_dir3_data_read(ds->dargs.trans, dp, rec_bno, -2, &bp); > - if (!xfs_scrub_fblock_process_error(ds->sc, XFS_DATA_FORK, rec_bno, > + if (!xchk_fblock_process_error(ds->sc, XFS_DATA_FORK, rec_bno, > &error)) > goto out; > if (!bp) { > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > goto out; > } > - xfs_scrub_buffer_recheck(ds->sc, bp); > + xchk_buffer_recheck(ds->sc, bp); > > if (ds->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out_relse; > @@ -240,7 +240,7 @@ xfs_scrub_dir_rec( > p = (char *)mp->m_dir_inode_ops->data_entry_p(bp->b_addr); > endp = xfs_dir3_data_endp(mp->m_dir_geo, bp->b_addr); > if (!endp) { > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > goto out_relse; > } > while (p < endp) { > @@ -258,7 +258,7 @@ xfs_scrub_dir_rec( > p += mp->m_dir_inode_ops->data_entsize(dep->namelen); > } > if (p >= endp) { > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > goto out_relse; > } > > @@ -267,14 +267,14 @@ xfs_scrub_dir_rec( > hash = be32_to_cpu(ent->hashval); > tag = be16_to_cpup(dp->d_ops->data_entry_tag_p(dent)); > if (!xfs_verify_dir_ino(mp, ino) || tag != off) > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > if (dent->namelen == 0) { > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > goto out_relse; > } > calc_hash = xfs_da_hashname(dent->name, dent->namelen); > if (calc_hash != hash) > - xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > + xchk_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno); > > out_relse: > xfs_trans_brelse(ds->dargs.trans, bp); > @@ -288,7 +288,7 @@ xfs_scrub_dir_rec( > * shortest, and that there aren't any bogus entries. > */ > STATIC void > -xfs_scrub_directory_check_free_entry( > +xchk_directory_check_free_entry( > struct xfs_scrub_context *sc, > xfs_dablk_t lblk, > struct xfs_dir2_data_free *bf, > @@ -308,12 +308,12 @@ xfs_scrub_directory_check_free_entry( > return; > > /* Unused entry should be in the bestfrees but wasn't found. */ > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > } > > /* Check free space info in a directory data block. */ > STATIC int > -xfs_scrub_directory_data_bestfree( > +xchk_directory_data_bestfree( > struct xfs_scrub_context *sc, > xfs_dablk_t lblk, > bool is_block) > @@ -339,15 +339,15 @@ xfs_scrub_directory_data_bestfree( > if (is_block) { > /* dir block format */ > if (lblk != XFS_B_TO_FSBT(mp, XFS_DIR2_DATA_OFFSET)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > error = xfs_dir3_block_read(sc->tp, sc->ip, &bp); > } else { > /* dir data format */ > error = xfs_dir3_data_read(sc->tp, sc->ip, lblk, -1, &bp); > } > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > goto out; > - xfs_scrub_buffer_recheck(sc, bp); > + xchk_buffer_recheck(sc, bp); > > /* XXX: Check xfs_dir3_data_hdr.pad is zero once we start setting it. */ > > @@ -362,7 +362,7 @@ xfs_scrub_directory_data_bestfree( > if (offset == 0) > continue; > if (offset >= mp->m_dir_geo->blksize) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out_buf; > } > dup = (struct xfs_dir2_data_unused *)(bp->b_addr + offset); > @@ -372,13 +372,13 @@ xfs_scrub_directory_data_bestfree( > if (dup->freetag != cpu_to_be16(XFS_DIR2_DATA_FREE_TAG) || > be16_to_cpu(dup->length) != be16_to_cpu(dfp->length) || > tag != ((char *)dup - (char *)bp->b_addr)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out_buf; > } > > /* bestfree records should be ordered largest to smallest */ > if (smallest_bestfree < be16_to_cpu(dfp->length)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out_buf; > } > > @@ -400,7 +400,7 @@ xfs_scrub_directory_data_bestfree( > dep = (struct xfs_dir2_data_entry *)ptr; > newlen = d_ops->data_entsize(dep->namelen); > if (newlen <= 0) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, > lblk); > goto out_buf; > } > @@ -411,7 +411,7 @@ xfs_scrub_directory_data_bestfree( > /* Spot check this free entry */ > tag = be16_to_cpu(*xfs_dir2_data_unused_tag_p(dup)); > if (tag != ((char *)dup - (char *)bp->b_addr)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out_buf; > } > > @@ -419,14 +419,14 @@ xfs_scrub_directory_data_bestfree( > * Either this entry is a bestfree or it's smaller than > * any of the bestfrees. > */ > - xfs_scrub_directory_check_free_entry(sc, lblk, bf, dup); > + xchk_directory_check_free_entry(sc, lblk, bf, dup); > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out_buf; > > /* Move on. */ > newlen = be16_to_cpu(dup->length); > if (newlen <= 0) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out_buf; > } > ptr += newlen; > @@ -436,11 +436,11 @@ xfs_scrub_directory_data_bestfree( > > /* We're required to fill all the space. */ > if (ptr != endptr) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > > /* Did we see at least as many free slots as there are bestfrees? */ > if (nr_frees < nr_bestfrees) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > out_buf: > xfs_trans_brelse(sc->tp, bp); > out: > @@ -454,7 +454,7 @@ xfs_scrub_directory_data_bestfree( > * array is in order. > */ > STATIC void > -xfs_scrub_directory_check_freesp( > +xchk_directory_check_freesp( > struct xfs_scrub_context *sc, > xfs_dablk_t lblk, > struct xfs_buf *dbp, > @@ -465,15 +465,15 @@ xfs_scrub_directory_check_freesp( > dfp = sc->ip->d_ops->data_bestfree_p(dbp->b_addr); > > if (len != be16_to_cpu(dfp->length)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > > if (len > 0 && be16_to_cpu(dfp->offset) == 0) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > } > > /* Check free space info in a directory leaf1 block. */ > STATIC int > -xfs_scrub_directory_leaf1_bestfree( > +xchk_directory_leaf1_bestfree( > struct xfs_scrub_context *sc, > struct xfs_da_args *args, > xfs_dablk_t lblk) > @@ -497,9 +497,9 @@ xfs_scrub_directory_leaf1_bestfree( > > /* Read the free space block. */ > error = xfs_dir3_leaf_read(sc->tp, sc->ip, lblk, -1, &bp); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > goto out; > - xfs_scrub_buffer_recheck(sc, bp); > + xchk_buffer_recheck(sc, bp); > > leaf = bp->b_addr; > d_ops->leaf_hdr_from_disk(&leafhdr, leaf); > @@ -512,7 +512,7 @@ xfs_scrub_directory_leaf1_bestfree( > struct xfs_dir3_leaf_hdr *hdr3 = bp->b_addr; > > if (hdr3->pad != cpu_to_be32(0)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > } > > /* > @@ -520,19 +520,19 @@ xfs_scrub_directory_leaf1_bestfree( > * blocks that can fit under i_size. > */ > if (bestcount != xfs_dir2_byte_to_db(geo, sc->ip->i_d.di_size)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > > /* Is the leaf count even remotely sane? */ > if (leafhdr.count > d_ops->leaf_max_ents(geo)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > > /* Leaves and bests don't overlap in leaf format. */ > if ((char *)&ents[leafhdr.count] > (char *)bestp) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > > @@ -540,13 +540,13 @@ xfs_scrub_directory_leaf1_bestfree( > for (i = 0; i < leafhdr.count; i++) { > hash = be32_to_cpu(ents[i].hashval); > if (i > 0 && lasthash > hash) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > lasthash = hash; > if (ents[i].address == cpu_to_be32(XFS_DIR2_NULL_DATAPTR)) > stale++; > } > if (leafhdr.stale != stale) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > > @@ -557,10 +557,10 @@ xfs_scrub_directory_leaf1_bestfree( > continue; > error = xfs_dir3_data_read(sc->tp, sc->ip, > i * args->geo->fsbcount, -1, &dbp); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, > &error)) > break; > - xfs_scrub_directory_check_freesp(sc, lblk, dbp, best); > + xchk_directory_check_freesp(sc, lblk, dbp, best); > xfs_trans_brelse(sc->tp, dbp); > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > @@ -571,7 +571,7 @@ xfs_scrub_directory_leaf1_bestfree( > > /* Check free space info in a directory freespace block. */ > STATIC int > -xfs_scrub_directory_free_bestfree( > +xchk_directory_free_bestfree( > struct xfs_scrub_context *sc, > struct xfs_da_args *args, > xfs_dablk_t lblk) > @@ -587,15 +587,15 @@ xfs_scrub_directory_free_bestfree( > > /* Read the free space block */ > error = xfs_dir2_free_read(sc->tp, sc->ip, lblk, &bp); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > goto out; > - xfs_scrub_buffer_recheck(sc, bp); > + xchk_buffer_recheck(sc, bp); > > if (xfs_sb_version_hascrc(&sc->mp->m_sb)) { > struct xfs_dir3_free_hdr *hdr3 = bp->b_addr; > > if (hdr3->pad != cpu_to_be32(0)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > } > > /* Check all the entries. */ > @@ -610,22 +610,22 @@ xfs_scrub_directory_free_bestfree( > error = xfs_dir3_data_read(sc->tp, sc->ip, > (freehdr.firstdb + i) * args->geo->fsbcount, > -1, &dbp); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, > &error)) > break; > - xfs_scrub_directory_check_freesp(sc, lblk, dbp, best); > + xchk_directory_check_freesp(sc, lblk, dbp, best); > xfs_trans_brelse(sc->tp, dbp); > } > > if (freehdr.nused + stale != freehdr.nvalid) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > out: > return error; > } > > /* Check free space information in directories. */ > STATIC int > -xfs_scrub_directory_blocks( > +xchk_directory_blocks( > struct xfs_scrub_context *sc) > { > struct xfs_bmbt_irec got; > @@ -656,7 +656,7 @@ xfs_scrub_directory_blocks( > args.geo = mp->m_dir_geo; > args.trans = sc->tp; > error = xfs_dir2_isblock(&args, &is_block); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) > goto out; > > /* Iterate all the data extents in the directory... */ > @@ -666,7 +666,7 @@ xfs_scrub_directory_blocks( > if (is_block && > (got.br_startoff > 0 || > got.br_blockcount != args.geo->fsbcount)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, > got.br_startoff); > break; > } > @@ -690,7 +690,7 @@ xfs_scrub_directory_blocks( > args.geo->fsbcount); > lblk < got.br_startoff + got.br_blockcount; > lblk += args.geo->fsbcount) { > - error = xfs_scrub_directory_data_bestfree(sc, lblk, > + error = xchk_directory_data_bestfree(sc, lblk, > is_block); > if (error) > goto out; > @@ -709,10 +709,10 @@ xfs_scrub_directory_blocks( > got.br_blockcount == args.geo->fsbcount && > !xfs_iext_next_extent(ifp, &icur, &got)) { > if (is_block) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > - error = xfs_scrub_directory_leaf1_bestfree(sc, &args, > + error = xchk_directory_leaf1_bestfree(sc, &args, > leaf_lblk); > if (error) > goto out; > @@ -731,11 +731,11 @@ xfs_scrub_directory_blocks( > */ > lblk = got.br_startoff; > if (lblk & ~0xFFFFFFFFULL) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > if (is_block) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); > goto out; > } > > @@ -754,7 +754,7 @@ xfs_scrub_directory_blocks( > args.geo->fsbcount); > lblk < got.br_startoff + got.br_blockcount; > lblk += args.geo->fsbcount) { > - error = xfs_scrub_directory_free_bestfree(sc, &args, > + error = xchk_directory_free_bestfree(sc, &args, > lblk); > if (error) > goto out; > @@ -769,11 +769,11 @@ xfs_scrub_directory_blocks( > > /* Scrub a whole directory. */ > int > -xfs_scrub_directory( > +xchk_directory( > struct xfs_scrub_context *sc) > { > - struct xfs_scrub_dir_ctx sdc = { > - .dir_iter.actor = xfs_scrub_dir_actor, > + struct xchk_dir_ctx sdc = { > + .dir_iter.actor = xchk_dir_actor, > .dir_iter.pos = 0, > .sc = sc, > }; > @@ -786,12 +786,12 @@ xfs_scrub_directory( > > /* Plausible size? */ > if (sc->ip->i_d.di_size < xfs_dir2_sf_hdr_size(0)) { > - xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_set_corrupt(sc, sc->ip->i_ino); > goto out; > } > > /* Check directory tree structure */ > - error = xfs_scrub_da_btree(sc, XFS_DATA_FORK, xfs_scrub_dir_rec, NULL); > + error = xchk_da_btree(sc, XFS_DATA_FORK, xchk_dir_rec, NULL); > if (error) > return error; > > @@ -799,7 +799,7 @@ xfs_scrub_directory( > return error; > > /* Check the freespace. */ > - error = xfs_scrub_directory_blocks(sc); > + error = xchk_directory_blocks(sc); > if (error) > return error; > > @@ -816,7 +816,7 @@ xfs_scrub_directory( > /* > * Look up every name in this directory by hash. > * > - * Use the xfs_readdir function to call xfs_scrub_dir_actor on > + * Use the xfs_readdir function to call xchk_dir_actor on > * every directory entry in this directory. In _actor, we check > * the name, inode number, and ftype (if applicable) of the > * entry. xfs_readdir uses the VFS filldir functions to provide > @@ -834,7 +834,7 @@ xfs_scrub_directory( > xfs_iunlock(sc->ip, XFS_ILOCK_EXCL); > while (true) { > error = xfs_readdir(sc->tp, sc->ip, &sdc.dir_iter, bufsize); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, > &error)) > goto out; > if (oldpos == sdc.dir_iter.pos) > diff --git a/fs/xfs/scrub/ialloc.c b/fs/xfs/scrub/ialloc.c > index 13d43d108574..69d652b7299c 100644 > --- a/fs/xfs/scrub/ialloc.c > +++ b/fs/xfs/scrub/ialloc.c > @@ -35,11 +35,11 @@ > * try again after forcing logged inode cores out to disk. > */ > int > -xfs_scrub_setup_ag_iallocbt( > +xchk_setup_ag_iallocbt( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_ag_btree(sc, ip, sc->try_harder); > + return xchk_setup_ag_btree(sc, ip, sc->try_harder); > } > > /* Inode btree scrubber. */ > @@ -50,7 +50,7 @@ xfs_scrub_setup_ag_iallocbt( > * we have a record or not depending on freecount. > */ > static inline void > -xfs_scrub_iallocbt_chunk_xref_other( > +xchk_iallocbt_chunk_xref_other( > struct xfs_scrub_context *sc, > struct xfs_inobt_rec_incore *irec, > xfs_agino_t agino) > @@ -66,16 +66,16 @@ xfs_scrub_iallocbt_chunk_xref_other( > if (!(*pcur)) > return; > error = xfs_ialloc_has_inode_record(*pcur, agino, agino, &has_irec); > - if (!xfs_scrub_should_check_xref(sc, &error, pcur)) > + if (!xchk_should_check_xref(sc, &error, pcur)) > return; > if (((irec->ir_freecount > 0 && !has_irec) || > (irec->ir_freecount == 0 && has_irec))) > - xfs_scrub_btree_xref_set_corrupt(sc, *pcur, 0); > + xchk_btree_xref_set_corrupt(sc, *pcur, 0); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_iallocbt_chunk_xref( > +xchk_iallocbt_chunk_xref( > struct xfs_scrub_context *sc, > struct xfs_inobt_rec_incore *irec, > xfs_agino_t agino, > @@ -87,17 +87,17 @@ xfs_scrub_iallocbt_chunk_xref( > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, len); > - xfs_scrub_iallocbt_chunk_xref_other(sc, irec, agino); > + xchk_xref_is_used_space(sc, agbno, len); > + xchk_iallocbt_chunk_xref_other(sc, irec, agino); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INODES); > - xfs_scrub_xref_is_owned_by(sc, agbno, len, &oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, len); > + xchk_xref_is_owned_by(sc, agbno, len, &oinfo); > + xchk_xref_is_not_shared(sc, agbno, len); > } > > /* Is this chunk worth checking? */ > STATIC bool > -xfs_scrub_iallocbt_chunk( > - struct xfs_scrub_btree *bs, > +xchk_iallocbt_chunk( > + struct xchk_btree *bs, > struct xfs_inobt_rec_incore *irec, > xfs_agino_t agino, > xfs_extlen_t len) > @@ -110,16 +110,16 @@ xfs_scrub_iallocbt_chunk( > if (bno + len <= bno || > !xfs_verify_agbno(mp, agno, bno) || > !xfs_verify_agbno(mp, agno, bno + len - 1)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > - xfs_scrub_iallocbt_chunk_xref(bs->sc, irec, agino, bno, len); > + xchk_iallocbt_chunk_xref(bs->sc, irec, agino, bno, len); > > return true; > } > > /* Count the number of free inodes. */ > static unsigned int > -xfs_scrub_iallocbt_freecount( > +xchk_iallocbt_freecount( > xfs_inofree_t freemask) > { > BUILD_BUG_ON(sizeof(freemask) != sizeof(__u64)); > @@ -128,8 +128,8 @@ xfs_scrub_iallocbt_freecount( > > /* Check a particular inode with ir_free. */ > STATIC int > -xfs_scrub_iallocbt_check_cluster_freemask( > - struct xfs_scrub_btree *bs, > +xchk_iallocbt_check_cluster_freemask( > + struct xchk_btree *bs, > xfs_ino_t fsino, > xfs_agino_t chunkino, > xfs_agino_t clusterino, > @@ -143,14 +143,14 @@ xfs_scrub_iallocbt_check_cluster_freemask( > bool inuse; > int error = 0; > > - if (xfs_scrub_should_terminate(bs->sc, &error)) > + if (xchk_should_terminate(bs->sc, &error)) > return error; > > dip = xfs_buf_offset(bp, clusterino * mp->m_sb.sb_inodesize); > if (be16_to_cpu(dip->di_magic) != XFS_DINODE_MAGIC || > (dip->di_version >= 3 && > be64_to_cpu(dip->di_ino) != fsino + clusterino)) { > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > goto out; > } > > @@ -175,15 +175,15 @@ xfs_scrub_iallocbt_check_cluster_freemask( > freemask_ok = inode_is_free ^ inuse; > } > if (!freemask_ok) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > out: > return 0; > } > > /* Make sure the free mask is consistent with what the inodes think. */ > STATIC int > -xfs_scrub_iallocbt_check_freemask( > - struct xfs_scrub_btree *bs, > +xchk_iallocbt_check_freemask( > + struct xchk_btree *bs, > struct xfs_inobt_rec_incore *irec) > { > struct xfs_owner_info oinfo; > @@ -223,18 +223,18 @@ xfs_scrub_iallocbt_check_freemask( > /* The whole cluster must be a hole or not a hole. */ > ir_holemask = (irec->ir_holemask & holemask); > if (ir_holemask != holemask && ir_holemask != 0) { > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > continue; > } > > /* If any part of this is a hole, skip it. */ > if (ir_holemask) { > - xfs_scrub_xref_is_not_owned_by(bs->sc, agbno, > + xchk_xref_is_not_owned_by(bs->sc, agbno, > blks_per_cluster, &oinfo); > continue; > } > > - xfs_scrub_xref_is_owned_by(bs->sc, agbno, blks_per_cluster, > + xchk_xref_is_owned_by(bs->sc, agbno, blks_per_cluster, > &oinfo); > > /* Grab the inode cluster buffer. */ > @@ -245,13 +245,13 @@ xfs_scrub_iallocbt_check_freemask( > > error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap, > &dip, &bp, 0, 0); > - if (!xfs_scrub_btree_xref_process_error(bs->sc, bs->cur, 0, > + if (!xchk_btree_xref_process_error(bs->sc, bs->cur, 0, > &error)) > continue; > > /* Which inodes are free? */ > for (clusterino = 0; clusterino < nr_inodes; clusterino++) { > - error = xfs_scrub_iallocbt_check_cluster_freemask(bs, > + error = xchk_iallocbt_check_cluster_freemask(bs, > fsino, chunkino, clusterino, irec, bp); > if (error) { > xfs_trans_brelse(bs->cur->bc_tp, bp); > @@ -267,8 +267,8 @@ xfs_scrub_iallocbt_check_freemask( > > /* Scrub an inobt/finobt record. */ > STATIC int > -xfs_scrub_iallocbt_rec( > - struct xfs_scrub_btree *bs, > +xchk_iallocbt_rec( > + struct xchk_btree *bs, > union xfs_btree_rec *rec) > { > struct xfs_mount *mp = bs->cur->bc_mp; > @@ -289,18 +289,18 @@ xfs_scrub_iallocbt_rec( > > if (irec.ir_count > XFS_INODES_PER_CHUNK || > irec.ir_freecount > XFS_INODES_PER_CHUNK) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > real_freecount = irec.ir_freecount + > (XFS_INODES_PER_CHUNK - irec.ir_count); > - if (real_freecount != xfs_scrub_iallocbt_freecount(irec.ir_free)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + if (real_freecount != xchk_iallocbt_freecount(irec.ir_free)) > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > agino = irec.ir_startino; > /* Record has to be properly aligned within the AG. */ > if (!xfs_verify_agino(mp, agno, agino) || > !xfs_verify_agino(mp, agno, agino + XFS_INODES_PER_CHUNK - 1)) { > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > goto out; > } > > @@ -308,7 +308,7 @@ xfs_scrub_iallocbt_rec( > agbno = XFS_AGINO_TO_AGBNO(mp, irec.ir_startino); > if ((agbno & (xfs_ialloc_cluster_alignment(mp) - 1)) || > (agbno & (xfs_icluster_size_fsb(mp) - 1))) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > *inode_blocks += XFS_B_TO_FSB(mp, > irec.ir_count * mp->m_sb.sb_inodesize); > @@ -318,9 +318,9 @@ xfs_scrub_iallocbt_rec( > len = XFS_B_TO_FSB(mp, > XFS_INODES_PER_CHUNK * mp->m_sb.sb_inodesize); > if (irec.ir_count != XFS_INODES_PER_CHUNK) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > - if (!xfs_scrub_iallocbt_chunk(bs, &irec, agino, len)) > + if (!xchk_iallocbt_chunk(bs, &irec, agino, len)) > goto out; > goto check_freemask; > } > @@ -333,12 +333,12 @@ xfs_scrub_iallocbt_rec( > holes = ~xfs_inobt_irec_to_allocmask(&irec); > if ((holes & irec.ir_free) != holes || > irec.ir_freecount > irec.ir_count) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > for (i = 0; i < XFS_INOBT_HOLEMASK_BITS; i++) { > if (holemask & 1) > holecount += XFS_INODES_PER_HOLEMASK_BIT; > - else if (!xfs_scrub_iallocbt_chunk(bs, &irec, agino, len)) > + else if (!xchk_iallocbt_chunk(bs, &irec, agino, len)) > break; > holemask >>= 1; > agino += XFS_INODES_PER_HOLEMASK_BIT; > @@ -346,10 +346,10 @@ xfs_scrub_iallocbt_rec( > > if (holecount > XFS_INODES_PER_CHUNK || > holecount + irec.ir_count != XFS_INODES_PER_CHUNK) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > check_freemask: > - error = xfs_scrub_iallocbt_check_freemask(bs, &irec); > + error = xchk_iallocbt_check_freemask(bs, &irec); > if (error) > goto out; > > @@ -362,7 +362,7 @@ xfs_scrub_iallocbt_rec( > * Don't bother if we're missing btree cursors, as we're already corrupt. > */ > STATIC void > -xfs_scrub_iallocbt_xref_rmap_btreeblks( > +xchk_iallocbt_xref_rmap_btreeblks( > struct xfs_scrub_context *sc, > int which) > { > @@ -374,27 +374,27 @@ xfs_scrub_iallocbt_xref_rmap_btreeblks( > > if (!sc->sa.ino_cur || !sc->sa.rmap_cur || > (xfs_sb_version_hasfinobt(&sc->mp->m_sb) && !sc->sa.fino_cur) || > - xfs_scrub_skip_xref(sc->sm)) > + xchk_skip_xref(sc->sm)) > return; > > /* Check that we saw as many inobt blocks as the rmap says. */ > error = xfs_btree_count_blocks(sc->sa.ino_cur, &inobt_blocks); > - if (!xfs_scrub_process_error(sc, 0, 0, &error)) > + if (!xchk_process_error(sc, 0, 0, &error)) > return; > > if (sc->sa.fino_cur) { > error = xfs_btree_count_blocks(sc->sa.fino_cur, &finobt_blocks); > - if (!xfs_scrub_process_error(sc, 0, 0, &error)) > + if (!xchk_process_error(sc, 0, 0, &error)) > return; > } > > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INOBT); > - error = xfs_scrub_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, &oinfo, > + error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, &oinfo, > &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (blocks != inobt_blocks + finobt_blocks) > - xfs_scrub_btree_set_corrupt(sc, sc->sa.ino_cur, 0); > + xchk_btree_set_corrupt(sc, sc->sa.ino_cur, 0); > } > > /* > @@ -402,7 +402,7 @@ xfs_scrub_iallocbt_xref_rmap_btreeblks( > * the rmap says are owned by inodes. > */ > STATIC void > -xfs_scrub_iallocbt_xref_rmap_inodes( > +xchk_iallocbt_xref_rmap_inodes( > struct xfs_scrub_context *sc, > int which, > xfs_filblks_t inode_blocks) > @@ -411,22 +411,22 @@ xfs_scrub_iallocbt_xref_rmap_inodes( > xfs_filblks_t blocks; > int error; > > - if (!sc->sa.rmap_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) > return; > > /* Check that we saw as many inode blocks as the rmap knows about. */ > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INODES); > - error = xfs_scrub_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, &oinfo, > + error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, &oinfo, > &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (blocks != inode_blocks) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > } > > /* Scrub the inode btrees for some AG. */ > STATIC int > -xfs_scrub_iallocbt( > +xchk_iallocbt( > struct xfs_scrub_context *sc, > xfs_btnum_t which) > { > @@ -437,12 +437,12 @@ xfs_scrub_iallocbt( > > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INOBT); > cur = which == XFS_BTNUM_INO ? sc->sa.ino_cur : sc->sa.fino_cur; > - error = xfs_scrub_btree(sc, cur, xfs_scrub_iallocbt_rec, &oinfo, > + error = xchk_btree(sc, cur, xchk_iallocbt_rec, &oinfo, > &inode_blocks); > if (error) > return error; > > - xfs_scrub_iallocbt_xref_rmap_btreeblks(sc, which); > + xchk_iallocbt_xref_rmap_btreeblks(sc, which); > > /* > * If we're scrubbing the inode btree, inode_blocks is the number of > @@ -452,28 +452,28 @@ xfs_scrub_iallocbt( > * to inode chunks with free inodes. > */ > if (which == XFS_BTNUM_INO) > - xfs_scrub_iallocbt_xref_rmap_inodes(sc, which, inode_blocks); > + xchk_iallocbt_xref_rmap_inodes(sc, which, inode_blocks); > > return error; > } > > int > -xfs_scrub_inobt( > +xchk_inobt( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_iallocbt(sc, XFS_BTNUM_INO); > + return xchk_iallocbt(sc, XFS_BTNUM_INO); > } > > int > -xfs_scrub_finobt( > +xchk_finobt( > struct xfs_scrub_context *sc) > { > - return xfs_scrub_iallocbt(sc, XFS_BTNUM_FINO); > + return xchk_iallocbt(sc, XFS_BTNUM_FINO); > } > > /* See if an inode btree has (or doesn't have) an inode chunk record. */ > static inline void > -xfs_scrub_xref_inode_check( > +xchk_xref_inode_check( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len, > @@ -483,33 +483,33 @@ xfs_scrub_xref_inode_check( > bool has_inodes; > int error; > > - if (!(*icur) || xfs_scrub_skip_xref(sc->sm)) > + if (!(*icur) || xchk_skip_xref(sc->sm)) > return; > > error = xfs_ialloc_has_inodes_at_extent(*icur, agbno, len, &has_inodes); > - if (!xfs_scrub_should_check_xref(sc, &error, icur)) > + if (!xchk_should_check_xref(sc, &error, icur)) > return; > if (has_inodes != should_have_inodes) > - xfs_scrub_btree_xref_set_corrupt(sc, *icur, 0); > + xchk_btree_xref_set_corrupt(sc, *icur, 0); > } > > /* xref check that the extent is not covered by inodes */ > void > -xfs_scrub_xref_is_not_inode_chunk( > +xchk_xref_is_not_inode_chunk( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > { > - xfs_scrub_xref_inode_check(sc, agbno, len, &sc->sa.ino_cur, false); > - xfs_scrub_xref_inode_check(sc, agbno, len, &sc->sa.fino_cur, false); > + xchk_xref_inode_check(sc, agbno, len, &sc->sa.ino_cur, false); > + xchk_xref_inode_check(sc, agbno, len, &sc->sa.fino_cur, false); > } > > /* xref check that the extent is covered by inodes */ > void > -xfs_scrub_xref_is_inode_chunk( > +xchk_xref_is_inode_chunk( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > { > - xfs_scrub_xref_inode_check(sc, agbno, len, &sc->sa.ino_cur, true); > + xchk_xref_inode_check(sc, agbno, len, &sc->sa.ino_cur, true); > } > diff --git a/fs/xfs/scrub/inode.c b/fs/xfs/scrub/inode.c > index 7a6208505980..d85fbec39e52 100644 > --- a/fs/xfs/scrub/inode.c > +++ b/fs/xfs/scrub/inode.c > @@ -37,7 +37,7 @@ > * the goal. > */ > int > -xfs_scrub_setup_inode( > +xchk_setup_inode( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > @@ -47,13 +47,13 @@ xfs_scrub_setup_inode( > * Try to get the inode. If the verifiers fail, we try again > * in raw mode. > */ > - error = xfs_scrub_get_inode(sc, ip); > + error = xchk_get_inode(sc, ip); > switch (error) { > case 0: > break; > case -EFSCORRUPTED: > case -EFSBADCRC: > - return xfs_scrub_trans_alloc(sc, 0); > + return xchk_trans_alloc(sc, 0); > default: > return error; > } > @@ -61,7 +61,7 @@ xfs_scrub_setup_inode( > /* Got the inode, lock it and we're ready to go. */ > sc->ilock_flags = XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL; > xfs_ilock(sc->ip, sc->ilock_flags); > - error = xfs_scrub_trans_alloc(sc, 0); > + error = xchk_trans_alloc(sc, 0); > if (error) > goto out; > sc->ilock_flags |= XFS_ILOCK_EXCL; > @@ -76,7 +76,7 @@ xfs_scrub_setup_inode( > > /* Validate di_extsize hint. */ > STATIC void > -xfs_scrub_inode_extsize( > +xchk_inode_extsize( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip, > xfs_ino_t ino, > @@ -88,7 +88,7 @@ xfs_scrub_inode_extsize( > fa = xfs_inode_validate_extsize(sc->mp, be32_to_cpu(dip->di_extsize), > mode, flags); > if (fa) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > /* > @@ -98,7 +98,7 @@ xfs_scrub_inode_extsize( > * These functions must be kept in sync with each other. > */ > STATIC void > -xfs_scrub_inode_cowextsize( > +xchk_inode_cowextsize( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip, > xfs_ino_t ino, > @@ -112,12 +112,12 @@ xfs_scrub_inode_cowextsize( > be32_to_cpu(dip->di_cowextsize), mode, flags, > flags2); > if (fa) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > /* Make sure the di_flags make sense for the inode. */ > STATIC void > -xfs_scrub_inode_flags( > +xchk_inode_flags( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip, > xfs_ino_t ino, > @@ -157,12 +157,12 @@ xfs_scrub_inode_flags( > > return; > bad: > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > /* Make sure the di_flags2 make sense for the inode. */ > STATIC void > -xfs_scrub_inode_flags2( > +xchk_inode_flags2( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip, > xfs_ino_t ino, > @@ -200,12 +200,12 @@ xfs_scrub_inode_flags2( > > return; > bad: > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > /* Scrub all the ondisk inode fields. */ > STATIC void > -xfs_scrub_dinode( > +xchk_dinode( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip, > xfs_ino_t ino) > @@ -237,7 +237,7 @@ xfs_scrub_dinode( > /* mode is recognized */ > break; > default: > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > } > > @@ -248,22 +248,22 @@ xfs_scrub_dinode( > * We autoconvert v1 inodes into v2 inodes on writeout, > * so just mark this inode for preening. > */ > - xfs_scrub_ino_set_preen(sc, ino); > + xchk_ino_set_preen(sc, ino); > break; > case 2: > case 3: > if (dip->di_onlink != 0) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > if (dip->di_mode == 0 && sc->ip) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > if (dip->di_projid_hi != 0 && > !xfs_sb_version_hasprojid32bit(&mp->m_sb)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > default: > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > return; > } > > @@ -273,40 +273,40 @@ xfs_scrub_dinode( > */ > if (dip->di_uid == cpu_to_be32(-1U) || > dip->di_gid == cpu_to_be32(-1U)) > - xfs_scrub_ino_set_warning(sc, ino); > + xchk_ino_set_warning(sc, ino); > > /* di_format */ > switch (dip->di_format) { > case XFS_DINODE_FMT_DEV: > if (!S_ISCHR(mode) && !S_ISBLK(mode) && > !S_ISFIFO(mode) && !S_ISSOCK(mode)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_LOCAL: > if (!S_ISDIR(mode) && !S_ISLNK(mode)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_EXTENTS: > if (!S_ISREG(mode) && !S_ISDIR(mode) && !S_ISLNK(mode)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_BTREE: > if (!S_ISREG(mode) && !S_ISDIR(mode)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_UUID: > default: > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > } > > /* di_[amc]time.nsec */ > if (be32_to_cpu(dip->di_atime.t_nsec) >= NSEC_PER_SEC) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > if (be32_to_cpu(dip->di_mtime.t_nsec) >= NSEC_PER_SEC) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > if (be32_to_cpu(dip->di_ctime.t_nsec) >= NSEC_PER_SEC) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* > * di_size. xfs_dinode_verify checks for things that screw up > @@ -315,19 +315,19 @@ xfs_scrub_dinode( > */ > isize = be64_to_cpu(dip->di_size); > if (isize & (1ULL << 63)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* Devices, fifos, and sockets must have zero size */ > if (!S_ISDIR(mode) && !S_ISREG(mode) && !S_ISLNK(mode) && isize != 0) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* Directories can't be larger than the data section size (32G) */ > if (S_ISDIR(mode) && (isize == 0 || isize >= XFS_DIR2_SPACE_SIZE)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* Symlinks can't be larger than SYMLINK_MAXLEN */ > if (S_ISLNK(mode) && (isize == 0 || isize >= XFS_SYMLINK_MAXLEN)) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* > * Warn if the running kernel can't handle the kinds of offsets > @@ -336,7 +336,7 @@ xfs_scrub_dinode( > * overly large offsets, flag the inode for admin review. > */ > if (isize >= mp->m_super->s_maxbytes) > - xfs_scrub_ino_set_warning(sc, ino); > + xchk_ino_set_warning(sc, ino); > > /* di_nblocks */ > if (flags2 & XFS_DIFLAG2_REFLINK) { > @@ -351,15 +351,15 @@ xfs_scrub_dinode( > */ > if (be64_to_cpu(dip->di_nblocks) >= > mp->m_sb.sb_dblocks + mp->m_sb.sb_rblocks) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } else { > if (be64_to_cpu(dip->di_nblocks) >= mp->m_sb.sb_dblocks) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > - xfs_scrub_inode_flags(sc, dip, ino, mode, flags); > + xchk_inode_flags(sc, dip, ino, mode, flags); > > - xfs_scrub_inode_extsize(sc, dip, ino, mode, flags); > + xchk_inode_extsize(sc, dip, ino, mode, flags); > > /* di_nextents */ > nextents = be32_to_cpu(dip->di_nextents); > @@ -367,31 +367,31 @@ xfs_scrub_dinode( > switch (dip->di_format) { > case XFS_DINODE_FMT_EXTENTS: > if (nextents > fork_recs) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_BTREE: > if (nextents <= fork_recs) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > default: > if (nextents != 0) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > } > > /* di_forkoff */ > if (XFS_DFORK_APTR(dip) >= (char *)dip + mp->m_sb.sb_inodesize) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > if (dip->di_anextents != 0 && dip->di_forkoff == 0) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > if (dip->di_forkoff == 0 && dip->di_aformat != XFS_DINODE_FMT_EXTENTS) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* di_aformat */ > if (dip->di_aformat != XFS_DINODE_FMT_LOCAL && > dip->di_aformat != XFS_DINODE_FMT_EXTENTS && > dip->di_aformat != XFS_DINODE_FMT_BTREE) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > > /* di_anextents */ > nextents = be16_to_cpu(dip->di_anextents); > @@ -399,22 +399,22 @@ xfs_scrub_dinode( > switch (dip->di_aformat) { > case XFS_DINODE_FMT_EXTENTS: > if (nextents > fork_recs) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > case XFS_DINODE_FMT_BTREE: > if (nextents <= fork_recs) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > break; > default: > if (nextents != 0) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > if (dip->di_version >= 3) { > if (be32_to_cpu(dip->di_crtime.t_nsec) >= NSEC_PER_SEC) > - xfs_scrub_ino_set_corrupt(sc, ino); > - xfs_scrub_inode_flags2(sc, dip, ino, mode, flags, flags2); > - xfs_scrub_inode_cowextsize(sc, dip, ino, mode, flags, > + xchk_ino_set_corrupt(sc, ino); > + xchk_inode_flags2(sc, dip, ino, mode, flags, flags2); > + xchk_inode_cowextsize(sc, dip, ino, mode, flags, > flags2); > } > } > @@ -425,7 +425,7 @@ xfs_scrub_dinode( > * IGET_UNTRUSTED, which checks the inobt for us. > */ > static void > -xfs_scrub_inode_xref_finobt( > +xchk_inode_xref_finobt( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > @@ -434,7 +434,7 @@ xfs_scrub_inode_xref_finobt( > int has_record; > int error; > > - if (!sc->sa.fino_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.fino_cur || xchk_skip_xref(sc->sm)) > return; > > agino = XFS_INO_TO_AGINO(sc->mp, ino); > @@ -445,12 +445,12 @@ xfs_scrub_inode_xref_finobt( > */ > error = xfs_inobt_lookup(sc->sa.fino_cur, agino, XFS_LOOKUP_LE, > &has_record); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.fino_cur) || > + if (!xchk_should_check_xref(sc, &error, &sc->sa.fino_cur) || > !has_record) > return; > > error = xfs_inobt_get_rec(sc->sa.fino_cur, &rec, &has_record); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.fino_cur) || > + if (!xchk_should_check_xref(sc, &error, &sc->sa.fino_cur) || > !has_record) > return; > > @@ -463,12 +463,12 @@ xfs_scrub_inode_xref_finobt( > return; > > if (rec.ir_free & XFS_INOBT_MASK(agino - rec.ir_startino)) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.fino_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.fino_cur, 0); > } > > /* Cross reference the inode fields with the forks. */ > STATIC void > -xfs_scrub_inode_xref_bmap( > +xchk_inode_xref_bmap( > struct xfs_scrub_context *sc, > struct xfs_dinode *dip) > { > @@ -477,32 +477,32 @@ xfs_scrub_inode_xref_bmap( > xfs_filblks_t acount; > int error; > > - if (xfs_scrub_skip_xref(sc->sm)) > + if (xchk_skip_xref(sc->sm)) > return; > > /* Walk all the extents to check nextents/naextents/nblocks. */ > error = xfs_bmap_count_blocks(sc->tp, sc->ip, XFS_DATA_FORK, > &nextents, &count); > - if (!xfs_scrub_should_check_xref(sc, &error, NULL)) > + if (!xchk_should_check_xref(sc, &error, NULL)) > return; > if (nextents < be32_to_cpu(dip->di_nextents)) > - xfs_scrub_ino_xref_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino); > > error = xfs_bmap_count_blocks(sc->tp, sc->ip, XFS_ATTR_FORK, > &nextents, &acount); > - if (!xfs_scrub_should_check_xref(sc, &error, NULL)) > + if (!xchk_should_check_xref(sc, &error, NULL)) > return; > if (nextents != be16_to_cpu(dip->di_anextents)) > - xfs_scrub_ino_xref_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino); > > /* Check nblocks against the inode. */ > if (count + acount != be64_to_cpu(dip->di_nblocks)) > - xfs_scrub_ino_xref_set_corrupt(sc, sc->ip->i_ino); > + xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_inode_xref( > +xchk_inode_xref( > struct xfs_scrub_context *sc, > xfs_ino_t ino, > struct xfs_dinode *dip) > @@ -518,18 +518,18 @@ xfs_scrub_inode_xref( > agno = XFS_INO_TO_AGNO(sc->mp, ino); > agbno = XFS_INO_TO_AGBNO(sc->mp, ino); > > - error = xfs_scrub_ag_init(sc, agno, &sc->sa); > - if (!xfs_scrub_xref_process_error(sc, agno, agbno, &error)) > + error = xchk_ag_init(sc, agno, &sc->sa); > + if (!xchk_xref_process_error(sc, agno, agbno, &error)) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, 1); > - xfs_scrub_inode_xref_finobt(sc, ino); > + xchk_xref_is_used_space(sc, agbno, 1); > + xchk_inode_xref_finobt(sc, ino); > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INODES); > - xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo); > - xfs_scrub_xref_is_not_shared(sc, agbno, 1); > - xfs_scrub_inode_xref_bmap(sc, dip); > + xchk_xref_is_owned_by(sc, agbno, 1, &oinfo); > + xchk_xref_is_not_shared(sc, agbno, 1); > + xchk_inode_xref_bmap(sc, dip); > > - xfs_scrub_ag_free(sc, &sc->sa); > + xchk_ag_free(sc, &sc->sa); > } > > /* > @@ -539,7 +539,7 @@ xfs_scrub_inode_xref( > * reflink filesystem. > */ > static void > -xfs_scrub_inode_check_reflink_iflag( > +xchk_inode_check_reflink_iflag( > struct xfs_scrub_context *sc, > xfs_ino_t ino) > { > @@ -552,18 +552,18 @@ xfs_scrub_inode_check_reflink_iflag( > > error = xfs_reflink_inode_has_shared_extents(sc->tp, sc->ip, > &has_shared); > - if (!xfs_scrub_xref_process_error(sc, XFS_INO_TO_AGNO(mp, ino), > + if (!xchk_xref_process_error(sc, XFS_INO_TO_AGNO(mp, ino), > XFS_INO_TO_AGBNO(mp, ino), &error)) > return; > if (xfs_is_reflink_inode(sc->ip) && !has_shared) > - xfs_scrub_ino_set_preen(sc, ino); > + xchk_ino_set_preen(sc, ino); > else if (!xfs_is_reflink_inode(sc->ip) && has_shared) > - xfs_scrub_ino_set_corrupt(sc, ino); > + xchk_ino_set_corrupt(sc, ino); > } > > /* Scrub an inode. */ > int > -xfs_scrub_inode( > +xchk_inode( > struct xfs_scrub_context *sc) > { > struct xfs_dinode di; > @@ -575,13 +575,13 @@ xfs_scrub_inode( > * and a NULL inode, so flag the corruption error and return. > */ > if (!sc->ip) { > - xfs_scrub_ino_set_corrupt(sc, sc->sm->sm_ino); > + xchk_ino_set_corrupt(sc, sc->sm->sm_ino); > return 0; > } > > /* Scrub the inode core. */ > xfs_inode_to_disk(sc->ip, &di, 0); > - xfs_scrub_dinode(sc, &di, sc->ip->i_ino); > + xchk_dinode(sc, &di, sc->ip->i_ino); > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > goto out; > > @@ -591,9 +591,9 @@ xfs_scrub_inode( > * we scrubbed the dinode. > */ > if (S_ISREG(VFS_I(sc->ip)->i_mode)) > - xfs_scrub_inode_check_reflink_iflag(sc, sc->ip->i_ino); > + xchk_inode_check_reflink_iflag(sc, sc->ip->i_ino); > > - xfs_scrub_inode_xref(sc, sc->ip->i_ino, &di); > + xchk_inode_xref(sc, sc->ip->i_ino, &di); > out: > return error; > } > diff --git a/fs/xfs/scrub/parent.c b/fs/xfs/scrub/parent.c > index e2bda58c32f0..0a78d8411f23 100644 > --- a/fs/xfs/scrub/parent.c > +++ b/fs/xfs/scrub/parent.c > @@ -27,18 +27,18 @@ > > /* Set us up to scrub parents. */ > int > -xfs_scrub_setup_parent( > +xchk_setup_parent( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_inode_contents(sc, ip, 0); > + return xchk_setup_inode_contents(sc, ip, 0); > } > > /* Parent pointers */ > > /* Look for an entry in a parent pointing to this inode. */ > > -struct xfs_scrub_parent_ctx { > +struct xchk_parent_ctx { > struct dir_context dc; > xfs_ino_t ino; > xfs_nlink_t nlink; > @@ -46,7 +46,7 @@ struct xfs_scrub_parent_ctx { > > /* Look for a single entry in a directory pointing to an inode. */ > STATIC int > -xfs_scrub_parent_actor( > +xchk_parent_actor( > struct dir_context *dc, > const char *name, > int namelen, > @@ -54,9 +54,9 @@ xfs_scrub_parent_actor( > u64 ino, > unsigned type) > { > - struct xfs_scrub_parent_ctx *spc; > + struct xchk_parent_ctx *spc; > > - spc = container_of(dc, struct xfs_scrub_parent_ctx, dc); > + spc = container_of(dc, struct xchk_parent_ctx, dc); > if (spc->ino == ino) > spc->nlink++; > return 0; > @@ -64,13 +64,13 @@ xfs_scrub_parent_actor( > > /* Count the number of dentries in the parent dir that point to this inode. */ > STATIC int > -xfs_scrub_parent_count_parent_dentries( > +xchk_parent_count_parent_dentries( > struct xfs_scrub_context *sc, > struct xfs_inode *parent, > xfs_nlink_t *nlink) > { > - struct xfs_scrub_parent_ctx spc = { > - .dc.actor = xfs_scrub_parent_actor, > + struct xchk_parent_ctx spc = { > + .dc.actor = xchk_parent_actor, > .dc.pos = 0, > .ino = sc->ip->i_ino, > .nlink = 0, > @@ -120,7 +120,7 @@ xfs_scrub_parent_count_parent_dentries( > * entry pointing back to the inode being scrubbed. > */ > STATIC int > -xfs_scrub_parent_validate( > +xchk_parent_validate( > struct xfs_scrub_context *sc, > xfs_ino_t dnum, > bool *try_again) > @@ -138,7 +138,7 @@ xfs_scrub_parent_validate( > > /* '..' must not point to ourselves. */ > if (sc->ip->i_ino == dnum) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > @@ -165,13 +165,13 @@ xfs_scrub_parent_validate( > error = xfs_iget(mp, sc->tp, dnum, XFS_IGET_UNTRUSTED, 0, &dp); > if (error == -EINVAL) { > error = -EFSCORRUPTED; > - xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, &error); > + xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error); > goto out; > } > - if (!xfs_scrub_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, &error)) > + if (!xchk_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out; > if (dp == sc->ip || !S_ISDIR(VFS_I(dp)->i_mode)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out_rele; > } > > @@ -183,12 +183,12 @@ xfs_scrub_parent_validate( > * the child inodes. > */ > if (xfs_ilock_nowait(dp, XFS_IOLOCK_SHARED)) { > - error = xfs_scrub_parent_count_parent_dentries(sc, dp, &nlink); > - if (!xfs_scrub_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, > + error = xchk_parent_count_parent_dentries(sc, dp, &nlink); > + if (!xchk_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, > &error)) > goto out_unlock; > if (nlink != expected_nlink) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out_unlock; > } > > @@ -200,18 +200,18 @@ xfs_scrub_parent_validate( > */ > xfs_iunlock(sc->ip, sc->ilock_flags); > sc->ilock_flags = 0; > - error = xfs_scrub_ilock_inverted(dp, XFS_IOLOCK_SHARED); > + error = xchk_ilock_inverted(dp, XFS_IOLOCK_SHARED); > if (error) > goto out_rele; > > /* Go looking for our dentry. */ > - error = xfs_scrub_parent_count_parent_dentries(sc, dp, &nlink); > - if (!xfs_scrub_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, &error)) > + error = xchk_parent_count_parent_dentries(sc, dp, &nlink); > + if (!xchk_fblock_xref_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out_unlock; > > /* Drop the parent lock, relock this inode. */ > xfs_iunlock(dp, XFS_IOLOCK_SHARED); > - error = xfs_scrub_ilock_inverted(sc->ip, XFS_IOLOCK_EXCL); > + error = xchk_ilock_inverted(sc->ip, XFS_IOLOCK_EXCL); > if (error) > goto out_rele; > sc->ilock_flags = XFS_IOLOCK_EXCL; > @@ -225,7 +225,7 @@ xfs_scrub_parent_validate( > > /* Look up '..' to see if the inode changed. */ > error = xfs_dir_lookup(sc->tp, sc->ip, &xfs_name_dotdot, &dnum, NULL); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out_rele; > > /* Drat, parent changed. Try again! */ > @@ -241,7 +241,7 @@ xfs_scrub_parent_validate( > * for us in the parent. > */ > if (nlink != expected_nlink) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > return error; > > out_unlock: > @@ -254,7 +254,7 @@ xfs_scrub_parent_validate( > > /* Scrub a parent pointer. */ > int > -xfs_scrub_parent( > +xchk_parent( > struct xfs_scrub_context *sc) > { > struct xfs_mount *mp = sc->mp; > @@ -272,7 +272,7 @@ xfs_scrub_parent( > > /* We're not a special inode, are we? */ > if (!xfs_verify_dir_ino(mp, sc->ip->i_ino)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > @@ -288,10 +288,10 @@ xfs_scrub_parent( > > /* Look up '..' */ > error = xfs_dir_lookup(sc->tp, sc->ip, &xfs_name_dotdot, &dnum, NULL); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out; > if (!xfs_verify_dir_ino(mp, dnum)) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > @@ -299,12 +299,12 @@ xfs_scrub_parent( > if (sc->ip == mp->m_rootip) { > if (sc->ip->i_ino != mp->m_sb.sb_rootino || > sc->ip->i_ino != dnum) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > do { > - error = xfs_scrub_parent_validate(sc, dnum, &try_again); > + error = xchk_parent_validate(sc, dnum, &try_again); > if (error) > goto out; > } while (try_again && ++tries < 20); > @@ -314,7 +314,7 @@ xfs_scrub_parent( > * incomplete. Userspace can decide if it wants to try again. > */ > if (try_again && tries == 20) > - xfs_scrub_set_incomplete(sc); > + xchk_set_incomplete(sc); > out: > /* > * If we failed to lock the parent inode even after a retry, just mark > @@ -322,7 +322,7 @@ xfs_scrub_parent( > */ > if (sc->try_harder && error == -EDEADLOCK) { > error = 0; > - xfs_scrub_set_incomplete(sc); > + xchk_set_incomplete(sc); > } > return error; > } > diff --git a/fs/xfs/scrub/quota.c b/fs/xfs/scrub/quota.c > index 6ff906aa0a3b..d1b52dd7efcd 100644 > --- a/fs/xfs/scrub/quota.c > +++ b/fs/xfs/scrub/quota.c > @@ -30,7 +30,7 @@ > > /* Convert a scrub type code to a DQ flag, or return 0 if error. */ > static inline uint > -xfs_scrub_quota_to_dqtype( > +xchk_quota_to_dqtype( > struct xfs_scrub_context *sc) > { > switch (sc->sm->sm_type) { > @@ -47,7 +47,7 @@ xfs_scrub_quota_to_dqtype( > > /* Set us up to scrub a quota. */ > int > -xfs_scrub_setup_quota( > +xchk_setup_quota( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > @@ -57,14 +57,14 @@ xfs_scrub_setup_quota( > if (!XFS_IS_QUOTA_RUNNING(sc->mp) || !XFS_IS_QUOTA_ON(sc->mp)) > return -ENOENT; > > - dqtype = xfs_scrub_quota_to_dqtype(sc); > + dqtype = xchk_quota_to_dqtype(sc); > if (dqtype == 0) > return -EINVAL; > sc->has_quotaofflock = true; > mutex_lock(&sc->mp->m_quotainfo->qi_quotaofflock); > if (!xfs_this_quota_on(sc->mp, dqtype)) > return -ENOENT; > - error = xfs_scrub_setup_fs(sc, ip); > + error = xchk_setup_fs(sc, ip); > if (error) > return error; > sc->ip = xfs_quota_inode(sc->mp, dqtype); > @@ -75,19 +75,19 @@ xfs_scrub_setup_quota( > > /* Quotas. */ > > -struct xfs_scrub_quota_info { > +struct xchk_quota_info { > struct xfs_scrub_context *sc; > xfs_dqid_t last_id; > }; > > /* Scrub the fields in an individual quota item. */ > STATIC int > -xfs_scrub_quota_item( > +xchk_quota_item( > struct xfs_dquot *dq, > uint dqtype, > void *priv) > { > - struct xfs_scrub_quota_info *sqi = priv; > + struct xchk_quota_info *sqi = priv; > struct xfs_scrub_context *sc = sqi->sc; > struct xfs_mount *mp = sc->mp; > struct xfs_disk_dquot *d = &dq->q_core; > @@ -111,16 +111,16 @@ xfs_scrub_quota_item( > */ > offset = id / qi->qi_dqperchunk; > if (id && id <= sqi->last_id) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > sqi->last_id = id; > > /* Did we get the dquot type we wanted? */ > if (dqtype != (d->d_flags & XFS_DQ_ALLTYPES)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > if (d->d_pad0 != cpu_to_be32(0) || d->d_pad != cpu_to_be16(0)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > /* Check the limits. */ > bhard = be64_to_cpu(d->d_blk_hardlimit); > @@ -140,19 +140,19 @@ xfs_scrub_quota_item( > * the hard limit. > */ > if (bhard > mp->m_sb.sb_dblocks) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > if (bsoft > bhard) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > if (ihard > mp->m_maxicount) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > if (isoft > ihard) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > if (rhard > mp->m_sb.sb_rblocks) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > if (rsoft > rhard) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > /* Check the resource counts. */ > bcount = be64_to_cpu(d->d_bcount); > @@ -167,15 +167,15 @@ xfs_scrub_quota_item( > */ > if (xfs_sb_version_hasreflink(&mp->m_sb)) { > if (mp->m_sb.sb_dblocks < bcount) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, > offset); > } else { > if (mp->m_sb.sb_dblocks < bcount) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, > offset); > } > if (icount > fs_icount || rcount > mp->m_sb.sb_rblocks) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); > > /* > * We can violate the hard limits if the admin suddenly sets a > @@ -183,18 +183,18 @@ xfs_scrub_quota_item( > * admin review. > */ > if (id != 0 && bhard != 0 && bcount > bhard) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > if (id != 0 && ihard != 0 && icount > ihard) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > if (id != 0 && rhard != 0 && rcount > rhard) > - xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); > + xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); > > return 0; > } > > /* Check the quota's data fork. */ > STATIC int > -xfs_scrub_quota_data_fork( > +xchk_quota_data_fork( > struct xfs_scrub_context *sc) > { > struct xfs_bmbt_irec irec = { 0 }; > @@ -205,7 +205,7 @@ xfs_scrub_quota_data_fork( > int error = 0; > > /* Invoke the fork scrubber. */ > - error = xfs_scrub_metadata_inode_forks(sc); > + error = xchk_metadata_inode_forks(sc); > if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > return error; > > @@ -213,7 +213,7 @@ xfs_scrub_quota_data_fork( > max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk; > ifp = XFS_IFORK_PTR(sc->ip, XFS_DATA_FORK); > for_each_xfs_iext(ifp, &icur, &irec) { > - if (xfs_scrub_should_terminate(sc, &error)) > + if (xchk_should_terminate(sc, &error)) > break; > /* > * delalloc extents or blocks mapped above the highest > @@ -222,7 +222,7 @@ xfs_scrub_quota_data_fork( > if (isnullstartblock(irec.br_startblock) || > irec.br_startoff > max_dqid_off || > irec.br_startoff + irec.br_blockcount - 1 > max_dqid_off) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, > irec.br_startoff); > break; > } > @@ -233,19 +233,19 @@ xfs_scrub_quota_data_fork( > > /* Scrub all of a quota type's items. */ > int > -xfs_scrub_quota( > +xchk_quota( > struct xfs_scrub_context *sc) > { > - struct xfs_scrub_quota_info sqi; > + struct xchk_quota_info sqi; > struct xfs_mount *mp = sc->mp; > struct xfs_quotainfo *qi = mp->m_quotainfo; > uint dqtype; > int error = 0; > > - dqtype = xfs_scrub_quota_to_dqtype(sc); > + dqtype = xchk_quota_to_dqtype(sc); > > /* Look for problem extents. */ > - error = xfs_scrub_quota_data_fork(sc); > + error = xchk_quota_data_fork(sc); > if (error) > goto out; > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > @@ -260,10 +260,10 @@ xfs_scrub_quota( > sc->ilock_flags = 0; > sqi.sc = sc; > sqi.last_id = 0; > - error = xfs_qm_dqiterate(mp, dqtype, xfs_scrub_quota_item, &sqi); > + error = xfs_qm_dqiterate(mp, dqtype, xchk_quota_item, &sqi); > sc->ilock_flags = XFS_ILOCK_EXCL; > xfs_ilock(sc->ip, sc->ilock_flags); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, > sqi.last_id * qi->qi_dqperchunk, &error)) > goto out; > > diff --git a/fs/xfs/scrub/refcount.c b/fs/xfs/scrub/refcount.c > index 607a9faa8ecc..274febc49b23 100644 > --- a/fs/xfs/scrub/refcount.c > +++ b/fs/xfs/scrub/refcount.c > @@ -28,11 +28,11 @@ > * Set us up to scrub reference count btrees. > */ > int > -xfs_scrub_setup_ag_refcountbt( > +xchk_setup_ag_refcountbt( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_ag_btree(sc, ip, false); > + return xchk_setup_ag_btree(sc, ip, false); > } > > /* Reference count btree scrubber. */ > @@ -73,12 +73,12 @@ xfs_scrub_setup_ag_refcountbt( > * If the refcount is correct, all the check conditions in the algorithm > * should always hold true. If not, the refcount is incorrect. > */ > -struct xfs_scrub_refcnt_frag { > +struct xchk_refcnt_frag { > struct list_head list; > struct xfs_rmap_irec rm; > }; > > -struct xfs_scrub_refcnt_check { > +struct xchk_refcnt_check { > struct xfs_scrub_context *sc; > struct list_head fragments; > > @@ -99,18 +99,18 @@ struct xfs_scrub_refcnt_check { > * fragments as the refcountbt says we should have. > */ > STATIC int > -xfs_scrub_refcountbt_rmap_check( > +xchk_refcountbt_rmap_check( > struct xfs_btree_cur *cur, > struct xfs_rmap_irec *rec, > void *priv) > { > - struct xfs_scrub_refcnt_check *refchk = priv; > - struct xfs_scrub_refcnt_frag *frag; > + struct xchk_refcnt_check *refchk = priv; > + struct xchk_refcnt_frag *frag; > xfs_agblock_t rm_last; > xfs_agblock_t rc_last; > int error = 0; > > - if (xfs_scrub_should_terminate(refchk->sc, &error)) > + if (xchk_should_terminate(refchk->sc, &error)) > return error; > > rm_last = rec->rm_startblock + rec->rm_blockcount - 1; > @@ -118,7 +118,7 @@ xfs_scrub_refcountbt_rmap_check( > > /* Confirm that a single-owner refc extent is a CoW stage. */ > if (refchk->refcount == 1 && rec->rm_owner != XFS_RMAP_OWN_COW) { > - xfs_scrub_btree_xref_set_corrupt(refchk->sc, cur, 0); > + xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); > return 0; > } > > @@ -135,7 +135,7 @@ xfs_scrub_refcountbt_rmap_check( > * is healthy each rmap_irec we see will be in agbno order > * so we don't need insertion sort here. > */ > - frag = kmem_alloc(sizeof(struct xfs_scrub_refcnt_frag), > + frag = kmem_alloc(sizeof(struct xchk_refcnt_frag), > KM_MAYFAIL); > if (!frag) > return -ENOMEM; > @@ -154,12 +154,12 @@ xfs_scrub_refcountbt_rmap_check( > * we have a refcountbt error. > */ > STATIC void > -xfs_scrub_refcountbt_process_rmap_fragments( > - struct xfs_scrub_refcnt_check *refchk) > +xchk_refcountbt_process_rmap_fragments( > + struct xchk_refcnt_check *refchk) > { > struct list_head worklist; > - struct xfs_scrub_refcnt_frag *frag; > - struct xfs_scrub_refcnt_frag *n; > + struct xchk_refcnt_frag *frag; > + struct xchk_refcnt_frag *n; > xfs_agblock_t bno; > xfs_agblock_t rbno; > xfs_agblock_t next_rbno; > @@ -277,13 +277,13 @@ xfs_scrub_refcountbt_process_rmap_fragments( > > /* Use the rmap entries covering this extent to verify the refcount. */ > STATIC void > -xfs_scrub_refcountbt_xref_rmap( > +xchk_refcountbt_xref_rmap( > struct xfs_scrub_context *sc, > xfs_agblock_t bno, > xfs_extlen_t len, > xfs_nlink_t refcount) > { > - struct xfs_scrub_refcnt_check refchk = { > + struct xchk_refcnt_check refchk = { > .sc = sc, > .bno = bno, > .len = len, > @@ -292,11 +292,11 @@ xfs_scrub_refcountbt_xref_rmap( > }; > struct xfs_rmap_irec low; > struct xfs_rmap_irec high; > - struct xfs_scrub_refcnt_frag *frag; > - struct xfs_scrub_refcnt_frag *n; > + struct xchk_refcnt_frag *frag; > + struct xchk_refcnt_frag *n; > int error; > > - if (!sc->sa.rmap_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) > return; > > /* Cross-reference with the rmapbt to confirm the refcount. */ > @@ -307,13 +307,13 @@ xfs_scrub_refcountbt_xref_rmap( > > INIT_LIST_HEAD(&refchk.fragments); > error = xfs_rmap_query_range(sc->sa.rmap_cur, &low, &high, > - &xfs_scrub_refcountbt_rmap_check, &refchk); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + &xchk_refcountbt_rmap_check, &refchk); > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > goto out_free; > > - xfs_scrub_refcountbt_process_rmap_fragments(&refchk); > + xchk_refcountbt_process_rmap_fragments(&refchk); > if (refcount != refchk.seen) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > > out_free: > list_for_each_entry_safe(frag, n, &refchk.fragments, list) { > @@ -324,7 +324,7 @@ xfs_scrub_refcountbt_xref_rmap( > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_refcountbt_xref( > +xchk_refcountbt_xref( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len, > @@ -333,15 +333,15 @@ xfs_scrub_refcountbt_xref( > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, len); > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, len); > - xfs_scrub_refcountbt_xref_rmap(sc, agbno, len, refcount); > + xchk_xref_is_used_space(sc, agbno, len); > + xchk_xref_is_not_inode_chunk(sc, agbno, len); > + xchk_refcountbt_xref_rmap(sc, agbno, len, refcount); > } > > /* Scrub a refcountbt record. */ > STATIC int > -xfs_scrub_refcountbt_rec( > - struct xfs_scrub_btree *bs, > +xchk_refcountbt_rec( > + struct xchk_btree *bs, > union xfs_btree_rec *rec) > { > struct xfs_mount *mp = bs->cur->bc_mp; > @@ -360,7 +360,7 @@ xfs_scrub_refcountbt_rec( > /* Only CoW records can have refcount == 1. */ > has_cowflag = (bno & XFS_REFC_COW_START); > if ((refcount == 1 && !has_cowflag) || (refcount != 1 && has_cowflag)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > if (has_cowflag) > (*cow_blocks) += len; > > @@ -369,19 +369,19 @@ xfs_scrub_refcountbt_rec( > if (bno + len <= bno || > !xfs_verify_agbno(mp, agno, bno) || > !xfs_verify_agbno(mp, agno, bno + len - 1)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (refcount == 0) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > - xfs_scrub_refcountbt_xref(bs->sc, bno, len, refcount); > + xchk_refcountbt_xref(bs->sc, bno, len, refcount); > > return error; > } > > /* Make sure we have as many refc blocks as the rmap says. */ > STATIC void > -xfs_scrub_refcount_xref_rmap( > +xchk_refcount_xref_rmap( > struct xfs_scrub_context *sc, > struct xfs_owner_info *oinfo, > xfs_filblks_t cow_blocks) > @@ -390,33 +390,33 @@ xfs_scrub_refcount_xref_rmap( > xfs_filblks_t blocks; > int error; > > - if (!sc->sa.rmap_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) > return; > > /* Check that we saw as many refcbt blocks as the rmap knows about. */ > error = xfs_btree_count_blocks(sc->sa.refc_cur, &refcbt_blocks); > - if (!xfs_scrub_btree_process_error(sc, sc->sa.refc_cur, 0, &error)) > + if (!xchk_btree_process_error(sc, sc->sa.refc_cur, 0, &error)) > return; > - error = xfs_scrub_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, > + error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, > &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (blocks != refcbt_blocks) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > > /* Check that we saw as many cow blocks as the rmap knows about. */ > xfs_rmap_ag_owner(oinfo, XFS_RMAP_OWN_COW); > - error = xfs_scrub_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, > + error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, > &blocks); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (blocks != cow_blocks) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > } > > /* Scrub the refcount btree for some AG. */ > int > -xfs_scrub_refcountbt( > +xchk_refcountbt( > struct xfs_scrub_context *sc) > { > struct xfs_owner_info oinfo; > @@ -424,19 +424,19 @@ xfs_scrub_refcountbt( > int error; > > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_REFC); > - error = xfs_scrub_btree(sc, sc->sa.refc_cur, xfs_scrub_refcountbt_rec, > + error = xchk_btree(sc, sc->sa.refc_cur, xchk_refcountbt_rec, > &oinfo, &cow_blocks); > if (error) > return error; > > - xfs_scrub_refcount_xref_rmap(sc, &oinfo, cow_blocks); > + xchk_refcount_xref_rmap(sc, &oinfo, cow_blocks); > > return 0; > } > > /* xref check that a cow staging extent is marked in the refcountbt. */ > void > -xfs_scrub_xref_is_cow_staging( > +xchk_xref_is_cow_staging( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > @@ -446,35 +446,35 @@ xfs_scrub_xref_is_cow_staging( > int has_refcount; > int error; > > - if (!sc->sa.refc_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) > return; > > /* Find the CoW staging extent. */ > error = xfs_refcount_lookup_le(sc->sa.refc_cur, > agbno + XFS_REFC_COW_START, &has_refcount); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) > return; > if (!has_refcount) { > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > return; > } > > error = xfs_refcount_get_rec(sc->sa.refc_cur, &rc, &has_refcount); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) > return; > if (!has_refcount) { > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > return; > } > > /* CoW flag must be set, refcount must be 1. */ > has_cowflag = (rc.rc_startblock & XFS_REFC_COW_START); > if (!has_cowflag || rc.rc_refcount != 1) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > > /* Must be at least as long as what was passed in */ > if (rc.rc_blockcount < len) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > } > > /* > @@ -482,7 +482,7 @@ xfs_scrub_xref_is_cow_staging( > * can have multiple owners. > */ > void > -xfs_scrub_xref_is_not_shared( > +xchk_xref_is_not_shared( > struct xfs_scrub_context *sc, > xfs_agblock_t agbno, > xfs_extlen_t len) > @@ -490,12 +490,12 @@ xfs_scrub_xref_is_not_shared( > bool shared; > int error; > > - if (!sc->sa.refc_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) > return; > > error = xfs_refcount_has_record(sc->sa.refc_cur, agbno, len, &shared); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) > return; > if (shared) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > } > diff --git a/fs/xfs/scrub/repair.c b/fs/xfs/scrub/repair.c > index ea39e2bdc96a..5eccd89c64a8 100644 > --- a/fs/xfs/scrub/repair.c > +++ b/fs/xfs/scrub/repair.c > @@ -50,7 +50,7 @@ xfs_repair_attempt( > > trace_xfs_repair_attempt(ip, sc->sm, error); > > - xfs_scrub_ag_btcur_free(&sc->sa); > + xchk_ag_btcur_free(&sc->sa); > > /* Repair whatever's broken. */ > ASSERT(sc->ops->repair); > @@ -110,7 +110,7 @@ xfs_repair_probe( > { > int error = 0; > > - if (xfs_scrub_should_terminate(sc, &error)) > + if (xchk_should_terminate(sc, &error)) > return error; > > return 0; > diff --git a/fs/xfs/scrub/rmap.c b/fs/xfs/scrub/rmap.c > index c6d763236ba7..4b75fc2f31f3 100644 > --- a/fs/xfs/scrub/rmap.c > +++ b/fs/xfs/scrub/rmap.c > @@ -29,18 +29,18 @@ > * Set us up to scrub reverse mapping btrees. > */ > int > -xfs_scrub_setup_ag_rmapbt( > +xchk_setup_ag_rmapbt( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > - return xfs_scrub_setup_ag_btree(sc, ip, false); > + return xchk_setup_ag_btree(sc, ip, false); > } > > /* Reverse-mapping scrubber. */ > > /* Cross-reference a rmap against the refcount btree. */ > STATIC void > -xfs_scrub_rmapbt_xref_refc( > +xchk_rmapbt_xref_refc( > struct xfs_scrub_context *sc, > struct xfs_rmap_irec *irec) > { > @@ -52,7 +52,7 @@ xfs_scrub_rmapbt_xref_refc( > bool is_unwritten; > int error; > > - if (!sc->sa.refc_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) > return; > > non_inode = XFS_RMAP_NON_INODE_OWNER(irec->rm_owner); > @@ -63,15 +63,15 @@ xfs_scrub_rmapbt_xref_refc( > /* If this is shared, must be a data fork extent. */ > error = xfs_refcount_find_shared(sc->sa.refc_cur, irec->rm_startblock, > irec->rm_blockcount, &fbno, &flen, false); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) > return; > if (flen != 0 && (non_inode || is_attr || is_bmbt || is_unwritten)) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); > } > > /* Cross-reference with the other btrees. */ > STATIC void > -xfs_scrub_rmapbt_xref( > +xchk_rmapbt_xref( > struct xfs_scrub_context *sc, > struct xfs_rmap_irec *irec) > { > @@ -81,22 +81,22 @@ xfs_scrub_rmapbt_xref( > if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) > return; > > - xfs_scrub_xref_is_used_space(sc, agbno, len); > + xchk_xref_is_used_space(sc, agbno, len); > if (irec->rm_owner == XFS_RMAP_OWN_INODES) > - xfs_scrub_xref_is_inode_chunk(sc, agbno, len); > + xchk_xref_is_inode_chunk(sc, agbno, len); > else > - xfs_scrub_xref_is_not_inode_chunk(sc, agbno, len); > + xchk_xref_is_not_inode_chunk(sc, agbno, len); > if (irec->rm_owner == XFS_RMAP_OWN_COW) > - xfs_scrub_xref_is_cow_staging(sc, irec->rm_startblock, > + xchk_xref_is_cow_staging(sc, irec->rm_startblock, > irec->rm_blockcount); > else > - xfs_scrub_rmapbt_xref_refc(sc, irec); > + xchk_rmapbt_xref_refc(sc, irec); > } > > /* Scrub an rmapbt record. */ > STATIC int > -xfs_scrub_rmapbt_rec( > - struct xfs_scrub_btree *bs, > +xchk_rmapbt_rec( > + struct xchk_btree *bs, > union xfs_btree_rec *rec) > { > struct xfs_mount *mp = bs->cur->bc_mp; > @@ -109,12 +109,12 @@ xfs_scrub_rmapbt_rec( > int error; > > error = xfs_rmap_btrec_to_irec(rec, &irec); > - if (!xfs_scrub_btree_process_error(bs->sc, bs->cur, 0, &error)) > + if (!xchk_btree_process_error(bs->sc, bs->cur, 0, &error)) > goto out; > > /* Check extent. */ > if (irec.rm_startblock + irec.rm_blockcount <= irec.rm_startblock) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (irec.rm_owner == XFS_RMAP_OWN_FS) { > /* > @@ -124,7 +124,7 @@ xfs_scrub_rmapbt_rec( > */ > if (irec.rm_startblock != 0 || > irec.rm_blockcount != XFS_AGFL_BLOCK(mp) + 1) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > } else { > /* > * Otherwise we must point somewhere past the static metadata > @@ -133,7 +133,7 @@ xfs_scrub_rmapbt_rec( > if (!xfs_verify_agbno(mp, agno, irec.rm_startblock) || > !xfs_verify_agbno(mp, agno, irec.rm_startblock + > irec.rm_blockcount - 1)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > } > > /* Check flags. */ > @@ -143,47 +143,47 @@ xfs_scrub_rmapbt_rec( > is_unwritten = irec.rm_flags & XFS_RMAP_UNWRITTEN; > > if (is_bmbt && irec.rm_offset != 0) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (non_inode && irec.rm_offset != 0) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (is_unwritten && (is_bmbt || non_inode || is_attr)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (non_inode && (is_bmbt || is_unwritten || is_attr)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > > if (!non_inode) { > if (!xfs_verify_ino(mp, irec.rm_owner)) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > } else { > /* Non-inode owner within the magic values? */ > if (irec.rm_owner <= XFS_RMAP_OWN_MIN || > irec.rm_owner > XFS_RMAP_OWN_FS) > - xfs_scrub_btree_set_corrupt(bs->sc, bs->cur, 0); > + xchk_btree_set_corrupt(bs->sc, bs->cur, 0); > } > > - xfs_scrub_rmapbt_xref(bs->sc, &irec); > + xchk_rmapbt_xref(bs->sc, &irec); > out: > return error; > } > > /* Scrub the rmap btree for some AG. */ > int > -xfs_scrub_rmapbt( > +xchk_rmapbt( > struct xfs_scrub_context *sc) > { > struct xfs_owner_info oinfo; > > xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_AG); > - return xfs_scrub_btree(sc, sc->sa.rmap_cur, xfs_scrub_rmapbt_rec, > + return xchk_btree(sc, sc->sa.rmap_cur, xchk_rmapbt_rec, > &oinfo, NULL); > } > > /* xref check that the extent is owned by a given owner */ > static inline void > -xfs_scrub_xref_check_owner( > +xchk_xref_check_owner( > struct xfs_scrub_context *sc, > xfs_agblock_t bno, > xfs_extlen_t len, > @@ -193,42 +193,42 @@ xfs_scrub_xref_check_owner( > bool has_rmap; > int error; > > - if (!sc->sa.rmap_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) > return; > > error = xfs_rmap_record_exists(sc->sa.rmap_cur, bno, len, oinfo, > &has_rmap); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (has_rmap != should_have_rmap) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > } > > /* xref check that the extent is owned by a given owner */ > void > -xfs_scrub_xref_is_owned_by( > +xchk_xref_is_owned_by( > struct xfs_scrub_context *sc, > xfs_agblock_t bno, > xfs_extlen_t len, > struct xfs_owner_info *oinfo) > { > - xfs_scrub_xref_check_owner(sc, bno, len, oinfo, true); > + xchk_xref_check_owner(sc, bno, len, oinfo, true); > } > > /* xref check that the extent is not owned by a given owner */ > void > -xfs_scrub_xref_is_not_owned_by( > +xchk_xref_is_not_owned_by( > struct xfs_scrub_context *sc, > xfs_agblock_t bno, > xfs_extlen_t len, > struct xfs_owner_info *oinfo) > { > - xfs_scrub_xref_check_owner(sc, bno, len, oinfo, false); > + xchk_xref_check_owner(sc, bno, len, oinfo, false); > } > > /* xref check that the extent has no reverse mapping at all */ > void > -xfs_scrub_xref_has_no_owner( > +xchk_xref_has_no_owner( > struct xfs_scrub_context *sc, > xfs_agblock_t bno, > xfs_extlen_t len) > @@ -236,12 +236,12 @@ xfs_scrub_xref_has_no_owner( > bool has_rmap; > int error; > > - if (!sc->sa.rmap_cur || xfs_scrub_skip_xref(sc->sm)) > + if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) > return; > > error = xfs_rmap_has_record(sc->sa.rmap_cur, bno, len, &has_rmap); > - if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > + if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) > return; > if (has_rmap) > - xfs_scrub_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > + xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); > } > diff --git a/fs/xfs/scrub/rtbitmap.c b/fs/xfs/scrub/rtbitmap.c > index 1f86e02a07ca..3f0fc83562ae 100644 > --- a/fs/xfs/scrub/rtbitmap.c > +++ b/fs/xfs/scrub/rtbitmap.c > @@ -25,13 +25,13 @@ > > /* Set us up with the realtime metadata locked. */ > int > -xfs_scrub_setup_rt( > +xchk_setup_rt( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > int error; > > - error = xfs_scrub_setup_fs(sc, ip); > + error = xchk_setup_fs(sc, ip); > if (error) > return error; > > @@ -46,7 +46,7 @@ xfs_scrub_setup_rt( > > /* Scrub a free extent record from the realtime bitmap. */ > STATIC int > -xfs_scrub_rtbitmap_rec( > +xchk_rtbitmap_rec( > struct xfs_trans *tp, > struct xfs_rtalloc_rec *rec, > void *priv) > @@ -61,24 +61,24 @@ xfs_scrub_rtbitmap_rec( > if (startblock + blockcount <= startblock || > !xfs_verify_rtbno(sc->mp, startblock) || > !xfs_verify_rtbno(sc->mp, startblock + blockcount - 1)) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > return 0; > } > > /* Scrub the realtime bitmap. */ > int > -xfs_scrub_rtbitmap( > +xchk_rtbitmap( > struct xfs_scrub_context *sc) > { > int error; > > /* Invoke the fork scrubber. */ > - error = xfs_scrub_metadata_inode_forks(sc); > + error = xchk_metadata_inode_forks(sc); > if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > return error; > > - error = xfs_rtalloc_query_all(sc->tp, xfs_scrub_rtbitmap_rec, sc); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > + error = xfs_rtalloc_query_all(sc->tp, xchk_rtbitmap_rec, sc); > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out; > > out: > @@ -87,7 +87,7 @@ xfs_scrub_rtbitmap( > > /* Scrub the realtime summary. */ > int > -xfs_scrub_rtsummary( > +xchk_rtsummary( > struct xfs_scrub_context *sc) > { > struct xfs_inode *rsumip = sc->mp->m_rsumip; > @@ -107,12 +107,12 @@ xfs_scrub_rtsummary( > xfs_ilock(sc->ip, sc->ilock_flags); > > /* Invoke the fork scrubber. */ > - error = xfs_scrub_metadata_inode_forks(sc); > + error = xchk_metadata_inode_forks(sc); > if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) > goto out; > > /* XXX: implement this some day */ > - xfs_scrub_set_incomplete(sc); > + xchk_set_incomplete(sc); > out: > /* Switch back to the rtbitmap inode and lock flags. */ > xfs_iunlock(sc->ip, sc->ilock_flags); > @@ -124,7 +124,7 @@ xfs_scrub_rtsummary( > > /* xref check that the extent is not free in the rtbitmap */ > void > -xfs_scrub_xref_is_used_rt_space( > +xchk_xref_is_used_rt_space( > struct xfs_scrub_context *sc, > xfs_rtblock_t fsbno, > xfs_extlen_t len) > @@ -135,7 +135,7 @@ xfs_scrub_xref_is_used_rt_space( > bool is_free; > int error; > > - if (xfs_scrub_skip_xref(sc->sm)) > + if (xchk_skip_xref(sc->sm)) > return; > > startext = fsbno; > @@ -147,10 +147,10 @@ xfs_scrub_xref_is_used_rt_space( > xfs_ilock(sc->mp->m_rbmip, XFS_ILOCK_SHARED | XFS_ILOCK_RTBITMAP); > error = xfs_rtalloc_extent_is_free(sc->mp, sc->tp, startext, extcount, > &is_free); > - if (!xfs_scrub_should_check_xref(sc, &error, NULL)) > + if (!xchk_should_check_xref(sc, &error, NULL)) > goto out_unlock; > if (is_free) > - xfs_scrub_ino_xref_set_corrupt(sc, sc->mp->m_rbmip->i_ino); > + xchk_ino_xref_set_corrupt(sc, sc->mp->m_rbmip->i_ino); > out_unlock: > xfs_iunlock(sc->mp->m_rbmip, XFS_ILOCK_SHARED | XFS_ILOCK_RTBITMAP); > } > diff --git a/fs/xfs/scrub/scrub.c b/fs/xfs/scrub/scrub.c > index 58ae76b3a421..ead97ab91a3c 100644 > --- a/fs/xfs/scrub/scrub.c > +++ b/fs/xfs/scrub/scrub.c > @@ -131,6 +131,12 @@ > * optimize the structure so that the rebuild knows what to do. The > * second check evaluates the completeness of the repair; that is what > * is reported to userspace. > + * > + * A quick note on symbol prefixes: > + * - "xfs_" are general XFS symbols. > + * - "xchk_" are symbols related to metadata checking. > + * - "xrep_" are symbols related to metadata repair. > + * - "xfs_scrub_" are symbols that tie online fsck to the rest of XFS. > */ > > /* > @@ -144,12 +150,12 @@ > * supported by the running kernel. > */ > static int > -xfs_scrub_probe( > +xchk_probe( > struct xfs_scrub_context *sc) > { > int error = 0; > > - if (xfs_scrub_should_terminate(sc, &error)) > + if (xchk_should_terminate(sc, &error)) > return error; > > return 0; > @@ -159,12 +165,12 @@ xfs_scrub_probe( > > /* Free all the resources and finish the transactions. */ > STATIC int > -xfs_scrub_teardown( > +xchk_teardown( > struct xfs_scrub_context *sc, > struct xfs_inode *ip_in, > int error) > { > - xfs_scrub_ag_free(sc, &sc->sa); > + xchk_ag_free(sc, &sc->sa); > if (sc->tp) { > if (error == 0 && (sc->sm->sm_flags & XFS_SCRUB_IFLAG_REPAIR)) > error = xfs_trans_commit(sc->tp); > @@ -191,165 +197,165 @@ xfs_scrub_teardown( > > /* Scrubbing dispatch. */ > > -static const struct xfs_scrub_meta_ops meta_scrub_ops[] = { > +static const struct xchk_meta_ops meta_scrub_ops[] = { > [XFS_SCRUB_TYPE_PROBE] = { /* ioctl presence test */ > .type = ST_NONE, > - .setup = xfs_scrub_setup_fs, > - .scrub = xfs_scrub_probe, > + .setup = xchk_setup_fs, > + .scrub = xchk_probe, > .repair = xfs_repair_probe, > }, > [XFS_SCRUB_TYPE_SB] = { /* superblock */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_fs, > - .scrub = xfs_scrub_superblock, > + .setup = xchk_setup_fs, > + .scrub = xchk_superblock, > .repair = xfs_repair_superblock, > }, > [XFS_SCRUB_TYPE_AGF] = { /* agf */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_fs, > - .scrub = xfs_scrub_agf, > + .setup = xchk_setup_fs, > + .scrub = xchk_agf, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_AGFL]= { /* agfl */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_fs, > - .scrub = xfs_scrub_agfl, > + .setup = xchk_setup_fs, > + .scrub = xchk_agfl, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_AGI] = { /* agi */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_fs, > - .scrub = xfs_scrub_agi, > + .setup = xchk_setup_fs, > + .scrub = xchk_agi, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_BNOBT] = { /* bnobt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_allocbt, > - .scrub = xfs_scrub_bnobt, > + .setup = xchk_setup_ag_allocbt, > + .scrub = xchk_bnobt, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_CNTBT] = { /* cntbt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_allocbt, > - .scrub = xfs_scrub_cntbt, > + .setup = xchk_setup_ag_allocbt, > + .scrub = xchk_cntbt, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_INOBT] = { /* inobt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_iallocbt, > - .scrub = xfs_scrub_inobt, > + .setup = xchk_setup_ag_iallocbt, > + .scrub = xchk_inobt, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_FINOBT] = { /* finobt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_iallocbt, > - .scrub = xfs_scrub_finobt, > + .setup = xchk_setup_ag_iallocbt, > + .scrub = xchk_finobt, > .has = xfs_sb_version_hasfinobt, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_RMAPBT] = { /* rmapbt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_rmapbt, > - .scrub = xfs_scrub_rmapbt, > + .setup = xchk_setup_ag_rmapbt, > + .scrub = xchk_rmapbt, > .has = xfs_sb_version_hasrmapbt, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_REFCNTBT] = { /* refcountbt */ > .type = ST_PERAG, > - .setup = xfs_scrub_setup_ag_refcountbt, > - .scrub = xfs_scrub_refcountbt, > + .setup = xchk_setup_ag_refcountbt, > + .scrub = xchk_refcountbt, > .has = xfs_sb_version_hasreflink, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_INODE] = { /* inode record */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_inode, > - .scrub = xfs_scrub_inode, > + .setup = xchk_setup_inode, > + .scrub = xchk_inode, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_BMBTD] = { /* inode data fork */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_inode_bmap, > - .scrub = xfs_scrub_bmap_data, > + .setup = xchk_setup_inode_bmap, > + .scrub = xchk_bmap_data, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_BMBTA] = { /* inode attr fork */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_inode_bmap, > - .scrub = xfs_scrub_bmap_attr, > + .setup = xchk_setup_inode_bmap, > + .scrub = xchk_bmap_attr, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_BMBTC] = { /* inode CoW fork */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_inode_bmap, > - .scrub = xfs_scrub_bmap_cow, > + .setup = xchk_setup_inode_bmap, > + .scrub = xchk_bmap_cow, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_DIR] = { /* directory */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_directory, > - .scrub = xfs_scrub_directory, > + .setup = xchk_setup_directory, > + .scrub = xchk_directory, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_XATTR] = { /* extended attributes */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_xattr, > - .scrub = xfs_scrub_xattr, > + .setup = xchk_setup_xattr, > + .scrub = xchk_xattr, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_SYMLINK] = { /* symbolic link */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_symlink, > - .scrub = xfs_scrub_symlink, > + .setup = xchk_setup_symlink, > + .scrub = xchk_symlink, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_PARENT] = { /* parent pointers */ > .type = ST_INODE, > - .setup = xfs_scrub_setup_parent, > - .scrub = xfs_scrub_parent, > + .setup = xchk_setup_parent, > + .scrub = xchk_parent, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_RTBITMAP] = { /* realtime bitmap */ > .type = ST_FS, > - .setup = xfs_scrub_setup_rt, > - .scrub = xfs_scrub_rtbitmap, > + .setup = xchk_setup_rt, > + .scrub = xchk_rtbitmap, > .has = xfs_sb_version_hasrealtime, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_RTSUM] = { /* realtime summary */ > .type = ST_FS, > - .setup = xfs_scrub_setup_rt, > - .scrub = xfs_scrub_rtsummary, > + .setup = xchk_setup_rt, > + .scrub = xchk_rtsummary, > .has = xfs_sb_version_hasrealtime, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_UQUOTA] = { /* user quota */ > .type = ST_FS, > - .setup = xfs_scrub_setup_quota, > - .scrub = xfs_scrub_quota, > + .setup = xchk_setup_quota, > + .scrub = xchk_quota, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_GQUOTA] = { /* group quota */ > .type = ST_FS, > - .setup = xfs_scrub_setup_quota, > - .scrub = xfs_scrub_quota, > + .setup = xchk_setup_quota, > + .scrub = xchk_quota, > .repair = xfs_repair_notsupported, > }, > [XFS_SCRUB_TYPE_PQUOTA] = { /* project quota */ > .type = ST_FS, > - .setup = xfs_scrub_setup_quota, > - .scrub = xfs_scrub_quota, > + .setup = xchk_setup_quota, > + .scrub = xchk_quota, > .repair = xfs_repair_notsupported, > }, > }; > > /* This isn't a stable feature, warn once per day. */ > static inline void > -xfs_scrub_experimental_warning( > +xchk_experimental_warning( > struct xfs_mount *mp) > { > static struct ratelimit_state scrub_warning = RATELIMIT_STATE_INIT( > - "xfs_scrub_warning", 86400 * HZ, 1); > + "xchk_warning", 86400 * HZ, 1); > ratelimit_set_flags(&scrub_warning, RATELIMIT_MSG_ON_RELEASE); > > if (__ratelimit(&scrub_warning)) > @@ -358,12 +364,12 @@ xfs_scrub_experimental_warning( > } > > static int > -xfs_scrub_validate_inputs( > +xchk_validate_inputs( > struct xfs_mount *mp, > struct xfs_scrub_metadata *sm) > { > int error; > - const struct xfs_scrub_meta_ops *ops; > + const struct xchk_meta_ops *ops; > > error = -EINVAL; > /* Check our inputs. */ > @@ -441,7 +447,7 @@ xfs_scrub_validate_inputs( > } > > #ifdef CONFIG_XFS_ONLINE_REPAIR > -static inline void xfs_scrub_postmortem(struct xfs_scrub_context *sc) > +static inline void xchk_postmortem(struct xfs_scrub_context *sc) > { > /* > * Userspace asked us to repair something, we repaired it, rescanned > @@ -454,7 +460,7 @@ static inline void xfs_scrub_postmortem(struct xfs_scrub_context *sc) > xfs_repair_failure(sc->mp); > } > #else > -static inline void xfs_scrub_postmortem(struct xfs_scrub_context *sc) > +static inline void xchk_postmortem(struct xfs_scrub_context *sc) > { > /* > * Userspace asked us to scrub something, it's broken, and we have no > @@ -480,9 +486,9 @@ xfs_scrub_metadata( > int error = 0; > > BUILD_BUG_ON(sizeof(meta_scrub_ops) != > - (sizeof(struct xfs_scrub_meta_ops) * XFS_SCRUB_TYPE_NR)); > + (sizeof(struct xchk_meta_ops) * XFS_SCRUB_TYPE_NR)); > > - trace_xfs_scrub_start(ip, sm, error); > + trace_xchk_start(ip, sm, error); > > /* Forbidden if we are shut down or mounted norecovery. */ > error = -ESHUTDOWN; > @@ -492,11 +498,11 @@ xfs_scrub_metadata( > if (mp->m_flags & XFS_MOUNT_NORECOVERY) > goto out; > > - error = xfs_scrub_validate_inputs(mp, sm); > + error = xchk_validate_inputs(mp, sm); > if (error) > goto out; > > - xfs_scrub_experimental_warning(mp); > + xchk_experimental_warning(mp); > > retry_op: > /* Set up for the operation. */ > @@ -518,7 +524,7 @@ xfs_scrub_metadata( > * Tear down everything we hold, then set up again with > * preparation for worst-case scenarios. > */ > - error = xfs_scrub_teardown(&sc, ip, 0); > + error = xchk_teardown(&sc, ip, 0); > if (error) > goto out; > try_harder = true; > @@ -553,7 +559,7 @@ xfs_scrub_metadata( > if (error == -EAGAIN) { > if (sc.try_harder) > try_harder = true; > - error = xfs_scrub_teardown(&sc, ip, 0); > + error = xchk_teardown(&sc, ip, 0); > if (error) { > xfs_repair_failure(mp); > goto out; > @@ -563,11 +569,11 @@ xfs_scrub_metadata( > } > > out_nofix: > - xfs_scrub_postmortem(&sc); > + xchk_postmortem(&sc); > out_teardown: > - error = xfs_scrub_teardown(&sc, ip, error); > + error = xchk_teardown(&sc, ip, error); > out: > - trace_xfs_scrub_done(ip, sm, error); > + trace_xchk_done(ip, sm, error); > if (error == -EFSCORRUPTED || error == -EFSBADCRC) { > sm->sm_flags |= XFS_SCRUB_OFLAG_CORRUPT; > error = 0; > diff --git a/fs/xfs/scrub/scrub.h b/fs/xfs/scrub/scrub.h > index b295edd5fc0e..0f59a47c4bb0 100644 > --- a/fs/xfs/scrub/scrub.h > +++ b/fs/xfs/scrub/scrub.h > @@ -9,14 +9,14 @@ > struct xfs_scrub_context; > > /* Type info and names for the scrub types. */ > -enum xfs_scrub_type { > +enum xchk_type { > ST_NONE = 1, /* disabled */ > ST_PERAG, /* per-AG metadata */ > ST_FS, /* per-FS metadata */ > ST_INODE, /* per-inode metadata */ > }; > > -struct xfs_scrub_meta_ops { > +struct xchk_meta_ops { > /* Acquire whatever resources are needed for the operation. */ > int (*setup)(struct xfs_scrub_context *, > struct xfs_inode *); > @@ -31,11 +31,11 @@ struct xfs_scrub_meta_ops { > bool (*has)(struct xfs_sb *); > > /* type describing required/allowed inputs */ > - enum xfs_scrub_type type; > + enum xchk_type type; > }; > > /* Buffer pointers and btree cursors for an entire AG. */ > -struct xfs_scrub_ag { > +struct xchk_ag { > xfs_agnumber_t agno; > struct xfs_perag *pag; > > @@ -57,7 +57,7 @@ struct xfs_scrub_context { > /* General scrub state. */ > struct xfs_mount *mp; > struct xfs_scrub_metadata *sm; > - const struct xfs_scrub_meta_ops *ops; > + const struct xchk_meta_ops *ops; > struct xfs_trans *tp; > struct xfs_inode *ip; > void *buf; > @@ -66,78 +66,78 @@ struct xfs_scrub_context { > bool has_quotaofflock; > > /* State tracking for single-AG operations. */ > - struct xfs_scrub_ag sa; > + struct xchk_ag sa; > }; > > /* Metadata scrubbers */ > -int xfs_scrub_tester(struct xfs_scrub_context *sc); > -int xfs_scrub_superblock(struct xfs_scrub_context *sc); > -int xfs_scrub_agf(struct xfs_scrub_context *sc); > -int xfs_scrub_agfl(struct xfs_scrub_context *sc); > -int xfs_scrub_agi(struct xfs_scrub_context *sc); > -int xfs_scrub_bnobt(struct xfs_scrub_context *sc); > -int xfs_scrub_cntbt(struct xfs_scrub_context *sc); > -int xfs_scrub_inobt(struct xfs_scrub_context *sc); > -int xfs_scrub_finobt(struct xfs_scrub_context *sc); > -int xfs_scrub_rmapbt(struct xfs_scrub_context *sc); > -int xfs_scrub_refcountbt(struct xfs_scrub_context *sc); > -int xfs_scrub_inode(struct xfs_scrub_context *sc); > -int xfs_scrub_bmap_data(struct xfs_scrub_context *sc); > -int xfs_scrub_bmap_attr(struct xfs_scrub_context *sc); > -int xfs_scrub_bmap_cow(struct xfs_scrub_context *sc); > -int xfs_scrub_directory(struct xfs_scrub_context *sc); > -int xfs_scrub_xattr(struct xfs_scrub_context *sc); > -int xfs_scrub_symlink(struct xfs_scrub_context *sc); > -int xfs_scrub_parent(struct xfs_scrub_context *sc); > +int xchk_tester(struct xfs_scrub_context *sc); > +int xchk_superblock(struct xfs_scrub_context *sc); > +int xchk_agf(struct xfs_scrub_context *sc); > +int xchk_agfl(struct xfs_scrub_context *sc); > +int xchk_agi(struct xfs_scrub_context *sc); > +int xchk_bnobt(struct xfs_scrub_context *sc); > +int xchk_cntbt(struct xfs_scrub_context *sc); > +int xchk_inobt(struct xfs_scrub_context *sc); > +int xchk_finobt(struct xfs_scrub_context *sc); > +int xchk_rmapbt(struct xfs_scrub_context *sc); > +int xchk_refcountbt(struct xfs_scrub_context *sc); > +int xchk_inode(struct xfs_scrub_context *sc); > +int xchk_bmap_data(struct xfs_scrub_context *sc); > +int xchk_bmap_attr(struct xfs_scrub_context *sc); > +int xchk_bmap_cow(struct xfs_scrub_context *sc); > +int xchk_directory(struct xfs_scrub_context *sc); > +int xchk_xattr(struct xfs_scrub_context *sc); > +int xchk_symlink(struct xfs_scrub_context *sc); > +int xchk_parent(struct xfs_scrub_context *sc); > #ifdef CONFIG_XFS_RT > -int xfs_scrub_rtbitmap(struct xfs_scrub_context *sc); > -int xfs_scrub_rtsummary(struct xfs_scrub_context *sc); > +int xchk_rtbitmap(struct xfs_scrub_context *sc); > +int xchk_rtsummary(struct xfs_scrub_context *sc); > #else > static inline int > -xfs_scrub_rtbitmap(struct xfs_scrub_context *sc) > +xchk_rtbitmap(struct xfs_scrub_context *sc) > { > return -ENOENT; > } > static inline int > -xfs_scrub_rtsummary(struct xfs_scrub_context *sc) > +xchk_rtsummary(struct xfs_scrub_context *sc) > { > return -ENOENT; > } > #endif > #ifdef CONFIG_XFS_QUOTA > -int xfs_scrub_quota(struct xfs_scrub_context *sc); > +int xchk_quota(struct xfs_scrub_context *sc); > #else > static inline int > -xfs_scrub_quota(struct xfs_scrub_context *sc) > +xchk_quota(struct xfs_scrub_context *sc) > { > return -ENOENT; > } > #endif > > /* cross-referencing helpers */ > -void xfs_scrub_xref_is_used_space(struct xfs_scrub_context *sc, > +void xchk_xref_is_used_space(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len); > -void xfs_scrub_xref_is_not_inode_chunk(struct xfs_scrub_context *sc, > +void xchk_xref_is_not_inode_chunk(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len); > -void xfs_scrub_xref_is_inode_chunk(struct xfs_scrub_context *sc, > +void xchk_xref_is_inode_chunk(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len); > -void xfs_scrub_xref_is_owned_by(struct xfs_scrub_context *sc, > +void xchk_xref_is_owned_by(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len, > struct xfs_owner_info *oinfo); > -void xfs_scrub_xref_is_not_owned_by(struct xfs_scrub_context *sc, > +void xchk_xref_is_not_owned_by(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len, > struct xfs_owner_info *oinfo); > -void xfs_scrub_xref_has_no_owner(struct xfs_scrub_context *sc, > +void xchk_xref_has_no_owner(struct xfs_scrub_context *sc, > xfs_agblock_t agbno, xfs_extlen_t len); > -void xfs_scrub_xref_is_cow_staging(struct xfs_scrub_context *sc, > +void xchk_xref_is_cow_staging(struct xfs_scrub_context *sc, > xfs_agblock_t bno, xfs_extlen_t len); > -void xfs_scrub_xref_is_not_shared(struct xfs_scrub_context *sc, > +void xchk_xref_is_not_shared(struct xfs_scrub_context *sc, > xfs_agblock_t bno, xfs_extlen_t len); > #ifdef CONFIG_XFS_RT > -void xfs_scrub_xref_is_used_rt_space(struct xfs_scrub_context *sc, > +void xchk_xref_is_used_rt_space(struct xfs_scrub_context *sc, > xfs_rtblock_t rtbno, xfs_extlen_t len); > #else > -# define xfs_scrub_xref_is_used_rt_space(sc, rtbno, len) do { } while (0) > +# define xchk_xref_is_used_rt_space(sc, rtbno, len) do { } while (0) > #endif > > #endif /* __XFS_SCRUB_SCRUB_H__ */ > diff --git a/fs/xfs/scrub/symlink.c b/fs/xfs/scrub/symlink.c > index 570a89812116..e2a288e34337 100644 > --- a/fs/xfs/scrub/symlink.c > +++ b/fs/xfs/scrub/symlink.c > @@ -25,7 +25,7 @@ > > /* Set us up to scrub a symbolic link. */ > int > -xfs_scrub_setup_symlink( > +xchk_setup_symlink( > struct xfs_scrub_context *sc, > struct xfs_inode *ip) > { > @@ -34,13 +34,13 @@ xfs_scrub_setup_symlink( > if (!sc->buf) > return -ENOMEM; > > - return xfs_scrub_setup_inode_contents(sc, ip, 0); > + return xchk_setup_inode_contents(sc, ip, 0); > } > > /* Symbolic links. */ > > int > -xfs_scrub_symlink( > +xchk_symlink( > struct xfs_scrub_context *sc) > { > struct xfs_inode *ip = sc->ip; > @@ -55,7 +55,7 @@ xfs_scrub_symlink( > > /* Plausible size? */ > if (len > XFS_SYMLINK_MAXLEN || len <= 0) { > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > @@ -63,16 +63,16 @@ xfs_scrub_symlink( > if (ifp->if_flags & XFS_IFINLINE) { > if (len > XFS_IFORK_DSIZE(ip) || > len > strnlen(ifp->if_u1.if_data, XFS_IFORK_DSIZE(ip))) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > goto out; > } > > /* Remote symlink; must read the contents. */ > error = xfs_readlink_bmap_ilocked(sc->ip, sc->buf); > - if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > + if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, 0, &error)) > goto out; > if (strnlen(sc->buf, XFS_SYMLINK_MAXLEN) < len) > - xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > + xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, 0); > out: > return error; > } > diff --git a/fs/xfs/scrub/trace.c b/fs/xfs/scrub/trace.c > index 7c76d8b5cb05..1ef1202a1e45 100644 > --- a/fs/xfs/scrub/trace.c > +++ b/fs/xfs/scrub/trace.c > @@ -22,7 +22,7 @@ > > /* Figure out which block the btree cursor was pointing to. */ > static inline xfs_fsblock_t > -xfs_scrub_btree_cur_fsbno( > +xchk_btree_cur_fsbno( > struct xfs_btree_cur *cur, > int level) > { > diff --git a/fs/xfs/scrub/trace.h b/fs/xfs/scrub/trace.h > index cec3e5ece5a1..11967d3942ac 100644 > --- a/fs/xfs/scrub/trace.h > +++ b/fs/xfs/scrub/trace.h > @@ -12,7 +12,7 @@ > #include <linux/tracepoint.h> > #include "xfs_bit.h" > > -DECLARE_EVENT_CLASS(xfs_scrub_class, > +DECLARE_EVENT_CLASS(xchk_class, > TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_metadata *sm, > int error), > TP_ARGS(ip, sm, error), > @@ -47,18 +47,18 @@ DECLARE_EVENT_CLASS(xfs_scrub_class, > __entry->error) > ) > #define DEFINE_SCRUB_EVENT(name) \ > -DEFINE_EVENT(xfs_scrub_class, name, \ > +DEFINE_EVENT(xchk_class, name, \ > TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_metadata *sm, \ > int error), \ > TP_ARGS(ip, sm, error)) > > -DEFINE_SCRUB_EVENT(xfs_scrub_start); > -DEFINE_SCRUB_EVENT(xfs_scrub_done); > -DEFINE_SCRUB_EVENT(xfs_scrub_deadlock_retry); > +DEFINE_SCRUB_EVENT(xchk_start); > +DEFINE_SCRUB_EVENT(xchk_done); > +DEFINE_SCRUB_EVENT(xchk_deadlock_retry); > DEFINE_SCRUB_EVENT(xfs_repair_attempt); > DEFINE_SCRUB_EVENT(xfs_repair_done); > > -TRACE_EVENT(xfs_scrub_op_error, > +TRACE_EVENT(xchk_op_error, > TP_PROTO(struct xfs_scrub_context *sc, xfs_agnumber_t agno, > xfs_agblock_t bno, int error, void *ret_ip), > TP_ARGS(sc, agno, bno, error, ret_ip), > @@ -87,7 +87,7 @@ TRACE_EVENT(xfs_scrub_op_error, > __entry->ret_ip) > ); > > -TRACE_EVENT(xfs_scrub_file_op_error, > +TRACE_EVENT(xchk_file_op_error, > TP_PROTO(struct xfs_scrub_context *sc, int whichfork, > xfs_fileoff_t offset, int error, void *ret_ip), > TP_ARGS(sc, whichfork, offset, error, ret_ip), > @@ -119,7 +119,7 @@ TRACE_EVENT(xfs_scrub_file_op_error, > __entry->ret_ip) > ); > > -DECLARE_EVENT_CLASS(xfs_scrub_block_error_class, > +DECLARE_EVENT_CLASS(xchk_block_error_class, > TP_PROTO(struct xfs_scrub_context *sc, xfs_daddr_t daddr, void *ret_ip), > TP_ARGS(sc, daddr, ret_ip), > TP_STRUCT__entry( > @@ -153,15 +153,15 @@ DECLARE_EVENT_CLASS(xfs_scrub_block_error_class, > ) > > #define DEFINE_SCRUB_BLOCK_ERROR_EVENT(name) \ > -DEFINE_EVENT(xfs_scrub_block_error_class, name, \ > +DEFINE_EVENT(xchk_block_error_class, name, \ > TP_PROTO(struct xfs_scrub_context *sc, xfs_daddr_t daddr, \ > void *ret_ip), \ > TP_ARGS(sc, daddr, ret_ip)) > > -DEFINE_SCRUB_BLOCK_ERROR_EVENT(xfs_scrub_block_error); > -DEFINE_SCRUB_BLOCK_ERROR_EVENT(xfs_scrub_block_preen); > +DEFINE_SCRUB_BLOCK_ERROR_EVENT(xchk_block_error); > +DEFINE_SCRUB_BLOCK_ERROR_EVENT(xchk_block_preen); > > -DECLARE_EVENT_CLASS(xfs_scrub_ino_error_class, > +DECLARE_EVENT_CLASS(xchk_ino_error_class, > TP_PROTO(struct xfs_scrub_context *sc, xfs_ino_t ino, void *ret_ip), > TP_ARGS(sc, ino, ret_ip), > TP_STRUCT__entry( > @@ -184,16 +184,16 @@ DECLARE_EVENT_CLASS(xfs_scrub_ino_error_class, > ) > > #define DEFINE_SCRUB_INO_ERROR_EVENT(name) \ > -DEFINE_EVENT(xfs_scrub_ino_error_class, name, \ > +DEFINE_EVENT(xchk_ino_error_class, name, \ > TP_PROTO(struct xfs_scrub_context *sc, xfs_ino_t ino, \ > void *ret_ip), \ > TP_ARGS(sc, ino, ret_ip)) > > -DEFINE_SCRUB_INO_ERROR_EVENT(xfs_scrub_ino_error); > -DEFINE_SCRUB_INO_ERROR_EVENT(xfs_scrub_ino_preen); > -DEFINE_SCRUB_INO_ERROR_EVENT(xfs_scrub_ino_warning); > +DEFINE_SCRUB_INO_ERROR_EVENT(xchk_ino_error); > +DEFINE_SCRUB_INO_ERROR_EVENT(xchk_ino_preen); > +DEFINE_SCRUB_INO_ERROR_EVENT(xchk_ino_warning); > > -DECLARE_EVENT_CLASS(xfs_scrub_fblock_error_class, > +DECLARE_EVENT_CLASS(xchk_fblock_error_class, > TP_PROTO(struct xfs_scrub_context *sc, int whichfork, > xfs_fileoff_t offset, void *ret_ip), > TP_ARGS(sc, whichfork, offset, ret_ip), > @@ -223,15 +223,15 @@ DECLARE_EVENT_CLASS(xfs_scrub_fblock_error_class, > ); > > #define DEFINE_SCRUB_FBLOCK_ERROR_EVENT(name) \ > -DEFINE_EVENT(xfs_scrub_fblock_error_class, name, \ > +DEFINE_EVENT(xchk_fblock_error_class, name, \ > TP_PROTO(struct xfs_scrub_context *sc, int whichfork, \ > xfs_fileoff_t offset, void *ret_ip), \ > TP_ARGS(sc, whichfork, offset, ret_ip)) > > -DEFINE_SCRUB_FBLOCK_ERROR_EVENT(xfs_scrub_fblock_error); > -DEFINE_SCRUB_FBLOCK_ERROR_EVENT(xfs_scrub_fblock_warning); > +DEFINE_SCRUB_FBLOCK_ERROR_EVENT(xchk_fblock_error); > +DEFINE_SCRUB_FBLOCK_ERROR_EVENT(xchk_fblock_warning); > > -TRACE_EVENT(xfs_scrub_incomplete, > +TRACE_EVENT(xchk_incomplete, > TP_PROTO(struct xfs_scrub_context *sc, void *ret_ip), > TP_ARGS(sc, ret_ip), > TP_STRUCT__entry( > @@ -250,7 +250,7 @@ TRACE_EVENT(xfs_scrub_incomplete, > __entry->ret_ip) > ); > > -TRACE_EVENT(xfs_scrub_btree_op_error, > +TRACE_EVENT(xchk_btree_op_error, > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > int level, int error, void *ret_ip), > TP_ARGS(sc, cur, level, error, ret_ip), > @@ -266,7 +266,7 @@ TRACE_EVENT(xfs_scrub_btree_op_error, > __field(void *, ret_ip) > ), > TP_fast_assign( > - xfs_fsblock_t fsbno = xfs_scrub_btree_cur_fsbno(cur, level); > + xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); > > __entry->dev = sc->mp->m_super->s_dev; > __entry->type = sc->sm->sm_type; > @@ -290,7 +290,7 @@ TRACE_EVENT(xfs_scrub_btree_op_error, > __entry->ret_ip) > ); > > -TRACE_EVENT(xfs_scrub_ifork_btree_op_error, > +TRACE_EVENT(xchk_ifork_btree_op_error, > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > int level, int error, void *ret_ip), > TP_ARGS(sc, cur, level, error, ret_ip), > @@ -308,7 +308,7 @@ TRACE_EVENT(xfs_scrub_ifork_btree_op_error, > __field(void *, ret_ip) > ), > TP_fast_assign( > - xfs_fsblock_t fsbno = xfs_scrub_btree_cur_fsbno(cur, level); > + xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); > __entry->dev = sc->mp->m_super->s_dev; > __entry->ino = sc->ip->i_ino; > __entry->whichfork = cur->bc_private.b.whichfork; > @@ -335,7 +335,7 @@ TRACE_EVENT(xfs_scrub_ifork_btree_op_error, > __entry->ret_ip) > ); > > -TRACE_EVENT(xfs_scrub_btree_error, > +TRACE_EVENT(xchk_btree_error, > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > int level, void *ret_ip), > TP_ARGS(sc, cur, level, ret_ip), > @@ -350,7 +350,7 @@ TRACE_EVENT(xfs_scrub_btree_error, > __field(void *, ret_ip) > ), > TP_fast_assign( > - xfs_fsblock_t fsbno = xfs_scrub_btree_cur_fsbno(cur, level); > + xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); > __entry->dev = sc->mp->m_super->s_dev; > __entry->type = sc->sm->sm_type; > __entry->btnum = cur->bc_btnum; > @@ -371,7 +371,7 @@ TRACE_EVENT(xfs_scrub_btree_error, > __entry->ret_ip) > ); > > -TRACE_EVENT(xfs_scrub_ifork_btree_error, > +TRACE_EVENT(xchk_ifork_btree_error, > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > int level, void *ret_ip), > TP_ARGS(sc, cur, level, ret_ip), > @@ -388,7 +388,7 @@ TRACE_EVENT(xfs_scrub_ifork_btree_error, > __field(void *, ret_ip) > ), > TP_fast_assign( > - xfs_fsblock_t fsbno = xfs_scrub_btree_cur_fsbno(cur, level); > + xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); > __entry->dev = sc->mp->m_super->s_dev; > __entry->ino = sc->ip->i_ino; > __entry->whichfork = cur->bc_private.b.whichfork; > @@ -413,7 +413,7 @@ TRACE_EVENT(xfs_scrub_ifork_btree_error, > __entry->ret_ip) > ); > > -DECLARE_EVENT_CLASS(xfs_scrub_sbtree_class, > +DECLARE_EVENT_CLASS(xchk_sbtree_class, > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, > int level), > TP_ARGS(sc, cur, level), > @@ -428,7 +428,7 @@ DECLARE_EVENT_CLASS(xfs_scrub_sbtree_class, > __field(int, ptr) > ), > TP_fast_assign( > - xfs_fsblock_t fsbno = xfs_scrub_btree_cur_fsbno(cur, level); > + xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); > > __entry->dev = sc->mp->m_super->s_dev; > __entry->type = sc->sm->sm_type; > @@ -450,15 +450,15 @@ DECLARE_EVENT_CLASS(xfs_scrub_sbtree_class, > __entry->ptr) > ) > #define DEFINE_SCRUB_SBTREE_EVENT(name) \ > -DEFINE_EVENT(xfs_scrub_sbtree_class, name, \ > +DEFINE_EVENT(xchk_sbtree_class, name, \ > TP_PROTO(struct xfs_scrub_context *sc, struct xfs_btree_cur *cur, \ > int level), \ > TP_ARGS(sc, cur, level)) > > -DEFINE_SCRUB_SBTREE_EVENT(xfs_scrub_btree_rec); > -DEFINE_SCRUB_SBTREE_EVENT(xfs_scrub_btree_key); > +DEFINE_SCRUB_SBTREE_EVENT(xchk_btree_rec); > +DEFINE_SCRUB_SBTREE_EVENT(xchk_btree_key); > > -TRACE_EVENT(xfs_scrub_xref_error, > +TRACE_EVENT(xchk_xref_error, > TP_PROTO(struct xfs_scrub_context *sc, int error, void *ret_ip), > TP_ARGS(sc, error, ret_ip), > TP_STRUCT__entry( > > -- > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html