mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 13:05:10 +07:00
72f76f7364
Fix smatch complaints about uninitialized return codes. Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> Reviewed-by: Dave Chinner <dchinner@redhat.com>
817 lines
23 KiB
C
817 lines
23 KiB
C
/*
|
|
* Copyright (C) 2017 Oracle. All Rights Reserved.
|
|
*
|
|
* Author: Darrick J. Wong <darrick.wong@oracle.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version 2
|
|
* of the License, or (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it would be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*/
|
|
#include "xfs.h"
|
|
#include "xfs_fs.h"
|
|
#include "xfs_shared.h"
|
|
#include "xfs_format.h"
|
|
#include "xfs_trans_resv.h"
|
|
#include "xfs_mount.h"
|
|
#include "xfs_defer.h"
|
|
#include "xfs_btree.h"
|
|
#include "xfs_bit.h"
|
|
#include "xfs_log_format.h"
|
|
#include "xfs_trans.h"
|
|
#include "xfs_sb.h"
|
|
#include "xfs_inode.h"
|
|
#include "xfs_icache.h"
|
|
#include "xfs_itable.h"
|
|
#include "xfs_da_format.h"
|
|
#include "xfs_da_btree.h"
|
|
#include "xfs_dir2.h"
|
|
#include "xfs_dir2_priv.h"
|
|
#include "xfs_ialloc.h"
|
|
#include "scrub/xfs_scrub.h"
|
|
#include "scrub/scrub.h"
|
|
#include "scrub/common.h"
|
|
#include "scrub/trace.h"
|
|
#include "scrub/dabtree.h"
|
|
|
|
/* Set us up to scrub directories. */
|
|
int
|
|
xfs_scrub_setup_directory(
|
|
struct xfs_scrub_context *sc,
|
|
struct xfs_inode *ip)
|
|
{
|
|
return xfs_scrub_setup_inode_contents(sc, ip, 0);
|
|
}
|
|
|
|
/* Directories */
|
|
|
|
/* Scrub a directory entry. */
|
|
|
|
struct xfs_scrub_dir_ctx {
|
|
/* VFS fill-directory iterator */
|
|
struct dir_context dir_iter;
|
|
|
|
struct xfs_scrub_context *sc;
|
|
};
|
|
|
|
/* Check that an inode's mode matches a given DT_ type. */
|
|
STATIC int
|
|
xfs_scrub_dir_check_ftype(
|
|
struct xfs_scrub_dir_ctx *sdc,
|
|
xfs_fileoff_t offset,
|
|
xfs_ino_t inum,
|
|
int dtype)
|
|
{
|
|
struct xfs_mount *mp = sdc->sc->mp;
|
|
struct xfs_inode *ip;
|
|
int ino_dtype;
|
|
int error = 0;
|
|
|
|
if (!xfs_sb_version_hasftype(&mp->m_sb)) {
|
|
if (dtype != DT_UNKNOWN && dtype != DT_DIR)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK,
|
|
offset);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Grab the inode pointed to by the dirent. We release the
|
|
* inode before we cancel the scrub transaction. Since we're
|
|
* don't know a priori that releasing the inode won't trigger
|
|
* eofblocks cleanup (which allocates what would be a nested
|
|
* transaction), we can't use DONTCACHE here because DONTCACHE
|
|
* inodes can trigger immediate inactive cleanup of the inode.
|
|
*/
|
|
error = xfs_iget(mp, sdc->sc->tp, inum, 0, 0, &ip);
|
|
if (!xfs_scrub_fblock_process_error(sdc->sc, XFS_DATA_FORK, offset,
|
|
&error))
|
|
goto out;
|
|
|
|
/* Convert mode to the DT_* values that dir_emit uses. */
|
|
ino_dtype = xfs_dir3_get_dtype(mp,
|
|
xfs_mode_to_ftype(VFS_I(ip)->i_mode));
|
|
if (ino_dtype != dtype)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset);
|
|
iput(VFS_I(ip));
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* Scrub a single directory entry.
|
|
*
|
|
* We use the VFS directory iterator (i.e. readdir) to call this
|
|
* function for every directory entry in a directory. Once we're here,
|
|
* we check the inode number to make sure it's sane, then we check that
|
|
* we can look up this filename. Finally, we check the ftype.
|
|
*/
|
|
STATIC int
|
|
xfs_scrub_dir_actor(
|
|
struct dir_context *dir_iter,
|
|
const char *name,
|
|
int namelen,
|
|
loff_t pos,
|
|
u64 ino,
|
|
unsigned type)
|
|
{
|
|
struct xfs_mount *mp;
|
|
struct xfs_inode *ip;
|
|
struct xfs_scrub_dir_ctx *sdc;
|
|
struct xfs_name xname;
|
|
xfs_ino_t lookup_ino;
|
|
xfs_dablk_t offset;
|
|
int error = 0;
|
|
|
|
sdc = container_of(dir_iter, struct xfs_scrub_dir_ctx, dir_iter);
|
|
ip = sdc->sc->ip;
|
|
mp = ip->i_mount;
|
|
offset = xfs_dir2_db_to_da(mp->m_dir_geo,
|
|
xfs_dir2_dataptr_to_db(mp->m_dir_geo, pos));
|
|
|
|
/* Does this inode number make sense? */
|
|
if (!xfs_verify_dir_ino(mp, ino)) {
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset);
|
|
goto out;
|
|
}
|
|
|
|
if (!strncmp(".", name, namelen)) {
|
|
/* If this is "." then check that the inum matches the dir. */
|
|
if (xfs_sb_version_hasftype(&mp->m_sb) && type != DT_DIR)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK,
|
|
offset);
|
|
if (ino != ip->i_ino)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK,
|
|
offset);
|
|
} else if (!strncmp("..", name, namelen)) {
|
|
/*
|
|
* If this is ".." in the root inode, check that the inum
|
|
* matches this dir.
|
|
*/
|
|
if (xfs_sb_version_hasftype(&mp->m_sb) && type != DT_DIR)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK,
|
|
offset);
|
|
if (ip->i_ino == mp->m_sb.sb_rootino && ino != ip->i_ino)
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK,
|
|
offset);
|
|
}
|
|
|
|
/* Verify that we can look up this name by hash. */
|
|
xname.name = name;
|
|
xname.len = namelen;
|
|
xname.type = XFS_DIR3_FT_UNKNOWN;
|
|
|
|
error = xfs_dir_lookup(sdc->sc->tp, ip, &xname, &lookup_ino, NULL);
|
|
if (!xfs_scrub_fblock_process_error(sdc->sc, XFS_DATA_FORK, offset,
|
|
&error))
|
|
goto fail_xref;
|
|
if (lookup_ino != ino) {
|
|
xfs_scrub_fblock_set_corrupt(sdc->sc, XFS_DATA_FORK, offset);
|
|
goto out;
|
|
}
|
|
|
|
/* Verify the file type. This function absorbs error codes. */
|
|
error = xfs_scrub_dir_check_ftype(sdc, offset, lookup_ino, type);
|
|
if (error)
|
|
goto out;
|
|
out:
|
|
return error;
|
|
fail_xref:
|
|
return error;
|
|
}
|
|
|
|
/* Scrub a directory btree record. */
|
|
STATIC int
|
|
xfs_scrub_dir_rec(
|
|
struct xfs_scrub_da_btree *ds,
|
|
int level,
|
|
void *rec)
|
|
{
|
|
struct xfs_mount *mp = ds->state->mp;
|
|
struct xfs_dir2_leaf_entry *ent = rec;
|
|
struct xfs_inode *dp = ds->dargs.dp;
|
|
struct xfs_dir2_data_entry *dent;
|
|
struct xfs_buf *bp;
|
|
xfs_ino_t ino;
|
|
xfs_dablk_t rec_bno;
|
|
xfs_dir2_db_t db;
|
|
xfs_dir2_data_aoff_t off;
|
|
xfs_dir2_dataptr_t ptr;
|
|
xfs_dahash_t calc_hash;
|
|
xfs_dahash_t hash;
|
|
unsigned int tag;
|
|
int error;
|
|
|
|
/* Check the hash of the entry. */
|
|
error = xfs_scrub_da_btree_hash(ds, level, &ent->hashval);
|
|
if (error)
|
|
goto out;
|
|
|
|
/* Valid hash pointer? */
|
|
ptr = be32_to_cpu(ent->address);
|
|
if (ptr == 0)
|
|
return 0;
|
|
|
|
/* Find the directory entry's location. */
|
|
db = xfs_dir2_dataptr_to_db(mp->m_dir_geo, ptr);
|
|
off = xfs_dir2_dataptr_to_off(mp->m_dir_geo, ptr);
|
|
rec_bno = xfs_dir2_db_to_da(mp->m_dir_geo, db);
|
|
|
|
if (rec_bno >= mp->m_dir_geo->leafblk) {
|
|
xfs_scrub_da_set_corrupt(ds, level);
|
|
goto out;
|
|
}
|
|
error = xfs_dir3_data_read(ds->dargs.trans, dp, rec_bno, -2, &bp);
|
|
if (!xfs_scrub_fblock_process_error(ds->sc, XFS_DATA_FORK, rec_bno,
|
|
&error))
|
|
goto out;
|
|
if (!bp) {
|
|
xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno);
|
|
goto out;
|
|
}
|
|
|
|
/* Retrieve the entry, sanity check it, and compare hashes. */
|
|
dent = (struct xfs_dir2_data_entry *)(((char *)bp->b_addr) + off);
|
|
ino = be64_to_cpu(dent->inumber);
|
|
hash = be32_to_cpu(ent->hashval);
|
|
tag = be16_to_cpup(dp->d_ops->data_entry_tag_p(dent));
|
|
if (!xfs_verify_dir_ino(mp, ino) || tag != off)
|
|
xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno);
|
|
if (dent->namelen == 0) {
|
|
xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno);
|
|
goto out_relse;
|
|
}
|
|
calc_hash = xfs_da_hashname(dent->name, dent->namelen);
|
|
if (calc_hash != hash)
|
|
xfs_scrub_fblock_set_corrupt(ds->sc, XFS_DATA_FORK, rec_bno);
|
|
|
|
out_relse:
|
|
xfs_trans_brelse(ds->dargs.trans, bp);
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* Is this unused entry either in the bestfree or smaller than all of
|
|
* them? We've already checked that the bestfrees are sorted longest to
|
|
* shortest, and that there aren't any bogus entries.
|
|
*/
|
|
STATIC void
|
|
xfs_scrub_directory_check_free_entry(
|
|
struct xfs_scrub_context *sc,
|
|
xfs_dablk_t lblk,
|
|
struct xfs_dir2_data_free *bf,
|
|
struct xfs_dir2_data_unused *dup)
|
|
{
|
|
struct xfs_dir2_data_free *dfp;
|
|
unsigned int dup_length;
|
|
|
|
dup_length = be16_to_cpu(dup->length);
|
|
|
|
/* Unused entry is shorter than any of the bestfrees */
|
|
if (dup_length < be16_to_cpu(bf[XFS_DIR2_DATA_FD_COUNT - 1].length))
|
|
return;
|
|
|
|
for (dfp = &bf[XFS_DIR2_DATA_FD_COUNT - 1]; dfp >= bf; dfp--)
|
|
if (dup_length == be16_to_cpu(dfp->length))
|
|
return;
|
|
|
|
/* Unused entry should be in the bestfrees but wasn't found. */
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
}
|
|
|
|
/* Check free space info in a directory data block. */
|
|
STATIC int
|
|
xfs_scrub_directory_data_bestfree(
|
|
struct xfs_scrub_context *sc,
|
|
xfs_dablk_t lblk,
|
|
bool is_block)
|
|
{
|
|
struct xfs_dir2_data_unused *dup;
|
|
struct xfs_dir2_data_free *dfp;
|
|
struct xfs_buf *bp;
|
|
struct xfs_dir2_data_free *bf;
|
|
struct xfs_mount *mp = sc->mp;
|
|
const struct xfs_dir_ops *d_ops;
|
|
char *ptr;
|
|
char *endptr;
|
|
u16 tag;
|
|
unsigned int nr_bestfrees = 0;
|
|
unsigned int nr_frees = 0;
|
|
unsigned int smallest_bestfree;
|
|
int newlen;
|
|
int offset;
|
|
int error;
|
|
|
|
d_ops = sc->ip->d_ops;
|
|
|
|
if (is_block) {
|
|
/* dir block format */
|
|
if (lblk != XFS_B_TO_FSBT(mp, XFS_DIR2_DATA_OFFSET))
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
error = xfs_dir3_block_read(sc->tp, sc->ip, &bp);
|
|
} else {
|
|
/* dir data format */
|
|
error = xfs_dir3_data_read(sc->tp, sc->ip, lblk, -1, &bp);
|
|
}
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error))
|
|
goto out;
|
|
|
|
/* XXX: Check xfs_dir3_data_hdr.pad is zero once we start setting it. */
|
|
|
|
/* Do the bestfrees correspond to actual free space? */
|
|
bf = d_ops->data_bestfree_p(bp->b_addr);
|
|
smallest_bestfree = UINT_MAX;
|
|
for (dfp = &bf[0]; dfp < &bf[XFS_DIR2_DATA_FD_COUNT]; dfp++) {
|
|
offset = be16_to_cpu(dfp->offset);
|
|
if (offset == 0)
|
|
continue;
|
|
if (offset >= mp->m_dir_geo->blksize) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out_buf;
|
|
}
|
|
dup = (struct xfs_dir2_data_unused *)(bp->b_addr + offset);
|
|
tag = be16_to_cpu(*xfs_dir2_data_unused_tag_p(dup));
|
|
|
|
/* bestfree doesn't match the entry it points at? */
|
|
if (dup->freetag != cpu_to_be16(XFS_DIR2_DATA_FREE_TAG) ||
|
|
be16_to_cpu(dup->length) != be16_to_cpu(dfp->length) ||
|
|
tag != ((char *)dup - (char *)bp->b_addr)) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out_buf;
|
|
}
|
|
|
|
/* bestfree records should be ordered largest to smallest */
|
|
if (smallest_bestfree < be16_to_cpu(dfp->length)) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out_buf;
|
|
}
|
|
|
|
smallest_bestfree = be16_to_cpu(dfp->length);
|
|
nr_bestfrees++;
|
|
}
|
|
|
|
/* Make sure the bestfrees are actually the best free spaces. */
|
|
ptr = (char *)d_ops->data_entry_p(bp->b_addr);
|
|
if (is_block) {
|
|
struct xfs_dir2_block_tail *btp;
|
|
|
|
btp = xfs_dir2_block_tail_p(mp->m_dir_geo, bp->b_addr);
|
|
endptr = (char *)xfs_dir2_block_leaf_p(btp);
|
|
} else
|
|
endptr = (char *)bp->b_addr + BBTOB(bp->b_length);
|
|
|
|
/* Iterate the entries, stopping when we hit or go past the end. */
|
|
while (ptr < endptr) {
|
|
dup = (struct xfs_dir2_data_unused *)ptr;
|
|
/* Skip real entries */
|
|
if (dup->freetag != cpu_to_be16(XFS_DIR2_DATA_FREE_TAG)) {
|
|
struct xfs_dir2_data_entry *dep;
|
|
|
|
dep = (struct xfs_dir2_data_entry *)ptr;
|
|
newlen = d_ops->data_entsize(dep->namelen);
|
|
if (newlen <= 0) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK,
|
|
lblk);
|
|
goto out_buf;
|
|
}
|
|
ptr += newlen;
|
|
continue;
|
|
}
|
|
|
|
/* Spot check this free entry */
|
|
tag = be16_to_cpu(*xfs_dir2_data_unused_tag_p(dup));
|
|
if (tag != ((char *)dup - (char *)bp->b_addr))
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
|
|
/*
|
|
* Either this entry is a bestfree or it's smaller than
|
|
* any of the bestfrees.
|
|
*/
|
|
xfs_scrub_directory_check_free_entry(sc, lblk, bf, dup);
|
|
|
|
/* Move on. */
|
|
newlen = be16_to_cpu(dup->length);
|
|
if (newlen <= 0) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out_buf;
|
|
}
|
|
ptr += newlen;
|
|
if (ptr <= endptr)
|
|
nr_frees++;
|
|
}
|
|
|
|
/* We're required to fill all the space. */
|
|
if (ptr != endptr)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
|
|
/* Did we see at least as many free slots as there are bestfrees? */
|
|
if (nr_frees < nr_bestfrees)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
out_buf:
|
|
xfs_trans_brelse(sc->tp, bp);
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* Does the free space length in the free space index block ($len) match
|
|
* the longest length in the directory data block's bestfree array?
|
|
* Assume that we've already checked that the data block's bestfree
|
|
* array is in order.
|
|
*/
|
|
STATIC void
|
|
xfs_scrub_directory_check_freesp(
|
|
struct xfs_scrub_context *sc,
|
|
xfs_dablk_t lblk,
|
|
struct xfs_buf *dbp,
|
|
unsigned int len)
|
|
{
|
|
struct xfs_dir2_data_free *dfp;
|
|
|
|
dfp = sc->ip->d_ops->data_bestfree_p(dbp->b_addr);
|
|
|
|
if (len != be16_to_cpu(dfp->length))
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
|
|
if (len > 0 && be16_to_cpu(dfp->offset) == 0)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
}
|
|
|
|
/* Check free space info in a directory leaf1 block. */
|
|
STATIC int
|
|
xfs_scrub_directory_leaf1_bestfree(
|
|
struct xfs_scrub_context *sc,
|
|
struct xfs_da_args *args,
|
|
xfs_dablk_t lblk)
|
|
{
|
|
struct xfs_dir3_icleaf_hdr leafhdr;
|
|
struct xfs_dir2_leaf_entry *ents;
|
|
struct xfs_dir2_leaf_tail *ltp;
|
|
struct xfs_dir2_leaf *leaf;
|
|
struct xfs_buf *dbp;
|
|
struct xfs_buf *bp;
|
|
const struct xfs_dir_ops *d_ops = sc->ip->d_ops;
|
|
struct xfs_da_geometry *geo = sc->mp->m_dir_geo;
|
|
__be16 *bestp;
|
|
__u16 best;
|
|
__u32 hash;
|
|
__u32 lasthash = 0;
|
|
__u32 bestcount;
|
|
unsigned int stale = 0;
|
|
int i;
|
|
int error;
|
|
|
|
/* Read the free space block. */
|
|
error = xfs_dir3_leaf_read(sc->tp, sc->ip, lblk, -1, &bp);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error))
|
|
goto out;
|
|
|
|
leaf = bp->b_addr;
|
|
d_ops->leaf_hdr_from_disk(&leafhdr, leaf);
|
|
ents = d_ops->leaf_ents_p(leaf);
|
|
ltp = xfs_dir2_leaf_tail_p(geo, leaf);
|
|
bestcount = be32_to_cpu(ltp->bestcount);
|
|
bestp = xfs_dir2_leaf_bests_p(ltp);
|
|
|
|
if (xfs_sb_version_hascrc(&sc->mp->m_sb)) {
|
|
struct xfs_dir3_leaf_hdr *hdr3 = bp->b_addr;
|
|
|
|
if (hdr3->pad != cpu_to_be32(0))
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
}
|
|
|
|
/*
|
|
* There should be as many bestfree slots as there are dir data
|
|
* blocks that can fit under i_size.
|
|
*/
|
|
if (bestcount != xfs_dir2_byte_to_db(geo, sc->ip->i_d.di_size)) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
|
|
/* Is the leaf count even remotely sane? */
|
|
if (leafhdr.count > d_ops->leaf_max_ents(geo)) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
|
|
/* Leaves and bests don't overlap in leaf format. */
|
|
if ((char *)&ents[leafhdr.count] > (char *)bestp) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
|
|
/* Check hash value order, count stale entries. */
|
|
for (i = 0; i < leafhdr.count; i++) {
|
|
hash = be32_to_cpu(ents[i].hashval);
|
|
if (i > 0 && lasthash > hash)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
lasthash = hash;
|
|
if (ents[i].address == cpu_to_be32(XFS_DIR2_NULL_DATAPTR))
|
|
stale++;
|
|
}
|
|
if (leafhdr.stale != stale)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
|
|
/* Check all the bestfree entries. */
|
|
for (i = 0; i < bestcount; i++, bestp++) {
|
|
best = be16_to_cpu(*bestp);
|
|
if (best == NULLDATAOFF)
|
|
continue;
|
|
error = xfs_dir3_data_read(sc->tp, sc->ip,
|
|
i * args->geo->fsbcount, -1, &dbp);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk,
|
|
&error))
|
|
continue;
|
|
xfs_scrub_directory_check_freesp(sc, lblk, dbp, best);
|
|
xfs_trans_brelse(sc->tp, dbp);
|
|
}
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/* Check free space info in a directory freespace block. */
|
|
STATIC int
|
|
xfs_scrub_directory_free_bestfree(
|
|
struct xfs_scrub_context *sc,
|
|
struct xfs_da_args *args,
|
|
xfs_dablk_t lblk)
|
|
{
|
|
struct xfs_dir3_icfree_hdr freehdr;
|
|
struct xfs_buf *dbp;
|
|
struct xfs_buf *bp;
|
|
__be16 *bestp;
|
|
__u16 best;
|
|
unsigned int stale = 0;
|
|
int i;
|
|
int error;
|
|
|
|
/* Read the free space block */
|
|
error = xfs_dir2_free_read(sc->tp, sc->ip, lblk, &bp);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error))
|
|
goto out;
|
|
|
|
if (xfs_sb_version_hascrc(&sc->mp->m_sb)) {
|
|
struct xfs_dir3_free_hdr *hdr3 = bp->b_addr;
|
|
|
|
if (hdr3->pad != cpu_to_be32(0))
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
}
|
|
|
|
/* Check all the entries. */
|
|
sc->ip->d_ops->free_hdr_from_disk(&freehdr, bp->b_addr);
|
|
bestp = sc->ip->d_ops->free_bests_p(bp->b_addr);
|
|
for (i = 0; i < freehdr.nvalid; i++, bestp++) {
|
|
best = be16_to_cpu(*bestp);
|
|
if (best == NULLDATAOFF) {
|
|
stale++;
|
|
continue;
|
|
}
|
|
error = xfs_dir3_data_read(sc->tp, sc->ip,
|
|
(freehdr.firstdb + i) * args->geo->fsbcount,
|
|
-1, &dbp);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk,
|
|
&error))
|
|
continue;
|
|
xfs_scrub_directory_check_freesp(sc, lblk, dbp, best);
|
|
xfs_trans_brelse(sc->tp, dbp);
|
|
}
|
|
|
|
if (freehdr.nused + stale != freehdr.nvalid)
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/* Check free space information in directories. */
|
|
STATIC int
|
|
xfs_scrub_directory_blocks(
|
|
struct xfs_scrub_context *sc)
|
|
{
|
|
struct xfs_bmbt_irec got;
|
|
struct xfs_da_args args;
|
|
struct xfs_ifork *ifp;
|
|
struct xfs_mount *mp = sc->mp;
|
|
xfs_fileoff_t leaf_lblk;
|
|
xfs_fileoff_t free_lblk;
|
|
xfs_fileoff_t lblk;
|
|
struct xfs_iext_cursor icur;
|
|
xfs_dablk_t dabno;
|
|
bool found;
|
|
int is_block = 0;
|
|
int error;
|
|
|
|
/* Ignore local format directories. */
|
|
if (sc->ip->i_d.di_format != XFS_DINODE_FMT_EXTENTS &&
|
|
sc->ip->i_d.di_format != XFS_DINODE_FMT_BTREE)
|
|
return 0;
|
|
|
|
ifp = XFS_IFORK_PTR(sc->ip, XFS_DATA_FORK);
|
|
lblk = XFS_B_TO_FSB(mp, XFS_DIR2_DATA_OFFSET);
|
|
leaf_lblk = XFS_B_TO_FSB(mp, XFS_DIR2_LEAF_OFFSET);
|
|
free_lblk = XFS_B_TO_FSB(mp, XFS_DIR2_FREE_OFFSET);
|
|
|
|
/* Is this a block dir? */
|
|
args.dp = sc->ip;
|
|
args.geo = mp->m_dir_geo;
|
|
args.trans = sc->tp;
|
|
error = xfs_dir2_isblock(&args, &is_block);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error))
|
|
goto out;
|
|
|
|
/* Iterate all the data extents in the directory... */
|
|
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
|
|
while (found) {
|
|
/* Block directories only have a single block at offset 0. */
|
|
if (is_block &&
|
|
(got.br_startoff > 0 ||
|
|
got.br_blockcount != args.geo->fsbcount)) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK,
|
|
got.br_startoff);
|
|
break;
|
|
}
|
|
|
|
/* No more data blocks... */
|
|
if (got.br_startoff >= leaf_lblk)
|
|
break;
|
|
|
|
/*
|
|
* Check each data block's bestfree data.
|
|
*
|
|
* Iterate all the fsbcount-aligned block offsets in
|
|
* this directory. The directory block reading code is
|
|
* smart enough to do its own bmap lookups to handle
|
|
* discontiguous directory blocks. When we're done
|
|
* with the extent record, re-query the bmap at the
|
|
* next fsbcount-aligned offset to avoid redundant
|
|
* block checks.
|
|
*/
|
|
for (lblk = roundup((xfs_dablk_t)got.br_startoff,
|
|
args.geo->fsbcount);
|
|
lblk < got.br_startoff + got.br_blockcount;
|
|
lblk += args.geo->fsbcount) {
|
|
error = xfs_scrub_directory_data_bestfree(sc, lblk,
|
|
is_block);
|
|
if (error)
|
|
goto out;
|
|
}
|
|
dabno = got.br_startoff + got.br_blockcount;
|
|
lblk = roundup(dabno, args.geo->fsbcount);
|
|
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
|
|
}
|
|
|
|
if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
|
|
goto out;
|
|
|
|
/* Look for a leaf1 block, which has free info. */
|
|
if (xfs_iext_lookup_extent(sc->ip, ifp, leaf_lblk, &icur, &got) &&
|
|
got.br_startoff == leaf_lblk &&
|
|
got.br_blockcount == args.geo->fsbcount &&
|
|
!xfs_iext_next_extent(ifp, &icur, &got)) {
|
|
if (is_block) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
error = xfs_scrub_directory_leaf1_bestfree(sc, &args,
|
|
leaf_lblk);
|
|
if (error)
|
|
goto out;
|
|
}
|
|
|
|
if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
|
|
goto out;
|
|
|
|
/* Scan for free blocks */
|
|
lblk = free_lblk;
|
|
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
|
|
while (found) {
|
|
/*
|
|
* Dirs can't have blocks mapped above 2^32.
|
|
* Single-block dirs shouldn't even be here.
|
|
*/
|
|
lblk = got.br_startoff;
|
|
if (lblk & ~0xFFFFFFFFULL) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
if (is_block) {
|
|
xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Check each dir free block's bestfree data.
|
|
*
|
|
* Iterate all the fsbcount-aligned block offsets in
|
|
* this directory. The directory block reading code is
|
|
* smart enough to do its own bmap lookups to handle
|
|
* discontiguous directory blocks. When we're done
|
|
* with the extent record, re-query the bmap at the
|
|
* next fsbcount-aligned offset to avoid redundant
|
|
* block checks.
|
|
*/
|
|
for (lblk = roundup((xfs_dablk_t)got.br_startoff,
|
|
args.geo->fsbcount);
|
|
lblk < got.br_startoff + got.br_blockcount;
|
|
lblk += args.geo->fsbcount) {
|
|
error = xfs_scrub_directory_free_bestfree(sc, &args,
|
|
lblk);
|
|
if (error)
|
|
goto out;
|
|
}
|
|
dabno = got.br_startoff + got.br_blockcount;
|
|
lblk = roundup(dabno, args.geo->fsbcount);
|
|
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
|
|
}
|
|
out:
|
|
return error;
|
|
}
|
|
|
|
/* Scrub a whole directory. */
|
|
int
|
|
xfs_scrub_directory(
|
|
struct xfs_scrub_context *sc)
|
|
{
|
|
struct xfs_scrub_dir_ctx sdc = {
|
|
.dir_iter.actor = xfs_scrub_dir_actor,
|
|
.dir_iter.pos = 0,
|
|
.sc = sc,
|
|
};
|
|
size_t bufsize;
|
|
loff_t oldpos;
|
|
int error = 0;
|
|
|
|
if (!S_ISDIR(VFS_I(sc->ip)->i_mode))
|
|
return -ENOENT;
|
|
|
|
/* Plausible size? */
|
|
if (sc->ip->i_d.di_size < xfs_dir2_sf_hdr_size(0)) {
|
|
xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino, NULL);
|
|
goto out;
|
|
}
|
|
|
|
/* Check directory tree structure */
|
|
error = xfs_scrub_da_btree(sc, XFS_DATA_FORK, xfs_scrub_dir_rec, NULL);
|
|
if (error)
|
|
return error;
|
|
|
|
if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
|
|
return error;
|
|
|
|
/* Check the freespace. */
|
|
error = xfs_scrub_directory_blocks(sc);
|
|
if (error)
|
|
return error;
|
|
|
|
if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
|
|
return error;
|
|
|
|
/*
|
|
* Check that every dirent we see can also be looked up by hash.
|
|
* Userspace usually asks for a 32k buffer, so we will too.
|
|
*/
|
|
bufsize = (size_t)min_t(loff_t, XFS_READDIR_BUFSIZE,
|
|
sc->ip->i_d.di_size);
|
|
|
|
/*
|
|
* Look up every name in this directory by hash.
|
|
*
|
|
* Use the xfs_readdir function to call xfs_scrub_dir_actor on
|
|
* every directory entry in this directory. In _actor, we check
|
|
* the name, inode number, and ftype (if applicable) of the
|
|
* entry. xfs_readdir uses the VFS filldir functions to provide
|
|
* iteration context.
|
|
*
|
|
* The VFS grabs a read or write lock via i_rwsem before it reads
|
|
* or writes to a directory. If we've gotten this far we've
|
|
* already obtained IOLOCK_EXCL, which (since 4.10) is the same as
|
|
* getting a write lock on i_rwsem. Therefore, it is safe for us
|
|
* to drop the ILOCK here in order to reuse the _readdir and
|
|
* _dir_lookup routines, which do their own ILOCK locking.
|
|
*/
|
|
oldpos = 0;
|
|
sc->ilock_flags &= ~XFS_ILOCK_EXCL;
|
|
xfs_iunlock(sc->ip, XFS_ILOCK_EXCL);
|
|
while (true) {
|
|
error = xfs_readdir(sc->tp, sc->ip, &sdc.dir_iter, bufsize);
|
|
if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, 0,
|
|
&error))
|
|
goto out;
|
|
if (oldpos == sdc.dir_iter.pos)
|
|
break;
|
|
oldpos = sdc.dir_iter.pos;
|
|
}
|
|
|
|
out:
|
|
return error;
|
|
}
|