]> asedeno.scripts.mit.edu Git - linux.git/blobdiff - fs/xfs/scrub/ialloc.c
xfs: hoist inode cluster checks out of loop
[linux.git] / fs / xfs / scrub / ialloc.c
index 882dc56c5c21e5c0f8d02ea705669eced2132664..0ce793d92995898086b8420b8e412834c99db674 100644 (file)
@@ -47,6 +47,12 @@ xchk_setup_ag_iallocbt(
 struct xchk_iallocbt {
        /* Number of inodes we see while scanning inobt. */
        unsigned long long      inodes;
+
+       /* Expected next startino, for big block filesystems. */
+       xfs_agino_t             next_startino;
+
+       /* Expected end of the current inode cluster. */
+       xfs_agino_t             next_cluster_ino;
 };
 
 /*
@@ -182,19 +188,19 @@ xchk_iallocbt_check_cluster_freemask(
        return 0;
 }
 
-/* Make sure the free mask is consistent with what the inodes think. */
+/* Check an inode cluster. */
 STATIC int
-xchk_iallocbt_check_freemask(
+xchk_iallocbt_check_cluster(
        struct xchk_btree               *bs,
-       struct xfs_inobt_rec_incore     *irec)
+       struct xfs_inobt_rec_incore     *irec,
+       xfs_agino_t                     agino)
 {
        struct xfs_imap                 imap;
        struct xfs_mount                *mp = bs->cur->bc_mp;
        struct xfs_dinode               *dip;
        struct xfs_buf                  *bp;
        xfs_ino_t                       fsino;
-       xfs_agino_t                     nr_inodes;
-       xfs_agino_t                     agino;
+       unsigned int                    nr_inodes;
        xfs_agino_t                     chunkino;
        xfs_agino_t                     clusterino;
        xfs_agblock_t                   agbno;
@@ -203,65 +209,157 @@ xchk_iallocbt_check_freemask(
        int                             error = 0;
 
        /* Make sure the freemask matches the inode records. */
-       nr_inodes = mp->m_inodes_per_cluster;
+       nr_inodes = min_t(unsigned int, XFS_INODES_PER_CHUNK,
+                       mp->m_inodes_per_cluster);
+
+       fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino);
+       chunkino = agino - irec->ir_startino;
+       agbno = XFS_AGINO_TO_AGBNO(mp, agino);
+
+       /* Compute the holemask mask for this cluster. */
+       for (clusterino = 0, holemask = 0; clusterino < nr_inodes;
+            clusterino += XFS_INODES_PER_HOLEMASK_BIT)
+               holemask |= XFS_INOBT_MASK((chunkino + clusterino) /
+                               XFS_INODES_PER_HOLEMASK_BIT);
+
+       /* The whole cluster must be a hole or not a hole. */
+       ir_holemask = (irec->ir_holemask & holemask);
+       if (ir_holemask != holemask && ir_holemask != 0) {
+               xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
+               return 0;
+       }
+
+       /* If any part of this is a hole, skip it. */
+       if (ir_holemask) {
+               xchk_xref_is_not_owned_by(bs->sc, agbno,
+                               mp->m_blocks_per_cluster,
+                               &XFS_RMAP_OINFO_INODES);
+               return 0;
+       }
+
+       xchk_xref_is_owned_by(bs->sc, agbno, mp->m_blocks_per_cluster,
+                       &XFS_RMAP_OINFO_INODES);
+
+       /* Grab the inode cluster buffer. */
+       imap.im_blkno = XFS_AGB_TO_DADDR(mp, bs->cur->bc_private.a.agno, agbno);
+       imap.im_len = XFS_FSB_TO_BB(mp, mp->m_blocks_per_cluster);
+       imap.im_boffset = 0;
+
+       error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap, &dip, &bp, 0, 0);
+       if (!xchk_btree_xref_process_error(bs->sc, bs->cur, 0, &error))
+               return 0;
+
+       /* Which inodes are free? */
+       for (clusterino = 0; clusterino < nr_inodes; clusterino++) {
+               error = xchk_iallocbt_check_cluster_freemask(bs, fsino,
+                               chunkino, clusterino, irec, bp);
+               if (error)
+                       break;
+       }
+
+       xfs_trans_brelse(bs->cur->bc_tp, bp);
+       return error;
+}
+
+/* Make sure the free mask is consistent with what the inodes think. */
+STATIC int
+xchk_iallocbt_check_freemask(
+       struct xchk_btree               *bs,
+       struct xfs_inobt_rec_incore     *irec)
+{
+       struct xfs_mount                *mp = bs->cur->bc_mp;
+       xfs_agino_t                     agino;
+       int                             error = 0;
 
        for (agino = irec->ir_startino;
             agino < irec->ir_startino + XFS_INODES_PER_CHUNK;
             agino += mp->m_inodes_per_cluster) {
-               fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino);
-               chunkino = agino - irec->ir_startino;
-               agbno = XFS_AGINO_TO_AGBNO(mp, agino);
-
-               /* Compute the holemask mask for this cluster. */
-               for (clusterino = 0, holemask = 0; clusterino < nr_inodes;
-                    clusterino += XFS_INODES_PER_HOLEMASK_BIT)
-                       holemask |= XFS_INOBT_MASK((chunkino + clusterino) /
-                                       XFS_INODES_PER_HOLEMASK_BIT);
-
-               /* The whole cluster must be a hole or not a hole. */
-               ir_holemask = (irec->ir_holemask & holemask);
-               if (ir_holemask != holemask && ir_holemask != 0) {
+               error = xchk_iallocbt_check_cluster(bs, irec, agino);
+               if (error)
+                       break;
+       }
+
+       return error;
+}
+
+/*
+ * Make sure this inode btree record is aligned properly.  Because a fs block
+ * contains multiple inodes, we check that the inobt record is aligned to the
+ * correct inode, not just the correct block on disk.  This results in a finer
+ * grained corruption check.
+ */
+STATIC void
+xchk_iallocbt_rec_alignment(
+       struct xchk_btree               *bs,
+       struct xfs_inobt_rec_incore     *irec)
+{
+       struct xfs_mount                *mp = bs->sc->mp;
+       struct xchk_iallocbt            *iabt = bs->private;
+
+       /*
+        * finobt records have different positioning requirements than inobt
+        * records: each finobt record must have a corresponding inobt record.
+        * That is checked in the xref function, so for now we only catch the
+        * obvious case where the record isn't at all aligned properly.
+        *
+        * Note that if a fs block contains more than a single chunk of inodes,
+        * we will have finobt records only for those chunks containing free
+        * inodes, and therefore expect chunk alignment of finobt records.
+        * Otherwise, we expect that the finobt record is aligned to the
+        * cluster alignment as told by the superblock.
+        */
+       if (bs->cur->bc_btnum == XFS_BTNUM_FINO) {
+               unsigned int    imask;
+
+               imask = min_t(unsigned int, XFS_INODES_PER_CHUNK,
+                               mp->m_cluster_align_inodes) - 1;
+               if (irec->ir_startino & imask)
                        xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
-                       continue;
-               }
+               return;
+       }
 
-               /* If any part of this is a hole, skip it. */
-               if (ir_holemask) {
-                       xchk_xref_is_not_owned_by(bs->sc, agbno,
-                                       mp->m_blocks_per_cluster,
-                                       &XFS_RMAP_OINFO_INODES);
-                       continue;
+       if (iabt->next_startino != NULLAGINO) {
+               /*
+                * We're midway through a cluster of inodes that is mapped by
+                * multiple inobt records.  Did we get the record for the next
+                * irec in the sequence?
+                */
+               if (irec->ir_startino != iabt->next_startino) {
+                       xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
+                       return;
                }
 
-               xchk_xref_is_owned_by(bs->sc, agbno, mp->m_blocks_per_cluster,
-                               &XFS_RMAP_OINFO_INODES);
+               iabt->next_startino += XFS_INODES_PER_CHUNK;
 
-               /* Grab the inode cluster buffer. */
-               imap.im_blkno = XFS_AGB_TO_DADDR(mp, bs->cur->bc_private.a.agno,
-                               agbno);
-               imap.im_len = XFS_FSB_TO_BB(mp, mp->m_blocks_per_cluster);
-               imap.im_boffset = 0;
-
-               error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap,
-                               &dip, &bp, 0, 0);
-               if (!xchk_btree_xref_process_error(bs->sc, bs->cur, 0,
-                               &error))
-                       continue;
-
-               /* Which inodes are free? */
-               for (clusterino = 0; clusterino < nr_inodes; clusterino++) {
-                       error = xchk_iallocbt_check_cluster_freemask(bs,
-                                       fsino, chunkino, clusterino, irec, bp);
-                       if (error) {
-                               xfs_trans_brelse(bs->cur->bc_tp, bp);
-                               return error;
-                       }
+               /* Are we done with the cluster? */
+               if (iabt->next_startino >= iabt->next_cluster_ino) {
+                       iabt->next_startino = NULLAGINO;
+                       iabt->next_cluster_ino = NULLAGINO;
                }
+               return;
+       }
 
-               xfs_trans_brelse(bs->cur->bc_tp, bp);
+       /* inobt records must be aligned to cluster and inoalignmnt size. */
+       if (irec->ir_startino & (mp->m_cluster_align_inodes - 1)) {
+               xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
+               return;
        }
 
-       return error;
+       if (irec->ir_startino & (mp->m_inodes_per_cluster - 1)) {
+               xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
+               return;
+       }
+
+       if (mp->m_inodes_per_cluster <= XFS_INODES_PER_CHUNK)
+               return;
+
+       /*
+        * If this is the start of an inode cluster that can be mapped by
+        * multiple inobt records, the next inobt record must follow exactly
+        * after this one.
+        */
+       iabt->next_startino = irec->ir_startino + XFS_INODES_PER_CHUNK;
+       iabt->next_cluster_ino = irec->ir_startino + mp->m_inodes_per_cluster;
 }
 
 /* Scrub an inobt/finobt record. */
@@ -276,7 +374,6 @@ xchk_iallocbt_rec(
        uint64_t                        holes;
        xfs_agnumber_t                  agno = bs->cur->bc_private.a.agno;
        xfs_agino_t                     agino;
-       xfs_agblock_t                   agbno;
        xfs_extlen_t                    len;
        int                             holecount;
        int                             i;
@@ -303,11 +400,9 @@ xchk_iallocbt_rec(
                goto out;
        }
 
-       /* Make sure this record is aligned to cluster and inoalignmnt size. */
-       agbno = XFS_AGINO_TO_AGBNO(mp, irec.ir_startino);
-       if ((agbno & (mp->m_cluster_align - 1)) ||
-           (agbno & (mp->m_blocks_per_cluster - 1)))
-               xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
+       xchk_iallocbt_rec_alignment(bs, &irec);
+       if (bs->sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
+               goto out;
 
        iabt->inodes += irec.ir_count;
 
@@ -429,6 +524,8 @@ xchk_iallocbt(
        struct xfs_btree_cur    *cur;
        struct xchk_iallocbt    iabt = {
                .inodes         = 0,
+               .next_startino  = NULLAGINO,
+               .next_cluster_ino = NULLAGINO,
        };
        int                     error;