cd57e594ad
- sanity check for NULL user buffer in xfs_ioc_bulkstat[_compat]() - remove the special case for XFS_IOC_FSBULKSTAT with count == 1. This special case causes bulkstat to fail because the special case uses xfs_bulkstat_single() instead of xfs_bulkstat() and the two functions have different semantics. xfs_bulkstat() will return the next inode after the one supplied while skipping internal inodes (ie quota inodes). xfs_bulkstate_single() will only lookup the inode supplied and return an error if it is an internal inode. - in xfs_bulkstat(), need to initialise 'lastino' to the inode supplied so in cases were we return without examining any inodes the scan wont restart back at zero. - sanity check for valid *ubcountp values. Cannot sanity check for valid ubuffer here because some users of xfs_bulkstat() don't supply a buffer. - checks against 'ubleft' (the space left in the user's buffer) should be against 'statstruct_size' which is the supplied minimum object size. The mixture of checks against statstruct_size and 0 was one of the reasons we were skipping inodes. - if the formatter function returns BULKSTAT_RV_NOTHING and an error and the error is not ENOENT or EINVAL then we need to abort the scan. ENOENT is for inodes that are no longer valid and we just skip them. EINVAL is returned if we try to lookup an internal inode so we skip them too. For a DMF scan if the inode and DMF attribute cannot fit into the space left in the user's buffer it would return ERANGE. We didn't handle this error and skipped the inode. We would continue to skip inodes until one fitted into the user's buffer or we completed the scan. - put back the recalculation of agino (that got removed with the last fix) at the end of the while loop. This is because the code at the start of the loop expects agino to be the last inode examined if it is non-zero. - if we found some inodes but then encountered an error, return success this time and the error next time. If the formatter aborted with ENOMEM we will now return this error but only if we couldn't read any inodes. Previously if we encountered ENOMEM without reading any inodes we returned a zero count and no error which falsely indicated the scan was complete. SGI-PV: 973431 SGI-Modid: xfs-linux-melb:xfs-kern:30089a Signed-off-by: Lachlan McIlroy <lachlan@sgi.com> Signed-off-by: David Chinner <dgc@sgi.com>
932 lines
26 KiB
C
932 lines
26 KiB
C
/*
|
|
* Copyright (c) 2000-2002,2005 Silicon Graphics, Inc.
|
|
* All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it would be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
#include "xfs.h"
|
|
#include "xfs_fs.h"
|
|
#include "xfs_types.h"
|
|
#include "xfs_bit.h"
|
|
#include "xfs_log.h"
|
|
#include "xfs_inum.h"
|
|
#include "xfs_trans.h"
|
|
#include "xfs_sb.h"
|
|
#include "xfs_ag.h"
|
|
#include "xfs_dir2.h"
|
|
#include "xfs_dmapi.h"
|
|
#include "xfs_mount.h"
|
|
#include "xfs_bmap_btree.h"
|
|
#include "xfs_alloc_btree.h"
|
|
#include "xfs_ialloc_btree.h"
|
|
#include "xfs_dir2_sf.h"
|
|
#include "xfs_attr_sf.h"
|
|
#include "xfs_dinode.h"
|
|
#include "xfs_inode.h"
|
|
#include "xfs_ialloc.h"
|
|
#include "xfs_itable.h"
|
|
#include "xfs_error.h"
|
|
#include "xfs_btree.h"
|
|
|
|
int
|
|
xfs_internal_inum(
|
|
xfs_mount_t *mp,
|
|
xfs_ino_t ino)
|
|
{
|
|
return (ino == mp->m_sb.sb_rbmino || ino == mp->m_sb.sb_rsumino ||
|
|
(XFS_SB_VERSION_HASQUOTA(&mp->m_sb) &&
|
|
(ino == mp->m_sb.sb_uquotino || ino == mp->m_sb.sb_gquotino)));
|
|
}
|
|
|
|
STATIC int
|
|
xfs_bulkstat_one_iget(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t ino, /* inode number to get data for */
|
|
xfs_daddr_t bno, /* starting bno of inode cluster */
|
|
xfs_bstat_t *buf, /* return buffer */
|
|
int *stat) /* BULKSTAT_RV_... */
|
|
{
|
|
xfs_icdinode_t *dic; /* dinode core info pointer */
|
|
xfs_inode_t *ip; /* incore inode pointer */
|
|
bhv_vnode_t *vp;
|
|
int error;
|
|
|
|
error = xfs_iget(mp, NULL, ino,
|
|
XFS_IGET_BULKSTAT, XFS_ILOCK_SHARED, &ip, bno);
|
|
if (error) {
|
|
*stat = BULKSTAT_RV_NOTHING;
|
|
return error;
|
|
}
|
|
|
|
ASSERT(ip != NULL);
|
|
ASSERT(ip->i_blkno != (xfs_daddr_t)0);
|
|
if (ip->i_d.di_mode == 0) {
|
|
*stat = BULKSTAT_RV_NOTHING;
|
|
error = XFS_ERROR(ENOENT);
|
|
goto out_iput;
|
|
}
|
|
|
|
vp = XFS_ITOV(ip);
|
|
dic = &ip->i_d;
|
|
|
|
/* xfs_iget returns the following without needing
|
|
* further change.
|
|
*/
|
|
buf->bs_nlink = dic->di_nlink;
|
|
buf->bs_projid = dic->di_projid;
|
|
buf->bs_ino = ino;
|
|
buf->bs_mode = dic->di_mode;
|
|
buf->bs_uid = dic->di_uid;
|
|
buf->bs_gid = dic->di_gid;
|
|
buf->bs_size = dic->di_size;
|
|
vn_atime_to_bstime(vp, &buf->bs_atime);
|
|
buf->bs_mtime.tv_sec = dic->di_mtime.t_sec;
|
|
buf->bs_mtime.tv_nsec = dic->di_mtime.t_nsec;
|
|
buf->bs_ctime.tv_sec = dic->di_ctime.t_sec;
|
|
buf->bs_ctime.tv_nsec = dic->di_ctime.t_nsec;
|
|
buf->bs_xflags = xfs_ip2xflags(ip);
|
|
buf->bs_extsize = dic->di_extsize << mp->m_sb.sb_blocklog;
|
|
buf->bs_extents = dic->di_nextents;
|
|
buf->bs_gen = dic->di_gen;
|
|
memset(buf->bs_pad, 0, sizeof(buf->bs_pad));
|
|
buf->bs_dmevmask = dic->di_dmevmask;
|
|
buf->bs_dmstate = dic->di_dmstate;
|
|
buf->bs_aextents = dic->di_anextents;
|
|
|
|
switch (dic->di_format) {
|
|
case XFS_DINODE_FMT_DEV:
|
|
buf->bs_rdev = ip->i_df.if_u2.if_rdev;
|
|
buf->bs_blksize = BLKDEV_IOSIZE;
|
|
buf->bs_blocks = 0;
|
|
break;
|
|
case XFS_DINODE_FMT_LOCAL:
|
|
case XFS_DINODE_FMT_UUID:
|
|
buf->bs_rdev = 0;
|
|
buf->bs_blksize = mp->m_sb.sb_blocksize;
|
|
buf->bs_blocks = 0;
|
|
break;
|
|
case XFS_DINODE_FMT_EXTENTS:
|
|
case XFS_DINODE_FMT_BTREE:
|
|
buf->bs_rdev = 0;
|
|
buf->bs_blksize = mp->m_sb.sb_blocksize;
|
|
buf->bs_blocks = dic->di_nblocks + ip->i_delayed_blks;
|
|
break;
|
|
}
|
|
|
|
out_iput:
|
|
xfs_iput(ip, XFS_ILOCK_SHARED);
|
|
return error;
|
|
}
|
|
|
|
STATIC int
|
|
xfs_bulkstat_one_dinode(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t ino, /* inode number to get data for */
|
|
xfs_dinode_t *dip, /* dinode inode pointer */
|
|
xfs_bstat_t *buf) /* return buffer */
|
|
{
|
|
xfs_dinode_core_t *dic; /* dinode core info pointer */
|
|
|
|
dic = &dip->di_core;
|
|
|
|
/*
|
|
* The inode format changed when we moved the link count and
|
|
* made it 32 bits long. If this is an old format inode,
|
|
* convert it in memory to look like a new one. If it gets
|
|
* flushed to disk we will convert back before flushing or
|
|
* logging it. We zero out the new projid field and the old link
|
|
* count field. We'll handle clearing the pad field (the remains
|
|
* of the old uuid field) when we actually convert the inode to
|
|
* the new format. We don't change the version number so that we
|
|
* can distinguish this from a real new format inode.
|
|
*/
|
|
if (dic->di_version == XFS_DINODE_VERSION_1) {
|
|
buf->bs_nlink = be16_to_cpu(dic->di_onlink);
|
|
buf->bs_projid = 0;
|
|
} else {
|
|
buf->bs_nlink = be32_to_cpu(dic->di_nlink);
|
|
buf->bs_projid = be16_to_cpu(dic->di_projid);
|
|
}
|
|
|
|
buf->bs_ino = ino;
|
|
buf->bs_mode = be16_to_cpu(dic->di_mode);
|
|
buf->bs_uid = be32_to_cpu(dic->di_uid);
|
|
buf->bs_gid = be32_to_cpu(dic->di_gid);
|
|
buf->bs_size = be64_to_cpu(dic->di_size);
|
|
buf->bs_atime.tv_sec = be32_to_cpu(dic->di_atime.t_sec);
|
|
buf->bs_atime.tv_nsec = be32_to_cpu(dic->di_atime.t_nsec);
|
|
buf->bs_mtime.tv_sec = be32_to_cpu(dic->di_mtime.t_sec);
|
|
buf->bs_mtime.tv_nsec = be32_to_cpu(dic->di_mtime.t_nsec);
|
|
buf->bs_ctime.tv_sec = be32_to_cpu(dic->di_ctime.t_sec);
|
|
buf->bs_ctime.tv_nsec = be32_to_cpu(dic->di_ctime.t_nsec);
|
|
buf->bs_xflags = xfs_dic2xflags(dic);
|
|
buf->bs_extsize = be32_to_cpu(dic->di_extsize) << mp->m_sb.sb_blocklog;
|
|
buf->bs_extents = be32_to_cpu(dic->di_nextents);
|
|
buf->bs_gen = be32_to_cpu(dic->di_gen);
|
|
memset(buf->bs_pad, 0, sizeof(buf->bs_pad));
|
|
buf->bs_dmevmask = be32_to_cpu(dic->di_dmevmask);
|
|
buf->bs_dmstate = be16_to_cpu(dic->di_dmstate);
|
|
buf->bs_aextents = be16_to_cpu(dic->di_anextents);
|
|
|
|
switch (dic->di_format) {
|
|
case XFS_DINODE_FMT_DEV:
|
|
buf->bs_rdev = be32_to_cpu(dip->di_u.di_dev);
|
|
buf->bs_blksize = BLKDEV_IOSIZE;
|
|
buf->bs_blocks = 0;
|
|
break;
|
|
case XFS_DINODE_FMT_LOCAL:
|
|
case XFS_DINODE_FMT_UUID:
|
|
buf->bs_rdev = 0;
|
|
buf->bs_blksize = mp->m_sb.sb_blocksize;
|
|
buf->bs_blocks = 0;
|
|
break;
|
|
case XFS_DINODE_FMT_EXTENTS:
|
|
case XFS_DINODE_FMT_BTREE:
|
|
buf->bs_rdev = 0;
|
|
buf->bs_blksize = mp->m_sb.sb_blocksize;
|
|
buf->bs_blocks = be64_to_cpu(dic->di_nblocks);
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
STATIC int
|
|
xfs_bulkstat_one_fmt(
|
|
void __user *ubuffer,
|
|
const xfs_bstat_t *buffer)
|
|
{
|
|
if (copy_to_user(ubuffer, buffer, sizeof(*buffer)))
|
|
return -EFAULT;
|
|
return sizeof(*buffer);
|
|
}
|
|
|
|
/*
|
|
* Return stat information for one inode.
|
|
* Return 0 if ok, else errno.
|
|
*/
|
|
int /* error status */
|
|
xfs_bulkstat_one(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t ino, /* inode number to get data for */
|
|
void __user *buffer, /* buffer to place output in */
|
|
int ubsize, /* size of buffer */
|
|
void *private_data, /* my private data */
|
|
xfs_daddr_t bno, /* starting bno of inode cluster */
|
|
int *ubused, /* bytes used by me */
|
|
void *dibuff, /* on-disk inode buffer */
|
|
int *stat) /* BULKSTAT_RV_... */
|
|
{
|
|
xfs_bstat_t *buf; /* return buffer */
|
|
int error = 0; /* error value */
|
|
xfs_dinode_t *dip; /* dinode inode pointer */
|
|
bulkstat_one_fmt_pf formatter = private_data ? : xfs_bulkstat_one_fmt;
|
|
|
|
dip = (xfs_dinode_t *)dibuff;
|
|
*stat = BULKSTAT_RV_NOTHING;
|
|
|
|
if (!buffer || xfs_internal_inum(mp, ino))
|
|
return XFS_ERROR(EINVAL);
|
|
if (ubsize < sizeof(*buf))
|
|
return XFS_ERROR(ENOMEM);
|
|
|
|
buf = kmem_alloc(sizeof(*buf), KM_SLEEP);
|
|
|
|
if (dip == NULL) {
|
|
/* We're not being passed a pointer to a dinode. This happens
|
|
* if BULKSTAT_FG_IGET is selected. Do the iget.
|
|
*/
|
|
error = xfs_bulkstat_one_iget(mp, ino, bno, buf, stat);
|
|
if (error)
|
|
goto out_free;
|
|
} else {
|
|
xfs_bulkstat_one_dinode(mp, ino, dip, buf);
|
|
}
|
|
|
|
error = formatter(buffer, buf);
|
|
if (error < 0) {
|
|
error = EFAULT;
|
|
goto out_free;
|
|
}
|
|
|
|
*stat = BULKSTAT_RV_DIDONE;
|
|
if (ubused)
|
|
*ubused = error;
|
|
|
|
out_free:
|
|
kmem_free(buf, sizeof(*buf));
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* Test to see whether we can use the ondisk inode directly, based
|
|
* on the given bulkstat flags, filling in dipp accordingly.
|
|
* Returns zero if the inode is dodgey.
|
|
*/
|
|
STATIC int
|
|
xfs_bulkstat_use_dinode(
|
|
xfs_mount_t *mp,
|
|
int flags,
|
|
xfs_buf_t *bp,
|
|
int clustidx,
|
|
xfs_dinode_t **dipp)
|
|
{
|
|
xfs_dinode_t *dip;
|
|
unsigned int aformat;
|
|
|
|
*dipp = NULL;
|
|
if (!bp || (flags & BULKSTAT_FG_IGET))
|
|
return 1;
|
|
dip = (xfs_dinode_t *)
|
|
xfs_buf_offset(bp, clustidx << mp->m_sb.sb_inodelog);
|
|
/*
|
|
* Check the buffer containing the on-disk inode for di_nlink == 0.
|
|
* This is to prevent xfs_bulkstat from picking up just reclaimed
|
|
* inodes that have their in-core state initialized but not flushed
|
|
* to disk yet. This is a temporary hack that would require a proper
|
|
* fix in the future.
|
|
*/
|
|
if (be16_to_cpu(dip->di_core.di_magic) != XFS_DINODE_MAGIC ||
|
|
!XFS_DINODE_GOOD_VERSION(dip->di_core.di_version) ||
|
|
!dip->di_core.di_nlink)
|
|
return 0;
|
|
if (flags & BULKSTAT_FG_QUICK) {
|
|
*dipp = dip;
|
|
return 1;
|
|
}
|
|
/* BULKSTAT_FG_INLINE: if attr fork is local, or not there, use it */
|
|
aformat = dip->di_core.di_aformat;
|
|
if ((XFS_CFORK_Q(&dip->di_core) == 0) ||
|
|
(aformat == XFS_DINODE_FMT_LOCAL) ||
|
|
(aformat == XFS_DINODE_FMT_EXTENTS && !dip->di_core.di_anextents)) {
|
|
*dipp = dip;
|
|
return 1;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
#define XFS_BULKSTAT_UBLEFT(ubleft) ((ubleft) >= statstruct_size)
|
|
|
|
/*
|
|
* Return stat information in bulk (by-inode) for the filesystem.
|
|
*/
|
|
int /* error status */
|
|
xfs_bulkstat(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t *lastinop, /* last inode returned */
|
|
int *ubcountp, /* size of buffer/count returned */
|
|
bulkstat_one_pf formatter, /* func that'd fill a single buf */
|
|
void *private_data,/* private data for formatter */
|
|
size_t statstruct_size, /* sizeof struct filling */
|
|
char __user *ubuffer, /* buffer with inode stats */
|
|
int flags, /* defined in xfs_itable.h */
|
|
int *done) /* 1 if there are more stats to get */
|
|
{
|
|
xfs_agblock_t agbno=0;/* allocation group block number */
|
|
xfs_buf_t *agbp; /* agi header buffer */
|
|
xfs_agi_t *agi; /* agi header data */
|
|
xfs_agino_t agino; /* inode # in allocation group */
|
|
xfs_agnumber_t agno; /* allocation group number */
|
|
xfs_daddr_t bno; /* inode cluster start daddr */
|
|
int chunkidx; /* current index into inode chunk */
|
|
int clustidx; /* current index into inode cluster */
|
|
xfs_btree_cur_t *cur; /* btree cursor for ialloc btree */
|
|
int end_of_ag; /* set if we've seen the ag end */
|
|
int error; /* error code */
|
|
int fmterror;/* bulkstat formatter result */
|
|
__int32_t gcnt; /* current btree rec's count */
|
|
xfs_inofree_t gfree; /* current btree rec's free mask */
|
|
xfs_agino_t gino; /* current btree rec's start inode */
|
|
int i; /* loop index */
|
|
int icount; /* count of inodes good in irbuf */
|
|
size_t irbsize; /* size of irec buffer in bytes */
|
|
xfs_ino_t ino; /* inode number (filesystem) */
|
|
xfs_inobt_rec_incore_t *irbp; /* current irec buffer pointer */
|
|
xfs_inobt_rec_incore_t *irbuf; /* start of irec buffer */
|
|
xfs_inobt_rec_incore_t *irbufend; /* end of good irec buffer entries */
|
|
xfs_ino_t lastino; /* last inode number returned */
|
|
int nbcluster; /* # of blocks in a cluster */
|
|
int nicluster; /* # of inodes in a cluster */
|
|
int nimask; /* mask for inode clusters */
|
|
int nirbuf; /* size of irbuf */
|
|
int rval; /* return value error code */
|
|
int tmp; /* result value from btree calls */
|
|
int ubcount; /* size of user's buffer */
|
|
int ubleft; /* bytes left in user's buffer */
|
|
char __user *ubufp; /* pointer into user's buffer */
|
|
int ubelem; /* spaces used in user's buffer */
|
|
int ubused; /* bytes used by formatter */
|
|
xfs_buf_t *bp; /* ptr to on-disk inode cluster buf */
|
|
xfs_dinode_t *dip; /* ptr into bp for specific inode */
|
|
xfs_inode_t *ip; /* ptr to in-core inode struct */
|
|
|
|
/*
|
|
* Get the last inode value, see if there's nothing to do.
|
|
*/
|
|
ino = (xfs_ino_t)*lastinop;
|
|
lastino = ino;
|
|
dip = NULL;
|
|
agno = XFS_INO_TO_AGNO(mp, ino);
|
|
agino = XFS_INO_TO_AGINO(mp, ino);
|
|
if (agno >= mp->m_sb.sb_agcount ||
|
|
ino != XFS_AGINO_TO_INO(mp, agno, agino)) {
|
|
*done = 1;
|
|
*ubcountp = 0;
|
|
return 0;
|
|
}
|
|
if (!ubcountp || *ubcountp <= 0) {
|
|
return EINVAL;
|
|
}
|
|
ubcount = *ubcountp; /* statstruct's */
|
|
ubleft = ubcount * statstruct_size; /* bytes */
|
|
*ubcountp = ubelem = 0;
|
|
*done = 0;
|
|
fmterror = 0;
|
|
ubufp = ubuffer;
|
|
nicluster = mp->m_sb.sb_blocksize >= XFS_INODE_CLUSTER_SIZE(mp) ?
|
|
mp->m_sb.sb_inopblock :
|
|
(XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog);
|
|
nimask = ~(nicluster - 1);
|
|
nbcluster = nicluster >> mp->m_sb.sb_inopblog;
|
|
irbuf = kmem_zalloc_greedy(&irbsize, NBPC, NBPC * 4,
|
|
KM_SLEEP | KM_MAYFAIL | KM_LARGE);
|
|
nirbuf = irbsize / sizeof(*irbuf);
|
|
|
|
/*
|
|
* Loop over the allocation groups, starting from the last
|
|
* inode returned; 0 means start of the allocation group.
|
|
*/
|
|
rval = 0;
|
|
while (XFS_BULKSTAT_UBLEFT(ubleft) && agno < mp->m_sb.sb_agcount) {
|
|
cond_resched();
|
|
bp = NULL;
|
|
down_read(&mp->m_peraglock);
|
|
error = xfs_ialloc_read_agi(mp, NULL, agno, &agbp);
|
|
up_read(&mp->m_peraglock);
|
|
if (error) {
|
|
/*
|
|
* Skip this allocation group and go to the next one.
|
|
*/
|
|
agno++;
|
|
agino = 0;
|
|
continue;
|
|
}
|
|
agi = XFS_BUF_TO_AGI(agbp);
|
|
/*
|
|
* Allocate and initialize a btree cursor for ialloc btree.
|
|
*/
|
|
cur = xfs_btree_init_cursor(mp, NULL, agbp, agno, XFS_BTNUM_INO,
|
|
(xfs_inode_t *)0, 0);
|
|
irbp = irbuf;
|
|
irbufend = irbuf + nirbuf;
|
|
end_of_ag = 0;
|
|
/*
|
|
* If we're returning in the middle of an allocation group,
|
|
* we need to get the remainder of the chunk we're in.
|
|
*/
|
|
if (agino > 0) {
|
|
/*
|
|
* Lookup the inode chunk that this inode lives in.
|
|
*/
|
|
error = xfs_inobt_lookup_le(cur, agino, 0, 0, &tmp);
|
|
if (!error && /* no I/O error */
|
|
tmp && /* lookup succeeded */
|
|
/* got the record, should always work */
|
|
!(error = xfs_inobt_get_rec(cur, &gino, &gcnt,
|
|
&gfree, &i)) &&
|
|
i == 1 &&
|
|
/* this is the right chunk */
|
|
agino < gino + XFS_INODES_PER_CHUNK &&
|
|
/* lastino was not last in chunk */
|
|
(chunkidx = agino - gino + 1) <
|
|
XFS_INODES_PER_CHUNK &&
|
|
/* there are some left allocated */
|
|
XFS_INOBT_MASKN(chunkidx,
|
|
XFS_INODES_PER_CHUNK - chunkidx) & ~gfree) {
|
|
/*
|
|
* Grab the chunk record. Mark all the
|
|
* uninteresting inodes (because they're
|
|
* before our start point) free.
|
|
*/
|
|
for (i = 0; i < chunkidx; i++) {
|
|
if (XFS_INOBT_MASK(i) & ~gfree)
|
|
gcnt++;
|
|
}
|
|
gfree |= XFS_INOBT_MASKN(0, chunkidx);
|
|
irbp->ir_startino = gino;
|
|
irbp->ir_freecount = gcnt;
|
|
irbp->ir_free = gfree;
|
|
irbp++;
|
|
agino = gino + XFS_INODES_PER_CHUNK;
|
|
icount = XFS_INODES_PER_CHUNK - gcnt;
|
|
} else {
|
|
/*
|
|
* If any of those tests failed, bump the
|
|
* inode number (just in case).
|
|
*/
|
|
agino++;
|
|
icount = 0;
|
|
}
|
|
/*
|
|
* In any case, increment to the next record.
|
|
*/
|
|
if (!error)
|
|
error = xfs_inobt_increment(cur, 0, &tmp);
|
|
} else {
|
|
/*
|
|
* Start of ag. Lookup the first inode chunk.
|
|
*/
|
|
error = xfs_inobt_lookup_ge(cur, 0, 0, 0, &tmp);
|
|
icount = 0;
|
|
}
|
|
/*
|
|
* Loop through inode btree records in this ag,
|
|
* until we run out of inodes or space in the buffer.
|
|
*/
|
|
while (irbp < irbufend && icount < ubcount) {
|
|
/*
|
|
* Loop as long as we're unable to read the
|
|
* inode btree.
|
|
*/
|
|
while (error) {
|
|
agino += XFS_INODES_PER_CHUNK;
|
|
if (XFS_AGINO_TO_AGBNO(mp, agino) >=
|
|
be32_to_cpu(agi->agi_length))
|
|
break;
|
|
error = xfs_inobt_lookup_ge(cur, agino, 0, 0,
|
|
&tmp);
|
|
cond_resched();
|
|
}
|
|
/*
|
|
* If ran off the end of the ag either with an error,
|
|
* or the normal way, set end and stop collecting.
|
|
*/
|
|
if (error ||
|
|
(error = xfs_inobt_get_rec(cur, &gino, &gcnt,
|
|
&gfree, &i)) ||
|
|
i == 0) {
|
|
end_of_ag = 1;
|
|
break;
|
|
}
|
|
/*
|
|
* If this chunk has any allocated inodes, save it.
|
|
* Also start read-ahead now for this chunk.
|
|
*/
|
|
if (gcnt < XFS_INODES_PER_CHUNK) {
|
|
/*
|
|
* Loop over all clusters in the next chunk.
|
|
* Do a readahead if there are any allocated
|
|
* inodes in that cluster.
|
|
*/
|
|
for (agbno = XFS_AGINO_TO_AGBNO(mp, gino),
|
|
chunkidx = 0;
|
|
chunkidx < XFS_INODES_PER_CHUNK;
|
|
chunkidx += nicluster,
|
|
agbno += nbcluster) {
|
|
if (XFS_INOBT_MASKN(chunkidx,
|
|
nicluster) & ~gfree)
|
|
xfs_btree_reada_bufs(mp, agno,
|
|
agbno, nbcluster);
|
|
}
|
|
irbp->ir_startino = gino;
|
|
irbp->ir_freecount = gcnt;
|
|
irbp->ir_free = gfree;
|
|
irbp++;
|
|
icount += XFS_INODES_PER_CHUNK - gcnt;
|
|
}
|
|
/*
|
|
* Set agino to after this chunk and bump the cursor.
|
|
*/
|
|
agino = gino + XFS_INODES_PER_CHUNK;
|
|
error = xfs_inobt_increment(cur, 0, &tmp);
|
|
cond_resched();
|
|
}
|
|
/*
|
|
* Drop the btree buffers and the agi buffer.
|
|
* We can't hold any of the locks these represent
|
|
* when calling iget.
|
|
*/
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
|
|
xfs_buf_relse(agbp);
|
|
/*
|
|
* Now format all the good inodes into the user's buffer.
|
|
*/
|
|
irbufend = irbp;
|
|
for (irbp = irbuf;
|
|
irbp < irbufend && XFS_BULKSTAT_UBLEFT(ubleft); irbp++) {
|
|
/*
|
|
* Now process this chunk of inodes.
|
|
*/
|
|
for (agino = irbp->ir_startino, chunkidx = clustidx = 0;
|
|
XFS_BULKSTAT_UBLEFT(ubleft) &&
|
|
irbp->ir_freecount < XFS_INODES_PER_CHUNK;
|
|
chunkidx++, clustidx++, agino++) {
|
|
ASSERT(chunkidx < XFS_INODES_PER_CHUNK);
|
|
/*
|
|
* Recompute agbno if this is the
|
|
* first inode of the cluster.
|
|
*
|
|
* Careful with clustidx. There can be
|
|
* multple clusters per chunk, a single
|
|
* cluster per chunk or a cluster that has
|
|
* inodes represented from several different
|
|
* chunks (if blocksize is large).
|
|
*
|
|
* Because of this, the starting clustidx is
|
|
* initialized to zero in this loop but must
|
|
* later be reset after reading in the cluster
|
|
* buffer.
|
|
*/
|
|
if ((chunkidx & (nicluster - 1)) == 0) {
|
|
agbno = XFS_AGINO_TO_AGBNO(mp,
|
|
irbp->ir_startino) +
|
|
((chunkidx & nimask) >>
|
|
mp->m_sb.sb_inopblog);
|
|
|
|
if (flags & (BULKSTAT_FG_QUICK |
|
|
BULKSTAT_FG_INLINE)) {
|
|
ino = XFS_AGINO_TO_INO(mp, agno,
|
|
agino);
|
|
bno = XFS_AGB_TO_DADDR(mp, agno,
|
|
agbno);
|
|
|
|
/*
|
|
* Get the inode cluster buffer
|
|
*/
|
|
ASSERT(xfs_inode_zone != NULL);
|
|
ip = kmem_zone_zalloc(xfs_inode_zone,
|
|
KM_SLEEP);
|
|
ip->i_ino = ino;
|
|
ip->i_mount = mp;
|
|
spin_lock_init(&ip->i_flags_lock);
|
|
if (bp)
|
|
xfs_buf_relse(bp);
|
|
error = xfs_itobp(mp, NULL, ip,
|
|
&dip, &bp, bno,
|
|
XFS_IMAP_BULKSTAT);
|
|
if (!error)
|
|
clustidx = ip->i_boffset / mp->m_sb.sb_inodesize;
|
|
kmem_zone_free(xfs_inode_zone, ip);
|
|
if (XFS_TEST_ERROR(error != 0,
|
|
mp, XFS_ERRTAG_BULKSTAT_READ_CHUNK,
|
|
XFS_RANDOM_BULKSTAT_READ_CHUNK)) {
|
|
bp = NULL;
|
|
ubleft = 0;
|
|
rval = error;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
ino = XFS_AGINO_TO_INO(mp, agno, agino);
|
|
bno = XFS_AGB_TO_DADDR(mp, agno, agbno);
|
|
/*
|
|
* Skip if this inode is free.
|
|
*/
|
|
if (XFS_INOBT_MASK(chunkidx) & irbp->ir_free) {
|
|
lastino = ino;
|
|
continue;
|
|
}
|
|
/*
|
|
* Count used inodes as free so we can tell
|
|
* when the chunk is used up.
|
|
*/
|
|
irbp->ir_freecount++;
|
|
if (!xfs_bulkstat_use_dinode(mp, flags, bp,
|
|
clustidx, &dip)) {
|
|
lastino = ino;
|
|
continue;
|
|
}
|
|
/*
|
|
* If we need to do an iget, cannot hold bp.
|
|
* Drop it, until starting the next cluster.
|
|
*/
|
|
if ((flags & BULKSTAT_FG_INLINE) && !dip) {
|
|
if (bp)
|
|
xfs_buf_relse(bp);
|
|
bp = NULL;
|
|
}
|
|
|
|
/*
|
|
* Get the inode and fill in a single buffer.
|
|
* BULKSTAT_FG_QUICK uses dip to fill it in.
|
|
* BULKSTAT_FG_IGET uses igets.
|
|
* BULKSTAT_FG_INLINE uses dip if we have an
|
|
* inline attr fork, else igets.
|
|
* See: xfs_bulkstat_one & xfs_dm_bulkstat_one.
|
|
* This is also used to count inodes/blks, etc
|
|
* in xfs_qm_quotacheck.
|
|
*/
|
|
ubused = statstruct_size;
|
|
error = formatter(mp, ino, ubufp,
|
|
ubleft, private_data,
|
|
bno, &ubused, dip, &fmterror);
|
|
if (fmterror == BULKSTAT_RV_NOTHING) {
|
|
if (error && error != ENOENT &&
|
|
error != EINVAL) {
|
|
ubleft = 0;
|
|
rval = error;
|
|
break;
|
|
}
|
|
lastino = ino;
|
|
continue;
|
|
}
|
|
if (fmterror == BULKSTAT_RV_GIVEUP) {
|
|
ubleft = 0;
|
|
ASSERT(error);
|
|
rval = error;
|
|
break;
|
|
}
|
|
if (ubufp)
|
|
ubufp += ubused;
|
|
ubleft -= ubused;
|
|
ubelem++;
|
|
lastino = ino;
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
|
|
if (bp)
|
|
xfs_buf_relse(bp);
|
|
|
|
/*
|
|
* Set up for the next loop iteration.
|
|
*/
|
|
if (XFS_BULKSTAT_UBLEFT(ubleft)) {
|
|
if (end_of_ag) {
|
|
agno++;
|
|
agino = 0;
|
|
} else
|
|
agino = XFS_INO_TO_AGINO(mp, lastino);
|
|
} else
|
|
break;
|
|
}
|
|
/*
|
|
* Done, we're either out of filesystem or space to put the data.
|
|
*/
|
|
kmem_free(irbuf, irbsize);
|
|
*ubcountp = ubelem;
|
|
/*
|
|
* Found some inodes, return them now and return the error next time.
|
|
*/
|
|
if (ubelem)
|
|
rval = 0;
|
|
if (agno >= mp->m_sb.sb_agcount) {
|
|
/*
|
|
* If we ran out of filesystem, mark lastino as off
|
|
* the end of the filesystem, so the next call
|
|
* will return immediately.
|
|
*/
|
|
*lastinop = (xfs_ino_t)XFS_AGINO_TO_INO(mp, agno, 0);
|
|
*done = 1;
|
|
} else
|
|
*lastinop = (xfs_ino_t)lastino;
|
|
|
|
return rval;
|
|
}
|
|
|
|
/*
|
|
* Return stat information in bulk (by-inode) for the filesystem.
|
|
* Special case for non-sequential one inode bulkstat.
|
|
*/
|
|
int /* error status */
|
|
xfs_bulkstat_single(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t *lastinop, /* inode to return */
|
|
char __user *buffer, /* buffer with inode stats */
|
|
int *done) /* 1 if there are more stats to get */
|
|
{
|
|
int count; /* count value for bulkstat call */
|
|
int error; /* return value */
|
|
xfs_ino_t ino; /* filesystem inode number */
|
|
int res; /* result from bs1 */
|
|
|
|
/*
|
|
* note that requesting valid inode numbers which are not allocated
|
|
* to inodes will most likely cause xfs_itobp to generate warning
|
|
* messages about bad magic numbers. This is ok. The fact that
|
|
* the inode isn't actually an inode is handled by the
|
|
* error check below. Done this way to make the usual case faster
|
|
* at the expense of the error case.
|
|
*/
|
|
|
|
ino = (xfs_ino_t)*lastinop;
|
|
error = xfs_bulkstat_one(mp, ino, buffer, sizeof(xfs_bstat_t),
|
|
NULL, 0, NULL, NULL, &res);
|
|
if (error) {
|
|
/*
|
|
* Special case way failed, do it the "long" way
|
|
* to see if that works.
|
|
*/
|
|
(*lastinop)--;
|
|
count = 1;
|
|
if (xfs_bulkstat(mp, lastinop, &count, xfs_bulkstat_one,
|
|
NULL, sizeof(xfs_bstat_t), buffer,
|
|
BULKSTAT_FG_IGET, done))
|
|
return error;
|
|
if (count == 0 || (xfs_ino_t)*lastinop != ino)
|
|
return error == EFSCORRUPTED ?
|
|
XFS_ERROR(EINVAL) : error;
|
|
else
|
|
return 0;
|
|
}
|
|
*done = 0;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
xfs_inumbers_fmt(
|
|
void __user *ubuffer, /* buffer to write to */
|
|
const xfs_inogrp_t *buffer, /* buffer to read from */
|
|
long count, /* # of elements to read */
|
|
long *written) /* # of bytes written */
|
|
{
|
|
if (copy_to_user(ubuffer, buffer, count * sizeof(*buffer)))
|
|
return -EFAULT;
|
|
*written = count * sizeof(*buffer);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Return inode number table for the filesystem.
|
|
*/
|
|
int /* error status */
|
|
xfs_inumbers(
|
|
xfs_mount_t *mp, /* mount point for filesystem */
|
|
xfs_ino_t *lastino, /* last inode returned */
|
|
int *count, /* size of buffer/count returned */
|
|
void __user *ubuffer,/* buffer with inode descriptions */
|
|
inumbers_fmt_pf formatter)
|
|
{
|
|
xfs_buf_t *agbp;
|
|
xfs_agino_t agino;
|
|
xfs_agnumber_t agno;
|
|
int bcount;
|
|
xfs_inogrp_t *buffer;
|
|
int bufidx;
|
|
xfs_btree_cur_t *cur;
|
|
int error;
|
|
__int32_t gcnt;
|
|
xfs_inofree_t gfree;
|
|
xfs_agino_t gino;
|
|
int i;
|
|
xfs_ino_t ino;
|
|
int left;
|
|
int tmp;
|
|
|
|
ino = (xfs_ino_t)*lastino;
|
|
agno = XFS_INO_TO_AGNO(mp, ino);
|
|
agino = XFS_INO_TO_AGINO(mp, ino);
|
|
left = *count;
|
|
*count = 0;
|
|
bcount = MIN(left, (int)(NBPP / sizeof(*buffer)));
|
|
buffer = kmem_alloc(bcount * sizeof(*buffer), KM_SLEEP);
|
|
error = bufidx = 0;
|
|
cur = NULL;
|
|
agbp = NULL;
|
|
while (left > 0 && agno < mp->m_sb.sb_agcount) {
|
|
if (agbp == NULL) {
|
|
down_read(&mp->m_peraglock);
|
|
error = xfs_ialloc_read_agi(mp, NULL, agno, &agbp);
|
|
up_read(&mp->m_peraglock);
|
|
if (error) {
|
|
/*
|
|
* If we can't read the AGI of this ag,
|
|
* then just skip to the next one.
|
|
*/
|
|
ASSERT(cur == NULL);
|
|
agbp = NULL;
|
|
agno++;
|
|
agino = 0;
|
|
continue;
|
|
}
|
|
cur = xfs_btree_init_cursor(mp, NULL, agbp, agno,
|
|
XFS_BTNUM_INO, (xfs_inode_t *)0, 0);
|
|
error = xfs_inobt_lookup_ge(cur, agino, 0, 0, &tmp);
|
|
if (error) {
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
|
|
cur = NULL;
|
|
xfs_buf_relse(agbp);
|
|
agbp = NULL;
|
|
/*
|
|
* Move up the last inode in the current
|
|
* chunk. The lookup_ge will always get
|
|
* us the first inode in the next chunk.
|
|
*/
|
|
agino += XFS_INODES_PER_CHUNK - 1;
|
|
continue;
|
|
}
|
|
}
|
|
if ((error = xfs_inobt_get_rec(cur, &gino, &gcnt, &gfree,
|
|
&i)) ||
|
|
i == 0) {
|
|
xfs_buf_relse(agbp);
|
|
agbp = NULL;
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
|
|
cur = NULL;
|
|
agno++;
|
|
agino = 0;
|
|
continue;
|
|
}
|
|
agino = gino + XFS_INODES_PER_CHUNK - 1;
|
|
buffer[bufidx].xi_startino = XFS_AGINO_TO_INO(mp, agno, gino);
|
|
buffer[bufidx].xi_alloccount = XFS_INODES_PER_CHUNK - gcnt;
|
|
buffer[bufidx].xi_allocmask = ~gfree;
|
|
bufidx++;
|
|
left--;
|
|
if (bufidx == bcount) {
|
|
long written;
|
|
if (formatter(ubuffer, buffer, bufidx, &written)) {
|
|
error = XFS_ERROR(EFAULT);
|
|
break;
|
|
}
|
|
ubuffer += written;
|
|
*count += bufidx;
|
|
bufidx = 0;
|
|
}
|
|
if (left) {
|
|
error = xfs_inobt_increment(cur, 0, &tmp);
|
|
if (error) {
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
|
|
cur = NULL;
|
|
xfs_buf_relse(agbp);
|
|
agbp = NULL;
|
|
/*
|
|
* The agino value has already been bumped.
|
|
* Just try to skip up to it.
|
|
*/
|
|
agino += XFS_INODES_PER_CHUNK;
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
if (!error) {
|
|
if (bufidx) {
|
|
long written;
|
|
if (formatter(ubuffer, buffer, bufidx, &written))
|
|
error = XFS_ERROR(EFAULT);
|
|
else
|
|
*count += bufidx;
|
|
}
|
|
*lastino = XFS_AGINO_TO_INO(mp, agno, agino);
|
|
}
|
|
kmem_free(buffer, bcount * sizeof(*buffer));
|
|
if (cur)
|
|
xfs_btree_del_cursor(cur, (error ? XFS_BTREE_ERROR :
|
|
XFS_BTREE_NOERROR));
|
|
if (agbp)
|
|
xfs_buf_relse(agbp);
|
|
return error;
|
|
}
|