forked from luck/tmp_suning_uos_patched
7989cb8ef5
gcc-4.1 and more recent aggressively inline static functions which increases XFS stack usage by ~15% in critical paths. Prevent this from occurring by adding noinline to the STATIC definition. Also uninline some functions that are too large to be inlined and were causing problems with CONFIG_FORCED_INLINING=y. Finally, clean up all the different users of inline, __inline and __inline__ and put them under one STATIC_INLINE macro. For debug kernels the STATIC_INLINE macro uninlines those functions. SGI-PV: 957159 SGI-Modid: xfs-linux-melb:xfs-kern:27585a Signed-off-by: David Chinner <dgc@sgi.com> Signed-off-by: David Chatterton <chatz@sgi.com> Signed-off-by: Tim Shimmin <tes@sgi.com>
239 lines
5.3 KiB
C
239 lines
5.3 KiB
C
/*
|
|
* Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
|
|
* All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it would be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
#include "xfs.h"
|
|
|
|
uint64_t vn_generation; /* vnode generation number */
|
|
DEFINE_SPINLOCK(vnumber_lock);
|
|
|
|
/*
|
|
* Dedicated vnode inactive/reclaim sync semaphores.
|
|
* Prime number of hash buckets since address is used as the key.
|
|
*/
|
|
#define NVSYNC 37
|
|
#define vptosync(v) (&vsync[((unsigned long)v) % NVSYNC])
|
|
static wait_queue_head_t vsync[NVSYNC];
|
|
|
|
void
|
|
vn_init(void)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < NVSYNC; i++)
|
|
init_waitqueue_head(&vsync[i]);
|
|
}
|
|
|
|
void
|
|
vn_iowait(
|
|
bhv_vnode_t *vp)
|
|
{
|
|
wait_queue_head_t *wq = vptosync(vp);
|
|
|
|
wait_event(*wq, (atomic_read(&vp->v_iocount) == 0));
|
|
}
|
|
|
|
void
|
|
vn_iowake(
|
|
bhv_vnode_t *vp)
|
|
{
|
|
if (atomic_dec_and_test(&vp->v_iocount))
|
|
wake_up(vptosync(vp));
|
|
}
|
|
|
|
/*
|
|
* Volume managers supporting multiple paths can send back ENODEV when the
|
|
* final path disappears. In this case continuing to fill the page cache
|
|
* with dirty data which cannot be written out is evil, so prevent that.
|
|
*/
|
|
void
|
|
vn_ioerror(
|
|
bhv_vnode_t *vp,
|
|
int error,
|
|
char *f,
|
|
int l)
|
|
{
|
|
if (unlikely(error == -ENODEV))
|
|
bhv_vfs_force_shutdown(vp->v_vfsp, SHUTDOWN_DEVICE_REQ, f, l);
|
|
}
|
|
|
|
bhv_vnode_t *
|
|
vn_initialize(
|
|
struct inode *inode)
|
|
{
|
|
bhv_vnode_t *vp = vn_from_inode(inode);
|
|
|
|
XFS_STATS_INC(vn_active);
|
|
XFS_STATS_INC(vn_alloc);
|
|
|
|
vp->v_flag = VMODIFIED;
|
|
spinlock_init(&vp->v_lock, "v_lock");
|
|
|
|
spin_lock(&vnumber_lock);
|
|
if (!++vn_generation) /* v_number shouldn't be zero */
|
|
vn_generation++;
|
|
vp->v_number = vn_generation;
|
|
spin_unlock(&vnumber_lock);
|
|
|
|
ASSERT(VN_CACHED(vp) == 0);
|
|
|
|
/* Initialize the first behavior and the behavior chain head. */
|
|
vn_bhv_head_init(VN_BHV_HEAD(vp), "vnode");
|
|
|
|
atomic_set(&vp->v_iocount, 0);
|
|
|
|
#ifdef XFS_VNODE_TRACE
|
|
vp->v_trace = ktrace_alloc(VNODE_TRACE_SIZE, KM_SLEEP);
|
|
#endif /* XFS_VNODE_TRACE */
|
|
|
|
vn_trace_exit(vp, __FUNCTION__, (inst_t *)__return_address);
|
|
return vp;
|
|
}
|
|
|
|
/*
|
|
* Revalidate the Linux inode from the vattr.
|
|
* Note: i_size _not_ updated; we must hold the inode
|
|
* semaphore when doing that - callers responsibility.
|
|
*/
|
|
void
|
|
vn_revalidate_core(
|
|
bhv_vnode_t *vp,
|
|
bhv_vattr_t *vap)
|
|
{
|
|
struct inode *inode = vn_to_inode(vp);
|
|
|
|
inode->i_mode = vap->va_mode;
|
|
inode->i_nlink = vap->va_nlink;
|
|
inode->i_uid = vap->va_uid;
|
|
inode->i_gid = vap->va_gid;
|
|
inode->i_blocks = vap->va_nblocks;
|
|
inode->i_mtime = vap->va_mtime;
|
|
inode->i_ctime = vap->va_ctime;
|
|
if (vap->va_xflags & XFS_XFLAG_IMMUTABLE)
|
|
inode->i_flags |= S_IMMUTABLE;
|
|
else
|
|
inode->i_flags &= ~S_IMMUTABLE;
|
|
if (vap->va_xflags & XFS_XFLAG_APPEND)
|
|
inode->i_flags |= S_APPEND;
|
|
else
|
|
inode->i_flags &= ~S_APPEND;
|
|
if (vap->va_xflags & XFS_XFLAG_SYNC)
|
|
inode->i_flags |= S_SYNC;
|
|
else
|
|
inode->i_flags &= ~S_SYNC;
|
|
if (vap->va_xflags & XFS_XFLAG_NOATIME)
|
|
inode->i_flags |= S_NOATIME;
|
|
else
|
|
inode->i_flags &= ~S_NOATIME;
|
|
}
|
|
|
|
/*
|
|
* Revalidate the Linux inode from the vnode.
|
|
*/
|
|
int
|
|
__vn_revalidate(
|
|
bhv_vnode_t *vp,
|
|
bhv_vattr_t *vattr)
|
|
{
|
|
int error;
|
|
|
|
vn_trace_entry(vp, __FUNCTION__, (inst_t *)__return_address);
|
|
vattr->va_mask = XFS_AT_STAT | XFS_AT_XFLAGS;
|
|
error = bhv_vop_getattr(vp, vattr, 0, NULL);
|
|
if (likely(!error)) {
|
|
vn_revalidate_core(vp, vattr);
|
|
VUNMODIFY(vp);
|
|
}
|
|
return -error;
|
|
}
|
|
|
|
int
|
|
vn_revalidate(
|
|
bhv_vnode_t *vp)
|
|
{
|
|
bhv_vattr_t vattr;
|
|
|
|
return __vn_revalidate(vp, &vattr);
|
|
}
|
|
|
|
/*
|
|
* Add a reference to a referenced vnode.
|
|
*/
|
|
bhv_vnode_t *
|
|
vn_hold(
|
|
bhv_vnode_t *vp)
|
|
{
|
|
struct inode *inode;
|
|
|
|
XFS_STATS_INC(vn_hold);
|
|
|
|
VN_LOCK(vp);
|
|
inode = igrab(vn_to_inode(vp));
|
|
ASSERT(inode);
|
|
VN_UNLOCK(vp, 0);
|
|
|
|
return vp;
|
|
}
|
|
|
|
#ifdef XFS_VNODE_TRACE
|
|
|
|
#define KTRACE_ENTER(vp, vk, s, line, ra) \
|
|
ktrace_enter( (vp)->v_trace, \
|
|
/* 0 */ (void *)(__psint_t)(vk), \
|
|
/* 1 */ (void *)(s), \
|
|
/* 2 */ (void *)(__psint_t) line, \
|
|
/* 3 */ (void *)(__psint_t)(vn_count(vp)), \
|
|
/* 4 */ (void *)(ra), \
|
|
/* 5 */ (void *)(__psunsigned_t)(vp)->v_flag, \
|
|
/* 6 */ (void *)(__psint_t)current_cpu(), \
|
|
/* 7 */ (void *)(__psint_t)current_pid(), \
|
|
/* 8 */ (void *)__return_address, \
|
|
/* 9 */ NULL, NULL, NULL, NULL, NULL, NULL, NULL)
|
|
|
|
/*
|
|
* Vnode tracing code.
|
|
*/
|
|
void
|
|
vn_trace_entry(bhv_vnode_t *vp, const char *func, inst_t *ra)
|
|
{
|
|
KTRACE_ENTER(vp, VNODE_KTRACE_ENTRY, func, 0, ra);
|
|
}
|
|
|
|
void
|
|
vn_trace_exit(bhv_vnode_t *vp, const char *func, inst_t *ra)
|
|
{
|
|
KTRACE_ENTER(vp, VNODE_KTRACE_EXIT, func, 0, ra);
|
|
}
|
|
|
|
void
|
|
vn_trace_hold(bhv_vnode_t *vp, char *file, int line, inst_t *ra)
|
|
{
|
|
KTRACE_ENTER(vp, VNODE_KTRACE_HOLD, file, line, ra);
|
|
}
|
|
|
|
void
|
|
vn_trace_ref(bhv_vnode_t *vp, char *file, int line, inst_t *ra)
|
|
{
|
|
KTRACE_ENTER(vp, VNODE_KTRACE_REF, file, line, ra);
|
|
}
|
|
|
|
void
|
|
vn_trace_rele(bhv_vnode_t *vp, char *file, int line, inst_t *ra)
|
|
{
|
|
KTRACE_ENTER(vp, VNODE_KTRACE_RELE, file, line, ra);
|
|
}
|
|
#endif /* XFS_VNODE_TRACE */
|