Merge branch 'for-linus' of git://oss.sgi.com/xfs/xfs

* 'for-linus' of git://oss.sgi.com/xfs/xfs: (69 commits)
  xfs: add AIL pushing tracepoints
  xfs: put in missed fix for merge problem
  xfs: do not flush data workqueues in xfs_flush_buftarg
  xfs: remove XFS_bflush
  xfs: remove xfs_buf_target_name
  xfs: use xfs_ioerror_alert in xfs_buf_iodone_callbacks
  xfs: clean up xfs_ioerror_alert
  xfs: clean up buffer allocation
  xfs: remove buffers from the delwri list in xfs_buf_stale
  xfs: remove XFS_BUF_STALE and XFS_BUF_SUPER_STALE
  xfs: remove XFS_BUF_SET_VTYPE and XFS_BUF_SET_VTYPE_REF
  xfs: remove XFS_BUF_FINISH_IOWAIT
  xfs: remove xfs_get_buftarg_list
  xfs: fix buffer flushing during unmount
  xfs: optimize fsync on directories
  xfs: reduce the number of log forces from tail pushing
  xfs: Don't allocate new buffers on every call to _xfs_buf_find
  xfs: simplify xfs_trans_ijoin* again
  xfs: unlock the inode before log force in xfs_change_file_space
  xfs: unlock the inode before log force in xfs_fs_nfs_commit_metadata
  ...
This commit is contained in:
Linus Torvalds 2011-10-28 10:31:42 -07:00
commit 5619a69396
47 changed files with 2041 additions and 2332 deletions

View File

@ -452,7 +452,7 @@ xfs_alloc_read_agfl(
if (error) if (error)
return error; return error;
ASSERT(!xfs_buf_geterror(bp)); ASSERT(!xfs_buf_geterror(bp));
XFS_BUF_SET_VTYPE_REF(bp, B_FS_AGFL, XFS_AGFL_REF); xfs_buf_set_ref(bp, XFS_AGFL_REF);
*bpp = bp; *bpp = bp;
return 0; return 0;
} }
@ -2139,7 +2139,7 @@ xfs_read_agf(
xfs_trans_brelse(tp, *bpp); xfs_trans_brelse(tp, *bpp);
return XFS_ERROR(EFSCORRUPTED); return XFS_ERROR(EFSCORRUPTED);
} }
XFS_BUF_SET_VTYPE_REF(*bpp, B_FS_AGF, XFS_AGF_REF); xfs_buf_set_ref(*bpp, XFS_AGF_REF);
return 0; return 0;
} }

View File

@ -38,40 +38,6 @@
#include <linux/pagevec.h> #include <linux/pagevec.h>
#include <linux/writeback.h> #include <linux/writeback.h>
/*
* Prime number of hash buckets since address is used as the key.
*/
#define NVSYNC 37
#define to_ioend_wq(v) (&xfs_ioend_wq[((unsigned long)v) % NVSYNC])
static wait_queue_head_t xfs_ioend_wq[NVSYNC];
void __init
xfs_ioend_init(void)
{
int i;
for (i = 0; i < NVSYNC; i++)
init_waitqueue_head(&xfs_ioend_wq[i]);
}
void
xfs_ioend_wait(
xfs_inode_t *ip)
{
wait_queue_head_t *wq = to_ioend_wq(ip);
wait_event(*wq, (atomic_read(&ip->i_iocount) == 0));
}
STATIC void
xfs_ioend_wake(
xfs_inode_t *ip)
{
if (atomic_dec_and_test(&ip->i_iocount))
wake_up(to_ioend_wq(ip));
}
void void
xfs_count_page_state( xfs_count_page_state(
struct page *page, struct page *page,
@ -115,25 +81,20 @@ xfs_destroy_ioend(
xfs_ioend_t *ioend) xfs_ioend_t *ioend)
{ {
struct buffer_head *bh, *next; struct buffer_head *bh, *next;
struct xfs_inode *ip = XFS_I(ioend->io_inode);
for (bh = ioend->io_buffer_head; bh; bh = next) { for (bh = ioend->io_buffer_head; bh; bh = next) {
next = bh->b_private; next = bh->b_private;
bh->b_end_io(bh, !ioend->io_error); bh->b_end_io(bh, !ioend->io_error);
} }
/* if (ioend->io_iocb) {
* Volume managers supporting multiple paths can send back ENODEV if (ioend->io_isasync) {
* when the final path disappears. In this case continuing to fill aio_complete(ioend->io_iocb, ioend->io_error ?
* the page cache with dirty data which cannot be written out is ioend->io_error : ioend->io_result, 0);
* evil, so prevent that. }
*/ inode_dio_done(ioend->io_inode);
if (unlikely(ioend->io_error == -ENODEV)) {
xfs_do_force_shutdown(ip->i_mount, SHUTDOWN_DEVICE_REQ,
__FILE__, __LINE__);
} }
xfs_ioend_wake(ip);
mempool_free(ioend, xfs_ioend_pool); mempool_free(ioend, xfs_ioend_pool);
} }
@ -155,6 +116,15 @@ xfs_ioend_new_eof(
return isize > ip->i_d.di_size ? isize : 0; return isize > ip->i_d.di_size ? isize : 0;
} }
/*
* Fast and loose check if this write could update the on-disk inode size.
*/
static inline bool xfs_ioend_is_append(struct xfs_ioend *ioend)
{
return ioend->io_offset + ioend->io_size >
XFS_I(ioend->io_inode)->i_d.di_size;
}
/* /*
* Update on-disk file size now that data has been written to disk. The * Update on-disk file size now that data has been written to disk. The
* current in-memory file size is i_size. If a write is beyond eof i_new_size * current in-memory file size is i_size. If a write is beyond eof i_new_size
@ -173,9 +143,6 @@ xfs_setfilesize(
xfs_inode_t *ip = XFS_I(ioend->io_inode); xfs_inode_t *ip = XFS_I(ioend->io_inode);
xfs_fsize_t isize; xfs_fsize_t isize;
if (unlikely(ioend->io_error))
return 0;
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL)) if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL))
return EAGAIN; return EAGAIN;
@ -192,6 +159,9 @@ xfs_setfilesize(
/* /*
* Schedule IO completion handling on the final put of an ioend. * Schedule IO completion handling on the final put of an ioend.
*
* If there is no work to do we might as well call it a day and free the
* ioend right now.
*/ */
STATIC void STATIC void
xfs_finish_ioend( xfs_finish_ioend(
@ -200,8 +170,10 @@ xfs_finish_ioend(
if (atomic_dec_and_test(&ioend->io_remaining)) { if (atomic_dec_and_test(&ioend->io_remaining)) {
if (ioend->io_type == IO_UNWRITTEN) if (ioend->io_type == IO_UNWRITTEN)
queue_work(xfsconvertd_workqueue, &ioend->io_work); queue_work(xfsconvertd_workqueue, &ioend->io_work);
else else if (xfs_ioend_is_append(ioend))
queue_work(xfsdatad_workqueue, &ioend->io_work); queue_work(xfsdatad_workqueue, &ioend->io_work);
else
xfs_destroy_ioend(ioend);
} }
} }
@ -216,17 +188,24 @@ xfs_end_io(
struct xfs_inode *ip = XFS_I(ioend->io_inode); struct xfs_inode *ip = XFS_I(ioend->io_inode);
int error = 0; int error = 0;
if (XFS_FORCED_SHUTDOWN(ip->i_mount)) {
error = -EIO;
goto done;
}
if (ioend->io_error)
goto done;
/* /*
* For unwritten extents we need to issue transactions to convert a * For unwritten extents we need to issue transactions to convert a
* range to normal written extens after the data I/O has finished. * range to normal written extens after the data I/O has finished.
*/ */
if (ioend->io_type == IO_UNWRITTEN && if (ioend->io_type == IO_UNWRITTEN) {
likely(!ioend->io_error && !XFS_FORCED_SHUTDOWN(ip->i_mount))) {
error = xfs_iomap_write_unwritten(ip, ioend->io_offset, error = xfs_iomap_write_unwritten(ip, ioend->io_offset,
ioend->io_size); ioend->io_size);
if (error) if (error) {
ioend->io_error = error; ioend->io_error = -error;
goto done;
}
} }
/* /*
@ -236,6 +215,7 @@ xfs_end_io(
error = xfs_setfilesize(ioend); error = xfs_setfilesize(ioend);
ASSERT(!error || error == EAGAIN); ASSERT(!error || error == EAGAIN);
done:
/* /*
* If we didn't complete processing of the ioend, requeue it to the * If we didn't complete processing of the ioend, requeue it to the
* tail of the workqueue for another attempt later. Otherwise destroy * tail of the workqueue for another attempt later. Otherwise destroy
@ -247,8 +227,6 @@ xfs_end_io(
/* ensure we don't spin on blocked ioends */ /* ensure we don't spin on blocked ioends */
delay(1); delay(1);
} else { } else {
if (ioend->io_iocb)
aio_complete(ioend->io_iocb, ioend->io_result, 0);
xfs_destroy_ioend(ioend); xfs_destroy_ioend(ioend);
} }
} }
@ -285,13 +263,13 @@ xfs_alloc_ioend(
* all the I/O from calling the completion routine too early. * all the I/O from calling the completion routine too early.
*/ */
atomic_set(&ioend->io_remaining, 1); atomic_set(&ioend->io_remaining, 1);
ioend->io_isasync = 0;
ioend->io_error = 0; ioend->io_error = 0;
ioend->io_list = NULL; ioend->io_list = NULL;
ioend->io_type = type; ioend->io_type = type;
ioend->io_inode = inode; ioend->io_inode = inode;
ioend->io_buffer_head = NULL; ioend->io_buffer_head = NULL;
ioend->io_buffer_tail = NULL; ioend->io_buffer_tail = NULL;
atomic_inc(&XFS_I(ioend->io_inode)->i_iocount);
ioend->io_offset = 0; ioend->io_offset = 0;
ioend->io_size = 0; ioend->io_size = 0;
ioend->io_iocb = NULL; ioend->io_iocb = NULL;
@ -337,8 +315,8 @@ xfs_map_blocks(
count = mp->m_maxioffset - offset; count = mp->m_maxioffset - offset;
end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + count); end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + count);
offset_fsb = XFS_B_TO_FSBT(mp, offset); offset_fsb = XFS_B_TO_FSBT(mp, offset);
error = xfs_bmapi(NULL, ip, offset_fsb, end_fsb - offset_fsb, error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb,
bmapi_flags, NULL, 0, imap, &nimaps, NULL); imap, &nimaps, bmapi_flags);
xfs_iunlock(ip, XFS_ILOCK_SHARED); xfs_iunlock(ip, XFS_ILOCK_SHARED);
if (error) if (error)
@ -551,7 +529,6 @@ xfs_cancel_ioend(
unlock_buffer(bh); unlock_buffer(bh);
} while ((bh = next_bh) != NULL); } while ((bh = next_bh) != NULL);
xfs_ioend_wake(XFS_I(ioend->io_inode));
mempool_free(ioend, xfs_ioend_pool); mempool_free(ioend, xfs_ioend_pool);
} while ((ioend = next) != NULL); } while ((ioend = next) != NULL);
} }
@ -1161,8 +1138,8 @@ __xfs_get_blocks(
end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + size); end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + size);
offset_fsb = XFS_B_TO_FSBT(mp, offset); offset_fsb = XFS_B_TO_FSBT(mp, offset);
error = xfs_bmapi(NULL, ip, offset_fsb, end_fsb - offset_fsb, error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb,
XFS_BMAPI_ENTIRE, NULL, 0, &imap, &nimaps, NULL); &imap, &nimaps, XFS_BMAPI_ENTIRE);
if (error) if (error)
goto out_unlock; goto out_unlock;
@ -1300,7 +1277,6 @@ xfs_end_io_direct_write(
bool is_async) bool is_async)
{ {
struct xfs_ioend *ioend = iocb->private; struct xfs_ioend *ioend = iocb->private;
struct inode *inode = ioend->io_inode;
/* /*
* blockdev_direct_IO can return an error even after the I/O * blockdev_direct_IO can return an error even after the I/O
@ -1311,28 +1287,17 @@ xfs_end_io_direct_write(
ioend->io_offset = offset; ioend->io_offset = offset;
ioend->io_size = size; ioend->io_size = size;
ioend->io_iocb = iocb;
ioend->io_result = ret;
if (private && size > 0) if (private && size > 0)
ioend->io_type = IO_UNWRITTEN; ioend->io_type = IO_UNWRITTEN;
if (is_async) { if (is_async) {
/* ioend->io_isasync = 1;
* If we are converting an unwritten extent we need to delay
* the AIO completion until after the unwrittent extent
* conversion has completed, otherwise do it ASAP.
*/
if (ioend->io_type == IO_UNWRITTEN) {
ioend->io_iocb = iocb;
ioend->io_result = ret;
} else {
aio_complete(iocb, ret, 0);
}
xfs_finish_ioend(ioend); xfs_finish_ioend(ioend);
} else { } else {
xfs_finish_ioend_sync(ioend); xfs_finish_ioend_sync(ioend);
} }
/* XXX: probably should move into the real I/O completion handler */
inode_dio_done(inode);
} }
STATIC ssize_t STATIC ssize_t

View File

@ -47,6 +47,7 @@ typedef struct xfs_ioend {
unsigned int io_type; /* delalloc / unwritten */ unsigned int io_type; /* delalloc / unwritten */
int io_error; /* I/O error code */ int io_error; /* I/O error code */
atomic_t io_remaining; /* hold count */ atomic_t io_remaining; /* hold count */
unsigned int io_isasync : 1; /* needs aio_complete */
struct inode *io_inode; /* file being written to */ struct inode *io_inode; /* file being written to */
struct buffer_head *io_buffer_head;/* buffer linked list head */ struct buffer_head *io_buffer_head;/* buffer linked list head */
struct buffer_head *io_buffer_tail;/* buffer linked list tail */ struct buffer_head *io_buffer_tail;/* buffer linked list tail */
@ -60,9 +61,6 @@ typedef struct xfs_ioend {
extern const struct address_space_operations xfs_address_space_operations; extern const struct address_space_operations xfs_address_space_operations;
extern int xfs_get_blocks(struct inode *, sector_t, struct buffer_head *, int); extern int xfs_get_blocks(struct inode *, sector_t, struct buffer_head *, int);
extern void xfs_ioend_init(void);
extern void xfs_ioend_wait(struct xfs_inode *);
extern void xfs_count_page_state(struct page *, int *, int *); extern void xfs_count_page_state(struct page *, int *, int *);
#endif /* __XFS_AOPS_H__ */ #endif /* __XFS_AOPS_H__ */

View File

@ -319,7 +319,7 @@ xfs_attr_set_int(
return (error); return (error);
} }
xfs_trans_ijoin(args.trans, dp); xfs_trans_ijoin(args.trans, dp, 0);
/* /*
* If the attribute list is non-existent or a shortform list, * If the attribute list is non-existent or a shortform list,
@ -389,7 +389,7 @@ xfs_attr_set_int(
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args.trans, dp); xfs_trans_ijoin(args.trans, dp, 0);
/* /*
* Commit the leaf transformation. We'll need another (linked) * Commit the leaf transformation. We'll need another (linked)
@ -537,7 +537,7 @@ xfs_attr_remove_int(xfs_inode_t *dp, struct xfs_name *name, int flags)
* No need to make quota reservations here. We expect to release some * No need to make quota reservations here. We expect to release some
* blocks not allocate in the common case. * blocks not allocate in the common case.
*/ */
xfs_trans_ijoin(args.trans, dp); xfs_trans_ijoin(args.trans, dp, 0);
/* /*
* Decide on what work routines to call based on the inode size. * Decide on what work routines to call based on the inode size.
@ -809,7 +809,7 @@ xfs_attr_inactive(xfs_inode_t *dp)
* No need to make quota reservations here. We expect to release some * No need to make quota reservations here. We expect to release some
* blocks, not allocate, in the common case. * blocks, not allocate, in the common case.
*/ */
xfs_trans_ijoin(trans, dp); xfs_trans_ijoin(trans, dp, 0);
/* /*
* Decide on what work routines to call based on the inode size. * Decide on what work routines to call based on the inode size.
@ -823,18 +823,6 @@ xfs_attr_inactive(xfs_inode_t *dp)
if (error) if (error)
goto out; goto out;
/*
* Signal synchronous inactive transactions unless this is a
* synchronous mount filesystem in which case we know that we're here
* because we've been called out of xfs_inactive which means that the
* last reference is gone and the unlink transaction has already hit
* the disk so async inactive transactions are safe.
*/
if (!(mp->m_flags & XFS_MOUNT_WSYNC)) {
if (dp->i_d.di_anextents > 0)
xfs_trans_set_sync(trans);
}
error = xfs_itruncate_extents(&trans, dp, XFS_ATTR_FORK, 0); error = xfs_itruncate_extents(&trans, dp, XFS_ATTR_FORK, 0);
if (error) if (error)
goto out; goto out;
@ -973,7 +961,7 @@ xfs_attr_leaf_addname(xfs_da_args_t *args)
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
/* /*
* Commit the current trans (including the inode) and start * Commit the current trans (including the inode) and start
@ -1075,7 +1063,7 @@ xfs_attr_leaf_addname(xfs_da_args_t *args)
* in all transactions. * in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
} else } else
xfs_da_buf_done(bp); xfs_da_buf_done(bp);
@ -1149,7 +1137,7 @@ xfs_attr_leaf_removename(xfs_da_args_t *args)
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
} else } else
xfs_da_buf_done(bp); xfs_da_buf_done(bp);
return(0); return(0);
@ -1303,7 +1291,7 @@ restart:
* in all transactions. * in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
/* /*
* Commit the node conversion and start the next * Commit the node conversion and start the next
@ -1340,7 +1328,7 @@ restart:
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
} else { } else {
/* /*
* Addition succeeded, update Btree hashvals. * Addition succeeded, update Btree hashvals.
@ -1452,7 +1440,7 @@ restart:
* in all transactions. * in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
} }
/* /*
@ -1584,7 +1572,7 @@ xfs_attr_node_removename(xfs_da_args_t *args)
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
/* /*
* Commit the Btree join operation and start a new trans. * Commit the Btree join operation and start a new trans.
@ -1635,7 +1623,7 @@ xfs_attr_node_removename(xfs_da_args_t *args)
* in all transactions. * in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
} else } else
xfs_da_brelse(args->trans, bp); xfs_da_brelse(args->trans, bp);
} }
@ -1975,10 +1963,9 @@ xfs_attr_rmtval_get(xfs_da_args_t *args)
lblkno = args->rmtblkno; lblkno = args->rmtblkno;
while (valuelen > 0) { while (valuelen > 0) {
nmap = ATTR_RMTVALUE_MAPSIZE; nmap = ATTR_RMTVALUE_MAPSIZE;
error = xfs_bmapi(args->trans, args->dp, (xfs_fileoff_t)lblkno, error = xfs_bmapi_read(args->dp, (xfs_fileoff_t)lblkno,
args->rmtblkcnt, args->rmtblkcnt, map, &nmap,
XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, XFS_BMAPI_ATTRFORK);
NULL, 0, map, &nmap, NULL);
if (error) if (error)
return(error); return(error);
ASSERT(nmap >= 1); ASSERT(nmap >= 1);
@ -2052,10 +2039,9 @@ xfs_attr_rmtval_set(xfs_da_args_t *args)
*/ */
xfs_bmap_init(args->flist, args->firstblock); xfs_bmap_init(args->flist, args->firstblock);
nmap = 1; nmap = 1;
error = xfs_bmapi(args->trans, dp, (xfs_fileoff_t)lblkno, error = xfs_bmapi_write(args->trans, dp, (xfs_fileoff_t)lblkno,
blkcnt, blkcnt,
XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA | XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA,
XFS_BMAPI_WRITE,
args->firstblock, args->total, &map, &nmap, args->firstblock, args->total, &map, &nmap,
args->flist); args->flist);
if (!error) { if (!error) {
@ -2074,7 +2060,7 @@ xfs_attr_rmtval_set(xfs_da_args_t *args)
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, dp); xfs_trans_ijoin(args->trans, dp, 0);
ASSERT(nmap == 1); ASSERT(nmap == 1);
ASSERT((map.br_startblock != DELAYSTARTBLOCK) && ASSERT((map.br_startblock != DELAYSTARTBLOCK) &&
@ -2104,14 +2090,11 @@ xfs_attr_rmtval_set(xfs_da_args_t *args)
*/ */
xfs_bmap_init(args->flist, args->firstblock); xfs_bmap_init(args->flist, args->firstblock);
nmap = 1; nmap = 1;
error = xfs_bmapi(NULL, dp, (xfs_fileoff_t)lblkno, error = xfs_bmapi_read(dp, (xfs_fileoff_t)lblkno,
args->rmtblkcnt, args->rmtblkcnt, &map, &nmap,
XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, XFS_BMAPI_ATTRFORK);
args->firstblock, 0, &map, &nmap, if (error)
NULL);
if (error) {
return(error); return(error);
}
ASSERT(nmap == 1); ASSERT(nmap == 1);
ASSERT((map.br_startblock != DELAYSTARTBLOCK) && ASSERT((map.br_startblock != DELAYSTARTBLOCK) &&
(map.br_startblock != HOLESTARTBLOCK)); (map.br_startblock != HOLESTARTBLOCK));
@ -2121,16 +2104,17 @@ xfs_attr_rmtval_set(xfs_da_args_t *args)
bp = xfs_buf_get(mp->m_ddev_targp, dblkno, blkcnt, bp = xfs_buf_get(mp->m_ddev_targp, dblkno, blkcnt,
XBF_LOCK | XBF_DONT_BLOCK); XBF_LOCK | XBF_DONT_BLOCK);
ASSERT(!xfs_buf_geterror(bp)); if (!bp)
return ENOMEM;
tmp = (valuelen < XFS_BUF_SIZE(bp)) ? valuelen : tmp = (valuelen < XFS_BUF_SIZE(bp)) ? valuelen :
XFS_BUF_SIZE(bp); XFS_BUF_SIZE(bp);
xfs_buf_iomove(bp, 0, tmp, src, XBRW_WRITE); xfs_buf_iomove(bp, 0, tmp, src, XBRW_WRITE);
if (tmp < XFS_BUF_SIZE(bp)) if (tmp < XFS_BUF_SIZE(bp))
xfs_buf_zero(bp, tmp, XFS_BUF_SIZE(bp) - tmp); xfs_buf_zero(bp, tmp, XFS_BUF_SIZE(bp) - tmp);
if ((error = xfs_bwrite(mp, bp))) {/* GROT: NOTE: synchronous write */ error = xfs_bwrite(bp); /* GROT: NOTE: synchronous write */
return (error); xfs_buf_relse(bp);
} if (error)
return error;
src += tmp; src += tmp;
valuelen -= tmp; valuelen -= tmp;
@ -2166,16 +2150,12 @@ xfs_attr_rmtval_remove(xfs_da_args_t *args)
/* /*
* Try to remember where we decided to put the value. * Try to remember where we decided to put the value.
*/ */
xfs_bmap_init(args->flist, args->firstblock);
nmap = 1; nmap = 1;
error = xfs_bmapi(NULL, args->dp, (xfs_fileoff_t)lblkno, error = xfs_bmapi_read(args->dp, (xfs_fileoff_t)lblkno,
args->rmtblkcnt, args->rmtblkcnt, &map, &nmap,
XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, XFS_BMAPI_ATTRFORK);
args->firstblock, 0, &map, &nmap, if (error)
args->flist);
if (error) {
return(error); return(error);
}
ASSERT(nmap == 1); ASSERT(nmap == 1);
ASSERT((map.br_startblock != DELAYSTARTBLOCK) && ASSERT((map.br_startblock != DELAYSTARTBLOCK) &&
(map.br_startblock != HOLESTARTBLOCK)); (map.br_startblock != HOLESTARTBLOCK));
@ -2188,8 +2168,7 @@ xfs_attr_rmtval_remove(xfs_da_args_t *args)
*/ */
bp = xfs_incore(mp->m_ddev_targp, dblkno, blkcnt, XBF_TRYLOCK); bp = xfs_incore(mp->m_ddev_targp, dblkno, blkcnt, XBF_TRYLOCK);
if (bp) { if (bp) {
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
XFS_BUF_UNDELAYWRITE(bp);
xfs_buf_relse(bp); xfs_buf_relse(bp);
bp = NULL; bp = NULL;
} }
@ -2227,7 +2206,7 @@ xfs_attr_rmtval_remove(xfs_da_args_t *args)
* a new one. We need the inode to be in all transactions. * a new one. We need the inode to be in all transactions.
*/ */
if (committed) if (committed)
xfs_trans_ijoin(args->trans, args->dp); xfs_trans_ijoin(args->trans, args->dp, 0);
/* /*
* Close out trans and start the next one in the chain. * Close out trans and start the next one in the chain.

View File

@ -2926,9 +2926,8 @@ xfs_attr_leaf_freextent(xfs_trans_t **trans, xfs_inode_t *dp,
* Try to remember where we decided to put the value. * Try to remember where we decided to put the value.
*/ */
nmap = 1; nmap = 1;
error = xfs_bmapi(*trans, dp, (xfs_fileoff_t)tblkno, tblkcnt, error = xfs_bmapi_read(dp, (xfs_fileoff_t)tblkno, tblkcnt,
XFS_BMAPI_ATTRFORK | XFS_BMAPI_METADATA, &map, &nmap, XFS_BMAPI_ATTRFORK);
NULL, 0, &map, &nmap, NULL);
if (error) { if (error) {
return(error); return(error);
} }
@ -2948,6 +2947,8 @@ xfs_attr_leaf_freextent(xfs_trans_t **trans, xfs_inode_t *dp,
bp = xfs_trans_get_buf(*trans, bp = xfs_trans_get_buf(*trans,
dp->i_mount->m_ddev_targp, dp->i_mount->m_ddev_targp,
dblkno, dblkcnt, XBF_LOCK); dblkno, dblkcnt, XBF_LOCK);
if (!bp)
return ENOMEM;
xfs_trans_binval(*trans, bp); xfs_trans_binval(*trans, bp);
/* /*
* Roll to next transaction. * Roll to next transaction.

File diff suppressed because it is too large Load Diff

View File

@ -62,27 +62,23 @@ typedef struct xfs_bmap_free
#define XFS_BMAP_MAX_NMAP 4 #define XFS_BMAP_MAX_NMAP 4
/* /*
* Flags for xfs_bmapi * Flags for xfs_bmapi_*
*/ */
#define XFS_BMAPI_WRITE 0x001 /* write operation: allocate space */ #define XFS_BMAPI_ENTIRE 0x001 /* return entire extent, not trimmed */
#define XFS_BMAPI_DELAY 0x002 /* delayed write operation */ #define XFS_BMAPI_METADATA 0x002 /* mapping metadata not user data */
#define XFS_BMAPI_ENTIRE 0x004 /* return entire extent, not trimmed */ #define XFS_BMAPI_ATTRFORK 0x004 /* use attribute fork not data */
#define XFS_BMAPI_METADATA 0x008 /* mapping metadata not user data */ #define XFS_BMAPI_PREALLOC 0x008 /* preallocation op: unwritten space */
#define XFS_BMAPI_ATTRFORK 0x010 /* use attribute fork not data */ #define XFS_BMAPI_IGSTATE 0x010 /* Ignore state - */
#define XFS_BMAPI_PREALLOC 0x040 /* preallocation op: unwritten space */
#define XFS_BMAPI_IGSTATE 0x080 /* Ignore state - */
/* combine contig. space */ /* combine contig. space */
#define XFS_BMAPI_CONTIG 0x100 /* must allocate only one extent */ #define XFS_BMAPI_CONTIG 0x020 /* must allocate only one extent */
/* /*
* unwritten extent conversion - this needs write cache flushing and no additional * unwritten extent conversion - this needs write cache flushing and no additional
* allocation alignments. When specified with XFS_BMAPI_PREALLOC it converts * allocation alignments. When specified with XFS_BMAPI_PREALLOC it converts
* from written to unwritten, otherwise convert from unwritten to written. * from written to unwritten, otherwise convert from unwritten to written.
*/ */
#define XFS_BMAPI_CONVERT 0x200 #define XFS_BMAPI_CONVERT 0x040
#define XFS_BMAPI_FLAGS \ #define XFS_BMAPI_FLAGS \
{ XFS_BMAPI_WRITE, "WRITE" }, \
{ XFS_BMAPI_DELAY, "DELAY" }, \
{ XFS_BMAPI_ENTIRE, "ENTIRE" }, \ { XFS_BMAPI_ENTIRE, "ENTIRE" }, \
{ XFS_BMAPI_METADATA, "METADATA" }, \ { XFS_BMAPI_METADATA, "METADATA" }, \
{ XFS_BMAPI_ATTRFORK, "ATTRFORK" }, \ { XFS_BMAPI_ATTRFORK, "ATTRFORK" }, \
@ -113,21 +109,28 @@ static inline void xfs_bmap_init(xfs_bmap_free_t *flp, xfs_fsblock_t *fbp)
* Argument structure for xfs_bmap_alloc. * Argument structure for xfs_bmap_alloc.
*/ */
typedef struct xfs_bmalloca { typedef struct xfs_bmalloca {
xfs_fsblock_t firstblock; /* i/o first block allocated */ xfs_fsblock_t *firstblock; /* i/o first block allocated */
xfs_fsblock_t rval; /* starting block of new extent */ struct xfs_bmap_free *flist; /* bmap freelist */
xfs_fileoff_t off; /* offset in file filling in */
struct xfs_trans *tp; /* transaction pointer */ struct xfs_trans *tp; /* transaction pointer */
struct xfs_inode *ip; /* incore inode pointer */ struct xfs_inode *ip; /* incore inode pointer */
struct xfs_bmbt_irec *prevp; /* extent before the new one */ struct xfs_bmbt_irec prev; /* extent before the new one */
struct xfs_bmbt_irec *gotp; /* extent after, or delayed */ struct xfs_bmbt_irec got; /* extent after, or delayed */
xfs_extlen_t alen; /* i/o length asked/allocated */
xfs_fileoff_t offset; /* offset in file filling in */
xfs_extlen_t length; /* i/o length asked/allocated */
xfs_fsblock_t blkno; /* starting block of new extent */
struct xfs_btree_cur *cur; /* btree cursor */
xfs_extnum_t idx; /* current extent index */
int nallocs;/* number of extents alloc'd */
int logflags;/* flags for transaction logging */
xfs_extlen_t total; /* total blocks needed for xaction */ xfs_extlen_t total; /* total blocks needed for xaction */
xfs_extlen_t minlen; /* minimum allocation size (blocks) */ xfs_extlen_t minlen; /* minimum allocation size (blocks) */
xfs_extlen_t minleft; /* amount must be left after alloc */ xfs_extlen_t minleft; /* amount must be left after alloc */
char eof; /* set if allocating past last extent */ char eof; /* set if allocating past last extent */
char wasdel; /* replacing a delayed allocation */ char wasdel; /* replacing a delayed allocation */
char userdata;/* set if is user data */ char userdata;/* set if is user data */
char low; /* low on space, using seq'l ags */
char aeof; /* allocated space at eof */ char aeof; /* allocated space at eof */
char conv; /* overwriting unwritten extents */ char conv; /* overwriting unwritten extents */
} xfs_bmalloca_t; } xfs_bmalloca_t;
@ -152,251 +155,62 @@ typedef struct xfs_bmalloca {
{ BMAP_RIGHT_FILLING, "RF" }, \ { BMAP_RIGHT_FILLING, "RF" }, \
{ BMAP_ATTRFORK, "ATTR" } { BMAP_ATTRFORK, "ATTR" }
/*
* Add bmap trace insert entries for all the contents of the extent list.
*
* Quite excessive tracing. Only do this for debug builds.
*/
#if defined(__KERNEL) && defined(DEBUG) #if defined(__KERNEL) && defined(DEBUG)
void void xfs_bmap_trace_exlist(struct xfs_inode *ip, xfs_extnum_t cnt,
xfs_bmap_trace_exlist( int whichfork, unsigned long caller_ip);
struct xfs_inode *ip, /* incore inode pointer */
xfs_extnum_t cnt, /* count of entries in list */
int whichfork,
unsigned long caller_ip); /* data or attr fork */
#define XFS_BMAP_TRACE_EXLIST(ip,c,w) \ #define XFS_BMAP_TRACE_EXLIST(ip,c,w) \
xfs_bmap_trace_exlist(ip,c,w, _THIS_IP_) xfs_bmap_trace_exlist(ip,c,w, _THIS_IP_)
#else #else
#define XFS_BMAP_TRACE_EXLIST(ip,c,w) #define XFS_BMAP_TRACE_EXLIST(ip,c,w)
#endif #endif
/* int xfs_bmap_add_attrfork(struct xfs_inode *ip, int size, int rsvd);
* Convert inode from non-attributed to attributed. void xfs_bmap_add_free(xfs_fsblock_t bno, xfs_filblks_t len,
* Must not be in a transaction, ip must not be locked. struct xfs_bmap_free *flist, struct xfs_mount *mp);
*/ void xfs_bmap_cancel(struct xfs_bmap_free *flist);
int /* error code */ void xfs_bmap_compute_maxlevels(struct xfs_mount *mp, int whichfork);
xfs_bmap_add_attrfork( int xfs_bmap_first_unused(struct xfs_trans *tp, struct xfs_inode *ip,
struct xfs_inode *ip, /* incore inode pointer */ xfs_extlen_t len, xfs_fileoff_t *unused, int whichfork);
int size, /* space needed for new attribute */ int xfs_bmap_last_before(struct xfs_trans *tp, struct xfs_inode *ip,
int rsvd); /* flag for reserved block allocation */ xfs_fileoff_t *last_block, int whichfork);
int xfs_bmap_last_offset(struct xfs_trans *tp, struct xfs_inode *ip,
/* xfs_fileoff_t *unused, int whichfork);
* Add the extent to the list of extents to be free at transaction end. int xfs_bmap_one_block(struct xfs_inode *ip, int whichfork);
* The list is maintained sorted (by block number). int xfs_bmap_read_extents(struct xfs_trans *tp, struct xfs_inode *ip,
*/ int whichfork);
void int xfs_bmapi_read(struct xfs_inode *ip, xfs_fileoff_t bno,
xfs_bmap_add_free( xfs_filblks_t len, struct xfs_bmbt_irec *mval,
xfs_fsblock_t bno, /* fs block number of extent */ int *nmap, int flags);
xfs_filblks_t len, /* length of extent */ int xfs_bmapi_delay(struct xfs_inode *ip, xfs_fileoff_t bno,
xfs_bmap_free_t *flist, /* list of extents */ xfs_filblks_t len, struct xfs_bmbt_irec *mval,
struct xfs_mount *mp); /* mount point structure */ int *nmap, int flags);
int xfs_bmapi_write(struct xfs_trans *tp, struct xfs_inode *ip,
/* xfs_fileoff_t bno, xfs_filblks_t len, int flags,
* Routine to clean up the free list data structure when xfs_fsblock_t *firstblock, xfs_extlen_t total,
* an error occurs during a transaction. struct xfs_bmbt_irec *mval, int *nmap,
*/ struct xfs_bmap_free *flist);
void int xfs_bunmapi(struct xfs_trans *tp, struct xfs_inode *ip,
xfs_bmap_cancel( xfs_fileoff_t bno, xfs_filblks_t len, int flags,
xfs_bmap_free_t *flist); /* free list to clean up */ xfs_extnum_t nexts, xfs_fsblock_t *firstblock,
struct xfs_bmap_free *flist, int *done);
/* int xfs_check_nostate_extents(struct xfs_ifork *ifp, xfs_extnum_t idx,
* Compute and fill in the value of the maximum depth of a bmap btree xfs_extnum_t num);
* in this filesystem. Done once, during mount. uint xfs_default_attroffset(struct xfs_inode *ip);
*/
void
xfs_bmap_compute_maxlevels(
struct xfs_mount *mp, /* file system mount structure */
int whichfork); /* data or attr fork */
/*
* Returns the file-relative block number of the first unused block in the file.
* This is the lowest-address hole if the file has holes, else the first block
* past the end of file.
*/
int /* error */
xfs_bmap_first_unused(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
xfs_extlen_t len, /* size of hole to find */
xfs_fileoff_t *unused, /* unused block num */
int whichfork); /* data or attr fork */
/*
* Returns the file-relative block number of the last block + 1 before
* last_block (input value) in the file.
* This is not based on i_size, it is based on the extent list.
* Returns 0 for local files, as they do not have an extent list.
*/
int /* error */
xfs_bmap_last_before(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
xfs_fileoff_t *last_block, /* last block */
int whichfork); /* data or attr fork */
/*
* Returns the file-relative block number of the first block past eof in
* the file. This is not based on i_size, it is based on the extent list.
* Returns 0 for local files, as they do not have an extent list.
*/
int /* error */
xfs_bmap_last_offset(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
xfs_fileoff_t *unused, /* last block num */
int whichfork); /* data or attr fork */
/*
* Returns whether the selected fork of the inode has exactly one
* block or not. For the data fork we check this matches di_size,
* implying the file's range is 0..bsize-1.
*/
int
xfs_bmap_one_block(
struct xfs_inode *ip, /* incore inode */
int whichfork); /* data or attr fork */
/*
* Read in the extents to iu_extents.
* All inode fields are set up by caller, we just traverse the btree
* and copy the records in.
*/
int /* error */
xfs_bmap_read_extents(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
int whichfork); /* data or attr fork */
/*
* Map file blocks to filesystem blocks.
* File range is given by the bno/len pair.
* Adds blocks to file if a write ("flags & XFS_BMAPI_WRITE" set)
* into a hole or past eof.
* Only allocates blocks from a single allocation group,
* to avoid locking problems.
* The returned value in "firstblock" from the first call in a transaction
* must be remembered and presented to subsequent calls in "firstblock".
* An upper bound for the number of blocks to be allocated is supplied to
* the first call in "total"; if no allocation group has that many free
* blocks then the call will fail (return NULLFSBLOCK in "firstblock").
*/
int /* error */
xfs_bmapi(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
xfs_fileoff_t bno, /* starting file offs. mapped */
xfs_filblks_t len, /* length to map in file */
int flags, /* XFS_BMAPI_... */
xfs_fsblock_t *firstblock, /* first allocated block
controls a.g. for allocs */
xfs_extlen_t total, /* total blocks needed */
struct xfs_bmbt_irec *mval, /* output: map values */
int *nmap, /* i/o: mval size/count */
xfs_bmap_free_t *flist); /* i/o: list extents to free */
/*
* Map file blocks to filesystem blocks, simple version.
* One block only, read-only.
* For flags, only the XFS_BMAPI_ATTRFORK flag is examined.
* For the other flag values, the effect is as if XFS_BMAPI_METADATA
* was set and all the others were clear.
*/
int /* error */
xfs_bmapi_single(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
int whichfork, /* data or attr fork */
xfs_fsblock_t *fsb, /* output: mapped block */
xfs_fileoff_t bno); /* starting file offs. mapped */
/*
* Unmap (remove) blocks from a file.
* If nexts is nonzero then the number of extents to remove is limited to
* that value. If not all extents in the block range can be removed then
* *done is set.
*/
int /* error */
xfs_bunmapi(
struct xfs_trans *tp, /* transaction pointer */
struct xfs_inode *ip, /* incore inode */
xfs_fileoff_t bno, /* starting offset to unmap */
xfs_filblks_t len, /* length to unmap in file */
int flags, /* XFS_BMAPI_... */
xfs_extnum_t nexts, /* number of extents max */
xfs_fsblock_t *firstblock, /* first allocated block
controls a.g. for allocs */
xfs_bmap_free_t *flist, /* i/o: list extents to free */
int *done); /* set if not done yet */
/*
* Check an extent list, which has just been read, for
* any bit in the extent flag field.
*/
int
xfs_check_nostate_extents(
struct xfs_ifork *ifp,
xfs_extnum_t idx,
xfs_extnum_t num);
uint
xfs_default_attroffset(
struct xfs_inode *ip);
#ifdef __KERNEL__ #ifdef __KERNEL__
/*
* Routine to be called at transaction's end by xfs_bmapi, xfs_bunmapi
* caller. Frees all the extents that need freeing, which must be done
* last due to locking considerations.
*
* Return 1 if the given transaction was committed and a new one allocated,
* and 0 otherwise.
*/
int /* error */
xfs_bmap_finish(
struct xfs_trans **tp, /* transaction pointer addr */
xfs_bmap_free_t *flist, /* i/o: list extents to free */
int *committed); /* xact committed or not */
/* bmap to userspace formatter - copy to user & advance pointer */ /* bmap to userspace formatter - copy to user & advance pointer */
typedef int (*xfs_bmap_format_t)(void **, struct getbmapx *, int *); typedef int (*xfs_bmap_format_t)(void **, struct getbmapx *, int *);
/* int xfs_bmap_finish(struct xfs_trans **tp, struct xfs_bmap_free *flist,
* Get inode's extents as described in bmv, and format for output. int *committed);
*/ int xfs_getbmap(struct xfs_inode *ip, struct getbmapx *bmv,
int /* error code */ xfs_bmap_format_t formatter, void *arg);
xfs_getbmap( int xfs_bmap_eof(struct xfs_inode *ip, xfs_fileoff_t endoff,
xfs_inode_t *ip, int whichfork, int *eof);
struct getbmapx *bmv, /* user bmap structure */ int xfs_bmap_count_blocks(struct xfs_trans *tp, struct xfs_inode *ip,
xfs_bmap_format_t formatter, /* format to user */ int whichfork, int *count);
void *arg); /* formatter arg */ int xfs_bmap_punch_delalloc_range(struct xfs_inode *ip,
xfs_fileoff_t start_fsb, xfs_fileoff_t length);
/*
* Check if the endoff is outside the last extent. If so the caller will grow
* the allocation to a stripe unit boundary
*/
int
xfs_bmap_eof(
struct xfs_inode *ip,
xfs_fileoff_t endoff,
int whichfork,
int *eof);
/*
* Count fsblocks of the given fork.
*/
int
xfs_bmap_count_blocks(
xfs_trans_t *tp,
struct xfs_inode *ip,
int whichfork,
int *count);
int
xfs_bmap_punch_delalloc_range(
struct xfs_inode *ip,
xfs_fileoff_t start_fsb,
xfs_fileoff_t length);
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif /* __XFS_BMAP_H__ */ #endif /* __XFS_BMAP_H__ */

View File

@ -631,7 +631,7 @@ xfs_btree_read_bufl(
} }
ASSERT(!xfs_buf_geterror(bp)); ASSERT(!xfs_buf_geterror(bp));
if (bp) if (bp)
XFS_BUF_SET_VTYPE_REF(bp, B_FS_MAP, refval); xfs_buf_set_ref(bp, refval);
*bpp = bp; *bpp = bp;
return 0; return 0;
} }
@ -939,13 +939,13 @@ xfs_btree_set_refs(
switch (cur->bc_btnum) { switch (cur->bc_btnum) {
case XFS_BTNUM_BNO: case XFS_BTNUM_BNO:
case XFS_BTNUM_CNT: case XFS_BTNUM_CNT:
XFS_BUF_SET_VTYPE_REF(bp, B_FS_MAP, XFS_ALLOC_BTREE_REF); xfs_buf_set_ref(bp, XFS_ALLOC_BTREE_REF);
break; break;
case XFS_BTNUM_INO: case XFS_BTNUM_INO:
XFS_BUF_SET_VTYPE_REF(bp, B_FS_INOMAP, XFS_INO_BTREE_REF); xfs_buf_set_ref(bp, XFS_INO_BTREE_REF);
break; break;
case XFS_BTNUM_BMAP: case XFS_BTNUM_BMAP:
XFS_BUF_SET_VTYPE_REF(bp, B_FS_MAP, XFS_BMAP_BTREE_REF); xfs_buf_set_ref(bp, XFS_BMAP_BTREE_REF);
break; break;
default: default:
ASSERT(0); ASSERT(0);
@ -970,7 +970,8 @@ xfs_btree_get_buf_block(
*bpp = xfs_trans_get_buf(cur->bc_tp, mp->m_ddev_targp, d, *bpp = xfs_trans_get_buf(cur->bc_tp, mp->m_ddev_targp, d,
mp->m_bsize, flags); mp->m_bsize, flags);
ASSERT(!xfs_buf_geterror(*bpp)); if (!*bpp)
return ENOMEM;
*block = XFS_BUF_TO_BLOCK(*bpp); *block = XFS_BUF_TO_BLOCK(*bpp);
return 0; return 0;

View File

@ -43,7 +43,6 @@
static kmem_zone_t *xfs_buf_zone; static kmem_zone_t *xfs_buf_zone;
STATIC int xfsbufd(void *); STATIC int xfsbufd(void *);
STATIC void xfs_buf_delwri_queue(xfs_buf_t *, int);
static struct workqueue_struct *xfslogd_workqueue; static struct workqueue_struct *xfslogd_workqueue;
struct workqueue_struct *xfsdatad_workqueue; struct workqueue_struct *xfsdatad_workqueue;
@ -66,10 +65,6 @@ struct workqueue_struct *xfsconvertd_workqueue;
#define xb_to_km(flags) \ #define xb_to_km(flags) \
(((flags) & XBF_DONT_BLOCK) ? KM_NOFS : KM_SLEEP) (((flags) & XBF_DONT_BLOCK) ? KM_NOFS : KM_SLEEP)
#define xfs_buf_allocate(flags) \
kmem_zone_alloc(xfs_buf_zone, xb_to_km(flags))
#define xfs_buf_deallocate(bp) \
kmem_zone_free(xfs_buf_zone, (bp));
static inline int static inline int
xfs_buf_is_vmapped( xfs_buf_is_vmapped(
@ -152,6 +147,7 @@ xfs_buf_stale(
struct xfs_buf *bp) struct xfs_buf *bp)
{ {
bp->b_flags |= XBF_STALE; bp->b_flags |= XBF_STALE;
xfs_buf_delwri_dequeue(bp);
atomic_set(&(bp)->b_lru_ref, 0); atomic_set(&(bp)->b_lru_ref, 0);
if (!list_empty(&bp->b_lru)) { if (!list_empty(&bp->b_lru)) {
struct xfs_buftarg *btp = bp->b_target; struct xfs_buftarg *btp = bp->b_target;
@ -167,14 +163,19 @@ xfs_buf_stale(
ASSERT(atomic_read(&bp->b_hold) >= 1); ASSERT(atomic_read(&bp->b_hold) >= 1);
} }
STATIC void struct xfs_buf *
_xfs_buf_initialize( xfs_buf_alloc(
xfs_buf_t *bp, struct xfs_buftarg *target,
xfs_buftarg_t *target,
xfs_off_t range_base, xfs_off_t range_base,
size_t range_length, size_t range_length,
xfs_buf_flags_t flags) xfs_buf_flags_t flags)
{ {
struct xfs_buf *bp;
bp = kmem_zone_alloc(xfs_buf_zone, xb_to_km(flags));
if (unlikely(!bp))
return NULL;
/* /*
* We don't want certain flags to appear in b_flags. * We don't want certain flags to appear in b_flags.
*/ */
@ -203,8 +204,9 @@ _xfs_buf_initialize(
init_waitqueue_head(&bp->b_waiters); init_waitqueue_head(&bp->b_waiters);
XFS_STATS_INC(xb_create); XFS_STATS_INC(xb_create);
trace_xfs_buf_init(bp, _RET_IP_); trace_xfs_buf_init(bp, _RET_IP_);
return bp;
} }
/* /*
@ -277,7 +279,7 @@ xfs_buf_free(
} else if (bp->b_flags & _XBF_KMEM) } else if (bp->b_flags & _XBF_KMEM)
kmem_free(bp->b_addr); kmem_free(bp->b_addr);
_xfs_buf_free_pages(bp); _xfs_buf_free_pages(bp);
xfs_buf_deallocate(bp); kmem_zone_free(xfs_buf_zone, bp);
} }
/* /*
@ -416,10 +418,7 @@ _xfs_buf_map_pages(
/* /*
* Look up, and creates if absent, a lockable buffer for * Look up, and creates if absent, a lockable buffer for
* a given range of an inode. The buffer is returned * a given range of an inode. The buffer is returned
* locked. If other overlapping buffers exist, they are * locked. No I/O is implied by this call.
* released before the new buffer is created and locked,
* which may imply that this call will block until those buffers
* are unlocked. No I/O is implied by this call.
*/ */
xfs_buf_t * xfs_buf_t *
_xfs_buf_find( _xfs_buf_find(
@ -481,8 +480,6 @@ _xfs_buf_find(
/* No match found */ /* No match found */
if (new_bp) { if (new_bp) {
_xfs_buf_initialize(new_bp, btp, range_base,
range_length, flags);
rb_link_node(&new_bp->b_rbnode, parent, rbp); rb_link_node(&new_bp->b_rbnode, parent, rbp);
rb_insert_color(&new_bp->b_rbnode, &pag->pag_buf_tree); rb_insert_color(&new_bp->b_rbnode, &pag->pag_buf_tree);
/* the buffer keeps the perag reference until it is freed */ /* the buffer keeps the perag reference until it is freed */
@ -525,35 +522,51 @@ found:
} }
/* /*
* Assembles a buffer covering the specified range. * Assembles a buffer covering the specified range. The code is optimised for
* Storage in memory for all portions of the buffer will be allocated, * cache hits, as metadata intensive workloads will see 3 orders of magnitude
* although backing storage may not be. * more hits than misses.
*/ */
xfs_buf_t * struct xfs_buf *
xfs_buf_get( xfs_buf_get(
xfs_buftarg_t *target,/* target for buffer */ xfs_buftarg_t *target,/* target for buffer */
xfs_off_t ioff, /* starting offset of range */ xfs_off_t ioff, /* starting offset of range */
size_t isize, /* length of range */ size_t isize, /* length of range */
xfs_buf_flags_t flags) xfs_buf_flags_t flags)
{ {
xfs_buf_t *bp, *new_bp; struct xfs_buf *bp;
struct xfs_buf *new_bp;
int error = 0; int error = 0;
new_bp = xfs_buf_allocate(flags); bp = _xfs_buf_find(target, ioff, isize, flags, NULL);
if (likely(bp))
goto found;
new_bp = xfs_buf_alloc(target, ioff << BBSHIFT, isize << BBSHIFT,
flags);
if (unlikely(!new_bp)) if (unlikely(!new_bp))
return NULL; return NULL;
bp = _xfs_buf_find(target, ioff, isize, flags, new_bp); bp = _xfs_buf_find(target, ioff, isize, flags, new_bp);
if (!bp) {
kmem_zone_free(xfs_buf_zone, new_bp);
return NULL;
}
if (bp == new_bp) { if (bp == new_bp) {
error = xfs_buf_allocate_memory(bp, flags); error = xfs_buf_allocate_memory(bp, flags);
if (error) if (error)
goto no_buffer; goto no_buffer;
} else { } else
xfs_buf_deallocate(new_bp); kmem_zone_free(xfs_buf_zone, new_bp);
if (unlikely(bp == NULL))
return NULL;
}
/*
* Now we have a workable buffer, fill in the block number so
* that we can do IO on it.
*/
bp->b_bn = ioff;
bp->b_count_desired = bp->b_buffer_length;
found:
if (!(bp->b_flags & XBF_MAPPED)) { if (!(bp->b_flags & XBF_MAPPED)) {
error = _xfs_buf_map_pages(bp, flags); error = _xfs_buf_map_pages(bp, flags);
if (unlikely(error)) { if (unlikely(error)) {
@ -564,18 +577,10 @@ xfs_buf_get(
} }
XFS_STATS_INC(xb_get); XFS_STATS_INC(xb_get);
/*
* Always fill in the block number now, the mapped cases can do
* their own overlay of this later.
*/
bp->b_bn = ioff;
bp->b_count_desired = bp->b_buffer_length;
trace_xfs_buf_get(bp, flags, _RET_IP_); trace_xfs_buf_get(bp, flags, _RET_IP_);
return bp; return bp;
no_buffer: no_buffer:
if (flags & (XBF_LOCK | XBF_TRYLOCK)) if (flags & (XBF_LOCK | XBF_TRYLOCK))
xfs_buf_unlock(bp); xfs_buf_unlock(bp);
xfs_buf_rele(bp); xfs_buf_rele(bp);
@ -689,19 +694,6 @@ xfs_buf_read_uncached(
return bp; return bp;
} }
xfs_buf_t *
xfs_buf_get_empty(
size_t len,
xfs_buftarg_t *target)
{
xfs_buf_t *bp;
bp = xfs_buf_allocate(0);
if (bp)
_xfs_buf_initialize(bp, target, 0, len, 0);
return bp;
}
/* /*
* Return a buffer allocated as an empty buffer and associated to external * Return a buffer allocated as an empty buffer and associated to external
* memory via xfs_buf_associate_memory() back to it's empty state. * memory via xfs_buf_associate_memory() back to it's empty state.
@ -787,10 +779,9 @@ xfs_buf_get_uncached(
int error, i; int error, i;
xfs_buf_t *bp; xfs_buf_t *bp;
bp = xfs_buf_allocate(0); bp = xfs_buf_alloc(target, 0, len, 0);
if (unlikely(bp == NULL)) if (unlikely(bp == NULL))
goto fail; goto fail;
_xfs_buf_initialize(bp, target, 0, len, 0);
error = _xfs_buf_get_pages(bp, page_count, 0); error = _xfs_buf_get_pages(bp, page_count, 0);
if (error) if (error)
@ -818,7 +809,7 @@ xfs_buf_get_uncached(
__free_page(bp->b_pages[i]); __free_page(bp->b_pages[i]);
_xfs_buf_free_pages(bp); _xfs_buf_free_pages(bp);
fail_free_buf: fail_free_buf:
xfs_buf_deallocate(bp); kmem_zone_free(xfs_buf_zone, bp);
fail: fail:
return NULL; return NULL;
} }
@ -937,12 +928,6 @@ void
xfs_buf_unlock( xfs_buf_unlock(
struct xfs_buf *bp) struct xfs_buf *bp)
{ {
if ((bp->b_flags & (XBF_DELWRI|_XBF_DELWRI_Q)) == XBF_DELWRI) {
atomic_inc(&bp->b_hold);
bp->b_flags |= XBF_ASYNC;
xfs_buf_delwri_queue(bp, 0);
}
XB_CLEAR_OWNER(bp); XB_CLEAR_OWNER(bp);
up(&bp->b_sema); up(&bp->b_sema);
@ -1019,9 +1004,19 @@ xfs_buf_ioerror(
trace_xfs_buf_ioerror(bp, error, _RET_IP_); trace_xfs_buf_ioerror(bp, error, _RET_IP_);
} }
void
xfs_buf_ioerror_alert(
struct xfs_buf *bp,
const char *func)
{
xfs_alert(bp->b_target->bt_mount,
"metadata I/O error: block 0x%llx (\"%s\") error %d buf count %zd",
(__uint64_t)XFS_BUF_ADDR(bp), func,
bp->b_error, XFS_BUF_COUNT(bp));
}
int int
xfs_bwrite( xfs_bwrite(
struct xfs_mount *mp,
struct xfs_buf *bp) struct xfs_buf *bp)
{ {
int error; int error;
@ -1033,25 +1028,13 @@ xfs_bwrite(
xfs_bdstrat_cb(bp); xfs_bdstrat_cb(bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) if (error) {
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR); xfs_force_shutdown(bp->b_target->bt_mount,
xfs_buf_relse(bp); SHUTDOWN_META_IO_ERROR);
}
return error; return error;
} }
void
xfs_bdwrite(
void *mp,
struct xfs_buf *bp)
{
trace_xfs_buf_bdwrite(bp, _RET_IP_);
bp->b_flags &= ~XBF_READ;
bp->b_flags |= (XBF_DELWRI | XBF_ASYNC);
xfs_buf_delwri_queue(bp, 1);
}
/* /*
* Called when we want to stop a buffer from getting written or read. * Called when we want to stop a buffer from getting written or read.
* We attach the EIO error, muck with its flags, and call xfs_buf_ioend * We attach the EIO error, muck with its flags, and call xfs_buf_ioend
@ -1074,9 +1057,8 @@ xfs_bioerror(
* We're calling xfs_buf_ioend, so delete XBF_DONE flag. * We're calling xfs_buf_ioend, so delete XBF_DONE flag.
*/ */
XFS_BUF_UNREAD(bp); XFS_BUF_UNREAD(bp);
XFS_BUF_UNDELAYWRITE(bp);
XFS_BUF_UNDONE(bp); XFS_BUF_UNDONE(bp);
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
xfs_buf_ioend(bp, 0); xfs_buf_ioend(bp, 0);
@ -1103,9 +1085,8 @@ xfs_bioerror_relse(
* change that interface. * change that interface.
*/ */
XFS_BUF_UNREAD(bp); XFS_BUF_UNREAD(bp);
XFS_BUF_UNDELAYWRITE(bp);
XFS_BUF_DONE(bp); XFS_BUF_DONE(bp);
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
bp->b_iodone = NULL; bp->b_iodone = NULL;
if (!(fl & XBF_ASYNC)) { if (!(fl & XBF_ASYNC)) {
/* /*
@ -1115,7 +1096,7 @@ xfs_bioerror_relse(
* ASYNC buffers. * ASYNC buffers.
*/ */
xfs_buf_ioerror(bp, EIO); xfs_buf_ioerror(bp, EIO);
XFS_BUF_FINISH_IOWAIT(bp); complete(&bp->b_iowait);
} else { } else {
xfs_buf_relse(bp); xfs_buf_relse(bp);
} }
@ -1275,15 +1256,10 @@ xfs_buf_iorequest(
{ {
trace_xfs_buf_iorequest(bp, _RET_IP_); trace_xfs_buf_iorequest(bp, _RET_IP_);
if (bp->b_flags & XBF_DELWRI) { ASSERT(!(bp->b_flags & XBF_DELWRI));
xfs_buf_delwri_queue(bp, 1);
return 0;
}
if (bp->b_flags & XBF_WRITE) { if (bp->b_flags & XBF_WRITE)
xfs_buf_wait_unpin(bp); xfs_buf_wait_unpin(bp);
}
xfs_buf_hold(bp); xfs_buf_hold(bp);
/* Set the count to 1 initially, this will stop an I/O /* Set the count to 1 initially, this will stop an I/O
@ -1481,9 +1457,13 @@ xfs_setsize_buftarg_flags(
btp->bt_smask = sectorsize - 1; btp->bt_smask = sectorsize - 1;
if (set_blocksize(btp->bt_bdev, sectorsize)) { if (set_blocksize(btp->bt_bdev, sectorsize)) {
char name[BDEVNAME_SIZE];
bdevname(btp->bt_bdev, name);
xfs_warn(btp->bt_mount, xfs_warn(btp->bt_mount,
"Cannot set_blocksize to %u on device %s\n", "Cannot set_blocksize to %u on device %s\n",
sectorsize, xfs_buf_target_name(btp)); sectorsize, name);
return EINVAL; return EINVAL;
} }
@ -1514,12 +1494,12 @@ xfs_setsize_buftarg(
} }
STATIC int STATIC int
xfs_alloc_delwrite_queue( xfs_alloc_delwri_queue(
xfs_buftarg_t *btp, xfs_buftarg_t *btp,
const char *fsname) const char *fsname)
{ {
INIT_LIST_HEAD(&btp->bt_delwrite_queue); INIT_LIST_HEAD(&btp->bt_delwri_queue);
spin_lock_init(&btp->bt_delwrite_lock); spin_lock_init(&btp->bt_delwri_lock);
btp->bt_flags = 0; btp->bt_flags = 0;
btp->bt_task = kthread_run(xfsbufd, btp, "xfsbufd/%s", fsname); btp->bt_task = kthread_run(xfsbufd, btp, "xfsbufd/%s", fsname);
if (IS_ERR(btp->bt_task)) if (IS_ERR(btp->bt_task))
@ -1549,7 +1529,7 @@ xfs_alloc_buftarg(
spin_lock_init(&btp->bt_lru_lock); spin_lock_init(&btp->bt_lru_lock);
if (xfs_setsize_buftarg_early(btp, bdev)) if (xfs_setsize_buftarg_early(btp, bdev))
goto error; goto error;
if (xfs_alloc_delwrite_queue(btp, fsname)) if (xfs_alloc_delwri_queue(btp, fsname))
goto error; goto error;
btp->bt_shrinker.shrink = xfs_buftarg_shrink; btp->bt_shrinker.shrink = xfs_buftarg_shrink;
btp->bt_shrinker.seeks = DEFAULT_SEEKS; btp->bt_shrinker.seeks = DEFAULT_SEEKS;
@ -1565,56 +1545,48 @@ error:
/* /*
* Delayed write buffer handling * Delayed write buffer handling
*/ */
STATIC void void
xfs_buf_delwri_queue( xfs_buf_delwri_queue(
xfs_buf_t *bp, xfs_buf_t *bp)
int unlock)
{ {
struct list_head *dwq = &bp->b_target->bt_delwrite_queue; struct xfs_buftarg *btp = bp->b_target;
spinlock_t *dwlk = &bp->b_target->bt_delwrite_lock;
trace_xfs_buf_delwri_queue(bp, _RET_IP_); trace_xfs_buf_delwri_queue(bp, _RET_IP_);
ASSERT((bp->b_flags&(XBF_DELWRI|XBF_ASYNC)) == (XBF_DELWRI|XBF_ASYNC)); ASSERT(!(bp->b_flags & XBF_READ));
spin_lock(dwlk); spin_lock(&btp->bt_delwri_lock);
/* If already in the queue, dequeue and place at tail */
if (!list_empty(&bp->b_list)) { if (!list_empty(&bp->b_list)) {
/* if already in the queue, move it to the tail */
ASSERT(bp->b_flags & _XBF_DELWRI_Q); ASSERT(bp->b_flags & _XBF_DELWRI_Q);
if (unlock) list_move_tail(&bp->b_list, &btp->bt_delwri_queue);
atomic_dec(&bp->b_hold); } else {
list_del(&bp->b_list);
}
if (list_empty(dwq)) {
/* start xfsbufd as it is about to have something to do */ /* start xfsbufd as it is about to have something to do */
wake_up_process(bp->b_target->bt_task); if (list_empty(&btp->bt_delwri_queue))
wake_up_process(bp->b_target->bt_task);
atomic_inc(&bp->b_hold);
bp->b_flags |= XBF_DELWRI | _XBF_DELWRI_Q | XBF_ASYNC;
list_add_tail(&bp->b_list, &btp->bt_delwri_queue);
} }
bp->b_flags |= _XBF_DELWRI_Q;
list_add_tail(&bp->b_list, dwq);
bp->b_queuetime = jiffies; bp->b_queuetime = jiffies;
spin_unlock(dwlk); spin_unlock(&btp->bt_delwri_lock);
if (unlock)
xfs_buf_unlock(bp);
} }
void void
xfs_buf_delwri_dequeue( xfs_buf_delwri_dequeue(
xfs_buf_t *bp) xfs_buf_t *bp)
{ {
spinlock_t *dwlk = &bp->b_target->bt_delwrite_lock;
int dequeued = 0; int dequeued = 0;
spin_lock(dwlk); spin_lock(&bp->b_target->bt_delwri_lock);
if ((bp->b_flags & XBF_DELWRI) && !list_empty(&bp->b_list)) { if ((bp->b_flags & XBF_DELWRI) && !list_empty(&bp->b_list)) {
ASSERT(bp->b_flags & _XBF_DELWRI_Q); ASSERT(bp->b_flags & _XBF_DELWRI_Q);
list_del_init(&bp->b_list); list_del_init(&bp->b_list);
dequeued = 1; dequeued = 1;
} }
bp->b_flags &= ~(XBF_DELWRI|_XBF_DELWRI_Q); bp->b_flags &= ~(XBF_DELWRI|_XBF_DELWRI_Q);
spin_unlock(dwlk); spin_unlock(&bp->b_target->bt_delwri_lock);
if (dequeued) if (dequeued)
xfs_buf_rele(bp); xfs_buf_rele(bp);
@ -1646,16 +1618,9 @@ xfs_buf_delwri_promote(
if (bp->b_queuetime < jiffies - age) if (bp->b_queuetime < jiffies - age)
return; return;
bp->b_queuetime = jiffies - age; bp->b_queuetime = jiffies - age;
spin_lock(&btp->bt_delwrite_lock); spin_lock(&btp->bt_delwri_lock);
list_move(&bp->b_list, &btp->bt_delwrite_queue); list_move(&bp->b_list, &btp->bt_delwri_queue);
spin_unlock(&btp->bt_delwrite_lock); spin_unlock(&btp->bt_delwri_lock);
}
STATIC void
xfs_buf_runall_queues(
struct workqueue_struct *queue)
{
flush_workqueue(queue);
} }
/* /*
@ -1669,15 +1634,13 @@ xfs_buf_delwri_split(
unsigned long age) unsigned long age)
{ {
xfs_buf_t *bp, *n; xfs_buf_t *bp, *n;
struct list_head *dwq = &target->bt_delwrite_queue;
spinlock_t *dwlk = &target->bt_delwrite_lock;
int skipped = 0; int skipped = 0;
int force; int force;
force = test_and_clear_bit(XBT_FORCE_FLUSH, &target->bt_flags); force = test_and_clear_bit(XBT_FORCE_FLUSH, &target->bt_flags);
INIT_LIST_HEAD(list); INIT_LIST_HEAD(list);
spin_lock(dwlk); spin_lock(&target->bt_delwri_lock);
list_for_each_entry_safe(bp, n, dwq, b_list) { list_for_each_entry_safe(bp, n, &target->bt_delwri_queue, b_list) {
ASSERT(bp->b_flags & XBF_DELWRI); ASSERT(bp->b_flags & XBF_DELWRI);
if (!xfs_buf_ispinned(bp) && xfs_buf_trylock(bp)) { if (!xfs_buf_ispinned(bp) && xfs_buf_trylock(bp)) {
@ -1694,10 +1657,9 @@ xfs_buf_delwri_split(
} else } else
skipped++; skipped++;
} }
spin_unlock(dwlk);
spin_unlock(&target->bt_delwri_lock);
return skipped; return skipped;
} }
/* /*
@ -1747,7 +1709,7 @@ xfsbufd(
} }
/* sleep for a long time if there is nothing to do. */ /* sleep for a long time if there is nothing to do. */
if (list_empty(&target->bt_delwrite_queue)) if (list_empty(&target->bt_delwri_queue))
tout = MAX_SCHEDULE_TIMEOUT; tout = MAX_SCHEDULE_TIMEOUT;
schedule_timeout_interruptible(tout); schedule_timeout_interruptible(tout);
@ -1783,9 +1745,7 @@ xfs_flush_buftarg(
LIST_HEAD(wait_list); LIST_HEAD(wait_list);
struct blk_plug plug; struct blk_plug plug;
xfs_buf_runall_queues(xfsconvertd_workqueue); flush_workqueue(xfslogd_workqueue);
xfs_buf_runall_queues(xfsdatad_workqueue);
xfs_buf_runall_queues(xfslogd_workqueue);
set_bit(XBT_FORCE_FLUSH, &target->bt_flags); set_bit(XBT_FORCE_FLUSH, &target->bt_flags);
pincount = xfs_buf_delwri_split(target, &tmp_list, 0); pincount = xfs_buf_delwri_split(target, &tmp_list, 0);
@ -1866,11 +1826,3 @@ xfs_buf_terminate(void)
destroy_workqueue(xfslogd_workqueue); destroy_workqueue(xfslogd_workqueue);
kmem_zone_destroy(xfs_buf_zone); kmem_zone_destroy(xfs_buf_zone);
} }
#ifdef CONFIG_KDB_MODULES
struct list_head *
xfs_get_buftarg_list(void)
{
return &xfs_buftarg_list;
}
#endif

View File

@ -105,8 +105,8 @@ typedef struct xfs_buftarg {
/* per device delwri queue */ /* per device delwri queue */
struct task_struct *bt_task; struct task_struct *bt_task;
struct list_head bt_delwrite_queue; struct list_head bt_delwri_queue;
spinlock_t bt_delwrite_lock; spinlock_t bt_delwri_lock;
unsigned long bt_flags; unsigned long bt_flags;
/* LRU control structures */ /* LRU control structures */
@ -175,7 +175,8 @@ extern xfs_buf_t *xfs_buf_get(xfs_buftarg_t *, xfs_off_t, size_t,
extern xfs_buf_t *xfs_buf_read(xfs_buftarg_t *, xfs_off_t, size_t, extern xfs_buf_t *xfs_buf_read(xfs_buftarg_t *, xfs_off_t, size_t,
xfs_buf_flags_t); xfs_buf_flags_t);
extern xfs_buf_t *xfs_buf_get_empty(size_t, xfs_buftarg_t *); struct xfs_buf *xfs_buf_alloc(struct xfs_buftarg *, xfs_off_t, size_t,
xfs_buf_flags_t);
extern void xfs_buf_set_empty(struct xfs_buf *bp, size_t len); extern void xfs_buf_set_empty(struct xfs_buf *bp, size_t len);
extern xfs_buf_t *xfs_buf_get_uncached(struct xfs_buftarg *, size_t, int); extern xfs_buf_t *xfs_buf_get_uncached(struct xfs_buftarg *, size_t, int);
extern int xfs_buf_associate_memory(xfs_buf_t *, void *, size_t); extern int xfs_buf_associate_memory(xfs_buf_t *, void *, size_t);
@ -197,14 +198,14 @@ extern void xfs_buf_unlock(xfs_buf_t *);
((bp)->b_sema.count <= 0) ((bp)->b_sema.count <= 0)
/* Buffer Read and Write Routines */ /* Buffer Read and Write Routines */
extern int xfs_bwrite(struct xfs_mount *mp, struct xfs_buf *bp); extern int xfs_bwrite(struct xfs_buf *bp);
extern void xfs_bdwrite(void *mp, xfs_buf_t *bp);
extern void xfsbdstrat(struct xfs_mount *, struct xfs_buf *); extern void xfsbdstrat(struct xfs_mount *, struct xfs_buf *);
extern int xfs_bdstrat_cb(struct xfs_buf *); extern int xfs_bdstrat_cb(struct xfs_buf *);
extern void xfs_buf_ioend(xfs_buf_t *, int); extern void xfs_buf_ioend(xfs_buf_t *, int);
extern void xfs_buf_ioerror(xfs_buf_t *, int); extern void xfs_buf_ioerror(xfs_buf_t *, int);
extern void xfs_buf_ioerror_alert(struct xfs_buf *, const char *func);
extern int xfs_buf_iorequest(xfs_buf_t *); extern int xfs_buf_iorequest(xfs_buf_t *);
extern int xfs_buf_iowait(xfs_buf_t *); extern int xfs_buf_iowait(xfs_buf_t *);
extern void xfs_buf_iomove(xfs_buf_t *, size_t, size_t, void *, extern void xfs_buf_iomove(xfs_buf_t *, size_t, size_t, void *,
@ -221,38 +222,22 @@ static inline int xfs_buf_geterror(xfs_buf_t *bp)
extern xfs_caddr_t xfs_buf_offset(xfs_buf_t *, size_t); extern xfs_caddr_t xfs_buf_offset(xfs_buf_t *, size_t);
/* Delayed Write Buffer Routines */ /* Delayed Write Buffer Routines */
extern void xfs_buf_delwri_dequeue(xfs_buf_t *); extern void xfs_buf_delwri_queue(struct xfs_buf *);
extern void xfs_buf_delwri_promote(xfs_buf_t *); extern void xfs_buf_delwri_dequeue(struct xfs_buf *);
extern void xfs_buf_delwri_promote(struct xfs_buf *);
/* Buffer Daemon Setup Routines */ /* Buffer Daemon Setup Routines */
extern int xfs_buf_init(void); extern int xfs_buf_init(void);
extern void xfs_buf_terminate(void); extern void xfs_buf_terminate(void);
static inline const char *
xfs_buf_target_name(struct xfs_buftarg *target)
{
static char __b[BDEVNAME_SIZE];
return bdevname(target->bt_bdev, __b);
}
#define XFS_BUF_ZEROFLAGS(bp) \ #define XFS_BUF_ZEROFLAGS(bp) \
((bp)->b_flags &= ~(XBF_READ|XBF_WRITE|XBF_ASYNC|XBF_DELWRI| \ ((bp)->b_flags &= ~(XBF_READ|XBF_WRITE|XBF_ASYNC|XBF_DELWRI| \
XBF_SYNCIO|XBF_FUA|XBF_FLUSH)) XBF_SYNCIO|XBF_FUA|XBF_FLUSH))
void xfs_buf_stale(struct xfs_buf *bp); void xfs_buf_stale(struct xfs_buf *bp);
#define XFS_BUF_STALE(bp) xfs_buf_stale(bp);
#define XFS_BUF_UNSTALE(bp) ((bp)->b_flags &= ~XBF_STALE) #define XFS_BUF_UNSTALE(bp) ((bp)->b_flags &= ~XBF_STALE)
#define XFS_BUF_ISSTALE(bp) ((bp)->b_flags & XBF_STALE) #define XFS_BUF_ISSTALE(bp) ((bp)->b_flags & XBF_STALE)
#define XFS_BUF_SUPER_STALE(bp) do { \
XFS_BUF_STALE(bp); \
xfs_buf_delwri_dequeue(bp); \
XFS_BUF_DONE(bp); \
} while (0)
#define XFS_BUF_DELAYWRITE(bp) ((bp)->b_flags |= XBF_DELWRI)
#define XFS_BUF_UNDELAYWRITE(bp) xfs_buf_delwri_dequeue(bp)
#define XFS_BUF_ISDELAYWRITE(bp) ((bp)->b_flags & XBF_DELWRI) #define XFS_BUF_ISDELAYWRITE(bp) ((bp)->b_flags & XBF_DELWRI)
#define XFS_BUF_DONE(bp) ((bp)->b_flags |= XBF_DONE) #define XFS_BUF_DONE(bp) ((bp)->b_flags |= XBF_DONE)
@ -280,23 +265,16 @@ void xfs_buf_stale(struct xfs_buf *bp);
#define XFS_BUF_SIZE(bp) ((bp)->b_buffer_length) #define XFS_BUF_SIZE(bp) ((bp)->b_buffer_length)
#define XFS_BUF_SET_SIZE(bp, cnt) ((bp)->b_buffer_length = (cnt)) #define XFS_BUF_SET_SIZE(bp, cnt) ((bp)->b_buffer_length = (cnt))
static inline void static inline void xfs_buf_set_ref(struct xfs_buf *bp, int lru_ref)
xfs_buf_set_ref(
struct xfs_buf *bp,
int lru_ref)
{ {
atomic_set(&bp->b_lru_ref, lru_ref); atomic_set(&bp->b_lru_ref, lru_ref);
} }
#define XFS_BUF_SET_VTYPE_REF(bp, type, ref) xfs_buf_set_ref(bp, ref)
#define XFS_BUF_SET_VTYPE(bp, type) do { } while (0)
static inline int xfs_buf_ispinned(struct xfs_buf *bp) static inline int xfs_buf_ispinned(struct xfs_buf *bp)
{ {
return atomic_read(&bp->b_pin_count); return atomic_read(&bp->b_pin_count);
} }
#define XFS_BUF_FINISH_IOWAIT(bp) complete(&bp->b_iowait);
static inline void xfs_buf_relse(xfs_buf_t *bp) static inline void xfs_buf_relse(xfs_buf_t *bp)
{ {
xfs_buf_unlock(bp); xfs_buf_unlock(bp);
@ -313,14 +291,7 @@ extern void xfs_wait_buftarg(xfs_buftarg_t *);
extern int xfs_setsize_buftarg(xfs_buftarg_t *, unsigned int, unsigned int); extern int xfs_setsize_buftarg(xfs_buftarg_t *, unsigned int, unsigned int);
extern int xfs_flush_buftarg(xfs_buftarg_t *, int); extern int xfs_flush_buftarg(xfs_buftarg_t *, int);
#ifdef CONFIG_KDB_MODULES
extern struct list_head *xfs_get_buftarg_list(void);
#endif
#define xfs_getsize_buftarg(buftarg) block_size((buftarg)->bt_bdev) #define xfs_getsize_buftarg(buftarg) block_size((buftarg)->bt_bdev)
#define xfs_readonly_buftarg(buftarg) bdev_read_only((buftarg)->bt_bdev) #define xfs_readonly_buftarg(buftarg) bdev_read_only((buftarg)->bt_bdev)
#define xfs_binval(buftarg) xfs_flush_buftarg(buftarg, 1)
#define XFS_bflush(buftarg) xfs_flush_buftarg(buftarg, 1)
#endif /* __XFS_BUF_H__ */ #endif /* __XFS_BUF_H__ */

View File

@ -967,7 +967,8 @@ xfs_buf_iodone_callbacks(
* I/O errors, there's no point in giving this a retry. * I/O errors, there's no point in giving this a retry.
*/ */
if (XFS_FORCED_SHUTDOWN(mp)) { if (XFS_FORCED_SHUTDOWN(mp)) {
XFS_BUF_SUPER_STALE(bp); xfs_buf_stale(bp);
XFS_BUF_DONE(bp);
trace_xfs_buf_item_iodone(bp, _RET_IP_); trace_xfs_buf_item_iodone(bp, _RET_IP_);
goto do_callbacks; goto do_callbacks;
} }
@ -975,9 +976,7 @@ xfs_buf_iodone_callbacks(
if (bp->b_target != lasttarg || if (bp->b_target != lasttarg ||
time_after(jiffies, (lasttime + 5*HZ))) { time_after(jiffies, (lasttime + 5*HZ))) {
lasttime = jiffies; lasttime = jiffies;
xfs_alert(mp, "Device %s: metadata write error block 0x%llx", xfs_buf_ioerror_alert(bp, __func__);
xfs_buf_target_name(bp->b_target),
(__uint64_t)XFS_BUF_ADDR(bp));
} }
lasttarg = bp->b_target; lasttarg = bp->b_target;
@ -993,7 +992,7 @@ xfs_buf_iodone_callbacks(
xfs_buf_ioerror(bp, 0); /* errno of 0 unsets the flag */ xfs_buf_ioerror(bp, 0); /* errno of 0 unsets the flag */
if (!XFS_BUF_ISSTALE(bp)) { if (!XFS_BUF_ISSTALE(bp)) {
XFS_BUF_DELAYWRITE(bp); xfs_buf_delwri_queue(bp);
XFS_BUF_DONE(bp); XFS_BUF_DONE(bp);
} }
ASSERT(bp->b_iodone != NULL); ASSERT(bp->b_iodone != NULL);
@ -1006,9 +1005,8 @@ xfs_buf_iodone_callbacks(
* If the write of the buffer was synchronous, we want to make * If the write of the buffer was synchronous, we want to make
* sure to return the error to the caller of xfs_bwrite(). * sure to return the error to the caller of xfs_bwrite().
*/ */
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
XFS_BUF_DONE(bp); XFS_BUF_DONE(bp);
XFS_BUF_UNDELAYWRITE(bp);
trace_xfs_buf_error_relse(bp, _RET_IP_); trace_xfs_buf_error_relse(bp, _RET_IP_);

View File

@ -1578,9 +1578,8 @@ xfs_da_grow_inode_int(
*/ */
nmap = 1; nmap = 1;
ASSERT(args->firstblock != NULL); ASSERT(args->firstblock != NULL);
error = xfs_bmapi(tp, dp, *bno, count, error = xfs_bmapi_write(tp, dp, *bno, count,
xfs_bmapi_aflag(w)|XFS_BMAPI_WRITE|XFS_BMAPI_METADATA| xfs_bmapi_aflag(w)|XFS_BMAPI_METADATA|XFS_BMAPI_CONTIG,
XFS_BMAPI_CONTIG,
args->firstblock, args->total, &map, &nmap, args->firstblock, args->total, &map, &nmap,
args->flist); args->flist);
if (error) if (error)
@ -1602,9 +1601,8 @@ xfs_da_grow_inode_int(
for (b = *bno, mapi = 0; b < *bno + count; ) { for (b = *bno, mapi = 0; b < *bno + count; ) {
nmap = MIN(XFS_BMAP_MAX_NMAP, count); nmap = MIN(XFS_BMAP_MAX_NMAP, count);
c = (int)(*bno + count - b); c = (int)(*bno + count - b);
error = xfs_bmapi(tp, dp, b, c, error = xfs_bmapi_write(tp, dp, b, c,
xfs_bmapi_aflag(w)|XFS_BMAPI_WRITE| xfs_bmapi_aflag(w)|XFS_BMAPI_METADATA,
XFS_BMAPI_METADATA,
args->firstblock, args->total, args->firstblock, args->total,
&mapp[mapi], &nmap, args->flist); &mapp[mapi], &nmap, args->flist);
if (error) if (error)
@ -1975,33 +1973,16 @@ xfs_da_do_buf(
/* /*
* Optimize the one-block case. * Optimize the one-block case.
*/ */
if (nfsb == 1) { if (nfsb == 1)
xfs_fsblock_t fsb;
if ((error =
xfs_bmapi_single(trans, dp, whichfork, &fsb,
(xfs_fileoff_t)bno))) {
return error;
}
mapp = &map; mapp = &map;
if (fsb == NULLFSBLOCK) { else
nmap = 0;
} else {
map.br_startblock = fsb;
map.br_startoff = (xfs_fileoff_t)bno;
map.br_blockcount = 1;
nmap = 1;
}
} else {
mapp = kmem_alloc(sizeof(*mapp) * nfsb, KM_SLEEP); mapp = kmem_alloc(sizeof(*mapp) * nfsb, KM_SLEEP);
nmap = nfsb;
if ((error = xfs_bmapi(trans, dp, (xfs_fileoff_t)bno, nmap = nfsb;
nfsb, error = xfs_bmapi_read(dp, (xfs_fileoff_t)bno, nfsb, mapp,
XFS_BMAPI_METADATA | &nmap, xfs_bmapi_aflag(whichfork));
xfs_bmapi_aflag(whichfork), if (error)
NULL, 0, mapp, &nmap, NULL))) goto exit0;
goto exit0;
}
} else { } else {
map.br_startblock = XFS_DADDR_TO_FSB(mp, mappedbno); map.br_startblock = XFS_DADDR_TO_FSB(mp, mappedbno);
map.br_startoff = (xfs_fileoff_t)bno; map.br_startoff = (xfs_fileoff_t)bno;
@ -2072,13 +2053,10 @@ xfs_da_do_buf(
if (!bp) if (!bp)
continue; continue;
if (caller == 1) { if (caller == 1) {
if (whichfork == XFS_ATTR_FORK) { if (whichfork == XFS_ATTR_FORK)
XFS_BUF_SET_VTYPE_REF(bp, B_FS_ATTR_BTREE, xfs_buf_set_ref(bp, XFS_ATTR_BTREE_REF);
XFS_ATTR_BTREE_REF); else
} else { xfs_buf_set_ref(bp, XFS_DIR_BTREE_REF);
XFS_BUF_SET_VTYPE_REF(bp, B_FS_DIR_BTREE,
XFS_DIR_BTREE_REF);
}
} }
if (bplist) { if (bplist) {
bplist[nbplist++] = bp; bplist[nbplist++] = bp;

View File

@ -425,8 +425,8 @@ xfs_swap_extents(
} }
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL);
xfs_trans_ijoin_ref(tp, tip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL); xfs_trans_ijoin(tp, tip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL);
xfs_trans_log_inode(tp, ip, ilf_fields); xfs_trans_log_inode(tp, ip, ilf_fields);
xfs_trans_log_inode(tp, tip, tilf_fields); xfs_trans_log_inode(tp, tip, tilf_fields);
@ -438,7 +438,7 @@ xfs_swap_extents(
if (mp->m_flags & XFS_MOUNT_WSYNC) if (mp->m_flags & XFS_MOUNT_WSYNC)
xfs_trans_set_sync(tp); xfs_trans_set_sync(tp);
error = xfs_trans_commit(tp, XFS_TRANS_SWAPEXT); error = xfs_trans_commit(tp, 0);
trace_xfs_swap_extent_after(ip, 0); trace_xfs_swap_extent_after(ip, 0);
trace_xfs_swap_extent_after(tip, 1); trace_xfs_swap_extent_after(tip, 1);

View File

@ -888,12 +888,10 @@ xfs_dir2_leaf_getdents(
* we already have in the table. * we already have in the table.
*/ */
nmap = map_size - map_valid; nmap = map_size - map_valid;
error = xfs_bmapi(NULL, dp, error = xfs_bmapi_read(dp, map_off,
map_off,
xfs_dir2_byte_to_da(mp, xfs_dir2_byte_to_da(mp,
XFS_DIR2_LEAF_OFFSET) - map_off, XFS_DIR2_LEAF_OFFSET) - map_off,
XFS_BMAPI_METADATA, NULL, 0, &map[map_valid], &nmap, 0);
&map[map_valid], &nmap, NULL);
/* /*
* Don't know if we should ignore this or * Don't know if we should ignore this or
* try to return an error. * try to return an error.

View File

@ -38,7 +38,7 @@ xfs_trim_extents(
struct xfs_mount *mp, struct xfs_mount *mp,
xfs_agnumber_t agno, xfs_agnumber_t agno,
xfs_fsblock_t start, xfs_fsblock_t start,
xfs_fsblock_t len, xfs_fsblock_t end,
xfs_fsblock_t minlen, xfs_fsblock_t minlen,
__uint64_t *blocks_trimmed) __uint64_t *blocks_trimmed)
{ {
@ -100,7 +100,7 @@ xfs_trim_extents(
* down partially overlapping ranges for now. * down partially overlapping ranges for now.
*/ */
if (XFS_AGB_TO_FSB(mp, agno, fbno) + flen < start || if (XFS_AGB_TO_FSB(mp, agno, fbno) + flen < start ||
XFS_AGB_TO_FSB(mp, agno, fbno) >= start + len) { XFS_AGB_TO_FSB(mp, agno, fbno) > end) {
trace_xfs_discard_exclude(mp, agno, fbno, flen); trace_xfs_discard_exclude(mp, agno, fbno, flen);
goto next_extent; goto next_extent;
} }
@ -145,7 +145,7 @@ xfs_ioc_trim(
struct request_queue *q = mp->m_ddev_targp->bt_bdev->bd_disk->queue; struct request_queue *q = mp->m_ddev_targp->bt_bdev->bd_disk->queue;
unsigned int granularity = q->limits.discard_granularity; unsigned int granularity = q->limits.discard_granularity;
struct fstrim_range range; struct fstrim_range range;
xfs_fsblock_t start, len, minlen; xfs_fsblock_t start, end, minlen;
xfs_agnumber_t start_agno, end_agno, agno; xfs_agnumber_t start_agno, end_agno, agno;
__uint64_t blocks_trimmed = 0; __uint64_t blocks_trimmed = 0;
int error, last_error = 0; int error, last_error = 0;
@ -165,19 +165,19 @@ xfs_ioc_trim(
* matter as trimming blocks is an advisory interface. * matter as trimming blocks is an advisory interface.
*/ */
start = XFS_B_TO_FSBT(mp, range.start); start = XFS_B_TO_FSBT(mp, range.start);
len = XFS_B_TO_FSBT(mp, range.len); end = start + XFS_B_TO_FSBT(mp, range.len) - 1;
minlen = XFS_B_TO_FSB(mp, max_t(u64, granularity, range.minlen)); minlen = XFS_B_TO_FSB(mp, max_t(u64, granularity, range.minlen));
start_agno = XFS_FSB_TO_AGNO(mp, start); if (start >= mp->m_sb.sb_dblocks)
if (start_agno >= mp->m_sb.sb_agcount)
return -XFS_ERROR(EINVAL); return -XFS_ERROR(EINVAL);
if (end > mp->m_sb.sb_dblocks - 1)
end = mp->m_sb.sb_dblocks - 1;
end_agno = XFS_FSB_TO_AGNO(mp, start + len); start_agno = XFS_FSB_TO_AGNO(mp, start);
if (end_agno >= mp->m_sb.sb_agcount) end_agno = XFS_FSB_TO_AGNO(mp, end);
end_agno = mp->m_sb.sb_agcount - 1;
for (agno = start_agno; agno <= end_agno; agno++) { for (agno = start_agno; agno <= end_agno; agno++) {
error = -xfs_trim_extents(mp, agno, start, len, minlen, error = -xfs_trim_extents(mp, agno, start, end, minlen,
&blocks_trimmed); &blocks_trimmed);
if (error) if (error)
last_error = error; last_error = error;

View File

@ -377,16 +377,14 @@ xfs_qm_dqalloc(
return (ESRCH); return (ESRCH);
} }
xfs_trans_ijoin_ref(tp, quotip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, quotip, XFS_ILOCK_EXCL);
nmaps = 1; nmaps = 1;
if ((error = xfs_bmapi(tp, quotip, error = xfs_bmapi_write(tp, quotip, offset_fsb,
offset_fsb, XFS_DQUOT_CLUSTER_SIZE_FSB, XFS_DQUOT_CLUSTER_SIZE_FSB, XFS_BMAPI_METADATA,
XFS_BMAPI_METADATA | XFS_BMAPI_WRITE, &firstblock, XFS_QM_DQALLOC_SPACE_RES(mp),
&firstblock, &map, &nmaps, &flist);
XFS_QM_DQALLOC_SPACE_RES(mp), if (error)
&map, &nmaps, &flist))) {
goto error0; goto error0;
}
ASSERT(map.br_blockcount == XFS_DQUOT_CLUSTER_SIZE_FSB); ASSERT(map.br_blockcount == XFS_DQUOT_CLUSTER_SIZE_FSB);
ASSERT(nmaps == 1); ASSERT(nmaps == 1);
ASSERT((map.br_startblock != DELAYSTARTBLOCK) && ASSERT((map.br_startblock != DELAYSTARTBLOCK) &&
@ -402,8 +400,11 @@ xfs_qm_dqalloc(
dqp->q_blkno, dqp->q_blkno,
mp->m_quotainfo->qi_dqchunklen, mp->m_quotainfo->qi_dqchunklen,
0); 0);
if (!bp || (error = xfs_buf_geterror(bp)))
error = xfs_buf_geterror(bp);
if (error)
goto error1; goto error1;
/* /*
* Make a chunk of dquots out of this buffer and log * Make a chunk of dquots out of this buffer and log
* the entire thing. * the entire thing.
@ -485,9 +486,8 @@ xfs_qm_dqtobp(
/* /*
* Find the block map; no allocations yet * Find the block map; no allocations yet
*/ */
error = xfs_bmapi(NULL, quotip, dqp->q_fileoffset, error = xfs_bmapi_read(quotip, dqp->q_fileoffset,
XFS_DQUOT_CLUSTER_SIZE_FSB, XFS_BMAPI_METADATA, XFS_DQUOT_CLUSTER_SIZE_FSB, &map, &nmaps, 0);
NULL, 0, &map, &nmaps, NULL);
xfs_iunlock(quotip, XFS_ILOCK_SHARED); xfs_iunlock(quotip, XFS_ILOCK_SHARED);
if (error) if (error)
@ -605,7 +605,7 @@ xfs_qm_dqread(
dqp->q_res_rtbcount = be64_to_cpu(ddqp->d_rtbcount); dqp->q_res_rtbcount = be64_to_cpu(ddqp->d_rtbcount);
/* Mark the buf so that this will stay incore a little longer */ /* Mark the buf so that this will stay incore a little longer */
XFS_BUF_SET_VTYPE_REF(bp, B_FS_DQUOT, XFS_DQUOT_REF); xfs_buf_set_ref(bp, XFS_DQUOT_REF);
/* /*
* We got the buffer with a xfs_trans_read_buf() (in dqtobp()) * We got the buffer with a xfs_trans_read_buf() (in dqtobp())
@ -1242,9 +1242,11 @@ xfs_qm_dqflush(
} }
if (flags & SYNC_WAIT) if (flags & SYNC_WAIT)
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
else else
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
trace_xfs_dqflush_done(dqp); trace_xfs_dqflush_done(dqp);

View File

@ -229,16 +229,16 @@ xfs_fs_nfs_commit_metadata(
{ {
struct xfs_inode *ip = XFS_I(inode); struct xfs_inode *ip = XFS_I(inode);
struct xfs_mount *mp = ip->i_mount; struct xfs_mount *mp = ip->i_mount;
int error = 0; xfs_lsn_t lsn = 0;
xfs_ilock(ip, XFS_ILOCK_SHARED); xfs_ilock(ip, XFS_ILOCK_SHARED);
if (xfs_ipincount(ip)) { if (xfs_ipincount(ip))
error = _xfs_log_force_lsn(mp, ip->i_itemp->ili_last_lsn, lsn = ip->i_itemp->ili_last_lsn;
XFS_LOG_SYNC, NULL);
}
xfs_iunlock(ip, XFS_ILOCK_SHARED); xfs_iunlock(ip, XFS_ILOCK_SHARED);
return error; if (!lsn)
return 0;
return _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, NULL);
} }
const struct export_operations xfs_export_operations = { const struct export_operations xfs_export_operations = {

View File

@ -124,6 +124,35 @@ xfs_iozero(
return (-status); return (-status);
} }
/*
* Fsync operations on directories are much simpler than on regular files,
* as there is no file data to flush, and thus also no need for explicit
* cache flush operations, and there are no non-transaction metadata updates
* on directories either.
*/
STATIC int
xfs_dir_fsync(
struct file *file,
loff_t start,
loff_t end,
int datasync)
{
struct xfs_inode *ip = XFS_I(file->f_mapping->host);
struct xfs_mount *mp = ip->i_mount;
xfs_lsn_t lsn = 0;
trace_xfs_dir_fsync(ip);
xfs_ilock(ip, XFS_ILOCK_SHARED);
if (xfs_ipincount(ip))
lsn = ip->i_itemp->ili_last_lsn;
xfs_iunlock(ip, XFS_ILOCK_SHARED);
if (!lsn)
return 0;
return _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, NULL);
}
STATIC int STATIC int
xfs_file_fsync( xfs_file_fsync(
struct file *file, struct file *file,
@ -137,6 +166,7 @@ xfs_file_fsync(
struct xfs_trans *tp; struct xfs_trans *tp;
int error = 0; int error = 0;
int log_flushed = 0; int log_flushed = 0;
xfs_lsn_t lsn = 0;
trace_xfs_file_fsync(ip); trace_xfs_file_fsync(ip);
@ -149,10 +179,6 @@ xfs_file_fsync(
xfs_iflags_clear(ip, XFS_ITRUNCATED); xfs_iflags_clear(ip, XFS_ITRUNCATED);
xfs_ilock(ip, XFS_IOLOCK_SHARED);
xfs_ioend_wait(ip);
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
if (mp->m_flags & XFS_MOUNT_BARRIER) { if (mp->m_flags & XFS_MOUNT_BARRIER) {
/* /*
* If we have an RT and/or log subvolume we need to make sure * If we have an RT and/or log subvolume we need to make sure
@ -216,11 +242,11 @@ xfs_file_fsync(
* transaction. So we play it safe and fire off the * transaction. So we play it safe and fire off the
* transaction anyway. * transaction anyway.
*/ */
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
xfs_trans_set_sync(tp); error = xfs_trans_commit(tp, 0);
error = _xfs_trans_commit(tp, 0, &log_flushed);
lsn = ip->i_itemp->ili_last_lsn;
xfs_iunlock(ip, XFS_ILOCK_EXCL); xfs_iunlock(ip, XFS_ILOCK_EXCL);
} else { } else {
/* /*
@ -231,14 +257,14 @@ xfs_file_fsync(
* disk yet, the inode will be still be pinned. If it is, * disk yet, the inode will be still be pinned. If it is,
* force the log. * force the log.
*/ */
if (xfs_ipincount(ip)) { if (xfs_ipincount(ip))
error = _xfs_log_force_lsn(mp, lsn = ip->i_itemp->ili_last_lsn;
ip->i_itemp->ili_last_lsn,
XFS_LOG_SYNC, &log_flushed);
}
xfs_iunlock(ip, XFS_ILOCK_SHARED); xfs_iunlock(ip, XFS_ILOCK_SHARED);
} }
if (!error && lsn)
error = _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, &log_flushed);
/* /*
* If we only have a single device, and the log force about was * If we only have a single device, and the log force about was
* a no-op we might have to flush the data device cache here. * a no-op we might have to flush the data device cache here.
@ -317,7 +343,19 @@ xfs_file_aio_read(
if (XFS_FORCED_SHUTDOWN(mp)) if (XFS_FORCED_SHUTDOWN(mp))
return -EIO; return -EIO;
if (unlikely(ioflags & IO_ISDIRECT)) { /*
* Locking is a bit tricky here. If we take an exclusive lock
* for direct IO, we effectively serialise all new concurrent
* read IO to this file and block it behind IO that is currently in
* progress because IO in progress holds the IO lock shared. We only
* need to hold the lock exclusive to blow away the page cache, so
* only take lock exclusively if the page cache needs invalidation.
* This allows the normal direct IO case of no page cache pages to
* proceeed concurrently without serialisation.
*/
xfs_rw_ilock(ip, XFS_IOLOCK_SHARED);
if ((ioflags & IO_ISDIRECT) && inode->i_mapping->nrpages) {
xfs_rw_iunlock(ip, XFS_IOLOCK_SHARED);
xfs_rw_ilock(ip, XFS_IOLOCK_EXCL); xfs_rw_ilock(ip, XFS_IOLOCK_EXCL);
if (inode->i_mapping->nrpages) { if (inode->i_mapping->nrpages) {
@ -330,8 +368,7 @@ xfs_file_aio_read(
} }
} }
xfs_rw_ilock_demote(ip, XFS_IOLOCK_EXCL); xfs_rw_ilock_demote(ip, XFS_IOLOCK_EXCL);
} else }
xfs_rw_ilock(ip, XFS_IOLOCK_SHARED);
trace_xfs_file_read(ip, size, iocb->ki_pos, ioflags); trace_xfs_file_read(ip, size, iocb->ki_pos, ioflags);
@ -407,11 +444,13 @@ xfs_aio_write_isize_update(
*/ */
STATIC void STATIC void
xfs_aio_write_newsize_update( xfs_aio_write_newsize_update(
struct xfs_inode *ip) struct xfs_inode *ip,
xfs_fsize_t new_size)
{ {
if (ip->i_new_size) { if (new_size == ip->i_new_size) {
xfs_rw_ilock(ip, XFS_ILOCK_EXCL); xfs_rw_ilock(ip, XFS_ILOCK_EXCL);
ip->i_new_size = 0; if (new_size == ip->i_new_size)
ip->i_new_size = 0;
if (ip->i_d.di_size > ip->i_size) if (ip->i_d.di_size > ip->i_size)
ip->i_d.di_size = ip->i_size; ip->i_d.di_size = ip->i_size;
xfs_rw_iunlock(ip, XFS_ILOCK_EXCL); xfs_rw_iunlock(ip, XFS_ILOCK_EXCL);
@ -462,7 +501,7 @@ xfs_file_splice_write(
ret = generic_file_splice_write(pipe, outfilp, ppos, count, flags); ret = generic_file_splice_write(pipe, outfilp, ppos, count, flags);
xfs_aio_write_isize_update(inode, ppos, ret); xfs_aio_write_isize_update(inode, ppos, ret);
xfs_aio_write_newsize_update(ip); xfs_aio_write_newsize_update(ip, new_size);
xfs_iunlock(ip, XFS_IOLOCK_EXCL); xfs_iunlock(ip, XFS_IOLOCK_EXCL);
return ret; return ret;
} }
@ -500,11 +539,9 @@ xfs_zero_last_block(
last_fsb = XFS_B_TO_FSBT(mp, isize); last_fsb = XFS_B_TO_FSBT(mp, isize);
nimaps = 1; nimaps = 1;
error = xfs_bmapi(NULL, ip, last_fsb, 1, 0, NULL, 0, &imap, error = xfs_bmapi_read(ip, last_fsb, 1, &imap, &nimaps, 0);
&nimaps, NULL); if (error)
if (error) {
return error; return error;
}
ASSERT(nimaps > 0); ASSERT(nimaps > 0);
/* /*
* If the block underlying isize is just a hole, then there * If the block underlying isize is just a hole, then there
@ -595,8 +632,8 @@ xfs_zero_eof(
while (start_zero_fsb <= end_zero_fsb) { while (start_zero_fsb <= end_zero_fsb) {
nimaps = 1; nimaps = 1;
zero_count_fsb = end_zero_fsb - start_zero_fsb + 1; zero_count_fsb = end_zero_fsb - start_zero_fsb + 1;
error = xfs_bmapi(NULL, ip, start_zero_fsb, zero_count_fsb, error = xfs_bmapi_read(ip, start_zero_fsb, zero_count_fsb,
0, NULL, 0, &imap, &nimaps, NULL); &imap, &nimaps, 0);
if (error) { if (error) {
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL|XFS_IOLOCK_EXCL)); ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL|XFS_IOLOCK_EXCL));
return error; return error;
@ -659,6 +696,7 @@ xfs_file_aio_write_checks(
struct file *file, struct file *file,
loff_t *pos, loff_t *pos,
size_t *count, size_t *count,
xfs_fsize_t *new_sizep,
int *iolock) int *iolock)
{ {
struct inode *inode = file->f_mapping->host; struct inode *inode = file->f_mapping->host;
@ -666,6 +704,9 @@ xfs_file_aio_write_checks(
xfs_fsize_t new_size; xfs_fsize_t new_size;
int error = 0; int error = 0;
xfs_rw_ilock(ip, XFS_ILOCK_EXCL);
*new_sizep = 0;
restart:
error = generic_write_checks(file, pos, count, S_ISBLK(inode->i_mode)); error = generic_write_checks(file, pos, count, S_ISBLK(inode->i_mode));
if (error) { if (error) {
xfs_rw_iunlock(ip, XFS_ILOCK_EXCL | *iolock); xfs_rw_iunlock(ip, XFS_ILOCK_EXCL | *iolock);
@ -673,20 +714,41 @@ xfs_file_aio_write_checks(
return error; return error;
} }
new_size = *pos + *count;
if (new_size > ip->i_size)
ip->i_new_size = new_size;
if (likely(!(file->f_mode & FMODE_NOCMTIME))) if (likely(!(file->f_mode & FMODE_NOCMTIME)))
file_update_time(file); file_update_time(file);
/* /*
* If the offset is beyond the size of the file, we need to zero any * If the offset is beyond the size of the file, we need to zero any
* blocks that fall between the existing EOF and the start of this * blocks that fall between the existing EOF and the start of this
* write. * write. There is no need to issue zeroing if another in-flght IO ends
* at or before this one If zeronig is needed and we are currently
* holding the iolock shared, we need to update it to exclusive which
* involves dropping all locks and relocking to maintain correct locking
* order. If we do this, restart the function to ensure all checks and
* values are still valid.
*/ */
if (*pos > ip->i_size) if ((ip->i_new_size && *pos > ip->i_new_size) ||
(!ip->i_new_size && *pos > ip->i_size)) {
if (*iolock == XFS_IOLOCK_SHARED) {
xfs_rw_iunlock(ip, XFS_ILOCK_EXCL | *iolock);
*iolock = XFS_IOLOCK_EXCL;
xfs_rw_ilock(ip, XFS_ILOCK_EXCL | *iolock);
goto restart;
}
error = -xfs_zero_eof(ip, *pos, ip->i_size); error = -xfs_zero_eof(ip, *pos, ip->i_size);
}
/*
* If this IO extends beyond EOF, we may need to update ip->i_new_size.
* We have already zeroed space beyond EOF (if necessary). Only update
* ip->i_new_size if this IO ends beyond any other in-flight writes.
*/
new_size = *pos + *count;
if (new_size > ip->i_size) {
if (new_size > ip->i_new_size)
ip->i_new_size = new_size;
*new_sizep = new_size;
}
xfs_rw_iunlock(ip, XFS_ILOCK_EXCL); xfs_rw_iunlock(ip, XFS_ILOCK_EXCL);
if (error) if (error)
@ -721,7 +783,7 @@ xfs_file_aio_write_checks(
* the dio layer. To avoid the problem with aio, we also need to wait for * the dio layer. To avoid the problem with aio, we also need to wait for
* outstanding IOs to complete so that unwritten extent conversion is completed * outstanding IOs to complete so that unwritten extent conversion is completed
* before we try to map the overlapping block. This is currently implemented by * before we try to map the overlapping block. This is currently implemented by
* hitting it with a big hammer (i.e. xfs_ioend_wait()). * hitting it with a big hammer (i.e. inode_dio_wait()).
* *
* Returns with locks held indicated by @iolock and errors indicated by * Returns with locks held indicated by @iolock and errors indicated by
* negative return values. * negative return values.
@ -733,6 +795,7 @@ xfs_file_dio_aio_write(
unsigned long nr_segs, unsigned long nr_segs,
loff_t pos, loff_t pos,
size_t ocount, size_t ocount,
xfs_fsize_t *new_size,
int *iolock) int *iolock)
{ {
struct file *file = iocb->ki_filp; struct file *file = iocb->ki_filp;
@ -753,18 +816,35 @@ xfs_file_dio_aio_write(
if ((pos & mp->m_blockmask) || ((pos + count) & mp->m_blockmask)) if ((pos & mp->m_blockmask) || ((pos + count) & mp->m_blockmask))
unaligned_io = 1; unaligned_io = 1;
if (unaligned_io || mapping->nrpages || pos > ip->i_size) /*
* We don't need to take an exclusive lock unless there page cache needs
* to be invalidated or unaligned IO is being executed. We don't need to
* consider the EOF extension case here because
* xfs_file_aio_write_checks() will relock the inode as necessary for
* EOF zeroing cases and fill out the new inode size as appropriate.
*/
if (unaligned_io || mapping->nrpages)
*iolock = XFS_IOLOCK_EXCL; *iolock = XFS_IOLOCK_EXCL;
else else
*iolock = XFS_IOLOCK_SHARED; *iolock = XFS_IOLOCK_SHARED;
xfs_rw_ilock(ip, XFS_ILOCK_EXCL | *iolock); xfs_rw_ilock(ip, *iolock);
ret = xfs_file_aio_write_checks(file, &pos, &count, iolock); /*
* Recheck if there are cached pages that need invalidate after we got
* the iolock to protect against other threads adding new pages while
* we were waiting for the iolock.
*/
if (mapping->nrpages && *iolock == XFS_IOLOCK_SHARED) {
xfs_rw_iunlock(ip, *iolock);
*iolock = XFS_IOLOCK_EXCL;
xfs_rw_ilock(ip, *iolock);
}
ret = xfs_file_aio_write_checks(file, &pos, &count, new_size, iolock);
if (ret) if (ret)
return ret; return ret;
if (mapping->nrpages) { if (mapping->nrpages) {
WARN_ON(*iolock != XFS_IOLOCK_EXCL);
ret = -xfs_flushinval_pages(ip, (pos & PAGE_CACHE_MASK), -1, ret = -xfs_flushinval_pages(ip, (pos & PAGE_CACHE_MASK), -1,
FI_REMAPF_LOCKED); FI_REMAPF_LOCKED);
if (ret) if (ret)
@ -776,7 +856,7 @@ xfs_file_dio_aio_write(
* otherwise demote the lock if we had to flush cached pages * otherwise demote the lock if we had to flush cached pages
*/ */
if (unaligned_io) if (unaligned_io)
xfs_ioend_wait(ip); inode_dio_wait(inode);
else if (*iolock == XFS_IOLOCK_EXCL) { else if (*iolock == XFS_IOLOCK_EXCL) {
xfs_rw_ilock_demote(ip, XFS_IOLOCK_EXCL); xfs_rw_ilock_demote(ip, XFS_IOLOCK_EXCL);
*iolock = XFS_IOLOCK_SHARED; *iolock = XFS_IOLOCK_SHARED;
@ -798,6 +878,7 @@ xfs_file_buffered_aio_write(
unsigned long nr_segs, unsigned long nr_segs,
loff_t pos, loff_t pos,
size_t ocount, size_t ocount,
xfs_fsize_t *new_size,
int *iolock) int *iolock)
{ {
struct file *file = iocb->ki_filp; struct file *file = iocb->ki_filp;
@ -809,9 +890,9 @@ xfs_file_buffered_aio_write(
size_t count = ocount; size_t count = ocount;
*iolock = XFS_IOLOCK_EXCL; *iolock = XFS_IOLOCK_EXCL;
xfs_rw_ilock(ip, XFS_ILOCK_EXCL | *iolock); xfs_rw_ilock(ip, *iolock);
ret = xfs_file_aio_write_checks(file, &pos, &count, iolock); ret = xfs_file_aio_write_checks(file, &pos, &count, new_size, iolock);
if (ret) if (ret)
return ret; return ret;
@ -851,6 +932,7 @@ xfs_file_aio_write(
ssize_t ret; ssize_t ret;
int iolock; int iolock;
size_t ocount = 0; size_t ocount = 0;
xfs_fsize_t new_size = 0;
XFS_STATS_INC(xs_write_calls); XFS_STATS_INC(xs_write_calls);
@ -870,10 +952,10 @@ xfs_file_aio_write(
if (unlikely(file->f_flags & O_DIRECT)) if (unlikely(file->f_flags & O_DIRECT))
ret = xfs_file_dio_aio_write(iocb, iovp, nr_segs, pos, ret = xfs_file_dio_aio_write(iocb, iovp, nr_segs, pos,
ocount, &iolock); ocount, &new_size, &iolock);
else else
ret = xfs_file_buffered_aio_write(iocb, iovp, nr_segs, pos, ret = xfs_file_buffered_aio_write(iocb, iovp, nr_segs, pos,
ocount, &iolock); ocount, &new_size, &iolock);
xfs_aio_write_isize_update(inode, &iocb->ki_pos, ret); xfs_aio_write_isize_update(inode, &iocb->ki_pos, ret);
@ -894,7 +976,7 @@ xfs_file_aio_write(
} }
out_unlock: out_unlock:
xfs_aio_write_newsize_update(ip); xfs_aio_write_newsize_update(ip, new_size);
xfs_rw_iunlock(ip, iolock); xfs_rw_iunlock(ip, iolock);
return ret; return ret;
} }
@ -1087,7 +1169,7 @@ const struct file_operations xfs_dir_file_operations = {
#ifdef CONFIG_COMPAT #ifdef CONFIG_COMPAT
.compat_ioctl = xfs_file_compat_ioctl, .compat_ioctl = xfs_file_compat_ioctl,
#endif #endif
.fsync = xfs_file_fsync, .fsync = xfs_dir_fsync,
}; };
static const struct vm_operations_struct xfs_file_vm_ops = { static const struct vm_operations_struct xfs_file_vm_ops = {

View File

@ -682,7 +682,7 @@ xfs_filestream_new_ag(
ip = ap->ip; ip = ap->ip;
mp = ip->i_mount; mp = ip->i_mount;
cache = mp->m_filestream; cache = mp->m_filestream;
minlen = ap->alen; minlen = ap->length;
*agp = NULLAGNUMBER; *agp = NULLAGNUMBER;
/* /*
@ -761,7 +761,7 @@ xfs_filestream_new_ag(
*/ */
ag = (ag == NULLAGNUMBER) ? 0 : (ag + 1) % mp->m_sb.sb_agcount; ag = (ag == NULLAGNUMBER) ? 0 : (ag + 1) % mp->m_sb.sb_agcount;
flags = (ap->userdata ? XFS_PICK_USERDATA : 0) | flags = (ap->userdata ? XFS_PICK_USERDATA : 0) |
(ap->low ? XFS_PICK_LOWSPACE : 0); (ap->flist->xbf_low ? XFS_PICK_LOWSPACE : 0);
err = _xfs_filestream_pick_ag(mp, ag, agp, flags, minlen); err = _xfs_filestream_pick_ag(mp, ag, agp, flags, minlen);
if (err || *agp == NULLAGNUMBER) if (err || *agp == NULLAGNUMBER)

View File

@ -194,6 +194,10 @@ xfs_growfs_data_private(
bp = xfs_buf_get(mp->m_ddev_targp, bp = xfs_buf_get(mp->m_ddev_targp,
XFS_AG_DADDR(mp, agno, XFS_AGF_DADDR(mp)), XFS_AG_DADDR(mp, agno, XFS_AGF_DADDR(mp)),
XFS_FSS_TO_BB(mp, 1), XBF_LOCK | XBF_MAPPED); XFS_FSS_TO_BB(mp, 1), XBF_LOCK | XBF_MAPPED);
if (!bp) {
error = ENOMEM;
goto error0;
}
agf = XFS_BUF_TO_AGF(bp); agf = XFS_BUF_TO_AGF(bp);
memset(agf, 0, mp->m_sb.sb_sectsize); memset(agf, 0, mp->m_sb.sb_sectsize);
agf->agf_magicnum = cpu_to_be32(XFS_AGF_MAGIC); agf->agf_magicnum = cpu_to_be32(XFS_AGF_MAGIC);
@ -216,16 +220,21 @@ xfs_growfs_data_private(
tmpsize = agsize - XFS_PREALLOC_BLOCKS(mp); tmpsize = agsize - XFS_PREALLOC_BLOCKS(mp);
agf->agf_freeblks = cpu_to_be32(tmpsize); agf->agf_freeblks = cpu_to_be32(tmpsize);
agf->agf_longest = cpu_to_be32(tmpsize); agf->agf_longest = cpu_to_be32(tmpsize);
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
if (error) { xfs_buf_relse(bp);
if (error)
goto error0; goto error0;
}
/* /*
* AG inode header block * AG inode header block
*/ */
bp = xfs_buf_get(mp->m_ddev_targp, bp = xfs_buf_get(mp->m_ddev_targp,
XFS_AG_DADDR(mp, agno, XFS_AGI_DADDR(mp)), XFS_AG_DADDR(mp, agno, XFS_AGI_DADDR(mp)),
XFS_FSS_TO_BB(mp, 1), XBF_LOCK | XBF_MAPPED); XFS_FSS_TO_BB(mp, 1), XBF_LOCK | XBF_MAPPED);
if (!bp) {
error = ENOMEM;
goto error0;
}
agi = XFS_BUF_TO_AGI(bp); agi = XFS_BUF_TO_AGI(bp);
memset(agi, 0, mp->m_sb.sb_sectsize); memset(agi, 0, mp->m_sb.sb_sectsize);
agi->agi_magicnum = cpu_to_be32(XFS_AGI_MAGIC); agi->agi_magicnum = cpu_to_be32(XFS_AGI_MAGIC);
@ -240,10 +249,11 @@ xfs_growfs_data_private(
agi->agi_dirino = cpu_to_be32(NULLAGINO); agi->agi_dirino = cpu_to_be32(NULLAGINO);
for (bucket = 0; bucket < XFS_AGI_UNLINKED_BUCKETS; bucket++) for (bucket = 0; bucket < XFS_AGI_UNLINKED_BUCKETS; bucket++)
agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO);
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
if (error) { xfs_buf_relse(bp);
if (error)
goto error0; goto error0;
}
/* /*
* BNO btree root block * BNO btree root block
*/ */
@ -251,6 +261,10 @@ xfs_growfs_data_private(
XFS_AGB_TO_DADDR(mp, agno, XFS_BNO_BLOCK(mp)), XFS_AGB_TO_DADDR(mp, agno, XFS_BNO_BLOCK(mp)),
BTOBB(mp->m_sb.sb_blocksize), BTOBB(mp->m_sb.sb_blocksize),
XBF_LOCK | XBF_MAPPED); XBF_LOCK | XBF_MAPPED);
if (!bp) {
error = ENOMEM;
goto error0;
}
block = XFS_BUF_TO_BLOCK(bp); block = XFS_BUF_TO_BLOCK(bp);
memset(block, 0, mp->m_sb.sb_blocksize); memset(block, 0, mp->m_sb.sb_blocksize);
block->bb_magic = cpu_to_be32(XFS_ABTB_MAGIC); block->bb_magic = cpu_to_be32(XFS_ABTB_MAGIC);
@ -262,10 +276,11 @@ xfs_growfs_data_private(
arec->ar_startblock = cpu_to_be32(XFS_PREALLOC_BLOCKS(mp)); arec->ar_startblock = cpu_to_be32(XFS_PREALLOC_BLOCKS(mp));
arec->ar_blockcount = cpu_to_be32( arec->ar_blockcount = cpu_to_be32(
agsize - be32_to_cpu(arec->ar_startblock)); agsize - be32_to_cpu(arec->ar_startblock));
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
if (error) { xfs_buf_relse(bp);
if (error)
goto error0; goto error0;
}
/* /*
* CNT btree root block * CNT btree root block
*/ */
@ -273,6 +288,10 @@ xfs_growfs_data_private(
XFS_AGB_TO_DADDR(mp, agno, XFS_CNT_BLOCK(mp)), XFS_AGB_TO_DADDR(mp, agno, XFS_CNT_BLOCK(mp)),
BTOBB(mp->m_sb.sb_blocksize), BTOBB(mp->m_sb.sb_blocksize),
XBF_LOCK | XBF_MAPPED); XBF_LOCK | XBF_MAPPED);
if (!bp) {
error = ENOMEM;
goto error0;
}
block = XFS_BUF_TO_BLOCK(bp); block = XFS_BUF_TO_BLOCK(bp);
memset(block, 0, mp->m_sb.sb_blocksize); memset(block, 0, mp->m_sb.sb_blocksize);
block->bb_magic = cpu_to_be32(XFS_ABTC_MAGIC); block->bb_magic = cpu_to_be32(XFS_ABTC_MAGIC);
@ -285,10 +304,11 @@ xfs_growfs_data_private(
arec->ar_blockcount = cpu_to_be32( arec->ar_blockcount = cpu_to_be32(
agsize - be32_to_cpu(arec->ar_startblock)); agsize - be32_to_cpu(arec->ar_startblock));
nfree += be32_to_cpu(arec->ar_blockcount); nfree += be32_to_cpu(arec->ar_blockcount);
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
if (error) { xfs_buf_relse(bp);
if (error)
goto error0; goto error0;
}
/* /*
* INO btree root block * INO btree root block
*/ */
@ -296,6 +316,10 @@ xfs_growfs_data_private(
XFS_AGB_TO_DADDR(mp, agno, XFS_IBT_BLOCK(mp)), XFS_AGB_TO_DADDR(mp, agno, XFS_IBT_BLOCK(mp)),
BTOBB(mp->m_sb.sb_blocksize), BTOBB(mp->m_sb.sb_blocksize),
XBF_LOCK | XBF_MAPPED); XBF_LOCK | XBF_MAPPED);
if (!bp) {
error = ENOMEM;
goto error0;
}
block = XFS_BUF_TO_BLOCK(bp); block = XFS_BUF_TO_BLOCK(bp);
memset(block, 0, mp->m_sb.sb_blocksize); memset(block, 0, mp->m_sb.sb_blocksize);
block->bb_magic = cpu_to_be32(XFS_IBT_MAGIC); block->bb_magic = cpu_to_be32(XFS_IBT_MAGIC);
@ -303,10 +327,10 @@ xfs_growfs_data_private(
block->bb_numrecs = 0; block->bb_numrecs = 0;
block->bb_u.s.bb_leftsib = cpu_to_be32(NULLAGBLOCK); block->bb_u.s.bb_leftsib = cpu_to_be32(NULLAGBLOCK);
block->bb_u.s.bb_rightsib = cpu_to_be32(NULLAGBLOCK); block->bb_u.s.bb_rightsib = cpu_to_be32(NULLAGBLOCK);
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
if (error) { xfs_buf_relse(bp);
if (error)
goto error0; goto error0;
}
} }
xfs_trans_agblocks_delta(tp, nfree); xfs_trans_agblocks_delta(tp, nfree);
/* /*
@ -396,9 +420,9 @@ xfs_growfs_data_private(
* just issue a warning and continue. The real work is * just issue a warning and continue. The real work is
* already done and committed. * already done and committed.
*/ */
if (!(error = xfs_bwrite(mp, bp))) { error = xfs_bwrite(bp);
continue; xfs_buf_relse(bp);
} else { if (error) {
xfs_warn(mp, xfs_warn(mp,
"write error %d updating secondary superblock for ag %d", "write error %d updating secondary superblock for ag %d",
error, agno); error, agno);

View File

@ -150,7 +150,7 @@ xfs_check_agi_freecount(
/* /*
* Initialise a new set of inodes. * Initialise a new set of inodes.
*/ */
STATIC void STATIC int
xfs_ialloc_inode_init( xfs_ialloc_inode_init(
struct xfs_mount *mp, struct xfs_mount *mp,
struct xfs_trans *tp, struct xfs_trans *tp,
@ -202,8 +202,8 @@ xfs_ialloc_inode_init(
fbuf = xfs_trans_get_buf(tp, mp->m_ddev_targp, d, fbuf = xfs_trans_get_buf(tp, mp->m_ddev_targp, d,
mp->m_bsize * blks_per_cluster, mp->m_bsize * blks_per_cluster,
XBF_LOCK); XBF_LOCK);
ASSERT(!xfs_buf_geterror(fbuf)); if (!fbuf)
return ENOMEM;
/* /*
* Initialize all inodes in this buffer and then log them. * Initialize all inodes in this buffer and then log them.
* *
@ -225,6 +225,7 @@ xfs_ialloc_inode_init(
} }
xfs_trans_inode_alloc_buf(tp, fbuf); xfs_trans_inode_alloc_buf(tp, fbuf);
} }
return 0;
} }
/* /*
@ -369,9 +370,11 @@ xfs_ialloc_ag_alloc(
* rather than a linear progression to prevent the next generation * rather than a linear progression to prevent the next generation
* number from being easily guessable. * number from being easily guessable.
*/ */
xfs_ialloc_inode_init(args.mp, tp, agno, args.agbno, args.len, error = xfs_ialloc_inode_init(args.mp, tp, agno, args.agbno,
random32()); args.len, random32());
if (error)
return error;
/* /*
* Convert the results. * Convert the results.
*/ */
@ -1502,7 +1505,7 @@ xfs_read_agi(
return XFS_ERROR(EFSCORRUPTED); return XFS_ERROR(EFSCORRUPTED);
} }
XFS_BUF_SET_VTYPE_REF(*bpp, B_FS_AGI, XFS_AGI_REF); xfs_buf_set_ref(*bpp, XFS_AGI_REF);
xfs_check_agi_unlinked(agi); xfs_check_agi_unlinked(agi);
return 0; return 0;

View File

@ -75,7 +75,6 @@ xfs_inode_alloc(
return NULL; return NULL;
} }
ASSERT(atomic_read(&ip->i_iocount) == 0);
ASSERT(atomic_read(&ip->i_pincount) == 0); ASSERT(atomic_read(&ip->i_pincount) == 0);
ASSERT(!spin_is_locked(&ip->i_flags_lock)); ASSERT(!spin_is_locked(&ip->i_flags_lock));
ASSERT(completion_done(&ip->i_flush)); ASSERT(completion_done(&ip->i_flush));
@ -150,7 +149,6 @@ xfs_inode_free(
} }
/* asserts to verify all state is correct here */ /* asserts to verify all state is correct here */
ASSERT(atomic_read(&ip->i_iocount) == 0);
ASSERT(atomic_read(&ip->i_pincount) == 0); ASSERT(atomic_read(&ip->i_pincount) == 0);
ASSERT(!spin_is_locked(&ip->i_flags_lock)); ASSERT(!spin_is_locked(&ip->i_flags_lock));
ASSERT(completion_done(&ip->i_flush)); ASSERT(completion_done(&ip->i_flush));

View File

@ -190,12 +190,6 @@ xfs_imap_to_bp(
} }
xfs_inobp_check(mp, bp); xfs_inobp_check(mp, bp);
/*
* Mark the buffer as an inode buffer now that it looks good
*/
XFS_BUF_SET_VTYPE(bp, B_FS_INO);
*bpp = bp; *bpp = bp;
return 0; return 0;
} }
@ -1152,7 +1146,7 @@ xfs_ialloc(
/* /*
* Log the new values stuffed into the inode. * Log the new values stuffed into the inode.
*/ */
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
xfs_trans_log_inode(tp, ip, flags); xfs_trans_log_inode(tp, ip, flags);
/* now that we have an i_mode we can setup inode ops and unlock */ /* now that we have an i_mode we can setup inode ops and unlock */
@ -1187,6 +1181,7 @@ xfs_isize_check(
xfs_fileoff_t map_first; xfs_fileoff_t map_first;
int nimaps; int nimaps;
xfs_bmbt_irec_t imaps[2]; xfs_bmbt_irec_t imaps[2];
int error;
if (!S_ISREG(ip->i_d.di_mode)) if (!S_ISREG(ip->i_d.di_mode))
return; return;
@ -1203,13 +1198,12 @@ xfs_isize_check(
* The filesystem could be shutting down, so bmapi may return * The filesystem could be shutting down, so bmapi may return
* an error. * an error.
*/ */
if (xfs_bmapi(NULL, ip, map_first, error = xfs_bmapi_read(ip, map_first,
(XFS_B_TO_FSB(mp, (XFS_B_TO_FSB(mp,
(xfs_ufsize_t)XFS_MAXIOFFSET(mp)) - (xfs_ufsize_t)XFS_MAXIOFFSET(mp)) - map_first),
map_first), imaps, &nimaps, XFS_BMAPI_ENTIRE);
XFS_BMAPI_ENTIRE, NULL, 0, imaps, &nimaps, if (error)
NULL)) return;
return;
ASSERT(nimaps == 1); ASSERT(nimaps == 1);
ASSERT(imaps[0].br_startblock == HOLESTARTBLOCK); ASSERT(imaps[0].br_startblock == HOLESTARTBLOCK);
} }
@ -1297,7 +1291,7 @@ xfs_itruncate_extents(
*/ */
error = xfs_bmap_finish(&tp, &free_list, &committed); error = xfs_bmap_finish(&tp, &free_list, &committed);
if (committed) if (committed)
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
if (error) if (error)
goto out_bmap_cancel; goto out_bmap_cancel;
@ -1313,7 +1307,7 @@ xfs_itruncate_extents(
error = xfs_trans_commit(tp, 0); error = xfs_trans_commit(tp, 0);
tp = ntp; tp = ntp;
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
if (error) if (error)
goto out; goto out;
@ -1644,7 +1638,7 @@ xfs_iunlink_remove(
* inodes that are in memory - they all must be marked stale and attached to * inodes that are in memory - they all must be marked stale and attached to
* the cluster buffer. * the cluster buffer.
*/ */
STATIC void STATIC int
xfs_ifree_cluster( xfs_ifree_cluster(
xfs_inode_t *free_ip, xfs_inode_t *free_ip,
xfs_trans_t *tp, xfs_trans_t *tp,
@ -1690,6 +1684,8 @@ xfs_ifree_cluster(
mp->m_bsize * blks_per_cluster, mp->m_bsize * blks_per_cluster,
XBF_LOCK); XBF_LOCK);
if (!bp)
return ENOMEM;
/* /*
* Walk the inodes already attached to the buffer and mark them * Walk the inodes already attached to the buffer and mark them
* stale. These will all have the flush locks held, so an * stale. These will all have the flush locks held, so an
@ -1799,6 +1795,7 @@ retry:
} }
xfs_perag_put(pag); xfs_perag_put(pag);
return 0;
} }
/* /*
@ -1878,10 +1875,10 @@ xfs_ifree(
dip->di_mode = 0; dip->di_mode = 0;
if (delete) { if (delete) {
xfs_ifree_cluster(ip, tp, first_ino); error = xfs_ifree_cluster(ip, tp, first_ino);
} }
return 0; return error;
} }
/* /*
@ -2472,11 +2469,11 @@ cluster_corrupt_out:
*/ */
if (bp->b_iodone) { if (bp->b_iodone) {
XFS_BUF_UNDONE(bp); XFS_BUF_UNDONE(bp);
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
xfs_buf_ioerror(bp, EIO); xfs_buf_ioerror(bp, EIO);
xfs_buf_ioend(bp, 0); xfs_buf_ioend(bp, 0);
} else { } else {
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
xfs_buf_relse(bp); xfs_buf_relse(bp);
} }
} }
@ -2597,9 +2594,11 @@ xfs_iflush(
goto cluster_corrupt_out; goto cluster_corrupt_out;
if (flags & SYNC_WAIT) if (flags & SYNC_WAIT)
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
else else
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
return error; return error;
corrupt_out: corrupt_out:

View File

@ -257,7 +257,6 @@ typedef struct xfs_inode {
xfs_fsize_t i_size; /* in-memory size */ xfs_fsize_t i_size; /* in-memory size */
xfs_fsize_t i_new_size; /* size when write completes */ xfs_fsize_t i_new_size; /* size when write completes */
atomic_t i_iocount; /* outstanding I/O count */
/* VFS inode */ /* VFS inode */
struct inode i_vnode; /* embedded VFS inode */ struct inode i_vnode; /* embedded VFS inode */

View File

@ -658,10 +658,8 @@ xfs_inode_item_unlock(
lock_flags = iip->ili_lock_flags; lock_flags = iip->ili_lock_flags;
iip->ili_lock_flags = 0; iip->ili_lock_flags = 0;
if (lock_flags) { if (lock_flags)
xfs_iunlock(ip, lock_flags); xfs_iunlock(ip, lock_flags);
IRELE(ip);
}
} }
/* /*

View File

@ -1069,7 +1069,7 @@ xfs_ioctl_setattr(
} }
} }
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/* /*
* Change file ownership. Must be the owner or privileged. * Change file ownership. Must be the owner or privileged.

View File

@ -208,22 +208,20 @@ xfs_iomap_write_direct(
if (error) if (error)
goto error1; goto error1;
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
bmapi_flag = XFS_BMAPI_WRITE; bmapi_flag = 0;
if (offset < ip->i_size || extsz) if (offset < ip->i_size || extsz)
bmapi_flag |= XFS_BMAPI_PREALLOC; bmapi_flag |= XFS_BMAPI_PREALLOC;
/* /*
* Issue the xfs_bmapi() call to allocate the blocks.
*
* From this point onwards we overwrite the imap pointer that the * From this point onwards we overwrite the imap pointer that the
* caller gave to us. * caller gave to us.
*/ */
xfs_bmap_init(&free_list, &firstfsb); xfs_bmap_init(&free_list, &firstfsb);
nimaps = 1; nimaps = 1;
error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, bmapi_flag, error = xfs_bmapi_write(tp, ip, offset_fsb, count_fsb, bmapi_flag,
&firstfsb, 0, imap, &nimaps, &free_list); &firstfsb, 0, imap, &nimaps, &free_list);
if (error) if (error)
goto error0; goto error0;
@ -300,8 +298,8 @@ xfs_iomap_eof_want_preallocate(
while (count_fsb > 0) { while (count_fsb > 0) {
imaps = nimaps; imaps = nimaps;
firstblock = NULLFSBLOCK; firstblock = NULLFSBLOCK;
error = xfs_bmapi(NULL, ip, start_fsb, count_fsb, 0, error = xfs_bmapi_read(ip, start_fsb, count_fsb, imap, &imaps,
&firstblock, 0, imap, &imaps, NULL); 0);
if (error) if (error)
return error; return error;
for (n = 0; n < imaps; n++) { for (n = 0; n < imaps; n++) {
@ -381,7 +379,6 @@ xfs_iomap_write_delay(
xfs_fileoff_t last_fsb; xfs_fileoff_t last_fsb;
xfs_off_t aligned_offset; xfs_off_t aligned_offset;
xfs_fileoff_t ioalign; xfs_fileoff_t ioalign;
xfs_fsblock_t firstblock;
xfs_extlen_t extsz; xfs_extlen_t extsz;
int nimaps; int nimaps;
xfs_bmbt_irec_t imap[XFS_WRITE_IMAPS]; xfs_bmbt_irec_t imap[XFS_WRITE_IMAPS];
@ -425,12 +422,8 @@ retry:
} }
nimaps = XFS_WRITE_IMAPS; nimaps = XFS_WRITE_IMAPS;
firstblock = NULLFSBLOCK; error = xfs_bmapi_delay(ip, offset_fsb, last_fsb - offset_fsb,
error = xfs_bmapi(NULL, ip, offset_fsb, imap, &nimaps, XFS_BMAPI_ENTIRE);
(xfs_filblks_t)(last_fsb - offset_fsb),
XFS_BMAPI_DELAY | XFS_BMAPI_WRITE |
XFS_BMAPI_ENTIRE, &firstblock, 1, imap,
&nimaps, NULL);
switch (error) { switch (error) {
case 0: case 0:
case ENOSPC: case ENOSPC:
@ -535,7 +528,7 @@ xfs_iomap_write_allocate(
return XFS_ERROR(error); return XFS_ERROR(error);
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
xfs_bmap_init(&free_list, &first_block); xfs_bmap_init(&free_list, &first_block);
@ -587,14 +580,12 @@ xfs_iomap_write_allocate(
} }
/* /*
* Go get the actual blocks.
*
* From this point onwards we overwrite the imap * From this point onwards we overwrite the imap
* pointer that the caller gave to us. * pointer that the caller gave to us.
*/ */
error = xfs_bmapi(tp, ip, map_start_fsb, count_fsb, error = xfs_bmapi_write(tp, ip, map_start_fsb,
XFS_BMAPI_WRITE, &first_block, 1, count_fsb, 0, &first_block, 1,
imap, &nimaps, &free_list); imap, &nimaps, &free_list);
if (error) if (error)
goto trans_cancel; goto trans_cancel;
@ -701,15 +692,15 @@ xfs_iomap_write_unwritten(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/* /*
* Modify the unwritten extent state of the buffer. * Modify the unwritten extent state of the buffer.
*/ */
xfs_bmap_init(&free_list, &firstfsb); xfs_bmap_init(&free_list, &firstfsb);
nimaps = 1; nimaps = 1;
error = xfs_bmapi(tp, ip, offset_fsb, count_fsb, error = xfs_bmapi_write(tp, ip, offset_fsb, count_fsb,
XFS_BMAPI_WRITE|XFS_BMAPI_CONVERT, &firstfsb, XFS_BMAPI_CONVERT, &firstfsb,
1, &imap, &nimaps, &free_list); 1, &imap, &nimaps, &free_list);
if (error) if (error)
goto error_on_bmapi_transaction; goto error_on_bmapi_transaction;

View File

@ -466,7 +466,7 @@ xfs_vn_getattr(
trace_xfs_getattr(ip); trace_xfs_getattr(ip);
if (XFS_FORCED_SHUTDOWN(mp)) if (XFS_FORCED_SHUTDOWN(mp))
return XFS_ERROR(EIO); return -XFS_ERROR(EIO);
stat->size = XFS_ISIZE(ip); stat->size = XFS_ISIZE(ip);
stat->dev = inode->i_sb->s_dev; stat->dev = inode->i_sb->s_dev;
@ -612,7 +612,7 @@ xfs_setattr_nonsize(
} }
} }
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/* /*
* Change file ownership. Must be the owner or privileged. * Change file ownership. Must be the owner or privileged.
@ -834,16 +834,16 @@ xfs_setattr_size(
* care about here. * care about here.
*/ */
if (ip->i_size != ip->i_d.di_size && iattr->ia_size > ip->i_d.di_size) { if (ip->i_size != ip->i_d.di_size && iattr->ia_size > ip->i_d.di_size) {
error = xfs_flush_pages(ip, ip->i_d.di_size, iattr->ia_size, error = xfs_flush_pages(ip, ip->i_d.di_size, iattr->ia_size, 0,
XBF_ASYNC, FI_NONE); FI_NONE);
if (error) if (error)
goto out_unlock; goto out_unlock;
} }
/* /*
* Wait for all I/O to complete. * Wait for all direct I/O to complete.
*/ */
xfs_ioend_wait(ip); inode_dio_wait(inode);
error = -block_truncate_page(inode->i_mapping, iattr->ia_size, error = -block_truncate_page(inode->i_mapping, iattr->ia_size,
xfs_get_blocks); xfs_get_blocks);
@ -864,7 +864,7 @@ xfs_setattr_size(
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/* /*
* Only change the c/mtime if we are changing the size or we are * Only change the c/mtime if we are changing the size or we are

View File

@ -880,8 +880,8 @@ xlog_iodone(xfs_buf_t *bp)
*/ */
if (XFS_TEST_ERROR((xfs_buf_geterror(bp)), l->l_mp, if (XFS_TEST_ERROR((xfs_buf_geterror(bp)), l->l_mp,
XFS_ERRTAG_IODONE_IOERR, XFS_RANDOM_IODONE_IOERR)) { XFS_ERRTAG_IODONE_IOERR, XFS_RANDOM_IODONE_IOERR)) {
xfs_ioerror_alert("xlog_iodone", l->l_mp, bp, XFS_BUF_ADDR(bp)); xfs_buf_ioerror_alert(bp, __func__);
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
xfs_force_shutdown(l->l_mp, SHUTDOWN_LOG_IO_ERROR); xfs_force_shutdown(l->l_mp, SHUTDOWN_LOG_IO_ERROR);
/* /*
* This flag will be propagated to the trans-committed * This flag will be propagated to the trans-committed
@ -1047,7 +1047,7 @@ xlog_alloc_log(xfs_mount_t *mp,
xlog_get_iclog_buffer_size(mp, log); xlog_get_iclog_buffer_size(mp, log);
error = ENOMEM; error = ENOMEM;
bp = xfs_buf_get_empty(log->l_iclog_size, mp->m_logdev_targp); bp = xfs_buf_alloc(mp->m_logdev_targp, 0, log->l_iclog_size, 0);
if (!bp) if (!bp)
goto out_free_log; goto out_free_log;
bp->b_iodone = xlog_iodone; bp->b_iodone = xlog_iodone;
@ -1247,7 +1247,7 @@ xlog_bdstrat(
if (iclog->ic_state & XLOG_STATE_IOERROR) { if (iclog->ic_state & XLOG_STATE_IOERROR) {
xfs_buf_ioerror(bp, EIO); xfs_buf_ioerror(bp, EIO);
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
xfs_buf_ioend(bp, 0); xfs_buf_ioend(bp, 0);
/* /*
* It would seem logical to return EIO here, but we rely on * It would seem logical to return EIO here, but we rely on
@ -1387,9 +1387,9 @@ xlog_sync(xlog_t *log,
*/ */
XFS_BUF_WRITE(bp); XFS_BUF_WRITE(bp);
if ((error = xlog_bdstrat(bp))) { error = xlog_bdstrat(bp);
xfs_ioerror_alert("xlog_sync", log->l_mp, bp, if (error) {
XFS_BUF_ADDR(bp)); xfs_buf_ioerror_alert(bp, "xlog_sync");
return error; return error;
} }
if (split) { if (split) {
@ -1423,9 +1423,9 @@ xlog_sync(xlog_t *log,
/* account for internal log which doesn't start at block #0 */ /* account for internal log which doesn't start at block #0 */
XFS_BUF_SET_ADDR(bp, XFS_BUF_ADDR(bp) + log->l_logBBstart); XFS_BUF_SET_ADDR(bp, XFS_BUF_ADDR(bp) + log->l_logBBstart);
XFS_BUF_WRITE(bp); XFS_BUF_WRITE(bp);
if ((error = xlog_bdstrat(bp))) { error = xlog_bdstrat(bp);
xfs_ioerror_alert("xlog_sync (split)", log->l_mp, if (error) {
bp, XFS_BUF_ADDR(bp)); xfs_buf_ioerror_alert(bp, "xlog_sync (split)");
return error; return error;
} }
} }

View File

@ -183,8 +183,7 @@ xlog_bread_noalign(
xfsbdstrat(log->l_mp, bp); xfsbdstrat(log->l_mp, bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) if (error)
xfs_ioerror_alert("xlog_bread", log->l_mp, xfs_buf_ioerror_alert(bp, __func__);
bp, XFS_BUF_ADDR(bp));
return error; return error;
} }
@ -268,9 +267,10 @@ xlog_bwrite(
xfs_buf_lock(bp); xfs_buf_lock(bp);
XFS_BUF_SET_COUNT(bp, BBTOB(nbblks)); XFS_BUF_SET_COUNT(bp, BBTOB(nbblks));
if ((error = xfs_bwrite(log->l_mp, bp))) error = xfs_bwrite(bp);
xfs_ioerror_alert("xlog_bwrite", log->l_mp, if (error)
bp, XFS_BUF_ADDR(bp)); xfs_buf_ioerror_alert(bp, __func__);
xfs_buf_relse(bp);
return error; return error;
} }
@ -361,9 +361,7 @@ xlog_recover_iodone(
* We're not going to bother about retrying * We're not going to bother about retrying
* this during recovery. One strike! * this during recovery. One strike!
*/ */
xfs_ioerror_alert("xlog_recover_iodone", xfs_buf_ioerror_alert(bp, __func__);
bp->b_target->bt_mount, bp,
XFS_BUF_ADDR(bp));
xfs_force_shutdown(bp->b_target->bt_mount, xfs_force_shutdown(bp->b_target->bt_mount,
SHUTDOWN_META_IO_ERROR); SHUTDOWN_META_IO_ERROR);
} }
@ -2135,8 +2133,7 @@ xlog_recover_buffer_pass2(
return XFS_ERROR(ENOMEM); return XFS_ERROR(ENOMEM);
error = bp->b_error; error = bp->b_error;
if (error) { if (error) {
xfs_ioerror_alert("xlog_recover_do..(read#1)", mp, xfs_buf_ioerror_alert(bp, "xlog_recover_do..(read#1)");
bp, buf_f->blf_blkno);
xfs_buf_relse(bp); xfs_buf_relse(bp);
return error; return error;
} }
@ -2171,15 +2168,16 @@ xlog_recover_buffer_pass2(
be16_to_cpu(*((__be16 *)xfs_buf_offset(bp, 0))) && be16_to_cpu(*((__be16 *)xfs_buf_offset(bp, 0))) &&
(XFS_BUF_COUNT(bp) != MAX(log->l_mp->m_sb.sb_blocksize, (XFS_BUF_COUNT(bp) != MAX(log->l_mp->m_sb.sb_blocksize,
(__uint32_t)XFS_INODE_CLUSTER_SIZE(log->l_mp)))) { (__uint32_t)XFS_INODE_CLUSTER_SIZE(log->l_mp)))) {
XFS_BUF_STALE(bp); xfs_buf_stale(bp);
error = xfs_bwrite(mp, bp); error = xfs_bwrite(bp);
} else { } else {
ASSERT(bp->b_target->bt_mount == mp); ASSERT(bp->b_target->bt_mount == mp);
bp->b_iodone = xlog_recover_iodone; bp->b_iodone = xlog_recover_iodone;
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
} }
return (error); xfs_buf_relse(bp);
return error;
} }
STATIC int STATIC int
@ -2230,8 +2228,7 @@ xlog_recover_inode_pass2(
} }
error = bp->b_error; error = bp->b_error;
if (error) { if (error) {
xfs_ioerror_alert("xlog_recover_do..(read#2)", mp, xfs_buf_ioerror_alert(bp, "xlog_recover_do..(read#2)");
bp, in_f->ilf_blkno);
xfs_buf_relse(bp); xfs_buf_relse(bp);
goto error; goto error;
} }
@ -2439,7 +2436,8 @@ xlog_recover_inode_pass2(
write_inode_buffer: write_inode_buffer:
ASSERT(bp->b_target->bt_mount == mp); ASSERT(bp->b_target->bt_mount == mp);
bp->b_iodone = xlog_recover_iodone; bp->b_iodone = xlog_recover_iodone;
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
error: error:
if (need_free) if (need_free)
kmem_free(in_f); kmem_free(in_f);
@ -2537,8 +2535,7 @@ xlog_recover_dquot_pass2(
XFS_FSB_TO_BB(mp, dq_f->qlf_len), XFS_FSB_TO_BB(mp, dq_f->qlf_len),
0, &bp); 0, &bp);
if (error) { if (error) {
xfs_ioerror_alert("xlog_recover_do..(read#3)", mp, xfs_buf_ioerror_alert(bp, "xlog_recover_do..(read#3)");
bp, dq_f->qlf_blkno);
return error; return error;
} }
ASSERT(bp); ASSERT(bp);
@ -2561,7 +2558,8 @@ xlog_recover_dquot_pass2(
ASSERT(dq_f->qlf_size == 2); ASSERT(dq_f->qlf_size == 2);
ASSERT(bp->b_target->bt_mount == mp); ASSERT(bp->b_target->bt_mount == mp);
bp->b_iodone = xlog_recover_iodone; bp->b_iodone = xlog_recover_iodone;
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
return (0); return (0);
} }
@ -3656,7 +3654,7 @@ xlog_do_recover(
return error; return error;
} }
XFS_bflush(log->l_mp->m_ddev_targp); xfs_flush_buftarg(log->l_mp->m_ddev_targp, 1);
/* /*
* If IO errors happened during recovery, bail out. * If IO errors happened during recovery, bail out.
@ -3689,8 +3687,7 @@ xlog_do_recover(
xfsbdstrat(log->l_mp, bp); xfsbdstrat(log->l_mp, bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) { if (error) {
xfs_ioerror_alert("xlog_do_recover", xfs_buf_ioerror_alert(bp, __func__);
log->l_mp, bp, XFS_BUF_ADDR(bp));
ASSERT(0); ASSERT(0);
xfs_buf_relse(bp); xfs_buf_relse(bp);
return error; return error;

View File

@ -44,9 +44,6 @@
#include "xfs_trace.h" #include "xfs_trace.h"
STATIC void xfs_unmountfs_wait(xfs_mount_t *);
#ifdef HAVE_PERCPU_SB #ifdef HAVE_PERCPU_SB
STATIC void xfs_icsb_balance_counter(xfs_mount_t *, xfs_sb_field_t, STATIC void xfs_icsb_balance_counter(xfs_mount_t *, xfs_sb_field_t,
int); int);
@ -1484,7 +1481,7 @@ xfs_unmountfs(
* state as much as possible. * state as much as possible.
*/ */
xfs_reclaim_inodes(mp, 0); xfs_reclaim_inodes(mp, 0);
XFS_bflush(mp->m_ddev_targp); xfs_flush_buftarg(mp->m_ddev_targp, 1);
xfs_reclaim_inodes(mp, SYNC_WAIT); xfs_reclaim_inodes(mp, SYNC_WAIT);
xfs_qm_unmount(mp); xfs_qm_unmount(mp);
@ -1496,11 +1493,6 @@ xfs_unmountfs(
*/ */
xfs_log_force(mp, XFS_LOG_SYNC); xfs_log_force(mp, XFS_LOG_SYNC);
xfs_binval(mp->m_ddev_targp);
if (mp->m_rtdev_targp) {
xfs_binval(mp->m_rtdev_targp);
}
/* /*
* Unreserve any blocks we have so that when we unmount we don't account * Unreserve any blocks we have so that when we unmount we don't account
* the reserved free space as used. This is really only necessary for * the reserved free space as used. This is really only necessary for
@ -1526,7 +1518,16 @@ xfs_unmountfs(
xfs_warn(mp, "Unable to update superblock counters. " xfs_warn(mp, "Unable to update superblock counters. "
"Freespace may not be correct on next mount."); "Freespace may not be correct on next mount.");
xfs_unmountfs_writesb(mp); xfs_unmountfs_writesb(mp);
xfs_unmountfs_wait(mp); /* wait for async bufs */
/*
* Make sure all buffers have been flushed and completed before
* unmounting the log.
*/
error = xfs_flush_buftarg(mp->m_ddev_targp, 1);
if (error)
xfs_warn(mp, "%d busy buffers during unmount.", error);
xfs_wait_buftarg(mp->m_ddev_targp);
xfs_log_unmount_write(mp); xfs_log_unmount_write(mp);
xfs_log_unmount(mp); xfs_log_unmount(mp);
xfs_uuid_unmount(mp); xfs_uuid_unmount(mp);
@ -1537,16 +1538,6 @@ xfs_unmountfs(
xfs_free_perag(mp); xfs_free_perag(mp);
} }
STATIC void
xfs_unmountfs_wait(xfs_mount_t *mp)
{
if (mp->m_logdev_targp != mp->m_ddev_targp)
xfs_wait_buftarg(mp->m_logdev_targp);
if (mp->m_rtdev_targp)
xfs_wait_buftarg(mp->m_rtdev_targp);
xfs_wait_buftarg(mp->m_ddev_targp);
}
int int
xfs_fs_writable(xfs_mount_t *mp) xfs_fs_writable(xfs_mount_t *mp)
{ {
@ -1612,15 +1603,14 @@ xfs_unmountfs_writesb(xfs_mount_t *mp)
XFS_BUF_UNDONE(sbp); XFS_BUF_UNDONE(sbp);
XFS_BUF_UNREAD(sbp); XFS_BUF_UNREAD(sbp);
XFS_BUF_UNDELAYWRITE(sbp); xfs_buf_delwri_dequeue(sbp);
XFS_BUF_WRITE(sbp); XFS_BUF_WRITE(sbp);
XFS_BUF_UNASYNC(sbp); XFS_BUF_UNASYNC(sbp);
ASSERT(sbp->b_target == mp->m_ddev_targp); ASSERT(sbp->b_target == mp->m_ddev_targp);
xfsbdstrat(mp, sbp); xfsbdstrat(mp, sbp);
error = xfs_buf_iowait(sbp); error = xfs_buf_iowait(sbp);
if (error) if (error)
xfs_ioerror_alert("xfs_unmountfs_writesb", xfs_buf_ioerror_alert(sbp, __func__);
mp, sbp, XFS_BUF_ADDR(sbp));
xfs_buf_relse(sbp); xfs_buf_relse(sbp);
} }
return error; return error;

View File

@ -1296,7 +1296,8 @@ xfs_qm_dqiter_bufs(
break; break;
xfs_qm_reset_dqcounts(mp, bp, firstid, type); xfs_qm_reset_dqcounts(mp, bp, firstid, type);
xfs_bdwrite(mp, bp); xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
/* /*
* goto the next block. * goto the next block.
*/ */
@ -1346,11 +1347,8 @@ xfs_qm_dqiterate(
* the inode is never added to the transaction. * the inode is never added to the transaction.
*/ */
xfs_ilock(qip, XFS_ILOCK_SHARED); xfs_ilock(qip, XFS_ILOCK_SHARED);
error = xfs_bmapi(NULL, qip, lblkno, error = xfs_bmapi_read(qip, lblkno, maxlblkcnt - lblkno,
maxlblkcnt - lblkno, map, &nmaps, 0);
XFS_BMAPI_METADATA,
NULL,
0, map, &nmaps, NULL);
xfs_iunlock(qip, XFS_ILOCK_SHARED); xfs_iunlock(qip, XFS_ILOCK_SHARED);
if (error) if (error)
break; break;
@ -1683,7 +1681,7 @@ xfs_qm_quotacheck(
* quotacheck'd stamp on the superblock. So, here we do a synchronous * quotacheck'd stamp on the superblock. So, here we do a synchronous
* flush. * flush.
*/ */
XFS_bflush(mp->m_ddev_targp); xfs_flush_buftarg(mp->m_ddev_targp, 1);
/* /*
* If one type of quotas is off, then it will lose its * If one type of quotas is off, then it will lose its

View File

@ -261,7 +261,7 @@ xfs_qm_scall_trunc_qfile(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
error = xfs_itruncate_data(&tp, ip, 0); error = xfs_itruncate_data(&tp, ip, 0);
if (error) { if (error) {

View File

@ -170,12 +170,12 @@ xfs_rename(
* we can rely on either trans_commit or trans_cancel to unlock * we can rely on either trans_commit or trans_cancel to unlock
* them. * them.
*/ */
xfs_trans_ijoin_ref(tp, src_dp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, src_dp, XFS_ILOCK_EXCL);
if (new_parent) if (new_parent)
xfs_trans_ijoin_ref(tp, target_dp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, target_dp, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, src_ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, src_ip, XFS_ILOCK_EXCL);
if (target_ip) if (target_ip)
xfs_trans_ijoin_ref(tp, target_ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, target_ip, XFS_ILOCK_EXCL);
/* /*
* If we are using project inheritance, we only allow renames * If we are using project inheritance, we only allow renames

View File

@ -112,7 +112,7 @@ xfs_growfs_rt_alloc(
* Lock the inode. * Lock the inode.
*/ */
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
xfs_bmap_init(&flist, &firstblock); xfs_bmap_init(&flist, &firstblock);
/* /*
@ -120,9 +120,9 @@ xfs_growfs_rt_alloc(
*/ */
nmap = 1; nmap = 1;
cancelflags |= XFS_TRANS_ABORT; cancelflags |= XFS_TRANS_ABORT;
error = xfs_bmapi(tp, ip, oblocks, nblocks - oblocks, error = xfs_bmapi_write(tp, ip, oblocks, nblocks - oblocks,
XFS_BMAPI_WRITE | XFS_BMAPI_METADATA, &firstblock, XFS_BMAPI_METADATA, &firstblock,
resblks, &map, &nmap, &flist); resblks, &map, &nmap, &flist);
if (!error && nmap < 1) if (!error && nmap < 1)
error = XFS_ERROR(ENOSPC); error = XFS_ERROR(ENOSPC);
if (error) if (error)
@ -155,7 +155,7 @@ xfs_growfs_rt_alloc(
* Lock the bitmap inode. * Lock the bitmap inode.
*/ */
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
/* /*
* Get a buffer for the block. * Get a buffer for the block.
*/ */
@ -856,33 +856,23 @@ xfs_rtbuf_get(
xfs_buf_t **bpp) /* output: buffer for the block */ xfs_buf_t **bpp) /* output: buffer for the block */
{ {
xfs_buf_t *bp; /* block buffer, result */ xfs_buf_t *bp; /* block buffer, result */
xfs_daddr_t d; /* disk addr of block */
int error; /* error value */
xfs_fsblock_t fsb; /* fs block number for block */
xfs_inode_t *ip; /* bitmap or summary inode */ xfs_inode_t *ip; /* bitmap or summary inode */
xfs_bmbt_irec_t map;
int nmap;
int error; /* error value */
ip = issum ? mp->m_rsumip : mp->m_rbmip; ip = issum ? mp->m_rsumip : mp->m_rbmip;
/*
* Map from the file offset (block) and inode number to the error = xfs_bmapi_read(ip, block, 1, &map, &nmap, XFS_DATA_FORK);
* file system block. if (error)
*/
error = xfs_bmapi_single(tp, ip, XFS_DATA_FORK, &fsb, block);
if (error) {
return error; return error;
}
ASSERT(fsb != NULLFSBLOCK); ASSERT(map.br_startblock != NULLFSBLOCK);
/* error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
* Convert to disk address for buffer cache. XFS_FSB_TO_DADDR(mp, map.br_startblock),
*/
d = XFS_FSB_TO_DADDR(mp, fsb);
/*
* Read the buffer.
*/
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, d,
mp->m_bsize, 0, &bp); mp->m_bsize, 0, &bp);
if (error) { if (error)
return error; return error;
}
ASSERT(!xfs_buf_geterror(bp)); ASSERT(!xfs_buf_geterror(bp));
*bpp = bp; *bpp = bp;
return 0; return 0;
@ -1970,7 +1960,7 @@ xfs_growfs_rt(
* Lock out other callers by grabbing the bitmap inode lock. * Lock out other callers by grabbing the bitmap inode lock.
*/ */
xfs_ilock(mp->m_rbmip, XFS_ILOCK_EXCL); xfs_ilock(mp->m_rbmip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, mp->m_rbmip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, mp->m_rbmip, XFS_ILOCK_EXCL);
/* /*
* Update the bitmap inode's size. * Update the bitmap inode's size.
*/ */
@ -1982,7 +1972,7 @@ xfs_growfs_rt(
* Get the summary inode into the transaction. * Get the summary inode into the transaction.
*/ */
xfs_ilock(mp->m_rsumip, XFS_ILOCK_EXCL); xfs_ilock(mp->m_rsumip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, mp->m_rsumip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, mp->m_rsumip, XFS_ILOCK_EXCL);
/* /*
* Update the summary inode's size. * Update the summary inode's size.
*/ */
@ -2153,7 +2143,7 @@ xfs_rtfree_extent(
* Synchronize by locking the bitmap inode. * Synchronize by locking the bitmap inode.
*/ */
xfs_ilock(mp->m_rbmip, XFS_ILOCK_EXCL); xfs_ilock(mp->m_rbmip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, mp->m_rbmip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, mp->m_rbmip, XFS_ILOCK_EXCL);
#if defined(__KERNEL__) && defined(DEBUG) #if defined(__KERNEL__) && defined(DEBUG)
/* /*

View File

@ -91,24 +91,6 @@ xfs_do_force_shutdown(
} }
} }
/*
* Prints out an ALERT message about I/O error.
*/
void
xfs_ioerror_alert(
char *func,
struct xfs_mount *mp,
xfs_buf_t *bp,
xfs_daddr_t blkno)
{
xfs_alert(mp,
"I/O error occurred: meta-data dev %s block 0x%llx"
" (\"%s\") error %d buf count %zd",
xfs_buf_target_name(bp->b_target),
(__uint64_t)blkno, func,
bp->b_error, XFS_BUF_COUNT(bp));
}
/* /*
* This isn't an absolute requirement, but it is * This isn't an absolute requirement, but it is
* just a good idea to call xfs_read_buf instead of * just a good idea to call xfs_read_buf instead of
@ -143,14 +125,13 @@ xfs_read_buf(
} else { } else {
*bpp = NULL; *bpp = NULL;
if (error) { if (error) {
xfs_ioerror_alert("xfs_read_buf", mp, bp, XFS_BUF_ADDR(bp)); xfs_buf_ioerror_alert(bp, __func__);
} else { } else {
error = XFS_ERROR(EIO); error = XFS_ERROR(EIO);
} }
if (bp) { if (bp) {
XFS_BUF_UNDONE(bp); XFS_BUF_UNDONE(bp);
XFS_BUF_UNDELAYWRITE(bp); xfs_buf_stale(bp);
XFS_BUF_STALE(bp);
/* /*
* brelse clears B_ERROR and b_error * brelse clears B_ERROR and b_error
*/ */

View File

@ -42,8 +42,6 @@ xfs_fsb_to_db(struct xfs_inode *ip, xfs_fsblock_t fsb)
extern int xfs_read_buf(struct xfs_mount *mp, xfs_buftarg_t *btp, extern int xfs_read_buf(struct xfs_mount *mp, xfs_buftarg_t *btp,
xfs_daddr_t blkno, int len, uint flags, xfs_daddr_t blkno, int len, uint flags,
struct xfs_buf **bpp); struct xfs_buf **bpp);
extern void xfs_ioerror_alert(char *func, struct xfs_mount *mp,
xfs_buf_t *bp, xfs_daddr_t blkno);
extern xfs_extlen_t xfs_get_extsz_hint(struct xfs_inode *ip); extern xfs_extlen_t xfs_get_extsz_hint(struct xfs_inode *ip);
#endif /* __XFS_RW_H__ */ #endif /* __XFS_RW_H__ */

View File

@ -796,8 +796,6 @@ xfs_fs_destroy_inode(
if (is_bad_inode(inode)) if (is_bad_inode(inode))
goto out_reclaim; goto out_reclaim;
xfs_ioend_wait(ip);
ASSERT(XFS_FORCED_SHUTDOWN(ip->i_mount) || ip->i_delayed_blks == 0); ASSERT(XFS_FORCED_SHUTDOWN(ip->i_mount) || ip->i_delayed_blks == 0);
/* /*
@ -837,7 +835,6 @@ xfs_fs_inode_init_once(
inode_init_once(VFS_I(ip)); inode_init_once(VFS_I(ip));
/* xfs inode */ /* xfs inode */
atomic_set(&ip->i_iocount, 0);
atomic_set(&ip->i_pincount, 0); atomic_set(&ip->i_pincount, 0);
spin_lock_init(&ip->i_flags_lock); spin_lock_init(&ip->i_flags_lock);
init_waitqueue_head(&ip->i_ipin_wait); init_waitqueue_head(&ip->i_ipin_wait);
@ -887,7 +884,7 @@ xfs_log_inode(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
return xfs_trans_commit(tp, 0); return xfs_trans_commit(tp, 0);
} }
@ -914,9 +911,8 @@ xfs_fs_write_inode(
* of forcing it all the way to stable storage using a * of forcing it all the way to stable storage using a
* synchronous transaction we let the log force inside the * synchronous transaction we let the log force inside the
* ->sync_fs call do that for thus, which reduces the number * ->sync_fs call do that for thus, which reduces the number
* of synchronous log foces dramatically. * of synchronous log forces dramatically.
*/ */
xfs_ioend_wait(ip);
error = xfs_log_inode(ip); error = xfs_log_inode(ip);
if (error) if (error)
goto out; goto out;
@ -1019,7 +1015,7 @@ xfs_fs_put_super(
*/ */
xfs_filestream_unmount(mp); xfs_filestream_unmount(mp);
XFS_bflush(mp->m_ddev_targp); xfs_flush_buftarg(mp->m_ddev_targp, 1);
xfs_unmountfs(mp); xfs_unmountfs(mp);
xfs_freesb(mp); xfs_freesb(mp);
@ -1443,7 +1439,7 @@ xfs_fs_fill_super(
*/ */
xfs_filestream_unmount(mp); xfs_filestream_unmount(mp);
XFS_bflush(mp->m_ddev_targp); xfs_flush_buftarg(mp->m_ddev_targp, 1);
xfs_unmountfs(mp); xfs_unmountfs(mp);
goto out_free_sb; goto out_free_sb;
@ -1670,7 +1666,6 @@ init_xfs_fs(void)
printk(KERN_INFO XFS_VERSION_STRING " with " printk(KERN_INFO XFS_VERSION_STRING " with "
XFS_BUILD_OPTIONS " enabled\n"); XFS_BUILD_OPTIONS " enabled\n");
xfs_ioend_init();
xfs_dir_startup(); xfs_dir_startup();
error = xfs_init_zones(); error = xfs_init_zones();

View File

@ -227,21 +227,17 @@ xfs_sync_inode_data(
int error = 0; int error = 0;
if (!mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) if (!mapping_tagged(mapping, PAGECACHE_TAG_DIRTY))
goto out_wait; return 0;
if (!xfs_ilock_nowait(ip, XFS_IOLOCK_SHARED)) { if (!xfs_ilock_nowait(ip, XFS_IOLOCK_SHARED)) {
if (flags & SYNC_TRYLOCK) if (flags & SYNC_TRYLOCK)
goto out_wait; return 0;
xfs_ilock(ip, XFS_IOLOCK_SHARED); xfs_ilock(ip, XFS_IOLOCK_SHARED);
} }
error = xfs_flush_pages(ip, 0, -1, (flags & SYNC_WAIT) ? error = xfs_flush_pages(ip, 0, -1, (flags & SYNC_WAIT) ?
0 : XBF_ASYNC, FI_NONE); 0 : XBF_ASYNC, FI_NONE);
xfs_iunlock(ip, XFS_IOLOCK_SHARED); xfs_iunlock(ip, XFS_IOLOCK_SHARED);
out_wait:
if (flags & SYNC_WAIT)
xfs_ioend_wait(ip);
return error; return error;
} }
@ -322,6 +318,7 @@ xfs_sync_fsdata(
struct xfs_mount *mp) struct xfs_mount *mp)
{ {
struct xfs_buf *bp; struct xfs_buf *bp;
int error;
/* /*
* If the buffer is pinned then push on the log so we won't get stuck * If the buffer is pinned then push on the log so we won't get stuck
@ -334,8 +331,9 @@ xfs_sync_fsdata(
bp = xfs_getsb(mp, 0); bp = xfs_getsb(mp, 0);
if (xfs_buf_ispinned(bp)) if (xfs_buf_ispinned(bp))
xfs_log_force(mp, 0); xfs_log_force(mp, 0);
error = xfs_bwrite(bp);
return xfs_bwrite(mp, bp); xfs_buf_relse(bp);
return error;
} }
/* /*
@ -379,7 +377,7 @@ xfs_quiesce_data(
/* flush data-only devices */ /* flush data-only devices */
if (mp->m_rtdev_targp) if (mp->m_rtdev_targp)
XFS_bflush(mp->m_rtdev_targp); xfs_flush_buftarg(mp->m_rtdev_targp, 1);
return error ? error : error2; return error ? error : error2;
} }

View File

@ -30,6 +30,7 @@ struct xfs_buf_log_item;
struct xfs_da_args; struct xfs_da_args;
struct xfs_da_node_entry; struct xfs_da_node_entry;
struct xfs_dquot; struct xfs_dquot;
struct xfs_log_item;
struct xlog_ticket; struct xlog_ticket;
struct log; struct log;
struct xlog_recover; struct xlog_recover;
@ -320,7 +321,6 @@ DEFINE_BUF_EVENT(xfs_buf_rele);
DEFINE_BUF_EVENT(xfs_buf_iodone); DEFINE_BUF_EVENT(xfs_buf_iodone);
DEFINE_BUF_EVENT(xfs_buf_iorequest); DEFINE_BUF_EVENT(xfs_buf_iorequest);
DEFINE_BUF_EVENT(xfs_buf_bawrite); DEFINE_BUF_EVENT(xfs_buf_bawrite);
DEFINE_BUF_EVENT(xfs_buf_bdwrite);
DEFINE_BUF_EVENT(xfs_buf_lock); DEFINE_BUF_EVENT(xfs_buf_lock);
DEFINE_BUF_EVENT(xfs_buf_lock_done); DEFINE_BUF_EVENT(xfs_buf_lock_done);
DEFINE_BUF_EVENT(xfs_buf_trylock); DEFINE_BUF_EVENT(xfs_buf_trylock);
@ -577,6 +577,7 @@ DEFINE_INODE_EVENT(xfs_vm_bmap);
DEFINE_INODE_EVENT(xfs_file_ioctl); DEFINE_INODE_EVENT(xfs_file_ioctl);
DEFINE_INODE_EVENT(xfs_file_compat_ioctl); DEFINE_INODE_EVENT(xfs_file_compat_ioctl);
DEFINE_INODE_EVENT(xfs_ioctl_setattr); DEFINE_INODE_EVENT(xfs_ioctl_setattr);
DEFINE_INODE_EVENT(xfs_dir_fsync);
DEFINE_INODE_EVENT(xfs_file_fsync); DEFINE_INODE_EVENT(xfs_file_fsync);
DEFINE_INODE_EVENT(xfs_destroy_inode); DEFINE_INODE_EVENT(xfs_destroy_inode);
DEFINE_INODE_EVENT(xfs_write_inode); DEFINE_INODE_EVENT(xfs_write_inode);
@ -853,6 +854,42 @@ DEFINE_LOGGRANT_EVENT(xfs_log_ungrant_enter);
DEFINE_LOGGRANT_EVENT(xfs_log_ungrant_exit); DEFINE_LOGGRANT_EVENT(xfs_log_ungrant_exit);
DEFINE_LOGGRANT_EVENT(xfs_log_ungrant_sub); DEFINE_LOGGRANT_EVENT(xfs_log_ungrant_sub);
DECLARE_EVENT_CLASS(xfs_log_item_class,
TP_PROTO(struct xfs_log_item *lip),
TP_ARGS(lip),
TP_STRUCT__entry(
__field(dev_t, dev)
__field(void *, lip)
__field(uint, type)
__field(uint, flags)
__field(xfs_lsn_t, lsn)
),
TP_fast_assign(
__entry->dev = lip->li_mountp->m_super->s_dev;
__entry->lip = lip;
__entry->type = lip->li_type;
__entry->flags = lip->li_flags;
__entry->lsn = lip->li_lsn;
),
TP_printk("dev %d:%d lip 0x%p lsn %d/%d type %s flags %s",
MAJOR(__entry->dev), MINOR(__entry->dev),
__entry->lip,
CYCLE_LSN(__entry->lsn), BLOCK_LSN(__entry->lsn),
__print_symbolic(__entry->type, XFS_LI_TYPE_DESC),
__print_flags(__entry->flags, "|", XFS_LI_FLAGS))
)
#define DEFINE_LOG_ITEM_EVENT(name) \
DEFINE_EVENT(xfs_log_item_class, name, \
TP_PROTO(struct xfs_log_item *lip), \
TP_ARGS(lip))
DEFINE_LOG_ITEM_EVENT(xfs_ail_push);
DEFINE_LOG_ITEM_EVENT(xfs_ail_pushbuf);
DEFINE_LOG_ITEM_EVENT(xfs_ail_pushbuf_pinned);
DEFINE_LOG_ITEM_EVENT(xfs_ail_pinned);
DEFINE_LOG_ITEM_EVENT(xfs_ail_locked);
DECLARE_EVENT_CLASS(xfs_file_class, DECLARE_EVENT_CLASS(xfs_file_class,
TP_PROTO(struct xfs_inode *ip, size_t count, loff_t offset, int flags), TP_PROTO(struct xfs_inode *ip, size_t count, loff_t offset, int flags),
TP_ARGS(ip, count, offset, flags), TP_ARGS(ip, count, offset, flags),

View File

@ -1790,9 +1790,7 @@ xfs_trans_commit_cil(
} }
/* /*
* xfs_trans_commit * Commit the given transaction to the log.
*
* Commit the given transaction to the log a/synchronously.
* *
* XFS disk error handling mechanism is not based on a typical * XFS disk error handling mechanism is not based on a typical
* transaction abort mechanism. Logically after the filesystem * transaction abort mechanism. Logically after the filesystem
@ -1804,10 +1802,9 @@ xfs_trans_commit_cil(
* Do not reference the transaction structure after this call. * Do not reference the transaction structure after this call.
*/ */
int int
_xfs_trans_commit( xfs_trans_commit(
struct xfs_trans *tp, struct xfs_trans *tp,
uint flags, uint flags)
int *log_flushed)
{ {
struct xfs_mount *mp = tp->t_mountp; struct xfs_mount *mp = tp->t_mountp;
xfs_lsn_t commit_lsn = -1; xfs_lsn_t commit_lsn = -1;
@ -1866,7 +1863,7 @@ _xfs_trans_commit(
if (sync) { if (sync) {
if (!error) { if (!error) {
error = _xfs_log_force_lsn(mp, commit_lsn, error = _xfs_log_force_lsn(mp, commit_lsn,
XFS_LOG_SYNC, log_flushed); XFS_LOG_SYNC, NULL);
} }
XFS_STATS_INC(xs_trans_sync); XFS_STATS_INC(xs_trans_sync);
} else { } else {
@ -2021,6 +2018,6 @@ xfs_trans_roll(
if (error) if (error)
return error; return error;
xfs_trans_ijoin(trans, dp); xfs_trans_ijoin(trans, dp, 0);
return 0; return 0;
} }

View File

@ -470,8 +470,7 @@ void xfs_trans_stale_inode_buf(xfs_trans_t *, struct xfs_buf *);
void xfs_trans_dquot_buf(xfs_trans_t *, struct xfs_buf *, uint); void xfs_trans_dquot_buf(xfs_trans_t *, struct xfs_buf *, uint);
void xfs_trans_inode_alloc_buf(xfs_trans_t *, struct xfs_buf *); void xfs_trans_inode_alloc_buf(xfs_trans_t *, struct xfs_buf *);
void xfs_trans_ichgtime(struct xfs_trans *, struct xfs_inode *, int); void xfs_trans_ichgtime(struct xfs_trans *, struct xfs_inode *, int);
void xfs_trans_ijoin_ref(struct xfs_trans *, struct xfs_inode *, uint); void xfs_trans_ijoin(struct xfs_trans *, struct xfs_inode *, uint);
void xfs_trans_ijoin(struct xfs_trans *, struct xfs_inode *);
void xfs_trans_log_buf(xfs_trans_t *, struct xfs_buf *, uint, uint); void xfs_trans_log_buf(xfs_trans_t *, struct xfs_buf *, uint, uint);
void xfs_trans_log_inode(xfs_trans_t *, struct xfs_inode *, uint); void xfs_trans_log_inode(xfs_trans_t *, struct xfs_inode *, uint);
struct xfs_efi_log_item *xfs_trans_get_efi(xfs_trans_t *, uint); struct xfs_efi_log_item *xfs_trans_get_efi(xfs_trans_t *, uint);
@ -487,10 +486,7 @@ void xfs_trans_log_efd_extent(xfs_trans_t *,
struct xfs_efd_log_item *, struct xfs_efd_log_item *,
xfs_fsblock_t, xfs_fsblock_t,
xfs_extlen_t); xfs_extlen_t);
int _xfs_trans_commit(xfs_trans_t *, int xfs_trans_commit(xfs_trans_t *, uint flags);
uint flags,
int *);
#define xfs_trans_commit(tp, flags) _xfs_trans_commit(tp, flags, NULL)
void xfs_trans_cancel(xfs_trans_t *, int); void xfs_trans_cancel(xfs_trans_t *, int);
int xfs_trans_ail_init(struct xfs_mount *); int xfs_trans_ail_init(struct xfs_mount *);
void xfs_trans_ail_destroy(struct xfs_mount *); void xfs_trans_ail_destroy(struct xfs_mount *);

View File

@ -26,6 +26,7 @@
#include "xfs_ag.h" #include "xfs_ag.h"
#include "xfs_mount.h" #include "xfs_mount.h"
#include "xfs_trans_priv.h" #include "xfs_trans_priv.h"
#include "xfs_trace.h"
#include "xfs_error.h" #include "xfs_error.h"
#ifdef DEBUG #ifdef DEBUG
@ -364,12 +365,24 @@ xfsaild_push(
xfs_lsn_t lsn; xfs_lsn_t lsn;
xfs_lsn_t target; xfs_lsn_t target;
long tout = 10; long tout = 10;
int flush_log = 0;
int stuck = 0; int stuck = 0;
int count = 0; int count = 0;
int push_xfsbufd = 0; int push_xfsbufd = 0;
/*
* If last time we ran we encountered pinned items, force the log first
* and wait for it before pushing again.
*/
spin_lock(&ailp->xa_lock); spin_lock(&ailp->xa_lock);
if (ailp->xa_last_pushed_lsn == 0 && ailp->xa_log_flush &&
!list_empty(&ailp->xa_ail)) {
ailp->xa_log_flush = 0;
spin_unlock(&ailp->xa_lock);
XFS_STATS_INC(xs_push_ail_flush);
xfs_log_force(mp, XFS_LOG_SYNC);
spin_lock(&ailp->xa_lock);
}
target = ailp->xa_target; target = ailp->xa_target;
lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->xa_last_pushed_lsn); lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->xa_last_pushed_lsn);
if (!lip || XFS_FORCED_SHUTDOWN(mp)) { if (!lip || XFS_FORCED_SHUTDOWN(mp)) {
@ -413,16 +426,20 @@ xfsaild_push(
switch (lock_result) { switch (lock_result) {
case XFS_ITEM_SUCCESS: case XFS_ITEM_SUCCESS:
XFS_STATS_INC(xs_push_ail_success); XFS_STATS_INC(xs_push_ail_success);
trace_xfs_ail_push(lip);
IOP_PUSH(lip); IOP_PUSH(lip);
ailp->xa_last_pushed_lsn = lsn; ailp->xa_last_pushed_lsn = lsn;
break; break;
case XFS_ITEM_PUSHBUF: case XFS_ITEM_PUSHBUF:
XFS_STATS_INC(xs_push_ail_pushbuf); XFS_STATS_INC(xs_push_ail_pushbuf);
trace_xfs_ail_pushbuf(lip);
if (!IOP_PUSHBUF(lip)) { if (!IOP_PUSHBUF(lip)) {
trace_xfs_ail_pushbuf_pinned(lip);
stuck++; stuck++;
flush_log = 1; ailp->xa_log_flush++;
} else { } else {
ailp->xa_last_pushed_lsn = lsn; ailp->xa_last_pushed_lsn = lsn;
} }
@ -431,12 +448,15 @@ xfsaild_push(
case XFS_ITEM_PINNED: case XFS_ITEM_PINNED:
XFS_STATS_INC(xs_push_ail_pinned); XFS_STATS_INC(xs_push_ail_pinned);
trace_xfs_ail_pinned(lip);
stuck++; stuck++;
flush_log = 1; ailp->xa_log_flush++;
break; break;
case XFS_ITEM_LOCKED: case XFS_ITEM_LOCKED:
XFS_STATS_INC(xs_push_ail_locked); XFS_STATS_INC(xs_push_ail_locked);
trace_xfs_ail_locked(lip);
stuck++; stuck++;
break; break;
@ -476,16 +496,6 @@ xfsaild_push(
xfs_trans_ail_cursor_done(ailp, &cur); xfs_trans_ail_cursor_done(ailp, &cur);
spin_unlock(&ailp->xa_lock); spin_unlock(&ailp->xa_lock);
if (flush_log) {
/*
* If something we need to push out was pinned, then
* push out the log so it will become unpinned and
* move forward in the AIL.
*/
XFS_STATS_INC(xs_push_ail_flush);
xfs_log_force(mp, 0);
}
if (push_xfsbufd) { if (push_xfsbufd) {
/* we've got delayed write buffers to flush */ /* we've got delayed write buffers to flush */
wake_up_process(mp->m_ddev_targp->bt_task); wake_up_process(mp->m_ddev_targp->bt_task);
@ -496,6 +506,7 @@ out_done:
if (!count) { if (!count) {
/* We're past our target or empty, so idle */ /* We're past our target or empty, so idle */
ailp->xa_last_pushed_lsn = 0; ailp->xa_last_pushed_lsn = 0;
ailp->xa_log_flush = 0;
tout = 50; tout = 50;
} else if (XFS_LSN_CMP(lsn, target) >= 0) { } else if (XFS_LSN_CMP(lsn, target) >= 0) {
@ -514,9 +525,13 @@ out_done:
* were stuck. * were stuck.
* *
* Backoff a bit more to allow some I/O to complete before * Backoff a bit more to allow some I/O to complete before
* continuing from where we were. * restarting from the start of the AIL. This prevents us
* from spinning on the same items, and if they are pinned will
* all the restart to issue a log force to unpin the stuck
* items.
*/ */
tout = 20; tout = 20;
ailp->xa_last_pushed_lsn = 0;
} }
return tout; return tout;

View File

@ -160,8 +160,10 @@ xfs_trans_get_buf(xfs_trans_t *tp,
bp = xfs_trans_buf_item_match(tp, target_dev, blkno, len); bp = xfs_trans_buf_item_match(tp, target_dev, blkno, len);
if (bp != NULL) { if (bp != NULL) {
ASSERT(xfs_buf_islocked(bp)); ASSERT(xfs_buf_islocked(bp));
if (XFS_FORCED_SHUTDOWN(tp->t_mountp)) if (XFS_FORCED_SHUTDOWN(tp->t_mountp)) {
XFS_BUF_SUPER_STALE(bp); xfs_buf_stale(bp);
XFS_BUF_DONE(bp);
}
/* /*
* If the buffer is stale then it was binval'ed * If the buffer is stale then it was binval'ed
@ -294,8 +296,7 @@ xfs_trans_read_buf(
if (bp->b_error) { if (bp->b_error) {
error = bp->b_error; error = bp->b_error;
xfs_ioerror_alert("xfs_trans_read_buf", mp, xfs_buf_ioerror_alert(bp, __func__);
bp, blkno);
xfs_buf_relse(bp); xfs_buf_relse(bp);
return error; return error;
} }
@ -337,8 +338,7 @@ xfs_trans_read_buf(
xfsbdstrat(tp->t_mountp, bp); xfsbdstrat(tp->t_mountp, bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) { if (error) {
xfs_ioerror_alert("xfs_trans_read_buf", mp, xfs_buf_ioerror_alert(bp, __func__);
bp, blkno);
xfs_buf_relse(bp); xfs_buf_relse(bp);
/* /*
* We can gracefully recover from most read * We can gracefully recover from most read
@ -387,9 +387,9 @@ xfs_trans_read_buf(
} }
if (bp->b_error) { if (bp->b_error) {
error = bp->b_error; error = bp->b_error;
XFS_BUF_SUPER_STALE(bp); xfs_buf_stale(bp);
xfs_ioerror_alert("xfs_trans_read_buf", mp, XFS_BUF_DONE(bp);
bp, blkno); xfs_buf_ioerror_alert(bp, __func__);
if (tp->t_flags & XFS_TRANS_DIRTY) if (tp->t_flags & XFS_TRANS_DIRTY)
xfs_force_shutdown(tp->t_mountp, SHUTDOWN_META_IO_ERROR); xfs_force_shutdown(tp->t_mountp, SHUTDOWN_META_IO_ERROR);
xfs_buf_relse(bp); xfs_buf_relse(bp);
@ -643,13 +643,14 @@ xfs_trans_log_buf(xfs_trans_t *tp,
* inside the b_bdstrat callback so that this won't get written to * inside the b_bdstrat callback so that this won't get written to
* disk. * disk.
*/ */
XFS_BUF_DELAYWRITE(bp);
XFS_BUF_DONE(bp); XFS_BUF_DONE(bp);
ASSERT(atomic_read(&bip->bli_refcount) > 0); ASSERT(atomic_read(&bip->bli_refcount) > 0);
bp->b_iodone = xfs_buf_iodone_callbacks; bp->b_iodone = xfs_buf_iodone_callbacks;
bip->bli_item.li_cb = xfs_buf_iodone; bip->bli_item.li_cb = xfs_buf_iodone;
xfs_buf_delwri_queue(bp);
trace_xfs_trans_log_buf(bip); trace_xfs_trans_log_buf(bip);
/* /*
@ -738,8 +739,7 @@ xfs_trans_binval(
* We set the stale bit in the buffer as well since we're getting * We set the stale bit in the buffer as well since we're getting
* rid of it. * rid of it.
*/ */
XFS_BUF_UNDELAYWRITE(bp); xfs_buf_stale(bp);
XFS_BUF_STALE(bp);
bip->bli_flags |= XFS_BLI_STALE; bip->bli_flags |= XFS_BLI_STALE;
bip->bli_flags &= ~(XFS_BLI_INODE_BUF | XFS_BLI_LOGGED | XFS_BLI_DIRTY); bip->bli_flags &= ~(XFS_BLI_INODE_BUF | XFS_BLI_LOGGED | XFS_BLI_DIRTY);
bip->bli_format.blf_flags &= ~XFS_BLF_INODE_BUF; bip->bli_format.blf_flags &= ~XFS_BLF_INODE_BUF;

View File

@ -47,11 +47,13 @@ xfs_trans_inode_broot_debug(
* Add a locked inode to the transaction. * Add a locked inode to the transaction.
* *
* The inode must be locked, and it cannot be associated with any transaction. * The inode must be locked, and it cannot be associated with any transaction.
* If lock_flags is non-zero the inode will be unlocked on transaction commit.
*/ */
void void
xfs_trans_ijoin( xfs_trans_ijoin(
struct xfs_trans *tp, struct xfs_trans *tp,
struct xfs_inode *ip) struct xfs_inode *ip,
uint lock_flags)
{ {
xfs_inode_log_item_t *iip; xfs_inode_log_item_t *iip;
@ -59,7 +61,9 @@ xfs_trans_ijoin(
if (ip->i_itemp == NULL) if (ip->i_itemp == NULL)
xfs_inode_item_init(ip, ip->i_mount); xfs_inode_item_init(ip, ip->i_mount);
iip = ip->i_itemp; iip = ip->i_itemp;
ASSERT(iip->ili_lock_flags == 0); ASSERT(iip->ili_lock_flags == 0);
iip->ili_lock_flags = lock_flags;
/* /*
* Get a log_item_desc to point at the new item. * Get a log_item_desc to point at the new item.
@ -69,25 +73,6 @@ xfs_trans_ijoin(
xfs_trans_inode_broot_debug(ip); xfs_trans_inode_broot_debug(ip);
} }
/*
* Add a locked inode to the transaction.
*
*
* Grabs a reference to the inode which will be dropped when the transaction
* is committed. The inode will also be unlocked at that point. The inode
* must be locked, and it cannot be associated with any transaction.
*/
void
xfs_trans_ijoin_ref(
struct xfs_trans *tp,
struct xfs_inode *ip,
uint lock_flags)
{
xfs_trans_ijoin(tp, ip);
IHOLD(ip);
ip->i_itemp->ili_lock_flags = lock_flags;
}
/* /*
* Transactional inode timestamp update. Requires the inode to be locked and * Transactional inode timestamp update. Requires the inode to be locked and
* joined to the transaction supplied. Relies on the transaction subsystem to * joined to the transaction supplied. Relies on the transaction subsystem to

View File

@ -70,6 +70,7 @@ struct xfs_ail {
struct list_head xa_cursors; struct list_head xa_cursors;
spinlock_t xa_lock; spinlock_t xa_lock;
xfs_lsn_t xa_last_pushed_lsn; xfs_lsn_t xa_last_pushed_lsn;
int xa_log_flush;
}; };
/* /*

View File

@ -72,8 +72,8 @@ xfs_readlink_bmap(
xfs_buf_t *bp; xfs_buf_t *bp;
int error = 0; int error = 0;
error = xfs_bmapi(NULL, ip, 0, XFS_B_TO_FSB(mp, pathlen), 0, NULL, 0, error = xfs_bmapi_read(ip, 0, XFS_B_TO_FSB(mp, pathlen), mval, &nmaps,
mval, &nmaps, NULL); 0);
if (error) if (error)
goto out; goto out;
@ -87,8 +87,7 @@ xfs_readlink_bmap(
return XFS_ERROR(ENOMEM); return XFS_ERROR(ENOMEM);
error = bp->b_error; error = bp->b_error;
if (error) { if (error) {
xfs_ioerror_alert("xfs_readlink", xfs_buf_ioerror_alert(bp, __func__);
ip->i_mount, bp, XFS_BUF_ADDR(bp));
xfs_buf_relse(bp); xfs_buf_relse(bp);
goto out; goto out;
} }
@ -178,8 +177,7 @@ xfs_free_eofblocks(
nimaps = 1; nimaps = 1;
xfs_ilock(ip, XFS_ILOCK_SHARED); xfs_ilock(ip, XFS_ILOCK_SHARED);
error = xfs_bmapi(NULL, ip, end_fsb, map_len, 0, error = xfs_bmapi_read(ip, end_fsb, map_len, &imap, &nimaps, 0);
NULL, 0, &imap, &nimaps, NULL);
xfs_iunlock(ip, XFS_ILOCK_SHARED); xfs_iunlock(ip, XFS_ILOCK_SHARED);
if (!error && (nimaps != 0) && if (!error && (nimaps != 0) &&
@ -220,7 +218,7 @@ xfs_free_eofblocks(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
error = xfs_itruncate_data(&tp, ip, ip->i_size); error = xfs_itruncate_data(&tp, ip, ip->i_size);
if (error) { if (error) {
@ -289,7 +287,7 @@ xfs_inactive_symlink_rmt(
xfs_ilock(ip, XFS_IOLOCK_EXCL | XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_IOLOCK_EXCL | XFS_ILOCK_EXCL);
size = (int)ip->i_d.di_size; size = (int)ip->i_d.di_size;
ip->i_d.di_size = 0; ip->i_d.di_size = 0;
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
/* /*
* Find the block(s) so we can inval and unmap them. * Find the block(s) so we can inval and unmap them.
@ -297,9 +295,9 @@ xfs_inactive_symlink_rmt(
done = 0; done = 0;
xfs_bmap_init(&free_list, &first_block); xfs_bmap_init(&free_list, &first_block);
nmaps = ARRAY_SIZE(mval); nmaps = ARRAY_SIZE(mval);
if ((error = xfs_bmapi(tp, ip, 0, XFS_B_TO_FSB(mp, size), error = xfs_bmapi_read(ip, 0, XFS_B_TO_FSB(mp, size),
XFS_BMAPI_METADATA, &first_block, 0, mval, &nmaps, mval, &nmaps, 0);
&free_list))) if (error)
goto error0; goto error0;
/* /*
* Invalidate the block(s). * Invalidate the block(s).
@ -308,6 +306,10 @@ xfs_inactive_symlink_rmt(
bp = xfs_trans_get_buf(tp, mp->m_ddev_targp, bp = xfs_trans_get_buf(tp, mp->m_ddev_targp,
XFS_FSB_TO_DADDR(mp, mval[i].br_startblock), XFS_FSB_TO_DADDR(mp, mval[i].br_startblock),
XFS_FSB_TO_BB(mp, mval[i].br_blockcount), 0); XFS_FSB_TO_BB(mp, mval[i].br_blockcount), 0);
if (!bp) {
error = ENOMEM;
goto error1;
}
xfs_trans_binval(tp, bp); xfs_trans_binval(tp, bp);
} }
/* /*
@ -333,7 +335,7 @@ xfs_inactive_symlink_rmt(
* Mark it dirty so it will be logged and moved forward in the log as * Mark it dirty so it will be logged and moved forward in the log as
* part of every commit. * part of every commit.
*/ */
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
/* /*
* Get a new, empty transaction to return to our caller. * Get a new, empty transaction to return to our caller.
@ -466,7 +468,7 @@ xfs_inactive_attrs(
goto error_cancel; goto error_cancel;
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
xfs_idestroy_fork(ip, XFS_ATTR_FORK); xfs_idestroy_fork(ip, XFS_ATTR_FORK);
ASSERT(ip->i_d.di_anextents == 0); ASSERT(ip->i_d.di_anextents == 0);
@ -647,8 +649,6 @@ xfs_inactive(
if (truncate) { if (truncate) {
xfs_ilock(ip, XFS_IOLOCK_EXCL); xfs_ilock(ip, XFS_IOLOCK_EXCL);
xfs_ioend_wait(ip);
error = xfs_trans_reserve(tp, 0, error = xfs_trans_reserve(tp, 0,
XFS_ITRUNCATE_LOG_RES(mp), XFS_ITRUNCATE_LOG_RES(mp),
0, XFS_TRANS_PERM_LOG_RES, 0, XFS_TRANS_PERM_LOG_RES,
@ -662,7 +662,7 @@ xfs_inactive(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
error = xfs_itruncate_data(&tp, ip, 0); error = xfs_itruncate_data(&tp, ip, 0);
if (error) { if (error) {
@ -686,7 +686,7 @@ xfs_inactive(
return VN_INACTIVE_CACHE; return VN_INACTIVE_CACHE;
} }
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
} else { } else {
error = xfs_trans_reserve(tp, 0, error = xfs_trans_reserve(tp, 0,
XFS_IFREE_LOG_RES(mp), XFS_IFREE_LOG_RES(mp),
@ -699,7 +699,7 @@ xfs_inactive(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL);
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
} }
/* /*
@ -939,7 +939,7 @@ xfs_create(
* the transaction cancel unlocking dp so don't do it explicitly in the * the transaction cancel unlocking dp so don't do it explicitly in the
* error path. * error path.
*/ */
xfs_trans_ijoin_ref(tp, dp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, dp, XFS_ILOCK_EXCL);
unlock_dp_on_error = B_FALSE; unlock_dp_on_error = B_FALSE;
error = xfs_dir_createname(tp, dp, name, ip->i_ino, error = xfs_dir_createname(tp, dp, name, ip->i_ino,
@ -1260,8 +1260,8 @@ xfs_remove(
xfs_lock_two_inodes(dp, ip, XFS_ILOCK_EXCL); xfs_lock_two_inodes(dp, ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, dp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, dp, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
/* /*
* If we're removing a directory perform some additional validation. * If we're removing a directory perform some additional validation.
@ -1406,8 +1406,8 @@ xfs_link(
xfs_lock_two_inodes(sip, tdp, XFS_ILOCK_EXCL); xfs_lock_two_inodes(sip, tdp, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, sip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, sip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, tdp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, tdp, XFS_ILOCK_EXCL);
/* /*
* If the source has too many links, we can't make any more to it. * If the source has too many links, we can't make any more to it.
@ -1601,7 +1601,7 @@ xfs_symlink(
* transaction cancel unlocking dp so don't do it explicitly in the * transaction cancel unlocking dp so don't do it explicitly in the
* error path. * error path.
*/ */
xfs_trans_ijoin_ref(tp, dp, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, dp, XFS_ILOCK_EXCL);
unlock_dp_on_error = B_FALSE; unlock_dp_on_error = B_FALSE;
/* /*
@ -1632,10 +1632,9 @@ xfs_symlink(
first_fsb = 0; first_fsb = 0;
nmaps = SYMLINK_MAPS; nmaps = SYMLINK_MAPS;
error = xfs_bmapi(tp, ip, first_fsb, fs_blocks, error = xfs_bmapi_write(tp, ip, first_fsb, fs_blocks,
XFS_BMAPI_WRITE | XFS_BMAPI_METADATA, XFS_BMAPI_METADATA, &first_block, resblks,
&first_block, resblks, mval, &nmaps, mval, &nmaps, &free_list);
&free_list);
if (error) if (error)
goto error2; goto error2;
@ -1650,7 +1649,10 @@ xfs_symlink(
byte_cnt = XFS_FSB_TO_B(mp, mval[n].br_blockcount); byte_cnt = XFS_FSB_TO_B(mp, mval[n].br_blockcount);
bp = xfs_trans_get_buf(tp, mp->m_ddev_targp, d, bp = xfs_trans_get_buf(tp, mp->m_ddev_targp, d,
BTOBB(byte_cnt), 0); BTOBB(byte_cnt), 0);
ASSERT(!xfs_buf_geterror(bp)); if (!bp) {
error = ENOMEM;
goto error2;
}
if (pathlen < byte_cnt) { if (pathlen < byte_cnt) {
byte_cnt = pathlen; byte_cnt = pathlen;
} }
@ -1732,7 +1734,7 @@ xfs_set_dmattrs(
return error; return error;
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin_ref(tp, ip, XFS_ILOCK_EXCL); xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
ip->i_d.di_dmevmask = evmask; ip->i_d.di_dmevmask = evmask;
ip->i_d.di_dmstate = state; ip->i_d.di_dmstate = state;
@ -1778,7 +1780,6 @@ xfs_alloc_file_space(
xfs_fileoff_t startoffset_fsb; xfs_fileoff_t startoffset_fsb;
xfs_fsblock_t firstfsb; xfs_fsblock_t firstfsb;
int nimaps; int nimaps;
int bmapi_flag;
int quota_flag; int quota_flag;
int rt; int rt;
xfs_trans_t *tp; xfs_trans_t *tp;
@ -1806,7 +1807,6 @@ xfs_alloc_file_space(
count = len; count = len;
imapp = &imaps[0]; imapp = &imaps[0];
nimaps = 1; nimaps = 1;
bmapi_flag = XFS_BMAPI_WRITE | alloc_type;
startoffset_fsb = XFS_B_TO_FSBT(mp, offset); startoffset_fsb = XFS_B_TO_FSBT(mp, offset);
allocatesize_fsb = XFS_B_TO_FSB(mp, count); allocatesize_fsb = XFS_B_TO_FSB(mp, count);
@ -1877,16 +1877,12 @@ xfs_alloc_file_space(
if (error) if (error)
goto error1; goto error1;
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/*
* Issue the xfs_bmapi() call to allocate the blocks
*/
xfs_bmap_init(&free_list, &firstfsb); xfs_bmap_init(&free_list, &firstfsb);
error = xfs_bmapi(tp, ip, startoffset_fsb, error = xfs_bmapi_write(tp, ip, startoffset_fsb,
allocatesize_fsb, bmapi_flag, allocatesize_fsb, alloc_type, &firstfsb,
&firstfsb, 0, imapp, &nimaps, 0, imapp, &nimaps, &free_list);
&free_list);
if (error) { if (error) {
goto error0; goto error0;
} }
@ -1976,8 +1972,7 @@ xfs_zero_remaining_bytes(
for (offset = startoff; offset <= endoff; offset = lastoffset + 1) { for (offset = startoff; offset <= endoff; offset = lastoffset + 1) {
offset_fsb = XFS_B_TO_FSBT(mp, offset); offset_fsb = XFS_B_TO_FSBT(mp, offset);
nimap = 1; nimap = 1;
error = xfs_bmapi(NULL, ip, offset_fsb, 1, 0, error = xfs_bmapi_read(ip, offset_fsb, 1, &imap, &nimap, 0);
NULL, 0, &imap, &nimap, NULL);
if (error || nimap < 1) if (error || nimap < 1)
break; break;
ASSERT(imap.br_blockcount >= 1); ASSERT(imap.br_blockcount >= 1);
@ -1997,8 +1992,8 @@ xfs_zero_remaining_bytes(
xfsbdstrat(mp, bp); xfsbdstrat(mp, bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) { if (error) {
xfs_ioerror_alert("xfs_zero_remaining_bytes(read)", xfs_buf_ioerror_alert(bp,
mp, bp, XFS_BUF_ADDR(bp)); "xfs_zero_remaining_bytes(read)");
break; break;
} }
memset(bp->b_addr + memset(bp->b_addr +
@ -2010,8 +2005,8 @@ xfs_zero_remaining_bytes(
xfsbdstrat(mp, bp); xfsbdstrat(mp, bp);
error = xfs_buf_iowait(bp); error = xfs_buf_iowait(bp);
if (error) { if (error) {
xfs_ioerror_alert("xfs_zero_remaining_bytes(write)", xfs_buf_ioerror_alert(bp,
mp, bp, XFS_BUF_ADDR(bp)); "xfs_zero_remaining_bytes(write)");
break; break;
} }
} }
@ -2076,7 +2071,7 @@ xfs_free_file_space(
if (need_iolock) { if (need_iolock) {
xfs_ilock(ip, XFS_IOLOCK_EXCL); xfs_ilock(ip, XFS_IOLOCK_EXCL);
/* wait for the completion of any pending DIOs */ /* wait for the completion of any pending DIOs */
xfs_ioend_wait(ip); inode_dio_wait(VFS_I(ip));
} }
rounding = max_t(uint, 1 << mp->m_sb.sb_blocklog, PAGE_CACHE_SIZE); rounding = max_t(uint, 1 << mp->m_sb.sb_blocklog, PAGE_CACHE_SIZE);
@ -2096,8 +2091,8 @@ xfs_free_file_space(
*/ */
if (rt && !xfs_sb_version_hasextflgbit(&mp->m_sb)) { if (rt && !xfs_sb_version_hasextflgbit(&mp->m_sb)) {
nimap = 1; nimap = 1;
error = xfs_bmapi(NULL, ip, startoffset_fsb, error = xfs_bmapi_read(ip, startoffset_fsb, 1,
1, 0, NULL, 0, &imap, &nimap, NULL); &imap, &nimap, 0);
if (error) if (error)
goto out_unlock_iolock; goto out_unlock_iolock;
ASSERT(nimap == 0 || nimap == 1); ASSERT(nimap == 0 || nimap == 1);
@ -2111,8 +2106,8 @@ xfs_free_file_space(
startoffset_fsb += mp->m_sb.sb_rextsize - mod; startoffset_fsb += mp->m_sb.sb_rextsize - mod;
} }
nimap = 1; nimap = 1;
error = xfs_bmapi(NULL, ip, endoffset_fsb - 1, error = xfs_bmapi_read(ip, endoffset_fsb - 1, 1,
1, 0, NULL, 0, &imap, &nimap, NULL); &imap, &nimap, 0);
if (error) if (error)
goto out_unlock_iolock; goto out_unlock_iolock;
ASSERT(nimap == 0 || nimap == 1); ASSERT(nimap == 0 || nimap == 1);
@ -2180,7 +2175,7 @@ xfs_free_file_space(
if (error) if (error)
goto error1; goto error1;
xfs_trans_ijoin(tp, ip); xfs_trans_ijoin(tp, ip, 0);
/* /*
* issue the bunmapi() call to free the blocks * issue the bunmapi() call to free the blocks
@ -2353,8 +2348,7 @@ xfs_change_file_space(
} }
xfs_ilock(ip, XFS_ILOCK_EXCL); xfs_ilock(ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
xfs_trans_ijoin(tp, ip);
if ((attr_flags & XFS_ATTR_DMI) == 0) { if ((attr_flags & XFS_ATTR_DMI) == 0) {
ip->i_d.di_mode &= ~S_ISUID; ip->i_d.di_mode &= ~S_ISUID;
@ -2379,10 +2373,5 @@ xfs_change_file_space(
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
if (attr_flags & XFS_ATTR_SYNC) if (attr_flags & XFS_ATTR_SYNC)
xfs_trans_set_sync(tp); xfs_trans_set_sync(tp);
return xfs_trans_commit(tp, 0);
error = xfs_trans_commit(tp, 0);
xfs_iunlock(ip, XFS_ILOCK_EXCL);
return error;
} }