2018-06-06 10:42:14 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0+
|
2016-08-03 09:12:25 +08:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2016 Oracle. All Rights Reserved.
|
|
|
|
* Author: Darrick J. Wong <darrick.wong@oracle.com>
|
|
|
|
*/
|
|
|
|
#include "xfs.h"
|
|
|
|
#include "xfs_fs.h"
|
|
|
|
#include "xfs_shared.h"
|
|
|
|
#include "xfs_format.h"
|
|
|
|
#include "xfs_log_format.h"
|
|
|
|
#include "xfs_trans_resv.h"
|
|
|
|
#include "xfs_mount.h"
|
|
|
|
#include "xfs_defer.h"
|
|
|
|
#include "xfs_trans.h"
|
2018-08-01 22:20:32 +08:00
|
|
|
#include "xfs_buf_item.h"
|
2018-08-01 22:20:32 +08:00
|
|
|
#include "xfs_inode.h"
|
|
|
|
#include "xfs_inode_item.h"
|
2016-08-03 09:12:25 +08:00
|
|
|
#include "xfs_trace.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Deferred Operations in XFS
|
|
|
|
*
|
|
|
|
* Due to the way locking rules work in XFS, certain transactions (block
|
|
|
|
* mapping and unmapping, typically) have permanent reservations so that
|
|
|
|
* we can roll the transaction to adhere to AG locking order rules and
|
|
|
|
* to unlock buffers between metadata updates. Prior to rmap/reflink,
|
|
|
|
* the mapping code had a mechanism to perform these deferrals for
|
|
|
|
* extents that were going to be freed; this code makes that facility
|
|
|
|
* more generic.
|
|
|
|
*
|
|
|
|
* When adding the reverse mapping and reflink features, it became
|
|
|
|
* necessary to perform complex remapping multi-transactions to comply
|
|
|
|
* with AG locking order rules, and to be able to spread a single
|
|
|
|
* refcount update operation (an operation on an n-block extent can
|
|
|
|
* update as many as n records!) among multiple transactions. XFS can
|
|
|
|
* roll a transaction to facilitate this, but using this facility
|
|
|
|
* requires us to log "intent" items in case log recovery needs to
|
|
|
|
* redo the operation, and to log "done" items to indicate that redo
|
|
|
|
* is not necessary.
|
|
|
|
*
|
|
|
|
* Deferred work is tracked in xfs_defer_pending items. Each pending
|
|
|
|
* item tracks one type of deferred work. Incoming work items (which
|
|
|
|
* have not yet had an intent logged) are attached to a pending item
|
|
|
|
* on the dop_intake list, where they wait for the caller to finish
|
|
|
|
* the deferred operations.
|
|
|
|
*
|
|
|
|
* Finishing a set of deferred operations is an involved process. To
|
|
|
|
* start, we define "rolling a deferred-op transaction" as follows:
|
|
|
|
*
|
|
|
|
* > For each xfs_defer_pending item on the dop_intake list,
|
|
|
|
* - Sort the work items in AG order. XFS locking
|
|
|
|
* order rules require us to lock buffers in AG order.
|
|
|
|
* - Create a log intent item for that type.
|
|
|
|
* - Attach it to the pending item.
|
|
|
|
* - Move the pending item from the dop_intake list to the
|
|
|
|
* dop_pending list.
|
|
|
|
* > Roll the transaction.
|
|
|
|
*
|
|
|
|
* NOTE: To avoid exceeding the transaction reservation, we limit the
|
|
|
|
* number of items that we attach to a given xfs_defer_pending.
|
|
|
|
*
|
|
|
|
* The actual finishing process looks like this:
|
|
|
|
*
|
|
|
|
* > For each xfs_defer_pending in the dop_pending list,
|
|
|
|
* - Roll the deferred-op transaction as above.
|
|
|
|
* - Create a log done item for that type, and attach it to the
|
|
|
|
* log intent item.
|
|
|
|
* - For each work item attached to the log intent item,
|
|
|
|
* * Perform the described action.
|
|
|
|
* * Attach the work item to the log done item.
|
2016-09-19 08:26:25 +08:00
|
|
|
* * If the result of doing the work was -EAGAIN, ->finish work
|
|
|
|
* wants a new transaction. See the "Requesting a Fresh
|
|
|
|
* Transaction while Finishing Deferred Work" section below for
|
|
|
|
* details.
|
2016-08-03 09:12:25 +08:00
|
|
|
*
|
|
|
|
* The key here is that we must log an intent item for all pending
|
|
|
|
* work items every time we roll the transaction, and that we must log
|
|
|
|
* a done item as soon as the work is completed. With this mechanism
|
|
|
|
* we can perform complex remapping operations, chaining intent items
|
|
|
|
* as needed.
|
|
|
|
*
|
2016-09-19 08:26:25 +08:00
|
|
|
* Requesting a Fresh Transaction while Finishing Deferred Work
|
|
|
|
*
|
|
|
|
* If ->finish_item decides that it needs a fresh transaction to
|
|
|
|
* finish the work, it must ask its caller (xfs_defer_finish) for a
|
|
|
|
* continuation. The most likely cause of this circumstance are the
|
|
|
|
* refcount adjust functions deciding that they've logged enough items
|
|
|
|
* to be at risk of exceeding the transaction reservation.
|
|
|
|
*
|
|
|
|
* To get a fresh transaction, we want to log the existing log done
|
|
|
|
* item to prevent the log intent item from replaying, immediately log
|
|
|
|
* a new log intent item with the unfinished work items, roll the
|
|
|
|
* transaction, and re-call ->finish_item wherever it left off. The
|
|
|
|
* log done item and the new log intent item must be in the same
|
|
|
|
* transaction or atomicity cannot be guaranteed; defer_finish ensures
|
|
|
|
* that this happens.
|
|
|
|
*
|
|
|
|
* This requires some coordination between ->finish_item and
|
|
|
|
* defer_finish. Upon deciding to request a new transaction,
|
|
|
|
* ->finish_item should update the current work item to reflect the
|
|
|
|
* unfinished work. Next, it should reset the log done item's list
|
|
|
|
* count to the number of items finished, and return -EAGAIN.
|
|
|
|
* defer_finish sees the -EAGAIN, logs the new log intent item
|
|
|
|
* with the remaining work items, and leaves the xfs_defer_pending
|
|
|
|
* item at the head of the dop_work queue. Then it rolls the
|
|
|
|
* transaction and picks up processing where it left off. It is
|
|
|
|
* required that ->finish_item must be careful to leave enough
|
|
|
|
* transaction reservation to fit the new log intent item.
|
|
|
|
*
|
2016-08-03 09:12:25 +08:00
|
|
|
* This is an example of remapping the extent (E, E+B) into file X at
|
|
|
|
* offset A and dealing with the extent (C, C+B) already being mapped
|
|
|
|
* there:
|
|
|
|
* +-------------------------------------------------+
|
|
|
|
* | Unmap file X startblock C offset A length B | t0
|
|
|
|
* | Intent to reduce refcount for extent (C, B) |
|
|
|
|
* | Intent to remove rmap (X, C, A, B) |
|
|
|
|
* | Intent to free extent (D, 1) (bmbt block) |
|
|
|
|
* | Intent to map (X, A, B) at startblock E |
|
|
|
|
* +-------------------------------------------------+
|
|
|
|
* | Map file X startblock E offset A length B | t1
|
|
|
|
* | Done mapping (X, E, A, B) |
|
|
|
|
* | Intent to increase refcount for extent (E, B) |
|
|
|
|
* | Intent to add rmap (X, E, A, B) |
|
|
|
|
* +-------------------------------------------------+
|
|
|
|
* | Reduce refcount for extent (C, B) | t2
|
2016-09-19 08:26:25 +08:00
|
|
|
* | Done reducing refcount for extent (C, 9) |
|
|
|
|
* | Intent to reduce refcount for extent (C+9, B-9) |
|
|
|
|
* | (ran out of space after 9 refcount updates) |
|
|
|
|
* +-------------------------------------------------+
|
|
|
|
* | Reduce refcount for extent (C+9, B+9) | t3
|
|
|
|
* | Done reducing refcount for extent (C+9, B-9) |
|
2016-08-03 09:12:25 +08:00
|
|
|
* | Increase refcount for extent (E, B) |
|
|
|
|
* | Done increasing refcount for extent (E, B) |
|
|
|
|
* | Intent to free extent (C, B) |
|
|
|
|
* | Intent to free extent (F, 1) (refcountbt block) |
|
|
|
|
* | Intent to remove rmap (F, 1, REFC) |
|
|
|
|
* +-------------------------------------------------+
|
2016-09-19 08:26:25 +08:00
|
|
|
* | Remove rmap (X, C, A, B) | t4
|
2016-08-03 09:12:25 +08:00
|
|
|
* | Done removing rmap (X, C, A, B) |
|
|
|
|
* | Add rmap (X, E, A, B) |
|
|
|
|
* | Done adding rmap (X, E, A, B) |
|
|
|
|
* | Remove rmap (F, 1, REFC) |
|
|
|
|
* | Done removing rmap (F, 1, REFC) |
|
|
|
|
* +-------------------------------------------------+
|
2016-09-19 08:26:25 +08:00
|
|
|
* | Free extent (C, B) | t5
|
2016-08-03 09:12:25 +08:00
|
|
|
* | Done freeing extent (C, B) |
|
|
|
|
* | Free extent (D, 1) |
|
|
|
|
* | Done freeing extent (D, 1) |
|
|
|
|
* | Free extent (F, 1) |
|
|
|
|
* | Done freeing extent (F, 1) |
|
|
|
|
* +-------------------------------------------------+
|
|
|
|
*
|
|
|
|
* If we should crash before t2 commits, log recovery replays
|
|
|
|
* the following intent items:
|
|
|
|
*
|
|
|
|
* - Intent to reduce refcount for extent (C, B)
|
|
|
|
* - Intent to remove rmap (X, C, A, B)
|
|
|
|
* - Intent to free extent (D, 1) (bmbt block)
|
|
|
|
* - Intent to increase refcount for extent (E, B)
|
|
|
|
* - Intent to add rmap (X, E, A, B)
|
|
|
|
*
|
|
|
|
* In the process of recovering, it should also generate and take care
|
|
|
|
* of these intent items:
|
|
|
|
*
|
|
|
|
* - Intent to free extent (C, B)
|
|
|
|
* - Intent to free extent (F, 1) (refcountbt block)
|
|
|
|
* - Intent to remove rmap (F, 1, REFC)
|
2016-09-19 08:26:25 +08:00
|
|
|
*
|
|
|
|
* Note that the continuation requested between t2 and t3 is likely to
|
|
|
|
* reoccur.
|
2016-08-03 09:12:25 +08:00
|
|
|
*/
|
|
|
|
|
2018-12-13 00:46:22 +08:00
|
|
|
static const struct xfs_defer_op_type *defer_op_types[] = {
|
|
|
|
[XFS_DEFER_OPS_TYPE_BMAP] = &xfs_bmap_update_defer_type,
|
|
|
|
[XFS_DEFER_OPS_TYPE_REFCOUNT] = &xfs_refcount_update_defer_type,
|
|
|
|
[XFS_DEFER_OPS_TYPE_RMAP] = &xfs_rmap_update_defer_type,
|
|
|
|
[XFS_DEFER_OPS_TYPE_FREE] = &xfs_extent_free_defer_type,
|
|
|
|
[XFS_DEFER_OPS_TYPE_AGFL_FREE] = &xfs_agfl_free_defer_type,
|
|
|
|
};
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2020-05-01 03:52:20 +08:00
|
|
|
static void
|
|
|
|
xfs_defer_create_intent(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_defer_pending *dfp,
|
|
|
|
bool sort)
|
|
|
|
{
|
|
|
|
const struct xfs_defer_op_type *ops = defer_op_types[dfp->dfp_type];
|
|
|
|
|
2020-09-22 00:15:09 +08:00
|
|
|
if (!dfp->dfp_intent)
|
|
|
|
dfp->dfp_intent = ops->create_intent(tp, &dfp->dfp_work,
|
|
|
|
dfp->dfp_count, sort);
|
2020-05-01 03:52:20 +08:00
|
|
|
}
|
|
|
|
|
2016-08-03 09:12:25 +08:00
|
|
|
/*
|
|
|
|
* For each pending item in the intake list, log its intent item and the
|
|
|
|
* associated extents, then add the entire intake list to the end of
|
|
|
|
* the pending list.
|
|
|
|
*/
|
|
|
|
STATIC void
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_defer_create_intents(
|
2018-08-01 22:20:33 +08:00
|
|
|
struct xfs_trans *tp)
|
2016-08-03 09:12:25 +08:00
|
|
|
{
|
|
|
|
struct xfs_defer_pending *dfp;
|
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
list_for_each_entry(dfp, &tp->t_dfops, dfp_list) {
|
2018-08-01 22:20:34 +08:00
|
|
|
trace_xfs_defer_create_intent(tp->t_mountp, dfp);
|
2020-05-01 03:52:20 +08:00
|
|
|
xfs_defer_create_intent(tp, dfp, true);
|
2016-08-03 09:12:25 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Abort all the intents that were committed. */
|
|
|
|
STATIC void
|
|
|
|
xfs_defer_trans_abort(
|
|
|
|
struct xfs_trans *tp,
|
2018-08-01 22:20:34 +08:00
|
|
|
struct list_head *dop_pending)
|
2016-08-03 09:12:25 +08:00
|
|
|
{
|
|
|
|
struct xfs_defer_pending *dfp;
|
2018-12-13 00:46:22 +08:00
|
|
|
const struct xfs_defer_op_type *ops;
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
trace_xfs_defer_trans_abort(tp, _RET_IP_);
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2016-10-24 11:21:18 +08:00
|
|
|
/* Abort intent items that don't have a done item. */
|
2018-08-01 22:20:34 +08:00
|
|
|
list_for_each_entry(dfp, dop_pending, dfp_list) {
|
2018-12-13 00:46:22 +08:00
|
|
|
ops = defer_op_types[dfp->dfp_type];
|
2016-08-03 09:13:02 +08:00
|
|
|
trace_xfs_defer_pending_abort(tp->t_mountp, dfp);
|
2016-10-24 11:21:18 +08:00
|
|
|
if (dfp->dfp_intent && !dfp->dfp_done) {
|
2018-12-13 00:46:22 +08:00
|
|
|
ops->abort_intent(dfp->dfp_intent);
|
2016-10-24 11:21:18 +08:00
|
|
|
dfp->dfp_intent = NULL;
|
|
|
|
}
|
2016-08-03 09:12:25 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Roll a transaction so we can do some deferred op processing. */
|
|
|
|
STATIC int
|
|
|
|
xfs_defer_trans_roll(
|
2018-08-01 22:20:34 +08:00
|
|
|
struct xfs_trans **tpp)
|
2016-08-03 09:12:25 +08:00
|
|
|
{
|
2018-08-01 22:20:34 +08:00
|
|
|
struct xfs_trans *tp = *tpp;
|
2018-08-01 22:20:32 +08:00
|
|
|
struct xfs_buf_log_item *bli;
|
2018-08-01 22:20:32 +08:00
|
|
|
struct xfs_inode_log_item *ili;
|
2018-08-01 22:20:32 +08:00
|
|
|
struct xfs_log_item *lip;
|
|
|
|
struct xfs_buf *bplist[XFS_DEFER_OPS_NR_BUFS];
|
2018-08-01 22:20:32 +08:00
|
|
|
struct xfs_inode *iplist[XFS_DEFER_OPS_NR_INODES];
|
xfs: use ordered buffers to initialize dquot buffers during quotacheck
While QAing the new xfs_repair quotacheck code, I uncovered a quota
corruption bug resulting from a bad interaction between dquot buffer
initialization and quotacheck. The bug can be reproduced with the
following sequence:
# mkfs.xfs -f /dev/sdf
# mount /dev/sdf /opt -o usrquota
# su nobody -s /bin/bash -c 'touch /opt/barf'
# sync
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 3 0 0 00 [------]
nobody 1 0 0 00 [------]
# xfs_io -x -c 'shutdown' /opt
# umount /opt
# mount /dev/sdf /opt -o usrquota
# touch /opt/man2
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 1 0 0 00 [------]
nobody 1 0 0 00 [------]
# umount /opt
Notice how the initial quotacheck set the root dquot icount to 3
(rootino, rbmino, rsumino), but after shutdown -> remount -> recovery,
xfs_quota reports that the root dquot has only 1 icount. We haven't
deleted anything from the filesystem, which means that quota is now
under-counting. This behavior is not limited to icount or the root
dquot, but this is the shortest reproducer.
I traced the cause of this discrepancy to the way that we handle ondisk
dquot updates during quotacheck vs. regular fs activity. Normally, when
we allocate a disk block for a dquot, we log the buffer as a regular
(dquot) buffer. Subsequent updates to the dquots backed by that block
are done via separate dquot log item updates, which means that they
depend on the logged buffer update being written to disk before the
dquot items. Because individual dquots have their own LSN fields, that
initial dquot buffer must always be recovered.
However, the story changes for quotacheck, which can cause dquot block
allocations but persists the final dquot counter values via a delwri
list. Because recovery doesn't gate dquot buffer replay on an LSN, this
means that the initial dquot buffer can be replayed over the (newer)
contents that were delwritten at the end of quotacheck. In effect, this
re-initializes the dquot counters after they've been updated. If the
log does not contain any other dquot items to recover, the obsolete
dquot contents will not be corrected by log recovery.
Because quotacheck uses a transaction to log the setting of the CHKD
flags in the superblock, we skip quotacheck during the second mount
call, which allows the incorrect icount to remain.
Fix this by changing the ondisk dquot initialization function to use
ordered buffers to write out fresh dquot blocks if it detects that we're
running quotacheck. If the system goes down before quotacheck can
complete, the CHKD flags will not be set in the superblock and the next
mount will run quotacheck again, which can fix uninitialized dquot
buffers. This requires amending the defer code to maintaine ordered
buffer state across defer rolls for the sake of the dquot allocation
code.
For regular operations we preserve the current behavior since the dquot
items require properly initialized ondisk dquot records.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-05-14 06:33:27 +08:00
|
|
|
unsigned int ordered = 0; /* bitmap */
|
2018-08-01 22:20:32 +08:00
|
|
|
int bpcount = 0, ipcount = 0;
|
2016-08-03 09:12:25 +08:00
|
|
|
int i;
|
|
|
|
int error;
|
|
|
|
|
xfs: use ordered buffers to initialize dquot buffers during quotacheck
While QAing the new xfs_repair quotacheck code, I uncovered a quota
corruption bug resulting from a bad interaction between dquot buffer
initialization and quotacheck. The bug can be reproduced with the
following sequence:
# mkfs.xfs -f /dev/sdf
# mount /dev/sdf /opt -o usrquota
# su nobody -s /bin/bash -c 'touch /opt/barf'
# sync
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 3 0 0 00 [------]
nobody 1 0 0 00 [------]
# xfs_io -x -c 'shutdown' /opt
# umount /opt
# mount /dev/sdf /opt -o usrquota
# touch /opt/man2
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 1 0 0 00 [------]
nobody 1 0 0 00 [------]
# umount /opt
Notice how the initial quotacheck set the root dquot icount to 3
(rootino, rbmino, rsumino), but after shutdown -> remount -> recovery,
xfs_quota reports that the root dquot has only 1 icount. We haven't
deleted anything from the filesystem, which means that quota is now
under-counting. This behavior is not limited to icount or the root
dquot, but this is the shortest reproducer.
I traced the cause of this discrepancy to the way that we handle ondisk
dquot updates during quotacheck vs. regular fs activity. Normally, when
we allocate a disk block for a dquot, we log the buffer as a regular
(dquot) buffer. Subsequent updates to the dquots backed by that block
are done via separate dquot log item updates, which means that they
depend on the logged buffer update being written to disk before the
dquot items. Because individual dquots have their own LSN fields, that
initial dquot buffer must always be recovered.
However, the story changes for quotacheck, which can cause dquot block
allocations but persists the final dquot counter values via a delwri
list. Because recovery doesn't gate dquot buffer replay on an LSN, this
means that the initial dquot buffer can be replayed over the (newer)
contents that were delwritten at the end of quotacheck. In effect, this
re-initializes the dquot counters after they've been updated. If the
log does not contain any other dquot items to recover, the obsolete
dquot contents will not be corrected by log recovery.
Because quotacheck uses a transaction to log the setting of the CHKD
flags in the superblock, we skip quotacheck during the second mount
call, which allows the incorrect icount to remain.
Fix this by changing the ondisk dquot initialization function to use
ordered buffers to write out fresh dquot blocks if it detects that we're
running quotacheck. If the system goes down before quotacheck can
complete, the CHKD flags will not be set in the superblock and the next
mount will run quotacheck again, which can fix uninitialized dquot
buffers. This requires amending the defer code to maintaine ordered
buffer state across defer rolls for the sake of the dquot allocation
code.
For regular operations we preserve the current behavior since the dquot
items require properly initialized ondisk dquot records.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-05-14 06:33:27 +08:00
|
|
|
BUILD_BUG_ON(NBBY * sizeof(ordered) < XFS_DEFER_OPS_NR_BUFS);
|
|
|
|
|
2018-08-01 22:20:34 +08:00
|
|
|
list_for_each_entry(lip, &tp->t_items, li_trans) {
|
2018-08-01 22:20:32 +08:00
|
|
|
switch (lip->li_type) {
|
|
|
|
case XFS_LI_BUF:
|
|
|
|
bli = container_of(lip, struct xfs_buf_log_item,
|
|
|
|
bli_item);
|
|
|
|
if (bli->bli_flags & XFS_BLI_HOLD) {
|
|
|
|
if (bpcount >= XFS_DEFER_OPS_NR_BUFS) {
|
|
|
|
ASSERT(0);
|
|
|
|
return -EFSCORRUPTED;
|
|
|
|
}
|
xfs: use ordered buffers to initialize dquot buffers during quotacheck
While QAing the new xfs_repair quotacheck code, I uncovered a quota
corruption bug resulting from a bad interaction between dquot buffer
initialization and quotacheck. The bug can be reproduced with the
following sequence:
# mkfs.xfs -f /dev/sdf
# mount /dev/sdf /opt -o usrquota
# su nobody -s /bin/bash -c 'touch /opt/barf'
# sync
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 3 0 0 00 [------]
nobody 1 0 0 00 [------]
# xfs_io -x -c 'shutdown' /opt
# umount /opt
# mount /dev/sdf /opt -o usrquota
# touch /opt/man2
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 1 0 0 00 [------]
nobody 1 0 0 00 [------]
# umount /opt
Notice how the initial quotacheck set the root dquot icount to 3
(rootino, rbmino, rsumino), but after shutdown -> remount -> recovery,
xfs_quota reports that the root dquot has only 1 icount. We haven't
deleted anything from the filesystem, which means that quota is now
under-counting. This behavior is not limited to icount or the root
dquot, but this is the shortest reproducer.
I traced the cause of this discrepancy to the way that we handle ondisk
dquot updates during quotacheck vs. regular fs activity. Normally, when
we allocate a disk block for a dquot, we log the buffer as a regular
(dquot) buffer. Subsequent updates to the dquots backed by that block
are done via separate dquot log item updates, which means that they
depend on the logged buffer update being written to disk before the
dquot items. Because individual dquots have their own LSN fields, that
initial dquot buffer must always be recovered.
However, the story changes for quotacheck, which can cause dquot block
allocations but persists the final dquot counter values via a delwri
list. Because recovery doesn't gate dquot buffer replay on an LSN, this
means that the initial dquot buffer can be replayed over the (newer)
contents that were delwritten at the end of quotacheck. In effect, this
re-initializes the dquot counters after they've been updated. If the
log does not contain any other dquot items to recover, the obsolete
dquot contents will not be corrected by log recovery.
Because quotacheck uses a transaction to log the setting of the CHKD
flags in the superblock, we skip quotacheck during the second mount
call, which allows the incorrect icount to remain.
Fix this by changing the ondisk dquot initialization function to use
ordered buffers to write out fresh dquot blocks if it detects that we're
running quotacheck. If the system goes down before quotacheck can
complete, the CHKD flags will not be set in the superblock and the next
mount will run quotacheck again, which can fix uninitialized dquot
buffers. This requires amending the defer code to maintaine ordered
buffer state across defer rolls for the sake of the dquot allocation
code.
For regular operations we preserve the current behavior since the dquot
items require properly initialized ondisk dquot records.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-05-14 06:33:27 +08:00
|
|
|
if (bli->bli_flags & XFS_BLI_ORDERED)
|
|
|
|
ordered |= (1U << bpcount);
|
|
|
|
else
|
|
|
|
xfs_trans_dirty_buf(tp, bli->bli_buf);
|
2018-08-01 22:20:32 +08:00
|
|
|
bplist[bpcount++] = bli->bli_buf;
|
|
|
|
}
|
|
|
|
break;
|
2018-08-01 22:20:32 +08:00
|
|
|
case XFS_LI_INODE:
|
|
|
|
ili = container_of(lip, struct xfs_inode_log_item,
|
|
|
|
ili_item);
|
|
|
|
if (ili->ili_lock_flags == 0) {
|
|
|
|
if (ipcount >= XFS_DEFER_OPS_NR_INODES) {
|
|
|
|
ASSERT(0);
|
|
|
|
return -EFSCORRUPTED;
|
|
|
|
}
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_trans_log_inode(tp, ili->ili_inode,
|
2018-08-01 22:20:32 +08:00
|
|
|
XFS_ILOG_CORE);
|
|
|
|
iplist[ipcount++] = ili->ili_inode;
|
|
|
|
}
|
|
|
|
break;
|
2018-08-01 22:20:32 +08:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2017-12-08 11:07:02 +08:00
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
trace_xfs_defer_trans_roll(tp, _RET_IP_);
|
2016-08-03 09:13:02 +08:00
|
|
|
|
2019-04-25 00:27:41 +08:00
|
|
|
/*
|
|
|
|
* Roll the transaction. Rolling always given a new transaction (even
|
|
|
|
* if committing the old one fails!) to hand back to the caller, so we
|
|
|
|
* join the held resources to the new transaction so that we always
|
|
|
|
* return with the held resources joined to @tpp, no matter what
|
|
|
|
* happened.
|
|
|
|
*/
|
2018-08-01 22:20:34 +08:00
|
|
|
error = xfs_trans_roll(tpp);
|
|
|
|
tp = *tpp;
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2017-08-29 01:21:03 +08:00
|
|
|
/* Rejoin the joined inodes. */
|
2018-08-01 22:20:32 +08:00
|
|
|
for (i = 0; i < ipcount; i++)
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_trans_ijoin(tp, iplist[i], 0);
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2017-12-08 11:07:02 +08:00
|
|
|
/* Rejoin the buffers and dirty them so the log moves forward. */
|
2018-08-01 22:20:32 +08:00
|
|
|
for (i = 0; i < bpcount; i++) {
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_trans_bjoin(tp, bplist[i]);
|
xfs: use ordered buffers to initialize dquot buffers during quotacheck
While QAing the new xfs_repair quotacheck code, I uncovered a quota
corruption bug resulting from a bad interaction between dquot buffer
initialization and quotacheck. The bug can be reproduced with the
following sequence:
# mkfs.xfs -f /dev/sdf
# mount /dev/sdf /opt -o usrquota
# su nobody -s /bin/bash -c 'touch /opt/barf'
# sync
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 3 0 0 00 [------]
nobody 1 0 0 00 [------]
# xfs_io -x -c 'shutdown' /opt
# umount /opt
# mount /dev/sdf /opt -o usrquota
# touch /opt/man2
# xfs_quota -x -c 'report -ahi' /opt
User quota on /opt (/dev/sdf)
Inodes
User ID Used Soft Hard Warn/Grace
---------- ---------------------------------
root 1 0 0 00 [------]
nobody 1 0 0 00 [------]
# umount /opt
Notice how the initial quotacheck set the root dquot icount to 3
(rootino, rbmino, rsumino), but after shutdown -> remount -> recovery,
xfs_quota reports that the root dquot has only 1 icount. We haven't
deleted anything from the filesystem, which means that quota is now
under-counting. This behavior is not limited to icount or the root
dquot, but this is the shortest reproducer.
I traced the cause of this discrepancy to the way that we handle ondisk
dquot updates during quotacheck vs. regular fs activity. Normally, when
we allocate a disk block for a dquot, we log the buffer as a regular
(dquot) buffer. Subsequent updates to the dquots backed by that block
are done via separate dquot log item updates, which means that they
depend on the logged buffer update being written to disk before the
dquot items. Because individual dquots have their own LSN fields, that
initial dquot buffer must always be recovered.
However, the story changes for quotacheck, which can cause dquot block
allocations but persists the final dquot counter values via a delwri
list. Because recovery doesn't gate dquot buffer replay on an LSN, this
means that the initial dquot buffer can be replayed over the (newer)
contents that were delwritten at the end of quotacheck. In effect, this
re-initializes the dquot counters after they've been updated. If the
log does not contain any other dquot items to recover, the obsolete
dquot contents will not be corrected by log recovery.
Because quotacheck uses a transaction to log the setting of the CHKD
flags in the superblock, we skip quotacheck during the second mount
call, which allows the incorrect icount to remain.
Fix this by changing the ondisk dquot initialization function to use
ordered buffers to write out fresh dquot blocks if it detects that we're
running quotacheck. If the system goes down before quotacheck can
complete, the CHKD flags will not be set in the superblock and the next
mount will run quotacheck again, which can fix uninitialized dquot
buffers. This requires amending the defer code to maintaine ordered
buffer state across defer rolls for the sake of the dquot allocation
code.
For regular operations we preserve the current behavior since the dquot
items require properly initialized ondisk dquot records.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-05-14 06:33:27 +08:00
|
|
|
if (ordered & (1U << i))
|
|
|
|
xfs_trans_ordered_buf(tp, bplist[i]);
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_trans_bhold(tp, bplist[i]);
|
2017-12-08 11:07:02 +08:00
|
|
|
}
|
|
|
|
|
2019-04-25 00:27:41 +08:00
|
|
|
if (error)
|
|
|
|
trace_xfs_defer_trans_roll_error(tp, error);
|
2016-08-03 09:12:25 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2018-08-01 22:20:34 +08:00
|
|
|
/*
|
|
|
|
* Free up any items left in the list.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
xfs_defer_cancel_list(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct list_head *dop_list)
|
|
|
|
{
|
|
|
|
struct xfs_defer_pending *dfp;
|
|
|
|
struct xfs_defer_pending *pli;
|
|
|
|
struct list_head *pwi;
|
|
|
|
struct list_head *n;
|
2018-12-13 00:46:22 +08:00
|
|
|
const struct xfs_defer_op_type *ops;
|
2018-08-01 22:20:34 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Free the pending items. Caller should already have arranged
|
|
|
|
* for the intent items to be released.
|
|
|
|
*/
|
|
|
|
list_for_each_entry_safe(dfp, pli, dop_list, dfp_list) {
|
2018-12-13 00:46:22 +08:00
|
|
|
ops = defer_op_types[dfp->dfp_type];
|
2018-08-01 22:20:34 +08:00
|
|
|
trace_xfs_defer_cancel_list(mp, dfp);
|
|
|
|
list_del(&dfp->dfp_list);
|
|
|
|
list_for_each_safe(pwi, n, &dfp->dfp_work) {
|
|
|
|
list_del(pwi);
|
|
|
|
dfp->dfp_count--;
|
2018-12-13 00:46:22 +08:00
|
|
|
ops->cancel_item(pwi);
|
2018-08-01 22:20:34 +08:00
|
|
|
}
|
|
|
|
ASSERT(dfp->dfp_count == 0);
|
|
|
|
kmem_free(dfp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-05-01 03:52:21 +08:00
|
|
|
/*
|
|
|
|
* Log an intent-done item for the first pending intent, and finish the work
|
|
|
|
* items.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
xfs_defer_finish_one(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_defer_pending *dfp)
|
|
|
|
{
|
|
|
|
const struct xfs_defer_op_type *ops = defer_op_types[dfp->dfp_type];
|
2020-05-01 03:52:22 +08:00
|
|
|
struct xfs_btree_cur *state = NULL;
|
2020-05-01 03:52:21 +08:00
|
|
|
struct list_head *li, *n;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
trace_xfs_defer_pending_finish(tp->t_mountp, dfp);
|
|
|
|
|
|
|
|
dfp->dfp_done = ops->create_done(tp, dfp->dfp_intent, dfp->dfp_count);
|
|
|
|
list_for_each_safe(li, n, &dfp->dfp_work) {
|
|
|
|
list_del(li);
|
|
|
|
dfp->dfp_count--;
|
2020-05-01 03:52:22 +08:00
|
|
|
error = ops->finish_item(tp, dfp->dfp_done, li, &state);
|
2020-05-01 03:52:21 +08:00
|
|
|
if (error == -EAGAIN) {
|
|
|
|
/*
|
|
|
|
* Caller wants a fresh transaction; put the work item
|
|
|
|
* back on the list and log a new log intent item to
|
|
|
|
* replace the old one. See "Requesting a Fresh
|
|
|
|
* Transaction while Finishing Deferred Work" above.
|
|
|
|
*/
|
|
|
|
list_add(li, &dfp->dfp_work);
|
|
|
|
dfp->dfp_count++;
|
|
|
|
dfp->dfp_done = NULL;
|
2020-09-22 00:15:09 +08:00
|
|
|
dfp->dfp_intent = NULL;
|
2020-05-01 03:52:21 +08:00
|
|
|
xfs_defer_create_intent(tp, dfp, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (error)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Done with the dfp, free it. */
|
|
|
|
list_del(&dfp->dfp_list);
|
|
|
|
kmem_free(dfp);
|
|
|
|
out:
|
|
|
|
if (ops->finish_cleanup)
|
|
|
|
ops->finish_cleanup(tp, state, error);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2016-08-03 09:12:25 +08:00
|
|
|
/*
|
|
|
|
* Finish all the pending work. This involves logging intent items for
|
|
|
|
* any work items that wandered in since the last transaction roll (if
|
|
|
|
* one has even happened), rolling the transaction, and finishing the
|
|
|
|
* work items in the first item on the logged-and-pending list.
|
|
|
|
*
|
|
|
|
* If an inode is provided, relog it to the new transaction.
|
|
|
|
*/
|
|
|
|
int
|
2018-07-25 04:43:15 +08:00
|
|
|
xfs_defer_finish_noroll(
|
2018-07-25 04:43:15 +08:00
|
|
|
struct xfs_trans **tp)
|
2016-08-03 09:12:25 +08:00
|
|
|
{
|
|
|
|
struct xfs_defer_pending *dfp;
|
|
|
|
int error = 0;
|
2018-08-01 22:20:34 +08:00
|
|
|
LIST_HEAD(dop_pending);
|
2016-08-03 09:12:25 +08:00
|
|
|
|
|
|
|
ASSERT((*tp)->t_flags & XFS_TRANS_PERM_LOG_RES);
|
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
trace_xfs_defer_finish(*tp, _RET_IP_);
|
2016-08-03 09:13:02 +08:00
|
|
|
|
2016-08-03 09:12:25 +08:00
|
|
|
/* Until we run out of pending work to finish... */
|
2018-08-01 22:20:35 +08:00
|
|
|
while (!list_empty(&dop_pending) || !list_empty(&(*tp)->t_dfops)) {
|
2018-08-01 22:20:34 +08:00
|
|
|
xfs_defer_create_intents(*tp);
|
2018-08-01 22:20:35 +08:00
|
|
|
list_splice_tail_init(&(*tp)->t_dfops, &dop_pending);
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2018-07-25 04:43:09 +08:00
|
|
|
error = xfs_defer_trans_roll(tp);
|
2016-08-03 09:12:25 +08:00
|
|
|
if (error)
|
2020-05-01 03:52:21 +08:00
|
|
|
goto out_shutdown;
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2018-08-01 22:20:34 +08:00
|
|
|
dfp = list_first_entry(&dop_pending, struct xfs_defer_pending,
|
|
|
|
dfp_list);
|
2020-05-01 03:52:21 +08:00
|
|
|
error = xfs_defer_finish_one(*tp, dfp);
|
|
|
|
if (error && error != -EAGAIN)
|
|
|
|
goto out_shutdown;
|
2018-08-01 22:20:33 +08:00
|
|
|
}
|
2018-07-25 04:43:10 +08:00
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
trace_xfs_defer_finish_done(*tp, _RET_IP_);
|
2018-08-01 22:20:33 +08:00
|
|
|
return 0;
|
2020-05-01 03:52:21 +08:00
|
|
|
|
|
|
|
out_shutdown:
|
|
|
|
xfs_defer_trans_abort(*tp, &dop_pending);
|
|
|
|
xfs_force_shutdown((*tp)->t_mountp, SHUTDOWN_CORRUPT_INCORE);
|
|
|
|
trace_xfs_defer_finish_error(*tp, error);
|
|
|
|
xfs_defer_cancel_list((*tp)->t_mountp, &dop_pending);
|
|
|
|
xfs_defer_cancel(*tp);
|
|
|
|
return error;
|
2016-08-03 09:12:25 +08:00
|
|
|
}
|
|
|
|
|
2018-07-25 04:43:15 +08:00
|
|
|
int
|
|
|
|
xfs_defer_finish(
|
|
|
|
struct xfs_trans **tp)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Finish and roll the transaction once more to avoid returning to the
|
|
|
|
* caller with a dirty transaction.
|
|
|
|
*/
|
|
|
|
error = xfs_defer_finish_noroll(tp);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
if ((*tp)->t_flags & XFS_TRANS_DIRTY) {
|
|
|
|
error = xfs_defer_trans_roll(tp);
|
2018-08-01 22:20:34 +08:00
|
|
|
if (error) {
|
|
|
|
xfs_force_shutdown((*tp)->t_mountp,
|
|
|
|
SHUTDOWN_CORRUPT_INCORE);
|
2018-07-25 04:43:15 +08:00
|
|
|
return error;
|
2018-08-01 22:20:34 +08:00
|
|
|
}
|
2018-07-25 04:43:15 +08:00
|
|
|
}
|
2020-09-26 08:39:27 +08:00
|
|
|
|
|
|
|
/* Reset LOWMODE now that we've finished all the dfops. */
|
|
|
|
ASSERT(list_empty(&(*tp)->t_dfops));
|
|
|
|
(*tp)->t_flags &= ~XFS_TRANS_LOWMODE;
|
2018-07-25 04:43:15 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-08-03 09:12:25 +08:00
|
|
|
void
|
2018-08-01 22:20:30 +08:00
|
|
|
xfs_defer_cancel(
|
2018-08-01 22:20:34 +08:00
|
|
|
struct xfs_trans *tp)
|
2016-08-03 09:12:25 +08:00
|
|
|
{
|
2018-08-01 22:20:34 +08:00
|
|
|
struct xfs_mount *mp = tp->t_mountp;
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2018-08-01 22:20:35 +08:00
|
|
|
trace_xfs_defer_cancel(tp, _RET_IP_);
|
|
|
|
xfs_defer_cancel_list(mp, &tp->t_dfops);
|
2016-08-03 09:12:25 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Add an item for later deferred processing. */
|
|
|
|
void
|
|
|
|
xfs_defer_add(
|
2018-08-01 22:20:34 +08:00
|
|
|
struct xfs_trans *tp,
|
2016-08-03 09:12:25 +08:00
|
|
|
enum xfs_defer_ops_type type,
|
|
|
|
struct list_head *li)
|
|
|
|
{
|
|
|
|
struct xfs_defer_pending *dfp = NULL;
|
2018-12-13 00:46:22 +08:00
|
|
|
const struct xfs_defer_op_type *ops;
|
2016-08-03 09:12:25 +08:00
|
|
|
|
2018-08-01 22:20:34 +08:00
|
|
|
ASSERT(tp->t_flags & XFS_TRANS_PERM_LOG_RES);
|
2018-12-13 00:46:22 +08:00
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(defer_op_types) != XFS_DEFER_OPS_TYPE_MAX);
|
2018-08-01 22:20:34 +08:00
|
|
|
|
2016-08-03 09:12:25 +08:00
|
|
|
/*
|
|
|
|
* Add the item to a pending item at the end of the intake list.
|
|
|
|
* If the last pending item has the same type, reuse it. Else,
|
|
|
|
* create a new pending item at the end of the intake list.
|
|
|
|
*/
|
2018-08-01 22:20:35 +08:00
|
|
|
if (!list_empty(&tp->t_dfops)) {
|
|
|
|
dfp = list_last_entry(&tp->t_dfops,
|
2016-08-03 09:12:25 +08:00
|
|
|
struct xfs_defer_pending, dfp_list);
|
2018-12-13 00:46:22 +08:00
|
|
|
ops = defer_op_types[dfp->dfp_type];
|
|
|
|
if (dfp->dfp_type != type ||
|
|
|
|
(ops->max_items && dfp->dfp_count >= ops->max_items))
|
2016-08-03 09:12:25 +08:00
|
|
|
dfp = NULL;
|
|
|
|
}
|
|
|
|
if (!dfp) {
|
|
|
|
dfp = kmem_alloc(sizeof(struct xfs_defer_pending),
|
2019-08-27 03:06:22 +08:00
|
|
|
KM_NOFS);
|
2018-12-13 00:46:22 +08:00
|
|
|
dfp->dfp_type = type;
|
2016-08-03 09:12:25 +08:00
|
|
|
dfp->dfp_intent = NULL;
|
2016-08-30 11:51:39 +08:00
|
|
|
dfp->dfp_done = NULL;
|
2016-08-03 09:12:25 +08:00
|
|
|
dfp->dfp_count = 0;
|
|
|
|
INIT_LIST_HEAD(&dfp->dfp_work);
|
2018-08-01 22:20:35 +08:00
|
|
|
list_add_tail(&dfp->dfp_list, &tp->t_dfops);
|
2016-08-03 09:12:25 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
list_add_tail(li, &dfp->dfp_work);
|
|
|
|
dfp->dfp_count++;
|
|
|
|
}
|
|
|
|
|
2018-07-25 04:43:11 +08:00
|
|
|
/*
|
2018-08-01 22:20:35 +08:00
|
|
|
* Move deferred ops from one transaction to another and reset the source to
|
|
|
|
* initial state. This is primarily used to carry state forward across
|
|
|
|
* transaction rolls with pending dfops.
|
2018-07-25 04:43:11 +08:00
|
|
|
*/
|
|
|
|
void
|
|
|
|
xfs_defer_move(
|
2018-08-01 22:20:30 +08:00
|
|
|
struct xfs_trans *dtp,
|
|
|
|
struct xfs_trans *stp)
|
2018-07-25 04:43:11 +08:00
|
|
|
{
|
2018-08-01 22:20:35 +08:00
|
|
|
list_splice_init(&stp->t_dfops, &dtp->t_dfops);
|
2018-07-25 04:43:11 +08:00
|
|
|
|
2018-08-01 22:20:31 +08:00
|
|
|
/*
|
|
|
|
* Low free space mode was historically controlled by a dfops field.
|
|
|
|
* This meant that low mode state potentially carried across multiple
|
|
|
|
* transaction rolls. Transfer low mode on a dfops move to preserve
|
|
|
|
* that behavior.
|
|
|
|
*/
|
|
|
|
dtp->t_flags |= (stp->t_flags & XFS_TRANS_LOWMODE);
|
2020-09-26 08:39:27 +08:00
|
|
|
stp->t_flags &= ~XFS_TRANS_LOWMODE;
|
2018-07-25 04:43:11 +08:00
|
|
|
}
|
2020-09-22 00:15:09 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Prepare a chain of fresh deferred ops work items to be completed later. Log
|
|
|
|
* recovery requires the ability to put off until later the actual finishing
|
|
|
|
* work so that it can process unfinished items recovered from the log in
|
|
|
|
* correct order.
|
|
|
|
*
|
|
|
|
* Create and log intent items for all the work that we're capturing so that we
|
|
|
|
* can be assured that the items will get replayed if the system goes down
|
xfs: proper replay of deferred ops queued during log recovery
When we replay unfinished intent items that have been recovered from the
log, it's possible that the replay will cause the creation of more
deferred work items. As outlined in commit 509955823cc9c ("xfs: log
recovery should replay deferred ops in order"), later work items have an
implicit ordering dependency on earlier work items. Therefore, recovery
must replay the items (both recovered and created) in the same order
that they would have been during normal operation.
For log recovery, we enforce this ordering by using an empty transaction
to collect deferred ops that get created in the process of recovering a
log intent item to prevent them from being committed before the rest of
the recovered intent items. After we finish committing all the
recovered log items, we allocate a transaction with an enormous block
reservation, splice our huge list of created deferred ops into that
transaction, and commit it, thereby finishing all those ops.
This is /really/ hokey -- it's the one place in XFS where we allow
nested transactions; the splicing of the defer ops list is is inelegant
and has to be done twice per recovery function; and the broken way we
handle inode pointers and block reservations cause subtle use-after-free
and allocator problems that will be fixed by this patch and the two
patches after it.
Therefore, replace the hokey empty transaction with a structure designed
to capture each chain of deferred ops that are created as part of
recovering a single unfinished log intent. Finally, refactor the loop
that replays those chains to do so using one transaction per chain.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-09-26 08:39:37 +08:00
|
|
|
* before log recovery gets a chance to finish the work it put off. The entire
|
|
|
|
* deferred ops state is transferred to the capture structure and the
|
|
|
|
* transaction is then ready for the caller to commit it. If there are no
|
|
|
|
* intent items to capture, this function returns NULL.
|
2020-09-22 00:15:09 +08:00
|
|
|
*/
|
xfs: proper replay of deferred ops queued during log recovery
When we replay unfinished intent items that have been recovered from the
log, it's possible that the replay will cause the creation of more
deferred work items. As outlined in commit 509955823cc9c ("xfs: log
recovery should replay deferred ops in order"), later work items have an
implicit ordering dependency on earlier work items. Therefore, recovery
must replay the items (both recovered and created) in the same order
that they would have been during normal operation.
For log recovery, we enforce this ordering by using an empty transaction
to collect deferred ops that get created in the process of recovering a
log intent item to prevent them from being committed before the rest of
the recovered intent items. After we finish committing all the
recovered log items, we allocate a transaction with an enormous block
reservation, splice our huge list of created deferred ops into that
transaction, and commit it, thereby finishing all those ops.
This is /really/ hokey -- it's the one place in XFS where we allow
nested transactions; the splicing of the defer ops list is is inelegant
and has to be done twice per recovery function; and the broken way we
handle inode pointers and block reservations cause subtle use-after-free
and allocator problems that will be fixed by this patch and the two
patches after it.
Therefore, replace the hokey empty transaction with a structure designed
to capture each chain of deferred ops that are created as part of
recovering a single unfinished log intent. Finally, refactor the loop
that replays those chains to do so using one transaction per chain.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-09-26 08:39:37 +08:00
|
|
|
static struct xfs_defer_capture *
|
|
|
|
xfs_defer_ops_capture(
|
|
|
|
struct xfs_trans *tp)
|
|
|
|
{
|
|
|
|
struct xfs_defer_capture *dfc;
|
|
|
|
|
|
|
|
if (list_empty(&tp->t_dfops))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* Create an object to capture the defer ops. */
|
|
|
|
dfc = kmem_zalloc(sizeof(*dfc), KM_NOFS);
|
|
|
|
INIT_LIST_HEAD(&dfc->dfc_list);
|
|
|
|
INIT_LIST_HEAD(&dfc->dfc_dfops);
|
|
|
|
|
|
|
|
xfs_defer_create_intents(tp);
|
|
|
|
|
|
|
|
/* Move the dfops chain and transaction state to the capture struct. */
|
|
|
|
list_splice_init(&tp->t_dfops, &dfc->dfc_dfops);
|
|
|
|
dfc->dfc_tpflags = tp->t_flags & XFS_TRANS_LOWMODE;
|
|
|
|
tp->t_flags &= ~XFS_TRANS_LOWMODE;
|
|
|
|
|
2020-09-26 08:39:49 +08:00
|
|
|
/* Capture the remaining block reservations along with the dfops. */
|
|
|
|
dfc->dfc_blkres = tp->t_blk_res - tp->t_blk_res_used;
|
|
|
|
dfc->dfc_rtxres = tp->t_rtx_res - tp->t_rtx_res_used;
|
|
|
|
|
xfs: proper replay of deferred ops queued during log recovery
When we replay unfinished intent items that have been recovered from the
log, it's possible that the replay will cause the creation of more
deferred work items. As outlined in commit 509955823cc9c ("xfs: log
recovery should replay deferred ops in order"), later work items have an
implicit ordering dependency on earlier work items. Therefore, recovery
must replay the items (both recovered and created) in the same order
that they would have been during normal operation.
For log recovery, we enforce this ordering by using an empty transaction
to collect deferred ops that get created in the process of recovering a
log intent item to prevent them from being committed before the rest of
the recovered intent items. After we finish committing all the
recovered log items, we allocate a transaction with an enormous block
reservation, splice our huge list of created deferred ops into that
transaction, and commit it, thereby finishing all those ops.
This is /really/ hokey -- it's the one place in XFS where we allow
nested transactions; the splicing of the defer ops list is is inelegant
and has to be done twice per recovery function; and the broken way we
handle inode pointers and block reservations cause subtle use-after-free
and allocator problems that will be fixed by this patch and the two
patches after it.
Therefore, replace the hokey empty transaction with a structure designed
to capture each chain of deferred ops that are created as part of
recovering a single unfinished log intent. Finally, refactor the loop
that replays those chains to do so using one transaction per chain.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-09-26 08:39:37 +08:00
|
|
|
return dfc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Release all resources that we used to capture deferred ops. */
|
2020-09-22 00:15:09 +08:00
|
|
|
void
|
xfs: proper replay of deferred ops queued during log recovery
When we replay unfinished intent items that have been recovered from the
log, it's possible that the replay will cause the creation of more
deferred work items. As outlined in commit 509955823cc9c ("xfs: log
recovery should replay deferred ops in order"), later work items have an
implicit ordering dependency on earlier work items. Therefore, recovery
must replay the items (both recovered and created) in the same order
that they would have been during normal operation.
For log recovery, we enforce this ordering by using an empty transaction
to collect deferred ops that get created in the process of recovering a
log intent item to prevent them from being committed before the rest of
the recovered intent items. After we finish committing all the
recovered log items, we allocate a transaction with an enormous block
reservation, splice our huge list of created deferred ops into that
transaction, and commit it, thereby finishing all those ops.
This is /really/ hokey -- it's the one place in XFS where we allow
nested transactions; the splicing of the defer ops list is is inelegant
and has to be done twice per recovery function; and the broken way we
handle inode pointers and block reservations cause subtle use-after-free
and allocator problems that will be fixed by this patch and the two
patches after it.
Therefore, replace the hokey empty transaction with a structure designed
to capture each chain of deferred ops that are created as part of
recovering a single unfinished log intent. Finally, refactor the loop
that replays those chains to do so using one transaction per chain.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-09-26 08:39:37 +08:00
|
|
|
xfs_defer_ops_release(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct xfs_defer_capture *dfc)
|
|
|
|
{
|
|
|
|
xfs_defer_cancel_list(mp, &dfc->dfc_dfops);
|
|
|
|
kmem_free(dfc);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Capture any deferred ops and commit the transaction. This is the last step
|
|
|
|
* needed to finish a log intent item that we recovered from the log.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
xfs_defer_ops_capture_and_commit(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct list_head *capture_list)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = tp->t_mountp;
|
|
|
|
struct xfs_defer_capture *dfc;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
/* If we don't capture anything, commit transaction and exit. */
|
|
|
|
dfc = xfs_defer_ops_capture(tp);
|
|
|
|
if (!dfc)
|
|
|
|
return xfs_trans_commit(tp);
|
|
|
|
|
|
|
|
/* Commit the transaction and add the capture structure to the list. */
|
|
|
|
error = xfs_trans_commit(tp);
|
|
|
|
if (error) {
|
|
|
|
xfs_defer_ops_release(mp, dfc);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_add_tail(&dfc->dfc_list, capture_list);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Attach a chain of captured deferred ops to a new transaction and free the
|
|
|
|
* capture structure.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xfs_defer_ops_continue(
|
|
|
|
struct xfs_defer_capture *dfc,
|
|
|
|
struct xfs_trans *tp)
|
2020-09-22 00:15:09 +08:00
|
|
|
{
|
xfs: proper replay of deferred ops queued during log recovery
When we replay unfinished intent items that have been recovered from the
log, it's possible that the replay will cause the creation of more
deferred work items. As outlined in commit 509955823cc9c ("xfs: log
recovery should replay deferred ops in order"), later work items have an
implicit ordering dependency on earlier work items. Therefore, recovery
must replay the items (both recovered and created) in the same order
that they would have been during normal operation.
For log recovery, we enforce this ordering by using an empty transaction
to collect deferred ops that get created in the process of recovering a
log intent item to prevent them from being committed before the rest of
the recovered intent items. After we finish committing all the
recovered log items, we allocate a transaction with an enormous block
reservation, splice our huge list of created deferred ops into that
transaction, and commit it, thereby finishing all those ops.
This is /really/ hokey -- it's the one place in XFS where we allow
nested transactions; the splicing of the defer ops list is is inelegant
and has to be done twice per recovery function; and the broken way we
handle inode pointers and block reservations cause subtle use-after-free
and allocator problems that will be fixed by this patch and the two
patches after it.
Therefore, replace the hokey empty transaction with a structure designed
to capture each chain of deferred ops that are created as part of
recovering a single unfinished log intent. Finally, refactor the loop
that replays those chains to do so using one transaction per chain.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2020-09-26 08:39:37 +08:00
|
|
|
ASSERT(tp->t_flags & XFS_TRANS_PERM_LOG_RES);
|
|
|
|
ASSERT(!(tp->t_flags & XFS_TRANS_DIRTY));
|
|
|
|
|
|
|
|
/* Move captured dfops chain and state to the transaction. */
|
|
|
|
list_splice_init(&dfc->dfc_dfops, &tp->t_dfops);
|
|
|
|
tp->t_flags |= dfc->dfc_tpflags;
|
|
|
|
|
|
|
|
kmem_free(dfc);
|
2020-09-22 00:15:09 +08:00
|
|
|
}
|