mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 20:23:57 +08:00
6ca444cfd6
Split out the btree level information into a separate struct and put it at the end of the cursor structure as a VLA. Files with huge data forks (and in the future, the realtime rmap btree) will require the ability to support many more levels than a per-AG btree cursor, which means that we're going to create per-btree type cursor caches to conserve memory for the more common case. Note that a subsequent patch actually introduces dynamic cursor heights. This one merely rearranges the structure to prepare for that. Signed-off-by: Darrick J. Wong <djwong@kernel.org> Reviewed-by: Chandan Babu R <chandan.babu@oracle.com> Reviewed-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Dave Chinner <dchinner@redhat.com>
315 lines
7.8 KiB
C
315 lines
7.8 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/*
|
|
* Copyright (C) 2018 Oracle. All Rights Reserved.
|
|
* Author: Darrick J. Wong <darrick.wong@oracle.com>
|
|
*/
|
|
#include "xfs.h"
|
|
#include "xfs_fs.h"
|
|
#include "xfs_shared.h"
|
|
#include "xfs_format.h"
|
|
#include "xfs_trans_resv.h"
|
|
#include "xfs_mount.h"
|
|
#include "xfs_btree.h"
|
|
#include "scrub/bitmap.h"
|
|
|
|
/*
|
|
* Set a range of this bitmap. Caller must ensure the range is not set.
|
|
*
|
|
* This is the logical equivalent of bitmap |= mask(start, len).
|
|
*/
|
|
int
|
|
xbitmap_set(
|
|
struct xbitmap *bitmap,
|
|
uint64_t start,
|
|
uint64_t len)
|
|
{
|
|
struct xbitmap_range *bmr;
|
|
|
|
bmr = kmem_alloc(sizeof(struct xbitmap_range), KM_MAYFAIL);
|
|
if (!bmr)
|
|
return -ENOMEM;
|
|
|
|
INIT_LIST_HEAD(&bmr->list);
|
|
bmr->start = start;
|
|
bmr->len = len;
|
|
list_add_tail(&bmr->list, &bitmap->list);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Free everything related to this bitmap. */
|
|
void
|
|
xbitmap_destroy(
|
|
struct xbitmap *bitmap)
|
|
{
|
|
struct xbitmap_range *bmr;
|
|
struct xbitmap_range *n;
|
|
|
|
for_each_xbitmap_extent(bmr, n, bitmap) {
|
|
list_del(&bmr->list);
|
|
kmem_free(bmr);
|
|
}
|
|
}
|
|
|
|
/* Set up a per-AG block bitmap. */
|
|
void
|
|
xbitmap_init(
|
|
struct xbitmap *bitmap)
|
|
{
|
|
INIT_LIST_HEAD(&bitmap->list);
|
|
}
|
|
|
|
/* Compare two btree extents. */
|
|
static int
|
|
xbitmap_range_cmp(
|
|
void *priv,
|
|
const struct list_head *a,
|
|
const struct list_head *b)
|
|
{
|
|
struct xbitmap_range *ap;
|
|
struct xbitmap_range *bp;
|
|
|
|
ap = container_of(a, struct xbitmap_range, list);
|
|
bp = container_of(b, struct xbitmap_range, list);
|
|
|
|
if (ap->start > bp->start)
|
|
return 1;
|
|
if (ap->start < bp->start)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Remove all the blocks mentioned in @sub from the extents in @bitmap.
|
|
*
|
|
* The intent is that callers will iterate the rmapbt for all of its records
|
|
* for a given owner to generate @bitmap; and iterate all the blocks of the
|
|
* metadata structures that are not being rebuilt and have the same rmapbt
|
|
* owner to generate @sub. This routine subtracts all the extents
|
|
* mentioned in sub from all the extents linked in @bitmap, which leaves
|
|
* @bitmap as the list of blocks that are not accounted for, which we assume
|
|
* are the dead blocks of the old metadata structure. The blocks mentioned in
|
|
* @bitmap can be reaped.
|
|
*
|
|
* This is the logical equivalent of bitmap &= ~sub.
|
|
*/
|
|
#define LEFT_ALIGNED (1 << 0)
|
|
#define RIGHT_ALIGNED (1 << 1)
|
|
int
|
|
xbitmap_disunion(
|
|
struct xbitmap *bitmap,
|
|
struct xbitmap *sub)
|
|
{
|
|
struct list_head *lp;
|
|
struct xbitmap_range *br;
|
|
struct xbitmap_range *new_br;
|
|
struct xbitmap_range *sub_br;
|
|
uint64_t sub_start;
|
|
uint64_t sub_len;
|
|
int state;
|
|
int error = 0;
|
|
|
|
if (list_empty(&bitmap->list) || list_empty(&sub->list))
|
|
return 0;
|
|
ASSERT(!list_empty(&sub->list));
|
|
|
|
list_sort(NULL, &bitmap->list, xbitmap_range_cmp);
|
|
list_sort(NULL, &sub->list, xbitmap_range_cmp);
|
|
|
|
/*
|
|
* Now that we've sorted both lists, we iterate bitmap once, rolling
|
|
* forward through sub and/or bitmap as necessary until we find an
|
|
* overlap or reach the end of either list. We do not reset lp to the
|
|
* head of bitmap nor do we reset sub_br to the head of sub. The
|
|
* list traversal is similar to merge sort, but we're deleting
|
|
* instead. In this manner we avoid O(n^2) operations.
|
|
*/
|
|
sub_br = list_first_entry(&sub->list, struct xbitmap_range,
|
|
list);
|
|
lp = bitmap->list.next;
|
|
while (lp != &bitmap->list) {
|
|
br = list_entry(lp, struct xbitmap_range, list);
|
|
|
|
/*
|
|
* Advance sub_br and/or br until we find a pair that
|
|
* intersect or we run out of extents.
|
|
*/
|
|
while (sub_br->start + sub_br->len <= br->start) {
|
|
if (list_is_last(&sub_br->list, &sub->list))
|
|
goto out;
|
|
sub_br = list_next_entry(sub_br, list);
|
|
}
|
|
if (sub_br->start >= br->start + br->len) {
|
|
lp = lp->next;
|
|
continue;
|
|
}
|
|
|
|
/* trim sub_br to fit the extent we have */
|
|
sub_start = sub_br->start;
|
|
sub_len = sub_br->len;
|
|
if (sub_br->start < br->start) {
|
|
sub_len -= br->start - sub_br->start;
|
|
sub_start = br->start;
|
|
}
|
|
if (sub_len > br->len)
|
|
sub_len = br->len;
|
|
|
|
state = 0;
|
|
if (sub_start == br->start)
|
|
state |= LEFT_ALIGNED;
|
|
if (sub_start + sub_len == br->start + br->len)
|
|
state |= RIGHT_ALIGNED;
|
|
switch (state) {
|
|
case LEFT_ALIGNED:
|
|
/* Coincides with only the left. */
|
|
br->start += sub_len;
|
|
br->len -= sub_len;
|
|
break;
|
|
case RIGHT_ALIGNED:
|
|
/* Coincides with only the right. */
|
|
br->len -= sub_len;
|
|
lp = lp->next;
|
|
break;
|
|
case LEFT_ALIGNED | RIGHT_ALIGNED:
|
|
/* Total overlap, just delete ex. */
|
|
lp = lp->next;
|
|
list_del(&br->list);
|
|
kmem_free(br);
|
|
break;
|
|
case 0:
|
|
/*
|
|
* Deleting from the middle: add the new right extent
|
|
* and then shrink the left extent.
|
|
*/
|
|
new_br = kmem_alloc(sizeof(struct xbitmap_range),
|
|
KM_MAYFAIL);
|
|
if (!new_br) {
|
|
error = -ENOMEM;
|
|
goto out;
|
|
}
|
|
INIT_LIST_HEAD(&new_br->list);
|
|
new_br->start = sub_start + sub_len;
|
|
new_br->len = br->start + br->len - new_br->start;
|
|
list_add(&new_br->list, &br->list);
|
|
br->len = sub_start - br->start;
|
|
lp = lp->next;
|
|
break;
|
|
default:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
out:
|
|
return error;
|
|
}
|
|
#undef LEFT_ALIGNED
|
|
#undef RIGHT_ALIGNED
|
|
|
|
/*
|
|
* Record all btree blocks seen while iterating all records of a btree.
|
|
*
|
|
* We know that the btree query_all function starts at the left edge and walks
|
|
* towards the right edge of the tree. Therefore, we know that we can walk up
|
|
* the btree cursor towards the root; if the pointer for a given level points
|
|
* to the first record/key in that block, we haven't seen this block before;
|
|
* and therefore we need to remember that we saw this block in the btree.
|
|
*
|
|
* So if our btree is:
|
|
*
|
|
* 4
|
|
* / | \
|
|
* 1 2 3
|
|
*
|
|
* Pretend for this example that each leaf block has 100 btree records. For
|
|
* the first btree record, we'll observe that bc_levels[0].ptr == 1, so we
|
|
* record that we saw block 1. Then we observe that bc_levels[1].ptr == 1, so
|
|
* we record block 4. The list is [1, 4].
|
|
*
|
|
* For the second btree record, we see that bc_levels[0].ptr == 2, so we exit
|
|
* the loop. The list remains [1, 4].
|
|
*
|
|
* For the 101st btree record, we've moved onto leaf block 2. Now
|
|
* bc_levels[0].ptr == 1 again, so we record that we saw block 2. We see that
|
|
* bc_levels[1].ptr == 2, so we exit the loop. The list is now [1, 4, 2].
|
|
*
|
|
* For the 102nd record, bc_levels[0].ptr == 2, so we continue.
|
|
*
|
|
* For the 201st record, we've moved on to leaf block 3.
|
|
* bc_levels[0].ptr == 1, so we add 3 to the list. Now it is [1, 4, 2, 3].
|
|
*
|
|
* For the 300th record we just exit, with the list being [1, 4, 2, 3].
|
|
*/
|
|
|
|
/*
|
|
* Record all the buffers pointed to by the btree cursor. Callers already
|
|
* engaged in a btree walk should call this function to capture the list of
|
|
* blocks going from the leaf towards the root.
|
|
*/
|
|
int
|
|
xbitmap_set_btcur_path(
|
|
struct xbitmap *bitmap,
|
|
struct xfs_btree_cur *cur)
|
|
{
|
|
struct xfs_buf *bp;
|
|
xfs_fsblock_t fsb;
|
|
int i;
|
|
int error;
|
|
|
|
for (i = 0; i < cur->bc_nlevels && cur->bc_levels[i].ptr == 1; i++) {
|
|
xfs_btree_get_block(cur, i, &bp);
|
|
if (!bp)
|
|
continue;
|
|
fsb = XFS_DADDR_TO_FSB(cur->bc_mp, xfs_buf_daddr(bp));
|
|
error = xbitmap_set(bitmap, fsb, 1);
|
|
if (error)
|
|
return error;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Collect a btree's block in the bitmap. */
|
|
STATIC int
|
|
xbitmap_collect_btblock(
|
|
struct xfs_btree_cur *cur,
|
|
int level,
|
|
void *priv)
|
|
{
|
|
struct xbitmap *bitmap = priv;
|
|
struct xfs_buf *bp;
|
|
xfs_fsblock_t fsbno;
|
|
|
|
xfs_btree_get_block(cur, level, &bp);
|
|
if (!bp)
|
|
return 0;
|
|
|
|
fsbno = XFS_DADDR_TO_FSB(cur->bc_mp, xfs_buf_daddr(bp));
|
|
return xbitmap_set(bitmap, fsbno, 1);
|
|
}
|
|
|
|
/* Walk the btree and mark the bitmap wherever a btree block is found. */
|
|
int
|
|
xbitmap_set_btblocks(
|
|
struct xbitmap *bitmap,
|
|
struct xfs_btree_cur *cur)
|
|
{
|
|
return xfs_btree_visit_blocks(cur, xbitmap_collect_btblock,
|
|
XFS_BTREE_VISIT_ALL, bitmap);
|
|
}
|
|
|
|
/* How many bits are set in this bitmap? */
|
|
uint64_t
|
|
xbitmap_hweight(
|
|
struct xbitmap *bitmap)
|
|
{
|
|
struct xbitmap_range *bmr;
|
|
struct xbitmap_range *n;
|
|
uint64_t ret = 0;
|
|
|
|
for_each_xbitmap_extent(bmr, n, bitmap)
|
|
ret += bmr->len;
|
|
|
|
return ret;
|
|
}
|