mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-11 12:28:41 +08:00
btrfs: rename btrfs_bio to btrfs_io_context
The structure btrfs_bio is used by two different sites: - bio->bi_private for mirror based profiles For those profiles (SINGLE/DUP/RAID1*/RAID10), this structures records how many mirrors are still pending, and save the original endio function of the bio. - RAID56 code In that case, RAID56 only utilize the stripes info, and no long uses that to trace the pending mirrors. So btrfs_bio is not always bind to a bio, and contains more info for IO context, thus renaming it will make the naming less confusing. Signed-off-by: Qu Wenruo <wqu@suse.com> Reviewed-by: David Sterba <dsterba@suse.com> Signed-off-by: David Sterba <dsterba@suse.com>
This commit is contained in:
parent
dc2872247e
commit
4c66461179
@ -1455,7 +1455,7 @@ static int btrfsic_map_block(struct btrfsic_state *state, u64 bytenr, u32 len,
|
||||
struct btrfs_fs_info *fs_info = state->fs_info;
|
||||
int ret;
|
||||
u64 length;
|
||||
struct btrfs_bio *multi = NULL;
|
||||
struct btrfs_io_context *multi = NULL;
|
||||
struct btrfs_device *device;
|
||||
|
||||
length = len;
|
||||
|
@ -1266,7 +1266,7 @@ static int btrfs_issue_discard(struct block_device *bdev, u64 start, u64 len,
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int do_discard_extent(struct btrfs_bio_stripe *stripe, u64 *bytes)
|
||||
static int do_discard_extent(struct btrfs_io_stripe *stripe, u64 *bytes)
|
||||
{
|
||||
struct btrfs_device *dev = stripe->dev;
|
||||
struct btrfs_fs_info *fs_info = dev->fs_info;
|
||||
@ -1313,22 +1313,21 @@ int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
u64 discarded_bytes = 0;
|
||||
u64 end = bytenr + num_bytes;
|
||||
u64 cur = bytenr;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
|
||||
/*
|
||||
* Avoid races with device replace and make sure our bbio has devices
|
||||
* Avoid races with device replace and make sure our bioc has devices
|
||||
* associated to its stripes that don't go away while we are discarding.
|
||||
*/
|
||||
btrfs_bio_counter_inc_blocked(fs_info);
|
||||
while (cur < end) {
|
||||
struct btrfs_bio_stripe *stripe;
|
||||
struct btrfs_io_stripe *stripe;
|
||||
int i;
|
||||
|
||||
num_bytes = end - cur;
|
||||
/* Tell the block device(s) that the sectors can be discarded */
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_DISCARD, cur,
|
||||
&num_bytes, &bbio, 0);
|
||||
&num_bytes, &bioc, 0);
|
||||
/*
|
||||
* Error can be -ENOMEM, -ENOENT (no such chunk mapping) or
|
||||
* -EOPNOTSUPP. For any such error, @num_bytes is not updated,
|
||||
@ -1337,8 +1336,8 @@ int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
|
||||
stripe = bbio->stripes;
|
||||
for (i = 0; i < bbio->num_stripes; i++, stripe++) {
|
||||
stripe = bioc->stripes;
|
||||
for (i = 0; i < bioc->num_stripes; i++, stripe++) {
|
||||
u64 bytes;
|
||||
struct btrfs_device *device = stripe->dev;
|
||||
|
||||
@ -1361,7 +1360,7 @@ int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
* And since there are two loops, explicitly
|
||||
* go to out to avoid confusion.
|
||||
*/
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
goto out;
|
||||
}
|
||||
|
||||
@ -1372,7 +1371,7 @@ int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
*/
|
||||
ret = 0;
|
||||
}
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
cur += num_bytes;
|
||||
}
|
||||
out:
|
||||
|
@ -2290,7 +2290,7 @@ static int repair_io_failure(struct btrfs_fs_info *fs_info, u64 ino, u64 start,
|
||||
struct btrfs_device *dev;
|
||||
u64 map_length = 0;
|
||||
u64 sector;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
int ret;
|
||||
|
||||
ASSERT(!(fs_info->sb->s_flags & SB_RDONLY));
|
||||
@ -2304,7 +2304,7 @@ static int repair_io_failure(struct btrfs_fs_info *fs_info, u64 ino, u64 start,
|
||||
map_length = length;
|
||||
|
||||
/*
|
||||
* Avoid races with device replace and make sure our bbio has devices
|
||||
* Avoid races with device replace and make sure our bioc has devices
|
||||
* associated to its stripes that don't go away while we are doing the
|
||||
* read repair operation.
|
||||
*/
|
||||
@ -2317,28 +2317,28 @@ static int repair_io_failure(struct btrfs_fs_info *fs_info, u64 ino, u64 start,
|
||||
* stripe's dev and sector.
|
||||
*/
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_READ, logical,
|
||||
&map_length, &bbio, 0);
|
||||
&map_length, &bioc, 0);
|
||||
if (ret) {
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
bio_put(bio);
|
||||
return -EIO;
|
||||
}
|
||||
ASSERT(bbio->mirror_num == 1);
|
||||
ASSERT(bioc->mirror_num == 1);
|
||||
} else {
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_WRITE, logical,
|
||||
&map_length, &bbio, mirror_num);
|
||||
&map_length, &bioc, mirror_num);
|
||||
if (ret) {
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
bio_put(bio);
|
||||
return -EIO;
|
||||
}
|
||||
BUG_ON(mirror_num != bbio->mirror_num);
|
||||
BUG_ON(mirror_num != bioc->mirror_num);
|
||||
}
|
||||
|
||||
sector = bbio->stripes[bbio->mirror_num - 1].physical >> 9;
|
||||
sector = bioc->stripes[bioc->mirror_num - 1].physical >> 9;
|
||||
bio->bi_iter.bi_sector = sector;
|
||||
dev = bbio->stripes[bbio->mirror_num - 1].dev;
|
||||
btrfs_put_bbio(bbio);
|
||||
dev = bioc->stripes[bioc->mirror_num - 1].dev;
|
||||
btrfs_put_bioc(bioc);
|
||||
if (!dev || !dev->bdev ||
|
||||
!test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state)) {
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
|
@ -360,7 +360,7 @@ static void extent_map_device_set_bits(struct extent_map *em, unsigned bits)
|
||||
int i;
|
||||
|
||||
for (i = 0; i < map->num_stripes; i++) {
|
||||
struct btrfs_bio_stripe *stripe = &map->stripes[i];
|
||||
struct btrfs_io_stripe *stripe = &map->stripes[i];
|
||||
struct btrfs_device *device = stripe->dev;
|
||||
|
||||
set_extent_bits_nowait(&device->alloc_state, stripe->physical,
|
||||
@ -375,7 +375,7 @@ static void extent_map_device_clear_bits(struct extent_map *em, unsigned bits)
|
||||
int i;
|
||||
|
||||
for (i = 0; i < map->num_stripes; i++) {
|
||||
struct btrfs_bio_stripe *stripe = &map->stripes[i];
|
||||
struct btrfs_io_stripe *stripe = &map->stripes[i];
|
||||
struct btrfs_device *device = stripe->dev;
|
||||
|
||||
__clear_extent_bit(&device->alloc_state, stripe->physical,
|
||||
|
@ -61,7 +61,7 @@ enum btrfs_rbio_ops {
|
||||
|
||||
struct btrfs_raid_bio {
|
||||
struct btrfs_fs_info *fs_info;
|
||||
struct btrfs_bio *bbio;
|
||||
struct btrfs_io_context *bioc;
|
||||
|
||||
/* while we're doing rmw on a stripe
|
||||
* we put it into a hash table so we can
|
||||
@ -271,7 +271,7 @@ static void cache_rbio_pages(struct btrfs_raid_bio *rbio)
|
||||
*/
|
||||
static int rbio_bucket(struct btrfs_raid_bio *rbio)
|
||||
{
|
||||
u64 num = rbio->bbio->raid_map[0];
|
||||
u64 num = rbio->bioc->raid_map[0];
|
||||
|
||||
/*
|
||||
* we shift down quite a bit. We're using byte
|
||||
@ -559,8 +559,7 @@ static int rbio_can_merge(struct btrfs_raid_bio *last,
|
||||
test_bit(RBIO_CACHE_BIT, &cur->flags))
|
||||
return 0;
|
||||
|
||||
if (last->bbio->raid_map[0] !=
|
||||
cur->bbio->raid_map[0])
|
||||
if (last->bioc->raid_map[0] != cur->bioc->raid_map[0])
|
||||
return 0;
|
||||
|
||||
/* we can't merge with different operations */
|
||||
@ -673,7 +672,7 @@ static noinline int lock_stripe_add(struct btrfs_raid_bio *rbio)
|
||||
|
||||
spin_lock_irqsave(&h->lock, flags);
|
||||
list_for_each_entry(cur, &h->hash_list, hash_list) {
|
||||
if (cur->bbio->raid_map[0] != rbio->bbio->raid_map[0])
|
||||
if (cur->bioc->raid_map[0] != rbio->bioc->raid_map[0])
|
||||
continue;
|
||||
|
||||
spin_lock(&cur->bio_list_lock);
|
||||
@ -838,7 +837,7 @@ static void __free_raid_bio(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
}
|
||||
|
||||
btrfs_put_bbio(rbio->bbio);
|
||||
btrfs_put_bioc(rbio->bioc);
|
||||
kfree(rbio);
|
||||
}
|
||||
|
||||
@ -906,7 +905,7 @@ static void raid_write_end_io(struct bio *bio)
|
||||
|
||||
/* OK, we have read all the stripes we need to. */
|
||||
max_errors = (rbio->operation == BTRFS_RBIO_PARITY_SCRUB) ?
|
||||
0 : rbio->bbio->max_errors;
|
||||
0 : rbio->bioc->max_errors;
|
||||
if (atomic_read(&rbio->error) > max_errors)
|
||||
err = BLK_STS_IOERR;
|
||||
|
||||
@ -961,12 +960,12 @@ static unsigned long rbio_nr_pages(unsigned long stripe_len, int nr_stripes)
|
||||
* this does not allocate any pages for rbio->pages.
|
||||
*/
|
||||
static struct btrfs_raid_bio *alloc_rbio(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_bio *bbio,
|
||||
struct btrfs_io_context *bioc,
|
||||
u64 stripe_len)
|
||||
{
|
||||
struct btrfs_raid_bio *rbio;
|
||||
int nr_data = 0;
|
||||
int real_stripes = bbio->num_stripes - bbio->num_tgtdevs;
|
||||
int real_stripes = bioc->num_stripes - bioc->num_tgtdevs;
|
||||
int num_pages = rbio_nr_pages(stripe_len, real_stripes);
|
||||
int stripe_npages = DIV_ROUND_UP(stripe_len, PAGE_SIZE);
|
||||
void *p;
|
||||
@ -987,7 +986,7 @@ static struct btrfs_raid_bio *alloc_rbio(struct btrfs_fs_info *fs_info,
|
||||
spin_lock_init(&rbio->bio_list_lock);
|
||||
INIT_LIST_HEAD(&rbio->stripe_cache);
|
||||
INIT_LIST_HEAD(&rbio->hash_list);
|
||||
rbio->bbio = bbio;
|
||||
rbio->bioc = bioc;
|
||||
rbio->fs_info = fs_info;
|
||||
rbio->stripe_len = stripe_len;
|
||||
rbio->nr_pages = num_pages;
|
||||
@ -1015,9 +1014,9 @@ static struct btrfs_raid_bio *alloc_rbio(struct btrfs_fs_info *fs_info,
|
||||
CONSUME_ALLOC(rbio->finish_pbitmap, BITS_TO_LONGS(stripe_npages));
|
||||
#undef CONSUME_ALLOC
|
||||
|
||||
if (bbio->map_type & BTRFS_BLOCK_GROUP_RAID5)
|
||||
if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID5)
|
||||
nr_data = real_stripes - 1;
|
||||
else if (bbio->map_type & BTRFS_BLOCK_GROUP_RAID6)
|
||||
else if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID6)
|
||||
nr_data = real_stripes - 2;
|
||||
else
|
||||
BUG();
|
||||
@ -1077,10 +1076,10 @@ static int rbio_add_io_page(struct btrfs_raid_bio *rbio,
|
||||
struct bio *last = bio_list->tail;
|
||||
int ret;
|
||||
struct bio *bio;
|
||||
struct btrfs_bio_stripe *stripe;
|
||||
struct btrfs_io_stripe *stripe;
|
||||
u64 disk_start;
|
||||
|
||||
stripe = &rbio->bbio->stripes[stripe_nr];
|
||||
stripe = &rbio->bioc->stripes[stripe_nr];
|
||||
disk_start = stripe->physical + (page_index << PAGE_SHIFT);
|
||||
|
||||
/* if the device is missing, just fail this stripe */
|
||||
@ -1155,7 +1154,7 @@ static void index_rbio_pages(struct btrfs_raid_bio *rbio)
|
||||
int i = 0;
|
||||
|
||||
start = bio->bi_iter.bi_sector << 9;
|
||||
stripe_offset = start - rbio->bbio->raid_map[0];
|
||||
stripe_offset = start - rbio->bioc->raid_map[0];
|
||||
page_index = stripe_offset >> PAGE_SHIFT;
|
||||
|
||||
if (bio_flagged(bio, BIO_CLONED))
|
||||
@ -1179,7 +1178,7 @@ static void index_rbio_pages(struct btrfs_raid_bio *rbio)
|
||||
*/
|
||||
static noinline void finish_rmw(struct btrfs_raid_bio *rbio)
|
||||
{
|
||||
struct btrfs_bio *bbio = rbio->bbio;
|
||||
struct btrfs_io_context *bioc = rbio->bioc;
|
||||
void **pointers = rbio->finish_pointers;
|
||||
int nr_data = rbio->nr_data;
|
||||
int stripe;
|
||||
@ -1284,11 +1283,11 @@ static noinline void finish_rmw(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
}
|
||||
|
||||
if (likely(!bbio->num_tgtdevs))
|
||||
if (likely(!bioc->num_tgtdevs))
|
||||
goto write_data;
|
||||
|
||||
for (stripe = 0; stripe < rbio->real_stripes; stripe++) {
|
||||
if (!bbio->tgtdev_map[stripe])
|
||||
if (!bioc->tgtdev_map[stripe])
|
||||
continue;
|
||||
|
||||
for (pagenr = 0; pagenr < rbio->stripe_npages; pagenr++) {
|
||||
@ -1302,7 +1301,7 @@ static noinline void finish_rmw(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
|
||||
ret = rbio_add_io_page(rbio, &bio_list, page,
|
||||
rbio->bbio->tgtdev_map[stripe],
|
||||
rbio->bioc->tgtdev_map[stripe],
|
||||
pagenr, rbio->stripe_len);
|
||||
if (ret)
|
||||
goto cleanup;
|
||||
@ -1339,12 +1338,12 @@ static int find_bio_stripe(struct btrfs_raid_bio *rbio,
|
||||
{
|
||||
u64 physical = bio->bi_iter.bi_sector;
|
||||
int i;
|
||||
struct btrfs_bio_stripe *stripe;
|
||||
struct btrfs_io_stripe *stripe;
|
||||
|
||||
physical <<= 9;
|
||||
|
||||
for (i = 0; i < rbio->bbio->num_stripes; i++) {
|
||||
stripe = &rbio->bbio->stripes[i];
|
||||
for (i = 0; i < rbio->bioc->num_stripes; i++) {
|
||||
stripe = &rbio->bioc->stripes[i];
|
||||
if (in_range(physical, stripe->physical, rbio->stripe_len) &&
|
||||
stripe->dev->bdev && bio->bi_bdev == stripe->dev->bdev) {
|
||||
return i;
|
||||
@ -1365,7 +1364,7 @@ static int find_logical_bio_stripe(struct btrfs_raid_bio *rbio,
|
||||
int i;
|
||||
|
||||
for (i = 0; i < rbio->nr_data; i++) {
|
||||
u64 stripe_start = rbio->bbio->raid_map[i];
|
||||
u64 stripe_start = rbio->bioc->raid_map[i];
|
||||
|
||||
if (in_range(logical, stripe_start, rbio->stripe_len))
|
||||
return i;
|
||||
@ -1456,7 +1455,7 @@ static void raid_rmw_end_io(struct bio *bio)
|
||||
if (!atomic_dec_and_test(&rbio->stripes_pending))
|
||||
return;
|
||||
|
||||
if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
|
||||
if (atomic_read(&rbio->error) > rbio->bioc->max_errors)
|
||||
goto cleanup;
|
||||
|
||||
/*
|
||||
@ -1538,8 +1537,8 @@ static int raid56_rmw_stripe(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
|
||||
/*
|
||||
* the bbio may be freed once we submit the last bio. Make sure
|
||||
* not to touch it after that
|
||||
* The bioc may be freed once we submit the last bio. Make sure not to
|
||||
* touch it after that.
|
||||
*/
|
||||
atomic_set(&rbio->stripes_pending, bios_to_read);
|
||||
while ((bio = bio_list_pop(&bio_list))) {
|
||||
@ -1720,16 +1719,16 @@ static void btrfs_raid_unplug(struct blk_plug_cb *cb, bool from_schedule)
|
||||
* our main entry point for writes from the rest of the FS.
|
||||
*/
|
||||
int raid56_parity_write(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len)
|
||||
struct btrfs_io_context *bioc, u64 stripe_len)
|
||||
{
|
||||
struct btrfs_raid_bio *rbio;
|
||||
struct btrfs_plug_cb *plug = NULL;
|
||||
struct blk_plug_cb *cb;
|
||||
int ret;
|
||||
|
||||
rbio = alloc_rbio(fs_info, bbio, stripe_len);
|
||||
rbio = alloc_rbio(fs_info, bioc, stripe_len);
|
||||
if (IS_ERR(rbio)) {
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
return PTR_ERR(rbio);
|
||||
}
|
||||
bio_list_add(&rbio->bio_list, bio);
|
||||
@ -1842,7 +1841,7 @@ static void __raid_recover_end_io(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
|
||||
/* all raid6 handling here */
|
||||
if (rbio->bbio->map_type & BTRFS_BLOCK_GROUP_RAID6) {
|
||||
if (rbio->bioc->map_type & BTRFS_BLOCK_GROUP_RAID6) {
|
||||
/*
|
||||
* single failure, rebuild from parity raid5
|
||||
* style
|
||||
@ -1874,8 +1873,8 @@ static void __raid_recover_end_io(struct btrfs_raid_bio *rbio)
|
||||
* here due to a crc mismatch and we can't give them the
|
||||
* data they want
|
||||
*/
|
||||
if (rbio->bbio->raid_map[failb] == RAID6_Q_STRIPE) {
|
||||
if (rbio->bbio->raid_map[faila] ==
|
||||
if (rbio->bioc->raid_map[failb] == RAID6_Q_STRIPE) {
|
||||
if (rbio->bioc->raid_map[faila] ==
|
||||
RAID5_P_STRIPE) {
|
||||
err = BLK_STS_IOERR;
|
||||
goto cleanup;
|
||||
@ -1887,7 +1886,7 @@ static void __raid_recover_end_io(struct btrfs_raid_bio *rbio)
|
||||
goto pstripe;
|
||||
}
|
||||
|
||||
if (rbio->bbio->raid_map[failb] == RAID5_P_STRIPE) {
|
||||
if (rbio->bioc->raid_map[failb] == RAID5_P_STRIPE) {
|
||||
raid6_datap_recov(rbio->real_stripes,
|
||||
PAGE_SIZE, faila, pointers);
|
||||
} else {
|
||||
@ -2006,7 +2005,7 @@ static void raid_recover_end_io(struct bio *bio)
|
||||
if (!atomic_dec_and_test(&rbio->stripes_pending))
|
||||
return;
|
||||
|
||||
if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
|
||||
if (atomic_read(&rbio->error) > rbio->bioc->max_errors)
|
||||
rbio_orig_end_io(rbio, BLK_STS_IOERR);
|
||||
else
|
||||
__raid_recover_end_io(rbio);
|
||||
@ -2074,7 +2073,7 @@ static int __raid56_parity_recover(struct btrfs_raid_bio *rbio)
|
||||
* were up to date, or we might have no bios to read because
|
||||
* the devices were gone.
|
||||
*/
|
||||
if (atomic_read(&rbio->error) <= rbio->bbio->max_errors) {
|
||||
if (atomic_read(&rbio->error) <= rbio->bioc->max_errors) {
|
||||
__raid_recover_end_io(rbio);
|
||||
return 0;
|
||||
} else {
|
||||
@ -2083,8 +2082,8 @@ static int __raid56_parity_recover(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
|
||||
/*
|
||||
* the bbio may be freed once we submit the last bio. Make sure
|
||||
* not to touch it after that
|
||||
* The bioc may be freed once we submit the last bio. Make sure not to
|
||||
* touch it after that.
|
||||
*/
|
||||
atomic_set(&rbio->stripes_pending, bios_to_read);
|
||||
while ((bio = bio_list_pop(&bio_list))) {
|
||||
@ -2117,21 +2116,21 @@ cleanup:
|
||||
* of the drive.
|
||||
*/
|
||||
int raid56_parity_recover(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len,
|
||||
struct btrfs_io_context *bioc, u64 stripe_len,
|
||||
int mirror_num, int generic_io)
|
||||
{
|
||||
struct btrfs_raid_bio *rbio;
|
||||
int ret;
|
||||
|
||||
if (generic_io) {
|
||||
ASSERT(bbio->mirror_num == mirror_num);
|
||||
ASSERT(bioc->mirror_num == mirror_num);
|
||||
btrfs_io_bio(bio)->mirror_num = mirror_num;
|
||||
}
|
||||
|
||||
rbio = alloc_rbio(fs_info, bbio, stripe_len);
|
||||
rbio = alloc_rbio(fs_info, bioc, stripe_len);
|
||||
if (IS_ERR(rbio)) {
|
||||
if (generic_io)
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
return PTR_ERR(rbio);
|
||||
}
|
||||
|
||||
@ -2142,11 +2141,11 @@ int raid56_parity_recover(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
rbio->faila = find_logical_bio_stripe(rbio, bio);
|
||||
if (rbio->faila == -1) {
|
||||
btrfs_warn(fs_info,
|
||||
"%s could not find the bad stripe in raid56 so that we cannot recover any more (bio has logical %llu len %llu, bbio has map_type %llu)",
|
||||
"%s could not find the bad stripe in raid56 so that we cannot recover any more (bio has logical %llu len %llu, bioc has map_type %llu)",
|
||||
__func__, bio->bi_iter.bi_sector << 9,
|
||||
(u64)bio->bi_iter.bi_size, bbio->map_type);
|
||||
(u64)bio->bi_iter.bi_size, bioc->map_type);
|
||||
if (generic_io)
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
kfree(rbio);
|
||||
return -EIO;
|
||||
}
|
||||
@ -2155,7 +2154,7 @@ int raid56_parity_recover(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
btrfs_bio_counter_inc_noblocked(fs_info);
|
||||
rbio->generic_bio_cnt = 1;
|
||||
} else {
|
||||
btrfs_get_bbio(bbio);
|
||||
btrfs_get_bioc(bioc);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -2214,7 +2213,7 @@ static void read_rebuild_work(struct btrfs_work *work)
|
||||
/*
|
||||
* The following code is used to scrub/replace the parity stripe
|
||||
*
|
||||
* Caller must have already increased bio_counter for getting @bbio.
|
||||
* Caller must have already increased bio_counter for getting @bioc.
|
||||
*
|
||||
* Note: We need make sure all the pages that add into the scrub/replace
|
||||
* raid bio are correct and not be changed during the scrub/replace. That
|
||||
@ -2223,14 +2222,14 @@ static void read_rebuild_work(struct btrfs_work *work)
|
||||
|
||||
struct btrfs_raid_bio *
|
||||
raid56_parity_alloc_scrub_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len,
|
||||
struct btrfs_io_context *bioc, u64 stripe_len,
|
||||
struct btrfs_device *scrub_dev,
|
||||
unsigned long *dbitmap, int stripe_nsectors)
|
||||
{
|
||||
struct btrfs_raid_bio *rbio;
|
||||
int i;
|
||||
|
||||
rbio = alloc_rbio(fs_info, bbio, stripe_len);
|
||||
rbio = alloc_rbio(fs_info, bioc, stripe_len);
|
||||
if (IS_ERR(rbio))
|
||||
return NULL;
|
||||
bio_list_add(&rbio->bio_list, bio);
|
||||
@ -2242,12 +2241,12 @@ raid56_parity_alloc_scrub_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
rbio->operation = BTRFS_RBIO_PARITY_SCRUB;
|
||||
|
||||
/*
|
||||
* After mapping bbio with BTRFS_MAP_WRITE, parities have been sorted
|
||||
* After mapping bioc with BTRFS_MAP_WRITE, parities have been sorted
|
||||
* to the end position, so this search can start from the first parity
|
||||
* stripe.
|
||||
*/
|
||||
for (i = rbio->nr_data; i < rbio->real_stripes; i++) {
|
||||
if (bbio->stripes[i].dev == scrub_dev) {
|
||||
if (bioc->stripes[i].dev == scrub_dev) {
|
||||
rbio->scrubp = i;
|
||||
break;
|
||||
}
|
||||
@ -2260,7 +2259,7 @@ raid56_parity_alloc_scrub_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
bitmap_copy(rbio->dbitmap, dbitmap, stripe_nsectors);
|
||||
|
||||
/*
|
||||
* We have already increased bio_counter when getting bbio, record it
|
||||
* We have already increased bio_counter when getting bioc, record it
|
||||
* so we can free it at rbio_orig_end_io().
|
||||
*/
|
||||
rbio->generic_bio_cnt = 1;
|
||||
@ -2275,10 +2274,10 @@ void raid56_add_scrub_pages(struct btrfs_raid_bio *rbio, struct page *page,
|
||||
int stripe_offset;
|
||||
int index;
|
||||
|
||||
ASSERT(logical >= rbio->bbio->raid_map[0]);
|
||||
ASSERT(logical + PAGE_SIZE <= rbio->bbio->raid_map[0] +
|
||||
ASSERT(logical >= rbio->bioc->raid_map[0]);
|
||||
ASSERT(logical + PAGE_SIZE <= rbio->bioc->raid_map[0] +
|
||||
rbio->stripe_len * rbio->nr_data);
|
||||
stripe_offset = (int)(logical - rbio->bbio->raid_map[0]);
|
||||
stripe_offset = (int)(logical - rbio->bioc->raid_map[0]);
|
||||
index = stripe_offset >> PAGE_SHIFT;
|
||||
rbio->bio_pages[index] = page;
|
||||
}
|
||||
@ -2312,7 +2311,7 @@ static int alloc_rbio_essential_pages(struct btrfs_raid_bio *rbio)
|
||||
static noinline void finish_parity_scrub(struct btrfs_raid_bio *rbio,
|
||||
int need_check)
|
||||
{
|
||||
struct btrfs_bio *bbio = rbio->bbio;
|
||||
struct btrfs_io_context *bioc = rbio->bioc;
|
||||
void **pointers = rbio->finish_pointers;
|
||||
unsigned long *pbitmap = rbio->finish_pbitmap;
|
||||
int nr_data = rbio->nr_data;
|
||||
@ -2335,7 +2334,7 @@ static noinline void finish_parity_scrub(struct btrfs_raid_bio *rbio,
|
||||
else
|
||||
BUG();
|
||||
|
||||
if (bbio->num_tgtdevs && bbio->tgtdev_map[rbio->scrubp]) {
|
||||
if (bioc->num_tgtdevs && bioc->tgtdev_map[rbio->scrubp]) {
|
||||
is_replace = 1;
|
||||
bitmap_copy(pbitmap, rbio->dbitmap, rbio->stripe_npages);
|
||||
}
|
||||
@ -2435,7 +2434,7 @@ writeback:
|
||||
|
||||
page = rbio_stripe_page(rbio, rbio->scrubp, pagenr);
|
||||
ret = rbio_add_io_page(rbio, &bio_list, page,
|
||||
bbio->tgtdev_map[rbio->scrubp],
|
||||
bioc->tgtdev_map[rbio->scrubp],
|
||||
pagenr, rbio->stripe_len);
|
||||
if (ret)
|
||||
goto cleanup;
|
||||
@ -2483,7 +2482,7 @@ static inline int is_data_stripe(struct btrfs_raid_bio *rbio, int stripe)
|
||||
*/
|
||||
static void validate_rbio_for_parity_scrub(struct btrfs_raid_bio *rbio)
|
||||
{
|
||||
if (atomic_read(&rbio->error) > rbio->bbio->max_errors)
|
||||
if (atomic_read(&rbio->error) > rbio->bioc->max_errors)
|
||||
goto cleanup;
|
||||
|
||||
if (rbio->faila >= 0 || rbio->failb >= 0) {
|
||||
@ -2504,7 +2503,7 @@ static void validate_rbio_for_parity_scrub(struct btrfs_raid_bio *rbio)
|
||||
* the data, so the capability of the repair is declined.
|
||||
* (In the case of RAID5, we can not repair anything)
|
||||
*/
|
||||
if (dfail > rbio->bbio->max_errors - 1)
|
||||
if (dfail > rbio->bioc->max_errors - 1)
|
||||
goto cleanup;
|
||||
|
||||
/*
|
||||
@ -2625,8 +2624,8 @@ static void raid56_parity_scrub_stripe(struct btrfs_raid_bio *rbio)
|
||||
}
|
||||
|
||||
/*
|
||||
* the bbio may be freed once we submit the last bio. Make sure
|
||||
* not to touch it after that
|
||||
* The bioc may be freed once we submit the last bio. Make sure not to
|
||||
* touch it after that.
|
||||
*/
|
||||
atomic_set(&rbio->stripes_pending, bios_to_read);
|
||||
while ((bio = bio_list_pop(&bio_list))) {
|
||||
@ -2671,11 +2670,11 @@ void raid56_parity_submit_scrub_rbio(struct btrfs_raid_bio *rbio)
|
||||
|
||||
struct btrfs_raid_bio *
|
||||
raid56_alloc_missing_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 length)
|
||||
struct btrfs_io_context *bioc, u64 length)
|
||||
{
|
||||
struct btrfs_raid_bio *rbio;
|
||||
|
||||
rbio = alloc_rbio(fs_info, bbio, length);
|
||||
rbio = alloc_rbio(fs_info, bioc, length);
|
||||
if (IS_ERR(rbio))
|
||||
return NULL;
|
||||
|
||||
@ -2695,7 +2694,7 @@ raid56_alloc_missing_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
}
|
||||
|
||||
/*
|
||||
* When we get bbio, we have already increased bio_counter, record it
|
||||
* When we get bioc, we have already increased bio_counter, record it
|
||||
* so we can free it at rbio_orig_end_io()
|
||||
*/
|
||||
rbio->generic_bio_cnt = 1;
|
||||
|
@ -31,24 +31,24 @@ struct btrfs_raid_bio;
|
||||
struct btrfs_device;
|
||||
|
||||
int raid56_parity_recover(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len,
|
||||
struct btrfs_io_context *bioc, u64 stripe_len,
|
||||
int mirror_num, int generic_io);
|
||||
int raid56_parity_write(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len);
|
||||
struct btrfs_io_context *bioc, u64 stripe_len);
|
||||
|
||||
void raid56_add_scrub_pages(struct btrfs_raid_bio *rbio, struct page *page,
|
||||
u64 logical);
|
||||
|
||||
struct btrfs_raid_bio *
|
||||
raid56_parity_alloc_scrub_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 stripe_len,
|
||||
struct btrfs_io_context *bioc, u64 stripe_len,
|
||||
struct btrfs_device *scrub_dev,
|
||||
unsigned long *dbitmap, int stripe_nsectors);
|
||||
void raid56_parity_submit_scrub_rbio(struct btrfs_raid_bio *rbio);
|
||||
|
||||
struct btrfs_raid_bio *
|
||||
raid56_alloc_missing_rbio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
struct btrfs_bio *bbio, u64 length);
|
||||
struct btrfs_io_context *bioc, u64 length);
|
||||
void raid56_submit_missing_rbio(struct btrfs_raid_bio *rbio);
|
||||
|
||||
int btrfs_alloc_stripe_hash_table(struct btrfs_fs_info *info);
|
||||
|
@ -227,7 +227,7 @@ start_machine:
|
||||
}
|
||||
|
||||
static struct reada_zone *reada_find_zone(struct btrfs_device *dev, u64 logical,
|
||||
struct btrfs_bio *bbio)
|
||||
struct btrfs_io_context *bioc)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = dev->fs_info;
|
||||
int ret;
|
||||
@ -275,11 +275,11 @@ static struct reada_zone *reada_find_zone(struct btrfs_device *dev, u64 logical,
|
||||
kref_init(&zone->refcnt);
|
||||
zone->elems = 0;
|
||||
zone->device = dev; /* our device always sits at index 0 */
|
||||
for (i = 0; i < bbio->num_stripes; ++i) {
|
||||
for (i = 0; i < bioc->num_stripes; ++i) {
|
||||
/* bounds have already been checked */
|
||||
zone->devs[i] = bbio->stripes[i].dev;
|
||||
zone->devs[i] = bioc->stripes[i].dev;
|
||||
}
|
||||
zone->ndevs = bbio->num_stripes;
|
||||
zone->ndevs = bioc->num_stripes;
|
||||
|
||||
spin_lock(&fs_info->reada_lock);
|
||||
ret = radix_tree_insert(&dev->reada_zones,
|
||||
@ -309,7 +309,7 @@ static struct reada_extent *reada_find_extent(struct btrfs_fs_info *fs_info,
|
||||
int ret;
|
||||
struct reada_extent *re = NULL;
|
||||
struct reada_extent *re_exist = NULL;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
struct btrfs_device *dev;
|
||||
struct btrfs_device *prev_dev;
|
||||
u64 length;
|
||||
@ -345,28 +345,28 @@ static struct reada_extent *reada_find_extent(struct btrfs_fs_info *fs_info,
|
||||
*/
|
||||
length = fs_info->nodesize;
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_GET_READ_MIRRORS, logical,
|
||||
&length, &bbio, 0);
|
||||
if (ret || !bbio || length < fs_info->nodesize)
|
||||
&length, &bioc, 0);
|
||||
if (ret || !bioc || length < fs_info->nodesize)
|
||||
goto error;
|
||||
|
||||
if (bbio->num_stripes > BTRFS_MAX_MIRRORS) {
|
||||
if (bioc->num_stripes > BTRFS_MAX_MIRRORS) {
|
||||
btrfs_err(fs_info,
|
||||
"readahead: more than %d copies not supported",
|
||||
BTRFS_MAX_MIRRORS);
|
||||
goto error;
|
||||
}
|
||||
|
||||
real_stripes = bbio->num_stripes - bbio->num_tgtdevs;
|
||||
real_stripes = bioc->num_stripes - bioc->num_tgtdevs;
|
||||
for (nzones = 0; nzones < real_stripes; ++nzones) {
|
||||
struct reada_zone *zone;
|
||||
|
||||
dev = bbio->stripes[nzones].dev;
|
||||
dev = bioc->stripes[nzones].dev;
|
||||
|
||||
/* cannot read ahead on missing device. */
|
||||
if (!dev->bdev)
|
||||
continue;
|
||||
|
||||
zone = reada_find_zone(dev, logical, bbio);
|
||||
zone = reada_find_zone(dev, logical, bioc);
|
||||
if (!zone)
|
||||
continue;
|
||||
|
||||
@ -464,7 +464,7 @@ static struct reada_extent *reada_find_extent(struct btrfs_fs_info *fs_info,
|
||||
if (!have_zone)
|
||||
goto error;
|
||||
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
return re;
|
||||
|
||||
error:
|
||||
@ -488,7 +488,7 @@ error:
|
||||
kref_put(&zone->refcnt, reada_zone_release);
|
||||
spin_unlock(&fs_info->reada_lock);
|
||||
}
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
kfree(re);
|
||||
return re_exist;
|
||||
}
|
||||
|
115
fs/btrfs/scrub.c
115
fs/btrfs/scrub.c
@ -57,7 +57,7 @@ struct scrub_ctx;
|
||||
|
||||
struct scrub_recover {
|
||||
refcount_t refs;
|
||||
struct btrfs_bio *bbio;
|
||||
struct btrfs_io_context *bioc;
|
||||
u64 map_length;
|
||||
};
|
||||
|
||||
@ -254,7 +254,7 @@ static void scrub_put_ctx(struct scrub_ctx *sctx);
|
||||
static inline int scrub_is_page_on_raid56(struct scrub_page *spage)
|
||||
{
|
||||
return spage->recover &&
|
||||
(spage->recover->bbio->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK);
|
||||
(spage->recover->bioc->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK);
|
||||
}
|
||||
|
||||
static void scrub_pending_bio_inc(struct scrub_ctx *sctx)
|
||||
@ -798,7 +798,7 @@ static inline void scrub_put_recover(struct btrfs_fs_info *fs_info,
|
||||
{
|
||||
if (refcount_dec_and_test(&recover->refs)) {
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
btrfs_put_bbio(recover->bbio);
|
||||
btrfs_put_bioc(recover->bioc);
|
||||
kfree(recover);
|
||||
}
|
||||
}
|
||||
@ -1027,8 +1027,7 @@ static int scrub_handle_errored_block(struct scrub_block *sblock_to_check)
|
||||
sblock_other = sblocks_for_recheck + mirror_index;
|
||||
} else {
|
||||
struct scrub_recover *r = sblock_bad->pagev[0]->recover;
|
||||
int max_allowed = r->bbio->num_stripes -
|
||||
r->bbio->num_tgtdevs;
|
||||
int max_allowed = r->bioc->num_stripes - r->bioc->num_tgtdevs;
|
||||
|
||||
if (mirror_index >= max_allowed)
|
||||
break;
|
||||
@ -1218,14 +1217,14 @@ out:
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline int scrub_nr_raid_mirrors(struct btrfs_bio *bbio)
|
||||
static inline int scrub_nr_raid_mirrors(struct btrfs_io_context *bioc)
|
||||
{
|
||||
if (bbio->map_type & BTRFS_BLOCK_GROUP_RAID5)
|
||||
if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID5)
|
||||
return 2;
|
||||
else if (bbio->map_type & BTRFS_BLOCK_GROUP_RAID6)
|
||||
else if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID6)
|
||||
return 3;
|
||||
else
|
||||
return (int)bbio->num_stripes;
|
||||
return (int)bioc->num_stripes;
|
||||
}
|
||||
|
||||
static inline void scrub_stripe_index_and_offset(u64 logical, u64 map_type,
|
||||
@ -1269,7 +1268,7 @@ static int scrub_setup_recheck_block(struct scrub_block *original_sblock,
|
||||
u64 flags = original_sblock->pagev[0]->flags;
|
||||
u64 have_csum = original_sblock->pagev[0]->have_csum;
|
||||
struct scrub_recover *recover;
|
||||
struct btrfs_bio *bbio;
|
||||
struct btrfs_io_context *bioc;
|
||||
u64 sublen;
|
||||
u64 mapped_length;
|
||||
u64 stripe_offset;
|
||||
@ -1288,7 +1287,7 @@ static int scrub_setup_recheck_block(struct scrub_block *original_sblock,
|
||||
while (length > 0) {
|
||||
sublen = min_t(u64, length, fs_info->sectorsize);
|
||||
mapped_length = sublen;
|
||||
bbio = NULL;
|
||||
bioc = NULL;
|
||||
|
||||
/*
|
||||
* With a length of sectorsize, each returned stripe represents
|
||||
@ -1296,27 +1295,27 @@ static int scrub_setup_recheck_block(struct scrub_block *original_sblock,
|
||||
*/
|
||||
btrfs_bio_counter_inc_blocked(fs_info);
|
||||
ret = btrfs_map_sblock(fs_info, BTRFS_MAP_GET_READ_MIRRORS,
|
||||
logical, &mapped_length, &bbio);
|
||||
if (ret || !bbio || mapped_length < sublen) {
|
||||
btrfs_put_bbio(bbio);
|
||||
logical, &mapped_length, &bioc);
|
||||
if (ret || !bioc || mapped_length < sublen) {
|
||||
btrfs_put_bioc(bioc);
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
return -EIO;
|
||||
}
|
||||
|
||||
recover = kzalloc(sizeof(struct scrub_recover), GFP_NOFS);
|
||||
if (!recover) {
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
refcount_set(&recover->refs, 1);
|
||||
recover->bbio = bbio;
|
||||
recover->bioc = bioc;
|
||||
recover->map_length = mapped_length;
|
||||
|
||||
BUG_ON(page_index >= SCRUB_MAX_PAGES_PER_BLOCK);
|
||||
|
||||
nmirrors = min(scrub_nr_raid_mirrors(bbio), BTRFS_MAX_MIRRORS);
|
||||
nmirrors = min(scrub_nr_raid_mirrors(bioc), BTRFS_MAX_MIRRORS);
|
||||
|
||||
for (mirror_index = 0; mirror_index < nmirrors;
|
||||
mirror_index++) {
|
||||
@ -1348,17 +1347,17 @@ leave_nomem:
|
||||
sctx->fs_info->csum_size);
|
||||
|
||||
scrub_stripe_index_and_offset(logical,
|
||||
bbio->map_type,
|
||||
bbio->raid_map,
|
||||
bioc->map_type,
|
||||
bioc->raid_map,
|
||||
mapped_length,
|
||||
bbio->num_stripes -
|
||||
bbio->num_tgtdevs,
|
||||
bioc->num_stripes -
|
||||
bioc->num_tgtdevs,
|
||||
mirror_index,
|
||||
&stripe_index,
|
||||
&stripe_offset);
|
||||
spage->physical = bbio->stripes[stripe_index].physical +
|
||||
spage->physical = bioc->stripes[stripe_index].physical +
|
||||
stripe_offset;
|
||||
spage->dev = bbio->stripes[stripe_index].dev;
|
||||
spage->dev = bioc->stripes[stripe_index].dev;
|
||||
|
||||
BUG_ON(page_index >= original_sblock->page_count);
|
||||
spage->physical_for_dev_replace =
|
||||
@ -1401,7 +1400,7 @@ static int scrub_submit_raid56_bio_wait(struct btrfs_fs_info *fs_info,
|
||||
bio->bi_end_io = scrub_bio_wait_endio;
|
||||
|
||||
mirror_num = spage->sblock->pagev[0]->mirror_num;
|
||||
ret = raid56_parity_recover(fs_info, bio, spage->recover->bbio,
|
||||
ret = raid56_parity_recover(fs_info, bio, spage->recover->bioc,
|
||||
spage->recover->map_length,
|
||||
mirror_num, 0);
|
||||
if (ret)
|
||||
@ -2203,7 +2202,7 @@ static void scrub_missing_raid56_pages(struct scrub_block *sblock)
|
||||
struct btrfs_fs_info *fs_info = sctx->fs_info;
|
||||
u64 length = sblock->page_count * PAGE_SIZE;
|
||||
u64 logical = sblock->pagev[0]->logical;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
struct bio *bio;
|
||||
struct btrfs_raid_bio *rbio;
|
||||
int ret;
|
||||
@ -2211,19 +2210,19 @@ static void scrub_missing_raid56_pages(struct scrub_block *sblock)
|
||||
|
||||
btrfs_bio_counter_inc_blocked(fs_info);
|
||||
ret = btrfs_map_sblock(fs_info, BTRFS_MAP_GET_READ_MIRRORS, logical,
|
||||
&length, &bbio);
|
||||
if (ret || !bbio || !bbio->raid_map)
|
||||
goto bbio_out;
|
||||
&length, &bioc);
|
||||
if (ret || !bioc || !bioc->raid_map)
|
||||
goto bioc_out;
|
||||
|
||||
if (WARN_ON(!sctx->is_dev_replace ||
|
||||
!(bbio->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK))) {
|
||||
!(bioc->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK))) {
|
||||
/*
|
||||
* We shouldn't be scrubbing a missing device. Even for dev
|
||||
* replace, we should only get here for RAID 5/6. We either
|
||||
* managed to mount something with no mirrors remaining or
|
||||
* there's a bug in scrub_remap_extent()/btrfs_map_block().
|
||||
*/
|
||||
goto bbio_out;
|
||||
goto bioc_out;
|
||||
}
|
||||
|
||||
bio = btrfs_io_bio_alloc(0);
|
||||
@ -2231,7 +2230,7 @@ static void scrub_missing_raid56_pages(struct scrub_block *sblock)
|
||||
bio->bi_private = sblock;
|
||||
bio->bi_end_io = scrub_missing_raid56_end_io;
|
||||
|
||||
rbio = raid56_alloc_missing_rbio(fs_info, bio, bbio, length);
|
||||
rbio = raid56_alloc_missing_rbio(fs_info, bio, bioc, length);
|
||||
if (!rbio)
|
||||
goto rbio_out;
|
||||
|
||||
@ -2249,9 +2248,9 @@ static void scrub_missing_raid56_pages(struct scrub_block *sblock)
|
||||
|
||||
rbio_out:
|
||||
bio_put(bio);
|
||||
bbio_out:
|
||||
bioc_out:
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
spin_lock(&sctx->stat_lock);
|
||||
sctx->stat.malloc_errors++;
|
||||
spin_unlock(&sctx->stat_lock);
|
||||
@ -2826,7 +2825,7 @@ static void scrub_parity_check_and_repair(struct scrub_parity *sparity)
|
||||
struct btrfs_fs_info *fs_info = sctx->fs_info;
|
||||
struct bio *bio;
|
||||
struct btrfs_raid_bio *rbio;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
u64 length;
|
||||
int ret;
|
||||
|
||||
@ -2838,16 +2837,16 @@ static void scrub_parity_check_and_repair(struct scrub_parity *sparity)
|
||||
|
||||
btrfs_bio_counter_inc_blocked(fs_info);
|
||||
ret = btrfs_map_sblock(fs_info, BTRFS_MAP_WRITE, sparity->logic_start,
|
||||
&length, &bbio);
|
||||
if (ret || !bbio || !bbio->raid_map)
|
||||
goto bbio_out;
|
||||
&length, &bioc);
|
||||
if (ret || !bioc || !bioc->raid_map)
|
||||
goto bioc_out;
|
||||
|
||||
bio = btrfs_io_bio_alloc(0);
|
||||
bio->bi_iter.bi_sector = sparity->logic_start >> 9;
|
||||
bio->bi_private = sparity;
|
||||
bio->bi_end_io = scrub_parity_bio_endio;
|
||||
|
||||
rbio = raid56_parity_alloc_scrub_rbio(fs_info, bio, bbio,
|
||||
rbio = raid56_parity_alloc_scrub_rbio(fs_info, bio, bioc,
|
||||
length, sparity->scrub_dev,
|
||||
sparity->dbitmap,
|
||||
sparity->nsectors);
|
||||
@ -2860,9 +2859,9 @@ static void scrub_parity_check_and_repair(struct scrub_parity *sparity)
|
||||
|
||||
rbio_out:
|
||||
bio_put(bio);
|
||||
bbio_out:
|
||||
bioc_out:
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
bitmap_or(sparity->ebitmap, sparity->ebitmap, sparity->dbitmap,
|
||||
sparity->nsectors);
|
||||
spin_lock(&sctx->stat_lock);
|
||||
@ -2901,7 +2900,7 @@ static noinline_for_stack int scrub_raid56_parity(struct scrub_ctx *sctx,
|
||||
struct btrfs_root *root = fs_info->extent_root;
|
||||
struct btrfs_root *csum_root = fs_info->csum_root;
|
||||
struct btrfs_extent_item *extent;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
u64 flags;
|
||||
int ret;
|
||||
int slot;
|
||||
@ -3044,22 +3043,22 @@ again:
|
||||
extent_len);
|
||||
|
||||
mapped_length = extent_len;
|
||||
bbio = NULL;
|
||||
bioc = NULL;
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_READ,
|
||||
extent_logical, &mapped_length, &bbio,
|
||||
extent_logical, &mapped_length, &bioc,
|
||||
0);
|
||||
if (!ret) {
|
||||
if (!bbio || mapped_length < extent_len)
|
||||
if (!bioc || mapped_length < extent_len)
|
||||
ret = -EIO;
|
||||
}
|
||||
if (ret) {
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
goto out;
|
||||
}
|
||||
extent_physical = bbio->stripes[0].physical;
|
||||
extent_mirror_num = bbio->mirror_num;
|
||||
extent_dev = bbio->stripes[0].dev;
|
||||
btrfs_put_bbio(bbio);
|
||||
extent_physical = bioc->stripes[0].physical;
|
||||
extent_mirror_num = bioc->mirror_num;
|
||||
extent_dev = bioc->stripes[0].dev;
|
||||
btrfs_put_bioc(bioc);
|
||||
|
||||
ret = btrfs_lookup_csums_range(csum_root,
|
||||
extent_logical,
|
||||
@ -4309,20 +4308,20 @@ static void scrub_remap_extent(struct btrfs_fs_info *fs_info,
|
||||
int *extent_mirror_num)
|
||||
{
|
||||
u64 mapped_length;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
int ret;
|
||||
|
||||
mapped_length = extent_len;
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_READ, extent_logical,
|
||||
&mapped_length, &bbio, 0);
|
||||
if (ret || !bbio || mapped_length < extent_len ||
|
||||
!bbio->stripes[0].dev->bdev) {
|
||||
btrfs_put_bbio(bbio);
|
||||
&mapped_length, &bioc, 0);
|
||||
if (ret || !bioc || mapped_length < extent_len ||
|
||||
!bioc->stripes[0].dev->bdev) {
|
||||
btrfs_put_bioc(bioc);
|
||||
return;
|
||||
}
|
||||
|
||||
*extent_physical = bbio->stripes[0].physical;
|
||||
*extent_mirror_num = bbio->mirror_num;
|
||||
*extent_dev = bbio->stripes[0].dev;
|
||||
btrfs_put_bbio(bbio);
|
||||
*extent_physical = bioc->stripes[0].physical;
|
||||
*extent_mirror_num = bioc->mirror_num;
|
||||
*extent_dev = bioc->stripes[0].dev;
|
||||
btrfs_put_bioc(bioc);
|
||||
}
|
||||
|
@ -250,7 +250,7 @@ static void btrfs_dev_stat_print_on_load(struct btrfs_device *device);
|
||||
static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret,
|
||||
struct btrfs_io_context **bioc_ret,
|
||||
int mirror_num, int need_raid_map);
|
||||
|
||||
/*
|
||||
@ -5789,7 +5789,7 @@ static int find_live_mirror(struct btrfs_fs_info *fs_info,
|
||||
}
|
||||
|
||||
/* Bubble-sort the stripe set to put the parity/syndrome stripes last */
|
||||
static void sort_parity_stripes(struct btrfs_bio *bbio, int num_stripes)
|
||||
static void sort_parity_stripes(struct btrfs_io_context *bioc, int num_stripes)
|
||||
{
|
||||
int i;
|
||||
int again = 1;
|
||||
@ -5798,52 +5798,53 @@ static void sort_parity_stripes(struct btrfs_bio *bbio, int num_stripes)
|
||||
again = 0;
|
||||
for (i = 0; i < num_stripes - 1; i++) {
|
||||
/* Swap if parity is on a smaller index */
|
||||
if (bbio->raid_map[i] > bbio->raid_map[i + 1]) {
|
||||
swap(bbio->stripes[i], bbio->stripes[i + 1]);
|
||||
swap(bbio->raid_map[i], bbio->raid_map[i + 1]);
|
||||
if (bioc->raid_map[i] > bioc->raid_map[i + 1]) {
|
||||
swap(bioc->stripes[i], bioc->stripes[i + 1]);
|
||||
swap(bioc->raid_map[i], bioc->raid_map[i + 1]);
|
||||
again = 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static struct btrfs_bio *alloc_btrfs_bio(int total_stripes, int real_stripes)
|
||||
static struct btrfs_io_context *alloc_btrfs_io_context(int total_stripes,
|
||||
int real_stripes)
|
||||
{
|
||||
struct btrfs_bio *bbio = kzalloc(
|
||||
/* the size of the btrfs_bio */
|
||||
sizeof(struct btrfs_bio) +
|
||||
/* plus the variable array for the stripes */
|
||||
sizeof(struct btrfs_bio_stripe) * (total_stripes) +
|
||||
/* plus the variable array for the tgt dev */
|
||||
struct btrfs_io_context *bioc = kzalloc(
|
||||
/* The size of btrfs_io_context */
|
||||
sizeof(struct btrfs_io_context) +
|
||||
/* Plus the variable array for the stripes */
|
||||
sizeof(struct btrfs_io_stripe) * (total_stripes) +
|
||||
/* Plus the variable array for the tgt dev */
|
||||
sizeof(int) * (real_stripes) +
|
||||
/*
|
||||
* plus the raid_map, which includes both the tgt dev
|
||||
* and the stripes
|
||||
* Plus the raid_map, which includes both the tgt dev
|
||||
* and the stripes.
|
||||
*/
|
||||
sizeof(u64) * (total_stripes),
|
||||
GFP_NOFS|__GFP_NOFAIL);
|
||||
|
||||
atomic_set(&bbio->error, 0);
|
||||
refcount_set(&bbio->refs, 1);
|
||||
atomic_set(&bioc->error, 0);
|
||||
refcount_set(&bioc->refs, 1);
|
||||
|
||||
bbio->tgtdev_map = (int *)(bbio->stripes + total_stripes);
|
||||
bbio->raid_map = (u64 *)(bbio->tgtdev_map + real_stripes);
|
||||
bioc->tgtdev_map = (int *)(bioc->stripes + total_stripes);
|
||||
bioc->raid_map = (u64 *)(bioc->tgtdev_map + real_stripes);
|
||||
|
||||
return bbio;
|
||||
return bioc;
|
||||
}
|
||||
|
||||
void btrfs_get_bbio(struct btrfs_bio *bbio)
|
||||
void btrfs_get_bioc(struct btrfs_io_context *bioc)
|
||||
{
|
||||
WARN_ON(!refcount_read(&bbio->refs));
|
||||
refcount_inc(&bbio->refs);
|
||||
WARN_ON(!refcount_read(&bioc->refs));
|
||||
refcount_inc(&bioc->refs);
|
||||
}
|
||||
|
||||
void btrfs_put_bbio(struct btrfs_bio *bbio)
|
||||
void btrfs_put_bioc(struct btrfs_io_context *bioc)
|
||||
{
|
||||
if (!bbio)
|
||||
if (!bioc)
|
||||
return;
|
||||
if (refcount_dec_and_test(&bbio->refs))
|
||||
kfree(bbio);
|
||||
if (refcount_dec_and_test(&bioc->refs))
|
||||
kfree(bioc);
|
||||
}
|
||||
|
||||
/* can REQ_OP_DISCARD be sent with other REQ like REQ_OP_WRITE? */
|
||||
@ -5853,11 +5854,11 @@ void btrfs_put_bbio(struct btrfs_bio *bbio)
|
||||
*/
|
||||
static int __btrfs_map_block_for_discard(struct btrfs_fs_info *fs_info,
|
||||
u64 logical, u64 *length_ret,
|
||||
struct btrfs_bio **bbio_ret)
|
||||
struct btrfs_io_context **bioc_ret)
|
||||
{
|
||||
struct extent_map *em;
|
||||
struct map_lookup *map;
|
||||
struct btrfs_bio *bbio;
|
||||
struct btrfs_io_context *bioc;
|
||||
u64 length = *length_ret;
|
||||
u64 offset;
|
||||
u64 stripe_nr;
|
||||
@ -5876,8 +5877,8 @@ static int __btrfs_map_block_for_discard(struct btrfs_fs_info *fs_info,
|
||||
int ret = 0;
|
||||
int i;
|
||||
|
||||
/* discard always return a bbio */
|
||||
ASSERT(bbio_ret);
|
||||
/* Discard always returns a bioc. */
|
||||
ASSERT(bioc_ret);
|
||||
|
||||
em = btrfs_get_chunk_map(fs_info, logical, length);
|
||||
if (IS_ERR(em))
|
||||
@ -5940,26 +5941,25 @@ static int __btrfs_map_block_for_discard(struct btrfs_fs_info *fs_info,
|
||||
&stripe_index);
|
||||
}
|
||||
|
||||
bbio = alloc_btrfs_bio(num_stripes, 0);
|
||||
if (!bbio) {
|
||||
bioc = alloc_btrfs_io_context(num_stripes, 0);
|
||||
if (!bioc) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
for (i = 0; i < num_stripes; i++) {
|
||||
bbio->stripes[i].physical =
|
||||
bioc->stripes[i].physical =
|
||||
map->stripes[stripe_index].physical +
|
||||
stripe_offset + stripe_nr * map->stripe_len;
|
||||
bbio->stripes[i].dev = map->stripes[stripe_index].dev;
|
||||
bioc->stripes[i].dev = map->stripes[stripe_index].dev;
|
||||
|
||||
if (map->type & (BTRFS_BLOCK_GROUP_RAID0 |
|
||||
BTRFS_BLOCK_GROUP_RAID10)) {
|
||||
bbio->stripes[i].length = stripes_per_dev *
|
||||
bioc->stripes[i].length = stripes_per_dev *
|
||||
map->stripe_len;
|
||||
|
||||
if (i / sub_stripes < remaining_stripes)
|
||||
bbio->stripes[i].length +=
|
||||
map->stripe_len;
|
||||
bioc->stripes[i].length += map->stripe_len;
|
||||
|
||||
/*
|
||||
* Special for the first stripe and
|
||||
@ -5970,19 +5970,17 @@ static int __btrfs_map_block_for_discard(struct btrfs_fs_info *fs_info,
|
||||
* off end_off
|
||||
*/
|
||||
if (i < sub_stripes)
|
||||
bbio->stripes[i].length -=
|
||||
stripe_offset;
|
||||
bioc->stripes[i].length -= stripe_offset;
|
||||
|
||||
if (stripe_index >= last_stripe &&
|
||||
stripe_index <= (last_stripe +
|
||||
sub_stripes - 1))
|
||||
bbio->stripes[i].length -=
|
||||
stripe_end_offset;
|
||||
bioc->stripes[i].length -= stripe_end_offset;
|
||||
|
||||
if (i == sub_stripes - 1)
|
||||
stripe_offset = 0;
|
||||
} else {
|
||||
bbio->stripes[i].length = length;
|
||||
bioc->stripes[i].length = length;
|
||||
}
|
||||
|
||||
stripe_index++;
|
||||
@ -5992,9 +5990,9 @@ static int __btrfs_map_block_for_discard(struct btrfs_fs_info *fs_info,
|
||||
}
|
||||
}
|
||||
|
||||
*bbio_ret = bbio;
|
||||
bbio->map_type = map->type;
|
||||
bbio->num_stripes = num_stripes;
|
||||
*bioc_ret = bioc;
|
||||
bioc->map_type = map->type;
|
||||
bioc->num_stripes = num_stripes;
|
||||
out:
|
||||
free_extent_map(em);
|
||||
return ret;
|
||||
@ -6018,7 +6016,7 @@ static int get_extra_mirror_from_replace(struct btrfs_fs_info *fs_info,
|
||||
u64 srcdev_devid, int *mirror_num,
|
||||
u64 *physical)
|
||||
{
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
int num_stripes;
|
||||
int index_srcdev = 0;
|
||||
int found = 0;
|
||||
@ -6027,20 +6025,20 @@ static int get_extra_mirror_from_replace(struct btrfs_fs_info *fs_info,
|
||||
int ret = 0;
|
||||
|
||||
ret = __btrfs_map_block(fs_info, BTRFS_MAP_GET_READ_MIRRORS,
|
||||
logical, &length, &bbio, 0, 0);
|
||||
logical, &length, &bioc, 0, 0);
|
||||
if (ret) {
|
||||
ASSERT(bbio == NULL);
|
||||
ASSERT(bioc == NULL);
|
||||
return ret;
|
||||
}
|
||||
|
||||
num_stripes = bbio->num_stripes;
|
||||
num_stripes = bioc->num_stripes;
|
||||
if (*mirror_num > num_stripes) {
|
||||
/*
|
||||
* BTRFS_MAP_GET_READ_MIRRORS does not contain this mirror,
|
||||
* that means that the requested area is not left of the left
|
||||
* cursor
|
||||
*/
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
return -EIO;
|
||||
}
|
||||
|
||||
@ -6050,7 +6048,7 @@ static int get_extra_mirror_from_replace(struct btrfs_fs_info *fs_info,
|
||||
* pointer to the one of the target drive.
|
||||
*/
|
||||
for (i = 0; i < num_stripes; i++) {
|
||||
if (bbio->stripes[i].dev->devid != srcdev_devid)
|
||||
if (bioc->stripes[i].dev->devid != srcdev_devid)
|
||||
continue;
|
||||
|
||||
/*
|
||||
@ -6058,15 +6056,15 @@ static int get_extra_mirror_from_replace(struct btrfs_fs_info *fs_info,
|
||||
* mirror with the lowest physical address
|
||||
*/
|
||||
if (found &&
|
||||
physical_of_found <= bbio->stripes[i].physical)
|
||||
physical_of_found <= bioc->stripes[i].physical)
|
||||
continue;
|
||||
|
||||
index_srcdev = i;
|
||||
found = 1;
|
||||
physical_of_found = bbio->stripes[i].physical;
|
||||
physical_of_found = bioc->stripes[i].physical;
|
||||
}
|
||||
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
|
||||
ASSERT(found);
|
||||
if (!found)
|
||||
@ -6097,12 +6095,12 @@ static bool is_block_group_to_copy(struct btrfs_fs_info *fs_info, u64 logical)
|
||||
}
|
||||
|
||||
static void handle_ops_on_dev_replace(enum btrfs_map_op op,
|
||||
struct btrfs_bio **bbio_ret,
|
||||
struct btrfs_io_context **bioc_ret,
|
||||
struct btrfs_dev_replace *dev_replace,
|
||||
u64 logical,
|
||||
int *num_stripes_ret, int *max_errors_ret)
|
||||
{
|
||||
struct btrfs_bio *bbio = *bbio_ret;
|
||||
struct btrfs_io_context *bioc = *bioc_ret;
|
||||
u64 srcdev_devid = dev_replace->srcdev->devid;
|
||||
int tgtdev_indexes = 0;
|
||||
int num_stripes = *num_stripes_ret;
|
||||
@ -6132,17 +6130,17 @@ static void handle_ops_on_dev_replace(enum btrfs_map_op op,
|
||||
*/
|
||||
index_where_to_add = num_stripes;
|
||||
for (i = 0; i < num_stripes; i++) {
|
||||
if (bbio->stripes[i].dev->devid == srcdev_devid) {
|
||||
if (bioc->stripes[i].dev->devid == srcdev_devid) {
|
||||
/* write to new disk, too */
|
||||
struct btrfs_bio_stripe *new =
|
||||
bbio->stripes + index_where_to_add;
|
||||
struct btrfs_bio_stripe *old =
|
||||
bbio->stripes + i;
|
||||
struct btrfs_io_stripe *new =
|
||||
bioc->stripes + index_where_to_add;
|
||||
struct btrfs_io_stripe *old =
|
||||
bioc->stripes + i;
|
||||
|
||||
new->physical = old->physical;
|
||||
new->length = old->length;
|
||||
new->dev = dev_replace->tgtdev;
|
||||
bbio->tgtdev_map[i] = index_where_to_add;
|
||||
bioc->tgtdev_map[i] = index_where_to_add;
|
||||
index_where_to_add++;
|
||||
max_errors++;
|
||||
tgtdev_indexes++;
|
||||
@ -6162,30 +6160,29 @@ static void handle_ops_on_dev_replace(enum btrfs_map_op op,
|
||||
* full copy of the source drive.
|
||||
*/
|
||||
for (i = 0; i < num_stripes; i++) {
|
||||
if (bbio->stripes[i].dev->devid == srcdev_devid) {
|
||||
if (bioc->stripes[i].dev->devid == srcdev_devid) {
|
||||
/*
|
||||
* In case of DUP, in order to keep it simple,
|
||||
* only add the mirror with the lowest physical
|
||||
* address
|
||||
*/
|
||||
if (found &&
|
||||
physical_of_found <=
|
||||
bbio->stripes[i].physical)
|
||||
physical_of_found <= bioc->stripes[i].physical)
|
||||
continue;
|
||||
index_srcdev = i;
|
||||
found = 1;
|
||||
physical_of_found = bbio->stripes[i].physical;
|
||||
physical_of_found = bioc->stripes[i].physical;
|
||||
}
|
||||
}
|
||||
if (found) {
|
||||
struct btrfs_bio_stripe *tgtdev_stripe =
|
||||
bbio->stripes + num_stripes;
|
||||
struct btrfs_io_stripe *tgtdev_stripe =
|
||||
bioc->stripes + num_stripes;
|
||||
|
||||
tgtdev_stripe->physical = physical_of_found;
|
||||
tgtdev_stripe->length =
|
||||
bbio->stripes[index_srcdev].length;
|
||||
bioc->stripes[index_srcdev].length;
|
||||
tgtdev_stripe->dev = dev_replace->tgtdev;
|
||||
bbio->tgtdev_map[index_srcdev] = num_stripes;
|
||||
bioc->tgtdev_map[index_srcdev] = num_stripes;
|
||||
|
||||
tgtdev_indexes++;
|
||||
num_stripes++;
|
||||
@ -6194,8 +6191,8 @@ static void handle_ops_on_dev_replace(enum btrfs_map_op op,
|
||||
|
||||
*num_stripes_ret = num_stripes;
|
||||
*max_errors_ret = max_errors;
|
||||
bbio->num_tgtdevs = tgtdev_indexes;
|
||||
*bbio_ret = bbio;
|
||||
bioc->num_tgtdevs = tgtdev_indexes;
|
||||
*bioc_ret = bioc;
|
||||
}
|
||||
|
||||
static bool need_full_stripe(enum btrfs_map_op op)
|
||||
@ -6298,7 +6295,7 @@ int btrfs_get_io_geometry(struct btrfs_fs_info *fs_info, struct extent_map *em,
|
||||
static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret,
|
||||
struct btrfs_io_context **bioc_ret,
|
||||
int mirror_num, int need_raid_map)
|
||||
{
|
||||
struct extent_map *em;
|
||||
@ -6313,7 +6310,7 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
int num_stripes;
|
||||
int max_errors = 0;
|
||||
int tgtdev_indexes = 0;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
|
||||
int dev_replace_is_ongoing = 0;
|
||||
int num_alloc_stripes;
|
||||
@ -6322,7 +6319,7 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
u64 raid56_full_stripe_start = (u64)-1;
|
||||
struct btrfs_io_geometry geom;
|
||||
|
||||
ASSERT(bbio_ret);
|
||||
ASSERT(bioc_ret);
|
||||
ASSERT(op != BTRFS_MAP_DISCARD);
|
||||
|
||||
em = btrfs_get_chunk_map(fs_info, logical, *length);
|
||||
@ -6466,20 +6463,20 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
tgtdev_indexes = num_stripes;
|
||||
}
|
||||
|
||||
bbio = alloc_btrfs_bio(num_alloc_stripes, tgtdev_indexes);
|
||||
if (!bbio) {
|
||||
bioc = alloc_btrfs_io_context(num_alloc_stripes, tgtdev_indexes);
|
||||
if (!bioc) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
for (i = 0; i < num_stripes; i++) {
|
||||
bbio->stripes[i].physical = map->stripes[stripe_index].physical +
|
||||
bioc->stripes[i].physical = map->stripes[stripe_index].physical +
|
||||
stripe_offset + stripe_nr * map->stripe_len;
|
||||
bbio->stripes[i].dev = map->stripes[stripe_index].dev;
|
||||
bioc->stripes[i].dev = map->stripes[stripe_index].dev;
|
||||
stripe_index++;
|
||||
}
|
||||
|
||||
/* build raid_map */
|
||||
/* Build raid_map */
|
||||
if (map->type & BTRFS_BLOCK_GROUP_RAID56_MASK && need_raid_map &&
|
||||
(need_full_stripe(op) || mirror_num > 1)) {
|
||||
u64 tmp;
|
||||
@ -6491,15 +6488,15 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
/* Fill in the logical address of each stripe */
|
||||
tmp = stripe_nr * data_stripes;
|
||||
for (i = 0; i < data_stripes; i++)
|
||||
bbio->raid_map[(i+rot) % num_stripes] =
|
||||
bioc->raid_map[(i + rot) % num_stripes] =
|
||||
em->start + (tmp + i) * map->stripe_len;
|
||||
|
||||
bbio->raid_map[(i+rot) % map->num_stripes] = RAID5_P_STRIPE;
|
||||
bioc->raid_map[(i + rot) % map->num_stripes] = RAID5_P_STRIPE;
|
||||
if (map->type & BTRFS_BLOCK_GROUP_RAID6)
|
||||
bbio->raid_map[(i+rot+1) % num_stripes] =
|
||||
bioc->raid_map[(i + rot + 1) % num_stripes] =
|
||||
RAID6_Q_STRIPE;
|
||||
|
||||
sort_parity_stripes(bbio, num_stripes);
|
||||
sort_parity_stripes(bioc, num_stripes);
|
||||
}
|
||||
|
||||
if (need_full_stripe(op))
|
||||
@ -6507,15 +6504,15 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
|
||||
if (dev_replace_is_ongoing && dev_replace->tgtdev != NULL &&
|
||||
need_full_stripe(op)) {
|
||||
handle_ops_on_dev_replace(op, &bbio, dev_replace, logical,
|
||||
handle_ops_on_dev_replace(op, &bioc, dev_replace, logical,
|
||||
&num_stripes, &max_errors);
|
||||
}
|
||||
|
||||
*bbio_ret = bbio;
|
||||
bbio->map_type = map->type;
|
||||
bbio->num_stripes = num_stripes;
|
||||
bbio->max_errors = max_errors;
|
||||
bbio->mirror_num = mirror_num;
|
||||
*bioc_ret = bioc;
|
||||
bioc->map_type = map->type;
|
||||
bioc->num_stripes = num_stripes;
|
||||
bioc->max_errors = max_errors;
|
||||
bioc->mirror_num = mirror_num;
|
||||
|
||||
/*
|
||||
* this is the case that REQ_READ && dev_replace_is_ongoing &&
|
||||
@ -6524,9 +6521,9 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
*/
|
||||
if (patch_the_first_stripe_for_dev_replace && num_stripes > 0) {
|
||||
WARN_ON(num_stripes > 1);
|
||||
bbio->stripes[0].dev = dev_replace->tgtdev;
|
||||
bbio->stripes[0].physical = physical_to_patch_in_first_stripe;
|
||||
bbio->mirror_num = map->num_stripes + 1;
|
||||
bioc->stripes[0].dev = dev_replace->tgtdev;
|
||||
bioc->stripes[0].physical = physical_to_patch_in_first_stripe;
|
||||
bioc->mirror_num = map->num_stripes + 1;
|
||||
}
|
||||
out:
|
||||
if (dev_replace_is_ongoing) {
|
||||
@ -6540,40 +6537,40 @@ out:
|
||||
|
||||
int btrfs_map_block(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret, int mirror_num)
|
||||
struct btrfs_io_context **bioc_ret, int mirror_num)
|
||||
{
|
||||
if (op == BTRFS_MAP_DISCARD)
|
||||
return __btrfs_map_block_for_discard(fs_info, logical,
|
||||
length, bbio_ret);
|
||||
length, bioc_ret);
|
||||
|
||||
return __btrfs_map_block(fs_info, op, logical, length, bbio_ret,
|
||||
return __btrfs_map_block(fs_info, op, logical, length, bioc_ret,
|
||||
mirror_num, 0);
|
||||
}
|
||||
|
||||
/* For Scrub/replace */
|
||||
int btrfs_map_sblock(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret)
|
||||
struct btrfs_io_context **bioc_ret)
|
||||
{
|
||||
return __btrfs_map_block(fs_info, op, logical, length, bbio_ret, 0, 1);
|
||||
return __btrfs_map_block(fs_info, op, logical, length, bioc_ret, 0, 1);
|
||||
}
|
||||
|
||||
static inline void btrfs_end_bbio(struct btrfs_bio *bbio, struct bio *bio)
|
||||
static inline void btrfs_end_bioc(struct btrfs_io_context *bioc, struct bio *bio)
|
||||
{
|
||||
bio->bi_private = bbio->private;
|
||||
bio->bi_end_io = bbio->end_io;
|
||||
bio->bi_private = bioc->private;
|
||||
bio->bi_end_io = bioc->end_io;
|
||||
bio_endio(bio);
|
||||
|
||||
btrfs_put_bbio(bbio);
|
||||
btrfs_put_bioc(bioc);
|
||||
}
|
||||
|
||||
static void btrfs_end_bio(struct bio *bio)
|
||||
{
|
||||
struct btrfs_bio *bbio = bio->bi_private;
|
||||
struct btrfs_io_context *bioc = bio->bi_private;
|
||||
int is_orig_bio = 0;
|
||||
|
||||
if (bio->bi_status) {
|
||||
atomic_inc(&bbio->error);
|
||||
atomic_inc(&bioc->error);
|
||||
if (bio->bi_status == BLK_STS_IOERR ||
|
||||
bio->bi_status == BLK_STS_TARGET) {
|
||||
struct btrfs_device *dev = btrfs_io_bio(bio)->device;
|
||||
@ -6591,22 +6588,22 @@ static void btrfs_end_bio(struct bio *bio)
|
||||
}
|
||||
}
|
||||
|
||||
if (bio == bbio->orig_bio)
|
||||
if (bio == bioc->orig_bio)
|
||||
is_orig_bio = 1;
|
||||
|
||||
btrfs_bio_counter_dec(bbio->fs_info);
|
||||
btrfs_bio_counter_dec(bioc->fs_info);
|
||||
|
||||
if (atomic_dec_and_test(&bbio->stripes_pending)) {
|
||||
if (atomic_dec_and_test(&bioc->stripes_pending)) {
|
||||
if (!is_orig_bio) {
|
||||
bio_put(bio);
|
||||
bio = bbio->orig_bio;
|
||||
bio = bioc->orig_bio;
|
||||
}
|
||||
|
||||
btrfs_io_bio(bio)->mirror_num = bbio->mirror_num;
|
||||
btrfs_io_bio(bio)->mirror_num = bioc->mirror_num;
|
||||
/* only send an error to the higher layers if it is
|
||||
* beyond the tolerance of the btrfs bio
|
||||
*/
|
||||
if (atomic_read(&bbio->error) > bbio->max_errors) {
|
||||
if (atomic_read(&bioc->error) > bioc->max_errors) {
|
||||
bio->bi_status = BLK_STS_IOERR;
|
||||
} else {
|
||||
/*
|
||||
@ -6616,18 +6613,18 @@ static void btrfs_end_bio(struct bio *bio)
|
||||
bio->bi_status = BLK_STS_OK;
|
||||
}
|
||||
|
||||
btrfs_end_bbio(bbio, bio);
|
||||
btrfs_end_bioc(bioc, bio);
|
||||
} else if (!is_orig_bio) {
|
||||
bio_put(bio);
|
||||
}
|
||||
}
|
||||
|
||||
static void submit_stripe_bio(struct btrfs_bio *bbio, struct bio *bio,
|
||||
static void submit_stripe_bio(struct btrfs_io_context *bioc, struct bio *bio,
|
||||
u64 physical, struct btrfs_device *dev)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = bbio->fs_info;
|
||||
struct btrfs_fs_info *fs_info = bioc->fs_info;
|
||||
|
||||
bio->bi_private = bbio;
|
||||
bio->bi_private = bioc;
|
||||
btrfs_io_bio(bio)->device = dev;
|
||||
bio->bi_end_io = btrfs_end_bio;
|
||||
bio->bi_iter.bi_sector = physical >> 9;
|
||||
@ -6657,20 +6654,20 @@ static void submit_stripe_bio(struct btrfs_bio *bbio, struct bio *bio,
|
||||
btrfsic_submit_bio(bio);
|
||||
}
|
||||
|
||||
static void bbio_error(struct btrfs_bio *bbio, struct bio *bio, u64 logical)
|
||||
static void bioc_error(struct btrfs_io_context *bioc, struct bio *bio, u64 logical)
|
||||
{
|
||||
atomic_inc(&bbio->error);
|
||||
if (atomic_dec_and_test(&bbio->stripes_pending)) {
|
||||
atomic_inc(&bioc->error);
|
||||
if (atomic_dec_and_test(&bioc->stripes_pending)) {
|
||||
/* Should be the original bio. */
|
||||
WARN_ON(bio != bbio->orig_bio);
|
||||
WARN_ON(bio != bioc->orig_bio);
|
||||
|
||||
btrfs_io_bio(bio)->mirror_num = bbio->mirror_num;
|
||||
btrfs_io_bio(bio)->mirror_num = bioc->mirror_num;
|
||||
bio->bi_iter.bi_sector = logical >> 9;
|
||||
if (atomic_read(&bbio->error) > bbio->max_errors)
|
||||
if (atomic_read(&bioc->error) > bioc->max_errors)
|
||||
bio->bi_status = BLK_STS_IOERR;
|
||||
else
|
||||
bio->bi_status = BLK_STS_OK;
|
||||
btrfs_end_bbio(bbio, bio);
|
||||
btrfs_end_bioc(bioc, bio);
|
||||
}
|
||||
}
|
||||
|
||||
@ -6685,35 +6682,35 @@ blk_status_t btrfs_map_bio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
int ret;
|
||||
int dev_nr;
|
||||
int total_devs;
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
|
||||
length = bio->bi_iter.bi_size;
|
||||
map_length = length;
|
||||
|
||||
btrfs_bio_counter_inc_blocked(fs_info);
|
||||
ret = __btrfs_map_block(fs_info, btrfs_op(bio), logical,
|
||||
&map_length, &bbio, mirror_num, 1);
|
||||
&map_length, &bioc, mirror_num, 1);
|
||||
if (ret) {
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
return errno_to_blk_status(ret);
|
||||
}
|
||||
|
||||
total_devs = bbio->num_stripes;
|
||||
bbio->orig_bio = first_bio;
|
||||
bbio->private = first_bio->bi_private;
|
||||
bbio->end_io = first_bio->bi_end_io;
|
||||
bbio->fs_info = fs_info;
|
||||
atomic_set(&bbio->stripes_pending, bbio->num_stripes);
|
||||
total_devs = bioc->num_stripes;
|
||||
bioc->orig_bio = first_bio;
|
||||
bioc->private = first_bio->bi_private;
|
||||
bioc->end_io = first_bio->bi_end_io;
|
||||
bioc->fs_info = fs_info;
|
||||
atomic_set(&bioc->stripes_pending, bioc->num_stripes);
|
||||
|
||||
if ((bbio->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK) &&
|
||||
if ((bioc->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK) &&
|
||||
((btrfs_op(bio) == BTRFS_MAP_WRITE) || (mirror_num > 1))) {
|
||||
/* In this case, map_length has been set to the length of
|
||||
a single stripe; not the whole write */
|
||||
if (btrfs_op(bio) == BTRFS_MAP_WRITE) {
|
||||
ret = raid56_parity_write(fs_info, bio, bbio,
|
||||
ret = raid56_parity_write(fs_info, bio, bioc,
|
||||
map_length);
|
||||
} else {
|
||||
ret = raid56_parity_recover(fs_info, bio, bbio,
|
||||
ret = raid56_parity_recover(fs_info, bio, bioc,
|
||||
map_length, mirror_num, 1);
|
||||
}
|
||||
|
||||
@ -6729,12 +6726,12 @@ blk_status_t btrfs_map_bio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
}
|
||||
|
||||
for (dev_nr = 0; dev_nr < total_devs; dev_nr++) {
|
||||
dev = bbio->stripes[dev_nr].dev;
|
||||
dev = bioc->stripes[dev_nr].dev;
|
||||
if (!dev || !dev->bdev || test_bit(BTRFS_DEV_STATE_MISSING,
|
||||
&dev->dev_state) ||
|
||||
(btrfs_op(first_bio) == BTRFS_MAP_WRITE &&
|
||||
!test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))) {
|
||||
bbio_error(bbio, first_bio, logical);
|
||||
bioc_error(bioc, first_bio, logical);
|
||||
continue;
|
||||
}
|
||||
|
||||
@ -6743,7 +6740,7 @@ blk_status_t btrfs_map_bio(struct btrfs_fs_info *fs_info, struct bio *bio,
|
||||
else
|
||||
bio = first_bio;
|
||||
|
||||
submit_stripe_bio(bbio, bio, bbio->stripes[dev_nr].physical, dev);
|
||||
submit_stripe_bio(bioc, bio, bioc->stripes[dev_nr].physical, dev);
|
||||
}
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
return BLK_STS_OK;
|
||||
|
@ -306,11 +306,11 @@ struct btrfs_fs_devices {
|
||||
/*
|
||||
* we need the mirror number and stripe index to be passed around
|
||||
* the call chain while we are processing end_io (especially errors).
|
||||
* Really, what we need is a btrfs_bio structure that has this info
|
||||
* Really, what we need is a btrfs_io_context structure that has this info
|
||||
* and is properly sized with its stripe array, but we're not there
|
||||
* quite yet. We have our own btrfs bioset, and all of the bios
|
||||
* we allocate are actually btrfs_io_bios. We'll cram as much of
|
||||
* struct btrfs_bio as we can into this over time.
|
||||
* struct btrfs_io_context as we can into this over time.
|
||||
*/
|
||||
struct btrfs_io_bio {
|
||||
unsigned int mirror_num;
|
||||
@ -339,13 +339,29 @@ static inline void btrfs_io_bio_free_csum(struct btrfs_io_bio *io_bio)
|
||||
}
|
||||
}
|
||||
|
||||
struct btrfs_bio_stripe {
|
||||
struct btrfs_io_stripe {
|
||||
struct btrfs_device *dev;
|
||||
u64 physical;
|
||||
u64 length; /* only used for discard mappings */
|
||||
};
|
||||
|
||||
struct btrfs_bio {
|
||||
/*
|
||||
* Context for IO subsmission for device stripe.
|
||||
*
|
||||
* - Track the unfinished mirrors for mirror based profiles
|
||||
* Mirror based profiles are SINGLE/DUP/RAID1/RAID10.
|
||||
*
|
||||
* - Contain the logical -> physical mapping info
|
||||
* Used by submit_stripe_bio() for mapping logical bio
|
||||
* into physical device address.
|
||||
*
|
||||
* - Contain device replace info
|
||||
* Used by handle_ops_on_dev_replace() to copy logical bios
|
||||
* into the new device.
|
||||
*
|
||||
* - Contain RAID56 full stripe logical bytenrs
|
||||
*/
|
||||
struct btrfs_io_context {
|
||||
refcount_t refs;
|
||||
atomic_t stripes_pending;
|
||||
struct btrfs_fs_info *fs_info;
|
||||
@ -365,7 +381,7 @@ struct btrfs_bio {
|
||||
* so raid_map[0] is the start of our full stripe
|
||||
*/
|
||||
u64 *raid_map;
|
||||
struct btrfs_bio_stripe stripes[];
|
||||
struct btrfs_io_stripe stripes[];
|
||||
};
|
||||
|
||||
struct btrfs_device_info {
|
||||
@ -400,11 +416,11 @@ struct map_lookup {
|
||||
int num_stripes;
|
||||
int sub_stripes;
|
||||
int verified_stripes; /* For mount time dev extent verification */
|
||||
struct btrfs_bio_stripe stripes[];
|
||||
struct btrfs_io_stripe stripes[];
|
||||
};
|
||||
|
||||
#define map_lookup_size(n) (sizeof(struct map_lookup) + \
|
||||
(sizeof(struct btrfs_bio_stripe) * (n)))
|
||||
(sizeof(struct btrfs_io_stripe) * (n)))
|
||||
|
||||
struct btrfs_balance_args;
|
||||
struct btrfs_balance_progress;
|
||||
@ -441,14 +457,14 @@ static inline enum btrfs_map_op btrfs_op(struct bio *bio)
|
||||
}
|
||||
}
|
||||
|
||||
void btrfs_get_bbio(struct btrfs_bio *bbio);
|
||||
void btrfs_put_bbio(struct btrfs_bio *bbio);
|
||||
void btrfs_get_bioc(struct btrfs_io_context *bioc);
|
||||
void btrfs_put_bioc(struct btrfs_io_context *bioc);
|
||||
int btrfs_map_block(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret, int mirror_num);
|
||||
struct btrfs_io_context **bioc_ret, int mirror_num);
|
||||
int btrfs_map_sblock(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
|
||||
u64 logical, u64 *length,
|
||||
struct btrfs_bio **bbio_ret);
|
||||
struct btrfs_io_context **bioc_ret);
|
||||
int btrfs_get_io_geometry(struct btrfs_fs_info *fs_info, struct extent_map *map,
|
||||
enum btrfs_map_op op, u64 logical,
|
||||
struct btrfs_io_geometry *io_geom);
|
||||
|
@ -1637,27 +1637,27 @@ int btrfs_zoned_issue_zeroout(struct btrfs_device *device, u64 physical, u64 len
|
||||
static int read_zone_info(struct btrfs_fs_info *fs_info, u64 logical,
|
||||
struct blk_zone *zone)
|
||||
{
|
||||
struct btrfs_bio *bbio = NULL;
|
||||
struct btrfs_io_context *bioc = NULL;
|
||||
u64 mapped_length = PAGE_SIZE;
|
||||
unsigned int nofs_flag;
|
||||
int nmirrors;
|
||||
int i, ret;
|
||||
|
||||
ret = btrfs_map_sblock(fs_info, BTRFS_MAP_GET_READ_MIRRORS, logical,
|
||||
&mapped_length, &bbio);
|
||||
if (ret || !bbio || mapped_length < PAGE_SIZE) {
|
||||
btrfs_put_bbio(bbio);
|
||||
&mapped_length, &bioc);
|
||||
if (ret || !bioc || mapped_length < PAGE_SIZE) {
|
||||
btrfs_put_bioc(bioc);
|
||||
return -EIO;
|
||||
}
|
||||
|
||||
if (bbio->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK)
|
||||
if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK)
|
||||
return -EINVAL;
|
||||
|
||||
nofs_flag = memalloc_nofs_save();
|
||||
nmirrors = (int)bbio->num_stripes;
|
||||
nmirrors = (int)bioc->num_stripes;
|
||||
for (i = 0; i < nmirrors; i++) {
|
||||
u64 physical = bbio->stripes[i].physical;
|
||||
struct btrfs_device *dev = bbio->stripes[i].dev;
|
||||
u64 physical = bioc->stripes[i].physical;
|
||||
struct btrfs_device *dev = bioc->stripes[i].dev;
|
||||
|
||||
/* Missing device */
|
||||
if (!dev->bdev)
|
||||
|
Loading…
Reference in New Issue
Block a user