2017-03-17 14:18:50 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/*
|
|
|
|
* Some low level IO code, and hacks for various block layer limitations
|
|
|
|
*
|
|
|
|
* Copyright 2010, 2011 Kent Overstreet <kent.overstreet@gmail.com>
|
|
|
|
* Copyright 2012 Google, Inc.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "bcachefs.h"
|
2020-10-17 09:39:16 +08:00
|
|
|
#include "alloc_background.h"
|
2018-10-06 12:46:55 +08:00
|
|
|
#include "alloc_foreground.h"
|
2020-12-18 04:08:58 +08:00
|
|
|
#include "bkey_buf.h"
|
2017-03-17 14:18:50 +08:00
|
|
|
#include "bset.h"
|
|
|
|
#include "btree_update.h"
|
|
|
|
#include "buckets.h"
|
|
|
|
#include "checksum.h"
|
|
|
|
#include "compress.h"
|
|
|
|
#include "clock.h"
|
|
|
|
#include "debug.h"
|
|
|
|
#include "disk_groups.h"
|
2018-11-02 03:13:19 +08:00
|
|
|
#include "ec.h"
|
2017-03-17 14:18:50 +08:00
|
|
|
#include "error.h"
|
2019-11-16 04:52:28 +08:00
|
|
|
#include "extent_update.h"
|
2019-10-10 00:11:00 +08:00
|
|
|
#include "inode.h"
|
2017-03-17 14:18:50 +08:00
|
|
|
#include "io.h"
|
|
|
|
#include "journal.h"
|
|
|
|
#include "keylist.h"
|
|
|
|
#include "move.h"
|
|
|
|
#include "rebalance.h"
|
|
|
|
#include "super.h"
|
|
|
|
#include "super-io.h"
|
|
|
|
#include "trace.h"
|
|
|
|
|
|
|
|
#include <linux/blkdev.h>
|
|
|
|
#include <linux/random.h>
|
2020-07-21 01:00:15 +08:00
|
|
|
#include <linux/sched/mm.h>
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2020-07-22 01:34:22 +08:00
|
|
|
const char *bch2_blk_status_to_str(blk_status_t status)
|
|
|
|
{
|
|
|
|
if (status == BLK_STS_REMOVED)
|
|
|
|
return "device removed";
|
|
|
|
return blk_status_to_str(status);
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
#ifndef CONFIG_BCACHEFS_NO_LATENCY_ACCT
|
|
|
|
|
|
|
|
static bool bch2_target_congested(struct bch_fs *c, u16 target)
|
|
|
|
{
|
|
|
|
const struct bch_devs_mask *devs;
|
|
|
|
unsigned d, nr = 0, total = 0;
|
|
|
|
u64 now = local_clock(), last;
|
|
|
|
s64 congested;
|
|
|
|
struct bch_dev *ca;
|
|
|
|
|
|
|
|
if (!target)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
2020-08-07 03:22:24 +08:00
|
|
|
devs = bch2_target_to_mask(c, target) ?:
|
|
|
|
&c->rw_devs[BCH_DATA_user];
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
for_each_set_bit(d, devs->d, BCH_SB_MEMBERS_MAX) {
|
|
|
|
ca = rcu_dereference(c->devs[d]);
|
|
|
|
if (!ca)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
congested = atomic_read(&ca->congested);
|
|
|
|
last = READ_ONCE(ca->congested_last);
|
|
|
|
if (time_after64(now, last))
|
|
|
|
congested -= (now - last) >> 12;
|
|
|
|
|
|
|
|
total += max(congested, 0LL);
|
|
|
|
nr++;
|
|
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return bch2_rand_range(nr * CONGESTED_MAX) < total;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void bch2_congested_acct(struct bch_dev *ca, u64 io_latency,
|
|
|
|
u64 now, int rw)
|
|
|
|
{
|
|
|
|
u64 latency_capable =
|
|
|
|
ca->io_latency[rw].quantiles.entries[QUANTILE_IDX(1)].m;
|
|
|
|
/* ideally we'd be taking into account the device's variance here: */
|
|
|
|
u64 latency_threshold = latency_capable << (rw == READ ? 2 : 3);
|
|
|
|
s64 latency_over = io_latency - latency_threshold;
|
|
|
|
|
|
|
|
if (latency_threshold && latency_over > 0) {
|
|
|
|
/*
|
|
|
|
* bump up congested by approximately latency_over * 4 /
|
|
|
|
* latency_threshold - we don't need much accuracy here so don't
|
|
|
|
* bother with the divide:
|
|
|
|
*/
|
|
|
|
if (atomic_read(&ca->congested) < CONGESTED_MAX)
|
|
|
|
atomic_add(latency_over >>
|
|
|
|
max_t(int, ilog2(latency_threshold) - 2, 0),
|
|
|
|
&ca->congested);
|
|
|
|
|
|
|
|
ca->congested_last = now;
|
|
|
|
} else if (atomic_read(&ca->congested) > 0) {
|
|
|
|
atomic_dec(&ca->congested);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void bch2_latency_acct(struct bch_dev *ca, u64 submit_time, int rw)
|
|
|
|
{
|
|
|
|
atomic64_t *latency = &ca->cur_latency[rw];
|
|
|
|
u64 now = local_clock();
|
|
|
|
u64 io_latency = time_after64(now, submit_time)
|
|
|
|
? now - submit_time
|
|
|
|
: 0;
|
|
|
|
u64 old, new, v = atomic64_read(latency);
|
|
|
|
|
|
|
|
do {
|
|
|
|
old = v;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the io latency was reasonably close to the current
|
|
|
|
* latency, skip doing the update and atomic operation - most of
|
|
|
|
* the time:
|
|
|
|
*/
|
|
|
|
if (abs((int) (old - io_latency)) < (old >> 1) &&
|
|
|
|
now & ~(~0 << 5))
|
|
|
|
break;
|
|
|
|
|
|
|
|
new = ewma_add(old, io_latency, 5);
|
|
|
|
} while ((v = atomic64_cmpxchg(latency, old, new)) != old);
|
|
|
|
|
|
|
|
bch2_congested_acct(ca, io_latency, now, rw);
|
|
|
|
|
|
|
|
__bch2_time_stats_update(&ca->io_latency[rw], submit_time, now);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
|
|
|
static bool bch2_target_congested(struct bch_fs *c, u16 target)
|
|
|
|
{
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* Allocate, free from mempool: */
|
|
|
|
|
|
|
|
void bch2_bio_free_pages_pool(struct bch_fs *c, struct bio *bio)
|
|
|
|
{
|
|
|
|
struct bvec_iter_all iter;
|
|
|
|
struct bio_vec *bv;
|
|
|
|
|
|
|
|
bio_for_each_segment_all(bv, bio, iter)
|
|
|
|
if (bv->bv_page != ZERO_PAGE(0))
|
|
|
|
mempool_free(bv->bv_page, &c->bio_bounce_pages);
|
|
|
|
bio->bi_vcnt = 0;
|
|
|
|
}
|
|
|
|
|
2019-07-04 07:27:42 +08:00
|
|
|
static struct page *__bio_alloc_page_pool(struct bch_fs *c, bool *using_mempool)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
2019-07-04 07:27:42 +08:00
|
|
|
struct page *page;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (likely(!*using_mempool)) {
|
2019-07-04 07:27:42 +08:00
|
|
|
page = alloc_page(GFP_NOIO);
|
|
|
|
if (unlikely(!page)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
mutex_lock(&c->bio_bounce_pages_lock);
|
|
|
|
*using_mempool = true;
|
|
|
|
goto pool_alloc;
|
|
|
|
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
pool_alloc:
|
2019-07-04 07:27:42 +08:00
|
|
|
page = mempool_alloc(&c->bio_bounce_pages, GFP_NOIO);
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2019-07-04 07:27:42 +08:00
|
|
|
return page;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void bch2_bio_alloc_pages_pool(struct bch_fs *c, struct bio *bio,
|
2019-07-04 07:27:42 +08:00
|
|
|
size_t size)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
|
|
|
bool using_mempool = false;
|
|
|
|
|
2019-07-04 07:27:42 +08:00
|
|
|
while (size) {
|
|
|
|
struct page *page = __bio_alloc_page_pool(c, &using_mempool);
|
2020-11-06 01:16:05 +08:00
|
|
|
unsigned len = min_t(size_t, PAGE_SIZE, size);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-04 07:27:42 +08:00
|
|
|
BUG_ON(!bio_add_page(bio, page, len, 0));
|
|
|
|
size -= len;
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (using_mempool)
|
|
|
|
mutex_unlock(&c->bio_bounce_pages_lock);
|
|
|
|
}
|
|
|
|
|
2019-10-10 00:11:00 +08:00
|
|
|
/* Extent update path: */
|
|
|
|
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
int bch2_sum_sector_overwrites(struct btree_trans *trans,
|
|
|
|
struct btree_iter *extent_iter,
|
|
|
|
struct bkey_i *new,
|
|
|
|
bool *maybe_extending,
|
|
|
|
bool *should_check_enospc,
|
|
|
|
s64 *i_sectors_delta,
|
|
|
|
s64 *disk_sectors_delta)
|
2019-10-10 00:11:00 +08:00
|
|
|
{
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
struct bch_fs *c = trans->c;
|
2019-10-10 00:11:00 +08:00
|
|
|
struct btree_iter *iter;
|
|
|
|
struct bkey_s_c old;
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
unsigned new_replicas = bch2_bkey_replicas(c, bkey_i_to_s_c(new));
|
|
|
|
bool new_compressed = bch2_bkey_sectors_compressed(bkey_i_to_s_c(new));
|
2019-10-10 00:11:00 +08:00
|
|
|
int ret = 0;
|
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
*maybe_extending = true;
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
*should_check_enospc = false;
|
2020-12-11 02:38:54 +08:00
|
|
|
*i_sectors_delta = 0;
|
|
|
|
*disk_sectors_delta = 0;
|
2019-10-10 00:11:00 +08:00
|
|
|
|
|
|
|
iter = bch2_trans_copy_iter(trans, extent_iter);
|
|
|
|
|
|
|
|
for_each_btree_key_continue(iter, BTREE_ITER_SLOTS, old, ret) {
|
2020-12-11 02:38:54 +08:00
|
|
|
s64 sectors = min(new->k.p.offset, old.k->p.offset) -
|
|
|
|
max(bkey_start_offset(&new->k),
|
|
|
|
bkey_start_offset(old.k));
|
2019-10-10 00:11:00 +08:00
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
*i_sectors_delta += sectors *
|
2019-10-10 00:11:00 +08:00
|
|
|
(bkey_extent_is_allocation(&new->k) -
|
|
|
|
bkey_extent_is_allocation(old.k));
|
|
|
|
|
2021-03-21 12:03:34 +08:00
|
|
|
*disk_sectors_delta += sectors * bch2_bkey_nr_ptrs_allocated(bkey_i_to_s_c(new));
|
|
|
|
*disk_sectors_delta -= new->k.p.snapshot == old.k->p.snapshot
|
|
|
|
? sectors * bch2_bkey_nr_ptrs_fully_allocated(old)
|
|
|
|
: 0;
|
2020-12-11 02:38:54 +08:00
|
|
|
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
if (!*should_check_enospc &&
|
|
|
|
(new_replicas > bch2_bkey_replicas(c, old) ||
|
|
|
|
(!new_compressed && bch2_bkey_sectors_compressed(old))))
|
|
|
|
*should_check_enospc = true;
|
|
|
|
|
2019-10-10 00:11:00 +08:00
|
|
|
if (bkey_cmp(old.k->p, new->k.p) >= 0) {
|
|
|
|
/*
|
|
|
|
* Check if there's already data above where we're
|
|
|
|
* going to be writing to - this means we're definitely
|
|
|
|
* not extending the file:
|
|
|
|
*
|
|
|
|
* Note that it's not sufficient to check if there's
|
|
|
|
* data up to the sector offset we're going to be
|
|
|
|
* writing to, because i_size could be up to one block
|
|
|
|
* less:
|
|
|
|
*/
|
|
|
|
if (!bkey_cmp(old.k->p, new->k.p))
|
|
|
|
old = bch2_btree_iter_next(iter);
|
|
|
|
|
|
|
|
if (old.k && !bkey_err(old) &&
|
|
|
|
old.k->p.inode == extent_iter->pos.inode &&
|
|
|
|
bkey_extent_is_data(old.k))
|
|
|
|
*maybe_extending = false;
|
|
|
|
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bch2_trans_iter_put(trans, iter);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
int bch2_extent_update(struct btree_trans *trans,
|
|
|
|
struct btree_iter *iter,
|
|
|
|
struct bkey_i *k,
|
|
|
|
struct disk_reservation *disk_res,
|
|
|
|
u64 *journal_seq,
|
|
|
|
u64 new_i_size,
|
2020-12-11 02:38:54 +08:00
|
|
|
s64 *i_sectors_delta_total)
|
2019-10-10 00:11:00 +08:00
|
|
|
{
|
|
|
|
/* this must live until after bch2_trans_commit(): */
|
|
|
|
struct bkey_inode_buf inode_p;
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
bool extending = false, should_check_enospc;
|
2020-12-11 02:38:54 +08:00
|
|
|
s64 i_sectors_delta = 0, disk_sectors_delta = 0;
|
2019-10-10 00:11:00 +08:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = bch2_extent_trim_atomic(k, iter);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
ret = bch2_sum_sector_overwrites(trans, iter, k,
|
2020-12-11 02:38:54 +08:00
|
|
|
&extending,
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
&should_check_enospc,
|
2020-12-11 02:38:54 +08:00
|
|
|
&i_sectors_delta,
|
|
|
|
&disk_sectors_delta);
|
2019-10-10 00:11:00 +08:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
if (disk_res &&
|
|
|
|
disk_sectors_delta > (s64) disk_res->sectors) {
|
|
|
|
ret = bch2_disk_reservation_add(trans->c, disk_res,
|
|
|
|
disk_sectors_delta - disk_res->sectors,
|
bcachefs: Change when we allow overwrites
Originally, we'd check for -ENOSPC when getting a disk reservation
whenever the new extent took up more space on disk than the old extent.
Erasure coding screwed this up, because with erasure coding writes are
initially replicated, and then in the background the extra replicas are
dropped when the stripe is created. This means that with erasure coding
enabled, writes will always take up more space on disk than the data
they're overwriting - but, according to posix, overwrites aren't
supposed to return ENOSPC.
So, in this patch we fudge things: if the new extent has more replicas
than the _effective_ replicas of the old extent, or if the old extent is
compressed and the new one isn't, we check for ENOSPC when getting the
disk reservation - otherwise, we don't.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-12-15 10:59:33 +08:00
|
|
|
!should_check_enospc
|
|
|
|
? BCH_DISK_RESERVATION_NOFAIL : 0);
|
2020-12-11 02:38:54 +08:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-10-10 00:11:00 +08:00
|
|
|
new_i_size = extending
|
|
|
|
? min(k->k.p.offset << 9, new_i_size)
|
|
|
|
: 0;
|
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
if (i_sectors_delta || new_i_size) {
|
2019-10-10 00:11:00 +08:00
|
|
|
struct btree_iter *inode_iter;
|
|
|
|
struct bch_inode_unpacked inode_u;
|
|
|
|
|
|
|
|
inode_iter = bch2_inode_peek(trans, &inode_u,
|
|
|
|
k->k.p.inode, BTREE_ITER_INTENT);
|
|
|
|
if (IS_ERR(inode_iter))
|
|
|
|
return PTR_ERR(inode_iter);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* XXX:
|
|
|
|
* writeback can race a bit with truncate, because truncate
|
|
|
|
* first updates the inode then truncates the pagecache. This is
|
|
|
|
* ugly, but lets us preserve the invariant that the in memory
|
|
|
|
* i_size is always >= the on disk i_size.
|
|
|
|
*
|
|
|
|
BUG_ON(new_i_size > inode_u.bi_size &&
|
|
|
|
(inode_u.bi_flags & BCH_INODE_I_SIZE_DIRTY));
|
|
|
|
*/
|
|
|
|
BUG_ON(new_i_size > inode_u.bi_size && !extending);
|
|
|
|
|
|
|
|
if (!(inode_u.bi_flags & BCH_INODE_I_SIZE_DIRTY) &&
|
|
|
|
new_i_size > inode_u.bi_size)
|
|
|
|
inode_u.bi_size = new_i_size;
|
|
|
|
else
|
|
|
|
new_i_size = 0;
|
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
inode_u.bi_sectors += i_sectors_delta;
|
2019-10-10 00:11:00 +08:00
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
if (i_sectors_delta || new_i_size) {
|
2020-11-06 12:39:33 +08:00
|
|
|
bch2_inode_pack(trans->c, &inode_p, &inode_u);
|
bcachefs: Start using bpos.snapshot field
This patch starts treating the bpos.snapshot field like part of the key
in the btree code:
* bpos_successor() and bpos_predecessor() now include the snapshot field
* Keys in btrees that will be using snapshots (extents, inodes, dirents
and xattrs) now always have their snapshot field set to U32_MAX
The btree iterator code gets a new flag, BTREE_ITER_ALL_SNAPSHOTS, that
determines whether we're iterating over keys in all snapshots or not -
internally, this controlls whether bkey_(successor|predecessor)
increment/decrement the snapshot field, or only the higher bits of the
key.
We add a new member to struct btree_iter, iter->snapshot: when
BTREE_ITER_ALL_SNAPSHOTS is not set, iter->pos.snapshot should always
equal iter->snapshot, which will be 0 for btrees that don't use
snapshots, and alsways U32_MAX for btrees that will use snapshots
(until we enable snapshot creation).
This patch also introduces a new metadata version number, and compat
code for reading from/writing to older versions - this isn't a forced
upgrade (yet).
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2021-03-25 06:02:16 +08:00
|
|
|
|
|
|
|
inode_p.inode.k.p.snapshot = iter->snapshot;
|
|
|
|
|
2019-10-10 00:11:00 +08:00
|
|
|
bch2_trans_update(trans, inode_iter,
|
2020-01-01 05:17:42 +08:00
|
|
|
&inode_p.inode.k_i, 0);
|
2019-10-10 00:11:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bch2_trans_iter_put(trans, inode_iter);
|
|
|
|
}
|
|
|
|
|
2020-01-01 05:17:42 +08:00
|
|
|
bch2_trans_update(trans, iter, k, 0);
|
2019-10-10 00:11:00 +08:00
|
|
|
|
|
|
|
ret = bch2_trans_commit(trans, disk_res, journal_seq,
|
2019-10-20 07:03:23 +08:00
|
|
|
BTREE_INSERT_NOCHECK_RW|
|
2020-12-22 06:17:18 +08:00
|
|
|
BTREE_INSERT_NOFAIL);
|
2020-12-11 02:38:54 +08:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
2019-10-10 00:11:00 +08:00
|
|
|
|
2020-12-11 02:38:54 +08:00
|
|
|
if (i_sectors_delta_total)
|
|
|
|
*i_sectors_delta_total += i_sectors_delta;
|
|
|
|
return 0;
|
2019-10-10 00:11:00 +08:00
|
|
|
}
|
|
|
|
|
2019-10-11 00:47:22 +08:00
|
|
|
int bch2_fpunch_at(struct btree_trans *trans, struct btree_iter *iter,
|
|
|
|
struct bpos end, u64 *journal_seq,
|
|
|
|
s64 *i_sectors_delta)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = trans->c;
|
|
|
|
unsigned max_sectors = KEY_SIZE_MAX & (~0 << c->block_bits);
|
|
|
|
struct bkey_s_c k;
|
|
|
|
int ret = 0, ret2 = 0;
|
|
|
|
|
|
|
|
while ((k = bch2_btree_iter_peek(iter)).k &&
|
|
|
|
bkey_cmp(iter->pos, end) < 0) {
|
|
|
|
struct disk_reservation disk_res =
|
|
|
|
bch2_disk_reservation_init(c, 0);
|
|
|
|
struct bkey_i delete;
|
|
|
|
|
2020-02-27 04:39:46 +08:00
|
|
|
bch2_trans_begin(trans);
|
2019-12-21 05:35:24 +08:00
|
|
|
|
2019-10-11 00:47:22 +08:00
|
|
|
ret = bkey_err(k);
|
|
|
|
if (ret)
|
|
|
|
goto btree_err;
|
|
|
|
|
|
|
|
bkey_init(&delete.k);
|
|
|
|
delete.k.p = iter->pos;
|
|
|
|
|
|
|
|
/* create the biggest key we can */
|
|
|
|
bch2_key_resize(&delete.k, max_sectors);
|
2019-11-10 08:02:48 +08:00
|
|
|
bch2_cut_back(end, &delete);
|
2019-10-11 00:47:22 +08:00
|
|
|
|
|
|
|
ret = bch2_extent_update(trans, iter, &delete,
|
|
|
|
&disk_res, journal_seq,
|
|
|
|
0, i_sectors_delta);
|
|
|
|
bch2_disk_reservation_put(c, &disk_res);
|
|
|
|
btree_err:
|
|
|
|
if (ret == -EINTR) {
|
|
|
|
ret2 = ret;
|
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (bkey_cmp(iter->pos, end) > 0) {
|
|
|
|
bch2_btree_iter_set_pos(iter, end);
|
|
|
|
ret = bch2_btree_iter_traverse(iter);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret ?: ret2;
|
|
|
|
}
|
|
|
|
|
|
|
|
int bch2_fpunch(struct bch_fs *c, u64 inum, u64 start, u64 end,
|
|
|
|
u64 *journal_seq, s64 *i_sectors_delta)
|
|
|
|
{
|
|
|
|
struct btree_trans trans;
|
|
|
|
struct btree_iter *iter;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
bch2_trans_init(&trans, c, BTREE_ITER_MAX, 1024);
|
2021-02-21 08:27:37 +08:00
|
|
|
iter = bch2_trans_get_iter(&trans, BTREE_ID_extents,
|
2019-10-11 00:47:22 +08:00
|
|
|
POS(inum, start),
|
|
|
|
BTREE_ITER_INTENT);
|
|
|
|
|
|
|
|
ret = bch2_fpunch_at(&trans, iter, POS(inum, end),
|
|
|
|
journal_seq, i_sectors_delta);
|
2021-03-20 08:29:11 +08:00
|
|
|
|
|
|
|
bch2_trans_iter_put(&trans, iter);
|
2019-10-11 00:47:22 +08:00
|
|
|
bch2_trans_exit(&trans);
|
|
|
|
|
|
|
|
if (ret == -EINTR)
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-10-10 00:50:39 +08:00
|
|
|
int bch2_write_index_default(struct bch_write_op *op)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = op->c;
|
2020-12-18 04:08:58 +08:00
|
|
|
struct bkey_buf sk;
|
2019-10-10 00:50:39 +08:00
|
|
|
struct keylist *keys = &op->insert_keys;
|
|
|
|
struct bkey_i *k = bch2_keylist_front(keys);
|
|
|
|
struct btree_trans trans;
|
|
|
|
struct btree_iter *iter;
|
|
|
|
int ret;
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_init(&sk);
|
2019-10-10 00:50:39 +08:00
|
|
|
bch2_trans_init(&trans, c, BTREE_ITER_MAX, 1024);
|
|
|
|
|
2021-02-21 08:27:37 +08:00
|
|
|
iter = bch2_trans_get_iter(&trans, BTREE_ID_extents,
|
2019-10-10 00:50:39 +08:00
|
|
|
bkey_start_pos(&k->k),
|
|
|
|
BTREE_ITER_SLOTS|BTREE_ITER_INTENT);
|
|
|
|
|
|
|
|
do {
|
2020-02-27 04:39:46 +08:00
|
|
|
bch2_trans_begin(&trans);
|
2019-12-21 05:35:24 +08:00
|
|
|
|
2019-11-10 05:01:15 +08:00
|
|
|
k = bch2_keylist_front(keys);
|
2019-10-10 00:50:39 +08:00
|
|
|
|
bcachefs: Start using bpos.snapshot field
This patch starts treating the bpos.snapshot field like part of the key
in the btree code:
* bpos_successor() and bpos_predecessor() now include the snapshot field
* Keys in btrees that will be using snapshots (extents, inodes, dirents
and xattrs) now always have their snapshot field set to U32_MAX
The btree iterator code gets a new flag, BTREE_ITER_ALL_SNAPSHOTS, that
determines whether we're iterating over keys in all snapshots or not -
internally, this controlls whether bkey_(successor|predecessor)
increment/decrement the snapshot field, or only the higher bits of the
key.
We add a new member to struct btree_iter, iter->snapshot: when
BTREE_ITER_ALL_SNAPSHOTS is not set, iter->pos.snapshot should always
equal iter->snapshot, which will be 0 for btrees that don't use
snapshots, and alsways U32_MAX for btrees that will use snapshots
(until we enable snapshot creation).
This patch also introduces a new metadata version number, and compat
code for reading from/writing to older versions - this isn't a forced
upgrade (yet).
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2021-03-25 06:02:16 +08:00
|
|
|
k->k.p.snapshot = iter->snapshot;
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_realloc(&sk, c, k->k.u64s);
|
2019-11-10 05:01:15 +08:00
|
|
|
bkey_copy(sk.k, k);
|
2019-11-10 08:02:48 +08:00
|
|
|
bch2_cut_front(iter->pos, sk.k);
|
2019-10-10 00:50:39 +08:00
|
|
|
|
2019-11-10 05:01:15 +08:00
|
|
|
ret = bch2_extent_update(&trans, iter, sk.k,
|
2019-10-10 00:50:39 +08:00
|
|
|
&op->res, op_journal_seq(op),
|
|
|
|
op->new_i_size, &op->i_sectors_delta);
|
|
|
|
if (ret == -EINTR)
|
|
|
|
continue;
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
|
2019-11-10 08:02:48 +08:00
|
|
|
if (bkey_cmp(iter->pos, k->k.p) >= 0)
|
2019-10-10 00:50:39 +08:00
|
|
|
bch2_keylist_pop_front(keys);
|
|
|
|
} while (!bch2_keylist_empty(keys));
|
|
|
|
|
2021-03-20 08:29:11 +08:00
|
|
|
bch2_trans_iter_put(&trans, iter);
|
2019-10-10 00:50:39 +08:00
|
|
|
bch2_trans_exit(&trans);
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_exit(&sk, c);
|
2019-10-10 00:50:39 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
/* Writes */
|
|
|
|
|
|
|
|
void bch2_submit_wbio_replicas(struct bch_write_bio *wbio, struct bch_fs *c,
|
|
|
|
enum bch_data_type type,
|
|
|
|
const struct bkey_i *k)
|
|
|
|
{
|
2018-11-02 03:10:01 +08:00
|
|
|
struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(bkey_i_to_s_c(k));
|
2017-03-17 14:18:50 +08:00
|
|
|
const struct bch_extent_ptr *ptr;
|
|
|
|
struct bch_write_bio *n;
|
|
|
|
struct bch_dev *ca;
|
|
|
|
|
|
|
|
BUG_ON(c->opts.nochanges);
|
|
|
|
|
2018-11-02 03:10:01 +08:00
|
|
|
bkey_for_each_ptr(ptrs, ptr) {
|
2017-03-17 14:18:50 +08:00
|
|
|
BUG_ON(ptr->dev >= BCH_SB_MEMBERS_MAX ||
|
|
|
|
!c->devs[ptr->dev]);
|
|
|
|
|
|
|
|
ca = bch_dev_bkey_exists(c, ptr->dev);
|
|
|
|
|
2018-11-02 03:10:01 +08:00
|
|
|
if (to_entry(ptr + 1) < ptrs.end) {
|
2017-03-17 14:18:50 +08:00
|
|
|
n = to_wbio(bio_alloc_clone(NULL, &wbio->bio,
|
|
|
|
GFP_NOIO, &ca->replica_set));
|
|
|
|
|
|
|
|
n->bio.bi_end_io = wbio->bio.bi_end_io;
|
|
|
|
n->bio.bi_private = wbio->bio.bi_private;
|
|
|
|
n->parent = wbio;
|
|
|
|
n->split = true;
|
|
|
|
n->bounce = false;
|
|
|
|
n->put_bio = true;
|
|
|
|
n->bio.bi_opf = wbio->bio.bi_opf;
|
|
|
|
bio_inc_remaining(&wbio->bio);
|
|
|
|
} else {
|
|
|
|
n = wbio;
|
|
|
|
n->split = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
n->c = c;
|
|
|
|
n->dev = ptr->dev;
|
2020-08-04 01:58:36 +08:00
|
|
|
n->have_ioref = bch2_dev_get_ioref(ca,
|
|
|
|
type == BCH_DATA_btree ? READ : WRITE);
|
2017-03-17 14:18:50 +08:00
|
|
|
n->submit_time = local_clock();
|
|
|
|
n->bio.bi_iter.bi_sector = ptr->offset;
|
|
|
|
|
|
|
|
if (likely(n->have_ioref)) {
|
|
|
|
this_cpu_add(ca->io_done->sectors[WRITE][type],
|
|
|
|
bio_sectors(&n->bio));
|
|
|
|
|
|
|
|
bio_set_dev(&n->bio, ca->disk_sb.bdev);
|
|
|
|
|
2020-07-10 06:28:11 +08:00
|
|
|
if (type != BCH_DATA_btree && unlikely(c->opts.no_data_io)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bio_endio(&n->bio);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
submit_bio(&n->bio);
|
|
|
|
} else {
|
|
|
|
n->bio.bi_status = BLK_STS_REMOVED;
|
|
|
|
bio_endio(&n->bio);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __bch2_write(struct closure *);
|
|
|
|
|
|
|
|
static void bch2_write_done(struct closure *cl)
|
|
|
|
{
|
|
|
|
struct bch_write_op *op = container_of(cl, struct bch_write_op, cl);
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
|
|
|
|
if (!op->error && (op->flags & BCH_WRITE_FLUSH))
|
|
|
|
op->error = bch2_journal_error(&c->journal);
|
|
|
|
|
2020-06-30 06:22:06 +08:00
|
|
|
bch2_disk_reservation_put(c, &op->res);
|
2017-03-17 14:18:50 +08:00
|
|
|
percpu_ref_put(&c->writes);
|
|
|
|
bch2_keylist_free(&op->insert_keys, op->inline_keys);
|
|
|
|
|
|
|
|
bch2_time_stats_update(&c->times[BCH_TIME_data_write], op->start_time);
|
|
|
|
|
2019-12-28 02:44:03 +08:00
|
|
|
if (op->end_io) {
|
|
|
|
EBUG_ON(cl->parent);
|
|
|
|
closure_debug_destroy(cl);
|
2019-11-02 09:16:51 +08:00
|
|
|
op->end_io(op);
|
2019-12-28 02:44:03 +08:00
|
|
|
} else {
|
2019-11-02 09:16:51 +08:00
|
|
|
closure_return(cl);
|
2019-12-28 02:44:03 +08:00
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* bch_write_index - after a write, update index to point to new data
|
|
|
|
*/
|
|
|
|
static void __bch2_write_index(struct bch_write_op *op)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
struct keylist *keys = &op->insert_keys;
|
|
|
|
struct bch_extent_ptr *ptr;
|
|
|
|
struct bkey_i *src, *dst = keys->keys, *n, *k;
|
2018-11-02 03:13:19 +08:00
|
|
|
unsigned dev;
|
2017-03-17 14:18:50 +08:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
for (src = keys->keys; src != keys->top; src = n) {
|
|
|
|
n = bkey_next(src);
|
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
if (bkey_extent_is_direct_data(&src->k)) {
|
|
|
|
bch2_bkey_drop_ptrs(bkey_i_to_s(src), ptr,
|
|
|
|
test_bit(ptr->dev, op->failed.d));
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
if (!bch2_bkey_nr_ptrs(bkey_i_to_s_c(src))) {
|
|
|
|
ret = -EIO;
|
|
|
|
goto err;
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
if (dst != src)
|
|
|
|
memmove_u64s_down(dst, src, src->u64s);
|
2017-03-17 14:18:50 +08:00
|
|
|
dst = bkey_next(dst);
|
|
|
|
}
|
|
|
|
|
|
|
|
keys->top = dst;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* probably not the ideal place to hook this in, but I don't
|
|
|
|
* particularly want to plumb io_opts all the way through the btree
|
|
|
|
* update stack right now
|
|
|
|
*/
|
2018-02-24 05:26:10 +08:00
|
|
|
for_each_keylist_key(keys, k) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_rebalance_add_key(c, bkey_i_to_s_c(k), &op->opts);
|
|
|
|
|
2018-02-24 05:26:10 +08:00
|
|
|
if (bch2_bkey_is_incompressible(bkey_i_to_s_c(k)))
|
|
|
|
bch2_check_set_feature(op->c, BCH_FEATURE_incompressible);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
if (!bch2_keylist_empty(keys)) {
|
|
|
|
u64 sectors_start = keylist_sectors(keys);
|
|
|
|
int ret = op->index_update_fn(op);
|
|
|
|
|
2019-09-27 10:21:39 +08:00
|
|
|
BUG_ON(ret == -EINTR);
|
2017-03-17 14:18:50 +08:00
|
|
|
BUG_ON(keylist_sectors(keys) && !ret);
|
|
|
|
|
|
|
|
op->written += sectors_start - keylist_sectors(keys);
|
|
|
|
|
|
|
|
if (ret) {
|
2020-12-04 02:57:22 +08:00
|
|
|
bch_err_inum_ratelimited(c, op->pos.inode,
|
|
|
|
"write error %i from btree update", ret);
|
2017-03-17 14:18:50 +08:00
|
|
|
op->error = ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
out:
|
2018-11-02 03:13:19 +08:00
|
|
|
/* If some a bucket wasn't written, we can't erasure code it: */
|
|
|
|
for_each_set_bit(dev, op->failed.d, BCH_SB_MEMBERS_MAX)
|
|
|
|
bch2_open_bucket_write_error(c, &op->open_buckets, dev);
|
|
|
|
|
2018-10-06 16:12:42 +08:00
|
|
|
bch2_open_buckets_put(c, &op->open_buckets);
|
2017-03-17 14:18:50 +08:00
|
|
|
return;
|
|
|
|
err:
|
|
|
|
keys->top = keys->keys;
|
|
|
|
op->error = ret;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_write_index(struct closure *cl)
|
|
|
|
{
|
|
|
|
struct bch_write_op *op = container_of(cl, struct bch_write_op, cl);
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
|
|
|
|
__bch2_write_index(op);
|
|
|
|
|
2020-05-13 12:15:28 +08:00
|
|
|
if (!(op->flags & BCH_WRITE_DONE)) {
|
|
|
|
continue_at(cl, __bch2_write, index_update_wq(op));
|
|
|
|
} else if (!op->error && (op->flags & BCH_WRITE_FLUSH)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_journal_flush_seq_async(&c->journal,
|
|
|
|
*op_journal_seq(op),
|
|
|
|
cl);
|
|
|
|
continue_at(cl, bch2_write_done, index_update_wq(op));
|
|
|
|
} else {
|
|
|
|
continue_at_nobarrier(cl, bch2_write_done, NULL);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_write_endio(struct bio *bio)
|
|
|
|
{
|
|
|
|
struct closure *cl = bio->bi_private;
|
|
|
|
struct bch_write_op *op = container_of(cl, struct bch_write_op, cl);
|
|
|
|
struct bch_write_bio *wbio = to_wbio(bio);
|
|
|
|
struct bch_write_bio *parent = wbio->split ? wbio->parent : NULL;
|
|
|
|
struct bch_fs *c = wbio->c;
|
|
|
|
struct bch_dev *ca = bch_dev_bkey_exists(c, wbio->dev);
|
|
|
|
|
2020-12-04 02:57:22 +08:00
|
|
|
if (bch2_dev_inum_io_err_on(bio->bi_status, ca,
|
|
|
|
op->pos.inode,
|
|
|
|
op->pos.offset - bio_sectors(bio), /* XXX definitely wrong */
|
|
|
|
"data write error: %s",
|
2020-07-22 01:34:22 +08:00
|
|
|
bch2_blk_status_to_str(bio->bi_status)))
|
2017-03-17 14:18:50 +08:00
|
|
|
set_bit(wbio->dev, op->failed.d);
|
|
|
|
|
|
|
|
if (wbio->have_ioref) {
|
|
|
|
bch2_latency_acct(ca, wbio->submit_time, WRITE);
|
|
|
|
percpu_ref_put(&ca->io_ref);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (wbio->bounce)
|
|
|
|
bch2_bio_free_pages_pool(c, bio);
|
|
|
|
|
|
|
|
if (wbio->put_bio)
|
|
|
|
bio_put(bio);
|
|
|
|
|
|
|
|
if (parent)
|
|
|
|
bio_endio(&parent->bio);
|
2019-11-02 09:16:51 +08:00
|
|
|
else if (!(op->flags & BCH_WRITE_SKIP_CLOSURE_PUT))
|
2017-03-17 14:18:50 +08:00
|
|
|
closure_put(cl);
|
2019-11-02 09:16:51 +08:00
|
|
|
else
|
|
|
|
continue_at_nobarrier(cl, bch2_write_index, index_update_wq(op));
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void init_append_extent(struct bch_write_op *op,
|
|
|
|
struct write_point *wp,
|
|
|
|
struct bversion version,
|
|
|
|
struct bch_extent_crc_unpacked crc)
|
|
|
|
{
|
2019-10-29 15:57:58 +08:00
|
|
|
struct bkey_i_extent *e;
|
2018-12-06 01:28:35 +08:00
|
|
|
struct bch_extent_ptr *ptr;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
op->pos.offset += crc.uncompressed_size;
|
2019-10-29 15:57:58 +08:00
|
|
|
|
|
|
|
e = bkey_extent_init(op->insert_keys.top);
|
2018-09-28 09:08:39 +08:00
|
|
|
e->k.p = op->pos;
|
|
|
|
e->k.size = crc.uncompressed_size;
|
|
|
|
e->k.version = version;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2018-09-28 09:08:39 +08:00
|
|
|
if (crc.csum_type ||
|
|
|
|
crc.compression_type ||
|
|
|
|
crc.nonce)
|
2019-07-26 01:52:14 +08:00
|
|
|
bch2_extent_crc_append(&e->k_i, crc);
|
2018-09-28 09:08:39 +08:00
|
|
|
|
2018-11-02 03:10:01 +08:00
|
|
|
bch2_alloc_sectors_append_ptrs(op->c, wp, &e->k_i, crc.compressed_size);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2018-12-06 01:28:35 +08:00
|
|
|
if (op->flags & BCH_WRITE_CACHED)
|
|
|
|
extent_for_each_ptr(extent_i_to_s(e), ptr)
|
|
|
|
ptr->cached = true;
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_keylist_push(&op->insert_keys);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct bio *bch2_write_bio_alloc(struct bch_fs *c,
|
|
|
|
struct write_point *wp,
|
|
|
|
struct bio *src,
|
2018-11-02 03:13:19 +08:00
|
|
|
bool *page_alloc_failed,
|
|
|
|
void *buf)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
|
|
|
struct bch_write_bio *wbio;
|
|
|
|
struct bio *bio;
|
|
|
|
unsigned output_available =
|
|
|
|
min(wp->sectors_free << 9, src->bi_iter.bi_size);
|
2019-08-23 04:41:50 +08:00
|
|
|
unsigned pages = DIV_ROUND_UP(output_available +
|
|
|
|
(buf
|
|
|
|
? ((unsigned long) buf & (PAGE_SIZE - 1))
|
|
|
|
: 0), PAGE_SIZE);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
bio = bio_alloc_bioset(NULL, pages, 0,
|
|
|
|
GFP_NOIO, &c->bio_write);
|
|
|
|
wbio = wbio_init(bio);
|
|
|
|
wbio->put_bio = true;
|
|
|
|
/* copy WRITE_SYNC flag */
|
|
|
|
wbio->bio.bi_opf = src->bi_opf;
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (buf) {
|
2019-07-04 07:27:42 +08:00
|
|
|
bch2_bio_map(bio, buf, output_available);
|
2018-11-02 03:13:19 +08:00
|
|
|
return bio;
|
|
|
|
}
|
|
|
|
|
|
|
|
wbio->bounce = true;
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
/*
|
|
|
|
* We can't use mempool for more than c->sb.encoded_extent_max
|
|
|
|
* worth of pages, but we'd like to allocate more if we can:
|
|
|
|
*/
|
2019-07-04 07:27:42 +08:00
|
|
|
bch2_bio_alloc_pages_pool(c, bio,
|
|
|
|
min_t(unsigned, output_available,
|
|
|
|
c->sb.encoded_extent_max << 9));
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-04 07:27:42 +08:00
|
|
|
if (bio->bi_iter.bi_size < output_available)
|
|
|
|
*page_alloc_failed =
|
|
|
|
bch2_bio_alloc_pages(bio,
|
|
|
|
output_available -
|
|
|
|
bio->bi_iter.bi_size,
|
|
|
|
GFP_NOFS) != 0;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
return bio;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int bch2_write_rechecksum(struct bch_fs *c,
|
|
|
|
struct bch_write_op *op,
|
|
|
|
unsigned new_csum_type)
|
|
|
|
{
|
|
|
|
struct bio *bio = &op->wbio.bio;
|
|
|
|
struct bch_extent_crc_unpacked new_crc;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/* bch2_rechecksum_bio() can't encrypt or decrypt data: */
|
|
|
|
|
|
|
|
if (bch2_csum_type_is_encryption(op->crc.csum_type) !=
|
|
|
|
bch2_csum_type_is_encryption(new_csum_type))
|
|
|
|
new_csum_type = op->crc.csum_type;
|
|
|
|
|
|
|
|
ret = bch2_rechecksum_bio(c, bio, op->version, op->crc,
|
|
|
|
NULL, &new_crc,
|
|
|
|
op->crc.offset, op->crc.live_size,
|
|
|
|
new_csum_type);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
bio_advance(bio, op->crc.offset << 9);
|
|
|
|
bio->bi_iter.bi_size = op->crc.live_size << 9;
|
|
|
|
op->crc = new_crc;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int bch2_write_decrypt(struct bch_write_op *op)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
struct nonce nonce = extent_nonce(op->version, op->crc);
|
|
|
|
struct bch_csum csum;
|
|
|
|
|
|
|
|
if (!bch2_csum_type_is_encryption(op->crc.csum_type))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we need to decrypt data in the write path, we'll no longer be able
|
|
|
|
* to verify the existing checksum (poly1305 mac, in this case) after
|
|
|
|
* it's decrypted - this is the last point we'll be able to reverify the
|
|
|
|
* checksum:
|
|
|
|
*/
|
|
|
|
csum = bch2_checksum_bio(c, op->crc.csum_type, nonce, &op->wbio.bio);
|
|
|
|
if (bch2_crc_cmp(op->crc.csum, csum))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
bch2_encrypt_bio(c, op->crc.csum_type, nonce, &op->wbio.bio);
|
|
|
|
op->crc.csum_type = 0;
|
|
|
|
op->crc.csum = (struct bch_csum) { 0, 0 };
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum prep_encoded_ret {
|
|
|
|
PREP_ENCODED_OK,
|
|
|
|
PREP_ENCODED_ERR,
|
|
|
|
PREP_ENCODED_CHECKSUM_ERR,
|
|
|
|
PREP_ENCODED_DO_WRITE,
|
|
|
|
} bch2_write_prep_encoded_data(struct bch_write_op *op, struct write_point *wp)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
struct bio *bio = &op->wbio.bio;
|
|
|
|
|
|
|
|
if (!(op->flags & BCH_WRITE_DATA_ENCODED))
|
|
|
|
return PREP_ENCODED_OK;
|
|
|
|
|
|
|
|
BUG_ON(bio_sectors(bio) != op->crc.compressed_size);
|
|
|
|
|
|
|
|
/* Can we just write the entire extent as is? */
|
|
|
|
if (op->crc.uncompressed_size == op->crc.live_size &&
|
|
|
|
op->crc.compressed_size <= wp->sectors_free &&
|
2018-02-24 05:26:10 +08:00
|
|
|
(op->crc.compression_type == op->compression_type ||
|
|
|
|
op->incompressible)) {
|
|
|
|
if (!crc_is_compressed(op->crc) &&
|
2017-03-17 14:18:50 +08:00
|
|
|
op->csum_type != op->crc.csum_type &&
|
|
|
|
bch2_write_rechecksum(c, op, op->csum_type))
|
|
|
|
return PREP_ENCODED_CHECKSUM_ERR;
|
|
|
|
|
|
|
|
return PREP_ENCODED_DO_WRITE;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the data is compressed and we couldn't write the entire extent as
|
|
|
|
* is, we have to decompress it:
|
|
|
|
*/
|
2018-02-24 05:26:10 +08:00
|
|
|
if (crc_is_compressed(op->crc)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_csum csum;
|
|
|
|
|
|
|
|
if (bch2_write_decrypt(op))
|
|
|
|
return PREP_ENCODED_CHECKSUM_ERR;
|
|
|
|
|
|
|
|
/* Last point we can still verify checksum: */
|
|
|
|
csum = bch2_checksum_bio(c, op->crc.csum_type,
|
|
|
|
extent_nonce(op->version, op->crc),
|
|
|
|
bio);
|
|
|
|
if (bch2_crc_cmp(op->crc.csum, csum))
|
|
|
|
return PREP_ENCODED_CHECKSUM_ERR;
|
|
|
|
|
|
|
|
if (bch2_bio_uncompress_inplace(c, bio, &op->crc))
|
|
|
|
return PREP_ENCODED_ERR;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* No longer have compressed data after this point - data might be
|
|
|
|
* encrypted:
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the data is checksummed and we're only writing a subset,
|
|
|
|
* rechecksum and adjust bio to point to currently live data:
|
|
|
|
*/
|
|
|
|
if ((op->crc.live_size != op->crc.uncompressed_size ||
|
|
|
|
op->crc.csum_type != op->csum_type) &&
|
|
|
|
bch2_write_rechecksum(c, op, op->csum_type))
|
|
|
|
return PREP_ENCODED_CHECKSUM_ERR;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we want to compress the data, it has to be decrypted:
|
|
|
|
*/
|
|
|
|
if ((op->compression_type ||
|
|
|
|
bch2_csum_type_is_encryption(op->crc.csum_type) !=
|
|
|
|
bch2_csum_type_is_encryption(op->csum_type)) &&
|
|
|
|
bch2_write_decrypt(op))
|
|
|
|
return PREP_ENCODED_CHECKSUM_ERR;
|
|
|
|
|
|
|
|
return PREP_ENCODED_OK;
|
|
|
|
}
|
|
|
|
|
2019-11-02 09:16:51 +08:00
|
|
|
static int bch2_write_extent(struct bch_write_op *op, struct write_point *wp,
|
|
|
|
struct bio **_dst)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
struct bio *src = &op->wbio.bio, *dst = src;
|
|
|
|
struct bvec_iter saved_iter;
|
2018-11-02 03:13:19 +08:00
|
|
|
void *ec_buf;
|
2019-11-02 09:16:51 +08:00
|
|
|
struct bpos ec_pos = op->pos;
|
2018-11-02 03:13:19 +08:00
|
|
|
unsigned total_output = 0, total_input = 0;
|
|
|
|
bool bounce = false;
|
|
|
|
bool page_alloc_failed = false;
|
2017-03-17 14:18:50 +08:00
|
|
|
int ret, more = 0;
|
|
|
|
|
|
|
|
BUG_ON(!bio_sectors(src));
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
ec_buf = bch2_writepoint_ec_buf(c, wp);
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
switch (bch2_write_prep_encoded_data(op, wp)) {
|
|
|
|
case PREP_ENCODED_OK:
|
|
|
|
break;
|
|
|
|
case PREP_ENCODED_ERR:
|
|
|
|
ret = -EIO;
|
|
|
|
goto err;
|
|
|
|
case PREP_ENCODED_CHECKSUM_ERR:
|
|
|
|
goto csum_err;
|
|
|
|
case PREP_ENCODED_DO_WRITE:
|
2019-11-02 09:16:51 +08:00
|
|
|
/* XXX look for bug here */
|
2018-11-02 03:13:19 +08:00
|
|
|
if (ec_buf) {
|
|
|
|
dst = bch2_write_bio_alloc(c, wp, src,
|
|
|
|
&page_alloc_failed,
|
|
|
|
ec_buf);
|
|
|
|
bio_copy_data(dst, src);
|
|
|
|
bounce = true;
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
init_append_extent(op, wp, op->version, op->crc);
|
|
|
|
goto do_write;
|
|
|
|
}
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (ec_buf ||
|
|
|
|
op->compression_type ||
|
2017-03-17 14:18:50 +08:00
|
|
|
(op->csum_type &&
|
|
|
|
!(op->flags & BCH_WRITE_PAGES_STABLE)) ||
|
|
|
|
(bch2_csum_type_is_encryption(op->csum_type) &&
|
|
|
|
!(op->flags & BCH_WRITE_PAGES_OWNED))) {
|
2018-11-02 03:13:19 +08:00
|
|
|
dst = bch2_write_bio_alloc(c, wp, src,
|
|
|
|
&page_alloc_failed,
|
|
|
|
ec_buf);
|
2017-03-17 14:18:50 +08:00
|
|
|
bounce = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
saved_iter = dst->bi_iter;
|
|
|
|
|
|
|
|
do {
|
|
|
|
struct bch_extent_crc_unpacked crc =
|
|
|
|
(struct bch_extent_crc_unpacked) { 0 };
|
|
|
|
struct bversion version = op->version;
|
|
|
|
size_t dst_len, src_len;
|
|
|
|
|
|
|
|
if (page_alloc_failed &&
|
|
|
|
bio_sectors(dst) < wp->sectors_free &&
|
|
|
|
bio_sectors(dst) < c->sb.encoded_extent_max)
|
|
|
|
break;
|
|
|
|
|
|
|
|
BUG_ON(op->compression_type &&
|
|
|
|
(op->flags & BCH_WRITE_DATA_ENCODED) &&
|
|
|
|
bch2_csum_type_is_encryption(op->crc.csum_type));
|
|
|
|
BUG_ON(op->compression_type && !bounce);
|
|
|
|
|
2018-02-24 05:26:10 +08:00
|
|
|
crc.compression_type = op->incompressible
|
|
|
|
? BCH_COMPRESSION_TYPE_incompressible
|
|
|
|
: op->compression_type
|
|
|
|
? bch2_bio_compress(c, dst, &dst_len, src, &src_len,
|
|
|
|
op->compression_type)
|
2017-03-17 14:18:50 +08:00
|
|
|
: 0;
|
2018-02-24 05:26:10 +08:00
|
|
|
if (!crc_is_compressed(crc)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
dst_len = min(dst->bi_iter.bi_size, src->bi_iter.bi_size);
|
|
|
|
dst_len = min_t(unsigned, dst_len, wp->sectors_free << 9);
|
|
|
|
|
|
|
|
if (op->csum_type)
|
|
|
|
dst_len = min_t(unsigned, dst_len,
|
|
|
|
c->sb.encoded_extent_max << 9);
|
|
|
|
|
|
|
|
if (bounce) {
|
|
|
|
swap(dst->bi_iter.bi_size, dst_len);
|
|
|
|
bio_copy_data(dst, src);
|
|
|
|
swap(dst->bi_iter.bi_size, dst_len);
|
|
|
|
}
|
|
|
|
|
|
|
|
src_len = dst_len;
|
|
|
|
}
|
|
|
|
|
|
|
|
BUG_ON(!src_len || !dst_len);
|
|
|
|
|
|
|
|
if (bch2_csum_type_is_encryption(op->csum_type)) {
|
|
|
|
if (bversion_zero(version)) {
|
2020-02-19 09:02:41 +08:00
|
|
|
version.lo = atomic64_inc_return(&c->key_version);
|
2017-03-17 14:18:50 +08:00
|
|
|
} else {
|
|
|
|
crc.nonce = op->nonce;
|
|
|
|
op->nonce += src_len >> 9;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((op->flags & BCH_WRITE_DATA_ENCODED) &&
|
2018-02-24 05:26:10 +08:00
|
|
|
!crc_is_compressed(crc) &&
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_csum_type_is_encryption(op->crc.csum_type) ==
|
|
|
|
bch2_csum_type_is_encryption(op->csum_type)) {
|
|
|
|
/*
|
|
|
|
* Note: when we're using rechecksum(), we need to be
|
|
|
|
* checksumming @src because it has all the data our
|
|
|
|
* existing checksum covers - if we bounced (because we
|
|
|
|
* were trying to compress), @dst will only have the
|
|
|
|
* part of the data the new checksum will cover.
|
|
|
|
*
|
|
|
|
* But normally we want to be checksumming post bounce,
|
|
|
|
* because part of the reason for bouncing is so the
|
|
|
|
* data can't be modified (by userspace) while it's in
|
|
|
|
* flight.
|
|
|
|
*/
|
|
|
|
if (bch2_rechecksum_bio(c, src, version, op->crc,
|
|
|
|
&crc, &op->crc,
|
|
|
|
src_len >> 9,
|
|
|
|
bio_sectors(src) - (src_len >> 9),
|
|
|
|
op->csum_type))
|
|
|
|
goto csum_err;
|
|
|
|
} else {
|
|
|
|
if ((op->flags & BCH_WRITE_DATA_ENCODED) &&
|
|
|
|
bch2_rechecksum_bio(c, src, version, op->crc,
|
|
|
|
NULL, &op->crc,
|
|
|
|
src_len >> 9,
|
|
|
|
bio_sectors(src) - (src_len >> 9),
|
|
|
|
op->crc.csum_type))
|
|
|
|
goto csum_err;
|
|
|
|
|
|
|
|
crc.compressed_size = dst_len >> 9;
|
|
|
|
crc.uncompressed_size = src_len >> 9;
|
|
|
|
crc.live_size = src_len >> 9;
|
|
|
|
|
|
|
|
swap(dst->bi_iter.bi_size, dst_len);
|
|
|
|
bch2_encrypt_bio(c, op->csum_type,
|
|
|
|
extent_nonce(version, crc), dst);
|
|
|
|
crc.csum = bch2_checksum_bio(c, op->csum_type,
|
|
|
|
extent_nonce(version, crc), dst);
|
|
|
|
crc.csum_type = op->csum_type;
|
|
|
|
swap(dst->bi_iter.bi_size, dst_len);
|
|
|
|
}
|
|
|
|
|
|
|
|
init_append_extent(op, wp, version, crc);
|
|
|
|
|
|
|
|
if (dst != src)
|
|
|
|
bio_advance(dst, dst_len);
|
|
|
|
bio_advance(src, src_len);
|
2018-11-02 03:13:19 +08:00
|
|
|
total_output += dst_len;
|
|
|
|
total_input += src_len;
|
2017-03-17 14:18:50 +08:00
|
|
|
} while (dst->bi_iter.bi_size &&
|
|
|
|
src->bi_iter.bi_size &&
|
|
|
|
wp->sectors_free &&
|
|
|
|
!bch2_keylist_realloc(&op->insert_keys,
|
|
|
|
op->inline_keys,
|
|
|
|
ARRAY_SIZE(op->inline_keys),
|
|
|
|
BKEY_EXTENT_U64s_MAX));
|
|
|
|
|
|
|
|
more = src->bi_iter.bi_size != 0;
|
|
|
|
|
|
|
|
dst->bi_iter = saved_iter;
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (dst == src && more) {
|
|
|
|
BUG_ON(total_output != total_input);
|
|
|
|
|
|
|
|
dst = bio_split(src, total_input >> 9,
|
2017-03-17 14:18:50 +08:00
|
|
|
GFP_NOIO, &c->bio_write);
|
2018-11-02 03:13:19 +08:00
|
|
|
wbio_init(dst)->put_bio = true;
|
|
|
|
/* copy WRITE_SYNC flag */
|
|
|
|
dst->bi_opf = src->bi_opf;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
dst->bi_iter.bi_size = total_output;
|
|
|
|
do_write:
|
|
|
|
/* might have done a realloc... */
|
2019-11-02 09:16:51 +08:00
|
|
|
bch2_ec_add_backpointer(c, wp, ec_pos, total_input >> 9);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-02 09:16:51 +08:00
|
|
|
*_dst = dst;
|
2017-03-17 14:18:50 +08:00
|
|
|
return more;
|
|
|
|
csum_err:
|
|
|
|
bch_err(c, "error verifying existing checksum while "
|
|
|
|
"rewriting existing data (memory corruption?)");
|
|
|
|
ret = -EIO;
|
|
|
|
err:
|
2018-11-02 03:13:19 +08:00
|
|
|
if (to_wbio(dst)->bounce)
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_bio_free_pages_pool(c, dst);
|
2018-11-02 03:13:19 +08:00
|
|
|
if (to_wbio(dst)->put_bio)
|
2017-03-17 14:18:50 +08:00
|
|
|
bio_put(dst);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __bch2_write(struct closure *cl)
|
|
|
|
{
|
|
|
|
struct bch_write_op *op = container_of(cl, struct bch_write_op, cl);
|
|
|
|
struct bch_fs *c = op->c;
|
|
|
|
struct write_point *wp;
|
2019-11-02 09:16:51 +08:00
|
|
|
struct bio *bio;
|
|
|
|
bool skip_put = true;
|
2020-07-21 01:00:15 +08:00
|
|
|
unsigned nofs_flags;
|
2017-03-17 14:18:50 +08:00
|
|
|
int ret;
|
2020-07-21 01:00:15 +08:00
|
|
|
|
|
|
|
nofs_flags = memalloc_nofs_save();
|
2017-03-17 14:18:50 +08:00
|
|
|
again:
|
2018-11-02 03:13:19 +08:00
|
|
|
memset(&op->failed, 0, sizeof(op->failed));
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
do {
|
2019-11-02 09:16:51 +08:00
|
|
|
struct bkey_i *key_to_write;
|
|
|
|
unsigned key_to_write_offset = op->insert_keys.top_p -
|
|
|
|
op->insert_keys.keys_p;
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
/* +1 for possible cache device: */
|
2018-10-06 16:12:42 +08:00
|
|
|
if (op->open_buckets.nr + op->nr_replicas + 1 >
|
|
|
|
ARRAY_SIZE(op->open_buckets.v))
|
2017-03-17 14:18:50 +08:00
|
|
|
goto flush_io;
|
|
|
|
|
|
|
|
if (bch2_keylist_realloc(&op->insert_keys,
|
|
|
|
op->inline_keys,
|
|
|
|
ARRAY_SIZE(op->inline_keys),
|
|
|
|
BKEY_EXTENT_U64s_MAX))
|
|
|
|
goto flush_io;
|
|
|
|
|
2020-03-18 23:46:46 +08:00
|
|
|
if ((op->flags & BCH_WRITE_FROM_INTERNAL) &&
|
|
|
|
percpu_ref_is_dying(&c->writes)) {
|
|
|
|
ret = -EROFS;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
2020-08-13 01:48:02 +08:00
|
|
|
/*
|
|
|
|
* The copygc thread is now global, which means it's no longer
|
|
|
|
* freeing up space on specific disks, which means that
|
|
|
|
* allocations for specific disks may hang arbitrarily long:
|
|
|
|
*/
|
2017-03-17 14:18:50 +08:00
|
|
|
wp = bch2_alloc_sectors_start(c,
|
|
|
|
op->target,
|
2018-11-02 03:13:19 +08:00
|
|
|
op->opts.erasure_code,
|
2017-03-17 14:18:50 +08:00
|
|
|
op->write_point,
|
|
|
|
&op->devs_have,
|
|
|
|
op->nr_replicas,
|
|
|
|
op->nr_replicas_required,
|
|
|
|
op->alloc_reserve,
|
|
|
|
op->flags,
|
2020-08-13 01:48:02 +08:00
|
|
|
(op->flags & (BCH_WRITE_ALLOC_NOWAIT|
|
|
|
|
BCH_WRITE_ONLY_SPECIFIED_DEVS)) ? NULL : cl);
|
2017-03-17 14:18:50 +08:00
|
|
|
EBUG_ON(!wp);
|
|
|
|
|
|
|
|
if (unlikely(IS_ERR(wp))) {
|
|
|
|
if (unlikely(PTR_ERR(wp) != -EAGAIN)) {
|
|
|
|
ret = PTR_ERR(wp);
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
goto flush_io;
|
|
|
|
}
|
|
|
|
|
2020-07-23 10:40:32 +08:00
|
|
|
/*
|
|
|
|
* It's possible for the allocator to fail, put us on the
|
|
|
|
* freelist waitlist, and then succeed in one of various retry
|
|
|
|
* paths: if that happens, we need to disable the skip_put
|
|
|
|
* optimization because otherwise there won't necessarily be a
|
|
|
|
* barrier before we free the bch_write_op:
|
|
|
|
*/
|
|
|
|
if (atomic_read(&cl->remaining) & CLOSURE_WAITING)
|
|
|
|
skip_put = false;
|
|
|
|
|
2018-10-06 16:12:42 +08:00
|
|
|
bch2_open_bucket_get(c, wp, &op->open_buckets);
|
2019-11-02 09:16:51 +08:00
|
|
|
ret = bch2_write_extent(op, wp, &bio);
|
|
|
|
bch2_alloc_sectors_done(c, wp);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (ret < 0)
|
|
|
|
goto err;
|
2019-11-02 09:16:51 +08:00
|
|
|
|
2020-05-13 12:15:28 +08:00
|
|
|
if (ret) {
|
2019-11-02 09:16:51 +08:00
|
|
|
skip_put = false;
|
2020-05-13 12:15:28 +08:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* for the skip_put optimization this has to be set
|
|
|
|
* before we submit the bio:
|
|
|
|
*/
|
|
|
|
op->flags |= BCH_WRITE_DONE;
|
|
|
|
}
|
2019-11-02 09:16:51 +08:00
|
|
|
|
|
|
|
bio->bi_end_io = bch2_write_endio;
|
|
|
|
bio->bi_private = &op->cl;
|
2019-11-10 05:43:16 +08:00
|
|
|
bio->bi_opf |= REQ_OP_WRITE;
|
2019-11-02 09:16:51 +08:00
|
|
|
|
|
|
|
if (!skip_put)
|
|
|
|
closure_get(bio->bi_private);
|
|
|
|
else
|
|
|
|
op->flags |= BCH_WRITE_SKIP_CLOSURE_PUT;
|
|
|
|
|
|
|
|
key_to_write = (void *) (op->insert_keys.keys_p +
|
|
|
|
key_to_write_offset);
|
|
|
|
|
2020-07-10 06:28:11 +08:00
|
|
|
bch2_submit_wbio_replicas(to_wbio(bio), c, BCH_DATA_user,
|
2019-11-02 09:16:51 +08:00
|
|
|
key_to_write);
|
2017-03-17 14:18:50 +08:00
|
|
|
} while (ret);
|
|
|
|
|
2019-11-02 09:16:51 +08:00
|
|
|
if (!skip_put)
|
|
|
|
continue_at(cl, bch2_write_index, index_update_wq(op));
|
2020-07-21 01:00:15 +08:00
|
|
|
out:
|
|
|
|
memalloc_nofs_restore(nofs_flags);
|
2017-03-17 14:18:50 +08:00
|
|
|
return;
|
|
|
|
err:
|
|
|
|
op->error = ret;
|
2020-05-13 12:15:28 +08:00
|
|
|
op->flags |= BCH_WRITE_DONE;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-02 09:16:51 +08:00
|
|
|
continue_at(cl, bch2_write_index, index_update_wq(op));
|
2020-07-21 01:00:15 +08:00
|
|
|
goto out;
|
2017-03-17 14:18:50 +08:00
|
|
|
flush_io:
|
2020-05-13 12:15:28 +08:00
|
|
|
/*
|
|
|
|
* If the write can't all be submitted at once, we generally want to
|
|
|
|
* block synchronously as that signals backpressure to the caller.
|
|
|
|
*
|
|
|
|
* However, if we're running out of a workqueue, we can't block here
|
|
|
|
* because we'll be blocking other work items from completing:
|
|
|
|
*/
|
|
|
|
if (current->flags & PF_WQ_WORKER) {
|
|
|
|
continue_at(cl, bch2_write_index, index_update_wq(op));
|
2020-07-21 01:00:15 +08:00
|
|
|
goto out;
|
2020-05-13 12:15:28 +08:00
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
closure_sync(cl);
|
|
|
|
|
|
|
|
if (!bch2_keylist_empty(&op->insert_keys)) {
|
|
|
|
__bch2_write_index(op);
|
|
|
|
|
|
|
|
if (op->error) {
|
2020-05-13 12:15:28 +08:00
|
|
|
op->flags |= BCH_WRITE_DONE;
|
2017-03-17 14:18:50 +08:00
|
|
|
continue_at_nobarrier(cl, bch2_write_done, NULL);
|
2020-07-21 01:00:15 +08:00
|
|
|
goto out;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
goto again;
|
|
|
|
}
|
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
static void bch2_write_data_inline(struct bch_write_op *op, unsigned data_len)
|
|
|
|
{
|
|
|
|
struct closure *cl = &op->cl;
|
|
|
|
struct bio *bio = &op->wbio.bio;
|
|
|
|
struct bvec_iter iter;
|
|
|
|
struct bkey_i_inline_data *id;
|
|
|
|
unsigned sectors;
|
|
|
|
int ret;
|
|
|
|
|
2019-12-29 09:17:06 +08:00
|
|
|
bch2_check_set_feature(op->c, BCH_FEATURE_inline_data);
|
2019-11-30 02:47:42 +08:00
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
ret = bch2_keylist_realloc(&op->insert_keys, op->inline_keys,
|
|
|
|
ARRAY_SIZE(op->inline_keys),
|
|
|
|
BKEY_U64s + DIV_ROUND_UP(data_len, 8));
|
|
|
|
if (ret) {
|
|
|
|
op->error = ret;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
sectors = bio_sectors(bio);
|
|
|
|
op->pos.offset += sectors;
|
|
|
|
|
|
|
|
id = bkey_inline_data_init(op->insert_keys.top);
|
|
|
|
id->k.p = op->pos;
|
|
|
|
id->k.version = op->version;
|
|
|
|
id->k.size = sectors;
|
|
|
|
|
|
|
|
iter = bio->bi_iter;
|
|
|
|
iter.bi_size = data_len;
|
|
|
|
memcpy_from_bio(id->v.data, bio, iter);
|
|
|
|
|
|
|
|
while (data_len & 7)
|
|
|
|
id->v.data[data_len++] = '\0';
|
|
|
|
set_bkey_val_bytes(&id->k, data_len);
|
|
|
|
bch2_keylist_push(&op->insert_keys);
|
|
|
|
|
|
|
|
op->flags |= BCH_WRITE_WROTE_DATA_INLINE;
|
2020-05-13 12:15:28 +08:00
|
|
|
op->flags |= BCH_WRITE_DONE;
|
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
continue_at_nobarrier(cl, bch2_write_index, NULL);
|
|
|
|
return;
|
|
|
|
err:
|
|
|
|
bch2_write_done(&op->cl);
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
/**
|
|
|
|
* bch_write - handle a write to a cache device or flash only volume
|
|
|
|
*
|
|
|
|
* This is the starting point for any data to end up in a cache device; it could
|
|
|
|
* be from a normal write, or a writeback write, or a write to a flash only
|
|
|
|
* volume - it's also used by the moving garbage collector to compact data in
|
|
|
|
* mostly empty buckets.
|
|
|
|
*
|
|
|
|
* It first writes the data to the cache, creating a list of keys to be inserted
|
|
|
|
* (if the data won't fit in a single open bucket, there will be multiple keys);
|
|
|
|
* after the data is written it calls bch_journal, and after the keys have been
|
|
|
|
* added to the next journal write they're inserted into the btree.
|
|
|
|
*
|
|
|
|
* If op->discard is true, instead of inserting the data it invalidates the
|
|
|
|
* region of the cache represented by op->bio and op->inode.
|
|
|
|
*/
|
|
|
|
void bch2_write(struct closure *cl)
|
|
|
|
{
|
|
|
|
struct bch_write_op *op = container_of(cl, struct bch_write_op, cl);
|
2019-08-22 08:16:42 +08:00
|
|
|
struct bio *bio = &op->wbio.bio;
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_fs *c = op->c;
|
2019-11-10 05:43:16 +08:00
|
|
|
unsigned data_len;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
BUG_ON(!op->nr_replicas);
|
|
|
|
BUG_ON(!op->write_point.v);
|
|
|
|
BUG_ON(!bkey_cmp(op->pos, POS_MAX));
|
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
op->start_time = local_clock();
|
|
|
|
bch2_keylist_init(&op->insert_keys, op->inline_keys);
|
|
|
|
wbio_init(bio)->put_bio = false;
|
|
|
|
|
2019-08-22 08:16:42 +08:00
|
|
|
if (bio_sectors(bio) & (c->opts.block_size - 1)) {
|
2020-12-04 02:57:22 +08:00
|
|
|
bch_err_inum_ratelimited(c, op->pos.inode,
|
|
|
|
"misaligned write");
|
2019-08-22 08:16:42 +08:00
|
|
|
op->error = -EIO;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
if (c->opts.nochanges ||
|
|
|
|
!percpu_ref_tryget(&c->writes)) {
|
|
|
|
op->error = -EROFS;
|
2019-08-22 08:16:42 +08:00
|
|
|
goto err;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2019-08-22 08:16:42 +08:00
|
|
|
bch2_increment_clock(c, bio_sectors(bio), WRITE);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-10 05:43:16 +08:00
|
|
|
data_len = min_t(u64, bio->bi_iter.bi_size,
|
|
|
|
op->new_i_size - (op->pos.offset << 9));
|
|
|
|
|
2019-11-30 02:48:09 +08:00
|
|
|
if (c->opts.inline_data &&
|
|
|
|
data_len <= min(block_bytes(c) / 2, 1024U)) {
|
2019-11-10 05:43:16 +08:00
|
|
|
bch2_write_data_inline(op, data_len);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
continue_at_nobarrier(cl, __bch2_write, NULL);
|
2019-08-22 08:16:42 +08:00
|
|
|
return;
|
|
|
|
err:
|
2020-06-30 06:22:06 +08:00
|
|
|
bch2_disk_reservation_put(c, &op->res);
|
2019-12-28 02:44:03 +08:00
|
|
|
|
|
|
|
if (op->end_io) {
|
|
|
|
EBUG_ON(cl->parent);
|
|
|
|
closure_debug_destroy(cl);
|
2019-11-12 02:42:10 +08:00
|
|
|
op->end_io(op);
|
2019-12-28 02:44:03 +08:00
|
|
|
} else {
|
2019-11-12 02:42:10 +08:00
|
|
|
closure_return(cl);
|
2019-12-28 02:44:03 +08:00
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Cache promotion on read */
|
|
|
|
|
|
|
|
struct promote_op {
|
|
|
|
struct closure cl;
|
2018-12-19 03:43:00 +08:00
|
|
|
struct rcu_head rcu;
|
2017-03-17 14:18:50 +08:00
|
|
|
u64 start_time;
|
|
|
|
|
|
|
|
struct rhash_head hash;
|
|
|
|
struct bpos pos;
|
|
|
|
|
|
|
|
struct migrate_write write;
|
|
|
|
struct bio_vec bi_inline_vecs[0]; /* must be last */
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct rhashtable_params bch_promote_params = {
|
|
|
|
.head_offset = offsetof(struct promote_op, hash),
|
|
|
|
.key_offset = offsetof(struct promote_op, pos),
|
|
|
|
.key_len = sizeof(struct bpos),
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline bool should_promote(struct bch_fs *c, struct bkey_s_c k,
|
|
|
|
struct bpos pos,
|
|
|
|
struct bch_io_opts opts,
|
|
|
|
unsigned flags)
|
|
|
|
{
|
|
|
|
if (!(flags & BCH_READ_MAY_PROMOTE))
|
|
|
|
return false;
|
|
|
|
|
2019-06-10 23:31:07 +08:00
|
|
|
if (!opts.promote_target)
|
2017-03-17 14:18:50 +08:00
|
|
|
return false;
|
|
|
|
|
2019-07-26 01:52:14 +08:00
|
|
|
if (bch2_bkey_has_target(c, k, opts.promote_target))
|
2017-03-17 14:18:50 +08:00
|
|
|
return false;
|
|
|
|
|
2019-06-10 23:31:07 +08:00
|
|
|
if (bch2_target_congested(c, opts.promote_target)) {
|
|
|
|
/* XXX trace this */
|
2017-03-17 14:18:50 +08:00
|
|
|
return false;
|
2019-06-10 23:31:07 +08:00
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (rhashtable_lookup_fast(&c->promote_table, &pos,
|
|
|
|
bch_promote_params))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void promote_free(struct bch_fs *c, struct promote_op *op)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = rhashtable_remove_fast(&c->promote_table, &op->hash,
|
|
|
|
bch_promote_params);
|
|
|
|
BUG_ON(ret);
|
|
|
|
percpu_ref_put(&c->writes);
|
2018-12-19 03:43:00 +08:00
|
|
|
kfree_rcu(op, rcu);
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void promote_done(struct closure *cl)
|
|
|
|
{
|
|
|
|
struct promote_op *op =
|
|
|
|
container_of(cl, struct promote_op, cl);
|
|
|
|
struct bch_fs *c = op->write.op.c;
|
|
|
|
|
|
|
|
bch2_time_stats_update(&c->times[BCH_TIME_data_promote],
|
|
|
|
op->start_time);
|
|
|
|
|
|
|
|
bch2_bio_free_pages_pool(c, &op->write.op.wbio.bio);
|
|
|
|
promote_free(c, op);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void promote_start(struct promote_op *op, struct bch_read_bio *rbio)
|
|
|
|
{
|
|
|
|
struct bch_fs *c = rbio->c;
|
|
|
|
struct closure *cl = &op->cl;
|
|
|
|
struct bio *bio = &op->write.op.wbio.bio;
|
|
|
|
|
|
|
|
trace_promote(&rbio->bio);
|
|
|
|
|
|
|
|
/* we now own pages: */
|
|
|
|
BUG_ON(!rbio->bounce);
|
|
|
|
BUG_ON(rbio->bio.bi_vcnt > bio->bi_max_vecs);
|
|
|
|
|
|
|
|
memcpy(bio->bi_io_vec, rbio->bio.bi_io_vec,
|
|
|
|
sizeof(struct bio_vec) * rbio->bio.bi_vcnt);
|
|
|
|
swap(bio->bi_vcnt, rbio->bio.bi_vcnt);
|
|
|
|
|
|
|
|
bch2_migrate_read_done(&op->write, rbio);
|
|
|
|
|
|
|
|
closure_init(cl, NULL);
|
|
|
|
closure_call(&op->write.op.cl, bch2_write, c->wq, cl);
|
|
|
|
closure_return_with_destructor(cl, promote_done);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct promote_op *__promote_alloc(struct bch_fs *c,
|
2019-08-16 21:59:56 +08:00
|
|
|
enum btree_id btree_id,
|
2018-02-24 05:26:10 +08:00
|
|
|
struct bkey_s_c k,
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bpos pos,
|
2018-10-02 23:03:39 +08:00
|
|
|
struct extent_ptr_decoded *pick,
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_io_opts opts,
|
2019-07-09 23:16:33 +08:00
|
|
|
unsigned sectors,
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_read_bio **rbio)
|
|
|
|
{
|
|
|
|
struct promote_op *op = NULL;
|
|
|
|
struct bio *bio;
|
2019-07-09 23:16:33 +08:00
|
|
|
unsigned pages = DIV_ROUND_UP(sectors, PAGE_SECTORS);
|
2017-03-17 14:18:50 +08:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (!percpu_ref_tryget(&c->writes))
|
|
|
|
return NULL;
|
|
|
|
|
2019-07-09 23:16:33 +08:00
|
|
|
op = kzalloc(sizeof(*op) + sizeof(struct bio_vec) * pages, GFP_NOIO);
|
2017-03-17 14:18:50 +08:00
|
|
|
if (!op)
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
op->start_time = local_clock();
|
|
|
|
op->pos = pos;
|
|
|
|
|
|
|
|
/*
|
2019-07-09 23:16:33 +08:00
|
|
|
* We don't use the mempool here because extents that aren't
|
|
|
|
* checksummed or compressed can be too big for the mempool:
|
2017-03-17 14:18:50 +08:00
|
|
|
*/
|
2019-07-09 23:16:33 +08:00
|
|
|
*rbio = kzalloc(sizeof(struct bch_read_bio) +
|
|
|
|
sizeof(struct bio_vec) * pages,
|
|
|
|
GFP_NOIO);
|
|
|
|
if (!*rbio)
|
|
|
|
goto err;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-09 23:16:33 +08:00
|
|
|
rbio_init(&(*rbio)->bio, opts);
|
|
|
|
bio_init(&(*rbio)->bio, NULL, (*rbio)->bio.bi_inline_vecs, pages, 0);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-09 23:16:33 +08:00
|
|
|
if (bch2_bio_alloc_pages(&(*rbio)->bio, sectors << 9,
|
|
|
|
GFP_NOIO))
|
|
|
|
goto err;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-09 23:16:33 +08:00
|
|
|
(*rbio)->bounce = true;
|
|
|
|
(*rbio)->split = true;
|
|
|
|
(*rbio)->kmalloc = true;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (rhashtable_lookup_insert_fast(&c->promote_table, &op->hash,
|
|
|
|
bch_promote_params))
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
bio = &op->write.op.wbio.bio;
|
2019-07-09 23:16:33 +08:00
|
|
|
bio_init(bio, NULL, bio->bi_inline_vecs, pages, 0);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
ret = bch2_migrate_write_init(c, &op->write,
|
|
|
|
writepoint_hashed((unsigned long) current),
|
|
|
|
opts,
|
|
|
|
DATA_PROMOTE,
|
|
|
|
(struct data_opts) {
|
2020-07-23 11:11:48 +08:00
|
|
|
.target = opts.promote_target,
|
|
|
|
.nr_replicas = 1,
|
2017-03-17 14:18:50 +08:00
|
|
|
},
|
2018-02-24 05:26:10 +08:00
|
|
|
btree_id, k);
|
2017-03-17 14:18:50 +08:00
|
|
|
BUG_ON(ret);
|
|
|
|
|
|
|
|
return op;
|
|
|
|
err:
|
|
|
|
if (*rbio)
|
|
|
|
bio_free_pages(&(*rbio)->bio);
|
|
|
|
kfree(*rbio);
|
|
|
|
*rbio = NULL;
|
|
|
|
kfree(op);
|
|
|
|
percpu_ref_put(&c->writes);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2019-11-08 04:04:13 +08:00
|
|
|
noinline
|
|
|
|
static struct promote_op *promote_alloc(struct bch_fs *c,
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bvec_iter iter,
|
|
|
|
struct bkey_s_c k,
|
2018-10-02 23:03:39 +08:00
|
|
|
struct extent_ptr_decoded *pick,
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_io_opts opts,
|
|
|
|
unsigned flags,
|
|
|
|
struct bch_read_bio **rbio,
|
|
|
|
bool *bounce,
|
|
|
|
bool *read_full)
|
|
|
|
{
|
|
|
|
bool promote_full = *read_full || READ_ONCE(c->promote_whole_extents);
|
2019-07-09 23:16:33 +08:00
|
|
|
/* data might have to be decompressed in the write path: */
|
2017-03-17 14:18:50 +08:00
|
|
|
unsigned sectors = promote_full
|
2019-07-09 23:16:33 +08:00
|
|
|
? max(pick->crc.compressed_size, pick->crc.live_size)
|
2017-03-17 14:18:50 +08:00
|
|
|
: bvec_iter_sectors(iter);
|
|
|
|
struct bpos pos = promote_full
|
|
|
|
? bkey_start_pos(k.k)
|
|
|
|
: POS(k.k->p.inode, iter.bi_sector);
|
|
|
|
struct promote_op *promote;
|
|
|
|
|
|
|
|
if (!should_promote(c, k, pos, opts, flags))
|
|
|
|
return NULL;
|
|
|
|
|
2019-08-16 21:59:56 +08:00
|
|
|
promote = __promote_alloc(c,
|
|
|
|
k.k->type == KEY_TYPE_reflink_v
|
2021-02-21 08:27:37 +08:00
|
|
|
? BTREE_ID_reflink
|
|
|
|
: BTREE_ID_extents,
|
2018-02-24 05:26:10 +08:00
|
|
|
k, pos, pick, opts, sectors, rbio);
|
2017-03-17 14:18:50 +08:00
|
|
|
if (!promote)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
*bounce = true;
|
|
|
|
*read_full = promote_full;
|
|
|
|
return promote;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Read */
|
|
|
|
|
|
|
|
#define READ_RETRY_AVOID 1
|
|
|
|
#define READ_RETRY 2
|
|
|
|
#define READ_ERR 3
|
|
|
|
|
|
|
|
enum rbio_context {
|
|
|
|
RBIO_CONTEXT_NULL,
|
|
|
|
RBIO_CONTEXT_HIGHPRI,
|
|
|
|
RBIO_CONTEXT_UNBOUND,
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline struct bch_read_bio *
|
|
|
|
bch2_rbio_parent(struct bch_read_bio *rbio)
|
|
|
|
{
|
|
|
|
return rbio->split ? rbio->parent : rbio;
|
|
|
|
}
|
|
|
|
|
|
|
|
__always_inline
|
|
|
|
static void bch2_rbio_punt(struct bch_read_bio *rbio, work_func_t fn,
|
|
|
|
enum rbio_context context,
|
|
|
|
struct workqueue_struct *wq)
|
|
|
|
{
|
|
|
|
if (context <= rbio->context) {
|
|
|
|
fn(&rbio->work);
|
|
|
|
} else {
|
|
|
|
rbio->work.func = fn;
|
|
|
|
rbio->context = context;
|
|
|
|
queue_work(wq, &rbio->work);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct bch_read_bio *bch2_rbio_free(struct bch_read_bio *rbio)
|
|
|
|
{
|
|
|
|
BUG_ON(rbio->bounce && !rbio->split);
|
|
|
|
|
|
|
|
if (rbio->promote)
|
|
|
|
promote_free(rbio->c, rbio->promote);
|
|
|
|
rbio->promote = NULL;
|
|
|
|
|
|
|
|
if (rbio->bounce)
|
|
|
|
bch2_bio_free_pages_pool(rbio->c, &rbio->bio);
|
|
|
|
|
|
|
|
if (rbio->split) {
|
|
|
|
struct bch_read_bio *parent = rbio->parent;
|
|
|
|
|
|
|
|
if (rbio->kmalloc)
|
|
|
|
kfree(rbio);
|
|
|
|
else
|
|
|
|
bio_put(&rbio->bio);
|
|
|
|
|
|
|
|
rbio = parent;
|
|
|
|
}
|
|
|
|
|
|
|
|
return rbio;
|
|
|
|
}
|
|
|
|
|
2019-09-08 01:16:41 +08:00
|
|
|
/*
|
|
|
|
* Only called on a top level bch_read_bio to complete an entire read request,
|
|
|
|
* not a split:
|
|
|
|
*/
|
2017-03-17 14:18:50 +08:00
|
|
|
static void bch2_rbio_done(struct bch_read_bio *rbio)
|
|
|
|
{
|
2019-09-08 01:16:41 +08:00
|
|
|
if (rbio->start_time)
|
|
|
|
bch2_time_stats_update(&rbio->c->times[BCH_TIME_data_read],
|
|
|
|
rbio->start_time);
|
2017-03-17 14:18:50 +08:00
|
|
|
bio_endio(&rbio->bio);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_read_retry_nodecode(struct bch_fs *c, struct bch_read_bio *rbio,
|
|
|
|
struct bvec_iter bvec_iter, u64 inode,
|
2018-11-02 03:28:45 +08:00
|
|
|
struct bch_io_failures *failed,
|
|
|
|
unsigned flags)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
2019-03-26 03:10:15 +08:00
|
|
|
struct btree_trans trans;
|
|
|
|
struct btree_iter *iter;
|
2020-12-18 04:08:58 +08:00
|
|
|
struct bkey_buf sk;
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bkey_s_c k;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
flags &= ~BCH_READ_LAST_FRAGMENT;
|
2019-10-25 02:22:29 +08:00
|
|
|
flags |= BCH_READ_MUST_CLONE;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_init(&sk);
|
2019-05-15 22:54:43 +08:00
|
|
|
bch2_trans_init(&trans, c, 0, 0);
|
2019-03-26 03:10:15 +08:00
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
iter = bch2_trans_get_iter(&trans, rbio->data_btree,
|
|
|
|
rbio->read_pos, BTREE_ITER_SLOTS);
|
2017-03-17 14:18:50 +08:00
|
|
|
retry:
|
|
|
|
rbio->bio.bi_status = 0;
|
|
|
|
|
2019-03-26 03:10:15 +08:00
|
|
|
k = bch2_btree_iter_peek_slot(iter);
|
2019-03-28 10:03:30 +08:00
|
|
|
if (bkey_err(k))
|
2017-03-17 14:18:50 +08:00
|
|
|
goto err;
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_reassemble(&sk, c, k);
|
2019-11-10 05:01:15 +08:00
|
|
|
k = bkey_i_to_s_c(sk.k);
|
2019-03-26 03:10:15 +08:00
|
|
|
bch2_trans_unlock(&trans);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-10 05:01:15 +08:00
|
|
|
if (!bch2_bkey_matches_ptr(c, k,
|
2019-07-26 01:52:14 +08:00
|
|
|
rbio->pick.ptr,
|
2021-03-15 09:30:08 +08:00
|
|
|
rbio->data_pos.offset -
|
2019-07-26 01:52:14 +08:00
|
|
|
rbio->pick.crc.offset)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
/* extent we wanted to read no longer exists: */
|
|
|
|
rbio->hole = true;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
ret = __bch2_read_extent(&trans, rbio, bvec_iter,
|
|
|
|
rbio->read_pos,
|
|
|
|
rbio->data_btree,
|
|
|
|
k, 0, failed, flags);
|
2017-03-17 14:18:50 +08:00
|
|
|
if (ret == READ_RETRY)
|
|
|
|
goto retry;
|
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
out:
|
|
|
|
bch2_rbio_done(rbio);
|
2021-03-20 08:29:11 +08:00
|
|
|
bch2_trans_iter_put(&trans, iter);
|
2019-03-26 03:10:15 +08:00
|
|
|
bch2_trans_exit(&trans);
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_exit(&sk, c);
|
2019-03-26 03:10:15 +08:00
|
|
|
return;
|
|
|
|
err:
|
|
|
|
rbio->bio.bi_status = BLK_STS_IOERR;
|
|
|
|
goto out;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_rbio_retry(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct bch_read_bio *rbio =
|
|
|
|
container_of(work, struct bch_read_bio, work);
|
|
|
|
struct bch_fs *c = rbio->c;
|
|
|
|
struct bvec_iter iter = rbio->bvec_iter;
|
|
|
|
unsigned flags = rbio->flags;
|
2021-03-15 09:30:08 +08:00
|
|
|
u64 inode = rbio->read_pos.inode;
|
2018-11-02 03:28:45 +08:00
|
|
|
struct bch_io_failures failed = { .nr = 0 };
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
trace_read_retry(&rbio->bio);
|
|
|
|
|
|
|
|
if (rbio->retry == READ_RETRY_AVOID)
|
2018-11-02 03:28:45 +08:00
|
|
|
bch2_mark_io_failure(&failed, &rbio->pick);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
rbio->bio.bi_status = 0;
|
|
|
|
|
|
|
|
rbio = bch2_rbio_free(rbio);
|
|
|
|
|
|
|
|
flags |= BCH_READ_IN_RETRY;
|
|
|
|
flags &= ~BCH_READ_MAY_PROMOTE;
|
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
if (flags & BCH_READ_NODECODE) {
|
2018-11-02 03:28:45 +08:00
|
|
|
bch2_read_retry_nodecode(c, rbio, iter, inode, &failed, flags);
|
2021-03-13 09:29:28 +08:00
|
|
|
} else {
|
|
|
|
flags &= ~BCH_READ_LAST_FRAGMENT;
|
|
|
|
flags |= BCH_READ_MUST_CLONE;
|
|
|
|
|
|
|
|
__bch2_read(c, rbio, iter, inode, &failed, flags);
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_rbio_error(struct bch_read_bio *rbio, int retry,
|
|
|
|
blk_status_t error)
|
|
|
|
{
|
|
|
|
rbio->retry = retry;
|
|
|
|
|
|
|
|
if (rbio->flags & BCH_READ_IN_RETRY)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (retry == READ_ERR) {
|
|
|
|
rbio = bch2_rbio_free(rbio);
|
|
|
|
|
|
|
|
rbio->bio.bi_status = error;
|
|
|
|
bch2_rbio_done(rbio);
|
|
|
|
} else {
|
|
|
|
bch2_rbio_punt(rbio, bch2_rbio_retry,
|
|
|
|
RBIO_CONTEXT_UNBOUND, system_unbound_wq);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-03-17 03:48:58 +08:00
|
|
|
static int __bch2_rbio_narrow_crcs(struct btree_trans *trans,
|
|
|
|
struct bch_read_bio *rbio)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
|
|
|
struct bch_fs *c = rbio->c;
|
2021-03-15 09:30:08 +08:00
|
|
|
u64 data_offset = rbio->data_pos.offset - rbio->pick.crc.offset;
|
2020-03-17 03:48:58 +08:00
|
|
|
struct bch_extent_crc_unpacked new_crc;
|
|
|
|
struct btree_iter *iter = NULL;
|
|
|
|
struct bkey_i *new;
|
|
|
|
struct bkey_s_c k;
|
|
|
|
int ret = 0;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2018-02-24 05:26:10 +08:00
|
|
|
if (crc_is_compressed(rbio->pick.crc))
|
2020-03-17 03:48:58 +08:00
|
|
|
return 0;
|
2019-03-14 08:49:16 +08:00
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
iter = bch2_trans_get_iter(trans, rbio->data_btree, rbio->data_pos,
|
2019-09-23 07:35:12 +08:00
|
|
|
BTREE_ITER_SLOTS|BTREE_ITER_INTENT);
|
|
|
|
k = bch2_btree_iter_peek_slot(iter);
|
2020-03-17 03:48:58 +08:00
|
|
|
if ((ret = bkey_err(k)))
|
|
|
|
goto out;
|
|
|
|
|
2019-07-26 01:52:14 +08:00
|
|
|
if (bversion_cmp(k.k->version, rbio->version) ||
|
|
|
|
!bch2_bkey_matches_ptr(c, k, rbio->pick.ptr, data_offset))
|
2017-03-17 14:18:50 +08:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
/* Extent was merged? */
|
2019-07-26 01:52:14 +08:00
|
|
|
if (bkey_start_offset(k.k) < data_offset ||
|
|
|
|
k.k->p.offset > data_offset + rbio->pick.crc.uncompressed_size)
|
2017-03-17 14:18:50 +08:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
if (bch2_rechecksum_bio(c, &rbio->bio, rbio->version,
|
2019-05-16 00:24:39 +08:00
|
|
|
rbio->pick.crc, NULL, &new_crc,
|
2019-07-26 01:52:14 +08:00
|
|
|
bkey_start_offset(k.k) - data_offset, k.k->size,
|
2019-05-16 00:24:39 +08:00
|
|
|
rbio->pick.crc.csum_type)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bch_err(c, "error verifying existing checksum while narrowing checksum (memory corruption?)");
|
2020-03-17 03:48:58 +08:00
|
|
|
ret = 0;
|
2017-03-17 14:18:50 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
/*
|
|
|
|
* going to be temporarily appending another checksum entry:
|
|
|
|
*/
|
|
|
|
new = bch2_trans_kmalloc(trans, bkey_bytes(k.k) +
|
|
|
|
sizeof(struct bch_extent_crc128));
|
|
|
|
if ((ret = PTR_ERR_OR_ZERO(new)))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
bkey_reassemble(new, k);
|
|
|
|
|
2020-03-17 03:48:58 +08:00
|
|
|
if (!bch2_bkey_narrow_crcs(new, new_crc))
|
2017-03-17 14:18:50 +08:00
|
|
|
goto out;
|
|
|
|
|
2020-03-17 03:48:58 +08:00
|
|
|
bch2_trans_update(trans, iter, new, 0);
|
2017-03-17 14:18:50 +08:00
|
|
|
out:
|
2020-03-17 03:48:58 +08:00
|
|
|
bch2_trans_iter_put(trans, iter);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static noinline void bch2_rbio_narrow_crcs(struct bch_read_bio *rbio)
|
|
|
|
{
|
|
|
|
bch2_trans_do(rbio->c, NULL, NULL, BTREE_INSERT_NOFAIL,
|
|
|
|
__bch2_rbio_narrow_crcs(&trans, rbio));
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Inner part that may run in process context */
|
|
|
|
static void __bch2_read_endio(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct bch_read_bio *rbio =
|
|
|
|
container_of(work, struct bch_read_bio, work);
|
|
|
|
struct bch_fs *c = rbio->c;
|
|
|
|
struct bch_dev *ca = bch_dev_bkey_exists(c, rbio->pick.ptr.dev);
|
|
|
|
struct bio *src = &rbio->bio;
|
|
|
|
struct bio *dst = &bch2_rbio_parent(rbio)->bio;
|
|
|
|
struct bvec_iter dst_iter = rbio->bvec_iter;
|
|
|
|
struct bch_extent_crc_unpacked crc = rbio->pick.crc;
|
|
|
|
struct nonce nonce = extent_nonce(rbio->version, crc);
|
|
|
|
struct bch_csum csum;
|
|
|
|
|
|
|
|
/* Reset iterator for checksumming and copying bounced data: */
|
|
|
|
if (rbio->bounce) {
|
|
|
|
src->bi_iter.bi_size = crc.compressed_size << 9;
|
|
|
|
src->bi_iter.bi_idx = 0;
|
|
|
|
src->bi_iter.bi_bvec_done = 0;
|
|
|
|
} else {
|
|
|
|
src->bi_iter = rbio->bvec_iter;
|
|
|
|
}
|
|
|
|
|
|
|
|
csum = bch2_checksum_bio(c, crc.csum_type, nonce, src);
|
|
|
|
if (bch2_crc_cmp(csum, rbio->pick.crc.csum) && !c->opts.no_data_io)
|
|
|
|
goto csum_err;
|
|
|
|
|
|
|
|
if (unlikely(rbio->narrow_crcs))
|
|
|
|
bch2_rbio_narrow_crcs(rbio);
|
|
|
|
|
|
|
|
if (rbio->flags & BCH_READ_NODECODE)
|
|
|
|
goto nodecode;
|
|
|
|
|
|
|
|
/* Adjust crc to point to subset of data we want: */
|
2019-07-10 00:56:43 +08:00
|
|
|
crc.offset += rbio->offset_into_extent;
|
2017-03-17 14:18:50 +08:00
|
|
|
crc.live_size = bvec_iter_sectors(rbio->bvec_iter);
|
|
|
|
|
2018-02-24 05:26:10 +08:00
|
|
|
if (crc_is_compressed(crc)) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_encrypt_bio(c, crc.csum_type, nonce, src);
|
|
|
|
if (bch2_bio_uncompress(c, src, dst, dst_iter, crc))
|
|
|
|
goto decompression_err;
|
|
|
|
} else {
|
|
|
|
/* don't need to decrypt the entire bio: */
|
|
|
|
nonce = nonce_add(nonce, crc.offset << 9);
|
|
|
|
bio_advance(src, crc.offset << 9);
|
|
|
|
|
|
|
|
BUG_ON(src->bi_iter.bi_size < dst_iter.bi_size);
|
|
|
|
src->bi_iter.bi_size = dst_iter.bi_size;
|
|
|
|
|
|
|
|
bch2_encrypt_bio(c, crc.csum_type, nonce, src);
|
|
|
|
|
|
|
|
if (rbio->bounce) {
|
|
|
|
struct bvec_iter src_iter = src->bi_iter;
|
|
|
|
bio_copy_data_iter(dst, &dst_iter, src, &src_iter);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rbio->promote) {
|
|
|
|
/*
|
|
|
|
* Re encrypt data we decrypted, so it's consistent with
|
|
|
|
* rbio->crc:
|
|
|
|
*/
|
|
|
|
bch2_encrypt_bio(c, crc.csum_type, nonce, src);
|
|
|
|
promote_start(rbio->promote, rbio);
|
|
|
|
rbio->promote = NULL;
|
|
|
|
}
|
|
|
|
nodecode:
|
|
|
|
if (likely(!(rbio->flags & BCH_READ_IN_RETRY))) {
|
|
|
|
rbio = bch2_rbio_free(rbio);
|
|
|
|
bch2_rbio_done(rbio);
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
csum_err:
|
|
|
|
/*
|
|
|
|
* Checksum error: if the bio wasn't bounced, we may have been
|
|
|
|
* reading into buffers owned by userspace (that userspace can
|
|
|
|
* scribble over) - retry the read, bouncing it this time:
|
|
|
|
*/
|
|
|
|
if (!rbio->bounce && (rbio->flags & BCH_READ_USER_MAPPED)) {
|
|
|
|
rbio->flags |= BCH_READ_MUST_BOUNCE;
|
|
|
|
bch2_rbio_error(rbio, READ_RETRY, BLK_STS_IOERR);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
bch2_dev_inum_io_error(ca, rbio->read_pos.inode, (u64) rbio->bvec_iter.bi_sector,
|
2020-12-04 02:57:22 +08:00
|
|
|
"data checksum error: expected %0llx:%0llx got %0llx:%0llx (type %u)",
|
2017-03-17 14:18:50 +08:00
|
|
|
rbio->pick.crc.csum.hi, rbio->pick.crc.csum.lo,
|
|
|
|
csum.hi, csum.lo, crc.csum_type);
|
|
|
|
bch2_rbio_error(rbio, READ_RETRY_AVOID, BLK_STS_IOERR);
|
|
|
|
return;
|
|
|
|
decompression_err:
|
2021-03-15 09:30:08 +08:00
|
|
|
bch_err_inum_ratelimited(c, rbio->read_pos.inode,
|
2020-12-04 02:57:22 +08:00
|
|
|
"decompression error");
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_rbio_error(rbio, READ_ERR, BLK_STS_IOERR);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void bch2_read_endio(struct bio *bio)
|
|
|
|
{
|
|
|
|
struct bch_read_bio *rbio =
|
|
|
|
container_of(bio, struct bch_read_bio, bio);
|
|
|
|
struct bch_fs *c = rbio->c;
|
|
|
|
struct bch_dev *ca = bch_dev_bkey_exists(c, rbio->pick.ptr.dev);
|
|
|
|
struct workqueue_struct *wq = NULL;
|
|
|
|
enum rbio_context context = RBIO_CONTEXT_NULL;
|
|
|
|
|
|
|
|
if (rbio->have_ioref) {
|
|
|
|
bch2_latency_acct(ca, rbio->submit_time, READ);
|
|
|
|
percpu_ref_put(&ca->io_ref);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!rbio->split)
|
|
|
|
rbio->bio.bi_end_io = rbio->end_io;
|
|
|
|
|
2020-12-04 02:57:22 +08:00
|
|
|
if (bch2_dev_inum_io_err_on(bio->bi_status, ca,
|
2021-03-15 09:30:08 +08:00
|
|
|
rbio->read_pos.inode,
|
|
|
|
rbio->read_pos.offset,
|
2020-12-04 02:57:22 +08:00
|
|
|
"data read error: %s",
|
2020-07-22 01:34:22 +08:00
|
|
|
bch2_blk_status_to_str(bio->bi_status))) {
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_rbio_error(rbio, READ_RETRY_AVOID, bio->bi_status);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rbio->pick.ptr.cached &&
|
|
|
|
(((rbio->flags & BCH_READ_RETRY_IF_STALE) && race_fault()) ||
|
|
|
|
ptr_stale(ca, &rbio->pick.ptr))) {
|
|
|
|
atomic_long_inc(&c->read_realloc_races);
|
|
|
|
|
|
|
|
if (rbio->flags & BCH_READ_RETRY_IF_STALE)
|
|
|
|
bch2_rbio_error(rbio, READ_RETRY, BLK_STS_AGAIN);
|
|
|
|
else
|
|
|
|
bch2_rbio_error(rbio, READ_ERR, BLK_STS_AGAIN);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rbio->narrow_crcs ||
|
2018-02-24 05:26:10 +08:00
|
|
|
crc_is_compressed(rbio->pick.crc) ||
|
2017-03-17 14:18:50 +08:00
|
|
|
bch2_csum_type_is_encryption(rbio->pick.crc.csum_type))
|
|
|
|
context = RBIO_CONTEXT_UNBOUND, wq = system_unbound_wq;
|
|
|
|
else if (rbio->pick.crc.csum_type)
|
|
|
|
context = RBIO_CONTEXT_HIGHPRI, wq = system_highpri_wq;
|
|
|
|
|
|
|
|
bch2_rbio_punt(rbio, __bch2_read_endio, context, wq);
|
|
|
|
}
|
|
|
|
|
2019-08-23 04:07:37 +08:00
|
|
|
int __bch2_read_indirect_extent(struct btree_trans *trans,
|
|
|
|
unsigned *offset_into_extent,
|
2020-12-18 04:08:58 +08:00
|
|
|
struct bkey_buf *orig_k)
|
2019-08-16 21:59:56 +08:00
|
|
|
{
|
|
|
|
struct btree_iter *iter;
|
|
|
|
struct bkey_s_c k;
|
|
|
|
u64 reflink_offset;
|
|
|
|
int ret;
|
|
|
|
|
2020-05-22 22:50:05 +08:00
|
|
|
reflink_offset = le64_to_cpu(bkey_i_to_reflink_p(orig_k->k)->v.idx) +
|
2019-08-16 21:59:56 +08:00
|
|
|
*offset_into_extent;
|
|
|
|
|
2021-02-21 08:27:37 +08:00
|
|
|
iter = bch2_trans_get_iter(trans, BTREE_ID_reflink,
|
2019-09-27 10:21:39 +08:00
|
|
|
POS(0, reflink_offset),
|
|
|
|
BTREE_ITER_SLOTS);
|
2019-08-16 21:59:56 +08:00
|
|
|
k = bch2_btree_iter_peek_slot(iter);
|
|
|
|
ret = bkey_err(k);
|
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
|
2020-10-25 07:51:34 +08:00
|
|
|
if (k.k->type != KEY_TYPE_reflink_v &&
|
|
|
|
k.k->type != KEY_TYPE_indirect_inline_data) {
|
2020-12-04 02:57:22 +08:00
|
|
|
bch_err_inum_ratelimited(trans->c, orig_k->k->k.p.inode,
|
2019-08-16 21:59:56 +08:00
|
|
|
"pointer to nonexistent indirect extent");
|
|
|
|
ret = -EIO;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
*offset_into_extent = iter->pos.offset - bkey_start_offset(k.k);
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_reassemble(orig_k, trans->c, k);
|
2019-08-16 21:59:56 +08:00
|
|
|
err:
|
|
|
|
bch2_trans_iter_put(trans, iter);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-10-17 09:39:16 +08:00
|
|
|
int __bch2_read_extent(struct btree_trans *trans, struct bch_read_bio *orig,
|
2021-03-15 09:30:08 +08:00
|
|
|
struct bvec_iter iter, struct bpos read_pos,
|
|
|
|
enum btree_id data_btree, struct bkey_s_c k,
|
2019-07-10 00:56:43 +08:00
|
|
|
unsigned offset_into_extent,
|
2018-11-02 03:28:45 +08:00
|
|
|
struct bch_io_failures *failed, unsigned flags)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
2020-10-17 09:39:16 +08:00
|
|
|
struct bch_fs *c = trans->c;
|
2018-10-02 23:03:39 +08:00
|
|
|
struct extent_ptr_decoded pick;
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bch_read_bio *rbio = NULL;
|
|
|
|
struct bch_dev *ca;
|
|
|
|
struct promote_op *promote = NULL;
|
|
|
|
bool bounce = false, read_full = false, narrow_crcs = false;
|
2021-03-15 09:30:08 +08:00
|
|
|
struct bpos data_pos = bkey_start_pos(k.k);
|
2017-03-17 14:18:50 +08:00
|
|
|
int pick_ret;
|
|
|
|
|
2020-10-25 07:51:34 +08:00
|
|
|
if (bkey_extent_is_inline_data(k.k)) {
|
2019-11-10 05:43:16 +08:00
|
|
|
unsigned bytes = min_t(unsigned, iter.bi_size,
|
2020-10-25 07:51:34 +08:00
|
|
|
bkey_inline_data_bytes(k.k));
|
2019-11-10 05:43:16 +08:00
|
|
|
|
|
|
|
swap(iter.bi_size, bytes);
|
2020-10-25 07:51:34 +08:00
|
|
|
memcpy_to_bio(&orig->bio, iter, bkey_inline_data_p(k));
|
2019-11-10 05:43:16 +08:00
|
|
|
swap(iter.bi_size, bytes);
|
|
|
|
bio_advance_iter(&orig->bio, &iter, bytes);
|
|
|
|
zero_fill_bio_iter(&orig->bio, iter);
|
|
|
|
goto out_read_done;
|
|
|
|
}
|
|
|
|
|
2018-11-02 03:10:01 +08:00
|
|
|
pick_ret = bch2_bkey_pick_read_device(c, k, failed, &pick);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
/* hole or reservation - just zero fill: */
|
|
|
|
if (!pick_ret)
|
|
|
|
goto hole;
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (pick_ret < 0) {
|
2020-12-04 02:57:22 +08:00
|
|
|
bch_err_inum_ratelimited(c, k.k->p.inode,
|
|
|
|
"no device to read from");
|
2018-11-02 03:13:19 +08:00
|
|
|
goto err;
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (pick_ret > 0)
|
|
|
|
ca = bch_dev_bkey_exists(c, pick.ptr.dev);
|
|
|
|
|
|
|
|
if (flags & BCH_READ_NODECODE) {
|
|
|
|
/*
|
|
|
|
* can happen if we retry, and the extent we were going to read
|
|
|
|
* has been merged in the meantime:
|
|
|
|
*/
|
|
|
|
if (pick.crc.compressed_size > orig->bio.bi_vcnt * PAGE_SECTORS)
|
|
|
|
goto hole;
|
|
|
|
|
|
|
|
iter.bi_size = pick.crc.compressed_size << 9;
|
2020-01-04 05:03:10 +08:00
|
|
|
goto get_bio;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!(flags & BCH_READ_LAST_FRAGMENT) ||
|
|
|
|
bio_flagged(&orig->bio, BIO_CHAIN))
|
|
|
|
flags |= BCH_READ_MUST_CLONE;
|
|
|
|
|
2019-07-26 01:52:14 +08:00
|
|
|
narrow_crcs = !(flags & BCH_READ_IN_RETRY) &&
|
|
|
|
bch2_can_narrow_extent_crcs(k, pick.crc);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (narrow_crcs && (flags & BCH_READ_USER_MAPPED))
|
|
|
|
flags |= BCH_READ_MUST_BOUNCE;
|
|
|
|
|
2019-11-08 04:04:13 +08:00
|
|
|
EBUG_ON(offset_into_extent + bvec_iter_sectors(iter) > k.k->size);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2018-02-24 05:26:10 +08:00
|
|
|
if (crc_is_compressed(pick.crc) ||
|
2017-03-17 14:18:50 +08:00
|
|
|
(pick.crc.csum_type != BCH_CSUM_NONE &&
|
|
|
|
(bvec_iter_sectors(iter) != pick.crc.uncompressed_size ||
|
|
|
|
(bch2_csum_type_is_encryption(pick.crc.csum_type) &&
|
|
|
|
(flags & BCH_READ_USER_MAPPED)) ||
|
|
|
|
(flags & BCH_READ_MUST_BOUNCE)))) {
|
|
|
|
read_full = true;
|
|
|
|
bounce = true;
|
|
|
|
}
|
|
|
|
|
2019-11-08 04:04:13 +08:00
|
|
|
if (orig->opts.promote_target)
|
|
|
|
promote = promote_alloc(c, iter, k, &pick, orig->opts, flags,
|
|
|
|
&rbio, &bounce, &read_full);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (!read_full) {
|
2018-02-24 05:26:10 +08:00
|
|
|
EBUG_ON(crc_is_compressed(pick.crc));
|
2017-03-17 14:18:50 +08:00
|
|
|
EBUG_ON(pick.crc.csum_type &&
|
|
|
|
(bvec_iter_sectors(iter) != pick.crc.uncompressed_size ||
|
|
|
|
bvec_iter_sectors(iter) != pick.crc.live_size ||
|
|
|
|
pick.crc.offset ||
|
2019-07-10 00:56:43 +08:00
|
|
|
offset_into_extent));
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
data_pos.offset += offset_into_extent;
|
2017-03-17 14:18:50 +08:00
|
|
|
pick.ptr.offset += pick.crc.offset +
|
2019-07-10 00:56:43 +08:00
|
|
|
offset_into_extent;
|
2019-08-16 21:59:56 +08:00
|
|
|
offset_into_extent = 0;
|
2017-03-17 14:18:50 +08:00
|
|
|
pick.crc.compressed_size = bvec_iter_sectors(iter);
|
|
|
|
pick.crc.uncompressed_size = bvec_iter_sectors(iter);
|
|
|
|
pick.crc.offset = 0;
|
|
|
|
pick.crc.live_size = bvec_iter_sectors(iter);
|
2019-07-10 00:56:43 +08:00
|
|
|
offset_into_extent = 0;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
2020-01-04 05:03:10 +08:00
|
|
|
get_bio:
|
2017-03-17 14:18:50 +08:00
|
|
|
if (rbio) {
|
2019-07-09 23:16:33 +08:00
|
|
|
/*
|
|
|
|
* promote already allocated bounce rbio:
|
|
|
|
* promote needs to allocate a bio big enough for uncompressing
|
|
|
|
* data in the write path, but we're not going to use it all
|
|
|
|
* here:
|
|
|
|
*/
|
2019-11-08 04:04:13 +08:00
|
|
|
EBUG_ON(rbio->bio.bi_iter.bi_size <
|
2019-07-09 23:16:33 +08:00
|
|
|
pick.crc.compressed_size << 9);
|
|
|
|
rbio->bio.bi_iter.bi_size =
|
|
|
|
pick.crc.compressed_size << 9;
|
2017-03-17 14:18:50 +08:00
|
|
|
} else if (bounce) {
|
|
|
|
unsigned sectors = pick.crc.compressed_size;
|
|
|
|
|
|
|
|
rbio = rbio_init(bio_alloc_bioset(NULL,
|
|
|
|
DIV_ROUND_UP(sectors, PAGE_SECTORS),
|
|
|
|
0,
|
|
|
|
GFP_NOIO,
|
|
|
|
&c->bio_read_split),
|
|
|
|
orig->opts);
|
|
|
|
|
|
|
|
bch2_bio_alloc_pages_pool(c, &rbio->bio, sectors << 9);
|
|
|
|
rbio->bounce = true;
|
|
|
|
rbio->split = true;
|
|
|
|
} else if (flags & BCH_READ_MUST_CLONE) {
|
|
|
|
/*
|
|
|
|
* Have to clone if there were any splits, due to error
|
|
|
|
* reporting issues (if a split errored, and retrying didn't
|
|
|
|
* work, when it reports the error to its parent (us) we don't
|
|
|
|
* know if the error was from our bio, and we should retry, or
|
|
|
|
* from the whole bio, in which case we don't want to retry and
|
|
|
|
* lose the error)
|
|
|
|
*/
|
|
|
|
rbio = rbio_init(bio_alloc_clone(NULL, &orig->bio, GFP_NOIO,
|
|
|
|
&c->bio_read_split),
|
|
|
|
orig->opts);
|
|
|
|
rbio->bio.bi_iter = iter;
|
|
|
|
rbio->split = true;
|
|
|
|
} else {
|
|
|
|
rbio = orig;
|
|
|
|
rbio->bio.bi_iter = iter;
|
2019-11-08 04:04:13 +08:00
|
|
|
EBUG_ON(bio_flagged(&rbio->bio, BIO_CHAIN));
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2019-11-08 04:04:13 +08:00
|
|
|
EBUG_ON(bio_sectors(&rbio->bio) != pick.crc.compressed_size);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
rbio->c = c;
|
|
|
|
rbio->submit_time = local_clock();
|
|
|
|
if (rbio->split)
|
|
|
|
rbio->parent = orig;
|
|
|
|
else
|
|
|
|
rbio->end_io = orig->bio.bi_end_io;
|
|
|
|
rbio->bvec_iter = iter;
|
2019-07-10 00:56:43 +08:00
|
|
|
rbio->offset_into_extent= offset_into_extent;
|
2017-03-17 14:18:50 +08:00
|
|
|
rbio->flags = flags;
|
|
|
|
rbio->have_ioref = pick_ret > 0 && bch2_dev_get_ioref(ca, READ);
|
|
|
|
rbio->narrow_crcs = narrow_crcs;
|
|
|
|
rbio->hole = 0;
|
|
|
|
rbio->retry = 0;
|
|
|
|
rbio->context = 0;
|
2019-11-08 04:04:13 +08:00
|
|
|
/* XXX: only initialize this if needed */
|
2017-03-17 14:18:50 +08:00
|
|
|
rbio->devs_have = bch2_bkey_devs(k);
|
|
|
|
rbio->pick = pick;
|
2021-03-15 09:30:08 +08:00
|
|
|
rbio->read_pos = read_pos;
|
|
|
|
rbio->data_btree = data_btree;
|
|
|
|
rbio->data_pos = data_pos;
|
2017-03-17 14:18:50 +08:00
|
|
|
rbio->version = k.k->version;
|
|
|
|
rbio->promote = promote;
|
|
|
|
INIT_WORK(&rbio->work, NULL);
|
|
|
|
|
|
|
|
rbio->bio.bi_opf = orig->bio.bi_opf;
|
|
|
|
rbio->bio.bi_iter.bi_sector = pick.ptr.offset;
|
|
|
|
rbio->bio.bi_end_io = bch2_read_endio;
|
|
|
|
|
|
|
|
if (rbio->bounce)
|
|
|
|
trace_read_bounce(&rbio->bio);
|
|
|
|
|
|
|
|
bch2_increment_clock(c, bio_sectors(&rbio->bio), READ);
|
|
|
|
|
2020-12-21 10:42:19 +08:00
|
|
|
/*
|
|
|
|
* If it's being moved internally, we don't want to flag it as a cache
|
|
|
|
* hit:
|
|
|
|
*/
|
|
|
|
if (pick.ptr.cached && !(flags & BCH_READ_NODECODE))
|
2020-10-17 09:39:16 +08:00
|
|
|
bch2_bucket_io_time_reset(trans, pick.ptr.dev,
|
|
|
|
PTR_BUCKET_NR(ca, &pick.ptr), READ);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-11-08 04:04:13 +08:00
|
|
|
if (!(flags & (BCH_READ_IN_RETRY|BCH_READ_LAST_FRAGMENT))) {
|
2018-11-02 03:13:19 +08:00
|
|
|
bio_inc_remaining(&orig->bio);
|
|
|
|
trace_read_split(&orig->bio);
|
|
|
|
}
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (!rbio->pick.idx) {
|
|
|
|
if (!rbio->have_ioref) {
|
2020-12-04 02:57:22 +08:00
|
|
|
bch_err_inum_ratelimited(c, k.k->p.inode,
|
|
|
|
"no device to read from");
|
2018-11-02 03:13:19 +08:00
|
|
|
bch2_rbio_error(rbio, READ_RETRY_AVOID, BLK_STS_IOERR);
|
|
|
|
goto out;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2020-07-10 06:28:11 +08:00
|
|
|
this_cpu_add(ca->io_done->sectors[READ][BCH_DATA_user],
|
2018-11-02 03:13:19 +08:00
|
|
|
bio_sectors(&rbio->bio));
|
|
|
|
bio_set_dev(&rbio->bio, ca->disk_sb.bdev);
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
if (unlikely(c->opts.no_data_io)) {
|
2018-11-02 03:13:19 +08:00
|
|
|
if (likely(!(flags & BCH_READ_IN_RETRY)))
|
|
|
|
bio_endio(&rbio->bio);
|
|
|
|
} else {
|
|
|
|
if (likely(!(flags & BCH_READ_IN_RETRY)))
|
|
|
|
submit_bio(&rbio->bio);
|
|
|
|
else
|
|
|
|
submit_bio_wait(&rbio->bio);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* Attempting reconstruct read: */
|
|
|
|
if (bch2_ec_read_extent(c, rbio)) {
|
|
|
|
bch2_rbio_error(rbio, READ_RETRY_AVOID, BLK_STS_IOERR);
|
|
|
|
goto out;
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
if (likely(!(flags & BCH_READ_IN_RETRY)))
|
|
|
|
bio_endio(&rbio->bio);
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
if (likely(!(flags & BCH_READ_IN_RETRY))) {
|
2017-03-17 14:18:50 +08:00
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
rbio->context = RBIO_CONTEXT_UNBOUND;
|
|
|
|
bch2_read_endio(&rbio->bio);
|
|
|
|
|
|
|
|
ret = rbio->retry;
|
|
|
|
rbio = bch2_rbio_free(rbio);
|
|
|
|
|
|
|
|
if (ret == READ_RETRY_AVOID) {
|
2018-11-02 03:28:45 +08:00
|
|
|
bch2_mark_io_failure(failed, &pick);
|
2017-03-17 14:18:50 +08:00
|
|
|
ret = READ_RETRY;
|
|
|
|
}
|
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
if (!ret)
|
|
|
|
goto out_read_done;
|
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-11-02 03:13:19 +08:00
|
|
|
err:
|
|
|
|
if (flags & BCH_READ_IN_RETRY)
|
2017-03-17 14:18:50 +08:00
|
|
|
return READ_ERR;
|
2018-11-02 03:13:19 +08:00
|
|
|
|
|
|
|
orig->bio.bi_status = BLK_STS_IOERR;
|
|
|
|
goto out_read_done;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
hole:
|
|
|
|
/*
|
|
|
|
* won't normally happen in the BCH_READ_NODECODE
|
|
|
|
* (bch2_move_extent()) path, but if we retry and the extent we wanted
|
|
|
|
* to read no longer exists we have to signal that:
|
|
|
|
*/
|
|
|
|
if (flags & BCH_READ_NODECODE)
|
|
|
|
orig->hole = true;
|
|
|
|
|
|
|
|
zero_fill_bio_iter(&orig->bio, iter);
|
2018-11-02 03:13:19 +08:00
|
|
|
out_read_done:
|
2017-03-17 14:18:50 +08:00
|
|
|
if (flags & BCH_READ_LAST_FRAGMENT)
|
|
|
|
bch2_rbio_done(orig);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
void __bch2_read(struct bch_fs *c, struct bch_read_bio *rbio,
|
|
|
|
struct bvec_iter bvec_iter, u64 inode,
|
|
|
|
struct bch_io_failures *failed, unsigned flags)
|
2017-03-17 14:18:50 +08:00
|
|
|
{
|
2019-03-26 03:10:15 +08:00
|
|
|
struct btree_trans trans;
|
|
|
|
struct btree_iter *iter;
|
2020-12-18 04:08:58 +08:00
|
|
|
struct bkey_buf sk;
|
2017-03-17 14:18:50 +08:00
|
|
|
struct bkey_s_c k;
|
2019-04-18 03:49:28 +08:00
|
|
|
int ret;
|
2019-03-26 03:10:15 +08:00
|
|
|
|
2017-03-17 14:18:50 +08:00
|
|
|
BUG_ON(flags & BCH_READ_NODECODE);
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_init(&sk);
|
2019-09-27 10:21:39 +08:00
|
|
|
bch2_trans_init(&trans, c, 0, 0);
|
|
|
|
retry:
|
|
|
|
bch2_trans_begin(&trans);
|
|
|
|
|
2021-02-21 08:27:37 +08:00
|
|
|
iter = bch2_trans_get_iter(&trans, BTREE_ID_extents,
|
2021-03-13 09:29:28 +08:00
|
|
|
POS(inode, bvec_iter.bi_sector),
|
2019-08-16 21:59:56 +08:00
|
|
|
BTREE_ITER_SLOTS);
|
|
|
|
while (1) {
|
|
|
|
unsigned bytes, sectors, offset_into_extent;
|
2021-03-15 09:30:08 +08:00
|
|
|
enum btree_id data_btree = BTREE_ID_extents;
|
2019-08-16 21:59:56 +08:00
|
|
|
|
|
|
|
bch2_btree_iter_set_pos(iter,
|
2021-03-13 09:29:28 +08:00
|
|
|
POS(inode, bvec_iter.bi_sector));
|
2019-08-16 21:59:56 +08:00
|
|
|
|
|
|
|
k = bch2_btree_iter_peek_slot(iter);
|
|
|
|
ret = bkey_err(k);
|
|
|
|
if (ret)
|
2021-03-20 08:29:11 +08:00
|
|
|
break;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2019-07-10 00:56:43 +08:00
|
|
|
offset_into_extent = iter->pos.offset -
|
|
|
|
bkey_start_offset(k.k);
|
2019-08-16 21:59:56 +08:00
|
|
|
sectors = k.k->size - offset_into_extent;
|
|
|
|
|
2020-12-18 04:08:58 +08:00
|
|
|
bch2_bkey_buf_reassemble(&sk, c, k);
|
2019-11-10 05:01:15 +08:00
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
ret = bch2_read_indirect_extent(&trans, &data_btree,
|
2020-05-22 22:50:05 +08:00
|
|
|
&offset_into_extent, &sk);
|
2019-08-16 21:59:56 +08:00
|
|
|
if (ret)
|
2021-03-20 08:29:11 +08:00
|
|
|
break;
|
2019-08-16 21:59:56 +08:00
|
|
|
|
2020-10-25 08:56:47 +08:00
|
|
|
k = bkey_i_to_s_c(sk.k);
|
|
|
|
|
2019-08-16 21:59:56 +08:00
|
|
|
/*
|
|
|
|
* With indirect extents, the amount of data to read is the min
|
|
|
|
* of the original extent and the indirect extent:
|
|
|
|
*/
|
|
|
|
sectors = min(sectors, k.k->size - offset_into_extent);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Unlock the iterator while the btree node's lock is still in
|
|
|
|
* cache, before doing the IO:
|
|
|
|
*/
|
|
|
|
bch2_trans_unlock(&trans);
|
2019-07-10 00:56:43 +08:00
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
bytes = min(sectors, bvec_iter_sectors(bvec_iter)) << 9;
|
|
|
|
swap(bvec_iter.bi_size, bytes);
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
if (bvec_iter.bi_size == bytes)
|
2017-03-17 14:18:50 +08:00
|
|
|
flags |= BCH_READ_LAST_FRAGMENT;
|
|
|
|
|
2021-03-15 09:30:08 +08:00
|
|
|
ret = __bch2_read_extent(&trans, rbio, bvec_iter, iter->pos,
|
|
|
|
data_btree, k,
|
2021-03-13 09:29:28 +08:00
|
|
|
offset_into_extent, failed, flags);
|
2021-03-20 08:29:11 +08:00
|
|
|
if (ret)
|
|
|
|
break;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
|
|
|
if (flags & BCH_READ_LAST_FRAGMENT)
|
2019-08-16 21:59:56 +08:00
|
|
|
break;
|
2017-03-17 14:18:50 +08:00
|
|
|
|
2021-03-13 09:29:28 +08:00
|
|
|
swap(bvec_iter.bi_size, bytes);
|
|
|
|
bio_advance_iter(&rbio->bio, &bvec_iter, bytes);
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
2021-03-20 08:29:11 +08:00
|
|
|
bch2_trans_iter_put(&trans, iter);
|
|
|
|
|
|
|
|
if (ret == -EINTR || ret == READ_RETRY || ret == READ_RETRY_AVOID)
|
2019-09-27 10:21:39 +08:00
|
|
|
goto retry;
|
|
|
|
|
2021-03-20 08:29:11 +08:00
|
|
|
if (ret) {
|
|
|
|
bch_err_inum_ratelimited(c, inode,
|
|
|
|
"read error %i from btree lookup", ret);
|
|
|
|
rbio->bio.bi_status = BLK_STS_IOERR;
|
|
|
|
bch2_rbio_done(rbio);
|
|
|
|
}
|
|
|
|
bch2_trans_exit(&trans);
|
|
|
|
bch2_bkey_buf_exit(&sk, c);
|
2017-03-17 14:18:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void bch2_fs_io_exit(struct bch_fs *c)
|
|
|
|
{
|
|
|
|
if (c->promote_table.tbl)
|
|
|
|
rhashtable_destroy(&c->promote_table);
|
|
|
|
mempool_exit(&c->bio_bounce_pages);
|
|
|
|
bioset_exit(&c->bio_write);
|
|
|
|
bioset_exit(&c->bio_read_split);
|
|
|
|
bioset_exit(&c->bio_read);
|
|
|
|
}
|
|
|
|
|
|
|
|
int bch2_fs_io_init(struct bch_fs *c)
|
|
|
|
{
|
|
|
|
if (bioset_init(&c->bio_read, 1, offsetof(struct bch_read_bio, bio),
|
|
|
|
BIOSET_NEED_BVECS) ||
|
|
|
|
bioset_init(&c->bio_read_split, 1, offsetof(struct bch_read_bio, bio),
|
|
|
|
BIOSET_NEED_BVECS) ||
|
|
|
|
bioset_init(&c->bio_write, 1, offsetof(struct bch_write_bio, bio),
|
|
|
|
BIOSET_NEED_BVECS) ||
|
|
|
|
mempool_init_page_pool(&c->bio_bounce_pages,
|
|
|
|
max_t(unsigned,
|
|
|
|
c->opts.btree_node_size,
|
|
|
|
c->sb.encoded_extent_max) /
|
|
|
|
PAGE_SECTORS, 0) ||
|
|
|
|
rhashtable_init(&c->promote_table, &bch_promote_params))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|