mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-19 02:34:01 +08:00
bcache: Move keylist out of btree_op
Slowly working on pruning struct btree_op - the aim is for it to only contain things that are actually necessary for traversing the btree. Signed-off-by: Kent Overstreet <kmo@daterainc.com>
This commit is contained in:
parent
a34a8bfd4e
commit
0b93207abb
@ -113,7 +113,6 @@ void bch_btree_op_init_stack(struct btree_op *op)
|
||||
memset(op, 0, sizeof(struct btree_op));
|
||||
closure_init_stack(&op->cl);
|
||||
op->lock = -1;
|
||||
bch_keylist_init(&op->keys);
|
||||
}
|
||||
|
||||
/* Btree key manipulation */
|
||||
|
@ -267,10 +267,6 @@ struct btree_op {
|
||||
unsigned lookup_done:1;
|
||||
unsigned insert_collision:1;
|
||||
|
||||
/* Anything after this point won't get zeroed in do_bio_hook() */
|
||||
|
||||
/* Keys to be inserted */
|
||||
struct keylist keys;
|
||||
BKEY_PADDED(replace);
|
||||
};
|
||||
|
||||
|
@ -301,6 +301,9 @@ int bch_journal_replay(struct cache_set *s, struct list_head *list,
|
||||
list_entry(list->prev, struct journal_replay, list);
|
||||
|
||||
uint64_t start = i->j.last_seq, end = i->j.seq, n = start;
|
||||
struct keylist keylist;
|
||||
|
||||
bch_keylist_init(&keylist);
|
||||
|
||||
list_for_each_entry(i, list, list) {
|
||||
BUG_ON(i->pin && atomic_read(i->pin) != 1);
|
||||
@ -314,16 +317,16 @@ int bch_journal_replay(struct cache_set *s, struct list_head *list,
|
||||
k = bkey_next(k)) {
|
||||
trace_bcache_journal_replay_key(k);
|
||||
|
||||
bkey_copy(op->keys.top, k);
|
||||
bch_keylist_push(&op->keys);
|
||||
bkey_copy(keylist.top, k);
|
||||
bch_keylist_push(&keylist);
|
||||
|
||||
op->journal = i->pin;
|
||||
|
||||
ret = bch_btree_insert(op, s, &op->keys);
|
||||
ret = bch_btree_insert(op, s, &keylist);
|
||||
if (ret)
|
||||
goto err;
|
||||
|
||||
BUG_ON(!bch_keylist_empty(&op->keys));
|
||||
BUG_ON(!bch_keylist_empty(&keylist));
|
||||
keys++;
|
||||
|
||||
cond_resched();
|
||||
|
@ -231,11 +231,11 @@ static void bch_data_insert_keys(struct closure *cl)
|
||||
#endif
|
||||
|
||||
if (s->write)
|
||||
op->journal = bch_journal(op->c, &op->keys,
|
||||
op->journal = bch_journal(op->c, &s->insert_keys,
|
||||
op->flush_journal
|
||||
? &s->cl : NULL);
|
||||
|
||||
if (bch_btree_insert(op, op->c, &op->keys)) {
|
||||
if (bch_btree_insert(op, op->c, &s->insert_keys)) {
|
||||
s->error = -ENOMEM;
|
||||
op->insert_data_done = true;
|
||||
}
|
||||
@ -247,7 +247,7 @@ static void bch_data_insert_keys(struct closure *cl)
|
||||
if (!op->insert_data_done)
|
||||
continue_at(cl, bch_data_insert_start, bcache_wq);
|
||||
|
||||
bch_keylist_free(&op->keys);
|
||||
bch_keylist_free(&s->insert_keys);
|
||||
closure_return(cl);
|
||||
}
|
||||
|
||||
@ -439,6 +439,7 @@ static bool bch_alloc_sectors(struct bkey *k, unsigned sectors,
|
||||
static void bch_data_invalidate(struct closure *cl)
|
||||
{
|
||||
struct btree_op *op = container_of(cl, struct btree_op, cl);
|
||||
struct search *s = container_of(op, struct search, op);
|
||||
struct bio *bio = op->cache_bio;
|
||||
|
||||
pr_debug("invalidating %i sectors from %llu",
|
||||
@ -447,14 +448,14 @@ static void bch_data_invalidate(struct closure *cl)
|
||||
while (bio_sectors(bio)) {
|
||||
unsigned len = min(bio_sectors(bio), 1U << 14);
|
||||
|
||||
if (bch_keylist_realloc(&op->keys, 0, op->c))
|
||||
if (bch_keylist_realloc(&s->insert_keys, 0, op->c))
|
||||
goto out;
|
||||
|
||||
bio->bi_sector += len;
|
||||
bio->bi_size -= len << 9;
|
||||
|
||||
bch_keylist_add(&op->keys, &KEY(op->inode,
|
||||
bio->bi_sector, len));
|
||||
bch_keylist_add(&s->insert_keys,
|
||||
&KEY(op->inode, bio->bi_sector, len));
|
||||
}
|
||||
|
||||
op->insert_data_done = true;
|
||||
@ -466,6 +467,7 @@ out:
|
||||
static void bch_data_insert_error(struct closure *cl)
|
||||
{
|
||||
struct btree_op *op = container_of(cl, struct btree_op, cl);
|
||||
struct search *s = container_of(op, struct search, op);
|
||||
|
||||
/*
|
||||
* Our data write just errored, which means we've got a bunch of keys to
|
||||
@ -476,9 +478,9 @@ static void bch_data_insert_error(struct closure *cl)
|
||||
* from the keys we'll accomplish just that.
|
||||
*/
|
||||
|
||||
struct bkey *src = op->keys.keys, *dst = op->keys.keys;
|
||||
struct bkey *src = s->insert_keys.keys, *dst = s->insert_keys.keys;
|
||||
|
||||
while (src != op->keys.top) {
|
||||
while (src != s->insert_keys.top) {
|
||||
struct bkey *n = bkey_next(src);
|
||||
|
||||
SET_KEY_PTRS(src, 0);
|
||||
@ -488,7 +490,7 @@ static void bch_data_insert_error(struct closure *cl)
|
||||
src = n;
|
||||
}
|
||||
|
||||
op->keys.top = dst;
|
||||
s->insert_keys.top = dst;
|
||||
|
||||
bch_data_insert_keys(cl);
|
||||
}
|
||||
@ -539,12 +541,12 @@ static void bch_data_insert_start(struct closure *cl)
|
||||
? s->d->bio_split : op->c->bio_split;
|
||||
|
||||
/* 1 for the device pointer and 1 for the chksum */
|
||||
if (bch_keylist_realloc(&op->keys,
|
||||
if (bch_keylist_realloc(&s->insert_keys,
|
||||
1 + (op->csum ? 1 : 0),
|
||||
op->c))
|
||||
continue_at(cl, bch_data_insert_keys, bcache_wq);
|
||||
|
||||
k = op->keys.top;
|
||||
k = s->insert_keys.top;
|
||||
bkey_init(k);
|
||||
SET_KEY_INODE(k, op->inode);
|
||||
SET_KEY_OFFSET(k, bio->bi_sector);
|
||||
@ -570,7 +572,7 @@ static void bch_data_insert_start(struct closure *cl)
|
||||
bio_csum(n, k);
|
||||
|
||||
trace_bcache_cache_insert(k);
|
||||
bch_keylist_push(&op->keys);
|
||||
bch_keylist_push(&s->insert_keys);
|
||||
|
||||
n->bi_rw |= REQ_WRITE;
|
||||
bch_submit_bbio(n, op->c, k, 0);
|
||||
@ -605,7 +607,7 @@ err:
|
||||
op->insert_data_done = true;
|
||||
bio_put(bio);
|
||||
|
||||
if (!bch_keylist_empty(&op->keys))
|
||||
if (!bch_keylist_empty(&s->insert_keys))
|
||||
continue_at(cl, bch_data_insert_keys, bcache_wq);
|
||||
else
|
||||
closure_return(cl);
|
||||
@ -634,8 +636,9 @@ err:
|
||||
void bch_data_insert(struct closure *cl)
|
||||
{
|
||||
struct btree_op *op = container_of(cl, struct btree_op, cl);
|
||||
struct search *s = container_of(op, struct search, op);
|
||||
|
||||
bch_keylist_init(&op->keys);
|
||||
bch_keylist_init(&s->insert_keys);
|
||||
bio_get(op->cache_bio);
|
||||
bch_data_insert_start(cl);
|
||||
}
|
||||
@ -724,9 +727,11 @@ static void search_free(struct closure *cl)
|
||||
|
||||
static struct search *search_alloc(struct bio *bio, struct bcache_device *d)
|
||||
{
|
||||
struct search *s;
|
||||
struct bio_vec *bv;
|
||||
struct search *s = mempool_alloc(d->c->search, GFP_NOIO);
|
||||
memset(s, 0, offsetof(struct search, op.keys));
|
||||
|
||||
s = mempool_alloc(d->c->search, GFP_NOIO);
|
||||
memset(s, 0, offsetof(struct search, insert_keys));
|
||||
|
||||
__closure_init(&s->cl, NULL);
|
||||
|
||||
|
@ -25,8 +25,10 @@ struct search {
|
||||
short error;
|
||||
unsigned long start_time;
|
||||
|
||||
/* Anything past op->keys won't get zeroed in do_bio_hook */
|
||||
struct btree_op op;
|
||||
|
||||
/* Anything past this point won't get zeroed in search_alloc() */
|
||||
struct keylist insert_keys;
|
||||
};
|
||||
|
||||
void bch_cache_read_endio(struct bio *, int);
|
||||
|
@ -300,18 +300,21 @@ static void write_dirty_finish(struct closure *cl)
|
||||
if (KEY_DIRTY(&w->key)) {
|
||||
unsigned i;
|
||||
struct btree_op op;
|
||||
struct keylist keys;
|
||||
|
||||
bch_btree_op_init_stack(&op);
|
||||
bch_keylist_init(&keys);
|
||||
|
||||
op.type = BTREE_REPLACE;
|
||||
bkey_copy(&op.replace, &w->key);
|
||||
|
||||
SET_KEY_DIRTY(&w->key, false);
|
||||
bch_keylist_add(&op.keys, &w->key);
|
||||
bch_keylist_add(&keys, &w->key);
|
||||
|
||||
for (i = 0; i < KEY_PTRS(&w->key); i++)
|
||||
atomic_inc(&PTR_BUCKET(dc->disk.c, &w->key, i)->pin);
|
||||
|
||||
bch_btree_insert(&op, dc->disk.c, &op.keys);
|
||||
bch_btree_insert(&op, dc->disk.c, &keys);
|
||||
closure_sync(&op.cl);
|
||||
|
||||
if (op.insert_collision)
|
||||
|
Loading…
Reference in New Issue
Block a user