License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 22:07:57 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2018-04-04 01:23:33 +08:00
|
|
|
|
2008-01-25 05:13:08 +08:00
|
|
|
#include <linux/err.h>
|
|
|
|
#include <linux/slab.h>
|
2007-08-28 04:49:44 +08:00
|
|
|
#include <linux/spinlock.h>
|
2022-10-19 22:50:49 +08:00
|
|
|
#include "messages.h"
|
2010-12-17 14:21:50 +08:00
|
|
|
#include "ctree.h"
|
2019-03-27 20:24:12 +08:00
|
|
|
#include "volumes.h"
|
2007-08-28 04:49:44 +08:00
|
|
|
#include "extent_map.h"
|
2016-03-10 17:26:59 +08:00
|
|
|
#include "compression.h"
|
2022-09-19 22:06:29 +08:00
|
|
|
#include "btrfs_inode.h"
|
2007-08-28 04:49:44 +08:00
|
|
|
|
2007-09-11 07:58:16 +08:00
|
|
|
|
2007-08-28 04:49:44 +08:00
|
|
|
static struct kmem_cache *extent_map_cache;
|
2007-11-28 00:16:35 +08:00
|
|
|
|
2007-11-19 23:22:33 +08:00
|
|
|
int __init extent_map_init(void)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2012-09-07 17:00:48 +08:00
|
|
|
extent_map_cache = kmem_cache_create("btrfs_extent_map",
|
2009-04-13 21:33:09 +08:00
|
|
|
sizeof(struct extent_map), 0,
|
2016-06-24 02:17:08 +08:00
|
|
|
SLAB_MEM_SPREAD, NULL);
|
2007-11-19 23:22:33 +08:00
|
|
|
if (!extent_map_cache)
|
|
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
|
|
|
|
2018-02-20 00:24:18 +08:00
|
|
|
void __cold extent_map_exit(void)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2016-01-29 21:36:35 +08:00
|
|
|
kmem_cache_destroy(extent_map_cache);
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Initialize the extent tree @tree. Should be called for each new inode or
|
|
|
|
* other user of the extent_map interface.
|
2008-06-12 09:52:17 +08:00
|
|
|
*/
|
2011-04-21 06:34:43 +08:00
|
|
|
void extent_map_tree_init(struct extent_map_tree *tree)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2018-08-23 03:51:52 +08:00
|
|
|
tree->map = RB_ROOT_CACHED;
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
INIT_LIST_HEAD(&tree->modified_extents);
|
2009-09-03 04:24:52 +08:00
|
|
|
rwlock_init(&tree->lock);
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Allocate a new extent_map structure. The new structure is returned with a
|
|
|
|
* reference count of one and needs to be freed using free_extent_map()
|
2008-06-12 09:52:17 +08:00
|
|
|
*/
|
2011-04-21 06:48:27 +08:00
|
|
|
struct extent_map *alloc_extent_map(void)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
|
|
|
struct extent_map *em;
|
2012-10-12 04:54:30 +08:00
|
|
|
em = kmem_cache_zalloc(extent_map_cache, GFP_NOFS);
|
2011-02-14 08:45:29 +08:00
|
|
|
if (!em)
|
|
|
|
return NULL;
|
2014-02-25 22:15:12 +08:00
|
|
|
RB_CLEAR_NODE(&em->rb_node);
|
2010-12-17 14:21:50 +08:00
|
|
|
em->compress_type = BTRFS_COMPRESS_NONE;
|
2017-03-03 16:55:12 +08:00
|
|
|
refcount_set(&em->refs, 1);
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
INIT_LIST_HEAD(&em->list);
|
2007-08-28 04:49:44 +08:00
|
|
|
return em;
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Drop the reference out on @em by one and free the structure if the reference
|
|
|
|
* count hits zero.
|
2008-06-12 09:52:17 +08:00
|
|
|
*/
|
2007-08-28 04:49:44 +08:00
|
|
|
void free_extent_map(struct extent_map *em)
|
|
|
|
{
|
2007-08-30 23:54:02 +08:00
|
|
|
if (!em)
|
|
|
|
return;
|
2017-03-03 16:55:12 +08:00
|
|
|
if (refcount_dec_and_test(&em->refs)) {
|
2014-02-25 22:15:12 +08:00
|
|
|
WARN_ON(extent_map_in_tree(em));
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
WARN_ON(!list_empty(&em->list));
|
2014-06-19 10:42:52 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags))
|
2015-06-03 22:55:48 +08:00
|
|
|
kfree(em->map_lookup);
|
2007-08-28 04:49:44 +08:00
|
|
|
kmem_cache_free(extent_map_cache, em);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/* Do the math around the end of an extent, handling wrapping. */
|
2013-11-25 11:23:51 +08:00
|
|
|
static u64 range_end(u64 start, u64 len)
|
|
|
|
{
|
|
|
|
if (start + len < start)
|
|
|
|
return (u64)-1;
|
|
|
|
return start + len;
|
|
|
|
}
|
|
|
|
|
2018-08-23 03:51:52 +08:00
|
|
|
static int tree_insert(struct rb_root_cached *root, struct extent_map *em)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2018-08-23 03:51:52 +08:00
|
|
|
struct rb_node **p = &root->rb_root.rb_node;
|
2009-01-06 10:25:51 +08:00
|
|
|
struct rb_node *parent = NULL;
|
2013-11-25 11:23:51 +08:00
|
|
|
struct extent_map *entry = NULL;
|
|
|
|
struct rb_node *orig_parent = NULL;
|
|
|
|
u64 end = range_end(em->start, em->len);
|
2018-08-23 03:51:52 +08:00
|
|
|
bool leftmost = true;
|
2007-08-28 04:49:44 +08:00
|
|
|
|
2009-01-06 10:25:51 +08:00
|
|
|
while (*p) {
|
2007-08-28 04:49:44 +08:00
|
|
|
parent = *p;
|
2008-01-25 05:13:08 +08:00
|
|
|
entry = rb_entry(parent, struct extent_map, rb_node);
|
|
|
|
|
2018-08-23 03:51:52 +08:00
|
|
|
if (em->start < entry->start) {
|
2007-08-28 04:49:44 +08:00
|
|
|
p = &(*p)->rb_left;
|
2018-08-23 03:51:52 +08:00
|
|
|
} else if (em->start >= extent_map_end(entry)) {
|
2007-08-28 04:49:44 +08:00
|
|
|
p = &(*p)->rb_right;
|
2018-08-23 03:51:52 +08:00
|
|
|
leftmost = false;
|
|
|
|
} else {
|
2013-11-25 11:23:51 +08:00
|
|
|
return -EEXIST;
|
2018-08-23 03:51:52 +08:00
|
|
|
}
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
|
|
|
|
2013-11-25 11:23:51 +08:00
|
|
|
orig_parent = parent;
|
|
|
|
while (parent && em->start >= extent_map_end(entry)) {
|
|
|
|
parent = rb_next(parent);
|
|
|
|
entry = rb_entry(parent, struct extent_map, rb_node);
|
|
|
|
}
|
|
|
|
if (parent)
|
|
|
|
if (end > entry->start && em->start < extent_map_end(entry))
|
|
|
|
return -EEXIST;
|
|
|
|
|
|
|
|
parent = orig_parent;
|
|
|
|
entry = rb_entry(parent, struct extent_map, rb_node);
|
|
|
|
while (parent && em->start < entry->start) {
|
|
|
|
parent = rb_prev(parent);
|
|
|
|
entry = rb_entry(parent, struct extent_map, rb_node);
|
|
|
|
}
|
|
|
|
if (parent)
|
|
|
|
if (end > entry->start && em->start < extent_map_end(entry))
|
|
|
|
return -EEXIST;
|
|
|
|
|
|
|
|
rb_link_node(&em->rb_node, orig_parent, p);
|
2018-08-23 03:51:52 +08:00
|
|
|
rb_insert_color_cached(&em->rb_node, root, leftmost);
|
2013-11-25 11:23:51 +08:00
|
|
|
return 0;
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
|
|
|
|
2008-09-30 03:18:18 +08:00
|
|
|
/*
|
2022-10-27 20:21:42 +08:00
|
|
|
* Search through the tree for an extent_map with a given offset. If it can't
|
|
|
|
* be found, try to find some neighboring extents
|
2008-09-30 03:18:18 +08:00
|
|
|
*/
|
2007-08-28 04:49:44 +08:00
|
|
|
static struct rb_node *__tree_search(struct rb_root *root, u64 offset,
|
2022-09-19 22:06:38 +08:00
|
|
|
struct rb_node **prev_or_next_ret)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2009-01-06 10:25:51 +08:00
|
|
|
struct rb_node *n = root->rb_node;
|
2007-08-28 04:49:44 +08:00
|
|
|
struct rb_node *prev = NULL;
|
2008-01-23 05:47:59 +08:00
|
|
|
struct rb_node *orig_prev = NULL;
|
2008-01-25 05:13:08 +08:00
|
|
|
struct extent_map *entry;
|
|
|
|
struct extent_map *prev_entry = NULL;
|
2007-08-28 04:49:44 +08:00
|
|
|
|
2022-09-19 22:06:38 +08:00
|
|
|
ASSERT(prev_or_next_ret);
|
2022-09-19 22:06:37 +08:00
|
|
|
|
2009-01-06 10:25:51 +08:00
|
|
|
while (n) {
|
2008-01-25 05:13:08 +08:00
|
|
|
entry = rb_entry(n, struct extent_map, rb_node);
|
2007-08-28 04:49:44 +08:00
|
|
|
prev = n;
|
|
|
|
prev_entry = entry;
|
|
|
|
|
|
|
|
if (offset < entry->start)
|
|
|
|
n = n->rb_left;
|
2008-01-25 05:13:08 +08:00
|
|
|
else if (offset >= extent_map_end(entry))
|
2007-08-28 04:49:44 +08:00
|
|
|
n = n->rb_right;
|
|
|
|
else
|
|
|
|
return n;
|
|
|
|
}
|
2008-01-23 05:47:59 +08:00
|
|
|
|
2022-09-19 22:06:37 +08:00
|
|
|
orig_prev = prev;
|
|
|
|
while (prev && offset >= extent_map_end(prev_entry)) {
|
|
|
|
prev = rb_next(prev);
|
|
|
|
prev_entry = rb_entry(prev, struct extent_map, rb_node);
|
2008-01-23 05:47:59 +08:00
|
|
|
}
|
|
|
|
|
2022-09-19 22:06:38 +08:00
|
|
|
/*
|
|
|
|
* Previous extent map found, return as in this case the caller does not
|
|
|
|
* care about the next one.
|
|
|
|
*/
|
|
|
|
if (prev) {
|
|
|
|
*prev_or_next_ret = prev;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
prev = orig_prev;
|
2022-09-19 22:06:37 +08:00
|
|
|
prev_entry = rb_entry(prev, struct extent_map, rb_node);
|
|
|
|
while (prev && offset < prev_entry->start) {
|
|
|
|
prev = rb_prev(prev);
|
2008-01-25 05:13:08 +08:00
|
|
|
prev_entry = rb_entry(prev, struct extent_map, rb_node);
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
2022-09-19 22:06:38 +08:00
|
|
|
*prev_or_next_ret = prev;
|
2022-09-19 22:06:37 +08:00
|
|
|
|
2007-08-28 04:49:44 +08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/* Check to see if two extent_map structs are adjacent and safe to merge. */
|
2008-01-25 05:13:08 +08:00
|
|
|
static int mergable_maps(struct extent_map *prev, struct extent_map *next)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2008-07-19 00:01:11 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_PINNED, &prev->flags))
|
|
|
|
return 0;
|
|
|
|
|
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-30 02:49:59 +08:00
|
|
|
/*
|
|
|
|
* don't merge compressed extents, we need to know their
|
|
|
|
* actual size
|
|
|
|
*/
|
|
|
|
if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags))
|
|
|
|
return 0;
|
|
|
|
|
2013-01-25 01:02:07 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) ||
|
|
|
|
test_bit(EXTENT_FLAG_LOGGING, &next->flags))
|
|
|
|
return 0;
|
|
|
|
|
2013-04-06 04:51:15 +08:00
|
|
|
/*
|
|
|
|
* We don't want to merge stuff that hasn't been written to the log yet
|
|
|
|
* since it may not reflect exactly what is on disk, and that would be
|
|
|
|
* bad.
|
|
|
|
*/
|
|
|
|
if (!list_empty(&prev->list) || !list_empty(&next->list))
|
|
|
|
return 0;
|
|
|
|
|
2019-01-08 22:53:46 +08:00
|
|
|
ASSERT(next->block_start != EXTENT_MAP_DELALLOC &&
|
|
|
|
prev->block_start != EXTENT_MAP_DELALLOC);
|
|
|
|
|
2019-05-10 23:48:30 +08:00
|
|
|
if (prev->map_lookup || next->map_lookup)
|
|
|
|
ASSERT(test_bit(EXTENT_FLAG_FS_MAPPING, &prev->flags) &&
|
|
|
|
test_bit(EXTENT_FLAG_FS_MAPPING, &next->flags));
|
|
|
|
|
2008-01-25 05:13:08 +08:00
|
|
|
if (extent_map_end(prev) == next->start &&
|
|
|
|
prev->flags == next->flags &&
|
2019-05-10 23:48:30 +08:00
|
|
|
prev->map_lookup == next->map_lookup &&
|
2008-01-25 05:13:08 +08:00
|
|
|
((next->block_start == EXTENT_MAP_HOLE &&
|
|
|
|
prev->block_start == EXTENT_MAP_HOLE) ||
|
|
|
|
(next->block_start == EXTENT_MAP_INLINE &&
|
|
|
|
prev->block_start == EXTENT_MAP_INLINE) ||
|
|
|
|
(next->block_start < EXTENT_MAP_LAST_BYTE - 1 &&
|
|
|
|
next->block_start == extent_map_block_end(prev)))) {
|
|
|
|
return 1;
|
|
|
|
}
|
2007-08-28 04:49:44 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2011-07-14 11:18:33 +08:00
|
|
|
static void try_merge_map(struct extent_map_tree *tree, struct extent_map *em)
|
2009-09-12 00:27:37 +08:00
|
|
|
{
|
|
|
|
struct extent_map *merge = NULL;
|
|
|
|
struct rb_node *rb;
|
|
|
|
|
Btrfs: fix race between using extent maps and merging them
We have a few cases where we allow an extent map that is in an extent map
tree to be merged with other extents in the tree. Such cases include the
unpinning of an extent after the respective ordered extent completed or
after logging an extent during a fast fsync. This can lead to subtle and
dangerous problems because when doing the merge some other task might be
using the same extent map and as consequence see an inconsistent state of
the extent map - for example sees the new length but has seen the old start
offset.
With luck this triggers a BUG_ON(), and not some silent bug, such as the
following one in __do_readpage():
$ cat -n fs/btrfs/extent_io.c
3061 static int __do_readpage(struct extent_io_tree *tree,
3062 struct page *page,
(...)
3127 em = __get_extent_map(inode, page, pg_offset, cur,
3128 end - cur + 1, get_extent, em_cached);
3129 if (IS_ERR_OR_NULL(em)) {
3130 SetPageError(page);
3131 unlock_extent(tree, cur, end);
3132 break;
3133 }
3134 extent_offset = cur - em->start;
3135 BUG_ON(extent_map_end(em) <= cur);
(...)
Consider the following example scenario, where we end up hitting the
BUG_ON() in __do_readpage().
We have an inode with a size of 8KiB and 2 extent maps:
extent A: file offset 0, length 4KiB, disk_bytenr = X, persisted on disk by
a previous transaction
extent B: file offset 4KiB, length 4KiB, disk_bytenr = X + 4KiB, not yet
persisted but writeback started for it already. The extent map
is pinned since there's writeback and an ordered extent in
progress, so it can not be merged with extent map A yet
The following sequence of steps leads to the BUG_ON():
1) The ordered extent for extent B completes, the respective page gets its
writeback bit cleared and the extent map is unpinned, at that point it
is not yet merged with extent map A because it's in the list of modified
extents;
2) Due to memory pressure, or some other reason, the MM subsystem releases
the page corresponding to extent B - btrfs_releasepage() is called and
returns 1, meaning the page can be released as it's not dirty, not under
writeback anymore and the extent range is not locked in the inode's
iotree. However the extent map is not released, either because we are
not in a context that allows memory allocations to block or because the
inode's size is smaller than 16MiB - in this case our inode has a size
of 8KiB;
3) Task B needs to read extent B and ends up __do_readpage() through the
btrfs_readpage() callback. At __do_readpage() it gets a reference to
extent map B;
4) Task A, doing a fast fsync, calls clear_em_loggin() against extent map B
while holding the write lock on the inode's extent map tree - this
results in try_merge_map() being called and since it's possible to merge
extent map B with extent map A now (the extent map B was removed from
the list of modified extents), the merging begins - it sets extent map
B's start offset to 0 (was 4KiB), but before it increments the map's
length to 8KiB (4kb + 4KiB), task A is at:
BUG_ON(extent_map_end(em) <= cur);
The call to extent_map_end() sees the extent map has a start of 0
and a length still at 4KiB, so it returns 4KiB and 'cur' is 4KiB, so
the BUG_ON() is triggered.
So it's dangerous to modify an extent map that is in the tree, because some
other task might have got a reference to it before and still using it, and
needs to see a consistent map while using it. Generally this is very rare
since most paths that lookup and use extent maps also have the file range
locked in the inode's iotree. The fsync path is pretty much the only
exception where we don't do it to avoid serialization with concurrent
reads.
Fix this by not allowing an extent map do be merged if if it's being used
by tasks other then the one attempting to merge the extent map (when the
reference count of the extent map is greater than 2).
Reported-by: ryusuke1925 <st13s20@gm.ibaraki-ct.ac.jp>
Reported-by: Koki Mitani <koki.mitani.xg@hco.ntt.co.jp>
Bugzilla: https://bugzilla.kernel.org/show_bug.cgi?id=206211
CC: stable@vger.kernel.org # 4.4+
Reviewed-by: Josef Bacik <josef@toxicpanda.com>
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2020-01-31 22:06:07 +08:00
|
|
|
/*
|
|
|
|
* We can't modify an extent map that is in the tree and that is being
|
|
|
|
* used by another task, as it can cause that other task to see it in
|
|
|
|
* inconsistent state during the merging. We always have 1 reference for
|
|
|
|
* the tree and 1 for this task (which is unpinning the extent map or
|
|
|
|
* clearing the logging flag), so anything > 2 means it's being used by
|
|
|
|
* other tasks too.
|
|
|
|
*/
|
|
|
|
if (refcount_read(&em->refs) > 2)
|
|
|
|
return;
|
|
|
|
|
2009-09-12 00:27:37 +08:00
|
|
|
if (em->start != 0) {
|
|
|
|
rb = rb_prev(&em->rb_node);
|
|
|
|
if (rb)
|
|
|
|
merge = rb_entry(rb, struct extent_map, rb_node);
|
|
|
|
if (rb && mergable_maps(merge, em)) {
|
|
|
|
em->start = merge->start;
|
2012-10-12 04:54:30 +08:00
|
|
|
em->orig_start = merge->orig_start;
|
2009-09-12 00:27:37 +08:00
|
|
|
em->len += merge->len;
|
|
|
|
em->block_len += merge->block_len;
|
|
|
|
em->block_start = merge->block_start;
|
2012-10-12 04:54:30 +08:00
|
|
|
em->mod_len = (em->mod_len + em->mod_start) - merge->mod_start;
|
|
|
|
em->mod_start = merge->mod_start;
|
|
|
|
em->generation = max(em->generation, merge->generation);
|
btrfs: defrag: don't use merged extent map for their generation check
For extent maps, if they are not compressed extents and are adjacent by
logical addresses and file offsets, they can be merged into one larger
extent map.
Such merged extent map will have the higher generation of all the
original ones.
But this brings a problem for autodefrag, as it relies on accurate
extent_map::generation to determine if one extent should be defragged.
For merged extent maps, their higher generation can mark some older
extents to be defragged while the original extent map doesn't meet the
minimal generation threshold.
Thus this will cause extra IO.
So solve the problem, here we introduce a new flag, EXTENT_FLAG_MERGED,
to indicate if the extent map is merged from one or more ems.
And for autodefrag, if we find a merged extent map, and its generation
meets the generation requirement, we just don't use this one, and go
back to defrag_get_extent() to read extent maps from subvolume trees.
This could cause more read IO, but should result less defrag data write,
so in the long run it should be a win for autodefrag.
Reviewed-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: Qu Wenruo <wqu@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-02-11 14:46:13 +08:00
|
|
|
set_bit(EXTENT_FLAG_MERGED, &em->flags);
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
|
2018-08-23 03:51:52 +08:00
|
|
|
rb_erase_cached(&merge->rb_node, &tree->map);
|
2014-02-25 22:15:12 +08:00
|
|
|
RB_CLEAR_NODE(&merge->rb_node);
|
2009-09-12 00:27:37 +08:00
|
|
|
free_extent_map(merge);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
rb = rb_next(&em->rb_node);
|
|
|
|
if (rb)
|
|
|
|
merge = rb_entry(rb, struct extent_map, rb_node);
|
|
|
|
if (rb && mergable_maps(em, merge)) {
|
|
|
|
em->len += merge->len;
|
2013-11-30 19:28:35 +08:00
|
|
|
em->block_len += merge->block_len;
|
2018-08-23 03:51:52 +08:00
|
|
|
rb_erase_cached(&merge->rb_node, &tree->map);
|
2014-02-25 22:15:12 +08:00
|
|
|
RB_CLEAR_NODE(&merge->rb_node);
|
2012-10-12 04:54:30 +08:00
|
|
|
em->mod_len = (merge->mod_start + merge->mod_len) - em->mod_start;
|
|
|
|
em->generation = max(em->generation, merge->generation);
|
btrfs: defrag: don't use merged extent map for their generation check
For extent maps, if they are not compressed extents and are adjacent by
logical addresses and file offsets, they can be merged into one larger
extent map.
Such merged extent map will have the higher generation of all the
original ones.
But this brings a problem for autodefrag, as it relies on accurate
extent_map::generation to determine if one extent should be defragged.
For merged extent maps, their higher generation can mark some older
extents to be defragged while the original extent map doesn't meet the
minimal generation threshold.
Thus this will cause extra IO.
So solve the problem, here we introduce a new flag, EXTENT_FLAG_MERGED,
to indicate if the extent map is merged from one or more ems.
And for autodefrag, if we find a merged extent map, and its generation
meets the generation requirement, we just don't use this one, and go
back to defrag_get_extent() to read extent maps from subvolume trees.
This could cause more read IO, but should result less defrag data write,
so in the long run it should be a win for autodefrag.
Reviewed-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: Qu Wenruo <wqu@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-02-11 14:46:13 +08:00
|
|
|
set_bit(EXTENT_FLAG_MERGED, &em->flags);
|
2009-09-12 00:27:37 +08:00
|
|
|
free_extent_map(merge);
|
|
|
|
}
|
2011-07-14 11:18:33 +08:00
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Unpin an extent from the cache.
|
|
|
|
*
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
* @tree: tree to unpin the extent in
|
|
|
|
* @start: logical offset in the file
|
|
|
|
* @len: length of the extent
|
|
|
|
* @gen: generation that this extent has been modified in
|
|
|
|
*
|
|
|
|
* Called after an extent has been written to disk properly. Set the generation
|
|
|
|
* to the generation that actually added the file item to the inode so we know
|
|
|
|
* we need to sync this extent when we call fsync().
|
|
|
|
*/
|
|
|
|
int unpin_extent_cache(struct extent_map_tree *tree, u64 start, u64 len,
|
|
|
|
u64 gen)
|
2011-07-14 11:18:33 +08:00
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
struct extent_map *em;
|
2012-08-28 00:52:20 +08:00
|
|
|
bool prealloc = false;
|
2011-07-14 11:18:33 +08:00
|
|
|
|
|
|
|
write_lock(&tree->lock);
|
|
|
|
em = lookup_extent_mapping(tree, start, len);
|
|
|
|
|
|
|
|
WARN_ON(!em || em->start != start);
|
|
|
|
|
|
|
|
if (!em)
|
|
|
|
goto out;
|
|
|
|
|
Btrfs: turbo charge fsync
At least for the vm workload. Currently on fsync we will
1) Truncate all items in the log tree for the given inode if they exist
and
2) Copy all items for a given inode into the log
The problem with this is that for things like VMs you can have lots of
extents from the fragmented writing behavior, and worst yet you may have
only modified a few extents, not the entire thing. This patch fixes this
problem by tracking which transid modified our extent, and then when we do
the tree logging we find all of the extents we've modified in our current
transaction, sort them and commit them. We also only truncate up to the
xattrs of the inode and copy that stuff in normally, and then just drop any
extents in the range we have that exist in the log already. Here are some
numbers of a 50 meg fio job that does random writes and fsync()s after every
write
Original Patched
SATA drive 82KB/s 140KB/s
Fusion drive 431KB/s 2532KB/s
So around 2-6 times faster depending on your hardware. There are a few
corner cases, for example if you truncate at all we have to do it the old
way since there is no way to be sure what is in the log is ok. This
probably could be done smarter, but if you write-fsync-truncate-write-fsync
you deserve what you get. All this work is in RAM of course so if your
inode gets evicted from cache and you read it in and fsync it we'll do it
the slow way if we are still in the same transaction that we last modified
the inode in.
The biggest cool part of this is that it requires no changes to the recovery
code, so if you fsync with this patch and crash and load an old kernel, it
will run the recovery and be a-ok. I have tested this pretty thoroughly
with an fsync tester and everything comes back fine, as well as xfstests.
Thanks,
Signed-off-by: Josef Bacik <jbacik@fusionio.com>
2012-08-18 01:14:17 +08:00
|
|
|
em->generation = gen;
|
2011-07-14 11:18:33 +08:00
|
|
|
clear_bit(EXTENT_FLAG_PINNED, &em->flags);
|
2012-08-28 00:52:20 +08:00
|
|
|
em->mod_start = em->start;
|
|
|
|
em->mod_len = em->len;
|
|
|
|
|
2012-12-03 23:58:15 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_FILLING, &em->flags)) {
|
2012-08-28 00:52:20 +08:00
|
|
|
prealloc = true;
|
2012-12-03 23:58:15 +08:00
|
|
|
clear_bit(EXTENT_FLAG_FILLING, &em->flags);
|
2012-08-28 00:52:20 +08:00
|
|
|
}
|
2011-07-14 11:18:33 +08:00
|
|
|
|
|
|
|
try_merge_map(tree, em);
|
2012-08-28 00:52:20 +08:00
|
|
|
|
|
|
|
if (prealloc) {
|
|
|
|
em->mod_start = em->start;
|
|
|
|
em->mod_len = em->len;
|
|
|
|
}
|
|
|
|
|
2009-09-12 00:27:37 +08:00
|
|
|
free_extent_map(em);
|
|
|
|
out:
|
|
|
|
write_unlock(&tree->lock);
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2013-01-25 01:02:07 +08:00
|
|
|
void clear_em_logging(struct extent_map_tree *tree, struct extent_map *em)
|
|
|
|
{
|
2022-09-19 22:06:36 +08:00
|
|
|
lockdep_assert_held_write(&tree->lock);
|
|
|
|
|
2013-01-25 01:02:07 +08:00
|
|
|
clear_bit(EXTENT_FLAG_LOGGING, &em->flags);
|
2014-02-25 22:15:12 +08:00
|
|
|
if (extent_map_in_tree(em))
|
2013-01-28 22:45:20 +08:00
|
|
|
try_merge_map(tree, em);
|
2013-01-25 01:02:07 +08:00
|
|
|
}
|
|
|
|
|
2014-02-25 22:15:13 +08:00
|
|
|
static inline void setup_extent_mapping(struct extent_map_tree *tree,
|
|
|
|
struct extent_map *em,
|
|
|
|
int modified)
|
|
|
|
{
|
2017-03-03 16:55:12 +08:00
|
|
|
refcount_inc(&em->refs);
|
2014-02-25 22:15:13 +08:00
|
|
|
em->mod_start = em->start;
|
|
|
|
em->mod_len = em->len;
|
|
|
|
|
|
|
|
if (modified)
|
|
|
|
list_move(&em->list, &tree->modified_extents);
|
|
|
|
else
|
|
|
|
try_merge_map(tree, em);
|
|
|
|
}
|
|
|
|
|
2019-03-27 20:24:12 +08:00
|
|
|
static void extent_map_device_set_bits(struct extent_map *em, unsigned bits)
|
|
|
|
{
|
|
|
|
struct map_lookup *map = em->map_lookup;
|
|
|
|
u64 stripe_size = em->orig_block_len;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < map->num_stripes; i++) {
|
2021-09-15 15:17:16 +08:00
|
|
|
struct btrfs_io_stripe *stripe = &map->stripes[i];
|
2019-03-27 20:24:12 +08:00
|
|
|
struct btrfs_device *device = stripe->dev;
|
|
|
|
|
2023-05-25 07:04:30 +08:00
|
|
|
set_extent_bit(&device->alloc_state, stripe->physical,
|
2023-05-25 07:04:37 +08:00
|
|
|
stripe->physical + stripe_size - 1,
|
2023-05-25 07:04:39 +08:00
|
|
|
bits | EXTENT_NOWAIT, NULL);
|
2019-03-27 20:24:12 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void extent_map_device_clear_bits(struct extent_map *em, unsigned bits)
|
|
|
|
{
|
|
|
|
struct map_lookup *map = em->map_lookup;
|
|
|
|
u64 stripe_size = em->orig_block_len;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < map->num_stripes; i++) {
|
2021-09-15 15:17:16 +08:00
|
|
|
struct btrfs_io_stripe *stripe = &map->stripes[i];
|
2019-03-27 20:24:12 +08:00
|
|
|
struct btrfs_device *device = stripe->dev;
|
|
|
|
|
|
|
|
__clear_extent_bit(&device->alloc_state, stripe->physical,
|
2023-05-25 07:04:37 +08:00
|
|
|
stripe->physical + stripe_size - 1,
|
|
|
|
bits | EXTENT_NOWAIT,
|
2023-05-25 07:04:39 +08:00
|
|
|
NULL, NULL);
|
2019-03-27 20:24:12 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
2021-01-22 17:57:52 +08:00
|
|
|
* Add new extent map to the extent tree
|
|
|
|
*
|
2008-06-12 09:52:17 +08:00
|
|
|
* @tree: tree to insert new map in
|
|
|
|
* @em: map to insert
|
2021-01-22 17:57:52 +08:00
|
|
|
* @modified: indicate whether the given @em should be added to the
|
|
|
|
* modified list, which indicates the extent needs to be logged
|
2008-06-12 09:52:17 +08:00
|
|
|
*
|
|
|
|
* Insert @em into @tree or perform a simple forward/backward merge with
|
|
|
|
* existing mappings. The extent_map struct passed in will be inserted
|
|
|
|
* into the tree directly, with an additional reference taken, or a
|
2011-03-31 09:57:33 +08:00
|
|
|
* reference dropped if the merge attempt was successful.
|
2007-08-28 04:49:44 +08:00
|
|
|
*/
|
|
|
|
int add_extent_mapping(struct extent_map_tree *tree,
|
2013-04-06 04:51:15 +08:00
|
|
|
struct extent_map *em, int modified)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
2019-03-27 23:19:55 +08:00
|
|
|
lockdep_assert_held_write(&tree->lock);
|
|
|
|
|
2013-11-25 11:23:51 +08:00
|
|
|
ret = tree_insert(&tree->map, em);
|
|
|
|
if (ret)
|
2007-08-28 04:49:44 +08:00
|
|
|
goto out;
|
2013-11-25 11:23:51 +08:00
|
|
|
|
2014-02-25 22:15:13 +08:00
|
|
|
setup_extent_mapping(tree, em, modified);
|
2019-03-27 20:24:16 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags)) {
|
2019-03-27 20:24:12 +08:00
|
|
|
extent_map_device_set_bits(em, CHUNK_ALLOCATED);
|
2019-03-27 20:24:16 +08:00
|
|
|
extent_map_device_clear_bits(em, CHUNK_TRIMMED);
|
|
|
|
}
|
2007-08-28 04:49:44 +08:00
|
|
|
out:
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2013-04-26 04:41:01 +08:00
|
|
|
static struct extent_map *
|
|
|
|
__lookup_extent_mapping(struct extent_map_tree *tree,
|
|
|
|
u64 start, u64 len, int strict)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
|
|
|
struct extent_map *em;
|
|
|
|
struct rb_node *rb_node;
|
2022-09-19 22:06:38 +08:00
|
|
|
struct rb_node *prev_or_next = NULL;
|
2008-06-10 22:21:04 +08:00
|
|
|
u64 end = range_end(start, len);
|
|
|
|
|
2022-09-19 22:06:38 +08:00
|
|
|
rb_node = __tree_search(&tree->map.rb_root, start, &prev_or_next);
|
2007-08-28 04:49:44 +08:00
|
|
|
if (!rb_node) {
|
2022-09-19 22:06:38 +08:00
|
|
|
if (prev_or_next)
|
|
|
|
rb_node = prev_or_next;
|
2011-07-14 11:18:15 +08:00
|
|
|
else
|
|
|
|
return NULL;
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
2011-07-14 11:18:15 +08:00
|
|
|
|
2007-08-28 04:49:44 +08:00
|
|
|
em = rb_entry(rb_node, struct extent_map, rb_node);
|
2008-01-25 05:13:08 +08:00
|
|
|
|
2011-07-14 11:18:15 +08:00
|
|
|
if (strict && !(end > em->start && start < extent_map_end(em)))
|
|
|
|
return NULL;
|
2008-01-25 05:13:08 +08:00
|
|
|
|
2017-03-03 16:55:12 +08:00
|
|
|
refcount_inc(&em->refs);
|
2007-08-28 04:49:44 +08:00
|
|
|
return em;
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Lookup extent_map that intersects @start + @len range.
|
|
|
|
*
|
2011-07-14 11:18:15 +08:00
|
|
|
* @tree: tree to lookup in
|
|
|
|
* @start: byte offset to start the search
|
|
|
|
* @len: length of the lookup range
|
|
|
|
*
|
|
|
|
* Find and return the first extent_map struct in @tree that intersects the
|
|
|
|
* [start, len] range. There may be additional objects in the tree that
|
|
|
|
* intersect, so check the object returned carefully to make sure that no
|
|
|
|
* additional lookups are needed.
|
|
|
|
*/
|
|
|
|
struct extent_map *lookup_extent_mapping(struct extent_map_tree *tree,
|
|
|
|
u64 start, u64 len)
|
|
|
|
{
|
|
|
|
return __lookup_extent_mapping(tree, start, len, 1);
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Find a nearby extent map intersecting @start + @len (not an exact search).
|
|
|
|
*
|
2009-09-19 04:07:03 +08:00
|
|
|
* @tree: tree to lookup in
|
|
|
|
* @start: byte offset to start the search
|
|
|
|
* @len: length of the lookup range
|
|
|
|
*
|
|
|
|
* Find and return the first extent_map struct in @tree that intersects the
|
|
|
|
* [start, len] range.
|
|
|
|
*
|
|
|
|
* If one can't be found, any nearby extent may be returned
|
|
|
|
*/
|
|
|
|
struct extent_map *search_extent_mapping(struct extent_map_tree *tree,
|
|
|
|
u64 start, u64 len)
|
|
|
|
{
|
2011-07-14 11:18:15 +08:00
|
|
|
return __lookup_extent_mapping(tree, start, len, 0);
|
2009-09-19 04:07:03 +08:00
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Remove an extent_map from the extent tree.
|
|
|
|
*
|
2008-06-12 09:52:17 +08:00
|
|
|
* @tree: extent tree to remove from
|
2016-03-05 03:23:12 +08:00
|
|
|
* @em: extent map being removed
|
2008-06-12 09:52:17 +08:00
|
|
|
*
|
2022-10-27 20:21:42 +08:00
|
|
|
* Remove @em from @tree. No reference counts are dropped, and no checks
|
|
|
|
* are done to see if the range is in use.
|
2007-08-28 04:49:44 +08:00
|
|
|
*/
|
2018-09-13 11:01:15 +08:00
|
|
|
void remove_extent_mapping(struct extent_map_tree *tree, struct extent_map *em)
|
2007-08-28 04:49:44 +08:00
|
|
|
{
|
2022-02-03 23:36:45 +08:00
|
|
|
lockdep_assert_held_write(&tree->lock);
|
|
|
|
|
2008-07-19 00:01:11 +08:00
|
|
|
WARN_ON(test_bit(EXTENT_FLAG_PINNED, &em->flags));
|
2018-08-23 03:51:52 +08:00
|
|
|
rb_erase_cached(&em->rb_node, &tree->map);
|
2012-09-15 00:59:20 +08:00
|
|
|
if (!test_bit(EXTENT_FLAG_LOGGING, &em->flags))
|
|
|
|
list_del_init(&em->list);
|
2019-03-27 20:24:12 +08:00
|
|
|
if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags))
|
|
|
|
extent_map_device_clear_bits(em, CHUNK_ALLOCATED);
|
2014-02-25 22:15:12 +08:00
|
|
|
RB_CLEAR_NODE(&em->rb_node);
|
2007-08-28 04:49:44 +08:00
|
|
|
}
|
2014-02-25 22:15:13 +08:00
|
|
|
|
2023-05-24 23:03:09 +08:00
|
|
|
static void replace_extent_mapping(struct extent_map_tree *tree,
|
|
|
|
struct extent_map *cur,
|
|
|
|
struct extent_map *new,
|
|
|
|
int modified)
|
2014-02-25 22:15:13 +08:00
|
|
|
{
|
2022-02-03 23:36:45 +08:00
|
|
|
lockdep_assert_held_write(&tree->lock);
|
|
|
|
|
2014-02-25 22:15:13 +08:00
|
|
|
WARN_ON(test_bit(EXTENT_FLAG_PINNED, &cur->flags));
|
|
|
|
ASSERT(extent_map_in_tree(cur));
|
|
|
|
if (!test_bit(EXTENT_FLAG_LOGGING, &cur->flags))
|
|
|
|
list_del_init(&cur->list);
|
2018-08-23 03:51:52 +08:00
|
|
|
rb_replace_node_cached(&cur->rb_node, &new->rb_node, &tree->map);
|
2014-02-25 22:15:13 +08:00
|
|
|
RB_CLEAR_NODE(&cur->rb_node);
|
|
|
|
|
|
|
|
setup_extent_mapping(tree, new, modified);
|
|
|
|
}
|
2018-01-06 03:51:11 +08:00
|
|
|
|
2022-10-11 20:16:54 +08:00
|
|
|
static struct extent_map *next_extent_map(const struct extent_map *em)
|
2018-01-06 03:51:11 +08:00
|
|
|
{
|
|
|
|
struct rb_node *next;
|
|
|
|
|
|
|
|
next = rb_next(&em->rb_node);
|
|
|
|
if (!next)
|
|
|
|
return NULL;
|
|
|
|
return container_of(next, struct extent_map, rb_node);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct extent_map *prev_extent_map(struct extent_map *em)
|
|
|
|
{
|
|
|
|
struct rb_node *prev;
|
|
|
|
|
|
|
|
prev = rb_prev(&em->rb_node);
|
|
|
|
if (!prev)
|
|
|
|
return NULL;
|
|
|
|
return container_of(prev, struct extent_map, rb_node);
|
|
|
|
}
|
|
|
|
|
2018-11-28 19:05:13 +08:00
|
|
|
/*
|
|
|
|
* Helper for btrfs_get_extent. Given an existing extent in the tree,
|
2018-01-06 03:51:11 +08:00
|
|
|
* the existing extent is the nearest extent to map_start,
|
|
|
|
* and an extent that you want to insert, deal with overlap and insert
|
|
|
|
* the best fitted new extent into the tree.
|
|
|
|
*/
|
2018-01-06 03:51:17 +08:00
|
|
|
static noinline int merge_extent_mapping(struct extent_map_tree *em_tree,
|
|
|
|
struct extent_map *existing,
|
|
|
|
struct extent_map *em,
|
|
|
|
u64 map_start)
|
2018-01-06 03:51:11 +08:00
|
|
|
{
|
|
|
|
struct extent_map *prev;
|
|
|
|
struct extent_map *next;
|
|
|
|
u64 start;
|
|
|
|
u64 end;
|
|
|
|
u64 start_diff;
|
|
|
|
|
|
|
|
BUG_ON(map_start < em->start || map_start >= extent_map_end(em));
|
|
|
|
|
|
|
|
if (existing->start > map_start) {
|
|
|
|
next = existing;
|
|
|
|
prev = prev_extent_map(next);
|
|
|
|
} else {
|
|
|
|
prev = existing;
|
|
|
|
next = next_extent_map(prev);
|
|
|
|
}
|
|
|
|
|
|
|
|
start = prev ? extent_map_end(prev) : em->start;
|
|
|
|
start = max_t(u64, start, em->start);
|
|
|
|
end = next ? next->start : extent_map_end(em);
|
|
|
|
end = min_t(u64, end, extent_map_end(em));
|
|
|
|
start_diff = start - em->start;
|
|
|
|
em->start = start;
|
|
|
|
em->len = end - start;
|
|
|
|
if (em->block_start < EXTENT_MAP_LAST_BYTE &&
|
|
|
|
!test_bit(EXTENT_FLAG_COMPRESSED, &em->flags)) {
|
|
|
|
em->block_start += start_diff;
|
|
|
|
em->block_len = em->len;
|
|
|
|
}
|
|
|
|
return add_extent_mapping(em_tree, em, 0);
|
|
|
|
}
|
|
|
|
|
2022-10-27 20:21:42 +08:00
|
|
|
/*
|
|
|
|
* Add extent mapping into em_tree.
|
2021-01-22 17:57:53 +08:00
|
|
|
*
|
|
|
|
* @fs_info: the filesystem
|
|
|
|
* @em_tree: extent tree into which we want to insert the extent mapping
|
|
|
|
* @em_in: extent we are inserting
|
|
|
|
* @start: start of the logical range btrfs_get_extent() is requesting
|
|
|
|
* @len: length of the logical range btrfs_get_extent() is requesting
|
2018-01-06 03:51:11 +08:00
|
|
|
*
|
|
|
|
* Note that @em_in's range may be different from [start, start+len),
|
|
|
|
* but they must be overlapped.
|
|
|
|
*
|
|
|
|
* Insert @em_in into @em_tree. In case there is an overlapping range, handle
|
|
|
|
* the -EEXIST by either:
|
|
|
|
* a) Returning the existing extent in @em_in if @start is within the
|
|
|
|
* existing em.
|
|
|
|
* b) Merge the existing extent with @em_in passed in.
|
|
|
|
*
|
|
|
|
* Return 0 on success, otherwise -EEXIST.
|
|
|
|
*
|
|
|
|
*/
|
2018-04-04 03:45:57 +08:00
|
|
|
int btrfs_add_extent_mapping(struct btrfs_fs_info *fs_info,
|
|
|
|
struct extent_map_tree *em_tree,
|
2018-01-06 03:51:11 +08:00
|
|
|
struct extent_map **em_in, u64 start, u64 len)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
struct extent_map *em = *em_in;
|
|
|
|
|
2022-09-16 15:28:35 +08:00
|
|
|
/*
|
|
|
|
* Tree-checker should have rejected any inline extent with non-zero
|
|
|
|
* file offset. Here just do a sanity check.
|
|
|
|
*/
|
|
|
|
if (em->block_start == EXTENT_MAP_INLINE)
|
|
|
|
ASSERT(em->start == 0);
|
|
|
|
|
2018-01-06 03:51:11 +08:00
|
|
|
ret = add_extent_mapping(em_tree, em, 0);
|
|
|
|
/* it is possible that someone inserted the extent into the tree
|
|
|
|
* while we had the lock dropped. It is also possible that
|
|
|
|
* an overlapping map exists in the tree
|
|
|
|
*/
|
|
|
|
if (ret == -EEXIST) {
|
|
|
|
struct extent_map *existing;
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
existing = search_extent_mapping(em_tree, start, len);
|
2018-01-06 03:51:16 +08:00
|
|
|
|
2018-04-04 03:45:57 +08:00
|
|
|
trace_btrfs_handle_em_exist(fs_info, existing, em, start, len);
|
2018-01-06 03:51:16 +08:00
|
|
|
|
2018-01-06 03:51:11 +08:00
|
|
|
/*
|
|
|
|
* existing will always be non-NULL, since there must be
|
|
|
|
* extent causing the -EEXIST.
|
|
|
|
*/
|
|
|
|
if (start >= existing->start &&
|
|
|
|
start < extent_map_end(existing)) {
|
|
|
|
free_extent_map(em);
|
|
|
|
*em_in = existing;
|
|
|
|
ret = 0;
|
|
|
|
} else {
|
2018-01-06 03:51:15 +08:00
|
|
|
u64 orig_start = em->start;
|
|
|
|
u64 orig_len = em->len;
|
|
|
|
|
2018-01-06 03:51:11 +08:00
|
|
|
/*
|
|
|
|
* The existing extent map is the one nearest to
|
|
|
|
* the [start, start + len) range which overlaps
|
|
|
|
*/
|
|
|
|
ret = merge_extent_mapping(em_tree, existing,
|
|
|
|
em, start);
|
|
|
|
if (ret) {
|
|
|
|
free_extent_map(em);
|
|
|
|
*em_in = NULL;
|
2018-01-06 03:51:15 +08:00
|
|
|
WARN_ONCE(ret,
|
|
|
|
"unexpected error %d: merge existing(start %llu len %llu) with em(start %llu len %llu)\n",
|
|
|
|
ret, existing->start, existing->len,
|
|
|
|
orig_start, orig_len);
|
2018-01-06 03:51:11 +08:00
|
|
|
}
|
2018-01-06 03:51:15 +08:00
|
|
|
free_extent_map(existing);
|
2018-01-06 03:51:11 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT(ret == 0 || ret == -EEXIST);
|
|
|
|
return ret;
|
|
|
|
}
|
2022-09-19 22:06:29 +08:00
|
|
|
|
2022-09-19 22:06:32 +08:00
|
|
|
/*
|
|
|
|
* Drop all extent maps from a tree in the fastest possible way, rescheduling
|
|
|
|
* if needed. This avoids searching the tree, from the root down to the first
|
|
|
|
* extent map, before each deletion.
|
|
|
|
*/
|
|
|
|
static void drop_all_extent_maps_fast(struct extent_map_tree *tree)
|
|
|
|
{
|
|
|
|
write_lock(&tree->lock);
|
|
|
|
while (!RB_EMPTY_ROOT(&tree->map.rb_root)) {
|
|
|
|
struct extent_map *em;
|
|
|
|
struct rb_node *node;
|
|
|
|
|
|
|
|
node = rb_first_cached(&tree->map);
|
|
|
|
em = rb_entry(node, struct extent_map, rb_node);
|
|
|
|
clear_bit(EXTENT_FLAG_PINNED, &em->flags);
|
|
|
|
clear_bit(EXTENT_FLAG_LOGGING, &em->flags);
|
|
|
|
remove_extent_mapping(tree, em);
|
|
|
|
free_extent_map(em);
|
|
|
|
cond_resched_rwlock_write(&tree->lock);
|
|
|
|
}
|
|
|
|
write_unlock(&tree->lock);
|
|
|
|
}
|
|
|
|
|
2022-09-19 22:06:29 +08:00
|
|
|
/*
|
|
|
|
* Drop all extent maps in a given range.
|
|
|
|
*
|
|
|
|
* @inode: The target inode.
|
|
|
|
* @start: Start offset of the range.
|
|
|
|
* @end: End offset of the range (inclusive value).
|
|
|
|
* @skip_pinned: Indicate if pinned extent maps should be ignored or not.
|
|
|
|
*
|
|
|
|
* This drops all the extent maps that intersect the given range [@start, @end].
|
|
|
|
* Extent maps that partially overlap the range and extend behind or beyond it,
|
|
|
|
* are split.
|
|
|
|
* The caller should have locked an appropriate file range in the inode's io
|
|
|
|
* tree before calling this function.
|
|
|
|
*/
|
|
|
|
void btrfs_drop_extent_map_range(struct btrfs_inode *inode, u64 start, u64 end,
|
|
|
|
bool skip_pinned)
|
|
|
|
{
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
struct extent_map *split;
|
|
|
|
struct extent_map *split2;
|
|
|
|
struct extent_map *em;
|
2022-09-19 22:06:29 +08:00
|
|
|
struct extent_map_tree *em_tree = &inode->extent_tree;
|
|
|
|
u64 len = end - start + 1;
|
|
|
|
|
|
|
|
WARN_ON(end < start);
|
|
|
|
if (end == (u64)-1) {
|
2022-09-19 22:06:32 +08:00
|
|
|
if (start == 0 && !skip_pinned) {
|
|
|
|
drop_all_extent_maps_fast(em_tree);
|
|
|
|
return;
|
|
|
|
}
|
2022-09-19 22:06:29 +08:00
|
|
|
len = (u64)-1;
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
} else {
|
|
|
|
/* Make end offset exclusive for use in the loop below. */
|
|
|
|
end++;
|
2022-09-19 22:06:29 +08:00
|
|
|
}
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* It's ok if we fail to allocate the extent maps, see the comment near
|
|
|
|
* the bottom of the loop below. We only need two spare extent maps in
|
|
|
|
* the worst case, where the first extent map that intersects our range
|
|
|
|
* starts before the range and the last extent map that intersects our
|
|
|
|
* range ends after our range (and they might be the same extent map),
|
|
|
|
* because we need to split those two extent maps at the boundaries.
|
|
|
|
*/
|
|
|
|
split = alloc_extent_map();
|
|
|
|
split2 = alloc_extent_map();
|
|
|
|
|
|
|
|
write_lock(&em_tree->lock);
|
|
|
|
em = lookup_extent_mapping(em_tree, start, len);
|
|
|
|
|
|
|
|
while (em) {
|
|
|
|
/* extent_map_end() returns exclusive value (last byte + 1). */
|
|
|
|
const u64 em_end = extent_map_end(em);
|
|
|
|
struct extent_map *next_em = NULL;
|
2022-09-19 22:06:29 +08:00
|
|
|
u64 gen;
|
|
|
|
unsigned long flags;
|
|
|
|
bool modified;
|
|
|
|
bool compressed;
|
|
|
|
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
if (em_end < end) {
|
|
|
|
next_em = next_extent_map(em);
|
|
|
|
if (next_em) {
|
|
|
|
if (next_em->start < end)
|
|
|
|
refcount_inc(&next_em->refs);
|
|
|
|
else
|
|
|
|
next_em = NULL;
|
|
|
|
}
|
2022-09-19 22:06:29 +08:00
|
|
|
}
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
|
2022-09-19 22:06:29 +08:00
|
|
|
if (skip_pinned && test_bit(EXTENT_FLAG_PINNED, &em->flags)) {
|
2022-09-19 22:06:30 +08:00
|
|
|
start = em_end;
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
goto next;
|
2022-09-19 22:06:29 +08:00
|
|
|
}
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
|
2023-02-27 20:53:56 +08:00
|
|
|
flags = em->flags;
|
2022-09-19 22:06:29 +08:00
|
|
|
clear_bit(EXTENT_FLAG_PINNED, &em->flags);
|
2023-02-27 20:53:56 +08:00
|
|
|
/*
|
|
|
|
* In case we split the extent map, we want to preserve the
|
|
|
|
* EXTENT_FLAG_LOGGING flag on our extent map, but we don't want
|
|
|
|
* it on the new extent maps.
|
|
|
|
*/
|
2022-09-19 22:06:29 +08:00
|
|
|
clear_bit(EXTENT_FLAG_LOGGING, &flags);
|
|
|
|
modified = !list_empty(&em->list);
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The extent map does not cross our target range, so no need to
|
|
|
|
* split it, we can remove it directly.
|
|
|
|
*/
|
|
|
|
if (em->start >= start && em_end <= end)
|
|
|
|
goto remove_em;
|
|
|
|
|
|
|
|
gen = em->generation;
|
|
|
|
compressed = test_bit(EXTENT_FLAG_COMPRESSED, &em->flags);
|
2022-09-19 22:06:29 +08:00
|
|
|
|
|
|
|
if (em->start < start) {
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
if (!split) {
|
|
|
|
split = split2;
|
|
|
|
split2 = NULL;
|
|
|
|
if (!split)
|
|
|
|
goto remove_em;
|
|
|
|
}
|
2022-09-19 22:06:29 +08:00
|
|
|
split->start = em->start;
|
|
|
|
split->len = start - em->start;
|
|
|
|
|
|
|
|
if (em->block_start < EXTENT_MAP_LAST_BYTE) {
|
|
|
|
split->orig_start = em->orig_start;
|
|
|
|
split->block_start = em->block_start;
|
|
|
|
|
|
|
|
if (compressed)
|
|
|
|
split->block_len = em->block_len;
|
|
|
|
else
|
|
|
|
split->block_len = split->len;
|
|
|
|
split->orig_block_len = max(split->block_len,
|
|
|
|
em->orig_block_len);
|
|
|
|
split->ram_bytes = em->ram_bytes;
|
|
|
|
} else {
|
|
|
|
split->orig_start = split->start;
|
|
|
|
split->block_len = 0;
|
|
|
|
split->block_start = em->block_start;
|
|
|
|
split->orig_block_len = 0;
|
|
|
|
split->ram_bytes = split->len;
|
|
|
|
}
|
|
|
|
|
|
|
|
split->generation = gen;
|
|
|
|
split->flags = flags;
|
|
|
|
split->compress_type = em->compress_type;
|
|
|
|
replace_extent_mapping(em_tree, em, split, modified);
|
|
|
|
free_extent_map(split);
|
|
|
|
split = split2;
|
|
|
|
split2 = NULL;
|
|
|
|
}
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
if (em_end > end) {
|
|
|
|
if (!split) {
|
|
|
|
split = split2;
|
|
|
|
split2 = NULL;
|
|
|
|
if (!split)
|
|
|
|
goto remove_em;
|
|
|
|
}
|
btrfs: fix incorrect splitting in btrfs_drop_extent_map_range
In production we were seeing a variety of WARN_ON()'s in the extent_map
code, specifically in btrfs_drop_extent_map_range() when we have to call
add_extent_mapping() for our second split.
Consider the following extent map layout
PINNED
[0 16K) [32K, 48K)
and then we call btrfs_drop_extent_map_range for [0, 36K), with
skip_pinned == true. The initial loop will have
start = 0
end = 36K
len = 36K
we will find the [0, 16k) extent, but since we are pinned we will skip
it, which has this code
start = em_end;
if (end != (u64)-1)
len = start + len - em_end;
em_end here is 16K, so now the values are
start = 16K
len = 16K + 36K - 16K = 36K
len should instead be 20K. This is a problem when we find the next
extent at [32K, 48K), we need to split this extent to leave [36K, 48k),
however the code for the split looks like this
split->start = start + len;
split->len = em_end - (start + len);
In this case we have
em_end = 48K
split->start = 16K + 36K // this should be 16K + 20K
split->len = 48K - (16K + 36K) // this overflows as 16K + 36K is 52K
and now we have an invalid extent_map in the tree that potentially
overlaps other entries in the extent map. Even in the non-overlapping
case we will have split->start set improperly, which will cause problems
with any block related calculations.
We don't actually need len in this loop, we can simply use end as our
end point, and only adjust start up when we find a pinned extent we need
to skip.
Adjust the logic to do this, which keeps us from inserting an invalid
extent map.
We only skip_pinned in the relocation case, so this is relatively rare,
except in the case where you are running relocation a lot, which can
happen with auto relocation on.
Fixes: 55ef68990029 ("Btrfs: Fix btrfs_drop_extent_cache for skip pinned case")
CC: stable@vger.kernel.org # 4.14+
Reviewed-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: Josef Bacik <josef@toxicpanda.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2023-08-18 04:57:30 +08:00
|
|
|
split->start = end;
|
|
|
|
split->len = em_end - end;
|
2022-09-19 22:06:29 +08:00
|
|
|
split->block_start = em->block_start;
|
|
|
|
split->flags = flags;
|
|
|
|
split->compress_type = em->compress_type;
|
|
|
|
split->generation = gen;
|
|
|
|
|
|
|
|
if (em->block_start < EXTENT_MAP_LAST_BYTE) {
|
|
|
|
split->orig_block_len = max(em->block_len,
|
|
|
|
em->orig_block_len);
|
|
|
|
|
|
|
|
split->ram_bytes = em->ram_bytes;
|
|
|
|
if (compressed) {
|
|
|
|
split->block_len = em->block_len;
|
|
|
|
split->orig_start = em->orig_start;
|
|
|
|
} else {
|
|
|
|
const u64 diff = start + len - em->start;
|
|
|
|
|
|
|
|
split->block_len = split->len;
|
|
|
|
split->block_start += diff;
|
|
|
|
split->orig_start = em->orig_start;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
split->ram_bytes = split->len;
|
|
|
|
split->orig_start = split->start;
|
|
|
|
split->block_len = 0;
|
|
|
|
split->orig_block_len = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (extent_map_in_tree(em)) {
|
|
|
|
replace_extent_mapping(em_tree, em, split,
|
|
|
|
modified);
|
|
|
|
} else {
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = add_extent_mapping(em_tree, split,
|
|
|
|
modified);
|
|
|
|
/* Logic error, shouldn't happen. */
|
|
|
|
ASSERT(ret == 0);
|
|
|
|
if (WARN_ON(ret != 0) && modified)
|
|
|
|
btrfs_set_inode_full_sync(inode);
|
|
|
|
}
|
|
|
|
free_extent_map(split);
|
|
|
|
split = NULL;
|
|
|
|
}
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
remove_em:
|
2022-09-19 22:06:29 +08:00
|
|
|
if (extent_map_in_tree(em)) {
|
|
|
|
/*
|
|
|
|
* If the extent map is still in the tree it means that
|
|
|
|
* either of the following is true:
|
|
|
|
*
|
|
|
|
* 1) It fits entirely in our range (doesn't end beyond
|
|
|
|
* it or starts before it);
|
|
|
|
*
|
|
|
|
* 2) It starts before our range and/or ends after our
|
|
|
|
* range, and we were not able to allocate the extent
|
|
|
|
* maps for split operations, @split and @split2.
|
|
|
|
*
|
|
|
|
* If we are at case 2) then we just remove the entire
|
|
|
|
* extent map - this is fine since if anyone needs it to
|
|
|
|
* access the subranges outside our range, will just
|
|
|
|
* load it again from the subvolume tree's file extent
|
|
|
|
* item. However if the extent map was in the list of
|
|
|
|
* modified extents, then we must mark the inode for a
|
|
|
|
* full fsync, otherwise a fast fsync will miss this
|
|
|
|
* extent if it's new and needs to be logged.
|
|
|
|
*/
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
if ((em->start < start || em_end > end) && modified) {
|
|
|
|
ASSERT(!split);
|
2022-09-19 22:06:29 +08:00
|
|
|
btrfs_set_inode_full_sync(inode);
|
|
|
|
}
|
|
|
|
remove_extent_mapping(em_tree, em);
|
|
|
|
}
|
|
|
|
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
/*
|
|
|
|
* Once for the tree reference (we replaced or removed the
|
|
|
|
* extent map from the tree).
|
|
|
|
*/
|
2022-09-19 22:06:29 +08:00
|
|
|
free_extent_map(em);
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
next:
|
|
|
|
/* Once for us (for our lookup reference). */
|
2022-09-19 22:06:29 +08:00
|
|
|
free_extent_map(em);
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
|
|
|
|
em = next_em;
|
2022-09-19 22:06:29 +08:00
|
|
|
}
|
|
|
|
|
btrfs: drop extent map range more efficiently
Currently when dropping extent maps for a file range, through
btrfs_drop_extent_map_range(), we do the following non-optimal things:
1) We lookup for extent maps one by one, always starting the search from
the root of the extent map tree. This is not efficient if we have
multiple extent maps in the range;
2) We check on every iteration if we have the 'split' and 'split2' spare
extent maps in case we need to split an extent map that intersects our
range but also crosses its boundaries (to the left, to the right or
both cases). If our target range is for example:
[2M, 8M)
And we have 3 extents maps in the range:
[1M, 3M) [3M, 6M) [6M, 10M[
The on the first iteration we allocate two extent maps for 'split' and
'split2', and use the 'split' to split the first extent map, so after
the split we set 'split' to 'split2' and then set 'split2' to NULL.
On the second iteration, we don't need to split the second extent map,
but because 'split2' is now NULL, we allocate a new extent map for
'split2'.
On the third iteration we need to split the third extent map, so we
use the extent map pointed by 'split'.
So we ended up allocating 3 extent maps for splitting, but all we
needed was 2 extent maps. We never need to allocate more than 2,
because extent maps that need to be split are always the first one
and the last one in the target range.
Improve on this by:
1) Using rb_next() to move on to the next extent map. This results in
iterating over less nodes of the tree and it does not require comparing
the ranges of nodes to our start/end offset;
2) Allocate the 2 extent maps for splitting before entering the loop and
never allocate more than 2. In practice it's very rare to have the
combination of both extent map allocations fail, since we have a
dedicated slab for extent maps, and also have the need to split two
extent maps.
This patch is part of a patchset comprised of the following patches:
btrfs: fix missed extent on fsync after dropping extent maps
btrfs: move btrfs_drop_extent_cache() to extent_map.c
btrfs: use extent_map_end() at btrfs_drop_extent_map_range()
btrfs: use cond_resched_rwlock_write() during inode eviction
btrfs: move open coded extent map tree deletion out of inode eviction
btrfs: add helper to replace extent map range with a new extent map
btrfs: remove the refcount warning/check at free_extent_map()
btrfs: remove unnecessary extent map initializations
btrfs: assert tree is locked when clearing extent map from logging
btrfs: remove unnecessary NULL pointer checks when searching extent maps
btrfs: remove unnecessary next extent map search
btrfs: avoid pointless extent map tree search when flushing delalloc
btrfs: drop extent map range more efficiently
And the following fio test was done before and after applying the whole
patchset, on a non-debug kernel (Debian's default kernel config) on a 12
cores Intel box with 64G of ram:
$ cat test.sh
#!/bin/bash
DEV=/dev/nvme0n1
MNT=/mnt/nvme0n1
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS="-R free-space-tree -O no-holes"
cat <<EOF > /tmp/fio-job.ini
[writers]
rw=randwrite
fsync=8
fallocate=none
group_reporting=1
direct=0
bssplit=4k/20:8k/20:16k/20:32k/10:64k/10:128k/5:256k/5:512k/5:1m/5
ioengine=psync
filesize=2G
runtime=300
time_based
directory=$MNT
numjobs=8
thread
EOF
echo performance | \
tee /sys/devices/system/cpu/cpu*/cpufreq/scaling_governor
echo
echo "Using config:"
echo
cat /tmp/fio-job.ini
echo
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Result before applying the patchset:
WRITE: bw=197MiB/s (206MB/s), 197MiB/s-197MiB/s (206MB/s-206MB/s), io=57.7GiB (61.9GB), run=300188-300188msec
Result after applying the patchset:
WRITE: bw=203MiB/s (213MB/s), 203MiB/s-203MiB/s (213MB/s-213MB/s), io=59.5GiB (63.9GB), run=300019-300019msec
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
2022-09-19 22:06:40 +08:00
|
|
|
write_unlock(&em_tree->lock);
|
|
|
|
|
2022-09-19 22:06:29 +08:00
|
|
|
free_extent_map(split);
|
|
|
|
free_extent_map(split2);
|
|
|
|
}
|
2022-09-19 22:06:33 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Replace a range in the inode's extent map tree with a new extent map.
|
|
|
|
*
|
|
|
|
* @inode: The target inode.
|
|
|
|
* @new_em: The new extent map to add to the inode's extent map tree.
|
|
|
|
* @modified: Indicate if the new extent map should be added to the list of
|
|
|
|
* modified extents (for fast fsync tracking).
|
|
|
|
*
|
|
|
|
* Drops all the extent maps in the inode's extent map tree that intersect the
|
|
|
|
* range of the new extent map and adds the new extent map to the tree.
|
|
|
|
* The caller should have locked an appropriate file range in the inode's io
|
|
|
|
* tree before calling this function.
|
|
|
|
*/
|
|
|
|
int btrfs_replace_extent_map_range(struct btrfs_inode *inode,
|
|
|
|
struct extent_map *new_em,
|
|
|
|
bool modified)
|
|
|
|
{
|
|
|
|
const u64 end = new_em->start + new_em->len - 1;
|
|
|
|
struct extent_map_tree *tree = &inode->extent_tree;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ASSERT(!extent_map_in_tree(new_em));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The caller has locked an appropriate file range in the inode's io
|
|
|
|
* tree, but getting -EEXIST when adding the new extent map can still
|
|
|
|
* happen in case there are extents that partially cover the range, and
|
|
|
|
* this is due to two tasks operating on different parts of the extent.
|
|
|
|
* See commit 18e83ac75bfe67 ("Btrfs: fix unexpected EEXIST from
|
|
|
|
* btrfs_get_extent") for an example and details.
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
btrfs_drop_extent_map_range(inode, new_em->start, end, false);
|
|
|
|
write_lock(&tree->lock);
|
|
|
|
ret = add_extent_mapping(tree, new_em, modified);
|
|
|
|
write_unlock(&tree->lock);
|
|
|
|
} while (ret == -EEXIST);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2023-05-24 23:03:09 +08:00
|
|
|
|
|
|
|
/*
|
2023-05-24 23:03:17 +08:00
|
|
|
* Split off the first pre bytes from the extent_map at [start, start + len],
|
|
|
|
* and set the block_start for it to new_logical.
|
2023-05-24 23:03:09 +08:00
|
|
|
*
|
|
|
|
* This function is used when an ordered_extent needs to be split.
|
|
|
|
*/
|
2023-05-24 23:03:17 +08:00
|
|
|
int split_extent_map(struct btrfs_inode *inode, u64 start, u64 len, u64 pre,
|
|
|
|
u64 new_logical)
|
2023-05-24 23:03:09 +08:00
|
|
|
{
|
|
|
|
struct extent_map_tree *em_tree = &inode->extent_tree;
|
|
|
|
struct extent_map *em;
|
|
|
|
struct extent_map *split_pre = NULL;
|
|
|
|
struct extent_map *split_mid = NULL;
|
|
|
|
int ret = 0;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
ASSERT(pre != 0);
|
|
|
|
ASSERT(pre < len);
|
|
|
|
|
|
|
|
split_pre = alloc_extent_map();
|
|
|
|
if (!split_pre)
|
|
|
|
return -ENOMEM;
|
|
|
|
split_mid = alloc_extent_map();
|
|
|
|
if (!split_mid) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto out_free_pre;
|
|
|
|
}
|
|
|
|
|
|
|
|
lock_extent(&inode->io_tree, start, start + len - 1, NULL);
|
|
|
|
write_lock(&em_tree->lock);
|
|
|
|
em = lookup_extent_mapping(em_tree, start, len);
|
|
|
|
if (!em) {
|
|
|
|
ret = -EIO;
|
|
|
|
goto out_unlock;
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT(em->len == len);
|
|
|
|
ASSERT(!test_bit(EXTENT_FLAG_COMPRESSED, &em->flags));
|
|
|
|
ASSERT(em->block_start < EXTENT_MAP_LAST_BYTE);
|
|
|
|
ASSERT(test_bit(EXTENT_FLAG_PINNED, &em->flags));
|
|
|
|
ASSERT(!test_bit(EXTENT_FLAG_LOGGING, &em->flags));
|
|
|
|
ASSERT(!list_empty(&em->list));
|
|
|
|
|
|
|
|
flags = em->flags;
|
|
|
|
clear_bit(EXTENT_FLAG_PINNED, &em->flags);
|
|
|
|
|
|
|
|
/* First, replace the em with a new extent_map starting from * em->start */
|
|
|
|
split_pre->start = em->start;
|
|
|
|
split_pre->len = pre;
|
|
|
|
split_pre->orig_start = split_pre->start;
|
2023-05-24 23:03:17 +08:00
|
|
|
split_pre->block_start = new_logical;
|
2023-05-24 23:03:09 +08:00
|
|
|
split_pre->block_len = split_pre->len;
|
|
|
|
split_pre->orig_block_len = split_pre->block_len;
|
|
|
|
split_pre->ram_bytes = split_pre->len;
|
|
|
|
split_pre->flags = flags;
|
|
|
|
split_pre->compress_type = em->compress_type;
|
|
|
|
split_pre->generation = em->generation;
|
|
|
|
|
|
|
|
replace_extent_mapping(em_tree, em, split_pre, 1);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Now we only have an extent_map at:
|
|
|
|
* [em->start, em->start + pre]
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Insert the middle extent_map. */
|
|
|
|
split_mid->start = em->start + pre;
|
|
|
|
split_mid->len = em->len - pre;
|
|
|
|
split_mid->orig_start = split_mid->start;
|
|
|
|
split_mid->block_start = em->block_start + pre;
|
|
|
|
split_mid->block_len = split_mid->len;
|
|
|
|
split_mid->orig_block_len = split_mid->block_len;
|
|
|
|
split_mid->ram_bytes = split_mid->len;
|
|
|
|
split_mid->flags = flags;
|
|
|
|
split_mid->compress_type = em->compress_type;
|
|
|
|
split_mid->generation = em->generation;
|
|
|
|
add_extent_mapping(em_tree, split_mid, 1);
|
|
|
|
|
|
|
|
/* Once for us */
|
|
|
|
free_extent_map(em);
|
|
|
|
/* Once for the tree */
|
|
|
|
free_extent_map(em);
|
|
|
|
|
|
|
|
out_unlock:
|
|
|
|
write_unlock(&em_tree->lock);
|
|
|
|
unlock_extent(&inode->io_tree, start, start + len - 1, NULL);
|
|
|
|
free_extent_map(split_mid);
|
|
|
|
out_free_pre:
|
|
|
|
free_extent_map(split_pre);
|
|
|
|
return ret;
|
|
|
|
}
|