2007-08-08 04:15:09 +08:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2007 Oracle. All rights reserved.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public
|
|
|
|
* License v2 as published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public
|
|
|
|
* License along with this program; if not, write to the
|
|
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
|
|
* Boston, MA 021110-1307, USA.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/sched.h>
|
|
|
|
#include "ctree.h"
|
|
|
|
#include "disk-io.h"
|
|
|
|
#include "print-tree.h"
|
|
|
|
#include "transaction.h"
|
2008-06-26 04:01:31 +08:00
|
|
|
#include "locking.h"
|
2007-08-08 04:15:09 +08:00
|
|
|
|
2009-01-06 10:25:51 +08:00
|
|
|
/* defrag all the leaves in a given btree. If cache_only == 1, don't read
|
|
|
|
* things from disk, otherwise read all the leaves and try to get key order to
|
2008-09-30 03:18:18 +08:00
|
|
|
* better reflect disk order
|
|
|
|
*/
|
2009-01-06 10:25:51 +08:00
|
|
|
|
2007-08-08 04:15:09 +08:00
|
|
|
int btrfs_defrag_leaves(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root, int cache_only)
|
|
|
|
{
|
|
|
|
struct btrfs_path *path = NULL;
|
2008-06-26 04:01:31 +08:00
|
|
|
struct btrfs_key key;
|
2007-08-08 04:15:09 +08:00
|
|
|
int ret = 0;
|
|
|
|
int wret;
|
|
|
|
int level;
|
2007-08-11 02:06:19 +08:00
|
|
|
int is_extent = 0;
|
2008-06-26 04:01:31 +08:00
|
|
|
int next_key_ret = 0;
|
2007-08-11 02:06:19 +08:00
|
|
|
u64 last_ret = 0;
|
2008-06-26 04:01:31 +08:00
|
|
|
u64 min_trans = 0;
|
|
|
|
|
|
|
|
if (cache_only)
|
|
|
|
goto out;
|
2007-08-11 02:06:19 +08:00
|
|
|
|
2008-06-26 04:01:31 +08:00
|
|
|
if (root->fs_info->extent_root == root) {
|
2008-06-26 04:01:31 +08:00
|
|
|
/*
|
|
|
|
* there's recursion here right now in the tree locking,
|
|
|
|
* we can't defrag the extent root without deadlock
|
|
|
|
*/
|
|
|
|
goto out;
|
2008-06-26 04:01:31 +08:00
|
|
|
}
|
2008-06-26 04:01:30 +08:00
|
|
|
|
2007-08-11 02:06:19 +08:00
|
|
|
if (root->ref_cows == 0 && !is_extent)
|
2007-08-08 04:15:09 +08:00
|
|
|
goto out;
|
2007-10-16 04:14:19 +08:00
|
|
|
|
2008-02-16 02:19:35 +08:00
|
|
|
if (btrfs_test_opt(root, SSD))
|
|
|
|
goto out;
|
|
|
|
|
2007-08-08 04:15:09 +08:00
|
|
|
path = btrfs_alloc_path();
|
|
|
|
if (!path)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2007-10-16 04:14:19 +08:00
|
|
|
level = btrfs_header_level(root->node);
|
2007-10-16 04:19:46 +08:00
|
|
|
|
2009-01-06 10:25:51 +08:00
|
|
|
if (level == 0)
|
2007-08-08 04:15:09 +08:00
|
|
|
goto out;
|
2009-01-06 10:25:51 +08:00
|
|
|
|
2007-08-08 04:15:09 +08:00
|
|
|
if (root->defrag_progress.objectid == 0) {
|
2008-06-26 04:01:31 +08:00
|
|
|
struct extent_buffer *root_node;
|
2008-05-25 02:04:53 +08:00
|
|
|
u32 nritems;
|
|
|
|
|
2008-06-26 04:01:31 +08:00
|
|
|
root_node = btrfs_lock_root_node(root);
|
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 22:25:08 +08:00
|
|
|
btrfs_set_lock_blocking(root_node);
|
2008-06-26 04:01:31 +08:00
|
|
|
nritems = btrfs_header_nritems(root_node);
|
2008-05-25 02:04:53 +08:00
|
|
|
root->defrag_max.objectid = 0;
|
|
|
|
/* from above we know this is not a leaf */
|
2008-06-26 04:01:31 +08:00
|
|
|
btrfs_node_key_to_cpu(root_node, &root->defrag_max,
|
2008-05-25 02:04:53 +08:00
|
|
|
nritems - 1);
|
2008-06-26 04:01:31 +08:00
|
|
|
btrfs_tree_unlock(root_node);
|
|
|
|
free_extent_buffer(root_node);
|
|
|
|
memset(&key, 0, sizeof(key));
|
2007-08-08 04:15:09 +08:00
|
|
|
} else {
|
2008-06-26 04:01:31 +08:00
|
|
|
memcpy(&key, &root->defrag_progress, sizeof(key));
|
2007-08-08 04:15:09 +08:00
|
|
|
}
|
|
|
|
|
2008-06-26 04:01:31 +08:00
|
|
|
path->keep_locks = 1;
|
2008-06-26 04:01:31 +08:00
|
|
|
if (cache_only)
|
|
|
|
min_trans = root->defrag_trans_start;
|
|
|
|
|
2008-09-06 04:13:11 +08:00
|
|
|
ret = btrfs_search_forward(root, &key, NULL, path,
|
|
|
|
cache_only, min_trans);
|
2008-06-26 04:01:31 +08:00
|
|
|
if (ret < 0)
|
|
|
|
goto out;
|
|
|
|
if (ret > 0) {
|
|
|
|
ret = 0;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
btrfs_release_path(root, path);
|
2008-06-26 04:01:31 +08:00
|
|
|
wret = btrfs_search_slot(trans, root, &key, path, 0, 1);
|
2007-08-08 04:15:09 +08:00
|
|
|
|
2008-06-26 04:01:31 +08:00
|
|
|
if (wret < 0) {
|
|
|
|
ret = wret;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
if (!path->nodes[1]) {
|
|
|
|
ret = 0;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
path->slots[1] = btrfs_header_nritems(path->nodes[1]);
|
2008-06-26 04:01:31 +08:00
|
|
|
next_key_ret = btrfs_find_next_key(root, path, &key, 1, cache_only,
|
|
|
|
min_trans);
|
2008-06-26 04:01:31 +08:00
|
|
|
ret = btrfs_realloc_node(trans, root,
|
|
|
|
path->nodes[1], 0,
|
|
|
|
cache_only, &last_ret,
|
|
|
|
&root->defrag_progress);
|
2010-05-16 22:49:58 +08:00
|
|
|
if (ret) {
|
|
|
|
WARN_ON(ret == -EAGAIN);
|
|
|
|
goto out;
|
|
|
|
}
|
2008-06-26 04:01:31 +08:00
|
|
|
if (next_key_ret == 0) {
|
|
|
|
memcpy(&root->defrag_progress, &key, sizeof(key));
|
|
|
|
ret = -EAGAIN;
|
2007-08-08 04:15:09 +08:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
if (path)
|
|
|
|
btrfs_free_path(path);
|
2008-05-25 02:04:53 +08:00
|
|
|
if (ret == -EAGAIN) {
|
|
|
|
if (root->defrag_max.objectid > root->defrag_progress.objectid)
|
|
|
|
goto done;
|
|
|
|
if (root->defrag_max.type > root->defrag_progress.type)
|
|
|
|
goto done;
|
|
|
|
if (root->defrag_max.offset > root->defrag_progress.offset)
|
|
|
|
goto done;
|
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
done:
|
2007-08-08 04:15:09 +08:00
|
|
|
if (ret != -EAGAIN) {
|
|
|
|
memset(&root->defrag_progress, 0,
|
|
|
|
sizeof(root->defrag_progress));
|
2008-06-26 04:01:31 +08:00
|
|
|
root->defrag_trans_start = trans->transid;
|
2007-08-08 04:15:09 +08:00
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|