btrfs: unexport the temporary exported functions

These were renamed and exported to facilitate logical migration of
different code chunks into block-group.c.  Now that all the users are in
one file go ahead and rename them back, move the code around, and make
them static.

Signed-off-by: Josef Bacik <josef@toxicpanda.com>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
This commit is contained in:
Josef Bacik 2019-06-20 15:38:07 -04:00 committed by David Sterba
parent 3e43c279e8
commit e11c0406ad
2 changed files with 14 additions and 18 deletions

View File

@ -21,7 +21,7 @@
* *
* Should be called with balance_lock held * Should be called with balance_lock held
*/ */
u64 btrfs_get_restripe_target(struct btrfs_fs_info *fs_info, u64 flags) static u64 get_restripe_target(struct btrfs_fs_info *fs_info, u64 flags)
{ {
struct btrfs_balance_control *bctl = fs_info->balance_ctl; struct btrfs_balance_control *bctl = fs_info->balance_ctl;
u64 target = 0; u64 target = 0;
@ -62,7 +62,7 @@ static u64 btrfs_reduce_alloc_profile(struct btrfs_fs_info *fs_info, u64 flags)
* reduce to the target profile * reduce to the target profile
*/ */
spin_lock(&fs_info->balance_lock); spin_lock(&fs_info->balance_lock);
target = btrfs_get_restripe_target(fs_info, flags); target = get_restripe_target(fs_info, flags);
if (target) { if (target) {
/* Pick target profile only if it's already available */ /* Pick target profile only if it's already available */
if ((flags & target) & BTRFS_EXTENDED_PROFILE_MASK) { if ((flags & target) & BTRFS_EXTENDED_PROFILE_MASK) {
@ -424,7 +424,7 @@ int btrfs_wait_block_group_cache_done(struct btrfs_block_group_cache *cache)
} }
#ifdef CONFIG_BTRFS_DEBUG #ifdef CONFIG_BTRFS_DEBUG
void btrfs_fragment_free_space(struct btrfs_block_group_cache *block_group) static void fragment_free_space(struct btrfs_block_group_cache *block_group)
{ {
struct btrfs_fs_info *fs_info = block_group->fs_info; struct btrfs_fs_info *fs_info = block_group->fs_info;
u64 start = block_group->key.objectid; u64 start = block_group->key.objectid;
@ -661,7 +661,7 @@ static noinline void caching_thread(struct btrfs_work *work)
block_group->space_info->bytes_used += bytes_used >> 1; block_group->space_info->bytes_used += bytes_used >> 1;
spin_unlock(&block_group->lock); spin_unlock(&block_group->lock);
spin_unlock(&block_group->space_info->lock); spin_unlock(&block_group->space_info->lock);
btrfs_fragment_free_space(block_group); fragment_free_space(block_group);
} }
#endif #endif
@ -768,7 +768,7 @@ int btrfs_cache_block_group(struct btrfs_block_group_cache *cache,
cache->space_info->bytes_used += bytes_used >> 1; cache->space_info->bytes_used += bytes_used >> 1;
spin_unlock(&cache->lock); spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock); spin_unlock(&cache->space_info->lock);
btrfs_fragment_free_space(cache); fragment_free_space(cache);
} }
#endif #endif
mutex_unlock(&caching_ctl->mutex); mutex_unlock(&caching_ctl->mutex);
@ -1180,7 +1180,7 @@ struct btrfs_trans_handle *btrfs_start_trans_remove_block_group(
* data in this block group. That check should be done by relocation routine, * data in this block group. That check should be done by relocation routine,
* not this function. * not this function.
*/ */
int __btrfs_inc_block_group_ro(struct btrfs_block_group_cache *cache, int force) static int inc_block_group_ro(struct btrfs_block_group_cache *cache, int force)
{ {
struct btrfs_space_info *sinfo = cache->space_info; struct btrfs_space_info *sinfo = cache->space_info;
u64 num_bytes; u64 num_bytes;
@ -1296,7 +1296,7 @@ void btrfs_delete_unused_bgs(struct btrfs_fs_info *fs_info)
spin_unlock(&block_group->lock); spin_unlock(&block_group->lock);
/* We don't want to force the issue, only flip if it's ok. */ /* We don't want to force the issue, only flip if it's ok. */
ret = __btrfs_inc_block_group_ro(block_group, 0); ret = inc_block_group_ro(block_group, 0);
up_write(&space_info->groups_sem); up_write(&space_info->groups_sem);
if (ret < 0) { if (ret < 0) {
ret = 0; ret = 0;
@ -1822,7 +1822,7 @@ int btrfs_read_block_groups(struct btrfs_fs_info *info)
set_avail_alloc_bits(info, cache->flags); set_avail_alloc_bits(info, cache->flags);
if (btrfs_chunk_readonly(info, cache->key.objectid)) { if (btrfs_chunk_readonly(info, cache->key.objectid)) {
__btrfs_inc_block_group_ro(cache, 1); inc_block_group_ro(cache, 1);
} else if (btrfs_block_group_used(&cache->item) == 0) { } else if (btrfs_block_group_used(&cache->item) == 0) {
ASSERT(list_empty(&cache->bg_list)); ASSERT(list_empty(&cache->bg_list));
btrfs_mark_bg_unused(cache); btrfs_mark_bg_unused(cache);
@ -1843,11 +1843,11 @@ int btrfs_read_block_groups(struct btrfs_fs_info *info)
list_for_each_entry(cache, list_for_each_entry(cache,
&space_info->block_groups[BTRFS_RAID_RAID0], &space_info->block_groups[BTRFS_RAID_RAID0],
list) list)
__btrfs_inc_block_group_ro(cache, 1); inc_block_group_ro(cache, 1);
list_for_each_entry(cache, list_for_each_entry(cache,
&space_info->block_groups[BTRFS_RAID_SINGLE], &space_info->block_groups[BTRFS_RAID_SINGLE],
list) list)
__btrfs_inc_block_group_ro(cache, 1); inc_block_group_ro(cache, 1);
} }
btrfs_init_global_block_rsv(info); btrfs_init_global_block_rsv(info);
@ -1936,7 +1936,7 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, u64 bytes_used,
u64 new_bytes_used = size - bytes_used; u64 new_bytes_used = size - bytes_used;
bytes_used += new_bytes_used >> 1; bytes_used += new_bytes_used >> 1;
btrfs_fragment_free_space(cache); fragment_free_space(cache);
} }
#endif #endif
/* /*
@ -1982,7 +1982,7 @@ static u64 update_block_group_flags(struct btrfs_fs_info *fs_info, u64 flags)
* if restripe for this chunk_type is on pick target profile and * if restripe for this chunk_type is on pick target profile and
* return, otherwise do the usual balance * return, otherwise do the usual balance
*/ */
stripped = btrfs_get_restripe_target(fs_info, flags); stripped = get_restripe_target(fs_info, flags);
if (stripped) if (stripped)
return extended_to_chunk(stripped); return extended_to_chunk(stripped);
@ -2070,14 +2070,14 @@ again:
goto out; goto out;
} }
ret = __btrfs_inc_block_group_ro(cache, 0); ret = inc_block_group_ro(cache, 0);
if (!ret) if (!ret)
goto out; goto out;
alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags);
ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE);
if (ret < 0) if (ret < 0)
goto out; goto out;
ret = __btrfs_inc_block_group_ro(cache, 0); ret = inc_block_group_ro(cache, 0);
out: out:
if (cache->flags & BTRFS_BLOCK_GROUP_SYSTEM) { if (cache->flags & BTRFS_BLOCK_GROUP_SYSTEM) {
alloc_flags = update_block_group_flags(fs_info, cache->flags); alloc_flags = update_block_group_flags(fs_info, cache->flags);

View File

@ -166,7 +166,6 @@ static inline int btrfs_should_fragment_free_space(
(btrfs_test_opt(fs_info, FRAGMENT_DATA) && (btrfs_test_opt(fs_info, FRAGMENT_DATA) &&
block_group->flags & BTRFS_BLOCK_GROUP_DATA); block_group->flags & BTRFS_BLOCK_GROUP_DATA);
} }
void btrfs_fragment_free_space(struct btrfs_block_group_cache *block_group);
#endif #endif
struct btrfs_block_group_cache *btrfs_lookup_first_block_group( struct btrfs_block_group_cache *btrfs_lookup_first_block_group(
@ -246,7 +245,4 @@ static inline int btrfs_block_group_cache_done(
cache->cached == BTRFS_CACHE_ERROR; cache->cached == BTRFS_CACHE_ERROR;
} }
int __btrfs_inc_block_group_ro(struct btrfs_block_group_cache *cache, int force);
u64 btrfs_get_restripe_target(struct btrfs_fs_info *fs_info, u64 flags);
#endif /* BTRFS_BLOCK_GROUP_H */ #endif /* BTRFS_BLOCK_GROUP_H */