for-5.4-tag
-----BEGIN PGP SIGNATURE----- iQIzBAABCgAdFiEE8rQSAMVO+zA4DBdWxWXV+ddtWDsFAl1/hCoACgkQxWXV+ddt WDs0lQ//flGLX4fvaY2vuWA26t1elITnIatyX8S+xP4pUsT1Tyy1egeGpR8Jku/7 sCOgUlEM2MNXqveOdkQqPJuFPp3B6tInz4S/fowtLlz4enp7uTXw2SFuS3bhOJ+b rpxK9VTc6QV3aipBCG31m8fnDiMaj2Hcspp0oej3V2mBhLUvzn69+P4eo7WN+46w r2F605+lfURauHE6WjM09HINx3NGSfPqdSA5rJvHSm0jlxhb9l3DJOX8cYkbf8lo MAbLDZmtiDiQAqRcsQPi6LZ1LKBkOYaeSnVvnXnH23FI04LBra3duk03qpvWCW2R c1tFnKF5vACCyBQp1z8WYP9GjjoW5WT33R2iXufgwXP6pkLpS/12qLLeXqO2K4p5 zINKrIkF3P+GHxiDsQZE3G9A4UpKWFHCxKdxyWIV8LQDEBrgE2Mo3NThEyRBbP+8 1dia4j+qFHvPTMNBvBCjCZMqDwbCe9H70WOXKGE36JITW2le91mn4qHl4SuWReUP IoHYDVcC/eBGRegc9X+bLJNjJYqo+XFo6u32/fUC5YVhngycQEi2vg1vv8fWQ7dB g/Ruo3Inrk8h5kPmrHvbOzGazgANIt5ELHrYMRMA5WSgaq29jtGt9oTnsrd+I88G aPJtwAZfLwdSjl/pwJw8atEPrf04DA2w+gO7rZ/AmeLshnGfOTc= =bY+a -----END PGP SIGNATURE----- Merge tag 'for-5.4-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux Pull btrfs updates from David Sterba: "This continues with work on code refactoring, sanity checks and space handling. There are some less user visible changes, nothing that would particularly stand out. User visible changes: - tree checker, more sanity checks of: - ROOT_ITEM (key, size, generation, level, alignment, flags) - EXTENT_ITEM and METADATA_ITEM checks (key, size, offset, alignment, refs) - tree block reference items - EXTENT_DATA_REF (key, hash, offset) - deprecate flag BTRFS_SUBVOL_CREATE_ASYNC for subvolume creation ioctl, scheduled removal in 5.7 - delete stale and unused UAPI definitions BTRFS_DEV_REPLACE_ITEM_STATE_* - improved export of debugging information available via existing sysfs directory structure - try harder to delete relations between qgroups and allow to delete orphan entries - remove unreliable space checks before relocation starts Core: - space handling: - improved ticket reservations and other high level logic in order to remove special cases - factor flushing infrastructure and use it for different contexts, allows to remove some special case handling - reduce metadata reservation when only updating inodes - reduce global block reserve minimum size (affects small filesystems) - improved overcommit logic wrt global block reserve - tests: - fix memory leaks in extent IO tree - catch all TRIM range Fixes: - fix ENOSPC errors, leading to transaction aborts, when cloning extents - several fixes for inode number cache (mount option inode_cache) - fix potential soft lockups during send when traversing large trees - fix unaligned access to space cache pages with SLUB debug on (PowerPC) Other: - refactoring public/private functions, moving to new or more appropriate files - defines converted to enums - error handling improvements - more assertions and comments - old code deletion" * tag 'for-5.4-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux: (138 commits) btrfs: Relinquish CPUs in btrfs_compare_trees btrfs: Don't assign retval of btrfs_try_tree_write_lock/btrfs_tree_read_lock_atomic btrfs: create structure to encode checksum type and length btrfs: turn checksum type define into an enum btrfs: add enospc debug messages for ticket failure btrfs: do not account global reserve in can_overcommit btrfs: use btrfs_try_granting_tickets in update_global_rsv btrfs: always reserve our entire size for the global reserve btrfs: change the minimum global reserve size btrfs: rename btrfs_space_info_add_old_bytes btrfs: remove orig_bytes from reserve_ticket btrfs: fix may_commit_transaction to deal with no partial filling btrfs: rework wake_all_tickets btrfs: refactor the ticket wakeup code btrfs: stop partially refilling tickets when releasing space btrfs: add space reservation tracepoint for reserved bytes btrfs: roll tracepoint into btrfs_space_info_update helper btrfs: do not allow reservations if we have pending tickets btrfs: stop clearing EXTENT_DIRTY in inode I/O tree btrfs: treat RWF_{,D}SYNC writes as sync for CRCs ...
This commit is contained in:
commit
7d14df2d28
|
@ -11,7 +11,7 @@ btrfs-y += super.o ctree.o extent-tree.o print-tree.o root-tree.o dir-item.o \
|
|||
compression.o delayed-ref.o relocation.o delayed-inode.o scrub.o \
|
||||
reada.o backref.o ulist.o qgroup.o send.o dev-replace.o raid56.o \
|
||||
uuid-tree.o props.o free-space-tree.o tree-checker.o space-info.o \
|
||||
block-rsv.o delalloc-space.o
|
||||
block-rsv.o delalloc-space.o block-group.o
|
||||
|
||||
btrfs-$(CONFIG_BTRFS_FS_POSIX_ACL) += acl.o
|
||||
btrfs-$(CONFIG_BTRFS_FS_CHECK_INTEGRITY) += check-integrity.o
|
||||
|
|
|
@ -12,9 +12,11 @@
|
|||
#include "async-thread.h"
|
||||
#include "ctree.h"
|
||||
|
||||
#define WORK_DONE_BIT 0
|
||||
#define WORK_ORDER_DONE_BIT 1
|
||||
#define WORK_HIGH_PRIO_BIT 2
|
||||
enum {
|
||||
WORK_DONE_BIT,
|
||||
WORK_ORDER_DONE_BIT,
|
||||
WORK_HIGH_PRIO_BIT,
|
||||
};
|
||||
|
||||
#define NO_THRESHOLD (-1)
|
||||
#define DFT_THRESHOLD (32)
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,250 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
|
||||
#ifndef BTRFS_BLOCK_GROUP_H
|
||||
#define BTRFS_BLOCK_GROUP_H
|
||||
|
||||
#include "free-space-cache.h"
|
||||
|
||||
enum btrfs_disk_cache_state {
|
||||
BTRFS_DC_WRITTEN,
|
||||
BTRFS_DC_ERROR,
|
||||
BTRFS_DC_CLEAR,
|
||||
BTRFS_DC_SETUP,
|
||||
};
|
||||
|
||||
/*
|
||||
* Control flags for do_chunk_alloc's force field CHUNK_ALLOC_NO_FORCE means to
|
||||
* only allocate a chunk if we really need one.
|
||||
*
|
||||
* CHUNK_ALLOC_LIMITED means to only try and allocate one if we have very few
|
||||
* chunks already allocated. This is used as part of the clustering code to
|
||||
* help make sure we have a good pool of storage to cluster in, without filling
|
||||
* the FS with empty chunks
|
||||
*
|
||||
* CHUNK_ALLOC_FORCE means it must try to allocate one
|
||||
*/
|
||||
enum btrfs_chunk_alloc_enum {
|
||||
CHUNK_ALLOC_NO_FORCE,
|
||||
CHUNK_ALLOC_LIMITED,
|
||||
CHUNK_ALLOC_FORCE,
|
||||
};
|
||||
|
||||
struct btrfs_caching_control {
|
||||
struct list_head list;
|
||||
struct mutex mutex;
|
||||
wait_queue_head_t wait;
|
||||
struct btrfs_work work;
|
||||
struct btrfs_block_group_cache *block_group;
|
||||
u64 progress;
|
||||
refcount_t count;
|
||||
};
|
||||
|
||||
/* Once caching_thread() finds this much free space, it will wake up waiters. */
|
||||
#define CACHING_CTL_WAKE_UP SZ_2M
|
||||
|
||||
struct btrfs_block_group_cache {
|
||||
struct btrfs_key key;
|
||||
struct btrfs_block_group_item item;
|
||||
struct btrfs_fs_info *fs_info;
|
||||
struct inode *inode;
|
||||
spinlock_t lock;
|
||||
u64 pinned;
|
||||
u64 reserved;
|
||||
u64 delalloc_bytes;
|
||||
u64 bytes_super;
|
||||
u64 flags;
|
||||
u64 cache_generation;
|
||||
|
||||
/*
|
||||
* If the free space extent count exceeds this number, convert the block
|
||||
* group to bitmaps.
|
||||
*/
|
||||
u32 bitmap_high_thresh;
|
||||
|
||||
/*
|
||||
* If the free space extent count drops below this number, convert the
|
||||
* block group back to extents.
|
||||
*/
|
||||
u32 bitmap_low_thresh;
|
||||
|
||||
/*
|
||||
* It is just used for the delayed data space allocation because
|
||||
* only the data space allocation and the relative metadata update
|
||||
* can be done cross the transaction.
|
||||
*/
|
||||
struct rw_semaphore data_rwsem;
|
||||
|
||||
/* For raid56, this is a full stripe, without parity */
|
||||
unsigned long full_stripe_len;
|
||||
|
||||
unsigned int ro;
|
||||
unsigned int iref:1;
|
||||
unsigned int has_caching_ctl:1;
|
||||
unsigned int removed:1;
|
||||
|
||||
int disk_cache_state;
|
||||
|
||||
/* Cache tracking stuff */
|
||||
int cached;
|
||||
struct btrfs_caching_control *caching_ctl;
|
||||
u64 last_byte_to_unpin;
|
||||
|
||||
struct btrfs_space_info *space_info;
|
||||
|
||||
/* Free space cache stuff */
|
||||
struct btrfs_free_space_ctl *free_space_ctl;
|
||||
|
||||
/* Block group cache stuff */
|
||||
struct rb_node cache_node;
|
||||
|
||||
/* For block groups in the same raid type */
|
||||
struct list_head list;
|
||||
|
||||
/* Usage count */
|
||||
atomic_t count;
|
||||
|
||||
/*
|
||||
* List of struct btrfs_free_clusters for this block group.
|
||||
* Today it will only have one thing on it, but that may change
|
||||
*/
|
||||
struct list_head cluster_list;
|
||||
|
||||
/* For delayed block group creation or deletion of empty block groups */
|
||||
struct list_head bg_list;
|
||||
|
||||
/* For read-only block groups */
|
||||
struct list_head ro_list;
|
||||
|
||||
atomic_t trimming;
|
||||
|
||||
/* For dirty block groups */
|
||||
struct list_head dirty_list;
|
||||
struct list_head io_list;
|
||||
|
||||
struct btrfs_io_ctl io_ctl;
|
||||
|
||||
/*
|
||||
* Incremented when doing extent allocations and holding a read lock
|
||||
* on the space_info's groups_sem semaphore.
|
||||
* Decremented when an ordered extent that represents an IO against this
|
||||
* block group's range is created (after it's added to its inode's
|
||||
* root's list of ordered extents) or immediately after the allocation
|
||||
* if it's a metadata extent or fallocate extent (for these cases we
|
||||
* don't create ordered extents).
|
||||
*/
|
||||
atomic_t reservations;
|
||||
|
||||
/*
|
||||
* Incremented while holding the spinlock *lock* by a task checking if
|
||||
* it can perform a nocow write (incremented if the value for the *ro*
|
||||
* field is 0). Decremented by such tasks once they create an ordered
|
||||
* extent or before that if some error happens before reaching that step.
|
||||
* This is to prevent races between block group relocation and nocow
|
||||
* writes through direct IO.
|
||||
*/
|
||||
atomic_t nocow_writers;
|
||||
|
||||
/* Lock for free space tree operations. */
|
||||
struct mutex free_space_lock;
|
||||
|
||||
/*
|
||||
* Does the block group need to be added to the free space tree?
|
||||
* Protected by free_space_lock.
|
||||
*/
|
||||
int needs_free_space;
|
||||
|
||||
/* Record locked full stripes for RAID5/6 block group */
|
||||
struct btrfs_full_stripe_locks_tree full_stripe_locks_root;
|
||||
};
|
||||
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
static inline int btrfs_should_fragment_free_space(
|
||||
struct btrfs_block_group_cache *block_group)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = block_group->fs_info;
|
||||
|
||||
return (btrfs_test_opt(fs_info, FRAGMENT_METADATA) &&
|
||||
block_group->flags & BTRFS_BLOCK_GROUP_METADATA) ||
|
||||
(btrfs_test_opt(fs_info, FRAGMENT_DATA) &&
|
||||
block_group->flags & BTRFS_BLOCK_GROUP_DATA);
|
||||
}
|
||||
#endif
|
||||
|
||||
struct btrfs_block_group_cache *btrfs_lookup_first_block_group(
|
||||
struct btrfs_fs_info *info, u64 bytenr);
|
||||
struct btrfs_block_group_cache *btrfs_lookup_block_group(
|
||||
struct btrfs_fs_info *info, u64 bytenr);
|
||||
struct btrfs_block_group_cache *btrfs_next_block_group(
|
||||
struct btrfs_block_group_cache *cache);
|
||||
void btrfs_get_block_group(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_put_block_group(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_dec_block_group_reservations(struct btrfs_fs_info *fs_info,
|
||||
const u64 start);
|
||||
void btrfs_wait_block_group_reservations(struct btrfs_block_group_cache *bg);
|
||||
bool btrfs_inc_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
void btrfs_dec_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
void btrfs_wait_nocow_writers(struct btrfs_block_group_cache *bg);
|
||||
void btrfs_wait_block_group_cache_progress(struct btrfs_block_group_cache *cache,
|
||||
u64 num_bytes);
|
||||
int btrfs_wait_block_group_cache_done(struct btrfs_block_group_cache *cache);
|
||||
int btrfs_cache_block_group(struct btrfs_block_group_cache *cache,
|
||||
int load_cache_only);
|
||||
void btrfs_put_caching_control(struct btrfs_caching_control *ctl);
|
||||
struct btrfs_caching_control *btrfs_get_caching_control(
|
||||
struct btrfs_block_group_cache *cache);
|
||||
u64 add_new_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 start, u64 end);
|
||||
struct btrfs_trans_handle *btrfs_start_trans_remove_block_group(
|
||||
struct btrfs_fs_info *fs_info,
|
||||
const u64 chunk_offset);
|
||||
int btrfs_remove_block_group(struct btrfs_trans_handle *trans,
|
||||
u64 group_start, struct extent_map *em);
|
||||
void btrfs_delete_unused_bgs(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_mark_bg_unused(struct btrfs_block_group_cache *bg);
|
||||
int btrfs_read_block_groups(struct btrfs_fs_info *info);
|
||||
int btrfs_make_block_group(struct btrfs_trans_handle *trans, u64 bytes_used,
|
||||
u64 type, u64 chunk_offset, u64 size);
|
||||
void btrfs_create_pending_block_groups(struct btrfs_trans_handle *trans);
|
||||
int btrfs_inc_block_group_ro(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_dec_block_group_ro(struct btrfs_block_group_cache *cache);
|
||||
int btrfs_start_dirty_block_groups(struct btrfs_trans_handle *trans);
|
||||
int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans);
|
||||
int btrfs_setup_space_cache(struct btrfs_trans_handle *trans);
|
||||
int btrfs_update_block_group(struct btrfs_trans_handle *trans,
|
||||
u64 bytenr, u64 num_bytes, int alloc);
|
||||
int btrfs_add_reserved_bytes(struct btrfs_block_group_cache *cache,
|
||||
u64 ram_bytes, u64 num_bytes, int delalloc);
|
||||
void btrfs_free_reserved_bytes(struct btrfs_block_group_cache *cache,
|
||||
u64 num_bytes, int delalloc);
|
||||
int btrfs_chunk_alloc(struct btrfs_trans_handle *trans, u64 flags,
|
||||
enum btrfs_chunk_alloc_enum force);
|
||||
int btrfs_force_chunk_alloc(struct btrfs_trans_handle *trans, u64 type);
|
||||
void check_system_chunk(struct btrfs_trans_handle *trans, const u64 type);
|
||||
u64 btrfs_get_alloc_profile(struct btrfs_fs_info *fs_info, u64 orig_flags);
|
||||
void btrfs_put_block_group_cache(struct btrfs_fs_info *info);
|
||||
int btrfs_free_block_groups(struct btrfs_fs_info *info);
|
||||
|
||||
static inline u64 btrfs_data_alloc_profile(struct btrfs_fs_info *fs_info)
|
||||
{
|
||||
return btrfs_get_alloc_profile(fs_info, BTRFS_BLOCK_GROUP_DATA);
|
||||
}
|
||||
|
||||
static inline u64 btrfs_metadata_alloc_profile(struct btrfs_fs_info *fs_info)
|
||||
{
|
||||
return btrfs_get_alloc_profile(fs_info, BTRFS_BLOCK_GROUP_METADATA);
|
||||
}
|
||||
|
||||
static inline u64 btrfs_system_alloc_profile(struct btrfs_fs_info *fs_info)
|
||||
{
|
||||
return btrfs_get_alloc_profile(fs_info, BTRFS_BLOCK_GROUP_SYSTEM);
|
||||
}
|
||||
|
||||
static inline int btrfs_block_group_cache_done(
|
||||
struct btrfs_block_group_cache *cache)
|
||||
{
|
||||
smp_mb();
|
||||
return cache->cached == BTRFS_CACHE_FINISHED ||
|
||||
cache->cached == BTRFS_CACHE_ERROR;
|
||||
}
|
||||
|
||||
#endif /* BTRFS_BLOCK_GROUP_H */
|
|
@ -1,9 +1,9 @@
|
|||
// SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "block-rsv.h"
|
||||
#include "space-info.h"
|
||||
#include "math.h"
|
||||
#include "transaction.h"
|
||||
|
||||
static u64 block_rsv_release_bytes(struct btrfs_fs_info *fs_info,
|
||||
|
@ -54,8 +54,9 @@ static u64 block_rsv_release_bytes(struct btrfs_fs_info *fs_info,
|
|||
spin_unlock(&dest->lock);
|
||||
}
|
||||
if (num_bytes)
|
||||
btrfs_space_info_add_old_bytes(fs_info, space_info,
|
||||
num_bytes);
|
||||
btrfs_space_info_free_bytes_may_use(fs_info,
|
||||
space_info,
|
||||
num_bytes);
|
||||
}
|
||||
if (qgroup_to_release_ret)
|
||||
*qgroup_to_release_ret = qgroup_to_release;
|
||||
|
@ -258,6 +259,7 @@ void btrfs_update_global_block_rsv(struct btrfs_fs_info *fs_info)
|
|||
struct btrfs_block_rsv *block_rsv = &fs_info->global_block_rsv;
|
||||
struct btrfs_space_info *sinfo = block_rsv->space_info;
|
||||
u64 num_bytes;
|
||||
unsigned min_items;
|
||||
|
||||
/*
|
||||
* The global block rsv is based on the size of the extent tree, the
|
||||
|
@ -267,7 +269,26 @@ void btrfs_update_global_block_rsv(struct btrfs_fs_info *fs_info)
|
|||
num_bytes = btrfs_root_used(&fs_info->extent_root->root_item) +
|
||||
btrfs_root_used(&fs_info->csum_root->root_item) +
|
||||
btrfs_root_used(&fs_info->tree_root->root_item);
|
||||
num_bytes = max_t(u64, num_bytes, SZ_16M);
|
||||
|
||||
/*
|
||||
* We at a minimum are going to modify the csum root, the tree root, and
|
||||
* the extent root.
|
||||
*/
|
||||
min_items = 3;
|
||||
|
||||
/*
|
||||
* But we also want to reserve enough space so we can do the fallback
|
||||
* global reserve for an unlink, which is an additional 5 items (see the
|
||||
* comment in __unlink_start_trans for what we're modifying.)
|
||||
*
|
||||
* But we also need space for the delayed ref updates from the unlink,
|
||||
* so its 10, 5 for the actual operation, and 5 for the delayed ref
|
||||
* updates.
|
||||
*/
|
||||
min_items += 10;
|
||||
|
||||
num_bytes = max_t(u64, num_bytes,
|
||||
btrfs_calc_insert_metadata_size(fs_info, min_items));
|
||||
|
||||
spin_lock(&sinfo->lock);
|
||||
spin_lock(&block_rsv->lock);
|
||||
|
@ -275,25 +296,16 @@ void btrfs_update_global_block_rsv(struct btrfs_fs_info *fs_info)
|
|||
block_rsv->size = min_t(u64, num_bytes, SZ_512M);
|
||||
|
||||
if (block_rsv->reserved < block_rsv->size) {
|
||||
num_bytes = btrfs_space_info_used(sinfo, true);
|
||||
if (sinfo->total_bytes > num_bytes) {
|
||||
num_bytes = sinfo->total_bytes - num_bytes;
|
||||
num_bytes = min(num_bytes,
|
||||
block_rsv->size - block_rsv->reserved);
|
||||
block_rsv->reserved += num_bytes;
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, sinfo,
|
||||
num_bytes);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
sinfo->flags, num_bytes,
|
||||
1);
|
||||
}
|
||||
num_bytes = block_rsv->size - block_rsv->reserved;
|
||||
block_rsv->reserved += num_bytes;
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, sinfo,
|
||||
num_bytes);
|
||||
} else if (block_rsv->reserved > block_rsv->size) {
|
||||
num_bytes = block_rsv->reserved - block_rsv->size;
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, sinfo,
|
||||
-num_bytes);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
sinfo->flags, num_bytes, 0);
|
||||
block_rsv->reserved = block_rsv->size;
|
||||
btrfs_try_granting_tickets(fs_info, sinfo);
|
||||
}
|
||||
|
||||
if (block_rsv->reserved == block_rsv->size)
|
||||
|
|
|
@ -940,7 +940,7 @@ static void btrfsic_stack_frame_free(struct btrfsic_stack_frame *sf)
|
|||
kfree(sf);
|
||||
}
|
||||
|
||||
static int btrfsic_process_metablock(
|
||||
static noinline_for_stack int btrfsic_process_metablock(
|
||||
struct btrfsic_state *state,
|
||||
struct btrfsic_block *const first_block,
|
||||
struct btrfsic_block_data_ctx *const first_block_ctx,
|
||||
|
@ -1706,8 +1706,9 @@ static void btrfsic_dump_database(struct btrfsic_state *state)
|
|||
* Test whether the disk block contains a tree block (leaf or node)
|
||||
* (note that this test fails for the super block)
|
||||
*/
|
||||
static int btrfsic_test_for_metadata(struct btrfsic_state *state,
|
||||
char **datav, unsigned int num_pages)
|
||||
static noinline_for_stack int btrfsic_test_for_metadata(
|
||||
struct btrfsic_state *state,
|
||||
char **datav, unsigned int num_pages)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = state->fs_info;
|
||||
SHASH_DESC_ON_STACK(shash, fs_info->csum_shash);
|
||||
|
|
|
@ -18,6 +18,7 @@
|
|||
#include <linux/sched/mm.h>
|
||||
#include <linux/log2.h>
|
||||
#include <crypto/hash.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "disk-io.h"
|
||||
#include "transaction.h"
|
||||
|
@ -1039,7 +1040,7 @@ int btrfs_compress_pages(unsigned int type_level, struct address_space *mapping,
|
|||
struct list_head *workspace;
|
||||
int ret;
|
||||
|
||||
level = btrfs_compress_op[type]->set_level(level);
|
||||
level = btrfs_compress_set_level(type, level);
|
||||
workspace = get_workspace(type, level);
|
||||
ret = btrfs_compress_op[type]->compress_pages(workspace, mapping,
|
||||
start, pages,
|
||||
|
@ -1611,7 +1612,23 @@ unsigned int btrfs_compress_str2level(unsigned int type, const char *str)
|
|||
level = 0;
|
||||
}
|
||||
|
||||
level = btrfs_compress_op[type]->set_level(level);
|
||||
level = btrfs_compress_set_level(type, level);
|
||||
|
||||
return level;
|
||||
}
|
||||
|
||||
/*
|
||||
* Adjust @level according to the limits of the compression algorithm or
|
||||
* fallback to default
|
||||
*/
|
||||
unsigned int btrfs_compress_set_level(int type, unsigned level)
|
||||
{
|
||||
const struct btrfs_compress_op *ops = btrfs_compress_op[type];
|
||||
|
||||
if (level == 0)
|
||||
level = ops->default_level;
|
||||
else
|
||||
level = min(level, ops->max_level);
|
||||
|
||||
return level;
|
||||
}
|
||||
|
|
|
@ -156,12 +156,9 @@ struct btrfs_compress_op {
|
|||
unsigned long start_byte,
|
||||
size_t srclen, size_t destlen);
|
||||
|
||||
/*
|
||||
* This bounds the level set by the user to be within range of a
|
||||
* particular compression type. It returns the level that will be used
|
||||
* if the level is out of bounds or the default if 0 is passed in.
|
||||
*/
|
||||
unsigned int (*set_level)(unsigned int level);
|
||||
/* Maximum level supported by the compression algorithm */
|
||||
unsigned int max_level;
|
||||
unsigned int default_level;
|
||||
};
|
||||
|
||||
/* The heuristic workspaces are managed via the 0th workspace manager */
|
||||
|
@ -175,6 +172,8 @@ extern const struct btrfs_compress_op btrfs_zstd_compress;
|
|||
const char* btrfs_compress_type2str(enum btrfs_compression_type type);
|
||||
bool btrfs_compress_is_valid_type(const char *str, size_t len);
|
||||
|
||||
unsigned int btrfs_compress_set_level(int type, unsigned level);
|
||||
|
||||
int btrfs_compress_heuristic(struct inode *inode, u64 start, u64 end);
|
||||
|
||||
#endif
|
||||
|
|
452
fs/btrfs/ctree.c
452
fs/btrfs/ctree.c
|
@ -29,6 +29,28 @@ static int balance_node_right(struct btrfs_trans_handle *trans,
|
|||
static void del_ptr(struct btrfs_root *root, struct btrfs_path *path,
|
||||
int level, int slot);
|
||||
|
||||
static const struct btrfs_csums {
|
||||
u16 size;
|
||||
const char *name;
|
||||
} btrfs_csums[] = {
|
||||
[BTRFS_CSUM_TYPE_CRC32] = { .size = 4, .name = "crc32c" },
|
||||
};
|
||||
|
||||
int btrfs_super_csum_size(const struct btrfs_super_block *s)
|
||||
{
|
||||
u16 t = btrfs_super_csum_type(s);
|
||||
/*
|
||||
* csum type is validated at mount time
|
||||
*/
|
||||
return btrfs_csums[t].size;
|
||||
}
|
||||
|
||||
const char *btrfs_super_csum_name(u16 csum_type)
|
||||
{
|
||||
/* csum type is validated at mount time */
|
||||
return btrfs_csums[csum_type].name;
|
||||
}
|
||||
|
||||
struct btrfs_path *btrfs_alloc_path(void)
|
||||
{
|
||||
return kmem_cache_zalloc(btrfs_path_cachep, GFP_NOFS);
|
||||
|
@ -376,8 +398,6 @@ void btrfs_put_tree_mod_seq(struct btrfs_fs_info *fs_info,
|
|||
* The 'start address' is the logical address of the *new* root node
|
||||
* for root replace operations, or the logical address of the affected
|
||||
* block for all other operations.
|
||||
*
|
||||
* Note: must be called with write lock for fs_info::tree_mod_log_lock.
|
||||
*/
|
||||
static noinline int
|
||||
__tree_mod_log_insert(struct btrfs_fs_info *fs_info, struct tree_mod_elem *tm)
|
||||
|
@ -387,6 +407,8 @@ __tree_mod_log_insert(struct btrfs_fs_info *fs_info, struct tree_mod_elem *tm)
|
|||
struct rb_node *parent = NULL;
|
||||
struct tree_mod_elem *cur;
|
||||
|
||||
lockdep_assert_held_write(&fs_info->tree_mod_log_lock);
|
||||
|
||||
tm->seq = btrfs_inc_tree_mod_seq(fs_info);
|
||||
|
||||
tm_root = &fs_info->tree_mod_log;
|
||||
|
@ -1343,6 +1365,7 @@ get_old_root(struct btrfs_root *root, u64 time_seq)
|
|||
struct tree_mod_elem *tm;
|
||||
struct extent_buffer *eb = NULL;
|
||||
struct extent_buffer *eb_root;
|
||||
u64 eb_root_owner = 0;
|
||||
struct extent_buffer *old;
|
||||
struct tree_mod_root *old_root = NULL;
|
||||
u64 old_generation = 0;
|
||||
|
@ -1380,6 +1403,7 @@ get_old_root(struct btrfs_root *root, u64 time_seq)
|
|||
free_extent_buffer(old);
|
||||
}
|
||||
} else if (old_root) {
|
||||
eb_root_owner = btrfs_header_owner(eb_root);
|
||||
btrfs_tree_read_unlock(eb_root);
|
||||
free_extent_buffer(eb_root);
|
||||
eb = alloc_dummy_extent_buffer(fs_info, logical);
|
||||
|
@ -1396,7 +1420,7 @@ get_old_root(struct btrfs_root *root, u64 time_seq)
|
|||
if (old_root) {
|
||||
btrfs_set_header_bytenr(eb, eb->start);
|
||||
btrfs_set_header_backref_rev(eb, BTRFS_MIXED_BACKREF_REV);
|
||||
btrfs_set_header_owner(eb, btrfs_header_owner(eb_root));
|
||||
btrfs_set_header_owner(eb, eb_root_owner);
|
||||
btrfs_set_header_level(eb, old_root->level);
|
||||
btrfs_set_header_generation(eb, old_generation);
|
||||
}
|
||||
|
@ -1790,8 +1814,8 @@ static void root_sub_used(struct btrfs_root *root, u32 size)
|
|||
/* given a node and slot number, this reads the blocks it points to. The
|
||||
* extent buffer is returned with a reference taken (but unlocked).
|
||||
*/
|
||||
static noinline struct extent_buffer *read_node_slot(
|
||||
struct extent_buffer *parent, int slot)
|
||||
struct extent_buffer *btrfs_read_node_slot(struct extent_buffer *parent,
|
||||
int slot)
|
||||
{
|
||||
int level = btrfs_header_level(parent);
|
||||
struct extent_buffer *eb;
|
||||
|
@ -1860,7 +1884,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans,
|
|||
return 0;
|
||||
|
||||
/* promote the child to a root */
|
||||
child = read_node_slot(mid, 0);
|
||||
child = btrfs_read_node_slot(mid, 0);
|
||||
if (IS_ERR(child)) {
|
||||
ret = PTR_ERR(child);
|
||||
btrfs_handle_fs_error(fs_info, ret, NULL);
|
||||
|
@ -1900,7 +1924,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans,
|
|||
BTRFS_NODEPTRS_PER_BLOCK(fs_info) / 4)
|
||||
return 0;
|
||||
|
||||
left = read_node_slot(parent, pslot - 1);
|
||||
left = btrfs_read_node_slot(parent, pslot - 1);
|
||||
if (IS_ERR(left))
|
||||
left = NULL;
|
||||
|
||||
|
@ -1915,7 +1939,7 @@ static noinline int balance_level(struct btrfs_trans_handle *trans,
|
|||
}
|
||||
}
|
||||
|
||||
right = read_node_slot(parent, pslot + 1);
|
||||
right = btrfs_read_node_slot(parent, pslot + 1);
|
||||
if (IS_ERR(right))
|
||||
right = NULL;
|
||||
|
||||
|
@ -2075,7 +2099,7 @@ static noinline int push_nodes_for_insert(struct btrfs_trans_handle *trans,
|
|||
if (!parent)
|
||||
return 1;
|
||||
|
||||
left = read_node_slot(parent, pslot - 1);
|
||||
left = btrfs_read_node_slot(parent, pslot - 1);
|
||||
if (IS_ERR(left))
|
||||
left = NULL;
|
||||
|
||||
|
@ -2127,7 +2151,7 @@ static noinline int push_nodes_for_insert(struct btrfs_trans_handle *trans,
|
|||
btrfs_tree_unlock(left);
|
||||
free_extent_buffer(left);
|
||||
}
|
||||
right = read_node_slot(parent, pslot + 1);
|
||||
right = btrfs_read_node_slot(parent, pslot + 1);
|
||||
if (IS_ERR(right))
|
||||
right = NULL;
|
||||
|
||||
|
@ -2889,15 +2913,13 @@ cow_done:
|
|||
if (!p->skip_locking) {
|
||||
level = btrfs_header_level(b);
|
||||
if (level <= write_lock_level) {
|
||||
err = btrfs_try_tree_write_lock(b);
|
||||
if (!err) {
|
||||
if (!btrfs_try_tree_write_lock(b)) {
|
||||
btrfs_set_path_blocking(p);
|
||||
btrfs_tree_lock(b);
|
||||
}
|
||||
p->locks[level] = BTRFS_WRITE_LOCK;
|
||||
} else {
|
||||
err = btrfs_tree_read_lock_atomic(b);
|
||||
if (!err) {
|
||||
if (!btrfs_tree_read_lock_atomic(b)) {
|
||||
btrfs_set_path_blocking(p);
|
||||
btrfs_tree_read_lock(b);
|
||||
}
|
||||
|
@ -3031,8 +3053,7 @@ again:
|
|||
}
|
||||
|
||||
level = btrfs_header_level(b);
|
||||
err = btrfs_tree_read_lock_atomic(b);
|
||||
if (!err) {
|
||||
if (!btrfs_tree_read_lock_atomic(b)) {
|
||||
btrfs_set_path_blocking(p);
|
||||
btrfs_tree_read_lock(b);
|
||||
}
|
||||
|
@ -3572,7 +3593,7 @@ static int leaf_space_used(struct extent_buffer *l, int start, int nr)
|
|||
|
||||
if (!nr)
|
||||
return 0;
|
||||
btrfs_init_map_token(&token);
|
||||
btrfs_init_map_token(&token, l);
|
||||
start_item = btrfs_item_nr(start);
|
||||
end_item = btrfs_item_nr(end);
|
||||
data_len = btrfs_token_item_offset(l, start_item, &token) +
|
||||
|
@ -3630,8 +3651,6 @@ static noinline int __push_leaf_right(struct btrfs_path *path,
|
|||
u32 data_end;
|
||||
u32 this_item_size;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
if (empty)
|
||||
nr = 0;
|
||||
else
|
||||
|
@ -3704,6 +3723,7 @@ static noinline int __push_leaf_right(struct btrfs_path *path,
|
|||
push_items * sizeof(struct btrfs_item));
|
||||
|
||||
/* update the item pointers */
|
||||
btrfs_init_map_token(&token, right);
|
||||
right_nritems += push_items;
|
||||
btrfs_set_header_nritems(right, right_nritems);
|
||||
push_space = BTRFS_LEAF_DATA_SIZE(fs_info);
|
||||
|
@ -3781,7 +3801,7 @@ static int push_leaf_right(struct btrfs_trans_handle *trans, struct btrfs_root
|
|||
|
||||
btrfs_assert_tree_locked(path->nodes[1]);
|
||||
|
||||
right = read_node_slot(upper, slot + 1);
|
||||
right = btrfs_read_node_slot(upper, slot + 1);
|
||||
/*
|
||||
* slot + 1 is not valid or we fail to read the right node,
|
||||
* no big deal, just return.
|
||||
|
@ -3858,8 +3878,6 @@ static noinline int __push_leaf_left(struct btrfs_path *path, int data_size,
|
|||
u32 old_left_item_size;
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
if (empty)
|
||||
nr = min(right_nritems, max_slot);
|
||||
else
|
||||
|
@ -3913,6 +3931,7 @@ static noinline int __push_leaf_left(struct btrfs_path *path, int data_size,
|
|||
old_left_nritems = btrfs_header_nritems(left);
|
||||
BUG_ON(old_left_nritems <= 0);
|
||||
|
||||
btrfs_init_map_token(&token, left);
|
||||
old_left_item_size = btrfs_item_offset_nr(left, old_left_nritems - 1);
|
||||
for (i = old_left_nritems; i < old_left_nritems + push_items; i++) {
|
||||
u32 ioff;
|
||||
|
@ -3944,6 +3963,8 @@ static noinline int __push_leaf_left(struct btrfs_path *path, int data_size,
|
|||
(btrfs_header_nritems(right) - push_items) *
|
||||
sizeof(struct btrfs_item));
|
||||
}
|
||||
|
||||
btrfs_init_map_token(&token, right);
|
||||
right_nritems -= push_items;
|
||||
btrfs_set_header_nritems(right, right_nritems);
|
||||
push_space = BTRFS_LEAF_DATA_SIZE(fs_info);
|
||||
|
@ -4015,7 +4036,7 @@ static int push_leaf_left(struct btrfs_trans_handle *trans, struct btrfs_root
|
|||
|
||||
btrfs_assert_tree_locked(path->nodes[1]);
|
||||
|
||||
left = read_node_slot(path->nodes[1], slot - 1);
|
||||
left = btrfs_read_node_slot(path->nodes[1], slot - 1);
|
||||
/*
|
||||
* slot - 1 is not valid or we fail to read the left node,
|
||||
* no big deal, just return.
|
||||
|
@ -4074,8 +4095,6 @@ static noinline void copy_for_split(struct btrfs_trans_handle *trans,
|
|||
struct btrfs_disk_key disk_key;
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
nritems = nritems - mid;
|
||||
btrfs_set_header_nritems(right, nritems);
|
||||
data_copy_size = btrfs_item_end_nr(l, mid) - leaf_data_end(l);
|
||||
|
@ -4091,6 +4110,7 @@ static noinline void copy_for_split(struct btrfs_trans_handle *trans,
|
|||
|
||||
rt_data_off = BTRFS_LEAF_DATA_SIZE(fs_info) - btrfs_item_end_nr(l, mid);
|
||||
|
||||
btrfs_init_map_token(&token, right);
|
||||
for (i = 0; i < nritems; i++) {
|
||||
struct btrfs_item *item = btrfs_item_nr(i);
|
||||
u32 ioff;
|
||||
|
@ -4574,8 +4594,6 @@ void btrfs_truncate_item(struct btrfs_path *path, u32 new_size, int from_end)
|
|||
int i;
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
leaf = path->nodes[0];
|
||||
slot = path->slots[0];
|
||||
|
||||
|
@ -4597,6 +4615,7 @@ void btrfs_truncate_item(struct btrfs_path *path, u32 new_size, int from_end)
|
|||
* item0..itemN ... dataN.offset..dataN.size .. data0.size
|
||||
*/
|
||||
/* first correct the data pointers */
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
for (i = slot; i < nritems; i++) {
|
||||
u32 ioff;
|
||||
item = btrfs_item_nr(i);
|
||||
|
@ -4671,8 +4690,6 @@ void btrfs_extend_item(struct btrfs_path *path, u32 data_size)
|
|||
int i;
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
leaf = path->nodes[0];
|
||||
|
||||
nritems = btrfs_header_nritems(leaf);
|
||||
|
@ -4697,6 +4714,7 @@ void btrfs_extend_item(struct btrfs_path *path, u32 data_size)
|
|||
* item0..itemN ... dataN.offset..dataN.size .. data0.size
|
||||
*/
|
||||
/* first correct the data pointers */
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
for (i = slot; i < nritems; i++) {
|
||||
u32 ioff;
|
||||
item = btrfs_item_nr(i);
|
||||
|
@ -4748,8 +4766,6 @@ void setup_items_for_insert(struct btrfs_root *root, struct btrfs_path *path,
|
|||
}
|
||||
btrfs_unlock_up_safe(path, 1);
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
leaf = path->nodes[0];
|
||||
slot = path->slots[0];
|
||||
|
||||
|
@ -4763,6 +4779,7 @@ void setup_items_for_insert(struct btrfs_root *root, struct btrfs_path *path,
|
|||
BUG();
|
||||
}
|
||||
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
if (slot != nritems) {
|
||||
unsigned int old_data = btrfs_item_end_nr(leaf, slot);
|
||||
|
||||
|
@ -4969,9 +4986,6 @@ int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root,
|
|||
int wret;
|
||||
int i;
|
||||
u32 nritems;
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
leaf = path->nodes[0];
|
||||
last_off = btrfs_item_offset_nr(leaf, slot + nr - 1);
|
||||
|
@ -4983,12 +4997,14 @@ int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root,
|
|||
|
||||
if (slot + nr != nritems) {
|
||||
int data_end = leaf_data_end(leaf);
|
||||
struct btrfs_map_token token;
|
||||
|
||||
memmove_extent_buffer(leaf, BTRFS_LEAF_DATA_OFFSET +
|
||||
data_end + dsize,
|
||||
BTRFS_LEAF_DATA_OFFSET + data_end,
|
||||
last_off - data_end);
|
||||
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
for (i = slot + nr; i < nritems; i++) {
|
||||
u32 ioff;
|
||||
|
||||
|
@ -5222,7 +5238,7 @@ find_next_key:
|
|||
goto out;
|
||||
}
|
||||
btrfs_set_path_blocking(path);
|
||||
cur = read_node_slot(cur, slot);
|
||||
cur = btrfs_read_node_slot(cur, slot);
|
||||
if (IS_ERR(cur)) {
|
||||
ret = PTR_ERR(cur);
|
||||
goto out;
|
||||
|
@ -5244,368 +5260,6 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static int tree_move_down(struct btrfs_path *path, int *level)
|
||||
{
|
||||
struct extent_buffer *eb;
|
||||
|
||||
BUG_ON(*level == 0);
|
||||
eb = read_node_slot(path->nodes[*level], path->slots[*level]);
|
||||
if (IS_ERR(eb))
|
||||
return PTR_ERR(eb);
|
||||
|
||||
path->nodes[*level - 1] = eb;
|
||||
path->slots[*level - 1] = 0;
|
||||
(*level)--;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tree_move_next_or_upnext(struct btrfs_path *path,
|
||||
int *level, int root_level)
|
||||
{
|
||||
int ret = 0;
|
||||
int nritems;
|
||||
nritems = btrfs_header_nritems(path->nodes[*level]);
|
||||
|
||||
path->slots[*level]++;
|
||||
|
||||
while (path->slots[*level] >= nritems) {
|
||||
if (*level == root_level)
|
||||
return -1;
|
||||
|
||||
/* move upnext */
|
||||
path->slots[*level] = 0;
|
||||
free_extent_buffer(path->nodes[*level]);
|
||||
path->nodes[*level] = NULL;
|
||||
(*level)++;
|
||||
path->slots[*level]++;
|
||||
|
||||
nritems = btrfs_header_nritems(path->nodes[*level]);
|
||||
ret = 1;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Returns 1 if it had to move up and next. 0 is returned if it moved only next
|
||||
* or down.
|
||||
*/
|
||||
static int tree_advance(struct btrfs_path *path,
|
||||
int *level, int root_level,
|
||||
int allow_down,
|
||||
struct btrfs_key *key)
|
||||
{
|
||||
int ret;
|
||||
|
||||
if (*level == 0 || !allow_down) {
|
||||
ret = tree_move_next_or_upnext(path, level, root_level);
|
||||
} else {
|
||||
ret = tree_move_down(path, level);
|
||||
}
|
||||
if (ret >= 0) {
|
||||
if (*level == 0)
|
||||
btrfs_item_key_to_cpu(path->nodes[*level], key,
|
||||
path->slots[*level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(path->nodes[*level], key,
|
||||
path->slots[*level]);
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int tree_compare_item(struct btrfs_path *left_path,
|
||||
struct btrfs_path *right_path,
|
||||
char *tmp_buf)
|
||||
{
|
||||
int cmp;
|
||||
int len1, len2;
|
||||
unsigned long off1, off2;
|
||||
|
||||
len1 = btrfs_item_size_nr(left_path->nodes[0], left_path->slots[0]);
|
||||
len2 = btrfs_item_size_nr(right_path->nodes[0], right_path->slots[0]);
|
||||
if (len1 != len2)
|
||||
return 1;
|
||||
|
||||
off1 = btrfs_item_ptr_offset(left_path->nodes[0], left_path->slots[0]);
|
||||
off2 = btrfs_item_ptr_offset(right_path->nodes[0],
|
||||
right_path->slots[0]);
|
||||
|
||||
read_extent_buffer(left_path->nodes[0], tmp_buf, off1, len1);
|
||||
|
||||
cmp = memcmp_extent_buffer(right_path->nodes[0], tmp_buf, off2, len1);
|
||||
if (cmp)
|
||||
return 1;
|
||||
return 0;
|
||||
}
|
||||
|
||||
#define ADVANCE 1
|
||||
#define ADVANCE_ONLY_NEXT -1
|
||||
|
||||
/*
|
||||
* This function compares two trees and calls the provided callback for
|
||||
* every changed/new/deleted item it finds.
|
||||
* If shared tree blocks are encountered, whole subtrees are skipped, making
|
||||
* the compare pretty fast on snapshotted subvolumes.
|
||||
*
|
||||
* This currently works on commit roots only. As commit roots are read only,
|
||||
* we don't do any locking. The commit roots are protected with transactions.
|
||||
* Transactions are ended and rejoined when a commit is tried in between.
|
||||
*
|
||||
* This function checks for modifications done to the trees while comparing.
|
||||
* If it detects a change, it aborts immediately.
|
||||
*/
|
||||
int btrfs_compare_trees(struct btrfs_root *left_root,
|
||||
struct btrfs_root *right_root,
|
||||
btrfs_changed_cb_t changed_cb, void *ctx)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = left_root->fs_info;
|
||||
int ret;
|
||||
int cmp;
|
||||
struct btrfs_path *left_path = NULL;
|
||||
struct btrfs_path *right_path = NULL;
|
||||
struct btrfs_key left_key;
|
||||
struct btrfs_key right_key;
|
||||
char *tmp_buf = NULL;
|
||||
int left_root_level;
|
||||
int right_root_level;
|
||||
int left_level;
|
||||
int right_level;
|
||||
int left_end_reached;
|
||||
int right_end_reached;
|
||||
int advance_left;
|
||||
int advance_right;
|
||||
u64 left_blockptr;
|
||||
u64 right_blockptr;
|
||||
u64 left_gen;
|
||||
u64 right_gen;
|
||||
|
||||
left_path = btrfs_alloc_path();
|
||||
if (!left_path) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
right_path = btrfs_alloc_path();
|
||||
if (!right_path) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
tmp_buf = kvmalloc(fs_info->nodesize, GFP_KERNEL);
|
||||
if (!tmp_buf) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
left_path->search_commit_root = 1;
|
||||
left_path->skip_locking = 1;
|
||||
right_path->search_commit_root = 1;
|
||||
right_path->skip_locking = 1;
|
||||
|
||||
/*
|
||||
* Strategy: Go to the first items of both trees. Then do
|
||||
*
|
||||
* If both trees are at level 0
|
||||
* Compare keys of current items
|
||||
* If left < right treat left item as new, advance left tree
|
||||
* and repeat
|
||||
* If left > right treat right item as deleted, advance right tree
|
||||
* and repeat
|
||||
* If left == right do deep compare of items, treat as changed if
|
||||
* needed, advance both trees and repeat
|
||||
* If both trees are at the same level but not at level 0
|
||||
* Compare keys of current nodes/leafs
|
||||
* If left < right advance left tree and repeat
|
||||
* If left > right advance right tree and repeat
|
||||
* If left == right compare blockptrs of the next nodes/leafs
|
||||
* If they match advance both trees but stay at the same level
|
||||
* and repeat
|
||||
* If they don't match advance both trees while allowing to go
|
||||
* deeper and repeat
|
||||
* If tree levels are different
|
||||
* Advance the tree that needs it and repeat
|
||||
*
|
||||
* Advancing a tree means:
|
||||
* If we are at level 0, try to go to the next slot. If that's not
|
||||
* possible, go one level up and repeat. Stop when we found a level
|
||||
* where we could go to the next slot. We may at this point be on a
|
||||
* node or a leaf.
|
||||
*
|
||||
* If we are not at level 0 and not on shared tree blocks, go one
|
||||
* level deeper.
|
||||
*
|
||||
* If we are not at level 0 and on shared tree blocks, go one slot to
|
||||
* the right if possible or go up and right.
|
||||
*/
|
||||
|
||||
down_read(&fs_info->commit_root_sem);
|
||||
left_level = btrfs_header_level(left_root->commit_root);
|
||||
left_root_level = left_level;
|
||||
left_path->nodes[left_level] =
|
||||
btrfs_clone_extent_buffer(left_root->commit_root);
|
||||
if (!left_path->nodes[left_level]) {
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
right_level = btrfs_header_level(right_root->commit_root);
|
||||
right_root_level = right_level;
|
||||
right_path->nodes[right_level] =
|
||||
btrfs_clone_extent_buffer(right_root->commit_root);
|
||||
if (!right_path->nodes[right_level]) {
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
|
||||
if (left_level == 0)
|
||||
btrfs_item_key_to_cpu(left_path->nodes[left_level],
|
||||
&left_key, left_path->slots[left_level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(left_path->nodes[left_level],
|
||||
&left_key, left_path->slots[left_level]);
|
||||
if (right_level == 0)
|
||||
btrfs_item_key_to_cpu(right_path->nodes[right_level],
|
||||
&right_key, right_path->slots[right_level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(right_path->nodes[right_level],
|
||||
&right_key, right_path->slots[right_level]);
|
||||
|
||||
left_end_reached = right_end_reached = 0;
|
||||
advance_left = advance_right = 0;
|
||||
|
||||
while (1) {
|
||||
if (advance_left && !left_end_reached) {
|
||||
ret = tree_advance(left_path, &left_level,
|
||||
left_root_level,
|
||||
advance_left != ADVANCE_ONLY_NEXT,
|
||||
&left_key);
|
||||
if (ret == -1)
|
||||
left_end_reached = ADVANCE;
|
||||
else if (ret < 0)
|
||||
goto out;
|
||||
advance_left = 0;
|
||||
}
|
||||
if (advance_right && !right_end_reached) {
|
||||
ret = tree_advance(right_path, &right_level,
|
||||
right_root_level,
|
||||
advance_right != ADVANCE_ONLY_NEXT,
|
||||
&right_key);
|
||||
if (ret == -1)
|
||||
right_end_reached = ADVANCE;
|
||||
else if (ret < 0)
|
||||
goto out;
|
||||
advance_right = 0;
|
||||
}
|
||||
|
||||
if (left_end_reached && right_end_reached) {
|
||||
ret = 0;
|
||||
goto out;
|
||||
} else if (left_end_reached) {
|
||||
if (right_level == 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&right_key,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
}
|
||||
advance_right = ADVANCE;
|
||||
continue;
|
||||
} else if (right_end_reached) {
|
||||
if (left_level == 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key,
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
}
|
||||
advance_left = ADVANCE;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (left_level == 0 && right_level == 0) {
|
||||
cmp = btrfs_comp_cpu_keys(&left_key, &right_key);
|
||||
if (cmp < 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key,
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_left = ADVANCE;
|
||||
} else if (cmp > 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&right_key,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
enum btrfs_compare_tree_result result;
|
||||
|
||||
WARN_ON(!extent_buffer_uptodate(left_path->nodes[0]));
|
||||
ret = tree_compare_item(left_path, right_path,
|
||||
tmp_buf);
|
||||
if (ret)
|
||||
result = BTRFS_COMPARE_TREE_CHANGED;
|
||||
else
|
||||
result = BTRFS_COMPARE_TREE_SAME;
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key, result, ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_left = ADVANCE;
|
||||
advance_right = ADVANCE;
|
||||
}
|
||||
} else if (left_level == right_level) {
|
||||
cmp = btrfs_comp_cpu_keys(&left_key, &right_key);
|
||||
if (cmp < 0) {
|
||||
advance_left = ADVANCE;
|
||||
} else if (cmp > 0) {
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
left_blockptr = btrfs_node_blockptr(
|
||||
left_path->nodes[left_level],
|
||||
left_path->slots[left_level]);
|
||||
right_blockptr = btrfs_node_blockptr(
|
||||
right_path->nodes[right_level],
|
||||
right_path->slots[right_level]);
|
||||
left_gen = btrfs_node_ptr_generation(
|
||||
left_path->nodes[left_level],
|
||||
left_path->slots[left_level]);
|
||||
right_gen = btrfs_node_ptr_generation(
|
||||
right_path->nodes[right_level],
|
||||
right_path->slots[right_level]);
|
||||
if (left_blockptr == right_blockptr &&
|
||||
left_gen == right_gen) {
|
||||
/*
|
||||
* As we're on a shared block, don't
|
||||
* allow to go deeper.
|
||||
*/
|
||||
advance_left = ADVANCE_ONLY_NEXT;
|
||||
advance_right = ADVANCE_ONLY_NEXT;
|
||||
} else {
|
||||
advance_left = ADVANCE;
|
||||
advance_right = ADVANCE;
|
||||
}
|
||||
}
|
||||
} else if (left_level < right_level) {
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
advance_left = ADVANCE;
|
||||
}
|
||||
}
|
||||
|
||||
out:
|
||||
btrfs_free_path(left_path);
|
||||
btrfs_free_path(right_path);
|
||||
kvfree(tmp_buf);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* this is similar to btrfs_next_leaf, but does not try to preserve
|
||||
* and fixup the path. It looks for and returns the next key in the
|
||||
|
@ -5623,7 +5277,7 @@ int btrfs_find_next_key(struct btrfs_root *root, struct btrfs_path *path,
|
|||
int slot;
|
||||
struct extent_buffer *c;
|
||||
|
||||
WARN_ON(!path->keep_locks);
|
||||
WARN_ON(!path->keep_locks && !path->skip_locking);
|
||||
while (level < BTRFS_MAX_LEVEL) {
|
||||
if (!path->nodes[level])
|
||||
return 1;
|
||||
|
@ -5639,7 +5293,7 @@ next:
|
|||
!path->nodes[level + 1])
|
||||
return 1;
|
||||
|
||||
if (path->locks[level + 1]) {
|
||||
if (path->locks[level + 1] || path->skip_locking) {
|
||||
level++;
|
||||
continue;
|
||||
}
|
||||
|
|
417
fs/btrfs/ctree.h
417
fs/btrfs/ctree.h
|
@ -16,7 +16,6 @@
|
|||
#include <linux/backing-dev.h>
|
||||
#include <linux/wait.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/kobject.h>
|
||||
#include <trace/events/btrfs.h>
|
||||
#include <asm/kmap_types.h>
|
||||
#include <asm/unaligned.h>
|
||||
|
@ -39,10 +38,12 @@ struct btrfs_transaction;
|
|||
struct btrfs_pending_snapshot;
|
||||
struct btrfs_delayed_ref_root;
|
||||
struct btrfs_space_info;
|
||||
struct btrfs_block_group_cache;
|
||||
extern struct kmem_cache *btrfs_trans_handle_cachep;
|
||||
extern struct kmem_cache *btrfs_bit_radix_cachep;
|
||||
extern struct kmem_cache *btrfs_path_cachep;
|
||||
extern struct kmem_cache *btrfs_free_space_cachep;
|
||||
extern struct kmem_cache *btrfs_free_space_bitmap_cachep;
|
||||
struct btrfs_ordered_sum;
|
||||
struct btrfs_ref;
|
||||
|
||||
|
@ -82,10 +83,6 @@ struct btrfs_ref;
|
|||
*/
|
||||
#define BTRFS_LINK_MAX 65535U
|
||||
|
||||
/* four bytes for CRC32 */
|
||||
static const int btrfs_csum_sizes[] = { 4 };
|
||||
static const char *btrfs_csum_names[] = { "crc32c" };
|
||||
|
||||
#define BTRFS_EMPTY_DIR_SIZE 0
|
||||
|
||||
/* ioprio of readahead is set to idle */
|
||||
|
@ -397,12 +394,6 @@ struct btrfs_dev_replace {
|
|||
wait_queue_head_t replace_wait;
|
||||
};
|
||||
|
||||
/* For raid type sysfs entries */
|
||||
struct raid_kobject {
|
||||
u64 flags;
|
||||
struct kobject kobj;
|
||||
};
|
||||
|
||||
/*
|
||||
* free clusters are used to claim free space in relatively large chunks,
|
||||
* allowing us to do less seeky writes. They are used for all metadata
|
||||
|
@ -439,40 +430,6 @@ enum btrfs_caching_type {
|
|||
BTRFS_CACHE_ERROR,
|
||||
};
|
||||
|
||||
enum btrfs_disk_cache_state {
|
||||
BTRFS_DC_WRITTEN,
|
||||
BTRFS_DC_ERROR,
|
||||
BTRFS_DC_CLEAR,
|
||||
BTRFS_DC_SETUP,
|
||||
};
|
||||
|
||||
struct btrfs_caching_control {
|
||||
struct list_head list;
|
||||
struct mutex mutex;
|
||||
wait_queue_head_t wait;
|
||||
struct btrfs_work work;
|
||||
struct btrfs_block_group_cache *block_group;
|
||||
u64 progress;
|
||||
refcount_t count;
|
||||
};
|
||||
|
||||
/* Once caching_thread() finds this much free space, it will wake up waiters. */
|
||||
#define CACHING_CTL_WAKE_UP SZ_2M
|
||||
|
||||
struct btrfs_io_ctl {
|
||||
void *cur, *orig;
|
||||
struct page *page;
|
||||
struct page **pages;
|
||||
struct btrfs_fs_info *fs_info;
|
||||
struct inode *inode;
|
||||
unsigned long size;
|
||||
int index;
|
||||
int num_pages;
|
||||
int entries;
|
||||
int bitmaps;
|
||||
unsigned check_crcs:1;
|
||||
};
|
||||
|
||||
/*
|
||||
* Tree to record all locked full stripes of a RAID5/6 block group
|
||||
*/
|
||||
|
@ -481,120 +438,6 @@ struct btrfs_full_stripe_locks_tree {
|
|||
struct mutex lock;
|
||||
};
|
||||
|
||||
struct btrfs_block_group_cache {
|
||||
struct btrfs_key key;
|
||||
struct btrfs_block_group_item item;
|
||||
struct btrfs_fs_info *fs_info;
|
||||
struct inode *inode;
|
||||
spinlock_t lock;
|
||||
u64 pinned;
|
||||
u64 reserved;
|
||||
u64 delalloc_bytes;
|
||||
u64 bytes_super;
|
||||
u64 flags;
|
||||
u64 cache_generation;
|
||||
|
||||
/*
|
||||
* If the free space extent count exceeds this number, convert the block
|
||||
* group to bitmaps.
|
||||
*/
|
||||
u32 bitmap_high_thresh;
|
||||
|
||||
/*
|
||||
* If the free space extent count drops below this number, convert the
|
||||
* block group back to extents.
|
||||
*/
|
||||
u32 bitmap_low_thresh;
|
||||
|
||||
/*
|
||||
* It is just used for the delayed data space allocation because
|
||||
* only the data space allocation and the relative metadata update
|
||||
* can be done cross the transaction.
|
||||
*/
|
||||
struct rw_semaphore data_rwsem;
|
||||
|
||||
/* for raid56, this is a full stripe, without parity */
|
||||
unsigned long full_stripe_len;
|
||||
|
||||
unsigned int ro;
|
||||
unsigned int iref:1;
|
||||
unsigned int has_caching_ctl:1;
|
||||
unsigned int removed:1;
|
||||
|
||||
int disk_cache_state;
|
||||
|
||||
/* cache tracking stuff */
|
||||
int cached;
|
||||
struct btrfs_caching_control *caching_ctl;
|
||||
u64 last_byte_to_unpin;
|
||||
|
||||
struct btrfs_space_info *space_info;
|
||||
|
||||
/* free space cache stuff */
|
||||
struct btrfs_free_space_ctl *free_space_ctl;
|
||||
|
||||
/* block group cache stuff */
|
||||
struct rb_node cache_node;
|
||||
|
||||
/* for block groups in the same raid type */
|
||||
struct list_head list;
|
||||
|
||||
/* usage count */
|
||||
atomic_t count;
|
||||
|
||||
/* List of struct btrfs_free_clusters for this block group.
|
||||
* Today it will only have one thing on it, but that may change
|
||||
*/
|
||||
struct list_head cluster_list;
|
||||
|
||||
/* For delayed block group creation or deletion of empty block groups */
|
||||
struct list_head bg_list;
|
||||
|
||||
/* For read-only block groups */
|
||||
struct list_head ro_list;
|
||||
|
||||
atomic_t trimming;
|
||||
|
||||
/* For dirty block groups */
|
||||
struct list_head dirty_list;
|
||||
struct list_head io_list;
|
||||
|
||||
struct btrfs_io_ctl io_ctl;
|
||||
|
||||
/*
|
||||
* Incremented when doing extent allocations and holding a read lock
|
||||
* on the space_info's groups_sem semaphore.
|
||||
* Decremented when an ordered extent that represents an IO against this
|
||||
* block group's range is created (after it's added to its inode's
|
||||
* root's list of ordered extents) or immediately after the allocation
|
||||
* if it's a metadata extent or fallocate extent (for these cases we
|
||||
* don't create ordered extents).
|
||||
*/
|
||||
atomic_t reservations;
|
||||
|
||||
/*
|
||||
* Incremented while holding the spinlock *lock* by a task checking if
|
||||
* it can perform a nocow write (incremented if the value for the *ro*
|
||||
* field is 0). Decremented by such tasks once they create an ordered
|
||||
* extent or before that if some error happens before reaching that step.
|
||||
* This is to prevent races between block group relocation and nocow
|
||||
* writes through direct IO.
|
||||
*/
|
||||
atomic_t nocow_writers;
|
||||
|
||||
/* Lock for free space tree operations. */
|
||||
struct mutex free_space_lock;
|
||||
|
||||
/*
|
||||
* Does the block group need to be added to the free space tree?
|
||||
* Protected by free_space_lock.
|
||||
*/
|
||||
int needs_free_space;
|
||||
|
||||
/* Record locked full stripes for RAID5/6 block group */
|
||||
struct btrfs_full_stripe_locks_tree full_stripe_locks_root;
|
||||
};
|
||||
|
||||
/* delayed seq elem */
|
||||
struct seq_list {
|
||||
struct list_head list;
|
||||
|
@ -610,22 +453,6 @@ enum btrfs_orphan_cleanup_state {
|
|||
ORPHAN_CLEANUP_DONE = 2,
|
||||
};
|
||||
|
||||
/* used by the raid56 code to lock stripes for read/modify/write */
|
||||
struct btrfs_stripe_hash {
|
||||
struct list_head hash_list;
|
||||
spinlock_t lock;
|
||||
};
|
||||
|
||||
/* used by the raid56 code to lock stripes for read/modify/write */
|
||||
struct btrfs_stripe_hash_table {
|
||||
struct list_head stripe_cache;
|
||||
spinlock_t cache_lock;
|
||||
int cache_size;
|
||||
struct btrfs_stripe_hash table[];
|
||||
};
|
||||
|
||||
#define BTRFS_STRIPE_HASH_TABLE_BITS 11
|
||||
|
||||
void btrfs_init_async_reclaim_work(struct work_struct *work);
|
||||
|
||||
/* fs_info */
|
||||
|
@ -1279,6 +1106,16 @@ struct btrfs_root {
|
|||
#endif
|
||||
};
|
||||
|
||||
struct btrfs_clone_extent_info {
|
||||
u64 disk_offset;
|
||||
u64 disk_len;
|
||||
u64 data_offset;
|
||||
u64 data_len;
|
||||
u64 file_offset;
|
||||
char *extent_buf;
|
||||
u32 item_size;
|
||||
};
|
||||
|
||||
struct btrfs_file_private {
|
||||
void *filldir_buf;
|
||||
};
|
||||
|
@ -1377,19 +1214,6 @@ static inline u32 BTRFS_MAX_XATTR_SIZE(const struct btrfs_fs_info *info)
|
|||
btrfs_clear_opt(fs_info->mount_opt, opt); \
|
||||
}
|
||||
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
static inline int
|
||||
btrfs_should_fragment_free_space(struct btrfs_block_group_cache *block_group)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = block_group->fs_info;
|
||||
|
||||
return (btrfs_test_opt(fs_info, FRAGMENT_METADATA) &&
|
||||
block_group->flags & BTRFS_BLOCK_GROUP_METADATA) ||
|
||||
(btrfs_test_opt(fs_info, FRAGMENT_DATA) &&
|
||||
block_group->flags & BTRFS_BLOCK_GROUP_DATA);
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Requests for changes that need to be done during transaction commit.
|
||||
*
|
||||
|
@ -1475,8 +1299,10 @@ struct btrfs_map_token {
|
|||
#define BTRFS_BYTES_TO_BLKS(fs_info, bytes) \
|
||||
((bytes) >> (fs_info)->sb->s_blocksize_bits)
|
||||
|
||||
static inline void btrfs_init_map_token (struct btrfs_map_token *token)
|
||||
static inline void btrfs_init_map_token(struct btrfs_map_token *token,
|
||||
struct extent_buffer *eb)
|
||||
{
|
||||
token->eb = eb;
|
||||
token->kaddr = NULL;
|
||||
}
|
||||
|
||||
|
@ -1507,17 +1333,10 @@ u##bits btrfs_get_token_##bits(const struct extent_buffer *eb, \
|
|||
void btrfs_set_token_##bits(struct extent_buffer *eb, const void *ptr, \
|
||||
unsigned long off, u##bits val, \
|
||||
struct btrfs_map_token *token); \
|
||||
static inline u##bits btrfs_get_##bits(const struct extent_buffer *eb, \
|
||||
const void *ptr, \
|
||||
unsigned long off) \
|
||||
{ \
|
||||
return btrfs_get_token_##bits(eb, ptr, off, NULL); \
|
||||
} \
|
||||
static inline void btrfs_set_##bits(struct extent_buffer *eb, void *ptr,\
|
||||
unsigned long off, u##bits val) \
|
||||
{ \
|
||||
btrfs_set_token_##bits(eb, ptr, off, val, NULL); \
|
||||
}
|
||||
u##bits btrfs_get_##bits(const struct extent_buffer *eb, \
|
||||
const void *ptr, unsigned long off); \
|
||||
void btrfs_set_##bits(struct extent_buffer *eb, void *ptr, \
|
||||
unsigned long off, u##bits val);
|
||||
|
||||
DECLARE_BTRFS_SETGET_BITS(8)
|
||||
DECLARE_BTRFS_SETGET_BITS(16)
|
||||
|
@ -2059,16 +1878,6 @@ static inline void btrfs_dir_item_key_to_cpu(const struct extent_buffer *eb,
|
|||
btrfs_disk_key_to_cpu(key, &disk_key);
|
||||
}
|
||||
|
||||
static inline u8 btrfs_key_type(const struct btrfs_key *key)
|
||||
{
|
||||
return key->type;
|
||||
}
|
||||
|
||||
static inline void btrfs_set_key_type(struct btrfs_key *key, u8 val)
|
||||
{
|
||||
key->type = val;
|
||||
}
|
||||
|
||||
/* struct btrfs_header */
|
||||
BTRFS_SETGET_HEADER_FUNCS(header_bytenr, struct btrfs_header, bytenr, 64);
|
||||
BTRFS_SETGET_HEADER_FUNCS(header_generation, struct btrfs_header,
|
||||
|
@ -2354,20 +2163,8 @@ BTRFS_SETGET_STACK_FUNCS(super_magic, struct btrfs_super_block, magic, 64);
|
|||
BTRFS_SETGET_STACK_FUNCS(super_uuid_tree_generation, struct btrfs_super_block,
|
||||
uuid_tree_generation, 64);
|
||||
|
||||
static inline int btrfs_super_csum_size(const struct btrfs_super_block *s)
|
||||
{
|
||||
u16 t = btrfs_super_csum_type(s);
|
||||
/*
|
||||
* csum type is validated at mount time
|
||||
*/
|
||||
return btrfs_csum_sizes[t];
|
||||
}
|
||||
|
||||
static inline const char *btrfs_super_csum_name(u16 csum_type)
|
||||
{
|
||||
/* csum type is validated at mount time */
|
||||
return btrfs_csum_names[csum_type];
|
||||
}
|
||||
int btrfs_super_csum_size(const struct btrfs_super_block *s);
|
||||
const char *btrfs_super_csum_name(u16 csum_type);
|
||||
|
||||
/*
|
||||
* The leaf data grows from end-to-front in the node.
|
||||
|
@ -2440,30 +2237,6 @@ static inline u32 btrfs_file_extent_inline_item_len(
|
|||
return btrfs_item_size(eb, e) - BTRFS_FILE_EXTENT_INLINE_DATA_START;
|
||||
}
|
||||
|
||||
/* btrfs_dev_stats_item */
|
||||
static inline u64 btrfs_dev_stats_value(const struct extent_buffer *eb,
|
||||
const struct btrfs_dev_stats_item *ptr,
|
||||
int index)
|
||||
{
|
||||
u64 val;
|
||||
|
||||
read_extent_buffer(eb, &val,
|
||||
offsetof(struct btrfs_dev_stats_item, values) +
|
||||
((unsigned long)ptr) + (index * sizeof(u64)),
|
||||
sizeof(val));
|
||||
return val;
|
||||
}
|
||||
|
||||
static inline void btrfs_set_dev_stats_value(struct extent_buffer *eb,
|
||||
struct btrfs_dev_stats_item *ptr,
|
||||
int index, u64 val)
|
||||
{
|
||||
write_extent_buffer(eb, &val,
|
||||
offsetof(struct btrfs_dev_stats_item, values) +
|
||||
((unsigned long)ptr) + (index * sizeof(u64)),
|
||||
sizeof(val));
|
||||
}
|
||||
|
||||
/* btrfs_qgroup_status_item */
|
||||
BTRFS_SETGET_FUNCS(qgroup_status_generation, struct btrfs_qgroup_status_item,
|
||||
generation, 64);
|
||||
|
@ -2600,32 +2373,33 @@ enum btrfs_inline_ref_type {
|
|||
int btrfs_get_extent_inline_ref_type(const struct extent_buffer *eb,
|
||||
struct btrfs_extent_inline_ref *iref,
|
||||
enum btrfs_inline_ref_type is_data);
|
||||
u64 hash_extent_data_ref(u64 root_objectid, u64 owner, u64 offset);
|
||||
|
||||
u64 btrfs_csum_bytes_to_leaves(struct btrfs_fs_info *fs_info, u64 csum_bytes);
|
||||
|
||||
static inline u64 btrfs_calc_trans_metadata_size(struct btrfs_fs_info *fs_info,
|
||||
unsigned num_items)
|
||||
/*
|
||||
* Use this if we would be adding new items, as we could split nodes as we cow
|
||||
* down the tree.
|
||||
*/
|
||||
static inline u64 btrfs_calc_insert_metadata_size(struct btrfs_fs_info *fs_info,
|
||||
unsigned num_items)
|
||||
{
|
||||
return (u64)fs_info->nodesize * BTRFS_MAX_LEVEL * 2 * num_items;
|
||||
}
|
||||
|
||||
/*
|
||||
* Doing a truncate won't result in new nodes or leaves, just what we need for
|
||||
* COW.
|
||||
* Doing a truncate or a modification won't result in new nodes or leaves, just
|
||||
* what we need for COW.
|
||||
*/
|
||||
static inline u64 btrfs_calc_trunc_metadata_size(struct btrfs_fs_info *fs_info,
|
||||
static inline u64 btrfs_calc_metadata_size(struct btrfs_fs_info *fs_info,
|
||||
unsigned num_items)
|
||||
{
|
||||
return (u64)fs_info->nodesize * BTRFS_MAX_LEVEL * num_items;
|
||||
}
|
||||
|
||||
void btrfs_dec_block_group_reservations(struct btrfs_fs_info *fs_info,
|
||||
const u64 start);
|
||||
void btrfs_wait_block_group_reservations(struct btrfs_block_group_cache *bg);
|
||||
bool btrfs_inc_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
void btrfs_dec_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
void btrfs_wait_nocow_writers(struct btrfs_block_group_cache *bg);
|
||||
void btrfs_put_block_group(struct btrfs_block_group_cache *cache);
|
||||
int btrfs_add_excluded_extent(struct btrfs_fs_info *fs_info,
|
||||
u64 start, u64 num_bytes);
|
||||
void btrfs_free_excluded_extents(struct btrfs_block_group_cache *cache);
|
||||
int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans,
|
||||
unsigned long count);
|
||||
void btrfs_cleanup_ref_head_accounting(struct btrfs_fs_info *fs_info,
|
||||
|
@ -2642,11 +2416,6 @@ int btrfs_pin_extent_for_log_replay(struct btrfs_fs_info *fs_info,
|
|||
int btrfs_exclude_logged_extents(struct extent_buffer *eb);
|
||||
int btrfs_cross_ref_exist(struct btrfs_root *root,
|
||||
u64 objectid, u64 offset, u64 bytenr);
|
||||
struct btrfs_block_group_cache *btrfs_lookup_block_group(
|
||||
struct btrfs_fs_info *info,
|
||||
u64 bytenr);
|
||||
void btrfs_get_block_group(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_put_block_group(struct btrfs_block_group_cache *cache);
|
||||
struct extent_buffer *btrfs_alloc_tree_block(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root,
|
||||
u64 parent, u64 root_objectid,
|
||||
|
@ -2685,28 +2454,9 @@ int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans);
|
|||
int btrfs_inc_extent_ref(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_ref *generic_ref);
|
||||
|
||||
int btrfs_start_dirty_block_groups(struct btrfs_trans_handle *trans);
|
||||
int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans);
|
||||
int btrfs_setup_space_cache(struct btrfs_trans_handle *trans);
|
||||
int btrfs_extent_readonly(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
int btrfs_free_block_groups(struct btrfs_fs_info *info);
|
||||
int btrfs_read_block_groups(struct btrfs_fs_info *info);
|
||||
int btrfs_can_relocate(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
int btrfs_make_block_group(struct btrfs_trans_handle *trans,
|
||||
u64 bytes_used, u64 type, u64 chunk_offset,
|
||||
u64 size);
|
||||
struct btrfs_trans_handle *btrfs_start_trans_remove_block_group(
|
||||
struct btrfs_fs_info *fs_info,
|
||||
const u64 chunk_offset);
|
||||
int btrfs_remove_block_group(struct btrfs_trans_handle *trans,
|
||||
u64 group_start, struct extent_map *em);
|
||||
void btrfs_delete_unused_bgs(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_get_block_group_trimming(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_put_block_group_trimming(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_create_pending_block_groups(struct btrfs_trans_handle *trans);
|
||||
u64 btrfs_data_alloc_profile(struct btrfs_fs_info *fs_info);
|
||||
u64 btrfs_metadata_alloc_profile(struct btrfs_fs_info *fs_info);
|
||||
u64 btrfs_system_alloc_profile(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_clear_space_info_full(struct btrfs_fs_info *info);
|
||||
|
||||
enum btrfs_reserve_flush_enum {
|
||||
|
@ -2717,6 +2467,7 @@ enum btrfs_reserve_flush_enum {
|
|||
* case, use FLUSH LIMIT
|
||||
*/
|
||||
BTRFS_RESERVE_FLUSH_LIMIT,
|
||||
BTRFS_RESERVE_FLUSH_EVICT,
|
||||
BTRFS_RESERVE_FLUSH_ALL,
|
||||
};
|
||||
|
||||
|
@ -2729,31 +2480,10 @@ enum btrfs_flush_state {
|
|||
FLUSH_DELALLOC_WAIT = 6,
|
||||
ALLOC_CHUNK = 7,
|
||||
ALLOC_CHUNK_FORCE = 8,
|
||||
COMMIT_TRANS = 9,
|
||||
RUN_DELAYED_IPUTS = 9,
|
||||
COMMIT_TRANS = 10,
|
||||
};
|
||||
|
||||
/*
|
||||
* control flags for do_chunk_alloc's force field
|
||||
* CHUNK_ALLOC_NO_FORCE means to only allocate a chunk
|
||||
* if we really need one.
|
||||
*
|
||||
* CHUNK_ALLOC_LIMITED means to only try and allocate one
|
||||
* if we have very few chunks already allocated. This is
|
||||
* used as part of the clustering code to help make sure
|
||||
* we have a good pool of storage to cluster in, without
|
||||
* filling the FS with empty chunks
|
||||
*
|
||||
* CHUNK_ALLOC_FORCE means it must try to allocate one
|
||||
*
|
||||
*/
|
||||
enum btrfs_chunk_alloc_enum {
|
||||
CHUNK_ALLOC_NO_FORCE,
|
||||
CHUNK_ALLOC_LIMITED,
|
||||
CHUNK_ALLOC_FORCE,
|
||||
};
|
||||
|
||||
int btrfs_chunk_alloc(struct btrfs_trans_handle *trans, u64 flags,
|
||||
enum btrfs_chunk_alloc_enum force);
|
||||
int btrfs_subvolume_reserve_metadata(struct btrfs_root *root,
|
||||
struct btrfs_block_rsv *rsv,
|
||||
int nitems, bool use_global_rsv);
|
||||
|
@ -2763,15 +2493,11 @@ void btrfs_delalloc_release_extents(struct btrfs_inode *inode, u64 num_bytes,
|
|||
bool qgroup_free);
|
||||
|
||||
int btrfs_delalloc_reserve_metadata(struct btrfs_inode *inode, u64 num_bytes);
|
||||
int btrfs_inc_block_group_ro(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_dec_block_group_ro(struct btrfs_block_group_cache *cache);
|
||||
void btrfs_put_block_group_cache(struct btrfs_fs_info *info);
|
||||
u64 btrfs_account_ro_block_groups_free_space(struct btrfs_space_info *sinfo);
|
||||
int btrfs_error_unpin_extent_range(struct btrfs_fs_info *fs_info,
|
||||
u64 start, u64 end);
|
||||
int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
u64 num_bytes, u64 *actual_bytes);
|
||||
int btrfs_force_chunk_alloc(struct btrfs_trans_handle *trans, u64 type);
|
||||
int btrfs_trim_fs(struct btrfs_fs_info *fs_info, struct fstrim_range *range);
|
||||
|
||||
int btrfs_init_space_info(struct btrfs_fs_info *fs_info);
|
||||
|
@ -2780,10 +2506,6 @@ int btrfs_delayed_refs_qgroup_accounting(struct btrfs_trans_handle *trans,
|
|||
int btrfs_start_write_no_snapshotting(struct btrfs_root *root);
|
||||
void btrfs_end_write_no_snapshotting(struct btrfs_root *root);
|
||||
void btrfs_wait_for_snapshot_creation(struct btrfs_root *root);
|
||||
void check_system_chunk(struct btrfs_trans_handle *trans, const u64 type);
|
||||
u64 add_new_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 start, u64 end);
|
||||
void btrfs_mark_bg_unused(struct btrfs_block_group_cache *bg);
|
||||
|
||||
/* ctree.c */
|
||||
int btrfs_bin_search(struct extent_buffer *eb, const struct btrfs_key *key,
|
||||
|
@ -2806,20 +2528,9 @@ int btrfs_find_next_key(struct btrfs_root *root, struct btrfs_path *path,
|
|||
int btrfs_search_forward(struct btrfs_root *root, struct btrfs_key *min_key,
|
||||
struct btrfs_path *path,
|
||||
u64 min_trans);
|
||||
enum btrfs_compare_tree_result {
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
BTRFS_COMPARE_TREE_CHANGED,
|
||||
BTRFS_COMPARE_TREE_SAME,
|
||||
};
|
||||
typedef int (*btrfs_changed_cb_t)(struct btrfs_path *left_path,
|
||||
struct btrfs_path *right_path,
|
||||
struct btrfs_key *key,
|
||||
enum btrfs_compare_tree_result result,
|
||||
void *ctx);
|
||||
int btrfs_compare_trees(struct btrfs_root *left_root,
|
||||
struct btrfs_root *right_root,
|
||||
btrfs_changed_cb_t cb, void *ctx);
|
||||
struct extent_buffer *btrfs_read_node_slot(struct extent_buffer *parent,
|
||||
int slot);
|
||||
|
||||
int btrfs_cow_block(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root, struct extent_buffer *buf,
|
||||
struct extent_buffer *parent, int parent_slot,
|
||||
|
@ -3068,14 +2779,12 @@ btrfs_lookup_inode_extref(struct btrfs_trans_handle *trans,
|
|||
u64 inode_objectid, u64 ref_objectid, int ins_len,
|
||||
int cow);
|
||||
|
||||
int btrfs_find_name_in_backref(struct extent_buffer *leaf, int slot,
|
||||
const char *name,
|
||||
int name_len, struct btrfs_inode_ref **ref_ret);
|
||||
int btrfs_find_name_in_ext_backref(struct extent_buffer *leaf, int slot,
|
||||
u64 ref_objectid, const char *name,
|
||||
int name_len,
|
||||
struct btrfs_inode_extref **extref_ret);
|
||||
|
||||
struct btrfs_inode_ref *btrfs_find_name_in_backref(struct extent_buffer *leaf,
|
||||
int slot, const char *name,
|
||||
int name_len);
|
||||
struct btrfs_inode_extref *btrfs_find_name_in_ext_backref(
|
||||
struct extent_buffer *leaf, int slot, u64 ref_objectid,
|
||||
const char *name, int name_len);
|
||||
/* file-item.c */
|
||||
struct btrfs_dio_private;
|
||||
int btrfs_del_csums(struct btrfs_trans_handle *trans,
|
||||
|
@ -3137,7 +2846,7 @@ int btrfs_start_delalloc_snapshot(struct btrfs_root *root);
|
|||
int btrfs_start_delalloc_roots(struct btrfs_fs_info *fs_info, int nr);
|
||||
int btrfs_set_extent_delalloc(struct inode *inode, u64 start, u64 end,
|
||||
unsigned int extra_bits,
|
||||
struct extent_state **cached_state, int dedupe);
|
||||
struct extent_state **cached_state);
|
||||
int btrfs_create_subvol_root(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *new_root,
|
||||
struct btrfs_root *parent_root,
|
||||
|
@ -3233,6 +2942,10 @@ int __btrfs_drop_extents(struct btrfs_trans_handle *trans,
|
|||
int btrfs_drop_extents(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root, struct inode *inode, u64 start,
|
||||
u64 end, int drop_cache);
|
||||
int btrfs_punch_hole_range(struct inode *inode, struct btrfs_path *path,
|
||||
const u64 start, const u64 end,
|
||||
struct btrfs_clone_extent_info *clone_info,
|
||||
struct btrfs_trans_handle **trans_out);
|
||||
int btrfs_mark_extent_written(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_inode *inode, u64 start, u64 end);
|
||||
int btrfs_release_file(struct inode *inode, struct file *file);
|
||||
|
@ -3248,12 +2961,6 @@ loff_t btrfs_remap_file_range(struct file *file_in, loff_t pos_in,
|
|||
int btrfs_defrag_leaves(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root);
|
||||
|
||||
/* sysfs.c */
|
||||
int __init btrfs_init_sysfs(void);
|
||||
void __cold btrfs_exit_sysfs(void);
|
||||
int btrfs_sysfs_add_mounted(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_sysfs_remove_mounted(struct btrfs_fs_info *fs_info);
|
||||
|
||||
/* super.c */
|
||||
int btrfs_parse_options(struct btrfs_fs_info *info, char *options,
|
||||
unsigned long new_flags);
|
||||
|
@ -3722,26 +3429,4 @@ static inline int btrfs_is_testing(struct btrfs_fs_info *fs_info)
|
|||
}
|
||||
#endif
|
||||
|
||||
static inline void cond_wake_up(struct wait_queue_head *wq)
|
||||
{
|
||||
/*
|
||||
* This implies a full smp_mb barrier, see comments for
|
||||
* waitqueue_active why.
|
||||
*/
|
||||
if (wq_has_sleeper(wq))
|
||||
wake_up(wq);
|
||||
}
|
||||
|
||||
static inline void cond_wake_up_nomb(struct wait_queue_head *wq)
|
||||
{
|
||||
/*
|
||||
* Special case for conditional wakeup where the barrier required for
|
||||
* waitqueue_active is implied by some of the preceding code. Eg. one
|
||||
* of such atomic operations (atomic_dec_and_return, ...), or a
|
||||
* unlock/lock sequence, etc.
|
||||
*/
|
||||
if (waitqueue_active(wq))
|
||||
wake_up(wq);
|
||||
}
|
||||
|
||||
#endif
|
||||
|
|
|
@ -1,12 +0,0 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
/*
|
||||
* Copyright (C) 2016 Fujitsu. All rights reserved.
|
||||
*/
|
||||
|
||||
#ifndef BTRFS_DEDUPE_H
|
||||
#define BTRFS_DEDUPE_H
|
||||
|
||||
/* later in-band dedupe will expand this struct */
|
||||
struct btrfs_dedupe_hash;
|
||||
|
||||
#endif
|
|
@ -7,6 +7,7 @@
|
|||
#include "space-info.h"
|
||||
#include "transaction.h"
|
||||
#include "qgroup.h"
|
||||
#include "block-group.h"
|
||||
|
||||
int btrfs_alloc_data_chunk_ondemand(struct btrfs_inode *inode, u64 bytes)
|
||||
{
|
||||
|
@ -129,8 +130,6 @@ commit_trans:
|
|||
return -ENOSPC;
|
||||
}
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, data_sinfo, bytes);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
data_sinfo->flags, bytes, 1);
|
||||
spin_unlock(&data_sinfo->lock);
|
||||
|
||||
return 0;
|
||||
|
@ -182,8 +181,6 @@ void btrfs_free_reserved_data_space_noquota(struct inode *inode, u64 start,
|
|||
data_sinfo = fs_info->data_sinfo;
|
||||
spin_lock(&data_sinfo->lock);
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, data_sinfo, -len);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
data_sinfo->flags, len, 0);
|
||||
spin_unlock(&data_sinfo->lock);
|
||||
}
|
||||
|
||||
|
@ -254,13 +251,20 @@ static void btrfs_calculate_inode_block_rsv_size(struct btrfs_fs_info *fs_info,
|
|||
|
||||
lockdep_assert_held(&inode->lock);
|
||||
outstanding_extents = inode->outstanding_extents;
|
||||
if (outstanding_extents)
|
||||
reserve_size = btrfs_calc_trans_metadata_size(fs_info,
|
||||
outstanding_extents + 1);
|
||||
|
||||
/*
|
||||
* Insert size for the number of outstanding extents, 1 normal size for
|
||||
* updating the inode.
|
||||
*/
|
||||
if (outstanding_extents) {
|
||||
reserve_size = btrfs_calc_insert_metadata_size(fs_info,
|
||||
outstanding_extents);
|
||||
reserve_size += btrfs_calc_metadata_size(fs_info, 1);
|
||||
}
|
||||
csum_leaves = btrfs_csum_bytes_to_leaves(fs_info,
|
||||
inode->csum_bytes);
|
||||
reserve_size += btrfs_calc_trans_metadata_size(fs_info,
|
||||
csum_leaves);
|
||||
reserve_size += btrfs_calc_insert_metadata_size(fs_info,
|
||||
csum_leaves);
|
||||
/*
|
||||
* For qgroup rsv, the calculation is very simple:
|
||||
* account one nodesize for each outstanding extent
|
||||
|
@ -281,10 +285,16 @@ static void calc_inode_reservations(struct btrfs_fs_info *fs_info,
|
|||
{
|
||||
u64 nr_extents = count_max_extents(num_bytes);
|
||||
u64 csum_leaves = btrfs_csum_bytes_to_leaves(fs_info, num_bytes);
|
||||
u64 inode_update = btrfs_calc_metadata_size(fs_info, 1);
|
||||
|
||||
/* We add one for the inode update at finish ordered time */
|
||||
*meta_reserve = btrfs_calc_trans_metadata_size(fs_info,
|
||||
nr_extents + csum_leaves + 1);
|
||||
*meta_reserve = btrfs_calc_insert_metadata_size(fs_info,
|
||||
nr_extents + csum_leaves);
|
||||
|
||||
/*
|
||||
* finish_ordered_io has to update the inode, so add the space required
|
||||
* for an inode update.
|
||||
*/
|
||||
*meta_reserve += inode_update;
|
||||
*qgroup_reserve = nr_extents * fs_info->nodesize;
|
||||
}
|
||||
|
||||
|
|
|
@ -6,6 +6,7 @@
|
|||
|
||||
#include <linux/slab.h>
|
||||
#include <linux/iversion.h>
|
||||
#include "misc.h"
|
||||
#include "delayed-inode.h"
|
||||
#include "disk-io.h"
|
||||
#include "transaction.h"
|
||||
|
@ -474,6 +475,9 @@ static void __btrfs_remove_delayed_item(struct btrfs_delayed_item *delayed_item)
|
|||
struct rb_root_cached *root;
|
||||
struct btrfs_delayed_root *delayed_root;
|
||||
|
||||
/* Not associated with any delayed_node */
|
||||
if (!delayed_item->delayed_node)
|
||||
return;
|
||||
delayed_root = delayed_item->delayed_node->root->fs_info->delayed_root;
|
||||
|
||||
BUG_ON(!delayed_root);
|
||||
|
@ -555,7 +559,7 @@ static int btrfs_delayed_item_reserve_metadata(struct btrfs_trans_handle *trans,
|
|||
src_rsv = trans->block_rsv;
|
||||
dst_rsv = &fs_info->delayed_block_rsv;
|
||||
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
|
||||
/*
|
||||
* Here we migrate space rsv from transaction rsv, since have already
|
||||
|
@ -609,7 +613,7 @@ static int btrfs_delayed_inode_reserve_metadata(
|
|||
src_rsv = trans->block_rsv;
|
||||
dst_rsv = &fs_info->delayed_block_rsv;
|
||||
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
num_bytes = btrfs_calc_metadata_size(fs_info, 1);
|
||||
|
||||
/*
|
||||
* btrfs_dirty_inode will update the inode under btrfs_join_transaction
|
||||
|
@ -1525,7 +1529,12 @@ int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans,
|
|||
* we have reserved enough space when we start a new transaction,
|
||||
* so reserving metadata failure is impossible.
|
||||
*/
|
||||
BUG_ON(ret);
|
||||
if (ret < 0) {
|
||||
btrfs_err(trans->fs_info,
|
||||
"metadata reservation failed for delayed dir item deltiona, should have been reserved");
|
||||
btrfs_release_delayed_item(item);
|
||||
goto end;
|
||||
}
|
||||
|
||||
mutex_lock(&node->mutex);
|
||||
ret = __btrfs_add_delayed_deletion_item(node, item);
|
||||
|
@ -1534,7 +1543,8 @@ int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans,
|
|||
"err add delayed dir index item(index: %llu) into the deletion tree of the delayed node(root id: %llu, inode id: %llu, errno: %d)",
|
||||
index, node->root->root_key.objectid,
|
||||
node->inode_id, ret);
|
||||
BUG();
|
||||
btrfs_delayed_item_release_metadata(dir->root, item);
|
||||
btrfs_release_delayed_item(item);
|
||||
}
|
||||
mutex_unlock(&node->mutex);
|
||||
end:
|
||||
|
|
|
@ -79,7 +79,7 @@ int btrfs_should_throttle_delayed_refs(struct btrfs_trans_handle *trans)
|
|||
void btrfs_delayed_refs_rsv_release(struct btrfs_fs_info *fs_info, int nr)
|
||||
{
|
||||
struct btrfs_block_rsv *block_rsv = &fs_info->delayed_refs_rsv;
|
||||
u64 num_bytes = btrfs_calc_trans_metadata_size(fs_info, nr);
|
||||
u64 num_bytes = btrfs_calc_insert_metadata_size(fs_info, nr);
|
||||
u64 released = 0;
|
||||
|
||||
released = __btrfs_block_rsv_release(fs_info, block_rsv, num_bytes,
|
||||
|
@ -105,8 +105,8 @@ void btrfs_update_delayed_refs_rsv(struct btrfs_trans_handle *trans)
|
|||
if (!trans->delayed_ref_updates)
|
||||
return;
|
||||
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info,
|
||||
trans->delayed_ref_updates);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info,
|
||||
trans->delayed_ref_updates);
|
||||
spin_lock(&delayed_rsv->lock);
|
||||
delayed_rsv->size += num_bytes;
|
||||
delayed_rsv->full = 0;
|
||||
|
@ -158,7 +158,7 @@ void btrfs_migrate_to_delayed_refs_rsv(struct btrfs_fs_info *fs_info,
|
|||
trace_btrfs_space_reservation(fs_info, "delayed_refs_rsv",
|
||||
0, num_bytes, 1);
|
||||
if (to_free)
|
||||
btrfs_space_info_add_old_bytes(fs_info,
|
||||
btrfs_space_info_free_bytes_may_use(fs_info,
|
||||
delayed_refs_rsv->space_info, to_free);
|
||||
}
|
||||
|
||||
|
@ -174,7 +174,7 @@ int btrfs_delayed_refs_rsv_refill(struct btrfs_fs_info *fs_info,
|
|||
enum btrfs_reserve_flush_enum flush)
|
||||
{
|
||||
struct btrfs_block_rsv *block_rsv = &fs_info->delayed_refs_rsv;
|
||||
u64 limit = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
u64 limit = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
u64 num_bytes = 0;
|
||||
int ret = -ENOSPC;
|
||||
|
||||
|
|
|
@ -9,6 +9,7 @@
|
|||
#include <linux/blkdev.h>
|
||||
#include <linux/kthread.h>
|
||||
#include <linux/math64.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "extent_map.h"
|
||||
#include "disk-io.h"
|
||||
|
@ -56,7 +57,7 @@ int btrfs_init_dev_replace(struct btrfs_fs_info *fs_info)
|
|||
no_valid_dev_replace_entry_found:
|
||||
ret = 0;
|
||||
dev_replace->replace_state =
|
||||
BTRFS_DEV_REPLACE_ITEM_STATE_NEVER_STARTED;
|
||||
BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED;
|
||||
dev_replace->cont_reading_from_srcdev_mode =
|
||||
BTRFS_DEV_REPLACE_ITEM_CONT_READING_FROM_SRCDEV_MODE_ALWAYS;
|
||||
dev_replace->time_started = 0;
|
||||
|
|
|
@ -40,6 +40,7 @@
|
|||
#include "compression.h"
|
||||
#include "tree-checker.h"
|
||||
#include "ref-verify.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#define BTRFS_SUPER_FLAG_SUPP (BTRFS_HEADER_FLAG_WRITTEN |\
|
||||
BTRFS_HEADER_FLAG_RELOC |\
|
||||
|
@ -416,6 +417,16 @@ int btrfs_verify_level_key(struct extent_buffer *eb, int level,
|
|||
*/
|
||||
if (btrfs_header_generation(eb) > fs_info->last_trans_committed)
|
||||
return 0;
|
||||
|
||||
/* We have @first_key, so this @eb must have at least one item */
|
||||
if (btrfs_header_nritems(eb) == 0) {
|
||||
btrfs_err(fs_info,
|
||||
"invalid tree nritems, bytenr=%llu nritems=0 expect >0",
|
||||
eb->start);
|
||||
WARN_ON(IS_ENABLED(CONFIG_BTRFS_DEBUG));
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
if (found_level)
|
||||
btrfs_node_key_to_cpu(eb, &found_key, 0);
|
||||
else
|
||||
|
@ -1037,35 +1048,6 @@ void readahead_tree_block(struct btrfs_fs_info *fs_info, u64 bytenr)
|
|||
free_extent_buffer(buf);
|
||||
}
|
||||
|
||||
int reada_tree_block_flagged(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
int mirror_num, struct extent_buffer **eb)
|
||||
{
|
||||
struct extent_buffer *buf = NULL;
|
||||
int ret;
|
||||
|
||||
buf = btrfs_find_create_tree_block(fs_info, bytenr);
|
||||
if (IS_ERR(buf))
|
||||
return 0;
|
||||
|
||||
set_bit(EXTENT_BUFFER_READAHEAD, &buf->bflags);
|
||||
|
||||
ret = read_extent_buffer_pages(buf, WAIT_PAGE_LOCK, mirror_num);
|
||||
if (ret) {
|
||||
free_extent_buffer_stale(buf);
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (test_bit(EXTENT_BUFFER_CORRUPT, &buf->bflags)) {
|
||||
free_extent_buffer_stale(buf);
|
||||
return -EIO;
|
||||
} else if (extent_buffer_uptodate(buf)) {
|
||||
*eb = buf;
|
||||
} else {
|
||||
free_extent_buffer(buf);
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
struct extent_buffer *btrfs_find_create_tree_block(
|
||||
struct btrfs_fs_info *fs_info,
|
||||
u64 bytenr)
|
||||
|
|
|
@ -45,8 +45,6 @@ struct extent_buffer *read_tree_block(struct btrfs_fs_info *fs_info, u64 bytenr,
|
|||
u64 parent_transid, int level,
|
||||
struct btrfs_key *first_key);
|
||||
void readahead_tree_block(struct btrfs_fs_info *fs_info, u64 bytenr);
|
||||
int reada_tree_block_flagged(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
int mirror_num, struct extent_buffer **eb);
|
||||
struct extent_buffer *btrfs_find_create_tree_block(
|
||||
struct btrfs_fs_info *fs_info,
|
||||
u64 bytenr);
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -1938,9 +1938,9 @@ out:
|
|||
}
|
||||
|
||||
void extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end,
|
||||
u64 delalloc_end, struct page *locked_page,
|
||||
unsigned clear_bits,
|
||||
unsigned long page_ops)
|
||||
struct page *locked_page,
|
||||
unsigned clear_bits,
|
||||
unsigned long page_ops)
|
||||
{
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, start, end, clear_bits, 1, 0,
|
||||
NULL);
|
||||
|
@ -4339,10 +4339,8 @@ int extent_invalidatepage(struct extent_io_tree *tree,
|
|||
|
||||
lock_extent_bits(tree, start, end, &cached_state);
|
||||
wait_on_page_writeback(page);
|
||||
clear_extent_bit(tree, start, end,
|
||||
EXTENT_LOCKED | EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING,
|
||||
1, 1, &cached_state);
|
||||
clear_extent_bit(tree, start, end, EXTENT_LOCKED | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING, 1, 1, &cached_state);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -494,9 +494,9 @@ int map_private_extent_buffer(const struct extent_buffer *eb,
|
|||
void extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end);
|
||||
void extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end);
|
||||
void extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end,
|
||||
u64 delalloc_end, struct page *locked_page,
|
||||
unsigned bits_to_clear,
|
||||
unsigned long page_ops);
|
||||
struct page *locked_page,
|
||||
unsigned bits_to_clear,
|
||||
unsigned long page_ops);
|
||||
struct bio *btrfs_bio_alloc(u64 first_byte);
|
||||
struct bio *btrfs_io_bio_alloc(unsigned int nr_iovecs);
|
||||
struct bio *btrfs_bio_clone(struct bio *bio);
|
||||
|
|
|
@ -384,6 +384,8 @@ int add_extent_mapping(struct extent_map_tree *tree,
|
|||
{
|
||||
int ret = 0;
|
||||
|
||||
lockdep_assert_held_write(&tree->lock);
|
||||
|
||||
ret = tree_insert(&tree->map, em);
|
||||
if (ret)
|
||||
goto out;
|
||||
|
|
435
fs/btrfs/file.c
435
fs/btrfs/file.c
|
@ -537,8 +537,8 @@ int btrfs_dirty_pages(struct inode *inode, struct page **pages,
|
|||
* we can set things up properly
|
||||
*/
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, start_pos, end_of_last_block,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG, 0, 0, cached);
|
||||
EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG,
|
||||
0, 0, cached);
|
||||
|
||||
if (!btrfs_is_free_space_inode(BTRFS_I(inode))) {
|
||||
if (start_pos >= isize &&
|
||||
|
@ -559,7 +559,7 @@ int btrfs_dirty_pages(struct inode *inode, struct page **pages,
|
|||
}
|
||||
|
||||
err = btrfs_set_extent_delalloc(inode, start_pos, end_of_last_block,
|
||||
extra_bits, cached, 0);
|
||||
extra_bits, cached);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
|
@ -1882,10 +1882,10 @@ static ssize_t btrfs_file_write_iter(struct kiocb *iocb,
|
|||
u64 start_pos;
|
||||
u64 end_pos;
|
||||
ssize_t num_written = 0;
|
||||
bool sync = (file->f_flags & O_DSYNC) || IS_SYNC(file->f_mapping->host);
|
||||
const bool sync = iocb->ki_flags & IOCB_DSYNC;
|
||||
ssize_t err;
|
||||
loff_t pos;
|
||||
size_t count = iov_iter_count(from);
|
||||
size_t count;
|
||||
loff_t oldsize;
|
||||
int clean_page = 0;
|
||||
|
||||
|
@ -1906,6 +1906,7 @@ static ssize_t btrfs_file_write_iter(struct kiocb *iocb,
|
|||
}
|
||||
|
||||
pos = iocb->ki_pos;
|
||||
count = iov_iter_count(from);
|
||||
if (iocb->ki_flags & IOCB_NOWAIT) {
|
||||
/*
|
||||
* We will allocate space in case nodatacow is not set,
|
||||
|
@ -2439,27 +2440,286 @@ static int btrfs_punch_hole_lock_range(struct inode *inode,
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int btrfs_insert_clone_extent(struct btrfs_trans_handle *trans,
|
||||
struct inode *inode,
|
||||
struct btrfs_path *path,
|
||||
struct btrfs_clone_extent_info *clone_info,
|
||||
const u64 clone_len)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
struct btrfs_file_extent_item *extent;
|
||||
struct extent_buffer *leaf;
|
||||
struct btrfs_key key;
|
||||
int slot;
|
||||
struct btrfs_ref ref = { 0 };
|
||||
u64 ref_offset;
|
||||
int ret;
|
||||
|
||||
if (clone_len == 0)
|
||||
return 0;
|
||||
|
||||
if (clone_info->disk_offset == 0 &&
|
||||
btrfs_fs_incompat(fs_info, NO_HOLES))
|
||||
return 0;
|
||||
|
||||
key.objectid = btrfs_ino(BTRFS_I(inode));
|
||||
key.type = BTRFS_EXTENT_DATA_KEY;
|
||||
key.offset = clone_info->file_offset;
|
||||
ret = btrfs_insert_empty_item(trans, root, path, &key,
|
||||
clone_info->item_size);
|
||||
if (ret)
|
||||
return ret;
|
||||
leaf = path->nodes[0];
|
||||
slot = path->slots[0];
|
||||
write_extent_buffer(leaf, clone_info->extent_buf,
|
||||
btrfs_item_ptr_offset(leaf, slot),
|
||||
clone_info->item_size);
|
||||
extent = btrfs_item_ptr(leaf, slot, struct btrfs_file_extent_item);
|
||||
btrfs_set_file_extent_offset(leaf, extent, clone_info->data_offset);
|
||||
btrfs_set_file_extent_num_bytes(leaf, extent, clone_len);
|
||||
btrfs_mark_buffer_dirty(leaf);
|
||||
btrfs_release_path(path);
|
||||
|
||||
/* If it's a hole, nothing more needs to be done. */
|
||||
if (clone_info->disk_offset == 0)
|
||||
return 0;
|
||||
|
||||
inode_add_bytes(inode, clone_len);
|
||||
btrfs_init_generic_ref(&ref, BTRFS_ADD_DELAYED_REF,
|
||||
clone_info->disk_offset,
|
||||
clone_info->disk_len, 0);
|
||||
ref_offset = clone_info->file_offset - clone_info->data_offset;
|
||||
btrfs_init_data_ref(&ref, root->root_key.objectid,
|
||||
btrfs_ino(BTRFS_I(inode)), ref_offset);
|
||||
ret = btrfs_inc_extent_ref(trans, &ref);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* The respective range must have been previously locked, as well as the inode.
|
||||
* The end offset is inclusive (last byte of the range).
|
||||
* @clone_info is NULL for fallocate's hole punching and non-NULL for extent
|
||||
* cloning.
|
||||
* When cloning, we don't want to end up in a state where we dropped extents
|
||||
* without inserting a new one, so we must abort the transaction to avoid a
|
||||
* corruption.
|
||||
*/
|
||||
int btrfs_punch_hole_range(struct inode *inode, struct btrfs_path *path,
|
||||
const u64 start, const u64 end,
|
||||
struct btrfs_clone_extent_info *clone_info,
|
||||
struct btrfs_trans_handle **trans_out)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
u64 min_size = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
u64 ino_size = round_up(inode->i_size, fs_info->sectorsize);
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
struct btrfs_trans_handle *trans = NULL;
|
||||
struct btrfs_block_rsv *rsv;
|
||||
unsigned int rsv_count;
|
||||
u64 cur_offset;
|
||||
u64 drop_end;
|
||||
u64 len = end - start;
|
||||
int ret = 0;
|
||||
|
||||
if (end <= start)
|
||||
return -EINVAL;
|
||||
|
||||
rsv = btrfs_alloc_block_rsv(fs_info, BTRFS_BLOCK_RSV_TEMP);
|
||||
if (!rsv) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
rsv->size = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
rsv->failfast = 1;
|
||||
|
||||
/*
|
||||
* 1 - update the inode
|
||||
* 1 - removing the extents in the range
|
||||
* 1 - adding the hole extent if no_holes isn't set or if we are cloning
|
||||
* an extent
|
||||
*/
|
||||
if (!btrfs_fs_incompat(fs_info, NO_HOLES) || clone_info)
|
||||
rsv_count = 3;
|
||||
else
|
||||
rsv_count = 2;
|
||||
|
||||
trans = btrfs_start_transaction(root, rsv_count);
|
||||
if (IS_ERR(trans)) {
|
||||
ret = PTR_ERR(trans);
|
||||
trans = NULL;
|
||||
goto out_free;
|
||||
}
|
||||
|
||||
ret = btrfs_block_rsv_migrate(&fs_info->trans_block_rsv, rsv,
|
||||
min_size, false);
|
||||
BUG_ON(ret);
|
||||
trans->block_rsv = rsv;
|
||||
|
||||
cur_offset = start;
|
||||
while (cur_offset < end) {
|
||||
ret = __btrfs_drop_extents(trans, root, inode, path,
|
||||
cur_offset, end + 1, &drop_end,
|
||||
1, 0, 0, NULL);
|
||||
if (ret != -ENOSPC) {
|
||||
/*
|
||||
* When cloning we want to avoid transaction aborts when
|
||||
* nothing was done and we are attempting to clone parts
|
||||
* of inline extents, in such cases -EOPNOTSUPP is
|
||||
* returned by __btrfs_drop_extents() without having
|
||||
* changed anything in the file.
|
||||
*/
|
||||
if (clone_info && ret && ret != -EOPNOTSUPP)
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
break;
|
||||
}
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
|
||||
if (!clone_info && cur_offset < drop_end &&
|
||||
cur_offset < ino_size) {
|
||||
ret = fill_holes(trans, BTRFS_I(inode), path,
|
||||
cur_offset, drop_end);
|
||||
if (ret) {
|
||||
/*
|
||||
* If we failed then we didn't insert our hole
|
||||
* entries for the area we dropped, so now the
|
||||
* fs is corrupted, so we must abort the
|
||||
* transaction.
|
||||
*/
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (clone_info) {
|
||||
u64 clone_len = drop_end - cur_offset;
|
||||
|
||||
ret = btrfs_insert_clone_extent(trans, inode, path,
|
||||
clone_info, clone_len);
|
||||
if (ret) {
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
break;
|
||||
}
|
||||
clone_info->data_len -= clone_len;
|
||||
clone_info->data_offset += clone_len;
|
||||
clone_info->file_offset += clone_len;
|
||||
}
|
||||
|
||||
cur_offset = drop_end;
|
||||
|
||||
ret = btrfs_update_inode(trans, root, inode);
|
||||
if (ret)
|
||||
break;
|
||||
|
||||
btrfs_end_transaction(trans);
|
||||
btrfs_btree_balance_dirty(fs_info);
|
||||
|
||||
trans = btrfs_start_transaction(root, rsv_count);
|
||||
if (IS_ERR(trans)) {
|
||||
ret = PTR_ERR(trans);
|
||||
trans = NULL;
|
||||
break;
|
||||
}
|
||||
|
||||
ret = btrfs_block_rsv_migrate(&fs_info->trans_block_rsv,
|
||||
rsv, min_size, false);
|
||||
BUG_ON(ret); /* shouldn't happen */
|
||||
trans->block_rsv = rsv;
|
||||
|
||||
if (!clone_info) {
|
||||
ret = find_first_non_hole(inode, &cur_offset, &len);
|
||||
if (unlikely(ret < 0))
|
||||
break;
|
||||
if (ret && !len) {
|
||||
ret = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* If we were cloning, force the next fsync to be a full one since we
|
||||
* we replaced (or just dropped in the case of cloning holes when
|
||||
* NO_HOLES is enabled) extents and extent maps.
|
||||
* This is for the sake of simplicity, and cloning into files larger
|
||||
* than 16Mb would force the full fsync any way (when
|
||||
* try_release_extent_mapping() is invoked during page cache truncation.
|
||||
*/
|
||||
if (clone_info)
|
||||
set_bit(BTRFS_INODE_NEEDS_FULL_SYNC,
|
||||
&BTRFS_I(inode)->runtime_flags);
|
||||
|
||||
if (ret)
|
||||
goto out_trans;
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
/*
|
||||
* If we are using the NO_HOLES feature we might have had already an
|
||||
* hole that overlaps a part of the region [lockstart, lockend] and
|
||||
* ends at (or beyond) lockend. Since we have no file extent items to
|
||||
* represent holes, drop_end can be less than lockend and so we must
|
||||
* make sure we have an extent map representing the existing hole (the
|
||||
* call to __btrfs_drop_extents() might have dropped the existing extent
|
||||
* map representing the existing hole), otherwise the fast fsync path
|
||||
* will not record the existence of the hole region
|
||||
* [existing_hole_start, lockend].
|
||||
*/
|
||||
if (drop_end <= end)
|
||||
drop_end = end + 1;
|
||||
/*
|
||||
* Don't insert file hole extent item if it's for a range beyond eof
|
||||
* (because it's useless) or if it represents a 0 bytes range (when
|
||||
* cur_offset == drop_end).
|
||||
*/
|
||||
if (!clone_info && cur_offset < ino_size && cur_offset < drop_end) {
|
||||
ret = fill_holes(trans, BTRFS_I(inode), path,
|
||||
cur_offset, drop_end);
|
||||
if (ret) {
|
||||
/* Same comment as above. */
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
goto out_trans;
|
||||
}
|
||||
}
|
||||
if (clone_info) {
|
||||
ret = btrfs_insert_clone_extent(trans, inode, path, clone_info,
|
||||
clone_info->data_len);
|
||||
if (ret) {
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
goto out_trans;
|
||||
}
|
||||
}
|
||||
|
||||
out_trans:
|
||||
if (!trans)
|
||||
goto out_free;
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
if (ret)
|
||||
btrfs_end_transaction(trans);
|
||||
else
|
||||
*trans_out = trans;
|
||||
out_free:
|
||||
btrfs_free_block_rsv(fs_info, rsv);
|
||||
out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int btrfs_punch_hole(struct inode *inode, loff_t offset, loff_t len)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
struct extent_state *cached_state = NULL;
|
||||
struct btrfs_path *path;
|
||||
struct btrfs_block_rsv *rsv;
|
||||
struct btrfs_trans_handle *trans;
|
||||
struct btrfs_trans_handle *trans = NULL;
|
||||
u64 lockstart;
|
||||
u64 lockend;
|
||||
u64 tail_start;
|
||||
u64 tail_len;
|
||||
u64 orig_start = offset;
|
||||
u64 cur_offset;
|
||||
u64 min_size = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
u64 drop_end;
|
||||
int ret = 0;
|
||||
int err = 0;
|
||||
unsigned int rsv_count;
|
||||
bool same_block;
|
||||
bool no_holes = btrfs_fs_incompat(fs_info, NO_HOLES);
|
||||
u64 ino_size;
|
||||
bool truncated_block = false;
|
||||
bool updated_inode = false;
|
||||
|
@ -2566,145 +2826,24 @@ static int btrfs_punch_hole(struct inode *inode, loff_t offset, loff_t len)
|
|||
goto out;
|
||||
}
|
||||
|
||||
rsv = btrfs_alloc_block_rsv(fs_info, BTRFS_BLOCK_RSV_TEMP);
|
||||
if (!rsv) {
|
||||
ret = -ENOMEM;
|
||||
goto out_free;
|
||||
}
|
||||
rsv->size = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
rsv->failfast = 1;
|
||||
|
||||
/*
|
||||
* 1 - update the inode
|
||||
* 1 - removing the extents in the range
|
||||
* 1 - adding the hole extent if no_holes isn't set
|
||||
*/
|
||||
rsv_count = no_holes ? 2 : 3;
|
||||
trans = btrfs_start_transaction(root, rsv_count);
|
||||
if (IS_ERR(trans)) {
|
||||
err = PTR_ERR(trans);
|
||||
goto out_free;
|
||||
}
|
||||
|
||||
ret = btrfs_block_rsv_migrate(&fs_info->trans_block_rsv, rsv,
|
||||
min_size, false);
|
||||
BUG_ON(ret);
|
||||
trans->block_rsv = rsv;
|
||||
|
||||
cur_offset = lockstart;
|
||||
len = lockend - cur_offset;
|
||||
while (cur_offset < lockend) {
|
||||
ret = __btrfs_drop_extents(trans, root, inode, path,
|
||||
cur_offset, lockend + 1,
|
||||
&drop_end, 1, 0, 0, NULL);
|
||||
if (ret != -ENOSPC)
|
||||
break;
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
|
||||
if (cur_offset < drop_end && cur_offset < ino_size) {
|
||||
ret = fill_holes(trans, BTRFS_I(inode), path,
|
||||
cur_offset, drop_end);
|
||||
if (ret) {
|
||||
/*
|
||||
* If we failed then we didn't insert our hole
|
||||
* entries for the area we dropped, so now the
|
||||
* fs is corrupted, so we must abort the
|
||||
* transaction.
|
||||
*/
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
err = ret;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
cur_offset = drop_end;
|
||||
|
||||
ret = btrfs_update_inode(trans, root, inode);
|
||||
if (ret) {
|
||||
err = ret;
|
||||
break;
|
||||
}
|
||||
|
||||
btrfs_end_transaction(trans);
|
||||
btrfs_btree_balance_dirty(fs_info);
|
||||
|
||||
trans = btrfs_start_transaction(root, rsv_count);
|
||||
if (IS_ERR(trans)) {
|
||||
ret = PTR_ERR(trans);
|
||||
trans = NULL;
|
||||
break;
|
||||
}
|
||||
|
||||
ret = btrfs_block_rsv_migrate(&fs_info->trans_block_rsv,
|
||||
rsv, min_size, false);
|
||||
BUG_ON(ret); /* shouldn't happen */
|
||||
trans->block_rsv = rsv;
|
||||
|
||||
ret = find_first_non_hole(inode, &cur_offset, &len);
|
||||
if (unlikely(ret < 0))
|
||||
break;
|
||||
if (ret && !len) {
|
||||
ret = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (ret) {
|
||||
err = ret;
|
||||
goto out_trans;
|
||||
}
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
/*
|
||||
* If we are using the NO_HOLES feature we might have had already an
|
||||
* hole that overlaps a part of the region [lockstart, lockend] and
|
||||
* ends at (or beyond) lockend. Since we have no file extent items to
|
||||
* represent holes, drop_end can be less than lockend and so we must
|
||||
* make sure we have an extent map representing the existing hole (the
|
||||
* call to __btrfs_drop_extents() might have dropped the existing extent
|
||||
* map representing the existing hole), otherwise the fast fsync path
|
||||
* will not record the existence of the hole region
|
||||
* [existing_hole_start, lockend].
|
||||
*/
|
||||
if (drop_end <= lockend)
|
||||
drop_end = lockend + 1;
|
||||
/*
|
||||
* Don't insert file hole extent item if it's for a range beyond eof
|
||||
* (because it's useless) or if it represents a 0 bytes range (when
|
||||
* cur_offset == drop_end).
|
||||
*/
|
||||
if (cur_offset < ino_size && cur_offset < drop_end) {
|
||||
ret = fill_holes(trans, BTRFS_I(inode), path,
|
||||
cur_offset, drop_end);
|
||||
if (ret) {
|
||||
/* Same comment as above. */
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
err = ret;
|
||||
goto out_trans;
|
||||
}
|
||||
}
|
||||
|
||||
out_trans:
|
||||
if (!trans)
|
||||
goto out_free;
|
||||
ret = btrfs_punch_hole_range(inode, path, lockstart, lockend, NULL,
|
||||
&trans);
|
||||
btrfs_free_path(path);
|
||||
if (ret)
|
||||
goto out;
|
||||
|
||||
ASSERT(trans != NULL);
|
||||
inode_inc_iversion(inode);
|
||||
inode->i_mtime = inode->i_ctime = current_time(inode);
|
||||
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
ret = btrfs_update_inode(trans, root, inode);
|
||||
updated_inode = true;
|
||||
btrfs_end_transaction(trans);
|
||||
btrfs_btree_balance_dirty(fs_info);
|
||||
out_free:
|
||||
btrfs_free_path(path);
|
||||
btrfs_free_block_rsv(fs_info, rsv);
|
||||
out:
|
||||
unlock_extent_cached(&BTRFS_I(inode)->io_tree, lockstart, lockend,
|
||||
&cached_state);
|
||||
out_only_mutex:
|
||||
if (!updated_inode && truncated_block && !ret && !err) {
|
||||
if (!updated_inode && truncated_block && !ret) {
|
||||
/*
|
||||
* If we only end up zeroing part of a page, we still need to
|
||||
* update the inode item, so that all the time fields are
|
||||
|
@ -2719,16 +2858,18 @@ out_only_mutex:
|
|||
inode->i_ctime = now;
|
||||
trans = btrfs_start_transaction(root, 1);
|
||||
if (IS_ERR(trans)) {
|
||||
err = PTR_ERR(trans);
|
||||
ret = PTR_ERR(trans);
|
||||
} else {
|
||||
err = btrfs_update_inode(trans, root, inode);
|
||||
ret = btrfs_end_transaction(trans);
|
||||
int ret2;
|
||||
|
||||
ret = btrfs_update_inode(trans, root, inode);
|
||||
ret2 = btrfs_end_transaction(trans);
|
||||
if (!ret)
|
||||
ret = ret2;
|
||||
}
|
||||
}
|
||||
inode_unlock(inode);
|
||||
if (ret && !err)
|
||||
err = ret;
|
||||
return err;
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Helper structure to record which range is already reserved */
|
||||
|
|
|
@ -20,6 +20,7 @@
|
|||
#include "volumes.h"
|
||||
#include "space-info.h"
|
||||
#include "delalloc-space.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#define BITS_PER_BITMAP (PAGE_SIZE * 8UL)
|
||||
#define MAX_CACHE_BYTES_PER_GIG SZ_32K
|
||||
|
@ -210,8 +211,8 @@ int btrfs_check_trunc_cache_free_space(struct btrfs_fs_info *fs_info,
|
|||
int ret;
|
||||
|
||||
/* 1 for slack space, 1 for updating the inode */
|
||||
needed_bytes = btrfs_calc_trunc_metadata_size(fs_info, 1) +
|
||||
btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
needed_bytes = btrfs_calc_insert_metadata_size(fs_info, 1) +
|
||||
btrfs_calc_metadata_size(fs_info, 1);
|
||||
|
||||
spin_lock(&rsv->lock);
|
||||
if (rsv->reserved < needed_bytes)
|
||||
|
@ -764,7 +765,8 @@ static int __load_free_space_cache(struct btrfs_root *root, struct inode *inode,
|
|||
} else {
|
||||
ASSERT(num_bitmaps);
|
||||
num_bitmaps--;
|
||||
e->bitmap = kzalloc(PAGE_SIZE, GFP_NOFS);
|
||||
e->bitmap = kmem_cache_zalloc(
|
||||
btrfs_free_space_bitmap_cachep, GFP_NOFS);
|
||||
if (!e->bitmap) {
|
||||
kmem_cache_free(
|
||||
btrfs_free_space_cachep, e);
|
||||
|
@ -1004,7 +1006,7 @@ update_cache_item(struct btrfs_trans_handle *trans,
|
|||
ret = btrfs_search_slot(trans, root, &key, path, 0, 1);
|
||||
if (ret < 0) {
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, inode->i_size - 1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC, 0, 0, NULL);
|
||||
EXTENT_DELALLOC, 0, 0, NULL);
|
||||
goto fail;
|
||||
}
|
||||
leaf = path->nodes[0];
|
||||
|
@ -1016,9 +1018,8 @@ update_cache_item(struct btrfs_trans_handle *trans,
|
|||
if (found_key.objectid != BTRFS_FREE_SPACE_OBJECTID ||
|
||||
found_key.offset != offset) {
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, 0,
|
||||
inode->i_size - 1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC, 0, 0,
|
||||
NULL);
|
||||
inode->i_size - 1, EXTENT_DELALLOC, 0,
|
||||
0, NULL);
|
||||
btrfs_release_path(path);
|
||||
goto fail;
|
||||
}
|
||||
|
@ -1114,7 +1115,7 @@ static int flush_dirty_cache(struct inode *inode)
|
|||
ret = btrfs_wait_ordered_range(inode, 0, (u64)-1);
|
||||
if (ret)
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, inode->i_size - 1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC, 0, 0, NULL);
|
||||
EXTENT_DELALLOC, 0, 0, NULL);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
@ -1881,7 +1882,7 @@ static void free_bitmap(struct btrfs_free_space_ctl *ctl,
|
|||
struct btrfs_free_space *bitmap_info)
|
||||
{
|
||||
unlink_free_space(ctl, bitmap_info);
|
||||
kfree(bitmap_info->bitmap);
|
||||
kmem_cache_free(btrfs_free_space_bitmap_cachep, bitmap_info->bitmap);
|
||||
kmem_cache_free(btrfs_free_space_cachep, bitmap_info);
|
||||
ctl->total_bitmaps--;
|
||||
ctl->op->recalc_thresholds(ctl);
|
||||
|
@ -2135,7 +2136,8 @@ new_bitmap:
|
|||
}
|
||||
|
||||
/* allocate the bitmap */
|
||||
info->bitmap = kzalloc(PAGE_SIZE, GFP_NOFS);
|
||||
info->bitmap = kmem_cache_zalloc(btrfs_free_space_bitmap_cachep,
|
||||
GFP_NOFS);
|
||||
spin_lock(&ctl->tree_lock);
|
||||
if (!info->bitmap) {
|
||||
ret = -ENOMEM;
|
||||
|
@ -2146,7 +2148,9 @@ new_bitmap:
|
|||
|
||||
out:
|
||||
if (info) {
|
||||
kfree(info->bitmap);
|
||||
if (info->bitmap)
|
||||
kmem_cache_free(btrfs_free_space_bitmap_cachep,
|
||||
info->bitmap);
|
||||
kmem_cache_free(btrfs_free_space_cachep, info);
|
||||
}
|
||||
|
||||
|
@ -2376,6 +2380,14 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
int btrfs_add_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 bytenr, u64 size)
|
||||
{
|
||||
return __btrfs_add_free_space(block_group->fs_info,
|
||||
block_group->free_space_ctl,
|
||||
bytenr, size);
|
||||
}
|
||||
|
||||
int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 offset, u64 bytes)
|
||||
{
|
||||
|
@ -2802,7 +2814,8 @@ out:
|
|||
if (entry->bytes == 0) {
|
||||
ctl->free_extents--;
|
||||
if (entry->bitmap) {
|
||||
kfree(entry->bitmap);
|
||||
kmem_cache_free(btrfs_free_space_bitmap_cachep,
|
||||
entry->bitmap);
|
||||
ctl->total_bitmaps--;
|
||||
ctl->op->recalc_thresholds(ctl);
|
||||
}
|
||||
|
@ -3606,7 +3619,7 @@ again:
|
|||
}
|
||||
|
||||
if (!map) {
|
||||
map = kzalloc(PAGE_SIZE, GFP_NOFS);
|
||||
map = kmem_cache_zalloc(btrfs_free_space_bitmap_cachep, GFP_NOFS);
|
||||
if (!map) {
|
||||
kmem_cache_free(btrfs_free_space_cachep, info);
|
||||
return -ENOMEM;
|
||||
|
@ -3635,7 +3648,8 @@ again:
|
|||
|
||||
if (info)
|
||||
kmem_cache_free(btrfs_free_space_cachep, info);
|
||||
kfree(map);
|
||||
if (map)
|
||||
kmem_cache_free(btrfs_free_space_bitmap_cachep, map);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -36,7 +36,19 @@ struct btrfs_free_space_op {
|
|||
struct btrfs_free_space *info);
|
||||
};
|
||||
|
||||
struct btrfs_io_ctl;
|
||||
struct btrfs_io_ctl {
|
||||
void *cur, *orig;
|
||||
struct page *page;
|
||||
struct page **pages;
|
||||
struct btrfs_fs_info *fs_info;
|
||||
struct inode *inode;
|
||||
unsigned long size;
|
||||
int index;
|
||||
int num_pages;
|
||||
int entries;
|
||||
int bitmaps;
|
||||
unsigned check_crcs:1;
|
||||
};
|
||||
|
||||
struct inode *lookup_free_space_inode(
|
||||
struct btrfs_block_group_cache *block_group,
|
||||
|
@ -73,14 +85,8 @@ void btrfs_init_free_space_ctl(struct btrfs_block_group_cache *block_group);
|
|||
int __btrfs_add_free_space(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_free_space_ctl *ctl,
|
||||
u64 bytenr, u64 size);
|
||||
static inline int
|
||||
btrfs_add_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 bytenr, u64 size)
|
||||
{
|
||||
return __btrfs_add_free_space(block_group->fs_info,
|
||||
block_group->free_space_ctl,
|
||||
bytenr, size);
|
||||
}
|
||||
int btrfs_add_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 bytenr, u64 size);
|
||||
int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group,
|
||||
u64 bytenr, u64 size);
|
||||
void __btrfs_remove_free_space_cache(struct btrfs_free_space_ctl *ctl);
|
||||
|
|
|
@ -10,6 +10,7 @@
|
|||
#include "locking.h"
|
||||
#include "free-space-tree.h"
|
||||
#include "transaction.h"
|
||||
#include "block-group.h"
|
||||
|
||||
static int __add_block_group_free_space(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_block_group_cache *block_group,
|
||||
|
|
|
@ -6,6 +6,8 @@
|
|||
#ifndef BTRFS_FREE_SPACE_TREE_H
|
||||
#define BTRFS_FREE_SPACE_TREE_H
|
||||
|
||||
struct btrfs_caching_control;
|
||||
|
||||
/*
|
||||
* The default size for new free space bitmap items. The last bitmap in a block
|
||||
* group may be truncated, and none of the free space tree code assumes that
|
||||
|
|
|
@ -8,9 +8,9 @@
|
|||
#include "transaction.h"
|
||||
#include "print-tree.h"
|
||||
|
||||
int btrfs_find_name_in_backref(struct extent_buffer *leaf, int slot,
|
||||
const char *name,
|
||||
int name_len, struct btrfs_inode_ref **ref_ret)
|
||||
struct btrfs_inode_ref *btrfs_find_name_in_backref(struct extent_buffer *leaf,
|
||||
int slot, const char *name,
|
||||
int name_len)
|
||||
{
|
||||
struct btrfs_inode_ref *ref;
|
||||
unsigned long ptr;
|
||||
|
@ -28,19 +28,15 @@ int btrfs_find_name_in_backref(struct extent_buffer *leaf, int slot,
|
|||
cur_offset += len + sizeof(*ref);
|
||||
if (len != name_len)
|
||||
continue;
|
||||
if (memcmp_extent_buffer(leaf, name, name_ptr, name_len) == 0) {
|
||||
if (ref_ret)
|
||||
*ref_ret = ref;
|
||||
return 1;
|
||||
}
|
||||
if (memcmp_extent_buffer(leaf, name, name_ptr, name_len) == 0)
|
||||
return ref;
|
||||
}
|
||||
return 0;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
int btrfs_find_name_in_ext_backref(struct extent_buffer *leaf, int slot,
|
||||
u64 ref_objectid,
|
||||
const char *name, int name_len,
|
||||
struct btrfs_inode_extref **extref_ret)
|
||||
struct btrfs_inode_extref *btrfs_find_name_in_ext_backref(
|
||||
struct extent_buffer *leaf, int slot, u64 ref_objectid,
|
||||
const char *name, int name_len)
|
||||
{
|
||||
struct btrfs_inode_extref *extref;
|
||||
unsigned long ptr;
|
||||
|
@ -65,15 +61,12 @@ int btrfs_find_name_in_ext_backref(struct extent_buffer *leaf, int slot,
|
|||
|
||||
if (ref_name_len == name_len &&
|
||||
btrfs_inode_extref_parent(leaf, extref) == ref_objectid &&
|
||||
(memcmp_extent_buffer(leaf, name, name_ptr, name_len) == 0)) {
|
||||
if (extref_ret)
|
||||
*extref_ret = extref;
|
||||
return 1;
|
||||
}
|
||||
(memcmp_extent_buffer(leaf, name, name_ptr, name_len) == 0))
|
||||
return extref;
|
||||
|
||||
cur_offset += ref_name_len + sizeof(*extref);
|
||||
}
|
||||
return 0;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/* Returns NULL if no extref found */
|
||||
|
@ -87,7 +80,6 @@ btrfs_lookup_inode_extref(struct btrfs_trans_handle *trans,
|
|||
{
|
||||
int ret;
|
||||
struct btrfs_key key;
|
||||
struct btrfs_inode_extref *extref;
|
||||
|
||||
key.objectid = inode_objectid;
|
||||
key.type = BTRFS_INODE_EXTREF_KEY;
|
||||
|
@ -98,11 +90,9 @@ btrfs_lookup_inode_extref(struct btrfs_trans_handle *trans,
|
|||
return ERR_PTR(ret);
|
||||
if (ret > 0)
|
||||
return NULL;
|
||||
if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0],
|
||||
ref_objectid, name, name_len,
|
||||
&extref))
|
||||
return NULL;
|
||||
return extref;
|
||||
return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0],
|
||||
ref_objectid, name, name_len);
|
||||
|
||||
}
|
||||
|
||||
static int btrfs_del_inode_extref(struct btrfs_trans_handle *trans,
|
||||
|
@ -142,9 +132,9 @@ static int btrfs_del_inode_extref(struct btrfs_trans_handle *trans,
|
|||
* This should always succeed so error here will make the FS
|
||||
* readonly.
|
||||
*/
|
||||
if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0],
|
||||
ref_objectid,
|
||||
name, name_len, &extref)) {
|
||||
extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0],
|
||||
ref_objectid, name, name_len);
|
||||
if (!extref) {
|
||||
btrfs_handle_fs_error(root->fs_info, -ENOENT, NULL);
|
||||
ret = -EROFS;
|
||||
goto out;
|
||||
|
@ -213,8 +203,10 @@ int btrfs_del_inode_ref(struct btrfs_trans_handle *trans,
|
|||
} else if (ret < 0) {
|
||||
goto out;
|
||||
}
|
||||
if (!btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
|
||||
name, name_len, &ref)) {
|
||||
|
||||
ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name,
|
||||
name_len);
|
||||
if (!ref) {
|
||||
ret = -ENOENT;
|
||||
search_ext_refs = 1;
|
||||
goto out;
|
||||
|
@ -285,7 +277,7 @@ static int btrfs_insert_inode_extref(struct btrfs_trans_handle *trans,
|
|||
if (btrfs_find_name_in_ext_backref(path->nodes[0],
|
||||
path->slots[0],
|
||||
ref_objectid,
|
||||
name, name_len, NULL))
|
||||
name, name_len))
|
||||
goto out;
|
||||
|
||||
btrfs_extend_item(path, ins_len);
|
||||
|
@ -341,9 +333,9 @@ int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans,
|
|||
ins_len);
|
||||
if (ret == -EEXIST) {
|
||||
u32 old_size;
|
||||
|
||||
if (btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
|
||||
name, name_len, &ref))
|
||||
ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
|
||||
name, name_len);
|
||||
if (ref)
|
||||
goto out;
|
||||
|
||||
old_size = btrfs_item_size_nr(path->nodes[0], path->slots[0]);
|
||||
|
@ -359,7 +351,7 @@ int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans,
|
|||
if (ret == -EOVERFLOW) {
|
||||
if (btrfs_find_name_in_backref(path->nodes[0],
|
||||
path->slots[0],
|
||||
name, name_len, &ref))
|
||||
name, name_len))
|
||||
ret = -EEXIST;
|
||||
else
|
||||
ret = -EMLINK;
|
||||
|
|
|
@ -13,6 +13,19 @@
|
|||
#include "transaction.h"
|
||||
#include "delalloc-space.h"
|
||||
|
||||
static void fail_caching_thread(struct btrfs_root *root)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = root->fs_info;
|
||||
|
||||
btrfs_warn(fs_info, "failed to start inode caching task");
|
||||
btrfs_clear_pending_and_info(fs_info, INODE_MAP_CACHE,
|
||||
"disabling inode map caching");
|
||||
spin_lock(&root->ino_cache_lock);
|
||||
root->ino_cache_state = BTRFS_CACHE_ERROR;
|
||||
spin_unlock(&root->ino_cache_lock);
|
||||
wake_up(&root->ino_cache_wait);
|
||||
}
|
||||
|
||||
static int caching_kthread(void *data)
|
||||
{
|
||||
struct btrfs_root *root = data;
|
||||
|
@ -29,8 +42,10 @@ static int caching_kthread(void *data)
|
|||
return 0;
|
||||
|
||||
path = btrfs_alloc_path();
|
||||
if (!path)
|
||||
if (!path) {
|
||||
fail_caching_thread(root);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
/* Since the commit root is read-only, we can safely skip locking. */
|
||||
path->skip_locking = 1;
|
||||
|
@ -146,6 +161,7 @@ static void start_caching(struct btrfs_root *root)
|
|||
spin_lock(&root->ino_cache_lock);
|
||||
root->ino_cache_state = BTRFS_CACHE_FINISHED;
|
||||
spin_unlock(&root->ino_cache_lock);
|
||||
wake_up(&root->ino_cache_wait);
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -160,15 +176,13 @@ static void start_caching(struct btrfs_root *root)
|
|||
if (!ret && objectid <= BTRFS_LAST_FREE_OBJECTID) {
|
||||
__btrfs_add_free_space(fs_info, ctl, objectid,
|
||||
BTRFS_LAST_FREE_OBJECTID - objectid + 1);
|
||||
wake_up(&root->ino_cache_wait);
|
||||
}
|
||||
|
||||
tsk = kthread_run(caching_kthread, root, "btrfs-ino-cache-%llu",
|
||||
root->root_key.objectid);
|
||||
if (IS_ERR(tsk)) {
|
||||
btrfs_warn(fs_info, "failed to start inode caching task");
|
||||
btrfs_clear_pending_and_info(fs_info, INODE_MAP_CACHE,
|
||||
"disabling inode map caching");
|
||||
}
|
||||
if (IS_ERR(tsk))
|
||||
fail_caching_thread(root);
|
||||
}
|
||||
|
||||
int btrfs_find_free_ino(struct btrfs_root *root, u64 *objectid)
|
||||
|
@ -186,11 +200,14 @@ again:
|
|||
|
||||
wait_event(root->ino_cache_wait,
|
||||
root->ino_cache_state == BTRFS_CACHE_FINISHED ||
|
||||
root->ino_cache_state == BTRFS_CACHE_ERROR ||
|
||||
root->free_ino_ctl->free_space > 0);
|
||||
|
||||
if (root->ino_cache_state == BTRFS_CACHE_FINISHED &&
|
||||
root->free_ino_ctl->free_space == 0)
|
||||
return -ENOSPC;
|
||||
else if (root->ino_cache_state == BTRFS_CACHE_ERROR)
|
||||
return btrfs_find_free_objectid(root, objectid);
|
||||
else
|
||||
goto again;
|
||||
}
|
||||
|
@ -419,7 +436,7 @@ int btrfs_save_ino_cache(struct btrfs_root *root,
|
|||
* 1 item for free space object
|
||||
* 3 items for pre-allocation
|
||||
*/
|
||||
trans->bytes_reserved = btrfs_calc_trans_metadata_size(fs_info, 10);
|
||||
trans->bytes_reserved = btrfs_calc_insert_metadata_size(fs_info, 10);
|
||||
ret = btrfs_block_rsv_add(root, trans->block_rsv,
|
||||
trans->bytes_reserved,
|
||||
BTRFS_RESERVE_NO_FLUSH);
|
||||
|
@ -485,6 +502,7 @@ again:
|
|||
prealloc, prealloc, &alloc_hint);
|
||||
if (ret) {
|
||||
btrfs_delalloc_release_extents(BTRFS_I(inode), prealloc, true);
|
||||
btrfs_delalloc_release_metadata(BTRFS_I(inode), prealloc, true);
|
||||
goto out_put;
|
||||
}
|
||||
|
||||
|
|
399
fs/btrfs/inode.c
399
fs/btrfs/inode.c
|
@ -30,6 +30,7 @@
|
|||
#include <linux/swap.h>
|
||||
#include <linux/sched/mm.h>
|
||||
#include <asm/unaligned.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "disk-io.h"
|
||||
#include "transaction.h"
|
||||
|
@ -46,8 +47,8 @@
|
|||
#include "backref.h"
|
||||
#include "props.h"
|
||||
#include "qgroup.h"
|
||||
#include "dedupe.h"
|
||||
#include "delalloc-space.h"
|
||||
#include "block-group.h"
|
||||
|
||||
struct btrfs_iget_args {
|
||||
struct btrfs_key *location;
|
||||
|
@ -74,15 +75,15 @@ static struct kmem_cache *btrfs_inode_cachep;
|
|||
struct kmem_cache *btrfs_trans_handle_cachep;
|
||||
struct kmem_cache *btrfs_path_cachep;
|
||||
struct kmem_cache *btrfs_free_space_cachep;
|
||||
struct kmem_cache *btrfs_free_space_bitmap_cachep;
|
||||
|
||||
static int btrfs_setsize(struct inode *inode, struct iattr *attr);
|
||||
static int btrfs_truncate(struct inode *inode, bool skip_writeback);
|
||||
static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent);
|
||||
static noinline int cow_file_range(struct inode *inode,
|
||||
struct page *locked_page,
|
||||
u64 start, u64 end, u64 delalloc_end,
|
||||
int *page_started, unsigned long *nr_written,
|
||||
int unlock, struct btrfs_dedupe_hash *hash);
|
||||
u64 start, u64 end, int *page_started,
|
||||
unsigned long *nr_written, int unlock);
|
||||
static struct extent_map *create_io_em(struct inode *inode, u64 start, u64 len,
|
||||
u64 orig_start, u64 block_start,
|
||||
u64 block_len, u64 orig_block_len,
|
||||
|
@ -178,6 +179,9 @@ static int insert_inline_extent(struct btrfs_trans_handle *trans,
|
|||
size_t cur_size = size;
|
||||
unsigned long offset;
|
||||
|
||||
ASSERT((compressed_size > 0 && compressed_pages) ||
|
||||
(compressed_size == 0 && !compressed_pages));
|
||||
|
||||
if (compressed_size && compressed_pages)
|
||||
cur_size = compressed_size;
|
||||
|
||||
|
@ -462,8 +466,7 @@ static inline void inode_should_defrag(struct btrfs_inode *inode,
|
|||
* are written in the same order that the flusher thread sent them
|
||||
* down.
|
||||
*/
|
||||
static noinline void compress_file_range(struct async_chunk *async_chunk,
|
||||
int *num_added)
|
||||
static noinline int compress_file_range(struct async_chunk *async_chunk)
|
||||
{
|
||||
struct inode *inode = async_chunk->inode;
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
|
@ -479,6 +482,7 @@ static noinline void compress_file_range(struct async_chunk *async_chunk,
|
|||
int i;
|
||||
int will_compress;
|
||||
int compress_type = fs_info->compress_type;
|
||||
int compressed_extents = 0;
|
||||
int redirty = 0;
|
||||
|
||||
inode_should_defrag(BTRFS_I(inode), start, end, end - start + 1,
|
||||
|
@ -615,14 +619,21 @@ cont:
|
|||
* our outstanding extent for clearing delalloc for this
|
||||
* range.
|
||||
*/
|
||||
extent_clear_unlock_delalloc(inode, start, end, end,
|
||||
NULL, clear_flags,
|
||||
extent_clear_unlock_delalloc(inode, start, end, NULL,
|
||||
clear_flags,
|
||||
PAGE_UNLOCK |
|
||||
PAGE_CLEAR_DIRTY |
|
||||
PAGE_SET_WRITEBACK |
|
||||
page_error_op |
|
||||
PAGE_END_WRITEBACK);
|
||||
goto free_pages_out;
|
||||
|
||||
for (i = 0; i < nr_pages; i++) {
|
||||
WARN_ON(pages[i]->mapping);
|
||||
put_page(pages[i]);
|
||||
}
|
||||
kfree(pages);
|
||||
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -641,7 +652,7 @@ cont:
|
|||
*/
|
||||
total_in = ALIGN(total_in, PAGE_SIZE);
|
||||
if (total_compressed + blocksize <= total_in) {
|
||||
*num_added += 1;
|
||||
compressed_extents++;
|
||||
|
||||
/*
|
||||
* The async work queues will take care of doing actual
|
||||
|
@ -658,7 +669,7 @@ cont:
|
|||
cond_resched();
|
||||
goto again;
|
||||
}
|
||||
return;
|
||||
return compressed_extents;
|
||||
}
|
||||
}
|
||||
if (pages) {
|
||||
|
@ -697,16 +708,9 @@ cleanup_and_bail_uncompressed:
|
|||
extent_range_redirty_for_io(inode, start, end);
|
||||
add_async_extent(async_chunk, start, end - start + 1, 0, NULL, 0,
|
||||
BTRFS_COMPRESS_NONE);
|
||||
*num_added += 1;
|
||||
compressed_extents++;
|
||||
|
||||
return;
|
||||
|
||||
free_pages_out:
|
||||
for (i = 0; i < nr_pages; i++) {
|
||||
WARN_ON(pages[i]->mapping);
|
||||
put_page(pages[i]);
|
||||
}
|
||||
kfree(pages);
|
||||
return compressed_extents;
|
||||
}
|
||||
|
||||
static void free_async_extent_pages(struct async_extent *async_extent)
|
||||
|
@ -762,10 +766,7 @@ retry:
|
|||
async_extent->start,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
&page_started, &nr_written, 0,
|
||||
NULL);
|
||||
&page_started, &nr_written, 0);
|
||||
|
||||
/* JDM XXX */
|
||||
|
||||
|
@ -853,8 +854,6 @@ retry:
|
|||
* clear dirty, set writeback and unlock the pages.
|
||||
*/
|
||||
extent_clear_unlock_delalloc(inode, async_extent->start,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
NULL, EXTENT_LOCKED | EXTENT_DELALLOC,
|
||||
|
@ -875,7 +874,7 @@ retry:
|
|||
btrfs_writepage_endio_finish_ordered(p, start, end, 0);
|
||||
|
||||
p->mapping = NULL;
|
||||
extent_clear_unlock_delalloc(inode, start, end, end,
|
||||
extent_clear_unlock_delalloc(inode, start, end,
|
||||
NULL, 0,
|
||||
PAGE_END_WRITEBACK |
|
||||
PAGE_SET_ERROR);
|
||||
|
@ -891,8 +890,6 @@ out_free_reserve:
|
|||
btrfs_free_reserved_extent(fs_info, ins.objectid, ins.offset, 1);
|
||||
out_free:
|
||||
extent_clear_unlock_delalloc(inode, async_extent->start,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
async_extent->start +
|
||||
async_extent->ram_size - 1,
|
||||
NULL, EXTENT_LOCKED | EXTENT_DELALLOC |
|
||||
|
@ -953,9 +950,8 @@ static u64 get_extent_allocation_hint(struct inode *inode, u64 start,
|
|||
*/
|
||||
static noinline int cow_file_range(struct inode *inode,
|
||||
struct page *locked_page,
|
||||
u64 start, u64 end, u64 delalloc_end,
|
||||
int *page_started, unsigned long *nr_written,
|
||||
int unlock, struct btrfs_dedupe_hash *hash)
|
||||
u64 start, u64 end, int *page_started,
|
||||
unsigned long *nr_written, int unlock)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
|
@ -994,8 +990,7 @@ static noinline int cow_file_range(struct inode *inode,
|
|||
* our outstanding extent for clearing delalloc for this
|
||||
* range.
|
||||
*/
|
||||
extent_clear_unlock_delalloc(inode, start, end,
|
||||
delalloc_end, NULL,
|
||||
extent_clear_unlock_delalloc(inode, start, end, NULL,
|
||||
EXTENT_LOCKED | EXTENT_DELALLOC |
|
||||
EXTENT_DELALLOC_NEW | EXTENT_DEFRAG |
|
||||
EXTENT_DO_ACCOUNTING, PAGE_UNLOCK |
|
||||
|
@ -1078,7 +1073,7 @@ static noinline int cow_file_range(struct inode *inode,
|
|||
|
||||
extent_clear_unlock_delalloc(inode, start,
|
||||
start + ram_size - 1,
|
||||
delalloc_end, locked_page,
|
||||
locked_page,
|
||||
EXTENT_LOCKED | EXTENT_DELALLOC,
|
||||
page_ops);
|
||||
if (num_bytes < cur_alloc_size)
|
||||
|
@ -1122,7 +1117,6 @@ out_unlock:
|
|||
*/
|
||||
if (extent_reserved) {
|
||||
extent_clear_unlock_delalloc(inode, start,
|
||||
start + cur_alloc_size,
|
||||
start + cur_alloc_size,
|
||||
locked_page,
|
||||
clear_bits,
|
||||
|
@ -1131,8 +1125,7 @@ out_unlock:
|
|||
if (start >= end)
|
||||
goto out;
|
||||
}
|
||||
extent_clear_unlock_delalloc(inode, start, end, delalloc_end,
|
||||
locked_page,
|
||||
extent_clear_unlock_delalloc(inode, start, end, locked_page,
|
||||
clear_bits | EXTENT_CLEAR_DATA_RESV,
|
||||
page_ops);
|
||||
goto out;
|
||||
|
@ -1144,12 +1137,12 @@ out_unlock:
|
|||
static noinline void async_cow_start(struct btrfs_work *work)
|
||||
{
|
||||
struct async_chunk *async_chunk;
|
||||
int num_added = 0;
|
||||
int compressed_extents;
|
||||
|
||||
async_chunk = container_of(work, struct async_chunk, work);
|
||||
|
||||
compress_file_range(async_chunk, &num_added);
|
||||
if (num_added == 0) {
|
||||
compressed_extents = compress_file_range(async_chunk);
|
||||
if (compressed_extents == 0) {
|
||||
btrfs_add_delayed_iput(async_chunk->inode);
|
||||
async_chunk->inode = NULL;
|
||||
}
|
||||
|
@ -1235,7 +1228,7 @@ static int cow_file_range_async(struct inode *inode, struct page *locked_page,
|
|||
PAGE_SET_WRITEBACK | PAGE_END_WRITEBACK |
|
||||
PAGE_SET_ERROR;
|
||||
|
||||
extent_clear_unlock_delalloc(inode, start, end, 0, locked_page,
|
||||
extent_clear_unlock_delalloc(inode, start, end, locked_page,
|
||||
clear_bits, page_ops);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
@ -1310,36 +1303,25 @@ static noinline int csum_exist_in_range(struct btrfs_fs_info *fs_info,
|
|||
*/
|
||||
static noinline int run_delalloc_nocow(struct inode *inode,
|
||||
struct page *locked_page,
|
||||
u64 start, u64 end, int *page_started, int force,
|
||||
unsigned long *nr_written)
|
||||
const u64 start, const u64 end,
|
||||
int *page_started, int force,
|
||||
unsigned long *nr_written)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
struct extent_buffer *leaf;
|
||||
struct btrfs_path *path;
|
||||
struct btrfs_file_extent_item *fi;
|
||||
struct btrfs_key found_key;
|
||||
struct extent_map *em;
|
||||
u64 cow_start;
|
||||
u64 cur_offset;
|
||||
u64 extent_end;
|
||||
u64 extent_offset;
|
||||
u64 disk_bytenr;
|
||||
u64 num_bytes;
|
||||
u64 disk_num_bytes;
|
||||
u64 ram_bytes;
|
||||
int extent_type;
|
||||
u64 cow_start = (u64)-1;
|
||||
u64 cur_offset = start;
|
||||
int ret;
|
||||
int type;
|
||||
int nocow;
|
||||
int check_prev = 1;
|
||||
bool nolock;
|
||||
bool check_prev = true;
|
||||
const bool freespace_inode = btrfs_is_free_space_inode(BTRFS_I(inode));
|
||||
u64 ino = btrfs_ino(BTRFS_I(inode));
|
||||
bool nocow = false;
|
||||
u64 disk_bytenr = 0;
|
||||
|
||||
path = btrfs_alloc_path();
|
||||
if (!path) {
|
||||
extent_clear_unlock_delalloc(inode, start, end, end,
|
||||
locked_page,
|
||||
extent_clear_unlock_delalloc(inode, start, end, locked_page,
|
||||
EXTENT_LOCKED | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING |
|
||||
EXTENT_DEFRAG, PAGE_UNLOCK |
|
||||
|
@ -1349,15 +1331,29 @@ static noinline int run_delalloc_nocow(struct inode *inode,
|
|||
return -ENOMEM;
|
||||
}
|
||||
|
||||
nolock = btrfs_is_free_space_inode(BTRFS_I(inode));
|
||||
|
||||
cow_start = (u64)-1;
|
||||
cur_offset = start;
|
||||
while (1) {
|
||||
struct btrfs_key found_key;
|
||||
struct btrfs_file_extent_item *fi;
|
||||
struct extent_buffer *leaf;
|
||||
u64 extent_end;
|
||||
u64 extent_offset;
|
||||
u64 num_bytes = 0;
|
||||
u64 disk_num_bytes;
|
||||
u64 ram_bytes;
|
||||
int extent_type;
|
||||
|
||||
nocow = false;
|
||||
|
||||
ret = btrfs_lookup_file_extent(NULL, root, path, ino,
|
||||
cur_offset, 0);
|
||||
if (ret < 0)
|
||||
goto error;
|
||||
|
||||
/*
|
||||
* If there is no extent for our range when doing the initial
|
||||
* search, then go back to the previous slot as it will be the
|
||||
* one containing the search offset
|
||||
*/
|
||||
if (ret > 0 && path->slots[0] > 0 && check_prev) {
|
||||
leaf = path->nodes[0];
|
||||
btrfs_item_key_to_cpu(leaf, &found_key,
|
||||
|
@ -1366,8 +1362,9 @@ static noinline int run_delalloc_nocow(struct inode *inode,
|
|||
found_key.type == BTRFS_EXTENT_DATA_KEY)
|
||||
path->slots[0]--;
|
||||
}
|
||||
check_prev = 0;
|
||||
check_prev = false;
|
||||
next_slot:
|
||||
/* Go to next leaf if we have exhausted the current one */
|
||||
leaf = path->nodes[0];
|
||||
if (path->slots[0] >= btrfs_header_nritems(leaf)) {
|
||||
ret = btrfs_next_leaf(root, path);
|
||||
|
@ -1381,28 +1378,40 @@ next_slot:
|
|||
leaf = path->nodes[0];
|
||||
}
|
||||
|
||||
nocow = 0;
|
||||
disk_bytenr = 0;
|
||||
num_bytes = 0;
|
||||
btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]);
|
||||
|
||||
/* Didn't find anything for our INO */
|
||||
if (found_key.objectid > ino)
|
||||
break;
|
||||
/*
|
||||
* Keep searching until we find an EXTENT_ITEM or there are no
|
||||
* more extents for this inode
|
||||
*/
|
||||
if (WARN_ON_ONCE(found_key.objectid < ino) ||
|
||||
found_key.type < BTRFS_EXTENT_DATA_KEY) {
|
||||
path->slots[0]++;
|
||||
goto next_slot;
|
||||
}
|
||||
|
||||
/* Found key is not EXTENT_DATA_KEY or starts after req range */
|
||||
if (found_key.type > BTRFS_EXTENT_DATA_KEY ||
|
||||
found_key.offset > end)
|
||||
break;
|
||||
|
||||
/*
|
||||
* If the found extent starts after requested offset, then
|
||||
* adjust extent_end to be right before this extent begins
|
||||
*/
|
||||
if (found_key.offset > cur_offset) {
|
||||
extent_end = found_key.offset;
|
||||
extent_type = 0;
|
||||
goto out_check;
|
||||
}
|
||||
|
||||
/*
|
||||
* Found extent which begins before our range and potentially
|
||||
* intersect it
|
||||
*/
|
||||
fi = btrfs_item_ptr(leaf, path->slots[0],
|
||||
struct btrfs_file_extent_item);
|
||||
extent_type = btrfs_file_extent_type(leaf, fi);
|
||||
|
@ -1416,26 +1425,36 @@ next_slot:
|
|||
btrfs_file_extent_num_bytes(leaf, fi);
|
||||
disk_num_bytes =
|
||||
btrfs_file_extent_disk_num_bytes(leaf, fi);
|
||||
/*
|
||||
* If extent we got ends before our range starts, skip
|
||||
* to next extent
|
||||
*/
|
||||
if (extent_end <= start) {
|
||||
path->slots[0]++;
|
||||
goto next_slot;
|
||||
}
|
||||
/* Skip holes */
|
||||
if (disk_bytenr == 0)
|
||||
goto out_check;
|
||||
/* Skip compressed/encrypted/encoded extents */
|
||||
if (btrfs_file_extent_compression(leaf, fi) ||
|
||||
btrfs_file_extent_encryption(leaf, fi) ||
|
||||
btrfs_file_extent_other_encoding(leaf, fi))
|
||||
goto out_check;
|
||||
/*
|
||||
* Do the same check as in btrfs_cross_ref_exist but
|
||||
* without the unnecessary search.
|
||||
* If extent is created before the last volume's snapshot
|
||||
* this implies the extent is shared, hence we can't do
|
||||
* nocow. This is the same check as in
|
||||
* btrfs_cross_ref_exist but without calling
|
||||
* btrfs_search_slot.
|
||||
*/
|
||||
if (!nolock &&
|
||||
if (!freespace_inode &&
|
||||
btrfs_file_extent_generation(leaf, fi) <=
|
||||
btrfs_root_last_snapshot(&root->root_item))
|
||||
goto out_check;
|
||||
if (extent_type == BTRFS_FILE_EXTENT_REG && !force)
|
||||
goto out_check;
|
||||
/* If extent is RO, we must COW it */
|
||||
if (btrfs_extent_readonly(fs_info, disk_bytenr))
|
||||
goto out_check;
|
||||
ret = btrfs_cross_ref_exist(root, ino,
|
||||
|
@ -1452,17 +1471,17 @@ next_slot:
|
|||
goto error;
|
||||
}
|
||||
|
||||
WARN_ON_ONCE(nolock);
|
||||
WARN_ON_ONCE(freespace_inode);
|
||||
goto out_check;
|
||||
}
|
||||
disk_bytenr += extent_offset;
|
||||
disk_bytenr += cur_offset - found_key.offset;
|
||||
num_bytes = min(end + 1, extent_end) - cur_offset;
|
||||
/*
|
||||
* if there are pending snapshots for this root,
|
||||
* we fall into common COW way.
|
||||
* If there are pending snapshots for this root, we
|
||||
* fall into common COW way
|
||||
*/
|
||||
if (!nolock && atomic_read(&root->snapshot_force_cow))
|
||||
if (!freespace_inode && atomic_read(&root->snapshot_force_cow))
|
||||
goto out_check;
|
||||
/*
|
||||
* force cow if csum exists in the range.
|
||||
|
@ -1481,27 +1500,29 @@ next_slot:
|
|||
cur_offset = cow_start;
|
||||
goto error;
|
||||
}
|
||||
WARN_ON_ONCE(nolock);
|
||||
WARN_ON_ONCE(freespace_inode);
|
||||
goto out_check;
|
||||
}
|
||||
if (!btrfs_inc_nocow_writers(fs_info, disk_bytenr))
|
||||
goto out_check;
|
||||
nocow = 1;
|
||||
nocow = true;
|
||||
} else if (extent_type == BTRFS_FILE_EXTENT_INLINE) {
|
||||
extent_end = found_key.offset +
|
||||
btrfs_file_extent_ram_bytes(leaf, fi);
|
||||
extent_end = ALIGN(extent_end,
|
||||
fs_info->sectorsize);
|
||||
extent_end = found_key.offset + ram_bytes;
|
||||
extent_end = ALIGN(extent_end, fs_info->sectorsize);
|
||||
/* Skip extents outside of our requested range */
|
||||
if (extent_end <= start) {
|
||||
path->slots[0]++;
|
||||
goto next_slot;
|
||||
}
|
||||
} else {
|
||||
/* If this triggers then we have a memory corruption */
|
||||
BUG();
|
||||
}
|
||||
out_check:
|
||||
if (extent_end <= start) {
|
||||
path->slots[0]++;
|
||||
if (nocow)
|
||||
btrfs_dec_nocow_writers(fs_info, disk_bytenr);
|
||||
goto next_slot;
|
||||
}
|
||||
/*
|
||||
* If nocow is false then record the beginning of the range
|
||||
* that needs to be COWed
|
||||
*/
|
||||
if (!nocow) {
|
||||
if (cow_start == (u64)-1)
|
||||
cow_start = cur_offset;
|
||||
|
@ -1513,11 +1534,16 @@ out_check:
|
|||
}
|
||||
|
||||
btrfs_release_path(path);
|
||||
|
||||
/*
|
||||
* COW range from cow_start to found_key.offset - 1. As the key
|
||||
* will contain the beginning of the first extent that can be
|
||||
* NOCOW, following one which needs to be COW'ed
|
||||
*/
|
||||
if (cow_start != (u64)-1) {
|
||||
ret = cow_file_range(inode, locked_page,
|
||||
cow_start, found_key.offset - 1,
|
||||
end, page_started, nr_written, 1,
|
||||
NULL);
|
||||
page_started, nr_written, 1);
|
||||
if (ret) {
|
||||
if (nocow)
|
||||
btrfs_dec_nocow_writers(fs_info,
|
||||
|
@ -1529,6 +1555,7 @@ out_check:
|
|||
|
||||
if (extent_type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
u64 orig_start = found_key.offset - extent_offset;
|
||||
struct extent_map *em;
|
||||
|
||||
em = create_io_em(inode, cur_offset, num_bytes,
|
||||
orig_start,
|
||||
|
@ -1545,19 +1572,29 @@ out_check:
|
|||
goto error;
|
||||
}
|
||||
free_extent_map(em);
|
||||
}
|
||||
|
||||
if (extent_type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
type = BTRFS_ORDERED_PREALLOC;
|
||||
ret = btrfs_add_ordered_extent(inode, cur_offset,
|
||||
disk_bytenr, num_bytes,
|
||||
num_bytes,
|
||||
BTRFS_ORDERED_PREALLOC);
|
||||
if (ret) {
|
||||
btrfs_drop_extent_cache(BTRFS_I(inode),
|
||||
cur_offset,
|
||||
cur_offset + num_bytes - 1,
|
||||
0);
|
||||
goto error;
|
||||
}
|
||||
} else {
|
||||
type = BTRFS_ORDERED_NOCOW;
|
||||
ret = btrfs_add_ordered_extent(inode, cur_offset,
|
||||
disk_bytenr, num_bytes,
|
||||
num_bytes,
|
||||
BTRFS_ORDERED_NOCOW);
|
||||
if (ret)
|
||||
goto error;
|
||||
}
|
||||
|
||||
ret = btrfs_add_ordered_extent(inode, cur_offset, disk_bytenr,
|
||||
num_bytes, num_bytes, type);
|
||||
if (nocow)
|
||||
btrfs_dec_nocow_writers(fs_info, disk_bytenr);
|
||||
BUG_ON(ret); /* -ENOMEM */
|
||||
nocow = false;
|
||||
|
||||
if (root->root_key.objectid ==
|
||||
BTRFS_DATA_RELOC_TREE_OBJECTID)
|
||||
|
@ -1570,7 +1607,7 @@ out_check:
|
|||
num_bytes);
|
||||
|
||||
extent_clear_unlock_delalloc(inode, cur_offset,
|
||||
cur_offset + num_bytes - 1, end,
|
||||
cur_offset + num_bytes - 1,
|
||||
locked_page, EXTENT_LOCKED |
|
||||
EXTENT_DELALLOC |
|
||||
EXTENT_CLEAR_DATA_RESV,
|
||||
|
@ -1595,15 +1632,18 @@ out_check:
|
|||
|
||||
if (cow_start != (u64)-1) {
|
||||
cur_offset = end;
|
||||
ret = cow_file_range(inode, locked_page, cow_start, end, end,
|
||||
page_started, nr_written, 1, NULL);
|
||||
ret = cow_file_range(inode, locked_page, cow_start, end,
|
||||
page_started, nr_written, 1);
|
||||
if (ret)
|
||||
goto error;
|
||||
}
|
||||
|
||||
error:
|
||||
if (nocow)
|
||||
btrfs_dec_nocow_writers(fs_info, disk_bytenr);
|
||||
|
||||
if (ret && cur_offset < end)
|
||||
extent_clear_unlock_delalloc(inode, cur_offset, end, end,
|
||||
extent_clear_unlock_delalloc(inode, cur_offset, end,
|
||||
locked_page, EXTENT_LOCKED |
|
||||
EXTENT_DELALLOC | EXTENT_DEFRAG |
|
||||
EXTENT_DO_ACCOUNTING, PAGE_UNLOCK |
|
||||
|
@ -1654,8 +1694,8 @@ int btrfs_run_delalloc_range(struct inode *inode, struct page *locked_page,
|
|||
page_started, 0, nr_written);
|
||||
} else if (!inode_can_compress(inode) ||
|
||||
!inode_need_compress(inode, start, end)) {
|
||||
ret = cow_file_range(inode, locked_page, start, end, end,
|
||||
page_started, nr_written, 1, NULL);
|
||||
ret = cow_file_range(inode, locked_page, start, end,
|
||||
page_started, nr_written, 1);
|
||||
} else {
|
||||
set_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
||||
&BTRFS_I(inode)->runtime_flags);
|
||||
|
@ -2090,7 +2130,7 @@ static noinline int add_pending_csums(struct btrfs_trans_handle *trans,
|
|||
|
||||
int btrfs_set_extent_delalloc(struct inode *inode, u64 start, u64 end,
|
||||
unsigned int extra_bits,
|
||||
struct extent_state **cached_state, int dedupe)
|
||||
struct extent_state **cached_state)
|
||||
{
|
||||
WARN_ON(PAGE_ALIGNED(end));
|
||||
return set_extent_delalloc(&BTRFS_I(inode)->io_tree, start, end,
|
||||
|
@ -2156,7 +2196,7 @@ again:
|
|||
}
|
||||
|
||||
ret = btrfs_set_extent_delalloc(inode, page_start, page_end, 0,
|
||||
&cached_state, 0);
|
||||
&cached_state);
|
||||
if (ret) {
|
||||
mapping_set_error(page->mapping, ret);
|
||||
end_extent_writepage(page, ret, page_start, page_end);
|
||||
|
@ -3850,7 +3890,7 @@ static void fill_inode_item(struct btrfs_trans_handle *trans,
|
|||
{
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
|
||||
btrfs_set_token_inode_uid(leaf, item, i_uid_read(inode), &token);
|
||||
btrfs_set_token_inode_gid(leaf, item, i_gid_read(inode), &token);
|
||||
|
@ -4946,12 +4986,11 @@ again:
|
|||
}
|
||||
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, block_start, block_end,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG,
|
||||
0, 0, &cached_state);
|
||||
EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG,
|
||||
0, 0, &cached_state);
|
||||
|
||||
ret = btrfs_set_extent_delalloc(inode, block_start, block_end, 0,
|
||||
&cached_state, 0);
|
||||
&cached_state);
|
||||
if (ret) {
|
||||
unlock_extent_cached(io_tree, block_start, block_end,
|
||||
&cached_state);
|
||||
|
@ -5332,9 +5371,9 @@ static void evict_inode_truncate_pages(struct inode *inode)
|
|||
btrfs_qgroup_free_data(inode, NULL, start, end - start + 1);
|
||||
|
||||
clear_extent_bit(io_tree, start, end,
|
||||
EXTENT_LOCKED | EXTENT_DIRTY |
|
||||
EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING |
|
||||
EXTENT_DEFRAG, 1, 1, &cached_state);
|
||||
EXTENT_LOCKED | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG, 1, 1,
|
||||
&cached_state);
|
||||
|
||||
cond_resched();
|
||||
spin_lock(&io_tree->lock);
|
||||
|
@ -5347,59 +5386,50 @@ static struct btrfs_trans_handle *evict_refill_and_join(struct btrfs_root *root,
|
|||
{
|
||||
struct btrfs_fs_info *fs_info = root->fs_info;
|
||||
struct btrfs_block_rsv *global_rsv = &fs_info->global_block_rsv;
|
||||
u64 delayed_refs_extra = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
int failures = 0;
|
||||
|
||||
for (;;) {
|
||||
struct btrfs_trans_handle *trans;
|
||||
int ret;
|
||||
|
||||
ret = btrfs_block_rsv_refill(root, rsv,
|
||||
rsv->size + delayed_refs_extra,
|
||||
BTRFS_RESERVE_FLUSH_LIMIT);
|
||||
|
||||
if (ret && ++failures > 2) {
|
||||
btrfs_warn(fs_info,
|
||||
"could not allocate space for a delete; will truncate on mount");
|
||||
return ERR_PTR(-ENOSPC);
|
||||
}
|
||||
|
||||
/*
|
||||
* Evict can generate a large amount of delayed refs without
|
||||
* having a way to add space back since we exhaust our temporary
|
||||
* block rsv. We aren't allowed to do FLUSH_ALL in this case
|
||||
* because we could deadlock with so many things in the flushing
|
||||
* code, so we have to try and hold some extra space to
|
||||
* compensate for our delayed ref generation. If we can't get
|
||||
* that space then we need see if we can steal our minimum from
|
||||
* the global reserve. We will be ratelimited by the amount of
|
||||
* space we have for the delayed refs rsv, so we'll end up
|
||||
* committing and trying again.
|
||||
*/
|
||||
trans = btrfs_join_transaction(root);
|
||||
if (IS_ERR(trans) || !ret) {
|
||||
if (!IS_ERR(trans)) {
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
trans->bytes_reserved = delayed_refs_extra;
|
||||
btrfs_block_rsv_migrate(rsv, trans->block_rsv,
|
||||
delayed_refs_extra, 1);
|
||||
}
|
||||
return trans;
|
||||
}
|
||||
struct btrfs_trans_handle *trans;
|
||||
u64 delayed_refs_extra = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Eviction should be taking place at some place safe because of our
|
||||
* delayed iputs. However the normal flushing code will run delayed
|
||||
* iputs, so we cannot use FLUSH_ALL otherwise we'll deadlock.
|
||||
*
|
||||
* We reserve the delayed_refs_extra here again because we can't use
|
||||
* btrfs_start_transaction(root, 0) for the same deadlocky reason as
|
||||
* above. We reserve our extra bit here because we generate a ton of
|
||||
* delayed refs activity by truncating.
|
||||
*
|
||||
* If we cannot make our reservation we'll attempt to steal from the
|
||||
* global reserve, because we really want to be able to free up space.
|
||||
*/
|
||||
ret = btrfs_block_rsv_refill(root, rsv, rsv->size + delayed_refs_extra,
|
||||
BTRFS_RESERVE_FLUSH_EVICT);
|
||||
if (ret) {
|
||||
/*
|
||||
* Try to steal from the global reserve if there is space for
|
||||
* it.
|
||||
*/
|
||||
if (!btrfs_check_space_for_delayed_refs(fs_info) &&
|
||||
!btrfs_block_rsv_migrate(global_rsv, rsv, rsv->size, 0))
|
||||
return trans;
|
||||
|
||||
/* If not, commit and try again. */
|
||||
ret = btrfs_commit_transaction(trans);
|
||||
if (ret)
|
||||
return ERR_PTR(ret);
|
||||
if (btrfs_check_space_for_delayed_refs(fs_info) ||
|
||||
btrfs_block_rsv_migrate(global_rsv, rsv, rsv->size, 0)) {
|
||||
btrfs_warn(fs_info,
|
||||
"could not allocate space for delete; will truncate on mount");
|
||||
return ERR_PTR(-ENOSPC);
|
||||
}
|
||||
delayed_refs_extra = 0;
|
||||
}
|
||||
|
||||
trans = btrfs_join_transaction(root);
|
||||
if (IS_ERR(trans))
|
||||
return trans;
|
||||
|
||||
if (delayed_refs_extra) {
|
||||
trans->block_rsv = &fs_info->trans_block_rsv;
|
||||
trans->bytes_reserved = delayed_refs_extra;
|
||||
btrfs_block_rsv_migrate(rsv, trans->block_rsv,
|
||||
delayed_refs_extra, 1);
|
||||
}
|
||||
return trans;
|
||||
}
|
||||
|
||||
void btrfs_evict_inode(struct inode *inode)
|
||||
|
@ -5446,7 +5476,7 @@ void btrfs_evict_inode(struct inode *inode)
|
|||
rsv = btrfs_alloc_block_rsv(fs_info, BTRFS_BLOCK_RSV_TEMP);
|
||||
if (!rsv)
|
||||
goto no_delete;
|
||||
rsv->size = btrfs_calc_trunc_metadata_size(fs_info, 1);
|
||||
rsv->size = btrfs_calc_metadata_size(fs_info, 1);
|
||||
rsv->failfast = 1;
|
||||
|
||||
btrfs_i_size_write(BTRFS_I(inode), 0);
|
||||
|
@ -7701,12 +7731,9 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
|
|||
u64 start = iblock << inode->i_blkbits;
|
||||
u64 lockstart, lockend;
|
||||
u64 len = bh_result->b_size;
|
||||
int unlock_bits = EXTENT_LOCKED;
|
||||
int ret = 0;
|
||||
|
||||
if (create)
|
||||
unlock_bits |= EXTENT_DIRTY;
|
||||
else
|
||||
if (!create)
|
||||
len = min_t(u64, len, fs_info->sectorsize);
|
||||
|
||||
lockstart = start;
|
||||
|
@ -7765,9 +7792,8 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
|
|||
if (ret < 0)
|
||||
goto unlock_err;
|
||||
|
||||
/* clear and unlock the entire range */
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, lockstart, lockend,
|
||||
unlock_bits, 1, 0, &cached_state);
|
||||
unlock_extent_cached(&BTRFS_I(inode)->io_tree, lockstart,
|
||||
lockend, &cached_state);
|
||||
} else {
|
||||
ret = btrfs_get_blocks_direct_read(em, bh_result, inode,
|
||||
start, len);
|
||||
|
@ -7783,9 +7809,8 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
|
|||
*/
|
||||
lockstart = start + bh_result->b_size;
|
||||
if (lockstart < lockend) {
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, lockstart,
|
||||
lockend, unlock_bits, 1, 0,
|
||||
&cached_state);
|
||||
unlock_extent_cached(&BTRFS_I(inode)->io_tree,
|
||||
lockstart, lockend, &cached_state);
|
||||
} else {
|
||||
free_extent_state(cached_state);
|
||||
}
|
||||
|
@ -7796,8 +7821,8 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
|
|||
return 0;
|
||||
|
||||
unlock_err:
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, lockstart, lockend,
|
||||
unlock_bits, 1, 0, &cached_state);
|
||||
unlock_extent_cached(&BTRFS_I(inode)->io_tree, lockstart, lockend,
|
||||
&cached_state);
|
||||
err:
|
||||
if (dio_data)
|
||||
current->journal_info = dio_data;
|
||||
|
@ -8812,8 +8837,7 @@ again:
|
|||
*/
|
||||
if (!inode_evicting)
|
||||
clear_extent_bit(tree, start, end,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DELALLOC_NEW |
|
||||
EXTENT_DELALLOC | EXTENT_DELALLOC_NEW |
|
||||
EXTENT_LOCKED | EXTENT_DO_ACCOUNTING |
|
||||
EXTENT_DEFRAG, 1, 0, &cached_state);
|
||||
/*
|
||||
|
@ -8868,8 +8892,7 @@ again:
|
|||
if (PageDirty(page))
|
||||
btrfs_qgroup_free_data(inode, NULL, page_start, PAGE_SIZE);
|
||||
if (!inode_evicting) {
|
||||
clear_extent_bit(tree, page_start, page_end,
|
||||
EXTENT_LOCKED | EXTENT_DIRTY |
|
||||
clear_extent_bit(tree, page_start, page_end, EXTENT_LOCKED |
|
||||
EXTENT_DELALLOC | EXTENT_DELALLOC_NEW |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG, 1, 1,
|
||||
&cached_state);
|
||||
|
@ -8997,12 +9020,11 @@ again:
|
|||
* reserve data&meta space before lock_page() (see above comments).
|
||||
*/
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, page_start, end,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG,
|
||||
0, 0, &cached_state);
|
||||
EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING |
|
||||
EXTENT_DEFRAG, 0, 0, &cached_state);
|
||||
|
||||
ret2 = btrfs_set_extent_delalloc(inode, page_start, end, 0,
|
||||
&cached_state, 0);
|
||||
&cached_state);
|
||||
if (ret2) {
|
||||
unlock_extent_cached(io_tree, page_start, page_end,
|
||||
&cached_state);
|
||||
|
@ -9060,7 +9082,7 @@ static int btrfs_truncate(struct inode *inode, bool skip_writeback)
|
|||
int ret;
|
||||
struct btrfs_trans_handle *trans;
|
||||
u64 mask = fs_info->sectorsize - 1;
|
||||
u64 min_size = btrfs_calc_trunc_metadata_size(fs_info, 1);
|
||||
u64 min_size = btrfs_calc_metadata_size(fs_info, 1);
|
||||
|
||||
if (!skip_writeback) {
|
||||
ret = btrfs_wait_ordered_range(inode, inode->i_size & (~mask),
|
||||
|
@ -9380,6 +9402,7 @@ void __cold btrfs_destroy_cachep(void)
|
|||
kmem_cache_destroy(btrfs_trans_handle_cachep);
|
||||
kmem_cache_destroy(btrfs_path_cachep);
|
||||
kmem_cache_destroy(btrfs_free_space_cachep);
|
||||
kmem_cache_destroy(btrfs_free_space_bitmap_cachep);
|
||||
}
|
||||
|
||||
int __init btrfs_init_cachep(void)
|
||||
|
@ -9409,6 +9432,12 @@ int __init btrfs_init_cachep(void)
|
|||
if (!btrfs_free_space_cachep)
|
||||
goto fail;
|
||||
|
||||
btrfs_free_space_bitmap_cachep = kmem_cache_create("btrfs_free_space_bitmap",
|
||||
PAGE_SIZE, PAGE_SIZE,
|
||||
SLAB_RED_ZONE, NULL);
|
||||
if (!btrfs_free_space_bitmap_cachep)
|
||||
goto fail;
|
||||
|
||||
return 0;
|
||||
fail:
|
||||
btrfs_destroy_cachep();
|
||||
|
|
436
fs/btrfs/ioctl.c
436
fs/btrfs/ioctl.c
|
@ -45,6 +45,7 @@
|
|||
#include "compression.h"
|
||||
#include "space-info.h"
|
||||
#include "delalloc-space.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
/* If we have a 32-bit userspace and 64-bit kernel, then the UAPI
|
||||
|
@ -1332,9 +1333,8 @@ again:
|
|||
lock_extent_bits(&BTRFS_I(inode)->io_tree,
|
||||
page_start, page_end - 1, &cached_state);
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, page_start,
|
||||
page_end - 1, EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_DO_ACCOUNTING | EXTENT_DEFRAG, 0, 0,
|
||||
&cached_state);
|
||||
page_end - 1, EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING |
|
||||
EXTENT_DEFRAG, 0, 0, &cached_state);
|
||||
|
||||
if (i_done != page_cnt) {
|
||||
spin_lock(&BTRFS_I(inode)->lock);
|
||||
|
@ -1840,8 +1840,15 @@ static noinline int btrfs_ioctl_snap_create_v2(struct file *file,
|
|||
goto free_args;
|
||||
}
|
||||
|
||||
if (vol_args->flags & BTRFS_SUBVOL_CREATE_ASYNC)
|
||||
if (vol_args->flags & BTRFS_SUBVOL_CREATE_ASYNC) {
|
||||
struct inode *inode = file_inode(file);
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
|
||||
|
||||
btrfs_warn(fs_info,
|
||||
"SNAP_CREATE_V2 ioctl with CREATE_ASYNC is deprecated and will be removed in kernel 5.7");
|
||||
|
||||
ptr = &transid;
|
||||
}
|
||||
if (vol_args->flags & BTRFS_SUBVOL_RDONLY)
|
||||
readonly = true;
|
||||
if (vol_args->flags & BTRFS_SUBVOL_QGROUP_INHERIT) {
|
||||
|
@ -3324,61 +3331,6 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static void clone_update_extent_map(struct btrfs_inode *inode,
|
||||
const struct btrfs_trans_handle *trans,
|
||||
const struct btrfs_path *path,
|
||||
const u64 hole_offset,
|
||||
const u64 hole_len)
|
||||
{
|
||||
struct extent_map_tree *em_tree = &inode->extent_tree;
|
||||
struct extent_map *em;
|
||||
int ret;
|
||||
|
||||
em = alloc_extent_map();
|
||||
if (!em) {
|
||||
set_bit(BTRFS_INODE_NEEDS_FULL_SYNC, &inode->runtime_flags);
|
||||
return;
|
||||
}
|
||||
|
||||
if (path) {
|
||||
struct btrfs_file_extent_item *fi;
|
||||
|
||||
fi = btrfs_item_ptr(path->nodes[0], path->slots[0],
|
||||
struct btrfs_file_extent_item);
|
||||
btrfs_extent_item_to_extent_map(inode, path, fi, false, em);
|
||||
em->generation = -1;
|
||||
if (btrfs_file_extent_type(path->nodes[0], fi) ==
|
||||
BTRFS_FILE_EXTENT_INLINE)
|
||||
set_bit(BTRFS_INODE_NEEDS_FULL_SYNC,
|
||||
&inode->runtime_flags);
|
||||
} else {
|
||||
em->start = hole_offset;
|
||||
em->len = hole_len;
|
||||
em->ram_bytes = em->len;
|
||||
em->orig_start = hole_offset;
|
||||
em->block_start = EXTENT_MAP_HOLE;
|
||||
em->block_len = 0;
|
||||
em->orig_block_len = 0;
|
||||
em->compress_type = BTRFS_COMPRESS_NONE;
|
||||
em->generation = trans->transid;
|
||||
}
|
||||
|
||||
while (1) {
|
||||
write_lock(&em_tree->lock);
|
||||
ret = add_extent_mapping(em_tree, em, 1);
|
||||
write_unlock(&em_tree->lock);
|
||||
if (ret != -EEXIST) {
|
||||
free_extent_map(em);
|
||||
break;
|
||||
}
|
||||
btrfs_drop_extent_cache(inode, em->start,
|
||||
em->start + em->len - 1, 0);
|
||||
}
|
||||
|
||||
if (ret)
|
||||
set_bit(BTRFS_INODE_NEEDS_FULL_SYNC, &inode->runtime_flags);
|
||||
}
|
||||
|
||||
/*
|
||||
* Make sure we do not end up inserting an inline extent into a file that has
|
||||
* already other (non-inline) extents. If a file has an inline extent it can
|
||||
|
@ -3519,6 +3471,7 @@ copy_inline_extent:
|
|||
path->slots[0]),
|
||||
size);
|
||||
inode_add_bytes(dst, datal);
|
||||
set_bit(BTRFS_INODE_NEEDS_FULL_SYNC, &BTRFS_I(dst)->runtime_flags);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -3570,6 +3523,14 @@ static int btrfs_clone(struct inode *src, struct inode *inode,
|
|||
|
||||
while (1) {
|
||||
u64 next_key_min_offset = key.offset + 1;
|
||||
struct btrfs_file_extent_item *extent;
|
||||
int type;
|
||||
u32 size;
|
||||
struct btrfs_key new_key;
|
||||
u64 disko = 0, diskl = 0;
|
||||
u64 datao = 0, datal = 0;
|
||||
u8 comp;
|
||||
u64 drop_start;
|
||||
|
||||
/*
|
||||
* note the key will change type as we walk through the
|
||||
|
@ -3610,75 +3571,115 @@ process_slot:
|
|||
key.objectid != btrfs_ino(BTRFS_I(src)))
|
||||
break;
|
||||
|
||||
if (key.type == BTRFS_EXTENT_DATA_KEY) {
|
||||
struct btrfs_file_extent_item *extent;
|
||||
int type;
|
||||
u32 size;
|
||||
struct btrfs_key new_key;
|
||||
u64 disko = 0, diskl = 0;
|
||||
u64 datao = 0, datal = 0;
|
||||
u8 comp;
|
||||
u64 drop_start;
|
||||
ASSERT(key.type == BTRFS_EXTENT_DATA_KEY);
|
||||
|
||||
extent = btrfs_item_ptr(leaf, slot,
|
||||
struct btrfs_file_extent_item);
|
||||
comp = btrfs_file_extent_compression(leaf, extent);
|
||||
type = btrfs_file_extent_type(leaf, extent);
|
||||
if (type == BTRFS_FILE_EXTENT_REG ||
|
||||
type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
disko = btrfs_file_extent_disk_bytenr(leaf,
|
||||
extent);
|
||||
diskl = btrfs_file_extent_disk_num_bytes(leaf,
|
||||
extent);
|
||||
datao = btrfs_file_extent_offset(leaf, extent);
|
||||
datal = btrfs_file_extent_num_bytes(leaf,
|
||||
extent);
|
||||
} else if (type == BTRFS_FILE_EXTENT_INLINE) {
|
||||
/* take upper bound, may be compressed */
|
||||
datal = btrfs_file_extent_ram_bytes(leaf,
|
||||
extent);
|
||||
extent = btrfs_item_ptr(leaf, slot,
|
||||
struct btrfs_file_extent_item);
|
||||
comp = btrfs_file_extent_compression(leaf, extent);
|
||||
type = btrfs_file_extent_type(leaf, extent);
|
||||
if (type == BTRFS_FILE_EXTENT_REG ||
|
||||
type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
disko = btrfs_file_extent_disk_bytenr(leaf, extent);
|
||||
diskl = btrfs_file_extent_disk_num_bytes(leaf, extent);
|
||||
datao = btrfs_file_extent_offset(leaf, extent);
|
||||
datal = btrfs_file_extent_num_bytes(leaf, extent);
|
||||
} else if (type == BTRFS_FILE_EXTENT_INLINE) {
|
||||
/* Take upper bound, may be compressed */
|
||||
datal = btrfs_file_extent_ram_bytes(leaf, extent);
|
||||
}
|
||||
|
||||
/*
|
||||
* The first search might have left us at an extent item that
|
||||
* ends before our target range's start, can happen if we have
|
||||
* holes and NO_HOLES feature enabled.
|
||||
*/
|
||||
if (key.offset + datal <= off) {
|
||||
path->slots[0]++;
|
||||
goto process_slot;
|
||||
} else if (key.offset >= off + len) {
|
||||
break;
|
||||
}
|
||||
next_key_min_offset = key.offset + datal;
|
||||
size = btrfs_item_size_nr(leaf, slot);
|
||||
read_extent_buffer(leaf, buf, btrfs_item_ptr_offset(leaf, slot),
|
||||
size);
|
||||
|
||||
btrfs_release_path(path);
|
||||
path->leave_spinning = 0;
|
||||
|
||||
memcpy(&new_key, &key, sizeof(new_key));
|
||||
new_key.objectid = btrfs_ino(BTRFS_I(inode));
|
||||
if (off <= key.offset)
|
||||
new_key.offset = key.offset + destoff - off;
|
||||
else
|
||||
new_key.offset = destoff;
|
||||
|
||||
/*
|
||||
* Deal with a hole that doesn't have an extent item that
|
||||
* represents it (NO_HOLES feature enabled).
|
||||
* This hole is either in the middle of the cloning range or at
|
||||
* the beginning (fully overlaps it or partially overlaps it).
|
||||
*/
|
||||
if (new_key.offset != last_dest_end)
|
||||
drop_start = last_dest_end;
|
||||
else
|
||||
drop_start = new_key.offset;
|
||||
|
||||
if (type == BTRFS_FILE_EXTENT_REG ||
|
||||
type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
struct btrfs_clone_extent_info clone_info;
|
||||
|
||||
/*
|
||||
* a | --- range to clone ---| b
|
||||
* | ------------- extent ------------- |
|
||||
*/
|
||||
|
||||
/* Subtract range b */
|
||||
if (key.offset + datal > off + len)
|
||||
datal = off + len - key.offset;
|
||||
|
||||
/* Subtract range a */
|
||||
if (off > key.offset) {
|
||||
datao += off - key.offset;
|
||||
datal -= off - key.offset;
|
||||
}
|
||||
|
||||
/*
|
||||
* The first search might have left us at an extent
|
||||
* item that ends before our target range's start, can
|
||||
* happen if we have holes and NO_HOLES feature enabled.
|
||||
*/
|
||||
if (key.offset + datal <= off) {
|
||||
path->slots[0]++;
|
||||
goto process_slot;
|
||||
} else if (key.offset >= off + len) {
|
||||
break;
|
||||
clone_info.disk_offset = disko;
|
||||
clone_info.disk_len = diskl;
|
||||
clone_info.data_offset = datao;
|
||||
clone_info.data_len = datal;
|
||||
clone_info.file_offset = new_key.offset;
|
||||
clone_info.extent_buf = buf;
|
||||
clone_info.item_size = size;
|
||||
ret = btrfs_punch_hole_range(inode, path,
|
||||
drop_start,
|
||||
new_key.offset + datal - 1,
|
||||
&clone_info, &trans);
|
||||
if (ret)
|
||||
goto out;
|
||||
} else if (type == BTRFS_FILE_EXTENT_INLINE) {
|
||||
u64 skip = 0;
|
||||
u64 trim = 0;
|
||||
|
||||
if (off > key.offset) {
|
||||
skip = off - key.offset;
|
||||
new_key.offset += skip;
|
||||
}
|
||||
next_key_min_offset = key.offset + datal;
|
||||
size = btrfs_item_size_nr(leaf, slot);
|
||||
read_extent_buffer(leaf, buf,
|
||||
btrfs_item_ptr_offset(leaf, slot),
|
||||
size);
|
||||
|
||||
btrfs_release_path(path);
|
||||
path->leave_spinning = 0;
|
||||
if (key.offset + datal > off + len)
|
||||
trim = key.offset + datal - (off + len);
|
||||
|
||||
memcpy(&new_key, &key, sizeof(new_key));
|
||||
new_key.objectid = btrfs_ino(BTRFS_I(inode));
|
||||
if (off <= key.offset)
|
||||
new_key.offset = key.offset + destoff - off;
|
||||
else
|
||||
new_key.offset = destoff;
|
||||
|
||||
/*
|
||||
* Deal with a hole that doesn't have an extent item
|
||||
* that represents it (NO_HOLES feature enabled).
|
||||
* This hole is either in the middle of the cloning
|
||||
* range or at the beginning (fully overlaps it or
|
||||
* partially overlaps it).
|
||||
*/
|
||||
if (new_key.offset != last_dest_end)
|
||||
drop_start = last_dest_end;
|
||||
else
|
||||
drop_start = new_key.offset;
|
||||
if (comp && (skip || trim)) {
|
||||
ret = -EINVAL;
|
||||
goto out;
|
||||
}
|
||||
size -= skip + trim;
|
||||
datal -= skip + trim;
|
||||
|
||||
/*
|
||||
* If our extent is inline, we know we will drop or
|
||||
* adjust at most 1 extent item in the destination root.
|
||||
*
|
||||
* 1 - adjusting old extent (we may have to split it)
|
||||
* 1 - add new extent
|
||||
* 1 - inode update
|
||||
|
@ -3689,140 +3690,28 @@ process_slot:
|
|||
goto out;
|
||||
}
|
||||
|
||||
if (type == BTRFS_FILE_EXTENT_REG ||
|
||||
type == BTRFS_FILE_EXTENT_PREALLOC) {
|
||||
/*
|
||||
* a | --- range to clone ---| b
|
||||
* | ------------- extent ------------- |
|
||||
*/
|
||||
|
||||
/* subtract range b */
|
||||
if (key.offset + datal > off + len)
|
||||
datal = off + len - key.offset;
|
||||
|
||||
/* subtract range a */
|
||||
if (off > key.offset) {
|
||||
datao += off - key.offset;
|
||||
datal -= off - key.offset;
|
||||
}
|
||||
|
||||
ret = btrfs_drop_extents(trans, root, inode,
|
||||
drop_start,
|
||||
new_key.offset + datal,
|
||||
1);
|
||||
if (ret) {
|
||||
if (ret != -EOPNOTSUPP)
|
||||
btrfs_abort_transaction(trans,
|
||||
ret);
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
}
|
||||
|
||||
ret = btrfs_insert_empty_item(trans, root, path,
|
||||
&new_key, size);
|
||||
if (ret) {
|
||||
ret = clone_copy_inline_extent(inode, trans, path,
|
||||
&new_key, drop_start,
|
||||
datal, skip, size, buf);
|
||||
if (ret) {
|
||||
if (ret != -EOPNOTSUPP)
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
}
|
||||
|
||||
leaf = path->nodes[0];
|
||||
slot = path->slots[0];
|
||||
write_extent_buffer(leaf, buf,
|
||||
btrfs_item_ptr_offset(leaf, slot),
|
||||
size);
|
||||
|
||||
extent = btrfs_item_ptr(leaf, slot,
|
||||
struct btrfs_file_extent_item);
|
||||
|
||||
/* disko == 0 means it's a hole */
|
||||
if (!disko)
|
||||
datao = 0;
|
||||
|
||||
btrfs_set_file_extent_offset(leaf, extent,
|
||||
datao);
|
||||
btrfs_set_file_extent_num_bytes(leaf, extent,
|
||||
datal);
|
||||
|
||||
if (disko) {
|
||||
struct btrfs_ref ref = { 0 };
|
||||
inode_add_bytes(inode, datal);
|
||||
btrfs_init_generic_ref(&ref,
|
||||
BTRFS_ADD_DELAYED_REF, disko,
|
||||
diskl, 0);
|
||||
btrfs_init_data_ref(&ref,
|
||||
root->root_key.objectid,
|
||||
btrfs_ino(BTRFS_I(inode)),
|
||||
new_key.offset - datao);
|
||||
ret = btrfs_inc_extent_ref(trans, &ref);
|
||||
if (ret) {
|
||||
btrfs_abort_transaction(trans,
|
||||
ret);
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
|
||||
}
|
||||
}
|
||||
} else if (type == BTRFS_FILE_EXTENT_INLINE) {
|
||||
u64 skip = 0;
|
||||
u64 trim = 0;
|
||||
|
||||
if (off > key.offset) {
|
||||
skip = off - key.offset;
|
||||
new_key.offset += skip;
|
||||
}
|
||||
|
||||
if (key.offset + datal > off + len)
|
||||
trim = key.offset + datal - (off + len);
|
||||
|
||||
if (comp && (skip || trim)) {
|
||||
ret = -EINVAL;
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
}
|
||||
size -= skip + trim;
|
||||
datal -= skip + trim;
|
||||
|
||||
ret = clone_copy_inline_extent(inode,
|
||||
trans, path,
|
||||
&new_key,
|
||||
drop_start,
|
||||
datal,
|
||||
skip, size, buf);
|
||||
if (ret) {
|
||||
if (ret != -EOPNOTSUPP)
|
||||
btrfs_abort_transaction(trans,
|
||||
ret);
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
}
|
||||
leaf = path->nodes[0];
|
||||
slot = path->slots[0];
|
||||
}
|
||||
|
||||
/* If we have an implicit hole (NO_HOLES feature). */
|
||||
if (drop_start < new_key.offset)
|
||||
clone_update_extent_map(BTRFS_I(inode), trans,
|
||||
NULL, drop_start,
|
||||
new_key.offset - drop_start);
|
||||
|
||||
clone_update_extent_map(BTRFS_I(inode), trans,
|
||||
path, 0, 0);
|
||||
|
||||
btrfs_mark_buffer_dirty(leaf);
|
||||
btrfs_release_path(path);
|
||||
|
||||
last_dest_end = ALIGN(new_key.offset + datal,
|
||||
fs_info->sectorsize);
|
||||
ret = clone_finish_inode_update(trans, inode,
|
||||
last_dest_end,
|
||||
destoff, olen,
|
||||
no_time_update);
|
||||
if (ret)
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
if (new_key.offset + datal >= destoff + len)
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
btrfs_release_path(path);
|
||||
|
||||
last_dest_end = ALIGN(new_key.offset + datal,
|
||||
fs_info->sectorsize);
|
||||
ret = clone_finish_inode_update(trans, inode, last_dest_end,
|
||||
destoff, olen, no_time_update);
|
||||
if (ret)
|
||||
goto out;
|
||||
if (new_key.offset + datal >= destoff + len)
|
||||
break;
|
||||
|
||||
btrfs_release_path(path);
|
||||
key.offset = next_key_min_offset;
|
||||
|
||||
|
@ -3834,32 +3723,27 @@ process_slot:
|
|||
ret = 0;
|
||||
|
||||
if (last_dest_end < destoff + len) {
|
||||
struct btrfs_clone_extent_info clone_info = { 0 };
|
||||
/*
|
||||
* We have an implicit hole (NO_HOLES feature is enabled) that
|
||||
* fully or partially overlaps our cloning range at its end.
|
||||
*/
|
||||
btrfs_release_path(path);
|
||||
path->leave_spinning = 0;
|
||||
|
||||
/*
|
||||
* 1 - remove extent(s)
|
||||
* 1 - inode update
|
||||
* We are dealing with a hole and our clone_info already has a
|
||||
* disk_offset of 0, we only need to fill the data length and
|
||||
* file offset.
|
||||
*/
|
||||
trans = btrfs_start_transaction(root, 2);
|
||||
if (IS_ERR(trans)) {
|
||||
ret = PTR_ERR(trans);
|
||||
clone_info.data_len = destoff + len - last_dest_end;
|
||||
clone_info.file_offset = last_dest_end;
|
||||
ret = btrfs_punch_hole_range(inode, path,
|
||||
last_dest_end, destoff + len - 1,
|
||||
&clone_info, &trans);
|
||||
if (ret)
|
||||
goto out;
|
||||
}
|
||||
ret = btrfs_drop_extents(trans, root, inode,
|
||||
last_dest_end, destoff + len, 1);
|
||||
if (ret) {
|
||||
if (ret != -EOPNOTSUPP)
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
btrfs_end_transaction(trans);
|
||||
goto out;
|
||||
}
|
||||
clone_update_extent_map(BTRFS_I(inode), trans, NULL,
|
||||
last_dest_end,
|
||||
destoff + len - last_dest_end);
|
||||
|
||||
ret = clone_finish_inode_update(trans, inode, destoff + len,
|
||||
destoff, olen, no_time_update);
|
||||
}
|
||||
|
@ -4313,6 +4197,9 @@ static noinline long btrfs_ioctl_start_sync(struct btrfs_root *root,
|
|||
u64 transid;
|
||||
int ret;
|
||||
|
||||
btrfs_warn(root->fs_info,
|
||||
"START_SYNC ioctl is deprecated and will be removed in kernel 5.7");
|
||||
|
||||
trans = btrfs_attach_transaction_barrier(root);
|
||||
if (IS_ERR(trans)) {
|
||||
if (PTR_ERR(trans) != -ENOENT)
|
||||
|
@ -4340,6 +4227,9 @@ static noinline long btrfs_ioctl_wait_sync(struct btrfs_fs_info *fs_info,
|
|||
{
|
||||
u64 transid;
|
||||
|
||||
btrfs_warn(fs_info,
|
||||
"WAIT_SYNC ioctl is deprecated and will be removed in kernel 5.7");
|
||||
|
||||
if (argp) {
|
||||
if (copy_from_user(&transid, argp, sizeof(transid)))
|
||||
return -EFAULT;
|
||||
|
@ -5381,7 +5271,7 @@ static int check_feature_bits(struct btrfs_fs_info *fs_info,
|
|||
u64 change_mask, u64 flags, u64 supported_flags,
|
||||
u64 safe_set, u64 safe_clear)
|
||||
{
|
||||
const char *type = btrfs_feature_set_names[set];
|
||||
const char *type = btrfs_feature_set_name(set);
|
||||
char *names;
|
||||
u64 disallowed, unsupported;
|
||||
u64 set_mask = flags & change_mask;
|
||||
|
@ -5562,6 +5452,10 @@ long btrfs_ioctl(struct file *file, unsigned int
|
|||
return btrfs_ioctl_setflags(file, argp);
|
||||
case FS_IOC_GETVERSION:
|
||||
return btrfs_ioctl_getversion(file, argp);
|
||||
case FS_IOC_GETFSLABEL:
|
||||
return btrfs_ioctl_get_fslabel(file, argp);
|
||||
case FS_IOC_SETFSLABEL:
|
||||
return btrfs_ioctl_set_fslabel(file, argp);
|
||||
case FITRIM:
|
||||
return btrfs_ioctl_fitrim(file, argp);
|
||||
case BTRFS_IOC_SNAP_CREATE:
|
||||
|
@ -5673,10 +5567,6 @@ long btrfs_ioctl(struct file *file, unsigned int
|
|||
return btrfs_ioctl_quota_rescan_wait(file, argp);
|
||||
case BTRFS_IOC_DEV_REPLACE:
|
||||
return btrfs_ioctl_dev_replace(fs_info, argp);
|
||||
case BTRFS_IOC_GET_FSLABEL:
|
||||
return btrfs_ioctl_get_fslabel(file, argp);
|
||||
case BTRFS_IOC_SET_FSLABEL:
|
||||
return btrfs_ioctl_set_fslabel(file, argp);
|
||||
case BTRFS_IOC_GET_SUPPORTED_FEATURES:
|
||||
return btrfs_ioctl_get_supported_features(argp);
|
||||
case BTRFS_IOC_GET_FEATURES:
|
||||
|
|
|
@ -8,6 +8,7 @@
|
|||
#include <linux/spinlock.h>
|
||||
#include <linux/page-flags.h>
|
||||
#include <asm/bug.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "extent_io.h"
|
||||
#include "locking.h"
|
||||
|
@ -119,42 +120,6 @@ void btrfs_set_lock_blocking_write(struct extent_buffer *eb)
|
|||
}
|
||||
}
|
||||
|
||||
void btrfs_clear_lock_blocking_read(struct extent_buffer *eb)
|
||||
{
|
||||
trace_btrfs_clear_lock_blocking_read(eb);
|
||||
/*
|
||||
* No lock is required. The lock owner may change if we have a read
|
||||
* lock, but it won't change to or away from us. If we have the write
|
||||
* lock, we are the owner and it'll never change.
|
||||
*/
|
||||
if (eb->lock_nested && current->pid == eb->lock_owner)
|
||||
return;
|
||||
BUG_ON(atomic_read(&eb->blocking_readers) == 0);
|
||||
read_lock(&eb->lock);
|
||||
btrfs_assert_spinning_readers_get(eb);
|
||||
/* atomic_dec_and_test implies a barrier */
|
||||
if (atomic_dec_and_test(&eb->blocking_readers))
|
||||
cond_wake_up_nomb(&eb->read_lock_wq);
|
||||
}
|
||||
|
||||
void btrfs_clear_lock_blocking_write(struct extent_buffer *eb)
|
||||
{
|
||||
trace_btrfs_clear_lock_blocking_write(eb);
|
||||
/*
|
||||
* no lock is required. The lock owner may change if
|
||||
* we have a read lock, but it won't change to or away
|
||||
* from us. If we have the write lock, we are the owner
|
||||
* and it'll never change.
|
||||
*/
|
||||
if (eb->lock_nested && current->pid == eb->lock_owner)
|
||||
return;
|
||||
write_lock(&eb->lock);
|
||||
BUG_ON(eb->blocking_writers != 1);
|
||||
btrfs_assert_spinning_writers_get(eb);
|
||||
if (--eb->blocking_writers == 0)
|
||||
cond_wake_up(&eb->write_lock_wq);
|
||||
}
|
||||
|
||||
/*
|
||||
* take a spinning read lock. This will wait for any blocking
|
||||
* writers
|
||||
|
|
|
@ -19,8 +19,6 @@ void btrfs_tree_read_unlock(struct extent_buffer *eb);
|
|||
void btrfs_tree_read_unlock_blocking(struct extent_buffer *eb);
|
||||
void btrfs_set_lock_blocking_read(struct extent_buffer *eb);
|
||||
void btrfs_set_lock_blocking_write(struct extent_buffer *eb);
|
||||
void btrfs_clear_lock_blocking_read(struct extent_buffer *eb);
|
||||
void btrfs_clear_lock_blocking_write(struct extent_buffer *eb);
|
||||
void btrfs_assert_tree_locked(struct extent_buffer *eb);
|
||||
int btrfs_try_tree_read_lock(struct extent_buffer *eb);
|
||||
int btrfs_try_tree_write_lock(struct extent_buffer *eb);
|
||||
|
|
|
@ -507,11 +507,6 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static unsigned int lzo_set_level(unsigned int level)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
const struct btrfs_compress_op btrfs_lzo_compress = {
|
||||
.init_workspace_manager = lzo_init_workspace_manager,
|
||||
.cleanup_workspace_manager = lzo_cleanup_workspace_manager,
|
||||
|
@ -522,5 +517,6 @@ const struct btrfs_compress_op btrfs_lzo_compress = {
|
|||
.compress_pages = lzo_compress_pages,
|
||||
.decompress_bio = lzo_decompress_bio,
|
||||
.decompress = lzo_decompress,
|
||||
.set_level = lzo_set_level,
|
||||
.max_level = 1,
|
||||
.default_level = 1,
|
||||
};
|
||||
|
|
|
@ -1,28 +0,0 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
/*
|
||||
* Copyright (C) 2012 Fujitsu. All rights reserved.
|
||||
* Written by Miao Xie <miaox@cn.fujitsu.com>
|
||||
*/
|
||||
|
||||
#ifndef BTRFS_MATH_H
|
||||
#define BTRFS_MATH_H
|
||||
|
||||
#include <asm/div64.h>
|
||||
|
||||
static inline u64 div_factor(u64 num, int factor)
|
||||
{
|
||||
if (factor == 10)
|
||||
return num;
|
||||
num *= factor;
|
||||
return div_u64(num, 10);
|
||||
}
|
||||
|
||||
static inline u64 div_factor_fine(u64 num, int factor)
|
||||
{
|
||||
if (factor == 100)
|
||||
return num;
|
||||
num *= factor;
|
||||
return div_u64(num, 100);
|
||||
}
|
||||
|
||||
#endif
|
|
@ -0,0 +1,50 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
|
||||
#ifndef BTRFS_MISC_H
|
||||
#define BTRFS_MISC_H
|
||||
|
||||
#include <linux/sched.h>
|
||||
#include <linux/wait.h>
|
||||
#include <asm/div64.h>
|
||||
|
||||
#define in_range(b, first, len) ((b) >= (first) && (b) < (first) + (len))
|
||||
|
||||
static inline void cond_wake_up(struct wait_queue_head *wq)
|
||||
{
|
||||
/*
|
||||
* This implies a full smp_mb barrier, see comments for
|
||||
* waitqueue_active why.
|
||||
*/
|
||||
if (wq_has_sleeper(wq))
|
||||
wake_up(wq);
|
||||
}
|
||||
|
||||
static inline void cond_wake_up_nomb(struct wait_queue_head *wq)
|
||||
{
|
||||
/*
|
||||
* Special case for conditional wakeup where the barrier required for
|
||||
* waitqueue_active is implied by some of the preceding code. Eg. one
|
||||
* of such atomic operations (atomic_dec_and_return, ...), or a
|
||||
* unlock/lock sequence, etc.
|
||||
*/
|
||||
if (waitqueue_active(wq))
|
||||
wake_up(wq);
|
||||
}
|
||||
|
||||
static inline u64 div_factor(u64 num, int factor)
|
||||
{
|
||||
if (factor == 10)
|
||||
return num;
|
||||
num *= factor;
|
||||
return div_u64(num, 10);
|
||||
}
|
||||
|
||||
static inline u64 div_factor_fine(u64 num, int factor)
|
||||
{
|
||||
if (factor == 100)
|
||||
return num;
|
||||
num *= factor;
|
||||
return div_u64(num, 100);
|
||||
}
|
||||
|
||||
#endif
|
|
@ -7,6 +7,7 @@
|
|||
#include <linux/blkdev.h>
|
||||
#include <linux/writeback.h>
|
||||
#include <linux/sched/mm.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "transaction.h"
|
||||
#include "btrfs_inode.h"
|
||||
|
|
|
@ -362,7 +362,7 @@ static int inherit_props(struct btrfs_trans_handle *trans,
|
|||
* reservations if we do add more properties in the future.
|
||||
*/
|
||||
if (need_reserve) {
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
ret = btrfs_block_rsv_add(root, trans->block_rsv,
|
||||
num_bytes, BTRFS_RESERVE_NO_FLUSH);
|
||||
if (ret)
|
||||
|
|
|
@ -21,7 +21,7 @@
|
|||
#include "backref.h"
|
||||
#include "extent_io.h"
|
||||
#include "qgroup.h"
|
||||
|
||||
#include "block-group.h"
|
||||
|
||||
/* TODO XXX FIXME
|
||||
* - subvol delete -> delete when ref goes to 0? delete limits also?
|
||||
|
@ -1312,8 +1312,9 @@ static int __del_qgroup_relation(struct btrfs_trans_handle *trans, u64 src,
|
|||
struct btrfs_qgroup *member;
|
||||
struct btrfs_qgroup_list *list;
|
||||
struct ulist *tmp;
|
||||
bool found = false;
|
||||
int ret = 0;
|
||||
int err;
|
||||
int ret2;
|
||||
|
||||
tmp = ulist_alloc(GFP_KERNEL);
|
||||
if (!tmp)
|
||||
|
@ -1327,28 +1328,39 @@ static int __del_qgroup_relation(struct btrfs_trans_handle *trans, u64 src,
|
|||
|
||||
member = find_qgroup_rb(fs_info, src);
|
||||
parent = find_qgroup_rb(fs_info, dst);
|
||||
if (!member || !parent) {
|
||||
ret = -EINVAL;
|
||||
goto out;
|
||||
}
|
||||
/*
|
||||
* The parent/member pair doesn't exist, then try to delete the dead
|
||||
* relation items only.
|
||||
*/
|
||||
if (!member || !parent)
|
||||
goto delete_item;
|
||||
|
||||
/* check if such qgroup relation exist firstly */
|
||||
list_for_each_entry(list, &member->groups, next_group) {
|
||||
if (list->group == parent)
|
||||
goto exist;
|
||||
if (list->group == parent) {
|
||||
found = true;
|
||||
break;
|
||||
}
|
||||
}
|
||||
ret = -ENOENT;
|
||||
goto out;
|
||||
exist:
|
||||
ret = del_qgroup_relation_item(trans, src, dst);
|
||||
err = del_qgroup_relation_item(trans, dst, src);
|
||||
if (err && !ret)
|
||||
ret = err;
|
||||
|
||||
spin_lock(&fs_info->qgroup_lock);
|
||||
del_relation_rb(fs_info, src, dst);
|
||||
ret = quick_update_accounting(fs_info, tmp, src, dst, -1);
|
||||
spin_unlock(&fs_info->qgroup_lock);
|
||||
delete_item:
|
||||
ret = del_qgroup_relation_item(trans, src, dst);
|
||||
if (ret < 0 && ret != -ENOENT)
|
||||
goto out;
|
||||
ret2 = del_qgroup_relation_item(trans, dst, src);
|
||||
if (ret2 < 0 && ret2 != -ENOENT)
|
||||
goto out;
|
||||
|
||||
/* At least one deletion succeeded, return 0 */
|
||||
if (!ret || !ret2)
|
||||
ret = 0;
|
||||
|
||||
if (found) {
|
||||
spin_lock(&fs_info->qgroup_lock);
|
||||
del_relation_rb(fs_info, src, dst);
|
||||
ret = quick_update_accounting(fs_info, tmp, src, dst, -1);
|
||||
spin_unlock(&fs_info->qgroup_lock);
|
||||
}
|
||||
out:
|
||||
ulist_free(tmp);
|
||||
return ret;
|
||||
|
|
|
@ -35,6 +35,22 @@
|
|||
|
||||
#define RBIO_CACHE_SIZE 1024
|
||||
|
||||
#define BTRFS_STRIPE_HASH_TABLE_BITS 11
|
||||
|
||||
/* Used by the raid56 code to lock stripes for read/modify/write */
|
||||
struct btrfs_stripe_hash {
|
||||
struct list_head hash_list;
|
||||
spinlock_t lock;
|
||||
};
|
||||
|
||||
/* Used by the raid56 code to lock stripes for read/modify/write */
|
||||
struct btrfs_stripe_hash_table {
|
||||
struct list_head stripe_cache;
|
||||
spinlock_t cache_lock;
|
||||
int cache_size;
|
||||
struct btrfs_stripe_hash table[];
|
||||
};
|
||||
|
||||
enum btrfs_rbio_ops {
|
||||
BTRFS_RBIO_WRITE,
|
||||
BTRFS_RBIO_READ_REBUILD,
|
||||
|
|
|
@ -14,6 +14,7 @@
|
|||
#include "disk-io.h"
|
||||
#include "transaction.h"
|
||||
#include "dev-replace.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#undef DEBUG
|
||||
|
||||
|
@ -638,6 +639,35 @@ static int reada_pick_zone(struct btrfs_device *dev)
|
|||
return 1;
|
||||
}
|
||||
|
||||
static int reada_tree_block_flagged(struct btrfs_fs_info *fs_info, u64 bytenr,
|
||||
int mirror_num, struct extent_buffer **eb)
|
||||
{
|
||||
struct extent_buffer *buf = NULL;
|
||||
int ret;
|
||||
|
||||
buf = btrfs_find_create_tree_block(fs_info, bytenr);
|
||||
if (IS_ERR(buf))
|
||||
return 0;
|
||||
|
||||
set_bit(EXTENT_BUFFER_READAHEAD, &buf->bflags);
|
||||
|
||||
ret = read_extent_buffer_pages(buf, WAIT_PAGE_LOCK, mirror_num);
|
||||
if (ret) {
|
||||
free_extent_buffer_stale(buf);
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (test_bit(EXTENT_BUFFER_CORRUPT, &buf->bflags)) {
|
||||
free_extent_buffer_stale(buf);
|
||||
return -EIO;
|
||||
} else if (extent_buffer_uptodate(buf)) {
|
||||
*eb = buf;
|
||||
} else {
|
||||
free_extent_buffer(buf);
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int reada_start_machine_dev(struct btrfs_device *dev)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = dev->fs_info;
|
||||
|
|
|
@ -21,6 +21,7 @@
|
|||
#include "qgroup.h"
|
||||
#include "print-tree.h"
|
||||
#include "delalloc-space.h"
|
||||
#include "block-group.h"
|
||||
|
||||
/*
|
||||
* backref_node, mapping_node and tree_block start with this
|
||||
|
@ -3311,7 +3312,7 @@ static int relocate_file_extent_cluster(struct inode *inode,
|
|||
}
|
||||
|
||||
ret = btrfs_set_extent_delalloc(inode, page_start, page_end, 0,
|
||||
NULL, 0);
|
||||
NULL);
|
||||
if (ret) {
|
||||
unlock_page(page);
|
||||
put_page(page);
|
||||
|
|
|
@ -533,7 +533,7 @@ int btrfs_subvolume_reserve_metadata(struct btrfs_root *root,
|
|||
return ret;
|
||||
}
|
||||
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, items);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info, items);
|
||||
rsv->space_info = btrfs_find_space_info(fs_info,
|
||||
BTRFS_BLOCK_GROUP_METADATA);
|
||||
ret = btrfs_block_rsv_add(root, rsv, num_bytes,
|
||||
|
|
|
@ -18,6 +18,7 @@
|
|||
#include "check-integrity.h"
|
||||
#include "rcu-string.h"
|
||||
#include "raid56.h"
|
||||
#include "block-group.h"
|
||||
|
||||
/*
|
||||
* This is only the first step towards a full-features scrub. It reads all
|
||||
|
|
375
fs/btrfs/send.c
375
fs/btrfs/send.c
|
@ -260,6 +260,21 @@ struct name_cache_entry {
|
|||
char name[];
|
||||
};
|
||||
|
||||
#define ADVANCE 1
|
||||
#define ADVANCE_ONLY_NEXT -1
|
||||
|
||||
enum btrfs_compare_tree_result {
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
BTRFS_COMPARE_TREE_CHANGED,
|
||||
BTRFS_COMPARE_TREE_SAME,
|
||||
};
|
||||
typedef int (*btrfs_changed_cb_t)(struct btrfs_path *left_path,
|
||||
struct btrfs_path *right_path,
|
||||
struct btrfs_key *key,
|
||||
enum btrfs_compare_tree_result result,
|
||||
void *ctx);
|
||||
|
||||
__cold
|
||||
static void inconsistent_snapshot_error(struct send_ctx *sctx,
|
||||
enum btrfs_compare_tree_result result,
|
||||
|
@ -6514,6 +6529,366 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static int tree_move_down(struct btrfs_path *path, int *level)
|
||||
{
|
||||
struct extent_buffer *eb;
|
||||
|
||||
BUG_ON(*level == 0);
|
||||
eb = btrfs_read_node_slot(path->nodes[*level], path->slots[*level]);
|
||||
if (IS_ERR(eb))
|
||||
return PTR_ERR(eb);
|
||||
|
||||
path->nodes[*level - 1] = eb;
|
||||
path->slots[*level - 1] = 0;
|
||||
(*level)--;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tree_move_next_or_upnext(struct btrfs_path *path,
|
||||
int *level, int root_level)
|
||||
{
|
||||
int ret = 0;
|
||||
int nritems;
|
||||
nritems = btrfs_header_nritems(path->nodes[*level]);
|
||||
|
||||
path->slots[*level]++;
|
||||
|
||||
while (path->slots[*level] >= nritems) {
|
||||
if (*level == root_level)
|
||||
return -1;
|
||||
|
||||
/* move upnext */
|
||||
path->slots[*level] = 0;
|
||||
free_extent_buffer(path->nodes[*level]);
|
||||
path->nodes[*level] = NULL;
|
||||
(*level)++;
|
||||
path->slots[*level]++;
|
||||
|
||||
nritems = btrfs_header_nritems(path->nodes[*level]);
|
||||
ret = 1;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Returns 1 if it had to move up and next. 0 is returned if it moved only next
|
||||
* or down.
|
||||
*/
|
||||
static int tree_advance(struct btrfs_path *path,
|
||||
int *level, int root_level,
|
||||
int allow_down,
|
||||
struct btrfs_key *key)
|
||||
{
|
||||
int ret;
|
||||
|
||||
if (*level == 0 || !allow_down) {
|
||||
ret = tree_move_next_or_upnext(path, level, root_level);
|
||||
} else {
|
||||
ret = tree_move_down(path, level);
|
||||
}
|
||||
if (ret >= 0) {
|
||||
if (*level == 0)
|
||||
btrfs_item_key_to_cpu(path->nodes[*level], key,
|
||||
path->slots[*level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(path->nodes[*level], key,
|
||||
path->slots[*level]);
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int tree_compare_item(struct btrfs_path *left_path,
|
||||
struct btrfs_path *right_path,
|
||||
char *tmp_buf)
|
||||
{
|
||||
int cmp;
|
||||
int len1, len2;
|
||||
unsigned long off1, off2;
|
||||
|
||||
len1 = btrfs_item_size_nr(left_path->nodes[0], left_path->slots[0]);
|
||||
len2 = btrfs_item_size_nr(right_path->nodes[0], right_path->slots[0]);
|
||||
if (len1 != len2)
|
||||
return 1;
|
||||
|
||||
off1 = btrfs_item_ptr_offset(left_path->nodes[0], left_path->slots[0]);
|
||||
off2 = btrfs_item_ptr_offset(right_path->nodes[0],
|
||||
right_path->slots[0]);
|
||||
|
||||
read_extent_buffer(left_path->nodes[0], tmp_buf, off1, len1);
|
||||
|
||||
cmp = memcmp_extent_buffer(right_path->nodes[0], tmp_buf, off2, len1);
|
||||
if (cmp)
|
||||
return 1;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* This function compares two trees and calls the provided callback for
|
||||
* every changed/new/deleted item it finds.
|
||||
* If shared tree blocks are encountered, whole subtrees are skipped, making
|
||||
* the compare pretty fast on snapshotted subvolumes.
|
||||
*
|
||||
* This currently works on commit roots only. As commit roots are read only,
|
||||
* we don't do any locking. The commit roots are protected with transactions.
|
||||
* Transactions are ended and rejoined when a commit is tried in between.
|
||||
*
|
||||
* This function checks for modifications done to the trees while comparing.
|
||||
* If it detects a change, it aborts immediately.
|
||||
*/
|
||||
static int btrfs_compare_trees(struct btrfs_root *left_root,
|
||||
struct btrfs_root *right_root,
|
||||
btrfs_changed_cb_t changed_cb, void *ctx)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = left_root->fs_info;
|
||||
int ret;
|
||||
int cmp;
|
||||
struct btrfs_path *left_path = NULL;
|
||||
struct btrfs_path *right_path = NULL;
|
||||
struct btrfs_key left_key;
|
||||
struct btrfs_key right_key;
|
||||
char *tmp_buf = NULL;
|
||||
int left_root_level;
|
||||
int right_root_level;
|
||||
int left_level;
|
||||
int right_level;
|
||||
int left_end_reached;
|
||||
int right_end_reached;
|
||||
int advance_left;
|
||||
int advance_right;
|
||||
u64 left_blockptr;
|
||||
u64 right_blockptr;
|
||||
u64 left_gen;
|
||||
u64 right_gen;
|
||||
|
||||
left_path = btrfs_alloc_path();
|
||||
if (!left_path) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
right_path = btrfs_alloc_path();
|
||||
if (!right_path) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
tmp_buf = kvmalloc(fs_info->nodesize, GFP_KERNEL);
|
||||
if (!tmp_buf) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
left_path->search_commit_root = 1;
|
||||
left_path->skip_locking = 1;
|
||||
right_path->search_commit_root = 1;
|
||||
right_path->skip_locking = 1;
|
||||
|
||||
/*
|
||||
* Strategy: Go to the first items of both trees. Then do
|
||||
*
|
||||
* If both trees are at level 0
|
||||
* Compare keys of current items
|
||||
* If left < right treat left item as new, advance left tree
|
||||
* and repeat
|
||||
* If left > right treat right item as deleted, advance right tree
|
||||
* and repeat
|
||||
* If left == right do deep compare of items, treat as changed if
|
||||
* needed, advance both trees and repeat
|
||||
* If both trees are at the same level but not at level 0
|
||||
* Compare keys of current nodes/leafs
|
||||
* If left < right advance left tree and repeat
|
||||
* If left > right advance right tree and repeat
|
||||
* If left == right compare blockptrs of the next nodes/leafs
|
||||
* If they match advance both trees but stay at the same level
|
||||
* and repeat
|
||||
* If they don't match advance both trees while allowing to go
|
||||
* deeper and repeat
|
||||
* If tree levels are different
|
||||
* Advance the tree that needs it and repeat
|
||||
*
|
||||
* Advancing a tree means:
|
||||
* If we are at level 0, try to go to the next slot. If that's not
|
||||
* possible, go one level up and repeat. Stop when we found a level
|
||||
* where we could go to the next slot. We may at this point be on a
|
||||
* node or a leaf.
|
||||
*
|
||||
* If we are not at level 0 and not on shared tree blocks, go one
|
||||
* level deeper.
|
||||
*
|
||||
* If we are not at level 0 and on shared tree blocks, go one slot to
|
||||
* the right if possible or go up and right.
|
||||
*/
|
||||
|
||||
down_read(&fs_info->commit_root_sem);
|
||||
left_level = btrfs_header_level(left_root->commit_root);
|
||||
left_root_level = left_level;
|
||||
left_path->nodes[left_level] =
|
||||
btrfs_clone_extent_buffer(left_root->commit_root);
|
||||
if (!left_path->nodes[left_level]) {
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
right_level = btrfs_header_level(right_root->commit_root);
|
||||
right_root_level = right_level;
|
||||
right_path->nodes[right_level] =
|
||||
btrfs_clone_extent_buffer(right_root->commit_root);
|
||||
if (!right_path->nodes[right_level]) {
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
up_read(&fs_info->commit_root_sem);
|
||||
|
||||
if (left_level == 0)
|
||||
btrfs_item_key_to_cpu(left_path->nodes[left_level],
|
||||
&left_key, left_path->slots[left_level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(left_path->nodes[left_level],
|
||||
&left_key, left_path->slots[left_level]);
|
||||
if (right_level == 0)
|
||||
btrfs_item_key_to_cpu(right_path->nodes[right_level],
|
||||
&right_key, right_path->slots[right_level]);
|
||||
else
|
||||
btrfs_node_key_to_cpu(right_path->nodes[right_level],
|
||||
&right_key, right_path->slots[right_level]);
|
||||
|
||||
left_end_reached = right_end_reached = 0;
|
||||
advance_left = advance_right = 0;
|
||||
|
||||
while (1) {
|
||||
cond_resched();
|
||||
if (advance_left && !left_end_reached) {
|
||||
ret = tree_advance(left_path, &left_level,
|
||||
left_root_level,
|
||||
advance_left != ADVANCE_ONLY_NEXT,
|
||||
&left_key);
|
||||
if (ret == -1)
|
||||
left_end_reached = ADVANCE;
|
||||
else if (ret < 0)
|
||||
goto out;
|
||||
advance_left = 0;
|
||||
}
|
||||
if (advance_right && !right_end_reached) {
|
||||
ret = tree_advance(right_path, &right_level,
|
||||
right_root_level,
|
||||
advance_right != ADVANCE_ONLY_NEXT,
|
||||
&right_key);
|
||||
if (ret == -1)
|
||||
right_end_reached = ADVANCE;
|
||||
else if (ret < 0)
|
||||
goto out;
|
||||
advance_right = 0;
|
||||
}
|
||||
|
||||
if (left_end_reached && right_end_reached) {
|
||||
ret = 0;
|
||||
goto out;
|
||||
} else if (left_end_reached) {
|
||||
if (right_level == 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&right_key,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
}
|
||||
advance_right = ADVANCE;
|
||||
continue;
|
||||
} else if (right_end_reached) {
|
||||
if (left_level == 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key,
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
}
|
||||
advance_left = ADVANCE;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (left_level == 0 && right_level == 0) {
|
||||
cmp = btrfs_comp_cpu_keys(&left_key, &right_key);
|
||||
if (cmp < 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key,
|
||||
BTRFS_COMPARE_TREE_NEW,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_left = ADVANCE;
|
||||
} else if (cmp > 0) {
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&right_key,
|
||||
BTRFS_COMPARE_TREE_DELETED,
|
||||
ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
enum btrfs_compare_tree_result result;
|
||||
|
||||
WARN_ON(!extent_buffer_uptodate(left_path->nodes[0]));
|
||||
ret = tree_compare_item(left_path, right_path,
|
||||
tmp_buf);
|
||||
if (ret)
|
||||
result = BTRFS_COMPARE_TREE_CHANGED;
|
||||
else
|
||||
result = BTRFS_COMPARE_TREE_SAME;
|
||||
ret = changed_cb(left_path, right_path,
|
||||
&left_key, result, ctx);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
advance_left = ADVANCE;
|
||||
advance_right = ADVANCE;
|
||||
}
|
||||
} else if (left_level == right_level) {
|
||||
cmp = btrfs_comp_cpu_keys(&left_key, &right_key);
|
||||
if (cmp < 0) {
|
||||
advance_left = ADVANCE;
|
||||
} else if (cmp > 0) {
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
left_blockptr = btrfs_node_blockptr(
|
||||
left_path->nodes[left_level],
|
||||
left_path->slots[left_level]);
|
||||
right_blockptr = btrfs_node_blockptr(
|
||||
right_path->nodes[right_level],
|
||||
right_path->slots[right_level]);
|
||||
left_gen = btrfs_node_ptr_generation(
|
||||
left_path->nodes[left_level],
|
||||
left_path->slots[left_level]);
|
||||
right_gen = btrfs_node_ptr_generation(
|
||||
right_path->nodes[right_level],
|
||||
right_path->slots[right_level]);
|
||||
if (left_blockptr == right_blockptr &&
|
||||
left_gen == right_gen) {
|
||||
/*
|
||||
* As we're on a shared block, don't
|
||||
* allow to go deeper.
|
||||
*/
|
||||
advance_left = ADVANCE_ONLY_NEXT;
|
||||
advance_right = ADVANCE_ONLY_NEXT;
|
||||
} else {
|
||||
advance_left = ADVANCE;
|
||||
advance_right = ADVANCE;
|
||||
}
|
||||
}
|
||||
} else if (left_level < right_level) {
|
||||
advance_right = ADVANCE;
|
||||
} else {
|
||||
advance_left = ADVANCE;
|
||||
}
|
||||
}
|
||||
|
||||
out:
|
||||
btrfs_free_path(left_path);
|
||||
btrfs_free_path(right_path);
|
||||
kvfree(tmp_buf);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int send_subvol(struct send_ctx *sctx)
|
||||
{
|
||||
int ret;
|
||||
|
|
|
@ -1,5 +1,6 @@
|
|||
// SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "space-info.h"
|
||||
#include "sysfs.h"
|
||||
|
@ -7,7 +8,7 @@
|
|||
#include "free-space-cache.h"
|
||||
#include "ordered-data.h"
|
||||
#include "transaction.h"
|
||||
#include "math.h"
|
||||
#include "block-group.h"
|
||||
|
||||
u64 btrfs_space_info_used(struct btrfs_space_info *s_info,
|
||||
bool may_use_included)
|
||||
|
@ -33,23 +34,6 @@ void btrfs_clear_space_info_full(struct btrfs_fs_info *info)
|
|||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
static const char *alloc_name(u64 flags)
|
||||
{
|
||||
switch (flags) {
|
||||
case BTRFS_BLOCK_GROUP_METADATA|BTRFS_BLOCK_GROUP_DATA:
|
||||
return "mixed";
|
||||
case BTRFS_BLOCK_GROUP_METADATA:
|
||||
return "metadata";
|
||||
case BTRFS_BLOCK_GROUP_DATA:
|
||||
return "data";
|
||||
case BTRFS_BLOCK_GROUP_SYSTEM:
|
||||
return "system";
|
||||
default:
|
||||
WARN_ON(1);
|
||||
return "invalid-combination";
|
||||
};
|
||||
}
|
||||
|
||||
static int create_space_info(struct btrfs_fs_info *info, u64 flags)
|
||||
{
|
||||
|
||||
|
@ -79,13 +63,9 @@ static int create_space_info(struct btrfs_fs_info *info, u64 flags)
|
|||
INIT_LIST_HEAD(&space_info->tickets);
|
||||
INIT_LIST_HEAD(&space_info->priority_tickets);
|
||||
|
||||
ret = kobject_init_and_add(&space_info->kobj, &space_info_ktype,
|
||||
info->space_info_kobj, "%s",
|
||||
alloc_name(space_info->flags));
|
||||
if (ret) {
|
||||
kobject_put(&space_info->kobj);
|
||||
ret = btrfs_sysfs_add_space_info_type(info, space_info);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
list_add_rcu(&space_info->list, &info->space_info);
|
||||
if (flags & BTRFS_BLOCK_GROUP_DATA)
|
||||
|
@ -151,9 +131,7 @@ void btrfs_update_space_info(struct btrfs_fs_info *info, u64 flags,
|
|||
found->bytes_readonly += bytes_readonly;
|
||||
if (total_bytes > 0)
|
||||
found->full = 0;
|
||||
btrfs_space_info_add_new_bytes(info, found,
|
||||
total_bytes - bytes_used -
|
||||
bytes_readonly);
|
||||
btrfs_try_granting_tickets(info, found);
|
||||
spin_unlock(&found->lock);
|
||||
*space_info = found;
|
||||
}
|
||||
|
@ -187,9 +165,7 @@ static int can_overcommit(struct btrfs_fs_info *fs_info,
|
|||
enum btrfs_reserve_flush_enum flush,
|
||||
bool system_chunk)
|
||||
{
|
||||
struct btrfs_block_rsv *global_rsv = &fs_info->global_block_rsv;
|
||||
u64 profile;
|
||||
u64 space_size;
|
||||
u64 avail;
|
||||
u64 used;
|
||||
int factor;
|
||||
|
@ -203,22 +179,7 @@ static int can_overcommit(struct btrfs_fs_info *fs_info,
|
|||
else
|
||||
profile = btrfs_metadata_alloc_profile(fs_info);
|
||||
|
||||
used = btrfs_space_info_used(space_info, false);
|
||||
|
||||
/*
|
||||
* We only want to allow over committing if we have lots of actual space
|
||||
* free, but if we don't have enough space to handle the global reserve
|
||||
* space then we could end up having a real enospc problem when trying
|
||||
* to allocate a chunk or some other such important allocation.
|
||||
*/
|
||||
spin_lock(&global_rsv->lock);
|
||||
space_size = calc_global_rsv_need_space(global_rsv);
|
||||
spin_unlock(&global_rsv->lock);
|
||||
if (used + space_size >= space_info->total_bytes)
|
||||
return 0;
|
||||
|
||||
used += space_info->bytes_may_use;
|
||||
|
||||
used = btrfs_space_info_used(space_info, true);
|
||||
avail = atomic64_read(&fs_info->free_chunk_space);
|
||||
|
||||
/*
|
||||
|
@ -249,103 +210,41 @@ static int can_overcommit(struct btrfs_fs_info *fs_info,
|
|||
* This is for space we already have accounted in space_info->bytes_may_use, so
|
||||
* basically when we're returning space from block_rsv's.
|
||||
*/
|
||||
void btrfs_space_info_add_old_bytes(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
u64 num_bytes)
|
||||
void btrfs_try_granting_tickets(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info)
|
||||
{
|
||||
struct reserve_ticket *ticket;
|
||||
struct list_head *head;
|
||||
u64 used;
|
||||
enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_NO_FLUSH;
|
||||
bool check_overcommit = false;
|
||||
|
||||
spin_lock(&space_info->lock);
|
||||
lockdep_assert_held(&space_info->lock);
|
||||
|
||||
head = &space_info->priority_tickets;
|
||||
|
||||
/*
|
||||
* If we are over our limit then we need to check and see if we can
|
||||
* overcommit, and if we can't then we just need to free up our space
|
||||
* and not satisfy any requests.
|
||||
*/
|
||||
used = btrfs_space_info_used(space_info, true);
|
||||
if (used - num_bytes >= space_info->total_bytes)
|
||||
check_overcommit = true;
|
||||
again:
|
||||
while (!list_empty(head) && num_bytes) {
|
||||
ticket = list_first_entry(head, struct reserve_ticket,
|
||||
list);
|
||||
/*
|
||||
* We use 0 bytes because this space is already reserved, so
|
||||
* adding the ticket space would be a double count.
|
||||
*/
|
||||
if (check_overcommit &&
|
||||
!can_overcommit(fs_info, space_info, 0, flush, false))
|
||||
break;
|
||||
if (num_bytes >= ticket->bytes) {
|
||||
list_del_init(&ticket->list);
|
||||
num_bytes -= ticket->bytes;
|
||||
ticket->bytes = 0;
|
||||
space_info->tickets_id++;
|
||||
wake_up(&ticket->wait);
|
||||
} else {
|
||||
ticket->bytes -= num_bytes;
|
||||
num_bytes = 0;
|
||||
}
|
||||
}
|
||||
while (!list_empty(head)) {
|
||||
struct reserve_ticket *ticket;
|
||||
u64 used = btrfs_space_info_used(space_info, true);
|
||||
|
||||
if (num_bytes && head == &space_info->priority_tickets) {
|
||||
head = &space_info->tickets;
|
||||
flush = BTRFS_RESERVE_FLUSH_ALL;
|
||||
goto again;
|
||||
}
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, space_info, -num_bytes);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
space_info->flags, num_bytes, 0);
|
||||
spin_unlock(&space_info->lock);
|
||||
}
|
||||
ticket = list_first_entry(head, struct reserve_ticket, list);
|
||||
|
||||
/*
|
||||
* This is for newly allocated space that isn't accounted in
|
||||
* space_info->bytes_may_use yet. So if we allocate a chunk or unpin an extent
|
||||
* we use this helper.
|
||||
*/
|
||||
void btrfs_space_info_add_new_bytes(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
u64 num_bytes)
|
||||
{
|
||||
struct reserve_ticket *ticket;
|
||||
struct list_head *head = &space_info->priority_tickets;
|
||||
|
||||
again:
|
||||
while (!list_empty(head) && num_bytes) {
|
||||
ticket = list_first_entry(head, struct reserve_ticket,
|
||||
list);
|
||||
if (num_bytes >= ticket->bytes) {
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
space_info->flags,
|
||||
ticket->bytes, 1);
|
||||
list_del_init(&ticket->list);
|
||||
num_bytes -= ticket->bytes;
|
||||
/* Check and see if our ticket can be satisified now. */
|
||||
if ((used + ticket->bytes <= space_info->total_bytes) ||
|
||||
can_overcommit(fs_info, space_info, ticket->bytes, flush,
|
||||
false)) {
|
||||
btrfs_space_info_update_bytes_may_use(fs_info,
|
||||
space_info,
|
||||
ticket->bytes);
|
||||
list_del_init(&ticket->list);
|
||||
ticket->bytes = 0;
|
||||
space_info->tickets_id++;
|
||||
wake_up(&ticket->wait);
|
||||
} else {
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
space_info->flags,
|
||||
num_bytes, 1);
|
||||
btrfs_space_info_update_bytes_may_use(fs_info,
|
||||
space_info,
|
||||
num_bytes);
|
||||
ticket->bytes -= num_bytes;
|
||||
num_bytes = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (num_bytes && head == &space_info->priority_tickets) {
|
||||
if (head == &space_info->priority_tickets) {
|
||||
head = &space_info->tickets;
|
||||
flush = BTRFS_RESERVE_FLUSH_ALL;
|
||||
goto again;
|
||||
}
|
||||
}
|
||||
|
@ -359,14 +258,11 @@ do { \
|
|||
spin_unlock(&__rsv->lock); \
|
||||
} while (0)
|
||||
|
||||
void btrfs_dump_space_info(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *info, u64 bytes,
|
||||
int dump_block_groups)
|
||||
static void __btrfs_dump_space_info(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *info)
|
||||
{
|
||||
struct btrfs_block_group_cache *cache;
|
||||
int index = 0;
|
||||
lockdep_assert_held(&info->lock);
|
||||
|
||||
spin_lock(&info->lock);
|
||||
btrfs_info(fs_info, "space_info %llu has %llu free, is %sfull",
|
||||
info->flags,
|
||||
info->total_bytes - btrfs_space_info_used(info, true),
|
||||
|
@ -376,7 +272,6 @@ void btrfs_dump_space_info(struct btrfs_fs_info *fs_info,
|
|||
info->total_bytes, info->bytes_used, info->bytes_pinned,
|
||||
info->bytes_reserved, info->bytes_may_use,
|
||||
info->bytes_readonly);
|
||||
spin_unlock(&info->lock);
|
||||
|
||||
DUMP_BLOCK_RSV(fs_info, global_block_rsv);
|
||||
DUMP_BLOCK_RSV(fs_info, trans_block_rsv);
|
||||
|
@ -384,6 +279,19 @@ void btrfs_dump_space_info(struct btrfs_fs_info *fs_info,
|
|||
DUMP_BLOCK_RSV(fs_info, delayed_block_rsv);
|
||||
DUMP_BLOCK_RSV(fs_info, delayed_refs_rsv);
|
||||
|
||||
}
|
||||
|
||||
void btrfs_dump_space_info(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *info, u64 bytes,
|
||||
int dump_block_groups)
|
||||
{
|
||||
struct btrfs_block_group_cache *cache;
|
||||
int index = 0;
|
||||
|
||||
spin_lock(&info->lock);
|
||||
__btrfs_dump_space_info(fs_info, info);
|
||||
spin_unlock(&info->lock);
|
||||
|
||||
if (!dump_block_groups)
|
||||
return;
|
||||
|
||||
|
@ -432,7 +340,7 @@ static inline u64 calc_reclaim_items_nr(struct btrfs_fs_info *fs_info,
|
|||
u64 bytes;
|
||||
u64 nr;
|
||||
|
||||
bytes = btrfs_calc_trans_metadata_size(fs_info, 1);
|
||||
bytes = btrfs_calc_insert_metadata_size(fs_info, 1);
|
||||
nr = div64_u64(to_reclaim, bytes);
|
||||
if (!nr)
|
||||
nr = 1;
|
||||
|
@ -557,12 +465,19 @@ static int may_commit_transaction(struct btrfs_fs_info *fs_info,
|
|||
struct btrfs_trans_handle *trans;
|
||||
u64 bytes_needed;
|
||||
u64 reclaim_bytes = 0;
|
||||
u64 cur_free_bytes = 0;
|
||||
|
||||
trans = (struct btrfs_trans_handle *)current->journal_info;
|
||||
if (trans)
|
||||
return -EAGAIN;
|
||||
|
||||
spin_lock(&space_info->lock);
|
||||
cur_free_bytes = btrfs_space_info_used(space_info, true);
|
||||
if (cur_free_bytes < space_info->total_bytes)
|
||||
cur_free_bytes = space_info->total_bytes - cur_free_bytes;
|
||||
else
|
||||
cur_free_bytes = 0;
|
||||
|
||||
if (!list_empty(&space_info->priority_tickets))
|
||||
ticket = list_first_entry(&space_info->priority_tickets,
|
||||
struct reserve_ticket, list);
|
||||
|
@ -570,6 +485,11 @@ static int may_commit_transaction(struct btrfs_fs_info *fs_info,
|
|||
ticket = list_first_entry(&space_info->tickets,
|
||||
struct reserve_ticket, list);
|
||||
bytes_needed = (ticket) ? ticket->bytes : 0;
|
||||
|
||||
if (bytes_needed > cur_free_bytes)
|
||||
bytes_needed -= cur_free_bytes;
|
||||
else
|
||||
bytes_needed = 0;
|
||||
spin_unlock(&space_info->lock);
|
||||
|
||||
if (!bytes_needed)
|
||||
|
@ -684,7 +604,7 @@ static void flush_space(struct btrfs_fs_info *fs_info,
|
|||
if (ret > 0 || ret == -ENOSPC)
|
||||
ret = 0;
|
||||
break;
|
||||
case COMMIT_TRANS:
|
||||
case RUN_DELAYED_IPUTS:
|
||||
/*
|
||||
* If we have pending delayed iputs then we could free up a
|
||||
* bunch of pinned space, so make sure we run the iputs before
|
||||
|
@ -692,7 +612,8 @@ static void flush_space(struct btrfs_fs_info *fs_info,
|
|||
*/
|
||||
btrfs_run_delayed_iputs(fs_info);
|
||||
btrfs_wait_on_delayed_iputs(fs_info);
|
||||
|
||||
break;
|
||||
case COMMIT_TRANS:
|
||||
ret = may_commit_transaction(fs_info, space_info);
|
||||
break;
|
||||
default:
|
||||
|
@ -762,19 +683,70 @@ static inline int need_do_async_reclaim(struct btrfs_fs_info *fs_info,
|
|||
!test_bit(BTRFS_FS_STATE_REMOUNTING, &fs_info->fs_state));
|
||||
}
|
||||
|
||||
static bool wake_all_tickets(struct list_head *head)
|
||||
/*
|
||||
* maybe_fail_all_tickets - we've exhausted our flushing, start failing tickets
|
||||
* @fs_info - fs_info for this fs
|
||||
* @space_info - the space info we were flushing
|
||||
*
|
||||
* We call this when we've exhausted our flushing ability and haven't made
|
||||
* progress in satisfying tickets. The reservation code handles tickets in
|
||||
* order, so if there is a large ticket first and then smaller ones we could
|
||||
* very well satisfy the smaller tickets. This will attempt to wake up any
|
||||
* tickets in the list to catch this case.
|
||||
*
|
||||
* This function returns true if it was able to make progress by clearing out
|
||||
* other tickets, or if it stumbles across a ticket that was smaller than the
|
||||
* first ticket.
|
||||
*/
|
||||
static bool maybe_fail_all_tickets(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info)
|
||||
{
|
||||
struct reserve_ticket *ticket;
|
||||
u64 tickets_id = space_info->tickets_id;
|
||||
u64 first_ticket_bytes = 0;
|
||||
|
||||
if (btrfs_test_opt(fs_info, ENOSPC_DEBUG)) {
|
||||
btrfs_info(fs_info, "cannot satisfy tickets, dumping space info");
|
||||
__btrfs_dump_space_info(fs_info, space_info);
|
||||
}
|
||||
|
||||
while (!list_empty(&space_info->tickets) &&
|
||||
tickets_id == space_info->tickets_id) {
|
||||
ticket = list_first_entry(&space_info->tickets,
|
||||
struct reserve_ticket, list);
|
||||
|
||||
/*
|
||||
* may_commit_transaction will avoid committing the transaction
|
||||
* if it doesn't feel like the space reclaimed by the commit
|
||||
* would result in the ticket succeeding. However if we have a
|
||||
* smaller ticket in the queue it may be small enough to be
|
||||
* satisified by committing the transaction, so if any
|
||||
* subsequent ticket is smaller than the first ticket go ahead
|
||||
* and send us back for another loop through the enospc flushing
|
||||
* code.
|
||||
*/
|
||||
if (first_ticket_bytes == 0)
|
||||
first_ticket_bytes = ticket->bytes;
|
||||
else if (first_ticket_bytes > ticket->bytes)
|
||||
return true;
|
||||
|
||||
if (btrfs_test_opt(fs_info, ENOSPC_DEBUG))
|
||||
btrfs_info(fs_info, "failing ticket with %llu bytes",
|
||||
ticket->bytes);
|
||||
|
||||
while (!list_empty(head)) {
|
||||
ticket = list_first_entry(head, struct reserve_ticket, list);
|
||||
list_del_init(&ticket->list);
|
||||
ticket->error = -ENOSPC;
|
||||
wake_up(&ticket->wait);
|
||||
if (ticket->bytes != ticket->orig_bytes)
|
||||
return true;
|
||||
|
||||
/*
|
||||
* We're just throwing tickets away, so more flushing may not
|
||||
* trip over btrfs_try_granting_tickets, so we need to call it
|
||||
* here to see if we can make progress with the next ticket in
|
||||
* the list.
|
||||
*/
|
||||
btrfs_try_granting_tickets(fs_info, space_info);
|
||||
}
|
||||
return false;
|
||||
return (tickets_id != space_info->tickets_id);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -842,7 +814,7 @@ static void btrfs_async_reclaim_metadata_space(struct work_struct *work)
|
|||
if (flush_state > COMMIT_TRANS) {
|
||||
commit_cycles++;
|
||||
if (commit_cycles > 2) {
|
||||
if (wake_all_tickets(&space_info->tickets)) {
|
||||
if (maybe_fail_all_tickets(fs_info, space_info)) {
|
||||
flush_state = FLUSH_DELAYED_ITEMS_NR;
|
||||
commit_cycles--;
|
||||
} else {
|
||||
|
@ -867,9 +839,22 @@ static const enum btrfs_flush_state priority_flush_states[] = {
|
|||
ALLOC_CHUNK,
|
||||
};
|
||||
|
||||
static const enum btrfs_flush_state evict_flush_states[] = {
|
||||
FLUSH_DELAYED_ITEMS_NR,
|
||||
FLUSH_DELAYED_ITEMS,
|
||||
FLUSH_DELAYED_REFS_NR,
|
||||
FLUSH_DELAYED_REFS,
|
||||
FLUSH_DELALLOC,
|
||||
FLUSH_DELALLOC_WAIT,
|
||||
ALLOC_CHUNK,
|
||||
COMMIT_TRANS,
|
||||
};
|
||||
|
||||
static void priority_reclaim_metadata_space(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
struct reserve_ticket *ticket)
|
||||
struct btrfs_space_info *space_info,
|
||||
struct reserve_ticket *ticket,
|
||||
const enum btrfs_flush_state *states,
|
||||
int states_nr)
|
||||
{
|
||||
u64 to_reclaim;
|
||||
int flush_state;
|
||||
|
@ -885,8 +870,7 @@ static void priority_reclaim_metadata_space(struct btrfs_fs_info *fs_info,
|
|||
|
||||
flush_state = 0;
|
||||
do {
|
||||
flush_space(fs_info, space_info, to_reclaim,
|
||||
priority_flush_states[flush_state]);
|
||||
flush_space(fs_info, space_info, to_reclaim, states[flush_state]);
|
||||
flush_state++;
|
||||
spin_lock(&space_info->lock);
|
||||
if (ticket->bytes == 0) {
|
||||
|
@ -894,23 +878,22 @@ static void priority_reclaim_metadata_space(struct btrfs_fs_info *fs_info,
|
|||
return;
|
||||
}
|
||||
spin_unlock(&space_info->lock);
|
||||
} while (flush_state < ARRAY_SIZE(priority_flush_states));
|
||||
} while (flush_state < states_nr);
|
||||
}
|
||||
|
||||
static int wait_reserve_ticket(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
struct reserve_ticket *ticket)
|
||||
static void wait_reserve_ticket(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
struct reserve_ticket *ticket)
|
||||
|
||||
{
|
||||
DEFINE_WAIT(wait);
|
||||
u64 reclaim_bytes = 0;
|
||||
int ret = 0;
|
||||
|
||||
spin_lock(&space_info->lock);
|
||||
while (ticket->bytes > 0 && ticket->error == 0) {
|
||||
ret = prepare_to_wait_event(&ticket->wait, &wait, TASK_KILLABLE);
|
||||
if (ret) {
|
||||
ret = -EINTR;
|
||||
ticket->error = -EINTR;
|
||||
break;
|
||||
}
|
||||
spin_unlock(&space_info->lock);
|
||||
|
@ -920,17 +903,54 @@ static int wait_reserve_ticket(struct btrfs_fs_info *fs_info,
|
|||
finish_wait(&ticket->wait, &wait);
|
||||
spin_lock(&space_info->lock);
|
||||
}
|
||||
if (!ret)
|
||||
ret = ticket->error;
|
||||
if (!list_empty(&ticket->list))
|
||||
list_del_init(&ticket->list);
|
||||
if (ticket->bytes && ticket->bytes < ticket->orig_bytes)
|
||||
reclaim_bytes = ticket->orig_bytes - ticket->bytes;
|
||||
spin_unlock(&space_info->lock);
|
||||
}
|
||||
|
||||
if (reclaim_bytes)
|
||||
btrfs_space_info_add_old_bytes(fs_info, space_info,
|
||||
reclaim_bytes);
|
||||
/**
|
||||
* handle_reserve_ticket - do the appropriate flushing and waiting for a ticket
|
||||
* @fs_info - the fs
|
||||
* @space_info - the space_info for the reservation
|
||||
* @ticket - the ticket for the reservation
|
||||
* @flush - how much we can flush
|
||||
*
|
||||
* This does the work of figuring out how to flush for the ticket, waiting for
|
||||
* the reservation, and returning the appropriate error if there is one.
|
||||
*/
|
||||
static int handle_reserve_ticket(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
struct reserve_ticket *ticket,
|
||||
enum btrfs_reserve_flush_enum flush)
|
||||
{
|
||||
int ret;
|
||||
|
||||
switch (flush) {
|
||||
case BTRFS_RESERVE_FLUSH_ALL:
|
||||
wait_reserve_ticket(fs_info, space_info, ticket);
|
||||
break;
|
||||
case BTRFS_RESERVE_FLUSH_LIMIT:
|
||||
priority_reclaim_metadata_space(fs_info, space_info, ticket,
|
||||
priority_flush_states,
|
||||
ARRAY_SIZE(priority_flush_states));
|
||||
break;
|
||||
case BTRFS_RESERVE_FLUSH_EVICT:
|
||||
priority_reclaim_metadata_space(fs_info, space_info, ticket,
|
||||
evict_flush_states,
|
||||
ARRAY_SIZE(evict_flush_states));
|
||||
break;
|
||||
default:
|
||||
ASSERT(0);
|
||||
break;
|
||||
}
|
||||
|
||||
spin_lock(&space_info->lock);
|
||||
ret = ticket->error;
|
||||
if (ticket->bytes || ticket->error) {
|
||||
list_del_init(&ticket->list);
|
||||
if (!ret)
|
||||
ret = -ENOSPC;
|
||||
}
|
||||
spin_unlock(&space_info->lock);
|
||||
ASSERT(list_empty(&ticket->list));
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -956,8 +976,8 @@ static int __reserve_metadata_bytes(struct btrfs_fs_info *fs_info,
|
|||
{
|
||||
struct reserve_ticket ticket;
|
||||
u64 used;
|
||||
u64 reclaim_bytes = 0;
|
||||
int ret = 0;
|
||||
bool pending_tickets;
|
||||
|
||||
ASSERT(orig_bytes);
|
||||
ASSERT(!current->journal_info || flush != BTRFS_RESERVE_FLUSH_ALL);
|
||||
|
@ -965,18 +985,19 @@ static int __reserve_metadata_bytes(struct btrfs_fs_info *fs_info,
|
|||
spin_lock(&space_info->lock);
|
||||
ret = -ENOSPC;
|
||||
used = btrfs_space_info_used(space_info, true);
|
||||
pending_tickets = !list_empty(&space_info->tickets) ||
|
||||
!list_empty(&space_info->priority_tickets);
|
||||
|
||||
/*
|
||||
* Carry on if we have enough space (short-circuit) OR call
|
||||
* can_overcommit() to ensure we can overcommit to continue.
|
||||
*/
|
||||
if ((used + orig_bytes <= space_info->total_bytes) ||
|
||||
can_overcommit(fs_info, space_info, orig_bytes, flush,
|
||||
system_chunk)) {
|
||||
if (!pending_tickets &&
|
||||
((used + orig_bytes <= space_info->total_bytes) ||
|
||||
can_overcommit(fs_info, space_info, orig_bytes, flush,
|
||||
system_chunk))) {
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, space_info,
|
||||
orig_bytes);
|
||||
trace_btrfs_space_reservation(fs_info, "space_info",
|
||||
space_info->flags, orig_bytes, 1);
|
||||
ret = 0;
|
||||
}
|
||||
|
||||
|
@ -988,7 +1009,6 @@ static int __reserve_metadata_bytes(struct btrfs_fs_info *fs_info,
|
|||
* the list and we will do our own flushing further down.
|
||||
*/
|
||||
if (ret && flush != BTRFS_RESERVE_NO_FLUSH) {
|
||||
ticket.orig_bytes = orig_bytes;
|
||||
ticket.bytes = orig_bytes;
|
||||
ticket.error = 0;
|
||||
init_waitqueue_head(&ticket.wait);
|
||||
|
@ -1028,25 +1048,7 @@ static int __reserve_metadata_bytes(struct btrfs_fs_info *fs_info,
|
|||
if (!ret || flush == BTRFS_RESERVE_NO_FLUSH)
|
||||
return ret;
|
||||
|
||||
if (flush == BTRFS_RESERVE_FLUSH_ALL)
|
||||
return wait_reserve_ticket(fs_info, space_info, &ticket);
|
||||
|
||||
ret = 0;
|
||||
priority_reclaim_metadata_space(fs_info, space_info, &ticket);
|
||||
spin_lock(&space_info->lock);
|
||||
if (ticket.bytes) {
|
||||
if (ticket.bytes < orig_bytes)
|
||||
reclaim_bytes = orig_bytes - ticket.bytes;
|
||||
list_del_init(&ticket.list);
|
||||
ret = -ENOSPC;
|
||||
}
|
||||
spin_unlock(&space_info->lock);
|
||||
|
||||
if (reclaim_bytes)
|
||||
btrfs_space_info_add_old_bytes(fs_info, space_info,
|
||||
reclaim_bytes);
|
||||
ASSERT(list_empty(&ticket.list));
|
||||
return ret;
|
||||
return handle_reserve_ticket(fs_info, space_info, &ticket, flush);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -70,7 +70,6 @@ struct btrfs_space_info {
|
|||
};
|
||||
|
||||
struct reserve_ticket {
|
||||
u64 orig_bytes;
|
||||
u64 bytes;
|
||||
int error;
|
||||
struct list_head list;
|
||||
|
@ -87,14 +86,18 @@ static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info)
|
|||
*
|
||||
* Declare a helper function to detect underflow of various space info members
|
||||
*/
|
||||
#define DECLARE_SPACE_INFO_UPDATE(name) \
|
||||
#define DECLARE_SPACE_INFO_UPDATE(name, trace_name) \
|
||||
static inline void \
|
||||
btrfs_space_info_update_##name(struct btrfs_fs_info *fs_info, \
|
||||
struct btrfs_space_info *sinfo, \
|
||||
s64 bytes) \
|
||||
{ \
|
||||
const u64 abs_bytes = (bytes < 0) ? -bytes : bytes; \
|
||||
lockdep_assert_held(&sinfo->lock); \
|
||||
trace_update_##name(fs_info, sinfo, sinfo->name, bytes); \
|
||||
trace_btrfs_space_reservation(fs_info, trace_name, \
|
||||
sinfo->flags, abs_bytes, \
|
||||
bytes > 0); \
|
||||
if (bytes < 0 && sinfo->name < -bytes) { \
|
||||
WARN_ON(1); \
|
||||
sinfo->name = 0; \
|
||||
|
@ -103,15 +106,9 @@ btrfs_space_info_update_##name(struct btrfs_fs_info *fs_info, \
|
|||
sinfo->name += bytes; \
|
||||
}
|
||||
|
||||
DECLARE_SPACE_INFO_UPDATE(bytes_may_use);
|
||||
DECLARE_SPACE_INFO_UPDATE(bytes_pinned);
|
||||
DECLARE_SPACE_INFO_UPDATE(bytes_may_use, "space_info");
|
||||
DECLARE_SPACE_INFO_UPDATE(bytes_pinned, "pinned");
|
||||
|
||||
void btrfs_space_info_add_new_bytes(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
u64 num_bytes);
|
||||
void btrfs_space_info_add_old_bytes(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
u64 num_bytes);
|
||||
int btrfs_init_space_info(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_update_space_info(struct btrfs_fs_info *info, u64 flags,
|
||||
u64 total_bytes, u64 bytes_used,
|
||||
|
@ -129,5 +126,18 @@ int btrfs_reserve_metadata_bytes(struct btrfs_root *root,
|
|||
struct btrfs_block_rsv *block_rsv,
|
||||
u64 orig_bytes,
|
||||
enum btrfs_reserve_flush_enum flush);
|
||||
void btrfs_try_granting_tickets(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info);
|
||||
|
||||
static inline void btrfs_space_info_free_bytes_may_use(
|
||||
struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info,
|
||||
u64 num_bytes)
|
||||
{
|
||||
spin_lock(&space_info->lock);
|
||||
btrfs_space_info_update_bytes_may_use(fs_info, space_info, -num_bytes);
|
||||
btrfs_try_granting_tickets(fs_info, space_info);
|
||||
spin_unlock(&space_info->lock);
|
||||
}
|
||||
|
||||
#endif /* BTRFS_SPACE_INFO_H */
|
||||
|
|
|
@ -33,6 +33,8 @@ static inline void put_unaligned_le8(u8 val, void *p)
|
|||
*
|
||||
* The extent buffer api is used to do the page spanning work required to
|
||||
* have a metadata blocksize different from the page size.
|
||||
*
|
||||
* There are 2 variants defined, one with a token pointer and one without.
|
||||
*/
|
||||
|
||||
#define DEFINE_BTRFS_SETGET_BITS(bits) \
|
||||
|
@ -50,8 +52,10 @@ u##bits btrfs_get_token_##bits(const struct extent_buffer *eb, \
|
|||
int size = sizeof(u##bits); \
|
||||
u##bits res; \
|
||||
\
|
||||
if (token && token->kaddr && token->offset <= offset && \
|
||||
token->eb == eb && \
|
||||
ASSERT(token); \
|
||||
ASSERT(token->eb == eb); \
|
||||
\
|
||||
if (token->kaddr && token->offset <= offset && \
|
||||
(token->offset + PAGE_SIZE >= offset + size)) { \
|
||||
kaddr = token->kaddr; \
|
||||
p = kaddr + part_offset - token->offset; \
|
||||
|
@ -68,11 +72,33 @@ u##bits btrfs_get_token_##bits(const struct extent_buffer *eb, \
|
|||
} \
|
||||
p = kaddr + part_offset - map_start; \
|
||||
res = get_unaligned_le##bits(p + off); \
|
||||
if (token) { \
|
||||
token->kaddr = kaddr; \
|
||||
token->offset = map_start; \
|
||||
token->eb = eb; \
|
||||
token->kaddr = kaddr; \
|
||||
token->offset = map_start; \
|
||||
return res; \
|
||||
} \
|
||||
u##bits btrfs_get_##bits(const struct extent_buffer *eb, \
|
||||
const void *ptr, unsigned long off) \
|
||||
{ \
|
||||
unsigned long part_offset = (unsigned long)ptr; \
|
||||
unsigned long offset = part_offset + off; \
|
||||
void *p; \
|
||||
int err; \
|
||||
char *kaddr; \
|
||||
unsigned long map_start; \
|
||||
unsigned long map_len; \
|
||||
int size = sizeof(u##bits); \
|
||||
u##bits res; \
|
||||
\
|
||||
err = map_private_extent_buffer(eb, offset, size, \
|
||||
&kaddr, &map_start, &map_len); \
|
||||
if (err) { \
|
||||
__le##bits leres; \
|
||||
\
|
||||
read_extent_buffer(eb, &leres, offset, size); \
|
||||
return le##bits##_to_cpu(leres); \
|
||||
} \
|
||||
p = kaddr + part_offset - map_start; \
|
||||
res = get_unaligned_le##bits(p + off); \
|
||||
return res; \
|
||||
} \
|
||||
void btrfs_set_token_##bits(struct extent_buffer *eb, \
|
||||
|
@ -89,8 +115,10 @@ void btrfs_set_token_##bits(struct extent_buffer *eb, \
|
|||
unsigned long map_len; \
|
||||
int size = sizeof(u##bits); \
|
||||
\
|
||||
if (token && token->kaddr && token->offset <= offset && \
|
||||
token->eb == eb && \
|
||||
ASSERT(token); \
|
||||
ASSERT(token->eb == eb); \
|
||||
\
|
||||
if (token->kaddr && token->offset <= offset && \
|
||||
(token->offset + PAGE_SIZE >= offset + size)) { \
|
||||
kaddr = token->kaddr; \
|
||||
p = kaddr + part_offset - token->offset; \
|
||||
|
@ -108,11 +136,32 @@ void btrfs_set_token_##bits(struct extent_buffer *eb, \
|
|||
} \
|
||||
p = kaddr + part_offset - map_start; \
|
||||
put_unaligned_le##bits(val, p + off); \
|
||||
if (token) { \
|
||||
token->kaddr = kaddr; \
|
||||
token->offset = map_start; \
|
||||
token->eb = eb; \
|
||||
token->kaddr = kaddr; \
|
||||
token->offset = map_start; \
|
||||
} \
|
||||
void btrfs_set_##bits(struct extent_buffer *eb, void *ptr, \
|
||||
unsigned long off, u##bits val) \
|
||||
{ \
|
||||
unsigned long part_offset = (unsigned long)ptr; \
|
||||
unsigned long offset = part_offset + off; \
|
||||
void *p; \
|
||||
int err; \
|
||||
char *kaddr; \
|
||||
unsigned long map_start; \
|
||||
unsigned long map_len; \
|
||||
int size = sizeof(u##bits); \
|
||||
\
|
||||
err = map_private_extent_buffer(eb, offset, size, \
|
||||
&kaddr, &map_start, &map_len); \
|
||||
if (err) { \
|
||||
__le##bits val2; \
|
||||
\
|
||||
val2 = cpu_to_le##bits(val); \
|
||||
write_extent_buffer(eb, &val2, offset, size); \
|
||||
return; \
|
||||
} \
|
||||
p = kaddr + part_offset - map_start; \
|
||||
put_unaligned_le##bits(val, p + off); \
|
||||
}
|
||||
|
||||
DEFINE_BTRFS_SETGET_BITS(8)
|
||||
|
|
|
@ -43,7 +43,9 @@
|
|||
#include "free-space-cache.h"
|
||||
#include "backref.h"
|
||||
#include "space-info.h"
|
||||
#include "sysfs.h"
|
||||
#include "tests/btrfs-tests.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#include "qgroup.h"
|
||||
#define CREATE_TRACE_POINTS
|
||||
|
@ -1899,11 +1901,10 @@ static inline int btrfs_calc_avail_data_space(struct btrfs_fs_info *fs_info,
|
|||
struct btrfs_device_info *devices_info;
|
||||
struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
|
||||
struct btrfs_device *device;
|
||||
u64 skip_space;
|
||||
u64 type;
|
||||
u64 avail_space;
|
||||
u64 min_stripe_size;
|
||||
int min_stripes, num_stripes = 1;
|
||||
int num_stripes = 1;
|
||||
int i = 0, nr_devices;
|
||||
const struct btrfs_raid_attr *rattr;
|
||||
|
||||
|
@ -1930,7 +1931,6 @@ static inline int btrfs_calc_avail_data_space(struct btrfs_fs_info *fs_info,
|
|||
/* calc min stripe number for data space allocation */
|
||||
type = btrfs_data_alloc_profile(fs_info);
|
||||
rattr = &btrfs_raid_array[btrfs_bg_flags_to_raid_index(type)];
|
||||
min_stripes = rattr->devs_min;
|
||||
|
||||
if (type & BTRFS_BLOCK_GROUP_RAID0)
|
||||
num_stripes = nr_devices;
|
||||
|
@ -1956,28 +1956,21 @@ static inline int btrfs_calc_avail_data_space(struct btrfs_fs_info *fs_info,
|
|||
avail_space = device->total_bytes - device->bytes_used;
|
||||
|
||||
/* align with stripe_len */
|
||||
avail_space = div_u64(avail_space, BTRFS_STRIPE_LEN);
|
||||
avail_space *= BTRFS_STRIPE_LEN;
|
||||
avail_space = rounddown(avail_space, BTRFS_STRIPE_LEN);
|
||||
|
||||
/*
|
||||
* In order to avoid overwriting the superblock on the drive,
|
||||
* btrfs starts at an offset of at least 1MB when doing chunk
|
||||
* allocation.
|
||||
*
|
||||
* This ensures we have at least min_stripe_size free space
|
||||
* after excluding 1MB.
|
||||
*/
|
||||
skip_space = SZ_1M;
|
||||
|
||||
/*
|
||||
* we can use the free space in [0, skip_space - 1], subtract
|
||||
* it from the total.
|
||||
*/
|
||||
if (avail_space && avail_space >= skip_space)
|
||||
avail_space -= skip_space;
|
||||
else
|
||||
avail_space = 0;
|
||||
|
||||
if (avail_space < min_stripe_size)
|
||||
if (avail_space <= SZ_1M + min_stripe_size)
|
||||
continue;
|
||||
|
||||
avail_space -= SZ_1M;
|
||||
|
||||
devices_info[i].dev = device;
|
||||
devices_info[i].max_avail = avail_space;
|
||||
|
||||
|
@ -1991,9 +1984,8 @@ static inline int btrfs_calc_avail_data_space(struct btrfs_fs_info *fs_info,
|
|||
|
||||
i = nr_devices - 1;
|
||||
avail_space = 0;
|
||||
while (nr_devices >= min_stripes) {
|
||||
if (num_stripes > nr_devices)
|
||||
num_stripes = nr_devices;
|
||||
while (nr_devices >= rattr->devs_min) {
|
||||
num_stripes = min(num_stripes, nr_devices);
|
||||
|
||||
if (devices_info[i].max_avail >= min_stripe_size) {
|
||||
int j;
|
||||
|
|
270
fs/btrfs/sysfs.c
270
fs/btrfs/sysfs.c
|
@ -4,12 +4,11 @@
|
|||
*/
|
||||
|
||||
#include <linux/sched.h>
|
||||
#include <linux/sched/mm.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/completion.h>
|
||||
#include <linux/kobject.h>
|
||||
#include <linux/bug.h>
|
||||
#include <linux/debugfs.h>
|
||||
|
||||
#include "ctree.h"
|
||||
#include "disk-io.h"
|
||||
|
@ -17,10 +16,75 @@
|
|||
#include "sysfs.h"
|
||||
#include "volumes.h"
|
||||
#include "space-info.h"
|
||||
#include "block-group.h"
|
||||
|
||||
struct btrfs_feature_attr {
|
||||
struct kobj_attribute kobj_attr;
|
||||
enum btrfs_feature_set feature_set;
|
||||
u64 feature_bit;
|
||||
};
|
||||
|
||||
/* For raid type sysfs entries */
|
||||
struct raid_kobject {
|
||||
u64 flags;
|
||||
struct kobject kobj;
|
||||
};
|
||||
|
||||
#define __INIT_KOBJ_ATTR(_name, _mode, _show, _store) \
|
||||
{ \
|
||||
.attr = { .name = __stringify(_name), .mode = _mode }, \
|
||||
.show = _show, \
|
||||
.store = _store, \
|
||||
}
|
||||
|
||||
#define BTRFS_ATTR_RW(_prefix, _name, _show, _store) \
|
||||
static struct kobj_attribute btrfs_attr_##_prefix##_##_name = \
|
||||
__INIT_KOBJ_ATTR(_name, 0644, _show, _store)
|
||||
|
||||
#define BTRFS_ATTR(_prefix, _name, _show) \
|
||||
static struct kobj_attribute btrfs_attr_##_prefix##_##_name = \
|
||||
__INIT_KOBJ_ATTR(_name, 0444, _show, NULL)
|
||||
|
||||
#define BTRFS_ATTR_PTR(_prefix, _name) \
|
||||
(&btrfs_attr_##_prefix##_##_name.attr)
|
||||
|
||||
#define BTRFS_FEAT_ATTR(_name, _feature_set, _feature_prefix, _feature_bit) \
|
||||
static struct btrfs_feature_attr btrfs_attr_features_##_name = { \
|
||||
.kobj_attr = __INIT_KOBJ_ATTR(_name, S_IRUGO, \
|
||||
btrfs_feature_attr_show, \
|
||||
btrfs_feature_attr_store), \
|
||||
.feature_set = _feature_set, \
|
||||
.feature_bit = _feature_prefix ##_## _feature_bit, \
|
||||
}
|
||||
#define BTRFS_FEAT_ATTR_PTR(_name) \
|
||||
(&btrfs_attr_features_##_name.kobj_attr.attr)
|
||||
|
||||
#define BTRFS_FEAT_ATTR_COMPAT(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_COMPAT, BTRFS_FEATURE_COMPAT, feature)
|
||||
#define BTRFS_FEAT_ATTR_COMPAT_RO(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_COMPAT_RO, BTRFS_FEATURE_COMPAT_RO, feature)
|
||||
#define BTRFS_FEAT_ATTR_INCOMPAT(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_INCOMPAT, BTRFS_FEATURE_INCOMPAT, feature)
|
||||
|
||||
static inline struct btrfs_fs_info *to_fs_info(struct kobject *kobj);
|
||||
static inline struct btrfs_fs_devices *to_fs_devs(struct kobject *kobj);
|
||||
|
||||
static struct btrfs_feature_attr *to_btrfs_feature_attr(struct kobj_attribute *a)
|
||||
{
|
||||
return container_of(a, struct btrfs_feature_attr, kobj_attr);
|
||||
}
|
||||
|
||||
static struct kobj_attribute *attr_to_btrfs_attr(struct attribute *attr)
|
||||
{
|
||||
return container_of(attr, struct kobj_attribute, attr);
|
||||
}
|
||||
|
||||
static struct btrfs_feature_attr *attr_to_btrfs_feature_attr(
|
||||
struct attribute *attr)
|
||||
{
|
||||
return to_btrfs_feature_attr(attr_to_btrfs_attr(attr));
|
||||
}
|
||||
|
||||
static u64 get_features(struct btrfs_fs_info *fs_info,
|
||||
enum btrfs_feature_set set)
|
||||
{
|
||||
|
@ -247,6 +311,25 @@ static const struct attribute_group btrfs_static_feature_attr_group = {
|
|||
.attrs = btrfs_supported_static_feature_attrs,
|
||||
};
|
||||
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
|
||||
/*
|
||||
* Runtime debugging exported via sysfs
|
||||
*
|
||||
* /sys/fs/btrfs/debug - applies to module or all filesystems
|
||||
* /sys/fs/btrfs/UUID - applies only to the given filesystem
|
||||
*/
|
||||
static struct attribute *btrfs_debug_feature_attrs[] = {
|
||||
NULL
|
||||
};
|
||||
|
||||
static const struct attribute_group btrfs_debug_feature_attr_group = {
|
||||
.name = "debug",
|
||||
.attrs = btrfs_debug_feature_attrs,
|
||||
};
|
||||
|
||||
#endif
|
||||
|
||||
static ssize_t btrfs_show_u64(u64 *value_ptr, spinlock_t *lock, char *buf)
|
||||
{
|
||||
u64 val;
|
||||
|
@ -316,7 +399,7 @@ static void release_raid_kobj(struct kobject *kobj)
|
|||
kfree(to_raid_kobj(kobj));
|
||||
}
|
||||
|
||||
struct kobj_type btrfs_raid_ktype = {
|
||||
static struct kobj_type btrfs_raid_ktype = {
|
||||
.sysfs_ops = &kobj_sysfs_ops,
|
||||
.release = release_raid_kobj,
|
||||
.default_groups = raid_groups,
|
||||
|
@ -375,7 +458,7 @@ static void space_info_release(struct kobject *kobj)
|
|||
kfree(sinfo);
|
||||
}
|
||||
|
||||
struct kobj_type space_info_ktype = {
|
||||
static struct kobj_type space_info_ktype = {
|
||||
.sysfs_ops = &kobj_sysfs_ops,
|
||||
.release = space_info_release,
|
||||
.default_groups = space_info_groups,
|
||||
|
@ -655,12 +738,17 @@ void btrfs_sysfs_remove_mounted(struct btrfs_fs_info *fs_info)
|
|||
btrfs_sysfs_rm_device_link(fs_info->fs_devices, NULL);
|
||||
}
|
||||
|
||||
const char * const btrfs_feature_set_names[FEAT_MAX] = {
|
||||
static const char * const btrfs_feature_set_names[FEAT_MAX] = {
|
||||
[FEAT_COMPAT] = "compat",
|
||||
[FEAT_COMPAT_RO] = "compat_ro",
|
||||
[FEAT_INCOMPAT] = "incompat",
|
||||
};
|
||||
|
||||
const char * const btrfs_feature_set_name(enum btrfs_feature_set set)
|
||||
{
|
||||
return btrfs_feature_set_names[set];
|
||||
}
|
||||
|
||||
char *btrfs_printable_features(enum btrfs_feature_set set, u64 flags)
|
||||
{
|
||||
size_t bufsize = 4096; /* safe max, 64 names * 64 bytes */
|
||||
|
@ -730,6 +818,110 @@ static void init_feature_attrs(void)
|
|||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Create a sysfs entry for a given block group type at path
|
||||
* /sys/fs/btrfs/UUID/allocation/data/TYPE
|
||||
*/
|
||||
void btrfs_sysfs_add_block_group_type(struct btrfs_block_group_cache *cache)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = cache->fs_info;
|
||||
struct btrfs_space_info *space_info = cache->space_info;
|
||||
struct raid_kobject *rkobj;
|
||||
const int index = btrfs_bg_flags_to_raid_index(cache->flags);
|
||||
unsigned int nofs_flag;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Setup a NOFS context because kobject_add(), deep in its call chain,
|
||||
* does GFP_KERNEL allocations, and we are often called in a context
|
||||
* where if reclaim is triggered we can deadlock (we are either holding
|
||||
* a transaction handle or some lock required for a transaction
|
||||
* commit).
|
||||
*/
|
||||
nofs_flag = memalloc_nofs_save();
|
||||
|
||||
rkobj = kzalloc(sizeof(*rkobj), GFP_NOFS);
|
||||
if (!rkobj) {
|
||||
memalloc_nofs_restore(nofs_flag);
|
||||
btrfs_warn(cache->fs_info,
|
||||
"couldn't alloc memory for raid level kobject");
|
||||
return;
|
||||
}
|
||||
|
||||
rkobj->flags = cache->flags;
|
||||
kobject_init(&rkobj->kobj, &btrfs_raid_ktype);
|
||||
ret = kobject_add(&rkobj->kobj, &space_info->kobj, "%s",
|
||||
btrfs_bg_type_to_raid_name(rkobj->flags));
|
||||
memalloc_nofs_restore(nofs_flag);
|
||||
if (ret) {
|
||||
kobject_put(&rkobj->kobj);
|
||||
btrfs_warn(fs_info,
|
||||
"failed to add kobject for block cache, ignoring");
|
||||
return;
|
||||
}
|
||||
|
||||
space_info->block_group_kobjs[index] = &rkobj->kobj;
|
||||
}
|
||||
|
||||
/*
|
||||
* Remove sysfs directories for all block group types of a given space info and
|
||||
* the space info as well
|
||||
*/
|
||||
void btrfs_sysfs_remove_space_info(struct btrfs_space_info *space_info)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < BTRFS_NR_RAID_TYPES; i++) {
|
||||
struct kobject *kobj;
|
||||
|
||||
kobj = space_info->block_group_kobjs[i];
|
||||
space_info->block_group_kobjs[i] = NULL;
|
||||
if (kobj) {
|
||||
kobject_del(kobj);
|
||||
kobject_put(kobj);
|
||||
}
|
||||
}
|
||||
kobject_del(&space_info->kobj);
|
||||
kobject_put(&space_info->kobj);
|
||||
}
|
||||
|
||||
static const char *alloc_name(u64 flags)
|
||||
{
|
||||
switch (flags) {
|
||||
case BTRFS_BLOCK_GROUP_METADATA | BTRFS_BLOCK_GROUP_DATA:
|
||||
return "mixed";
|
||||
case BTRFS_BLOCK_GROUP_METADATA:
|
||||
return "metadata";
|
||||
case BTRFS_BLOCK_GROUP_DATA:
|
||||
return "data";
|
||||
case BTRFS_BLOCK_GROUP_SYSTEM:
|
||||
return "system";
|
||||
default:
|
||||
WARN_ON(1);
|
||||
return "invalid-combination";
|
||||
};
|
||||
}
|
||||
|
||||
/*
|
||||
* Create a sysfs entry for a space info type at path
|
||||
* /sys/fs/btrfs/UUID/allocation/TYPE
|
||||
*/
|
||||
int btrfs_sysfs_add_space_info_type(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = kobject_init_and_add(&space_info->kobj, &space_info_ktype,
|
||||
fs_info->space_info_kobj, "%s",
|
||||
alloc_name(space_info->flags));
|
||||
if (ret) {
|
||||
kobject_put(&space_info->kobj);
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* when one_device is NULL, it removes all device links */
|
||||
|
||||
int btrfs_sysfs_rm_device_link(struct btrfs_fs_devices *fs_devices,
|
||||
|
@ -806,15 +998,35 @@ int btrfs_sysfs_add_device_link(struct btrfs_fs_devices *fs_devices,
|
|||
return error;
|
||||
}
|
||||
|
||||
void btrfs_kobject_uevent(struct block_device *bdev, enum kobject_action action)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = kobject_uevent(&disk_to_dev(bdev->bd_disk)->kobj, action);
|
||||
if (ret)
|
||||
pr_warn("BTRFS: Sending event '%d' to kobject: '%s' (%p): failed\n",
|
||||
action, kobject_name(&disk_to_dev(bdev->bd_disk)->kobj),
|
||||
&disk_to_dev(bdev->bd_disk)->kobj);
|
||||
}
|
||||
|
||||
void btrfs_sysfs_update_sprout_fsid(struct btrfs_fs_devices *fs_devices,
|
||||
const u8 *fsid)
|
||||
{
|
||||
char fsid_buf[BTRFS_UUID_UNPARSED_SIZE];
|
||||
|
||||
/*
|
||||
* Sprouting changes fsid of the mounted filesystem, rename the fsid
|
||||
* directory
|
||||
*/
|
||||
snprintf(fsid_buf, BTRFS_UUID_UNPARSED_SIZE, "%pU", fsid);
|
||||
if (kobject_rename(&fs_devices->fsid_kobj, fsid_buf))
|
||||
btrfs_warn(fs_devices->fs_info,
|
||||
"sysfs: failed to create fsid for sprout");
|
||||
}
|
||||
|
||||
/* /sys/fs/btrfs/ entry */
|
||||
static struct kset *btrfs_kset;
|
||||
|
||||
/* /sys/kernel/debug/btrfs */
|
||||
static struct dentry *btrfs_debugfs_root_dentry;
|
||||
|
||||
/* Debugging tunables and exported data */
|
||||
u64 btrfs_debugfs_test;
|
||||
|
||||
/*
|
||||
* Can be called by the device discovery thread.
|
||||
* And parent can be specified for seed device
|
||||
|
@ -859,6 +1071,13 @@ int btrfs_sysfs_add_mounted(struct btrfs_fs_info *fs_info)
|
|||
if (error)
|
||||
goto failure;
|
||||
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
error = sysfs_create_group(fsid_kobj,
|
||||
&btrfs_debug_feature_attr_group);
|
||||
if (error)
|
||||
goto failure;
|
||||
#endif
|
||||
|
||||
error = addrm_unknown_feature_attrs(fs_info, true);
|
||||
if (error)
|
||||
goto failure;
|
||||
|
@ -913,25 +1132,6 @@ void btrfs_sysfs_feature_update(struct btrfs_fs_info *fs_info,
|
|||
ret = sysfs_create_group(fsid_kobj, &btrfs_feature_attr_group);
|
||||
}
|
||||
|
||||
static void btrfs_init_debugfs(void)
|
||||
{
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
btrfs_debugfs_root_dentry = debugfs_create_dir("btrfs", NULL);
|
||||
|
||||
/*
|
||||
* Example code, how to export data through debugfs.
|
||||
*
|
||||
* file: /sys/kernel/debug/btrfs/test
|
||||
* contents of: btrfs_debugfs_test
|
||||
*/
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
debugfs_create_u64("test", S_IRUGO | S_IWUSR, btrfs_debugfs_root_dentry,
|
||||
&btrfs_debugfs_test);
|
||||
#endif
|
||||
|
||||
#endif
|
||||
}
|
||||
|
||||
int __init btrfs_init_sysfs(void)
|
||||
{
|
||||
int ret;
|
||||
|
@ -940,8 +1140,6 @@ int __init btrfs_init_sysfs(void)
|
|||
if (!btrfs_kset)
|
||||
return -ENOMEM;
|
||||
|
||||
btrfs_init_debugfs();
|
||||
|
||||
init_feature_attrs();
|
||||
ret = sysfs_create_group(&btrfs_kset->kobj, &btrfs_feature_attr_group);
|
||||
if (ret)
|
||||
|
@ -951,12 +1149,17 @@ int __init btrfs_init_sysfs(void)
|
|||
if (ret)
|
||||
goto out_remove_group;
|
||||
|
||||
#ifdef CONFIG_BTRFS_DEBUG
|
||||
ret = sysfs_create_group(&btrfs_kset->kobj, &btrfs_debug_feature_attr_group);
|
||||
if (ret)
|
||||
goto out2;
|
||||
#endif
|
||||
|
||||
return 0;
|
||||
|
||||
out_remove_group:
|
||||
sysfs_remove_group(&btrfs_kset->kobj, &btrfs_feature_attr_group);
|
||||
out2:
|
||||
debugfs_remove_recursive(btrfs_debugfs_root_dentry);
|
||||
kset_unregister(btrfs_kset);
|
||||
|
||||
return ret;
|
||||
|
@ -968,6 +1171,5 @@ void __cold btrfs_exit_sysfs(void)
|
|||
&btrfs_static_feature_attr_group);
|
||||
sysfs_remove_group(&btrfs_kset->kobj, &btrfs_feature_attr_group);
|
||||
kset_unregister(btrfs_kset);
|
||||
debugfs_remove_recursive(btrfs_debugfs_root_dentry);
|
||||
}
|
||||
|
||||
|
|
|
@ -3,10 +3,7 @@
|
|||
#ifndef BTRFS_SYSFS_H
|
||||
#define BTRFS_SYSFS_H
|
||||
|
||||
/*
|
||||
* Data exported through sysfs
|
||||
*/
|
||||
extern u64 btrfs_debugfs_test;
|
||||
#include <linux/kobject.h>
|
||||
|
||||
enum btrfs_feature_set {
|
||||
FEAT_COMPAT,
|
||||
|
@ -15,71 +12,8 @@ enum btrfs_feature_set {
|
|||
FEAT_MAX
|
||||
};
|
||||
|
||||
#define __INIT_KOBJ_ATTR(_name, _mode, _show, _store) \
|
||||
{ \
|
||||
.attr = { .name = __stringify(_name), .mode = _mode }, \
|
||||
.show = _show, \
|
||||
.store = _store, \
|
||||
}
|
||||
|
||||
#define BTRFS_ATTR_RW(_prefix, _name, _show, _store) \
|
||||
static struct kobj_attribute btrfs_attr_##_prefix##_##_name = \
|
||||
__INIT_KOBJ_ATTR(_name, 0644, _show, _store)
|
||||
|
||||
#define BTRFS_ATTR(_prefix, _name, _show) \
|
||||
static struct kobj_attribute btrfs_attr_##_prefix##_##_name = \
|
||||
__INIT_KOBJ_ATTR(_name, 0444, _show, NULL)
|
||||
|
||||
#define BTRFS_ATTR_PTR(_prefix, _name) \
|
||||
(&btrfs_attr_##_prefix##_##_name.attr)
|
||||
|
||||
|
||||
struct btrfs_feature_attr {
|
||||
struct kobj_attribute kobj_attr;
|
||||
enum btrfs_feature_set feature_set;
|
||||
u64 feature_bit;
|
||||
};
|
||||
|
||||
#define BTRFS_FEAT_ATTR(_name, _feature_set, _feature_prefix, _feature_bit) \
|
||||
static struct btrfs_feature_attr btrfs_attr_features_##_name = { \
|
||||
.kobj_attr = __INIT_KOBJ_ATTR(_name, S_IRUGO, \
|
||||
btrfs_feature_attr_show, \
|
||||
btrfs_feature_attr_store), \
|
||||
.feature_set = _feature_set, \
|
||||
.feature_bit = _feature_prefix ##_## _feature_bit, \
|
||||
}
|
||||
#define BTRFS_FEAT_ATTR_PTR(_name) \
|
||||
(&btrfs_attr_features_##_name.kobj_attr.attr)
|
||||
|
||||
#define BTRFS_FEAT_ATTR_COMPAT(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_COMPAT, BTRFS_FEATURE_COMPAT, feature)
|
||||
#define BTRFS_FEAT_ATTR_COMPAT_RO(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_COMPAT_RO, BTRFS_FEATURE_COMPAT_RO, feature)
|
||||
#define BTRFS_FEAT_ATTR_INCOMPAT(name, feature) \
|
||||
BTRFS_FEAT_ATTR(name, FEAT_INCOMPAT, BTRFS_FEATURE_INCOMPAT, feature)
|
||||
|
||||
/* convert from attribute */
|
||||
static inline struct btrfs_feature_attr *
|
||||
to_btrfs_feature_attr(struct kobj_attribute *a)
|
||||
{
|
||||
return container_of(a, struct btrfs_feature_attr, kobj_attr);
|
||||
}
|
||||
|
||||
static inline struct kobj_attribute *attr_to_btrfs_attr(struct attribute *attr)
|
||||
{
|
||||
return container_of(attr, struct kobj_attribute, attr);
|
||||
}
|
||||
|
||||
static inline struct btrfs_feature_attr *
|
||||
attr_to_btrfs_feature_attr(struct attribute *attr)
|
||||
{
|
||||
return to_btrfs_feature_attr(attr_to_btrfs_attr(attr));
|
||||
}
|
||||
|
||||
char *btrfs_printable_features(enum btrfs_feature_set set, u64 flags);
|
||||
extern const char * const btrfs_feature_set_names[FEAT_MAX];
|
||||
extern struct kobj_type space_info_ktype;
|
||||
extern struct kobj_type btrfs_raid_ktype;
|
||||
const char * const btrfs_feature_set_name(enum btrfs_feature_set set);
|
||||
int btrfs_sysfs_add_device_link(struct btrfs_fs_devices *fs_devices,
|
||||
struct btrfs_device *one_device);
|
||||
int btrfs_sysfs_rm_device_link(struct btrfs_fs_devices *fs_devices,
|
||||
|
@ -88,7 +22,19 @@ int btrfs_sysfs_add_fsid(struct btrfs_fs_devices *fs_devs,
|
|||
struct kobject *parent);
|
||||
int btrfs_sysfs_add_device(struct btrfs_fs_devices *fs_devs);
|
||||
void btrfs_sysfs_remove_fsid(struct btrfs_fs_devices *fs_devs);
|
||||
void btrfs_sysfs_update_sprout_fsid(struct btrfs_fs_devices *fs_devices,
|
||||
const u8 *fsid);
|
||||
void btrfs_sysfs_feature_update(struct btrfs_fs_info *fs_info,
|
||||
u64 bit, enum btrfs_feature_set set);
|
||||
void btrfs_kobject_uevent(struct block_device *bdev, enum kobject_action action);
|
||||
|
||||
int __init btrfs_init_sysfs(void);
|
||||
void __cold btrfs_exit_sysfs(void);
|
||||
int btrfs_sysfs_add_mounted(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_sysfs_remove_mounted(struct btrfs_fs_info *fs_info);
|
||||
void btrfs_sysfs_add_block_group_type(struct btrfs_block_group_cache *cache);
|
||||
int btrfs_sysfs_add_space_info_type(struct btrfs_fs_info *fs_info,
|
||||
struct btrfs_space_info *space_info);
|
||||
void btrfs_sysfs_remove_space_info(struct btrfs_space_info *space_info);
|
||||
|
||||
#endif
|
||||
|
|
|
@ -15,6 +15,7 @@
|
|||
#include "../volumes.h"
|
||||
#include "../disk-io.h"
|
||||
#include "../qgroup.h"
|
||||
#include "../block-group.h"
|
||||
|
||||
static struct vfsmount *test_mnt = NULL;
|
||||
|
||||
|
|
|
@ -438,6 +438,7 @@ static int test_find_first_clear_extent_bit(void)
|
|||
{
|
||||
struct extent_io_tree tree;
|
||||
u64 start, end;
|
||||
int ret = -EINVAL;
|
||||
|
||||
test_msg("running find_first_clear_extent_bit test");
|
||||
extent_io_tree_init(NULL, &tree, IO_TREE_SELFTEST, NULL);
|
||||
|
@ -452,9 +453,11 @@ static int test_find_first_clear_extent_bit(void)
|
|||
find_first_clear_extent_bit(&tree, SZ_512K, &start, &end,
|
||||
CHUNK_TRIMMED | CHUNK_ALLOCATED);
|
||||
|
||||
if (start != 0 || end != SZ_1M -1)
|
||||
if (start != 0 || end != SZ_1M - 1) {
|
||||
test_err("error finding beginning range: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/* Now add 32M-64M so that we have a hole between 4M-32M */
|
||||
set_extent_bits(&tree, SZ_32M, SZ_64M - 1,
|
||||
|
@ -466,9 +469,11 @@ static int test_find_first_clear_extent_bit(void)
|
|||
find_first_clear_extent_bit(&tree, 12 * SZ_1M, &start, &end,
|
||||
CHUNK_TRIMMED | CHUNK_ALLOCATED);
|
||||
|
||||
if (start != SZ_4M || end != SZ_32M - 1)
|
||||
if (start != SZ_4M || end != SZ_32M - 1) {
|
||||
test_err("error finding trimmed range: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/*
|
||||
* Search in the middle of allocated range, should get the next one
|
||||
|
@ -477,9 +482,11 @@ static int test_find_first_clear_extent_bit(void)
|
|||
find_first_clear_extent_bit(&tree, SZ_2M, &start, &end,
|
||||
CHUNK_TRIMMED | CHUNK_ALLOCATED);
|
||||
|
||||
if (start != SZ_4M || end != SZ_32M -1)
|
||||
if (start != SZ_4M || end != SZ_32M - 1) {
|
||||
test_err("error finding next unalloc range: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/*
|
||||
* Set 64M-72M with CHUNK_ALLOC flag, then search for CHUNK_TRIMMED flag
|
||||
|
@ -489,9 +496,11 @@ static int test_find_first_clear_extent_bit(void)
|
|||
find_first_clear_extent_bit(&tree, SZ_64M + SZ_1M, &start, &end,
|
||||
CHUNK_TRIMMED);
|
||||
|
||||
if (start != SZ_64M || end != SZ_64M + SZ_8M - 1)
|
||||
if (start != SZ_64M || end != SZ_64M + SZ_8M - 1) {
|
||||
test_err("error finding exact range: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
find_first_clear_extent_bit(&tree, SZ_64M - SZ_8M, &start, &end,
|
||||
CHUNK_TRIMMED);
|
||||
|
@ -500,21 +509,29 @@ static int test_find_first_clear_extent_bit(void)
|
|||
* Search in the middle of set range whose immediate neighbour doesn't
|
||||
* have the bits set so it must be returned
|
||||
*/
|
||||
if (start != SZ_64M || end != SZ_64M + SZ_8M - 1)
|
||||
if (start != SZ_64M || end != SZ_64M + SZ_8M - 1) {
|
||||
test_err("error finding next alloc range: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/*
|
||||
* Search beyond any known range, shall return after last known range
|
||||
* and end should be -1
|
||||
*/
|
||||
find_first_clear_extent_bit(&tree, -1, &start, &end, CHUNK_TRIMMED);
|
||||
if (start != SZ_64M + SZ_8M || end != -1)
|
||||
if (start != SZ_64M + SZ_8M || end != -1) {
|
||||
test_err(
|
||||
"error handling beyond end of range search: start %llu end %llu",
|
||||
start, end);
|
||||
goto out;
|
||||
}
|
||||
|
||||
return 0;
|
||||
ret = 0;
|
||||
out:
|
||||
clear_extent_bits(&tree, 0, (u64)-1, CHUNK_TRIMMED | CHUNK_ALLOCATED);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
int btrfs_test_extent_io(u32 sectorsize, u32 nodesize)
|
||||
|
|
|
@ -8,6 +8,7 @@
|
|||
#include "../ctree.h"
|
||||
#include "../disk-io.h"
|
||||
#include "../free-space-cache.h"
|
||||
#include "../block-group.h"
|
||||
|
||||
#define BITS_PER_BITMAP (PAGE_SIZE * 8UL)
|
||||
|
||||
|
|
|
@ -9,6 +9,7 @@
|
|||
#include "../disk-io.h"
|
||||
#include "../free-space-tree.h"
|
||||
#include "../transaction.h"
|
||||
#include "../block-group.h"
|
||||
|
||||
struct free_space_extent {
|
||||
u64 start;
|
||||
|
|
|
@ -957,7 +957,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
|
||||
/* [BTRFS_MAX_EXTENT_SIZE] */
|
||||
ret = btrfs_set_extent_delalloc(inode, 0, BTRFS_MAX_EXTENT_SIZE - 1, 0,
|
||||
NULL, 0);
|
||||
NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -972,7 +972,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
/* [BTRFS_MAX_EXTENT_SIZE][sectorsize] */
|
||||
ret = btrfs_set_extent_delalloc(inode, BTRFS_MAX_EXTENT_SIZE,
|
||||
BTRFS_MAX_EXTENT_SIZE + sectorsize - 1,
|
||||
0, NULL, 0);
|
||||
0, NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -988,8 +988,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
ret = clear_extent_bit(&BTRFS_I(inode)->io_tree,
|
||||
BTRFS_MAX_EXTENT_SIZE >> 1,
|
||||
(BTRFS_MAX_EXTENT_SIZE >> 1) + sectorsize - 1,
|
||||
EXTENT_DELALLOC | EXTENT_DIRTY |
|
||||
EXTENT_UPTODATE, 0, 0, NULL);
|
||||
EXTENT_DELALLOC | EXTENT_UPTODATE, 0, 0, NULL);
|
||||
if (ret) {
|
||||
test_err("clear_extent_bit returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1005,7 +1004,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
ret = btrfs_set_extent_delalloc(inode, BTRFS_MAX_EXTENT_SIZE >> 1,
|
||||
(BTRFS_MAX_EXTENT_SIZE >> 1)
|
||||
+ sectorsize - 1,
|
||||
0, NULL, 0);
|
||||
0, NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1023,7 +1022,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
ret = btrfs_set_extent_delalloc(inode,
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize,
|
||||
(BTRFS_MAX_EXTENT_SIZE << 1) + 3 * sectorsize - 1,
|
||||
0, NULL, 0);
|
||||
0, NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1040,7 +1039,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
*/
|
||||
ret = btrfs_set_extent_delalloc(inode,
|
||||
BTRFS_MAX_EXTENT_SIZE + sectorsize,
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize - 1, 0, NULL, 0);
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize - 1, 0, NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1056,8 +1055,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
ret = clear_extent_bit(&BTRFS_I(inode)->io_tree,
|
||||
BTRFS_MAX_EXTENT_SIZE + sectorsize,
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize - 1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_UPTODATE, 0, 0, NULL);
|
||||
EXTENT_DELALLOC | EXTENT_UPTODATE, 0, 0, NULL);
|
||||
if (ret) {
|
||||
test_err("clear_extent_bit returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1075,7 +1073,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
*/
|
||||
ret = btrfs_set_extent_delalloc(inode,
|
||||
BTRFS_MAX_EXTENT_SIZE + sectorsize,
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize - 1, 0, NULL, 0);
|
||||
BTRFS_MAX_EXTENT_SIZE + 2 * sectorsize - 1, 0, NULL);
|
||||
if (ret) {
|
||||
test_err("btrfs_set_extent_delalloc returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1089,8 +1087,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
|
||||
/* Empty */
|
||||
ret = clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, (u64)-1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_UPTODATE, 0, 0, NULL);
|
||||
EXTENT_DELALLOC | EXTENT_UPTODATE, 0, 0, NULL);
|
||||
if (ret) {
|
||||
test_err("clear_extent_bit returned %d", ret);
|
||||
goto out;
|
||||
|
@ -1105,8 +1102,7 @@ static int test_extent_accounting(u32 sectorsize, u32 nodesize)
|
|||
out:
|
||||
if (ret)
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, (u64)-1,
|
||||
EXTENT_DIRTY | EXTENT_DELALLOC |
|
||||
EXTENT_UPTODATE, 0, 0, NULL);
|
||||
EXTENT_DELALLOC | EXTENT_UPTODATE, 0, 0, NULL);
|
||||
iput(inode);
|
||||
btrfs_free_dummy_root(root);
|
||||
btrfs_free_dummy_fs_info(fs_info);
|
||||
|
|
|
@ -10,6 +10,7 @@
|
|||
#include <linux/pagemap.h>
|
||||
#include <linux/blkdev.h>
|
||||
#include <linux/uuid.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "disk-io.h"
|
||||
#include "transaction.h"
|
||||
|
@ -19,6 +20,7 @@
|
|||
#include "volumes.h"
|
||||
#include "dev-replace.h"
|
||||
#include "qgroup.h"
|
||||
#include "block-group.h"
|
||||
|
||||
#define BTRFS_ROOT_TRANS_TAG 0
|
||||
|
||||
|
@ -484,7 +486,7 @@ start_transaction(struct btrfs_root *root, unsigned int num_items,
|
|||
* worth of delayed refs updates in this trans handle, and
|
||||
* refill that amount for whatever is missing in the reserve.
|
||||
*/
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, num_items);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info, num_items);
|
||||
if (delayed_refs_rsv->full == 0) {
|
||||
delayed_refs_bytes = num_bytes;
|
||||
num_bytes <<= 1;
|
||||
|
@ -635,7 +637,7 @@ struct btrfs_trans_handle *btrfs_start_transaction_fallback_global_rsv(
|
|||
if (IS_ERR(trans))
|
||||
return trans;
|
||||
|
||||
num_bytes = btrfs_calc_trans_metadata_size(fs_info, num_items);
|
||||
num_bytes = btrfs_calc_insert_metadata_size(fs_info, num_items);
|
||||
ret = btrfs_cond_migrate_bytes(fs_info, &fs_info->trans_block_rsv,
|
||||
num_bytes, min_factor);
|
||||
if (ret) {
|
||||
|
|
|
@ -821,6 +821,417 @@ static int check_inode_item(struct extent_buffer *leaf,
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int check_root_item(struct extent_buffer *leaf, struct btrfs_key *key,
|
||||
int slot)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = leaf->fs_info;
|
||||
struct btrfs_root_item ri;
|
||||
const u64 valid_root_flags = BTRFS_ROOT_SUBVOL_RDONLY |
|
||||
BTRFS_ROOT_SUBVOL_DEAD;
|
||||
|
||||
/* No such tree id */
|
||||
if (key->objectid == 0) {
|
||||
generic_err(leaf, slot, "invalid root id 0");
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/*
|
||||
* Some older kernel may create ROOT_ITEM with non-zero offset, so here
|
||||
* we only check offset for reloc tree whose key->offset must be a
|
||||
* valid tree.
|
||||
*/
|
||||
if (key->objectid == BTRFS_TREE_RELOC_OBJECTID && key->offset == 0) {
|
||||
generic_err(leaf, slot, "invalid root id 0 for reloc tree");
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
if (btrfs_item_size_nr(leaf, slot) != sizeof(ri)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root item size, have %u expect %zu",
|
||||
btrfs_item_size_nr(leaf, slot), sizeof(ri));
|
||||
}
|
||||
|
||||
read_extent_buffer(leaf, &ri, btrfs_item_ptr_offset(leaf, slot),
|
||||
sizeof(ri));
|
||||
|
||||
/* Generation related */
|
||||
if (btrfs_root_generation(&ri) >
|
||||
btrfs_super_generation(fs_info->super_copy) + 1) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root generation, have %llu expect (0, %llu]",
|
||||
btrfs_root_generation(&ri),
|
||||
btrfs_super_generation(fs_info->super_copy) + 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (btrfs_root_generation_v2(&ri) >
|
||||
btrfs_super_generation(fs_info->super_copy) + 1) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root v2 generation, have %llu expect (0, %llu]",
|
||||
btrfs_root_generation_v2(&ri),
|
||||
btrfs_super_generation(fs_info->super_copy) + 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (btrfs_root_last_snapshot(&ri) >
|
||||
btrfs_super_generation(fs_info->super_copy) + 1) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root last_snapshot, have %llu expect (0, %llu]",
|
||||
btrfs_root_last_snapshot(&ri),
|
||||
btrfs_super_generation(fs_info->super_copy) + 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/* Alignment and level check */
|
||||
if (!IS_ALIGNED(btrfs_root_bytenr(&ri), fs_info->sectorsize)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root bytenr, have %llu expect to be aligned to %u",
|
||||
btrfs_root_bytenr(&ri), fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (btrfs_root_level(&ri) >= BTRFS_MAX_LEVEL) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root level, have %u expect [0, %u]",
|
||||
btrfs_root_level(&ri), BTRFS_MAX_LEVEL - 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (ri.drop_level >= BTRFS_MAX_LEVEL) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root level, have %u expect [0, %u]",
|
||||
ri.drop_level, BTRFS_MAX_LEVEL - 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/* Flags check */
|
||||
if (btrfs_root_flags(&ri) & ~valid_root_flags) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid root flags, have 0x%llx expect mask 0x%llx",
|
||||
btrfs_root_flags(&ri), valid_root_flags);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
__printf(3,4)
|
||||
__cold
|
||||
static void extent_err(const struct extent_buffer *eb, int slot,
|
||||
const char *fmt, ...)
|
||||
{
|
||||
struct btrfs_key key;
|
||||
struct va_format vaf;
|
||||
va_list args;
|
||||
u64 bytenr;
|
||||
u64 len;
|
||||
|
||||
btrfs_item_key_to_cpu(eb, &key, slot);
|
||||
bytenr = key.objectid;
|
||||
if (key.type == BTRFS_METADATA_ITEM_KEY ||
|
||||
key.type == BTRFS_TREE_BLOCK_REF_KEY ||
|
||||
key.type == BTRFS_SHARED_BLOCK_REF_KEY)
|
||||
len = eb->fs_info->nodesize;
|
||||
else
|
||||
len = key.offset;
|
||||
va_start(args, fmt);
|
||||
|
||||
vaf.fmt = fmt;
|
||||
vaf.va = &args;
|
||||
|
||||
btrfs_crit(eb->fs_info,
|
||||
"corrupt %s: block=%llu slot=%d extent bytenr=%llu len=%llu %pV",
|
||||
btrfs_header_level(eb) == 0 ? "leaf" : "node",
|
||||
eb->start, slot, bytenr, len, &vaf);
|
||||
va_end(args);
|
||||
}
|
||||
|
||||
static int check_extent_item(struct extent_buffer *leaf,
|
||||
struct btrfs_key *key, int slot)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = leaf->fs_info;
|
||||
struct btrfs_extent_item *ei;
|
||||
bool is_tree_block = false;
|
||||
unsigned long ptr; /* Current pointer inside inline refs */
|
||||
unsigned long end; /* Extent item end */
|
||||
const u32 item_size = btrfs_item_size_nr(leaf, slot);
|
||||
u64 flags;
|
||||
u64 generation;
|
||||
u64 total_refs; /* Total refs in btrfs_extent_item */
|
||||
u64 inline_refs = 0; /* found total inline refs */
|
||||
|
||||
if (key->type == BTRFS_METADATA_ITEM_KEY &&
|
||||
!btrfs_fs_incompat(fs_info, SKINNY_METADATA)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid key type, METADATA_ITEM type invalid when SKINNY_METADATA feature disabled");
|
||||
return -EUCLEAN;
|
||||
}
|
||||
/* key->objectid is the bytenr for both key types */
|
||||
if (!IS_ALIGNED(key->objectid, fs_info->sectorsize)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid key objectid, have %llu expect to be aligned to %u",
|
||||
key->objectid, fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/* key->offset is tree level for METADATA_ITEM_KEY */
|
||||
if (key->type == BTRFS_METADATA_ITEM_KEY &&
|
||||
key->offset >= BTRFS_MAX_LEVEL) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid tree level, have %llu expect [0, %u]",
|
||||
key->offset, BTRFS_MAX_LEVEL - 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/*
|
||||
* EXTENT/METADATA_ITEM consists of:
|
||||
* 1) One btrfs_extent_item
|
||||
* Records the total refs, type and generation of the extent.
|
||||
*
|
||||
* 2) One btrfs_tree_block_info (for EXTENT_ITEM and tree backref only)
|
||||
* Records the first key and level of the tree block.
|
||||
*
|
||||
* 2) Zero or more btrfs_extent_inline_ref(s)
|
||||
* Each inline ref has one btrfs_extent_inline_ref shows:
|
||||
* 2.1) The ref type, one of the 4
|
||||
* TREE_BLOCK_REF Tree block only
|
||||
* SHARED_BLOCK_REF Tree block only
|
||||
* EXTENT_DATA_REF Data only
|
||||
* SHARED_DATA_REF Data only
|
||||
* 2.2) Ref type specific data
|
||||
* Either using btrfs_extent_inline_ref::offset, or specific
|
||||
* data structure.
|
||||
*/
|
||||
if (item_size < sizeof(*ei)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid item size, have %u expect [%zu, %u)",
|
||||
item_size, sizeof(*ei),
|
||||
BTRFS_LEAF_DATA_SIZE(fs_info));
|
||||
return -EUCLEAN;
|
||||
}
|
||||
end = item_size + btrfs_item_ptr_offset(leaf, slot);
|
||||
|
||||
/* Checks against extent_item */
|
||||
ei = btrfs_item_ptr(leaf, slot, struct btrfs_extent_item);
|
||||
flags = btrfs_extent_flags(leaf, ei);
|
||||
total_refs = btrfs_extent_refs(leaf, ei);
|
||||
generation = btrfs_extent_generation(leaf, ei);
|
||||
if (generation > btrfs_super_generation(fs_info->super_copy) + 1) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid generation, have %llu expect (0, %llu]",
|
||||
generation,
|
||||
btrfs_super_generation(fs_info->super_copy) + 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (!is_power_of_2(flags & (BTRFS_EXTENT_FLAG_DATA |
|
||||
BTRFS_EXTENT_FLAG_TREE_BLOCK))) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent flag, have 0x%llx expect 1 bit set in 0x%llx",
|
||||
flags, BTRFS_EXTENT_FLAG_DATA |
|
||||
BTRFS_EXTENT_FLAG_TREE_BLOCK);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
is_tree_block = !!(flags & BTRFS_EXTENT_FLAG_TREE_BLOCK);
|
||||
if (is_tree_block) {
|
||||
if (key->type == BTRFS_EXTENT_ITEM_KEY &&
|
||||
key->offset != fs_info->nodesize) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent length, have %llu expect %u",
|
||||
key->offset, fs_info->nodesize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
} else {
|
||||
if (key->type != BTRFS_EXTENT_ITEM_KEY) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid key type, have %u expect %u for data backref",
|
||||
key->type, BTRFS_EXTENT_ITEM_KEY);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (!IS_ALIGNED(key->offset, fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent length, have %llu expect aligned to %u",
|
||||
key->offset, fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
}
|
||||
ptr = (unsigned long)(struct btrfs_extent_item *)(ei + 1);
|
||||
|
||||
/* Check the special case of btrfs_tree_block_info */
|
||||
if (is_tree_block && key->type != BTRFS_METADATA_ITEM_KEY) {
|
||||
struct btrfs_tree_block_info *info;
|
||||
|
||||
info = (struct btrfs_tree_block_info *)ptr;
|
||||
if (btrfs_tree_block_level(leaf, info) >= BTRFS_MAX_LEVEL) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid tree block info level, have %u expect [0, %u]",
|
||||
btrfs_tree_block_level(leaf, info),
|
||||
BTRFS_MAX_LEVEL - 1);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
ptr = (unsigned long)(struct btrfs_tree_block_info *)(info + 1);
|
||||
}
|
||||
|
||||
/* Check inline refs */
|
||||
while (ptr < end) {
|
||||
struct btrfs_extent_inline_ref *iref;
|
||||
struct btrfs_extent_data_ref *dref;
|
||||
struct btrfs_shared_data_ref *sref;
|
||||
u64 dref_offset;
|
||||
u64 inline_offset;
|
||||
u8 inline_type;
|
||||
|
||||
if (ptr + sizeof(*iref) > end) {
|
||||
extent_err(leaf, slot,
|
||||
"inline ref item overflows extent item, ptr %lu iref size %zu end %lu",
|
||||
ptr, sizeof(*iref), end);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
iref = (struct btrfs_extent_inline_ref *)ptr;
|
||||
inline_type = btrfs_extent_inline_ref_type(leaf, iref);
|
||||
inline_offset = btrfs_extent_inline_ref_offset(leaf, iref);
|
||||
if (ptr + btrfs_extent_inline_ref_size(inline_type) > end) {
|
||||
extent_err(leaf, slot,
|
||||
"inline ref item overflows extent item, ptr %lu iref size %u end %lu",
|
||||
ptr, inline_type, end);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
switch (inline_type) {
|
||||
/* inline_offset is subvolid of the owner, no need to check */
|
||||
case BTRFS_TREE_BLOCK_REF_KEY:
|
||||
inline_refs++;
|
||||
break;
|
||||
/* Contains parent bytenr */
|
||||
case BTRFS_SHARED_BLOCK_REF_KEY:
|
||||
if (!IS_ALIGNED(inline_offset, fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid tree parent bytenr, have %llu expect aligned to %u",
|
||||
inline_offset, fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
inline_refs++;
|
||||
break;
|
||||
/*
|
||||
* Contains owner subvolid, owner key objectid, adjusted offset.
|
||||
* The only obvious corruption can happen in that offset.
|
||||
*/
|
||||
case BTRFS_EXTENT_DATA_REF_KEY:
|
||||
dref = (struct btrfs_extent_data_ref *)(&iref->offset);
|
||||
dref_offset = btrfs_extent_data_ref_offset(leaf, dref);
|
||||
if (!IS_ALIGNED(dref_offset, fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid data ref offset, have %llu expect aligned to %u",
|
||||
dref_offset, fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
inline_refs += btrfs_extent_data_ref_count(leaf, dref);
|
||||
break;
|
||||
/* Contains parent bytenr and ref count */
|
||||
case BTRFS_SHARED_DATA_REF_KEY:
|
||||
sref = (struct btrfs_shared_data_ref *)(iref + 1);
|
||||
if (!IS_ALIGNED(inline_offset, fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid data parent bytenr, have %llu expect aligned to %u",
|
||||
inline_offset, fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
inline_refs += btrfs_shared_data_ref_count(leaf, sref);
|
||||
break;
|
||||
default:
|
||||
extent_err(leaf, slot, "unknown inline ref type: %u",
|
||||
inline_type);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
ptr += btrfs_extent_inline_ref_size(inline_type);
|
||||
}
|
||||
/* No padding is allowed */
|
||||
if (ptr != end) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent item size, padding bytes found");
|
||||
return -EUCLEAN;
|
||||
}
|
||||
|
||||
/* Finally, check the inline refs against total refs */
|
||||
if (inline_refs > total_refs) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent refs, have %llu expect >= inline %llu",
|
||||
total_refs, inline_refs);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int check_simple_keyed_refs(struct extent_buffer *leaf,
|
||||
struct btrfs_key *key, int slot)
|
||||
{
|
||||
u32 expect_item_size = 0;
|
||||
|
||||
if (key->type == BTRFS_SHARED_DATA_REF_KEY)
|
||||
expect_item_size = sizeof(struct btrfs_shared_data_ref);
|
||||
|
||||
if (btrfs_item_size_nr(leaf, slot) != expect_item_size) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid item size, have %u expect %u for key type %u",
|
||||
btrfs_item_size_nr(leaf, slot),
|
||||
expect_item_size, key->type);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (!IS_ALIGNED(key->objectid, leaf->fs_info->sectorsize)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid key objectid for shared block ref, have %llu expect aligned to %u",
|
||||
key->objectid, leaf->fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (key->type != BTRFS_TREE_BLOCK_REF_KEY &&
|
||||
!IS_ALIGNED(key->offset, leaf->fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid tree parent bytenr, have %llu expect aligned to %u",
|
||||
key->offset, leaf->fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int check_extent_data_ref(struct extent_buffer *leaf,
|
||||
struct btrfs_key *key, int slot)
|
||||
{
|
||||
struct btrfs_extent_data_ref *dref;
|
||||
unsigned long ptr = btrfs_item_ptr_offset(leaf, slot);
|
||||
const unsigned long end = ptr + btrfs_item_size_nr(leaf, slot);
|
||||
|
||||
if (btrfs_item_size_nr(leaf, slot) % sizeof(*dref) != 0) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid item size, have %u expect aligned to %zu for key type %u",
|
||||
btrfs_item_size_nr(leaf, slot),
|
||||
sizeof(*dref), key->type);
|
||||
}
|
||||
if (!IS_ALIGNED(key->objectid, leaf->fs_info->sectorsize)) {
|
||||
generic_err(leaf, slot,
|
||||
"invalid key objectid for shared block ref, have %llu expect aligned to %u",
|
||||
key->objectid, leaf->fs_info->sectorsize);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
for (; ptr < end; ptr += sizeof(*dref)) {
|
||||
u64 root_objectid;
|
||||
u64 owner;
|
||||
u64 offset;
|
||||
u64 hash;
|
||||
|
||||
dref = (struct btrfs_extent_data_ref *)ptr;
|
||||
root_objectid = btrfs_extent_data_ref_root(leaf, dref);
|
||||
owner = btrfs_extent_data_ref_objectid(leaf, dref);
|
||||
offset = btrfs_extent_data_ref_offset(leaf, dref);
|
||||
hash = hash_extent_data_ref(root_objectid, owner, offset);
|
||||
if (hash != key->offset) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent data ref hash, item has 0x%016llx key has 0x%016llx",
|
||||
hash, key->offset);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
if (!IS_ALIGNED(offset, leaf->fs_info->sectorsize)) {
|
||||
extent_err(leaf, slot,
|
||||
"invalid extent data backref offset, have %llu expect aligned to %u",
|
||||
offset, leaf->fs_info->sectorsize);
|
||||
}
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Common point to switch the item-specific validation.
|
||||
*/
|
||||
|
@ -856,6 +1267,21 @@ static int check_leaf_item(struct extent_buffer *leaf,
|
|||
case BTRFS_INODE_ITEM_KEY:
|
||||
ret = check_inode_item(leaf, key, slot);
|
||||
break;
|
||||
case BTRFS_ROOT_ITEM_KEY:
|
||||
ret = check_root_item(leaf, key, slot);
|
||||
break;
|
||||
case BTRFS_EXTENT_ITEM_KEY:
|
||||
case BTRFS_METADATA_ITEM_KEY:
|
||||
ret = check_extent_item(leaf, key, slot);
|
||||
break;
|
||||
case BTRFS_TREE_BLOCK_REF_KEY:
|
||||
case BTRFS_SHARED_DATA_REF_KEY:
|
||||
case BTRFS_SHARED_BLOCK_REF_KEY:
|
||||
ret = check_simple_keyed_refs(leaf, key, slot);
|
||||
break;
|
||||
case BTRFS_EXTENT_DATA_REF_KEY:
|
||||
ret = check_extent_data_ref(leaf, key, slot);
|
||||
break;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
@ -899,6 +1325,12 @@ static int check_leaf(struct extent_buffer *leaf, bool check_item_data)
|
|||
owner);
|
||||
return -EUCLEAN;
|
||||
}
|
||||
/* Unknown tree */
|
||||
if (owner == 0) {
|
||||
generic_err(leaf, 0,
|
||||
"invalid owner, root 0 is not defined");
|
||||
return -EUCLEAN;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -8,6 +8,7 @@
|
|||
#include <linux/blkdev.h>
|
||||
#include <linux/list_sort.h>
|
||||
#include <linux/iversion.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "tree-log.h"
|
||||
#include "disk-io.h"
|
||||
|
@ -24,10 +25,12 @@
|
|||
* LOG_INODE_EXISTS means to log just enough to recreate the inode
|
||||
* during log replay
|
||||
*/
|
||||
#define LOG_INODE_ALL 0
|
||||
#define LOG_INODE_EXISTS 1
|
||||
#define LOG_OTHER_INODE 2
|
||||
#define LOG_OTHER_INODE_ALL 3
|
||||
enum {
|
||||
LOG_INODE_ALL,
|
||||
LOG_INODE_EXISTS,
|
||||
LOG_OTHER_INODE,
|
||||
LOG_OTHER_INODE_ALL,
|
||||
};
|
||||
|
||||
/*
|
||||
* directory trouble cases
|
||||
|
@ -81,10 +84,12 @@
|
|||
* The last stage is to deal with directories and links and extents
|
||||
* and all the other fun semantics
|
||||
*/
|
||||
#define LOG_WALK_PIN_ONLY 0
|
||||
#define LOG_WALK_REPLAY_INODES 1
|
||||
#define LOG_WALK_REPLAY_DIR_INDEX 2
|
||||
#define LOG_WALK_REPLAY_ALL 3
|
||||
enum {
|
||||
LOG_WALK_PIN_ONLY,
|
||||
LOG_WALK_REPLAY_INODES,
|
||||
LOG_WALK_REPLAY_DIR_INDEX,
|
||||
LOG_WALK_REPLAY_ALL,
|
||||
};
|
||||
|
||||
static int btrfs_log_inode(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root, struct btrfs_inode *inode,
|
||||
|
@ -188,10 +193,6 @@ static int join_running_log_trans(struct btrfs_root *root)
|
|||
{
|
||||
int ret = -ENOENT;
|
||||
|
||||
smp_mb();
|
||||
if (!root->log_root)
|
||||
return -ENOENT;
|
||||
|
||||
mutex_lock(&root->log_mutex);
|
||||
if (root->log_root) {
|
||||
ret = 0;
|
||||
|
@ -505,7 +506,7 @@ insert:
|
|||
ino_size != 0) {
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
btrfs_init_map_token(&token, dst_eb);
|
||||
btrfs_set_token_inode_size(dst_eb, dst_item,
|
||||
ino_size, &token);
|
||||
}
|
||||
|
@ -967,7 +968,7 @@ static noinline int backref_in_log(struct btrfs_root *log,
|
|||
if (btrfs_find_name_in_ext_backref(path->nodes[0],
|
||||
path->slots[0],
|
||||
ref_objectid,
|
||||
name, namelen, NULL))
|
||||
name, namelen))
|
||||
match = 1;
|
||||
|
||||
goto out;
|
||||
|
@ -1266,12 +1267,12 @@ again:
|
|||
goto out;
|
||||
|
||||
if (key->type == BTRFS_INODE_EXTREF_KEY)
|
||||
ret = btrfs_find_name_in_ext_backref(log_eb, log_slot,
|
||||
parent_id, name,
|
||||
namelen, NULL);
|
||||
ret = !!btrfs_find_name_in_ext_backref(log_eb, log_slot,
|
||||
parent_id, name,
|
||||
namelen);
|
||||
else
|
||||
ret = btrfs_find_name_in_backref(log_eb, log_slot, name,
|
||||
namelen, NULL);
|
||||
ret = !!btrfs_find_name_in_backref(log_eb, log_slot,
|
||||
name, namelen);
|
||||
|
||||
if (!ret) {
|
||||
struct inode *dir;
|
||||
|
@ -1333,12 +1334,11 @@ static int btrfs_inode_ref_exists(struct inode *inode, struct inode *dir,
|
|||
goto out;
|
||||
}
|
||||
if (key.type == BTRFS_INODE_EXTREF_KEY)
|
||||
ret = btrfs_find_name_in_ext_backref(path->nodes[0],
|
||||
path->slots[0], parent_id,
|
||||
name, namelen, NULL);
|
||||
ret = !!btrfs_find_name_in_ext_backref(path->nodes[0],
|
||||
path->slots[0], parent_id, name, namelen);
|
||||
else
|
||||
ret = btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
|
||||
name, namelen, NULL);
|
||||
ret = !!btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
|
||||
name, namelen);
|
||||
|
||||
out:
|
||||
btrfs_free_path(path);
|
||||
|
@ -3842,7 +3842,7 @@ static void fill_inode_item(struct btrfs_trans_handle *trans,
|
|||
{
|
||||
struct btrfs_map_token token;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
|
||||
if (log_inode_only) {
|
||||
/* set the generation to zero so the recover code
|
||||
|
@ -4302,8 +4302,6 @@ static int log_one_extent(struct btrfs_trans_handle *trans,
|
|||
if (ret)
|
||||
return ret;
|
||||
|
||||
btrfs_init_map_token(&token);
|
||||
|
||||
ret = __btrfs_drop_extents(trans, log, &inode->vfs_inode, path, em->start,
|
||||
em->start + em->len, NULL, 0, 1,
|
||||
sizeof(*fi), &extent_inserted);
|
||||
|
@ -4321,6 +4319,7 @@ static int log_one_extent(struct btrfs_trans_handle *trans,
|
|||
return ret;
|
||||
}
|
||||
leaf = path->nodes[0];
|
||||
btrfs_init_map_token(&token, leaf);
|
||||
fi = btrfs_item_ptr(leaf, path->slots[0],
|
||||
struct btrfs_file_extent_item);
|
||||
|
||||
|
@ -6233,7 +6232,7 @@ int btrfs_recover_log_trees(struct btrfs_root *log_root_tree)
|
|||
struct btrfs_fs_info *fs_info = log_root_tree->fs_info;
|
||||
struct walk_control wc = {
|
||||
.process_func = process_one_buffer,
|
||||
.stage = 0,
|
||||
.stage = LOG_WALK_PIN_ONLY,
|
||||
};
|
||||
|
||||
path = btrfs_alloc_path();
|
||||
|
|
|
@ -14,6 +14,7 @@
|
|||
#include <linux/semaphore.h>
|
||||
#include <linux/uuid.h>
|
||||
#include <linux/list_sort.h>
|
||||
#include "misc.h"
|
||||
#include "ctree.h"
|
||||
#include "extent_map.h"
|
||||
#include "disk-io.h"
|
||||
|
@ -24,11 +25,11 @@
|
|||
#include "async-thread.h"
|
||||
#include "check-integrity.h"
|
||||
#include "rcu-string.h"
|
||||
#include "math.h"
|
||||
#include "dev-replace.h"
|
||||
#include "sysfs.h"
|
||||
#include "tree-checker.h"
|
||||
#include "space-info.h"
|
||||
#include "block-group.h"
|
||||
|
||||
const struct btrfs_raid_attr btrfs_raid_array[BTRFS_NR_RAID_TYPES] = {
|
||||
[BTRFS_RAID_RAID10] = {
|
||||
|
@ -190,7 +191,6 @@ out_overflow:;
|
|||
|
||||
static int init_first_rw_device(struct btrfs_trans_handle *trans);
|
||||
static int btrfs_relocate_sys_chunks(struct btrfs_fs_info *fs_info);
|
||||
static void __btrfs_reset_dev_stats(struct btrfs_device *dev);
|
||||
static void btrfs_dev_stat_print_on_error(struct btrfs_device *dev);
|
||||
static void btrfs_dev_stat_print_on_load(struct btrfs_device *device);
|
||||
static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
||||
|
@ -358,19 +358,6 @@ static void free_fs_devices(struct btrfs_fs_devices *fs_devices)
|
|||
kfree(fs_devices);
|
||||
}
|
||||
|
||||
static void btrfs_kobject_uevent(struct block_device *bdev,
|
||||
enum kobject_action action)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = kobject_uevent(&disk_to_dev(bdev->bd_disk)->kobj, action);
|
||||
if (ret)
|
||||
pr_warn("BTRFS: Sending event '%d' to kobject: '%s' (%p): failed\n",
|
||||
action,
|
||||
kobject_name(&disk_to_dev(bdev->bd_disk)->kobj),
|
||||
&disk_to_dev(bdev->bd_disk)->kobj);
|
||||
}
|
||||
|
||||
void __exit btrfs_cleanup_fs_uuids(void)
|
||||
{
|
||||
struct btrfs_fs_devices *fs_devices;
|
||||
|
@ -1128,6 +1115,7 @@ static struct btrfs_fs_devices *clone_fs_devices(struct btrfs_fs_devices *orig)
|
|||
struct btrfs_fs_devices *fs_devices;
|
||||
struct btrfs_device *device;
|
||||
struct btrfs_device *orig_dev;
|
||||
int ret = 0;
|
||||
|
||||
fs_devices = alloc_fs_devices(orig->fsid, NULL);
|
||||
if (IS_ERR(fs_devices))
|
||||
|
@ -1141,8 +1129,10 @@ static struct btrfs_fs_devices *clone_fs_devices(struct btrfs_fs_devices *orig)
|
|||
|
||||
device = btrfs_alloc_device(NULL, &orig_dev->devid,
|
||||
orig_dev->uuid);
|
||||
if (IS_ERR(device))
|
||||
if (IS_ERR(device)) {
|
||||
ret = PTR_ERR(device);
|
||||
goto error;
|
||||
}
|
||||
|
||||
/*
|
||||
* This is ok to do without rcu read locked because we hold the
|
||||
|
@ -1153,6 +1143,7 @@ static struct btrfs_fs_devices *clone_fs_devices(struct btrfs_fs_devices *orig)
|
|||
GFP_KERNEL);
|
||||
if (!name) {
|
||||
btrfs_free_device(device);
|
||||
ret = -ENOMEM;
|
||||
goto error;
|
||||
}
|
||||
rcu_assign_pointer(device->name, name);
|
||||
|
@ -1167,7 +1158,7 @@ static struct btrfs_fs_devices *clone_fs_devices(struct btrfs_fs_devices *orig)
|
|||
error:
|
||||
mutex_unlock(&orig->device_list_mutex);
|
||||
free_fs_devices(fs_devices);
|
||||
return ERR_PTR(-ENOMEM);
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -1551,9 +1542,16 @@ static bool contains_pending_extent(struct btrfs_device *device, u64 *start,
|
|||
* @len is used to store the size of the free space that we find.
|
||||
* But if we don't find suitable free space, it is used to store the size of
|
||||
* the max free space.
|
||||
*
|
||||
* NOTE: This function will search *commit* root of device tree, and does extra
|
||||
* check to ensure dev extents are not double allocated.
|
||||
* This makes the function safe to allocate dev extents but may not report
|
||||
* correct usable device space, as device extent freed in current transaction
|
||||
* is not reported as avaiable.
|
||||
*/
|
||||
int find_free_dev_extent_start(struct btrfs_device *device, u64 num_bytes,
|
||||
u64 search_start, u64 *start, u64 *len)
|
||||
static int find_free_dev_extent_start(struct btrfs_device *device,
|
||||
u64 num_bytes, u64 search_start, u64 *start,
|
||||
u64 *len)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = device->fs_info;
|
||||
struct btrfs_root *root = fs_info->dev_root;
|
||||
|
@ -1855,7 +1853,12 @@ static noinline int find_next_devid(struct btrfs_fs_info *fs_info,
|
|||
if (ret < 0)
|
||||
goto error;
|
||||
|
||||
BUG_ON(ret == 0); /* Corruption */
|
||||
if (ret == 0) {
|
||||
/* Corruption */
|
||||
btrfs_err(fs_info, "corrupted chunk tree devid -1 matched");
|
||||
ret = -EUCLEAN;
|
||||
goto error;
|
||||
}
|
||||
|
||||
ret = btrfs_previous_item(fs_info->chunk_root, path,
|
||||
BTRFS_DEV_ITEMS_OBJECTID,
|
||||
|
@ -2686,22 +2689,14 @@ int btrfs_init_new_device(struct btrfs_fs_info *fs_info, const char *device_path
|
|||
}
|
||||
|
||||
if (seeding_dev) {
|
||||
char fsid_buf[BTRFS_UUID_UNPARSED_SIZE];
|
||||
|
||||
ret = btrfs_finish_sprout(trans);
|
||||
if (ret) {
|
||||
btrfs_abort_transaction(trans, ret);
|
||||
goto error_sysfs;
|
||||
}
|
||||
|
||||
/* Sprouting would change fsid of the mounted root,
|
||||
* so rename the fsid on the sysfs
|
||||
*/
|
||||
snprintf(fsid_buf, BTRFS_UUID_UNPARSED_SIZE, "%pU",
|
||||
fs_info->fs_devices->fsid);
|
||||
if (kobject_rename(&fs_devices->fsid_kobj, fsid_buf))
|
||||
btrfs_warn(fs_info,
|
||||
"sysfs: failed to create fsid for sprout");
|
||||
btrfs_sysfs_update_sprout_fsid(fs_devices,
|
||||
fs_info->fs_devices->fsid);
|
||||
}
|
||||
|
||||
ret = btrfs_commit_transaction(trans);
|
||||
|
@ -3076,10 +3071,6 @@ static int btrfs_relocate_chunk(struct btrfs_fs_info *fs_info, u64 chunk_offset)
|
|||
*/
|
||||
lockdep_assert_held(&fs_info->delete_unused_bgs_mutex);
|
||||
|
||||
ret = btrfs_can_relocate(fs_info, chunk_offset);
|
||||
if (ret)
|
||||
return -ENOSPC;
|
||||
|
||||
/* step one, relocate all the extents inside this chunk */
|
||||
btrfs_scrub_pause(fs_info);
|
||||
ret = btrfs_relocate_block_group(fs_info, chunk_offset);
|
||||
|
@ -6011,7 +6002,6 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
|||
{
|
||||
struct extent_map *em;
|
||||
struct map_lookup *map;
|
||||
u64 offset;
|
||||
u64 stripe_offset;
|
||||
u64 stripe_nr;
|
||||
u64 stripe_len;
|
||||
|
@ -6042,11 +6032,10 @@ static int __btrfs_map_block(struct btrfs_fs_info *fs_info,
|
|||
return ret;
|
||||
|
||||
em = btrfs_get_chunk_map(fs_info, logical, *length);
|
||||
ASSERT(em);
|
||||
ASSERT(!IS_ERR(em));
|
||||
map = em->map_lookup;
|
||||
|
||||
*length = geom.len;
|
||||
offset = geom.offset;
|
||||
stripe_len = geom.stripe_len;
|
||||
stripe_nr = geom.stripe_nr;
|
||||
stripe_offset = geom.stripe_offset;
|
||||
|
@ -7296,18 +7285,32 @@ void btrfs_init_devices_late(struct btrfs_fs_info *fs_info)
|
|||
}
|
||||
}
|
||||
|
||||
static void __btrfs_reset_dev_stats(struct btrfs_device *dev)
|
||||
static u64 btrfs_dev_stats_value(const struct extent_buffer *eb,
|
||||
const struct btrfs_dev_stats_item *ptr,
|
||||
int index)
|
||||
{
|
||||
int i;
|
||||
u64 val;
|
||||
|
||||
for (i = 0; i < BTRFS_DEV_STAT_VALUES_MAX; i++)
|
||||
btrfs_dev_stat_reset(dev, i);
|
||||
read_extent_buffer(eb, &val,
|
||||
offsetof(struct btrfs_dev_stats_item, values) +
|
||||
((unsigned long)ptr) + (index * sizeof(u64)),
|
||||
sizeof(val));
|
||||
return val;
|
||||
}
|
||||
|
||||
static void btrfs_set_dev_stats_value(struct extent_buffer *eb,
|
||||
struct btrfs_dev_stats_item *ptr,
|
||||
int index, u64 val)
|
||||
{
|
||||
write_extent_buffer(eb, &val,
|
||||
offsetof(struct btrfs_dev_stats_item, values) +
|
||||
((unsigned long)ptr) + (index * sizeof(u64)),
|
||||
sizeof(val));
|
||||
}
|
||||
|
||||
int btrfs_init_dev_stats(struct btrfs_fs_info *fs_info)
|
||||
{
|
||||
struct btrfs_key key;
|
||||
struct btrfs_key found_key;
|
||||
struct btrfs_root *dev_root = fs_info->dev_root;
|
||||
struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
|
||||
struct extent_buffer *eb;
|
||||
|
@ -7318,10 +7321,8 @@ int btrfs_init_dev_stats(struct btrfs_fs_info *fs_info)
|
|||
int i;
|
||||
|
||||
path = btrfs_alloc_path();
|
||||
if (!path) {
|
||||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
if (!path)
|
||||
return -ENOMEM;
|
||||
|
||||
mutex_lock(&fs_devices->device_list_mutex);
|
||||
list_for_each_entry(device, &fs_devices->devices, dev_list) {
|
||||
|
@ -7333,14 +7334,14 @@ int btrfs_init_dev_stats(struct btrfs_fs_info *fs_info)
|
|||
key.offset = device->devid;
|
||||
ret = btrfs_search_slot(NULL, dev_root, &key, path, 0, 0);
|
||||
if (ret) {
|
||||
__btrfs_reset_dev_stats(device);
|
||||
for (i = 0; i < BTRFS_DEV_STAT_VALUES_MAX; i++)
|
||||
btrfs_dev_stat_set(device, i, 0);
|
||||
device->dev_stats_valid = 1;
|
||||
btrfs_release_path(path);
|
||||
continue;
|
||||
}
|
||||
slot = path->slots[0];
|
||||
eb = path->nodes[0];
|
||||
btrfs_item_key_to_cpu(eb, &found_key, slot);
|
||||
item_size = btrfs_item_size_nr(eb, slot);
|
||||
|
||||
ptr = btrfs_item_ptr(eb, slot,
|
||||
|
@ -7351,7 +7352,7 @@ int btrfs_init_dev_stats(struct btrfs_fs_info *fs_info)
|
|||
btrfs_dev_stat_set(device, i,
|
||||
btrfs_dev_stats_value(eb, ptr, i));
|
||||
else
|
||||
btrfs_dev_stat_reset(device, i);
|
||||
btrfs_dev_stat_set(device, i, 0);
|
||||
}
|
||||
|
||||
device->dev_stats_valid = 1;
|
||||
|
@ -7360,7 +7361,6 @@ int btrfs_init_dev_stats(struct btrfs_fs_info *fs_info)
|
|||
}
|
||||
mutex_unlock(&fs_devices->device_list_mutex);
|
||||
|
||||
out:
|
||||
btrfs_free_path(path);
|
||||
return ret < 0 ? ret : 0;
|
||||
}
|
||||
|
@ -7534,7 +7534,7 @@ int btrfs_get_dev_stats(struct btrfs_fs_info *fs_info,
|
|||
stats->values[i] =
|
||||
btrfs_dev_stat_read_and_reset(dev, i);
|
||||
else
|
||||
btrfs_dev_stat_reset(dev, i);
|
||||
btrfs_dev_stat_set(dev, i, 0);
|
||||
}
|
||||
} else {
|
||||
for (i = 0; i < BTRFS_DEV_STAT_VALUES_MAX; i++)
|
||||
|
|
|
@ -82,7 +82,6 @@ struct btrfs_device {
|
|||
|
||||
unsigned long dev_state;
|
||||
blk_status_t last_flush_error;
|
||||
int flush_bio_sent;
|
||||
|
||||
#ifdef __BTRFS_NEED_DEVICE_DATA_ORDERED
|
||||
seqcount_t data_seqcount;
|
||||
|
@ -475,8 +474,6 @@ int btrfs_cancel_balance(struct btrfs_fs_info *fs_info);
|
|||
int btrfs_create_uuid_tree(struct btrfs_fs_info *fs_info);
|
||||
int btrfs_check_uuid_tree(struct btrfs_fs_info *fs_info);
|
||||
int btrfs_chunk_readonly(struct btrfs_fs_info *fs_info, u64 chunk_offset);
|
||||
int find_free_dev_extent_start(struct btrfs_device *device, u64 num_bytes,
|
||||
u64 search_start, u64 *start, u64 *max_avail);
|
||||
int find_free_dev_extent(struct btrfs_device *device, u64 num_bytes,
|
||||
u64 *start, u64 *max_avail);
|
||||
void btrfs_dev_stat_inc_and_print(struct btrfs_device *dev, int index);
|
||||
|
@ -550,12 +547,6 @@ static inline void btrfs_dev_stat_set(struct btrfs_device *dev,
|
|||
atomic_inc(&dev->dev_stats_ccnt);
|
||||
}
|
||||
|
||||
static inline void btrfs_dev_stat_reset(struct btrfs_device *dev,
|
||||
int index)
|
||||
{
|
||||
btrfs_dev_stat_set(dev, index, 0);
|
||||
}
|
||||
|
||||
/*
|
||||
* Convert block group flags (BTRFS_BLOCK_GROUP_*) to btrfs_raid_types, which
|
||||
* can be used as index to access btrfs_raid_array[].
|
||||
|
|
|
@ -418,14 +418,6 @@ next:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static unsigned int zlib_set_level(unsigned int level)
|
||||
{
|
||||
if (!level)
|
||||
return BTRFS_ZLIB_DEFAULT_LEVEL;
|
||||
|
||||
return min_t(unsigned int, level, 9);
|
||||
}
|
||||
|
||||
const struct btrfs_compress_op btrfs_zlib_compress = {
|
||||
.init_workspace_manager = zlib_init_workspace_manager,
|
||||
.cleanup_workspace_manager = zlib_cleanup_workspace_manager,
|
||||
|
@ -436,5 +428,6 @@ const struct btrfs_compress_op btrfs_zlib_compress = {
|
|||
.compress_pages = zlib_compress_pages,
|
||||
.decompress_bio = zlib_decompress_bio,
|
||||
.decompress = zlib_decompress,
|
||||
.set_level = zlib_set_level,
|
||||
.max_level = 9,
|
||||
.default_level = BTRFS_ZLIB_DEFAULT_LEVEL,
|
||||
};
|
||||
|
|
|
@ -17,6 +17,7 @@
|
|||
#include <linux/sched.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/zstd.h>
|
||||
#include "misc.h"
|
||||
#include "compression.h"
|
||||
#include "ctree.h"
|
||||
|
||||
|
@ -710,14 +711,6 @@ finish:
|
|||
return ret;
|
||||
}
|
||||
|
||||
static unsigned int zstd_set_level(unsigned int level)
|
||||
{
|
||||
if (!level)
|
||||
return ZSTD_BTRFS_DEFAULT_LEVEL;
|
||||
|
||||
return min_t(unsigned int, level, ZSTD_BTRFS_MAX_LEVEL);
|
||||
}
|
||||
|
||||
const struct btrfs_compress_op btrfs_zstd_compress = {
|
||||
.init_workspace_manager = zstd_init_workspace_manager,
|
||||
.cleanup_workspace_manager = zstd_cleanup_workspace_manager,
|
||||
|
@ -728,5 +721,6 @@ const struct btrfs_compress_op btrfs_zstd_compress = {
|
|||
.compress_pages = zstd_compress_pages,
|
||||
.decompress_bio = zstd_decompress_bio,
|
||||
.decompress = zstd_decompress,
|
||||
.set_level = zstd_set_level,
|
||||
.max_level = ZSTD_BTRFS_MAX_LEVEL,
|
||||
.default_level = ZSTD_BTRFS_DEFAULT_LEVEL,
|
||||
};
|
||||
|
|
|
@ -1088,6 +1088,7 @@ TRACE_EVENT(btrfs_trigger_flush,
|
|||
{ FLUSH_DELAYED_REFS, "FLUSH_ELAYED_REFS"}, \
|
||||
{ ALLOC_CHUNK, "ALLOC_CHUNK"}, \
|
||||
{ ALLOC_CHUNK_FORCE, "ALLOC_CHUNK_FORCE"}, \
|
||||
{ RUN_DELAYED_IPUTS, "RUN_DELAYED_IPUTS"}, \
|
||||
{ COMMIT_TRANS, "COMMIT_TRANS"})
|
||||
|
||||
TRACE_EVENT(btrfs_flush_space,
|
||||
|
@ -2086,8 +2087,6 @@ DEFINE_BTRFS_LOCK_EVENT(btrfs_tree_read_unlock);
|
|||
DEFINE_BTRFS_LOCK_EVENT(btrfs_tree_read_unlock_blocking);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_set_lock_blocking_read);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_set_lock_blocking_write);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_clear_lock_blocking_read);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_clear_lock_blocking_write);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_try_tree_read_lock);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_try_tree_write_lock);
|
||||
DEFINE_BTRFS_LOCK_EVENT(btrfs_tree_read_lock_atomic);
|
||||
|
|
|
@ -665,7 +665,12 @@ struct btrfs_ioctl_get_dev_stats {
|
|||
/* out values: */
|
||||
__u64 values[BTRFS_DEV_STAT_VALUES_MAX];
|
||||
|
||||
__u64 unused[128 - 2 - BTRFS_DEV_STAT_VALUES_MAX]; /* pad to 1k */
|
||||
/*
|
||||
* This pads the struct to 1032 bytes. It was originally meant to pad to
|
||||
* 1024 bytes, but when adding the flags field, the padding calculation
|
||||
* was not adjusted.
|
||||
*/
|
||||
__u64 unused[128 - 2 - BTRFS_DEV_STAT_VALUES_MAX];
|
||||
};
|
||||
|
||||
#define BTRFS_QUOTA_CTL_ENABLE 1
|
||||
|
@ -917,10 +922,8 @@ enum btrfs_err_code {
|
|||
#define BTRFS_IOC_QUOTA_RESCAN_STATUS _IOR(BTRFS_IOCTL_MAGIC, 45, \
|
||||
struct btrfs_ioctl_quota_rescan_args)
|
||||
#define BTRFS_IOC_QUOTA_RESCAN_WAIT _IO(BTRFS_IOCTL_MAGIC, 46)
|
||||
#define BTRFS_IOC_GET_FSLABEL _IOR(BTRFS_IOCTL_MAGIC, 49, \
|
||||
char[BTRFS_LABEL_SIZE])
|
||||
#define BTRFS_IOC_SET_FSLABEL _IOW(BTRFS_IOCTL_MAGIC, 50, \
|
||||
char[BTRFS_LABEL_SIZE])
|
||||
#define BTRFS_IOC_GET_FSLABEL FS_IOC_GETFSLABEL
|
||||
#define BTRFS_IOC_SET_FSLABEL FS_IOC_SETFSLABEL
|
||||
#define BTRFS_IOC_GET_DEV_STATS _IOWR(BTRFS_IOCTL_MAGIC, 52, \
|
||||
struct btrfs_ioctl_get_dev_stats)
|
||||
#define BTRFS_IOC_DEV_REPLACE _IOWR(BTRFS_IOCTL_MAGIC, 53, \
|
||||
|
|
|
@ -300,7 +300,9 @@
|
|||
#define BTRFS_CSUM_SIZE 32
|
||||
|
||||
/* csum types */
|
||||
#define BTRFS_CSUM_TYPE_CRC32 0
|
||||
enum btrfs_csum_type {
|
||||
BTRFS_CSUM_TYPE_CRC32 = 0,
|
||||
};
|
||||
|
||||
/*
|
||||
* flags definitions for directory entry item type
|
||||
|
@ -806,11 +808,6 @@ struct btrfs_dev_stats_item {
|
|||
|
||||
#define BTRFS_DEV_REPLACE_ITEM_CONT_READING_FROM_SRCDEV_MODE_ALWAYS 0
|
||||
#define BTRFS_DEV_REPLACE_ITEM_CONT_READING_FROM_SRCDEV_MODE_AVOID 1
|
||||
#define BTRFS_DEV_REPLACE_ITEM_STATE_NEVER_STARTED 0
|
||||
#define BTRFS_DEV_REPLACE_ITEM_STATE_STARTED 1
|
||||
#define BTRFS_DEV_REPLACE_ITEM_STATE_SUSPENDED 2
|
||||
#define BTRFS_DEV_REPLACE_ITEM_STATE_FINISHED 3
|
||||
#define BTRFS_DEV_REPLACE_ITEM_STATE_CANCELED 4
|
||||
|
||||
struct btrfs_dev_replace_item {
|
||||
/*
|
||||
|
|
Loading…
Reference in New Issue