f2fs: clean up lfs/adaptive mount option
This patch removes F2FS_MOUNT_ADAPTIVE and F2FS_MOUNT_LFS mount options, and add F2FS_OPTION.fs_mode with below two status to indicate filesystem mode. enum { FS_MODE_ADAPTIVE, /* use both lfs/ssr allocation */ FS_MODE_LFS, /* use lfs allocation only */ }; It can enhance code readability and fs mode's scalability. Signed-off-by: Chao Yu <yuchao0@huawei.com> Signed-off-by: Jaegeuk Kim <jaegeuk@kernel.org>
This commit is contained in:
parent
a9117eca1d
commit
b0332a0f95
|
@ -442,7 +442,7 @@ static inline void __submit_bio(struct f2fs_sb_info *sbi,
|
|||
if (type != DATA && type != NODE)
|
||||
goto submit_io;
|
||||
|
||||
if (test_opt(sbi, LFS) && current->plug)
|
||||
if (f2fs_lfs_mode(sbi) && current->plug)
|
||||
blk_finish_plug(current->plug);
|
||||
|
||||
if (F2FS_IO_ALIGNED(sbi))
|
||||
|
@ -1413,7 +1413,7 @@ int f2fs_map_blocks(struct inode *inode, struct f2fs_map_blocks *map,
|
|||
end = pgofs + maxblocks;
|
||||
|
||||
if (!create && f2fs_lookup_extent_cache(inode, pgofs, &ei)) {
|
||||
if (test_opt(sbi, LFS) && flag == F2FS_GET_BLOCK_DIO &&
|
||||
if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO &&
|
||||
map->m_may_create)
|
||||
goto next_dnode;
|
||||
|
||||
|
@ -1468,7 +1468,7 @@ next_block:
|
|||
|
||||
if (__is_valid_data_blkaddr(blkaddr)) {
|
||||
/* use out-place-update for driect IO under LFS mode */
|
||||
if (test_opt(sbi, LFS) && flag == F2FS_GET_BLOCK_DIO &&
|
||||
if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO &&
|
||||
map->m_may_create) {
|
||||
err = __allocate_data_block(&dn, map->m_seg_type);
|
||||
if (err)
|
||||
|
@ -2388,7 +2388,7 @@ bool f2fs_should_update_outplace(struct inode *inode, struct f2fs_io_info *fio)
|
|||
{
|
||||
struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
|
||||
|
||||
if (test_opt(sbi, LFS))
|
||||
if (f2fs_lfs_mode(sbi))
|
||||
return true;
|
||||
if (S_ISDIR(inode->i_mode))
|
||||
return true;
|
||||
|
|
|
@ -91,8 +91,6 @@ extern const char *f2fs_fault_name[FAULT_MAX];
|
|||
#define F2FS_MOUNT_FORCE_FG_GC 0x00004000
|
||||
#define F2FS_MOUNT_DATA_FLUSH 0x00008000
|
||||
#define F2FS_MOUNT_FAULT_INJECTION 0x00010000
|
||||
#define F2FS_MOUNT_ADAPTIVE 0x00020000
|
||||
#define F2FS_MOUNT_LFS 0x00040000
|
||||
#define F2FS_MOUNT_USRQUOTA 0x00080000
|
||||
#define F2FS_MOUNT_GRPQUOTA 0x00100000
|
||||
#define F2FS_MOUNT_PRJQUOTA 0x00200000
|
||||
|
@ -139,6 +137,7 @@ struct f2fs_mount_info {
|
|||
int whint_mode;
|
||||
int alloc_mode; /* segment allocation policy */
|
||||
int fsync_mode; /* fsync policy */
|
||||
int fs_mode; /* fs mode: LFS or ADAPTIVE */
|
||||
bool test_dummy_encryption; /* test dummy encryption */
|
||||
block_t unusable_cap; /* Amount of space allowed to be
|
||||
* unusable when disabling checkpoint
|
||||
|
@ -1173,6 +1172,11 @@ enum {
|
|||
GC_URGENT,
|
||||
};
|
||||
|
||||
enum {
|
||||
FS_MODE_ADAPTIVE, /* use both lfs/ssr allocation */
|
||||
FS_MODE_LFS, /* use lfs allocation only */
|
||||
};
|
||||
|
||||
enum {
|
||||
WHINT_MODE_OFF, /* not pass down write hints */
|
||||
WHINT_MODE_USER, /* try to pass down hints given by users */
|
||||
|
@ -3910,20 +3914,9 @@ static inline bool f2fs_hw_is_readonly(struct f2fs_sb_info *sbi)
|
|||
return false;
|
||||
}
|
||||
|
||||
|
||||
static inline void set_opt_mode(struct f2fs_sb_info *sbi, unsigned int mt)
|
||||
static inline bool f2fs_lfs_mode(struct f2fs_sb_info *sbi)
|
||||
{
|
||||
clear_opt(sbi, ADAPTIVE);
|
||||
clear_opt(sbi, LFS);
|
||||
|
||||
switch (mt) {
|
||||
case F2FS_MOUNT_ADAPTIVE:
|
||||
set_opt(sbi, ADAPTIVE);
|
||||
break;
|
||||
case F2FS_MOUNT_LFS:
|
||||
set_opt(sbi, LFS);
|
||||
break;
|
||||
}
|
||||
return F2FS_OPTION(sbi).fs_mode == FS_MODE_LFS;
|
||||
}
|
||||
|
||||
static inline bool f2fs_may_encrypt(struct inode *inode)
|
||||
|
@ -3978,7 +3971,7 @@ static inline int allow_outplace_dio(struct inode *inode,
|
|||
struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
|
||||
int rw = iov_iter_rw(iter);
|
||||
|
||||
return (test_opt(sbi, LFS) && (rw == WRITE) &&
|
||||
return (f2fs_lfs_mode(sbi) && (rw == WRITE) &&
|
||||
!block_unaligned_IO(inode, iocb, iter));
|
||||
}
|
||||
|
||||
|
@ -4000,7 +3993,7 @@ static inline bool f2fs_force_buffered_io(struct inode *inode,
|
|||
*/
|
||||
if (f2fs_sb_has_blkzoned(sbi))
|
||||
return true;
|
||||
if (test_opt(sbi, LFS) && (rw == WRITE)) {
|
||||
if (f2fs_lfs_mode(sbi) && (rw == WRITE)) {
|
||||
if (block_unaligned_IO(inode, iocb, iter))
|
||||
return true;
|
||||
if (F2FS_IO_ALIGNED(sbi))
|
||||
|
|
|
@ -1126,7 +1126,7 @@ next_dnode:
|
|||
|
||||
if (!f2fs_is_checkpointed_data(sbi, *blkaddr)) {
|
||||
|
||||
if (test_opt(sbi, LFS)) {
|
||||
if (f2fs_lfs_mode(sbi)) {
|
||||
f2fs_put_dnode(&dn);
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
|
|
@ -765,7 +765,7 @@ static int move_data_block(struct inode *inode, block_t bidx,
|
|||
struct page *page, *mpage;
|
||||
block_t newaddr;
|
||||
int err = 0;
|
||||
bool lfs_mode = test_opt(fio.sbi, LFS);
|
||||
bool lfs_mode = f2fs_lfs_mode(fio.sbi);
|
||||
|
||||
/* do not read out */
|
||||
page = f2fs_grab_cache_page(inode->i_mapping, bidx, false);
|
||||
|
|
|
@ -172,7 +172,7 @@ bool f2fs_need_SSR(struct f2fs_sb_info *sbi)
|
|||
int dent_secs = get_blocktype_secs(sbi, F2FS_DIRTY_DENTS);
|
||||
int imeta_secs = get_blocktype_secs(sbi, F2FS_DIRTY_IMETA);
|
||||
|
||||
if (test_opt(sbi, LFS))
|
||||
if (f2fs_lfs_mode(sbi))
|
||||
return false;
|
||||
if (sbi->gc_mode == GC_URGENT)
|
||||
return true;
|
||||
|
@ -1940,7 +1940,7 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi,
|
|||
unsigned int start = 0, end = -1;
|
||||
unsigned int secno, start_segno;
|
||||
bool force = (cpc->reason & CP_DISCARD);
|
||||
bool need_align = test_opt(sbi, LFS) && __is_large_section(sbi);
|
||||
bool need_align = f2fs_lfs_mode(sbi) && __is_large_section(sbi);
|
||||
|
||||
mutex_lock(&dirty_i->seglist_lock);
|
||||
|
||||
|
@ -1972,7 +1972,7 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi,
|
|||
(end - 1) <= cpc->trim_end)
|
||||
continue;
|
||||
|
||||
if (!test_opt(sbi, LFS) || !__is_large_section(sbi)) {
|
||||
if (!f2fs_lfs_mode(sbi) || !__is_large_section(sbi)) {
|
||||
f2fs_issue_discard(sbi, START_BLOCK(sbi, start),
|
||||
(end - start) << sbi->log_blocks_per_seg);
|
||||
continue;
|
||||
|
@ -2830,7 +2830,7 @@ int f2fs_trim_fs(struct f2fs_sb_info *sbi, struct fstrim_range *range)
|
|||
struct discard_policy dpolicy;
|
||||
unsigned long long trimmed = 0;
|
||||
int err = 0;
|
||||
bool need_align = test_opt(sbi, LFS) && __is_large_section(sbi);
|
||||
bool need_align = f2fs_lfs_mode(sbi) && __is_large_section(sbi);
|
||||
|
||||
if (start >= MAX_BLKADDR(sbi) || range->len < sbi->blocksize)
|
||||
return -EINVAL;
|
||||
|
@ -3193,7 +3193,7 @@ static void update_device_state(struct f2fs_io_info *fio)
|
|||
static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio)
|
||||
{
|
||||
int type = __get_segment_type(fio);
|
||||
bool keep_order = (test_opt(fio->sbi, LFS) && type == CURSEG_COLD_DATA);
|
||||
bool keep_order = (f2fs_lfs_mode(fio->sbi) && type == CURSEG_COLD_DATA);
|
||||
|
||||
if (keep_order)
|
||||
down_read(&fio->sbi->io_order_lock);
|
||||
|
@ -4678,7 +4678,7 @@ int f2fs_build_segment_manager(struct f2fs_sb_info *sbi)
|
|||
if (sm_info->rec_prefree_segments > DEF_MAX_RECLAIM_PREFREE_SEGMENTS)
|
||||
sm_info->rec_prefree_segments = DEF_MAX_RECLAIM_PREFREE_SEGMENTS;
|
||||
|
||||
if (!test_opt(sbi, LFS))
|
||||
if (!f2fs_lfs_mode(sbi))
|
||||
sm_info->ipu_policy = 1 << F2FS_IPU_FSYNC;
|
||||
sm_info->min_ipu_util = DEF_MIN_IPU_UTIL;
|
||||
sm_info->min_fsync_blocks = DEF_MIN_FSYNC_BLOCKS;
|
||||
|
|
|
@ -600,10 +600,10 @@ static int parse_options(struct super_block *sb, char *options)
|
|||
kvfree(name);
|
||||
return -EINVAL;
|
||||
}
|
||||
set_opt_mode(sbi, F2FS_MOUNT_ADAPTIVE);
|
||||
F2FS_OPTION(sbi).fs_mode = FS_MODE_ADAPTIVE;
|
||||
} else if (strlen(name) == 3 &&
|
||||
!strncmp(name, "lfs", 3)) {
|
||||
set_opt_mode(sbi, F2FS_MOUNT_LFS);
|
||||
F2FS_OPTION(sbi).fs_mode = FS_MODE_LFS;
|
||||
} else {
|
||||
kvfree(name);
|
||||
return -EINVAL;
|
||||
|
@ -904,7 +904,7 @@ static int parse_options(struct super_block *sb, char *options)
|
|||
}
|
||||
#endif
|
||||
|
||||
if (F2FS_IO_SIZE_BITS(sbi) && !test_opt(sbi, LFS)) {
|
||||
if (F2FS_IO_SIZE_BITS(sbi) && !f2fs_lfs_mode(sbi)) {
|
||||
f2fs_err(sbi, "Should set mode=lfs with %uKB-sized IO",
|
||||
F2FS_IO_SIZE_KB(sbi));
|
||||
return -EINVAL;
|
||||
|
@ -934,7 +934,7 @@ static int parse_options(struct super_block *sb, char *options)
|
|||
}
|
||||
}
|
||||
|
||||
if (test_opt(sbi, DISABLE_CHECKPOINT) && test_opt(sbi, LFS)) {
|
||||
if (test_opt(sbi, DISABLE_CHECKPOINT) && f2fs_lfs_mode(sbi)) {
|
||||
f2fs_err(sbi, "LFS not compatible with checkpoint=disable\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
@ -1500,9 +1500,9 @@ static int f2fs_show_options(struct seq_file *seq, struct dentry *root)
|
|||
seq_puts(seq, ",data_flush");
|
||||
|
||||
seq_puts(seq, ",mode=");
|
||||
if (test_opt(sbi, ADAPTIVE))
|
||||
if (F2FS_OPTION(sbi).fs_mode == FS_MODE_ADAPTIVE)
|
||||
seq_puts(seq, "adaptive");
|
||||
else if (test_opt(sbi, LFS))
|
||||
else if (F2FS_OPTION(sbi).fs_mode == FS_MODE_LFS)
|
||||
seq_puts(seq, "lfs");
|
||||
seq_printf(seq, ",active_logs=%u", F2FS_OPTION(sbi).active_logs);
|
||||
if (test_opt(sbi, RESERVE_ROOT))
|
||||
|
@ -1589,9 +1589,9 @@ static void default_options(struct f2fs_sb_info *sbi)
|
|||
set_opt(sbi, FLUSH_MERGE);
|
||||
set_opt(sbi, DISCARD);
|
||||
if (f2fs_sb_has_blkzoned(sbi))
|
||||
set_opt_mode(sbi, F2FS_MOUNT_LFS);
|
||||
F2FS_OPTION(sbi).fs_mode = FS_MODE_LFS;
|
||||
else
|
||||
set_opt_mode(sbi, F2FS_MOUNT_ADAPTIVE);
|
||||
F2FS_OPTION(sbi).fs_mode = FS_MODE_ADAPTIVE;
|
||||
|
||||
#ifdef CONFIG_F2FS_FS_XATTR
|
||||
set_opt(sbi, XATTR_USER);
|
||||
|
|
Loading…
Reference in New Issue