[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <64926da7-05ba-c731-b425-1984cd394015@huawei.com>
Date: Tue, 7 Apr 2020 11:21:04 +0800
From: Chao Yu <yuchao0@...wei.com>
To: Jaegeuk Kim <jaegeuk@...nel.org>
CC: Sahitya Tummala <stummala@...eaurora.org>,
<linux-kernel@...r.kernel.org>,
<linux-f2fs-devel@...ts.sourceforge.net>
Subject: Re: [f2fs-dev] [PATCH] f2fs: prevent meta updates while checkpoint is
in progress
On 2020/4/7 10:58, Jaegeuk Kim wrote:
> On 04/07, Chao Yu wrote:
>> On 2020/4/4 1:27, Jaegeuk Kim wrote:
>>> On 04/03, Jaegeuk Kim wrote:
>>>> On 04/01, Sahitya Tummala wrote:
>>>>> Hi Jaegeuk,
>>>>>
>>>>> Got it.
>>>>> The diff below looks good to me.
>>>>> Would you like me to test it and put a patch for this?
>>>>
>>>> Sahitya, Chao,
>>>>
>>>> Could you please take a look at this patch and test intensively?
>>>>
>>>> Thanks,
>>>
>>> v2:
>>>
>>> >From 6bf7d5b227d466b0fe90d4957af29bd184fb646e Mon Sep 17 00:00:00 2001
>>> From: Jaegeuk Kim <jaegeuk@...nel.org>
>>> Date: Tue, 31 Mar 2020 11:43:07 -0700
>>> Subject: [PATCH] f2fs: refactor resize_fs to avoid meta updates in progress
>>>
>>> Sahitya raised an issue:
>>> - prevent meta updates while checkpoint is in progress
>>>
>>> allocate_segment_for_resize() can cause metapage updates if
>>> it requires to change the current node/data segments for resizing.
>>> Stop these meta updates when there is a checkpoint already
>>> in progress to prevent inconsistent CP data.
>>>
>>> Signed-off-by: Sahitya Tummala <stummala@...eaurora.org>
>>> Signed-off-by: Jaegeuk Kim <jaegeuk@...nel.org>
>>> ---
>>> fs/f2fs/checkpoint.c | 6 +-
>>> fs/f2fs/f2fs.h | 2 +-
>>> fs/f2fs/file.c | 5 +-
>>> fs/f2fs/gc.c | 107 +++++++++++++++++++-----------------
>>> fs/f2fs/super.c | 1 -
>>> include/trace/events/f2fs.h | 4 +-
>>> 6 files changed, 67 insertions(+), 58 deletions(-)
>>>
>>> diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
>>> index 852890b72d6ac..531995192b714 100644
>>> --- a/fs/f2fs/checkpoint.c
>>> +++ b/fs/f2fs/checkpoint.c
>>> @@ -1553,7 +1553,8 @@ int f2fs_write_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc)
>>> return 0;
>>> f2fs_warn(sbi, "Start checkpoint disabled!");
>>> }
>>> - mutex_lock(&sbi->cp_mutex);
>>> + if (cpc->reason != CP_RESIZE)
>>> + mutex_lock(&sbi->cp_mutex);
>>>
>>> if (!is_sbi_flag_set(sbi, SBI_IS_DIRTY) &&
>>> ((cpc->reason & CP_FASTBOOT) || (cpc->reason & CP_SYNC) ||
>>> @@ -1622,7 +1623,8 @@ int f2fs_write_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc)
>>> f2fs_update_time(sbi, CP_TIME);
>>> trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, "finish checkpoint");
>>> out:
>>> - mutex_unlock(&sbi->cp_mutex);
>>> + if (cpc->reason != CP_RESIZE)
>>> + mutex_unlock(&sbi->cp_mutex);
>>> return err;
>>> }
>>>
>>> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
>>> index be02a5cadd944..f9b2caa2135bd 100644
>>> --- a/fs/f2fs/f2fs.h
>>> +++ b/fs/f2fs/f2fs.h
>>> @@ -193,6 +193,7 @@ enum {
>>> #define CP_DISCARD 0x00000010
>>> #define CP_TRIMMED 0x00000020
>>> #define CP_PAUSE 0x00000040
>>> +#define CP_RESIZE 0x00000080
>>>
>>> #define MAX_DISCARD_BLOCKS(sbi) BLKS_PER_SEC(sbi)
>>> #define DEF_MAX_DISCARD_REQUEST 8 /* issue 8 discards per round */
>>> @@ -1421,7 +1422,6 @@ struct f2fs_sb_info {
>>> unsigned int segs_per_sec; /* segments per section */
>>> unsigned int secs_per_zone; /* sections per zone */
>>> unsigned int total_sections; /* total section count */
>>> - struct mutex resize_mutex; /* for resize exclusion */
>>> unsigned int total_node_count; /* total node block count */
>>> unsigned int total_valid_node_count; /* valid node block count */
>>> loff_t max_file_blocks; /* max block index of file */
>>> diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c
>>> index 257e61d0afffb..b4c12370bb3d6 100644
>>> --- a/fs/f2fs/file.c
>>> +++ b/fs/f2fs/file.c
>>> @@ -3305,7 +3305,6 @@ static int f2fs_ioc_resize_fs(struct file *filp, unsigned long arg)
>>> {
>>> struct f2fs_sb_info *sbi = F2FS_I_SB(file_inode(filp));
>>> __u64 block_count;
>>> - int ret;
>>>
>>> if (!capable(CAP_SYS_ADMIN))
>>> return -EPERM;
>>> @@ -3317,9 +3316,7 @@ static int f2fs_ioc_resize_fs(struct file *filp, unsigned long arg)
>>> sizeof(block_count)))
>>> return -EFAULT;
>>>
>>> - ret = f2fs_resize_fs(sbi, block_count);
>>> -
>>> - return ret;
>>> + return f2fs_resize_fs(sbi, block_count);
>>> }
>>>
>>> static int f2fs_ioc_enable_verity(struct file *filp, unsigned long arg)
>>> diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c
>>> index 26248c8936db0..46c75ecb64a2e 100644
>>> --- a/fs/f2fs/gc.c
>>> +++ b/fs/f2fs/gc.c
>>> @@ -1399,12 +1399,29 @@ void f2fs_build_gc_manager(struct f2fs_sb_info *sbi)
>>> GET_SEGNO(sbi, FDEV(0).end_blk) + 1;
>>> }
>>>
>>> -static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
>>> - unsigned int end)
>>> +static int free_segment_range(struct f2fs_sb_info *sbi,
>>> + unsigned int secs, bool gc_only)
>>> {
>>> - int type;
>>> - unsigned int segno, next_inuse;
>>> + unsigned int segno, next_inuse, start, end;
>>> + struct cp_control cpc = { CP_RESIZE, 0, 0, 0 };
>>> + int gc_mode, gc_type;
>>> int err = 0;
>>> + int type;
>>> +
>>> + /* Force block allocation for GC */
>>> + MAIN_SECS(sbi) -= secs;
>>> + start = MAIN_SECS(sbi) * sbi->segs_per_sec;
>>> + end = MAIN_SEGS(sbi) - 1;
>>> +
>>> + mutex_lock(&DIRTY_I(sbi)->seglist_lock);
>>> + for (gc_mode = 0; gc_mode < MAX_GC_POLICY; gc_mode++)
>>> + if (SIT_I(sbi)->last_victim[gc_mode] >= start)
>>> + SIT_I(sbi)->last_victim[gc_mode] = 0;
>>> +
>>> + for (gc_type = BG_GC; gc_type <= FG_GC; gc_type++)
>>> + if (sbi->next_victim_seg[gc_type] >= start)
>>> + sbi->next_victim_seg[gc_type] = NULL_SEGNO;
>>> + mutex_unlock(&DIRTY_I(sbi)->seglist_lock);
>>>
>>> /* Move out cursegs from the target range */
>>> for (type = CURSEG_HOT_DATA; type < NR_CURSEG_TYPE; type++)
>>> @@ -1417,18 +1434,20 @@ static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
>>> .iroot = RADIX_TREE_INIT(gc_list.iroot, GFP_NOFS),
>>> };
>>>
>>> - down_write(&sbi->gc_lock);
>>> do_garbage_collect(sbi, segno, &gc_list, FG_GC);
>>> - up_write(&sbi->gc_lock);
>>> put_gc_inode(&gc_list);
>>>
>>> - if (get_valid_blocks(sbi, segno, true))
>>> - return -EAGAIN;
>>> + if (!gc_only && get_valid_blocks(sbi, segno, true)) {
>>> + err = -EAGAIN;
>>> + goto out;
>>> + }
>>> }
>>> + if (gc_only)
>>> + goto out;
>>>
>>> - err = f2fs_sync_fs(sbi->sb, 1);
>>> + err = f2fs_write_checkpoint(sbi, &cpc);
>>> if (err)
>>> - return err;
>>> + goto out;
>>>
>>> next_inuse = find_next_inuse(FREE_I(sbi), end + 1, start);
>>> if (next_inuse <= end) {
>>> @@ -1436,6 +1455,8 @@ static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
>>> next_inuse);
>>> f2fs_bug_on(sbi, 1);
>>> }
>>> +out:
>>> + MAIN_SECS(sbi) -= secs;
>>> return err;
>>> }
>>>
>>> @@ -1481,6 +1502,7 @@ static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs)
>>>
>>> SM_I(sbi)->segment_count = (int)SM_I(sbi)->segment_count + segs;
>>> MAIN_SEGS(sbi) = (int)MAIN_SEGS(sbi) + segs;
>>> + MAIN_SECS(sbi) += secs;
>>> FREE_I(sbi)->free_sections = (int)FREE_I(sbi)->free_sections + secs;
>>> FREE_I(sbi)->free_segments = (int)FREE_I(sbi)->free_segments + segs;
>>> F2FS_CKPT(sbi)->user_block_count = cpu_to_le64(user_block_count + blks);
>>> @@ -1502,8 +1524,8 @@ static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs)
>>> int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
>>> {
>>> __u64 old_block_count, shrunk_blocks;
>>> + struct cp_control cpc = { CP_RESIZE, 0, 0, 0 };
>>> unsigned int secs;
>>> - int gc_mode, gc_type;
>>> int err = 0;
>>> __u32 rem;
>>>
>>> @@ -1538,10 +1560,22 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
>>> return -EINVAL;
>>> }
>>>
>>> - freeze_bdev(sbi->sb->s_bdev);
>>> -
>>> shrunk_blocks = old_block_count - block_count;
>>> secs = div_u64(shrunk_blocks, BLKS_PER_SEC(sbi));
>>> +
>>> + /* protect MAIN_SEC in free_segment_range */
>>> + f2fs_lock_op(sbi);
>>> + err = free_segment_range(sbi, secs, true);
>>
>> For this path, we break the rule that we need hold gc_lock during
>> do_garbage_collect().
>
> I don't get the point.
> In free_segment_range(), gc_lock is held before/after do_garbage_collect().
@@ -1417,18 +1434,20 @@ static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
.iroot = RADIX_TREE_INIT(gc_list.iroot, GFP_NOFS),
};
- down_write(&sbi->gc_lock);
do_garbage_collect(sbi, segno, &gc_list, FG_GC);
- up_write(&sbi->gc_lock);
put_gc_inode(&gc_list);
They were removed, right?
>
>>
>> One other concern is that still the granularity of lock_op is too large,
>> to avoid potential hang if it triggers heavy gc migration, how about using
>> a timeout mechanism in free_segment_range() like we did in
>> f2fs_disable_checkpoint()?
>
> We can do first round GC without f2fs_lock_op().
Yup, that makes sense to me.
>
>>
>>> + f2fs_unlock_op(sbi);
>>> + if (err)
>>> + return err;
>>> +
>>> + set_sbi_flag(sbi, SBI_IS_RESIZEFS);
>>> +
>>> + freeze_super(sbi->sb);
>>> + down_write(&sbi->gc_lock);
>>> + mutex_lock(&sbi->cp_mutex);
>>> +
>>> spin_lock(&sbi->stat_lock);
>>> if (shrunk_blocks + valid_user_blocks(sbi) +
>>> sbi->current_reserved_blocks + sbi->unusable_block_count +
>>> @@ -1550,69 +1584,44 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
>>> else
>>> sbi->user_block_count -= shrunk_blocks;
>>> spin_unlock(&sbi->stat_lock);
>>> - if (err) {
>>> - thaw_bdev(sbi->sb->s_bdev, sbi->sb);
>>> - return err;
>>> - }
>>> -
>>> - mutex_lock(&sbi->resize_mutex);
>>> - set_sbi_flag(sbi, SBI_IS_RESIZEFS);
>>> -
>>> - mutex_lock(&DIRTY_I(sbi)->seglist_lock);
>>> -
>>> - MAIN_SECS(sbi) -= secs;
>>> -
>>> - for (gc_mode = 0; gc_mode < MAX_GC_POLICY; gc_mode++)
>>> - if (SIT_I(sbi)->last_victim[gc_mode] >=
>>> - MAIN_SECS(sbi) * sbi->segs_per_sec)
>>> - SIT_I(sbi)->last_victim[gc_mode] = 0;
>>> -
>>> - for (gc_type = BG_GC; gc_type <= FG_GC; gc_type++)
>>> - if (sbi->next_victim_seg[gc_type] >=
>>> - MAIN_SECS(sbi) * sbi->segs_per_sec)
>>> - sbi->next_victim_seg[gc_type] = NULL_SEGNO;
>>> -
>>> - mutex_unlock(&DIRTY_I(sbi)->seglist_lock);
>>> + if (err)
>>> + goto out_err;
>>>
>>> - err = free_segment_range(sbi, MAIN_SECS(sbi) * sbi->segs_per_sec,
>>> - MAIN_SEGS(sbi) - 1);
>>> + err = free_segment_range(sbi, secs, false);
>>
>> Lock coverage is still large here, what about just checking the resize condition
>> with find_next_inuse(, end + 1, start), if the migration finished, then let's call
>> write_checkpoint(), otherwise, returning -EAGAIN.
>
> We did GC above, so how much time do you expect to complete here? Basically
Just to consider handling corner case here, which is in between round 1 and round 2
GC, there may occur a fair mount of data writes locating in the tail of device.
> I hesitate to return EAGAIN, since there's no context with respect to how
> many times user need to retry to get succeess. Disabling checkpoint has
I think user can accept EAGAIN if we describe clearly about that the interface
may fail due to concurrent IO on device, if the administrator want to resize
successfully, concurrent operation on device should be avoided.
> some ways to get a sense tho. Nevertheless, if we want to return EAGAIN,
> it'd be better to give # of bytes that it requires to migrate further?
Hmm.. I don't think there is such number, as resize should be atomic.
Thanks,
>
>>
>>> if (err)
>>> - goto out;
>>> + goto recover_out;
>>>
>>> update_sb_metadata(sbi, -secs);
>>>
>>> err = f2fs_commit_super(sbi, false);
>>> if (err) {
>>> update_sb_metadata(sbi, secs);
>>> - goto out;
>>> + goto recover_out;
>>> }
>>>
>>> - mutex_lock(&sbi->cp_mutex);
>>> update_fs_metadata(sbi, -secs);
>>> clear_sbi_flag(sbi, SBI_IS_RESIZEFS);
>>> set_sbi_flag(sbi, SBI_IS_DIRTY);
>>> - mutex_unlock(&sbi->cp_mutex);
>>>
>>> - err = f2fs_sync_fs(sbi->sb, 1);
>>> + err = f2fs_write_checkpoint(sbi, &cpc);
>>> if (err) {
>>> - mutex_lock(&sbi->cp_mutex);
>>> update_fs_metadata(sbi, secs);
>>> - mutex_unlock(&sbi->cp_mutex);
>>> update_sb_metadata(sbi, secs);
>>> f2fs_commit_super(sbi, false);
>>> }
>>> -out:
>>> +recover_out:
>>> if (err) {
>>> set_sbi_flag(sbi, SBI_NEED_FSCK);
>>> f2fs_err(sbi, "resize_fs failed, should run fsck to repair!");
>>>
>>> - MAIN_SECS(sbi) += secs;
>>> spin_lock(&sbi->stat_lock);
>>> sbi->user_block_count += shrunk_blocks;
>>> spin_unlock(&sbi->stat_lock);
>>> }
>>> +out_err:
>>> + mutex_unlock(&sbi->cp_mutex);
>>> + up_write(&sbi->gc_lock);
>>> + thaw_super(sbi->sb);
>>> clear_sbi_flag(sbi, SBI_IS_RESIZEFS);
>>> - mutex_unlock(&sbi->resize_mutex);
>>> - thaw_bdev(sbi->sb->s_bdev, sbi->sb);
>>> return err;
>>> }
>>> diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c
>>> index b83b17b54a0a6..1e7b1d21d0177 100644
>>> --- a/fs/f2fs/super.c
>>> +++ b/fs/f2fs/super.c
>>> @@ -3412,7 +3412,6 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
>>> init_rwsem(&sbi->gc_lock);
>>> mutex_init(&sbi->writepages);
>>> mutex_init(&sbi->cp_mutex);
>>> - mutex_init(&sbi->resize_mutex);
>>> init_rwsem(&sbi->node_write);
>>> init_rwsem(&sbi->node_change);
>>>
>>> diff --git a/include/trace/events/f2fs.h b/include/trace/events/f2fs.h
>>> index 4d7d4c391879d..5d1a72001fdb4 100644
>>> --- a/include/trace/events/f2fs.h
>>> +++ b/include/trace/events/f2fs.h
>>> @@ -50,6 +50,7 @@ TRACE_DEFINE_ENUM(CP_RECOVERY);
>>> TRACE_DEFINE_ENUM(CP_DISCARD);
>>> TRACE_DEFINE_ENUM(CP_TRIMMED);
>>> TRACE_DEFINE_ENUM(CP_PAUSE);
>>> +TRACE_DEFINE_ENUM(CP_RESIZE);
>>>
>>> #define show_block_type(type) \
>>> __print_symbolic(type, \
>>> @@ -126,7 +127,8 @@ TRACE_DEFINE_ENUM(CP_PAUSE);
>>> { CP_RECOVERY, "Recovery" }, \
>>> { CP_DISCARD, "Discard" }, \
>>> { CP_PAUSE, "Pause" }, \
>>> - { CP_TRIMMED, "Trimmed" })
>>> + { CP_TRIMMED, "Trimmed" }, \
>>> + { CP_RESIZE, "Resize" })
>>>
>>> #define show_fsync_cpreason(type) \
>>> __print_symbolic(type, \
>>>
> .
>
Powered by blists - more mailing lists