lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200407025818.GA137081@google.com>
Date:   Mon, 6 Apr 2020 19:58:18 -0700
From:   Jaegeuk Kim <jaegeuk@...nel.org>
To:     Chao Yu <yuchao0@...wei.com>
Cc:     Sahitya Tummala <stummala@...eaurora.org>,
        linux-kernel@...r.kernel.org,
        linux-f2fs-devel@...ts.sourceforge.net
Subject: Re: [f2fs-dev] [PATCH] f2fs: prevent meta updates while checkpoint
 is in progress

On 04/07, Chao Yu wrote:
> On 2020/4/4 1:27, Jaegeuk Kim wrote:
> > On 04/03, Jaegeuk Kim wrote:
> >> On 04/01, Sahitya Tummala wrote:
> >>> Hi Jaegeuk,
> >>>
> >>> Got it.
> >>> The diff below looks good to me.
> >>> Would you like me to test it and put a patch for this?
> >>
> >> Sahitya, Chao,
> >>
> >> Could you please take a look at this patch and test intensively?
> >>
> >> Thanks,
> > 
> > v2:
> > 
> >>From 6bf7d5b227d466b0fe90d4957af29bd184fb646e Mon Sep 17 00:00:00 2001
> > From: Jaegeuk Kim <jaegeuk@...nel.org>
> > Date: Tue, 31 Mar 2020 11:43:07 -0700
> > Subject: [PATCH] f2fs: refactor resize_fs to avoid meta updates in progress
> > 
> > Sahitya raised an issue:
> > - prevent meta updates while checkpoint is in progress
> > 
> > allocate_segment_for_resize() can cause metapage updates if
> > it requires to change the current node/data segments for resizing.
> > Stop these meta updates when there is a checkpoint already
> > in progress to prevent inconsistent CP data.
> > 
> > Signed-off-by: Sahitya Tummala <stummala@...eaurora.org>
> > Signed-off-by: Jaegeuk Kim <jaegeuk@...nel.org>
> > ---
> >  fs/f2fs/checkpoint.c        |   6 +-
> >  fs/f2fs/f2fs.h              |   2 +-
> >  fs/f2fs/file.c              |   5 +-
> >  fs/f2fs/gc.c                | 107 +++++++++++++++++++-----------------
> >  fs/f2fs/super.c             |   1 -
> >  include/trace/events/f2fs.h |   4 +-
> >  6 files changed, 67 insertions(+), 58 deletions(-)
> > 
> > diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c
> > index 852890b72d6ac..531995192b714 100644
> > --- a/fs/f2fs/checkpoint.c
> > +++ b/fs/f2fs/checkpoint.c
> > @@ -1553,7 +1553,8 @@ int f2fs_write_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc)
> >  			return 0;
> >  		f2fs_warn(sbi, "Start checkpoint disabled!");
> >  	}
> > -	mutex_lock(&sbi->cp_mutex);
> > +	if (cpc->reason != CP_RESIZE)
> > +		mutex_lock(&sbi->cp_mutex);
> >  
> >  	if (!is_sbi_flag_set(sbi, SBI_IS_DIRTY) &&
> >  		((cpc->reason & CP_FASTBOOT) || (cpc->reason & CP_SYNC) ||
> > @@ -1622,7 +1623,8 @@ int f2fs_write_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc)
> >  	f2fs_update_time(sbi, CP_TIME);
> >  	trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, "finish checkpoint");
> >  out:
> > -	mutex_unlock(&sbi->cp_mutex);
> > +	if (cpc->reason != CP_RESIZE)
> > +		mutex_unlock(&sbi->cp_mutex);
> >  	return err;
> >  }
> >  
> > diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> > index be02a5cadd944..f9b2caa2135bd 100644
> > --- a/fs/f2fs/f2fs.h
> > +++ b/fs/f2fs/f2fs.h
> > @@ -193,6 +193,7 @@ enum {
> >  #define	CP_DISCARD	0x00000010
> >  #define CP_TRIMMED	0x00000020
> >  #define CP_PAUSE	0x00000040
> > +#define CP_RESIZE 	0x00000080
> >  
> >  #define MAX_DISCARD_BLOCKS(sbi)		BLKS_PER_SEC(sbi)
> >  #define DEF_MAX_DISCARD_REQUEST		8	/* issue 8 discards per round */
> > @@ -1421,7 +1422,6 @@ struct f2fs_sb_info {
> >  	unsigned int segs_per_sec;		/* segments per section */
> >  	unsigned int secs_per_zone;		/* sections per zone */
> >  	unsigned int total_sections;		/* total section count */
> > -	struct mutex resize_mutex;		/* for resize exclusion */
> >  	unsigned int total_node_count;		/* total node block count */
> >  	unsigned int total_valid_node_count;	/* valid node block count */
> >  	loff_t max_file_blocks;			/* max block index of file */
> > diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c
> > index 257e61d0afffb..b4c12370bb3d6 100644
> > --- a/fs/f2fs/file.c
> > +++ b/fs/f2fs/file.c
> > @@ -3305,7 +3305,6 @@ static int f2fs_ioc_resize_fs(struct file *filp, unsigned long arg)
> >  {
> >  	struct f2fs_sb_info *sbi = F2FS_I_SB(file_inode(filp));
> >  	__u64 block_count;
> > -	int ret;
> >  
> >  	if (!capable(CAP_SYS_ADMIN))
> >  		return -EPERM;
> > @@ -3317,9 +3316,7 @@ static int f2fs_ioc_resize_fs(struct file *filp, unsigned long arg)
> >  			   sizeof(block_count)))
> >  		return -EFAULT;
> >  
> > -	ret = f2fs_resize_fs(sbi, block_count);
> > -
> > -	return ret;
> > +	return f2fs_resize_fs(sbi, block_count);
> >  }
> >  
> >  static int f2fs_ioc_enable_verity(struct file *filp, unsigned long arg)
> > diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c
> > index 26248c8936db0..46c75ecb64a2e 100644
> > --- a/fs/f2fs/gc.c
> > +++ b/fs/f2fs/gc.c
> > @@ -1399,12 +1399,29 @@ void f2fs_build_gc_manager(struct f2fs_sb_info *sbi)
> >  				GET_SEGNO(sbi, FDEV(0).end_blk) + 1;
> >  }
> >  
> > -static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
> > -							unsigned int end)
> > +static int free_segment_range(struct f2fs_sb_info *sbi,
> > +				unsigned int secs, bool gc_only)
> >  {
> > -	int type;
> > -	unsigned int segno, next_inuse;
> > +	unsigned int segno, next_inuse, start, end;
> > +	struct cp_control cpc = { CP_RESIZE, 0, 0, 0 };
> > +	int gc_mode, gc_type;
> >  	int err = 0;
> > +	int type;
> > +
> > +	/* Force block allocation for GC */
> > +	MAIN_SECS(sbi) -= secs;
> > +	start = MAIN_SECS(sbi) * sbi->segs_per_sec;
> > +	end = MAIN_SEGS(sbi) - 1;
> > +
> > +	mutex_lock(&DIRTY_I(sbi)->seglist_lock);
> > +	for (gc_mode = 0; gc_mode < MAX_GC_POLICY; gc_mode++)
> > +		if (SIT_I(sbi)->last_victim[gc_mode] >= start)
> > +			SIT_I(sbi)->last_victim[gc_mode] = 0;
> > +
> > +	for (gc_type = BG_GC; gc_type <= FG_GC; gc_type++)
> > +		if (sbi->next_victim_seg[gc_type] >= start)
> > +			sbi->next_victim_seg[gc_type] = NULL_SEGNO;
> > +	mutex_unlock(&DIRTY_I(sbi)->seglist_lock);
> >  
> >  	/* Move out cursegs from the target range */
> >  	for (type = CURSEG_HOT_DATA; type < NR_CURSEG_TYPE; type++)
> > @@ -1417,18 +1434,20 @@ static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
> >  			.iroot = RADIX_TREE_INIT(gc_list.iroot, GFP_NOFS),
> >  		};
> >  
> > -		down_write(&sbi->gc_lock);
> >  		do_garbage_collect(sbi, segno, &gc_list, FG_GC);
> > -		up_write(&sbi->gc_lock);
> >  		put_gc_inode(&gc_list);
> >  
> > -		if (get_valid_blocks(sbi, segno, true))
> > -			return -EAGAIN;
> > +		if (!gc_only && get_valid_blocks(sbi, segno, true)) {
> > +			err = -EAGAIN;
> > +			goto out;
> > +		}
> >  	}
> > +	if (gc_only)
> > +		goto out;
> >  
> > -	err = f2fs_sync_fs(sbi->sb, 1);
> > +	err = f2fs_write_checkpoint(sbi, &cpc);
> >  	if (err)
> > -		return err;
> > +		goto out;
> >  
> >  	next_inuse = find_next_inuse(FREE_I(sbi), end + 1, start);
> >  	if (next_inuse <= end) {
> > @@ -1436,6 +1455,8 @@ static int free_segment_range(struct f2fs_sb_info *sbi, unsigned int start,
> >  			 next_inuse);
> >  		f2fs_bug_on(sbi, 1);
> >  	}
> > +out:
> > +	MAIN_SECS(sbi) -= secs;
> >  	return err;
> >  }
> >  
> > @@ -1481,6 +1502,7 @@ static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs)
> >  
> >  	SM_I(sbi)->segment_count = (int)SM_I(sbi)->segment_count + segs;
> >  	MAIN_SEGS(sbi) = (int)MAIN_SEGS(sbi) + segs;
> > +	MAIN_SECS(sbi) += secs;
> >  	FREE_I(sbi)->free_sections = (int)FREE_I(sbi)->free_sections + secs;
> >  	FREE_I(sbi)->free_segments = (int)FREE_I(sbi)->free_segments + segs;
> >  	F2FS_CKPT(sbi)->user_block_count = cpu_to_le64(user_block_count + blks);
> > @@ -1502,8 +1524,8 @@ static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs)
> >  int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
> >  {
> >  	__u64 old_block_count, shrunk_blocks;
> > +	struct cp_control cpc = { CP_RESIZE, 0, 0, 0 };
> >  	unsigned int secs;
> > -	int gc_mode, gc_type;
> >  	int err = 0;
> >  	__u32 rem;
> >  
> > @@ -1538,10 +1560,22 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
> >  		return -EINVAL;
> >  	}
> >  
> > -	freeze_bdev(sbi->sb->s_bdev);
> > -
> >  	shrunk_blocks = old_block_count - block_count;
> >  	secs = div_u64(shrunk_blocks, BLKS_PER_SEC(sbi));
> > +
> > +	/* protect MAIN_SEC in free_segment_range */
> > +	f2fs_lock_op(sbi);
> > +	err = free_segment_range(sbi, secs, true);
> 
> For this path, we break the rule that we need hold gc_lock during
> do_garbage_collect().

I don't get the point.
In free_segment_range(), gc_lock is held before/after do_garbage_collect().

> 
> One other concern is that still the granularity of lock_op is too large,
> to avoid potential hang if it triggers heavy gc migration, how about using
> a timeout mechanism in free_segment_range() like we did in
> f2fs_disable_checkpoint()?

We can do first round GC without f2fs_lock_op().

> 
> > +	f2fs_unlock_op(sbi);
> > +	if (err)
> > +		return err;
> > +
> > +	set_sbi_flag(sbi, SBI_IS_RESIZEFS);
> > +
> > +	freeze_super(sbi->sb);
> > +	down_write(&sbi->gc_lock);
> > +	mutex_lock(&sbi->cp_mutex);
> > +
> >  	spin_lock(&sbi->stat_lock);
> >  	if (shrunk_blocks + valid_user_blocks(sbi) +
> >  		sbi->current_reserved_blocks + sbi->unusable_block_count +
> > @@ -1550,69 +1584,44 @@ int f2fs_resize_fs(struct f2fs_sb_info *sbi, __u64 block_count)
> >  	else
> >  		sbi->user_block_count -= shrunk_blocks;
> >  	spin_unlock(&sbi->stat_lock);
> > -	if (err) {
> > -		thaw_bdev(sbi->sb->s_bdev, sbi->sb);
> > -		return err;
> > -	}
> > -
> > -	mutex_lock(&sbi->resize_mutex);
> > -	set_sbi_flag(sbi, SBI_IS_RESIZEFS);
> > -
> > -	mutex_lock(&DIRTY_I(sbi)->seglist_lock);
> > -
> > -	MAIN_SECS(sbi) -= secs;
> > -
> > -	for (gc_mode = 0; gc_mode < MAX_GC_POLICY; gc_mode++)
> > -		if (SIT_I(sbi)->last_victim[gc_mode] >=
> > -					MAIN_SECS(sbi) * sbi->segs_per_sec)
> > -			SIT_I(sbi)->last_victim[gc_mode] = 0;
> > -
> > -	for (gc_type = BG_GC; gc_type <= FG_GC; gc_type++)
> > -		if (sbi->next_victim_seg[gc_type] >=
> > -					MAIN_SECS(sbi) * sbi->segs_per_sec)
> > -			sbi->next_victim_seg[gc_type] = NULL_SEGNO;
> > -
> > -	mutex_unlock(&DIRTY_I(sbi)->seglist_lock);
> > +	if (err)
> > +		goto out_err;
> >  
> > -	err = free_segment_range(sbi, MAIN_SECS(sbi) * sbi->segs_per_sec,
> > -			MAIN_SEGS(sbi) - 1);
> > +	err = free_segment_range(sbi, secs, false);
> 
> Lock coverage is still large here, what about just checking the resize condition
> with find_next_inuse(, end + 1, start), if the migration finished, then let's call
> write_checkpoint(), otherwise, returning -EAGAIN.

We did GC above, so how much time do you expect to complete here? Basically
I hesitate to return EAGAIN, since there's no context with respect to how
many times user need to retry to get succeess. Disabling checkpoint has
some ways to get a sense tho. Nevertheless, if we want to return EAGAIN,
it'd be better to give # of bytes that it requires to migrate further?

> 
> >  	if (err)
> > -		goto out;
> > +		goto recover_out;
> >  
> >  	update_sb_metadata(sbi, -secs);
> >  
> >  	err = f2fs_commit_super(sbi, false);
> >  	if (err) {
> >  		update_sb_metadata(sbi, secs);
> > -		goto out;
> > +		goto recover_out;
> >  	}
> >  
> > -	mutex_lock(&sbi->cp_mutex);
> >  	update_fs_metadata(sbi, -secs);
> >  	clear_sbi_flag(sbi, SBI_IS_RESIZEFS);
> >  	set_sbi_flag(sbi, SBI_IS_DIRTY);
> > -	mutex_unlock(&sbi->cp_mutex);
> >  
> > -	err = f2fs_sync_fs(sbi->sb, 1);
> > +	err = f2fs_write_checkpoint(sbi, &cpc);
> >  	if (err) {
> > -		mutex_lock(&sbi->cp_mutex);
> >  		update_fs_metadata(sbi, secs);
> > -		mutex_unlock(&sbi->cp_mutex);
> >  		update_sb_metadata(sbi, secs);
> >  		f2fs_commit_super(sbi, false);
> >  	}
> > -out:
> > +recover_out:
> >  	if (err) {
> >  		set_sbi_flag(sbi, SBI_NEED_FSCK);
> >  		f2fs_err(sbi, "resize_fs failed, should run fsck to repair!");
> >  
> > -		MAIN_SECS(sbi) += secs;
> >  		spin_lock(&sbi->stat_lock);
> >  		sbi->user_block_count += shrunk_blocks;
> >  		spin_unlock(&sbi->stat_lock);
> >  	}
> > +out_err:
> > +	mutex_unlock(&sbi->cp_mutex);
> > +	up_write(&sbi->gc_lock);
> > +	thaw_super(sbi->sb);
> >  	clear_sbi_flag(sbi, SBI_IS_RESIZEFS);
> > -	mutex_unlock(&sbi->resize_mutex);
> > -	thaw_bdev(sbi->sb->s_bdev, sbi->sb);
> >  	return err;
> >  }
> > diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c
> > index b83b17b54a0a6..1e7b1d21d0177 100644
> > --- a/fs/f2fs/super.c
> > +++ b/fs/f2fs/super.c
> > @@ -3412,7 +3412,6 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> >  	init_rwsem(&sbi->gc_lock);
> >  	mutex_init(&sbi->writepages);
> >  	mutex_init(&sbi->cp_mutex);
> > -	mutex_init(&sbi->resize_mutex);
> >  	init_rwsem(&sbi->node_write);
> >  	init_rwsem(&sbi->node_change);
> >  
> > diff --git a/include/trace/events/f2fs.h b/include/trace/events/f2fs.h
> > index 4d7d4c391879d..5d1a72001fdb4 100644
> > --- a/include/trace/events/f2fs.h
> > +++ b/include/trace/events/f2fs.h
> > @@ -50,6 +50,7 @@ TRACE_DEFINE_ENUM(CP_RECOVERY);
> >  TRACE_DEFINE_ENUM(CP_DISCARD);
> >  TRACE_DEFINE_ENUM(CP_TRIMMED);
> >  TRACE_DEFINE_ENUM(CP_PAUSE);
> > +TRACE_DEFINE_ENUM(CP_RESIZE);
> >  
> >  #define show_block_type(type)						\
> >  	__print_symbolic(type,						\
> > @@ -126,7 +127,8 @@ TRACE_DEFINE_ENUM(CP_PAUSE);
> >  		{ CP_RECOVERY,	"Recovery" },				\
> >  		{ CP_DISCARD,	"Discard" },				\
> >  		{ CP_PAUSE,	"Pause" },				\
> > -		{ CP_TRIMMED,	"Trimmed" })
> > +		{ CP_TRIMMED,	"Trimmed" },				\
> > +		{ CP_RESIZE,	"Resize" })
> >  
> >  #define show_fsync_cpreason(type)					\
> >  	__print_symbolic(type,						\
> > 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ