[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <YTvWd1/EzwFu6I4T@google.com>
Date: Fri, 10 Sep 2021 15:04:39 -0700
From: Jaegeuk Kim <jaegeuk@...nel.org>
To: Chao Yu <chao@...nel.org>
Cc: linux-f2fs-devel@...ts.sourceforge.net,
linux-kernel@...r.kernel.org, Chao Yu <chao.yu@...ux.dev>
Subject: Re: [PATCH] f2fs: introduce excess_dirty_threshold()
On 09/01, Chao Yu wrote:
> This patch enables f2fs_balance_fs_bg() to check all metadatas' dirty
> threshold rather than just checking node block's, so that checkpoint()
> from background can be triggered more frequently to avoid heaping up
> too much dirty metadatas.
>
> Threshold value by default:
> race with foreground ops single type global
> No 16MB 24MB
> Yes 24MB 36MB
>
> In addtion, let f2fs_balance_fs_bg() be aware of roll-forward sapce
> as well as fsync().
>
> Signed-off-by: Chao Yu <chao@...nel.org>
> ---
> fs/f2fs/f2fs.h | 3 +++
> fs/f2fs/node.h | 5 -----
> fs/f2fs/segment.c | 23 +++++++++++++++++++++--
> 3 files changed, 24 insertions(+), 7 deletions(-)
>
> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> index 6c5e75f86da4..5ae2ca6dba96 100644
> --- a/fs/f2fs/f2fs.h
> +++ b/fs/f2fs/f2fs.h
> @@ -562,6 +562,9 @@ enum {
>
> #define MAX_DIR_RA_PAGES 4 /* maximum ra pages of dir */
>
> +/* dirty segments threshold for triggering CP */
> +#define DEFAULT_DIRTY_THRESHOLD 4
> +
> /* for in-memory extent cache entry */
> #define F2FS_MIN_EXTENT_LEN 64 /* minimum extent length */
>
> diff --git a/fs/f2fs/node.h b/fs/f2fs/node.h
> index ff14a6e5ac1c..18b98cf0465b 100644
> --- a/fs/f2fs/node.h
> +++ b/fs/f2fs/node.h
> @@ -138,11 +138,6 @@ static inline bool excess_cached_nats(struct f2fs_sb_info *sbi)
> return NM_I(sbi)->nat_cnt[TOTAL_NAT] >= DEF_NAT_CACHE_THRESHOLD;
> }
>
> -static inline bool excess_dirty_nodes(struct f2fs_sb_info *sbi)
> -{
> - return get_pages(sbi, F2FS_DIRTY_NODES) >= sbi->blocks_per_seg * 8;
> -}
> -
> enum mem_type {
> FREE_NIDS, /* indicates the free nid list */
> NAT_ENTRIES, /* indicates the cached nat entry */
> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
> index 7358342652ec..ffd148429a9d 100644
> --- a/fs/f2fs/segment.c
> +++ b/fs/f2fs/segment.c
> @@ -535,6 +535,25 @@ void f2fs_balance_fs(struct f2fs_sb_info *sbi, bool need)
> }
> }
>
> +static inline bool excess_dirty_threshold(struct f2fs_sb_info *sbi)
> +{
> + int factor = rwsem_is_locked(&sbi->cp_rwsem) ? 3 : 2;
> + unsigned int dents = get_pages(sbi, F2FS_DIRTY_DENTS);
> + unsigned int qdata = get_pages(sbi, F2FS_DIRTY_QDATA);
> + unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES);
> + unsigned int meta = get_pages(sbi, F2FS_DIRTY_META);
> + unsigned int imeta = get_pages(sbi, F2FS_DIRTY_IMETA);
> + unsigned int threshold = sbi->blocks_per_seg * factor *
> + DEFAULT_DIRTY_THRESHOLD;
> + unsigned int global_threshold = threshold * 3 / 2;
> +
> + if (dents >= threshold || qdata >= threshold ||
> + nodes >= threshold || meta >= threshold ||
> + imeta >= threshold)
> + return true;
> + return dents + qdata + nodes + meta + imeta > global_threshold;
> +}
> +
> void f2fs_balance_fs_bg(struct f2fs_sb_info *sbi, bool from_bg)
> {
> if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
> @@ -553,8 +572,8 @@ void f2fs_balance_fs_bg(struct f2fs_sb_info *sbi, bool from_bg)
> else
> f2fs_build_free_nids(sbi, false, false);
>
> - if (excess_dirty_nats(sbi) || excess_dirty_nodes(sbi) ||
> - excess_prefree_segs(sbi))
> + if (excess_dirty_nats(sbi) || excess_dirty_threshold(sbi) ||
> + excess_prefree_segs(sbi) || f2fs_space_for_roll_forward(sbi))
f2fs_space_for_roll_forward() == 0?
> goto do_sync;
>
> /* there is background inflight IO or foreground operation recently */
> --
> 2.32.0
Powered by blists - more mailing lists