lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20140623214053.GA3088@jmac.local>
Date:	Tue, 24 Jun 2014 06:40:53 +0900
From:	Jaegeuk Kim <jaegeuk@...nel.org>
To:	Chao Yu <chao2.yu@...sung.com>
Cc:	Changman Lee <cm224.lee@...sung.com>,
	linux-f2fs-devel@...ts.sourceforge.net,
	linux-fsdevel@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [f2fs-dev] [PATCH v2 RESEND] f2fs: refactor flush_nat_entries
 codes for reducing NAT writes

Hi Chao,

Thank you for the patch. :)
Just one minor suggestion.

[snip]

>  /*
> @@ -1792,80 +1864,87 @@ void flush_nat_entries(struct f2fs_sb_info *sbi)
>  	struct f2fs_nm_info *nm_i = NM_I(sbi);
>  	struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
>  	struct f2fs_summary_block *sum = curseg->sum_blk;
> -	struct nat_entry *ne, *cur;
> -	struct page *page = NULL;
> -	struct f2fs_nat_block *nat_blk = NULL;
> -	nid_t start_nid = 0, end_nid = 0;
> -	bool flushed;
> -
> -	flushed = flush_nats_in_journal(sbi);
> -
> -	if (!flushed)
> -		mutex_lock(&curseg->curseg_mutex);
> -
> -	/* 1) flush dirty nat caches */
> -	list_for_each_entry_safe(ne, cur, &nm_i->dirty_nat_entries, list) {
> -		nid_t nid;
> -		struct f2fs_nat_entry raw_ne;
> -		int offset = -1;
> -
> -		if (nat_get_blkaddr(ne) == NEW_ADDR)
> -			continue;
> +	struct nat_entry_set *nes, *tmp;
> +	struct list_head *head = &nm_i->nat_entry_set;
> +	bool to_journal = true;
>  
> -		nid = nat_get_nid(ne);
> +	/* merge nat entries of dirty list to nat entry set temporarily */
> +	merge_nats_in_set(sbi);
>  
> -		if (flushed)
> -			goto to_nat_page;
> +	if (!__has_cursum_space(sum, nm_i->dirty_nat_cnt)) {
> +		/*
> +		 * if there are no enough space in journal to store dirty nat
> +		 * entries, remove all entries from journal and merge them
> +		 * into nat entry set.
> +		 */
> +		remove_nats_in_journal(sbi);
> +		merge_nats_in_set(sbi);
> +	}

How about this?

	/*
	 * if there are no enough space in journal to store dirty nat
	 * entries, remove all entries from journal and merge them
	 * into nat entry set.
	 */
	if (!__has_cursum_space(sum, nm_i->dirty_nat_cnt))
		remove_nats_in_journal(sbi);
		
	/* merge nat entries of dirty list to nat entry set temporarily */
	merge_nats_in_set(sbi);

Thanks,

>  
> -		/* if there is room for nat enries in curseg->sumpage */
> -		offset = lookup_journal_in_cursum(sum, NAT_JOURNAL, nid, 1);
> -		if (offset >= 0) {
> -			raw_ne = nat_in_journal(sum, offset);
> -			goto flush_now;
> -		}
> -to_nat_page:
> -		if (!page || (start_nid > nid || nid > end_nid)) {
> -			if (page) {
> -				f2fs_put_page(page, 1);
> -				page = NULL;
> -			}
> -			start_nid = START_NID(nid);
> -			end_nid = start_nid + NAT_ENTRY_PER_BLOCK - 1;
> +	if (!nm_i->dirty_nat_cnt)
> +		return;
>  
> -			/*
> -			 * get nat block with dirty flag, increased reference
> -			 * count, mapped and lock
> -			 */
> +	/*
> +	 * there are two steps to flush nat entries:
> +	 * #1, flush nat entries to journal in current hot data summary block.
> +	 * #2, flush nat entries to nat page.
> +	 */
> +	list_for_each_entry_safe(nes, tmp, head, set_list) {
> +		struct f2fs_nat_block *nat_blk;
> +		struct nat_entry *ne, *cur;
> +		struct page *page;
> +		nid_t start_nid = nes->start_nid;
> +
> +		if (to_journal && !__has_cursum_space(sum, nes->entry_cnt))
> +			to_journal = false;
> +
> +		if (to_journal) {
> +			mutex_lock(&curseg->curseg_mutex);
> +		} else {
>  			page = get_next_nat_page(sbi, start_nid);
>  			nat_blk = page_address(page);
> +			f2fs_bug_on(!nat_blk);
>  		}
>  
> -		f2fs_bug_on(!nat_blk);
> -		raw_ne = nat_blk->entries[nid - start_nid];
> -flush_now:
> -		raw_nat_from_node_info(&raw_ne, &ne->ni);
> -
> -		if (offset < 0) {
> -			nat_blk->entries[nid - start_nid] = raw_ne;
> -		} else {
> -			nat_in_journal(sum, offset) = raw_ne;
> -			nid_in_journal(sum, offset) = cpu_to_le32(nid);
> -		}
> +		/* flush dirty nats in nat entry set */
> +		list_for_each_entry_safe(ne, cur, &nes->entry_list, list) {
> +			struct f2fs_nat_entry *raw_ne;
> +			nid_t nid = nat_get_nid(ne);
> +			int offset;
> +
> +			if (to_journal) {
> +				offset = lookup_journal_in_cursum(sum,
> +							NAT_JOURNAL, nid, 1);
> +				f2fs_bug_on(offset < 0);
> +				raw_ne = &nat_in_journal(sum, offset);
> +				nid_in_journal(sum, offset) = cpu_to_le32(nid);
> +			} else {
> +				raw_ne = &nat_blk->entries[nid - start_nid];
> +			}
> +			raw_nat_from_node_info(raw_ne, &ne->ni);
>  
> -		if (nat_get_blkaddr(ne) == NULL_ADDR &&
> +			if (nat_get_blkaddr(ne) == NULL_ADDR &&
>  				add_free_nid(sbi, nid, false) <= 0) {
> -			write_lock(&nm_i->nat_tree_lock);
> -			__del_from_nat_cache(nm_i, ne);
> -			write_unlock(&nm_i->nat_tree_lock);
> -		} else {
> -			write_lock(&nm_i->nat_tree_lock);
> -			__clear_nat_cache_dirty(nm_i, ne);
> -			write_unlock(&nm_i->nat_tree_lock);
> +				write_lock(&nm_i->nat_tree_lock);
> +				__del_from_nat_cache(nm_i, ne);
> +				write_unlock(&nm_i->nat_tree_lock);
> +			} else {
> +				write_lock(&nm_i->nat_tree_lock);
> +				__clear_nat_cache_dirty(nm_i, ne);
> +				write_unlock(&nm_i->nat_tree_lock);
> +			}
>  		}
> +
> +		if (to_journal)
> +			mutex_unlock(&curseg->curseg_mutex);
> +		else
> +			f2fs_put_page(page, 1);
> +
> +		release_nat_entry_set(nes, nm_i);
>  	}
> -	if (!flushed)
> -		mutex_unlock(&curseg->curseg_mutex);
> -	f2fs_put_page(page, 1);
> +
> +	f2fs_bug_on(!list_empty(head));
> +	f2fs_bug_on(nm_i->dirty_nat_cnt);
>  }
>  
>  static int init_node_manager(struct f2fs_sb_info *sbi)
> @@ -1894,6 +1973,7 @@ static int init_node_manager(struct f2fs_sb_info *sbi)
>  	INIT_RADIX_TREE(&nm_i->nat_root, GFP_ATOMIC);
>  	INIT_LIST_HEAD(&nm_i->nat_entries);
>  	INIT_LIST_HEAD(&nm_i->dirty_nat_entries);
> +	INIT_LIST_HEAD(&nm_i->nat_entry_set);
>  
>  	mutex_init(&nm_i->build_lock);
>  	spin_lock_init(&nm_i->free_nid_list_lock);
> @@ -1974,19 +2054,30 @@ int __init create_node_manager_caches(void)
>  	nat_entry_slab = f2fs_kmem_cache_create("nat_entry",
>  			sizeof(struct nat_entry));
>  	if (!nat_entry_slab)
> -		return -ENOMEM;
> +		goto fail;
>  
>  	free_nid_slab = f2fs_kmem_cache_create("free_nid",
>  			sizeof(struct free_nid));
> -	if (!free_nid_slab) {
> -		kmem_cache_destroy(nat_entry_slab);
> -		return -ENOMEM;
> -	}
> +	if (!free_nid_slab)
> +		goto destory_nat_entry;
> +
> +	nat_entry_set_slab = f2fs_kmem_cache_create("nat_entry_set",
> +			sizeof(struct nat_entry_set));
> +	if (!nat_entry_set_slab)
> +		goto destory_free_nid;
>  	return 0;
> +
> +destory_free_nid:
> +	kmem_cache_destroy(free_nid_slab);
> +destory_nat_entry:
> +	kmem_cache_destroy(nat_entry_slab);
> +fail:
> +	return -ENOMEM;
>  }
>  
>  void destroy_node_manager_caches(void)
>  {
> +	kmem_cache_destroy(nat_entry_set_slab);
>  	kmem_cache_destroy(free_nid_slab);
>  	kmem_cache_destroy(nat_entry_slab);
>  }
> diff --git a/fs/f2fs/node.h b/fs/f2fs/node.h
> index 7281112..8a116a4 100644
> --- a/fs/f2fs/node.h
> +++ b/fs/f2fs/node.h
> @@ -89,6 +89,13 @@ enum mem_type {
>  	DIRTY_DENTS	/* indicates dirty dentry pages */
>  };
>  
> +struct nat_entry_set {
> +	struct list_head set_list;	/* link with all nat sets */
> +	struct list_head entry_list;	/* link with dirty nat entries */
> +	nid_t start_nid;		/* start nid of nats in set */
> +	unsigned int entry_cnt;		/* the # of nat entries in set */
> +};
> +
>  /*
>   * For free nid mangement
>   */
> -- 
> 1.7.9.5

-- 
Jaegeuk Kim
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ