lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <0D1876A8-BB77-4C1A-BE4F-B4A0E81DD4EA@dilger.ca>
Date:   Wed, 9 Nov 2016 15:57:53 -0700
From:   Andreas Dilger <adilger@...ger.ca>
To:     Jaegeuk Kim <jaegeuk@...nel.org>
Cc:     LKML <linux-kernel@...r.kernel.org>,
        Lustre Development <lustre-devel@...ts.lustre.org>,
        linux-fsdevel <linux-fsdevel@...r.kernel.org>,
        linux-f2fs-devel@...ts.sourceforge.net,
        linux-btrfs <linux-btrfs@...r.kernel.org>
Subject: Re: [PATCH] f2fs: support multiple devices

On Nov 9, 2016, at 1:56 PM, Jaegeuk Kim <jaegeuk@...nel.org> wrote:
> 
> This patch implements multiple devices support for f2fs.
> Given multiple devices by mkfs.f2fs, f2fs shows them entirely as one big
> volume under one f2fs instance.
> 
> Internal block management is very simple, but we will modify block
> allocation and background GC policy to boost IO speed by exploiting them
> accoording to each device speed.

How will you integrate this into FIEMAP, since it is now possible if a
file is split across multiple devices then it will return ambiguous block
numbers for a file.  I've been meaning to merge the FIEMAP handling in
Lustre to support multiple devices in a single filesystem, so that this
can be detected in userspace.

struct ll_fiemap_extent {
        __u64 fe_logical;  /* logical offset in bytes for the start of
                            * the extent from the beginning of the file
                            */
        __u64 fe_physical; /* physical offset in bytes for the start
                            * of the extent from the beginning of the disk
                            */
        __u64 fe_length;   /* length in bytes for this extent */
        __u64 fe_reserved64[2];
        __u32 fe_flags;    /* FIEMAP_EXTENT_* flags for this extent */
        __u32 fe_device;   /* device number for this extent */
        __u32 fe_reserved[2];
};

This adds the 32-bit "fe_device" field, which would optionally be filled
in by the filesystem (zero otherwise).  It would return the kernel device
number (i.e. st_dev), or for network filesystem (with FIEMAP_EXTENT_NET
set) this could just return an integer device number since the device
number is meaningless (and may conflict) on a remote system.

Since AFAIK Btrfs also has multiple device support there are an increasing
number of places where this would be useful.

Cheers, Andreas

> 
> Signed-off-by: Jaegeuk Kim <jaegeuk@...nel.org>
> ---
> fs/f2fs/data.c          |  55 ++++++++++++++++---
> fs/f2fs/f2fs.h          |  29 ++++++++--
> fs/f2fs/segment.c       | 119 +++++++++++++++++++++++++++++------------
> fs/f2fs/super.c         | 138 ++++++++++++++++++++++++++++++++++++++----------
> include/linux/f2fs_fs.h |  10 +++-
> 5 files changed, 277 insertions(+), 74 deletions(-)
> 
> diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c
> index 47ded0c..e2be24e 100644
> --- a/fs/f2fs/data.c
> +++ b/fs/f2fs/data.c
> @@ -88,6 +88,46 @@ static void f2fs_write_end_io(struct bio *bio)
> }
> 
> /*
> + * Return true, if pre_bio's bdev is same as its target device.
> + */
> +struct block_device *f2fs_target_device(struct f2fs_sb_info *sbi,
> +				block_t blk_addr, struct bio *bio)
> +{
> +	struct block_device *bdev = sbi->sb->s_bdev;
> +	int i;
> +
> +	for (i = 0; i < sbi->s_ndevs; i++) {
> +		if (FDEV(i).start_blk <= blk_addr &&
> +					FDEV(i).end_blk >= blk_addr) {
> +			blk_addr -= FDEV(i).start_blk;
> +			bdev = FDEV(i).bdev;
> +			break;
> +		}
> +	}
> +	if (bio) {
> +		bio->bi_bdev = bdev;
> +		bio->bi_iter.bi_sector = SECTOR_FROM_BLOCK(blk_addr);
> +	}
> +	return bdev;
> +}
> +
> +int f2fs_target_device_index(struct f2fs_sb_info *sbi, block_t blkaddr)
> +{
> +	int i;
> +
> +	for (i = 0; i < sbi->s_ndevs; i++)
> +		if (FDEV(i).start_blk <= blkaddr && FDEV(i).end_blk >= blkaddr)
> +			return i;
> +	return 0;
> +}
> +
> +static bool __same_bdev(struct f2fs_sb_info *sbi,
> +				block_t blk_addr, struct bio *bio)
> +{
> +	return f2fs_target_device(sbi, blk_addr, NULL) == bio->bi_bdev;
> +}
> +
> +/*
>  * Low-level block read/write IO operations.
>  */
> static struct bio *__bio_alloc(struct f2fs_sb_info *sbi, block_t blk_addr,
> @@ -97,8 +137,7 @@ static struct bio *__bio_alloc(struct f2fs_sb_info *sbi, block_t blk_addr,
> 
> 	bio = f2fs_bio_alloc(npages);
> 
> -	bio->bi_bdev = sbi->sb->s_bdev;
> -	bio->bi_iter.bi_sector = SECTOR_FROM_BLOCK(blk_addr);
> +	f2fs_target_device(sbi, blk_addr, bio);
> 	bio->bi_end_io = is_read ? f2fs_read_end_io : f2fs_write_end_io;
> 	bio->bi_private = is_read ? NULL : sbi;
> 
> @@ -273,7 +312,8 @@ void f2fs_submit_page_mbio(struct f2fs_io_info *fio)
> 	down_write(&io->io_rwsem);
> 
> 	if (io->bio && (io->last_block_in_bio != fio->new_blkaddr - 1 ||
> -	    (io->fio.op != fio->op || io->fio.op_flags != fio->op_flags)))
> +	    (io->fio.op != fio->op || io->fio.op_flags != fio->op_flags) ||
> +			!__same_bdev(sbi, fio->new_blkaddr, io->bio)))
> 		__submit_merged_bio(io);
> alloc_new:
> 	if (io->bio == NULL) {
> @@ -965,7 +1005,6 @@ static struct bio *f2fs_grab_bio(struct inode *inode, block_t blkaddr,
> {
> 	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
> 	struct fscrypt_ctx *ctx = NULL;
> -	struct block_device *bdev = sbi->sb->s_bdev;
> 	struct bio *bio;
> 
> 	if (f2fs_encrypted_inode(inode) && S_ISREG(inode->i_mode)) {
> @@ -983,8 +1022,7 @@ static struct bio *f2fs_grab_bio(struct inode *inode, block_t blkaddr,
> 			fscrypt_release_ctx(ctx);
> 		return ERR_PTR(-ENOMEM);
> 	}
> -	bio->bi_bdev = bdev;
> -	bio->bi_iter.bi_sector = SECTOR_FROM_BLOCK(blkaddr);
> +	f2fs_target_device(sbi, blkaddr, bio);
> 	bio->bi_end_io = f2fs_read_end_io;
> 	bio->bi_private = ctx;
> 
> @@ -1079,7 +1117,8 @@ static int f2fs_mpage_readpages(struct address_space *mapping,
> 		 * This page will go to BIO.  Do we need to send this
> 		 * BIO off first?
> 		 */
> -		if (bio && (last_block_in_bio != block_nr - 1)) {
> +		if (bio && (last_block_in_bio != block_nr - 1 ||
> +			!__same_bdev(F2FS_I_SB(inode), block_nr, bio))) {
> submit_and_realloc:
> 			__submit_bio(F2FS_I_SB(inode), bio, DATA);
> 			bio = NULL;
> @@ -1738,6 +1777,8 @@ static ssize_t f2fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
> 		return 0;
> 	if (test_opt(F2FS_I_SB(inode), LFS))
> 		return 0;
> +	if (F2FS_I_SB(inode)->s_ndevs)
> +		return 0;
> 
> 	trace_f2fs_direct_IO_enter(inode, offset, count, rw);
> 
> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
> index 9650514..1737c45 100644
> --- a/fs/f2fs/f2fs.h
> +++ b/fs/f2fs/f2fs.h
> @@ -730,6 +730,20 @@ struct f2fs_bio_info {
> 	struct rw_semaphore io_rwsem;	/* blocking op for bio */
> };
> 
> +#define FDEV(i)				(sbi->devs[i])
> +#define RDEV(i)				(raw_super->devs[i])
> +struct f2fs_dev_info {
> +	struct block_device *bdev;
> +	char path[MAX_PATH_LEN];
> +	unsigned int total_segments;
> +	block_t start_blk;
> +	block_t end_blk;
> +#ifdef CONFIG_BLK_DEV_ZONED
> +	unsigned int nr_blkz;			/* Total number of zones */
> +	u8 *blkz_type;				/* Array of zones type */
> +#endif
> +};
> +
> enum inode_type {
> 	DIR_INODE,			/* for dirty dir inode */
> 	FILE_INODE,			/* for dirty regular/symlink inode */
> @@ -778,10 +792,8 @@ struct f2fs_sb_info {
> #endif
> 
> #ifdef CONFIG_BLK_DEV_ZONED
> -	unsigned int nr_blkz;			/* Total number of zones */
> 	unsigned int blocks_per_blkz;		/* F2FS blocks per zone */
> 	unsigned int log_blocks_per_blkz;	/* log2 F2FS blocks per zone */
> -	u8 *blkz_type;				/* Array of zones type */
> #endif
> 
> 	/* for node-related operations */
> @@ -897,6 +909,8 @@ struct f2fs_sb_info {
> 
> 	/* For shrinker support */
> 	struct list_head s_list;
> +	int s_ndevs;				/* number of devices */
> +	struct f2fs_dev_info *devs;		/* for device list */
> 	struct mutex umount_mutex;
> 	unsigned int shrinker_run_no;
> 
> @@ -2159,6 +2173,9 @@ void f2fs_submit_merged_bio_cond(struct f2fs_sb_info *, struct inode *,
> void f2fs_flush_merged_bios(struct f2fs_sb_info *);
> int f2fs_submit_page_bio(struct f2fs_io_info *);
> void f2fs_submit_page_mbio(struct f2fs_io_info *);
> +struct block_device *f2fs_target_device(struct f2fs_sb_info *,
> +				block_t, struct bio *);
> +int f2fs_target_device_index(struct f2fs_sb_info *, block_t);
> void set_data_blkaddr(struct dnode_of_data *);
> void f2fs_update_data_blkaddr(struct dnode_of_data *, block_t);
> int reserve_new_blocks(struct dnode_of_data *, blkcnt_t);
> @@ -2446,11 +2463,15 @@ static inline int f2fs_sb_mounted_blkzoned(struct super_block *sb)
> 
> #ifdef CONFIG_BLK_DEV_ZONED
> static inline int get_blkz_type(struct f2fs_sb_info *sbi,
> -				block_t blkaddr)
> +			struct block_device *bdev, block_t blkaddr)
> {
> 	unsigned int zno = blkaddr >> sbi->log_blocks_per_blkz;
> +	int i;
> 
> -	return sbi->blkz_type[zno];
> +	for (i = 0; i < sbi->s_ndevs; i++)
> +		if (FDEV(i).bdev == bdev)
> +			return FDEV(i).blkz_type[zno];
> +	return -EINVAL;
> }
> #endif
> 
> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
> index 7fb7dd3..ef727d1 100644
> --- a/fs/f2fs/segment.c
> +++ b/fs/f2fs/segment.c
> @@ -403,6 +403,33 @@ void f2fs_balance_fs_bg(struct f2fs_sb_info *sbi)
> 	}
> }
> 
> +static int __submit_flush_wait(struct block_device *bdev)
> +{
> +	struct bio *bio = f2fs_bio_alloc(0);
> +	int ret;
> +
> +	bio_set_op_attrs(bio, REQ_OP_WRITE, WRITE_FLUSH);
> +	bio->bi_bdev = bdev;
> +	ret = submit_bio_wait(bio);
> +	bio_put(bio);
> +	return ret;
> +}
> +
> +static int submit_flush_wait(struct f2fs_sb_info *sbi)
> +{
> +	int ret = __submit_flush_wait(sbi->sb->s_bdev);
> +	int i;
> +
> +	if (sbi->s_ndevs && !ret) {
> +		for (i = 1; i < sbi->s_ndevs; i++) {
> +			ret = __submit_flush_wait(FDEV(i).bdev);
> +			if (ret)
> +				break;
> +		}
> +	}
> +	return ret;
> +}
> +
> static int issue_flush_thread(void *data)
> {
> 	struct f2fs_sb_info *sbi = data;
> @@ -413,25 +440,18 @@ static int issue_flush_thread(void *data)
> 		return 0;
> 
> 	if (!llist_empty(&fcc->issue_list)) {
> -		struct bio *bio;
> 		struct flush_cmd *cmd, *next;
> 		int ret;
> 
> -		bio = f2fs_bio_alloc(0);
> -
> 		fcc->dispatch_list = llist_del_all(&fcc->issue_list);
> 		fcc->dispatch_list = llist_reverse_order(fcc->dispatch_list);
> 
> -		bio->bi_bdev = sbi->sb->s_bdev;
> -		bio_set_op_attrs(bio, REQ_OP_WRITE, WRITE_FLUSH);
> -		ret = submit_bio_wait(bio);
> -
> +		ret = submit_flush_wait(sbi);
> 		llist_for_each_entry_safe(cmd, next,
> 					  fcc->dispatch_list, llnode) {
> 			cmd->ret = ret;
> 			complete(&cmd->wait);
> 		}
> -		bio_put(bio);
> 		fcc->dispatch_list = NULL;
> 	}
> 
> @@ -452,15 +472,11 @@ int f2fs_issue_flush(struct f2fs_sb_info *sbi)
> 		return 0;
> 
> 	if (!test_opt(sbi, FLUSH_MERGE) || !atomic_read(&fcc->submit_flush)) {
> -		struct bio *bio = f2fs_bio_alloc(0);
> 		int ret;
> 
> 		atomic_inc(&fcc->submit_flush);
> -		bio->bi_bdev = sbi->sb->s_bdev;
> -		bio_set_op_attrs(bio, REQ_OP_WRITE, WRITE_FLUSH);
> -		ret = submit_bio_wait(bio);
> +		ret = submit_flush_wait(sbi);
> 		atomic_dec(&fcc->submit_flush);
> -		bio_put(bio);
> 		return ret;
> 	}
> 
> @@ -637,14 +653,18 @@ static void f2fs_submit_bio_wait_endio(struct bio *bio)
> 
> /* this function is copied from blkdev_issue_discard from block/blk-lib.c */
> static int __f2fs_issue_discard_async(struct f2fs_sb_info *sbi,
> -				block_t blkstart, block_t blklen)
> +		struct block_device *bdev, block_t blkstart, block_t blklen)
> {
> -	struct block_device *bdev = sbi->sb->s_bdev;
> 	struct bio *bio = NULL;
> 	int err;
> 
> 	trace_f2fs_issue_discard(sbi->sb, blkstart, blklen);
> 
> +	if (sbi->s_ndevs) {
> +		int devi = f2fs_target_device_index(sbi, blkstart);
> +
> +		blkstart -= FDEV(devi).start_blk;
> +	}
> 	err = __blkdev_issue_discard(bdev,
> 				SECTOR_FROM_BLOCK(blkstart),
> 				SECTOR_FROM_BLOCK(blklen),
> @@ -662,18 +682,24 @@ static int __f2fs_issue_discard_async(struct f2fs_sb_info *sbi,
> }
> 
> #ifdef CONFIG_BLK_DEV_ZONED
> -static int f2fs_issue_discard_zone(struct f2fs_sb_info *sbi,
> -					block_t blkstart, block_t blklen)
> +static int __f2fs_issue_discard_zone(struct f2fs_sb_info *sbi,
> +		struct block_device *bdev, block_t blkstart, block_t blklen)
> {
> -	sector_t sector = SECTOR_FROM_BLOCK(blkstart);
> 	sector_t nr_sects = SECTOR_FROM_BLOCK(blklen);
> -	struct block_device *bdev = sbi->sb->s_bdev;
> +	sector_t sector;
> +	int devi = 0;
> 
> -	if (nr_sects != bdev_zone_size(bdev)) {
> +	if (sbi->s_ndevs) {
> +		devi = f2fs_target_device_index(sbi, blkstart);
> +		blkstart -= FDEV(devi).start_blk;
> +	}
> +	sector = SECTOR_FROM_BLOCK(blkstart);
> +
> +	if (sector % bdev_zone_size(bdev) || nr_sects != bdev_zone_size(bdev)) {
> 		f2fs_msg(sbi->sb, KERN_INFO,
> -			 "Unaligned discard attempted (sector %llu + %llu)",
> -			 (unsigned long long)sector,
> -			 (unsigned long long)nr_sects);
> +			"(%d) %s: Unaligned discard attempted (block %x + %x)",
> +			devi, sbi->s_ndevs ? FDEV(devi).path: "",
> +			blkstart, blklen);
> 		return -EIO;
> 	}
> 
> @@ -682,14 +708,12 @@ static int f2fs_issue_discard_zone(struct f2fs_sb_info *sbi,
> 	 * use regular discard if the drive supports it. For sequential
> 	 * zones, reset the zone write pointer.
> 	 */
> -	switch (get_blkz_type(sbi, blkstart)) {
> +	switch (get_blkz_type(sbi, bdev, blkstart)) {
> 
> 	case BLK_ZONE_TYPE_CONVENTIONAL:
> 		if (!blk_queue_discard(bdev_get_queue(bdev)))
> 			return 0;
> -		return __f2fs_issue_discard_async(sbi, blkstart,
> -						  blklen);
> -
> +		return __f2fs_issue_discard_async(sbi, bdev, blkstart, blklen);
> 	case BLK_ZONE_TYPE_SEQWRITE_REQ:
> 	case BLK_ZONE_TYPE_SEQWRITE_PREF:
> 		trace_f2fs_issue_reset_zone(sbi->sb, blkstart);
> @@ -702,14 +726,45 @@ static int f2fs_issue_discard_zone(struct f2fs_sb_info *sbi,
> }
> #endif
> 
> +static int __issue_discard_async(struct f2fs_sb_info *sbi,
> +		struct block_device *bdev, block_t blkstart, block_t blklen)
> +{
> +#ifdef CONFIG_BLK_DEV_ZONED
> +	if (f2fs_sb_mounted_blkzoned(sbi->sb) &&
> +				bdev_zoned_model(bdev) != BLK_ZONED_NONE)
> +		return __f2fs_issue_discard_zone(sbi, bdev, blkstart, blklen);
> +#endif
> +	return __f2fs_issue_discard_async(sbi, bdev, blkstart, blklen);
> +}
> +
> static int f2fs_issue_discard(struct f2fs_sb_info *sbi,
> 				block_t blkstart, block_t blklen)
> {
> +	sector_t start = blkstart, len = 0;
> +	struct block_device *bdev;
> 	struct seg_entry *se;
> 	unsigned int offset;
> 	block_t i;
> +	int err = 0;
> +
> +	bdev = f2fs_target_device(sbi, blkstart, NULL);
> +
> +	for (i = blkstart; i < blkstart + blklen; i++, len++) {
> +		if (i != start) {
> +			struct block_device *bdev2 =
> +				f2fs_target_device(sbi, i, NULL);
> +
> +			if (bdev2 != bdev) {
> +				err = __issue_discard_async(sbi, bdev,
> +						start, len);
> +				if (err)
> +					return err;
> +				bdev = bdev2;
> +				start = i;
> +				len = 0;
> +			}
> +		}
> 
> -	for (i = blkstart; i < blkstart + blklen; i++) {
> 		se = get_seg_entry(sbi, GET_SEGNO(sbi, i));
> 		offset = GET_BLKOFF_FROM_SEG0(sbi, i);
> 
> @@ -717,11 +772,9 @@ static int f2fs_issue_discard(struct f2fs_sb_info *sbi,
> 			sbi->discard_blks--;
> 	}
> 
> -#ifdef CONFIG_BLK_DEV_ZONED
> -	if (f2fs_sb_mounted_blkzoned(sbi->sb))
> -		return f2fs_issue_discard_zone(sbi, blkstart, blklen);
> -#endif
> -	return __f2fs_issue_discard_async(sbi, blkstart, blklen);
> +	if (len)
> +		err = __issue_discard_async(sbi, bdev, start, len);
> +	return err;
> }
> 
> static void __add_discard_entry(struct f2fs_sb_info *sbi,
> diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c
> index eca9aea..4ccbb86 100644
> --- a/fs/f2fs/super.c
> +++ b/fs/f2fs/super.c
> @@ -713,6 +713,19 @@ static void destroy_percpu_info(struct f2fs_sb_info *sbi)
> 	percpu_counter_destroy(&sbi->total_valid_inode_count);
> }
> 
> +static void destroy_device_list(struct f2fs_sb_info *sbi)
> +{
> +	int i;
> +
> +	for (i = 0; i < sbi->s_ndevs; i++) {
> +		blkdev_put(FDEV(i).bdev, FMODE_EXCL);
> +#ifdef CONFIG_BLK_DEV_ZONED
> +		kfree(FDEV(i).blkz_type);
> +#endif
> +	}
> +	kfree(sbi->devs);
> +}
> +
> static void f2fs_put_super(struct super_block *sb)
> {
> 	struct f2fs_sb_info *sbi = F2FS_SB(sb);
> @@ -773,6 +786,8 @@ static void f2fs_put_super(struct super_block *sb)
> 		crypto_free_shash(sbi->s_chksum_driver);
> 	kfree(sbi->raw_super);
> 
> +	destroy_device_list(sbi);
> +
> 	destroy_percpu_info(sbi);
> 	kfree(sbi);
> }
> @@ -1516,9 +1531,9 @@ static int init_percpu_info(struct f2fs_sb_info *sbi)
> }
> 
> #ifdef CONFIG_BLK_DEV_ZONED
> -static int init_blkz_info(struct f2fs_sb_info *sbi)
> +static int init_blkz_info(struct f2fs_sb_info *sbi, int devi)
> {
> -	struct block_device *bdev = sbi->sb->s_bdev;
> +	struct block_device *bdev = FDEV(devi).bdev;
> 	sector_t nr_sectors = bdev->bd_part->nr_sects;
> 	sector_t sector = 0;
> 	struct blk_zone *zones;
> @@ -1529,15 +1544,21 @@ static int init_blkz_info(struct f2fs_sb_info *sbi)
> 	if (!f2fs_sb_mounted_blkzoned(sbi->sb))
> 		return 0;
> 
> +	if (sbi->blocks_per_blkz && sbi->blocks_per_blkz !=
> +				SECTOR_TO_BLOCK(bdev_zone_size(bdev)))
> +		return -EINVAL;
> 	sbi->blocks_per_blkz = SECTOR_TO_BLOCK(bdev_zone_size(bdev));
> +	if (sbi->log_blocks_per_blkz && sbi->log_blocks_per_blkz !=
> +				__ilog2_u32(sbi->blocks_per_blkz))
> +		return -EINVAL;
> 	sbi->log_blocks_per_blkz = __ilog2_u32(sbi->blocks_per_blkz);
> -	sbi->nr_blkz = SECTOR_TO_BLOCK(nr_sectors) >>
> -		sbi->log_blocks_per_blkz;
> +	FDEV(devi).nr_blkz = SECTOR_TO_BLOCK(nr_sectors) >>
> +					sbi->log_blocks_per_blkz;
> 	if (nr_sectors & (bdev_zone_size(bdev) - 1))
> -		sbi->nr_blkz++;
> +		FDEV(devi).nr_blkz++;
> 
> -	sbi->blkz_type = kmalloc(sbi->nr_blkz, GFP_KERNEL);
> -	if (!sbi->blkz_type)
> +	FDEV(devi).blkz_type = kmalloc(FDEV(devi).nr_blkz, GFP_KERNEL);
> +	if (!FDEV(devi).blkz_type)
> 		return -ENOMEM;
> 
> #define F2FS_REPORT_NR_ZONES   4096
> @@ -1562,7 +1583,7 @@ static int init_blkz_info(struct f2fs_sb_info *sbi)
> 		}
> 
> 		for (i = 0; i < nr_zones; i++) {
> -			sbi->blkz_type[n] = zones[i].type;
> +			FDEV(devi).blkz_type[n] = zones[i].type;
> 			sector += zones[i].len;
> 			n++;
> 		}
> @@ -1666,6 +1687,77 @@ int f2fs_commit_super(struct f2fs_sb_info *sbi, bool recover)
> 	return err;
> }
> 
> +static int f2fs_scan_devices(struct f2fs_sb_info *sbi)
> +{
> +	struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi);
> +	int i;
> +
> +	for (i = 0; i < MAX_DEVICES; i++) {
> +		if (!RDEV(i).path[0])
> +			return 0;
> +
> +		if (i == 0) {
> +			sbi->devs = kzalloc(sizeof(struct f2fs_dev_info) *
> +						MAX_DEVICES, GFP_KERNEL);
> +			if (!sbi->devs)
> +				return -ENOMEM;
> +		}
> +
> +		memcpy(FDEV(i).path, RDEV(i).path, MAX_PATH_LEN);
> +		FDEV(i).total_segments = le32_to_cpu(RDEV(i).total_segments);
> +		if (i == 0) {
> +			FDEV(i).start_blk = 0;
> +			FDEV(i).end_blk = FDEV(i).start_blk +
> +				(FDEV(i).total_segments <<
> +				sbi->log_blocks_per_seg) - 1 +
> +				le32_to_cpu(raw_super->segment0_blkaddr);
> +		} else {
> +			FDEV(i).start_blk = FDEV(i - 1).end_blk + 1;
> +			FDEV(i).end_blk = FDEV(i).start_blk +
> +				(FDEV(i).total_segments <<
> +				sbi->log_blocks_per_seg) - 1;
> +		}
> +
> +		FDEV(i).bdev = blkdev_get_by_path(FDEV(i).path,
> +					sbi->sb->s_mode, sbi->sb->s_type);
> +		if (IS_ERR(FDEV(i).bdev))
> +			return PTR_ERR(FDEV(i).bdev);
> +
> +		/* to release errored devices */
> +		sbi->s_ndevs = i + 1;
> +
> +#ifdef CONFIG_BLK_DEV_ZONED
> +		if (bdev_zoned_model(FDEV(i).bdev) == BLK_ZONED_HM &&
> +				!f2fs_sb_mounted_blkzoned(sbi->sb)) {
> +			f2fs_msg(sbi->sb, KERN_ERR,
> +				"Zoned block device feature not enabled\n");
> +			return -EINVAL;
> +		}
> +		if (bdev_zoned_model(FDEV(i).bdev) != BLK_ZONED_NONE) {
> +			if (init_blkz_info(sbi, i)) {
> +				f2fs_msg(sbi->sb, KERN_ERR,
> +					"Failed to initialize F2FS blkzone information");
> +				return -EINVAL;
> +			}
> +			f2fs_msg(sbi->sb, KERN_INFO,
> +				"Mount Device [%2d]: %20s, %8u, %8x - %8x (zone: %s)",
> +				i, FDEV(i).path,
> +				FDEV(i).total_segments,
> +				FDEV(i).start_blk, FDEV(i).end_blk,
> +				bdev_zoned_model(FDEV(i).bdev) == BLK_ZONED_HA ?
> +				"Host-aware" : "Host-managed");
> +			continue;
> +		}
> +#endif
> +		f2fs_msg(sbi->sb, KERN_INFO,
> +			"Mount Device [%2d]: %20s, %8u, %8x - %8x",
> +				i, FDEV(i).path,
> +				FDEV(i).total_segments,
> +				FDEV(i).start_blk, FDEV(i).end_blk);
> +	}
> +	return 0;
> +}
> +
> static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> {
> 	struct f2fs_sb_info *sbi;
> @@ -1724,15 +1816,7 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> 			 "Zoned block device support is not enabled\n");
> 		goto free_sb_buf;
> 	}
> -#else
> -	if (bdev_zoned_model(sb->s_bdev) == BLK_ZONED_HM &&
> -	    !f2fs_sb_mounted_blkzoned(sb)) {
> -		f2fs_msg(sb, KERN_ERR,
> -			 "Zoned block device feature not enabled\n");
> -		goto free_sb_buf;
> -	}
> #endif
> -
> 	default_options(sbi);
> 	/* parse mount options */
> 	options = kstrdup((const char *)data, GFP_KERNEL);
> @@ -1802,6 +1886,13 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> 		goto free_meta_inode;
> 	}
> 
> +	/* Initialize device list */
> +	err = f2fs_scan_devices(sbi);
> +	if (err) {
> +		f2fs_msg(sb, KERN_ERR, "Failed to find devices");
> +		goto free_devices;
> +	}
> +
> 	sbi->total_valid_node_count =
> 				le32_to_cpu(sbi->ckpt->valid_node_count);
> 	percpu_counter_set(&sbi->total_valid_inode_count,
> @@ -1820,15 +1911,6 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> 
> 	init_ino_entry_info(sbi);
> 
> -#ifdef CONFIG_BLK_DEV_ZONED
> -	err = init_blkz_info(sbi);
> -	if (err) {
> -		f2fs_msg(sb, KERN_ERR,
> -			"Failed to initialize F2FS blkzone information");
> -		goto free_blkz;
> -	}
> -#endif
> -
> 	/* setup f2fs internal modules */
> 	err = build_segment_manager(sbi);
> 	if (err) {
> @@ -2007,10 +2089,8 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
> 	destroy_node_manager(sbi);
> free_sm:
> 	destroy_segment_manager(sbi);
> -#ifdef CONFIG_BLK_DEV_ZONED
> -free_blkz:
> -	kfree(sbi->blkz_type);
> -#endif
> +free_devices:
> +	destroy_device_list(sbi);
> 	kfree(sbi->ckpt);
> free_meta_inode:
> 	make_bad_inode(sbi->meta_inode);
> diff --git a/include/linux/f2fs_fs.h b/include/linux/f2fs_fs.h
> index 422630b..cea41a1 100644
> --- a/include/linux/f2fs_fs.h
> +++ b/include/linux/f2fs_fs.h
> @@ -52,10 +52,17 @@
> 
> #define VERSION_LEN	256
> #define MAX_VOLUME_NAME		512
> +#define MAX_PATH_LEN		64
> +#define MAX_DEVICES		8
> 
> /*
>  * For superblock
>  */
> +struct f2fs_device {
> +	__u8 path[MAX_PATH_LEN];
> +	__le32 total_segments;
> +} __packed;
> +
> struct f2fs_super_block {
> 	__le32 magic;			/* Magic Number */
> 	__le16 major_ver;		/* Major Version */
> @@ -94,7 +101,8 @@ struct f2fs_super_block {
> 	__le32 feature;			/* defined features */
> 	__u8 encryption_level;		/* versioning level for encryption */
> 	__u8 encrypt_pw_salt[16];	/* Salt used for string2key algorithm */
> -	__u8 reserved[871];		/* valid reserved region */
> +	struct f2fs_device devs[MAX_DEVICES];	/* device list */
> +	__u8 reserved[327];		/* valid reserved region */
> } __packed;
> 
> /*
> --
> 2.8.3

Cheers, Andreas

Download attachment "signature.asc" of type "application/pgp-signature" (834 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ