lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Wed, 13 Sep 2023 05:41:52 +0000
From:   Johannes Thumshirn <Johannes.Thumshirn@....com>
To:     "dsterba@...e.cz" <dsterba@...e.cz>
CC:     Chris Mason <clm@...com>, Josef Bacik <josef@...icpanda.com>,
        David Sterba <dsterba@...e.com>,
        Christoph Hellwig <hch@....de>,
        Naohiro Aota <Naohiro.Aota@....com>, Qu Wenruo <wqu@...e.com>,
        Damien Le Moal <dlemoal@...nel.org>,
        "linux-btrfs@...r.kernel.org" <linux-btrfs@...r.kernel.org>,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v8 07/11] btrfs: zoned: allow zoned RAID

On 12.09.23 22:49, David Sterba wrote:
> On Mon, Sep 11, 2023 at 05:52:08AM -0700, Johannes Thumshirn wrote:
>> When we have a raid-stripe-tree, we can do RAID0/1/10 on zoned devices for
>> data block-groups. For meta-data block-groups, we don't actually need
>> anything special, as all meta-data I/O is protected by the
>> btrfs_zoned_meta_io_lock() already.
>>
>> Signed-off-by: Johannes Thumshirn <johannes.thumshirn@....com>
>> ---
>>   fs/btrfs/raid-stripe-tree.h |   7 ++-
>>   fs/btrfs/volumes.c          |   2 +
>>   fs/btrfs/zoned.c            | 113 +++++++++++++++++++++++++++++++++++++++++++-
>>   3 files changed, 119 insertions(+), 3 deletions(-)
>>
>> diff --git a/fs/btrfs/raid-stripe-tree.h b/fs/btrfs/raid-stripe-tree.h
>> index 40aa553ae8aa..30c7d5981890 100644
>> --- a/fs/btrfs/raid-stripe-tree.h
>> +++ b/fs/btrfs/raid-stripe-tree.h
>> @@ -8,6 +8,11 @@
>>   
>>   #include "disk-io.h"
>>   
>> +#define BTRFS_RST_SUPP_BLOCK_GROUP_MASK		(BTRFS_BLOCK_GROUP_DUP |\
>> +						 BTRFS_BLOCK_GROUP_RAID1_MASK |\
>> +						 BTRFS_BLOCK_GROUP_RAID0 |\
>> +						 BTRFS_BLOCK_GROUP_RAID10)
>> +
>>   struct btrfs_io_context;
>>   struct btrfs_io_stripe;
>>   
>> @@ -32,7 +37,7 @@ static inline bool btrfs_need_stripe_tree_update(struct btrfs_fs_info *fs_info,
>>   	if (type != BTRFS_BLOCK_GROUP_DATA)
>>   		return false;
>>   
>> -	if (profile & BTRFS_BLOCK_GROUP_RAID1_MASK)
>> +	if (profile & BTRFS_RST_SUPP_BLOCK_GROUP_MASK)
>>   		return true;
>>   
>>   	return false;
>> diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c
>> index 7c25f5c77788..9f17e5f290f4 100644
>> --- a/fs/btrfs/volumes.c
>> +++ b/fs/btrfs/volumes.c
>> @@ -6438,6 +6438,8 @@ int btrfs_map_block(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
>>   	 * I/O context structure.
>>   	 */
>>   	if (smap && num_alloc_stripes == 1 &&
>> +	    !(btrfs_need_stripe_tree_update(fs_info, map->type) &&
>> +	      op != BTRFS_MAP_READ) &&
>>   	    !((map->type & BTRFS_BLOCK_GROUP_RAID56_MASK) && mirror_num > 1)) {
>>   		ret = set_io_stripe(fs_info, op, logical, length, smap, map,
>>   				    stripe_index, stripe_offset, stripe_nr);
>> diff --git a/fs/btrfs/zoned.c b/fs/btrfs/zoned.c
>> index c6eedf4bfba9..4ca36875058c 100644
>> --- a/fs/btrfs/zoned.c
>> +++ b/fs/btrfs/zoned.c
>> @@ -1481,8 +1481,9 @@ int btrfs_load_block_group_zone_info(struct btrfs_block_group *cache, bool new)
>>   			set_bit(BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE, &cache->runtime_flags);
>>   		break;
>>   	case BTRFS_BLOCK_GROUP_DUP:
>> -		if (map->type & BTRFS_BLOCK_GROUP_DATA) {
>> -			btrfs_err(fs_info, "zoned: profile DUP not yet supported on data bg");
>> +		if (map->type & BTRFS_BLOCK_GROUP_DATA &&
>> +		    !btrfs_stripe_tree_root(fs_info)) {
>> +			btrfs_err(fs_info, "zoned: data DUP profile needs stripe_root");
>>   			ret = -EINVAL;
>>   			goto out;
>>   		}
>> @@ -1520,8 +1521,116 @@ int btrfs_load_block_group_zone_info(struct btrfs_block_group *cache, bool new)
>>   		cache->zone_capacity = min(caps[0], caps[1]);
>>   		break;
>>   	case BTRFS_BLOCK_GROUP_RAID1:
>> +	case BTRFS_BLOCK_GROUP_RAID1C3:
>> +	case BTRFS_BLOCK_GROUP_RAID1C4:
> 
> This
> 
>> +		if (map->type & BTRFS_BLOCK_GROUP_DATA &&
>> +		    !btrfs_stripe_tree_root(fs_info)) {
>> +			btrfs_err(fs_info,
>> +				  "zoned: data %s needs stripe_root",
>> +				  btrfs_bg_type_to_raid_name(map->type));
>> +			ret = -EIO;
>> +			goto out;
>> +
>> +		}
>> +
>> +		for (i = 0; i < map->num_stripes; i++) {
>> +			if (alloc_offsets[i] == WP_MISSING_DEV ||
>> +			    alloc_offsets[i] == WP_CONVENTIONAL)
>> +				continue;
>> +
>> +			if ((alloc_offsets[0] != alloc_offsets[i]) &&
>> +			    !btrfs_test_opt(fs_info, DEGRADED)) {
>> +				btrfs_err(fs_info,
>> +					  "zoned: write pointer offset mismatch of zones in %s profile",
>> +					  btrfs_bg_type_to_raid_name(map->type));
>> +				ret = -EIO;
>> +				goto out;
>> +			}
>> +			if (test_bit(0, active) != test_bit(i, active)) {
>> +				if (!btrfs_test_opt(fs_info, DEGRADED) &&
>> +				    !btrfs_zone_activate(cache)) {
>> +					ret = -EIO;
>> +					goto out;
>> +				}
>> +			} else {
>> +				if (test_bit(0, active))
>> +					set_bit(BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE,
>> +						&cache->runtime_flags);
>> +			}
>> +			/*
>> +			 * In case a device is missing we have a cap of 0, so don't
>> +			 * use it.
>> +			 */
>> +			cache->zone_capacity = min_not_zero(caps[0], caps[i]);
>> +		}
>> +
>> +		if (alloc_offsets[0] != WP_MISSING_DEV)
>> +			cache->alloc_offset = alloc_offsets[0];
>> +		else
>> +			cache->alloc_offset = alloc_offsets[i - 1];
> 
> whole block
> 
>> +		break;
>>   	case BTRFS_BLOCK_GROUP_RAID0:
> 
> and
> 
>> +		if (map->type & BTRFS_BLOCK_GROUP_DATA &&
>> +		    !btrfs_stripe_tree_root(fs_info)) {
>> +			btrfs_err(fs_info,
>> +				  "zoned: data %s needs stripe_root",
>> +				  btrfs_bg_type_to_raid_name(map->type));
>> +			ret = -EIO;
>> +			goto out;
>> +
>> +		}
>> +		for (i = 0; i < map->num_stripes; i++) {
>> +			if (alloc_offsets[i] == WP_MISSING_DEV ||
>> +			    alloc_offsets[i] == WP_CONVENTIONAL)
>> +				continue;
>> +
>> +			if (test_bit(0, active) != test_bit(i, active)) {
>> +				if (!btrfs_zone_activate(cache)) {
>> +					ret = -EIO;
>> +					goto out;
>> +				}
>> +			} else {
>> +				if (test_bit(0, active))
>> +					set_bit(BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE,
>> +						&cache->runtime_flags);
>> +			}
>> +			cache->zone_capacity += caps[i];
>> +			cache->alloc_offset += alloc_offsets[i];
>> +
>> +		}
>> +		break;
>>   	case BTRFS_BLOCK_GROUP_RAID10:
>> +		if (map->type & BTRFS_BLOCK_GROUP_DATA &&
>> +		    !btrfs_stripe_tree_root(fs_info)) {
>> +			btrfs_err(fs_info,
>> +				  "zoned: data %s needs stripe_root",
>> +				  btrfs_bg_type_to_raid_name(map->type));
>> +			ret = -EIO;
>> +			goto out;
>> +
>> +		}
>> +		for (i = 0; i < map->num_stripes; i++) {
>> +			if (alloc_offsets[i] == WP_MISSING_DEV ||
>> +			    alloc_offsets[i] == WP_CONVENTIONAL)
>> +				continue;
>> +
>> +			if (test_bit(0, active) != test_bit(i, active)) {
>> +				if (!btrfs_zone_activate(cache)) {
>> +					ret = -EIO;
>> +					goto out;
>> +				}
>> +			} else {
>> +				if (test_bit(0, active))
>> +					set_bit(BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE,
>> +						&cache->runtime_flags);
>> +			}
>> +			if ((i % map->sub_stripes) == 0) {
>> +				cache->zone_capacity += caps[i];
>> +				cache->alloc_offset += alloc_offsets[i];
>> +			}
>> +
>> +		}
>> +		break;
> 
> Seem to be quite long and nested for a case, can they be factored to
> helpers?

Sure, but I'd love to have
https://lore.kernel.org/all/20230605085108.580976-1-hch@lst.de/
pulled in first. This patchset handles (among other things) the DUP and 
single cases as well.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ