lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Tue, 19 Jul 2022 10:50:24 +0800
From:   Coly Li <colyli@...e.de>
To:     williamsukatube@....com
Cc:     Kent Overstreet <kent.overstreet@...il.com>,
        linux-bcache@...r.kernel.org, linux-kernel@...r.kernel.org,
        William Dean <williamsukatube@...il.com>,
        Hacash Robot <hacashRobot@...tino.com>
Subject: Re: [PATCH -next] bcache: Fix spelling mistakes and cleanup code



> 2022年7月15日 15:15,williamsukatube@....com 写道:
> 
> From: William Dean <williamsukatube@...il.com>
> 
> First, fix follow spelling misktakes:
> 	automatical  ==> automatic
> 	individial  ==> around
> 	embeddded  ==> embedded
> 	addionally  ==> addtionally
> 	unncessary  ==> unnecessary
> 	definitly  ==> definitely
> Second, delete extra space or tab where
> appropriate to make code format more standardized.
> 
> Reported-by: Hacash Robot <hacashRobot@...tino.com>
> Signed-off-by: William Dean <williamsukatube@...il.com>

Hi William,

It should be better to split this patch into 2 parts, one is for typo fixes, one is for extra space or tab fixes. Then it can be more convenient for me to review and maintain.

And BTW, addionally  ==> addtionally, I guess you meant ‘additionally’, wasn’t it?

Thanks.

Coly Li

> ---
> drivers/md/bcache/bcache.h    | 2 +-
> drivers/md/bcache/bset.h      | 2 +-
> drivers/md/bcache/btree.c     | 6 +++---
> drivers/md/bcache/btree.h     | 2 +-
> drivers/md/bcache/journal.c   | 2 +-
> drivers/md/bcache/request.c   | 5 ++---
> drivers/md/bcache/stats.c     | 2 +-
> drivers/md/bcache/super.c     | 2 +-
> drivers/md/bcache/writeback.c | 2 +-
> drivers/md/bcache/writeback.h | 2 +-
> 10 files changed, 13 insertions(+), 14 deletions(-)
> 
> diff --git a/drivers/md/bcache/bcache.h b/drivers/md/bcache/bcache.h
> index 2acda9cea0f9..2b35c0a14d4d 100644
> --- a/drivers/md/bcache/bcache.h
> +++ b/drivers/md/bcache/bcache.h
> @@ -635,7 +635,7 @@ struct cache_set {
> 	struct bkey		gc_done;
> 
> 	/*
> -	 * For automatical garbage collection after writeback completed, this
> +	 * For automatic garbage collection after writeback completed, this
> 	 * varialbe is used as bit fields,
> 	 * - 0000 0001b (BCH_ENABLE_AUTO_GC): enable gc after writeback
> 	 * - 0000 0010b (BCH_DO_AUTO_GC):     do gc after writeback
> diff --git a/drivers/md/bcache/bset.h b/drivers/md/bcache/bset.h
> index d795c84246b0..76f75bbcb731 100644
> --- a/drivers/md/bcache/bset.h
> +++ b/drivers/md/bcache/bset.h
> @@ -45,7 +45,7 @@
>  * 4 in memory - we lazily resort as needed.
>  *
>  * We implement code here for creating and maintaining auxiliary search trees
> - * (described below) for searching an individial bset, and on top of that we
> + * (described below) for searching an individual bset, and on top of that we
>  * implement a btree iterator.
>  *
>  * BTREE ITERATOR:
> diff --git a/drivers/md/bcache/btree.c b/drivers/md/bcache/btree.c
> index e136d6edc1ed..3d9fec07e862 100644
> --- a/drivers/md/bcache/btree.c
> +++ b/drivers/md/bcache/btree.c
> @@ -154,7 +154,7 @@ void bch_btree_node_read_done(struct btree *b)
> 	/*
> 	 * c->fill_iter can allocate an iterator with more memory space
> 	 * than static MAX_BSETS.
> -	 * See the comment arount cache_set->fill_iter.
> +	 * See the comment around cache_set->fill_iter.
> 	 */
> 	iter = mempool_alloc(&b->c->fill_iter, GFP_NOIO);
> 	iter->size = b->c->cache->sb.bucket_size / b->c->cache->sb.block_size;
> @@ -2529,8 +2529,8 @@ int __bch_btree_map_nodes(struct btree_op *op, struct cache_set *c,
> }
> 
> int bch_btree_map_keys_recurse(struct btree *b, struct btree_op *op,
> -				      struct bkey *from, btree_map_keys_fn *fn,
> -				      int flags)
> +			       struct bkey *from, btree_map_keys_fn *fn,
> +			       int flags)
> {
> 	int ret = MAP_CONTINUE;
> 	struct bkey *k;
> diff --git a/drivers/md/bcache/btree.h b/drivers/md/bcache/btree.h
> index 1b5fdbc0d83e..b46bf6268aca 100644
> --- a/drivers/md/bcache/btree.h
> +++ b/drivers/md/bcache/btree.h
> @@ -54,7 +54,7 @@
>  * Btree nodes never have to be explicitly read in; bch_btree_node_get() handles
>  * this.
>  *
> - * For writing, we have two btree_write structs embeddded in struct btree - one
> + * For writing, we have two btree_write structs embedded in struct btree - one
>  * write in flight, and one being set up, and we toggle between them.
>  *
>  * Writing is done with a single function -  bch_btree_write() really serves two
> diff --git a/drivers/md/bcache/journal.c b/drivers/md/bcache/journal.c
> index e5da469a4235..c7f6d2611037 100644
> --- a/drivers/md/bcache/journal.c
> +++ b/drivers/md/bcache/journal.c
> @@ -790,7 +790,7 @@ static void journal_write_unlocked(struct closure *cl)
> 
> 		atomic_long_add(sectors, &ca->meta_sectors_written);
> 
> -		bio_reset(bio, ca->bdev, REQ_OP_WRITE | 
> +		bio_reset(bio, ca->bdev, REQ_OP_WRITE |
> 			  REQ_SYNC | REQ_META | REQ_PREFLUSH | REQ_FUA);
> 		bio->bi_iter.bi_sector	= PTR_OFFSET(k, i);
> 		bio->bi_iter.bi_size = sectors << 9;
> diff --git a/drivers/md/bcache/request.c b/drivers/md/bcache/request.c
> index f2c5a7e06fa9..ceba536e0880 100644
> --- a/drivers/md/bcache/request.c
> +++ b/drivers/md/bcache/request.c
> @@ -338,7 +338,6 @@ unsigned int bch_get_congested(const struct cache_set *c)
> 		return 0;
> 
> 	i += CONGESTED_MAX;
> -
> 	if (i > 0)
> 		i = fract_exp_two(i, 6);
> 
> @@ -1326,8 +1325,8 @@ static int flash_dev_ioctl(struct bcache_device *d, fmode_t mode,
> 
> void bch_flash_dev_request_init(struct bcache_device *d)
> {
> -	d->cache_miss				= flash_dev_cache_miss;
> -	d->ioctl				= flash_dev_ioctl;
> +	d->cache_miss	= flash_dev_cache_miss;
> +	d->ioctl	= flash_dev_ioctl;
> }
> 
> void bch_request_exit(void)
> diff --git a/drivers/md/bcache/stats.c b/drivers/md/bcache/stats.c
> index 68b02216033d..983d16822e53 100644
> --- a/drivers/md/bcache/stats.c
> +++ b/drivers/md/bcache/stats.c
> @@ -11,7 +11,7 @@
> #include "sysfs.h"
> 
> /*
> - * We keep absolute totals of various statistics, and addionally a set of three
> + * We keep absolute totals of various statistics, and addtionally a set of three
>  * rolling averages.
>  *
>  * Every so often, a timer goes off and rescales the rolling averages.
> diff --git a/drivers/md/bcache/super.c b/drivers/md/bcache/super.c
> index 3563d15dbaf2..aca27386cca2 100644
> --- a/drivers/md/bcache/super.c
> +++ b/drivers/md/bcache/super.c
> @@ -686,6 +686,7 @@ int bch_prio_write(struct cache *ca, bool wait)
> 
> 		ca->prio_last_buckets[i] = ca->prio_buckets[i];
> 	}
> +
> 	return 0;
> }
> 
> @@ -2663,7 +2664,6 @@ static ssize_t register_bcache(struct kobject *k, struct kobj_attribute *attr,
> 	return ret;
> }
> 
> -
> struct pdev {
> 	struct list_head list;
> 	struct cached_dev *dc;
> diff --git a/drivers/md/bcache/writeback.c b/drivers/md/bcache/writeback.c
> index 3f0ff3aab6f2..bd83a33b8a2f 100644
> --- a/drivers/md/bcache/writeback.c
> +++ b/drivers/md/bcache/writeback.c
> @@ -238,7 +238,7 @@ static void update_writeback_rate(struct work_struct *work)
> 	/*
> 	 * If the whole cache set is idle, set_at_max_writeback_rate()
> 	 * will set writeback rate to a max number. Then it is
> -	 * unncessary to update writeback rate for an idle cache set
> +	 * unnecessary to update writeback rate for an idle cache set
> 	 * in maximum writeback rate number(s).
> 	 */
> 	if (atomic_read(&dc->has_dirty) && dc->writeback_percent &&
> diff --git a/drivers/md/bcache/writeback.h b/drivers/md/bcache/writeback.h
> index 31df716951f6..37f66bea522f 100644
> --- a/drivers/md/bcache/writeback.h
> +++ b/drivers/md/bcache/writeback.h
> @@ -69,7 +69,7 @@ static inline int offset_to_stripe(struct bcache_device *d,
> 	}
> 
> 	/*
> -	 * Here offset is definitly smaller than INT_MAX,
> +	 * Here offset is definitely smaller than INT_MAX,
> 	 * return it as int will never overflow.
> 	 */
> 	return offset;
> -- 
> 2.25.1
> 

Powered by blists - more mailing lists