lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Thu, 4 Aug 2016 20:20:16 +0200
From:	Jan Kara <jack@...e.cz>
To:	Dmitry Monakhov <dmonakhov@...nvz.org>
Cc:	linux-ext4@...r.kernel.org, tytso@....edu
Subject: Re: [PATCH] ext4: improve ext4lazyinit scalability

Hi Dmitry!

Some spelling fixes below:

On Tue 19-07-16 16:30:32, Dmitry Monakhov wrote:
> ext4lazyinit is global thread. This thread performs itable initalization
> under
        ^^ li_list_mtx mutex.

> It basically does followes:
                    ^^^^ following

> ext4_lazyinit_thread
>   ->mutex_lock(&eli->li_list_mtx);
>   ->ext4_run_li_request(elr)
>     ->ext4_init_inode_table-> Do a lot of IO if list is large
						^^ the

> And when new mounts/umount arrives they have to block on ->li_list_mtx
               ^^^^ mount      ^^ arrive

> because  lazy_thread holds it during full walk procedure.
> ext4_fill_super
>  ->ext4_register_li_request
>    ->mutex_lock(&ext4_li_info->li_list_mtx);
>    ->list_add(&elr->lr_request, &ext4_li_info >li_request_list);
> In my case mount takes 40minutes on server with 36 * 4Tb HDD.
> Convenient user may face this in case of very slow dev ( /dev/mmcblkXXX)
   ^^^ Common?

> Even more. I one of filesystem was frozen lazyinit_thread will simply blocks
             ^^ If     ^^^ filesystems				    block ^^

> on sb_start_write() so other mount/umounts will suck forever.
					^^ umount ^^^ be stuck

> This patch changes logic like follows:
> - grap ->s_umount read sem before process new li_request after that it is safe
    ^^ grab                         ^^ processing         ^^^^. After
>   to drop list_mtx because all callers of li_remove_requers are holds ->s_umount
            ^^ li_list_mtx                   ^^ li_remove_request  ^^ holding
>   for write.
> - li_thread skip frozen SB's
              ^^ skips

> Locking:
> Locking order is asserted by umout path like follows: s_umount ->li_list_mtx
                               ^^ umount

> so the only way to to grab ->s_mount inside li_thread is via down_read_trylock
                  ^^^^^ should be just one 'to'

> Signed-off-by: Dmitry Monakhov <dmonakhov@...nvz.org>

> ---
>  fs/ext4/super.c | 53 ++++++++++++++++++++++++++++++++++++-----------------
>  1 file changed, 36 insertions(+), 17 deletions(-)
> 
> diff --git a/fs/ext4/super.c b/fs/ext4/super.c
> index 3822a5a..0ee193f 100644
> --- a/fs/ext4/super.c
> +++ b/fs/ext4/super.c
> @@ -2635,7 +2635,6 @@ static int ext4_run_li_request(struct ext4_li_request *elr)
>  	sb = elr->lr_super;
>  	ngroups = EXT4_SB(sb)->s_groups_count;
>  
> -	sb_start_write(sb);
>  	for (group = elr->lr_next_group; group < ngroups; group++) {
>  		gdp = ext4_get_group_desc(sb, group, NULL);
>  		if (!gdp) {
> @@ -2662,8 +2661,6 @@ static int ext4_run_li_request(struct ext4_li_request *elr)
>  		elr->lr_next_sched = jiffies + elr->lr_timeout;
>  		elr->lr_next_group = group + 1;
>  	}
> -	sb_end_write(sb);
> -
>  	return ret;
>  }
>  
> @@ -2713,9 +2710,9 @@ static struct task_struct *ext4_lazyinit_task;
>  static int ext4_lazyinit_thread(void *arg)
>  {
>  	struct ext4_lazy_init *eli = (struct ext4_lazy_init *)arg;
> -	struct list_head *pos, *n;
>  	struct ext4_li_request *elr;
>  	unsigned long next_wakeup, cur;
> +	LIST_HEAD(request_list);
>  
>  	BUG_ON(NULL == eli);
>  
> @@ -2728,21 +2725,43 @@ cont_thread:
>  			mutex_unlock(&eli->li_list_mtx);
>  			goto exit_thread;
>  		}
> -
> -		list_for_each_safe(pos, n, &eli->li_request_list) {
> -			elr = list_entry(pos, struct ext4_li_request,
> -					 lr_request);
> -
> -			if (time_after_eq(jiffies, elr->lr_next_sched)) {
> -				if (ext4_run_li_request(elr) != 0) {
> -					/* error, remove the lazy_init job */
> -					ext4_remove_li_request(elr);
> -					continue;
> +		list_splice_init(&eli->li_request_list, &request_list);

Do you really need this temporary list? You could as well iterate through
the original list if you fetch the next entry after you reacquire
li_list_mtx and before you remove current entry from the list...

> +		while (!list_empty(&request_list)) {
> +			int err = 0;
> +			int progress = 0;
> +
> +			elr = list_entry(request_list.next,
> +					 struct ext4_li_request, lr_request);
> +			list_move(request_list.next, &eli->li_request_list);
> +			if (time_before(jiffies, elr->lr_next_sched)) {
> +				if (time_before(elr->lr_next_sched, next_wakeup))
> +					next_wakeup = elr->lr_next_sched;
> +				continue;
> +			}
> +			if (down_read_trylock(&elr->lr_super->s_umount)) {
> +				if (sb_start_write_trylock(elr->lr_super)) {
> +					progress = 1;
> +					/* We holds sb->s_umount, sb can not
						^^ hold

Also we use the following comment style in ext4:

/*
 * text here
 * text here
 */

> +					 * be removed from the list, it is
> +					 * now safe to drop li_list_mtx
> +					 */
> +					mutex_unlock(&eli->li_list_mtx);
> +					err = ext4_run_li_request(elr);
> +					sb_end_write(elr->lr_super);
> +					mutex_lock(&eli->li_list_mtx);
>  				}
> +				up_read((&elr->lr_super->s_umount));
> +			}
> +			/* error, remove the lazy_init job */
> +			if (err) {
> +				ext4_remove_li_request(elr);
> +				continue;
> +			}
> +			if (!progress) {
> +				elr->lr_next_sched = jiffies +
> +					(prandom_u32()
> +					 % (EXT4_DEF_LI_MAX_START_DELAY * HZ));
>  			}
> -
> -			if (time_before(elr->lr_next_sched, next_wakeup))
> -				next_wakeup = elr->lr_next_sched;
>  		}
>  		mutex_unlock(&eli->li_list_mtx);

Otherwise the patch looks good to me.

									Honza
-- 
Jan Kara <jack@...e.com>
SUSE Labs, CR
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ