[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20160804182016.GC12861@quack2.suse.cz>
Date: Thu, 4 Aug 2016 20:20:16 +0200
From: Jan Kara <jack@...e.cz>
To: Dmitry Monakhov <dmonakhov@...nvz.org>
Cc: linux-ext4@...r.kernel.org, tytso@....edu
Subject: Re: [PATCH] ext4: improve ext4lazyinit scalability
Hi Dmitry!
Some spelling fixes below:
On Tue 19-07-16 16:30:32, Dmitry Monakhov wrote:
> ext4lazyinit is global thread. This thread performs itable initalization
> under
^^ li_list_mtx mutex.
> It basically does followes:
^^^^ following
> ext4_lazyinit_thread
> ->mutex_lock(&eli->li_list_mtx);
> ->ext4_run_li_request(elr)
> ->ext4_init_inode_table-> Do a lot of IO if list is large
^^ the
> And when new mounts/umount arrives they have to block on ->li_list_mtx
^^^^ mount ^^ arrive
> because lazy_thread holds it during full walk procedure.
> ext4_fill_super
> ->ext4_register_li_request
> ->mutex_lock(&ext4_li_info->li_list_mtx);
> ->list_add(&elr->lr_request, &ext4_li_info >li_request_list);
> In my case mount takes 40minutes on server with 36 * 4Tb HDD.
> Convenient user may face this in case of very slow dev ( /dev/mmcblkXXX)
^^^ Common?
> Even more. I one of filesystem was frozen lazyinit_thread will simply blocks
^^ If ^^^ filesystems block ^^
> on sb_start_write() so other mount/umounts will suck forever.
^^ umount ^^^ be stuck
> This patch changes logic like follows:
> - grap ->s_umount read sem before process new li_request after that it is safe
^^ grab ^^ processing ^^^^. After
> to drop list_mtx because all callers of li_remove_requers are holds ->s_umount
^^ li_list_mtx ^^ li_remove_request ^^ holding
> for write.
> - li_thread skip frozen SB's
^^ skips
> Locking:
> Locking order is asserted by umout path like follows: s_umount ->li_list_mtx
^^ umount
> so the only way to to grab ->s_mount inside li_thread is via down_read_trylock
^^^^^ should be just one 'to'
> Signed-off-by: Dmitry Monakhov <dmonakhov@...nvz.org>
> ---
> fs/ext4/super.c | 53 ++++++++++++++++++++++++++++++++++++-----------------
> 1 file changed, 36 insertions(+), 17 deletions(-)
>
> diff --git a/fs/ext4/super.c b/fs/ext4/super.c
> index 3822a5a..0ee193f 100644
> --- a/fs/ext4/super.c
> +++ b/fs/ext4/super.c
> @@ -2635,7 +2635,6 @@ static int ext4_run_li_request(struct ext4_li_request *elr)
> sb = elr->lr_super;
> ngroups = EXT4_SB(sb)->s_groups_count;
>
> - sb_start_write(sb);
> for (group = elr->lr_next_group; group < ngroups; group++) {
> gdp = ext4_get_group_desc(sb, group, NULL);
> if (!gdp) {
> @@ -2662,8 +2661,6 @@ static int ext4_run_li_request(struct ext4_li_request *elr)
> elr->lr_next_sched = jiffies + elr->lr_timeout;
> elr->lr_next_group = group + 1;
> }
> - sb_end_write(sb);
> -
> return ret;
> }
>
> @@ -2713,9 +2710,9 @@ static struct task_struct *ext4_lazyinit_task;
> static int ext4_lazyinit_thread(void *arg)
> {
> struct ext4_lazy_init *eli = (struct ext4_lazy_init *)arg;
> - struct list_head *pos, *n;
> struct ext4_li_request *elr;
> unsigned long next_wakeup, cur;
> + LIST_HEAD(request_list);
>
> BUG_ON(NULL == eli);
>
> @@ -2728,21 +2725,43 @@ cont_thread:
> mutex_unlock(&eli->li_list_mtx);
> goto exit_thread;
> }
> -
> - list_for_each_safe(pos, n, &eli->li_request_list) {
> - elr = list_entry(pos, struct ext4_li_request,
> - lr_request);
> -
> - if (time_after_eq(jiffies, elr->lr_next_sched)) {
> - if (ext4_run_li_request(elr) != 0) {
> - /* error, remove the lazy_init job */
> - ext4_remove_li_request(elr);
> - continue;
> + list_splice_init(&eli->li_request_list, &request_list);
Do you really need this temporary list? You could as well iterate through
the original list if you fetch the next entry after you reacquire
li_list_mtx and before you remove current entry from the list...
> + while (!list_empty(&request_list)) {
> + int err = 0;
> + int progress = 0;
> +
> + elr = list_entry(request_list.next,
> + struct ext4_li_request, lr_request);
> + list_move(request_list.next, &eli->li_request_list);
> + if (time_before(jiffies, elr->lr_next_sched)) {
> + if (time_before(elr->lr_next_sched, next_wakeup))
> + next_wakeup = elr->lr_next_sched;
> + continue;
> + }
> + if (down_read_trylock(&elr->lr_super->s_umount)) {
> + if (sb_start_write_trylock(elr->lr_super)) {
> + progress = 1;
> + /* We holds sb->s_umount, sb can not
^^ hold
Also we use the following comment style in ext4:
/*
* text here
* text here
*/
> + * be removed from the list, it is
> + * now safe to drop li_list_mtx
> + */
> + mutex_unlock(&eli->li_list_mtx);
> + err = ext4_run_li_request(elr);
> + sb_end_write(elr->lr_super);
> + mutex_lock(&eli->li_list_mtx);
> }
> + up_read((&elr->lr_super->s_umount));
> + }
> + /* error, remove the lazy_init job */
> + if (err) {
> + ext4_remove_li_request(elr);
> + continue;
> + }
> + if (!progress) {
> + elr->lr_next_sched = jiffies +
> + (prandom_u32()
> + % (EXT4_DEF_LI_MAX_START_DELAY * HZ));
> }
> -
> - if (time_before(elr->lr_next_sched, next_wakeup))
> - next_wakeup = elr->lr_next_sched;
> }
> mutex_unlock(&eli->li_list_mtx);
Otherwise the patch looks good to me.
Honza
--
Jan Kara <jack@...e.com>
SUSE Labs, CR
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists