lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 8 Nov 2010 21:26:07 -0500
From:	Vivek Goyal <vgoyal@...hat.com>
To:	Shaohua Li <shaohua.li@...el.com>
Cc:	lkml <linux-kernel@...r.kernel.org>,
	Jens Axboe <jaxboe@...ionio.com>,
	"czoccolo@...il.com" <czoccolo@...il.com>
Subject: Re: [patch 2/3]cfq-iosched: schedule dispatch for noidle queue

On Mon, Nov 08, 2010 at 09:15:40PM -0500, Vivek Goyal wrote:
> On Tue, Nov 09, 2010 at 09:31:07AM +0800, Shaohua Li wrote:
> > On Mon, 2010-11-08 at 22:28 +0800, Vivek Goyal wrote:
> > > On Mon, Nov 08, 2010 at 10:07:18AM +0800, Shaohua Li wrote:
> > > > A queue is idle at cfq_dispatch_requests(), but it gets noidle later. Unless
> > > > other task explictly does unplug or all requests are drained, we will not
> > > > deliever requests to the disk even cfq_arm_slice_timer doesn't make the
> > > > queue idle. For example, cfq_should_idle() returns true because of
> > > > service_tree->count == 1, and then other queues are added. Note, I didn't
> > > > see obvious performance impacts so far with the patch, but just thought
> > > > this could be a problem.
> > > > 
> > > > Signed-off-by: Shaohua Li <shaohua.li@...el.com>
> > > > 
> > > > ---
> > > >  block/cfq-iosched.c |   23 ++++++++++++++++++++++-
> > > >  1 file changed, 22 insertions(+), 1 deletion(-)
> > > > 
> > > > Index: linux/block/cfq-iosched.c
> > > > ===================================================================
> > > > --- linux.orig/block/cfq-iosched.c	2010-11-08 08:41:20.000000000 +0800
> > > > +++ linux/block/cfq-iosched.c	2010-11-08 08:43:51.000000000 +0800
> > > > @@ -3265,6 +3265,10 @@ cfq_should_preempt(struct cfq_data *cfqd
> > > >  	if (cfq_class_rt(new_cfqq) && !cfq_class_rt(cfqq))
> > > >  		return true;
> > > >  
> > > > +	/* An idle queue should not be idle now for some reason */
> > > > +	if (RB_EMPTY_ROOT(&cfqq->sort_list) && !cfq_should_idle(cfqd, cfqq))
> > > > +		return true;
> > > > +
> > > >  	if (!cfqd->active_cic || !cfq_cfqq_wait_request(cfqq))
> > > >  		return false;
> > > >  
> > > > @@ -3508,8 +3512,25 @@ static void cfq_completed_request(struct
> > > >  		}
> > > >  	}
> > > >  
> > > > -	if (!cfqd->rq_in_driver)
> > > > +	if (!cfqd->rq_in_driver) {
> > > >  		cfq_schedule_dispatch(cfqd);
> > > > +		return;
> > > > +	}
> > > > +	/*
> > > > +	 * A queue is idle at cfq_dispatch_requests(), but it gets noidle
> > > > +	 * later. We schedule a dispatch if the queue has no requests,
> > > > +	 * otherwise the disk is actually in idle till all requests
> > > > +	 * are finished even cfq_arm_slice_timer doesn't make the queue idle
> > > > +	 * */
> > > 
> > > Why do we have to wait for all requests to finish in device? Will driver
> > > most likely not ask for next request when 1-2 requests have completed
> > > and at that time we should expire the queue if queue is no more marked
> > > as "noidle"?
> > The issue is a queue is idle just because it's the last queue of the
> > service tree. Then a new queue is added and the idled queue should not
> > idle now. we should preempt the idled queue soon. does this make sense
> > to you?
> 
> If that's the case then you should just modify should_preempt() so that
> addition of a new queue could preempt an empty queue which has now become
> noidle.
> 
> You have also modified cfq_completed_request() function, which will wake
> up the worker thread and then try to dispatch a request. IMHO, in practice
> driver asks for new request almost immediately and you don't gain much
> by this additional wakeup.
> 
> So my point being, that we increased the code complexity for no visible
> performance improvement also increased thread wakeups resulting in more
> cpu consumption.
> 
> If there was a visible performance gain in your testing then it would have
> made sense but you said that you did not notice any improvements. Then
> why to increase the complexity.

Thinking more about it. I suspect modification to cfq_completed_request()
is a dead code. In the sense, status of a queue will change from idle to
noidle, when a new request comes in (for most of the cases). If that's
the case then we don't have to modify cfq_completed_request() code at
all. An extra condition in cfq_should_preempt() should cover the case of
preempting the existing empty queue which is noidle.

Can you give me an example in what cases cfq_completed_request() will
really wake up the thread which will result in early dispatch of request
then it would have been if we relied on driver asking for next request.
One such example could be that somebody set slice_idle=0 but that's one
time event and I really don't think we should be optimizing for that 
corner case.

Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ