lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20100409021054.13511.52377.stgit@austin.mtv.corp.google.com>
Date:	Thu, 08 Apr 2010 19:19:00 -0700
From:	Divyesh Shah <dpshah@...gle.com>
To:	jens.axboe@...cle.com, vgoyal@...hat.com
Cc:	linux-kernel@...r.kernel.org, nauman@...gle.com,
	ctalbott@...gle.com
Subject: [PATCH] cfq-iosched: Fix the incorrect timeslice accounting with
	forced_dispatch.

When CFQ dispatches requests forcefully due to a barrier or changing iosched,
it runs through all cfqq's dispatching requests and then expires each queue.
However, it does not activate a cfqq before flushing its IOs resulting in
using stale values for computing slice_used.
This patch fixes it by calling activate queue before flushing reuqests from
each queue.

This is useful mostly for barrier requests because when the iosched is changing
it really doesnt matter if we have incorrect accounting since we're going to
break down all structures anyway.

We also now expire the current timeslice before moving on with the dispatch
to accurately account slice used for that cfqq.

Signed-off-by: Divyesh Shah<dpshah@...gle.com>
---
Side question that is related:
(W/o the change to expire the current timeslice) If there is a currently active
queue which has no requests pending and is idling and we enter forced_dispatch,
it seems to me that it is pure chance that we are not hitting the BUG_ON for
cfqd->busy_queues in cfq_forced_dispatch(). The current active queue (which is
on rr and included in the busy_queues count) can appear in any order from
get_next_queue_forced() and it may well happen that all non-empty cfqqs were
dispatched from before it and by the time we get to this queue cfqd->rq_queued
goes down to zero and we bail out. Hence, __cfq_slice_expired() never gets
called for this cfqq and it is not taken off from rr which should result in
cfqd->busy_queues to be non-zero and the BUG_ON I mentioned earlier should be
hit.
Does this sound correct?

 block/cfq-iosched.c |    7 +++++--
 1 files changed, 5 insertions(+), 2 deletions(-)

diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index 9102ffc..39b9a36 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -2196,10 +2196,13 @@ static int cfq_forced_dispatch(struct cfq_data *cfqd)
 	struct cfq_queue *cfqq;
 	int dispatched = 0;
 
-	while ((cfqq = cfq_get_next_queue_forced(cfqd)) != NULL)
+	/* Expire the timeslice of the current active queue first */
+	cfq_slice_expired(cfqd, 0);
+	while ((cfqq = cfq_get_next_queue_forced(cfqd)) != NULL) {
+		__cfq_set_active_queue(cfqd, cfqq);
 		dispatched += __cfq_forced_dispatch_cfqq(cfqq);
+	}
 
-	cfq_slice_expired(cfqd, 0);
 	BUG_ON(cfqd->busy_queues);
 
 	cfq_log(cfqd, "forced_dispatch=%d", dispatched);

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ