[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20110509143702.GK1661@htj.dyndns.org>
Date: Mon, 9 May 2011 16:37:02 +0200
From: Tejun Heo <tj@...nel.org>
To: Vivek Goyal <vgoyal@...hat.com>
Cc: Shaohua Li <shaohua.li@...el.com>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"linux-ide@...r.kernel.org" <linux-ide@...r.kernel.org>,
"jaxboe@...ionio.com" <jaxboe@...ionio.com>,
"hch@...radead.org" <hch@...radead.org>,
"jgarzik@...ox.com" <jgarzik@...ox.com>,
"djwong@...ibm.com" <djwong@...ibm.com>,
"sshtylyov@...sta.com" <sshtylyov@...sta.com>,
James Bottomley <James.Bottomley@...senPartnership.com>,
"linux-scsi@...r.kernel.org" <linux-scsi@...r.kernel.org>,
"ricwheeler@...il.com" <ricwheeler@...il.com>
Subject: Re: [patch v3 2/3] block: hold queue if flush is running for
non-queueable flush drive
Hello, Vivek, Shaohua.
On Mon, May 09, 2011 at 09:58:51AM -0400, Vivek Goyal wrote:
> > > I am assuming that these back-to-back flushes are independent of each
> > > other otherwise write request will anyway get between two flushes.
Yeap.
> > > If that's the case, then should we solve the problem by improving flush
> > > merge logic a bit better. (Say idle a bit before issuing a flush only
> > > if request queue is not empty).
Maybe, I don't know. Maybe we can implement some sort of adaptive
delay logic which detects parallel stream of flush requests and try to
insert artificial delays between flush issues to maximize throughput.
With such workload, such optimization probably wouldn't hurt latency
that much either.
However, such heuristics tend to be fairly finicky and result in
unexpected behavior when workload isn't of the expected pattern and
that's why I stuck with mostly mechanical mechanism we currently have
for the initial implementation, but please feel free to play with it.
It would be awesome if such logic can be implemented with something
mechanical which doesn't involve a lot of magic tunables, timers and
pattern detection logics.
> If we try to get rid of WRITE completely between these 10 flushes
> then we run the risk of starving other READS/WRITES as long as
> flushes are going on.
Back-to-back flushes are very cheap. It just costs command
issue/completion roundtrip latency and if that's a problem we can
optimize that out too, so I don't think stream of back-to-back empty
flushes can be a problem. For flushes with data, nothing prevents
other IOs from issued while flush writes are being sequenced. It will
look like,
FLUSH -> FLUSH and other data -> FLUSH -> FLUSH -> FLUSH and other data...
Before this patch, it was
FLUSH -> FLUSH and other data -> FLUSH -> other data -> FLUSH -> FLUSH and other data...
The change isn't gonna cause starvation.
> I did not understand this one. With improved back to back merge logic
> we have already optimized the flush case. So for 10 flush and one write
> we seem to be issuing following (as per your mail).
>
> 1 flush (6 flush merged)--> WRITE --> 1flush (4 flush merged).
>
> So where is the opprotinutiy for drive (non flush queuing drive) to optimize
> flush here?
I suspect the workload doesn't really swamp the device with fsync's
and there frequently are brief holes without any pending flush. In
such cases, flush sequencer will start sequencing immediately and the
following flushes would lose a chance to be merged and the intervening
write causes the performance hit. It would be interesting to
investigate this deeper tho.
> IOW, if flush merging is already working well, do we really want to move
> in a direction where we can potentially starve other READ/WRITE happening
> in an attempt to improve throughput for a sepecific workload.
I don't think it's gonna starve anything and actually like it as it
effectively implements another side of mechnical merging. Before the
patch, it only tried to merge requests which already shared a FLUSH.
ie. It delayed post-FLUSH until all writes which shared pre-FLUSH
finished. This change doesn't introduce any latency on idle queue
while allowing effective merging of flushes which immediately follows
the first one, and there's no knobs to adjust - timings are regulated
by how fast the device can handle flushes and IOs.
So, whether we add further merge strategy or not, I think this one can
serve nicely as one of the base logics and would really like to see
how it affects higher end devices.
Please note that really highend devices with battery backed buffer
wouldn't care about all of this anyway. I think we would mostly be
looking at SCSI hard drives or cheap arrays where I don't think
issuing other IOs together with flush would bring a lot of benefits.
So, I'd like to enable the logic by default unless it actively hurts
those devices.
Thank you.
--
tejun
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists