[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <1303202686.3981.216.camel@sli10-conroe>
Date: Tue, 19 Apr 2011 16:44:46 +0800
From: Shaohua Li <shaohua.li@...el.com>
To: lkml <linux-kernel@...r.kernel.org>,
linux-ide <linux-ide@...r.kernel.org>
Cc: Jens Axboe <jaxboe@...ionio.com>, Jeff Garzik <jgarzik@...ox.com>,
Christoph Hellwig <hch@...radead.org>,
Tejun Heo <htejun@...il.com>
Subject: [PATCH 1/2]block: optimize non-queueable flush request drive
In some drives, flush requests are non-queueable. This means when a flush request
is running, normal read/write requests are not. In such drive, when running flush
requests finish, we can make pending flush requests finish. Since normal write
requests are not running, pending flush requests also flush required drive cache
out. This reduces some flush requests running and improve performance.
This patch allows block core utilizes the optimization. Next patch will enable it
for SATA.
Signed-off-by: Shaohua Li <shaohua.li@...el.com>
---
block/blk-flush.c | 15 +++++++++++++--
include/linux/blkdev.h | 11 +++++++++++
2 files changed, 24 insertions(+), 2 deletions(-)
Index: linux/block/blk-flush.c
===================================================================
--- linux.orig/block/blk-flush.c 2011-04-19 09:21:47.000000000 +0800
+++ linux/block/blk-flush.c 2011-04-19 16:38:22.000000000 +0800
@@ -193,18 +193,29 @@ static bool blk_flush_complete_seq(struc
static void flush_end_io(struct request *flush_rq, int error)
{
struct request_queue *q = flush_rq->q;
- struct list_head *running = &q->flush_queue[q->flush_running_idx];
+ LIST_HEAD(proceed_list);
bool queued = false;
struct request *rq, *n;
BUG_ON(q->flush_pending_idx == q->flush_running_idx);
+ list_splice_init(&q->flush_queue[q->flush_running_idx], &proceed_list);
+ /*
+ * If queue doesn't support queueable flush request, we can push the
+ * pending requests to the next stage too. For such queue, there are no
+ * normal requests running when flush request is running, so this still
+ * guarantees the correctness.
+ */
+ if (!blk_queue_flush_queueable(q))
+ list_splice_tail_init(&q->flush_queue[q->flush_pending_idx],
+ &proceed_list);
+
/* account completion of the flush request */
q->flush_running_idx ^= 1;
elv_completed_request(q, flush_rq);
/* and push the waiting requests to the next stage */
- list_for_each_entry_safe(rq, n, running, flush.list) {
+ list_for_each_entry_safe(rq, n, &proceed_list, flush.list) {
unsigned int seq = blk_flush_cur_seq(rq);
BUG_ON(seq != REQ_FSEQ_PREFLUSH && seq != REQ_FSEQ_POSTFLUSH);
Index: linux/include/linux/blkdev.h
===================================================================
--- linux.orig/include/linux/blkdev.h 2011-04-19 09:15:15.000000000 +0800
+++ linux/include/linux/blkdev.h 2011-04-19 10:04:46.000000000 +0800
@@ -366,6 +366,7 @@ struct request_queue
* for flush operations
*/
unsigned int flush_flags;
+ unsigned int flush_not_queueable:1;
unsigned int flush_pending_idx:1;
unsigned int flush_running_idx:1;
unsigned long flush_pending_since;
@@ -552,6 +553,16 @@ static inline void blk_clear_queue_full(
queue_flag_clear(QUEUE_FLAG_ASYNCFULL, q);
}
+static inline void blk_set_queue_flush_queueable(struct request_queue *q,
+ bool queueable)
+{
+ q->flush_not_queueable = !queueable;
+}
+
+static inline bool blk_queue_flush_queueable(struct request_queue *q)
+{
+ return !q->flush_not_queueable;
+}
/*
* mergeable request must not have _NOMERGE or _BARRIER bit set, nor may
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists