[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <200607252218.01663.jesper.juhl@gmail.com>
Date: Tue, 25 Jul 2006 22:18:01 +0200
From: Jesper Juhl <jesper.juhl@...il.com>
To: linux-kernel@...r.kernel.org
Cc: Jens Axboe <axboe@...e.de>, Andrea Arcangeli <andrea@...e.de>,
Roman Zippel <zippel@...ux-m68k.org>,
Sam Ravnborg <sam@...nborg.org>
Subject: [PATCH] update I/O sched Kconfig help texts - CFQ is now default, not AS.
Change I/O scheduler description to correctly show CFQ as being the default
scheduler and not the anticipatory scheduler that previously was default.
Signed-off-by: Jesper Juhl <jesper.juhl@...il.com>
---
block/Kconfig.iosched | 9 +++++----
1 files changed, 5 insertions(+), 4 deletions(-)
--- linux-2.6.18-rc2-git5-orig/block/Kconfig.iosched 2006-07-18 18:46:18.000000000 +0200
+++ linux-2.6.18-rc2-git5/block/Kconfig.iosched 2006-07-25 22:12:08.000000000 +0200
@@ -15,10 +15,10 @@ config IOSCHED_AS
tristate "Anticipatory I/O scheduler"
default y
---help---
- The anticipatory I/O scheduler is the default disk scheduler. It is
- generally a good choice for most environments, but is quite large and
- complex when compared to the deadline I/O scheduler, it can also be
- slower in some cases especially some database loads.
+ The anticipatory I/O scheduler is generally a good choice for most
+ environments, but is quite large and complex when compared to the
+ deadline I/O scheduler, it can also be slower in some cases
+ especially some database loads.
config IOSCHED_DEADLINE
tristate "Deadline I/O scheduler"
@@ -37,6 +37,7 @@ config IOSCHED_CFQ
The CFQ I/O scheduler tries to distribute bandwidth equally
among all processes in the system. It should provide a fair
working environment, suitable for desktop systems.
+ This is the default I/O scheduler.
choice
prompt "Default I/O scheduler"
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists