[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <201305252130.27645.amitkale@geeksofpune.in>
Date: Sat, 25 May 2013 21:30:27 +0530
From: Amit Kale <amitkale@...ksofpune.in>
To: Jens Axboe <axboe@...nel.dk>
Cc: Amit Kale <akale@...c-inc.com>,
OS Engineering <osengineering@...c-inc.com>,
LKML <linux-kernel@...r.kernel.org>,
Padmini Balasubramaniyan <padminib@...c-inc.com>,
Amit Phansalkar <aphansalkar@...c-inc.com>,
koverstreet@...gle.com, linux-bcache@...r.kernel.org,
thornber@...hat.com, dm-devel@...hat.com
Subject: Re: EnhanceIO(TM) caching driver features [1/3]
On Saturday 25 May 2013, Jens Axboe wrote:
> Please don't top post!
Got to use a different email client for that. Note that I am writing this from
my personal email address. This email and any future emails I write from this
address are my personal views and sTec can't be held responsible for them.
> On Sat, May 25 2013, Amit Kale wrote:
> > Hi Jens,
> >
> > I by mistake dropped the weblink to demartek study while composing my
> > email. The demartek study is published here:
> > http://www.demartek.com/Demartek_STEC_S1120_PCIe_Evaluation_2013-02.html.
> > It's an independent study. Here are a few numbers taken from this
> > report. In a database comparison using transactions per second
> >
> > HDD baseline (40 disks) - 2570 tps
> > 240GB Cache - 9844 tps
> > 480GB cache - 19758 tps
> > RAID5 pure SSD - 32380 tps
> > RAID0 pure SSD - 40467 tps
> >
> > There are two types of performance comparisons, application based and
> > IO pattern based. Application based tests measure efficiency of cache
> > replacement algorithms. These are time consuming. Above tests were
> > done by demartek over a period of time. I don't have performance
> > comparisons between EnhanceIO(TM) driver, bcache and dm-cache. I'll
> > try to get them done in-house.
>
> Unless I'm badly mistaken, that study is only on enhanceio, it does not
> compare it to any other solutions.
That's correct. I haven't seen any application level benchmark based
comparisons between different caching solutions on any platform.
> Additionally, it's running on
> Windows?!
Yes. However as I have said above, application level testing is primarly a
test of cache replacement algorithm. So the effect of a platform is less,
although not zero.
> I don't think it's too much to ask to see results on the
> operating system for which you are submitting the changes.
Agreed that's a fair thing to ask.
>
> > IO pattern based tests can be done quickly. However since IO pattern
> > is fixed prior to the test, output tends to depend on whether the IO
> > pattern suits the caching algorithm. These are relatively easy. I can
> > definitely post this comparison.
>
> It's fairly trivial to do some synthetic cache testing with fio, using
> eg the zipf distribution. That'll get you data reuse, for both reads and
> writes (if you want), in the selected distribution.
While the running a test is trivial deciding what IO patterns to run is a
difficult problem. The bottom line for sequential, random, zipf and pareto is
the same - they all test a fixed IO pattern, which at best is very unlike an
application pattern. Cache behavior affects IO addresses that an application
issues. The block list of IOs requested by an application is different when
running on HDD, SSD and cache. Memory usage by cache is one significant factor
in this effect. IO latency (more and less) also affects when multiple threads
are processing transactions.
Regardless of which IO pattern is used following characteristics to a large
extent measure efficiency of a cache engine minus cache replacement algorithm.
Hit rate - Can be varied between 0 to 100%. 90% and above are of interest for
caching.
Read versus Write mix - Can be varied from 0/100, 10/90, ...., 10/90, 0/100.
IO block size - Fixed equal to or a multiple of cache block size. Variable
complicates analysis of results.
Does following comparison sound interesting? I welcome others to propose
modifications or other ways.
Cache block size 4kB.
HDD size = 500GB
SSD size = 100GB and 500GB.
HDD equal to SSD is only to study cache behavior so all of the tests below
need not be performed.
For read-only cache mode
This works best for write intensive loads R/W mix - 10/90 and 30/70
Write intensive loads are usually long writes Block size - 64kB
Cache hit ratio 0%, 90%, 95%, 100%.
For write-through cache mode
This works best for read intensive loads R/W mix 100/0, 90/10
Block size - 4kB, 16kB, 128kB.
Cache hit ratio 90%, 95%, 100%
For write-back cache mode
This works best for fluctuating loads R/W mix 90/10
Block size - 4kB
Cache hit ratio 95%.
Thanks.
-Amit
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists