lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-id: <4D34FC5D.5080605@samsung.com>
Date:	Tue, 18 Jan 2011 11:35:09 +0900
From:	Jaehoon Chung <jh80.chung@...sung.com>
To:	Per Forlin <per.forlin@...aro.org>
Cc:	linux-mmc@...r.kernel.org, linux-arm-kernel@...ts.infradead.org,
	linux-kernel@...r.kernel.org, dev@...ts.linaro.org,
	Chris Ball <cjb@...top.org>,
	Kyungmin Park <kyungmin.park@...sung.com>
Subject: Re: [PATCH 0/5] mmc: add double buffering for mmc block requests

Hi Per..

it is interesting approach..so
we want to test your double buffering in our environment(Samsung SoC).

Did you test with SDHCI?
If you tested with SDHCI, i want to know how much increase the performance.

Thanks,
Jaehoon Chung

Per Forlin wrote:
> Add support to prepare one MMC request while another is active on
> the host. This is done by making the issue_rw_rq() asynchronous.
> The increase in throughput is proportional to the time it takes to
> prepare a request and how fast the memory is. The faster the MMC/SD is
> the more significant the prepare request time becomes. Measurements on U5500
> and U8500 on eMMC shows significant performance gain for DMA on MMC for large
> reads. In the PIO case there is some gain in performance for large reads too.
> There seems to be no or small performance gain for write, don't have a good
> explanation for this yet.
> 
> There are two optional hooks pre_req() and post_req() that the host driver
> may implement in order to improve double buffering. In the DMA case pre_req()
> may do dma_map_sg() and prepare the dma descriptor and post_req runs the
> dma_unmap_sg.
> 
> The mmci host driver implementation for double buffering is not intended
> nor ready for mainline yet. It is only an example of how to implement
> pre_req() and post_req(). The reason for this is that the basic DMA support
> for MMCI is not complete yet. The mmci patches are sent in a separate patch
> series "[FYI 0/4] arm: mmci: example implementation of double buffering".
> 
> Issues/Questions for issue_rw_rq() in block.c:
> * Is it safe to claim the host for the first MMC request and wait to release
>   it until the MMC queue is empty again? Or must the host be claimed and
>   released for every request?
> * Is it possible to predict the result from __blk_end_request().
>   If there are no errors for a completed MMC request and the
>   blk_rq_bytes(req) == data.bytes_xfered, will it be guaranteed that
>   __blk_end_request will return 0?
> 
> Here follows the IOZone results for u8500 v1.1 on eMMC.
> The numbers for DMA are a bit to good here due to the fact that the
> CPU speed is decreased compared to u8500 v2. This makes the cache handling
> even more significant.
> 
> Command line used: ./iozone -az -i0 -i1 -i2 -s 50m -I -f /iozone.tmp -e -R -+u
> 
> Relative diff: VANILLA-MMC-PIO -> 2BUF-MMC-PIO
> cpu load is abs diff
>                                                         random  random
>         KB      reclen  write   rewrite read    reread  read    write
>         51200   4       +0%     +0%     +0%     +0%     +0%     +0%
>         cpu:            +0.1    -0.1    -0.5    -0.3    -0.1    -0.0
> 
>         51200   8       +0%     +0%     +6%     +6%     +8%     +0%
>         cpu:            +0.1    -0.1    -0.3    -0.4    -0.8    +0.0
> 
>         51200   16      +0%     -2%     +0%     +0%     -3%     +0%
>         cpu:            +0.0    -0.2    +0.0    +0.0    -0.2    +0.0
> 
>         51200   32      +0%     +1%     +0%     +0%     +0%     +0%
>         cpu:            +0.1    +0.0    -0.3    +0.0    +0.0    +0.0
> 
>         51200   64      +0%     +0%     +0%     +0%     +0%     +0%
>         cpu:            +0.1    +0.0    +0.0    +0.0    +0.0    +0.0
> 
>         51200   128     +0%     +1%     +1%     +1%     +1%     +0%
>         cpu:            +0.0    +0.2    +0.1    -0.3    +0.4    +0.0
> 
>         51200   256     +0%     +0%     +1%     +1%     +1%     +0%
>         cpu:            +0.0    -0.0    +0.1    +0.1    +0.1    +0.0
> 
>         51200   512     +0%     +1%     +2%     +2%     +2%     +0%
>         cpu:            +0.1    +0.0    +0.2    +0.2    +0.2    +0.1
> 
>         51200   1024    +0%     +2%     +2%     +2%     +3%     +0%
>         cpu:            +0.2    +0.1    +0.2    +0.5    -0.8    +0.0
> 
>         51200   2048    +0%     +2%     +3%     +3%     +3%     +0%
>         cpu:            +0.0    -0.2    +0.4    +0.8    -0.5    +0.2
> 
>         51200   4096    +0%     +1%     +3%     +3%     +3%     +1%
>         cpu:            +0.2    +0.1    +0.9    +0.9    +0.5    +0.1
> 
>         51200   8192    +1%     +0%     +3%     +3%     +3%     +1%
>         cpu:            +0.2    +0.2    +1.3    +1.3    +1.0    +0.0
> 
>         51200   16384   +0%     +1%     +3%     +3%     +3%     +1%
>         cpu:            +0.2    +0.1    +1.0    +1.3    +1.0    +0.5
> 
> Relative diff: VANILLA-MMC-DMA -> 2BUF-MMC-MMCI-DMA
> cpu load is abs diff
>                                                         random  random
>         KB      reclen  write   rewrite read    reread  read    write
>         51200   4       +0%     -3%     +6%     +5%     +5%     +0%
>         cpu:            +0.0    -0.2    -0.6    -0.1    +0.3    +0.0
> 
>         51200   8       +0%     +0%     +7%     +7%     +7%     +0%
>         cpu:            +0.0    +0.1    +0.8    +0.6    +0.9    +0.0
> 
>         51200   16      +0%     +0%     +7%     +7%     +8%     +0%
>         cpu:            +0.0    -0.0    +0.7    +0.7    +0.8    +0.0
> 
>         51200   32      +0%     +0%     +8%     +8%     +9%     +0%
>         cpu:            +0.0    +0.1    +0.7    +0.7    +0.3    +0.0
> 
>         51200   64      +0%     +1%     +9%     +9%     +9%     +0%
>         cpu:            +0.0    +0.0    +0.8    +0.7    +0.8    +0.0
> 
>         51200   128     +1%     +0%     +13%    +13%    +14%    +0%
>         cpu:            +0.2    +0.0    +1.0    +1.0    +1.1    +0.0
> 
>         51200   256     +1%     +2%     +8%     +8%     +11%    +0%
>         cpu:            +0.0    +0.3    +0.0    +0.7    +1.5    +0.0
> 
>         51200   512     +1%     +2%     +16%    +16%    +17%    +0%
>         cpu:            +0.2    +0.2    +2.2    +2.1    +2.2    +0.1
> 
>         51200   1024    +1%     +2%     +20%    +20%    +20%    +1%
>         cpu:            +0.2    +0.1    +2.6    +1.9    +2.6    +0.0
> 
>         51200   2048    +0%     +2%     +22%    +22%    +21%    +0%
>         cpu:            +0.0    +0.3    +2.3    +2.9    +2.1    -0.0
> 
>         51200   4096    +1%     +2%     +23%    +23%    +23%    +1%
>         cpu:            +0.2    +0.1    +2.0    +3.2    +3.1    +0.0
> 
>         51200   8192    +1%     +5%     +24%    +24%    +24%    +1%
>         cpu:            +1.4    -0.0    +4.2    +3.0    +2.8    +0.1
> 
>         51200   16384   +1%     +3%     +24%    +24%    +24%    +2%
>         cpu:            +0.0    +0.3    +3.4    +3.8    +3.7    +0.1
> 
> Here follows the IOZone results for u5500 on eMMC.
> These numbers for DMA are more as expected.
> 
> Command line used: ./iozone -az -i0 -i1 -i2 -s 50m -I -f /iozone.tmp -e -R -+u
> 
> Relative diff: VANILLA-MMC-DMA -> 2BUF-MMC-MMCI-DMA
> cpu load is abs diff
>                                                         random  random
>         KB      reclen  write   rewrite read    reread  read    write
>         51200   128     +1%     +1%     +10%    +9%     +10%    +0%
>         cpu:            +0.1    +0.0    +1.3    +0.1    +0.8    +0.1
> 
>         51200   256     +2%     +2%     +7%     +7%     +9%     +0%
>         cpu:            +0.1    +0.4    +0.5    +0.6    +0.7    +0.0
> 
>         51200   512     +2%     +2%     +12%    +12%    +12%    +1%
>         cpu:            +0.4    +0.6    +1.8    +2.4    +2.4    +0.2
> 
>         51200   1024    +2%     +3%     +14%    +14%    +14%    +0%
>         cpu:            +0.3    +0.1    +2.1    +1.4    +1.4    +0.2
> 
>         51200   2048    +3%     +3%     +16%    +16%    +16%    +1%
>         cpu:            +0.2    +0.2    +2.5    +1.8    +2.4    -0.2
> 
>         51200   4096    +3%     +3%     +17%    +17%    +18%    +3%
>         cpu:            +0.1    -0.1    +2.7    +2.0    +2.7    -0.1
> 
>         51200   8192    +3%     +3%     +18%    +18%    +18%    +3%
>         cpu:            -0.1    +0.2    +3.0    +2.3    +2.2    +0.2
> 
>         51200   16384   +3%     +3%     +18%    +18%    +18%    +4%
>         cpu:            +0.2    +0.2    +2.8    +3.5    +2.4    -0.0
> 
> Per Forlin (5):
>   mmc: add member in mmc queue struct to hold request data
>   mmc: Add a block request prepare function
>   mmc: Add a second mmc queue request member
>   mmc: Store the mmc block request struct in mmc queue
>   mmc: Add double buffering for mmc block requests
> 
>  drivers/mmc/card/block.c |  337 ++++++++++++++++++++++++++++++----------------
>  drivers/mmc/card/queue.c |  171 +++++++++++++++---------
>  drivers/mmc/card/queue.h |   31 +++-
>  drivers/mmc/core/core.c  |   77 +++++++++--
>  include/linux/mmc/core.h |    7 +-
>  include/linux/mmc/host.h |    8 +
>  6 files changed, 432 insertions(+), 199 deletions(-)
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-mmc" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ