lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <8b471c83-8e62-bf93-b502-0bb7e9c271d6@linaro.org>
Date:   Thu, 5 Jan 2017 23:48:34 +0800
From:   Alex Shi <alex.shi@...aro.org>
To:     Daniel Lezcano <daniel.lezcano@...aro.org>,
        "Rafael J . Wysocki" <rafael.j.wysocki@...el.com>,
        vincent.guittot@...aro.org,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
        "linux-pm@...r.kernel.org" <linux-pm@...r.kernel.org>
Subject: Re: [PATCH 0/3] per cpu resume latency

Sorry for missing the mailing list.

Add linux-kernel and linux-pm.


On 01/05/2017 11:29 PM, Alex Shi wrote:
> cpu_dma_latency is designed to keep all cpu awake from deep c-state.
> That is good keep system with short response latency. But sometime we
> don't need all cpu power especially in a more and more multi-core day.
> So set all cpu restless that lead to a big power waste.
> 
> A better way is to keep the short cpu response latency on needed cpu, 
> while let other unnecesscary cpus go to deep idle. That is this 
> patchset. We just use the pm_qos_resume_latency on cpu. Giving the 
> short cpu latency on appointed cpu via setting value on
> /sys/devices/system/cpu/cpuX/power/pm_qos_resume_latency_us
> We can set we wanted latency value according to the value of
> /sys/devices/system/cpu/cpuX/cpuidle/stateX/latency. to just a bit
> less related state's latency value. Then cpu can get to this state or
> higher.
> 
> Here is some testing data on my dragonboard 410c, the latency of state1
> is 280us. It has 4 cores.
> 
> Benchmark: cyclictest -t 1  -n -i 10000 -l 1000 -q --latency=10000
> 
> without the patch:
> Latency (us) Min:     87 Act:  209 Avg:  205 Max:     239
> With the patch and cpu0/power/pm_qos_resume_latency_us is lower than
> 280us, like set to 279
> benchmark result on cpu0:
> Latency (us) Min:     82 Act:   91 Avg:   95 Max:     110
> In repeat testing, the Avg latency always drop to half of vanilla kernel
> value, as well as Max latency value, although sometime the Max latency
> is similar with vanilla kernel. 
> 
> Also we could use the cpu_dma_latency to get the similar short latency.
> But 'idlestate' show all cpu are restless. Here is the idle status 
> compression between cpu_dma_latency and this feature:
> 
> To record idlestate
> #./idlestat --trace -t 10 -f /tmp/mytracepmlat -p -c -w -- cyclictest -t 1  -n -i 10000 -l 1000 -q --latency=10000
> 
> To compare the idle state, the 'total' colum show cpu1~3 nearly stay
> in WFI state with cpu_dma_latency. but w/ my patch, they can get about
> 10 second sleep in 'spc' state.
> # ./idlestat --import -f /tmp/mytracepmlat -b /tmp/mytrace -r comparison
> Log is 10.055305 secs long with 7514 events
> Log is 10.055370 secs long with 7545 events
> --------------------------------------------------------------------------------
> | C-state  |   min    |   max    |   avg    |   total  | hits  |  over | under |
> --------------------------------------------------------------------------------
> | clusterA                                                                     |
> --------------------------------------------------------------------------------
> |      WFI |      2us |  12.88ms |   4.18ms |    9.76s |  2334 |     0 |     0 |
> |          |     -2us |  -14.4ms |    -17us |  -72.5ms |    -8 |     0 |     0 |
> --------------------------------------------------------------------------------
> |             cpu0                                                             |
> --------------------------------------------------------------------------------
> |      WFI |      3us | 100.98ms |  26.81ms |   10.03s |   374 |     0 |     0 |
> |          |     -1us |     -1us |   -350us |   +5.0ms |    +5 |     0 |     0 |
> --------------------------------------------------------------------------------
> |             cpu1                                                             |
> --------------------------------------------------------------------------------
> |      WFI |    280us |   3.96ms |   1.96ms |  19.64ms |    10 |     0 |     5 |
> |          |   +221us | -891.7ms |   -9.1ms |    -9.9s |  -889 |     0 |     0 |
> |      spc |    234us |  19.71ms |   9.79ms |    9.91s |  1012 |     4 |     0 |
> |          |   +167us |  +17.9ms |   +8.6ms |    +9.9s | +1009 |    +1 |     0 |
> --------------------------------------------------------------------------------
> |             cpu2                                                             |
> --------------------------------------------------------------------------------
> |      WFI |     86us |   1.01ms |    637us |   1.91ms |     3 |     0 |     0 |
> |          |    -16us |  -26.5ms |   -8.8ms |   -10.0s | -1057 |     0 |     0 |
> |      spc |    930us |  47.67ms |  10.05ms |    9.92s |   987 |     2 |     0 |
> |          |   -1.4ms |  +43.7ms |   +6.9ms |    +9.9s |  +985 |    +2 |     0 |
> --------------------------------------------------------------------------------
> |             cpu3                                                             |
> --------------------------------------------------------------------------------
> |      WFI |      0us |      0us |      0us |      0us |     0 |     0 |     0 |
> |          |          |    -4.0s | -152.1ms |   -10.0s |   -66 |     0 |     0 |
> |      spc |    420us |    3.50s | 913.74ms |   10.05s |    11 |     3 |     0 |
> |          |   -891us |    +3.5s | +911.0ms |   +10.0s |    +8 |    +1 |     0 |
> --------------------------------------------------------------------------------
> 
> 
> Thanks
> Alex
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ