lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 10 Feb 2012 19:15:10 -0800
From:	Saravana Kannan <skannan@...eaurora.org>
To:	Ingo Molnar <mingo@...e.hu>
CC:	Benjamin Herrenschmidt <benh@...nel.crashing.org>,
	Todd Poynor <toddpoynor@...gle.com>,
	Russell King <linux@....linux.org.uk>,
	Peter Zijlstra <a.p.zijlstra@...llo.nl>,
	Nicolas Pitre <nico@...xnic.net>,
	Oleg Nesterov <oleg@...hat.com>, cpufreq@...r.kernel.org,
	linux-kernel@...r.kernel.org,
	Anton Vorontsov <anton.vorontsov@...aro.org>,
	linaro-kernel@...ts.linaro.org, Mike Chan <mike@...roid.com>,
	Dave Jones <davej@...hat.com>,
	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
	kernel-team@...roid.com, linux-arm-kernel@...ts.infradead.org,
	Arjan Van De Ven <arjan@...radead.org>
Subject: Re: [PATCH RFC 0/4] Scheduler idle notifiers and users

On 02/08/2012 11:51 PM, Ingo Molnar wrote:
>
> * Benjamin Herrenschmidt<benh@...nel.crashing.org>  wrote:
>
>> On Wed, 2012-02-08 at 15:23 -0500, Dave Jones wrote:
>>> I think the biggest mistake we ever made with cpufreq was making it
>>> so configurable. If we redesign it, just say no to plugin governors,
>>> and
>>> yes to a lot fewer sysfs knobs.
>>>
>>> So, provide mechanism to kill off all the governors, and there's a
>>> migration path from what we have now to something that just works
>>> in a lot more cases, while remaining configurable enough for the
>>> corner-cases.
>>
>> On the other hand, the need for schedulable contxts may not
>> necessarily go away.
>
> We will support it, but the *sane* hw solution is where
> frequency transitions can be done atomically.

I'm not sure atomicity has much to do with this. From what I can tell, 
it's about the physical characteristics of the voltage source and the 
load on said source.

After a quick digging around for some info for one of our platforms 
(ARM/MSM), it looks like it will take 200us to ramp up the power rail 
from the voltage for the lowest CPU freq to voltage for the highest CPU 
freq. And that's ignoring any communication delay. The 200us is purely 
how long it takes for the PMIC output to settle given the power load 
from the CPU. I would think other PMICs from different manufacturers 
would be in the same ballpark.

200us is a lot of time to add to a context switch or to busy wait on 
when the processors today can run at GHz speeds.

So, with what I know this doesn't look like a matter of broken HW unless 
the PMIC I'm looking up data for is a really crappy one. I'm sure other 
in the community know more about PMICs than I do and they can correct me 
if the general PMIC voltage settling characteristic is much better than 
the one I'm look at.

> Most workloads
> change their characteristics very quickly, and so does their
> power management profile change.
>
> The user-space driven policy model failed for that reason: it
> was *way* too slow in reacting) - and slow hardware transitions
> suck for a similar reason as well.

I think we all agree on this.

> We accomodate all hardware as well as we can, but we *design*
> for proper hardware. So Peter is right, this should be done
> properly.

When you say accommodate all hardware, does it mean we will keep around 
CPUfreq and allow attempts at improving it? Or we will completely move 
to scheduler based CPU freq scaling, but won't try to force atomicity? 
Say, may be queue up a notification to a CPU driver to scale up the 
frequency as soon as it can?

IMHO, I think the problem with CPUfreq and its dynamic governors today 
is that they do a timer based sampling of the CPU load instead of 
getting some hints from the scheduler when the scheduler knows that the 
load average is quite high.

-Saravana

-- 
Sent by an employee of the Qualcomm Innovation Center, Inc.
The Qualcomm Innovation Center, Inc. is a member of the Code Aurora Forum.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ