[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <alpine.DEB.2.02.1306120916300.5954@nftneq.ynat.uz>
Date: Wed, 12 Jun 2013 09:22:06 -0700 (PDT)
From: David Lang <david@...g.hm>
To: Amit Kucheria <amit.kucheria@...aro.org>
cc: Arjan van de Ven <arjan@...ux.intel.com>,
"len.brown@...el.com" <len.brown@...el.com>,
"alex.shi@...el.com" <alex.shi@...el.com>,
"corbet@....net" <corbet@....net>,
Peter Zijlstra <peterz@...radead.org>,
Catalin Marinas <catalin.marinas@....com>,
Linux PM list <linux-pm@...r.kernel.org>,
"Rafael J. Wysocki" <rjw@...ysocki.net>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
Morten Rasmussen <Morten.Rasmussen@....com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
linaro-kernel <linaro-kernel@...ts.linaro.org>,
Mike Galbraith <efault@....de>,
Preeti U Murthy <preeti@...ux.vnet.ibm.com>,
Andrew Morton <akpm@...ux-foundation.org>,
"pjt@...gle.com" <pjt@...gle.com>, Ingo Molnar <mingo@...nel.org>
Subject: Re: power-efficient scheduling design
On Wed, 12 Jun 2013, Amit Kucheria wrote:
> On Wed, Jun 12, 2013 at 7:18 AM, Arjan van de Ven <arjan@...ux.intel.com> wrote:
>> On 6/11/2013 5:27 PM, David Lang wrote:
>>>
>>>
>>> Nobody is saying that this sort of thing should be in the fastpath of the
>>> scheduler.
>>>
>>> But if the scheduler has a table that tells it the possible states, and
>>> the cost to get from the current state to each of these states (and to get
>>> back and/or wake up to
>>> full power), then the scheduler can make the decision on what to do,
>>> invoke a routine to make the change (and in the meantime, not be fighting
>>> the change by trying to
>>> schedule processes on a core that's about to be powered off), and then
>>> when the change happens, the scheduler will have a new version of the table
>>> of possible states and costs
>>>
>>> This isn't in the fastpath, it's in the rebalancing logic.
>>
>>
>> the reality is much more complex unfortunately.
>> C and P states hang together tightly, and even C state on
>> one core impacts other cores' performance, just like P state selection
>> on one core impacts other cores.
>>
>> (at least for x86, we should really stop talking as if the OS picks the
>> "frequency",
>> that's just not the case anymore)
>
> This is true of ARM platforms too. As Daniel pointed out in an earlier
> email, the operating point (frequency, voltage) has a bearing on the
> c-state latency too.
>
> An additional complexity is thermal constraints. E.g. On a quad-core
> Cortex-A15 processor capable of say 1.5GHz, you won't be able to run
> all 4 cores at that speed for very long w/o exceeding the thermal
> envelope. These overdrive frequencies (turbo in x86-speak) impact the
> rest of the system by either constraining the frequency of other cores
> or requiring aggresive thermal management.
>
> Do we really want to track these details in the scheduler or just let
> the scheduler provide notifications to the existing subsystems
> (cpufreq, cpuidle, thermal, etc.) with some sort of feedback going
> back to the scheduler to influence future decisions?
>
> Feeback to the scheduler could be something like the following (pardon
> the names):
>
> 1. ISOLATE_CORE: Don't schedule anything on this core - cpuidle might
> use this to synchronise cores for a cluster shutdown, thermal
> framework could use this as idle injection to reduce temperature
> 2. CAP_CAPACITY: Don't expect cpufreq to raise the frequency on this
> core - cpufreq might use this to cap overall energy since overdrive
> operating points are very expensive, thermal might use this to slow
> down rate of increase of die temperature
How much data are you going to have to move back and forth between the different
systems?
do you really only want the all-or-nothing "use this core as much as possible"
vs "don't use this core at all"? or do you need the ability to indicate how much
to use a particular core (something that is needed anyway for asymetrical cores
I think)
If there is too much information that needs to be moved back and forth between
these 'subsystems' for the 'right' thing to happen, then it would seem like it
makes more sense to combine them.
Even combined, there are parts that are still pretty modular (like the details
of shifting from one state to another, and the different high level strategies
to follow for different modes of operation), but having access to all the
information rather than only bits and pieces of the information at lower
granularity would seem like an improvement.
David Lang
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists