lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <77df6298-945b-7663-0cb9-cd24fa583c6a@linaro.org>
Date:   Wed, 31 May 2017 18:40:37 +0200
From:   Daniel Lezcano <daniel.lezcano@...aro.org>
To:     Sudeep Holla <sudeep.holla@....com>
Cc:     rjw@...ysocki.net, lorenzo.pieralisi@....com, leo.yan@...aro.org,
        "open list:CPUIDLE DRIVERS" <linux-pm@...r.kernel.org>,
        open list <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] ARM: cpuidle: Support asymmetric idle definition

On 22/05/2017 17:02, Sudeep Holla wrote:
> 
> 
> On 22/05/17 15:48, Daniel Lezcano wrote:
>> On 22/05/2017 15:02, Sudeep Holla wrote:
>>
>> [ ... ]
>>
>>>>>>>> +		drv->cpumask = &cpu_topology[cpu].core_sibling;
>>>>>>>> +
>>>>>>>
>>>>>>> This is not always true and not architecturally guaranteed. So instead
>>>>>>> of introducing this broken dependency, better to extract information
>>>>>>> from the device tree.
>>>>>>
>>>>>> Can you give an example of a broken dependency ?
>>>>>>
>>>>>> The cpu topology information is extracted from the device tree. So
>>>>>> if the topology is broken, the DT is broken also. Otherwise, the
>>>>>> topology code must fix the broken dependency from the DT.
>>>>>>
>>>>>
>>>>> No, I meant there's no guarantee that all designs must follow this rule.
>>>>> I don't mean CPU topology code or binding is broken. What I meant is
>>>>> linking CPU topology to CPU power domains is wrong. We should make use
>>>>> of DT you infer this information as it's already there. Topology bindings
>>>>> makes no reference to power and hence you simply can't infer that
>>>>> information from it.
>>>>
>>>> Ok, I will have a look how power domains can fit in this.
>>>>
>>>> However I'm curious to know a platform with a cluster idle state
>>>> powering down only a subset of CPUs belonging to the cluster.
>>>>
>>>
>>> We can't reuse CPU topology for power domains:
>>> 1. As I mentioned earlier for sure, it won't be same with ARM DynamIQ.
>>> 2. Topology bindings strictly restrict themselves with topology and not
>>> connected with power-domains. We also have separate power domain
>>> bindings.
>>
>> Yes, the theory is valid, but practically nowadays I don't see where we
>> have a cluster defined by a topology with a different cluster power domain.
>>
> 
> While I agree that it's true in current practice, but in past we have
> seen "innovative designs". We initially had 2 clusters(big and little)
> then we saw 3 cluster(big little and tiny or whatever you what to call)
> So as it's not architecturally guaranteed, it's not nice to make this
> assumption in a generic driver.
> 
>> By the way, if you have any pointer to documentation for DynamIQ PM and
>> design? I would be interested to have a look.
>>
> 
> I don't have anything in detail. Excerpts from the link I sent earlier
> indicate that it's possible and highly likely.
> 
> "DynamIQ supports multiple, configurable, performance domains within a
> single cluster. These domains, consisting of single or multiple ARM
> CPUs, can scale in performance and power with finer granularity than
> previous quad-core clusters."
> 
>>> We need to separate topology and power domains. We have some dependency
>>> like this in big little drivers(both CPUfreq and CPUIdle) but that
>>> dependencies must be removed as they are not architecturally guaranteed.
>>> Lorenzo had a patch[1] to solve this issue, I can post the latest
>>> version of it again and continue the discussion after some basic
>>> rebase/testing.
>>
>> Actually, I am not convinced by the approach proposed in this patch.
>>
>> Let me have a look at the idle power domain before, I do believe we can
>> do something much more simple.
>>
> 
> OK, if you think so.

Hi Sudeep, Lorenzo,

I have been thinking and looking at the domain-idle-state and I don't
see an obvious connection between what is describing the power domain,
the cpu idle driver and what we are trying to achieve.

I would like to suggest something much more simple, register a cpuidle
driver per cpu, so every cpu can have its own idle definitions, that
should work for dynamiQ, smp and hmp. The impact on the driver will be
minimal.



-- 
 <http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs

Follow Linaro:  <http://www.facebook.com/pages/Linaro> Facebook |
<http://twitter.com/#!/linaroorg> Twitter |
<http://www.linaro.org/linaro-blog/> Blog

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ