[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <632321a8-d7f0-49a6-9577-95fac4c87b1c@arm.com>
Date: Thu, 2 May 2019 11:44:57 +0100
From: Ionela Voinescu <ionela.voinescu@....com>
To: Thara Gopinath <thara.gopinath@...aro.org>, mingo@...hat.com,
peterz@...radead.org, rui.zhang@...el.com
Cc: linux-kernel@...r.kernel.org, amit.kachhap@...il.com,
viresh.kumar@...aro.org, javi.merino@...nel.org,
edubezval@...il.com, daniel.lezcano@...aro.org,
vincent.guittot@...aro.org, nicolas.dechesne@...aro.org,
bjorn.andersson@...aro.org, dietmar.eggemann@....com
Subject: Re: [PATCH V2 0/3] Introduce Thermal Pressure
Hi Thara,
>> After cleaning it up I'm getting results around 5.6s for this test case.
>> I've run 50 iterations for each test, with 90s cool down period between
>> them.
>>
>>
>> Hackbench: (1 group , 30000 loops, 50 runs)
>> Result Standard Deviation
>> (Time Secs) (% of mean)
>>
>> No Thermal Pressure(step_wise) 5.644 7.760%
>> No Thermal Pressure(IPA) 5.677 9.062%
>>
>> Thermal Pressure Averaging
>> non-PELT Algo. Decay : 250 ms 5.627 5.593%
>> (step-wise, bigs capped only)
>>
>> Thermal Pressure Averaging
>> non-PELT Algo. Decay : 250 ms 5.690 3.738%
>> (IPA)
>>
>> All of the results above are within 1.1% difference with a
>> significantly higher standard deviation.
>
> Hi Ionela,
>
> I have replied to your original emails without seeing this one. So,
> interesting results. I see IPA is worse off (Slightly) than step wise in
> both thermal pressure and non-thermal pressure scenarios. Did you try
> 500 ms decay period by any chance?
>
I don't think we can draw a conclusion on that given how close the
results are and given the high standard deviation. Probably if I run
them again the tables will be turned :).
I did not run experiments with different decay periods yet, as I want to
have first a list of experiments that are relevant for thermal pressure,
that can help later with refining the solution, which can mean either
deciding on a decay period or possibly going with the instantaneous
thermal pressure. Please find more details below.
>>
>> I wanted to run this initially to validate my setup and understand
>> if there is any conclusion we can draw from a test like this, that
>> floods the CPUs with tasks. Looking over the traces, the tasks are
>> running almost back to back, trying to use all available resources,
>> on all the CPUs.
>> Therefore, I doubt that there could be better decisions that could be
>> made, knowing about thermal pressure, for this usecase.
>>
>> I'll try next some capacity inversion usecase and post the results when
>> they are ready.
>
I've started looking into this, starting from the most obvious case of
capacity inversion: using the user-space thermal governor and capping
the bigs to their lowest OPP. The LITTLEs are left uncapped.
This was not enough on the Hikey960 as the bigs at their lowest OPP were
in the capacity margin of the LITTLEs at their highest OPP. That meant
that LITTLEs would not pull tasks from the bigs, even if they had higher
capacity, as the capacity was in within the 25% margin. So another
change I've made was to set the capacity margin in fair.c to 10%.
I've run both sysbench and dhrystone. I'll put here only the results for
sysbench, interleaved, with and without considering thermal pressure (TP
and !TP).
As before, the TP solution uses averaging with a 250ms decay period.
Sysbench: (500000 req, 4 runs)
Result Standard Deviation
(Time Secs) (% of mean)
!TP/4 threads 146.46 0.063%
TP/4 threads 136.36 0.002%
!TP/5 threads 115.38 0.028%
TP/5 threads 110.62 0.006%
!TP/6 threads 95.38 0.051%
TP/6 threads 93.07 0.054%
!TP/7 threads 81.19 0.012%
TP/7 threads 80.32 0.028%
!TP/8 threads 72.58 2.295%
TP/8 threads 71.37 0.044%
As expected, the results are significantly improved when the scheduler
is let know of reduced capacity on the bigs which results in tasks being
placed or migrated to the littles which are able to provide better
performance. Traces nicely confirm this.
To be noted that these results only show that reflecting thermal
pressure in the capacity of the CPUs is useful and the scheduler is
equipped to make proper use of this information.
Possibly a thing to consider is whether or not to reduce the capacity
margin, but that's for another discussion.
This does not reflect the benefits of averaging, as, with the bigs
always being capped to the lowest OPP, the thermal pressure value will
be constant over the duration of the workload. The same results would
have been obtained with instantaneous thermal pressure.
Secondly, I've tried to use the step-wise governor, modified to only
cap the big CPUs, with the intention to obtain smaller periods of
capacity inversion for which a thermal pressure solution would show its
benefits.
Unfortunately dhrystone was misbehaving for some reason and was
giving me a high variation between results for the same test case.
Also, sysbench, ran with the same arguments as above, was not creating
enough load and thermal capping as to show the benefits of considering
thermal pressure.
So my recommendation is continue exploiting more test-cases like these.
I would continue with sysbench as it looks more stable, but modify the
the temperature threshold to determine periods of drastic capping of the
bigs. Once a dynamic test case and setup like this (no fixing
frequencies) is identified, it can be used to understand if averaging
is needed and to refine the decay period, and establish a good default.
What do you think? Does this make sense as a direction for obtaining
test cases? In my opinion the previous test cases were not triggering
the right behaviors that can help prove the need for thermal pressure,
or help refine it.
I will try to continue in this direction, but I won't be able to get to
in for a few days.
You'll find more results at:
https://docs.google.com/spreadsheets/d/1ibxDSSSLTodLzihNAw6jM36eVZABuPMMnjvV-Xh4NEo/edit?usp=sharing
> Sure. let me know if I can help.
Any test results or recommendations for test cases would be helpful.
The need for thermal pressure is obvious, but the way that thermal
pressure is reflected in the capacity of the CPUs could be supported by
more thorough testing.
Regards,
Ionela.
>
> Regards
> Thara
>
>>
>> Hope it helps,
>> Ionela.
>>
>>
>>> Thank you,
>>> Ionela.
>>>
>
>
Powered by blists - more mailing lists