[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <e3421211-c59d-a372-2e27-68f452f4f8b1@linaro.org>
Date: Mon, 25 Feb 2019 16:01:09 +0100
From: Daniel Lezcano <daniel.lezcano@...aro.org>
To: "Rafael J. Wysocki" <rafael@...nel.org>
Cc: "Rafael J. Wysocki" <rjw@...ysocki.net>,
Ulf Hansson <ulf.hansson@...aro.org>,
Linux PM <linux-pm@...r.kernel.org>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] cpuidle: Add a predict callback for the governors
Hi Rafael,
On 22/02/2019 11:35, Rafael J. Wysocki wrote:
> On Thu, Feb 21, 2019 at 6:40 PM Daniel Lezcano
> <daniel.lezcano@...aro.org> wrote:
>>
>> On 21/02/2019 17:18, Rafael J. Wysocki wrote:
>>> On Thu, Feb 21, 2019 at 3:56 PM Daniel Lezcano
>>> <daniel.lezcano@...aro.org> wrote:
>>>>
>>>> Predicting the next event on the current CPU is implemented in the
>>>> idle state selection function, thus the selection logic and the
>>>> prediction are tied together and it is hard to decorrelate both.
>>>>
>>>> The following change introduces the cpuidle function to give the
>>>> opportunity to the governor to store the guess estimate of the
>>>> different source of wakeup and then reuse them in the selection
>>>> process. Consequently we end up with two separate operations clearly
>>>> identified.
>>>>
>>>> As the next events are stored in the cpuidle device structure it is
>>>> easy to propagate them in the different governor callbacks.
>>>
>>> Can you explain a bit how you would use this new callback in a governor?
>>
>> Sure.
>>
>> Today we have the selection and the prediction tied together. The
>> prediction is modulated with some inputs coming from the governor's
>> policy (eg. performance multiplier).
>>
>> It is hard to know if the prediction is correct or not, hard to know the
>> duration of the computation for the next event and hard to know if the
>> idle state selection succeeded because of a good prediction or a good
>> governor policy.
>>
>> I propose to provide the callback where we fill the guess estimated next
>> events on the system, so we can trace them and benchmark the computation
>> time.
>>
>> The selection of the idle state becomes an separate action where we can
>> apply any specific governor heuristic or policy.
>>
>> By separating the selection and the prediction, we can identify where
>> the weakness is in our test scenario: the prediction or the governor
>> selection policy.
>
> I'm not quite convinced about the idea that the "prediction" and
> "selection" parts can be entirely separate.
>
> Besides, this new callback doesn't return anything, it goes before
> ->select and the latter is invoked unconditionally anyway. That's
> extra overhead (even if small) for no real gain. I don't see why it
> is better to do the split in the core rather than in the governor
> itself. You can always design a governor ->select() to call two
> separate functions, one after another, internally, so why to do that
> in the core and inflict that particular design choice on everybody?
It is a way to clearly identify what is part of the prediction and what
is part of the decision in order to easily spot the weakness of the
governor. We may be doing good predictions but bad decisions or the
opposite.
I agree we are not forced to create a new callback for this and we can
create a prediction function directly inside the governor. I don't have
a strong preference actually and I'm fine with your proposal.
> For example, there's no "predict" part running before the "select" one
> in the TEO governor at all. There is something like that in menu, but
> splitting it off would be rather artificial IMO.
>
> Next, the cpuidle_predict structure. I'm not even sure why it is there at all.
>
> Presumably, the purpose of it is to pass some data from ->predict to
> ->select, but why does it have to be there in struct cpuidle_device?
> The governor appears to be the only user of it anyway, so wouldn't it
> be better to store that information in the governor's internal data
> structures if really needed?
At one moment we will need the prediction information from the scheduler
in order to optimize the cpu selection for the wakeup. I thought the
cpuidle device structure can be a place where we can store it.
There are a lot of optimizations we can do after knowing when a CPU is
expected to wakeup.
Do you have a suggestion where to store the next wakeup for a CPU?
> Moreover, why does it include the next_hrtimer and next_timer fields?
> It looks like you'd need to do some nohz code surgery to separate the
> two and would that be useful even? And even so, these values need not
> be predicted.
I agree they don't need to be predicted but they are part of the sources
of wake up with a deterministic behavior and fall under the prediction
umbrella as they are part of the equation for the next event.
The purpose of filling these fields is to give the select callback all
the clues to take its decision.
But regarding your POV, which is valid, we can consider using an
internal prediction function and just export the next event without a
full description of the different wake up source categories.
> It is known when the next timer event is going to
> occur, both with and without the scheduler tick included, and that's
> why there is tick_nohz_get_sleep_length().
The next hrtimer and the next timer are the deadline versions of
tick_nohz_get_sleep_length(), respectively the delta_time parameter and
the returned value.
I massaged the tick_nohz_get_sleep_length() to return the deadline
rather than the remaining time. The changes will come with the patches
Ulf is about to send. IMHO, it helps to understand the code by splitting
in two functions rather than passing an extra parameter.
> If you want to do a similar thing for non-timer interrupts in order to
> use the IRQ stats, you can simply provide a function similar to
> tick_nohz_get_sleep_length() for that, but that's available already,
> isn't it?
Actually, I'm interested in deadlines not in relative remaining time
because we need later to compare those when we are about to wake up a
CPU or enter the cluster state.
Indeed, the irq timings is already available but with the limitation of
predicting regular intervals. I rewrote the code to handle both regular
and repeating pattern. I'll post the series as soon as I have the numbers.
> Also, I'm not really sure what next_resched is and how exactly it is
> going to be computed.
The next_resched estimation is to handle the situation where you have
one CPU handling IO related interrupts for a specific device waking up
another CPU with the io blocked task. It is very similar to the avg_idle
value.
--
<http://www.linaro.org/> Linaro.org │ Open source software for ARM SoCs
Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook |
<http://twitter.com/#!/linaroorg> Twitter |
<http://www.linaro.org/linaro-blog/> Blog
Powered by blists - more mailing lists