[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1474570233.3916.197.camel@linux.intel.com>
Date: Thu, 22 Sep 2016 11:50:33 -0700
From: Tim Chen <tim.c.chen@...ux.intel.com>
To: "Rafael J. Wysocki" <rafael@...nel.org>,
Srinivas Pandruvada <srinivas.pandruvada@...ux.intel.com>
Cc: "Rafael J. Wysocki" <rjw@...ysocki.net>,
Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...hat.com>, Borislav Petkov <bp@...e.de>,
the arch/x86 maintainers <x86@...nel.org>,
Linux PM <linux-pm@...r.kernel.org>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
ACPI Devel Maling List <linux-acpi@...r.kernel.org>,
Peter Zijlstra <peterz@...radead.org>, jolsa@...hat.com
Subject: Re: [PATCH v4 10/10] cpufreq: intel_pstate: Use CPPC to get max
performance
On Wed, 2016-09-21 at 22:30 +0200, Rafael J. Wysocki wrote:
> On Wed, Sep 21, 2016 at 9:19 PM, Srinivas Pandruvada
> <srinivas.pandruvada@...ux.intel.com> wrote:
> >
> >
> > +
> > +static void intel_pstate_check_and_enable_itmt(int cpu)
> > +{
> > + /*
> > + * For checking whether there is any difference in the maximum
> > + * performance for each CPU, need to wait till we have CPPC
> > + * data from all CPUs called from the cpufreq core. If there is a
> > + * difference in the maximum performance, then we have ITMT support.
> > + * If ITMT is supported, update the scheduler core priority for each
> > + * CPU and call to enable the ITMT feature.
> > + */
> > + if (cpumask_subset(topology_core_cpumask(cpu), &cppc_read_cpu_mask)) {
> > + int cpu_index;
> > + int max_prio;
> > + struct cpudata *cpu;
> > + bool itmt_support = false;
> > +
> > + cpu = all_cpu_data[cpumask_first(&cppc_read_cpu_mask)];
> > + max_prio = cpu->cppc_perf->highest_perf;
> > + for_each_cpu(cpu_index, &cppc_read_cpu_mask) {
> > + cpu = all_cpu_data[cpu_index];
> > + if (max_prio != cpu->cppc_perf->highest_perf) {
> > + itmt_support = true;
> > + break;
> > + }
> > + }
> > +
> > + if (!itmt_support)
> > + return;
> > +
> > + for_each_cpu(cpu_index, &cppc_read_cpu_mask) {
> > + cpu = all_cpu_data[cpu_index];
> > + sched_set_itmt_core_prio(cpu->cppc_perf->highest_perf,
> > + cpu_index);
> > + }
> My current understanding is that we need to rebuild sched domains
> after setting the priorities,
No, that's not true. We need to rebuild the sched domains only
when the sched domain flags are changed, not when we are changing
the priorities. Only the sched domain flag is a property of
the sched domain. CPU priority values are not part of sched domain.
Morten had similar question about whether we need to rebuild sched domain
when we change cpu priorities when we first post the patches.
Peter has explained that it wasn't necessary.
http://lkml.iu.edu/hypermail/linux/kernel/1608.3/01753.html
> so what if there are two CPU packages
> and there are highest_perf differences in both, and we first enumerate
> the first package entirely before getting to the second one?
>
> In that case we'll schedule the work item after enumerating the first
> package and it may rebuild the sched domains before all priorities are
> set for the second package, may it not?
That is not a problem. For the second package, all the cpu priorities
are initialized to the same value. So even if we start to do
asym_packing in the scheduler for the whole system,
on the second package, all the cpus are treated equally by the scheduler.
We will operate as if there is no favored core till we update the
priorities of the cpu on the second package.
That said, we don't enable ITMT automatically for 2 package system.
So the explicit sysctl command to enable ITMT and cause the sched domain
rebuild for 2 package system is most likely to come after
we have discovered and set all the cpu priorities.
>
> This seems to require some more consideration.
>
> >
> > + /*
> > + * Since this function is in the hotcpu notifier callback
> > + * path, submit a task to workqueue to call
> > + * sched_set_itmt_support().
> > + */
> > + schedule_work(&sched_itmt_work);
> It doesn't make sense to do this more than once IMO and what if we
> attempt to schedule the work item again when it has been scheduled
> once already? Don't we need any protection here?
It is not a problem for sched_set_itmt_support to be called more than
once.
First, we will ignore the second call if sched_itmt_capable has already
been set to the same value in the previous sched_set_itmt_support call.
Secondly, the call to update sched_itmt_capable
is protected by the itmt_update_mutex.
Thanks.
Tim
Powered by blists - more mailing lists