lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <9c57c92c-3e0c-b8c5-4be9-8f4df344a347@linux.vnet.ibm.com>
Date:   Tue, 14 Feb 2023 20:54:09 +0530
From:   shrikanth hegde <sshegde@...ux.vnet.ibm.com>
To:     mingo@...hat.com, peterz@...radead.org,
        Vincent Guittot <vincent.guittot@...aro.org>
Cc:     dietmar.eggemann@....com, bsegall@...gle.com,
        Thomas Gleixner <tglx@...utronix.de>,
        Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
        Arjan van de Ven <arjan@...ux.intel.com>,
        svaidy@...ux.ibm.com, linux-kernel@...r.kernel.org,
        shrikanth hegde <sshegde@...ux.vnet.ibm.com>
Subject: [RFC PATCH] sched/fair: Interleave cfs bandwidth timers for improved
 single thread performance at low utilization

CPU cfs bandwidth controller uses hrtimer called period timer. Quota is
refilled upon the timer expiry and re-started when there are running tasks
within the cgroup. Each cgroup has a separate period timer which manages
the period and quota for that cgroup.

start_cfs_bandwidth calls hrtimer_forward_now which set the expiry value
based on the below logic. expiry = $initial_value + $N * $period

However, start_cfs_bandwidth doesn't set any initial value. Hence
multiple such timers would align on expiry if their period value is
same. This happens when there are multiple cgroups and each has runnable
task. Upon expiry each timer will unthrottle respective rq's and all the
rq would start at same time, competing for CPU time and use all
the SMT threads likely.

There is performance gain that can be achieved here if the timers are
interleaved when the utilization of each CPU cgroup is low and total
utilization of all the CPU cgroup's is less than 50%. This is likely
true when using containers. If the timers are interleaved, then the
unthrottled cgroup can run freely without many context switches and can
also benefit from SMT Folding[1]. This effect will be further amplified in
SPLPAR environment[2] as this would cause less hypervisor preemptions.
There can be benefit due to less IPI storm as well. Docker provides a
config option of period timer value, whereas the kubernetes only
provides millicore option. Hence with typical deployment period values
will be set to 100ms as kubernetes millicore will set the quota
accordingly without altering period values.

[1] SMT folding is a mechanism where processor core is reconfigured to
lower SMT mode to improve performance when some sibling threads are
idle. In a SMT8 core, when only one or two threads are running on a
core, we get the best throughput compared to running all 8 threads.

[2] SPLPAR is an Shared Processor Logical PARtition. There can be many
SPLPARs running on the same physical machine sharing the CPU resources.
One SPLPAR can consume all CPU resource it can, if the other SPLPARs are
idle. Processors within the SPLPAR are called vCPU. vCPU can be higher
than CPU.  Hence at an instance of time if there are more requested vCPU
than CPU, then vCPU can be preempted. When the timers align, there will
be spike in requested vCPU when the timers expire. This can lead to
preemption when the other SPLPARs are not idle.

Since we are trading off between the performance vs power here,
benchmarked both the numbers. Frequency is set to 3.00Ghz and
socket power has been measured. Ran the stress-ng with two
cgroups. The numbers are with patch and without patch on a Power
system with SMT=8. Below table shows time taken by each group to
complete. Here each cgroup is assigned 25% runtime. period value is
set to 100ms.

workload: stress-ng --cpu=4 --cpu-ops=50000
data shows time it took to complete in seconds for each run.
Tried to interleave by best effort with the patch.
1CG - time to finish when only 1 cgroup is running.
2CG - time to finish when 2 cgroups are running together.
power - power consumed in Watts for the socket running the workload.
Performance gain is indicated in +ve percentage numbers and power
increase is indicated in -ve numbers. 1CG numbers are same as expected.
We are looking at improvement in 2CG Mainly.

             6.2.rc5                           with patch
        1CG    power   2CG    power   | 1CG  power     2CG        power
1Core   218     44     315      46    | 219    45    277(+12%)    47(-2%)
        219     43     315      45    | 219    44    244(+22%)    48(-6%)
	                              |
2Core   108     48     158      52    | 109    50    114(+26%)    59(-13%)
        109     49     157      52    | 109    49    136(+13%)    56(-7%)
                                      |
4Core    60     59      89      65    |  62    58     72(+19%)    68(-5%)
         61     61      90      65    |  62    60     68(+24%)    73(-12%)
                                      |
8Core    33     77      48      83    |  33    77     37(+23%)    91(-10%)
         33     77      48      84    |  33    77     38(+21%)    90(-7%)

There is no benefit at higher utilization of 50% or more. There is no
degradation also.

This is RFC PATCH V2, where the code has been shifted from hrtimer to
sched. This patch sets an initial value as multiple of period/10.
Here timers can still align if the time started the cgroup is within the
period/10 interval. On a real life workload, time gives sufficient
randomness. There can be a better interleaving by being more
deterministic. For example, when there are 2 cgroups, they should
have initial value of 0/50ms or 10/60ms so on. When there are 3 cgroups,
0/3/6ms or 1/4/7ms etc. That is more complicated as it has to account
for cgroup addition/deletion and accuracy w.r.t to period/quota.
If that approach is better here, then will come up with that patch.

Signed-off-by: Shrikanth Hegde<sshegde@...ux.vnet.ibm.com>
---
 kernel/sched/fair.c | 17 ++++++++++++++---
 1 file changed, 14 insertions(+), 3 deletions(-)

diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index ff4dbbae3b10..7b69c329e05d 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -5939,14 +5939,25 @@ static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq)

 void start_cfs_bandwidth(struct cfs_bandwidth *cfs_b)
 {
-	lockdep_assert_held(&cfs_b->lock);
+	struct hrtimer *period_timer = &cfs_b->period_timer;
+	s64 incr = ktime_to_ns(cfs_b->period) / 10;
+	ktime_t delta;
+	u64 orun = 1;

+	lockdep_assert_held(&cfs_b->lock);
 	if (cfs_b->period_active)
 		return;

 	cfs_b->period_active = 1;
-	hrtimer_forward_now(&cfs_b->period_timer, cfs_b->period);
-	hrtimer_start_expires(&cfs_b->period_timer, HRTIMER_MODE_ABS_PINNED);
+	delta = ktime_sub(period_timer->base->get_time(),
+			hrtimer_get_expires(period_timer));
+	if (unlikely(delta >= cfs_b->period)) {
+		orun = ktime_divns(delta, incr);
+		hrtimer_add_expires_ns(period_timer, incr * orun);
+	}
+
+	hrtimer_forward_now(period_timer, cfs_b->period);
+	hrtimer_start_expires(period_timer, HRTIMER_MODE_ABS_PINNED);
 }

 static void destroy_cfs_bandwidth(struct cfs_bandwidth *cfs_b)
--
2.31.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ