[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20220419213244.2104972-4-void@manifault.com>
Date: Tue, 19 Apr 2022 14:32:44 -0700
From: David Vernet <void@...ifault.com>
To: tj@...nel.org, lizefan.x@...edance.com, hannes@...xchg.org,
cgroups@...r.kernel.org, linux-kernel@...r.kernel.org
Cc: kernel-team@...com
Subject: [PATCH 3/4] cgroup: Add test_cgcpu_weight_overprovisioned() testcase
test_cpu.c includes testcases that validate the cgroup cpu controller.
This patch adds a new testcase called test_cgcpu_weight_overprovisioned()
that verifies the expected behavior of creating multiple processes with
different cpu.weight, on a system that is overprovisioned.
So as to avoid code duplication, this patch also updates cpu_hog_func_param
to take a new hog_clock_type enum which informs how time is counted in
hog_cpus_timed() (either process time or wall clock time).
Signed-off-by: David Vernet <void@...ifault.com>
---
tools/testing/selftests/cgroup/cgroup_util.c | 12 ++
tools/testing/selftests/cgroup/cgroup_util.h | 1 +
tools/testing/selftests/cgroup/test_cpu.c | 138 ++++++++++++++++++-
3 files changed, 144 insertions(+), 7 deletions(-)
diff --git a/tools/testing/selftests/cgroup/cgroup_util.c b/tools/testing/selftests/cgroup/cgroup_util.c
index 0cf7e90c0052..b690fdc8b4cd 100644
--- a/tools/testing/selftests/cgroup/cgroup_util.c
+++ b/tools/testing/selftests/cgroup/cgroup_util.c
@@ -190,6 +190,18 @@ int cg_write(const char *cgroup, const char *control, char *buf)
return -1;
}
+int cg_write_numeric(const char *cgroup, const char *control, long value)
+{
+ char buf[64];
+ int ret;
+
+ ret = sprintf(buf, "%lu", value);
+ if (ret < 0)
+ return ret;
+
+ return cg_write(cgroup, control, buf);
+}
+
int cg_find_unified_root(char *root, size_t len)
{
char buf[10 * PAGE_SIZE];
diff --git a/tools/testing/selftests/cgroup/cgroup_util.h b/tools/testing/selftests/cgroup/cgroup_util.h
index 1df13dc8b8aa..0f79156697cf 100644
--- a/tools/testing/selftests/cgroup/cgroup_util.h
+++ b/tools/testing/selftests/cgroup/cgroup_util.h
@@ -35,6 +35,7 @@ extern long cg_read_long(const char *cgroup, const char *control);
long cg_read_key_long(const char *cgroup, const char *control, const char *key);
extern long cg_read_lc(const char *cgroup, const char *control);
extern int cg_write(const char *cgroup, const char *control, char *buf);
+int cg_write_numeric(const char *cgroup, const char *control, long value);
extern int cg_run(const char *cgroup,
int (*fn)(const char *cgroup, void *arg),
void *arg);
diff --git a/tools/testing/selftests/cgroup/test_cpu.c b/tools/testing/selftests/cgroup/test_cpu.c
index 57f6308b1ef4..2afac9f9e1e2 100644
--- a/tools/testing/selftests/cgroup/test_cpu.c
+++ b/tools/testing/selftests/cgroup/test_cpu.c
@@ -2,6 +2,8 @@
#define _GNU_SOURCE
#include <linux/limits.h>
+#include <sys/sysinfo.h>
+#include <sys/wait.h>
#include <errno.h>
#include <pthread.h>
#include <stdio.h>
@@ -10,9 +12,17 @@
#include "../kselftest.h"
#include "cgroup_util.h"
+enum hog_clock_type {
+ // Count elapsed time using the CLOCK_PROCESS_CPUTIME_ID clock.
+ CPU_HOG_CLOCK_PROCESS,
+ // Count elapsed time using system wallclock time.
+ CPU_HOG_CLOCK_WALL,
+};
+
struct cpu_hog_func_param {
int nprocs;
long runtime_nsec;
+ enum hog_clock_type clock_type;
};
/*
@@ -90,8 +100,14 @@ static int hog_cpus_timed(const char *cgroup, void *arg)
{
const struct cpu_hog_func_param *param =
(struct cpu_hog_func_param *)arg;
+ long start_time;
long nsecs_remaining = param->runtime_nsec;
int i, ret;
+ struct timespec ts;
+
+ ret = clock_gettime(CLOCK_MONOTONIC, &ts);
+ if (ret != 0)
+ return ret;
for (i = 0; i < param->nprocs; i++) {
pthread_t tid;
@@ -101,22 +117,29 @@ static int hog_cpus_timed(const char *cgroup, void *arg)
return ret;
}
+ start_time = ts.tv_nsec + ts.tv_sec * NSEC_PER_SEC;
while (nsecs_remaining > 0) {
- long nsecs_so_far;
- struct timespec ts = {
- .tv_sec = nsecs_remaining / NSEC_PER_SEC,
- .tv_nsec = nsecs_remaining % NSEC_PER_SEC,
- };
+ long nsecs_so_far, baseline;
+ clockid_t clock_id;
+ ts.tv_sec = nsecs_remaining / NSEC_PER_SEC;
+ ts.tv_nsec = nsecs_remaining % NSEC_PER_SEC;
ret = nanosleep(&ts, NULL);
if (ret && errno != EINTR)
return ret;
- ret = clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &ts);
+ if (param->clock_type == CPU_HOG_CLOCK_PROCESS) {
+ clock_id = CLOCK_PROCESS_CPUTIME_ID;
+ baseline = 0;
+ } else {
+ clock_id = CLOCK_MONOTONIC;
+ baseline = start_time;
+ }
+ ret = clock_gettime(clock_id, &ts);
if (ret != 0)
return ret;
- nsecs_so_far = ts.tv_sec * NSEC_PER_SEC + ts.tv_nsec;
+ nsecs_so_far = ts.tv_sec * NSEC_PER_SEC + ts.tv_nsec - baseline;
nsecs_remaining = nsecs_so_far > param->runtime_nsec
? 0
: param->runtime_nsec - nsecs_so_far;
@@ -153,6 +176,7 @@ static int test_cgcpu_stats(const char *root)
struct cpu_hog_func_param param = {
.nprocs = 1,
.runtime_nsec = usage_seconds * NSEC_PER_SEC,
+ .clock_type = CPU_HOG_CLOCK_PROCESS,
};
if (cg_run(cgcpu, hog_cpus_timed, (void *)¶m))
goto cleanup;
@@ -174,6 +198,105 @@ static int test_cgcpu_stats(const char *root)
return ret;
}
+/*
+ * First, this test creates the following hierarchy:
+ * A
+ * A/B cpu.weight = 50
+ * A/C cpu.weight = 100
+ * A/D cpu.weight = 150
+ *
+ * A separate process is then created for each child cgroup which spawns as
+ * many threads as there are cores, and hogs each CPU as much as possible
+ * for some time interval.
+ *
+ * Once all of the children have exited, we verify that each child cgroup
+ * was given proportional runtime as informed by their cpu.weight.
+ */
+static int test_cgcpu_weight_overprovisioned(const char *root)
+{
+ struct child {
+ char *cgroup;
+ pid_t pid;
+ long usage;
+ };
+ int ret = KSFT_FAIL, i;
+ char *parent = NULL;
+ struct child children[3] = {NULL};
+ long usage_seconds = 10;
+
+ parent = cg_name(root, "cgcpu_test_0");
+ if (!parent)
+ goto cleanup;
+
+ if (cg_create(parent))
+ goto cleanup;
+
+ if (cg_write(parent, "cgroup.subtree_control", "+cpu"))
+ goto cleanup;
+
+ for (i = 0; i < ARRAY_SIZE(children); i++) {
+ children[i].cgroup = cg_name_indexed(parent, "cgcpu_child", i);
+ if (!children[i].cgroup)
+ goto cleanup;
+
+ if (cg_create(children[i].cgroup))
+ goto cleanup;
+
+ if (cg_write_numeric(children[i].cgroup, "cpu.weight",
+ 50 * (i + 1)))
+ goto cleanup;
+ }
+
+ for (i = 0; i < ARRAY_SIZE(children); i++) {
+ struct cpu_hog_func_param param = {
+ .nprocs = get_nprocs(),
+ .runtime_nsec = usage_seconds * NSEC_PER_SEC,
+ .clock_type = CPU_HOG_CLOCK_WALL,
+ };
+ pid_t pid = cg_run_nowait(children[i].cgroup, hog_cpus_timed,
+ (void *)¶m);
+ if (pid <= 0)
+ goto cleanup;
+ children[i].pid = pid;
+ }
+
+ for (i = 0; i < ARRAY_SIZE(children); i++) {
+ int retcode;
+
+ waitpid(children[i].pid, &retcode, 0);
+ if (!WIFEXITED(retcode))
+ goto cleanup;
+ if (WEXITSTATUS(retcode))
+ goto cleanup;
+ }
+
+ for (i = 0; i < ARRAY_SIZE(children); i++)
+ children[i].usage = cg_read_key_long(children[i].cgroup,
+ "cpu.stat", "usage_usec");
+
+ for (i = 0; i < ARRAY_SIZE(children) - 1; i++) {
+ long delta;
+
+ if (children[i + 1].usage <= children[i].usage)
+ goto cleanup;
+
+ delta = children[i + 1].usage - children[i].usage;
+ if (!values_close(delta, children[0].usage, 35))
+ goto cleanup;
+ }
+
+ ret = KSFT_PASS;
+cleanup:
+ for (i = 0; i < ARRAY_SIZE(children); i++) {
+ cg_destroy(children[i].cgroup);
+ free(children[i].cgroup);
+ }
+ cg_destroy(parent);
+ free(parent);
+
+ return ret;
+}
+
#define T(x) { x, #x }
struct cgcpu_test {
int (*fn)(const char *root);
@@ -181,6 +304,7 @@ struct cgcpu_test {
} tests[] = {
T(test_cgcpu_subtree_control),
T(test_cgcpu_stats),
+ T(test_cgcpu_weight_overprovisioned),
};
#undef T
--
2.30.2
Powered by blists - more mailing lists