lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20211104145713.4419-3-laoar.shao@gmail.com>
Date:   Thu,  4 Nov 2021 14:57:11 +0000
From:   Yafang Shao <laoar.shao@...il.com>
To:     mingo@...hat.com, peterz@...radead.org, juri.lelli@...hat.com,
        vincent.guittot@...aro.org, dietmar.eggemann@....com,
        rostedt@...dmis.org, bsegall@...gle.com, mgorman@...e.de,
        bristot@...hat.com
Cc:     linux-kernel@...r.kernel.org, Yafang Shao <laoar.shao@...il.com>,
        Valentin Schneider <valentin.schneider@....com>
Subject: [RFC PATCH 2/4] sched/fair: Introduce cfs_migration

A new per-cpu kthread named "cfs_migration/N" is introduced to do
cfs specific balance works. It is a FIFO task with priority FIFO-1,
which means it can preempt any cfs tasks but can't preempt other FIFO
tasks. The kthreads as follows,

    PID     COMMAND
    13      [cfs_migration/0]
    20      [cfs_migration/1]
    25      [cfs_migration/2]
    32      [cfs_migration/3]
    38      [cfs_migration/4]
    ...

    $ cat /proc/13/sched
    ...
    policy                                       :                    1
    prio                                         :                   98
    ...

    $ cat /proc/13/status
    ...
    Cpus_allowed:	0001
    Cpus_allowed_list:	0
    ...

All the works need to be done will be queued into a singly linked list,
in which the first queued will be first serviced.

Signed-off-by: Yafang Shao <laoar.shao@...il.com>
Cc: Valentin Schneider <valentin.schneider@....com>
Cc: Vincent Guittot <vincent.guittot@...aro.org>
Cc: Dietmar Eggemann <dietmar.eggemann@....com>
---
 kernel/sched/fair.c | 93 +++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 93 insertions(+)

diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 87db481e8a56..56b3fa91828b 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -20,6 +20,8 @@
  *  Adaptive scheduling granularity, math enhancements by Peter Zijlstra
  *  Copyright (C) 2007 Red Hat, Inc., Peter Zijlstra
  */
+#include <linux/smpboot.h>
+#include <linux/stop_machine.h>
 #include "sched.h"
 
 /*
@@ -11915,3 +11917,94 @@ int sched_trace_rq_nr_running(struct rq *rq)
         return rq ? rq->nr_running : -1;
 }
 EXPORT_SYMBOL_GPL(sched_trace_rq_nr_running);
+
+#ifdef CONFIG_SMP
+struct cfs_migrater {
+	struct task_struct *thread;
+	struct list_head works;
+	raw_spinlock_t lock;
+};
+
+DEFINE_PER_CPU(struct cfs_migrater, cfs_migrater);
+
+static int cfs_migration_should_run(unsigned int cpu)
+{
+	struct cfs_migrater *migrater = &per_cpu(cfs_migrater, cpu);
+	unsigned long flags;
+	int run;
+
+	raw_spin_lock_irqsave(&migrater->lock, flags);
+	run = !list_empty(&migrater->works);
+	raw_spin_unlock_irqrestore(&migrater->lock, flags);
+
+	return run;
+}
+
+static void cfs_migration_setup(unsigned int cpu)
+{
+	/* cfs_migration should have a higher priority than normal tasks,
+	 * but a lower priority than other FIFO tasks.
+	 */
+	sched_set_fifo_low(current);
+}
+
+static void cfs_migrater_thread(unsigned int cpu)
+{
+	struct cfs_migrater *migrater = &per_cpu(cfs_migrater, cpu);
+	struct cpu_stop_work *work;
+
+repeat:
+	work = NULL;
+	raw_spin_lock_irq(&migrater->lock);
+	if (!list_empty(&migrater->works)) {
+		work = list_first_entry(&migrater->works,
+					struct cpu_stop_work, list);
+		list_del_init(&work->list);
+	}
+	raw_spin_unlock_irq(&migrater->lock);
+
+	if (work) {
+		struct cpu_stop_done *done = work->done;
+		cpu_stop_fn_t fn = work->fn;
+		void *arg = work->arg;
+		int ret;
+
+		preempt_count_inc();
+		ret = fn(arg);
+		if (done) {
+			if (ret)
+				done->ret = ret;
+			cpu_stop_signal_done(done);
+		}
+		preempt_count_dec();
+		goto repeat;
+	}
+}
+
+static struct smp_hotplug_thread cfs_migration_threads = {
+	.store			= &cfs_migrater.thread,
+	.setup			= cfs_migration_setup,
+	.thread_fn		= cfs_migrater_thread,
+	.thread_comm		= "cfs_migration/%u",
+	.thread_should_run	= cfs_migration_should_run,
+};
+
+static int __init cfs_migration_init(void)
+{
+	struct cfs_migrater *cm = &per_cpu(cfs_migrater, raw_smp_processor_id());
+	unsigned int cpu;
+
+	for_each_possible_cpu(cpu) {
+		struct cfs_migrater *migrater = &per_cpu(cfs_migrater, cpu);
+
+		raw_spin_lock_init(&migrater->lock);
+		INIT_LIST_HEAD(&migrater->works);
+	}
+
+	WARN_ON_ONCE(smpboot_register_percpu_thread(&cfs_migration_threads));
+	kthread_unpark(cm->thread);
+
+	return 0;
+}
+early_initcall(cfs_migration_init)
+#endif
-- 
2.17.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ