[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <YdLaMvaM9vq4W6f1@gmail.com>
Date: Mon, 3 Jan 2022 12:12:50 +0100
From: Ingo Molnar <mingo@...nel.org>
To: Greg Kroah-Hartman <gregkh@...uxfoundation.org>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>,
linux-kernel@...r.kernel.org, linux-arch@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>,
Peter Zijlstra <peterz@...radead.org>,
Thomas Gleixner <tglx@...utronix.de>,
"David S. Miller" <davem@...emloft.net>,
Ard Biesheuvel <ardb@...nel.org>,
Josh Poimboeuf <jpoimboe@...hat.com>,
Jonathan Corbet <corbet@....net>,
Al Viro <viro@...iv.linux.org.uk>
Subject: Re: [PATCH 0000/2297] [ANNOUNCE, RFC] "Fast Kernel Headers" Tree
-v1: Eliminate the Linux kernel's "Dependency Hell"
* Greg Kroah-Hartman <gregkh@...uxfoundation.org> wrote:
> > Before going into details about how this tree solves 'dependency hell'
> > exactly, here's the current kernel build performance gain with
> > CONFIG_FAST_HEADERS=y enabled, (and with CONFIG_KALLSYMS_FAST=y enabled as
> > well - see below), using a stock x86 Linux distribution's .config with all
> > modules built into the vmlinux:
> >
> > #
> > # Performance counter stats for 'make -j96 vmlinux' (3 runs):
> > #
> > # (Elapsed time in seconds):
> > #
> >
> > v5.16-rc7: 231.34 +- 0.60 secs, 15.5 builds/hour # [ vanilla baseline ]
> > -fast-headers-v1: 129.97 +- 0.51 secs, 27.7 builds/hour # +78.0% improvement
> >
> > Or in terms of CPU time utilized:
> >
> > v5.16-rc7: 11,474,982.05 msec cpu-clock # 49.601 CPUs utilized
> > -fast-headers-v1: 7,100,730.37 msec cpu-clock # 54.635 CPUs utilized # +61.6% improvement
>
> Speed up is very impressive, nice job!
Thanks! :-)
> > Techniques used by the fast-headers tree to reduce header size & dependencies:
> >
> > - Aggressive decoupling of high level headers from each other, starting
> > with <linux/sched.h>. Since 'struct task_struct' is a union of many
> > subsystems, there's a new "per_task" infrastructure modeled after the
> > per_cpu framework, which creates fields in task_struct without having
> > to modify sched.h or the 'struct task_struct' type:
> >
> > DECLARE_PER_TASK(type, name);
> > ...
> > per_task(current, name) = val;
> >
> > The per_task() facility then seamlessly creates an offset into the
> > task_struct->per_task_area[] array, and uses the asm-offsets.h
> > mechanism to create offsets into it early in the build.
> >
> > There's no runtime overhead disadvantage from using per_task() framework,
> > the generated code is functionally equivalent to types embedded in
> > task_struct.
>
> This is "interesting", but how are you going to keep the
> kernel/sched/per_task_area_struct_defs.h and struct task_struct_per_task
> definition in sync?
I have plans to clean this up further - see below - but in general I'd
*discourage* the embedding of new complex types to task_struct.
In practice, most new task_struct fields are either simple types or
pointers to structs, which can be added to task_struct without having to
define a complex type for <linux/sched.h>.
For example here's the list of the last 5 extensions of task_struct, since
November 2020 - I copy & pasted them out of git log -p include/linux/sched.h:
+ unsigned in_eventfd_signal:1;
+ cpumask_t *user_cpus_ptr;
+ unsigned int saved_state;
+ unsigned long saved_state_change;
+ struct bpf_run_ctx *bpf_ctx;
All of those new fields are either simple C types or struct pointers - none
of those extensions need per_task() handling per se.
The overall policy to extend task_struct, going forward, would be to:
- Either make simple-type or struct-pointer additions to task_struct, that
don't couple <linux/sched.h> to other subsystems.
- Or, if you absolutely must - and we don't want to forbid this - use the
per_task() machinery to create a simple accessor to a complex embedded
type.
> [...] It seems that you manually created this (which is great for
> testing), but over the long-term, trying to manually determine what needs
> to be done here to keep everything lined up properly is going to be a
> major pain.
Note that under the policy above - and even according to the practice of
the last ~1.5 years - it should be exceedingly rare having to extend the
per_task() facility.
There's one thing ugly about it, the fixed PER_TASK_BYTES limit, I plan to
make ->per_task_array[] the last field of task_struct, i.e. change it to:
u8 per_task_area[];
This actually became possible through the fixing of the x86 FPU code in the
following fast-headers commit:
4ae0f28bc1c8 headers/deps: x86/fpu: Make task_struct::thread constant size
In the last ~1 year existence of the per_task() facility I didn't have any
maintenance troubles with these fields getting out of sync, but we could
also auto-generate kernel/sched/per_task_area_struct_defs.h from
kernel/sched/per_task_area_struct.h via a build-time script, and make
kernel/sched/per_task_area_struct.h the only method to define such fields.
> That issue aside, I took a glance at the tree, and overall it looks like
> a lot of nice cleanups. Most of these can probably go through the
> various subsystem trees, after you split them out, for the "major" .h
> cleanups. Is that something you are going to be planning on doing?
Yeah, I absolutely plan on doing that too:
- About ~70% of the commits can be split up & parallelized through
maintainer trees.
- With the exception of the untangling of sched.h, per_task and the
"Optimize Headers" series, where a lot of patches are dependent on each
other. These are actually needed to get any measurable benefits from this
tree (!). We can do these through the scheduler tree, or through the
dedicated headers tree I posted.
The latter monolithic series is pretty much unavoidable, it's the result of
30 years of coupling a lot of kernel subsystems to task_struct via embedded
structs & other complex types, that needed quite a bit of effort to
untangle, and that untangling needed to happen in-order.
Do these plans this sound good to you?
Thanks,
Ingo
Powered by blists - more mailing lists