lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 3 Sep 2020 14:36:36 -0400
From:   Phil Auld <pauld@...hat.com>
To:     Marcelo Tosatti <mtosatti@...hat.com>
Cc:     Frederic Weisbecker <frederic@...nel.org>,
        Peter Zijlstra <peterz@...radead.org>,
        Thomas Gleixner <tglx@...utronix.de>,
        "Paul E. McKenney" <paulmck@...nel.org>,
        Joel Fernandes <joel@...lfernandes.org>,
        linux-kernel@...r.kernel.org
Subject: Re: Requirements to control kernel isolation/nohz_full at runtime

On Thu, Sep 03, 2020 at 03:30:15PM -0300 Marcelo Tosatti wrote:
> On Thu, Sep 03, 2020 at 03:23:59PM -0300, Marcelo Tosatti wrote:
> > On Tue, Sep 01, 2020 at 12:46:41PM +0200, Frederic Weisbecker wrote:
> > > Hi,
> > 
> > Hi Frederic,
> > 
> > Thanks for the summary! Looking forward to your comments...
> > 
> > > I'm currently working on making nohz_full/nohz_idle runtime toggable
> > > and some other people seem to be interested as well. So I've dumped
> > > a few thoughts about some pre-requirements to achieve that for those
> > > interested.
> > > 
> > > As you can see, there is a bit of hard work in the way. I'm iterating
> > > that in https://pad.kernel.org/p/isolation, feel free to edit:
> > > 
> > > 
> > > == RCU nocb ==
> > > 
> > > Currently controllable with "rcu_nocbs=" boot parameter and/or through nohz_full=/isolcpus=nohz
> > > We need to make it toggeable at runtime. Currently handling that:
> > > v1: https://lwn.net/Articles/820544/
> > > v2: coming soon
> > 
> > Nice.
> > 
> > > == TIF_NOHZ ==
> > > 
> > > Need to get rid of that in order not to trigger syscall slowpath on CPUs that don't want nohz_full.
> > > Also we don't want to iterate all threads and clear the flag when the last nohz_full CPU exits nohz_full
> > > mode. Prefer static keys to call context tracking on archs. x86 does that well.
> > > 
> > > == Proper entry code ==
> > > 
> > > We must make sure that a given arch never calls exception_enter() / exception_exit().
> > > This saves the previous state of context tracking and switch to kernel mode (from context tracking POV)
> > > temporarily. Since this state is saved on the stack, this prevents us from turning off context tracking
> > > entirely on a CPU: The tracking must be done on all CPUs and that takes some cycles.
> > > 
> > > This means that, considering early entry code (before the call to context tracking upon kernel entry,
> > > and after the call to context tracking upon kernel exit), we must take care of few things:
> > > 
> > > 1) Make sure early entry code can't trigger exceptions. Or if it does, the given exception can't schedule
> > > or use RCU (unless it calls rcu_nmi_enter()). Otherwise the exception must call exception_enter()/exception_exit()
> > > which we don't want.
> > > 
> > > 2) No call to schedule_user().
> > > 
> > > 3) Make sure early entry code is not interruptible or preempt_schedule_irq() would rely on
> > > exception_entry()/exception_exit()
> > > 
> > > 4) Make sure early entry code can't be traced (no call to preempt_schedule_notrace()), or if it does it
> > > can't schedule
> > > 
> > > I believe x86 does most of that well. In the end we should remove exception_enter()/exit implementations
> > > in x86 and replace it with a check that makes sure context_tracking state is not in USER. An arch meeting
> > > all the above conditions would earn a CONFIG_ARCH_HAS_SANE_CONTEXT_TRACKING. Being able to toggle nohz_full
> > > at runtime would depend on that.
> > > 
> > > 
> > > == Cputime accounting ==
> > > 
> > > Both write and read side must switch to tick based accounting and drop the use of seqlock in task_cputime(),
> > > task_gtime(), kcpustat_field(), kcpustat_cpu_fetch(). Special ordering/state machine is required to make that without races.
> > > 
> > > == Nohz ==
> > > 
> > > Switch from nohz_full to nohz_idle. Mind a few details:
> > >     
> > >     1) Turn off 1Hz offlined tick handled in housekeeping
> > >     2) Handle tick dependencies, take care of racing CPUs setting/clearing tick dependency. It's much trickier when
> > >     we switch from nohz_idle to nohz_full
> > >     
> > > == Unbound affinity ==
> > > 
> > > Restore kernel threads, workqueue, timers, etc... wide affinity. But take care of cpumasks that have been set through other
> > > interfaces: sysfs, procfs, etc...
> > 
> > We were looking at a userspace interface: what would be a proper
> > (unified, similar to isolcpus= interface) and its implementation:
> > 
> > The simplest idea for interface seemed to be exposing the integer list of
> > CPUs and isolation flags to userspace (probably via sysfs).
> > 
> > The scheme would allow flags to be separately enabled/disabled, 
> > with not all flags being necessary toggable (could for example
> > disallow nohz_full= toggling until it is implemented, but allow for
> > other isolation features to be toggable).
> > 
> > This would require per flag housekeeping_masks (instead of a single).
> > 
> > Back to the userspace interface, you mentioned earlier that cpusets
> > was a possibility for it. However:
> > 
> > "Cpusets provide a Linux kernel mechanism to constrain which CPUs and
> > Memory Nodes are used by a process or set of processes.
> > 
> > The Linux kernel already has a pair of mechanisms to specify on which
> > CPUs a task may be scheduled (sched_setaffinity) and on which Memory
> > Nodes it may obtain memory (mbind, set_mempolicy).
> > 
> > Cpusets extends these two mechanisms as follows:"
> > 
> > The isolation flags do not necessarily have anything to do with
> > tasks, but with CPUs: a given feature is disabled or enabled on a
> > given CPU. 
> > No?
> 
> One cpumask per feature, implemented separately in sysfs, also 
> seems OK (modulo documentation about the RCU update and users
> of the previous versions).
> 
> This is what is being done for rcu_nocbs= already...
> 

exclusive cpusets is used now to control scheduler load balancing on
a group of cpus.  It seems to me that this is the same idea and is part
of the isolation concept.  Having a toggle for each subsystem/feature in
cpusets could provide the needed userspace api. 

Under the covers it might be implemented as twiddling various cpumasks.

We need to be shifting to managing load balancing with cpusets anyway.



Cheers,
Phil

-- 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ