[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87pna98ajt.fsf@nanos.tec.linutronix.de>
Date: Tue, 09 Jun 2020 00:20:06 +0200
From: Thomas Gleixner <tglx@...utronix.de>
To: Qian Cai <cai@....pw>, Peter Zijlstra <peterz@...radead.org>
Cc: LKML <linux-kernel@...r.kernel.org>,
Andy Lutomirski <luto@...nel.org>,
Andrew Cooper <andrew.cooper3@...rix.com>,
X86 ML <x86@...nel.org>,
"Paul E. McKenney" <paulmck@...nel.org>,
Alexandre Chartre <alexandre.chartre@...cle.com>,
Frederic Weisbecker <frederic@...nel.org>,
Paolo Bonzini <pbonzini@...hat.com>,
Sean Christopherson <sean.j.christopherson@...el.com>,
Masami Hiramatsu <mhiramat@...nel.org>,
Petr Mladek <pmladek@...e.com>,
Steven Rostedt <rostedt@...dmis.org>,
Joel Fernandes <joel@...lfernandes.org>,
Boris Ostrovsky <boris.ostrovsky@...cle.com>,
Juergen Gross <jgross@...e.com>,
Brian Gerst <brgerst@...il.com>,
Mathieu Desnoyers <mathieu.desnoyers@...icios.com>,
Josh Poimboeuf <jpoimboe@...hat.com>,
Will Deacon <will@...nel.org>,
Tom Lendacky <thomas.lendacky@....com>,
Wei Liu <wei.liu@...nel.org>,
Michael Kelley <mikelley@...rosoft.com>,
Jason Chen CJ <jason.cj.chen@...el.com>,
Zhao Yakui <yakui.zhao@...el.com>,
Alexander Potapenko <glider@...gle.com>
Subject: Re: [patch V9 10/39] x86/entry: Provide helpers for execute on irqstack
Qian,
can you please ensure that people who got cc'ed because the problem
affects their subsystem are included on your replies even if you are
replying to a different subthread?
I explicitely did:
Cc:+ Alexander
at the very beginning of my reply:
https://lore.kernel.org/r/87v9k3jdc6.fsf@nanos.tec.linutronix.de
to make you aware of that.
Yes, email sucks, but it sucks even more when people are careless.
Qian Cai <cai@....pw> writes:
> On Fri, Jun 05, 2020 at 07:36:22PM +0200, Peter Zijlstra wrote:
>
> Even after I trimmed the .config [1] to the barely minimal, a subset of LTP
> test still unable to finish on those AMD servers with page_owner=on.
What a surprise...
> [1]:
> https://raw.githubusercontent.com/cailca/linux-mm/master/x86.config
>
> It looks like because this new IRQ entry introduced by this patch,
>
> </IRQ>
> asm_call_on_stack at arch/x86/entry/entry_64.S:710
> handle_edge_irq at kernel/irq/chip.c:832
>
> which will running out of the stack depot limit due to nested loops
> below.
>
> which has this loop,
>
> do {
> ...
> handle_irq_event(desc);
> ...
> } while ((desc->istate & IRQS_PENDING) &&
> !irqd_irq_disabled(&desc->irq_data));
This loop has absolutely nothing to do with stack entry usage.
foo()
do {
bar();
} while (condition);
}
If you take a stack trace inside bar() it will be the same stack trace
for every single loop iteration. And that stack trace will not be any
different from:
foo()
{
bar():
}
assumed that the call chain leading to foo() is the same in both cases.
And you can add even more loops in subsequent call chains within
bar(). They do not matter at all.
> Here has a nested loop,
>
> for_each_action_of_desc(desc, action) {
> ...
> res = action->handler(irq, action->dev_id);
> ...
> }
And this one is completely irrelevant because the interrupt which we are
looking at is a PCI interrupt which CANNOT be shared. IOW, the number of
loop iterations and the number of handlers invoked is exactly ONE.
I seriously have no idea what you are trying to demonstrate by finding
loops in a SINGLE callchain.
> Here has one more nested loop,
>
> while (table->orig_nents) {
> ...
> free_fn(sgl, alloc_size);
> ...
> }
>
> free_fn() will call kmem_cache_free(). Since we have page_owner=on, it
> will call save_stack() to save the each free stack trace.
That stack trace for each invocation of free_fn() in this loop is
exactly the same stack trace. The same stack trace is not eating up any
memory because the hash matches, i.e. the stack trace in the depot is
already known.
Here is the simplified difference between the old code and the new code:
Old New
handle_edge_irq handle_edge_irq
do_IRQ asm_call_on_stack
common_interrupt common_interrupt
asm_common_interrupt
IOW, for every _UNIQUE_ interrupt related call chain, there is exactly
ONE stack entry more than before.
For a loop which generates the exact same stack trace for every
iteration this extra entry is not a problem.
But what matters is that interrupts can hit any random code path. So the
amount of possible non-unique call chains is pretty much unlimited. And
with a high number of non-unique call chains the extra entry starts to
matter.
It's trival math, isn't it?
TS = Total Size of depot
as = average size of all stored unique stack traces
ms = maximum number of unqiue stack traces which fit in TS
ms_old = TS / as
Lets further assume that the vast majority of stack traces are taken
from interrupt context. That means with the new code this results in:
ms_new = TS / (as + 1)
==> ms_new = ms_old * as / (as + 1)
Depending on the value of 'as' the +1 can shave off a significant
percentage of capacity. IOW, the capacity is simply too small now for
the test scenario you are running. Truly a surprising outcome, right?
To get facts instead of useless loop theories, can you please apply the
patch below, enable DEBUGFS and provide the output of
/sys/kernel/debug/stackdepot/info
for a kernel before that change and after? Please read out that file at
periodically roughly the same amounts of time after starting your test
scenario.
Note, that I doubled the size of the stack depot so that we get real
numbers and not the cutoff by the size limit. IOW, the warning should
not trigger anymore. If it triggers nevertheless then the numbers will
still tell us an interesting story.
Thanks,
tglx
---
lib/stackdepot.c | 43 ++++++++++++++++++++++++++++++++++++++++++-
1 file changed, 42 insertions(+), 1 deletion(-)
--- a/lib/stackdepot.c
+++ b/lib/stackdepot.c
@@ -31,6 +31,7 @@
#include <linux/stackdepot.h>
#include <linux/string.h>
#include <linux/types.h>
+#include <linux/debugfs.h>
#define DEPOT_STACK_BITS (sizeof(depot_stack_handle_t) * 8)
@@ -42,7 +43,7 @@
STACK_ALLOC_ALIGN)
#define STACK_ALLOC_INDEX_BITS (DEPOT_STACK_BITS - \
STACK_ALLOC_NULL_PROTECTION_BITS - STACK_ALLOC_OFFSET_BITS)
-#define STACK_ALLOC_SLABS_CAP 8192
+#define STACK_ALLOC_SLABS_CAP 16384
#define STACK_ALLOC_MAX_SLABS \
(((1LL << (STACK_ALLOC_INDEX_BITS)) < STACK_ALLOC_SLABS_CAP) ? \
(1LL << (STACK_ALLOC_INDEX_BITS)) : STACK_ALLOC_SLABS_CAP)
@@ -70,6 +71,7 @@ static void *stack_slabs[STACK_ALLOC_MAX
static int depot_index;
static int next_slab_inited;
static size_t depot_offset;
+static unsigned long unique_stacks;
static DEFINE_SPINLOCK(depot_lock);
static bool init_stack_slab(void **prealloc)
@@ -138,6 +140,7 @@ static struct stack_record *depot_alloc_
stack->handle.valid = 1;
memcpy(stack->entries, entries, size * sizeof(unsigned long));
depot_offset += required_size;
+ unique_stacks++;
return stack;
}
@@ -340,3 +343,41 @@ unsigned int filter_irq_stacks(unsigned
return nr_entries;
}
EXPORT_SYMBOL_GPL(filter_irq_stacks);
+
+static int debug_show(struct seq_file *m, void *p)
+{
+ unsigned long unst;
+ int didx, doff;
+
+ spin_lock_irq(&depot_lock);
+ unst = unique_stacks;
+ didx = depot_index;
+ doff = depot_offset;
+ spin_unlock_irq(&depot_lock);
+
+ seq_printf(m, "Unique stacks: %lu\n", unst);
+ seq_printf(m, "Depot index: %d\n", didx);
+ seq_printf(m, "Depot offset: %d\n", doff);
+ return 0;
+}
+
+static int debug_open(struct inode *inode, struct file *file)
+{
+ return single_open(file, debug_show, inode->i_private);
+}
+
+static const struct file_operations dfs_ops = {
+ .open = debug_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = single_release,
+};
+
+static int __init debugfs_init(void)
+{
+ struct dentry *root_dir = debugfs_create_dir("stackdepot", NULL);
+
+ debugfs_create_file("info", 0444, root_dir, NULL, &dfs_ops);
+ return 0;
+}
+__initcall(debugfs_init);
Powered by blists - more mailing lists