[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <56F1F4A6.2060400@lab126.com>
Date: Tue, 22 Mar 2016 18:43:02 -0700
From: Joel Fernandes <joelaf@...126.com>
To: <linux-mm@...ck.org>, <linux-kernel@...r.kernel.org>,
<tj@...nel.org>, <linux-rt-users@...r.kernel.org>,
Nick Piggin <npiggin@...e.de>, Andi Kleen <ak@...ux.intel.com>
Subject: [RFC] high preempt off latency in vfree path
Hi,
I'm seeing on my system with some real time audio requirements, I'm seeing the preemptirqsoff
tracer complaining that preempt was off for 17ms in the vfree path. Since we have requirements
of 8ms scheduling this seems awfully bad.
The tracer output showed __free_vmap_area was about 7300 times. Can we do better here? I have
proposed 2 potential fixes here, any thoughts on which one's better?
Here's the path that blocks preempt (full latency ftrace output uploaded to
http://raw.codepile.net/pile/OWNpvKkB.js)
=> preempt_count_sub
=> _raw_spin_unlock
=> __purge_vmap_area_lazy
=> free_vmap_area_noflush
=> remove_vm_area
=> __vunmap
=> vfree
=> n_tty_close
=> tty_ldisc_close.isra.1
=> tty_ldisc_kill
=> tty_ldisc_release
=> tty_release
Here are the approaches:
(1)
One is we reduce the number of lazy_max_pages (right now its around 32MB per core worth of pages).
diff --git a/mm/vmalloc.c b/mm/vmalloc.c
index aa3891e..2720f4f 100644
--- a/mm/vmalloc.c
+++ b/mm/vmalloc.c
@@ -564,7 +564,7 @@ static unsigned long lazy_max_pages(void)
log = fls(num_online_cpus());
- return log * (32UL * 1024 * 1024 / PAGE_SIZE);
+ return log * (8UL * 1024 * 1024 / PAGE_SIZE);
}
(2) Second alternative approach I am thinking is to change purge_lock into a mutex and then
move the vmap_area spinlock around the free_vmap_area call. Thus giving the scheduler a chance
to put something else on the CPU in between free_vmap_area calls. That would look like:
diff --git a/mm/vmalloc.c b/mm/vmalloc.c
index aa3891e..9565d72 100644
--- a/mm/vmalloc.c
+++ b/mm/vmalloc.c
@@ -594,7 +594,7 @@ void set_iounmap_nonlazy(void)
static void __purge_vmap_area_lazy(unsigned long *start, unsigned long *end,
int sync, int force_flush)
{
- static DEFINE_SPINLOCK(purge_lock);
+ static DEFINE_MUTEX(purge_lock);
LIST_HEAD(valist);
struct vmap_area *va;
struct vmap_area *n_va;
@@ -606,10 +606,10 @@ static void __purge_vmap_area_lazy(unsigned long *start, unsigned long *end,
* the case that isn't actually used at the moment anyway.
*/
if (!sync && !force_flush) {
- if (!spin_trylock(&purge_lock))
+ if (!mutex_trylock(&purge_lock))
return;
} else
- spin_lock(&purge_lock);
+ mutex_lock(&purge_lock);
if (sync)
purge_fragmented_blocks_allcpus();
@@ -636,12 +636,13 @@ static void __purge_vmap_area_lazy(unsigned long *start, unsigned long *end,
flush_tlb_kernel_range(*start, *end);
if (nr) {
- spin_lock(&vmap_area_lock);
- list_for_each_entry_safe(va, n_va, &valist, purge_list)
+ list_for_each_entry_safe(va, n_va, &valist, purge_list) {
+ spin_lock(&vmap_area_lock);
__free_vmap_area(va);
+ spin_unlock(&vmap_area_lock);
+ }
- spin_unlock(&vmap_area_lock);
}
- spin_unlock(&purge_lock);
+ mutex_unlock(&purge_lock);
}
/*
Thanks!
Joel
Powered by blists - more mailing lists