[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4B2C4727.1010302@gmail.com>
Date: Sat, 19 Dec 2009 12:23:19 +0900
From: Minchan Kim <minchan.kim@...il.com>
To: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
CC: Andi Kleen <andi@...stfloor.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"linux-mm@...ck.org" <linux-mm@...ck.org>, cl@...ux-foundation.org,
"akpm@...ux-foundation.org" <akpm@...ux-foundation.org>,
"mingo@...e.hu" <mingo@...e.hu>
Subject: Re: [RFC 2/4] add mm event counter
Hi, Kame.
KAMEZAWA Hiroyuki wrote:
> From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
>
> Add version counter to mm_struct. It's updated when
> write_lock is held and released. And this patch also adds
> task->mm_version. By this, mm_semaphore can provides some
> operation like seqlock.
>
> Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
> ---
> include/linux/mm_types.h | 1 +
> include/linux/sched.h | 2 +-
> mm/mm_accessor.c | 29 ++++++++++++++++++++++++++---
> 3 files changed, 28 insertions(+), 4 deletions(-)
>
> Index: mmotm-mm-accessor/include/linux/mm_types.h
> ===================================================================
> --- mmotm-mm-accessor.orig/include/linux/mm_types.h
> +++ mmotm-mm-accessor/include/linux/mm_types.h
> @@ -216,6 +216,7 @@ struct mm_struct {
> atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */
> int map_count; /* number of VMAs */
> struct rw_semaphore mmap_sem;
How about ?
struct map_sem {
struct rw_semaphore;
#ifdef CONFIG_PER_THREAD_VMACACHE
int version;
#endif
};
struct mm_struct {
...
struct map_sem mmap_sem
...
};
void mm_read_lock(struct mem_sem *mmap_sem)
{
down_read(mmap_sem);
#ifdef CONFIG_PER_THREAD_VMACACHE
if (current->mm_version != mmap_sem->version)
current->mm_version = mmap_sem->version;
#endif
}
We know many place(ex, page fault patch) are matched (current->mm == mm).
Let's compare it just case of get_user_pages and few cases before calling mm_xxx_lock.
Why I suggest above is that i don't want regression about single thread app.
(Of course. you use the cache hit well but we can't ignore compare and
one cache line invalidation by store).
If we can configure CONFIG_PER_THREAD_VMACACHE, we can prevent it.
As a matter of fact, I want to control speculative page fault and older in runtime.
For example, if any process starts to have many threads, VM turns on speculative about
the process. But It's not easy to determine the threshold
(ex, the number of thread >> NR_CPU .
I think mm_accessor patch is valuable as above.
1. It doesn't hide lock instance.
2. It can be configurable.
3. code is more simple.
If you can do it well without mm_accessor, I don't mind it.
I think this is a good start point.
> + int version;
> spinlock_t page_table_lock; /* Protects page tables and some counters */
>
> struct list_head mmlist; /* List of maybe swapped mm's. These are globally strung
> Index: mmotm-mm-accessor/include/linux/sched.h
> ===================================================================
> --- mmotm-mm-accessor.orig/include/linux/sched.h
> +++ mmotm-mm-accessor/include/linux/sched.h
> @@ -1276,7 +1276,7 @@ struct task_struct {
> struct plist_node pushable_tasks;
>
> struct mm_struct *mm, *active_mm;
> -
> + int mm_version;
> /* task state */
> int exit_state;
> int exit_code, exit_signal;
> Index: mmotm-mm-accessor/mm/mm_accessor.c
> ===================================================================
> --- mmotm-mm-accessor.orig/mm/mm_accessor.c
> +++ mmotm-mm-accessor/mm/mm_accessor.c
> @@ -1,15 +1,20 @@
> -#include <linux/mm_types.h>
> +#include <linux/sched.h>
> #include <linux/module.h>
>
> void mm_read_lock(struct mm_struct *mm)
> {
> down_read(&mm->mmap_sem);
> + if (current->mm == mm && current->mm_version != mm->version)
> + current->mm_version = mm->version;
> }
> EXPORT_SYMBOL(mm_read_lock);
>
..
<snip>
..
> void mm_write_unlock(struct mm_struct *mm)
> {
> + mm->version++;
> up_write(&mm->mmap_sem);
Don't we need to increase version in unlock case?
> }
> EXPORT_SYMBOL(mm_write_unlock);
>
> int mm_write_trylock(struct mm_struct *mm)
> {
> - return down_write_trylock(&mm->mmap_sem);
> + int ret = down_write_trylock(&mm->mmap_sem);
> +
> + if (ret)
> + mm->version++;
> + return ret;
> }
> EXPORT_SYMBOL(mm_write_trylock);
>
> @@ -45,6 +56,7 @@ EXPORT_SYMBOL(mm_is_locked);
>
> void mm_write_to_read_lock(struct mm_struct *mm)
> {
> + mm->version++;
> downgrade_write(&mm->mmap_sem);
> }
> EXPORT_SYMBOL(mm_write_to_read_lock);
> @@ -78,3 +90,14 @@ void mm_read_might_lock(struct mm_struct
> might_lock_read(&mm->mmap_sem);
> }
> EXPORT_SYMBOL(mm_read_might_lock);
> +
> +/*
> + * Called when mm is accessed without read-lock or for chekcing
^^^^^^^^
checking :)
> + * per-thread cached value is stale or not.
> + */
> +int mm_version_check(struct mm_struct *mm)
Nitpick:
How about "int vma_cache_stale(struct mm_struct *mm)"?
> +{
> + if ((current->mm == mm) && (current->mm_version != mm->version))
> + return 0;
> + return 1;
> +}
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists