lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4B2C4727.1010302@gmail.com>
Date:	Sat, 19 Dec 2009 12:23:19 +0900
From:	Minchan Kim <minchan.kim@...il.com>
To:	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
CC:	Andi Kleen <andi@...stfloor.org>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	"linux-mm@...ck.org" <linux-mm@...ck.org>, cl@...ux-foundation.org,
	"akpm@...ux-foundation.org" <akpm@...ux-foundation.org>,
	"mingo@...e.hu" <mingo@...e.hu>
Subject: Re: [RFC 2/4] add mm event counter

Hi, Kame. 

KAMEZAWA Hiroyuki wrote:
> From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
> 
> Add version counter to mm_struct. It's updated when
> write_lock is held and released. And this patch also adds
> task->mm_version. By this, mm_semaphore can provides some
> operation like seqlock.
> 
> Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
> ---
>  include/linux/mm_types.h |    1 +
>  include/linux/sched.h    |    2 +-
>  mm/mm_accessor.c         |   29 ++++++++++++++++++++++++++---
>  3 files changed, 28 insertions(+), 4 deletions(-)
> 
> Index: mmotm-mm-accessor/include/linux/mm_types.h
> ===================================================================
> --- mmotm-mm-accessor.orig/include/linux/mm_types.h
> +++ mmotm-mm-accessor/include/linux/mm_types.h
> @@ -216,6 +216,7 @@ struct mm_struct {
>  	atomic_t mm_count;			/* How many references to "struct mm_struct" (users count as 1) */
>  	int map_count;				/* number of VMAs */
>  	struct rw_semaphore mmap_sem;

How about ?

struct map_sem {
	struct rw_semaphore;
#ifdef CONFIG_PER_THREAD_VMACACHE
	int version;
#endif
};

struct mm_struct {
	...
	struct map_sem mmap_sem
	...
};

void mm_read_lock(struct mem_sem *mmap_sem)
{
	down_read(mmap_sem);
#ifdef CONFIG_PER_THREAD_VMACACHE
	if (current->mm_version != mmap_sem->version)
		current->mm_version = mmap_sem->version;
#endif
}

We know many place(ex, page fault patch) are matched (current->mm == mm).
Let's compare it just case of get_user_pages and few cases before calling mm_xxx_lock.

Why I suggest above is that i don't want regression about single thread app.
(Of course. you use the cache hit well but we can't ignore compare and 
one cache line invalidation by store).

If we can configure CONFIG_PER_THREAD_VMACACHE, we can prevent it.

As a matter of fact, I want to control speculative page fault and older in runtime.
For example, if any process starts to have many threads, VM turns on speculative about
the process. But It's not easy to determine the threshold
(ex, the number of thread >> NR_CPU  .

I think mm_accessor patch is valuable as above.

1. It doesn't hide lock instance.  
2. It can be configurable.
3. code is more simple. 

If you can do it well without mm_accessor, I don't mind it. 

I think this is a good start point.

> +	int version;
>  	spinlock_t page_table_lock;		/* Protects page tables and some counters */
>  
>  	struct list_head mmlist;		/* List of maybe swapped mm's.	These are globally strung
> Index: mmotm-mm-accessor/include/linux/sched.h
> ===================================================================
> --- mmotm-mm-accessor.orig/include/linux/sched.h
> +++ mmotm-mm-accessor/include/linux/sched.h
> @@ -1276,7 +1276,7 @@ struct task_struct {
>  	struct plist_node pushable_tasks;
>  
>  	struct mm_struct *mm, *active_mm;
> -
> +	int mm_version;
>  /* task state */
>  	int exit_state;
>  	int exit_code, exit_signal;
> Index: mmotm-mm-accessor/mm/mm_accessor.c
> ===================================================================
> --- mmotm-mm-accessor.orig/mm/mm_accessor.c
> +++ mmotm-mm-accessor/mm/mm_accessor.c
> @@ -1,15 +1,20 @@
> -#include <linux/mm_types.h>
> +#include <linux/sched.h>
>  #include <linux/module.h>
>  
>  void mm_read_lock(struct mm_struct *mm)
>  {
>  	down_read(&mm->mmap_sem);
> +	if (current->mm == mm && current->mm_version != mm->version)
> +		current->mm_version = mm->version;
>  }
>  EXPORT_SYMBOL(mm_read_lock);
>  
..
<snip>
..

>  void mm_write_unlock(struct mm_struct *mm)
>  {
> +	mm->version++;
>  	up_write(&mm->mmap_sem);

Don't we need to increase version in unlock case?

>  }
>  EXPORT_SYMBOL(mm_write_unlock);
>  
>  int mm_write_trylock(struct mm_struct *mm)
>  {
> -	return down_write_trylock(&mm->mmap_sem);
> +	int ret = down_write_trylock(&mm->mmap_sem);
> +
> +	if (ret)
> +		mm->version++;
> +	return ret;
>  }
>  EXPORT_SYMBOL(mm_write_trylock);
>  
> @@ -45,6 +56,7 @@ EXPORT_SYMBOL(mm_is_locked);
>  
>  void mm_write_to_read_lock(struct mm_struct *mm)
>  {
> +	mm->version++;
>  	downgrade_write(&mm->mmap_sem);
>  }
>  EXPORT_SYMBOL(mm_write_to_read_lock);
> @@ -78,3 +90,14 @@ void mm_read_might_lock(struct mm_struct
>  	might_lock_read(&mm->mmap_sem);
>  }
>  EXPORT_SYMBOL(mm_read_might_lock);
> +
> +/*
> + * Called when mm is accessed without read-lock or for chekcing
							  ^^^^^^^^
							  checking :)
> + * per-thread cached value is stale or not.
> + */
> +int mm_version_check(struct mm_struct *mm)
Nitpick:

How about "int vma_cache_stale(struct mm_struct *mm)"?

> +{
> +	if ((current->mm == mm) && (current->mm_version != mm->version))
> +		return 0;
> +	return 1;
> +}
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ