lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20061125035526.GF14076@parisc-linux.org>
Date:	Fri, 24 Nov 2006 20:55:27 -0700
From:	Matthew Wilcox <matthew@....cx>
To:	linux-kernel@...r.kernel.org
Cc:	Ingo Molnar <mingo@...e.hu>
Subject: Re: [PATCH 1/2] Introduce mutex_lock_timeout


No comment on this?

On Thu, Nov 09, 2006 at 11:27:21AM -0700, Matthew Wilcox wrote:
> 
> We have a couple of places in the tree that really could do with a
> down_timeout() function.  I noticed it in qla2xxx and ACPI, but maybe
> there are others that don't have the courtesy of putting "I wish we had
> a down_timeout" comment beside their implementation.
> 
> I don't really want to poke at 25 different implementations of
> down_timeout, but doing a mutex_timeout() seems easy enough, and at
> least qla2xxx can use it.
> 
> I mused on whether to put the timeout argument into the struct
> mutex, but on the whole, I thought it better to somewhat penalise
> mutex_lock_interruptible(), mutex_lock_nested and the mutex lock slowpaths
> in time than it was to penalise all mutex users in space.
> 
> Here's the patch to add mutex_lock_timeout().  I booted the resulting
> kernel on a parisc machine, but I wouldn't claim to have done any real
> testing.  It's fairly mechanical though.
> 
> diff --git a/include/asm-arm/mutex.h b/include/asm-arm/mutex.h
> index cb29d84..6fa4e5f 100644
> --- a/include/asm-arm/mutex.h
> +++ b/include/asm-arm/mutex.h
> @@ -44,7 +44,8 @@ __mutex_fastpath_lock(atomic_t *count, f
>  }
>  
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies,
> +			fastcall int (*fail_fn)(atomic_t *, long jiffies))
>  {
>  	int __ex_flag, __res;
>  
> @@ -60,7 +61,7 @@ __mutex_fastpath_lock_retval(atomic_t *c
>  
>  	__res |= __ex_flag;
>  	if (unlikely(__res != 0))
> -		__res = fail_fn(count);
> +		__res = fail_fn(count, jiffies);
>  	return __res;
>  }
>  
> diff --git a/include/asm-generic/mutex-dec.h b/include/asm-generic/mutex-dec.h
> index 0134151..92a93df 100644
> --- a/include/asm-generic/mutex-dec.h
> +++ b/include/asm-generic/mutex-dec.h
> @@ -37,10 +37,11 @@ __mutex_fastpath_lock(atomic_t *count, f
>   * or anything the slow path function returns.
>   */
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies,
> +				fastcall int (*fail_fn)(atomic_t *, long))
>  {
>  	if (unlikely(atomic_dec_return(count) < 0))
> -		return fail_fn(count);
> +		return fail_fn(count, jiffies);
>  	else {
>  		smp_mb();
>  		return 0;
> diff --git a/include/asm-generic/mutex-null.h b/include/asm-generic/mutex-null.h
> index e1bbbc7..5a65210 100644
> --- a/include/asm-generic/mutex-null.h
> +++ b/include/asm-generic/mutex-null.h
> @@ -11,7 +11,8 @@ #ifndef _ASM_GENERIC_MUTEX_NULL_H
>  #define _ASM_GENERIC_MUTEX_NULL_H
>  
>  #define __mutex_fastpath_lock(count, fail_fn)		fail_fn(count)
> -#define __mutex_fastpath_lock_retval(count, fail_fn)	fail_fn(count)
> +#define __mutex_fastpath_lock_retval(count, jiffies, fail_fn)	\
> +						fail_fn(count, jiffies)
>  #define __mutex_fastpath_unlock(count, fail_fn)		fail_fn(count)
>  #define __mutex_fastpath_trylock(count, fail_fn)	fail_fn(count)
>  #define __mutex_slowpath_needs_to_unlock()		1
> diff --git a/include/asm-generic/mutex-xchg.h b/include/asm-generic/mutex-xchg.h
> index 6a7e8c1..b5eee4e 100644
> --- a/include/asm-generic/mutex-xchg.h
> +++ b/include/asm-generic/mutex-xchg.h
> @@ -42,10 +42,11 @@ __mutex_fastpath_lock(atomic_t *count, f
>   * or anything the slow path function returns
>   */
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count, fastcall int (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies,
> +			fastcall int (*fail_fn)(atomic_t *, long))
>  {
>  	if (unlikely(atomic_xchg(count, 0) != 1))
> -		return fail_fn(count);
> +		return fail_fn(count, jiffies);
>  	else {
>  		smp_mb();
>  		return 0;
> diff --git a/include/asm-i386/mutex.h b/include/asm-i386/mutex.h
> index 7a17d9e..7e13ded 100644
> --- a/include/asm-i386/mutex.h
> +++ b/include/asm-i386/mutex.h
> @@ -51,11 +51,11 @@ do {									\
>   * or anything the slow path function returns
>   */
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count,
> -			     int fastcall (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies,
> +			     int fastcall (*fail_fn)(atomic_t *, long))
>  {
>  	if (unlikely(atomic_dec_return(count) < 0))
> -		return fail_fn(count);
> +		return fail_fn(count, jiffies);
>  	else
>  		return 0;
>  }
> diff --git a/include/asm-ia64/mutex.h b/include/asm-ia64/mutex.h
> index bed73a6..151f067 100644
> --- a/include/asm-ia64/mutex.h
> +++ b/include/asm-ia64/mutex.h
> @@ -36,10 +36,11 @@ __mutex_fastpath_lock(atomic_t *count, v
>   * or anything the slow path function returns.
>   */
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies,
> +				int (*fail_fn)(atomic_t *, long))
>  {
>  	if (unlikely(ia64_fetchadd4_acq(count, -1) != 1))
> -		return fail_fn(count);
> +		return fail_fn(count, jiffies);
>  	return 0;
>  }
>  
> diff --git a/include/asm-x86_64/mutex.h b/include/asm-x86_64/mutex.h
> index 16396b1..18668fa 100644
> --- a/include/asm-x86_64/mutex.h
> +++ b/include/asm-x86_64/mutex.h
> @@ -46,11 +46,11 @@ do {									\
>   * or anything the slow path function returns
>   */
>  static inline int
> -__mutex_fastpath_lock_retval(atomic_t *count,
> -			     int fastcall (*fail_fn)(atomic_t *))
> +__mutex_fastpath_lock_retval(atomic_t *count, long jiffies
> +			     int fastcall (*fail_fn)(atomic_t *, long))
>  {
>  	if (unlikely(atomic_dec_return(count) < 0))
> -		return fail_fn(count);
> +		return fail_fn(count, jiffies);
>  	else
>  		return 0;
>  }
> diff --git a/include/linux/mutex.h b/include/linux/mutex.h
> index 27c48da..b70caf2 100644
> --- a/include/linux/mutex.h
> +++ b/include/linux/mutex.h
> @@ -122,6 +122,7 @@ static inline int fastcall mutex_is_lock
>   */
>  extern void fastcall mutex_lock(struct mutex *lock);
>  extern int fastcall mutex_lock_interruptible(struct mutex *lock);
> +extern int fastcall mutex_lock_timeout(struct mutex *lock, long jiffies);
>  
>  #ifdef CONFIG_DEBUG_LOCK_ALLOC
>  extern void mutex_lock_nested(struct mutex *lock, unsigned int subclass);
> diff --git a/kernel/mutex.c b/kernel/mutex.c
> index 8c71cf7..d2ec4d3 100644
> --- a/kernel/mutex.c
> +++ b/kernel/mutex.c
> @@ -122,7 +122,8 @@ EXPORT_SYMBOL(mutex_unlock);
>   * Lock a mutex (possibly interruptible), slowpath:
>   */
>  static inline int __sched
> -__mutex_lock_common(struct mutex *lock, long state, unsigned int subclass)
> +__mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
> +			long jiffies)
>  {
>  	struct task_struct *task = current;
>  	struct mutex_waiter waiter;
> @@ -158,19 +159,19 @@ __mutex_lock_common(struct mutex *lock, 
>  		 * TASK_UNINTERRUPTIBLE case.)
>  		 */
>  		if (unlikely(state == TASK_INTERRUPTIBLE &&
> -						signal_pending(task))) {
> +				(signal_pending(task) || jiffies == 0))) {
>  			mutex_remove_waiter(lock, &waiter, task->thread_info);
>  			mutex_release(&lock->dep_map, 1, _RET_IP_);
>  			spin_unlock_mutex(&lock->wait_lock, flags);
>  
>  			debug_mutex_free_waiter(&waiter);
> -			return -EINTR;
> +			return jiffies ? -EINTR : -ETIMEDOUT;
>  		}
>  		__set_task_state(task, state);
>  
>  		/* didnt get the lock, go to sleep: */
>  		spin_unlock_mutex(&lock->wait_lock, flags);
> -		schedule();
> +		jiffies = schedule_timeout(jiffies);
>  		spin_lock_mutex(&lock->wait_lock, flags);
>  	}
>  
> @@ -194,7 +195,8 @@ __mutex_lock_slowpath(atomic_t *lock_cou
>  {
>  	struct mutex *lock = container_of(lock_count, struct mutex, count);
>  
> -	__mutex_lock_common(lock, TASK_UNINTERRUPTIBLE, 0);
> +	__mutex_lock_common(lock, TASK_UNINTERRUPTIBLE, 0,
> +						MAX_SCHEDULE_TIMEOUT);
>  }
>  
>  #ifdef CONFIG_DEBUG_LOCK_ALLOC
> @@ -202,7 +204,8 @@ void __sched
>  mutex_lock_nested(struct mutex *lock, unsigned int subclass)
>  {
>  	might_sleep();
> -	__mutex_lock_common(lock, TASK_UNINTERRUPTIBLE, subclass);
> +	__mutex_lock_common(lock, TASK_UNINTERRUPTIBLE, subclass,
> +						MAX_SCHEDULE_TIMEOUT);
>  }
>  
>  EXPORT_SYMBOL_GPL(mutex_lock_nested);
> @@ -256,10 +259,10 @@ __mutex_unlock_slowpath(atomic_t *lock_c
>  
>  /*
>   * Here come the less common (and hence less performance-critical) APIs:
> - * mutex_lock_interruptible() and mutex_trylock().
> + * mutex_lock_interruptible(), mutex_lock_timeout and mutex_trylock().
>   */
>  static int fastcall noinline __sched
> -__mutex_lock_interruptible_slowpath(atomic_t *lock_count);
> +__mutex_lock_interruptible_slowpath(atomic_t *lock_count, long jiffies);
>  
>  /***
>   * mutex_lock_interruptible - acquire the mutex, interruptable
> @@ -275,18 +278,39 @@ __mutex_lock_interruptible_slowpath(atom
>  int fastcall __sched mutex_lock_interruptible(struct mutex *lock)
>  {
>  	might_sleep();
> -	return __mutex_fastpath_lock_retval
> -			(&lock->count, __mutex_lock_interruptible_slowpath);
> +	return __mutex_fastpath_lock_retval(&lock->count,
> +		MAX_SCHEDULE_TIMEOUT, __mutex_lock_interruptible_slowpath);
>  }
>  
>  EXPORT_SYMBOL(mutex_lock_interruptible);
>  
> +/**
> + * mutex_lock_timeout - try to acquire the mutex and fail if it takes too long
> + * @lock: the mutex to be acquired
> + * @timeout: the number of jiffies to wait
> + *
> + * Lock the mutex like mutex_lock(), and return 0 if the mutex has
> + * been acquired or sleep until the mutex becomes available. If a
> + * signal arrives while waiting for the lock then this function
> + * returns -EINTR.  If the timeout expires, it returns -ETIMEDOUT.
> + *
> + * This function is similar to (but not equivalent to) down_interruptible().
> + */
> +int fastcall __sched mutex_lock_timeout(struct mutex *lock, long jiffies)
> +{
> +	might_sleep();
> +	return __mutex_fastpath_lock_retval(&lock->count, jiffies,
> +					__mutex_lock_interruptible_slowpath);
> +}
> +
> +EXPORT_SYMBOL(mutex_lock_timeout);
> +
>  static int fastcall noinline __sched
> -__mutex_lock_interruptible_slowpath(atomic_t *lock_count)
> +__mutex_lock_interruptible_slowpath(atomic_t *lock_count, long jiffies)
>  {
>  	struct mutex *lock = container_of(lock_count, struct mutex, count);
>  
> -	return __mutex_lock_common(lock, TASK_INTERRUPTIBLE, 0);
> +	return __mutex_lock_common(lock, TASK_INTERRUPTIBLE, 0, jiffies);
>  }
>  
>  /*
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ