[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <4B975A00.3090404@dcl.info.waseda.ac.jp>
Date: Wed, 10 Mar 2010 17:36:16 +0900
From: Hitoshi Mitake <mitake@....info.waseda.ac.jp>
To: Peter Zijlstra <peterz@...radead.org>
CC: Ingo Molnar <mingo@...e.hu>,
Frederic Weisbecker <fweisbec@...il.com>,
LKML <linux-kernel@...r.kernel.org>,
Jason Baron <jbaron@...hat.com>
Subject: Re: Question about policy of calling lockdep functions in trylocks
On 03/02/10 22:43, Hitoshi Mitake wrote:
> On 03/02/10 18:13, Peter Zijlstra wrote:
>> On Tue, 2010-03-02 at 17:44 +0900, Hitoshi Mitake wrote:
>>> Hi,
>>>
>>> I have a question about policy of callings lockdep functions in
>>> trylocks.
>>>
>>> Normal locks like __raw_spin_lock are defined like this:
>>>
>>> static inline void __raw_spin_lock(raw_spinlock_t *lock)
>>> {
>>> preempt_disable();
>>> spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
>>> LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
>>> }
>>>
>>> And LOCK_CONTENDED is defined:
>>> #define LOCK_CONTENDED(_lock, try, lock) \
>>> do { \
>>> if (!try(_lock)) { \
>>> lock_contended(&(_lock)->dep_map, _RET_IP_); \
>>> lock(_lock); \
>>> } \
>>> lock_acquired(&(_lock)->dep_map, _RET_IP_); \
>>> } while (0)
>>>
>>> So, acquiring and releasing lock with no contention calls lockdep
>>> functions like this:
>>>
>>> lock_acquire -> lock_acquired -> lock_release
>>>
>>> And acquiring and releasing lock with contention calls lockdep
functions
>>> like this:
>>>
>>> lock_acquire -> lock_contended -> lock_acquired -> lock_release
>>>
>>> But I found that locks with try like __raw_spin_trylock is defined like
>>> this:
>>>
>>> static inline int __raw_spin_trylock(raw_spinlock_t *lock)
>>> {
>>> preempt_disable();
>>> if (do_raw_spin_trylock(lock)) {
>>> spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
>>> return 1;
>>> }
>>> preempt_enable();
>>> return 0;
>>> }
>>>
>>> So, trying acquiring and releasing lock with no contention calls
lockdep
>>> functions like this:
>>>
>>> lock_acquire -> lock_release
>>>
>>> And failed trying acquiring calls no lockdep function.
>>>
>>> I felt that policy of calling lockdep functions is strange.
>>> Trylocks should be like this:
>>>
>>> static inline int __raw_spin_trylock(raw_spinlock_t *lock)
>>> {
>>> preempt_disable();
>>> spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
>>> if (do_raw_spin_trylock(lock)) {
>>> spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
>>> lock_acquired(&lock->dep_map, _RET_IP_);
>>> return 1;
>>> }
>>> lock_contended(&lock->dep_map, _RET_IP_);
>>> preempt_enable();
>>> return 0;
>>> }
>>
>> Did you mean to call acquire twice?
>
> Sorry, this is my mistake.
> I've forgot to remove spin_acquire() after do_raw_spin_trylock().
>
>>
>>>
>>> This is my question.
>>> Are there some reasons current calling lockdep functions of trylocks?
>>> If not, can I change these trylocks like I described above?
>>>
>>> The reason why I'm asking about it is perf lock.
>>> For state machine of perf lock, these event sequenses are very
>>> confusable.
>>> Because sequence of trylock is subset of normal lock. This is
ambiguity.
>>
>> Well, trylocks cannot contend, so the lock_contended() call doesn't make
>> sense, I don't think it will confuse lockstat, since acquire will simply
>> reset the state again, but it will waste cycles, nor is there a reason
>> to call acquired without first having call contended. So no.
>>
>> What exactly is the problem, the lack of callbacks for a failed trylock?
>> Why would you want one?
>>
>> Because other than that I see no problem, you get an acquire(.try=1) and
>> a release() to match and lockstat measure and accounts the hold-time.
>
> Ah, I've forgot about try and read parameters of lock_acquire().
> These parameters are enough things for me.
> perf lock can separate event sequences of normal locks and trylocks with
> these.
>
Sorry, I have another suggestion.
Some subsystems of kernel use functions of lockdep such as lock_acquire()
only for lockdep_map objects without actual locks.
e.g. rcu_read_lock() of RCU: Frederic did lots of work on recursion of
this function.
It seems that they are using functions of lockdep purely for validation
of dependency.
These are not actual locks. But produces trace events of locks,
and lockdep calculates their holded time (in /proc/lock_stats).
I think this situation is not so good.
Because
* overhead of validation is equal to sum of ones of validation and
holdtime,
* overhead of holdtime is equal to sum of ones of validation and holdtime,
* and overhead of tracing lock events is equal to sum of ones of
validation, holdtime and tracing :(
So I'd like to suggest that separating these features.
The way is like this,
1) current struct of locks are holding lockdep_map,
changing this lockdep_map to more general structure.
(Temporary, I'd like to call it "struct lock_monitor")
2) Rename current functions of lockdep.
e.g. lock_acquire() -> lockdep_acquire()
3) Prepare more general functions for hooking lock events,
void lock_acquire(struct lockdep_map *lock, ...)
->
void lock_acquire(struct lock_monitor *lock, ...)
New lock_acquire() should call hook functions
of validation, holdtime and tracing.
If so, overhead of each features can be separated.
There are at least 3 requested features related to analyzing lock,
* validation ... what lockdep does
* in kernel histogram ... more rich /proc/lock_stat, Jason Baron may
be interested in this
* tracing and analyzing ... what perf lock wants to do
So I think this is worth to do.
I'm doing it (not finished yet), and noticed that
patch series for this will be very big one.
Because many lockdep users are in kernel here and there.
So I'd like to ask your opinion before implementing and posting, Peter.
How do you think?
Thanks,
Hitoshi
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists