[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <alpine.LFD.1.10.0809222106090.3265@nehalem.linux-foundation.org>
Date: Mon, 22 Sep 2008 21:19:01 -0700 (PDT)
From: Linus Torvalds <torvalds@...ux-foundation.org>
To: Steven Rostedt <rostedt@...dmis.org>
cc: Roland Dreier <rdreier@...co.com>,
Masami Hiramatsu <mhiramat@...hat.com>,
Martin Bligh <mbligh@...gle.com>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
Thomas Gleixner <tglx@...utronix.de>,
Mathieu Desnoyers <compudj@...stal.dyndns.org>,
darren@...art.com, "Frank Ch. Eigler" <fche@...hat.com>,
systemtap-ml <systemtap@...rces.redhat.com>
Subject: Re: Unified tracing buffer
On Mon, 22 Sep 2008, Steven Rostedt wrote:
>
> But, with that, with a global atomic counter, and the following trace:
>
> cpu 0: trace_point_a
> cpu 1: trace_point_c
> cpu 0: trace_point_b
> cpu 1: trace_point_d
>
> Could the event a really come after event d, even though we already hit
> event b?
Each tracepoint will basically give a partial ordering (if you make it so,
of course - and on x86 it's hard to avoid it).
And with many trace-points, you can narrow down ordering if you're lucky.
But say that you have code like
CPU#1 CPU#2
trace_a trace_c
.. ..
trace_b trace_d
and since each CPU itself is obviously strictly ordered, you a priori know
that a < b, and c < d. But your trace buffer can look many different ways:
- a -> b -> c -> d
c -> d -> a -> b
Now you do know that what happened between c and d must all have
happened entirely after/before the things that happened between
a and b, and there is no overlap.
This is only assuming the x86 full memory barrier from a "lock xadd" of
course, but those are the semantics you'd get on x86. On others, the
ordering might not be that strong.
- a -> c -> b -> d
a -> c -> d -> b
With these trace point orderings, you really don't know anything at all
about the order of any access that happened in between. CPU#1 might
have gone first. Or not. Or partially. You simply do not know.
> But I guess you are stating the fact that what the computer does
> internally, no one really knows. Without the help of real memory barriers,
> ording of memory accesses is mostly determined by tarot cards.
Well, x86 defines a memory order. But what I'm trying to explain is that
memory order still doesn't actually specify what happens to the code that
actually does tracing! The trace is only going to show the order of the
tracepoints, not the _other_ memory accesses. So you'll have *some*
information, but it's very partial.
And the thing is, all those other memory accesses are the ones that do all
the real work. You'll know they happened _somewhere_ between two
tracepoints, but not much more than that.
This is why timestamps aren't really any worse than sequence numbers in
all practical matters. They'll get you close enough that you can consider
them equivalent to a cache-coherent counter, just one that you don't have
to take a cache miss for, and that increments on its own!
Quite a lot of CPU's have nice, dependable, TSC's that run at constant
frequency.
And quite a lot of traces care a _lot_ about real time. When you do IO
tracing, the problem is almost never about lock ordering or anything like
that. You want to see how long a request took. You don't care AT ALL how
many tracepoints were in between the beginning and end, you care about how
many microseconds there were!
Linus
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists