[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090903210139.GA25581@elte.hu>
Date: Thu, 3 Sep 2009 23:01:39 +0200
From: Ingo Molnar <mingo@...e.hu>
To: Jason Baron <jbaron@...hat.com>
Cc: linux-kernel@...r.kernel.org, mathieu.desnoyers@...ymtl.ca,
roland@...hat.com, rth@...hat.com
Subject: Re: [PATCH 0/4] RFC: jump label - (tracepoint optimizations)
* Jason Baron <jbaron@...hat.com> wrote:
> hi,
>
> Problem:
>
> Currenly, tracepoints are implemented using a conditional. The
> conditional check requires checking a global variable for each
> tracepoint. Although, the overhead of this check is small, it
> increases under memory pressure. As we increase the number of
> tracepoints in the kernel this may become more of an issue. In
> addition, tracepoints are often dormant (disabled), and provide no
> direct kernel functionality. Thus, it is highly desirable to
> reduce their impact as much as possible. Mathieu Desnoyers has
> already suggested a number of requirements for a solution to this
> issue.
>
> Solution:
>
> In discussing this problem with Roland McGrath and Richard
> Henderson, we came up with a new 'asm goto' statement that allows
> branching to a label. Thus, this patch set introdues a
> 'STATIC_JUMP_IF()' macro as follows:
>
> #ifdef HAVE_STATIC_JUMP
>
> #define STATIC_JUMP_IF(tag, label, cond) \
> asm goto ("1:" /* 5-byte insn */ \
> P6_NOP5 \
> ".pushsection __jump_table, \"a\" \n\t" \
> _ASM_PTR "1b, %l[" #label "], %c0 \n\t" \
> ".popsection \n\t" \
> : : "i" (__sjstrtab_##tag) : : label)
>
> #else
>
> #define STATIC_JUMP_IF(tag, label, cond) \
> if (unlikely(cond)) \
> goto label;
>
> #endif /* !HAVE_STATIC_JUMP */
>
>
> which can be used as:
>
> STATIC_JUMP_IF(trace, trace_label, jump_enabled);
> printk("not doing tracing\n");
> if (0) {
> trace_label:
> printk("doing tracing: %d\n", file);
> }
>
> ---------------------------------------
>
> Thus, if 'HAVE_STATIC_JUMP' is defined (which will depend
> ultimately on the existence of 'asm goto' in the compiler
> version), we simply have a no-op followed by a jump around the
> dormant (disabled) tracing code. The 'STATIC_JUMP_IF()' macro,
> produces a 'jump_table' which has the following format:
>
> [instruction address] [jump target] [tracepoint name]
>
> Thus, to enable a tracepoint, we simply patch the 'instruction
> address' with a jump to the 'jump target'. The current
> implementation is using ftrace infrastructure to accomplish the
> patching, which uses 'stop_machine'. In subsequent versions, we
> will update the mechanism to use more efficient code patching
> techniques.
>
> I've tested the performance of this using 'get_cycles()' calls
> around the tracepoint call sites. For an Intel Core 2 Quad cpu (in
> cycles, averages):
>
> idle after tbench run
> ---- ----------------
> old code 32 88
> new code 2 4
>
>
> The performance improvement can be reproduced very reliably (using
> patch 4 in this series) on both Intel and AMD hardware.
>
> In terms of code analysis the current code for the disabled case
> is a 'cmpl' followed by a 'je' around the tracepoint code. so:
>
> cmpl - 83 3d 0e 77 87 00 00 - 7 bytes
> je - 74 3e - 2 bytes
>
> total of 9 instruction bytes.
>
> The new code is a 'nopl' followed by a 'jmp'. Thus:
>
> nopl - 0f 1f 44 00 00 - 5 bytes
> jmp - eb 3e - 2 bytes
>
> total of 7 instruction bytes.
>
> So, the new code also accounts for 2 less bytes in the instruction
> cache per tracepoint.
>
> here's a link to the gcc thread introducing this feature:
>
> http://gcc.gnu.org/ml/gcc-patches/2009-07/msg01556.html
This looks really interesting and desired. Once GCC adds this (or an
equivalent) feature, i'd love to have your optimization in the
kernel.
> Todo:
>
> - convert the patching to a more optimal implementation (not using stop machine)
> - expand infrastructure for modules
> - other use cases?
[...]
Other usecases might be kernel features that are turned on/off via
some slowpath. For example SLAB statistics could be patched in/out
using this method. Or scheduler statistics.
Basically everything that is optional and touches some very hot
codepath would be eligible - not just tracepoints.
Ingo
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists