[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <0f8fe661-c450-ccd8-761f-dbfff449c533@huawei.com>
Date: Thu, 26 May 2022 17:45:03 +0800
From: Xu Kuohai <xukuohai@...wei.com>
To: Mark Rutland <mark.rutland@....com>
CC: <bpf@...r.kernel.org>, <linux-arm-kernel@...ts.infradead.org>,
<linux-kernel@...r.kernel.org>, <netdev@...r.kernel.org>,
<linux-kselftest@...r.kernel.org>,
Catalin Marinas <catalin.marinas@....com>,
Will Deacon <will@...nel.org>,
Steven Rostedt <rostedt@...dmis.org>,
Ingo Molnar <mingo@...hat.com>,
Daniel Borkmann <daniel@...earbox.net>,
Alexei Starovoitov <ast@...nel.org>,
Zi Shen Lim <zlim.lnx@...il.com>,
Andrii Nakryiko <andrii@...nel.org>,
Martin KaFai Lau <kafai@...com>,
Song Liu <songliubraving@...com>, Yonghong Song <yhs@...com>,
John Fastabend <john.fastabend@...il.com>,
KP Singh <kpsingh@...nel.org>,
"David S . Miller" <davem@...emloft.net>,
Hideaki YOSHIFUJI <yoshfuji@...ux-ipv6.org>,
David Ahern <dsahern@...nel.org>,
Thomas Gleixner <tglx@...utronix.de>,
Borislav Petkov <bp@...en8.de>,
Dave Hansen <dave.hansen@...ux.intel.com>, <x86@...nel.org>,
<hpa@...or.com>, Shuah Khan <shuah@...nel.org>,
Jakub Kicinski <kuba@...nel.org>,
Jesper Dangaard Brouer <hawk@...nel.org>,
Pasha Tatashin <pasha.tatashin@...een.com>,
Ard Biesheuvel <ardb@...nel.org>,
Daniel Kiss <daniel.kiss@....com>,
Steven Price <steven.price@....com>,
Sudeep Holla <sudeep.holla@....com>,
Marc Zyngier <maz@...nel.org>,
Peter Collingbourne <pcc@...gle.com>,
Mark Brown <broonie@...nel.org>,
Delyan Kratunov <delyank@...com>,
Kumar Kartikeya Dwivedi <memxor@...il.com>,
Wang ShaoBo <bobo.shaobowang@...wei.com>,
<cj.chengjian@...wei.com>, <huawei.libin@...wei.com>,
<xiexiuqi@...wei.com>, <liwei391@...wei.com>
Subject: Re: [PATCH bpf-next v5 1/6] arm64: ftrace: Add ftrace direct call
support
On 5/25/2022 9:38 PM, Mark Rutland wrote:
> On Wed, May 18, 2022 at 09:16:33AM -0400, Xu Kuohai wrote:
>> Add ftrace direct support for arm64.
>>
>> 1. When there is custom trampoline only, replace the fentry nop to a
>> jump instruction that jumps directly to the custom trampoline.
>>
>> 2. When ftrace trampoline and custom trampoline coexist, jump from
>> fentry to ftrace trampoline first, then jump to custom trampoline
>> when ftrace trampoline exits. The current unused register
>> pt_regs->orig_x0 is used as an intermediary for jumping from ftrace
>> trampoline to custom trampoline.
>
> For those of us not all that familiar with BPF, can you explain *why* you want
> this? The above explains what the patch implements, but not why that's useful.
>
> e.g. is this just to avoid the overhead of the ops list processing in the
> regular ftrace code, or is the custom trampoline there to allow you to do
> something special?
IIUC, ftrace direct call was designed to *remove* the unnecessary
overhead of saving regs completely [1][2].
[1]
https://lore.kernel.org/all/20191022175052.frjzlnjjfwwfov64@ast-mbp.dhcp.thefacebook.com/
[2] https://lore.kernel.org/all/20191108212834.594904349@goodmis.org/
This patch itself is just a variant of [3].
[3] https://lore.kernel.org/all/20191108213450.891579507@goodmis.org/
>
> There is another patch series on the list from some of your colleagues which
> uses dynamic trampolines to try to avoid that ops list overhead, and it's not
> clear to me whether these are trying to solve the largely same problem or
> something different. That other thread is at:
>
> https://lore.kernel.org/linux-arm-kernel/20220316100132.244849-1-bobo.shaobowang@huawei.com/
>
> ... and I've added the relevant parties to CC here, since there doesn't seem to
> be any overlap in the CC lists of the two threads.
We're not working to solve the same problem. The trampoline introduced
in this series helps us to monitor kernel function or another bpf prog
with bpf, and also helps us to use bpf prog like a normal kernel
function pointer.
>
> In that other thread I've suggested a general approach we could follow at:
>
> https://lore.kernel.org/linux-arm-kernel/YmGF%2FOpIhAF8YeVq@lakrids/
>
Is it possible for a kernel function to take a long jump to common
trampoline when we get a huge kernel image?
> As noted in that thread, I have a few concerns which equally apply here:
>
> * Due to the limited range of BL instructions, it's not always possible to
> patch an ftrace call-site to branch to an arbitrary trampoline. The way this
> works for ftrace today relies upon knowingthe set of trampolines at
> compile-time, and allocating module PLTs for those, and that approach cannot
> work reliably for dynanically allocated trampolines.
Currently patch 5 returns -ENOTSUPP when long jump is detected, so no
bpf trampoline is constructed for out of range patch-site:
if (is_long_jump(orig_call, image))
return -ENOTSUPP;
>
> I'd strongly prefer to avoid custom tramplines unless they're strictly
> necessary for functional reasons, so that we can have this work reliably and
> consistently.
bpf trampoline is needed by bpf itself, not to replace ftrace trampolines.
>> * If this is mostly about avoiding the ops list processing overhead, I
beleive
> we can implement some custom ops support more generally in ftrace which would
> still use a common trampoline but could directly call into those custom ops.
> I would strongly prefer this over custom trampolines.
>
> * I'm looking to minimize the set of regs ftrace saves, and never save a full
> pt_regs, since today we (incompletely) fill that with bogus values and cannot
> acquire some state reliably (e.g. PSTATE). I'd like to avoid usage of pt_regs
> unless necessary, and I don't want to add additional reliance upon that
> structure.
Even if such a common trampoline is used, bpf trampoline is still
necessary since we need to construct custom instructions to implement
bpf functions, for example, to implement kernel function pointer with a
bpf prog.
>
>> Signed-off-by: Xu Kuohai <xukuohai@...wei.com>
>> Acked-by: Song Liu <songliubraving@...com>
>> ---
>> arch/arm64/Kconfig | 2 ++
>> arch/arm64/include/asm/ftrace.h | 12 ++++++++++++
>> arch/arm64/kernel/asm-offsets.c | 1 +
>> arch/arm64/kernel/entry-ftrace.S | 18 +++++++++++++++---
>> 4 files changed, 30 insertions(+), 3 deletions(-)
>>
>> diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
>> index 57c4c995965f..81cc330daafc 100644
>> --- a/arch/arm64/Kconfig
>> +++ b/arch/arm64/Kconfig
>> @@ -177,6 +177,8 @@ config ARM64
>> select HAVE_DYNAMIC_FTRACE
>> select HAVE_DYNAMIC_FTRACE_WITH_REGS \
>> if $(cc-option,-fpatchable-function-entry=2)
>> + select HAVE_DYNAMIC_FTRACE_WITH_DIRECT_CALLS \
>> + if DYNAMIC_FTRACE_WITH_REGS
>> select FTRACE_MCOUNT_USE_PATCHABLE_FUNCTION_ENTRY \
>> if DYNAMIC_FTRACE_WITH_REGS
>> select HAVE_EFFICIENT_UNALIGNED_ACCESS
>> diff --git a/arch/arm64/include/asm/ftrace.h b/arch/arm64/include/asm/ftrace.h
>> index 1494cfa8639b..14a35a5df0a1 100644
>> --- a/arch/arm64/include/asm/ftrace.h
>> +++ b/arch/arm64/include/asm/ftrace.h
>> @@ -78,6 +78,18 @@ static inline unsigned long ftrace_call_adjust(unsigned long addr)
>> return addr;
>> }
>>
>> +#ifdef CONFIG_HAVE_DYNAMIC_FTRACE_WITH_DIRECT_CALLS
>> +static inline void arch_ftrace_set_direct_caller(struct pt_regs *regs,
>> + unsigned long addr)
>> +{
>> + /*
>> + * Place custom trampoline address in regs->orig_x0 to let ftrace
>> + * trampoline jump to it.
>> + */
>> + regs->orig_x0 = addr;
>> +}
>> +#endif /* CONFIG_HAVE_DYNAMIC_FTRACE_WITH_DIRECT_CALLS */
>
> Please, let's not abuse pt_regs::orig_x0 for this. That's at best unnecessarily
> confusing, and if we really need a field to place a value like this it implies
> we should add an ftrace-specific structure to hold the ftrace-specific context
> information.
>
Sorry for this confusion, this was modified in the x86 way:
https://lore.kernel.org/all/20191108213450.891579507@goodmis.org/
> Thanks,
> Mark.
>
>> +
>> #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
>> struct dyn_ftrace;g w
>> int ftrace_init_nop(struct module *mod, struct dyn_ftrace *rec);
>> diff --git a/arch/arm64/kernel/asm-offsets.c b/arch/arm64/kernel/asm-offsets.c
>> index 1197e7679882..b1ed0bf01c59 100644
>> --- a/arch/arm64/kernel/asm-offsets.c
>> +++ b/arch/arm64/kernel/asm-offsets.c
>> @@ -80,6 +80,7 @@ int main(void)
>> DEFINE(S_SDEI_TTBR1, offsetof(struct pt_regs, sdei_ttbr1));
>> DEFINE(S_PMR_SAVE, offsetof(struct pt_regs, pmr_save));
>> DEFINE(S_STACKFRAME, offsetof(struct pt_regs, stackframe));
>> + DEFINE(S_ORIG_X0, offsetof(struct pt_regs, orig_x0));
>> DEFINE(PT_REGS_SIZE, sizeof(struct pt_regs));
>> BLANK();
>> #ifdef CONFIG_COMPAT
>> diff --git a/arch/arm64/kernel/entry-ftrace.S b/arch/arm64/kernel/entry-ftrace.S
>> index e535480a4069..dfe62c55e3a2 100644
>> --- a/arch/arm64/kernel/entry-ftrace.S
>> +++ b/arch/arm64/kernel/entry-ftrace.S
>> @@ -60,6 +60,9 @@
>> str x29, [sp, #S_FP]
>> .endif
>>
>> + /* Set orig_x0 to zero */
>> + str xzr, [sp, #S_ORIG_X0]
>> +
>> /* Save the callsite's SP and LR */
>> add x10, sp, #(PT_REGS_SIZE + 16)
>> stp x9, x10, [sp, #S_LR]
>> @@ -119,12 +122,21 @@ ftrace_common_return:
>> /* Restore the callsite's FP, LR, PC */
>> ldr x29, [sp, #S_FP]
>> ldr x30, [sp, #S_LR]
>> - ldr x9, [sp, #S_PC]
>> -
>> + ldr x10, [sp, #S_PC]
>> +
>> + ldr x11, [sp, #S_ORIG_X0]
>> + cbz x11, 1f
>> + /* Set x9 to parent ip before jump to custom trampoline */
>> + mov x9, x30
>> + /* Set lr to self ip */
>> + ldr x30, [sp, #S_PC]
>> + /* Set x10 (used for return address) to custom trampoline */
>> + mov x10, x11
>> +1:
>> /* Restore the callsite's SP */
>> add sp, sp, #PT_REGS_SIZE + 16
>>
>> - ret x9
>> + ret x10
>> SYM_CODE_END(ftrace_common)
>>
>> #ifdef CONFIG_FUNCTION_GRAPH_TRACER
>> --
>> 2.30.2
>>
> .
Powered by blists - more mailing lists