[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CACT4Y+Zaov2rynD0T_SVbZ4_s+fqMnUg961PcaJ=mg40D34BPQ@mail.gmail.com>
Date: Fri, 25 Jul 2025 12:21:03 +0200
From: Dmitry Vyukov <dvyukov@...gle.com>
To: Alexander Potapenko <glider@...gle.com>
Cc: quic_jiangenj@...cinc.com, linux-kernel@...r.kernel.org,
kasan-dev@...glegroups.com, x86@...nel.org,
Aleksandr Nogikh <nogikh@...gle.com>, Andrey Konovalov <andreyknvl@...il.com>, Borislav Petkov <bp@...en8.de>,
Dave Hansen <dave.hansen@...ux.intel.com>, Ingo Molnar <mingo@...hat.com>,
Josh Poimboeuf <jpoimboe@...nel.org>, Marco Elver <elver@...gle.com>,
Peter Zijlstra <peterz@...radead.org>, Thomas Gleixner <tglx@...utronix.de>
Subject: Re: [PATCH v2 06/11] kcov: x86: introduce CONFIG_KCOV_UNIQUE
On Fri, 25 Jul 2025 at 12:07, Alexander Potapenko <glider@...gle.com> wrote:
>
> On Wed, Jul 9, 2025 at 5:01 PM Dmitry Vyukov <dvyukov@...gle.com> wrote:
> >
> > On Thu, 26 Jun 2025 at 15:42, Alexander Potapenko <glider@...gle.com> wrote:
> > >
> > > The new config switches coverage instrumentation to using
> > > __sanitizer_cov_trace_pc_guard(u32 *guard)
> > > instead of
> > > __sanitizer_cov_trace_pc(void)
> > >
> > > This relies on Clang's -fsanitize-coverage=trace-pc-guard flag [1].
> > >
> > > Each callback receives a unique 32-bit guard variable residing in the
> > > __sancov_guards section. Those guards can be used by kcov to deduplicate
> > > the coverage on the fly.
> > >
> > > As a first step, we make the new instrumentation mode 1:1 compatible
> > > with the old one.
> > >
> > > [1] https://clang.llvm.org/docs/SanitizerCoverage.html#tracing-pcs-with-guards
> > >
> > > Cc: x86@...nel.org
> > > Signed-off-by: Alexander Potapenko <glider@...gle.com>
> > >
> > > ---
> > > Change-Id: Iacb1e71fd061a82c2acadf2347bba4863b9aec39
> > >
> > > v2:
> > > - Address comments by Dmitry Vyukov
> > > - rename CONFIG_KCOV_ENABLE_GUARDS to CONFIG_KCOV_UNIQUE
> > > - update commit description and config description
> > > - Address comments by Marco Elver
> > > - rename sanitizer_cov_write_subsequent() to kcov_append_to_buffer()
> > > - make config depend on X86_64 (via ARCH_HAS_KCOV_UNIQUE)
> > > - swap #ifdef branches
> > > - tweak config description
> > > - remove redundant check for CONFIG_CC_HAS_SANCOV_TRACE_PC_GUARD
> > > ---
> > > arch/x86/Kconfig | 1 +
> > > arch/x86/kernel/vmlinux.lds.S | 1 +
> > > include/asm-generic/vmlinux.lds.h | 14 ++++++-
> > > include/linux/kcov.h | 2 +
> > > kernel/kcov.c | 61 +++++++++++++++++++++----------
> > > lib/Kconfig.debug | 24 ++++++++++++
> > > scripts/Makefile.kcov | 4 ++
> > > scripts/module.lds.S | 23 ++++++++++++
> > > tools/objtool/check.c | 1 +
> > > 9 files changed, 110 insertions(+), 21 deletions(-)
> > >
> > > diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig
> > > index e21cca404943e..d104c5a193bdf 100644
> > > --- a/arch/x86/Kconfig
> > > +++ b/arch/x86/Kconfig
> > > @@ -93,6 +93,7 @@ config X86
> > > select ARCH_HAS_FORTIFY_SOURCE
> > > select ARCH_HAS_GCOV_PROFILE_ALL
> > > select ARCH_HAS_KCOV if X86_64
> > > + select ARCH_HAS_KCOV_UNIQUE if X86_64
> > > select ARCH_HAS_KERNEL_FPU_SUPPORT
> > > select ARCH_HAS_MEM_ENCRYPT
> > > select ARCH_HAS_MEMBARRIER_SYNC_CORE
> > > diff --git a/arch/x86/kernel/vmlinux.lds.S b/arch/x86/kernel/vmlinux.lds.S
> > > index cda5f8362e9da..8076e8953fddc 100644
> > > --- a/arch/x86/kernel/vmlinux.lds.S
> > > +++ b/arch/x86/kernel/vmlinux.lds.S
> > > @@ -372,6 +372,7 @@ SECTIONS
> > > . = ALIGN(PAGE_SIZE);
> > > __bss_stop = .;
> > > }
> > > + SANCOV_GUARDS_BSS
> > >
> > > /*
> > > * The memory occupied from _text to here, __end_of_kernel_reserve, is
> > > diff --git a/include/asm-generic/vmlinux.lds.h b/include/asm-generic/vmlinux.lds.h
> > > index 58a635a6d5bdf..875c4deb66208 100644
> > > --- a/include/asm-generic/vmlinux.lds.h
> > > +++ b/include/asm-generic/vmlinux.lds.h
> > > @@ -102,7 +102,8 @@
> > > * sections to be brought in with rodata.
> > > */
> > > #if defined(CONFIG_LD_DEAD_CODE_DATA_ELIMINATION) || defined(CONFIG_LTO_CLANG) || \
> > > -defined(CONFIG_AUTOFDO_CLANG) || defined(CONFIG_PROPELLER_CLANG)
> > > + defined(CONFIG_AUTOFDO_CLANG) || defined(CONFIG_PROPELLER_CLANG) || \
> > > + defined(CONFIG_KCOV_UNIQUE)
> > > #define TEXT_MAIN .text .text.[0-9a-zA-Z_]*
> > > #else
> > > #define TEXT_MAIN .text
> > > @@ -121,6 +122,17 @@ defined(CONFIG_AUTOFDO_CLANG) || defined(CONFIG_PROPELLER_CLANG)
> > > #define SBSS_MAIN .sbss
> > > #endif
> > >
> > > +#if defined(CONFIG_KCOV_UNIQUE)
> > > +#define SANCOV_GUARDS_BSS \
> > > + __sancov_guards(NOLOAD) : { \
> > > + __start___sancov_guards = .; \
> > > + *(__sancov_guards); \
> > > + __stop___sancov_guards = .; \
> > > + }
> > > +#else
> > > +#define SANCOV_GUARDS_BSS
> > > +#endif
> > > +
> > > /*
> > > * GCC 4.5 and later have a 32 bytes section alignment for structures.
> > > * Except GCC 4.9, that feels the need to align on 64 bytes.
> > > diff --git a/include/linux/kcov.h b/include/linux/kcov.h
> > > index 0e425c3524b86..dd8bbee6fe274 100644
> > > --- a/include/linux/kcov.h
> > > +++ b/include/linux/kcov.h
> > > @@ -107,6 +107,8 @@ typedef unsigned long long kcov_u64;
> > > #endif
> > >
> > > void __sanitizer_cov_trace_pc(void);
> > > +void __sanitizer_cov_trace_pc_guard(u32 *guard);
> > > +void __sanitizer_cov_trace_pc_guard_init(uint32_t *start, uint32_t *stop);
> > > void __sanitizer_cov_trace_cmp1(u8 arg1, u8 arg2);
> > > void __sanitizer_cov_trace_cmp2(u16 arg1, u16 arg2);
> > > void __sanitizer_cov_trace_cmp4(u32 arg1, u32 arg2);
> > > diff --git a/kernel/kcov.c b/kernel/kcov.c
> > > index ff7f118644f49..8e98ca8d52743 100644
> > > --- a/kernel/kcov.c
> > > +++ b/kernel/kcov.c
> > > @@ -195,27 +195,15 @@ static notrace unsigned long canonicalize_ip(unsigned long ip)
> > > return ip;
> > > }
> > >
> > > -/*
> > > - * Entry point from instrumented code.
> > > - * This is called once per basic-block/edge.
> > > - */
> > > -void notrace __sanitizer_cov_trace_pc(void)
> > > +static notrace void kcov_append_to_buffer(unsigned long *area, int size,
> > > + unsigned long ip)
> > > {
> > > - struct task_struct *t;
> > > - unsigned long *area;
> > > - unsigned long ip = canonicalize_ip(_RET_IP_);
> > > - unsigned long pos;
> > > -
> > > - t = current;
> > > - if (!check_kcov_mode(KCOV_MODE_TRACE_PC, t))
> > > - return;
> > > -
> > > - area = t->kcov_state.area;
> > > /* The first 64-bit word is the number of subsequent PCs. */
> > > - pos = READ_ONCE(area[0]) + 1;
> > > - if (likely(pos < t->kcov_state.size)) {
> > > - /* Previously we write pc before updating pos. However, some
> > > - * early interrupt code could bypass check_kcov_mode() check
> > > + unsigned long pos = READ_ONCE(area[0]) + 1;
> > > +
> > > + if (likely(pos < size)) {
> > > + /*
> > > + * Some early interrupt code could bypass check_kcov_mode() check
> > > * and invoke __sanitizer_cov_trace_pc(). If such interrupt is
> > > * raised between writing pc and updating pos, the pc could be
> > > * overitten by the recursive __sanitizer_cov_trace_pc().
> > > @@ -226,7 +214,40 @@ void notrace __sanitizer_cov_trace_pc(void)
> > > area[pos] = ip;
> > > }
> > > }
> > > +
> > > +/*
> > > + * Entry point from instrumented code.
> > > + * This is called once per basic-block/edge.
> > > + */
> > > +#ifdef CONFIG_KCOV_UNIQUE
> > > +void notrace __sanitizer_cov_trace_pc_guard(u32 *guard)
> > > +{
> > > + if (!check_kcov_mode(KCOV_MODE_TRACE_PC, current))
> > > + return;
> > > +
> > > + kcov_append_to_buffer(current->kcov_state.area,
> > > + current->kcov_state.size,
> > > + canonicalize_ip(_RET_IP_));
> > > +}
> > > +EXPORT_SYMBOL(__sanitizer_cov_trace_pc_guard);
> > > +
> > > +void notrace __sanitizer_cov_trace_pc_guard_init(uint32_t *start,
> > > + uint32_t *stop)
> > > +{
> > > +}
> > > +EXPORT_SYMBOL(__sanitizer_cov_trace_pc_guard_init);
> > > +#else /* !CONFIG_KCOV_UNIQUE */
> > > +void notrace __sanitizer_cov_trace_pc(void)
> > > +{
> > > + if (!check_kcov_mode(KCOV_MODE_TRACE_PC, current))
> > > + return;
> > > +
> > > + kcov_append_to_buffer(current->kcov_state.area,
> > > + current->kcov_state.size,
> > > + canonicalize_ip(_RET_IP_));
> > > +}
> > > EXPORT_SYMBOL(__sanitizer_cov_trace_pc);
> > > +#endif
> > >
> > > #ifdef CONFIG_KCOV_ENABLE_COMPARISONS
> > > static void notrace write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip)
> > > @@ -254,7 +275,7 @@ static void notrace write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip)
> > > start_index = 1 + count * KCOV_WORDS_PER_CMP;
> > > end_pos = (start_index + KCOV_WORDS_PER_CMP) * sizeof(u64);
> > > if (likely(end_pos <= max_pos)) {
> > > - /* See comment in __sanitizer_cov_trace_pc(). */
> > > + /* See comment in kcov_append_to_buffer(). */
> > > WRITE_ONCE(area[0], count + 1);
> > > barrier();
> > > area[start_index] = type;
> > > diff --git a/lib/Kconfig.debug b/lib/Kconfig.debug
> > > index f9051ab610d54..24dcb721dbb0b 100644
> > > --- a/lib/Kconfig.debug
> > > +++ b/lib/Kconfig.debug
> > > @@ -2156,6 +2156,8 @@ config ARCH_HAS_KCOV
> > > config CC_HAS_SANCOV_TRACE_PC
> > > def_bool $(cc-option,-fsanitize-coverage=trace-pc)
> > >
> > > +config CC_HAS_SANCOV_TRACE_PC_GUARD
> > > + def_bool $(cc-option,-fsanitize-coverage=trace-pc-guard)
> > >
> > > config KCOV
> > > bool "Code coverage for fuzzing"
> > > @@ -2172,6 +2174,28 @@ config KCOV
> > >
> > > For more details, see Documentation/dev-tools/kcov.rst.
> > >
> > > +config ARCH_HAS_KCOV_UNIQUE
> > > + bool
> > > + help
> > > + An architecture should select this when it can successfully
> > > + build and run with CONFIG_KCOV_UNIQUE.
> > > +
> > > +config KCOV_UNIQUE
> > > + depends on KCOV
> > > + depends on CC_HAS_SANCOV_TRACE_PC_GUARD && ARCH_HAS_KCOV_UNIQUE
> > > + bool "Use coverage guards for KCOV"
> > > + help
> > > + Use coverage guards instrumentation for KCOV, passing
> > > + -fsanitize-coverage=trace-pc-guard to the compiler.
> >
> > I think this should talk about the new mode, the new ioctl's, and
> > visible differences for end users first.
>
> Something like this, maybe?
>
> This option enables KCOV's unique program counter (PC)
> collection mode,
> which deduplicates PCs on the fly when the KCOV_UNIQUE_ENABLE ioctl is
> used.
>
> This significantly reduces the memory footprint for coverage data
> collection compared to trace mode, as it prevents the kernel from
> storing the same PC multiple times.
> Enabling this mode incurs a slight increase in kernel binary size.
>
>
Looks good to me.
> > > + Every coverage callback is associated with a global variable that
> > > + allows to efficiently deduplicate coverage at collection time.
> > > + This drastically reduces the buffer size required for coverage
> > > + collection.
> > > +
> > > + This config comes at a cost of increased binary size (4 bytes of .bss
> > > + plus 1-2 instructions to pass an extra parameter, per basic block).
> > > +
> > > config KCOV_ENABLE_COMPARISONS
> > > bool "Enable comparison operands collection by KCOV"
> > > depends on KCOV
> > > diff --git a/scripts/Makefile.kcov b/scripts/Makefile.kcov
> > > index 67e8cfe3474b7..0b17533ef35f6 100644
> > > --- a/scripts/Makefile.kcov
> > > +++ b/scripts/Makefile.kcov
> > > @@ -1,5 +1,9 @@
> > > # SPDX-License-Identifier: GPL-2.0-only
> > > +ifeq ($(CONFIG_KCOV_UNIQUE),y)
> > > +kcov-flags-y += -fsanitize-coverage=trace-pc-guard
> > > +else
> > > kcov-flags-$(CONFIG_CC_HAS_SANCOV_TRACE_PC) += -fsanitize-coverage=trace-pc
> > > +endif
> > > kcov-flags-$(CONFIG_KCOV_ENABLE_COMPARISONS) += -fsanitize-coverage=trace-cmp
> > > kcov-flags-$(CONFIG_GCC_PLUGIN_SANCOV) += -fplugin=$(objtree)/scripts/gcc-plugins/sancov_plugin.so
> > >
> > > diff --git a/scripts/module.lds.S b/scripts/module.lds.S
> > > index 450f1088d5fd3..314b56680ea1a 100644
> > > --- a/scripts/module.lds.S
> > > +++ b/scripts/module.lds.S
> > > @@ -64,6 +64,29 @@ SECTIONS {
> > > MOD_CODETAG_SECTIONS()
> > > }
> > > #endif
> > > +
> > > +#ifdef CONFIG_KCOV_UNIQUE
> > > + __sancov_guards(NOLOAD) : {
> > > + __start___sancov_guards = .;
> > > + *(__sancov_guards);
> > > + __stop___sancov_guards = .;
> > > + }
> > > +
> > > + .text : {
> > > + *(.text .text.[0-9a-zA-Z_]*)
> > > + *(.text..L*)
> > > + }
> >
> > Why do we need these here? .text does not look specific to CONFIG_KCOV_UNIQUE.
> > Is it because of constructors/destructors emitted by the compiler, and
> > .init.text/.exit.text don't work w/o .text?
> > A comment here would be useful.
>
> This is because the compiler creates duplicate .init.text/.exit.text,
> making the module loader unhappy.
> I'll add a comment.
Yes, a comment would help.
Powered by blists - more mailing lists