[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CACT4Y+Y_PfAhjV26xYf8wcEv0MYygC14c_92hBN8gqOACK7Oow@mail.gmail.com>
Date: Fri, 16 Apr 2021 10:42:32 +0200
From: Dmitry Vyukov <dvyukov@...gle.com>
To: Andrey Konovalov <andreyknvl@...il.com>
Cc: Alexander Lochmann <info@...xander-lochmann.de>,
Andrey Konovalov <andreyknvl@...gle.com>,
Jonathan Corbet <corbet@....net>,
Randy Dunlap <rdunlap@...radead.org>,
Andrew Klychkov <andrew.a.klychkov@...il.com>,
Miguel Ojeda <ojeda@...nel.org>,
Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Jakub Kicinski <kuba@...nel.org>,
Aleksandr Nogikh <nogikh@...gle.com>,
Wei Yongjun <weiyongjun1@...wei.com>,
Maciej Grochowski <maciej.grochowski@...me>,
kasan-dev <kasan-dev@...glegroups.com>,
"open list:DOCUMENTATION" <linux-doc@...r.kernel.org>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCHv3] Introduced new tracing mode KCOV_MODE_UNIQUE.
On Sat, Mar 27, 2021 at 3:56 PM Andrey Konovalov <andreyknvl@...il.com> wrote:
>
> On Fri, Mar 26, 2021 at 9:52 PM Alexander Lochmann
> <info@...xander-lochmann.de> wrote:
> >
>
> Hi Alexander,
>
> > It simply stores the executed PCs.
> > The execution order is discarded.
> > Each bit in the shared buffer represents every fourth
> > byte of the text segment.
> > Since a call instruction on every supported
> > architecture is at least four bytes, it is safe
> > to just store every fourth byte of the text segment.
>
> What about jumps?
KCOV adds call __sanitizer_cov_trace_pc per coverage point. So besides
the instructions in the original code, we also always have this call.
> [...]
>
> > -#define KCOV_IN_CTXSW (1 << 30)
> > +#define KCOV_IN_CTXSW (1 << 31)
>
> This change needs to be mentioned and explained in the changelog.
>
> > -static notrace bool check_kcov_mode(enum kcov_mode needed_mode, struct task_struct *t)
> > +static __always_inline notrace bool check_kcov_mode(enum kcov_mode needed_mode,
> > + struct task_struct *t,
> > + unsigned int *mode)
> > {
> > - unsigned int mode;
> > -
> > /*
> > * We are interested in code coverage as a function of a syscall inputs,
> > * so we ignore code executed in interrupts, unless we are in a remote
> > @@ -162,7 +163,7 @@ static notrace bool check_kcov_mode(enum kcov_mode needed_mode, struct task_stru
> > */
> > if (!in_task() && !(in_serving_softirq() && t->kcov_softirq))
> > return false;
> > - mode = READ_ONCE(t->kcov_mode);
> > + *mode = READ_ONCE(t->kcov_mode);
> > /*
> > * There is some code that runs in interrupts but for which
> > * in_interrupt() returns false (e.g. preempt_schedule_irq()).
> > @@ -171,7 +172,7 @@ static notrace bool check_kcov_mode(enum kcov_mode needed_mode, struct task_stru
> > * kcov_start().
> > */
> > barrier();
> > - return mode == needed_mode;
> > + return ((int)(*mode & (KCOV_IN_CTXSW | needed_mode))) > 0;
>
> This change needs to be mentioned and explained in the changelog.
>
> [...]
>
> > static notrace unsigned long canonicalize_ip(unsigned long ip)
> > @@ -191,18 +192,27 @@ void notrace __sanitizer_cov_trace_pc(void)
> > struct task_struct *t;
> > unsigned long *area;
> > unsigned long ip = canonicalize_ip(_RET_IP_);
> > - unsigned long pos;
> > + unsigned long pos, idx;
> > + unsigned int mode;
> >
> > t = current;
> > - if (!check_kcov_mode(KCOV_MODE_TRACE_PC, t))
> > + if (!check_kcov_mode(KCOV_MODE_TRACE_PC | KCOV_MODE_UNIQUE_PC, t, &mode))
> > return;
> >
> > area = t->kcov_area;
> > - /* The first 64-bit word is the number of subsequent PCs. */
> > - pos = READ_ONCE(area[0]) + 1;
> > - if (likely(pos < t->kcov_size)) {
> > - area[pos] = ip;
> > - WRITE_ONCE(area[0], pos);
> > + if (likely(mode == KCOV_MODE_TRACE_PC)) {
> > + /* The first 64-bit word is the number of subsequent PCs. */
> > + pos = READ_ONCE(area[0]) + 1;
> > + if (likely(pos < t->kcov_size)) {
> > + area[pos] = ip;
> > + WRITE_ONCE(area[0], pos);
> > + }
> > + } else {
> > + idx = (ip - canonicalize_ip((unsigned long)&_stext)) / 4;
> > + pos = idx % BITS_PER_LONG;
> > + idx /= BITS_PER_LONG;
> > + if (likely(idx < t->kcov_size))
> > + WRITE_ONCE(area[idx], READ_ONCE(area[idx]) | 1L << pos);
>
> This is confusing: for KCOV_MODE_TRACE_PC, pos is used to index area,
> and for else, idx is used to index area. You should swap idx and pos.
>
> [...]
>
> > @@ -213,9 +223,10 @@ static void notrace write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip)
> > struct task_struct *t;
> > u64 *area;
> > u64 count, start_index, end_pos, max_pos;
> > + unsigned int mode;
> >
> > t = current;
> > - if (!check_kcov_mode(KCOV_MODE_TRACE_CMP, t))
> > + if (!check_kcov_mode(KCOV_MODE_TRACE_CMP, t, &mode))
> > return;
>
> mode isn't used here, right? No need for it then.
>
> > @@ -562,12 +576,14 @@ static int kcov_ioctl_locked(struct kcov *kcov, unsigned int cmd,
> > {
> > struct task_struct *t;
> > unsigned long size, unused;
> > - int mode, i;
> > + int mode, i, text_size, ret = 0;
> > struct kcov_remote_arg *remote_arg;
> > struct kcov_remote *remote;
> > unsigned long flags;
> >
> > switch (cmd) {
> > + case KCOV_INIT_UNIQUE:
> > + fallthrough;
> > case KCOV_INIT_TRACE:
> > /*
> > * Enable kcov in trace mode and setup buffer size.
> > @@ -581,11 +597,42 @@ static int kcov_ioctl_locked(struct kcov *kcov, unsigned int cmd,
> > * that must not overflow.
> > */
> > size = arg;
> > - if (size < 2 || size > INT_MAX / sizeof(unsigned long))
> > - return -EINVAL;
> > - kcov->size = size;
> > - kcov->mode = KCOV_MODE_INIT;
> > - return 0;
> > + if (cmd == KCOV_INIT_UNIQUE) {
>
> Let's put this code under KCOV_INIT_UNIQUE in the switch. This
> internal if only saves duplicating two lines of code, which isn't
> worth it.
>
> > + if (size != 0)
> > + return -EINVAL;
> > + text_size = (canonicalize_ip((unsigned long)&_etext)
> > + - canonicalize_ip((unsigned long)&_stext));
> > + /**
> > + * A call instr is at least four bytes on every supported architecture.
> > + * Hence, just every fourth instruction can potentially be a call.
> > + */
> > + text_size = roundup(text_size, 4);
> > + text_size /= 4;
> > + /*
> > + * Round up size of text segment to multiple of BITS_PER_LONG.
> > + * Otherwise, we cannot track
> > + * the last (text_size % BITS_PER_LONG) addresses.
> > + */
> > + text_size = roundup(text_size, BITS_PER_LONG);
> > + /* Get the amount of bytes needed */
> > + text_size = text_size / 8;
> > + /* mmap() requires size to be a multiple of PAGE_SIZE */
> > + text_size = roundup(text_size, PAGE_SIZE);
> > + /* Get the cover size (= amount of bytes stored) */
> > + ret = text_size;
> > + kcov->size = text_size / sizeof(unsigned long);
> > + kcov_debug("text size = 0x%lx, roundup = 0x%x, kcov->size = 0x%x\n",
> > + ((unsigned long)&_etext) - ((unsigned long)&_stext),
> > + text_size,
> > + kcov->size);
> > + kcov->mode = KCOV_MODE_INIT_UNIQUE;
> > + } else {
> > + if (size < 2 || size > INT_MAX / sizeof(unsigned long))
> > + return -EINVAL;
> > + kcov->size = size;
> > + kcov->mode = KCOV_MODE_INIT_TRACE;
> > + }
> > + return ret;
>
> Thanks!
>
> --
> You received this message because you are subscribed to the Google Groups "kasan-dev" group.
> To unsubscribe from this group and stop receiving emails from it, send an email to kasan-dev+unsubscribe@...glegroups.com.
> To view this discussion on the web visit https://groups.google.com/d/msgid/kasan-dev/CA%2BfCnZcTi%3DQLGC_LCdhs%2BfMrxkqX66kXEuM5ewOmjVjifKzUrw%40mail.gmail.com.
Powered by blists - more mailing lists