lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CACT4Y+Z+7zGD4_d-snEMLTin1a0C-yYQGKc0n+E6y7dBzY-OgQ@mail.gmail.com>
Date:   Thu, 4 Feb 2021 14:37:21 +0100
From:   Dmitry Vyukov <dvyukov@...gle.com>
To:     Peter Zijlstra <peterz@...radead.org>
Cc:     Ingo Molnar <mingo@...hat.com>,
        Arnaldo Carvalho de Melo <acme@...nel.org>,
        Mark Rutland <mark.rutland@....com>,
        Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
        Jiri Olsa <jolsa@...hat.com>,
        Namhyung Kim <namhyung@...nel.org>,
        Will Deacon <will@...nel.org>,
        LKML <linux-kernel@...r.kernel.org>,
        Matt Morehouse <mascasa@...gle.com>
Subject: Re: Process-wide watchpoints

On Thu, Feb 4, 2021 at 2:33 PM Peter Zijlstra <peterz@...radead.org> wrote:
>
> On Thu, Feb 04, 2021 at 01:53:59PM +0100, Dmitry Vyukov wrote:
> > On Thu, Feb 4, 2021 at 1:09 PM Peter Zijlstra <peterz@...radead.org> wrote:
>
> > > What do we do then? The advantage of IOC_REFRESH is that it disables the
> > > event until it gets explicitly re-armed, avoiding recursion issues etc.
> > > Do you want those semantics? If so, we'd need to have IOC_REFRESH find
> > > the actual event for the current task, which should be doable I suppose.
> >
> > Frankly, I don't know. I didn't use it in my prototype, nor I fully
> > understand what it's doing. Does it make sense for breakpoints?
> > I see IOC_REFRESH has a check for !attr.inherit, so it will fail for
> > my use case currently. I would say we just leave it as is for now.
>
> Well, the way it works is that currently you set event_limit > 0. Then
> each event will decrement, when we hit 0 we disable and raise a signal.
>
> REFRESH will increment event_limit and re-enable.
>
> This means you're guaranteed not to get another signal until you're
> ready for it. It allows leaving the signal handler context to handle the
> signal.
>
> I suppose you're looking for something like this, which goes in top of
> that thread_only thing.
>
> --- a/include/uapi/linux/perf_event.h
> +++ b/include/uapi/linux/perf_event.h
> @@ -389,7 +389,8 @@ struct perf_event_attr {
>                                 cgroup         :  1, /* include cgroup events */
>                                 text_poke      :  1, /* include text poke events */
>                                 thread_only    :  1, /* only inherit on threads */
> -                               __reserved_1   : 29;
> +                               sigtrap        :  1, /* foo */
> +                               __reserved_1   : 28;
>
>         union {
>                 __u32           wakeup_events;    /* wakeup every n events */
> --- a/kernel/events/core.c
> +++ b/kernel/events/core.c
> @@ -6273,6 +6273,13 @@ static void perf_pending_event_disable(s
>
>         if (cpu == smp_processor_id()) {
>                 WRITE_ONCE(event->pending_disable, -1);
> +
> +               if (event->attr.sigtrap) {
> +                       atomic_inc(&event->event_limit); /* rearm */
> +                       send_sig_info(SIGTRAP, SEND_SIG_PRIV, current);
> +                       return;
> +               }
> +
>                 perf_event_disable_local(event);
>                 return;
>         }
> @@ -8936,6 +8943,7 @@ static int __perf_event_overflow(struct
>                                    int throttle, struct perf_sample_data *data,
>                                    struct pt_regs *regs)
>  {
> +       perf_overflow_handler_t ovf;
>         int events = atomic_read(&event->event_limit);
>         int ret = 0;
>
> @@ -8961,7 +8969,15 @@ static int __perf_event_overflow(struct
>                 perf_event_disable_inatomic(event);
>         }
>
> -       READ_ONCE(event->overflow_handler)(event, data, regs);
> +       ovf = READ_ONCE(event->overflow_handler);
> +#ifdef CONFIG_RETPOLINE
> +       if (ovf == perf_event_output_forward) {
> +               perf_event_output_forward(event, data, regs);
> +       } else if (ovf == perf_event_output_backward) {
> +               perf_event_output_backward(event, data, regs);
> +       } else
> +#endif
> +               ovf(event, data, regs);
>
>         if (*perf_event_fasync(event) && event->pending_kill) {
>                 event->pending_wakeup = 1;
> @@ -11281,6 +11297,9 @@ perf_event_alloc(struct perf_event_attr
>
>         event->state            = PERF_EVENT_STATE_INACTIVE;
>
> +       if (event->attr.sigtrap)
> +               event->event_limit = ATOMIC_INIT(1);
> +
>         if (task) {
>                 event->attach_state = PERF_ATTACH_TASK;
>                 /*
> @@ -11556,6 +11575,9 @@ static int perf_copy_attr(struct perf_ev
>         if (attr->thread_only && !attr->inherit)
>                 return -EINVAL;
>
> +       if (attr->sigtrap && attr->inherit && !attr->thread_only)
> +               return -EINVAL;
> +
>  out:
>         return ret;


Thanks. Let me see if this will work for us and test.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ