lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 29 Sep 2010 15:56:08 +0200
From:	Stephane Eranian <eranian@...gle.com>
To:	Robert Richter <robert.richter@....com>
Cc:	"mingo@...hat.com" <mingo@...hat.com>,
	"hpa@...or.com" <hpa@...or.com>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	"yinghai@...nel.org" <yinghai@...nel.org>,
	"andi@...stfloor.org" <andi@...stfloor.org>,
	"peterz@...radead.org" <peterz@...radead.org>,
	"gorcunov@...il.com" <gorcunov@...il.com>,
	"ying.huang@...el.com" <ying.huang@...el.com>,
	"fweisbec@...il.com" <fweisbec@...il.com>,
	"ming.m.lin@...el.com" <ming.m.lin@...el.com>,
	"tglx@...utronix.de" <tglx@...utronix.de>,
	"dzickus@...hat.com" <dzickus@...hat.com>,
	"mingo@...e.hu" <mingo@...e.hu>
Subject: Re: [tip:perf/urgent] perf, x86: Catch spurious interrupts after
 disabling counters

On Wed, Sep 29, 2010 at 3:39 PM, Robert Richter <robert.richter@....com> wrote:
> On 29.09.10 09:13:30, Stephane Eranian wrote:
>
>> for_each_set_bit(bit, (unsigned long *)&status, X86_PMC_IDX_MAX) {
>>                 struct perf_event *event = cpuc->events[bit];
>>
>>                 handled++;
>>
>>                 if (!test_bit(bit, cpuc->active_mask))
>
>                        /* spurious interrupt here */
>
>>                     continue;
>> }
>>
>> I think the logic is similar. What makes the difference, it seems, is that
>> handled is incremented unconditionally if the ovfl_mask says it has
>> an overflow, i.e., before active_mask is checked.
>
> Note that we can use here for_each_set_bit() since we have the status
> mask. So we may increment handled always.
>
> On AMD we use for_each_counter(), but only check active counters to
> avoid unnecessary rdmsrl()s for unused counters. But here, we only can
> increment handled if we detect an overflow or if we know a counter was
> disabled.
>
>> On Westmere, we've seen situations where the overflow mask and active
>> mask did not agree.
>
> It's the 'spurious interrupt' branch above.
>
>> On counter disable, the overflow mask bit is not cleared, thus one may iterate
>> in the loop and fail the active_mask. But handled would be incremented in that
>> case, so that would behave like in your patch.
>
> Right, spurious interrupts are counted and a 'handled' is returned.
>
Ok, I think we agree on this. It is handled in the Intel case, though it
is not clearly explained with a comment.
The P4 case needs to be fixed.

Here is another difference I noticed in x86_handle_irq() vs.
intel_pmu_handle_irq().
For Intel, handled is incremented even if there is no 64-bit overflow.

With generic X86, it is incremented only when you have a 64-bit
overflow. I think that's wrong. You don't hit that condition very often
on AMD because counters are 47 bits wide, but this is generic code
and on P6 you definitively will. I believe you need to hoist handled++
just after the check on active_mask.


What do you think?

> -Robert
>
> --
> Advanced Micro Devices, Inc.
> Operating System Research Center
>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ