lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <2a400ffa-9653-5805-3cd5-937b102ef056@redhat.com>
Date:   Mon, 16 Mar 2020 10:20:09 -0400
From:   Waiman Long <longman@...hat.com>
To:     Guenter Roeck <linux@...ck-us.net>
Cc:     Frederic Weisbecker <fweisbec@...il.com>,
        Thomas Gleixner <tglx@...utronix.de>,
        Ingo Molnar <mingo@...nel.org>, linux-kernel@...r.kernel.org,
        Jeremy Linton <jeremy.linton@....com>, pbunyan@...hat.com
Subject: Re: [RFC PATCH v2] tick: Make tick_periodic() check for missing ticks

On 3/15/20 10:57 PM, Guenter Roeck wrote:
> On 3/15/20 7:43 PM, Waiman Long wrote:
>> On 3/15/20 10:20 PM, Guenter Roeck wrote:
>>> Hi,
>>>
>>> On Fri, Feb 07, 2020 at 02:39:29PM -0500, Waiman Long wrote:
>>>> The tick_periodic() function is used at the beginning part of the
>>>> bootup process for time keeping while the other clock sources are
>>>> being initialized.
>>>>
>>>> The current code assumes that all the timer interrupts are handled in
>>>> a timely manner with no missing ticks. That is not actually true. Some
>>>> ticks are missed and there are some discrepancies between the tick time
>>>> (jiffies) and the timestamp reported in the kernel log.  Some systems,
>>>> however, are more prone to missing ticks than the others.  In the extreme
>>>> case, the discrepancy can actually cause a soft lockup message to be
>>>> printed by the watchdog kthread. For example, on a Cavium ThunderX2
>>>> Sabre arm64 system:
>>>>
>>>>  [   25.496379] watchdog: BUG: soft lockup - CPU#14 stuck for 22s!
>>>>
>>>> On that system, the missing ticks are especially prevalent during the
>>>> smp_init() phase of the boot process. With an instrumented kernel,
>>>> it was found that it took about 24s as reported by the timestamp for
>>>> the tick to accumulate 4s of time.
>>>>
>>>> Investigation and bisection done by others seemed to point to the
>>>> commit 73f381660959 ("arm64: Advertise mitigation of Spectre-v2, or
>>>> lack thereof") as the culprit. It could also be a firmware issue as
>>>> new firmware was promised that would fix the issue.
>>>>
>>>> To properly address this problem, we cannot assume that there will
>>>> be no missing tick in tick_periodic(). This function is now modified
>>>> to follow the example of tick_do_update_jiffies64() by using another
>>>> reference clock to check for missing ticks. Since the watchdog timer
>>>> uses running_clock(), it is used here as the reference. With this patch
>>>> applied, the soft lockup problem in the arm64 system is gone and tick
>>>> time tracks much more closely to the timestamp time.
>>>>
>>>> Signed-off-by: Waiman Long <longman@...hat.com>
>>> Since this patch is in linux-next, roughly 10% of my x86 and x86_64
>>> qemu emulation boots are stalling. Typical log:
>>>
>>> [    0.002016] smpboot: Total of 1 processors activated (7576.40 BogoMIPS)
>>> [    0.002016] devtmpfs: initialized
>>> [    0.002016] clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
>>> [    0.002016] futex hash table entries: 256 (order: 3, 32768 bytes, linear)
>>> [    0.002016] xor: measuring software checksum speed
>>>
>>> another:
>>>
>>> [    0.002653] Freeing SMP alternatives memory: 44K
>>> [    0.002653] smpboot: CPU0: Intel Westmere E56xx/L56xx/X56xx (IBRS update) (family: 0x6, model: 0x2c, stepping: 0x1)
>>> [    0.002653] Performance Events: unsupported p6 CPU model 44 no PMU driver, software events only.
>>> [    0.002653] rcu: Hierarchical SRCU implementation.
>>> [    0.002653] smp: Bringing up secondary CPUs ...
>>> [    0.002653] x86: Booting SMP configuration:
>>> [    0.002653] .... node  #0, CPUs:      #1
>>> [    0.000000] smpboot: CPU 1 Converting physical 0 to logical die 1
>>>
>>> ... and then there is silence until the test aborts.
>>>
>>> This is only (or at least predominantly) seen if the system running
>>> the emulation is under load.
>>>
>>> Reverting this patch fixes the problem.
>> I was aware that there are some problem with this patch, but it is hard
>> to reproduce it. Do you have a more consistent way to reproduce it.
>> When  you say under load, you mean that the host system is also busy so
>> that there are a lot of vcpu preemption. Right? Could you give me the
> Correct. I am able to reproduce the problem quite reliably (ie 2-3 boots
> out of ~25 fail) if I run a kernel compilation in parallel, but not (or
> rarely) if the system is otherwise idle.
>
>> x86-64 .config file that you use?
>>
> Attached. It is pretty much defconfig with various debug and test options
> enabled.
>
> Guenter

Thanks,
Longman

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ