lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Thu, 08 Jul 2010 14:52:28 -0400
From:	Prarit Bhargava <prarit@...hat.com>
To:	Linux Kernel <linux-kernel@...r.kernel.org>
CC:	jens.axboe@...cle.com, the arch/x86 maintainers <x86@...nel.org>,
	Don Zickus <dzickus@...hat.com>,
	Suresh Siddha <suresh.b.siddha@...el.com>
Subject: cpu softplug kernel hang

The panic below is from an 2.6.32-based kernel, however, AFAICT the same
issue exists with the latest 2.6.35-rc3+ kernel.

I have diagnosed the issue as being identical to the issue that I fixed
with the Intel rngd
driver sometime ago:

http://marc.info/?l=linux-kernel&m=117275119001289&w=2

When doing the following,

while true; do
        for i in `seq 12 23`; do echo 0 >
/sys/devices/system/cpu/cpu$i/online; done
        sleep 5
        for i in `seq 12 23`; do echo 1 >
/sys/devices/system/cpu/cpu$i/online; done
        sleep 5
done

I see (with the nmi_watchdog enabled)

BUG: NMI Watchdog detected LOCKUP on CPU11, ip ffffffff81029e72, registers:
CPU 11 
Modules linked in: nfs lockd fscache nfs_acl auth_rpcgss autofs4 sunrpc
cpufreq_ondemand acpi_cpufreq freq_table ipv6 dm_mirror dm_region_hash dm_log
uinput sg serio_raw i2c_i801 iTCO_wdt iTCO_vendor_support ioatdma i7core_edac
edac_core shpchp igb dca ext4 mbcache jbd2 sr_mod cdrom sd_mod crc_t10dif ahci
pata_acpi ata_generic pata_jmicron radeon ttm drm_kms_helper drm i2c_algo_bit
i2c_core dm_mod [last unloaded: microcode]

Modules linked in: nfs lockd fscache nfs_acl auth_rpcgss autofs4 sunrpc
cpufreq_ondemand acpi_cpufreq freq_table ipv6 dm_mirror dm_region_hash dm_log
uinput sg serio_raw i2c_i801 iTCO_wdt iTCO_vendor_support ioatdma i7core_edac
edac_core shpchp igb dca ext4 mbcache jbd2 sr_mod cdrom sd_mod crc_t10dif ahci
pata_acpi ata_generic pata_jmicron radeon ttm drm_kms_helper drm i2c_algo_bit
i2c_core dm_mod [last unloaded: microcode]
Pid: 704, comm: kexec Not tainted 2.6.32 #1 X8DTN
RIP: 0010:[<ffffffff81029e72>]  [<ffffffff81029e72>] ipi_handler+0x32/0xa0
RSP: 0000:ffff8801474a3f58  EFLAGS: 00000046
RAX: 0000000000000000 RBX: ffff880337393ea8 RCX: ffff88013ae41580
RDX: 00000000ffffffff RSI: 0000000000000000 RDI: ffff880337393ea8
RBP: ffff8801474a3f68 R08: 0000000061c941a6 R09: 00000000578070b9
R10: 0000000080507210 R11: 0000000025410601 R12: 0000000000000086
R13: 00000000ffffffff R14: ffffffff817491d0 R15: 0000000090793245
FS:  00007fefd5f3d700(0000) GS:ffff8801474a0000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 000000000040d000 CR3: 0000000316a8e000 CR4: 00000000000006e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process kexec (pid: 704, threadinfo ffff880314caa000, task ffff880335a0b500)
Stack:
 ffff880147571f40 000000000000000b ffff8801474a3f98 ffffffff810a6d28
<0> 00000000aa149910 00000000f21570f0 000000000008b495 00000000521ebd53
<0> ffff8801474a3fa8 ffffffff8102ea57 ffff880314cabf80 ffffffff81013e53
Call Trace:
 <IRQ> 
 [<ffffffff810a6d28>] generic_smp_call_function_interrupt+0x78/0x130
 [<ffffffff8102ea57>] smp_call_function_interrupt+0x27/0x40
 [<ffffffff81013e53>] call_function_interrupt+0x13/0x20
 <EOI> 
Code: 0f 1f 44 00 00 48 89 fb 9c 58 0f 1f 44 00 00 49 89 c4 fa 66 0f 1f 44 00
00 f0 ff 0f 8b 47 04 85 c0 75 0f 66 0f 1f 44 00 00 f3 90 <8b> 43 04 85 c0 74 f7
8b 7b 18 83 ff ff 74 47 48 8b 05 08 25 a1 

Since this is a panic, I get traces from all other cpus.

CPU 14 is in _write_lock_irq
CPU  2 is in _read_lock
CPU  6 has called smp_call_function() with the ipi_handler to sync mtrr's on
the new cpu

The problem is that ipi_handler does this:

static void ipi_handler(void *info)
{
#ifdef CONFIG_SMP
        struct set_mtrr_data *data = info;
        unsigned long flags;

        local_irq_save(flags);

        atomic_dec(&data->count); << global value that each processor entering
ipi_handler decrements
        while (!atomic_read(&data->gate)) << 1 when data->count != 0
                cpu_relax();

So what happens is that CPU 2 is in _read_lock and has acquired a lock.  CPU 14
is waiting for the release of that lock with IRQs *off*.

CPU 6 launches smp_call_function, and CPU 2 answers and runs the ipi_handler()
and waits (as do all other processors).

CPU 14, however, does not see the IPI because it is waiting with interrupts off
for the lock that CPU 2 is holding.

Boom.  Deadlock.

I'm planning on fixing this by rewriting this code and using stop_machine_run()
to synchronize the cpus instead of doing it in the ipi_handler.  

Unless, of course, someone has a better suggestion.

P.




--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ