lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAHk-=whuO3U90x_i6nq+xmVymwqcc=kkb5=gv4vyLScQn7ZwBw@mail.gmail.com>
Date:   Sat, 29 Aug 2020 10:54:47 -0700
From:   Linus Torvalds <torvalds@...ux-foundation.org>
To:     Daniel Exner <dex@...gonslave.de>, Christoph Hellwig <hch@....de>
Cc:     Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
        crazy@...galware.org
Subject: Re: Kernel 5.9-rc Regression: Boot failure with nvme

Just adding Christoph to the participants list, since at a guess it's
due to his changes whether they came from the nvme side or the dma
side..

Christoph?

                  Linus

On Sat, Aug 29, 2020 at 10:06 AM Daniel Exner <dex@...gonslave.de> wrote:
>
> Hi,
>
> (please keep me in the loop, as I'm currently not suscribed)
>
> both 5.9-rc1 and -rc2 fail to boot with my TOSHIBA-RD400 NVME:
>
> [    1.015590] ------------[ cut here ]------------
> [    1.015594] WARNING: CPU: 7 PID: 99 at mm/page_alloc.c:4864
> __alloc_pages_nodemask+0x299/0x330
> [    1.015594] Modules linked in: syscopyarea xhci_pci(+)
> xhci_pci_renesas sysfillrect xhci_hcd nvme aesni_intel(+) crypto_simd
> sysimgblt fb_sys_fops cryptd nvme_core t10_pi glue_helper drm hwmon
> scsi_mod agpgart i2c_core usbcore video wmi button dm_mirror
> dm_region_hash dm_log dm_mod unix ipv6 autofs4
> [    1.015602] CPU: 7 PID: 99 Comm: kworker/u16:1 Not tainted
> 5.9.0-rc2-dirty #12
> [    1.015603] Hardware name: To Be Filled By O.E.M. To Be Filled By
> O.E.M./Z170 Gaming K6, BIOS P7.50 10/18/2018
> [    1.015607] Workqueue: nvme-reset-wq nvme_reset_work [nvme]
> [    1.015608] RIP: 0010:__alloc_pages_nodemask+0x299/0x330
> [    1.015609] Code: 66 0f 85 46 ff ff ff e8 24 46 dd ff e9 3c ff ff ff
> e8 4b 4f fc ff 48 89 c7 e9 ad fe ff ff 81 e5 00 20 00 00 0f 85 7b ff ff
> ff <0f> 0b e9 74 ff ff ff 31 c0 e9 1b fe ff ff 65 48 8b 04 25 00 6d 01
> [    1.015610] RSP: 0000:ffffb3ed002abcb8 EFLAGS: 00010246
> [    1.015611] RAX: 0000000000000000 RBX: ffff9c8e827c4118 RCX:
> 0000000000000000
> [    1.015611] RDX: 0000000000000000 RSI: 0000000000000034 RDI:
> 0000000000000cc0
> [    1.015612] RBP: 0000000000000000 R08: 0000000000000000 R09:
> ffffffffffffffff
> [    1.015612] R10: 0000000000000006 R11: ffffb3ee002abb97 R12:
> 0000000000000000
> [    1.015613] R13: 0000000000000000 R14: ffff9c8e921060b0 R15:
> 0000000000000cc0
> [    1.015614] FS:  0000000000000000(0000) GS:ffff9c8e96bc0000(0000)
> knlGS:0000000000000000
> [    1.015615] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [    1.015615] CR2: 0000559988e2f4d8 CR3: 00000008433e4004 CR4:
> 00000000003706e0
> [    1.015616] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> 0000000000000000
> [    1.015617] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
> 0000000000000400
> [    1.015617] Call Trace:
> [    1.015621]  dma_direct_alloc_pages+0x1e9/0x2c0
> [    1.015623]  ? pci_alloc_irq_vectors_affinity+0xa5/0x100
> [    1.015626]  nvme_alloc_queue+0x10a/0x170 [nvme]
> [    1.015629]  nvme_reset_work+0x70b/0x12b0 [nvme]
> [    1.015631]  ? nvme_irq_check+0x30/0x30 [nvme]
> [    1.015634]  process_one_work+0x1da/0x3d0
> [    1.015636]  worker_thread+0x4a/0x3c0
> [    1.015637]  ? process_one_work+0x3d0/0x3d0
> [    1.015638]  kthread+0x114/0x160
> [    1.015640]  ? kthread_park+0x90/0x90
> [    1.015641]  ret_from_fork+0x22/0x30
> [    1.015643] ---[ end trace 268d4f4db1ef121e ]---
>
>
> Resulting in:
> [    1.015644] nvme nvme0: Removing after probe failure status: -12
>
> If you need more infos I can try to provide them.
>
> Greetings
> Daniel
>
>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ