[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID:
<PNZPR01MB4528980B1F286C5843B70DBFFE322@PNZPR01MB4528.INDPRD01.PROD.OUTLOOK.COM>
Date: Sat, 7 Dec 2024 12:26:14 +0800
From: Chen Wang <unicorn_wang@...look.com>
To: kbusch@...nel.org, axboe@...nel.dk, hch@....de, sagi@...mberg.me,
linux-nvme@...ts.infradead.org,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Cc: akpm@...ux-foundation.org, linux-mm@...ck.org, fengchun.li@...hgo.com,
Inochi Amaoto <inochiama@...look.com>
Subject: Re: [bug report] WARNING when nvme_probe
Adding CC to Inochi, who might be intereseted in too.
On 2024/12/7 12:22, Chen Wang wrote:
> Hello,
>
> When I test with latest 6.13-rc1, I see kernel complains as below, but
> I don't see this in 6.12 serials (6.12-rc1 & 6.12-rc7).
>
> I'm developing PCIe driver and there is a RC switch which has some
> devices conneted to, including a nvme device.
>
> Can somebody help double-check and comment?
>
> Sent to NVME related mails and CCed to MM, who might be interested in
> this question.
>
> Thanks,
>
> Chen
>
>
> [ 2.075408] ------------[ cut here ]------------
> [ 2.075414] WARNING: CPU: 51 PID: 403 at mm/page_alloc.c:4727
> __alloc_pages_noprof+0x788/0xd08
> [ 2.075432] Modules linked in:
> [ 2.075442] CPU: 51 UID: 0 PID: 403 Comm: kworker/u257:0 Not
> tainted 6.13.0-rc1+ #1
> [ 2.075449] Hardware name: Milk-V Pioneer (DT)
> [ 2.075454] Workqueue: async async_run_entry_fn
> [ 2.075467] epc : __alloc_pages_noprof+0x788/0xd08
> [ 2.075473] ra : __dma_direct_alloc_pages.isra.0+0x1f8/0x434
> [ 2.075480] epc : ffffffff8029e252 ra : ffffffff800d1f96 sp :
> ffffffd7017bb840
> [ 2.075485] gp : ffffffff82072a28 tp : ffffffd701761c00 t0 :
> ffffffd703b023b8
> [ 2.075488] t1 : 0000000000000000 t2 : 6d7768206e6f6d77 s0 :
> ffffffd7017bb9e0
> [ 2.075491] s1 : 0000000000000cc0 a0 : 0000000000000cc0 a1 :
> 000000000000000e
> [ 2.075495] a2 : 0000000000000000 a3 : 0000000000000000 a4 :
> 0000000000000001
> [ 2.075498] a5 : ffffffff82013d7a a6 : 3fffffffffffffff a7 :
> 000000000000003e
> [ 2.075501] s2 : ffffffffffffffff s3 : 0000000000000000 s4 :
> 0000000004000000
> [ 2.075504] s5 : 0000000000003fff s6 : ffffffffffffffff s7 :
> 000000000000000e
> [ 2.075508] s8 : ffffffff8213d190 s9 : 000000000000000e s10:
> 0000000000000100
> [ 2.075510] s11: ffffffd70034d3c0 t3 : ffffffd700260f00 t4 :
> ffffffd700260f00
> [ 2.075514] t5 : ffffffd700260000 t6 : ffffffd7017bb6d8
> [ 2.075516] status: 0000000200000120 badaddr: 0000000000000000
> cause: 0000000000000003
> [ 2.075521] [<ffffffff8029e252>] __alloc_pages_noprof+0x788/0xd08
> [ 2.075527] [<ffffffff800d1f96>]
> __dma_direct_alloc_pages.isra.0+0x1f8/0x434
> [ 2.075533] [<ffffffff800d27ac>] dma_direct_alloc_pages+0x28/0x11c
> [ 2.075537] [<ffffffff800cdc54>] __dma_alloc_pages+0x3a/0x54
> [ 2.075547] [<ffffffff800d199c>] dma_alloc_noncontiguous+0x8e/0x1c6
> [ 2.075551] [<ffffffff80a3f352>] nvme_setup_host_mem+0x2a8/0x446
> [ 2.075558] [<ffffffff80a42dbc>] nvme_probe+0x730/0x802
> [ 2.075564] [<ffffffff806fb4dc>] local_pci_probe+0x40/0x82
> [ 2.075572] [<ffffffff806fc1a6>] pci_device_probe+0xca/0x1f0
> [ 2.075578] [<ffffffff809a2964>] really_probe+0x9e/0x33e
> [ 2.075589] [<ffffffff809a2c80>] __driver_probe_device+0x7c/0x132
> [ 2.075595] [<ffffffff809a2e0c>] driver_probe_device+0x38/0xc6
> [ 2.075601] [<ffffffff809a3182>]
> __driver_attach_async_helper+0x4a/0xc4
> [ 2.075607] [<ffffffff8006616a>] async_run_entry_fn+0x34/0x126
> [ 2.075613] [<ffffffff80055a16>] process_one_work+0x172/0x316
> [ 2.075621] [<ffffffff80056626>] worker_thread+0x2e2/0x448
> [ 2.075626] [<ffffffff8005ee4a>] kthread+0xc2/0xe2
> [ 2.075633] [<ffffffff80e8de56>] ret_from_fork+0xe/0x1c
> [ 2.075645] ---[ end trace 0000000000000000 ]---
> [ 2.094302] nvme nvme0: allocated 64 MiB host memory buffer (16
> segments).
> [ 2.223519] nvme nvme0: 1/0/0 default/read/poll queues
>
>
Powered by blists - more mailing lists