lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <031340ee-7c64-42ef-9459-604f0543708a@huaweicloud.com>
Date:   Fri, 19 May 2023 15:21:41 +0800
From:   Yu Kuai <yukuai1@...weicloud.com>
To:     yang lan <lanyang0908@...il.com>, Yu Kuai <yukuai1@...weicloud.com>
Cc:     Matthew Wilcox <willy@...radead.org>, viro@...iv.linux.org.uk,
        linux-kernel@...r.kernel.org, josef@...icpanda.com,
        linux-block@...r.kernel.org, nbd@...er.debian.org,
        syzkaller-bugs@...glegroups.com, linux-fsdevel@...r.kernel.org,
        axboe@...nel.dk, haris.iqbal@...os.com, jinpu.wang@...os.com,
        brauner@...nel.org, "yukuai (C)" <yukuai3@...wei.com>
Subject: Re: INFO: task hung in blkdev_open bug

Hi,

在 2023/05/19 15:12, yang lan 写道:
> Hi,
> 
> ./rqos/wbt/wb_background:4
> ./rqos/wbt/wb_normal:8
> ./rqos/wbt/unknown_cnt:0
> ./rqos/wbt/min_lat_nsec:2000000
> ./rqos/wbt/inflight:0: inflight 0
> ./rqos/wbt/inflight:1: inflight 0
> ./rqos/wbt/inflight:2: inflight 0
> ./rqos/wbt/id:0
> ./rqos/wbt/enabled:1
> ./rqos/wbt/curr_win_nsec:0
> ./hctx0/type:default
> ./hctx0/dispatch_busy:0
> ./hctx0/active:0
> ./hctx0/run:1
> ./hctx0/sched_tags_bitmap:00000000: 0100 0000 0000 0000 0000 0000 0000 0000
> ./hctx0/sched_tags_bitmap:00000010: 0000 0000 0000 0000 0000 0000 0000 0000
> ./hctx0/sched_tags:nr_tags=256
> ./hctx0/sched_tags:nr_reserved_tags=0
> ./hctx0/sched_tags:active_queues=0
> ./hctx0/sched_tags:bitmap_tags:
> ./hctx0/sched_tags:depth=256
> ./hctx0/sched_tags:busy=1
> ./hctx0/sched_tags:cleared=0
> ./hctx0/sched_tags:bits_per_word=64
> ./hctx0/sched_tags:map_nr=4
> ./hctx0/sched_tags:alloc_hint={245, 45}
> ./hctx0/sched_tags:wake_batch=8
> ./hctx0/sched_tags:wake_index=0
> ./hctx0/sched_tags:ws_active=0
> ./hctx0/sched_tags:ws={
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:     {.wait=inactive},
> ./hctx0/sched_tags:}
> ./hctx0/sched_tags:round_robin=0
> ./hctx0/sched_tags:min_shallow_depth=192
> ./hctx0/tags_bitmap:00000000: 0000 0000 0100 0000 0000 0000 0000 0000
> ./hctx0/tags:nr_tags=128
> ./hctx0/tags:nr_reserved_tags=0
> ./hctx0/tags:active_queues=0
> ./hctx0/tags:bitmap_tags:
> ./hctx0/tags:depth=128
> ./hctx0/tags:busy=1

tags:busy and sched_tags:busy shows that one io is inflight.

> ./hctx0/tags:cleared=0
> ./hctx0/tags:bits_per_word=32
> ./hctx0/tags:map_nr=4
> ./hctx0/tags:alloc_hint={123, 51}
> ./hctx0/tags:wake_batch=8
> ./hctx0/tags:wake_index=0
> ./hctx0/tags:ws_active=0
> ./hctx0/tags:ws={
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:   {.wait=inactive},
> ./hctx0/tags:}
> ./hctx0/tags:round_robin=0
> ./hctx0/tags:min_shallow_depth=4294967295
> ./hctx0/ctx_map:00000000: 00
> ./hctx0/busy:ffff888016860000 {.op=READ, .cmd_flags=,
> .rq_flags=STARTED|ELVPRIV|IO_STAT|STATS|ELV, .state=in_flight,
> .tag=32, .internal_tag=0}

About "busy" means this io is issued to driver and yet not finished. So
next step is to dig out where is the io in the driver and why it's not
done.

Thanks,
Kuai
> ./hctx0/flags:alloc_policy=FIFO SHOULD_MERGE|BLOCKING
> ./sched/queued:0 1 0
> ./sched/owned_by_driver:0 1 0
> ./sched/async_depth:192
> ./sched/starved:0
> ./sched/batching:1
> ./state:SAME_COMP|NONROT|IO_STAT|INIT_DONE|STATS|REGISTERED|NOWAIT|30
> ./pm_only:0
> 
> So how can we know where the io is?
> 
> Regards,
> 
> Yang
> 
> Yu Kuai <yukuai1@...weicloud.com> 于2023年5月18日周四 11:30写道:
>>
>> Hi,
>>
>> 在 2023/05/18 0:27, yang lan 写道:
>>> Hi,
>>>
>>> Thank you for your response.
>>>
>>>> Does this reproduce on current kernels, eg 6.4-rc2?
>>>
>>> Yeah, it can be reproduced on kernel 6.4-rc2.
>>>
>>
>> Below log shows that io hang, can you collect following debugfs so
>> that we can know where is the io now.
>>
>> cd /sys/kernel/debug/block/[test_device] && find . -type f -exec grep
>> -aH . {} \;
>>
>> Thanks,
>> Kuai
>>> root@...kaller:~# uname -a
>>> Linux syzkaller 6.4.0-rc2 #1 SMP PREEMPT_DYNAMIC Wed May 17 22:58:52
>>> CST 2023 x86_64 GNU/Linux
>>> root@...kaller:~# gcc poc_blkdev.c -o poc_blkdev
>>> root@...kaller:~# ./poc_blkdev
>>> [  128.718051][ T7121] nbd0: detected capacity change from 0 to 4
>>> [  158.917678][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 30 seconds
>>> [  188.997677][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 60 seconds
>>> [  219.077191][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 90 seconds
>>> [  249.157312][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 120 seconds
>>> [  279.237409][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 150 seconds
>>> [  309.317843][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 180 seconds
>>> [  339.397950][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 210 seconds
>>> [  369.478031][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 240 seconds
>>> [  399.558253][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 270 seconds
>>> [  429.638372][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 300 seconds
>>> [  459.718454][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 330 seconds
>>> [  489.798571][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 360 seconds
>>> [  519.878643][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 390 seconds
>>> [  549.958966][  T998] block nbd0: Possible stuck request
>>> ffff888016f08000: control (read@0,2048B). Runtime 420 seconds
>>> [  571.719145][   T30] INFO: task systemd-udevd:7123 blocked for more
>>> than 143 seconds.
>>> [  571.719652][   T30]       Not tainted 6.4.0-rc2 #1
>>> [  571.719900][   T30] "echo 0 >
>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>>> [  571.720307][   T30] task:systemd-udevd   state:D stack:26224
>>> pid:7123  ppid:3998   flags:0x00004004
>>> [  571.720756][   T30] Call Trace:
>>> [  571.720923][   T30]  <TASK>
>>> [  571.721073][   T30]  __schedule+0x9ca/0x2630
>>> [  571.721348][   T30]  ? firmware_map_remove+0x1e0/0x1e0
>>> [  571.721618][   T30]  ? find_held_lock+0x33/0x1c0
>>> [  571.721866][   T30]  ? lock_release+0x3b9/0x690
>>> [  571.722108][   T30]  ? do_read_cache_folio+0x4ff/0xb20
>>> [  571.722447][   T30]  ? lock_downgrade+0x6b0/0x6b0
>>> [  571.722785][   T30]  ? mark_held_locks+0xb0/0x110
>>> [  571.723044][   T30]  schedule+0xd3/0x1b0
>>> [  571.723264][   T30]  io_schedule+0x1b/0x70
>>> [  571.723489][   T30]  ? do_read_cache_folio+0x58c/0xb20
>>> [  571.723760][   T30]  do_read_cache_folio+0x58c/0xb20
>>> [  571.724036][   T30]  ? blkdev_readahead+0x20/0x20
>>> [  571.724319][   T30]  ? __filemap_get_folio+0x8e0/0x8e0
>>> [  571.724588][   T30]  ? __sanitizer_cov_trace_switch+0x53/0x90
>>> [  571.724885][   T30]  ? __sanitizer_cov_trace_pc+0x1e/0x50
>>> [  571.725246][   T30]  ? format_decode+0x1cf/0xb50
>>> [  571.725547][   T30]  ? __sanitizer_cov_trace_pc+0x1e/0x50
>>> [  571.725837][   T30]  ? fill_ptr_key+0x30/0x30
>>> [  571.726072][   T30]  ? default_pointer+0x4a0/0x4a0
>>> [  571.726335][   T30]  ? __isolate_free_page+0x220/0x220
>>> [  571.726608][   T30]  ? filemap_fdatawrite_wbc+0x1c0/0x1c0
>>> [  571.726888][   T30]  ? __sanitizer_cov_trace_pc+0x1e/0x50
>>> [  571.727172][   T30]  ? read_part_sector+0x229/0x420
>>> [  571.727434][   T30]  ? adfspart_check_ADFS+0x560/0x560
>>> [  571.727707][   T30]  read_part_sector+0xfa/0x420
>>> [  571.727963][   T30]  adfspart_check_POWERTEC+0x90/0x690
>>> [  571.728244][   T30]  ? adfspart_check_ADFS+0x560/0x560
>>> [  571.728520][   T30]  ? __kasan_slab_alloc+0x33/0x70
>>> [  571.728780][   T30]  ? adfspart_check_ICS+0x8f0/0x8f0
>>> [  571.729889][   T30]  ? snprintf+0xb2/0xe0
>>> [  571.730145][   T30]  ? vsprintf+0x30/0x30
>>> [  571.730374][   T30]  ? __sanitizer_cov_trace_pc+0x1e/0x50
>>> [  571.730659][   T30]  ? adfspart_check_ICS+0x8f0/0x8f0
>>> [  571.730928][   T30]  bdev_disk_changed+0x674/0x1260
>>> [  571.731189][   T30]  ? write_comp_data+0x1f/0x70
>>> [  571.731439][   T30]  ? iput+0xd0/0x780
>>> [  571.731646][   T30]  blkdev_get_whole+0x186/0x260
>>> [  571.731886][   T30]  blkdev_get_by_dev+0x4ce/0xae0
>>> [  571.732139][   T30]  blkdev_open+0x140/0x2c0
>>> [  571.732366][   T30]  do_dentry_open+0x6de/0x1450
>>> [  571.732612][   T30]  ? blkdev_close+0x80/0x80
>>> [  571.732848][   T30]  path_openat+0xd6d/0x26d0
>>> [  571.733084][   T30]  ? lock_downgrade+0x6b0/0x6b0
>>> [  571.733336][   T30]  ? vfs_path_lookup+0x110/0x110
>>> [  571.733591][   T30]  do_filp_open+0x1bb/0x290
>>> [  571.733824][   T30]  ? may_open_dev+0xf0/0xf0
>>> [  571.734061][   T30]  ? __phys_addr_symbol+0x30/0x70
>>> [  571.734324][   T30]  ? do_raw_spin_unlock+0x176/0x260
>>> [  571.734595][   T30]  do_sys_openat2+0x5fd/0x980
>>> [  571.734837][   T30]  ? file_open_root+0x3f0/0x3f0
>>> [  571.735087][   T30]  ? seccomp_notify_ioctl+0xff0/0xff0
>>> [  571.735368][   T30]  do_sys_open+0xce/0x140
>>> [  571.735596][   T30]  ? filp_open+0x80/0x80
>>> [  571.735820][   T30]  ? __secure_computing+0x1e3/0x340
>>> [  571.736090][   T30]  do_syscall_64+0x38/0x80
>>> [  571.736325][   T30]  entry_SYSCALL_64_after_hwframe+0x63/0xcd
>>> [  571.736626][   T30] RIP: 0033:0x7fb212210840
>>> [  571.736857][   T30] RSP: 002b:00007fffb37bbbe8 EFLAGS: 00000246
>>> ORIG_RAX: 0000000000000002
>>> [  571.737269][   T30] RAX: ffffffffffffffda RBX: 0000560e09072e10
>>> RCX: 00007fb212210840
>>> [  571.737651][   T30] RDX: 0000560e08e39fe3 RSI: 00000000000a0800
>>> RDI: 0000560e090813b0
>>> [  571.738037][   T30] RBP: 00007fffb37bbd60 R08: 0000560e08e39670
>>> R09: 0000000000000010
>>> [  571.738432][   T30] R10: 0000560e08e39d0c R11: 0000000000000246
>>> R12: 00007fffb37bbcb0
>>> [  571.739563][   T30] R13: 0000560e09087a70 R14: 0000000000000003
>>> R15: 000000000000000e
>>> [  571.739973][   T30]  </TASK>
>>> [  571.740133][   T30]
>>> [  571.740133][   T30] Showing all locks held in the system:
>>> [  571.740495][   T30] 1 lock held by rcu_tasks_kthre/13:
>>> [  571.740758][   T30]  #0: ffffffff8b6badd0
>>> (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at:
>>> rcu_tasks_one_gp+0x2b/0xdb0
>>> [  571.741301][   T30] 1 lock held by rcu_tasks_trace/14:
>>> [  571.741571][   T30]  #0: ffffffff8b6baad0
>>> (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at:
>>> rcu_tasks_one_gp+0x2b/0xdb0
>>> [  571.742134][   T30] 1 lock held by khungtaskd/30:
>>> [  571.742385][   T30]  #0: ffffffff8b6bb960
>>> (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x5b/0x300
>>> [  571.742947][   T30] 2 locks held by kworker/u8:0/50:
>>> [  571.743198][   T30]  #0: ffff888016e7b138
>>> ((wq_completion)nbd0-recv){+.+.}-{0:0}, at:
>>> process_one_work+0x94b/0x17b0
>>> [  571.743809][   T30]  #1: ffff888011e4fdd0
>>> ((work_completion)(&args->work)){+.+.}-{0:0}, at:
>>> process_one_work+0x984/0x17b0
>>> [  571.744393][   T30] 1 lock held by in:imklog/6784:
>>> [  571.744643][   T30]  #0: ffff88801106e368
>>> (&f->f_pos_lock){+.+.}-{3:3}, at: __fdget_pos+0xe9/0x100
>>> [  571.745122][   T30] 1 lock held by systemd-udevd/7123:
>>> [  571.745381][   T30]  #0: ffff8880431854c8
>>> (&disk->open_mutex){+.+.}-{3:3}, at: blkdev_get_by_dev+0x24b/0xae0
>>> [  571.745885][   T30]
>>> [  571.746008][   T30] =============================================
>>> [  571.746008][   T30]
>>> [  571.746424][   T30] NMI backtrace for cpu 1
>>> [  571.746642][   T30] CPU: 1 PID: 30 Comm: khungtaskd Not tainted 6.4.0-rc2 #1
>>> [  571.746989][   T30] Hardware name: QEMU Standard PC (i440FX + PIIX,
>>> 1996), BIOS 1.12.0-1 04/01/2014
>>> [  571.747440][   T30] Call Trace:
>>> [  571.747606][   T30]  <TASK>
>>> [  571.747764][   T30]  dump_stack_lvl+0x91/0xf0
>>> [  571.747997][   T30]  nmi_cpu_backtrace+0x21a/0x2b0
>>> [  571.748257][   T30]  ? lapic_can_unplug_cpu+0xa0/0xa0
>>> [  571.748525][   T30]  nmi_trigger_cpumask_backtrace+0x28c/0x2f0
>>> [  571.748830][   T30]  watchdog+0xe4b/0x10c0
>>> [  571.749057][   T30]  ? proc_dohung_task_timeout_secs+0x90/0x90
>>> [  571.749366][   T30]  kthread+0x33b/0x430
>>> [  571.749596][   T30]  ? kthread_complete_and_exit+0x40/0x40
>>> [  571.749891][   T30]  ret_from_fork+0x1f/0x30
>>> [  571.750126][   T30]  </TASK>
>>> [  571.750347][   T30] Sending NMI from CPU 1 to CPUs 0:
>>> [  571.750620][    C0] NMI backtrace for cpu 0
>>> [  571.750626][    C0] CPU: 0 PID: 3987 Comm: systemd-journal Not
>>> tainted 6.4.0-rc2 #1
>>> [  571.750637][    C0] Hardware name: QEMU Standard PC (i440FX + PIIX,
>>> 1996), BIOS 1.12.0-1 04/01/2014
>>> [  571.750643][    C0] RIP: 0033:0x7fb1d8c34bd1
>>> [  571.750652][    C0] Code: ed 4d 89 cf 75 a3 0f 1f 00 48 85 ed 75 4b
>>> 48 8b 54 24 28 48 8b 44 24 18 48 8b 7c 24 20 48 29 da 48 8b 70 20 48
>>> 0f af 54 24 08 <48> 83 c4 38 5b 5d 41 5c 41 5d 41 5e 41 5f e9 ac f2 04
>>> 00 0f 1f 40
>>> [  571.750662][    C0] RSP: 002b:00007ffff9686c30 EFLAGS: 00000202
>>> [  571.750670][    C0] RAX: 00007ffff9686e50 RBX: 0000000000000002
>>> RCX: 0000000000000010
>>> [  571.750677][    C0] RDX: 0000000000000010 RSI: 00007ffff9686d80
>>> RDI: 00007ffff9686f20
>>> [  571.750683][    C0] RBP: 0000000000000000 R08: 0000000000000010
>>> R09: 00007ffff9686d90
>>> [  571.750689][    C0] R10: 00007ffff9686fb0 R11: 00007fb1d8d6a060
>>> R12: 00007ffff9686f30
>>> [  571.750696][    C0] R13: 00007fb1d9d20ee0 R14: 00007ffff9686f30
>>> R15: 00007ffff9686d90
>>> [  571.750703][    C0] FS:  00007fb1da33d8c0 GS:  0000000000000000
>>> [  571.752358][   T30] Kernel panic - not syncing: hung_task: blocked tasks
>>> [  571.757337][   T30] CPU: 1 PID: 30 Comm: khungtaskd Not tainted 6.4.0-rc2 #1
>>> [  571.757686][   T30] Hardware name: QEMU Standard PC (i440FX + PIIX,
>>> 1996), BIOS 1.12.0-1 04/01/2014
>>> [  571.758131][   T30] Call Trace:
>>> [  571.758302][   T30]  <TASK>
>>> [  571.758462][   T30]  dump_stack_lvl+0x91/0xf0
>>> [  571.758714][   T30]  panic+0x62d/0x6a0
>>> [  571.758926][   T30]  ? panic_smp_self_stop+0x90/0x90
>>> [  571.759188][   T30]  ? preempt_schedule_common+0x1a/0xc0
>>> [  571.759486][   T30]  ? preempt_schedule_thunk+0x1a/0x20
>>> [  571.759785][   T30]  ? watchdog+0xc21/0x10c0
>>> [  571.760020][   T30]  watchdog+0xc32/0x10c0
>>> [  571.760240][   T30]  ? proc_dohung_task_timeout_secs+0x90/0x90
>>> [  571.760541][   T30]  kthread+0x33b/0x430
>>> [  571.760753][   T30]  ? kthread_complete_and_exit+0x40/0x40
>>> [  571.761052][   T30]  ret_from_fork+0x1f/0x30
>>> [  571.761286][   T30]  </TASK>
>>> [  571.761814][   T30] Kernel Offset: disabled
>>> [  571.762047][   T30] Rebooting in 86400 seconds..
>>>
>>>> You need to include poc_blkdev.c as part of your report.
>>>
>>> It's a little confusing and I'm sorry for that.
>>> The poc_blkdev.c is exactly the C reproducer
>>> (https://pastebin.com/raw/6mg7uF8W).
>>>
>>>> I suspect you've done something that is known to not work (as root,
>>>> so we won't necessarily care).  But I can't really say without seeing
>>>> what you've done.  Running syzkaller is an art, and most people aren't
>>>> good at it.  It takes a lot of work to submit good quality bug reports,
>>>> see this article:
>>>>
>>>> https://blog.regehr.org/archives/2037
>>>
>>> I have read this article and thanks for your recommendations.
>>> I'm not familiar with this module and I haven't figured out the root
>>> cause of this bug yet.
>>>
>>> Regards,
>>>
>>> Yang
>>>
>>> Matthew Wilcox <willy@...radead.org> 于2023年5月17日周三 20:20写道:
>>>>
>>>> On Wed, May 17, 2023 at 07:12:23PM +0800, yang lan wrote:
>>>>> root@...kaller:~# uname -a
>>>>> Linux syzkaller 5.10.179 #1 SMP PREEMPT Thu Apr 27 16:22:48 CST 2023
>>>>
>>>> Does this reproduce on current kernels, eg 6.4-rc2?
>>>>
>>>>> root@...kaller:~# gcc poc_blkdev.c -o poc_blkdev
>>>>
>>>> You need to include poc_blkdev.c as part of your report.
>>>>
>>>>> Please let me know if I can provide any more information, and I hope I
>>>>> didn't mess up this bug report.
>>>>
>>>> I suspect you've done something that is known to not work (as root,
>>>> so we won't necessarily care).  But I can't really say without seeing
>>>> what you've done.  Running syzkaller is an art, and most people aren't
>>>> good at it.  It takes a lot of work to submit good quality bug reports,
>>>> see this article:
>>>>
>>>> https://blog.regehr.org/archives/2037
>>>
>>> Matthew Wilcox <willy@...radead.org> 于2023年5月17日周三 20:20写道:
>>>>
>>>> On Wed, May 17, 2023 at 07:12:23PM +0800, yang lan wrote:
>>>>> root@...kaller:~# uname -a
>>>>> Linux syzkaller 5.10.179 #1 SMP PREEMPT Thu Apr 27 16:22:48 CST 2023
>>>>
>>>> Does this reproduce on current kernels, eg 6.4-rc2?
>>>>
>>>>> root@...kaller:~# gcc poc_blkdev.c -o poc_blkdev
>>>>
>>>> You need to include poc_blkdev.c as part of your report.
>>>>
>>>>> Please let me know if I can provide any more information, and I hope I
>>>>> didn't mess up this bug report.
>>>>
>>>> I suspect you've done something that is known to not work (as root,
>>>> so we won't necessarily care).  But I can't really say without seeing
>>>> what you've done.  Running syzkaller is an art, and most people aren't
>>>> good at it.  It takes a lot of work to submit good quality bug reports,
>>>> see this article:
>>>>
>>>> https://blog.regehr.org/archives/2037
>>> .
>>>
>>
> .
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ