lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAOuPNLhmZ5mT2qY_iD9g5v3taSVuZ6OnReiW1usFhEH8mkw1Kw@mail.gmail.com>
Date: Wed, 3 Jul 2024 19:40:03 +0530
From: Pintu Agarwal <pintu.ping@...il.com>
To: dm-devel@...hat.com, open list <linux-kernel@...r.kernel.org>, snitzer@...hat.com, 
	agk@...hat.com
Subject: Re: device-mapper: verity: 251:0: metadata block 18398 is corrupted

Hi,
Any comments on the issue below ?

On Sat, 29 Jun 2024 at 00:16, Pintu Agarwal <pintu.ping@...il.com> wrote:
>
> Hi,
> In one of our NAND products (arm64) we are having Kernel 5.15 along
> with squashfs + ubiblock + dm-verity on rootfs along with ramdisk.
>
> Recently we have enabled "NAND Page Cache" in our nand driver to
> improve nand read performance.
> But after enabling this feature, the dm-verity detects metadata
> corruption during boot-up and the device is crashing.
> If we disable dm-verity (from ramdisk) then we don't see this problem
> and the device boots fine.
>
> Are there any dm-verity specific changes (in higher Kernel version)
> that we are missing here ?
>
> This is the failure log.
> What does the below metadata corruption indicate ?
> Does this mean, the root_hash is corrupted ?
>
> {{{
> [    7.731295][  T136] device-mapper: verity: 251:0: metadata block
> 18398 is corrupted
> [    7.742723][  T136] ------------[ cut here ]------------
> [    7.748206][  T136] workqueue: WQ_MEM_RECLAIM kverityd:verity_work
> is flushing !WQ_MEM_RECLAIM k_sm_usb:0x0
> [    7.754840][  T136] WARNING: CPU: 3 PID: 136 at
> kernel/workqueue.c:2660 check_flush_dependency+0x12c/0x134
> [...]
> [    7.809215][  T136] pc : check_flush_dependency+0x12c/0x134
> [    7.814933][  T136] lr : check_flush_dependency+0x12c/0x134
> [...]
> [    7.905120][  T136] Call trace:
> [    7.908345][  T136]  check_flush_dependency+0x12c/0x134
> [    7.913710][  T136]  flush_workqueue+0x1cc/0x4dc
> [    7.918452][  T136]  dwc3_msm_shutdown+0x48/0x58
> [    7.923195][  T136]  platform_shutdown+0x24/0x30
> [    7.927937][  T136]  device_shutdown+0x170/0x220
> [    7.932680][  T136]  kernel_restart+0x40/0xfc
> [    7.937152][  T136]  verity_handle_err+0x11c/0x1b0
> [    7.942071][  T136]  verity_hash_for_block+0x260/0x2d8
> [    7.947343][  T136]  verity_verify_io+0xe8/0x568
> [    7.952085][  T136]  verity_work+0x24/0x74
> [    7.956297][  T136]  process_one_work+0x1a8/0x3a0
> [    7.961133][  T136]  worker_thread+0x22c/0x490
> [    7.965700][  T136]  kthread+0x154/0x218
> [    7.969725][  T136]  ret_from_fork+0x10/0x20
> [    7.974116][  T136] ---[ end trace 166e4069e91d0a01 ]---
> }}}

Note, this issue occurs only some times during boot (1/6 reboots).
We applied all patches from 6.9 Kernel and still the issue exists.
Is it possible that the root_hash got corrupted after the image loaded
into RAM ?

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ