lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <844e5cd4-462e-4b88-b3b5-816465a3b7e3@linux.intel.com>
Date: Fri, 17 Oct 2025 15:24:05 +0200
From: Karol Wachowski <karol.wachowski@...ux.intel.com>
To: yi.zhang@...wei.com, tytso@....edu
Cc: linux-ext4@...r.kernel.org, adilger.kernel@...ger.ca
Subject: Possible regression in pin_user_pages_fast() behavior after commit
 7ac67301e82f ("ext4: enable large folio for regular file")

Hi,

I’m not entirely sure if this is right way to report this.

I’ve encountered what appears to be a regression (or at least a
behavioral change) related to pin_user_pages_fast() when used with
FOLL_LONGTERM on a Copy-on-Write (CoW) mapping (i.e. VM_MAYWRITE without
VM_SHARED). Specifically, the call never finishes when the requested
size exceeds 8 MB.

The same scenario works correctly prior to the following change:
commit 7ac67301e82f02b77a5c8e7377a1f414ef108b84
Author: Zhang Yi <yi.zhang@...wei.com>
Date:   Mon May 12 14:33:19 2025 +0800

    ext4: enable large folio for regular file

It seems the issue manifests when pin_user_pages_fast() falls back to
_gup_longterm_locked(). In that case, we end up calling
handle_mm_fault() with FAULT_FLAG_UNSHARE, which splits the PMD. 
>From ftrace, it looks like the kernel enters an apparent infinite loop
of handle_mm_fault() which in turn invokes filemap_map_pages() from the
ext4 ops.

  1)   1.553 us    |      handle_mm_fault();
  1)   0.126 us    |      __cond_resched();
  1)   0.055 us    |      vma_pgtable_walk_begin();
  1)   0.057 us    |      _raw_spin_lock();
  1)   0.111 us    |      _raw_spin_unlock();
  1)   0.050 us    |      vma_pgtable_walk_end();
  1)   1.521 us    |      handle_mm_fault();
  1)   0.122 us    |      __cond_resched();
  1)   0.055 us    |      vma_pgtable_walk_begin();
  1)   0.288 us    |      _raw_spin_lock();
  1)   0.053 us    |      _raw_spin_unlock();
  1)   0.048 us    |      vma_pgtable_walk_end();
  1)   1.484 us    |      handle_mm_fault();
  1)   0.124 us    |      __cond_resched();
  1)   0.056 us    |      vma_pgtable_walk_begin();
  1)   0.272 us    |      _raw_spin_lock();
  1)   0.051 us    |      _raw_spin_unlock();
  1)   0.050 us    |      vma_pgtable_walk_end();
  1)   1.566 us    |      handle_mm_fault();
  1)   0.211 us    |      __cond_resched();
  1)   0.107 us    |      vma_pgtable_walk_begin();
  1)   0.054 us    |      _raw_spin_lock();
  1)   0.052 us    |      _raw_spin_unlock();
  1)   0.049 us    |      vma_pgtable_walk_end();

I haven’t been able to gather more detailed diagnostics yet, but I’d
appreciate any guidance on whether this is a known issue, or if
additional debugging information would be helpful.

-
Karol


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ