lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <a4bcf251-e1cf-47ef-84df-5c43b2b288c0@gmail.com>
Date: Tue, 21 Oct 2025 16:12:18 +0900
From: Jeuk Kim <jeuk20.kim@...il.com>
To: Chao Yu <chao@...nel.org>, jaegeuk@...nel.org
Cc: linux-f2fs-devel@...ts.sourceforge.net, linux-kernel@...r.kernel.org,
 jeuk20.kim@...sung.com, d_hyun.kwon@...sung.com, gyusun.lee@...sung.com,
 hyenc.jeong@...sung.com, j-young.choi@...sung.com, jaemyung.lee@...sung.com,
 jieon.seol@...sung.com, keosung.park@...sung.com, wone.jung@...sung.com
Subject: Re: [PATCH] f2fs: serialize writeback for inline-crypto inodes


On 10/21/2025 3:51 PM, Chao Yu wrote:
> On 10/21/25 11:33, Jeuk Kim wrote:
>> On 10/16/2025 7:12 PM, Chao Yu wrote:
>>> On 10/16/2025 1:16 PM, Jeuk Kim wrote:
>>>> From: Jeuk Kim <jeuk20.kim@...sung.com>
>>>>
>>>> Inline encryption derives DUN from <inode, file offset>,
>>>> so bios from different inodes can't merge. With multi-threaded
>>>> buffered O_SYNC writes where each thread writes to its own file,
>>>> 4KiB-per-page LBA allocation interleaves across inodes and
>>>> causes bio split. Serialize writeback for fscrypt inline-crypto
>>>> inodes via __should_serialize_io() to keep foreground writeback
>>>> focused on one inode and avoid split.
>>>>
>>>> Test: fio --name=wb_osync --rw=write --bs=1M \
>>>>         --time_based=1 --runtime=60s --size=2G \
>>>>         --ioengine=psync --direct=0 --sync=1 \
>>>>         --numjobs=8 --thread=1 --nrfiles=1 \
>>>>         --filename_format='wb_osync.$jobnum'
>>>>
>>>> device: UFS
>>>>
>>>> Before -
>>>>     write throughput: 675MiB/s
>>>>     device I/O size distribution (by count, total 1027414):
>>>>       4 KiB:  923139 (89.9%)
>>>>       8 KiB:  84798 (8.3%)
>>>>       ≥512 KiB: 453 (0.0%)
>>>>
>>>> After -
>>>>     write throughput: 1760MiB/s
>>>>     device I/O size distribution (by count, total 231750):
>>>>       4 KiB:  16904 (7.3%)
>>>>       8 KiB:  72128 (31.1%)
>>>>       ≥512 KiB: 118900 (51.3%)
>>>>
>>>> Signed-off-by: Jeuk Kim <jeuk20.kim@...sung.com>
>>>> ---
>>>>    fs/f2fs/data.c | 2 ++
>>>>    1 file changed, 2 insertions(+)
>>>>
>>>> diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c
>>>> index ef38e62cda8f..ae6fb435d576 100644
>>>> --- a/fs/f2fs/data.c
>>>> +++ b/fs/f2fs/data.c
>>>> @@ -3217,6 +3217,8 @@ static inline bool __should_serialize_io(struct inode *inode,
>>>>          if (f2fs_need_compress_data(inode))
>>>>            return true;
>>>> +    if (fscrypt_inode_uses_inline_crypto(inode))
>>>> +        return true;
>>>>        if (wbc->sync_mode != WB_SYNC_ALL)
>>>>            return true;
>>>>        if (get_dirty_pages(inode) >= SM_I(F2FS_I_SB(inode))->min_seq_blocks)
>>> Jeuk,
>>>
>>> Can you please try tuning /sys/fs/f2fs/<dev>/min_seq_blocks to see whether it
>>> can achive the goal?
>>>
>>> Thanks,
>>>
>> Hi Chao,
>>
>> Thanks a lot for the suggestion.
>> I tried tuning `/sys/fs/f2fs/<dev>/min_seq_blocks` as you mentioned, and it also achieved similar performance improvement on my setup.
>>
>> Your approach looks cleaner and better than the one I proposed.
>>
>>  From what I see, even after reducing this value from the default (2MB) to 512 KB on my local system, there doesn’t seem to be any noticeable performance drop or other side effects.
>> Do you see any possible downsides with lowering this value that I might have missed?
> Hi Jeuk,
>
> We're using sbi->writepages to serialize large IOs, once you tuned default
> value from 2MB to 512KB, in Android, there are threads issue [512K, 2M)
> sized IOs, they will join into racing on grabbing the .writepages lock,
> I guess that will cause potential performance regression, right?

That's right, that could happen.

I’ll run some tests to check that, including a few other cases that 
might be affected.
I’ll share the results here if I find anything noticeable.

Thanks for your help!

> Thanks,
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ