lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4c9c1833-0f5e-4978-8204-9195009edb33@gmail.com>
Date: Sat, 1 Mar 2025 15:14:01 +0100
From: Heiner Kallweit <hkallweit1@...il.com>
To: Rui Salvaterra <rsalvaterra@...il.com>
Cc: nic_swsd@...ltek.com, netdev@...r.kernel.org
Subject: Re: [PATCH] r8169: add support for 16K jumbo frames on RTL8125B

On 01.03.2025 12:45, Rui Salvaterra wrote:
> Hi, Heiner,
> 
> On Fri, 28 Feb 2025 at 20:22, Heiner Kallweit <hkallweit1@...il.com> wrote:
>>
>> This has been proposed and discussed before. Decision was to not increase
>> the max jumbo packet size, as vendor drivers r8125/r8126 also support max 9k.
> 
> I did a cursory search around the mailing list, but didn't find
> anything specific. Maybe I didn't look hard enough. However…
> 
>> And in general it's not clear whether you would gain anything from jumbo packets,
>> because hw TSO and c'summing aren't supported for jumbo packets.
> 
> … I actually have numbers to justify it. For my use case, jumbo frames
> make a *huge* difference. I have an Atom 330-based file server, this
> CPU is too slow to saturate the link with a MTU of 1500 bytes. The
> situation, however, changes dramatically when I use jumbo frames. Case
> in point…
> 
> 
> MTU = 1500 bytes:
> 
> Accepted connection from 192.168.17.20, port 55514
> [  5] local 192.168.17.16 port 5201 connected to 192.168.17.20 port 55524
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-1.00   sec   241 MBytes  2.02 Gbits/sec
> [  5]   1.00-2.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   2.00-3.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   3.00-4.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   4.00-5.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   5.00-6.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   6.00-7.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   7.00-8.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   8.00-9.00   sec   242 MBytes  2.03 Gbits/sec
> [  5]   9.00-10.00  sec   242 MBytes  2.03 Gbits/sec
> [  5]  10.00-10.00  sec   128 KBytes  1.27 Gbits/sec
> - - - - - - - - - - - - - - - - - - - - - - - - -
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-10.00  sec  2.36 GBytes  2.03 Gbits/sec                  receiver
> 
Depending on the kernel version HW TSO may be be off per default.
Use ethtool to check/enable HW TSO, and see whether speed improves.

> 
> MTU = 9000 bytes:
> 
> Accepted connection from 192.168.17.20, port 53474
> [  5] local 192.168.17.16 port 5201 connected to 192.168.17.20 port 53490
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-1.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   1.00-2.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   2.00-3.00   sec   294 MBytes  2.47 Gbits/sec
> [  5]   3.00-4.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   4.00-5.00   sec   294 MBytes  2.47 Gbits/sec
> [  5]   5.00-6.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   6.00-7.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   7.00-8.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   8.00-9.00   sec   295 MBytes  2.47 Gbits/sec
> [  5]   9.00-10.00  sec   295 MBytes  2.47 Gbits/sec
> [  5]  10.00-10.00  sec   384 KBytes  2.38 Gbits/sec
> - - - - - - - - - - - - - - - - - - - - - - - - -
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-10.00  sec  2.88 GBytes  2.47 Gbits/sec                  receiver
> 
> 
> MTU = 12000 bytes (with my patch):
> 
> Accepted connection from 192.168.17.20, port 59378
> [  5] local 192.168.17.16 port 5201 connected to 192.168.17.20 port 59388
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-1.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   1.00-2.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   2.00-3.00   sec   295 MBytes  2.48 Gbits/sec
> [  5]   3.00-4.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   4.00-5.00   sec   295 MBytes  2.48 Gbits/sec
> [  5]   5.00-6.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   6.00-7.00   sec   295 MBytes  2.48 Gbits/sec
> [  5]   7.00-8.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   8.00-9.00   sec   296 MBytes  2.48 Gbits/sec
> [  5]   9.00-10.00  sec   294 MBytes  2.47 Gbits/sec
> [  5]  10.00-10.00  sec   512 KBytes  2.49 Gbits/sec
> - - - - - - - - - - - - - - - - - - - - - - - - -
> [ ID] Interval           Transfer     Bitrate
> [  5]   0.00-10.00  sec  2.89 GBytes  2.48 Gbits/sec                  receiver
> 
> 
> This demonstrates that the bottleneck is in the frame processing. With
> a larger frame size, the number of checksum calculations is also
> lower, for the same amount of payload data, and the CPU is able to
> handle them.
> 
> 
> Kind regards,
> 
> Rui Salvaterra


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ