lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 02 Dec 2019 11:53:22 +0100 From: Paolo Abeni <pabeni@...hat.com> To: Paweł Staszewski <pstaszewski@...are.pl>, David Ahern <dsahern@...il.com>, netdev@...r.kernel.org, Jesper Dangaard Brouer <brouer@...hat.com> Subject: Re: Linux kernel - 5.4.0+ (net-next from 27.11.2019) routing/network performance On Mon, 2019-12-02 at 11:09 +0100, Paweł Staszewski wrote: > W dniu 01.12.2019 o 17:05, David Ahern pisze: > > On 11/29/19 4:00 PM, Paweł Staszewski wrote: > > > As always - each year i need to summarize network performance for > > > routing applications like linux router on native Linux kernel (without > > > xdp/dpdk/vpp etc) :) > > > > > Do you keep past profiles? How does this profile (and traffic rates) > > compare to older kernels - e.g., 5.0 or 4.19? > > > > > Yes - so for 4.19: > > Max bandwidth was about 40-42Gbit/s RX / 40-42Gbit/s TX of > forwarded(routed) traffic > > And after "order-0 pages" patches - max was 50Gbit/s RX + 50Gbit/s TX > (forwarding - bandwidth max) > > (current kernel almost doubled this) Looks like we are on the good track ;) [...] > After "order-0 pages" patch > > PerfTop: 104692 irqs/sec kernel:99.5% exact: 0.0% [4000Hz > cycles], (all, 56 CPUs) > --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- > > > 9.06% [kernel] [k] mlx5e_skb_from_cqe_mpwrq_linear > 6.43% [kernel] [k] tasklet_action_common.isra.21 > 5.68% [kernel] [k] fib_table_lookup > 4.89% [kernel] [k] irq_entries_start > 4.53% [kernel] [k] mlx5_eq_int > 4.10% [kernel] [k] build_skb > 3.39% [kernel] [k] mlx5e_poll_tx_cq > 3.38% [kernel] [k] mlx5e_sq_xmit > 2.73% [kernel] [k] mlx5e_poll_rx_cq Compared to the current kernel perf figures, it looks like most of the gains come from driver changes. [... current perf figures follow ...] > --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- > > > 7.56% [kernel] [k] __dev_queue_xmit This is a bit surprising to me. I guess this is due '__dev_queue_xmit()' being calling twice per packet (team, NIC) and due to the retpoline overhead. > 1.74% [kernel] [k] tcp_gro_receive If the reference use-case is with a quite large number of cuncurrent flows, I guess you can try disabling GRO Cheers, Paolo
Powered by blists - more mailing lists