[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <Z6LYjHJxx0pI45WU@LQ3V64L9R2>
Date: Tue, 4 Feb 2025 19:18:36 -0800
From: Joe Damato <jdamato@...tly.com>
To: Samiullah Khawaja <skhawaja@...gle.com>
Cc: Jakub Kicinski <kuba@...nel.org>,
"David S . Miller " <davem@...emloft.net>,
Eric Dumazet <edumazet@...gle.com>, Paolo Abeni <pabeni@...hat.com>,
almasrymina@...gle.com, netdev@...r.kernel.org
Subject: Re: [PATCH net-next v3 0/4] Add support to do threaded napi busy poll
On Wed, Feb 05, 2025 at 12:10:48AM +0000, Samiullah Khawaja wrote:
> Extend the already existing support of threaded napi poll to do continuous
> busy polling.
[...]
Overall, +1 to everything Martin said in his response. I think I'd
like to try to reproduce this myself to better understand the stated
numbers below.
IMHO: the cover letter needs more details.
>
> Setup:
>
> - Running on Google C3 VMs with idpf driver with following configurations.
> - IRQ affinity and coalascing is common for both experiments.
As Martin suggested, a lot more detail here would be helpful.
> - There is only 1 RX/TX queue configured.
> - First experiment enables busy poll using sysctl for both epoll and
> socket APIs.
> - Second experiment enables NAPI threaded busy poll for the full device
> using sysctl.
>
> Non threaded NAPI busy poll enabled using sysctl.
> ```
> echo 400 | sudo tee /proc/sys/net/core/busy_poll
> echo 400 | sudo tee /proc/sys/net/core/busy_read
I'm not sure why busy_read is enabled here?
Maybe more details on how exactly the internals of onload+neper work
would explain it, but I presume it's an epoll_wait loop with
non-blocking reads so busy_read wouldn't do anything?
> echo 2 | sudo tee /sys/class/net/eth0/napi_defer_hard_irqs
> echo 15000 | sudo tee /sys/class/net/eth0/gro_flush_timeout
> ```
The deferral amounts above are relatively small, which makes me
wonder if you are seeing IRQ and softIRQ interference in the base
case?
I ask because it seems like in the test case (if I read the patch
correctly) the processing of packets happens when BH is disabled.
Did I get that right?
If so, then:
- In the base case, IRQs can be generated and softirq can interfere
with packet processing.
- In the test case, packet processing happens but BH is disabled,
reducing interference.
If I got that right, it sounds like IRQ suspension would show good
results in this case, too, and it's probably worth comparing IRQ
suspension in the onload+neper setup.
It seems like it shouldn't be too difficult to get onload+neper
using it and the data would be very enlightening.
Powered by blists - more mailing lists