[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1189171370.4234.38.camel@localhost>
Date: Fri, 07 Sep 2007 09:22:50 -0400
From: jamal <hadi@...erus.ca>
To: James Chapman <jchapman@...alix.com>
Cc: netdev@...r.kernel.org, davem@...emloft.net, jeff@...zik.org,
mandeep.baines@...il.com, ossthema@...ibm.com,
Stephen Hemminger <shemminger@...l.org>
Subject: Re: RFC: possible NAPI improvements to reduce interrupt rates for
low traffic rates
On Fri, 2007-07-09 at 10:31 +0100, James Chapman wrote:
> Not really. I used 3-year-old, single CPU x86 boxes with e100
> interfaces.
> The idle poll change keeps them in polled mode. Without idle
> poll, I get twice as many interrupts as packets, one for txdone and one
> for rx. NAPI is continuously scheduled in/out.
Certainly faster than the machine in the paper (which was about 2 years
old in 2005).
I could never get ping -f to do that for me - so things must be getting
worse with newer machines then.
> No. Since I did a flood ping from the machine under test, the improved
> latency meant that the ping response was handled more quickly, causing
> the next packet to be sent sooner. So more packets were transmitted in
> the allotted time (10 seconds).
ok.
> With current NAPI:
> rtt min/avg/max/mdev = 0.902/1.843/101.727/4.659 ms, pipe 9, ipg/ewma
> 1.611/1.421 ms
>
> With idle poll changes:
> rtt min/avg/max/mdev = 0.898/1.117/28.371/0.689 ms, pipe 3, ipg/ewma
> 1.175/1.236 ms
Not bad in terms of latency. The deviation certainly looks better.
> But the CPU has done more work.
I am going to be the devil's advocate[1]:
If the problem i am trying to solve is "reduce cpu use at lower rate",
then this is not the right answer because your cpu use has gone up.
Your latency numbers have not improved that much (looking at the avg)
and your throughput is not that much higher. Will i be willing to pay
more cpu (of an already piggish cpu use by NAPI at that rate with 2
interupts per packet)?
Another test: try a simple ping and compare the rtts.
> The problem I started thinking about was the one where NAPI thrashes
> in/out of polled mode at higher and higher rates as network interface
> speeds and CPU speeds increase. A flood ping demonstrates this even on
> 100M links on my boxes.
things must be getting worse in the state of average hardware out there.
It will be worthwile exercise to compare on an even faster machine
and see what transpires there.
> Networking boxes want consistent
> performance/latency for all traffic patterns and they need to avoid
> interrupt livelock. Current practice seems to be to use hardware
> interrupt mitigation or timers to limit interrupt rate but this just
> hurts latency, as you noted. So I'm trying to find a way to limit the
> NAPI interrupt rate without increasing latency. My comment about this
> approach being suitable for routers and networked servers is that these
> boxes care more about minimizing packet latency than they do about
> wasting CPU cycles by polling idle devices.
I think the arguement of "who cares about a little more cpu" is valid
for the case of routers. It is a double edged sword, because it applies
to the case of "who cares if NAPI uses a little more cpu at low rates"
and "who cares if James turns on polling and abuses a little more-more
cpu". Since NAPI is the incumbent, the onus(sp?) is to do better. You
must do better sir!
Look at the timers, she said - that way you may be able to cut the cpu
abuse.
cheers,
jamal
[1] historically the devils advocate was a farce really ;->
-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists