[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20070827.140251.95055210.davem@davemloft.net>
Date: Mon, 27 Aug 2007 14:02:51 -0700 (PDT)
From: David Miller <davem@...emloft.net>
To: jchapman@...alix.com
Cc: shemminger@...ux-foundation.org, ossthema@...ibm.com,
akepner@....com, netdev@...r.kernel.org, raisch@...ibm.com,
themann@...ibm.com, linux-kernel@...r.kernel.org,
linuxppc-dev@...abs.org, meder@...ibm.com, tklein@...ibm.com,
stefan.roscher@...ibm.com
Subject: Re: RFC: issues concerning the next NAPI interface
From: James Chapman <jchapman@...alix.com>
Date: Mon, 27 Aug 2007 16:51:29 +0100
> To implement this, there's no need for timers, hrtimers or generic NAPI
> support that others have suggested. A driver's poll() would set an
> internal flag and record the current jiffies value when finding
> workdone=0 rather than doing an immediate napi_complete(). Early in
> poll() it would test this flag and if set, do a low-cost test to see if
> it had any work to do. If no work, it would check the saved jiffies
> value and do the napi_complete() only if no work has been done for a
> configurable number of jiffies. This keeps interrupts disabled longer at
> the expense of many more calls to poll() where no work is done. So
> critical to this scheme is modifying the driver's poll() to fastpath the
> case of having no work to do while waiting for its local jiffy count to
> expire.
>
> Here's an untested patch for tg3 that illustrates the idea.
It's only going to work with hrtimers, these interfaces can
process at least 100,000 per jiffies tick.
And the hrtimer granularity is going to need to be significantly low,
and futhermore you're adding a guaranteed extra interrupt (for the
hrtimer firing) in these cases where we're exactly trying to avoid is
more interrupts.
If you can make it work, fine, but it's going to need to be at a
minimum disabled when the hrtimer granularity is not sufficient.
But there are huger fish to fry for you I think. Talk to your
platform maintainers and ask for an interface for obtaining
a flat static distribution of interrupts to cpus in order to
support multiqueue NAPI better.
In your previous postings you made arguments saying that the
automatic placement of interrupts to cpus made everything
bunch of to a single cpu and you wanted to propagate the
NAPI work to other cpu's software interrupts from there.
That logic is bogus, because it merely proves that the hardware
interrupt distribution is broken. If it's a bad cpu to run
software interrupts on, it's also a bad cpu to run hardware
interrupts on.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists