[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAKgT0Uc2OdABa3LVkOVY=jLRhnqOou1-7YW0hE2YRwphv8bVJA@mail.gmail.com>
Date: Thu, 26 Jul 2012 11:06:45 -0700
From: Alexander Duyck <alexander.duyck@...il.com>
To: Eric Dumazet <eric.dumazet@...il.com>
Cc: David Miller <davem@...emloft.net>, netdev@...r.kernel.org
Subject: Re: [PATCH 00/16] Remove the ipv4 routing cache
On Thu, Jul 26, 2012 at 10:43 AM, Eric Dumazet <eric.dumazet@...il.com> wrote:
> On Thu, 2012-07-26 at 19:36 +0200, Eric Dumazet wrote:
>> On Thu, 2012-07-26 at 19:31 +0200, Eric Dumazet wrote:
>> > On Thu, 2012-07-26 at 10:18 -0700, Alexander Duyck wrote:
>> >
>> > > I tested this patch and it looks like it runs, but still has the same
>> > > performance issue. I did some digging into the annotation for
>> > > ip_route_intput_noref and it seems like the issue is that I am hitting
>> > > the dst_hold call in __mkroute_input.
>> >
>> > David suggested a percpu cache.
>> >
>> > nh_rth_input would be allocated by alloc_percpu(struct dst *)
>> >
>> > I can work on this.
>>
>> Wait a minute, on input we should use the noref trick too.
>>
>
> Something like : (on top of latest David patch)
>
> diff --git a/net/ipv4/route.c b/net/ipv4/route.c
> index 7a591aa..d5d2ad1 100644
> --- a/net/ipv4/route.c
> +++ b/net/ipv4/route.c
> @@ -1371,8 +1371,7 @@ static void ip_handle_martian_source(struct net_device *dev,
> static int __mkroute_input(struct sk_buff *skb,
> const struct fib_result *res,
> struct in_device *in_dev,
> - __be32 daddr, __be32 saddr, u32 tos,
> - struct rtable **result)
> + __be32 daddr, __be32 saddr, u32 tos)
> {
> struct rtable *rth;
> int err;
> @@ -1423,7 +1422,7 @@ static int __mkroute_input(struct sk_buff *skb,
> if (!itag) {
> rth = FIB_RES_NH(*res).nh_rth_input;
> if (rt_cache_valid(rth)) {
> - dst_hold(&rth->dst);
> + skb_dst_set_noref(skb, &rth->dst);
> goto out;
> }
> do_cache = true;
> @@ -1451,7 +1450,6 @@ static int __mkroute_input(struct sk_buff *skb,
>
> rt_set_nexthop(rth, daddr, res, NULL, res->fi, res->type, itag);
> out:
> - *result = rth;
> err = 0;
> cleanup:
> return err;
> @@ -1463,21 +1461,13 @@ static int ip_mkroute_input(struct sk_buff *skb,
> struct in_device *in_dev,
> __be32 daddr, __be32 saddr, u32 tos)
> {
> - struct rtable *rth = NULL;
> - int err;
> -
> #ifdef CONFIG_IP_ROUTE_MULTIPATH
> if (res->fi && res->fi->fib_nhs > 1)
> fib_select_multipath(res);
> #endif
>
> /* create a routing cache entry */
> - err = __mkroute_input(skb, res, in_dev, daddr, saddr, tos, &rth);
> - if (err)
> - return err;
> -
> - skb_dst_set(skb, &rth->dst);
> - return 0;
> + return __mkroute_input(skb, res, in_dev, daddr, saddr, tos);
> }
>
> /*
>
With your changes in place I see an increase from 7.5Mpps to 9.9Mpps
for 8 queues, and increasing the queues to 9 gets me up to 11Mpps even
if the 9th queue is on another node. This is a HUGE improvement over
what we had before.
The only remaining overhead that has been introduced with the recent
changes appears to be the fib_table_lookup which doesn't have any hot
spots that jump out at me. The performance is in-line with what I was
seeing when I was randomly generating source IPs from a fairly large
set so I suspect this is just the expected behaviour without a routing
cache in place.
Thanks,
Alex
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists