[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1354290335.11754.447.camel@localhost>
Date: Fri, 30 Nov 2012 16:45:35 +0100
From: Jesper Dangaard Brouer <brouer@...hat.com>
To: Eric Dumazet <eric.dumazet@...il.com>
Cc: David Miller <davem@...emloft.net>, fw@...len.de,
netdev@...r.kernel.org, pablo@...filter.org, tgraf@...g.ch,
amwang@...hat.com, kaber@...sh.net, paulmck@...ux.vnet.ibm.com,
herbert@...dor.hengli.com.au
Subject: Re: [net-next PATCH V2 1/9] net: frag evictor, avoid killing warm
frag queues
On Fri, 2012-11-30 at 06:52 -0800, Eric Dumazet wrote:
> On Fri, 2012-11-30 at 11:04 +0100, Jesper Dangaard Brouer wrote:
> > So, let me instead show, with tests, that the evictor strategy is
> > broken, while keeping the original default thresh settings:
> >
> > # grep . /proc/sys/net/ipv4/ipfrag_*_thresh
> > /proc/sys/net/ipv4/ipfrag_high_thresh:262144
> > /proc/sys/net/ipv4/ipfrag_low_thresh:196608
> >
> > Test purpose, I will on a single 10G link demonstrate, that starting
> > several "N" netperf UDP fragmentation flows, will hurt performance, and
> > then claim this is caused by the bad evictor strategy.
> >
> > Test setup:
> > - Disable Ethernet flow control
> > - netperf packet size 65507
> > - Run netserver on one NUMA node
> > - Start netperf clients against a NIC on the other NUMA node
> > - (The NUMA imbalance helps the effect occur at lower N)
> >
> > Result: N=1 8040 Mbit/s
> > Result: N=2 9584 Mbit/s (4739+4845)
> > Result: N=3 4055 Mbit/s (1436+1371+1248)
> > Result: N=4 2247 Mbit/s (1538+29+54+626)
> > Result: N=5 879 Mbit/s (78+152+226+125+298)
> > Result: N=6 293 Mbit/s (85+55+32+57+46+18)
> > Result: N=7 354 Mbit/s (70+47+33+80+20+72+32)
> >
> > Can we, now, agree that the current evictor strategy is broken?!?
>
> Your setup is broken for sure.
No, its not.
> I dont know how you expect that many
> datagrams being correctly reassembled with ipfrag_high_thresh=262144
That's my point... I'm showing that its not possible, with out current
implementation!
> No matter strategy is implemented, an attacker knows it and can send
> frags so that regular workload is denied. Kernel cant decide which
> packets are more likely to be completed.
Our current evictor implementation will allow the attacker to kill ALL
frag traffic to the machine (and cause high CPU load, with CPUs
spinning).
My implementation will guarantee that some fragments will be, allowed to
finish and complete. A far better choice, than our current situation.
> BTW, install fq_codel at the sender side, so that frags are nicely
> interleaved. Because on real networks, frags of an UDP datagram rarely
> come to destination in a single train with no alien packets inside the
> train.
You are arguing in my favor. I have taken great care that my test will
interleave UDP datagrams (by CPU pinning netperf clients on the sender
side). If I just naively start all netperf client on one CPU on the
sender, then I will have packet trains... and the result will be:
Packet trains result N=5 8884 Mbit/s (1775+1775+1790+1789+1755)
That test would be "broken for sure".
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists