[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110504144622.GA15823@redhat.com>
Date: Wed, 4 May 2011 17:46:22 +0300
From: "Michael S. Tsirkin" <mst@...hat.com>
To: Krishna Kumar <krkumar2@...ibm.com>
Cc: davem@...emloft.net, eric.dumazet@...il.com, kvm@...r.kernel.org,
netdev@...r.kernel.org, rusty@...tcorp.com.au
Subject: Re: [PATCH 0/4] [RFC] virtio-net: Improve small packet performance
On Wed, May 04, 2011 at 07:32:58PM +0530, Krishna Kumar wrote:
> Earlier approach to improving small packet performance went
> along the lines of dropping packets when the txq is full to
> avoid stop/start of the txq. Though performance improved
> significantly (upto 3x) for a single thread, multiple netperf
> sessions showed a regression of upto -17% (starting from 4
> sessions).
>
> This patch proposes a different approach with the following
> changes:
>
> A. virtio:
> - Provide a API to get available number of slots.
>
> B. virtio-net:
> - Remove stop/start txq's and associated callback.
> - Pre-calculate the number of slots needed to transmit
> the skb in xmit_skb and bail out early if enough space
> is not available. My testing shows that 2.5-3% of
> packets are benefited by using this API.
> - Do not drop skbs but instead return TX_BUSY like other
> drivers.
> - When returning EBUSY, set a per-txq variable to indicate
> to dev_queue_xmit() whether to restart xmits on this txq.
>
> C. net/sched/sch_generic.c:
> Since virtio-net now returns EBUSY, the skb is requeued to
> gso_skb. This allows adding the addional check for restart
> xmits in just the slow-path (the first re-queued packet
> case of dequeue_skb, where it checks for gso_skb) before
> deciding whether to call the driver or not.
>
> Patch was also tested between two servers with Emulex OneConnect
> 10G cards to confirm there is no regression. Though the patch is
> an attempt to improve only small packet performance, there was
> improvement for 1K, 2K and also 16K both in BW and SD. Results
> from Guest -> Remote Host (BW in Mbps) for 1K and 16K I/O sizes:
>
> ________________________________________________________
> I/O Size: 1K
> # BW1 BW2 (%) SD1 SD2 (%)
> ________________________________________________________
> 1 1226 3313 (170.2) 6.6 1.9 (-71.2)
> 2 3223 7705 (139.0) 18.0 7.1 (-60.5)
> 4 7223 8716 (20.6) 36.5 29.7 (-18.6)
> 8 8689 8693 (0) 131.5 123.0 (-6.4)
> 16 8059 8285 (2.8) 578.3 506.2 (-12.4)
> 32 7758 7955 (2.5) 2281.4 2244.2 (-1.6)
> 64 7503 7895 (5.2) 9734.0 9424.4 (-3.1)
> 96 7496 7751 (3.4) 21980.9 20169.3 (-8.2)
> 128 7389 7741 (4.7) 40467.5 34995.5 (-13.5)
> ________________________________________________________
> Summary: BW: 16.2% SD: -10.2%
>
> ________________________________________________________
> I/O Size: 16K
> # BW1 BW2 (%) SD1 SD2 (%)
> ________________________________________________________
> 1 6684 7019 (5.0) 1.1 1.1 (0)
> 2 7674 7196 (-6.2) 5.0 4.8 (-4.0)
> 4 7358 8032 (9.1) 21.3 20.4 (-4.2)
> 8 7393 8015 (8.4) 82.7 82.0 (-.8)
> 16 7958 8366 (5.1) 283.2 310.7 (9.7)
> 32 7792 8113 (4.1) 1257.5 1363.0 (8.3)
> 64 7673 8040 (4.7) 5723.1 5812.4 (1.5)
> 96 7462 7883 (5.6) 12731.8 12119.8 (-4.8)
> 128 7338 7800 (6.2) 21331.7 21094.7 (-1.1)
> ________________________________________________________
> Summary: BW: 4.6% SD: -1.5%
>
> Signed-off-by: Krishna Kumar <krkumar2@...ibm.com>
> ---
So IIUC, we delay transmit by an arbitrary value and hope
that the host is done with the packets by then?
Interesting.
I am currently testing an approach where
we tell the host explicitly to interrupt us only after
a large part of the queue is empty.
With 256 entries in a queue, we should get 1 interrupt per
on the order of 100 packets which does not seem like a lot.
I can post it, mind testing this?
--
MST
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists