[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <4E42F112.4020300@hp.com>
Date: Wed, 10 Aug 2011 13:58:58 -0700
From: Rick Jones <rick.jones2@...com>
To: "J.Hwan Kim" <frog1120@...il.com>
CC: netdev <netdev@...r.kernel.org>, rizzo@....unipi.it
Subject: Re: Intel 82599 ixgbe driver performance
On 08/09/2011 11:19 PM, J.Hwan Kim wrote:
> Hi, everyone
>
> I'm testing our network card which includes intel 82599 based on
> ixgbe driver. I wonder what is the Rx performance of i82599 without
> network stack only with 64Byte frames. Our driver reads the packet
> directly from DMA packet buffer and push to the application without
> passing through linux kernel stack. It seems that the intel 82599
> cannot push 64B frames to DMA area in 10G. Is it right?
Does your driver perform a copy of that 64B frame to user space?
Is this a single-threaded test running?
What does an lat_mem_rd -t (-t for random stride) test from lmbench give
for your system's memory latency? (Perhaps using numactl to ensure
local, or remote memory access, as you desire)
At line rate for minimum sized frames over 10 GbE, you have a frame
arriving every 60-odd nanoseconds. At that speed, you cannot take even
one cache miss per frame (*) in a single-threaded path and still achieve
line-rate PPS.
As it happens, there was a presentation at HP Labs recently, given by
Luigi Rizzo on his netmap work. The slides can be found at
http://info.iet.unipi.it/~luigi/netmap/talk-hp.html . As it happens,
Luigi presented some performance figures using an Intel 82599.
happy benchmarking,
rick jones
(*) much of my time has been spent in a world where a cache miss is
three digits worth of nanoseconds (to the left of the decimal),
sometimes high two digits.
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists