[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20100903101437.GA31575@redhat.com>
Date: Fri, 3 Sep 2010 13:14:37 +0300
From: "Michael S. Tsirkin" <mst@...hat.com>
To: Shirley Ma <mashirle@...ibm.com>
Cc: xiaohui.xin@...el.com, netdev@...r.kernel.org, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org, mingo@...e.hu, davem@...emloft.net,
herbert@...dor.hengli.com.au, jdike@...ux.intel.com
Subject: Re: [RFC PATCH v9 00/16] Provide a zero-copy method on KVM
virtio-net.
On Tue, Aug 10, 2010 at 06:23:24PM -0700, Shirley Ma wrote:
> Hello Xiaohui,
>
> On Fri, 2010-08-06 at 17:23 +0800, xiaohui.xin@...el.com wrote:
> > Our goal is to improve the bandwidth and reduce the CPU usage.
> > Exact performance data will be provided later.
>
> Have you had any performance data to share here? I tested my
> experimental macvtap zero copy for TX only. The performance I have seen
> as below without any tuning, (default setting):
>
> Before: netperf 16K message size results with 60 secs run is 7.5Gb/s
> over ixgbe 10GbE card. perf top shows:
>
> 2103.00 12.9% copy_user_generic_string
> 1541.00 9.4% handle_tx
> 1490.00 9.1% _raw_spin_unlock_irqrestore
> 1361.00 8.3% _raw_spin_lock_irqsave
> 1288.00 7.9% _raw_spin_lock
> 924.00 5.7% vhost_worker
>
> After: netperf results with 60 secs run is 8.1Gb/s, perf output:
>
> 1093.00 9.9% _raw_spin_unlock_irqrestore
> 1048.00 9.5% handle_tx
> 934.00 8.5% _raw_spin_lock_irqsave
> 864.00 7.9% _raw_spin_lock
> 644.00 5.9% vhost_worker
> 387.00 3.5% use_mm
>
> I am still working on collecting more data (latency, cpu
> utilization...). I will let you know once I get all data for macvtap TX
> zero copy. Also I found some vhost performance regression on the new
> kernel with tuning. I used to get 9.4Gb/s, now I couldn't get it.
>
> Shirley
Could you please try disabling mergeable buffers, and see if this gets
you back where you were?
-global virtio-net-pci.mrg_rxbuf=off
--
MST
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists