lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Wed, 20 Aug 2014 13:05:15 +0200 From: "Michael S. Tsirkin" <mst@...hat.com> To: Razya Ladelsky <RAZYA@...ibm.com> Cc: abel.gordon@...il.com, Alex Glikson <GLIKSON@...ibm.com>, David Miller <davem@...emloft.net>, Eran Raichstein <ERANRA@...ibm.com>, Joel Nider <JOELN@...ibm.com>, kvm@...r.kernel.org, kvm-owner@...r.kernel.org, linux-kernel@...r.kernel.org, netdev@...r.kernel.org, virtualization@...ts.linux-foundation.org, Yossi Kuperman1 <YOSSIKU@...ibm.com> Subject: Re: [PATCH] vhost: Add polling mode On Tue, Aug 19, 2014 at 11:36:31AM +0300, Razya Ladelsky wrote: > > That was just one example. There many other possibilities. Either > > actually make the systems load all host CPUs equally, or divide > > throughput by host CPU. > > > > The polling patch adds this capability to vhost, reducing costly exit > overhead when the vm is loaded. > > In order to load the vm I ran netperf with msg size of 256: > > Without polling: 2480 Mbits/sec, utilization: vm - 100% vhost - 64% > With Polling: 4160 Mbits/sec, utilization: vm - 100% vhost - 100% > > Therefore, throughput/cpu without polling is 15.1, and 20.8 with polling. > Can you please present results in a form that makes it possible to see the effect on various configurations and workloads? Here's one example where this was done: https://lkml.org/lkml/2014/8/14/495 You really should also provide data about your host configuration (missing in the above link). > My intention was to load vhost as close as possible to 100% utilization > without polling, in order to compare it to the polling utilization case > (where vhost is always 100%). > The best use case, of course, would be when the shared vhost thread work > (TBD) is integrated and then vhost will actually be using its polling > cycles to handle requests of multiple devices (even from multiple vms). > > Thanks, > Razya -- MST -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists