[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4D46F9AE.80606@goop.org>
Date: Mon, 31 Jan 2011 10:04:30 -0800
From: Jeremy Fitzhardinge <jeremy@...p.org>
To: Kaushik Barde <kbarde@...wei.com>
CC: 'Avi Kivity' <avi@...hat.com>, 'Jan Beulich' <JBeulich@...ell.com>,
'Xiaowei Yang' <xiaowei.yang@...wei.com>,
'Nick Piggin' <npiggin@...nel.dk>,
'Peter Zijlstra' <a.p.zijlstra@...llo.nl>,
fanhenglong@...wei.com, 'Kenneth Lee' <liguozhu@...wei.com>,
'linqaingmin' <linqiangmin@...wei.com>, wangzhenguo@...wei.com,
'Wu Fengguang' <fengguang.wu@...el.com>,
xen-devel@...ts.xensource.com, linux-kernel@...r.kernel.org,
'Marcelo Tosatti' <mtosatti@...hat.com>
Subject: Re: One (possible) x86 get_user_pages bug
On 01/30/2011 02:21 PM, Kaushik Barde wrote:
> I agree i.e. deviation from underlying arch consideration is not a good
> idea.
>
> Also, agreed, hypervisor knows which page entries are ready for TLB flush
> across vCPUs.
>
> But, using above knowledge, along with TLB flush based on IPI is a better
> solution. Its ability to synchronize it with pCPU based IPI and TLB flush
> across vCPU. is key.
I'm not sure I follow you here. The issue with TLB flush IPIs is that
the hypervisor doesn't know the purpose of the IPI and ends up
(potentially) waking up a sleeping VCPU just to flush its tlb - but
since it was sleeping there were no stale TLB entries to flush.
Xen's TLB flush hypercalls can optimise that case by only sending a real
IPI to PCPUs which are actually running target VCPUs. In other cases,
where a PCPU is known to have stale entries but it isn't running a
relevant VCPU, it can just mark a deferred TLB flush which gets executed
before the VCPU runs again.
In other words, Xen can take significant advantage of getting a
higher-level call ("flush these TLBs") compared just a simple IPI.
Are you suggesting that the hypervisor should export some kind of "known
dirty TLB" table to the guest, and have the guest work out which VCPUs
need IPIs sent to them? How would that work?
> IPIs themselves should be in few hundred uSecs in terms latency. Also, why
> should pCPU be in sleep state for active vCPU scheduled page workload?
A "few hundred uSecs" is really very slow - that's nearly a
millisecond. It's worth spending some effort to avoid those kinds of
delays.
J
> -Kaushik
>
> -----Original Message-----
> From: Avi Kivity [mailto:avi@...hat.com]
> Sent: Sunday, January 30, 2011 5:02 AM
> To: Jeremy Fitzhardinge
> Cc: Jan Beulich; Xiaowei Yang; Nick Piggin; Peter Zijlstra;
> fanhenglong@...wei.com; Kaushik Barde; Kenneth Lee; linqaingmin;
> wangzhenguo@...wei.com; Wu Fengguang; xen-devel@...ts.xensource.com;
> linux-kernel@...r.kernel.org; Marcelo Tosatti
> Subject: Re: One (possible) x86 get_user_pages bug
>
> On 01/27/2011 08:27 PM, Jeremy Fitzhardinge wrote:
>> And even just considering virtualization, having non-IPI-based tlb
>> shootdown is a measurable performance win, since a hypervisor can
>> optimise away a cross-VCPU shootdown if it knows no physical TLB
>> contains the target VCPU's entries. I can imagine the KVM folks could
>> get some benefit from that as well.
> It's nice to avoid the IPI (and waking up a cpu if it happens to be
> asleep) but I think the risk of deviating too much from the baremetal
> arch is too large, as demonstrated by this bug.
>
> (well, async page faults is a counterexample, I wonder if/when it will
> bite us)
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists