[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <200704242325.23447.ak@suse.de>
Date: Tue, 24 Apr 2007 23:25:23 +0200
From: Andi Kleen <ak@...e.de>
To: Jeremy Fitzhardinge <jeremy@...p.org>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
virtualization@...ts.osdl.org, lkml <linux-kernel@...r.kernel.org>,
Ian Pratt <ian.pratt@...source.com>,
Christian Limpach <Christian.Limpach@...cam.ac.uk>,
Adrian Bunk <bunk@...sta.de>
Subject: Re: [PATCH 06/25] xen: Core Xen implementation
On Monday 23 April 2007 23:56:44 Jeremy Fitzhardinge wrote:
> Core Xen Implementation
>
> This patch is a rollup of all the core pieces of the Xen
> implementation, including booting, memory management, interrupts, time
> and so on.
The patch is definitely too big.
> +#ifdef CONFIG_XEN
> +/* Xen only supports sysenter/sysexit in ring0 guests,
> + and only if it the guest asks for it. So for now,
> + this should never be used. */
> +ENTRY(xen_sti_sysexit)
> + CFI_STARTPROC
> + ud2
> + CFI_ENDPROC
> +ENDPROC(xen_sti_sysexit)
Put that elsewhere? It doesn't need to be here.
> +++ b/arch/i386/xen/enlighten.c
> @@ -0,0 +1,727 @@
Comments describing what all the files do?
> + unsigned maskedx = ~0;
> + if (*eax == 1)
> + maskedx = ~((1 << X86_FEATURE_APIC) |
> + (1 << X86_FEATURE_ACPI) |
> + (1 << X86_FEATURE_ACC));
Why ACC?
And why doesn't Xen mask those by itself?
And you got apic functions later which would be never called?
Why are the hooks needed then?
> +
> +static unsigned long xen_save_fl(void)
> +{
> + struct vcpu_info *vcpu;
> + unsigned long flags;
> +
> + preempt_disable();
> + vcpu = x86_read_percpu(xen_vcpu);
> + /* flag has opposite sense of mask */
> + flags = !vcpu->evtchn_upcall_mask;
> + preempt_enable();
If you use get_cpu/put_cpu it will be optimized away on PREEMPT && !SMP
(more occurrences)
> +static void xen_restore_fl(unsigned long flags)
> +{
> + struct vcpu_info *vcpu;
> +
> + preempt_disable();
> +
> + /* convert from IF type flag */
> + flags = !(flags & X86_EFLAGS_IF);
> + vcpu = x86_read_percpu(xen_vcpu);
> + vcpu->evtchn_upcall_mask = flags;
> + if (flags == 0) {
> + barrier(); /* unmask then check (avoid races) */
Don't you need a rmb() here then? The CPU could speculate reads
(more occurrences)
> + if (unlikely(vcpu->evtchn_upcall_pending))
> + force_evtchn_callback();
> + preempt_enable();
> + } else
> + preempt_enable_no_resched();
> +}
> +
> +static void xen_irq_disable(void)
> +{
> + struct vcpu_info *vcpu;
> + preempt_disable();
> + vcpu = x86_read_percpu(xen_vcpu);
> + vcpu->evtchn_upcall_mask = 1;
> + preempt_enable_no_resched();
First with the new per cpu the preempt disable shouldn't be needed
anymore because the thing is atomic. In the worst case you do
the change on the previous CPU, but that can happen anyways after
preempt_enable
And then when you have enabled who transfers the irq off state to the
new CPU?
> +static void xen_halt(void)
> +{
> +#if 0
> + if (irqs_disabled())
> + HYPERVISOR_vcpu_op(VCPUOP_down, smp_processor_id(), NULL);
> +#endif
> +}
Who halts then?
> +static void xen_load_gdt(const struct Xgt_desc_struct *dtr)
> +{
> + unsigned long *frames;
> + unsigned long va = dtr->address;
> + unsigned int size = dtr->size + 1;
> + int f;
> + struct multicall_space mcs;
> +
> + BUG_ON(size > 16*PAGE_SIZE);
Why 16?
> + count = desc->size / 8;
> + BUG_ON(count > 256);
should be >= ?
> +static void xen_set_iopl_mask(unsigned mask)
> +{
> +#if 0
> + struct physdev_set_iopl set_iopl;
> +
> + /* Force the change at ring 0. */
> + set_iopl.iopl = (mask == 0) ? 1 : (mask >> 12) & 3;
> + HYPERVISOR_physdev_op(PHYSDEVOP_set_iopl, &set_iopl);
> +#endif
And who does iopl then?
> + * Page-directory addresses above 4GB do not fit into architectural %cr3.
> + * When accessing %cr3, or equivalent field in vcpu_guest_context, guests
> + * must use the following accessor macros to pack/unpack valid MFNs.
> + */
> +#define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20))
> +#define xen_cr3_to_pfn(cr3) (((unsigned)(cr3) >> 12) | ((unsigned)(cr3) << 20))
Don't you lose bits >4GB here due to the casts before shift?
> + BUG_ON(memcmp(xen_start_info->magic, "xen-3.0", 7) != 0);
Hopefully Xen 4.0 can handle this then.
> +
> + /* Poke various useful things into boot_params */
> + LOADER_TYPE = (9 << 4) | 0;
| 0 ? Probably should be something symbolic
> +unsigned long xen_pmd_val(pmd_t pmd)
> +{
> + BUG();
> + return 0;
> +}
Why do we have pmd_val hooks then?
> +pmd_t xen_make_pmd(unsigned long pmd)
> +{
> + BUG();
> + return __pmd(0);
> +}
and make_pmd hooks?
> --- /dev/null
> +++ b/arch/i386/xen/time.c
... will review the rest later.
Can you please split it up a bit?
-Andi
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists