[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <m1bq15gxml.fsf@frodo.ebiederm.org>
Date: Thu, 10 Jul 2008 15:52:50 -0700
From: ebiederm@...ssion.com (Eric W. Biederman)
To: Suresh Siddha <suresh.b.siddha@...el.com>
Cc: "mingo@...e.hu" <mingo@...e.hu>, "hpa@...or.com" <hpa@...or.com>,
"tglx@...utronix.de" <tglx@...utronix.de>,
"akpm@...ux-foundation.org" <akpm@...ux-foundation.org>,
"arjan@...ux.intel.com" <arjan@...ux.intel.com>,
"andi@...stfloor.org" <andi@...stfloor.org>,
"jbarnes@...tuousgeek.org" <jbarnes@...tuousgeek.org>,
"steiner@....com" <steiner@....com>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [patch 00/26] x64, x2apic/intr-remap: Interrupt-remapping and x2apic support
Suresh Siddha <suresh.b.siddha@...el.com> writes:
> Flushing the interrupt entry cache will take care of this. We modify the IRTE
> and then flush the interrupt entry cache before cleaning up the original
> vector allocated.
>
> Any new interrupts from the device will see the new entry. Old in flight
> interrupts will be registered at the CPU before the flush of the cache is
> complete.
That sounds nice in principle. I saw cpu cache flushes, I saw writes.
I did not see any reads which is necessary to get that behavior with
the standard pci transaction rules.
Having seen enough little races and misbehaving hardware I'm very paranoid
about irq migration. The current implementation is belt and suspenders
and I still think there are races that I have missed.
>> You are sizing an array as NR_IRQS this is something there should be
> sufficient
>> existing infrastructure to avoid. Arrays sized by NR_IRQS is a significant
>> problem both for scaling the system up and down so ultimately we need to kill
>> this. For now we should not introduce any new arrays.
>
> Ok. Ideally dynamic_irq_init()/cleanup() can take care of this. or
> create_irq()/destroy_irq() and embed this as a pointer somewhere inside
> irq_desc. I need to take a look at this more closer and post a fix up patch.
Sounds good. Ultimately we are looking at handler_data or chip_data.
There are very specific rules that meant I could not use them for
the msi data but otherwise I don't remember exactly what the are for.
IOMMU are covered though.
>> A lot of your code is generic, and some of it is for just x86_64. Since the
>> cpus are capable of running in 32bit mode. We really need to implement x86_32
>> and x86_64 support in the same code base. Which I believe means factoring out
>> pieces of io_apic_N.c into things such as msi.c that can be shared between the
>> two architectures.
>
> Yes, As you and Ingo mentioned, there is nothing 64bit specific and one
> can easily add the 32bit support. But before that we need, some more
> x86 unification and I am very short on resources currently :(
At least for msi the code you are working on was essentially unified
when it was written, it just happened to have two copies. I don't
think I'm asking for heaving lifting. Mostly just putting code that
is touched into something other then the growing monstrosity that is
ioapic.c
Further can we please see some better abstractions. In particular can
we generate a token for the irq destination. And have the msi and
ioapic setup read that token and program it into the hardware. The
rules for which bits go where is exactly the same both with and
without irq_remapping so having an if statement there seems to obscure
what is really happening. Especially if as it appears that we may be used
the new token format with x2apics without remapping.
My primary concern is that the end result be well factored irq handling code
so it is possible to get in there and look at the code and maintain it.
A small part of that is the 32bit support. Another part are the missing
abstractions I described. I don't know what else since I have barely scratched the surface patch
review wise.
Eric
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists