[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20070823091556.GA18456@skynet.ie>
Date: Thu, 23 Aug 2007 10:15:56 +0100
From: mel@...net.ie (Mel Gorman)
To: "Luck, Tony" <tony.luck@...el.com>
Cc: Jeremy Higdon <jeremy@....com>,
Kamalesh Babulal <kamalesh@...ux.vnet.ibm.com>,
Andi Kleen <ak@...e.de>,
Andrew Morton <akpm@...ux-foundation.org>,
linux-kernel@...r.kernel.org,
Balbir Singh <balbir@...ux.vnet.ibm.com>,
linux-ia64@...r.kernel.org
Subject: Re: [BUG] 2.6.23-rc3-mm1 Kernel panic - not syncing: DMA: Memory would be corrupted
On (22/08/07 16:27), Luck, Tony didst pronounce:
> > The more ioc's you have, the more space you will use.
>
> Default SW IOTLB allocation is 64MB ... how much should we see
> used per ioc?
>
> Kamelesh: You could try increasing the amount of sw iotlb space
> available by booting with a swiotlb=131072 argument (argument
> value is the number of 2K slabs to allocate ... 131072 would
> give you four times as much space as the default allocation).
>
I tried that value and just in case swiotlb=262144. An IA-64 machines I
have here fails with the same message anyway. i.e.
[ 19.834906] mptbase: Initiating ioc1 bringup
[ 20.317152] ioc1: LSI53C1030 C0: Capabilities={Initiator}
[ 15.474303] scsi1 : ioc1: LSI53C1030 C0, FwRev=01032821h, Ports=1, MaxQ=222, IRQ=72
[ 20.669730] GSI 142 (level, low) -> CPU 5 (0x1200) vector 73
[ 20.675602] ACPI: PCI Interrupt 0000:41:03.0[A] -> GSI 142 (level, low) -> IRQ 73
[ 20.683508] mptbase: Initiating ioc2 bringup
[ 21.166796] ioc2: LSI53C1030 C0: Capabilities={Initiator}
[ 21.180539] DMA: Out of SW-IOMMU space for 263200 bytes at device ?
[ 21.187018] Kernel panic - not syncing: DMA: Memory would be corrupted
--
Mel Gorman
Part-time Phd Student Linux Technology Center
University of Limerick IBM Dublin Software Lab
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists