lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <54F17FB6.8040308@linux.intel.com>
Date:	Sat, 28 Feb 2015 16:43:34 +0800
From:	Jiang Liu <jiang.liu@...ux.intel.com>
To:	Bjorn Helgaas <bhelgaas@...gle.com>,
	Dave Airlie <airlied@...il.com>
CC:	LKML <linux-kernel@...r.kernel.org>,
	Rafael Wysocki <rjw@...ysocki.net>,
	"linux-pci@...r.kernel.org" <linux-pci@...r.kernel.org>,
	"linux-acpi@...r.kernel.org" <linux-acpi@...r.kernel.org>,
	Thomas Gleixner <tglx@...utronix.de>,
	Lv Zheng <lv.zheng@...el.com>
Subject: Re: regression in 4.0.0-rc1 with r8169 ethernet

On 2015/2/25 15:10, Bjorn Helgaas wrote:
> [+cc Jiang, Thomas, Lv, Rafael, linux-pci, linux-acpi]
> 
> On Tue, Feb 24, 2015 at 9:47 PM, Dave Airlie <airlied@...il.com> wrote:
>> Hey,
>>
>> just booted an old AMD rs780 box with new kernel and my ethernet
>> failed to come up!
>>
>> Looks like the MAC addr is all ff's because the PCI bridge windows are
>> messed up.
>>
>> I've attached two dmesg one from a 3.19.0-rc6 I had on it, and one
>> failing from the 4.0.0-rc1 time.
>>
>> b24e2bdde4af656bb0679a101265ebb8f8735d3c is latest Linus commit in
>> that tree (I have some radeon patches on top).
>>
>> motherboard is a Gigabyte GA-MA78GM-S2H, lspci also attached.
> 
> Here's the dmesg diff that looks relevant to me:
> 
> -pci_bus 0000:00: root bus resource [io  0x0000-0x0cf7]
> -pci_bus 0000:00: root bus resource [io  0x0d00-0xffff]
> -pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff]
> -pci_bus 0000:00: root bus resource [mem 0x000c0000-0x000dffff]
> -pci_bus 0000:00: root bus resource [mem 0xfed40000-0xfed44fff]
> -pci_bus 0000:00: root bus resource [mem 0x7ff00000-0xfebfffff]
> +pci_bus 0000:00: root bus resource [io  0x0cf8-0x0cff]
> +pci_bus 0000:00: root bus resource [io  0x0000-0x0cf7 window]
> +pci_bus 0000:00: root bus resource [io  0x0d00-0xffff window]
> +pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
> +pci_bus 0000:00: root bus resource [mem 0x000c0000-0x000dffff window]
> +pci_bus 0000:00: root bus resource [mem 0xfed40000-0xfed44fff window]
> 
> What's interesting is:
> 
> * v3.19 ignored [io  0x0cf8-0x0cff], but v4.0 includes it.  I think
> it's wrong to include it because that's the configuration space
> address/data registers, so it's consumed by the host bridge and not
> produced on the downstream side.
Hi Bjorn,
	We should ignore resources occupied by host bridge itself.
Will work out a patch to fix this issue.

> 
> * v3.19 includes [mem 0x7ff00000-0xfebfffff], but v4.0 does not.  This
> is what's screwing up the devices.
I need more information to figure out why resource [mem
0x7ff00000-0xfebfffff] is ignored by new ACPI parser.
Could you please help to forward the original dmesg attachments?
Regards!
Gerry

> 
> I think all the windows should be marked as ACPI_PRODUCER in _CRS
> since the space is "produced" on the downstream side of the bridge.
> The [io  0x0cf8-0x0cff] region should probably be marked
> ACPI_CONSUMER, and maybe that accounts for why v3.19 ignores it.  But
> I haven't found the code that does that yet.
> 
> I suspect this is all related to the ACPI resource parsing rework.  I
> looked through that briefly, but no issues jumped out at me, so this
> is just a heads-up in case it is obvious to you guys.
> 
> Dave, it'd be useful if you could collect an acpidump so we can look
> at the _CRS data in more detail.
> 
> Bjorn
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ