[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <E8C008223DD5F64485DFBDF6D4B7F71D0289EDC2@msgswbmnmsp25.wellsfargo.com>
Date: Sun, 18 Mar 2007 22:48:55 -0500
From: <Greg.Chandler@...lsfargo.com>
To: <auke-jan.h.kok@...el.com>
Cc: <hancockr@...w.ca>, <linux-kernel@...r.kernel.org>,
<e1000-devel@...ts.sourceforge.net>
Subject: RE: 1000xf bus problem
That would completely and uttly suck if it were the case.
So in theory, the card should be talking at full speed, but since the
bridge is 66 then it would be the bottleneck right?
There are 3 pci busses in the server I am working on, so I know there is
at least _a_ bridge there.....
-----Original Message-----
From: Kok, Auke [mailto:auke-jan.h.kok@...el.com]
Sent: Sunday, March 18, 2007 10:40 PM
To: Chandler, Greg
Cc: auke-jan.h.kok@...el.com; hancockr@...w.ca;
linux-kernel@...r.kernel.org; e1000-devel@...ts.sourceforge.net
Subject: Re: 1000xf bus problem
Greg.Chandler@...lsfargo.com wrote:
> If you mean dmesg it says this:
> e1000: 0000:0d:02.0: e1000_probe: (PCI-X:100MHz:64-bit) {macaddress}
>
> That's weird... dmesg shows one thing, lspci shows another, and my
> data transfers seem to point to the lspci info...
>
> Any idea which I should trust?
Both, the e1000 driver asks the card what it sees from it's side of the
connection, and lspci tells you what the cpu side of it is connected to.
Since stuff like pci bridges exist, both could very well be correct!
I highly suspect that that is exactly the case.
Auke
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists