[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <433343.88181.qm@web82107.mail.mud.yahoo.com>
Date: Mon, 11 Aug 2008 09:04:12 -0700 (PDT)
From: David Witbrodt <dawitbro@...global.net>
To: paulmck@...ux.vnet.ibm.com
Cc: Peter Zijlstra <peterz@...radead.org>,
linux-kernel@...r.kernel.org, Yinghai Lu <yhlu.kernel@...il.com>,
Ingo Molnar <mingo@...e.hu>,
Thomas Gleixner <tglx@...utronix.de>,
"H. Peter Anvin" <hpa@...or.com>, netdev <netdev@...r.kernel.org>
Subject: Re: [PATCH diagnostic] Re: HPET regression in 2.6.26 versus 2.6.25 -- RCU problem
I didn't check my email yesterday -- sorry about that, but sometimes life
intervenes -- so I'm a bit late replying.
Also, I see several messages relevant to this thread in my inbox: I have
decided to address each in order, so that I don't mix something up and do
something foolish.
> And here is the patch. It is still a bit raw, so the results should
> be viewed with some suspicion. It adds a default-off kernel parameter
> CONFIG_RCU_CPU_STALL which must be enabled.
Thanks for the patch. I had a problem applying the patch because I
have not yet transitioned my email system from my old machine to my new
3-system home network setup. (I used to share a data partition between
Windows and Linux so that my archives would stay in sync; my new setup
will allow keeping the POP downloads on one machine, and sharing the
archives via IMAP, but even since May I still haven't gotten around to
it.)
My ISP's webmail interface altered the whitespace, and I'm so new to
git that I couldn't figure out how to keep it from rejecting the
patch. I had updated Linus' git tree to 2.6.27-rc2, and when I saw
that your patch was against something in 2.6.27-rc1 I thought this
might be the problem. Visually inspecting the files, I saw that the
lines matched perfectly, other than whitespace, so I just gave up and
applied the patches manually.
I ran 'make menuconfig', but nothing about your new feature was asked.
Then I realized that I had changed the .config to CONFIG_PREEMPT because
of an experiment you had my try a few days ago. When I disabled that,
I was able to see the new option and enable it.
The kernel built fine, so I installed and rebooted...
> Rather than exponential backoff, it backs off to once per 30 seconds.
> My feeling upon thinking on it was that if you have stalled RCU grace
> periods for that long, a few extra printk() messages are probably the
> least of your worries...
Well, I was hoping to see something interesting. I ran it with parameters
"debug initcall_debug", and it locked up at the same place. I let it for
15 minutes, in case of some delayed reaction. Nada.
The output was nearly identical to what I posted last Tuesday (see
http://www.uwsg.indiana.edu/hypermail/linux/kernel/0808.0/2224.html).
Here are the last few lines:
==================================
[snip]
calling pci_bios_assign_resources+0x0/0x8b
pci 0000:00:01.0: PCI bridge, secondary bus 0000:01
pci 0000:00:01.0: IO window: 0xe000-0xefff
pci 0000:00:01.0: MEM window: 0xfdd00000-0xfdefffff
pci 0000:00:01.0: PREFETCH window: 0x000000d8000000-0x000000dfffffff
pci 0000:00:14.4: PCI bridge, secondary bus 0000:02
pci 0000:00:14.4: IO window: 0xd000-0xdfff
pci 0000:00:14.4: MEM window: 0xfdc00000-0xfdcfffff
pci 0000:00:14.4: PREFETCH window: 0x000000fdf00000-0x000000fdffffff
initcall pci_bios_assign_resources returned 0 after 285702 msecs
calling inet_init+0x0/0x250
NET: Registered protocol family 2
===== END OUTPUT =================
The only difference in the output was trivial: "285696 msecs" became
"285702 msecs". None of the printk()'s from your driver were executed.
(As I mentioned on Tuesday, that number of milliseconds is WAY off, and
it still bothers me. The total time from the GRUB screen disappearing
to the last line printed is < 5 secs (maybe < 3 secs), not 285 secs!)
Moving on to the other LKML messages....
Thanks,
Dave W.
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists