lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20060925233750.GA9791@in.ibm.com>
Date:	Mon, 25 Sep 2006 19:37:50 -0400
From:	Vivek Goyal <vgoyal@...ibm.com>
To:	Jean-Marc Saffroy <saffroy@...il.com>
Cc:	Andrew Morton <akpm@...l.org>, linux-kernel@...r.kernel.org,
	Jaroslav Kysela <perex@...e.cz>, Takashi Iwai <tiwai@...e.de>,
	Dave Anderson <anderson@...hat.com>,
	"Eric W. Biederman" <ebiederm@...ssion.com>
Subject: Re: oops in :snd_pcm_oss:resample_expand+0x19c/0x1f0

On Mon, Sep 25, 2006 at 11:01:52PM +0200, Jean-Marc Saffroy wrote:
> Hi Vivek,
> 
> Thanks for all the good news on the future of kdump, it's nice to know 
> that people are working on improving the user experience.
> 
> On Mon, 25 Sep 2006, Vivek Goyal wrote:
> 
> >>>>Oh and I wish I could use gdb on a kdump core. :-)
> >
> >Currently we can use gdb but only for linearly mapped region. You are 
> >right its just a matter of re-generating the elf headers and remap the 
> >vmalloc areas to enable module debugging in gdb. I can not do it after 
> >the crash so probably the best place would be do it in user space. A 
> >program can read /proc/vmcore and regenerate the headers for enabling 
> >module debugging with gdb.
> 
> I assume that "after the crash" means "in the kernel crash handler", 
> AFAICT the current dump from vmcore has all what's needed.
> 

Yes. I mean kernel crash handler. Yes all the contents are present
in the vmcore file. It is just a matter of somebody reading vmcore
and walk through either the page tables or may be vmlist and export
these mappings through ELF headers so that gdb can read it.
 
> >Hmm.. Crash vs gdb is an interesting issue. I have not used gdb very 
> >extensively on core dumps, but with my limited experiece, I found 
> >"crash" to be more friendly.
> 
> One thing I like *very much* in gdb is its ability to display function 
> params and local variables in any stack frame, and I haven't found out how 
> to do it with crash.
> 

Agreed. AFAIK, crash does not display the function params and local
variables. gdb has got this advantage, though last time I looked
at local variables dispalyed by gdb, they seemed to be junk. Not very
sure why it was so?

> >Crash has got so many in-built commands tailored for kernel debugging 
> >and gdb lacks all those. Yes, we can write gdb scripts to implement 
> >those, but last time Alaxender Nyberg wrote few gdb scripts to dump all 
> >the threads and it was so slow.
> 
> I agree that gdb is sometimes very slow, but maybe it's easier to optimize 
> gdb than to make crash smarter?
> 

I beg to differ here. Not sure why it is easier to optimize gdb. If we
try to optimize gdb by writing scripts, then IMHO, writing C program
is easier and faster. If the idea is to optimize gdb by modifying gdb
code then its no different than crash. In fact, why to reinvent the 
wheel if crash already does so many things for us. Yes, but probably we
need to modify crash to be able to obtain function parameters and local
variables.   

> For this particular problem (listing threads), the real fix would be to 
> add the PT_NOTE entry that each thread deserves, then gdb would let you do 
> "info threads" instead, and dump nice backtraces of each.
> 

Displaying even the currently non executing threads using "info threads"
and their backtraces is interesting. Crash can already do that. I am 
apprehensive about traversing through the task list and dumping every
thread's register state after a crash. There is no gurantee that list
is in a sane state. And in general, we are trying to make crash handler
as small as possible to improve the reliability of dumping operation.

Register state of every non-executing thread is already present in the
vmcore and IMHO, we should write scripts to get info about other
threads then doing it in kernel.

Thanks
Vivek
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ