lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Thu, 23 Apr 2009 09:37:09 +0200
From:	Andi Kleen <andi@...stfloor.org>
To:	Ingo Molnar <mingo@...e.hu>
Cc:	Andi Kleen <andi@...stfloor.org>,
	Linus Torvalds <torvalds@...ux-foundation.org>,
	Jeff Garzik <jeff@...zik.org>,
	LKML <linux-kernel@...r.kernel.org>,
	Thomas Gleixner <tglx@...utronix.de>,
	Ingo Molnar <mingo@...hat.com>,
	"H. Peter Anvin" <hpa@...or.com>, x86@...nel.org
Subject: Re: fresh data was Re: [PATCH] X86-32: Let gcc decide whether to inline memcpy was Re: New x86 warning

On Thu, Apr 23, 2009 at 08:36:25AM +0200, Ingo Molnar wrote:
> 
> * Andi Kleen <andi@...stfloor.org> wrote:
> 
> > Andi Kleen <andi@...stfloor.org> writes:
> > 
> > >> > Quick test here:
> > >> 
> > >> How about you just compile the kernel with gcc-3.2 and compare the number 
> > >> of calls to memcpy before-and-after instead? That's the real test.
> > >
> > > I waited over 10 minutes for the full vmlinux objdumps to finish. sorry lost
> > > patience. If someone has a fast disassembler we can try it. I'll leave
> > > them running over night, maybe there are exact numbers tomorrow.
> > >
> > > But from a quick check (find -name '*.o' | xargs nm | grep memcpy) there are
> > > very little files which call it with the patch, so there's some
> > > evidence that there isn't a dramatic increase.
> > 
> > I let the objdumps finish over night. [...]
> 
> objdump -d never took me more than a minute - let alone a full 

I use objdump -S. Maybe that's slower than -d.

Hmm quick test, yes -S seems to be much slower than -d. Thanks for
the hint. I guess I should switch to -d for these cases, unfortunately
-S seems to be hardcoded in my fingers and of course it gives much
nicer output if you have debug info.

> night. You must be doing something really wrong there. Looking at 
> objdump -d is an essential, unavoidable component of my workflow 
> with x86 architecture patches, you need to find a way to do it 

I do it all the time too, but only for specific functions, not
for full kernels. I have a objdump-symbol script for that that
looks up a symbol in the symbol table and only disassembles
the function I'm interested in 
(ftp://firstfloor.org/pub/ak/perl/objdump-symbol) 
I normally don't look at full listings of the complete kernel.

> > [...] On my setup (defconfig + some additions) there are actually 
> > less calls to out of line memcpy/__memcpy with the patch. I see 
> > only one for my defconfig, while there are ~10 without the patch. 
> > So it makes very little difference. The code size savings must 
> > come from more efficient code generation for the inline case. I 
> > haven't investigated that in detail though.
> > 
> > So the patch seems like a overall win.
> 
> It's a clear loss here with GCC 3.4, and it took me less than 5 
> minutes to figure that out.

Loss in what way?

> 
> With what precise compiler version did you test (please paste the 
> gcc -v output), and could you send me the precise .config you used, 

See the 2nd previous mail: 3.2.3

I didn't do tests with later versions, assuming there are no 
regressions.

> and describe the method you used to determine the number of 
> out-of-line memcpy calls? I'd like to double-check your numbers.

objdump -S ... | grep call.*memcpy         (gives some false positives,
you have to weed them out)

In addition I did a quick find -name '*.o' | xargs nm | grep 'U.*memcpy$'
to (under) estimate the calls  

-Andi
-- 
ak@...ux.intel.com -- Speaking for myself only.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ