lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Sat, 8 Dec 2007 17:16:24 -0500 (EST) From: Steven Rostedt <rostedt@...dmis.org> To: Linus Torvalds <torvalds@...ux-foundation.org> cc: LKML <linux-kernel@...r.kernel.org>, Christoph Lameter <clameter@....com>, Ingo Molnar <mingo@...e.hu>, Peter Zijlstra <a.p.zijlstra@...llo.nl>, Christoph Hellwig <hch@...radead.org> Subject: Re: Major regression on hackbench with SLUB Hi Linus, > On Fri, 7 Dec 2007, Linus Torvalds wrote: > > > > Can you do one run with oprofile, and see exactly where the cost is? It > > should hopefully be pretty darn obvious, considering your timing. The results are here: http://people.redhat.com/srostedt/slub/results/slab.op http://people.redhat.com/srostedt/slub/results/slub.op > > Anyway, it's unclear whether the reason it goes into the slow-path because > the freelist is just always empty, or whether it hits the > > ... || !node_match(c, node) I did two things. First I tried making node_match always return true, the second was just commenting out the above check. They both had pretty much the exact same results. It slowed down by double! Instead of taking 15 seconds to complete, both took 30 seconds to complete. Here's the oprofiles of those runs. node_match return 1: http://people.redhat.com/srostedt/slub/results/slub-nonode.op comment out node_match check: http://people.redhat.com/srostedt/slub/results/slub-nochecknode.op > > [ The whole node match thing is likely totally bogus. I suspect we pay > *more* in trying to match nodes than we'd ever likely pay in just > returning the wrong node for an allocation, but that's neither here nor > there ] I don't think it's bogus by the result that removing it slowes it down by half. I'm wondering if there's not some other cache thrashing happening somewhere else in the slub code. I'll start adding some stats in the code to see where the congestion might lie. I'll look into this on Monday. Seems that both SLAB and SLUB run kernel compiles the same. Here's the results of my compile test. (make -j256 and chrt -f 10 make -j256) http://people.redhat.com/srostedt/slub/ -- Steve -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists