lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 19 Nov 2012 17:05:12 -0800 (PST) From: David Rientjes <rientjes@...gle.com> To: Mel Gorman <mgorman@...e.de> cc: Ingo Molnar <mingo@...nel.org>, linux-kernel@...r.kernel.org, linux-mm@...ck.org, Peter Zijlstra <a.p.zijlstra@...llo.nl>, Paul Turner <pjt@...gle.com>, Lee Schermerhorn <Lee.Schermerhorn@...com>, Christoph Lameter <cl@...ux.com>, Rik van Riel <riel@...hat.com>, Andrew Morton <akpm@...ux-foundation.org>, Andrea Arcangeli <aarcange@...hat.com>, Linus Torvalds <torvalds@...ux-foundation.org>, Thomas Gleixner <tglx@...utronix.de>, Johannes Weiner <hannes@...xchg.org>, Hugh Dickins <hughd@...gle.com> Subject: Re: [PATCH 00/27] Latest numa/core release, v16 On Mon, 19 Nov 2012, David Rientjes wrote: > I confirm that SPECjbb2005 1.07 -Xmx4g regresses in terms of throughput on > my 16-way, 4 node system with 32GB of memory using 16 warehouses and 240 > measurement seconds. I averaged the throughput for five runs on each > kernel. > > Java(TM) SE Runtime Environment (build 1.6.0_06-b02) > Java HotSpot(TM) 64-Bit Server VM (build 10.0-b22, mixed mode) > > Both kernels have > CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE=y > CONFIG_TRANSPARENT_HUGEPAGE=y > CONFIG_TRANSPARENT_HUGEPAGE_ALWAYS=y > > numa/core at 01aa90068b12 ("sched: Use the best-buddy 'ideal cpu' in > balancing decisions") with > > CONFIG_ARCH_SUPPORTS_NUMA_BALANCING=y > CONFIG_ARCH_WANTS_NUMA_GENERIC_PGPROT=y > CONFIG_NUMA_BALANCING=y > CONFIG_NUMA_BALANCING_HUGEPAGE=y > CONFIG_ARCH_USES_NUMA_GENERIC_PGPROT=y > CONFIG_ARCH_USES_NUMA_GENERIC_PGPROT_HUGEPAGE=y > > had a throughput of 128315.19 SPECjbb2005 bops. > > numa/core at ec05a2311c35 ("Merge branch 'sched/urgent' into sched/core") > had an average throughput of 136918.34 SPECjbb2005 bops, which is a 6.3% > regression. > perftop during the run on numa/core at 01aa90068b12 ("sched: Use the best-buddy 'ideal cpu' in balancing decisions"): 15.99% [kernel] [k] page_fault 4.05% [kernel] [k] getnstimeofday 3.96% [kernel] [k] _raw_spin_lock 3.20% [kernel] [k] rcu_check_callbacks 2.93% [kernel] [k] generic_smp_call_function_interrupt 2.90% [kernel] [k] __do_page_fault 2.82% [kernel] [k] ktime_get 2.62% [kernel] [k] read_tsc 2.41% [kernel] [k] handle_mm_fault 2.01% [kernel] [k] flush_tlb_func 1.99% [kernel] [k] retint_swapgs 1.83% [kernel] [k] emulate_vsyscall 1.71% [kernel] [k] handle_pte_fault 1.63% [kernel] [k] task_tick_fair 1.57% [kernel] [k] clear_page_c 1.55% [kernel] [k] down_read_trylock 1.54% [kernel] [k] copy_user_generic_string 1.48% [kernel] [k] ktime_get_update_offsets 1.37% [kernel] [k] find_vma 1.23% [kernel] [k] mpol_misplaced 1.14% [kernel] [k] task_numa_fault 1.10% [kernel] [k] run_timer_softirq 1.06% [kernel] [k] up_read 0.87% [kernel] [k] __bad_area_nosemaphore 0.82% [kernel] [k] write_ok_or_segv 0.77% [kernel] [k] update_cfs_shares 0.76% [kernel] [k] update_curr 0.75% [kernel] [k] error_sti 0.75% [kernel] [k] get_vma_policy 0.73% [kernel] [k] smp_call_function_many 0.66% [kernel] [k] do_wp_page 0.60% [kernel] [k] error_entry 0.60% [kernel] [k] call_function_interrupt 0.59% [kernel] [k] error_exit 0.58% [kernel] [k] _raw_spin_lock_irqsave 0.58% [kernel] [k] tick_sched_timer 0.57% [kernel] [k] __do_softirq 0.57% [kernel] [k] mem_cgroup_count_vm_event 0.56% [kernel] [k] account_user_time 0.56% [kernel] [k] spurious_fault 0.54% [kernel] [k] acct_update_integrals 0.54% [kernel] [k] bad_area_nosemaphore [ Both kernels for this test were booted with cgroup_disable=memory on the command line, why mem_cgroup_count_vm_event shows up at all here is strange... ] -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists