lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 17 Jun 2011 11:22:43 -0700
From:	Tim Chen <tim.c.chen@...ux.intel.com>
To:	Linus Torvalds <torvalds@...ux-foundation.org>
Cc:	Peter Zijlstra <a.p.zijlstra@...llo.nl>,
	Andi Kleen <ak@...ux.intel.com>,
	Shaohua Li <shaohua.li@...el.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	Hugh Dickins <hughd@...gle.com>,
	KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>,
	Benjamin Herrenschmidt <benh@...nel.crashing.org>,
	David Miller <davem@...emloft.net>,
	Martin Schwidefsky <schwidefsky@...ibm.com>,
	Russell King <rmk@....linux.org.uk>,
	Paul Mundt <lethal@...ux-sh.org>,
	Jeff Dike <jdike@...toit.com>,
	Richard Weinberger <richard@....at>,
	"Luck, Tony" <tony.luck@...el.com>,
	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
	Mel Gorman <mel@....ul.ie>, Nick Piggin <npiggin@...nel.dk>,
	Namhyung Kim <namhyung@...il.com>,
	"Shi, Alex" <alex.shi@...el.com>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	"linux-mm@...ck.org" <linux-mm@...ck.org>,
	"Rafael J. Wysocki" <rjw@...k.pl>
Subject: Re: REGRESSION: Performance regressions from switching
 anon_vma->lock to mutex

On Thu, 2011-06-16 at 20:58 -0700, Linus Torvalds wrote:

> 
> So Tim, I'd like you to test out my first patch (that only does the
> anon_vma_clone() case) once again, but now in the cleaned-up version.
> Does this patch really make a big improvement for you? If so, this
> first step is probably worth doing regardless of the more complicated
> second step, but I'd want to really make sure it's ok, and that the
> performance improvement you saw is consistent and not a fluke.
> 
>                  Linus

Linus,

For this patch, I've run it 10 times and got an average throughput of
104.9% compared with 2.6.39 vanilla baseline.  Wide variations are seen
run to run and the difference between max and min throughput is 52% of
average value.

So to recap,

                        Throughput
2.6.39(vanilla)         100.0%
2.6.39+ra-patch         166.7%  (+66.7%)        
3.0-rc2(vanilla)         68.0%  (-32%)
3.0-rc2+linus           115.7%  (+15.7%)
3.0-rc2+linus+softirq    86.2%  (-17.3%)
3.0-rc2+linus (v2)      104.9%  (+4.9%)

The time spent in the anon_vma mutex seems to directly affect
throughput.

In one run on your patch, I got a low throughput of 90.1% vs 2.6.39
throughput. The mutex_lock occupied 15.6% of cpu.

In another run, I got a high throughput of 120.8% vs 2.6.39 throughput.
The mutex lock occupied 7.5% of cpu.

I've attached the profiles of the two runs and a 3.0-rc2 vanilla run for
your reference.

I will follow up later with numbers that has Peter's patch added.

Thanks.

Tim

----------Profiles Below-------------------------

3.0-rc2+linus(v2) run 1 (90.1% throughput vs 2.6.39)

-     15.60%          exim  [kernel.kallsyms]             [k] __mutex_lock_common.clone.5
   - __mutex_lock_common.clone.5
      - 99.99% __mutex_lock_slowpath
         - mutex_lock
            + 75.52% anon_vma_lock.clone.10
            + 23.88% anon_vma_clone
-      4.38%          exim  [kernel.kallsyms]             [k] _raw_spin_lock_irqsave
   - _raw_spin_lock_irqsave
      + 82.83% cpupri_set
      + 6.75% try_to_wake_up
      + 5.35% release_pages
      + 1.72% pagevec_lru_move_fn
      + 0.93% get_page_from_freelist
      + 0.51% lock_timer_base.clone.20
+      3.22%          exim  [kernel.kallsyms]             [k] page_fault
+      2.62%          exim  [kernel.kallsyms]             [k] do_raw_spin_lock
+      2.30%          exim  [kernel.kallsyms]             [k] mutex_unlock
+      2.02%          exim  [kernel.kallsyms]             [k] unmap_vmas


3.0-rc2_linus(v2) run 2 (120.8% throughput vs 2.6.39)

-      7.53%             exim  [kernel.kallsyms]             [k] __mutex_lock_common.clone.5
   - __mutex_lock_common.clone.5
      - 99.99% __mutex_lock_slowpath
         - mutex_lock
            + 75.99% anon_vma_lock.clone.10
            + 22.68% anon_vma_clone
            + 0.70% unlink_file_vma
-      4.15%             exim  [kernel.kallsyms]             [k] _raw_spin_lock_irqsave
   - _raw_spin_lock_irqsave
      + 83.37% cpupri_set
      + 7.06% release_pages
      + 2.74% pagevec_lru_move_fn
      + 2.18% try_to_wake_up
      + 0.99% get_page_from_freelist
      + 0.59% lock_timer_base.clone.20
      + 0.58% lock_hrtimer_base.clone.16
+      4.06%             exim  [kernel.kallsyms]             [k] page_fault
+      2.33%             exim  [kernel.kallsyms]             [k] unmap_vmas
+      2.22%             exim  [kernel.kallsyms]             [k] do_raw_spin_lock
+      2.05%             exim  [kernel.kallsyms]             [k] page_cache_get_speculative
+      1.98%             exim  [kernel.kallsyms]             [k] mutex_unlock


3.0-rc2 vanilla run 

-     18.60%          exim  [kernel.kallsyms]        [k] __mutex_lock_common.clone.5                                            ↑
   - __mutex_lock_common.clone.5                                                                                                ▮
      - 99.99% __mutex_lock_slowpath                                                                                            ▒
         - mutex_lock                                                                                                           ▒
            - 99.54% anon_vma_lock.clone.10                                                                                     ▒
               + 38.99% anon_vma_clone                                                                                          ▒
               + 37.56% unlink_anon_vmas                                                                                        ▒
               + 11.92% anon_vma_fork                                                                                           ▒
               + 11.53% anon_vma_free                                                                                           ▒
-      4.03%          exim  [kernel.kallsyms]        [k] _raw_spin_lock_irqsave                                                 ▒
   - _raw_spin_lock_irqsave                                                                                                     ▒
      + 87.25% cpupri_set                                                                                                       ▒
      + 4.75% release_pages                                                                                                     ▒
      + 3.68% try_to_wake_up                                                                                                    ▒
      + 1.17% pagevec_lru_move_fn                                                                                               ▒
      + 0.71% get_page_from_freelist                                                                                            ▒
+      3.00%          exim  [kernel.kallsyms]        [k] do_raw_spin_lock                                                       ▒
+      2.90%          exim  [kernel.kallsyms]        [k] page_fault                                                             ▒
+      2.25%          exim  [kernel.kallsyms]        [k] mutex_unlock                                                           ▒
+      1.82%          exim  [kernel.kallsyms]        [k] unmap_vmas                                                             ▒
+      1.62%          exim  [kernel.kallsyms]        [k] copy_page_c                                                            ▒


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ