lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Sun, 04 Jan 2015 08:39:34 +0800
From:	Huang Ying <ying.huang@...el.com>
To:	Kirill Tkhai <tkhai@...dex.ru>
Cc:	Kirill Tkhai <ktkhai@...allels.com>,
	Ingo Molnar <mingo@...nel.org>,
	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>,
	Wu Fengguang <fengguang.wu@...el.com>
Subject: Re: [LKP] [sched] a15b12ac36a: -46.9%
 time.voluntary_context_switches +1.5% will-it-scale.per_process_ops

Hi, Kirill,

Sorry for late.

On Tue, 2014-12-23 at 11:57 +0300, Kirill Tkhai wrote:
> Hi, Huang,
> 
> what do these digits mean? What test does?
> 
> 23.12.2014, 08:16, "Huang Ying" <ying.huang@...el.com>:
> > FYI, we noticed the below changes on
> >
> > commit a15b12ac36ad4e7b856a4ae54937ae26a51aebad ("sched: Do not stop cpu in set_cpus_allowed_ptr() if task is not running")
> >
> > testbox/testcase/testparams: lkp-g5/will-it-scale/performance-lock1
> >
> > 1ba93d42727c4400  a15b12ac36ad4e7b856a4ae549
> > ----------------  --------------------------

Above is the good commit and the bad commit.

> >          %stddev     %change         %stddev
> >              \          |                \
> >    1517261 ±  0%      +1.5%    1539994 ±  0%  will-it-scale.per_process_ops

We have basic description of data above, where %stddev is standard
deviation.

What's more do you want?

Best Regards,
Huang, Ying

> >        247 ± 30%    +131.8%        573 ± 49%  sched_debug.cpu#61.ttwu_count
> >        225 ± 22%    +142.8%        546 ± 34%  sched_debug.cpu#81.ttwu_local
> >      15115 ± 44%     +37.3%      20746 ± 40%  numa-meminfo.node7.Active
> >       1028 ± 38%    +115.3%       2214 ± 36%  sched_debug.cpu#16.ttwu_local
> >          2 ± 19%    +133.3%          5 ± 43%  sched_debug.cpu#89.cpu_load[3]
> >         21 ± 45%     +88.2%         40 ± 23%  sched_debug.cfs_rq[99]:/.tg_load_contrib
> >        414 ± 33%     +98.6%        823 ± 28%  sched_debug.cpu#81.ttwu_count
> >          4 ± 10%     +88.2%          8 ± 12%  sched_debug.cfs_rq[33]:/.runnable_load_avg
> >         22 ± 26%     +80.9%         40 ± 24%  sched_debug.cfs_rq[103]:/.tg_load_contrib
> >          7 ± 17%     -41.4%          4 ± 25%  sched_debug.cfs_rq[41]:/.load
> >          7 ± 17%     -37.9%          4 ± 19%  sched_debug.cpu#41.load
> >          3 ± 22%    +106.7%          7 ± 10%  sched_debug.cfs_rq[36]:/.runnable_load_avg
> >        174 ± 13%     +48.7%        259 ± 31%  sched_debug.cpu#112.ttwu_count
> >          4 ± 19%     +88.9%          8 ±  5%  sched_debug.cfs_rq[35]:/.runnable_load_avg
> >        260 ± 10%     +55.6%        405 ± 26%  numa-vmstat.node3.nr_anon_pages
> >       1042 ± 10%     +56.0%       1626 ± 26%  numa-meminfo.node3.AnonPages
> >         26 ± 22%     +74.3%         45 ± 16%  sched_debug.cfs_rq[65]:/.tg_load_contrib
> >         21 ± 43%     +71.3%         37 ± 26%  sched_debug.cfs_rq[100]:/.tg_load_contrib
> >       3686 ± 21%     +40.2%       5167 ± 19%  sched_debug.cpu#16.ttwu_count
> >        142 ±  9%     +34.4%        191 ± 24%  sched_debug.cpu#112.ttwu_local
> >          5 ± 18%     +69.6%          9 ± 15%  sched_debug.cfs_rq[35]:/.load
> >          2 ± 30%    +100.0%          5 ± 37%  sched_debug.cpu#106.cpu_load[1]
> >          3 ± 23%    +100.0%          6 ± 48%  sched_debug.cpu#106.cpu_load[2]
> >          5 ± 18%     +69.6%          9 ± 15%  sched_debug.cpu#35.load
> >          9 ± 20%     +48.6%         13 ± 16%  sched_debug.cfs_rq[7]:/.runnable_load_avg
> >       1727 ± 15%     +43.9%       2484 ± 30%  sched_debug.cpu#34.ttwu_local
> >         10 ± 17%     -40.5%          6 ± 13%  sched_debug.cpu#41.cpu_load[0]
> >         10 ± 14%     -29.3%          7 ±  5%  sched_debug.cpu#45.cpu_load[4]
> >         10 ± 17%     -33.3%          7 ± 10%  sched_debug.cpu#41.cpu_load[1]
> >       6121 ±  8%     +56.7%       9595 ± 30%  sched_debug.cpu#13.sched_goidle
> >         13 ±  8%     -25.9%         10 ± 17%  sched_debug.cpu#39.cpu_load[2]
> >         12 ± 16%     -24.0%          9 ± 15%  sched_debug.cpu#37.cpu_load[2]
> >        492 ± 17%     -21.3%        387 ± 24%  sched_debug.cpu#46.ttwu_count
> >       3761 ± 11%     -23.9%       2863 ± 15%  sched_debug.cpu#93.curr->pid
> >        570 ± 19%     +43.2%        816 ± 17%  sched_debug.cpu#86.ttwu_count
> >       5279 ±  8%     +63.5%       8631 ± 33%  sched_debug.cpu#13.ttwu_count
> >        377 ± 22%     -28.6%        269 ± 14%  sched_debug.cpu#46.ttwu_local
> >       5396 ± 10%     +29.9%       7007 ± 14%  sched_debug.cpu#16.sched_goidle
> >       1959 ± 12%     +36.9%       2683 ± 15%  numa-vmstat.node2.nr_slab_reclaimable
> >       7839 ± 12%     +37.0%      10736 ± 15%  numa-meminfo.node2.SReclaimable
> >          5 ± 15%     +66.7%          8 ±  9%  sched_debug.cfs_rq[33]:/.load
> >          5 ± 25%     +47.8%          8 ± 10%  sched_debug.cfs_rq[37]:/.load
> >          2 ±  0%     +87.5%          3 ± 34%  sched_debug.cpu#89.cpu_load[4]
> >          5 ± 15%     +66.7%          8 ±  9%  sched_debug.cpu#33.load
> >          6 ± 23%     +41.7%          8 ± 10%  sched_debug.cpu#37.load
> >          8 ± 10%     -26.5%          6 ±  6%  sched_debug.cpu#51.cpu_load[1]
> >       7300 ± 37%     +63.6%      11943 ± 16%  softirqs.TASKLET
> >       2984 ±  6%     +43.1%       4271 ± 23%  sched_debug.cpu#20.ttwu_count
> >        328 ±  4%     +40.5%        462 ± 25%  sched_debug.cpu#26.ttwu_local
> >         10 ±  7%     -27.5%          7 ±  5%  sched_debug.cpu#43.cpu_load[3]
> >          9 ±  8%     -30.8%          6 ±  6%  sched_debug.cpu#41.cpu_load[3]
> >          9 ±  8%     -27.0%          6 ±  6%  sched_debug.cpu#41.cpu_load[4]
> >         10 ± 14%     -32.5%          6 ±  6%  sched_debug.cpu#41.cpu_load[2]
> >      16292 ±  6%     +42.8%      23260 ± 25%  sched_debug.cpu#13.nr_switches
> >         14 ± 28%     +55.9%         23 ±  8%  sched_debug.cpu#99.cpu_load[0]
> >          5 ±  8%     +28.6%          6 ± 12%  sched_debug.cpu#17.load
> >         13 ±  7%     -23.1%         10 ± 12%  sched_debug.cpu#39.cpu_load[3]
> >          7 ± 10%     -35.7%          4 ± 11%  sched_debug.cfs_rq[45]:/.runnable_load_avg
> >       5076 ± 13%     -21.8%       3970 ± 11%  numa-vmstat.node0.nr_slab_unreclaimable
> >      20306 ± 13%     -21.8%      15886 ± 11%  numa-meminfo.node0.SUnreclaim
> >         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#45.cpu_load[3]
> >         11 ± 11%     -29.5%          7 ± 14%  sched_debug.cpu#45.cpu_load[1]
> >         10 ± 12%     -26.8%          7 ±  6%  sched_debug.cpu#44.cpu_load[1]
> >         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#44.cpu_load[0]
> >          7 ± 17%     +48.3%         10 ±  7%  sched_debug.cfs_rq[11]:/.runnable_load_avg
> >         11 ± 12%     -34.1%          7 ± 11%  sched_debug.cpu#47.cpu_load[0]
> >         10 ± 10%     -27.9%          7 ±  5%  sched_debug.cpu#47.cpu_load[1]
> >         10 ±  8%     -26.8%          7 ± 11%  sched_debug.cpu#47.cpu_load[2]
> >         10 ±  8%     -28.6%          7 ± 14%  sched_debug.cpu#43.cpu_load[0]
> >         10 ± 10%     -27.9%          7 ± 10%  sched_debug.cpu#43.cpu_load[1]
> >         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#43.cpu_load[2]
> >      12940 ±  3%     +49.8%      19387 ± 35%  numa-meminfo.node2.Active(anon)
> >       3235 ±  2%     +49.8%       4844 ± 35%  numa-vmstat.node2.nr_active_anon
> >         17 ± 17%     +36.6%         24 ±  9%  sched_debug.cpu#97.cpu_load[2]
> >      14725 ±  8%     +21.8%      17928 ± 11%  sched_debug.cpu#16.nr_switches
> >        667 ± 10%     +45.3%        969 ± 22%  sched_debug.cpu#17.ttwu_local
> >       3257 ±  5%     +22.4%       3988 ± 11%  sched_debug.cpu#118.curr->pid
> >       3144 ± 15%     -20.7%       2493 ±  8%  sched_debug.cpu#95.curr->pid
> >       2192 ± 11%     +50.9%       3308 ± 37%  sched_debug.cpu#18.ttwu_count
> >          6 ± 11%     +37.5%          8 ± 19%  sched_debug.cfs_rq[22]:/.load
> >         12 ±  5%     +27.1%         15 ±  8%  sched_debug.cpu#5.cpu_load[1]
> >         11 ± 12%     -23.4%          9 ± 13%  sched_debug.cpu#37.cpu_load[3]
> >          6 ± 11%     +37.5%          8 ± 19%  sched_debug.cpu#22.load
> >          8 ±  8%     -25.0%          6 ±  0%  sched_debug.cpu#51.cpu_load[2]
> >          7 ±  6%     -20.0%          6 ± 11%  sched_debug.cpu#55.cpu_load[3]
> >         11 ±  9%     -17.4%          9 ±  9%  sched_debug.cpu#39.cpu_load[4]
> >         12 ±  5%     -22.9%          9 ± 11%  sched_debug.cpu#38.cpu_load[3]
> >        420 ± 13%     +43.0%        601 ±  9%  sched_debug.cpu#30.ttwu_local
> >       1682 ± 14%     +38.5%       2329 ± 17%  numa-meminfo.node7.AnonPages
> >        423 ± 13%     +37.0%        579 ± 16%  numa-vmstat.node7.nr_anon_pages
> >         15 ± 13%     +41.9%         22 ±  5%  sched_debug.cpu#99.cpu_load[1]
> >          6 ± 20%     +44.0%          9 ± 13%  sched_debug.cfs_rq[19]:/.runnable_load_avg
> >          9 ±  4%     -24.3%          7 ±  0%  sched_debug.cpu#43.cpu_load[4]
> >       6341 ±  7%     -19.6%       5100 ± 16%  sched_debug.cpu#43.curr->pid
> >       2577 ± 11%     -11.9%       2270 ± 10%  sched_debug.cpu#33.ttwu_count
> >         13 ±  6%     -18.5%         11 ± 12%  sched_debug.cpu#40.cpu_load[2]
> >       4828 ±  6%     +23.8%       5979 ±  6%  sched_debug.cpu#34.curr->pid
> >       4351 ± 12%     +33.9%       5824 ± 12%  sched_debug.cpu#36.curr->pid
> >         10 ±  8%     -23.8%          8 ±  8%  sched_debug.cpu#37.cpu_load[4]
> >         10 ± 14%     -28.6%          7 ±  6%  sched_debug.cpu#45.cpu_load[2]
> >         17 ± 22%     +40.6%         24 ±  7%  sched_debug.cpu#97.cpu_load[1]
> >         11 ±  9%     +21.3%         14 ±  5%  sched_debug.cpu#7.cpu_load[2]
> >         10 ±  8%     -26.2%          7 ± 10%  sched_debug.cpu#36.cpu_load[4]
> >      12853 ±  2%     +20.0%      15429 ± 11%  numa-meminfo.node2.AnonPages
> >       4744 ±  8%     +30.8%       6204 ± 11%  sched_debug.cpu#35.curr->pid
> >       3214 ±  2%     +20.0%       3856 ± 11%  numa-vmstat.node2.nr_anon_pages
> >       6181 ±  6%     +24.9%       7718 ±  9%  sched_debug.cpu#13.curr->pid
> >       6675 ± 23%     +27.5%       8510 ± 10%  sched_debug.cfs_rq[91]:/.tg_load_avg
> >     171261 ±  5%     -22.2%     133177 ± 15%  numa-numastat.node0.local_node
> >       6589 ± 21%     +29.3%       8522 ± 11%  sched_debug.cfs_rq[89]:/.tg_load_avg
> >       6508 ± 20%     +28.0%       8331 ±  8%  sched_debug.cfs_rq[88]:/.tg_load_avg
> >       6598 ± 22%     +29.2%       8525 ± 11%  sched_debug.cfs_rq[90]:/.tg_load_avg
> >        590 ± 13%     -21.4%        464 ±  7%  sched_debug.cpu#105.ttwu_local
> >     175392 ±  5%     -21.7%     137308 ± 14%  numa-numastat.node0.numa_hit
> >         11 ±  6%     -18.2%          9 ±  7%  sched_debug.cpu#38.cpu_load[4]
> >       6643 ± 23%     +27.4%       8465 ± 10%  sched_debug.cfs_rq[94]:/.tg_load_avg
> >       6764 ±  7%     +13.8%       7695 ±  7%  sched_debug.cpu#12.curr->pid
> >         29 ± 28%     +34.5%         39 ±  5%  sched_debug.cfs_rq[98]:/.tg_load_contrib
> >       1776 ±  7%     +29.4%       2298 ± 13%  sched_debug.cpu#11.ttwu_local
> >         13 ±  0%     -19.2%         10 ±  8%  sched_debug.cpu#40.cpu_load[3]
> >          7 ±  5%     -17.2%          6 ±  0%  sched_debug.cpu#51.cpu_load[3]
> >       7371 ± 20%     -18.0%       6045 ±  3%  sched_debug.cpu#1.sched_goidle
> >      26560 ±  2%     +14.0%      30287 ±  7%  numa-meminfo.node2.Slab
> >      16161 ±  6%      -9.4%      14646 ±  1%  sched_debug.cfs_rq[27]:/.avg->runnable_avg_sum
> >        351 ±  6%      -9.3%        318 ±  1%  sched_debug.cfs_rq[27]:/.tg_runnable_contrib
> >       7753 ± 27%     -22.9%       5976 ±  5%  sched_debug.cpu#2.sched_goidle
> >       3828 ±  9%     +17.3%       4490 ±  6%  sched_debug.cpu#23.sched_goidle
> >      23925 ±  2%     +23.0%      29419 ± 23%  numa-meminfo.node2.Active
> >         47 ±  6%     -15.8%         40 ± 19%  sched_debug.cpu#42.cpu_load[1]
> >        282 ±  5%      -9.7%        254 ±  7%  sched_debug.cfs_rq[109]:/.tg_runnable_contrib
> >        349 ±  5%      -9.3%        317 ±  1%  sched_debug.cfs_rq[26]:/.tg_runnable_contrib
> >       6941 ±  3%      +8.9%       7558 ±  7%  sched_debug.cpu#61.nr_switches
> >      16051 ±  5%      -8.9%      14618 ±  1%  sched_debug.cfs_rq[26]:/.avg->runnable_avg_sum
> >     238944 ±  3%      +9.2%     260958 ±  5%  numa-vmstat.node2.numa_local
> >      12966 ±  5%      -9.5%      11732 ±  6%  sched_debug.cfs_rq[109]:/.avg->runnable_avg_sum
> >       1004 ±  3%      +8.2%       1086 ±  4%  sched_debug.cpu#118.sched_goidle
> >      20746 ±  4%      -8.4%      19000 ±  1%  sched_debug.cfs_rq[45]:/.avg->runnable_avg_sum
> >        451 ±  4%      -8.3%        413 ±  1%  sched_debug.cfs_rq[45]:/.tg_runnable_contrib
> >       3538 ±  4%     +17.2%       4147 ±  8%  sched_debug.cpu#26.ttwu_count
> >         16 ±  9%     +13.8%         18 ±  2%  sched_debug.cpu#99.cpu_load[3]
> >       1531 ±  0%     +11.3%       1704 ±  1%  numa-meminfo.node7.KernelStack
> >       3569 ±  3%     +17.2%       4182 ± 10%  sched_debug.cpu#24.sched_goidle
> >       1820 ±  3%     -12.5%       1594 ±  8%  slabinfo.taskstats.num_objs
> >       1819 ±  3%     -12.4%       1594 ±  8%  slabinfo.taskstats.active_objs
> >       4006 ±  5%     +19.1%       4769 ±  8%  sched_debug.cpu#17.sched_goidle
> >      21412 ± 19%     -17.0%      17779 ±  3%  sched_debug.cpu#2.nr_switches
> >         16 ±  9%     +24.2%         20 ±  4%  sched_debug.cpu#99.cpu_load[2]
> >      10493 ±  7%     +13.3%      11890 ±  4%  sched_debug.cpu#23.nr_switches
> >       1207 ±  2%     -46.9%        640 ±  4%  time.voluntary_context_switches
> >
> >                           time.voluntary_context_switches
> >
> >   1300 ++-----------*--*--------------------*-------------------------------+
> >        *..*.*..*.. +      *.*..*..*.*..*..*     .*..*..*.  .*..*.*..*..     |
> >   1200 ++         *                            *         *.            *.*..*
> >   1100 ++                                                                   |
> >        |                                                                    |
> >   1000 ++                                                                   |
> >        |                                                                    |
> >    900 ++                                                                   |
> >        |                                                                    |
> >    800 ++                                                                   |
> >    700 ++                                                                   |
> >        O    O     O       O O  O       O  O O  O O       O       O          |
> >    600 ++ O    O    O  O          O O                  O    O  O            |
> >        |                                            O                       |
> >    500 ++-------------------------------------------------------------------+
> >
> >         [*] bisect-good sample
> >         [O] bisect-bad  sample
> >
> > To reproduce:
> >
> >         apt-get install ruby ruby-oj
> >         git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
> >         cd lkp-tests
> >         bin/setup-local job.yaml # the job file attached in this email
> >         bin/run-local   job.yaml
> >
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> >
> 
> Regards,
> Kirill


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ