lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Tue, 23 Dec 2014 11:57:13 +0300
From:	Kirill Tkhai <tkhai@...dex.ru>
To:	Huang Ying <ying.huang@...el.com>,
	Kirill Tkhai <ktkhai@...allels.com>
Cc:	Ingo Molnar <mingo@...nel.org>,
	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: Re: [LKP] [sched] a15b12ac36a: -46.9% time.voluntary_context_switches +1.5% will-it-scale.per_process_ops

Hi, Huang,

what do these digits mean? What test does?

23.12.2014, 08:16, "Huang Ying" <ying.huang@...el.com>:
> FYI, we noticed the below changes on
>
> commit a15b12ac36ad4e7b856a4ae54937ae26a51aebad ("sched: Do not stop cpu in set_cpus_allowed_ptr() if task is not running")
>
> testbox/testcase/testparams: lkp-g5/will-it-scale/performance-lock1
>
> 1ba93d42727c4400  a15b12ac36ad4e7b856a4ae549
> ----------------  --------------------------
>          %stddev     %change         %stddev
>              \          |                \
>    1517261 ±  0%      +1.5%    1539994 ±  0%  will-it-scale.per_process_ops
>        247 ± 30%    +131.8%        573 ± 49%  sched_debug.cpu#61.ttwu_count
>        225 ± 22%    +142.8%        546 ± 34%  sched_debug.cpu#81.ttwu_local
>      15115 ± 44%     +37.3%      20746 ± 40%  numa-meminfo.node7.Active
>       1028 ± 38%    +115.3%       2214 ± 36%  sched_debug.cpu#16.ttwu_local
>          2 ± 19%    +133.3%          5 ± 43%  sched_debug.cpu#89.cpu_load[3]
>         21 ± 45%     +88.2%         40 ± 23%  sched_debug.cfs_rq[99]:/.tg_load_contrib
>        414 ± 33%     +98.6%        823 ± 28%  sched_debug.cpu#81.ttwu_count
>          4 ± 10%     +88.2%          8 ± 12%  sched_debug.cfs_rq[33]:/.runnable_load_avg
>         22 ± 26%     +80.9%         40 ± 24%  sched_debug.cfs_rq[103]:/.tg_load_contrib
>          7 ± 17%     -41.4%          4 ± 25%  sched_debug.cfs_rq[41]:/.load
>          7 ± 17%     -37.9%          4 ± 19%  sched_debug.cpu#41.load
>          3 ± 22%    +106.7%          7 ± 10%  sched_debug.cfs_rq[36]:/.runnable_load_avg
>        174 ± 13%     +48.7%        259 ± 31%  sched_debug.cpu#112.ttwu_count
>          4 ± 19%     +88.9%          8 ±  5%  sched_debug.cfs_rq[35]:/.runnable_load_avg
>        260 ± 10%     +55.6%        405 ± 26%  numa-vmstat.node3.nr_anon_pages
>       1042 ± 10%     +56.0%       1626 ± 26%  numa-meminfo.node3.AnonPages
>         26 ± 22%     +74.3%         45 ± 16%  sched_debug.cfs_rq[65]:/.tg_load_contrib
>         21 ± 43%     +71.3%         37 ± 26%  sched_debug.cfs_rq[100]:/.tg_load_contrib
>       3686 ± 21%     +40.2%       5167 ± 19%  sched_debug.cpu#16.ttwu_count
>        142 ±  9%     +34.4%        191 ± 24%  sched_debug.cpu#112.ttwu_local
>          5 ± 18%     +69.6%          9 ± 15%  sched_debug.cfs_rq[35]:/.load
>          2 ± 30%    +100.0%          5 ± 37%  sched_debug.cpu#106.cpu_load[1]
>          3 ± 23%    +100.0%          6 ± 48%  sched_debug.cpu#106.cpu_load[2]
>          5 ± 18%     +69.6%          9 ± 15%  sched_debug.cpu#35.load
>          9 ± 20%     +48.6%         13 ± 16%  sched_debug.cfs_rq[7]:/.runnable_load_avg
>       1727 ± 15%     +43.9%       2484 ± 30%  sched_debug.cpu#34.ttwu_local
>         10 ± 17%     -40.5%          6 ± 13%  sched_debug.cpu#41.cpu_load[0]
>         10 ± 14%     -29.3%          7 ±  5%  sched_debug.cpu#45.cpu_load[4]
>         10 ± 17%     -33.3%          7 ± 10%  sched_debug.cpu#41.cpu_load[1]
>       6121 ±  8%     +56.7%       9595 ± 30%  sched_debug.cpu#13.sched_goidle
>         13 ±  8%     -25.9%         10 ± 17%  sched_debug.cpu#39.cpu_load[2]
>         12 ± 16%     -24.0%          9 ± 15%  sched_debug.cpu#37.cpu_load[2]
>        492 ± 17%     -21.3%        387 ± 24%  sched_debug.cpu#46.ttwu_count
>       3761 ± 11%     -23.9%       2863 ± 15%  sched_debug.cpu#93.curr->pid
>        570 ± 19%     +43.2%        816 ± 17%  sched_debug.cpu#86.ttwu_count
>       5279 ±  8%     +63.5%       8631 ± 33%  sched_debug.cpu#13.ttwu_count
>        377 ± 22%     -28.6%        269 ± 14%  sched_debug.cpu#46.ttwu_local
>       5396 ± 10%     +29.9%       7007 ± 14%  sched_debug.cpu#16.sched_goidle
>       1959 ± 12%     +36.9%       2683 ± 15%  numa-vmstat.node2.nr_slab_reclaimable
>       7839 ± 12%     +37.0%      10736 ± 15%  numa-meminfo.node2.SReclaimable
>          5 ± 15%     +66.7%          8 ±  9%  sched_debug.cfs_rq[33]:/.load
>          5 ± 25%     +47.8%          8 ± 10%  sched_debug.cfs_rq[37]:/.load
>          2 ±  0%     +87.5%          3 ± 34%  sched_debug.cpu#89.cpu_load[4]
>          5 ± 15%     +66.7%          8 ±  9%  sched_debug.cpu#33.load
>          6 ± 23%     +41.7%          8 ± 10%  sched_debug.cpu#37.load
>          8 ± 10%     -26.5%          6 ±  6%  sched_debug.cpu#51.cpu_load[1]
>       7300 ± 37%     +63.6%      11943 ± 16%  softirqs.TASKLET
>       2984 ±  6%     +43.1%       4271 ± 23%  sched_debug.cpu#20.ttwu_count
>        328 ±  4%     +40.5%        462 ± 25%  sched_debug.cpu#26.ttwu_local
>         10 ±  7%     -27.5%          7 ±  5%  sched_debug.cpu#43.cpu_load[3]
>          9 ±  8%     -30.8%          6 ±  6%  sched_debug.cpu#41.cpu_load[3]
>          9 ±  8%     -27.0%          6 ±  6%  sched_debug.cpu#41.cpu_load[4]
>         10 ± 14%     -32.5%          6 ±  6%  sched_debug.cpu#41.cpu_load[2]
>      16292 ±  6%     +42.8%      23260 ± 25%  sched_debug.cpu#13.nr_switches
>         14 ± 28%     +55.9%         23 ±  8%  sched_debug.cpu#99.cpu_load[0]
>          5 ±  8%     +28.6%          6 ± 12%  sched_debug.cpu#17.load
>         13 ±  7%     -23.1%         10 ± 12%  sched_debug.cpu#39.cpu_load[3]
>          7 ± 10%     -35.7%          4 ± 11%  sched_debug.cfs_rq[45]:/.runnable_load_avg
>       5076 ± 13%     -21.8%       3970 ± 11%  numa-vmstat.node0.nr_slab_unreclaimable
>      20306 ± 13%     -21.8%      15886 ± 11%  numa-meminfo.node0.SUnreclaim
>         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#45.cpu_load[3]
>         11 ± 11%     -29.5%          7 ± 14%  sched_debug.cpu#45.cpu_load[1]
>         10 ± 12%     -26.8%          7 ±  6%  sched_debug.cpu#44.cpu_load[1]
>         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#44.cpu_load[0]
>          7 ± 17%     +48.3%         10 ±  7%  sched_debug.cfs_rq[11]:/.runnable_load_avg
>         11 ± 12%     -34.1%          7 ± 11%  sched_debug.cpu#47.cpu_load[0]
>         10 ± 10%     -27.9%          7 ±  5%  sched_debug.cpu#47.cpu_load[1]
>         10 ±  8%     -26.8%          7 ± 11%  sched_debug.cpu#47.cpu_load[2]
>         10 ±  8%     -28.6%          7 ± 14%  sched_debug.cpu#43.cpu_load[0]
>         10 ± 10%     -27.9%          7 ± 10%  sched_debug.cpu#43.cpu_load[1]
>         10 ± 10%     -28.6%          7 ±  6%  sched_debug.cpu#43.cpu_load[2]
>      12940 ±  3%     +49.8%      19387 ± 35%  numa-meminfo.node2.Active(anon)
>       3235 ±  2%     +49.8%       4844 ± 35%  numa-vmstat.node2.nr_active_anon
>         17 ± 17%     +36.6%         24 ±  9%  sched_debug.cpu#97.cpu_load[2]
>      14725 ±  8%     +21.8%      17928 ± 11%  sched_debug.cpu#16.nr_switches
>        667 ± 10%     +45.3%        969 ± 22%  sched_debug.cpu#17.ttwu_local
>       3257 ±  5%     +22.4%       3988 ± 11%  sched_debug.cpu#118.curr->pid
>       3144 ± 15%     -20.7%       2493 ±  8%  sched_debug.cpu#95.curr->pid
>       2192 ± 11%     +50.9%       3308 ± 37%  sched_debug.cpu#18.ttwu_count
>          6 ± 11%     +37.5%          8 ± 19%  sched_debug.cfs_rq[22]:/.load
>         12 ±  5%     +27.1%         15 ±  8%  sched_debug.cpu#5.cpu_load[1]
>         11 ± 12%     -23.4%          9 ± 13%  sched_debug.cpu#37.cpu_load[3]
>          6 ± 11%     +37.5%          8 ± 19%  sched_debug.cpu#22.load
>          8 ±  8%     -25.0%          6 ±  0%  sched_debug.cpu#51.cpu_load[2]
>          7 ±  6%     -20.0%          6 ± 11%  sched_debug.cpu#55.cpu_load[3]
>         11 ±  9%     -17.4%          9 ±  9%  sched_debug.cpu#39.cpu_load[4]
>         12 ±  5%     -22.9%          9 ± 11%  sched_debug.cpu#38.cpu_load[3]
>        420 ± 13%     +43.0%        601 ±  9%  sched_debug.cpu#30.ttwu_local
>       1682 ± 14%     +38.5%       2329 ± 17%  numa-meminfo.node7.AnonPages
>        423 ± 13%     +37.0%        579 ± 16%  numa-vmstat.node7.nr_anon_pages
>         15 ± 13%     +41.9%         22 ±  5%  sched_debug.cpu#99.cpu_load[1]
>          6 ± 20%     +44.0%          9 ± 13%  sched_debug.cfs_rq[19]:/.runnable_load_avg
>          9 ±  4%     -24.3%          7 ±  0%  sched_debug.cpu#43.cpu_load[4]
>       6341 ±  7%     -19.6%       5100 ± 16%  sched_debug.cpu#43.curr->pid
>       2577 ± 11%     -11.9%       2270 ± 10%  sched_debug.cpu#33.ttwu_count
>         13 ±  6%     -18.5%         11 ± 12%  sched_debug.cpu#40.cpu_load[2]
>       4828 ±  6%     +23.8%       5979 ±  6%  sched_debug.cpu#34.curr->pid
>       4351 ± 12%     +33.9%       5824 ± 12%  sched_debug.cpu#36.curr->pid
>         10 ±  8%     -23.8%          8 ±  8%  sched_debug.cpu#37.cpu_load[4]
>         10 ± 14%     -28.6%          7 ±  6%  sched_debug.cpu#45.cpu_load[2]
>         17 ± 22%     +40.6%         24 ±  7%  sched_debug.cpu#97.cpu_load[1]
>         11 ±  9%     +21.3%         14 ±  5%  sched_debug.cpu#7.cpu_load[2]
>         10 ±  8%     -26.2%          7 ± 10%  sched_debug.cpu#36.cpu_load[4]
>      12853 ±  2%     +20.0%      15429 ± 11%  numa-meminfo.node2.AnonPages
>       4744 ±  8%     +30.8%       6204 ± 11%  sched_debug.cpu#35.curr->pid
>       3214 ±  2%     +20.0%       3856 ± 11%  numa-vmstat.node2.nr_anon_pages
>       6181 ±  6%     +24.9%       7718 ±  9%  sched_debug.cpu#13.curr->pid
>       6675 ± 23%     +27.5%       8510 ± 10%  sched_debug.cfs_rq[91]:/.tg_load_avg
>     171261 ±  5%     -22.2%     133177 ± 15%  numa-numastat.node0.local_node
>       6589 ± 21%     +29.3%       8522 ± 11%  sched_debug.cfs_rq[89]:/.tg_load_avg
>       6508 ± 20%     +28.0%       8331 ±  8%  sched_debug.cfs_rq[88]:/.tg_load_avg
>       6598 ± 22%     +29.2%       8525 ± 11%  sched_debug.cfs_rq[90]:/.tg_load_avg
>        590 ± 13%     -21.4%        464 ±  7%  sched_debug.cpu#105.ttwu_local
>     175392 ±  5%     -21.7%     137308 ± 14%  numa-numastat.node0.numa_hit
>         11 ±  6%     -18.2%          9 ±  7%  sched_debug.cpu#38.cpu_load[4]
>       6643 ± 23%     +27.4%       8465 ± 10%  sched_debug.cfs_rq[94]:/.tg_load_avg
>       6764 ±  7%     +13.8%       7695 ±  7%  sched_debug.cpu#12.curr->pid
>         29 ± 28%     +34.5%         39 ±  5%  sched_debug.cfs_rq[98]:/.tg_load_contrib
>       1776 ±  7%     +29.4%       2298 ± 13%  sched_debug.cpu#11.ttwu_local
>         13 ±  0%     -19.2%         10 ±  8%  sched_debug.cpu#40.cpu_load[3]
>          7 ±  5%     -17.2%          6 ±  0%  sched_debug.cpu#51.cpu_load[3]
>       7371 ± 20%     -18.0%       6045 ±  3%  sched_debug.cpu#1.sched_goidle
>      26560 ±  2%     +14.0%      30287 ±  7%  numa-meminfo.node2.Slab
>      16161 ±  6%      -9.4%      14646 ±  1%  sched_debug.cfs_rq[27]:/.avg->runnable_avg_sum
>        351 ±  6%      -9.3%        318 ±  1%  sched_debug.cfs_rq[27]:/.tg_runnable_contrib
>       7753 ± 27%     -22.9%       5976 ±  5%  sched_debug.cpu#2.sched_goidle
>       3828 ±  9%     +17.3%       4490 ±  6%  sched_debug.cpu#23.sched_goidle
>      23925 ±  2%     +23.0%      29419 ± 23%  numa-meminfo.node2.Active
>         47 ±  6%     -15.8%         40 ± 19%  sched_debug.cpu#42.cpu_load[1]
>        282 ±  5%      -9.7%        254 ±  7%  sched_debug.cfs_rq[109]:/.tg_runnable_contrib
>        349 ±  5%      -9.3%        317 ±  1%  sched_debug.cfs_rq[26]:/.tg_runnable_contrib
>       6941 ±  3%      +8.9%       7558 ±  7%  sched_debug.cpu#61.nr_switches
>      16051 ±  5%      -8.9%      14618 ±  1%  sched_debug.cfs_rq[26]:/.avg->runnable_avg_sum
>     238944 ±  3%      +9.2%     260958 ±  5%  numa-vmstat.node2.numa_local
>      12966 ±  5%      -9.5%      11732 ±  6%  sched_debug.cfs_rq[109]:/.avg->runnable_avg_sum
>       1004 ±  3%      +8.2%       1086 ±  4%  sched_debug.cpu#118.sched_goidle
>      20746 ±  4%      -8.4%      19000 ±  1%  sched_debug.cfs_rq[45]:/.avg->runnable_avg_sum
>        451 ±  4%      -8.3%        413 ±  1%  sched_debug.cfs_rq[45]:/.tg_runnable_contrib
>       3538 ±  4%     +17.2%       4147 ±  8%  sched_debug.cpu#26.ttwu_count
>         16 ±  9%     +13.8%         18 ±  2%  sched_debug.cpu#99.cpu_load[3]
>       1531 ±  0%     +11.3%       1704 ±  1%  numa-meminfo.node7.KernelStack
>       3569 ±  3%     +17.2%       4182 ± 10%  sched_debug.cpu#24.sched_goidle
>       1820 ±  3%     -12.5%       1594 ±  8%  slabinfo.taskstats.num_objs
>       1819 ±  3%     -12.4%       1594 ±  8%  slabinfo.taskstats.active_objs
>       4006 ±  5%     +19.1%       4769 ±  8%  sched_debug.cpu#17.sched_goidle
>      21412 ± 19%     -17.0%      17779 ±  3%  sched_debug.cpu#2.nr_switches
>         16 ±  9%     +24.2%         20 ±  4%  sched_debug.cpu#99.cpu_load[2]
>      10493 ±  7%     +13.3%      11890 ±  4%  sched_debug.cpu#23.nr_switches
>       1207 ±  2%     -46.9%        640 ±  4%  time.voluntary_context_switches
>
>                           time.voluntary_context_switches
>
>   1300 ++-----------*--*--------------------*-------------------------------+
>        *..*.*..*.. +      *.*..*..*.*..*..*     .*..*..*.  .*..*.*..*..     |
>   1200 ++         *                            *         *.            *.*..*
>   1100 ++                                                                   |
>        |                                                                    |
>   1000 ++                                                                   |
>        |                                                                    |
>    900 ++                                                                   |
>        |                                                                    |
>    800 ++                                                                   |
>    700 ++                                                                   |
>        O    O     O       O O  O       O  O O  O O       O       O          |
>    600 ++ O    O    O  O          O O                  O    O  O            |
>        |                                            O                       |
>    500 ++-------------------------------------------------------------------+
>
>         [*] bisect-good sample
>         [O] bisect-bad  sample
>
> To reproduce:
>
>         apt-get install ruby ruby-oj
>         git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
>         cd lkp-tests
>         bin/setup-local job.yaml # the job file attached in this email
>         bin/run-local   job.yaml
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>

Regards,
Kirill
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ