[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <1420331974.8652.2.camel@intel.com>
Date: Sun, 04 Jan 2015 08:39:34 +0800
From: Huang Ying <ying.huang@...el.com>
To: Kirill Tkhai <tkhai@...dex.ru>
Cc: Kirill Tkhai <ktkhai@...allels.com>,
Ingo Molnar <mingo@...nel.org>,
LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>,
Wu Fengguang <fengguang.wu@...el.com>
Subject: Re: [LKP] [sched] a15b12ac36a: -46.9%
time.voluntary_context_switches +1.5% will-it-scale.per_process_ops
Hi, Kirill,
Sorry for late.
On Tue, 2014-12-23 at 11:57 +0300, Kirill Tkhai wrote:
> Hi, Huang,
>
> what do these digits mean? What test does?
>
> 23.12.2014, 08:16, "Huang Ying" <ying.huang@...el.com>:
> > FYI, we noticed the below changes on
> >
> > commit a15b12ac36ad4e7b856a4ae54937ae26a51aebad ("sched: Do not stop cpu in set_cpus_allowed_ptr() if task is not running")
> >
> > testbox/testcase/testparams: lkp-g5/will-it-scale/performance-lock1
> >
> > 1ba93d42727c4400 a15b12ac36ad4e7b856a4ae549
> > ---------------- --------------------------
Above is the good commit and the bad commit.
> > %stddev %change %stddev
> > \ | \
> > 1517261 ± 0% +1.5% 1539994 ± 0% will-it-scale.per_process_ops
We have basic description of data above, where %stddev is standard
deviation.
What's more do you want?
Best Regards,
Huang, Ying
> > 247 ± 30% +131.8% 573 ± 49% sched_debug.cpu#61.ttwu_count
> > 225 ± 22% +142.8% 546 ± 34% sched_debug.cpu#81.ttwu_local
> > 15115 ± 44% +37.3% 20746 ± 40% numa-meminfo.node7.Active
> > 1028 ± 38% +115.3% 2214 ± 36% sched_debug.cpu#16.ttwu_local
> > 2 ± 19% +133.3% 5 ± 43% sched_debug.cpu#89.cpu_load[3]
> > 21 ± 45% +88.2% 40 ± 23% sched_debug.cfs_rq[99]:/.tg_load_contrib
> > 414 ± 33% +98.6% 823 ± 28% sched_debug.cpu#81.ttwu_count
> > 4 ± 10% +88.2% 8 ± 12% sched_debug.cfs_rq[33]:/.runnable_load_avg
> > 22 ± 26% +80.9% 40 ± 24% sched_debug.cfs_rq[103]:/.tg_load_contrib
> > 7 ± 17% -41.4% 4 ± 25% sched_debug.cfs_rq[41]:/.load
> > 7 ± 17% -37.9% 4 ± 19% sched_debug.cpu#41.load
> > 3 ± 22% +106.7% 7 ± 10% sched_debug.cfs_rq[36]:/.runnable_load_avg
> > 174 ± 13% +48.7% 259 ± 31% sched_debug.cpu#112.ttwu_count
> > 4 ± 19% +88.9% 8 ± 5% sched_debug.cfs_rq[35]:/.runnable_load_avg
> > 260 ± 10% +55.6% 405 ± 26% numa-vmstat.node3.nr_anon_pages
> > 1042 ± 10% +56.0% 1626 ± 26% numa-meminfo.node3.AnonPages
> > 26 ± 22% +74.3% 45 ± 16% sched_debug.cfs_rq[65]:/.tg_load_contrib
> > 21 ± 43% +71.3% 37 ± 26% sched_debug.cfs_rq[100]:/.tg_load_contrib
> > 3686 ± 21% +40.2% 5167 ± 19% sched_debug.cpu#16.ttwu_count
> > 142 ± 9% +34.4% 191 ± 24% sched_debug.cpu#112.ttwu_local
> > 5 ± 18% +69.6% 9 ± 15% sched_debug.cfs_rq[35]:/.load
> > 2 ± 30% +100.0% 5 ± 37% sched_debug.cpu#106.cpu_load[1]
> > 3 ± 23% +100.0% 6 ± 48% sched_debug.cpu#106.cpu_load[2]
> > 5 ± 18% +69.6% 9 ± 15% sched_debug.cpu#35.load
> > 9 ± 20% +48.6% 13 ± 16% sched_debug.cfs_rq[7]:/.runnable_load_avg
> > 1727 ± 15% +43.9% 2484 ± 30% sched_debug.cpu#34.ttwu_local
> > 10 ± 17% -40.5% 6 ± 13% sched_debug.cpu#41.cpu_load[0]
> > 10 ± 14% -29.3% 7 ± 5% sched_debug.cpu#45.cpu_load[4]
> > 10 ± 17% -33.3% 7 ± 10% sched_debug.cpu#41.cpu_load[1]
> > 6121 ± 8% +56.7% 9595 ± 30% sched_debug.cpu#13.sched_goidle
> > 13 ± 8% -25.9% 10 ± 17% sched_debug.cpu#39.cpu_load[2]
> > 12 ± 16% -24.0% 9 ± 15% sched_debug.cpu#37.cpu_load[2]
> > 492 ± 17% -21.3% 387 ± 24% sched_debug.cpu#46.ttwu_count
> > 3761 ± 11% -23.9% 2863 ± 15% sched_debug.cpu#93.curr->pid
> > 570 ± 19% +43.2% 816 ± 17% sched_debug.cpu#86.ttwu_count
> > 5279 ± 8% +63.5% 8631 ± 33% sched_debug.cpu#13.ttwu_count
> > 377 ± 22% -28.6% 269 ± 14% sched_debug.cpu#46.ttwu_local
> > 5396 ± 10% +29.9% 7007 ± 14% sched_debug.cpu#16.sched_goidle
> > 1959 ± 12% +36.9% 2683 ± 15% numa-vmstat.node2.nr_slab_reclaimable
> > 7839 ± 12% +37.0% 10736 ± 15% numa-meminfo.node2.SReclaimable
> > 5 ± 15% +66.7% 8 ± 9% sched_debug.cfs_rq[33]:/.load
> > 5 ± 25% +47.8% 8 ± 10% sched_debug.cfs_rq[37]:/.load
> > 2 ± 0% +87.5% 3 ± 34% sched_debug.cpu#89.cpu_load[4]
> > 5 ± 15% +66.7% 8 ± 9% sched_debug.cpu#33.load
> > 6 ± 23% +41.7% 8 ± 10% sched_debug.cpu#37.load
> > 8 ± 10% -26.5% 6 ± 6% sched_debug.cpu#51.cpu_load[1]
> > 7300 ± 37% +63.6% 11943 ± 16% softirqs.TASKLET
> > 2984 ± 6% +43.1% 4271 ± 23% sched_debug.cpu#20.ttwu_count
> > 328 ± 4% +40.5% 462 ± 25% sched_debug.cpu#26.ttwu_local
> > 10 ± 7% -27.5% 7 ± 5% sched_debug.cpu#43.cpu_load[3]
> > 9 ± 8% -30.8% 6 ± 6% sched_debug.cpu#41.cpu_load[3]
> > 9 ± 8% -27.0% 6 ± 6% sched_debug.cpu#41.cpu_load[4]
> > 10 ± 14% -32.5% 6 ± 6% sched_debug.cpu#41.cpu_load[2]
> > 16292 ± 6% +42.8% 23260 ± 25% sched_debug.cpu#13.nr_switches
> > 14 ± 28% +55.9% 23 ± 8% sched_debug.cpu#99.cpu_load[0]
> > 5 ± 8% +28.6% 6 ± 12% sched_debug.cpu#17.load
> > 13 ± 7% -23.1% 10 ± 12% sched_debug.cpu#39.cpu_load[3]
> > 7 ± 10% -35.7% 4 ± 11% sched_debug.cfs_rq[45]:/.runnable_load_avg
> > 5076 ± 13% -21.8% 3970 ± 11% numa-vmstat.node0.nr_slab_unreclaimable
> > 20306 ± 13% -21.8% 15886 ± 11% numa-meminfo.node0.SUnreclaim
> > 10 ± 10% -28.6% 7 ± 6% sched_debug.cpu#45.cpu_load[3]
> > 11 ± 11% -29.5% 7 ± 14% sched_debug.cpu#45.cpu_load[1]
> > 10 ± 12% -26.8% 7 ± 6% sched_debug.cpu#44.cpu_load[1]
> > 10 ± 10% -28.6% 7 ± 6% sched_debug.cpu#44.cpu_load[0]
> > 7 ± 17% +48.3% 10 ± 7% sched_debug.cfs_rq[11]:/.runnable_load_avg
> > 11 ± 12% -34.1% 7 ± 11% sched_debug.cpu#47.cpu_load[0]
> > 10 ± 10% -27.9% 7 ± 5% sched_debug.cpu#47.cpu_load[1]
> > 10 ± 8% -26.8% 7 ± 11% sched_debug.cpu#47.cpu_load[2]
> > 10 ± 8% -28.6% 7 ± 14% sched_debug.cpu#43.cpu_load[0]
> > 10 ± 10% -27.9% 7 ± 10% sched_debug.cpu#43.cpu_load[1]
> > 10 ± 10% -28.6% 7 ± 6% sched_debug.cpu#43.cpu_load[2]
> > 12940 ± 3% +49.8% 19387 ± 35% numa-meminfo.node2.Active(anon)
> > 3235 ± 2% +49.8% 4844 ± 35% numa-vmstat.node2.nr_active_anon
> > 17 ± 17% +36.6% 24 ± 9% sched_debug.cpu#97.cpu_load[2]
> > 14725 ± 8% +21.8% 17928 ± 11% sched_debug.cpu#16.nr_switches
> > 667 ± 10% +45.3% 969 ± 22% sched_debug.cpu#17.ttwu_local
> > 3257 ± 5% +22.4% 3988 ± 11% sched_debug.cpu#118.curr->pid
> > 3144 ± 15% -20.7% 2493 ± 8% sched_debug.cpu#95.curr->pid
> > 2192 ± 11% +50.9% 3308 ± 37% sched_debug.cpu#18.ttwu_count
> > 6 ± 11% +37.5% 8 ± 19% sched_debug.cfs_rq[22]:/.load
> > 12 ± 5% +27.1% 15 ± 8% sched_debug.cpu#5.cpu_load[1]
> > 11 ± 12% -23.4% 9 ± 13% sched_debug.cpu#37.cpu_load[3]
> > 6 ± 11% +37.5% 8 ± 19% sched_debug.cpu#22.load
> > 8 ± 8% -25.0% 6 ± 0% sched_debug.cpu#51.cpu_load[2]
> > 7 ± 6% -20.0% 6 ± 11% sched_debug.cpu#55.cpu_load[3]
> > 11 ± 9% -17.4% 9 ± 9% sched_debug.cpu#39.cpu_load[4]
> > 12 ± 5% -22.9% 9 ± 11% sched_debug.cpu#38.cpu_load[3]
> > 420 ± 13% +43.0% 601 ± 9% sched_debug.cpu#30.ttwu_local
> > 1682 ± 14% +38.5% 2329 ± 17% numa-meminfo.node7.AnonPages
> > 423 ± 13% +37.0% 579 ± 16% numa-vmstat.node7.nr_anon_pages
> > 15 ± 13% +41.9% 22 ± 5% sched_debug.cpu#99.cpu_load[1]
> > 6 ± 20% +44.0% 9 ± 13% sched_debug.cfs_rq[19]:/.runnable_load_avg
> > 9 ± 4% -24.3% 7 ± 0% sched_debug.cpu#43.cpu_load[4]
> > 6341 ± 7% -19.6% 5100 ± 16% sched_debug.cpu#43.curr->pid
> > 2577 ± 11% -11.9% 2270 ± 10% sched_debug.cpu#33.ttwu_count
> > 13 ± 6% -18.5% 11 ± 12% sched_debug.cpu#40.cpu_load[2]
> > 4828 ± 6% +23.8% 5979 ± 6% sched_debug.cpu#34.curr->pid
> > 4351 ± 12% +33.9% 5824 ± 12% sched_debug.cpu#36.curr->pid
> > 10 ± 8% -23.8% 8 ± 8% sched_debug.cpu#37.cpu_load[4]
> > 10 ± 14% -28.6% 7 ± 6% sched_debug.cpu#45.cpu_load[2]
> > 17 ± 22% +40.6% 24 ± 7% sched_debug.cpu#97.cpu_load[1]
> > 11 ± 9% +21.3% 14 ± 5% sched_debug.cpu#7.cpu_load[2]
> > 10 ± 8% -26.2% 7 ± 10% sched_debug.cpu#36.cpu_load[4]
> > 12853 ± 2% +20.0% 15429 ± 11% numa-meminfo.node2.AnonPages
> > 4744 ± 8% +30.8% 6204 ± 11% sched_debug.cpu#35.curr->pid
> > 3214 ± 2% +20.0% 3856 ± 11% numa-vmstat.node2.nr_anon_pages
> > 6181 ± 6% +24.9% 7718 ± 9% sched_debug.cpu#13.curr->pid
> > 6675 ± 23% +27.5% 8510 ± 10% sched_debug.cfs_rq[91]:/.tg_load_avg
> > 171261 ± 5% -22.2% 133177 ± 15% numa-numastat.node0.local_node
> > 6589 ± 21% +29.3% 8522 ± 11% sched_debug.cfs_rq[89]:/.tg_load_avg
> > 6508 ± 20% +28.0% 8331 ± 8% sched_debug.cfs_rq[88]:/.tg_load_avg
> > 6598 ± 22% +29.2% 8525 ± 11% sched_debug.cfs_rq[90]:/.tg_load_avg
> > 590 ± 13% -21.4% 464 ± 7% sched_debug.cpu#105.ttwu_local
> > 175392 ± 5% -21.7% 137308 ± 14% numa-numastat.node0.numa_hit
> > 11 ± 6% -18.2% 9 ± 7% sched_debug.cpu#38.cpu_load[4]
> > 6643 ± 23% +27.4% 8465 ± 10% sched_debug.cfs_rq[94]:/.tg_load_avg
> > 6764 ± 7% +13.8% 7695 ± 7% sched_debug.cpu#12.curr->pid
> > 29 ± 28% +34.5% 39 ± 5% sched_debug.cfs_rq[98]:/.tg_load_contrib
> > 1776 ± 7% +29.4% 2298 ± 13% sched_debug.cpu#11.ttwu_local
> > 13 ± 0% -19.2% 10 ± 8% sched_debug.cpu#40.cpu_load[3]
> > 7 ± 5% -17.2% 6 ± 0% sched_debug.cpu#51.cpu_load[3]
> > 7371 ± 20% -18.0% 6045 ± 3% sched_debug.cpu#1.sched_goidle
> > 26560 ± 2% +14.0% 30287 ± 7% numa-meminfo.node2.Slab
> > 16161 ± 6% -9.4% 14646 ± 1% sched_debug.cfs_rq[27]:/.avg->runnable_avg_sum
> > 351 ± 6% -9.3% 318 ± 1% sched_debug.cfs_rq[27]:/.tg_runnable_contrib
> > 7753 ± 27% -22.9% 5976 ± 5% sched_debug.cpu#2.sched_goidle
> > 3828 ± 9% +17.3% 4490 ± 6% sched_debug.cpu#23.sched_goidle
> > 23925 ± 2% +23.0% 29419 ± 23% numa-meminfo.node2.Active
> > 47 ± 6% -15.8% 40 ± 19% sched_debug.cpu#42.cpu_load[1]
> > 282 ± 5% -9.7% 254 ± 7% sched_debug.cfs_rq[109]:/.tg_runnable_contrib
> > 349 ± 5% -9.3% 317 ± 1% sched_debug.cfs_rq[26]:/.tg_runnable_contrib
> > 6941 ± 3% +8.9% 7558 ± 7% sched_debug.cpu#61.nr_switches
> > 16051 ± 5% -8.9% 14618 ± 1% sched_debug.cfs_rq[26]:/.avg->runnable_avg_sum
> > 238944 ± 3% +9.2% 260958 ± 5% numa-vmstat.node2.numa_local
> > 12966 ± 5% -9.5% 11732 ± 6% sched_debug.cfs_rq[109]:/.avg->runnable_avg_sum
> > 1004 ± 3% +8.2% 1086 ± 4% sched_debug.cpu#118.sched_goidle
> > 20746 ± 4% -8.4% 19000 ± 1% sched_debug.cfs_rq[45]:/.avg->runnable_avg_sum
> > 451 ± 4% -8.3% 413 ± 1% sched_debug.cfs_rq[45]:/.tg_runnable_contrib
> > 3538 ± 4% +17.2% 4147 ± 8% sched_debug.cpu#26.ttwu_count
> > 16 ± 9% +13.8% 18 ± 2% sched_debug.cpu#99.cpu_load[3]
> > 1531 ± 0% +11.3% 1704 ± 1% numa-meminfo.node7.KernelStack
> > 3569 ± 3% +17.2% 4182 ± 10% sched_debug.cpu#24.sched_goidle
> > 1820 ± 3% -12.5% 1594 ± 8% slabinfo.taskstats.num_objs
> > 1819 ± 3% -12.4% 1594 ± 8% slabinfo.taskstats.active_objs
> > 4006 ± 5% +19.1% 4769 ± 8% sched_debug.cpu#17.sched_goidle
> > 21412 ± 19% -17.0% 17779 ± 3% sched_debug.cpu#2.nr_switches
> > 16 ± 9% +24.2% 20 ± 4% sched_debug.cpu#99.cpu_load[2]
> > 10493 ± 7% +13.3% 11890 ± 4% sched_debug.cpu#23.nr_switches
> > 1207 ± 2% -46.9% 640 ± 4% time.voluntary_context_switches
> >
> > time.voluntary_context_switches
> >
> > 1300 ++-----------*--*--------------------*-------------------------------+
> > *..*.*..*.. + *.*..*..*.*..*..* .*..*..*. .*..*.*..*.. |
> > 1200 ++ * * *. *.*..*
> > 1100 ++ |
> > | |
> > 1000 ++ |
> > | |
> > 900 ++ |
> > | |
> > 800 ++ |
> > 700 ++ |
> > O O O O O O O O O O O O O |
> > 600 ++ O O O O O O O O O |
> > | O |
> > 500 ++-------------------------------------------------------------------+
> >
> > [*] bisect-good sample
> > [O] bisect-bad sample
> >
> > To reproduce:
> >
> > apt-get install ruby ruby-oj
> > git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
> > cd lkp-tests
> > bin/setup-local job.yaml # the job file attached in this email
> > bin/run-local job.yaml
> >
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> >
>
> Regards,
> Kirill
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists