lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Thu, 21 Feb 2019 22:31:54 +0000
From:   Wei Yang <richard.weiyang@...il.com>
To:     Wei Yang <richardw.yang@...ux.intel.com>
Cc:     "Huang, Ying" <ying.huang@...el.com>,
        Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
        kernel test robot <rong.a.chen@...el.com>,
        Stephen Rothwell <sfr@...b.auug.org.au>,
        "Rafael J. Wysocki" <rafael.j.wysocki@...el.com>, lkp@...org,
        LKML <linux-kernel@...r.kernel.org>
Subject: Re: [LKP] [driver core] 570d020012: will-it-scale.per_thread_ops
 -12.2% regression

On Thu, Feb 21, 2019 at 03:53:13PM +0800, Wei Yang wrote:
>On Thu, Feb 21, 2019 at 03:18:22PM +0800, Huang, Ying wrote:
>>Greg Kroah-Hartman <gregkh@...uxfoundation.org> writes:
>>
>>> On Thu, Feb 21, 2019 at 11:10:49AM +0800, kernel test robot wrote:
>>>> On Tue, Feb 19, 2019 at 01:19:04PM +0100, Greg Kroah-Hartman wrote:
>>>> > On Tue, Feb 19, 2019 at 08:59:45AM +0800, Wei Yang wrote:
>>>> > > On Mon, Feb 18, 2019 at 03:54:42PM +0800, kernel test robot wrote:
>>>> > > >Greeting,
>>>> > > >
>>>> > > >FYI, we noticed a -12.2% regression of will-it-scale.per_thread_ops due to commit:
>>>> > > >
>>>> > > >
>>>> > > >commit: 570d0200123fb4f809aa2f6226e93a458d664d70 ("driver core: move device->knode_class to device_private")
>>>> > > >https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
>>>> > > >
>>>> > > 
>>>> > > This is interesting.
>>>> > > 
>>>> > > I didn't expect the move of this field will impact the performance.
>>>> > > 
>>>> > > The reason is struct device is a hotter memory than device->device_private?
>>>> > > 
>>>> > > >in testcase: will-it-scale
>>>> > > >on test machine: 288 threads Knights Mill with 80G memory
>>>> > > >with following parameters:
>>>> > > >
>>>> > > >	nr_task: 100%
>>>> > > >	mode: thread
>>>> > > >	test: unlink2
>>>> > > >	cpufreq_governor: performance
>>>> > > >
>>>> > > >test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
>>>> > > >test-url: https://github.com/antonblanchard/will-it-scale
>>>> > > >
>>>> > > >In addition to that, the commit also has significant impact on the following tests:
>>>> > > >
>>>> > > >+------------------+---------------------------------------------------------------+
>>>> > > >| testcase: change | will-it-scale: will-it-scale.per_thread_ops -29.9% regression |
>>>> > > >| test machine     | 288 threads Knights Mill with 80G memory                      |
>>>> > > >| test parameters  | cpufreq_governor=performance                                  |
>>>> > > >|                  | mode=thread                                                   |
>>>> > > >|                  | nr_task=100%                                                  |
>>>> > > >|                  | test=signal1                                                  |
>>>> > 
>>>> > Ok, I'm going to blame your testing system, or something here, and not
>>>> > the above patch.
>>>> > 
>>>> > All this test does is call raise(3).  That does not touch the driver
>>>> > core at all.
>>>> > 
>>>> > > >+------------------+---------------------------------------------------------------+
>>>> > > >| testcase: change | will-it-scale: will-it-scale.per_thread_ops -16.5% regression |
>>>> > > >| test machine     | 288 threads Knights Mill with 80G memory                      |
>>>> > > >| test parameters  | cpufreq_governor=performance                                  |
>>>> > > >|                  | mode=thread                                                   |
>>>> > > >|                  | nr_task=100%                                                  |
>>>> > > >|                  | test=open1                                                    |
>>>> > > >+------------------+---------------------------------------------------------------+
>>>> > 
>>>> > Same here, open1 just calls open/close a lot.  No driver core
>>>> > interaction at all there either.
>>>> > 
>>>> > So are you _sure_ this is the offending patch?
>>>> 
>>>> Hi Greg,
>>>> 
>>>> We did an experiment, recovered the layout of struct device. and we
>>>> found the regression is gone. I guess the regession is not from the
>>>> patch but related to the struct layout.
>>>> 
>>>> 
>>>> tests: 1
>>>> testcase/path_params/tbox_group/run: will-it-scale/performance-thread-100%-unlink2/lkp-knm01
>>>> 
>>>> 570d0200123fb4f8  a36dc70b810afe9183de2ea18f  
>>>> ----------------  --------------------------  
>>>>          %stddev      change         %stddev
>>>>              \          |                \  
>>>>     237096              14%     270789        will-it-scale.workload
>>>>        823              14%        939        will-it-scale.per_thread_ops
>>>> 
>>>> 
>>>> tests: 1
>>>> testcase/path_params/tbox_group/run: will-it-scale/performance-thread-100%-signal1/lkp-knm01
>>>> 
>>>> 570d0200123fb4f8  a36dc70b810afe9183de2ea18f  
>>>> ----------------  --------------------------  
>>>>          %stddev      change         %stddev
>>>>              \          |                \  
>>>>      93.51   3%        48%     138.53   3%  will-it-scale.time.user_time
>>>>        186              40%        261        will-it-scale.per_thread_ops
>>>>      53909              40%      75507        will-it-scale.workload
>>>> 
>>>> 
>>>> tests: 1
>>>> testcase/path_params/tbox_group/run: will-it-scale/performance-thread-100%-open1/lkp-knm01
>>>> 
>>>> 570d0200123fb4f8  a36dc70b810afe9183de2ea18f  
>>>> ----------------  --------------------------  
>>>>          %stddev      change         %stddev
>>>>              \          |                \  
>>>>     447722              22%     546258  10%  will-it-scale.time.involuntary_context_switches
>>>>     226995              19%     269751        will-it-scale.workload
>>>>        787              19%        936        will-it-scale.per_thread_ops
>>>> 
>>>> 
>>>> 
>>>> commit a36dc70b810afe9183de2ea18faa4c0939c139ac
>>>> Author: 0day robot <lkp@...el.com>
>>>> Date:   Wed Feb 20 14:21:19 2019 +0800
>>>> 
>>>>     backfile klist_node in struct device for debugging
>>>>     
>>>>     Signed-off-by: 0day robot <lkp@...el.com>
>>>> 
>>>> diff --git a/include/linux/device.h b/include/linux/device.h
>>>> index d0e452fd0bff2..31666cb72b3ba 100644
>>>> --- a/include/linux/device.h
>>>> +++ b/include/linux/device.h
>>>> @@ -1035,6 +1035,7 @@ struct device {
>>>>  	spinlock_t		devres_lock;
>>>>  	struct list_head	devres_head;
>>>>  
>>>> +	struct klist_node       knode_class_test_by_rongc;
>>>>  	struct class		*class;
>>>>  	const struct attribute_group **groups;	/* optional groups */
>>>
>>> While this is fun to worry about alignment and structure size of 'struct
>>> device' I find it odd given that the syscalls and userspace load of
>>> those test programs have nothing to do with 'struct device' at all.
>>>
>>> So I can work on fixing up the alignment of struct device, as that's a
>>> nice thing to do for systems with 30k of these in memory, but that
>>> shouldn't affect a workload of a constant string of signal calls.
>>
>>Hi, Greg,
>>
>>I don't think this is an issues of struct device.  As you said, struct
>>device isn't access much during test.  Struct device may share slab page
>>with some other data structures (signal related, or fd related (as in
>>some other test cases)), so that the alignment of these data structures
>>are affected, so caused the performance regression.
>>
>
>I didn't get the point here neither.
>
>slab allocator ask memory from page allocator Page by Page and split the page
>into pre-defined size. For example, 128B, 512B... Just as shown in
>/proc/slabinfo. 
>
>Per my understanding, each struct device / device_private will sits in its own
>aligned space. struct device would sits in 1K slab and struct device_private
>would sits in 256B slab, both before and after this patch if I am correct.
>

As Greg mentioned, device is embedded in other structure. My analysis
here is not correct. The change in size of device, may affect the size
of struct who wraps device.

In case this struct is allocated by kmem_cache, this may affect the
number of objects allocated each time, or even the number of pages
allocated each time.

This means with this patch may affect the system in the following two
aspects:

    * the times for struct allocation
    * the order of page it asks from page allocator

One place we may take a look is the /proc/slabinfo. To see whether the
change in size of device affect kmem_cach objects.

Hmm... while as Greg mentioned, those cases will not involve struct
device allocation. So the above aspects may not take effect?

>Hmm... I am just curious about how this alignment is affected. Maybe I lost
>some point?
>
>>Best Regards,
>>Huang, Ying
>>
>>> thanks,
>>>
>>> greg k-h
>
>-- 
>Wei Yang
>Help you, Help me

-- 
Wei Yang
Help you, Help me

Powered by blists - more mailing lists