[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <bdec481f-b402-64b6-75b0-350b370f3eac@suse.cz>
Date: Tue, 6 Mar 2018 08:55:57 +0100
From: Vlastimil Babka <vbabka@...e.cz>
To: Aaron Lu <aaron.lu@...el.com>
Cc: Michal Hocko <mhocko@...nel.org>, linux-mm@...ck.org,
linux-kernel@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>,
Huang Ying <ying.huang@...el.com>,
Dave Hansen <dave.hansen@...el.com>,
Kemi Wang <kemi.wang@...el.com>,
Tim Chen <tim.c.chen@...ux.intel.com>,
Andi Kleen <ak@...ux.intel.com>,
Mel Gorman <mgorman@...hsingularity.net>,
Matthew Wilcox <willy@...radead.org>,
David Rientjes <rientjes@...gle.com>
Subject: Re: [PATCH v4 3/3] mm/free_pcppages_bulk: prefetch buddy while not
holding lock
On 03/05/2018 12:41 PM, Aaron Lu wrote:
> On Fri, Mar 02, 2018 at 06:55:25PM +0100, Vlastimil Babka wrote:
>> On 03/01/2018 03:00 PM, Michal Hocko wrote:
>>>
>>> I am really surprised that this has such a big impact.
>>
>> It's even stranger to me. Struct page is 64 bytes these days, exactly a
>> a cache line. Unless that changed, Intel CPUs prefetched a "buddy" cache
>> line (that forms an aligned 128 bytes block with the one we touch).
>> Which is exactly a order-0 buddy struct page! Maybe that implicit
>> prefetching stopped at L2 and explicit goes all the way to L1, can't
>
> The Intel Architecture Optimization Manual section 7.3.2 says:
>
> prefetchT0 - fetch data into all cache levels
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: 1st, 2nd and 3rd level cache.
>
> prefetchT2 - fetch data into 2nd and 3rd level caches (identical to
> prefetchT1)
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: 2nd and 3rd level cache.
>
> prefetchNTA - fetch data into non-temporal cache close to the processor,
> minimizing cache pollution
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: must fetch into 3rd level cache with fast replacement.
>
> I tried 'prefetcht0' and 'prefetcht2' instead of the default
> 'prefetchNTA' on a 2 sockets Intel Skylake, the two ended up with about
> the same performance number as prefetchNTA. I had expected prefetchT0 to
> deliver a better score if it was indeed due to L1D since prefetchT2 will
> not place data into L1 while prefetchT0 will, but looks like it is not
> the case here.
>
> It feels more like the buddy cacheline isn't in any level of the caches
> without prefetch for some reason.
So the adjacent line prefetch might be disabled? Could you check bios or
the MSR mentioned in
https://software.intel.com/en-us/articles/disclosure-of-hw-prefetcher-control-on-some-intel-processors
>> remember. Would that make such a difference? It would be nice to do some
>> perf tests with cache counters to see what is really going on...
>
> Compare prefetchT2 to no-prefetch, I saw these metrics change:
>
> no-prefetch change prefetchT2 metrics
> \ \
> stddev stddev
> ------------------------------------------------------------------------
> 0.18 +0.0 0.18 perf-stat.branch-miss-rate%
> 8.268e+09 +3.8% 8.585e+09 perf-stat.branch-misses
> 2.333e+10 +4.7% 2.443e+10 perf-stat.cache-misses
> 2.402e+11 +5.0% 2.522e+11 perf-stat.cache-references
> 3.52 -1.1% 3.48 perf-stat.cpi
> 0.02 -0.0 0.01 ±3% perf-stat.dTLB-load-miss-rate%
> 8.677e+08 -7.3% 8.048e+08 ±3% perf-stat.dTLB-load-misses
> 1.18 +0.0 1.19 perf-stat.dTLB-store-miss-rate%
> 2.359e+10 +6.0% 2.502e+10 perf-stat.dTLB-store-misses
> 1.979e+12 +5.0% 2.078e+12 perf-stat.dTLB-stores
> 6.126e+09 +10.1% 6.745e+09 ±3% perf-stat.iTLB-load-misses
> 3464 -8.4% 3172 ±3% perf-stat.instructions-per-iTLB-miss
> 0.28 +1.1% 0.29 perf-stat.ipc
> 2.929e+09 +5.1% 3.077e+09 perf-stat.minor-faults
> 9.244e+09 +4.7% 9.681e+09 perf-stat.node-loads
> 2.491e+08 +5.8% 2.634e+08 perf-stat.node-store-misses
> 6.472e+09 +6.1% 6.869e+09 perf-stat.node-stores
> 2.929e+09 +5.1% 3.077e+09 perf-stat.page-faults
> 2182469 -4.2% 2090977 perf-stat.path-length
>
> Not sure if this is useful though...
Looks like most stats increased in absolute values as the work done
increased and this is a time-limited benchmark? Although number of
instructions (calculated from itlb misses and insns-per-itlb-miss) shows
less than 1% increase, so dunno. And the improvement comes from reduced
dTLB-load-misses? That makes no sense for order-0 buddy struct pages
which always share a page. And the memmap mapping should use huge pages.
BTW what is path-length?
Powered by blists - more mailing lists