lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date: Tue, 12 Mar 2024 16:02:19 +0530
From: Sandipan Das <sandipan.das@....com>
To: Ian Rogers <irogers@...gle.com>
Cc: linux-perf-users@...r.kernel.org, linux-kernel@...r.kernel.org,
 peterz@...radead.org, mingo@...hat.com, acme@...nel.org,
 namhyung@...nel.org, mark.rutland@....com,
 alexander.shishkin@...ux.intel.com, jolsa@...nel.org,
 adrian.hunter@...el.com, eranian@...gle.com, ravi.bangoria@....com,
 ananth.narayan@....com
Subject: Re: [PATCH 3/4] perf vendor events amd: Add Zen 5 metrics

On 3/12/2024 12:16 AM, Ian Rogers wrote:
> On Sun, Mar 10, 2024 at 10:24 PM Sandipan Das <sandipan.das@....com> wrote:
>>
>> Add metrics taken from Section 1.2 "Performance Measurement" of the
>> Performance Monitor Counters for AMD Family 1Ah Model 00h-0Fh Processors
>> document available at the link below.
>>
>> The recommended metrics are sourced from Table 1 "Guidance for Common
>> Performance Statistics with Complex Event Selects".
>>
>> The pipeline utilization metrics are sourced from Table 2 "Guidance
>> for Pipeline Utilization Analysis Statistics". These are useful for
>> finding performance bottlenecks by analyzing activity at different
>> stages of the pipeline. There are metric groups available for Level 1
>> and Level 2 analysis.
>>
>> Link: https://bugzilla.kernel.org/attachment.cgi?id=305974
>> Signed-off-by: Sandipan Das <sandipan.das@....com>
> 
> Could you consider reviewing:
> https://lore.kernel.org/lkml/20240301184737.2660108-1-irogers@google.com/
> 

Sure. I did start going over the series a few days back.

>> ---
>>  .../pmu-events/arch/x86/amdzen5/pipeline.json |  98 +++++
>>  .../arch/x86/amdzen5/recommended.json         | 357 ++++++++++++++++++
>>  2 files changed, 455 insertions(+)
>>  create mode 100644 tools/perf/pmu-events/arch/x86/amdzen5/pipeline.json
>>  create mode 100644 tools/perf/pmu-events/arch/x86/amdzen5/recommended.json
>>
>> diff --git a/tools/perf/pmu-events/arch/x86/amdzen5/pipeline.json b/tools/perf/pmu-events/arch/x86/amdzen5/pipeline.json
>> new file mode 100644
>> index 000000000000..36dc76b793ae
>> --- /dev/null
>> +++ b/tools/perf/pmu-events/arch/x86/amdzen5/pipeline.json
>> @@ -0,0 +1,98 @@
>> +[
>> +  {
>> +    "MetricName": "total_dispatch_slots",
>> +    "BriefDescription": "Total dispatch slots (up to 8 instructions can be dispatched in each cycle).",
>> +    "MetricExpr": "8 * ls_not_halted_cyc"
> 
> Should the unit be slots?
> 

Yes, I'll update this.

>> +  },
>> +  {
>> +    "MetricName": "frontend_bound",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because the frontend did not supply enough instructions/ops.",
> 
> Given the output is in percent, is fraction an accurate description?
> Wouldn't "percentage" be better? This issue repeats below, but I'll
> just highlight the first instance.
> 

Will fix this and other instances.

>> +    "MetricExpr": "d_ratio(de_no_dispatch_per_slot.no_ops_from_frontend, total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL1",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots" ?
> 
>> +  },
>> +  {
>> +    "MetricName": "bad_speculation",
>> +    "BriefDescription": "Fraction of dispatched ops that did not retire.",
>> +    "MetricExpr": "d_ratio(de_src_op_disp.all - ex_ret_ops, total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL1",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% ops"
> 
>> +  },
>> +  {
>> +    "MetricName": "backend_bound",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because of backend stalls.",
>> +    "MetricExpr": "d_ratio(de_no_dispatch_per_slot.backend_stalls, total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL1",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots"
> 
>> +  },
>> +  {
>> +    "MetricName": "smt_contention",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because the other thread was selected.",
>> +    "MetricExpr": "d_ratio(de_no_dispatch_per_slot.smt_contention, total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL1",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots"
> 
>> +  },
>> +  {
>> +    "MetricName": "retiring",
>> +    "BriefDescription": "Fraction of dispatch slots used by ops that retired.",
>> +    "MetricExpr": "d_ratio(ex_ret_ops, total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL1",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots"
> 
>> +  },
>> +  {
>> +    "MetricName": "frontend_bound_latency",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because of a latency bottleneck in the frontend (such as instruction cache or TLB misses).",
>> +    "MetricExpr": "d_ratio((8 * cpu@...no_dispatch_per_slot.no_ops_from_frontend\\,cmask\\=0x8@), total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL2;frontend_bound_group",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots"
> 
>> +  },
>> +  {
>> +    "MetricName": "frontend_bound_bandwidth",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because of a bandwidth bottleneck in the frontend (such as decode or op cache fetch bandwidth).",
>> +    "MetricExpr": "d_ratio(de_no_dispatch_per_slot.no_ops_from_frontend - (8 * cpu@...no_dispatch_per_slot.no_ops_from_frontend\\,cmask\\=0x8@), total_dispatch_slots)",
>> +    "MetricGroup": "PipelineL2;frontend_bound_group",
>> +    "ScaleUnit": "100%"
> 
> Perhaps "100% slots"
> 
> It seems unexpected that a latency (above) and  bandwidth metric would
> be reporting a percentage, perhaps this needs capturing in the metric
> name.
> 
> Same issues repeat below...
> 

Will update these.

>> +  },
>> +  {
>> +    "MetricName": "bad_speculation_mispredicts",
>> +    "BriefDescription": "Fraction of dispatched ops that were flushed due to branch mispredicts.",
>> +    "MetricExpr": "d_ratio(bad_speculation * ex_ret_brn_misp, ex_ret_brn_misp + bp_redirects.resync)",
>> +    "MetricGroup": "PipelineL2;bad_speculation_group",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "bad_speculation_pipeline_restarts",
>> +    "BriefDescription": "Fraction of dispatched ops that were flushed due to pipeline restarts (resyncs).",
>> +    "MetricExpr": "d_ratio(bad_speculation * bp_redirects.resync, ex_ret_brn_misp + bp_redirects.resync)",
>> +    "MetricGroup": "PipelineL2;bad_speculation_group",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "backend_bound_memory",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because of stalls due to the memory subsystem.",
>> +    "MetricExpr": "backend_bound * d_ratio(ex_no_retire.load_not_complete, ex_no_retire.not_complete)",
>> +    "MetricGroup": "PipelineL2;backend_bound_group",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "backend_bound_cpu",
>> +    "BriefDescription": "Fraction of dispatch slots that remained unused because of stalls not related to the memory subsystem.",
>> +    "MetricExpr": "backend_bound * (1 - d_ratio(ex_no_retire.load_not_complete, ex_no_retire.not_complete))",
>> +    "MetricGroup": "PipelineL2;backend_bound_group",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "retiring_fastpath",
>> +    "BriefDescription": "Fraction of dispatch slots used by fastpath ops that retired.",
>> +    "MetricExpr": "retiring * (1 - d_ratio(ex_ret_ucode_ops, ex_ret_ops))",
>> +    "MetricGroup": "PipelineL2;retiring_group",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "retiring_microcode",
>> +    "BriefDescription": "Fraction of dispatch slots used by microcode ops that retired.",
>> +    "MetricExpr": "retiring * d_ratio(ex_ret_ucode_ops, ex_ret_ops)",
>> +    "MetricGroup": "PipelineL2;retiring_group",
>> +    "ScaleUnit": "100%"
>> +  }
>> +]
>> diff --git a/tools/perf/pmu-events/arch/x86/amdzen5/recommended.json b/tools/perf/pmu-events/arch/x86/amdzen5/recommended.json
>> new file mode 100644
>> index 000000000000..986f8b2b2d5b
>> --- /dev/null
>> +++ b/tools/perf/pmu-events/arch/x86/amdzen5/recommended.json
>> @@ -0,0 +1,357 @@
>> +[
>> +  {
>> +    "MetricName": "branch_misprediction_ratio",
>> +    "BriefDescription": "Execution-time branch misprediction ratio (non-speculative).",
> 
> Is ratio or rate better?
> ```
> $ grep -r MetricName tools/perf/pmu-events/arch/| grep _rate |wc -l
> 246
> $ grep -r MetricName tools/perf/pmu-events/arch/| grep _ratio |wc -l
> 135
> ```
> 

If having both helps, I can add a rate metric as well.

>> +    "MetricExpr": "d_ratio(ex_ret_brn_misp, ex_ret_brn)",
>> +    "MetricGroup": "branch_prediction",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "all_data_cache_accesses_pti",
>> +    "BriefDescription": "All data cache accesses per thousand instructions.",
>> +    "MetricExpr": "ls_dispatch.all / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
> 
> Perhaps "1e3instructions", and below.
> 

Will fix these.

>> +  },
>> +  {
>> +    "MetricName": "all_l2_cache_accesses_pti",
>> +    "BriefDescription": "All L2 cache accesses per thousand instructions.",
>> +    "MetricExpr": "(l2_request_g1.all_no_prefetch + l2_pf_hit_l2.l2_hwpf + l2_pf_miss_l2_hit_l3.l2_hwpf + l2_pf_miss_l2_l3.l2_hwpf) / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_accesses_from_l1_ic_misses_pti",
>> +    "BriefDescription": "L2 cache accesses from L1 instruction cache misses (including prefetch) per thousand instructions.",
>> +    "MetricExpr": "l2_request_g1.cacheable_ic_read / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_accesses_from_l1_dc_misses_pti",
>> +    "BriefDescription": "L2 cache accesses from L1 data cache misses (including prefetch) per thousand instructions.",
>> +    "MetricExpr": "l2_request_g1.all_dc / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_accesses_from_l2_hwpf_pti",
>> +    "BriefDescription": "L2 cache accesses from L2 cache hardware prefetcher per thousand instructions.",
>> +    "MetricExpr": "(l2_pf_hit_l2.l1_dc_l2_hwpf + l2_pf_miss_l2_hit_l3.l1_dc_l2_hwpf + l2_pf_miss_l2_l3.l1_dc_l2_hwpf) / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "all_l2_cache_misses_pti",
>> +    "BriefDescription": "All L2 cache misses per thousand instructions.",
>> +    "MetricExpr": "(l2_cache_req_stat.ic_dc_miss_in_l2 + l2_pf_miss_l2_hit_l3.l2_hwpf + l2_pf_miss_l2_l3.l2_hwpf) / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_misses_from_l1_ic_miss_pti",
>> +    "BriefDescription": "L2 cache misses from L1 instruction cache misses per thousand instructions.",
>> +    "MetricExpr": "l2_cache_req_stat.ic_fill_miss / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_misses_from_l1_dc_miss_pti",
>> +    "BriefDescription": "L2 cache misses from L1 data cache misses per thousand instructions.",
>> +    "MetricExpr": "l2_cache_req_stat.ls_rd_blk_c / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_misses_from_l2_hwpf_pti",
>> +    "BriefDescription": "L2 cache misses from L2 cache hardware prefetcher per thousand instructions.",
>> +    "MetricExpr": "(l2_pf_miss_l2_hit_l3.l1_dc_l2_hwpf + l2_pf_miss_l2_l3.l1_dc_l2_hwpf) / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "all_l2_cache_hits_pti",
>> +    "BriefDescription": "All L2 cache hits per thousand instructions.",
>> +    "MetricExpr": "(l2_cache_req_stat.ic_dc_hit_in_l2 + l2_pf_hit_l2.l2_hwpf) / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_hits_from_l1_ic_miss_pti",
>> +    "BriefDescription": "L2 cache hits from L1 instruction cache misses per thousand instructions.",
>> +    "MetricExpr": "l2_cache_req_stat.ic_hit_in_l2 / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_hits_from_l1_dc_miss_pti",
>> +    "BriefDescription": "L2 cache hits from L1 data cache misses per thousand instructions.",
>> +    "MetricExpr": "l2_cache_req_stat.dc_hit_in_l2 / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_cache_hits_from_l2_hwpf_pti",
>> +    "BriefDescription": "L2 cache hits from L2 cache hardware prefetcher per thousand instructions.",
>> +    "MetricExpr": "l2_pf_hit_l2.l1_dc_l2_hwpf / instructions",
>> +    "MetricGroup": "l2_cache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l3_cache_accesses",
>> +    "BriefDescription": "L3 cache accesses.",
>> +    "MetricExpr": "l3_lookup_state.all_coherent_accesses_to_l3",
>> +    "MetricGroup": "l3_cache"
>> +  },
>> +  {
>> +    "MetricName": "l3_misses",
>> +    "BriefDescription": "L3 misses (including cacheline state change requests).",
> 
> local vs remote?
> 

Unfortunately, the underlying event doesn't provide a local vs. remote breakdown.

>> +    "MetricExpr": "l3_lookup_state.l3_miss",
>> +    "MetricGroup": "l3_cache"
>> +  },
>> +  {
>> +    "MetricName": "l3_read_miss_latency",
>> +    "BriefDescription": "Average L3 read miss latency (in core clocks).",
>> +    "MetricExpr": "(l3_xi_sampled_latency.all * 10) / l3_xi_sampled_latency_requests.all",
>> +    "MetricGroup": "l3_cache",
>> +    "ScaleUnit": "1core clocks"
>> +  },
>> +  {
>> +    "MetricName": "l3_read_miss_latency_for_local_dram",
>> +    "BriefDescription": "Average L3 read miss latency (in core clocks) for local DRAM.",
>> +    "MetricExpr": "(l3_xi_sampled_latency.dram_near * 10) / l3_xi_sampled_latency_requests.dram_near",
>> +    "MetricGroup": "l3_cache",
>> +    "ScaleUnit": "1core clocks"
> 
> "core clocks" isn't defined in the attached documentation. How can one
> look up the different clock types? If "core" is basically all clock
> types in the metrics then consider dropping "core" here.
> 

The correct unit for these metrics should be nanoseconds. I'll fix them.

>> +  },
>> +  {
>> +    "MetricName": "l3_read_miss_latency_for_remote_dram",
>> +    "BriefDescription": "Average L3 read miss latency (in core clocks) for remote DRAM.",
>> +    "MetricExpr": "(l3_xi_sampled_latency.dram_far * 10) / l3_xi_sampled_latency_requests.dram_far",
>> +    "MetricGroup": "l3_cache",
>> +    "ScaleUnit": "1core clocks"
>> +  },
>> +  {
>> +    "MetricName": "op_cache_fetch_miss_ratio",
>> +    "BriefDescription": "Op cache miss ratio for all fetches.",
>> +    "MetricExpr": "d_ratio(op_cache_hit_miss.op_cache_miss, op_cache_hit_miss.all_op_cache_accesses)",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "ic_fetch_miss_ratio",
>> +    "BriefDescription": "Instruction cache miss ratio for all fetches. An instruction cache miss will not be counted by this metric if it is an OC hit.",
>> +    "MetricExpr": "d_ratio(ic_tag_hit_miss.instruction_cache_miss, ic_tag_hit_miss.all_instruction_cache_accesses)",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "l1_data_cache_fills_from_memory_pti",
>> +    "BriefDescription": "L1 data cache fills from DRAM or MMIO in any NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_any_fills_from_sys.dram_io_all / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_data_cache_fills_from_remote_node_pti",
>> +    "BriefDescription": "L1 data cache fills from a different NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_any_fills_from_sys.far_all / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_data_cache_fills_from_same_ccx_pti",
>> +    "BriefDescription": "L1 data cache fills from within the same CCX per thousand instructions.",
>> +    "MetricExpr": "ls_any_fills_from_sys.local_all / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_data_cache_fills_from_different_ccx_pti",
>> +    "BriefDescription": "L1 data cache fills from another CCX cache in any NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_any_fills_from_sys.remote_cache / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "all_l1_data_cache_fills_pti",
>> +    "BriefDescription": "All L1 data cache fills per thousand instructions.",
>> +    "MetricExpr": "ls_any_fills_from_sys.all / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_local_l2_pti",
>> +    "BriefDescription": "L1 demand data cache fills from local L2 cache per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.local_l2 / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_same_ccx_pti",
>> +    "BriefDescription": "L1 demand data cache fills from within the same CCX per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.local_ccx / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_near_cache_pti",
>> +    "BriefDescription": "L1 demand data cache fills from another CCX cache in the same NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.near_cache / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_near_memory_pti",
>> +    "BriefDescription": "L1 demand data cache fills from DRAM or MMIO in the same NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.dram_io_near / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_far_cache_pti",
>> +    "BriefDescription": "L1 demand data cache fills from another CCX cache in a different NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.far_cache / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_demand_data_cache_fills_from_far_memory_pti",
>> +    "BriefDescription": "L1 demand data cache fills from DRAM or MMIO in a different NUMA node per thousand instructions.",
>> +    "MetricExpr": "ls_dmnd_fills_from_sys.dram_io_far / instructions",
>> +    "MetricGroup": "l1_dcache",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_itlb_misses_pti",
>> +    "BriefDescription": "L1 instruction TLB misses per thousand instructions.",
>> +    "MetricExpr": "(bp_l1_tlb_miss_l2_tlb_hit + bp_l1_tlb_miss_l2_tlb_miss.all) / instructions",
>> +    "MetricGroup": "tlb",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_itlb_misses_pti",
>> +    "BriefDescription": "L2 instruction TLB misses and instruction page walks per thousand instructions.",
>> +    "MetricExpr": "bp_l1_tlb_miss_l2_tlb_miss.all / instructions",
>> +    "MetricGroup": "tlb",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l1_dtlb_misses_pti",
>> +    "BriefDescription": "L1 data TLB misses per thousand instructions.",
>> +    "MetricExpr": "ls_l1_d_tlb_miss.all / instructions",
>> +    "MetricGroup": "tlb",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "l2_dtlb_misses_pti",
>> +    "BriefDescription": "L2 data TLB misses and data page walks per thousand instructions.",
>> +    "MetricExpr": "ls_l1_d_tlb_miss.all_l2_miss / instructions",
>> +    "MetricGroup": "tlb",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "all_tlbs_flushed_pti",
>> +    "BriefDescription": "All TLBs flushed per thousand instructions.",
>> +    "MetricExpr": "ls_tlb_flush.all / instructions",
>> +    "MetricGroup": "tlb",
>> +    "ScaleUnit": "1e3"
>> +  },
>> +  {
>> +    "MetricName": "macro_ops_dispatched",
>> +    "BriefDescription": "Macro-ops dispatched.",
>> +    "MetricExpr": "de_src_op_disp.all",
>> +    "MetricGroup": "decoder"
>> +  },
>> +  {
>> +    "MetricName": "sse_avx_stalls",
>> +    "BriefDescription": "Mixed SSE/AVX stalls.",
>> +    "MetricExpr": "fp_disp_faults.sse_avx_all"
>> +  },
>> +  {
>> +    "MetricName": "macro_ops_retired",
>> +    "BriefDescription": "Macro-ops retired.",
>> +    "MetricExpr": "ex_ret_ops"
>> +  },
>> +  {
>> +    "MetricName": "umc_data_bus_utilization",
>> +    "BriefDescription": "Memory controller data bus utilization.",
>> +    "MetricExpr": "d_ratio(umc_data_slot_clks.all / 2, umc_mem_clk)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "umc_cas_cmd_rate",
>> +    "BriefDescription": "Memory controller CAS command rate.",
>> +    "MetricExpr": "d_ratio(umc_cas_cmd.all * 1000, umc_mem_clk)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1"
>> +  },
>> +  {
>> +    "MetricName": "umc_cas_cmd_read_ratio",
>> +    "BriefDescription": "Ratio of memory controller CAS commands for reads.",
>> +    "MetricExpr": "d_ratio(umc_cas_cmd.rd, umc_cas_cmd.all)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "umc_cas_cmd_write_ratio",
>> +    "BriefDescription": "Ratio of memory controller CAS commands for writes.",
>> +    "MetricExpr": "d_ratio(umc_cas_cmd.wr, umc_cas_cmd.all)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "umc_mem_read_bandwidth",
>> +    "BriefDescription": "Estimated memory read bandwidth.",
>> +    "MetricExpr": "(umc_cas_cmd.rd * 64) / 1e6 / duration_time",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "1MB/s"
>> +  },
>> +  {
>> +    "MetricName": "umc_mem_write_bandwidth",
>> +    "BriefDescription": "Estimated memory write bandwidth.",
>> +    "MetricExpr": "(umc_cas_cmd.wr * 64) / 1e6 / duration_time",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "1MB/s"
>> +  },
>> +  {
>> +    "MetricName": "umc_mem_bandwidth",
>> +    "BriefDescription": "Estimated combined memory bandwidth.",
>> +    "MetricExpr": "(umc_cas_cmd.all * 64) / 1e6 / duration_time",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "1MB/s"
>> +  },
>> +  {
>> +    "MetricName": "umc_cas_cmd_read_ratio",
>> +    "BriefDescription": "Ratio of memory controller CAS commands for reads.",
>> +    "MetricExpr": "d_ratio(umc_cas_cmd.rd, umc_cas_cmd.all)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1",
>> +    "ScaleUnit": "100%"
>> +  },
>> +  {
>> +    "MetricName": "umc_cas_cmd_rate",
>> +    "BriefDescription": "Memory controller CAS command rate.",
>> +    "MetricExpr": "d_ratio(umc_cas_cmd.all * 1000, umc_mem_clk)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1"
>> +  },
>> +  {
>> +    "MetricName": "umc_activate_cmd_rate",
>> +    "BriefDescription": "Memory controller ACTIVATE command rate.",
>> +    "MetricExpr": "d_ratio(umc_act_cmd.all * 1000, umc_mem_clk)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1"
>> +  },
>> +  {
>> +    "MetricName": "umc_precharge_cmd_rate",
>> +    "BriefDescription": "Memory controller PRECHARGE command rate.",
>> +    "MetricExpr": "d_ratio(umc_pchg_cmd.all * 1000, umc_mem_clk)",
>> +    "MetricGroup": "memory_controller",
>> +    "PerPkg": "1"
> 
> Units of umc_mem_clk?
> 

Yes, per memory clock cycle i.e. umc_mem_clk. Will fix these.


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ