lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <22a067f5-67b6-5a2c-f8c4-d4f92806ae4c@huawei.com>
Date:   Wed, 27 Apr 2022 10:07:17 +0800
From:   "huangguangbin (A)" <huangguangbin2@...wei.com>
To:     John Garry <john.garry@...wei.com>, <will@...nel.org>,
        <mark.rutland@....com>
CC:     <linux-arm-kernel@...ts.infradead.org>,
        <linux-kernel@...r.kernel.org>, <linuxarm@...wei.com>,
        <liuqi115@...wei.com>, <zhangshaokun@...ilicon.com>,
        <f.fangjian@...wei.com>, <lipeng321@...wei.com>,
        <shenjian15@...wei.com>
Subject: Re: [PATCH V5 2/2] drivers/perf: hisi: add driver for HNS3 PMU



On 2022/4/26 21:48, John Garry wrote:
> On 26/04/2022 14:28, Guangbin Huang wrote:
>> HNS3(HiSilicon Network System 3) PMU is RCiEP device in HiSilicon SoC NIC,
>> supports collection of performance statistics such as bandwidth, latency,
>> packet rate and interrupt rate.
>>
>> NIC of each SICL has one PMU device for it. Driver registers each PMU
>> device to perf, and exports information of supported events, filter mode of
>> each event, bdf range, hardware clock frequency, identifier and so on via
>> sysfs.
>>
>> Each PMU device has its own registers of control, counters and interrupt,
>> and it supports 8 hardware events, each hardward event has its own
>> registers for configuration, counters and interrupt.
>>
>> Filter options contains:
>> event        - select event
>> port         - select physical port of nic
>> tc           - select tc(must be used with port)
>> func         - select PF/VF
>> queue        - select queue of PF/VF(must be used with func)
>> intr         - select interrupt number(must be used with func)
>> global       - select all functions of IO DIE
>>
>> Signed-off-by: Guangbin Huang <huangguangbin2@...wei.com>
> 
> Apart from some nits, below, feel free to add:
> Reviewed-by: John Garry <john.garry@...wei.com>
> 
Ok, thanks.
> 
>> ---
>>   MAINTAINERS                       |    6 +
>>   drivers/perf/hisilicon/Kconfig    |   10 +
>>   drivers/perf/hisilicon/Makefile   |    1 +
>>   drivers/perf/hisilicon/hns3_pmu.c | 1660 +++++++++++++++++++++++++++++
>>   include/linux/cpuhotplug.h        |    1 +
>>   5 files changed, 1678 insertions(+)
>>   create mode 100644 drivers/perf/hisilicon/hns3_pmu.c
>>
>> diff --git a/MAINTAINERS b/MAINTAINERS
>> index 69a2935daf6c..34b87348503a 100644
>> --- a/MAINTAINERS
>> +++ b/MAINTAINERS
>> @@ -8691,6 +8691,12 @@ F:    Documentation/admin-guide/perf/hisi-pcie-pmu.rst
>>   F:    Documentation/admin-guide/perf/hisi-pmu.rst
>>   F:    drivers/perf/hisilicon
>> +HISILICON HNS3 PMU DRIVER
>> +M:    Guangbin Huang <huangguangbin2@...wei.com>
>> +S:    Supported
>> +F:    Documentation/admin-guide/perf/hns3-pmu.rst
>> +F:    drivers/perf/hisilicon/hns3_pmu.c
>> +
>>   HISILICON QM AND ZIP Controller DRIVER
>>   M:    Zhou Wang <wangzhou1@...ilicon.com>
>>   L:    linux-crypto@...r.kernel.org
>> diff --git a/drivers/perf/hisilicon/Kconfig b/drivers/perf/hisilicon/Kconfig
>> index 5546218b5598..171bfc1b6bc2 100644
>> --- a/drivers/perf/hisilicon/Kconfig
>> +++ b/drivers/perf/hisilicon/Kconfig
>> @@ -14,3 +14,13 @@ config HISI_PCIE_PMU
>>         RCiEP devices.
>>         Adds the PCIe PMU into perf events system for monitoring latency,
>>         bandwidth etc.
>> +
>> +config HNS3_PMU
>> +    tristate "HNS3 PERF PMU"
>> +    depends on ARM64 || COMPILE_TEST
>> +    depends on PCI
>> +    help
>> +      Provide support for HNS3 performance monitoring unit (PMU) RCiEP
>> +      devices.
>> +      Adds the HNS3 PMU into perf events system for monitoring latency,
>> +      bandwidth etc.
>> diff --git a/drivers/perf/hisilicon/Makefile b/drivers/perf/hisilicon/Makefile
>> index 506ed39e3266..13297ec2798f 100644
>> --- a/drivers/perf/hisilicon/Makefile
>> +++ b/drivers/perf/hisilicon/Makefile
>> @@ -4,3 +4,4 @@ obj-$(CONFIG_HISI_PMU) += hisi_uncore_pmu.o hisi_uncore_l3c_pmu.o \
>>                 hisi_uncore_pa_pmu.o
>>   obj-$(CONFIG_HISI_PCIE_PMU) += hisi_pcie_pmu.o
>> +obj-$(CONFIG_HNS3_PMU) += hns3_pmu.o
>> diff --git a/drivers/perf/hisilicon/hns3_pmu.c b/drivers/perf/hisilicon/hns3_pmu.c
>> new file mode 100644
>> index 000000000000..43c0993a1e12
>> --- /dev/null
>> +++ b/drivers/perf/hisilicon/hns3_pmu.c
>> @@ -0,0 +1,1660 @@
>> +// SPDX-License-Identifier: GPL-2.0-only
>> +/*
>> + * This driver adds support for HNS3 PMU iEP device. Related perf events are
>> + * bandwidth, latency, packet rate, interrupt rate etc.
>> + *
>> + * Copyright (C) 2022 HiSilicon Limited
>> + */
>> +#include <linux/bitfield.h>
>> +#include <linux/bitmap.h>
>> +#include <linux/bug.h>
>> +#include <linux/cpuhotplug.h>
>> +#include <linux/cpumask.h>
>> +#include <linux/delay.h>
>> +#include <linux/device.h>
>> +#include <linux/err.h>
>> +#include <linux/interrupt.h>
>> +#include <linux/iopoll.h>
>> +#include <linux/io-64-nonatomic-hi-lo.h>
>> +#include <linux/irq.h>
>> +#include <linux/kernel.h>
>> +#include <linux/list.h>
>> +#include <linux/module.h>
>> +#include <linux/pci.h>
>> +#include <linux/pci-epf.h>
>> +#include <linux/perf_event.h>
>> +#include <linux/smp.h>
>> +
>> +/* registers offset address */
>> +#define HNS3_PMU_REG_GLOBAL_CTRL        0x0000
>> +#define HNS3_PMU_REG_CLOCK_FREQ            0x0020
>> +#define HNS3_PMU_REG_BDF            0x0fe0
>> +#define HNS3_PMU_REG_VERSION            0x0fe4
>> +#define HNS3_PMU_REG_DEVICE_ID            0x0fe8
>> +
>> +#define HNS3_PMU_REG_EVENT_OFFSET        0x1000
>> +#define HNS3_PMU_REG_EVENT_SIZE            0x1000
>> +#define HNS3_PMU_REG_EVENT_CTRL_LOW        0x00
>> +#define HNS3_PMU_REG_EVENT_CTRL_HIGH        0x04
>> +#define HNS3_PMU_REG_EVENT_INTR_STATUS        0x08
>> +#define HNS3_PMU_REG_EVENT_INTR_MASK        0x0c
>> +#define HNS3_PMU_REG_EVENT_COUNTER        0x10
>> +#define HNS3_PMU_REG_EVENT_EXT_COUNTER        0x18
>> +#define HNS3_PMU_REG_EVENT_QID_CTRL        0x28
>> +#define HNS3_PMU_REG_EVENT_QID_PARA        0x2c
>> +
>> +#define HNS3_PMU_FILTER_SUPPORT_GLOBAL        BIT(0)
>> +#define HNS3_PMU_FILTER_SUPPORT_PORT        BIT(1)
>> +#define HNS3_PMU_FILTER_SUPPORT_PORT_TC        BIT(2)
>> +#define HNS3_PMU_FILTER_SUPPORT_FUNC        BIT(3)
>> +#define HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE    BIT(4)
>> +#define HNS3_PMU_FILTER_SUPPORT_FUNC_INTR    BIT(5)
>> +
>> +#define HNS3_PMU_FILTER_ALL_TC            0xf
>> +#define HNS3_PMU_FILTER_ALL_QUEUE        0xffff
>> +
>> +#define HNS3_PMU_CTRL_SUBEVENT_S        4
>> +#define HNS3_PMU_CTRL_FILTER_MODE_S        24
>> +
>> +#define HNS3_PMU_GLOBAL_START            BIT(0)
>> +
>> +#define HNS3_PMU_EVENT_STATUS_RESET        BIT(11)
>> +#define HNS3_PMU_EVENT_EN            BIT(12)
>> +#define HNS3_PMU_EVENT_OVERFLOW_RESTART        BIT(15)
>> +
>> +#define HNS3_PMU_QID_PARA_FUNC_S        0
>> +#define HNS3_PMU_QID_PARA_QUEUE_S        16
>> +
>> +#define HNS3_PMU_QID_CTRL_REQ_ENABLE        BIT(0)
>> +#define HNS3_PMU_QID_CTRL_DONE            BIT(1)
>> +#define HNS3_PMU_QID_CTRL_MISS            BIT(2)
>> +
>> +#define HNS3_PMU_INTR_MASK_OVERFLOW        BIT(1)
>> +
>> +#define HNS3_PMU_MAX_HW_EVENTS            8
>> +
>> +/*
>> + * Each hardware event contains two registers (counter and ext_counter) for
>> + * bandwidth, packet rate, latency and interrupt rate. These two registers will
>> + * be triggered to run at the same when a hardware event is enabled. The meaning
>> + * of counter and ext_counter of different event type are different, their
>> + * meaning show as follow:
>> + *
>> + * +----------------+------------------+---------------+
>> + * |   event type   |     counter      |  ext_counter  |
>> + * +----------------+------------------+---------------+
>> + * | bandwidth      | byte number      | cycle number  |
>> + * +----------------+------------------+---------------+
>> + * | packet rate    | packet number    | cycle number  |
>> + * +----------------+------------------+---------------+
>> + * | latency        | cycle number     | packet number |
>> + * +----------------+------------------+---------------+
>> + * | interrupt rate | interrupt number | cycle number  |
>> + * +----------------+------------------+---------------+
>> + *
>> + * The cycle number indicates increment of counter of hardware timer, the
>> + * frequency of hardware timer can be read from hw_clk_freq file.
>> + *
>> + * Performance of each hardware event is calculated by: counter / ext_counter.
>> + *
>> + * Since processing of data is preferred to be done in userspace, we expose
>> + * ext_counter as a separate event for userspace and use bit 16 to indicate it.
>> + * For example, event 0x00001 and 0x10001 are actually one event for hardware
>> + * because bit 0-15 are same. If the bit 16 of one event is 0 means to read
>> + * counter register, otherwise means to read ext_counter register.
>> + */
>> +/* bandwidth events */
>> +#define HNS3_PMU_EVT_BW_SSU_EGU_BYTE_NUM        0x00001
>> +#define HNS3_PMU_EVT_BW_SSU_EGU_TIME            0x10001
>> +#define HNS3_PMU_EVT_BW_SSU_RPU_BYTE_NUM        0x00002
>> +#define HNS3_PMU_EVT_BW_SSU_RPU_TIME            0x10002
>> +#define HNS3_PMU_EVT_BW_SSU_ROCE_BYTE_NUM        0x00003
>> +#define HNS3_PMU_EVT_BW_SSU_ROCE_TIME            0x10003
>> +#define HNS3_PMU_EVT_BW_ROCE_SSU_BYTE_NUM        0x00004
>> +#define HNS3_PMU_EVT_BW_ROCE_SSU_TIME            0x10004
>> +#define HNS3_PMU_EVT_BW_TPU_SSU_BYTE_NUM        0x00005
>> +#define HNS3_PMU_EVT_BW_TPU_SSU_TIME            0x10005
>> +#define HNS3_PMU_EVT_BW_RPU_RCBRX_BYTE_NUM        0x00006
>> +#define HNS3_PMU_EVT_BW_RPU_RCBRX_TIME            0x10006
>> +#define HNS3_PMU_EVT_BW_RCBTX_TXSCH_BYTE_NUM        0x00008
>> +#define HNS3_PMU_EVT_BW_RCBTX_TXSCH_TIME        0x10008
>> +#define HNS3_PMU_EVT_BW_WR_FBD_BYTE_NUM            0x00009
>> +#define HNS3_PMU_EVT_BW_WR_FBD_TIME            0x10009
>> +#define HNS3_PMU_EVT_BW_WR_EBD_BYTE_NUM            0x0000a
>> +#define HNS3_PMU_EVT_BW_WR_EBD_TIME            0x1000a
>> +#define HNS3_PMU_EVT_BW_RD_FBD_BYTE_NUM            0x0000b
>> +#define HNS3_PMU_EVT_BW_RD_FBD_TIME            0x1000b
>> +#define HNS3_PMU_EVT_BW_RD_EBD_BYTE_NUM            0x0000c
>> +#define HNS3_PMU_EVT_BW_RD_EBD_TIME            0x1000c
>> +#define HNS3_PMU_EVT_BW_RD_PAY_M0_BYTE_NUM        0x0000d
>> +#define HNS3_PMU_EVT_BW_RD_PAY_M0_TIME            0x1000d
>> +#define HNS3_PMU_EVT_BW_RD_PAY_M1_BYTE_NUM        0x0000e
>> +#define HNS3_PMU_EVT_BW_RD_PAY_M1_TIME            0x1000e
>> +#define HNS3_PMU_EVT_BW_WR_PAY_M0_BYTE_NUM        0x0000f
>> +#define HNS3_PMU_EVT_BW_WR_PAY_M0_TIME            0x1000f
>> +#define HNS3_PMU_EVT_BW_WR_PAY_M1_BYTE_NUM        0x00010
>> +#define HNS3_PMU_EVT_BW_WR_PAY_M1_TIME            0x10010
>> +
>> +/* packet rate events */
>> +#define HNS3_PMU_EVT_PPS_IGU_SSU_PACKET_NUM        0x00100
>> +#define HNS3_PMU_EVT_PPS_IGU_SSU_TIME            0x10100
>> +#define HNS3_PMU_EVT_PPS_SSU_EGU_PACKET_NUM        0x00101
>> +#define HNS3_PMU_EVT_PPS_SSU_EGU_TIME            0x10101
>> +#define HNS3_PMU_EVT_PPS_SSU_RPU_PACKET_NUM        0x00102
>> +#define HNS3_PMU_EVT_PPS_SSU_RPU_TIME            0x10102
>> +#define HNS3_PMU_EVT_PPS_SSU_ROCE_PACKET_NUM        0x00103
>> +#define HNS3_PMU_EVT_PPS_SSU_ROCE_TIME            0x10103
>> +#define HNS3_PMU_EVT_PPS_ROCE_SSU_PACKET_NUM        0x00104
>> +#define HNS3_PMU_EVT_PPS_ROCE_SSU_TIME            0x10104
>> +#define HNS3_PMU_EVT_PPS_TPU_SSU_PACKET_NUM        0x00105
>> +#define HNS3_PMU_EVT_PPS_TPU_SSU_TIME            0x10105
>> +#define HNS3_PMU_EVT_PPS_RPU_RCBRX_PACKET_NUM        0x00106
>> +#define HNS3_PMU_EVT_PPS_RPU_RCBRX_TIME            0x10106
>> +#define HNS3_PMU_EVT_PPS_RCBTX_TPU_PACKET_NUM        0x00107
>> +#define HNS3_PMU_EVT_PPS_RCBTX_TPU_TIME            0x10107
>> +#define HNS3_PMU_EVT_PPS_RCBTX_TXSCH_PACKET_NUM        0x00108
>> +#define HNS3_PMU_EVT_PPS_RCBTX_TXSCH_TIME        0x10108
>> +#define HNS3_PMU_EVT_PPS_WR_FBD_PACKET_NUM        0x00109
>> +#define HNS3_PMU_EVT_PPS_WR_FBD_TIME            0x10109
>> +#define HNS3_PMU_EVT_PPS_WR_EBD_PACKET_NUM        0x0010a
>> +#define HNS3_PMU_EVT_PPS_WR_EBD_TIME            0x1010a
>> +#define HNS3_PMU_EVT_PPS_RD_FBD_PACKET_NUM        0x0010b
>> +#define HNS3_PMU_EVT_PPS_RD_FBD_TIME            0x1010b
>> +#define HNS3_PMU_EVT_PPS_RD_EBD_PACKET_NUM        0x0010c
>> +#define HNS3_PMU_EVT_PPS_RD_EBD_TIME            0x1010c
>> +#define HNS3_PMU_EVT_PPS_RD_PAY_M0_PACKET_NUM        0x0010d
>> +#define HNS3_PMU_EVT_PPS_RD_PAY_M0_TIME            0x1010d
>> +#define HNS3_PMU_EVT_PPS_RD_PAY_M1_PACKET_NUM        0x0010e
>> +#define HNS3_PMU_EVT_PPS_RD_PAY_M1_TIME            0x1010e
>> +#define HNS3_PMU_EVT_PPS_WR_PAY_M0_PACKET_NUM        0x0010f
>> +#define HNS3_PMU_EVT_PPS_WR_PAY_M0_TIME            0x1010f
>> +#define HNS3_PMU_EVT_PPS_WR_PAY_M1_PACKET_NUM        0x00110
>> +#define HNS3_PMU_EVT_PPS_WR_PAY_M1_TIME            0x10110
>> +#define HNS3_PMU_EVT_PPS_NICROH_TX_PRE_PACKET_NUM    0x00111
>> +#define HNS3_PMU_EVT_PPS_NICROH_TX_PRE_TIME        0x10111
>> +#define HNS3_PMU_EVT_PPS_NICROH_RX_PRE_PACKET_NUM    0x00112
>> +#define HNS3_PMU_EVT_PPS_NICROH_RX_PRE_TIME        0x10112
>> +
>> +/* latency events */
>> +#define HNS3_PMU_EVT_DLY_TX_PUSH_TIME            0x00202
>> +#define HNS3_PMU_EVT_DLY_TX_PUSH_PACKET_NUM        0x10202
>> +#define HNS3_PMU_EVT_DLY_TX_TIME            0x00204
>> +#define HNS3_PMU_EVT_DLY_TX_PACKET_NUM            0x10204
>> +#define HNS3_PMU_EVT_DLY_SSU_TX_NIC_TIME        0x00206
>> +#define HNS3_PMU_EVT_DLY_SSU_TX_NIC_PACKET_NUM        0x10206
>> +#define HNS3_PMU_EVT_DLY_SSU_TX_ROCE_TIME        0x00207
>> +#define HNS3_PMU_EVT_DLY_SSU_TX_ROCE_PACKET_NUM        0x10207
>> +#define HNS3_PMU_EVT_DLY_SSU_RX_NIC_TIME        0x00208
>> +#define HNS3_PMU_EVT_DLY_SSU_RX_NIC_PACKET_NUM        0x10208
>> +#define HNS3_PMU_EVT_DLY_SSU_RX_ROCE_TIME        0x00209
>> +#define HNS3_PMU_EVT_DLY_SSU_RX_ROCE_PACKET_NUM        0x10209
>> +#define HNS3_PMU_EVT_DLY_RPU_TIME            0x0020e
>> +#define HNS3_PMU_EVT_DLY_RPU_PACKET_NUM            0x1020e
>> +#define HNS3_PMU_EVT_DLY_TPU_TIME            0x0020f
>> +#define HNS3_PMU_EVT_DLY_TPU_PACKET_NUM            0x1020f
>> +#define HNS3_PMU_EVT_DLY_RPE_TIME            0x00210
>> +#define HNS3_PMU_EVT_DLY_RPE_PACKET_NUM            0x10210
>> +#define HNS3_PMU_EVT_DLY_TPE_TIME            0x00211
>> +#define HNS3_PMU_EVT_DLY_TPE_PACKET_NUM            0x10211
>> +#define HNS3_PMU_EVT_DLY_TPE_PUSH_TIME            0x00212
>> +#define HNS3_PMU_EVT_DLY_TPE_PUSH_PACKET_NUM        0x10212
>> +#define HNS3_PMU_EVT_DLY_WR_FBD_TIME            0x00213
>> +#define HNS3_PMU_EVT_DLY_WR_FBD_PACKET_NUM        0x10213
>> +#define HNS3_PMU_EVT_DLY_WR_EBD_TIME            0x00214
>> +#define HNS3_PMU_EVT_DLY_WR_EBD_PACKET_NUM        0x10214
>> +#define HNS3_PMU_EVT_DLY_RD_FBD_TIME            0x00215
>> +#define HNS3_PMU_EVT_DLY_RD_FBD_PACKET_NUM        0x10215
>> +#define HNS3_PMU_EVT_DLY_RD_EBD_TIME            0x00216
>> +#define HNS3_PMU_EVT_DLY_RD_EBD_PACKET_NUM        0x10216
>> +#define HNS3_PMU_EVT_DLY_RD_PAY_M0_TIME            0x00217
>> +#define HNS3_PMU_EVT_DLY_RD_PAY_M0_PACKET_NUM        0x10217
>> +#define HNS3_PMU_EVT_DLY_RD_PAY_M1_TIME            0x00218
>> +#define HNS3_PMU_EVT_DLY_RD_PAY_M1_PACKET_NUM        0x10218
>> +#define HNS3_PMU_EVT_DLY_WR_PAY_M0_TIME            0x00219
>> +#define HNS3_PMU_EVT_DLY_WR_PAY_M0_PACKET_NUM        0x10219
>> +#define HNS3_PMU_EVT_DLY_WR_PAY_M1_TIME            0x0021a
>> +#define HNS3_PMU_EVT_DLY_WR_PAY_M1_PACKET_NUM        0x1021a
>> +#define HNS3_PMU_EVT_DLY_MSIX_WRITE_TIME        0x0021c
>> +#define HNS3_PMU_EVT_DLY_MSIX_WRITE_PACKET_NUM        0x1021c
>> +
>> +/* interrupt rate events */
>> +#define HNS3_PMU_EVT_PPS_MSIX_NIC_INTR_NUM        0x00300
>> +#define HNS3_PMU_EVT_PPS_MSIX_NIC_TIME            0x10300
>> +
>> +/* filter mode supported by each bandwidth event */
>> +#define HNS3_PMU_FILTER_BW_SSU_EGU        0x07
>> +#define HNS3_PMU_FILTER_BW_SSU_RPU        0x1f
>> +#define HNS3_PMU_FILTER_BW_SSU_ROCE        0x0f
>> +#define HNS3_PMU_FILTER_BW_ROCE_SSU        0x0f
>> +#define HNS3_PMU_FILTER_BW_TPU_SSU        0x1f
>> +#define HNS3_PMU_FILTER_BW_RPU_RCBRX        0x11
>> +#define HNS3_PMU_FILTER_BW_RCBTX_TXSCH        0x11
>> +#define HNS3_PMU_FILTER_BW_WR_FBD        0x1b
>> +#define HNS3_PMU_FILTER_BW_WR_EBD        0x11
>> +#define HNS3_PMU_FILTER_BW_RD_FBD        0x01
>> +#define HNS3_PMU_FILTER_BW_RD_EBD        0x1b
>> +#define HNS3_PMU_FILTER_BW_RD_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_BW_RD_PAY_M1        0x01
>> +#define HNS3_PMU_FILTER_BW_WR_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_BW_WR_PAY_M1        0x01
>> +
>> +/* filter mode supported by each packet rate event */
>> +#define HNS3_PMU_FILTER_PPS_IGU_SSU        0x07
>> +#define HNS3_PMU_FILTER_PPS_SSU_EGU        0x07
>> +#define HNS3_PMU_FILTER_PPS_SSU_RPU        0x1f
>> +#define HNS3_PMU_FILTER_PPS_SSU_ROCE        0x0f
>> +#define HNS3_PMU_FILTER_PPS_ROCE_SSU        0x0f
>> +#define HNS3_PMU_FILTER_PPS_TPU_SSU        0x1f
>> +#define HNS3_PMU_FILTER_PPS_RPU_RCBRX        0x11
>> +#define HNS3_PMU_FILTER_PPS_RCBTX_TPU        0x1f
>> +#define HNS3_PMU_FILTER_PPS_RCBTX_TXSCH        0x11
>> +#define HNS3_PMU_FILTER_PPS_WR_FBD        0x1b
>> +#define HNS3_PMU_FILTER_PPS_WR_EBD        0x11
>> +#define HNS3_PMU_FILTER_PPS_RD_FBD        0x01
>> +#define HNS3_PMU_FILTER_PPS_RD_EBD        0x1b
>> +#define HNS3_PMU_FILTER_PPS_RD_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_PPS_RD_PAY_M1        0x01
>> +#define HNS3_PMU_FILTER_PPS_WR_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_PPS_WR_PAY_M1        0x01
>> +#define HNS3_PMU_FILTER_PPS_NICROH_TX_PRE    0x01
>> +#define HNS3_PMU_FILTER_PPS_NICROH_RX_PRE    0x01
>> +
>> +/* filter mode supported by each latency event */
>> +#define HNS3_PMU_FILTER_DLY_TX_PUSH        0x01
>> +#define HNS3_PMU_FILTER_DLY_TX            0x01
>> +#define HNS3_PMU_FILTER_DLY_SSU_TX_NIC        0x07
>> +#define HNS3_PMU_FILTER_DLY_SSU_TX_ROCE        0x07
>> +#define HNS3_PMU_FILTER_DLY_SSU_RX_NIC        0x07
>> +#define HNS3_PMU_FILTER_DLY_SSU_RX_ROCE        0x07
>> +#define HNS3_PMU_FILTER_DLY_RPU            0x11
>> +#define HNS3_PMU_FILTER_DLY_TPU            0x1f
>> +#define HNS3_PMU_FILTER_DLY_RPE            0x01
>> +#define HNS3_PMU_FILTER_DLY_TPE            0x0b
>> +#define HNS3_PMU_FILTER_DLY_TPE_PUSH        0x1b
>> +#define HNS3_PMU_FILTER_DLY_WR_FBD        0x1b
>> +#define HNS3_PMU_FILTER_DLY_WR_EBD        0x11
>> +#define HNS3_PMU_FILTER_DLY_RD_FBD        0x01
>> +#define HNS3_PMU_FILTER_DLY_RD_EBD        0x1b
>> +#define HNS3_PMU_FILTER_DLY_RD_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_DLY_RD_PAY_M1        0x01
>> +#define HNS3_PMU_FILTER_DLY_WR_PAY_M0        0x01
>> +#define HNS3_PMU_FILTER_DLY_WR_PAY_M1        0x01
>> +#define HNS3_PMU_FILTER_DLY_MSIX_WRITE        0x01
>> +
>> +/* filter mode supported by each interrupt rate event */
>> +#define HNS3_PMU_FILTER_INTR_MSIX_NIC        0x01
>> +
>> +enum hns3_pmu_hw_filter_mode {
>> +    HNS3_PMU_HW_FILTER_GLOBAL,
>> +    HNS3_PMU_HW_FILTER_PORT,
>> +    HNS3_PMU_HW_FILTER_PORT_TC,
>> +    HNS3_PMU_HW_FILTER_FUNC,
>> +    HNS3_PMU_HW_FILTER_FUNC_QUEUE,
>> +    HNS3_PMU_HW_FILTER_FUNC_INTR,
>> +};
>> +
>> +struct hns3_pmu_event_attr {
>> +    u32 event;
>> +    u16 filter_support;
>> +};
>> +
>> +struct hns3_pmu {
>> +    struct perf_event *hw_events[HNS3_PMU_MAX_HW_EVENTS];
>> +    struct hlist_node node;
>> +    struct pci_dev *pdev;
>> +    struct pmu pmu;
>> +    void __iomem *base;
>> +    int irq;
>> +    int on_cpu;
>> +    u32 identifier;
>> +    u32 hw_clk_freq; /* hardware clock frequency of PMU */
>> +    /* maximum and minimum bdf allowed by PMU */
>> +    u16 bdf_min;
>> +    u16 bdf_max;
>> +};
>> +
>> +#define to_hns3_pmu(p)  (container_of((p), struct hns3_pmu, pmu))
>> +
>> +#define GET_PCI_DEVFN(bdf)  ((bdf) & 0xff)
>> +
>> +#define FILTER_CONDITION_PORT(port) ((1 << (port)) & 0xff)
>> +#define FILTER_CONDITION_PORT_TC(port, tc) (((port) << 3) | ((tc) & 0x07))
>> +#define FILTER_CONDITION_FUNC_INTR(func, intr) (((intr) << 8) | (func))
>> +
>> +#define HNS3_PMU_FILTER_ATTR(_name, _config, _start, _end)            \
>> +    static inline u64 hns3_get_##_name(struct perf_event *event)  \
> 
> 
> nit: hns3_pmu would be better prefix
OK.

> 
>> +    {                                                             \
>> +        return FIELD_GET(GENMASK_ULL(_end, _start),           \
>> +                 event->attr._config);                \
>> +    }
>> +
>> +HNS3_PMU_FILTER_ATTR(event, config, 0, 16);
>> +HNS3_PMU_FILTER_ATTR(subevent, config, 0, 7);
>> +HNS3_PMU_FILTER_ATTR(event_type, config, 8, 15);
>> +HNS3_PMU_FILTER_ATTR(ext_counter_used, config, 16, 16);
>> +HNS3_PMU_FILTER_ATTR(real_event, config, 0, 15);
>> +HNS3_PMU_FILTER_ATTR(port, config1, 0, 3);
>> +HNS3_PMU_FILTER_ATTR(tc, config1, 4, 7);
>> +HNS3_PMU_FILTER_ATTR(bdf, config1, 8, 23);
>> +HNS3_PMU_FILTER_ATTR(queue, config1, 24, 39);
>> +HNS3_PMU_FILTER_ATTR(intr, config1, 40, 51);
>> +HNS3_PMU_FILTER_ATTR(global, config1, 52, 52);
>> +
>> +#define HNS3_BW_EVT_BYTE_NUM(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_BW_##_name##_BYTE_NUM,                \
>> +    HNS3_PMU_FILTER_BW_##_name})
>> +#define HNS3_BW_EVT_TIME(_name)        (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_BW_##_name##_TIME,                    \
>> +    HNS3_PMU_FILTER_BW_##_name})
>> +#define HNS3_PPS_EVT_PACKET_NUM(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_PPS_##_name##_PACKET_NUM,                \
>> +    HNS3_PMU_FILTER_PPS_##_name})
>> +#define HNS3_PPS_EVT_TIME(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_PPS_##_name##_TIME,                \
>> +    HNS3_PMU_FILTER_PPS_##_name})
>> +#define HNS3_DLY_EVT_TIME(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_DLY_##_name##_TIME,                \
>> +    HNS3_PMU_FILTER_DLY_##_name})
>> +#define HNS3_DLY_EVT_PACKET_NUM(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_DLY_##_name##_PACKET_NUM,                \
>> +    HNS3_PMU_FILTER_DLY_##_name})
>> +#define HNS3_INTR_EVT_INTR_NUM(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_PPS_##_name##_INTR_NUM,                \
>> +    HNS3_PMU_FILTER_INTR_##_name})
>> +#define HNS3_INTR_EVT_TIME(_name)    (&(struct hns3_pmu_event_attr) {\
>> +    HNS3_PMU_EVT_PPS_##_name##_TIME,                \
>> +    HNS3_PMU_FILTER_INTR_##_name})
>> +
>> +static ssize_t hns3_pmu_format_show(struct device *dev,
>> +                    struct device_attribute *attr, char *buf)
>> +{
>> +    struct dev_ext_attribute *eattr;
>> +
>> +    eattr = container_of(attr, struct dev_ext_attribute, attr);
>> +
>> +    return sysfs_emit(buf, "%s\n", (char *)eattr->var);
>> +}
>> +
>> +static ssize_t hns3_pmu_event_show(struct device *dev,
>> +                   struct device_attribute *attr, char *buf)
>> +{
>> +    struct hns3_pmu_event_attr *event;
>> +    struct dev_ext_attribute *eattr;
>> +
>> +    eattr = container_of(attr, struct dev_ext_attribute, attr);
>> +    event = eattr->var;
>> +
>> +    return sysfs_emit(buf, "config=0x%05x\n", event->event);
>> +}
>> +
>> +static ssize_t hns3_pmu_filter_mode_show(struct device *dev,
>> +                     struct device_attribute *attr,
>> +                     char *buf)
>> +{
>> +    struct hns3_pmu_event_attr *event;
>> +    struct dev_ext_attribute *eattr;
>> +    int len;
>> +
>> +    eattr = container_of(attr, struct dev_ext_attribute, attr);
>> +    event = eattr->var;
>> +
>> +    len = sysfs_emit_at(buf, 0, "filter mode supported: ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_GLOBAL)
>> +        len += sysfs_emit_at(buf, len, "global ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT)
>> +        len += sysfs_emit_at(buf, len, "port ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT_TC)
>> +        len += sysfs_emit_at(buf, len, "port-tc ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC)
>> +        len += sysfs_emit_at(buf, len, "func ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE)
>> +        len += sysfs_emit_at(buf, len, "func-queue ");
>> +    if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_INTR)
>> +        len += sysfs_emit_at(buf, len, "func-intr");
>> +
>> +    len += sysfs_emit_at(buf, len, "\n");
>> +
>> +    return len;
>> +}
>> +
>> +#define HNS3_PMU_ATTR(_name, _func, _config)                \
>> +    (&((struct dev_ext_attribute[]) {                \
>> +        { __ATTR(_name, 0444, _func, NULL), (void *)_config }    \
>> +    })[0].attr.attr)
>> +
>> +#define HNS3_PMU_FORMAT_ATTR(_name, _format) \
>> +    HNS3_PMU_ATTR(_name, hns3_pmu_format_show, (void *)_format)
>> +#define HNS3_PMU_EVENT_ATTR(_name, _event) \
>> +    HNS3_PMU_ATTR(_name, hns3_pmu_event_show, (void *)_event)
>> +#define HNS3_PMU_FLT_MODE_ATTR(_name, _event) \
>> +    HNS3_PMU_ATTR(_name, hns3_pmu_filter_mode_show, (void *)_event)
>> +
>> +#define HNS3_PMU_BW_EVT_PAIR(_name, _macro) \
>> +    HNS3_PMU_EVENT_ATTR(_name##_byte_num, HNS3_BW_EVT_BYTE_NUM(_macro)), \
>> +    HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_BW_EVT_TIME(_macro))
>> +#define HNS3_PMU_PPS_EVT_PAIR(_name, _macro) \
>> +    HNS3_PMU_EVENT_ATTR(_name##_packet_num, HNS3_PPS_EVT_PACKET_NUM(_macro)), \
>> +    HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_PPS_EVT_TIME(_macro))
>> +#define HNS3_PMU_DLY_EVT_PAIR(_name, _macro) \
>> +    HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_DLY_EVT_TIME(_macro)), \
>> +    HNS3_PMU_EVENT_ATTR(_name##_packet_num, HNS3_DLY_EVT_PACKET_NUM(_macro))
>> +#define HNS3_PMU_INTR_EVT_PAIR(_name, _macro) \
>> +    HNS3_PMU_EVENT_ATTR(_name##_intr_num, HNS3_INTR_EVT_INTR_NUM(_macro)), \
>> +    HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_INTR_EVT_TIME(_macro))
>> +
>> +#define HNS3_PMU_BW_FLT_MODE_PAIR(_name, _macro) \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_byte_num, HNS3_BW_EVT_BYTE_NUM(_macro)), \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_BW_EVT_TIME(_macro))
>> +#define HNS3_PMU_PPS_FLT_MODE_PAIR(_name, _macro) \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_packet_num, HNS3_PPS_EVT_PACKET_NUM(_macro)), \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_PPS_EVT_TIME(_macro))
>> +#define HNS3_PMU_DLY_FLT_MODE_PAIR(_name, _macro) \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_DLY_EVT_TIME(_macro)), \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_packet_num, HNS3_DLY_EVT_PACKET_NUM(_macro))
>> +#define HNS3_PMU_INTR_FLT_MODE_PAIR(_name, _macro) \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_intr_num, HNS3_INTR_EVT_INTR_NUM(_macro)), \
>> +    HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_INTR_EVT_TIME(_macro))
>> +
>> +static u8 hns3_pmu_hw_filter_modes[] = {
>> +    HNS3_PMU_HW_FILTER_GLOBAL,
>> +    HNS3_PMU_HW_FILTER_PORT,
>> +    HNS3_PMU_HW_FILTER_PORT_TC,
>> +    HNS3_PMU_HW_FILTER_FUNC,
>> +    HNS3_PMU_HW_FILTER_FUNC_QUEUE,
>> +    HNS3_PMU_HW_FILTER_FUNC_INTR,
>> +};
>> +
>> +#define HNS3_PMU_SET_HW_FILTER(_hwc, _mode) \
>> +    ((_hwc)->addr_filters = (void *)&hns3_pmu_hw_filter_modes[(_mode)])
>> +
>> +static ssize_t identifier_show(struct device *dev,
>> +                   struct device_attribute *attr, char *buf)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev));
>> +
>> +    return sysfs_emit(buf, "0x%x\n", hns3_pmu->identifier);
>> +}
>> +static DEVICE_ATTR_RO(identifier);
>> +
>> +static ssize_t cpumask_show(struct device *dev, struct device_attribute *attr,
>> +                char *buf)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev));
>> +
>> +    return sysfs_emit(buf, "%d\n", hns3_pmu->on_cpu);
>> +}
>> +static DEVICE_ATTR_RO(cpumask);
>> +
>> +static ssize_t bdf_min_show(struct device *dev, struct device_attribute *attr,
>> +                char *buf)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev));
>> +
>> +    return sysfs_emit(buf, "0x%4x\n", hns3_pmu->bdf_min);
>> +}
>> +static DEVICE_ATTR_RO(bdf_min);
>> +
>> +static ssize_t bdf_max_show(struct device *dev, struct device_attribute *attr,
>> +                char *buf)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev));
>> +
>> +    return sysfs_emit(buf, "0x%4x\n", hns3_pmu->bdf_max);
>> +}
>> +static DEVICE_ATTR_RO(bdf_max);
>> +
>> +static ssize_t hw_clk_freq_show(struct device *dev,
>> +                struct device_attribute *attr, char *buf)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev));
>> +
>> +    return sysfs_emit(buf, "%u\n", hns3_pmu->hw_clk_freq);
>> +}
>> +static DEVICE_ATTR_RO(hw_clk_freq);
>> +
>> +static struct attribute *hns3_pmu_events_attr[] = {
>> +    /* bandwidth events */
>> +    HNS3_PMU_BW_EVT_PAIR(bw_ssu_egu, SSU_EGU),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_ssu_rpu, SSU_RPU),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_ssu_roce, SSU_ROCE),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_roce_ssu, ROCE_SSU),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_tpu_ssu, TPU_SSU),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rpu_rcbrx, RPU_RCBRX),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rcbtx_txsch, RCBTX_TXSCH),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_wr_fbd, WR_FBD),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_wr_ebd, WR_EBD),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rd_fbd, RD_FBD),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rd_ebd, RD_EBD),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_BW_EVT_PAIR(bw_wr_pay_m1, WR_PAY_M1),
>> +
>> +    /* packet rate events */
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_igu_ssu, IGU_SSU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_ssu_egu, SSU_EGU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_ssu_rpu, SSU_RPU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_ssu_roce, SSU_ROCE),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_roce_ssu, ROCE_SSU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_tpu_ssu, TPU_SSU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rpu_rcbrx, RPU_RCBRX),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rcbtx_tpu, RCBTX_TPU),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rcbtx_txsch, RCBTX_TXSCH),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_wr_fbd, WR_FBD),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_wr_ebd, WR_EBD),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rd_fbd, RD_FBD),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rd_ebd, RD_EBD),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_wr_pay_m1, WR_PAY_M1),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_intr_nicroh_tx_pre, NICROH_TX_PRE),
>> +    HNS3_PMU_PPS_EVT_PAIR(pps_intr_nicroh_rx_pre, NICROH_RX_PRE),
>> +
>> +    /* latency events */
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_tx_push_to_mac, TX_PUSH),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_tx_normal_to_mac, TX),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_ssu_tx_th_nic, SSU_TX_NIC),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_ssu_tx_th_roce, SSU_TX_ROCE),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_ssu_rx_th_nic, SSU_RX_NIC),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_ssu_rx_th_roce, SSU_RX_ROCE),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rpu, RPU),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_tpu, TPU),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rpe, RPE),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_tpe_normal, TPE),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_tpe_push, TPE_PUSH),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_wr_fbd, WR_FBD),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_wr_ebd, WR_EBD),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rd_fbd, RD_FBD),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rd_ebd, RD_EBD),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_wr_pay_m1, WR_PAY_M1),
>> +    HNS3_PMU_DLY_EVT_PAIR(dly_msix_write, MSIX_WRITE),
>> +
>> +    /* interrupt rate events */
>> +    HNS3_PMU_INTR_EVT_PAIR(pps_intr_msix_nic, MSIX_NIC),
>> +
>> +    NULL
>> +};
>> +
>> +static struct attribute *hns3_pmu_filter_mode_attr[] = {
>> +    /* bandwidth events */
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_egu, SSU_EGU),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_rpu, SSU_RPU),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_roce, SSU_ROCE),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_roce_ssu, ROCE_SSU),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_tpu_ssu, TPU_SSU),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rpu_rcbrx, RPU_RCBRX),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rcbtx_txsch, RCBTX_TXSCH),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_fbd, WR_FBD),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_ebd, WR_EBD),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_fbd, RD_FBD),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_ebd, RD_EBD),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_pay_m1, WR_PAY_M1),
>> +
>> +    /* packet rate events */
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_igu_ssu, IGU_SSU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_egu, SSU_EGU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_rpu, SSU_RPU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_roce, SSU_ROCE),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_roce_ssu, ROCE_SSU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_tpu_ssu, TPU_SSU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rpu_rcbrx, RPU_RCBRX),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rcbtx_tpu, RCBTX_TPU),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rcbtx_txsch, RCBTX_TXSCH),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_fbd, WR_FBD),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_ebd, WR_EBD),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_fbd, RD_FBD),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_ebd, RD_EBD),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_pay_m1, WR_PAY_M1),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_intr_nicroh_tx_pre, NICROH_TX_PRE),
>> +    HNS3_PMU_PPS_FLT_MODE_PAIR(pps_intr_nicroh_rx_pre, NICROH_RX_PRE),
>> +
>> +    /* latency events */
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tx_push_to_mac, TX_PUSH),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tx_normal_to_mac, TX),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_tx_th_nic, SSU_TX_NIC),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_tx_th_roce, SSU_TX_ROCE),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_rx_th_nic, SSU_RX_NIC),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_rx_th_roce, SSU_RX_ROCE),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rpu, RPU),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpu, TPU),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rpe, RPE),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpe_normal, TPE),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpe_push, TPE_PUSH),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_fbd, WR_FBD),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_ebd, WR_EBD),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_fbd, RD_FBD),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_ebd, RD_EBD),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_pay_m0, RD_PAY_M0),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_pay_m1, RD_PAY_M1),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_pay_m0, WR_PAY_M0),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_pay_m1, WR_PAY_M1),
>> +    HNS3_PMU_DLY_FLT_MODE_PAIR(dly_msix_write, MSIX_WRITE),
>> +
>> +    /* interrupt rate events */
>> +    HNS3_PMU_INTR_FLT_MODE_PAIR(pps_intr_msix_nic, MSIX_NIC),
>> +
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_events_group = {
>> +    .name = "events",
>> +    .attrs = hns3_pmu_events_attr,
>> +};
>> +
>> +static struct attribute_group hns3_pmu_filter_mode_group = {
>> +    .name = "filtermode",
>> +    .attrs = hns3_pmu_filter_mode_attr,
>> +};
>> +
>> +static struct attribute *hns3_pmu_format_attr[] = {
>> +    HNS3_PMU_FORMAT_ATTR(event, "config:0-16"),
>> +    HNS3_PMU_FORMAT_ATTR(port, "config1:0-3"),
>> +    HNS3_PMU_FORMAT_ATTR(tc, "config1:4-7"),
>> +    HNS3_PMU_FORMAT_ATTR(bdf, "config1:8-23"),
>> +    HNS3_PMU_FORMAT_ATTR(queue, "config1:24-39"),
>> +    HNS3_PMU_FORMAT_ATTR(intr, "config1:40-51"),
>> +    HNS3_PMU_FORMAT_ATTR(global, "config1:52-52"),
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_format_group = {
>> +    .name = "format",
>> +    .attrs = hns3_pmu_format_attr,
>> +};
>> +
>> +static struct attribute *hns3_pmu_cpumask_attrs[] = {
>> +    &dev_attr_cpumask.attr,
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_cpumask_attr_group = {
>> +    .attrs = hns3_pmu_cpumask_attrs,
>> +};
>> +
>> +static struct attribute *hns3_pmu_identifier_attrs[] = {
>> +    &dev_attr_identifier.attr,
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_identifier_attr_group = {
>> +    .attrs = hns3_pmu_identifier_attrs,
>> +};
>> +
>> +static struct attribute *hns3_pmu_bdf_range_attrs[] = {
>> +    &dev_attr_bdf_min.attr,
>> +    &dev_attr_bdf_max.attr,
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_bdf_range_attr_group = {
>> +    .attrs = hns3_pmu_bdf_range_attrs,
>> +};
>> +
>> +static struct attribute *hns3_pmu_hw_clk_freq_attrs[] = {
>> +    &dev_attr_hw_clk_freq.attr,
>> +    NULL
>> +};
>> +
>> +static struct attribute_group hns3_pmu_hw_clk_freq_attr_group = {
>> +    .attrs = hns3_pmu_hw_clk_freq_attrs,
>> +};
>> +
>> +static const struct attribute_group *hns3_pmu_attr_groups[] = {
>> +    &hns3_pmu_events_group,
>> +    &hns3_pmu_filter_mode_group,
>> +    &hns3_pmu_format_group,
>> +    &hns3_pmu_cpumask_attr_group,
>> +    &hns3_pmu_identifier_attr_group,
>> +    &hns3_pmu_bdf_range_attr_group,
>> +    &hns3_pmu_hw_clk_freq_attr_group,
>> +    NULL
>> +};
>> +
>> +static u32 hns3_pmu_get_offset(u32 offset, u32 idx)
>> +{
>> +    return offset + HNS3_PMU_REG_EVENT_OFFSET +
>> +           HNS3_PMU_REG_EVENT_SIZE * idx;
>> +}
>> +
>> +static u32 hns3_pmu_readl(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx)
>> +{
>> +    u32 offset = hns3_pmu_get_offset(reg_offset, idx);
>> +
>> +    return readl(hns3_pmu->base + offset);
>> +}
>> +
>> +static void hns3_pmu_writel(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx,
>> +                u32 val)
>> +{
>> +    u32 offset = hns3_pmu_get_offset(reg_offset, idx);
>> +
>> +    writel(val, hns3_pmu->base + offset);
>> +}
>> +
>> +static u64 hns3_pmu_readq(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx)
>> +{
>> +    u32 offset = hns3_pmu_get_offset(reg_offset, idx);
>> +
>> +    return readq(hns3_pmu->base + offset);
>> +}
>> +
>> +static void hns3_pmu_writeq(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx,
>> +                u64 val)
>> +{
>> +    u32 offset = hns3_pmu_get_offset(reg_offset, idx);
>> +
>> +    writeq(val, hns3_pmu->base + offset);
>> +}
>> +
>> +static bool hns3_pmu_cmp_event(struct perf_event *target,
>> +                   struct perf_event *event)
>> +{
>> +    return hns3_get_real_event(target) == hns3_get_real_event(event);
>> +}
>> +
>> +static int hns3_pmu_find_related_event_idx(struct hns3_pmu *hns3_pmu,
>> +                       struct perf_event *event)
>> +{
>> +    struct perf_event *sibling;
>> +    int hw_event_used = 0;
>> +    int idx;
>> +
>> +    for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) {
>> +        sibling = hns3_pmu->hw_events[idx];
>> +        if (!sibling)
>> +            continue;
>> +
>> +        hw_event_used++;
>> +
>> +        if (!hns3_pmu_cmp_event(sibling, event))
>> +            continue;
>> +
>> +        /* Related events is used in group */
>> +        if (sibling->group_leader == event->group_leader)
>> +            return idx;
>> +    }
>> +
>> +    /* No related event and all hardware events are used up */
>> +    if (hw_event_used >= HNS3_PMU_MAX_HW_EVENTS)
>> +        return -EBUSY;
>> +
>> +    /* No related event and there is extra hardware events can be use */
>> +    return -ENOENT;
>> +}
>> +
>> +static int hns3_pmu_get_event_idx(struct hns3_pmu *hns3_pmu)
>> +{
>> +    int idx;
>> +
>> +    for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) {
>> +        if (!hns3_pmu->hw_events[idx])
>> +            return idx;
>> +    }
>> +
>> +    return -EBUSY;
>> +}
>> +
>> +static bool hns3_pmu_valid_bdf(struct hns3_pmu *hns3_pmu, u16 bdf)
>> +{
>> +    struct pci_dev *pdev;
>> +
>> +    if (bdf < hns3_pmu->bdf_min || bdf > hns3_pmu->bdf_max) {
>> +        pci_err(hns3_pmu->pdev, "Invalid EP device: %#x!\n", bdf);
>> +        return false;
>> +    }
>> +
>> +    pdev = pci_get_domain_bus_and_slot(pci_domain_nr(hns3_pmu->pdev->bus),
>> +                       PCI_BUS_NUM(bdf),
>> +                       GET_PCI_DEVFN(bdf));
>> +    if (!pdev) {
>> +        pci_err(hns3_pmu->pdev, "Nonexistent EP device: %#x!\n", bdf);
>> +        return false;
>> +    }
>> +
>> +    pci_dev_put(pdev);
>> +    return true;
>> +}
>> +
>> +static void hns3_pmu_set_qid_para(struct hns3_pmu *hns3_pmu, u32 idx, u16 bdf,
>> +                  u16 queue)
>> +{
>> +    u32 val;
>> +
>> +    val = GET_PCI_DEVFN(bdf);
>> +    val |= (u32)queue << HNS3_PMU_QID_PARA_QUEUE_S;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_PARA, idx, val);
>> +}
>> +
>> +static bool hns3_pmu_qid_req_start(struct hns3_pmu *hns3_pmu, u32 idx)
>> +{
>> +    bool queue_id_valid = false;
>> +    u32 reg_qid_ctrl, val;
>> +    int err;
>> +
>> +    /* enable queue id request */
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_CTRL, idx,
>> +            HNS3_PMU_QID_CTRL_REQ_ENABLE);
>> +
>> +    reg_qid_ctrl = hns3_pmu_get_offset(HNS3_PMU_REG_EVENT_QID_CTRL, idx);
>> +    err = readl_poll_timeout(hns3_pmu->base + reg_qid_ctrl, val,
>> +                 val & HNS3_PMU_QID_CTRL_DONE, 1, 1000);
>> +    if (err == -ETIMEDOUT) {
>> +        pci_err(hns3_pmu->pdev, "QID request timeout!\n");
>> +        goto out;
>> +    }
>> +
>> +    queue_id_valid = !(val & HNS3_PMU_QID_CTRL_MISS);
>> +
>> +out:
>> +    /* disable qid request and clear status */
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_CTRL, idx, 0);
>> +
>> +    return queue_id_valid;
>> +}
>> +
>> +static bool hns3_pmu_valid_queue(struct hns3_pmu *hns3_pmu, u32 idx, u16 bdf,
>> +                 u16 queue)
>> +{
>> +    hns3_pmu_set_qid_para(hns3_pmu, idx, bdf, queue);
>> +
>> +    return hns3_pmu_qid_req_start(hns3_pmu, idx);
>> +}
>> +
>> +static struct hns3_pmu_event_attr *hns3_pmu_get_pmu_event(u32 event)
>> +{
>> +    struct hns3_pmu_event_attr *pmu_event;
>> +    struct dev_ext_attribute *eattr;
>> +    struct device_attribute *dattr;
>> +    struct attribute *attr;
>> +    u32 i;
>> +
>> +    for (i = 0; i < ARRAY_SIZE(hns3_pmu_events_attr) - 1; i++) {
>> +        attr = hns3_pmu_events_attr[i];
>> +        dattr = container_of(attr, struct device_attribute, attr);
>> +        eattr = container_of(dattr, struct dev_ext_attribute, attr);
>> +        pmu_event = eattr->var;
>> +
>> +        if (event == pmu_event->event)
>> +            return pmu_event;
>> +    }
>> +
>> +    return NULL;
>> +}
>> +
>> +static int hns3_pmu_set_func_mode(struct perf_event *event,
>> +                  struct hns3_pmu *hns3_pmu)
>> +{
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u16 bdf = hns3_get_bdf(event);
>> +
>> +    if (!hns3_pmu_valid_bdf(hns3_pmu, bdf))
>> +        return -ENOENT;
>> +
>> +    HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC);
>> +
>> +    return 0;
>> +}
>> +
>> +static int hns3_pmu_set_func_queue_mode(struct perf_event *event,
>> +                    struct hns3_pmu *hns3_pmu)
>> +{
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u16 queue_id = hns3_get_queue(event);
>> +    u16 bdf = hns3_get_bdf(event);
>> +
>> +    if (!hns3_pmu_valid_bdf(hns3_pmu, bdf))
>> +        return -ENOENT;
>> +
>> +    if (!hns3_pmu_valid_queue(hns3_pmu, hwc->idx, bdf, queue_id)) {
>> +        pci_err(hns3_pmu->pdev, "Invalid queue: %u\n", queue_id);
>> +        return -ENOENT;
>> +    }
>> +
>> +    HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC_QUEUE);
>> +
>> +    return 0;
>> +}
>> +
>> +static bool
>> +hns3_pmu_is_enabled_global_mode(struct perf_event *event,
>> +                struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u8 global = hns3_get_global(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_GLOBAL))
>> +        return false;
>> +
>> +    return global;
>> +}
>> +
>> +static bool hns3_pmu_is_enabled_func_mode(struct perf_event *event,
>> +                      struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u16 queue_id = hns3_get_queue(event);
>> +    u16 bdf = hns3_get_bdf(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC))
>> +        return false;
>> +    else if (queue_id != HNS3_PMU_FILTER_ALL_QUEUE)
>> +        return false;
>> +
>> +    return bdf;
> 
> nit: I know I suggested this but it looks odd as normally we would take benefit of the compiler converting to a bool like this:
> 
> bool foo(int var)
> {
>      return var & 0x2;
> }
> 
> So not explicitly using u16 or similar for the return value.
> 
> Anyway, I'm fine with it as long as others are.
> 
Ok, then I keep it.


>> +}
>> +
>> +static bool
>> +hns3_pmu_is_enabled_func_queue_mode(struct perf_event *event,
>> +                    struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u16 queue_id = hns3_get_queue(event);
>> +    u16 bdf = hns3_get_bdf(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE))
>> +        return false;
>> +    else if (queue_id == HNS3_PMU_FILTER_ALL_QUEUE)
>> +        return false;
>> +
>> +    return bdf;
>> +}
>> +
>> +static bool hns3_pmu_is_enabled_port_mode(struct perf_event *event,
>> +                      struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u8 tc_id = hns3_get_tc(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT))
>> +        return false;
>> +
>> +    return tc_id == HNS3_PMU_FILTER_ALL_TC;
>> +}
>> +
>> +static bool
>> +hns3_pmu_is_enabled_port_tc_mode(struct perf_event *event,
>> +                 struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u8 tc_id = hns3_get_tc(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT_TC))
>> +        return false;
>> +
>> +    return tc_id != HNS3_PMU_FILTER_ALL_TC;
>> +}
>> +
>> +static bool
>> +hns3_pmu_is_enabled_func_intr_mode(struct perf_event *event,
>> +                   struct hns3_pmu *hns3_pmu,
>> +                   struct hns3_pmu_event_attr *pmu_event)
>> +{
>> +    u16 bdf = hns3_get_bdf(event);
>> +
>> +    if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_INTR))
>> +        return false;
>> +
>> +    return hns3_pmu_valid_bdf(hns3_pmu, bdf);
>> +}
>> +
>> +static int hns3_pmu_select_filter_mode(struct perf_event *event,
>> +                       struct hns3_pmu *hns3_pmu)
>> +{
>> +    struct hns3_pmu_event_attr *pmu_event;
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u32 event_id = hns3_get_event(event);
>> +
>> +    pmu_event = hns3_pmu_get_pmu_event(event_id);
>> +    if (!pmu_event) {
>> +        pci_err(hns3_pmu->pdev, "Invalid pmu event\n");
>> +        return -ENOENT;
>> +    }
>> +
>> +    if (hns3_pmu_is_enabled_global_mode(event, pmu_event)) {
>> +        HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_GLOBAL);
>> +        return 0;
>> +    }
>> +
>> +    if (hns3_pmu_is_enabled_func_mode(event, pmu_event))
>> +        return hns3_pmu_set_func_mode(event, hns3_pmu);
>> +
>> +    if (hns3_pmu_is_enabled_func_queue_mode(event, pmu_event))
>> +        return hns3_pmu_set_func_queue_mode(event, hns3_pmu);
>> +
>> +    if (hns3_pmu_is_enabled_port_mode(event, pmu_event)) {
>> +        HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_PORT);
>> +        return 0;
>> +    }
>> +
>> +    if (hns3_pmu_is_enabled_port_tc_mode(event, pmu_event)) {
>> +        HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_PORT_TC);
>> +        return 0;
>> +    }
>> +
>> +    if (hns3_pmu_is_enabled_func_intr_mode(event, hns3_pmu, pmu_event)) {
>> +        HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC_INTR);
>> +        return 0;
>> +    }
>> +
>> +    return -ENOENT;
>> +}
>> +
>> +static bool hns3_pmu_validate_event_group(struct perf_event *event)
>> +{
>> +    struct perf_event *sibling, *leader = event->group_leader;
>> +    struct perf_event *event_group[HNS3_PMU_MAX_HW_EVENTS];
>> +    int counters = 1;
>> +    int num;
>> +
>> +    event_group[0] = leader;
>> +    if (!is_software_event(leader)) {
>> +        if (leader->pmu != event->pmu)
>> +            return false;
>> +
>> +        if (leader != event && !hns3_pmu_cmp_event(leader, event))
>> +            event_group[counters++] = event;
>> +    }
>> +
>> +    for_each_sibling_event(sibling, event->group_leader) {
>> +        if (is_software_event(sibling))
>> +            continue;
>> +
>> +        if (sibling->pmu != event->pmu)
>> +            return false;
>> +
>> +        for (num = 0; num < counters; num++) {
>> +            if (hns3_pmu_cmp_event(event_group[num], sibling))
>> +                break;
>> +        }
>> +
>> +        if (num == counters)
>> +            event_group[counters++] = sibling;
>> +    }
>> +
>> +    return counters <= HNS3_PMU_MAX_HW_EVENTS;
>> +}
>> +
>> +static u32 hns3_pmu_get_filter_condition(struct perf_event *event)
>> +{
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u16 intr_id = hns3_get_intr(event);
>> +    u8 port_id = hns3_get_port(event);
>> +    u16 bdf = hns3_get_bdf(event);
>> +    u8 tc_id = hns3_get_tc(event);
>> +    u8 filter_mode;
>> +    u32 filter = 0;
>> +
>> +    filter_mode = *(u8 *)hwc->addr_filters;
>> +    switch (filter_mode) {
>> +    case HNS3_PMU_HW_FILTER_PORT:
>> +        filter = FILTER_CONDITION_PORT(port_id);
> 
> nit: this code can be simplified by returning directly
> 
Ok, all cases of this function will be returned directly.

>> +        break;
>> +    case HNS3_PMU_HW_FILTER_PORT_TC:
>> +        filter = FILTER_CONDITION_PORT_TC(port_id, tc_id);
>> +        break;
>> +    case HNS3_PMU_HW_FILTER_FUNC:
>> +    case HNS3_PMU_HW_FILTER_FUNC_QUEUE:
>> +        filter = GET_PCI_DEVFN(bdf);
> 
> and here and so on
> 
>> +        break;
>> +    case HNS3_PMU_HW_FILTER_FUNC_INTR:
>> +        filter = FILTER_CONDITION_FUNC_INTR(GET_PCI_DEVFN(bdf),
>> +                            intr_id);
>> +        break;
>> +    default:
>> +        break;
>> +    }
>> +
>> +    return filter;
>> +}
>> +
>> +static void hns3_pmu_config_filter(struct perf_event *event)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    u8 event_type = hns3_get_event_type(event);
>> +    u8 subevent_id = hns3_get_subevent(event);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u8 filter_mode = *(u8 *)hwc->addr_filters;
>> +    u16 queue_id = hns3_get_queue(event);
>> +    u16 bdf = hns3_get_bdf(event);
>> +    u32 idx = hwc->idx;
>> +    u32 val;
>> +
>> +    val = event_type;
>> +    val |= subevent_id << HNS3_PMU_CTRL_SUBEVENT_S;
>> +    val |= filter_mode << HNS3_PMU_CTRL_FILTER_MODE_S;
>> +    val |= HNS3_PMU_EVENT_OVERFLOW_RESTART;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val);
>> +
>> +    val = hns3_pmu_get_filter_condition(event);
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_HIGH, idx, val);
>> +
>> +    if (filter_mode == HNS3_PMU_HW_FILTER_FUNC_QUEUE)
>> +        hns3_pmu_set_qid_para(hns3_pmu, idx, bdf, queue_id);
>> +}
>> +
>> +static void hns3_pmu_enable_counter(struct hns3_pmu *hns3_pmu,
>> +                    struct hw_perf_event *hwc)
>> +{
>> +    u32 idx = hwc->idx;
>> +    u32 val;
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx);
>> +    val |= HNS3_PMU_EVENT_EN;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val);
>> +}
>> +
>> +static void hns3_pmu_disable_counter(struct hns3_pmu *hns3_pmu,
>> +                     struct hw_perf_event *hwc)
>> +{
>> +    u32 idx = hwc->idx;
>> +    u32 val;
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx);
>> +    val &= ~HNS3_PMU_EVENT_EN;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val);
>> +}
>> +
>> +static void hns3_pmu_enable_intr(struct hns3_pmu *hns3_pmu,
>> +                 struct hw_perf_event *hwc)
>> +{
>> +    u32 idx = hwc->idx;
>> +    u32 val;
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx);
>> +    val &= ~HNS3_PMU_INTR_MASK_OVERFLOW;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx, val);
>> +}
>> +
>> +static void hns3_pmu_disable_intr(struct hns3_pmu *hns3_pmu,
>> +                  struct hw_perf_event *hwc)
>> +{
>> +    u32 idx = hwc->idx;
>> +    u32 val;
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx);
>> +    val |= HNS3_PMU_INTR_MASK_OVERFLOW;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx, val);
>> +}
>> +
>> +static void hns3_pmu_clear_intr_status(struct hns3_pmu *hns3_pmu, u32 idx)
>> +{
>> +    u32 val;
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx);
>> +    val |= HNS3_PMU_EVENT_STATUS_RESET;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val);
>> +
>> +    val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx);
>> +    val &= ~HNS3_PMU_EVENT_STATUS_RESET;
>> +    hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val);
>> +}
>> +
>> +static u64 hns3_pmu_read_counter(struct perf_event *event)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +
>> +    return hns3_pmu_readq(hns3_pmu, event->hw.event_base, event->hw.idx);
>> +}
>> +
>> +static void hns3_pmu_write_counter(struct perf_event *event, u64 value)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    u32 idx = event->hw.idx;
>> +
>> +    hns3_pmu_writeq(hns3_pmu, HNS3_PMU_REG_EVENT_COUNTER, idx, value);
>> +    hns3_pmu_writeq(hns3_pmu, HNS3_PMU_REG_EVENT_EXT_COUNTER, idx, value);
>> +}
>> +
>> +static void hns3_pmu_init_counter(struct perf_event *event)
>> +{
>> +    struct hw_perf_event *hwc = &event->hw;
>> +
>> +    local64_set(&hwc->prev_count, 0);
>> +    hns3_pmu_write_counter(event, 0);
>> +}
>> +
>> +static int hns3_pmu_event_init(struct perf_event *event)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    int idx;
>> +    int ret;
>> +
>> +    if (event->attr.type != event->pmu->type)
>> +        return -ENOENT;
>> +
>> +    /* Sampling is not supported */
>> +    if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK)
>> +        return -EOPNOTSUPP;
>> +
>> +    event->cpu = hns3_pmu->on_cpu;
>> +
>> +    idx = hns3_pmu_get_event_idx(hns3_pmu);
>> +    if (idx < 0) {
>> +        pci_err(hns3_pmu->pdev, "Up to %u events are supported!\n",
>> +            HNS3_PMU_MAX_HW_EVENTS);
>> +        return -EBUSY;
>> +    }
>> +
>> +    hwc->idx = idx;
>> +
>> +    ret = hns3_pmu_select_filter_mode(event, hns3_pmu);
>> +    if (ret) {
>> +        pci_err(hns3_pmu->pdev, "Invalid filter, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    if (!hns3_pmu_validate_event_group(event)) {
>> +        pci_err(hns3_pmu->pdev, "Invalid event group.\n");
>> +        return -EINVAL;
>> +    }
>> +
>> +    if (hns3_get_ext_counter_used(event))
>> +        hwc->event_base = HNS3_PMU_REG_EVENT_EXT_COUNTER;
>> +    else
>> +        hwc->event_base = HNS3_PMU_REG_EVENT_COUNTER;
>> +
>> +    return 0;
>> +}
>> +
>> +static void hns3_pmu_read(struct perf_event *event)
>> +{
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    u64 new_cnt, prev_cnt, delta;
>> +
>> +    do {
>> +        prev_cnt = local64_read(&hwc->prev_count);
>> +        new_cnt = hns3_pmu_read_counter(event);
>> +    } while (local64_cmpxchg(&hwc->prev_count, prev_cnt, new_cnt) !=
>> +         prev_cnt);
>> +
>> +    delta = new_cnt - prev_cnt;
>> +    local64_add(delta, &event->count);
>> +}
>> +
>> +static void hns3_pmu_start(struct perf_event *event, int flags)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +
>> +    if (WARN_ON_ONCE(!(hwc->state & PERF_HES_STOPPED)))
>> +        return;
>> +
>> +    WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE));
>> +    hwc->state = 0;
>> +
>> +    hns3_pmu_config_filter(event);
>> +    hns3_pmu_init_counter(event);
>> +    hns3_pmu_enable_intr(hns3_pmu, hwc);
>> +    hns3_pmu_enable_counter(hns3_pmu, hwc);
>> +
>> +    perf_event_update_userpage(event);
>> +}
>> +
>> +static void hns3_pmu_stop(struct perf_event *event, int flags)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +
>> +    hns3_pmu_disable_counter(hns3_pmu, hwc);
>> +    hns3_pmu_disable_intr(hns3_pmu, hwc);
>> +
>> +    WARN_ON_ONCE(hwc->state & PERF_HES_STOPPED);
>> +    hwc->state |= PERF_HES_STOPPED;
>> +
>> +    if (hwc->state & PERF_HES_UPTODATE)
>> +        return;
>> +
>> +    /* Read hardware counter and update the perf counter statistics */
>> +    hns3_pmu_read(event);
>> +    hwc->state |= PERF_HES_UPTODATE;
>> +}
>> +
>> +static int hns3_pmu_add(struct perf_event *event, int flags)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +    int idx;
>> +
>> +    hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE;
>> +
>> +    /* Check all working events to find a related event. */
>> +    idx = hns3_pmu_find_related_event_idx(hns3_pmu, event);
>> +    if (idx < 0 && idx != -ENOENT)
>> +        return idx;
>> +
>> +    /* Current event shares an enabled hardware event with related event */
>> +    if (idx >= 0 && idx < HNS3_PMU_MAX_HW_EVENTS) {
>> +        hwc->idx = idx;
>> +        goto start_count;
>> +    }
>> +
>> +    idx = hns3_pmu_get_event_idx(hns3_pmu);
>> +    if (idx < 0)
>> +        return idx;
>> +
>> +    hwc->idx = idx;
>> +    hns3_pmu->hw_events[idx] = event;
>> +
>> +start_count:
>> +    if (flags & PERF_EF_START)
>> +        hns3_pmu_start(event, PERF_EF_RELOAD);
>> +
>> +    return 0;
>> +}
>> +
>> +static void hns3_pmu_del(struct perf_event *event, int flags)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu);
>> +    struct hw_perf_event *hwc = &event->hw;
>> +
>> +    hns3_pmu_stop(event, PERF_EF_UPDATE);
>> +    hns3_pmu->hw_events[hwc->idx] = NULL;
>> +    perf_event_update_userpage(event);
>> +}
>> +
>> +static void hns3_pmu_enable(struct pmu *pmu)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(pmu);
>> +    u32 val;
>> +
>> +    val = readl(hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL);
>> +    val |= HNS3_PMU_GLOBAL_START;
>> +    writel(val, hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL);
>> +}
>> +
>> +static void hns3_pmu_disable(struct pmu *pmu)
>> +{
>> +    struct hns3_pmu *hns3_pmu = to_hns3_pmu(pmu);
>> +    u32 val;
>> +
>> +    val = readl(hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL);
>> +    val &= ~HNS3_PMU_GLOBAL_START;
>> +    writel(val, hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL);
>> +}
>> +
>> +static int hns3_pmu_alloc_pmu(struct pci_dev *pdev, struct hns3_pmu *hns3_pmu)
>> +{
>> +    u16 device_id;
>> +    char *name;
>> +    u32 val;
>> +
>> +    hns3_pmu->base = pcim_iomap_table(pdev)[BAR_2];
>> +    if (!hns3_pmu->base) {
>> +        pci_err(pdev, "ioremap failed for hns3_pmu resource\n");
> 
> nit: I think that the driver name is mentioned in the pci dev print already so no need to mention hns3_pmu
> 
Ok.

>> +        return -ENOMEM;
>> +    }
>> +
>> +    hns3_pmu->hw_clk_freq = readl(hns3_pmu->base + HNS3_PMU_REG_CLOCK_FREQ);
>> +
>> +    val = readl(hns3_pmu->base + HNS3_PMU_REG_BDF);
>> +    hns3_pmu->bdf_min = val & 0xffff;
>> +    hns3_pmu->bdf_max = val >> 16;
>> +
>> +    val = readl(hns3_pmu->base + HNS3_PMU_REG_DEVICE_ID);
>> +    device_id = val & 0xffff;
>> +    name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "hns3_pmu_sicl_%u", device_id);
>> +    if (!name)
>> +        return -ENOMEM;
>> +
>> +    hns3_pmu->pdev = pdev;
>> +    hns3_pmu->on_cpu = -1;
>> +    hns3_pmu->identifier = readl(hns3_pmu->base + HNS3_PMU_REG_VERSION);
>> +    hns3_pmu->pmu = (struct pmu) {
>> +        .name        = name,
>> +        .module        = THIS_MODULE,
>> +        .event_init    = hns3_pmu_event_init,
>> +        .pmu_enable    = hns3_pmu_enable,
>> +        .pmu_disable    = hns3_pmu_disable,
>> +        .add        = hns3_pmu_add,
>> +        .del        = hns3_pmu_del,
>> +        .start        = hns3_pmu_start,
>> +        .stop        = hns3_pmu_stop,
>> +        .read        = hns3_pmu_read,
>> +        .task_ctx_nr    = perf_invalid_context,
>> +        .attr_groups    = hns3_pmu_attr_groups,
>> +        .capabilities    = PERF_PMU_CAP_NO_EXCLUDE,
>> +    };
>> +
>> +    return 0;
>> +}
>> +
>> +static irqreturn_t hns3_pmu_irq(int irq, void *data)
>> +{
>> +    struct hns3_pmu *hns3_pmu = data;
>> +    u32 intr_status, idx;
>> +
>> +    for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) {
>> +        intr_status = hns3_pmu_readl(hns3_pmu,
>> +                         HNS3_PMU_REG_EVENT_INTR_STATUS,
>> +                         idx);
>> +
>> +        /*
>> +         * As each counter will restart from 0 when it is overflowed,
>> +         * extra processing is no need, just clear interrupt status.
>> +         */
>> +        if (intr_status)
>> +            hns3_pmu_clear_intr_status(hns3_pmu, idx);
>> +    }
>> +
>> +    return IRQ_HANDLED;
>> +}
>> +
>> +static int hns3_pmu_online_cpu(unsigned int cpu, struct hlist_node *node)
>> +{
>> +    struct hns3_pmu *hns3_pmu;
>> +
>> +    hns3_pmu = hlist_entry_safe(node, struct hns3_pmu, node);
>> +    if (!hns3_pmu)
>> +        return -ENODEV;
>> +
>> +    if (hns3_pmu->on_cpu == -1) {
>> +        hns3_pmu->on_cpu = cpu;
>> +        irq_set_affinity(hns3_pmu->irq, cpumask_of(cpu));
>> +    }
>> +
>> +    return 0;
>> +}
>> +
>> +static int hns3_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
>> +{
>> +    struct hns3_pmu *hns3_pmu;
>> +    unsigned int target;
>> +
>> +    hns3_pmu = hlist_entry_safe(node, struct hns3_pmu, node);
>> +    if (!hns3_pmu)
>> +        return -ENODEV;
>> +
>> +    /* Nothing to do if this CPU doesn't own the PMU */
>> +    if (hns3_pmu->on_cpu != cpu)
>> +        return 0;
>> +
>> +    /* Choose a new CPU from all online cpus */
>> +    target = cpumask_any_but(cpu_online_mask, cpu);
>> +    if (target >= nr_cpu_ids)
>> +        return 0;
>> +
>> +    perf_pmu_migrate_context(&hns3_pmu->pmu, cpu, target);
>> +    hns3_pmu->on_cpu = target;
>> +    irq_set_affinity(hns3_pmu->irq, cpumask_of(target));
>> +
>> +    return 0;
>> +}
>> +
>> +static void hns3_pmu_free_irq(void *data)
>> +{
>> +    struct pci_dev *pdev = data;
>> +
>> +    pci_free_irq_vectors(pdev);
>> +}
>> +
>> +static int hns3_pmu_irq_register(struct pci_dev *pdev,
>> +                 struct hns3_pmu *hns3_pmu)
>> +{
>> +    int irq, ret;
>> +
>> +    ret = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_MSI);
>> +    if (ret < 0) {
>> +        pci_err(pdev, "failed to enable MSI vectors, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    ret = devm_add_action(&pdev->dev, hns3_pmu_free_irq, pdev);
>> +    if (ret) {
>> +        pci_err(pdev, "failed to add free irq action, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    irq = pci_irq_vector(pdev, 0);
>> +    ret = devm_request_irq(&pdev->dev, irq, hns3_pmu_irq, 0,
>> +                   hns3_pmu->pmu.name, hns3_pmu);
>> +    if (ret) {
>> +        pci_err(pdev, "failed to register irq, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    hns3_pmu->irq = irq;
>> +
>> +    return 0;
>> +}
>> +
>> +static int hns3_pmu_init_pmu(struct pci_dev *pdev, struct hns3_pmu *hns3_pmu)
>> +{
>> +    int ret;
>> +
>> +    ret = hns3_pmu_alloc_pmu(pdev, hns3_pmu);
>> +    if (ret)
>> +        return ret;
>> +
>> +    ret = hns3_pmu_irq_register(pdev, hns3_pmu);
>> +    if (ret)
>> +        return ret;
>> +
>> +    ret = cpuhp_state_add_instance(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
>> +                       &hns3_pmu->node);
>> +    if (ret) {
>> +        pci_err(pdev, "failed to register hotplug, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    ret = perf_pmu_register(&hns3_pmu->pmu, hns3_pmu->pmu.name, -1);
>> +    if (ret) {
>> +        pci_err(pdev, "failed to register perf PMU, ret = %d.\n", ret);
>> +        cpuhp_state_remove_instance(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
>> +                        &hns3_pmu->node);
>> +    }
>> +
>> +    return ret;
>> +}
>> +
>> +static void hns3_pmu_uninit_pmu(struct pci_dev *pdev)
>> +{
>> +    struct hns3_pmu *hns3_pmu = pci_get_drvdata(pdev);
>> +
>> +    perf_pmu_unregister(&hns3_pmu->pmu);
>> +    cpuhp_state_remove_instance(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
>> +                    &hns3_pmu->node);
>> +}
>> +
>> +static int hns3_pmu_init_dev(struct pci_dev *pdev)
>> +{
>> +    int ret;
>> +
>> +    ret = pcim_enable_device(pdev);
>> +    if (ret) {
>> +        pci_err(pdev, "failed to enable pci device, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    ret = pcim_iomap_regions(pdev, BIT(BAR_2), "hns3_pmu");
>> +    if (ret < 0) {
>> +        pci_err(pdev, "failed to request pci region, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    pci_set_master(pdev);
>> +
>> +    return 0;
>> +}
>> +
>> +static int hns3_pmu_probe(struct pci_dev *pdev, const struct pci_device_id *id)
>> +{
>> +    struct hns3_pmu *hns3_pmu;
>> +    int ret;
>> +
>> +    hns3_pmu = devm_kzalloc(&pdev->dev, sizeof(*hns3_pmu), GFP_KERNEL);
>> +    if (!hns3_pmu)
>> +        return -ENOMEM;
>> +
>> +    ret = hns3_pmu_init_dev(pdev);
>> +    if (ret)
>> +        return ret;
>> +
>> +    ret = hns3_pmu_init_pmu(pdev, hns3_pmu);
>> +    if (ret) {
>> +        pci_clear_master(pdev);
>> +        return ret;
>> +    }
>> +
>> +    pci_set_drvdata(pdev, hns3_pmu);
>> +
>> +    return ret;
>> +}
>> +
>> +static void hns3_pmu_remove(struct pci_dev *pdev)
>> +{
>> +    hns3_pmu_uninit_pmu(pdev);
>> +    pci_clear_master(pdev);
>> +    pci_set_drvdata(pdev, NULL);
>> +}
>> +
>> +static const struct pci_device_id hns3_pmu_ids[] = {
>> +    { PCI_DEVICE(PCI_VENDOR_ID_HUAWEI, 0xa22b) },
>> +    { 0, }
>> +};
>> +MODULE_DEVICE_TABLE(pci, hns3_pmu_ids);
>> +
>> +static struct pci_driver hns3_pmu_driver = {
>> +    .name = "hns3_pmu",
>> +    .id_table = hns3_pmu_ids,
>> +    .probe = hns3_pmu_probe,
>> +    .remove = hns3_pmu_remove,
>> +};
>> +
>> +static int __init hns3_pmu_module_init(void)
>> +{
>> +    int ret;
>> +
>> +    ret = cpuhp_setup_state_multi(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
>> +                      "AP_PERF_ARM_HNS3_PMU_ONLINE",
>> +                      hns3_pmu_online_cpu,
>> +                      hns3_pmu_offline_cpu);
>> +    if (ret) {
>> +        pr_err("failed to setup HNS3 PMU hotplug, ret = %d.\n", ret);
>> +        return ret;
>> +    }
>> +
>> +    ret = pci_register_driver(&hns3_pmu_driver);
>> +    if (ret) {
>> +        pr_err("failed to register pci driver, ret = %d.\n", ret);
>> +        cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE);
>> +    }
>> +
>> +    return ret;
>> +}
>> +module_init(hns3_pmu_module_init);
>> +
>> +static void __exit hns3_pmu_module_exit(void)
>> +{
>> +    pci_unregister_driver(&hns3_pmu_driver);
>> +    cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE);
>> +}
>> +module_exit(hns3_pmu_module_exit);
>> +
>> +MODULE_DESCRIPTION("HNS3 PMU driver");
>> +MODULE_LICENSE("GPL v2");
>> diff --git a/include/linux/cpuhotplug.h b/include/linux/cpuhotplug.h
>> index 411a428ace4d..9c2cb0600740 100644
>> --- a/include/linux/cpuhotplug.h
>> +++ b/include/linux/cpuhotplug.h
>> @@ -226,6 +226,7 @@ enum cpuhp_state {
>>       CPUHP_AP_PERF_ARM_HISI_PA_ONLINE,
>>       CPUHP_AP_PERF_ARM_HISI_SLLC_ONLINE,
>>       CPUHP_AP_PERF_ARM_HISI_PCIE_PMU_ONLINE,
>> +    CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
>>       CPUHP_AP_PERF_ARM_L2X0_ONLINE,
>>       CPUHP_AP_PERF_ARM_QCOM_L2_ONLINE,
>>       CPUHP_AP_PERF_ARM_QCOM_L3_ONLINE,
> 
> .

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ