[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <57738C93.8010900@huawei.com>
Date: Wed, 29 Jun 2016 16:53:39 +0800
From: oulijun <oulijun@...wei.com>
To: Leon Romanovsky <leon@...nel.org>
CC: Doug Ledford <dledford@...hat.com>, <sean.hefty@...el.com>,
<hal.rosenstock@...il.com>, <davem@...emloft.net>,
<jeffrey.t.kirsher@...el.com>, <jiri@...lanox.com>,
<ogerlitz@...lanox.com>, <linux-rdma@...r.kernel.org>,
<linux-kernel@...r.kernel.org>, <netdev@...r.kernel.org>,
<gongyangming@...wei.com>, <xiaokun@...wei.com>,
<tangchaofei@...wei.com>, <haifeng.wei@...wei.com>,
<yisen.zhuang@...wei.com>, <yankejian@...wei.com>,
<charles.chenxin@...wei.com>, <linuxarm@...wei.com>
Subject: Re: [PATCH v10 07/22] IB/hns: Add event queue support
Hi, Leon
在 2016/6/24 23:46, Leon Romanovsky 写道:
> On Thu, Jun 16, 2016 at 10:35:15PM +0800, Lijun Ou wrote:
>> This patch added event queue support for RoCE driver. It is used
>> for RoCE interrupt. RoCE includes 32 synchronous event irqs, 1
>> asynchronous event irq and 1 common overflow irq.
>>
>> Signed-off-by: Wei Hu <xavier.huwei@...wei.com>
>> Signed-off-by: Nenglong Zhao <zhaonenglong@...ilicon.com>
>> Signed-off-by: Lijun Ou <oulijun@...wei.com>
>> ---
>> PATCH v9/v8:
>> - No change over the PATCH v7
>>
>> PATCH v7:
>> This fixes the comments given by Doug Ledford over the PATCH v6:
>> Link: https://lkml.org/lkml/2016/5/13/510
>>
>> PATCH v6:
>> - No change over the PATCH v5
>>
>> PATCH v5:
>> - The initial patch which was redesigned based on the second patch
>> in PATCH v4
>> ---
>> ---
>> drivers/infiniband/hw/hns/hns_roce_cmd.c | 22 +
>> drivers/infiniband/hw/hns/hns_roce_common.h | 70 +++
>> drivers/infiniband/hw/hns/hns_roce_cq.c | 77 +++
>> drivers/infiniband/hw/hns/hns_roce_device.h | 135 +++++
>> drivers/infiniband/hw/hns/hns_roce_eq.c | 750 ++++++++++++++++++++++++++++
>> drivers/infiniband/hw/hns/hns_roce_eq.h | 130 +++++
>> drivers/infiniband/hw/hns/hns_roce_main.c | 24 +
>> drivers/infiniband/hw/hns/hns_roce_qp.c | 63 +++
>> 8 files changed, 1271 insertions(+)
>> create mode 100644 drivers/infiniband/hw/hns/hns_roce_cq.c
>> create mode 100644 drivers/infiniband/hw/hns/hns_roce_eq.c
>> create mode 100644 drivers/infiniband/hw/hns/hns_roce_eq.h
>> create mode 100644 drivers/infiniband/hw/hns/hns_roce_qp.c
>>
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_cmd.c b/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> index 64e84fe..67b3137 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> @@ -45,6 +45,28 @@
>>
>> #define CMD_MAX_NUM 32
>>
>> +static int hns_roce_status_to_errno(u8 orig_status)
>> +{
>> + if (orig_status == HNS_ROCE_CMD_SUCCESS)
>> + return 0;
>> + else
>> + return -EIO;
>> +}
>
> 1. Can orig_status be different from SUCCESS? You defined one enum only.
> 2. return (orig_status == HNS_ROCE_CMD_SUCCESS)?0:(-EIO);
>
Sure, I will modify it in patch v11.
>> +
>> +void hns_roce_cmd_event(struct hns_roce_dev *hr_dev, u16 token, u8 status,
>> + u64 out_param)
>> +{
>> + struct hns_roce_cmd_context
>> + *context = &hr_dev->cmd.context[token & hr_dev->cmd.token_mask];
>> +
>> + if (token != context->token)
>> + return;
>> +
>> + context->result = hns_roce_status_to_errno(status);
>> + context->out_param = out_param;
>> + complete(&context->done);
>> +}
>> +
>> int hns_roce_cmd_init(struct hns_roce_dev *hr_dev)
>> {
>> struct device *dev = &hr_dev->pdev->dev;
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_common.h b/drivers/infiniband/hw/hns/hns_roce_common.h
>> index 595cda9..4805852 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_common.h
>> +++ b/drivers/infiniband/hw/hns/hns_roce_common.h
>> @@ -33,7 +33,56 @@
>> #ifndef _HNS_ROCE_COMMON_H
>> #define _HNS_ROCE_COMMON_H
>>
>> +#define roce_write(dev, reg, val) writel((val), (dev)->reg_base + (reg))
>> #define roce_read(dev, reg) readl((dev)->reg_base + (reg))
>> +#define roce_raw_write(value, addr) \
>> + __raw_writel((__force u32)cpu_to_le32(value), (addr))
>> +
>> +#define roce_get_field(origin, mask, shift) \
>> + (((origin) & (mask)) >> (shift))
>> +
>> +#define roce_get_bit(origin, shift) \
>> + roce_get_field((origin), (1ul << (shift)), (shift))
>> +
>> +#define roce_set_field(origin, mask, shift, val) \
>> + do { \
>> + (origin) &= (~(mask)); \
>> + (origin) |= (((u32)(val) << (shift)) & (mask)); \
>> + } while (0)
>> +
>> +#define roce_set_bit(origin, shift, val) \
>> + roce_set_field((origin), (1ul << (shift)), (shift), (val))
>> +
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S 0
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M \
>> + (((1UL << 2) - 1) << ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S)
>> +
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S 8
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M \
>> + (((1UL << 4) - 1) << ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S)
>> +
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S 17
>> +
>> +#define ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S 0
>> +#define ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M \
>> + (((1UL << 5) - 1) << ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S)
>> +
>> +#define ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S 16
>> +#define ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M \
>> + (((1UL << 16) - 1) << ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S)
>> +
>> +#define ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S 0
>> +#define ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M \
>> + (((1UL << 16) - 1) << ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S)
>> +
>> +#define ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S 16
>> +#define ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S 1
>> +#define ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S 0
>> +
>> +#define ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S 0
>> +#define ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S 1
>> +
>> +#define ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S 0
>>
>> /*************ROCEE_REG DEFINITION****************/
>> #define ROCEE_VENDOR_ID_REG 0x0
>> @@ -44,8 +93,29 @@
>> #define ROCEE_SYS_IMAGE_GUID_L_REG 0xC
>> #define ROCEE_SYS_IMAGE_GUID_H_REG 0x10
>>
>> +#define ROCEE_CAEP_AEQE_CONS_IDX_REG 0x3AC
>> +#define ROCEE_CAEP_CEQC_CONS_IDX_0_REG 0x3BC
>> +
>> +#define ROCEE_ECC_UCERR_ALM1_REG 0xB38
>> +#define ROCEE_ECC_UCERR_ALM2_REG 0xB3C
>> +#define ROCEE_ECC_CERR_ALM1_REG 0xB44
>> +#define ROCEE_ECC_CERR_ALM2_REG 0xB48
>> +
>> #define ROCEE_ACK_DELAY_REG 0x14
>>
>> +#define ROCEE_CAEP_CE_INTERVAL_CFG_REG 0x190
>> +#define ROCEE_CAEP_CE_BURST_NUM_CFG_REG 0x194
>> +
>> #define ROCEE_MB1_REG 0x210
>>
>> +#define ROCEE_CAEP_AEQC_AEQE_SHIFT_REG 0x3A0
>> +#define ROCEE_CAEP_CEQC_SHIFT_0_REG 0x3B0
>> +#define ROCEE_CAEP_CE_IRQ_MASK_0_REG 0x3C0
>> +#define ROCEE_CAEP_CEQ_ALM_OVF_0_REG 0x3C4
>> +#define ROCEE_CAEP_AE_MASK_REG 0x6C8
>> +#define ROCEE_CAEP_AE_ST_REG 0x6CC
>> +
>> +#define ROCEE_ECC_UCERR_ALM0_REG 0xB34
>> +#define ROCEE_ECC_CERR_ALM0_REG 0xB40
>
> Indentation
>
Exactly as Doug said, no indentation in my editor and vi
>> +
>> #endif /* _HNS_ROCE_COMMON_H */
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_cq.c b/drivers/infiniband/hw/hns/hns_roce_cq.c
>> new file mode 100644
>> index 0000000..42a3c98
>> --- /dev/null
>> +++ b/drivers/infiniband/hw/hns/hns_roce_cq.c
>> @@ -0,0 +1,77 @@
>> +/*
>> + * Copyright (c) 2016 Hisilicon Limited.
>> + *
>> + * This software is available to you under a choice of one of two
>> + * licenses. You may choose to be licensed under the terms of the GNU
>> + * General Public License (GPL) Version 2, available from the file
>> + * COPYING in the main directory of this source tree, or the
>> + * OpenIB.org BSD license below:
>> + *
>> + * Redistribution and use in source and binary forms, with or
>> + * without modification, are permitted provided that the following
>> + * conditions are met:
>> + *
>> + * - Redistributions of source code must retain the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer.
>> + *
>> + * - Redistributions in binary form must reproduce the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer in the documentation and/or other materials
>> + * provided with the distribution.
>> + *
>> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> + * SOFTWARE.
>> + */
>> +
>> +#include <linux/hardirq.h>
>> +#include <linux/log2.h>
>> +#include <linux/slab.h>
>> +#include "hns_roce_device.h"
>> +
>> +void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn)
>> +{
>> + struct device *dev = &hr_dev->pdev->dev;
>> + struct hns_roce_cq *cq;
>> +
>> + cq = radix_tree_lookup(&hr_dev->cq_table.tree,
>> + cqn & (hr_dev->caps.num_cqs - 1));
>> + if (!cq) {
>> + dev_warn(dev, "Completion event for bogus CQ 0x%08x\n", cqn);
>> + return;
>> + }
>> +
>> + cq->comp(cq);
>> +}
>> +
>> +void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type)
>> +{
>> + struct hns_roce_cq_table *cq_table = &hr_dev->cq_table;
>> + struct device *dev = &hr_dev->pdev->dev;
>> + struct hns_roce_cq *cq;
>> +
>> + spin_lock(&cq_table->lock);
>
> What exactly do you protect here with lock?
> radix_tree_lookup? Why didn't you protect it in hns_roce_cq_completion
> function?
We have checked and analysed it. We think that here is not lock. Becasue maybe
lead to deadlock in event irq functions. We will remove it in patch v11
>
>> +
>> + cq = radix_tree_lookup(&cq_table->tree,
>> + cqn & (hr_dev->caps.num_cqs - 1));
>> + if (cq)
>> + atomic_inc(&cq->refcount);
>> +
>> + spin_unlock(&cq_table->lock);
>> +
>> + if (!cq) {
>> + dev_warn(dev, "Async event for bogus CQ %08x\n", cqn);
>> + return;
>> + }
>> +
>> + cq->event(cq, (enum hns_roce_event)event_type);
>> +
>> + if (atomic_dec_and_test(&cq->refcount))
>> + complete(&cq->free);
>> +}
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_device.h b/drivers/infiniband/hw/hns/hns_roce_device.h
>> index 23b7e17..57184ab 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_device.h
>> +++ b/drivers/infiniband/hw/hns/hns_roce_device.h
>> @@ -44,6 +44,8 @@
>>
>> #define DRV_NAME "hns_roce"
>>
>> +#define HNS_ROCE_BA_SIZE (32 * 4096)
>> +
>> #define HNS_ROCE_MAX_IRQ_NUM 34
>>
>> #define HNS_ROCE_COMP_VEC_NUM 32
>> @@ -53,8 +55,89 @@
>>
>> #define HNS_ROCE_MAX_PORTS 6
>>
>> +enum hns_roce_event {
>> + HNS_ROCE_EVENT_TYPE_PATH_MIG = 0x01,
>> + HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED = 0x02,
>> + HNS_ROCE_EVENT_TYPE_COMM_EST = 0x03,
>> + HNS_ROCE_EVENT_TYPE_SQ_DRAINED = 0x04,
>> + HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR = 0x05,
>> + HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR = 0x06,
>> + HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR = 0x07,
>> + HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH = 0x08,
>> + HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH = 0x09,
>> + HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR = 0x0a,
>> + HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR = 0x0b,
>> + HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW = 0x0c,
>> + HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID = 0x0d,
>> + HNS_ROCE_EVENT_TYPE_PORT_CHANGE = 0x0f,
>
> Please add comment here that 0x10 and 0x11 were skipped on purpose.
>
OK, I wll add it in patch v11
>> + HNS_ROCE_EVENT_TYPE_DB_OVERFLOW = 0x12,
>> + HNS_ROCE_EVENT_TYPE_MB = 0x13,
>> + HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW = 0x14,
>> +};
>> +
>> +/* Local Work Queue Catastrophic Error,SUBTYPE 0x5 */
>> +enum {
>> + HNS_ROCE_LWQCE_QPC_ERROR = 1,
>> + HNS_ROCE_LWQCE_MTU_ERROR = 2,
>> + HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR = 3,
>> + HNS_ROCE_LWQCE_WQE_ADDR_ERROR = 4,
>> + HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR = 5,
>> + HNS_ROCE_LWQCE_SL_ERROR = 6,
>> + HNS_ROCE_LWQCE_PORT_ERROR = 7,
>> +};
>> +
>> +/* Local Access Violation Work Queue Error,SUBTYPE 0x7 */
>> +enum {
>> + HNS_ROCE_LAVWQE_R_KEY_VIOLATION = 1,
>> + HNS_ROCE_LAVWQE_LENGTH_ERROR = 2,
>> + HNS_ROCE_LAVWQE_VA_ERROR = 3,
>> + HNS_ROCE_LAVWQE_PD_ERROR = 4,
>> + HNS_ROCE_LAVWQE_RW_ACC_ERROR = 5,
>> + HNS_ROCE_LAVWQE_KEY_STATE_ERROR = 6,
>> + HNS_ROCE_LAVWQE_MR_OPERATION_ERROR = 7,
>> +};
>> +
>> +/* DOORBELL overflow subtype */
>> +enum {
>> + HNS_ROCE_DB_SUBTYPE_SDB_OVF = 1,
>> + HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF = 2,
>> + HNS_ROCE_DB_SUBTYPE_ODB_OVF = 3,
>> + HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF = 4,
>> + HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP = 5,
>> + HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP = 6,
>> +};
>> +
>> +enum {
>> + HNS_ROCE_CMD_SUCCESS = 1,
>> +};
>> +
>> +struct hns_roce_buf_list {
>> + void *buf;
>> + dma_addr_t map;
>> +};
>> +
>> +struct hns_roce_cq {
>> + void (*comp)(struct hns_roce_cq *);
>> + void (*event)(struct hns_roce_cq *, enum hns_roce_event);
>> +
>> + atomic_t refcount;
>> + struct completion free;
>> +};
>> +
>> +struct hns_roce_qp_table {
>> + spinlock_t lock;
>> +};
>> +
>> +struct hns_roce_cq_table {
>> + spinlock_t lock;
>> + struct radix_tree_root tree;
>> +};
>> +
>> struct hns_roce_cmd_context {
>> + struct completion done;
>> + int result;
>> int next;
>> + u64 out_param;
>> u16 token;
>> };
>>
>> @@ -87,11 +170,42 @@ struct hns_roce_cmdq {
>> u8 toggle;
>> };
>>
>> +struct hns_roce_dev;
>> +
>> +struct hns_roce_qp {
>> + void (*event)(struct hns_roce_qp *,
>> + enum hns_roce_event);
>> +
>> + atomic_t refcount;
>> + struct completion free;
>> +};
>> +
>> struct hns_roce_ib_iboe {
>> struct net_device *netdevs[HNS_ROCE_MAX_PORTS];
>> u8 phy_port[HNS_ROCE_MAX_PORTS];
>> };
>>
>> +struct hns_roce_eq {
>> + struct hns_roce_dev *hr_dev;
>> + void __iomem *doorbell;
>> +
>> + int type_flag;/* Aeq:1 ceq:0 */
>> + int eqn;
>> + u32 entries;
>> + int log_entries;
>> + int eqe_size;
>> + int irq;
>> + u16 have_irq;
>> + int log_page_size;
>> + int cons_index;
>> + struct hns_roce_buf_list *buf_list;
>> +};
>> +
>> +struct hns_roce_eq_table {
>> + struct hns_roce_eq *eq;
>> + void __iomem **eqc_base;
>> +};
>> +
>> struct hns_roce_caps {
>> u64 fw_ver;
>> u8 num_ports;
>> @@ -150,6 +264,7 @@ struct hns_roce_dev {
>> int irq[HNS_ROCE_MAX_IRQ_NUM];
>> u8 __iomem *reg_base;
>> struct hns_roce_caps caps;
>> + struct radix_tree_root qp_table_tree;
>>
>> u64 fw_ver;
>> u64 sys_image_guid;
>> @@ -158,17 +273,37 @@ struct hns_roce_dev {
>> u32 hw_rev;
>>
>> struct hns_roce_cmdq cmd;
>> + struct hns_roce_cq_table cq_table;
>> + struct hns_roce_qp_table qp_table;
>> + struct hns_roce_eq_table eq_table;
>>
>> int cmd_mod;
>> int loop_idc;
>> struct hns_roce_hw *hw;
>> };
>>
>> +static inline struct hns_roce_qp
>> + *__hns_roce_qp_lookup(struct hns_roce_dev *hr_dev, u32 qpn)
>> +{
>> + return radix_tree_lookup(&hr_dev->qp_table_tree,
>> + qpn & (hr_dev->caps.num_qps - 1));
>> +}
>> +
>> int hns_roce_cmd_init(struct hns_roce_dev *hr_dev);
>> void hns_roce_cmd_cleanup(struct hns_roce_dev *hr_dev);
>> +void hns_roce_cmd_event(struct hns_roce_dev *hr_dev, u16 token, u8 status,
>> + u64 out_param);
>
> Please fix indentation.
>
I have checked it in my editor and vi in linux, it is not exit indentation question
>> int hns_roce_cmd_use_events(struct hns_roce_dev *hr_dev);
>> void hns_roce_cmd_use_polling(struct hns_roce_dev *hr_dev);
>>
>> +int hns_roce_init_eq_table(struct hns_roce_dev *hr_dev);
>> +
>> +void hns_roce_cleanup_eq_table(struct hns_roce_dev *hr_dev);
>> +
>> +void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn);
>> +void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type);
>> +void hns_roce_qp_event(struct hns_roce_dev *hr_dev, u32 qpn, int event_type);
>> +
>> extern struct hns_roce_hw hns_roce_hw_v1;
>>
>> #endif /* _HNS_ROCE_DEVICE_H */
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_eq.c b/drivers/infiniband/hw/hns/hns_roce_eq.c
>> new file mode 100644
>> index 0000000..6600a23
>> --- /dev/null
>> +++ b/drivers/infiniband/hw/hns/hns_roce_eq.c
>> @@ -0,0 +1,750 @@
>> +/*
>> + * Copyright (c) 2016 Hisilicon Limited.
>> + *
>> + * This software is available to you under a choice of one of two
>> + * licenses. You may choose to be licensed under the terms of the GNU
>> + * General Public License (GPL) Version 2, available from the file
>> + * COPYING in the main directory of this source tree, or the
>> + * OpenIB.org BSD license below:
>> + *
>> + * Redistribution and use in source and binary forms, with or
>> + * without modification, are permitted provided that the following
>> + * conditions are met:
>> + *
>> + * - Redistributions of source code must retain the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer.
>> + *
>> + * - Redistributions in binary form must reproduce the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer in the documentation and/or other materials
>> + * provided with the distribution.
>> + *
>> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> + * SOFTWARE.
>> + */
>> +
>> +#include <linux/dma-mapping.h>
>> +#include <linux/init.h>
>> +#include <linux/interrupt.h>
>> +#include <linux/slab.h>
>> +#include "hns_roce_common.h"
>> +#include "hns_roce_device.h"
>> +#include "hns_roce_eq.h"
>> +
>> +static void eq_set_cons_index(struct hns_roce_eq *eq, int req_not)
>> +{
>> + roce_raw_write((eq->cons_index & CONS_INDEX_MASK) |
>> + (req_not << eq->log_entries), eq->doorbell);
>> + /* Memory barrier */
>> + mb();
>> +}
>> +
>> +static struct hns_roce_aeqe *get_aeqe(struct hns_roce_eq *eq, u32 entry)
>> +{
>> + unsigned long off = (entry & (eq->entries - 1)) *
>> + HNS_ROCE_AEQ_ENTRY_SIZE;
>> +
>> + return (struct hns_roce_aeqe *)((u8 *)
>> + (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> + off % HNS_ROCE_BA_SIZE);
>> +}
>> +
>> +static struct hns_roce_aeqe *next_aeqe_sw(struct hns_roce_eq *eq)
>> +{
>> + struct hns_roce_aeqe *aeqe = get_aeqe(eq, eq->cons_index);
>> +
>> + return (roce_get_bit(aeqe->asyn, HNS_ROCE_AEQE_U32_4_OWNER_S) ^
>> + !!(eq->cons_index & eq->entries)) ? aeqe : NULL;
>> +}
>> +
>> +static int hns_roce_aeq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>
> Please refactor this function. It has switch in switch and it is almost 200 LOCs.
>
Yes, We will do this in patch v11
>> +{
>> + struct device *dev = &hr_dev->pdev->dev;
>> + struct hns_roce_aeqe *aeqe;
>> + int aeqes_found = 0;
>> + int qpn = 0;
>> +
>> + while ((aeqe = next_aeqe_sw(eq))) {
>> + dev_dbg(dev, "aeqe = %p, aeqe->asyn.event_type = 0x%lx\n", aeqe,
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + /* Memory barrier */
>> + rmb();
>> +
>> + switch (roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S)) {
>> + case HNS_ROCE_EVENT_TYPE_PATH_MIG:
>> + dev_warn(dev, "PATH MIG not supported\n");
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_COMM_EST:
>> + dev_warn(dev, "COMMUNICATION ESTABLISHED\n");
>
> CAPSLOCK????
>
We will modify it in patch v11
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_SQ_DRAINED:
>> + dev_warn(dev, "SQ DRAINED not supported\n");
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED:
>> + dev_warn(dev, "PATH MIG FAILED\n");
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
>> + dev_warn(dev, "qpn = 0x%lx\n",
>> + roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S));
>> + hns_roce_qp_event(hr_dev,
>> + roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
>> + qpn = roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
>> + dev_warn(dev, "Local Work Queue Catastrophic Error.\n");
>> + switch (roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> + case HNS_ROCE_LWQCE_QPC_ERROR:
>> + dev_warn(dev, "QP %d, QPC error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_MTU_ERROR:
>> + dev_warn(dev, "QP %d, MTU error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR:
>> + dev_warn(dev, "QP %d, WQE BA addr error.\n",
>> + qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_WQE_ADDR_ERROR:
>> + dev_warn(dev, "QP %d, WQE addr error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR:
>> + dev_warn(dev, "QP %d, WQE shift error\n", qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_SL_ERROR:
>> + dev_warn(dev, "QP %d, SL error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LWQCE_PORT_ERROR:
>> + dev_warn(dev, "QP %d, port error.\n", qpn);
>> + break;
>> + default:
>> + break;
>> + }
>> +
>> + hns_roce_qp_event(hr_dev,
>> + roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
>> + qpn = roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
>> + dev_warn(dev, "Local Access Violation Work Queue Error.\n");
>> + switch (roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> + case HNS_ROCE_LAVWQE_R_KEY_VIOLATION:
>> + dev_warn(dev, "QP %d, R_key violation.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_LENGTH_ERROR:
>> + dev_warn(dev, "QP %d, length error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_VA_ERROR:
>> + dev_warn(dev, "QP %d, VA error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_PD_ERROR:
>> + dev_err(dev, "QP %d, PD error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_RW_ACC_ERROR:
>> + dev_warn(dev, "QP %d, rw acc error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_KEY_STATE_ERROR:
>> + dev_warn(dev, "QP %d, key state error.\n", qpn);
>> + break;
>> + case HNS_ROCE_LAVWQE_MR_OPERATION_ERROR:
>> + dev_warn(dev, "QP %d, MR operation error.\n",
>> + qpn);
>> + break;
>> + default:
>> + break;
>> + }
>> +
>> + hns_roce_qp_event(hr_dev,
>> + roce_get_field(aeqe->event.qp_event.qp,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> + HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH:
>> + case HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR:
>> + case HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH:
>> + dev_warn(dev, "SRQ not support!\n");
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
>> + dev_warn(dev, "CQ 0x%lx access err.\n",
>> + roce_get_field(aeqe->event.cq_event.cq,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
>> + hns_roce_cq_event(hr_dev,
>> + le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
>> + dev_warn(dev, "CQ 0x%lx overflow\n",
>> + roce_get_field(aeqe->event.cq_event.cq,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
>> + hns_roce_cq_event(hr_dev,
>> + le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
>> + dev_warn(dev, "CQ ID invalid.\n");
>> + hns_roce_cq_event(hr_dev,
>> + le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> + HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)),
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_PORT_CHANGE:
>> + dev_warn(dev, "port change.\n");
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_MB:
>> + hns_roce_cmd_event(hr_dev,
>> + le16_to_cpu(aeqe->event.cmd.token),
>> + aeqe->event.cmd.status,
>> + le64_to_cpu(aeqe->event.cmd.out_param
>> + ));
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_DB_OVERFLOW:
>> + switch (roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> + case HNS_ROCE_DB_SUBTYPE_SDB_OVF:
>> + dev_warn(dev, "SDB overflow.\n");
>> + break;
>> + case HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF:
>> + dev_warn(dev, "SDB almost overflow.\n");
>> + break;
>> + case HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP:
>> + dev_warn(dev, "SDB almost empty.\n");
>> + break;
>> + case HNS_ROCE_DB_SUBTYPE_ODB_OVF:
>> + dev_warn(dev, "ODB overflow.\n");
>> + break;
>> + case HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF:
>> + dev_warn(dev, "ODB almost overflow.\n");
>> + break;
>> + case HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP:
>> + dev_warn(dev, "SDB almost empty.\n");
>> + break;
>> + default:
>> + break;
>> + }
>> +
>> + break;
>> + case HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW:
>> + dev_warn(dev, "CEQ 0x%lx OVERFLOW EVENT.\n",
>> + roce_get_field(aeqe->event.ce_event.ceqe,
>> + HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M,
>> + HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S));
>> + break;
>> + default:
>> + dev_warn(dev, "Unhandled event 0x%lx on EQ %d at index %u\n",
>> + roce_get_field(aeqe->asyn,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> + HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S),
>> + eq->eqn, eq->cons_index);
>> + break;
>> + };
>> +
>> + eq->cons_index++;
>> + aeqes_found = 1;
>> +
>> + if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
>> + dev_warn(dev, "cons_index overflow, set back to zero\n"
>> + );
>> + eq->cons_index = 0;
>> + }
>> + }
>> +
>> + eq_set_cons_index(eq, 0);
>> +
>> + return aeqes_found;
>> +}
>> +
>> +static struct hns_roce_ceqe *get_ceqe(struct hns_roce_eq *eq, u32 entry)
>> +{
>> + unsigned long off = (entry & (eq->entries - 1)) *
>> + HNS_ROCE_CEQ_ENTRY_SIZE;
>> +
>> + return (struct hns_roce_ceqe *)((u8 *)
>> + (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> + off % HNS_ROCE_BA_SIZE);
>> +}
>> +
>> +static struct hns_roce_ceqe *next_ceqe_sw(struct hns_roce_eq *eq)
>> +{
>> + struct hns_roce_ceqe *ceqe = get_ceqe(eq, eq->cons_index);
>> +
>> + return (!!(roce_get_bit(ceqe->ceqe.comp,
>> + HNS_ROCE_CEQE_CEQE_COMP_OWNER_S))) ^
>> + (!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
>> +}
>> +
>> +static int hns_roce_ceq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>> +{
>> + struct hns_roce_ceqe *ceqe;
>> + int ceqes_found = 0;
>> + u32 cqn;
>> +
>> + while ((ceqe = next_ceqe_sw(eq))) {
>> + /* Memory barrier */
>> + rmb();
>> + cqn = roce_get_field(ceqe->ceqe.comp,
>> + HNS_ROCE_CEQE_CEQE_COMP_CQN_M,
>> + HNS_ROCE_CEQE_CEQE_COMP_CQN_S);
>> + hns_roce_cq_completion(hr_dev, cqn);
>> +
>> + ++eq->cons_index;
>> + ceqes_found = 1;
>> +
>> + if (eq->cons_index > 2 * hr_dev->caps.ceqe_depth[eq->eqn] - 1) {
>> + dev_warn(&eq->hr_dev->pdev->dev,
>> + "cons_index overflow, set back to zero\n");
>> + eq->cons_index = 0;
>> + }
>> + }
>> +
>> + eq_set_cons_index(eq, 0);
>> +
>> + return ceqes_found;
>> +}
>> +
>> +static int hns_roce_aeq_ovf_int(struct hns_roce_dev *hr_dev,
>> + struct hns_roce_eq *eq)
>> +{
>> + struct device *dev = &eq->hr_dev->pdev->dev;
>> + int eqovf_found = 0;
>> + u32 caepaemask_val;
>> + u32 cealmovf_val;
>> + u32 caepaest_val;
>> + u32 aeshift_val;
>> + u32 ceshift_val;
>> + u32 cemask_val;
>> + int i = 0;
>> +
>> + /**
>> + * AEQ overflow ECC mult bit err CEQ overflow alarm
>> + * must clear interrupt, mask irq, clear irq, cancel mask operation
>> + */
>> + aeshift_val = roce_read(hr_dev, ROCEE_CAEP_AEQC_AEQE_SHIFT_REG);
>> +
>> + if (roce_get_bit(aeshift_val,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S) == 1) {
>> + dev_warn(dev, "AEQ overflow!\n");
>> +
>> + /* Set mask */
>> + caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> + roce_set_bit(caepaemask_val,
>> + ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> + HNS_ROCE_INT_MASK_ENABLE);
>> + roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> +
>> + /* Clear int state(INT_WC : write 1 clear) */
>> + caepaest_val = roce_read(hr_dev, ROCEE_CAEP_AE_ST_REG);
>> + roce_set_bit(caepaest_val,
>> + ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S, 1);
>> + roce_write(hr_dev, ROCEE_CAEP_AE_ST_REG, caepaest_val);
>> +
>> + /* Clear mask */
>> + caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> + roce_set_bit(caepaemask_val,
>> + ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> + HNS_ROCE_INT_MASK_DISABLE);
>> + roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> + }
>> +
>> + /* CEQ almost overflow */
>> + for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> + ceshift_val = roce_read(hr_dev, ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> + i * CEQ_REG_OFFSET);
>> +
>> + if (roce_get_bit(ceshift_val,
>> + ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S) == 1) {
>> + dev_warn(dev, "CEQ[%d] almost overflow!\n", i);
>> + eqovf_found++;
>> +
>> + /* Set mask */
>> + cemask_val = roce_read(hr_dev,
>> + ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> + i * CEQ_REG_OFFSET);
>> + roce_set_bit(cemask_val,
>> + ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> + HNS_ROCE_INT_MASK_ENABLE);
>> + roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> + i * CEQ_REG_OFFSET, cemask_val);
>> +
>> + /* Clear int state(INT_WC : write 1 clear) */
>> + cealmovf_val = roce_read(hr_dev,
>> + ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> + i * CEQ_REG_OFFSET);
>> + roce_set_bit(cealmovf_val,
>> + ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S,
>> + 1);
>> + roce_write(hr_dev, ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> + i * CEQ_REG_OFFSET, cealmovf_val);
>> +
>> + /* Clear mask */
>> + cemask_val = roce_read(hr_dev,
>> + ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> + i * CEQ_REG_OFFSET);
>> + roce_set_bit(cemask_val,
>> + ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> + HNS_ROCE_INT_MASK_DISABLE);
>> + roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> + i * CEQ_REG_OFFSET, cemask_val);
>> + }
>> + }
>> +
>> + /* ECC multi-bit error alarm */
>> + dev_warn(dev, "ECC UCERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> + roce_read(hr_dev, ROCEE_ECC_UCERR_ALM0_REG),
>> + roce_read(hr_dev, ROCEE_ECC_UCERR_ALM1_REG),
>> + roce_read(hr_dev, ROCEE_ECC_UCERR_ALM2_REG));
>> +
>> + dev_warn(dev, "ECC CERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> + roce_read(hr_dev, ROCEE_ECC_CERR_ALM0_REG),
>> + roce_read(hr_dev, ROCEE_ECC_CERR_ALM1_REG),
>> + roce_read(hr_dev, ROCEE_ECC_CERR_ALM2_REG));
>> +
>> + return eqovf_found;
>> +}
>> +
>> +static int hns_roce_eq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>> +{
>> + int eqes_found = 0;
>> +
>> + if (likely(eq->type_flag == HNS_ROCE_CEQ))
>> + /* CEQ irq routine, CEQ is pulse irq, not clear */
>> + eqes_found = hns_roce_ceq_int(hr_dev, eq);
>> + else if (likely(eq->type_flag == HNS_ROCE_AEQ))
>> + /* AEQ irq routine, AEQ is pulse irq, not clear */
>> + eqes_found = hns_roce_aeq_int(hr_dev, eq);
>> + else
>> + /* AEQ queue overflow irq */
>> + eqes_found = hns_roce_aeq_ovf_int(hr_dev, eq);
>> +
>> + return eqes_found;
>> +}
>> +
>> +static irqreturn_t hns_roce_msi_x_interrupt(int irq, void *eq_ptr)
>> +{
>> + int int_work = 0;
>> + struct hns_roce_eq *eq = eq_ptr;
>> + struct hns_roce_dev *hr_dev = eq->hr_dev;
>> +
>> + int_work = hns_roce_eq_int(hr_dev, eq);
>> +
>> + return IRQ_RETVAL(int_work);
>> +}
>> +
>> +static void hns_roce_enable_eq(struct hns_roce_dev *hr_dev, int eq_num,
>> + int enable_flag)
>> +{
>> + void __iomem *eqc = hr_dev->eq_table.eqc_base[eq_num];
>> + u32 val;
>> +
>> + val = readl(eqc);
>> +
>> + if (enable_flag)
>> + roce_set_field(val,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> + HNS_ROCE_EQ_STAT_VALID);
>> + else
>> + roce_set_field(val,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> + HNS_ROCE_EQ_STAT_INVALID);
>> + writel(val, eqc);
>> +}
>> +
>> +static int hns_roce_create_eq(struct hns_roce_dev *hr_dev,
>> + struct hns_roce_eq *eq)
>> +{
>> + void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
>> + struct device *dev = &hr_dev->pdev->dev;
>> + dma_addr_t tmp_dma_addr;
>> + u32 eqconsindx_val = 0;
>> + u32 eqcuridx_val = 0;
>> + u32 eqshift_val = 0;
>> + int num_bas = 0;
>> + int ret;
>> + int i;
>> +
>> + num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
>> + HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> +
>> + if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
>> + dev_err(dev, "[error]eq buf %d gt ba size(%d) need bas=%d\n",
>> + (eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
>> + num_bas);
>> + return -EINVAL;
>> + }
>> +
>> + eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
>> + if (!eq->buf_list)
>> + return -ENOMEM;
>> +
>> + for (i = 0; i < num_bas; ++i) {
>> + eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
>> + &tmp_dma_addr,
>> + GFP_KERNEL);
>> + if (!eq->buf_list[i].buf) {
>> + ret = -ENOMEM;
>> + goto err_out_free_pages;
>> + }
>> +
>> + eq->buf_list[i].map = tmp_dma_addr;
>> + memset(eq->buf_list[i].buf, 0, HNS_ROCE_BA_SIZE);
>> + }
>> + eq->cons_index = 0;
>> + roce_set_field(eqshift_val,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> + HNS_ROCE_EQ_STAT_INVALID);
>> + roce_set_field(eqshift_val,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M,
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S,
>> + eq->log_entries);
>> + writel(eqshift_val, eqc);
>> +
>> + /* Configure eq extended address 12~44bit */
>> + writel((u32)(eq->buf_list[0].map >> 12), (u8 *)eqc + 4);
>> +
>> + /*
>> + * Configure eq extended address 45~49 bit.
>> + * 44 = 32 + 12, When evaluating addr to hardware, shift 12 because of
>> + * using 4K page, and shift more 32 because of
>> + * caculating the high 32 bit value evaluated to hardware.
>> + */
>> + roce_set_field(eqcuridx_val, ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M,
>> + ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S,
>> + eq->buf_list[0].map >> 44);
>> + roce_set_field(eqcuridx_val,
>> + ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M,
>> + ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S, 0);
>> + writel(eqcuridx_val, (u8 *)eqc + 8);
>> +
>> + /* Configure eq consumer index */
>> + roce_set_field(eqconsindx_val,
>> + ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M,
>> + ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S, 0);
>> + writel(eqconsindx_val, (u8 *)eqc + 0xc);
>> +
>> + return 0;
>> +
>> +err_out_free_pages:
>> + for (i = 0; i < num_bas; ++i)
>> + if (eq->buf_list[i].buf)
>
> The two lines above can be replaced to something like this:
>
> for ( i = i-1 ; i >= 0: i--) {
>
Good, We think your advice is good for simplifing the lines and improving
perferance after analysized. We also modify the other place which exit the
similar questions
>> + dma_free_coherent(dev, HNS_ROCE_BA_SIZE,
>> + eq->buf_list[i].buf,
>> + eq->buf_list[i].map);
>> + kfree(eq->buf_list);
>> + return ret;
>> +}
>> +
>> +static void hns_roce_free_eq(struct hns_roce_dev *hr_dev,
>> + struct hns_roce_eq *eq)
>> +{
>> + int i = 0;
>> + int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
>> + HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> +
>> + if (!eq->buf_list)
>> + return;
>> +
>> + for (i = 0; i < npages; ++i)
>> + if (eq->buf_list[i].buf)
>
> Is it possible situation to have eq->buf_list[i].buf == NULL at the
> middle of iteration?
>
We have analysized it according to your reivews. We think that the eq->buf_list[i].buf
is not NULL at the middle of iteration because only the eq->buf_list[i].buf allocated
when the hns_roce_free_eq be called. As result, the if (eq->buf_list[i].buf) condition
should be deleted, right?
>> + dma_free_coherent(&hr_dev->pdev->dev, HNS_ROCE_BA_SIZE,
>> + eq->buf_list[i].buf,
>> + eq->buf_list[i].map);
>> + kfree(eq->buf_list);
>> +}
>> +
>> +static void hns_roce_int_mask_en(struct hns_roce_dev *hr_dev)
>> +{
>> + int i = 0;
>> + u32 aemask_val;
>> + int masken = 0;
>> +
>> + /* AEQ INT */
>> + aemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> + roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> + masken);
>> + roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S, masken);
>> + roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, aemask_val);
>> +
>> + /* CEQ INT */
>> + for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> + /* IRQ mask */
>> + roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> + i * CEQ_REG_OFFSET, masken);
>> + }
>> +}
>> +
>> +static void hns_roce_ce_int_default_cfg(struct hns_roce_dev *hr_dev)
>> +{
>> + /* Configure ce int interval */
>> + roce_write(hr_dev, ROCEE_CAEP_CE_INTERVAL_CFG_REG,
>> + HNS_ROCE_CEQ_DEFAULT_INTERVAL);
>> +
>> + /* Configure ce int burst num */
>> + roce_write(hr_dev, ROCEE_CAEP_CE_BURST_NUM_CFG_REG,
>> + HNS_ROCE_CEQ_DEFAULT_BURST_NUM);
>> +}
>> +
>> +int hns_roce_init_eq_table(struct hns_roce_dev *hr_dev)
>> +{
>> + struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> + struct device *dev = &hr_dev->pdev->dev;
>> + struct hns_roce_eq *eq = NULL;
>> + int eq_num = 0;
>> + int ret = 0;
>> + int i = 0;
>> +
>> + eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> + eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
>> + if (!eq_table->eq)
>> + return -ENOMEM;
>> +
>> + eq_table->eqc_base = kcalloc(eq_num, sizeof(*eq_table->eqc_base),
>> + GFP_KERNEL);
>> + if (!eq_table->eqc_base) {
>> + ret = -ENOMEM;
>> + goto err_eqc_base_alloc_fail;
>> + }
>> +
>> + for (i = 0; i < eq_num; i++) {
>> + eq = &eq_table->eq[i];
>> + eq->hr_dev = hr_dev;
>> + eq->eqn = i;
>> + eq->irq = hr_dev->irq[i];
>> + eq->log_page_size = PAGE_SHIFT;
>> +
>> + if (i < hr_dev->caps.num_comp_vectors) {
>> + /* CEQ */
>> + eq_table->eqc_base[i] = hr_dev->reg_base +
>> + ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> + HNS_ROCE_CEQC_REG_OFFSET * i;
>> + eq->type_flag = HNS_ROCE_CEQ;
>> + eq->doorbell = hr_dev->reg_base +
>> + ROCEE_CAEP_CEQC_CONS_IDX_0_REG +
>> + HNS_ROCE_CEQC_REG_OFFSET * i;
>> + eq->entries = hr_dev->caps.ceqe_depth[i];
>> + eq->log_entries = ilog2(eq->entries);
>> + eq->eqe_size = sizeof(struct hns_roce_ceqe);
>> + } else {
>> + /* AEQ */
>> + eq_table->eqc_base[i] = hr_dev->reg_base +
>> + ROCEE_CAEP_AEQC_AEQE_SHIFT_REG;
>> + eq->type_flag = HNS_ROCE_AEQ;
>> + eq->doorbell = hr_dev->reg_base +
>> + ROCEE_CAEP_AEQE_CONS_IDX_REG;
>> + eq->entries = hr_dev->caps.aeqe_depth;
>> + eq->log_entries = ilog2(eq->entries);
>> + eq->eqe_size = sizeof(struct hns_roce_aeqe);
>> + }
>> + }
>> +
>> + /* Disable irq */
>> + hns_roce_int_mask_en(hr_dev);
>> +
>> + /* Configure CE irq interval and burst num */
>> + hns_roce_ce_int_default_cfg(hr_dev);
>> +
>> + for (i = 0; i < eq_num; i++) {
>> + ret = hns_roce_create_eq(hr_dev, &eq_table->eq[i]);
>> + if (ret) {
>> + dev_err(dev, "eq create failed\n");
>> + goto err_create_eq_fail;
>> + }
>> +
>> + ret = request_irq(eq_table->eq[i].irq, hns_roce_msi_x_interrupt,
>> + 0, hr_dev->irq_names, eq_table->eq + i);
>> + if (ret) {
>> + dev_err(dev, "request irq error!\n");
>> + goto err_create_eq_fail;
>> + }
>> +
>> + eq_table->eq[i].have_irq = 1;
>> +
>> + hns_roce_enable_eq(hr_dev, i, EQ_ENABLE);
>> + }
>> +
>> + return 0;
>> +
>> +err_create_eq_fail:
>> + for (i = 0; i < eq_num; i++) {
>> + /* Disable EQ */
>> + hns_roce_enable_eq(hr_dev, i, EQ_DISABLE);
>> +
>> + if (eq_table->eq[i].have_irq)
>> + free_irq(eq_table->eq[i].irq, eq_table->eq + i);
>> +
>> + hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
>> + }
>> + kfree(eq_table->eqc_base);
>> +
>> +err_eqc_base_alloc_fail:
>> + kfree(eq_table->eq);
>> + return ret;
>> +}
>> +
>> +void hns_roce_cleanup_eq_table(struct hns_roce_dev *hr_dev)
>> +{
>> + int i;
>> + int eq_num;
>> + struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> +
>> + eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> + for (i = 0; i < eq_num; i++) {
>> + /* Disable EQ */
>> + hns_roce_enable_eq(hr_dev, i, EQ_DISABLE);
>> +
>> + if (eq_table->eq[i].have_irq)
>> + free_irq(eq_table->eq[i].irq, eq_table->eq + i);
>> +
>> + hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
>> + }
>> +
>> + kfree(eq_table->eqc_base);
>> + kfree(eq_table->eq);
>> +}
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_eq.h b/drivers/infiniband/hw/hns/hns_roce_eq.h
>> new file mode 100644
>> index 0000000..99906e3
>> --- /dev/null
>> +++ b/drivers/infiniband/hw/hns/hns_roce_eq.h
>> @@ -0,0 +1,130 @@
>> +/*
>> + * Copyright (c) 2016 Hisilicon Limited.
>> + *
>> + * This software is available to you under a choice of one of two
>> + * licenses. You may choose to be licensed under the terms of the GNU
>> + * General Public License (GPL) Version 2, available from the file
>> + * COPYING in the main directory of this source tree, or the
>> + * OpenIB.org BSD license below:
>> + *
>> + * Redistribution and use in source and binary forms, with or
>> + * without modification, are permitted provided that the following
>> + * conditions are met:
>> + *
>> + * - Redistributions of source code must retain the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer.
>> + *
>> + * - Redistributions in binary form must reproduce the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer in the documentation and/or other materials
>> + * provided with the distribution.
>> + *
>> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> + * SOFTWARE.
>> + */
>> +
>> +#ifndef _HNS_ROCE_EQ_H
>> +#define _HNS_ROCE_EQ_H
>> +
>> +#define HNS_ROCE_CEQ 1
>> +#define HNS_ROCE_AEQ 2
>> +
>> +#define HNS_ROCE_CEQ_ENTRY_SIZE 0x4
>> +#define HNS_ROCE_AEQ_ENTRY_SIZE 0x10
>> +#define HNS_ROCE_CEQC_REG_OFFSET 0x18
>
> Indentation
>
I have checked it, no indentation question
>> +
>> +#define HNS_ROCE_CEQ_DEFAULT_INTERVAL 0x10
>> +#define HNS_ROCE_CEQ_DEFAULT_BURST_NUM 0x10
>> +
>> +#define HNS_ROCE_INT_MASK_DISABLE 0
>> +#define HNS_ROCE_INT_MASK_ENABLE 1
>
> Indentation
>
>> +
>> +#define EQ_ENABLE 1
>> +#define EQ_DISABLE 0
>> +#define CONS_INDEX_MASK 0xffff
>> +
>> +#define CEQ_REG_OFFSET 0x18
>> +
>> +enum {
>> + HNS_ROCE_EQ_STAT_INVALID = 0,
>> + HNS_ROCE_EQ_STAT_VALID = 2,
>> +};
>> +
>> +struct hns_roce_aeqe {
>> + u32 asyn;
>> + union {
>> + struct {
>> + u32 qp;
>> + u32 rsv0;
>> + u32 rsv1;
>> + } qp_event;
>> +
>> + struct {
>> + u32 cq;
>> + u32 rsv0;
>> + u32 rsv1;
>> + } cq_event;
>> +
>> + struct {
>> + u32 port;
>> + u32 rsv0;
>> + u32 rsv1;
>> + } port_event;
>> +
>> + struct {
>> + u32 ceqe;
>> + u32 rsv0;
>> + u32 rsv1;
>> + } ce_event;
>> +
>> + struct {
>> + __le64 out_param;
>> + __le16 token;
>> + u8 status;
>> + u8 rsv0;
>> + } __packed cmd;
>> + } event;
>> +};
>> +
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S 16
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M \
>> + (((1UL << 8) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S)
>> +
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S 24
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M \
>> + (((1UL << 7) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)
>> +
>> +#define HNS_ROCE_AEQE_U32_4_OWNER_S 31
>> +
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M \
>> + (((1UL << 24) - 1) << HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S)
>> +
>> +#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M \
>> + (((1UL << 16) - 1) << HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)
>> +
>> +#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M \
>> + (((1UL << 5) - 1) << HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S)
>> +
>> +struct hns_roce_ceqe {
>> + union {
>> + int comp;
>> + } ceqe;
>> +};
>> +
>> +#define HNS_ROCE_CEQE_CEQE_COMP_OWNER_S 0
>> +
>> +#define HNS_ROCE_CEQE_CEQE_COMP_CQN_S 16
>> +#define HNS_ROCE_CEQE_CEQE_COMP_CQN_M \
>> + (((1UL << 16) - 1) << HNS_ROCE_CEQE_CEQE_COMP_CQN_S)
>> +
>> +#endif /* _HNS_ROCE_EQ_H */
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_main.c b/drivers/infiniband/hw/hns/hns_roce_main.c
>> index c9e6d37..0b9cee7 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_main.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_main.c
>> @@ -183,6 +183,26 @@ static int hns_roce_probe(struct platform_device *pdev)
>> goto error_failed_cmd_init;
>> }
>>
>> + ret = hns_roce_init_eq_table(hr_dev);
>> + if (ret) {
>> + dev_err(dev, "eq init failed!\n");
>> + goto error_failed_eq_table;
>> + }
>> +
>> + if (hr_dev->cmd_mod) {
>> + ret = hns_roce_cmd_use_events(hr_dev);
>> + if (ret) {
>> + dev_err(dev, "Switch to event-driven cmd failed!\n");
>> + goto error_failed_use_event;
>> + }
>> + }
>> +
>> +error_failed_use_event:
>> + hns_roce_cleanup_eq_table(hr_dev);
>> +
>> +error_failed_eq_table:
>> + hns_roce_cmd_cleanup(hr_dev);
>> +
>> error_failed_cmd_init:
>> ret = hns_roce_engine_reset(hr_dev, false);
>> if (ret)
>> @@ -202,6 +222,10 @@ static int hns_roce_remove(struct platform_device *pdev)
>> {
>> struct hns_roce_dev *hr_dev = platform_get_drvdata(pdev);
>>
>> + if (hr_dev->cmd_mod)
>> + hns_roce_cmd_use_polling(hr_dev);
>> +
>> + hns_roce_cleanup_eq_table(hr_dev);
>> hns_roce_cmd_cleanup(hr_dev);
>> (void)hns_roce_engine_reset(hr_dev, false);
>>
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_qp.c b/drivers/infiniband/hw/hns/hns_roce_qp.c
>> new file mode 100644
>> index 0000000..a826c11
>> --- /dev/null
>> +++ b/drivers/infiniband/hw/hns/hns_roce_qp.c
>> @@ -0,0 +1,63 @@
>> +/*
>> + * Copyright (c) 2016 Hisilicon Limited.
>> + * Copyright (c) 2007, 2008 Mellanox Technologies. All rights reserved.
>> + *
>> + * This software is available to you under a choice of one of two
>> + * licenses. You may choose to be licensed under the terms of the GNU
>> + * General Public License (GPL) Version 2, available from the file
>> + * COPYING in the main directory of this source tree, or the
>> + * OpenIB.org BSD license below:
>> + *
>> + * Redistribution and use in source and binary forms, with or
>> + * without modification, are permitted provided that the following
>> + * conditions are met:
>> + *
>> + * - Redistributions of source code must retain the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer.
>> + *
>> + * - Redistributions in binary form must reproduce the above
>> + * copyright notice, this list of conditions and the following
>> + * disclaimer in the documentation and/or other materials
>> + * provided with the distribution.
>> + *
>> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> + * SOFTWARE.
>> + */
>> +
>> +#include <linux/log2.h>
>> +#include <linux/slab.h>
>> +#include <rdma/ib_cache.h>
>> +#include <rdma/ib_pack.h>
>> +#include "hns_roce_device.h"
>> +
>> +void hns_roce_qp_event(struct hns_roce_dev *hr_dev, u32 qpn, int event_type)
>> +{
>> + struct hns_roce_qp_table *qp_table = &hr_dev->qp_table;
>> + struct device *dev = &hr_dev->pdev->dev;
>> + struct hns_roce_qp *qp;
>> +
>> + spin_lock(&qp_table->lock);
>> +
>> + qp = __hns_roce_qp_lookup(hr_dev, qpn);
>> + if (qp)
>> + atomic_inc(&qp->refcount);
>> +
>> + spin_unlock(&qp_table->lock);
>> +
>> + if (!qp) {
>> + dev_warn(dev, "Async event for bogus QP %08x\n", qpn);
>> + return;
>> + }
>> +
>> + qp->event(qp, (enum hns_roce_event)event_type);
>> +
>> + if (atomic_dec_and_test(&qp->refcount))
>> + complete(&qp->free);
>> +}
>> --
>> 1.9.1
>>
Powered by blists - more mailing lists