[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20210531225222.16992-25-smalin@marvell.com>
Date: Tue, 1 Jun 2021 01:52:19 +0300
From: Shai Malin <smalin@...vell.com>
To: <netdev@...r.kernel.org>, <linux-nvme@...ts.infradead.org>,
<davem@...emloft.net>, <kuba@...nel.org>, <sagi@...mberg.me>,
<hch@....de>, <axboe@...com>, <kbusch@...nel.org>
CC: <aelior@...vell.com>, <mkalderon@...vell.com>,
<okulkarni@...vell.com>, <pkushwaha@...vell.com>,
<prabhakar.pkin@...il.com>, <malin1024@...il.com>,
<smalin@...vell.com>
Subject: [RFC PATCH v7 24/27] qedn: Add support of NVME ICReq & ICResp
From: Prabhakar Kushwaha <pkushwaha@...vell.com>
Once a TCP connection established, the host sends an Initialize
Connection Request (ICReq) PDU to the controller.
Further Initialize Connection Response (ICResp) PDU received from
controller is processed by host to establish a connection and
exchange connection configuration parameters.
This patch present support of generation of ICReq and processing of
ICResp. It also update host configuration based on exchanged parameters.
Acked-by: Igor Russkikh <irusskikh@...vell.com>
Signed-off-by: Prabhakar Kushwaha <pkushwaha@...vell.com>
Signed-off-by: Omkar Kulkarni <okulkarni@...vell.com>
Signed-off-by: Michal Kalderon <mkalderon@...vell.com>
Signed-off-by: Ariel Elior <aelior@...vell.com>
Signed-off-by: Shai Malin <smalin@...vell.com>
Reviewed-by: Hannes Reinecke <hare@...e.de>
---
drivers/nvme/hw/qedn/qedn.h | 37 ++++
drivers/nvme/hw/qedn/qedn_conn.c | 325 ++++++++++++++++++++++++++++++-
drivers/nvme/hw/qedn/qedn_main.c | 14 ++
drivers/nvme/hw/qedn/qedn_task.c | 8 +-
4 files changed, 380 insertions(+), 4 deletions(-)
diff --git a/drivers/nvme/hw/qedn/qedn.h b/drivers/nvme/hw/qedn/qedn.h
index 5ed6dc108506..9584021c4db1 100644
--- a/drivers/nvme/hw/qedn/qedn.h
+++ b/drivers/nvme/hw/qedn/qedn.h
@@ -16,6 +16,7 @@
/* Driver includes */
#include "../../host/tcp-offload.h"
+#include <linux/nvme-tcp.h>
#define QEDN_MODULE_NAME "qedn"
@@ -42,6 +43,8 @@
/* Protocol defines */
#define QEDN_MAX_IO_SIZE QED_NVMETCP_MAX_IO_SIZE
+#define QEDN_MAX_PDU_SIZE 0x80000 /* 512KB */
+#define QEDN_MAX_OUTSTANDING_R2T_PDUS 0 /* 0 Based == 1 max R2T */
#define QEDN_FW_SLOW_IO_MIN_SGE_LIMIT (9700 / 6)
@@ -50,6 +53,13 @@
#define QEDN_INVALID_ITID 0xFFFF
+#define QEDN_ICREQ_FW_PAYLOAD (sizeof(struct nvme_tcp_icreq_pdu) -\
+ QED_NVMETCP_NON_IO_HDR_SIZE)
+#define QEDN_ICREQ_FW_PAYLOAD_START 8
+
+/* The FW will handle the ICReq as CCCID 0 (FW internal design) */
+#define QEDN_ICREQ_CCCID 0
+
/*
* TCP offload stack default configurations and defines.
* Future enhancements will allow controlling the configurable
@@ -120,6 +130,16 @@ struct qedn_fp_queue {
char irqname[QEDN_IRQ_NAME_LEN];
};
+struct qedn_negotiation_params {
+ u32 maxh2cdata; /* Negotiation */
+ u32 maxr2t; /* Validation */
+ u16 pfv; /* Validation */
+ bool hdr_digest; /* Negotiation */
+ bool data_digest; /* Negotiation */
+ u8 cpda; /* Negotiation */
+ u8 hpda; /* Validation */
+};
+
struct qedn_ctx {
struct pci_dev *pdev;
struct qed_dev *cdev;
@@ -176,6 +196,9 @@ struct qedn_endpoint {
struct nvmetcp_db_data db_data;
void __iomem *p_doorbell;
+ /* Spinlock for accessing FW queue */
+ spinlock_t doorbell_lock;
+
/* TCP Params */
__be32 dst_addr[4]; /* In network order */
__be32 src_addr[4]; /* In network order */
@@ -252,6 +275,12 @@ struct qedn_ctrl {
atomic_t host_num_active_conns;
};
+struct qedn_icreq_padding {
+ u32 *buffer;
+ dma_addr_t pa;
+ struct nvmetcp_sge sge;
+};
+
/* Connection level struct */
struct qedn_conn_ctx {
/* IO path */
@@ -300,6 +329,11 @@ struct qedn_conn_ctx {
size_t sq_depth;
+ struct qedn_negotiation_params required_params;
+ struct qedn_negotiation_params pdu_params;
+ struct nvme_tcp_icresp_pdu icresp;
+ struct qedn_icreq_padding *icreq_pad;
+
/* "dummy" socket */
struct socket *sock;
};
@@ -308,6 +342,7 @@ enum qedn_conn_resources_state {
QEDN_CONN_RESRC_FW_SQ,
QEDN_CONN_RESRC_ACQUIRE_CONN,
QEDN_CONN_RESRC_TASKS,
+ QEDN_CONN_RESRC_ICREQ_PAD,
QEDN_CONN_RESRC_CCCID_ITID_MAP,
QEDN_CONN_RESRC_TCP_PORT,
QEDN_CONN_RESRC_DB_ADD,
@@ -337,5 +372,7 @@ struct qedn_task_ctx *
qedn_get_free_task_from_pool(struct qedn_conn_ctx *conn_ctx, u16 cccid);
void qedn_destroy_free_tasks(struct qedn_fp_queue *fp_q,
struct qedn_io_resources *io_resrc);
+void qedn_prep_icresp(struct qedn_conn_ctx *conn_ctx, struct nvmetcp_fw_cqe *cqe);
+void qedn_ring_doorbell(struct qedn_conn_ctx *conn_ctx);
#endif /* _QEDN_H_ */
diff --git a/drivers/nvme/hw/qedn/qedn_conn.c b/drivers/nvme/hw/qedn/qedn_conn.c
index 02f36ef5d8e0..9d5a7377ff1a 100644
--- a/drivers/nvme/hw/qedn/qedn_conn.c
+++ b/drivers/nvme/hw/qedn/qedn_conn.c
@@ -34,6 +34,18 @@ inline int qedn_qid(struct nvme_tcp_ofld_queue *queue)
return queue - queue->ctrl->queues;
}
+void qedn_ring_doorbell(struct qedn_conn_ctx *conn_ctx)
+{
+ u16 prod_idx;
+
+ prod_idx = qed_chain_get_prod_idx(&conn_ctx->ep.fw_sq_chain);
+ conn_ctx->ep.db_data.sq_prod = cpu_to_le16(prod_idx);
+
+ /* wmb - Make sure fw idx is coherent */
+ wmb();
+ writel(*(u32 *)&conn_ctx->ep.db_data, conn_ctx->ep.p_doorbell);
+}
+
int qedn_set_con_state(struct qedn_conn_ctx *conn_ctx, enum qedn_conn_state new_state)
{
spin_lock_bh(&conn_ctx->conn_state_lock);
@@ -136,6 +148,71 @@ int qedn_initialize_endpoint(struct qedn_endpoint *ep, u8 *local_mac_addr,
return -1;
}
+static int qedn_alloc_icreq_pad(struct qedn_conn_ctx *conn_ctx)
+{
+ struct qedn_ctx *qedn = conn_ctx->qedn;
+ struct qedn_icreq_padding *icreq_pad;
+ u32 *buffer;
+ int rc = 0;
+
+ icreq_pad = kzalloc(sizeof(*icreq_pad), GFP_KERNEL);
+ if (!icreq_pad)
+ return -ENOMEM;
+
+ conn_ctx->icreq_pad = icreq_pad;
+ memset(&icreq_pad->sge, 0, sizeof(icreq_pad->sge));
+ buffer = dma_alloc_coherent(&qedn->pdev->dev,
+ QEDN_ICREQ_FW_PAYLOAD,
+ &icreq_pad->pa,
+ GFP_KERNEL);
+ if (!buffer) {
+ pr_err("Could not allocate icreq_padding SGE buffer.\n");
+ rc = -ENOMEM;
+ goto release_icreq_pad;
+ }
+
+ DMA_REGPAIR_LE(icreq_pad->sge.sge_addr, icreq_pad->pa);
+ icreq_pad->sge.sge_len = cpu_to_le32(QEDN_ICREQ_FW_PAYLOAD);
+ icreq_pad->buffer = buffer;
+ set_bit(QEDN_CONN_RESRC_ICREQ_PAD, &conn_ctx->resrc_state);
+
+ return 0;
+
+release_icreq_pad:
+ kfree(icreq_pad);
+ conn_ctx->icreq_pad = NULL;
+
+ return rc;
+}
+
+static void qedn_free_icreq_pad(struct qedn_conn_ctx *conn_ctx)
+{
+ struct qedn_ctx *qedn = conn_ctx->qedn;
+ struct qedn_icreq_padding *icreq_pad;
+ u32 *buffer;
+
+ icreq_pad = conn_ctx->icreq_pad;
+ if (unlikely(!icreq_pad)) {
+ pr_err("null ptr in icreq_pad in conn_ctx\n");
+ goto finally;
+ }
+
+ buffer = icreq_pad->buffer;
+ if (buffer) {
+ dma_free_coherent(&qedn->pdev->dev,
+ QEDN_ICREQ_FW_PAYLOAD,
+ (void *)buffer,
+ icreq_pad->pa);
+ icreq_pad->buffer = NULL;
+ }
+
+ kfree(icreq_pad);
+ conn_ctx->icreq_pad = NULL;
+
+finally:
+ clear_bit(QEDN_CONN_RESRC_ICREQ_PAD, &conn_ctx->resrc_state);
+}
+
static void qedn_release_conn_ctx(struct qedn_conn_ctx *conn_ctx)
{
struct qedn_ctx *qedn = conn_ctx->qedn;
@@ -168,6 +245,9 @@ static void qedn_release_conn_ctx(struct qedn_conn_ctx *conn_ctx)
clear_bit(QEDN_CONN_RESRC_ACQUIRE_CONN, &conn_ctx->resrc_state);
}
+ if (test_bit(QEDN_CONN_RESRC_ICREQ_PAD, &conn_ctx->resrc_state))
+ qedn_free_icreq_pad(conn_ctx);
+
if (test_bit(QEDN_CONN_RESRC_TASKS, &conn_ctx->resrc_state)) {
clear_bit(QEDN_CONN_RESRC_TASKS, &conn_ctx->resrc_state);
qedn_return_active_tasks(conn_ctx);
@@ -329,6 +409,208 @@ void qedn_terminate_connection(struct qedn_conn_ctx *conn_ctx)
queue_work(qctrl->sp_wq, &conn_ctx->sp_wq_entry);
}
+static int qedn_nvmetcp_update_conn(struct qedn_ctx *qedn, struct qedn_conn_ctx *conn_ctx)
+{
+ struct qedn_negotiation_params *pdu_params = &conn_ctx->pdu_params;
+ struct qed_nvmetcp_params_update *conn_info;
+ int rc;
+
+ conn_info = kzalloc(sizeof(*conn_info), GFP_KERNEL);
+ if (!conn_info)
+ return -ENOMEM;
+
+ conn_info->hdr_digest_en = pdu_params->hdr_digest;
+ conn_info->data_digest_en = pdu_params->data_digest;
+ conn_info->max_recv_pdu_length = QEDN_MAX_PDU_SIZE;
+ conn_info->max_io_size = QEDN_MAX_IO_SIZE;
+ conn_info->max_send_pdu_length = pdu_params->maxh2cdata;
+
+ rc = qed_ops->update_conn(qedn->cdev, conn_ctx->conn_handle, conn_info);
+ if (rc) {
+ pr_err("Could not update connection\n");
+ rc = -ENXIO;
+ }
+
+ kfree(conn_info);
+
+ return rc;
+}
+
+static int qedn_update_ramrod(struct qedn_conn_ctx *conn_ctx)
+{
+ struct qedn_ctx *qedn = conn_ctx->qedn;
+ int rc = 0;
+
+ rc = qedn_set_con_state(conn_ctx, CONN_STATE_WAIT_FOR_UPDATE_EQE);
+ if (rc)
+ return rc;
+
+ rc = qedn_nvmetcp_update_conn(qedn, conn_ctx);
+ if (rc)
+ return rc;
+
+ if (conn_ctx->state != CONN_STATE_WAIT_FOR_UPDATE_EQE) {
+ pr_err("cid 0x%x: Unexpected state 0x%x after update ramrod\n",
+ conn_ctx->fw_cid, conn_ctx->state);
+
+ return -EINVAL;
+ }
+
+ return rc;
+}
+
+static int qedn_send_icreq(struct qedn_conn_ctx *conn_ctx)
+{
+ struct storage_sgl_task_params *sgl_task_params;
+ struct nvmetcp_task_params task_params;
+ struct qedn_task_ctx *qedn_task = NULL;
+ struct nvme_tcp_icreq_pdu icreq;
+ struct nvmetcp_wqe *chain_sqe;
+ struct nvmetcp_wqe local_sqe;
+
+ qedn_task = qedn_get_free_task_from_pool(conn_ctx, QEDN_ICREQ_CCCID);
+ if (!qedn_task)
+ return -EINVAL;
+
+ memset(&icreq, 0, sizeof(icreq));
+ memset(&local_sqe, 0, sizeof(local_sqe));
+
+ /* Initialize ICReq */
+ icreq.hdr.type = nvme_tcp_icreq;
+ icreq.hdr.hlen = sizeof(icreq);
+ icreq.hdr.pdo = 0;
+ icreq.hdr.plen = cpu_to_le32(icreq.hdr.hlen);
+ icreq.pfv = cpu_to_le16(conn_ctx->required_params.pfv);
+ icreq.maxr2t = cpu_to_le32(conn_ctx->required_params.maxr2t);
+ icreq.hpda = conn_ctx->required_params.hpda;
+ if (conn_ctx->required_params.hdr_digest)
+ icreq.digest |= NVME_TCP_HDR_DIGEST_ENABLE;
+ if (conn_ctx->required_params.data_digest)
+ icreq.digest |= NVME_TCP_DATA_DIGEST_ENABLE;
+
+ /* Initialize task params */
+ task_params.opq.lo = cpu_to_le32(((u64)(qedn_task)) & 0xffffffff);
+ task_params.opq.hi = cpu_to_le32(((u64)(qedn_task)) >> 32);
+ task_params.context = qedn_task->fw_task_ctx;
+ task_params.sqe = &local_sqe;
+ task_params.conn_icid = (u16)conn_ctx->conn_handle;
+ task_params.itid = qedn_task->itid;
+ task_params.cq_rss_number = conn_ctx->default_cq;
+ task_params.tx_io_size = QEDN_ICREQ_FW_PAYLOAD;
+ task_params.rx_io_size = 0; /* Rx doesn't use SGL for icresp */
+
+ /* Init SGE for ICReq padding */
+ sgl_task_params = &qedn_task->sgl_task_params;
+ sgl_task_params->total_buffer_size = task_params.tx_io_size;
+ sgl_task_params->small_mid_sge = false;
+ sgl_task_params->num_sges = 1;
+ memcpy(sgl_task_params->sgl, &conn_ctx->icreq_pad->sge,
+ sizeof(conn_ctx->icreq_pad->sge));
+
+ /* ICReq is sent as two parts.
+ * First part: (16 bytes + First 8 bytes of icreq.rsvd2[]) are sent
+ * via task context which is initialized above in icreq
+ * Second part: Rest bytes are sent via SGE, happening here
+ */
+ memcpy(conn_ctx->icreq_pad->buffer,
+ &icreq.rsvd2[QEDN_ICREQ_FW_PAYLOAD_START],
+ QEDN_ICREQ_FW_PAYLOAD);
+
+ qed_ops->init_icreq_exchange(&task_params, &icreq, sgl_task_params, NULL);
+
+ qedn_set_con_state(conn_ctx, CONN_STATE_WAIT_FOR_IC_COMP);
+ atomic_inc(&conn_ctx->num_active_fw_tasks);
+
+ /* spin_lock - doorbell is accessed both Rx flow and response flow */
+ spin_lock(&conn_ctx->ep.doorbell_lock);
+ chain_sqe = qed_chain_produce(&conn_ctx->ep.fw_sq_chain);
+ memcpy(chain_sqe, &local_sqe, sizeof(local_sqe));
+ qedn_ring_doorbell(conn_ctx);
+ spin_unlock(&conn_ctx->ep.doorbell_lock);
+
+ return 0;
+}
+
+void qedn_prep_icresp(struct qedn_conn_ctx *conn_ctx, struct nvmetcp_fw_cqe *cqe)
+{
+ struct nvmetcp_icresp_mdata *icresp_from_cqe =
+ (struct nvmetcp_icresp_mdata *)&cqe->cqe_data.icresp_mdata;
+ struct nvme_tcp_icresp_pdu *icresp = &conn_ctx->icresp;
+ struct nvme_tcp_ofld_ctrl *ctrl = conn_ctx->ctrl;
+ struct qedn_ctrl *qctrl = NULL;
+
+ qctrl = (struct qedn_ctrl *)ctrl->private_data;
+ if (!qctrl)
+ return;
+
+ icresp->pfv = cpu_to_le16(icresp_from_cqe->pfv);
+ icresp->cpda = icresp_from_cqe->cpda;
+ icresp->digest = icresp_from_cqe->digest;
+ icresp->maxdata = cpu_to_le32(icresp_from_cqe->maxdata);
+
+ qedn_set_sp_wa(conn_ctx, HANDLE_ICRESP);
+ queue_work(qctrl->sp_wq, &conn_ctx->sp_wq_entry);
+}
+
+static int qedn_handle_icresp(struct qedn_conn_ctx *conn_ctx)
+{
+ struct nvme_tcp_icresp_pdu *icresp = &conn_ctx->icresp;
+ int rc = 0;
+ u16 pfv;
+
+ /* Swapping requirement will be removed in future FW versions */
+ pfv = __swab16(le16_to_cpu(icresp->pfv));
+
+ qedn_free_icreq_pad(conn_ctx);
+
+ /* Validate ICResp */
+ if (pfv != conn_ctx->required_params.pfv) {
+ pr_err("cid %u: unsupported pfv %u\n", conn_ctx->fw_cid, pfv);
+
+ return -EINVAL;
+ }
+
+ if (icresp->cpda > conn_ctx->required_params.cpda) {
+ pr_err("cid %u: unsupported cpda %u\n", conn_ctx->fw_cid, icresp->cpda);
+
+ return -EINVAL;
+ }
+
+ if ((NVME_TCP_HDR_DIGEST_ENABLE & icresp->digest) !=
+ conn_ctx->required_params.hdr_digest) {
+ if ((NVME_TCP_HDR_DIGEST_ENABLE & icresp->digest) >
+ conn_ctx->required_params.hdr_digest) {
+ pr_err("cid 0x%x: invalid header digest bit\n", conn_ctx->fw_cid);
+ }
+ }
+
+ if ((NVME_TCP_DATA_DIGEST_ENABLE & icresp->digest) !=
+ conn_ctx->required_params.data_digest) {
+ if ((NVME_TCP_DATA_DIGEST_ENABLE & icresp->digest) >
+ conn_ctx->required_params.data_digest) {
+ pr_err("cid 0x%x: invalid data digest bit\n", conn_ctx->fw_cid);
+ }
+ }
+
+ memset(&conn_ctx->pdu_params, 0, sizeof(conn_ctx->pdu_params));
+ /* Swapping requirement will be removed in future FW versions */
+ conn_ctx->pdu_params.maxh2cdata =
+ __swab32(le32_to_cpu(icresp->maxdata));
+ conn_ctx->pdu_params.maxh2cdata = QEDN_MAX_PDU_SIZE;
+ if (conn_ctx->pdu_params.maxh2cdata > QEDN_MAX_PDU_SIZE)
+ conn_ctx->pdu_params.maxh2cdata = QEDN_MAX_PDU_SIZE;
+
+ conn_ctx->pdu_params.pfv = pfv;
+ conn_ctx->pdu_params.cpda = icresp->cpda;
+ conn_ctx->pdu_params.hpda = conn_ctx->required_params.hpda;
+ conn_ctx->pdu_params.hdr_digest = NVME_TCP_HDR_DIGEST_ENABLE & icresp->digest;
+ conn_ctx->pdu_params.data_digest = NVME_TCP_DATA_DIGEST_ENABLE & icresp->digest;
+ conn_ctx->pdu_params.maxr2t = conn_ctx->required_params.maxr2t;
+ rc = qedn_update_ramrod(conn_ctx);
+
+ return rc;
+}
+
/* Slowpath EQ Callback */
int qedn_event_cb(void *context, u8 fw_event_code, void *event_ring_data)
{
@@ -385,7 +667,8 @@ int qedn_event_cb(void *context, u8 fw_event_code, void *event_ring_data)
if (rc)
return rc;
- /* Placeholder - for ICReq flow */
+ qedn_set_sp_wa(conn_ctx, SEND_ICREQ);
+ queue_work(qctrl->sp_wq, &conn_ctx->sp_wq_entry);
}
break;
@@ -433,6 +716,8 @@ static int qedn_prep_and_offload_queue(struct qedn_conn_ctx *conn_ctx)
set_bit(QEDN_CONN_RESRC_FW_SQ, &conn_ctx->resrc_state);
+ spin_lock_init(&conn_ctx->ep.doorbell_lock);
+
atomic_set(&conn_ctx->num_active_tasks, 0);
atomic_set(&conn_ctx->num_active_fw_tasks, 0);
@@ -497,6 +782,11 @@ static int qedn_prep_and_offload_queue(struct qedn_conn_ctx *conn_ctx)
memset(conn_ctx->host_cccid_itid, 0xFF, dma_size);
set_bit(QEDN_CONN_RESRC_CCCID_ITID_MAP, &conn_ctx->resrc_state);
+
+ rc = qedn_alloc_icreq_pad(conn_ctx);
+ if (rc)
+ goto rel_conn;
+
rc = qedn_set_con_state(conn_ctx, CONN_STATE_WAIT_FOR_CONNECT_DONE);
if (rc)
goto rel_conn;
@@ -569,6 +859,9 @@ void qedn_sp_wq_handler(struct work_struct *work)
qedn = conn_ctx->qedn;
if (test_bit(DESTROY_CONNECTION, &conn_ctx->agg_work_action)) {
+ if (test_bit(HANDLE_ICRESP, &conn_ctx->agg_work_action))
+ qedn_clr_sp_wa(conn_ctx, HANDLE_ICRESP);
+
qedn_destroy_connection(conn_ctx);
return;
@@ -583,6 +876,36 @@ void qedn_sp_wq_handler(struct work_struct *work)
return;
}
}
+
+ if (test_bit(SEND_ICREQ, &conn_ctx->agg_work_action)) {
+ qedn_clr_sp_wa(conn_ctx, SEND_ICREQ);
+ rc = qedn_send_icreq(conn_ctx);
+ if (rc)
+ return;
+
+ return;
+ }
+
+ if (test_bit(HANDLE_ICRESP, &conn_ctx->agg_work_action)) {
+ rc = qedn_handle_icresp(conn_ctx);
+
+ qedn_clr_sp_wa(conn_ctx, HANDLE_ICRESP);
+ if (rc) {
+ pr_err("IC handling returned with 0x%x\n", rc);
+ if (test_and_set_bit(DESTROY_CONNECTION, &conn_ctx->agg_work_action))
+ return;
+
+ qedn_destroy_connection(conn_ctx);
+
+ return;
+ }
+
+ atomic_inc(&conn_ctx->est_conn_indicator);
+ qedn_set_con_state(conn_ctx, CONN_STATE_NVMETCP_CONN_ESTABLISHED);
+ wake_up_interruptible(&conn_ctx->conn_waitq);
+
+ return;
+ }
}
/* Clear connection aggregative slowpath work action */
diff --git a/drivers/nvme/hw/qedn/qedn_main.c b/drivers/nvme/hw/qedn/qedn_main.c
index 471044351767..d262f9751322 100644
--- a/drivers/nvme/hw/qedn/qedn_main.c
+++ b/drivers/nvme/hw/qedn/qedn_main.c
@@ -275,6 +275,19 @@ static int qedn_release_ctrl(struct nvme_tcp_ofld_ctrl *ctrl)
return 0;
}
+static void qedn_set_pdu_params(struct qedn_conn_ctx *conn_ctx)
+{
+ /* Enable digest once supported */
+ conn_ctx->required_params.hdr_digest = 0;
+ conn_ctx->required_params.data_digest = 0;
+
+ conn_ctx->required_params.maxr2t = QEDN_MAX_OUTSTANDING_R2T_PDUS;
+ conn_ctx->required_params.pfv = NVME_TCP_PFV_1_0;
+ conn_ctx->required_params.cpda = 0;
+ conn_ctx->required_params.hpda = 0;
+ conn_ctx->required_params.maxh2cdata = QEDN_MAX_PDU_SIZE;
+}
+
static int qedn_create_queue(struct nvme_tcp_ofld_queue *queue, int qid,
size_t queue_size)
{
@@ -306,6 +319,7 @@ static int qedn_create_queue(struct nvme_tcp_ofld_queue *queue, int qid,
conn_ctx->ctrl = ctrl;
conn_ctx->sq_depth = queue_size;
mutex_init(&conn_ctx->send_mutex);
+ qedn_set_pdu_params(conn_ctx);
init_waitqueue_head(&conn_ctx->conn_waitq);
atomic_set(&conn_ctx->est_conn_indicator, 0);
diff --git a/drivers/nvme/hw/qedn/qedn_task.c b/drivers/nvme/hw/qedn/qedn_task.c
index 71d0cda89e39..1f484b6ffaca 100644
--- a/drivers/nvme/hw/qedn/qedn_task.c
+++ b/drivers/nvme/hw/qedn/qedn_task.c
@@ -386,9 +386,11 @@ void qedn_io_work_cq(struct qedn_ctx *qedn, struct nvmetcp_fw_cqe *cqe)
break;
case NVMETCP_TASK_TYPE_INIT_CONN_REQUEST:
-
- /* Placeholder - ICReq flow */
-
+ /* Clear ICReq-padding SGE from SGL */
+ qedn_common_clear_fw_sgl(&qedn_task->sgl_task_params);
+ /* Task is not required for icresp processing */
+ qedn_return_task_to_pool(conn_ctx, qedn_task);
+ qedn_prep_icresp(conn_ctx, cqe);
break;
default:
pr_info("Could not identify task type\n");
--
2.22.0
Powered by blists - more mailing lists