[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20220126130140.66316-2-tonylu@linux.alibaba.com>
Date: Wed, 26 Jan 2022 21:01:41 +0800
From: Tony Lu <tonylu@...ux.alibaba.com>
To: kgraul@...ux.ibm.com, kuba@...nel.org, davem@...emloft.net
Cc: netdev@...r.kernel.org, linux-s390@...r.kernel.org,
linux-rdma@...r.kernel.org
Subject: [PATCH net-next 1/2] net/smc: Introduce smc_ib_cq to bind link and cq
This patch introduces struct smc_ib_cq as a medium between smc_link and
ib_cq. Every smc_link can access ib_cq from their own, and unbinds
smc_link from smc_ib_device. This allows flexible mapping, prepares for
multiple CQs support.
Signed-off-by: Tony Lu <tonylu@...ux.alibaba.com>
---
net/smc/smc_core.h | 2 ++
net/smc/smc_ib.c | 85 +++++++++++++++++++++++++++++++++-------------
net/smc/smc_ib.h | 13 ++++---
net/smc/smc_wr.c | 34 +++++++++----------
4 files changed, 89 insertions(+), 45 deletions(-)
diff --git a/net/smc/smc_core.h b/net/smc/smc_core.h
index 521c64a3d8d3..fd10cad8fb77 100644
--- a/net/smc/smc_core.h
+++ b/net/smc/smc_core.h
@@ -86,6 +86,8 @@ struct smc_link {
struct ib_pd *roce_pd; /* IB protection domain,
* unique for every RoCE QP
*/
+ struct smc_ib_cq *smcibcq_recv; /* cq for recv */
+ struct smc_ib_cq *smcibcq_send; /* cq for send */
struct ib_qp *roce_qp; /* IB queue pair */
struct ib_qp_attr qp_attr; /* IB queue pair attributes */
diff --git a/net/smc/smc_ib.c b/net/smc/smc_ib.c
index a3e2d3b89568..0d98cf440adc 100644
--- a/net/smc/smc_ib.c
+++ b/net/smc/smc_ib.c
@@ -133,7 +133,7 @@ int smc_ib_ready_link(struct smc_link *lnk)
if (rc)
goto out;
smc_wr_remember_qp_attr(lnk);
- rc = ib_req_notify_cq(lnk->smcibdev->roce_cq_recv,
+ rc = ib_req_notify_cq(lnk->smcibcq_recv->ib_cq,
IB_CQ_SOLICITED_MASK);
if (rc)
goto out;
@@ -651,6 +651,8 @@ void smc_ib_destroy_queue_pair(struct smc_link *lnk)
if (lnk->roce_qp)
ib_destroy_qp(lnk->roce_qp);
lnk->roce_qp = NULL;
+ lnk->smcibcq_send = NULL;
+ lnk->smcibcq_recv = NULL;
}
/* create a queue pair within the protection domain for a link */
@@ -660,8 +662,8 @@ int smc_ib_create_queue_pair(struct smc_link *lnk)
struct ib_qp_init_attr qp_attr = {
.event_handler = smc_ib_qp_event_handler,
.qp_context = lnk,
- .send_cq = lnk->smcibdev->roce_cq_send,
- .recv_cq = lnk->smcibdev->roce_cq_recv,
+ .send_cq = lnk->smcibdev->ib_cq_send->ib_cq,
+ .recv_cq = lnk->smcibdev->ib_cq_recv->ib_cq,
.srq = NULL,
.cap = {
/* include unsolicited rdma_writes as well,
@@ -679,10 +681,13 @@ int smc_ib_create_queue_pair(struct smc_link *lnk)
lnk->roce_qp = ib_create_qp(lnk->roce_pd, &qp_attr);
rc = PTR_ERR_OR_ZERO(lnk->roce_qp);
- if (IS_ERR(lnk->roce_qp))
+ if (IS_ERR(lnk->roce_qp)) {
lnk->roce_qp = NULL;
- else
+ } else {
+ lnk->smcibcq_send = lnk->smcibdev->ib_cq_send;
+ lnk->smcibcq_recv = lnk->smcibdev->ib_cq_recv;
smc_wr_remember_qp_attr(lnk);
+ }
return rc;
}
@@ -799,10 +804,21 @@ void smc_ib_buf_unmap_sg(struct smc_link *lnk,
buf_slot->sgt[lnk->link_idx].sgl->dma_address = 0;
}
+static void smc_ib_cleanup_cq(struct smc_ib_device *smcibdev)
+{
+ ib_destroy_cq(smcibdev->ib_cq_send->ib_cq);
+ kfree(smcibdev->ib_cq_send);
+ smcibdev->ib_cq_send = NULL;
+
+ ib_destroy_cq(smcibdev->ib_cq_recv->ib_cq);
+ kfree(smcibdev->ib_cq_recv);
+ smcibdev->ib_cq_recv = NULL;
+}
+
long smc_ib_setup_per_ibdev(struct smc_ib_device *smcibdev)
{
- struct ib_cq_init_attr cqattr = {
- .cqe = SMC_MAX_CQE, .comp_vector = 0 };
+ struct ib_cq_init_attr cqattr = { .cqe = SMC_MAX_CQE };
+ struct smc_ib_cq *smcibcq_send, *smcibcq_recv;
int cqe_size_order, smc_order;
long rc;
@@ -815,28 +831,50 @@ long smc_ib_setup_per_ibdev(struct smc_ib_device *smcibdev)
smc_order = MAX_ORDER - cqe_size_order - 1;
if (SMC_MAX_CQE + 2 > (0x00000001 << smc_order) * PAGE_SIZE)
cqattr.cqe = (0x00000001 << smc_order) * PAGE_SIZE - 2;
- smcibdev->roce_cq_send = ib_create_cq(smcibdev->ibdev,
- smc_wr_tx_cq_handler, NULL,
- smcibdev, &cqattr);
- rc = PTR_ERR_OR_ZERO(smcibdev->roce_cq_send);
- if (IS_ERR(smcibdev->roce_cq_send)) {
- smcibdev->roce_cq_send = NULL;
+
+ smcibcq_send = kzalloc(sizeof(*smcibcq_send), GFP_KERNEL);
+ if (!smcibcq_send) {
+ rc = -ENOMEM;
+ goto out;
+ }
+ smcibcq_send->smcibdev = smcibdev;
+ smcibcq_send->is_send = 1;
+ cqattr.comp_vector = 0;
+ smcibcq_send->ib_cq = ib_create_cq(smcibdev->ibdev,
+ smc_wr_tx_cq_handler, NULL,
+ smcibcq_send, &cqattr);
+ rc = PTR_ERR_OR_ZERO(smcibdev->ib_cq_send);
+ if (IS_ERR(smcibdev->ib_cq_send)) {
+ smcibdev->ib_cq_send = NULL;
goto out;
}
- smcibdev->roce_cq_recv = ib_create_cq(smcibdev->ibdev,
- smc_wr_rx_cq_handler, NULL,
- smcibdev, &cqattr);
- rc = PTR_ERR_OR_ZERO(smcibdev->roce_cq_recv);
- if (IS_ERR(smcibdev->roce_cq_recv)) {
- smcibdev->roce_cq_recv = NULL;
- goto err;
+ smcibdev->ib_cq_send = smcibcq_send;
+
+ smcibcq_recv = kzalloc(sizeof(*smcibcq_recv), GFP_KERNEL);
+ if (!smcibcq_recv) {
+ rc = -ENOMEM;
+ goto err_send;
+ }
+ smcibcq_recv->smcibdev = smcibdev;
+ cqattr.comp_vector = 1;
+ smcibcq_recv->ib_cq = ib_create_cq(smcibdev->ibdev,
+ smc_wr_rx_cq_handler, NULL,
+ smcibcq_recv, &cqattr);
+ rc = PTR_ERR_OR_ZERO(smcibdev->ib_cq_recv);
+ if (IS_ERR(smcibdev->ib_cq_recv)) {
+ smcibdev->ib_cq_recv = NULL;
+ goto err_recv;
}
+ smcibdev->ib_cq_recv = smcibcq_recv;
smc_wr_add_dev(smcibdev);
smcibdev->initialized = 1;
goto out;
-err:
- ib_destroy_cq(smcibdev->roce_cq_send);
+err_recv:
+ kfree(smcibcq_recv);
+ ib_destroy_cq(smcibcq_send->ib_cq);
+err_send:
+ kfree(smcibcq_send);
out:
mutex_unlock(&smcibdev->mutex);
return rc;
@@ -848,8 +886,7 @@ static void smc_ib_cleanup_per_ibdev(struct smc_ib_device *smcibdev)
if (!smcibdev->initialized)
goto out;
smcibdev->initialized = 0;
- ib_destroy_cq(smcibdev->roce_cq_recv);
- ib_destroy_cq(smcibdev->roce_cq_send);
+ smc_ib_cleanup_cq(smcibdev);
smc_wr_remove_dev(smcibdev);
out:
mutex_unlock(&smcibdev->mutex);
diff --git a/net/smc/smc_ib.h b/net/smc/smc_ib.h
index 5d8b49c57f50..1dc567599977 100644
--- a/net/smc/smc_ib.h
+++ b/net/smc/smc_ib.h
@@ -32,15 +32,20 @@ struct smc_ib_devices { /* list of smc ib devices definition */
extern struct smc_ib_devices smc_ib_devices; /* list of smc ib devices */
extern struct smc_lgr_list smc_lgr_list; /* list of linkgroups */
+struct smc_ib_cq { /* ib_cq wrapper for smc */
+ struct smc_ib_device *smcibdev; /* parent ib device */
+ struct ib_cq *ib_cq; /* real ib_cq for link */
+ struct tasklet_struct tasklet; /* tasklet for wr */
+ bool is_send; /* send for recv cq */
+};
+
struct smc_ib_device { /* ib-device infos for smc */
struct list_head list;
struct ib_device *ibdev;
struct ib_port_attr pattr[SMC_MAX_PORTS]; /* ib dev. port attrs */
struct ib_event_handler event_handler; /* global ib_event handler */
- struct ib_cq *roce_cq_send; /* send completion queue */
- struct ib_cq *roce_cq_recv; /* recv completion queue */
- struct tasklet_struct send_tasklet; /* called by send cq handler */
- struct tasklet_struct recv_tasklet; /* called by recv cq handler */
+ struct smc_ib_cq *ib_cq_send; /* send completion queue */
+ struct smc_ib_cq *ib_cq_recv; /* recv completion queue */
char mac[SMC_MAX_PORTS][ETH_ALEN];
/* mac address per port*/
u8 pnetid[SMC_MAX_PORTS][SMC_MAX_PNETID_LEN];
diff --git a/net/smc/smc_wr.c b/net/smc/smc_wr.c
index 24be1d03fef9..ddb0ba67a851 100644
--- a/net/smc/smc_wr.c
+++ b/net/smc/smc_wr.c
@@ -135,7 +135,7 @@ static inline void smc_wr_tx_process_cqe(struct ib_wc *wc)
static void smc_wr_tx_tasklet_fn(struct tasklet_struct *t)
{
- struct smc_ib_device *dev = from_tasklet(dev, t, send_tasklet);
+ struct smc_ib_cq *smcibcq = from_tasklet(smcibcq, t, tasklet);
struct ib_wc wc[SMC_WR_MAX_POLL_CQE];
int i = 0, rc;
int polled = 0;
@@ -144,9 +144,9 @@ static void smc_wr_tx_tasklet_fn(struct tasklet_struct *t)
polled++;
do {
memset(&wc, 0, sizeof(wc));
- rc = ib_poll_cq(dev->roce_cq_send, SMC_WR_MAX_POLL_CQE, wc);
+ rc = ib_poll_cq(smcibcq->ib_cq, SMC_WR_MAX_POLL_CQE, wc);
if (polled == 1) {
- ib_req_notify_cq(dev->roce_cq_send,
+ ib_req_notify_cq(smcibcq->ib_cq,
IB_CQ_NEXT_COMP |
IB_CQ_REPORT_MISSED_EVENTS);
}
@@ -161,9 +161,9 @@ static void smc_wr_tx_tasklet_fn(struct tasklet_struct *t)
void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context)
{
- struct smc_ib_device *dev = (struct smc_ib_device *)cq_context;
+ struct smc_ib_cq *smcibcq = (struct smc_ib_cq *)cq_context;
- tasklet_schedule(&dev->send_tasklet);
+ tasklet_schedule(&smcibcq->tasklet);
}
/*---------------------------- request submission ---------------------------*/
@@ -306,7 +306,7 @@ int smc_wr_tx_send(struct smc_link *link, struct smc_wr_tx_pend_priv *priv)
struct smc_wr_tx_pend *pend;
int rc;
- ib_req_notify_cq(link->smcibdev->roce_cq_send,
+ ib_req_notify_cq(link->smcibcq_send->ib_cq,
IB_CQ_NEXT_COMP | IB_CQ_REPORT_MISSED_EVENTS);
pend = container_of(priv, struct smc_wr_tx_pend, priv);
rc = ib_post_send(link->roce_qp, &link->wr_tx_ibs[pend->idx], NULL);
@@ -323,7 +323,7 @@ int smc_wr_tx_v2_send(struct smc_link *link, struct smc_wr_tx_pend_priv *priv,
int rc;
link->wr_tx_v2_ib->sg_list[0].length = len;
- ib_req_notify_cq(link->smcibdev->roce_cq_send,
+ ib_req_notify_cq(link->smcibcq_send->ib_cq,
IB_CQ_NEXT_COMP | IB_CQ_REPORT_MISSED_EVENTS);
rc = ib_post_send(link->roce_qp, link->wr_tx_v2_ib, NULL);
if (rc) {
@@ -367,7 +367,7 @@ int smc_wr_reg_send(struct smc_link *link, struct ib_mr *mr)
{
int rc;
- ib_req_notify_cq(link->smcibdev->roce_cq_send,
+ ib_req_notify_cq(link->smcibcq_send->ib_cq,
IB_CQ_NEXT_COMP | IB_CQ_REPORT_MISSED_EVENTS);
link->wr_reg_state = POSTED;
link->wr_reg.wr.wr_id = (u64)(uintptr_t)mr;
@@ -476,7 +476,7 @@ static inline void smc_wr_rx_process_cqes(struct ib_wc wc[], int num)
static void smc_wr_rx_tasklet_fn(struct tasklet_struct *t)
{
- struct smc_ib_device *dev = from_tasklet(dev, t, recv_tasklet);
+ struct smc_ib_cq *smcibcq = from_tasklet(smcibcq, t, tasklet);
struct ib_wc wc[SMC_WR_MAX_POLL_CQE];
int polled = 0;
int rc;
@@ -485,9 +485,9 @@ static void smc_wr_rx_tasklet_fn(struct tasklet_struct *t)
polled++;
do {
memset(&wc, 0, sizeof(wc));
- rc = ib_poll_cq(dev->roce_cq_recv, SMC_WR_MAX_POLL_CQE, wc);
+ rc = ib_poll_cq(smcibcq->ib_cq, SMC_WR_MAX_POLL_CQE, wc);
if (polled == 1) {
- ib_req_notify_cq(dev->roce_cq_recv,
+ ib_req_notify_cq(smcibcq->ib_cq,
IB_CQ_SOLICITED_MASK
| IB_CQ_REPORT_MISSED_EVENTS);
}
@@ -501,9 +501,9 @@ static void smc_wr_rx_tasklet_fn(struct tasklet_struct *t)
void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context)
{
- struct smc_ib_device *dev = (struct smc_ib_device *)cq_context;
+ struct smc_ib_cq *smcibcq = (struct smc_ib_cq *)cq_context;
- tasklet_schedule(&dev->recv_tasklet);
+ tasklet_schedule(&smcibcq->tasklet);
}
int smc_wr_rx_post_init(struct smc_link *link)
@@ -830,14 +830,14 @@ int smc_wr_alloc_link_mem(struct smc_link *link)
void smc_wr_remove_dev(struct smc_ib_device *smcibdev)
{
- tasklet_kill(&smcibdev->recv_tasklet);
- tasklet_kill(&smcibdev->send_tasklet);
+ tasklet_kill(&smcibdev->ib_cq_recv->tasklet);
+ tasklet_kill(&smcibdev->ib_cq_send->tasklet);
}
void smc_wr_add_dev(struct smc_ib_device *smcibdev)
{
- tasklet_setup(&smcibdev->recv_tasklet, smc_wr_rx_tasklet_fn);
- tasklet_setup(&smcibdev->send_tasklet, smc_wr_tx_tasklet_fn);
+ tasklet_setup(&smcibdev->ib_cq_recv->tasklet, smc_wr_rx_tasklet_fn);
+ tasklet_setup(&smcibdev->ib_cq_send->tasklet, smc_wr_tx_tasklet_fn);
}
int smc_wr_create_link(struct smc_link *lnk)
--
2.32.0.3.g01195cf9f
Powered by blists - more mailing lists