[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20240621183947.4105278-8-allen.lkml@gmail.com>
Date: Fri, 21 Jun 2024 11:39:39 -0700
From: Allen Pais <allen.lkml@...il.com>
To: netdev@...r.kernel.org
Cc: Allen Pais <allen.lkml@...il.com>
Subject: [PATCH 07/15] net: thunderx: Convert tasklet API to new bottom half workqueue mechanism
Migrate tasklet APIs to the new bottom half workqueue mechanism. It
replaces all occurrences of tasklet usage with the appropriate workqueue
APIs throughout the cavium/thunderx driver. This transition ensures
compatibility with the latest design and enhances performance.
Signed-off-by: Allen Pais <allen.lkml@...il.com>
---
drivers/net/ethernet/cavium/thunder/nic.h | 5 ++--
.../net/ethernet/cavium/thunder/nicvf_main.c | 24 +++++++++----------
.../ethernet/cavium/thunder/nicvf_queues.c | 4 ++--
.../ethernet/cavium/thunder/nicvf_queues.h | 2 +-
4 files changed, 18 insertions(+), 17 deletions(-)
diff --git a/drivers/net/ethernet/cavium/thunder/nic.h b/drivers/net/ethernet/cavium/thunder/nic.h
index 090d6b83982a..ecc175b6e7fa 100644
--- a/drivers/net/ethernet/cavium/thunder/nic.h
+++ b/drivers/net/ethernet/cavium/thunder/nic.h
@@ -8,6 +8,7 @@
#include <linux/netdevice.h>
#include <linux/interrupt.h>
+#include <linux/workqueue.h>
#include <linux/pci.h>
#include "thunder_bgx.h"
@@ -295,7 +296,7 @@ struct nicvf {
bool rb_work_scheduled;
struct page *rb_page;
struct delayed_work rbdr_work;
- struct tasklet_struct rbdr_task;
+ struct work_struct rbdr_bh_work;
/* Secondary Qset */
u8 sqs_count;
@@ -319,7 +320,7 @@ struct nicvf {
bool loopback_supported;
struct nicvf_rss_info rss_info;
struct nicvf_pfc pfc;
- struct tasklet_struct qs_err_task;
+ struct work_struct qs_err_bh_work;
struct work_struct reset_task;
struct nicvf_work rx_mode_work;
/* spinlock to protect workqueue arguments from concurrent access */
diff --git a/drivers/net/ethernet/cavium/thunder/nicvf_main.c b/drivers/net/ethernet/cavium/thunder/nicvf_main.c
index aebb9fef3f6e..b0878bd25cf0 100644
--- a/drivers/net/ethernet/cavium/thunder/nicvf_main.c
+++ b/drivers/net/ethernet/cavium/thunder/nicvf_main.c
@@ -982,9 +982,9 @@ static int nicvf_poll(struct napi_struct *napi, int budget)
*
* As of now only CQ errors are handled
*/
-static void nicvf_handle_qs_err(struct tasklet_struct *t)
+static void nicvf_handle_qs_err(struct work_struct *work)
{
- struct nicvf *nic = from_tasklet(nic, t, qs_err_task);
+ struct nicvf *nic = from_work(nic, work, qs_err_bh_work);
struct queue_set *qs = nic->qs;
int qidx;
u64 status;
@@ -1069,7 +1069,7 @@ static irqreturn_t nicvf_rbdr_intr_handler(int irq, void *nicvf_irq)
if (!nicvf_is_intr_enabled(nic, NICVF_INTR_RBDR, qidx))
continue;
nicvf_disable_intr(nic, NICVF_INTR_RBDR, qidx);
- tasklet_hi_schedule(&nic->rbdr_task);
+ queue_work(system_bh_highpri_wq, &nic->rbdr_bh_work);
/* Clear interrupt */
nicvf_clear_intr(nic, NICVF_INTR_RBDR, qidx);
}
@@ -1085,7 +1085,7 @@ static irqreturn_t nicvf_qs_err_intr_handler(int irq, void *nicvf_irq)
/* Disable Qset err interrupt and schedule softirq */
nicvf_disable_intr(nic, NICVF_INTR_QS_ERR, 0);
- tasklet_hi_schedule(&nic->qs_err_task);
+ queue_work(system_bh_highpri_wq, &nic->qs_err_bh_work);
nicvf_clear_intr(nic, NICVF_INTR_QS_ERR, 0);
return IRQ_HANDLED;
@@ -1364,8 +1364,8 @@ int nicvf_stop(struct net_device *netdev)
for (irq = 0; irq < nic->num_vec; irq++)
synchronize_irq(pci_irq_vector(nic->pdev, irq));
- tasklet_kill(&nic->rbdr_task);
- tasklet_kill(&nic->qs_err_task);
+ cancel_work_sync(&nic->rbdr_bh_work);
+ cancel_work_sync(&nic->qs_err_bh_work);
if (nic->rb_work_scheduled)
cancel_delayed_work_sync(&nic->rbdr_work);
@@ -1488,11 +1488,11 @@ int nicvf_open(struct net_device *netdev)
nicvf_hw_set_mac_addr(nic, netdev);
}
- /* Init tasklet for handling Qset err interrupt */
- tasklet_setup(&nic->qs_err_task, nicvf_handle_qs_err);
+ /* Init bh_work for handling Qset err interrupt */
+ INIT_WORK(&nic->qs_err_bh_work, nicvf_handle_qs_err);
- /* Init RBDR tasklet which will refill RBDR */
- tasklet_setup(&nic->rbdr_task, nicvf_rbdr_task);
+ /* Init RBDR bh_work which will refill RBDR */
+ INIT_WORK(&nic->rbdr_bh_work, nicvf_rbdr_bh_work);
INIT_DELAYED_WORK(&nic->rbdr_work, nicvf_rbdr_work);
/* Configure CPI alorithm */
@@ -1561,8 +1561,8 @@ int nicvf_open(struct net_device *netdev)
cleanup:
nicvf_disable_intr(nic, NICVF_INTR_MBOX, 0);
nicvf_unregister_interrupts(nic);
- tasklet_kill(&nic->qs_err_task);
- tasklet_kill(&nic->rbdr_task);
+ cancel_work_sync(&nic->qs_err_bh_work);
+ cancel_work_sync(&nic->rbdr_bh_work);
napi_del:
for (qidx = 0; qidx < qs->cq_cnt; qidx++) {
cq_poll = nic->napi[qidx];
diff --git a/drivers/net/ethernet/cavium/thunder/nicvf_queues.c b/drivers/net/ethernet/cavium/thunder/nicvf_queues.c
index 06397cc8bb36..ad71160879e4 100644
--- a/drivers/net/ethernet/cavium/thunder/nicvf_queues.c
+++ b/drivers/net/ethernet/cavium/thunder/nicvf_queues.c
@@ -461,9 +461,9 @@ void nicvf_rbdr_work(struct work_struct *work)
}
/* In Softirq context, alloc rcv buffers in atomic mode */
-void nicvf_rbdr_task(struct tasklet_struct *t)
+void nicvf_rbdr_bh_work(struct work_struct *work)
{
- struct nicvf *nic = from_tasklet(nic, t, rbdr_task);
+ struct nicvf *nic = from_work(nic, work, rbdr_bh_work);
nicvf_refill_rbdr(nic, GFP_ATOMIC);
if (nic->rb_alloc_fail) {
diff --git a/drivers/net/ethernet/cavium/thunder/nicvf_queues.h b/drivers/net/ethernet/cavium/thunder/nicvf_queues.h
index 8453defc296c..c6f18fb7c50e 100644
--- a/drivers/net/ethernet/cavium/thunder/nicvf_queues.h
+++ b/drivers/net/ethernet/cavium/thunder/nicvf_queues.h
@@ -348,7 +348,7 @@ void nicvf_xdp_sq_doorbell(struct nicvf *nic, struct snd_queue *sq, int sq_num);
struct sk_buff *nicvf_get_rcv_skb(struct nicvf *nic,
struct cqe_rx_t *cqe_rx, bool xdp);
-void nicvf_rbdr_task(struct tasklet_struct *t);
+void nicvf_rbdr_bh_work(struct work_struct *work);
void nicvf_rbdr_work(struct work_struct *work);
void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx);
--
2.34.1
Powered by blists - more mailing lists