lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <aWZxNFIh2trMm04T@lizhi-Precision-Tower-5810>
Date: Tue, 13 Jan 2026 11:22:12 -0500
From: Frank Li <Frank.li@....com>
To: Wei Fang <wei.fang@....com>
Cc: shenwei.wang@....com, xiaoning.wang@....com, andrew+netdev@...n.ch,
	davem@...emloft.net, edumazet@...gle.com, kuba@...nel.org,
	pabeni@...hat.com, ast@...nel.org, daniel@...earbox.net,
	hawk@...nel.org, john.fastabend@...il.com, sdf@...ichev.me,
	netdev@...r.kernel.org, linux-kernel@...r.kernel.org,
	imx@...ts.linux.dev, bpf@...r.kernel.org
Subject: Re: [PATCH net-next 07/11] net: fec: use switch statement to check
 the type of tx_buf

On Tue, Jan 13, 2026 at 11:29:35AM +0800, Wei Fang wrote:
> The tx_buf has three types: FEC_TXBUF_T_SKB, FEC_TXBUF_T_XDP_NDO and
> FEC_TXBUF_T_XDP_TX. Currently, the driver uses 'if...else...' statements
> to check the type and perform the corresponding processing. This is very
> detrimental to future expansion. For example, if new types are added to
> support XDP zero copy in the future, continuing to use 'if...else...'
> would be a very bad coding style. So the 'if...else...' statements in
> the current driver are replaced with switch statements to support XDP
> zero copy in the future.
>
> Signed-off-by: Wei Fang <wei.fang@....com>
> ---
>  drivers/net/ethernet/freescale/fec_main.c | 167 +++++++++++-----------
>  1 file changed, 82 insertions(+), 85 deletions(-)
>
> diff --git a/drivers/net/ethernet/freescale/fec_main.c b/drivers/net/ethernet/freescale/fec_main.c
> index f3e93598a27c..3bd89d7f105b 100644
> --- a/drivers/net/ethernet/freescale/fec_main.c
> +++ b/drivers/net/ethernet/freescale/fec_main.c
> @@ -1023,33 +1023,33 @@ static void fec_enet_bd_init(struct net_device *dev)
>  		txq->bd.cur = bdp;
>
>  		for (i = 0; i < txq->bd.ring_size; i++) {
> +			dma_addr_t dma = fec32_to_cpu(bdp->cbd_bufaddr);
> +			struct page *page;
> +
>  			/* Initialize the BD for every fragment in the page. */
>  			bdp->cbd_sc = cpu_to_fec16(0);
> -			if (txq->tx_buf[i].type == FEC_TXBUF_T_SKB) {
> -				if (bdp->cbd_bufaddr &&
> -				    !IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr)))
> -					dma_unmap_single(&fep->pdev->dev,
> -							 fec32_to_cpu(bdp->cbd_bufaddr),
> -							 fec16_to_cpu(bdp->cbd_datlen),
> -							 DMA_TO_DEVICE);
> -				if (txq->tx_buf[i].buf_p)
> -					dev_kfree_skb_any(txq->tx_buf[i].buf_p);
> -			} else if (txq->tx_buf[i].type == FEC_TXBUF_T_XDP_NDO) {
> -				if (bdp->cbd_bufaddr)
> -					dma_unmap_single(&fep->pdev->dev,
> -							 fec32_to_cpu(bdp->cbd_bufaddr),
> +			switch (txq->tx_buf[i].type) {
> +			case FEC_TXBUF_T_SKB:
> +				if (dma && !IS_TSO_HEADER(txq, dma))
> +					dma_unmap_single(&fep->pdev->dev, dma,
>  							 fec16_to_cpu(bdp->cbd_datlen),
>  							 DMA_TO_DEVICE);
>
> -				if (txq->tx_buf[i].buf_p)
> -					xdp_return_frame(txq->tx_buf[i].buf_p);
> -			} else {
> -				struct page *page = txq->tx_buf[i].buf_p;
> -
> -				if (page)
> -					page_pool_put_page(pp_page_to_nmdesc(page)->pp,
> -							   page, 0,
> -							   false);
> +				dev_kfree_skb_any(txq->tx_buf[i].buf_p);
> +				break;
> +			case FEC_TXBUF_T_XDP_NDO:
> +				dma_unmap_single(&fep->pdev->dev, dma,
> +						 fec16_to_cpu(bdp->cbd_datlen),
> +						 DMA_TO_DEVICE);
> +				xdp_return_frame(txq->tx_buf[i].buf_p);

look like logic is not exactly same as original one

if (txq->tx_buf[i].type == FEC_TXBUF_T_XDP_NDO) {
	if (bdp->cbd_bufaddr)
		...

Frank

> +				break;
> +			case FEC_TXBUF_T_XDP_TX:
> +				page = txq->tx_buf[i].buf_p;
> +				page_pool_put_page(pp_page_to_nmdesc(page)->pp,
> +						   page, 0, false);
> +				break;
> +			default:
> +				break;
>  			}
>
>  			txq->tx_buf[i].buf_p = NULL;
> @@ -1514,45 +1514,66 @@ fec_enet_tx_queue(struct net_device *ndev, u16 queue_id, int budget)
>  			break;
>
>  		index = fec_enet_get_bd_index(bdp, &txq->bd);
> +		frame_len = fec16_to_cpu(bdp->cbd_datlen);
>
> -		if (txq->tx_buf[index].type == FEC_TXBUF_T_SKB) {
> -			skb = txq->tx_buf[index].buf_p;
> +		switch (txq->tx_buf[index].type) {
> +		case FEC_TXBUF_T_SKB:
>  			if (bdp->cbd_bufaddr &&
>  			    !IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr)))
>  				dma_unmap_single(&fep->pdev->dev,
>  						 fec32_to_cpu(bdp->cbd_bufaddr),
> -						 fec16_to_cpu(bdp->cbd_datlen),
> -						 DMA_TO_DEVICE);
> -			bdp->cbd_bufaddr = cpu_to_fec32(0);
> +						 frame_len, DMA_TO_DEVICE);
> +
> +			skb = txq->tx_buf[index].buf_p;
>  			if (!skb)
>  				goto tx_buf_done;
> -		} else {
> +
> +			frame_len = skb->len;
> +
> +			/* NOTE: SKBTX_IN_PROGRESS being set does not imply it's we who
> +			 * are to time stamp the packet, so we still need to check time
> +			 * stamping enabled flag.
> +			 */
> +			if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS &&
> +				     fep->hwts_tx_en) && fep->bufdesc_ex) {
> +				struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
> +				struct skb_shared_hwtstamps shhwtstamps;
> +
> +				fec_enet_hwtstamp(fep, fec32_to_cpu(ebdp->ts), &shhwtstamps);
> +				skb_tstamp_tx(skb, &shhwtstamps);
> +			}
> +
> +			/* Free the sk buffer associated with this last transmit */
> +			napi_consume_skb(skb, budget);
> +			break;
> +		case FEC_TXBUF_T_XDP_NDO:
>  			/* Tx processing cannot call any XDP (or page pool) APIs if
>  			 * the "budget" is 0. Because NAPI is called with budget of
>  			 * 0 (such as netpoll) indicates we may be in an IRQ context,
>  			 * however, we can't use the page pool from IRQ context.
>  			 */
>  			if (unlikely(!budget))
> -				break;
> +				goto out;
>
> -			if (txq->tx_buf[index].type == FEC_TXBUF_T_XDP_NDO) {
> -				xdpf = txq->tx_buf[index].buf_p;
> -				if (bdp->cbd_bufaddr)
> -					dma_unmap_single(&fep->pdev->dev,
> -							 fec32_to_cpu(bdp->cbd_bufaddr),
> -							 fec16_to_cpu(bdp->cbd_datlen),
> -							 DMA_TO_DEVICE);
> -			} else {
> -				page = txq->tx_buf[index].buf_p;
> -			}
> -
> -			bdp->cbd_bufaddr = cpu_to_fec32(0);
> -			if (unlikely(!txq->tx_buf[index].buf_p)) {
> -				txq->tx_buf[index].type = FEC_TXBUF_T_SKB;
> -				goto tx_buf_done;
> -			}
> +			xdpf = txq->tx_buf[index].buf_p;
> +			dma_unmap_single(&fep->pdev->dev,
> +					 fec32_to_cpu(bdp->cbd_bufaddr),
> +					 frame_len,  DMA_TO_DEVICE);
> +			xdp_return_frame_rx_napi(xdpf);
> +			break;
> +		case FEC_TXBUF_T_XDP_TX:
> +			if (unlikely(!budget))
> +				goto out;
>
> -			frame_len = fec16_to_cpu(bdp->cbd_datlen);
> +			page = txq->tx_buf[index].buf_p;
> +			/* The dma_sync_size = 0 as XDP_TX has already synced
> +			 * DMA for_device
> +			 */
> +			page_pool_put_page(pp_page_to_nmdesc(page)->pp, page,
> +					   0, true);
> +			break;
> +		default:
> +			break;
>  		}
>
>  		/* Check for errors. */
> @@ -1572,11 +1593,7 @@ fec_enet_tx_queue(struct net_device *ndev, u16 queue_id, int budget)
>  				ndev->stats.tx_carrier_errors++;
>  		} else {
>  			ndev->stats.tx_packets++;
> -
> -			if (txq->tx_buf[index].type == FEC_TXBUF_T_SKB)
> -				ndev->stats.tx_bytes += skb->len;
> -			else
> -				ndev->stats.tx_bytes += frame_len;
> +			ndev->stats.tx_bytes += frame_len;
>  		}
>
>  		/* Deferred means some collisions occurred during transmit,
> @@ -1585,35 +1602,12 @@ fec_enet_tx_queue(struct net_device *ndev, u16 queue_id, int budget)
>  		if (status & BD_ENET_TX_DEF)
>  			ndev->stats.collisions++;
>
> -		if (txq->tx_buf[index].type == FEC_TXBUF_T_SKB) {
> -			/* NOTE: SKBTX_IN_PROGRESS being set does not imply it's we who
> -			 * are to time stamp the packet, so we still need to check time
> -			 * stamping enabled flag.
> -			 */
> -			if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS &&
> -				     fep->hwts_tx_en) && fep->bufdesc_ex) {
> -				struct skb_shared_hwtstamps shhwtstamps;
> -				struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
> -
> -				fec_enet_hwtstamp(fep, fec32_to_cpu(ebdp->ts), &shhwtstamps);
> -				skb_tstamp_tx(skb, &shhwtstamps);
> -			}
> -
> -			/* Free the sk buffer associated with this last transmit */
> -			napi_consume_skb(skb, budget);
> -		} else if (txq->tx_buf[index].type == FEC_TXBUF_T_XDP_NDO) {
> -			xdp_return_frame_rx_napi(xdpf);
> -		} else { /* recycle pages of XDP_TX frames */
> -			/* The dma_sync_size = 0 as XDP_TX has already synced DMA for_device */
> -			page_pool_put_page(pp_page_to_nmdesc(page)->pp, page,
> -					   0, true);
> -		}
> -
>  		txq->tx_buf[index].buf_p = NULL;
>  		/* restore default tx buffer type: FEC_TXBUF_T_SKB */
>  		txq->tx_buf[index].type = FEC_TXBUF_T_SKB;
>
>  tx_buf_done:
> +		bdp->cbd_bufaddr = cpu_to_fec32(0);
>  		/* Make sure the update to bdp and tx_buf are performed
>  		 * before dirty_tx
>  		 */
> @@ -1632,6 +1626,8 @@ fec_enet_tx_queue(struct net_device *ndev, u16 queue_id, int budget)
>  		}
>  	}
>
> +out:
> +
>  	/* ERR006358: Keep the transmitter going */
>  	if (bdp != txq->bd.cur &&
>  	    readl(txq->bd.reg_desc_active) == 0)
> @@ -3413,6 +3409,7 @@ static void fec_enet_free_buffers(struct net_device *ndev)
>  	unsigned int i;
>  	struct fec_enet_priv_tx_q *txq;
>  	struct fec_enet_priv_rx_q *rxq;
> +	struct page *page;
>  	unsigned int q;
>
>  	for (q = 0; q < fep->num_rx_queues; q++) {
> @@ -3436,20 +3433,20 @@ static void fec_enet_free_buffers(struct net_device *ndev)
>  			kfree(txq->tx_bounce[i]);
>  			txq->tx_bounce[i] = NULL;
>
> -			if (!txq->tx_buf[i].buf_p) {
> -				txq->tx_buf[i].type = FEC_TXBUF_T_SKB;
> -				continue;
> -			}
> -
> -			if (txq->tx_buf[i].type == FEC_TXBUF_T_SKB) {
> +			switch (txq->tx_buf[i].type) {
> +			case FEC_TXBUF_T_SKB:
>  				dev_kfree_skb(txq->tx_buf[i].buf_p);
> -			} else if (txq->tx_buf[i].type == FEC_TXBUF_T_XDP_NDO) {
> +				break;
> +			case FEC_TXBUF_T_XDP_NDO:
>  				xdp_return_frame(txq->tx_buf[i].buf_p);
> -			} else {
> -				struct page *page = txq->tx_buf[i].buf_p;
> -
> +				break;
> +			case FEC_TXBUF_T_XDP_TX:
> +				page = txq->tx_buf[i].buf_p;
>  				page_pool_put_page(pp_page_to_nmdesc(page)->pp,
>  						   page, 0, false);
> +				break;
> +			default:
> +				break;
>  			}
>
>  			txq->tx_buf[i].buf_p = NULL;
> --
> 2.34.1
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ