lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Sun, 12 Apr 2015 14:52:21 +0300
From:	Ido Shamay <idos@....mellanox.co.il>
To:	Amir Vadai <amirv@...lanox.com>,
	"David S. Miller" <davem@...emloft.net>
CC:	netdev@...r.kernel.org, Yevgeny Petrilin <yevgenyp@...lanox.com>,
	Saeed Mahameed <saeedm@...lanox.com>,
	Or Gerlitz <ogerlitz@...lanox.com>,
	Achiad Shochat <achiad@...lanox.com>,
	Ido Shamay <idos@...lanox.com>
Subject: Re: [PATCH net-next 09/11] net/mlx5: Ethernet Datapath files

On 4/8/2015 5:51 PM, Amir Vadai wrote:
> Signed-off-by: Amir Vadai<amirv@...lanox.com>
> ---
>   drivers/net/ethernet/mellanox/mlx5/core/en_rx.c   | 249 +++++++++++++++
>   drivers/net/ethernet/mellanox/mlx5/core/en_tx.c   | 350 ++++++++++++++++++++++
>   drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c | 107 +++++++
>   3 files changed, 706 insertions(+)
>   create mode 100644 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
>   create mode 100644 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
>   create mode 100644 drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c
>
> diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
> new file mode 100644
> index 0000000..e567046
> --- /dev/null
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
> @@ -0,0 +1,249 @@
> +/*
> + * Copyright (c) 2015, Mellanox Technologies. All rights reserved.
> + *
> + * This software is available to you under a choice of one of two
> + * licenses.  You may choose to be licensed under the terms of the GNU
> + * General Public License (GPL) Version 2, available from the file
> + * COPYING in the main directory of this source tree, or the
> + * OpenIB.org BSD license below:
> + *
> + *     Redistribution and use in source and binary forms, with or
> + *     without modification, are permitted provided that the following
> + *     conditions are met:
> + *
> + *      - Redistributions of source code must retain the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer.
> + *
> + *      - Redistributions in binary form must reproduce the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer in the documentation and/or other materials
> + *        provided with the distribution.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
> + * SOFTWARE.
> + */
> +
> +#include <linux/ip.h>
> +#include <linux/ipv6.h>
> +#include <linux/tcp.h>
> +#include "en.h"
> +
> +static inline int mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq,
> +				     struct mlx5e_rx_wqe *wqe, u16 ix)
> +{
> +	struct sk_buff *skb;
> +	dma_addr_t dma_addr;
> +
> +	skb = netdev_alloc_skb(rq->netdev, rq->wqe_sz);
> +	if (unlikely(!skb))
> +		return -ENOMEM;
> +
> +	skb_reserve(skb, MLX5E_NET_IP_ALIGN);
> +
> +	dma_addr = dma_map_single(rq->pdev,
> +				  /* hw start padding */
> +				  skb->data - MLX5E_NET_IP_ALIGN,
> +				  /* hw   end padding */
> +				  skb_end_offset(skb),
> +				  DMA_FROM_DEVICE);
skb_end_offset  depends on NET_SKBUFF_DATA_USES_OFFSET, and may be a 
pointer.
Can use rq->wqe_sz instead.

> +
> +	if (unlikely(dma_mapping_error(rq->pdev, dma_addr)))
> +		goto err_free_skb;
> +
> +	*((dma_addr_t *)skb->cb) = dma_addr;
> +	wqe->data.addr = cpu_to_be64(dma_addr + MLX5E_NET_IP_ALIGN);
> +
> +	rq->skb[ix] = skb;
> +
> +	return 0;
> +
> +err_free_skb:
> +	dev_kfree_skb(skb);
> +
> +	return -ENOMEM;
> +}
> +
> +bool mlx5e_post_rx_wqes(struct mlx5e_rq *rq)
> +{
> +	struct mlx5_wq_ll *wq = &rq->wq;
> +
> +	if (unlikely(!test_bit(MLX5E_RQ_STATE_POST_WQES_ENABLE, &rq->state)))
> +		return false;
> +
> +	while (!mlx5_wq_ll_is_full(wq)) {
> +		struct mlx5e_rx_wqe *wqe = mlx5_wq_ll_get_wqe(wq, wq->head);
> +
> +		if (unlikely(mlx5e_alloc_rx_wqe(rq, wqe, wq->head)))
> +			break;
> +
> +		mlx5_wq_ll_push(wq, be16_to_cpu(wqe->next.next_wqe_index));
> +	}
> +
> +	/* ensure wqes are visible to device before updating doorbell record */
> +	wmb();
> +
> +	mlx5_wq_ll_update_db_record(wq);
> +
> +	return !mlx5_wq_ll_is_full(wq);
> +}
> +
> +static void mlx5e_lro_update_hdr(struct sk_buff *skb, struct mlx5_cqe64 *cqe)
> +{
> +	struct ethhdr	*eth	= (struct ethhdr *)(skb->data);
> +	struct iphdr	*ipv4	= (struct iphdr *)(skb->data + ETH_HLEN);
> +	struct ipv6hdr	*ipv6	= (struct ipv6hdr *)(skb->data + ETH_HLEN);
> +	struct tcphdr	*tcp;
> +
> +	u8 l4_hdr_type = get_cqe_l4_hdr_type(cqe);
> +	int tcp_ack = ((CQE_L4_HDR_TYPE_TCP_ACK_NO_DATA  == l4_hdr_type) ||
> +		       (CQE_L4_HDR_TYPE_TCP_ACK_AND_DATA == l4_hdr_type));
> +
> +	u16 tot_len = be32_to_cpu(cqe->byte_cnt) - ETH_HLEN;
> +
> +	if (eth->h_proto == htons(ETH_P_IP)) {
> +		tcp = (struct tcphdr *)(skb->data + ETH_HLEN +
> +					sizeof(struct iphdr));
> +		ipv6 = NULL;
> +	} else {
> +		tcp = (struct tcphdr *)(skb->data + ETH_HLEN +
> +					sizeof(struct ipv6hdr));
> +		ipv4 = NULL;
> +	}
> +
> +	if (get_cqe_lro_tcppsh(cqe))
> +		tcp->psh                = 1;
> +
> +	if (tcp_ack) {
> +		tcp->ack                = 1;
> +		tcp->ack_seq            = cqe->lro_ack_seq_num;
> +		tcp->window             = cqe->lro_tcp_win;
> +	}
> +
> +	if (ipv4) {
> +		ipv4->ttl               = cqe->lro_min_ttl;
> +		ipv4->tot_len           = cpu_to_be16(tot_len);
> +		ipv4->check             = 0;
> +		ipv4->check             = ip_fast_csum((unsigned char *)ipv4,
> +						       ipv4->ihl);
> +	} else {
> +		ipv6->hop_limit         = cqe->lro_min_ttl;
> +		ipv6->payload_len       = cpu_to_be16(tot_len -
> +						      sizeof(struct ipv6hdr));
> +	}
> +}
> +
> +static inline void mlx5e_skb_set_hash(struct mlx5_cqe64 *cqe,
> +				      struct sk_buff *skb)
> +{
> +	u8 cht = cqe->rss_hash_type;
> +	int ht = (cht & CQE_RSS_HTYPE_L4) ? PKT_HASH_TYPE_L4 :
> +		 (cht & CQE_RSS_HTYPE_IP) ? PKT_HASH_TYPE_L3 :
> +					    PKT_HASH_TYPE_NONE;
> +	skb_set_hash(skb, be32_to_cpu(cqe->rss_hash_result), ht);
> +}
> +
> +static inline void mlx5e_build_rx_skb(struct mlx5_cqe64 *cqe,
> +				      struct mlx5e_rq *rq,
> +				      struct sk_buff *skb)
> +{
> +	struct net_device *netdev = rq->netdev;
> +	u32 cqe_bcnt = be32_to_cpu(cqe->byte_cnt);
> +	int lro_num_seg;
> +
> +	skb_put(skb, cqe_bcnt);
> +
> +	lro_num_seg = be32_to_cpu(cqe->srqn) >> 24;
> +	if (lro_num_seg > 1) {
> +		mlx5e_lro_update_hdr(skb, cqe);
> +		skb_shinfo(skb)->gso_size = MLX5E_PARAMS_DEFAULT_LRO_WQE_SZ;
> +		rq->stats.lro_packets++;
> +		rq->stats.lro_bytes += cqe_bcnt;
> +	}
> +
> +	if (likely(netdev->features & NETIF_F_RXCSUM) &&
> +	    (cqe->hds_ip_ext & CQE_L2_OK) &&
> +	    (cqe->hds_ip_ext & CQE_L3_OK) &&
> +	    (cqe->hds_ip_ext & CQE_L4_OK)) {
> +		skb->ip_summed = CHECKSUM_UNNECESSARY;
> +	} else {
> +		skb->ip_summed = CHECKSUM_NONE;
> +		rq->stats.csum_none++;
> +	}
> +
> +	skb->protocol = eth_type_trans(skb, netdev);
> +
> +	skb_record_rx_queue(skb, rq->ix);
> +
> +	if (likely(netdev->features & NETIF_F_RXHASH))
> +		mlx5e_skb_set_hash(cqe, skb);
> +
> +	if (cqe_has_vlan(cqe))
> +		__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q),
> +				       be16_to_cpu(cqe->vlan_info));
> +}
> +
> +bool mlx5e_poll_rx_cq(struct mlx5e_cq *cq, int budget)
> +{
> +	struct mlx5e_rq *rq = cq->sqrq;
> +	int i;
> +
> +	/* avoid accessing cq (dma coherent memory) if not needed */
> +	if (!test_and_clear_bit(MLX5E_CQ_HAS_CQES, &cq->flags))
> +		return false;
> +
> +	for (i = 0; i < budget; i++) {
> +		struct mlx5e_rx_wqe *wqe;
> +		struct mlx5_cqe64 *cqe;
> +		struct sk_buff *skb;
> +		__be16 wqe_counter_be;
> +		u16 wqe_counter;
> +
> +		cqe = mlx5e_get_cqe(cq);
> +		if (!cqe)
> +			break;
> +
> +		wqe_counter_be = cqe->wqe_counter;
> +		wqe_counter    = be16_to_cpu(wqe_counter_be);
> +		wqe            = mlx5_wq_ll_get_wqe(&rq->wq, wqe_counter);
> +		skb            = rq->skb[wqe_counter];
> +		rq->skb[wqe_counter] = NULL;
> +
> +		dma_unmap_single(rq->pdev,
> +				 *((dma_addr_t *)skb->cb),
> +				 skb_end_offset(skb),
> +				 DMA_FROM_DEVICE);
> +
> +		if (unlikely((cqe->op_own >> 4) != MLX5_CQE_RESP_SEND)) {
> +			rq->stats.wqe_err++;
> +			dev_kfree_skb(skb);
> +			goto wq_ll_pop;
> +		}
> +
> +		mlx5e_build_rx_skb(cqe, rq, skb);
> +		rq->stats.packets++;
> +		napi_gro_receive(cq->napi, skb);
> +
> +wq_ll_pop:
> +		mlx5_wq_ll_pop(&rq->wq, wqe_counter_be,
> +			       &wqe->next.next_wqe_index);
> +	}
> +
> +	mlx5_cqwq_update_db_record(&cq->wq);
> +
> +	/* ensure cq space is freed before enabling more cqes */
> +	wmb();
> +
> +	if (i == budget) {
> +		set_bit(MLX5E_CQ_HAS_CQES, &cq->flags);
> +		return true;
> +	}
> +
> +	return false;
> +}
> diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
> new file mode 100644
> index 0000000..1bd2027
> --- /dev/null
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
> @@ -0,0 +1,350 @@
> +/*
> + * Copyright (c) 2015, Mellanox Technologies. All rights reserved.
> + *
> + * This software is available to you under a choice of one of two
> + * licenses.  You may choose to be licensed under the terms of the GNU
> + * General Public License (GPL) Version 2, available from the file
> + * COPYING in the main directory of this source tree, or the
> + * OpenIB.org BSD license below:
> + *
> + *     Redistribution and use in source and binary forms, with or
> + *     without modification, are permitted provided that the following
> + *     conditions are met:
> + *
> + *      - Redistributions of source code must retain the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer.
> + *
> + *      - Redistributions in binary form must reproduce the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer in the documentation and/or other materials
> + *        provided with the distribution.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
> + * SOFTWARE.
> + */
> +
> +#include <linux/tcp.h>
> +#include <linux/if_vlan.h>
> +#include "en.h"
> +
> +static void mlx5e_dma_pop_last_pushed(struct mlx5e_sq *sq, dma_addr_t *addr,
> +				      u32 *size)
> +{
> +	sq->dma_fifo_pc--;
> +	*addr = sq->dma_fifo[sq->dma_fifo_pc & sq->dma_fifo_mask].addr;
> +	*size = sq->dma_fifo[sq->dma_fifo_pc & sq->dma_fifo_mask].size;
> +}
> +
> +static void mlx5e_dma_unmap_wqe_err(struct mlx5e_sq *sq, struct sk_buff *skb)
> +{
> +	dma_addr_t addr;
> +	u32 size;
> +	int i;
> +
> +	for (i = 0; i < MLX5E_TX_SKB_CB(skb)->num_dma; i++) {
> +		mlx5e_dma_pop_last_pushed(sq, &addr, &size);
> +		dma_unmap_single(sq->pdev, addr, size, DMA_TO_DEVICE);
> +	}
> +}
> +
> +static inline void mlx5e_dma_push(struct mlx5e_sq *sq, dma_addr_t addr,
> +				  u32 size)
> +{
> +	sq->dma_fifo[sq->dma_fifo_pc & sq->dma_fifo_mask].addr = addr;
> +	sq->dma_fifo[sq->dma_fifo_pc & sq->dma_fifo_mask].size = size;
> +	sq->dma_fifo_pc++;
> +}
> +
> +static inline void mlx5e_dma_get(struct mlx5e_sq *sq, u32 i, dma_addr_t *addr,
> +				 u32 *size)
> +{
> +	*addr = sq->dma_fifo[i & sq->dma_fifo_mask].addr;
> +	*size = sq->dma_fifo[i & sq->dma_fifo_mask].size;
> +}
> +
> +u16 mlx5e_select_queue(struct net_device *dev, struct sk_buff *skb,
> +		       void *accel_priv, select_queue_fallback_t fallback)
> +{
> +	struct mlx5e_priv *priv = netdev_priv(dev);
> +	int channel_ix = fallback(dev, skb);
> +	int up = skb_vlan_tag_present(skb)        ?
> +		 skb->vlan_tci >> VLAN_PRIO_SHIFT :
> +		 priv->default_vlan_prio;
> +	int tc = netdev_get_prio_tc_map(dev, up);
> +
> +	return (tc << priv->order_base_2_num_channels) | channel_ix;
> +}
> +
> +static inline u16 mlx5e_get_inline_hdr_size(struct mlx5e_sq *sq,
> +					    struct sk_buff *skb)
> +{
> +#define MLX5E_MIN_INLINE 16 /* eth header with vlan (w/o next ethertype) */
> +#define MLX5E_MAX_INLINE (128 - sizeof(struct mlx5e_tx_wqe) +\
> +			  2/*sizeof(eseg->inline_hdr_start)*/)
> +
> +	if (!skb_shinfo(skb)->nr_frags &&
> +	    (skb_headlen(skb) <= MLX5E_MAX_INLINE))
> +		return skb_headlen(skb);
> +
> +	return MLX5E_MIN_INLINE;
> +}
> +
> +static inline void mlx5e_insert_vlan(void *start, struct sk_buff *skb, u16 ihs)
> +{
> +	struct vlan_ethhdr *vhdr = (struct vlan_ethhdr *)start;
> +	int cpy1_sz = 2 * ETH_ALEN;
> +	int cpy2_sz = ihs - cpy1_sz - VLAN_HLEN;
> +
> +	skb_copy_from_linear_data(skb, vhdr, cpy1_sz);
> +	skb_pull_inline(skb, cpy1_sz);
> +	vhdr->h_vlan_proto = skb->vlan_proto;
> +	vhdr->h_vlan_TCI = cpu_to_be16(skb_vlan_tag_get(skb));
> +	skb_copy_from_linear_data(skb, &vhdr->h_vlan_encapsulated_proto,
> +				  cpy2_sz);
> +	skb_pull_inline(skb, cpy2_sz);
> +}
> +
> +static netdev_tx_t mlx5e_sq_xmit(struct mlx5e_sq *sq, struct sk_buff *skb)
> +{
> +	struct mlx5_wq_cyc       *wq   = &sq->wq;
> +
> +	u16 pi = sq->pc & wq->sz_m1;
> +	struct mlx5e_tx_wqe      *wqe  = mlx5_wq_cyc_get_wqe(wq, pi);
> +
> +	struct mlx5_wqe_ctrl_seg *cseg = &wqe->ctrl;
> +	struct mlx5_wqe_eth_seg  *eseg = &wqe->eth;
> +	struct mlx5_wqe_data_seg *dseg;
> +
> +	u8  opcode = MLX5_OPCODE_SEND;
> +	dma_addr_t dma_addr = 0;
> +	u16 headlen;
> +	u16 ds_cnt;
> +	u16 ihs;
> +	int i;
> +
> +	memset(wqe, 0, sizeof(*wqe));
> +
> +	if (likely(skb->ip_summed == CHECKSUM_PARTIAL))
> +		eseg->cs_flags	= MLX5_ETH_WQE_L3_CSUM | MLX5_ETH_WQE_L4_CSUM;
> +	else
> +		sq->stats.csum_offload_none++;
> +
> +	if (skb_is_gso(skb)) {
> +		u32 payload_len;
> +		int num_pkts;
> +
> +		eseg->mss    = cpu_to_be16(skb_shinfo(skb)->gso_size);
> +		opcode       = MLX5_OPCODE_LSO;
> +		ihs          = skb_transport_offset(skb) + tcp_hdrlen(skb);
> +		payload_len  = skb->len - ihs;
> +		num_pkts     =    (payload_len / skb_shinfo(skb)->gso_size) +
> +				!!(payload_len % skb_shinfo(skb)->gso_size);
> +		MLX5E_TX_SKB_CB(skb)->num_bytes = skb->len +
> +						  (num_pkts - 1) * ihs;
> +		sq->stats.tso_packets++;
> +		sq->stats.tso_bytes += payload_len;
> +	} else {
> +		ihs             = mlx5e_get_inline_hdr_size(sq, skb);
> +		MLX5E_TX_SKB_CB(skb)->num_bytes = max_t(unsigned int, skb->len,
> +							ETH_ZLEN);
> +	}
> +
> +	if (skb_vlan_tag_present(skb)) {
> +		mlx5e_insert_vlan(eseg->inline_hdr_start, skb, ihs);
> +	} else {
> +		skb_copy_from_linear_data(skb, eseg->inline_hdr_start, ihs);
> +		skb_pull_inline(skb, ihs);
> +	}
> +
> +	eseg->inline_hdr_sz	= cpu_to_be16(ihs);
> +
> +	ds_cnt  = sizeof(*wqe) / MLX5_SEND_WQE_DS;
> +	ds_cnt += DIV_ROUND_UP(ihs - sizeof(eseg->inline_hdr_start),
> +			       MLX5_SEND_WQE_DS);
> +	dseg    = (struct mlx5_wqe_data_seg *)cseg + ds_cnt;
> +
> +	MLX5E_TX_SKB_CB(skb)->num_dma = 0;
> +
> +	headlen = skb_headlen(skb);
> +	if (headlen) {
> +		dma_addr = dma_map_single(sq->pdev, skb->data, headlen,
> +					  DMA_TO_DEVICE);
> +		if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
> +			goto dma_unmap_wqe_err;
> +
> +		dseg->addr       = cpu_to_be64(dma_addr);
> +		dseg->lkey       = sq->mkey_be;
> +		dseg->byte_count = cpu_to_be32(headlen);
> +
> +		mlx5e_dma_push(sq, dma_addr, headlen);
> +		MLX5E_TX_SKB_CB(skb)->num_dma++;
> +
> +		dseg++;
> +	}
> +
> +	for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
> +		struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i];
> +		int fsz = skb_frag_size(frag);
> +
> +		dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
> +					    DMA_TO_DEVICE);
> +		if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
> +			goto dma_unmap_wqe_err;
> +
> +		dseg->addr       = cpu_to_be64(dma_addr);
> +		dseg->lkey       = sq->mkey_be;
> +		dseg->byte_count = cpu_to_be32(fsz);
> +
> +		mlx5e_dma_push(sq, dma_addr, fsz);
> +		MLX5E_TX_SKB_CB(skb)->num_dma++;
> +
> +		dseg++;
> +	}
> +
> +	ds_cnt += MLX5E_TX_SKB_CB(skb)->num_dma;
> +
> +	cseg->opmod_idx_opcode	= cpu_to_be32((sq->pc << 8) | opcode);
> +	cseg->qpn_ds		= cpu_to_be32((sq->sqn << 8) | ds_cnt);
> +	cseg->fm_ce_se		= MLX5_WQE_CTRL_CQ_UPDATE;
> +
> +	sq->skb[pi] = skb;
> +
> +	MLX5E_TX_SKB_CB(skb)->num_wqebbs = DIV_ROUND_UP(ds_cnt,
> +							MLX5_SEND_WQEBB_NUM_DS);
> +	sq->pc += MLX5E_TX_SKB_CB(skb)->num_wqebbs;
> +
> +	netdev_tx_sent_queue(sq->txq, MLX5E_TX_SKB_CB(skb)->num_bytes);
> +
> +	if (unlikely(!mlx5e_sq_has_room_for(sq, MLX5_SEND_WQE_MAX_WQEBBS))) {
> +		netif_tx_stop_queue(sq->txq);
> +		sq->stats.stopped++;
> +	}
> +
> +	if (!skb->xmit_more || netif_xmit_stopped(sq->txq))
> +		mlx5e_tx_notify_hw(sq, wqe);
> +
> +	sq->stats.packets++;
> +	return NETDEV_TX_OK;
> +
> +dma_unmap_wqe_err:
> +	sq->stats.dropped++;
> +	mlx5e_dma_unmap_wqe_err(sq, skb);
> +
> +	dev_kfree_skb_any(skb);
> +
> +	return NETDEV_TX_OK;
> +}
> +
> +netdev_tx_t mlx5e_xmit(struct sk_buff *skb, struct net_device *dev)
> +{
> +	struct mlx5e_priv *priv = netdev_priv(dev);
> +	int ix = skb->queue_mapping;
> +	int tc = 0;
> +	struct mlx5e_channel *c = priv->channel[ix];
> +	struct mlx5e_sq *sq = &c->sq[tc];
> +
> +	return mlx5e_sq_xmit(sq, skb);
> +}
> +
> +netdev_tx_t mlx5e_xmit_multi_tc(struct sk_buff *skb, struct net_device *dev)
> +{
> +	struct mlx5e_priv *priv = netdev_priv(dev);
> +	int ix = skb->queue_mapping & priv->queue_mapping_channel_mask;
> +	int tc = skb->queue_mapping >> priv->order_base_2_num_channels;
> +	struct mlx5e_channel *c = priv->channel[ix];
> +	struct mlx5e_sq *sq = &c->sq[tc];
> +
> +	return mlx5e_sq_xmit(sq, skb);
> +}
> +
> +bool mlx5e_poll_tx_cq(struct mlx5e_cq *cq)
> +{
> +	struct mlx5e_sq *sq;
> +	u32 dma_fifo_cc;
> +	u32 nbytes;
> +	u16 npkts;
> +	u16 sqcc;
> +	int i;
> +
> +	/* avoid accessing cq (dma coherent memory) if not needed */
> +	if (!test_and_clear_bit(MLX5E_CQ_HAS_CQES, &cq->flags))
> +		return false;
> +
> +	sq = cq->sqrq;
> +
> +	npkts = 0;
> +	nbytes = 0;
> +
> +	/* sq->cc must be updated only after mlx5_cqwq_update_db_record(),
> +	 * otherwise a cq overrun may occur */
> +	sqcc = sq->cc;
> +
> +	/* avoid dirtying sq cache line every cqe */
> +	dma_fifo_cc = sq->dma_fifo_cc;
> +
> +	for (i = 0; i < MLX5E_TX_CQ_POLL_BUDGET; i++) {
> +		struct mlx5_cqe64 *cqe;
> +		struct sk_buff *skb;
> +		u16 ci;
> +		int j;
> +
> +		cqe = mlx5e_get_cqe(cq);
> +		if (!cqe)
> +			break;
> +
> +		ci = sqcc & sq->wq.sz_m1;
> +		skb = sq->skb[ci];
> +
> +		if (unlikely(!skb)) { /* nop */
> +			sq->stats.nop++;
> +			sqcc++;
> +			goto free_skb;
> +		}
> +
> +		for (j = 0; j < MLX5E_TX_SKB_CB(skb)->num_dma; j++) {
> +			dma_addr_t addr;
> +			u32 size;
> +
> +			mlx5e_dma_get(sq, dma_fifo_cc, &addr, &size);
> +			dma_fifo_cc++;
> +			dma_unmap_single(sq->pdev, addr, size, DMA_TO_DEVICE);
> +		}
> +
> +		npkts++;
> +		nbytes += MLX5E_TX_SKB_CB(skb)->num_bytes;
> +		sqcc += MLX5E_TX_SKB_CB(skb)->num_wqebbs;
> +
> +free_skb:
> +		dev_kfree_skb(skb);
> +	}
> +
> +	mlx5_cqwq_update_db_record(&cq->wq);
> +
> +	/* ensure cq space is freed before enabling more cqes */
> +	wmb();
> +
> +	sq->dma_fifo_cc = dma_fifo_cc;
> +	sq->cc = sqcc;
> +
> +	netdev_tx_completed_queue(sq->txq, npkts, nbytes);
> +
> +	if (netif_tx_queue_stopped(sq->txq) &&
> +	    mlx5e_sq_has_room_for(sq, MLX5_SEND_WQE_MAX_WQEBBS) &&
> +	    likely(test_bit(MLX5E_SQ_STATE_WAKE_TXQ_ENABLE, &sq->state))) {
> +				netif_tx_wake_queue(sq->txq);
> +				sq->stats.wake++;
> +	}
> +	if (i == MLX5E_TX_CQ_POLL_BUDGET) {
> +		set_bit(MLX5E_CQ_HAS_CQES, &cq->flags);
> +		return true;
> +	}
> +
> +	return false;
> +}
> diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c
> new file mode 100644
> index 0000000..088bc42
> --- /dev/null
> +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c
> @@ -0,0 +1,107 @@
> +/*
> + * Copyright (c) 2015, Mellanox Technologies. All rights reserved.
> + *
> + * This software is available to you under a choice of one of two
> + * licenses.  You may choose to be licensed under the terms of the GNU
> + * General Public License (GPL) Version 2, available from the file
> + * COPYING in the main directory of this source tree, or the
> + * OpenIB.org BSD license below:
> + *
> + *     Redistribution and use in source and binary forms, with or
> + *     without modification, are permitted provided that the following
> + *     conditions are met:
> + *
> + *      - Redistributions of source code must retain the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer.
> + *
> + *      - Redistributions in binary form must reproduce the above
> + *        copyright notice, this list of conditions and the following
> + *        disclaimer in the documentation and/or other materials
> + *        provided with the distribution.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
> + * SOFTWARE.
> + */
> +
> +#include "en.h"
> +
> +struct mlx5_cqe64 *mlx5e_get_cqe(struct mlx5e_cq *cq)
> +{
> +	struct mlx5_cqwq *wq = &cq->wq;
> +	u32 ci = mlx5_cqwq_get_ci(wq);
> +	struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci);
> +	int cqe_ownership_bit = cqe->op_own & MLX5_CQE_OWNER_MASK;
> +	int sw_ownership_val = mlx5_cqwq_get_wrap_cnt(wq) & 1;
> +
> +	if (cqe_ownership_bit != sw_ownership_val)
> +		return NULL;
> +
> +	mlx5_cqwq_pop(wq);
> +
> +	/* ensure cqe content is read after cqe ownership bit */
> +	rmb();
> +
> +	return cqe;
> +}
> +
> +int mlx5e_napi_poll(struct napi_struct *napi, int budget)
> +{
> +	struct mlx5e_channel *c = container_of(napi, struct mlx5e_channel,
> +					       napi);
> +	bool busy = false;
> +	int i;
> +
> +	clear_bit(MLX5E_CHANNEL_NAPI_SCHED, &c->flags);
> +
> +	for (i = 0; i < c->num_tc; i++)
> +		busy |= mlx5e_poll_tx_cq(&c->sq[i].cq);
> +
> +	busy |= mlx5e_poll_rx_cq(&c->rq.cq, budget);
> +
> +	busy |= mlx5e_post_rx_wqes(c->rq.cq.sqrq);
> +
> +	if (busy)
> +		return budget;
> +
> +	napi_complete(napi);
> +
> +	/* avoid losing completion event during/after polling cqs */
> +	if (test_bit(MLX5E_CHANNEL_NAPI_SCHED, &c->flags)) {
> +		napi_schedule(napi);
> +		return 0;
> +	}
> +
> +	for (i = 0; i < c->num_tc; i++)
> +		mlx5e_cq_arm(&c->sq[i].cq);
> +	mlx5e_cq_arm(&c->rq.cq);
> +
> +	return 0;
> +}
> +
> +void mlx5e_completion_event(struct mlx5_core_cq *mcq)
> +{
> +	struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq);
> +
> +	set_bit(MLX5E_CQ_HAS_CQES, &cq->flags);
> +	set_bit(MLX5E_CHANNEL_NAPI_SCHED, &cq->channel->flags);
> +	barrier();
> +	napi_schedule(cq->napi);
> +}
> +
> +void mlx5e_cq_error_event(struct mlx5_core_cq *mcq, enum mlx5_event event)
> +{
> +	struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq);
> +	struct mlx5e_channel *c = cq->channel;
> +	struct mlx5e_priv *priv = c->priv;
> +	struct net_device *netdev = priv->netdev;
> +
> +	netdev_err(netdev, "%s: cqn=0x%.6x event=0x%.2x\n",
> +		   __func__, mcq->cqn, event);
> +}

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ