[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <82588c26-465e-2caf-8f35-10b529faab36@redhat.com>
Date: Wed, 16 Jun 2021 15:35:06 +0800
From: Jason Wang <jasowang@...hat.com>
To: Xuan Zhuo <xuanzhuo@...ux.alibaba.com>, netdev@...r.kernel.org
Cc: "David S. Miller" <davem@...emloft.net>,
Jakub Kicinski <kuba@...nel.org>,
"Michael S. Tsirkin" <mst@...hat.com>,
Björn Töpel <bjorn@...nel.org>,
Magnus Karlsson <magnus.karlsson@...el.com>,
Jonathan Lemon <jonathan.lemon@...il.com>,
Alexei Starovoitov <ast@...nel.org>,
Daniel Borkmann <daniel@...earbox.net>,
Jesper Dangaard Brouer <hawk@...nel.org>,
John Fastabend <john.fastabend@...il.com>,
Andrii Nakryiko <andrii@...nel.org>,
Martin KaFai Lau <kafai@...com>,
Song Liu <songliubraving@...com>, Yonghong Song <yhs@...com>,
KP Singh <kpsingh@...nel.org>,
virtualization@...ts.linux-foundation.org, bpf@...r.kernel.org,
"dust . li" <dust.li@...ux.alibaba.com>
Subject: Re: [PATCH net-next v5 11/15] virtio-net: move to virtio_net.h
在 2021/6/10 下午4:22, Xuan Zhuo 写道:
> Move some structure definitions and inline functions into the
> virtio_net.h file.
>
> Signed-off-by: Xuan Zhuo <xuanzhuo@...ux.alibaba.com>
Acked-by: Jason Wang <jasowang@...hat.com>
> ---
> drivers/net/virtio/virtio_net.c | 225 +------------------------------
> drivers/net/virtio/virtio_net.h | 230 ++++++++++++++++++++++++++++++++
> 2 files changed, 232 insertions(+), 223 deletions(-)
> create mode 100644 drivers/net/virtio/virtio_net.h
>
> diff --git a/drivers/net/virtio/virtio_net.c b/drivers/net/virtio/virtio_net.c
> index 953739860563..395ec1f18331 100644
> --- a/drivers/net/virtio/virtio_net.c
> +++ b/drivers/net/virtio/virtio_net.c
> @@ -4,24 +4,8 @@
> * Copyright 2007 Rusty Russell <rusty@...tcorp.com.au> IBM Corporation
> */
> //#define DEBUG
> -#include <linux/netdevice.h>
> -#include <linux/etherdevice.h>
> -#include <linux/ethtool.h>
> -#include <linux/module.h>
> -#include <linux/virtio.h>
> -#include <linux/virtio_net.h>
> -#include <linux/bpf.h>
> -#include <linux/bpf_trace.h>
> -#include <linux/scatterlist.h>
> -#include <linux/if_vlan.h>
> -#include <linux/slab.h>
> -#include <linux/cpu.h>
> -#include <linux/average.h>
> -#include <linux/filter.h>
> -#include <linux/kernel.h>
> -#include <net/route.h>
> -#include <net/xdp.h>
> -#include <net/net_failover.h>
> +
> +#include "virtio_net.h"
>
> static int napi_weight = NAPI_POLL_WEIGHT;
> module_param(napi_weight, int, 0444);
> @@ -44,15 +28,6 @@ module_param(napi_tx, bool, 0644);
> #define VIRTIO_XDP_TX BIT(0)
> #define VIRTIO_XDP_REDIR BIT(1)
>
> -#define VIRTIO_XDP_FLAG BIT(0)
> -
> -/* RX packet size EWMA. The average packet size is used to determine the packet
> - * buffer size when refilling RX rings. As the entire RX ring may be refilled
> - * at once, the weight is chosen so that the EWMA will be insensitive to short-
> - * term, transient changes in packet size.
> - */
> -DECLARE_EWMA(pkt_len, 0, 64)
> -
> #define VIRTNET_DRIVER_VERSION "1.0.0"
>
> static const unsigned long guest_offloads[] = {
> @@ -68,35 +43,6 @@ static const unsigned long guest_offloads[] = {
> (1ULL << VIRTIO_NET_F_GUEST_ECN) | \
> (1ULL << VIRTIO_NET_F_GUEST_UFO))
>
> -struct virtnet_stat_desc {
> - char desc[ETH_GSTRING_LEN];
> - size_t offset;
> -};
> -
> -struct virtnet_sq_stats {
> - struct u64_stats_sync syncp;
> - u64 packets;
> - u64 bytes;
> - u64 xdp_tx;
> - u64 xdp_tx_drops;
> - u64 kicks;
> -};
> -
> -struct virtnet_rq_stats {
> - struct u64_stats_sync syncp;
> - u64 packets;
> - u64 bytes;
> - u64 drops;
> - u64 xdp_packets;
> - u64 xdp_tx;
> - u64 xdp_redirects;
> - u64 xdp_drops;
> - u64 kicks;
> -};
> -
> -#define VIRTNET_SQ_STAT(m) offsetof(struct virtnet_sq_stats, m)
> -#define VIRTNET_RQ_STAT(m) offsetof(struct virtnet_rq_stats, m)
> -
> static const struct virtnet_stat_desc virtnet_sq_stats_desc[] = {
> { "packets", VIRTNET_SQ_STAT(packets) },
> { "bytes", VIRTNET_SQ_STAT(bytes) },
> @@ -119,54 +65,6 @@ static const struct virtnet_stat_desc virtnet_rq_stats_desc[] = {
> #define VIRTNET_SQ_STATS_LEN ARRAY_SIZE(virtnet_sq_stats_desc)
> #define VIRTNET_RQ_STATS_LEN ARRAY_SIZE(virtnet_rq_stats_desc)
>
> -/* Internal representation of a send virtqueue */
> -struct send_queue {
> - /* Virtqueue associated with this send _queue */
> - struct virtqueue *vq;
> -
> - /* TX: fragments + linear part + virtio header */
> - struct scatterlist sg[MAX_SKB_FRAGS + 2];
> -
> - /* Name of the send queue: output.$index */
> - char name[40];
> -
> - struct virtnet_sq_stats stats;
> -
> - struct napi_struct napi;
> -};
> -
> -/* Internal representation of a receive virtqueue */
> -struct receive_queue {
> - /* Virtqueue associated with this receive_queue */
> - struct virtqueue *vq;
> -
> - struct napi_struct napi;
> -
> - struct bpf_prog __rcu *xdp_prog;
> -
> - struct virtnet_rq_stats stats;
> -
> - /* Chain pages by the private ptr. */
> - struct page *pages;
> -
> - /* Average packet length for mergeable receive buffers. */
> - struct ewma_pkt_len mrg_avg_pkt_len;
> -
> - /* Page frag for packet buffer allocation. */
> - struct page_frag alloc_frag;
> -
> - /* RX: fragments + linear part + virtio header */
> - struct scatterlist sg[MAX_SKB_FRAGS + 2];
> -
> - /* Min single buffer size for mergeable buffers case. */
> - unsigned int min_buf_len;
> -
> - /* Name of this receive queue: input.$index */
> - char name[40];
> -
> - struct xdp_rxq_info xdp_rxq;
> -};
> -
> /* Control VQ buffers: protected by the rtnl lock */
> struct control_buf {
> struct virtio_net_ctrl_hdr hdr;
> @@ -178,67 +76,6 @@ struct control_buf {
> __virtio64 offloads;
> };
>
> -struct virtnet_info {
> - struct virtio_device *vdev;
> - struct virtqueue *cvq;
> - struct net_device *dev;
> - struct send_queue *sq;
> - struct receive_queue *rq;
> - unsigned int status;
> -
> - /* Max # of queue pairs supported by the device */
> - u16 max_queue_pairs;
> -
> - /* # of queue pairs currently used by the driver */
> - u16 curr_queue_pairs;
> -
> - /* # of XDP queue pairs currently used by the driver */
> - u16 xdp_queue_pairs;
> -
> - /* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
> - bool xdp_enabled;
> -
> - /* I like... big packets and I cannot lie! */
> - bool big_packets;
> -
> - /* Host will merge rx buffers for big packets (shake it! shake it!) */
> - bool mergeable_rx_bufs;
> -
> - /* Has control virtqueue */
> - bool has_cvq;
> -
> - /* Host can handle any s/g split between our header and packet data */
> - bool any_header_sg;
> -
> - /* Packet virtio header size */
> - u8 hdr_len;
> -
> - /* Work struct for refilling if we run low on memory. */
> - struct delayed_work refill;
> -
> - /* Work struct for config space updates */
> - struct work_struct config_work;
> -
> - /* Does the affinity hint is set for virtqueues? */
> - bool affinity_hint_set;
> -
> - /* CPU hotplug instances for online & dead */
> - struct hlist_node node;
> - struct hlist_node node_dead;
> -
> - struct control_buf *ctrl;
> -
> - /* Ethtool settings */
> - u8 duplex;
> - u32 speed;
> -
> - unsigned long guest_offloads;
> - unsigned long guest_offloads_capable;
> -
> - /* failover when STANDBY feature enabled */
> - struct failover *failover;
> -};
> -
> struct padded_vnet_hdr {
> struct virtio_net_hdr_mrg_rxbuf hdr;
> /*
> @@ -249,21 +86,6 @@ struct padded_vnet_hdr {
> char padding[4];
> };
>
> -static bool is_xdp_frame(void *ptr)
> -{
> - return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> -}
> -
> -static void *xdp_to_ptr(struct xdp_frame *ptr)
> -{
> - return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);
> -}
> -
> -static struct xdp_frame *ptr_to_xdp(void *ptr)
> -{
> - return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
> -}
> -
> static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,
> int gfp)
> {
> @@ -280,30 +102,6 @@ static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,
> return buf;
> }
>
> -static void __free_old_xmit(struct send_queue *sq, bool in_napi,
> - struct virtnet_sq_stats *stats)
> -{
> - unsigned int len;
> - void *ptr;
> -
> - while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
> - if (!is_xdp_frame(ptr)) {
> - struct sk_buff *skb = ptr;
> -
> - pr_debug("Sent skb %p\n", skb);
> -
> - stats->bytes += skb->len;
> - napi_consume_skb(skb, in_napi);
> - } else {
> - struct xdp_frame *frame = ptr_to_xdp(ptr);
> -
> - stats->bytes += frame->len;
> - xdp_return_frame(frame);
> - }
> - stats->packets++;
> - }
> -}
> -
> /* Converting between virtqueue no. and kernel tx/rx queue no.
> * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq
> */
> @@ -359,15 +157,6 @@ static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)
> return p;
> }
>
> -static void virtqueue_napi_schedule(struct napi_struct *napi,
> - struct virtqueue *vq)
> -{
> - if (napi_schedule_prep(napi)) {
> - virtqueue_disable_cb(vq);
> - __napi_schedule(napi);
> - }
> -}
> -
> static void virtqueue_napi_complete(struct napi_struct *napi,
> struct virtqueue *vq, int processed)
> {
> @@ -1537,16 +1326,6 @@ static void free_old_xmit(struct send_queue *sq, bool in_napi)
> u64_stats_update_end(&sq->stats.syncp);
> }
>
> -static bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
> -{
> - if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
> - return false;
> - else if (q < vi->curr_queue_pairs)
> - return true;
> - else
> - return false;
> -}
> -
> static void virtnet_poll_cleantx(struct receive_queue *rq)
> {
> struct virtnet_info *vi = rq->vq->vdev->priv;
> diff --git a/drivers/net/virtio/virtio_net.h b/drivers/net/virtio/virtio_net.h
> new file mode 100644
> index 000000000000..931cc81f92fb
> --- /dev/null
> +++ b/drivers/net/virtio/virtio_net.h
> @@ -0,0 +1,230 @@
> +/* SPDX-License-Identifier: GPL-2.0-or-later */
> +
> +#ifndef __VIRTIO_NET_H__
> +#define __VIRTIO_NET_H__
> +#include <linux/netdevice.h>
> +#include <linux/etherdevice.h>
> +#include <linux/ethtool.h>
> +#include <linux/module.h>
> +#include <linux/virtio.h>
> +#include <linux/virtio_net.h>
> +#include <linux/bpf.h>
> +#include <linux/bpf_trace.h>
> +#include <linux/scatterlist.h>
> +#include <linux/if_vlan.h>
> +#include <linux/slab.h>
> +#include <linux/cpu.h>
> +#include <linux/average.h>
> +#include <linux/filter.h>
> +#include <linux/kernel.h>
> +#include <net/route.h>
> +#include <net/xdp.h>
> +#include <net/net_failover.h>
> +#include <net/xdp_sock_drv.h>
> +
> +#define VIRTIO_XDP_FLAG BIT(0)
> +
> +struct virtnet_info {
> + struct virtio_device *vdev;
> + struct virtqueue *cvq;
> + struct net_device *dev;
> + struct send_queue *sq;
> + struct receive_queue *rq;
> + unsigned int status;
> +
> + /* Max # of queue pairs supported by the device */
> + u16 max_queue_pairs;
> +
> + /* # of queue pairs currently used by the driver */
> + u16 curr_queue_pairs;
> +
> + /* # of XDP queue pairs currently used by the driver */
> + u16 xdp_queue_pairs;
> +
> + /* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
> + bool xdp_enabled;
> +
> + /* I like... big packets and I cannot lie! */
> + bool big_packets;
> +
> + /* Host will merge rx buffers for big packets (shake it! shake it!) */
> + bool mergeable_rx_bufs;
> +
> + /* Has control virtqueue */
> + bool has_cvq;
> +
> + /* Host can handle any s/g split between our header and packet data */
> + bool any_header_sg;
> +
> + /* Packet virtio header size */
> + u8 hdr_len;
> +
> + /* Work struct for refilling if we run low on memory. */
> + struct delayed_work refill;
> +
> + /* Work struct for config space updates */
> + struct work_struct config_work;
> +
> + /* Does the affinity hint is set for virtqueues? */
> + bool affinity_hint_set;
> +
> + /* CPU hotplug instances for online & dead */
> + struct hlist_node node;
> + struct hlist_node node_dead;
> +
> + struct control_buf *ctrl;
> +
> + /* Ethtool settings */
> + u8 duplex;
> + u32 speed;
> +
> + unsigned long guest_offloads;
> + unsigned long guest_offloads_capable;
> +
> + /* failover when STANDBY feature enabled */
> + struct failover *failover;
> +};
> +
> +/* RX packet size EWMA. The average packet size is used to determine the packet
> + * buffer size when refilling RX rings. As the entire RX ring may be refilled
> + * at once, the weight is chosen so that the EWMA will be insensitive to short-
> + * term, transient changes in packet size.
> + */
> +DECLARE_EWMA(pkt_len, 0, 64)
> +
> +struct virtnet_stat_desc {
> + char desc[ETH_GSTRING_LEN];
> + size_t offset;
> +};
> +
> +struct virtnet_sq_stats {
> + struct u64_stats_sync syncp;
> + u64 packets;
> + u64 bytes;
> + u64 xdp_tx;
> + u64 xdp_tx_drops;
> + u64 kicks;
> +};
> +
> +struct virtnet_rq_stats {
> + struct u64_stats_sync syncp;
> + u64 packets;
> + u64 bytes;
> + u64 drops;
> + u64 xdp_packets;
> + u64 xdp_tx;
> + u64 xdp_redirects;
> + u64 xdp_drops;
> + u64 kicks;
> +};
> +
> +#define VIRTNET_SQ_STAT(m) offsetof(struct virtnet_sq_stats, m)
> +#define VIRTNET_RQ_STAT(m) offsetof(struct virtnet_rq_stats, m)
> +
> +/* Internal representation of a send virtqueue */
> +struct send_queue {
> + /* Virtqueue associated with this send _queue */
> + struct virtqueue *vq;
> +
> + /* TX: fragments + linear part + virtio header */
> + struct scatterlist sg[MAX_SKB_FRAGS + 2];
> +
> + /* Name of the send queue: output.$index */
> + char name[40];
> +
> + struct virtnet_sq_stats stats;
> +
> + struct napi_struct napi;
> +};
> +
> +/* Internal representation of a receive virtqueue */
> +struct receive_queue {
> + /* Virtqueue associated with this receive_queue */
> + struct virtqueue *vq;
> +
> + struct napi_struct napi;
> +
> + struct bpf_prog __rcu *xdp_prog;
> +
> + struct virtnet_rq_stats stats;
> +
> + /* Chain pages by the private ptr. */
> + struct page *pages;
> +
> + /* Average packet length for mergeable receive buffers. */
> + struct ewma_pkt_len mrg_avg_pkt_len;
> +
> + /* Page frag for packet buffer allocation. */
> + struct page_frag alloc_frag;
> +
> + /* RX: fragments + linear part + virtio header */
> + struct scatterlist sg[MAX_SKB_FRAGS + 2];
> +
> + /* Min single buffer size for mergeable buffers case. */
> + unsigned int min_buf_len;
> +
> + /* Name of this receive queue: input.$index */
> + char name[40];
> +
> + struct xdp_rxq_info xdp_rxq;
> +};
> +
> +static inline bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
> +{
> + if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
> + return false;
> + else if (q < vi->curr_queue_pairs)
> + return true;
> + else
> + return false;
> +}
> +
> +static inline void virtqueue_napi_schedule(struct napi_struct *napi,
> + struct virtqueue *vq)
> +{
> + if (napi_schedule_prep(napi)) {
> + virtqueue_disable_cb(vq);
> + __napi_schedule(napi);
> + }
> +}
> +
> +static inline bool is_xdp_frame(void *ptr)
> +{
> + return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> +}
> +
> +static inline void *xdp_to_ptr(struct xdp_frame *ptr)
> +{
> + return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);
> +}
> +
> +static inline struct xdp_frame *ptr_to_xdp(void *ptr)
> +{
> + return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
> +}
> +
> +static inline void __free_old_xmit(struct send_queue *sq, bool in_napi,
> + struct virtnet_sq_stats *stats)
> +{
> + unsigned int len;
> + void *ptr;
> +
> + while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
> + if (!is_xdp_frame(ptr)) {
> + struct sk_buff *skb = ptr;
> +
> + pr_debug("Sent skb %p\n", skb);
> +
> + stats->bytes += skb->len;
> + napi_consume_skb(skb, in_napi);
> + } else {
> + struct xdp_frame *frame = ptr_to_xdp(ptr);
> +
> + stats->bytes += frame->len;
> + xdp_return_frame(frame);
> + }
> + stats->packets++;
> + }
> +}
> +
> +#endif
Powered by blists - more mailing lists