lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <CACGkMEvjzGS1fthS93aV0QKX0maVjNtM43-CYJA5GSw30rEwDw@mail.gmail.com> Date: Wed, 25 Oct 2023 11:03:32 +0800 From: Jason Wang <jasowang@...hat.com> To: Heng Qi <hengqi@...ux.alibaba.com> Cc: "Michael S. Tsirkin" <mst@...hat.com>, netdev@...r.kernel.org, virtualization@...ts.linux-foundation.org, Xuan Zhuo <xuanzhuo@...ux.alibaba.com>, Eric Dumazet <edumazet@...gle.com>, "David S. Miller" <davem@...emloft.net>, Paolo Abeni <pabeni@...hat.com>, Jesper Dangaard Brouer <hawk@...nel.org>, John Fastabend <john.fastabend@...il.com>, Alexei Starovoitov <ast@...nel.org>, Jakub Kicinski <kuba@...nel.org>, Simon Horman <horms@...nel.org>, "Liu, Yujie" <yujie.liu@...el.com> Subject: Re: [PATCH net-next 3/5] virtio-net: extract virtqueue coalescig cmd for reuse On Thu, Oct 12, 2023 at 3:44 PM Heng Qi <hengqi@...ux.alibaba.com> wrote: > > Extract commands to set virtqueue coalescing parameters for reuse > by ethtool -Q, vq resize and netdim. > > Signed-off-by: Heng Qi <hengqi@...ux.alibaba.com> Acked-by: Jason Wang <jasowang@...hat.com> Thanks > --- > drivers/net/virtio_net.c | 106 +++++++++++++++++++++++---------------- > 1 file changed, 64 insertions(+), 42 deletions(-) > > diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c > index 54b3fb8d0384..caef78bb3963 100644 > --- a/drivers/net/virtio_net.c > +++ b/drivers/net/virtio_net.c > @@ -2846,6 +2846,58 @@ static void virtnet_cpu_notif_remove(struct virtnet_info *vi) > &vi->node_dead); > } > > +static int virtnet_send_ctrl_coal_vq_cmd(struct virtnet_info *vi, > + u16 vqn, u32 max_usecs, u32 max_packets) > +{ > + struct scatterlist sgs; > + > + vi->ctrl->coal_vq.vqn = cpu_to_le16(vqn); > + vi->ctrl->coal_vq.coal.max_usecs = cpu_to_le32(max_usecs); > + vi->ctrl->coal_vq.coal.max_packets = cpu_to_le32(max_packets); > + sg_init_one(&sgs, &vi->ctrl->coal_vq, sizeof(vi->ctrl->coal_vq)); > + > + if (!virtnet_send_command(vi, VIRTIO_NET_CTRL_NOTF_COAL, > + VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET, > + &sgs)) > + return -EINVAL; > + > + return 0; > +} > + > +static int virtnet_send_rx_ctrl_coal_vq_cmd(struct virtnet_info *vi, > + u16 queue, u32 max_usecs, > + u32 max_packets) > +{ > + int err; > + > + err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(queue), > + max_usecs, max_packets); > + if (err) > + return err; > + > + vi->rq[queue].intr_coal.max_usecs = max_usecs; > + vi->rq[queue].intr_coal.max_packets = max_packets; > + > + return 0; > +} > + > +static int virtnet_send_tx_ctrl_coal_vq_cmd(struct virtnet_info *vi, > + u16 queue, u32 max_usecs, > + u32 max_packets) > +{ > + int err; > + > + err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(queue), > + max_usecs, max_packets); > + if (err) > + return err; > + > + vi->sq[queue].intr_coal.max_usecs = max_usecs; > + vi->sq[queue].intr_coal.max_packets = max_packets; > + > + return 0; > +} > + > static void virtnet_get_ringparam(struct net_device *dev, > struct ethtool_ringparam *ring, > struct kernel_ethtool_ringparam *kernel_ring, > @@ -2903,14 +2955,11 @@ static int virtnet_set_ringparam(struct net_device *dev, > * through the VIRTIO_NET_CTRL_NOTF_COAL_TX_SET command, or, if the driver > * did not set any TX coalescing parameters, to 0. > */ > - err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(i), > - vi->intr_coal_tx.max_usecs, > - vi->intr_coal_tx.max_packets); > + err = virtnet_send_tx_ctrl_coal_vq_cmd(vi, i, > + vi->intr_coal_tx.max_usecs, > + vi->intr_coal_tx.max_packets); > if (err) > return err; > - > - vi->sq[i].intr_coal.max_usecs = vi->intr_coal_tx.max_usecs; > - vi->sq[i].intr_coal.max_packets = vi->intr_coal_tx.max_packets; > } > > if (ring->rx_pending != rx_pending) { > @@ -2919,14 +2968,11 @@ static int virtnet_set_ringparam(struct net_device *dev, > return err; > > /* The reason is same as the transmit virtqueue reset */ > - err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(i), > - vi->intr_coal_rx.max_usecs, > - vi->intr_coal_rx.max_packets); > + err = virtnet_send_rx_ctrl_coal_vq_cmd(vi, i, > + vi->intr_coal_rx.max_usecs, > + vi->intr_coal_rx.max_packets); > if (err) > return err; > - > - vi->rq[i].intr_coal.max_usecs = vi->intr_coal_rx.max_usecs; > - vi->rq[i].intr_coal.max_packets = vi->intr_coal_rx.max_packets; > } > } > > @@ -3327,48 +3373,24 @@ static int virtnet_send_notf_coal_cmds(struct virtnet_info *vi, > return 0; > } > > -static int virtnet_send_ctrl_coal_vq_cmd(struct virtnet_info *vi, > - u16 vqn, u32 max_usecs, u32 max_packets) > -{ > - struct scatterlist sgs; > - > - vi->ctrl->coal_vq.vqn = cpu_to_le16(vqn); > - vi->ctrl->coal_vq.coal.max_usecs = cpu_to_le32(max_usecs); > - vi->ctrl->coal_vq.coal.max_packets = cpu_to_le32(max_packets); > - sg_init_one(&sgs, &vi->ctrl->coal_vq, sizeof(vi->ctrl->coal_vq)); > - > - if (!virtnet_send_command(vi, VIRTIO_NET_CTRL_NOTF_COAL, > - VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET, > - &sgs)) > - return -EINVAL; > - > - return 0; > -} > - > static int virtnet_send_notf_coal_vq_cmds(struct virtnet_info *vi, > struct ethtool_coalesce *ec, > u16 queue) > { > int err; > > - err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(queue), > - ec->rx_coalesce_usecs, > - ec->rx_max_coalesced_frames); > + err = virtnet_send_rx_ctrl_coal_vq_cmd(vi, queue, > + ec->rx_coalesce_usecs, > + ec->rx_max_coalesced_frames); > if (err) > return err; > > - vi->rq[queue].intr_coal.max_usecs = ec->rx_coalesce_usecs; > - vi->rq[queue].intr_coal.max_packets = ec->rx_max_coalesced_frames; > - > - err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(queue), > - ec->tx_coalesce_usecs, > - ec->tx_max_coalesced_frames); > + err = virtnet_send_tx_ctrl_coal_vq_cmd(vi, queue, > + ec->tx_coalesce_usecs, > + ec->tx_max_coalesced_frames); > if (err) > return err; > > - vi->sq[queue].intr_coal.max_usecs = ec->tx_coalesce_usecs; > - vi->sq[queue].intr_coal.max_packets = ec->tx_max_coalesced_frames; > - > return 0; > } > > -- > 2.19.1.6.gb485710b >
Powered by blists - more mailing lists