[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20220330180436.24644-19-gdawar@xilinx.com>
Date: Wed, 30 Mar 2022 23:33:58 +0530
From: Gautam Dawar <gautam.dawar@...inx.com>
To: "Michael S. Tsirkin" <mst@...hat.com>,
Jason Wang <jasowang@...hat.com>,
Gautam Dawar <gdawar@...inx.com>, Eli Cohen <elic@...dia.com>,
Stefano Garzarella <sgarzare@...hat.com>,
Parav Pandit <parav@...dia.com>,
<virtualization@...ts.linux-foundation.org>,
<linux-kernel@...r.kernel.org>, <kvm@...r.kernel.org>,
<netdev@...r.kernel.org>
CC: <martinh@...inx.com>, <hanand@...inx.com>, <martinpo@...inx.com>,
<pabloc@...inx.com>, <dinang@...inx.com>, <tanuj.kamde@....com>,
<habetsm.xilinx@...il.com>, <ecree.xilinx@...il.com>,
<eperezma@...hat.com>, Wu Zongyong <wuzongyong@...ux.alibaba.com>,
Christophe JAILLET <christophe.jaillet@...adoo.fr>,
Zhu Lingshan <lingshan.zhu@...el.com>,
Xie Yongji <xieyongji@...edance.com>,
Si-Wei Liu <si-wei.liu@...cle.com>,
Longpeng <longpeng2@...wei.com>,
Dan Carpenter <dan.carpenter@...cle.com>,
Zhang Min <zhang.min9@....com.cn>
Subject: [PATCH v2 18/19] vdpa_sim: filter destination mac address
This patch implements a simple unicast filter for vDPA simulator.
Signed-off-by: Jason Wang <jasowang@...hat.com>
Signed-off-by: Gautam Dawar <gdawar@...inx.com>
---
drivers/vdpa/vdpa_sim/vdpa_sim_net.c | 49 ++++++++++++++++++----------
1 file changed, 31 insertions(+), 18 deletions(-)
diff --git a/drivers/vdpa/vdpa_sim/vdpa_sim_net.c b/drivers/vdpa/vdpa_sim/vdpa_sim_net.c
index f4607172b0b8..5fa59d4fddc8 100644
--- a/drivers/vdpa/vdpa_sim/vdpa_sim_net.c
+++ b/drivers/vdpa/vdpa_sim/vdpa_sim_net.c
@@ -47,13 +47,28 @@ static void vdpasim_net_complete(struct vdpasim_virtqueue *vq, size_t len)
local_bh_enable();
}
+static bool receive_filter(struct vdpasim *vdpasim, size_t len)
+{
+ bool modern = vdpasim->features & (1ULL << VIRTIO_F_VERSION_1);
+ size_t hdr_len = modern ? sizeof(struct virtio_net_hdr_v1) :
+ sizeof(struct virtio_net_hdr);
+ struct virtio_net_config *vio_config = vdpasim->config;
+
+ if (len < ETH_ALEN + hdr_len)
+ return false;
+
+ if (!strncmp(vdpasim->buffer + hdr_len, vio_config->mac, ETH_ALEN))
+ return true;
+
+ return false;
+}
+
static void vdpasim_net_work(struct work_struct *work)
{
struct vdpasim *vdpasim = container_of(work, struct vdpasim, work);
struct vdpasim_virtqueue *txq = &vdpasim->vqs[1];
struct vdpasim_virtqueue *rxq = &vdpasim->vqs[0];
ssize_t read, write;
- size_t total_write;
int pkts = 0;
int err;
@@ -66,36 +81,34 @@ static void vdpasim_net_work(struct work_struct *work)
goto out;
while (true) {
- total_write = 0;
err = vringh_getdesc_iotlb(&txq->vring, &txq->out_iov, NULL,
&txq->head, GFP_ATOMIC);
if (err <= 0)
break;
+ read = vringh_iov_pull_iotlb(&txq->vring, &txq->out_iov,
+ vdpasim->buffer,
+ PAGE_SIZE);
+
+ if (!receive_filter(vdpasim, read)) {
+ vdpasim_net_complete(txq, 0);
+ continue;
+ }
+
err = vringh_getdesc_iotlb(&rxq->vring, NULL, &rxq->in_iov,
&rxq->head, GFP_ATOMIC);
if (err <= 0) {
- vringh_complete_iotlb(&txq->vring, txq->head, 0);
+ vdpasim_net_complete(txq, 0);
break;
}
- while (true) {
- read = vringh_iov_pull_iotlb(&txq->vring, &txq->out_iov,
- vdpasim->buffer,
- PAGE_SIZE);
- if (read <= 0)
- break;
-
- write = vringh_iov_push_iotlb(&rxq->vring, &rxq->in_iov,
- vdpasim->buffer, read);
- if (write <= 0)
- break;
-
- total_write += write;
- }
+ write = vringh_iov_push_iotlb(&rxq->vring, &rxq->in_iov,
+ vdpasim->buffer, read);
+ if (write <= 0)
+ break;
vdpasim_net_complete(txq, 0);
- vdpasim_net_complete(rxq, total_write);
+ vdpasim_net_complete(rxq, write);
if (++pkts > 4) {
schedule_work(&vdpasim->work);
--
2.30.1
Powered by blists - more mailing lists