lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20140114082516.GB1101@redhat.com>
Date:	Tue, 14 Jan 2014 10:25:16 +0200
From:	"Michael S. Tsirkin" <mst@...hat.com>
To:	Jason Wang <jasowang@...hat.com>
Cc:	davem@...emloft.net, netdev@...r.kernel.org,
	linux-kernel@...r.kernel.org, Vlad Yasevich <vyasevic@...hat.com>,
	John Fastabend <john.r.fastabend@...el.com>,
	Stephen Hemminger <stephen@...workplumber.org>,
	Herbert Xu <herbert@...dor.apana.org.au>
Subject: Re: [PATCH net-next] tun/macvtap: limit the packets queued through
 rcvbuf

On Tue, Jan 14, 2014 at 02:53:07PM +0800, Jason Wang wrote:
> We used to limit the number of packets queued through tx_queue_length. This
> has several issues:
> 
> - tx_queue_length is the control of qdisc queue length, simply reusing it
>   to control the packets queued by device may cause confusion.
> - After commit 6acf54f1cf0a6747bac9fea26f34cfc5a9029523 ("macvtap: Add
>   support of packet capture on macvtap device."), an unexpected qdisc
>   caused by non-zero tx_queue_length will lead qdisc lock contention for
>   multiqueue deivce.
> - What we really want is to limit the total amount of memory occupied not
>   the number of packets.
> 
> So this patch tries to solve the above issues by using socket rcvbuf to
> limit the packets could be queued for tun/macvtap. This was done by using
> sock_queue_rcv_skb() instead of a direct call to skb_queue_tail(). Also two
> new ioctl() were introduced for userspace to change the rcvbuf like what we
> have done for sndbuf.
> 
> With this fix, we can safely change the tx_queue_len of macvtap to
> zero. This will make multiqueue works without extra lock contention.
> 
> Cc: Vlad Yasevich <vyasevic@...hat.com>
> Cc: Michael S. Tsirkin <mst@...hat.com>
> Cc: John Fastabend <john.r.fastabend@...el.com>
> Cc: Stephen Hemminger <stephen@...workplumber.org>
> Cc: Herbert Xu <herbert@...dor.apana.org.au>
> Signed-off-by: Jason Wang <jasowang@...hat.com>

No, I don't think we can change userspace-visible behaviour like that.

This will break any existing user that tries to control
queue length through sysfs,netlink or device ioctl.

Take a look at my patch in msg ID 20140109071721.GD19559@...hat.com
which gives one way to set tx_queue_len to zero without
breaking userspace.



> ---
>  drivers/net/macvtap.c       | 31 ++++++++++++++++++++---------
>  drivers/net/tun.c           | 48 +++++++++++++++++++++++++++++++++------------
>  include/uapi/linux/if_tun.h |  3 +++
>  3 files changed, 60 insertions(+), 22 deletions(-)
> 
> diff --git a/drivers/net/macvtap.c b/drivers/net/macvtap.c
> index a2c3a89..c429c56 100644
> --- a/drivers/net/macvtap.c
> +++ b/drivers/net/macvtap.c
> @@ -292,9 +292,6 @@ static rx_handler_result_t macvtap_handle_frame(struct sk_buff **pskb)
>  	if (!q)
>  		return RX_HANDLER_PASS;
>  
> -	if (skb_queue_len(&q->sk.sk_receive_queue) >= dev->tx_queue_len)
> -		goto drop;
> -
>  	skb_push(skb, ETH_HLEN);
>  
>  	/* Apply the forward feature mask so that we perform segmentation
> @@ -310,8 +307,10 @@ static rx_handler_result_t macvtap_handle_frame(struct sk_buff **pskb)
>  			goto drop;
>  
>  		if (!segs) {
> -			skb_queue_tail(&q->sk.sk_receive_queue, skb);
> -			goto wake_up;
> +			if (sock_queue_rcv_skb(&q->sk, skb))
> +				goto drop;
> +			else
> +				goto wake_up;
>  		}
>  
>  		kfree_skb(skb);
> @@ -319,11 +318,17 @@ static rx_handler_result_t macvtap_handle_frame(struct sk_buff **pskb)
>  			struct sk_buff *nskb = segs->next;
>  
>  			segs->next = NULL;
> -			skb_queue_tail(&q->sk.sk_receive_queue, segs);
> +			if (sock_queue_rcv_skb(&q->sk, segs)) {
> +				skb = segs;
> +				skb->next = nskb;
> +				goto drop;
> +			}
> +
>  			segs = nskb;
>  		}
>  	} else {
> -		skb_queue_tail(&q->sk.sk_receive_queue, skb);
> +		if (sock_queue_rcv_skb(&q->sk, skb))
> +			goto drop;
>  	}
>  
>  wake_up:
> @@ -333,7 +338,7 @@ wake_up:
>  drop:
>  	/* Count errors/drops only here, thus don't care about args. */
>  	macvlan_count_rx(vlan, 0, 0, 0);
> -	kfree_skb(skb);
> +	kfree_skb_list(skb);
>  	return RX_HANDLER_CONSUMED;
>  }
>  
> @@ -414,7 +419,7 @@ static void macvtap_dellink(struct net_device *dev,
>  static void macvtap_setup(struct net_device *dev)
>  {
>  	macvlan_common_setup(dev);
> -	dev->tx_queue_len = TUN_READQ_SIZE;
> +	dev->tx_queue_len = 0;
>  }
>  
>  static struct rtnl_link_ops macvtap_link_ops __read_mostly = {
> @@ -469,6 +474,7 @@ static int macvtap_open(struct inode *inode, struct file *file)
>  	sock_init_data(&q->sock, &q->sk);
>  	q->sk.sk_write_space = macvtap_sock_write_space;
>  	q->sk.sk_destruct = macvtap_sock_destruct;
> +	q->sk.sk_rcvbuf = TUN_RCVBUF;
>  	q->flags = IFF_VNET_HDR | IFF_NO_PI | IFF_TAP;
>  	q->vnet_hdr_sz = sizeof(struct virtio_net_hdr);
>  
> @@ -1040,6 +1046,13 @@ static long macvtap_ioctl(struct file *file, unsigned int cmd,
>  		q->sk.sk_sndbuf = u;
>  		return 0;
>  
> +	case TUNSETRCVBUF:
> +		if (get_user(u, up))
> +			return -EFAULT;
> +
> +		q->sk.sk_rcvbuf = u;
> +		return 0;
> +
>  	case TUNGETVNETHDRSZ:
>  		s = q->vnet_hdr_sz;
>  		if (put_user(s, sp))
> diff --git a/drivers/net/tun.c b/drivers/net/tun.c
> index 09f6662..7a08fa3 100644
> --- a/drivers/net/tun.c
> +++ b/drivers/net/tun.c
> @@ -177,6 +177,7 @@ struct tun_struct {
>  
>  	int			vnet_hdr_sz;
>  	int			sndbuf;
> +	int			rcvbuf;
>  	struct tap_filter	txflt;
>  	struct sock_fprog	fprog;
>  	/* protected by rtnl lock */
> @@ -771,17 +772,6 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev)
>  	if (!check_filter(&tun->txflt, skb))
>  		goto drop;
>  
> -	if (tfile->socket.sk->sk_filter &&
> -	    sk_filter(tfile->socket.sk, skb))
> -		goto drop;
> -
> -	/* Limit the number of packets queued by dividing txq length with the
> -	 * number of queues.
> -	 */
> -	if (skb_queue_len(&tfile->socket.sk->sk_receive_queue)
> -			  >= dev->tx_queue_len / tun->numqueues)
> -		goto drop;
> -
>  	if (unlikely(skb_orphan_frags(skb, GFP_ATOMIC)))
>  		goto drop;
>  
> @@ -798,7 +788,8 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev)
>  	nf_reset(skb);
>  
>  	/* Enqueue packet */
> -	skb_queue_tail(&tfile->socket.sk->sk_receive_queue, skb);
> +	if (sock_queue_rcv_skb(tfile->socket.sk, skb))
> +		goto drop;
>  
>  	/* Notify and wake up reader process */
>  	if (tfile->flags & TUN_FASYNC)
> @@ -1668,6 +1659,7 @@ static int tun_set_iff(struct net *net, struct file *file, struct ifreq *ifr)
>  
>  		tun->filter_attached = false;
>  		tun->sndbuf = tfile->socket.sk->sk_sndbuf;
> +		tun->rcvbuf = tfile->socket.sk->sk_rcvbuf;
>  
>  		spin_lock_init(&tun->lock);
>  
> @@ -1837,6 +1829,17 @@ static void tun_set_sndbuf(struct tun_struct *tun)
>  	}
>  }
>  
> +static void tun_set_rcvbuf(struct tun_struct *tun)
> +{
> +	struct tun_file *tfile;
> +	int i;
> +
> +	for (i = 0; i < tun->numqueues; i++) {
> +		tfile = rtnl_dereference(tun->tfiles[i]);
> +		tfile->socket.sk->sk_sndbuf = tun->sndbuf;
> +	}
> +}
> +
>  static int tun_set_queue(struct file *file, struct ifreq *ifr)
>  {
>  	struct tun_file *tfile = file->private_data;
> @@ -1878,7 +1881,7 @@ static long __tun_chr_ioctl(struct file *file, unsigned int cmd,
>  	struct ifreq ifr;
>  	kuid_t owner;
>  	kgid_t group;
> -	int sndbuf;
> +	int sndbuf, rcvbuf;
>  	int vnet_hdr_sz;
>  	unsigned int ifindex;
>  	int ret;
> @@ -2061,6 +2064,22 @@ static long __tun_chr_ioctl(struct file *file, unsigned int cmd,
>  		tun_set_sndbuf(tun);
>  		break;
>  
> +	case TUNGETRCVBUF:
> +		rcvbuf = tfile->socket.sk->sk_rcvbuf;
> +		if (copy_to_user(argp, &rcvbuf, sizeof(rcvbuf)))
> +			ret = -EFAULT;
> +		break;
> +
> +	case TUNSETRCVBUF:
> +		if (copy_from_user(&rcvbuf, argp, sizeof(rcvbuf))) {
> +			ret = -EFAULT;
> +			break;
> +		}
> +
> +		tun->rcvbuf = rcvbuf;
> +		tun_set_rcvbuf(tun);
> +		break;
> +
>  	case TUNGETVNETHDRSZ:
>  		vnet_hdr_sz = tun->vnet_hdr_sz;
>  		if (copy_to_user(argp, &vnet_hdr_sz, sizeof(vnet_hdr_sz)))
> @@ -2139,6 +2158,8 @@ static long tun_chr_compat_ioctl(struct file *file,
>  	case TUNSETTXFILTER:
>  	case TUNGETSNDBUF:
>  	case TUNSETSNDBUF:
> +	case TUNGETRCVBUF:
> +	case TUNSETRCVBUF:
>  	case SIOCGIFHWADDR:
>  	case SIOCSIFHWADDR:
>  		arg = (unsigned long)compat_ptr(arg);
> @@ -2204,6 +2225,7 @@ static int tun_chr_open(struct inode *inode, struct file * file)
>  
>  	tfile->sk.sk_write_space = tun_sock_write_space;
>  	tfile->sk.sk_sndbuf = INT_MAX;
> +	tfile->sk.sk_rcvbuf = TUN_RCVBUF;
>  
>  	file->private_data = tfile;
>  	set_bit(SOCK_EXTERNALLY_ALLOCATED, &tfile->socket.flags);
> diff --git a/include/uapi/linux/if_tun.h b/include/uapi/linux/if_tun.h
> index e9502dd..8e04657 100644
> --- a/include/uapi/linux/if_tun.h
> +++ b/include/uapi/linux/if_tun.h
> @@ -22,6 +22,7 @@
>  
>  /* Read queue size */
>  #define TUN_READQ_SIZE	500
> +#define TUN_RCVBUF	(512 * PAGE_SIZE)
>  
>  /* TUN device flags */
>  #define TUN_TUN_DEV 	0x0001	

That's about 16x less than we were able to queue previously
by default.
How can you be sure this won't break any applications?

> @@ -58,6 +59,8 @@
>  #define TUNSETQUEUE  _IOW('T', 217, int)
>  #define TUNSETIFINDEX	_IOW('T', 218, unsigned int)
>  #define TUNGETFILTER _IOR('T', 219, struct sock_fprog)
> +#define TUNGETRCVBUF   _IOR('T', 220, int)
> +#define TUNSETRCVBUF   _IOW('T', 221, int)
>  
>  /* TUNSETIFF ifr flags */
>  #define IFF_TUN		0x0001
> -- 
> 1.8.3.2
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ