[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20220811152012.319641-1-alexander.mikhalitsyn@virtuozzo.com>
Date: Thu, 11 Aug 2022 18:20:10 +0300
From: Alexander Mikhalitsyn <alexander.mikhalitsyn@...tuozzo.com>
To: netdev@...r.kernel.org
Cc: Alexander Mikhalitsyn <alexander.mikhalitsyn@...tuozzo.com>,
"David S. Miller" <davem@...emloft.net>,
Eric Dumazet <edumazet@...gle.com>,
Jakub Kicinski <kuba@...nel.org>,
Paolo Abeni <pabeni@...hat.com>,
Daniel Borkmann <daniel@...earbox.net>,
David Ahern <dsahern@...nel.org>,
Yajun Deng <yajun.deng@...ux.dev>,
Roopa Prabhu <roopa@...dia.com>,
Christian Brauner <brauner@...nel.org>,
linux-kernel@...r.kernel.org, "Denis V . Lunev" <den@...nvz.org>,
Alexey Kuznetsov <kuznet@....inr.ac.ru>,
Konstantin Khorenko <khorenko@...tuozzo.com>,
Pavel Tikhomirov <ptikhomirov@...tuozzo.com>,
Andrey Zhadchenko <andrey.zhadchenko@...tuozzo.com>,
Alexander Mikhalitsyn <alexander@...alicyn.com>,
kernel@...nvz.org, devel@...nvz.org
Subject: [PATCH v3 0/2] neighbour: fix possible DoS due to net iface start/stop loop
Dear friends,
Recently one of OpenVZ users reported that they have issues with network
availability of some containers. It was discovered that the reason is absence
of ARP replies from the Host Node on the requests about container IPs.
Of course, we started from tcpdump analysis and noticed that ARP requests
successfuly comes to the problematic node external interface. So, something
was wrong from the kernel side.
I've played a lot with arping and perf in attempts to understand what's
happening. And the key observation was that we experiencing issues only
with ARP requests with broadcast source ip (skb->pkt_type == PACKET_BROADCAST).
But for packets skb->pkt_type == PACKET_HOST everything works flawlessly.
Let me show a small piece of code:
static int arp_process(struct sock *sk, struct sk_buff *skb)
...
if (NEIGH_CB(skb)->flags & LOCALLY_ENQUEUED ||
skb->pkt_type == PACKET_HOST ||
NEIGH_VAR(in_dev->arp_parms, PROXY_DELAY) == 0) { // reply instantly
arp_send_dst(ARPOP_REPLY, ETH_P_ARP,
sip, dev, tip, sha,
dev->dev_addr, sha,
reply_dst);
} else {
pneigh_enqueue(&arp_tbl, // reply with delay
in_dev->arp_parms, skb);
goto out_free_dst;
}
The problem was that for PACKET_BROADCAST packets we delaying replies and use pneigh_enqueue() function.
For some reason, queued packets were lost almost all the time! The reason for such behaviour is pneigh_queue_purge()
function which cleanups all the queue, and this function called everytime once some network device in the system
gets link down.
neigh_ifdown -> pneigh_queue_purge
Now imagine that we have a node with 500+ containers with microservices. And some of that microservices are buggy
and always restarting... in this case, pneigh_queue_purge function will be called very frequently.
This problem is reproducible only with so-called "host routed" setup. The classical scheme bridge + veth
is not affected.
Minimal reproducer
Suppose that we have a network 172.29.1.1/16 brd 172.29.255.255
and we have free-to-use IP, let it be 172.29.128.3
1. Network configuration. I showing the minimal configuration, it makes no sense
as we have both veth devices stay at the same net namespace, but for demonstation and simplicity sake it's okay.
ip l a veth31427 type veth peer name veth314271
ip l s veth31427 up
ip l s veth314271 up
# setup static arp entry and publish it
arp -Ds -i br0 172.29.128.3 veth31427 pub
# setup static route for this address
route add 172.29.128.3/32 dev veth31427
2. "attacker" side (kubernetes pod with buggy microservice :) )
unshare -n
ip l a type veth
ip l s veth0 up
ip l s veth1 up
for i in {1..100000}; do ip link set veth0 down; sleep 0.01; ip link set veth0 up; done
This will totaly block ARP replies for 172.29.128.3 address. Just try
# arping -I eth0 172.29.128.3 -c 4
Our proposal is simple:
1. Let's cleanup queue partially. Remove only skb's that related to the net namespace
of the adapter which link is down.
2. Let's account proxy_queue limit properly per-device. Current limitation looks
not fully correct because we comparing per-device configurable limit with the
"global" qlen of proxy_queue.
Thanks,
Alex
v2:
- only ("neigh: fix possible DoS due to net iface start/stop") is changed
do del_timer_sync() if queue is empty after pneigh_queue_purge()
v3:
- rebase to net tree
Cc: "David S. Miller" <davem@...emloft.net>
Cc: Eric Dumazet <edumazet@...gle.com>
Cc: Jakub Kicinski <kuba@...nel.org>
Cc: Paolo Abeni <pabeni@...hat.com>
Cc: Daniel Borkmann <daniel@...earbox.net>
Cc: David Ahern <dsahern@...nel.org>
Cc: Yajun Deng <yajun.deng@...ux.dev>
Cc: Roopa Prabhu <roopa@...dia.com>
Cc: Christian Brauner <brauner@...nel.org>
Cc: netdev@...r.kernel.org
Cc: linux-kernel@...r.kernel.org
Cc: Denis V. Lunev <den@...nvz.org>
Cc: Alexey Kuznetsov <kuznet@....inr.ac.ru>
Cc: Konstantin Khorenko <khorenko@...tuozzo.com>
Cc: Pavel Tikhomirov <ptikhomirov@...tuozzo.com>
Cc: Andrey Zhadchenko <andrey.zhadchenko@...tuozzo.com>
Cc: Alexander Mikhalitsyn <alexander@...alicyn.com>
Cc: kernel@...nvz.org
Cc: devel@...nvz.org
Signed-off-by: Denis V. Lunev <den@...nvz.org>
Signed-off-by: Alexander Mikhalitsyn <alexander.mikhalitsyn@...tuozzo.com>
Alexander Mikhalitsyn (1):
neighbour: make proxy_queue.qlen limit per-device
Denis V. Lunev (1):
neigh: fix possible DoS due to net iface start/stop loop
include/net/neighbour.h | 1 +
net/core/neighbour.c | 46 +++++++++++++++++++++++++++++++++--------
2 files changed, 38 insertions(+), 9 deletions(-)
--
2.36.1
Powered by blists - more mailing lists