[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CANn89i+6Hc7q-a=zh_jcTn9_GM5xP6fzv2RcHY+tneqzE3UnHw@mail.gmail.com>
Date: Tue, 15 Feb 2022 22:24:57 -0800
From: Eric Dumazet <edumazet@...gle.com>
To: Jason Xing <kerneljasonxing@...il.com>
Cc: David Miller <davem@...emloft.net>,
Jakub Kicinski <kuba@...nel.org>,
Alexei Starovoitov <ast@...nel.org>,
Daniel Borkmann <daniel@...earbox.net>,
Andrii Nakryiko <andrii@...nel.org>,
Martin KaFai Lau <kafai@...com>,
Song Liu <songliubraving@...com>, Yonghong Song <yhs@...com>,
John Fastabend <john.fastabend@...il.com>,
KP Singh <kpsingh@...nel.org>, Paolo Abeni <pabeni@...hat.com>,
Wei Wang <weiwan@...gle.com>,
Alexander Aring <aahringo@...hat.com>,
Yangbo Lu <yangbo.lu@....com>, Florian Westphal <fw@...len.de>,
Tonghao Zhang <xiangxia.m.yue@...il.com>,
Thomas Gleixner <tglx@...utronix.de>,
netdev <netdev@...r.kernel.org>,
LKML <linux-kernel@...r.kernel.org>, bpf <bpf@...r.kernel.org>,
Jason Xing <xingwanli@...ishou.com>
Subject: Re: [PATCH v2 net-next] net: introduce SO_RCVBUFAUTO to let the
rcv_buf tune automatically
On Tue, Feb 15, 2022 at 9:03 PM <kerneljasonxing@...il.com> wrote:
>
> From: Jason Xing <xingwanli@...ishou.com>
>
> Normally, user doesn't care the logic behind the kernel if they're
> trying to set receive buffer via setsockopt. However, once the new
> value of the receive buffer is set even though it's not smaller than
> the initial value which is sysctl_tcp_rmem[1] implemented in
> tcp_rcv_space_adjust(),, the server's wscale will shrink and then
> lead to the bad bandwidth as intended.
Quite confusing changelog, honestly.
Users of SO_RCVBUF specifically told the kernel : I want to use _this_
buffer size, I do not want the kernel to decide for me.
Also, I think your changelog does not really explain that _if_ you set
SO_RCVBUF to a small value before
connect() or in general the 3WHS, the chosen wscale will be small, and
this won't allow future 10x increase
of the effective RWIN.
>
> For now, introducing a new socket option to let the receive buffer
> grow automatically no matter what the new value is can solve
> the bad bandwidth issue meanwhile it's not breaking the application
> with SO_RCVBUF option set.
>
> Here are some numbers:
> $ sysctl -a | grep rmem
> net.core.rmem_default = 212992
> net.core.rmem_max = 40880000
> net.ipv4.tcp_rmem = 4096 425984 40880000
>
> Case 1
> on the server side
> # iperf -s -p 5201
> on the client side
> # iperf -c [client ip] -p 5201
> It turns out that the bandwidth is 9.34 Gbits/sec while the wscale of
> server side is 10. It's good.
>
> Case 2
> on the server side
> #iperf -s -p 5201 -w 425984
> on the client side
> # iperf -c [client ip] -p 5201
> It turns out that the bandwidth is reduced to 2.73 Gbits/sec while the
> wcale is 2, even though the receive buffer is not changed at all at the
> very beginning.
>
> After this patch is applied, the bandwidth of case 2 is recovered to
> 9.34 Gbits/sec as expected at the cost of consuming more memory per
> socket.
How does your patch allow wscale to increase after flow is established ?
I would remove from the changelog these experimental numbers that look
quite wrong,
maybe copy/pasted from your prior version.
Instead I would describe why an application might want to clear the
'receive buffer size is locked' socket attribute.
>
> Signed-off-by: Jason Xing <xingwanli@...ishou.com>
> --
> v2: suggested by Eric
> - introduce new socket option instead of breaking the logic in SO_RCVBUF
> - Adjust the title and description of this patch
> link: https://lore.kernel.org/lkml/CANn89iL8vOUOH9bZaiA-cKcms+PotuKCxv7LpVx3RF0dDDSnmg@mail.gmail.com/
> ---
>
I think adding another parallel SO_RCVBUF option is not good. It is
adding confusion (and net/core/filter.c has been unchanged)
Also we want CRIU to work correctly.
So if you have a SO_XXXX setsockopt() call, you also need to provide
getsockopt() implementation.
I would suggest an option to clear or set SOCK_RCVBUF_LOCK, and
getsockopt() would return if the bit is currently set or not.
Something clearly describing the intent, like SO_RCVBUF_LOCK maybe.
Powered by blists - more mailing lists