lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <7121e4be-0e25-dd5f-9d29-0fb02cdbe8de@grimberg.me> Date: Sun, 13 Mar 2022 13:40:52 +0200 From: Sagi Grimberg <sagi@...mberg.me> To: Mingbao Sun <sunmingbao@....com>, Keith Busch <kbusch@...nel.org>, Jens Axboe <axboe@...com>, Christoph Hellwig <hch@....de>, Chaitanya Kulkarni <kch@...dia.com>, linux-nvme@...ts.infradead.org, linux-kernel@...r.kernel.org, Eric Dumazet <edumazet@...gle.com>, "David S . Miller" <davem@...emloft.net>, Hideaki YOSHIFUJI <yoshfuji@...ux-ipv6.org>, David Ahern <dsahern@...nel.org>, Jakub Kicinski <kuba@...nel.org>, netdev@...r.kernel.org Cc: tyler.sun@...l.com, ping.gan@...l.com, yanxiu.cai@...l.com, libin.zhang@...l.com, ao.sun@...l.com Subject: Re: [PATCH v2 2/3] nvme-tcp: support specifying the congestion-control > From: Mingbao Sun <tyler.sun@...l.com> Hey Mingbao, > congestion-control could have a noticeable impaction on the > performance of TCP-based communications. This is of course true > to NVMe_over_TCP. > > Different congestion-controls (e.g., cubic, dctcp) are suitable for > different scenarios. Proper adoption of congestion control would benefit > the performance. On the contrary, the performance could be destroyed. > > Though we can specify the congestion-control of NVMe_over_TCP via > writing '/proc/sys/net/ipv4/tcp_congestion_control', but this also > changes the congestion-control of all the future TCP sockets that > have not been explicitly assigned the congestion-control, thus bringing > potential impaction on their performance. > > So it makes sense to make NVMe_over_TCP support specifying the > congestion-control. And this commit addresses the host side. Thanks for this patchset. Generally, I'm not opposed to allow users to customize what they want to do, but in order to add something like this we need a few justifications. 1. Can you please provide your measurements that support your claims? 2. Can you please provide a real, existing use-case where this provides true, measureable value? And more specifically, please clarify how the use-case needs a local tuning for nvme-tcp that would not hold for other tcp streams that are running on the host (and vice-versa). 3. There are quite a few of TCP tuning knobs that will affect how nvme-tcp performs, just like any TCP application that running on Linux. However, Application level TCP tuning is not widespread at all, what makes nvme-tcp special to allow this, and why the TCP congestion is more important than other tuning knobs? I am not supportive of exporting all or some TCP level knobs as a local shadow for sysctl. Adding tunables, especially ones that are address niche use-cases in nature, can easily become a slippery slope for a set of rarely touched code and interface we are left stuck with for a long time... But while this feels a bit random to me, I'm not objecting to add this to the driver. I just want to make sure that this is something that is a) really required and b) does not backfire on us nor the user. > Implementation approach: > a new option called 'tcp_congestion' was created in fabrics opt_tokens > for 'nvme connect' command to passed in the congestion-control > specified by the user. > Then later in nvme_tcp_alloc_queue, the specified congestion-control > would be applied to the relevant sockets of the host side. Specifically to the interface, I'm wandering if this is the right interface... The user is used to sysctl with the semantics that it provides, wouldn't it be better to expose the exact same interface just for nvme-tcp sockets? Something like sysctl nvme.tcp_congestion_control ?
Powered by blists - more mailing lists