lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-Id: <1462198956-30181-2-git-send-email-jon.maloy@ericsson.com> Date: Mon, 2 May 2016 10:22:34 -0400 From: Jon Maloy <jon.maloy@...csson.com> To: davem@...emloft.net Cc: netdev@...r.kernel.org, Paul Gortmaker <paul.gortmaker@...driver.com>, parthasarathy.bhuvaragan@...csson.com, richard.alpe@...csson.com, ying.xue@...driver.com, maloy@...jonn.com, tipc-discussion@...ts.sourceforge.net, Jon Maloy <jon.maloy@...csson.com> Subject: [PATCH net-next 1/3] tipc: re-enable compensation for socket receive buffer double counting In the refactoring commit d570d86497ee ("tipc: enqueue arrived buffers in socket in separate function") we did by accident replace the test if (sk->sk_backlog.len == 0) atomic_set(&tsk->dupl_rcvcnt, 0); with if (sk->sk_backlog.len) atomic_set(&tsk->dupl_rcvcnt, 0); This effectively disables the compensation we have for the double receive buffer accounting that occurs temporarily when buffers are moved from the backlog to the socket receive queue. Until now, this has gone unnoticed because of the large receive buffer limits we are applying, but becomes indispensable when we reduce this buffer limit later in this series. We now fix this by inverting the mentioned condition. Acked-by: Ying Xue <ying.xue@...driver.com> Signed-off-by: Jon Maloy <jon.maloy@...csson.com> --- net/tipc/socket.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/net/tipc/socket.c b/net/tipc/socket.c index 3eeb50a..d37a940 100644 --- a/net/tipc/socket.c +++ b/net/tipc/socket.c @@ -1748,7 +1748,7 @@ static void tipc_sk_enqueue(struct sk_buff_head *inputq, struct sock *sk, /* Try backlog, compensating for double-counted bytes */ dcnt = &tipc_sk(sk)->dupl_rcvcnt; - if (sk->sk_backlog.len) + if (!sk->sk_backlog.len) atomic_set(dcnt, 0); lim = rcvbuf_limit(sk, skb) + atomic_read(dcnt); if (likely(!sk_add_backlog(sk, skb, lim))) -- 1.9.1
Powered by blists - more mailing lists