lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20200625.161636.1311963505288019596.davem@davemloft.net>
Date:   Thu, 25 Jun 2020 16:16:36 -0700 (PDT)
From:   David Miller <davem@...emloft.net>
To:     Jason@...c4.com
Cc:     netdev@...r.kernel.org
Subject: Re: [PATCH net 0/4] napi_gro_receive caller return value cleanups

From: "Jason A. Donenfeld" <Jason@...c4.com>
Date: Wed, 24 Jun 2020 16:06:02 -0600

> In 6570bc79c0df ("net: core: use listified Rx for GRO_NORMAL in
> napi_gro_receive()"), the GRO_NORMAL case stopped calling
> netif_receive_skb_internal, checking its return value, and returning
> GRO_DROP in case it failed. Instead, it calls into
> netif_receive_skb_list_internal (after a bit of indirection), which
> doesn't return any error. Therefore, napi_gro_receive will never return
> GRO_DROP, making handling GRO_DROP dead code.
> 
> I emailed the author of 6570bc79c0df on netdev [1] to see if this change
> was intentional, but the dlink.ru email address has been disconnected,
> and looking a bit further myself, it seems somewhat infeasible to start
> propagating return values backwards from the internal machinations of
> netif_receive_skb_list_internal.
> 
> Taking a look at all the callers of napi_gro_receive, it appears that
> three are checking the return value for the purpose of comparing it to
> the now never-happening GRO_DROP, and one just casts it to (void), a
> likely historical leftover. Every other of the 120 callers does not
> bother checking the return value.
> 
> And it seems like these remaining 116 callers are doing the right thing:
> after calling napi_gro_receive, the packet is now in the hands of the
> upper layers of the newtworking, and the device driver itself has no
> business now making decisions based on what the upper layers choose to
> do. Incrementing stats counters on GRO_DROP seems like a mistake, made
> by these three drivers, but not by the remaining 117.
> 
> It would seem, therefore, that after rectifying these four callers of
> napi_gro_receive, that I should go ahead and just remove returning the
> value from napi_gro_receive all together. However, napi_gro_receive has
> a function event tracer, and being able to introspect into the
> networking stack to see how often napi_gro_receive is returning whatever
> interesting GRO status (aside from _DROP) remains an interesting
> data point worth keeping for debugging.
> 
> So, this series simply gets rid of the return value checking for the
> four useless places where that check never evaluates to anything
> meaningful.
> 
> [1] https://lore.kernel.org/netdev/20200624210606.GA1362687@zx2c4.com/

Seires applied, thank you.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ