[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAK8P3a0brT0zyZGNWiS2R0RMHHFF2JG=_ixQyvjhj3Ky39o0UA@mail.gmail.com>
Date: Mon, 3 Jun 2019 12:04:18 +0200
From: Arnd Bergmann <arnd@...db.de>
To: Subash Abhinov Kasiviswanathan <subashab@...eaurora.org>
Cc: Bjorn Andersson <bjorn.andersson@...aro.org>,
Alex Elder <elder@...aro.org>, Dan Williams <dcbw@...hat.com>,
David Miller <davem@...emloft.net>,
Ilias Apalodimas <ilias.apalodimas@...aro.org>,
evgreen@...omium.org, Ben Chan <benchan@...gle.com>,
Eric Caruso <ejcaruso@...gle.com>, cpratapa@...eaurora.org,
syadagir@...eaurora.org, abhishek.esse@...il.com,
Networking <netdev@...r.kernel.org>,
DTML <devicetree@...r.kernel.org>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
linux-soc@...r.kernel.org,
Linux ARM <linux-arm-kernel@...ts.infradead.org>,
linux-arm-msm@...r.kernel.org
Subject: Re: [PATCH v2 00/17] net: introduce Qualcomm IPA driver
On Sat, Jun 1, 2019 at 1:59 AM Subash Abhinov Kasiviswanathan
<subashab@...eaurora.org> wrote:
> On 2019-05-31 17:33, Bjorn Andersson wrote:
> > On Fri 31 May 13:47 PDT 2019, Alex Elder wrote:
> >> On 5/31/19 2:19 PM, Arnd Bergmann wrote:
> > But any such changes would either be years into the future or for
> > specific devices and as such not applicable to any/most of devices on
> > the market now or in the coming years.
> >
> >
> > But as Arnd points out, if the software split between IPA and rmnet is
> > suboptimal your are encouraged to fix that.
>
> The split rmnet design was chosen because we could place rmnet
> over any transport - IPA, PCIe (https://lkml.org/lkml/2018/4/26/1159)
> or USB.
>
> rmnet registers a rx handler, so the rmnet packet processing itself
> happens in the same softirq when packets are queued to network stack
> by IPA.
I've read up on the implementation some more, and concluded that
it's mostly a regular protocol wrapper, doing IP over QMAP. There
is nothing wrong with the basic concept I think, and as you describe
this is an abstraction to keep the common bits in one place, and
have them configured consistently.
A few observations on more details here:
- What I'm worried about most here is the flow control handling on the
transmit side. The IPA driver now uses the modern BQL method to
control how much data gets submitted to the hardware at any time.
The rmnet driver also uses flow control using the
rmnet_map_command() function, that blocks tx on the higher
level device when the remote side asks us to.
I fear that doing flow control for a single physical device on two
separate netdev instances is counterproductive and confuses
both sides.
- I was a little confused by the location of the rmnet driver in
drivers/net/ethernet/... More conventionally, I think as a protocol
handler it should go into net/qmap/, with the ipa driver going
into drivers/net/qmap/ipa/, similar to what we have fo ethernet,
wireless, ppp, appletalk, etc.
- The rx_handler uses gro_cells, which as I understand is meant
for generic tunnelling setups and takes another loop through
NAPI to aggregate data from multiple queues, but in case of
IPA's single-queue receive calling gro directly would be simpler
and more efficient.
- I'm still not sure I understand the purpose of the layering with
using an rx_handler as opposed to just using
EXPORT_SYMBOL(rmnet_rx_handler) and calling that from
the hardware driver directly.
From the overall design and the rmnet Kconfig description, it
appears as though the intention as that rmnet could be a
generic wrapper on top of any device, but from the
implementation it seems that IPA is not actually usable that
way and would always go through IPA.
Arnd
Powered by blists - more mailing lists