lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Wed, 19 Oct 2022 17:04:31 +0200
From:   Bjørn Mork <bjorn@...k.no>
To:     Daniele Palmas <dnlplm@...il.com>
Cc:     David Miller <davem@...emloft.net>,
        Jakub Kicinski <kuba@...nel.org>,
        Paolo Abeni <pabeni@...hat.com>,
        Eric Dumazet <edumazet@...gle.com>, linux-usb@...r.kernel.org,
        netdev@...r.kernel.org
Subject: Re: [PATCH net-next 0/2] net: usb: qmi_wwan implement tx packets
 aggregation

Daniele Palmas <dnlplm@...il.com> writes:

> I verified this problem by using a MDM9207 Cat. 4 based modem (50Mbps/150Mbps
> max throughput). What is actually happening is pictured at
> https://drive.google.com/file/d/1xuAuDBfBEIM3Cdg2zHYQJ5tdk-JkfQn7/view?usp=sharing
>
> When rx and tx flows are tested singularly there's no issue in tx and minor
> issues in rx (a few spikes). When there are concurrent tx and rx flows, tx
> throughput has an huge drop. rx a minor one, but still present.
>
> The same scenario with tx aggregation enabled is pictured at
> https://drive.google.com/file/d/1Kw8TVFLVgr31o841fRu4fuMX9DNZqJB5/view?usp=sharing
> showing a regular graph.

That's pretty extreme.  Are these tcp tests?  Did you experiment with
qdisc options? What about latency here?

> This issue does not happen with high-cat modems (e.g. SDX20), or at least it
> does not happen at the throughputs I'm able to test currently: maybe the same
> could happen when moving close to the maximum rates supported by those modems.
> Anyway, having the tx aggregation enabled should not hurt.
>
> It is interesting to note that, for what I can understand, rmnet too does not
> support tx aggregation.

Looks like that is missing, yes. Did you consider implementing it there
instead?

> I'm aware that rmnet should be the preferred way for qmap, but I think there's
> still value in adding this feature to qmi_wwan qmap implementation since there
> are in the field many users of that.
>
> Moreover, having this in mainline could simplify backporting for those who are
> using qmi_wwan qmap feature but are stuck with old kernel versions.
>
> I'm also aware of the fact that sysfs files for configuration are not the
> preferred way, but it would feel odd changing the way for configuring the driver
> just for this feature, having it different from the previous knobs.

It's not just that it's not the preferred way.. I believe I promised
that we wouldn't add anything more to this interface.  And then I broke
that promise, promising that it would never happen again.  So much for
my integrity.

This all looks very nice to me, and the results are great, and it's just
another knob...

But I don't think we can continue adding this stuff.  The QMAP handling
should be done in the rmnet driver. Unless there is some reason it can't
be there? Wouldn't the same code work there?

Luckily I can chicken out here, and leave final the decision to Jakub
and David.



Bjørn

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ