lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <8a1ff94c-8689-0d4c-cc33-7b495daa065a@deltatee.com>
Date:   Thu, 22 Jun 2017 12:40:31 -0600
From:   Logan Gunthorpe <logang@...tatee.com>
To:     Allen Hubbe <Allen.Hubbe@...l.com>, 'Jon Mason' <jdmason@...zu.us>
Cc:     linux-ntb@...glegroups.com, linux-kernel@...r.kernel.org,
        'Dave Jiang' <dave.jiang@...el.com>,
        'Serge Semin' <fancer.lancer@...il.com>,
        'Kurt Schwemmer' <kurt.schwemmer@...rosemi.com>,
        'Stephen Bates' <sbates@...thlin.com>,
        'Greg Kroah-Hartman' <gregkh@...uxfoundation.org>
Subject: Re: New NTB API Issue

On 6/22/2017 12:32 PM, Allen Hubbe wrote:
> From: Logan Gunthorpe
>> Hey Guys,
>>
>> I've run into some subtle issues with the new API:
>>
>> It has to do with splitting mw_get_range into mw_get_align and
>> peer_mw_get_addr.
>>
>> The original mw_get_range returned the size of the /local/ memory
>> window's size, address and alignment requirements. The ntb clients then
>> take the local size and transmit it via spads to the peer which would
>> use it in setting up the memory window. However, it made the assumption
>> that the alignment restrictions were symmetric on both hosts seeing they
>> were not sent across the link.
>>
>> The new API makes a sensible change for this in that mw_get_align
>> appears to be intended to return the alignment restrictions (and now
>> size) of the peer. This helps a bit for the Switchtec driver but appears
>> to be a semantic change that wasn't really reflected in the changes to
>> the other NTB code. So, I see a couple of issues:
>>
>> 1) With our hardware, we can't actually know anything about the peer's
>> memory windows until the peer has finished its setup (ie. the link is
>> up). However, all the clients call the function during probe, before the
>> link is ready. There's really no good reason for this, so I think we
>> should change the clients so that mw_get_align is called only when the
>> link is up.
>>
>> 2) The changes to the Intel and AMD driver for mw_get_align sets
>> *max_size to the local pci resource size. (Thus making the assumption
>> that the local is the same as the peer, which is wrong). max_size isn't
>> actually used for anything so it's not _really_ an issue, but I do think
>> it's confusing and incorrect. I'd suggest we remove max_size until
>> something actually needs it, or at least set it to zero in cases where
>> the hardware doesn't support returning the size of the peer's memory
>> window (ie. in the Intel and AMD drivers).
>
> You're right, and the b2b_split in the Intel driver even makes use of different primary/secondary bar sizes. For Intel and AMD, it would make more sense to use the secondary bar size here.  The size of the secondary bar still not necessarily valid end-to-end, because in b2b the peer's primary bar size could be even smaller.
>
> I'm not entirely convinced that this should represent the end-to-end size of local and peer memory window configurations.  I think it should represent the largest side that would be valid to pass to ntb_mw_set_trans().  Then, the peers should communicate their respective max sizes (along with translation addresses, etc) before setting up the translations, and that exchange will ensure that the size finally used is valid end-to-end.

But why would the client ever need to use the max_size instead of the 
actual size of the bar as retrieved and exchanged from peer_mw_get_addr?

Logan

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ