[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CADvbK_cArD4QjPzHC1W4ZY=e5dur=uJsQd9WwSB2yYhnq_YbOg@mail.gmail.com>
Date: Fri, 18 Dec 2015 16:18:38 +0800
From: Xin Long <lucien.xin@...il.com>
To: Herbert Xu <herbert@...dor.apana.org.au>
Cc: Phil Sutter <phil@....cc>, davem <davem@...emloft.net>,
network dev <netdev@...r.kernel.org>,
linux-kernel@...r.kernel.org, tgraf@...g.ch,
fengguang.wu@...el.com, wfg@...ux.intel.com, lkp@...org
Subject: Re: rhashtable: Prevent spurious EBUSY errors on insertion
On Fri, Dec 18, 2015 at 10:26 AM, Herbert Xu
<herbert@...dor.apana.org.au> wrote:
> On Fri, Dec 18, 2015 at 12:07:08AM +0800, Xin Long wrote:
>>
>> I'm just wondering, why do not we handle the genuine double rehash
>> issue inside rhashtable? i mean it's just a temporary error that a
>> simple retry may fix it.
>
> Because a double rehash means that someone has cracked your hash
> function and there is no point in trying anymore.
ok, get your point, is it possible to be triggered by some cases under
a big stress insertion, but they are all legal cases. like we use rhash in
nftables, if there are a big batch sets to insert, may this issue happen?
>
> Cheers,
> --
> Email: Herbert Xu <herbert@...dor.apana.org.au>
> Home Page: http://gondor.apana.org.au/~herbert/
> PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists