[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20241031133017.682be72b.pasic@linux.ibm.com>
Date: Thu, 31 Oct 2024 13:30:17 +0100
From: Halil Pasic <pasic@...ux.ibm.com>
To: Dust Li <dust.li@...ux.alibaba.com>
Cc: Wenjia Zhang <wenjia@...ux.ibm.com>, Wen Gu <guwen@...ux.alibaba.com>,
"D. Wythe" <alibuda@...ux.alibaba.com>,
Tony Lu <tonylu@...ux.alibaba.com>, David Miller <davem@...emloft.net>,
Jakub Kicinski <kuba@...nel.org>, Eric
Dumazet <edumazet@...gle.com>,
Paolo Abeni <pabeni@...hat.com>, netdev@...r.kernel.org,
linux-s390@...r.kernel.org, Heiko Carstens
<hca@...ux.ibm.com>,
Jan Karcher <jaka@...ux.ibm.com>, Gerd Bayer
<gbayer@...ux.ibm.com>,
Alexandra Winter <wintera@...ux.ibm.com>,
Nils
Hoppmann <niho@...ux.ibm.com>,
Niklas Schnell <schnelle@...ux.ibm.com>,
Thorsten Winkler <twinkler@...ux.ibm.com>,
Karsten Graul
<kgraul@...ux.ibm.com>,
Stefan Raspl <raspl@...ux.ibm.com>, Halil Pasic
<pasic@...ux.ibm.com>
Subject: Re: [PATCH net-next] net/smc: increase SMC_WR_BUF_CNT
On Sat, 26 Oct 2024 07:58:39 +0800
Dust Li <dust.li@...ux.alibaba.com> wrote:
> >For some reason include/linux/wait.h does not offer a top level wrapper
> >macro for wait_event with interruptible, exclusive and timeout. I did
> >not spend too many cycles on thinking if that is even a combination that
> >makes sense (on the quick I don't see why not) and conversely I
> >refrained from making an attempt to accomplish the interruptible,
> >exclusive and timeout combo by using the abstraction-wise lower
> >level __wait_event interface.
> >
> >To alleviate the tx performance bottleneck and the CPU overhead due to
> >the spinlock contention, let us increase SMC_WR_BUF_CNT to 256.
>
> Hi,
>
> Have you tested other values, such as 64? In our internal version, we
> have used 64 for some time.
Yes we have, but I'm not sure the data is still to be found. Let me do
some digging.
>
> Increasing this to 256 will require a 36K continuous physical memory
> allocation in smc_wr_alloc_link_mem(). Based on my experience, this may
> fail on servers that have been running for a long time and have
> fragmented memory.
Good point! It is possible that I did not give sufficient thought to
this aspect.
Regards,
Halil
Powered by blists - more mailing lists