[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAMZdPi_GPxTT8Wky84V_FtGxC7cXLeeAi7tqkU9Xj_XxdFcL8Q@mail.gmail.com>
Date: Thu, 18 Mar 2021 17:43:31 +0100
From: Loic Poulain <loic.poulain@...aro.org>
To: Jeffrey Hugo <jhugo@...eaurora.org>,
Bhaumik Bhatt <bbhatt@...eaurora.org>
Cc: Manivannan Sadhasivam <manivannan.sadhasivam@...aro.org>,
linux-arm-msm <linux-arm-msm@...r.kernel.org>,
Hemant Kumar <hemantk@...eaurora.org>,
open list <linux-kernel@...r.kernel.org>,
Carl Yin(殷张成) <carl.yin@...ctel.com>,
Naveen Kumar <naveen.kumar@...ctel.com>,
jhugo=codeaurora.org@...eaurora.org
Subject: Re: [PATCH v4 1/3] bus: mhi: core: Introduce internal register poll
helper function
On Thu, 18 Mar 2021 at 17:13, Jeffrey Hugo <jhugo@...eaurora.org> wrote:
>
> On 3/17/2021 3:26 PM, Bhaumik Bhatt wrote:
> > On 2021-03-11 11:59 AM, Jeffrey Hugo wrote:
> >> On 3/11/2021 1:00 AM, Loic Poulain wrote:
> >>> Hi Bhaumik,
> >>>
> >>> On Thu, 11 Mar 2021 at 00:31, Bhaumik Bhatt <bbhatt@...eaurora.org>
> >>> wrote:
> >>>>
> >>>> Introduce helper function to allow MHI core driver to poll for
> >>>> a value in a register field. This helps reach a common path to
> >>>> read and poll register values along with a retry time interval.
> >>>>
> >>>> Signed-off-by: Bhaumik Bhatt <bbhatt@...eaurora.org>
> >>>> ---
> >>>> drivers/bus/mhi/core/internal.h | 3 +++
> >>>> drivers/bus/mhi/core/main.c | 23 +++++++++++++++++++++++
> >>>> 2 files changed, 26 insertions(+)
> >>>>
> >>>> diff --git a/drivers/bus/mhi/core/internal.h
> >>>> b/drivers/bus/mhi/core/internal.h
> >>>> index 6f80ec3..005286b 100644
> >>>> --- a/drivers/bus/mhi/core/internal.h
> >>>> +++ b/drivers/bus/mhi/core/internal.h
> >>>> @@ -643,6 +643,9 @@ int __must_check mhi_read_reg(struct
> >>>> mhi_controller *mhi_cntrl,
> >>>> int __must_check mhi_read_reg_field(struct mhi_controller *mhi_cntrl,
> >>>> void __iomem *base, u32 offset,
> >>>> u32 mask,
> >>>> u32 shift, u32 *out);
> >>>> +int __must_check mhi_poll_reg_field(struct mhi_controller *mhi_cntrl,
> >>>> + void __iomem *base, u32 offset,
> >>>> u32 mask,
> >>>> + u32 shift, u32 val, u32 delayus);
> >>>> void mhi_write_reg(struct mhi_controller *mhi_cntrl, void __iomem
> >>>> *base,
> >>>> u32 offset, u32 val);
> >>>> void mhi_write_reg_field(struct mhi_controller *mhi_cntrl, void
> >>>> __iomem *base,
> >>>> diff --git a/drivers/bus/mhi/core/main.c b/drivers/bus/mhi/core/main.c
> >>>> index 4e0131b..7c7f41a 100644
> >>>> --- a/drivers/bus/mhi/core/main.c
> >>>> +++ b/drivers/bus/mhi/core/main.c
> >>>> @@ -4,6 +4,7 @@
> >>>> *
> >>>> */
> >>>>
> >>>> +#include <linux/delay.h>
> >>>> #include <linux/device.h>
> >>>> #include <linux/dma-direction.h>
> >>>> #include <linux/dma-mapping.h>
> >>>> @@ -37,6 +38,28 @@ int __must_check mhi_read_reg_field(struct
> >>>> mhi_controller *mhi_cntrl,
> >>>> return 0;
> >>>> }
> >>>>
> >>>> +int __must_check mhi_poll_reg_field(struct mhi_controller *mhi_cntrl,
> >>>> + void __iomem *base, u32 offset,
> >>>> + u32 mask, u32 shift, u32 val,
> >>>> u32 delayus)
> >>>> +{
> >>>> + int ret;
> >>>> + u32 out, retry = (mhi_cntrl->timeout_ms * 1000) / delayus;
> >>>> +
> >>>> + while (retry--) {
> >>>> + ret = mhi_read_reg_field(mhi_cntrl, base, offset,
> >>>> mask, shift,
> >>>> + &out);
> >>>> + if (ret)
> >>>> + return ret;
> >>>> +
> >>>> + if (out == val)
> >>>> + return 0;
> >>>> +
> >>>> + udelay(delayus);
> >>>
> >>> Have you read my previous comment?
> >>> Do you really want to risk hogging the CPU for several seconds? we
> >>> know that some devices take several seconds to start/boot.
> >>> Why not using msleep variant here?
> >>
> >> usleep_range() if there is a desire to stay in us units?
> >>
> >> Given that the use of this function is for 25ms in one case, I wonder
> >> if this warning is applicable:
> >> https://elixir.bootlin.com/linux/latest/source/include/linux/delay.h#L28
> >>
> >> Counter point, 1ms latency over PCIe is not unusual. I know we've
> >> removed the PCIe dependencies from MHI, but PCIe is the real usecase
> >> at this time. Seems like this function could behave a bit weird if
> >> the parameter to udelay is something like "100", but the
> >> mhi_read_reg_field() call takes significantly longer than that. Feels
> >> like in some scenarios, we could actually exceed the timeout by a
> >> non-trivial margin.
> >>
> >> I guess I'm going back and forth in determining if us scale timing is
> >> a benefit in any way.
> > Thanks for all the inputs. I think a good idea here would be to use
> > fsleep()
> > API as we need to allow any timeout the caller specifies. Also, plan is to
> > drop the patch #3 in this series since that will require a busywait due to
> > the code being in panic path.
> >
> > I don't wish to accommodate another variable here for busywait but that
> > would be an option to pick sleep or delay depending on the caller's path.
> >
> > Please respond if there are any concerns.
>
> fsleep() would be some improvement, but I think there is still the issue
> Loic points out where if delayus is small, but timeout_ms is large (say
> 50us and 25s), this function will end up burning a lot of cpu cycles
> However that is likely an edge case, and if your poll cycle is that
> small, I think it should be assumed that the event is expected to happen
> quickly, so the full timeout should not be hit.
Well, my point is that during initial power_up, with a device
cold-booting, it can take several seconds for it to reach ready state
(not a corner case). That why timeout_ms can be as large as 20 seconds
for mhi_pci_modem. If polling is based on busy-wait, that means the
while loop will not let the CPU running anything else for several
seconds. Not sure what is the expected meaning of this timeout_ms in
first place... maybe I just use it badly.
Moreover, do we need microsecond latency on detecting ready
transition, this is not a critical path, right?
Regards,
Loic
Powered by blists - more mailing lists