lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20191108185503.GB3384779@ulmo>
Date:   Fri, 8 Nov 2019 19:55:03 +0100
From:   Thierry Reding <thierry.reding@...il.com>
To:     Stephen Boyd <sboyd@...nel.org>
Cc:     Dmitry Osipenko <digetx@...il.com>,
        Sowjanya Komatineni <skomatineni@...dia.com>,
        jason@...edaemon.net, jonathanh@...dia.com,
        linus.walleij@...aro.org, marc.zyngier@....com,
        mark.rutland@....com, stefan@...er.ch, tglx@...utronix.de,
        pdeschrijver@...dia.com, pgaikwad@...dia.com,
        linux-clk@...r.kernel.org, linux-gpio@...r.kernel.org,
        jckuo@...dia.com, josephl@...dia.com, talho@...dia.com,
        linux-tegra@...r.kernel.org, linux-kernel@...r.kernel.org,
        mperttunen@...dia.com, spatra@...dia.com, robh+dt@...nel.org,
        devicetree@...r.kernel.org, rjw@...ysocki.net,
        viresh.kumar@...aro.org, linux-pm@...r.kernel.org
Subject: Re: [PATCH v9 07/22] clk: Add API to get index of the clock parent

On Fri, Nov 08, 2019 at 10:12:49AM -0800, Stephen Boyd wrote:
> Quoting Thierry Reding (2019-11-08 02:11:16)
> > On Thu, Nov 07, 2019 at 11:19:32AM -0800, Stephen Boyd wrote:
> > > Quoting Thierry Reding (2019-11-07 07:21:15)
> > > > On Thu, Nov 07, 2019 at 03:54:03AM +0300, Dmitry Osipenko wrote:
> > > > > 07.11.2019 02:10, Stephen Boyd пишет:
> > > > > > Quoting Sowjanya Komatineni (2019-08-16 12:41:52)
> > > > > >> This patch adds an API clk_hw_get_parent_index to get index of the
> > > > > >> clock parent to use during the clock restore operations on system
> > > > > >> resume.
> > > > > >  
> > > > > > Is there a reason we can't save the clk hw index at suspend time by
> > > > > > reading the hardware to understand the current parent? The parent index
> > > > > > typically doesn't matter unless we're trying to communicate something
> > > > > > from the framework to the provider driver. Put another way, I would
> > > > > > think the provider driver can figure out the index itself without having
> > > > > > to go through the framework to do so.
> > > > > 
> > > > > Isn't it a bit wasteful to duplicate information about the parent within
> > > > > a provider if framework already has that info? The whole point of this
> > > > > new API is to allow providers to avoid that unnecessary duplication.
> > > > > 
> > > > > Please note that clk_hw_get_parent_index is getting used only at the
> > > > > resume time and not at suspend.
> > > > 
> > > > I agree with this. All of the information that we need is already cached
> > > > in the framework. Doing this in the driver would mean essentially adding
> > > > a "saved parent" field along with code to read the value at suspend time
> > > > to the three types of clocks that currently use this core helper.
> > > 
> > > Don't we already have a "saved parent" field by storing the pointer to
> > > the clk_hw?
> > > 
> > > > 
> > > > That's certainly something that we *can* do, but it doesn't sound like a
> > > > better option than simply querying the framework for the value that we
> > > > need.
> > > > 
> > > 
> > > Let me say this another way. Why does this driver want to know the index
> > > that the framework uses for some clk_hw pointer? Perhaps it happens to
> > > align with the same value that hardware uses, but I still don't
> > > understand why the driver wants to know what the framework has decided
> > > is the index for some clk_hw pointer.
> > > 
> > > Or is this something like "give me the index for the parent that the
> > > framework thinks I currently have but in reality don't have anymore
> > > because the register contents were wiped and we need to reparent it"?
> > 
> > Yeah, that's exactly what this is being used for. It's used to restore
> > the parent/child relationship during resume after the registers have
> > been wiped during supend.
> 
> Ok cool. Our whole suspend/resume and save/restore story hasn't really
> been well thought out so we may want to pull all this logic into the
> core one day. For now it's OK to do the heavy lifting from provider
> drivers until someone gets a better grasp on how this should all work.

Ah, that would explain why I was scratching my head trying to understand
how exactly this was supposed to work. It did feel like there was some
infrastructure there, but looking around there wasn't a very consistent
usage pattern that I could find.

I think suspend/resume is always a little tricky. For example the clocks
may required a slightly different logical sequences between SoCs. Maybe
even different types of clocks have different needs. We seem to have a
bit of that on Tegra alone already. Without having delved into this too
much, it seems to me like the core can't do a whole lot without stepping
(potentially) on drivers' toes.

The current save_context/restore_context seems to be mostly that,
though, so I think it's a good starting point. You're right that we may
eventually see clearer patterns appear.

> > > A generic API to get any index for this question is overkill and we should
> > > consider adding some sort of API like clk_hw_get_current_parent_index(),
> > > or a framework flag that tells the framework this parent is incorrect
> > > and we need to call the .set_parent() op again to reconfigure it.
> > 
> > Okay, I think I see what you're saying. The current implementation does
> > carry a bit of a risk because users could be calling this function with
> > any arbitrary pair of struct clk_hw *, even completely unrelated ones.
> > 
> > How about we turn it into this instead:
> > 
> >         /**
> >          * clk_hw_get_parent_index - return the index of the parent clock
> >          * @hw: clk_hw associated with the clk being consumed
> >          *
> >          * Fetches and returns the index of parent clock. Returns -EINVAL if the given
> >          * clock does not have a current parent.
> >          */
> >         int clk_hw_get_parent_index(struct clk_hw *hw)
> >         {
> >                 struct clk_hw *parent = clk_hw_get_parent(hw);
> > 
> >                 if (!parent)
> >                         return -EINVAL;
> > 
> >                 return clk_fetch_parent_index(hw->core, parent->core);
> >         }
> >         EXPORT_SYMBOL_GPL(clk_hw_get_parent_index);
> > 
> > I think that has the advantage that we can't pass it a parent that's not
> > really a parent. There's still the slightly weird case where the clock
> > doesn't have a current parent, but hopefully that's something we are not
> > going to encounter much. After all this only makes sense to be called on
> > mux clocks and they always do have a parent by definition.
> 
> Right.
> 
> > 
> > Perhaps we should be more explicit and wrap that !parent conditional in
> > a WARN_ON()? In my local patches I do that at the call sites because
> > they are all functions returning void, so we'd be silently ignoring the
> > cases, but I think it may make sense to have it in the core.
> > 
> 
> Sure a WARN_ON() sounds fair. That will not take the whole task down
> and makes sure that drivers aren't doing something incorrect. Otherwise,
> this looks good and we can optimize by caching the parent index later if
> we really need to.

Okay, great. I'll go replace the above patch in the branch that I have.
I'm not sure if you saw it, but I had sent this in a pull request for
v5.5-rc1 about a week ago because I've got Tegra clock driver patches
that depend on this. I can replace this patch with the above proposal
and update the Tegra clock driver branch and then resend the two pull
requests.

Does that sound like a plan?

Thierry

Download attachment "signature.asc" of type "application/pgp-signature" (834 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ