lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <6siwtqeeqmyg62mqlxpckhopkjzl24qvrjuk6p7ccysaeg7ltw@pnzzf5hjlu3i>
Date: Fri, 28 Feb 2025 12:16:51 +0100
From: Uwe Kleine-König <u.kleine-koenig@...libre.com>
To: Abel Vesa <abel.vesa@...aro.org>
Cc: Sebastian Reichel <sre@...nel.org>, Lee Jones <lee@...nel.org>, 
	Pavel Machek <pavel@...nel.org>, Anjelique Melendez <quic_amelende@...cinc.com>, 
	Kamal Wadhwa <quic_kamalw@...cinc.com>, Jishnu Prakash <jishnu.prakash@....qualcomm.com>, 
	Bjorn Andersson <andersson@...nel.org>, Konrad Dybcio <konradybcio@...nel.org>, 
	Johan Hovold <johan@...nel.org>, Pavel Machek <pavel@....cz>, linux-leds@...r.kernel.org, 
	linux-kernel@...r.kernel.org, stable@...r.kernel.org
Subject: Re: [PATCH] leds: rgb: leds-qcom-lpg: Fix pwm resolution for Hi-Res
 PWMs

Hello Abel,

On Fri, Feb 28, 2025 at 10:59:09AM +0200, Abel Vesa wrote:
> On 25-02-27 19:09:39, Uwe Kleine-König wrote:
> > On Thu, Feb 27, 2025 at 07:05:14PM +0200, Abel Vesa wrote:
> > > On 25-02-27 18:32:41, Abel Vesa wrote:
> > > > On 25-02-27 17:44:35, Abel Vesa wrote:
> > > > > On 25-02-27 16:25:06, Uwe Kleine-König wrote:
> > > > > > Hello Abel,
> > > > > > 
> > > > > > On Thu, Feb 27, 2025 at 04:26:14PM +0200, Abel Vesa wrote:
> > > > > > > On 25-02-27 10:58:47, Uwe Kleine-König wrote:
> > > > > > > > Can you please enable CONFIG_PWM_DEBUG, enable pwm tracing (
> > > > > > > > 
> > > > > > > > 	echo 1 > /sys/kernel/debug/tracing/events/pwm/enable
> > > > > > > > 
> > > > > > > > ) then reproduce the problem and provide the output of
> > > > > > > > 
> > > > > > > > 	cat /sys/kernel/debug/tracing/trace
> > > > > > > > 
> > > > > > > > .
> > > > > > > 
> > > > > > > $ cat trace
> > > > > > > # tracer: nop
> > > > > > > #
> > > > > > > # entries-in-buffer/entries-written: 13/13   #P:12
> > > > > > > #
> > > > > > > #                                _-----=> irqs-off/BH-disabled
> > > > > > > #                               / _----=> need-resched
> > > > > > > #                              | / _---=> hardirq/softirq
> > > > > > > #                              || / _--=> preempt-depth
> > > > > > > #                              ||| / _-=> migrate-disable
> > > > > > > #                              |||| /     delay
> > > > > > > #           TASK-PID     CPU#  |||||  TIMESTAMP  FUNCTION
> > > > > > > #              | |         |   |||||     |         |
> > > > > > >         modprobe-203     [000] .....     0.938668: pwm_get: pwmchip0.0: period=1066407 duty_cycle=533334 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.938775: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=0 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.938821: pwm_get: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.938936: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.938982: pwm_get: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.939274: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=921458 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.939320: pwm_get: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.939434: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0
> > > > > > >         modprobe-203     [000] .....     0.939480: pwm_get: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0
> > > > > > >  systemd-backlig-724     [006] .....     9.079538: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=5000000 polarity=0 enabled=1 err=0
> > > > > > >  systemd-backlig-724     [006] .....     9.079585: pwm_get: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0
> > > > > > >  systemd-backlig-724     [006] .....     9.079698: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0
> > > > > > >  systemd-backlig-724     [006] .....     9.079750: pwm_get: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0
> > > > > > > $
> > > > > > > 
> > > > > > > > 
> > > > > > > > I didn't take a deeper dive in this driver combination, but here is a
> > > > > > > > description about what *should* happen:
> > > > > > > > 
> > > > > > > > You're talking about period in MHz, the PWM abstraction uses
> > > > > > > > nanoseconds. So your summary translated to the PWM wording is (to the
> > > > > > > > best of my understanding):
> > > > > > > > 
> > > > > > > >   1. PWM backlight driver requests PWM with .period = 200 ns and
> > > > > > > >      .duty_cycle = 200 ns.
> > > > > > > > 
> > > > > > > >   2. leds-qcom-lpg cannot pick 200 ns exactly and then chooses .period =
> > > > > > > >      1000000000 / 4.26666 MHz = 234.375 ns
> > > > > > > >      
> > > > > > > >   3. leds-qcom-lpg then determines setting for requested .duty_cycle
> > > > > > > >      based on .period = 200 ns which then ends up with something bogus.
> > > > > > 
> > > > > > The trace looks better than what I expected. 2. is fine here because it
> > > > > > seems when Sebastian wrote "driver requests PWM with 5 MHz period" that
> > > > > > meant period = 5000000 ns. That was then rounded down to 4266537 ns. And
> > > > > > the request for period = 5000000 ns + duty_cycle = 5000000 ns was
> > > > > > serviced by configuring period = 4266537 ns + duty_cycle = 4266537 ns.
> > > > > > So that's a 100 % relative duty configuration as intended.
> > > > > > 
> > > > > > So just from the traces I don't spot a problem. Do these logs not match
> > > > > > what actually happens on the signal?
> > > > > 
> > > > > What I do not get is why do we expect 2 pwm_get() and 2 pwm_apply()
> > > > > calls each time ?
> > > > 
> > > > OK, so the second pwm_apply() is due to CONFIG_PWM_DEBUG.
> > 
> > ack. This is done just for the tests implemented in CONFIG_PWM_DEBUG, as
> > are the two pwm_get()s.
> > 
> > > > But still, the first pwm_apply() requests duty cycle of 5MHz:
> > 
> > 5 ms, yes. But it cannot give you 5 ms and so you get 4.266 ns.
> > 
> > > > systemd-backlig-724     [006] .....     9.079538: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=5000000 polarity=0 enabled=1 err=0
> > > > 
> > > > So since the period is 4.26MHz, due to the knobs selected by the
> > > > provider, this duty cycle will result in a PWM value that is above the
> > > > selected resolution, as I already mentioned.
> > 
> > "above the selected resolution"? Do you mean you don't get the exact
> > value that you requested?
> 
> I think I understand your point now.
> 
> You expectation is that the provider would remap the entire range of the
> period to whatever the HW can do.

If I understand you correctly, that's right. For a given hardware there
is a set of possible periods P. .apply() should pick 
max{ p ∈ P | p ≤ state->period }.

And similar for duty_cycle: After choosing a possible period p ∈ P,
there is a set D(p) of duty_cycles that the hardware can implement in
combination to period p. .apply() should pick
max{ d ∈ D(p) | d ≤ state->duty_cycle }.

> So in this case, when 5ms is requested as duty cycle from consumer, the 
> provider will select the max value.

Yes.

> What the current implementation of the leds-qcom-lpg does is that will
> expect a duty cycle request of up to 4.26ms. And according to you, even
> if the consumer requests 5ms, the leds-qcom-lpg driver should write the
> value of 255 (which is what the selected resolution allows (1 << 8) ) and
> not compute a higher value.

If the period is 4.26 ms, duty_cycle cannot be bigger than 4.26 ms. So
yes, that's what the driver should do.

> I think this is wrong though. The fact that the pwm generic framework
> reports 5ms when it is actually 4.26ms should be considered wrong.

After pwm_apply_might_sleep(mypwm, { .period = 5000000, .duty_cycle =
5000000, .enabled = true }), pwm_get_state() gives you 5000000 and
pwm_get_state_hw() gives you 4266537. You could argue that the
functions's names and semantic are not optimal. Changing that is hard,
see my failed attempt in 01ccf903edd6 ("pwm: Let pwm_get_state() return
the last implemented state") + 40a6b9a00930 ("Revert "pwm: Let
pwm_get_state() return the last implemented state"")

So I don't see how the PWM framework is wrong here. Depending on what
value you want to get, pick pwm_get_state() or pwm_get_state_hw().

> For cases where the exact value of the duty cycle matters, this would
> not even make sense.

What is "this"? pwm_get_state() returning the last requested value? If
you're interested in the last requested value, it does make sense.
 
> Correct me if I'm wrong, but the pwm API should behave more like:
> The consumer should ask for the closest period the HW can actually do
> and then use that closest period from there on for every duty cycle
> request.

You can do that today using pwm_round_waveform_might_sleep() (however
that needs some glue in the leds-qcom-lpg driver).

And note that most in-kernel users don't care about exactness a lot. So
the fire-and-forget approach is fine and it shouldn't be made more
complicated for those.

> This way, if the consumer initially wants 5ms but the provider
> can do only 4.26ms instead, at least the consumer would be able to
> correct its duty cycle requests based on what the HW says it can do.

I agree that the consumer should be able to make an informed choice, and
that was my focus when designing the waveform API. But I intend to not
force that on (e.g.) the leds-pwm driver if that doesn't care about
getting 4.26 ms or 5 ms.

> > > On top of that, the duty cycle in debugfs is also reported as 5000000ns
> > > when in fact it is 4266666ns, as the trace shows.
> > 
> > Yes. Consider that a relict from the times when there was no
> > pwm_get_state_hw(). Both values are interesting in different situations.
> > So just telling the real parameters isn't the optimal way forward
> > either.
> > 
> > Something like the patch I showed in
> > https://lore.kernel.org/all/7bcnckef23w6g47ll5l3bktygedrcfvr7fk3qjuq2swtoffhec@zs4w4tuh6qvm/
> 
> And this patchset only adds the info of actual value that the HW is actually doing.

"only"? Yes, that's the intention of that patch. What should it do more?

> So basically, the already existing state in this case will represent the
> "desired" state.

Yes, pwm->state tracks the state that was last passed to
pwm_apply_might_sleep() (most of the time).
 
> > would make you a bit luckier I guess. Feel free to polish that one a bit
> > (e.g.  by checking the return value of pwm_get_state_hw() and acting
> > sensible in reply to it) and send a proper patch. (A Suggested-by for me
> > is enough for such a patch, grab authorship yourself.)
> > 
> > > > > Need to dig a bit further.
> > > > > 
> > > > > But meanwhile, if the first pwm_apply() call goes all the way to the
> > > > > provider, then the duty cycle value, when translated to the actual PWM
> > > > > value that gets written to reg, will overflow.
> > 
> > No it will not. The .duty_cycle value (also 5000000 ns) will reach the
> > lowlevel PWM driver together with .period = 5000000 ns. Both are rounded
> > down to 4266666ns. I see no overflow. 
> 
> Again, the consumer is being lied to. It expects 5ms and gets 4.26ms
> instead.

I see what you mean, but I don't agree. The semantic of
pwm_apply_might_sleep() is: "Configure the state that is nearest to the
passed state" (for some metric that defines "nearest"). The function
returning 0 means: The hardware now has this nearest state.

The semantic of pwm_get_state() is approximately: "What state was
requested before?" So it will give you .period = 5000000 ns and
.duty_cycle = 5000000 ns.

The semantic of pwm_get_state_hs() is: "What state is the hardware in?"
So it will give you .period = 4266666 ns and .duty_cycle = 4266666 ns.

So there are no lies, just wrong expectations about the semantic of
these functions.

And if you think that pwm_apply_might_sleep() should fail when 5000000
ns is requested and it can only do 4266537 ns: Where should the line
drawn that decides between "4977777 ns is still ok" and "4977777 ns is
too far from 5000000 ns"?

> Imagine a device that is controlled via PWM and needs exact duty cycle
> values in ms, what would the consumer driver do in this case?

Traditionally it would need some hardware specific extra information.
Today it could work out the needed details with the waveform API
functions (though this is hard because there are only two supported
lowlevel drivers and no helper functions yet).

> And to make things worse, when the consumer is asking for duty cycle of
> 4ms while the period requested is 5ms (which would be 80%), the period
> the provider will do is actually 4.26ms while the duty cycle would be
> ~3.41ms, which if the pwm step (reg value) doesn't allow, it will probably
> result in an actual value that is even further than what the consumer
> is expecting.

Where does ~3.41 ms come from? (I guess that's 0.8 * 4.26 ms.) Note that
if you request .period = 5 ms and .duty_cycle = 4 ms, you get .period =
4.26 ms and the biggest duty_cycle not bigger than 4 ms that is possible
with .period = 4.26 ms. So most likely not a 80% relative duty_cycle.

> So I'm thinking maybe the pwm should probably even ask the provider
> for what duty cycle it will provide based on provider's provided period
> and then decide if the resulting duty cycle is what it really wants.

Look into the waveform functions. The basic building blocks for what you
want should be there.

> IIRC, this is more in line with what the CCF (common clocks framework)
> currently does.

It does? There is clk_round_rate() but that is really hard to use
because there are virtually no promises in that function. Consider you
want a clock to run at 666666 Hz and clk_round_rate(yourclk, 666666)
gives you 500000 Hz. What would you do? Even: What is the rate above
666666 Hz that is as good as 500000 Hz for your usecase? Is it 833332 Hz
or 888888 Hz? And do you want 666666 Hz or 666666.666666667 Hz and how
does that influence your search for the right clkrate? And it has the
same problem as the pwm waveform functions: Just because
clk_round_rate(yourclk, 666666) returned 500000 200 ms ago, it doesn't
mean that if I ask for 666666 now the world didn't change.

Best regards
Uwe

Download attachment "signature.asc" of type "application/pgp-signature" (489 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ