Hello Abel, On Fri, Feb 28, 2025 at 10:59:09AM +0200, Abel Vesa wrote: > On 25-02-27 19:09:39, Uwe Kleine-König wrote: > > On Thu, Feb 27, 2025 at 07:05:14PM +0200, Abel Vesa wrote: > > > On 25-02-27 18:32:41, Abel Vesa wrote: > > > > On 25-02-27 17:44:35, Abel Vesa wrote: > > > > > On 25-02-27 16:25:06, Uwe Kleine-König wrote: > > > > > > Hello Abel, > > > > > > > > > > > > On Thu, Feb 27, 2025 at 04:26:14PM +0200, Abel Vesa wrote: > > > > > > > On 25-02-27 10:58:47, Uwe Kleine-König wrote: > > > > > > > > Can you please enable CONFIG_PWM_DEBUG, enable pwm tracing ( > > > > > > > > > > > > > > > > echo 1 > /sys/kernel/debug/tracing/events/pwm/enable > > > > > > > > > > > > > > > > ) then reproduce the problem and provide the output of > > > > > > > > > > > > > > > > cat /sys/kernel/debug/tracing/trace > > > > > > > > > > > > > > > > . > > > > > > > > > > > > > > $ cat trace > > > > > > > # tracer: nop > > > > > > > # > > > > > > > # entries-in-buffer/entries-written: 13/13 #P:12 > > > > > > > # > > > > > > > # _-----=> irqs-off/BH-disabled > > > > > > > # / _----=> need-resched > > > > > > > # | / _---=> hardirq/softirq > > > > > > > # || / _--=> preempt-depth > > > > > > > # ||| / _-=> migrate-disable > > > > > > > # |||| / delay > > > > > > > # TASK-PID CPU# ||||| TIMESTAMP FUNCTION > > > > > > > # | | | ||||| | | > > > > > > > modprobe-203 [000] ..... 0.938668: pwm_get: pwmchip0.0: period=1066407 duty_cycle=533334 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.938775: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=0 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.938821: pwm_get: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.938936: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.938982: pwm_get: pwmchip0.0: period=4266537 duty_cycle=0 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.939274: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=921458 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.939320: pwm_get: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.939434: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0 > > > > > > > modprobe-203 [000] ..... 0.939480: pwm_get: pwmchip0.0: period=4266537 duty_cycle=921355 polarity=0 enabled=1 err=0 > > > > > > > systemd-backlig-724 [006] ..... 9.079538: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=5000000 polarity=0 enabled=1 err=0 > > > > > > > systemd-backlig-724 [006] ..... 9.079585: pwm_get: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0 > > > > > > > systemd-backlig-724 [006] ..... 9.079698: pwm_apply: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0 > > > > > > > systemd-backlig-724 [006] ..... 9.079750: pwm_get: pwmchip0.0: period=4266537 duty_cycle=4266537 polarity=0 enabled=1 err=0 > > > > > > > $ > > > > > > > > > > > > > > > > > > > > > > > I didn't take a deeper dive in this driver combination, but here is a > > > > > > > > description about what *should* happen: > > > > > > > > > > > > > > > > You're talking about period in MHz, the PWM abstraction uses > > > > > > > > nanoseconds. So your summary translated to the PWM wording is (to the > > > > > > > > best of my understanding): > > > > > > > > > > > > > > > > 1. PWM backlight driver requests PWM with .period = 200 ns and > > > > > > > > .duty_cycle = 200 ns. > > > > > > > > > > > > > > > > 2. leds-qcom-lpg cannot pick 200 ns exactly and then chooses .period = > > > > > > > > 1000000000 / 4.26666 MHz = 234.375 ns > > > > > > > > > > > > > > > > 3. leds-qcom-lpg then determines setting for requested .duty_cycle > > > > > > > > based on .period = 200 ns which then ends up with something bogus. > > > > > > > > > > > > The trace looks better than what I expected. 2. is fine here because it > > > > > > seems when Sebastian wrote "driver requests PWM with 5 MHz period" that > > > > > > meant period = 5000000 ns. That was then rounded down to 4266537 ns. And > > > > > > the request for period = 5000000 ns + duty_cycle = 5000000 ns was > > > > > > serviced by configuring period = 4266537 ns + duty_cycle = 4266537 ns. > > > > > > So that's a 100 % relative duty configuration as intended. > > > > > > > > > > > > So just from the traces I don't spot a problem. Do these logs not match > > > > > > what actually happens on the signal? > > > > > > > > > > What I do not get is why do we expect 2 pwm_get() and 2 pwm_apply() > > > > > calls each time ? > > > > > > > > OK, so the second pwm_apply() is due to CONFIG_PWM_DEBUG. > > > > ack. This is done just for the tests implemented in CONFIG_PWM_DEBUG, as > > are the two pwm_get()s. > > > > > > But still, the first pwm_apply() requests duty cycle of 5MHz: > > > > 5 ms, yes. But it cannot give you 5 ms and so you get 4.266 ns. > > > > > > systemd-backlig-724 [006] ..... 9.079538: pwm_apply: pwmchip0.0: period=5000000 duty_cycle=5000000 polarity=0 enabled=1 err=0 > > > > > > > > So since the period is 4.26MHz, due to the knobs selected by the > > > > provider, this duty cycle will result in a PWM value that is above the > > > > selected resolution, as I already mentioned. > > > > "above the selected resolution"? Do you mean you don't get the exact > > value that you requested? > > I think I understand your point now. > > You expectation is that the provider would remap the entire range of the > period to whatever the HW can do. If I understand you correctly, that's right. For a given hardware there is a set of possible periods P. .apply() should pick max{ p ∈ P | p ≤ state->period }. And similar for duty_cycle: After choosing a possible period p ∈ P, there is a set D(p) of duty_cycles that the hardware can implement in combination to period p. .apply() should pick max{ d ∈ D(p) | d ≤ state->duty_cycle }. > So in this case, when 5ms is requested as duty cycle from consumer, the > provider will select the max value. Yes. > What the current implementation of the leds-qcom-lpg does is that will > expect a duty cycle request of up to 4.26ms. And according to you, even > if the consumer requests 5ms, the leds-qcom-lpg driver should write the > value of 255 (which is what the selected resolution allows (1 << 8) ) and > not compute a higher value. If the period is 4.26 ms, duty_cycle cannot be bigger than 4.26 ms. So yes, that's what the driver should do. > I think this is wrong though. The fact that the pwm generic framework > reports 5ms when it is actually 4.26ms should be considered wrong. After pwm_apply_might_sleep(mypwm, { .period = 5000000, .duty_cycle = 5000000, .enabled = true }), pwm_get_state() gives you 5000000 and pwm_get_state_hw() gives you 4266537. You could argue that the functions's names and semantic are not optimal. Changing that is hard, see my failed attempt in 01ccf903edd6 ("pwm: Let pwm_get_state() return the last implemented state") + 40a6b9a00930 ("Revert "pwm: Let pwm_get_state() return the last implemented state"") So I don't see how the PWM framework is wrong here. Depending on what value you want to get, pick pwm_get_state() or pwm_get_state_hw(). > For cases where the exact value of the duty cycle matters, this would > not even make sense. What is "this"? pwm_get_state() returning the last requested value? If you're interested in the last requested value, it does make sense. > Correct me if I'm wrong, but the pwm API should behave more like: > The consumer should ask for the closest period the HW can actually do > and then use that closest period from there on for every duty cycle > request. You can do that today using pwm_round_waveform_might_sleep() (however that needs some glue in the leds-qcom-lpg driver). And note that most in-kernel users don't care about exactness a lot. So the fire-and-forget approach is fine and it shouldn't be made more complicated for those. > This way, if the consumer initially wants 5ms but the provider > can do only 4.26ms instead, at least the consumer would be able to > correct its duty cycle requests based on what the HW says it can do. I agree that the consumer should be able to make an informed choice, and that was my focus when designing the waveform API. But I intend to not force that on (e.g.) the leds-pwm driver if that doesn't care about getting 4.26 ms or 5 ms. > > > On top of that, the duty cycle in debugfs is also reported as 5000000ns > > > when in fact it is 4266666ns, as the trace shows. > > > > Yes. Consider that a relict from the times when there was no > > pwm_get_state_hw(). Both values are interesting in different situations. > > So just telling the real parameters isn't the optimal way forward > > either. > > > > Something like the patch I showed in > > https://lore.kernel.org/all/7bcnckef23w6g47ll5l3bktygedrcfvr7fk3qjuq2swtoffhec@zs4w4tuh6qvm/ > > And this patchset only adds the info of actual value that the HW is actually doing. "only"? Yes, that's the intention of that patch. What should it do more? > So basically, the already existing state in this case will represent the > "desired" state. Yes, pwm->state tracks the state that was last passed to pwm_apply_might_sleep() (most of the time). > > would make you a bit luckier I guess. Feel free to polish that one a bit > > (e.g. by checking the return value of pwm_get_state_hw() and acting > > sensible in reply to it) and send a proper patch. (A Suggested-by for me > > is enough for such a patch, grab authorship yourself.) > > > > > > > Need to dig a bit further. > > > > > > > > > > But meanwhile, if the first pwm_apply() call goes all the way to the > > > > > provider, then the duty cycle value, when translated to the actual PWM > > > > > value that gets written to reg, will overflow. > > > > No it will not. The .duty_cycle value (also 5000000 ns) will reach the > > lowlevel PWM driver together with .period = 5000000 ns. Both are rounded > > down to 4266666ns. I see no overflow. > > Again, the consumer is being lied to. It expects 5ms and gets 4.26ms > instead. I see what you mean, but I don't agree. The semantic of pwm_apply_might_sleep() is: "Configure the state that is nearest to the passed state" (for some metric that defines "nearest"). The function returning 0 means: The hardware now has this nearest state. The semantic of pwm_get_state() is approximately: "What state was requested before?" So it will give you .period = 5000000 ns and .duty_cycle = 5000000 ns. The semantic of pwm_get_state_hs() is: "What state is the hardware in?" So it will give you .period = 4266666 ns and .duty_cycle = 4266666 ns. So there are no lies, just wrong expectations about the semantic of these functions. And if you think that pwm_apply_might_sleep() should fail when 5000000 ns is requested and it can only do 4266537 ns: Where should the line drawn that decides between "4977777 ns is still ok" and "4977777 ns is too far from 5000000 ns"? > Imagine a device that is controlled via PWM and needs exact duty cycle > values in ms, what would the consumer driver do in this case? Traditionally it would need some hardware specific extra information. Today it could work out the needed details with the waveform API functions (though this is hard because there are only two supported lowlevel drivers and no helper functions yet). > And to make things worse, when the consumer is asking for duty cycle of > 4ms while the period requested is 5ms (which would be 80%), the period > the provider will do is actually 4.26ms while the duty cycle would be > ~3.41ms, which if the pwm step (reg value) doesn't allow, it will probably > result in an actual value that is even further than what the consumer > is expecting. Where does ~3.41 ms come from? (I guess that's 0.8 * 4.26 ms.) Note that if you request .period = 5 ms and .duty_cycle = 4 ms, you get .period = 4.26 ms and the biggest duty_cycle not bigger than 4 ms that is possible with .period = 4.26 ms. So most likely not a 80% relative duty_cycle. > So I'm thinking maybe the pwm should probably even ask the provider > for what duty cycle it will provide based on provider's provided period > and then decide if the resulting duty cycle is what it really wants. Look into the waveform functions. The basic building blocks for what you want should be there. > IIRC, this is more in line with what the CCF (common clocks framework) > currently does. It does? There is clk_round_rate() but that is really hard to use because there are virtually no promises in that function. Consider you want a clock to run at 666666 Hz and clk_round_rate(yourclk, 666666) gives you 500000 Hz. What would you do? Even: What is the rate above 666666 Hz that is as good as 500000 Hz for your usecase? Is it 833332 Hz or 888888 Hz? And do you want 666666 Hz or 666666.666666667 Hz and how does that influence your search for the right clkrate? And it has the same problem as the pwm waveform functions: Just because clk_round_rate(yourclk, 666666) returned 500000 200 ms ago, it doesn't mean that if I ask for 666666 now the world didn't change. Best regards Uwe
Attachment:
signature.asc
Description: PGP signature