Re: Questions over DSI within DRM.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Dave,

(Expanding the CC list a bit)

On Fri, Jul 02, 2021 at 12:03:31PM +0100, Dave Stevenson wrote:
> Hi All
> 
> I'm trying to get DSI devices working reliably on the Raspberry Pi,
> but I'm hitting a number of places where it isn't clear as to the
> expected behaviour within DRM.

Not a surprise. I dread reading the rest of this e-mail though :-)

> Power on state. Many devices want the DSI clock and/or data lanes in
> LP-11 state when they are powered up.

When they are powered up, or when they are enabled ?

> With the normal calling sequence of:
> - panel/bridge pre_enable calls from connector towards the encoder.
> - encoder enable which also enables video.
> - panel/bridge enable calls from encoder to connector.
> there is no point at which the DSI tx is initialised but not
> transmitting video. What DSI states are expected to be adopted at each
> point?

That's undefined I'm afraid, and it should be documented. The upside is
that you can propose the behaviour that you need :-)

> On a similar theme, some devices want the clock lane in HS mode early
> so they can use it in place of an external oscillator, but the data
> lanes still in LP-11. There appears to be no way for the
> display/bridge to signal this requirement or it be achieved.

You're right. A loooong time ago, the omapdrm driver had an internal
infrastructure that didn't use drm_bridge or drm_panel and instead
required omapdrm-specific drivers for those components. It used to model
the display pipeline in a different way than drm_bridge, with the sync
explicitly setting the source state. A DSI sink could thus control its
enable sequence, interleaving programming of the sink with control of
the source.

Migrating omapdrm to the drm_bridge model took a really large effort,
which makes me believe that transitioning the whole subsystem to
sink-controlled sources would be close to impossible. We could add
DSI-specific operations, or add another enable bridge operation
(post_pre_enable ? :-D). Neither would scale, but it may be enough.

> host_transfer calls can supposedly be made at any time, however unless
> MIPI_DSI_MSG_USE_LPM is set in the message then we're meant to send it
> in high speed mode. If this is before a mode has been set, what
> defines the link frequency parameters at this point? Adopting a random
> default sounds like a good way to get undefined behaviour.
> 
> DSI burst mode needs to set the DSI link frequency independently of
> the display mode. How is that meant to be configured? I would have
> expected it to come from DT due to link frequency often being chosen
> based on EMC restrictions, but I don't see such a thing in any
> binding.

Undefined too. DSI support was added to DRM without any design effort,
it's more a hack than a real solution. The issue with devices that can
be controlled over both DSI and I2C is completely unhandled. So far
nobody has really cared about implementing DSI right as far as I can
tell.

> As a follow on, bridge devices can support burst mode (eg TI's
> SN65DSI83 that's just been merged), so it needs to know the desired
> panel timings for the output side of the bridge, but the DSI link
> timings to set up the bridge's PLL. What's the correct way for
> signalling that? drm_crtc_state->adjusted_mode vs
> drm_crtc_state->mode? Except mode is userspace's request, not what has
> been validated/updated by the panel/bridge.

adjusted_mode is also a bit of a hack, it solves very specific issues,
and its design assumes a single encoder in the chain with no extra
bridges. We should instead add modes to the bridge state, and negotiate
modes along the pipeline the same way we negotiate formats.

> vc4 has constraints that the DSI host interface is fed off an integer
> divider from a typically 3GHz clock, so the host interface needs to
> signal that burst mode is in use even if the panel/bridge doesn't need
> to run in burst mode. (This does mean that displays that require a
> very precise link frequency can not be supported).
> It currently updates the adjusted_mode via drm_encoder_helper_funcs
> mode_fixup, but is that the correct thing to do, or is there a better
> solution?
> I'd have expected the DSI tx to be responsible for configuring burst
> mode parameters anyway, so the mechanism required would seem to be
> just the normal approach for adopting burst mode if that is defined.
> 
> Some DSI host interfaces are implemented as bridges, others are
> encoders. Pro's and con's of each? I suspect I'm just missing the
> history here.

It's indeed history. drm_encoder can't go away as it has been erronously
exposed to userspace, but going forward, everything should be a bridge.
The drm_encoder will still be required, but should just be a dummy,
representing the chain of bridges.

> When it comes to the MIPI_DSI_MODE_* flags, which ones are mutually
> exclusive, or are assumed based on others? Does a burst mode DSI sink
> set both MIPI_DSI_MODE_VIDEO and MIPI_DSI_MODE_VIDEO_BURST, or just
> the latter?
> Presumably !MIPI_DSI_MODE_VIDEO signals the of use command mode for
> conveying video. So looking at panel-ilitek-ili9881c where it sets
> just MIPI_DSI_MODE_VIDEO_SYNC_PULSE means command mode video with sync
> pulses? That sounds unlikely.

I haven't looked at that, I'm afraid I don't know.

> I have looked for any information that covers this, but failed to find
> such, hence calling on all your expertise.

I'm sorry for the lack of solutions to your issues. I can try to help
solving them though.

-- 
Regards,

Laurent Pinchart



[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux