On Fri, Jul 2, 2021 at 10:12 PM Laurent Pinchart <laurent.pinchart@xxxxxxxxxxxxxxxx> wrote: > > Hi Dave, > > (Expanding the CC list a bit) > > On Fri, Jul 02, 2021 at 12:03:31PM +0100, Dave Stevenson wrote: > > Hi All > > > > I'm trying to get DSI devices working reliably on the Raspberry Pi, > > but I'm hitting a number of places where it isn't clear as to the > > expected behaviour within DRM. > > Not a surprise. I dread reading the rest of this e-mail though :-) > > > Power on state. Many devices want the DSI clock and/or data lanes in > > LP-11 state when they are powered up. > > When they are powered up, or when they are enabled ? > > > With the normal calling sequence of: > > - panel/bridge pre_enable calls from connector towards the encoder. > > - encoder enable which also enables video. > > - panel/bridge enable calls from encoder to connector. > > there is no point at which the DSI tx is initialised but not > > transmitting video. What DSI states are expected to be adopted at each > > point? > > That's undefined I'm afraid, and it should be documented. The upside is > that you can propose the behaviour that you need :-) > > > On a similar theme, some devices want the clock lane in HS mode early > > so they can use it in place of an external oscillator, but the data > > lanes still in LP-11. There appears to be no way for the > > display/bridge to signal this requirement or it be achieved. > > You're right. A loooong time ago, the omapdrm driver had an internal > infrastructure that didn't use drm_bridge or drm_panel and instead > required omapdrm-specific drivers for those components. It used to model > the display pipeline in a different way than drm_bridge, with the sync > explicitly setting the source state. A DSI sink could thus control its > enable sequence, interleaving programming of the sink with control of > the source. > > Migrating omapdrm to the drm_bridge model took a really large effort, > which makes me believe that transitioning the whole subsystem to > sink-controlled sources would be close to impossible. We could add > DSI-specific operations, or add another enable bridge operation > (post_pre_enable ? :-D). Neither would scale, but it may be enough. > > > host_transfer calls can supposedly be made at any time, however unless > > MIPI_DSI_MSG_USE_LPM is set in the message then we're meant to send it > > in high speed mode. If this is before a mode has been set, what > > defines the link frequency parameters at this point? Adopting a random > > default sounds like a good way to get undefined behaviour. > > > > DSI burst mode needs to set the DSI link frequency independently of > > the display mode. How is that meant to be configured? I would have > > expected it to come from DT due to link frequency often being chosen > > based on EMC restrictions, but I don't see such a thing in any > > binding. > > Undefined too. DSI support was added to DRM without any design effort, > it's more a hack than a real solution. The issue with devices that can > be controlled over both DSI and I2C is completely unhandled. So far > nobody has really cared about implementing DSI right as far as I can > tell. > > > As a follow on, bridge devices can support burst mode (eg TI's > > SN65DSI83 that's just been merged), so it needs to know the desired > > panel timings for the output side of the bridge, but the DSI link > > timings to set up the bridge's PLL. What's the correct way for > > signalling that? drm_crtc_state->adjusted_mode vs > > drm_crtc_state->mode? Except mode is userspace's request, not what has > > been validated/updated by the panel/bridge. > > adjusted_mode is also a bit of a hack, it solves very specific issues, > and its design assumes a single encoder in the chain with no extra > bridges. We should instead add modes to the bridge state, and negotiate > modes along the pipeline the same way we negotiate formats. > > > vc4 has constraints that the DSI host interface is fed off an integer > > divider from a typically 3GHz clock, so the host interface needs to > > signal that burst mode is in use even if the panel/bridge doesn't need > > to run in burst mode. (This does mean that displays that require a > > very precise link frequency can not be supported). > > It currently updates the adjusted_mode via drm_encoder_helper_funcs > > mode_fixup, but is that the correct thing to do, or is there a better > > solution? > > I'd have expected the DSI tx to be responsible for configuring burst > > mode parameters anyway, so the mechanism required would seem to be > > just the normal approach for adopting burst mode if that is defined. > > > > Some DSI host interfaces are implemented as bridges, others are > > encoders. Pro's and con's of each? I suspect I'm just missing the > > history here. > > It's indeed history. drm_encoder can't go away as it has been erronously > exposed to userspace, but going forward, everything should be a bridge. > The drm_encoder will still be required, but should just be a dummy, > representing the chain of bridges. Just to add further on this with some example, thought it might help. NWL DSI initially started with encoder and eventually become bridge driven in mainline. https://lore.kernel.org/lkml/13373313.BzCyiC4ED7@jernej-laptop/t/ Exynos DSI becoming Samsung MIPI DSIM bridge https://patchwork.kernel.org/project/dri-devel/cover/20210704090230.26489-1-jagan@xxxxxxxxxxxxxxxxxxxx/ Thanks, Jagan.