On 12/5/18 2:25 AM, james qian wang (Arm Technology China) wrote: > Signed-off-by: James (Qian) Wang <james.qian.wang@xxxxxxx> > --- > Documentation/gpu/drivers.rst | 1 + > Documentation/gpu/komeda-kms.rst | 483 +++++++++++++++++++++++++++++++ > 2 files changed, 484 insertions(+) > create mode 100644 Documentation/gpu/komeda-kms.rst Hi, I have some editing changes for you to consider, although I did have a few problems with reading the text. > diff --git a/Documentation/gpu/komeda-kms.rst b/Documentation/gpu/komeda-kms.rst > new file mode 100644 > index 000000000000..8af925ca0869 > --- /dev/null > +++ b/Documentation/gpu/komeda-kms.rst > @@ -0,0 +1,483 @@ > +.. SPDX-License-Identifier: GPL-2.0 > + > +============================== > + drm/komeda ARM display driver > +============================== > + > +The drm/komeda driver supports for the ARM display processor D71 and later driver supports the ARM > +IPs, this document is for giving a brief overview of driver design: how it IPs. This document gives a brief overview of the driver design: (although I hate using "IPs" like that.) > +works and why design it like that. > + > +Overview of D71 like display IPs > +================================ > + > +From D71, Arm display IP begins to adopt a flexible and modularized ARM > +architecture. A display pipeline is made up of multiple individual and > +functional pipeline stages called components, and every component has some > +specific capabilities that can give the flowed pipeline pixel data a > +particular processing. > + > +Typical D71 components: > + > +Layer > +----- > +Layer is the first pipeline stage, which is for preparing the pixel data > +for the next stage. It fetches the pixel from memory, decodes it if it's > +AFBC, rotates the source image, unpacks or converts YUV pixels to the device > +internal RGB pixels, then adjust the color_space of pixels if need. adjusts if needed. > + > +Scaler > +------ > +As its name, scaler is taking responsability for scaling, and D71 also As its name suggests, scaler takes (or has) responsibility for > +supports image enhancements by scaler. > +The usage of scaler is very flexible and can be connected to layer output > +for layer scaling, or connected to compositor and scale the whole display > +frame and then feed the output data into wb_layer which will then write it > +into memory. > + > +Compositor (compiz) > +------------------- > +Compositor is for blending multiple layers or pixel data flows into one > +single display frame, and its output frame can be fed into post image > +processor and then on the monitor or fed into wb_layer and written to memory > +at the same time. And user also can insert a scaler between compositor and > +wb_layer to down scale the display frame first and then writing to memory. and then write to memory. > + > +Writeback Layer (wb_layer) > +-------------------------- > +Writeback layer do the opposite things of Layer, Which connect to compiz and try does which > +to write the composition result to memory. > + > +Post image processor (improc) > +----------------------------- > +Post image processor is for adjusting frame data like gamma and color space > +to fit the requirements of the monitor. > + > +Timing controller (timing_ctrlr) > +-------------------------------- > +Final stage of display pipeline, Timing controller is not for the pixel > +handling, but only for controlling the display timing. > + > +Merger > +------ > +D71 scaler mostly only has half the horizontal input/output capabilities compare compared > +with Layer, Like if Layer supports 4K input size, the scaler only can supports like support > +2K input/output in some time. To achieve the fully frame scaling, D71 introduce full introduces > +Layer Split, which split the whole image to two half part and feed them to two splits parts and feeds > +Layers A and B, and do the scaling independently, after scaling the result need does independently. After scaling the result needs > +to be feed to merger to merge two part image, and then output to compiz. fed > + > +Splitter > +-------- > +Similar to Layer Split, but Splitter is used for writeback, which split the splits > +compiz result to two part and then feed them to two scaler. parts and then feeds them to two scalers. > + > +Possible D71 Pipeline usage > +=========================== > + > +Benefit from the modularized architecture, D71 pipelines can be easily adjusted Benefitting from > +to fit different usages, following are some typical pipeline data flow usages. Following > +configurations: > + > +Single pipeline data flow > +------------------------- > + > +.. kernel-render:: DOT > + :alt: Single pipeline digraph > + :caption: Single pipeline data flow > + > + digraph single_ppl { > + rankdir=LR; > + > + subgraph { > + "Memory"; > + "Monitor"; > + } > + > + subgraph cluster_pipeline { > + style=dashed > + node [shape=box] > + { > + node [bgcolor=grey style=dashed] > + "Scaler-0"; > + "Scaler-1"; > + "Scaler-0/1" > + } > + > + node [bgcolor=grey style=filled] > + "Layer-0" -> "Scaler-0" > + "Layer-1" -> "Scaler-0" > + "Layer-2" -> "Scaler-1" > + "Layer-3" -> "Scaler-1" > + > + "Layer-0" -> "Compiz" > + "Layer-1" -> "Compiz" > + "Layer-2" -> "Compiz" > + "Layer-3" -> "Compiz" > + "Scaler-0" -> "Compiz" > + "Scaler-1" -> "Compiz" > + > + "Compiz" -> "Scaler-0/1" -> "Wb_layer" > + "Compiz" -> "Improc" -> "Timing Controller" > + } > + > + "Wb_layer" -> "Memory" > + "Timing Controller" -> "Monitor" > + } > + > +Dual pipeline with Slave enabled > +-------------------------------- > + > +If pipeline_B is free, D71 supports redirect its compositor output to D71 supports redirecting > +pipeline_A as an input of compositor of pipeline_A, then pipeline_B doesn't > +have its output and work as a slave of pipeline_A. confusing! > +NOTE: Since the compiz component doesn't output alpha value, the slave pipeline > +only can be used for bottom layers composition. > + > +.. kernel-render:: DOT > + :alt: Slave pipeline digraph > + :caption: Slave pipeline enabled data flow > + > + digraph slave_ppl { > + rankdir=LR; > + > + subgraph { > + "Memory"; > + "Monitor"; > + } > + node [shape=box] > + subgraph cluster_pipeline_slave { > + style=dashed > + label="Slave Pipeline_B" > + node [shape=box] > + { > + node [bgcolor=grey style=dashed] > + "Slave.Scaler-0"; > + "Slave.Scaler-1"; > + } > + > + node [bgcolor=grey style=filled] > + "Slave.Layer-0" -> "Slave.Scaler-0" > + "Slave.Layer-1" -> "Slave.Scaler-0" > + "Slave.Layer-2" -> "Slave.Scaler-1" > + "Slave.Layer-3" -> "Slave.Scaler-1" > + > + "Slave.Layer-0" -> "Slave.Compiz" > + "Slave.Layer-1" -> "Slave.Compiz" > + "Slave.Layer-2" -> "Slave.Compiz" > + "Slave.Layer-3" -> "Slave.Compiz" > + "Slave.Scaler-0" -> "Slave.Compiz" > + "Slave.Scaler-1" -> "Slave.Compiz" > + } > + > + subgraph cluster_pipeline_master { > + style=dashed > + label="Master Pipeline_A" > + node [shape=box] > + { > + node [bgcolor=grey style=dashed] > + "Scaler-0"; > + "Scaler-1"; > + "Scaler-0/1" > + } > + > + node [bgcolor=grey style=filled] > + "Layer-0" -> "Scaler-0" > + "Layer-1" -> "Scaler-0" > + "Layer-2" -> "Scaler-1" > + "Layer-3" -> "Scaler-1" > + > + "Slave.Compiz" -> "Compiz" > + "Layer-0" -> "Compiz" > + "Layer-1" -> "Compiz" > + "Layer-2" -> "Compiz" > + "Layer-3" -> "Compiz" > + "Scaler-0" -> "Compiz" > + "Scaler-1" -> "Compiz" > + > + "Compiz" -> "Scaler-0/1" -> "Wb_layer" > + "Compiz" -> "Improc" -> "Timing Controller" > + } > + > + "Wb_layer" -> "Memory" > + "Timing Controller" -> "Monitor" > + } > + > +Sub-pipelines for input and output > +---------------------------------- > + > +A complete display pipeline can be easily dividied into three sub-pipelines divided > +according to the in/out usage. > + > +Layer(input) pipeline > +~~~~~~~~~~~~~~~~~~~~~ > + > +.. kernel-render:: DOT > + :alt: Layer data digraph > + :caption: Layer (input) data flow > + > + digraph layer_data_flow { > + rankdir=LR; > + node [shape=box] > + > + { > + node [bgcolor=grey style=dashed] > + "Scaler-n"; > + } > + > + "Layer-n" -> "Scaler-n" -> "Compiz" > + } > + > +.. kernel-render:: DOT > + :alt: Layer Split digraph > + :caption: Layer Split pipeline > + > + digraph layer_data_flow { > + rankdir=LR; > + node [shape=box] > + > + "Layer-0/1" -> "Scaler-0" -> "Merger" > + "Layer-2/3" -> "Scaler-1" -> "Merger" > + "Merger" -> "Compiz" > + } > + > +Writeback(output) pipeline > +~~~~~~~~~~~~~~~~~~~~~~~~~~ > +.. kernel-render:: DOT > + :alt: writeback digraph > + :caption: Writeback(output) data flow > + > + digraph writeback_data_flow { > + rankdir=LR; > + node [shape=box] > + > + { > + node [bgcolor=grey style=dashed] > + "Scaler-n"; > + } > + > + "Compiz" -> "Scaler-n" -> "Wb_layer" > + } > + > +.. kernel-render:: DOT > + :alt: split writeback digraph > + :caption: Writeback(output) Split data flow > + > + digraph writeback_data_flow { > + rankdir=LR; > + node [shape=box] > + > + "Compiz" -> "Splitter" > + "Splitter" -> "Scaler-0" -> "Merger" > + "Splitter" -> "Scaler-1" -> "Merger" > + "Merger" -> "Wb_layer" > + } > + > +Display output pipeline > +~~~~~~~~~~~~~~~~~~~~~~~ > +.. kernel-render:: DOT > + :alt: display digraph > + :caption: display output data flow > + > + digraph single_ppl { > + rankdir=LR; > + node [shape=box] > + > + "Compiz" -> "Improc" -> "Timing Controller" > + } > + > +In the following section we'll see these three sub-pipelines will be handled > +by KMS-plane/wb_conn/crtc respectively. > + > +Komeda Resource abstraction > +=========================== > + > +struct komeda_pipeline/component > +-------------------------------- > + > +To fully utilize and easily access/configure the HW, driver side also use a the driver side also uses a > +similar architecture: Pipeline/Component to describe the HW features and > +capabilities. and a specific component includes two part: capabilities, parts: > + > +- Data flow controlling. > +- Specific component capabilities and features. > + > +So driver defines common header struct komeda_component to describe the data So the driver defines a common header > +flow controlling. and all specific component are subclass of this base structure. flow control and all specific components are a subclass of > + > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_pipeline.h > + :internal: > + > +Resource discovery and initialization > +===================================== > + > +Pipeline and component are used to describe how to handle the pixel data. > +We still need a @struct komeda_dev to describe the whole view of the device, and > +the control-abilites of device. > + > +Now we have &komeda_dev, &komeda_pipeline, &komeda_component, but how to fill dev Maybe: We have &komeda_dev, &komeda_pipeline, and &komeda_component. Now to fill the device > +with pipelines. Since komeda is not only for D71 but also intended for later IPs, > +of course we’d better share as much as possible between different IPs, to different IPs. To > +achieve it split komeda device into two layers: CORE and CHIP. achieve this, split the > + > +- CORE: for common features and capabilities handling. > +- CHIP: for register program and HW specific feature (limitition) handling. (limitation) > + > +CORE access to CHIP by three chip func: CORE can access CHIP by three chip function structures: > + > +- struct komeda_dev_funcs > +- struct komeda_pipeline_funcs > +- struct komeda_component_funcs > + > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_dev.h > + :internal: > + > +Format handling > +=============== > + > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_format_caps.h > + :internal: > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_framebuffer.h > + :internal: > + > +Attach komeda_dev to DRM-KMS > +============================ > + > +Komeda abstracts resources by pipeline/component, but DRM-KMS uses > +crtc/plane/connector. one KMS-obj can not represent only one single component, One cannot [preferably] > +since the requirements of a single KMS object cannot simply be achieved by a > +single component, usually that needs multiple components to fit the requirement. > +Like set mode, gamma, ctm for KMS all target on CRTC-obj, but komeda needs > +compiz, improc and timing_ctrlr to work together to fit these requirements. > +And a KMS-Plane may requires multiple komeda resources: layer/scaler/compiz. require > + > +So, One KMS-Obj represents a sub-pipeline of komeda resources. one > + > +- Plane: `Layer(input) pipeline`_ > +- Wb_connector: `Writeback(output) pipeline`_ > +- Crtc: `Display output pipeline`_ > + > +So, for komeda, we treat KMS crtc/plane/connector as users of pipeline and > +component, and one time a pipeline/component only can be used by one user. and at any one time > +And pipeline/component will be treated as private object of DRM-KMS, the state of DRM-KMS; the state > +will be managed drm_atomic_state as well. will be managed by drm_atomic_state as well. > + > +How to map plane to Layer(input) pipeline > +----------------------------------------- > + > +Komeda has mutiple Layer input pipelines, see: > +- `Single pipeline data flow`_ > +- `Dual pipeline with Slave enabled`_ > + > +the easist way is binding a plane to a fixed Layer pipeline. but consider the The easiest pipeline, > +komeda capabilities: > + > +- Layer Split, See `Layer(input) pipeline`_ > + > + Which actually handles a image by two Layer pipelines. if one plane only If > + represents a fixed Layer, then the Layer_Split need to be handled in user needs > + mode, for configure the pipeline as Layer_Split, user need to so much mode: to configure the pipeline as Layer_Split, the user needs much > + details of HW, and pass down lots of private properties. detail of the HW, and must pass down lots of private properties. > + But if handle it in kernel the logic will be smooth and simple. But if handled in the kernel, the logic will be smooth and simple. So above, I tried to write what I thought you meant, but saying that Layer_Split needs to be handled in user mode ... and then saying that if it is handled in the kernel, it would be smooth and simple -- that seems to be contradictory to me, so I guess I didn't understand it very well. > + > +- Slave pipeline, See `Dual pipeline with Slave enabled`_ > + > + Since the compiz component doesn't output alpha value, the slave pipeline > + only can be used for bottom layers composition, komeda driver wants to hide composition. The komeda driver wants to hide > + this limitation to user, the way is pickup suitable Layer according to to the user. The way to do this is to pick a suitable Layer according to Plane->zpos. > + Plane->zpos. > + > + Slave pipeline capabilities of komeda also means the plane in komeda is CRTC > + shareable. assign real Layer to plane in kernal according to the real HW Assign kernel > + usage and state can also easy the user driver for the sharable plane. ??confused!! shareable > + > +So for komeda, the KMS-plane doesn't represent a fixed komeda layer pipeline, > +but mutiple Layers with same capabilities, komeda will select one or more Layers multiple capabilities. > +to fit the requirement of one KMS-plane. > + > +Make component/pipeline to be drm_private_obj > +--------------------------------------------- > + > +Add :c:type:`drm_private_obj` to :c:type:`komeda_component`, :c:type:`komeda_pipeline` > + > +.. code-block:: c > + > + struct komeda_component { > + struct drm_private_obj obj; > + … Prefer ASCII "...". > + } > + > + struct komeda_pipeline { > + struct drm_private_obj obj; > + … same. > + } > + > +Tracking component_state/pipeline_state by drm_atomic_state > +----------------------------------------------------------- > + > +Add :c:type:`drm_private_state` and user to :c:type:`komeda_component_state`, > +:c:type:`komeda_pipeline_state` > + > +.. code-block:: c > + > + struct komeda_component_state { > + struct drm_private_state obj; > + void *binding_user; > + … same. > + } > + > + struct komeda_pipeline_state { > + struct drm_private_state obj; > + struct drm_crtc *crtc; > + ... like that :) > + } > + > +komeda component validation > +--------------------------- > + > +Komeda has multiple types of components, but the process of validation are is > +similar, usually include following steps: including the following steps: > + > +.. code-block:: c > + > + int komeda_xxxx_validate(struct komeda_component_xxx xxx_comp, > + struct komeda_component_output *input_dflow, > + struct drm_plane/crtc/connector *user, > + struct drm_plane/crtc/connector_state, *user_state) > + { > + setup 1: check if component is needed, like the scaler is optional depend depending > + on the user_state, if unneeded, just return, the caller will put user_state; if unneeded, just return, and the caller will put > + the data flow into next stage. > + Setup 2: check user_state with component features and capabilities to see > + if can meet, if not return fail. if requirements can be met; if not, return fail. > + Setup 3: get component_state from drm_atomic_state, and try set user to try to set > + component, fail if component has been assigned to another user already. component; > + Setup 3: configure the component_state, like set its input component, > + convert user_state to component specific state. > + Setup 4: adjust the input_dflow and prepare it for the next stage. > + } > + > +komeda_kms Abstraction > +---------------------- > + > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_kms.h > + :internal: > + > +komde_kms Functions > +------------------- > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_crtc.c > + :internal: > +.. kernel-doc:: drivers/gpu/drm/arm/display/komeda/komeda_plane.c > + :internal: > + > +Build komeda to be a linux module driver Linux > +======================================== > + > +Now we have two devices: > + > +- komeda_dev is for describing the real display hardware. > +- komeda_kms_dev is for attaching or connecting komeda_dev to DRM-KMS. > + > +all komeda operations are supplied or operated by komeda_dev or komeda_kms_dev, All > +the module driver is only a simple wrapper to pass the linux command Linux > +(probe/remove/pm) into komeda_dev or komeda_kms_dev. > HTH. -- ~Randy