Maxime Ripard <maxime@xxxxxxxxxx> 于2021年4月30日周五 下午5:35写道: > > Hi, > > On Sun, Apr 25, 2021 at 08:36:07PM +0800, Kevin Tang wrote: > > Adds dsi host controller support for the Unisoc's display subsystem. > > Adds dsi phy support for the Unisoc's display subsystem. > > Only MIPI DSI Displays supported, DP/TV/HMDI will be support > > in the feature. > > > > v1: > > - Remove dphy and dsi graph binding, merge the dphy driver into the dsi. > > > > v2: > > - Use drm_xxx to replace all DRM_XXX. > > - Use kzalloc to replace devm_kzalloc for sprd_dsi structure init. > > > > v4: > > - Use drmm_helpers to allocate encoder. > > - Move allocate encoder and connector to bind function. > > > > v5: > > - Drop the dsi ip file prefix. > > - Fix the checkpatch warnings. > > - Add Signed-off-by for dsi&dphy patch. > > - Use the mode_flags of mipi_dsi_device to setup crtc DPI and EDPI > > mode. > > > > Cc: Orson Zhai <orsonzhai@xxxxxxxxx> > > Cc: Chunyan Zhang <zhang.lyra@xxxxxxxxx> > > Signed-off-by: Kevin Tang <kevin.tang@xxxxxxxxxx> > > Output from checkpatch: > total: 0 errors, 3 warnings, 100 checks, 4207 lines checked > > > --- > > drivers/gpu/drm/sprd/Kconfig | 1 + > > drivers/gpu/drm/sprd/Makefile | 6 +- > > drivers/gpu/drm/sprd/dsi_ctrl.c | 794 ++++++++++++++ > > drivers/gpu/drm/sprd/dsi_ctrl.h | 1475 ++++++++++++++++++++++++++ > > drivers/gpu/drm/sprd/dsi_ctrl_ppi.c | 157 +++ > > drivers/gpu/drm/sprd/dsi_ctrl_ppi.h | 26 + > > drivers/gpu/drm/sprd/megacores_pll.c | 317 ++++++ > > drivers/gpu/drm/sprd/megacores_pll.h | 146 +++ > > drivers/gpu/drm/sprd/sprd_dpu.c | 17 + > > drivers/gpu/drm/sprd/sprd_drm.c | 1 + > > drivers/gpu/drm/sprd/sprd_drm.h | 1 + > > drivers/gpu/drm/sprd/sprd_dsi.c | 1124 ++++++++++++++++++++ > > drivers/gpu/drm/sprd/sprd_dsi.h | 107 ++ > > 13 files changed, 4171 insertions(+), 1 deletion(-) > > create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl.c > > create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl.h > > create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl_ppi.c > > create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl_ppi.h > > create mode 100644 drivers/gpu/drm/sprd/megacores_pll.c > > create mode 100644 drivers/gpu/drm/sprd/megacores_pll.h > > create mode 100644 drivers/gpu/drm/sprd/sprd_dsi.c > > create mode 100644 drivers/gpu/drm/sprd/sprd_dsi.h > > > > diff --git a/drivers/gpu/drm/sprd/Kconfig b/drivers/gpu/drm/sprd/Kconfig > > index 37762c333..3edeaeca0 100644 > > --- a/drivers/gpu/drm/sprd/Kconfig > > +++ b/drivers/gpu/drm/sprd/Kconfig > > @@ -5,6 +5,7 @@ config DRM_SPRD > > select DRM_GEM_CMA_HELPER > > select DRM_KMS_CMA_HELPER > > select DRM_KMS_HELPER > > + select DRM_MIPI_DSI > > select VIDEOMODE_HELPERS > > help > > Choose this option if you have a Unisoc chipset. > > diff --git a/drivers/gpu/drm/sprd/Makefile b/drivers/gpu/drm/sprd/Makefile > > index ab12b95e6..d49f4977b 100644 > > --- a/drivers/gpu/drm/sprd/Makefile > > +++ b/drivers/gpu/drm/sprd/Makefile > > @@ -1,4 +1,8 @@ > > # SPDX-License-Identifier: GPL-2.0 > > > > obj-y := sprd_drm.o \ > > - sprd_dpu.o > > + sprd_dpu.o \ > > + sprd_dsi.o \ > > + dw_dsi_ctrl.o \ > > + dw_dsi_ctrl_ppi.o \ > > + megacores_pll.o > > diff --git a/drivers/gpu/drm/sprd/dsi_ctrl.c b/drivers/gpu/drm/sprd/dsi_ctrl.c > > new file mode 100644 > > index 000000000..7eccf9654 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/dsi_ctrl.c > > @@ -0,0 +1,794 @@ > > +// SPDX-License-Identifier: GPL-2.0 > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#include <linux/io.h> > > +#include <linux/init.h> > > +#include <linux/delay.h> > > +#include <linux/module.h> > > + > > +#include "dsi_ctrl.h" > > + > > +/* > > + * Modify power status of DSI Host core > > + */ > > +void dsi_power_enable(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(enable, ®->SOFT_RESET); > > +} > > +/* > > + * Enable/disable DPI video mode > > + */ > > +void dsi_video_mode(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(0, ®->DSI_MODE_CFG); > > +} > > +/* > > + * Enable command mode (Generic interface) > > + */ > > +void dsi_cmd_mode(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(1, ®->DSI_MODE_CFG); > > +} > > + > > +bool dsi_is_cmd_mode(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + return readl(®->DSI_MODE_CFG); > > +} > > +/* > > + * Configure the read back virtual channel for the generic interface > > + */ > > +void dsi_rx_vcid(struct dsi_context *ctx, u8 vc) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x1C virtual_channel_id; > > + > > + virtual_channel_id.val = readl(®->VIRTUAL_CHANNEL_ID); > > + virtual_channel_id.bits.gen_rx_vcid = vc; > > + > > + writel(virtual_channel_id.val, ®->VIRTUAL_CHANNEL_ID); > > +} > > +/* > > + * Write the DPI video virtual channel destination > > + */ > > +void dsi_video_vcid(struct dsi_context *ctx, u8 vc) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x1C virtual_channel_id; > > + > > + virtual_channel_id.val = readl(®->VIRTUAL_CHANNEL_ID); > > + virtual_channel_id.bits.video_pkt_vcid = vc; > > + > > + writel(virtual_channel_id.val, ®->VIRTUAL_CHANNEL_ID); > > +} > > +/* > > + * Set DPI video mode type (burst/non-burst - with sync pulses or events) > > + */ > > +void dsi_dpi_video_burst_mode(struct dsi_context *ctx, int mode) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x38 vid_mode_cfg; > > + > > + vid_mode_cfg.val = readl(®->VID_MODE_CFG); > > + vid_mode_cfg.bits.vid_mode_type = mode; > > + > > + writel(vid_mode_cfg.val, ®->VID_MODE_CFG); > > +} > > +/* > > + * Set DPI video color coding > > + */ > > +void dsi_dpi_color_coding(struct dsi_context *ctx, int coding) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x20 dpi_video_format; > > + > > + dpi_video_format.val = readl(®->DPI_VIDEO_FORMAT); > > + dpi_video_format.bits.dpi_video_mode_format = coding; > > + > > + writel(dpi_video_format.val, ®->DPI_VIDEO_FORMAT); > > +} > > +/* > > + * Configure the Horizontal Line time > > + * param "byte_cycle" taken to transmit the total of the horizontal line > > + */ > > +void dsi_dpi_hline_time(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x2C video_line_time; > > + > > + video_line_time.val = readl(®->VIDEO_LINE_TIME); > > + video_line_time.bits.video_line_time = byte_cycle; > > + > > + writel(video_line_time.val, ®->VIDEO_LINE_TIME); > > +} > > +/* > > + * Configure the Horizontal back porch time > > + * param "byte_cycle" taken to transmit the horizontal back porch > > + */ > > +void dsi_dpi_hbp_time(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x28 video_line_hblk_time; > > + > > + video_line_hblk_time.val = readl(®->VIDEO_LINE_HBLK_TIME); > > + video_line_hblk_time.bits.video_line_hbp_time = byte_cycle; > > + > > + writel(video_line_hblk_time.val, ®->VIDEO_LINE_HBLK_TIME); > > +} > > +/* > > + * Configure the Horizontal sync time, > > + * param "byte_cycle" taken to transmit the horizontal sync > > + */ > > +void dsi_dpi_hsync_time(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x28 video_line_hblk_time; > > + > > + video_line_hblk_time.val = readl(®->VIDEO_LINE_HBLK_TIME); > > + video_line_hblk_time.bits.video_line_hsa_time = byte_cycle; > > + > > + writel(video_line_hblk_time.val, ®->VIDEO_LINE_HBLK_TIME); > > +} > > +/* > > + * Configure the vertical active lines of the video stream > > + */ > > +void dsi_dpi_vact(struct dsi_context *ctx, u16 lines) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x34 video_active_lines; > > + > > + video_active_lines.val = readl(®->VIDEO_VACTIVE_LINES); > > + video_active_lines.bits.vactive_lines = lines; > > + > > + writel(video_active_lines.val, ®->VIDEO_VACTIVE_LINES); > > +} > > + > > +void dsi_dpi_vfp(struct dsi_context *ctx, u16 lines) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x30 video_vblk_lines; > > + > > + video_vblk_lines.val = readl(®->VIDEO_VBLK_LINES); > > + video_vblk_lines.bits.vfp_lines = lines; > > + > > + writel(video_vblk_lines.val, ®->VIDEO_VBLK_LINES); > > +} > > + > > +void dsi_dpi_vbp(struct dsi_context *ctx, u16 lines) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x30 video_vblk_lines; > > + > > + video_vblk_lines.val = readl(®->VIDEO_VBLK_LINES); > > + video_vblk_lines.bits.vbp_lines = lines; > > + > > + writel(video_vblk_lines.val, ®->VIDEO_VBLK_LINES); > > +} > > + > > +void dsi_dpi_vsync(struct dsi_context *ctx, u16 lines) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x30 video_vblk_lines; > > + > > + video_vblk_lines.val = readl(®->VIDEO_VBLK_LINES); > > + video_vblk_lines.bits.vsa_lines = lines; > > + > > + writel(video_vblk_lines.val, ®->VIDEO_VBLK_LINES); > > +} > > + > > +void dsi_dpi_hporch_lp_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x38 vid_mode_cfg; > > + > > + vid_mode_cfg.val = readl(®->VID_MODE_CFG); > > + > > + vid_mode_cfg.bits.lp_hfp_en = enable; > > + vid_mode_cfg.bits.lp_hbp_en = enable; > > + > > + writel(vid_mode_cfg.val, ®->VID_MODE_CFG); > > +} > > +/* > > + * Enable return to low power mode inside vertical active lines periods when > > + * timing allows > > + */ > > +void dsi_dpi_vporch_lp_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x38 vid_mode_cfg; > > + > > + vid_mode_cfg.val = readl(®->VID_MODE_CFG); > > + > > + vid_mode_cfg.bits.lp_vact_en = enable; > > + vid_mode_cfg.bits.lp_vfp_en = enable; > > + vid_mode_cfg.bits.lp_vbp_en = enable; > > + vid_mode_cfg.bits.lp_vsa_en = enable; > > + > > + writel(vid_mode_cfg.val, ®->VID_MODE_CFG); > > +} > > +/* > > + * Enable FRAME BTA ACK > > + */ > > +void dsi_dpi_frame_ack_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x38 vid_mode_cfg; > > + > > + vid_mode_cfg.val = readl(®->VID_MODE_CFG); > > + vid_mode_cfg.bits.frame_bta_ack_en = enable; > > + > > + writel(vid_mode_cfg.val, ®->VID_MODE_CFG); > > +} > > +/* > > + * Write no of chunks to core - taken into consideration only when multi packet > > + * is enabled > > + */ > > +void dsi_dpi_chunk_num(struct dsi_context *ctx, u16 num) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x24 video_pkt_config; > > + > > + video_pkt_config.val = readl(®->VIDEO_PKT_CONFIG); > > + video_pkt_config.bits.video_line_chunk_num = num; > > + > > + writel(video_pkt_config.val, ®->VIDEO_PKT_CONFIG); > > +} > > +/* > > + * Write the null packet size - will only be taken into account when null > > + * packets are enabled. > > + */ > > +void dsi_dpi_null_packet_size(struct dsi_context *ctx, u16 size) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xC0 video_nullpkt_size; > > + > > + video_nullpkt_size.val = readl(®->VIDEO_NULLPKT_SIZE); > > + video_nullpkt_size.bits.video_nullpkt_size = size; > > + > > + writel(video_nullpkt_size.val, ®->VIDEO_NULLPKT_SIZE); > > +} > > +/* > > + * Write video packet size. obligatory for sending video > > + */ > > +void dsi_dpi_video_packet_size(struct dsi_context *ctx, u16 size) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x24 video_pkt_config; > > + > > + video_pkt_config.val = readl(®->VIDEO_PKT_CONFIG); > > + video_pkt_config.bits.video_pkt_size = size; > > + > > + writel(video_pkt_config.val, ®->VIDEO_PKT_CONFIG); > > +} > > +/* > > + * Specifiy the size of the packet memory write start/continue > > + */ > > +void dsi_edpi_max_pkt_size(struct dsi_context *ctx, u16 size) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xC4 dcs_wm_pkt_size; > > + > > + dcs_wm_pkt_size.val = readl(®->DCS_WM_PKT_SIZE); > > + dcs_wm_pkt_size.bits.dcs_wm_pkt_size = size; > > + > > + writel(dcs_wm_pkt_size.val, ®->DCS_WM_PKT_SIZE); > > +} > > +/* > > + * Enable tear effect acknowledge > > + */ > > +void dsi_tear_effect_ack_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x68 cmd_mode_cfg; > > + > > + cmd_mode_cfg.val = readl(®->CMD_MODE_CFG); > > + cmd_mode_cfg.bits.tear_fx_en = enable; > > + > > + writel(cmd_mode_cfg.val, ®->CMD_MODE_CFG); > > +} > > +/* > > + * Set DCS command packet transmission to transmission type > > + */ > > +void dsi_cmd_mode_lp_cmd_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x68 cmd_mode_cfg; > > + > > + cmd_mode_cfg.val = readl(®->CMD_MODE_CFG); > > + > > + cmd_mode_cfg.bits.gen_sw_0p_tx = enable; > > + cmd_mode_cfg.bits.gen_sw_1p_tx = enable; > > + cmd_mode_cfg.bits.gen_sw_2p_tx = enable; > > + cmd_mode_cfg.bits.gen_lw_tx = enable; > > + cmd_mode_cfg.bits.dcs_sw_0p_tx = enable; > > + cmd_mode_cfg.bits.dcs_sw_1p_tx = enable; > > + cmd_mode_cfg.bits.dcs_lw_tx = enable; > > + cmd_mode_cfg.bits.max_rd_pkt_size = enable; > > + > > + cmd_mode_cfg.bits.gen_sr_0p_tx = enable; > > + cmd_mode_cfg.bits.gen_sr_1p_tx = enable; > > + cmd_mode_cfg.bits.gen_sr_2p_tx = enable; > > + cmd_mode_cfg.bits.dcs_sr_0p_tx = enable; > > + > > + writel(cmd_mode_cfg.val, ®->CMD_MODE_CFG); > > +} > > +/* > > + * Set DCS read command packet transmission to transmission type > > + */ > > +void dsi_video_mode_lp_cmd_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x38 vid_mode_cfg; > > + > > + vid_mode_cfg.val = readl(®->VID_MODE_CFG); > > + vid_mode_cfg.bits.lp_cmd_en = enable; > > + > > + writel(vid_mode_cfg.val, ®->VID_MODE_CFG); > > +} > > + > > +/* > > + * Write command header in the generic interface (which also sends DCS commands) as a subset > > + */ > > +void dsi_set_packet_header(struct dsi_context *ctx, > > + u8 vc, > > + u8 type, > > + u8 wc_lsb, > > + u8 wc_msb) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x6C gen_hdr; > > + > > + gen_hdr.bits.gen_dt = type; > > + gen_hdr.bits.gen_vc = vc; > > + gen_hdr.bits.gen_wc_lsbyte = wc_lsb; > > + gen_hdr.bits.gen_wc_msbyte = wc_msb; > > + > > + writel(gen_hdr.val, ®->GEN_HDR); > > +} > > +/* > > + * Write the payload of the long packet commands > > + */ > > +void dsi_set_packet_payload(struct dsi_context *ctx, u32 payload) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(payload, ®->GEN_PLD_DATA); > > +} > > +/* > > + * Read the payload of the long packet commands > > + */ > > +u32 dsi_get_rx_payload(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + return readl(®->GEN_PLD_DATA); > > +} > > + > > +/* > > + * Enable Bus Turn-around request > > + */ > > +void dsi_bta_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(enable, ®->TA_EN); > > +} > > +/* > > + * Enable EOTp reception > > + */ > > +void dsi_eotp_rx_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xBC eotp_en; > > + > > + eotp_en.val = readl(®->EOTP_EN); > > + eotp_en.bits.rx_eotp_en = enable; > > + > > + writel(eotp_en.val, ®->EOTP_EN); > > +} > > +/* > > + * Enable EOTp transmission > > + */ > > +void dsi_eotp_tx_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xBC eotp_en; > > + > > + eotp_en.val = readl(®->EOTP_EN); > > + eotp_en.bits.tx_eotp_en = enable; > > + > > + writel(eotp_en.val, ®->EOTP_EN); > > +} > > +/* > > + * Enable ECC reception, error correction and reporting > > + */ > > +void dsi_ecc_rx_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xB4 rx_pkt_check_config; > > + > > + rx_pkt_check_config.val = readl(®->RX_PKT_CHECK_CONFIG); > > + rx_pkt_check_config.bits.rx_pkt_ecc_en = enable; > > + > > + writel(rx_pkt_check_config.val, ®->RX_PKT_CHECK_CONFIG); > > +} > > +/* > > + * Enable CRC reception, error reporting > > + */ > > +void dsi_crc_rx_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xB4 rx_pkt_check_config; > > + > > + rx_pkt_check_config.val = readl(®->RX_PKT_CHECK_CONFIG); > > + rx_pkt_check_config.bits.rx_pkt_crc_en = enable; > > + > > + writel(rx_pkt_check_config.val, ®->RX_PKT_CHECK_CONFIG); > > +} > > +/* > > + * Get status of read command > > + */ > > +bool dsi_is_bta_returned(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_rdcmd_done; > > +} > > +/* > > + * Get the FULL status of generic read payload fifo > > + */ > > +bool dsi_is_rx_payload_fifo_full(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_rdata_fifo_full; > > +} > > +/* > > + * Get the EMPTY status of generic read payload fifo > > + */ > > +bool dsi_is_rx_payload_fifo_empty(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_rdata_fifo_empty; > > +} > > +/* > > + * Get the FULL status of generic write payload fifo > > + */ > > +bool dsi_is_tx_payload_fifo_full(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_wdata_fifo_full; > > +} > > +/* > > + * Get the EMPTY status of generic write payload fifo > > + */ > > +bool dsi_is_tx_payload_fifo_empty(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_wdata_fifo_empty; > > +} > > +/* > > + * Get the EMPTY status of generic command fifo > > + */ > > +bool dsi_is_tx_cmd_fifo_empty(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x98 cmd_mode_status; > > + > > + cmd_mode_status.val = readl(®->CMD_MODE_STATUS); > > + > > + return cmd_mode_status.bits.gen_cmd_cmd_fifo_empty; > > +} > > +/* > > + * DPI interface signal delay config > > + * param byte_cycle period for waiting after controller receiving HSYNC from > > + * DPI interface to start read pixel data from memory. > > + */ > > +void dsi_dpi_sig_delay(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xD0 video_sig_delay_config; > > + > > + video_sig_delay_config.val = readl(®->VIDEO_SIG_DELAY_CONFIG); > > + video_sig_delay_config.bits.video_sig_delay = byte_cycle; > > + > > + writel(video_sig_delay_config.val, ®->VIDEO_SIG_DELAY_CONFIG); > > +} > > +/* > > + * Configure how many cycles of byte clock would the PHY module take > > + * to switch data lane from high speed to low power > > + */ > > +void dsi_datalane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xAC phy_datalane_time_config; > > + > > + phy_datalane_time_config.val = readl(®->PHY_DATALANE_TIME_CONFIG); > > + phy_datalane_time_config.bits.phy_datalane_hs_to_lp_time = byte_cycle; > > + > > + writel(phy_datalane_time_config.val, ®->PHY_DATALANE_TIME_CONFIG); > > +} > > +/* > > + * Configure how many cycles of byte clock would the PHY module take > > + * to switch the data lane from to low power high speed > > + */ > > +void dsi_datalane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xAC phy_datalane_time_config; > > + > > + phy_datalane_time_config.val = readl(®->PHY_DATALANE_TIME_CONFIG); > > + phy_datalane_time_config.bits.phy_datalane_lp_to_hs_time = byte_cycle; > > + > > + writel(phy_datalane_time_config.val, ®->PHY_DATALANE_TIME_CONFIG); > > +} > > +/* > > + * Configure how many cycles of byte clock would the PHY module take > > + * to switch clock lane from high speed to low power > > + */ > > +void dsi_clklane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xA8 phy_clklane_time_config; > > + > > + phy_clklane_time_config.val = readl(®->PHY_CLKLANE_TIME_CONFIG); > > + phy_clklane_time_config.bits.phy_clklane_hs_to_lp_time = byte_cycle; > > + > > + writel(phy_clklane_time_config.val, ®->PHY_CLKLANE_TIME_CONFIG); > > +} > > +/* > > + * Configure how many cycles of byte clock would the PHY module take > > + * to switch clock lane from to low power high speed > > + */ > > +void dsi_clklane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xA8 phy_clklane_time_config; > > + > > + phy_clklane_time_config.val = readl(®->PHY_CLKLANE_TIME_CONFIG); > > + phy_clklane_time_config.bits.phy_clklane_lp_to_hs_time = byte_cycle; > > + > > + writel(phy_clklane_time_config.val, ®->PHY_CLKLANE_TIME_CONFIG); > > +} > > +/* > > + * Configure how many cycles of byte clock would the PHY module take > > + * to turn the bus around to start receiving > > + */ > > +void dsi_max_read_time(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(byte_cycle, ®->MAX_READ_TIME); > > +} > > +/* > > + * Enable the automatic mechanism to stop providing clock in the clock > > + * lane when time allows > > + */ > > +void dsi_nc_clk_en(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x74 phy_clk_lane_lp_ctrl; > > + > > + phy_clk_lane_lp_ctrl.val = readl(®->PHY_CLK_LANE_LP_CTRL); > > + phy_clk_lane_lp_ctrl.bits.auto_clklane_ctrl_en = enable; > > + > > + writel(phy_clk_lane_lp_ctrl.val, ®->PHY_CLK_LANE_LP_CTRL); > > +} > > +/* > > + * Write transmission escape timeout > > + * a safe guard so that the state machine would reset if transmission > > + * takes too long > > + */ > > +void dsi_tx_escape_division(struct dsi_context *ctx, u8 div) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(div, ®->TX_ESC_CLK_CONFIG); > > +} > > +/* > > + * Configure timeout divisions (so they would have more clock ticks) > > + * div no of hs cycles before transiting back to LP in > > + * (lane_clk / div) > > + */ > > +void dsi_timeout_clock_division(struct dsi_context *ctx, u8 div) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(div, ®->TIMEOUT_CNT_CLK_CONFIG); > > +} > > +/* > > + * Configure the Low power receive time out > > + */ > > +void dsi_lp_rx_timeout(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(byte_cycle, ®->LRX_H_TO_CONFIG); > > +} > > +/* > > + * Configure a high speed transmission time out > > + */ > > +void dsi_hs_tx_timeout(struct dsi_context *ctx, u16 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(byte_cycle, ®->HTX_TO_CONFIG); > > +} > > +/* > > + * Get the error 0 interrupt register status > > + */ > > +u32 dsi_int0_status(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x08 protocol_int_sts; > > + > > + protocol_int_sts.val = readl(®->PROTOCOL_INT_STS); > > + writel(protocol_int_sts.val, ®->PROTOCOL_INT_CLR); > > + > > + if (protocol_int_sts.bits.dphy_errors_0) > > + drm_err(dsi->drm, "dphy_err: escape entry error\n"); > > + > > + if (protocol_int_sts.bits.dphy_errors_1) > > + drm_err(dsi->drm, "dphy_err: lp data transmission sync error\n"); > > + > > + if (protocol_int_sts.bits.dphy_errors_2) > > + drm_err(dsi->drm, "dphy_err: control error\n"); > > + > > + if (protocol_int_sts.bits.dphy_errors_3) > > + drm_err(dsi->drm, "dphy_err: LP0 contention error\n"); > > + > > + if (protocol_int_sts.bits.dphy_errors_4) > > + drm_err(dsi->drm, "dphy_err: LP1 contention error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_0) > > + drm_err(dsi->drm, "ack_err: SoT error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_1) > > + drm_err(dsi->drm, "ack_err: SoT Sync error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_2) > > + drm_err(dsi->drm, "ack_err: EoT Sync error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_3) > > + drm_err(dsi->drm, "ack_err: Escape Mode Entry Command error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_4) > > + drm_err(dsi->drm, "ack_err: LP Transmit Sync error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_5) > > + drm_err(dsi->drm, "ack_err: Peripheral Timeout error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_6) > > + drm_err(dsi->drm, "ack_err: False Control error\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_7) > > + drm_err(dsi->drm, "ack_err: reserved (specific to device)\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_8) > > + drm_err(dsi->drm, "ack_err: ECC error, single-bit (corrected)\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_9) > > + drm_err(dsi->drm, "ack_err: ECC error, multi-bit (not corrected)\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_10) > > + drm_err(dsi->drm, "ack_err: checksum error (long packet only)\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_11) > > + drm_err(dsi->drm, "ack_err: not recognized DSI data type\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_12) > > + drm_err(dsi->drm, "ack_err: DSI VC ID Invalid\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_13) > > + drm_err(dsi->drm, "ack_err: invalid transmission length\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_14) > > + drm_err(dsi->drm, "ack_err: reserved (specific to device)\n"); > > + > > + if (protocol_int_sts.bits.ack_with_err_15) > > + drm_err(dsi->drm, "ack_err: DSI protocol violation\n"); > > + > > + return 0; > > +} > > +/* > > + * Get the error 1 interrupt register status > > + */ > > +u32 dsi_int1_status(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x10 internal_int_sts; > > + u32 status = 0; > > + > > + internal_int_sts.val = readl(®->INTERNAL_INT_STS); > > + writel(internal_int_sts.val, ®->INTERNAL_INT_CLR); > > + > > + if (internal_int_sts.bits.receive_pkt_size_err) > > + drm_err(dsi->drm, "receive packet size error\n"); > > + > > + if (internal_int_sts.bits.eotp_not_receive_err) > > + drm_err(dsi->drm, "EoTp packet is not received\n"); > > + > > + if (internal_int_sts.bits.gen_cmd_cmd_fifo_wr_err) > > + drm_err(dsi->drm, "cmd header-fifo is full\n"); > > + > > + if (internal_int_sts.bits.gen_cmd_rdata_fifo_rd_err) > > + drm_err(dsi->drm, "cmd read-payload-fifo is empty\n"); > > + > > + if (internal_int_sts.bits.gen_cmd_rdata_fifo_wr_err) > > + drm_err(dsi->drm, "cmd read-payload-fifo is full\n"); > > + > > + if (internal_int_sts.bits.gen_cmd_wdata_fifo_wr_err) > > + drm_err(dsi->drm, "cmd write-payload-fifo is full\n"); > > + > > + if (internal_int_sts.bits.gen_cmd_wdata_fifo_rd_err) > > + drm_err(dsi->drm, "cmd write-payload-fifo is empty\n"); > > + > > + if (internal_int_sts.bits.dpi_pix_fifo_wr_err) { > > + drm_err(dsi->drm, "DPI pixel-fifo is full\n"); > > + status |= DSI_INT_STS_NEED_SOFT_RESET; > > + } > > + > > + if (internal_int_sts.bits.ecc_single_err) > > + drm_err(dsi->drm, "ECC single error in a received packet\n"); > > + > > + if (internal_int_sts.bits.ecc_multi_err) > > + drm_err(dsi->drm, "ECC multiple error in a received packet\n"); > > + > > + if (internal_int_sts.bits.crc_err) > > + drm_err(dsi->drm, "CRC error in the received packet payload\n"); > > + > > + if (internal_int_sts.bits.hs_tx_timeout) > > + drm_err(dsi->drm, "high-speed transmission timeout\n"); > > + > > + if (internal_int_sts.bits.lp_rx_timeout) > > + drm_err(dsi->drm, "low-power reception timeout\n"); > > + > > + return status; > > +} > > +/* > > + * Configure MASK (hiding) of interrupts coming from error 0 source > > + */ > > +void dsi_int0_mask(struct dsi_context *ctx, u32 mask) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(mask, ®->MASK_PROTOCOL_INT); > > +} > > +/* > > + * Configure MASK (hiding) of interrupts coming from error 1 source > > + */ > > +void dsi_int1_mask(struct dsi_context *ctx, u32 mask) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(mask, ®->MASK_INTERNAL_INT); > > +} > > diff --git a/drivers/gpu/drm/sprd/dsi_ctrl.h b/drivers/gpu/drm/sprd/dsi_ctrl.h > > new file mode 100644 > > index 000000000..88069227f > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/dsi_ctrl.h > > @@ -0,0 +1,1475 @@ > > +/* SPDX-License-Identifier: GPL-2.0 */ > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#ifndef _DSI_CTRL_H_ > > +#define _DSI_CTRL_H_ > > + > > +#include <asm/types.h> > > + > > +#include "sprd_dsi.h" > > + > > +struct dsi_reg { > > + union _0x00 { > > + u32 val; > > + struct _DSI_VERSION { > > + u32 dsi_version: 16; > > + u32 reserved: 16; > > + } bits; > > + } DSI_VERSION; > > Using unions and structures to define the register is really frowned > upon in favor of defines, like you rightfully did in the crtc driver. This workload is too big, this design has been used for many years, so I actually want to keep it the same, but if it really doesn’t meet the current design. I can change the design, but it may take a lot of time...... > > > + union _0x04 { > > + u32 val; > > + struct _SOFT_RESET { > > + /* > > + * This bit configures the core either to work normal or to > > + * reset. It's default value is 0. After the core configur- > > + * ation, to enable the mipi_dsi_host, set this register to 1. > > + * 1: power up 0: reset core > > + */ > > + u32 dsi_soft_reset: 1; > > + > > + u32 reserved: 31; > > + } bits; > > + } SOFT_RESET; > > + > > + union _0x08 { > > + u32 val; > > + struct _PROTOCOL_INT_STS { > > + /* ErrEsc escape entry error from Lane 0 */ > > + u32 dphy_errors_0: 1; > > + > > + /* ErrSyncEsc low-power data transmission synchronization > > + * error from Lane 0 > > + */ > > + u32 dphy_errors_1: 1; > > + > > + /* ErrControl error from Lane 0 */ > > + u32 dphy_errors_2: 1; > > + > > + /* ErrContentionLP0 LP0 contention error from Lane 0 */ > > + u32 dphy_errors_3: 1; > > + > > + /* ErrContentionLP1 LP1 contention error from Lane 0 */ > > + u32 dphy_errors_4: 1; > > + > > + /* debug mode protocol errors */ > > + u32 protocol_debug_err: 11; > > + > > + /* SoT error from the Acknowledge error report */ > > + u32 ack_with_err_0: 1; > > + > > + /* SoT Sync error from the Acknowledge error report */ > > + u32 ack_with_err_1: 1; > > + > > + /* EoT Sync error from the Acknowledge error report */ > > + u32 ack_with_err_2: 1; > > + > > + /* Escape Mode Entry Command error from the Acknowledge > > + * error report > > + */ > > + u32 ack_with_err_3: 1; > > + > > + /* LP Transmit Sync error from the Acknowledge error report */ > > + u32 ack_with_err_4: 1; > > + > > + /* Peripheral Timeout error from the Acknowledge error report */ > > + u32 ack_with_err_5: 1; > > + > > + /* False Control error from the Acknowledge error report */ > > + u32 ack_with_err_6: 1; > > + > > + /* reserved (specific to device) from the Acknowledge error > > + * report > > + */ > > + u32 ack_with_err_7: 1; > > + > > + /* ECC error, single-bit (detected and corrected) from the > > + * Acknowledge error report > > + */ > > + u32 ack_with_err_8: 1; > > + > > + /* ECC error, multi-bit (detected, not corrected) from the > > + * Acknowledge error report > > + */ > > + u32 ack_with_err_9: 1; > > + > > + /* checksum error (long packet only) from the Acknowledge > > + * error report > > + */ > > + u32 ack_with_err_10: 1; > > + > > + /* not recognized DSI data type from the Acknowledge error > > + * report > > + */ > > + u32 ack_with_err_11: 1; > > + > > + /* DSI VC ID Invalid from the Acknowledge error report */ > > + u32 ack_with_err_12: 1; > > + > > + /* invalid transmission length from the Acknowledge error > > + * report > > + */ > > + u32 ack_with_err_13: 1; > > + > > + /* reserved (specific to device) from the Acknowledge error > > + * report > > + */ > > + u32 ack_with_err_14: 1; > > + > > + /* DSI protocol violation from the Acknowledge error report */ > > + u32 ack_with_err_15: 1; > > + > > + } bits; > > + } PROTOCOL_INT_STS; > > + > > + union _0x0C { > > + u32 val; > > + struct _MASK_PROTOCOL_INT { > > + u32 mask_dphy_errors_0: 1; > > + u32 mask_dphy_errors_1: 1; > > + u32 mask_dphy_errors_2: 1; > > + u32 mask_dphy_errors_3: 1; > > + u32 mask_dphy_errors_4: 1; > > + u32 mask_protocol_debug_err: 11; > > + u32 mask_ack_with_err_0: 1; > > + u32 mask_ack_with_err_1: 1; > > + u32 mask_ack_with_err_2: 1; > > + u32 mask_ack_with_err_3: 1; > > + u32 mask_ack_with_err_4: 1; > > + u32 mask_ack_with_err_5: 1; > > + u32 mask_ack_with_err_6: 1; > > + u32 mask_ack_with_err_7: 1; > > + u32 mask_ack_with_err_8: 1; > > + u32 mask_ack_with_err_9: 1; > > + u32 mask_ack_with_err_10: 1; > > + u32 mask_ack_with_err_11: 1; > > + u32 mask_ack_with_err_12: 1; > > + u32 mask_ack_with_err_13: 1; > > + u32 mask_ack_with_err_14: 1; > > + u32 mask_ack_with_err_15: 1; > > + } bits; > > + } MASK_PROTOCOL_INT; > > + > > + union _0x10 { > > + u32 val; > > + struct _INTERNAL_INT_STS { > > + /* This bit indicates that the packet size error is detected > > + * during the packet reception. > > + */ > > + u32 receive_pkt_size_err: 1; > > + > > + /* This bit indicates that the EoTp packet is not received at > > + * the end of the incoming peripheral transmission > > + */ > > + u32 eotp_not_receive_err: 1; > > + > > + /* This bit indicates that the system tried to write a command > > + * through the Generic interface and the FIFO is full. There- > > + * fore, the command is not written. > > + */ > > + u32 gen_cmd_cmd_fifo_wr_err: 1; > > + > > + /* This bit indicates that during a DCS read data, the payload > > + * FIFO becomes empty and the data sent to the interface is > > + * corrupted. > > + */ > > + u32 gen_cmd_rdata_fifo_rd_err: 1; > > + > > + /* This bit indicates that during a generic interface packet > > + * read back, the payload FIFO becomes full and the received > > + * data is corrupted. > > + */ > > + u32 gen_cmd_rdata_fifo_wr_err: 1; > > + > > + /* This bit indicates that the system tried to write a payload > > + * data through the Generic interface and the FIFO is full. > > + * Therefore, the payload is not written. > > + */ > > + u32 gen_cmd_wdata_fifo_wr_err: 1; > > + > > + /* This bit indicates that during a Generic interface packet > > + * build, the payload FIFO becomes empty and corrupt data is > > + * sent. > > + */ > > + u32 gen_cmd_wdata_fifo_rd_err: 1; > > + > > + /* This bit indicates that during a DPI pixel line storage, > > + * the payload FIFO becomes full and the data stored is > > + * corrupted. > > + */ > > + u32 dpi_pix_fifo_wr_err: 1; > > + > > + /* internal debug error */ > > + u32 internal_debug_err: 19; > > + > > + /* This bit indicates that the ECC single error is detected > > + * and corrected in a received packet. > > + */ > > + u32 ecc_single_err: 1; > > + > > + /* This bit indicates that the ECC multiple error is detected > > + * in a received packet. > > + */ > > + u32 ecc_multi_err: 1; > > + > > + /* This bit indicates that the CRC error is detected in the > > + * received packet payload. > > + */ > > + u32 crc_err: 1; > > + > > + /* This bit indicates that the high-speed transmission timeout > > + * counter reached the end and contention is detected. > > + */ > > + u32 hs_tx_timeout: 1; > > + > > + /* This bit indicates that the low-power reception timeout > > + * counter reached the end and contention is detected. > > + */ > > + u32 lp_rx_timeout: 1; > > + > > + } bits; > > + } INTERNAL_INT_STS; > > + > > + union _0x14 { > > + u32 val; > > + struct _MASK_INTERNAL_INT { > > + u32 mask_receive_pkt_size_err: 1; > > + u32 mask_eopt_not_receive_err: 1; > > + u32 mask_gen_cmd_cmd_fifo_wr_err: 1; > > + u32 mask_gen_cmd_rdata_fifo_rd_err: 1; > > + u32 mask_gen_cmd_rdata_fifo_wr_err: 1; > > + u32 mask_gen_cmd_wdata_fifo_wr_err: 1; > > + u32 mask_gen_cmd_wdata_fifo_rd_err: 1; > > + u32 mask_dpi_pix_fifo_wr_err: 1; > > + u32 mask_internal_debug_err: 19; > > + u32 mask_ecc_single_err: 1; > > + u32 mask_ecc_multi_err: 1; > > + u32 mask_crc_err: 1; > > + u32 mask_hs_tx_timeout: 1; > > + u32 mask_lp_rx_timeout: 1; > > + } bits; > > + } MASK_INTERNAL_INT; > > + > > + union _0x18 { > > + u32 val; > > + struct _DSI_MODE_CFG { > > + /* This bit configures the operation mode > > + * 0: Video mode ; 1: Command mode > > + */ > > + u32 cmd_video_mode: 1; > > + > > + u32 reserved: 31; > > + > > + } bits; > > + } DSI_MODE_CFG; > > + > > + union _0x1C { > > + u32 val; > > + struct _VIRTUAL_CHANNEL_ID { > > + /* This field indicates the Generic interface read-back > > + * virtual channel identification > > + */ > > + u32 gen_rx_vcid: 2; > > + > > + /* This field configures the DPI virtual channel id that > > + * is indexed to the VIDEO mode packets > > + */ > > + u32 video_pkt_vcid: 2; > > + > > + u32 reserved: 28; > > + > > + } bits; > > + } VIRTUAL_CHANNEL_ID; > > + > > + union _0x20 { > > + u32 val; > > + struct _DPI_VIDEO_FORMAT { > > + /* > > + * This field configures the DPI color coding as follows: > > + * 0000: 16-bit configuration 1 > > + * 0001: 16-bit configuration 2 > > + * 0010: 16-bit configuration 3 > > + * 0011: 18-bit configuration 1 > > + * 0100: 18-bit configuration 2 > > + * 0101: 24-bit > > + * 0110: 20-bit YCbCr 4:2:2 loosely packed > > + * 0111: 24-bit YCbCr 4:2:2 > > + * 1000: 16-bit YCbCr 4:2:2 > > + * 1001: 30-bit > > + * 1010: 36-bit > > + * 1011: 12-bit YCbCr 4:2:0 > > + * 1100: Compression Display Stream > > + * 1101-1111: 12-bit YCbCr 4:2:0 > > + */ > > + u32 dpi_video_mode_format: 6; > > + > > + /* When set to 1, this bit activates loosely packed > > + * variant to 18-bit configurations > > + */ > > + u32 loosely18_en: 1; > > + > > + u32 reserved: 25; > > + > > + } bits; > > + } DPI_VIDEO_FORMAT; > > + > > + union _0x24 { > > + u32 val; > > + struct _VIDEO_PKT_CONFIG { > > + /* > > + * This field configures the number of pixels in a single > > + * video packet. For 18-bit not loosely packed data types, > > + * this number must be a multiple of 4. For YCbCr data > > + * types, it must be a multiple of 2, as described in the > > + * DSI specification. > > + */ > > + u32 video_pkt_size: 16; > > + > > + /* > > + * This register configures the number of chunks to be > > + * transmitted during a Line period (a chunk consists of > > + * a video packet and a null packet). If set to 0 or 1, > > + * the video line is transmitted in a single packet. If > > + * set to 1, the packet is part of a chunk, so a null packet > > + * follows it if vid_null_size > 0. Otherwise, multiple chunks > > + * are used to transmit each video line. > > + */ > > + u32 video_line_chunk_num: 16; > > + > > + } bits; > > + } VIDEO_PKT_CONFIG; > > + > > + union _0x28 { > > + u32 val; > > + struct _VIDEO_LINE_HBLK_TIME { > > + /* This field configures the Horizontal Back Porch period > > + * in lane byte clock cycles > > + */ > > + u32 video_line_hbp_time: 16; > > + > > + /* This field configures the Horizontal Synchronism Active > > + * period in lane byte clock cycles > > + */ > > + u32 video_line_hsa_time: 16; > > + > > + } bits; > > + } VIDEO_LINE_HBLK_TIME; > > + > > + union _0x2C { > > + u32 val; > > + struct _VIDEO_LINE_TIME { > > + /* This field configures the size of the total line time > > + * (HSA+HBP+HACT+HFP) counted in lane byte clock cycles > > + */ > > + u32 video_line_time: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } VIDEO_LINE_TIME; > > + > > + union _0x30 { > > + u32 val; > > + struct _VIDEO_VBLK_LINES { > > + /* This field configures the Vertical Front Porch period > > + * measured in number of horizontal lines > > + */ > > + u32 vfp_lines: 10; > > + > > + /* This field configures the Vertical Back Porch period > > + * measured in number of horizontal lines > > + */ > > + u32 vbp_lines: 10; > > + > > + /* This field configures the Vertical Synchronism Active > > + * period measured in number of horizontal lines > > + */ > > + u32 vsa_lines: 10; > > + > > + u32 reserved: 2; > > + > > + } bits; > > + } VIDEO_VBLK_LINES; > > + > > + union _0x34 { > > + u32 val; > > + struct _VIDEO_VACTIVE_LINES { > > + /* This field configures the Vertical Active period measured > > + * in number of horizontal lines > > + */ > > + u32 vactive_lines: 14; > > + > > + u32 reserved: 18; > > + > > + } bits; > > + } VIDEO_VACTIVE_LINES; > > + > > + union _0x38 { > > + u32 val; > > + struct _VID_MODE_CFG { > > + /* > > + * This field indicates the video mode transmission type as > > + * follows: > > + * 00: Non-burst with sync pulses > > + * 01: Non-burst with sync events > > + * 10 and 11: Burst mode > > + */ > > + u32 vid_mode_type: 2; > > + > > + u32 reserved_0: 6; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the VSA period when timing allows. > > + */ > > + u32 lp_vsa_en: 1; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the VBP period when timing allows. > > + */ > > + u32 lp_vbp_en: 1; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the VFP period when timing allows. > > + */ > > + u32 lp_vfp_en: 1; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the VACT period when timing allows. > > + */ > > + u32 lp_vact_en: 1; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the HBP period when timing allows. > > + */ > > + u32 lp_hbp_en: 1; > > + > > + /* When set to 1, this bit enables the return to low-power > > + * inside the HFP period when timing allows. > > + */ > > + u32 lp_hfp_en: 1; > > + > > + /* When set to 1, this bit enables the request for an ack- > > + * nowledge response at the end of a frame. > > + */ > > + u32 frame_bta_ack_en: 1; > > + > > + /* When set to 1, this bit enables the command transmission > > + * only in low-power mode. > > + */ > > + u32 lp_cmd_en: 1; > > + > > + u32 reserved_1: 16; > > + > > + } bits; > > + } VID_MODE_CFG; > > + > > + union _0x3C { > > + u32 val; > > + struct _SDF_MODE_CONFIG { > > + /* > > + * This field defines the 3D mode on/off & display orientation: > > + * 00: 3D mode off (2D mode on) > > + * 01: 3D mode on, portrait orientation > > + * 10: 3D mode on, landscape orientation > > + * 11: Reserved > > + */ > > + u32 rf_3d_mode: 2; > > + > > + /* > > + * This field defines the 3D image format: > > + * 00: Line (alternating lines of left and right data) > > + * 01: Frame (alternating frames of left and right data) > > + * 10: Pixel (alternating pixels of left and right data) > > + * 11: Reserved > > + */ > > + u32 rf_3d_format: 2; > > + > > + /* > > + * This field defines whether there is a second VSYNC pulse > > + * between Left and Right Images, when 3D Image Format is > > + * Frame-based: > > + * 0: No sync pulses between left and right data > > + * 1: Sync pulse (HSYNC, VSYNC, blanking) between left and > > + * right data > > + */ > > + u32 second_vsync_en: 1; > > + > > + /* > > + * This bit defines the left or right order: > > + * 0: Left eye data is sent first, and then the right eye data > > + * is sent. > > + * 1: Right eye data is sent first, and then the left eye data > > + * is sent. > > + */ > > + u32 left_right_order: 1; > > + > > + u32 reserved_0: 2; > > + > > + /* > > + * When set, causes the next VSS packet to include 3D control > > + * payload in every VSS packet. > > + */ > > + u32 rf_3d_payload_en: 1; > > + > > + u32 reserved_1: 23; > > + > > + } bits; > > + } SDF_MODE_CONFIG; > > + > > + union _0x40 { > > + u32 val; > > + struct _TIMEOUT_CNT_CLK_CONFIG { > > + /* > > + * This field indicates the division factor for the Time Out > > + * clock used as the timing unit in the configuration of HS to > > + * LP and LP to HS transition error. > > + */ > > + u32 timeout_cnt_clk_config: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } TIMEOUT_CNT_CLK_CONFIG; > > + > > + union _0x44 { > > + u32 val; > > + struct _HTX_TO_CONFIG { > > + /* > > + * This field configures the timeout counter that triggers > > + * a high speed transmission timeout contention detection > > + * (measured in TO_CLK_DIVISION cycles). > > + * > > + * If using the non-burst mode and there is no sufficient > > + * time to switch from HS to LP and back in the period which > > + * is from one line data finishing to the next line sync > > + * start, the DSI link returns the LP state once per frame, > > + * then you should configure the TO_CLK_DIVISION and > > + * hstx_to_cnt to be in accordance with: > > + * hstx_to_cnt * lanebyteclkperiod * TO_CLK_DIVISION >= the > > + * time of one FRAME data transmission * (1 + 10%) > > + * > > + * In burst mode, RGB pixel packets are time-compressed, > > + * leaving more time during a scan line. Therefore, if in > > + * burst mode and there is sufficient time to switch from HS > > + * to LP and back in the period of time from one line data > > + * finishing to the next line sync start, the DSI link can > > + * return LP mode and back in this time interval to save power. > > + * For this, configure the TO_CLK_DIVISION and hstx_to_cnt > > + * to be in accordance with: > > + * hstx_to_cnt * lanebyteclkperiod * TO_CLK_DIVISION >= the > > + * time of one LINE data transmission * (1 + 10%) > > + */ > > + u32 htx_to_cnt_limit: 32; > > + } bits; > > + } HTX_TO_CONFIG; > > + > > + union _0x48 { > > + u32 val; > > + struct _LRX_H_TO_CONFIG { > > + /* > > + * This field configures the timeout counter that triggers > > + * a low-power reception timeout contention detection (measured > > + * in TO_CLK_DIVISION cycles). > > + */ > > + u32 lrx_h_to_cnt_limit: 32; > > + } bits; > > + } LRX_H_TO_CONFIG; > > + > > + union _0x4C { > > + u32 val; > > + struct _RD_PRESP_TO_CONFIG { > > + /* > > + * This field sets a period for which the DWC_mipi_dsi_host > > + * keeps the link still, after sending a low-power read oper- > > + * ation. This period is measured in cycles of lanebyteclk. > > + * The counting starts when the D-PHY enters the Stop state > > + * and causes no interrupts. > > + */ > > + u32 lprd_presp_to_cnt_limit: 16; > > + > > + /* > > + * This field sets a period for which the DWC_mipi_dsi_host > > + * keeps the link still, after sending a high-speed read oper- > > + * ation. This period is measured in cycles of lanebyteclk. > > + * The counting starts when the D-PHY enters the Stop state > > + * and causes no interrupts. > > + */ > > + u32 hsrd_presp_to_cnt_limit: 16; > > + > > + } bits; > > + } RD_PRESP_TO_CONFIG; > > + > > + union _0x50 { > > + u32 val; > > + struct _HSWR_PRESP_TO_CONFIG { > > + /* > > + * This field sets a period for which the DWC_mipi_dsi_host > > + * keeps the link inactive after sending a high-speed write > > + * operation. This period is measured in cycles of lanebyteclk. > > + * The counting starts when the D-PHY enters the Stop state > > + * and causes no interrupts. > > + */ > > + u32 hswr_presp_to_cnt_limit: 16; > > + > > + u32 reserved_0: 8; > > + > > + /* > > + * When set to 1, this bit ensures that the peripheral response > > + * timeout caused by hs_wr_to_cnt is used only once per eDPI > > + * frame, when both the following conditions are met: > > + * dpivsync_edpiwms has risen and fallen. > > + * Packets originated from eDPI have been transmitted and its > > + * FIFO is empty again In this scenario no non-eDPI requests > > + * are sent to the D-PHY, even if there is traffic from generic > > + * or DBI ready to be sent, making it return to stop state. > > + * When it does so, PRESP_TO counter is activated and only when > > + * it finishes does the controller send any other traffic that > > + * is ready. > > + */ > > + u32 hswr_presp_to_mode: 1; > > + > > + u32 reserved_1: 7; > > + > > + } bits; > > + } HSWR_PRESP_TO_CONFIG; > > + > > + union _0x54 { > > + u32 val; > > + struct _LPWR_PRESP_TO_CONFIG { > > + /* > > + * This field sets a period for which the DWC_mipi_dsi_host > > + * keeps the link still, after sending a low-power write oper- > > + * ation. This period is measured in cycles of lanebyteclk. > > + * The counting starts when the D-PHY enters the Stop state > > + * and causes no interrupts. > > + */ > > + u32 lpwr_presp_to_cnt_limit: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } LPWR_PRESP_TO_CONFIG; > > + > > + union _0x58 { > > + u32 val; > > + struct _BTA_PRESP_TO_CONFIG { > > + /* > > + * This field sets a period for which the DWC_mipi_dsi_host > > + * keeps the link still, after completing a Bus Turn-Around. > > + * This period is measured in cycles of lanebyteclk. The > > + * counting starts when the D-PHY enters the Stop state and > > + * causes no interrupts. > > + */ > > + u32 bta_presp_to_cnt_limit: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } BTA_PRESP_TO_CONFIG; > > + > > + union _0x5C { > > + u32 val; > > + struct _TX_ESC_CLK_CONFIG { > > + /* > > + * This field indicates the division factor for the TX Escape > > + * clock source (lanebyteclk). The values 0 and 1 stop the > > + * TX_ESC clock generation. > > + */ > > + u32 tx_esc_clk_config: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } TX_ESC_CLK_CONFIG; > > + > > + union _0x60 { > > + u32 val; > > + struct _VACT_CMD_TRANS_LIMIT { > > + /* > > + * This field is used for the transmission of commands in > > + * low-power mode. It defines the size, in bytes, of the > > + * largest packet that can fit in a line during the VACT > > + * region. > > + */ > > + u32 vact_cmd_trans_limit: 8; > > + > > + u32 reserved: 24; > > + > > + } bits; > > + } VACT_CMD_TRANS_LIMIT; > > + > > + union _0x64 { > > + u32 val; > > + struct _VBLK_CMD_TRANS_LIMIT { > > + /* > > + * This field is used for the transmission of commands in > > + * low-power mode. It defines the size, in bytes, of the > > + * largest packet that can fit in a line during the VSA, VBP, > > + * and VFP regions. > > + */ > > + u32 vblk_cmd_trans_limit: 8; > > + > > + u32 reserved: 24; > > + > > + } bits; > > + } VBLK_CMD_TRANS_LIMIT; > > + > > + union _0x68 { > > + u32 val; > > + struct _CMD_MODE_CFG { > > + /* > > + * When set to 1, this bit enables the tearing effect > > + * acknowledge request. > > + */ > > + u32 tear_fx_en: 1; > > + > > + /* > > + * When set to 1, this bit enables the acknowledge request > > + * after each packet transmission. > > + */ > > + u32 ack_rqst_en: 1; > > + > > + u32 reserved_0: 3; > > + > > + u32 pps_tx: 1; > > + u32 exq_tx: 1; > > + u32 cmc_tx: 1; > > + > > + /* > > + * This bit configures the Generic short write packet with > > + * zero parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sw_0p_tx: 1; > > + > > + /* > > + * This bit configures the Generic short write packet with > > + * one parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sw_1p_tx: 1; > > + > > + /* > > + * This bit configures the Generic short write packet with > > + * two parameters command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sw_2p_tx: 1; > > + > > + /* > > + * This bit configures the Generic short read packet with > > + * zero parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sr_0p_tx: 1; > > + > > + /* > > + * This bit configures the Generic short read packet with > > + * one parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sr_1p_tx: 1; > > + > > + /* > > + * This bit configures the Generic short read packet with > > + * two parameters command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_sr_2p_tx: 1; > > + > > + /* > > + * This bit configures the Generic long write packet command > > + * transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 gen_lw_tx: 1; > > + > > + u32 reserved_1: 1; > > + > > + /* > > + * This bit configures the DCS short write packet with zero > > + * parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 dcs_sw_0p_tx: 1; > > + > > + /* > > + * This bit configures the DCS short write packet with one > > + * parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 dcs_sw_1p_tx: 1; > > + > > + /* > > + * This bit configures the DCS short read packet with zero > > + * parameter command transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 dcs_sr_0p_tx: 1; > > + > > + /* > > + * This bit configures the DCS long write packet command > > + * transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 dcs_lw_tx: 1; > > + > > + u32 reserved_2: 4; > > + > > + /* > > + * This bit configures the maximum read packet size command > > + * transmission type: > > + * 0: High-speed 1: Low-power > > + */ > > + u32 max_rd_pkt_size: 1; > > + > > + u32 reserved_3: 7; > > + > > + } bits; > > + } CMD_MODE_CFG; > > + > > + union _0x6C { > > + u32 val; > > + struct _GEN_HDR { > > + /* > > + * This field configures the packet data type of the header > > + * packet. > > + */ > > + u32 gen_dt: 6; > > + > > + /* > > + * This field configures the virtual channel id of the header > > + * packet. > > + */ > > + u32 gen_vc: 2; > > + > > + /* > > + * This field configures the least significant byte of the > > + * header packet's Word count for long packets or data 0 for > > + * short packets. > > + */ > > + u32 gen_wc_lsbyte: 8; > > + > > + /* > > + * This field configures the most significant byte of the > > + * header packet's word count for long packets or data 1 for > > + * short packets. > > + */ > > + u32 gen_wc_msbyte: 8; > > + > > + u32 reserved: 8; > > + > > + } bits; > > + } GEN_HDR; > > + > > + union _0x70 { > > + u32 val; > > + struct _GEN_PLD_DATA { > > + /* This field indicates byte 1 of the packet payload. */ > > + u32 gen_pld_b1: 8; > > + > > + /* This field indicates byte 2 of the packet payload. */ > > + u32 gen_pld_b2: 8; > > + > > + /* This field indicates byte 3 of the packet payload. */ > > + u32 gen_pld_b3: 8; > > + > > + /* This field indicates byte 4 of the packet payload. */ > > + u32 gen_pld_b4: 8; > > + > > + } bits; > > + } GEN_PLD_DATA; > > + > > + union _0x74 { > > + u32 val; > > + struct _PHY_CLK_LANE_LP_CTRL { > > + /* This bit controls the D-PHY PPI txrequestclkhs signal */ > > + u32 phy_clklane_tx_req_hs: 1; > > + > > + /* This bit enables the automatic mechanism to stop providing > > + * clock in the clock lane when time allows. > > + */ > > + u32 auto_clklane_ctrl_en: 1; > > + > > + u32 reserved: 30; > > + } bits; > > + } PHY_CLK_LANE_LP_CTRL; > > + > > + union _0x78 { > > + u32 val; > > + struct _PHY_INTERFACE_CTRL { > > + /* When set to 0, this bit places the D-PHY macro in power- > > + * down state. > > + */ > > + u32 rf_phy_shutdown: 1; > > + > > + /* When set to 0, this bit places the digital section of the > > + * D-PHY in the reset state. > > + */ > > + u32 rf_phy_reset_n: 1; > > + > > + /* When set to 1, this bit enables the D-PHY Clock Lane > > + * module. > > + */ > > + u32 rf_phy_clk_en: 1; > > + > > + /* When the D-PHY is in ULPS, this bit enables the D-PHY PLL. */ > > + u32 rf_phy_force_pll: 1; > > + > > + /* ULPS mode Request on clock lane */ > > + u32 rf_phy_clk_txrequlps: 1; > > + > > + /* ULPS mode Exit on clock lane */ > > + u32 rf_phy_clk_txexitulps: 1; > > + > > + /* ULPS mode Request on all active data lanes */ > > + u32 rf_phy_data_txrequlps: 1; > > + > > + /* ULPS mode Exit on all active data lanes */ > > + u32 rf_phy_data_txexitulps: 1; > > + > > + u32 reserved: 24; > > + } bits; > > + } PHY_INTERFACE_CTRL; > > + > > + union _0x7C { > > + u32 val; > > + struct _PHY_TX_TRIGGERS { > > + /* This field controls the trigger transmissions. */ > > + u32 phy_tx_triggers: 4; > > + > > + u32 reserved: 28; > > + } bits; > > + } PHY_TX_TRIGGERS; > > + > > + union _0x80 { > > + u32 val; > > + struct _DESKEW_START { > > + u32 deskew_start: 1; > > + u32 reserved: 31; > > + } bits; > > + } DESKEW_START; > > + > > + union _0x84 { > > + u32 val; > > + struct _DESKEW_MODE { > > + u32 deskew_mode: 2; > > + u32 reserved: 30; > > + } bits; > > + } DESKEW_MODE; > > + > > + union _0x88 { > > + u32 val; > > + struct _DESKEW_TIME { > > + u32 deskew_time: 32; > > + } bits; > > + } DESKEW_TIME; > > + > > + union _0x8C { > > + u32 val; > > + struct _DESKEW_PERIOD { > > + u32 deskew_period: 32; > > + } bits; > > + } DESKEW_PERIOD; > > + > > + union _0x90 { > > + u32 val; > > + struct _DESKEW_BUSY { > > + u32 deskew_busy: 1; > > + u32 reserved: 31; > > + } bits; > > + } DESKEW_BUSY; > > + > > + union _0x94 { > > + u32 val; > > + struct _DESKEW_LANE_MASK { > > + u32 deskew_lane0_mask: 1; > > + u32 deskew_lane1_mask: 1; > > + u32 deskew_lane2_mask: 1; > > + u32 deskew_lane3_mask: 1; > > + u32 reserved: 28; > > + } bits; > > + } DESKEW_LANE_MASK; > > + > > + union _0x98 { > > + u32 val; > > + struct _CMD_MODE_STATUS { > > + /* > > + * This bit is set when a read command is issued and cleared > > + * when the entire response is stored in the FIFO. > > + * Value after reset: 0x0 > > + * > > + * NOTE: > > + * For mipi-dsi-r1p0 IP, this bit is set immediately when > > + * the read cmd is set to the GEN_HDR register. > > + * > > + * For dsi-ctrl-r1p0 IP, this bit is set only after the read > > + * cmd was actually sent out from the controller. > > + */ > > + u32 gen_cmd_rdcmd_ongoing: 1; > > + > > + /* > > + * This bit indicates the empty status of the generic read > > + * payload FIFO. > > + * Value after reset: 0x1 > > + */ > > + u32 gen_cmd_rdata_fifo_empty: 1; > > + > > + /* > > + * This bit indicates the full status of the generic read > > + * payload FIFO. > > + * Value after reset: 0x0 > > + */ > > + u32 gen_cmd_rdata_fifo_full: 1; > > + > > + /* > > + * This bit indicates the empty status of the generic write > > + * payload FIFO. > > + * Value after reset: 0x1 > > + */ > > + u32 gen_cmd_wdata_fifo_empty: 1; > > + > > + /* > > + * This bit indicates the full status of the generic write > > + * payload FIFO. > > + * Value after reset: 0x0 > > + */ > > + u32 gen_cmd_wdata_fifo_full: 1; > > + > > + /* > > + * This bit indicates the empty status of the generic > > + * command FIFO. > > + * Value after reset: 0x1 > > + */ > > + u32 gen_cmd_cmd_fifo_empty: 1; > > + > > + /* > > + * This bit indicates the full status of the generic > > + * command FIFO. > > + * Value after reset: 0x0 > > + */ > > + u32 gen_cmd_cmd_fifo_full: 1; > > + > > + /* > > + * This bit is set when the entire response of read is > > + * stored in the rx payload FIFO. And it will be cleared > > + * automaticlly after read this bit each time. > > + * Value after reset: 0x0 > > + * > > + * NOTE: this bit is just supported for dsi-ctrl-r1p0 IP > > + */ > > + u32 gen_cmd_rdcmd_done: 1; > > + > > + u32 reserved : 24; > > + > > + } bits; > > + } CMD_MODE_STATUS; > > + > > + union _0x9C { > > + u32 val; > > + struct _PHY_STATUS { > > + /* the status of phydirection D-PHY signal */ > > + u32 phy_direction: 1; > > + > > + /* the status of phylock D-PHY signal */ > > + u32 phy_lock: 1; > > + > > + /* the status of rxulpsesc0lane D-PHY signal */ > > + u32 phy_rxulpsesc0lane: 1; > > + > > + /* the status of phystopstateclklane D-PHY signal */ > > + u32 phy_stopstateclklane: 1; > > + > > + /* the status of phystopstate0lane D-PHY signal */ > > + u32 phy_stopstate0lane: 1; > > + > > + /* the status of phystopstate1lane D-PHY signal */ > > + u32 phy_stopstate1lane: 1; > > + > > + /* the status of phystopstate2lane D-PHY signal */ > > + u32 phy_stopstate2lane: 1; > > + > > + /* the status of phystopstate3lane D-PHY signal */ > > + u32 phy_stopstate3lane: 1; > > + > > + /* the status of phyulpsactivenotclk D-PHY signal */ > > + u32 phy_ulpsactivenotclk: 1; > > + > > + /* the status of ulpsactivenot0lane D-PHY signal */ > > + u32 phy_ulpsactivenot0lane: 1; > > + > > + /* the status of ulpsactivenot1lane D-PHY signal */ > > + u32 phy_ulpsactivenot1lane: 1; > > + > > + /* the status of ulpsactivenot2lane D-PHY signal */ > > + u32 phy_ulpsactivenot2lane: 1; > > + > > + /* the status of ulpsactivenot3lane D-PHY signal */ > > + u32 phy_ulpsactivenot3lane: 1; > > + > > + u32 reserved: 19; > > + > > + } bits; > > + } PHY_STATUS; > > + > > + union _0xA0 { > > + u32 val; > > + struct _PHY_MIN_STOP_TIME { > > + /* This field configures the minimum wait period to request > > + * a high-speed transmission after the Stop state. > > + */ > > + u32 phy_min_stop_time: 8; > > + > > + u32 reserved: 24; > > + } bits; > > + } PHY_MIN_STOP_TIME; > > + > > + union _0xA4 { > > + u32 val; > > + struct _PHY_LANE_NUM_CONFIG { > > + /* > > + * This field configures the number of active data lanes: > > + * 00: One data lane (lane 0) > > + * 01: Two data lanes (lanes 0 and 1) > > + * 10: Three data lanes (lanes 0, 1, and 2) > > + * 11: Four data lanes (lanes 0, 1, 2, and 3) > > + */ > > + u32 phy_lane_num: 2; > > + > > + u32 reserved: 30; > > + > > + } bits; > > + } PHY_LANE_NUM_CONFIG; > > + > > + union _0xA8 { > > + u32 val; > > + struct _PHY_CLKLANE_TIME_CONFIG { > > + /* > > + * This field configures the maximum time that the D-PHY > > + * clock lane takes to go from low-power to high-speed > > + * transmission measured in lane byte clock cycles. > > + */ > > + u32 phy_clklane_lp_to_hs_time: 16; > > + > > + /* > > + * This field configures the maximum time that the D-PHY > > + * clock lane takes to go from high-speed to low-power > > + * transmission measured in lane byte clock cycles. > > + */ > > + u32 phy_clklane_hs_to_lp_time: 16; > > + > > + } bits; > > + } PHY_CLKLANE_TIME_CONFIG; > > + > > + union _0xAC { > > + u32 val; > > + struct _PHY_DATALANE_TIME_CONFIG { > > + /* > > + * This field configures the maximum time that the D-PHY data > > + * lanes take to go from low-power to high-speed transmission > > + * measured in lane byte clock cycles. > > + */ > > + u32 phy_datalane_lp_to_hs_time: 16; > > + > > + /* > > + * This field configures the maximum time that the D-PHY data > > + * lanes take to go from high-speed to low-power transmission > > + * measured in lane byte clock cycles. > > + */ > > + u32 phy_datalane_hs_to_lp_time: 16; > > + > > + } bits; > > + } PHY_DATALANE_TIME_CONFIG; > > + > > + union _0xB0 { > > + u32 val; > > + struct _MAX_READ_TIME { > > + /* > > + * This field configures the maximum time required to perform > > + * a read command in lane byte clock cycles. This register can > > + * only be modified when no read command is in progress. > > + */ > > + u32 max_rd_time: 16; > > + > > + u32 reserved: 16; > > + > > + } bits; > > + } MAX_READ_TIME; > > + > > + union _0xB4 { > > + u32 val; > > + struct _RX_PKT_CHECK_CONFIG { > > + /* When set to 1, this bit enables the ECC reception, error > > + * correction, and reporting. > > + */ > > + u32 rx_pkt_ecc_en: 1; > > + > > + /* When set to 1, this bit enables the CRC reception and error > > + * reporting. > > + */ > > + u32 rx_pkt_crc_en: 1; > > + > > + u32 reserved: 30; > > + > > + } bits; > > + } RX_PKT_CHECK_CONFIG; > > + > > + union _0xB8 { > > + u32 val; > > + struct _TA_EN { > > + /* When set to 1, this bit enables the Bus Turn-Around (BTA) > > + * request. > > + */ > > + u32 ta_en: 1; > > + > > + u32 reserved: 31; > > + > > + } bits; > > + } TA_EN; > > + > > + union _0xBC { > > + u32 val; > > + struct _EOTP_EN { > > + /* When set to 1, this bit enables the EoTp transmission */ > > + u32 tx_eotp_en: 1; > > + > > + /* When set to 1, this bit enables the EoTp reception. */ > > + u32 rx_eotp_en: 1; > > + > > + u32 reserved: 30; > > + > > + } bits; > > + } EOTP_EN; > > + > > + union _0xC0 { > > + u32 val; > > + struct _VIDEO_NULLPKT_SIZE { > > + /* > > + * This register configures the number of bytes inside a null > > + * packet. Setting it to 0 disables the null packets. > > + */ > > + u32 video_nullpkt_size: 13; > > + > > + u32 reserved: 19; > > + > > + } bits; > > + } VIDEO_NULLPKT_SIZE; > > + > > + union _0xC4 { > > + u32 val; > > + struct _DCS_WM_PKT_SIZE { > > + /* > > + * This field configures the maximum allowed size for an eDPI > > + * write memory command, measured in pixels. Automatic parti- > > + * tioning of data obtained from eDPI is permanently enabled. > > + */ > > + u32 dcs_wm_pkt_size: 16; > > + > > + u32 reserved: 16; > > + } bits; > > + } DCS_WM_PKT_SIZE; > > + > > + union _0xC8 { > > + u32 val; > > + struct _PROTOCOL_INT_CLR { > > + u32 clr_dphy_errors_0: 1; > > + u32 clr_dphy_errors_1: 1; > > + u32 clr_dphy_errors_2: 1; > > + u32 clr_dphy_errors_3: 1; > > + u32 clr_dphy_errors_4: 1; > > + u32 clr_protocol_debug_err: 11; > > + u32 clr_ack_with_err_0: 1; > > + u32 clr_ack_with_err_1: 1; > > + u32 clr_ack_with_err_2: 1; > > + u32 clr_ack_with_err_3: 1; > > + u32 clr_ack_with_err_4: 1; > > + u32 clr_ack_with_err_5: 1; > > + u32 clr_ack_with_err_6: 1; > > + u32 clr_ack_with_err_7: 1; > > + u32 clr_ack_with_err_8: 1; > > + u32 clr_ack_with_err_9: 1; > > + u32 clr_ack_with_err_10: 1; > > + u32 clr_ack_with_err_11: 1; > > + u32 clr_ack_with_err_12: 1; > > + u32 clr_ack_with_err_13: 1; > > + u32 clr_ack_with_err_14: 1; > > + u32 clr_ack_with_err_15: 1; > > + } bits; > > + } PROTOCOL_INT_CLR; > > + > > + union _0xCC { > > + u32 val; > > + struct _INTERNAL_INT_CLR { > > + u32 clr_receive_pkt_size_err: 1; > > + u32 clr_eopt_not_receive_err: 1; > > + u32 clr_gen_cmd_cmd_fifo_wr_err: 1; > > + u32 clr_gen_cmd_rdata_fifo_rd_err: 1; > > + u32 clr_gen_cmd_rdata_fifo_wr_err: 1; > > + u32 clr_gen_cmd_wdata_fifo_wr_err: 1; > > + u32 clr_gen_cmd_wdata_fifo_rd_err: 1; > > + u32 clr_dpi_pix_fifo_wr_err: 1; > > + u32 clr_internal_debug_err: 19; > > + u32 clr_ecc_single_err: 1; > > + u32 clr_ecc_multi_err: 1; > > + u32 clr_crc_err: 1; > > + u32 clr_hs_tx_timeout: 1; > > + u32 clr_lp_rx_timeout: 1; > > + } bits; > > + } INTERNAL_INT_CLR; > > + > > + union _0xD0 { > > + u32 val; > > + struct _VIDEO_SIG_DELAY_CONFIG { > > + > > + /* > > + * DPI interface signal delay to be used in clk lanebyte > > + * domain for control logic to read video data from pixel > > + * memory in mannal mode, measured in clk_lanebyte cycles > > + */ > > + u32 video_sig_delay: 24; > > + > > + /* > > + * 1'b1: mannal mode > > + * dsi controller will use video_sig_delay value as > > + * the delay for the packet handle logic to read video > > + * data from pixel memory. > > + * > > + * 1'b0: auto mode > > + * dsi controller will auto calculate the delay for > > + * the packet handle logic to read video data from > > + * pixel memory. > > + */ > > + u32 video_sig_delay_mode: 1; > > + > > + u32 reserved: 7; > > + } bits; > > + } VIDEO_SIG_DELAY_CONFIG; > > + > > + u32 reservedD4_EC[7]; > > + > > + union _0xF0 { > > + u32 val; > > + struct _PHY_TST_CTRL0 { > > + /* PHY test interface clear (active high) */ > > + u32 phy_testclr: 1; > > + > > + /* This bit is used to clock the TESTDIN bus into the D-PHY */ > > + u32 phy_testclk: 1; > > + > > + u32 reserved: 30; > > + } bits; > > + } PHY_TST_CTRL0; > > + > > + union _0xF4 { > > + u32 val; > > + struct _PHY_TST_CTRL1 { > > + /* PHY test interface input 8-bit data bus for internal > > + * register programming and test functionalities access. > > + */ > > + u32 phy_testdin: 8; > > + > > + /* PHY output 8-bit data bus for read-back and internal > > + * probing functionalities. > > + */ > > + u32 phy_testdout: 8; > > + > > + /* > > + * PHY test interface operation selector: > > + * 1: The address write operation is set on the falling edge > > + * of the testclk signal. > > + * 0: The data write operation is set on the rising edge of > > + * the testclk signal. > > + */ > > + u32 phy_testen: 1; > > + > > + u32 reserved: 15; > > + } bits; > > + } PHY_TST_CTRL1; > > + > > + u32 reservedF8_1FC[66]; > > + > > + union _0x200 { > > + u32 val; > > + struct _INT_PLL_STS { > > + u32 int_pll_sts: 1; > > + u32 reserved: 31; > > + } bits; > > + } INT_PLL_STS; > > + > > + union _0x204 { > > + u32 val; > > + struct _INT_PLL_MSK { > > + u32 int_pll_msk: 1; > > + u32 reserved: 31; > > + } bits; > > + } INT_PLL_MSK; > > + > > + union _0x208 { > > + u32 val; > > + struct _INT_PLL_CLR { > > + u32 int_pll_clr: 1; > > + u32 reserved: 31; > > + } bits; > > + } INT_PLL_CLR; > > + > > +}; > > + > > +void dsi_power_enable(struct dsi_context *ctx, int enable); > > +void dsi_video_mode(struct dsi_context *ctx); > > +void dsi_cmd_mode(struct dsi_context *ctx); > > +bool dsi_is_cmd_mode(struct dsi_context *ctx); > > +void dsi_rx_vcid(struct dsi_context *ctx, u8 vc); > > +void dsi_video_vcid(struct dsi_context *ctx, u8 vc); > > +void dsi_dpi_video_burst_mode(struct dsi_context *ctx, int mode); > > +void dsi_dpi_color_coding(struct dsi_context *ctx, int coding); > > +void dsi_dpi_sig_delay(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_dpi_hline_time(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_dpi_hsync_time(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_dpi_hbp_time(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_dpi_vact(struct dsi_context *ctx, u16 lines); > > +void dsi_dpi_vfp(struct dsi_context *ctx, u16 lines); > > +void dsi_dpi_vbp(struct dsi_context *ctx, u16 lines); > > +void dsi_dpi_vsync(struct dsi_context *ctx, u16 lines); > > +void dsi_dpi_hporch_lp_en(struct dsi_context *ctx, int enable); > > +void dsi_dpi_vporch_lp_en(struct dsi_context *ctx, int enable); > > +void dsi_dpi_frame_ack_en(struct dsi_context *ctx, int enable); > > +void dsi_dpi_chunk_num(struct dsi_context *ctx, u16 no); > > +void dsi_dpi_null_packet_size(struct dsi_context *ctx, u16 size); > > +void dsi_dpi_video_packet_size(struct dsi_context *ctx, u16 size); > > +void dsi_edpi_max_pkt_size(struct dsi_context *ctx, u16 size); > > +void dsi_tear_effect_ack_en(struct dsi_context *ctx, int enable); > > +void dsi_cmd_mode_lp_cmd_en(struct dsi_context *ctx, int enable); > > +void dsi_video_mode_lp_cmd_en(struct dsi_context *ctx, int enable); > > +void dsi_set_packet_header(struct dsi_context *ctx, u8 vc, u8 type, > > + u8 wc_lsb, u8 wc_msb); > > +void dsi_set_packet_payload(struct dsi_context *ctx, u32 payload); > > +u32 dsi_get_rx_payload(struct dsi_context *ctx); > > +void dsi_bta_en(struct dsi_context *ctx, int enable); > > +void dsi_eotp_rx_en(struct dsi_context *ctx, int enable); > > +void dsi_eotp_tx_en(struct dsi_context *ctx, int enable); > > +void dsi_ecc_rx_en(struct dsi_context *ctx, int enable); > > +void dsi_crc_rx_en(struct dsi_context *ctx, int enable); > > +bool dsi_is_bta_returned(struct dsi_context *ctx); > > +bool dsi_is_rx_payload_fifo_full(struct dsi_context *ctx); > > +bool dsi_is_rx_payload_fifo_empty(struct dsi_context *ctx); > > +bool dsi_is_tx_payload_fifo_full(struct dsi_context *ctx); > > +bool dsi_is_tx_payload_fifo_empty(struct dsi_context *ctx); > > +bool dsi_is_tx_cmd_fifo_empty(struct dsi_context *ctx); > > +void dsi_datalane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_datalane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_clklane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_clklane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_max_read_time(struct dsi_context *ctx, u16 byte_cycle); > > +void dsi_nc_clk_en(struct dsi_context *ctx, int enable); > > +void dsi_tx_escape_division(struct dsi_context *ctx, u8 div); > > +void dsi_timeout_clock_division(struct dsi_context *ctx, u8 div); > > +void dsi_lp_rx_timeout(struct dsi_context *ctx, u16 count); > > +void dsi_hs_tx_timeout(struct dsi_context *ctx, u16 count); > > +u32 dsi_int0_status(struct dsi_context *ctx); > > +u32 dsi_int1_status(struct dsi_context *ctx); > > +void dsi_int0_mask(struct dsi_context *ctx, u32 mask); > > +void dsi_int1_mask(struct dsi_context *ctx, u32 mask); > > + > > +#endif /* _DSI_CTRL_H_ */ > > diff --git a/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c > > new file mode 100644 > > index 000000000..6b3259ae5 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c > > @@ -0,0 +1,157 @@ > > +// SPDX-License-Identifier: GPL-2.0 > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#include <linux/io.h> > > +#include <linux/init.h> > > +#include <linux/module.h> > > + > > +#include "dsi_ctrl.h" > > +#include "dsi_ctrl_ppi.h" > > + > > +/* > > + * Reset D-PHY module > > + */ > > +void dsi_phy_rstz(struct dsi_context *ctx, int level) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x78 phy_interface_ctrl; > > + > > + phy_interface_ctrl.val = readl(®->PHY_INTERFACE_CTRL); > > + phy_interface_ctrl.bits.rf_phy_reset_n = level; > > + > > + writel(phy_interface_ctrl.val, ®->PHY_INTERFACE_CTRL); > > +} > > + > > +/* > > + * Power up/down D-PHY module > > + */ > > +void dsi_phy_shutdownz(struct dsi_context *ctx, int level) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x78 phy_interface_ctrl; > > + > > + phy_interface_ctrl.val = readl(®->PHY_INTERFACE_CTRL); > > + phy_interface_ctrl.bits.rf_phy_shutdown = level; > > + > > + writel(phy_interface_ctrl.val, ®->PHY_INTERFACE_CTRL); > > +} > > + > > +/* > > + * Configure minimum wait period for HS transmission request after a stop state > > + */ > > +void dsi_phy_stop_wait_time(struct dsi_context *ctx, u8 byte_cycle) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(byte_cycle, ®->PHY_MIN_STOP_TIME); > > +} > > + > > +/* > > + * Set number of active lanes > > + */ > > +void dsi_phy_datalane_en(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + > > + writel(ctx->lanes - 1, ®->PHY_LANE_NUM_CONFIG); > > +} > > + > > +/* > > + * Enable clock lane module > > + */ > > +void dsi_phy_clklane_en(struct dsi_context *ctx, int en) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x78 phy_interface_ctrl; > > + > > + phy_interface_ctrl.val = readl(®->PHY_INTERFACE_CTRL); > > + phy_interface_ctrl.bits.rf_phy_clk_en = en; > > + > > + writel(phy_interface_ctrl.val, ®->PHY_INTERFACE_CTRL); > > +} > > + > > +/* > > + * Request the PHY module to start transmission of high speed clock. > > + * This causes the clock lane to start transmitting DDR clock on the > > + * lane interconnect. > > + */ > > +void dsi_phy_clk_hs_rqst(struct dsi_context *ctx, int enable) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x74 phy_clk_lane_lp_ctrl; > > + > > + phy_clk_lane_lp_ctrl.val = readl(®->PHY_CLK_LANE_LP_CTRL); > > + phy_clk_lane_lp_ctrl.bits.auto_clklane_ctrl_en = 0; > > + phy_clk_lane_lp_ctrl.bits.phy_clklane_tx_req_hs = enable; > > + > > + writel(phy_clk_lane_lp_ctrl.val, ®->PHY_CLK_LANE_LP_CTRL); > > +} > > + > > +/* > > + * Get D-PHY PPI status > > + */ > > +u8 dsi_phy_is_pll_locked(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0x9C phy_status; > > + > > + phy_status.val = readl(®->PHY_STATUS); > > + > > + return phy_status.bits.phy_lock; > > +} > > + > > +void dsi_phy_test_clk(struct dsi_context *ctx, u8 value) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xF0 phy_tst_ctrl0; > > + > > + phy_tst_ctrl0.val = readl(®->PHY_TST_CTRL0); > > + phy_tst_ctrl0.bits.phy_testclk = value; > > + > > + writel(phy_tst_ctrl0.val, ®->PHY_TST_CTRL0); > > +} > > + > > +void dsi_phy_test_clr(struct dsi_context *ctx, u8 value) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xF0 phy_tst_ctrl0; > > + > > + phy_tst_ctrl0.val = readl(®->PHY_TST_CTRL0); > > + phy_tst_ctrl0.bits.phy_testclr = value; > > + > > + writel(phy_tst_ctrl0.val, ®->PHY_TST_CTRL0); > > +} > > + > > +void dsi_phy_test_en(struct dsi_context *ctx, u8 value) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xF4 phy_tst_ctrl1; > > + > > + phy_tst_ctrl1.val = readl(®->PHY_TST_CTRL1); > > + phy_tst_ctrl1.bits.phy_testen = value; > > + > > + writel(phy_tst_ctrl1.val, ®->PHY_TST_CTRL1); > > +} > > + > > +u8 dsi_phy_test_dout(struct dsi_context *ctx) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xF4 phy_tst_ctrl1; > > + > > + phy_tst_ctrl1.val = readl(®->PHY_TST_CTRL1); > > + > > + return phy_tst_ctrl1.bits.phy_testdout; > > +} > > + > > +void dsi_phy_test_din(struct dsi_context *ctx, u8 data) > > +{ > > + struct dsi_reg *reg = (struct dsi_reg *)ctx->base; > > + union _0xF4 phy_tst_ctrl1; > > + > > + phy_tst_ctrl1.val = readl(®->PHY_TST_CTRL1); > > + phy_tst_ctrl1.bits.phy_testdin = data; > > + > > + writel(phy_tst_ctrl1.val, ®->PHY_TST_CTRL1); > > +} > > diff --git a/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h > > new file mode 100644 > > index 000000000..6fb3fad89 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h > > @@ -0,0 +1,26 @@ > > +/* SPDX-License-Identifier: GPL-2.0 */ > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#ifndef _DSI_CTRL_PPI_H_ > > +#define _DSI_CTRL_PPI_H_ > > + > > +#include "sprd_dsi.h" > > + > > +void dsi_phy_rstz(struct dsi_context *ctx, int level); > > +void dsi_phy_shutdownz(struct dsi_context *ctx, int level); > > +void dsi_phy_force_pll(struct dsi_context *ctx, int force); > > +void dsi_phy_stop_wait_time(struct dsi_context *ctx, u8 byte_clk); > > +void dsi_phy_datalane_en(struct dsi_context *ctx); > > +void dsi_phy_clklane_en(struct dsi_context *ctx, int en); > > +void dsi_phy_clk_hs_rqst(struct dsi_context *ctx, int en); > > +u8 dsi_phy_is_pll_locked(struct dsi_context *ctx); > > +void dsi_phy_test_clk(struct dsi_context *ctx, u8 level); > > +void dsi_phy_test_clr(struct dsi_context *ctx, u8 level); > > +void dsi_phy_test_en(struct dsi_context *ctx, u8 level); > > +u8 dsi_phy_test_dout(struct dsi_context *ctx); > > +void dsi_phy_test_din(struct dsi_context *ctx, u8 data); > > +void dsi_phy_bist_en(struct dsi_context *ctx, int en); > > + > > +#endif /* _DSI_CTRL_PPI_H_ */ > > diff --git a/drivers/gpu/drm/sprd/megacores_pll.c b/drivers/gpu/drm/sprd/megacores_pll.c > > new file mode 100644 > > index 000000000..03c1f0f00 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/megacores_pll.c > > @@ -0,0 +1,317 @@ > > +// SPDX-License-Identifier: GPL-2.0 > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#include <asm/div64.h> > > +#include <linux/delay.h> > > +#include <linux/init.h> > > +#include <linux/kernel.h> > > +#include <linux/regmap.h> > > +#include <linux/string.h> > > + > > +#include "megacores_pll.h" > > + > > +#define L 0 > > +#define H 1 > > +#define CLK 0 > > +#define DATA 1 > > +#define INFINITY 0xffffffff > > +#define MIN_OUTPUT_FREQ (100) > > + > > +#define AVERAGE(a, b) (min(a, b) + abs((b) - (a)) / 2) > > + > > +/* sharkle */ > > +#define VCO_BAND_LOW 750 > > +#define VCO_BAND_MID 1100 > > +#define VCO_BAND_HIGH 1500 > > +#define PHY_REF_CLK 26000 > > + > > +static int dphy_calc_pll_param(struct dphy_pll *pll) > > +{ > > + const u32 khz = 1000; > > + const u32 mhz = 1000000; > > + const unsigned long long factor = 100; > > + unsigned long long tmp; > > + int i; > > + > > + pll->potential_fvco = pll->freq / khz; > > + pll->ref_clk = PHY_REF_CLK / khz; > > + > > + for (i = 0; i < 4; ++i) { > > + if (pll->potential_fvco >= VCO_BAND_LOW && > > + pll->potential_fvco <= VCO_BAND_HIGH) { > > + pll->fvco = pll->potential_fvco; > > + pll->out_sel = BIT(i); > > + break; > > + } > > + pll->potential_fvco <<= 1; > > + } > > + if (pll->fvco == 0) > > + return -EINVAL; > > + > > + if (pll->fvco >= VCO_BAND_LOW && pll->fvco <= VCO_BAND_MID) { > > + /* vco band control */ > > + pll->vco_band = 0x0; > > + /* low pass filter control */ > > + pll->lpf_sel = 1; > > + } else if (pll->fvco > VCO_BAND_MID && pll->fvco <= VCO_BAND_HIGH) { > > + pll->vco_band = 0x1; > > + pll->lpf_sel = 0; > > + } else > > + return -EINVAL; > > + > > + pll->nint = pll->fvco / pll->ref_clk; > > + tmp = pll->fvco * factor * mhz; > > + do_div(tmp, pll->ref_clk); > > + tmp = tmp - pll->nint * factor * mhz; > > + tmp *= BIT(20); > > + do_div(tmp, 100000000); > > + pll->kint = (u32)tmp; > > + pll->refin = 3; /* pre-divider bypass */ > > + pll->sdm_en = true; /* use fraction N PLL */ > > + pll->fdk_s = 0x1; /* fraction */ > > + pll->cp_s = 0x0; > > + pll->det_delay = 0x1; > > + > > + return 0; > > +} > > + > > +static void dphy_set_pll_reg(struct dphy_pll *pll, struct regmap *regmap) > > +{ > > + struct pll_reg *reg = &pll->reg; > > + u8 *val; > > + int i; > > + > > + u8 reg_addr[] = { > > + 0x03, 0x04, 0x06, 0x08, 0x09, > > + 0x0a, 0x0b, 0x0e, 0x0f > > + }; > > + > > + reg->_03.bits.prbs_bist = 1; > > + reg->_03.bits.en_lp_treot = true; > > + reg->_03.bits.lpf_sel = pll->lpf_sel; > > + reg->_03.bits.txfifo_bypass = 0; > > + reg->_04.bits.div = pll->div; > > + reg->_04.bits.masterof8lane = 1; > > + reg->_04.bits.cp_s = pll->cp_s; > > + reg->_04.bits.fdk_s = pll->fdk_s; > > + reg->_06.bits.nint = pll->nint; > > + reg->_08.bits.vco_band = pll->vco_band; > > + reg->_08.bits.sdm_en = pll->sdm_en; > > + reg->_08.bits.refin = pll->refin; > > + reg->_09.bits.kint_h = pll->kint >> 12; > > + reg->_0a.bits.kint_m = (pll->kint >> 4) & 0xff; > > + reg->_0b.bits.out_sel = pll->out_sel; > > + reg->_0b.bits.kint_l = pll->kint & 0xf; > > + reg->_0e.bits.pll_pu_byp = 0; > > + reg->_0e.bits.pll_pu = 0; > > + reg->_0e.bits.stopstate_sel = 1; > > + reg->_0f.bits.det_delay = pll->det_delay; > > + > > + val = (u8 *)® > > + > > + for (i = 0; i < sizeof(reg_addr); ++i) { > > + regmap_write(regmap, reg_addr[i], val[i]); > > + DRM_DEBUG("%02x: %02x\n", reg_addr[i], val[i]); > > + } > > +} > > + > > +int dphy_pll_config(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + struct regmap *regmap = ctx->regmap; > > + struct dphy_pll *pll = ctx->pll; > > + int ret; > > + > > + pll->freq = ctx->byte_clk * 8; > > + > > + /* FREQ = 26M * (NINT + KINT / 2^20) / out_sel */ > > + ret = dphy_calc_pll_param(pll); > > + if (ret) { > > + drm_err(dsi->drm, "failed to calculate dphy pll parameters\n"); > > + return ret; > > + } > > + dphy_set_pll_reg(pll, regmap); > > + > > + return 0; > > +} > > + > > +static void dphy_set_timing_reg(struct regmap *regmap, int type, u8 val[]) > > +{ > > + switch (type) { > > + case REQUEST_TIME: > > + regmap_write(regmap, 0x31, val[CLK]); > > + regmap_write(regmap, 0x41, val[DATA]); > > + regmap_write(regmap, 0x51, val[DATA]); > > + regmap_write(regmap, 0x61, val[DATA]); > > + regmap_write(regmap, 0x71, val[DATA]); > > + > > + regmap_write(regmap, 0x90, val[CLK]); > > + regmap_write(regmap, 0xa0, val[DATA]); > > + regmap_write(regmap, 0xb0, val[DATA]); > > + regmap_write(regmap, 0xc0, val[DATA]); > > + regmap_write(regmap, 0xd0, val[DATA]); > > + break; > > + case PREPARE_TIME: > > + regmap_write(regmap, 0x32, val[CLK]); > > + regmap_write(regmap, 0x42, val[DATA]); > > + regmap_write(regmap, 0x52, val[DATA]); > > + regmap_write(regmap, 0x62, val[DATA]); > > + regmap_write(regmap, 0x72, val[DATA]); > > + > > + regmap_write(regmap, 0x91, val[CLK]); > > + regmap_write(regmap, 0xa1, val[DATA]); > > + regmap_write(regmap, 0xb1, val[DATA]); > > + regmap_write(regmap, 0xc1, val[DATA]); > > + regmap_write(regmap, 0xd1, val[DATA]); > > + break; > > + case ZERO_TIME: > > + regmap_write(regmap, 0x33, val[CLK]); > > + regmap_write(regmap, 0x43, val[DATA]); > > + regmap_write(regmap, 0x53, val[DATA]); > > + regmap_write(regmap, 0x63, val[DATA]); > > + regmap_write(regmap, 0x73, val[DATA]); > > + > > + regmap_write(regmap, 0x92, val[CLK]); > > + regmap_write(regmap, 0xa2, val[DATA]); > > + regmap_write(regmap, 0xb2, val[DATA]); > > + regmap_write(regmap, 0xc2, val[DATA]); > > + regmap_write(regmap, 0xd2, val[DATA]); > > + break; > > + case TRAIL_TIME: > > + regmap_write(regmap, 0x34, val[CLK]); > > + regmap_write(regmap, 0x44, val[DATA]); > > + regmap_write(regmap, 0x54, val[DATA]); > > + regmap_write(regmap, 0x64, val[DATA]); > > + regmap_write(regmap, 0x74, val[DATA]); > > + > > + regmap_write(regmap, 0x93, val[CLK]); > > + regmap_write(regmap, 0xa3, val[DATA]); > > + regmap_write(regmap, 0xb3, val[DATA]); > > + regmap_write(regmap, 0xc3, val[DATA]); > > + regmap_write(regmap, 0xd3, val[DATA]); > > + break; > > + case EXIT_TIME: > > + regmap_write(regmap, 0x36, val[CLK]); > > + regmap_write(regmap, 0x46, val[DATA]); > > + regmap_write(regmap, 0x56, val[DATA]); > > + regmap_write(regmap, 0x66, val[DATA]); > > + regmap_write(regmap, 0x76, val[DATA]); > > + > > + regmap_write(regmap, 0x95, val[CLK]); > > + regmap_write(regmap, 0xA5, val[DATA]); > > + regmap_write(regmap, 0xB5, val[DATA]); > > + regmap_write(regmap, 0xc5, val[DATA]); > > + regmap_write(regmap, 0xd5, val[DATA]); > > + break; > > + case CLKPOST_TIME: > > + regmap_write(regmap, 0x35, val[CLK]); > > + regmap_write(regmap, 0x94, val[CLK]); > > + break; > > + > > + /* the following just use default value */ > > + case SETTLE_TIME: > > + case TA_GET: > > + case TA_GO: > > + case TA_SURE: > > + break; > > + default: > > + break; > > + } > > +} > > + > > +void dphy_timing_config(struct dsi_context *ctx) > > +{ > > + struct regmap *regmap = ctx->regmap; > > + struct dphy_pll *pll = ctx->pll; > > + const u32 factor = 2; > > + const u32 scale = 100; > > + u32 t_ui, t_byteck, t_half_byteck; > > + u32 range[2], constant; > > + u8 val[2]; > > + u32 tmp = 0; > > + > > + /* t_ui: 1 ui, byteck: 8 ui, half byteck: 4 ui */ > > + t_ui = 1000 * scale / (pll->freq / 1000); > > + t_byteck = t_ui << 3; > > + t_half_byteck = t_ui << 2; > > + constant = t_ui << 1; > > + > > + /* REQUEST_TIME: HS T-LPX: LP-01 > > + * For T-LPX, mipi spec defined min value is 50ns, > > + * but maybe it shouldn't be too small, because BTA, > > + * LP-10, LP-00, LP-01, all of this is related to T-LPX. > > + */ > > + range[L] = 50 * scale; > > + range[H] = INFINITY; > > + val[CLK] = DIV_ROUND_UP(range[L] * (factor << 1), t_byteck) - 2; > > + val[DATA] = val[CLK]; > > + dphy_set_timing_reg(regmap, REQUEST_TIME, val); > > + > > + /* PREPARE_TIME: HS sequence: LP-00 */ > > + range[L] = 38 * scale; > > + range[H] = 95 * scale; > > + tmp = AVERAGE(range[L], range[H]); > > + val[CLK] = DIV_ROUND_UP(AVERAGE(range[L], range[H]), > > + t_half_byteck) - 1; > > + range[L] = 40 * scale + 4 * t_ui; > > + range[H] = 85 * scale + 6 * t_ui; > > + tmp |= AVERAGE(range[L], range[H]) << 16; > > + val[DATA] = DIV_ROUND_UP(AVERAGE(range[L], range[H]), > > + t_half_byteck) - 1; > > + dphy_set_timing_reg(regmap, PREPARE_TIME, val); > > + > > + /* ZERO_TIME: HS-ZERO */ > > + range[L] = 300 * scale; > > + range[H] = INFINITY; > > + val[CLK] = DIV_ROUND_UP(range[L] * factor + (tmp & 0xffff) > > + - 525 * t_byteck / 100, t_byteck) - 2; > > + range[L] = 145 * scale + 10 * t_ui; > > + val[DATA] = DIV_ROUND_UP(range[L] * factor > > + + ((tmp >> 16) & 0xffff) - 525 * t_byteck / 100, > > + t_byteck) - 2; > > + dphy_set_timing_reg(regmap, ZERO_TIME, val); > > + > > + /* TRAIL_TIME: HS-TRAIL */ > > + range[L] = 60 * scale; > > + range[H] = INFINITY; > > + val[CLK] = DIV_ROUND_UP(range[L] * factor - constant, t_half_byteck); > > + range[L] = max(8 * t_ui, 60 * scale + 4 * t_ui); > > + val[DATA] = DIV_ROUND_UP(range[L] * 3 / 2 - constant, t_half_byteck) - 2; > > + dphy_set_timing_reg(regmap, TRAIL_TIME, val); > > + > > + /* EXIT_TIME: */ > > + range[L] = 100 * scale; > > + range[H] = INFINITY; > > + val[CLK] = DIV_ROUND_UP(range[L] * factor, t_byteck) - 2; > > + val[DATA] = val[CLK]; > > + dphy_set_timing_reg(regmap, EXIT_TIME, val); > > + > > + /* CLKPOST_TIME: */ > > + range[L] = 60 * scale + 52 * t_ui; > > + range[H] = INFINITY; > > + val[CLK] = DIV_ROUND_UP(range[L] * factor, t_byteck) - 2; > > + val[DATA] = val[CLK]; > > + dphy_set_timing_reg(regmap, CLKPOST_TIME, val); > > + > > + /* SETTLE_TIME: > > + * This time is used for receiver. So for transmitter, > > + * it can be ignored. > > + */ > > + > > + /* TA_GO: > > + * transmitter drives bridge state(LP-00) before releasing control, > > + * reg 0x1f default value: 0x04, which is good. > > + */ > > + > > + /* TA_SURE: > > + * After LP-10 state and before bridge state(LP-00), > > + * reg 0x20 default value: 0x01, which is good. > > + */ > > + > > + /* TA_GET: > > + * receiver drives Bridge state(LP-00) before releasing control > > + * reg 0x21 default value: 0x03, which is good. > > + */ > > +} > > diff --git a/drivers/gpu/drm/sprd/megacores_pll.h b/drivers/gpu/drm/sprd/megacores_pll.h > > new file mode 100644 > > index 000000000..bf20aae65 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/megacores_pll.h > > @@ -0,0 +1,146 @@ > > +/* SPDX-License-Identifier: GPL-2.0 */ > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#ifndef _MEGACORES_PLL_H_ > > +#define _MEGACORES_PLL_H_ > > + > > +#include "sprd_dsi.h" > > + > > +enum PLL_TIMING { > > + NONE, > > + REQUEST_TIME, > > + PREPARE_TIME, > > + SETTLE_TIME, > > + ZERO_TIME, > > + TRAIL_TIME, > > + EXIT_TIME, > > + CLKPOST_TIME, > > + TA_GET, > > + TA_GO, > > + TA_SURE, > > + TA_WAIT, > > +}; > > + > > +struct pll_reg { > > + union __reg_03__ { > > + struct __03 { > > + u8 prbs_bist: 1; > > + u8 en_lp_treot: 1; > > + u8 lpf_sel: 4; > > + u8 txfifo_bypass: 1; > > + u8 freq_hopping: 1; > > + } bits; > > + u8 val; > > + } _03; > > + union __reg_04__ { > > + struct __04 { > > + u8 div: 3; > > + u8 masterof8lane: 1; > > + u8 hop_trig: 1; > > + u8 cp_s: 2; > > + u8 fdk_s: 1; > > + } bits; > > + u8 val; > > + } _04; > > + union __reg_06__ { > > + struct __06 { > > + u8 nint: 7; > > + u8 mod_en: 1; > > + } bits; > > + u8 val; > > + } _06; > > + union __reg_07__ { > > + struct __07 { > > + u8 kdelta_h: 8; > > + } bits; > > + u8 val; > > + } _07; > > + union __reg_08__ { > > + struct __08 { > > + u8 vco_band: 1; > > + u8 sdm_en: 1; > > + u8 refin: 2; > > + u8 kdelta_l: 4; > > + } bits; > > + u8 val; > > + } _08; > > + union __reg_09__ { > > + struct __09 { > > + u8 kint_h: 8; > > + } bits; > > + u8 val; > > + } _09; > > + union __reg_0a__ { > > + struct __0a { > > + u8 kint_m: 8; > > + } bits; > > + u8 val; > > + } _0a; > > + union __reg_0b__ { > > + struct __0b { > > + u8 out_sel: 4; > > + u8 kint_l: 4; > > + } bits; > > + u8 val; > > + } _0b; > > + union __reg_0c__ { > > + struct __0c { > > + u8 kstep_h: 8; > > + } bits; > > + u8 val; > > + } _0c; > > + union __reg_0d__ { > > + struct __0d { > > + u8 kstep_m: 8; > > + } bits; > > + u8 val; > > + } _0d; > > + union __reg_0e__ { > > + struct __0e { > > + u8 pll_pu_byp: 1; > > + u8 pll_pu: 1; > > + u8 hsbist_len: 2; > > + u8 stopstate_sel: 1; > > + u8 kstep_l: 3; > > + } bits; > > + u8 val; > > + } _0e; > > + union __reg_0f__ { > > + struct __0f { > > + u8 det_delay:2; > > + u8 kdelta: 4; > > + u8 ldo0p4:2; > > + } bits; > > + u8 val; > > + } _0f; > > +}; > > + > > +struct dphy_pll { > > + u8 refin; /* Pre-divider control signal */ > > + u8 cp_s; /* 00: SDM_EN=1, 10: SDM_EN=0 */ > > + u8 fdk_s; /* PLL mode control: integer or fraction */ > > + u8 sdm_en; > > + u8 div; > > + u8 int_n; /* integer N PLL */ > > + u32 ref_clk; /* dphy reference clock, unit: MHz */ > > + u32 freq; /* panel config, unit: KHz */ > > + u32 fvco; > > + u32 potential_fvco; > > + u32 nint; /* sigma delta modulator NINT control */ > > + u32 kint; /* sigma delta modulator KINT control */ > > + u8 lpf_sel; /* low pass filter control */ > > + u8 out_sel; /* post divider control */ > > + u8 vco_band; /* vco range */ > > + u8 det_delay; > > + > > + struct pll_reg reg; > > +}; > > + > > +struct dsi_context; > > + > > +int dphy_pll_config(struct dsi_context *ctx); > > +void dphy_timing_config(struct dsi_context *ctx); > > + > > +#endif /* _MEGACORES_PLL_H_ */ > > diff --git a/drivers/gpu/drm/sprd/sprd_dpu.c b/drivers/gpu/drm/sprd/sprd_dpu.c > > index e74c3dbb3..10f6038e6 100644 > > --- a/drivers/gpu/drm/sprd/sprd_dpu.c > > +++ b/drivers/gpu/drm/sprd/sprd_dpu.c > > @@ -25,6 +25,7 @@ > > > > #include "sprd_drm.h" > > #include "sprd_dpu.h" > > +#include "sprd_dsi.h" > > > > /* Global control registers */ > > #define REG_DPU_CTRL 0x04 > > @@ -687,9 +688,25 @@ static void sprd_crtc_mode_set_nofb(struct drm_crtc *crtc) > > { > > struct sprd_dpu *dpu = to_sprd_crtc(crtc); > > struct drm_display_mode *mode = &crtc->state->adjusted_mode; > > + struct drm_encoder *encoder; > > + struct mipi_dsi_device *slave; > > + struct sprd_dsi *dsi; > > > > if (mode->type & DRM_MODE_TYPE_PREFERRED) > > drm_display_mode_to_videomode(mode, &dpu->ctx.vm); > > + > > + drm_for_each_encoder(encoder, crtc->dev) { > > + if (encoder->crtc != crtc) > > + continue; > > + > > + dsi = encoder_to_dsi(encoder); > > + slave = dsi->slave; > > + > > + if (slave->mode_flags & MIPI_DSI_MODE_VIDEO) > > + dpu->ctx.if_type = SPRD_DPU_IF_DPI; > > + else > > + dpu->ctx.if_type = SPRD_DPU_IF_EDPI; > > + } > > This would make more sense to store in a custom crtc_state and to fill > at atomic_check time. That way you can properly fail if you don't find > the encoder, or if it's not a DSI encoder, etc. Ok, i will remove it to atomic_check ops. > > > } > > > > static void sprd_crtc_atomic_enable(struct drm_crtc *crtc, > > diff --git a/drivers/gpu/drm/sprd/sprd_drm.c b/drivers/gpu/drm/sprd/sprd_drm.c > > index 549b71278..68cf0933c 100644 > > --- a/drivers/gpu/drm/sprd/sprd_drm.c > > +++ b/drivers/gpu/drm/sprd/sprd_drm.c > > @@ -183,6 +183,7 @@ static struct platform_driver sprd_drm_driver = { > > static struct platform_driver *sprd_drm_drivers[] = { > > &sprd_drm_driver, > > &sprd_dpu_driver, > > + &sprd_dsi_driver, > > }; > > > > static int __init sprd_drm_init(void) > > diff --git a/drivers/gpu/drm/sprd/sprd_drm.h b/drivers/gpu/drm/sprd/sprd_drm.h > > index 85d4a8b9f..95d1b972f 100644 > > --- a/drivers/gpu/drm/sprd/sprd_drm.h > > +++ b/drivers/gpu/drm/sprd/sprd_drm.h > > @@ -14,5 +14,6 @@ struct sprd_drm { > > }; > > > > extern struct platform_driver sprd_dpu_driver; > > +extern struct platform_driver sprd_dsi_driver; > > > > #endif /* _SPRD_DRM_H_ */ > > diff --git a/drivers/gpu/drm/sprd/sprd_dsi.c b/drivers/gpu/drm/sprd/sprd_dsi.c > > new file mode 100644 > > index 000000000..f26a37e50 > > --- /dev/null > > +++ b/drivers/gpu/drm/sprd/sprd_dsi.c > > @@ -0,0 +1,1124 @@ > > +// SPDX-License-Identifier: GPL-2.0 > > +/* > > + * Copyright (C) 2020 Unisoc Inc. > > + */ > > + > > +#include <linux/component.h> > > +#include <linux/module.h> > > +#include <linux/of_address.h> > > +#include <linux/of_device.h> > > +#include <linux/of_irq.h> > > +#include <linux/of_graph.h> > > +#include <video/mipi_display.h> > > + > > +#include <drm/drm_atomic_helper.h> > > +#include <drm/drm_crtc_helper.h> > > +#include <drm/drm_of.h> > > +#include <drm/drm_probe_helper.h> > > + > > +#include "sprd_drm.h" > > +#include "sprd_dpu.h" > > +#include "sprd_dsi.h" > > +#include "dsi_ctrl.h" > > +#include "dsi_ctrl_ppi.h" > > + > > +#define host_to_dsi(host) \ > > + container_of(host, struct sprd_dsi, host) > > +#define connector_to_dsi(connector) \ > > + container_of(connector, struct sprd_dsi, connector) > > + > > +static int regmap_tst_io_write(void *context, u32 reg, u32 val) > > +{ > > + struct sprd_dsi *dsi = context; > > + struct dsi_context *ctx = &dsi->ctx; > > + > > + if (val > 0xff || reg > 0xff) > > + return -EINVAL; > > + > > + drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, val); > > + > > + dsi_phy_test_en(ctx, 1); > > + dsi_phy_test_din(ctx, reg); > > + dsi_phy_test_clk(ctx, 1); > > + dsi_phy_test_clk(ctx, 0); > > + dsi_phy_test_en(ctx, 0); > > + dsi_phy_test_din(ctx, val); > > + dsi_phy_test_clk(ctx, 1); > > + dsi_phy_test_clk(ctx, 0); > > + > > + return 0; > > +} > > + > > +static int regmap_tst_io_read(void *context, u32 reg, u32 *val) > > +{ > > + struct sprd_dsi *dsi = context; > > + struct dsi_context *ctx = &dsi->ctx; > > + int ret; > > + > > + if (reg > 0xff) > > + return -EINVAL; > > + > > + dsi_phy_test_en(ctx, 1); > > + dsi_phy_test_din(ctx, reg); > > + dsi_phy_test_clk(ctx, 1); > > + dsi_phy_test_clk(ctx, 0); > > + dsi_phy_test_en(ctx, 0); > > + > > + udelay(1); > > + > > + ret = dsi_phy_test_dout(ctx); > > + if (ret < 0) > > + return ret; > > + > > + *val = ret; > > + > > + drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, *val); > > + return 0; > > +} > > + > > +static struct regmap_bus regmap_tst_io = { > > + .reg_write = regmap_tst_io_write, > > + .reg_read = regmap_tst_io_read, > > +}; > > + > > +static const struct regmap_config byte_config = { > > + .reg_bits = 8, > > + .val_bits = 8, > > +}; > > + > > +static int dphy_wait_pll_locked(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + int i; > > + > > + for (i = 0; i < 50000; i++) { > > + if (dsi_phy_is_pll_locked(ctx)) > > + return 0; > > + udelay(3); > > + } > > + > > + drm_err(dsi->drm, "dphy pll can not be locked\n"); > > + return -ETIMEDOUT; > > +} > > + > > +static int dsi_wait_tx_payload_fifo_empty(struct dsi_context *ctx) > > +{ > > + int i; > > + > > + for (i = 0; i < 5000; i++) { > > + if (dsi_is_tx_payload_fifo_empty(ctx)) > > + return 0; > > + udelay(1); > > + } > > + > > + return -ETIMEDOUT; > > +} > > + > > +static int dsi_wait_tx_cmd_fifo_empty(struct dsi_context *ctx) > > +{ > > + int i; > > + > > + for (i = 0; i < 5000; i++) { > > + if (dsi_is_tx_cmd_fifo_empty(ctx)) > > + return 0; > > + udelay(1); > > + } > > + > > + return -ETIMEDOUT; > > +} > > + > > +static int dsi_wait_rd_resp_completed(struct dsi_context *ctx) > > +{ > > + int i; > > + > > + for (i = 0; i < 10000; i++) { > > + if (dsi_is_bta_returned(ctx)) > > + return 0; > > + udelay(10); > > + } > > + > > + return -ETIMEDOUT; > > +} > > + > > +static u16 calc_bytes_per_pixel_x100(int coding) > > +{ > > + u16 Bpp_x100; > > + > > + switch (coding) { > > + case COLOR_CODE_16BIT_CONFIG1: > > + case COLOR_CODE_16BIT_CONFIG2: > > + case COLOR_CODE_16BIT_CONFIG3: > > + Bpp_x100 = 200; > > + break; > > + case COLOR_CODE_18BIT_CONFIG1: > > + case COLOR_CODE_18BIT_CONFIG2: > > + Bpp_x100 = 225; > > + break; > > + case COLOR_CODE_24BIT: > > + Bpp_x100 = 300; > > + break; > > + case COLOR_CODE_COMPRESSTION: > > + Bpp_x100 = 100; > > + break; > > + case COLOR_CODE_20BIT_YCC422_LOOSELY: > > + Bpp_x100 = 250; > > + break; > > + case COLOR_CODE_24BIT_YCC422: > > + Bpp_x100 = 300; > > + break; > > + case COLOR_CODE_16BIT_YCC422: > > + Bpp_x100 = 200; > > + break; > > + case COLOR_CODE_30BIT: > > + Bpp_x100 = 375; > > + break; > > + case COLOR_CODE_36BIT: > > + Bpp_x100 = 450; > > + break; > > + case COLOR_CODE_12BIT_YCC420: > > + Bpp_x100 = 150; > > + break; > > + default: > > + DRM_ERROR("invalid color coding"); > > + Bpp_x100 = 0; > > + break; > > + } > > + > > + return Bpp_x100; > > +} > > + > > +static u8 calc_video_size_step(int coding) > > +{ > > + u8 video_size_step; > > + > > + switch (coding) { > > + case COLOR_CODE_16BIT_CONFIG1: > > + case COLOR_CODE_16BIT_CONFIG2: > > + case COLOR_CODE_16BIT_CONFIG3: > > + case COLOR_CODE_18BIT_CONFIG1: > > + case COLOR_CODE_18BIT_CONFIG2: > > + case COLOR_CODE_24BIT: > > + case COLOR_CODE_COMPRESSTION: > > + return video_size_step = 1; > > + case COLOR_CODE_20BIT_YCC422_LOOSELY: > > + case COLOR_CODE_24BIT_YCC422: > > + case COLOR_CODE_16BIT_YCC422: > > + case COLOR_CODE_30BIT: > > + case COLOR_CODE_36BIT: > > + case COLOR_CODE_12BIT_YCC420: > > + return video_size_step = 2; > > + default: > > + DRM_ERROR("invalid color coding"); > > + return 0; > > + } > > +} > > + > > +static u16 round_video_size(int coding, u16 video_size) > > +{ > > + switch (coding) { > > + case COLOR_CODE_16BIT_YCC422: > > + case COLOR_CODE_24BIT_YCC422: > > + case COLOR_CODE_20BIT_YCC422_LOOSELY: > > + case COLOR_CODE_12BIT_YCC420: > > + /* round up active H pixels to a multiple of 2 */ > > + if ((video_size % 2) != 0) > > + video_size += 1; > > + break; > > + default: > > + break; > > + } > > + > > + return video_size; > > +} > > + > > +#define SPRD_MIPI_DSI_FMT_DSC 0xff > > +static u32 fmt_to_coding(u32 fmt) > > +{ > > + switch (fmt) { > > + case MIPI_DSI_FMT_RGB565: > > + return COLOR_CODE_16BIT_CONFIG1; > > + case MIPI_DSI_FMT_RGB666: > > + case MIPI_DSI_FMT_RGB666_PACKED: > > + return COLOR_CODE_18BIT_CONFIG1; > > + case MIPI_DSI_FMT_RGB888: > > + return COLOR_CODE_24BIT; > > + case SPRD_MIPI_DSI_FMT_DSC: > > + return COLOR_CODE_COMPRESSTION; > > + default: > > + DRM_ERROR("Unsupported format (%d)\n", fmt); > > + return COLOR_CODE_24BIT; > > + } > > +} > > + > > +#define ns_to_cycle(ns, byte_clk) \ > > + DIV_ROUND_UP((ns) * (byte_clk), 1000000) > > + > > +static void sprd_dsi_init(struct dsi_context *ctx) > > +{ > > + u16 data_hs2lp, data_lp2hs, clk_hs2lp, clk_lp2hs; > > + u16 max_rd_time; > > + int div; > > + > > + dsi_power_enable(ctx, 0); > > + dsi_int0_mask(ctx, 0xffffffff); > > + dsi_int1_mask(ctx, 0xffffffff); > > + dsi_cmd_mode(ctx); > > + dsi_eotp_rx_en(ctx, 0); > > + dsi_eotp_tx_en(ctx, 0); > > + dsi_ecc_rx_en(ctx, 1); > > + dsi_crc_rx_en(ctx, 1); > > + dsi_bta_en(ctx, 1); > > + dsi_video_vcid(ctx, 0); > > + dsi_rx_vcid(ctx, 0); > > + > > + div = DIV_ROUND_UP(ctx->byte_clk, ctx->esc_clk); > > + dsi_tx_escape_division(ctx, div); > > + > > + max_rd_time = ns_to_cycle(ctx->max_rd_time, ctx->byte_clk); > > + dsi_max_read_time(ctx, max_rd_time); > > + > > + data_hs2lp = ns_to_cycle(ctx->data_hs2lp, ctx->byte_clk); > > + data_lp2hs = ns_to_cycle(ctx->data_lp2hs, ctx->byte_clk); > > + clk_hs2lp = ns_to_cycle(ctx->clk_hs2lp, ctx->byte_clk); > > + clk_lp2hs = ns_to_cycle(ctx->clk_lp2hs, ctx->byte_clk); > > + dsi_datalane_hs2lp_config(ctx, data_hs2lp); > > + dsi_datalane_lp2hs_config(ctx, data_lp2hs); > > + dsi_clklane_hs2lp_config(ctx, clk_hs2lp); > > + dsi_clklane_lp2hs_config(ctx, clk_lp2hs); > > + > > + dsi_power_enable(ctx, 1); > > +} > > + > > +/* > > + * Free up resources and shutdown host controller and PHY > > + */ > > +static void sprd_dsi_fini(struct dsi_context *ctx) > > +{ > > + dsi_int0_mask(ctx, 0xffffffff); > > + dsi_int1_mask(ctx, 0xffffffff); > > + dsi_power_enable(ctx, 0); > > +} > > + > > +/* > > + * If not in burst mode, it will compute the video and null packet sizes > > + * according to necessity. > > + * Configure timers for data lanes and/or clock lane to return to LP when > > + * bandwidth is not filled by data. > > + */ > > +static int sprd_dsi_dpi_video(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + struct videomode *vm = &ctx->vm; > > + u16 Bpp_x100; > > + u16 video_size; > > + u32 ratio_x1000; > > + u16 null_pkt_size = 0; > > + u8 video_size_step; > > + u32 hs_to; > > + u32 total_bytes; > > + u32 bytes_per_chunk; > > + u32 chunks = 0; > > + u32 bytes_left = 0; > > + u32 chunk_overhead; > > + const u8 pkt_header = 6; > > + u8 coding; > > + int div; > > + u16 hline; > > + > > + coding = fmt_to_coding(ctx->format); > > + video_size = round_video_size(coding, vm->hactive); > > + Bpp_x100 = calc_bytes_per_pixel_x100(coding); > > + video_size_step = calc_video_size_step(coding); > > + ratio_x1000 = ctx->byte_clk * 1000 / (vm->pixelclock / 1000); > > + hline = vm->hactive + vm->hsync_len + vm->hfront_porch + > > + vm->hback_porch; > > + > > + dsi_power_enable(ctx, 0); > > + dsi_dpi_frame_ack_en(ctx, ctx->frame_ack_en); > > + dsi_dpi_color_coding(ctx, coding); > > + dsi_dpi_video_burst_mode(ctx, ctx->burst_mode); > > + dsi_dpi_sig_delay(ctx, 95 * hline * ratio_x1000 / 100000); > > + dsi_dpi_hline_time(ctx, hline * ratio_x1000 / 1000); > > + dsi_dpi_hsync_time(ctx, vm->hsync_len * ratio_x1000 / 1000); > > + dsi_dpi_hbp_time(ctx, vm->hback_porch * ratio_x1000 / 1000); > > + dsi_dpi_vact(ctx, vm->vactive); > > + dsi_dpi_vfp(ctx, vm->vfront_porch); > > + dsi_dpi_vbp(ctx, vm->vback_porch); > > + dsi_dpi_vsync(ctx, vm->vsync_len); > > + dsi_dpi_hporch_lp_en(ctx, 1); > > + dsi_dpi_vporch_lp_en(ctx, 1); > > + > > + hs_to = (hline * vm->vactive) + (2 * Bpp_x100) / 100; > > + for (div = 0x80; (div < hs_to) && (div > 2); div--) { > > + if ((hs_to % div) == 0) { > > + dsi_timeout_clock_division(ctx, div); > > + dsi_lp_rx_timeout(ctx, hs_to / div); > > + dsi_hs_tx_timeout(ctx, hs_to / div); > > + break; > > + } > > + } > > + > > + if (ctx->burst_mode == VIDEO_BURST_WITH_SYNC_PULSES) { > > + dsi_dpi_video_packet_size(ctx, video_size); > > + dsi_dpi_null_packet_size(ctx, 0); > > + dsi_dpi_chunk_num(ctx, 0); > > + } else { > > + /* non burst transmission */ > > + null_pkt_size = 0; > > + > > + /* bytes to be sent - first as one chunk */ > > + bytes_per_chunk = vm->hactive * Bpp_x100 / 100 + pkt_header; > > + > > + /* hline total bytes from the DPI interface */ > > + total_bytes = (vm->hactive + vm->hfront_porch) * > > + ratio_x1000 / ctx->lanes / 1000; > > + > > + /* check if the pixels actually fit on the DSI link */ > > + if (total_bytes < bytes_per_chunk) { > > + drm_err(dsi->drm, "current resolution can not be set\n"); > > + return -EINVAL; > > + } > > + > > + chunk_overhead = total_bytes - bytes_per_chunk; > > + > > + /* overhead higher than 1 -> enable multi packets */ > > + if (chunk_overhead > 1) { > > + > > + /* multi packets */ > > + for (video_size = video_size_step; > > + video_size < vm->hactive; > > + video_size += video_size_step) { > > + > > + if (vm->hactive * 1000 / video_size % 1000) > > + continue; > > + > > + chunks = vm->hactive / video_size; > > + bytes_per_chunk = Bpp_x100 * video_size / 100 > > + + pkt_header; > > + if (total_bytes >= (bytes_per_chunk * chunks)) { > > + bytes_left = total_bytes - > > + bytes_per_chunk * chunks; > > + break; > > + } > > + } > > + > > + /* prevent overflow (unsigned - unsigned) */ > > + if (bytes_left > (pkt_header * chunks)) { > > + null_pkt_size = (bytes_left - > > + pkt_header * chunks) / chunks; > > + /* avoid register overflow */ > > + if (null_pkt_size > 1023) > > + null_pkt_size = 1023; > > + } > > + > > + } else { > > + > > + /* single packet */ > > + chunks = 1; > > + > > + /* must be a multiple of 4 except 18 loosely */ > > + for (video_size = vm->hactive; > > + (video_size % video_size_step) != 0; > > + video_size++) > > + ; > > + } > > + > > + dsi_dpi_video_packet_size(ctx, video_size); > > + dsi_dpi_null_packet_size(ctx, null_pkt_size); > > + dsi_dpi_chunk_num(ctx, chunks); > > + } > > + > > + dsi_int0_mask(ctx, ctx->int0_mask); > > + dsi_int1_mask(ctx, ctx->int1_mask); > > + dsi_power_enable(ctx, 1); > > + > > + return 0; > > +} > > + > > +static void sprd_dsi_edpi_video(struct dsi_context *ctx) > > +{ > > + const u32 fifo_depth = 1096; > > + const u32 word_length = 4; > > + u32 hactive = ctx->vm.hactive; > > + u32 Bpp_x100; > > + u32 max_fifo_len; > > + u8 coding; > > + > > + coding = fmt_to_coding(ctx->format); > > + Bpp_x100 = calc_bytes_per_pixel_x100(coding); > > + max_fifo_len = word_length * fifo_depth * 100 / Bpp_x100; > > + > > + dsi_power_enable(ctx, 0); > > + dsi_dpi_color_coding(ctx, coding); > > + dsi_tear_effect_ack_en(ctx, ctx->te_ack_en); > > + > > + if (max_fifo_len > hactive) > > + dsi_edpi_max_pkt_size(ctx, hactive); > > + else > > + dsi_edpi_max_pkt_size(ctx, max_fifo_len); > > + > > + dsi_int0_mask(ctx, ctx->int0_mask); > > + dsi_int1_mask(ctx, ctx->int1_mask); > > + dsi_power_enable(ctx, 1); > > +} > > + > > +/* > > + * Send a packet on the generic interface, > > + * this function has an active delay to wait for the buffer to clear. > > + * The delay is limited to: > > + * (param_length / 4) x DSIH_FIFO_ACTIVE_WAIT x register access time > > + * the controller restricts the sending of. > > + * > > + * This function will not be able to send Null and Blanking packets due to > > + * controller restriction > > + */ > > +static int sprd_dsi_wr_pkt(struct dsi_context *ctx, u8 vc, u8 type, > > + const u8 *param, u16 len) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + u8 wc_lsbyte, wc_msbyte; > > + u32 payload; > > + int i, j, ret; > > + > > + if (vc > 3) > > + return -EINVAL; > > + > > + > > + /* 1st: for long packet, must config payload first */ > > + ret = dsi_wait_tx_payload_fifo_empty(ctx); > > + if (ret) { > > + drm_err(dsi->drm, "tx payload fifo is not empty\n"); > > + return ret; > > + } > > + > > + if (len > 2) { > > + for (i = 0, j = 0; i < len; i += j) { > > + payload = 0; > > + for (j = 0; (j < 4) && ((j + i) < (len)); j++) > > + payload |= param[i + j] << (j * 8); > > + > > + dsi_set_packet_payload(ctx, payload); > > + } > > + wc_lsbyte = len & 0xff; > > + wc_msbyte = len >> 8; > > + } else { > > + wc_lsbyte = (len > 0) ? param[0] : 0; > > + wc_msbyte = (len > 1) ? param[1] : 0; > > + } > > + > > + /* 2nd: then set packet header */ > > + ret = dsi_wait_tx_cmd_fifo_empty(ctx); > > + if (ret) { > > + drm_err(dsi->drm, "tx cmd fifo is not empty\n"); > > + return ret; > > + } > > + > > + dsi_set_packet_header(ctx, vc, type, wc_lsbyte, wc_msbyte); > > + > > + return 0; > > +} > > + > > +/* > > + * Send READ packet to peripheral using the generic interface, > > + * this will force command mode and stop video mode (because of BTA). > > + * > > + * This function has an active delay to wait for the buffer to clear, > > + * the delay is limited to 2 x DSIH_FIFO_ACTIVE_WAIT > > + * (waiting for command buffer, and waiting for receiving) > > + * @note this function will enable BTA > > + */ > > +static int sprd_dsi_rd_pkt(struct dsi_context *ctx, u8 vc, u8 type, > > + u8 msb_byte, u8 lsb_byte, > > + u8 *buffer, u8 bytes_to_read) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + int i, ret; > > + int count = 0; > > + u32 temp; > > + > > + if (vc > 3) > > + return -EINVAL; > > + > > + /* 1st: send read command to peripheral */ > > + if (!dsi_is_tx_cmd_fifo_empty(ctx)) > > + return -EIO; > > + > > + dsi_set_packet_header(ctx, vc, type, lsb_byte, msb_byte); > > + > > + /* 2nd: wait peripheral response completed */ > > + ret = dsi_wait_rd_resp_completed(ctx); > > + if (ret) { > > + drm_err(dsi->drm, "wait read response time out\n"); > > + return ret; > > + } > > + > > + /* 3rd: get data from rx payload fifo */ > > + if (dsi_is_rx_payload_fifo_empty(ctx)) { > > + drm_err(dsi->drm, "rx payload fifo empty\n"); > > + return -EIO; > > + } > > + > > + for (i = 0; i < 100; i++) { > > + temp = dsi_get_rx_payload(ctx); > > + > > + if (count < bytes_to_read) > > + buffer[count++] = temp & 0xff; > > + if (count < bytes_to_read) > > + buffer[count++] = (temp >> 8) & 0xff; > > + if (count < bytes_to_read) > > + buffer[count++] = (temp >> 16) & 0xff; > > + if (count < bytes_to_read) > > + buffer[count++] = (temp >> 24) & 0xff; > > + > > + if (dsi_is_rx_payload_fifo_empty(ctx)) > > + return count; > > + } > > + > > + return 0; > > +} > > + > > +static void sprd_dsi_set_work_mode(struct dsi_context *ctx, u8 mode) > > +{ > > + if (mode == DSI_MODE_CMD) > > + dsi_cmd_mode(ctx); > > + else > > + dsi_video_mode(ctx); > > +} > > + > > +static void sprd_dsi_lp_cmd_enable(struct dsi_context *ctx, bool enable) > > +{ > > + if (dsi_is_cmd_mode(ctx)) > > + dsi_cmd_mode_lp_cmd_en(ctx, enable); > > + else > > + dsi_video_mode_lp_cmd_en(ctx, enable); > > +} > > + > > +static void sprd_dsi_state_reset(struct dsi_context *ctx) > > +{ > > + dsi_power_enable(ctx, 0); > > + udelay(100); > > + dsi_power_enable(ctx, 1); > > +} > > + > > +static u32 sprd_dsi_int_status(struct dsi_context *ctx, int index) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + u32 status; > > + > > + if (index == 0) > > + status = dsi_int0_status(ctx); > > + else if (index == 1) > > + status = dsi_int1_status(ctx); > > + else { > > + drm_err(dsi->drm, "invalid dsi IRQ index %d\n", index); > > + status = -EINVAL; > > + } > > + > > + return status; > > +} > > + > > +static int sprd_dphy_init(struct dsi_context *ctx) > > +{ > > + struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); > > + int ret; > > + > > + dsi_phy_rstz(ctx, 0); > > + dsi_phy_shutdownz(ctx, 0); > > + dsi_phy_clklane_en(ctx, 0); > > + > > + dsi_phy_test_clr(ctx, 0); > > + dsi_phy_test_clr(ctx, 1); > > + dsi_phy_test_clr(ctx, 0); > > + > > + dphy_pll_config(ctx); > > + dphy_timing_config(ctx); > > + > > + dsi_phy_shutdownz(ctx, 1); > > + dsi_phy_rstz(ctx, 1); > > + dsi_phy_stop_wait_time(ctx, 0x1C); > > + dsi_phy_clklane_en(ctx, 1); > > + dsi_phy_datalane_en(ctx); > > + > > + ret = dphy_wait_pll_locked(ctx); > > + if (ret) { > > + drm_err(dsi->drm, "dphy initial failed\n"); > > + return ret; > > + } > > + > > + return 0; > > +} > > + > > +static void sprd_dphy_fini(struct dsi_context *ctx) > > +{ > > + dsi_phy_rstz(ctx, 0); > > + dsi_phy_shutdownz(ctx, 0); > > + dsi_phy_rstz(ctx, 1); > > +} > > + > > +static void sprd_dsi_encoder_enable(struct drm_encoder *encoder) > > +{ > > + struct sprd_dsi *dsi = encoder_to_dsi(encoder); > > + struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc); > > + struct dsi_context *ctx = &dsi->ctx; > > + > > + if (ctx->enabled) { > > + drm_warn(dsi->drm, "dsi is initialized\n"); > > + return; > > + } > > + > > + sprd_dsi_init(ctx); > > + if (ctx->work_mode == DSI_MODE_VIDEO) > > + sprd_dsi_dpi_video(ctx); > > + else > > + sprd_dsi_edpi_video(ctx); > > + > > + sprd_dphy_init(ctx); > > + > > + sprd_dsi_lp_cmd_enable(ctx, true); > > + > > + if (dsi->panel) { > > + drm_panel_prepare(dsi->panel); > > + drm_panel_enable(dsi->panel); > > + } > > + > > + sprd_dsi_set_work_mode(ctx, ctx->work_mode); > > + sprd_dsi_state_reset(ctx); > > + > > + if (ctx->nc_clk_en) > > + dsi_nc_clk_en(ctx, true); > > + else { > > + dsi_phy_clk_hs_rqst(ctx, true); > > + dphy_wait_pll_locked(ctx); > > + } > > + > > + sprd_dpu_run(dpu); > > + > > + ctx->enabled = true; > > +} > > + > > +static void sprd_dsi_encoder_disable(struct drm_encoder *encoder) > > +{ > > + struct sprd_dsi *dsi = encoder_to_dsi(encoder); > > + struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc); > > + struct dsi_context *ctx = &dsi->ctx; > > + > > + if (!ctx->enabled) { > > + drm_warn(dsi->drm, "dsi isn't initialized\n"); > > + return; > > + } > > + > > + sprd_dpu_stop(dpu); > > + sprd_dsi_set_work_mode(ctx, DSI_MODE_CMD); > > + sprd_dsi_lp_cmd_enable(ctx, true); > > + > > + if (dsi->panel) { > > + drm_panel_disable(dsi->panel); > > + drm_panel_unprepare(dsi->panel); > > + } > > + > > + sprd_dphy_fini(ctx); > > + sprd_dsi_fini(ctx); > > + > > + ctx->enabled = false; > > +} > > + > > +static void sprd_dsi_encoder_mode_set(struct drm_encoder *encoder, > > + struct drm_display_mode *mode, > > + struct drm_display_mode *adj_mode) > > +{ > > + struct sprd_dsi *dsi = encoder_to_dsi(encoder); > > + > > + drm_dbg(dsi->drm, "%s() set mode: %s\n", __func__, dsi->mode->name); > > +} > > You can remove that function Ok, i will remove it. > > > +static int sprd_dsi_encoder_atomic_check(struct drm_encoder *encoder, > > + struct drm_crtc_state *crtc_state, > > + struct drm_connector_state *conn_state) > > +{ > > + return 0; > > +} > > And this one too Ok, i will remove it. > > > +static const struct drm_encoder_helper_funcs sprd_encoder_helper_funcs = { > > + .atomic_check = sprd_dsi_encoder_atomic_check, > > + .mode_set = sprd_dsi_encoder_mode_set, > > + .enable = sprd_dsi_encoder_enable, > > + .disable = sprd_dsi_encoder_disable > > +}; > > + > > +static const struct drm_encoder_funcs sprd_encoder_funcs = { > > + .destroy = drm_encoder_cleanup, > > +}; > > + > > +static struct sprd_dsi *sprd_dsi_encoder_init(struct drm_device *drm, > > + struct device *dev) > > +{ > > + struct sprd_dsi *dsi; > > + u32 crtc_mask; > > + > > + crtc_mask = drm_of_find_possible_crtcs(drm, dev->of_node); > > + if (!crtc_mask) { > > + drm_err(drm, "failed to find crtc mask\n"); > > + return ERR_PTR(-EINVAL); > > + } > > + > > + drm_dbg(drm, "find possible crtcs: 0x%08x\n", crtc_mask); > > + > > + dsi = drmm_encoder_alloc(drm, struct sprd_dsi, encoder, > > + &sprd_encoder_funcs, DRM_MODE_ENCODER_DSI, NULL); > > + if (IS_ERR(dsi)) { > > + drm_err(drm, "failed to init dsi encoder.\n"); > > + return dsi; > > + } > > + > > + dsi->encoder.possible_crtcs = crtc_mask; > > + drm_encoder_helper_add(&dsi->encoder, &sprd_encoder_helper_funcs); > > + > > + return dsi; > > +} > > + > > +static int sprd_dsi_find_panel(struct sprd_dsi *dsi) > > +{ > > + struct device *dev = dsi->host.dev; > > + struct device_node *child, *lcds_node; > > + struct drm_panel *panel; > > + > > + /* search /lcds child node first */ > > + lcds_node = of_find_node_by_path("/lcds"); > > + for_each_child_of_node(lcds_node, child) { > > + panel = of_drm_find_panel(child); > > + if (!IS_ERR(panel)) { > > + dsi->panel = panel; > > + return 0; > > + } > > + } > > That's not part of your binding Ok, i will remove it. > > > + > > + /* > > + * If /lcds child node search failed, we search > > + * the child of dsi host node. > > + */ > > + for_each_child_of_node(dev->of_node, child) { > > + panel = of_drm_find_panel(child); > > + if (!IS_ERR(panel)) { > > + dsi->panel = panel; > > + return 0; > > + } > > + } > > And you don't need this either. You'll register a mipi_dsi_host, that > will in turn probe all the devices under that bus, and will then call > the .attach callback. This should be move to the .attach callback? > > > + drm_err(dsi->drm, "of_drm_find_panel() failed\n"); > > + return -ENODEV; > > +} > > + > > +static int sprd_dsi_host_attach(struct mipi_dsi_host *host, > > + struct mipi_dsi_device *slave) > > +{ > > + struct sprd_dsi *dsi = host_to_dsi(host); > > + struct dsi_context *ctx = &dsi->ctx; > > + int ret; > > + > > + dsi->slave = slave; > > + ctx->lanes = slave->lanes; > > + ctx->format = slave->format; > > + ctx->byte_clk = slave->hs_rate / 8; > > + ctx->esc_clk = slave->lp_rate; > > + > > + if (slave->mode_flags & MIPI_DSI_MODE_VIDEO) > > + ctx->work_mode = DSI_MODE_VIDEO; > > + else > > + ctx->work_mode = DSI_MODE_CMD; > > + > > + if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_BURST) > > + ctx->burst_mode = VIDEO_BURST_WITH_SYNC_PULSES; > > + else if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_SYNC_PULSE) > > + ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_PULSES; > > + else > > + ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_EVENTS; > > + > > + if (slave->mode_flags & MIPI_DSI_CLOCK_NON_CONTINUOUS) > > + ctx->nc_clk_en = true; > > I'm not sure why you need to duplicate all this, can't you just > dereference the dsi->slave pointer when you need it? Sorry, can you help me with a demo? > > > + ret = sprd_dsi_find_panel(dsi); > > + if (ret) > > + return ret; > > The panel is your slave here > > > + return 0; > > +} > > + > > +static int sprd_dsi_host_detach(struct mipi_dsi_host *host, > > + struct mipi_dsi_device *slave) > > +{ > > + /* do nothing */ > > + return 0; > > +} > > + > > +static ssize_t sprd_dsi_host_transfer(struct mipi_dsi_host *host, > > + const struct mipi_dsi_msg *msg) > > +{ > > + struct sprd_dsi *dsi = host_to_dsi(host); > > + const u8 *tx_buf = msg->tx_buf; > > + > > + if (msg->rx_buf && msg->rx_len) { > > + u8 lsb = (msg->tx_len > 0) ? tx_buf[0] : 0; > > + u8 msb = (msg->tx_len > 1) ? tx_buf[1] : 0; > > + > > + return sprd_dsi_rd_pkt(&dsi->ctx, msg->channel, msg->type, > > + msb, lsb, msg->rx_buf, msg->rx_len); > > + } > > + > > + if (msg->tx_buf && msg->tx_len) > > + return sprd_dsi_wr_pkt(&dsi->ctx, msg->channel, msg->type, > > + tx_buf, msg->tx_len); > > + > > + return 0; > > +} > > + > > +static const struct mipi_dsi_host_ops sprd_dsi_host_ops = { > > + .attach = sprd_dsi_host_attach, > > + .detach = sprd_dsi_host_detach, > > + .transfer = sprd_dsi_host_transfer, > > +}; > > + > > +static int sprd_dsi_host_init(struct sprd_dsi *dsi, struct device *dev) > > +{ > > + int ret; > > + > > + dsi->host.dev = dev; > > + dsi->host.ops = &sprd_dsi_host_ops; > > + > > + ret = mipi_dsi_host_register(&dsi->host); > > + if (ret) > > + drm_err(dsi->drm, "failed to register dsi host\n"); > > + > > + return ret; > > +} > > + > > +static int sprd_dsi_connector_get_modes(struct drm_connector *connector) > > +{ > > + struct sprd_dsi *dsi = connector_to_dsi(connector); > > + > > + return drm_panel_get_modes(dsi->panel, connector); > > +} > > + > > +static enum drm_mode_status > > +sprd_dsi_connector_mode_valid(struct drm_connector *connector, > > + struct drm_display_mode *mode) > > +{ > > + struct sprd_dsi *dsi = connector_to_dsi(connector); > > + > > + drm_dbg(dsi->drm, "%s() mode: "DRM_MODE_FMT"\n", __func__, DRM_MODE_ARG(mode)); > > + > > + if (mode->type & DRM_MODE_TYPE_PREFERRED) { > > + dsi->mode = mode; > > + drm_display_mode_to_videomode(dsi->mode, &dsi->ctx.vm); > > + } > > Again, what happens if the mode isn't the preferred one? We hope to restore the low-resolution image to the original resolution through the scaling algorithm while keeping the resolution unchanged. So whether it's dpu or dsi, must be keeping on preferred mode. > > You shouldn't be testing this but rather just use the crtc_state > adjusted_mode. > > Maxime