On 08/07/2020 13:16, Tomasz Figa wrote: > On Wed, Jul 8, 2020 at 9:14 AM Xia Jiang <xia.jiang@xxxxxxxxxxxx> wrote: >> >> On Tue, 2020-07-07 at 15:35 +0200, Tomasz Figa wrote: >>> On Tue, Jul 7, 2020 at 8:47 AM Xia Jiang <xia.jiang@xxxxxxxxxxxx> wrote: >>>> >>>> On Tue, 2020-06-30 at 16:53 +0000, Tomasz Figa wrote: >>>>> Hi Xia, >>>>> >>>>> On Tue, Jun 30, 2020 at 10:56:21AM +0800, Xia Jiang wrote: >>>>>> On Thu, 2020-06-11 at 18:46 +0000, Tomasz Figa wrote: >>>>>>> Hi Xia, >>>>>>> >>>>>>> On Thu, Jun 04, 2020 at 05:05:53PM +0800, Xia Jiang wrote: >>>>> [snip] >>>>>>>> +static void mtk_jpeg_enc_device_run(void *priv) >>>>>>>> +{ >>>>>>>> + struct mtk_jpeg_ctx *ctx = priv; >>>>>>>> + struct mtk_jpeg_dev *jpeg = ctx->jpeg; >>>>>>>> + struct vb2_v4l2_buffer *src_buf, *dst_buf; >>>>>>>> + enum vb2_buffer_state buf_state = VB2_BUF_STATE_ERROR; >>>>>>>> + unsigned long flags; >>>>>>>> + struct mtk_jpeg_src_buf *jpeg_src_buf; >>>>>>>> + struct mtk_jpeg_enc_bs enc_bs; >>>>>>>> + int ret; >>>>>>>> + >>>>>>>> + src_buf = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); >>>>>>>> + dst_buf = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); >>>>>>>> + jpeg_src_buf = mtk_jpeg_vb2_to_srcbuf(&src_buf->vb2_buf); >>>>>>>> + >>>>>>>> + ret = pm_runtime_get_sync(jpeg->dev); >>>>>>>> + if (ret < 0) >>>>>>>> + goto enc_end; >>>>>>>> + >>>>>>>> + spin_lock_irqsave(&jpeg->hw_lock, flags); >>>>>>>> + >>>>>>>> + /* >>>>>>>> + * Resetting the hardware every frame is to ensure that all the >>>>>>>> + * registers are cleared. This is a hardware requirement. >>>>>>>> + */ >>>>>>>> + mtk_jpeg_enc_reset(jpeg->reg_base); >>>>>>>> + >>>>>>>> + mtk_jpeg_set_enc_dst(ctx, jpeg->reg_base, &dst_buf->vb2_buf, &enc_bs); >>>>>>>> + mtk_jpeg_set_enc_src(ctx, jpeg->reg_base, &src_buf->vb2_buf); >>>>>>>> + mtk_jpeg_enc_set_config(jpeg->reg_base, ctx->out_q.fmt->hw_format, >>>>>>>> + ctx->enable_exif, ctx->enc_quality, >>>>>>>> + ctx->restart_interval); >>>>>>>> + mtk_jpeg_enc_start(jpeg->reg_base); >>>>>>> >>>>>>> Could we just move the above 5 functions into one function inside >>>>>>> mtk_jpeg_enc_hw.c that takes mtk_jpeg_dev pointer as its argument, let's >>>>>>> say mtk_jpeg_enc_hw_run() and simply program all the data to the registers >>>>>>> directly, without the extra level of abstractions? >>>>>> I can move the 5 functions into one function(mtk_jpeg_enc_hw_run()), but >>>>>> this function will be very long, because it contains computation code >>>>>> such as setting dst addr, blk_num, quality. >>>>>> In v4, you have adviced the following architecture: >>>>>> How about the following model, as used by many other drivers: >>>>>> >>>>>> mtk_jpeg_enc_set_src() >>>>>> { >>>>>> // Set any registers related to source format and buffer >>>>>> } >>>>>> >>>>>> mtk_jpeg_enc_set_dst() >>>>>> { >>>>>> // Set any registers related to destination format and buffer >>>>>> } >>>>>> >>>>>> mtk_jpeg_enc_set_params() >>>>>> { >>>>>> // Set any registers related to additional encoding parameters >>>>>> } >>>>>> >>>>>> mtk_jpeg_enc_device_run(enc, ctx) >>>>>> { >>>>>> mtk_jpeg_enc_set_src(enc, src_buf, src_fmt); >>>>>> mtk_jpeg_enc_set_dst(enc, dst_buf, dst_fmt); >>>>>> mtk_jpeg_enc_set_params(enc, ctx); >>>>>> // Trigger the hardware run >>>>>> } >>>>>> I think that this architecture is more clear(mtk_jpeg_enc_set_config is >>>>>> equivalent to mtk_jpeg_enc_set_params). >>>>>> Should I keep the original architecture or move 5 functions into >>>>>> mtk_jpeg_enc_hw_run? >>>>> >>>>> Sounds good to me. >>>>> >>>>> My biggest issue with the code that it ends up introducing one more >>>>> level of abstraction, but with the approach you suggested, the arguments >>>>> just accept standard structs, which avoids that problem. >>>>> >>>>> [snip] >>>>>>>> + >>>>>>>> + ctx->fh.ctrl_handler = &ctx->ctrl_hdl; >>>>>>>> + ctx->colorspace = V4L2_COLORSPACE_JPEG, >>>>>>>> + ctx->ycbcr_enc = V4L2_YCBCR_ENC_DEFAULT; >>>>>>>> + ctx->quantization = V4L2_QUANTIZATION_DEFAULT; >>>>>>>> + ctx->xfer_func = V4L2_XFER_FUNC_DEFAULT; >>>>>>> >>>>>>> Since we already have a v4l2_pix_format_mplane struct which has fields for >>>>>>> the above 4 values, could we just store them there? >>>>>>> >>>>>>> Also, I don't see this driver handling the colorspaces in any way, but it >>>>>>> seems to allow changing them from the userspace. This is incorrect, because >>>>>>> the userspace has no way to know that the colorspace is not handled. >>>>>>> Instead, the try_fmt implementation should always override the >>>>>>> userspace-provided colorspace configuration with the ones that the driver >>>>>>> assumes. >>>>>>> >>>>>>>> + pix_mp->width = MTK_JPEG_MIN_WIDTH; >>>>>>>> + pix_mp->height = MTK_JPEG_MIN_HEIGHT; >>>>>>>> + >>>>>>>> + q->fmt = mtk_jpeg_find_format(V4L2_PIX_FMT_YUYV, >>>>>>>> + MTK_JPEG_FMT_FLAG_ENC_OUTPUT); >>>>>>>> + vidioc_try_fmt(container_of(pix_mp, struct v4l2_format, >>>>>>>> + fmt.pix_mp), q->fmt); >>>>>>>> + q->w = pix_mp->width; >>>>>>>> + q->h = pix_mp->height; >>>>>>>> + q->crop_rect.width = pix_mp->width; >>>>>>>> + q->crop_rect.height = pix_mp->height; >>>>>>>> + q->sizeimage[0] = pix_mp->plane_fmt[0].sizeimage; >>>>>>>> + q->bytesperline[0] = pix_mp->plane_fmt[0].bytesperline; >>>>>>> >>>>>>> Actually, do we need this custom mtk_jpeg_q_data struct? Why couldn't we >>>>>>> just keep the same values inside the standard v4l2_pix_format_mplane >>>>>>> struct? >>>>>> I think that we need mtk_jpeg_q_data struct.If we delete it, how can we >>>>>> know these values(w, h, sizeimage, bytesperline, mtk_jpeg_fmt) belong to >>>>>> output or capture(output and capture's sizeimages are different, width >>>>>> and height are differnt too for jpeg dec )?We have >>>>>> s_fmt_vid_out_mplane/cap_mplane function to set these values. >>>>>> >>>>>> But we can use standard v4l2_pix_format_mplane struct replacing the w, h >>>>>> bytesperline, sizeimage in mtk_jpeg_q_data struct like this: >>>>>> struct mtk_jpeg_q_data{ >>>>>> struct mtk_jpeg_fmt *fmt; >>>>>> struct v4l2_pix_format_mplane pix_mp; >>>>>> struct v4l2_rect enc_crop_rect >>>>>> } >>>>>> Then delete ctx->colorspace ctx->ycbcr_enc ctx->quantization >>>>>> ctx->xfer_func, becuase v4l2_pix_format_mplane in q_data has contained >>>>>> them and assign them for out_q and cap_q separately. >>>>>> >>>>>> WDYT? >>>>> >>>>> Sounds good to me. I was considering just making it like >>>>> >>>>> struct mtk_jpeg_ctx { >>>>> struct mtk_jpeg_fmt *src_fmt; >>>>> struct v4l2_pix_format_mplane src_pix_mp; >>>>> struct v4l2_rect src_crop; >>>>> >>>>> struct mtk_jpeg_fmt *dst_fmt; >>>>> struct v4l2_pix_format_mplane dst_pix_mp; >>>>> struct v4l2_rect dst_crop; >>>>> }; >>>>> >>>>> but I like your suggestion as well, as long as custom data structures >>>>> are not used to store standard information. >>>> >>>> Dear Tomasz, >>>> >>>> I used the structure like below: >>>> struct mtk_jpeg_q_data{ >>>> struct mtk_jpeg_fmt *fmt; >>>> struct v4l2_pix_format_mplane pix_mp; >>>> struct v4l2_rect enc_crop_rect >>>> } >>>> Then delete ctx->colorspace ctx->ycbcr_enc ctx->quantization >>>> ctx->xfer_func, becuase v4l2_pix_format_mplane in q_data has contained >>>> them and assign them for out_q and cap_q separately. >>>> >>>> Then the v4l2_compliance test will fail, the fail log as below: >>>> Format ioctls: >>>> test VIDIOC_ENUM_FMT/FRAMESIZES/FRAMEINTERVALS: OK >>>> test VIDIOC_G/S_PARM: OK (Not Supported) >>>> test VIDIOC_G_FBUF: OK (Not Supported) >>>> test VIDIOC_G_FMT: OK >>>> test VIDIOC_TRY_FMT: OK >>>> fail: v4l2-test-formats.cpp(836): >>>> fmt_cap.g_colorspace() != col >>>> test VIDIOC_S_FMT: FAIL >>>> test VIDIOC_G_SLICED_VBI_CAP: OK (Not Supported) >>>> test Cropping: OK >>>> test Composing: OK (Not Supported) >>>> test Scaling: OK (Not Supported) >>>> >>>> The source code of v4l2-test-formats.cpp as below: >>>> >>>> static int testM2MFormats(struct node *node) >>>> { >>>> cv4l_fmt fmt_out; >>>> cv4l_fmt fmt; >>>> cv4l_fmt fmt_cap; >>>> __u32 cap_type = node->g_type(); >>>> __u32 out_type = v4l_type_invert(cap_type); >>>> __u32 col, ycbcr_enc, quant, xfer_func; >>>> >>>> fail_on_test(node->g_fmt(fmt_out, out_type)); >>>> node->g_fmt(fmt_cap, cap_type); >>>> fail_on_test(fmt_cap.g_colorspace() != fmt_out.g_colorspace()); >>>> fail_on_test(fmt_cap.g_ycbcr_enc() != fmt_out.g_ycbcr_enc()); >>>> fail_on_test(fmt_cap.g_quantization() != fmt_out.g_quantization()); >>>> fail_on_test(fmt_cap.g_xfer_func() != fmt_out.g_xfer_func()); >>>> col = fmt_out.g_colorspace() == V4L2_COLORSPACE_SMPTE170M ? >>>> V4L2_COLORSPACE_REC709 : V4L2_COLORSPACE_SMPTE170M; >>>> ycbcr_enc = fmt_out.g_ycbcr_enc() == V4L2_YCBCR_ENC_601 ? >>>> V4L2_YCBCR_ENC_709 : V4L2_YCBCR_ENC_601; >>>> quant = fmt_out.g_quantization() == V4L2_QUANTIZATION_LIM_RANGE ? >>>> V4L2_QUANTIZATION_FULL_RANGE : V4L2_QUANTIZATION_LIM_RANGE; >>>> xfer_func = fmt_out.g_xfer_func() == V4L2_XFER_FUNC_SRGB ? >>>> V4L2_XFER_FUNC_709 : V4L2_XFER_FUNC_SRGB; >>>> fmt_out.s_colorspace(col); >>>> fmt_out.s_xfer_func(xfer_func); >>>> fmt_out.s_ycbcr_enc(ycbcr_enc); >>>> fmt_out.s_quantization(quant); >>>> node->s_fmt(fmt_out); >>>> fail_on_test(fmt_out.g_colorspace() != col); >>>> fail_on_test(fmt_out.g_xfer_func() != xfer_func); >>>> fail_on_test(fmt_out.g_ycbcr_enc() != ycbcr_enc); >>>> fail_on_test(fmt_out.g_quantization() != quant); >>>> node->g_fmt(fmt_cap); >>>> fail_on_test(fmt_cap.g_colorspace() != col); // line 836 >>>> fail_on_test(fmt_cap.g_xfer_func() != xfer_func); >>>> fail_on_test(fmt_cap.g_ycbcr_enc() != ycbcr_enc); >>>> fail_on_test(fmt_cap.g_quantization() != quant); >>>> } >>>> >>>> It needs that cap's colorspace equals out's colorspace when userspace >>>> just set out's colorspace and then get cap's colorspace. However, cap's >>>> colorspace which getted from driver equals V4L2_COLORSPACE_JPEG, because >>>> the code in g_fmt() like this: >>>> pix_mp->colorspace = q_data->pix_mp.colorspace; >>>> pix_mp->ycbcr_enc = q_data->pix_mp.ycbcr_enc; >>>> pix_mp->xfer_func = q_data->pix_mp.xfer_func; >>>> pix_mp->quantization = q_data->pix_mp.quantization; >>>> >>>> How should I handle this case? Should I store them(colorspace, >>>> sfer_func, ycbcr_enc, quatization) in ctx as the orinal desin? Then I >>>> can get them from g_fmt() like this: >>>> pix_mp->colorspace = ctx->colorspace; >>>> pix_mp->ycbcr_enc = ctx->ycbcr_enc; >>>> pix_mp->xfer_func = ctx->xfer_func; >>>> pix_mp->quantization = ctx->quantization; >>> >>> Why would there be any other colorspace accepted? I suppose that the >>> hardware only supports the JPEG color space, so it shouldn't accept >>> any other colorspace in TRY_FMT (and thus S_FMT) anyway. >>> >>> Still, for correctness, I would suggest propagating the colorspace >>> (and related) information from OUTPUT format to CAPTURE format in >>> S_FMT(OUTPUT). >> Dear Tomasz, >> If the driver doesn't accept any other colorspace from userspace(means >> the colorspace always equals V4L2_COLORSPACE_JPEG on my understanding), >> the v4l2_compliance will fail at line 831: >> fail_on_test(fmt_out.g_colorspace() != col); // line 831 >> >> The v4l2_compliance needs driver can accept other colorspace and capture >> colorspace equals output colorspace. >> >> I try to propagate the colorspace from OUTPUT format to CAPTURE foramt >> in S_FMT, the compliance test succeed. >> >> Should the jpeg driver can accept other colorspace? > > Hans, could you advise on what is the proper behavior regarding > colorimetry support? The JPEG encoder doesn't do any colorspace conversion as far as I know, so to get a valid JPEG out of the encoder the colorspace of the OUTPUT format has to be the same as how the captured JPEG data is interpreted. So when userspace sets the output format, then the driver should replace the colorspace field by V4L2_COLORSPACE_SRGB, xfer_func by V4L2_XFER_FUNC and ycbcr_enc to V4L2_YCBCR_ENC_601. What I don't know is if this encoder driver supports both limited and full range for the YUV data. If so, then it supports both values for that field. If it supports only limited or full range, then it should replace the quantization field for the output format with a fixed value as well. The compliance test doesn't handle JPEG codecs very well when it comes to colorspace. I'll try to fix this. Regards, Hans