Hi Laurent and Jacopo. I'd started this reply before Laurent's follow up - I'll respond to that in a moment. On Mon, 4 Mar 2024 at 17:12, Jacopo Mondi <jacopo.mondi@xxxxxxxxxxxxxxxx> wrote: > > Hi Laurent > > On Fri, Mar 01, 2024 at 11:32:24PM +0200, Laurent Pinchart wrote: > > From: Dave Stevenson <dave.stevenson@xxxxxxxxxxxxxxx> > > > > Add a driver for the Unicam camera receiver block on BCM283x processors. > > It is represented as two video device nodes: unicam-image and > > unicam-embedded which are connected to an internal subdev (named > > unicam-subdev) in order to manage streams routing. > > > > Signed-off-by: Dave Stevenson <dave.stevenson@xxxxxxxxxxxxxxx> > > Co-developed-by: Naushir Patuck <naush@xxxxxxxxxxxxxxx> > > Signed-off-by: Naushir Patuck <naush@xxxxxxxxxxxxxxx> > > Co-developed-by: Jean-Michel Hautbois <jeanmichel.hautbois@xxxxxxxxxxxxxxxx> > > Signed-off-by: Jean-Michel Hautbois <jeanmichel.hautbois@xxxxxxxxxxxxxxxx> > > Co-developed-by: Laurent Pinchart <laurent.pinchart@xxxxxxxxxxxxxxxx> > > Signed-off-by: Laurent Pinchart <laurent.pinchart@xxxxxxxxxxxxxxxx> > > --- > > Changes since v5: > > > > - Move to drivers/media/platform/broadcom/ > > - Port to the upstream V4L2 streams API > > - Rebase on latest metadata API proposal > > - Add missing error message > > - Drop unneeded documentation block for unicam_isr() > > - Drop unneeded dev_dbg() and dev_err() messages > > - Drop unneeded streams_mask and fmt checks > > - Drop unused unicam_sd_pad_is_sink() > > - Drop unneeded includes > > - Drop v4l2_ctrl_subscribe_event() call > > - Use pm_runtime_resume_and_get() > > - Indentation and line wrap fixes > > - Let the framework set bus_info > > - Use v4l2_fwnode_endpoint_parse() > > - Fix media device cleanup > > - Drop lane reordering checks > > - Fix subdev state locking > > - Drop extra debug messages > > - Move clock handling to runtime PM handlers > > - Reorder functions > > - Rename init functions for more clarity > > - Initialize runtime PM earlier > > - Clarify error messages > > - Simplify subdev init with local variable > > - Fix subdev cleanup > > - Fix typos and indentation > > - Don't initialize local variables needlessly > > - Simplify num lanes check > > - Fix metadata handling in subdev set_fmt > > - Drop manual fallback to .s_stream() > > - Pass v4l2_pix_format to unicam_calc_format_size_bpl() > > - Simplify unicam_set_default_format() > > - Fix default format settings > > - Add busy check in unicam_s_fmt_meta() > > - Add missing \n at end of format strings > > - Fix metadata handling in subdev set_fmt > > - Fix locking when starting streaming > > - Return buffers from start streaming fails > > - Fix format validation for metadata node > > - Use video_device_pipeline_{start,stop}() helpers > > - Simplify format enumeration > > - Drop unset variable > > - Update MAINTAINERS entry > > - Update to the upstream v4l2_async_nf API > > - Update to the latest subdev routing API > > - Update to the latest subdev state API > > - Move from subdev .init_cfg() to .init_state() > > - Update to the latest videobuf2 API > > - Fix v4l2_subdev_enable_streams() error check > > - Use correct pad for the connected subdev > > - Return buffers to vb2 when start streaming fails > > - Improve debugging in start streaming handler > > - Simplify DMA address management > > - Drop comment about bcm2835-camera driver > > - Clarify comments that explain min/max sizes > > - Pass v4l2_pix_format to unicam_try_fmt() > > - Drop unneeded local variables > > - Rename image-related constants and functions > > - Turn unicam_fmt.metadata_fmt into bool > > - Rename unicam_fmt to unicam_format_info > > - Rename unicam_format_info variables to fmtinfo > > - Rename unicam_node.v_fmt to fmt > > - Add metadata formats for RAW10, RAW12 and RAW14 > > - Make metadata formats line-based > > - Validate format on metadata video device > > - Add Co-devlopped-by tags > > > > Changes since v3: > > > > - Add the vendor prefix for DT name > > - Use the reg-names in DT parsing > > - Remove MAINTAINERS entry > > > > Changes since v2: > > > > - Change code organization > > - Remove unused variables > > - Correct the fmt_meta functions > > - Rewrite the start/stop streaming > > - You can now start the image node alone, but not the metadata one > > - The buffers are allocated per-node > > - only the required stream is started, if the route exists and is > > enabled > > - Prefix the macros with UNICAM_ to not have too generic names > > - Drop colorspace support > > > > Changes since v1: > > > > - Replace the unicam_{info,debug,error} macros with dev_*() > > --- > > MAINTAINERS | 1 + > > drivers/media/platform/Kconfig | 1 + > > drivers/media/platform/Makefile | 1 + > > drivers/media/platform/broadcom/Kconfig | 23 + > > drivers/media/platform/broadcom/Makefile | 3 + > > .../platform/broadcom/bcm2835-unicam-regs.h | 255 ++ > > .../media/platform/broadcom/bcm2835-unicam.c | 2607 +++++++++++++++++ > > 7 files changed, 2891 insertions(+) > > create mode 100644 drivers/media/platform/broadcom/Kconfig > > create mode 100644 drivers/media/platform/broadcom/Makefile > > create mode 100644 drivers/media/platform/broadcom/bcm2835-unicam-regs.h > > create mode 100644 drivers/media/platform/broadcom/bcm2835-unicam.c > > > > diff --git a/MAINTAINERS b/MAINTAINERS > > index e50a59654e6e..cc350729f467 100644 > > --- a/MAINTAINERS > > +++ b/MAINTAINERS > > @@ -4002,6 +4002,7 @@ M: Raspberry Pi Kernel Maintenance <kernel-list@xxxxxxxxxxxxxxx> > > L: linux-media@xxxxxxxxxxxxxxx > > S: Maintained > > F: Documentation/devicetree/bindings/media/brcm,bcm2835-unicam.yaml > > +F: drivers/media/platform/bcm2835/ > > > > BROADCOM BCM47XX MIPS ARCHITECTURE > > M: Hauke Mehrtens <hauke@xxxxxxxxxx> > > diff --git a/drivers/media/platform/Kconfig b/drivers/media/platform/Kconfig > > index 91e54215de3a..2d79bfc68c15 100644 > > --- a/drivers/media/platform/Kconfig > > +++ b/drivers/media/platform/Kconfig > > @@ -67,6 +67,7 @@ source "drivers/media/platform/amlogic/Kconfig" > > source "drivers/media/platform/amphion/Kconfig" > > source "drivers/media/platform/aspeed/Kconfig" > > source "drivers/media/platform/atmel/Kconfig" > > +source "drivers/media/platform/broadcom/Kconfig" > > source "drivers/media/platform/cadence/Kconfig" > > source "drivers/media/platform/chips-media/Kconfig" > > source "drivers/media/platform/intel/Kconfig" > > diff --git a/drivers/media/platform/Makefile b/drivers/media/platform/Makefile > > index 3296ec1ebe16..da17301f7439 100644 > > --- a/drivers/media/platform/Makefile > > +++ b/drivers/media/platform/Makefile > > @@ -10,6 +10,7 @@ obj-y += amlogic/ > > obj-y += amphion/ > > obj-y += aspeed/ > > obj-y += atmel/ > > +obj-y += broadcom/ > > obj-y += cadence/ > > obj-y += chips-media/ > > obj-y += intel/ > > diff --git a/drivers/media/platform/broadcom/Kconfig b/drivers/media/platform/broadcom/Kconfig > > new file mode 100644 > > index 000000000000..cc2c9afcc948 > > --- /dev/null > > +++ b/drivers/media/platform/broadcom/Kconfig > > @@ -0,0 +1,23 @@ > > +# SPDX-License-Identifier: GPL-2.0 > > + > > +config VIDEO_BCM2835_UNICAM > > + tristate "Broadcom BCM283x/BCM271x Unicam video capture driver" > > + depends on ARCH_BCM2835 || COMPILE_TEST > > + depends on PM > > + depends on VIDEO_DEV > > + select MEDIA_CONTROLLER > > + select V4L2_FWNODE > > + select VIDEO_V4L2_SUBDEV_API > > + select VIDEOBUF2_DMA_CONTIG > > + help > > + Say Y here to enable support for the BCM283x/BCM271x CSI-2 receiver. > > + This is a V4L2 driver that controls the CSI-2 receiver directly, > > + independently from the VC4 firmware. > > + > > + This driver is mutually exclusive with the use of bcm2835-camera. The > > + firmware will disable all access to the peripheral from within the > > + firmware if it finds a DT node using it, and bcm2835-camera will > > + therefore fail to probe. > > + > > + To compile this driver as a module, choose M here. The module will be > > + called bcm2835-unicam. > > diff --git a/drivers/media/platform/broadcom/Makefile b/drivers/media/platform/broadcom/Makefile > > new file mode 100644 > > index 000000000000..03d2045aba2e > > --- /dev/null > > +++ b/drivers/media/platform/broadcom/Makefile > > @@ -0,0 +1,3 @@ > > +# SPDX-License-Identifier: GPL-2.0 > > + > > +obj-$(CONFIG_VIDEO_BCM2835_UNICAM) += bcm2835-unicam.o > > diff --git a/drivers/media/platform/broadcom/bcm2835-unicam-regs.h b/drivers/media/platform/broadcom/bcm2835-unicam-regs.h > > new file mode 100644 > > index 000000000000..84775fd2fac5 > > --- /dev/null > > +++ b/drivers/media/platform/broadcom/bcm2835-unicam-regs.h > > @@ -0,0 +1,255 @@ > > +/* SPDX-License-Identifier: GPL-2.0-only */ > > + > > +/* > > + * Copyright (C) 2017-2020 Raspberry Pi Trading. > > + * Dave Stevenson <dave.stevenson@xxxxxxxxxxxxxxx> > > + */ > > + > > +#ifndef VC4_REGS_UNICAM_H > > +#define VC4_REGS_UNICAM_H > > + > > +#include <linux/bits.h> > > + > > +/* > > + * The following values are taken from files found within the code drop > > + * made by Broadcom for the BCM21553 Graphics Driver, predominantly in > > + * brcm_usrlib/dag/vmcsx/vcinclude/hardware_vc4.h. > > + * They have been modified to be only the register offset. > > + */ > > +#define UNICAM_CTRL 0x000 > > +#define UNICAM_STA 0x004 > > +#define UNICAM_ANA 0x008 > > +#define UNICAM_PRI 0x00c > > +#define UNICAM_CLK 0x010 > > +#define UNICAM_CLT 0x014 > > +#define UNICAM_DAT0 0x018 > > +#define UNICAM_DAT1 0x01c > > +#define UNICAM_DAT2 0x020 > > +#define UNICAM_DAT3 0x024 > > +#define UNICAM_DLT 0x028 > > +#define UNICAM_CMP0 0x02c > > +#define UNICAM_CMP1 0x030 > > +#define UNICAM_CAP0 0x034 > > +#define UNICAM_CAP1 0x038 > > +#define UNICAM_ICTL 0x100 > > +#define UNICAM_ISTA 0x104 > > +#define UNICAM_IDI0 0x108 > > +#define UNICAM_IPIPE 0x10c > > +#define UNICAM_IBSA0 0x110 > > +#define UNICAM_IBEA0 0x114 > > +#define UNICAM_IBLS 0x118 > > +#define UNICAM_IBWP 0x11c > > +#define UNICAM_IHWIN 0x120 > > +#define UNICAM_IHSTA 0x124 > > +#define UNICAM_IVWIN 0x128 > > +#define UNICAM_IVSTA 0x12c > > +#define UNICAM_ICC 0x130 > > +#define UNICAM_ICS 0x134 > > +#define UNICAM_IDC 0x138 > > +#define UNICAM_IDPO 0x13c > > +#define UNICAM_IDCA 0x140 > > +#define UNICAM_IDCD 0x144 > > +#define UNICAM_IDS 0x148 > > +#define UNICAM_DCS 0x200 > > +#define UNICAM_DBSA0 0x204 > > +#define UNICAM_DBEA0 0x208 > > +#define UNICAM_DBWP 0x20c > > +#define UNICAM_DBCTL 0x300 > > +#define UNICAM_IBSA1 0x304 > > +#define UNICAM_IBEA1 0x308 > > +#define UNICAM_IDI1 0x30c > > +#define UNICAM_DBSA1 0x310 > > +#define UNICAM_DBEA1 0x314 > > +#define UNICAM_MISC 0x400 > > + > > +/* > > + * The following bitmasks are from the kernel released by Broadcom > > + * for Android - https://android.googlesource.com/kernel/bcm/ > > + * The Rhea, Hawaii, and Java chips all contain the same VideoCore4 > > + * Unicam block as BCM2835, as defined in eg > > + * arch/arm/mach-rhea/include/mach/rdb_A0/brcm_rdb_cam.h and similar. > > + * Values reworked to use the kernel BIT and GENMASK macros. > > + * > > + * Some of the bit mnenomics have been amended to match the datasheet. > > + */ > > +/* UNICAM_CTRL Register */ > > +#define UNICAM_CPE BIT(0) > > +#define UNICAM_MEM BIT(1) > > +#define UNICAM_CPR BIT(2) > > +#define UNICAM_CPM_MASK GENMASK(3, 3) > > +#define UNICAM_CPM_CSI2 0 > > +#define UNICAM_CPM_CCP2 1 > > +#define UNICAM_SOE BIT(4) > > +#define UNICAM_DCM_MASK GENMASK(5, 5) > > +#define UNICAM_DCM_STROBE 0 > > +#define UNICAM_DCM_DATA 1 > > +#define UNICAM_SLS BIT(6) > > +#define UNICAM_PFT_MASK GENMASK(11, 8) > > +#define UNICAM_OET_MASK GENMASK(20, 12) > > + > > +/* UNICAM_STA Register */ > > +#define UNICAM_SYN BIT(0) > > +#define UNICAM_CS BIT(1) > > +#define UNICAM_SBE BIT(2) > > +#define UNICAM_PBE BIT(3) > > +#define UNICAM_HOE BIT(4) > > +#define UNICAM_PLE BIT(5) > > +#define UNICAM_SSC BIT(6) > > +#define UNICAM_CRCE BIT(7) > > +#define UNICAM_OES BIT(8) > > +#define UNICAM_IFO BIT(9) > > +#define UNICAM_OFO BIT(10) > > +#define UNICAM_BFO BIT(11) > > +#define UNICAM_DL BIT(12) > > +#define UNICAM_PS BIT(13) > > +#define UNICAM_IS BIT(14) > > +#define UNICAM_PI0 BIT(15) > > +#define UNICAM_PI1 BIT(16) > > +#define UNICAM_FSI_S BIT(17) > > +#define UNICAM_FEI_S BIT(18) > > +#define UNICAM_LCI_S BIT(19) > > +#define UNICAM_BUF0_RDY BIT(20) > > +#define UNICAM_BUF0_NO BIT(21) > > +#define UNICAM_BUF1_RDY BIT(22) > > +#define UNICAM_BUF1_NO BIT(23) > > +#define UNICAM_DI BIT(24) > > + > > +#define UNICAM_STA_MASK_ALL \ > > + (UNICAM_DL | \ > > + UNICAM_SBE | \ > > + UNICAM_PBE | \ > > + UNICAM_HOE | \ > > + UNICAM_PLE | \ > > + UNICAM_SSC | \ > > + UNICAM_CRCE | \ > > + UNICAM_IFO | \ > > + UNICAM_OFO | \ > > + UNICAM_PS | \ > > + UNICAM_PI0 | \ > > + UNICAM_PI1) > > + > > +/* UNICAM_ANA Register */ > > +#define UNICAM_APD BIT(0) > > +#define UNICAM_BPD BIT(1) > > +#define UNICAM_AR BIT(2) > > +#define UNICAM_DDL BIT(3) > > +#define UNICAM_CTATADJ_MASK GENMASK(7, 4) > > +#define UNICAM_PTATADJ_MASK GENMASK(11, 8) > > + > > +/* UNICAM_PRI Register */ > > +#define UNICAM_PE BIT(0) > > +#define UNICAM_PT_MASK GENMASK(2, 1) > > +#define UNICAM_NP_MASK GENMASK(7, 4) > > +#define UNICAM_PP_MASK GENMASK(11, 8) > > +#define UNICAM_BS_MASK GENMASK(15, 12) > > +#define UNICAM_BL_MASK GENMASK(17, 16) > > + > > +/* UNICAM_CLK Register */ > > +#define UNICAM_CLE BIT(0) > > +#define UNICAM_CLPD BIT(1) > > +#define UNICAM_CLLPE BIT(2) > > +#define UNICAM_CLHSE BIT(3) > > +#define UNICAM_CLTRE BIT(4) > > +#define UNICAM_CLAC_MASK GENMASK(8, 5) > > +#define UNICAM_CLSTE BIT(29) > > + > > +/* UNICAM_CLT Register */ > > +#define UNICAM_CLT1_MASK GENMASK(7, 0) > > +#define UNICAM_CLT2_MASK GENMASK(15, 8) > > + > > +/* UNICAM_DATn Registers */ > > +#define UNICAM_DLE BIT(0) > > +#define UNICAM_DLPD BIT(1) > > +#define UNICAM_DLLPE BIT(2) > > +#define UNICAM_DLHSE BIT(3) > > +#define UNICAM_DLTRE BIT(4) > > +#define UNICAM_DLSM BIT(5) > > +#define UNICAM_DLFO BIT(28) > > +#define UNICAM_DLSTE BIT(29) > > + > > +#define UNICAM_DAT_MASK_ALL (UNICAM_DLSTE | UNICAM_DLFO) > > + > > +/* UNICAM_DLT Register */ > > +#define UNICAM_DLT1_MASK GENMASK(7, 0) > > +#define UNICAM_DLT2_MASK GENMASK(15, 8) > > +#define UNICAM_DLT3_MASK GENMASK(23, 16) > > + > > +/* UNICAM_ICTL Register */ > > +#define UNICAM_FSIE BIT(0) > > +#define UNICAM_FEIE BIT(1) > > +#define UNICAM_IBOB BIT(2) > > +#define UNICAM_FCM BIT(3) > > +#define UNICAM_TFC BIT(4) > > +#define UNICAM_LIP_MASK GENMASK(6, 5) > > +#define UNICAM_LCIE_MASK GENMASK(28, 16) > > + > > +/* UNICAM_IDI0/1 Register */ > > +#define UNICAM_ID0_MASK GENMASK(7, 0) > > +#define UNICAM_ID1_MASK GENMASK(15, 8) > > +#define UNICAM_ID2_MASK GENMASK(23, 16) > > +#define UNICAM_ID3_MASK GENMASK(31, 24) > > + > > +/* UNICAM_ISTA Register */ > > +#define UNICAM_FSI BIT(0) > > +#define UNICAM_FEI BIT(1) > > +#define UNICAM_LCI BIT(2) > > + > > +#define UNICAM_ISTA_MASK_ALL (UNICAM_FSI | UNICAM_FEI | UNICAM_LCI) > > + > > +/* UNICAM_IPIPE Register */ > > +#define UNICAM_PUM_MASK GENMASK(2, 0) > > +/* Unpacking modes */ > > +#define UNICAM_PUM_NONE 0 > > +#define UNICAM_PUM_UNPACK6 1 > > +#define UNICAM_PUM_UNPACK7 2 > > +#define UNICAM_PUM_UNPACK8 3 > > +#define UNICAM_PUM_UNPACK10 4 > > +#define UNICAM_PUM_UNPACK12 5 > > +#define UNICAM_PUM_UNPACK14 6 > > +#define UNICAM_PUM_UNPACK16 7 > > +#define UNICAM_DDM_MASK GENMASK(6, 3) > > +#define UNICAM_PPM_MASK GENMASK(9, 7) > > +/* Packing modes */ > > +#define UNICAM_PPM_NONE 0 > > +#define UNICAM_PPM_PACK8 1 > > +#define UNICAM_PPM_PACK10 2 > > +#define UNICAM_PPM_PACK12 3 > > +#define UNICAM_PPM_PACK14 4 > > +#define UNICAM_PPM_PACK16 5 > > +#define UNICAM_DEM_MASK GENMASK(11, 10) > > +#define UNICAM_DEBL_MASK GENMASK(14, 12) > > +#define UNICAM_ICM_MASK GENMASK(16, 15) > > +#define UNICAM_IDM_MASK GENMASK(17, 17) > > + > > +/* UNICAM_ICC Register */ > > +#define UNICAM_ICFL_MASK GENMASK(4, 0) > > +#define UNICAM_ICFH_MASK GENMASK(9, 5) > > +#define UNICAM_ICST_MASK GENMASK(12, 10) > > +#define UNICAM_ICLT_MASK GENMASK(15, 13) > > +#define UNICAM_ICLL_MASK GENMASK(31, 16) > > + > > +/* UNICAM_DCS Register */ > > +#define UNICAM_DIE BIT(0) > > +#define UNICAM_DIM BIT(1) > > +#define UNICAM_DBOB BIT(3) > > +#define UNICAM_FDE BIT(4) > > +#define UNICAM_LDP BIT(5) > > +#define UNICAM_EDL_MASK GENMASK(15, 8) > > + > > +/* UNICAM_DBCTL Register */ > > +#define UNICAM_DBEN BIT(0) > > +#define UNICAM_BUF0_IE BIT(1) > > +#define UNICAM_BUF1_IE BIT(2) > > + > > +/* UNICAM_CMP[0,1] register */ > > +#define UNICAM_PCE BIT(31) > > +#define UNICAM_GI BIT(9) > > +#define UNICAM_CPH BIT(8) > > +#define UNICAM_PCVC_MASK GENMASK(7, 6) > > +#define UNICAM_PCDT_MASK GENMASK(5, 0) > > + > > +/* UNICAM_MISC register */ > > +#define UNICAM_FL0 BIT(6) > > +#define UNICAM_FL1 BIT(9) > > + > > +#endif > > diff --git a/drivers/media/platform/broadcom/bcm2835-unicam.c b/drivers/media/platform/broadcom/bcm2835-unicam.c > > new file mode 100644 > > index 000000000000..716c89b8a217 > > --- /dev/null > > +++ b/drivers/media/platform/broadcom/bcm2835-unicam.c > > @@ -0,0 +1,2607 @@ > > +// SPDX-License-Identifier: GPL-2.0-only > > +/* > > + * BCM283x / BCM271x Unicam Capture Driver > > + * > > + * Copyright (C) 2017-2020 - Raspberry Pi (Trading) Ltd. > > + * > > + * Dave Stevenson <dave.stevenson@xxxxxxxxxxxxxxx> > > + * > > + * Based on TI am437x driver by > > + * Benoit Parrot <bparrot@xxxxxx> > > + * Lad, Prabhakar <prabhakar.csengg@xxxxxxxxx> > > + * > > + * and TI CAL camera interface driver by > > + * Benoit Parrot <bparrot@xxxxxx> > > + * > > + * > > + * There are two camera drivers in the kernel for BCM283x - this one and > > + * bcm2835-camera (currently in staging). > > + * > > + * This driver directly controls the Unicam peripheral - there is no > > + * involvement with the VideoCore firmware. Unicam receives CSI-2 or CCP2 data > > + * and writes it into SDRAM. The only potential processing options are to > > + * repack Bayer data into an alternate format, and applying windowing. The > > + * repacking does not shift the data, so can repack V4L2_PIX_FMT_Sxxxx10P to > > + * V4L2_PIX_FMT_Sxxxx10, or V4L2_PIX_FMT_Sxxxx12P to V4L2_PIX_FMT_Sxxxx12, but > > + * not generically up to V4L2_PIX_FMT_Sxxxx16. The driver will add both formats > > + * where the relevant formats are defined, and will automatically configure the > > + * repacking as required. Support for windowing may be added later. > > Does this last paragraph applies ? I see all the supported packed raw > format having an 'unpacked_fourcc' associated (all but the 8 bit ones > ofc) The penultimate sentence ("The driver will add both formats where the relevant formats are defined") can be dropped as you've dropped the video-node centric mode of operation, therefore all formats (whether packed or unpacked) are always advertised. The rest of the paragraph still applies - if you request a V4L2_PIX_FMT_SxxxxBB (not packed) format, then it the incoming data will be unpacked for you. > > + * > > + * It should be possible to connect this driver to any sensor with a suitable > > + * output interface and V4L2 subdevice driver. > > + */ > > + > > +#include <linux/clk.h> > > +#include <linux/delay.h> > > +#include <linux/device.h> > > +#include <linux/dma-mapping.h> > > +#include <linux/err.h> > > +#include <linux/interrupt.h> > > +#include <linux/io.h> > > +#include <linux/module.h> > > +#include <linux/of.h> > > +#include <linux/of_device.h> > > +#include <linux/platform_device.h> > > +#include <linux/pm_runtime.h> > > +#include <linux/slab.h> > > +#include <linux/videodev2.h> > > + > > +#include <media/v4l2-async.h> > > +#include <media/v4l2-common.h> > > +#include <media/v4l2-dev.h> > > +#include <media/v4l2-device.h> > > +#include <media/v4l2-event.h> > > +#include <media/v4l2-ioctl.h> > > +#include <media/v4l2-fwnode.h> > > +#include <media/v4l2-mc.h> > > +#include <media/videobuf2-dma-contig.h> > > + > > +#include "bcm2835-unicam-regs.h" > > + > > +#define UNICAM_MODULE_NAME "unicam" > > + > > +/* > > + * Unicam must request a minimum of 250Mhz from the VPU clock. > > + * Otherwise the input FIFOs overrun and cause image corruption. > > + */ > > +#define UNICAM_MIN_VPU_CLOCK_RATE (250 * 1000 * 1000) > > Shouldn't this be set in DT ? > > > + > > +/* Unicam has an internal DMA alignment constraint of 16 bytes for each line. */ > > +#define UNICAM_DMA_BPL_ALIGNMENT 16 > > + > > +/* > > + * The image stride is stored in a 16 bit register, and needs to be aligned to > > + * the DMA constraint. As the ISP in the same SoC has a 32 bytes alignment > > + * constraint on its input, set the image stride alignment to 32 bytes here as > > + * well to avoid incompatible configurations. > > + */ > > +#define UNICAM_IMAGE_BPL_ALIGNMENT 32 > > +#define UNICAM_IMAGE_MAX_BPL ((1 << 16) - UNICAM_IMAGE_BPL_ALIGNMENT) > > + > > +/* > > + * Max width is therefore determined by the max stride divided by the number of > > + * bits per pixel. Take 32bpp as a worst case. No imposed limit on the height, > > + * so adopt a square image for want of anything better. > > + */ > > +#define UNICAM_IMAGE_MIN_WIDTH 16 > > +#define UNICAM_IMAGE_MIN_HEIGHT 16 > > +#define UNICAM_IMAGE_MAX_WIDTH (UNICAM_IMAGE_MAX_BPL / 4) > > +#define UNICAM_IMAGE_MAX_HEIGHT UNICAM_IMAGE_MAX_WIDTH > > + > > +/* > > + * There's no intrinsic limits on the width and height for embedded dat. Use > > + * the same maximum values as for the image, to avoid overflows in the image > > + * size computation. > > + */ > > +#define UNICAM_META_MIN_WIDTH 1 > > +#define UNICAM_META_MIN_HEIGHT 1 > > +#define UNICAM_META_MAX_WIDTH UNICAM_IMAGE_MAX_WIDTH > > +#define UNICAM_META_MAX_HEIGHT UNICAM_IMAGE_MAX_HEIGHT > > + > > +/* > > + * Size of the dummy buffer. Can be any size really, but the DMA > > + * allocation works in units of page sizes. > > + */ > > +#define UNICAM_DUMMY_BUF_SIZE PAGE_SIZE > > + > > +#define UNICAM_SD_PAD_SINK 0 > > +#define UNICAM_SD_PAD_SOURCE_IMAGE 1 > > +#define UNICAM_SD_PAD_SOURCE_METADATA 2 > > +#define UNICAM_SD_NUM_PADS (1 + UNICAM_SD_PAD_SOURCE_METADATA) > > How about an enum then ? > > > + > > +enum unicam_node_type { > > + UNICAM_IMAGE_NODE, > > + UNICAM_METADATA_NODE, > > + UNICAM_MAX_NODES > > +}; > > + > > +/* > > + * struct unicam_format_info - Unicam media bus format information > > + * @fourcc: V4L2 pixel format FCC identifier. 0 if n/a. > > + * @unpacked_fourcc: V4L2 pixel format FCC identifier if the data is expanded > > + * out to 16bpp. 0 if n/a. > > + * @code: V4L2 media bus format code. > > + * @depth: Bits per pixel as delivered from the source. > > + * @csi_dt: CSI data type. > > + * @metadata_fmt: This format only applies to the metadata pad. > > + */ > > +struct unicam_format_info { > > + u32 fourcc; > > + u32 unpacked_fourcc; > > + u32 code; > > + u8 depth; > > + u8 csi_dt; > > + bool metadata_fmt; > > +}; > > + > > +struct unicam_buffer { > > + struct vb2_v4l2_buffer vb; > > + struct list_head list; > > + dma_addr_t dma_addr; > > + unsigned int size; > > +}; > > + > > +static inline struct unicam_buffer *to_unicam_buffer(struct vb2_buffer *vb) > > +{ > > + return container_of(vb, struct unicam_buffer, vb.vb2_buf); > > +} > > + > > +struct unicam_node { > > + bool registered; > > + bool streaming; > > + unsigned int id; > > + > > + /* Pointer to the current v4l2_buffer */ > > + struct unicam_buffer *cur_frm; > > + /* Pointer to the next v4l2_buffer */ > > + struct unicam_buffer *next_frm; > > + /* video capture */ > > + const struct unicam_format_info *fmtinfo; > > + /* Used to store current pixel format */ > > + struct v4l2_format fmt; > > + /* Buffer queue used in video-buf */ > > + struct vb2_queue buffer_queue; > > + /* Queue of filled frames */ > > + struct list_head dma_queue; > > + /* IRQ lock for DMA queue */ > > + spinlock_t dma_queue_lock; > > + /* lock used to access this structure */ > > + struct mutex lock; > > + /* Identifies video device for this channel */ > > + struct video_device video_dev; > > + /* Pointer to the parent handle */ > > + struct unicam_device *dev; > > + struct media_pad pad; > > + /* > > + * Dummy buffer intended to be used by unicam > > + * if we have no other queued buffers to swap to. > > + */ > > + struct unicam_buffer dummy_buf; > > + void *dummy_buf_cpu_addr; > > +}; > > + > > +struct unicam_device { > > + struct kref kref; > > + > > + /* peripheral base address */ > > + void __iomem *base; > > + /* clock gating base address */ > > + void __iomem *clk_gate_base; > > Is the clock gating part of unicam or should this be expressed as a > clock provide and be referenced in DT through a phandle ? Covered in previous reviews and accepted as part of unicam. https://patchwork.kernel.org/project/linux-media/patch/fae3d29bba67825030c0077dd9c79534b6888512.1505916622.git.dave.stevenson@xxxxxxxxxxxxxxx/ > > + /* lp clock handle */ > > + struct clk *clock; > > + /* vpu clock handle */ > > + struct clk *vpu_clock; > > + /* V4l2 device */ > > + struct v4l2_device v4l2_dev; > > + struct media_device mdev; > > + > > + /* parent device */ > > + struct device *dev; > > + /* subdevice async Notifier */ > > s/Notifier/notifier > > > + struct v4l2_async_notifier notifier; > > + unsigned int sequence; > > + > > + /* Sensor node */ > > + struct { > > + struct v4l2_subdev *subdev; > > + struct media_pad *pad; > > + } sensor; > > + > > + /* Internal subdev */ > > + struct { > > + struct v4l2_subdev sd; > > + struct media_pad pads[UNICAM_SD_NUM_PADS]; > > + bool streaming; > > + } subdev; > > + > > + enum v4l2_mbus_type bus_type; > > + /* > > + * Stores bus.mipi_csi2.flags for CSI2 sensors, or > > + * bus.mipi_csi1.strobe for CCP2. > > + */ > > + unsigned int bus_flags; > > + unsigned int max_data_lanes; > > + unsigned int active_data_lanes; > > + > > + struct media_pipeline pipe; > > + > > + struct unicam_node node[UNICAM_MAX_NODES]; > > +}; > > + > > +static inline struct unicam_device * > > +notifier_to_unicam_device(struct v4l2_async_notifier *notifier) > > +{ > > + return container_of(notifier, struct unicam_device, notifier); > > +} > > + > > +static inline struct unicam_device * > > I thought using inline is disouraged as the compiler knows what to do > > > +sd_to_unicam_device(struct v4l2_subdev *sd) > > +{ > > + return container_of(sd, struct unicam_device, subdev.sd); > > You can also use container_of_const() if you need it > > > +} > > + > > +static void unicam_release(struct kref *kref) > > +{ > > + struct unicam_device *unicam = > > + container_of(kref, struct unicam_device, kref); > > + > > + if (unicam->mdev.dev) > > + media_device_cleanup(&unicam->mdev); > > + > > + kfree(unicam); > > +} > > + > > +static struct unicam_device *unicam_get(struct unicam_device *unicam) > > +{ > > + kref_get(&unicam->kref); > > + return unicam; > > +} > > + > > +static void unicam_put(struct unicam_device *unicam) > > +{ > > + kref_put(&unicam->kref, unicam_release); > > +} > > + > > Does this handle a bit more gracefully the pesky lifetime management > issue in v4l2 ? > > > +/* ----------------------------------------------------------------------------- > > + * Misc helper functions > > + */ > > + > > +static inline bool unicam_sd_pad_is_source(u32 pad) > > +{ > > + /* Camera RX has 1 sink pad, and N source pads */ > > + return pad != UNICAM_SD_PAD_SINK; > > +} > > + > > +static inline bool is_metadata_node(struct unicam_node *node) > > +{ > > + return node->video_dev.device_caps & V4L2_CAP_META_CAPTURE; > > +} > > + > > +static inline bool is_image_node(struct unicam_node *node) > > +{ > > + return node->video_dev.device_caps & V4L2_CAP_VIDEO_CAPTURE; > > +} > > + > > +/* ----------------------------------------------------------------------------- > > + * Format data table and helper functions > > + */ > > + > > +static const struct v4l2_mbus_framefmt unicam_default_image_format = { > > + .width = 640, > > + .height = 480, > > + .code = MEDIA_BUS_FMT_UYVY8_1X16, > > + .field = V4L2_FIELD_NONE, > > + .colorspace = V4L2_COLORSPACE_SRGB, > > + .ycbcr_enc = V4L2_YCBCR_ENC_601, > > + .quantization = V4L2_QUANTIZATION_LIM_RANGE, > > + .xfer_func = V4L2_XFER_FUNC_SRGB, > > + .flags = 0, > > +}; > > + > > +static const struct v4l2_mbus_framefmt unicam_default_meta_format = { > > + .width = 640, > > + .height = 2, > > + .code = MEDIA_BUS_FMT_META_8, > > + .field = V4L2_FIELD_NONE, > > +}; > > + > > +static const struct unicam_format_info unicam_image_formats[] = { > > + /* YUV Formats */ > > + { > > + .fourcc = V4L2_PIX_FMT_YUYV, > > + .code = MEDIA_BUS_FMT_YUYV8_1X16, > > + .depth = 16, > > + .csi_dt = 0x1e, > > + }, { > > + .fourcc = V4L2_PIX_FMT_UYVY, > > + .code = MEDIA_BUS_FMT_UYVY8_1X16, > > + .depth = 16, > > + .csi_dt = 0x1e, > > + }, { > > + .fourcc = V4L2_PIX_FMT_YVYU, > > + .code = MEDIA_BUS_FMT_YVYU8_1X16, > > + .depth = 16, > > + .csi_dt = 0x1e, > > + }, { > > + .fourcc = V4L2_PIX_FMT_VYUY, > > + .code = MEDIA_BUS_FMT_VYUY8_1X16, > > + .depth = 16, > > + .csi_dt = 0x1e, > > + }, { > > + /* RGB Formats */ > > + .fourcc = V4L2_PIX_FMT_RGB565, /* gggbbbbb rrrrrggg */ > > + .code = MEDIA_BUS_FMT_RGB565_1X16, > > + .depth = 16, > > + .csi_dt = 0x22, > > + }, { > > + .fourcc = V4L2_PIX_FMT_RGB24, /* rgb */ > > + .code = MEDIA_BUS_FMT_RGB888_1X24, > > + .depth = 24, > > + .csi_dt = 0x24, > > + }, { > > + .fourcc = V4L2_PIX_FMT_BGR24, /* bgr */ > > + .code = MEDIA_BUS_FMT_BGR888_1X24, > > + .depth = 24, > > + .csi_dt = 0x24, > > + }, { > > + /* Bayer Formats */ > > + .fourcc = V4L2_PIX_FMT_SBGGR8, > > + .code = MEDIA_BUS_FMT_SBGGR8_1X8, > > + .depth = 8, > > + .csi_dt = 0x2a, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGBRG8, > > + .code = MEDIA_BUS_FMT_SGBRG8_1X8, > > + .depth = 8, > > + .csi_dt = 0x2a, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGRBG8, > > + .code = MEDIA_BUS_FMT_SGRBG8_1X8, > > + .depth = 8, > > + .csi_dt = 0x2a, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SRGGB8, > > + .code = MEDIA_BUS_FMT_SRGGB8_1X8, > > + .depth = 8, > > + .csi_dt = 0x2a, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SBGGR10P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SBGGR10, > > + .code = MEDIA_BUS_FMT_SBGGR10_1X10, > > + .depth = 10, > > + .csi_dt = 0x2b, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGBRG10P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGBRG10, > > + .code = MEDIA_BUS_FMT_SGBRG10_1X10, > > + .depth = 10, > > + .csi_dt = 0x2b, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGRBG10P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGRBG10, > > + .code = MEDIA_BUS_FMT_SGRBG10_1X10, > > + .depth = 10, > > + .csi_dt = 0x2b, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SRGGB10P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SRGGB10, > > + .code = MEDIA_BUS_FMT_SRGGB10_1X10, > > + .depth = 10, > > + .csi_dt = 0x2b, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SBGGR12P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SBGGR12, > > + .code = MEDIA_BUS_FMT_SBGGR12_1X12, > > + .depth = 12, > > + .csi_dt = 0x2c, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGBRG12P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGBRG12, > > + .code = MEDIA_BUS_FMT_SGBRG12_1X12, > > + .depth = 12, > > + .csi_dt = 0x2c, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGRBG12P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGRBG12, > > + .code = MEDIA_BUS_FMT_SGRBG12_1X12, > > + .depth = 12, > > + .csi_dt = 0x2c, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SRGGB12P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SRGGB12, > > + .code = MEDIA_BUS_FMT_SRGGB12_1X12, > > + .depth = 12, > > + .csi_dt = 0x2c, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SBGGR14P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SBGGR14, > > + .code = MEDIA_BUS_FMT_SBGGR14_1X14, > > + .depth = 14, > > + .csi_dt = 0x2d, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGBRG14P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGBRG14, > > + .code = MEDIA_BUS_FMT_SGBRG14_1X14, > > + .depth = 14, > > + .csi_dt = 0x2d, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SGRBG14P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SGRBG14, > > + .code = MEDIA_BUS_FMT_SGRBG14_1X14, > > + .depth = 14, > > + .csi_dt = 0x2d, > > + }, { > > + .fourcc = V4L2_PIX_FMT_SRGGB14P, > > + .unpacked_fourcc = V4L2_PIX_FMT_SRGGB14, > > + .code = MEDIA_BUS_FMT_SRGGB14_1X14, > > + .depth = 14, > > + .csi_dt = 0x2d, > > + }, { > > + /* 16 bit Bayer formats could be supported. */ > > + > > + /* Greyscale formats */ > > + .fourcc = V4L2_PIX_FMT_GREY, > > + .code = MEDIA_BUS_FMT_Y8_1X8, > > + .depth = 8, > > + .csi_dt = 0x2a, > > + }, { > > + .fourcc = V4L2_PIX_FMT_Y10P, > > + .unpacked_fourcc = V4L2_PIX_FMT_Y10, > > + .code = MEDIA_BUS_FMT_Y10_1X10, > > + .depth = 10, > > + .csi_dt = 0x2b, > > + }, { > > + .fourcc = V4L2_PIX_FMT_Y12P, > > + .unpacked_fourcc = V4L2_PIX_FMT_Y12, > > + .code = MEDIA_BUS_FMT_Y12_1X12, > > + .depth = 12, > > + .csi_dt = 0x2c, > > + }, { > > + .fourcc = V4L2_PIX_FMT_Y14P, > > + .unpacked_fourcc = V4L2_PIX_FMT_Y14, > > + .code = MEDIA_BUS_FMT_Y14_1X14, > > + .depth = 14, > > + .csi_dt = 0x2d, > > + }, > > +}; > > + > > +static const struct unicam_format_info unicam_meta_formats[] = { > > + { > > + .fourcc = V4L2_META_FMT_GENERIC_8, > > + .code = MEDIA_BUS_FMT_META_8, > > + .depth = 8, > > + .metadata_fmt = true, > > + }, { > > + .fourcc = V4L2_META_FMT_GENERIC_CSI2_10, > > + .code = MEDIA_BUS_FMT_META_10, > > + .depth = 10, > > + .metadata_fmt = true, > > + }, { > > + .fourcc = V4L2_META_FMT_GENERIC_CSI2_12, > > + .code = MEDIA_BUS_FMT_META_12, > > + .depth = 12, > > + .metadata_fmt = true, > > + }, { > > + .fourcc = V4L2_META_FMT_GENERIC_CSI2_14, > > + .code = MEDIA_BUS_FMT_META_14, > > + .depth = 14, > > + .metadata_fmt = true, > > + }, > > +}; > > + > > +/* Format setup functions */ > > +static const struct unicam_format_info * > > +unicam_find_format_by_code(u32 code, u32 pad) > > +{ > > + const struct unicam_format_info *formats; > > + unsigned int num_formats; > > + unsigned int i; > > + > > + if (pad == UNICAM_SD_PAD_SOURCE_IMAGE) { > > + formats = unicam_image_formats; > > + num_formats = ARRAY_SIZE(unicam_image_formats); > > + } else { > > + formats = unicam_meta_formats; > > + num_formats = ARRAY_SIZE(unicam_meta_formats); > > + } > > + > > + for (i = 0; i < num_formats; i++) { > > + if (formats[i].code == code) > > + return &formats[i]; > > + } > > + > > + return NULL; > > +} > > + > > +static const struct unicam_format_info * > > +unicam_find_format_by_fourcc(u32 fourcc, u32 pad) > > +{ > > + const struct unicam_format_info *formats; > > + unsigned int num_formats; > > + unsigned int i; > > + > > + if (pad == UNICAM_SD_PAD_SOURCE_IMAGE) { > > + formats = unicam_image_formats; > > + num_formats = ARRAY_SIZE(unicam_image_formats); > > + } else { > > + formats = unicam_meta_formats; > > + num_formats = ARRAY_SIZE(unicam_meta_formats); > > + } > > + > > + for (i = 0; i < num_formats; ++i) { > > + if (formats[i].fourcc == fourcc) > > + return &formats[i]; > > + } > > + > > + return NULL; > > +} > > + > > +static void unicam_calc_image_size_bpl(struct unicam_device *unicam, > > + const struct unicam_format_info *fmtinfo, > > + struct v4l2_pix_format *pix) > > +{ > > + u32 min_bpl; > > + > > + v4l_bound_align_image(&pix->width, UNICAM_IMAGE_MIN_WIDTH, > > + UNICAM_IMAGE_MAX_WIDTH, 2, > > + &pix->height, UNICAM_IMAGE_MIN_HEIGHT, > > + UNICAM_IMAGE_MAX_HEIGHT, 0, 0); > > + > > + /* Unpacking always goes to 16bpp */ > > + if (pix->pixelformat == fmtinfo->unpacked_fourcc) > > + min_bpl = pix->width * 2; > > + else > > + min_bpl = pix->width * fmtinfo->depth / 8; > > + min_bpl = ALIGN(min_bpl, UNICAM_IMAGE_BPL_ALIGNMENT); > > + > > + pix->bytesperline = ALIGN(pix->bytesperline, UNICAM_IMAGE_BPL_ALIGNMENT); > > + pix->bytesperline = clamp_t(unsigned int, pix->bytesperline, min_bpl, > > + UNICAM_IMAGE_MAX_BPL); > > + > > + pix->sizeimage = pix->height * pix->bytesperline; > > +} > > + > > +static void unicam_calc_meta_size_bpl(struct unicam_device *unicam, > > + const struct unicam_format_info *fmtinfo, > > + struct v4l2_meta_format *meta) > > +{ > > + v4l_bound_align_image(&meta->width, UNICAM_META_MIN_WIDTH, > > + UNICAM_META_MAX_WIDTH, 0, > > + &meta->height, UNICAM_META_MIN_HEIGHT, > > + UNICAM_META_MAX_HEIGHT, 0, 0); > > + > > + meta->bytesperline = ALIGN(meta->width * fmtinfo->depth / 8, > > + UNICAM_DMA_BPL_ALIGNMENT); > > + meta->buffersize = meta->height * meta->bytesperline; > > +} > > + > > +/* ----------------------------------------------------------------------------- > > + * Hardware handling > > + */ > > + > > As this has been used for years in BSP driver, I won't go in detail in > the HW handling functions > > > +static inline void unicam_clk_write(struct unicam_device *unicam, u32 val) > > +{ > > + /* Pass the CM_PASSWORD along with the value. */ > > + writel(val | 0x5a000000, unicam->clk_gate_base); > > +} > > + > > +static inline u32 unicam_reg_read(struct unicam_device *unicam, u32 offset) > > +{ > > + return readl(unicam->base + offset); > > +} > > + > > +static inline void unicam_reg_write(struct unicam_device *unicam, u32 offset, u32 val) > > +{ > > + writel(val, unicam->base + offset); > > +} > > + > > +static inline int unicam_get_field(u32 value, u32 mask) > > +{ > > + return (value & mask) >> __ffs(mask); > > +} > > + > > +static inline void unicam_set_field(u32 *valp, u32 field, u32 mask) > > +{ > > + u32 val = *valp; > > + > > + val &= ~mask; > > + val |= (field << __ffs(mask)) & mask; > > + *valp = val; > > +} > > + > > +static inline void unicam_reg_write_field(struct unicam_device *unicam, u32 offset, > > + u32 field, u32 mask) > > +{ > > + u32 val = unicam_reg_read(unicam, offset); > > + > > + unicam_set_field(&val, field, mask); > > + unicam_reg_write(unicam, offset, val); > > +} > > + > > +static void unicam_wr_dma_addr(struct unicam_node *node, > > + struct unicam_buffer *buf) > > +{ > > + dma_addr_t endaddr = buf->dma_addr + buf->size; > > + > > + if (node->id == UNICAM_IMAGE_NODE) { > > + unicam_reg_write(node->dev, UNICAM_IBSA0, buf->dma_addr); > > + unicam_reg_write(node->dev, UNICAM_IBEA0, endaddr); > > + } else { > > + unicam_reg_write(node->dev, UNICAM_DBSA0, buf->dma_addr); > > + unicam_reg_write(node->dev, UNICAM_DBEA0, endaddr); > > + } > > +} > > + > > +static unsigned int unicam_get_lines_done(struct unicam_device *unicam) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_IMAGE_NODE]; > > + unsigned int stride = node->fmt.fmt.pix.bytesperline; > > + struct unicam_buffer *frm = node->cur_frm; > > + dma_addr_t cur_addr; > > + > > + if (!frm) > > + return 0; > > + > > + cur_addr = unicam_reg_read(unicam, UNICAM_IBWP); > > + return (unsigned int)(cur_addr - frm->dma_addr) / stride; > > +} > > + > > +static void unicam_schedule_next_buffer(struct unicam_node *node) > > +{ > > + struct unicam_buffer *buf; > > + > > + buf = list_first_entry(&node->dma_queue, struct unicam_buffer, list); > > + node->next_frm = buf; > > + list_del(&buf->list); > > + > > + unicam_wr_dma_addr(node, buf); > > +} > > + > > +static void unicam_schedule_dummy_buffer(struct unicam_node *node) > > +{ > > + int node_id = is_image_node(node) ? UNICAM_IMAGE_NODE : UNICAM_METADATA_NODE; > > + > > + dev_dbg(node->dev->dev, "Scheduling dummy buffer for node %d\n", node_id); > > + > > + unicam_wr_dma_addr(node, &node->dummy_buf); > > + > > + node->next_frm = NULL; > > +} > > + > > +static void unicam_process_buffer_complete(struct unicam_node *node, > > + unsigned int sequence) > > +{ > > + node->cur_frm->vb.field = node->fmt.fmt.pix.field; > > + node->cur_frm->vb.sequence = sequence; > > + > > + vb2_buffer_done(&node->cur_frm->vb.vb2_buf, VB2_BUF_STATE_DONE); > > +} > > + > > +static void unicam_queue_event_sof(struct unicam_device *unicam) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_IMAGE_NODE]; > > + > > + struct v4l2_event event = { > > + .type = V4L2_EVENT_FRAME_SYNC, > > + .u.frame_sync.frame_sequence = unicam->sequence, > > + }; > > + > > + v4l2_event_queue(&node->video_dev, &event); > > +} > > + > > +static irqreturn_t unicam_isr(int irq, void *dev) > > +{ > > + struct unicam_device *unicam = dev; > > + unsigned int lines_done = unicam_get_lines_done(dev); > > + unsigned int sequence = unicam->sequence; > > + unsigned int i; > > + u32 ista, sta; > > + bool fe; > > + u64 ts; > > + > > + sta = unicam_reg_read(unicam, UNICAM_STA); > > + /* Write value back to clear the interrupts */ > > + unicam_reg_write(unicam, UNICAM_STA, sta); > > + > > + ista = unicam_reg_read(unicam, UNICAM_ISTA); > > + /* Write value back to clear the interrupts */ > > + unicam_reg_write(unicam, UNICAM_ISTA, ista); > > + > > + dev_dbg(unicam->dev, "ISR: ISTA: 0x%X, STA: 0x%X, sequence %d, lines done %d\n", > > + ista, sta, sequence, lines_done); > > + > > + if (!(sta & (UNICAM_IS | UNICAM_PI0))) > > + return IRQ_HANDLED; > > + > > + /* > > + * Look for either the Frame End interrupt or the Packet Capture status > > + * to signal a frame end. > > + */ > > + fe = ista & UNICAM_FEI || sta & UNICAM_PI0; > > + > > + /* > > + * We must run the frame end handler first. If we have a valid next_frm > > + * and we get a simultaneout FE + FS interrupt, running the FS handler > > + * first would null out the next_frm ptr and we would have lost the > > + * buffer forever. > > + */ > > + if (fe) { > > + /* > > + * Ensure we have swapped buffers already as we can't > > + * stop the peripheral. If no buffer is available, use a > > + * dummy buffer to dump out frames until we get a new buffer > > + * to use. > > + */ > > + for (i = 0; i < ARRAY_SIZE(unicam->node); i++) { > > + if (!unicam->node[i].streaming) > > + continue; > > + > > + /* > > + * If cur_frm == next_frm, it means we have not had > > + * a chance to swap buffers, likely due to having > > + * multiple interrupts occurring simultaneously (like FE > > + * + FS + LS). In this case, we cannot signal the buffer > > + * as complete, as the HW will reuse that buffer. > > + */ > > + if (unicam->node[i].cur_frm && > > + unicam->node[i].cur_frm != unicam->node[i].next_frm) > > + unicam_process_buffer_complete(&unicam->node[i], > > + sequence); > > + unicam->node[i].cur_frm = unicam->node[i].next_frm; > > + } > > + unicam->sequence++; > > + } > > + > > + if (ista & UNICAM_FSI) { > > + /* > > + * Timestamp is to be when the first data byte was captured, > > + * aka frame start. > > + */ > > + ts = ktime_get_ns(); > > + for (i = 0; i < ARRAY_SIZE(unicam->node); i++) { > > + if (!unicam->node[i].streaming) > > + continue; > > + > > + if (unicam->node[i].cur_frm) > > + unicam->node[i].cur_frm->vb.vb2_buf.timestamp = > > + ts; > > + else > > + dev_dbg(unicam->v4l2_dev.dev, > > + "ISR: [%d] Dropping frame, buffer not available at FS\n", > > + i); > > + /* > > + * Set the next frame output to go to a dummy frame > > + * if we have not managed to obtain another frame > > + * from the queue. > > + */ > > + unicam_schedule_dummy_buffer(&unicam->node[i]); > > + } > > + > > + unicam_queue_event_sof(unicam); > > + } > > + > > + /* > > + * Cannot swap buffer at frame end, there may be a race condition > > + * where the HW does not actually swap it if the new frame has > > + * already started. > > + */ > > + if (ista & (UNICAM_FSI | UNICAM_LCI) && !fe) { > > + for (i = 0; i < ARRAY_SIZE(unicam->node); i++) { > > + if (!unicam->node[i].streaming) > > + continue; > > + > > + spin_lock(&unicam->node[i].dma_queue_lock); > > + if (!list_empty(&unicam->node[i].dma_queue) && > > + !unicam->node[i].next_frm) > > + unicam_schedule_next_buffer(&unicam->node[i]); > > + spin_unlock(&unicam->node[i].dma_queue_lock); > > + } > > + } > > + > > + if (unicam_reg_read(unicam, UNICAM_ICTL) & UNICAM_FCM) { > > + /* Switch out of trigger mode if selected */ > > + unicam_reg_write_field(unicam, UNICAM_ICTL, 1, UNICAM_TFC); > > + unicam_reg_write_field(unicam, UNICAM_ICTL, 0, UNICAM_FCM); > > + } > > + return IRQ_HANDLED; > > +} > > + > > +static void unicam_set_packing_config(struct unicam_device *unicam) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_IMAGE_NODE]; > > + u32 pack, unpack; > > + u32 val; > > + > > + if (node->fmt.fmt.pix.pixelformat == node->fmtinfo->fourcc) { > > + unpack = UNICAM_PUM_NONE; > > + pack = UNICAM_PPM_NONE; > > + } else { > > + switch (node->fmtinfo->depth) { > > + case 8: > > + unpack = UNICAM_PUM_UNPACK8; > > + break; > > + case 10: > > + unpack = UNICAM_PUM_UNPACK10; > > + break; > > + case 12: > > + unpack = UNICAM_PUM_UNPACK12; > > + break; > > + case 14: > > + unpack = UNICAM_PUM_UNPACK14; > > + break; > > + case 16: > > + unpack = UNICAM_PUM_UNPACK16; > > + break; > > + default: > > + unpack = UNICAM_PUM_NONE; > > + break; > > + } > > + > > + /* Repacking is always to 16bpp */ > > + pack = UNICAM_PPM_PACK16; > > + } > > + > > + val = 0; > > + unicam_set_field(&val, unpack, UNICAM_PUM_MASK); > > + unicam_set_field(&val, pack, UNICAM_PPM_MASK); > > + unicam_reg_write(unicam, UNICAM_IPIPE, val); > > +} > > + > > +static void unicam_cfg_image_id(struct unicam_device *unicam) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_IMAGE_NODE]; > > + > > + if (unicam->bus_type == V4L2_MBUS_CSI2_DPHY) { > > + /* CSI2 mode, hardcode VC 0 for now. */ > > + unicam_reg_write(unicam, UNICAM_IDI0, > > + (0 << 6) | node->fmtinfo->csi_dt); > > + } else { > > + /* CCP2 mode */ > > + unicam_reg_write(unicam, UNICAM_IDI0, > > + 0x80 | node->fmtinfo->csi_dt); > > + } > > +} > > + > > +static void unicam_enable_ed(struct unicam_device *unicam) > > +{ > > + u32 val = unicam_reg_read(unicam, UNICAM_DCS); > > + > > + unicam_set_field(&val, 2, UNICAM_EDL_MASK); > > + /* Do not wrap at the end of the embedded data buffer */ > > + unicam_set_field(&val, 0, UNICAM_DBOB); > > + > > + unicam_reg_write(unicam, UNICAM_DCS, val); > > +} > > + > > +static void unicam_start_rx(struct unicam_device *unicam, > > + struct unicam_buffer *buf) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_IMAGE_NODE]; > > + int line_int_freq = node->fmt.fmt.pix.height >> 2; > > + unsigned int i; > > + u32 val; > > + > > + if (line_int_freq < 128) > > + line_int_freq = 128; > > + > > + /* Enable lane clocks */ > > + val = 1; > > + for (i = 0; i < unicam->active_data_lanes; i++) > > + val = val << 2 | 1; > > + unicam_clk_write(unicam, val); > > + > > + /* Basic init */ > > + unicam_reg_write(unicam, UNICAM_CTRL, UNICAM_MEM); > > + > > + /* Enable analogue control, and leave in reset. */ > > + val = UNICAM_AR; > > + unicam_set_field(&val, 7, UNICAM_CTATADJ_MASK); > > + unicam_set_field(&val, 7, UNICAM_PTATADJ_MASK); > > + unicam_reg_write(unicam, UNICAM_ANA, val); > > + usleep_range(1000, 2000); > > + > > + /* Come out of reset */ > > + unicam_reg_write_field(unicam, UNICAM_ANA, 0, UNICAM_AR); > > + > > + /* Peripheral reset */ > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 1, UNICAM_CPR); > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 0, UNICAM_CPR); > > + > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 0, UNICAM_CPE); > > + > > + /* Enable Rx control. */ > > + val = unicam_reg_read(unicam, UNICAM_CTRL); > > + if (unicam->bus_type == V4L2_MBUS_CSI2_DPHY) { > > + unicam_set_field(&val, UNICAM_CPM_CSI2, UNICAM_CPM_MASK); > > + unicam_set_field(&val, UNICAM_DCM_STROBE, UNICAM_DCM_MASK); > > + } else { > > + unicam_set_field(&val, UNICAM_CPM_CCP2, UNICAM_CPM_MASK); > > + unicam_set_field(&val, unicam->bus_flags, UNICAM_DCM_MASK); > > + } > > + /* Packet framer timeout */ > > + unicam_set_field(&val, 0xf, UNICAM_PFT_MASK); > > + unicam_set_field(&val, 128, UNICAM_OET_MASK); > > + unicam_reg_write(unicam, UNICAM_CTRL, val); > > + > > + unicam_reg_write(unicam, UNICAM_IHWIN, 0); > > + unicam_reg_write(unicam, UNICAM_IVWIN, 0); > > + > > + /* AXI bus access QoS setup */ > > + val = unicam_reg_read(unicam, UNICAM_PRI); > > + unicam_set_field(&val, 0, UNICAM_BL_MASK); > > + unicam_set_field(&val, 0, UNICAM_BS_MASK); > > + unicam_set_field(&val, 0xe, UNICAM_PP_MASK); > > + unicam_set_field(&val, 8, UNICAM_NP_MASK); > > + unicam_set_field(&val, 2, UNICAM_PT_MASK); > > + unicam_set_field(&val, 1, UNICAM_PE); > > + unicam_reg_write(unicam, UNICAM_PRI, val); > > + > > + unicam_reg_write_field(unicam, UNICAM_ANA, 0, UNICAM_DDL); > > + > > + /* Always start in trigger frame capture mode (UNICAM_FCM set) */ > > + val = UNICAM_FSIE | UNICAM_FEIE | UNICAM_FCM | UNICAM_IBOB; > > + unicam_set_field(&val, line_int_freq, UNICAM_LCIE_MASK); > > + unicam_reg_write(unicam, UNICAM_ICTL, val); > > + unicam_reg_write(unicam, UNICAM_STA, UNICAM_STA_MASK_ALL); > > + unicam_reg_write(unicam, UNICAM_ISTA, UNICAM_ISTA_MASK_ALL); > > + > > + /* tclk_term_en */ > > + unicam_reg_write_field(unicam, UNICAM_CLT, 2, UNICAM_CLT1_MASK); > > + /* tclk_settle */ > > + unicam_reg_write_field(unicam, UNICAM_CLT, 6, UNICAM_CLT2_MASK); > > + /* td_term_en */ > > + unicam_reg_write_field(unicam, UNICAM_DLT, 2, UNICAM_DLT1_MASK); > > + /* ths_settle */ > > + unicam_reg_write_field(unicam, UNICAM_DLT, 6, UNICAM_DLT2_MASK); > > + /* trx_enable */ > > + unicam_reg_write_field(unicam, UNICAM_DLT, 0, UNICAM_DLT3_MASK); > > + > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 0, UNICAM_SOE); > > + > > + /* Packet compare setup - required to avoid missing frame ends */ > > + val = 0; > > + unicam_set_field(&val, 1, UNICAM_PCE); > > + unicam_set_field(&val, 1, UNICAM_GI); > > + unicam_set_field(&val, 1, UNICAM_CPH); > > + unicam_set_field(&val, 0, UNICAM_PCVC_MASK); > > + unicam_set_field(&val, 1, UNICAM_PCDT_MASK); > > + unicam_reg_write(unicam, UNICAM_CMP0, val); > > + > > + /* Enable clock lane and set up terminations */ > > + val = 0; > > + if (unicam->bus_type == V4L2_MBUS_CSI2_DPHY) { > > + /* CSI2 */ > > + unicam_set_field(&val, 1, UNICAM_CLE); > > + unicam_set_field(&val, 1, UNICAM_CLLPE); > > + if (!(unicam->bus_flags & V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK)) { > > + unicam_set_field(&val, 1, UNICAM_CLTRE); > > + unicam_set_field(&val, 1, UNICAM_CLHSE); > > + } > > + } else { > > + /* CCP2 */ > > + unicam_set_field(&val, 1, UNICAM_CLE); > > + unicam_set_field(&val, 1, UNICAM_CLHSE); > > + unicam_set_field(&val, 1, UNICAM_CLTRE); > > + } > > + unicam_reg_write(unicam, UNICAM_CLK, val); > > + > > + /* > > + * Enable required data lanes with appropriate terminations. > > + * The same value needs to be written to UNICAM_DATn registers for > > + * the active lanes, and 0 for inactive ones. > > + */ > > + val = 0; > > + if (unicam->bus_type == V4L2_MBUS_CSI2_DPHY) { > > + /* CSI2 */ > > + unicam_set_field(&val, 1, UNICAM_DLE); > > + unicam_set_field(&val, 1, UNICAM_DLLPE); > > + if (!(unicam->bus_flags & V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK)) { > > + unicam_set_field(&val, 1, UNICAM_DLTRE); > > + unicam_set_field(&val, 1, UNICAM_DLHSE); > > + } > > + } else { > > + /* CCP2 */ > > + unicam_set_field(&val, 1, UNICAM_DLE); > > + unicam_set_field(&val, 1, UNICAM_DLHSE); > > + unicam_set_field(&val, 1, UNICAM_DLTRE); > > + } > > + unicam_reg_write(unicam, UNICAM_DAT0, val); > > + > > + if (unicam->active_data_lanes == 1) > > + val = 0; > > + unicam_reg_write(unicam, UNICAM_DAT1, val); > > + > > + if (unicam->max_data_lanes > 2) { > > + /* > > + * Registers UNICAM_DAT2 and UNICAM_DAT3 only valid if the > > + * instance supports more than 2 data lanes. > > + */ > > + if (unicam->active_data_lanes == 2) > > + val = 0; > > + unicam_reg_write(unicam, UNICAM_DAT2, val); > > + > > + if (unicam->active_data_lanes == 3) > > + val = 0; > > + unicam_reg_write(unicam, UNICAM_DAT3, val); > > + } > > + > > + unicam_reg_write(unicam, UNICAM_IBLS, > > + node->fmt.fmt.pix.bytesperline); > > + unicam_wr_dma_addr(&unicam->node[UNICAM_IMAGE_NODE], buf); > > + unicam_set_packing_config(unicam); > > + unicam_cfg_image_id(unicam); > > + > > + val = unicam_reg_read(unicam, UNICAM_MISC); > > + unicam_set_field(&val, 1, UNICAM_FL0); > > + unicam_set_field(&val, 1, UNICAM_FL1); > > + unicam_reg_write(unicam, UNICAM_MISC, val); > > + > > + /* Enable peripheral */ > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 1, UNICAM_CPE); > > + > > + /* Load image pointers */ > > + unicam_reg_write_field(unicam, UNICAM_ICTL, 1, UNICAM_LIP_MASK); > > + > > + /* > > + * Enable trigger only for the first frame to > > + * sync correctly to the FS from the source. > > + */ > > + unicam_reg_write_field(unicam, UNICAM_ICTL, 1, UNICAM_TFC); > > +} > > + > > +static void unicam_start_metadata(struct unicam_device *unicam, > > + struct unicam_buffer *buf) > > +{ > > + struct unicam_node *node = &unicam->node[UNICAM_METADATA_NODE]; > > + > > + unicam_enable_ed(unicam); > > + unicam_wr_dma_addr(node, buf); > > + unicam_reg_write_field(unicam, UNICAM_DCS, 1, UNICAM_LDP); > > +} > > + > > +static void unicam_disable(struct unicam_device *unicam) > > +{ > > + /* Analogue lane control disable */ > > + unicam_reg_write_field(unicam, UNICAM_ANA, 1, UNICAM_DDL); > > + > > + /* Stop the output engine */ > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 1, UNICAM_SOE); > > + > > + /* Disable the data lanes. */ > > + unicam_reg_write(unicam, UNICAM_DAT0, 0); > > + unicam_reg_write(unicam, UNICAM_DAT1, 0); > > + > > + if (unicam->max_data_lanes > 2) { > > + unicam_reg_write(unicam, UNICAM_DAT2, 0); > > + unicam_reg_write(unicam, UNICAM_DAT3, 0); > > + } > > + > > + /* Peripheral reset */ > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 1, UNICAM_CPR); > > + usleep_range(50, 100); > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 0, UNICAM_CPR); > > + > > + /* Disable peripheral */ > > + unicam_reg_write_field(unicam, UNICAM_CTRL, 0, UNICAM_CPE); > > + > > + /* Clear ED setup */ > > + unicam_reg_write(unicam, UNICAM_DCS, 0); > > + > > + /* Disable all lane clocks */ > > + unicam_clk_write(unicam, 0); > > +} > > + > > +/* ----------------------------------------------------------------------------- > > + * V4L2 subdev operations > > + */ > > + > > +static int __unicam_subdev_set_routing(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, > > + struct v4l2_subdev_krouting *routing) > > +{ > > + struct v4l2_subdev_route *route; > > + int ret; > > + > > + ret = v4l2_subdev_routing_validate(sd, routing, > > + V4L2_SUBDEV_ROUTING_ONLY_1_TO_1); > > + if (ret) > > + return ret; > > + > > + ret = v4l2_subdev_set_routing(sd, state, routing); > > + if (ret) > > + return ret; > > + > > + for_each_active_route(&state->routing, route) { > > + const struct v4l2_mbus_framefmt *def_fmt; > > + struct v4l2_mbus_framefmt *fmt; > > + > > + if (route->source_pad == UNICAM_SD_PAD_SOURCE_IMAGE) > > + def_fmt = &unicam_default_image_format; > > + else > > + def_fmt = &unicam_default_meta_format; > > + > > + fmt = v4l2_subdev_state_get_format(state, route->sink_pad, > > + route->sink_stream); > > + *fmt = *def_fmt; > > + fmt = v4l2_subdev_state_get_format(state, route->source_pad, > > + route->source_stream); > > + *fmt = *def_fmt; > > + } > > + > > + return 0; > > +} > > + > > +static int unicam_subdev_init_state(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state) > > +{ > > + struct v4l2_subdev_route routes[] = { > > + { > > + .sink_pad = UNICAM_SD_PAD_SINK, > > + .sink_stream = 0, > > + .source_pad = UNICAM_SD_PAD_SOURCE_IMAGE, > > + .source_stream = 0, > > + .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE, > > + }, > > + }; > > + > > + struct v4l2_subdev_krouting routing = { > > + .len_routes = ARRAY_SIZE(routes), > > + .num_routes = ARRAY_SIZE(routes), > > + .routes = routes, > > + }; > > + > > + /* Initialize routing to single route to the fist source pad. */ > > + return __unicam_subdev_set_routing(sd, state, &routing); > > +} > > + > > +static int unicam_subdev_enum_mbus_code(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, > > + struct v4l2_subdev_mbus_code_enum *code) > > +{ > > + u32 pad, stream; > > + int ret; > > + > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, > > + code->pad, code->stream, > > + &pad, &stream); > > + if (ret) > > + return ret; > > + > > + if (unicam_sd_pad_is_source(code->pad)) { > > + /* No transcoding, source and sink codes must match. */ > > + const struct v4l2_mbus_framefmt *fmt; > > + > > + fmt = v4l2_subdev_state_get_format(state, pad, stream); > > + if (!fmt) > > + return -EINVAL; > > + > > + if (code->index > 0) > > + return -EINVAL; > > + > > + code->code = fmt->code; > > + } else { > > + const struct unicam_format_info *formats; > > + unsigned int num_formats; > > + > > + if (pad == UNICAM_SD_PAD_SOURCE_IMAGE) { > > + formats = unicam_image_formats; > > + num_formats = ARRAY_SIZE(unicam_image_formats); > > + } else { > > + formats = unicam_meta_formats; > > + num_formats = ARRAY_SIZE(unicam_meta_formats); > > + } > > + > > + if (code->index >= num_formats) > > + return -EINVAL; > > + > > + code->code = formats[code->index].code; > > + } > > + > > + return 0; > > +} > > + > > +static int unicam_subdev_enum_frame_size(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, > > + struct v4l2_subdev_frame_size_enum *fse) > > +{ > > + u32 pad, stream; > > + int ret; > > + > > + if (fse->index > 0) > > + return -EINVAL; > > + > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, fse->pad, > > + fse->stream, &pad, > > + &stream); > > + if (ret) > > + return ret; > > + > > + if (unicam_sd_pad_is_source(fse->pad)) { > > + /* No transcoding, source and sink formats must match. */ > > + const struct v4l2_mbus_framefmt *fmt; > > + > > + fmt = v4l2_subdev_state_get_format(state, pad, stream); > > + if (!fmt) > > + return -EINVAL; > > + > > + if (fse->code != fmt->code) > > + return -EINVAL; > > + > > + fse->min_width = fmt->width; > > + fse->max_width = fmt->width; > > + fse->min_height = fmt->height; > > + fse->max_height = fmt->height; > > + } else { > > + const struct unicam_format_info *fmtinfo; > > + > > + fmtinfo = unicam_find_format_by_code(fse->code, pad); > > + if (!fmtinfo) > > + return -EINVAL; > > + > > + if (pad == UNICAM_SD_PAD_SOURCE_IMAGE) { > > + fse->min_width = UNICAM_IMAGE_MIN_WIDTH; > > + fse->max_width = UNICAM_IMAGE_MAX_WIDTH; > > + fse->min_height = UNICAM_IMAGE_MIN_HEIGHT; > > + fse->max_height = UNICAM_IMAGE_MAX_HEIGHT; > > + } else { > > + fse->min_width = UNICAM_META_MIN_WIDTH; > > + fse->max_width = UNICAM_META_MAX_WIDTH; > > + fse->min_height = UNICAM_META_MIN_HEIGHT; > > + fse->max_height = UNICAM_META_MAX_HEIGHT; > > + } > > + } > > + > > + return 0; > > +} > > + > > +static int unicam_subdev_set_format(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, > > + struct v4l2_subdev_format *format) > > +{ > > + struct unicam_device *unicam = sd_to_unicam_device(sd); > > + struct v4l2_mbus_framefmt *sink_format, *source_format; > > + const struct unicam_format_info *fmtinfo; > > + u32 source_pad, source_stream; > > + int ret; > > + > > + if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && > > + unicam->subdev.streaming) > > + return -EBUSY; > > + > > + /* No transcoding, source and sink formats must match. */ > > + if (unicam_sd_pad_is_source(format->pad)) > > + return v4l2_subdev_get_fmt(sd, state, format); > > + > > + /* > > + * Allowed formats for the stream on the sink pad depend on what source > > + * pad the stream is routed to. Find the corresponding source pad and > > + * use it to validate the media bus code. > > + */ > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, > > + format->pad, format->stream, > > + &source_pad, &source_stream); > > + if (ret) > > + return ret; > > + > > + fmtinfo = unicam_find_format_by_code(format->format.code, source_pad); > > + if (!fmtinfo) { > > + fmtinfo = source_pad == UNICAM_SD_PAD_SOURCE_IMAGE > > + ? &unicam_image_formats[0] : &unicam_meta_formats[0]; > > + format->format.code = fmtinfo->code; > > + } > > + > > + if (source_pad == UNICAM_SD_PAD_SOURCE_IMAGE) { > > + format->format.width = clamp_t(unsigned int, > > + format->format.width, > > + UNICAM_IMAGE_MIN_WIDTH, > > + UNICAM_IMAGE_MAX_WIDTH); > > + format->format.height = clamp_t(unsigned int, > > + format->format.height, > > + UNICAM_IMAGE_MIN_HEIGHT, > > + UNICAM_IMAGE_MAX_HEIGHT); > > + format->format.field = V4L2_FIELD_NONE; > > + } else { > > + format->format.width = clamp_t(unsigned int, > > + format->format.width, > > + UNICAM_META_MIN_WIDTH, > > + UNICAM_META_MAX_WIDTH); > > + format->format.height = clamp_t(unsigned int, > > + format->format.height, > > + UNICAM_META_MIN_HEIGHT, > > + UNICAM_META_MAX_HEIGHT); > > + format->format.field = V4L2_FIELD_NONE; > > + > > + /* Colorspace don't apply to metadata. */ > > + format->format.colorspace = 0; > > + format->format.ycbcr_enc = 0; > > + format->format.quantization = 0; > > + format->format.xfer_func = 0; > > + } > > + > > + sink_format = v4l2_subdev_state_get_format(state, format->pad, > > + format->stream); > > + source_format = v4l2_subdev_state_get_format(state, source_pad, > > + source_stream); > > + *sink_format = format->format; > > + *source_format = format->format; > > + > > + return 0; > > +} > > + > > +static int unicam_subdev_set_routing(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, > > + enum v4l2_subdev_format_whence which, > > + struct v4l2_subdev_krouting *routing) > > +{ > > + struct unicam_device *unicam = sd_to_unicam_device(sd); > > + > > + if (which == V4L2_SUBDEV_FORMAT_ACTIVE && unicam->subdev.streaming) > > + return -EBUSY; > > + > > + return __unicam_subdev_set_routing(sd, state, routing); > > +} > > + > > +static int unicam_sd_enable_streams(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, u32 pad, > > + u64 streams_mask) > > +{ > > + struct unicam_device *unicam = sd_to_unicam_device(sd); > > + u32 other_pad, other_stream; > > + int ret; > > + > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, pad, 0, > > + &other_pad, &other_stream); > > + if (ret) > > + return ret; > > + > > + unicam->sequence = 0; > > + > > + ret = v4l2_subdev_enable_streams(unicam->sensor.subdev, > > + unicam->sensor.pad->index, > > + BIT(other_stream)); > > + if (ret) { > > + dev_err(unicam->dev, "stream on failed in subdev\n"); > > + return ret; > > + } > > + > > + unicam->subdev.streaming = true; > > + > > + return 0; > > +} > > + > > +static int unicam_sd_disable_streams(struct v4l2_subdev *sd, > > + struct v4l2_subdev_state *state, u32 pad, > > + u64 streams_mask) > > +{ > > + struct unicam_device *unicam = sd_to_unicam_device(sd); > > + u32 other_pad, other_stream; > > + int ret; > > + > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, pad, 0, > > + &other_pad, &other_stream); > > + if (ret) > > + return ret; > > + > > + v4l2_subdev_disable_streams(unicam->sensor.subdev, > > + unicam->sensor.pad->index, > > + BIT(other_stream)); > > + > > + unicam->subdev.streaming = false; > > + > > + return 0; > > +} > > + > > +static const struct v4l2_subdev_pad_ops unicam_subdev_pad_ops = { > > + .enum_mbus_code = unicam_subdev_enum_mbus_code, > > + .enum_frame_size = unicam_subdev_enum_frame_size, > > + .get_fmt = v4l2_subdev_get_fmt, > > + .set_fmt = unicam_subdev_set_format, > > + .set_routing = unicam_subdev_set_routing, > > + .enable_streams = unicam_sd_enable_streams, > > + .disable_streams = unicam_sd_disable_streams, > > +}; > > + > > +static const struct v4l2_subdev_ops unicam_subdev_ops = { > > + .pad = &unicam_subdev_pad_ops, > > +}; > > + > > +static const struct v4l2_subdev_internal_ops unicam_subdev_internal_ops = { > > + .init_state = unicam_subdev_init_state, > > +}; > > + > > +static const struct media_entity_operations unicam_subdev_media_ops = { > > + .link_validate = v4l2_subdev_link_validate, > > + .has_pad_interdep = v4l2_subdev_has_pad_interdep, > > +}; > > + > > +static int unicam_subdev_init(struct unicam_device *unicam) > > +{ > > + struct v4l2_subdev *sd = &unicam->subdev.sd; > > + int ret; > > + > > + v4l2_subdev_init(sd, &unicam_subdev_ops); > > + sd->internal_ops = &unicam_subdev_internal_ops; > > + v4l2_set_subdevdata(sd, unicam); > > + > > + sd->entity.function = MEDIA_ENT_F_VID_IF_BRIDGE; > > + sd->entity.ops = &unicam_subdev_media_ops; > > + sd->dev = unicam->dev; > > + sd->owner = THIS_MODULE; > > + sd->flags = V4L2_SUBDEV_FL_HAS_DEVNODE | V4L2_SUBDEV_FL_STREAMS; > > + > > + strscpy(sd->name, "unicam", sizeof(sd->name)); > > + > > + unicam->subdev.pads[UNICAM_SD_PAD_SINK].flags = MEDIA_PAD_FL_SINK; > > + unicam->subdev.pads[UNICAM_SD_PAD_SOURCE_IMAGE].flags = MEDIA_PAD_FL_SOURCE; > > + unicam->subdev.pads[UNICAM_SD_PAD_SOURCE_METADATA].flags = MEDIA_PAD_FL_SOURCE; > > + > > + ret = media_entity_pads_init(&sd->entity, ARRAY_SIZE(unicam->subdev.pads), > > + unicam->subdev.pads); > > + if (ret) { > > + dev_err(unicam->dev, "Failed to initialize media entity: %d\n", > > + ret); > > + return ret; > > + } > > + > > + ret = v4l2_subdev_init_finalize(sd); > > + if (ret) { > > + dev_err(unicam->dev, "Failed to initialize subdev: %d\n", ret); > > + goto err_entity; > > + } > > + > > + ret = v4l2_device_register_subdev(&unicam->v4l2_dev, sd); > > + if (ret) { > > + dev_err(unicam->dev, "Failed to register subdev: %d\n", ret); > > + goto err_subdev; > > + } > > + > > + return 0; > > + > > +err_subdev: > > + v4l2_subdev_cleanup(sd); > > +err_entity: > > + media_entity_cleanup(&sd->entity); > > + return ret; > > +} > > + > > +static void unicam_subdev_cleanup(struct unicam_device *unicam) > > +{ > > + v4l2_subdev_cleanup(&unicam->subdev.sd); > > + media_entity_cleanup(&unicam->subdev.sd.entity); > > +} > > + > > +/* ----------------------------------------------------------------------------- > > + * Videobuf2 queue operations > > + */ > > + > > +static int unicam_queue_setup(struct vb2_queue *vq, > > + unsigned int *nbuffers, > > + unsigned int *nplanes, > > + unsigned int sizes[], > > + struct device *alloc_devs[]) > > +{ > > + struct unicam_node *node = vb2_get_drv_priv(vq); > > + u32 size = is_image_node(node) ? node->fmt.fmt.pix.sizeimage > > + : node->fmt.fmt.meta.buffersize; > > + > > + if (vq->num_buffers + *nbuffers < 3) > > Why 3 ? Are these dummies ? > > > + *nbuffers = 3 - vq->num_buffers; > > + > > + if (*nplanes) { > > + if (sizes[0] < size) { > > + dev_dbg(node->dev->dev, "sizes[0] %i < size %u\n", > > + sizes[0], size); > > + return -EINVAL; > > + } > > + size = sizes[0]; > > + } > > + > > + *nplanes = 1; > > + sizes[0] = size; > > + > > + return 0; > > +} > > + > > +static int unicam_buffer_prepare(struct vb2_buffer *vb) > > +{ > > + struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue); > > + struct unicam_buffer *buf = to_unicam_buffer(vb); > > + u32 size = is_image_node(node) ? node->fmt.fmt.pix.sizeimage > > + : node->fmt.fmt.meta.buffersize; > > + > > + if (vb2_plane_size(vb, 0) < size) { > > + dev_dbg(node->dev->dev, > > + "data will not fit into plane (%lu < %u)\n", > > + vb2_plane_size(vb, 0), size); > > + return -EINVAL; > > + } > > + > > + buf->dma_addr = vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0); > > + buf->size = size; > > + > > + vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size); > > + > > + return 0; > > +} > > + > > +static void unicam_return_buffers(struct unicam_node *node, > > + enum vb2_buffer_state state) > > +{ > > + struct unicam_buffer *buf, *tmp; > > + > > + list_for_each_entry_safe(buf, tmp, &node->dma_queue, list) { > > + list_del(&buf->list); > > + vb2_buffer_done(&buf->vb.vb2_buf, state); > > + } > > + > > + if (node->cur_frm) > > + vb2_buffer_done(&node->cur_frm->vb.vb2_buf, > > + state); > > + if (node->next_frm && node->cur_frm != node->next_frm) > > + vb2_buffer_done(&node->next_frm->vb.vb2_buf, > > + state); > > + > > + node->cur_frm = NULL; > > + node->next_frm = NULL; > > +} > > + > > +static int unicam_start_streaming(struct vb2_queue *vq, unsigned int count) > > +{ > > + struct unicam_node *node = vb2_get_drv_priv(vq); > > + struct unicam_device *unicam = node->dev; > > + struct v4l2_subdev_state *state; > > + struct unicam_buffer *buf; > > + unsigned long flags; > > + int ret; > > + u32 pad, stream; > > + u32 remote_pad = is_image_node(node) ? UNICAM_SD_PAD_SOURCE_IMAGE > > + : UNICAM_SD_PAD_SOURCE_METADATA; > > + > > + /* Look for the route for the given pad and stream. */ > > + state = v4l2_subdev_lock_and_get_active_state(&unicam->subdev.sd); > > + ret = v4l2_subdev_routing_find_opposite_end(&state->routing, > > + remote_pad, 0, > > + &pad, &stream); > > + v4l2_subdev_unlock_state(state); > > + > > + if (ret) > > + goto err_return_buffers; You've dropped calling get_mbus_config from in or around this area, therefore any device that supports dynamic switching on number of data lanes (eg tc358743) is broken. Docs guidance is "Callers should make sure they get the most up-to-date as possible configuration from the remote end, likely calling this operation as close as possible to stream on time" https://www.kernel.org/doc/html/latest/driver-api/media/v4l2-subdev.html?highlight=get_mbus_config > > + > > + dev_dbg(unicam->dev, "Starting stream on %s: %u/%u -> %u/%u (%s)\n", > > + unicam->subdev.sd.name, pad, stream, remote_pad, 0, > > + is_metadata_node(node) ? "metadata" : "image"); > > + > > + /* The metadata node can't be started alone. */ > > + if (is_metadata_node(node)) { > > + if (!unicam->node[UNICAM_IMAGE_NODE].streaming) { > > Does it mean the metadata node has to be started second, or should > this be made a nop and the metadata node gets started once the image > node is started too ? I'm fine with the constraint of having the > metadata node being started second fwiw This has been flipped from the downstream implementation. Downstream we expected the metadata node to be enabled first, but it wouldn't get started until the image node was enabled. Things like sequence numbers come to mind now that this is enabling the image node first. If the image node has produced the first frame before the metadata node has been enabled, then as I read it the sequence numbers for the two nodes should be different - there haven't really been any dropped frames as you hadn't enabled the node. I'd also want to confirm what the behaviour is if the frame has already started before the metadata node is enabled. Embedded data is almost always sent before the image data, so have we potentially missed it for the current frame, but at the end of frame are going to return both buffers? > > + dev_err(unicam->dev, > > + "Can't start metadata without image\n"); > > + ret = -EINVAL; > > + goto err_return_buffers; > > + } > > + > > + spin_lock_irqsave(&node->dma_queue_lock, flags); > > + buf = list_first_entry(&node->dma_queue, > > + struct unicam_buffer, list); > > + dev_dbg(unicam->dev, "buffer %p\n", buf); > > Is this useful ? > > > + node->cur_frm = buf; > > + node->next_frm = buf; > > + list_del(&buf->list); > > + spin_unlock_irqrestore(&node->dma_queue_lock, flags); > > + > > + unicam_start_metadata(unicam, buf); > > + node->streaming = true; > > + return 0; > > + } > > + > > + ret = pm_runtime_resume_and_get(unicam->dev); > > + if (ret < 0) { > > + dev_err(unicam->dev, "PM runtime resume failed: %d\n", ret); > > + goto err_return_buffers; > > + } > > + > > + ret = video_device_pipeline_start(&node->video_dev, &unicam->pipe); > > + if (ret < 0) { > > + dev_dbg(unicam->dev, "Failed to start media pipeline: %d\n", ret); > > Isn't this an err ? > > > + goto err_pm_put; > > + } > > + > > + spin_lock_irqsave(&node->dma_queue_lock, flags); > > + buf = list_first_entry(&node->dma_queue, > > + struct unicam_buffer, list); > > + dev_dbg(unicam->dev, "buffer %p\n", buf); > > + node->cur_frm = buf; > > + node->next_frm = buf; > > + list_del(&buf->list); > > + spin_unlock_irqrestore(&node->dma_queue_lock, flags); > > + > > + unicam_start_rx(unicam, buf); > > + > > + ret = v4l2_subdev_enable_streams(&unicam->subdev.sd, remote_pad, BIT(0)); > > A bit confused by the API here, shouldn't we also do this for embedded > data ? > > > + if (ret < 0) { > > + dev_err(unicam->dev, "stream on failed in subdev\n"); > > + goto error_pipeline; > > + } > > + > > + node->streaming = true; > > + > > + return 0; > > + > > +error_pipeline: > > + video_device_pipeline_stop(&node->video_dev); > > +err_pm_put: > > + pm_runtime_put_sync(unicam->dev); > > +err_return_buffers: > > + unicam_return_buffers(node, VB2_BUF_STATE_QUEUED); > > + return ret; > > +} > > + > > +static void unicam_stop_streaming(struct vb2_queue *vq) > > +{ > > + struct unicam_node *node = vb2_get_drv_priv(vq); > > + struct unicam_device *unicam = node->dev; > > + u32 remote_pad_index = is_image_node(node) ? UNICAM_SD_PAD_SOURCE_IMAGE > > + : UNICAM_SD_PAD_SOURCE_METADATA; > > + > > + node->streaming = false; > > + > > + v4l2_subdev_disable_streams(&unicam->subdev.sd, remote_pad_index, > > + BIT(0)); > > + > > + /* We can stream only with the image node. */ > > + if (is_metadata_node(node)) { > > + /* > > + * Allow the hardware to spin in the dummy buffer. > > + * This is only really needed if the embedded data pad is > > + * disabled before the image pad. > > + */ > > + unicam_wr_dma_addr(node, &node->dummy_buf); > > + goto dequeue_buffers; > > + } > > + > > + unicam_disable(unicam); > > + > > + video_device_pipeline_stop(&node->video_dev); > > + pm_runtime_put(unicam->dev); > > + > > +dequeue_buffers: > > + /* Clear all queued buffers for the node */ > > + unicam_return_buffers(node, VB2_BUF_STATE_ERROR); > > +} > > + > > +static void unicam_buffer_queue(struct vb2_buffer *vb) > > +{ > > + struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue); > > + struct unicam_buffer *buf = to_unicam_buffer(vb); > > + > > + spin_lock_irq(&node->dma_queue_lock); > > + list_add_tail(&buf->list, &node->dma_queue); > > + spin_unlock_irq(&node->dma_queue_lock); > > +} > > + > > +static const struct vb2_ops unicam_video_qops = { > > + .queue_setup = unicam_queue_setup, > > + .wait_prepare = vb2_ops_wait_prepare, > > + .wait_finish = vb2_ops_wait_finish, > > + .buf_prepare = unicam_buffer_prepare, > > + .start_streaming = unicam_start_streaming, > > + .stop_streaming = unicam_stop_streaming, > > + .buf_queue = unicam_buffer_queue, > > +}; > > + > > +/* ----------------------------------------------------------------------------- > > + * V4L2 video device operations > > + */ > > + > > +static int unicam_querycap(struct file *file, void *priv, > > + struct v4l2_capability *cap) > > +{ > > + strscpy(cap->driver, UNICAM_MODULE_NAME, sizeof(cap->driver)); > > + strscpy(cap->card, UNICAM_MODULE_NAME, sizeof(cap->card)); > > + > > + cap->capabilities |= V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_META_CAPTURE; > > + > > + return 0; > > +} > > + > > +static int unicam_enum_fmt_vid(struct file *file, void *priv, > > + struct v4l2_fmtdesc *f) > > +{ > > + unsigned int index; > > + unsigned int i; > > + > > + for (i = 0, index = 0; i < ARRAY_SIZE(unicam_image_formats); i++) { > > + if (f->mbus_code && unicam_image_formats[i].code != f->mbus_code) > > + continue; > > + > > + if (index == f->index) { > > + f->pixelformat = unicam_image_formats[i].fourcc; > > + return 0; > > + } > > + > > + index++; > > + > > + if (!unicam_image_formats[i].unpacked_fourcc) > > + continue; > > + > > + if (index == f->index) { > > + f->pixelformat = unicam_image_formats[i].unpacked_fourcc; > > + return 0; > > + } > > + > > + index++; > > + } > > + > > + return -EINVAL; > > +} > > + > > +static int unicam_g_fmt_vid(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + *f = node->fmt; > > + > > + return 0; > > +} > > + > > +static const struct unicam_format_info * > > +__unicam_try_fmt_vid(struct unicam_node *node, struct v4l2_pix_format *pix) > > +{ > > + const struct unicam_format_info *fmtinfo; > > + > > + /* > > + * Default to the first format if the requested pixel format code isn't > > + * supported. > > + */ > > + fmtinfo = unicam_find_format_by_fourcc(pix->pixelformat, > > + UNICAM_SD_PAD_SOURCE_IMAGE); > > + if (!fmtinfo) { > > + fmtinfo = &unicam_image_formats[0]; > > + pix->pixelformat = fmtinfo->fourcc; > > + } > > + > > + unicam_calc_image_size_bpl(node->dev, fmtinfo, pix); > > + > > + if (pix->field == V4L2_FIELD_ANY) > > + pix->field = V4L2_FIELD_NONE; > > + > > + return fmtinfo; > > +} > > + > > +static int unicam_try_fmt_vid(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + __unicam_try_fmt_vid(node, &f->fmt.pix); > > + return 0; > > +} > > + > > +static int unicam_s_fmt_vid(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + if (vb2_is_busy(&node->buffer_queue)) > > + return -EBUSY; > > + > > + node->fmtinfo = __unicam_try_fmt_vid(node, &f->fmt.pix); > > + node->fmt = *f; > > + > > + return 0; > > +} > > + > > +static int unicam_enum_fmt_meta(struct file *file, void *priv, > > + struct v4l2_fmtdesc *f) > > +{ > > + unsigned int i, index; > > + > > + for (i = 0, index = 0; i < ARRAY_SIZE(unicam_meta_formats); i++) { > > + if (f->mbus_code && unicam_meta_formats[i].code != f->mbus_code) > > Do we want to allow mbus_code filtering for metadata ? There's a > 1-to-1 relationship between mbus codes and pixel formats > > > + continue; > > + if (!unicam_meta_formats[i].metadata_fmt) > > + continue; > > How can this be false if we're iterating on unicam_meta_formats[] ? > > > + > > + if (index == f->index) { > > + f->pixelformat = unicam_meta_formats[i].fourcc; > > + f->type = V4L2_BUF_TYPE_META_CAPTURE; > > + return 0; > > + } > > + index++; > > + } > > + > > + return -EINVAL; > > +} > > + > > +static int unicam_g_fmt_meta(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + f->fmt.meta = node->fmt.fmt.meta; > > + > > + return 0; > > +} > > + > > +static const struct unicam_format_info * > > +__unicam_try_fmt_meta(struct unicam_node *node, struct v4l2_meta_format *meta) > > +{ > > + const struct unicam_format_info *fmtinfo; > > + > > + /* > > + * Default to the first format if the requested pixel format code isn't > > + * supported. > > + */ > > + fmtinfo = unicam_find_format_by_fourcc(meta->dataformat, > > + UNICAM_SD_PAD_SOURCE_METADATA); > > + if (!fmtinfo) { > > + fmtinfo = &unicam_meta_formats[0]; > > + meta->dataformat = fmtinfo->fourcc; > > + } > > + > > + unicam_calc_meta_size_bpl(node->dev, fmtinfo, meta); > > + > > + return fmtinfo; > > +} > > + > > +static int unicam_try_fmt_meta(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + __unicam_try_fmt_vid(node, &f->fmt.pix); > > + return 0; > > +} > > + > > +static int unicam_s_fmt_meta(struct file *file, void *priv, > > + struct v4l2_format *f) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + > > + if (vb2_is_busy(&node->buffer_queue)) > > + return -EBUSY; > > + > > + node->fmtinfo = __unicam_try_fmt_meta(node, &f->fmt.meta); > > + node->fmt = *f; > > + > > + return 0; > > +} > > + > > +static int unicam_enum_framesizes(struct file *file, void *fh, > > + struct v4l2_frmsizeenum *fsize) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + int ret = -EINVAL; > > + > > + if (fsize->index > 0) > > + return ret; > > + > > + if (is_image_node(node)) { > > + if (!unicam_find_format_by_fourcc(fsize->pixel_format, > > + UNICAM_SD_PAD_SOURCE_IMAGE)) > > + return ret; > > + > > + fsize->type = V4L2_FRMSIZE_TYPE_STEPWISE; > > + fsize->stepwise.min_width = UNICAM_IMAGE_MIN_WIDTH; > > + fsize->stepwise.max_width = UNICAM_IMAGE_MAX_WIDTH; > > + fsize->stepwise.step_width = 1; > > + fsize->stepwise.min_height = UNICAM_IMAGE_MIN_HEIGHT; > > + fsize->stepwise.max_height = UNICAM_IMAGE_MAX_HEIGHT; > > + fsize->stepwise.step_height = 1; > > + } else { > > + if (!unicam_find_format_by_fourcc(fsize->pixel_format, > > + UNICAM_SD_PAD_SOURCE_METADATA)) > > + return ret; > > + > Isn't this V4L2_FRMSIZE_TYPE_STEPWISE as well ? > > > + fsize->stepwise.min_width = UNICAM_META_MIN_WIDTH; > > + fsize->stepwise.max_width = UNICAM_META_MAX_WIDTH; > > + fsize->stepwise.step_width = 1; > > + fsize->stepwise.min_height = UNICAM_META_MIN_HEIGHT; > > + fsize->stepwise.max_height = UNICAM_META_MAX_HEIGHT; > > + fsize->stepwise.step_height = 1; > > + } > > + > > + return 0; > > +} > > + > > +static int unicam_log_status(struct file *file, void *fh) > > +{ > > + struct unicam_node *node = video_drvdata(file); > > + struct unicam_device *unicam = node->dev; > > + u32 reg; > > + > > + /* status for sub devices */ > > + v4l2_device_call_all(&unicam->v4l2_dev, 0, core, log_status); > > + > > + dev_info(unicam->dev, "-----Receiver status-----\n"); > > + dev_info(unicam->dev, "V4L2 width/height: %ux%u\n", > > + node->fmt.fmt.pix.width, node->fmt.fmt.pix.height); > > + dev_info(unicam->dev, "Mediabus format: %08x\n", > > + node->fmtinfo->code); > > + dev_info(unicam->dev, "V4L2 format: %08x\n", > > + node->fmt.fmt.pix.pixelformat); > > + reg = unicam_reg_read(unicam, UNICAM_IPIPE); > > + dev_info(unicam->dev, "Unpacking/packing: %u / %u\n", > > + unicam_get_field(reg, UNICAM_PUM_MASK), > > + unicam_get_field(reg, UNICAM_PPM_MASK)); > > + dev_info(unicam->dev, "----Live data----\n"); > > + dev_info(unicam->dev, "Programmed stride: %4u\n", > > + unicam_reg_read(unicam, UNICAM_IBLS)); > > + dev_info(unicam->dev, "Detected resolution: %ux%u\n", > > + unicam_reg_read(unicam, UNICAM_IHSTA), > > + unicam_reg_read(unicam, UNICAM_IVSTA)); > > + dev_info(unicam->dev, "Write pointer: %08x\n", > > + unicam_reg_read(unicam, UNICAM_IBWP)); > > + > > + return 0; > > +} > > + > > +static int unicam_subscribe_event(struct v4l2_fh *fh, > > + const struct v4l2_event_subscription *sub) > > +{ > > + switch (sub->type) { > > + case V4L2_EVENT_FRAME_SYNC: > > + return v4l2_event_subscribe(fh, sub, 2, NULL); > > + default: > > + return -EINVAL; > > + } > > +} > > + > > +static const struct v4l2_ioctl_ops unicam_ioctl_ops = { > > + .vidioc_querycap = unicam_querycap, > > + > > + .vidioc_enum_fmt_vid_cap = unicam_enum_fmt_vid, > > + .vidioc_g_fmt_vid_cap = unicam_g_fmt_vid, > > + .vidioc_try_fmt_vid_cap = unicam_try_fmt_vid, > > + .vidioc_s_fmt_vid_cap = unicam_s_fmt_vid, > > + > > + .vidioc_enum_fmt_meta_cap = unicam_enum_fmt_meta, > > + .vidioc_g_fmt_meta_cap = unicam_g_fmt_meta, > > + .vidioc_try_fmt_meta_cap = unicam_try_fmt_meta, > > + .vidioc_s_fmt_meta_cap = unicam_s_fmt_meta, > > + > > + .vidioc_enum_framesizes = unicam_enum_framesizes, > > + > > + .vidioc_reqbufs = vb2_ioctl_reqbufs, > > + .vidioc_create_bufs = vb2_ioctl_create_bufs, > > + .vidioc_prepare_buf = vb2_ioctl_prepare_buf, > > + .vidioc_querybuf = vb2_ioctl_querybuf, > > + .vidioc_qbuf = vb2_ioctl_qbuf, > > + .vidioc_dqbuf = vb2_ioctl_dqbuf, > > + .vidioc_expbuf = vb2_ioctl_expbuf, > > + .vidioc_streamon = vb2_ioctl_streamon, > > + .vidioc_streamoff = vb2_ioctl_streamoff, > > + > > + .vidioc_log_status = unicam_log_status, > > + .vidioc_subscribe_event = unicam_subscribe_event, > > + .vidioc_unsubscribe_event = v4l2_event_unsubscribe, > > +}; > > + > > +/* unicam capture driver file operations */ > > +static const struct v4l2_file_operations unicam_fops = { > > + .owner = THIS_MODULE, > > + .open = v4l2_fh_open, > > + .release = vb2_fop_release, > > + .poll = vb2_fop_poll, > > + .unlocked_ioctl = video_ioctl2, > > + .mmap = vb2_fop_mmap, > > +}; > > + > > +static int unicam_video_link_validate(struct media_link *link) > > +{ > > + struct video_device *vdev = > > + media_entity_to_video_device(link->sink->entity); > > + struct v4l2_subdev *sd = > > + media_entity_to_v4l2_subdev(link->source->entity); > > + struct unicam_node *node = video_get_drvdata(vdev); > > + const u32 pad = is_image_node(node) ? UNICAM_SD_PAD_SOURCE_IMAGE > > + : UNICAM_SD_PAD_SOURCE_METADATA; > > + const struct v4l2_mbus_framefmt *format; > > + struct v4l2_subdev_state *state; > > + int ret = 0; > > + > > + state = v4l2_subdev_lock_and_get_active_state(sd); > > + > > + format = v4l2_subdev_state_get_format(state, pad, 0); > > + if (!format) { > > + ret = -EINVAL; > > + goto out; > > + } > > + > > + if (is_image_node(node)) { > > + const struct v4l2_pix_format *fmt = &node->fmt.fmt.pix; > > + > > + if (node->fmtinfo->code != format->code || > > + fmt->height != format->height || > > + fmt->width != format->width || > > + fmt->field != format->field) { > > + dev_dbg(node->dev->dev, > > + "image: (%u x %u) 0x%08x %s != (%u x %u) 0x%08x %s\n", > > + fmt->width, fmt->height, node->fmtinfo->code, > > + v4l2_field_names[fmt->field], > > + format->width, format->height, format->code, > > + v4l2_field_names[format->field]); > > + ret = -EPIPE; > > + }; > > No need for ; here and in the next loop > > > + } else { > > + const struct v4l2_meta_format *fmt = &node->fmt.fmt.meta; > > + > > + if (node->fmtinfo->code != format->code || > > + fmt->height != format->height || > > + fmt->width != format->width) { > > + dev_dbg(node->dev->dev, > > + "meta: (%u x %u) 0x%04x != (%u x %u) 0x%04x\n", > > + fmt->width, fmt->height, node->fmtinfo->code, > > + format->width, format->height, format->code); > > + ret = -EPIPE; > > + }; > > + } > > + > > +out: > > + v4l2_subdev_unlock_state(state); > > + return ret; > > +} > > + > > +static const struct media_entity_operations unicam_video_media_ops = { > > + .link_validate = unicam_video_link_validate, > > +}; > > + > > +static void unicam_node_release(struct video_device *vdev) > > +{ > > + struct unicam_node *node = video_get_drvdata(vdev); > > + > > + unicam_put(node->dev); > > +} > > + > > +static void unicam_set_default_format(struct unicam_node *node) > > +{ > > + if (is_image_node(node)) { > > + struct v4l2_pix_format *fmt = &node->fmt.fmt.pix; > > + > > + node->fmtinfo = &unicam_image_formats[0]; > > + node->fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; > > + > > + v4l2_fill_pix_format(fmt, &unicam_default_image_format); > > + fmt->pixelformat = node->fmtinfo->fourcc; > > + unicam_calc_image_size_bpl(node->dev, node->fmtinfo, fmt); > > + } else { > > + struct v4l2_meta_format *fmt = &node->fmt.fmt.meta; > > + > > + node->fmtinfo = &unicam_meta_formats[0]; > > + node->fmt.type = V4L2_BUF_TYPE_META_CAPTURE; > > + > > + fmt->dataformat = node->fmtinfo->fourcc; > > + fmt->width = unicam_default_meta_format.width; > > + fmt->height = unicam_default_meta_format.height; > > + unicam_calc_meta_size_bpl(node->dev, node->fmtinfo, fmt); > > + } > > +} > > + > > +static int unicam_register_node(struct unicam_device *unicam, > > + enum unicam_node_type type) > > +{ > > + const u32 pad_index = type == UNICAM_IMAGE_NODE > > + ? UNICAM_SD_PAD_SOURCE_IMAGE > > + : UNICAM_SD_PAD_SOURCE_METADATA; > > + struct unicam_node *node = &unicam->node[type]; > > + struct video_device *vdev = &node->video_dev; > > + struct vb2_queue *q = &node->buffer_queue; > > + int ret; > > + > > + node->dev = unicam_get(unicam); > > + node->id = type; > > + > > + spin_lock_init(&node->dma_queue_lock); > > + mutex_init(&node->lock); > > + > > + INIT_LIST_HEAD(&node->dma_queue); > > + > > + /* Initialize the videobuf2 queue. */ > > + q->type = type == UNICAM_IMAGE_NODE ? V4L2_BUF_TYPE_VIDEO_CAPTURE > > + : V4L2_BUF_TYPE_META_CAPTURE; > > + q->io_modes = VB2_MMAP | VB2_DMABUF; > > + q->drv_priv = node; > > + q->ops = &unicam_video_qops; > > + q->mem_ops = &vb2_dma_contig_memops; > > + q->buf_struct_size = sizeof(struct unicam_buffer); > > + q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; > > + q->lock = &node->lock; > > + q->min_queued_buffers = 1; > > + q->dev = unicam->dev; > > + > > + ret = vb2_queue_init(q); > > + if (ret) { > > + dev_err(unicam->dev, "vb2_queue_init() failed\n"); > > + goto err_unicam_put; > > + } > > + > > + /* Initialize the video device. */ > > + vdev->release = unicam_node_release; > > + vdev->fops = &unicam_fops; > > + vdev->ioctl_ops = &unicam_ioctl_ops; > > + vdev->v4l2_dev = &unicam->v4l2_dev; > > + vdev->vfl_dir = VFL_DIR_RX; > > + vdev->queue = q; > > + vdev->lock = &node->lock; > > + vdev->device_caps = type == UNICAM_IMAGE_NODE > > + ? V4L2_CAP_VIDEO_CAPTURE : V4L2_CAP_META_CAPTURE; > > + vdev->device_caps |= V4L2_CAP_STREAMING | V4L2_CAP_IO_MC; > > + vdev->entity.ops = &unicam_video_media_ops; > > + > > + snprintf(vdev->name, sizeof(vdev->name), "%s-%s", UNICAM_MODULE_NAME, > > + type == UNICAM_IMAGE_NODE ? "image" : "embedded"); > > + > > + video_set_drvdata(vdev, node); > > + > > + if (type == UNICAM_IMAGE_NODE) > > + vdev->entity.flags |= MEDIA_ENT_FL_DEFAULT; > > + > > + node->pad.flags = MEDIA_PAD_FL_SINK; > > + > > + ret = media_entity_pads_init(&vdev->entity, 1, &node->pad); > > + if (ret) > > + goto err_unicam_put; > > + > > + node->dummy_buf.size = UNICAM_DUMMY_BUF_SIZE; > > + node->dummy_buf_cpu_addr = dma_alloc_coherent(unicam->dev, > > + node->dummy_buf.size, > > + &node->dummy_buf.dma_addr, > > + GFP_KERNEL); > > + if (!node->dummy_buf_cpu_addr) { > > + dev_err(unicam->dev, "Unable to allocate dummy buffer.\n"); > > + ret = -ENOMEM; > > + goto err_entity_cleanup; > > + } > > + > > + unicam_set_default_format(node); > > + > > + ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1); > > + if (ret) { > > + dev_err(unicam->dev, "Unable to register video device %s\n", > > + vdev->name); > > + goto err_dma_free; > > + } > > + > > + node->registered = true; > > + > > + ret = media_create_pad_link(&unicam->subdev.sd.entity, > > + pad_index, > > + &node->video_dev.entity, > > + 0, > > + MEDIA_LNK_FL_ENABLED | > > + MEDIA_LNK_FL_IMMUTABLE); > > + if (ret) { > > + /* > > + * No need for cleanup, the caller will unregister the > > + * video device, which will drop the reference on the > > + * device and trigger the cleanup. > > + */ > > + dev_err(unicam->dev, "Unable to create pad link for %s\n", > > + unicam->sensor.subdev->name); > > + return ret; > > + } > > + > > + return 0; > > + > > +err_dma_free: > > + dma_free_coherent(unicam->dev, node->dummy_buf.size, > > + node->dummy_buf_cpu_addr, > > + node->dummy_buf.dma_addr); > > +err_entity_cleanup: > > + media_entity_cleanup(&vdev->entity); > > +err_unicam_put: > > + unicam_put(unicam); > > + return ret; > > +} > > + > > +static void unicam_unregister_nodes(struct unicam_device *unicam) > > +{ > > + unsigned int i; > > + > > + for (i = 0; i < ARRAY_SIZE(unicam->node); i++) { > > + struct unicam_node *node = &unicam->node[i]; > > + > > + if (node->dummy_buf_cpu_addr) > > + dma_free_coherent(unicam->dev, node->dummy_buf.size, > > + node->dummy_buf_cpu_addr, > > + node->dummy_buf.dma_addr); > > + > > + if (node->registered) { > > + video_unregister_device(&node->video_dev); > > + node->registered = false; > > + } > > + } > > +} > > + > > +/* ----------------------------------------------------------------------------- > > + * Power management > > + */ > > + > > +static int unicam_runtime_resume(struct device *dev) > > +{ > > + struct unicam_device *unicam = dev_get_drvdata(dev); > > + int ret; > > + > > + ret = clk_set_min_rate(unicam->vpu_clock, UNICAM_MIN_VPU_CLOCK_RATE); > > + if (ret) { > > + dev_err(unicam->dev, "failed to set up VPU clock\n"); > > + return ret; > > + } > > Same question as above, shouldn't this be an 'assigned-clock-rates in > dts ? > > > + > > + ret = clk_prepare_enable(unicam->vpu_clock); > > + if (ret) { > > + dev_err(unicam->dev, "Failed to enable VPU clock: %d\n", ret); > > + goto err_vpu_clock; > > + } > > + > > + ret = clk_set_rate(unicam->clock, 100 * 1000 * 1000); > > + if (ret) { > > + dev_err(unicam->dev, "failed to set up CSI clock\n"); > > + goto err_vpu_prepare; > > + } > > same question > > > + > > + ret = clk_prepare_enable(unicam->clock); > > + if (ret) { > > + dev_err(unicam->dev, "Failed to enable CSI clock: %d\n", ret); > > + goto err_vpu_prepare; > > Do you need to 'unset' the rate as done for the VPU clock ? > > > + } > > + > > + return 0; > > + > > +err_vpu_prepare: > > + clk_disable_unprepare(unicam->vpu_clock); > > +err_vpu_clock: > > + if (clk_set_min_rate(unicam->vpu_clock, 0)) > > + dev_err(unicam->dev, "failed to reset the VPU clock\n"); > > + > > + return ret; > > +} > > + > > +static int unicam_runtime_suspend(struct device *dev) > > +{ > > + struct unicam_device *unicam = dev_get_drvdata(dev); > > + > > + clk_disable_unprepare(unicam->clock); > > (continuing with the above question: probably not, as the clock is > not 'unset' neither here) > > The rest looks good, nice to see this progressing! > > Thanks > j > <snip>