> -----Original Message----- > From: Eunchul Kim [mailto:chulspro.kim@xxxxxxxxxxx] > Sent: Wednesday, December 12, 2012 4:35 PM > To: dri-devel@xxxxxxxxxxxxxxxxxxxxx; inki.dae@xxxxxxxxxxx > Cc: jy0.jeon@xxxxxxxxxxx; yj44.cho@xxxxxxxxxxx; jmock.shin@xxxxxxxxxxx; > jaejoon.seo@xxxxxxxxxxx; kyungmin.park@xxxxxxxxxxx; > chulspro.kim@xxxxxxxxxxx > Subject: [RFC v3 1/5] drm/exynos: add ipp subsystem > > IPP stand for Image Post Processing and supports image scaler/rotator > /crop/flip/csc(color space conversion) and input/output DMA operations > using ipp drivers. > also supports writeback and display output operations. > ipp driver include FIMC, Rotator, GSC, SC, so on. > and ipp is integration device driver for each hardware. > > Signed-off-by: Eunchul Kim <chulspro.kim@xxxxxxxxxxx> > Signed-off-by: Jinyoung Jeon <jy0.jeon@xxxxxxxxxxx> > --- > drivers/gpu/drm/exynos/Kconfig | 6 + > drivers/gpu/drm/exynos/Makefile | 1 + > drivers/gpu/drm/exynos/exynos_drm_drv.c | 24 + > drivers/gpu/drm/exynos/exynos_drm_drv.h | 7 + > drivers/gpu/drm/exynos/exynos_drm_ipp.c | 2059 > +++++++++++++++++++++++++++++++ > drivers/gpu/drm/exynos/exynos_drm_ipp.h | 266 ++++ > include/uapi/drm/exynos_drm.h | 190 +++ > 7 files changed, 2553 insertions(+), 0 deletions(-) > create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.c > create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.h > > diff --git a/drivers/gpu/drm/exynos/Kconfig > b/drivers/gpu/drm/exynos/Kconfig > index 4ea8cdc..bcf1c9d 100644 > --- a/drivers/gpu/drm/exynos/Kconfig > +++ b/drivers/gpu/drm/exynos/Kconfig > @@ -45,3 +45,9 @@ config DRM_EXYNOS_G2D > depends on DRM_EXYNOS && !VIDEO_SAMSUNG_S5P_G2D > help > Choose this option if you want to use Exynos G2D for DRM. > + > +config DRM_EXYNOS_IPP > + bool "Exynos DRM IPP" > + depends on DRM_EXYNOS > + help > + Choose this option if you want to use IPP feature for DRM. > diff --git a/drivers/gpu/drm/exynos/Makefile > b/drivers/gpu/drm/exynos/Makefile > index 26813b8..6c536ce 100644 > --- a/drivers/gpu/drm/exynos/Makefile > +++ b/drivers/gpu/drm/exynos/Makefile > @@ -16,5 +16,6 @@ exynosdrm-$(CONFIG_DRM_EXYNOS_HDMI) += > exynos_hdmi.o exynos_mixer.o \ > exynos_drm_hdmi.o > exynosdrm-$(CONFIG_DRM_EXYNOS_VIDI) += exynos_drm_vidi.o > exynosdrm-$(CONFIG_DRM_EXYNOS_G2D) += exynos_drm_g2d.o > +exynosdrm-$(CONFIG_DRM_EXYNOS_IPP) += exynos_drm_ipp.o > > obj-$(CONFIG_DRM_EXYNOS) += exynosdrm.o > diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.c > b/drivers/gpu/drm/exynos/exynos_drm_drv.c > index 4a1168d..0eb8a97 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_drv.c > +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.c > @@ -40,6 +40,7 @@ > #include "exynos_drm_vidi.h" > #include "exynos_drm_dmabuf.h" > #include "exynos_drm_g2d.h" > +#include "exynos_drm_ipp.h" > #include "exynos_drm_iommu.h" > > #define DRIVER_NAME "exynos" > @@ -249,6 +250,14 @@ static struct drm_ioctl_desc exynos_ioctls[] = { > exynos_g2d_set_cmdlist_ioctl, DRM_UNLOCKED | DRM_AUTH), > DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, > exynos_g2d_exec_ioctl, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, > + exynos_drm_ipp_get_property, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, > + exynos_drm_ipp_set_property, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, > + exynos_drm_ipp_queue_buf, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, > + exynos_drm_ipp_cmd_ctrl, DRM_UNLOCKED | DRM_AUTH), > }; > > static const struct file_operations exynos_drm_driver_fops = { > @@ -363,6 +372,12 @@ static int __init exynos_drm_init(void) > goto out_g2d; > #endif > > +#ifdef CONFIG_DRM_EXYNOS_IPP > + ret = platform_driver_register(&ipp_driver); > + if (ret < 0) > + goto out_ipp; > +#endif > + > ret = platform_driver_register(&exynos_drm_platform_driver); > if (ret < 0) > goto out_drm; > @@ -380,6 +395,11 @@ out: > platform_driver_unregister(&exynos_drm_platform_driver); > > out_drm: > +#ifdef CONFIG_DRM_EXYNOS_IPP > + platform_driver_unregister(&ipp_driver); > +out_ipp: > +#endif > + > #ifdef CONFIG_DRM_EXYNOS_G2D > platform_driver_unregister(&g2d_driver); > out_g2d: > @@ -416,6 +436,10 @@ static void __exit exynos_drm_exit(void) > > platform_driver_unregister(&exynos_drm_platform_driver); > > +#ifdef CONFIG_DRM_EXYNOS_IPP > + platform_driver_unregister(&ipp_driver); > +#endif > + > #ifdef CONFIG_DRM_EXYNOS_G2D > platform_driver_unregister(&g2d_driver); > #endif > diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.h > b/drivers/gpu/drm/exynos/exynos_drm_drv.h > index a9db025..a365788 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_drv.h > +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.h > @@ -235,8 +235,14 @@ struct exynos_drm_g2d_private { > unsigned int gem_nr; > }; > > +struct exynos_drm_ipp_private { > + struct device *dev; > + struct list_head event_list; > +}; > + > struct drm_exynos_file_private { > struct exynos_drm_g2d_private *g2d_priv; > + struct exynos_drm_ipp_private *ipp_priv; > }; > > /* > @@ -346,4 +352,5 @@ extern struct platform_driver mixer_driver; > extern struct platform_driver exynos_drm_common_hdmi_driver; > extern struct platform_driver vidi_driver; > extern struct platform_driver g2d_driver; > +extern struct platform_driver ipp_driver; > #endif > diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.c > b/drivers/gpu/drm/exynos/exynos_drm_ipp.c > new file mode 100644 > index 0000000..eb12cb5 > --- /dev/null > +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.c > @@ -0,0 +1,2059 @@ > +/* > + * Copyright (C) 2012 Samsung Electronics Co.Ltd > + * Authors: > + * Eunchul Kim <chulspro.kim@xxxxxxxxxxx> > + * Jinyoung Jeon <jy0.jeon@xxxxxxxxxxx> > + * Sangmin Lee <lsmin.lee@xxxxxxxxxxx> > + * > + * This program is free software; you can redistribute it and/or modify > it > + * under the terms of the GNU General Public License as published by > the > + * Free Software Foundation; either version 2 of the License, or (at > your > + * option) any later version. > + * > + */ > +#include <linux/kernel.h> > +#include <linux/module.h> > +#include <linux/platform_device.h> > +#include <linux/clk.h> > +#include <linux/pm_runtime.h> > +#include <plat/map-base.h> > + > +#include <drm/drmP.h> > +#include <drm/exynos_drm.h> > +#include "exynos_drm_drv.h" > +#include "exynos_drm_gem.h" > +#include "exynos_drm_ipp.h" > + > +/* > + * IPP is stand for Image Post Processing and > + * supports image scaler/rotator and input/output DMA operations. > + * using FIMC, GSC, Rotator, so on. > + * IPP is integration device driver of same attribute h/w > + */ > + > +#define get_ipp_context(dev) > platform_get_drvdata(to_platform_device(dev)) > +#define ipp_is_m2m_cmd(c) (c == IPP_CMD_M2M) > + > +/* > + * A structure of event. > + * > + * @base: base of event. > + * @event: ipp event. > + */ > +struct drm_exynos_ipp_send_event { > + struct drm_pending_event base; > + struct drm_exynos_ipp_event event; > +}; > + > +/* > + * A structure of memory node. > + * > + * @list: list head to memory queue information. > + * @ops_id: id of operations. > + * @prop_id: id of property. > + * @buf_id: id of buffer. > + * @buf_info: gem objects and dma address, size. > + * @filp: a pointer to drm_file. > + */ > +struct drm_exynos_ipp_mem_node { > + struct list_head list; > + enum drm_exynos_ops_id ops_id; > + u32 prop_id; > + u32 buf_id; > + struct drm_exynos_ipp_buf_info buf_info; > + struct drm_file *filp; > +}; > + > +/* > + * A structure of ipp context. > + * > + * @subdrv: prepare initialization using subdrv. > + * @ipp_lock: lock for synchronization of access to ipp_idr. > + * @prop_lock: lock for synchronization of access to prop_idr. > + * @ipp_idr: ipp driver idr. > + * @prop_idr: property idr. > + * @event_workq: event work queue. > + * @cmd_workq: command work queue. > + */ > +struct ipp_context { > + struct exynos_drm_subdrv subdrv; > + struct mutex ipp_lock; > + struct mutex prop_lock; > + struct idr ipp_idr; > + struct idr prop_idr; > + struct workqueue_struct *event_workq; > + struct workqueue_struct *cmd_workq; > +}; > + > +static LIST_HEAD(exynos_drm_ippdrv_list); > +static DEFINE_MUTEX(exynos_drm_ippdrv_lock); > +static BLOCKING_NOTIFIER_HEAD(exynos_drm_ippnb_list); > + > +int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ippdrv) > + return -EINVAL; > + > + mutex_lock(&exynos_drm_ippdrv_lock); > + list_add_tail(&ippdrv->drv_list, &exynos_drm_ippdrv_list); > + mutex_unlock(&exynos_drm_ippdrv_lock); > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_register); > + > +int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ippdrv) > + return -EINVAL; > + > + mutex_lock(&exynos_drm_ippdrv_lock); > + list_del(&ippdrv->drv_list); > + mutex_unlock(&exynos_drm_ippdrv_lock); > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_unregister); > + > +static int ipp_create_id(struct idr *id_idr, struct mutex *lock, void > *obj, > + u32 *idp) > +{ > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > +again: > + /* ensure there is space available to allocate a handle */ > + if (idr_pre_get(id_idr, GFP_KERNEL) == 0) { > + DRM_ERROR("failed to get idr.\n"); > + return -ENOMEM; > + } > + > + /* do the allocation under our mutexlock */ > + mutex_lock(lock); > + ret = idr_get_new_above(id_idr, obj, 1, (int *)idp); > + mutex_unlock(lock); > + if (ret == -EAGAIN) > + goto again; > + > + return ret; > +} > + > +static void *ipp_find_obj(struct idr *id_idr, struct mutex *lock, u32 id) > +{ > + void *obj; > + > + DRM_DEBUG_KMS("%s:id[%d]\n", __func__, id); > + > + mutex_lock(lock); > + > + /* find object using handle */ > + obj = idr_find(id_idr, id); > + if (!obj) { > + DRM_ERROR("failed to find object.\n"); > + mutex_unlock(lock); > + return ERR_PTR(-ENODEV); > + } > + > + mutex_unlock(lock); > + > + return obj; > +} > + > +static inline bool ipp_check_dedicated(struct exynos_drm_ippdrv *ippdrv, > + enum drm_exynos_ipp_cmd cmd) > +{ > + /* > + * check dedicated flag and WB, OUTPUT operation with > + * power on state. > + */ > + if (ippdrv->dedicated || (!ipp_is_m2m_cmd(cmd) && > + !pm_runtime_suspended(ippdrv->dev))) > + return true; > + > + return false; > +} > + > +static struct exynos_drm_ippdrv *ipp_find_driver(struct ipp_context *ctx, > + struct drm_exynos_ipp_property *property) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + u32 ipp_id = property->ipp_id; > + > + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, ipp_id); > + > + if (ipp_id) { > + /* find ipp driver using idr */ > + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, > + ipp_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("not found ipp%d driver.\n", ipp_id); > + goto err_null; Just return ippdrv; > + } > + > + /* > + * WB, OUTPUT opertion not supported multi-operation. > + * so, make dedicated state at set property ioctl. > + * when ipp driver finished operations, clear dedicated flags. > + */ > + if (ipp_check_dedicated(ippdrv, property->cmd)) { > + DRM_ERROR("already used choose device.\n"); > + goto err_null; Just return proper error; > + } > + > + /* > + * This is necessary to find correct device in ipp drivers. > + * ipp drivers have different abilities, > + * so need to check property. > + */ > + if (ippdrv->check_property && > + ippdrv->check_property(ippdrv->dev, property)) { > + DRM_ERROR("not support property.\n"); > + goto err_null; Just return proper error; > + } > + > + return ippdrv; > + } else { > + /* > + * This case is search all ipp driver for finding. > + * user application don't set ipp_id in this case, > + * so ipp subsystem search correct driver in driver list. > + */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, > drv_list) { > + if (ipp_check_dedicated(ippdrv, property->cmd)) { > + DRM_DEBUG_KMS("%s:used device.\n", __func__); > + continue; > + } > + > + if (ippdrv->check_property && > + ippdrv->check_property(ippdrv->dev, property)) { > + DRM_DEBUG_KMS("%s:not support property.\n", > + __func__); > + continue; > + } > + > + return ippdrv; > + } > + > + DRM_ERROR("not support ipp driver operations.\n"); > + } > + > + return ERR_PTR(-ENODEV); > + > +err_null: > + return NULL; And remove the above codes. > +} > + > +static struct exynos_drm_ippdrv *ipp_find_drv_by_handle(u32 prop_id) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); > + > + if (list_empty(&exynos_drm_ippdrv_list)) { > + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); > + return ERR_PTR(-ENODEV); > + } > + > + /* > + * This case is search ipp driver by prop_id handle. > + * sometimes, ipp subsystem find driver by prop_id. > + * e.g PAUSE state, queue buf, command contro. > + */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", __func__, > + count++, (int)ippdrv); > + > + if (!list_empty(&ippdrv->cmd_list)) { > + list_for_each_entry(c_node, &ippdrv->cmd_list, list) > + if (c_node->property.prop_id == prop_id) > + return ippdrv; > + } > + } > + > + return ERR_PTR(-ENODEV); > +} > + > +int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_prop_list *prop_list = data; > + struct exynos_drm_ippdrv *ippdrv; > + int count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!prop_list) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, prop_list->ipp_id); > + > + if (!prop_list->ipp_id) { > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, > drv_list) > + count++; > + /* > + * Supports ippdrv list count for user application. > + * First step user application getting ippdrv count. > + * and second step getting ippdrv capability using ipp_id. > + */ > + prop_list->count = count; > + } else { > + /* > + * Getting ippdrv capability by ipp_id. > + * some deivce not supported wb, output interface. > + * so, user application detect correct ipp driver > + * using this ioctl. > + */ > + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, > + prop_list->ipp_id); > + if (!ippdrv) { > + DRM_ERROR("not found ipp%d driver.\n", > + prop_list->ipp_id); > + return -EINVAL; > + } > + > + prop_list = ippdrv->prop_list; > + } > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_get_property); Should this function really be exported? Which external module calls this function? If not so, remove it. > + > +static void ipp_print_property(struct drm_exynos_ipp_property *property, > + int idx) > +{ > + struct drm_exynos_ipp_config *config = &property->config[idx]; > + struct drm_exynos_pos *pos = &config->pos; > + struct drm_exynos_sz *sz = &config->sz; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]ops[%s]fmt[0x%x]\n", > + __func__, property->prop_id, idx ? "dst" : "src", config- > >fmt); > + > + DRM_DEBUG_KMS("%s:pos[%d %d %d %d]sz[%d %d]f[%d]r[%d]\n", > + __func__, pos->x, pos->y, pos->w, pos->h, > + sz->hsize, sz->vsize, config->flip, config->degree); > +} > + > +static int ipp_find_and_set_property(struct drm_exynos_ipp_property > *property) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + u32 prop_id = property->prop_id; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); > + > + ippdrv = ipp_find_drv_by_handle(prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; > + } > + > + /* > + * Find command node using command list in ippdrv. > + * when we find this command no using prop_id. > + * return property information set in this command node. > + */ > + list_for_each_entry(c_node, &ippdrv->cmd_list, list) { > + if ((c_node->property.prop_id == prop_id) && > + (c_node->state == IPP_STATE_STOP)) { > + DRM_DEBUG_KMS("%s:found cmd[%d]ippdrv[0x%x]\n", > + __func__, property->cmd, (int)ippdrv); > + > + c_node->property = *property; > + return 0; > + } > + } > + > + DRM_ERROR("failed to search property.\n"); > + > + return -EINVAL; > +} > + > +static struct drm_exynos_ipp_cmd_work *ipp_create_cmd_work(void) > +{ > + struct drm_exynos_ipp_cmd_work *cmd_work; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + cmd_work = kzalloc(sizeof(*cmd_work), GFP_KERNEL); > + if (!cmd_work) { > + DRM_ERROR("failed to alloc cmd_work.\n"); > + return ERR_PTR(-ENOMEM); > + } > + > + INIT_WORK((struct work_struct *)cmd_work, ipp_sched_cmd); > + > + return cmd_work; > +} > + > +static struct drm_exynos_ipp_event_work *ipp_create_event_work(void) > +{ > + struct drm_exynos_ipp_event_work *event_work; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + event_work = kzalloc(sizeof(*event_work), GFP_KERNEL); > + if (!event_work) { > + DRM_ERROR("failed to alloc event_work.\n"); > + return ERR_PTR(-ENOMEM); > + } > + > + INIT_WORK((struct work_struct *)event_work, ipp_sched_event); > + > + return event_work; > +} > + > +int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_property *property = data; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret, i; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!property) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + /* > + * This is log print for user application property. > + * user application set various property. > + */ > + for_each_ipp_ops(i) > + ipp_print_property(property, i); > + > + /* > + * set property ioctl generated new prop_id. > + * but in this case already asigned prop_id using old set property. > + * e.g PAUSE state. this case supports find current prop_id and use > it > + * instead of allocation. > + */ > + if (property->prop_id) { > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property- > >prop_id); > + return ipp_find_and_set_property(property); > + } > + > + /* find ipp driver using ipp id */ > + ippdrv = ipp_find_driver(ctx, property); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; > + } > + > + /* allocate command node */ > + c_node = kzalloc(sizeof(*c_node), GFP_KERNEL); > + if (!c_node) { > + DRM_ERROR("failed to allocate map node.\n"); > + return -ENOMEM; > + } > + > + /* create property id */ > + ret = ipp_create_id(&ctx->prop_idr, &ctx->prop_lock, c_node, > + &property->prop_id); > + if (ret) { > + DRM_ERROR("failed to create id.\n"); > + goto err_clear; > + } > + > + DRM_DEBUG_KMS("%s:created prop_id[%d]cmd[%d]ippdrv[0x%x]\n", > + __func__, property->prop_id, property->cmd, (int)ippdrv); > + > + /* stored property information and ippdrv in private data */ > + c_node->priv = priv; > + c_node->property = *property; > + c_node->state = IPP_STATE_IDLE; > + > + c_node->start_work = ipp_create_cmd_work(); > + if (IS_ERR_OR_NULL(c_node->start_work)) { > + DRM_ERROR("failed to create start work.\n"); > + goto err_clear; > + } > + > + c_node->stop_work = ipp_create_cmd_work(); > + if (IS_ERR_OR_NULL(c_node->stop_work)) { > + DRM_ERROR("failed to create stop work.\n"); > + goto err_free_start; > + } > + > + c_node->event_work = ipp_create_event_work(); > + if (IS_ERR_OR_NULL(c_node->event_work)) { > + DRM_ERROR("failed to create event work.\n"); > + goto err_free_stop; > + } > + > + mutex_init(&c_node->cmd_lock); > + mutex_init(&c_node->mem_lock); > + mutex_init(&c_node->event_lock); > + > + init_completion(&c_node->start_complete); > + init_completion(&c_node->stop_complete); > + > + for_each_ipp_ops(i) > + INIT_LIST_HEAD(&c_node->mem_list[i]); > + > + INIT_LIST_HEAD(&c_node->event_list); > + list_splice_init(&priv->event_list, &c_node->event_list); > + list_add_tail(&c_node->list, &ippdrv->cmd_list); > + > + /* make dedicated state without m2m */ > + if (!ipp_is_m2m_cmd(property->cmd)) > + ippdrv->dedicated = true; > + > + return 0; > + > +err_free_stop: > + kfree(c_node->stop_work); > +err_free_start: > + kfree(c_node->start_work); > +err_clear: > + kfree(c_node); > + return ret; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_set_property); > + > +static void ipp_clean_cmd_node(struct drm_exynos_ipp_cmd_node *c_node) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* delete list */ > + list_del(&c_node->list); > + > + /* destroy mutex */ > + mutex_destroy(&c_node->cmd_lock); > + mutex_destroy(&c_node->mem_lock); > + mutex_destroy(&c_node->event_lock); > + > + /* free command node */ > + kfree(c_node->start_work); > + kfree(c_node->stop_work); > + kfree(c_node->event_work); > + kfree(c_node); > +} > + > +static int ipp_check_mem_list(struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct drm_exynos_ipp_mem_node *m_node; > + struct list_head *head; > + int ret, i, count[EXYNOS_DRM_OPS_MAX] = { 0, }; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_lock(&c_node->mem_lock); > + > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:%s memory empty.\n", __func__, > + i ? "dst" : "src"); > + continue; > + } > + > + /* find memory node entry */ > + list_for_each_entry(m_node, head, list) { > + DRM_DEBUG_KMS("%s:%s,count[%d]m_node[0x%x]\n", > __func__, > + i ? "dst" : "src", count[i], (int)m_node); > + count[i]++; > + } > + } > + > + DRM_DEBUG_KMS("%s:min[%d]max[%d]\n", __func__, > + min(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST]), > + max(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST])); > + > + /* > + * M2M operations should be need paired memory address. > + * so, need to check minimum count about src, dst. > + * other case not use paired memory, so use maximum count > + */ > + if (ipp_is_m2m_cmd(property->cmd)) > + ret = min(count[EXYNOS_DRM_OPS_SRC], > + count[EXYNOS_DRM_OPS_DST]); > + else > + ret = max(count[EXYNOS_DRM_OPS_SRC], > + count[EXYNOS_DRM_OPS_DST]); > + > + mutex_unlock(&c_node->mem_lock); > + > + return ret; > +} > + > +static struct drm_exynos_ipp_mem_node > + *ipp_find_mem_node(struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct list_head *head; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, qbuf->buf_id); > + > + /* source/destination memory list */ > + head = &c_node->mem_list[qbuf->ops_id]; > + > + /* find memory node from memory list */ > + list_for_each_entry(m_node, head, list) { > + DRM_DEBUG_KMS("%s:count[%d]m_node[0x%x]\n", > + __func__, count++, (int)m_node); > + > + /* compare buffer id */ > + if (m_node->buf_id == qbuf->buf_id) > + return m_node; > + } > + > + return NULL; > +} > + > +static int ipp_set_mem_node(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node) > +{ > + struct exynos_drm_ipp_ops *ops = NULL; > + int ret = 0; > + > + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); > + > + if (!m_node) { > + DRM_ERROR("invalid queue node.\n"); > + return -EFAULT; > + } > + > + mutex_lock(&c_node->mem_lock); > + > + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); > + > + /* get operations callback */ > + ops = ippdrv->ops[m_node->ops_id]; > + if (!ops) { > + DRM_ERROR("not support ops.\n"); > + ret = -EFAULT; > + goto err_unlock; > + } > + > + /* set address and enable irq */ > + if (ops->set_addr) { > + ret = ops->set_addr(ippdrv->dev, &m_node->buf_info, > + m_node->buf_id, IPP_BUF_ENQUEUE); > + if (ret) { > + DRM_ERROR("failed to set addr.\n"); > + goto err_unlock; > + } > + } > + > +err_unlock: > + mutex_unlock(&c_node->mem_lock); > + return ret; > +} > + > +static struct drm_exynos_ipp_mem_node > + *ipp_get_mem_node(struct drm_device *drm_dev, > + struct drm_file *file, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_buf_info buf_info; > + void *addr; > + int i; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_lock(&c_node->mem_lock); > + > + m_node = kzalloc(sizeof(*m_node), GFP_KERNEL); > + if (!m_node) { > + DRM_ERROR("failed to allocate queue node.\n"); > + goto err_unlock; > + } > + > + /* clear base address for error handling */ > + memset(&buf_info, 0x0, sizeof(buf_info)); > + > + /* operations, buffer id */ > + m_node->ops_id = qbuf->ops_id; > + m_node->prop_id = qbuf->prop_id; > + m_node->buf_id = qbuf->buf_id; > + > + DRM_DEBUG_KMS("%s:m_node[0x%x]ops_id[%d]\n", __func__, > + (int)m_node, qbuf->ops_id); > + DRM_DEBUG_KMS("%s:prop_id[%d]buf_id[%d]\n", __func__, > + qbuf->prop_id, m_node->buf_id); > + > + for_each_ipp_planar(i) { > + DRM_DEBUG_KMS("%s:i[%d]handle[0x%x]\n", __func__, > + i, qbuf->handle[i]); > + > + /* get dma address by handle */ > + if (qbuf->handle[i] != 0) { if (qbuf->handle[i]) { > + addr = exynos_drm_gem_get_dma_addr(drm_dev, > + qbuf->handle[i], file); > + if (!addr) { if (IS_ERR(addr)) > + DRM_ERROR("failed to get addr.\n"); > + goto err_clear; > + } > + > + buf_info.handles[i] = qbuf->handle[i]; > + buf_info.base[i] = *(dma_addr_t *) addr; > + DRM_DEBUG_KMS("%s:i[%d]base[0x%x]hd[0x%x]\n", > + __func__, i, buf_info.base[i], > + (int)buf_info.handles[i]); > + } > + } > + > + m_node->filp = file; > + m_node->buf_info = buf_info; > + list_add_tail(&m_node->list, &c_node->mem_list[qbuf->ops_id]); > + > + mutex_unlock(&c_node->mem_lock); > + return m_node; > + > +err_clear: > + kfree(m_node); > +err_unlock: > + mutex_unlock(&c_node->mem_lock); > + return ERR_PTR(-EFAULT); > +} > + > +static int ipp_put_mem_node(struct drm_device *drm_dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node) > +{ > + int i; > + > + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); > + > + if (!m_node) { > + DRM_ERROR("invalid dequeue node.\n"); > + return -EFAULT; > + } > + > + if (list_empty(&m_node->list)) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + mutex_lock(&c_node->mem_lock); > + > + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); > + > + /* put gem buffer */ > + for_each_ipp_planar(i) { > + unsigned long handle = m_node->buf_info.handles[i]; > + if (handle) > + exynos_drm_gem_put_dma_addr(drm_dev, handle, > + m_node->filp); > + } > + > + /* delete list in queue */ > + list_del(&m_node->list); > + kfree(m_node); > + > + mutex_unlock(&c_node->mem_lock); > + > + return 0; > +} > + > +static void ipp_free_event(struct drm_pending_event *event) Does this function need? This function just calls kfree. Just call kfree instead of wrapper. > +{ > + kfree(event); > +} > + > +static int ipp_get_event(struct drm_device *drm_dev, > + struct drm_file *file, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_send_event *e; > + unsigned long flags; > + > + DRM_DEBUG_KMS("%s:ops_id[%d]buf_id[%d]\n", __func__, > + qbuf->ops_id, qbuf->buf_id); > + > + e = kzalloc(sizeof(*e), GFP_KERNEL); > + > + if (!e) { > + DRM_ERROR("failed to allocate event.\n"); > + spin_lock_irqsave(&drm_dev->event_lock, flags); > + file->event_space += sizeof(e->event); > + spin_unlock_irqrestore(&drm_dev->event_lock, flags); > + return -ENOMEM; > + } > + > + /* make event */ > + e->event.base.type = DRM_EXYNOS_IPP_EVENT; > + e->event.base.length = sizeof(e->event); > + e->event.user_data = qbuf->user_data; > + e->event.prop_id = qbuf->prop_id; > + e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; > + e->base.event = &e->event.base; > + e->base.file_priv = file; > + e->base.destroy = ipp_free_event; > + list_add_tail(&e->base.link, &c_node->event_list); > + > + return 0; > +} > + > +static void ipp_put_event(struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_send_event *e, *te; > + int count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (list_empty(&c_node->event_list)) { > + DRM_DEBUG_KMS("%s:event_list is empty.\n", __func__); > + return; > + } > + > + list_for_each_entry_safe(e, te, &c_node->event_list, base.link) { > + DRM_DEBUG_KMS("%s:count[%d]e[0x%x]\n", > + __func__, count++, (int)e); > + > + /* > + * quf == NULL condition means all event deletion. > + * stop operations want to delete all event list. > + * another case delete only same buf id. > + */ > + if (!qbuf) { > + /* delete list */ > + list_del(&e->base.link); > + kfree(e); > + } > + > + /* compare buffer id */ > + if (qbuf && (qbuf->buf_id == > + e->event.buf_id[EXYNOS_DRM_OPS_DST])) { > + /* delete list */ > + list_del(&e->base.link); > + kfree(e); > + return; > + } > + } > +} > + > +void ipp_handle_cmd_work(struct device *dev, > + struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_work *cmd_work, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + cmd_work->ippdrv = ippdrv; > + cmd_work->c_node = c_node; > + queue_work(ctx->cmd_workq, (struct work_struct *)cmd_work); > +} > + > +static int ipp_queue_buf_with_run(struct device *dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_property *property; > + struct exynos_drm_ipp_ops *ops; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + ippdrv = ipp_find_drv_by_handle(qbuf->prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EFAULT; > + } > + > + ops = ippdrv->ops[qbuf->ops_id]; > + if (!ops) { > + DRM_ERROR("failed to get ops.\n"); > + return -EFAULT; > + } > + > + property = &c_node->property; > + if (!property) { Property couldn't be NULL. remove it. > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + if (c_node->state != IPP_STATE_START) { > + DRM_DEBUG_KMS("%s:bypass for invalid state.\n" , __func__); > + return 0; No error? > + } > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return 0; No error? > + } > + > + /* > + * If set destination buffer and enabled clock, > + * then m2m operations need start operations at queue_buf > + */ > + if (ipp_is_m2m_cmd(property->cmd)) { > + struct drm_exynos_ipp_cmd_work *cmd_work = c_node- > >start_work; > + > + cmd_work->ctrl = IPP_CTRL_PLAY; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + } else { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +static void ipp_clean_queue_buf(struct drm_device *drm_dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node, *tm_node; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!list_empty(&c_node->mem_list[qbuf->ops_id])) { > + /* delete list */ > + list_for_each_entry_safe(m_node, tm_node, > + &c_node->mem_list[qbuf->ops_id], list) { > + if (m_node->buf_id == qbuf->buf_id && > + m_node->ops_id == qbuf->ops_id) > + ipp_put_mem_node(drm_dev, c_node, m_node); > + } > + } > +} > + > +int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_queue_buf *qbuf = data; > + struct drm_exynos_ipp_cmd_node *c_node; > + struct drm_exynos_ipp_mem_node *m_node; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!qbuf) { > + DRM_ERROR("invalid buf parameter.\n"); > + return -EINVAL; > + } > + > + if (qbuf->ops_id >= EXYNOS_DRM_OPS_MAX) { > + DRM_ERROR("invalid ops parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", > + __func__, qbuf->prop_id, qbuf->ops_id ? "dst" : "src", > + qbuf->buf_id, qbuf->buf_type); > + > + /* find command node */ > + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, > + qbuf->prop_id); > + if (!c_node) { > + DRM_ERROR("failed to get command node.\n"); > + return -EFAULT; > + } > + > + /* buffer control */ > + switch (qbuf->buf_type) { > + case IPP_BUF_ENQUEUE: > + /* get memory node */ > + m_node = ipp_get_mem_node(drm_dev, file, c_node, qbuf); > + if (IS_ERR_OR_NULL(m_node)) { This function never returns error. Use IS_ERR(m_node) instead. > + DRM_ERROR("failed to get m_node.\n"); > + return -EINVAL; And return PTR_ERR(m_node); > + } > + > + /* > + * first step get event for destination buffer. > + * and second step when M2M case run with destination buffer > + * if needed. > + */ > + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) { > + /* get event for destination buffer */ > + ret = ipp_get_event(drm_dev, file, c_node, qbuf); > + if (ret) { > + DRM_ERROR("failed to get event.\n"); > + goto err_clean_node; > + } > + > + /* > + * M2M case run play control for streaming feature. > + * other case set address and waiting. > + */ > + ret = ipp_queue_buf_with_run(dev, c_node, m_node, > qbuf); > + if (ret) { > + DRM_ERROR("failed to run command.\n"); > + goto err_clean_node; > + } > + } > + break; > + case IPP_BUF_DEQUEUE: > + mutex_lock(&c_node->cmd_lock); > + > + /* put event for destination buffer */ > + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) > + ipp_put_event(c_node, qbuf); > + > + ipp_clean_queue_buf(drm_dev, c_node, qbuf); > + > + mutex_unlock(&c_node->cmd_lock); > + break; > + default: > + DRM_ERROR("invalid buffer control.\n"); > + return -EINVAL; > + } > + > + return 0; > + > +err_clean_node: > + DRM_ERROR("clean memory nodes.\n"); > + > + ipp_clean_queue_buf(drm_dev, c_node, qbuf); > + return ret; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_queue_buf); Remove EXPORT_SYMBOL_GPL. > + > +static bool exynos_drm_ipp_check_valid(struct device *dev, > + enum drm_exynos_ipp_ctrl ctrl, enum drm_exynos_ipp_state > state) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (ctrl != IPP_CTRL_PLAY) { > + if (pm_runtime_suspended(dev)) { > + DRM_ERROR("pm:runtime_suspended.\n"); > + goto err_status; > + } > + } > + > + switch (ctrl) { > + case IPP_CTRL_PLAY: > + if (state != IPP_STATE_IDLE) > + goto err_status; > + break; > + case IPP_CTRL_STOP: > + if (state == IPP_STATE_STOP) > + goto err_status; > + break; > + case IPP_CTRL_PAUSE: > + if (state != IPP_STATE_START) > + goto err_status; > + break; > + case IPP_CTRL_RESUME: > + if (state != IPP_STATE_STOP) > + goto err_status; > + break; > + default: > + DRM_ERROR("invalid state.\n"); > + goto err_status; > + break; > + } > + > + return true; > + > +err_status: > + DRM_ERROR("invalid status:ctrl[%d]state[%d]\n", ctrl, state); > + return false; > +} > + > +int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct exynos_drm_ippdrv *ippdrv = NULL; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_cmd_ctrl *cmd_ctrl = data; > + struct drm_exynos_ipp_cmd_work *cmd_work; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!cmd_ctrl) { > + DRM_ERROR("invalid control parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:ctrl[%d]prop_id[%d]\n", __func__, > + cmd_ctrl->ctrl, cmd_ctrl->prop_id); > + > + ippdrv = ipp_find_drv_by_handle(cmd_ctrl->prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { Never return NULL. just use IS_ERR(ippdrv) > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; And return PTR_ERR(ippdrv); > + } > + > + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, > + cmd_ctrl->prop_id); > + if (!c_node) { > + DRM_ERROR("invalid command node list.\n"); > + return -EINVAL; > + } > + > + if (!exynos_drm_ipp_check_valid(ippdrv->dev, cmd_ctrl->ctrl, > + c_node->state)) { > + DRM_ERROR("invalid state.\n"); > + return -EINVAL; > + } > + > + switch (cmd_ctrl->ctrl) { > + case IPP_CTRL_PLAY: > + if (pm_runtime_suspended(ippdrv->dev)) > + pm_runtime_get_sync(ippdrv->dev); > + c_node->state = IPP_STATE_START; > + > + cmd_work = c_node->start_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + c_node->state = IPP_STATE_START; > + break; > + case IPP_CTRL_STOP: > + cmd_work = c_node->stop_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + > + if (!wait_for_completion_timeout(&c_node->stop_complete, > + msecs_to_jiffies(300))) { > + DRM_ERROR("timeout stop:prop_id[%d]\n", > + c_node->property.prop_id); > + } > + > + c_node->state = IPP_STATE_STOP; > + ippdrv->dedicated = false; > + ipp_clean_cmd_node(c_node); > + > + if (list_empty(&ippdrv->cmd_list)) > + pm_runtime_put_sync(ippdrv->dev); > + break; > + case IPP_CTRL_PAUSE: > + cmd_work = c_node->stop_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + > + if (!wait_for_completion_timeout(&c_node->stop_complete, > + msecs_to_jiffies(200))) { > + DRM_ERROR("timeout stop:prop_id[%d]\n", > + c_node->property.prop_id); > + } > + > + c_node->state = IPP_STATE_STOP; > + break; > + case IPP_CTRL_RESUME: > + c_node->state = IPP_STATE_START; > + cmd_work = c_node->start_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + break; > + default: > + /* ToDo: expand ctrl operation */ > + DRM_ERROR("could not support this state currently.\n"); > + goto err_clear; Just return error; > + } > + > + DRM_DEBUG_KMS("%s:done ctrl[%d]prop_id[%d]\n", __func__, > + cmd_ctrl->ctrl, cmd_ctrl->prop_id); > + > + return 0; > + > +err_clear: > + return ret; And remove the above label. > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_cmd_ctrl); Remove it. > + > +int exynos_drm_ippnb_register(struct notifier_block *nb) > +{ > + return blocking_notifier_chain_register( > + &exynos_drm_ippnb_list, nb); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_register); Remove it. > + > +int exynos_drm_ippnb_unregister(struct notifier_block *nb) > +{ > + return blocking_notifier_chain_unregister( > + &exynos_drm_ippnb_list, nb); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_unregister); Remove it. > + > +int exynos_drm_ippnb_send_event(unsigned long val, void *v) > +{ > + return blocking_notifier_call_chain( > + &exynos_drm_ippnb_list, val, v); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_send_event); Remove it. > + > +static int ipp_set_property(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_property *property) > +{ > + struct exynos_drm_ipp_ops *ops = NULL; > + int ret, i, swap = 0; > + > + if (!property) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* reset h/w block */ > + if (ippdrv->reset && > + ippdrv->reset(ippdrv->dev)) { > + DRM_ERROR("failed to reset.\n"); > + return -EINVAL; > + } > + > + /* set source,destination operations */ > + for_each_ipp_ops(i) { > + /* ToDo: integrate property and config */ > + struct drm_exynos_ipp_config *config = > + &property->config[i]; > + > + ops = ippdrv->ops[i]; > + if (!ops || !config) { > + DRM_ERROR("not support ops and config.\n"); > + return -EINVAL; > + } > + > + /* set format */ > + if (ops->set_fmt) { > + ret = ops->set_fmt(ippdrv->dev, config->fmt); > + if (ret) { > + DRM_ERROR("not support format.\n"); > + return ret; > + } > + } > + > + /* set transform for rotation, flip */ > + if (ops->set_transf) { > + swap = ops->set_transf(ippdrv->dev, config->degree, > + config->flip); > + if (swap < 0) { > + DRM_ERROR("not support tranf.\n"); > + return -EINVAL; > + } > + } > + > + /* set size */ > + if (ops->set_size) { > + ret = ops->set_size(ippdrv->dev, swap, &config->pos, > + &config->sz); > + if (ret) { > + DRM_ERROR("not support size.\n"); > + return ret; > + } > + } > + } > + > + return 0; > +} > + > +static int ipp_start_property(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct list_head *head; > + int ret, i; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* store command info in ippdrv */ > + ippdrv->cmd = c_node; > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return -ENOMEM; > + } > + > + /* set current property in ippdrv */ > + ret = ipp_set_property(ippdrv, property); > + if (ret) { > + DRM_ERROR("failed to set property.\n"); > + ippdrv->cmd = NULL; > + return ret; > + } > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("failed to get node.\n"); > + ret = -EFAULT; > + return ret; > + } > + > + DRM_DEBUG_KMS("%s:m_node[0x%x]\n", > + __func__, (int)m_node); > + > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + case IPP_CMD_WB: > + /* destination memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; > + > + list_for_each_entry(m_node, head, list) { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + list_for_each_entry(m_node, head, list) { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:cmd[%d]\n", __func__, property->cmd); > + > + /* start operations */ > + if (ippdrv->start) { > + ret = ippdrv->start(ippdrv->dev, property->cmd); > + if (ret) { > + DRM_ERROR("failed to start ops.\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +static int ipp_stop_property(struct drm_device *drm_dev, > + struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_mem_node *m_node, *tm_node; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct list_head *head; > + int ret = 0, i; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* put event */ > + ipp_put_event(c_node, NULL); > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", > + __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, > + head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, > + m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + } > + break; > + case IPP_CMD_WB: > + /* destination memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + ret = -EINVAL; > + goto err_clear; > + } > + > +err_clear: > + /* stop operations */ > + if (ippdrv->stop) > + ippdrv->stop(ippdrv->dev, property->cmd); > + > + return ret; > +} > + > +void ipp_sched_cmd(struct work_struct *work) > +{ > + struct drm_exynos_ipp_cmd_work *cmd_work = > + (struct drm_exynos_ipp_cmd_work *)work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + struct drm_exynos_ipp_property *property; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + ippdrv = cmd_work->ippdrv; > + if (!ippdrv) { > + DRM_ERROR("invalid ippdrv list.\n"); > + return; > + } > + > + c_node = cmd_work->c_node; > + if (!c_node) { > + DRM_ERROR("invalid command node list.\n"); > + return; > + } > + > + mutex_lock(&c_node->cmd_lock); > + > + property = &c_node->property; > + if (!property) { > + DRM_ERROR("failed to get property:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + > + switch (cmd_work->ctrl) { > + case IPP_CTRL_PLAY: > + case IPP_CTRL_RESUME: > + ret = ipp_start_property(ippdrv, c_node); > + if (ret) { > + DRM_ERROR("failed to start property:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + > + /* > + * M2M case supports wait_completion of transfer. > + * because M2M case supports single unit operation > + * with multiple queue. > + * M2M need to wait completion of data transfer. > + */ > + if (ipp_is_m2m_cmd(property->cmd)) { > + if (!wait_for_completion_timeout > + (&c_node->start_complete, msecs_to_jiffies(200))) { > + DRM_ERROR("timeout event:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + } > + break; > + case IPP_CTRL_STOP: > + case IPP_CTRL_PAUSE: > + ret = ipp_stop_property(ippdrv->drm_dev, ippdrv, > + c_node); > + if (ret) { > + DRM_ERROR("failed to stop property.\n"); > + goto err_unlock; > + } > + > + complete(&c_node->stop_complete); > + break; > + default: > + DRM_ERROR("unknown control type\n"); > + break; > + } > + > + DRM_DEBUG_KMS("%s:ctrl[%d] done.\n", __func__, cmd_work->ctrl); > + > +err_unlock: > + mutex_unlock(&c_node->cmd_lock); > +} > + > +static int ipp_send_event(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) > +{ > + struct drm_device *drm_dev = ippdrv->drm_dev; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_queue_buf qbuf; > + struct drm_exynos_ipp_send_event *e; > + struct list_head *head; > + struct timeval now; > + unsigned long flags; > + u32 tbuf_id[EXYNOS_DRM_OPS_MAX] = {0, }; > + int ret, i; > + > + for_each_ipp_ops(i) > + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, > + i ? "dst" : "src", buf_id[i]); > + > + if (!drm_dev) { > + DRM_ERROR("failed to get drm_dev.\n"); > + return -EINVAL; > + } > + > + if (!property) { > + DRM_ERROR("failed to get property.\n"); > + return -EINVAL; > + } > + > + if (list_empty(&c_node->event_list)) { > + DRM_DEBUG_KMS("%s:event list is empty.\n", __func__); > + return 0; > + } > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return 0; > + } > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[i] = m_node->buf_id; > + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, > + i ? "dst" : "src", tbuf_id[i]); > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + } > + break; > + case IPP_CMD_WB: > + /* clear buf for finding */ > + memset(&qbuf, 0x0, sizeof(qbuf)); > + qbuf.ops_id = EXYNOS_DRM_OPS_DST; > + qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST]; > + > + /* get memory node entry */ > + m_node = ipp_find_mem_node(c_node, &qbuf); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id; > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id; > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + return -EINVAL; > + } > + > + /* ToDo: Fix buffer id */ > + if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST]) > + DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n", > + tbuf_id[1], buf_id[1], property->prop_id); > + > + /* > + * command node have event list of destination buffer > + * If destination buffer enqueue to mem list, > + * then we make event and link to event list tail. > + * so, we get first event for first enqueued buffer. > + */ > + e = list_first_entry(&c_node->event_list, > + struct drm_exynos_ipp_send_event, base.link); > + > + if (!e) { > + DRM_ERROR("empty event.\n"); > + return -EINVAL; > + } > + > + do_gettimeofday(&now); > + DRM_DEBUG_KMS("%s:tv_sec[%ld]tv_usec[%ld]\n" > + , __func__, now.tv_sec, now.tv_usec); > + e->event.tv_sec = now.tv_sec; > + e->event.tv_usec = now.tv_usec; > + e->event.prop_id = property->prop_id; > + > + /* set buffer id about source destination */ > + for_each_ipp_ops(i) > + e->event.buf_id[i] = tbuf_id[i]; > + > + /* ToDo: compare index. If needed */ > + > + spin_lock_irqsave(&drm_dev->event_lock, flags); > + list_move_tail(&e->base.link, &e->base.file_priv->event_list); > + wake_up_interruptible(&e->base.file_priv->event_wait); > + spin_unlock_irqrestore(&drm_dev->event_lock, flags); > + > + DRM_DEBUG_KMS("%s:done cmd[%d]prop_id[%d]buf_id[%d]\n", __func__, > + property->cmd, property->prop_id, > tbuf_id[EXYNOS_DRM_OPS_DST]); > + > + return 0; > +} > + > +void ipp_sched_event(struct work_struct *work) > +{ > + struct drm_exynos_ipp_event_work *event_work = > + (struct drm_exynos_ipp_event_work *)work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret; > + > + if (!event_work) { > + DRM_ERROR("failed to get event_work.\n"); > + return; > + } > + > + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, > + event_work->buf_id[EXYNOS_DRM_OPS_DST]); > + > + ippdrv = event_work->ippdrv; > + if (!ippdrv) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return; > + } > + > + c_node = ippdrv->cmd; > + if (!c_node) { > + DRM_ERROR("failed to get command node.\n"); > + return; > + } > + > + /* > + * IPP supports command thread, event thread synchronization. > + * If IPP close immediately from user land, then IPP make > + * synchronization with command thread, so make complete event. > + * or going out operations. > + */ > + if (c_node->state != IPP_STATE_START) { > + DRM_DEBUG_KMS("%s:bypass state[%d]prop_id[%d]\n", > + __func__, c_node->state, c_node->property.prop_id); > + goto err_completion; > + } > + > + mutex_lock(&c_node->event_lock); > + > + ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); > + if (ret) { > + DRM_ERROR("failed to send event.\n"); > + goto err_completion; > + } > + > +err_completion: > + if (ipp_is_m2m_cmd(c_node->property.cmd)) > + complete(&c_node->start_complete); > + > + mutex_unlock(&c_node->event_lock); > +} > + > +static int ipp_subdrv_probe(struct drm_device *drm_dev, struct device > *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + struct exynos_drm_ippdrv *ippdrv; > + int ret, count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* get ipp driver entry */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + ippdrv->drm_dev = drm_dev; > + > + ret = ipp_create_id(&ctx->ipp_idr, &ctx->ipp_lock, ippdrv, > + &ippdrv->ipp_id); > + if (ret) { > + DRM_ERROR("failed to create id.\n"); > + goto err_idr; > + } > + > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]ipp_id[%d]\n", > __func__, > + count++, (int)ippdrv, ippdrv->ipp_id); > + > + if (ippdrv->ipp_id == 0) { > + DRM_ERROR("failed to get ipp_id[%d]\n", > + ippdrv->ipp_id); > + goto err_idr; > + } > + > + /* store parent device for node */ > + ippdrv->parent_dev = dev; > + > + /* store event work queue and handler */ > + ippdrv->event_workq = ctx->event_workq; > + ippdrv->sched_event = ipp_sched_event; > + INIT_LIST_HEAD(&ippdrv->cmd_list); > + } > + > + return 0; > + > +err_idr: > + idr_remove_all(&ctx->ipp_idr); > + idr_remove_all(&ctx->prop_idr); > + idr_destroy(&ctx->ipp_idr); > + idr_destroy(&ctx->prop_idr); > + return ret; > +} > + > +static void ipp_subdrv_remove(struct drm_device *drm_dev, struct device > *dev) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* get ipp driver entry */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + ippdrv->drm_dev = NULL; > + exynos_drm_ippdrv_unregister(ippdrv); > + } > + > + /* ToDo: free notifier callback list if needed */ > +} > + > +static int ipp_subdrv_open(struct drm_device *drm_dev, struct device *dev, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* ToDo: multi device open */ > + > + priv = kzalloc(sizeof(*priv), GFP_KERNEL); > + if (!priv) { > + DRM_ERROR("failed to allocate priv.\n"); > + return -ENOMEM; > + } > + priv->dev = dev; > + file_priv->ipp_priv = priv; > + > + INIT_LIST_HEAD(&priv->event_list); > + > + DRM_DEBUG_KMS("%s:done priv[0x%x]\n", __func__, (int)priv); > + > + return 0; > +} > + > +static void ipp_subdrv_close(struct drm_device *drm_dev, struct device > *dev, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct exynos_drm_ippdrv *ippdrv = NULL; > + struct drm_exynos_ipp_cmd_node *c_node, *tc_node; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:for priv[0x%x]\n", __func__, (int)priv); > + > + if (list_empty(&exynos_drm_ippdrv_list)) { > + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); > + goto err_clear; > + } > + > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + if (list_empty(&ippdrv->cmd_list)) > + continue; > + > + list_for_each_entry_safe(c_node, tc_node, > + &ippdrv->cmd_list, list) { > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", > + __func__, count++, (int)ippdrv); > + > + if (c_node->priv == priv) { > + /* > + * userland goto unnormal state. process killed. > + * and close the file. > + * so, IPP didn't called stop cmd ctrl. > + * so, we are make stop operation in this state. > + */ > + if (c_node->state == IPP_STATE_START) { > + ipp_stop_property(drm_dev, ippdrv, > + c_node); > + c_node->state = IPP_STATE_STOP; > + } > + > + ippdrv->dedicated = false; > + ipp_clean_cmd_node(c_node); > + if (list_empty(&ippdrv->cmd_list)) > + pm_runtime_put_sync(ippdrv->dev); > + } > + } > + } > + > +err_clear: > + kfree(priv); > + return; > +} > + > +static int __devinit ipp_probe(struct platform_device *pdev) > +{ > + struct device *dev = &pdev->dev; > + struct ipp_context *ctx; > + struct exynos_drm_subdrv *subdrv; > + int ret; > + > + ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); > + if (!ctx) > + return -ENOMEM; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_init(&ctx->ipp_lock); > + mutex_init(&ctx->prop_lock); > + > + idr_init(&ctx->ipp_idr); > + idr_init(&ctx->prop_idr); > + > + /* > + * create single thread for ipp event > + * IPP supports event thread for IPP drivers. > + * IPP driver send event_work to this thread. > + * and IPP event thread send event to user process. > + */ > + ctx->event_workq = create_singlethread_workqueue("ipp_event"); > + if (!ctx->event_workq) { > + dev_err(dev, "failed to create event workqueue\n"); > + ret = -EINVAL; > + goto err_clear; > + } > + > + /* > + * create single thread for ipp command > + * IPP supports command thread for user process. > + * user process make command node using set property ioctl. > + * and make start_work and send this work to command thread. > + * and then this command thread start property. > + */ > + ctx->cmd_workq = create_singlethread_workqueue("ipp_cmd"); > + if (!ctx->cmd_workq) { > + dev_err(dev, "failed to create cmd workqueue\n"); > + ret = -EINVAL; > + goto err_event_workq; > + } > + > + /* set sub driver informations */ > + subdrv = &ctx->subdrv; > + subdrv->dev = dev; > + subdrv->probe = ipp_subdrv_probe; > + subdrv->remove = ipp_subdrv_remove; > + subdrv->open = ipp_subdrv_open; > + subdrv->close = ipp_subdrv_close; > + > + platform_set_drvdata(pdev, ctx); > + > + ret = exynos_drm_subdrv_register(subdrv); > + if (ret < 0) { > + DRM_ERROR("failed to register drm ipp device.\n"); > + goto err_cmd_workq; > + } > + > + dev_info(&pdev->dev, "drm ipp registered successfully.\n"); > + > + return 0; > + > +err_cmd_workq: > + destroy_workqueue(ctx->cmd_workq); > +err_event_workq: > + destroy_workqueue(ctx->event_workq); > +err_clear: > + kfree(ctx); > + return ret; > +} > + > +static int __devexit ipp_remove(struct platform_device *pdev) > +{ > + struct ipp_context *ctx = platform_get_drvdata(pdev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* unregister sub driver */ > + exynos_drm_subdrv_unregister(&ctx->subdrv); > + > + /* remove,destroy ipp idr */ > + idr_remove_all(&ctx->ipp_idr); > + idr_remove_all(&ctx->prop_idr); > + idr_destroy(&ctx->ipp_idr); > + idr_destroy(&ctx->prop_idr); > + > + mutex_destroy(&ctx->ipp_lock); > + mutex_destroy(&ctx->prop_lock); > + > + /* destroy command, event work queue */ > + destroy_workqueue(ctx->cmd_workq); > + destroy_workqueue(ctx->event_workq); > + > + kfree(ctx); > + > + return 0; > +} > + > +static int ipp_power_ctrl(struct ipp_context *ctx, bool enable) > +{ > + /* ToDo: Need to implement power and sysmmu ctrl. */ > + DRM_DEBUG_KMS("%s:enable[%d]\n", __func__, enable); > + > + return 0; > +} > + > +#ifdef CONFIG_PM_SLEEP > +static int ipp_suspend(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (pm_runtime_suspended(dev)) > + return 0; > + > + return ipp_power_ctrl(ctx, false); > +} > + > +static int ipp_resume(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!pm_runtime_suspended(dev)) > + return ipp_power_ctrl(ctx, true); > + > + return 0; > +} > +#endif > + > +#ifdef CONFIG_PM_RUNTIME > +static int ipp_runtime_suspend(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + return ipp_power_ctrl(ctx, false); > +} > + > +static int ipp_runtime_resume(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + return ipp_power_ctrl(ctx, true); > +} > +#endif > + > +static const struct dev_pm_ops ipp_pm_ops = { > + SET_SYSTEM_SLEEP_PM_OPS(ipp_suspend, ipp_resume) > + SET_RUNTIME_PM_OPS(ipp_runtime_suspend, ipp_runtime_resume, NULL) > +}; > + > +struct platform_driver ipp_driver = { > + .probe = ipp_probe, > + .remove = __devexit_p(ipp_remove), > + .driver = { > + .name = "exynos-drm-ipp", > + .owner = THIS_MODULE, > + .pm = &ipp_pm_ops, > + }, > +}; > + > diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.h > b/drivers/gpu/drm/exynos/exynos_drm_ipp.h > new file mode 100644 > index 0000000..baab1f0 > --- /dev/null > +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.h > @@ -0,0 +1,266 @@ > +/* > + * Copyright (c) 2012 Samsung Electronics Co., Ltd. > + * > + * Authors: > + * Eunchul Kim <chulspro.kim@xxxxxxxxxxx> > + * Jinyoung Jeon <jy0.jeon@xxxxxxxxxxx> > + * Sangmin Lee <lsmin.lee@xxxxxxxxxxx> > + * > + * Permission is hereby granted, free of charge, to any person obtaining > a > + * copy of this software and associated documentation files (the > "Software"), > + * to deal in the Software without restriction, including without > limitation > + * the rights to use, copy, modify, merge, publish, distribute, > sublicense, > + * and/or sell copies of the Software, and to permit persons to whom the > + * Software is furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice (including the > next > + * paragraph) shall be included in all copies or substantial portions of > the > + * Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, > EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF > MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT > SHALL > + * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES > OR > + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, > + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR > + * OTHER DEALINGS IN THE SOFTWARE. > + */ > + > +#ifndef _EXYNOS_DRM_IPP_H_ > +#define _EXYNOS_DRM_IPP_H_ > + > +#define for_each_ipp_ops(pos) \ > + for (pos = 0; pos < EXYNOS_DRM_OPS_MAX; pos++) > +#define for_each_ipp_planar(pos) \ > + for (pos = 0; pos < EXYNOS_DRM_PLANAR_MAX; pos++) > + > +#define IPP_GET_LCD_WIDTH _IOR('F', 302, int) > +#define IPP_GET_LCD_HEIGHT _IOR('F', 303, int) > +#define IPP_SET_WRITEBACK _IOW('F', 304, u32) > + > +/* definition of state */ > +enum drm_exynos_ipp_state { > + IPP_STATE_IDLE, > + IPP_STATE_START, > + IPP_STATE_STOP, > +}; > + > +/* > + * A structure of command work information. > + * @work: work structure. > + * @ippdrv: current work ippdrv. > + * @c_node: command node information. > + * @ctrl: command control. > + */ > +struct drm_exynos_ipp_cmd_work { > + struct work_struct work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + enum drm_exynos_ipp_ctrl ctrl; > +}; > + > +/* > + * A structure of command node. > + * > + * @priv: IPP private infomation. > + * @list: list head to command queue information. > + * @event_list: list head of event. > + * @mem_list: list head to source,destination memory queue information. > + * @cmd_lock: lock for synchronization of access to ioctl. > + * @mem_lock: lock for synchronization of access to memory nodes. > + * @event_lock: lock for synchronization of access to scheduled event. > + * @start_complete: completion of start of command. > + * @stop_complete: completion of stop of command. > + * @property: property information. > + * @start_work: start command work structure. > + * @stop_work: stop command work structure. > + * @event_work: event work structure. > + * @state: state of command node. > + */ > +struct drm_exynos_ipp_cmd_node { > + struct exynos_drm_ipp_private *priv; > + struct list_head list; > + struct list_head event_list; > + struct list_head mem_list[EXYNOS_DRM_OPS_MAX]; > + struct mutex cmd_lock; > + struct mutex mem_lock; > + struct mutex event_lock; > + struct completion start_complete; > + struct completion stop_complete; > + struct drm_exynos_ipp_property property; > + struct drm_exynos_ipp_cmd_work *start_work; > + struct drm_exynos_ipp_cmd_work *stop_work; > + struct drm_exynos_ipp_event_work *event_work; > + enum drm_exynos_ipp_state state; > +}; > + > +/* > + * A structure of buffer information. > + * > + * @gem_objs: Y, Cb, Cr each gem object. > + * @base: Y, Cb, Cr each planar address. > + */ > +struct drm_exynos_ipp_buf_info { > + unsigned long handles[EXYNOS_DRM_PLANAR_MAX]; > + dma_addr_t base[EXYNOS_DRM_PLANAR_MAX]; > +}; > + > +/* > + * A structure of wb setting infomation. > + * > + * @enable: enable flag for wb. > + * @refresh: HZ of the refresh rate. > + */ > +struct drm_exynos_ipp_set_wb { > + __u32 enable; > + __u32 refresh; > +}; > + > +/* > + * A structure of event work information. > + * > + * @work: work structure. > + * @ippdrv: current work ippdrv. > + * @buf_id: id of src, dst buffer. > + */ > +struct drm_exynos_ipp_event_work { > + struct work_struct work; > + struct exynos_drm_ippdrv *ippdrv; > + u32 buf_id[EXYNOS_DRM_OPS_MAX]; > +}; > + > +/* > + * A structure of source,destination operations. > + * > + * @set_fmt: set format of image. > + * @set_transf: set transform(rotations, flip). > + * @set_size: set size of region. > + * @set_addr: set address for dma. > + */ > +struct exynos_drm_ipp_ops { > + int (*set_fmt)(struct device *dev, u32 fmt); > + int (*set_transf)(struct device *dev, > + enum drm_exynos_degree degree, > + enum drm_exynos_flip flip); > + int (*set_size)(struct device *dev, int swap, > + struct drm_exynos_pos *pos, struct drm_exynos_sz *sz); > + int (*set_addr)(struct device *dev, > + struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, > + enum drm_exynos_ipp_buf_type buf_type); > +}; > + > +/* > + * A structure of ipp driver. > + * > + * @drv_list: list head for registed sub driver information. > + * @parent_dev: parent device information. > + * @dev: platform device. > + * @drm_dev: drm device. > + * @ipp_id: id of ipp driver. > + * @dedicated: dedicated ipp device. > + * @ops: source, destination operations. > + * @event_workq: event work queue. > + * @cmd: current command information. > + * @cmd_list: list head for command information. > + * @prop_list: property informations of current ipp driver. > + * @check_property: check property about format, size, buffer. > + * @reset: reset ipp block. > + * @start: ipp each device start. > + * @stop: ipp each device stop. > + * @sched_event: work schedule handler. > + */ > +struct exynos_drm_ippdrv { > + struct list_head drv_list; > + struct device *parent_dev; > + struct device *dev; > + struct drm_device *drm_dev; > + u32 ipp_id; > + bool dedicated; > + struct exynos_drm_ipp_ops *ops[EXYNOS_DRM_OPS_MAX]; > + struct workqueue_struct *event_workq; > + struct drm_exynos_ipp_cmd_node *cmd; > + struct list_head cmd_list; > + struct drm_exynos_ipp_prop_list *prop_list; > + > + int (*check_property)(struct device *dev, > + struct drm_exynos_ipp_property *property); > + int (*reset)(struct device *dev); > + int (*start)(struct device *dev, enum drm_exynos_ipp_cmd cmd); > + void (*stop)(struct device *dev, enum drm_exynos_ipp_cmd cmd); > + void (*sched_event)(struct work_struct *work); > +}; > + > +#ifdef CONFIG_DRM_EXYNOS_IPP > +extern int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv); > +extern int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv); > +extern int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, > + struct drm_file *file); > +extern int exynos_drm_ippnb_register(struct notifier_block *nb); > +extern int exynos_drm_ippnb_unregister(struct notifier_block *nb); > +extern int exynos_drm_ippnb_send_event(unsigned long val, void *v); > +extern void ipp_sched_cmd(struct work_struct *work); > +extern void ipp_sched_event(struct work_struct *work); > + > +#else > +static inline int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv > *ippdrv) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv > *ippdrv) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ipp_get_property(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file_priv) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_set_property(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file_priv) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ippnb_register(struct notifier_block *nb) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippnb_unregister(struct notifier_block *nb) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippnb_send_event(unsigned long val, void *v) > +{ > + return -ENOTTY; > +} > +#endif > + > +#endif /* _EXYNOS_DRM_IPP_H_ */ > + > diff --git a/include/uapi/drm/exynos_drm.h b/include/uapi/drm/exynos_drm.h > index c0494d5..82772d7 100644 > --- a/include/uapi/drm/exynos_drm.h > +++ b/include/uapi/drm/exynos_drm.h > @@ -154,6 +154,170 @@ struct drm_exynos_g2d_exec { > __u64 async; > }; > > +enum drm_exynos_ops_id { > + EXYNOS_DRM_OPS_SRC, > + EXYNOS_DRM_OPS_DST, > + EXYNOS_DRM_OPS_MAX, > +}; > + > +struct drm_exynos_sz { > + __u32 hsize; > + __u32 vsize; > +}; > + > +struct drm_exynos_pos { > + __u32 x; > + __u32 y; > + __u32 w; > + __u32 h; > +}; > + > +enum drm_exynos_flip { > + EXYNOS_DRM_FLIP_NONE = (0 << 0), > + EXYNOS_DRM_FLIP_VERTICAL = (1 << 0), > + EXYNOS_DRM_FLIP_HORIZONTAL = (1 << 1), > +}; > + > +enum drm_exynos_degree { > + EXYNOS_DRM_DEGREE_0, > + EXYNOS_DRM_DEGREE_90, > + EXYNOS_DRM_DEGREE_180, > + EXYNOS_DRM_DEGREE_270, > +}; > + > +enum drm_exynos_planer { > + EXYNOS_DRM_PLANAR_Y, > + EXYNOS_DRM_PLANAR_CB, > + EXYNOS_DRM_PLANAR_CR, > + EXYNOS_DRM_PLANAR_MAX, > +}; > + > +/** > + * A structure for ipp supported property list. > + * > + * @version: version of this structure. > + * @ipp_id: id of ipp driver. > + * @count: count of ipp driver. > + * @writeback: flag of writeback supporting. > + * @flip: flag of flip supporting. > + * @degree: flag of degree information. > + * @csc: flag of csc supporting. > + * @crop: flag of crop supporting. > + * @scale: flag of scale supporting. > + * @refresh_min: min hz of refresh. > + * @refresh_max: max hz of refresh. > + * @crop_min: crop min resolution. > + * @crop_max: crop max resolution. > + * @scale_min: scale min resolution. > + * @scale_max: scale max resolution. > + */ > +struct drm_exynos_ipp_prop_list { > + __u32 version; > + __u32 ipp_id; > + __u32 count; > + __u32 writeback; > + __u32 flip; > + __u32 degree; > + __u32 csc; > + __u32 crop; > + __u32 scale; > + __u32 refresh_min; > + __u32 refresh_max; > + __u32 reserved; > + struct drm_exynos_sz crop_min; > + struct drm_exynos_sz crop_max; > + struct drm_exynos_sz scale_min; > + struct drm_exynos_sz scale_max; > +}; > + > +/** > + * A structure for ipp config. > + * > + * @ops_id: property of operation directions. > + * @flip: property of mirror, flip. > + * @degree: property of rotation degree. > + * @fmt: property of image format. > + * @sz: property of image size. > + * @pos: property of image position(src-cropped,dst-scaler). > + */ > +struct drm_exynos_ipp_config { > + enum drm_exynos_ops_id ops_id; > + enum drm_exynos_flip flip; > + enum drm_exynos_degree degree; > + __u32 fmt; > + struct drm_exynos_sz sz; > + struct drm_exynos_pos pos; > +}; > + > +enum drm_exynos_ipp_cmd { > + IPP_CMD_NONE, > + IPP_CMD_M2M, > + IPP_CMD_WB, > + IPP_CMD_OUTPUT, > + IPP_CMD_MAX, > +}; > + > +/** > + * A structure for ipp property. > + * > + * @config: source, destination config. > + * @cmd: definition of command. > + * @ipp_id: id of ipp driver. > + * @prop_id: id of property. > + * @refresh_rate: refresh rate. > + */ > +struct drm_exynos_ipp_property { > + struct drm_exynos_ipp_config config[EXYNOS_DRM_OPS_MAX]; > + enum drm_exynos_ipp_cmd cmd; > + __u32 ipp_id; > + __u32 prop_id; > + __u32 refresh_rate; > +}; > + > +enum drm_exynos_ipp_buf_type { > + IPP_BUF_ENQUEUE, > + IPP_BUF_DEQUEUE, > +}; > + > +/** > + * A structure for ipp buffer operations. > + * > + * @ops_id: operation directions. > + * @buf_type: definition of buffer. > + * @prop_id: id of property. > + * @buf_id: id of buffer. > + * @handle: Y, Cb, Cr each planar handle. > + * @user_data: user data. > + */ > +struct drm_exynos_ipp_queue_buf { > + enum drm_exynos_ops_id ops_id; > + enum drm_exynos_ipp_buf_type buf_type; > + __u32 prop_id; > + __u32 buf_id; > + __u32 handle[EXYNOS_DRM_PLANAR_MAX]; > + __u32 reserved; > + __u64 user_data; > +}; > + > +enum drm_exynos_ipp_ctrl { > + IPP_CTRL_PLAY, > + IPP_CTRL_STOP, > + IPP_CTRL_PAUSE, > + IPP_CTRL_RESUME, > + IPP_CTRL_MAX, > +}; > + > +/** > + * A structure for ipp start/stop operations. > + * > + * @prop_id: id of property. > + * @ctrl: definition of control. > + */ > +struct drm_exynos_ipp_cmd_ctrl { > + __u32 prop_id; > + enum drm_exynos_ipp_ctrl ctrl; > +}; > + > #define DRM_EXYNOS_GEM_CREATE 0x00 > #define DRM_EXYNOS_GEM_MAP_OFFSET 0x01 > #define DRM_EXYNOS_GEM_MMAP 0x02 > @@ -166,6 +330,12 @@ struct drm_exynos_g2d_exec { > #define DRM_EXYNOS_G2D_SET_CMDLIST 0x21 > #define DRM_EXYNOS_G2D_EXEC 0x22 > > +/* IPP - Image Post Processing */ > +#define DRM_EXYNOS_IPP_GET_PROPERTY 0x30 > +#define DRM_EXYNOS_IPP_SET_PROPERTY 0x31 > +#define DRM_EXYNOS_IPP_QUEUE_BUF 0x32 > +#define DRM_EXYNOS_IPP_CMD_CTRL 0x33 > + > #define DRM_IOCTL_EXYNOS_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + > \ > DRM_EXYNOS_GEM_CREATE, struct drm_exynos_gem_create) > > @@ -188,8 +358,18 @@ struct drm_exynos_g2d_exec { > #define DRM_IOCTL_EXYNOS_G2D_EXEC DRM_IOWR(DRM_COMMAND_BASE + > \ > DRM_EXYNOS_G2D_EXEC, struct drm_exynos_g2d_exec) > > +#define DRM_IOCTL_EXYNOS_IPP_GET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_GET_PROPERTY, struct drm_exynos_ipp_prop_list) > +#define DRM_IOCTL_EXYNOS_IPP_SET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_SET_PROPERTY, struct drm_exynos_ipp_property) > +#define DRM_IOCTL_EXYNOS_IPP_QUEUE_BUF DRM_IOWR(DRM_COMMAND_BASE + \ > + DRM_EXYNOS_IPP_QUEUE_BUF, struct drm_exynos_ipp_queue_buf) > +#define DRM_IOCTL_EXYNOS_IPP_CMD_CTRL DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_CMD_CTRL, struct drm_exynos_ipp_cmd_ctrl) > + > /* EXYNOS specific events */ > #define DRM_EXYNOS_G2D_EVENT 0x80000000 > +#define DRM_EXYNOS_IPP_EVENT 0x80000001 > > struct drm_exynos_g2d_event { > struct drm_event base; > @@ -200,4 +380,14 @@ struct drm_exynos_g2d_event { > __u32 reserved; > }; > > +struct drm_exynos_ipp_event { > + struct drm_event base; > + __u64 user_data; > + __u32 tv_sec; > + __u32 tv_usec; > + __u32 prop_id; > + __u32 reserved; > + __u32 buf_id[EXYNOS_DRM_OPS_MAX]; > +}; > + > #endif /* _UAPI_EXYNOS_DRM_H_ */ > -- > 1.7.0.4 _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx http://lists.freedesktop.org/mailman/listinfo/dri-devel