[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <85edcca3-cd51-441e-9bf8-6b86f8fe949f@amd.com>
Date: Fri, 13 Jun 2025 10:42:53 +0800
From: "Du, Bin" <bin.du@....com>
To: mchehab@...nel.org, hverkuil@...all.nl,
laurent.pinchart+renesas@...asonboard.com, bryan.odonoghue@...aro.org,
sakari.ailus@...ux.intel.com, prabhakar.mahadev-lad.rj@...renesas.com,
linux-media@...r.kernel.org, linux-kernel@...r.kernel.org
Cc: pratap.nirujogi@....com, benjamin.chan@....com, king.li@....com,
gjorgji.rosikopulos@....com, Phil.Jawich@....com, Dominic.Antony@....com,
Svetoslav.Stoilov@....com
Subject: Re: [PATCH v1 6/9] media: platform: amd: isp4 video node and buffers
handling added
Change-Id should be removed and will be fixed in V2
On 6/8/2025 10:49 PM, Bin Du wrote:
> Isp video implements v4l2 video interface and supports NV12 and YUYV.
> It manages buffers, pipeline power and state.
>
> Signed-off-by: Bin Du <Bin.Du@....com>
> Signed-off-by: Svetoslav Stoilov <Svetoslav.Stoilov@....com>
> Change-Id: I42178ebe877dd1901bd801a23d8f825827282e56
> ---
> drivers/media/platform/amd/isp4/Makefile | 1 +
> drivers/media/platform/amd/isp4/isp4.c | 13 +-
> drivers/media/platform/amd/isp4/isp4_subdev.c | 128 +-
> drivers/media/platform/amd/isp4/isp4_subdev.h | 2 +
> drivers/media/platform/amd/isp4/isp4_video.c | 1457 +++++++++++++++++
> drivers/media/platform/amd/isp4/isp4_video.h | 93 ++
> 6 files changed, 1684 insertions(+), 10 deletions(-)
> create mode 100644 drivers/media/platform/amd/isp4/isp4_video.c
> create mode 100644 drivers/media/platform/amd/isp4/isp4_video.h
>
> diff --git a/drivers/media/platform/amd/isp4/Makefile b/drivers/media/platform/amd/isp4/Makefile
> index d2f1523ad07a..a68f18fe79b4 100644
> --- a/drivers/media/platform/amd/isp4/Makefile
> +++ b/drivers/media/platform/amd/isp4/Makefile
> @@ -7,6 +7,7 @@ amd_capture-objs := isp4_subdev.o \
> isp4_phy.o \
> isp4_interface.o \
> isp4.o \
> + isp4_video.o \
> isp4_hw.o \
>
> ccflags-y += -I$(srctree)/drivers/media/platform/amd/isp4
> diff --git a/drivers/media/platform/amd/isp4/isp4.c b/drivers/media/platform/amd/isp4/isp4.c
> index 2dc7ea3b02e8..3beb35293504 100644
> --- a/drivers/media/platform/amd/isp4/isp4.c
> +++ b/drivers/media/platform/amd/isp4/isp4.c
> @@ -310,6 +310,16 @@ static int isp4_capture_probe(struct platform_device *pdev)
> goto err_isp4_deinit;
> }
>
> + ret = media_create_pad_link(&isp_dev->isp_sdev.sdev.entity,
> + 1, &isp_dev->isp_sdev.isp_vdev.vdev.entity,
> + 0,
> + MEDIA_LNK_FL_ENABLED |
> + MEDIA_LNK_FL_IMMUTABLE);
> + if (ret) {
> + dev_err(dev, "fail to create pad link %d\n", ret);
> + goto err_unreg_video_dev_notifier;
> + }
> +
> platform_set_drvdata(pdev, isp_dev);
>
> pm_runtime_set_suspended(dev);
> @@ -317,9 +327,10 @@ static int isp4_capture_probe(struct platform_device *pdev)
>
> return 0;
>
> -err_isp4_deinit:
> +err_unreg_video_dev_notifier:
> v4l2_async_nf_unregister(&isp_dev->notifier);
> v4l2_async_nf_cleanup(&isp_dev->notifier);
> +err_isp4_deinit:
> isp4sd_deinit(&isp_dev->isp_sdev);
> err_unreg_v4l2:
> v4l2_device_unregister(&isp_dev->v4l2_dev);
> diff --git a/drivers/media/platform/amd/isp4/isp4_subdev.c b/drivers/media/platform/amd/isp4/isp4_subdev.c
> index 4b28fbd6867b..816fa3a127f5 100644
> --- a/drivers/media/platform/amd/isp4/isp4_subdev.c
> +++ b/drivers/media/platform/amd/isp4/isp4_subdev.c
> @@ -167,6 +167,24 @@ static int isp4sd_enum_mbus_code(struct v4l2_subdev *isp_subdev,
> return 0;
> }
>
> +static int isp4sd_enum_frame_size(struct v4l2_subdev *isp_subdev,
> + struct v4l2_subdev_state *state,
> + struct v4l2_subdev_frame_size_enum *fse)
> +{
> + struct v4l2_frmsize_discrete min, max;
> +
> + if (fse->index >= ARRAY_SIZE(isp4sd_image_formats))
> + return -EINVAL;
> +
> + isp4vid_frmsize_range(&min, &max);
> + fse->min_width = min.width;
> + fse->max_width = max.width;
> + fse->min_height = min.height;
> + fse->max_height = max.height;
> +
> + return 0;
> +}
> +
> static bool isp4sd_get_str_out_prop(struct isp4_subdev *isp_subdev,
> struct isp4fw_image_prop *out_prop,
> struct v4l2_subdev_state *state, u32 pad)
> @@ -509,17 +527,14 @@ isp4sd_get_meta_by_mc(struct isp4_subdev *isp_subdev,
>
> static struct isp4if_img_buf_node *
> isp4sd_preview_done(struct isp4_subdev *isp_subdev,
> - struct isp4fw_meta_info *meta)
> + struct isp4fw_meta_info *meta,
> + struct isp4vid_framedone_param *pcb)
> {
> struct isp4_interface *ispif = &isp_subdev->ispif;
> struct isp4if_img_buf_node *prev = NULL;
> struct device *dev = isp_subdev->dev;
>
> - if (!meta) {
> - dev_err(dev, "fail bad param for preview done\n");
> - return prev;
> - }
> -
> + pcb->preview.status = ISP4VID_BUF_DONE_STATUS_ABSENT;
> if (meta->preview.enabled &&
> (meta->preview.status == BUFFER_STATUS_SKIPPED ||
> meta->preview.status == BUFFER_STATUS_DONE ||
> @@ -529,9 +544,12 @@ isp4sd_preview_done(struct isp4_subdev *isp_subdev,
> str_info = &isp_subdev->sensor_info.output_info;
> prev = isp4if_dequeue_buffer(ispif);
>
> - if (!prev)
> + if (!prev) {
> dev_err(dev, "fail null prev buf\n");
> -
> + } else {
> + pcb->preview.buf = prev->buf_info;
> + pcb->preview.status = ISP4VID_BUF_DONE_STATUS_SUCCESS;
> + }
> } else if (meta->preview.enabled) {
> dev_err(dev, "fail bad preview status %u\n",
> meta->preview.status);
> @@ -578,6 +596,7 @@ static void isp4sd_fw_resp_frame_done(struct isp4_subdev *isp_subdev,
> enum isp4if_stream_id stream_id,
> struct isp4fw_resp_param_package *para)
> {
> + struct isp4vid_framedone_param pcb = {0};
> struct isp4if_img_buf_node *prev = NULL;
> struct device *dev = isp_subdev->dev;
> struct isp4fw_meta_info *meta;
> @@ -590,12 +609,17 @@ static void isp4sd_fw_resp_frame_done(struct isp4_subdev *isp_subdev,
> return;
> }
>
> + pcb.poc = meta->poc;
> + pcb.cam_id = 0;
> +
> dev_dbg(dev, "ts:%llu,streamId:%d,poc:%u,preview_en:%u,(%i)\n",
> ktime_get_ns(), stream_id, meta->poc,
> meta->preview.enabled,
> meta->preview.status);
>
> - prev = isp4sd_preview_done(isp_subdev, meta);
> + prev = isp4sd_preview_done(isp_subdev, meta, &pcb);
> + if (pcb.preview.status != ISP4VID_BUF_DONE_STATUS_ABSENT)
> + isp4vid_notify(&isp_subdev->isp_vdev, &pcb);
>
> isp4if_dealloc_buffer_node(prev);
>
> @@ -974,6 +998,84 @@ static int isp4sd_start_stream(struct isp4_subdev *isp_subdev,
> return ret;
> }
>
> +static int ispsd4_subdev_link_validate(struct media_link *link)
> +{
> + return 0;
> +}
> +
> +static const struct media_entity_operations isp4sd_subdev_ent_ops = {
> + .link_validate = ispsd4_subdev_link_validate,
> +};
> +
> +static int isp4sd_ioc_send_img_buf(struct isp4vid_dev *sd,
> + struct isp4if_img_buf_info *buf_info)
> +{
> + struct isp4_subdev *isp_subdev = to_isp4_vdev(sd);
> + struct isp4_interface *ispif = &isp_subdev->ispif;
> + struct isp4if_img_buf_node *buf_node = NULL;
> + struct device *dev = isp_subdev->dev;
> + int ret = -EINVAL;
> +
> + mutex_lock(&isp_subdev->ops_mutex);
> + /* TODO: remove isp_status */
> + if (ispif->status != ISP4IF_STATUS_FW_RUNNING) {
> + dev_err(dev, "fail send img buf for bad fsm %d\n",
> + ispif->status);
> + mutex_unlock(&isp_subdev->ops_mutex);
> + return -EINVAL;
> + }
> +
> + buf_node = isp4if_alloc_buffer_node(buf_info);
> + if (!buf_node) {
> + dev_err(dev, "fail alloc sys img buf info node\n");
> + ret = -ENOMEM;
> + goto unlock_and_return;
> + }
> +
> + ret = isp4if_queue_buffer(ispif, buf_node);
> + if (ret) {
> + dev_err(dev, "fail to queue image buf, %d\n", ret);
> + goto error_release_buf_node;
> + }
> +
> + if (!isp_subdev->sensor_info.start_stream_cmd_sent) {
> + isp_subdev->sensor_info.buf_sent_cnt++;
> +
> + if (isp_subdev->sensor_info.buf_sent_cnt >=
> + ISP4SD_MIN_BUF_CNT_BEF_START_STREAM) {
> + ret = isp4if_send_command(ispif, CMD_ID_START_STREAM,
> + NULL, 0);
> +
> + if (ret) {
> + dev_err(dev, "fail to START_STREAM");
> + goto error_release_buf_node;
> + }
> + isp_subdev->sensor_info.start_stream_cmd_sent = true;
> + isp_subdev->sensor_info.output_info.start_status =
> + ISP4SD_START_STATUS_STARTED;
> + isp_subdev->sensor_info.status =
> + ISP4SD_START_STATUS_STARTED;
> + } else {
> + dev_dbg(dev,
> + "no send start,required %u,buf sent %u\n",
> + ISP4SD_MIN_BUF_CNT_BEF_START_STREAM,
> + isp_subdev->sensor_info.buf_sent_cnt);
> + }
> + }
> +
> + mutex_unlock(&isp_subdev->ops_mutex);
> +
> + return 0;
> +
> +error_release_buf_node:
> + isp4if_dealloc_buffer_node(buf_node);
> +
> +unlock_and_return:
> + mutex_unlock(&isp_subdev->ops_mutex);
> +
> + return ret;
> +}
> +
> static int isp4sd_subdev_pre_streamon(struct v4l2_subdev *sd, u32 flags)
> {
> struct isp4_subdev *isp_subdev = to_isp4_subdev(sd);
> @@ -1098,6 +1200,7 @@ static int isp4sd_disable_streams(struct v4l2_subdev *sd,
>
> static const struct v4l2_subdev_pad_ops isp4sd_pad_ops = {
> .enum_mbus_code = isp4sd_enum_mbus_code,
> + .enum_frame_size = isp4sd_enum_frame_size,
> .get_fmt = v4l2_subdev_get_fmt,
> .set_fmt = isp4sd_set_pad_format,
> .enable_streams = isp4sd_enable_streams,
> @@ -1123,6 +1226,10 @@ static const struct media_entity_operations isp4sd_sdev_ent_ops = {
> .link_validate = isp4sd_sdev_link_validate,
> };
>
> +static const struct isp4vid_ops isp4sd_isp4vid_ops = {
> + .send_buffer = isp4sd_ioc_send_img_buf,
> +};
> +
> int isp4sd_init(struct isp4_subdev *isp_subdev,
> struct v4l2_device *v4l2_dev,
> void *amdgpu_dev)
> @@ -1188,6 +1295,8 @@ int isp4sd_init(struct isp4_subdev *isp_subdev,
> isp_subdev->host2fw_seq_num = 1;
> ispif->status = ISP4IF_STATUS_PWR_OFF;
>
> + ret = isp4vid_dev_init(&isp_subdev->isp_vdev, &isp_subdev->sdev,
> + &isp4sd_isp4vid_ops, amdgpu_dev);
> if (ret)
> goto err_media_clean_up;
> return ret;
> @@ -1203,6 +1312,7 @@ void isp4sd_deinit(struct isp4_subdev *isp_subdev)
> struct isp4sd_output_info *output_info;
>
> output_info = &isp_subdev->sensor_info.output_info;
> + isp4vid_dev_deinit(&isp_subdev->isp_vdev);
> media_entity_cleanup(&isp_subdev->sdev.entity);
>
> isp4if_deinit(ispif);
> diff --git a/drivers/media/platform/amd/isp4/isp4_subdev.h b/drivers/media/platform/amd/isp4/isp4_subdev.h
> index 99ec914a95ce..e04811dd2591 100644
> --- a/drivers/media/platform/amd/isp4/isp4_subdev.h
> +++ b/drivers/media/platform/amd/isp4/isp4_subdev.h
> @@ -18,6 +18,7 @@
> #include "isp4_fw_cmd_resp.h"
> #include "isp4_hw_reg.h"
> #include "isp4_interface.h"
> +#include "isp4_video.h"
>
> /*
> * one is for none sesnor specefic response which is not used now
> @@ -97,6 +98,7 @@ struct isp4_subdev_thread_param {
> struct isp4_subdev {
> struct v4l2_subdev sdev;
> struct isp4_interface ispif;
> + struct isp4vid_dev isp_vdev;
>
> /*
> * sdev_pad[0] is sink pad connected to sensor
> diff --git a/drivers/media/platform/amd/isp4/isp4_video.c b/drivers/media/platform/amd/isp4/isp4_video.c
> new file mode 100644
> index 000000000000..39f0b6e49713
> --- /dev/null
> +++ b/drivers/media/platform/amd/isp4/isp4_video.c
> @@ -0,0 +1,1457 @@
> +// SPDX-License-Identifier: GPL-2.0+
> +/*
> + * Copyright (C) 2025 Advanced Micro Devices, Inc.
> + */
> +
> +#include <linux/pm_runtime.h>
> +#include <linux/vmalloc.h>
> +#include <media/v4l2-ioctl.h>
> +#include <media/v4l2-mc.h>
> +
> +#include "amdgpu_object.h"
> +#include "isp4_interface.h"
> +#include "isp4_subdev.h"
> +#include "isp4_video.h"
> +
> +#define ISP4VID_ISP_DRV_NAME "amd_isp_capture"
> +#define ISP4VID_VIDEO_BUF_NUM 5
> +#define ISP4VID_MAX_PREVIEW_FPS 30
> +#define ISP4VID_DEFAULT_FMT isp4vid_formats[0]
> +
> +#define ISP4VID_PAD_VIDEO_OUTPUT 1
> +
> +/* timeperframe default */
> +#define ISP4VID_ISP_TPF_DEFAULT isp4vid_tpfs[0]
> +
> +/* amdisp buffer for vb2 operations */
> +struct isp4vid_vb2_buf {
> + struct device *dev;
> + void *vaddr;
> + struct frame_vector *vec;
> + enum dma_data_direction dma_dir;
> + unsigned long size;
> + refcount_t refcount;
> + struct dma_buf *dbuf;
> + void *bo;
> + u64 gpu_addr;
> + struct vb2_vmarea_handler handler;
> + bool is_expbuf;
> +};
> +
> +static int isp4vid_vb2_mmap(void *buf_priv, struct vm_area_struct *vma);
> +
> +static void isp4vid_vb2_put(void *buf_priv);
> +
> +static const char *isp4vid_video_dev_name = "Preview";
> +
> +/* Sizes must be in increasing order */
> +static const struct v4l2_frmsize_discrete isp4vid_frmsize[] = {
> + {640, 360},
> + {640, 480},
> + {1280, 720},
> + {1280, 960},
> + {1920, 1080},
> + {1920, 1440},
> + {2560, 1440},
> + {2880, 1620},
> + {2880, 1624},
> + {2888, 1808},
> +};
> +
> +static const u32 isp4vid_formats[] = {
> + V4L2_PIX_FMT_NV12,
> + V4L2_PIX_FMT_YUYV
> +};
> +
> +/* timeperframe list */
> +static const struct v4l2_fract isp4vid_tpfs[] = {
> + {.numerator = 1, .denominator = ISP4VID_MAX_PREVIEW_FPS}
> +};
> +
> +static void
> +isp4vid_handle_frame_done(struct isp4vid_dev *isp_vdev,
> + const struct isp4if_img_buf_info *img_buf,
> + bool done_suc)
> +{
> + struct isp4vid_capture_buffer *isp4vid_buf;
> + enum vb2_buffer_state state;
> + void *vbuf;
> +
> + mutex_lock(&isp_vdev->buf_list_lock);
> +
> + /* Get the first entry of the list */
> + isp4vid_buf = list_first_entry_or_null(&isp_vdev->buf_list,
> + typeof(*isp4vid_buf), list);
> + if (!isp4vid_buf) {
> + mutex_unlock(&isp_vdev->buf_list_lock);
> + return;
> + }
> +
> + vbuf = vb2_plane_vaddr(&isp4vid_buf->vb2.vb2_buf, 0);
> +
> + if (vbuf != img_buf->planes[0].sys_addr) {
> + dev_err(isp_vdev->dev, "Invalid vbuf");
> + mutex_unlock(&isp_vdev->buf_list_lock);
> + return;
> + }
> +
> + /* Remove this entry from the list */
> + list_del(&isp4vid_buf->list);
> +
> + mutex_unlock(&isp_vdev->buf_list_lock);
> +
> + /* Fill the buffer */
> + isp4vid_buf->vb2.vb2_buf.timestamp = ktime_get_ns();
> + isp4vid_buf->vb2.sequence = isp_vdev->sequence++;
> + isp4vid_buf->vb2.field = V4L2_FIELD_ANY;
> +
> + /* at most 2 planes */
> + vb2_set_plane_payload(&isp4vid_buf->vb2.vb2_buf,
> + 0, isp_vdev->format.sizeimage);
> +
> + state = done_suc ? VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
> + vb2_buffer_done(&isp4vid_buf->vb2.vb2_buf, state);
> +
> + dev_dbg(isp_vdev->dev, "call vb2_buffer_done(size=%u)\n",
> + isp_vdev->format.sizeimage);
> +}
> +
> +s32 isp4vid_notify(void *cb_ctx, struct isp4vid_framedone_param *evt_param)
> +{
> + struct isp4vid_dev *isp4vid_vdev = cb_ctx;
> +
> + if (evt_param->preview.status != ISP4VID_BUF_DONE_STATUS_ABSENT) {
> + bool suc;
> +
> + suc = (evt_param->preview.status ==
> + ISP4VID_BUF_DONE_STATUS_SUCCESS);
> + isp4vid_handle_frame_done(isp4vid_vdev,
> + &evt_param->preview.buf,
> + suc);
> + }
> +
> + return 0;
> +}
> +
> +void isp4vid_frmsize_range(struct v4l2_frmsize_discrete *min,
> + struct v4l2_frmsize_discrete *max)
> +{
> + if (!min || !max) {
> + pr_err("fail, null param\n");
> + return;
> + }
> +
> + min->width = isp4vid_frmsize[0].width;
> + min->height = isp4vid_frmsize[0].height;
> + max->width = isp4vid_frmsize[ARRAY_SIZE(isp4vid_frmsize) - 1].width;
> + max->height = isp4vid_frmsize[ARRAY_SIZE(isp4vid_frmsize) - 1].height;
> +}
> +
> +static unsigned int isp4vid_vb2_num_users(void *buf_priv)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> +
> + if (!buf) {
> + pr_err("fail null buf handle\n");
> + return 0;
> + }
> + return refcount_read(&buf->refcount);
> +}
> +
> +static int isp4vid_vb2_mmap(void *buf_priv, struct vm_area_struct *vma)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> + int ret;
> +
> + if (!buf) {
> + pr_err("fail no memory to map\n");
> + return -EINVAL;
> + }
> +
> + ret = remap_vmalloc_range(vma, buf->vaddr, 0);
> + if (ret) {
> + dev_err(buf->dev, "fail remap vmalloc mem, %d\n", ret);
> + return ret;
> + }
> +
> + /*
> + * Make sure that vm_areas for 2 buffers won't be merged together
> + */
> + vm_flags_set(vma, VM_DONTEXPAND);
> +
> + dev_dbg(buf->dev, "mmap isp user bo 0x%llx size %ld refcount %d\n",
> + buf->gpu_addr, buf->size, buf->refcount.refs.counter);
> +
> + return 0;
> +}
> +
> +static void *isp4vid_vb2_vaddr(struct vb2_buffer *vb, void *buf_priv)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> +
> + if (!buf) {
> + pr_err("fail for empty buf\n");
> + return NULL;
> + }
> +
> + if (!buf->vaddr) {
> + dev_err(buf->dev,
> + "fail for buf vaddr is null\n");
> + return NULL;
> + }
> + return buf->vaddr;
> +}
> +
> +static void isp4vid_vb2_detach_dmabuf(void *mem_priv)
> +{
> + struct isp4vid_vb2_buf *buf = mem_priv;
> +
> + if (!buf) {
> + pr_err("fail invalid buf handle\n");
> + return;
> + }
> +
> + struct iosys_map map = IOSYS_MAP_INIT_VADDR(buf->vaddr);
> +
> + dev_dbg(buf->dev, "detach dmabuf of isp user bo 0x%llx size %ld",
> + buf->gpu_addr, buf->size);
> +
> + if (buf->vaddr)
> + dma_buf_vunmap_unlocked(buf->dbuf, &map);
> +
> + // put dmabuf for exported ones
> + dma_buf_put(buf->dbuf);
> +
> + kfree(buf);
> +}
> +
> +static void *isp4vid_vb2_attach_dmabuf(struct vb2_buffer *vb,
> + struct device *dev,
> + struct dma_buf *dbuf,
> + unsigned long size)
> +{
> + struct isp4vid_vb2_buf *buf;
> +
> + if (dbuf->size < size) {
> + dev_err(dev, "Invalid dmabuf size %ld %ld", dbuf->size, size);
> + return ERR_PTR(-EFAULT);
> + }
> +
> + buf = kzalloc(sizeof(*buf), GFP_KERNEL);
> + if (!buf)
> + return ERR_PTR(-ENOMEM);
> +
> + struct isp4vid_vb2_buf *dbg_buf = (struct isp4vid_vb2_buf *)dbuf->priv;
> +
> + buf->dev = dev;
> + buf->dbuf = dbuf;
> + buf->dma_dir = vb->vb2_queue->dma_dir;
> + buf->size = size;
> +
> + dev_dbg(dev, "attach dmabuf of isp user bo 0x%llx size %ld",
> + dbg_buf->gpu_addr, dbg_buf->size);
> +
> + return buf;
> +}
> +
> +static void isp4vid_vb2_unmap_dmabuf(void *mem_priv)
> +{
> + struct isp4vid_vb2_buf *buf = mem_priv;
> +
> + if (!buf) {
> + pr_err("fail invalid buf handle\n");
> + return;
> + }
> +
> + struct iosys_map map = IOSYS_MAP_INIT_VADDR(buf->vaddr);
> +
> + dev_dbg(buf->dev, "unmap dmabuf of isp user bo 0x%llx size %ld",
> + buf->gpu_addr, buf->size);
> +
> + dma_buf_vunmap_unlocked(buf->dbuf, &map);
> + buf->vaddr = NULL;
> +}
> +
> +static int isp4vid_vb2_map_dmabuf(void *mem_priv)
> +{
> + struct isp4vid_vb2_buf *mmap_buf = NULL;
> + struct isp4vid_vb2_buf *buf = mem_priv;
> + struct iosys_map map;
> + int ret;
> +
> + memset(&map, 0x0, sizeof(map));
> +
> + if (!buf) {
> + pr_err("fail invalid buf handle\n");
> + return -EINVAL;
> + }
> +
> + ret = dma_buf_vmap_unlocked(buf->dbuf, &map);
> + if (ret) {
> + dev_err(buf->dev, "vmap_unlocked fail");
> + return -EFAULT;
> + }
> + buf->vaddr = map.vaddr;
> +
> + mmap_buf = (struct isp4vid_vb2_buf *)buf->dbuf->priv;
> + buf->gpu_addr = mmap_buf->gpu_addr;
> +
> + dev_dbg(buf->dev, "map dmabuf of isp user bo 0x%llx size %ld",
> + buf->gpu_addr, buf->size);
> +
> + return 0;
> +}
> +
> +#ifdef CONFIG_HAS_DMA
> +struct isp4vid_vb2_amdgpu_attachment {
> + struct sg_table sgt;
> + enum dma_data_direction dma_dir;
> +};
> +
> +static int isp4vid_vb2_dmabuf_ops_attach(struct dma_buf *dbuf,
> + struct dma_buf_attachment *dbuf_attach)
> +{
> + struct isp4vid_vb2_buf *buf = dbuf->priv;
> + int num_pages = PAGE_ALIGN(buf->size) / PAGE_SIZE;
> + struct isp4vid_vb2_amdgpu_attachment *attach;
> + void *vaddr = buf->vaddr;
> + struct scatterlist *sg;
> + struct sg_table *sgt;
> + int ret;
> + int i;
> +
> + attach = kzalloc(sizeof(*attach), GFP_KERNEL);
> + if (!attach)
> + return -ENOMEM;
> +
> + sgt = &attach->sgt;
> + ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL);
> + if (ret) {
> + kfree(attach);
> + return ret;
> + }
> + for_each_sgtable_sg(sgt, sg, i) {
> + struct page *page = vmalloc_to_page(vaddr);
> +
> + if (!page) {
> + sg_free_table(sgt);
> + kfree(attach);
> + return -ENOMEM;
> + }
> + sg_set_page(sg, page, PAGE_SIZE, 0);
> + vaddr = ((char *)vaddr) + PAGE_SIZE;
> + }
> +
> + attach->dma_dir = DMA_NONE;
> + dbuf_attach->priv = attach;
> +
> + return 0;
> +}
> +
> +static void
> +isp4vid_vb2_dmabuf_ops_detach(struct dma_buf *dbuf,
> + struct dma_buf_attachment *dbuf_attach)
> +{
> + struct isp4vid_vb2_amdgpu_attachment *attach = dbuf_attach->priv;
> + struct sg_table *sgt;
> +
> + if (!attach) {
> + pr_err("fail invalid attach handler\n");
> + return;
> + }
> +
> + sgt = &attach->sgt;
> +
> + /* release the scatterlist cache */
> + if (attach->dma_dir != DMA_NONE)
> + dma_unmap_sgtable(dbuf_attach->dev, sgt, attach->dma_dir, 0);
> + sg_free_table(sgt);
> + kfree(attach);
> + dbuf_attach->priv = NULL;
> +}
> +
> +static struct sg_table
> +*isp4vid_vb2_dmabuf_ops_map(struct dma_buf_attachment *dbuf_attach,
> + enum dma_data_direction dma_dir)
> +{
> + struct isp4vid_vb2_amdgpu_attachment *attach = dbuf_attach->priv;
> + struct sg_table *sgt;
> +
> + sgt = &attach->sgt;
> + /* return previously mapped sg table */
> + if (attach->dma_dir == dma_dir)
> + return sgt;
> +
> + /* release any previous cache */
> + if (attach->dma_dir != DMA_NONE) {
> + dma_unmap_sgtable(dbuf_attach->dev, sgt, attach->dma_dir, 0);
> + attach->dma_dir = DMA_NONE;
> + }
> +
> + /* mapping to the client with new direction */
> + if (dma_map_sgtable(dbuf_attach->dev, sgt, dma_dir, 0)) {
> + dev_err(dbuf_attach->dev, "fail to map scatterlist");
> + return ERR_PTR(-EIO);
> + }
> +
> + attach->dma_dir = dma_dir;
> +
> + return sgt;
> +}
> +
> +static void isp4vid_vb2_dmabuf_ops_unmap(struct dma_buf_attachment *dbuf_attach,
> + struct sg_table *sgt,
> + enum dma_data_direction dma_dir)
> +{
> + /* nothing to be done here */
> +}
> +
> +static int isp4vid_vb2_dmabuf_ops_vmap(struct dma_buf *dbuf,
> + struct iosys_map *map)
> +{
> + struct isp4vid_vb2_buf *buf = dbuf->priv;
> +
> + iosys_map_set_vaddr(map, buf->vaddr);
> +
> + return 0;
> +}
> +
> +static void isp4vid_vb2_dmabuf_ops_release(struct dma_buf *dbuf)
> +{
> + struct isp4vid_vb2_buf *buf = dbuf->priv;
> +
> + /* drop reference obtained in vb2_isp4vid_get_dmabuf */
> + if (buf->is_expbuf)
> + isp4vid_vb2_put(dbuf->priv);
> + else
> + dev_dbg(buf->dev, "ignore buf release for implicit case");
> +}
> +
> +static int isp4vid_vb2_dmabuf_ops_mmap(struct dma_buf *dbuf,
> + struct vm_area_struct *vma)
> +{
> + return isp4vid_vb2_mmap(dbuf->priv, vma);
> +}
> +
> +static const struct dma_buf_ops vb2_isp4vid_dmabuf_ops = {
> + .attach = isp4vid_vb2_dmabuf_ops_attach,
> + .detach = isp4vid_vb2_dmabuf_ops_detach,
> + .map_dma_buf = isp4vid_vb2_dmabuf_ops_map,
> + .unmap_dma_buf = isp4vid_vb2_dmabuf_ops_unmap,
> + .vmap = isp4vid_vb2_dmabuf_ops_vmap,
> + .mmap = isp4vid_vb2_dmabuf_ops_mmap,
> + .release = isp4vid_vb2_dmabuf_ops_release,
> +};
> +
> +static struct dma_buf *isp4vid_get_dmabuf(struct vb2_buffer *vb,
> + void *buf_priv,
> + unsigned long flags)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> + DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
> + struct dma_buf *dbuf;
> +
> + if (!buf) {
> + pr_err("fail invalid buf handle\n");
> + return ERR_PTR(-EINVAL);
> + }
> +
> + exp_info.ops = &vb2_isp4vid_dmabuf_ops;
> + exp_info.size = buf->size;
> + exp_info.flags = flags;
> + exp_info.priv = buf;
> +
> + if (WARN_ON(!buf->vaddr))
> + return NULL;
> +
> + dbuf = dma_buf_export(&exp_info);
> + if (IS_ERR(dbuf))
> + return NULL;
> +
> + return dbuf;
> +}
> +
> +static struct dma_buf *isp4vid_vb2_get_dmabuf(struct vb2_buffer *vb,
> + void *buf_priv,
> + unsigned long flags)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> + struct dma_buf *dbuf;
> +
> + if (buf->dbuf) {
> + dev_dbg(buf->dev,
> + "dbuf already created, reuse implicit dbuf\n");
> + dbuf = buf->dbuf;
> + } else {
> + dbuf = isp4vid_get_dmabuf(vb, buf_priv, flags);
> + dev_dbg(buf->dev, "created new dbuf\n");
> + }
> + buf->is_expbuf = true;
> + refcount_inc(&buf->refcount);
> +
> + dev_dbg(buf->dev, "buf exported, refcount %d\n",
> + buf->refcount.refs.counter);
> +
> + return dbuf;
> +}
> +
> +#endif
> +
> +static void isp4vid_vb2_put_userptr(void *buf_priv)
> +{
> + struct isp4vid_vb2_buf *buf = buf_priv;
> +
> + if (!buf->vec->is_pfns) {
> + unsigned long vaddr = (unsigned long)buf->vaddr & PAGE_MASK;
> + unsigned int n_pages;
> +
> + n_pages = frame_vector_count(buf->vec);
> + if (vaddr)
> + vm_unmap_ram((void *)vaddr, n_pages);
> + if (buf->dma_dir == DMA_FROM_DEVICE ||
> + buf->dma_dir == DMA_BIDIRECTIONAL) {
> + struct page **pages;
> +
> + pages = frame_vector_pages(buf->vec);
> + if (!WARN_ON_ONCE(IS_ERR(pages))) {
> + unsigned int i;
> +
> + for (i = 0; i < n_pages; i++)
> + set_page_dirty_lock(pages[i]);
> + }
> + }
> + } else {
> + iounmap((__force void __iomem *)buf->vaddr);
> + }
> + vb2_destroy_framevec(buf->vec);
> + kfree(buf);
> +}
> +
> +static void *isp4vid_vb2_get_userptr(struct vb2_buffer *vb, struct device *dev,
> + unsigned long vaddr, unsigned long size)
> +{
> + struct isp4vid_vb2_buf *buf;
> + struct frame_vector *vec;
> + int n_pages, offset, i;
> + int ret = -ENOMEM;
> +
> + buf = kzalloc(sizeof(*buf), GFP_KERNEL);
> + if (!buf)
> + return ERR_PTR(-ENOMEM);
> +
> + buf->dev = dev;
> + buf->dma_dir = vb->vb2_queue->dma_dir;
> + offset = vaddr & ~PAGE_MASK;
> + buf->size = size;
> + vec = vb2_create_framevec(vaddr, size,
> + buf->dma_dir == DMA_FROM_DEVICE ||
> + buf->dma_dir == DMA_BIDIRECTIONAL);
> + if (IS_ERR(vec)) {
> + kfree(buf);
> + return vec;
> + }
> + buf->vec = vec;
> + n_pages = frame_vector_count(vec);
> + if (frame_vector_to_pages(vec) < 0) {
> + unsigned long *nums = frame_vector_pfns(vec);
> +
> + /*
> + * We cannot get page pointers for these pfns. Check memory is
> + * physically contiguous and use direct mapping.
> + */
> + for (i = 1; i < n_pages; i++)
> + if (nums[i - 1] + 1 != nums[i])
> + goto err_destroy_free;
> + buf->vaddr = (__force void *)
> + ioremap(__pfn_to_phys(nums[0]), size + offset);
> + } else {
> + buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1);
> + }
> +
> + if (!buf->vaddr)
> + goto err_destroy_free;
> +
> + buf->vaddr = ((char *)buf->vaddr) + offset;
> + return buf;
> +
> +err_destroy_free:
> + vb2_destroy_framevec(vec);
> + kfree(buf);
> + return ERR_PTR(ret);
> +}
> +
> +static void isp4vid_vb2_put(void *buf_priv)
> +{
> + struct isp4vid_vb2_buf *buf = (struct isp4vid_vb2_buf *)buf_priv;
> + struct amdgpu_bo *bo = (struct amdgpu_bo *)buf->bo;
> +
> + dev_dbg(buf->dev,
> + "release isp user bo 0x%llx size %ld refcount %d is_expbuf %d",
> + buf->gpu_addr, buf->size,
> + buf->refcount.refs.counter, buf->is_expbuf);
> +
> + if (refcount_dec_and_test(&buf->refcount)) {
> + amdgpu_bo_free_isp_user(bo);
> +
> + // put implicit dmabuf here, detach_dmabuf will not be called
> + if (!buf->is_expbuf)
> + dma_buf_put(buf->dbuf);
> +
> + vfree(buf->vaddr);
> + kfree(buf);
> + buf = NULL;
> + } else {
> + dev_warn(buf->dev, "ignore buffer free, refcount %u > 0",
> + refcount_read(&buf->refcount));
> + }
> +}
> +
> +static void *isp4vid_vb2_alloc(struct vb2_buffer *vb, struct device *dev,
> + unsigned long size)
> +{
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vb->vb2_queue);
> + struct isp4vid_vb2_buf *buf = NULL;
> + struct amdgpu_bo *bo;
> + u64 gpu_addr;
> + u32 ret;
> +
> + buf = kzalloc(sizeof(*buf), GFP_KERNEL | vb->vb2_queue->gfp_flags);
> + if (!buf)
> + return ERR_PTR(-ENOMEM);
> +
> + buf->dev = dev;
> + buf->size = size;
> + buf->vaddr = vmalloc_user(buf->size);
> + if (!buf->vaddr) {
> + kfree(buf);
> + return ERR_PTR(-ENOMEM);
> + }
> +
> + buf->dma_dir = vb->vb2_queue->dma_dir;
> + buf->handler.refcount = &buf->refcount;
> + buf->handler.put = isp4vid_vb2_put;
> + buf->handler.arg = buf;
> +
> + // get implicit dmabuf
> + buf->dbuf = isp4vid_get_dmabuf(vb, buf, 0);
> + if (!buf->dbuf) {
> + dev_err(dev, "fail to get dmabuf\n");
> + return ERR_PTR(-EINVAL);
> + }
> +
> + // create isp user BO and obtain gpu_addr
> + ret = amdgpu_bo_create_isp_user(isp_vdev->amdgpu_dev, buf->dbuf,
> + AMDGPU_GEM_DOMAIN_GTT, &bo, &gpu_addr);
> + if (ret) {
> + dev_err(dev, "fail to create BO\n");
> + return ERR_PTR(-EINVAL);
> + }
> +
> + buf->bo = (void *)bo;
> + buf->gpu_addr = gpu_addr;
> +
> + refcount_set(&buf->refcount, 1);
> +
> + dev_dbg(dev, "allocated isp user bo 0x%llx size %ld refcount %d",
> + buf->gpu_addr, buf->size, buf->refcount.refs.counter);
> +
> + return buf;
> +}
> +
> +const struct vb2_mem_ops isp4vid_vb2_memops = {
> + .alloc = isp4vid_vb2_alloc,
> + .put = isp4vid_vb2_put,
> + .get_userptr = isp4vid_vb2_get_userptr,
> + .put_userptr = isp4vid_vb2_put_userptr,
> +#ifdef CONFIG_HAS_DMA
> + .get_dmabuf = isp4vid_vb2_get_dmabuf,
> +#endif
> + .map_dmabuf = isp4vid_vb2_map_dmabuf,
> + .unmap_dmabuf = isp4vid_vb2_unmap_dmabuf,
> + .attach_dmabuf = isp4vid_vb2_attach_dmabuf,
> + .detach_dmabuf = isp4vid_vb2_detach_dmabuf,
> + .vaddr = isp4vid_vb2_vaddr,
> + .mmap = isp4vid_vb2_mmap,
> + .num_users = isp4vid_vb2_num_users,
> +};
> +
> +static const struct v4l2_pix_format isp4vid_fmt_default = {
> + .width = 1920,
> + .height = 1080,
> + .pixelformat = V4L2_PIX_FMT_NV12,
> + .field = V4L2_FIELD_NONE,
> + .colorspace = V4L2_COLORSPACE_SRGB,
> +};
> +
> +static void isp4vid_capture_return_all_buffers(struct isp4vid_dev *isp_vdev,
> + enum vb2_buffer_state state)
> +{
> + struct isp4vid_capture_buffer *vbuf, *node;
> +
> + mutex_lock(&isp_vdev->buf_list_lock);
> +
> + list_for_each_entry_safe(vbuf, node, &isp_vdev->buf_list, list) {
> + list_del(&vbuf->list);
> + vb2_buffer_done(&vbuf->vb2.vb2_buf, state);
> + }
> + mutex_unlock(&isp_vdev->buf_list_lock);
> +
> + dev_dbg(isp_vdev->dev, "call vb2_buffer_done(%d)\n", state);
> +}
> +
> +static int isp4vid_vdev_link_validate(struct media_link *link)
> +{
> + return 0;
> +}
> +
> +static const struct media_entity_operations isp4vid_vdev_ent_ops = {
> + .link_validate = isp4vid_vdev_link_validate,
> +};
> +
> +static const struct v4l2_file_operations isp4vid_vdev_fops = {
> + .owner = THIS_MODULE,
> + .open = v4l2_fh_open,
> + .release = vb2_fop_release,
> + .read = vb2_fop_read,
> + .poll = vb2_fop_poll,
> + .unlocked_ioctl = video_ioctl2,
> + .mmap = vb2_fop_mmap,
> +};
> +
> +static int isp4vid_ioctl_querycap(struct file *file, void *fh,
> + struct v4l2_capability *cap)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> +
> + strscpy(cap->driver, ISP4VID_ISP_DRV_NAME, sizeof(cap->driver));
> + snprintf(cap->card, sizeof(cap->card), "%s", ISP4VID_ISP_DRV_NAME);
> + snprintf(cap->bus_info, sizeof(cap->bus_info),
> + "platform:%s", ISP4VID_ISP_DRV_NAME);
> +
> + cap->capabilities |= (V4L2_CAP_STREAMING |
> + V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_IO_MC);
> +
> + dev_dbg(isp_vdev->dev, "%s|capabilities=0x%X",
> + isp_vdev->vdev.name, cap->capabilities);
> +
> + return 0;
> +}
> +
> +static int isp4vid_g_fmt_vid_cap(struct file *file, void *priv,
> + struct v4l2_format *f)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> +
> + f->fmt.pix = isp_vdev->format;
> +
> + return 0;
> +}
> +
> +static int isp4vid_try_fmt_vid_cap(struct file *file, void *priv,
> + struct v4l2_format *f)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> + struct v4l2_pix_format *format = &f->fmt.pix;
> + unsigned int i;
> +
> + if (f->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
> + return -EINVAL;
> +
> + /*
> + * Check if the hardware supports the requested format, use the default
> + * format otherwise.
> + */
> + for (i = 0; i < ARRAY_SIZE(isp4vid_formats); i++)
> + if (isp4vid_formats[i] == format->pixelformat)
> + break;
> +
> + if (i == ARRAY_SIZE(isp4vid_formats))
> + format->pixelformat = ISP4VID_DEFAULT_FMT;
> +
> + switch (format->pixelformat) {
> + case V4L2_PIX_FMT_NV12: {
> + const struct v4l2_frmsize_discrete *fsz =
> + v4l2_find_nearest_size(isp4vid_frmsize,
> + ARRAY_SIZE(isp4vid_frmsize),
> + width, height,
> + format->width, format->height);
> +
> + format->width = fsz->width;
> + format->height = fsz->height;
> +
> + format->bytesperline = format->width;
> + format->sizeimage = format->bytesperline *
> + format->height * 3 / 2;
> + }
> + break;
> + case V4L2_PIX_FMT_YUYV: {
> + const struct v4l2_frmsize_discrete *fsz =
> + v4l2_find_nearest_size(isp4vid_frmsize,
> + ARRAY_SIZE(isp4vid_frmsize),
> + width, height,
> + format->width, format->height);
> +
> + format->width = fsz->width;
> + format->height = fsz->height;
> +
> + format->bytesperline = format->width * 2;
> + format->sizeimage = format->bytesperline * format->height;
> + }
> + break;
> + default:
> + dev_err(isp_vdev->dev, "%s|unsupported fmt=%u",
> + isp_vdev->vdev.name, format->pixelformat);
> + return -EINVAL;
> + }
> +
> + if (format->field == V4L2_FIELD_ANY)
> + format->field = isp4vid_fmt_default.field;
> +
> + if (format->colorspace == V4L2_COLORSPACE_DEFAULT)
> + format->colorspace = isp4vid_fmt_default.colorspace;
> +
> + return 0;
> +}
> +
> +static int isp4vid_set_fmt_2_isp(struct v4l2_subdev *sdev,
> + struct v4l2_pix_format *pix_fmt)
> +{
> + struct v4l2_subdev_format fmt = {0};
> +
> + switch (pix_fmt->pixelformat) {
> + case V4L2_PIX_FMT_NV12:
> + fmt.format.code = MEDIA_BUS_FMT_YUYV8_1_5X8;
> + break;
> + case V4L2_PIX_FMT_YUYV:
> + fmt.format.code = MEDIA_BUS_FMT_YUYV8_1X16;
> + break;
> + default:
> + return -EINVAL;
> + };
> + fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE;
> + fmt.pad = ISP4VID_PAD_VIDEO_OUTPUT;
> + fmt.format.width = pix_fmt->width;
> + fmt.format.height = pix_fmt->height;
> + return v4l2_subdev_call(sdev, pad, set_fmt, NULL, &fmt);
> +}
> +
> +static int isp4vid_s_fmt_vid_cap(struct file *file, void *priv,
> + struct v4l2_format *f)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> + u32 pitch;
> + int ret;
> +
> + /* Do not change the format while stream is on */
> + if (vb2_is_busy(&isp_vdev->vbq))
> + return -EBUSY;
> +
> + if (f->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
> + return -EINVAL;
> +
> + ret = isp4vid_try_fmt_vid_cap(file, priv, f);
> + if (ret)
> + return ret;
> +
> + dev_dbg(isp_vdev->dev, "%s|width height:%ux%u->%ux%u",
> + isp_vdev->vdev.name,
> + isp_vdev->format.width, isp_vdev->format.height,
> + f->fmt.pix.width, f->fmt.pix.height);
> + dev_dbg(isp_vdev->dev, "%s|pixelformat:0x%x-0x%x",
> + isp_vdev->vdev.name, isp_vdev->format.pixelformat,
> + f->fmt.pix.pixelformat);
> + dev_dbg(isp_vdev->dev, "%s|bytesperline:%u->%u",
> + isp_vdev->vdev.name, isp_vdev->format.bytesperline,
> + f->fmt.pix.bytesperline);
> + dev_dbg(isp_vdev->dev, "%s|sizeimage:%u->%u",
> + isp_vdev->vdev.name, isp_vdev->format.sizeimage,
> + f->fmt.pix.sizeimage);
> +
> + isp_vdev->format = f->fmt.pix;
> +
> + switch (isp_vdev->format.pixelformat) {
> + case V4L2_PIX_FMT_NV12:
> + pitch = isp_vdev->format.width;
> + break;
> + case V4L2_PIX_FMT_YUYV:
> + pitch = isp_vdev->format.width * 2;
> + break;
> + default:
> + dev_err(isp_vdev->dev, "%s|unsupported fmt=0x%x\n",
> + isp_vdev->vdev.name, isp_vdev->format.pixelformat);
> + return -EINVAL;
> + }
> +
> + ret = isp4vid_set_fmt_2_isp(isp_vdev->isp_sdev, &isp_vdev->format);
> +
> + return ret;
> +}
> +
> +static int isp4vid_enum_fmt_vid_cap(struct file *file, void *priv,
> + struct v4l2_fmtdesc *f)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> +
> + switch (f->index) {
> + case 0:
> + f->pixelformat = V4L2_PIX_FMT_NV12;
> + break;
> + case 1:
> + f->pixelformat = V4L2_PIX_FMT_YUYV;
> + break;
> + default:
> + return -EINVAL;
> + }
> +
> + dev_dbg(isp_vdev->dev, "%s|index=%d, pixelformat=0x%X",
> + isp_vdev->vdev.name, f->index, f->pixelformat);
> +
> + return 0;
> +}
> +
> +static int isp4vid_enum_framesizes(struct file *file, void *fh,
> + struct v4l2_frmsizeenum *fsize)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> + unsigned int i;
> +
> + for (i = 0; i < ARRAY_SIZE(isp4vid_formats); i++) {
> + if (isp4vid_formats[i] == fsize->pixel_format)
> + break;
> + }
> + if (i == ARRAY_SIZE(isp4vid_formats))
> + return -EINVAL;
> +
> + if (fsize->index < ARRAY_SIZE(isp4vid_frmsize)) {
> + fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE;
> + fsize->discrete = isp4vid_frmsize[fsize->index];
> + dev_dbg(isp_vdev->dev, "%s|size[%d]=%dx%d",
> + isp_vdev->vdev.name, fsize->index,
> + fsize->discrete.width, fsize->discrete.height);
> + } else {
> + return -EINVAL;
> + }
> +
> + return 0;
> +}
> +
> +static int isp4vid_ioctl_enum_frameintervals(struct file *file, void *priv,
> + struct v4l2_frmivalenum *fival)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> + int i;
> +
> + if (fival->index >= ARRAY_SIZE(isp4vid_tpfs))
> + return -EINVAL;
> +
> + for (i = 0; i < ARRAY_SIZE(isp4vid_formats); i++)
> + if (isp4vid_formats[i] == fival->pixel_format)
> + break;
> + if (i == ARRAY_SIZE(isp4vid_formats))
> + return -EINVAL;
> +
> + for (i = 0; i < ARRAY_SIZE(isp4vid_frmsize); i++)
> + if (isp4vid_frmsize[i].width == fival->width &&
> + isp4vid_frmsize[i].height == fival->height)
> + break;
> + if (i == ARRAY_SIZE(isp4vid_frmsize))
> + return -EINVAL;
> +
> + fival->type = V4L2_FRMIVAL_TYPE_DISCRETE;
> + fival->discrete = isp4vid_tpfs[fival->index];
> + v4l2_simplify_fraction(&fival->discrete.numerator,
> + &fival->discrete.denominator, 8, 333);
> +
> + dev_dbg(isp_vdev->dev, "%s|interval[%d]=%d/%d",
> + isp_vdev->vdev.name, fival->index,
> + fival->discrete.numerator,
> + fival->discrete.denominator);
> +
> + return 0;
> +}
> +
> +static int isp4vid_ioctl_g_param(struct file *file, void *priv,
> + struct v4l2_streamparm *param)
> +{
> + struct isp4vid_dev *isp_vdev = video_drvdata(file);
> +
> + if (param->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
> + return -EINVAL;
> +
> + param->parm.capture.capability = V4L2_CAP_TIMEPERFRAME;
> + param->parm.capture.timeperframe = isp_vdev->timeperframe;
> + param->parm.capture.readbuffers = 0;
> +
> + dev_dbg(isp_vdev->dev, "%s|timeperframe=%d/%d", isp_vdev->vdev.name,
> + param->parm.capture.timeperframe.numerator,
> + param->parm.capture.timeperframe.denominator);
> + return 0;
> +}
> +
> +static const struct v4l2_ioctl_ops isp4vid_vdev_ioctl_ops = {
> + /* VIDIOC_QUERYCAP handler */
> + .vidioc_querycap = isp4vid_ioctl_querycap,
> +
> + /* VIDIOC_ENUM_FMT handlers */
> + .vidioc_enum_fmt_vid_cap = isp4vid_enum_fmt_vid_cap,
> +
> + /* VIDIOC_G_FMT handlers */
> + .vidioc_g_fmt_vid_cap = isp4vid_g_fmt_vid_cap,
> +
> + /* VIDIOC_S_FMT handlers */
> + .vidioc_s_fmt_vid_cap = isp4vid_s_fmt_vid_cap,
> +
> + /* VIDIOC_TRY_FMT handlers */
> + .vidioc_try_fmt_vid_cap = isp4vid_try_fmt_vid_cap,
> +
> + /* Buffer handlers */
> + .vidioc_reqbufs = vb2_ioctl_reqbufs,
> + .vidioc_querybuf = vb2_ioctl_querybuf,
> + .vidioc_qbuf = vb2_ioctl_qbuf,
> + .vidioc_expbuf = vb2_ioctl_expbuf,
> + .vidioc_dqbuf = vb2_ioctl_dqbuf,
> + .vidioc_create_bufs = vb2_ioctl_create_bufs,
> + .vidioc_prepare_buf = vb2_ioctl_prepare_buf,
> +
> + /* Stream on/off */
> + .vidioc_streamon = vb2_ioctl_streamon,
> + .vidioc_streamoff = vb2_ioctl_streamoff,
> +
> + /* Stream type-dependent parameter ioctls */
> + .vidioc_g_parm = isp4vid_ioctl_g_param,
> +
> + /* Debugging ioctls */
> + .vidioc_enum_framesizes = isp4vid_enum_framesizes,
> +
> + /* VIDIOC_ENUM_FRAMEINTERVALS */
> + .vidioc_enum_frameintervals = isp4vid_ioctl_enum_frameintervals,
> +
> +};
> +
> +static unsigned int isp4vid_get_image_size(struct v4l2_pix_format *fmt)
> +{
> + switch (fmt->pixelformat) {
> + case V4L2_PIX_FMT_NV12:
> + return fmt->width * fmt->height * 3 / 2;
> + case V4L2_PIX_FMT_YUYV:
> + return fmt->width * fmt->height * 2;
> + default:
> + return 0;
> + }
> +};
> +
> +static int isp4vid_qops_queue_setup(struct vb2_queue *vq,
> + unsigned int *nbuffers,
> + unsigned int *nplanes, unsigned int sizes[],
> + struct device *alloc_devs[])
> +{
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vq);
> + unsigned int q_num_bufs = vb2_get_num_buffers(vq);
> +
> + if (*nplanes > 1) {
> + dev_err(isp_vdev->dev,
> + "fail to setup queue, no mplane supported %u\n",
> + *nplanes);
> + return -EINVAL;
> + };
> +
> + if (*nplanes == 1) {
> + unsigned int size;
> +
> + size = isp4vid_get_image_size(&isp_vdev->format);
> + if (sizes[0] < size) {
> + dev_err(isp_vdev->dev,
> + "fail for small plane size %u, %u expected\n",
> + sizes[0], size);
> + return -EINVAL;
> + }
> + }
> +
> + if (q_num_bufs + *nbuffers < ISP4VID_VIDEO_BUF_NUM)
> + *nbuffers = ISP4VID_VIDEO_BUF_NUM - q_num_bufs;
> +
> + switch (isp_vdev->format.pixelformat) {
> + case V4L2_PIX_FMT_NV12:
> + case V4L2_PIX_FMT_YUYV: {
> + *nplanes = 1;
> + sizes[0] = max(sizes[0], isp_vdev->format.sizeimage);
> + isp_vdev->format.sizeimage = sizes[0];
> + }
> + break;
> + default:
> + dev_err(isp_vdev->dev, "%s|unsupported fmt=%u\n",
> + isp_vdev->vdev.name, isp_vdev->format.pixelformat);
> + return -EINVAL;
> + }
> +
> + dev_dbg(isp_vdev->dev, "%s|*nbuffers=%u *nplanes=%u sizes[0]=%u\n",
> + isp_vdev->vdev.name,
> + *nbuffers, *nplanes, sizes[0]);
> +
> + return 0;
> +}
> +
> +static void isp4vid_qops_buffer_queue(struct vb2_buffer *vb)
> +{
> + struct isp4vid_capture_buffer *buf =
> + container_of(vb, struct isp4vid_capture_buffer, vb2.vb2_buf);
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vb->vb2_queue);
> +
> + struct isp4vid_vb2_buf *priv_buf = vb->planes[0].mem_priv;
> + struct isp4if_img_buf_info *img_buf = &buf->img_buf;
> +
> + dev_dbg(isp_vdev->dev, "%s|index=%u", isp_vdev->vdev.name, vb->index);
> +
> + dev_dbg(isp_vdev->dev, "queue isp user bo 0x%llx size=%lu",
> + priv_buf->gpu_addr,
> + priv_buf->size);
> +
> + switch (isp_vdev->format.pixelformat) {
> + case V4L2_PIX_FMT_NV12: {
> + u32 y_size = isp_vdev->format.sizeimage / 3 * 2;
> + u32 uv_size = isp_vdev->format.sizeimage / 3;
> +
> + img_buf->planes[0].len = y_size;
> + img_buf->planes[0].sys_addr = priv_buf->vaddr;
> + img_buf->planes[0].mc_addr = priv_buf->gpu_addr;
> +
> + dev_dbg(isp_vdev->dev, "img_buf[0]: mc=0x%llx size=%u",
> + img_buf->planes[0].mc_addr,
> + img_buf->planes[0].len);
> +
> + img_buf->planes[1].len = uv_size;
> + img_buf->planes[1].sys_addr =
> + (void *)((u64)priv_buf->vaddr + y_size);
> + img_buf->planes[1].mc_addr = priv_buf->gpu_addr + y_size;
> +
> + dev_dbg(isp_vdev->dev, "img_buf[1]: mc=0x%llx size=%u",
> + img_buf->planes[1].mc_addr,
> + img_buf->planes[1].len);
> +
> + img_buf->planes[2].len = 0;
> + }
> + break;
> + case V4L2_PIX_FMT_YUYV: {
> + img_buf->planes[0].len = isp_vdev->format.sizeimage;
> + img_buf->planes[0].sys_addr = priv_buf->vaddr;
> + img_buf->planes[0].mc_addr = priv_buf->gpu_addr;
> +
> + dev_dbg(isp_vdev->dev, "img_buf[0]: mc=0x%llx size=%u",
> + img_buf->planes[0].mc_addr,
> + img_buf->planes[0].len);
> +
> + img_buf->planes[1].len = 0;
> + img_buf->planes[2].len = 0;
> + }
> + break;
> + default:
> + dev_err(isp_vdev->dev, "%s|unsupported fmt=%u",
> + isp_vdev->vdev.name, isp_vdev->format.pixelformat);
> + return;
> + }
> +
> + if (isp_vdev->stream_started)
> + isp_vdev->ops->send_buffer(isp_vdev, img_buf);
> +
> + mutex_lock(&isp_vdev->buf_list_lock);
> + list_add_tail(&buf->list, &isp_vdev->buf_list);
> + mutex_unlock(&isp_vdev->buf_list_lock);
> +}
> +
> +static void isp4vid_qops_buffer_cleanup(struct vb2_buffer *vb)
> +{
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vb->vb2_queue);
> + struct isp4vid_vb2_buf *buf = vb->planes[0].mem_priv;
> +
> + dev_dbg(isp_vdev->dev, "%s|index=%u vb->memory %u",
> + isp_vdev->vdev.name, vb->index, vb->memory);
> +
> + if (!buf) {
> + dev_err(isp_vdev->dev, "Invalid buf handle");
> + return;
> + }
> +
> + // release implicit dmabuf reference here for vb2 buffer
> + // of type MMAP and is exported
> + if (vb->memory == VB2_MEMORY_MMAP && buf->is_expbuf) {
> + dma_buf_put(buf->dbuf);
> + dev_dbg(isp_vdev->dev,
> + "put dmabuf for vb->memory %d expbuf %d",
> + vb->memory,
> + buf->is_expbuf);
> + }
> +}
> +
> +static int isp4vid_qops_start_streaming(struct vb2_queue *vq,
> + unsigned int count)
> +{
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vq);
> + struct isp4vid_capture_buffer *isp_buf;
> + struct v4l2_subdev *isp_subdev = NULL;
> + struct media_entity *entity;
> + struct v4l2_subdev *subdev;
> + struct media_pad *pad;
> + int ret = 0;
> +
> + if (isp_vdev->stream_started) {
> + dev_dbg(isp_vdev->dev,
> + "%s do nothing for already streaming\n",
> + isp_vdev->vdev.name);
> + return 0;
> + }
> + isp_vdev->sequence = 0;
> + ret = v4l2_pipeline_pm_get(&isp_vdev->vdev.entity);
> + if (ret) {
> + dev_err(isp_vdev->dev, "power up isp fail %d\n", ret);
> + return ret;
> + }
> +
> + entity = &isp_vdev->vdev.entity;
> + while (1) {
> + pad = &entity->pads[0];
> + if (!(pad->flags & MEDIA_PAD_FL_SINK))
> + break;
> +
> + pad = media_pad_remote_pad_first(pad);
> + if (!pad || !is_media_entity_v4l2_subdev(pad->entity))
> + break;
> +
> + entity = pad->entity;
> + subdev = media_entity_to_v4l2_subdev(entity);
> +
> + if (entity->function == MEDIA_ENT_F_PROC_VIDEO_ISP) {
> + ret = v4l2_subdev_call(subdev, video, pre_streamon, 0);
> + /* The isp s_stream should be called last! */
> + isp_subdev = subdev;
> + } else {
> + ret = v4l2_subdev_call(subdev, video, s_stream, 1);
> + }
> +
> + if (ret < 0 && ret != -ENOIOCTLCMD) {
> + dev_dbg(isp_vdev->dev, "fail start streaming: %s %d\n",
> + subdev->name, ret);
> + return ret;
> + }
> + }
> +
> + if (isp_subdev) {
> + ret = v4l2_subdev_call(isp_subdev, video, s_stream, 1);
> + if (ret < 0 && ret != -ENOIOCTLCMD) {
> + dev_dbg(isp_vdev->dev, "fail start stream: %s %d\n",
> + isp_subdev->name, ret);
> + return ret;
> + }
> + }
> +
> + list_for_each_entry(isp_buf, &isp_vdev->buf_list, list) {
> + isp_vdev->ops->send_buffer(isp_vdev, &isp_buf->img_buf);
> + }
> +
> + /* Start the media pipeline */
> + ret = video_device_pipeline_start(&isp_vdev->vdev, &isp_vdev->pipe);
> + if (ret) {
> + dev_err(isp_vdev->dev, "video_device_pipeline_start fail:%d",
> + ret);
> + isp4vid_capture_return_all_buffers(isp_vdev,
> + VB2_BUF_STATE_QUEUED);
> + return ret;
> + }
> + isp_vdev->stream_started = true;
> +
> + return 0;
> +}
> +
> +static void isp4vid_qops_stop_streaming(struct vb2_queue *vq)
> +{
> + struct isp4vid_dev *isp_vdev = vb2_get_drv_priv(vq);
> + struct v4l2_subdev *subdev, *isp_subdev = NULL;
> + struct media_entity *entity;
> + struct media_pad *pad;
> + int ret;
> +
> + if (!isp_vdev->stream_started) {
> + dev_dbg(isp_vdev->dev,
> + "%s stop_streaming, do none for not started\n",
> + isp_vdev->vdev.name);
> + return;
> + }
> + dev_dbg(isp_vdev->dev, "%s stop_streaming\n",
> + isp_vdev->vdev.name);
> +
> + entity = &isp_vdev->vdev.entity;
> + while (1) {
> + pad = &entity->pads[0];
> + if (!(pad->flags & MEDIA_PAD_FL_SINK))
> + break;
> +
> + pad = media_pad_remote_pad_first(pad);
> + if (!pad || !is_media_entity_v4l2_subdev(pad->entity))
> + break;
> +
> + entity = pad->entity;
> + subdev = media_entity_to_v4l2_subdev(entity);
> +
> + if (entity->function == MEDIA_ENT_F_PROC_VIDEO_ISP) {
> + /*
> + * isp subdev to call isp post_streamoff
> + * after s_stream sequence
> + */
> + isp_subdev = subdev;
> + }
> + ret = v4l2_subdev_call(subdev, video, s_stream, 0);
> +
> + if (ret < 0 && ret != -ENOIOCTLCMD)
> + dev_dbg(isp_vdev->dev, "fail start streaming: %s %d\n",
> + subdev->name, ret);
> + }
> +
> + if (isp_subdev) {
> + ret = v4l2_subdev_call(isp_subdev, video, post_streamoff);
> + if (ret < 0 && ret != -ENOIOCTLCMD)
> + dev_dbg(isp_vdev->dev, "fail start stream: %s %d\n",
> + isp_subdev->name, ret);
> + }
> +
> + isp_vdev->stream_started = false;
> + v4l2_pipeline_pm_put(&isp_vdev->vdev.entity);
> +
> + /* Stop the media pipeline */
> + video_device_pipeline_stop(&isp_vdev->vdev);
> +
> + /* Release all active buffers */
> + isp4vid_capture_return_all_buffers(isp_vdev, VB2_BUF_STATE_ERROR);
> +}
> +
> +static int isp4vid_fill_buffer_size(struct isp4vid_dev *isp_vdev)
> +{
> + int ret = 0;
> +
> + switch (isp_vdev->format.pixelformat) {
> + case V4L2_PIX_FMT_NV12:
> + isp_vdev->format.bytesperline = isp_vdev->format.width;
> + isp_vdev->format.sizeimage = isp_vdev->format.bytesperline *
> + isp_vdev->format.height * 3 / 2;
> + break;
> + case V4L2_PIX_FMT_YUYV:
> + isp_vdev->format.bytesperline = isp_vdev->format.width;
> + isp_vdev->format.sizeimage = isp_vdev->format.bytesperline *
> + isp_vdev->format.height * 2;
> + break;
> + default:
> + dev_err(isp_vdev->dev, "fail for invalid default format 0x%x",
> + isp_vdev->format.pixelformat);
> + ret = -EINVAL;
> + break;
> + }
> +
> + return ret;
> +}
> +
> +static const struct vb2_ops isp4vid_qops = {
> + .queue_setup = isp4vid_qops_queue_setup,
> + .buf_cleanup = isp4vid_qops_buffer_cleanup,
> + .buf_queue = isp4vid_qops_buffer_queue,
> + .start_streaming = isp4vid_qops_start_streaming,
> + .stop_streaming = isp4vid_qops_stop_streaming,
> + .wait_prepare = vb2_ops_wait_prepare,
> + .wait_finish = vb2_ops_wait_finish,
> +};
> +
> +int isp4vid_dev_init(struct isp4vid_dev *isp_vdev,
> + struct v4l2_subdev *isp_sdev,
> + const struct isp4vid_ops *ops,
> + void *amdgpu_dev)
> +{
> + const char *vdev_name = isp4vid_video_dev_name;
> + struct v4l2_device *v4l2_dev;
> + struct video_device *vdev;
> + struct vb2_queue *q;
> + int ret;
> +
> + if (!isp_vdev || !isp_sdev || !isp_sdev->v4l2_dev || !amdgpu_dev)
> + return -EINVAL;
> +
> + v4l2_dev = isp_sdev->v4l2_dev;
> + vdev = &isp_vdev->vdev;
> +
> + isp_vdev->isp_sdev = isp_sdev;
> + isp_vdev->amdgpu_dev = amdgpu_dev;
> + isp_vdev->dev = v4l2_dev->dev;
> + isp_vdev->ops = ops;
> +
> + /* Initialize the vb2_queue struct */
> + mutex_init(&isp_vdev->vbq_lock);
> + q = &isp_vdev->vbq;
> + q->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> + q->io_modes = VB2_MMAP | VB2_DMABUF;
> + q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
> + q->buf_struct_size = sizeof(struct isp4vid_capture_buffer);
> + q->min_queued_buffers = 2;
> + q->ops = &isp4vid_qops;
> + q->drv_priv = isp_vdev;
> + q->mem_ops = &isp4vid_vb2_memops;
> + q->lock = &isp_vdev->vbq_lock;
> + q->dev = v4l2_dev->dev;
> + ret = vb2_queue_init(q);
> + if (ret) {
> + dev_err(v4l2_dev->dev, "vb2_queue_init error:%d", ret);
> + return ret;
> + }
> + /* Initialize buffer list and its lock */
> + mutex_init(&isp_vdev->buf_list_lock);
> + INIT_LIST_HEAD(&isp_vdev->buf_list);
> +
> + /* Set default frame format */
> + isp_vdev->format = isp4vid_fmt_default;
> + isp_vdev->timeperframe = ISP4VID_ISP_TPF_DEFAULT;
> + v4l2_simplify_fraction(&isp_vdev->timeperframe.numerator,
> + &isp_vdev->timeperframe.denominator, 8, 333);
> +
> + ret = isp4vid_fill_buffer_size(isp_vdev);
> + if (ret) {
> + dev_err(v4l2_dev->dev, "fail to fill buffer size: %d\n", ret);
> + return ret;
> + }
> +
> + ret = isp4vid_set_fmt_2_isp(isp_sdev, &isp_vdev->format);
> + if (ret) {
> + dev_err(v4l2_dev->dev, "fail init format :%d\n", ret);
> + return ret;
> + }
> +
> + /* Initialize the video_device struct */
> + isp_vdev->vdev.entity.name = vdev_name;
> + isp_vdev->vdev.entity.function = MEDIA_ENT_F_IO_V4L;
> + isp_vdev->vdev_pad.flags = MEDIA_PAD_FL_SINK;
> + ret = media_entity_pads_init(&isp_vdev->vdev.entity, 1,
> + &isp_vdev->vdev_pad);
> +
> + if (ret) {
> + dev_err(v4l2_dev->dev, "init media entity pad fail:%d\n", ret);
> + return ret;
> + }
> +
> + vdev->device_caps = V4L2_CAP_VIDEO_CAPTURE |
> + V4L2_CAP_STREAMING | V4L2_CAP_IO_MC;
> + vdev->entity.ops = &isp4vid_vdev_ent_ops;
> + vdev->release = video_device_release_empty;
> + vdev->fops = &isp4vid_vdev_fops;
> + vdev->ioctl_ops = &isp4vid_vdev_ioctl_ops;
> + vdev->lock = NULL;
> + vdev->queue = q;
> + vdev->v4l2_dev = v4l2_dev;
> + vdev->vfl_dir = VFL_DIR_RX;
> + strscpy(vdev->name, vdev_name, sizeof(vdev->name));
> + video_set_drvdata(vdev, isp_vdev);
> +
> + ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1);
> + if (ret)
> + dev_err(v4l2_dev->dev, "register video device fail:%d\n", ret);
> +
> + return ret;
> +}
> +
> +void isp4vid_dev_deinit(struct isp4vid_dev *isp_vdev)
> +{
> + video_unregister_device(&isp_vdev->vdev);
> +}
> diff --git a/drivers/media/platform/amd/isp4/isp4_video.h b/drivers/media/platform/amd/isp4/isp4_video.h
> new file mode 100644
> index 000000000000..4d6705174d34
> --- /dev/null
> +++ b/drivers/media/platform/amd/isp4/isp4_video.h
> @@ -0,0 +1,93 @@
> +/* SPDX-License-Identifier: GPL-2.0+ */
> +/*
> + * Copyright (C) 2025 Advanced Micro Devices, Inc.
> + */
> +
> +#ifndef _ISP4_VIDEO_H_
> +#define _ISP4_VIDEO_H_
> +
> +#include <linux/mutex.h>
> +#include <media/videobuf2-memops.h>
> +#include <media/videobuf2-vmalloc.h>
> +#include <media/v4l2-dev.h>
> +#include "isp4_interface.h"
> +
> +enum isp4vid_buf_done_status {
> + /* It means no corresponding image buf in fw response */
> + ISP4VID_BUF_DONE_STATUS_ABSENT,
> + ISP4VID_BUF_DONE_STATUS_SUCCESS,
> + ISP4VID_BUF_DONE_STATUS_FAILED
> +};
> +
> +struct isp4vid_buf_done_info {
> + enum isp4vid_buf_done_status status;
> + struct isp4if_img_buf_info buf;
> +};
> +
> +/* call back parameter for CB_EVT_ID_FRAME_DONE */
> +struct isp4vid_framedone_param {
> + s32 poc;
> + s32 cam_id;
> + s64 time_stamp;
> + struct isp4vid_buf_done_info preview;
> +};
> +
> +struct isp4vid_capture_buffer {
> + /*
> + * struct vb2_v4l2_buffer must be the first element
> + * the videobuf2 framework will allocate this struct based on
> + * buf_struct_size and use the first sizeof(struct vb2_buffer) bytes of
> + * memory as a vb2_buffer
> + */
> + struct vb2_v4l2_buffer vb2;
> + struct isp4if_img_buf_info img_buf;
> + struct list_head list;
> +};
> +
> +struct isp4vid_dev;
> +
> +struct isp4vid_ops {
> + int (*send_buffer)(struct isp4vid_dev *vid,
> + struct isp4if_img_buf_info *img_buf);
> +};
> +
> +struct isp4vid_dev {
> + struct video_device vdev;
> + struct media_pad vdev_pad;
> + struct v4l2_pix_format format;
> +
> + /* mutex that protects vbq */
> + struct mutex vbq_lock;
> + struct vb2_queue vbq;
> +
> + /* mutex that protects buf_list */
> + struct mutex buf_list_lock;
> + struct list_head buf_list;
> +
> + u32 sequence;
> + bool stream_started;
> + struct task_struct *kthread;
> +
> + struct media_pipeline pipe;
> + struct device *dev;
> + void *amdgpu_dev;
> + struct v4l2_subdev *isp_sdev;
> + struct v4l2_fract timeperframe;
> +
> + /* Callback operations */
> + const struct isp4vid_ops *ops;
> +};
> +
> +int isp4vid_dev_init(struct isp4vid_dev *isp_vdev,
> + struct v4l2_subdev *isp_sdev,
> + const struct isp4vid_ops *ops,
> + void *amdgpu_dev);
> +
> +void isp4vid_dev_deinit(struct isp4vid_dev *isp_vdev);
> +
> +s32 isp4vid_notify(void *cb_ctx, struct isp4vid_framedone_param *evt_param);
> +
> +void isp4vid_frmsize_range(struct v4l2_frmsize_discrete *min,
> + struct v4l2_frmsize_discrete *max);
> +
> +#endif
Powered by blists - more mailing lists