[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200331115221.GA4767@pendragon.ideasonboard.com>
Date: Tue, 31 Mar 2020 14:52:21 +0300
From: Laurent Pinchart <laurent.pinchart@...asonboard.com>
To: Hans Verkuil <hverkuil@...all.nl>
Cc: Sakari Ailus <sakari.ailus@....fi>,
Sowjanya Komatineni <skomatineni@...dia.com>,
thierry.reding@...il.com, jonathanh@...dia.com, frankc@...dia.com,
helen.koike@...labora.com, digetx@...il.com, sboyd@...nel.org,
linux-media@...r.kernel.org, devicetree@...r.kernel.org,
linux-clk@...r.kernel.org, linux-tegra@...r.kernel.org,
linux-kernel@...r.kernel.org
Subject: Re: [RFC PATCH v5 6/9] media: tegra: Add Tegra210 Video input driver
Hello,
On Tue, Mar 31, 2020 at 01:27:19PM +0200, Hans Verkuil wrote:
> On 3/31/20 1:10 PM, Sakari Ailus wrote:
> > On Tue, Mar 31, 2020 at 12:56:57PM +0200, Hans Verkuil wrote:
> >> On 3/31/20 12:32 PM, Sakari Ailus wrote:
> >>> On Mon, Mar 30, 2020 at 12:59:15PM +0200, Hans Verkuil wrote:
> >>>> On 3/25/20 12:03 PM, Sakari Ailus wrote:
> >>>>> On Mon, Mar 23, 2020 at 10:52:32AM -0700, Sowjanya Komatineni wrote:
> >>>>>> Tegra210 contains a powerful Video Input (VI) hardware controller
> >>>>>> which can support up to 6 MIPI CSI camera sensors.
> >>>>>>
> >>>>>> Each Tegra CSI port can be one-to-one mapped to VI channel and can
> >>>>>> capture from an external camera sensor connected to CSI or from
> >>>>>> built-in test pattern generator.
> >>>>>>
> >>>>>> Tegra210 supports built-in test pattern generator from CSI to VI.
> >>>>>>
> >>>>>> This patch adds a V4L2 media controller and capture driver support
> >>>>>> for Tegra210 built-in CSI to VI test pattern generator.
> >>>>>>
> >>>>>> Signed-off-by: Sowjanya Komatineni <skomatineni@...dia.com>
> >>>>>> ---
> >>>>>> drivers/staging/media/Kconfig | 2 +
> >>>>>> drivers/staging/media/Makefile | 1 +
> >>>>>> drivers/staging/media/tegra/Kconfig | 10 +
> >>>>>> drivers/staging/media/tegra/Makefile | 8 +
> >>>>>> drivers/staging/media/tegra/TODO | 10 +
> >>>>>> drivers/staging/media/tegra/tegra-common.h | 263 +++++++
> >>>>>> drivers/staging/media/tegra/tegra-csi.c | 522 ++++++++++++++
> >>>>>> drivers/staging/media/tegra/tegra-csi.h | 118 ++++
> >>>>>> drivers/staging/media/tegra/tegra-vi.c | 1058 ++++++++++++++++++++++++++++
> >>>>>> drivers/staging/media/tegra/tegra-vi.h | 83 +++
> >>>>>> drivers/staging/media/tegra/tegra-video.c | 129 ++++
> >>>>>> drivers/staging/media/tegra/tegra-video.h | 32 +
> >>>>>> drivers/staging/media/tegra/tegra210.c | 754 ++++++++++++++++++++
> >>>>>> drivers/staging/media/tegra/tegra210.h | 192 +++++
> >>>>>
> >>>>> Why staging? Are there reasons not to aim this to the kernel proper right
> >>>>> away? If you only support TPG, the driver may not have too many (if any)
> >>>>> real users anyway.
> >>>>>
> >>>>>> 14 files changed, 3182 insertions(+)
> >>>>>> create mode 100644 drivers/staging/media/tegra/Kconfig
> >>>>>> create mode 100644 drivers/staging/media/tegra/Makefile
> >>>>>> create mode 100644 drivers/staging/media/tegra/TODO
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-common.h
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-csi.c
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-csi.h
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-vi.c
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-vi.h
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-video.c
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra-video.h
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra210.c
> >>>>>> create mode 100644 drivers/staging/media/tegra/tegra210.h
> >>>>>>
> >>>>
> >>>> <snip>
> >>>>
> >>>>>> +static int tegra_channel_g_input(struct file *file, void *priv,
> >>>>>> + unsigned int *i)
> >>>>>> +{
> >>>>>> + *i = 0;
> >>>>>> + return 0;
> >>>>>> +}
> >>>>>> +
> >>>>>> +static int tegra_channel_s_input(struct file *file, void *priv,
> >>>>>> + unsigned int input)
> >>>>>> +{
> >>>>>> + if (input > 0)
> >>>>>> + return -EINVAL;
> >>>>>> +
> >>>>>> + return 0;
> >>>>>> +}
> >>>>>
> >>>>> Please see patchset on topic "v4l2-dev/ioctl: Add V4L2_CAP_IO_MC" on
> >>>>> linux-media; it's relevant here, too.
> >>>>
> >>>> No, it isn't. The pipeline is controlled by the driver, not by userspace.
> >>>> This is a regular video capture driver, not an ISP driver.
> >>>
> >>> I don't think that really makes a difference, whether a device is an ISP or
> >>> not, but instead what does is whether there is something to control in its
> >>> pipeline that cannot be generally done through the regular V4L2 interface.
> >>> Even plain CSI-2 receiver drivers should be media device centric these days
> >>> as doing otherwise excludes using a range of sensor drivers with them,
> >>> including any possible future support for e.g. sensor embedded data.
> >>>
> >>
> >> We've been back and forth on this before for this driver. I see no reason to make things
> >> complicated, these are simple video pipelines for video capture. Making this media
> >> device centric means that existing software using the BSP version of this driver require
> >> a full rewrite, which is not desirable.
> >>
> >> If we are going to require CSI receiver drivers to be media centric, then that's a
> >> major departure of existing practice. And something that needs to be discussed first,
> >
> > I'd be happy to discuss that.
> >
> > Either way, the current design is problematic as it excludes a range of
> > camera sensors being used with the driver --- addressing of which requires
> > converting the driver MC centric. If the driver is merged to mainline, then
> > the user might face a Kconfig option or a module parameter to choose
> > between the two --- this defines uAPI behaviour after all.
> >
> > The only way to avoid that in the future is to make it MC-centric right
> > away.
> >
> >> since that will require that support for each csi receiver driver is added to libcamera.
> >> Is libcamera ready for that? Are common applications using libcamera yet?
> >>
> >> Obviously, if NVIDIA decides that this is worth the effort, then I have no objection.
> >> But I don't think it is something we should require at this stage.
> >
> > Works for me. But in that case NVIDIA should also be aware that doing so
> > has consequences.
> >
> > We also haven't discussed what to do with old V4L2-centric drivers which
> > you'd use with sensors that expose their own subdevs. The proportion of all
> > sensors might not be large currently but it is almost certainly bound to
> > grow in the future.
> >
> > FWIW, Intel ipu3-cio2 CSI-2 receiver driver is MC-centric e.g. for the
> > above reasons. Libcamera supports it currently. I'll let Laurent (cc'd)
> > comment on the details.
>
> I think it would be good to at least describe in some detail what you gain
> by taking the media centric route, and what the obstacles are (loss of compatibility
> with existing applications, requiring libcamera support).
In this case the main gain is control of the camera sensor. Sensors can
appear as simple when you don't look too closely at them, but many
sensors (especially the ones modelled after SMIA++ and the now standard
- and open! - MIPI CCS specification) have 3 locations to perform
cropping (analog, digital and output), and 3 locations to perform
scaling (binning, skipping, and full-featured scaler). All of these need
to be controlled by userspace one way or another if you want to
implement proper camera algorithms, which those platforms target.
> My personal feeling has always been that for ISP drivers the pros of making
> a media-centric driver outweigh the cons, but that for a standard video capture
> pipeline without complex processing blocks the cons outweigh the pros.
>
> This might change if libcamera becomes widely used, but we're not there yet.
>
> To be honest, I am not opposed to having a kernel config option for drivers
> like this that select the media-centric API vs a regular API, if that can be
> done without too much work. If you need full control for your embedded system,
> then you enable the option. If you want full compatibility with existing
> applications, then disable it.
How would distributions be supposed to handle those ? That could in the
end need to be a per-driver option, and it would be very messy. Maybe
it's unavoidable, I'm trying to figure out a way to avoid such an option
for sensor drivers, to decide to expose them as a single subdev or
multiple subdevs in order to support multiple streams CSI-2 streams, and
I'm not sure I'll succeed.
--
Regards,
Laurent Pinchart
Powered by blists - more mailing lists