blob: dc224aec7b81deceb92cf4c016f7dd7043826e3a [file] [log] [blame]
/*
* Analog Devices video capture driver
*
* Copyright (c) 2011 Analog Devices Inc.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 as
* published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
*/
#include <linux/completion.h>
#include <linux/delay.h>
#include <linux/errno.h>
#include <linux/fs.h>
#include <linux/i2c.h>
#include <linux/init.h>
#include <linux/interrupt.h>
#include <linux/io.h>
#include <linux/mm.h>
#include <linux/module.h>
#include <linux/platform_device.h>
#include <linux/slab.h>
#include <linux/time.h>
#include <linux/types.h>
#include <media/v4l2-common.h>
#include <media/v4l2-ctrls.h>
#include <media/v4l2-device.h>
#include <media/v4l2-ioctl.h>
#include <media/videobuf2-dma-contig.h>
#include <asm/dma.h>
#include <media/blackfin/bfin_capture.h>
#include <media/blackfin/ppi.h>
#define CAPTURE_DRV_NAME "bfin_capture"
struct bcap_format {
char *desc;
u32 pixelformat;
u32 mbus_code;
int bpp; /* bits per pixel */
int dlen; /* data length for ppi in bits */
};
struct bcap_buffer {
struct vb2_v4l2_buffer vb;
struct list_head list;
};
struct bcap_device {
/* capture device instance */
struct v4l2_device v4l2_dev;
/* v4l2 control handler */
struct v4l2_ctrl_handler ctrl_handler;
/* device node data */
struct video_device video_dev;
/* sub device instance */
struct v4l2_subdev *sd;
/* capture config */
struct bfin_capture_config *cfg;
/* ppi interface */
struct ppi_if *ppi;
/* current input */
unsigned int cur_input;
/* current selected standard */
v4l2_std_id std;
/* current selected dv_timings */
struct v4l2_dv_timings dv_timings;
/* used to store pixel format */
struct v4l2_pix_format fmt;
/* bits per pixel*/
int bpp;
/* data length for ppi in bits */
int dlen;
/* used to store sensor supported format */
struct bcap_format *sensor_formats;
/* number of sensor formats array */
int num_sensor_formats;
/* pointing to current video buffer */
struct bcap_buffer *cur_frm;
/* buffer queue used in videobuf2 */
struct vb2_queue buffer_queue;
/* allocator-specific contexts for each plane */
struct vb2_alloc_ctx *alloc_ctx;
/* queue of filled frames */
struct list_head dma_queue;
/* used in videobuf2 callback */
spinlock_t lock;
/* used to access capture device */
struct mutex mutex;
/* used to wait ppi to complete one transfer */
struct completion comp;
/* prepare to stop */
bool stop;
/* vb2 buffer sequence counter */
unsigned sequence;
};
static const struct bcap_format bcap_formats[] = {
{
.desc = "YCbCr 4:2:2 Interleaved UYVY",
.pixelformat = V4L2_PIX_FMT_UYVY,
.mbus_code = MEDIA_BUS_FMT_UYVY8_2X8,
.bpp = 16,
.dlen = 8,
},
{
.desc = "YCbCr 4:2:2 Interleaved YUYV",
.pixelformat = V4L2_PIX_FMT_YUYV,
.mbus_code = MEDIA_BUS_FMT_YUYV8_2X8,
.bpp = 16,
.dlen = 8,
},
{
.desc = "YCbCr 4:2:2 Interleaved UYVY",
.pixelformat = V4L2_PIX_FMT_UYVY,
.mbus_code = MEDIA_BUS_FMT_UYVY8_1X16,
.bpp = 16,
.dlen = 16,
},
{
.desc = "RGB 565",
.pixelformat = V4L2_PIX_FMT_RGB565,
.mbus_code = MEDIA_BUS_FMT_RGB565_2X8_LE,
.bpp = 16,
.dlen = 8,
},
{
.desc = "RGB 444",
.pixelformat = V4L2_PIX_FMT_RGB444,
.mbus_code = MEDIA_BUS_FMT_RGB444_2X8_PADHI_LE,
.bpp = 16,
.dlen = 8,
},
};
#define BCAP_MAX_FMTS ARRAY_SIZE(bcap_formats)
static irqreturn_t bcap_isr(int irq, void *dev_id);
static struct bcap_buffer *to_bcap_vb(struct vb2_v4l2_buffer *vb)
{
return container_of(vb, struct bcap_buffer, vb);
}
static int bcap_init_sensor_formats(struct bcap_device *bcap_dev)
{
struct v4l2_subdev_mbus_code_enum code = {
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
};
struct bcap_format *sf;
unsigned int num_formats = 0;
int i, j;
while (!v4l2_subdev_call(bcap_dev->sd, pad,
enum_mbus_code, NULL, &code)) {
num_formats++;
code.index++;
}
if (!num_formats)
return -ENXIO;
sf = kzalloc(num_formats * sizeof(*sf), GFP_KERNEL);
if (!sf)
return -ENOMEM;
for (i = 0; i < num_formats; i++) {
code.index = i;
v4l2_subdev_call(bcap_dev->sd, pad,
enum_mbus_code, NULL, &code);
for (j = 0; j < BCAP_MAX_FMTS; j++)
if (code.code == bcap_formats[j].mbus_code)
break;
if (j == BCAP_MAX_FMTS) {
/* we don't allow this sensor working with our bridge */
kfree(sf);
return -EINVAL;
}
sf[i] = bcap_formats[j];
}
bcap_dev->sensor_formats = sf;
bcap_dev->num_sensor_formats = num_formats;
return 0;
}
static void bcap_free_sensor_formats(struct bcap_device *bcap_dev)
{
bcap_dev->num_sensor_formats = 0;
kfree(bcap_dev->sensor_formats);
bcap_dev->sensor_formats = NULL;
}
static int bcap_queue_setup(struct vb2_queue *vq,
unsigned int *nbuffers, unsigned int *nplanes,
unsigned int sizes[], void *alloc_ctxs[])
{
struct bcap_device *bcap_dev = vb2_get_drv_priv(vq);
if (vq->num_buffers + *nbuffers < 2)
*nbuffers = 2;
alloc_ctxs[0] = bcap_dev->alloc_ctx;
if (*nplanes)
return sizes[0] < bcap_dev->fmt.sizeimage ? -EINVAL : 0;
*nplanes = 1;
sizes[0] = bcap_dev->fmt.sizeimage;
return 0;
}
static int bcap_buffer_prepare(struct vb2_buffer *vb)
{
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
struct bcap_device *bcap_dev = vb2_get_drv_priv(vb->vb2_queue);
unsigned long size = bcap_dev->fmt.sizeimage;
if (vb2_plane_size(vb, 0) < size) {
v4l2_err(&bcap_dev->v4l2_dev, "buffer too small (%lu < %lu)\n",
vb2_plane_size(vb, 0), size);
return -EINVAL;
}
vb2_set_plane_payload(vb, 0, size);
vbuf->field = bcap_dev->fmt.field;
return 0;
}
static void bcap_buffer_queue(struct vb2_buffer *vb)
{
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
struct bcap_device *bcap_dev = vb2_get_drv_priv(vb->vb2_queue);
struct bcap_buffer *buf = to_bcap_vb(vbuf);
unsigned long flags;
spin_lock_irqsave(&bcap_dev->lock, flags);
list_add_tail(&buf->list, &bcap_dev->dma_queue);
spin_unlock_irqrestore(&bcap_dev->lock, flags);
}
static void bcap_buffer_cleanup(struct vb2_buffer *vb)
{
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
struct bcap_device *bcap_dev = vb2_get_drv_priv(vb->vb2_queue);
struct bcap_buffer *buf = to_bcap_vb(vbuf);
unsigned long flags;
spin_lock_irqsave(&bcap_dev->lock, flags);
list_del_init(&buf->list);
spin_unlock_irqrestore(&bcap_dev->lock, flags);
}
static int bcap_start_streaming(struct vb2_queue *vq, unsigned int count)
{
struct bcap_device *bcap_dev = vb2_get_drv_priv(vq);
struct ppi_if *ppi = bcap_dev->ppi;
struct bcap_buffer *buf, *tmp;
struct ppi_params params;
dma_addr_t addr;
int ret;
/* enable streamon on the sub device */
ret = v4l2_subdev_call(bcap_dev->sd, video, s_stream, 1);
if (ret && (ret != -ENOIOCTLCMD)) {
v4l2_err(&bcap_dev->v4l2_dev, "stream on failed in subdev\n");
goto err;
}
/* set ppi params */
params.width = bcap_dev->fmt.width;
params.height = bcap_dev->fmt.height;
params.bpp = bcap_dev->bpp;
params.dlen = bcap_dev->dlen;
params.ppi_control = bcap_dev->cfg->ppi_control;
params.int_mask = bcap_dev->cfg->int_mask;
if (bcap_dev->cfg->inputs[bcap_dev->cur_input].capabilities
& V4L2_IN_CAP_DV_TIMINGS) {
struct v4l2_bt_timings *bt = &bcap_dev->dv_timings.bt;
params.hdelay = bt->hsync + bt->hbackporch;
params.vdelay = bt->vsync + bt->vbackporch;
params.line = V4L2_DV_BT_FRAME_WIDTH(bt);
params.frame = V4L2_DV_BT_FRAME_HEIGHT(bt);
} else if (bcap_dev->cfg->inputs[bcap_dev->cur_input].capabilities
& V4L2_IN_CAP_STD) {
params.hdelay = 0;
params.vdelay = 0;
if (bcap_dev->std & V4L2_STD_525_60) {
params.line = 858;
params.frame = 525;
} else {
params.line = 864;
params.frame = 625;
}
} else {
params.hdelay = 0;
params.vdelay = 0;
params.line = params.width + bcap_dev->cfg->blank_pixels;
params.frame = params.height;
}
ret = ppi->ops->set_params(ppi, &params);
if (ret < 0) {
v4l2_err(&bcap_dev->v4l2_dev,
"Error in setting ppi params\n");
goto err;
}
/* attach ppi DMA irq handler */
ret = ppi->ops->attach_irq(ppi, bcap_isr);
if (ret < 0) {
v4l2_err(&bcap_dev->v4l2_dev,
"Error in attaching interrupt handler\n");
goto err;
}
bcap_dev->sequence = 0;
reinit_completion(&bcap_dev->comp);
bcap_dev->stop = false;
/* get the next frame from the dma queue */
bcap_dev->cur_frm = list_entry(bcap_dev->dma_queue.next,
struct bcap_buffer, list);
/* remove buffer from the dma queue */
list_del_init(&bcap_dev->cur_frm->list);
addr = vb2_dma_contig_plane_dma_addr(&bcap_dev->cur_frm->vb.vb2_buf,
0);
/* update DMA address */
ppi->ops->update_addr(ppi, (unsigned long)addr);
/* enable ppi */
ppi->ops->start(ppi);
return 0;
err:
list_for_each_entry_safe(buf, tmp, &bcap_dev->dma_queue, list) {
list_del(&buf->list);
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
}
return ret;
}
static void bcap_stop_streaming(struct vb2_queue *vq)
{
struct bcap_device *bcap_dev = vb2_get_drv_priv(vq);
struct ppi_if *ppi = bcap_dev->ppi;
int ret;
bcap_dev->stop = true;
wait_for_completion(&bcap_dev->comp);
ppi->ops->stop(ppi);
ppi->ops->detach_irq(ppi);
ret = v4l2_subdev_call(bcap_dev->sd, video, s_stream, 0);
if (ret && (ret != -ENOIOCTLCMD))
v4l2_err(&bcap_dev->v4l2_dev,
"stream off failed in subdev\n");
/* release all active buffers */
if (bcap_dev->cur_frm)
vb2_buffer_done(&bcap_dev->cur_frm->vb.vb2_buf,
VB2_BUF_STATE_ERROR);
while (!list_empty(&bcap_dev->dma_queue)) {
bcap_dev->cur_frm = list_entry(bcap_dev->dma_queue.next,
struct bcap_buffer, list);
list_del_init(&bcap_dev->cur_frm->list);
vb2_buffer_done(&bcap_dev->cur_frm->vb.vb2_buf,
VB2_BUF_STATE_ERROR);
}
}
static struct vb2_ops bcap_video_qops = {
.queue_setup = bcap_queue_setup,
.buf_prepare = bcap_buffer_prepare,
.buf_cleanup = bcap_buffer_cleanup,
.buf_queue = bcap_buffer_queue,
.wait_prepare = vb2_ops_wait_prepare,
.wait_finish = vb2_ops_wait_finish,
.start_streaming = bcap_start_streaming,
.stop_streaming = bcap_stop_streaming,
};
static irqreturn_t bcap_isr(int irq, void *dev_id)
{
struct ppi_if *ppi = dev_id;
struct bcap_device *bcap_dev = ppi->priv;
struct vb2_v4l2_buffer *vbuf = &bcap_dev->cur_frm->vb;
struct vb2_buffer *vb = &vbuf->vb2_buf;
dma_addr_t addr;
spin_lock(&bcap_dev->lock);
if (!list_empty(&bcap_dev->dma_queue)) {
vb->timestamp = ktime_get_ns();
if (ppi->err) {
vb2_buffer_done(vb, VB2_BUF_STATE_ERROR);
ppi->err = false;
} else {
vbuf->sequence = bcap_dev->sequence++;
vb2_buffer_done(vb, VB2_BUF_STATE_DONE);
}
bcap_dev->cur_frm = list_entry(bcap_dev->dma_queue.next,
struct bcap_buffer, list);
list_del_init(&bcap_dev->cur_frm->list);
} else {
/* clear error flag, we will get a new frame */
if (ppi->err)
ppi->err = false;
}
ppi->ops->stop(ppi);
if (bcap_dev->stop) {
complete(&bcap_dev->comp);
} else {
addr = vb2_dma_contig_plane_dma_addr(
&bcap_dev->cur_frm->vb.vb2_buf, 0);
ppi->ops->update_addr(ppi, (unsigned long)addr);
ppi->ops->start(ppi);
}
spin_unlock(&bcap_dev->lock);
return IRQ_HANDLED;
}
static int bcap_querystd(struct file *file, void *priv, v4l2_std_id *std)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_STD))
return -ENODATA;
return v4l2_subdev_call(bcap_dev->sd, video, querystd, std);
}
static int bcap_g_std(struct file *file, void *priv, v4l2_std_id *std)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_STD))
return -ENODATA;
*std = bcap_dev->std;
return 0;
}
static int bcap_s_std(struct file *file, void *priv, v4l2_std_id std)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
int ret;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_STD))
return -ENODATA;
if (vb2_is_busy(&bcap_dev->buffer_queue))
return -EBUSY;
ret = v4l2_subdev_call(bcap_dev->sd, video, s_std, std);
if (ret < 0)
return ret;
bcap_dev->std = std;
return 0;
}
static int bcap_enum_dv_timings(struct file *file, void *priv,
struct v4l2_enum_dv_timings *timings)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_DV_TIMINGS))
return -ENODATA;
timings->pad = 0;
return v4l2_subdev_call(bcap_dev->sd, pad,
enum_dv_timings, timings);
}
static int bcap_query_dv_timings(struct file *file, void *priv,
struct v4l2_dv_timings *timings)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_DV_TIMINGS))
return -ENODATA;
return v4l2_subdev_call(bcap_dev->sd, video,
query_dv_timings, timings);
}
static int bcap_g_dv_timings(struct file *file, void *priv,
struct v4l2_dv_timings *timings)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_DV_TIMINGS))
return -ENODATA;
*timings = bcap_dev->dv_timings;
return 0;
}
static int bcap_s_dv_timings(struct file *file, void *priv,
struct v4l2_dv_timings *timings)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_input input;
int ret;
input = bcap_dev->cfg->inputs[bcap_dev->cur_input];
if (!(input.capabilities & V4L2_IN_CAP_DV_TIMINGS))
return -ENODATA;
if (vb2_is_busy(&bcap_dev->buffer_queue))
return -EBUSY;
ret = v4l2_subdev_call(bcap_dev->sd, video, s_dv_timings, timings);
if (ret < 0)
return ret;
bcap_dev->dv_timings = *timings;
return 0;
}
static int bcap_enum_input(struct file *file, void *priv,
struct v4l2_input *input)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct bfin_capture_config *config = bcap_dev->cfg;
int ret;
u32 status;
if (input->index >= config->num_inputs)
return -EINVAL;
*input = config->inputs[input->index];
/* get input status */
ret = v4l2_subdev_call(bcap_dev->sd, video, g_input_status, &status);
if (!ret)
input->status = status;
return 0;
}
static int bcap_g_input(struct file *file, void *priv, unsigned int *index)
{
struct bcap_device *bcap_dev = video_drvdata(file);
*index = bcap_dev->cur_input;
return 0;
}
static int bcap_s_input(struct file *file, void *priv, unsigned int index)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct bfin_capture_config *config = bcap_dev->cfg;
struct bcap_route *route;
int ret;
if (vb2_is_busy(&bcap_dev->buffer_queue))
return -EBUSY;
if (index >= config->num_inputs)
return -EINVAL;
route = &config->routes[index];
ret = v4l2_subdev_call(bcap_dev->sd, video, s_routing,
route->input, route->output, 0);
if ((ret < 0) && (ret != -ENOIOCTLCMD)) {
v4l2_err(&bcap_dev->v4l2_dev, "Failed to set input\n");
return ret;
}
bcap_dev->cur_input = index;
/* if this route has specific config, update ppi control */
if (route->ppi_control)
config->ppi_control = route->ppi_control;
return 0;
}
static int bcap_try_format(struct bcap_device *bcap,
struct v4l2_pix_format *pixfmt,
struct bcap_format *bcap_fmt)
{
struct bcap_format *sf = bcap->sensor_formats;
struct bcap_format *fmt = NULL;
struct v4l2_subdev_pad_config pad_cfg;
struct v4l2_subdev_format format = {
.which = V4L2_SUBDEV_FORMAT_TRY,
};
int ret, i;
for (i = 0; i < bcap->num_sensor_formats; i++) {
fmt = &sf[i];
if (pixfmt->pixelformat == fmt->pixelformat)
break;
}
if (i == bcap->num_sensor_formats)
fmt = &sf[0];
v4l2_fill_mbus_format(&format.format, pixfmt, fmt->mbus_code);
ret = v4l2_subdev_call(bcap->sd, pad, set_fmt, &pad_cfg,
&format);
if (ret < 0)
return ret;
v4l2_fill_pix_format(pixfmt, &format.format);
if (bcap_fmt) {
for (i = 0; i < bcap->num_sensor_formats; i++) {
fmt = &sf[i];
if (format.format.code == fmt->mbus_code)
break;
}
*bcap_fmt = *fmt;
}
pixfmt->bytesperline = pixfmt->width * fmt->bpp / 8;
pixfmt->sizeimage = pixfmt->bytesperline * pixfmt->height;
return 0;
}
static int bcap_enum_fmt_vid_cap(struct file *file, void *priv,
struct v4l2_fmtdesc *fmt)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct bcap_format *sf = bcap_dev->sensor_formats;
if (fmt->index >= bcap_dev->num_sensor_formats)
return -EINVAL;
fmt->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
strlcpy(fmt->description,
sf[fmt->index].desc,
sizeof(fmt->description));
fmt->pixelformat = sf[fmt->index].pixelformat;
return 0;
}
static int bcap_try_fmt_vid_cap(struct file *file, void *priv,
struct v4l2_format *fmt)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_pix_format *pixfmt = &fmt->fmt.pix;
return bcap_try_format(bcap_dev, pixfmt, NULL);
}
static int bcap_g_fmt_vid_cap(struct file *file, void *priv,
struct v4l2_format *fmt)
{
struct bcap_device *bcap_dev = video_drvdata(file);
fmt->fmt.pix = bcap_dev->fmt;
return 0;
}
static int bcap_s_fmt_vid_cap(struct file *file, void *priv,
struct v4l2_format *fmt)
{
struct bcap_device *bcap_dev = video_drvdata(file);
struct v4l2_subdev_format format = {
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
};
struct bcap_format bcap_fmt;
struct v4l2_pix_format *pixfmt = &fmt->fmt.pix;
int ret;
if (vb2_is_busy(&bcap_dev->buffer_queue))
return -EBUSY;
/* see if format works */
ret = bcap_try_format(bcap_dev, pixfmt, &bcap_fmt);
if (ret < 0)
return ret;
v4l2_fill_mbus_format(&format.format, pixfmt, bcap_fmt.mbus_code);
ret = v4l2_subdev_call(bcap_dev->sd, pad, set_fmt, NULL, &format);
if (ret < 0)
return ret;
bcap_dev->fmt = *pixfmt;
bcap_dev->bpp = bcap_fmt.bpp;
bcap_dev->dlen = bcap_fmt.dlen;
return 0;
}
static int bcap_querycap(struct file *file, void *priv,
struct v4l2_capability *cap)
{
struct bcap_device *bcap_dev = video_drvdata(file);
cap->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING;
cap->capabilities = cap->device_caps | V4L2_CAP_DEVICE_CAPS;
strlcpy(cap->driver, CAPTURE_DRV_NAME, sizeof(cap->driver));
strlcpy(cap->bus_info, "Blackfin Platform", sizeof(cap->bus_info));
strlcpy(cap->card, bcap_dev->cfg->card_name, sizeof(cap->card));
return 0;
}
static int bcap_g_parm(struct file *file, void *fh,
struct v4l2_streamparm *a)
{
struct bcap_device *bcap_dev = video_drvdata(file);
if (a->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
return -EINVAL;
return v4l2_subdev_call(bcap_dev->sd, video, g_parm, a);
}
static int bcap_s_parm(struct file *file, void *fh,
struct v4l2_streamparm *a)
{
struct bcap_device *bcap_dev = video_drvdata(file);
if (a->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
return -EINVAL;
return v4l2_subdev_call(bcap_dev->sd, video, s_parm, a);
}
static int bcap_log_status(struct file *file, void *priv)
{
struct bcap_device *bcap_dev = video_drvdata(file);
/* status for sub devices */
v4l2_device_call_all(&bcap_dev->v4l2_dev, 0, core, log_status);
return 0;
}
static const struct v4l2_ioctl_ops bcap_ioctl_ops = {
.vidioc_querycap = bcap_querycap,
.vidioc_g_fmt_vid_cap = bcap_g_fmt_vid_cap,
.vidioc_enum_fmt_vid_cap = bcap_enum_fmt_vid_cap,
.vidioc_s_fmt_vid_cap = bcap_s_fmt_vid_cap,
.vidioc_try_fmt_vid_cap = bcap_try_fmt_vid_cap,
.vidioc_enum_input = bcap_enum_input,
.vidioc_g_input = bcap_g_input,
.vidioc_s_input = bcap_s_input,
.vidioc_querystd = bcap_querystd,
.vidioc_s_std = bcap_s_std,
.vidioc_g_std = bcap_g_std,
.vidioc_s_dv_timings = bcap_s_dv_timings,
.vidioc_g_dv_timings = bcap_g_dv_timings,
.vidioc_query_dv_timings = bcap_query_dv_timings,
.vidioc_enum_dv_timings = bcap_enum_dv_timings,
.vidioc_reqbufs = vb2_ioctl_reqbufs,
.vidioc_create_bufs = vb2_ioctl_create_bufs,
.vidioc_querybuf = vb2_ioctl_querybuf,
.vidioc_qbuf = vb2_ioctl_qbuf,
.vidioc_dqbuf = vb2_ioctl_dqbuf,
#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,5,0)
.vidioc_expbuf = vb2_ioctl_expbuf,
#endif
.vidioc_streamon = vb2_ioctl_streamon,
.vidioc_streamoff = vb2_ioctl_streamoff,
.vidioc_g_parm = bcap_g_parm,
.vidioc_s_parm = bcap_s_parm,
.vidioc_log_status = bcap_log_status,
};
static struct v4l2_file_operations bcap_fops = {
.owner = THIS_MODULE,
.open = v4l2_fh_open,
.release = vb2_fop_release,
.unlocked_ioctl = video_ioctl2,
.mmap = vb2_fop_mmap,
#ifndef CONFIG_MMU
.get_unmapped_area = vb2_fop_get_unmapped_area,
#endif
.poll = vb2_fop_poll
};
static int bcap_probe(struct platform_device *pdev)
{
struct bcap_device *bcap_dev;
struct video_device *vfd;
struct i2c_adapter *i2c_adap;
struct bfin_capture_config *config;
struct vb2_queue *q;
struct bcap_route *route;
int ret;
config = pdev->dev.platform_data;
if (!config || !config->num_inputs) {
v4l2_err(pdev->dev.driver, "Unable to get board config\n");
return -ENODEV;
}
bcap_dev = kzalloc(sizeof(*bcap_dev), GFP_KERNEL);
if (!bcap_dev) {
v4l2_err(pdev->dev.driver, "Unable to alloc bcap_dev\n");
return -ENOMEM;
}
bcap_dev->cfg = config;
bcap_dev->ppi = ppi_create_instance(pdev, config->ppi_info);
if (!bcap_dev->ppi) {
v4l2_err(pdev->dev.driver, "Unable to create ppi\n");
ret = -ENODEV;
goto err_free_dev;
}
bcap_dev->ppi->priv = bcap_dev;
bcap_dev->alloc_ctx = vb2_dma_contig_init_ctx(&pdev->dev);
if (IS_ERR(bcap_dev->alloc_ctx)) {
ret = PTR_ERR(bcap_dev->alloc_ctx);
goto err_free_ppi;
}
vfd = &bcap_dev->video_dev;
/* initialize field of video device */
vfd->release = video_device_release_empty;
vfd->fops = &bcap_fops;
vfd->ioctl_ops = &bcap_ioctl_ops;
vfd->tvnorms = 0;
vfd->v4l2_dev = &bcap_dev->v4l2_dev;
strncpy(vfd->name, CAPTURE_DRV_NAME, sizeof(vfd->name));
ret = v4l2_device_register(&pdev->dev, &bcap_dev->v4l2_dev);
if (ret) {
v4l2_err(pdev->dev.driver,
"Unable to register v4l2 device\n");
goto err_cleanup_ctx;
}
v4l2_info(&bcap_dev->v4l2_dev, "v4l2 device registered\n");
bcap_dev->v4l2_dev.ctrl_handler = &bcap_dev->ctrl_handler;
ret = v4l2_ctrl_handler_init(&bcap_dev->ctrl_handler, 0);
if (ret) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to init control handler\n");
goto err_unreg_v4l2;
}
spin_lock_init(&bcap_dev->lock);
/* initialize queue */
q = &bcap_dev->buffer_queue;
q->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
q->io_modes = VB2_MMAP | VB2_DMABUF;
q->drv_priv = bcap_dev;
q->buf_struct_size = sizeof(struct bcap_buffer);
q->ops = &bcap_video_qops;
q->mem_ops = &vb2_dma_contig_memops;
q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
q->lock = &bcap_dev->mutex;
q->min_buffers_needed = 1;
ret = vb2_queue_init(q);
if (ret)
goto err_free_handler;
mutex_init(&bcap_dev->mutex);
init_completion(&bcap_dev->comp);
/* init video dma queues */
INIT_LIST_HEAD(&bcap_dev->dma_queue);
vfd->lock = &bcap_dev->mutex;
vfd->queue = q;
/* register video device */
ret = video_register_device(&bcap_dev->video_dev, VFL_TYPE_GRABBER, -1);
if (ret) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to register video device\n");
goto err_free_handler;
}
video_set_drvdata(&bcap_dev->video_dev, bcap_dev);
v4l2_info(&bcap_dev->v4l2_dev, "video device registered as: %s\n",
video_device_node_name(vfd));
/* load up the subdevice */
i2c_adap = i2c_get_adapter(config->i2c_adapter_id);
if (!i2c_adap) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to find i2c adapter\n");
ret = -ENODEV;
goto err_unreg_vdev;
}
bcap_dev->sd = v4l2_i2c_new_subdev_board(&bcap_dev->v4l2_dev,
i2c_adap,
&config->board_info,
NULL);
if (bcap_dev->sd) {
int i;
/* update tvnorms from the sub devices */
for (i = 0; i < config->num_inputs; i++)
vfd->tvnorms |= config->inputs[i].std;
} else {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to register sub device\n");
ret = -ENODEV;
goto err_unreg_vdev;
}
v4l2_info(&bcap_dev->v4l2_dev, "v4l2 sub device registered\n");
/*
* explicitly set input, otherwise some boards
* may not work at the state as we expected
*/
route = &config->routes[0];
ret = v4l2_subdev_call(bcap_dev->sd, video, s_routing,
route->input, route->output, 0);
if ((ret < 0) && (ret != -ENOIOCTLCMD)) {
v4l2_err(&bcap_dev->v4l2_dev, "Failed to set input\n");
goto err_unreg_vdev;
}
bcap_dev->cur_input = 0;
/* if this route has specific config, update ppi control */
if (route->ppi_control)
config->ppi_control = route->ppi_control;
/* now we can probe the default state */
if (config->inputs[0].capabilities & V4L2_IN_CAP_STD) {
v4l2_std_id std;
ret = v4l2_subdev_call(bcap_dev->sd, video, g_std, &std);
if (ret) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to get std\n");
goto err_unreg_vdev;
}
bcap_dev->std = std;
}
if (config->inputs[0].capabilities & V4L2_IN_CAP_DV_TIMINGS) {
struct v4l2_dv_timings dv_timings;
ret = v4l2_subdev_call(bcap_dev->sd, video,
g_dv_timings, &dv_timings);
if (ret) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to get dv timings\n");
goto err_unreg_vdev;
}
bcap_dev->dv_timings = dv_timings;
}
ret = bcap_init_sensor_formats(bcap_dev);
if (ret) {
v4l2_err(&bcap_dev->v4l2_dev,
"Unable to create sensor formats table\n");
goto err_unreg_vdev;
}
return 0;
err_unreg_vdev:
video_unregister_device(&bcap_dev->video_dev);
err_free_handler:
v4l2_ctrl_handler_free(&bcap_dev->ctrl_handler);
err_unreg_v4l2:
v4l2_device_unregister(&bcap_dev->v4l2_dev);
err_cleanup_ctx:
vb2_dma_contig_cleanup_ctx(bcap_dev->alloc_ctx);
err_free_ppi:
ppi_delete_instance(bcap_dev->ppi);
err_free_dev:
kfree(bcap_dev);
return ret;
}
static int bcap_remove(struct platform_device *pdev)
{
struct v4l2_device *v4l2_dev = platform_get_drvdata(pdev);
struct bcap_device *bcap_dev = container_of(v4l2_dev,
struct bcap_device, v4l2_dev);
bcap_free_sensor_formats(bcap_dev);
video_unregister_device(&bcap_dev->video_dev);
v4l2_ctrl_handler_free(&bcap_dev->ctrl_handler);
v4l2_device_unregister(v4l2_dev);
vb2_dma_contig_cleanup_ctx(bcap_dev->alloc_ctx);
ppi_delete_instance(bcap_dev->ppi);
kfree(bcap_dev);
return 0;
}
static struct platform_driver bcap_driver = {
.driver = {
.name = CAPTURE_DRV_NAME,
},
.probe = bcap_probe,
.remove = bcap_remove,
};
module_platform_driver(bcap_driver);
MODULE_DESCRIPTION("Analog Devices blackfin video capture driver");
MODULE_AUTHOR("Scott Jiang <Scott.Jiang.Linux@gmail.com>");
MODULE_LICENSE("GPL v2");