2591 lines
69 KiB
C
2591 lines
69 KiB
C
/*
|
|
* V4L2 Driver for PXA camera host
|
|
*
|
|
* Copyright (C) 2006, Sascha Hauer, Pengutronix
|
|
* Copyright (C) 2008, Guennadi Liakhovetski <kernel@pengutronix.de>
|
|
* Copyright (C) 2016, Robert Jarzmik <robert.jarzmik@free.fr>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/module.h>
|
|
#include <linux/io.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/device.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/err.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_graph.h>
|
|
#include <linux/time.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/dma/pxa-dma.h>
|
|
|
|
#include <media/v4l2-async.h>
|
|
#include <media/v4l2-clk.h>
|
|
#include <media/v4l2-common.h>
|
|
#include <media/v4l2-ctrls.h>
|
|
#include <media/v4l2-device.h>
|
|
#include <media/v4l2-event.h>
|
|
#include <media/v4l2-ioctl.h>
|
|
#include <media/v4l2-fwnode.h>
|
|
|
|
#include <media/videobuf2-dma-sg.h>
|
|
|
|
#include <linux/videodev2.h>
|
|
|
|
#include <linux/platform_data/media/camera-pxa.h>
|
|
|
|
#define PXA_CAM_VERSION "0.0.6"
|
|
#define PXA_CAM_DRV_NAME "pxa27x-camera"
|
|
|
|
#define DEFAULT_WIDTH 640
|
|
#define DEFAULT_HEIGHT 480
|
|
|
|
/* Camera Interface */
|
|
#define CICR0 0x0000
|
|
#define CICR1 0x0004
|
|
#define CICR2 0x0008
|
|
#define CICR3 0x000C
|
|
#define CICR4 0x0010
|
|
#define CISR 0x0014
|
|
#define CIFR 0x0018
|
|
#define CITOR 0x001C
|
|
#define CIBR0 0x0028
|
|
#define CIBR1 0x0030
|
|
#define CIBR2 0x0038
|
|
|
|
#define CICR0_DMAEN (1 << 31) /* DMA request enable */
|
|
#define CICR0_PAR_EN (1 << 30) /* Parity enable */
|
|
#define CICR0_SL_CAP_EN (1 << 29) /* Capture enable for slave mode */
|
|
#define CICR0_ENB (1 << 28) /* Camera interface enable */
|
|
#define CICR0_DIS (1 << 27) /* Camera interface disable */
|
|
#define CICR0_SIM (0x7 << 24) /* Sensor interface mode mask */
|
|
#define CICR0_TOM (1 << 9) /* Time-out mask */
|
|
#define CICR0_RDAVM (1 << 8) /* Receive-data-available mask */
|
|
#define CICR0_FEM (1 << 7) /* FIFO-empty mask */
|
|
#define CICR0_EOLM (1 << 6) /* End-of-line mask */
|
|
#define CICR0_PERRM (1 << 5) /* Parity-error mask */
|
|
#define CICR0_QDM (1 << 4) /* Quick-disable mask */
|
|
#define CICR0_CDM (1 << 3) /* Disable-done mask */
|
|
#define CICR0_SOFM (1 << 2) /* Start-of-frame mask */
|
|
#define CICR0_EOFM (1 << 1) /* End-of-frame mask */
|
|
#define CICR0_FOM (1 << 0) /* FIFO-overrun mask */
|
|
|
|
#define CICR1_TBIT (1 << 31) /* Transparency bit */
|
|
#define CICR1_RGBT_CONV (0x3 << 29) /* RGBT conversion mask */
|
|
#define CICR1_PPL (0x7ff << 15) /* Pixels per line mask */
|
|
#define CICR1_RGB_CONV (0x7 << 12) /* RGB conversion mask */
|
|
#define CICR1_RGB_F (1 << 11) /* RGB format */
|
|
#define CICR1_YCBCR_F (1 << 10) /* YCbCr format */
|
|
#define CICR1_RGB_BPP (0x7 << 7) /* RGB bis per pixel mask */
|
|
#define CICR1_RAW_BPP (0x3 << 5) /* Raw bis per pixel mask */
|
|
#define CICR1_COLOR_SP (0x3 << 3) /* Color space mask */
|
|
#define CICR1_DW (0x7 << 0) /* Data width mask */
|
|
|
|
#define CICR2_BLW (0xff << 24) /* Beginning-of-line pixel clock
|
|
wait count mask */
|
|
#define CICR2_ELW (0xff << 16) /* End-of-line pixel clock
|
|
wait count mask */
|
|
#define CICR2_HSW (0x3f << 10) /* Horizontal sync pulse width mask */
|
|
#define CICR2_BFPW (0x3f << 3) /* Beginning-of-frame pixel clock
|
|
wait count mask */
|
|
#define CICR2_FSW (0x7 << 0) /* Frame stabilization
|
|
wait count mask */
|
|
|
|
#define CICR3_BFW (0xff << 24) /* Beginning-of-frame line clock
|
|
wait count mask */
|
|
#define CICR3_EFW (0xff << 16) /* End-of-frame line clock
|
|
wait count mask */
|
|
#define CICR3_VSW (0x3f << 10) /* Vertical sync pulse width mask */
|
|
#define CICR3_BFPW (0x3f << 3) /* Beginning-of-frame pixel clock
|
|
wait count mask */
|
|
#define CICR3_LPF (0x7ff << 0) /* Lines per frame mask */
|
|
|
|
#define CICR4_MCLK_DLY (0x3 << 24) /* MCLK Data Capture Delay mask */
|
|
#define CICR4_PCLK_EN (1 << 23) /* Pixel clock enable */
|
|
#define CICR4_PCP (1 << 22) /* Pixel clock polarity */
|
|
#define CICR4_HSP (1 << 21) /* Horizontal sync polarity */
|
|
#define CICR4_VSP (1 << 20) /* Vertical sync polarity */
|
|
#define CICR4_MCLK_EN (1 << 19) /* MCLK enable */
|
|
#define CICR4_FR_RATE (0x7 << 8) /* Frame rate mask */
|
|
#define CICR4_DIV (0xff << 0) /* Clock divisor mask */
|
|
|
|
#define CISR_FTO (1 << 15) /* FIFO time-out */
|
|
#define CISR_RDAV_2 (1 << 14) /* Channel 2 receive data available */
|
|
#define CISR_RDAV_1 (1 << 13) /* Channel 1 receive data available */
|
|
#define CISR_RDAV_0 (1 << 12) /* Channel 0 receive data available */
|
|
#define CISR_FEMPTY_2 (1 << 11) /* Channel 2 FIFO empty */
|
|
#define CISR_FEMPTY_1 (1 << 10) /* Channel 1 FIFO empty */
|
|
#define CISR_FEMPTY_0 (1 << 9) /* Channel 0 FIFO empty */
|
|
#define CISR_EOL (1 << 8) /* End of line */
|
|
#define CISR_PAR_ERR (1 << 7) /* Parity error */
|
|
#define CISR_CQD (1 << 6) /* Camera interface quick disable */
|
|
#define CISR_CDD (1 << 5) /* Camera interface disable done */
|
|
#define CISR_SOF (1 << 4) /* Start of frame */
|
|
#define CISR_EOF (1 << 3) /* End of frame */
|
|
#define CISR_IFO_2 (1 << 2) /* FIFO overrun for Channel 2 */
|
|
#define CISR_IFO_1 (1 << 1) /* FIFO overrun for Channel 1 */
|
|
#define CISR_IFO_0 (1 << 0) /* FIFO overrun for Channel 0 */
|
|
|
|
#define CIFR_FLVL2 (0x7f << 23) /* FIFO 2 level mask */
|
|
#define CIFR_FLVL1 (0x7f << 16) /* FIFO 1 level mask */
|
|
#define CIFR_FLVL0 (0xff << 8) /* FIFO 0 level mask */
|
|
#define CIFR_THL_0 (0x3 << 4) /* Threshold Level for Channel 0 FIFO */
|
|
#define CIFR_RESET_F (1 << 3) /* Reset input FIFOs */
|
|
#define CIFR_FEN2 (1 << 2) /* FIFO enable for channel 2 */
|
|
#define CIFR_FEN1 (1 << 1) /* FIFO enable for channel 1 */
|
|
#define CIFR_FEN0 (1 << 0) /* FIFO enable for channel 0 */
|
|
|
|
#define CICR0_SIM_MP (0 << 24)
|
|
#define CICR0_SIM_SP (1 << 24)
|
|
#define CICR0_SIM_MS (2 << 24)
|
|
#define CICR0_SIM_EP (3 << 24)
|
|
#define CICR0_SIM_ES (4 << 24)
|
|
|
|
#define CICR1_DW_VAL(x) ((x) & CICR1_DW) /* Data bus width */
|
|
#define CICR1_PPL_VAL(x) (((x) << 15) & CICR1_PPL) /* Pixels per line */
|
|
#define CICR1_COLOR_SP_VAL(x) (((x) << 3) & CICR1_COLOR_SP) /* color space */
|
|
#define CICR1_RGB_BPP_VAL(x) (((x) << 7) & CICR1_RGB_BPP) /* bpp for rgb */
|
|
#define CICR1_RGBT_CONV_VAL(x) (((x) << 29) & CICR1_RGBT_CONV) /* rgbt conv */
|
|
|
|
#define CICR2_BLW_VAL(x) (((x) << 24) & CICR2_BLW) /* Beginning-of-line pixel clock wait count */
|
|
#define CICR2_ELW_VAL(x) (((x) << 16) & CICR2_ELW) /* End-of-line pixel clock wait count */
|
|
#define CICR2_HSW_VAL(x) (((x) << 10) & CICR2_HSW) /* Horizontal sync pulse width */
|
|
#define CICR2_BFPW_VAL(x) (((x) << 3) & CICR2_BFPW) /* Beginning-of-frame pixel clock wait count */
|
|
#define CICR2_FSW_VAL(x) (((x) << 0) & CICR2_FSW) /* Frame stabilization wait count */
|
|
|
|
#define CICR3_BFW_VAL(x) (((x) << 24) & CICR3_BFW) /* Beginning-of-frame line clock wait count */
|
|
#define CICR3_EFW_VAL(x) (((x) << 16) & CICR3_EFW) /* End-of-frame line clock wait count */
|
|
#define CICR3_VSW_VAL(x) (((x) << 11) & CICR3_VSW) /* Vertical sync pulse width */
|
|
#define CICR3_LPF_VAL(x) (((x) << 0) & CICR3_LPF) /* Lines per frame */
|
|
|
|
#define CICR0_IRQ_MASK (CICR0_TOM | CICR0_RDAVM | CICR0_FEM | CICR0_EOLM | \
|
|
CICR0_PERRM | CICR0_QDM | CICR0_CDM | CICR0_SOFM | \
|
|
CICR0_EOFM | CICR0_FOM)
|
|
|
|
#define sensor_call(cam, o, f, args...) \
|
|
v4l2_subdev_call(cam->sensor, o, f, ##args)
|
|
|
|
/*
|
|
* Format handling
|
|
*/
|
|
|
|
/**
|
|
* enum pxa_mbus_packing - data packing types on the media-bus
|
|
* @PXA_MBUS_PACKING_NONE: no packing, bit-for-bit transfer to RAM, one
|
|
* sample represents one pixel
|
|
* @PXA_MBUS_PACKING_2X8_PADHI: 16 bits transferred in 2 8-bit samples, in the
|
|
* possibly incomplete byte high bits are padding
|
|
* @PXA_MBUS_PACKING_EXTEND16: sample width (e.g., 10 bits) has to be extended
|
|
* to 16 bits
|
|
*/
|
|
enum pxa_mbus_packing {
|
|
PXA_MBUS_PACKING_NONE,
|
|
PXA_MBUS_PACKING_2X8_PADHI,
|
|
PXA_MBUS_PACKING_EXTEND16,
|
|
};
|
|
|
|
/**
|
|
* enum pxa_mbus_order - sample order on the media bus
|
|
* @PXA_MBUS_ORDER_LE: least significant sample first
|
|
* @PXA_MBUS_ORDER_BE: most significant sample first
|
|
*/
|
|
enum pxa_mbus_order {
|
|
PXA_MBUS_ORDER_LE,
|
|
PXA_MBUS_ORDER_BE,
|
|
};
|
|
|
|
/**
|
|
* enum pxa_mbus_layout - planes layout in memory
|
|
* @PXA_MBUS_LAYOUT_PACKED: color components packed
|
|
* @PXA_MBUS_LAYOUT_PLANAR_2Y_U_V: YUV components stored in 3 planes (4:2:2)
|
|
* @PXA_MBUS_LAYOUT_PLANAR_2Y_C: YUV components stored in a luma and a
|
|
* chroma plane (C plane is half the size
|
|
* of Y plane)
|
|
* @PXA_MBUS_LAYOUT_PLANAR_Y_C: YUV components stored in a luma and a
|
|
* chroma plane (C plane is the same size
|
|
* as Y plane)
|
|
*/
|
|
enum pxa_mbus_layout {
|
|
PXA_MBUS_LAYOUT_PACKED = 0,
|
|
PXA_MBUS_LAYOUT_PLANAR_2Y_U_V,
|
|
PXA_MBUS_LAYOUT_PLANAR_2Y_C,
|
|
PXA_MBUS_LAYOUT_PLANAR_Y_C,
|
|
};
|
|
|
|
/**
|
|
* struct pxa_mbus_pixelfmt - Data format on the media bus
|
|
* @name: Name of the format
|
|
* @fourcc: Fourcc code, that will be obtained if the data is
|
|
* stored in memory in the following way:
|
|
* @packing: Type of sample-packing, that has to be used
|
|
* @order: Sample order when storing in memory
|
|
* @layout: Planes layout in memory
|
|
* @bits_per_sample: How many bits the bridge has to sample
|
|
*/
|
|
struct pxa_mbus_pixelfmt {
|
|
const char *name;
|
|
u32 fourcc;
|
|
enum pxa_mbus_packing packing;
|
|
enum pxa_mbus_order order;
|
|
enum pxa_mbus_layout layout;
|
|
u8 bits_per_sample;
|
|
};
|
|
|
|
/**
|
|
* struct pxa_mbus_lookup - Lookup FOURCC IDs by mediabus codes for pass-through
|
|
* @code: mediabus pixel-code
|
|
* @fmt: pixel format description
|
|
*/
|
|
struct pxa_mbus_lookup {
|
|
u32 code;
|
|
struct pxa_mbus_pixelfmt fmt;
|
|
};
|
|
|
|
static const struct pxa_mbus_lookup mbus_fmt[] = {
|
|
{
|
|
.code = MEDIA_BUS_FMT_YUYV8_2X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_YUYV,
|
|
.name = "YUYV",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_YVYU8_2X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_YVYU,
|
|
.name = "YVYU",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_UYVY8_2X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_UYVY,
|
|
.name = "UYVY",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_VYUY8_2X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_VYUY,
|
|
.name = "VYUY",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_RGB555,
|
|
.name = "RGB555",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_BE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_RGB555X,
|
|
.name = "RGB555X",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_BE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_RGB565_2X8_LE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_RGB565,
|
|
.name = "RGB565",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_RGB565_2X8_BE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_RGB565X,
|
|
.name = "RGB565X",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_BE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SBGGR8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SBGGR8,
|
|
.name = "Bayer 8 BGGR",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGBRG8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGBRG8,
|
|
.name = "Bayer 8 GBRG",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGRBG8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGRBG8,
|
|
.name = "Bayer 8 GRBG",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SRGGB8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SRGGB8,
|
|
.name = "Bayer 8 RGGB",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SBGGR10_1X10,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SBGGR10,
|
|
.name = "Bayer 10 BGGR",
|
|
.bits_per_sample = 10,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_Y8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_GREY,
|
|
.name = "Grey",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_Y10_1X10,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_Y10,
|
|
.name = "Grey 10bit",
|
|
.bits_per_sample = 10,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SBGGR10_2X8_PADHI_LE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SBGGR10,
|
|
.name = "Bayer 10 BGGR",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SBGGR10_2X8_PADHI_BE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SBGGR10,
|
|
.name = "Bayer 10 BGGR",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_BE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_RGB444_2X8_PADHI_BE,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_RGB444,
|
|
.name = "RGB444",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_BE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_UYVY8_1X16,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_UYVY,
|
|
.name = "UYVY 16bit",
|
|
.bits_per_sample = 16,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_VYUY8_1X16,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_VYUY,
|
|
.name = "VYUY 16bit",
|
|
.bits_per_sample = 16,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_YUYV8_1X16,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_YUYV,
|
|
.name = "YUYV 16bit",
|
|
.bits_per_sample = 16,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_YVYU8_1X16,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_YVYU,
|
|
.name = "YVYU 16bit",
|
|
.bits_per_sample = 16,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGRBG10_DPCM8_1X8,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGRBG10DPCM8,
|
|
.name = "Bayer 10 BGGR DPCM 8",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_NONE,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGBRG10_1X10,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGBRG10,
|
|
.name = "Bayer 10 GBRG",
|
|
.bits_per_sample = 10,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGRBG10_1X10,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGRBG10,
|
|
.name = "Bayer 10 GRBG",
|
|
.bits_per_sample = 10,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SRGGB10_1X10,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SRGGB10,
|
|
.name = "Bayer 10 RGGB",
|
|
.bits_per_sample = 10,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SBGGR12_1X12,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SBGGR12,
|
|
.name = "Bayer 12 BGGR",
|
|
.bits_per_sample = 12,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGBRG12_1X12,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGBRG12,
|
|
.name = "Bayer 12 GBRG",
|
|
.bits_per_sample = 12,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SGRBG12_1X12,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SGRBG12,
|
|
.name = "Bayer 12 GRBG",
|
|
.bits_per_sample = 12,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
}, {
|
|
.code = MEDIA_BUS_FMT_SRGGB12_1X12,
|
|
.fmt = {
|
|
.fourcc = V4L2_PIX_FMT_SRGGB12,
|
|
.name = "Bayer 12 RGGB",
|
|
.bits_per_sample = 12,
|
|
.packing = PXA_MBUS_PACKING_EXTEND16,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PACKED,
|
|
},
|
|
},
|
|
};
|
|
|
|
static s32 pxa_mbus_bytes_per_line(u32 width, const struct pxa_mbus_pixelfmt *mf)
|
|
{
|
|
if (mf->layout != PXA_MBUS_LAYOUT_PACKED)
|
|
return width * mf->bits_per_sample / 8;
|
|
|
|
switch (mf->packing) {
|
|
case PXA_MBUS_PACKING_NONE:
|
|
return width * mf->bits_per_sample / 8;
|
|
case PXA_MBUS_PACKING_2X8_PADHI:
|
|
case PXA_MBUS_PACKING_EXTEND16:
|
|
return width * 2;
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
static s32 pxa_mbus_image_size(const struct pxa_mbus_pixelfmt *mf,
|
|
u32 bytes_per_line, u32 height)
|
|
{
|
|
if (mf->layout == PXA_MBUS_LAYOUT_PACKED)
|
|
return bytes_per_line * height;
|
|
|
|
switch (mf->packing) {
|
|
case PXA_MBUS_PACKING_2X8_PADHI:
|
|
return bytes_per_line * height * 2;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
static const struct pxa_mbus_pixelfmt *pxa_mbus_find_fmtdesc(
|
|
u32 code,
|
|
const struct pxa_mbus_lookup *lookup,
|
|
int n)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < n; i++)
|
|
if (lookup[i].code == code)
|
|
return &lookup[i].fmt;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const struct pxa_mbus_pixelfmt *pxa_mbus_get_fmtdesc(
|
|
u32 code)
|
|
{
|
|
return pxa_mbus_find_fmtdesc(code, mbus_fmt, ARRAY_SIZE(mbus_fmt));
|
|
}
|
|
|
|
static unsigned int pxa_mbus_config_compatible(const struct v4l2_mbus_config *cfg,
|
|
unsigned int flags)
|
|
{
|
|
unsigned long common_flags;
|
|
bool hsync = true, vsync = true, pclk, data, mode;
|
|
bool mipi_lanes, mipi_clock;
|
|
|
|
common_flags = cfg->flags & flags;
|
|
|
|
switch (cfg->type) {
|
|
case V4L2_MBUS_PARALLEL:
|
|
hsync = common_flags & (V4L2_MBUS_HSYNC_ACTIVE_HIGH |
|
|
V4L2_MBUS_HSYNC_ACTIVE_LOW);
|
|
vsync = common_flags & (V4L2_MBUS_VSYNC_ACTIVE_HIGH |
|
|
V4L2_MBUS_VSYNC_ACTIVE_LOW);
|
|
/* fall through */
|
|
case V4L2_MBUS_BT656:
|
|
pclk = common_flags & (V4L2_MBUS_PCLK_SAMPLE_RISING |
|
|
V4L2_MBUS_PCLK_SAMPLE_FALLING);
|
|
data = common_flags & (V4L2_MBUS_DATA_ACTIVE_HIGH |
|
|
V4L2_MBUS_DATA_ACTIVE_LOW);
|
|
mode = common_flags & (V4L2_MBUS_MASTER | V4L2_MBUS_SLAVE);
|
|
return (!hsync || !vsync || !pclk || !data || !mode) ?
|
|
0 : common_flags;
|
|
case V4L2_MBUS_CSI2:
|
|
mipi_lanes = common_flags & V4L2_MBUS_CSI2_LANES;
|
|
mipi_clock = common_flags & (V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK |
|
|
V4L2_MBUS_CSI2_CONTINUOUS_CLOCK);
|
|
return (!mipi_lanes || !mipi_clock) ? 0 : common_flags;
|
|
default:
|
|
WARN_ON(1);
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* struct soc_camera_format_xlate - match between host and sensor formats
|
|
* @code: code of a sensor provided format
|
|
* @host_fmt: host format after host translation from code
|
|
*
|
|
* Host and sensor translation structure. Used in table of host and sensor
|
|
* formats matchings in soc_camera_device. A host can override the generic list
|
|
* generation by implementing get_formats(), and use it for format checks and
|
|
* format setup.
|
|
*/
|
|
struct soc_camera_format_xlate {
|
|
u32 code;
|
|
const struct pxa_mbus_pixelfmt *host_fmt;
|
|
};
|
|
|
|
/*
|
|
* Structures
|
|
*/
|
|
enum pxa_camera_active_dma {
|
|
DMA_Y = 0x1,
|
|
DMA_U = 0x2,
|
|
DMA_V = 0x4,
|
|
};
|
|
|
|
/* buffer for one video frame */
|
|
struct pxa_buffer {
|
|
/* common v4l buffer stuff -- must be first */
|
|
struct vb2_v4l2_buffer vbuf;
|
|
struct list_head queue;
|
|
u32 code;
|
|
int nb_planes;
|
|
/* our descriptor lists for Y, U and V channels */
|
|
struct dma_async_tx_descriptor *descs[3];
|
|
dma_cookie_t cookie[3];
|
|
struct scatterlist *sg[3];
|
|
int sg_len[3];
|
|
size_t plane_sizes[3];
|
|
int inwork;
|
|
enum pxa_camera_active_dma active_dma;
|
|
};
|
|
|
|
struct pxa_camera_dev {
|
|
struct v4l2_device v4l2_dev;
|
|
struct video_device vdev;
|
|
struct v4l2_async_notifier notifier;
|
|
struct vb2_queue vb2_vq;
|
|
struct v4l2_subdev *sensor;
|
|
struct soc_camera_format_xlate *user_formats;
|
|
const struct soc_camera_format_xlate *current_fmt;
|
|
struct v4l2_pix_format current_pix;
|
|
|
|
struct v4l2_async_subdev asd;
|
|
struct v4l2_async_subdev *asds[1];
|
|
|
|
/*
|
|
* PXA27x is only supposed to handle one camera on its Quick Capture
|
|
* interface. If anyone ever builds hardware to enable more than
|
|
* one camera, they will have to modify this driver too
|
|
*/
|
|
struct clk *clk;
|
|
|
|
unsigned int irq;
|
|
void __iomem *base;
|
|
|
|
int channels;
|
|
struct dma_chan *dma_chans[3];
|
|
|
|
struct pxacamera_platform_data *pdata;
|
|
struct resource *res;
|
|
unsigned long platform_flags;
|
|
unsigned long ciclk;
|
|
unsigned long mclk;
|
|
u32 mclk_divisor;
|
|
struct v4l2_clk *mclk_clk;
|
|
u16 width_flags; /* max 10 bits */
|
|
|
|
struct list_head capture;
|
|
|
|
spinlock_t lock;
|
|
struct mutex mlock;
|
|
unsigned int buf_sequence;
|
|
|
|
struct pxa_buffer *active;
|
|
struct tasklet_struct task_eof;
|
|
|
|
u32 save_cicr[5];
|
|
};
|
|
|
|
struct pxa_cam {
|
|
unsigned long flags;
|
|
};
|
|
|
|
static const char *pxa_cam_driver_description = "PXA_Camera";
|
|
|
|
/*
|
|
* Format translation functions
|
|
*/
|
|
static const struct soc_camera_format_xlate
|
|
*pxa_mbus_xlate_by_fourcc(struct soc_camera_format_xlate *user_formats,
|
|
unsigned int fourcc)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; user_formats[i].code; i++)
|
|
if (user_formats[i].host_fmt->fourcc == fourcc)
|
|
return user_formats + i;
|
|
return NULL;
|
|
}
|
|
|
|
static struct soc_camera_format_xlate *pxa_mbus_build_fmts_xlate(
|
|
struct v4l2_device *v4l2_dev, struct v4l2_subdev *subdev,
|
|
int (*get_formats)(struct v4l2_device *, unsigned int,
|
|
struct soc_camera_format_xlate *xlate))
|
|
{
|
|
unsigned int i, fmts = 0, raw_fmts = 0;
|
|
int ret;
|
|
struct v4l2_subdev_mbus_code_enum code = {
|
|
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
|
|
};
|
|
struct soc_camera_format_xlate *user_formats;
|
|
|
|
while (!v4l2_subdev_call(subdev, pad, enum_mbus_code, NULL, &code)) {
|
|
raw_fmts++;
|
|
code.index++;
|
|
}
|
|
|
|
/*
|
|
* First pass - only count formats this host-sensor
|
|
* configuration can provide
|
|
*/
|
|
for (i = 0; i < raw_fmts; i++) {
|
|
ret = get_formats(v4l2_dev, i, NULL);
|
|
if (ret < 0)
|
|
return ERR_PTR(ret);
|
|
fmts += ret;
|
|
}
|
|
|
|
if (!fmts)
|
|
return ERR_PTR(-ENXIO);
|
|
|
|
user_formats = kcalloc(fmts + 1, sizeof(*user_formats), GFP_KERNEL);
|
|
if (!user_formats)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
/* Second pass - actually fill data formats */
|
|
fmts = 0;
|
|
for (i = 0; i < raw_fmts; i++) {
|
|
ret = get_formats(v4l2_dev, i, user_formats + fmts);
|
|
if (ret < 0)
|
|
goto egfmt;
|
|
fmts += ret;
|
|
}
|
|
user_formats[fmts].code = 0;
|
|
|
|
return user_formats;
|
|
egfmt:
|
|
kfree(user_formats);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
/*
|
|
* Videobuf operations
|
|
*/
|
|
static struct pxa_buffer *vb2_to_pxa_buffer(struct vb2_buffer *vb)
|
|
{
|
|
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
|
|
|
return container_of(vbuf, struct pxa_buffer, vbuf);
|
|
}
|
|
|
|
static struct device *pcdev_to_dev(struct pxa_camera_dev *pcdev)
|
|
{
|
|
return pcdev->v4l2_dev.dev;
|
|
}
|
|
|
|
static struct pxa_camera_dev *v4l2_dev_to_pcdev(struct v4l2_device *v4l2_dev)
|
|
{
|
|
return container_of(v4l2_dev, struct pxa_camera_dev, v4l2_dev);
|
|
}
|
|
|
|
static void pxa_camera_dma_irq(struct pxa_camera_dev *pcdev,
|
|
enum pxa_camera_active_dma act_dma);
|
|
|
|
static void pxa_camera_dma_irq_y(void *data)
|
|
{
|
|
struct pxa_camera_dev *pcdev = data;
|
|
|
|
pxa_camera_dma_irq(pcdev, DMA_Y);
|
|
}
|
|
|
|
static void pxa_camera_dma_irq_u(void *data)
|
|
{
|
|
struct pxa_camera_dev *pcdev = data;
|
|
|
|
pxa_camera_dma_irq(pcdev, DMA_U);
|
|
}
|
|
|
|
static void pxa_camera_dma_irq_v(void *data)
|
|
{
|
|
struct pxa_camera_dev *pcdev = data;
|
|
|
|
pxa_camera_dma_irq(pcdev, DMA_V);
|
|
}
|
|
|
|
/**
|
|
* pxa_init_dma_channel - init dma descriptors
|
|
* @pcdev: pxa camera device
|
|
* @buf: pxa camera buffer
|
|
* @channel: dma channel (0 => 'Y', 1 => 'U', 2 => 'V')
|
|
* @sg: dma scatter list
|
|
* @sglen: dma scatter list length
|
|
*
|
|
* Prepares the pxa dma descriptors to transfer one camera channel.
|
|
*
|
|
* Returns 0 if success or -ENOMEM if no memory is available
|
|
*/
|
|
static int pxa_init_dma_channel(struct pxa_camera_dev *pcdev,
|
|
struct pxa_buffer *buf, int channel,
|
|
struct scatterlist *sg, int sglen)
|
|
{
|
|
struct dma_chan *dma_chan = pcdev->dma_chans[channel];
|
|
struct dma_async_tx_descriptor *tx;
|
|
|
|
tx = dmaengine_prep_slave_sg(dma_chan, sg, sglen, DMA_DEV_TO_MEM,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_REUSE);
|
|
if (!tx) {
|
|
dev_err(pcdev_to_dev(pcdev),
|
|
"dmaengine_prep_slave_sg failed\n");
|
|
goto fail;
|
|
}
|
|
|
|
tx->callback_param = pcdev;
|
|
switch (channel) {
|
|
case 0:
|
|
tx->callback = pxa_camera_dma_irq_y;
|
|
break;
|
|
case 1:
|
|
tx->callback = pxa_camera_dma_irq_u;
|
|
break;
|
|
case 2:
|
|
tx->callback = pxa_camera_dma_irq_v;
|
|
break;
|
|
}
|
|
|
|
buf->descs[channel] = tx;
|
|
return 0;
|
|
fail:
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s (vb=%p) dma_tx=%p\n",
|
|
__func__, buf, tx);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void pxa_videobuf_set_actdma(struct pxa_camera_dev *pcdev,
|
|
struct pxa_buffer *buf)
|
|
{
|
|
buf->active_dma = DMA_Y;
|
|
if (buf->nb_planes == 3)
|
|
buf->active_dma |= DMA_U | DMA_V;
|
|
}
|
|
|
|
/**
|
|
* pxa_dma_start_channels - start DMA channel for active buffer
|
|
* @pcdev: pxa camera device
|
|
*
|
|
* Initialize DMA channels to the beginning of the active video buffer, and
|
|
* start these channels.
|
|
*/
|
|
static void pxa_dma_start_channels(struct pxa_camera_dev *pcdev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < pcdev->channels; i++) {
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s (channel=%d)\n", __func__, i);
|
|
dma_async_issue_pending(pcdev->dma_chans[i]);
|
|
}
|
|
}
|
|
|
|
static void pxa_dma_stop_channels(struct pxa_camera_dev *pcdev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < pcdev->channels; i++) {
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s (channel=%d)\n", __func__, i);
|
|
dmaengine_terminate_all(pcdev->dma_chans[i]);
|
|
}
|
|
}
|
|
|
|
static void pxa_dma_add_tail_buf(struct pxa_camera_dev *pcdev,
|
|
struct pxa_buffer *buf)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < pcdev->channels; i++) {
|
|
buf->cookie[i] = dmaengine_submit(buf->descs[i]);
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s (channel=%d) : submit vb=%p cookie=%d\n",
|
|
__func__, i, buf, buf->descs[i]->cookie);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* pxa_camera_start_capture - start video capturing
|
|
* @pcdev: camera device
|
|
*
|
|
* Launch capturing. DMA channels should not be active yet. They should get
|
|
* activated at the end of frame interrupt, to capture only whole frames, and
|
|
* never begin the capture of a partial frame.
|
|
*/
|
|
static void pxa_camera_start_capture(struct pxa_camera_dev *pcdev)
|
|
{
|
|
unsigned long cicr0;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev), "%s\n", __func__);
|
|
__raw_writel(__raw_readl(pcdev->base + CISR), pcdev->base + CISR);
|
|
/* Enable End-Of-Frame Interrupt */
|
|
cicr0 = __raw_readl(pcdev->base + CICR0) | CICR0_ENB;
|
|
cicr0 &= ~CICR0_EOFM;
|
|
__raw_writel(cicr0, pcdev->base + CICR0);
|
|
}
|
|
|
|
static void pxa_camera_stop_capture(struct pxa_camera_dev *pcdev)
|
|
{
|
|
unsigned long cicr0;
|
|
|
|
pxa_dma_stop_channels(pcdev);
|
|
|
|
cicr0 = __raw_readl(pcdev->base + CICR0) & ~CICR0_ENB;
|
|
__raw_writel(cicr0, pcdev->base + CICR0);
|
|
|
|
pcdev->active = NULL;
|
|
dev_dbg(pcdev_to_dev(pcdev), "%s\n", __func__);
|
|
}
|
|
|
|
static void pxa_camera_wakeup(struct pxa_camera_dev *pcdev,
|
|
struct pxa_buffer *buf,
|
|
enum vb2_buffer_state state)
|
|
{
|
|
struct vb2_buffer *vb = &buf->vbuf.vb2_buf;
|
|
struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
|
|
|
|
/* _init is used to debug races, see comment in pxa_camera_reqbufs() */
|
|
list_del_init(&buf->queue);
|
|
vb->timestamp = ktime_get_ns();
|
|
vbuf->sequence = pcdev->buf_sequence++;
|
|
vbuf->field = V4L2_FIELD_NONE;
|
|
vb2_buffer_done(vb, VB2_BUF_STATE_DONE);
|
|
dev_dbg(pcdev_to_dev(pcdev), "%s dequeued buffer (buf=0x%p)\n",
|
|
__func__, buf);
|
|
|
|
if (list_empty(&pcdev->capture)) {
|
|
pxa_camera_stop_capture(pcdev);
|
|
return;
|
|
}
|
|
|
|
pcdev->active = list_entry(pcdev->capture.next,
|
|
struct pxa_buffer, queue);
|
|
}
|
|
|
|
/**
|
|
* pxa_camera_check_link_miss - check missed DMA linking
|
|
* @pcdev: camera device
|
|
* @last_submitted: an opaque DMA cookie for last submitted
|
|
* @last_issued: an opaque DMA cookie for last issued
|
|
*
|
|
* The DMA chaining is done with DMA running. This means a tiny temporal window
|
|
* remains, where a buffer is queued on the chain, while the chain is already
|
|
* stopped. This means the tailed buffer would never be transferred by DMA.
|
|
* This function restarts the capture for this corner case, where :
|
|
* - DADR() == DADDR_STOP
|
|
* - a videobuffer is queued on the pcdev->capture list
|
|
*
|
|
* Please check the "DMA hot chaining timeslice issue" in
|
|
* Documentation/video4linux/pxa_camera.txt
|
|
*
|
|
* Context: should only be called within the dma irq handler
|
|
*/
|
|
static void pxa_camera_check_link_miss(struct pxa_camera_dev *pcdev,
|
|
dma_cookie_t last_submitted,
|
|
dma_cookie_t last_issued)
|
|
{
|
|
bool is_dma_stopped = last_submitted != last_issued;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s : top queued buffer=%p, is_dma_stopped=%d\n",
|
|
__func__, pcdev->active, is_dma_stopped);
|
|
|
|
if (pcdev->active && is_dma_stopped)
|
|
pxa_camera_start_capture(pcdev);
|
|
}
|
|
|
|
static void pxa_camera_dma_irq(struct pxa_camera_dev *pcdev,
|
|
enum pxa_camera_active_dma act_dma)
|
|
{
|
|
struct pxa_buffer *buf, *last_buf;
|
|
unsigned long flags;
|
|
u32 camera_status, overrun;
|
|
int chan;
|
|
enum dma_status last_status;
|
|
dma_cookie_t last_issued;
|
|
|
|
spin_lock_irqsave(&pcdev->lock, flags);
|
|
|
|
camera_status = __raw_readl(pcdev->base + CISR);
|
|
dev_dbg(pcdev_to_dev(pcdev), "camera dma irq, cisr=0x%x dma=%d\n",
|
|
camera_status, act_dma);
|
|
overrun = CISR_IFO_0;
|
|
if (pcdev->channels == 3)
|
|
overrun |= CISR_IFO_1 | CISR_IFO_2;
|
|
|
|
/*
|
|
* pcdev->active should not be NULL in DMA irq handler.
|
|
*
|
|
* But there is one corner case : if capture was stopped due to an
|
|
* overrun of channel 1, and at that same channel 2 was completed.
|
|
*
|
|
* When handling the overrun in DMA irq for channel 1, we'll stop the
|
|
* capture and restart it (and thus set pcdev->active to NULL). But the
|
|
* DMA irq handler will already be pending for channel 2. So on entering
|
|
* the DMA irq handler for channel 2 there will be no active buffer, yet
|
|
* that is normal.
|
|
*/
|
|
if (!pcdev->active)
|
|
goto out;
|
|
|
|
buf = pcdev->active;
|
|
WARN_ON(buf->inwork || list_empty(&buf->queue));
|
|
|
|
/*
|
|
* It's normal if the last frame creates an overrun, as there
|
|
* are no more DMA descriptors to fetch from QCI fifos
|
|
*/
|
|
switch (act_dma) {
|
|
case DMA_U:
|
|
chan = 1;
|
|
break;
|
|
case DMA_V:
|
|
chan = 2;
|
|
break;
|
|
default:
|
|
chan = 0;
|
|
break;
|
|
}
|
|
last_buf = list_entry(pcdev->capture.prev,
|
|
struct pxa_buffer, queue);
|
|
last_status = dma_async_is_tx_complete(pcdev->dma_chans[chan],
|
|
last_buf->cookie[chan],
|
|
NULL, &last_issued);
|
|
if (camera_status & overrun &&
|
|
last_status != DMA_COMPLETE) {
|
|
dev_dbg(pcdev_to_dev(pcdev), "FIFO overrun! CISR: %x\n",
|
|
camera_status);
|
|
pxa_camera_stop_capture(pcdev);
|
|
list_for_each_entry(buf, &pcdev->capture, queue)
|
|
pxa_dma_add_tail_buf(pcdev, buf);
|
|
pxa_camera_start_capture(pcdev);
|
|
goto out;
|
|
}
|
|
buf->active_dma &= ~act_dma;
|
|
if (!buf->active_dma) {
|
|
pxa_camera_wakeup(pcdev, buf, VB2_BUF_STATE_DONE);
|
|
pxa_camera_check_link_miss(pcdev, last_buf->cookie[chan],
|
|
last_issued);
|
|
}
|
|
|
|
out:
|
|
spin_unlock_irqrestore(&pcdev->lock, flags);
|
|
}
|
|
|
|
static u32 mclk_get_divisor(struct platform_device *pdev,
|
|
struct pxa_camera_dev *pcdev)
|
|
{
|
|
unsigned long mclk = pcdev->mclk;
|
|
u32 div;
|
|
unsigned long lcdclk;
|
|
|
|
lcdclk = clk_get_rate(pcdev->clk);
|
|
pcdev->ciclk = lcdclk;
|
|
|
|
/* mclk <= ciclk / 4 (27.4.2) */
|
|
if (mclk > lcdclk / 4) {
|
|
mclk = lcdclk / 4;
|
|
dev_warn(&pdev->dev,
|
|
"Limiting master clock to %lu\n", mclk);
|
|
}
|
|
|
|
/* We verify mclk != 0, so if anyone breaks it, here comes their Oops */
|
|
div = (lcdclk + 2 * mclk - 1) / (2 * mclk) - 1;
|
|
|
|
/* If we're not supplying MCLK, leave it at 0 */
|
|
if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN)
|
|
pcdev->mclk = lcdclk / (2 * (div + 1));
|
|
|
|
dev_dbg(&pdev->dev, "LCD clock %luHz, target freq %luHz, divisor %u\n",
|
|
lcdclk, mclk, div);
|
|
|
|
return div;
|
|
}
|
|
|
|
static void recalculate_fifo_timeout(struct pxa_camera_dev *pcdev,
|
|
unsigned long pclk)
|
|
{
|
|
/* We want a timeout > 1 pixel time, not ">=" */
|
|
u32 ciclk_per_pixel = pcdev->ciclk / pclk + 1;
|
|
|
|
__raw_writel(ciclk_per_pixel, pcdev->base + CITOR);
|
|
}
|
|
|
|
static void pxa_camera_activate(struct pxa_camera_dev *pcdev)
|
|
{
|
|
u32 cicr4 = 0;
|
|
|
|
/* disable all interrupts */
|
|
__raw_writel(0x3ff, pcdev->base + CICR0);
|
|
|
|
if (pcdev->platform_flags & PXA_CAMERA_PCLK_EN)
|
|
cicr4 |= CICR4_PCLK_EN;
|
|
if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN)
|
|
cicr4 |= CICR4_MCLK_EN;
|
|
if (pcdev->platform_flags & PXA_CAMERA_PCP)
|
|
cicr4 |= CICR4_PCP;
|
|
if (pcdev->platform_flags & PXA_CAMERA_HSP)
|
|
cicr4 |= CICR4_HSP;
|
|
if (pcdev->platform_flags & PXA_CAMERA_VSP)
|
|
cicr4 |= CICR4_VSP;
|
|
|
|
__raw_writel(pcdev->mclk_divisor | cicr4, pcdev->base + CICR4);
|
|
|
|
if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN)
|
|
/* Initialise the timeout under the assumption pclk = mclk */
|
|
recalculate_fifo_timeout(pcdev, pcdev->mclk);
|
|
else
|
|
/* "Safe default" - 13MHz */
|
|
recalculate_fifo_timeout(pcdev, 13000000);
|
|
|
|
clk_prepare_enable(pcdev->clk);
|
|
}
|
|
|
|
static void pxa_camera_deactivate(struct pxa_camera_dev *pcdev)
|
|
{
|
|
clk_disable_unprepare(pcdev->clk);
|
|
}
|
|
|
|
static void pxa_camera_eof(unsigned long arg)
|
|
{
|
|
struct pxa_camera_dev *pcdev = (struct pxa_camera_dev *)arg;
|
|
unsigned long cifr;
|
|
struct pxa_buffer *buf;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"Camera interrupt status 0x%x\n",
|
|
__raw_readl(pcdev->base + CISR));
|
|
|
|
/* Reset the FIFOs */
|
|
cifr = __raw_readl(pcdev->base + CIFR) | CIFR_RESET_F;
|
|
__raw_writel(cifr, pcdev->base + CIFR);
|
|
|
|
pcdev->active = list_first_entry(&pcdev->capture,
|
|
struct pxa_buffer, queue);
|
|
buf = pcdev->active;
|
|
pxa_videobuf_set_actdma(pcdev, buf);
|
|
|
|
pxa_dma_start_channels(pcdev);
|
|
}
|
|
|
|
static irqreturn_t pxa_camera_irq(int irq, void *data)
|
|
{
|
|
struct pxa_camera_dev *pcdev = data;
|
|
unsigned long status, cicr0;
|
|
|
|
status = __raw_readl(pcdev->base + CISR);
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"Camera interrupt status 0x%lx\n", status);
|
|
|
|
if (!status)
|
|
return IRQ_NONE;
|
|
|
|
__raw_writel(status, pcdev->base + CISR);
|
|
|
|
if (status & CISR_EOF) {
|
|
cicr0 = __raw_readl(pcdev->base + CICR0) | CICR0_EOFM;
|
|
__raw_writel(cicr0, pcdev->base + CICR0);
|
|
tasklet_schedule(&pcdev->task_eof);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int test_platform_param(struct pxa_camera_dev *pcdev,
|
|
unsigned char buswidth, unsigned long *flags)
|
|
{
|
|
/*
|
|
* Platform specified synchronization and pixel clock polarities are
|
|
* only a recommendation and are only used during probing. The PXA270
|
|
* quick capture interface supports both.
|
|
*/
|
|
*flags = (pcdev->platform_flags & PXA_CAMERA_MASTER ?
|
|
V4L2_MBUS_MASTER : V4L2_MBUS_SLAVE) |
|
|
V4L2_MBUS_HSYNC_ACTIVE_HIGH |
|
|
V4L2_MBUS_HSYNC_ACTIVE_LOW |
|
|
V4L2_MBUS_VSYNC_ACTIVE_HIGH |
|
|
V4L2_MBUS_VSYNC_ACTIVE_LOW |
|
|
V4L2_MBUS_DATA_ACTIVE_HIGH |
|
|
V4L2_MBUS_PCLK_SAMPLE_RISING |
|
|
V4L2_MBUS_PCLK_SAMPLE_FALLING;
|
|
|
|
/* If requested data width is supported by the platform, use it */
|
|
if ((1 << (buswidth - 1)) & pcdev->width_flags)
|
|
return 0;
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static void pxa_camera_setup_cicr(struct pxa_camera_dev *pcdev,
|
|
unsigned long flags, __u32 pixfmt)
|
|
{
|
|
unsigned long dw, bpp;
|
|
u32 cicr0, cicr1, cicr2, cicr3, cicr4 = 0, y_skip_top;
|
|
int ret = sensor_call(pcdev, sensor, g_skip_top_lines, &y_skip_top);
|
|
|
|
if (ret < 0)
|
|
y_skip_top = 0;
|
|
|
|
/*
|
|
* Datawidth is now guaranteed to be equal to one of the three values.
|
|
* We fix bit-per-pixel equal to data-width...
|
|
*/
|
|
switch (pcdev->current_fmt->host_fmt->bits_per_sample) {
|
|
case 10:
|
|
dw = 4;
|
|
bpp = 0x40;
|
|
break;
|
|
case 9:
|
|
dw = 3;
|
|
bpp = 0x20;
|
|
break;
|
|
default:
|
|
/*
|
|
* Actually it can only be 8 now,
|
|
* default is just to silence compiler warnings
|
|
*/
|
|
case 8:
|
|
dw = 2;
|
|
bpp = 0;
|
|
}
|
|
|
|
if (pcdev->platform_flags & PXA_CAMERA_PCLK_EN)
|
|
cicr4 |= CICR4_PCLK_EN;
|
|
if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN)
|
|
cicr4 |= CICR4_MCLK_EN;
|
|
if (flags & V4L2_MBUS_PCLK_SAMPLE_FALLING)
|
|
cicr4 |= CICR4_PCP;
|
|
if (flags & V4L2_MBUS_HSYNC_ACTIVE_LOW)
|
|
cicr4 |= CICR4_HSP;
|
|
if (flags & V4L2_MBUS_VSYNC_ACTIVE_LOW)
|
|
cicr4 |= CICR4_VSP;
|
|
|
|
cicr0 = __raw_readl(pcdev->base + CICR0);
|
|
if (cicr0 & CICR0_ENB)
|
|
__raw_writel(cicr0 & ~CICR0_ENB, pcdev->base + CICR0);
|
|
|
|
cicr1 = CICR1_PPL_VAL(pcdev->current_pix.width - 1) | bpp | dw;
|
|
|
|
switch (pixfmt) {
|
|
case V4L2_PIX_FMT_YUV422P:
|
|
pcdev->channels = 3;
|
|
cicr1 |= CICR1_YCBCR_F;
|
|
/*
|
|
* Normally, pxa bus wants as input UYVY format. We allow all
|
|
* reorderings of the YUV422 format, as no processing is done,
|
|
* and the YUV stream is just passed through without any
|
|
* transformation. Note that UYVY is the only format that
|
|
* should be used if pxa framebuffer Overlay2 is used.
|
|
*/
|
|
/* fall through */
|
|
case V4L2_PIX_FMT_UYVY:
|
|
case V4L2_PIX_FMT_VYUY:
|
|
case V4L2_PIX_FMT_YUYV:
|
|
case V4L2_PIX_FMT_YVYU:
|
|
cicr1 |= CICR1_COLOR_SP_VAL(2);
|
|
break;
|
|
case V4L2_PIX_FMT_RGB555:
|
|
cicr1 |= CICR1_RGB_BPP_VAL(1) | CICR1_RGBT_CONV_VAL(2) |
|
|
CICR1_TBIT | CICR1_COLOR_SP_VAL(1);
|
|
break;
|
|
case V4L2_PIX_FMT_RGB565:
|
|
cicr1 |= CICR1_COLOR_SP_VAL(1) | CICR1_RGB_BPP_VAL(2);
|
|
break;
|
|
}
|
|
|
|
cicr2 = 0;
|
|
cicr3 = CICR3_LPF_VAL(pcdev->current_pix.height - 1) |
|
|
CICR3_BFW_VAL(min((u32)255, y_skip_top));
|
|
cicr4 |= pcdev->mclk_divisor;
|
|
|
|
__raw_writel(cicr1, pcdev->base + CICR1);
|
|
__raw_writel(cicr2, pcdev->base + CICR2);
|
|
__raw_writel(cicr3, pcdev->base + CICR3);
|
|
__raw_writel(cicr4, pcdev->base + CICR4);
|
|
|
|
/* CIF interrupts are not used, only DMA */
|
|
cicr0 = (cicr0 & CICR0_ENB) | (pcdev->platform_flags & PXA_CAMERA_MASTER ?
|
|
CICR0_SIM_MP : (CICR0_SL_CAP_EN | CICR0_SIM_SP));
|
|
cicr0 |= CICR0_DMAEN | CICR0_IRQ_MASK;
|
|
__raw_writel(cicr0, pcdev->base + CICR0);
|
|
}
|
|
|
|
/*
|
|
* Videobuf2 section
|
|
*/
|
|
static void pxa_buffer_cleanup(struct pxa_buffer *buf)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < 3 && buf->descs[i]; i++) {
|
|
dmaengine_desc_free(buf->descs[i]);
|
|
kfree(buf->sg[i]);
|
|
buf->descs[i] = NULL;
|
|
buf->sg[i] = NULL;
|
|
buf->sg_len[i] = 0;
|
|
buf->plane_sizes[i] = 0;
|
|
}
|
|
buf->nb_planes = 0;
|
|
}
|
|
|
|
static int pxa_buffer_init(struct pxa_camera_dev *pcdev,
|
|
struct pxa_buffer *buf)
|
|
{
|
|
struct vb2_buffer *vb = &buf->vbuf.vb2_buf;
|
|
struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
|
|
int nb_channels = pcdev->channels;
|
|
int i, ret = 0;
|
|
unsigned long size = vb2_plane_size(vb, 0);
|
|
|
|
switch (nb_channels) {
|
|
case 1:
|
|
buf->plane_sizes[0] = size;
|
|
break;
|
|
case 3:
|
|
buf->plane_sizes[0] = size / 2;
|
|
buf->plane_sizes[1] = size / 4;
|
|
buf->plane_sizes[2] = size / 4;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
};
|
|
buf->nb_planes = nb_channels;
|
|
|
|
ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels,
|
|
buf->plane_sizes, buf->sg, buf->sg_len, GFP_KERNEL);
|
|
if (ret < 0) {
|
|
dev_err(pcdev_to_dev(pcdev),
|
|
"sg_split failed: %d\n", ret);
|
|
return ret;
|
|
}
|
|
for (i = 0; i < nb_channels; i++) {
|
|
ret = pxa_init_dma_channel(pcdev, buf, i,
|
|
buf->sg[i], buf->sg_len[i]);
|
|
if (ret) {
|
|
pxa_buffer_cleanup(buf);
|
|
return ret;
|
|
}
|
|
}
|
|
INIT_LIST_HEAD(&buf->queue);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void pxac_vb2_cleanup(struct vb2_buffer *vb)
|
|
{
|
|
struct pxa_buffer *buf = vb2_to_pxa_buffer(vb);
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue);
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s(vb=%p)\n", __func__, vb);
|
|
pxa_buffer_cleanup(buf);
|
|
}
|
|
|
|
static void pxac_vb2_queue(struct vb2_buffer *vb)
|
|
{
|
|
struct pxa_buffer *buf = vb2_to_pxa_buffer(vb);
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue);
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s(vb=%p) nb_channels=%d size=%lu active=%p\n",
|
|
__func__, vb, pcdev->channels, vb2_get_plane_payload(vb, 0),
|
|
pcdev->active);
|
|
|
|
list_add_tail(&buf->queue, &pcdev->capture);
|
|
|
|
pxa_dma_add_tail_buf(pcdev, buf);
|
|
}
|
|
|
|
/*
|
|
* Please check the DMA prepared buffer structure in :
|
|
* Documentation/video4linux/pxa_camera.txt
|
|
* Please check also in pxa_camera_check_link_miss() to understand why DMA chain
|
|
* modification while DMA chain is running will work anyway.
|
|
*/
|
|
static int pxac_vb2_prepare(struct vb2_buffer *vb)
|
|
{
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue);
|
|
struct pxa_buffer *buf = vb2_to_pxa_buffer(vb);
|
|
int ret = 0;
|
|
|
|
switch (pcdev->channels) {
|
|
case 1:
|
|
case 3:
|
|
vb2_set_plane_payload(vb, 0, pcdev->current_pix.sizeimage);
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s (vb=%p) nb_channels=%d size=%lu\n",
|
|
__func__, vb, pcdev->channels, vb2_get_plane_payload(vb, 0));
|
|
|
|
WARN_ON(!pcdev->current_fmt);
|
|
|
|
#ifdef DEBUG
|
|
/*
|
|
* This can be useful if you want to see if we actually fill
|
|
* the buffer with something
|
|
*/
|
|
for (i = 0; i < vb->num_planes; i++)
|
|
memset((void *)vb2_plane_vaddr(vb, i),
|
|
0xaa, vb2_get_plane_payload(vb, i));
|
|
#endif
|
|
|
|
/*
|
|
* I think, in buf_prepare you only have to protect global data,
|
|
* the actual buffer is yours
|
|
*/
|
|
buf->inwork = 0;
|
|
pxa_videobuf_set_actdma(pcdev, buf);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int pxac_vb2_init(struct vb2_buffer *vb)
|
|
{
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue);
|
|
struct pxa_buffer *buf = vb2_to_pxa_buffer(vb);
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s(nb_channels=%d)\n",
|
|
__func__, pcdev->channels);
|
|
|
|
return pxa_buffer_init(pcdev, buf);
|
|
}
|
|
|
|
static int pxac_vb2_queue_setup(struct vb2_queue *vq,
|
|
unsigned int *nbufs,
|
|
unsigned int *num_planes, unsigned int sizes[],
|
|
struct device *alloc_devs[])
|
|
{
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq);
|
|
int size = pcdev->current_pix.sizeimage;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"%s(vq=%p nbufs=%d num_planes=%d size=%d)\n",
|
|
__func__, vq, *nbufs, *num_planes, size);
|
|
/*
|
|
* Called from VIDIOC_REQBUFS or in compatibility mode For YUV422P
|
|
* format, even if there are 3 planes Y, U and V, we reply there is only
|
|
* one plane, containing Y, U and V data, one after the other.
|
|
*/
|
|
if (*num_planes)
|
|
return sizes[0] < size ? -EINVAL : 0;
|
|
|
|
*num_planes = 1;
|
|
switch (pcdev->channels) {
|
|
case 1:
|
|
case 3:
|
|
sizes[0] = size;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!*nbufs)
|
|
*nbufs = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vb2_start_streaming(struct vb2_queue *vq, unsigned int count)
|
|
{
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq);
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev), "%s(count=%d) active=%p\n",
|
|
__func__, count, pcdev->active);
|
|
|
|
pcdev->buf_sequence = 0;
|
|
if (!pcdev->active)
|
|
pxa_camera_start_capture(pcdev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void pxac_vb2_stop_streaming(struct vb2_queue *vq)
|
|
{
|
|
struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq);
|
|
struct pxa_buffer *buf, *tmp;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev), "%s active=%p\n",
|
|
__func__, pcdev->active);
|
|
pxa_camera_stop_capture(pcdev);
|
|
|
|
list_for_each_entry_safe(buf, tmp, &pcdev->capture, queue)
|
|
pxa_camera_wakeup(pcdev, buf, VB2_BUF_STATE_ERROR);
|
|
}
|
|
|
|
static const struct vb2_ops pxac_vb2_ops = {
|
|
.queue_setup = pxac_vb2_queue_setup,
|
|
.buf_init = pxac_vb2_init,
|
|
.buf_prepare = pxac_vb2_prepare,
|
|
.buf_queue = pxac_vb2_queue,
|
|
.buf_cleanup = pxac_vb2_cleanup,
|
|
.start_streaming = pxac_vb2_start_streaming,
|
|
.stop_streaming = pxac_vb2_stop_streaming,
|
|
.wait_prepare = vb2_ops_wait_prepare,
|
|
.wait_finish = vb2_ops_wait_finish,
|
|
};
|
|
|
|
static int pxa_camera_init_videobuf2(struct pxa_camera_dev *pcdev)
|
|
{
|
|
int ret;
|
|
struct vb2_queue *vq = &pcdev->vb2_vq;
|
|
|
|
memset(vq, 0, sizeof(*vq));
|
|
vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
vq->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
|
|
vq->drv_priv = pcdev;
|
|
vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
|
|
vq->buf_struct_size = sizeof(struct pxa_buffer);
|
|
vq->dev = pcdev->v4l2_dev.dev;
|
|
|
|
vq->ops = &pxac_vb2_ops;
|
|
vq->mem_ops = &vb2_dma_sg_memops;
|
|
vq->lock = &pcdev->mlock;
|
|
|
|
ret = vb2_queue_init(vq);
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"vb2_queue_init(vq=%p): %d\n", vq, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Video ioctls section
|
|
*/
|
|
static int pxa_camera_set_bus_param(struct pxa_camera_dev *pcdev)
|
|
{
|
|
struct v4l2_mbus_config cfg = {.type = V4L2_MBUS_PARALLEL,};
|
|
u32 pixfmt = pcdev->current_fmt->host_fmt->fourcc;
|
|
unsigned long bus_flags, common_flags;
|
|
int ret;
|
|
|
|
ret = test_platform_param(pcdev,
|
|
pcdev->current_fmt->host_fmt->bits_per_sample,
|
|
&bus_flags);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = sensor_call(pcdev, video, g_mbus_config, &cfg);
|
|
if (!ret) {
|
|
common_flags = pxa_mbus_config_compatible(&cfg,
|
|
bus_flags);
|
|
if (!common_flags) {
|
|
dev_warn(pcdev_to_dev(pcdev),
|
|
"Flags incompatible: camera 0x%x, host 0x%lx\n",
|
|
cfg.flags, bus_flags);
|
|
return -EINVAL;
|
|
}
|
|
} else if (ret != -ENOIOCTLCMD) {
|
|
return ret;
|
|
} else {
|
|
common_flags = bus_flags;
|
|
}
|
|
|
|
pcdev->channels = 1;
|
|
|
|
/* Make choises, based on platform preferences */
|
|
if ((common_flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH) &&
|
|
(common_flags & V4L2_MBUS_HSYNC_ACTIVE_LOW)) {
|
|
if (pcdev->platform_flags & PXA_CAMERA_HSP)
|
|
common_flags &= ~V4L2_MBUS_HSYNC_ACTIVE_HIGH;
|
|
else
|
|
common_flags &= ~V4L2_MBUS_HSYNC_ACTIVE_LOW;
|
|
}
|
|
|
|
if ((common_flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH) &&
|
|
(common_flags & V4L2_MBUS_VSYNC_ACTIVE_LOW)) {
|
|
if (pcdev->platform_flags & PXA_CAMERA_VSP)
|
|
common_flags &= ~V4L2_MBUS_VSYNC_ACTIVE_HIGH;
|
|
else
|
|
common_flags &= ~V4L2_MBUS_VSYNC_ACTIVE_LOW;
|
|
}
|
|
|
|
if ((common_flags & V4L2_MBUS_PCLK_SAMPLE_RISING) &&
|
|
(common_flags & V4L2_MBUS_PCLK_SAMPLE_FALLING)) {
|
|
if (pcdev->platform_flags & PXA_CAMERA_PCP)
|
|
common_flags &= ~V4L2_MBUS_PCLK_SAMPLE_RISING;
|
|
else
|
|
common_flags &= ~V4L2_MBUS_PCLK_SAMPLE_FALLING;
|
|
}
|
|
|
|
cfg.flags = common_flags;
|
|
ret = sensor_call(pcdev, video, s_mbus_config, &cfg);
|
|
if (ret < 0 && ret != -ENOIOCTLCMD) {
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"camera s_mbus_config(0x%lx) returned %d\n",
|
|
common_flags, ret);
|
|
return ret;
|
|
}
|
|
|
|
pxa_camera_setup_cicr(pcdev, common_flags, pixfmt);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxa_camera_try_bus_param(struct pxa_camera_dev *pcdev,
|
|
unsigned char buswidth)
|
|
{
|
|
struct v4l2_mbus_config cfg = {.type = V4L2_MBUS_PARALLEL,};
|
|
unsigned long bus_flags, common_flags;
|
|
int ret = test_platform_param(pcdev, buswidth, &bus_flags);
|
|
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = sensor_call(pcdev, video, g_mbus_config, &cfg);
|
|
if (!ret) {
|
|
common_flags = pxa_mbus_config_compatible(&cfg,
|
|
bus_flags);
|
|
if (!common_flags) {
|
|
dev_warn(pcdev_to_dev(pcdev),
|
|
"Flags incompatible: camera 0x%x, host 0x%lx\n",
|
|
cfg.flags, bus_flags);
|
|
return -EINVAL;
|
|
}
|
|
} else if (ret == -ENOIOCTLCMD) {
|
|
ret = 0;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct pxa_mbus_pixelfmt pxa_camera_formats[] = {
|
|
{
|
|
.fourcc = V4L2_PIX_FMT_YUV422P,
|
|
.name = "Planar YUV422 16 bit",
|
|
.bits_per_sample = 8,
|
|
.packing = PXA_MBUS_PACKING_2X8_PADHI,
|
|
.order = PXA_MBUS_ORDER_LE,
|
|
.layout = PXA_MBUS_LAYOUT_PLANAR_2Y_U_V,
|
|
},
|
|
};
|
|
|
|
/* This will be corrected as we get more formats */
|
|
static bool pxa_camera_packing_supported(const struct pxa_mbus_pixelfmt *fmt)
|
|
{
|
|
return fmt->packing == PXA_MBUS_PACKING_NONE ||
|
|
(fmt->bits_per_sample == 8 &&
|
|
fmt->packing == PXA_MBUS_PACKING_2X8_PADHI) ||
|
|
(fmt->bits_per_sample > 8 &&
|
|
fmt->packing == PXA_MBUS_PACKING_EXTEND16);
|
|
}
|
|
|
|
static int pxa_camera_get_formats(struct v4l2_device *v4l2_dev,
|
|
unsigned int idx,
|
|
struct soc_camera_format_xlate *xlate)
|
|
{
|
|
struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(v4l2_dev);
|
|
int formats = 0, ret;
|
|
struct v4l2_subdev_mbus_code_enum code = {
|
|
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
|
|
.index = idx,
|
|
};
|
|
const struct pxa_mbus_pixelfmt *fmt;
|
|
|
|
ret = sensor_call(pcdev, pad, enum_mbus_code, NULL, &code);
|
|
if (ret < 0)
|
|
/* No more formats */
|
|
return 0;
|
|
|
|
fmt = pxa_mbus_get_fmtdesc(code.code);
|
|
if (!fmt) {
|
|
dev_err(pcdev_to_dev(pcdev),
|
|
"Invalid format code #%u: %d\n", idx, code.code);
|
|
return 0;
|
|
}
|
|
|
|
/* This also checks support for the requested bits-per-sample */
|
|
ret = pxa_camera_try_bus_param(pcdev, fmt->bits_per_sample);
|
|
if (ret < 0)
|
|
return 0;
|
|
|
|
switch (code.code) {
|
|
case MEDIA_BUS_FMT_UYVY8_2X8:
|
|
formats++;
|
|
if (xlate) {
|
|
xlate->host_fmt = &pxa_camera_formats[0];
|
|
xlate->code = code.code;
|
|
xlate++;
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"Providing format %s using code %d\n",
|
|
pxa_camera_formats[0].name, code.code);
|
|
}
|
|
/* fall through */
|
|
case MEDIA_BUS_FMT_VYUY8_2X8:
|
|
case MEDIA_BUS_FMT_YUYV8_2X8:
|
|
case MEDIA_BUS_FMT_YVYU8_2X8:
|
|
case MEDIA_BUS_FMT_RGB565_2X8_LE:
|
|
case MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE:
|
|
if (xlate)
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"Providing format %s packed\n",
|
|
fmt->name);
|
|
break;
|
|
default:
|
|
if (!pxa_camera_packing_supported(fmt))
|
|
return 0;
|
|
if (xlate)
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"Providing format %s in pass-through mode\n",
|
|
fmt->name);
|
|
break;
|
|
}
|
|
|
|
/* Generic pass-through */
|
|
formats++;
|
|
if (xlate) {
|
|
xlate->host_fmt = fmt;
|
|
xlate->code = code.code;
|
|
xlate++;
|
|
}
|
|
|
|
return formats;
|
|
}
|
|
|
|
static int pxa_camera_build_formats(struct pxa_camera_dev *pcdev)
|
|
{
|
|
struct soc_camera_format_xlate *xlate;
|
|
|
|
xlate = pxa_mbus_build_fmts_xlate(&pcdev->v4l2_dev, pcdev->sensor,
|
|
pxa_camera_get_formats);
|
|
if (IS_ERR(xlate))
|
|
return PTR_ERR(xlate);
|
|
|
|
pcdev->user_formats = xlate;
|
|
return 0;
|
|
}
|
|
|
|
static void pxa_camera_destroy_formats(struct pxa_camera_dev *pcdev)
|
|
{
|
|
kfree(pcdev->user_formats);
|
|
}
|
|
|
|
static int pxa_camera_check_frame(u32 width, u32 height)
|
|
{
|
|
/* limit to pxa hardware capabilities */
|
|
return height < 32 || height > 2048 || width < 48 || width > 2048 ||
|
|
(width & 0x01);
|
|
}
|
|
|
|
#ifdef CONFIG_VIDEO_ADV_DEBUG
|
|
static int pxac_vidioc_g_register(struct file *file, void *priv,
|
|
struct v4l2_dbg_register *reg)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(file);
|
|
|
|
if (reg->reg > CIBR2)
|
|
return -ERANGE;
|
|
|
|
reg->val = __raw_readl(pcdev->base + reg->reg);
|
|
reg->size = sizeof(__u32);
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_s_register(struct file *file, void *priv,
|
|
const struct v4l2_dbg_register *reg)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(file);
|
|
|
|
if (reg->reg > CIBR2)
|
|
return -ERANGE;
|
|
if (reg->size != sizeof(__u32))
|
|
return -EINVAL;
|
|
__raw_writel(reg->val, pcdev->base + reg->reg);
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int pxac_vidioc_enum_fmt_vid_cap(struct file *filp, void *priv,
|
|
struct v4l2_fmtdesc *f)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
const struct pxa_mbus_pixelfmt *format;
|
|
unsigned int idx;
|
|
|
|
for (idx = 0; pcdev->user_formats[idx].code; idx++);
|
|
if (f->index >= idx)
|
|
return -EINVAL;
|
|
|
|
format = pcdev->user_formats[f->index].host_fmt;
|
|
f->pixelformat = format->fourcc;
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_g_fmt_vid_cap(struct file *filp, void *priv,
|
|
struct v4l2_format *f)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
struct v4l2_pix_format *pix = &f->fmt.pix;
|
|
|
|
pix->width = pcdev->current_pix.width;
|
|
pix->height = pcdev->current_pix.height;
|
|
pix->bytesperline = pcdev->current_pix.bytesperline;
|
|
pix->sizeimage = pcdev->current_pix.sizeimage;
|
|
pix->field = pcdev->current_pix.field;
|
|
pix->pixelformat = pcdev->current_fmt->host_fmt->fourcc;
|
|
pix->colorspace = pcdev->current_pix.colorspace;
|
|
dev_dbg(pcdev_to_dev(pcdev), "current_fmt->fourcc: 0x%08x\n",
|
|
pcdev->current_fmt->host_fmt->fourcc);
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_try_fmt_vid_cap(struct file *filp, void *priv,
|
|
struct v4l2_format *f)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
const struct soc_camera_format_xlate *xlate;
|
|
struct v4l2_pix_format *pix = &f->fmt.pix;
|
|
struct v4l2_subdev_pad_config pad_cfg;
|
|
struct v4l2_subdev_format format = {
|
|
.which = V4L2_SUBDEV_FORMAT_TRY,
|
|
};
|
|
struct v4l2_mbus_framefmt *mf = &format.format;
|
|
__u32 pixfmt = pix->pixelformat;
|
|
int ret;
|
|
|
|
xlate = pxa_mbus_xlate_by_fourcc(pcdev->user_formats, pixfmt);
|
|
if (!xlate) {
|
|
dev_warn(pcdev_to_dev(pcdev), "Format %x not found\n", pixfmt);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* Limit to pxa hardware capabilities. YUV422P planar format requires
|
|
* images size to be a multiple of 16 bytes. If not, zeros will be
|
|
* inserted between Y and U planes, and U and V planes, which violates
|
|
* the YUV422P standard.
|
|
*/
|
|
v4l_bound_align_image(&pix->width, 48, 2048, 1,
|
|
&pix->height, 32, 2048, 0,
|
|
pixfmt == V4L2_PIX_FMT_YUV422P ? 4 : 0);
|
|
|
|
v4l2_fill_mbus_format(mf, pix, xlate->code);
|
|
ret = sensor_call(pcdev, pad, set_fmt, &pad_cfg, &format);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
v4l2_fill_pix_format(pix, mf);
|
|
|
|
/* Only progressive video supported so far */
|
|
switch (mf->field) {
|
|
case V4L2_FIELD_ANY:
|
|
case V4L2_FIELD_NONE:
|
|
pix->field = V4L2_FIELD_NONE;
|
|
break;
|
|
default:
|
|
/* TODO: support interlaced at least in pass-through mode */
|
|
dev_err(pcdev_to_dev(pcdev), "Field type %d unsupported.\n",
|
|
mf->field);
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = pxa_mbus_bytes_per_line(pix->width, xlate->host_fmt);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
pix->bytesperline = ret;
|
|
ret = pxa_mbus_image_size(xlate->host_fmt, pix->bytesperline,
|
|
pix->height);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
pix->sizeimage = ret;
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_s_fmt_vid_cap(struct file *filp, void *priv,
|
|
struct v4l2_format *f)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
const struct soc_camera_format_xlate *xlate;
|
|
struct v4l2_pix_format *pix = &f->fmt.pix;
|
|
struct v4l2_subdev_format format = {
|
|
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
|
|
};
|
|
unsigned long flags;
|
|
int ret, is_busy;
|
|
|
|
dev_dbg(pcdev_to_dev(pcdev),
|
|
"s_fmt_vid_cap(pix=%dx%d:%x)\n",
|
|
pix->width, pix->height, pix->pixelformat);
|
|
|
|
spin_lock_irqsave(&pcdev->lock, flags);
|
|
is_busy = pcdev->active || vb2_is_busy(&pcdev->vb2_vq);
|
|
spin_unlock_irqrestore(&pcdev->lock, flags);
|
|
|
|
if (is_busy)
|
|
return -EBUSY;
|
|
|
|
ret = pxac_vidioc_try_fmt_vid_cap(filp, priv, f);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xlate = pxa_mbus_xlate_by_fourcc(pcdev->user_formats,
|
|
pix->pixelformat);
|
|
v4l2_fill_mbus_format(&format.format, pix, xlate->code);
|
|
ret = sensor_call(pcdev, pad, set_fmt, NULL, &format);
|
|
if (ret < 0) {
|
|
dev_warn(pcdev_to_dev(pcdev),
|
|
"Failed to configure for format %x\n",
|
|
pix->pixelformat);
|
|
} else if (pxa_camera_check_frame(pix->width, pix->height)) {
|
|
dev_warn(pcdev_to_dev(pcdev),
|
|
"Camera driver produced an unsupported frame %dx%d\n",
|
|
pix->width, pix->height);
|
|
return -EINVAL;
|
|
}
|
|
|
|
pcdev->current_fmt = xlate;
|
|
pcdev->current_pix = *pix;
|
|
|
|
ret = pxa_camera_set_bus_param(pcdev);
|
|
return ret;
|
|
}
|
|
|
|
static int pxac_vidioc_querycap(struct file *file, void *priv,
|
|
struct v4l2_capability *cap)
|
|
{
|
|
strlcpy(cap->bus_info, "platform:pxa-camera", sizeof(cap->bus_info));
|
|
strlcpy(cap->driver, PXA_CAM_DRV_NAME, sizeof(cap->driver));
|
|
strlcpy(cap->card, pxa_cam_driver_description, sizeof(cap->card));
|
|
cap->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING;
|
|
cap->capabilities = cap->device_caps | V4L2_CAP_DEVICE_CAPS;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_enum_input(struct file *file, void *priv,
|
|
struct v4l2_input *i)
|
|
{
|
|
if (i->index > 0)
|
|
return -EINVAL;
|
|
|
|
i->type = V4L2_INPUT_TYPE_CAMERA;
|
|
strlcpy(i->name, "Camera", sizeof(i->name));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_g_input(struct file *file, void *priv, unsigned int *i)
|
|
{
|
|
*i = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_vidioc_s_input(struct file *file, void *priv, unsigned int i)
|
|
{
|
|
if (i > 0)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pxac_fops_camera_open(struct file *filp)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
int ret;
|
|
|
|
mutex_lock(&pcdev->mlock);
|
|
ret = v4l2_fh_open(filp);
|
|
if (ret < 0)
|
|
goto out;
|
|
|
|
ret = sensor_call(pcdev, core, s_power, 1);
|
|
if (ret)
|
|
v4l2_fh_release(filp);
|
|
out:
|
|
mutex_unlock(&pcdev->mlock);
|
|
return ret;
|
|
}
|
|
|
|
static int pxac_fops_camera_release(struct file *filp)
|
|
{
|
|
struct pxa_camera_dev *pcdev = video_drvdata(filp);
|
|
int ret;
|
|
|
|
ret = vb2_fop_release(filp);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
mutex_lock(&pcdev->mlock);
|
|
ret = sensor_call(pcdev, core, s_power, 0);
|
|
mutex_unlock(&pcdev->mlock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct v4l2_file_operations pxa_camera_fops = {
|
|
.owner = THIS_MODULE,
|
|
.open = pxac_fops_camera_open,
|
|
.release = pxac_fops_camera_release,
|
|
.read = vb2_fop_read,
|
|
.poll = vb2_fop_poll,
|
|
.mmap = vb2_fop_mmap,
|
|
.unlocked_ioctl = video_ioctl2,
|
|
};
|
|
|
|
static const struct v4l2_ioctl_ops pxa_camera_ioctl_ops = {
|
|
.vidioc_querycap = pxac_vidioc_querycap,
|
|
|
|
.vidioc_enum_input = pxac_vidioc_enum_input,
|
|
.vidioc_g_input = pxac_vidioc_g_input,
|
|
.vidioc_s_input = pxac_vidioc_s_input,
|
|
|
|
.vidioc_enum_fmt_vid_cap = pxac_vidioc_enum_fmt_vid_cap,
|
|
.vidioc_g_fmt_vid_cap = pxac_vidioc_g_fmt_vid_cap,
|
|
.vidioc_s_fmt_vid_cap = pxac_vidioc_s_fmt_vid_cap,
|
|
.vidioc_try_fmt_vid_cap = pxac_vidioc_try_fmt_vid_cap,
|
|
|
|
.vidioc_reqbufs = vb2_ioctl_reqbufs,
|
|
.vidioc_create_bufs = vb2_ioctl_create_bufs,
|
|
.vidioc_querybuf = vb2_ioctl_querybuf,
|
|
.vidioc_qbuf = vb2_ioctl_qbuf,
|
|
.vidioc_dqbuf = vb2_ioctl_dqbuf,
|
|
.vidioc_expbuf = vb2_ioctl_expbuf,
|
|
.vidioc_streamon = vb2_ioctl_streamon,
|
|
.vidioc_streamoff = vb2_ioctl_streamoff,
|
|
#ifdef CONFIG_VIDEO_ADV_DEBUG
|
|
.vidioc_g_register = pxac_vidioc_g_register,
|
|
.vidioc_s_register = pxac_vidioc_s_register,
|
|
#endif
|
|
.vidioc_subscribe_event = v4l2_ctrl_subscribe_event,
|
|
.vidioc_unsubscribe_event = v4l2_event_unsubscribe,
|
|
};
|
|
|
|
static const struct v4l2_clk_ops pxa_camera_mclk_ops = {
|
|
};
|
|
|
|
static const struct video_device pxa_camera_videodev_template = {
|
|
.name = "pxa-camera",
|
|
.minor = -1,
|
|
.fops = &pxa_camera_fops,
|
|
.ioctl_ops = &pxa_camera_ioctl_ops,
|
|
.release = video_device_release_empty,
|
|
.device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING,
|
|
};
|
|
|
|
static int pxa_camera_sensor_bound(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
int err;
|
|
struct v4l2_device *v4l2_dev = notifier->v4l2_dev;
|
|
struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(v4l2_dev);
|
|
struct video_device *vdev = &pcdev->vdev;
|
|
struct v4l2_pix_format *pix = &pcdev->current_pix;
|
|
struct v4l2_subdev_format format = {
|
|
.which = V4L2_SUBDEV_FORMAT_ACTIVE,
|
|
};
|
|
struct v4l2_mbus_framefmt *mf = &format.format;
|
|
|
|
dev_info(pcdev_to_dev(pcdev), "%s(): trying to bind a device\n",
|
|
__func__);
|
|
mutex_lock(&pcdev->mlock);
|
|
*vdev = pxa_camera_videodev_template;
|
|
vdev->v4l2_dev = v4l2_dev;
|
|
vdev->lock = &pcdev->mlock;
|
|
pcdev->sensor = subdev;
|
|
pcdev->vdev.queue = &pcdev->vb2_vq;
|
|
pcdev->vdev.v4l2_dev = &pcdev->v4l2_dev;
|
|
pcdev->vdev.ctrl_handler = subdev->ctrl_handler;
|
|
video_set_drvdata(&pcdev->vdev, pcdev);
|
|
|
|
err = pxa_camera_build_formats(pcdev);
|
|
if (err) {
|
|
dev_err(pcdev_to_dev(pcdev), "building formats failed: %d\n",
|
|
err);
|
|
goto out;
|
|
}
|
|
|
|
pcdev->current_fmt = pcdev->user_formats;
|
|
pix->field = V4L2_FIELD_NONE;
|
|
pix->width = DEFAULT_WIDTH;
|
|
pix->height = DEFAULT_HEIGHT;
|
|
pix->bytesperline =
|
|
pxa_mbus_bytes_per_line(pix->width,
|
|
pcdev->current_fmt->host_fmt);
|
|
pix->sizeimage =
|
|
pxa_mbus_image_size(pcdev->current_fmt->host_fmt,
|
|
pix->bytesperline, pix->height);
|
|
pix->pixelformat = pcdev->current_fmt->host_fmt->fourcc;
|
|
v4l2_fill_mbus_format(mf, pix, pcdev->current_fmt->code);
|
|
|
|
err = sensor_call(pcdev, core, s_power, 1);
|
|
if (err)
|
|
goto out;
|
|
|
|
err = sensor_call(pcdev, pad, set_fmt, NULL, &format);
|
|
if (err)
|
|
goto out_sensor_poweroff;
|
|
|
|
v4l2_fill_pix_format(pix, mf);
|
|
pr_info("%s(): colorspace=0x%x pixfmt=0x%x\n",
|
|
__func__, pix->colorspace, pix->pixelformat);
|
|
|
|
err = pxa_camera_init_videobuf2(pcdev);
|
|
if (err)
|
|
goto out_sensor_poweroff;
|
|
|
|
err = video_register_device(&pcdev->vdev, VFL_TYPE_GRABBER, -1);
|
|
if (err) {
|
|
v4l2_err(v4l2_dev, "register video device failed: %d\n", err);
|
|
pcdev->sensor = NULL;
|
|
} else {
|
|
dev_info(pcdev_to_dev(pcdev),
|
|
"PXA Camera driver attached to camera %s\n",
|
|
subdev->name);
|
|
}
|
|
|
|
out_sensor_poweroff:
|
|
err = sensor_call(pcdev, core, s_power, 0);
|
|
out:
|
|
mutex_unlock(&pcdev->mlock);
|
|
return err;
|
|
}
|
|
|
|
static void pxa_camera_sensor_unbind(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(notifier->v4l2_dev);
|
|
|
|
mutex_lock(&pcdev->mlock);
|
|
dev_info(pcdev_to_dev(pcdev),
|
|
"PXA Camera driver detached from camera %s\n",
|
|
subdev->name);
|
|
|
|
/* disable capture, disable interrupts */
|
|
__raw_writel(0x3ff, pcdev->base + CICR0);
|
|
|
|
/* Stop DMA engine */
|
|
pxa_dma_stop_channels(pcdev);
|
|
|
|
pxa_camera_destroy_formats(pcdev);
|
|
|
|
if (pcdev->mclk_clk) {
|
|
v4l2_clk_unregister(pcdev->mclk_clk);
|
|
pcdev->mclk_clk = NULL;
|
|
}
|
|
|
|
video_unregister_device(&pcdev->vdev);
|
|
pcdev->sensor = NULL;
|
|
|
|
mutex_unlock(&pcdev->mlock);
|
|
}
|
|
|
|
static const struct v4l2_async_notifier_operations pxa_camera_sensor_ops = {
|
|
.bound = pxa_camera_sensor_bound,
|
|
.unbind = pxa_camera_sensor_unbind,
|
|
};
|
|
|
|
/*
|
|
* Driver probe, remove, suspend and resume operations
|
|
*/
|
|
static int pxa_camera_suspend(struct device *dev)
|
|
{
|
|
struct pxa_camera_dev *pcdev = dev_get_drvdata(dev);
|
|
int i = 0, ret = 0;
|
|
|
|
pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR0);
|
|
pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR1);
|
|
pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR2);
|
|
pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR3);
|
|
pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR4);
|
|
|
|
if (pcdev->sensor) {
|
|
ret = sensor_call(pcdev, core, s_power, 0);
|
|
if (ret == -ENOIOCTLCMD)
|
|
ret = 0;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int pxa_camera_resume(struct device *dev)
|
|
{
|
|
struct pxa_camera_dev *pcdev = dev_get_drvdata(dev);
|
|
int i = 0, ret = 0;
|
|
|
|
__raw_writel(pcdev->save_cicr[i++] & ~CICR0_ENB, pcdev->base + CICR0);
|
|
__raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR1);
|
|
__raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR2);
|
|
__raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR3);
|
|
__raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR4);
|
|
|
|
if (pcdev->sensor) {
|
|
ret = sensor_call(pcdev, core, s_power, 1);
|
|
if (ret == -ENOIOCTLCMD)
|
|
ret = 0;
|
|
}
|
|
|
|
/* Restart frame capture if active buffer exists */
|
|
if (!ret && pcdev->active)
|
|
pxa_camera_start_capture(pcdev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int pxa_camera_pdata_from_dt(struct device *dev,
|
|
struct pxa_camera_dev *pcdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
u32 mclk_rate;
|
|
struct device_node *remote, *np = dev->of_node;
|
|
struct v4l2_fwnode_endpoint ep;
|
|
int err = of_property_read_u32(np, "clock-frequency",
|
|
&mclk_rate);
|
|
if (!err) {
|
|
pcdev->platform_flags |= PXA_CAMERA_MCLK_EN;
|
|
pcdev->mclk = mclk_rate;
|
|
}
|
|
|
|
np = of_graph_get_next_endpoint(np, NULL);
|
|
if (!np) {
|
|
dev_err(dev, "could not find endpoint\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = v4l2_fwnode_endpoint_parse(of_fwnode_handle(np), &ep);
|
|
if (err) {
|
|
dev_err(dev, "could not parse endpoint\n");
|
|
goto out;
|
|
}
|
|
|
|
switch (ep.bus.parallel.bus_width) {
|
|
case 4:
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_4;
|
|
break;
|
|
case 5:
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_5;
|
|
break;
|
|
case 8:
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_8;
|
|
break;
|
|
case 9:
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_9;
|
|
break;
|
|
case 10:
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_10;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (ep.bus.parallel.flags & V4L2_MBUS_MASTER)
|
|
pcdev->platform_flags |= PXA_CAMERA_MASTER;
|
|
if (ep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH)
|
|
pcdev->platform_flags |= PXA_CAMERA_HSP;
|
|
if (ep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH)
|
|
pcdev->platform_flags |= PXA_CAMERA_VSP;
|
|
if (ep.bus.parallel.flags & V4L2_MBUS_PCLK_SAMPLE_RISING)
|
|
pcdev->platform_flags |= PXA_CAMERA_PCLK_EN | PXA_CAMERA_PCP;
|
|
if (ep.bus.parallel.flags & V4L2_MBUS_PCLK_SAMPLE_FALLING)
|
|
pcdev->platform_flags |= PXA_CAMERA_PCLK_EN;
|
|
|
|
asd->match_type = V4L2_ASYNC_MATCH_FWNODE;
|
|
remote = of_graph_get_remote_port(np);
|
|
if (remote) {
|
|
asd->match.fwnode.fwnode = of_fwnode_handle(remote);
|
|
of_node_put(remote);
|
|
} else {
|
|
dev_notice(dev, "no remote for %pOF\n", np);
|
|
}
|
|
|
|
out:
|
|
of_node_put(np);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int pxa_camera_probe(struct platform_device *pdev)
|
|
{
|
|
struct pxa_camera_dev *pcdev;
|
|
struct resource *res;
|
|
void __iomem *base;
|
|
struct dma_slave_config config = {
|
|
.src_addr_width = 0,
|
|
.src_maxburst = 8,
|
|
.direction = DMA_DEV_TO_MEM,
|
|
};
|
|
dma_cap_mask_t mask;
|
|
struct pxad_param params;
|
|
char clk_name[V4L2_CLK_NAME_SIZE];
|
|
int irq;
|
|
int err = 0, i;
|
|
|
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
irq = platform_get_irq(pdev, 0);
|
|
if (!res || irq < 0)
|
|
return -ENODEV;
|
|
|
|
pcdev = devm_kzalloc(&pdev->dev, sizeof(*pcdev), GFP_KERNEL);
|
|
if (!pcdev) {
|
|
dev_err(&pdev->dev, "Could not allocate pcdev\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
pcdev->clk = devm_clk_get(&pdev->dev, NULL);
|
|
if (IS_ERR(pcdev->clk))
|
|
return PTR_ERR(pcdev->clk);
|
|
|
|
pcdev->res = res;
|
|
|
|
pcdev->pdata = pdev->dev.platform_data;
|
|
if (&pdev->dev.of_node && !pcdev->pdata) {
|
|
err = pxa_camera_pdata_from_dt(&pdev->dev, pcdev, &pcdev->asd);
|
|
} else {
|
|
pcdev->platform_flags = pcdev->pdata->flags;
|
|
pcdev->mclk = pcdev->pdata->mclk_10khz * 10000;
|
|
pcdev->asd.match_type = V4L2_ASYNC_MATCH_I2C;
|
|
pcdev->asd.match.i2c.adapter_id =
|
|
pcdev->pdata->sensor_i2c_adapter_id;
|
|
pcdev->asd.match.i2c.address = pcdev->pdata->sensor_i2c_address;
|
|
}
|
|
if (err < 0)
|
|
return err;
|
|
|
|
if (!(pcdev->platform_flags & (PXA_CAMERA_DATAWIDTH_8 |
|
|
PXA_CAMERA_DATAWIDTH_9 | PXA_CAMERA_DATAWIDTH_10))) {
|
|
/*
|
|
* Platform hasn't set available data widths. This is bad.
|
|
* Warn and use a default.
|
|
*/
|
|
dev_warn(&pdev->dev, "WARNING! Platform hasn't set available data widths, using default 10 bit\n");
|
|
pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_10;
|
|
}
|
|
if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_8)
|
|
pcdev->width_flags = 1 << 7;
|
|
if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_9)
|
|
pcdev->width_flags |= 1 << 8;
|
|
if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_10)
|
|
pcdev->width_flags |= 1 << 9;
|
|
if (!pcdev->mclk) {
|
|
dev_warn(&pdev->dev,
|
|
"mclk == 0! Please, fix your platform data. Using default 20MHz\n");
|
|
pcdev->mclk = 20000000;
|
|
}
|
|
|
|
pcdev->mclk_divisor = mclk_get_divisor(pdev, pcdev);
|
|
|
|
INIT_LIST_HEAD(&pcdev->capture);
|
|
spin_lock_init(&pcdev->lock);
|
|
mutex_init(&pcdev->mlock);
|
|
|
|
/*
|
|
* Request the regions.
|
|
*/
|
|
base = devm_ioremap_resource(&pdev->dev, res);
|
|
if (IS_ERR(base))
|
|
return PTR_ERR(base);
|
|
|
|
pcdev->irq = irq;
|
|
pcdev->base = base;
|
|
|
|
/* request dma */
|
|
dma_cap_zero(mask);
|
|
dma_cap_set(DMA_SLAVE, mask);
|
|
dma_cap_set(DMA_PRIVATE, mask);
|
|
|
|
params.prio = 0;
|
|
params.drcmr = 68;
|
|
pcdev->dma_chans[0] =
|
|
dma_request_slave_channel_compat(mask, pxad_filter_fn,
|
|
¶ms, &pdev->dev, "CI_Y");
|
|
if (!pcdev->dma_chans[0]) {
|
|
dev_err(&pdev->dev, "Can't request DMA for Y\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
params.drcmr = 69;
|
|
pcdev->dma_chans[1] =
|
|
dma_request_slave_channel_compat(mask, pxad_filter_fn,
|
|
¶ms, &pdev->dev, "CI_U");
|
|
if (!pcdev->dma_chans[1]) {
|
|
dev_err(&pdev->dev, "Can't request DMA for Y\n");
|
|
err = -ENODEV;
|
|
goto exit_free_dma_y;
|
|
}
|
|
|
|
params.drcmr = 70;
|
|
pcdev->dma_chans[2] =
|
|
dma_request_slave_channel_compat(mask, pxad_filter_fn,
|
|
¶ms, &pdev->dev, "CI_V");
|
|
if (!pcdev->dma_chans[2]) {
|
|
dev_err(&pdev->dev, "Can't request DMA for V\n");
|
|
err = -ENODEV;
|
|
goto exit_free_dma_u;
|
|
}
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
config.src_addr = pcdev->res->start + CIBR0 + i * 8;
|
|
err = dmaengine_slave_config(pcdev->dma_chans[i], &config);
|
|
if (err < 0) {
|
|
dev_err(&pdev->dev, "dma slave config failed: %d\n",
|
|
err);
|
|
goto exit_free_dma;
|
|
}
|
|
}
|
|
|
|
/* request irq */
|
|
err = devm_request_irq(&pdev->dev, pcdev->irq, pxa_camera_irq, 0,
|
|
PXA_CAM_DRV_NAME, pcdev);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "Camera interrupt register failed\n");
|
|
goto exit_free_dma;
|
|
}
|
|
|
|
tasklet_init(&pcdev->task_eof, pxa_camera_eof, (unsigned long)pcdev);
|
|
|
|
pxa_camera_activate(pcdev);
|
|
|
|
dev_set_drvdata(&pdev->dev, pcdev);
|
|
err = v4l2_device_register(&pdev->dev, &pcdev->v4l2_dev);
|
|
if (err)
|
|
goto exit_free_dma;
|
|
|
|
pcdev->asds[0] = &pcdev->asd;
|
|
pcdev->notifier.subdevs = pcdev->asds;
|
|
pcdev->notifier.num_subdevs = 1;
|
|
pcdev->notifier.ops = &pxa_camera_sensor_ops;
|
|
|
|
if (!of_have_populated_dt())
|
|
pcdev->asd.match_type = V4L2_ASYNC_MATCH_I2C;
|
|
|
|
err = pxa_camera_init_videobuf2(pcdev);
|
|
if (err)
|
|
goto exit_free_v4l2dev;
|
|
|
|
if (pcdev->mclk) {
|
|
v4l2_clk_name_i2c(clk_name, sizeof(clk_name),
|
|
pcdev->asd.match.i2c.adapter_id,
|
|
pcdev->asd.match.i2c.address);
|
|
|
|
pcdev->mclk_clk = v4l2_clk_register(&pxa_camera_mclk_ops,
|
|
clk_name, NULL);
|
|
if (IS_ERR(pcdev->mclk_clk)) {
|
|
err = PTR_ERR(pcdev->mclk_clk);
|
|
goto exit_free_v4l2dev;
|
|
}
|
|
}
|
|
|
|
err = v4l2_async_notifier_register(&pcdev->v4l2_dev, &pcdev->notifier);
|
|
if (err)
|
|
goto exit_free_clk;
|
|
|
|
return 0;
|
|
exit_free_clk:
|
|
v4l2_clk_unregister(pcdev->mclk_clk);
|
|
exit_free_v4l2dev:
|
|
v4l2_device_unregister(&pcdev->v4l2_dev);
|
|
exit_free_dma:
|
|
dma_release_channel(pcdev->dma_chans[2]);
|
|
exit_free_dma_u:
|
|
dma_release_channel(pcdev->dma_chans[1]);
|
|
exit_free_dma_y:
|
|
dma_release_channel(pcdev->dma_chans[0]);
|
|
return err;
|
|
}
|
|
|
|
static int pxa_camera_remove(struct platform_device *pdev)
|
|
{
|
|
struct pxa_camera_dev *pcdev = dev_get_drvdata(&pdev->dev);
|
|
|
|
pxa_camera_deactivate(pcdev);
|
|
dma_release_channel(pcdev->dma_chans[0]);
|
|
dma_release_channel(pcdev->dma_chans[1]);
|
|
dma_release_channel(pcdev->dma_chans[2]);
|
|
|
|
v4l2_async_notifier_unregister(&pcdev->notifier);
|
|
|
|
if (pcdev->mclk_clk) {
|
|
v4l2_clk_unregister(pcdev->mclk_clk);
|
|
pcdev->mclk_clk = NULL;
|
|
}
|
|
|
|
v4l2_device_unregister(&pcdev->v4l2_dev);
|
|
|
|
dev_info(&pdev->dev, "PXA Camera driver unloaded\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops pxa_camera_pm = {
|
|
.suspend = pxa_camera_suspend,
|
|
.resume = pxa_camera_resume,
|
|
};
|
|
|
|
static const struct of_device_id pxa_camera_of_match[] = {
|
|
{ .compatible = "marvell,pxa270-qci", },
|
|
{},
|
|
};
|
|
MODULE_DEVICE_TABLE(of, pxa_camera_of_match);
|
|
|
|
static struct platform_driver pxa_camera_driver = {
|
|
.driver = {
|
|
.name = PXA_CAM_DRV_NAME,
|
|
.pm = &pxa_camera_pm,
|
|
.of_match_table = of_match_ptr(pxa_camera_of_match),
|
|
},
|
|
.probe = pxa_camera_probe,
|
|
.remove = pxa_camera_remove,
|
|
};
|
|
|
|
module_platform_driver(pxa_camera_driver);
|
|
|
|
MODULE_DESCRIPTION("PXA27x SoC Camera Host driver");
|
|
MODULE_AUTHOR("Guennadi Liakhovetski <kernel@pengutronix.de>");
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_VERSION(PXA_CAM_VERSION);
|
|
MODULE_ALIAS("platform:" PXA_CAM_DRV_NAME);
|