// SPDX-License-Identifier: GPL-2.0-or-later /* * V4L2 Driver for PXA camera host * * Copyright (C) 2006, Sascha Hauer, Pengutronix * Copyright (C) 2008, Guennadi Liakhovetski * Copyright (C) 2016, Robert Jarzmik */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #define PXA_CAM_VERSION "0.0.6" #define PXA_CAM_DRV_NAME "pxa27x-camera" #define DEFAULT_WIDTH 640 #define DEFAULT_HEIGHT 480 /* Camera Interface */ #define CICR0 0x0000 #define CICR1 0x0004 #define CICR2 0x0008 #define CICR3 0x000C #define CICR4 0x0010 #define CISR 0x0014 #define CIFR 0x0018 #define CITOR 0x001C #define CIBR0 0x0028 #define CIBR1 0x0030 #define CIBR2 0x0038 #define CICR0_DMAEN (1UL << 31) /* DMA request enable */ #define CICR0_PAR_EN (1 << 30) /* Parity enable */ #define CICR0_SL_CAP_EN (1 << 29) /* Capture enable for slave mode */ #define CICR0_ENB (1 << 28) /* Camera interface enable */ #define CICR0_DIS (1 << 27) /* Camera interface disable */ #define CICR0_SIM (0x7 << 24) /* Sensor interface mode mask */ #define CICR0_TOM (1 << 9) /* Time-out mask */ #define CICR0_RDAVM (1 << 8) /* Receive-data-available mask */ #define CICR0_FEM (1 << 7) /* FIFO-empty mask */ #define CICR0_EOLM (1 << 6) /* End-of-line mask */ #define CICR0_PERRM (1 << 5) /* Parity-error mask */ #define CICR0_QDM (1 << 4) /* Quick-disable mask */ #define CICR0_CDM (1 << 3) /* Disable-done mask */ #define CICR0_SOFM (1 << 2) /* Start-of-frame mask */ #define CICR0_EOFM (1 << 1) /* End-of-frame mask */ #define CICR0_FOM (1 << 0) /* FIFO-overrun mask */ #define CICR1_TBIT (1UL << 31) /* Transparency bit */ #define CICR1_RGBT_CONV (0x3 << 29) /* RGBT conversion mask */ #define CICR1_PPL (0x7ff << 15) /* Pixels per line mask */ #define CICR1_RGB_CONV (0x7 << 12) /* RGB conversion mask */ #define CICR1_RGB_F (1 << 11) /* RGB format */ #define CICR1_YCBCR_F (1 << 10) /* YCbCr format */ #define CICR1_RGB_BPP (0x7 << 7) /* RGB bis per pixel mask */ #define CICR1_RAW_BPP (0x3 << 5) /* Raw bis per pixel mask */ #define CICR1_COLOR_SP (0x3 << 3) /* Color space mask */ #define CICR1_DW (0x7 << 0) /* Data width mask */ #define CICR2_BLW (0xff << 24) /* Beginning-of-line pixel clock wait count mask */ #define CICR2_ELW (0xff << 16) /* End-of-line pixel clock wait count mask */ #define CICR2_HSW (0x3f << 10) /* Horizontal sync pulse width mask */ #define CICR2_BFPW (0x3f << 3) /* Beginning-of-frame pixel clock wait count mask */ #define CICR2_FSW (0x7 << 0) /* Frame stabilization wait count mask */ #define CICR3_BFW (0xff << 24) /* Beginning-of-frame line clock wait count mask */ #define CICR3_EFW (0xff << 16) /* End-of-frame line clock wait count mask */ #define CICR3_VSW (0x3f << 10) /* Vertical sync pulse width mask */ #define CICR3_BFPW (0x3f << 3) /* Beginning-of-frame pixel clock wait count mask */ #define CICR3_LPF (0x7ff << 0) /* Lines per frame mask */ #define CICR4_MCLK_DLY (0x3 << 24) /* MCLK Data Capture Delay mask */ #define CICR4_PCLK_EN (1 << 23) /* Pixel clock enable */ #define CICR4_PCP (1 << 22) /* Pixel clock polarity */ #define CICR4_HSP (1 << 21) /* Horizontal sync polarity */ #define CICR4_VSP (1 << 20) /* Vertical sync polarity */ #define CICR4_MCLK_EN (1 << 19) /* MCLK enable */ #define CICR4_FR_RATE (0x7 << 8) /* Frame rate mask */ #define CICR4_DIV (0xff << 0) /* Clock divisor mask */ #define CISR_FTO (1 << 15) /* FIFO time-out */ #define CISR_RDAV_2 (1 << 14) /* Channel 2 receive data available */ #define CISR_RDAV_1 (1 << 13) /* Channel 1 receive data available */ #define CISR_RDAV_0 (1 << 12) /* Channel 0 receive data available */ #define CISR_FEMPTY_2 (1 << 11) /* Channel 2 FIFO empty */ #define CISR_FEMPTY_1 (1 << 10) /* Channel 1 FIFO empty */ #define CISR_FEMPTY_0 (1 << 9) /* Channel 0 FIFO empty */ #define CISR_EOL (1 << 8) /* End of line */ #define CISR_PAR_ERR (1 << 7) /* Parity error */ #define CISR_CQD (1 << 6) /* Camera interface quick disable */ #define CISR_CDD (1 << 5) /* Camera interface disable done */ #define CISR_SOF (1 << 4) /* Start of frame */ #define CISR_EOF (1 << 3) /* End of frame */ #define CISR_IFO_2 (1 << 2) /* FIFO overrun for Channel 2 */ #define CISR_IFO_1 (1 << 1) /* FIFO overrun for Channel 1 */ #define CISR_IFO_0 (1 << 0) /* FIFO overrun for Channel 0 */ #define CIFR_FLVL2 (0x7f << 23) /* FIFO 2 level mask */ #define CIFR_FLVL1 (0x7f << 16) /* FIFO 1 level mask */ #define CIFR_FLVL0 (0xff << 8) /* FIFO 0 level mask */ #define CIFR_THL_0 (0x3 << 4) /* Threshold Level for Channel 0 FIFO */ #define CIFR_RESET_F (1 << 3) /* Reset input FIFOs */ #define CIFR_FEN2 (1 << 2) /* FIFO enable for channel 2 */ #define CIFR_FEN1 (1 << 1) /* FIFO enable for channel 1 */ #define CIFR_FEN0 (1 << 0) /* FIFO enable for channel 0 */ #define CICR0_SIM_MP (0 << 24) #define CICR0_SIM_SP (1 << 24) #define CICR0_SIM_MS (2 << 24) #define CICR0_SIM_EP (3 << 24) #define CICR0_SIM_ES (4 << 24) #define CICR1_DW_VAL(x) ((x) & CICR1_DW) /* Data bus width */ #define CICR1_PPL_VAL(x) (((x) << 15) & CICR1_PPL) /* Pixels per line */ #define CICR1_COLOR_SP_VAL(x) (((x) << 3) & CICR1_COLOR_SP) /* color space */ #define CICR1_RGB_BPP_VAL(x) (((x) << 7) & CICR1_RGB_BPP) /* bpp for rgb */ #define CICR1_RGBT_CONV_VAL(x) (((x) << 29) & CICR1_RGBT_CONV) /* rgbt conv */ #define CICR2_BLW_VAL(x) (((x) << 24) & CICR2_BLW) /* Beginning-of-line pixel clock wait count */ #define CICR2_ELW_VAL(x) (((x) << 16) & CICR2_ELW) /* End-of-line pixel clock wait count */ #define CICR2_HSW_VAL(x) (((x) << 10) & CICR2_HSW) /* Horizontal sync pulse width */ #define CICR2_BFPW_VAL(x) (((x) << 3) & CICR2_BFPW) /* Beginning-of-frame pixel clock wait count */ #define CICR2_FSW_VAL(x) (((x) << 0) & CICR2_FSW) /* Frame stabilization wait count */ #define CICR3_BFW_VAL(x) (((x) << 24) & CICR3_BFW) /* Beginning-of-frame line clock wait count */ #define CICR3_EFW_VAL(x) (((x) << 16) & CICR3_EFW) /* End-of-frame line clock wait count */ #define CICR3_VSW_VAL(x) (((x) << 11) & CICR3_VSW) /* Vertical sync pulse width */ #define CICR3_LPF_VAL(x) (((x) << 0) & CICR3_LPF) /* Lines per frame */ #define CICR0_IRQ_MASK (CICR0_TOM | CICR0_RDAVM | CICR0_FEM | CICR0_EOLM | \ CICR0_PERRM | CICR0_QDM | CICR0_CDM | CICR0_SOFM | \ CICR0_EOFM | CICR0_FOM) #define sensor_call(cam, o, f, args...) \ v4l2_subdev_call(cam->sensor, o, f, ##args) /* * Format handling */ /** * enum pxa_mbus_packing - data packing types on the media-bus * @PXA_MBUS_PACKING_NONE: no packing, bit-for-bit transfer to RAM, one * sample represents one pixel * @PXA_MBUS_PACKING_2X8_PADHI: 16 bits transferred in 2 8-bit samples, in the * possibly incomplete byte high bits are padding * @PXA_MBUS_PACKING_EXTEND16: sample width (e.g., 10 bits) has to be extended * to 16 bits */ enum pxa_mbus_packing { PXA_MBUS_PACKING_NONE, PXA_MBUS_PACKING_2X8_PADHI, PXA_MBUS_PACKING_EXTEND16, }; /** * enum pxa_mbus_order - sample order on the media bus * @PXA_MBUS_ORDER_LE: least significant sample first * @PXA_MBUS_ORDER_BE: most significant sample first */ enum pxa_mbus_order { PXA_MBUS_ORDER_LE, PXA_MBUS_ORDER_BE, }; /** * enum pxa_mbus_layout - planes layout in memory * @PXA_MBUS_LAYOUT_PACKED: color components packed * @PXA_MBUS_LAYOUT_PLANAR_2Y_U_V: YUV components stored in 3 planes (4:2:2) * @PXA_MBUS_LAYOUT_PLANAR_2Y_C: YUV components stored in a luma and a * chroma plane (C plane is half the size * of Y plane) * @PXA_MBUS_LAYOUT_PLANAR_Y_C: YUV components stored in a luma and a * chroma plane (C plane is the same size * as Y plane) */ enum pxa_mbus_layout { PXA_MBUS_LAYOUT_PACKED = 0, PXA_MBUS_LAYOUT_PLANAR_2Y_U_V, PXA_MBUS_LAYOUT_PLANAR_2Y_C, PXA_MBUS_LAYOUT_PLANAR_Y_C, }; /** * struct pxa_mbus_pixelfmt - Data format on the media bus * @name: Name of the format * @fourcc: Fourcc code, that will be obtained if the data is * stored in memory in the following way: * @packing: Type of sample-packing, that has to be used * @order: Sample order when storing in memory * @layout: Planes layout in memory * @bits_per_sample: How many bits the bridge has to sample */ struct pxa_mbus_pixelfmt { const char *name; u32 fourcc; enum pxa_mbus_packing packing; enum pxa_mbus_order order; enum pxa_mbus_layout layout; u8 bits_per_sample; }; /** * struct pxa_mbus_lookup - Lookup FOURCC IDs by mediabus codes for pass-through * @code: mediabus pixel-code * @fmt: pixel format description */ struct pxa_mbus_lookup { u32 code; struct pxa_mbus_pixelfmt fmt; }; static const struct pxa_mbus_lookup mbus_fmt[] = { { .code = MEDIA_BUS_FMT_YUYV8_2X8, .fmt = { .fourcc = V4L2_PIX_FMT_YUYV, .name = "YUYV", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_YVYU8_2X8, .fmt = { .fourcc = V4L2_PIX_FMT_YVYU, .name = "YVYU", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_UYVY8_2X8, .fmt = { .fourcc = V4L2_PIX_FMT_UYVY, .name = "UYVY", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_VYUY8_2X8, .fmt = { .fourcc = V4L2_PIX_FMT_VYUY, .name = "VYUY", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE, .fmt = { .fourcc = V4L2_PIX_FMT_RGB555, .name = "RGB555", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_BE, .fmt = { .fourcc = V4L2_PIX_FMT_RGB555X, .name = "RGB555X", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_BE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_RGB565_2X8_LE, .fmt = { .fourcc = V4L2_PIX_FMT_RGB565, .name = "RGB565", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_RGB565_2X8_BE, .fmt = { .fourcc = V4L2_PIX_FMT_RGB565X, .name = "RGB565X", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_BE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SBGGR8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_SBGGR8, .name = "Bayer 8 BGGR", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGBRG8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_SGBRG8, .name = "Bayer 8 GBRG", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGRBG8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_SGRBG8, .name = "Bayer 8 GRBG", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SRGGB8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_SRGGB8, .name = "Bayer 8 RGGB", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SBGGR10_1X10, .fmt = { .fourcc = V4L2_PIX_FMT_SBGGR10, .name = "Bayer 10 BGGR", .bits_per_sample = 10, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_Y8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_GREY, .name = "Grey", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_Y10_1X10, .fmt = { .fourcc = V4L2_PIX_FMT_Y10, .name = "Grey 10bit", .bits_per_sample = 10, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SBGGR10_2X8_PADHI_LE, .fmt = { .fourcc = V4L2_PIX_FMT_SBGGR10, .name = "Bayer 10 BGGR", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SBGGR10_2X8_PADHI_BE, .fmt = { .fourcc = V4L2_PIX_FMT_SBGGR10, .name = "Bayer 10 BGGR", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_BE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_RGB444_2X8_PADHI_BE, .fmt = { .fourcc = V4L2_PIX_FMT_RGB444, .name = "RGB444", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_BE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_UYVY8_1X16, .fmt = { .fourcc = V4L2_PIX_FMT_UYVY, .name = "UYVY 16bit", .bits_per_sample = 16, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_VYUY8_1X16, .fmt = { .fourcc = V4L2_PIX_FMT_VYUY, .name = "VYUY 16bit", .bits_per_sample = 16, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_YUYV8_1X16, .fmt = { .fourcc = V4L2_PIX_FMT_YUYV, .name = "YUYV 16bit", .bits_per_sample = 16, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_YVYU8_1X16, .fmt = { .fourcc = V4L2_PIX_FMT_YVYU, .name = "YVYU 16bit", .bits_per_sample = 16, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGRBG10_DPCM8_1X8, .fmt = { .fourcc = V4L2_PIX_FMT_SGRBG10DPCM8, .name = "Bayer 10 BGGR DPCM 8", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_NONE, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGBRG10_1X10, .fmt = { .fourcc = V4L2_PIX_FMT_SGBRG10, .name = "Bayer 10 GBRG", .bits_per_sample = 10, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGRBG10_1X10, .fmt = { .fourcc = V4L2_PIX_FMT_SGRBG10, .name = "Bayer 10 GRBG", .bits_per_sample = 10, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SRGGB10_1X10, .fmt = { .fourcc = V4L2_PIX_FMT_SRGGB10, .name = "Bayer 10 RGGB", .bits_per_sample = 10, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .fmt = { .fourcc = V4L2_PIX_FMT_SBGGR12, .name = "Bayer 12 BGGR", .bits_per_sample = 12, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .fmt = { .fourcc = V4L2_PIX_FMT_SGBRG12, .name = "Bayer 12 GBRG", .bits_per_sample = 12, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .fmt = { .fourcc = V4L2_PIX_FMT_SGRBG12, .name = "Bayer 12 GRBG", .bits_per_sample = 12, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .fmt = { .fourcc = V4L2_PIX_FMT_SRGGB12, .name = "Bayer 12 RGGB", .bits_per_sample = 12, .packing = PXA_MBUS_PACKING_EXTEND16, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PACKED, }, }, }; static s32 pxa_mbus_bytes_per_line(u32 width, const struct pxa_mbus_pixelfmt *mf) { if (mf->layout != PXA_MBUS_LAYOUT_PACKED) return width * mf->bits_per_sample / 8; switch (mf->packing) { case PXA_MBUS_PACKING_NONE: return width * mf->bits_per_sample / 8; case PXA_MBUS_PACKING_2X8_PADHI: case PXA_MBUS_PACKING_EXTEND16: return width * 2; } return -EINVAL; } static s32 pxa_mbus_image_size(const struct pxa_mbus_pixelfmt *mf, u32 bytes_per_line, u32 height) { if (mf->layout == PXA_MBUS_LAYOUT_PACKED) return bytes_per_line * height; switch (mf->packing) { case PXA_MBUS_PACKING_2X8_PADHI: return bytes_per_line * height * 2; default: return -EINVAL; } } static const struct pxa_mbus_pixelfmt *pxa_mbus_find_fmtdesc( u32 code, const struct pxa_mbus_lookup *lookup, int n) { int i; for (i = 0; i < n; i++) if (lookup[i].code == code) return &lookup[i].fmt; return NULL; } static const struct pxa_mbus_pixelfmt *pxa_mbus_get_fmtdesc( u32 code) { return pxa_mbus_find_fmtdesc(code, mbus_fmt, ARRAY_SIZE(mbus_fmt)); } /** * struct pxa_camera_format_xlate - match between host and sensor formats * @code: code of a sensor provided format * @host_fmt: host format after host translation from code * * Host and sensor translation structure. Used in table of host and sensor * formats matchings in pxa_camera_device. A host can override the generic list * generation by implementing get_formats(), and use it for format checks and * format setup. */ struct pxa_camera_format_xlate { u32 code; const struct pxa_mbus_pixelfmt *host_fmt; }; /* * Structures */ enum pxa_camera_active_dma { DMA_Y = 0x1, DMA_U = 0x2, DMA_V = 0x4, }; /* buffer for one video frame */ struct pxa_buffer { /* common v4l buffer stuff -- must be first */ struct vb2_v4l2_buffer vbuf; struct list_head queue; u32 code; int nb_planes; /* our descriptor lists for Y, U and V channels */ struct dma_async_tx_descriptor *descs[3]; dma_cookie_t cookie[3]; struct scatterlist *sg[3]; int sg_len[3]; size_t plane_sizes[3]; int inwork; enum pxa_camera_active_dma active_dma; }; struct pxa_camera_dev { struct v4l2_device v4l2_dev; struct video_device vdev; struct v4l2_async_notifier notifier; struct vb2_queue vb2_vq; struct v4l2_subdev *sensor; struct pxa_camera_format_xlate *user_formats; const struct pxa_camera_format_xlate *current_fmt; struct v4l2_pix_format current_pix; struct v4l2_async_subdev asd; /* * PXA27x is only supposed to handle one camera on its Quick Capture * interface. If anyone ever builds hardware to enable more than * one camera, they will have to modify this driver too */ struct clk *clk; unsigned int irq; void __iomem *base; int channels; struct dma_chan *dma_chans[3]; struct pxacamera_platform_data *pdata; struct resource *res; unsigned long platform_flags; unsigned long ciclk; unsigned long mclk; u32 mclk_divisor; struct v4l2_clk *mclk_clk; u16 width_flags; /* max 10 bits */ struct list_head capture; spinlock_t lock; struct mutex mlock; unsigned int buf_sequence; struct pxa_buffer *active; struct tasklet_struct task_eof; u32 save_cicr[5]; }; struct pxa_cam { unsigned long flags; }; static const char *pxa_cam_driver_description = "PXA_Camera"; /* * Format translation functions */ static const struct pxa_camera_format_xlate *pxa_mbus_xlate_by_fourcc(struct pxa_camera_format_xlate *user_formats, unsigned int fourcc) { unsigned int i; for (i = 0; user_formats[i].code; i++) if (user_formats[i].host_fmt->fourcc == fourcc) return user_formats + i; return NULL; } static struct pxa_camera_format_xlate *pxa_mbus_build_fmts_xlate( struct v4l2_device *v4l2_dev, struct v4l2_subdev *subdev, int (*get_formats)(struct v4l2_device *, unsigned int, struct pxa_camera_format_xlate *xlate)) { unsigned int i, fmts = 0, raw_fmts = 0; int ret; struct v4l2_subdev_mbus_code_enum code = { .which = V4L2_SUBDEV_FORMAT_ACTIVE, }; struct pxa_camera_format_xlate *user_formats; while (!v4l2_subdev_call(subdev, pad, enum_mbus_code, NULL, &code)) { raw_fmts++; code.index++; } /* * First pass - only count formats this host-sensor * configuration can provide */ for (i = 0; i < raw_fmts; i++) { ret = get_formats(v4l2_dev, i, NULL); if (ret < 0) return ERR_PTR(ret); fmts += ret; } if (!fmts) return ERR_PTR(-ENXIO); user_formats = kcalloc(fmts + 1, sizeof(*user_formats), GFP_KERNEL); if (!user_formats) return ERR_PTR(-ENOMEM); /* Second pass - actually fill data formats */ fmts = 0; for (i = 0; i < raw_fmts; i++) { ret = get_formats(v4l2_dev, i, user_formats + fmts); if (ret < 0) goto egfmt; fmts += ret; } user_formats[fmts].code = 0; return user_formats; egfmt: kfree(user_formats); return ERR_PTR(ret); } /* * Videobuf operations */ static struct pxa_buffer *vb2_to_pxa_buffer(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); return container_of(vbuf, struct pxa_buffer, vbuf); } static struct device *pcdev_to_dev(struct pxa_camera_dev *pcdev) { return pcdev->v4l2_dev.dev; } static struct pxa_camera_dev *v4l2_dev_to_pcdev(struct v4l2_device *v4l2_dev) { return container_of(v4l2_dev, struct pxa_camera_dev, v4l2_dev); } static void pxa_camera_dma_irq(struct pxa_camera_dev *pcdev, enum pxa_camera_active_dma act_dma); static void pxa_camera_dma_irq_y(void *data) { struct pxa_camera_dev *pcdev = data; pxa_camera_dma_irq(pcdev, DMA_Y); } static void pxa_camera_dma_irq_u(void *data) { struct pxa_camera_dev *pcdev = data; pxa_camera_dma_irq(pcdev, DMA_U); } static void pxa_camera_dma_irq_v(void *data) { struct pxa_camera_dev *pcdev = data; pxa_camera_dma_irq(pcdev, DMA_V); } /** * pxa_init_dma_channel - init dma descriptors * @pcdev: pxa camera device * @buf: pxa camera buffer * @channel: dma channel (0 => 'Y', 1 => 'U', 2 => 'V') * @sg: dma scatter list * @sglen: dma scatter list length * * Prepares the pxa dma descriptors to transfer one camera channel. * * Returns 0 if success or -ENOMEM if no memory is available */ static int pxa_init_dma_channel(struct pxa_camera_dev *pcdev, struct pxa_buffer *buf, int channel, struct scatterlist *sg, int sglen) { struct dma_chan *dma_chan = pcdev->dma_chans[channel]; struct dma_async_tx_descriptor *tx; tx = dmaengine_prep_slave_sg(dma_chan, sg, sglen, DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT | DMA_CTRL_REUSE); if (!tx) { dev_err(pcdev_to_dev(pcdev), "dmaengine_prep_slave_sg failed\n"); goto fail; } tx->callback_param = pcdev; switch (channel) { case 0: tx->callback = pxa_camera_dma_irq_y; break; case 1: tx->callback = pxa_camera_dma_irq_u; break; case 2: tx->callback = pxa_camera_dma_irq_v; break; } buf->descs[channel] = tx; return 0; fail: dev_dbg(pcdev_to_dev(pcdev), "%s (vb=%p) dma_tx=%p\n", __func__, buf, tx); return -ENOMEM; } static void pxa_videobuf_set_actdma(struct pxa_camera_dev *pcdev, struct pxa_buffer *buf) { buf->active_dma = DMA_Y; if (buf->nb_planes == 3) buf->active_dma |= DMA_U | DMA_V; } /** * pxa_dma_start_channels - start DMA channel for active buffer * @pcdev: pxa camera device * * Initialize DMA channels to the beginning of the active video buffer, and * start these channels. */ static void pxa_dma_start_channels(struct pxa_camera_dev *pcdev) { int i; for (i = 0; i < pcdev->channels; i++) { dev_dbg(pcdev_to_dev(pcdev), "%s (channel=%d)\n", __func__, i); dma_async_issue_pending(pcdev->dma_chans[i]); } } static void pxa_dma_stop_channels(struct pxa_camera_dev *pcdev) { int i; for (i = 0; i < pcdev->channels; i++) { dev_dbg(pcdev_to_dev(pcdev), "%s (channel=%d)\n", __func__, i); dmaengine_terminate_all(pcdev->dma_chans[i]); } } static void pxa_dma_add_tail_buf(struct pxa_camera_dev *pcdev, struct pxa_buffer *buf) { int i; for (i = 0; i < pcdev->channels; i++) { buf->cookie[i] = dmaengine_submit(buf->descs[i]); dev_dbg(pcdev_to_dev(pcdev), "%s (channel=%d) : submit vb=%p cookie=%d\n", __func__, i, buf, buf->descs[i]->cookie); } } /** * pxa_camera_start_capture - start video capturing * @pcdev: camera device * * Launch capturing. DMA channels should not be active yet. They should get * activated at the end of frame interrupt, to capture only whole frames, and * never begin the capture of a partial frame. */ static void pxa_camera_start_capture(struct pxa_camera_dev *pcdev) { unsigned long cicr0; dev_dbg(pcdev_to_dev(pcdev), "%s\n", __func__); __raw_writel(__raw_readl(pcdev->base + CISR), pcdev->base + CISR); /* Enable End-Of-Frame Interrupt */ cicr0 = __raw_readl(pcdev->base + CICR0) | CICR0_ENB; cicr0 &= ~CICR0_EOFM; __raw_writel(cicr0, pcdev->base + CICR0); } static void pxa_camera_stop_capture(struct pxa_camera_dev *pcdev) { unsigned long cicr0; pxa_dma_stop_channels(pcdev); cicr0 = __raw_readl(pcdev->base + CICR0) & ~CICR0_ENB; __raw_writel(cicr0, pcdev->base + CICR0); pcdev->active = NULL; dev_dbg(pcdev_to_dev(pcdev), "%s\n", __func__); } static void pxa_camera_wakeup(struct pxa_camera_dev *pcdev, struct pxa_buffer *buf, enum vb2_buffer_state state) { struct vb2_buffer *vb = &buf->vbuf.vb2_buf; struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); /* _init is used to debug races, see comment in pxa_camera_reqbufs() */ list_del_init(&buf->queue); vb->timestamp = ktime_get_ns(); vbuf->sequence = pcdev->buf_sequence++; vbuf->field = V4L2_FIELD_NONE; vb2_buffer_done(vb, VB2_BUF_STATE_DONE); dev_dbg(pcdev_to_dev(pcdev), "%s dequeued buffer (buf=0x%p)\n", __func__, buf); if (list_empty(&pcdev->capture)) { pxa_camera_stop_capture(pcdev); return; } pcdev->active = list_entry(pcdev->capture.next, struct pxa_buffer, queue); } /** * pxa_camera_check_link_miss - check missed DMA linking * @pcdev: camera device * @last_submitted: an opaque DMA cookie for last submitted * @last_issued: an opaque DMA cookie for last issued * * The DMA chaining is done with DMA running. This means a tiny temporal window * remains, where a buffer is queued on the chain, while the chain is already * stopped. This means the tailed buffer would never be transferred by DMA. * This function restarts the capture for this corner case, where : * - DADR() == DADDR_STOP * - a videobuffer is queued on the pcdev->capture list * * Please check the "DMA hot chaining timeslice issue" in * Documentation/driver-api/media/drivers/pxa_camera.rst * * Context: should only be called within the dma irq handler */ static void pxa_camera_check_link_miss(struct pxa_camera_dev *pcdev, dma_cookie_t last_submitted, dma_cookie_t last_issued) { bool is_dma_stopped = last_submitted != last_issued; dev_dbg(pcdev_to_dev(pcdev), "%s : top queued buffer=%p, is_dma_stopped=%d\n", __func__, pcdev->active, is_dma_stopped); if (pcdev->active && is_dma_stopped) pxa_camera_start_capture(pcdev); } static void pxa_camera_dma_irq(struct pxa_camera_dev *pcdev, enum pxa_camera_active_dma act_dma) { struct pxa_buffer *buf, *last_buf; unsigned long flags; u32 camera_status, overrun; int chan; enum dma_status last_status; dma_cookie_t last_issued; spin_lock_irqsave(&pcdev->lock, flags); camera_status = __raw_readl(pcdev->base + CISR); dev_dbg(pcdev_to_dev(pcdev), "camera dma irq, cisr=0x%x dma=%d\n", camera_status, act_dma); overrun = CISR_IFO_0; if (pcdev->channels == 3) overrun |= CISR_IFO_1 | CISR_IFO_2; /* * pcdev->active should not be NULL in DMA irq handler. * * But there is one corner case : if capture was stopped due to an * overrun of channel 1, and at that same channel 2 was completed. * * When handling the overrun in DMA irq for channel 1, we'll stop the * capture and restart it (and thus set pcdev->active to NULL). But the * DMA irq handler will already be pending for channel 2. So on entering * the DMA irq handler for channel 2 there will be no active buffer, yet * that is normal. */ if (!pcdev->active) goto out; buf = pcdev->active; WARN_ON(buf->inwork || list_empty(&buf->queue)); /* * It's normal if the last frame creates an overrun, as there * are no more DMA descriptors to fetch from QCI fifos */ switch (act_dma) { case DMA_U: chan = 1; break; case DMA_V: chan = 2; break; default: chan = 0; break; } last_buf = list_entry(pcdev->capture.prev, struct pxa_buffer, queue); last_status = dma_async_is_tx_complete(pcdev->dma_chans[chan], last_buf->cookie[chan], NULL, &last_issued); if (camera_status & overrun && last_status != DMA_COMPLETE) { dev_dbg(pcdev_to_dev(pcdev), "FIFO overrun! CISR: %x\n", camera_status); pxa_camera_stop_capture(pcdev); list_for_each_entry(buf, &pcdev->capture, queue) pxa_dma_add_tail_buf(pcdev, buf); pxa_camera_start_capture(pcdev); goto out; } buf->active_dma &= ~act_dma; if (!buf->active_dma) { pxa_camera_wakeup(pcdev, buf, VB2_BUF_STATE_DONE); pxa_camera_check_link_miss(pcdev, last_buf->cookie[chan], last_issued); } out: spin_unlock_irqrestore(&pcdev->lock, flags); } static u32 mclk_get_divisor(struct platform_device *pdev, struct pxa_camera_dev *pcdev) { unsigned long mclk = pcdev->mclk; u32 div; unsigned long lcdclk; lcdclk = clk_get_rate(pcdev->clk); pcdev->ciclk = lcdclk; /* mclk <= ciclk / 4 (27.4.2) */ if (mclk > lcdclk / 4) { mclk = lcdclk / 4; dev_warn(&pdev->dev, "Limiting master clock to %lu\n", mclk); } /* We verify mclk != 0, so if anyone breaks it, here comes their Oops */ div = (lcdclk + 2 * mclk - 1) / (2 * mclk) - 1; /* If we're not supplying MCLK, leave it at 0 */ if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN) pcdev->mclk = lcdclk / (2 * (div + 1)); dev_dbg(&pdev->dev, "LCD clock %luHz, target freq %luHz, divisor %u\n", lcdclk, mclk, div); return div; } static void recalculate_fifo_timeout(struct pxa_camera_dev *pcdev, unsigned long pclk) { /* We want a timeout > 1 pixel time, not ">=" */ u32 ciclk_per_pixel = pcdev->ciclk / pclk + 1; __raw_writel(ciclk_per_pixel, pcdev->base + CITOR); } static void pxa_camera_activate(struct pxa_camera_dev *pcdev) { u32 cicr4 = 0; /* disable all interrupts */ __raw_writel(0x3ff, pcdev->base + CICR0); if (pcdev->platform_flags & PXA_CAMERA_PCLK_EN) cicr4 |= CICR4_PCLK_EN; if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN) cicr4 |= CICR4_MCLK_EN; if (pcdev->platform_flags & PXA_CAMERA_PCP) cicr4 |= CICR4_PCP; if (pcdev->platform_flags & PXA_CAMERA_HSP) cicr4 |= CICR4_HSP; if (pcdev->platform_flags & PXA_CAMERA_VSP) cicr4 |= CICR4_VSP; __raw_writel(pcdev->mclk_divisor | cicr4, pcdev->base + CICR4); if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN) /* Initialise the timeout under the assumption pclk = mclk */ recalculate_fifo_timeout(pcdev, pcdev->mclk); else /* "Safe default" - 13MHz */ recalculate_fifo_timeout(pcdev, 13000000); clk_prepare_enable(pcdev->clk); } static void pxa_camera_deactivate(struct pxa_camera_dev *pcdev) { clk_disable_unprepare(pcdev->clk); } static void pxa_camera_eof(struct tasklet_struct *t) { struct pxa_camera_dev *pcdev = from_tasklet(pcdev, t, task_eof); unsigned long cifr; struct pxa_buffer *buf; dev_dbg(pcdev_to_dev(pcdev), "Camera interrupt status 0x%x\n", __raw_readl(pcdev->base + CISR)); /* Reset the FIFOs */ cifr = __raw_readl(pcdev->base + CIFR) | CIFR_RESET_F; __raw_writel(cifr, pcdev->base + CIFR); pcdev->active = list_first_entry(&pcdev->capture, struct pxa_buffer, queue); buf = pcdev->active; pxa_videobuf_set_actdma(pcdev, buf); pxa_dma_start_channels(pcdev); } static irqreturn_t pxa_camera_irq(int irq, void *data) { struct pxa_camera_dev *pcdev = data; unsigned long status, cicr0; status = __raw_readl(pcdev->base + CISR); dev_dbg(pcdev_to_dev(pcdev), "Camera interrupt status 0x%lx\n", status); if (!status) return IRQ_NONE; __raw_writel(status, pcdev->base + CISR); if (status & CISR_EOF) { cicr0 = __raw_readl(pcdev->base + CICR0) | CICR0_EOFM; __raw_writel(cicr0, pcdev->base + CICR0); tasklet_schedule(&pcdev->task_eof); } return IRQ_HANDLED; } static void pxa_camera_setup_cicr(struct pxa_camera_dev *pcdev, unsigned long flags, __u32 pixfmt) { unsigned long dw, bpp; u32 cicr0, cicr1, cicr2, cicr3, cicr4 = 0, y_skip_top; int ret = sensor_call(pcdev, sensor, g_skip_top_lines, &y_skip_top); if (ret < 0) y_skip_top = 0; /* * Datawidth is now guaranteed to be equal to one of the three values. * We fix bit-per-pixel equal to data-width... */ switch (pcdev->current_fmt->host_fmt->bits_per_sample) { case 10: dw = 4; bpp = 0x40; break; case 9: dw = 3; bpp = 0x20; break; default: /* * Actually it can only be 8 now, * default is just to silence compiler warnings */ case 8: dw = 2; bpp = 0; } if (pcdev->platform_flags & PXA_CAMERA_PCLK_EN) cicr4 |= CICR4_PCLK_EN; if (pcdev->platform_flags & PXA_CAMERA_MCLK_EN) cicr4 |= CICR4_MCLK_EN; if (flags & V4L2_MBUS_PCLK_SAMPLE_FALLING) cicr4 |= CICR4_PCP; if (flags & V4L2_MBUS_HSYNC_ACTIVE_LOW) cicr4 |= CICR4_HSP; if (flags & V4L2_MBUS_VSYNC_ACTIVE_LOW) cicr4 |= CICR4_VSP; cicr0 = __raw_readl(pcdev->base + CICR0); if (cicr0 & CICR0_ENB) __raw_writel(cicr0 & ~CICR0_ENB, pcdev->base + CICR0); cicr1 = CICR1_PPL_VAL(pcdev->current_pix.width - 1) | bpp | dw; switch (pixfmt) { case V4L2_PIX_FMT_YUV422P: pcdev->channels = 3; cicr1 |= CICR1_YCBCR_F; /* * Normally, pxa bus wants as input UYVY format. We allow all * reorderings of the YUV422 format, as no processing is done, * and the YUV stream is just passed through without any * transformation. Note that UYVY is the only format that * should be used if pxa framebuffer Overlay2 is used. */ /* fall through */ case V4L2_PIX_FMT_UYVY: case V4L2_PIX_FMT_VYUY: case V4L2_PIX_FMT_YUYV: case V4L2_PIX_FMT_YVYU: cicr1 |= CICR1_COLOR_SP_VAL(2); break; case V4L2_PIX_FMT_RGB555: cicr1 |= CICR1_RGB_BPP_VAL(1) | CICR1_RGBT_CONV_VAL(2) | CICR1_TBIT | CICR1_COLOR_SP_VAL(1); break; case V4L2_PIX_FMT_RGB565: cicr1 |= CICR1_COLOR_SP_VAL(1) | CICR1_RGB_BPP_VAL(2); break; } cicr2 = 0; cicr3 = CICR3_LPF_VAL(pcdev->current_pix.height - 1) | CICR3_BFW_VAL(min((u32)255, y_skip_top)); cicr4 |= pcdev->mclk_divisor; __raw_writel(cicr1, pcdev->base + CICR1); __raw_writel(cicr2, pcdev->base + CICR2); __raw_writel(cicr3, pcdev->base + CICR3); __raw_writel(cicr4, pcdev->base + CICR4); /* CIF interrupts are not used, only DMA */ cicr0 = (cicr0 & CICR0_ENB) | (pcdev->platform_flags & PXA_CAMERA_MASTER ? CICR0_SIM_MP : (CICR0_SL_CAP_EN | CICR0_SIM_SP)); cicr0 |= CICR0_DMAEN | CICR0_IRQ_MASK; __raw_writel(cicr0, pcdev->base + CICR0); } /* * Videobuf2 section */ static void pxa_buffer_cleanup(struct pxa_buffer *buf) { int i; for (i = 0; i < 3 && buf->descs[i]; i++) { dmaengine_desc_free(buf->descs[i]); kfree(buf->sg[i]); buf->descs[i] = NULL; buf->sg[i] = NULL; buf->sg_len[i] = 0; buf->plane_sizes[i] = 0; } buf->nb_planes = 0; } static int pxa_buffer_init(struct pxa_camera_dev *pcdev, struct pxa_buffer *buf) { struct vb2_buffer *vb = &buf->vbuf.vb2_buf; struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); int nb_channels = pcdev->channels; int i, ret = 0; unsigned long size = vb2_plane_size(vb, 0); switch (nb_channels) { case 1: buf->plane_sizes[0] = size; break; case 3: buf->plane_sizes[0] = size / 2; buf->plane_sizes[1] = size / 4; buf->plane_sizes[2] = size / 4; break; default: return -EINVAL; } buf->nb_planes = nb_channels; ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels, buf->plane_sizes, buf->sg, buf->sg_len, GFP_KERNEL); if (ret < 0) { dev_err(pcdev_to_dev(pcdev), "sg_split failed: %d\n", ret); return ret; } for (i = 0; i < nb_channels; i++) { ret = pxa_init_dma_channel(pcdev, buf, i, buf->sg[i], buf->sg_len[i]); if (ret) { pxa_buffer_cleanup(buf); return ret; } } INIT_LIST_HEAD(&buf->queue); return ret; } static void pxac_vb2_cleanup(struct vb2_buffer *vb) { struct pxa_buffer *buf = vb2_to_pxa_buffer(vb); struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue); dev_dbg(pcdev_to_dev(pcdev), "%s(vb=%p)\n", __func__, vb); pxa_buffer_cleanup(buf); } static void pxac_vb2_queue(struct vb2_buffer *vb) { struct pxa_buffer *buf = vb2_to_pxa_buffer(vb); struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue); dev_dbg(pcdev_to_dev(pcdev), "%s(vb=%p) nb_channels=%d size=%lu active=%p\n", __func__, vb, pcdev->channels, vb2_get_plane_payload(vb, 0), pcdev->active); list_add_tail(&buf->queue, &pcdev->capture); pxa_dma_add_tail_buf(pcdev, buf); } /* * Please check the DMA prepared buffer structure in : * Documentation/driver-api/media/drivers/pxa_camera.rst * Please check also in pxa_camera_check_link_miss() to understand why DMA chain * modification while DMA chain is running will work anyway. */ static int pxac_vb2_prepare(struct vb2_buffer *vb) { struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue); struct pxa_buffer *buf = vb2_to_pxa_buffer(vb); int ret = 0; #ifdef DEBUG int i; #endif switch (pcdev->channels) { case 1: case 3: vb2_set_plane_payload(vb, 0, pcdev->current_pix.sizeimage); break; default: return -EINVAL; } dev_dbg(pcdev_to_dev(pcdev), "%s (vb=%p) nb_channels=%d size=%lu\n", __func__, vb, pcdev->channels, vb2_get_plane_payload(vb, 0)); WARN_ON(!pcdev->current_fmt); #ifdef DEBUG /* * This can be useful if you want to see if we actually fill * the buffer with something */ for (i = 0; i < vb->num_planes; i++) memset((void *)vb2_plane_vaddr(vb, i), 0xaa, vb2_get_plane_payload(vb, i)); #endif /* * I think, in buf_prepare you only have to protect global data, * the actual buffer is yours */ buf->inwork = 0; pxa_videobuf_set_actdma(pcdev, buf); return ret; } static int pxac_vb2_init(struct vb2_buffer *vb) { struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vb->vb2_queue); struct pxa_buffer *buf = vb2_to_pxa_buffer(vb); dev_dbg(pcdev_to_dev(pcdev), "%s(nb_channels=%d)\n", __func__, pcdev->channels); return pxa_buffer_init(pcdev, buf); } static int pxac_vb2_queue_setup(struct vb2_queue *vq, unsigned int *nbufs, unsigned int *num_planes, unsigned int sizes[], struct device *alloc_devs[]) { struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq); int size = pcdev->current_pix.sizeimage; dev_dbg(pcdev_to_dev(pcdev), "%s(vq=%p nbufs=%d num_planes=%d size=%d)\n", __func__, vq, *nbufs, *num_planes, size); /* * Called from VIDIOC_REQBUFS or in compatibility mode For YUV422P * format, even if there are 3 planes Y, U and V, we reply there is only * one plane, containing Y, U and V data, one after the other. */ if (*num_planes) return sizes[0] < size ? -EINVAL : 0; *num_planes = 1; switch (pcdev->channels) { case 1: case 3: sizes[0] = size; break; default: return -EINVAL; } if (!*nbufs) *nbufs = 1; return 0; } static int pxac_vb2_start_streaming(struct vb2_queue *vq, unsigned int count) { struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq); dev_dbg(pcdev_to_dev(pcdev), "%s(count=%d) active=%p\n", __func__, count, pcdev->active); pcdev->buf_sequence = 0; if (!pcdev->active) pxa_camera_start_capture(pcdev); return 0; } static void pxac_vb2_stop_streaming(struct vb2_queue *vq) { struct pxa_camera_dev *pcdev = vb2_get_drv_priv(vq); struct pxa_buffer *buf, *tmp; dev_dbg(pcdev_to_dev(pcdev), "%s active=%p\n", __func__, pcdev->active); pxa_camera_stop_capture(pcdev); list_for_each_entry_safe(buf, tmp, &pcdev->capture, queue) pxa_camera_wakeup(pcdev, buf, VB2_BUF_STATE_ERROR); } static const struct vb2_ops pxac_vb2_ops = { .queue_setup = pxac_vb2_queue_setup, .buf_init = pxac_vb2_init, .buf_prepare = pxac_vb2_prepare, .buf_queue = pxac_vb2_queue, .buf_cleanup = pxac_vb2_cleanup, .start_streaming = pxac_vb2_start_streaming, .stop_streaming = pxac_vb2_stop_streaming, .wait_prepare = vb2_ops_wait_prepare, .wait_finish = vb2_ops_wait_finish, }; static int pxa_camera_init_videobuf2(struct pxa_camera_dev *pcdev) { int ret; struct vb2_queue *vq = &pcdev->vb2_vq; memset(vq, 0, sizeof(*vq)); vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE; vq->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF; vq->drv_priv = pcdev; vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; vq->buf_struct_size = sizeof(struct pxa_buffer); vq->dev = pcdev->v4l2_dev.dev; vq->ops = &pxac_vb2_ops; vq->mem_ops = &vb2_dma_sg_memops; vq->lock = &pcdev->mlock; ret = vb2_queue_init(vq); dev_dbg(pcdev_to_dev(pcdev), "vb2_queue_init(vq=%p): %d\n", vq, ret); return ret; } /* * Video ioctls section */ static int pxa_camera_set_bus_param(struct pxa_camera_dev *pcdev) { unsigned int bus_width = pcdev->current_fmt->host_fmt->bits_per_sample; struct v4l2_mbus_config cfg = {.type = V4L2_MBUS_PARALLEL,}; u32 pixfmt = pcdev->current_fmt->host_fmt->fourcc; int mbus_config; int ret; if (!((1 << (bus_width - 1)) & pcdev->width_flags)) { dev_err(pcdev_to_dev(pcdev), "Unsupported bus width %u", bus_width); return -EINVAL; } pcdev->channels = 1; /* Make choices, based on platform preferences */ mbus_config = 0; if (pcdev->platform_flags & PXA_CAMERA_MASTER) mbus_config |= V4L2_MBUS_MASTER; else mbus_config |= V4L2_MBUS_SLAVE; if (pcdev->platform_flags & PXA_CAMERA_HSP) mbus_config |= V4L2_MBUS_HSYNC_ACTIVE_HIGH; else mbus_config |= V4L2_MBUS_HSYNC_ACTIVE_LOW; if (pcdev->platform_flags & PXA_CAMERA_VSP) mbus_config |= V4L2_MBUS_VSYNC_ACTIVE_HIGH; else mbus_config |= V4L2_MBUS_VSYNC_ACTIVE_LOW; if (pcdev->platform_flags & PXA_CAMERA_PCP) mbus_config |= V4L2_MBUS_PCLK_SAMPLE_RISING; else mbus_config |= V4L2_MBUS_PCLK_SAMPLE_FALLING; mbus_config |= V4L2_MBUS_DATA_ACTIVE_HIGH; cfg.flags = mbus_config; ret = sensor_call(pcdev, pad, set_mbus_config, 0, &cfg); if (ret < 0 && ret != -ENOIOCTLCMD) { dev_err(pcdev_to_dev(pcdev), "Failed to call set_mbus_config: %d\n", ret); return ret; } /* * If the requested media bus configuration has not been fully applied * make sure it is supported by the platform. * * PXA does not support V4L2_MBUS_DATA_ACTIVE_LOW and the bus mastering * roles should match. */ if (cfg.flags != mbus_config) { unsigned int pxa_mbus_role = mbus_config & (V4L2_MBUS_MASTER | V4L2_MBUS_SLAVE); if (pxa_mbus_role != (cfg.flags & (V4L2_MBUS_MASTER | V4L2_MBUS_SLAVE))) { dev_err(pcdev_to_dev(pcdev), "Unsupported mbus configuration: bus mastering\n"); return -EINVAL; } if (cfg.flags & V4L2_MBUS_DATA_ACTIVE_LOW) { dev_err(pcdev_to_dev(pcdev), "Unsupported mbus configuration: DATA_ACTIVE_LOW\n"); return -EINVAL; } } pxa_camera_setup_cicr(pcdev, cfg.flags, pixfmt); return 0; } static const struct pxa_mbus_pixelfmt pxa_camera_formats[] = { { .fourcc = V4L2_PIX_FMT_YUV422P, .name = "Planar YUV422 16 bit", .bits_per_sample = 8, .packing = PXA_MBUS_PACKING_2X8_PADHI, .order = PXA_MBUS_ORDER_LE, .layout = PXA_MBUS_LAYOUT_PLANAR_2Y_U_V, }, }; /* This will be corrected as we get more formats */ static bool pxa_camera_packing_supported(const struct pxa_mbus_pixelfmt *fmt) { return fmt->packing == PXA_MBUS_PACKING_NONE || (fmt->bits_per_sample == 8 && fmt->packing == PXA_MBUS_PACKING_2X8_PADHI) || (fmt->bits_per_sample > 8 && fmt->packing == PXA_MBUS_PACKING_EXTEND16); } static int pxa_camera_get_formats(struct v4l2_device *v4l2_dev, unsigned int idx, struct pxa_camera_format_xlate *xlate) { struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(v4l2_dev); int formats = 0, ret; struct v4l2_subdev_mbus_code_enum code = { .which = V4L2_SUBDEV_FORMAT_ACTIVE, .index = idx, }; const struct pxa_mbus_pixelfmt *fmt; ret = sensor_call(pcdev, pad, enum_mbus_code, NULL, &code); if (ret < 0) /* No more formats */ return 0; fmt = pxa_mbus_get_fmtdesc(code.code); if (!fmt) { dev_err(pcdev_to_dev(pcdev), "Invalid format code #%u: %d\n", idx, code.code); return 0; } switch (code.code) { case MEDIA_BUS_FMT_UYVY8_2X8: formats++; if (xlate) { xlate->host_fmt = &pxa_camera_formats[0]; xlate->code = code.code; xlate++; dev_dbg(pcdev_to_dev(pcdev), "Providing format %s using code %d\n", pxa_camera_formats[0].name, code.code); } /* fall through */ case MEDIA_BUS_FMT_VYUY8_2X8: case MEDIA_BUS_FMT_YUYV8_2X8: case MEDIA_BUS_FMT_YVYU8_2X8: case MEDIA_BUS_FMT_RGB565_2X8_LE: case MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE: if (xlate) dev_dbg(pcdev_to_dev(pcdev), "Providing format %s packed\n", fmt->name); break; default: if (!pxa_camera_packing_supported(fmt)) return 0; if (xlate) dev_dbg(pcdev_to_dev(pcdev), "Providing format %s in pass-through mode\n", fmt->name); break; } /* Generic pass-through */ formats++; if (xlate) { xlate->host_fmt = fmt; xlate->code = code.code; xlate++; } return formats; } static int pxa_camera_build_formats(struct pxa_camera_dev *pcdev) { struct pxa_camera_format_xlate *xlate; xlate = pxa_mbus_build_fmts_xlate(&pcdev->v4l2_dev, pcdev->sensor, pxa_camera_get_formats); if (IS_ERR(xlate)) return PTR_ERR(xlate); pcdev->user_formats = xlate; return 0; } static void pxa_camera_destroy_formats(struct pxa_camera_dev *pcdev) { kfree(pcdev->user_formats); } static int pxa_camera_check_frame(u32 width, u32 height) { /* limit to pxa hardware capabilities */ return height < 32 || height > 2048 || width < 48 || width > 2048 || (width & 0x01); } #ifdef CONFIG_VIDEO_ADV_DEBUG static int pxac_vidioc_g_register(struct file *file, void *priv, struct v4l2_dbg_register *reg) { struct pxa_camera_dev *pcdev = video_drvdata(file); if (reg->reg > CIBR2) return -ERANGE; reg->val = __raw_readl(pcdev->base + reg->reg); reg->size = sizeof(__u32); return 0; } static int pxac_vidioc_s_register(struct file *file, void *priv, const struct v4l2_dbg_register *reg) { struct pxa_camera_dev *pcdev = video_drvdata(file); if (reg->reg > CIBR2) return -ERANGE; if (reg->size != sizeof(__u32)) return -EINVAL; __raw_writel(reg->val, pcdev->base + reg->reg); return 0; } #endif static int pxac_vidioc_enum_fmt_vid_cap(struct file *filp, void *priv, struct v4l2_fmtdesc *f) { struct pxa_camera_dev *pcdev = video_drvdata(filp); const struct pxa_mbus_pixelfmt *format; unsigned int idx; for (idx = 0; pcdev->user_formats[idx].code; idx++); if (f->index >= idx) return -EINVAL; format = pcdev->user_formats[f->index].host_fmt; f->pixelformat = format->fourcc; return 0; } static int pxac_vidioc_g_fmt_vid_cap(struct file *filp, void *priv, struct v4l2_format *f) { struct pxa_camera_dev *pcdev = video_drvdata(filp); struct v4l2_pix_format *pix = &f->fmt.pix; pix->width = pcdev->current_pix.width; pix->height = pcdev->current_pix.height; pix->bytesperline = pcdev->current_pix.bytesperline; pix->sizeimage = pcdev->current_pix.sizeimage; pix->field = pcdev->current_pix.field; pix->pixelformat = pcdev->current_fmt->host_fmt->fourcc; pix->colorspace = pcdev->current_pix.colorspace; dev_dbg(pcdev_to_dev(pcdev), "current_fmt->fourcc: 0x%08x\n", pcdev->current_fmt->host_fmt->fourcc); return 0; } static int pxac_vidioc_try_fmt_vid_cap(struct file *filp, void *priv, struct v4l2_format *f) { struct pxa_camera_dev *pcdev = video_drvdata(filp); const struct pxa_camera_format_xlate *xlate; struct v4l2_pix_format *pix = &f->fmt.pix; struct v4l2_subdev_pad_config pad_cfg; struct v4l2_subdev_format format = { .which = V4L2_SUBDEV_FORMAT_TRY, }; struct v4l2_mbus_framefmt *mf = &format.format; __u32 pixfmt = pix->pixelformat; int ret; xlate = pxa_mbus_xlate_by_fourcc(pcdev->user_formats, pixfmt); if (!xlate) { dev_warn(pcdev_to_dev(pcdev), "Format %x not found\n", pixfmt); return -EINVAL; } /* * Limit to pxa hardware capabilities. YUV422P planar format requires * images size to be a multiple of 16 bytes. If not, zeros will be * inserted between Y and U planes, and U and V planes, which violates * the YUV422P standard. */ v4l_bound_align_image(&pix->width, 48, 2048, 1, &pix->height, 32, 2048, 0, pixfmt == V4L2_PIX_FMT_YUV422P ? 4 : 0); v4l2_fill_mbus_format(mf, pix, xlate->code); ret = sensor_call(pcdev, pad, set_fmt, &pad_cfg, &format); if (ret < 0) return ret; v4l2_fill_pix_format(pix, mf); /* Only progressive video supported so far */ switch (mf->field) { case V4L2_FIELD_ANY: case V4L2_FIELD_NONE: pix->field = V4L2_FIELD_NONE; break; default: /* TODO: support interlaced at least in pass-through mode */ dev_err(pcdev_to_dev(pcdev), "Field type %d unsupported.\n", mf->field); return -EINVAL; } ret = pxa_mbus_bytes_per_line(pix->width, xlate->host_fmt); if (ret < 0) return ret; pix->bytesperline = ret; ret = pxa_mbus_image_size(xlate->host_fmt, pix->bytesperline, pix->height); if (ret < 0) return ret; pix->sizeimage = ret; return 0; } static int pxac_vidioc_s_fmt_vid_cap(struct file *filp, void *priv, struct v4l2_format *f) { struct pxa_camera_dev *pcdev = video_drvdata(filp); const struct pxa_camera_format_xlate *xlate; struct v4l2_pix_format *pix = &f->fmt.pix; struct v4l2_subdev_format format = { .which = V4L2_SUBDEV_FORMAT_ACTIVE, }; unsigned long flags; int ret, is_busy; dev_dbg(pcdev_to_dev(pcdev), "s_fmt_vid_cap(pix=%dx%d:%x)\n", pix->width, pix->height, pix->pixelformat); spin_lock_irqsave(&pcdev->lock, flags); is_busy = pcdev->active || vb2_is_busy(&pcdev->vb2_vq); spin_unlock_irqrestore(&pcdev->lock, flags); if (is_busy) return -EBUSY; ret = pxac_vidioc_try_fmt_vid_cap(filp, priv, f); if (ret) return ret; xlate = pxa_mbus_xlate_by_fourcc(pcdev->user_formats, pix->pixelformat); v4l2_fill_mbus_format(&format.format, pix, xlate->code); ret = sensor_call(pcdev, pad, set_fmt, NULL, &format); if (ret < 0) { dev_warn(pcdev_to_dev(pcdev), "Failed to configure for format %x\n", pix->pixelformat); } else if (pxa_camera_check_frame(pix->width, pix->height)) { dev_warn(pcdev_to_dev(pcdev), "Camera driver produced an unsupported frame %dx%d\n", pix->width, pix->height); return -EINVAL; } pcdev->current_fmt = xlate; pcdev->current_pix = *pix; ret = pxa_camera_set_bus_param(pcdev); return ret; } static int pxac_vidioc_querycap(struct file *file, void *priv, struct v4l2_capability *cap) { strscpy(cap->bus_info, "platform:pxa-camera", sizeof(cap->bus_info)); strscpy(cap->driver, PXA_CAM_DRV_NAME, sizeof(cap->driver)); strscpy(cap->card, pxa_cam_driver_description, sizeof(cap->card)); return 0; } static int pxac_vidioc_enum_input(struct file *file, void *priv, struct v4l2_input *i) { if (i->index > 0) return -EINVAL; i->type = V4L2_INPUT_TYPE_CAMERA; strscpy(i->name, "Camera", sizeof(i->name)); return 0; } static int pxac_vidioc_g_input(struct file *file, void *priv, unsigned int *i) { *i = 0; return 0; } static int pxac_vidioc_s_input(struct file *file, void *priv, unsigned int i) { if (i > 0) return -EINVAL; return 0; } static int pxac_sensor_set_power(struct pxa_camera_dev *pcdev, int on) { int ret; ret = sensor_call(pcdev, core, s_power, on); if (ret == -ENOIOCTLCMD) ret = 0; if (ret) { dev_warn(pcdev_to_dev(pcdev), "Failed to put subdevice in %s mode: %d\n", on ? "normal operation" : "power saving", ret); } return ret; } static int pxac_fops_camera_open(struct file *filp) { struct pxa_camera_dev *pcdev = video_drvdata(filp); int ret; mutex_lock(&pcdev->mlock); ret = v4l2_fh_open(filp); if (ret < 0) goto out; if (!v4l2_fh_is_singular_file(filp)) goto out; ret = pxac_sensor_set_power(pcdev, 1); if (ret) v4l2_fh_release(filp); out: mutex_unlock(&pcdev->mlock); return ret; } static int pxac_fops_camera_release(struct file *filp) { struct pxa_camera_dev *pcdev = video_drvdata(filp); int ret; bool fh_singular; mutex_lock(&pcdev->mlock); fh_singular = v4l2_fh_is_singular_file(filp); ret = _vb2_fop_release(filp, NULL); if (fh_singular) ret = pxac_sensor_set_power(pcdev, 0); mutex_unlock(&pcdev->mlock); return ret; } static const struct v4l2_file_operations pxa_camera_fops = { .owner = THIS_MODULE, .open = pxac_fops_camera_open, .release = pxac_fops_camera_release, .read = vb2_fop_read, .poll = vb2_fop_poll, .mmap = vb2_fop_mmap, .unlocked_ioctl = video_ioctl2, }; static const struct v4l2_ioctl_ops pxa_camera_ioctl_ops = { .vidioc_querycap = pxac_vidioc_querycap, .vidioc_enum_input = pxac_vidioc_enum_input, .vidioc_g_input = pxac_vidioc_g_input, .vidioc_s_input = pxac_vidioc_s_input, .vidioc_enum_fmt_vid_cap = pxac_vidioc_enum_fmt_vid_cap, .vidioc_g_fmt_vid_cap = pxac_vidioc_g_fmt_vid_cap, .vidioc_s_fmt_vid_cap = pxac_vidioc_s_fmt_vid_cap, .vidioc_try_fmt_vid_cap = pxac_vidioc_try_fmt_vid_cap, .vidioc_reqbufs = vb2_ioctl_reqbufs, .vidioc_create_bufs = vb2_ioctl_create_bufs, .vidioc_querybuf = vb2_ioctl_querybuf, .vidioc_qbuf = vb2_ioctl_qbuf, .vidioc_dqbuf = vb2_ioctl_dqbuf, .vidioc_expbuf = vb2_ioctl_expbuf, .vidioc_streamon = vb2_ioctl_streamon, .vidioc_streamoff = vb2_ioctl_streamoff, #ifdef CONFIG_VIDEO_ADV_DEBUG .vidioc_g_register = pxac_vidioc_g_register, .vidioc_s_register = pxac_vidioc_s_register, #endif .vidioc_subscribe_event = v4l2_ctrl_subscribe_event, .vidioc_unsubscribe_event = v4l2_event_unsubscribe, }; static const struct v4l2_clk_ops pxa_camera_mclk_ops = { }; static const struct video_device pxa_camera_videodev_template = { .name = "pxa-camera", .minor = -1, .fops = &pxa_camera_fops, .ioctl_ops = &pxa_camera_ioctl_ops, .release = video_device_release_empty, .device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING, }; static int pxa_camera_sensor_bound(struct v4l2_async_notifier *notifier, struct v4l2_subdev *subdev, struct v4l2_async_subdev *asd) { int err; struct v4l2_device *v4l2_dev = notifier->v4l2_dev; struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(v4l2_dev); struct video_device *vdev = &pcdev->vdev; struct v4l2_pix_format *pix = &pcdev->current_pix; struct v4l2_subdev_format format = { .which = V4L2_SUBDEV_FORMAT_ACTIVE, }; struct v4l2_mbus_framefmt *mf = &format.format; dev_info(pcdev_to_dev(pcdev), "%s(): trying to bind a device\n", __func__); mutex_lock(&pcdev->mlock); *vdev = pxa_camera_videodev_template; vdev->v4l2_dev = v4l2_dev; vdev->lock = &pcdev->mlock; pcdev->sensor = subdev; pcdev->vdev.queue = &pcdev->vb2_vq; pcdev->vdev.v4l2_dev = &pcdev->v4l2_dev; pcdev->vdev.ctrl_handler = subdev->ctrl_handler; video_set_drvdata(&pcdev->vdev, pcdev); err = pxa_camera_build_formats(pcdev); if (err) { dev_err(pcdev_to_dev(pcdev), "building formats failed: %d\n", err); goto out; } pcdev->current_fmt = pcdev->user_formats; pix->field = V4L2_FIELD_NONE; pix->width = DEFAULT_WIDTH; pix->height = DEFAULT_HEIGHT; pix->bytesperline = pxa_mbus_bytes_per_line(pix->width, pcdev->current_fmt->host_fmt); pix->sizeimage = pxa_mbus_image_size(pcdev->current_fmt->host_fmt, pix->bytesperline, pix->height); pix->pixelformat = pcdev->current_fmt->host_fmt->fourcc; v4l2_fill_mbus_format(mf, pix, pcdev->current_fmt->code); err = pxac_sensor_set_power(pcdev, 1); if (err) goto out; err = sensor_call(pcdev, pad, set_fmt, NULL, &format); if (err) goto out_sensor_poweroff; v4l2_fill_pix_format(pix, mf); pr_info("%s(): colorspace=0x%x pixfmt=0x%x\n", __func__, pix->colorspace, pix->pixelformat); err = pxa_camera_init_videobuf2(pcdev); if (err) goto out_sensor_poweroff; err = video_register_device(&pcdev->vdev, VFL_TYPE_VIDEO, -1); if (err) { v4l2_err(v4l2_dev, "register video device failed: %d\n", err); pcdev->sensor = NULL; } else { dev_info(pcdev_to_dev(pcdev), "PXA Camera driver attached to camera %s\n", subdev->name); } out_sensor_poweroff: err = pxac_sensor_set_power(pcdev, 0); out: mutex_unlock(&pcdev->mlock); return err; } static void pxa_camera_sensor_unbind(struct v4l2_async_notifier *notifier, struct v4l2_subdev *subdev, struct v4l2_async_subdev *asd) { struct pxa_camera_dev *pcdev = v4l2_dev_to_pcdev(notifier->v4l2_dev); mutex_lock(&pcdev->mlock); dev_info(pcdev_to_dev(pcdev), "PXA Camera driver detached from camera %s\n", subdev->name); /* disable capture, disable interrupts */ __raw_writel(0x3ff, pcdev->base + CICR0); /* Stop DMA engine */ pxa_dma_stop_channels(pcdev); pxa_camera_destroy_formats(pcdev); if (pcdev->mclk_clk) { v4l2_clk_unregister(pcdev->mclk_clk); pcdev->mclk_clk = NULL; } video_unregister_device(&pcdev->vdev); pcdev->sensor = NULL; mutex_unlock(&pcdev->mlock); } static const struct v4l2_async_notifier_operations pxa_camera_sensor_ops = { .bound = pxa_camera_sensor_bound, .unbind = pxa_camera_sensor_unbind, }; /* * Driver probe, remove, suspend and resume operations */ static int pxa_camera_suspend(struct device *dev) { struct pxa_camera_dev *pcdev = dev_get_drvdata(dev); int i = 0, ret = 0; pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR0); pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR1); pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR2); pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR3); pcdev->save_cicr[i++] = __raw_readl(pcdev->base + CICR4); if (pcdev->sensor) ret = pxac_sensor_set_power(pcdev, 0); return ret; } static int pxa_camera_resume(struct device *dev) { struct pxa_camera_dev *pcdev = dev_get_drvdata(dev); int i = 0, ret = 0; __raw_writel(pcdev->save_cicr[i++] & ~CICR0_ENB, pcdev->base + CICR0); __raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR1); __raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR2); __raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR3); __raw_writel(pcdev->save_cicr[i++], pcdev->base + CICR4); if (pcdev->sensor) { ret = pxac_sensor_set_power(pcdev, 1); } /* Restart frame capture if active buffer exists */ if (!ret && pcdev->active) pxa_camera_start_capture(pcdev); return ret; } static int pxa_camera_pdata_from_dt(struct device *dev, struct pxa_camera_dev *pcdev, struct v4l2_async_subdev *asd) { u32 mclk_rate; struct device_node *remote, *np = dev->of_node; struct v4l2_fwnode_endpoint ep = { .bus_type = 0 }; int err = of_property_read_u32(np, "clock-frequency", &mclk_rate); if (!err) { pcdev->platform_flags |= PXA_CAMERA_MCLK_EN; pcdev->mclk = mclk_rate; } np = of_graph_get_next_endpoint(np, NULL); if (!np) { dev_err(dev, "could not find endpoint\n"); return -EINVAL; } err = v4l2_fwnode_endpoint_parse(of_fwnode_handle(np), &ep); if (err) { dev_err(dev, "could not parse endpoint\n"); goto out; } switch (ep.bus.parallel.bus_width) { case 4: pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_4; break; case 5: pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_5; break; case 8: pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_8; break; case 9: pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_9; break; case 10: pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_10; break; default: break; } if (ep.bus.parallel.flags & V4L2_MBUS_MASTER) pcdev->platform_flags |= PXA_CAMERA_MASTER; if (ep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH) pcdev->platform_flags |= PXA_CAMERA_HSP; if (ep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH) pcdev->platform_flags |= PXA_CAMERA_VSP; if (ep.bus.parallel.flags & V4L2_MBUS_PCLK_SAMPLE_RISING) pcdev->platform_flags |= PXA_CAMERA_PCLK_EN | PXA_CAMERA_PCP; if (ep.bus.parallel.flags & V4L2_MBUS_PCLK_SAMPLE_FALLING) pcdev->platform_flags |= PXA_CAMERA_PCLK_EN; asd->match_type = V4L2_ASYNC_MATCH_FWNODE; remote = of_graph_get_remote_port_parent(np); if (remote) asd->match.fwnode = of_fwnode_handle(remote); else dev_notice(dev, "no remote for %pOF\n", np); out: of_node_put(np); return err; } static int pxa_camera_probe(struct platform_device *pdev) { struct pxa_camera_dev *pcdev; struct resource *res; void __iomem *base; struct dma_slave_config config = { .src_addr_width = 0, .src_maxburst = 8, .direction = DMA_DEV_TO_MEM, }; char clk_name[V4L2_CLK_NAME_SIZE]; int irq; int err = 0, i; res = platform_get_resource(pdev, IORESOURCE_MEM, 0); irq = platform_get_irq(pdev, 0); if (!res || irq < 0) return -ENODEV; pcdev = devm_kzalloc(&pdev->dev, sizeof(*pcdev), GFP_KERNEL); if (!pcdev) { dev_err(&pdev->dev, "Could not allocate pcdev\n"); return -ENOMEM; } pcdev->clk = devm_clk_get(&pdev->dev, NULL); if (IS_ERR(pcdev->clk)) return PTR_ERR(pcdev->clk); pcdev->res = res; pcdev->pdata = pdev->dev.platform_data; if (pcdev->pdata) { pcdev->platform_flags = pcdev->pdata->flags; pcdev->mclk = pcdev->pdata->mclk_10khz * 10000; pcdev->asd.match_type = V4L2_ASYNC_MATCH_I2C; pcdev->asd.match.i2c.adapter_id = pcdev->pdata->sensor_i2c_adapter_id; pcdev->asd.match.i2c.address = pcdev->pdata->sensor_i2c_address; } else if (pdev->dev.of_node) { err = pxa_camera_pdata_from_dt(&pdev->dev, pcdev, &pcdev->asd); } else { return -ENODEV; } if (err < 0) return err; if (!(pcdev->platform_flags & (PXA_CAMERA_DATAWIDTH_8 | PXA_CAMERA_DATAWIDTH_9 | PXA_CAMERA_DATAWIDTH_10))) { /* * Platform hasn't set available data widths. This is bad. * Warn and use a default. */ dev_warn(&pdev->dev, "WARNING! Platform hasn't set available data widths, using default 10 bit\n"); pcdev->platform_flags |= PXA_CAMERA_DATAWIDTH_10; } if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_8) pcdev->width_flags = 1 << 7; if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_9) pcdev->width_flags |= 1 << 8; if (pcdev->platform_flags & PXA_CAMERA_DATAWIDTH_10) pcdev->width_flags |= 1 << 9; if (!pcdev->mclk) { dev_warn(&pdev->dev, "mclk == 0! Please, fix your platform data. Using default 20MHz\n"); pcdev->mclk = 20000000; } pcdev->mclk_divisor = mclk_get_divisor(pdev, pcdev); INIT_LIST_HEAD(&pcdev->capture); spin_lock_init(&pcdev->lock); mutex_init(&pcdev->mlock); /* * Request the regions. */ base = devm_ioremap_resource(&pdev->dev, res); if (IS_ERR(base)) return PTR_ERR(base); pcdev->irq = irq; pcdev->base = base; /* request dma */ pcdev->dma_chans[0] = dma_request_chan(&pdev->dev, "CI_Y"); if (IS_ERR(pcdev->dma_chans[0])) { dev_err(&pdev->dev, "Can't request DMA for Y\n"); return PTR_ERR(pcdev->dma_chans[0]); } pcdev->dma_chans[1] = dma_request_chan(&pdev->dev, "CI_U"); if (IS_ERR(pcdev->dma_chans[1])) { dev_err(&pdev->dev, "Can't request DMA for U\n"); err = PTR_ERR(pcdev->dma_chans[1]); goto exit_free_dma_y; } pcdev->dma_chans[2] = dma_request_chan(&pdev->dev, "CI_V"); if (IS_ERR(pcdev->dma_chans[2])) { dev_err(&pdev->dev, "Can't request DMA for V\n"); err = PTR_ERR(pcdev->dma_chans[2]); goto exit_free_dma_u; } for (i = 0; i < 3; i++) { config.src_addr = pcdev->res->start + CIBR0 + i * 8; err = dmaengine_slave_config(pcdev->dma_chans[i], &config); if (err < 0) { dev_err(&pdev->dev, "dma slave config failed: %d\n", err); goto exit_free_dma; } } /* request irq */ err = devm_request_irq(&pdev->dev, pcdev->irq, pxa_camera_irq, 0, PXA_CAM_DRV_NAME, pcdev); if (err) { dev_err(&pdev->dev, "Camera interrupt register failed\n"); goto exit_free_dma; } tasklet_setup(&pcdev->task_eof, pxa_camera_eof); pxa_camera_activate(pcdev); dev_set_drvdata(&pdev->dev, pcdev); err = v4l2_device_register(&pdev->dev, &pcdev->v4l2_dev); if (err) goto exit_deactivate; v4l2_async_notifier_init(&pcdev->notifier); err = v4l2_async_notifier_add_subdev(&pcdev->notifier, &pcdev->asd); if (err) { fwnode_handle_put(pcdev->asd.match.fwnode); goto exit_free_v4l2dev; } pcdev->notifier.ops = &pxa_camera_sensor_ops; if (!of_have_populated_dt()) pcdev->asd.match_type = V4L2_ASYNC_MATCH_I2C; err = pxa_camera_init_videobuf2(pcdev); if (err) goto exit_notifier_cleanup; v4l2_clk_name_i2c(clk_name, sizeof(clk_name), pcdev->asd.match.i2c.adapter_id, pcdev->asd.match.i2c.address); pcdev->mclk_clk = v4l2_clk_register(&pxa_camera_mclk_ops, clk_name, NULL); if (IS_ERR(pcdev->mclk_clk)) { err = PTR_ERR(pcdev->mclk_clk); goto exit_notifier_cleanup; } err = v4l2_async_notifier_register(&pcdev->v4l2_dev, &pcdev->notifier); if (err) goto exit_free_clk; return 0; exit_free_clk: v4l2_clk_unregister(pcdev->mclk_clk); exit_notifier_cleanup: v4l2_async_notifier_cleanup(&pcdev->notifier); exit_free_v4l2dev: v4l2_device_unregister(&pcdev->v4l2_dev); exit_deactivate: pxa_camera_deactivate(pcdev); tasklet_kill(&pcdev->task_eof); exit_free_dma: dma_release_channel(pcdev->dma_chans[2]); exit_free_dma_u: dma_release_channel(pcdev->dma_chans[1]); exit_free_dma_y: dma_release_channel(pcdev->dma_chans[0]); return err; } static int pxa_camera_remove(struct platform_device *pdev) { struct pxa_camera_dev *pcdev = dev_get_drvdata(&pdev->dev); pxa_camera_deactivate(pcdev); tasklet_kill(&pcdev->task_eof); dma_release_channel(pcdev->dma_chans[0]); dma_release_channel(pcdev->dma_chans[1]); dma_release_channel(pcdev->dma_chans[2]); v4l2_async_notifier_unregister(&pcdev->notifier); v4l2_async_notifier_cleanup(&pcdev->notifier); if (pcdev->mclk_clk) { v4l2_clk_unregister(pcdev->mclk_clk); pcdev->mclk_clk = NULL; } v4l2_device_unregister(&pcdev->v4l2_dev); dev_info(&pdev->dev, "PXA Camera driver unloaded\n"); return 0; } static const struct dev_pm_ops pxa_camera_pm = { .suspend = pxa_camera_suspend, .resume = pxa_camera_resume, }; static const struct of_device_id pxa_camera_of_match[] = { { .compatible = "marvell,pxa270-qci", }, {}, }; MODULE_DEVICE_TABLE(of, pxa_camera_of_match); static struct platform_driver pxa_camera_driver = { .driver = { .name = PXA_CAM_DRV_NAME, .pm = &pxa_camera_pm, .of_match_table = of_match_ptr(pxa_camera_of_match), }, .probe = pxa_camera_probe, .remove = pxa_camera_remove, }; module_platform_driver(pxa_camera_driver); MODULE_DESCRIPTION("PXA27x Camera Driver"); MODULE_AUTHOR("Guennadi Liakhovetski "); MODULE_LICENSE("GPL"); MODULE_VERSION(PXA_CAM_VERSION); MODULE_ALIAS("platform:" PXA_CAM_DRV_NAME);