/* * Rockchip isp1 driver * * Copyright (C) 2017 Rockchip Electronics Co., Ltd. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include #include #include #include #include /* for ISP statistics */ #include "dev.h" #include "regs.h" #define RKISP1_ISP_STATS_REQ_BUFS_MIN 2 #define RKISP1_ISP_STATS_REQ_BUFS_MAX 8 static int rkisp1_stats_enum_fmt_meta_cap(struct file *file, void *priv, struct v4l2_fmtdesc *f) { struct video_device *video = video_devdata(file); struct rkisp1_isp_stats_vdev *stats_vdev = video_get_drvdata(video); if (f->index > 0 || f->type != video->queue->type) return -EINVAL; f->pixelformat = stats_vdev->vdev_fmt.fmt.meta.dataformat; return 0; } static int rkisp1_stats_g_fmt_meta_cap(struct file *file, void *priv, struct v4l2_format *f) { struct video_device *video = video_devdata(file); struct rkisp1_isp_stats_vdev *stats_vdev = video_get_drvdata(video); struct v4l2_meta_format *meta = &f->fmt.meta; if (f->type != video->queue->type) return -EINVAL; memset(meta, 0, sizeof(*meta)); meta->dataformat = stats_vdev->vdev_fmt.fmt.meta.dataformat; meta->buffersize = stats_vdev->vdev_fmt.fmt.meta.buffersize; return 0; } static int rkisp1_stats_querycap(struct file *file, void *priv, struct v4l2_capability *cap) { struct video_device *vdev = video_devdata(file); struct rkisp1_isp_stats_vdev *stats_vdev = video_get_drvdata(vdev); strcpy(cap->driver, DRIVER_NAME); snprintf(cap->driver, sizeof(cap->driver), "%s_v%d", DRIVER_NAME, stats_vdev->dev->isp_ver >> 4); strlcpy(cap->card, vdev->name, sizeof(cap->card)); strlcpy(cap->bus_info, "platform: " DRIVER_NAME, sizeof(cap->bus_info)); return 0; } /* ISP video device IOCTLs */ static const struct v4l2_ioctl_ops rkisp1_stats_ioctl = { .vidioc_reqbufs = vb2_ioctl_reqbufs, .vidioc_querybuf = vb2_ioctl_querybuf, .vidioc_create_bufs = vb2_ioctl_create_bufs, .vidioc_qbuf = vb2_ioctl_qbuf, .vidioc_dqbuf = vb2_ioctl_dqbuf, .vidioc_prepare_buf = vb2_ioctl_prepare_buf, .vidioc_expbuf = vb2_ioctl_expbuf, .vidioc_streamon = vb2_ioctl_streamon, .vidioc_streamoff = vb2_ioctl_streamoff, .vidioc_enum_fmt_meta_cap = rkisp1_stats_enum_fmt_meta_cap, .vidioc_g_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_s_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_try_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_querycap = rkisp1_stats_querycap }; struct v4l2_file_operations rkisp1_stats_fops = { .mmap = vb2_fop_mmap, .unlocked_ioctl = video_ioctl2, .poll = vb2_fop_poll, .open = v4l2_fh_open, .release = vb2_fop_release }; static int rkisp1_stats_vb2_queue_setup(struct vb2_queue *vq, unsigned int *num_buffers, unsigned int *num_planes, unsigned int sizes[], struct device *alloc_ctxs[]) { struct rkisp1_isp_stats_vdev *stats_vdev = vq->drv_priv; *num_planes = 1; *num_buffers = clamp_t(u32, *num_buffers, RKISP1_ISP_STATS_REQ_BUFS_MIN, RKISP1_ISP_STATS_REQ_BUFS_MAX); sizes[0] = sizeof(struct rkisp1_stat_buffer); INIT_LIST_HEAD(&stats_vdev->stat); return 0; } static void rkisp1_stats_vb2_buf_queue(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); struct rkisp1_buffer *stats_buf = to_rkisp1_buffer(vbuf); struct vb2_queue *vq = vb->vb2_queue; struct rkisp1_isp_stats_vdev *stats_dev = vq->drv_priv; stats_buf->vaddr[0] = vb2_plane_vaddr(vb, 0); spin_lock_bh(&stats_dev->rd_lock); list_add_tail(&stats_buf->queue, &stats_dev->stat); spin_unlock_bh(&stats_dev->rd_lock); } static void rkisp1_stats_vb2_stop_streaming(struct vb2_queue *vq) { struct rkisp1_isp_stats_vdev *stats_vdev = vq->drv_priv; struct rkisp1_buffer *buf; unsigned long flags; int i; /* Make sure no new work queued in isr before draining wq */ spin_lock_irqsave(&stats_vdev->irq_lock, flags); stats_vdev->streamon = false; spin_unlock_irqrestore(&stats_vdev->irq_lock, flags); tasklet_disable(&stats_vdev->rd_tasklet); spin_lock_bh(&stats_vdev->rd_lock); for (i = 0; i < RKISP1_ISP_STATS_REQ_BUFS_MAX; i++) { if (list_empty(&stats_vdev->stat)) break; buf = list_first_entry(&stats_vdev->stat, struct rkisp1_buffer, queue); list_del(&buf->queue); vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR); } spin_unlock_bh(&stats_vdev->rd_lock); } static int rkisp1_stats_vb2_start_streaming(struct vb2_queue *queue, unsigned int count) { struct rkisp1_isp_stats_vdev *stats_vdev = queue->drv_priv; stats_vdev->streamon = true; kfifo_reset(&stats_vdev->rd_kfifo); tasklet_enable(&stats_vdev->rd_tasklet); return 0; } static struct vb2_ops rkisp1_stats_vb2_ops = { .queue_setup = rkisp1_stats_vb2_queue_setup, .buf_queue = rkisp1_stats_vb2_buf_queue, .wait_prepare = vb2_ops_wait_prepare, .wait_finish = vb2_ops_wait_finish, .stop_streaming = rkisp1_stats_vb2_stop_streaming, .start_streaming = rkisp1_stats_vb2_start_streaming, }; static int rkisp1_stats_init_vb2_queue(struct vb2_queue *q, struct rkisp1_isp_stats_vdev *stats_vdev) { q->type = V4L2_BUF_TYPE_META_CAPTURE; q->io_modes = VB2_MMAP | VB2_USERPTR; q->drv_priv = stats_vdev; q->ops = &rkisp1_stats_vb2_ops; q->mem_ops = &vb2_vmalloc_memops; q->buf_struct_size = sizeof(struct rkisp1_buffer); q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; q->lock = &stats_vdev->dev->iqlock; q->dev = stats_vdev->dev->dev; return vb2_queue_init(q); } static void rkisp1_stats_get_awb_meas_v10(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { /* Protect against concurrent access from ISR? */ u32 reg_val; pbuf->meas_type |= CIFISP_STAT_AWB; reg_val = readl(stats_vdev->dev->base_addr + CIF_ISP_AWB_WHITE_CNT_V10); pbuf->params.awb.awb_mean[0].cnt = CIF_ISP_AWB_GET_PIXEL_CNT(reg_val); reg_val = readl(stats_vdev->dev->base_addr + CIF_ISP_AWB_MEAN_V10); pbuf->params.awb.awb_mean[0].mean_cr_or_r = CIF_ISP_AWB_GET_MEAN_CR_R(reg_val); pbuf->params.awb.awb_mean[0].mean_cb_or_b = CIF_ISP_AWB_GET_MEAN_CB_B(reg_val); pbuf->params.awb.awb_mean[0].mean_y_or_g = CIF_ISP_AWB_GET_MEAN_Y_G(reg_val); } static void rkisp1_stats_get_awb_meas_v12(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { /* Protect against concurrent access from ISR? */ u32 reg_val; pbuf->meas_type |= CIFISP_STAT_AWB; reg_val = readl(stats_vdev->dev->base_addr + CIF_ISP_AWB_WHITE_CNT_V12); pbuf->params.awb.awb_mean[0].cnt = CIF_ISP_AWB_GET_PIXEL_CNT(reg_val); reg_val = readl(stats_vdev->dev->base_addr + CIF_ISP_AWB_MEAN_V12); pbuf->params.awb.awb_mean[0].mean_cr_or_r = CIF_ISP_AWB_GET_MEAN_CR_R(reg_val); pbuf->params.awb.awb_mean[0].mean_cb_or_b = CIF_ISP_AWB_GET_MEAN_CB_B(reg_val); pbuf->params.awb.awb_mean[0].mean_y_or_g = CIF_ISP_AWB_GET_MEAN_Y_G(reg_val); } static void rkisp1_stats_get_aec_meas_v10(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { unsigned int i; void __iomem *addr = stats_vdev->dev->base_addr + CIF_ISP_EXP_MEAN_00_V10; pbuf->meas_type |= CIFISP_STAT_AUTOEXP; for (i = 0; i < stats_vdev->config->ae_mean_max; i++) pbuf->params.ae.exp_mean[i] = (u8)readl(addr + i * 4); } static void rkisp1_stats_get_aec_meas_v12(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { int i; void __iomem *addr = stats_vdev->dev->base_addr + CIF_ISP_EXP_MEAN_V12; u32 value; pbuf->meas_type |= CIFISP_STAT_AUTOEXP; for (i = 0; i < stats_vdev->config->ae_mean_max / 4; i++) { value = readl(addr + i * 4); pbuf->params.ae.exp_mean[4 * i + 0] = CIF_ISP_EXP_GET_MEAN_xy0_V12(value); pbuf->params.ae.exp_mean[4 * i + 1] = CIF_ISP_EXP_GET_MEAN_xy1_V12(value); pbuf->params.ae.exp_mean[4 * i + 2] = CIF_ISP_EXP_GET_MEAN_xy2_V12(value); pbuf->params.ae.exp_mean[4 * i + 3] = CIF_ISP_EXP_GET_MEAN_xy3_V12(value); } value = readl(addr + i * 4); pbuf->params.ae.exp_mean[4 * i + 0] = CIF_ISP_EXP_GET_MEAN_xy0_V12(value); } static void rkisp1_stats_get_afc_meas(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { void __iomem *base_addr; struct cifisp_af_stat *af; pbuf->meas_type |= CIFISP_STAT_AFM_FIN; af = &pbuf->params.af; base_addr = stats_vdev->dev->base_addr; af->window[0].sum = readl(base_addr + CIF_ISP_AFM_SUM_A); af->window[0].lum = readl(base_addr + CIF_ISP_AFM_LUM_A); af->window[1].sum = readl(base_addr + CIF_ISP_AFM_SUM_B); af->window[1].lum = readl(base_addr + CIF_ISP_AFM_LUM_B); af->window[2].sum = readl(base_addr + CIF_ISP_AFM_SUM_C); af->window[2].lum = readl(base_addr + CIF_ISP_AFM_LUM_C); } static void rkisp1_stats_get_hst_meas_v10(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { int i; void __iomem *addr = stats_vdev->dev->base_addr + CIF_ISP_HIST_BIN_0_V10; pbuf->meas_type |= CIFISP_STAT_HIST; for (i = 0; i < stats_vdev->config->hist_bin_n_max; i++) pbuf->params.hist.hist_bins[i] = readl(addr + (i * 4)); } static void rkisp1_stats_get_hst_meas_v12(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { int i; void __iomem *addr = stats_vdev->dev->base_addr + CIF_ISP_HIST_BIN_V12; u32 value; pbuf->meas_type |= CIFISP_STAT_HIST; for (i = 0; i < stats_vdev->config->hist_bin_n_max / 2; i++) { value = readl(addr + (i * 4)); pbuf->params.hist.hist_bins[2 * i] = CIF_ISP_HIST_GET_BIN0_V12(value); pbuf->params.hist.hist_bins[2 * i + 1] = CIF_ISP_HIST_GET_BIN1_V12(value); } } static void rkisp1_stats_get_bls_meas(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *dev = stats_vdev->dev; const struct ispsd_in_fmt *in_fmt = rkisp1_get_ispsd_in_fmt(&dev->isp_sdev); void __iomem *base = stats_vdev->dev->base_addr; struct cifisp_bls_meas_val *bls_val; bls_val = &pbuf->params.ae.bls_val; if (in_fmt->bayer_pat == RAW_BGGR) { bls_val->meas_b = readl(base + CIF_ISP_BLS_A_MEASURED); bls_val->meas_gb = readl(base + CIF_ISP_BLS_B_MEASURED); bls_val->meas_gr = readl(base + CIF_ISP_BLS_C_MEASURED); bls_val->meas_r = readl(base + CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RAW_GBRG) { bls_val->meas_gb = readl(base + CIF_ISP_BLS_A_MEASURED); bls_val->meas_b = readl(base + CIF_ISP_BLS_B_MEASURED); bls_val->meas_r = readl(base + CIF_ISP_BLS_C_MEASURED); bls_val->meas_gr = readl(base + CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RAW_GRBG) { bls_val->meas_gr = readl(base + CIF_ISP_BLS_A_MEASURED); bls_val->meas_r = readl(base + CIF_ISP_BLS_B_MEASURED); bls_val->meas_b = readl(base + CIF_ISP_BLS_C_MEASURED); bls_val->meas_gb = readl(base + CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RAW_RGGB) { bls_val->meas_r = readl(base + CIF_ISP_BLS_A_MEASURED); bls_val->meas_gr = readl(base + CIF_ISP_BLS_B_MEASURED); bls_val->meas_gb = readl(base + CIF_ISP_BLS_C_MEASURED); bls_val->meas_b = readl(base + CIF_ISP_BLS_D_MEASURED); } } static void rkisp1_stats_get_emb_data(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_stat_buffer *pbuf) { unsigned int i; struct rkisp1_device *dev = stats_vdev->dev; unsigned int ph = 0, out = 0, packet_len = 0, playload_len = 0; unsigned int mipi_kfifo_len; unsigned int idx; unsigned char *fifo_data; idx = RKISP1_EMDDATA_FIFO_MAX; for (i = 0; i < RKISP1_EMDDATA_FIFO_MAX; i++) { if (dev->emd_data_fifo[i].frame_id == pbuf->frame_id) { idx = i; break; } } if (idx == RKISP1_EMDDATA_FIFO_MAX) return; if (kfifo_is_empty(&dev->emd_data_fifo[idx].mipi_kfifo)) return; mipi_kfifo_len = dev->emd_data_fifo[idx].data_len; fifo_data = &pbuf->params.emd.data[0]; for (i = 0; i < mipi_kfifo_len;) { /* handle the package header */ out = kfifo_out(&dev->emd_data_fifo[idx].mipi_kfifo, &ph, sizeof(ph)); if (!out) break; packet_len = (ph >> 8) & 0xfff; i += sizeof(ph); /* handle the package data */ out = kfifo_out(&dev->emd_data_fifo[idx].mipi_kfifo, fifo_data, packet_len); if (!out) break; i += packet_len; playload_len += packet_len; fifo_data += packet_len; v4l2_dbg(1, rkisp1_debug, &dev->v4l2_dev, "packet_len: 0x%x, ph: 0x%x\n", packet_len, ph); } pbuf->meas_type |= CIFISP_STAT_EMB_DATA; v4l2_dbg(1, rkisp1_debug, &dev->v4l2_dev, "playload_len: %d, pbuf->frame_id %d\n", playload_len, pbuf->frame_id); } static struct rkisp1_stats_ops rkisp1_v10_stats_ops = { .get_awb_meas = rkisp1_stats_get_awb_meas_v10, .get_aec_meas = rkisp1_stats_get_aec_meas_v10, .get_afc_meas = rkisp1_stats_get_afc_meas, .get_hst_meas = rkisp1_stats_get_hst_meas_v10, .get_bls_meas = rkisp1_stats_get_bls_meas, .get_emb_data = rkisp1_stats_get_emb_data, }; static struct rkisp1_stats_ops rkisp1_v12_stats_ops = { .get_awb_meas = rkisp1_stats_get_awb_meas_v12, .get_aec_meas = rkisp1_stats_get_aec_meas_v12, .get_afc_meas = rkisp1_stats_get_afc_meas, .get_hst_meas = rkisp1_stats_get_hst_meas_v12, .get_bls_meas = rkisp1_stats_get_bls_meas, }; static struct rkisp1_stats_config rkisp1_v10_stats_config = { .ae_mean_max = 25, .hist_bin_n_max = 16, }; static struct rkisp1_stats_config rkisp1_v12_stats_config = { .ae_mean_max = 81, .hist_bin_n_max = 32, }; static void rkisp1_stats_send_measurement(struct rkisp1_isp_stats_vdev *stats_vdev, struct rkisp1_isp_readout_work *meas_work) { unsigned int cur_frame_id = -1; struct rkisp1_stat_buffer *cur_stat_buf; struct rkisp1_buffer *cur_buf = NULL; struct rkisp1_stats_ops *ops = stats_vdev->ops; cur_frame_id = atomic_read(&stats_vdev->dev->isp_sdev.frm_sync_seq) - 1; if (cur_frame_id != meas_work->frame_id) { v4l2_warn(stats_vdev->vnode.vdev.v4l2_dev, "Measurement late(%d, %d)\n", cur_frame_id, meas_work->frame_id); cur_frame_id = meas_work->frame_id; } spin_lock(&stats_vdev->rd_lock); /* get one empty buffer */ if (!list_empty(&stats_vdev->stat)) { cur_buf = list_first_entry(&stats_vdev->stat, struct rkisp1_buffer, queue); list_del(&cur_buf->queue); } spin_unlock(&stats_vdev->rd_lock); if (!cur_buf) return; cur_stat_buf = (struct rkisp1_stat_buffer *)(cur_buf->vaddr[0]); memset(cur_stat_buf, 0, sizeof(*cur_stat_buf)); cur_stat_buf->frame_id = cur_frame_id; if (meas_work->isp_ris & CIF_ISP_AWB_DONE) { ops->get_awb_meas(stats_vdev, cur_stat_buf); cur_stat_buf->meas_type |= CIFISP_STAT_AWB; } if (meas_work->isp_ris & CIF_ISP_AFM_FIN) { ops->get_afc_meas(stats_vdev, cur_stat_buf); cur_stat_buf->meas_type |= CIFISP_STAT_AFM_FIN; } if (meas_work->isp_ris & CIF_ISP_EXP_END) { ops->get_aec_meas(stats_vdev, cur_stat_buf); ops->get_bls_meas(stats_vdev, cur_stat_buf); cur_stat_buf->meas_type |= CIFISP_STAT_AUTOEXP; } if (meas_work->isp_ris & CIF_ISP_HIST_MEASURE_RDY) { ops->get_hst_meas(stats_vdev, cur_stat_buf); cur_stat_buf->meas_type |= CIFISP_STAT_HIST; } if ((meas_work->isp_ris & CIF_ISP_FRAME) && ops->get_emb_data) ops->get_emb_data(stats_vdev, cur_stat_buf); vb2_set_plane_payload(&cur_buf->vb.vb2_buf, 0, sizeof(struct rkisp1_stat_buffer)); cur_buf->vb.sequence = cur_frame_id; cur_buf->vb.vb2_buf.timestamp = meas_work->timestamp; vb2_buffer_done(&cur_buf->vb.vb2_buf, VB2_BUF_STATE_DONE); } static void rkisp1_stats_readout_task(unsigned long data) { unsigned int out = 0; struct rkisp1_isp_readout_work work; struct rkisp1_isp_stats_vdev *vdev = (struct rkisp1_isp_stats_vdev *)data; while (!kfifo_is_empty(&vdev->rd_kfifo)) { out = kfifo_out(&vdev->rd_kfifo, &work, sizeof(work)); if (!out) break; if (work.readout == RKISP1_ISP_READOUT_MEAS) rkisp1_stats_send_measurement(vdev, &work); } } int rkisp1_stats_isr(struct rkisp1_isp_stats_vdev *stats_vdev, u32 isp_ris) { unsigned int isp_mis_tmp = 0; struct rkisp1_isp_readout_work work; unsigned int cur_frame_id = atomic_read(&stats_vdev->dev->isp_sdev.frm_sync_seq) - 1; #ifdef LOG_ISR_EXE_TIME ktime_t in_t = ktime_get(); #endif spin_lock(&stats_vdev->irq_lock); isp_mis_tmp = isp_ris & (CIF_ISP_AWB_DONE | CIF_ISP_AFM_FIN | CIF_ISP_EXP_END | CIF_ISP_HIST_MEASURE_RDY); if (isp_mis_tmp) { writel(isp_mis_tmp, stats_vdev->dev->base_addr + CIF_ISP_ICR); isp_mis_tmp &= readl(stats_vdev->dev->base_addr + CIF_ISP_MIS); if (isp_mis_tmp) v4l2_err(stats_vdev->vnode.vdev.v4l2_dev, "isp icr 3A info err: 0x%x 0x%x\n", isp_mis_tmp, isp_ris); } if (!stats_vdev->streamon) goto unlock; if (isp_ris & (CIF_ISP_FRAME | CIF_ISP_AWB_DONE | CIF_ISP_AFM_FIN | CIF_ISP_EXP_END | CIF_ISP_HIST_MEASURE_RDY)) { work.readout = RKISP1_ISP_READOUT_MEAS; work.frame_id = cur_frame_id; work.isp_ris = isp_ris; work.timestamp = ktime_get_ns(); if (!kfifo_is_full(&stats_vdev->rd_kfifo)) kfifo_in(&stats_vdev->rd_kfifo, &work, sizeof(work)); else v4l2_err(stats_vdev->vnode.vdev.v4l2_dev, "stats kfifo is full\n"); tasklet_schedule(&stats_vdev->rd_tasklet); } #ifdef LOG_ISR_EXE_TIME if (isp_ris & (CIF_ISP_EXP_END | CIF_ISP_AWB_DONE | CIF_ISP_FRAME | CIF_ISP_HIST_MEASURE_RDY)) { unsigned int diff_us = ktime_to_us(ktime_sub(ktime_get(), in_t)); if (diff_us > g_longest_isr_time) g_longest_isr_time = diff_us; v4l2_info(stats_vdev->vnode.vdev.v4l2_dev, "isp_isr time %d %d\n", diff_us, g_longest_isr_time); } #endif unlock: spin_unlock(&stats_vdev->irq_lock); return 0; } static void rkisp1_init_stats_vdev(struct rkisp1_isp_stats_vdev *stats_vdev) { stats_vdev->vdev_fmt.fmt.meta.dataformat = V4L2_META_FMT_RK_ISP1_STAT_3A; stats_vdev->vdev_fmt.fmt.meta.buffersize = sizeof(struct rkisp1_stat_buffer); if (stats_vdev->dev->isp_ver == ISP_V12 || stats_vdev->dev->isp_ver == ISP_V13) { stats_vdev->ops = &rkisp1_v12_stats_ops; stats_vdev->config = &rkisp1_v12_stats_config; } else { stats_vdev->ops = &rkisp1_v10_stats_ops; stats_vdev->config = &rkisp1_v10_stats_config; } } int rkisp1_register_stats_vdev(struct rkisp1_isp_stats_vdev *stats_vdev, struct v4l2_device *v4l2_dev, struct rkisp1_device *dev) { int ret; struct rkisp1_vdev_node *node = &stats_vdev->vnode; struct video_device *vdev = &node->vdev; stats_vdev->dev = dev; INIT_LIST_HEAD(&stats_vdev->stat); spin_lock_init(&stats_vdev->irq_lock); spin_lock_init(&stats_vdev->rd_lock); strlcpy(vdev->name, "rkisp1-statistics", sizeof(vdev->name)); video_set_drvdata(vdev, stats_vdev); vdev->ioctl_ops = &rkisp1_stats_ioctl; vdev->fops = &rkisp1_stats_fops; vdev->release = video_device_release_empty; vdev->lock = &dev->iqlock; vdev->v4l2_dev = v4l2_dev; vdev->queue = &node->buf_queue; vdev->device_caps = V4L2_CAP_META_CAPTURE | V4L2_CAP_STREAMING; vdev->vfl_dir = VFL_DIR_RX; rkisp1_stats_init_vb2_queue(vdev->queue, stats_vdev); rkisp1_init_stats_vdev(stats_vdev); video_set_drvdata(vdev, stats_vdev); node->pad.flags = MEDIA_PAD_FL_SINK; ret = media_entity_pads_init(&vdev->entity, 1, &node->pad); if (ret < 0) goto err_release_queue; ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1); if (ret < 0) { dev_err(&vdev->dev, "could not register Video for Linux device\n"); goto err_cleanup_media_entity; } ret = kfifo_alloc(&stats_vdev->rd_kfifo, RKISP1_READOUT_WORK_SIZE, GFP_KERNEL); if (ret) { dev_err(&vdev->dev, "kfifo_alloc failed with error %d\n", ret); goto err_cleanup_media_entity; } tasklet_init(&stats_vdev->rd_tasklet, rkisp1_stats_readout_task, (unsigned long)stats_vdev); tasklet_disable(&stats_vdev->rd_tasklet); return 0; err_cleanup_media_entity: media_entity_cleanup(&vdev->entity); err_release_queue: vb2_queue_release(vdev->queue); return ret; } void rkisp1_unregister_stats_vdev(struct rkisp1_isp_stats_vdev *stats_vdev) { struct rkisp1_vdev_node *node = &stats_vdev->vnode; struct video_device *vdev = &node->vdev; kfifo_free(&stats_vdev->rd_kfifo); tasklet_kill(&stats_vdev->rd_tasklet); video_unregister_device(vdev); media_entity_cleanup(&vdev->entity); vb2_queue_release(vdev->queue); }