1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (c) 2019 Fuzhou Rockchip Electronics Co., Ltd. */
3 
4 #include <linux/kfifo.h>
5 #include <media/v4l2-common.h>
6 #include <media/v4l2-ioctl.h>
7 #include <media/videobuf2-core.h>
8 #include <media/videobuf2-vmalloc.h>	/* for ISP statistics */
9 #include <media/videobuf2-dma-contig.h>
10 #include <media/videobuf2-dma-sg.h>
11 #include <media/v4l2-mc.h>
12 #include <uapi/linux/rk-video-format.h>
13 #include "dev.h"
14 #include "regs.h"
15 #include "stats.h"
16 
17 #define RKISPP_STATS_REQ_BUFS_MIN 2
18 #define RKISPP_STATS_REQ_BUFS_MAX 8
19 
update_addr(struct rkispp_stats_vdev *stats_vdev)20 static void update_addr(struct rkispp_stats_vdev *stats_vdev)
21 {
22 	struct rkispp_dummy_buffer *dummy_buf;
23 	u32 addr;
24 
25 	if (stats_vdev->next_buf) {
26 		addr = stats_vdev->next_buf->buff_addr[0];
27 		rkispp_write(stats_vdev->dev, RKISPP_ORB_WR_BASE, addr);
28 	}
29 
30 	if (!stats_vdev->next_buf) {
31 		dummy_buf = &stats_vdev->dev->hw_dev->dummy_buf;
32 		if (!dummy_buf->mem_priv)
33 			return;
34 
35 		rkispp_write(stats_vdev->dev, RKISPP_ORB_WR_BASE, dummy_buf->dma_addr);
36 	}
37 }
38 
rkispp_stats_frame_end(struct rkispp_stats_vdev *stats_vdev)39 static int rkispp_stats_frame_end(struct rkispp_stats_vdev *stats_vdev)
40 {
41 	void __iomem *base = stats_vdev->dev->hw_dev->base_addr;
42 	struct rkispp_device *dev = stats_vdev->dev;
43 	struct rkispp_buffer *curr_buf;
44 	struct rkispp_stats_buffer *cur_stat_buf;
45 	unsigned long lock_flags = 0;
46 
47 	if (stats_vdev->curr_buf) {
48 		u64 ns = ktime_get_ns();
49 		u32 total_num = readl(base + RKISPP_ORB_TOTAL_NUM);
50 		u32 cur_frame_id = dev->ispp_sdev.frm_sync_seq;
51 		void *vaddr;
52 
53 		curr_buf = stats_vdev->curr_buf;
54 		vaddr = vb2_plane_vaddr(&curr_buf->vb.vb2_buf, 0);
55 		cur_stat_buf = (struct rkispp_stats_buffer *)vaddr;
56 
57 		cur_stat_buf->total_num = total_num;
58 		cur_stat_buf->meas_type = ISPP_MODULE_ORB;
59 		cur_stat_buf->frame_id = cur_frame_id;
60 		curr_buf->vb.vb2_buf.timestamp = ns;
61 		curr_buf->vb.sequence = cur_frame_id;
62 		vb2_set_plane_payload(&curr_buf->vb.vb2_buf, 0,
63 					      sizeof(struct rkispp_stats_buffer));
64 
65 		vb2_buffer_done(&curr_buf->vb.vb2_buf,
66 				VB2_BUF_STATE_DONE);
67 		stats_vdev->curr_buf = NULL;
68 	}
69 
70 	stats_vdev->curr_buf = stats_vdev->next_buf;
71 	stats_vdev->next_buf = NULL;
72 	spin_lock_irqsave(&stats_vdev->irq_lock, lock_flags);
73 	if (!list_empty(&stats_vdev->stat)) {
74 		stats_vdev->next_buf = list_first_entry(&stats_vdev->stat,
75 					struct rkispp_buffer, queue);
76 		list_del(&stats_vdev->next_buf->queue);
77 	}
78 	spin_unlock_irqrestore(&stats_vdev->irq_lock, lock_flags);
79 
80 	update_addr(stats_vdev);
81 	return 0;
82 }
83 
rkispp_stats_enum_fmt_meta_cap(struct file *file, void *priv, struct v4l2_fmtdesc *f)84 static int rkispp_stats_enum_fmt_meta_cap(struct file *file, void *priv,
85 					  struct v4l2_fmtdesc *f)
86 {
87 	struct video_device *video = video_devdata(file);
88 	struct rkispp_stats_vdev *stats_vdev = video_get_drvdata(video);
89 
90 	if (f->index > 0 || f->type != video->queue->type)
91 		return -EINVAL;
92 
93 	f->pixelformat = stats_vdev->vdev_fmt.fmt.meta.dataformat;
94 	return 0;
95 }
96 
rkispp_stats_g_fmt_meta_cap(struct file *file, void *priv, struct v4l2_format *f)97 static int rkispp_stats_g_fmt_meta_cap(struct file *file, void *priv,
98 				       struct v4l2_format *f)
99 {
100 	struct video_device *video = video_devdata(file);
101 	struct rkispp_stats_vdev *stats_vdev = video_get_drvdata(video);
102 	struct v4l2_meta_format *meta = &f->fmt.meta;
103 
104 	if (f->type != video->queue->type)
105 		return -EINVAL;
106 
107 	memset(meta, 0, sizeof(*meta));
108 	meta->dataformat = stats_vdev->vdev_fmt.fmt.meta.dataformat;
109 	meta->buffersize = stats_vdev->vdev_fmt.fmt.meta.buffersize;
110 
111 	return 0;
112 }
113 
rkispp_stats_querycap(struct file *file, void *priv, struct v4l2_capability *cap)114 static int rkispp_stats_querycap(struct file *file,
115 				 void *priv, struct v4l2_capability *cap)
116 {
117 	struct video_device *vdev = video_devdata(file);
118 	struct rkispp_stats_vdev *stats_vdev = video_get_drvdata(vdev);
119 
120 	strcpy(cap->driver, DRIVER_NAME);
121 	snprintf(cap->driver, sizeof(cap->driver),
122 		 "%s_v%d", DRIVER_NAME,
123 		 stats_vdev->dev->ispp_ver >> 4);
124 	strlcpy(cap->card, vdev->name, sizeof(cap->card));
125 	strlcpy(cap->bus_info, "platform: " DRIVER_NAME, sizeof(cap->bus_info));
126 
127 	return 0;
128 }
129 
rkispp_stats_fh_open(struct file *filp)130 static int rkispp_stats_fh_open(struct file *filp)
131 {
132 	struct rkispp_stats_vdev *stats = video_drvdata(filp);
133 	struct rkispp_device *isppdev = stats->dev;
134 	int ret;
135 
136 	ret = v4l2_fh_open(filp);
137 	if (!ret) {
138 		ret = v4l2_pipeline_pm_get(&stats->vnode.vdev.entity);
139 		if (ret < 0) {
140 			v4l2_err(&isppdev->v4l2_dev,
141 				 "pipeline power on failed %d\n", ret);
142 			vb2_fop_release(filp);
143 		}
144 	}
145 	return ret;
146 }
147 
rkispp_stats_fh_release(struct file *filp)148 static int rkispp_stats_fh_release(struct file *filp)
149 {
150 	struct rkispp_stats_vdev *stats = video_drvdata(filp);
151 	int ret;
152 
153 	ret = vb2_fop_release(filp);
154 	if (!ret)
155 		v4l2_pipeline_pm_put(&stats->vnode.vdev.entity);
156 	return ret;
157 }
158 
159 /* ISP video device IOCTLs */
160 static const struct v4l2_ioctl_ops rkispp_stats_ioctl = {
161 	.vidioc_reqbufs = vb2_ioctl_reqbufs,
162 	.vidioc_querybuf = vb2_ioctl_querybuf,
163 	.vidioc_create_bufs = vb2_ioctl_create_bufs,
164 	.vidioc_qbuf = vb2_ioctl_qbuf,
165 	.vidioc_dqbuf = vb2_ioctl_dqbuf,
166 	.vidioc_prepare_buf = vb2_ioctl_prepare_buf,
167 	.vidioc_expbuf = vb2_ioctl_expbuf,
168 	.vidioc_streamon = vb2_ioctl_streamon,
169 	.vidioc_streamoff = vb2_ioctl_streamoff,
170 	.vidioc_enum_fmt_meta_cap = rkispp_stats_enum_fmt_meta_cap,
171 	.vidioc_g_fmt_meta_cap = rkispp_stats_g_fmt_meta_cap,
172 	.vidioc_s_fmt_meta_cap = rkispp_stats_g_fmt_meta_cap,
173 	.vidioc_try_fmt_meta_cap = rkispp_stats_g_fmt_meta_cap,
174 	.vidioc_querycap = rkispp_stats_querycap
175 };
176 
177 struct v4l2_file_operations rkispp_stats_fops = {
178 	.mmap = vb2_fop_mmap,
179 	.unlocked_ioctl = video_ioctl2,
180 	.poll = vb2_fop_poll,
181 	.open = rkispp_stats_fh_open,
182 	.release = rkispp_stats_fh_release,
183 };
184 
rkispp_stats_vb2_queue_setup(struct vb2_queue *vq, unsigned int *num_buffers, unsigned int *num_planes, unsigned int sizes[], struct device *alloc_ctxs[])185 static int rkispp_stats_vb2_queue_setup(struct vb2_queue *vq,
186 					unsigned int *num_buffers,
187 					unsigned int *num_planes,
188 					unsigned int sizes[],
189 					struct device *alloc_ctxs[])
190 {
191 	struct rkispp_stats_vdev *stats_vdev = vq->drv_priv;
192 
193 	*num_planes = 1;
194 
195 	*num_buffers = clamp_t(u32, *num_buffers, RKISPP_STATS_REQ_BUFS_MIN,
196 			       RKISPP_STATS_REQ_BUFS_MAX);
197 
198 	sizes[0] = sizeof(struct rkispp_stats_buffer);
199 
200 	INIT_LIST_HEAD(&stats_vdev->stat);
201 
202 	return 0;
203 }
204 
rkispp_stats_vb2_buf_queue(struct vb2_buffer *vb)205 static void rkispp_stats_vb2_buf_queue(struct vb2_buffer *vb)
206 {
207 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
208 	struct rkispp_buffer *buf = to_rkispp_buffer(vbuf);
209 	struct vb2_queue *vq = vb->vb2_queue;
210 	struct rkispp_stats_vdev *stats_dev = vq->drv_priv;
211 	unsigned long lock_flags = 0;
212 
213 	vb2_plane_vaddr(vb, 0);
214 	if (stats_dev->dev->hw_dev->is_mmu) {
215 		struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
216 
217 		buf->buff_addr[0] = sg_dma_address(sgt->sgl);
218 	} else {
219 		buf->buff_addr[0] = vb2_dma_contig_plane_dma_addr(vb, 0);
220 	}
221 	spin_lock_irqsave(&stats_dev->irq_lock, lock_flags);
222 	if (stats_dev->streamon &&
223 	    !stats_dev->next_buf) {
224 		stats_dev->next_buf = buf;
225 		update_addr(stats_dev);
226 	} else {
227 		list_add_tail(&buf->queue, &stats_dev->stat);
228 	}
229 	spin_unlock_irqrestore(&stats_dev->irq_lock, lock_flags);
230 }
231 
destroy_buf_queue(struct rkispp_stats_vdev *stats_vdev, enum vb2_buffer_state state)232 static void destroy_buf_queue(struct rkispp_stats_vdev *stats_vdev,
233 			      enum vb2_buffer_state state)
234 {
235 	struct rkispp_buffer *buf;
236 
237 	if (stats_vdev->curr_buf) {
238 		list_add_tail(&stats_vdev->curr_buf->queue, &stats_vdev->stat);
239 		stats_vdev->curr_buf = NULL;
240 	}
241 	if (stats_vdev->next_buf) {
242 		list_add_tail(&stats_vdev->next_buf->queue, &stats_vdev->stat);
243 		stats_vdev->next_buf = NULL;
244 	}
245 	while (!list_empty(&stats_vdev->stat)) {
246 		buf = list_first_entry(&stats_vdev->stat,
247 			struct rkispp_buffer, queue);
248 		list_del(&buf->queue);
249 		vb2_buffer_done(&buf->vb.vb2_buf, state);
250 	}
251 }
252 
rkispp_stats_vb2_stop_streaming(struct vb2_queue *vq)253 static void rkispp_stats_vb2_stop_streaming(struct vb2_queue *vq)
254 {
255 	struct rkispp_stats_vdev *stats_vdev = vq->drv_priv;
256 	unsigned long flags;
257 
258 	spin_lock_irqsave(&stats_vdev->irq_lock, flags);
259 	stats_vdev->streamon = false;
260 	destroy_buf_queue(stats_vdev, VB2_BUF_STATE_ERROR);
261 	spin_unlock_irqrestore(&stats_vdev->irq_lock, flags);
262 }
263 
264 static int
rkispp_stats_vb2_start_streaming(struct vb2_queue *queue, unsigned int count)265 rkispp_stats_vb2_start_streaming(struct vb2_queue *queue,
266 				 unsigned int count)
267 {
268 	struct rkispp_stats_vdev *stats_vdev = queue->drv_priv;
269 	unsigned long flags;
270 
271 	if (stats_vdev->streamon)
272 		return -EBUSY;
273 
274 	/* config first buf */
275 	rkispp_stats_frame_end(stats_vdev);
276 
277 	spin_lock_irqsave(&stats_vdev->irq_lock, flags);
278 	stats_vdev->streamon = true;
279 	spin_unlock_irqrestore(&stats_vdev->irq_lock, flags);
280 
281 	return 0;
282 }
283 
284 static struct vb2_ops rkispp_stats_vb2_ops = {
285 	.queue_setup = rkispp_stats_vb2_queue_setup,
286 	.buf_queue = rkispp_stats_vb2_buf_queue,
287 	.wait_prepare = vb2_ops_wait_prepare,
288 	.wait_finish = vb2_ops_wait_finish,
289 	.stop_streaming = rkispp_stats_vb2_stop_streaming,
290 	.start_streaming = rkispp_stats_vb2_start_streaming,
291 };
292 
rkispp_stats_init_vb2_queue(struct vb2_queue *q, struct rkispp_stats_vdev *stats_vdev)293 static int rkispp_stats_init_vb2_queue(struct vb2_queue *q,
294 				       struct rkispp_stats_vdev *stats_vdev)
295 {
296 	q->type = V4L2_BUF_TYPE_META_CAPTURE;
297 	q->io_modes = VB2_MMAP | VB2_USERPTR;
298 	q->drv_priv = stats_vdev;
299 	q->ops = &rkispp_stats_vb2_ops;
300 	q->mem_ops = stats_vdev->dev->hw_dev->mem_ops;
301 	q->buf_struct_size = sizeof(struct rkispp_buffer);
302 	q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
303 	q->lock = &stats_vdev->dev->iqlock;
304 	q->dev = stats_vdev->dev->hw_dev->dev;
305 	if (stats_vdev->dev->hw_dev->is_dma_contig)
306 		q->dma_attrs = DMA_ATTR_FORCE_CONTIGUOUS;
307 	q->gfp_flags = GFP_DMA32;
308 	return vb2_queue_init(q);
309 }
310 
rkispp_stats_isr(struct rkispp_stats_vdev *stats_vdev, u32 mis)311 void rkispp_stats_isr(struct rkispp_stats_vdev *stats_vdev, u32 mis)
312 {
313 	spin_lock(&stats_vdev->irq_lock);
314 	if (!stats_vdev->streamon) {
315 		spin_unlock(&stats_vdev->irq_lock);
316 		return;
317 	}
318 	spin_unlock(&stats_vdev->irq_lock);
319 
320 	if (mis & ORB_INT)
321 		rkispp_stats_frame_end(stats_vdev);
322 }
323 
rkispp_init_stats_vdev(struct rkispp_stats_vdev *stats_vdev)324 static void rkispp_init_stats_vdev(struct rkispp_stats_vdev *stats_vdev)
325 {
326 	stats_vdev->vdev_fmt.fmt.meta.dataformat =
327 		V4L2_META_FMT_RK_ISPP_STAT;
328 	stats_vdev->vdev_fmt.fmt.meta.buffersize =
329 		sizeof(struct rkispp_stats_buffer);
330 }
331 
rkispp_register_stats_vdev(struct rkispp_device *dev)332 int rkispp_register_stats_vdev(struct rkispp_device *dev)
333 {
334 	struct rkispp_stats_vdev *stats_vdev = &dev->stats_vdev;
335 	struct rkispp_vdev_node *node = &stats_vdev->vnode;
336 	struct video_device *vdev = &node->vdev;
337 	int ret;
338 
339 	stats_vdev->dev = dev;
340 	INIT_LIST_HEAD(&stats_vdev->stat);
341 	spin_lock_init(&stats_vdev->irq_lock);
342 
343 	strlcpy(vdev->name, "rkispp-stats", sizeof(vdev->name));
344 
345 	vdev->ioctl_ops = &rkispp_stats_ioctl;
346 	vdev->fops = &rkispp_stats_fops;
347 	vdev->release = video_device_release_empty;
348 	vdev->lock = &dev->iqlock;
349 	vdev->v4l2_dev = &dev->v4l2_dev;
350 	vdev->queue = &node->buf_queue;
351 	vdev->device_caps = V4L2_CAP_META_CAPTURE | V4L2_CAP_STREAMING;
352 	vdev->vfl_dir =  VFL_DIR_RX;
353 	rkispp_stats_init_vb2_queue(vdev->queue, stats_vdev);
354 	rkispp_init_stats_vdev(stats_vdev);
355 	video_set_drvdata(vdev, stats_vdev);
356 
357 	node->pad.flags = MEDIA_PAD_FL_SINK;
358 	ret = media_entity_pads_init(&vdev->entity, 1, &node->pad);
359 	if (ret < 0)
360 		goto err_release_queue;
361 
362 	ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1);
363 	if (ret < 0) {
364 		dev_err(&vdev->dev,
365 			"could not register Video for Linux device\n");
366 		goto err_cleanup_media_entity;
367 	}
368 
369 	return 0;
370 
371 err_cleanup_media_entity:
372 	media_entity_cleanup(&vdev->entity);
373 err_release_queue:
374 	vb2_queue_release(vdev->queue);
375 	return ret;
376 }
377 
rkispp_unregister_stats_vdev(struct rkispp_device *dev)378 void rkispp_unregister_stats_vdev(struct rkispp_device *dev)
379 {
380 	struct rkispp_stats_vdev *stats_vdev = &dev->stats_vdev;
381 	struct rkispp_vdev_node *node = &stats_vdev->vnode;
382 	struct video_device *vdev = &node->vdev;
383 
384 	video_unregister_device(vdev);
385 	media_entity_cleanup(&vdev->entity);
386 	vb2_queue_release(vdev->queue);
387 }
388 
389