1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Copyright (c) 2011 - 2012 Samsung Electronics Co., Ltd.
4 *		http://www.samsung.com
5 *
6 * Samsung EXYNOS5 SoC series G-Scaler driver
7 */
8
9#include <linux/module.h>
10#include <linux/kernel.h>
11#include <linux/types.h>
12#include <linux/errno.h>
13#include <linux/bug.h>
14#include <linux/interrupt.h>
15#include <linux/workqueue.h>
16#include <linux/device.h>
17#include <linux/platform_device.h>
18#include <linux/list.h>
19#include <linux/io.h>
20#include <linux/slab.h>
21#include <linux/clk.h>
22
23#include <media/v4l2-ioctl.h>
24
25#include "gsc-core.h"
26
27static int gsc_m2m_ctx_stop_req(struct gsc_ctx *ctx)
28{
29	struct gsc_ctx *curr_ctx;
30	struct gsc_dev *gsc = ctx->gsc_dev;
31	int ret;
32
33	curr_ctx = v4l2_m2m_get_curr_priv(gsc->m2m.m2m_dev);
34	if (!gsc_m2m_pending(gsc) || (curr_ctx != ctx))
35		return 0;
36
37	gsc_ctx_state_lock_set(GSC_CTX_STOP_REQ, ctx);
38	ret = wait_event_timeout(gsc->irq_queue,
39			!gsc_ctx_state_is_set(GSC_CTX_STOP_REQ, ctx),
40			GSC_SHUTDOWN_TIMEOUT);
41
42	return ret == 0 ? -ETIMEDOUT : ret;
43}
44
45static void __gsc_m2m_job_abort(struct gsc_ctx *ctx)
46{
47	int ret;
48
49	ret = gsc_m2m_ctx_stop_req(ctx);
50	if ((ret == -ETIMEDOUT) || (ctx->state & GSC_CTX_ABORT)) {
51		gsc_ctx_state_lock_clear(GSC_CTX_STOP_REQ | GSC_CTX_ABORT, ctx);
52		gsc_m2m_job_finish(ctx, VB2_BUF_STATE_ERROR);
53	}
54}
55
56static int gsc_m2m_start_streaming(struct vb2_queue *q, unsigned int count)
57{
58	struct gsc_ctx *ctx = q->drv_priv;
59
60	return pm_runtime_resume_and_get(&ctx->gsc_dev->pdev->dev);
61}
62
63static void __gsc_m2m_cleanup_queue(struct gsc_ctx *ctx)
64{
65	struct vb2_v4l2_buffer *src_vb, *dst_vb;
66
67	while (v4l2_m2m_num_src_bufs_ready(ctx->m2m_ctx) > 0) {
68		src_vb = v4l2_m2m_src_buf_remove(ctx->m2m_ctx);
69		v4l2_m2m_buf_done(src_vb, VB2_BUF_STATE_ERROR);
70	}
71
72	while (v4l2_m2m_num_dst_bufs_ready(ctx->m2m_ctx) > 0) {
73		dst_vb = v4l2_m2m_dst_buf_remove(ctx->m2m_ctx);
74		v4l2_m2m_buf_done(dst_vb, VB2_BUF_STATE_ERROR);
75	}
76}
77
78static void gsc_m2m_stop_streaming(struct vb2_queue *q)
79{
80	struct gsc_ctx *ctx = q->drv_priv;
81
82	__gsc_m2m_job_abort(ctx);
83
84	__gsc_m2m_cleanup_queue(ctx);
85
86	pm_runtime_put(&ctx->gsc_dev->pdev->dev);
87}
88
89void gsc_m2m_job_finish(struct gsc_ctx *ctx, int vb_state)
90{
91	struct vb2_v4l2_buffer *src_vb, *dst_vb;
92
93	if (!ctx || !ctx->m2m_ctx)
94		return;
95
96	src_vb = v4l2_m2m_src_buf_remove(ctx->m2m_ctx);
97	dst_vb = v4l2_m2m_dst_buf_remove(ctx->m2m_ctx);
98
99	if (src_vb && dst_vb) {
100		dst_vb->vb2_buf.timestamp = src_vb->vb2_buf.timestamp;
101		dst_vb->timecode = src_vb->timecode;
102		dst_vb->flags &= ~V4L2_BUF_FLAG_TSTAMP_SRC_MASK;
103		dst_vb->flags |=
104			src_vb->flags
105			& V4L2_BUF_FLAG_TSTAMP_SRC_MASK;
106
107		v4l2_m2m_buf_done(src_vb, vb_state);
108		v4l2_m2m_buf_done(dst_vb, vb_state);
109
110		v4l2_m2m_job_finish(ctx->gsc_dev->m2m.m2m_dev,
111				    ctx->m2m_ctx);
112	}
113}
114
115static void gsc_m2m_job_abort(void *priv)
116{
117	__gsc_m2m_job_abort((struct gsc_ctx *)priv);
118}
119
120static int gsc_get_bufs(struct gsc_ctx *ctx)
121{
122	struct gsc_frame *s_frame, *d_frame;
123	struct vb2_v4l2_buffer *src_vb, *dst_vb;
124	int ret;
125
126	s_frame = &ctx->s_frame;
127	d_frame = &ctx->d_frame;
128
129	src_vb = v4l2_m2m_next_src_buf(ctx->m2m_ctx);
130	ret = gsc_prepare_addr(ctx, &src_vb->vb2_buf, s_frame, &s_frame->addr);
131	if (ret)
132		return ret;
133
134	dst_vb = v4l2_m2m_next_dst_buf(ctx->m2m_ctx);
135	ret = gsc_prepare_addr(ctx, &dst_vb->vb2_buf, d_frame, &d_frame->addr);
136	if (ret)
137		return ret;
138
139	dst_vb->vb2_buf.timestamp = src_vb->vb2_buf.timestamp;
140
141	return 0;
142}
143
144static void gsc_m2m_device_run(void *priv)
145{
146	struct gsc_ctx *ctx = priv;
147	struct gsc_dev *gsc;
148	unsigned long flags;
149	int ret;
150	bool is_set = false;
151
152	if (WARN(!ctx, "null hardware context\n"))
153		return;
154
155	gsc = ctx->gsc_dev;
156	spin_lock_irqsave(&gsc->slock, flags);
157
158	set_bit(ST_M2M_PEND, &gsc->state);
159
160	/* Reconfigure hardware if the context has changed. */
161	if (gsc->m2m.ctx != ctx) {
162		pr_debug("gsc->m2m.ctx = 0x%p, current_ctx = 0x%p",
163				gsc->m2m.ctx, ctx);
164		ctx->state |= GSC_PARAMS;
165		gsc->m2m.ctx = ctx;
166	}
167
168	is_set = ctx->state & GSC_CTX_STOP_REQ;
169	if (is_set) {
170		ctx->state &= ~GSC_CTX_STOP_REQ;
171		ctx->state |= GSC_CTX_ABORT;
172		wake_up(&gsc->irq_queue);
173		goto put_device;
174	}
175
176	ret = gsc_get_bufs(ctx);
177	if (ret) {
178		pr_err("Wrong address");
179		goto put_device;
180	}
181
182	gsc_set_prefbuf(gsc, &ctx->s_frame);
183	gsc_hw_set_input_addr(gsc, &ctx->s_frame.addr, GSC_M2M_BUF_NUM);
184	gsc_hw_set_output_addr(gsc, &ctx->d_frame.addr, GSC_M2M_BUF_NUM);
185
186	if (ctx->state & GSC_PARAMS) {
187		gsc_hw_set_input_buf_masking(gsc, GSC_M2M_BUF_NUM, false);
188		gsc_hw_set_output_buf_masking(gsc, GSC_M2M_BUF_NUM, false);
189		gsc_hw_set_frm_done_irq_mask(gsc, false);
190		gsc_hw_set_gsc_irq_enable(gsc, true);
191
192		if (gsc_set_scaler_info(ctx)) {
193			pr_err("Scaler setup error");
194			goto put_device;
195		}
196
197		gsc_hw_set_input_path(ctx);
198		gsc_hw_set_in_size(ctx);
199		gsc_hw_set_in_image_format(ctx);
200
201		gsc_hw_set_output_path(ctx);
202		gsc_hw_set_out_size(ctx);
203		gsc_hw_set_out_image_format(ctx);
204
205		gsc_hw_set_prescaler(ctx);
206		gsc_hw_set_mainscaler(ctx);
207		gsc_hw_set_rotation(ctx);
208		gsc_hw_set_global_alpha(ctx);
209	}
210
211	/* update shadow registers */
212	gsc_hw_set_sfr_update(ctx);
213
214	ctx->state &= ~GSC_PARAMS;
215	gsc_hw_enable_control(gsc, true);
216
217	spin_unlock_irqrestore(&gsc->slock, flags);
218	return;
219
220put_device:
221	ctx->state &= ~GSC_PARAMS;
222	spin_unlock_irqrestore(&gsc->slock, flags);
223}
224
225static int gsc_m2m_queue_setup(struct vb2_queue *vq,
226			unsigned int *num_buffers, unsigned int *num_planes,
227			unsigned int sizes[], struct device *alloc_devs[])
228{
229	struct gsc_ctx *ctx = vb2_get_drv_priv(vq);
230	struct gsc_frame *frame;
231	int i;
232
233	frame = ctx_get_frame(ctx, vq->type);
234	if (IS_ERR(frame))
235		return PTR_ERR(frame);
236
237	if (!frame->fmt)
238		return -EINVAL;
239
240	*num_planes = frame->fmt->num_planes;
241	for (i = 0; i < frame->fmt->num_planes; i++)
242		sizes[i] = frame->payload[i];
243	return 0;
244}
245
246static int gsc_m2m_buf_prepare(struct vb2_buffer *vb)
247{
248	struct gsc_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue);
249	struct gsc_frame *frame;
250	int i;
251
252	frame = ctx_get_frame(ctx, vb->vb2_queue->type);
253	if (IS_ERR(frame))
254		return PTR_ERR(frame);
255
256	if (V4L2_TYPE_IS_CAPTURE(vb->vb2_queue->type)) {
257		for (i = 0; i < frame->fmt->num_planes; i++)
258			vb2_set_plane_payload(vb, i, frame->payload[i]);
259	}
260
261	return 0;
262}
263
264static void gsc_m2m_buf_queue(struct vb2_buffer *vb)
265{
266	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
267	struct gsc_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue);
268
269	pr_debug("ctx: %p, ctx->state: 0x%x", ctx, ctx->state);
270
271	if (ctx->m2m_ctx)
272		v4l2_m2m_buf_queue(ctx->m2m_ctx, vbuf);
273}
274
275static const struct vb2_ops gsc_m2m_qops = {
276	.queue_setup	 = gsc_m2m_queue_setup,
277	.buf_prepare	 = gsc_m2m_buf_prepare,
278	.buf_queue	 = gsc_m2m_buf_queue,
279	.wait_prepare	 = vb2_ops_wait_prepare,
280	.wait_finish	 = vb2_ops_wait_finish,
281	.stop_streaming	 = gsc_m2m_stop_streaming,
282	.start_streaming = gsc_m2m_start_streaming,
283};
284
285static int gsc_m2m_querycap(struct file *file, void *fh,
286			   struct v4l2_capability *cap)
287{
288	struct gsc_ctx *ctx = fh_to_ctx(fh);
289	struct gsc_dev *gsc = ctx->gsc_dev;
290
291	strscpy(cap->driver, GSC_MODULE_NAME, sizeof(cap->driver));
292	strscpy(cap->card, GSC_MODULE_NAME " gscaler", sizeof(cap->card));
293	snprintf(cap->bus_info, sizeof(cap->bus_info), "platform:%s",
294		 dev_name(&gsc->pdev->dev));
295	return 0;
296}
297
298static int gsc_m2m_enum_fmt(struct file *file, void *priv,
299			    struct v4l2_fmtdesc *f)
300{
301	return gsc_enum_fmt(f);
302}
303
304static int gsc_m2m_g_fmt_mplane(struct file *file, void *fh,
305			     struct v4l2_format *f)
306{
307	struct gsc_ctx *ctx = fh_to_ctx(fh);
308
309	return gsc_g_fmt_mplane(ctx, f);
310}
311
312static int gsc_m2m_try_fmt_mplane(struct file *file, void *fh,
313				  struct v4l2_format *f)
314{
315	struct gsc_ctx *ctx = fh_to_ctx(fh);
316
317	return gsc_try_fmt_mplane(ctx, f);
318}
319
320static int gsc_m2m_s_fmt_mplane(struct file *file, void *fh,
321				 struct v4l2_format *f)
322{
323	struct gsc_ctx *ctx = fh_to_ctx(fh);
324	struct vb2_queue *vq;
325	struct gsc_frame *frame;
326	struct v4l2_pix_format_mplane *pix;
327	int i, ret = 0;
328
329	ret = gsc_m2m_try_fmt_mplane(file, fh, f);
330	if (ret)
331		return ret;
332
333	vq = v4l2_m2m_get_vq(ctx->m2m_ctx, f->type);
334
335	if (vb2_is_streaming(vq)) {
336		pr_err("queue (%d) busy", f->type);
337		return -EBUSY;
338	}
339
340	if (V4L2_TYPE_IS_OUTPUT(f->type))
341		frame = &ctx->s_frame;
342	else
343		frame = &ctx->d_frame;
344
345	pix = &f->fmt.pix_mp;
346	frame->fmt = find_fmt(&pix->pixelformat, NULL, 0);
347	frame->colorspace = pix->colorspace;
348	if (!frame->fmt)
349		return -EINVAL;
350
351	for (i = 0; i < frame->fmt->num_planes; i++)
352		frame->payload[i] = pix->plane_fmt[i].sizeimage;
353
354	gsc_set_frame_size(frame, pix->width, pix->height);
355
356	if (f->type == V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE)
357		gsc_ctx_state_lock_set(GSC_PARAMS | GSC_DST_FMT, ctx);
358	else
359		gsc_ctx_state_lock_set(GSC_PARAMS | GSC_SRC_FMT, ctx);
360
361	pr_debug("f_w: %d, f_h: %d", frame->f_width, frame->f_height);
362
363	return 0;
364}
365
366static int gsc_m2m_reqbufs(struct file *file, void *fh,
367			  struct v4l2_requestbuffers *reqbufs)
368{
369	struct gsc_ctx *ctx = fh_to_ctx(fh);
370	struct gsc_dev *gsc = ctx->gsc_dev;
371	u32 max_cnt;
372
373	max_cnt = (reqbufs->type == V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE) ?
374		gsc->variant->in_buf_cnt : gsc->variant->out_buf_cnt;
375	if (reqbufs->count > max_cnt)
376		return -EINVAL;
377
378	return v4l2_m2m_reqbufs(file, ctx->m2m_ctx, reqbufs);
379}
380
381static int gsc_m2m_expbuf(struct file *file, void *fh,
382				struct v4l2_exportbuffer *eb)
383{
384	struct gsc_ctx *ctx = fh_to_ctx(fh);
385	return v4l2_m2m_expbuf(file, ctx->m2m_ctx, eb);
386}
387
388static int gsc_m2m_querybuf(struct file *file, void *fh,
389					struct v4l2_buffer *buf)
390{
391	struct gsc_ctx *ctx = fh_to_ctx(fh);
392	return v4l2_m2m_querybuf(file, ctx->m2m_ctx, buf);
393}
394
395static int gsc_m2m_qbuf(struct file *file, void *fh,
396			  struct v4l2_buffer *buf)
397{
398	struct gsc_ctx *ctx = fh_to_ctx(fh);
399	return v4l2_m2m_qbuf(file, ctx->m2m_ctx, buf);
400}
401
402static int gsc_m2m_dqbuf(struct file *file, void *fh,
403			   struct v4l2_buffer *buf)
404{
405	struct gsc_ctx *ctx = fh_to_ctx(fh);
406	return v4l2_m2m_dqbuf(file, ctx->m2m_ctx, buf);
407}
408
409static int gsc_m2m_streamon(struct file *file, void *fh,
410			   enum v4l2_buf_type type)
411{
412	struct gsc_ctx *ctx = fh_to_ctx(fh);
413
414	/* The source and target color format need to be set */
415	if (V4L2_TYPE_IS_OUTPUT(type)) {
416		if (!gsc_ctx_state_is_set(GSC_SRC_FMT, ctx))
417			return -EINVAL;
418	} else if (!gsc_ctx_state_is_set(GSC_DST_FMT, ctx)) {
419		return -EINVAL;
420	}
421
422	return v4l2_m2m_streamon(file, ctx->m2m_ctx, type);
423}
424
425static int gsc_m2m_streamoff(struct file *file, void *fh,
426			    enum v4l2_buf_type type)
427{
428	struct gsc_ctx *ctx = fh_to_ctx(fh);
429	return v4l2_m2m_streamoff(file, ctx->m2m_ctx, type);
430}
431
432/* Return 1 if rectangle a is enclosed in rectangle b, or 0 otherwise. */
433static int is_rectangle_enclosed(struct v4l2_rect *a, struct v4l2_rect *b)
434{
435	if (a->left < b->left || a->top < b->top)
436		return 0;
437
438	if (a->left + a->width > b->left + b->width)
439		return 0;
440
441	if (a->top + a->height > b->top + b->height)
442		return 0;
443
444	return 1;
445}
446
447static int gsc_m2m_g_selection(struct file *file, void *fh,
448			struct v4l2_selection *s)
449{
450	struct gsc_frame *frame;
451	struct gsc_ctx *ctx = fh_to_ctx(fh);
452
453	if ((s->type != V4L2_BUF_TYPE_VIDEO_CAPTURE) &&
454	    (s->type != V4L2_BUF_TYPE_VIDEO_OUTPUT))
455		return -EINVAL;
456
457	frame = ctx_get_frame(ctx, s->type);
458	if (IS_ERR(frame))
459		return PTR_ERR(frame);
460
461	switch (s->target) {
462	case V4L2_SEL_TGT_COMPOSE_DEFAULT:
463	case V4L2_SEL_TGT_COMPOSE_BOUNDS:
464	case V4L2_SEL_TGT_CROP_BOUNDS:
465	case V4L2_SEL_TGT_CROP_DEFAULT:
466		s->r.left = 0;
467		s->r.top = 0;
468		s->r.width = frame->f_width;
469		s->r.height = frame->f_height;
470		return 0;
471
472	case V4L2_SEL_TGT_COMPOSE:
473	case V4L2_SEL_TGT_CROP:
474		s->r.left = frame->crop.left;
475		s->r.top = frame->crop.top;
476		s->r.width = frame->crop.width;
477		s->r.height = frame->crop.height;
478		return 0;
479	}
480
481	return -EINVAL;
482}
483
484static int gsc_m2m_s_selection(struct file *file, void *fh,
485				struct v4l2_selection *s)
486{
487	struct gsc_frame *frame;
488	struct gsc_ctx *ctx = fh_to_ctx(fh);
489	struct gsc_variant *variant = ctx->gsc_dev->variant;
490	struct v4l2_selection sel = *s;
491	int ret;
492
493	if ((s->type != V4L2_BUF_TYPE_VIDEO_CAPTURE) &&
494	    (s->type != V4L2_BUF_TYPE_VIDEO_OUTPUT))
495		return -EINVAL;
496
497	ret = gsc_try_selection(ctx, &sel);
498	if (ret)
499		return ret;
500
501	if (s->flags & V4L2_SEL_FLAG_LE &&
502	    !is_rectangle_enclosed(&sel.r, &s->r))
503		return -ERANGE;
504
505	if (s->flags & V4L2_SEL_FLAG_GE &&
506	    !is_rectangle_enclosed(&s->r, &sel.r))
507		return -ERANGE;
508
509	s->r = sel.r;
510
511	switch (s->target) {
512	case V4L2_SEL_TGT_COMPOSE_BOUNDS:
513	case V4L2_SEL_TGT_COMPOSE_DEFAULT:
514	case V4L2_SEL_TGT_COMPOSE:
515		frame = &ctx->s_frame;
516		break;
517
518	case V4L2_SEL_TGT_CROP_BOUNDS:
519	case V4L2_SEL_TGT_CROP:
520	case V4L2_SEL_TGT_CROP_DEFAULT:
521		frame = &ctx->d_frame;
522		break;
523
524	default:
525		return -EINVAL;
526	}
527
528	/* Check to see if scaling ratio is within supported range */
529	if (gsc_ctx_state_is_set(GSC_DST_FMT | GSC_SRC_FMT, ctx)) {
530		if (s->type == V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE) {
531			ret = gsc_check_scaler_ratio(variant, sel.r.width,
532				sel.r.height, ctx->d_frame.crop.width,
533				ctx->d_frame.crop.height,
534				ctx->gsc_ctrls.rotate->val, ctx->out_path);
535		} else {
536			ret = gsc_check_scaler_ratio(variant,
537				ctx->s_frame.crop.width,
538				ctx->s_frame.crop.height, sel.r.width,
539				sel.r.height, ctx->gsc_ctrls.rotate->val,
540				ctx->out_path);
541		}
542
543		if (ret) {
544			pr_err("Out of scaler range");
545			return -EINVAL;
546		}
547	}
548
549	frame->crop = sel.r;
550
551	gsc_ctx_state_lock_set(GSC_PARAMS, ctx);
552	return 0;
553}
554
555static const struct v4l2_ioctl_ops gsc_m2m_ioctl_ops = {
556	.vidioc_querycap		= gsc_m2m_querycap,
557	.vidioc_enum_fmt_vid_cap	= gsc_m2m_enum_fmt,
558	.vidioc_enum_fmt_vid_out	= gsc_m2m_enum_fmt,
559	.vidioc_g_fmt_vid_cap_mplane	= gsc_m2m_g_fmt_mplane,
560	.vidioc_g_fmt_vid_out_mplane	= gsc_m2m_g_fmt_mplane,
561	.vidioc_try_fmt_vid_cap_mplane	= gsc_m2m_try_fmt_mplane,
562	.vidioc_try_fmt_vid_out_mplane	= gsc_m2m_try_fmt_mplane,
563	.vidioc_s_fmt_vid_cap_mplane	= gsc_m2m_s_fmt_mplane,
564	.vidioc_s_fmt_vid_out_mplane	= gsc_m2m_s_fmt_mplane,
565	.vidioc_reqbufs			= gsc_m2m_reqbufs,
566	.vidioc_expbuf                  = gsc_m2m_expbuf,
567	.vidioc_querybuf		= gsc_m2m_querybuf,
568	.vidioc_qbuf			= gsc_m2m_qbuf,
569	.vidioc_dqbuf			= gsc_m2m_dqbuf,
570	.vidioc_streamon		= gsc_m2m_streamon,
571	.vidioc_streamoff		= gsc_m2m_streamoff,
572	.vidioc_g_selection		= gsc_m2m_g_selection,
573	.vidioc_s_selection		= gsc_m2m_s_selection
574};
575
576static int queue_init(void *priv, struct vb2_queue *src_vq,
577			struct vb2_queue *dst_vq)
578{
579	struct gsc_ctx *ctx = priv;
580	int ret;
581
582	memset(src_vq, 0, sizeof(*src_vq));
583	src_vq->type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
584	src_vq->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
585	src_vq->drv_priv = ctx;
586	src_vq->ops = &gsc_m2m_qops;
587	src_vq->mem_ops = &vb2_dma_contig_memops;
588	src_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer);
589	src_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
590	src_vq->lock = &ctx->gsc_dev->lock;
591	src_vq->dev = &ctx->gsc_dev->pdev->dev;
592
593	ret = vb2_queue_init(src_vq);
594	if (ret)
595		return ret;
596
597	memset(dst_vq, 0, sizeof(*dst_vq));
598	dst_vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
599	dst_vq->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
600	dst_vq->drv_priv = ctx;
601	dst_vq->ops = &gsc_m2m_qops;
602	dst_vq->mem_ops = &vb2_dma_contig_memops;
603	dst_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer);
604	dst_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
605	dst_vq->lock = &ctx->gsc_dev->lock;
606	dst_vq->dev = &ctx->gsc_dev->pdev->dev;
607
608	return vb2_queue_init(dst_vq);
609}
610
611static int gsc_m2m_open(struct file *file)
612{
613	struct gsc_dev *gsc = video_drvdata(file);
614	struct gsc_ctx *ctx = NULL;
615	int ret;
616
617	pr_debug("pid: %d, state: 0x%lx", task_pid_nr(current), gsc->state);
618
619	if (mutex_lock_interruptible(&gsc->lock))
620		return -ERESTARTSYS;
621
622	ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
623	if (!ctx) {
624		ret = -ENOMEM;
625		goto unlock;
626	}
627
628	v4l2_fh_init(&ctx->fh, gsc->m2m.vfd);
629	ret = gsc_ctrls_create(ctx);
630	if (ret)
631		goto error_fh;
632
633	/* Use separate control handler per file handle */
634	ctx->fh.ctrl_handler = &ctx->ctrl_handler;
635	file->private_data = &ctx->fh;
636	v4l2_fh_add(&ctx->fh);
637
638	ctx->gsc_dev = gsc;
639	/* Default color format */
640	ctx->s_frame.fmt = get_format(0);
641	ctx->d_frame.fmt = get_format(0);
642	/* Setup the device context for mem2mem mode. */
643	ctx->state = GSC_CTX_M2M;
644	ctx->flags = 0;
645	ctx->in_path = GSC_DMA;
646	ctx->out_path = GSC_DMA;
647
648	ctx->m2m_ctx = v4l2_m2m_ctx_init(gsc->m2m.m2m_dev, ctx, queue_init);
649	if (IS_ERR(ctx->m2m_ctx)) {
650		pr_err("Failed to initialize m2m context");
651		ret = PTR_ERR(ctx->m2m_ctx);
652		goto error_ctrls;
653	}
654
655	if (gsc->m2m.refcnt++ == 0)
656		set_bit(ST_M2M_OPEN, &gsc->state);
657
658	pr_debug("gsc m2m driver is opened, ctx(0x%p)", ctx);
659
660	mutex_unlock(&gsc->lock);
661	return 0;
662
663error_ctrls:
664	gsc_ctrls_delete(ctx);
665	v4l2_fh_del(&ctx->fh);
666error_fh:
667	v4l2_fh_exit(&ctx->fh);
668	kfree(ctx);
669unlock:
670	mutex_unlock(&gsc->lock);
671	return ret;
672}
673
674static int gsc_m2m_release(struct file *file)
675{
676	struct gsc_ctx *ctx = fh_to_ctx(file->private_data);
677	struct gsc_dev *gsc = ctx->gsc_dev;
678
679	pr_debug("pid: %d, state: 0x%lx, refcnt= %d",
680		task_pid_nr(current), gsc->state, gsc->m2m.refcnt);
681
682	mutex_lock(&gsc->lock);
683
684	v4l2_m2m_ctx_release(ctx->m2m_ctx);
685	gsc_ctrls_delete(ctx);
686	v4l2_fh_del(&ctx->fh);
687	v4l2_fh_exit(&ctx->fh);
688
689	if (--gsc->m2m.refcnt <= 0)
690		clear_bit(ST_M2M_OPEN, &gsc->state);
691	kfree(ctx);
692
693	mutex_unlock(&gsc->lock);
694	return 0;
695}
696
697static __poll_t gsc_m2m_poll(struct file *file,
698					struct poll_table_struct *wait)
699{
700	struct gsc_ctx *ctx = fh_to_ctx(file->private_data);
701	struct gsc_dev *gsc = ctx->gsc_dev;
702	__poll_t ret;
703
704	if (mutex_lock_interruptible(&gsc->lock))
705		return EPOLLERR;
706
707	ret = v4l2_m2m_poll(file, ctx->m2m_ctx, wait);
708	mutex_unlock(&gsc->lock);
709
710	return ret;
711}
712
713static int gsc_m2m_mmap(struct file *file, struct vm_area_struct *vma)
714{
715	struct gsc_ctx *ctx = fh_to_ctx(file->private_data);
716	struct gsc_dev *gsc = ctx->gsc_dev;
717	int ret;
718
719	if (mutex_lock_interruptible(&gsc->lock))
720		return -ERESTARTSYS;
721
722	ret = v4l2_m2m_mmap(file, ctx->m2m_ctx, vma);
723	mutex_unlock(&gsc->lock);
724
725	return ret;
726}
727
728static const struct v4l2_file_operations gsc_m2m_fops = {
729	.owner		= THIS_MODULE,
730	.open		= gsc_m2m_open,
731	.release	= gsc_m2m_release,
732	.poll		= gsc_m2m_poll,
733	.unlocked_ioctl	= video_ioctl2,
734	.mmap		= gsc_m2m_mmap,
735};
736
737static const struct v4l2_m2m_ops gsc_m2m_ops = {
738	.device_run	= gsc_m2m_device_run,
739	.job_abort	= gsc_m2m_job_abort,
740};
741
742int gsc_register_m2m_device(struct gsc_dev *gsc)
743{
744	struct platform_device *pdev;
745	int ret;
746
747	if (!gsc)
748		return -ENODEV;
749
750	pdev = gsc->pdev;
751
752	gsc->vdev.fops		= &gsc_m2m_fops;
753	gsc->vdev.ioctl_ops	= &gsc_m2m_ioctl_ops;
754	gsc->vdev.release	= video_device_release_empty;
755	gsc->vdev.lock		= &gsc->lock;
756	gsc->vdev.vfl_dir	= VFL_DIR_M2M;
757	gsc->vdev.v4l2_dev	= &gsc->v4l2_dev;
758	gsc->vdev.device_caps	= V4L2_CAP_STREAMING |
759				  V4L2_CAP_VIDEO_M2M_MPLANE;
760	snprintf(gsc->vdev.name, sizeof(gsc->vdev.name), "%s.%d:m2m",
761					GSC_MODULE_NAME, gsc->id);
762
763	video_set_drvdata(&gsc->vdev, gsc);
764
765	gsc->m2m.vfd = &gsc->vdev;
766	gsc->m2m.m2m_dev = v4l2_m2m_init(&gsc_m2m_ops);
767	if (IS_ERR(gsc->m2m.m2m_dev)) {
768		dev_err(&pdev->dev, "failed to initialize v4l2-m2m device\n");
769		return PTR_ERR(gsc->m2m.m2m_dev);
770	}
771
772	ret = video_register_device(&gsc->vdev, VFL_TYPE_VIDEO, -1);
773	if (ret) {
774		dev_err(&pdev->dev,
775			 "%s(): failed to register video device\n", __func__);
776		goto err_m2m_release;
777	}
778
779	pr_debug("gsc m2m driver registered as /dev/video%d", gsc->vdev.num);
780	return 0;
781
782err_m2m_release:
783	v4l2_m2m_release(gsc->m2m.m2m_dev);
784
785	return ret;
786}
787
788void gsc_unregister_m2m_device(struct gsc_dev *gsc)
789{
790	if (gsc) {
791		v4l2_m2m_release(gsc->m2m.m2m_dev);
792		video_unregister_device(&gsc->vdev);
793	}
794}
795