1/*
2 * Xilinx Video DMA
3 *
4 * Copyright (C) 2013-2015 Ideas on Board
5 * Copyright (C) 2013-2015 Xilinx, Inc.
6 *
7 * Contacts: Hyun Kwon <hyun.kwon@xilinx.com>
8 *           Laurent Pinchart <laurent.pinchart@ideasonboard.com>
9 *
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License version 2 as
12 * published by the Free Software Foundation.
13 */
14
15#include <linux/dma/xilinx_dma.h>
16#include <linux/lcm.h>
17#include <linux/list.h>
18#include <linux/module.h>
19#include <linux/of.h>
20#include <linux/slab.h>
21
22#include <media/v4l2-dev.h>
23#include <media/v4l2-fh.h>
24#include <media/v4l2-ioctl.h>
25#include <media/videobuf2-core.h>
26#include <media/videobuf2-dma-contig.h>
27
28#include "xilinx-dma.h"
29#include "xilinx-vip.h"
30#include "xilinx-vipp.h"
31
32#define XVIP_DMA_DEF_FORMAT		V4L2_PIX_FMT_YUYV
33#define XVIP_DMA_DEF_WIDTH		1920
34#define XVIP_DMA_DEF_HEIGHT		1080
35
36/* Minimum and maximum widths are expressed in bytes */
37#define XVIP_DMA_MIN_WIDTH		1U
38#define XVIP_DMA_MAX_WIDTH		65535U
39#define XVIP_DMA_MIN_HEIGHT		1U
40#define XVIP_DMA_MAX_HEIGHT		8191U
41
42/* -----------------------------------------------------------------------------
43 * Helper functions
44 */
45
46static struct v4l2_subdev *
47xvip_dma_remote_subdev(struct media_pad *local, u32 *pad)
48{
49	struct media_pad *remote;
50
51	remote = media_entity_remote_pad(local);
52	if (remote == NULL ||
53	    media_entity_type(remote->entity) != MEDIA_ENT_T_V4L2_SUBDEV)
54		return NULL;
55
56	if (pad)
57		*pad = remote->index;
58
59	return media_entity_to_v4l2_subdev(remote->entity);
60}
61
62static int xvip_dma_verify_format(struct xvip_dma *dma)
63{
64	struct v4l2_subdev_format fmt;
65	struct v4l2_subdev *subdev;
66	int ret;
67
68	subdev = xvip_dma_remote_subdev(&dma->pad, &fmt.pad);
69	if (subdev == NULL)
70		return -EPIPE;
71
72	fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE;
73	ret = v4l2_subdev_call(subdev, pad, get_fmt, NULL, &fmt);
74	if (ret < 0)
75		return ret == -ENOIOCTLCMD ? -EINVAL : ret;
76
77	if (dma->fmtinfo->code != fmt.format.code ||
78	    dma->format.height != fmt.format.height ||
79	    dma->format.width != fmt.format.width ||
80	    dma->format.colorspace != fmt.format.colorspace)
81		return -EINVAL;
82
83	return 0;
84}
85
86/* -----------------------------------------------------------------------------
87 * Pipeline Stream Management
88 */
89
90/**
91 * xvip_pipeline_start_stop - Start ot stop streaming on a pipeline
92 * @pipe: The pipeline
93 * @start: Start (when true) or stop (when false) the pipeline
94 *
95 * Walk the entities chain starting at the pipeline output video node and start
96 * or stop all of them.
97 *
98 * Return: 0 if successful, or the return value of the failed video::s_stream
99 * operation otherwise.
100 */
101static int xvip_pipeline_start_stop(struct xvip_pipeline *pipe, bool start)
102{
103	struct xvip_dma *dma = pipe->output;
104	struct media_entity *entity;
105	struct media_pad *pad;
106	struct v4l2_subdev *subdev;
107	int ret;
108
109	entity = &dma->video.entity;
110	while (1) {
111		pad = &entity->pads[0];
112		if (!(pad->flags & MEDIA_PAD_FL_SINK))
113			break;
114
115		pad = media_entity_remote_pad(pad);
116		if (pad == NULL ||
117		    media_entity_type(pad->entity) != MEDIA_ENT_T_V4L2_SUBDEV)
118			break;
119
120		entity = pad->entity;
121		subdev = media_entity_to_v4l2_subdev(entity);
122
123		ret = v4l2_subdev_call(subdev, video, s_stream, start);
124		if (start && ret < 0 && ret != -ENOIOCTLCMD)
125			return ret;
126	}
127
128	return 0;
129}
130
131/**
132 * xvip_pipeline_set_stream - Enable/disable streaming on a pipeline
133 * @pipe: The pipeline
134 * @on: Turn the stream on when true or off when false
135 *
136 * The pipeline is shared between all DMA engines connect at its input and
137 * output. While the stream state of DMA engines can be controlled
138 * independently, pipelines have a shared stream state that enable or disable
139 * all entities in the pipeline. For this reason the pipeline uses a streaming
140 * counter that tracks the number of DMA engines that have requested the stream
141 * to be enabled.
142 *
143 * When called with the @on argument set to true, this function will increment
144 * the pipeline streaming count. If the streaming count reaches the number of
145 * DMA engines in the pipeline it will enable all entities that belong to the
146 * pipeline.
147 *
148 * Similarly, when called with the @on argument set to false, this function will
149 * decrement the pipeline streaming count and disable all entities in the
150 * pipeline when the streaming count reaches zero.
151 *
152 * Return: 0 if successful, or the return value of the failed video::s_stream
153 * operation otherwise. Stopping the pipeline never fails. The pipeline state is
154 * not updated when the operation fails.
155 */
156static int xvip_pipeline_set_stream(struct xvip_pipeline *pipe, bool on)
157{
158	int ret = 0;
159
160	mutex_lock(&pipe->lock);
161
162	if (on) {
163		if (pipe->stream_count == pipe->num_dmas - 1) {
164			ret = xvip_pipeline_start_stop(pipe, true);
165			if (ret < 0)
166				goto done;
167		}
168		pipe->stream_count++;
169	} else {
170		if (--pipe->stream_count == 0)
171			xvip_pipeline_start_stop(pipe, false);
172	}
173
174done:
175	mutex_unlock(&pipe->lock);
176	return ret;
177}
178
179static int xvip_pipeline_validate(struct xvip_pipeline *pipe,
180				  struct xvip_dma *start)
181{
182	struct media_entity_graph graph;
183	struct media_entity *entity = &start->video.entity;
184	struct media_device *mdev = entity->parent;
185	unsigned int num_inputs = 0;
186	unsigned int num_outputs = 0;
187
188	mutex_lock(&mdev->graph_mutex);
189
190	/* Walk the graph to locate the video nodes. */
191	media_entity_graph_walk_start(&graph, entity);
192
193	while ((entity = media_entity_graph_walk_next(&graph))) {
194		struct xvip_dma *dma;
195
196		if (entity->type != MEDIA_ENT_T_DEVNODE_V4L)
197			continue;
198
199		dma = to_xvip_dma(media_entity_to_video_device(entity));
200
201		if (dma->pad.flags & MEDIA_PAD_FL_SINK) {
202			pipe->output = dma;
203			num_outputs++;
204		} else {
205			num_inputs++;
206		}
207	}
208
209	mutex_unlock(&mdev->graph_mutex);
210
211	/* We need exactly one output and zero or one input. */
212	if (num_outputs != 1 || num_inputs > 1)
213		return -EPIPE;
214
215	pipe->num_dmas = num_inputs + num_outputs;
216
217	return 0;
218}
219
220static void __xvip_pipeline_cleanup(struct xvip_pipeline *pipe)
221{
222	pipe->num_dmas = 0;
223	pipe->output = NULL;
224}
225
226/**
227 * xvip_pipeline_cleanup - Cleanup the pipeline after streaming
228 * @pipe: the pipeline
229 *
230 * Decrease the pipeline use count and clean it up if we were the last user.
231 */
232static void xvip_pipeline_cleanup(struct xvip_pipeline *pipe)
233{
234	mutex_lock(&pipe->lock);
235
236	/* If we're the last user clean up the pipeline. */
237	if (--pipe->use_count == 0)
238		__xvip_pipeline_cleanup(pipe);
239
240	mutex_unlock(&pipe->lock);
241}
242
243/**
244 * xvip_pipeline_prepare - Prepare the pipeline for streaming
245 * @pipe: the pipeline
246 * @dma: DMA engine at one end of the pipeline
247 *
248 * Validate the pipeline if no user exists yet, otherwise just increase the use
249 * count.
250 *
251 * Return: 0 if successful or -EPIPE if the pipeline is not valid.
252 */
253static int xvip_pipeline_prepare(struct xvip_pipeline *pipe,
254				 struct xvip_dma *dma)
255{
256	int ret;
257
258	mutex_lock(&pipe->lock);
259
260	/* If we're the first user validate and initialize the pipeline. */
261	if (pipe->use_count == 0) {
262		ret = xvip_pipeline_validate(pipe, dma);
263		if (ret < 0) {
264			__xvip_pipeline_cleanup(pipe);
265			goto done;
266		}
267	}
268
269	pipe->use_count++;
270	ret = 0;
271
272done:
273	mutex_unlock(&pipe->lock);
274	return ret;
275}
276
277/* -----------------------------------------------------------------------------
278 * videobuf2 queue operations
279 */
280
281/**
282 * struct xvip_dma_buffer - Video DMA buffer
283 * @buf: vb2 buffer base object
284 * @queue: buffer list entry in the DMA engine queued buffers list
285 * @dma: DMA channel that uses the buffer
286 */
287struct xvip_dma_buffer {
288	struct vb2_buffer buf;
289	struct list_head queue;
290	struct xvip_dma *dma;
291};
292
293#define to_xvip_dma_buffer(vb)	container_of(vb, struct xvip_dma_buffer, buf)
294
295static void xvip_dma_complete(void *param)
296{
297	struct xvip_dma_buffer *buf = param;
298	struct xvip_dma *dma = buf->dma;
299
300	spin_lock(&dma->queued_lock);
301	list_del(&buf->queue);
302	spin_unlock(&dma->queued_lock);
303
304	buf->buf.v4l2_buf.field = V4L2_FIELD_NONE;
305	buf->buf.v4l2_buf.sequence = dma->sequence++;
306	v4l2_get_timestamp(&buf->buf.v4l2_buf.timestamp);
307	vb2_set_plane_payload(&buf->buf, 0, dma->format.sizeimage);
308	vb2_buffer_done(&buf->buf, VB2_BUF_STATE_DONE);
309}
310
311static int
312xvip_dma_queue_setup(struct vb2_queue *vq, const struct v4l2_format *fmt,
313		     unsigned int *nbuffers, unsigned int *nplanes,
314		     unsigned int sizes[], void *alloc_ctxs[])
315{
316	struct xvip_dma *dma = vb2_get_drv_priv(vq);
317
318	/* Make sure the image size is large enough. */
319	if (fmt && fmt->fmt.pix.sizeimage < dma->format.sizeimage)
320		return -EINVAL;
321
322	*nplanes = 1;
323
324	sizes[0] = fmt ? fmt->fmt.pix.sizeimage : dma->format.sizeimage;
325	alloc_ctxs[0] = dma->alloc_ctx;
326
327	return 0;
328}
329
330static int xvip_dma_buffer_prepare(struct vb2_buffer *vb)
331{
332	struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
333	struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vb);
334
335	buf->dma = dma;
336
337	return 0;
338}
339
340static void xvip_dma_buffer_queue(struct vb2_buffer *vb)
341{
342	struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
343	struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vb);
344	struct dma_async_tx_descriptor *desc;
345	dma_addr_t addr = vb2_dma_contig_plane_dma_addr(vb, 0);
346	u32 flags;
347
348	if (dma->queue.type == V4L2_BUF_TYPE_VIDEO_CAPTURE) {
349		flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK;
350		dma->xt.dir = DMA_DEV_TO_MEM;
351		dma->xt.src_sgl = false;
352		dma->xt.dst_sgl = true;
353		dma->xt.dst_start = addr;
354	} else {
355		flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK;
356		dma->xt.dir = DMA_MEM_TO_DEV;
357		dma->xt.src_sgl = true;
358		dma->xt.dst_sgl = false;
359		dma->xt.src_start = addr;
360	}
361
362	dma->xt.frame_size = 1;
363	dma->sgl[0].size = dma->format.width * dma->fmtinfo->bpp;
364	dma->sgl[0].icg = dma->format.bytesperline - dma->sgl[0].size;
365	dma->xt.numf = dma->format.height;
366
367	desc = dmaengine_prep_interleaved_dma(dma->dma, &dma->xt, flags);
368	if (!desc) {
369		dev_err(dma->xdev->dev, "Failed to prepare DMA transfer\n");
370		vb2_buffer_done(&buf->buf, VB2_BUF_STATE_ERROR);
371		return;
372	}
373	desc->callback = xvip_dma_complete;
374	desc->callback_param = buf;
375
376	spin_lock_irq(&dma->queued_lock);
377	list_add_tail(&buf->queue, &dma->queued_bufs);
378	spin_unlock_irq(&dma->queued_lock);
379
380	dmaengine_submit(desc);
381
382	if (vb2_is_streaming(&dma->queue))
383		dma_async_issue_pending(dma->dma);
384}
385
386static int xvip_dma_start_streaming(struct vb2_queue *vq, unsigned int count)
387{
388	struct xvip_dma *dma = vb2_get_drv_priv(vq);
389	struct xvip_dma_buffer *buf, *nbuf;
390	struct xvip_pipeline *pipe;
391	int ret;
392
393	dma->sequence = 0;
394
395	/*
396	 * Start streaming on the pipeline. No link touching an entity in the
397	 * pipeline can be activated or deactivated once streaming is started.
398	 *
399	 * Use the pipeline object embedded in the first DMA object that starts
400	 * streaming.
401	 */
402	pipe = dma->video.entity.pipe
403	     ? to_xvip_pipeline(&dma->video.entity) : &dma->pipe;
404
405	ret = media_entity_pipeline_start(&dma->video.entity, &pipe->pipe);
406	if (ret < 0)
407		goto error;
408
409	/* Verify that the configured format matches the output of the
410	 * connected subdev.
411	 */
412	ret = xvip_dma_verify_format(dma);
413	if (ret < 0)
414		goto error_stop;
415
416	ret = xvip_pipeline_prepare(pipe, dma);
417	if (ret < 0)
418		goto error_stop;
419
420	/* Start the DMA engine. This must be done before starting the blocks
421	 * in the pipeline to avoid DMA synchronization issues.
422	 */
423	dma_async_issue_pending(dma->dma);
424
425	/* Start the pipeline. */
426	xvip_pipeline_set_stream(pipe, true);
427
428	return 0;
429
430error_stop:
431	media_entity_pipeline_stop(&dma->video.entity);
432
433error:
434	/* Give back all queued buffers to videobuf2. */
435	spin_lock_irq(&dma->queued_lock);
436	list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
437		vb2_buffer_done(&buf->buf, VB2_BUF_STATE_QUEUED);
438		list_del(&buf->queue);
439	}
440	spin_unlock_irq(&dma->queued_lock);
441
442	return ret;
443}
444
445static void xvip_dma_stop_streaming(struct vb2_queue *vq)
446{
447	struct xvip_dma *dma = vb2_get_drv_priv(vq);
448	struct xvip_pipeline *pipe = to_xvip_pipeline(&dma->video.entity);
449	struct xvip_dma_buffer *buf, *nbuf;
450
451	/* Stop the pipeline. */
452	xvip_pipeline_set_stream(pipe, false);
453
454	/* Stop and reset the DMA engine. */
455	dmaengine_terminate_all(dma->dma);
456
457	/* Cleanup the pipeline and mark it as being stopped. */
458	xvip_pipeline_cleanup(pipe);
459	media_entity_pipeline_stop(&dma->video.entity);
460
461	/* Give back all queued buffers to videobuf2. */
462	spin_lock_irq(&dma->queued_lock);
463	list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
464		vb2_buffer_done(&buf->buf, VB2_BUF_STATE_ERROR);
465		list_del(&buf->queue);
466	}
467	spin_unlock_irq(&dma->queued_lock);
468}
469
470static struct vb2_ops xvip_dma_queue_qops = {
471	.queue_setup = xvip_dma_queue_setup,
472	.buf_prepare = xvip_dma_buffer_prepare,
473	.buf_queue = xvip_dma_buffer_queue,
474	.wait_prepare = vb2_ops_wait_prepare,
475	.wait_finish = vb2_ops_wait_finish,
476	.start_streaming = xvip_dma_start_streaming,
477	.stop_streaming = xvip_dma_stop_streaming,
478};
479
480/* -----------------------------------------------------------------------------
481 * V4L2 ioctls
482 */
483
484static int
485xvip_dma_querycap(struct file *file, void *fh, struct v4l2_capability *cap)
486{
487	struct v4l2_fh *vfh = file->private_data;
488	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
489
490	cap->capabilities = V4L2_CAP_DEVICE_CAPS | V4L2_CAP_STREAMING
491			  | dma->xdev->v4l2_caps;
492
493	if (dma->queue.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
494		cap->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING;
495	else
496		cap->device_caps = V4L2_CAP_VIDEO_OUTPUT | V4L2_CAP_STREAMING;
497
498	strlcpy(cap->driver, "xilinx-vipp", sizeof(cap->driver));
499	strlcpy(cap->card, dma->video.name, sizeof(cap->card));
500	snprintf(cap->bus_info, sizeof(cap->bus_info), "platform:%s:%u",
501		 dma->xdev->dev->of_node->name, dma->port);
502
503	return 0;
504}
505
506/* FIXME: without this callback function, some applications are not configured
507 * with correct formats, and it results in frames in wrong format. Whether this
508 * callback needs to be required is not clearly defined, so it should be
509 * clarified through the mailing list.
510 */
511static int
512xvip_dma_enum_format(struct file *file, void *fh, struct v4l2_fmtdesc *f)
513{
514	struct v4l2_fh *vfh = file->private_data;
515	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
516
517	if (f->index > 0)
518		return -EINVAL;
519
520	f->pixelformat = dma->format.pixelformat;
521	strlcpy(f->description, dma->fmtinfo->description,
522		sizeof(f->description));
523
524	return 0;
525}
526
527static int
528xvip_dma_get_format(struct file *file, void *fh, struct v4l2_format *format)
529{
530	struct v4l2_fh *vfh = file->private_data;
531	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
532
533	format->fmt.pix = dma->format;
534
535	return 0;
536}
537
538static void
539__xvip_dma_try_format(struct xvip_dma *dma, struct v4l2_pix_format *pix,
540		      const struct xvip_video_format **fmtinfo)
541{
542	const struct xvip_video_format *info;
543	unsigned int min_width;
544	unsigned int max_width;
545	unsigned int min_bpl;
546	unsigned int max_bpl;
547	unsigned int width;
548	unsigned int align;
549	unsigned int bpl;
550
551	/* Retrieve format information and select the default format if the
552	 * requested format isn't supported.
553	 */
554	info = xvip_get_format_by_fourcc(pix->pixelformat);
555	if (IS_ERR(info))
556		info = xvip_get_format_by_fourcc(XVIP_DMA_DEF_FORMAT);
557
558	pix->pixelformat = info->fourcc;
559	pix->field = V4L2_FIELD_NONE;
560
561	/* The transfer alignment requirements are expressed in bytes. Compute
562	 * the minimum and maximum values, clamp the requested width and convert
563	 * it back to pixels.
564	 */
565	align = lcm(dma->align, info->bpp);
566	min_width = roundup(XVIP_DMA_MIN_WIDTH, align);
567	max_width = rounddown(XVIP_DMA_MAX_WIDTH, align);
568	width = rounddown(pix->width * info->bpp, align);
569
570	pix->width = clamp(width, min_width, max_width) / info->bpp;
571	pix->height = clamp(pix->height, XVIP_DMA_MIN_HEIGHT,
572			    XVIP_DMA_MAX_HEIGHT);
573
574	/* Clamp the requested bytes per line value. If the maximum bytes per
575	 * line value is zero, the module doesn't support user configurable line
576	 * sizes. Override the requested value with the minimum in that case.
577	 */
578	min_bpl = pix->width * info->bpp;
579	max_bpl = rounddown(XVIP_DMA_MAX_WIDTH, dma->align);
580	bpl = rounddown(pix->bytesperline, dma->align);
581
582	pix->bytesperline = clamp(bpl, min_bpl, max_bpl);
583	pix->sizeimage = pix->bytesperline * pix->height;
584
585	if (fmtinfo)
586		*fmtinfo = info;
587}
588
589static int
590xvip_dma_try_format(struct file *file, void *fh, struct v4l2_format *format)
591{
592	struct v4l2_fh *vfh = file->private_data;
593	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
594
595	__xvip_dma_try_format(dma, &format->fmt.pix, NULL);
596	return 0;
597}
598
599static int
600xvip_dma_set_format(struct file *file, void *fh, struct v4l2_format *format)
601{
602	struct v4l2_fh *vfh = file->private_data;
603	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
604	const struct xvip_video_format *info;
605
606	__xvip_dma_try_format(dma, &format->fmt.pix, &info);
607
608	if (vb2_is_busy(&dma->queue))
609		return -EBUSY;
610
611	dma->format = format->fmt.pix;
612	dma->fmtinfo = info;
613
614	return 0;
615}
616
617static const struct v4l2_ioctl_ops xvip_dma_ioctl_ops = {
618	.vidioc_querycap		= xvip_dma_querycap,
619	.vidioc_enum_fmt_vid_cap	= xvip_dma_enum_format,
620	.vidioc_g_fmt_vid_cap		= xvip_dma_get_format,
621	.vidioc_g_fmt_vid_out		= xvip_dma_get_format,
622	.vidioc_s_fmt_vid_cap		= xvip_dma_set_format,
623	.vidioc_s_fmt_vid_out		= xvip_dma_set_format,
624	.vidioc_try_fmt_vid_cap		= xvip_dma_try_format,
625	.vidioc_try_fmt_vid_out		= xvip_dma_try_format,
626	.vidioc_reqbufs			= vb2_ioctl_reqbufs,
627	.vidioc_querybuf		= vb2_ioctl_querybuf,
628	.vidioc_qbuf			= vb2_ioctl_qbuf,
629	.vidioc_dqbuf			= vb2_ioctl_dqbuf,
630	.vidioc_create_bufs		= vb2_ioctl_create_bufs,
631	.vidioc_expbuf			= vb2_ioctl_expbuf,
632	.vidioc_streamon		= vb2_ioctl_streamon,
633	.vidioc_streamoff		= vb2_ioctl_streamoff,
634};
635
636/* -----------------------------------------------------------------------------
637 * V4L2 file operations
638 */
639
640static const struct v4l2_file_operations xvip_dma_fops = {
641	.owner		= THIS_MODULE,
642	.unlocked_ioctl	= video_ioctl2,
643	.open		= v4l2_fh_open,
644	.release	= vb2_fop_release,
645	.poll		= vb2_fop_poll,
646	.mmap		= vb2_fop_mmap,
647};
648
649/* -----------------------------------------------------------------------------
650 * Xilinx Video DMA Core
651 */
652
653int xvip_dma_init(struct xvip_composite_device *xdev, struct xvip_dma *dma,
654		  enum v4l2_buf_type type, unsigned int port)
655{
656	char name[14];
657	int ret;
658
659	dma->xdev = xdev;
660	dma->port = port;
661	mutex_init(&dma->lock);
662	mutex_init(&dma->pipe.lock);
663	INIT_LIST_HEAD(&dma->queued_bufs);
664	spin_lock_init(&dma->queued_lock);
665
666	dma->fmtinfo = xvip_get_format_by_fourcc(XVIP_DMA_DEF_FORMAT);
667	dma->format.pixelformat = dma->fmtinfo->fourcc;
668	dma->format.colorspace = V4L2_COLORSPACE_SRGB;
669	dma->format.field = V4L2_FIELD_NONE;
670	dma->format.width = XVIP_DMA_DEF_WIDTH;
671	dma->format.height = XVIP_DMA_DEF_HEIGHT;
672	dma->format.bytesperline = dma->format.width * dma->fmtinfo->bpp;
673	dma->format.sizeimage = dma->format.bytesperline * dma->format.height;
674
675	/* Initialize the media entity... */
676	dma->pad.flags = type == V4L2_BUF_TYPE_VIDEO_CAPTURE
677		       ? MEDIA_PAD_FL_SINK : MEDIA_PAD_FL_SOURCE;
678
679	ret = media_entity_init(&dma->video.entity, 1, &dma->pad, 0);
680	if (ret < 0)
681		goto error;
682
683	/* ... and the video node... */
684	dma->video.fops = &xvip_dma_fops;
685	dma->video.v4l2_dev = &xdev->v4l2_dev;
686	dma->video.queue = &dma->queue;
687	snprintf(dma->video.name, sizeof(dma->video.name), "%s %s %u",
688		 xdev->dev->of_node->name,
689		 type == V4L2_BUF_TYPE_VIDEO_CAPTURE ? "output" : "input",
690		 port);
691	dma->video.vfl_type = VFL_TYPE_GRABBER;
692	dma->video.vfl_dir = type == V4L2_BUF_TYPE_VIDEO_CAPTURE
693			   ? VFL_DIR_RX : VFL_DIR_TX;
694	dma->video.release = video_device_release_empty;
695	dma->video.ioctl_ops = &xvip_dma_ioctl_ops;
696	dma->video.lock = &dma->lock;
697
698	video_set_drvdata(&dma->video, dma);
699
700	/* ... and the buffers queue... */
701	dma->alloc_ctx = vb2_dma_contig_init_ctx(dma->xdev->dev);
702	if (IS_ERR(dma->alloc_ctx))
703		goto error;
704
705	/* Don't enable VB2_READ and VB2_WRITE, as using the read() and write()
706	 * V4L2 APIs would be inefficient. Testing on the command line with a
707	 * 'cat /dev/video?' thus won't be possible, but given that the driver
708	 * anyway requires a test tool to setup the pipeline before any video
709	 * stream can be started, requiring a specific V4L2 test tool as well
710	 * instead of 'cat' isn't really a drawback.
711	 */
712	dma->queue.type = type;
713	dma->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
714	dma->queue.lock = &dma->lock;
715	dma->queue.drv_priv = dma;
716	dma->queue.buf_struct_size = sizeof(struct xvip_dma_buffer);
717	dma->queue.ops = &xvip_dma_queue_qops;
718	dma->queue.mem_ops = &vb2_dma_contig_memops;
719	dma->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC
720				   | V4L2_BUF_FLAG_TSTAMP_SRC_EOF;
721	ret = vb2_queue_init(&dma->queue);
722	if (ret < 0) {
723		dev_err(dma->xdev->dev, "failed to initialize VB2 queue\n");
724		goto error;
725	}
726
727	/* ... and the DMA channel. */
728	sprintf(name, "port%u", port);
729	dma->dma = dma_request_slave_channel(dma->xdev->dev, name);
730	if (dma->dma == NULL) {
731		dev_err(dma->xdev->dev, "no VDMA channel found\n");
732		ret = -ENODEV;
733		goto error;
734	}
735
736	dma->align = 1 << dma->dma->device->copy_align;
737
738	ret = video_register_device(&dma->video, VFL_TYPE_GRABBER, -1);
739	if (ret < 0) {
740		dev_err(dma->xdev->dev, "failed to register video device\n");
741		goto error;
742	}
743
744	return 0;
745
746error:
747	xvip_dma_cleanup(dma);
748	return ret;
749}
750
751void xvip_dma_cleanup(struct xvip_dma *dma)
752{
753	if (video_is_registered(&dma->video))
754		video_unregister_device(&dma->video);
755
756	if (dma->dma)
757		dma_release_channel(dma->dma);
758
759	if (!IS_ERR_OR_NULL(dma->alloc_ctx))
760		vb2_dma_contig_cleanup_ctx(dma->alloc_ctx);
761
762	media_entity_cleanup(&dma->video.entity);
763
764	mutex_destroy(&dma->lock);
765	mutex_destroy(&dma->pipe.lock);
766}
767