2 * helper functions for physically contiguous capture buffers
4 * The functions support hardware lacking scatter gather support
5 * (i.e. the buffers must be linear in physical memory)
7 * Copyright (c) 2008 Magnus Damm
9 * Based on videobuf-vmalloc.c,
10 * (c) 2007 Mauro Carvalho Chehab, <mchehab@infradead.org>
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of the GNU General Public License as published by
14 * the Free Software Foundation; either version 2
17 #include <linux/init.h>
18 #include <linux/module.h>
20 #include <linux/dma-mapping.h>
21 #include <media/videobuf-dma-contig.h>
23 struct videobuf_dma_contig_memory {
26 dma_addr_t dma_handle;
30 #define MAGIC_DC_MEM 0x0733ac61
31 #define MAGIC_CHECK(is, should) \
32 if (unlikely((is) != (should))) { \
33 pr_err("magic mismatch: %x expected %x\n", is, should); \
38 videobuf_vm_open(struct vm_area_struct *vma)
40 struct videobuf_mapping *map = vma->vm_private_data;
42 dev_dbg(map->q->dev, "vm_open %p [count=%u,vma=%08lx-%08lx]\n",
43 map, map->count, vma->vm_start, vma->vm_end);
48 static void videobuf_vm_close(struct vm_area_struct *vma)
50 struct videobuf_mapping *map = vma->vm_private_data;
51 struct videobuf_queue *q = map->q;
54 dev_dbg(map->q->dev, "vm_close %p [count=%u,vma=%08lx-%08lx]\n",
55 map, map->count, vma->vm_start, vma->vm_end);
58 if (0 == map->count) {
59 struct videobuf_dma_contig_memory *mem;
61 dev_dbg(map->q->dev, "munmap %p q=%p\n", map, q);
62 mutex_lock(&q->vb_lock);
64 /* We need first to cancel streams, before unmapping */
66 videobuf_queue_cancel(q);
68 for (i = 0; i < VIDEO_MAX_FRAME; i++) {
69 if (NULL == q->bufs[i])
72 if (q->bufs[i]->map != map)
75 mem = q->bufs[i]->priv;
77 /* This callback is called only if kernel has
78 allocated memory and this memory is mmapped.
79 In this case, memory should be freed,
80 in order to do memory unmap.
83 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
85 /* vfree is not atomic - can't be
86 called with IRQ's disabled
88 dev_dbg(map->q->dev, "buf[%d] freeing %p\n",
91 dma_free_coherent(q->dev, mem->size,
92 mem->vaddr, mem->dma_handle);
96 q->bufs[i]->map = NULL;
97 q->bufs[i]->baddr = 0;
102 mutex_unlock(&q->vb_lock);
106 static struct vm_operations_struct videobuf_vm_ops = {
107 .open = videobuf_vm_open,
108 .close = videobuf_vm_close,
111 static void *__videobuf_alloc(size_t size)
113 struct videobuf_dma_contig_memory *mem;
114 struct videobuf_buffer *vb;
116 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL);
118 mem = vb->priv = ((char *)vb) + size;
119 mem->magic = MAGIC_DC_MEM;
125 static void *__videobuf_to_vmalloc(struct videobuf_buffer *buf)
127 struct videobuf_dma_contig_memory *mem = buf->priv;
130 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
135 static int __videobuf_iolock(struct videobuf_queue *q,
136 struct videobuf_buffer *vb,
137 struct v4l2_framebuffer *fbuf)
139 struct videobuf_dma_contig_memory *mem = vb->priv;
142 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
144 switch (vb->memory) {
145 case V4L2_MEMORY_MMAP:
146 dev_dbg(q->dev, "%s memory method MMAP\n", __func__);
148 /* All handling should be done by __videobuf_mmap_mapper() */
150 dev_err(q->dev, "memory is not alloced/mmapped.\n");
154 case V4L2_MEMORY_USERPTR:
155 dev_dbg(q->dev, "%s memory method USERPTR\n", __func__);
157 /* The only USERPTR currently supported is the one needed for
163 mem->size = PAGE_ALIGN(vb->size);
164 mem->vaddr = dma_alloc_coherent(q->dev, mem->size,
165 &mem->dma_handle, GFP_KERNEL);
167 dev_err(q->dev, "dma_alloc_coherent %ld failed\n",
172 dev_dbg(q->dev, "dma_alloc_coherent data is at %p (%ld)\n",
173 mem->vaddr, mem->size);
175 case V4L2_MEMORY_OVERLAY:
177 dev_dbg(q->dev, "%s memory method OVERLAY/unknown\n",
185 static int __videobuf_sync(struct videobuf_queue *q,
186 struct videobuf_buffer *buf)
188 struct videobuf_dma_contig_memory *mem = buf->priv;
191 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
193 dma_sync_single_for_cpu(q->dev, mem->dma_handle, mem->size,
198 static int __videobuf_mmap_free(struct videobuf_queue *q)
202 dev_dbg(q->dev, "%s\n", __func__);
203 for (i = 0; i < VIDEO_MAX_FRAME; i++) {
204 if (q->bufs[i] && q->bufs[i]->map)
211 static int __videobuf_mmap_mapper(struct videobuf_queue *q,
212 struct vm_area_struct *vma)
214 struct videobuf_dma_contig_memory *mem;
215 struct videobuf_mapping *map;
218 unsigned long size, offset = vma->vm_pgoff << PAGE_SHIFT;
220 dev_dbg(q->dev, "%s\n", __func__);
221 if (!(vma->vm_flags & VM_WRITE) || !(vma->vm_flags & VM_SHARED))
224 /* look for first buffer to map */
225 for (first = 0; first < VIDEO_MAX_FRAME; first++) {
229 if (V4L2_MEMORY_MMAP != q->bufs[first]->memory)
231 if (q->bufs[first]->boff == offset)
234 if (VIDEO_MAX_FRAME == first) {
235 dev_dbg(q->dev, "invalid user space offset [offset=0x%lx]\n",
240 /* create mapping + update buffer list */
241 map = kzalloc(sizeof(struct videobuf_mapping), GFP_KERNEL);
245 q->bufs[first]->map = map;
246 map->start = vma->vm_start;
247 map->end = vma->vm_end;
250 q->bufs[first]->baddr = vma->vm_start;
252 mem = q->bufs[first]->priv;
254 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
256 mem->size = PAGE_ALIGN(q->bufs[first]->bsize);
257 mem->vaddr = dma_alloc_coherent(q->dev, mem->size,
258 &mem->dma_handle, GFP_KERNEL);
260 dev_err(q->dev, "dma_alloc_coherent size %ld failed\n",
264 dev_dbg(q->dev, "dma_alloc_coherent data is at addr %p (size %ld)\n",
265 mem->vaddr, mem->size);
267 /* Try to remap memory */
269 size = vma->vm_end - vma->vm_start;
270 size = (size < mem->size) ? size : mem->size;
272 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
273 retval = remap_pfn_range(vma, vma->vm_start,
274 mem->dma_handle >> PAGE_SHIFT,
275 size, vma->vm_page_prot);
277 dev_err(q->dev, "mmap: remap failed with error %d. ", retval);
278 dma_free_coherent(q->dev, mem->size,
279 mem->vaddr, mem->dma_handle);
283 vma->vm_ops = &videobuf_vm_ops;
284 vma->vm_flags |= VM_DONTEXPAND;
285 vma->vm_private_data = map;
287 dev_dbg(q->dev, "mmap %p: q=%p %08lx-%08lx (%lx) pgoff %08lx buf %d\n",
288 map, q, vma->vm_start, vma->vm_end,
289 (long int) q->bufs[first]->bsize,
290 vma->vm_pgoff, first);
292 videobuf_vm_open(vma);
301 static int __videobuf_copy_to_user(struct videobuf_queue *q,
302 char __user *data, size_t count,
305 struct videobuf_dma_contig_memory *mem = q->read_buf->priv;
309 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
312 /* copy to userspace */
313 if (count > q->read_buf->size - q->read_off)
314 count = q->read_buf->size - q->read_off;
318 if (copy_to_user(data, vaddr + q->read_off, count))
324 static int __videobuf_copy_stream(struct videobuf_queue *q,
325 char __user *data, size_t count, size_t pos,
326 int vbihack, int nonblocking)
329 struct videobuf_dma_contig_memory *mem = q->read_buf->priv;
332 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
335 /* dirty, undocumented hack -- pass the frame counter
336 * within the last four bytes of each vbi data block.
337 * We need that one to maintain backward compatibility
338 * to all vbi decoding software out there ... */
339 fc = (unsigned int *)mem->vaddr;
340 fc += (q->read_buf->size >> 2) - 1;
341 *fc = q->read_buf->field_count >> 1;
342 dev_dbg(q->dev, "vbihack: %d\n", *fc);
345 /* copy stuff using the common method */
346 count = __videobuf_copy_to_user(q, data, count, nonblocking);
348 if ((count == -EFAULT) && (pos == 0))
354 static struct videobuf_qtype_ops qops = {
355 .magic = MAGIC_QTYPE_OPS,
357 .alloc = __videobuf_alloc,
358 .iolock = __videobuf_iolock,
359 .sync = __videobuf_sync,
360 .mmap_free = __videobuf_mmap_free,
361 .mmap_mapper = __videobuf_mmap_mapper,
362 .video_copy_to_user = __videobuf_copy_to_user,
363 .copy_stream = __videobuf_copy_stream,
364 .vmalloc = __videobuf_to_vmalloc,
367 void videobuf_queue_dma_contig_init(struct videobuf_queue *q,
368 struct videobuf_queue_ops *ops,
371 enum v4l2_buf_type type,
372 enum v4l2_field field,
376 videobuf_queue_core_init(q, ops, dev, irqlock, type, field, msize,
379 EXPORT_SYMBOL_GPL(videobuf_queue_dma_contig_init);
381 dma_addr_t videobuf_to_dma_contig(struct videobuf_buffer *buf)
383 struct videobuf_dma_contig_memory *mem = buf->priv;
386 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
388 return mem->dma_handle;
390 EXPORT_SYMBOL_GPL(videobuf_to_dma_contig);
392 void videobuf_dma_contig_free(struct videobuf_queue *q,
393 struct videobuf_buffer *buf)
395 struct videobuf_dma_contig_memory *mem = buf->priv;
397 /* mmapped memory can't be freed here, otherwise mmapped region
398 would be released, while still needed. In this case, the memory
399 release should happen inside videobuf_vm_close().
400 So, it should free memory only if the memory were allocated for
403 if ((buf->memory != V4L2_MEMORY_USERPTR) || !buf->baddr)
409 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM);
411 dma_free_coherent(q->dev, mem->size, mem->vaddr, mem->dma_handle);
414 EXPORT_SYMBOL_GPL(videobuf_dma_contig_free);
416 MODULE_DESCRIPTION("helper module to manage video4linux dma contig buffers");
417 MODULE_AUTHOR("Magnus Damm");
418 MODULE_LICENSE("GPL");