Searched refs:buffers (Results 1 - 200 of 2085) sorted by relevance

1234567891011

/linux-4.1.27/lib/xz/
H A Dxz_dec_test.c42 * Input and output buffers. The input buffer is used as a temporary safe
49 * Structure to pass the input and output buffers to the XZ decoder.
52 static struct xz_buf buffers = { variable in typeref:struct:xz_buf
75 buffers.in_pos = 0; xz_dec_test_open()
76 buffers.in_size = 0; xz_dec_test_open()
77 buffers.out_pos = 0; xz_dec_test_open()
120 while ((remaining > 0 || buffers.out_pos == buffers.out_size) xz_dec_test_write()
122 if (buffers.in_pos == buffers.in_size) { xz_dec_test_write()
123 buffers.in_pos = 0; xz_dec_test_write()
124 buffers.in_size = min(remaining, sizeof(buffer_in)); xz_dec_test_write()
125 if (copy_from_user(buffer_in, buf, buffers.in_size)) xz_dec_test_write()
128 buf += buffers.in_size; xz_dec_test_write()
129 remaining -= buffers.in_size; xz_dec_test_write()
132 buffers.out_pos = 0; xz_dec_test_write()
133 ret = xz_dec_run(state, &buffers); xz_dec_test_write()
134 crc = crc32(crc, buffer_out, buffers.out_pos); xz_dec_test_write()
145 return size - remaining - (buffers.in_size - buffers.in_pos); xz_dec_test_write()
/linux-4.1.27/drivers/media/pci/ivtv/
H A Divtv-queue.c47 q->buffers = 0; ivtv_queue_init()
65 q->buffers++; ivtv_enqueue()
80 q->buffers--; ivtv_dequeue()
94 from->buffers--; ivtv_queue_move_buf()
100 to->buffers++; ivtv_queue_move_buf()
105 /* Move 'needed_bytes' worth of buffers from queue 'from' into queue 'to'.
106 If 'needed_bytes' == 0, then move all buffers from 'from' into 'to'.
107 If 'steal' != NULL, then buffers may also taken from that queue if
111 also cleared if buffers need to be taken from the 'steal' queue and
119 -ENOMEM is returned if the buffers could not be obtained, 0 if all
120 buffers where obtained from the 'from' list and if non-zero then
121 the number of stolen buffers is returned. */ ivtv_queue_move()
148 /* move buffers from the tail of the 'steal' queue to the tail of the ivtv_queue_move()
149 'from' queue. Always copy all the buffers with the same dma_xfer_cnt ivtv_queue_move()
151 if one frame is stored in multiple buffers. */ ivtv_queue_move()
155 steal->buffers--; ivtv_queue_move()
159 from->buffers++; ivtv_queue_move()
196 int SGsize = sizeof(struct ivtv_sg_host_element) * s->buffers; ivtv_stream_alloc()
199 if (s->buffers == 0) ivtv_stream_alloc()
202 IVTV_DEBUG_INFO("Allocate %s%s stream: %d x %d buffers (%dkB total)\n", ivtv_stream_alloc()
204 s->name, s->buffers, s->buf_size, s->buffers * s->buf_size / 1024); ivtv_stream_alloc()
238 /* allocate stream buffers. Initially all buffers are in q_free. */ ivtv_stream_alloc()
239 for (i = 0; i < s->buffers; i++) { ivtv_stream_alloc()
258 if (i == s->buffers) ivtv_stream_alloc()
260 IVTV_ERR("Couldn't allocate buffers for %s stream\n", s->name); ivtv_stream_alloc()
269 /* move all buffers to q_free */ ivtv_stream_free()
/linux-4.1.27/drivers/media/platform/vivid/
H A Dvivid-vid-common.c54 .buffers = 1,
64 .buffers = 1,
73 .buffers = 1,
82 .buffers = 1,
91 .buffers = 1,
100 .buffers = 1,
109 .buffers = 1,
118 .buffers = 1,
127 .buffers = 1,
136 .buffers = 1,
145 .buffers = 1,
154 .buffers = 1,
163 .buffers = 1,
171 .buffers = 1,
180 .buffers = 1,
188 .buffers = 1,
197 .buffers = 1,
207 .buffers = 1,
215 .buffers = 1,
223 .buffers = 1,
232 .buffers = 1,
241 .buffers = 1,
249 .buffers = 1,
257 .buffers = 1,
266 .buffers = 1,
275 .buffers = 1,
284 .buffers = 1,
294 .buffers = 1,
302 .buffers = 1,
310 .buffers = 1,
319 .buffers = 1,
327 .buffers = 1,
335 .buffers = 1,
343 .buffers = 1,
351 .buffers = 1,
359 .buffers = 1,
367 .buffers = 1,
375 .buffers = 1,
384 .buffers = 1,
393 .buffers = 1,
401 .buffers = 1,
409 .buffers = 1,
417 .buffers = 1,
426 .buffers = 2,
436 .buffers = 2,
446 .buffers = 3,
455 .buffers = 3,
464 .buffers = 2,
473 .buffers = 2,
488 if (fmt->buffers == 1 || dev->multiplanar) vivid_get_format()
/linux-4.1.27/include/media/
H A Dvideobuf-dma-contig.h2 * helper functions for physically contiguous capture buffers
5 * (i.e. the buffers must be linear in physical memory)
H A Dvideobuf-dma-sg.h2 * helper functions for SG DMA video4linux capture buffers
5 * (i.e. the buffers are not linear in physical memory, but fragmented
28 * A small set of helper functions to manage buffers (both userland
54 /* for kernel buffers */
60 /* for overlay buffers (pci-pci dma) */
73 /* for mmap'ed buffers */
H A Dvideobuf-vmalloc.h2 * helper functions for vmalloc capture buffers
5 * (i.e. the buffers are not linear in physical memory, but fragmented
H A Dv4l2-mem2mem.h4 * Helper functions for devices that use memory buffers for both source
58 /* Queue for buffers ready to be processed as soon as this
160 * v4l2_m2m_num_src_bufs_ready() - return the number of source buffers ready for
170 * v4l2_m2m_num_src_bufs_ready() - return the number of destination buffers
183 * buffers
192 * ready buffers
200 * v4l2_m2m_get_src_vq() - return vb2_queue for source buffers
209 * v4l2_m2m_get_dst_vq() - return vb2_queue for destination buffers
221 * buffers and return it
230 * ready buffers and return it
H A Dvideobuf2-core.h179 * @queued_entry: entry on the queued buffers list, which holds all
180 * buffers queued from userspace
181 * @done_entry: entry on the list that stores all buffers ready to
239 * smaller number of buffers. Driver should return
240 * the required number of buffers in *num_buffers, the
247 * of buffers, that are being allocated. When called from
280 * buffers are being returned to their default DEQUEUED
287 * receive buffers with @buf_queue callback before
289 * of already queued buffers in count parameter; driver
291 * buffers that have been already given by the @buf_queue
294 * If you need a minimum number of buffers before you can
298 * many buffers have been queued up by userspace.
301 * finish and give back all buffers it got from buf_queue()
310 * pre-queued buffers before calling STREAMON.
346 * @owner: The filehandle that 'owns' the buffers, i.e. the filehandle
358 * @gfp_flags: additional gfp flags used when allocating the buffers.
361 * @min_buffers_needed: the minimum number of buffers needed before
363 * cannot be started unless at least this number of buffers
366 * @mmap_lock: private mutex used when buffers are allocated/freed/mmapped
369 * @num_buffers: number of allocated/used buffers
370 * @queued_list: list of buffers currently queued from userspace
371 * @queued_count: number of buffers queued and ready for streaming.
372 * @owned_by_drv_count: number of buffers owned by the driver
373 * @done_list: list of buffers ready to be dequeued to userspace
375 * @done_wq: waitqueue for processes waiting for buffers ready to be dequeued
382 * buffers. Only set for capture queues if qbuf has not yet been
541 * This function checks if queue has any buffers allocated.
H A Dvideobuf-core.h2 * generic helper functions for handling video4linux capture buffers
31 * A small set of helper functions to manage video4linux buffers.
34 * functions, additionally some commonly used fields for v4l buffers
40 * The mmaper function allows to map any subset of contingous buffers.
41 * This includes one mmap() call for all buffers (which the original
97 /* for mmap'ed buffers */
/linux-4.1.27/drivers/media/pci/zoran/
H A Dzoran_driver.c195 fh->buffers.buffer_size = v4l_bufsize; map_mode_raw()
196 fh->buffers.num_buffers = v4l_nbufs; map_mode_raw()
201 fh->buffers.buffer_size = jpg_bufsize; map_mode_jpg()
202 fh->buffers.num_buffers = jpg_nbufs; map_mode_jpg()
210 * Allocate the V4L grab buffers
221 for (i = 0; i < fh->buffers.num_buffers; i++) { v4l_fbuffer_alloc()
222 if (fh->buffers.buffer[i].v4l.fbuffer) v4l_fbuffer_alloc()
229 mem = kmalloc(fh->buffers.buffer_size, v4l_fbuffer_alloc()
239 fh->buffers.buffer[i].v4l.fbuffer = mem; v4l_fbuffer_alloc()
240 fh->buffers.buffer[i].v4l.fbuffer_phys = virt_to_phys(mem); v4l_fbuffer_alloc()
241 fh->buffers.buffer[i].v4l.fbuffer_bus = virt_to_bus(mem); v4l_fbuffer_alloc()
242 for (off = 0; off < fh->buffers.buffer_size; v4l_fbuffer_alloc()
252 fh->buffers.allocated = 1; v4l_fbuffer_alloc()
257 /* free the V4L grab buffers */ v4l_fbuffer_free()
266 for (i = 0; i < fh->buffers.num_buffers; i++) { v4l_fbuffer_free()
267 if (!fh->buffers.buffer[i].v4l.fbuffer) v4l_fbuffer_free()
270 mem = fh->buffers.buffer[i].v4l.fbuffer; v4l_fbuffer_free()
271 for (off = 0; off < fh->buffers.buffer_size; v4l_fbuffer_free()
274 kfree(fh->buffers.buffer[i].v4l.fbuffer); v4l_fbuffer_free()
275 fh->buffers.buffer[i].v4l.fbuffer = NULL; v4l_fbuffer_free()
278 fh->buffers.allocated = 0; v4l_fbuffer_free()
282 * Allocate the MJPEG grab buffers.
289 * RJ: The contents grab buffers needs never be accessed in the driver.
300 * and fragment buffers, the device expects to see little-
306 * and fragment buffers are not little-endian.
315 for (i = 0; i < fh->buffers.num_buffers; i++) { jpg_fbuffer_alloc()
316 if (fh->buffers.buffer[i].jpg.frag_tab) jpg_fbuffer_alloc()
333 fh->buffers.buffer[i].jpg.frag_tab = (__le32 *)mem; jpg_fbuffer_alloc()
334 fh->buffers.buffer[i].jpg.frag_tab_bus = virt_to_bus(mem); jpg_fbuffer_alloc()
336 if (fh->buffers.need_contiguous) { jpg_fbuffer_alloc()
337 mem = kmalloc(fh->buffers.buffer_size, GFP_KERNEL); jpg_fbuffer_alloc()
346 fh->buffers.buffer[i].jpg.frag_tab[0] = jpg_fbuffer_alloc()
348 fh->buffers.buffer[i].jpg.frag_tab[1] = jpg_fbuffer_alloc()
349 cpu_to_le32((fh->buffers.buffer_size >> 1) | 1); jpg_fbuffer_alloc()
350 for (off = 0; off < fh->buffers.buffer_size; off += PAGE_SIZE) jpg_fbuffer_alloc()
354 for (j = 0; j < fh->buffers.buffer_size / PAGE_SIZE; j++) { jpg_fbuffer_alloc()
365 fh->buffers.buffer[i].jpg.frag_tab[2 * j] = jpg_fbuffer_alloc()
367 fh->buffers.buffer[i].jpg.frag_tab[2 * j + 1] = jpg_fbuffer_alloc()
372 fh->buffers.buffer[i].jpg.frag_tab[2 * j - 1] |= cpu_to_le32(1); jpg_fbuffer_alloc()
379 (fh->buffers.num_buffers * fh->buffers.buffer_size) >> 10); jpg_fbuffer_alloc()
381 fh->buffers.allocated = 1; jpg_fbuffer_alloc()
386 /* free the MJPEG grab buffers */ jpg_fbuffer_free()
397 for (i = 0, buffer = &fh->buffers.buffer[0]; jpg_fbuffer_free()
398 i < fh->buffers.num_buffers; i++, buffer++) { jpg_fbuffer_free()
402 if (fh->buffers.need_contiguous) { jpg_fbuffer_free()
407 for (off = 0; off < fh->buffers.buffer_size; off += PAGE_SIZE) jpg_fbuffer_free()
414 for (j = 0; j < fh->buffers.buffer_size / PAGE_SIZE; j++) { jpg_fbuffer_free()
430 fh->buffers.allocated = 0; jpg_fbuffer_free()
460 if (height * width * bpp > fh->buffers.buffer_size) { zoran_v4l_set_format()
464 ZR_DEVNAME(zr), __func__, fh->buffers.buffer_size >> 10); zoran_v4l_set_format()
492 if (!fh->buffers.allocated) { zoran_v4l_queue_frame()
495 "%s: %s - buffers not yet allocated\n", zoran_v4l_queue_frame()
501 if (num >= fh->buffers.num_buffers || num < 0) { zoran_v4l_queue_frame()
511 if (fh->buffers.active == ZORAN_FREE) { zoran_v4l_queue_frame()
513 zr->v4l_buffers = fh->buffers; zoran_v4l_queue_frame()
514 fh->buffers.active = ZORAN_ACTIVE; zoran_v4l_queue_frame()
530 fh->buffers.active = ZORAN_FREE; zoran_v4l_queue_frame()
548 fh->buffers.buffer[num] = zr->v4l_buffers.buffer[num]; zoran_v4l_queue_frame()
556 zr->v4l_buffers.active = fh->buffers.active; zoran_v4l_queue_frame()
570 if (fh->buffers.active == ZORAN_FREE) { v4l_sync()
579 if (frame >= fh->buffers.num_buffers || frame < 0) { v4l_sync()
609 fh->buffers.buffer[frame] = zr->v4l_buffers.buffer[frame]; v4l_sync()
617 fh->buffers.active = zr->v4l_buffers.active = ZORAN_FREE; v4l_sync()
638 /* Check if buffers are allocated */ zoran_jpg_queue_frame()
639 if (!fh->buffers.allocated) { zoran_jpg_queue_frame()
642 "%s: %s - buffers not yet allocated\n", zoran_jpg_queue_frame()
648 if (num >= fh->buffers.num_buffers || num < 0) { zoran_jpg_queue_frame()
668 if (fh->buffers.active == ZORAN_FREE) { zoran_jpg_queue_frame()
670 zr->jpg_buffers = fh->buffers; zoran_jpg_queue_frame()
671 fh->buffers.active = ZORAN_ACTIVE; zoran_jpg_queue_frame()
700 fh->buffers.buffer[num] = zr->jpg_buffers.buffer[num]; zoran_jpg_queue_frame()
707 fh->buffers.active = ZORAN_FREE; zoran_jpg_queue_frame()
718 zr->jpg_buffers.active = fh->buffers.active; zoran_jpg_queue_frame()
731 if (fh->buffers.active == ZORAN_FREE) { jpg_qbuf()
738 fh->buffers.active = zr->jpg_buffers.active = ZORAN_FREE; jpg_qbuf()
771 if (fh->buffers.active == ZORAN_FREE) { jpg_sync()
823 fh->buffers.buffer[frame] = zr->jpg_buffers.buffer[frame]; jpg_sync()
849 /* buffers */ zoran_open_init_session()
850 memset(&fh->buffers, 0, sizeof(fh->buffers)); zoran_open_init_session()
852 fh->buffers.buffer[i].state = BUZ_STATE_USER; /* nothing going on */ zoran_open_init_session()
853 fh->buffers.buffer[i].bs.frame = i; zoran_open_init_session()
855 fh->buffers.allocated = 0; zoran_open_init_session()
856 fh->buffers.active = ZORAN_FREE; zoran_open_init_session()
874 if (fh->buffers.active != ZORAN_FREE) { zoran_close_end_session()
880 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_close_end_session()
884 /* v4l buffers */ zoran_close_end_session()
885 if (fh->buffers.allocated) zoran_close_end_session()
889 if (fh->buffers.active != ZORAN_FREE) { zoran_close_end_session()
892 zr->jpg_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_close_end_session()
895 /* jpg buffers */ zoran_close_end_session()
896 if (fh->buffers.allocated) zoran_close_end_session()
1342 if (num < 0 || num >= fh->buffers.num_buffers || zoran_v4l2_buffer_status()
1343 !fh->buffers.allocated) { zoran_v4l2_buffer_status()
1346 "%s: %s - wrong number or buffers not allocated\n", zoran_v4l2_buffer_status()
1356 "FAL"[fh->buffers.active], num, zoran_v4l2_buffer_status()
1358 fh->buffers.buffer[num].map ? 'Y' : 'N'); zoran_v4l2_buffer_status()
1362 buf->length = fh->buffers.buffer_size; zoran_v4l2_buffer_status()
1365 buf->bytesused = fh->buffers.buffer[num].bs.length; zoran_v4l2_buffer_status()
1366 if (fh->buffers.buffer[num].state == BUZ_STATE_DONE || zoran_v4l2_buffer_status()
1367 fh->buffers.buffer[num].state == BUZ_STATE_USER) { zoran_v4l2_buffer_status()
1368 buf->sequence = fh->buffers.buffer[num].bs.seq; zoran_v4l2_buffer_status()
1370 buf->timestamp = fh->buffers.buffer[num].bs.timestamp; zoran_v4l2_buffer_status()
1386 if (num < 0 || num >= fh->buffers.num_buffers || zoran_v4l2_buffer_status()
1387 !fh->buffers.allocated) { zoran_v4l2_buffer_status()
1390 "%s: %s - wrong number or buffers not allocated\n", zoran_v4l2_buffer_status()
1398 buf->length = fh->buffers.buffer_size; zoran_v4l2_buffer_status()
1401 if (fh->buffers.buffer[num].state == BUZ_STATE_DONE || zoran_v4l2_buffer_status()
1402 fh->buffers.buffer[num].state == BUZ_STATE_USER) { zoran_v4l2_buffer_status()
1403 buf->sequence = fh->buffers.buffer[num].bs.seq; zoran_v4l2_buffer_status()
1404 buf->timestamp = fh->buffers.buffer[num].bs.timestamp; zoran_v4l2_buffer_status()
1405 buf->bytesused = fh->buffers.buffer[num].bs.length; zoran_v4l2_buffer_status()
1813 if (fh->buffers.allocated) { zoran_s_fmt_vid_out()
1860 fh->buffers.buffer_size = zoran_v4l2_calc_bufsize(&fh->jpg_settings); zoran_s_fmt_vid_out()
1873 fmt->fmt.pix.sizeimage = fh->buffers.buffer_size; zoran_s_fmt_vid_out()
1903 if ((fh->map_mode != ZORAN_MAP_MODE_RAW && fh->buffers.allocated) || zoran_s_fmt_vid_cap()
1904 fh->buffers.active != ZORAN_FREE) { zoran_s_fmt_vid_cap()
2017 if (fh->buffers.allocated) { zoran_reqbufs()
2020 "%s: VIDIOC_REQBUFS - buffers already allocated\n", zoran_reqbufs()
2034 /* The next mmap will map the V4L buffers */ zoran_reqbufs()
2036 fh->buffers.num_buffers = req->count; zoran_reqbufs()
2050 /* The next mmap will map the MJPEG buffers */ zoran_reqbufs()
2052 fh->buffers.num_buffers = req->count; zoran_reqbufs()
2053 fh->buffers.buffer_size = zoran_v4l2_calc_bufsize(&fh->jpg_settings); zoran_reqbufs()
2107 if (!zr->v4l_memgrab_active && fh->buffers.active == ZORAN_LOCKED) zoran_qbuf()
2133 fh->buffers.active == ZORAN_LOCKED) zoran_qbuf()
2239 fh->buffers.active != ZORAN_ACTIVE) { zoran_streamon()
2244 zr->v4l_buffers.active = fh->buffers.active = ZORAN_LOCKED; zoran_streamon()
2258 fh->buffers.active != ZORAN_ACTIVE) { zoran_streamon()
2263 zr->jpg_buffers.active = fh->buffers.active = ZORAN_LOCKED; zoran_streamon()
2296 if (fh->buffers.active == ZORAN_FREE && zoran_streamoff()
2311 for (i = 0; i < fh->buffers.num_buffers; i++) zoran_streamoff()
2313 fh->buffers = zr->v4l_buffers; zoran_streamoff()
2315 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_streamoff()
2327 if (fh->buffers.active == ZORAN_FREE && zoran_streamoff()
2590 if (fh->buffers.allocated) { zoran_s_crop()
2668 if (fh->buffers.active != ZORAN_FREE) { zoran_s_jpegcomp()
2679 if (!fh->buffers.allocated) zoran_s_jpegcomp()
2680 fh->buffers.buffer_size = zoran_s_jpegcomp()
2698 /* we should check whether buffers are ready to be synced on zoran_poll()
2703 * if no buffers queued or so, return POLLNVAL zoran_poll()
2718 "FAL"[fh->buffers.active], zr->v4l_sync_tail, zoran_poll()
2722 if (fh->buffers.active != ZORAN_FREE && zoran_poll()
2740 "FAL"[fh->buffers.active], zr->jpg_que_tail, zoran_poll()
2743 if (fh->buffers.active != ZORAN_FREE && zoran_poll()
2769 * This maps the buffers to user space.
2772 * the V4L or the MJPEG buffers are mapped
2801 for (i = 0; i < fh->buffers.num_buffers; i++) { zoran_vm_close()
2802 if (fh->buffers.buffer[i].map == map) zoran_vm_close()
2803 fh->buffers.buffer[i].map = NULL; zoran_vm_close()
2807 /* Any buffers still mapped? */ zoran_vm_close()
2808 for (i = 0; i < fh->buffers.num_buffers; i++) { zoran_vm_close()
2809 if (fh->buffers.buffer[i].map) { zoran_vm_close()
2815 dprintk(3, KERN_INFO "%s: %s - free %s buffers\n", ZR_DEVNAME(zr), zoran_vm_close()
2820 if (fh->buffers.active != ZORAN_FREE) { zoran_vm_close()
2826 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_vm_close()
2831 if (fh->buffers.active != ZORAN_FREE) { zoran_vm_close()
2834 zr->jpg_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_vm_close()
2877 if (!fh->buffers.allocated) { zoran_mmap()
2880 "%s: %s(%s) - buffers not yet allocated\n", zoran_mmap()
2886 first = offset / fh->buffers.buffer_size; zoran_mmap()
2887 last = first - 1 + size / fh->buffers.buffer_size; zoran_mmap()
2888 if (offset % fh->buffers.buffer_size != 0 || zoran_mmap()
2889 size % fh->buffers.buffer_size != 0 || first < 0 || zoran_mmap()
2890 last < 0 || first >= fh->buffers.num_buffers || zoran_mmap()
2891 last >= fh->buffers.buffer_size) { zoran_mmap()
2896 fh->buffers.buffer_size, zoran_mmap()
2897 fh->buffers.num_buffers); zoran_mmap()
2902 /* Check if any buffers are already mapped */ zoran_mmap()
2904 if (fh->buffers.buffer[i].map) { zoran_mmap()
2914 /* map these buffers */ zoran_mmap()
2930 if (todo > fh->buffers.buffer_size) zoran_mmap()
2931 todo = fh->buffers.buffer_size; zoran_mmap()
2932 page = fh->buffers.buffer[i].v4l.fbuffer_phys; zoran_mmap()
2944 fh->buffers.buffer[i].map = map; zoran_mmap()
2951 j < fh->buffers.buffer_size / PAGE_SIZE; zoran_mmap()
2954 (le32_to_cpu(fh->buffers.buffer[i].jpg. zoran_mmap()
2960 le32_to_cpu(fh->buffers. zoran_mmap()
2978 if (le32_to_cpu(fh->buffers.buffer[i].jpg. zoran_mmap()
2982 fh->buffers.buffer[i].map = map; zoran_mmap()
H A Dzoran.h66 # error "Too many video frame buffers to handle"
209 struct zoran_buffer buffer[MAX_FRAME]; /* buffers */
210 u8 allocated; /* Flag if buffers are allocated */
211 u8 need_contiguous; /* Flag if contiguous buffers are needed */
212 /* only applies to jpg buffers, raw buffers are always contiguous */
227 struct zoran_buffer_col buffers; /* buffers' info */ member in struct:zoran_fh
328 struct zoran_buffer_col v4l_buffers; /* V4L buffers' info */
355 struct zoran_buffer_col jpg_buffers; /* MJPEG buffers' info */
/linux-4.1.27/drivers/staging/speakup/
H A DMakefile19 buffers.o \
/linux-4.1.27/arch/metag/include/asm/
H A Dcache.h13 * buffers have cache line alignment.
H A Dfixmap.h32 * these 'compile-time allocated' memory buffers are
37 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/drivers/net/wan/
H A Dx25_asy.h25 /* These are pointers to the malloc()ed frame buffers. */
31 int buffsize; /* Max buffers sizes */
/linux-4.1.27/drivers/staging/octeon/
H A Dethernet-mem.c42 * @elements: Number of buffers to allocate
44 * Returns the actual number of buffers allocated.
67 * @elements: Number of buffers to allocate
95 * @elements: Number of buffers to allocate
97 * Returns the actual number of buffers allocated.
134 * @elements: Number of buffers that should be in the pool
152 pr_warn("Freeing of pool %u had too many buffers (%d)\n", cvm_oct_free_hw_memory()
155 pr_warn("Warning: Freeing of pool %u is missing %d buffers\n", cvm_oct_free_hw_memory()
H A Dethernet-defines.h34 * allows 32bit userspace application to access the buffers, but also
37 * Tells the driver to populate the packet buffers with kernel skbuffs.
39 * means that 32bit userspace can't access the packet buffers.
H A Docteon-ethernet.h45 /* Hardware fetch and add to count outstanding tx buffers */
54 /* List of outstanding tx buffers per queue */
/linux-4.1.27/include/drm/ttm/
H A Dttm_execbuf_util.h74 * If the function returns 0, all buffers are marked as "unfenced",
78 * reserve the same buffers in reverse order, all threads except one will
81 * unreserve their buffers.
85 * buffers on the list will be reserved upon return.
87 * If dups is non NULL all buffers already reserved by the current thread
89 * on the first already reserved buffer and all buffers from the list are
107 * @fence: The new exclusive fence for the buffers.
111 * It also unreserves all buffers, putting them on lru lists.
H A Dttm_bo_api.h132 * @ttm_bo_type_device: These are 'normal' buffers that can
136 * @ttm_bo_type_kernel: These buffers are like ttm_bo_type_device buffers,
167 * @persistent_swap_storage: Usually the swap storage is deleted for buffers
457 * @persistent_swap_storage: Usually the swap storage is deleted for buffers
504 * @persistent_swap_storage: Usually the swap storage is deleted for buffers
553 * the LRU list to evict any buffers left alive.
556 * point there shouldn't be any buffers left created by user-space, since
560 * make sure that there are no NO_EVICT buffers present in this memory type
564 * there are no appications currently validating buffers before this
570 * -EBUSY: There are still buffers left in this memory type.
581 * Evicts all buffers on the lru list of the memory type.
585 * currently validating buffers, and can do that by taking the
/linux-4.1.27/drivers/staging/iio/
H A Dring_hw.h2 * ring_hw.h - common functionality for iio hardware ring buffers
/linux-4.1.27/arch/s390/include/uapi/asm/
H A Dvirtio-ccw.h15 /* Alignment of vring buffers. */
/linux-4.1.27/include/linux/
H A Djournal-head.h63 * transaction (if there is one). Only applies to buffers on a
80 * Doubly-linked list of buffers on a transaction's data, metadata or
87 * is checkpointed. Only dirty buffers can be checkpointed.
93 * Doubly-linked list of buffers still remaining to be flushed
H A Dpipe_fs_i.h32 * @nrbufs: the number of non-empty pipe buffers in this pipe
33 * @buffers: total number of buffers (should be a power of 2)
44 * @bufs: the circular array of pipe buffers
49 unsigned int nrbufs, curbuf, buffers; member in struct:pipe_inode_info
H A Djbd.h89 * The buffer credits field is used to account for journaled buffers
92 * number of outstanding buffers possible at any time. When the
94 * the transaction, so that at all times we know how many buffers the
300 * @h_buffer_credits: Number of remaining buffers we are allowed to dirty.
313 /* Number of remaining buffers we are allowed to dirty: */
340 * new buffers to modify (state not used for now)
345 * The transaction keeps track of all of the buffers modified by a
346 * running transaction, and all of the buffers committed but not yet
400 /* Number of buffers on the t_buffers list [j_list_lock] */
404 * Doubly-linked circular list of all buffers reserved but not yet
410 * Doubly-linked circular list of all buffers under writeout during
416 * Doubly-linked circular list of all metadata buffers owned by this
422 * Doubly-linked circular list of all data buffers still to be
428 * Doubly-linked circular list of all forget buffers (superseded
429 * buffers which we can un-checkpoint once this transaction commits)
435 * Doubly-linked circular list of all buffers still to be flushed before
441 * Doubly-linked circular list of all buffers submitted for IO while
447 * Doubly-linked circular list of temporary buffers currently undergoing
453 * Doubly-linked circular list of metadata buffers being shadowed by log
454 * IO. The IO buffers on the iobuf list and the shadow buffers on this
460 * Doubly-linked circular list of control buffers being written to the
477 * Number of buffers reserved for use by all handles in this transaction
553 * @j_max_transaction_buffers: Maximum number of metadata buffers to allow in a
734 * Maximum number of metadata buffers to allow in a single compound
798 /* Filing buffers */
H A Dcirc_buf.h2 * See Documentation/circular-buffers.txt for more information.
H A Djbd2.h82 * The buffer credits field is used to account for journaled buffers
85 * number of outstanding buffers possible at any time. When the
87 * the transaction, so that at all times we know how many buffers the
405 * @h_buffer_credits: Number of remaining buffers we are allowed to dirty.
429 /* Number of remaining buffers we are allowed to dirty: */
472 * new buffers to modify (state not used for now)
477 * The transaction keeps track of all of the buffers modified by a
478 * running transaction, and all of the buffers committed but not yet
534 /* Number of buffers on the t_buffers list [j_list_lock] */
538 * Doubly-linked circular list of all buffers reserved but not yet
544 * Doubly-linked circular list of all metadata buffers owned by this
550 * Doubly-linked circular list of all forget buffers (superseded
551 * buffers which we can un-checkpoint once this transaction commits)
557 * Doubly-linked circular list of all buffers still to be flushed before
563 * Doubly-linked circular list of all buffers submitted for IO while
569 * Doubly-linked circular list of metadata buffers being shadowed by log
570 * IO. The IO buffers on the iobuf list and the shadow buffers on this
613 * Number of buffers reserved for use by all handles in this transaction
721 * @j_reserved_credits: Number of buffers reserved from the running transaction
734 * @j_max_transaction_buffers: Maximum number of metadata buffers to allow in a
877 /* Number of buffers reserved from the running transaction */
924 * Maximum number of metadata buffers to allow in a single compound
1017 /* Filing buffers */
1068 * Fired during journal abort for dirty buffers that will not be
H A Darcdevice.h72 #define D_TIMING 4096 /* show time needed to copy buffers to card */
168 /* byte deposited into first address of buffers on reset */
259 * Buffer management: an ARCnet card has 4 x 512-byte buffers, each of
262 * buffers, and take them as they're needed. This way, we simplify
267 * The rules: only the interrupt handler is allowed to _add_ buffers to
269 * handler and the transmit function will want to _remove_ buffers, so
274 * only four possible buffers, the queue should never be full.
H A Drelay.h36 size_t subbufs_produced; /* count of sub-buffers produced */
37 size_t subbufs_consumed; /* count of sub-buffers consumed */
60 size_t n_subbufs; /* number of sub-buffers per buffer */
66 struct rchan_buf *buf[NR_CPUS]; /* per-cpu channel buffers */
139 * than the default set of per-cpu buffers.
H A Drpmsg.h189 * In case there are no TX buffers available, the function will block until
214 * In case there are no TX buffers available, the function will block until
242 * In case there are no TX buffers available, the function will block until
266 * In case there are no TX buffers available, the function will immediately
291 * In case there are no TX buffers available, the function will immediately
318 * In case there are no TX buffers available, the function will immediately
H A Dmroute.h77 struct sk_buff_head unresolved; /* Unresolved buffers */
/linux-4.1.27/drivers/md/
H A Ddm-bufio.h40 * buffers.
90 * Initiate writing of dirty buffers, without waiting for completion.
95 * Write all dirty buffers. Guarantees that all dirty buffers created prior
119 * Set the minimum number of buffers before cleanup happens.
H A Ddm-bufio.c24 * Limit the number of buffers to DM_BUFIO_MEMORY_PERCENT of main memory
26 * Always allocate at least DM_BUFIO_MIN_BUFFERS buffers.
28 * dirty buffers.
42 * Free buffers when they are older than this (seconds)
72 * Linking of buffers:
73 * All buffers are linked to cache_hash with their hash_list field.
75 * Clean buffers that are not being written (B_WRITING not set)
78 * Dirty and clean buffers that are being written are linked to
82 * context), so some clean-not-writing buffers can be held on
257 * A red/black tree acts as an index for all the buffers.
351 * Small buffers are allocated with kmem_cache, to use space optimally.
353 * For large buffers, we choose between get_free_pages and vmalloc.
362 * buffers and more I/O will be performed. Don't use __get_free_pages if it
655 * Writing dirty buffers
820 * For debugging, if we set the cache size to 1, no new buffers will __alloc_buffer_wait_no_callback()
864 * Free a buffer and wake other threads waiting for free buffers.
909 unsigned long buffers; __get_memory_limit() local
917 buffers = dm_bufio_cache_size_per_client >> __get_memory_limit()
920 if (buffers < c->minimum_buffers) __get_memory_limit()
921 buffers = c->minimum_buffers; __get_memory_limit()
923 *limit_buffers = buffers; __get_memory_limit()
924 *threshold_buffers = buffers * DM_BUFIO_WRITEBACK_PERCENT / 100; __get_memory_limit()
929 * If we are over threshold_buffers, start freeing buffers.
1215 * For performance, it is essential that the buffers are written asynchronously
1273 * dirtying buffers, so we count the number of buffers walked dm_bufio_write_dirty_buffers()
1274 * and if it exceeds the total number of buffers, it means that dm_bufio_write_dirty_buffers()
1463 * An optimization so that the buffers are not written one-by-one. drop_buffers()
1691 * It is required that there are no references on any buffers.
/linux-4.1.27/drivers/media/pci/cx18/
H A Dcx18-queue.c257 /* Reset MDL id's and move all buffers back to the stream's buf_pool */ cx18_unload_queues()
287 * Attach buffers to MDLs, give the MDLs ids, and add MDLs to q_free cx18_load_queues()
289 * Excess buffers are left in buf_pool and/or on an MDL in q_idle cx18_load_queues()
316 * case of a non-integral number of buffers to meet cx18_load_queues()
328 /* Not enough buffers for this MDL; we won't use it */ cx18_load_queues()
352 if (s->buffers == 0) cx18_stream_alloc()
355 CX18_DEBUG_INFO("Allocate %s stream: %d x %d buffers " cx18_stream_alloc()
357 s->name, s->buffers, s->buf_size, cx18_stream_alloc()
358 s->buffers * s->buf_size / 1024, cx18_stream_alloc()
359 (s->buffers * s->buf_size * 100 / 1024) % 100); cx18_stream_alloc()
361 if (((char __iomem *)&cx->scb->cpu_mdl[cx->free_mdl_idx + s->buffers] - cx18_stream_alloc()
366 CX18_ERR("Too many buffers, cannot fit in SCB area\n"); cx18_stream_alloc()
367 CX18_ERR("Max buffers = %zu\n", cx18_stream_alloc()
374 /* allocate stream buffers and MDLs */ cx18_stream_alloc()
375 for (i = 0; i < s->buffers; i++) { cx18_stream_alloc()
409 if (i == s->buffers) { cx18_stream_alloc()
410 cx->free_mdl_idx += s->buffers; cx18_stream_alloc()
413 CX18_ERR("Couldn't allocate buffers for %s stream\n", s->name); cx18_stream_alloc()
424 CX18_DEBUG_INFO("Deallocating buffers for %s stream\n", s->name); cx18_stream_free()
426 /* move all buffers to buf_pool and all MDLs to q_idle */ cx18_stream_free()
H A Dcx18-controls.c52 /* Allocate sliced VBI buffers if needed. */ cx18_s_stream_vbi_fmt()
66 CX18_WARN("Unable to allocate buffers for " cx18_s_stream_vbi_fmt()
/linux-4.1.27/drivers/usb/class/
H A Dcdc-acm.h55 * The only reason to have several buffers is to accommodate assumptions
59 * when processing onlcr, so we only need 2 buffers. These values must be
88 u8 *ctrl_buffer; /* buffers of urbs */
89 dma_addr_t ctrl_dma; /* dma handles of buffers */
100 int write_used; /* number of non-empty write buffers */
/linux-4.1.27/net/sunrpc/xprtrdma/
H A Dverbs.c344 * recv buffers to the pool. NOTE: this affects synchronization of
1082 * Send/recv buffers in req/rep need to be registered rpcrdma_buffer_create()
1324 * Get a set of request/reply buffers.
1330 * removing buffers, and decremented *before* returning them.
1333 rpcrdma_buffer_get(struct rpcrdma_buffer *buffers) rpcrdma_buffer_get() argument
1335 struct rpcrdma_ia *ia = rdmab_to_ia(buffers); rpcrdma_buffer_get()
1340 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_buffer_get()
1341 if (buffers->rb_send_index == buffers->rb_max_requests) { rpcrdma_buffer_get()
1342 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_get()
1343 dprintk("RPC: %s: out of request buffers\n", __func__); rpcrdma_buffer_get()
1347 req = buffers->rb_send_bufs[buffers->rb_send_index]; rpcrdma_buffer_get()
1348 if (buffers->rb_send_index < buffers->rb_recv_index) { rpcrdma_buffer_get()
1351 buffers->rb_recv_index - buffers->rb_send_index); rpcrdma_buffer_get()
1354 req->rl_reply = buffers->rb_recv_bufs[buffers->rb_recv_index]; rpcrdma_buffer_get()
1355 buffers->rb_recv_bufs[buffers->rb_recv_index++] = NULL; rpcrdma_buffer_get()
1357 buffers->rb_send_bufs[buffers->rb_send_index++] = NULL; rpcrdma_buffer_get()
1362 req = rpcrdma_buffer_get_frmrs(req, buffers, &stale); rpcrdma_buffer_get()
1365 req = rpcrdma_buffer_get_fmrs(req, buffers); rpcrdma_buffer_get()
1370 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_get()
1372 rpcrdma_retry_flushed_linv(&stale, buffers); rpcrdma_buffer_get()
1377 * Put request/reply buffers back into pool.
1383 struct rpcrdma_buffer *buffers = req->rl_buffer; rpcrdma_buffer_put() local
1384 struct rpcrdma_ia *ia = rdmab_to_ia(buffers); rpcrdma_buffer_put()
1387 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_buffer_put()
1388 rpcrdma_buffer_put_sendbuf(req, buffers); rpcrdma_buffer_put()
1392 rpcrdma_buffer_put_mrs(req, buffers); rpcrdma_buffer_put()
1397 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_put()
1401 * Recover reply buffers from pool.
1408 struct rpcrdma_buffer *buffers = req->rl_buffer; rpcrdma_recv_buffer_get() local
1411 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_recv_buffer_get()
1412 if (buffers->rb_recv_index < buffers->rb_max_requests) { rpcrdma_recv_buffer_get()
1413 req->rl_reply = buffers->rb_recv_bufs[buffers->rb_recv_index]; rpcrdma_recv_buffer_get()
1414 buffers->rb_recv_bufs[buffers->rb_recv_index++] = NULL; rpcrdma_recv_buffer_get()
1416 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_recv_buffer_get()
1420 * Put reply buffers back into pool when not attached to
1426 struct rpcrdma_buffer *buffers = rep->rr_buffer; rpcrdma_recv_buffer_put() local
1430 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_recv_buffer_put()
1431 buffers->rb_recv_bufs[--buffers->rb_recv_index] = rep; rpcrdma_recv_buffer_put()
1432 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_recv_buffer_put()
1517 * rpcrdma_alloc_regbuf - kmalloc and register memory for SEND/RECV buffers
1652 /* How many chunk list items fit within our inline buffers?
/linux-4.1.27/drivers/infiniband/hw/qib/
H A Dqib_wc_x86_64.c62 * Set the PIO buffers to be WCCOMB, so we get HT bursts to the qib_enable_wc()
66 * the PIO buffers themselves. qib_enable_wc()
67 * For chips with two sets of buffers, the calculations are qib_enable_wc()
70 * The buffers are still packed, so a single range covers both. qib_enable_wc()
142 * or AMD x86_64 processor. AMD x86_64 processors flush WC buffers out in
144 * correct ordering. Intel processors, however, will flush write buffers
H A Dqib_tx.c53 * qib_disarm_piobufs - cancel a range of PIO buffers
56 * @cnt: the number of PIO buffers to cancel
58 * Cancel a range of PIO buffers. Used at user process close,
165 * Disarm a set of send buffers. If the buffer might be actively being
222 * called whenever our local copy indicates we have run out of send buffers
245 * happens when all buffers are in use, so only cpu overhead, not update_send_bufs()
268 * Debugging code and stats updates if no pio buffers available.
301 * Minor optimization. If we had no buffers on last call, qib_getsendbuf_range()
303 * if no buffers were updated, to be paranoid. qib_getsendbuf_range()
339 * buffers available, try an update and then rescan. qib_getsendbuf_range()
379 * qib_chg_pioavailkernel - change which send buffers are available for kernel
382 * @len: the number of send buffers
383 * @avail: true if the buffers are available for kernel use, false otherwise
405 * the user buffers before we hand them back to the qib_chg_pioavailkernel()
452 * all pio buffers, and issuing an abort, which cleans up anything in the
467 * Tell PSM to disarm buffers again before trying to reuse them. qib_cancel_sends()
/linux-4.1.27/drivers/usb/gadget/function/
H A Duvc_queue.c29 * Video buffers queue management.
34 * Video buffers are managed by videobuf2. The driver uses a mutex to protect
36 * spinlock to protect the IRQ queue that holds the buffers to be processed by
141 * Free the video buffers.
149 * Allocate the video buffers.
223 * Cancel the video buffers queue.
225 * Cancelling the queue marks all buffers on the irq queue as erroneous,
259 * Enable or disable the video buffers queue.
262 * disabled after stopping it. This ensures that the video buffers queue
263 * state can be properly initialized before buffers are accessed from the
269 * Disabling the video queue cancels the queue and removes all buffers from
297 * applications will be able to queue buffers for the next uvcg_queue_enable()
H A Duvc_queue.h13 /* Maximum number of video buffers. */
/linux-4.1.27/drivers/scsi/isci/
H A Dunsolicited_frame_control.c69 * The Unsolicited Frame buffers are set at the start of the UF sci_unsolicited_frame_control_construct()
71 * placed after the buffers. sci_unsolicited_frame_control_construct()
106 * Program the actual used UF buffers into the UF address table and sci_unsolicited_frame_control_construct()
110 uf = &uf_control->buffers.array[i]; sci_unsolicited_frame_control_construct()
136 *frame_header = &uf_control->buffers.array[frame_index].header->data; sci_unsolicited_frame_control_get_header()
149 *frame_buffer = uf_control->buffers.array[frame_index].buffer; sci_unsolicited_frame_control_get_buffer()
184 uf_control->buffers.array[frame_index].state = UNSOLICITED_FRAME_RELEASED; sci_unsolicited_frame_control_release_frame()
198 while (uf_control->buffers.array[frame_get].state == UNSOLICITED_FRAME_RELEASED) { sci_unsolicited_frame_control_release_frame()
199 uf_control->buffers.array[frame_get].state = UNSOLICITED_FRAME_EMPTY; sci_unsolicited_frame_control_release_frame()
/linux-4.1.27/drivers/atm/
H A Dzatm.h21 #define LOW_MARK 12 /* start adding new buffers if less than 12 */
22 #define HIGH_MARK 30 /* stop adding buffers after reaching 30 */
49 struct sk_buff_head tx_queue; /* list of buffers in transit */
54 struct sk_buff_head backlog; /* list of buffers waiting for ring */
H A Dnicstar.h52 /* Number of buffers initially allocated */
55 #define NUM_HB 8 /* Pre-allocated huge buffers */
56 #define NUM_IOVB 48 /* Iovec buffers */
58 /* Lower level for count of buffers */
64 /* Upper level for count of buffers */
97 * BUF_XX distinguish the Rx buffers depending on their (small/large) size.
688 volatile int count; /* number of buffers in the queue */
721 skb_pool sbpool; /* Small buffers */
722 skb_pool lbpool; /* Large buffers */
723 skb_pool hbpool; /* Pre-allocated huge buffers */
724 skb_pool iovpool; /* iovector buffers */
H A Dfore200e.h8 #define SMALL_BUFFER_SIZE 384 /* size of small buffers (multiple of 48 (PCA) and 64 (SBA) bytes) */
9 #define LARGE_BUFFER_SIZE 4032 /* size of large buffers (multiple of 48 (PCA) and 64 (SBA) bytes) */
12 #define RBD_BLK_SIZE 32 /* nbr of supplied rx buffers per rbd */
18 #define BUFFER_S1_SIZE SMALL_BUFFER_SIZE /* size of small buffers, scheme 1 */
19 #define BUFFER_L1_SIZE LARGE_BUFFER_SIZE /* size of large buffers, scheme 1 */
21 #define BUFFER_S2_SIZE SMALL_BUFFER_SIZE /* size of small buffers, scheme 2 */
22 #define BUFFER_L2_SIZE LARGE_BUFFER_SIZE /* size of large buffers, scheme 2 */
47 then it uses a number of *large* buffers for the trailing data.
636 struct buffer* buffer; /* array of rx buffers */
637 struct buffer* freebuf; /* list of free rx buffers */
638 volatile int freebuf_count; /* count of free rx buffers */
763 FORE200E_STATE_ALLOC_BUF, /* receive buffers allocated */
803 int buffer_alignment; /* rx buffers DMA alignment requirement */
/linux-4.1.27/include/uapi/linux/hsi/
H A Dcs-protocol.h65 /* maximum number of TX/RX buffers */
69 /* Parameters for setting up the data buffers */
85 __u32 rx_bufs; /* # of RX buffers */
86 __u32 tx_bufs; /* # of TX buffers */
/linux-4.1.27/include/uapi/linux/
H A Dsysinfo.h13 __kernel_ulong_t bufferram; /* Memory used by buffers */
H A Dfanotify.h105 /* Helper functions to deal with fanotify_event_metadata buffers */
H A Dmic_ioctl.h29 * @iov: An array of IOVEC structures containing user space buffers.
H A Dpg.h32 read and write buffers contain a single character "magic" flag.
/linux-4.1.27/drivers/media/platform/s5p-mfc/
H A Ds5p_mfc_common.h501 * @vq_src: vb2 queue for source buffers
502 * @vq_dst: vb2 queue for destination buffers
503 * @src_queue: driver internal queue for source buffers
504 * @dst_queue: driver internal queue for destination buffers
505 * @src_queue_cnt: number of buffers queued on the source internal queue
506 * @dst_queue_cnt: number of buffers queued on the dest internal queue
519 * @dpb_flush_flag: flag used to indicate that a DPB buffers are being
523 * @bank1: handle to memory allocated for temporary buffers from
525 * @bank2: handle to memory allocated for temporary buffers from
527 * @capture_state: state of the capture buffers queue
528 * @output_state: state of the output buffers queue
529 * @src_bufs: information on allocated source buffers
530 * @dst_bufs: information on allocated destination buffers
532 * @dec_dst_flag: flags for buffers queued in the hardware
533 * @dec_src_buf_size: size of the buffer for source buffers in decoding
542 * @dpb_count: count of the DPB buffers required by MFC hw
543 * @total_dpb_count: count of DPB buffers with additional buffers
548 * @mv_count: number of MV buffers allocated for decoding
550 * @enc_dst_buf_size: size of the buffers for encoder output
556 * @ref_queue: list of the reference buffers for encoding
557 * @ref_queue_cnt: number of the buffers in the reference list
/linux-4.1.27/drivers/media/usb/tm6000/
H A Dtm6000-usb-isoc.h34 /* transfer buffers for isoc transfer */
48 /* Stores already requested buffers */
/linux-4.1.27/arch/um/include/asm/
H A Dfixmap.h20 * these 'compile-time allocated' memory buffers are
25 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/arch/microblaze/include/asm/
H A Dfixmap.h41 * these 'compile-time allocated' memory buffers are
46 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/arch/powerpc/include/asm/
H A Dfixmap.h37 * these 'compile-time allocated' memory buffers are
42 * TLB entries of such buffers will not be flushed across
H A Dcache.h54 * prefetched instructions from the ifetch buffers. We also need a sync
/linux-4.1.27/drivers/net/ethernet/amd/
H A Dmvme147.c29 /* We have 32K of RAM for the init block and buffers. This places
30 * an upper limit on the number of buffers we can use. NetBSD uses 8 Rx
31 * buffers and 2 Tx buffers, it takes (8 + 2) * 1544 bytes.
116 printk("%s: No memory for LANCE buffers\n", dev->name); mvme147lance_probe()
H A Dau1000_eth.h43 * Data Buffer Descriptor. Data buffers must be aligned on 32 byte
127 u32 vaddr; /* virtual address of rx/tx buffers */
128 dma_addr_t dma_addr; /* dma address of rx/tx buffers */
H A Dhplance.c30 /* We have 16392 bytes of RAM for the init block and buffers. This places
31 * an upper limit on the number of buffers we can use. NetBSD uses 8 Rx
32 * buffers and 2 Tx buffers, it takes (8 + 2) * 1544 bytes.
H A Dlance.c130 IIIa. Ring buffers
131 The LANCE uses ring buffers of Tx and Rx descriptors. Each entry describes
133 of these buffers is set by LANCE_LOG_{RX,TX}_BUFFERS, which is log_2() of
143 The LANCE has the capability to "chain" both Rx and Tx buffers, but this driver
144 statically allocates full-sized (slightly oversized -- PKT_BUF_SZ) buffers to
146 allocating full-sized buffers "just in case", at the expense of a
150 the buffers are only used when needed as low-memory bounce buffers.
154 the initialization block, Rx and Tx rings, and data buffers, must be
161 As mentioned before, low-memory "bounce-buffers" are used when needed.
183 /* Set the number of Tx and Rx buffers, using Log_2(# buffers).
184 Reasonable default values are 16 Tx buffers, and 16 Rx buffers.
245 unsigned long rx_buffs; /* Address of Rx and Tx buffers. */
296 /* Non-zero if lance_probe1() needs to allocate low-memory bounce buffers.
976 /* The old LANCE chips doesn't automatically pad buffers to min. size. */ lance_start_xmit()
1158 buffers it's possible for a jabber packet to use two lance_rx()
1159 buffers, with only the last correctly noting the error. */ lance_rx()
/linux-4.1.27/drivers/media/platform/exynos4-is/
H A Dfimc-isp.h119 * @pending_buf_q: pending buffers queue head
120 * @active_buf_q: a queue head of buffers scheduled in hardware
122 * @active_buf_count: number of video buffers scheduled in hardware
124 * @reqbufs_count: number of buffers requested with REQBUFS ioctl
139 struct isp_video_buf *buffers[FIMC_ISP_MAX_BUFS]; member in struct:fimc_is_video
H A Dfimc-lite.h98 * @list: list head for the buffers queue
137 * @pending_buf_q: pending buffers queue head
138 * @active_buf_q: the queue head of buffers scheduled in hardware
139 * @vb_queue: vb2 buffers queue
141 * @active_buf_count: number of video buffers scheduled in hardware
143 * @reqbufs_count: the number of buffers requested with REQBUFS ioctl
H A Dfimc-core.h296 * @active_buf_q: the queue head of buffers scheduled in hardware
298 * @active_buf_cnt: number of video buffers scheduled in hardware
299 * @buf_index: index for managing the output DMA buffers
301 * @reqbufs_count: the number of buffers requested in REQBUFS ioctl
386 * @out_buf_count: maximum number of output DMA buffers supported
677 * fimc_active_queue_add - add buffer to the capture active buffers queue
678 * @buf: buffer to add to the active buffers list
688 * fimc_active_queue_pop - pop buffer from the capture active buffers queue
704 * fimc_pending_queue_add - add buffer to the capture pending buffers queue
705 * @buf: buffer to add to the pending buffers list
714 * fimc_pending_queue_pop - pop buffer from the capture pending buffers queue
/linux-4.1.27/fs/ntfs/
H A Daops.c121 /* Async buffers must be locked. */ ntfs_end_buffer_async_read()
129 * If none of the buffers had errors then we can set the page uptodate, ntfs_end_buffer_async_read()
172 * We read each buffer asynchronously and when all buffers are read in, our io
225 * may leave some buffers unmapped which are now allocated. This is ntfs_read_block()
226 * not a problem since these buffers will just get mapped when a write ntfs_read_block()
244 /* Loop through all the buffers in the page. */ ntfs_read_block()
354 /* Lock the buffers. */ ntfs_read_block()
361 /* Finally, start i/o on the buffers. */ ntfs_read_block()
371 /* No i/o was scheduled on any of the buffers. */ ntfs_read_block()
387 * function, ntfs_read_block(), which in turn creates and reads in the buffers
536 * For a page with buffers, map and write the dirty buffers asynchronously
537 * under page writeback. For a page without buffers, create buffers for the
540 * If a page doesn't have buffers the page dirty state is definitive. If a page
541 * does have buffers, the page dirty state is just a hint, and the buffer dirty
542 * state is definitive. (A hint which has rules: dirty buffers against a clean
544 * particular a dirty page containing clean buffers for example.)
585 "buffers. Redirtying page so we try " ntfs_write_block()
589 * its buffers' dirty state as-is. ntfs_write_block()
617 * here, and the (potentially unmapped) buffers may become dirty at ntfs_write_block()
626 * Loop through all the buffers in the page, mapping all the dirty ntfs_write_block()
627 * buffers to disk addresses and handling any aliases from the ntfs_write_block()
637 * Mapped buffers outside i_size will occur, because ntfs_write_block()
639 * truncate in progress. The contents of such buffers ntfs_write_block()
652 /* Clean buffers are not written out, so no need to map them. */ ntfs_write_block()
683 * all non-uptodate buffers and set them uptodate (and ntfs_write_block()
684 * dirty?). Note, there aren't any non-uptodate buffers ntfs_write_block()
686 * FIXME: For an uptodate page, the buffers may need to ntfs_write_block()
692 // Zero any non-uptodate buffers up to i_size. ntfs_write_block()
707 // set_buffer_uptodate() on complete buffers as well as ntfs_write_block()
711 /* No need to map buffers that are already mapped. */ ntfs_write_block()
835 /* Setup all mapped, dirty buffers for async write i/o. */ ntfs_write_block()
875 /* Submit the prepared buffers for i/o. */ ntfs_write_block()
959 /* Make sure we have mapped buffers. */ ntfs_write_mst_block()
1089 * buffers in this record from the list of ntfs_write_mst_block()
1090 * buffers to write and clear their dirty bit ntfs_write_mst_block()
1115 /* If there were no dirty buffers, we are done. */ ntfs_write_mst_block()
1126 /* Skip buffers which are not at the beginning of records. */ ntfs_write_mst_block()
1149 * Remove the buffers in this mft record from ntfs_write_mst_block()
1150 * the list of buffers to write. ntfs_write_mst_block()
1177 * Mark all the buffers in this record clean as we do ntfs_write_mst_block()
1192 /* Lock buffers and start synchronous write i/o on them. */ ntfs_write_mst_block()
1211 /* Wait on i/o completion of buffers. */ ntfs_write_mst_block()
1241 * Skip buffers which are not at the beginning of ntfs_write_mst_block()
1247 /* Skip removed buffers (and hence records). */ ntfs_write_mst_block()
1341 * buffers associated with the page asynchronously.
1371 * The page may have dirty, unmapped buffers. Make them ntfs_writepage()
1520 * buffers' dirty state as-is. ntfs_writepage()
1704 without touching the buffers
1719 * Set the buffers and the page in which the ntfs record is located dirty.
1724 * If the page does not have buffers, we create them and set them uptodate.
1725 * The page may not be locked which is why we need to handle the buffers under
1726 * the mapping->private_lock. Once the buffers are marked dirty we no longer
1727 * need the lock since try_to_free_buffers() does not free dirty buffers.
/linux-4.1.27/drivers/media/v4l2-core/
H A Dvideobuf2-core.c341 * video buffer memory for all buffers/planes on the queue and initializes the
344 * Returns the number of buffers successfully allocated.
361 /* Length stores number of planes for multiplanar buffers */ __vb2_queue_alloc()
403 dprintk(1, "allocated %d buffers, %d plane(s) each\n", __vb2_queue_alloc()
412 static void __vb2_free_mem(struct vb2_queue *q, unsigned int buffers) __vb2_free_mem() argument
417 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_free_mem()
423 /* Free MMAP buffers or release USERPTR buffers */ __vb2_free_mem()
434 * __vb2_queue_free() - free buffers at the end of the queue - video memory and
435 * related information, if no buffers are left return the queue to an
438 static int __vb2_queue_free(struct vb2_queue *q, unsigned int buffers) __vb2_queue_free() argument
446 * buffers from underneath __buf_prepare is obviously a bad idea, so we __vb2_queue_free()
447 * check if any of the buffers is in the state PREPARING, and if so we __vb2_queue_free()
450 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free()
455 dprintk(1, "preparing buffers, cannot free\n"); __vb2_queue_free()
461 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free()
470 __vb2_free_mem(q, buffers); __vb2_queue_free()
534 /* Free videobuf buffers */ __vb2_queue_free()
535 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free()
541 q->num_buffers -= buffers; __vb2_queue_free()
637 * __buffers_in_use() - return true if any buffers on the queue are in use and
674 * single-planar buffers, but userspace does not. __fill_v4l2_buffer()
858 * 3) negotiates number of buffers and planes per buffer with the driver
884 * We already have buffers allocated, so first check if they __reqbufs()
895 * Call queue_cancel to clean up any buffers in the PREPARED or __reqbufs()
896 * QUEUED state which is possible if buffers were prepared or __reqbufs()
923 * Ask the driver how many buffers and planes per buffer it requires. __reqbufs()
931 /* Finally, allocate buffers and video memory */ __reqbufs()
940 * number of buffers needed by this vb2_queue. __reqbufs()
946 * Check if driver can handle the allocated number of buffers. __reqbufs()
958 * Either the driver has accepted a smaller number of buffers, __reqbufs()
978 * Return the number of successfully allocated buffers __reqbufs()
1002 * __create_bufs() - Allocate buffers and any required auxiliary structs
1022 dprintk(1, "maximum number of buffers already allocated\n"); __create_bufs()
1036 * Ask the driver, whether the requested number of buffers, planes per __create_bufs()
1044 /* Finally, allocate buffers and video memory */ __create_bufs()
1053 * Check if driver can handle the so far allocated number of buffers. __create_bufs()
1059 * q->num_buffers contains the total number of buffers, that the __create_bufs()
1069 * Either the driver has accepted a smaller number of buffers, __create_bufs()
1089 * Return the number of successfully allocated buffers __create_bufs()
1158 * If start_streaming fails then it should return buffers with state
1164 * by the means of buf_queue callback. Only buffers previously queued to the
1169 * be started for some reason. In that case the buffers should be returned with
1196 /* sync buffers */ vb2_buffer_done()
1200 /* Add the buffer to the done buffers list */ vb2_buffer_done()
1211 /* Inform any processes that may be waiting for buffers */ vb2_buffer_done()
1217 * vb2_discard_done() - discard all buffers marked as DONE
1221 * discards all 'done' buffers as they would be too old to be requested after
1320 * Single-planar buffers do not use planes array, __fill_vb2_buffer()
1323 * single-planar buffers as well, for simplicity. __fill_vb2_buffer()
1371 * For output buffers mask out the timecode flag: __fill_vb2_buffer()
1627 /* sync buffers */ __enqueue_in_driver()
1771 * at least q->min_buffers_needed buffers queued up (i.e. the minimum
1772 * number of buffers required for the DMA engine to function). If the
1774 * buffers back to vb2 in state QUEUED. Check if that happened and if
1783 * If any buffers were queued before streamon, vb2_start_streaming()
1802 * documentation in videobuf2-core.h for more information how buffers vb2_start_streaming()
1809 * Forcefully reclaim buffers if the driver did not vb2_start_streaming()
1856 * Add to the queued buffers list, a buffer will stay on it until vb2_internal_qbuf()
1865 * For output buffers copy the timestamp if needed, vb2_internal_qbuf()
1888 * start_streaming() since not enough buffers were queued, and vb2_internal_qbuf()
1889 * we now have reached the minimum number of queued buffers, vb2_internal_qbuf()
1941 * spinlock protection. However, buffers may be removed from __vb2_wait_for_done_vb()
1952 dprintk(1, "streaming off, will not wait for buffers\n"); __vb2_wait_for_done_vb()
1957 dprintk(1, "Queue in error state, will not wait for buffers\n"); __vb2_wait_for_done_vb()
1969 dprintk(1, "nonblocking and no buffers to dequeue, " __vb2_wait_for_done_vb()
1984 dprintk(3, "will sleep waiting for buffers\n"); __vb2_wait_for_done_vb()
2039 * vb2_wait_for_all_buffers() - wait until all buffers are given back to vb2
2042 * This function will wait until all buffers that have been given to the driver
2050 dprintk(1, "streaming off, will not wait for buffers\n"); vb2_wait_for_all_buffers()
2131 * buffers ready for dequeuing are present. Normally the driver
2159 * Removes all queued buffers from driver's queue and all buffers queued by
2168 * buffers. __vb2_queue_cancel()
2176 * videobuf2-core.h for more information how buffers should be returned __vb2_queue_cancel()
2193 * Remove all buffers from videobuf's list... __vb2_queue_cancel()
2197 * ...and done list; userspace will not receive any buffers it __vb2_queue_cancel()
2205 * Reinitialize all buffers for next use. __vb2_queue_cancel()
2206 * Make sure to call buf_finish for any queued buffers. Normally __vb2_queue_cancel()
2239 dprintk(1, "no buffers have been allocated\n"); vb2_internal_streamon()
2244 dprintk(1, "need at least %u allocated buffers\n", vb2_internal_streamon()
2250 * Tell driver to start streaming provided sufficient buffers vb2_internal_streamon()
2273 * buffers will return -EIO.
2296 * 2) passes any previously queued buffers to the driver and starts streaming
2319 * Cancel will pause streaming and remove all buffers from the driver vb2_internal_streamoff()
2323 * queue buffers, and then call streamoff without ever having called vb2_internal_streamoff()
2324 * streamon, you would still expect those buffers to be returned to vb2_internal_streamoff()
2342 * 2) stop streaming and dequeues any queued buffers, including those previously
2369 * Go over all buffers and their planes, comparing the given offset __find_plane_by_offset()
2467 * vb2_mmap() - map video buffers into application address space
2472 * This function maps one plane of one of the available video buffers to
2531 * MMAP requires page_aligned buffers. vb2_mmap()
2666 * buffers queued than there are buffers available. vb2_poll()
2761 * tracking context related to the buffers.
2776 * @initial_index: in the read() case all buffers are queued up immediately
2778 * buffers. However, in the write() case no buffers are initially
2780 * __vb2_perform_fileio(). Only once all available buffers have
2782 * buffers. This means that initially __vb2_perform_fileio()
2784 * the buffers for the first time. That initial index is stored
2786 * available buffers have been queued and __vb2_perform_fileio()
2826 * Check if device supports mapping buffers to kernel virtual space. __vb2_init_fileio()
2854 * Request buffers and use MMAP type to force driver __vb2_init_fileio()
2855 * to allocate buffers by itself. __vb2_init_fileio()
2867 * (multiplane buffers are not supported). __vb2_init_fileio()
2887 * Read mode requires pre queuing of all buffers. __vb2_init_fileio()
2893 * Queue all buffers. __vb2_init_fileio()
2913 * All buffers have been queued, so mark that by setting __vb2_init_fileio()
3112 * If we are queuing up buffers for the first time, then __vb2_perform_fileio()
3122 * all the 'first time' buffers. __vb2_perform_fileio()
3312 /* If count == 0, then the owner has released all buffers and he vb2_ioctl_reqbufs()
3357 /* No need to call vb2_queue_is_busy(), anyone can query buffers. */ vb2_ioctl_querybuf()
H A Dv4l2-mem2mem.c4 * Helper functions for devices that use videobuf buffers for both their
48 /* Offset base for buffers on the destination queue - used to distinguish
49 * between source and destination buffers when mmapping - they receive the same
96 * v4l2_m2m_next_buf() - return next buffer from the list of ready buffers
117 * v4l2_m2m_buf_remove() - take off a buffer from the list of ready buffers and
202 * to be drained before doing streamoff), allow scheduling without v4l2 buffers
246 dprintk("No input buffers available\n"); v4l2_m2m_try_schedule()
257 dprintk("No output buffers available\n"); v4l2_m2m_try_schedule()
345 /* This instance might have more buffers ready, but since we do not v4l2_m2m_job_finish()
518 * v4l2_m2m_poll() - poll replacement, for destination buffers only
608 * by adding a predefined offset to buffers from one of the queues and
738 * v4l2_m2m_buf_queue() - add a buffer to the proper ready buffers list.
H A Dvideobuf-core.c2 * generic helper functions for handling video4linux capture buffers
40 MODULE_DESCRIPTION("helper module to manage video4linux buffers");
224 * __videobuf_free() - free all the buffers and their control structures
226 * This function can only be called if streaming/reading is off, i.e. no buffers
239 dprintk(1, "Cannot free buffers when streaming or reading\n"); __videobuf_free()
247 dprintk(1, "Cannot free mmapped buffers\n"); __videobuf_free()
272 /* remove queued buffers from list */ videobuf_queue_cancel()
285 /* free all buffers + clear queue */ videobuf_queue_cancel()
397 /* Allocate and initialize buffers */ __videobuf_mmap_setup()
422 dprintk(1, "mmap setup: %d buffers, %d bytes each\n", i, bsize); __videobuf_mmap_setup()
646 dprintk(2, "next_buffer: no buffers to dequeue\n"); stream_next_buffer_check_queue()
/linux-4.1.27/arch/mips/include/asm/octeon/
H A Dcvmx-helper-util.h61 * this many packet buffers free in FPA 0.
64 * than this many free packet buffers in FPA 0.
75 * this many packet buffers free in FPA 0.
78 * than this many free packet buffers in FPA 0.
144 * Free the packet buffers contained in a work queue entry.
162 * Since the number of buffers is not zero, we know this is cvmx_helper_free_packet_data()
165 * true, we need to free all buffers except for the first cvmx_helper_free_packet_data()
H A Dcvmx-config.h33 /* Packet buffers */
39 /* PKO queue command buffers */
/linux-4.1.27/drivers/media/platform/xilinx/
H A Dxilinx-dma.h67 * @queue: vb2 buffers queue
69 * @sequence: V4L2 buffers sequence number
70 * @queued_bufs: list of queued buffers
/linux-4.1.27/drivers/media/usb/pvrusb2/
H A Dpvrusb2-io.h70 /* Find out how many buffers are idle or ready */
74 /* Kill all pending buffers and throw away any ready buffers as well */
H A Dpvrusb2-io.c62 /* Pointers to all buffers */
63 struct pvr2_buffer **buffers; member in struct:pvr2_stream
64 /* Array size of buffers */
66 /* Total buffers actually in circulation */
68 /* Designed number of buffers to be in circulation */
305 /* Allocate buffers pointer array in multiples of 32 entries */ pvr2_stream_buffer_count()
324 memcpy(nb,sp->buffers, pvr2_stream_buffer_count()
326 kfree(sp->buffers); pvr2_stream_buffer_count()
328 sp->buffers = nb; pvr2_stream_buffer_count()
340 sp->buffers[sp->buffer_total_count] = bp; pvr2_stream_buffer_count()
347 bp = sp->buffers[sp->buffer_total_count - 1]; pvr2_stream_buffer_count()
349 sp->buffers[sp->buffer_total_count - 1] = NULL; pvr2_stream_buffer_count()
357 nb = kmemdup(sp->buffers, scnt * sizeof(*nb), pvr2_stream_buffer_count()
361 kfree(sp->buffers); pvr2_stream_buffer_count()
362 sp->buffers = nb; pvr2_stream_buffer_count()
387 bp = sp->buffers[sp->buffer_total_count - (cnt + 1)]; pvr2_stream_achieve_buffer_count()
582 return sp->buffers[id]; pvr2_stream_get_buffer()
/linux-4.1.27/drivers/net/fddi/skfp/h/
H A Dosdef1st.h67 // default number of receive buffers.
70 // default number of transmit buffers.
73 // Number of SMT buffers (Mbufs).
/linux-4.1.27/fs/
H A Dbuffer.c79 * Returns if the page has dirty or writeback buffers. If all the buffers
188 * But it's the page lock which protects the buffers. To get around this,
230 /* we might be here because some of the buffers on this page are __find_get_block_slow()
233 * elsewhere, don't buffer_error if we had some unmapped buffers __find_get_block_slow()
322 * If none of the buffers had errors and they are all end_buffer_async_read()
387 * If a page's buffers are under async readin (end_buffer_async_read
389 * control could lock one of the buffers after it has completed
390 * but while some of the other buffers have not completed. This
395 * The page comes unlocked when it has no locked buffer_async buffers
399 * the buffers.
436 * management of a list of dependent buffers at ->i_mapping->private_list.
438 * Locking is a little subtle: try_to_free_buffers() will remove buffers
441 * at the time, not against the S_ISREG file which depends on those buffers.
443 * which backs the buffers. Which is different from the address_space
444 * against which the buffers are listed. So for a particular address_space,
449 * Which introduces a requirement: all buffers on an address_space's
452 * address_spaces which do not place buffers at ->private_list via these
463 * mark_buffer_dirty_fsync() to clearly define why those buffers are being
470 * that buffers are taken *off* the old inode's list when they are freed
499 * you dirty the buffers, and then use osync_inode_buffers to wait for
500 * completion. Any other dirty buffers which are not yet queued for
560 * sync_mapping_buffers - write out & wait upon a mapping's "associated" buffers
561 * @mapping: the mapping which wants those buffers written
563 * Starts I/O against the buffers at mapping->private_list, and waits upon
567 * @mapping is a file or directory which needs those buffers to be written for
649 * If the page has buffers, the uptodate buffers are set dirty, to preserve
650 * dirty-state coherency between the page and the buffers. It the page does
651 * not have buffers then when they are later attached they will all be set
654 * The buffers are dirtied before the page is dirtied. There's a small race
657 * before the buffers, a concurrent writepage caller could clear the page dirty
658 * bit, see a bunch of clean buffers and we'd end up with dirty buffers/clean
662 * page's buffer list. Also use this to protect against clean buffers being
696 * Write out and wait upon a list of buffers.
699 * initially dirty buffers get waited on, but that any subsequently
700 * dirtied buffers don't. After all, we don't want fsync to last
703 * Do this in two main stages: first we copy dirty buffers to a
711 * the osync code to catch these locked, dirty buffers without requeuing
712 * any newly dirty buffers for write.
794 * Invalidate any and all dirty buffers on a given inode. We are
796 * done a sync(). Just drop the buffers from the inode list.
799 * assumes that all the buffers are against the blockdev. Not true
818 * Remove any clean buffers from the inode's buffer list. This is called
819 * when we're trying to free the inode itself. Those buffers can pin it.
821 * Returns true if all buffers were removed.
847 * Create the appropriate buffers when given a page for data area and
849 * follow the buffers created. Return NULL if unable to create more
850 * buffers.
895 * partially complete buffers, so all were released above. alloc_page_buffers()
938 * Initialise the state of a blockdev page's buffers.
1014 * Allocate some buffers for this page grow_dev_page()
1021 * Link the page to the buffers and initialise them. Take the grow_dev_page()
1039 * Create buffers for the specified block device block's page. If
1040 * that page was dirty, the buffers are set dirty also.
1069 /* Create a page with the proper size buffers.. */ grow_buffers()
1107 * The relationship between dirty buffers and dirty pages:
1109 * Whenever a page has any dirty buffers, the page's dirty bit is set, and
1112 * At all times, the dirtiness of the buffers represents the dirtiness of
1113 * subsections of the page. If the page has buffers, the page dirty bit is
1116 * When a page is set dirty in its entirety, all its buffers are marked dirty
1117 * (if the page has buffers).
1120 * buffers are not.
1122 * Also. When blockdev buffers are explicitly read with bread(), they
1124 * uptodate - even if all of its buffers are uptodate. A subsequent
1126 * buffers, will set the page uptodate and will perform no I/O.
1171 * Decrement a buffer_head's reference count. If all buffers against a page
1173 * and unlocked then try_to_free_buffers() may strip the buffers from the page
1174 * in preparation for freeing it (sometimes, rarely, buffers are removed from
1175 * a page but it ends up not being freed, and buffers may later be reattached).
1503 * block_invalidatepage() does not have to release all buffers, but it must
1547 * We release buffers only if the entire page is being invalidated. block_invalidatepage()
1560 * We attach and possibly dirty the buffers atomically wrt
1606 * be writeout I/O going on against recently-freed buffers. We don't
1662 * While block_write_full_page is writing back the dirty buffers under
1663 * the page lock, whoever dirtied the buffers may decide to clean them
1695 * here, and the (potentially unmapped) buffers may become dirty at __block_write_full_page()
1711 * Get all the dirty buffers mapped to disk addresses and __block_write_full_page()
1717 * mapped buffers outside i_size will occur, because __block_write_full_page()
1768 * The page and its buffers are protected by PageWriteback(), so we can __block_write_full_page()
1788 * The page was marked dirty, but the buffers were __block_write_full_page()
1809 /* Recovery: lock and submit the mapped buffers */ __block_write_full_page()
1841 * If a page has any new buffers, zero them out here, and mark them uptodate
1992 * If this is a partial write which happened to make all buffers __block_commit_write()
2042 * The buffers that were written will now be uptodate, so we block_write_end()
2108 * block_is_partially_uptodate checks whether buffers within a page are
2111 * Returns true if all buffers which correspond to a file portion
2213 * All buffers are uptodate - we can set the page uptodate block_read_full_page()
2222 /* Stage two: lock the buffers */ block_read_full_page()
2472 * Attach the singly-linked list of buffers created by nobh_write_begin, to
2540 * Allocate buffers so that we can keep track of state, and potentially nobh_write_begin()
2600 * The page is locked, so these buffers are protected from nobh_write_begin()
2698 * The page may have dirty, unmapped buffers. For example, nobh_writepage()
2902 * The page may have dirty, unmapped buffers. For example, block_write_full_page()
3075 * request. Further it marks as clean buffers that are processed for
3083 * All of the buffers must be for the same device, and must also be a
3160 * try_to_free_buffers() checks if all the buffers on this particular page
3166 * If the page is dirty but all the buffers are clean then we need to
3168 * may be against a block device, and a later reattachment of buffers
3169 * to a dirty page will set *all* buffers dirty. Which would corrupt
3172 * The same applies to regular filesystem pages: if all the buffers are
3233 * If the filesystem writes its buffers by hand (eg ext3) try_to_free_buffers()
3234 * then we can have clean buffers against a dirty page. We try_to_free_buffers()
3239 * the page's buffers clean. We discover that here and clean try_to_free_buffers()
H A Dsplice.c9 * that transfers data buffers to or from a pipe buffer.
205 if (pipe->nrbufs < pipe->buffers) { splice_to_pipe()
206 int newbuf = (pipe->curbuf + pipe->nrbufs) & (pipe->buffers - 1); splice_to_pipe()
226 if (pipe->nrbufs < pipe->buffers) splice_to_pipe()
279 unsigned int buffers = ACCESS_ONCE(pipe->buffers); splice_grow_spd() local
281 spd->nr_pages_max = buffers; splice_grow_spd()
282 if (buffers <= PIPE_DEF_BUFFERS) splice_grow_spd()
285 spd->pages = kmalloc(buffers * sizeof(struct page *), GFP_KERNEL); splice_grow_spd()
286 spd->partial = kmalloc(buffers * sizeof(struct partial_page), GFP_KERNEL); splice_grow_spd()
743 * destination. It returns when there's no more buffers left in
751 * locking is required around copying the pipe buffers to the
789 pipe->curbuf = (pipe->curbuf + 1) & (pipe->buffers - 1); splice_from_pipe_feed()
809 * value (one) if pipe buffers are available. It will return zero
956 int nbufs = pipe->buffers; iter_file_splice_write()
976 if (unlikely(nbufs < pipe->buffers)) { iter_file_splice_write()
978 nbufs = pipe->buffers; iter_file_splice_write()
996 if (idx == pipe->buffers - 1) iter_file_splice_write()
1022 /* dismiss the fully eaten buffers, adjust the partial one */ iter_file_splice_write()
1031 pipe->curbuf = (pipe->curbuf + 1) & (pipe->buffers - 1); iter_file_splice_write()
1261 * the pipe buffers in question: splice_direct_to_actor()
1263 for (i = 0; i < pipe->buffers; i++) { splice_direct_to_actor()
1442 int buffers = 0, error = 0; get_iovec_page_array() local
1483 if (npages > pipe_buffers - buffers) get_iovec_page_array()
1484 npages = pipe_buffers - buffers; get_iovec_page_array()
1487 0, &pages[buffers]); get_iovec_page_array()
1498 partial[buffers].offset = off; get_iovec_page_array()
1499 partial[buffers].len = plen; get_iovec_page_array()
1503 buffers++; get_iovec_page_array()
1519 if (error < npages || buffers == pipe_buffers) get_iovec_page_array()
1526 if (buffers) get_iovec_page_array()
1527 return buffers; get_iovec_page_array()
1758 if (pipe->nrbufs < pipe->buffers) opipe_prep()
1764 while (pipe->nrbufs >= pipe->buffers) { opipe_prep()
1830 if (!ipipe->nrbufs || opipe->nrbufs >= opipe->buffers) { splice_pipe_to_pipe()
1831 /* Already processed some buffers, break */ splice_pipe_to_pipe()
1842 * managed to process any buffers. A zero return splice_pipe_to_pipe()
1851 nbuf = (opipe->curbuf + opipe->nrbufs) & (opipe->buffers - 1); splice_pipe_to_pipe()
1861 ipipe->curbuf = (ipipe->curbuf + 1) & (ipipe->buffers - 1); splice_pipe_to_pipe()
1928 * If we have iterated all input buffers or ran out of link_pipe()
1931 if (i >= ipipe->nrbufs || opipe->nrbufs >= opipe->buffers) link_pipe()
1934 ibuf = ipipe->bufs + ((ipipe->curbuf + i) & (ipipe->buffers-1)); link_pipe()
1935 nbuf = (opipe->curbuf + opipe->nrbufs) & (opipe->buffers - 1); link_pipe()
H A Dmpage.c93 * The idea is to avoid adding buffers to pages that don't already have
95 * this marks the page up to date instead of adding new buffers.
106 * don't make any buffers if there is only one buffer on map_buffer_to_page()
221 * we just collected from get_block into the page's buffers do_mpage_readpage()
327 * - encountering a page which has buffers
412 * If the page has buffers then they will be used for obtaining the disk
416 * If the page has no buffers (preferred) then the page is mapped here.
434 * We have our BIO, so we can now mark the buffers clean. Make
435 * sure to only clean buffers which we know we'll be writing.
496 * unmapped dirty buffers are created by __mpage_writepage()
528 * Page has buffers, but they are all unmapped. The page was __mpage_writepage()
537 * The page has no buffers: map it to disk __mpage_writepage()
H A Dpipe.c172 * in the tee() system call, when we duplicate the buffers in one
283 curbuf = (curbuf + 1) & (pipe->buffers - 1); pipe_read()
363 (pipe->buffers - 1); pipe_write()
396 if (bufs < pipe->buffers) { pipe_write()
397 int newbuf = (pipe->curbuf + bufs) & (pipe->buffers-1); pipe_write()
440 if (bufs < pipe->buffers) pipe_write()
489 buf = (buf+1) & (pipe->buffers - 1); pipe_ioctl()
519 mask |= (nrbufs < pipe->buffers) ? POLLOUT | POLLWRNORM : 0; pipe_poll()
597 pipe->buffers = PIPE_DEF_BUFFERS; alloc_pipe_info()
611 for (i = 0; i < pipe->buffers; i++) { free_pipe_info()
958 * Allocate a new array of pipe buffers and copy the info over. Returns the
969 * contains more buffers than arg, then return busy. pipe_set_size()
987 if (tail < pipe->buffers) pipe_set_size()
990 tail &= (pipe->buffers - 1); pipe_set_size()
1002 pipe->buffers = nr_pages; pipe_set_size()
1075 ret = pipe->buffers * PAGE_SIZE; pipe_fcntl()
/linux-4.1.27/drivers/rpmsg/
H A Dvirtio_rpmsg_bus.c42 * @rbufs: kernel address of rx buffers
43 * @sbufs: kernel address of tx buffers
44 * @num_bufs: total number of buffers for rx and tx
46 * @bufs_dma: dma base addr of the buffers
52 * @sendq: wait queue of sending contexts waiting for a tx buffers
91 * We're allocating buffers of 512 bytes each for communications. The
92 * number of buffers will be computed from the number of buffers supported
93 * by the vring, upto a maximum of 512 buffers (256 in each direction).
98 * This will utilize a maximum total space of 256KB for the buffers.
100 * We might also want to add support for user-provided buffers in time.
582 * (half of our buffers are used for sending messages) get_a_tx_buf()
616 /* are we the first sleeping context waiting for tx buffers ? */ rpmsg_upref_sleepers()
643 /* are we the last sleeping context waiting for tx buffers ? */ rpmsg_downref_sleepers()
658 * @wait: indicates whether caller should block in case no TX buffers available
666 * The message is sent using one of the TX buffers that are available for
676 * Otherwise, if @wait is false, and there are no TX buffers available,
701 * We currently use fixed-sized buffers, and therefore the payload rpmsg_send_offchannel_raw()
705 * user-provided buffers (and then we can also support zero-copy rpmsg_send_offchannel_raw()
795 * We currently use fixed-sized buffers, so trivially sanitize rpmsg_recv_single()
979 /* we need less buffers if vrings are small */ rpmsg_probe()
987 /* allocate coherent memory for the buffers */ rpmsg_probe()
996 dev_dbg(&vdev->dev, "buffers: va %p, dma 0x%llx\n", bufs_va, rpmsg_probe()
999 /* half of the buffers is dedicated for RX */ rpmsg_probe()
1005 /* set up the receive buffers */ rpmsg_probe()
/linux-4.1.27/drivers/gpu/drm/
H A Ddrm_dma.c75 * Free all pages associated with DMA buffers, the buffers and pages lists, and
91 /* Clear dma buffers */ drm_legacy_dma_takedown()
140 * Reclaim the buffers.
/linux-4.1.27/arch/metag/include/uapi/asm/
H A Dsigcontext.h18 * buffers have been saved and will be replayed on sigreturn.
/linux-4.1.27/drivers/tty/
H A Dtty_buffer.c25 * Byte threshold to limit memory consumption for flip buffers.
108 * tty_buffer_free_all - free buffers used by a tty
111 * Remove all the buffers pending on a tty whether queued with data
143 * We round our buffers off in 256 character chunks to get better
202 * tty_buffer_flush - flush full tty buffers
206 * flush all the buffers containing receive data. If ld != NULL,
263 /* This is the slow path - looking for new buffers to use */ __tty_buffer_request_room()
317 several buffers. If this is the case we must loop */ tty_insert_flip_string_fixed_flag()
352 several buffers. If this is the case we must loop */ tty_insert_flip_string_flags()
362 * Takes any pending buffers and transfers their ownership to the
494 * Queue a push of the terminal flip buffers to the line discipline.
/linux-4.1.27/drivers/isdn/hardware/eicon/
H A Dpr_pc.h35 byte RcOutput; /* number of Rc buffers received */
36 byte IndOutput; /* number of Ind buffers received */
/linux-4.1.27/drivers/iio/
H A Dindustrialio-triggered-buffer.c31 * buffers will be used.
102 MODULE_DESCRIPTION("IIO helper functions for setting up triggered buffers");
/linux-4.1.27/drivers/media/platform/omap/
H A Domap_vout_vrfb.c29 * Function for allocating video buffers
75 * Free VRFB buffers
182 * Allocate the buffers for the VRFB space. Data is copied from V4L2
183 * buffers to the VRFB buffers using the DMA engine.
197 /* Allocate the VRFB buffers only if the buffers are not omap_vout_vrfb_buffer_setup()
281 /* Store buffers physical address into an array. Addresses omap_vout_prepare_vrfb()
/linux-4.1.27/drivers/scsi/mpt3sas/
H A Dmpt3sas_trigger_diag.c155 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_master()
162 /* check to see if trace buffers are currently released */ mpt3sas_trigger_master()
228 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_event()
235 /* check to see if trace buffers are currently released */ mpt3sas_trigger_event()
308 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_scsi()
315 /* check to see if trace buffers are currently released */ mpt3sas_trigger_scsi()
382 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_mpi()
389 /* check to see if trace buffers are currently released */ mpt3sas_trigger_mpi()
/linux-4.1.27/drivers/net/wireless/ti/wlcore/
H A Devent.h31 * The event mechanism is based on a pair of event buffers (buffers A and
/linux-4.1.27/arch/tile/include/asm/
H A Dfixmap.h38 * these 'compile-time allocated' memory buffers are
43 * TLB entries of such buffers will not be flushed across
H A Dkgdb.h39 * buffers at least NUMREGBYTES*2 are needed for register packets,
/linux-4.1.27/arch/mips/include/asm/
H A Dfixmap.h33 * these 'compile-time allocated' memory buffers are
38 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/arch/score/include/asm/
H A Dfixmap.h28 * these 'compile-time allocated' memory buffers are
33 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/drivers/isdn/icn/
H A Dicn.h88 #define SHM_CBUF_OFFSET (0x200) /* Offset to comm-buffers in shm */
89 #define SHM_DBUF_OFFSET (0x2000) /* Offset to data-buffers in shm */
92 * Layout of card's data buffers
152 int rcvidx[ICN_BCH]; /* Index for above buffers */
175 unsigned long memaddr; /* Address of memory mapped buffers */
176 icn_shmem __iomem *shmem; /* Pointer to memory-mapped-buffers */
/linux-4.1.27/drivers/media/platform/marvell-ccic/
H A Dmcam-core.c37 * we must have physically contiguous buffers to bring frames into.
38 * These parameters control how many buffers we use, whether we
43 * The controller can cycle through three buffers. We could use
51 "Non-zero value causes DMA buffers to be allocated when the "
54 "successfully getting those buffers. This parameter is "
60 "The number of DMA buffers to allocate. Can be either two "
66 "The size of the allocated DMA buffers. If actual operating "
67 "parameters require larger buffers, an attempt to reallocate "
348 * Allocate in-kernel DMA buffers for vmalloc mode.
380 cam_err(cam, "Insufficient DMA buffers, cannot operate\n"); mcam_alloc_dma_bufs()
385 cam_warn(cam, "Will limp along with only 2 buffers\n"); mcam_alloc_dma_bufs()
405 * Set up DMA buffers when operating in vmalloc mode
410 * Store the first two Y buffers (we aren't supporting mcam_ctlr_dma_vmalloc()
446 if (list_empty(&cam->buffers)) { mcam_frame_tasklet()
452 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, mcam_frame_tasklet()
469 * Make sure our allocated buffers are up to the task.
538 * If there are no available buffers, go into single mode mcam_set_contig_buffer()
540 if (list_empty(&cam->buffers)) { mcam_set_contig_buffer()
548 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, mcam_set_contig_buffer()
627 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, queue); mcam_sg_next_buffer()
652 if (list_empty(&cam->buffers)) { mcam_ctlr_dma_sg()
665 * a descriptor chain on the fly, since the controller buffers it
689 if (!list_empty(&cam->buffers)) { mcam_dma_sg_done()
1024 * Configuration. If we still don't have DMA buffers, mcam_read_setup()
1096 start = (cam->state == S_BUFWAIT) && !list_empty(&cam->buffers); mcam_vb_buf_queue()
1097 list_add(&mvb->queue, &cam->buffers); mcam_vb_buf_queue()
1114 INIT_LIST_HEAD(&cam->buffers); mcam_vb_start_streaming()
1119 * Videobuf2 sneakily hoards all the buffers and won't mcam_vb_start_streaming()
1123 * give us buffers soon. mcam_vb_start_streaming()
1125 if (cam->buffer_mode != B_vmalloc && list_empty(&cam->buffers)) { mcam_vb_start_streaming()
1146 /* They never gave us buffers */ mcam_vb_stop_streaming()
1160 * VB2 reclaims the buffers, so we need to forget mcam_vb_stop_streaming()
1164 INIT_LIST_HEAD(&cam->buffers); mcam_vb_stop_streaming()
1249 INIT_LIST_HEAD(&cam->buffers); mcam_setup_vb2()
1448 * Also can't if there are streaming buffers in place. mcam_vidioc_s_fmt_vid_cap()
1470 * Make sure we have appropriate DMA buffers. mcam_vidioc_s_fmt_vid_cap()
1539 * the level which controls the number of read buffers.
1927 INIT_LIST_HEAD(&cam->buffers); mccic_register()
1963 * If so requested, try to get our DMA buffers now. mccic_register()
1967 cam_warn(cam, "Unable to alloc DMA buffers at load" mccic_register()
2043 list_add(&cam->vb_bufs[0]->queue, &cam->buffers); mccic_resume()
H A Dmcam-core.h42 S_BUFWAIT /* streaming requested but no buffers yet */
161 struct list_head buffers; /* Available frames */ member in struct:mcam_camera
166 /* DMA buffers - vmalloc mode */
176 /* DMA buffers - DMA modes */
360 #define C1_TWOBUFS 0x08000000 /* Use only two DMA buffers */
/linux-4.1.27/arch/tile/include/hv/
H A Ddrv_xgbe_impl.h124 /** A queue of buffers.
126 * This structure partially defines a queue of empty buffers which have been
144 * new buffers are outstanding; when the requested write offset is one greater
148 * buffers at decreasing addresses. This makes the check for "is it time to
150 * buffers, and means that the value which defines the queue size,
171 * This queue has 10 slots, and thus can hold 9 buffers (_last_buffer = 9).
173 * valid, unread buffers in slots 2, 1, 0, 9, 8, and 7. The requested write
174 * pointer is at 4; thus, requests have been made to the IPP for buffers which
200 /** Offset to queues of empty buffers, one per size. */
232 /** The next empty buffers to be read, one index per size. */
/linux-4.1.27/drivers/s390/cio/
H A Dqdio_main.c107 /* not all buffers processed */ qdio_check_ccq()
118 * @state: state of the extracted buffers
120 * @count: count of buffers to examine
121 * @auto_ack: automatically acknowledge buffers
124 * Stops processing if a state is different from the last buffers state.
154 * extracted buffers before trying again. qdio_do_eqbs()
172 * @state: new state of the buffers
174 * @count: how many buffers to change
176 * Returns the number of successfully changed buffers.
215 /* returns number of examined buffers and their common state in *state */ get_buf_states()
248 /* wrap-around safe setting of slsb states, returns number of changed buffers */ set_buf_states()
497 /* need to change ALL buffers to get more interrupts */ inbound_primed()
509 * Don't check 128 buffers, as otherwise qdio_inbound_q_moved get_inbound_buffer_frontier()
740 * Don't check 128 buffers, as otherwise qdio_inbound_q_moved get_outbound_buffer_frontier()
789 /* all buffers processed? */ qdio_outbound_q_done()
1489 * handle_inbound - reset processed input buffers
1490 * @q: queue containing the buffers
1493 * @count: how many buffers are emptied
1540 * handle_outbound - process filled outbound buffers
1541 * @q: queue containing the buffers
1544 * @count: how many buffers are filled
1578 /* try to fast requeue buffers */ handle_outbound()
1597 * do_QDIO - process input or output buffers
1602 * @count: how many buffers to process
1634 * qdio_start_irq - process input buffers
1676 * qdio_get_next_buffers - process input buffers
1680 * @error: buffers are in error state
1684 * = 0 - no new buffers found
1685 * > 0 - number of processed buffers
H A Dqdio_setup.c38 * qdio_free_buffers() - free qdio buffers
39 * @buf: array of pointers to qdio buffers
40 * @count: number of qdio buffers to free
52 * qdio_alloc_buffers() - allocate qdio buffers
53 * @buf: array of pointers to qdio buffers
54 * @count: number of qdio buffers to allocate
75 * qdio_reset_buffers() - reset qdio buffers
76 * @buf: array of pointers to qdio buffers
77 * @count: number of qdio buffers that will be zeroed
/linux-4.1.27/drivers/media/usb/uvc/
H A Duvc_queue.c28 * Video buffers queue management.
33 * Video buffers are managed by videobuf2. The driver uses a mutex to protect
35 * spinlock to protect the IRQ queue that holds the buffers to be processed by
46 * Return all queued buffers to videobuf2 in the requested state.
337 * Check if buffers have been allocated.
351 * Cancel the video buffers queue.
353 * Cancelling the queue marks all buffers on the irq queue as erroneous,
/linux-4.1.27/arch/mips/cavium-octeon/executive/
H A Dcvmx-helper-util.c176 * this many packet buffers free in FPA 0.
179 * than this many free packet buffers in FPA 0.
187 /* Set RED to begin dropping packets when there are pass_thresh buffers cvmx_helper_setup_red_queue()
189 buffers */ cvmx_helper_setup_red_queue()
211 * this many packet buffers free in FPA 0.
214 * than this many free packet buffers in FPA 0.
226 /* Disable backpressure based on queued buffers. It needs SW support */ cvmx_helper_setup_red()
/linux-4.1.27/drivers/media/usb/cpia2/
H A Dcpia2_v4l.c56 MODULE_PARM_DESC(num_buffers, "Number of frame buffers (1-"
190 struct framebuf *frame = &cam->buffers[frame_nr]; sync()
442 if (cam->buffers[frame].status == FRAME_READING) cpia2_s_fmt_vid_cap()
446 cam->buffers[frame].status = FRAME_EMPTY; cpia2_s_fmt_vid_cap()
783 * NOTE: The user's request is ignored. For now the buffers are fixed.
818 buf->m.offset = cam->buffers[buf->index].data - cam->frame_buffer; cpia2_querybuf()
830 switch (cam->buffers[buf->index].status) { cpia2_querybuf()
838 buf->bytesused = cam->buffers[buf->index].length; cpia2_querybuf()
839 buf->timestamp = cam->buffers[buf->index].timestamp; cpia2_querybuf()
840 buf->sequence = cam->buffers[buf->index].seq; cpia2_querybuf()
871 if(cam->buffers[buf->index].status == FRAME_READY) cpia2_qbuf()
872 cam->buffers[buf->index].status = FRAME_EMPTY; cpia2_qbuf()
890 if(cam->buffers[i].status == FRAME_READY) { find_earliest_filled_buffer()
896 tv1 = &cam->buffers[i].timestamp; find_earliest_filled_buffer()
897 tv2 = &cam->buffers[found].timestamp; find_earliest_filled_buffer()
947 buf->bytesused = cam->buffers[buf->index].length; cpia2_dqbuf()
951 buf->timestamp = cam->buffers[buf->index].timestamp; cpia2_dqbuf()
952 buf->sequence = cam->buffers[buf->index].seq; cpia2_dqbuf()
953 buf->m.offset = cam->buffers[buf->index].data - cam->frame_buffer; cpia2_dqbuf()
960 cam->buffers[buf->index].status, buf->sequence, buf->bytesused); cpia2_dqbuf()
1234 DBG("Using %d buffers, each %d bytes, alternate=%d\n", check_parameters()
H A Dcpia2_usb.c124 /* No junk in the buffers */ process_frame()
137 * mmapped, or we have few buffers, we need to process_frame()
141 * buffers, but it would take an EXTREMELY process_frame()
153 DBG("Changed buffers, work:%d, current:%d\n", process_frame()
257 break; /* No READING or EMPTY buffers left */ cpia2_usb_complete()
718 DBG("Flushing buffers\n"); cpia2_usb_stream_start()
720 cam->buffers[i].status = FRAME_EMPTY; cpia2_usb_stream_start()
721 cam->buffers[i].length = 0; cpia2_usb_stream_start()
723 cam->curbuff = &cam->buffers[0]; cpia2_usb_stream_start()
889 if(cam->buffers) { cpia2_usb_disconnect()
H A Dcpia2_core.c138 * the camera, copy the buffers into the proper param structures.
2237 if(!cam->buffers) { cpia2_allocate_buffers()
2239 cam->buffers = kmalloc(size, GFP_KERNEL); cpia2_allocate_buffers()
2240 if(!cam->buffers) { cpia2_allocate_buffers()
2250 kfree(cam->buffers); cpia2_allocate_buffers()
2251 cam->buffers = NULL; cpia2_allocate_buffers()
2257 cam->buffers[i].next = &cam->buffers[i+1]; cpia2_allocate_buffers()
2258 cam->buffers[i].data = cam->frame_buffer +i*cam->frame_size; cpia2_allocate_buffers()
2259 cam->buffers[i].status = FRAME_EMPTY; cpia2_allocate_buffers()
2260 cam->buffers[i].length = 0; cpia2_allocate_buffers()
2261 cam->buffers[i].max_length = 0; cpia2_allocate_buffers()
2262 cam->buffers[i].num = i; cpia2_allocate_buffers()
2264 cam->buffers[i].next = cam->buffers; cpia2_allocate_buffers()
2265 cam->buffers[i].data = cam->frame_buffer +i*cam->frame_size; cpia2_allocate_buffers()
2266 cam->buffers[i].status = FRAME_EMPTY; cpia2_allocate_buffers()
2267 cam->buffers[i].length = 0; cpia2_allocate_buffers()
2268 cam->buffers[i].max_length = 0; cpia2_allocate_buffers()
2269 cam->buffers[i].num = i; cpia2_allocate_buffers()
2270 cam->curbuff = cam->buffers; cpia2_allocate_buffers()
2272 DBG("buffers=%p, curbuff=%p, workbuff=%p\n", cam->buffers, cam->curbuff, cpia2_allocate_buffers()
2284 if(cam->buffers) { cpia2_free_buffers()
2285 kfree(cam->buffers); cpia2_free_buffers()
2286 cam->buffers = NULL; cpia2_free_buffers()
/linux-4.1.27/fs/jbd/
H A Dcommit.c43 * ->mapping, and with attached buffers. These pages are trivially reclaimable
48 * see if we can strip all buffers from the backing page.
116 * cleaned up our previous buffers by now, so if we are in abort
180 * Submit all the data buffers to disk
198 * Cleanup any flushed data buffers from the data list. Even in journal_submit_data_buffers()
375 * BJ_Reserved buffers. Note, it is _not_ permissible to assume journal_commit_transaction()
376 * that there are no such buffers: if a large filesystem journal_commit_transaction()
379 * there are still BJ_Reserved buffers outstanding. These must journal_commit_transaction()
384 * we do not require it to remember exactly which old buffers it journal_commit_transaction()
408 * Now try to drop any written-back buffers from the journal's journal_commit_transaction()
419 * Clear revoked flag to reflect there is no revoked buffers journal_commit_transaction()
511 * If we found any dirty or locked buffers, then we should have journal_commit_transaction()
549 * any descriptor buffers which may have been journal_commit_transaction()
615 rid of the BJ_IO/BJ_Shadow pairing of buffers. */ journal_commit_transaction()
704 complete. Control buffers being written are on the journal_commit_transaction()
705 transaction's t_log_list queue, and metadata buffers are on journal_commit_transaction()
708 Wait for the buffers in reverse order. That way we are journal_commit_transaction()
779 /* Here we wait for the revoke record and descriptor record buffers */ journal_commit_transaction()
822 processing: any buffers committed as a result of this journal_commit_transaction()
837 * As there are other places (journal_unmap_buffer()) adding buffers journal_commit_transaction()
962 * Now recheck if some buffers did not get attached to the transaction journal_commit_transaction()
H A Dcheckpoint.c197 * Clean up transaction's list of buffers submitted for io.
199 * buffers. Note that we take the buffers in the opposite ordering
202 * Return 0 on success, and return <0 if some buffers have failed
349 * list of transactions to be checkpointed and send all its buffers
461 * buffers which should be written-back to the filesystem.
550 * Find all the written-back checkpoint buffers in the given list and release
554 * Returns number of buffers reaped (for debug)
598 * Find all the written-back checkpoint buffers in the journal and release them.
602 * Returns number of buffers reaped (for debug)
634 * we can possibly see not yet submitted buffers on io_list __journal_clean_checkpoint_list()
652 * elsewhere on disk. To achieve this, all of the buffers in a
/linux-4.1.27/drivers/s390/char/
H A Dsclp_vt220.c63 /* List of empty pages to be used as write request buffers */
100 /* Registration structure for SCLP output event buffers */
106 /* Registration structure for SCLP input event buffers */
115 * buffers if necessary.
124 /* Put buffer back to list of empty buffers */ sclp_vt220_process_queue()
177 /* Remove processed buffers and requeue rest */ sclp_vt220_callback()
179 /* Not all buffers were processed */ sclp_vt220_callback()
231 /* Only emit buffers with content */ sclp_vt220_emit_current()
394 * more empty buffers available, wait until one gets emptied. If DO_SCHEDULE
481 * Called by the SCLP to report incoming event buffers.
561 * to change as output buffers get emptied, or if the output flow
605 * Pass on all buffers to the hardware. Return only when there are no more
606 * buffers pending.
H A Dsclp_rw.h60 /* The number of empty mto buffers that can be contained in a single sccb. */
/linux-4.1.27/arch/avr32/include/asm/
H A Ddma-mapping.h190 * dma_map_sg - map a set of SG buffers for streaming mode DMA
192 * @sg: list of buffers
193 * @nents: number of buffers to map
196 * Map a set of buffers described by scatterlist in streaming
229 * dma_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg
231 * @sg: list of buffers
232 * @nents: number of buffers to map
302 * @sg: list of buffers
303 * @nents: number of buffers to map
/linux-4.1.27/drivers/staging/goldfish/
H A Dgoldfish_audio.c54 * We will allocate two read buffers and two write buffers.
55 * Having two read buffers facilitate stereo -> mono conversion.
56 * Having two write buffers facilitate interleaved IO.
240 * if buffers are newly empty, wake up blocked goldfish_audio_interrupt()
/linux-4.1.27/drivers/infiniband/hw/ipath/
H A Dipath_wc_x86_64.c61 * Set the PIO buffers to be WCCOMB, so we get HT bursts to the ipath_enable_wc()
65 * the PIO buffers themselves. ipath_enable_wc()
66 * For chips with two sets of buffers, the calculations are ipath_enable_wc()
69 * The buffers are still packed, so a single range covers both. ipath_enable_wc()
159 ipath_cdbg(VERBOSE, "undoing WCCOMB on pio buffers\n"); ipath_disable_wc()
H A Dipath_init_chip.c45 * min buffers we want to have per port, after driver
51 * buffers per port, etc.) Zero means use chip value.
59 * Number of buffers reserved for driver (verbs and layered drivers.)
60 * Initialized based on number of PIO buffers if not set via module interface.
70 MODULE_PARM_DESC(kpiobufs, "Set number of PIO buffers for driver");
73 * create_port0_egr - allocate the eager TID buffers
80 * Allocate the eager TID buffers and program them into infinipath.
82 * memory, and either use the buffers as is for things like verbs
83 * packets, or pass the buffers up to the ipath layered driver and
315 * 4K buffers take 2 pages; we use roundup just to be init_chip_first()
549 * in initialization, to busy out buffers as needed enable_chip()
772 /* if not an even divisor, some user ports get extra buffers */ ipath_init_chip()
916 * PIO buffers in early_init to avoid false parity errors. ipath_init_chip()
/linux-4.1.27/fs/jbd2/
H A Dcommit.c56 * ->mapping, and with attached buffers. These pages are trivially reclaimable
61 * see if we can strip all buffers from the backing page.
113 * cleaned up our previous buffers by now, so if we are in abort
210 * Submit all the data buffers of inode associated with the transaction to
230 * submit the inode data buffers. We use writepage journal_submit_data_buffers()
475 * BJ_Reserved buffers. Note, it is _not_ permissible to assume jbd2_journal_commit_transaction()
476 * that there are no such buffers: if a large filesystem jbd2_journal_commit_transaction()
479 * there are still BJ_Reserved buffers outstanding. These must jbd2_journal_commit_transaction()
484 * we do not require it to remember exactly which old buffers it jbd2_journal_commit_transaction()
508 * Now try to drop any written-back buffers from the journal's jbd2_journal_commit_transaction()
519 * Clear revoked flag to reflect there is no revoked buffers jbd2_journal_commit_transaction()
605 * any descriptor buffers which may have been jbd2_journal_commit_transaction()
669 rid of the shadow pairing of buffers. */ jbd2_journal_commit_transaction()
812 complete. Control buffers being written are on the jbd2_journal_commit_transaction()
813 transaction's t_log_list queue, and metadata buffers are on jbd2_journal_commit_transaction()
816 Wait for the buffers in reverse order. That way we are jbd2_journal_commit_transaction()
865 /* Here we wait for the revoke record and descriptor record buffers */ jbd2_journal_commit_transaction()
919 processing: any buffers committed as a result of this jbd2_journal_commit_transaction()
932 * As there are other places (journal_unmap_buffer()) adding buffers jbd2_journal_commit_transaction()
1062 * Now recheck if some buffers did not get attached to the transaction jbd2_journal_commit_transaction()
H A Dcheckpoint.c202 * list of transactions to be checkpointed and send all its buffers
251 /* checkpoint all of the transaction's buffers */ jbd2_log_do_checkpoint()
326 * Now we issued all of the transaction's buffers, let's deal jbd2_log_do_checkpoint()
327 * with the buffers that are out for I/O. jbd2_log_do_checkpoint()
384 * buffers which should be written-back to the filesystem.
419 * Find all the written-back checkpoint buffers in the given list and
420 * release them. If 'destroy' is set, clean all buffers unconditionally.
464 * Find all the written-back checkpoint buffers in the journal and release them.
465 * If 'destroy' is set, release all buffers unconditionally.
497 * we can possibly see not yet submitted buffers on io_list __jbd2_journal_clean_checkpoint_list()
514 * Remove buffers from all checkpoint lists as journal is aborted and we just
542 * elsewhere on disk. To achieve this, all of the buffers in a
/linux-4.1.27/fs/nilfs2/
H A Dsegment.h103 * @sc_segbufs: List of segment buffers
104 * @sc_write_logs: List of segment buffers to hold logs under writing
105 * @sc_segbuf_nblocks: Number of available blocks in segment buffers.
129 * @sc_watermark: Watermark for the number of dirty buffers
151 /* Segment buffers */
H A Dsegbuf.h68 * @sb_segsum_buffers: List of buffers for segment summaries
69 * @sb_payload_buffers: List of buffers for segment payload
H A Dpage.h41 BUFFER_FNS(NILFS_Node, nilfs_node) /* nilfs node buffers */
/linux-4.1.27/include/linux/iio/
H A Dbuffer.h21 * struct iio_buffer_access_funcs - access functions for buffers.
73 * @buffer_list: [INTERN] entry in the devices list of current buffers.
126 * iio_push_to_buffers_with_timestamp() - push data and timestamp to buffers
131 * Pushes data to the IIO device's buffers. If timestamps are enabled for the
133 * the sample data buffer before pushing it to the device buffers. The sample
/linux-4.1.27/drivers/misc/carma/
H A Dcarma-fpga.c45 * capture rate, all buffers are pre-allocated to avoid any potentially long
49 * different states of data buffers.
52 * This list holds all empty data buffers which are ready to receive data.
59 * This list holds data buffers which have been filled, and are waiting to be
62 * All buffers start life on the free list, then move successively to the
67 * It should be noted that all buffers are mapped and ready for DMA when they
180 /* Correlation data buffers */
186 /* Information about data buffers */
334 * data_free_buffers() - free all allocated buffers
337 * Free all buffers allocated by the driver (except those currently in the
367 * data_alloc_buffers() - allocate 1 seconds worth of data buffers
370 * Allocate enough buffers for a whole second worth of data
373 * second worth of data buffers could not be allocated, as long as a minimum
403 /* add it to the list of free buffers */ data_alloc_buffers()
408 /* Make sure we allocated the minimum required number of buffers */ data_alloc_buffers()
410 dev_err(priv->dev, "Unable to allocate enough data buffers\n"); data_alloc_buffers()
418 "Unable to allocate %d buffers, using %d buffers instead\n", data_alloc_buffers()
708 * All buffers passed to this function should be ready and mapped data_submit_dma()
796 /* If there are no free buffers, drop this data */ data_irq()
837 * Enable the device for buffered dumping. Allocates buffers and hooks up
872 /* allocate the correlation data buffers */ data_device_enable()
875 dev_err(priv->dev, "unable to allocate buffers\n"); data_device_enable()
919 * all buffers.
962 /* free all buffers: the free and used lists are not being changed */ data_device_disable()
/linux-4.1.27/drivers/isdn/sc/
H A Dpacket.c45 pr_debug("%s: out of TX buffers\n", sndpkt()
164 pr_debug("%s: calculating buffer space: %d buffers, %d big\n", setup_buffers()
187 * Prep the receive buffers setup_buffers()
H A Dhardware.h24 #define BUFFER_OFFSET 0x1800 /* Beginning of buffers */
77 buffers per channel */
/linux-4.1.27/drivers/media/usb/stk1160/
H A Dstk1160.h36 /* Number of buffers for isoc transfers */
44 /* Max and min video buffers */
106 /* transfer buffers for isoc transfer */
H A Dstk1160-video.c328 /* Reset urb buffers */ stk1160_isoc_irq()
370 * Releases urb and transfer buffers
378 stk1160_dbg("freeing %d urb buffers...\n", num_bufs); stk1160_free_isoc()
408 stk1160_dbg("all urb buffers freed\n"); stk1160_free_isoc()
459 /* allocate urbs and transfer buffers */ stk1160_alloc_isoc()
479 /* Not enough transfer buffers, so just give up */ stk1160_alloc_isoc()
515 /* At last we can say we have some buffers */ stk1160_alloc_isoc()
536 /* Save the allocated buffers so far, so we can properly free them */ stk1160_alloc_isoc()
/linux-4.1.27/drivers/net/wireless/libertas/
H A Ddev.h32 u32 fwd_drop_nobuf; /* Fwd: Run out of internal buffers */
128 struct list_head cmdfreeq; /* free command buffers */
129 struct list_head cmdpendingq; /* pending command buffers */
/linux-4.1.27/drivers/gpu/drm/i915/
H A Di915_gem_batch_pool.c30 * In order to submit batch buffers as 'secure', the software command parser
34 * managing the address space bindings for such buffers.
37 * set of scratch buffers to use for this purpose. The framework can be
/linux-4.1.27/net/sunrpc/
H A Dbackchannel_rqst.c55 * buffers hanging off of it.
71 * Preallocate up to min_reqs structures and related buffers for use
74 * preallocated buffers are added to the pool of resources used by
100 * buffers. Once we're done building the list we splice it xprt_setup_backchannel()
237 * Return the preallocated rpc_rqst structure and XDR buffers
275 * backchannel setup. Buffer space for the send and private XDR buffers
H A Dsunrpc.h33 * Header for dynamically allocated rpc buffers.
/linux-4.1.27/drivers/net/ethernet/aeroflex/
H A Dgreth.h70 /* How many pages are needed for buffers */
76 * Set to 1520 to make all buffers word aligned for non-gbit MAC.
/linux-4.1.27/drivers/net/slip/
H A Dslip.h63 /* These are pointers to the malloc()ed frame buffers. */
70 int buffsize; /* Max buffers sizes */
/linux-4.1.27/drivers/media/usb/stkwebcam/
H A Dstk-webcam.h51 /* Streaming IO buffers */
120 /* Streaming buffers */
/linux-4.1.27/drivers/spi/
H A Dspi-fsl-lib.h53 /* functions to deal with different sized buffers */
84 /* functions to deal with different sized buffers */
/linux-4.1.27/drivers/staging/android/ion/
H A Dion_priv.h37 * @node: node in the ion_device buffers tree
47 * @lock: protects the buffers cnt fields
53 * handler and only valid for buffers that are faulted in
83 /* used to track orphaned buffers */
162 * Represents a pool of memory from which buffers can be made. In some
164 * On others, some blocks might require large physically contiguous buffers
198 * in, this can affect how buffers are allocated from the heap.
227 * some helpers for common operations on buffers using the sg_table
/linux-4.1.27/drivers/net/wireless/ti/wl1251/
H A Devent.h29 * The event mechanism is based on a pair of event buffers (buffers A and
H A Drx.h36 * The size of each of the two buffers is large enough to hold the longest
41 * 2) The host reads the received packet from one of the double buffers.
/linux-4.1.27/drivers/video/fbdev/mb862xx/
H A Dmb862xxfb.h61 unsigned long cap_buf; /* capture buffers offset */
62 size_t cap_len; /* length of capture buffers */
/linux-4.1.27/arch/sh/include/asm/
H A Dfixmap.h33 * these 'compile-time allocated' memory buffers are
38 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/Documentation/video4linux/
H A Dv4l2-pci-skeleton.c60 * @buf_list: list of buffers queued for DMA
161 * checks if sufficient buffers have been allocated. Usually 3 is a good
162 * minimum number: many DMA engines need a minimum of 2 buffers in the
244 * Start streaming. First check if the minimum number of buffers have been
247 * buffers are available to actually start the DMA engine.
260 * In case of an error, return all active buffers to the start_streaming()
269 * Stop the DMA engine. Any remaining buffers in the DMA queue are dequeued
278 /* Release all active buffers */ stop_streaming()
383 * It is not allowed to change the format while buffers for use with skeleton_s_fmt_vid_cap()
425 * the buffers for streaming have been set up, so we have to allow for skeleton_s_std()
433 * while buffers for use with streaming have already been allocated. skeleton_s_std()
517 * while buffers for use with streaming have already been allocated. skeleton_s_dv_timings()
636 * while buffers for use with streaming have already been allocated. skeleton_s_input()
699 * the buffers can call the streaming ioctls, all other filehandles will
833 * Assume that this DMA engine needs to have at least two buffers skeleton_probe()
835 * won't be called until at least this many buffers are queued up. skeleton_probe()
850 * the vb2 core will allocate the buffers in 32-bit DMA memory. skeleton_probe()
/linux-4.1.27/kernel/trace/
H A Dring_buffer.c119 * A fast way to enable or disable all ring buffers is to
120 * call tracing_on or tracing_off. Turning off the ring buffers
121 * prevents all ring buffers from being recorded to.
133 * will permantly disable all ring buffers.
137 * Global flag to disable all recording to ring buffers
142 * 0 0 : ring buffers are off
143 * 1 0 : ring buffers are on
144 * X 1 : ring buffers are permanently disabled
159 /* Used for individual buffers (after the counter) */
165 * tracing_off_permanent - permanently disable ring buffers
167 * This function, once called, will disable all ring buffers
504 struct ring_buffer_per_cpu **buffers; member in struct:ring_buffer
547 * as data is added to any of the @buffer's cpu buffers. Otherwise
569 cpu_buffer = buffer->buffers[cpu]; ring_buffer_wait()
648 * as data is added to any of the @buffer's cpu buffers. Otherwise
651 * Returns POLLIN | POLLRDNORM if data exists in the buffers,
666 cpu_buffer = buffer->buffers[cpu]; ring_buffer_poll_wait()
1368 buffer->buffers = kzalloc(ALIGN(bsize, cache_line_size()), __ring_buffer_alloc()
1370 if (!buffer->buffers) __ring_buffer_alloc()
1374 buffer->buffers[cpu] = for_each_buffer_cpu()
1376 if (!buffer->buffers[cpu]) for_each_buffer_cpu()
1393 if (buffer->buffers[cpu]) for_each_buffer_cpu()
1394 rb_free_cpu_buffer(buffer->buffers[cpu]); for_each_buffer_cpu()
1396 kfree(buffer->buffers);
1425 rb_free_cpu_buffer(buffer->buffers[cpu]); ring_buffer_free()
1431 kfree(buffer->buffers); ring_buffer_free()
1715 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
1744 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
1760 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
1775 cpu_buffer = buffer->buffers[cpu_id];
1824 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
1837 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
2754 cpu_buffer = buffer->buffers[cpu]; ring_buffer_lock_reserve()
2857 cpu_buffer = buffer->buffers[cpu]; ring_buffer_unlock_commit()
2953 cpu_buffer = buffer->buffers[cpu]; ring_buffer_discard_commit()
3017 cpu_buffer = buffer->buffers[cpu]; ring_buffer_write()
3163 cpu_buffer = buffer->buffers[cpu]; ring_buffer_record_disable_cpu()
3183 cpu_buffer = buffer->buffers[cpu]; ring_buffer_record_enable_cpu()
3216 cpu_buffer = buffer->buffers[cpu]; ring_buffer_oldest_event_ts()
3247 cpu_buffer = buffer->buffers[cpu]; ring_buffer_bytes_cpu()
3266 cpu_buffer = buffer->buffers[cpu]; ring_buffer_entries_cpu()
3286 cpu_buffer = buffer->buffers[cpu]; ring_buffer_overrun_cpu()
3309 cpu_buffer = buffer->buffers[cpu]; ring_buffer_commit_overrun_cpu()
3331 cpu_buffer = buffer->buffers[cpu]; ring_buffer_dropped_events_cpu()
3351 cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_events_cpu()
3371 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
3394 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
3893 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_peek()
3974 cpu_buffer = buffer->buffers[cpu]; ring_buffer_consume()
4032 cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_prepare()
4162 return BUF_PAGE_SIZE * buffer->buffers[cpu]->nr_pages; ring_buffer_size()
4215 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_reset_cpu()
4248 * @buffer: The ring buffer to reset all cpu buffers
4275 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu()
4309 cpu_buffer = buffer->buffers[cpu]; ring_buffer_empty_cpu()
4324 * ring_buffer_swap_cpu - swap a CPU buffer between two ring buffers
4344 cpu_buffer_a = buffer_a->buffers[cpu]; ring_buffer_swap_cpu()
4345 cpu_buffer_b = buffer_b->buffers[cpu]; ring_buffer_swap_cpu()
4347 /* At least make sure the two buffers are somewhat the same */ ring_buffer_swap_cpu()
4383 buffer_a->buffers[cpu] = cpu_buffer_b; ring_buffer_swap_cpu()
4384 buffer_b->buffers[cpu] = cpu_buffer_a; ring_buffer_swap_cpu()
4483 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_page()
4667 nr_pages = buffer->buffers[cpu_i]->nr_pages; for_each_buffer_cpu()
4668 if (nr_pages != buffer->buffers[cpu_i]->nr_pages) { for_each_buffer_cpu()
4676 buffer->buffers[cpu] =
4678 if (!buffer->buffers[cpu]) {
4715 * is displayed and all ring buffers are disabled.
/linux-4.1.27/arch/arm/mm/
H A Ddma-mapping.c886 * arm_dma_map_sg - map a set of SG buffers for streaming mode DMA
888 * @sg: list of buffers
889 * @nents: number of buffers to map
892 * Map a set of buffers described by scatterlist in streaming mode for DMA.
926 * arm_dma_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg
928 * @sg: list of buffers
929 * @nents: number of buffers to unmap (same as was passed to dma_map_sg)
950 * @sg: list of buffers
951 * @nents: number of buffers to map (returned from dma_map_sg)
969 * @sg: list of buffers
970 * @nents: number of buffers to map (returned from dma_map_sg)
1583 * arm_coherent_iommu_map_sg - map a set of SG buffers for streaming mode DMA
1585 * @sg: list of buffers
1586 * @nents: number of buffers to map
1589 * Map a set of i/o coherent buffers described by scatterlist in streaming
1601 * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA
1603 * @sg: list of buffers
1604 * @nents: number of buffers to map
1607 * Map a set of buffers described by scatterlist in streaming mode for DMA.
1637 * arm_coherent_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg
1639 * @sg: list of buffers
1640 * @nents: number of buffers to unmap (same as was passed to dma_map_sg)
1653 * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg
1655 * @sg: list of buffers
1656 * @nents: number of buffers to unmap (same as was passed to dma_map_sg)
1671 * @sg: list of buffers
1672 * @nents: number of buffers to map (returned from dma_map_sg)
1689 * @sg: list of buffers
1690 * @nents: number of buffers to map (returned from dma_map_sg)
/linux-4.1.27/drivers/isdn/hysdn/
H A Dhysdn_net.c29 #define MAX_SKB_BUFFERS 20 /* number of buffers for keeping TX-data */
47 int sk_count; /* number of buffers currently in ring */
84 /* flush the currently occupied tx-buffers */
304 flush_tx_buffers((struct net_local *) dev); /* empty buffers */ hysdn_net_release()
H A Dhysdn_proclog.c98 /* Flushes buffers not longer in use. */
236 /* count is decremented and all buffers are noticed of closing. If this file */
237 /* was the last one to be closed, all buffers are freed. */
266 inf->usage_cnt--; /* decrement usage count for buffers */ hysdn_log_close()
271 if (pd->if_used <= 0) /* delete buffers if last file closed */ hysdn_log_close()
/linux-4.1.27/drivers/media/usb/au0828/
H A Dau0828.h142 /* transfer buffers for isoc transfer */
156 /* Stores already requested buffers */
269 char *transfer_buffer[AU0828_MAX_ISO_BUFS];/* transfer buffers for isoc
276 /* Preallocated transfer digital transfer buffers */
/linux-4.1.27/drivers/net/ethernet/sfc/
H A Dworkarounds.h30 /* Spurious parity errors in TSORT buffers */
/linux-4.1.27/drivers/input/serio/
H A Di8042.h51 * the i8042 buffers.
/linux-4.1.27/arch/tile/lib/
H A Dchecksum.c55 * large buffers, but those should be impossible. do_csum()
/linux-4.1.27/arch/x86/include/asm/
H A Dkgdb.h11 * buffers at least NUMREGBYTES*2 are needed for register packets
H A Dfixmap.h60 * These 'compile-time allocated' memory buffers are
65 * TLB entries of such buffers will not be flushed across
/linux-4.1.27/drivers/gpu/drm/udl/
H A Dudl_transfer.c101 * multiple buffers to be potentially encoded and sent in parallel).
105 * rendering can be done directly to output buffers (e.g. USB URBs).
108 * buffers if the line will take several buffers to complete.
/linux-4.1.27/fs/xfs/libxfs/
H A Dxfs_log_recover.h35 * item headers are in ri_buf[0]. Additional buffers follow.
H A Dxfs_symlink_remote.c129 /* no verification of non-crc buffers */ xfs_symlink_read_verify()
149 /* no verification of non-crc buffers */ xfs_symlink_write_verify()
/linux-4.1.27/include/net/
H A Daf_rxrpc.h20 * the mark applied to socket buffers that may be intercepted
/linux-4.1.27/include/uapi/drm/
H A Dmsm_drm.h62 * GEM buffers:
173 /* Each cmdstream submit consists of a table of buffers involved, and
174 * one or more cmdstream buffers. This allows for conditional execution
175 * (context-restore), and IB buffers needed for per tile/bin draw cmds.
/linux-4.1.27/arch/powerpc/sysdev/
H A Dudbg_memcons.c3 * buffers.
/linux-4.1.27/arch/mips/pnx833x/common/
H A Dsetup.c41 /* fake pci bus to avoid bounce buffers */ plat_mem_setup()
/linux-4.1.27/arch/mn10300/include/asm/
H A Dkgdb.h17 * buffers at least NUMREGBYTES*2 are needed for register packets
/linux-4.1.27/arch/powerpc/include/uapi/asm/
H A Dps3fb.h42 __u32 num_frames; /* num of frame buffers */
/linux-4.1.27/arch/arm64/include/asm/
H A Dfixmap.h28 * These 'compile-time allocated' memory buffers are
/linux-4.1.27/arch/c6x/platforms/
H A Demif.c60 /* look for a dscr-based enable for emifa pin buffers */ c6x_emifa_init()
/linux-4.1.27/fs/ext4/
H A Dtruncate.h9 * pagecache as well so that corresponding buffers get properly unmapped.
/linux-4.1.27/include/drm/
H A Ddrm_gem_cma_helper.h11 * @sgt: scatter/gather table for imported PRIME buffers
/linux-4.1.27/drivers/net/caif/
H A Dcaif_virtio.c83 * @tx_release_tasklet: Tasklet for freeing consumed TX buffers
92 * @allocsz: size of dma memory reserved for TX buffers
93 * @alloc_addr: virtual address to dma memory for TX buffers
94 * @alloc_dma: dma address to dma memory for TX buffers
95 * @genpool: Gen Pool used for allocating TX buffers
176 /* Stop looping if there are no more buffers to free */ cfv_release_used_buf()
466 /* Release any TX buffers on both used and avilable rings */ cfv_netdev_close()
531 /* garbage collect released buffers */ cfv_netdev_tx()
538 * Flow-on is triggered when sufficient buffers are freed cfv_netdev_tx()
564 /* Enable notifications of recycled TX buffers */ cfv_netdev_tx()
/linux-4.1.27/drivers/net/wireless/iwlwifi/
H A Diwl-fh.h103 * circular buffers (CBs/queues) containing Transmit Frame Descriptors (TFDs)
139 * concatenate up to 20 DRAM buffers to form a Tx frame, each Receive Buffer
199 * buffers waiting to be filled.
389 * (channel's buffers empty | no pending requests).
392 * 31-24: 1 = Channel buffers empty (channel 7:0)
504 * Both driver and device share these circular buffers, each of which must be
510 * Each TFD contains pointer/size information for up to 20 data buffers
511 * in host DRAM. These buffers collectively contain the (one) frame described
513 * itself, but buffers may be scattered in host DRAM. Each buffer has max size
514 * of (4K - 4). The concatenates all of a TFD's buffers into a single
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
H A Dadapter.h105 unsigned int credits; /* # of available Rx buffers */
106 unsigned int pend_cred; /* new buffers since last FL DB ring */
119 unsigned long empty; /* # of times queue ran out of buffers */
211 struct timer_list tx_reclaim_timer; /* reclaims TX buffers */
212 struct timer_list rx_reclaim_timer; /* reclaims RX buffers */
/linux-4.1.27/arch/tile/include/arch/
H A Dmpipe.h106 * This MMIO region is used for posting or fetching buffers to/from the
152 * This MMIO region is used for posting or fetching buffers to/from the
184 * egressed, buffers with HWB set will be returned to the other mPIPE
193 * Ignored on writes since all buffers on a given stack are the same size.
308 * software to determine when buffers associated with a descriptor may be
/linux-4.1.27/sound/oss/dmasound/
H A Ddmasound_core.c418 if (sq->buffers) sq_allocate_buffers()
422 sq->buffers = kmalloc (num * sizeof(char *), GFP_KERNEL); sq_allocate_buffers()
423 if (!sq->buffers) sq_allocate_buffers()
426 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); sq_allocate_buffers()
427 if (!sq->buffers[i]) { sq_allocate_buffers()
429 dmasound.mach.dma_free(sq->buffers[i], size); sq_allocate_buffers()
430 kfree(sq->buffers); sq_allocate_buffers()
431 sq->buffers = NULL; sq_allocate_buffers()
442 if (sq->buffers) { sq_release_buffers()
444 dmasound.mach.dma_free(sq->buffers[i], sq->bufSize); sq_release_buffers()
445 kfree(sq->buffers); sq_release_buffers()
446 sq->buffers = NULL; sq_release_buffers()
609 dest = write_sq.buffers[write_sq.rear]; sq_write()
650 dest = write_sq.buffers[(write_sq.rear+1) % write_sq.max_count]; sq_write()
734 /* allocate the default number & size of buffers. sq_open2()
822 sound_silence(); /* this _must_ stop DMA, we might be about to lose the buffers */ sq_reset_output()
1007 read/write - the app doesn't care about our internal buffers. sq_ioctl()
1469 * buffers being other than powers of two. This is not checked yet. dmasound_setup()
1481 printk("dmasound_setup: invalid number of buffers, using default = %d\n", numWriteBufs); dmasound_setup()
/linux-4.1.27/sound/usb/misc/
H A Dua101.c131 } buffers[MAX_MEMORY_BUFFERS]; member in struct:ua101::ua101_stream
1071 for (i = 0; i < ARRAY_SIZE(stream->buffers); ++i) { alloc_stream_buffers()
1074 stream->buffers[i].addr = alloc_stream_buffers()
1076 &stream->buffers[i].dma); alloc_stream_buffers()
1077 if (!stream->buffers[i].addr) alloc_stream_buffers()
1079 stream->buffers[i].size = size; alloc_stream_buffers()
1095 for (i = 0; i < ARRAY_SIZE(stream->buffers); ++i) free_stream_buffers()
1097 stream->buffers[i].size, free_stream_buffers()
1098 stream->buffers[i].addr, free_stream_buffers()
1099 stream->buffers[i].dma); free_stream_buffers()
1109 for (b = 0; b < ARRAY_SIZE(stream->buffers); ++b) { alloc_stream_urbs()
1110 unsigned int size = stream->buffers[b].size; alloc_stream_urbs()
1111 u8 *addr = stream->buffers[b].addr; alloc_stream_urbs()
1112 dma_addr_t dma = stream->buffers[b].dma; alloc_stream_urbs()
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/
H A Dmyri10ge_mcp.h172 * only be done if the driver has already supplied buffers
227 /* data0 = number of big buffers to use. It must be 0 or a power of 2.
228 * 0 indicates that the NIC consumes as many buffers as they are required
231 * number of buffers for each big receive packet.
373 /* When set, small receive buffers can cross page boundaries.
374 * Both small and big receive buffers may start at any address.
/linux-4.1.27/drivers/net/wireless/iwlwifi/pcie/
H A Dinternal.h84 * @free_count: Number of pre-allocated buffers in rx_free
87 * @rx_used: List of Rx buffers with no SKB
154 * the software buffers (in the variables @meta, @txb in struct
155 * iwl_txq) only have 32 entries, while the HW buffers (@tfds in
208 * @scratchbufs: start of command headers, including scratch buffers, for
253 * @rx_replenish: work that will be called when buffers need to be allocated
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/
H A Dselftest.h282 /** total # request buffers */
284 /** # posted request buffers */
288 /** allocate more buffers if scd_buf_nposted < scd_buf_low */
290 /** increase/decrease some buffers */
292 /** posted message buffers */
301 /* extra buffers for tolerating buggy peers, or unbalanced number
/linux-4.1.27/drivers/firmware/
H A Dqcom_scm.c82 * There can be arbitrary padding between the headers and buffers so
84 * to access the buffers in a safe manner.
112 * and response headers as well as the command and response buffers.
269 * Note that any buffers that are expected to be accessed by the secure world
272 * and response buffers is taken care of by qcom_scm_call; however, callers are
273 * responsible for any other cached buffers passed over to the secure world.
/linux-4.1.27/drivers/gpu/drm/omapdrm/
H A Domap_gem.c86 * allocated for buffers originally allocated from contiguous memory
125 /* To deal with userspace mmap'ings of 2d tiled buffers, which (a) are
201 * contiguous buffers can be remapped in TILER/DMM if they need to be
212 * shmem buffers that are mapped cached can simulate coherency via using
242 /* for non-cached buffers, ensure the new pages are clean because omap_gem_attach_pages()
280 /* for non-cached buffers, ensure the new pages are clean because omap_gem_detach_pages()
340 /* for tiled buffers, the virtual size has stride rounded up omap_gem_mmap_size()
365 /* Normal handling for the case of faulting in non-tiled buffers */ fault_1d()
391 /* Special handling for the case of faulting in 2d tiled buffers */ fault_2d()
577 * We do have some private objects, at least for scanout buffers omap_gem_mmap_obj()
809 * could potentially unpin and unmap buffers from TILER
842 * (only valid for tiled 2d buffers)
860 /* Get tiler stride for the buffer (only valid for 2d tiled buffers) */ omap_gem_tiled_stride()
1353 dev_err(dev->dev, "Tiled buffers require DMM\n"); omap_gem_new()
1357 /* tiled buffers are always shmem paged backed.. when they are omap_gem_new()
1362 /* currently don't allow cached buffers.. there is some caching omap_gem_new()
1391 * have DMM for remappign discontiguous buffers omap_gem_new()
/linux-4.1.27/drivers/media/platform/omap3isp/
H A Dispstat.c44 * buffers the allocated memory should fit. For that case, a recover
49 * buffers making it extreamily unlikely to have an access to invalid memory
202 /* Don't select uninitialised buffers if it's not required */ __isp_stat_buf_find()
355 dev_dbg(stat->isp->dev, "%s: all buffers were freed.\n", isp_stat_bufs_free()
390 * in the DMA buffers being mapped through the ISP IOMMU.
394 * passes NULL instead, resulting in the buffers being mapped directly as
398 * buffers could be allocated with kmalloc/vmalloc, but we still use
412 /* Are the old buffers big enough? */ isp_stat_bufs_alloc()
608 * isp_stat_buf_process - Process statistic buffers.
701 * ignore next buffers even if it was disabled in the meantime. isp_stat_try_enable()
702 * On the other hand, Histogram shouldn't ignore buffers anymore isp_stat_try_enable()
729 * the correct memory address, 2 buffers must be ignored. omap3isp_stat_sbl_overflow()
847 * necessary to not inform too early the buffers aren't busy in case __stat_isr()
/linux-4.1.27/drivers/virtio/
H A Dvirtio_ring.c69 /* Host supports indirect buffers */
166 * buffers, then go indirect. FIXME: tune this threshold */ virtqueue_add()
203 /* We're about to use some buffers from the free list. */ virtqueue_add()
259 * virtqueue_add_sgs - expose buffers to other end
292 * virtqueue_add_outbuf - expose output buffers to other end
314 * virtqueue_add_inbuf - expose input buffers to other end
467 * Returns NULL if there are no used buffers, or the "data" token
485 pr_debug("No more buffers in queue\n"); virtqueue_get_buf()
576 * virtqueue_poll - query pending used buffers
580 * Returns "true" if there are pending used buffers in the queue.
598 * buffers in the queue, to detect a possible race between the driver
616 * interrupts until most of the available buffers have been processed;
617 * it returns "false" if there are many pending buffers in the queue,
/linux-4.1.27/drivers/mtd/nand/
H A Dmpc5121_nfc.c181 /* Send data from NFC buffers to NAND flash */ mpc5121_nfc_send_prog_page()
405 /* Copy data from/to NFC spare buffers. */ mpc5121_nfc_copy_spare()
414 * NAND spare area is available through NFC spare buffers. mpc5121_nfc_copy_spare()
457 /* Copy data from/to NFC main and spare buffers */ mpc5121_nfc_buf_copy()
497 /* Read data from NFC buffers */ mpc5121_nfc_read_buf()
503 /* Write data to NFC buffers */ mpc5121_nfc_write_buf()
510 /* Read byte from NFC buffers */ mpc5121_nfc_read_byte()
520 /* Read word from NFC buffers */ mpc5121_nfc_read_word()
/linux-4.1.27/drivers/net/ethernet/pasemi/
H A Dpasemi_mac.h51 u64 *buffers; /* RX interface buffer ring */ member in struct:pasemi_mac_rxring
114 #define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
/linux-4.1.27/drivers/media/pci/pt3/
H A Dpt3_dma.c122 /* mark the whole buffers as "not written yet" */ pt3_init_dmabuf()
177 /* build circular-linked pointers (xfer_desc) to the data buffers*/ pt3_alloc_dmabuf()
/linux-4.1.27/drivers/media/platform/
H A Dvia-camera.h29 #define VCR_CI_3BUFS 0x00000004 /* 1 = 3 buffers, 0 = 2 buffers */
/linux-4.1.27/drivers/misc/mic/host/
H A Dmic_fops.c132 * We return POLLIN | POLLOUT from poll when new buffers are enqueued, and
133 * not when previously enqueued buffers may be available. This means that
/linux-4.1.27/drivers/staging/media/dt3155v4l/
H A Ddt3155v4l.h165 * @free_bufs_empty: no free image buffers
186 * @dmaq queue for dma buffers
/linux-4.1.27/drivers/pcmcia/
H A Dtopic.h153 /* Disable write buffers to prevent lockups under load with numerous topic95_override()
162 dev_info(&socket->dev->dev, "Disabled ToPIC95 Cardbus write buffers.\n"); topic95_override()
/linux-4.1.27/drivers/gpu/drm/radeon/
H A Dradeon_cs.c56 /* Since buffers which appear sooner in the relocation list are radeon_cs_buckets_add()
57 * likely to be used more often than buffers which appear later radeon_cs_buckets_add()
58 * in the list, the sort mustn't change the ordering of buffers radeon_cs_buckets_add()
114 * Also, the buffers used for write have a higher priority than radeon_cs_parser_relocs()
115 * the buffers used for read only, which doubles the range radeon_cs_parser_relocs()
402 * which affects the order of buffers in the LRU list. radeon_cs_parser_fini()
403 * This assures that the smallest buffers are added first radeon_cs_parser_fini()
405 * first, instead of large buffers whose eviction is more radeon_cs_parser_fini()
/linux-4.1.27/drivers/gpu/drm/ttm/
H A Dttm_execbuf_util.c84 * Reserve buffers for validation.
92 * buffers in different orders.
/linux-4.1.27/fs/xfs/
H A Dxfs_icreate_item.c100 * Because we have ordered buffers being tracked in the AIL for the inode
125 /* Ordered buffers do the dependency tracking here, so this does nothing. */
H A Dxfs_buf_item.c121 * Discontiguous buffers need a format structure per region that that is being
123 * they came from separate buffers, just like would occur if multiple buffers
125 * discontiguous buffers to be in-memory constructs, completely transparent to
173 * buffers we need to track which segment the dirty bits correspond to, xfs_buf_item_size()
336 * For icreate item based inode allocation, the buffers aren't written xfs_buf_item_format()
587 * (cancelled) buffers at unpin time, but we'll never go through the xfs_buf_item_unlock()
622 * Ordered buffers are dirty but may have no recorded changes, so ensure xfs_buf_item_unlock()
638 * Clean buffers, by definition, cannot be in the AIL. However, aborted xfs_buf_item_unlock()
639 * buffers may be dirty and hence in the AIL. Therefore if we are xfs_buf_item_unlock()
671 * The one exception to this is for buffers full of newly allocated
672 * inodes. These buffers are only relogged with the XFS_BLI_INODE_BUF
674 * inodes in the buffers will be replayed during recovery. If the
1044 * This is the iodone() function for buffers which have had callbacks
1126 * This is the iodone() function for buffers which have been
1146 * log-committed callbacks to unpin these buffers. Or we may never xfs_buf_iodone()
H A Dxfs_buf_item.h50 * needed to log buffers. It tracks how many times the lock has been
/linux-4.1.27/include/net/caif/
H A Dcaif_hsi.h27 * Decides if HSI buffers should be prefilled with 0xFF pattern for easier
28 * debugging. Both TX and RX buffers will be filled before the transfer.
H A Dcaif_spi.h49 /* Decides if SPI buffers should be prefilled with 0xFF pattern for easier
50 * debugging. Both TX and RX buffers will be filled before the transfer.
/linux-4.1.27/arch/s390/appldata/
H A Dappldata_mem.c56 u64 bufferram; /* memory reserved for buffers, free cache */
57 u64 cached; /* size of (used) cache, w/o buffers */
/linux-4.1.27/drivers/media/pci/meye/
H A Dmeye.h272 /* Maximum number of buffers */
305 /* list of buffers */
309 struct kfifo grabq; /* queue for buffers to be grabbed */
311 struct kfifo doneq; /* queue for grabbed buffers */
/linux-4.1.27/drivers/media/platform/s3c-camif/
H A Dcamif-core.h193 * @pending_buf_q: pending (empty) buffers queue head
194 * @active_buf_q: active (being written) buffers queue head
198 * @reqbufs_count: the number of buffers requested
320 * @list: list head for the buffers queue
/linux-4.1.27/drivers/net/ethernet/sun/
H A Dsunvnet.h26 /* VNET packets are sent in buffers with the first 6 bytes skipped
/linux-4.1.27/drivers/staging/iio/adc/
H A Dad7606.h77 * transfer buffers to live in their own cache lines.
/linux-4.1.27/drivers/staging/iio/frequency/
H A Dad9834.h66 * transfer buffers to live in their own cache lines.
/linux-4.1.27/arch/x86/kernel/
H A Dpci-nommu.c41 /* Map a set of buffers described by scatterlist in streaming
/linux-4.1.27/arch/xtensa/include/asm/
H A Dfixmap.h32 * these 'compile-time allocated' memory buffers are
/linux-4.1.27/drivers/gpu/drm/qxl/
H A Dqxl_prime.c28 * device that might share buffers with qxl */
/linux-4.1.27/drivers/char/xillybus/
H A Dxillybus_core.c81 * buffers' end_offset fields against changes made by IRQ handler (and in
323 struct xilly_buffer **buffers, xilly_get_dma_buffers()
331 if (buffers) { /* Not the message buffer */ xilly_get_dma_buffers()
383 if (buffers) { /* Not the message buffer */ xilly_get_dma_buffers()
386 buffers[i] = this_buffer++; xilly_get_dma_buffers()
486 struct xilly_buffer **buffers = NULL; xilly_setupchannels() local
515 buffers = devm_kcalloc(dev, bufnum, xilly_setupchannels()
518 if (!buffers) xilly_setupchannels()
531 channel->rd_buffers = buffers; xilly_setupchannels()
532 rc = xilly_get_dma_buffers(ep, &rd_alloc, buffers, xilly_setupchannels()
544 channel->wr_buffers = buffers; xilly_setupchannels()
545 rc = xilly_get_dma_buffers(ep, &wr_alloc, buffers, xilly_setupchannels()
846 /* Don't request more than all buffers */ xillybus_read()
1106 while (1) { /* Loop waiting for draining of buffers */ xillybus_myflush()
1608 * we have no leftover buffers pending before wrapping xillybus_release()
2023 /* Bootstrap phase III: Allocate buffers according to IDT */ xillybus_endpoint_discovery()
321 xilly_get_dma_buffers(struct xilly_endpoint *ep, struct xilly_alloc_state *s, struct xilly_buffer **buffers, int bufnum, int bytebufsize) xilly_get_dma_buffers() argument
/linux-4.1.27/arch/blackfin/lib/
H A Dins.S71 * buffers in/out of FIFOs.
/linux-4.1.27/fs/ext2/
H A Dfile.c84 "detected IO error when writing metadata buffers"); ext2_fsync()
/linux-4.1.27/fs/gfs2/
H A Dmeta_io.c73 * The page and its buffers are protected by PageWriteback(), so we can gfs2_aspace_writepage()
298 * gfs2_meta_wipe - make inode's buffers so they aren't dirty/pinned anymore
299 * @ip: the inode who owns the buffers
301 * @blen: the number of buffers in the run
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c62 * buffers which the hardware consumes) and free list entries are
78 * Max number of Rx buffers we replenish at a time. Again keep this
79 * modest, allocating buffers isn't cheap either.
257 * Tests specified Free List to see whether the number of buffers
368 * free_tx_desc - reclaims TX descriptors and their buffers
372 * @unmap: whether the buffers should be unmapped for DMA
375 * TX buffers. Called with the TX queue lock held.
424 * @unmap: whether the buffers should be unmapped for DMA
427 * and frees the associated buffers if possible. Called with the TX
464 * free_rx_bufs - free RX buffers on an SGE Free List
466 * @fl: the SGE Free List to free buffers from
467 * @n: how many buffers to free
469 * Release the next @n buffers on an SGE Free List RX queue. The
470 * buffers must be made inaccessible to hardware before calling this
581 * Support for poisoning RX buffers ...
596 * @n: the number of new buffers to allocate
599 * (Re)populate an SGE free-buffer queue with up to @n new packet buffers,
603 * of buffers allocated. If afterwards the queue is found critically low,
617 * Sanity: ensure that the result of adding n Free List buffers refill_fl()
626 * If we support large pages, prefer large buffers and fail over to refill_fl()
708 * buffers, tell the hardware about them and return the number of refill_fl()
709 * buffers which we were able to allocate. refill_fl()
900 * Generates a Scatter/Gather List for the buffers that make up a packet.
1498 * buff! size, which is expected since buffers are at least t4vf_pktgl_to_skb()
1670 * restore_rx_bufs - put back a packet's RX buffers
1680 * We implement the suspension by putting all of the RX buffers associated
1681 * with the current packet back on the original Free List. The buffers
2033 * Normally new RX buffers are added when existing ones are consumed but
2045 * Lists in need of more free buffers. If we find one and it's not sge_rx_timer_cb()
2047 * to refill it. If we're successful in adding enough buffers to push sge_rx_timer_cb()
2633 /* A FL with <= fl_starve_thres buffers is starving and a periodic t4vf_sge_init()
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
H A Di40e_adminq.c127 * This assumes the posted send buffers have already been cleaned
139 * This assumes the posted receive buffers have already been cleaned
148 * i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
159 * allocate the mapped buffers for the event processing i40e_alloc_arq_bufs()
169 /* allocate the mapped buffers */ i40e_alloc_arq_bufs()
231 /* allocate the mapped buffers */ i40e_alloc_asq_bufs()
346 /* Update tail in the HW to post pre-allocated buffers */ i40e_config_arq_regs()
396 /* allocate buffers in the rings */ i40e_init_asq()
455 /* allocate buffers in the rings */ i40e_init_arq()
500 /* free ring buffers */ i40e_shutdown_asq()
533 /* free ring buffers */ i40e_shutdown_arq()
926 * @e: event info from the receive descriptor, includes any buffers
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
H A Di40e_adminq.c125 * This assumes the posted send buffers have already been cleaned
137 * This assumes the posted receive buffers have already been cleaned
146 * i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
157 * allocate the mapped buffers for the event processing i40e_alloc_arq_bufs()
167 /* allocate the mapped buffers */ i40e_alloc_arq_bufs()
229 /* allocate the mapped buffers */ i40e_alloc_asq_bufs()
344 /* Update tail in the HW to post pre-allocated buffers */ i40e_config_arq_regs()
394 /* allocate buffers in the rings */ i40e_init_asq()
453 /* allocate buffers in the rings */ i40e_init_arq()
498 /* free ring buffers */ i40e_shutdown_asq()
531 /* free ring buffers */ i40e_shutdown_arq()
878 * @e: event info from the receive descriptor, includes any buffers

Completed in 9428 milliseconds

1234567891011