/linux-4.1.27/lib/xz/ |
H A D | xz_dec_test.c | 42 * Input and output buffers. The input buffer is used as a temporary safe 49 * Structure to pass the input and output buffers to the XZ decoder. 52 static struct xz_buf buffers = { variable in typeref:struct:xz_buf 75 buffers.in_pos = 0; xz_dec_test_open() 76 buffers.in_size = 0; xz_dec_test_open() 77 buffers.out_pos = 0; xz_dec_test_open() 120 while ((remaining > 0 || buffers.out_pos == buffers.out_size) xz_dec_test_write() 122 if (buffers.in_pos == buffers.in_size) { xz_dec_test_write() 123 buffers.in_pos = 0; xz_dec_test_write() 124 buffers.in_size = min(remaining, sizeof(buffer_in)); xz_dec_test_write() 125 if (copy_from_user(buffer_in, buf, buffers.in_size)) xz_dec_test_write() 128 buf += buffers.in_size; xz_dec_test_write() 129 remaining -= buffers.in_size; xz_dec_test_write() 132 buffers.out_pos = 0; xz_dec_test_write() 133 ret = xz_dec_run(state, &buffers); xz_dec_test_write() 134 crc = crc32(crc, buffer_out, buffers.out_pos); xz_dec_test_write() 145 return size - remaining - (buffers.in_size - buffers.in_pos); xz_dec_test_write()
|
/linux-4.1.27/drivers/media/pci/ivtv/ |
H A D | ivtv-queue.c | 47 q->buffers = 0; ivtv_queue_init() 65 q->buffers++; ivtv_enqueue() 80 q->buffers--; ivtv_dequeue() 94 from->buffers--; ivtv_queue_move_buf() 100 to->buffers++; ivtv_queue_move_buf() 105 /* Move 'needed_bytes' worth of buffers from queue 'from' into queue 'to'. 106 If 'needed_bytes' == 0, then move all buffers from 'from' into 'to'. 107 If 'steal' != NULL, then buffers may also taken from that queue if 111 also cleared if buffers need to be taken from the 'steal' queue and 119 -ENOMEM is returned if the buffers could not be obtained, 0 if all 120 buffers where obtained from the 'from' list and if non-zero then 121 the number of stolen buffers is returned. */ ivtv_queue_move() 148 /* move buffers from the tail of the 'steal' queue to the tail of the ivtv_queue_move() 149 'from' queue. Always copy all the buffers with the same dma_xfer_cnt ivtv_queue_move() 151 if one frame is stored in multiple buffers. */ ivtv_queue_move() 155 steal->buffers--; ivtv_queue_move() 159 from->buffers++; ivtv_queue_move() 196 int SGsize = sizeof(struct ivtv_sg_host_element) * s->buffers; ivtv_stream_alloc() 199 if (s->buffers == 0) ivtv_stream_alloc() 202 IVTV_DEBUG_INFO("Allocate %s%s stream: %d x %d buffers (%dkB total)\n", ivtv_stream_alloc() 204 s->name, s->buffers, s->buf_size, s->buffers * s->buf_size / 1024); ivtv_stream_alloc() 238 /* allocate stream buffers. Initially all buffers are in q_free. */ ivtv_stream_alloc() 239 for (i = 0; i < s->buffers; i++) { ivtv_stream_alloc() 258 if (i == s->buffers) ivtv_stream_alloc() 260 IVTV_ERR("Couldn't allocate buffers for %s stream\n", s->name); ivtv_stream_alloc() 269 /* move all buffers to q_free */ ivtv_stream_free()
|
/linux-4.1.27/drivers/media/platform/vivid/ |
H A D | vivid-vid-common.c | 54 .buffers = 1, 64 .buffers = 1, 73 .buffers = 1, 82 .buffers = 1, 91 .buffers = 1, 100 .buffers = 1, 109 .buffers = 1, 118 .buffers = 1, 127 .buffers = 1, 136 .buffers = 1, 145 .buffers = 1, 154 .buffers = 1, 163 .buffers = 1, 171 .buffers = 1, 180 .buffers = 1, 188 .buffers = 1, 197 .buffers = 1, 207 .buffers = 1, 215 .buffers = 1, 223 .buffers = 1, 232 .buffers = 1, 241 .buffers = 1, 249 .buffers = 1, 257 .buffers = 1, 266 .buffers = 1, 275 .buffers = 1, 284 .buffers = 1, 294 .buffers = 1, 302 .buffers = 1, 310 .buffers = 1, 319 .buffers = 1, 327 .buffers = 1, 335 .buffers = 1, 343 .buffers = 1, 351 .buffers = 1, 359 .buffers = 1, 367 .buffers = 1, 375 .buffers = 1, 384 .buffers = 1, 393 .buffers = 1, 401 .buffers = 1, 409 .buffers = 1, 417 .buffers = 1, 426 .buffers = 2, 436 .buffers = 2, 446 .buffers = 3, 455 .buffers = 3, 464 .buffers = 2, 473 .buffers = 2, 488 if (fmt->buffers == 1 || dev->multiplanar) vivid_get_format()
|
/linux-4.1.27/include/media/ |
H A D | videobuf-dma-contig.h | 2 * helper functions for physically contiguous capture buffers 5 * (i.e. the buffers must be linear in physical memory)
|
H A D | videobuf-dma-sg.h | 2 * helper functions for SG DMA video4linux capture buffers 5 * (i.e. the buffers are not linear in physical memory, but fragmented 28 * A small set of helper functions to manage buffers (both userland 54 /* for kernel buffers */ 60 /* for overlay buffers (pci-pci dma) */ 73 /* for mmap'ed buffers */
|
H A D | videobuf-vmalloc.h | 2 * helper functions for vmalloc capture buffers 5 * (i.e. the buffers are not linear in physical memory, but fragmented
|
H A D | v4l2-mem2mem.h | 4 * Helper functions for devices that use memory buffers for both source 58 /* Queue for buffers ready to be processed as soon as this 160 * v4l2_m2m_num_src_bufs_ready() - return the number of source buffers ready for 170 * v4l2_m2m_num_src_bufs_ready() - return the number of destination buffers 183 * buffers 192 * ready buffers 200 * v4l2_m2m_get_src_vq() - return vb2_queue for source buffers 209 * v4l2_m2m_get_dst_vq() - return vb2_queue for destination buffers 221 * buffers and return it 230 * ready buffers and return it
|
H A D | videobuf2-core.h | 179 * @queued_entry: entry on the queued buffers list, which holds all 180 * buffers queued from userspace 181 * @done_entry: entry on the list that stores all buffers ready to 239 * smaller number of buffers. Driver should return 240 * the required number of buffers in *num_buffers, the 247 * of buffers, that are being allocated. When called from 280 * buffers are being returned to their default DEQUEUED 287 * receive buffers with @buf_queue callback before 289 * of already queued buffers in count parameter; driver 291 * buffers that have been already given by the @buf_queue 294 * If you need a minimum number of buffers before you can 298 * many buffers have been queued up by userspace. 301 * finish and give back all buffers it got from buf_queue() 310 * pre-queued buffers before calling STREAMON. 346 * @owner: The filehandle that 'owns' the buffers, i.e. the filehandle 358 * @gfp_flags: additional gfp flags used when allocating the buffers. 361 * @min_buffers_needed: the minimum number of buffers needed before 363 * cannot be started unless at least this number of buffers 366 * @mmap_lock: private mutex used when buffers are allocated/freed/mmapped 369 * @num_buffers: number of allocated/used buffers 370 * @queued_list: list of buffers currently queued from userspace 371 * @queued_count: number of buffers queued and ready for streaming. 372 * @owned_by_drv_count: number of buffers owned by the driver 373 * @done_list: list of buffers ready to be dequeued to userspace 375 * @done_wq: waitqueue for processes waiting for buffers ready to be dequeued 382 * buffers. Only set for capture queues if qbuf has not yet been 541 * This function checks if queue has any buffers allocated.
|
H A D | videobuf-core.h | 2 * generic helper functions for handling video4linux capture buffers 31 * A small set of helper functions to manage video4linux buffers. 34 * functions, additionally some commonly used fields for v4l buffers 40 * The mmaper function allows to map any subset of contingous buffers. 41 * This includes one mmap() call for all buffers (which the original 97 /* for mmap'ed buffers */
|
/linux-4.1.27/drivers/media/pci/zoran/ |
H A D | zoran_driver.c | 195 fh->buffers.buffer_size = v4l_bufsize; map_mode_raw() 196 fh->buffers.num_buffers = v4l_nbufs; map_mode_raw() 201 fh->buffers.buffer_size = jpg_bufsize; map_mode_jpg() 202 fh->buffers.num_buffers = jpg_nbufs; map_mode_jpg() 210 * Allocate the V4L grab buffers 221 for (i = 0; i < fh->buffers.num_buffers; i++) { v4l_fbuffer_alloc() 222 if (fh->buffers.buffer[i].v4l.fbuffer) v4l_fbuffer_alloc() 229 mem = kmalloc(fh->buffers.buffer_size, v4l_fbuffer_alloc() 239 fh->buffers.buffer[i].v4l.fbuffer = mem; v4l_fbuffer_alloc() 240 fh->buffers.buffer[i].v4l.fbuffer_phys = virt_to_phys(mem); v4l_fbuffer_alloc() 241 fh->buffers.buffer[i].v4l.fbuffer_bus = virt_to_bus(mem); v4l_fbuffer_alloc() 242 for (off = 0; off < fh->buffers.buffer_size; v4l_fbuffer_alloc() 252 fh->buffers.allocated = 1; v4l_fbuffer_alloc() 257 /* free the V4L grab buffers */ v4l_fbuffer_free() 266 for (i = 0; i < fh->buffers.num_buffers; i++) { v4l_fbuffer_free() 267 if (!fh->buffers.buffer[i].v4l.fbuffer) v4l_fbuffer_free() 270 mem = fh->buffers.buffer[i].v4l.fbuffer; v4l_fbuffer_free() 271 for (off = 0; off < fh->buffers.buffer_size; v4l_fbuffer_free() 274 kfree(fh->buffers.buffer[i].v4l.fbuffer); v4l_fbuffer_free() 275 fh->buffers.buffer[i].v4l.fbuffer = NULL; v4l_fbuffer_free() 278 fh->buffers.allocated = 0; v4l_fbuffer_free() 282 * Allocate the MJPEG grab buffers. 289 * RJ: The contents grab buffers needs never be accessed in the driver. 300 * and fragment buffers, the device expects to see little- 306 * and fragment buffers are not little-endian. 315 for (i = 0; i < fh->buffers.num_buffers; i++) { jpg_fbuffer_alloc() 316 if (fh->buffers.buffer[i].jpg.frag_tab) jpg_fbuffer_alloc() 333 fh->buffers.buffer[i].jpg.frag_tab = (__le32 *)mem; jpg_fbuffer_alloc() 334 fh->buffers.buffer[i].jpg.frag_tab_bus = virt_to_bus(mem); jpg_fbuffer_alloc() 336 if (fh->buffers.need_contiguous) { jpg_fbuffer_alloc() 337 mem = kmalloc(fh->buffers.buffer_size, GFP_KERNEL); jpg_fbuffer_alloc() 346 fh->buffers.buffer[i].jpg.frag_tab[0] = jpg_fbuffer_alloc() 348 fh->buffers.buffer[i].jpg.frag_tab[1] = jpg_fbuffer_alloc() 349 cpu_to_le32((fh->buffers.buffer_size >> 1) | 1); jpg_fbuffer_alloc() 350 for (off = 0; off < fh->buffers.buffer_size; off += PAGE_SIZE) jpg_fbuffer_alloc() 354 for (j = 0; j < fh->buffers.buffer_size / PAGE_SIZE; j++) { jpg_fbuffer_alloc() 365 fh->buffers.buffer[i].jpg.frag_tab[2 * j] = jpg_fbuffer_alloc() 367 fh->buffers.buffer[i].jpg.frag_tab[2 * j + 1] = jpg_fbuffer_alloc() 372 fh->buffers.buffer[i].jpg.frag_tab[2 * j - 1] |= cpu_to_le32(1); jpg_fbuffer_alloc() 379 (fh->buffers.num_buffers * fh->buffers.buffer_size) >> 10); jpg_fbuffer_alloc() 381 fh->buffers.allocated = 1; jpg_fbuffer_alloc() 386 /* free the MJPEG grab buffers */ jpg_fbuffer_free() 397 for (i = 0, buffer = &fh->buffers.buffer[0]; jpg_fbuffer_free() 398 i < fh->buffers.num_buffers; i++, buffer++) { jpg_fbuffer_free() 402 if (fh->buffers.need_contiguous) { jpg_fbuffer_free() 407 for (off = 0; off < fh->buffers.buffer_size; off += PAGE_SIZE) jpg_fbuffer_free() 414 for (j = 0; j < fh->buffers.buffer_size / PAGE_SIZE; j++) { jpg_fbuffer_free() 430 fh->buffers.allocated = 0; jpg_fbuffer_free() 460 if (height * width * bpp > fh->buffers.buffer_size) { zoran_v4l_set_format() 464 ZR_DEVNAME(zr), __func__, fh->buffers.buffer_size >> 10); zoran_v4l_set_format() 492 if (!fh->buffers.allocated) { zoran_v4l_queue_frame() 495 "%s: %s - buffers not yet allocated\n", zoran_v4l_queue_frame() 501 if (num >= fh->buffers.num_buffers || num < 0) { zoran_v4l_queue_frame() 511 if (fh->buffers.active == ZORAN_FREE) { zoran_v4l_queue_frame() 513 zr->v4l_buffers = fh->buffers; zoran_v4l_queue_frame() 514 fh->buffers.active = ZORAN_ACTIVE; zoran_v4l_queue_frame() 530 fh->buffers.active = ZORAN_FREE; zoran_v4l_queue_frame() 548 fh->buffers.buffer[num] = zr->v4l_buffers.buffer[num]; zoran_v4l_queue_frame() 556 zr->v4l_buffers.active = fh->buffers.active; zoran_v4l_queue_frame() 570 if (fh->buffers.active == ZORAN_FREE) { v4l_sync() 579 if (frame >= fh->buffers.num_buffers || frame < 0) { v4l_sync() 609 fh->buffers.buffer[frame] = zr->v4l_buffers.buffer[frame]; v4l_sync() 617 fh->buffers.active = zr->v4l_buffers.active = ZORAN_FREE; v4l_sync() 638 /* Check if buffers are allocated */ zoran_jpg_queue_frame() 639 if (!fh->buffers.allocated) { zoran_jpg_queue_frame() 642 "%s: %s - buffers not yet allocated\n", zoran_jpg_queue_frame() 648 if (num >= fh->buffers.num_buffers || num < 0) { zoran_jpg_queue_frame() 668 if (fh->buffers.active == ZORAN_FREE) { zoran_jpg_queue_frame() 670 zr->jpg_buffers = fh->buffers; zoran_jpg_queue_frame() 671 fh->buffers.active = ZORAN_ACTIVE; zoran_jpg_queue_frame() 700 fh->buffers.buffer[num] = zr->jpg_buffers.buffer[num]; zoran_jpg_queue_frame() 707 fh->buffers.active = ZORAN_FREE; zoran_jpg_queue_frame() 718 zr->jpg_buffers.active = fh->buffers.active; zoran_jpg_queue_frame() 731 if (fh->buffers.active == ZORAN_FREE) { jpg_qbuf() 738 fh->buffers.active = zr->jpg_buffers.active = ZORAN_FREE; jpg_qbuf() 771 if (fh->buffers.active == ZORAN_FREE) { jpg_sync() 823 fh->buffers.buffer[frame] = zr->jpg_buffers.buffer[frame]; jpg_sync() 849 /* buffers */ zoran_open_init_session() 850 memset(&fh->buffers, 0, sizeof(fh->buffers)); zoran_open_init_session() 852 fh->buffers.buffer[i].state = BUZ_STATE_USER; /* nothing going on */ zoran_open_init_session() 853 fh->buffers.buffer[i].bs.frame = i; zoran_open_init_session() 855 fh->buffers.allocated = 0; zoran_open_init_session() 856 fh->buffers.active = ZORAN_FREE; zoran_open_init_session() 874 if (fh->buffers.active != ZORAN_FREE) { zoran_close_end_session() 880 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_close_end_session() 884 /* v4l buffers */ zoran_close_end_session() 885 if (fh->buffers.allocated) zoran_close_end_session() 889 if (fh->buffers.active != ZORAN_FREE) { zoran_close_end_session() 892 zr->jpg_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_close_end_session() 895 /* jpg buffers */ zoran_close_end_session() 896 if (fh->buffers.allocated) zoran_close_end_session() 1342 if (num < 0 || num >= fh->buffers.num_buffers || zoran_v4l2_buffer_status() 1343 !fh->buffers.allocated) { zoran_v4l2_buffer_status() 1346 "%s: %s - wrong number or buffers not allocated\n", zoran_v4l2_buffer_status() 1356 "FAL"[fh->buffers.active], num, zoran_v4l2_buffer_status() 1358 fh->buffers.buffer[num].map ? 'Y' : 'N'); zoran_v4l2_buffer_status() 1362 buf->length = fh->buffers.buffer_size; zoran_v4l2_buffer_status() 1365 buf->bytesused = fh->buffers.buffer[num].bs.length; zoran_v4l2_buffer_status() 1366 if (fh->buffers.buffer[num].state == BUZ_STATE_DONE || zoran_v4l2_buffer_status() 1367 fh->buffers.buffer[num].state == BUZ_STATE_USER) { zoran_v4l2_buffer_status() 1368 buf->sequence = fh->buffers.buffer[num].bs.seq; zoran_v4l2_buffer_status() 1370 buf->timestamp = fh->buffers.buffer[num].bs.timestamp; zoran_v4l2_buffer_status() 1386 if (num < 0 || num >= fh->buffers.num_buffers || zoran_v4l2_buffer_status() 1387 !fh->buffers.allocated) { zoran_v4l2_buffer_status() 1390 "%s: %s - wrong number or buffers not allocated\n", zoran_v4l2_buffer_status() 1398 buf->length = fh->buffers.buffer_size; zoran_v4l2_buffer_status() 1401 if (fh->buffers.buffer[num].state == BUZ_STATE_DONE || zoran_v4l2_buffer_status() 1402 fh->buffers.buffer[num].state == BUZ_STATE_USER) { zoran_v4l2_buffer_status() 1403 buf->sequence = fh->buffers.buffer[num].bs.seq; zoran_v4l2_buffer_status() 1404 buf->timestamp = fh->buffers.buffer[num].bs.timestamp; zoran_v4l2_buffer_status() 1405 buf->bytesused = fh->buffers.buffer[num].bs.length; zoran_v4l2_buffer_status() 1813 if (fh->buffers.allocated) { zoran_s_fmt_vid_out() 1860 fh->buffers.buffer_size = zoran_v4l2_calc_bufsize(&fh->jpg_settings); zoran_s_fmt_vid_out() 1873 fmt->fmt.pix.sizeimage = fh->buffers.buffer_size; zoran_s_fmt_vid_out() 1903 if ((fh->map_mode != ZORAN_MAP_MODE_RAW && fh->buffers.allocated) || zoran_s_fmt_vid_cap() 1904 fh->buffers.active != ZORAN_FREE) { zoran_s_fmt_vid_cap() 2017 if (fh->buffers.allocated) { zoran_reqbufs() 2020 "%s: VIDIOC_REQBUFS - buffers already allocated\n", zoran_reqbufs() 2034 /* The next mmap will map the V4L buffers */ zoran_reqbufs() 2036 fh->buffers.num_buffers = req->count; zoran_reqbufs() 2050 /* The next mmap will map the MJPEG buffers */ zoran_reqbufs() 2052 fh->buffers.num_buffers = req->count; zoran_reqbufs() 2053 fh->buffers.buffer_size = zoran_v4l2_calc_bufsize(&fh->jpg_settings); zoran_reqbufs() 2107 if (!zr->v4l_memgrab_active && fh->buffers.active == ZORAN_LOCKED) zoran_qbuf() 2133 fh->buffers.active == ZORAN_LOCKED) zoran_qbuf() 2239 fh->buffers.active != ZORAN_ACTIVE) { zoran_streamon() 2244 zr->v4l_buffers.active = fh->buffers.active = ZORAN_LOCKED; zoran_streamon() 2258 fh->buffers.active != ZORAN_ACTIVE) { zoran_streamon() 2263 zr->jpg_buffers.active = fh->buffers.active = ZORAN_LOCKED; zoran_streamon() 2296 if (fh->buffers.active == ZORAN_FREE && zoran_streamoff() 2311 for (i = 0; i < fh->buffers.num_buffers; i++) zoran_streamoff() 2313 fh->buffers = zr->v4l_buffers; zoran_streamoff() 2315 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_streamoff() 2327 if (fh->buffers.active == ZORAN_FREE && zoran_streamoff() 2590 if (fh->buffers.allocated) { zoran_s_crop() 2668 if (fh->buffers.active != ZORAN_FREE) { zoran_s_jpegcomp() 2679 if (!fh->buffers.allocated) zoran_s_jpegcomp() 2680 fh->buffers.buffer_size = zoran_s_jpegcomp() 2698 /* we should check whether buffers are ready to be synced on zoran_poll() 2703 * if no buffers queued or so, return POLLNVAL zoran_poll() 2718 "FAL"[fh->buffers.active], zr->v4l_sync_tail, zoran_poll() 2722 if (fh->buffers.active != ZORAN_FREE && zoran_poll() 2740 "FAL"[fh->buffers.active], zr->jpg_que_tail, zoran_poll() 2743 if (fh->buffers.active != ZORAN_FREE && zoran_poll() 2769 * This maps the buffers to user space. 2772 * the V4L or the MJPEG buffers are mapped 2801 for (i = 0; i < fh->buffers.num_buffers; i++) { zoran_vm_close() 2802 if (fh->buffers.buffer[i].map == map) zoran_vm_close() 2803 fh->buffers.buffer[i].map = NULL; zoran_vm_close() 2807 /* Any buffers still mapped? */ zoran_vm_close() 2808 for (i = 0; i < fh->buffers.num_buffers; i++) { zoran_vm_close() 2809 if (fh->buffers.buffer[i].map) { zoran_vm_close() 2815 dprintk(3, KERN_INFO "%s: %s - free %s buffers\n", ZR_DEVNAME(zr), zoran_vm_close() 2820 if (fh->buffers.active != ZORAN_FREE) { zoran_vm_close() 2826 zr->v4l_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_vm_close() 2831 if (fh->buffers.active != ZORAN_FREE) { zoran_vm_close() 2834 zr->jpg_buffers.active = fh->buffers.active = ZORAN_FREE; zoran_vm_close() 2877 if (!fh->buffers.allocated) { zoran_mmap() 2880 "%s: %s(%s) - buffers not yet allocated\n", zoran_mmap() 2886 first = offset / fh->buffers.buffer_size; zoran_mmap() 2887 last = first - 1 + size / fh->buffers.buffer_size; zoran_mmap() 2888 if (offset % fh->buffers.buffer_size != 0 || zoran_mmap() 2889 size % fh->buffers.buffer_size != 0 || first < 0 || zoran_mmap() 2890 last < 0 || first >= fh->buffers.num_buffers || zoran_mmap() 2891 last >= fh->buffers.buffer_size) { zoran_mmap() 2896 fh->buffers.buffer_size, zoran_mmap() 2897 fh->buffers.num_buffers); zoran_mmap() 2902 /* Check if any buffers are already mapped */ zoran_mmap() 2904 if (fh->buffers.buffer[i].map) { zoran_mmap() 2914 /* map these buffers */ zoran_mmap() 2930 if (todo > fh->buffers.buffer_size) zoran_mmap() 2931 todo = fh->buffers.buffer_size; zoran_mmap() 2932 page = fh->buffers.buffer[i].v4l.fbuffer_phys; zoran_mmap() 2944 fh->buffers.buffer[i].map = map; zoran_mmap() 2951 j < fh->buffers.buffer_size / PAGE_SIZE; zoran_mmap() 2954 (le32_to_cpu(fh->buffers.buffer[i].jpg. zoran_mmap() 2960 le32_to_cpu(fh->buffers. zoran_mmap() 2978 if (le32_to_cpu(fh->buffers.buffer[i].jpg. zoran_mmap() 2982 fh->buffers.buffer[i].map = map; zoran_mmap()
|
H A D | zoran.h | 66 # error "Too many video frame buffers to handle" 209 struct zoran_buffer buffer[MAX_FRAME]; /* buffers */ 210 u8 allocated; /* Flag if buffers are allocated */ 211 u8 need_contiguous; /* Flag if contiguous buffers are needed */ 212 /* only applies to jpg buffers, raw buffers are always contiguous */ 227 struct zoran_buffer_col buffers; /* buffers' info */ member in struct:zoran_fh 328 struct zoran_buffer_col v4l_buffers; /* V4L buffers' info */ 355 struct zoran_buffer_col jpg_buffers; /* MJPEG buffers' info */
|
/linux-4.1.27/drivers/staging/speakup/ |
H A D | Makefile | 19 buffers.o \
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | cache.h | 13 * buffers have cache line alignment.
|
H A D | fixmap.h | 32 * these 'compile-time allocated' memory buffers are 37 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/drivers/net/wan/ |
H A D | x25_asy.h | 25 /* These are pointers to the malloc()ed frame buffers. */ 31 int buffsize; /* Max buffers sizes */
|
/linux-4.1.27/drivers/staging/octeon/ |
H A D | ethernet-mem.c | 42 * @elements: Number of buffers to allocate 44 * Returns the actual number of buffers allocated. 67 * @elements: Number of buffers to allocate 95 * @elements: Number of buffers to allocate 97 * Returns the actual number of buffers allocated. 134 * @elements: Number of buffers that should be in the pool 152 pr_warn("Freeing of pool %u had too many buffers (%d)\n", cvm_oct_free_hw_memory() 155 pr_warn("Warning: Freeing of pool %u is missing %d buffers\n", cvm_oct_free_hw_memory()
|
H A D | ethernet-defines.h | 34 * allows 32bit userspace application to access the buffers, but also 37 * Tells the driver to populate the packet buffers with kernel skbuffs. 39 * means that 32bit userspace can't access the packet buffers.
|
H A D | octeon-ethernet.h | 45 /* Hardware fetch and add to count outstanding tx buffers */ 54 /* List of outstanding tx buffers per queue */
|
/linux-4.1.27/include/drm/ttm/ |
H A D | ttm_execbuf_util.h | 74 * If the function returns 0, all buffers are marked as "unfenced", 78 * reserve the same buffers in reverse order, all threads except one will 81 * unreserve their buffers. 85 * buffers on the list will be reserved upon return. 87 * If dups is non NULL all buffers already reserved by the current thread 89 * on the first already reserved buffer and all buffers from the list are 107 * @fence: The new exclusive fence for the buffers. 111 * It also unreserves all buffers, putting them on lru lists.
|
H A D | ttm_bo_api.h | 132 * @ttm_bo_type_device: These are 'normal' buffers that can 136 * @ttm_bo_type_kernel: These buffers are like ttm_bo_type_device buffers, 167 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 457 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 504 * @persistent_swap_storage: Usually the swap storage is deleted for buffers 553 * the LRU list to evict any buffers left alive. 556 * point there shouldn't be any buffers left created by user-space, since 560 * make sure that there are no NO_EVICT buffers present in this memory type 564 * there are no appications currently validating buffers before this 570 * -EBUSY: There are still buffers left in this memory type. 581 * Evicts all buffers on the lru list of the memory type. 585 * currently validating buffers, and can do that by taking the
|
/linux-4.1.27/drivers/staging/iio/ |
H A D | ring_hw.h | 2 * ring_hw.h - common functionality for iio hardware ring buffers
|
/linux-4.1.27/arch/s390/include/uapi/asm/ |
H A D | virtio-ccw.h | 15 /* Alignment of vring buffers. */
|
/linux-4.1.27/include/linux/ |
H A D | journal-head.h | 63 * transaction (if there is one). Only applies to buffers on a 80 * Doubly-linked list of buffers on a transaction's data, metadata or 87 * is checkpointed. Only dirty buffers can be checkpointed. 93 * Doubly-linked list of buffers still remaining to be flushed
|
H A D | pipe_fs_i.h | 32 * @nrbufs: the number of non-empty pipe buffers in this pipe 33 * @buffers: total number of buffers (should be a power of 2) 44 * @bufs: the circular array of pipe buffers 49 unsigned int nrbufs, curbuf, buffers; member in struct:pipe_inode_info
|
H A D | jbd.h | 89 * The buffer credits field is used to account for journaled buffers 92 * number of outstanding buffers possible at any time. When the 94 * the transaction, so that at all times we know how many buffers the 300 * @h_buffer_credits: Number of remaining buffers we are allowed to dirty. 313 /* Number of remaining buffers we are allowed to dirty: */ 340 * new buffers to modify (state not used for now) 345 * The transaction keeps track of all of the buffers modified by a 346 * running transaction, and all of the buffers committed but not yet 400 /* Number of buffers on the t_buffers list [j_list_lock] */ 404 * Doubly-linked circular list of all buffers reserved but not yet 410 * Doubly-linked circular list of all buffers under writeout during 416 * Doubly-linked circular list of all metadata buffers owned by this 422 * Doubly-linked circular list of all data buffers still to be 428 * Doubly-linked circular list of all forget buffers (superseded 429 * buffers which we can un-checkpoint once this transaction commits) 435 * Doubly-linked circular list of all buffers still to be flushed before 441 * Doubly-linked circular list of all buffers submitted for IO while 447 * Doubly-linked circular list of temporary buffers currently undergoing 453 * Doubly-linked circular list of metadata buffers being shadowed by log 454 * IO. The IO buffers on the iobuf list and the shadow buffers on this 460 * Doubly-linked circular list of control buffers being written to the 477 * Number of buffers reserved for use by all handles in this transaction 553 * @j_max_transaction_buffers: Maximum number of metadata buffers to allow in a 734 * Maximum number of metadata buffers to allow in a single compound 798 /* Filing buffers */
|
H A D | circ_buf.h | 2 * See Documentation/circular-buffers.txt for more information.
|
H A D | jbd2.h | 82 * The buffer credits field is used to account for journaled buffers 85 * number of outstanding buffers possible at any time. When the 87 * the transaction, so that at all times we know how many buffers the 405 * @h_buffer_credits: Number of remaining buffers we are allowed to dirty. 429 /* Number of remaining buffers we are allowed to dirty: */ 472 * new buffers to modify (state not used for now) 477 * The transaction keeps track of all of the buffers modified by a 478 * running transaction, and all of the buffers committed but not yet 534 /* Number of buffers on the t_buffers list [j_list_lock] */ 538 * Doubly-linked circular list of all buffers reserved but not yet 544 * Doubly-linked circular list of all metadata buffers owned by this 550 * Doubly-linked circular list of all forget buffers (superseded 551 * buffers which we can un-checkpoint once this transaction commits) 557 * Doubly-linked circular list of all buffers still to be flushed before 563 * Doubly-linked circular list of all buffers submitted for IO while 569 * Doubly-linked circular list of metadata buffers being shadowed by log 570 * IO. The IO buffers on the iobuf list and the shadow buffers on this 613 * Number of buffers reserved for use by all handles in this transaction 721 * @j_reserved_credits: Number of buffers reserved from the running transaction 734 * @j_max_transaction_buffers: Maximum number of metadata buffers to allow in a 877 /* Number of buffers reserved from the running transaction */ 924 * Maximum number of metadata buffers to allow in a single compound 1017 /* Filing buffers */ 1068 * Fired during journal abort for dirty buffers that will not be
|
H A D | arcdevice.h | 72 #define D_TIMING 4096 /* show time needed to copy buffers to card */ 168 /* byte deposited into first address of buffers on reset */ 259 * Buffer management: an ARCnet card has 4 x 512-byte buffers, each of 262 * buffers, and take them as they're needed. This way, we simplify 267 * The rules: only the interrupt handler is allowed to _add_ buffers to 269 * handler and the transmit function will want to _remove_ buffers, so 274 * only four possible buffers, the queue should never be full.
|
H A D | relay.h | 36 size_t subbufs_produced; /* count of sub-buffers produced */ 37 size_t subbufs_consumed; /* count of sub-buffers consumed */ 60 size_t n_subbufs; /* number of sub-buffers per buffer */ 66 struct rchan_buf *buf[NR_CPUS]; /* per-cpu channel buffers */ 139 * than the default set of per-cpu buffers.
|
H A D | rpmsg.h | 189 * In case there are no TX buffers available, the function will block until 214 * In case there are no TX buffers available, the function will block until 242 * In case there are no TX buffers available, the function will block until 266 * In case there are no TX buffers available, the function will immediately 291 * In case there are no TX buffers available, the function will immediately 318 * In case there are no TX buffers available, the function will immediately
|
H A D | mroute.h | 77 struct sk_buff_head unresolved; /* Unresolved buffers */
|
/linux-4.1.27/drivers/md/ |
H A D | dm-bufio.h | 40 * buffers. 90 * Initiate writing of dirty buffers, without waiting for completion. 95 * Write all dirty buffers. Guarantees that all dirty buffers created prior 119 * Set the minimum number of buffers before cleanup happens.
|
H A D | dm-bufio.c | 24 * Limit the number of buffers to DM_BUFIO_MEMORY_PERCENT of main memory 26 * Always allocate at least DM_BUFIO_MIN_BUFFERS buffers. 28 * dirty buffers. 42 * Free buffers when they are older than this (seconds) 72 * Linking of buffers: 73 * All buffers are linked to cache_hash with their hash_list field. 75 * Clean buffers that are not being written (B_WRITING not set) 78 * Dirty and clean buffers that are being written are linked to 82 * context), so some clean-not-writing buffers can be held on 257 * A red/black tree acts as an index for all the buffers. 351 * Small buffers are allocated with kmem_cache, to use space optimally. 353 * For large buffers, we choose between get_free_pages and vmalloc. 362 * buffers and more I/O will be performed. Don't use __get_free_pages if it 655 * Writing dirty buffers 820 * For debugging, if we set the cache size to 1, no new buffers will __alloc_buffer_wait_no_callback() 864 * Free a buffer and wake other threads waiting for free buffers. 909 unsigned long buffers; __get_memory_limit() local 917 buffers = dm_bufio_cache_size_per_client >> __get_memory_limit() 920 if (buffers < c->minimum_buffers) __get_memory_limit() 921 buffers = c->minimum_buffers; __get_memory_limit() 923 *limit_buffers = buffers; __get_memory_limit() 924 *threshold_buffers = buffers * DM_BUFIO_WRITEBACK_PERCENT / 100; __get_memory_limit() 929 * If we are over threshold_buffers, start freeing buffers. 1215 * For performance, it is essential that the buffers are written asynchronously 1273 * dirtying buffers, so we count the number of buffers walked dm_bufio_write_dirty_buffers() 1274 * and if it exceeds the total number of buffers, it means that dm_bufio_write_dirty_buffers() 1463 * An optimization so that the buffers are not written one-by-one. drop_buffers() 1691 * It is required that there are no references on any buffers.
|
/linux-4.1.27/drivers/media/pci/cx18/ |
H A D | cx18-queue.c | 257 /* Reset MDL id's and move all buffers back to the stream's buf_pool */ cx18_unload_queues() 287 * Attach buffers to MDLs, give the MDLs ids, and add MDLs to q_free cx18_load_queues() 289 * Excess buffers are left in buf_pool and/or on an MDL in q_idle cx18_load_queues() 316 * case of a non-integral number of buffers to meet cx18_load_queues() 328 /* Not enough buffers for this MDL; we won't use it */ cx18_load_queues() 352 if (s->buffers == 0) cx18_stream_alloc() 355 CX18_DEBUG_INFO("Allocate %s stream: %d x %d buffers " cx18_stream_alloc() 357 s->name, s->buffers, s->buf_size, cx18_stream_alloc() 358 s->buffers * s->buf_size / 1024, cx18_stream_alloc() 359 (s->buffers * s->buf_size * 100 / 1024) % 100); cx18_stream_alloc() 361 if (((char __iomem *)&cx->scb->cpu_mdl[cx->free_mdl_idx + s->buffers] - cx18_stream_alloc() 366 CX18_ERR("Too many buffers, cannot fit in SCB area\n"); cx18_stream_alloc() 367 CX18_ERR("Max buffers = %zu\n", cx18_stream_alloc() 374 /* allocate stream buffers and MDLs */ cx18_stream_alloc() 375 for (i = 0; i < s->buffers; i++) { cx18_stream_alloc() 409 if (i == s->buffers) { cx18_stream_alloc() 410 cx->free_mdl_idx += s->buffers; cx18_stream_alloc() 413 CX18_ERR("Couldn't allocate buffers for %s stream\n", s->name); cx18_stream_alloc() 424 CX18_DEBUG_INFO("Deallocating buffers for %s stream\n", s->name); cx18_stream_free() 426 /* move all buffers to buf_pool and all MDLs to q_idle */ cx18_stream_free()
|
H A D | cx18-controls.c | 52 /* Allocate sliced VBI buffers if needed. */ cx18_s_stream_vbi_fmt() 66 CX18_WARN("Unable to allocate buffers for " cx18_s_stream_vbi_fmt()
|
/linux-4.1.27/drivers/usb/class/ |
H A D | cdc-acm.h | 55 * The only reason to have several buffers is to accommodate assumptions 59 * when processing onlcr, so we only need 2 buffers. These values must be 88 u8 *ctrl_buffer; /* buffers of urbs */ 89 dma_addr_t ctrl_dma; /* dma handles of buffers */ 100 int write_used; /* number of non-empty write buffers */
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
H A D | verbs.c | 344 * recv buffers to the pool. NOTE: this affects synchronization of 1082 * Send/recv buffers in req/rep need to be registered rpcrdma_buffer_create() 1324 * Get a set of request/reply buffers. 1330 * removing buffers, and decremented *before* returning them. 1333 rpcrdma_buffer_get(struct rpcrdma_buffer *buffers) rpcrdma_buffer_get() argument 1335 struct rpcrdma_ia *ia = rdmab_to_ia(buffers); rpcrdma_buffer_get() 1340 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_buffer_get() 1341 if (buffers->rb_send_index == buffers->rb_max_requests) { rpcrdma_buffer_get() 1342 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_get() 1343 dprintk("RPC: %s: out of request buffers\n", __func__); rpcrdma_buffer_get() 1347 req = buffers->rb_send_bufs[buffers->rb_send_index]; rpcrdma_buffer_get() 1348 if (buffers->rb_send_index < buffers->rb_recv_index) { rpcrdma_buffer_get() 1351 buffers->rb_recv_index - buffers->rb_send_index); rpcrdma_buffer_get() 1354 req->rl_reply = buffers->rb_recv_bufs[buffers->rb_recv_index]; rpcrdma_buffer_get() 1355 buffers->rb_recv_bufs[buffers->rb_recv_index++] = NULL; rpcrdma_buffer_get() 1357 buffers->rb_send_bufs[buffers->rb_send_index++] = NULL; rpcrdma_buffer_get() 1362 req = rpcrdma_buffer_get_frmrs(req, buffers, &stale); rpcrdma_buffer_get() 1365 req = rpcrdma_buffer_get_fmrs(req, buffers); rpcrdma_buffer_get() 1370 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_get() 1372 rpcrdma_retry_flushed_linv(&stale, buffers); rpcrdma_buffer_get() 1377 * Put request/reply buffers back into pool. 1383 struct rpcrdma_buffer *buffers = req->rl_buffer; rpcrdma_buffer_put() local 1384 struct rpcrdma_ia *ia = rdmab_to_ia(buffers); rpcrdma_buffer_put() 1387 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_buffer_put() 1388 rpcrdma_buffer_put_sendbuf(req, buffers); rpcrdma_buffer_put() 1392 rpcrdma_buffer_put_mrs(req, buffers); rpcrdma_buffer_put() 1397 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_buffer_put() 1401 * Recover reply buffers from pool. 1408 struct rpcrdma_buffer *buffers = req->rl_buffer; rpcrdma_recv_buffer_get() local 1411 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_recv_buffer_get() 1412 if (buffers->rb_recv_index < buffers->rb_max_requests) { rpcrdma_recv_buffer_get() 1413 req->rl_reply = buffers->rb_recv_bufs[buffers->rb_recv_index]; rpcrdma_recv_buffer_get() 1414 buffers->rb_recv_bufs[buffers->rb_recv_index++] = NULL; rpcrdma_recv_buffer_get() 1416 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_recv_buffer_get() 1420 * Put reply buffers back into pool when not attached to 1426 struct rpcrdma_buffer *buffers = rep->rr_buffer; rpcrdma_recv_buffer_put() local 1430 spin_lock_irqsave(&buffers->rb_lock, flags); rpcrdma_recv_buffer_put() 1431 buffers->rb_recv_bufs[--buffers->rb_recv_index] = rep; rpcrdma_recv_buffer_put() 1432 spin_unlock_irqrestore(&buffers->rb_lock, flags); rpcrdma_recv_buffer_put() 1517 * rpcrdma_alloc_regbuf - kmalloc and register memory for SEND/RECV buffers 1652 /* How many chunk list items fit within our inline buffers?
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_wc_x86_64.c | 62 * Set the PIO buffers to be WCCOMB, so we get HT bursts to the qib_enable_wc() 66 * the PIO buffers themselves. qib_enable_wc() 67 * For chips with two sets of buffers, the calculations are qib_enable_wc() 70 * The buffers are still packed, so a single range covers both. qib_enable_wc() 142 * or AMD x86_64 processor. AMD x86_64 processors flush WC buffers out in 144 * correct ordering. Intel processors, however, will flush write buffers
|
H A D | qib_tx.c | 53 * qib_disarm_piobufs - cancel a range of PIO buffers 56 * @cnt: the number of PIO buffers to cancel 58 * Cancel a range of PIO buffers. Used at user process close, 165 * Disarm a set of send buffers. If the buffer might be actively being 222 * called whenever our local copy indicates we have run out of send buffers 245 * happens when all buffers are in use, so only cpu overhead, not update_send_bufs() 268 * Debugging code and stats updates if no pio buffers available. 301 * Minor optimization. If we had no buffers on last call, qib_getsendbuf_range() 303 * if no buffers were updated, to be paranoid. qib_getsendbuf_range() 339 * buffers available, try an update and then rescan. qib_getsendbuf_range() 379 * qib_chg_pioavailkernel - change which send buffers are available for kernel 382 * @len: the number of send buffers 383 * @avail: true if the buffers are available for kernel use, false otherwise 405 * the user buffers before we hand them back to the qib_chg_pioavailkernel() 452 * all pio buffers, and issuing an abort, which cleans up anything in the 467 * Tell PSM to disarm buffers again before trying to reuse them. qib_cancel_sends()
|
/linux-4.1.27/drivers/usb/gadget/function/ |
H A D | uvc_queue.c | 29 * Video buffers queue management. 34 * Video buffers are managed by videobuf2. The driver uses a mutex to protect 36 * spinlock to protect the IRQ queue that holds the buffers to be processed by 141 * Free the video buffers. 149 * Allocate the video buffers. 223 * Cancel the video buffers queue. 225 * Cancelling the queue marks all buffers on the irq queue as erroneous, 259 * Enable or disable the video buffers queue. 262 * disabled after stopping it. This ensures that the video buffers queue 263 * state can be properly initialized before buffers are accessed from the 269 * Disabling the video queue cancels the queue and removes all buffers from 297 * applications will be able to queue buffers for the next uvcg_queue_enable()
|
H A D | uvc_queue.h | 13 /* Maximum number of video buffers. */
|
/linux-4.1.27/drivers/scsi/isci/ |
H A D | unsolicited_frame_control.c | 69 * The Unsolicited Frame buffers are set at the start of the UF sci_unsolicited_frame_control_construct() 71 * placed after the buffers. sci_unsolicited_frame_control_construct() 106 * Program the actual used UF buffers into the UF address table and sci_unsolicited_frame_control_construct() 110 uf = &uf_control->buffers.array[i]; sci_unsolicited_frame_control_construct() 136 *frame_header = &uf_control->buffers.array[frame_index].header->data; sci_unsolicited_frame_control_get_header() 149 *frame_buffer = uf_control->buffers.array[frame_index].buffer; sci_unsolicited_frame_control_get_buffer() 184 uf_control->buffers.array[frame_index].state = UNSOLICITED_FRAME_RELEASED; sci_unsolicited_frame_control_release_frame() 198 while (uf_control->buffers.array[frame_get].state == UNSOLICITED_FRAME_RELEASED) { sci_unsolicited_frame_control_release_frame() 199 uf_control->buffers.array[frame_get].state = UNSOLICITED_FRAME_EMPTY; sci_unsolicited_frame_control_release_frame()
|
/linux-4.1.27/drivers/atm/ |
H A D | zatm.h | 21 #define LOW_MARK 12 /* start adding new buffers if less than 12 */ 22 #define HIGH_MARK 30 /* stop adding buffers after reaching 30 */ 49 struct sk_buff_head tx_queue; /* list of buffers in transit */ 54 struct sk_buff_head backlog; /* list of buffers waiting for ring */
|
H A D | nicstar.h | 52 /* Number of buffers initially allocated */ 55 #define NUM_HB 8 /* Pre-allocated huge buffers */ 56 #define NUM_IOVB 48 /* Iovec buffers */ 58 /* Lower level for count of buffers */ 64 /* Upper level for count of buffers */ 97 * BUF_XX distinguish the Rx buffers depending on their (small/large) size. 688 volatile int count; /* number of buffers in the queue */ 721 skb_pool sbpool; /* Small buffers */ 722 skb_pool lbpool; /* Large buffers */ 723 skb_pool hbpool; /* Pre-allocated huge buffers */ 724 skb_pool iovpool; /* iovector buffers */
|
H A D | fore200e.h | 8 #define SMALL_BUFFER_SIZE 384 /* size of small buffers (multiple of 48 (PCA) and 64 (SBA) bytes) */ 9 #define LARGE_BUFFER_SIZE 4032 /* size of large buffers (multiple of 48 (PCA) and 64 (SBA) bytes) */ 12 #define RBD_BLK_SIZE 32 /* nbr of supplied rx buffers per rbd */ 18 #define BUFFER_S1_SIZE SMALL_BUFFER_SIZE /* size of small buffers, scheme 1 */ 19 #define BUFFER_L1_SIZE LARGE_BUFFER_SIZE /* size of large buffers, scheme 1 */ 21 #define BUFFER_S2_SIZE SMALL_BUFFER_SIZE /* size of small buffers, scheme 2 */ 22 #define BUFFER_L2_SIZE LARGE_BUFFER_SIZE /* size of large buffers, scheme 2 */ 47 then it uses a number of *large* buffers for the trailing data. 636 struct buffer* buffer; /* array of rx buffers */ 637 struct buffer* freebuf; /* list of free rx buffers */ 638 volatile int freebuf_count; /* count of free rx buffers */ 763 FORE200E_STATE_ALLOC_BUF, /* receive buffers allocated */ 803 int buffer_alignment; /* rx buffers DMA alignment requirement */
|
/linux-4.1.27/include/uapi/linux/hsi/ |
H A D | cs-protocol.h | 65 /* maximum number of TX/RX buffers */ 69 /* Parameters for setting up the data buffers */ 85 __u32 rx_bufs; /* # of RX buffers */ 86 __u32 tx_bufs; /* # of TX buffers */
|
/linux-4.1.27/include/uapi/linux/ |
H A D | sysinfo.h | 13 __kernel_ulong_t bufferram; /* Memory used by buffers */
|
H A D | fanotify.h | 105 /* Helper functions to deal with fanotify_event_metadata buffers */
|
H A D | mic_ioctl.h | 29 * @iov: An array of IOVEC structures containing user space buffers.
|
H A D | pg.h | 32 read and write buffers contain a single character "magic" flag.
|
/linux-4.1.27/drivers/media/platform/s5p-mfc/ |
H A D | s5p_mfc_common.h | 501 * @vq_src: vb2 queue for source buffers 502 * @vq_dst: vb2 queue for destination buffers 503 * @src_queue: driver internal queue for source buffers 504 * @dst_queue: driver internal queue for destination buffers 505 * @src_queue_cnt: number of buffers queued on the source internal queue 506 * @dst_queue_cnt: number of buffers queued on the dest internal queue 519 * @dpb_flush_flag: flag used to indicate that a DPB buffers are being 523 * @bank1: handle to memory allocated for temporary buffers from 525 * @bank2: handle to memory allocated for temporary buffers from 527 * @capture_state: state of the capture buffers queue 528 * @output_state: state of the output buffers queue 529 * @src_bufs: information on allocated source buffers 530 * @dst_bufs: information on allocated destination buffers 532 * @dec_dst_flag: flags for buffers queued in the hardware 533 * @dec_src_buf_size: size of the buffer for source buffers in decoding 542 * @dpb_count: count of the DPB buffers required by MFC hw 543 * @total_dpb_count: count of DPB buffers with additional buffers 548 * @mv_count: number of MV buffers allocated for decoding 550 * @enc_dst_buf_size: size of the buffers for encoder output 556 * @ref_queue: list of the reference buffers for encoding 557 * @ref_queue_cnt: number of the buffers in the reference list
|
/linux-4.1.27/drivers/media/usb/tm6000/ |
H A D | tm6000-usb-isoc.h | 34 /* transfer buffers for isoc transfer */ 48 /* Stores already requested buffers */
|
/linux-4.1.27/arch/um/include/asm/ |
H A D | fixmap.h | 20 * these 'compile-time allocated' memory buffers are 25 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/arch/microblaze/include/asm/ |
H A D | fixmap.h | 41 * these 'compile-time allocated' memory buffers are 46 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | fixmap.h | 37 * these 'compile-time allocated' memory buffers are 42 * TLB entries of such buffers will not be flushed across
|
H A D | cache.h | 54 * prefetched instructions from the ifetch buffers. We also need a sync
|
/linux-4.1.27/drivers/net/ethernet/amd/ |
H A D | mvme147.c | 29 /* We have 32K of RAM for the init block and buffers. This places 30 * an upper limit on the number of buffers we can use. NetBSD uses 8 Rx 31 * buffers and 2 Tx buffers, it takes (8 + 2) * 1544 bytes. 116 printk("%s: No memory for LANCE buffers\n", dev->name); mvme147lance_probe()
|
H A D | au1000_eth.h | 43 * Data Buffer Descriptor. Data buffers must be aligned on 32 byte 127 u32 vaddr; /* virtual address of rx/tx buffers */ 128 dma_addr_t dma_addr; /* dma address of rx/tx buffers */
|
H A D | hplance.c | 30 /* We have 16392 bytes of RAM for the init block and buffers. This places 31 * an upper limit on the number of buffers we can use. NetBSD uses 8 Rx 32 * buffers and 2 Tx buffers, it takes (8 + 2) * 1544 bytes.
|
H A D | lance.c | 130 IIIa. Ring buffers 131 The LANCE uses ring buffers of Tx and Rx descriptors. Each entry describes 133 of these buffers is set by LANCE_LOG_{RX,TX}_BUFFERS, which is log_2() of 143 The LANCE has the capability to "chain" both Rx and Tx buffers, but this driver 144 statically allocates full-sized (slightly oversized -- PKT_BUF_SZ) buffers to 146 allocating full-sized buffers "just in case", at the expense of a 150 the buffers are only used when needed as low-memory bounce buffers. 154 the initialization block, Rx and Tx rings, and data buffers, must be 161 As mentioned before, low-memory "bounce-buffers" are used when needed. 183 /* Set the number of Tx and Rx buffers, using Log_2(# buffers). 184 Reasonable default values are 16 Tx buffers, and 16 Rx buffers. 245 unsigned long rx_buffs; /* Address of Rx and Tx buffers. */ 296 /* Non-zero if lance_probe1() needs to allocate low-memory bounce buffers. 976 /* The old LANCE chips doesn't automatically pad buffers to min. size. */ lance_start_xmit() 1158 buffers it's possible for a jabber packet to use two lance_rx() 1159 buffers, with only the last correctly noting the error. */ lance_rx()
|
/linux-4.1.27/drivers/media/platform/exynos4-is/ |
H A D | fimc-isp.h | 119 * @pending_buf_q: pending buffers queue head 120 * @active_buf_q: a queue head of buffers scheduled in hardware 122 * @active_buf_count: number of video buffers scheduled in hardware 124 * @reqbufs_count: number of buffers requested with REQBUFS ioctl 139 struct isp_video_buf *buffers[FIMC_ISP_MAX_BUFS]; member in struct:fimc_is_video
|
H A D | fimc-lite.h | 98 * @list: list head for the buffers queue 137 * @pending_buf_q: pending buffers queue head 138 * @active_buf_q: the queue head of buffers scheduled in hardware 139 * @vb_queue: vb2 buffers queue 141 * @active_buf_count: number of video buffers scheduled in hardware 143 * @reqbufs_count: the number of buffers requested with REQBUFS ioctl
|
H A D | fimc-core.h | 296 * @active_buf_q: the queue head of buffers scheduled in hardware 298 * @active_buf_cnt: number of video buffers scheduled in hardware 299 * @buf_index: index for managing the output DMA buffers 301 * @reqbufs_count: the number of buffers requested in REQBUFS ioctl 386 * @out_buf_count: maximum number of output DMA buffers supported 677 * fimc_active_queue_add - add buffer to the capture active buffers queue 678 * @buf: buffer to add to the active buffers list 688 * fimc_active_queue_pop - pop buffer from the capture active buffers queue 704 * fimc_pending_queue_add - add buffer to the capture pending buffers queue 705 * @buf: buffer to add to the pending buffers list 714 * fimc_pending_queue_pop - pop buffer from the capture pending buffers queue
|
/linux-4.1.27/fs/ntfs/ |
H A D | aops.c | 121 /* Async buffers must be locked. */ ntfs_end_buffer_async_read() 129 * If none of the buffers had errors then we can set the page uptodate, ntfs_end_buffer_async_read() 172 * We read each buffer asynchronously and when all buffers are read in, our io 225 * may leave some buffers unmapped which are now allocated. This is ntfs_read_block() 226 * not a problem since these buffers will just get mapped when a write ntfs_read_block() 244 /* Loop through all the buffers in the page. */ ntfs_read_block() 354 /* Lock the buffers. */ ntfs_read_block() 361 /* Finally, start i/o on the buffers. */ ntfs_read_block() 371 /* No i/o was scheduled on any of the buffers. */ ntfs_read_block() 387 * function, ntfs_read_block(), which in turn creates and reads in the buffers 536 * For a page with buffers, map and write the dirty buffers asynchronously 537 * under page writeback. For a page without buffers, create buffers for the 540 * If a page doesn't have buffers the page dirty state is definitive. If a page 541 * does have buffers, the page dirty state is just a hint, and the buffer dirty 542 * state is definitive. (A hint which has rules: dirty buffers against a clean 544 * particular a dirty page containing clean buffers for example.) 585 "buffers. Redirtying page so we try " ntfs_write_block() 589 * its buffers' dirty state as-is. ntfs_write_block() 617 * here, and the (potentially unmapped) buffers may become dirty at ntfs_write_block() 626 * Loop through all the buffers in the page, mapping all the dirty ntfs_write_block() 627 * buffers to disk addresses and handling any aliases from the ntfs_write_block() 637 * Mapped buffers outside i_size will occur, because ntfs_write_block() 639 * truncate in progress. The contents of such buffers ntfs_write_block() 652 /* Clean buffers are not written out, so no need to map them. */ ntfs_write_block() 683 * all non-uptodate buffers and set them uptodate (and ntfs_write_block() 684 * dirty?). Note, there aren't any non-uptodate buffers ntfs_write_block() 686 * FIXME: For an uptodate page, the buffers may need to ntfs_write_block() 692 // Zero any non-uptodate buffers up to i_size. ntfs_write_block() 707 // set_buffer_uptodate() on complete buffers as well as ntfs_write_block() 711 /* No need to map buffers that are already mapped. */ ntfs_write_block() 835 /* Setup all mapped, dirty buffers for async write i/o. */ ntfs_write_block() 875 /* Submit the prepared buffers for i/o. */ ntfs_write_block() 959 /* Make sure we have mapped buffers. */ ntfs_write_mst_block() 1089 * buffers in this record from the list of ntfs_write_mst_block() 1090 * buffers to write and clear their dirty bit ntfs_write_mst_block() 1115 /* If there were no dirty buffers, we are done. */ ntfs_write_mst_block() 1126 /* Skip buffers which are not at the beginning of records. */ ntfs_write_mst_block() 1149 * Remove the buffers in this mft record from ntfs_write_mst_block() 1150 * the list of buffers to write. ntfs_write_mst_block() 1177 * Mark all the buffers in this record clean as we do ntfs_write_mst_block() 1192 /* Lock buffers and start synchronous write i/o on them. */ ntfs_write_mst_block() 1211 /* Wait on i/o completion of buffers. */ ntfs_write_mst_block() 1241 * Skip buffers which are not at the beginning of ntfs_write_mst_block() 1247 /* Skip removed buffers (and hence records). */ ntfs_write_mst_block() 1341 * buffers associated with the page asynchronously. 1371 * The page may have dirty, unmapped buffers. Make them ntfs_writepage() 1520 * buffers' dirty state as-is. ntfs_writepage() 1704 without touching the buffers 1719 * Set the buffers and the page in which the ntfs record is located dirty. 1724 * If the page does not have buffers, we create them and set them uptodate. 1725 * The page may not be locked which is why we need to handle the buffers under 1726 * the mapping->private_lock. Once the buffers are marked dirty we no longer 1727 * need the lock since try_to_free_buffers() does not free dirty buffers.
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf2-core.c | 341 * video buffer memory for all buffers/planes on the queue and initializes the 344 * Returns the number of buffers successfully allocated. 361 /* Length stores number of planes for multiplanar buffers */ __vb2_queue_alloc() 403 dprintk(1, "allocated %d buffers, %d plane(s) each\n", __vb2_queue_alloc() 412 static void __vb2_free_mem(struct vb2_queue *q, unsigned int buffers) __vb2_free_mem() argument 417 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_free_mem() 423 /* Free MMAP buffers or release USERPTR buffers */ __vb2_free_mem() 434 * __vb2_queue_free() - free buffers at the end of the queue - video memory and 435 * related information, if no buffers are left return the queue to an 438 static int __vb2_queue_free(struct vb2_queue *q, unsigned int buffers) __vb2_queue_free() argument 446 * buffers from underneath __buf_prepare is obviously a bad idea, so we __vb2_queue_free() 447 * check if any of the buffers is in the state PREPARING, and if so we __vb2_queue_free() 450 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free() 455 dprintk(1, "preparing buffers, cannot free\n"); __vb2_queue_free() 461 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free() 470 __vb2_free_mem(q, buffers); __vb2_queue_free() 534 /* Free videobuf buffers */ __vb2_queue_free() 535 for (buffer = q->num_buffers - buffers; buffer < q->num_buffers; __vb2_queue_free() 541 q->num_buffers -= buffers; __vb2_queue_free() 637 * __buffers_in_use() - return true if any buffers on the queue are in use and 674 * single-planar buffers, but userspace does not. __fill_v4l2_buffer() 858 * 3) negotiates number of buffers and planes per buffer with the driver 884 * We already have buffers allocated, so first check if they __reqbufs() 895 * Call queue_cancel to clean up any buffers in the PREPARED or __reqbufs() 896 * QUEUED state which is possible if buffers were prepared or __reqbufs() 923 * Ask the driver how many buffers and planes per buffer it requires. __reqbufs() 931 /* Finally, allocate buffers and video memory */ __reqbufs() 940 * number of buffers needed by this vb2_queue. __reqbufs() 946 * Check if driver can handle the allocated number of buffers. __reqbufs() 958 * Either the driver has accepted a smaller number of buffers, __reqbufs() 978 * Return the number of successfully allocated buffers __reqbufs() 1002 * __create_bufs() - Allocate buffers and any required auxiliary structs 1022 dprintk(1, "maximum number of buffers already allocated\n"); __create_bufs() 1036 * Ask the driver, whether the requested number of buffers, planes per __create_bufs() 1044 /* Finally, allocate buffers and video memory */ __create_bufs() 1053 * Check if driver can handle the so far allocated number of buffers. __create_bufs() 1059 * q->num_buffers contains the total number of buffers, that the __create_bufs() 1069 * Either the driver has accepted a smaller number of buffers, __create_bufs() 1089 * Return the number of successfully allocated buffers __create_bufs() 1158 * If start_streaming fails then it should return buffers with state 1164 * by the means of buf_queue callback. Only buffers previously queued to the 1169 * be started for some reason. In that case the buffers should be returned with 1196 /* sync buffers */ vb2_buffer_done() 1200 /* Add the buffer to the done buffers list */ vb2_buffer_done() 1211 /* Inform any processes that may be waiting for buffers */ vb2_buffer_done() 1217 * vb2_discard_done() - discard all buffers marked as DONE 1221 * discards all 'done' buffers as they would be too old to be requested after 1320 * Single-planar buffers do not use planes array, __fill_vb2_buffer() 1323 * single-planar buffers as well, for simplicity. __fill_vb2_buffer() 1371 * For output buffers mask out the timecode flag: __fill_vb2_buffer() 1627 /* sync buffers */ __enqueue_in_driver() 1771 * at least q->min_buffers_needed buffers queued up (i.e. the minimum 1772 * number of buffers required for the DMA engine to function). If the 1774 * buffers back to vb2 in state QUEUED. Check if that happened and if 1783 * If any buffers were queued before streamon, vb2_start_streaming() 1802 * documentation in videobuf2-core.h for more information how buffers vb2_start_streaming() 1809 * Forcefully reclaim buffers if the driver did not vb2_start_streaming() 1856 * Add to the queued buffers list, a buffer will stay on it until vb2_internal_qbuf() 1865 * For output buffers copy the timestamp if needed, vb2_internal_qbuf() 1888 * start_streaming() since not enough buffers were queued, and vb2_internal_qbuf() 1889 * we now have reached the minimum number of queued buffers, vb2_internal_qbuf() 1941 * spinlock protection. However, buffers may be removed from __vb2_wait_for_done_vb() 1952 dprintk(1, "streaming off, will not wait for buffers\n"); __vb2_wait_for_done_vb() 1957 dprintk(1, "Queue in error state, will not wait for buffers\n"); __vb2_wait_for_done_vb() 1969 dprintk(1, "nonblocking and no buffers to dequeue, " __vb2_wait_for_done_vb() 1984 dprintk(3, "will sleep waiting for buffers\n"); __vb2_wait_for_done_vb() 2039 * vb2_wait_for_all_buffers() - wait until all buffers are given back to vb2 2042 * This function will wait until all buffers that have been given to the driver 2050 dprintk(1, "streaming off, will not wait for buffers\n"); vb2_wait_for_all_buffers() 2131 * buffers ready for dequeuing are present. Normally the driver 2159 * Removes all queued buffers from driver's queue and all buffers queued by 2168 * buffers. __vb2_queue_cancel() 2176 * videobuf2-core.h for more information how buffers should be returned __vb2_queue_cancel() 2193 * Remove all buffers from videobuf's list... __vb2_queue_cancel() 2197 * ...and done list; userspace will not receive any buffers it __vb2_queue_cancel() 2205 * Reinitialize all buffers for next use. __vb2_queue_cancel() 2206 * Make sure to call buf_finish for any queued buffers. Normally __vb2_queue_cancel() 2239 dprintk(1, "no buffers have been allocated\n"); vb2_internal_streamon() 2244 dprintk(1, "need at least %u allocated buffers\n", vb2_internal_streamon() 2250 * Tell driver to start streaming provided sufficient buffers vb2_internal_streamon() 2273 * buffers will return -EIO. 2296 * 2) passes any previously queued buffers to the driver and starts streaming 2319 * Cancel will pause streaming and remove all buffers from the driver vb2_internal_streamoff() 2323 * queue buffers, and then call streamoff without ever having called vb2_internal_streamoff() 2324 * streamon, you would still expect those buffers to be returned to vb2_internal_streamoff() 2342 * 2) stop streaming and dequeues any queued buffers, including those previously 2369 * Go over all buffers and their planes, comparing the given offset __find_plane_by_offset() 2467 * vb2_mmap() - map video buffers into application address space 2472 * This function maps one plane of one of the available video buffers to 2531 * MMAP requires page_aligned buffers. vb2_mmap() 2666 * buffers queued than there are buffers available. vb2_poll() 2761 * tracking context related to the buffers. 2776 * @initial_index: in the read() case all buffers are queued up immediately 2778 * buffers. However, in the write() case no buffers are initially 2780 * __vb2_perform_fileio(). Only once all available buffers have 2782 * buffers. This means that initially __vb2_perform_fileio() 2784 * the buffers for the first time. That initial index is stored 2786 * available buffers have been queued and __vb2_perform_fileio() 2826 * Check if device supports mapping buffers to kernel virtual space. __vb2_init_fileio() 2854 * Request buffers and use MMAP type to force driver __vb2_init_fileio() 2855 * to allocate buffers by itself. __vb2_init_fileio() 2867 * (multiplane buffers are not supported). __vb2_init_fileio() 2887 * Read mode requires pre queuing of all buffers. __vb2_init_fileio() 2893 * Queue all buffers. __vb2_init_fileio() 2913 * All buffers have been queued, so mark that by setting __vb2_init_fileio() 3112 * If we are queuing up buffers for the first time, then __vb2_perform_fileio() 3122 * all the 'first time' buffers. __vb2_perform_fileio() 3312 /* If count == 0, then the owner has released all buffers and he vb2_ioctl_reqbufs() 3357 /* No need to call vb2_queue_is_busy(), anyone can query buffers. */ vb2_ioctl_querybuf()
|
H A D | v4l2-mem2mem.c | 4 * Helper functions for devices that use videobuf buffers for both their 48 /* Offset base for buffers on the destination queue - used to distinguish 49 * between source and destination buffers when mmapping - they receive the same 96 * v4l2_m2m_next_buf() - return next buffer from the list of ready buffers 117 * v4l2_m2m_buf_remove() - take off a buffer from the list of ready buffers and 202 * to be drained before doing streamoff), allow scheduling without v4l2 buffers 246 dprintk("No input buffers available\n"); v4l2_m2m_try_schedule() 257 dprintk("No output buffers available\n"); v4l2_m2m_try_schedule() 345 /* This instance might have more buffers ready, but since we do not v4l2_m2m_job_finish() 518 * v4l2_m2m_poll() - poll replacement, for destination buffers only 608 * by adding a predefined offset to buffers from one of the queues and 738 * v4l2_m2m_buf_queue() - add a buffer to the proper ready buffers list.
|
H A D | videobuf-core.c | 2 * generic helper functions for handling video4linux capture buffers 40 MODULE_DESCRIPTION("helper module to manage video4linux buffers"); 224 * __videobuf_free() - free all the buffers and their control structures 226 * This function can only be called if streaming/reading is off, i.e. no buffers 239 dprintk(1, "Cannot free buffers when streaming or reading\n"); __videobuf_free() 247 dprintk(1, "Cannot free mmapped buffers\n"); __videobuf_free() 272 /* remove queued buffers from list */ videobuf_queue_cancel() 285 /* free all buffers + clear queue */ videobuf_queue_cancel() 397 /* Allocate and initialize buffers */ __videobuf_mmap_setup() 422 dprintk(1, "mmap setup: %d buffers, %d bytes each\n", i, bsize); __videobuf_mmap_setup() 646 dprintk(2, "next_buffer: no buffers to dequeue\n"); stream_next_buffer_check_queue()
|
/linux-4.1.27/arch/mips/include/asm/octeon/ |
H A D | cvmx-helper-util.h | 61 * this many packet buffers free in FPA 0. 64 * than this many free packet buffers in FPA 0. 75 * this many packet buffers free in FPA 0. 78 * than this many free packet buffers in FPA 0. 144 * Free the packet buffers contained in a work queue entry. 162 * Since the number of buffers is not zero, we know this is cvmx_helper_free_packet_data() 165 * true, we need to free all buffers except for the first cvmx_helper_free_packet_data()
|
H A D | cvmx-config.h | 33 /* Packet buffers */ 39 /* PKO queue command buffers */
|
/linux-4.1.27/drivers/media/platform/xilinx/ |
H A D | xilinx-dma.h | 67 * @queue: vb2 buffers queue 69 * @sequence: V4L2 buffers sequence number 70 * @queued_bufs: list of queued buffers
|
/linux-4.1.27/drivers/media/usb/pvrusb2/ |
H A D | pvrusb2-io.h | 70 /* Find out how many buffers are idle or ready */ 74 /* Kill all pending buffers and throw away any ready buffers as well */
|
H A D | pvrusb2-io.c | 62 /* Pointers to all buffers */ 63 struct pvr2_buffer **buffers; member in struct:pvr2_stream 64 /* Array size of buffers */ 66 /* Total buffers actually in circulation */ 68 /* Designed number of buffers to be in circulation */ 305 /* Allocate buffers pointer array in multiples of 32 entries */ pvr2_stream_buffer_count() 324 memcpy(nb,sp->buffers, pvr2_stream_buffer_count() 326 kfree(sp->buffers); pvr2_stream_buffer_count() 328 sp->buffers = nb; pvr2_stream_buffer_count() 340 sp->buffers[sp->buffer_total_count] = bp; pvr2_stream_buffer_count() 347 bp = sp->buffers[sp->buffer_total_count - 1]; pvr2_stream_buffer_count() 349 sp->buffers[sp->buffer_total_count - 1] = NULL; pvr2_stream_buffer_count() 357 nb = kmemdup(sp->buffers, scnt * sizeof(*nb), pvr2_stream_buffer_count() 361 kfree(sp->buffers); pvr2_stream_buffer_count() 362 sp->buffers = nb; pvr2_stream_buffer_count() 387 bp = sp->buffers[sp->buffer_total_count - (cnt + 1)]; pvr2_stream_achieve_buffer_count() 582 return sp->buffers[id]; pvr2_stream_get_buffer()
|
/linux-4.1.27/drivers/net/fddi/skfp/h/ |
H A D | osdef1st.h | 67 // default number of receive buffers. 70 // default number of transmit buffers. 73 // Number of SMT buffers (Mbufs).
|
/linux-4.1.27/fs/ |
H A D | buffer.c | 79 * Returns if the page has dirty or writeback buffers. If all the buffers 188 * But it's the page lock which protects the buffers. To get around this, 230 /* we might be here because some of the buffers on this page are __find_get_block_slow() 233 * elsewhere, don't buffer_error if we had some unmapped buffers __find_get_block_slow() 322 * If none of the buffers had errors and they are all end_buffer_async_read() 387 * If a page's buffers are under async readin (end_buffer_async_read 389 * control could lock one of the buffers after it has completed 390 * but while some of the other buffers have not completed. This 395 * The page comes unlocked when it has no locked buffer_async buffers 399 * the buffers. 436 * management of a list of dependent buffers at ->i_mapping->private_list. 438 * Locking is a little subtle: try_to_free_buffers() will remove buffers 441 * at the time, not against the S_ISREG file which depends on those buffers. 443 * which backs the buffers. Which is different from the address_space 444 * against which the buffers are listed. So for a particular address_space, 449 * Which introduces a requirement: all buffers on an address_space's 452 * address_spaces which do not place buffers at ->private_list via these 463 * mark_buffer_dirty_fsync() to clearly define why those buffers are being 470 * that buffers are taken *off* the old inode's list when they are freed 499 * you dirty the buffers, and then use osync_inode_buffers to wait for 500 * completion. Any other dirty buffers which are not yet queued for 560 * sync_mapping_buffers - write out & wait upon a mapping's "associated" buffers 561 * @mapping: the mapping which wants those buffers written 563 * Starts I/O against the buffers at mapping->private_list, and waits upon 567 * @mapping is a file or directory which needs those buffers to be written for 649 * If the page has buffers, the uptodate buffers are set dirty, to preserve 650 * dirty-state coherency between the page and the buffers. It the page does 651 * not have buffers then when they are later attached they will all be set 654 * The buffers are dirtied before the page is dirtied. There's a small race 657 * before the buffers, a concurrent writepage caller could clear the page dirty 658 * bit, see a bunch of clean buffers and we'd end up with dirty buffers/clean 662 * page's buffer list. Also use this to protect against clean buffers being 696 * Write out and wait upon a list of buffers. 699 * initially dirty buffers get waited on, but that any subsequently 700 * dirtied buffers don't. After all, we don't want fsync to last 703 * Do this in two main stages: first we copy dirty buffers to a 711 * the osync code to catch these locked, dirty buffers without requeuing 712 * any newly dirty buffers for write. 794 * Invalidate any and all dirty buffers on a given inode. We are 796 * done a sync(). Just drop the buffers from the inode list. 799 * assumes that all the buffers are against the blockdev. Not true 818 * Remove any clean buffers from the inode's buffer list. This is called 819 * when we're trying to free the inode itself. Those buffers can pin it. 821 * Returns true if all buffers were removed. 847 * Create the appropriate buffers when given a page for data area and 849 * follow the buffers created. Return NULL if unable to create more 850 * buffers. 895 * partially complete buffers, so all were released above. alloc_page_buffers() 938 * Initialise the state of a blockdev page's buffers. 1014 * Allocate some buffers for this page grow_dev_page() 1021 * Link the page to the buffers and initialise them. Take the grow_dev_page() 1039 * Create buffers for the specified block device block's page. If 1040 * that page was dirty, the buffers are set dirty also. 1069 /* Create a page with the proper size buffers.. */ grow_buffers() 1107 * The relationship between dirty buffers and dirty pages: 1109 * Whenever a page has any dirty buffers, the page's dirty bit is set, and 1112 * At all times, the dirtiness of the buffers represents the dirtiness of 1113 * subsections of the page. If the page has buffers, the page dirty bit is 1116 * When a page is set dirty in its entirety, all its buffers are marked dirty 1117 * (if the page has buffers). 1120 * buffers are not. 1122 * Also. When blockdev buffers are explicitly read with bread(), they 1124 * uptodate - even if all of its buffers are uptodate. A subsequent 1126 * buffers, will set the page uptodate and will perform no I/O. 1171 * Decrement a buffer_head's reference count. If all buffers against a page 1173 * and unlocked then try_to_free_buffers() may strip the buffers from the page 1174 * in preparation for freeing it (sometimes, rarely, buffers are removed from 1175 * a page but it ends up not being freed, and buffers may later be reattached). 1503 * block_invalidatepage() does not have to release all buffers, but it must 1547 * We release buffers only if the entire page is being invalidated. block_invalidatepage() 1560 * We attach and possibly dirty the buffers atomically wrt 1606 * be writeout I/O going on against recently-freed buffers. We don't 1662 * While block_write_full_page is writing back the dirty buffers under 1663 * the page lock, whoever dirtied the buffers may decide to clean them 1695 * here, and the (potentially unmapped) buffers may become dirty at __block_write_full_page() 1711 * Get all the dirty buffers mapped to disk addresses and __block_write_full_page() 1717 * mapped buffers outside i_size will occur, because __block_write_full_page() 1768 * The page and its buffers are protected by PageWriteback(), so we can __block_write_full_page() 1788 * The page was marked dirty, but the buffers were __block_write_full_page() 1809 /* Recovery: lock and submit the mapped buffers */ __block_write_full_page() 1841 * If a page has any new buffers, zero them out here, and mark them uptodate 1992 * If this is a partial write which happened to make all buffers __block_commit_write() 2042 * The buffers that were written will now be uptodate, so we block_write_end() 2108 * block_is_partially_uptodate checks whether buffers within a page are 2111 * Returns true if all buffers which correspond to a file portion 2213 * All buffers are uptodate - we can set the page uptodate block_read_full_page() 2222 /* Stage two: lock the buffers */ block_read_full_page() 2472 * Attach the singly-linked list of buffers created by nobh_write_begin, to 2540 * Allocate buffers so that we can keep track of state, and potentially nobh_write_begin() 2600 * The page is locked, so these buffers are protected from nobh_write_begin() 2698 * The page may have dirty, unmapped buffers. For example, nobh_writepage() 2902 * The page may have dirty, unmapped buffers. For example, block_write_full_page() 3075 * request. Further it marks as clean buffers that are processed for 3083 * All of the buffers must be for the same device, and must also be a 3160 * try_to_free_buffers() checks if all the buffers on this particular page 3166 * If the page is dirty but all the buffers are clean then we need to 3168 * may be against a block device, and a later reattachment of buffers 3169 * to a dirty page will set *all* buffers dirty. Which would corrupt 3172 * The same applies to regular filesystem pages: if all the buffers are 3233 * If the filesystem writes its buffers by hand (eg ext3) try_to_free_buffers() 3234 * then we can have clean buffers against a dirty page. We try_to_free_buffers() 3239 * the page's buffers clean. We discover that here and clean try_to_free_buffers()
|
H A D | splice.c | 9 * that transfers data buffers to or from a pipe buffer. 205 if (pipe->nrbufs < pipe->buffers) { splice_to_pipe() 206 int newbuf = (pipe->curbuf + pipe->nrbufs) & (pipe->buffers - 1); splice_to_pipe() 226 if (pipe->nrbufs < pipe->buffers) splice_to_pipe() 279 unsigned int buffers = ACCESS_ONCE(pipe->buffers); splice_grow_spd() local 281 spd->nr_pages_max = buffers; splice_grow_spd() 282 if (buffers <= PIPE_DEF_BUFFERS) splice_grow_spd() 285 spd->pages = kmalloc(buffers * sizeof(struct page *), GFP_KERNEL); splice_grow_spd() 286 spd->partial = kmalloc(buffers * sizeof(struct partial_page), GFP_KERNEL); splice_grow_spd() 743 * destination. It returns when there's no more buffers left in 751 * locking is required around copying the pipe buffers to the 789 pipe->curbuf = (pipe->curbuf + 1) & (pipe->buffers - 1); splice_from_pipe_feed() 809 * value (one) if pipe buffers are available. It will return zero 956 int nbufs = pipe->buffers; iter_file_splice_write() 976 if (unlikely(nbufs < pipe->buffers)) { iter_file_splice_write() 978 nbufs = pipe->buffers; iter_file_splice_write() 996 if (idx == pipe->buffers - 1) iter_file_splice_write() 1022 /* dismiss the fully eaten buffers, adjust the partial one */ iter_file_splice_write() 1031 pipe->curbuf = (pipe->curbuf + 1) & (pipe->buffers - 1); iter_file_splice_write() 1261 * the pipe buffers in question: splice_direct_to_actor() 1263 for (i = 0; i < pipe->buffers; i++) { splice_direct_to_actor() 1442 int buffers = 0, error = 0; get_iovec_page_array() local 1483 if (npages > pipe_buffers - buffers) get_iovec_page_array() 1484 npages = pipe_buffers - buffers; get_iovec_page_array() 1487 0, &pages[buffers]); get_iovec_page_array() 1498 partial[buffers].offset = off; get_iovec_page_array() 1499 partial[buffers].len = plen; get_iovec_page_array() 1503 buffers++; get_iovec_page_array() 1519 if (error < npages || buffers == pipe_buffers) get_iovec_page_array() 1526 if (buffers) get_iovec_page_array() 1527 return buffers; get_iovec_page_array() 1758 if (pipe->nrbufs < pipe->buffers) opipe_prep() 1764 while (pipe->nrbufs >= pipe->buffers) { opipe_prep() 1830 if (!ipipe->nrbufs || opipe->nrbufs >= opipe->buffers) { splice_pipe_to_pipe() 1831 /* Already processed some buffers, break */ splice_pipe_to_pipe() 1842 * managed to process any buffers. A zero return splice_pipe_to_pipe() 1851 nbuf = (opipe->curbuf + opipe->nrbufs) & (opipe->buffers - 1); splice_pipe_to_pipe() 1861 ipipe->curbuf = (ipipe->curbuf + 1) & (ipipe->buffers - 1); splice_pipe_to_pipe() 1928 * If we have iterated all input buffers or ran out of link_pipe() 1931 if (i >= ipipe->nrbufs || opipe->nrbufs >= opipe->buffers) link_pipe() 1934 ibuf = ipipe->bufs + ((ipipe->curbuf + i) & (ipipe->buffers-1)); link_pipe() 1935 nbuf = (opipe->curbuf + opipe->nrbufs) & (opipe->buffers - 1); link_pipe()
|
H A D | mpage.c | 93 * The idea is to avoid adding buffers to pages that don't already have 95 * this marks the page up to date instead of adding new buffers. 106 * don't make any buffers if there is only one buffer on map_buffer_to_page() 221 * we just collected from get_block into the page's buffers do_mpage_readpage() 327 * - encountering a page which has buffers 412 * If the page has buffers then they will be used for obtaining the disk 416 * If the page has no buffers (preferred) then the page is mapped here. 434 * We have our BIO, so we can now mark the buffers clean. Make 435 * sure to only clean buffers which we know we'll be writing. 496 * unmapped dirty buffers are created by __mpage_writepage() 528 * Page has buffers, but they are all unmapped. The page was __mpage_writepage() 537 * The page has no buffers: map it to disk __mpage_writepage()
|
H A D | pipe.c | 172 * in the tee() system call, when we duplicate the buffers in one 283 curbuf = (curbuf + 1) & (pipe->buffers - 1); pipe_read() 363 (pipe->buffers - 1); pipe_write() 396 if (bufs < pipe->buffers) { pipe_write() 397 int newbuf = (pipe->curbuf + bufs) & (pipe->buffers-1); pipe_write() 440 if (bufs < pipe->buffers) pipe_write() 489 buf = (buf+1) & (pipe->buffers - 1); pipe_ioctl() 519 mask |= (nrbufs < pipe->buffers) ? POLLOUT | POLLWRNORM : 0; pipe_poll() 597 pipe->buffers = PIPE_DEF_BUFFERS; alloc_pipe_info() 611 for (i = 0; i < pipe->buffers; i++) { free_pipe_info() 958 * Allocate a new array of pipe buffers and copy the info over. Returns the 969 * contains more buffers than arg, then return busy. pipe_set_size() 987 if (tail < pipe->buffers) pipe_set_size() 990 tail &= (pipe->buffers - 1); pipe_set_size() 1002 pipe->buffers = nr_pages; pipe_set_size() 1075 ret = pipe->buffers * PAGE_SIZE; pipe_fcntl()
|
/linux-4.1.27/drivers/rpmsg/ |
H A D | virtio_rpmsg_bus.c | 42 * @rbufs: kernel address of rx buffers 43 * @sbufs: kernel address of tx buffers 44 * @num_bufs: total number of buffers for rx and tx 46 * @bufs_dma: dma base addr of the buffers 52 * @sendq: wait queue of sending contexts waiting for a tx buffers 91 * We're allocating buffers of 512 bytes each for communications. The 92 * number of buffers will be computed from the number of buffers supported 93 * by the vring, upto a maximum of 512 buffers (256 in each direction). 98 * This will utilize a maximum total space of 256KB for the buffers. 100 * We might also want to add support for user-provided buffers in time. 582 * (half of our buffers are used for sending messages) get_a_tx_buf() 616 /* are we the first sleeping context waiting for tx buffers ? */ rpmsg_upref_sleepers() 643 /* are we the last sleeping context waiting for tx buffers ? */ rpmsg_downref_sleepers() 658 * @wait: indicates whether caller should block in case no TX buffers available 666 * The message is sent using one of the TX buffers that are available for 676 * Otherwise, if @wait is false, and there are no TX buffers available, 701 * We currently use fixed-sized buffers, and therefore the payload rpmsg_send_offchannel_raw() 705 * user-provided buffers (and then we can also support zero-copy rpmsg_send_offchannel_raw() 795 * We currently use fixed-sized buffers, so trivially sanitize rpmsg_recv_single() 979 /* we need less buffers if vrings are small */ rpmsg_probe() 987 /* allocate coherent memory for the buffers */ rpmsg_probe() 996 dev_dbg(&vdev->dev, "buffers: va %p, dma 0x%llx\n", bufs_va, rpmsg_probe() 999 /* half of the buffers is dedicated for RX */ rpmsg_probe() 1005 /* set up the receive buffers */ rpmsg_probe()
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_dma.c | 75 * Free all pages associated with DMA buffers, the buffers and pages lists, and 91 /* Clear dma buffers */ drm_legacy_dma_takedown() 140 * Reclaim the buffers.
|
/linux-4.1.27/arch/metag/include/uapi/asm/ |
H A D | sigcontext.h | 18 * buffers have been saved and will be replayed on sigreturn.
|
/linux-4.1.27/drivers/tty/ |
H A D | tty_buffer.c | 25 * Byte threshold to limit memory consumption for flip buffers. 108 * tty_buffer_free_all - free buffers used by a tty 111 * Remove all the buffers pending on a tty whether queued with data 143 * We round our buffers off in 256 character chunks to get better 202 * tty_buffer_flush - flush full tty buffers 206 * flush all the buffers containing receive data. If ld != NULL, 263 /* This is the slow path - looking for new buffers to use */ __tty_buffer_request_room() 317 several buffers. If this is the case we must loop */ tty_insert_flip_string_fixed_flag() 352 several buffers. If this is the case we must loop */ tty_insert_flip_string_flags() 362 * Takes any pending buffers and transfers their ownership to the 494 * Queue a push of the terminal flip buffers to the line discipline.
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
H A D | pr_pc.h | 35 byte RcOutput; /* number of Rc buffers received */ 36 byte IndOutput; /* number of Ind buffers received */
|
/linux-4.1.27/drivers/iio/ |
H A D | industrialio-triggered-buffer.c | 31 * buffers will be used. 102 MODULE_DESCRIPTION("IIO helper functions for setting up triggered buffers");
|
/linux-4.1.27/drivers/media/platform/omap/ |
H A D | omap_vout_vrfb.c | 29 * Function for allocating video buffers 75 * Free VRFB buffers 182 * Allocate the buffers for the VRFB space. Data is copied from V4L2 183 * buffers to the VRFB buffers using the DMA engine. 197 /* Allocate the VRFB buffers only if the buffers are not omap_vout_vrfb_buffer_setup() 281 /* Store buffers physical address into an array. Addresses omap_vout_prepare_vrfb()
|
/linux-4.1.27/drivers/scsi/mpt3sas/ |
H A D | mpt3sas_trigger_diag.c | 155 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_master() 162 /* check to see if trace buffers are currently released */ mpt3sas_trigger_master() 228 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_event() 235 /* check to see if trace buffers are currently released */ mpt3sas_trigger_event() 308 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_scsi() 315 /* check to see if trace buffers are currently released */ mpt3sas_trigger_scsi() 382 /* check to see if trace buffers are currently registered */ mpt3sas_trigger_mpi() 389 /* check to see if trace buffers are currently released */ mpt3sas_trigger_mpi()
|
/linux-4.1.27/drivers/net/wireless/ti/wlcore/ |
H A D | event.h | 31 * The event mechanism is based on a pair of event buffers (buffers A and
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | fixmap.h | 38 * these 'compile-time allocated' memory buffers are 43 * TLB entries of such buffers will not be flushed across
|
H A D | kgdb.h | 39 * buffers at least NUMREGBYTES*2 are needed for register packets,
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | fixmap.h | 33 * these 'compile-time allocated' memory buffers are 38 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/arch/score/include/asm/ |
H A D | fixmap.h | 28 * these 'compile-time allocated' memory buffers are 33 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/drivers/isdn/icn/ |
H A D | icn.h | 88 #define SHM_CBUF_OFFSET (0x200) /* Offset to comm-buffers in shm */ 89 #define SHM_DBUF_OFFSET (0x2000) /* Offset to data-buffers in shm */ 92 * Layout of card's data buffers 152 int rcvidx[ICN_BCH]; /* Index for above buffers */ 175 unsigned long memaddr; /* Address of memory mapped buffers */ 176 icn_shmem __iomem *shmem; /* Pointer to memory-mapped-buffers */
|
/linux-4.1.27/drivers/media/platform/marvell-ccic/ |
H A D | mcam-core.c | 37 * we must have physically contiguous buffers to bring frames into. 38 * These parameters control how many buffers we use, whether we 43 * The controller can cycle through three buffers. We could use 51 "Non-zero value causes DMA buffers to be allocated when the " 54 "successfully getting those buffers. This parameter is " 60 "The number of DMA buffers to allocate. Can be either two " 66 "The size of the allocated DMA buffers. If actual operating " 67 "parameters require larger buffers, an attempt to reallocate " 348 * Allocate in-kernel DMA buffers for vmalloc mode. 380 cam_err(cam, "Insufficient DMA buffers, cannot operate\n"); mcam_alloc_dma_bufs() 385 cam_warn(cam, "Will limp along with only 2 buffers\n"); mcam_alloc_dma_bufs() 405 * Set up DMA buffers when operating in vmalloc mode 410 * Store the first two Y buffers (we aren't supporting mcam_ctlr_dma_vmalloc() 446 if (list_empty(&cam->buffers)) { mcam_frame_tasklet() 452 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, mcam_frame_tasklet() 469 * Make sure our allocated buffers are up to the task. 538 * If there are no available buffers, go into single mode mcam_set_contig_buffer() 540 if (list_empty(&cam->buffers)) { mcam_set_contig_buffer() 548 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, mcam_set_contig_buffer() 627 buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, queue); mcam_sg_next_buffer() 652 if (list_empty(&cam->buffers)) { mcam_ctlr_dma_sg() 665 * a descriptor chain on the fly, since the controller buffers it 689 if (!list_empty(&cam->buffers)) { mcam_dma_sg_done() 1024 * Configuration. If we still don't have DMA buffers, mcam_read_setup() 1096 start = (cam->state == S_BUFWAIT) && !list_empty(&cam->buffers); mcam_vb_buf_queue() 1097 list_add(&mvb->queue, &cam->buffers); mcam_vb_buf_queue() 1114 INIT_LIST_HEAD(&cam->buffers); mcam_vb_start_streaming() 1119 * Videobuf2 sneakily hoards all the buffers and won't mcam_vb_start_streaming() 1123 * give us buffers soon. mcam_vb_start_streaming() 1125 if (cam->buffer_mode != B_vmalloc && list_empty(&cam->buffers)) { mcam_vb_start_streaming() 1146 /* They never gave us buffers */ mcam_vb_stop_streaming() 1160 * VB2 reclaims the buffers, so we need to forget mcam_vb_stop_streaming() 1164 INIT_LIST_HEAD(&cam->buffers); mcam_vb_stop_streaming() 1249 INIT_LIST_HEAD(&cam->buffers); mcam_setup_vb2() 1448 * Also can't if there are streaming buffers in place. mcam_vidioc_s_fmt_vid_cap() 1470 * Make sure we have appropriate DMA buffers. mcam_vidioc_s_fmt_vid_cap() 1539 * the level which controls the number of read buffers. 1927 INIT_LIST_HEAD(&cam->buffers); mccic_register() 1963 * If so requested, try to get our DMA buffers now. mccic_register() 1967 cam_warn(cam, "Unable to alloc DMA buffers at load" mccic_register() 2043 list_add(&cam->vb_bufs[0]->queue, &cam->buffers); mccic_resume()
|
H A D | mcam-core.h | 42 S_BUFWAIT /* streaming requested but no buffers yet */ 161 struct list_head buffers; /* Available frames */ member in struct:mcam_camera 166 /* DMA buffers - vmalloc mode */ 176 /* DMA buffers - DMA modes */ 360 #define C1_TWOBUFS 0x08000000 /* Use only two DMA buffers */
|
/linux-4.1.27/arch/tile/include/hv/ |
H A D | drv_xgbe_impl.h | 124 /** A queue of buffers. 126 * This structure partially defines a queue of empty buffers which have been 144 * new buffers are outstanding; when the requested write offset is one greater 148 * buffers at decreasing addresses. This makes the check for "is it time to 150 * buffers, and means that the value which defines the queue size, 171 * This queue has 10 slots, and thus can hold 9 buffers (_last_buffer = 9). 173 * valid, unread buffers in slots 2, 1, 0, 9, 8, and 7. The requested write 174 * pointer is at 4; thus, requests have been made to the IPP for buffers which 200 /** Offset to queues of empty buffers, one per size. */ 232 /** The next empty buffers to be read, one index per size. */
|
/linux-4.1.27/drivers/s390/cio/ |
H A D | qdio_main.c | 107 /* not all buffers processed */ qdio_check_ccq() 118 * @state: state of the extracted buffers 120 * @count: count of buffers to examine 121 * @auto_ack: automatically acknowledge buffers 124 * Stops processing if a state is different from the last buffers state. 154 * extracted buffers before trying again. qdio_do_eqbs() 172 * @state: new state of the buffers 174 * @count: how many buffers to change 176 * Returns the number of successfully changed buffers. 215 /* returns number of examined buffers and their common state in *state */ get_buf_states() 248 /* wrap-around safe setting of slsb states, returns number of changed buffers */ set_buf_states() 497 /* need to change ALL buffers to get more interrupts */ inbound_primed() 509 * Don't check 128 buffers, as otherwise qdio_inbound_q_moved get_inbound_buffer_frontier() 740 * Don't check 128 buffers, as otherwise qdio_inbound_q_moved get_outbound_buffer_frontier() 789 /* all buffers processed? */ qdio_outbound_q_done() 1489 * handle_inbound - reset processed input buffers 1490 * @q: queue containing the buffers 1493 * @count: how many buffers are emptied 1540 * handle_outbound - process filled outbound buffers 1541 * @q: queue containing the buffers 1544 * @count: how many buffers are filled 1578 /* try to fast requeue buffers */ handle_outbound() 1597 * do_QDIO - process input or output buffers 1602 * @count: how many buffers to process 1634 * qdio_start_irq - process input buffers 1676 * qdio_get_next_buffers - process input buffers 1680 * @error: buffers are in error state 1684 * = 0 - no new buffers found 1685 * > 0 - number of processed buffers
|
H A D | qdio_setup.c | 38 * qdio_free_buffers() - free qdio buffers 39 * @buf: array of pointers to qdio buffers 40 * @count: number of qdio buffers to free 52 * qdio_alloc_buffers() - allocate qdio buffers 53 * @buf: array of pointers to qdio buffers 54 * @count: number of qdio buffers to allocate 75 * qdio_reset_buffers() - reset qdio buffers 76 * @buf: array of pointers to qdio buffers 77 * @count: number of qdio buffers that will be zeroed
|
/linux-4.1.27/drivers/media/usb/uvc/ |
H A D | uvc_queue.c | 28 * Video buffers queue management. 33 * Video buffers are managed by videobuf2. The driver uses a mutex to protect 35 * spinlock to protect the IRQ queue that holds the buffers to be processed by 46 * Return all queued buffers to videobuf2 in the requested state. 337 * Check if buffers have been allocated. 351 * Cancel the video buffers queue. 353 * Cancelling the queue marks all buffers on the irq queue as erroneous,
|
/linux-4.1.27/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-helper-util.c | 176 * this many packet buffers free in FPA 0. 179 * than this many free packet buffers in FPA 0. 187 /* Set RED to begin dropping packets when there are pass_thresh buffers cvmx_helper_setup_red_queue() 189 buffers */ cvmx_helper_setup_red_queue() 211 * this many packet buffers free in FPA 0. 214 * than this many free packet buffers in FPA 0. 226 /* Disable backpressure based on queued buffers. It needs SW support */ cvmx_helper_setup_red()
|
/linux-4.1.27/drivers/media/usb/cpia2/ |
H A D | cpia2_v4l.c | 56 MODULE_PARM_DESC(num_buffers, "Number of frame buffers (1-" 190 struct framebuf *frame = &cam->buffers[frame_nr]; sync() 442 if (cam->buffers[frame].status == FRAME_READING) cpia2_s_fmt_vid_cap() 446 cam->buffers[frame].status = FRAME_EMPTY; cpia2_s_fmt_vid_cap() 783 * NOTE: The user's request is ignored. For now the buffers are fixed. 818 buf->m.offset = cam->buffers[buf->index].data - cam->frame_buffer; cpia2_querybuf() 830 switch (cam->buffers[buf->index].status) { cpia2_querybuf() 838 buf->bytesused = cam->buffers[buf->index].length; cpia2_querybuf() 839 buf->timestamp = cam->buffers[buf->index].timestamp; cpia2_querybuf() 840 buf->sequence = cam->buffers[buf->index].seq; cpia2_querybuf() 871 if(cam->buffers[buf->index].status == FRAME_READY) cpia2_qbuf() 872 cam->buffers[buf->index].status = FRAME_EMPTY; cpia2_qbuf() 890 if(cam->buffers[i].status == FRAME_READY) { find_earliest_filled_buffer() 896 tv1 = &cam->buffers[i].timestamp; find_earliest_filled_buffer() 897 tv2 = &cam->buffers[found].timestamp; find_earliest_filled_buffer() 947 buf->bytesused = cam->buffers[buf->index].length; cpia2_dqbuf() 951 buf->timestamp = cam->buffers[buf->index].timestamp; cpia2_dqbuf() 952 buf->sequence = cam->buffers[buf->index].seq; cpia2_dqbuf() 953 buf->m.offset = cam->buffers[buf->index].data - cam->frame_buffer; cpia2_dqbuf() 960 cam->buffers[buf->index].status, buf->sequence, buf->bytesused); cpia2_dqbuf() 1234 DBG("Using %d buffers, each %d bytes, alternate=%d\n", check_parameters()
|
H A D | cpia2_usb.c | 124 /* No junk in the buffers */ process_frame() 137 * mmapped, or we have few buffers, we need to process_frame() 141 * buffers, but it would take an EXTREMELY process_frame() 153 DBG("Changed buffers, work:%d, current:%d\n", process_frame() 257 break; /* No READING or EMPTY buffers left */ cpia2_usb_complete() 718 DBG("Flushing buffers\n"); cpia2_usb_stream_start() 720 cam->buffers[i].status = FRAME_EMPTY; cpia2_usb_stream_start() 721 cam->buffers[i].length = 0; cpia2_usb_stream_start() 723 cam->curbuff = &cam->buffers[0]; cpia2_usb_stream_start() 889 if(cam->buffers) { cpia2_usb_disconnect()
|
H A D | cpia2_core.c | 138 * the camera, copy the buffers into the proper param structures. 2237 if(!cam->buffers) { cpia2_allocate_buffers() 2239 cam->buffers = kmalloc(size, GFP_KERNEL); cpia2_allocate_buffers() 2240 if(!cam->buffers) { cpia2_allocate_buffers() 2250 kfree(cam->buffers); cpia2_allocate_buffers() 2251 cam->buffers = NULL; cpia2_allocate_buffers() 2257 cam->buffers[i].next = &cam->buffers[i+1]; cpia2_allocate_buffers() 2258 cam->buffers[i].data = cam->frame_buffer +i*cam->frame_size; cpia2_allocate_buffers() 2259 cam->buffers[i].status = FRAME_EMPTY; cpia2_allocate_buffers() 2260 cam->buffers[i].length = 0; cpia2_allocate_buffers() 2261 cam->buffers[i].max_length = 0; cpia2_allocate_buffers() 2262 cam->buffers[i].num = i; cpia2_allocate_buffers() 2264 cam->buffers[i].next = cam->buffers; cpia2_allocate_buffers() 2265 cam->buffers[i].data = cam->frame_buffer +i*cam->frame_size; cpia2_allocate_buffers() 2266 cam->buffers[i].status = FRAME_EMPTY; cpia2_allocate_buffers() 2267 cam->buffers[i].length = 0; cpia2_allocate_buffers() 2268 cam->buffers[i].max_length = 0; cpia2_allocate_buffers() 2269 cam->buffers[i].num = i; cpia2_allocate_buffers() 2270 cam->curbuff = cam->buffers; cpia2_allocate_buffers() 2272 DBG("buffers=%p, curbuff=%p, workbuff=%p\n", cam->buffers, cam->curbuff, cpia2_allocate_buffers() 2284 if(cam->buffers) { cpia2_free_buffers() 2285 kfree(cam->buffers); cpia2_free_buffers() 2286 cam->buffers = NULL; cpia2_free_buffers()
|
/linux-4.1.27/fs/jbd/ |
H A D | commit.c | 43 * ->mapping, and with attached buffers. These pages are trivially reclaimable 48 * see if we can strip all buffers from the backing page. 116 * cleaned up our previous buffers by now, so if we are in abort 180 * Submit all the data buffers to disk 198 * Cleanup any flushed data buffers from the data list. Even in journal_submit_data_buffers() 375 * BJ_Reserved buffers. Note, it is _not_ permissible to assume journal_commit_transaction() 376 * that there are no such buffers: if a large filesystem journal_commit_transaction() 379 * there are still BJ_Reserved buffers outstanding. These must journal_commit_transaction() 384 * we do not require it to remember exactly which old buffers it journal_commit_transaction() 408 * Now try to drop any written-back buffers from the journal's journal_commit_transaction() 419 * Clear revoked flag to reflect there is no revoked buffers journal_commit_transaction() 511 * If we found any dirty or locked buffers, then we should have journal_commit_transaction() 549 * any descriptor buffers which may have been journal_commit_transaction() 615 rid of the BJ_IO/BJ_Shadow pairing of buffers. */ journal_commit_transaction() 704 complete. Control buffers being written are on the journal_commit_transaction() 705 transaction's t_log_list queue, and metadata buffers are on journal_commit_transaction() 708 Wait for the buffers in reverse order. That way we are journal_commit_transaction() 779 /* Here we wait for the revoke record and descriptor record buffers */ journal_commit_transaction() 822 processing: any buffers committed as a result of this journal_commit_transaction() 837 * As there are other places (journal_unmap_buffer()) adding buffers journal_commit_transaction() 962 * Now recheck if some buffers did not get attached to the transaction journal_commit_transaction()
|
H A D | checkpoint.c | 197 * Clean up transaction's list of buffers submitted for io. 199 * buffers. Note that we take the buffers in the opposite ordering 202 * Return 0 on success, and return <0 if some buffers have failed 349 * list of transactions to be checkpointed and send all its buffers 461 * buffers which should be written-back to the filesystem. 550 * Find all the written-back checkpoint buffers in the given list and release 554 * Returns number of buffers reaped (for debug) 598 * Find all the written-back checkpoint buffers in the journal and release them. 602 * Returns number of buffers reaped (for debug) 634 * we can possibly see not yet submitted buffers on io_list __journal_clean_checkpoint_list() 652 * elsewhere on disk. To achieve this, all of the buffers in a
|
/linux-4.1.27/drivers/s390/char/ |
H A D | sclp_vt220.c | 63 /* List of empty pages to be used as write request buffers */ 100 /* Registration structure for SCLP output event buffers */ 106 /* Registration structure for SCLP input event buffers */ 115 * buffers if necessary. 124 /* Put buffer back to list of empty buffers */ sclp_vt220_process_queue() 177 /* Remove processed buffers and requeue rest */ sclp_vt220_callback() 179 /* Not all buffers were processed */ sclp_vt220_callback() 231 /* Only emit buffers with content */ sclp_vt220_emit_current() 394 * more empty buffers available, wait until one gets emptied. If DO_SCHEDULE 481 * Called by the SCLP to report incoming event buffers. 561 * to change as output buffers get emptied, or if the output flow 605 * Pass on all buffers to the hardware. Return only when there are no more 606 * buffers pending.
|
H A D | sclp_rw.h | 60 /* The number of empty mto buffers that can be contained in a single sccb. */
|
/linux-4.1.27/arch/avr32/include/asm/ |
H A D | dma-mapping.h | 190 * dma_map_sg - map a set of SG buffers for streaming mode DMA 192 * @sg: list of buffers 193 * @nents: number of buffers to map 196 * Map a set of buffers described by scatterlist in streaming 229 * dma_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg 231 * @sg: list of buffers 232 * @nents: number of buffers to map 302 * @sg: list of buffers 303 * @nents: number of buffers to map
|
/linux-4.1.27/drivers/staging/goldfish/ |
H A D | goldfish_audio.c | 54 * We will allocate two read buffers and two write buffers. 55 * Having two read buffers facilitate stereo -> mono conversion. 56 * Having two write buffers facilitate interleaved IO. 240 * if buffers are newly empty, wake up blocked goldfish_audio_interrupt()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_wc_x86_64.c | 61 * Set the PIO buffers to be WCCOMB, so we get HT bursts to the ipath_enable_wc() 65 * the PIO buffers themselves. ipath_enable_wc() 66 * For chips with two sets of buffers, the calculations are ipath_enable_wc() 69 * The buffers are still packed, so a single range covers both. ipath_enable_wc() 159 ipath_cdbg(VERBOSE, "undoing WCCOMB on pio buffers\n"); ipath_disable_wc()
|
H A D | ipath_init_chip.c | 45 * min buffers we want to have per port, after driver 51 * buffers per port, etc.) Zero means use chip value. 59 * Number of buffers reserved for driver (verbs and layered drivers.) 60 * Initialized based on number of PIO buffers if not set via module interface. 70 MODULE_PARM_DESC(kpiobufs, "Set number of PIO buffers for driver"); 73 * create_port0_egr - allocate the eager TID buffers 80 * Allocate the eager TID buffers and program them into infinipath. 82 * memory, and either use the buffers as is for things like verbs 83 * packets, or pass the buffers up to the ipath layered driver and 315 * 4K buffers take 2 pages; we use roundup just to be init_chip_first() 549 * in initialization, to busy out buffers as needed enable_chip() 772 /* if not an even divisor, some user ports get extra buffers */ ipath_init_chip() 916 * PIO buffers in early_init to avoid false parity errors. ipath_init_chip()
|
/linux-4.1.27/fs/jbd2/ |
H A D | commit.c | 56 * ->mapping, and with attached buffers. These pages are trivially reclaimable 61 * see if we can strip all buffers from the backing page. 113 * cleaned up our previous buffers by now, so if we are in abort 210 * Submit all the data buffers of inode associated with the transaction to 230 * submit the inode data buffers. We use writepage journal_submit_data_buffers() 475 * BJ_Reserved buffers. Note, it is _not_ permissible to assume jbd2_journal_commit_transaction() 476 * that there are no such buffers: if a large filesystem jbd2_journal_commit_transaction() 479 * there are still BJ_Reserved buffers outstanding. These must jbd2_journal_commit_transaction() 484 * we do not require it to remember exactly which old buffers it jbd2_journal_commit_transaction() 508 * Now try to drop any written-back buffers from the journal's jbd2_journal_commit_transaction() 519 * Clear revoked flag to reflect there is no revoked buffers jbd2_journal_commit_transaction() 605 * any descriptor buffers which may have been jbd2_journal_commit_transaction() 669 rid of the shadow pairing of buffers. */ jbd2_journal_commit_transaction() 812 complete. Control buffers being written are on the jbd2_journal_commit_transaction() 813 transaction's t_log_list queue, and metadata buffers are on jbd2_journal_commit_transaction() 816 Wait for the buffers in reverse order. That way we are jbd2_journal_commit_transaction() 865 /* Here we wait for the revoke record and descriptor record buffers */ jbd2_journal_commit_transaction() 919 processing: any buffers committed as a result of this jbd2_journal_commit_transaction() 932 * As there are other places (journal_unmap_buffer()) adding buffers jbd2_journal_commit_transaction() 1062 * Now recheck if some buffers did not get attached to the transaction jbd2_journal_commit_transaction()
|
H A D | checkpoint.c | 202 * list of transactions to be checkpointed and send all its buffers 251 /* checkpoint all of the transaction's buffers */ jbd2_log_do_checkpoint() 326 * Now we issued all of the transaction's buffers, let's deal jbd2_log_do_checkpoint() 327 * with the buffers that are out for I/O. jbd2_log_do_checkpoint() 384 * buffers which should be written-back to the filesystem. 419 * Find all the written-back checkpoint buffers in the given list and 420 * release them. If 'destroy' is set, clean all buffers unconditionally. 464 * Find all the written-back checkpoint buffers in the journal and release them. 465 * If 'destroy' is set, release all buffers unconditionally. 497 * we can possibly see not yet submitted buffers on io_list __jbd2_journal_clean_checkpoint_list() 514 * Remove buffers from all checkpoint lists as journal is aborted and we just 542 * elsewhere on disk. To achieve this, all of the buffers in a
|
/linux-4.1.27/fs/nilfs2/ |
H A D | segment.h | 103 * @sc_segbufs: List of segment buffers 104 * @sc_write_logs: List of segment buffers to hold logs under writing 105 * @sc_segbuf_nblocks: Number of available blocks in segment buffers. 129 * @sc_watermark: Watermark for the number of dirty buffers 151 /* Segment buffers */
|
H A D | segbuf.h | 68 * @sb_segsum_buffers: List of buffers for segment summaries 69 * @sb_payload_buffers: List of buffers for segment payload
|
H A D | page.h | 41 BUFFER_FNS(NILFS_Node, nilfs_node) /* nilfs node buffers */
|
/linux-4.1.27/include/linux/iio/ |
H A D | buffer.h | 21 * struct iio_buffer_access_funcs - access functions for buffers. 73 * @buffer_list: [INTERN] entry in the devices list of current buffers. 126 * iio_push_to_buffers_with_timestamp() - push data and timestamp to buffers 131 * Pushes data to the IIO device's buffers. If timestamps are enabled for the 133 * the sample data buffer before pushing it to the device buffers. The sample
|
/linux-4.1.27/drivers/misc/carma/ |
H A D | carma-fpga.c | 45 * capture rate, all buffers are pre-allocated to avoid any potentially long 49 * different states of data buffers. 52 * This list holds all empty data buffers which are ready to receive data. 59 * This list holds data buffers which have been filled, and are waiting to be 62 * All buffers start life on the free list, then move successively to the 67 * It should be noted that all buffers are mapped and ready for DMA when they 180 /* Correlation data buffers */ 186 /* Information about data buffers */ 334 * data_free_buffers() - free all allocated buffers 337 * Free all buffers allocated by the driver (except those currently in the 367 * data_alloc_buffers() - allocate 1 seconds worth of data buffers 370 * Allocate enough buffers for a whole second worth of data 373 * second worth of data buffers could not be allocated, as long as a minimum 403 /* add it to the list of free buffers */ data_alloc_buffers() 408 /* Make sure we allocated the minimum required number of buffers */ data_alloc_buffers() 410 dev_err(priv->dev, "Unable to allocate enough data buffers\n"); data_alloc_buffers() 418 "Unable to allocate %d buffers, using %d buffers instead\n", data_alloc_buffers() 708 * All buffers passed to this function should be ready and mapped data_submit_dma() 796 /* If there are no free buffers, drop this data */ data_irq() 837 * Enable the device for buffered dumping. Allocates buffers and hooks up 872 /* allocate the correlation data buffers */ data_device_enable() 875 dev_err(priv->dev, "unable to allocate buffers\n"); data_device_enable() 919 * all buffers. 962 /* free all buffers: the free and used lists are not being changed */ data_device_disable()
|
/linux-4.1.27/drivers/isdn/sc/ |
H A D | packet.c | 45 pr_debug("%s: out of TX buffers\n", sndpkt() 164 pr_debug("%s: calculating buffer space: %d buffers, %d big\n", setup_buffers() 187 * Prep the receive buffers setup_buffers()
|
H A D | hardware.h | 24 #define BUFFER_OFFSET 0x1800 /* Beginning of buffers */ 77 buffers per channel */
|
/linux-4.1.27/drivers/media/usb/stk1160/ |
H A D | stk1160.h | 36 /* Number of buffers for isoc transfers */ 44 /* Max and min video buffers */ 106 /* transfer buffers for isoc transfer */
|
H A D | stk1160-video.c | 328 /* Reset urb buffers */ stk1160_isoc_irq() 370 * Releases urb and transfer buffers 378 stk1160_dbg("freeing %d urb buffers...\n", num_bufs); stk1160_free_isoc() 408 stk1160_dbg("all urb buffers freed\n"); stk1160_free_isoc() 459 /* allocate urbs and transfer buffers */ stk1160_alloc_isoc() 479 /* Not enough transfer buffers, so just give up */ stk1160_alloc_isoc() 515 /* At last we can say we have some buffers */ stk1160_alloc_isoc() 536 /* Save the allocated buffers so far, so we can properly free them */ stk1160_alloc_isoc()
|
/linux-4.1.27/drivers/net/wireless/libertas/ |
H A D | dev.h | 32 u32 fwd_drop_nobuf; /* Fwd: Run out of internal buffers */ 128 struct list_head cmdfreeq; /* free command buffers */ 129 struct list_head cmdpendingq; /* pending command buffers */
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
H A D | i915_gem_batch_pool.c | 30 * In order to submit batch buffers as 'secure', the software command parser 34 * managing the address space bindings for such buffers. 37 * set of scratch buffers to use for this purpose. The framework can be
|
/linux-4.1.27/net/sunrpc/ |
H A D | backchannel_rqst.c | 55 * buffers hanging off of it. 71 * Preallocate up to min_reqs structures and related buffers for use 74 * preallocated buffers are added to the pool of resources used by 100 * buffers. Once we're done building the list we splice it xprt_setup_backchannel() 237 * Return the preallocated rpc_rqst structure and XDR buffers 275 * backchannel setup. Buffer space for the send and private XDR buffers
|
H A D | sunrpc.h | 33 * Header for dynamically allocated rpc buffers.
|
/linux-4.1.27/drivers/net/ethernet/aeroflex/ |
H A D | greth.h | 70 /* How many pages are needed for buffers */ 76 * Set to 1520 to make all buffers word aligned for non-gbit MAC.
|
/linux-4.1.27/drivers/net/slip/ |
H A D | slip.h | 63 /* These are pointers to the malloc()ed frame buffers. */ 70 int buffsize; /* Max buffers sizes */
|
/linux-4.1.27/drivers/media/usb/stkwebcam/ |
H A D | stk-webcam.h | 51 /* Streaming IO buffers */ 120 /* Streaming buffers */
|
/linux-4.1.27/drivers/spi/ |
H A D | spi-fsl-lib.h | 53 /* functions to deal with different sized buffers */ 84 /* functions to deal with different sized buffers */
|
/linux-4.1.27/drivers/staging/android/ion/ |
H A D | ion_priv.h | 37 * @node: node in the ion_device buffers tree 47 * @lock: protects the buffers cnt fields 53 * handler and only valid for buffers that are faulted in 83 /* used to track orphaned buffers */ 162 * Represents a pool of memory from which buffers can be made. In some 164 * On others, some blocks might require large physically contiguous buffers 198 * in, this can affect how buffers are allocated from the heap. 227 * some helpers for common operations on buffers using the sg_table
|
/linux-4.1.27/drivers/net/wireless/ti/wl1251/ |
H A D | event.h | 29 * The event mechanism is based on a pair of event buffers (buffers A and
|
H A D | rx.h | 36 * The size of each of the two buffers is large enough to hold the longest 41 * 2) The host reads the received packet from one of the double buffers.
|
/linux-4.1.27/drivers/video/fbdev/mb862xx/ |
H A D | mb862xxfb.h | 61 unsigned long cap_buf; /* capture buffers offset */ 62 size_t cap_len; /* length of capture buffers */
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | fixmap.h | 33 * these 'compile-time allocated' memory buffers are 38 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/Documentation/video4linux/ |
H A D | v4l2-pci-skeleton.c | 60 * @buf_list: list of buffers queued for DMA 161 * checks if sufficient buffers have been allocated. Usually 3 is a good 162 * minimum number: many DMA engines need a minimum of 2 buffers in the 244 * Start streaming. First check if the minimum number of buffers have been 247 * buffers are available to actually start the DMA engine. 260 * In case of an error, return all active buffers to the start_streaming() 269 * Stop the DMA engine. Any remaining buffers in the DMA queue are dequeued 278 /* Release all active buffers */ stop_streaming() 383 * It is not allowed to change the format while buffers for use with skeleton_s_fmt_vid_cap() 425 * the buffers for streaming have been set up, so we have to allow for skeleton_s_std() 433 * while buffers for use with streaming have already been allocated. skeleton_s_std() 517 * while buffers for use with streaming have already been allocated. skeleton_s_dv_timings() 636 * while buffers for use with streaming have already been allocated. skeleton_s_input() 699 * the buffers can call the streaming ioctls, all other filehandles will 833 * Assume that this DMA engine needs to have at least two buffers skeleton_probe() 835 * won't be called until at least this many buffers are queued up. skeleton_probe() 850 * the vb2 core will allocate the buffers in 32-bit DMA memory. skeleton_probe()
|
/linux-4.1.27/kernel/trace/ |
H A D | ring_buffer.c | 119 * A fast way to enable or disable all ring buffers is to 120 * call tracing_on or tracing_off. Turning off the ring buffers 121 * prevents all ring buffers from being recorded to. 133 * will permantly disable all ring buffers. 137 * Global flag to disable all recording to ring buffers 142 * 0 0 : ring buffers are off 143 * 1 0 : ring buffers are on 144 * X 1 : ring buffers are permanently disabled 159 /* Used for individual buffers (after the counter) */ 165 * tracing_off_permanent - permanently disable ring buffers 167 * This function, once called, will disable all ring buffers 504 struct ring_buffer_per_cpu **buffers; member in struct:ring_buffer 547 * as data is added to any of the @buffer's cpu buffers. Otherwise 569 cpu_buffer = buffer->buffers[cpu]; ring_buffer_wait() 648 * as data is added to any of the @buffer's cpu buffers. Otherwise 651 * Returns POLLIN | POLLRDNORM if data exists in the buffers, 666 cpu_buffer = buffer->buffers[cpu]; ring_buffer_poll_wait() 1368 buffer->buffers = kzalloc(ALIGN(bsize, cache_line_size()), __ring_buffer_alloc() 1370 if (!buffer->buffers) __ring_buffer_alloc() 1374 buffer->buffers[cpu] = for_each_buffer_cpu() 1376 if (!buffer->buffers[cpu]) for_each_buffer_cpu() 1393 if (buffer->buffers[cpu]) for_each_buffer_cpu() 1394 rb_free_cpu_buffer(buffer->buffers[cpu]); for_each_buffer_cpu() 1396 kfree(buffer->buffers); 1425 rb_free_cpu_buffer(buffer->buffers[cpu]); ring_buffer_free() 1431 kfree(buffer->buffers); ring_buffer_free() 1715 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 1744 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 1760 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 1775 cpu_buffer = buffer->buffers[cpu_id]; 1824 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 1837 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 2754 cpu_buffer = buffer->buffers[cpu]; ring_buffer_lock_reserve() 2857 cpu_buffer = buffer->buffers[cpu]; ring_buffer_unlock_commit() 2953 cpu_buffer = buffer->buffers[cpu]; ring_buffer_discard_commit() 3017 cpu_buffer = buffer->buffers[cpu]; ring_buffer_write() 3163 cpu_buffer = buffer->buffers[cpu]; ring_buffer_record_disable_cpu() 3183 cpu_buffer = buffer->buffers[cpu]; ring_buffer_record_enable_cpu() 3216 cpu_buffer = buffer->buffers[cpu]; ring_buffer_oldest_event_ts() 3247 cpu_buffer = buffer->buffers[cpu]; ring_buffer_bytes_cpu() 3266 cpu_buffer = buffer->buffers[cpu]; ring_buffer_entries_cpu() 3286 cpu_buffer = buffer->buffers[cpu]; ring_buffer_overrun_cpu() 3309 cpu_buffer = buffer->buffers[cpu]; ring_buffer_commit_overrun_cpu() 3331 cpu_buffer = buffer->buffers[cpu]; ring_buffer_dropped_events_cpu() 3351 cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_events_cpu() 3371 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 3394 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 3893 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_peek() 3974 cpu_buffer = buffer->buffers[cpu]; ring_buffer_consume() 4032 cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_prepare() 4162 return BUF_PAGE_SIZE * buffer->buffers[cpu]->nr_pages; ring_buffer_size() 4215 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_reset_cpu() 4248 * @buffer: The ring buffer to reset all cpu buffers 4275 cpu_buffer = buffer->buffers[cpu]; for_each_buffer_cpu() 4309 cpu_buffer = buffer->buffers[cpu]; ring_buffer_empty_cpu() 4324 * ring_buffer_swap_cpu - swap a CPU buffer between two ring buffers 4344 cpu_buffer_a = buffer_a->buffers[cpu]; ring_buffer_swap_cpu() 4345 cpu_buffer_b = buffer_b->buffers[cpu]; ring_buffer_swap_cpu() 4347 /* At least make sure the two buffers are somewhat the same */ ring_buffer_swap_cpu() 4383 buffer_a->buffers[cpu] = cpu_buffer_b; ring_buffer_swap_cpu() 4384 buffer_b->buffers[cpu] = cpu_buffer_a; ring_buffer_swap_cpu() 4483 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; ring_buffer_read_page() 4667 nr_pages = buffer->buffers[cpu_i]->nr_pages; for_each_buffer_cpu() 4668 if (nr_pages != buffer->buffers[cpu_i]->nr_pages) { for_each_buffer_cpu() 4676 buffer->buffers[cpu] = 4678 if (!buffer->buffers[cpu]) { 4715 * is displayed and all ring buffers are disabled.
|
/linux-4.1.27/arch/arm/mm/ |
H A D | dma-mapping.c | 886 * arm_dma_map_sg - map a set of SG buffers for streaming mode DMA 888 * @sg: list of buffers 889 * @nents: number of buffers to map 892 * Map a set of buffers described by scatterlist in streaming mode for DMA. 926 * arm_dma_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg 928 * @sg: list of buffers 929 * @nents: number of buffers to unmap (same as was passed to dma_map_sg) 950 * @sg: list of buffers 951 * @nents: number of buffers to map (returned from dma_map_sg) 969 * @sg: list of buffers 970 * @nents: number of buffers to map (returned from dma_map_sg) 1583 * arm_coherent_iommu_map_sg - map a set of SG buffers for streaming mode DMA 1585 * @sg: list of buffers 1586 * @nents: number of buffers to map 1589 * Map a set of i/o coherent buffers described by scatterlist in streaming 1601 * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA 1603 * @sg: list of buffers 1604 * @nents: number of buffers to map 1607 * Map a set of buffers described by scatterlist in streaming mode for DMA. 1637 * arm_coherent_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg 1639 * @sg: list of buffers 1640 * @nents: number of buffers to unmap (same as was passed to dma_map_sg) 1653 * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg 1655 * @sg: list of buffers 1656 * @nents: number of buffers to unmap (same as was passed to dma_map_sg) 1671 * @sg: list of buffers 1672 * @nents: number of buffers to map (returned from dma_map_sg) 1689 * @sg: list of buffers 1690 * @nents: number of buffers to map (returned from dma_map_sg)
|
/linux-4.1.27/drivers/isdn/hysdn/ |
H A D | hysdn_net.c | 29 #define MAX_SKB_BUFFERS 20 /* number of buffers for keeping TX-data */ 47 int sk_count; /* number of buffers currently in ring */ 84 /* flush the currently occupied tx-buffers */ 304 flush_tx_buffers((struct net_local *) dev); /* empty buffers */ hysdn_net_release()
|
H A D | hysdn_proclog.c | 98 /* Flushes buffers not longer in use. */ 236 /* count is decremented and all buffers are noticed of closing. If this file */ 237 /* was the last one to be closed, all buffers are freed. */ 266 inf->usage_cnt--; /* decrement usage count for buffers */ hysdn_log_close() 271 if (pd->if_used <= 0) /* delete buffers if last file closed */ hysdn_log_close()
|
/linux-4.1.27/drivers/media/usb/au0828/ |
H A D | au0828.h | 142 /* transfer buffers for isoc transfer */ 156 /* Stores already requested buffers */ 269 char *transfer_buffer[AU0828_MAX_ISO_BUFS];/* transfer buffers for isoc 276 /* Preallocated transfer digital transfer buffers */
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | workarounds.h | 30 /* Spurious parity errors in TSORT buffers */
|
/linux-4.1.27/drivers/input/serio/ |
H A D | i8042.h | 51 * the i8042 buffers.
|
/linux-4.1.27/arch/tile/lib/ |
H A D | checksum.c | 55 * large buffers, but those should be impossible. do_csum()
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | kgdb.h | 11 * buffers at least NUMREGBYTES*2 are needed for register packets
|
H A D | fixmap.h | 60 * These 'compile-time allocated' memory buffers are 65 * TLB entries of such buffers will not be flushed across
|
/linux-4.1.27/drivers/gpu/drm/udl/ |
H A D | udl_transfer.c | 101 * multiple buffers to be potentially encoded and sent in parallel). 105 * rendering can be done directly to output buffers (e.g. USB URBs). 108 * buffers if the line will take several buffers to complete.
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_log_recover.h | 35 * item headers are in ri_buf[0]. Additional buffers follow.
|
H A D | xfs_symlink_remote.c | 129 /* no verification of non-crc buffers */ xfs_symlink_read_verify() 149 /* no verification of non-crc buffers */ xfs_symlink_write_verify()
|
/linux-4.1.27/include/net/ |
H A D | af_rxrpc.h | 20 * the mark applied to socket buffers that may be intercepted
|
/linux-4.1.27/include/uapi/drm/ |
H A D | msm_drm.h | 62 * GEM buffers: 173 /* Each cmdstream submit consists of a table of buffers involved, and 174 * one or more cmdstream buffers. This allows for conditional execution 175 * (context-restore), and IB buffers needed for per tile/bin draw cmds.
|
/linux-4.1.27/arch/powerpc/sysdev/ |
H A D | udbg_memcons.c | 3 * buffers.
|
/linux-4.1.27/arch/mips/pnx833x/common/ |
H A D | setup.c | 41 /* fake pci bus to avoid bounce buffers */ plat_mem_setup()
|
/linux-4.1.27/arch/mn10300/include/asm/ |
H A D | kgdb.h | 17 * buffers at least NUMREGBYTES*2 are needed for register packets
|
/linux-4.1.27/arch/powerpc/include/uapi/asm/ |
H A D | ps3fb.h | 42 __u32 num_frames; /* num of frame buffers */
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | fixmap.h | 28 * These 'compile-time allocated' memory buffers are
|
/linux-4.1.27/arch/c6x/platforms/ |
H A D | emif.c | 60 /* look for a dscr-based enable for emifa pin buffers */ c6x_emifa_init()
|
/linux-4.1.27/fs/ext4/ |
H A D | truncate.h | 9 * pagecache as well so that corresponding buffers get properly unmapped.
|
/linux-4.1.27/include/drm/ |
H A D | drm_gem_cma_helper.h | 11 * @sgt: scatter/gather table for imported PRIME buffers
|
/linux-4.1.27/drivers/net/caif/ |
H A D | caif_virtio.c | 83 * @tx_release_tasklet: Tasklet for freeing consumed TX buffers 92 * @allocsz: size of dma memory reserved for TX buffers 93 * @alloc_addr: virtual address to dma memory for TX buffers 94 * @alloc_dma: dma address to dma memory for TX buffers 95 * @genpool: Gen Pool used for allocating TX buffers 176 /* Stop looping if there are no more buffers to free */ cfv_release_used_buf() 466 /* Release any TX buffers on both used and avilable rings */ cfv_netdev_close() 531 /* garbage collect released buffers */ cfv_netdev_tx() 538 * Flow-on is triggered when sufficient buffers are freed cfv_netdev_tx() 564 /* Enable notifications of recycled TX buffers */ cfv_netdev_tx()
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/ |
H A D | iwl-fh.h | 103 * circular buffers (CBs/queues) containing Transmit Frame Descriptors (TFDs) 139 * concatenate up to 20 DRAM buffers to form a Tx frame, each Receive Buffer 199 * buffers waiting to be filled. 389 * (channel's buffers empty | no pending requests). 392 * 31-24: 1 = Channel buffers empty (channel 7:0) 504 * Both driver and device share these circular buffers, each of which must be 510 * Each TFD contains pointer/size information for up to 20 data buffers 511 * in host DRAM. These buffers collectively contain the (one) frame described 513 * itself, but buffers may be scattered in host DRAM. Each buffer has max size 514 * of (4K - 4). The concatenates all of a TFD's buffers into a single
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 105 unsigned int credits; /* # of available Rx buffers */ 106 unsigned int pend_cred; /* new buffers since last FL DB ring */ 119 unsigned long empty; /* # of times queue ran out of buffers */ 211 struct timer_list tx_reclaim_timer; /* reclaims TX buffers */ 212 struct timer_list rx_reclaim_timer; /* reclaims RX buffers */
|
/linux-4.1.27/arch/tile/include/arch/ |
H A D | mpipe.h | 106 * This MMIO region is used for posting or fetching buffers to/from the 152 * This MMIO region is used for posting or fetching buffers to/from the 184 * egressed, buffers with HWB set will be returned to the other mPIPE 193 * Ignored on writes since all buffers on a given stack are the same size. 308 * software to determine when buffers associated with a descriptor may be
|
/linux-4.1.27/sound/oss/dmasound/ |
H A D | dmasound_core.c | 418 if (sq->buffers) sq_allocate_buffers() 422 sq->buffers = kmalloc (num * sizeof(char *), GFP_KERNEL); sq_allocate_buffers() 423 if (!sq->buffers) sq_allocate_buffers() 426 sq->buffers[i] = dmasound.mach.dma_alloc(size, GFP_KERNEL); sq_allocate_buffers() 427 if (!sq->buffers[i]) { sq_allocate_buffers() 429 dmasound.mach.dma_free(sq->buffers[i], size); sq_allocate_buffers() 430 kfree(sq->buffers); sq_allocate_buffers() 431 sq->buffers = NULL; sq_allocate_buffers() 442 if (sq->buffers) { sq_release_buffers() 444 dmasound.mach.dma_free(sq->buffers[i], sq->bufSize); sq_release_buffers() 445 kfree(sq->buffers); sq_release_buffers() 446 sq->buffers = NULL; sq_release_buffers() 609 dest = write_sq.buffers[write_sq.rear]; sq_write() 650 dest = write_sq.buffers[(write_sq.rear+1) % write_sq.max_count]; sq_write() 734 /* allocate the default number & size of buffers. sq_open2() 822 sound_silence(); /* this _must_ stop DMA, we might be about to lose the buffers */ sq_reset_output() 1007 read/write - the app doesn't care about our internal buffers. sq_ioctl() 1469 * buffers being other than powers of two. This is not checked yet. dmasound_setup() 1481 printk("dmasound_setup: invalid number of buffers, using default = %d\n", numWriteBufs); dmasound_setup()
|
/linux-4.1.27/sound/usb/misc/ |
H A D | ua101.c | 131 } buffers[MAX_MEMORY_BUFFERS]; member in struct:ua101::ua101_stream 1071 for (i = 0; i < ARRAY_SIZE(stream->buffers); ++i) { alloc_stream_buffers() 1074 stream->buffers[i].addr = alloc_stream_buffers() 1076 &stream->buffers[i].dma); alloc_stream_buffers() 1077 if (!stream->buffers[i].addr) alloc_stream_buffers() 1079 stream->buffers[i].size = size; alloc_stream_buffers() 1095 for (i = 0; i < ARRAY_SIZE(stream->buffers); ++i) free_stream_buffers() 1097 stream->buffers[i].size, free_stream_buffers() 1098 stream->buffers[i].addr, free_stream_buffers() 1099 stream->buffers[i].dma); free_stream_buffers() 1109 for (b = 0; b < ARRAY_SIZE(stream->buffers); ++b) { alloc_stream_urbs() 1110 unsigned int size = stream->buffers[b].size; alloc_stream_urbs() 1111 u8 *addr = stream->buffers[b].addr; alloc_stream_urbs() 1112 dma_addr_t dma = stream->buffers[b].dma; alloc_stream_urbs()
|
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge_mcp.h | 172 * only be done if the driver has already supplied buffers 227 /* data0 = number of big buffers to use. It must be 0 or a power of 2. 228 * 0 indicates that the NIC consumes as many buffers as they are required 231 * number of buffers for each big receive packet. 373 /* When set, small receive buffers can cross page boundaries. 374 * Both small and big receive buffers may start at any address.
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/pcie/ |
H A D | internal.h | 84 * @free_count: Number of pre-allocated buffers in rx_free 87 * @rx_used: List of Rx buffers with no SKB 154 * the software buffers (in the variables @meta, @txb in struct 155 * iwl_txq) only have 32 entries, while the HW buffers (@tfds in 208 * @scratchbufs: start of command headers, including scratch buffers, for 253 * @rx_replenish: work that will be called when buffers need to be allocated
|
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/ |
H A D | selftest.h | 282 /** total # request buffers */ 284 /** # posted request buffers */ 288 /** allocate more buffers if scd_buf_nposted < scd_buf_low */ 290 /** increase/decrease some buffers */ 292 /** posted message buffers */ 301 /* extra buffers for tolerating buggy peers, or unbalanced number
|
/linux-4.1.27/drivers/firmware/ |
H A D | qcom_scm.c | 82 * There can be arbitrary padding between the headers and buffers so 84 * to access the buffers in a safe manner. 112 * and response headers as well as the command and response buffers. 269 * Note that any buffers that are expected to be accessed by the secure world 272 * and response buffers is taken care of by qcom_scm_call; however, callers are 273 * responsible for any other cached buffers passed over to the secure world.
|
/linux-4.1.27/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem.c | 86 * allocated for buffers originally allocated from contiguous memory 125 /* To deal with userspace mmap'ings of 2d tiled buffers, which (a) are 201 * contiguous buffers can be remapped in TILER/DMM if they need to be 212 * shmem buffers that are mapped cached can simulate coherency via using 242 /* for non-cached buffers, ensure the new pages are clean because omap_gem_attach_pages() 280 /* for non-cached buffers, ensure the new pages are clean because omap_gem_detach_pages() 340 /* for tiled buffers, the virtual size has stride rounded up omap_gem_mmap_size() 365 /* Normal handling for the case of faulting in non-tiled buffers */ fault_1d() 391 /* Special handling for the case of faulting in 2d tiled buffers */ fault_2d() 577 * We do have some private objects, at least for scanout buffers omap_gem_mmap_obj() 809 * could potentially unpin and unmap buffers from TILER 842 * (only valid for tiled 2d buffers) 860 /* Get tiler stride for the buffer (only valid for 2d tiled buffers) */ omap_gem_tiled_stride() 1353 dev_err(dev->dev, "Tiled buffers require DMM\n"); omap_gem_new() 1357 /* tiled buffers are always shmem paged backed.. when they are omap_gem_new() 1362 /* currently don't allow cached buffers.. there is some caching omap_gem_new() 1391 * have DMM for remappign discontiguous buffers omap_gem_new()
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
H A D | ispstat.c | 44 * buffers the allocated memory should fit. For that case, a recover 49 * buffers making it extreamily unlikely to have an access to invalid memory 202 /* Don't select uninitialised buffers if it's not required */ __isp_stat_buf_find() 355 dev_dbg(stat->isp->dev, "%s: all buffers were freed.\n", isp_stat_bufs_free() 390 * in the DMA buffers being mapped through the ISP IOMMU. 394 * passes NULL instead, resulting in the buffers being mapped directly as 398 * buffers could be allocated with kmalloc/vmalloc, but we still use 412 /* Are the old buffers big enough? */ isp_stat_bufs_alloc() 608 * isp_stat_buf_process - Process statistic buffers. 701 * ignore next buffers even if it was disabled in the meantime. isp_stat_try_enable() 702 * On the other hand, Histogram shouldn't ignore buffers anymore isp_stat_try_enable() 729 * the correct memory address, 2 buffers must be ignored. omap3isp_stat_sbl_overflow() 847 * necessary to not inform too early the buffers aren't busy in case __stat_isr()
|
/linux-4.1.27/drivers/virtio/ |
H A D | virtio_ring.c | 69 /* Host supports indirect buffers */ 166 * buffers, then go indirect. FIXME: tune this threshold */ virtqueue_add() 203 /* We're about to use some buffers from the free list. */ virtqueue_add() 259 * virtqueue_add_sgs - expose buffers to other end 292 * virtqueue_add_outbuf - expose output buffers to other end 314 * virtqueue_add_inbuf - expose input buffers to other end 467 * Returns NULL if there are no used buffers, or the "data" token 485 pr_debug("No more buffers in queue\n"); virtqueue_get_buf() 576 * virtqueue_poll - query pending used buffers 580 * Returns "true" if there are pending used buffers in the queue. 598 * buffers in the queue, to detect a possible race between the driver 616 * interrupts until most of the available buffers have been processed; 617 * it returns "false" if there are many pending buffers in the queue,
|
/linux-4.1.27/drivers/mtd/nand/ |
H A D | mpc5121_nfc.c | 181 /* Send data from NFC buffers to NAND flash */ mpc5121_nfc_send_prog_page() 405 /* Copy data from/to NFC spare buffers. */ mpc5121_nfc_copy_spare() 414 * NAND spare area is available through NFC spare buffers. mpc5121_nfc_copy_spare() 457 /* Copy data from/to NFC main and spare buffers */ mpc5121_nfc_buf_copy() 497 /* Read data from NFC buffers */ mpc5121_nfc_read_buf() 503 /* Write data to NFC buffers */ mpc5121_nfc_write_buf() 510 /* Read byte from NFC buffers */ mpc5121_nfc_read_byte() 520 /* Read word from NFC buffers */ mpc5121_nfc_read_word()
|
/linux-4.1.27/drivers/net/ethernet/pasemi/ |
H A D | pasemi_mac.h | 51 u64 *buffers; /* RX interface buffer ring */ member in struct:pasemi_mac_rxring 114 #define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
|
/linux-4.1.27/drivers/media/pci/pt3/ |
H A D | pt3_dma.c | 122 /* mark the whole buffers as "not written yet" */ pt3_init_dmabuf() 177 /* build circular-linked pointers (xfer_desc) to the data buffers*/ pt3_alloc_dmabuf()
|
/linux-4.1.27/drivers/media/platform/ |
H A D | via-camera.h | 29 #define VCR_CI_3BUFS 0x00000004 /* 1 = 3 buffers, 0 = 2 buffers */
|
/linux-4.1.27/drivers/misc/mic/host/ |
H A D | mic_fops.c | 132 * We return POLLIN | POLLOUT from poll when new buffers are enqueued, and 133 * not when previously enqueued buffers may be available. This means that
|
/linux-4.1.27/drivers/staging/media/dt3155v4l/ |
H A D | dt3155v4l.h | 165 * @free_bufs_empty: no free image buffers 186 * @dmaq queue for dma buffers
|
/linux-4.1.27/drivers/pcmcia/ |
H A D | topic.h | 153 /* Disable write buffers to prevent lockups under load with numerous topic95_override() 162 dev_info(&socket->dev->dev, "Disabled ToPIC95 Cardbus write buffers.\n"); topic95_override()
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | radeon_cs.c | 56 /* Since buffers which appear sooner in the relocation list are radeon_cs_buckets_add() 57 * likely to be used more often than buffers which appear later radeon_cs_buckets_add() 58 * in the list, the sort mustn't change the ordering of buffers radeon_cs_buckets_add() 114 * Also, the buffers used for write have a higher priority than radeon_cs_parser_relocs() 115 * the buffers used for read only, which doubles the range radeon_cs_parser_relocs() 402 * which affects the order of buffers in the LRU list. radeon_cs_parser_fini() 403 * This assures that the smallest buffers are added first radeon_cs_parser_fini() 405 * first, instead of large buffers whose eviction is more radeon_cs_parser_fini()
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_execbuf_util.c | 84 * Reserve buffers for validation. 92 * buffers in different orders.
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_icreate_item.c | 100 * Because we have ordered buffers being tracked in the AIL for the inode 125 /* Ordered buffers do the dependency tracking here, so this does nothing. */
|
H A D | xfs_buf_item.c | 121 * Discontiguous buffers need a format structure per region that that is being 123 * they came from separate buffers, just like would occur if multiple buffers 125 * discontiguous buffers to be in-memory constructs, completely transparent to 173 * buffers we need to track which segment the dirty bits correspond to, xfs_buf_item_size() 336 * For icreate item based inode allocation, the buffers aren't written xfs_buf_item_format() 587 * (cancelled) buffers at unpin time, but we'll never go through the xfs_buf_item_unlock() 622 * Ordered buffers are dirty but may have no recorded changes, so ensure xfs_buf_item_unlock() 638 * Clean buffers, by definition, cannot be in the AIL. However, aborted xfs_buf_item_unlock() 639 * buffers may be dirty and hence in the AIL. Therefore if we are xfs_buf_item_unlock() 671 * The one exception to this is for buffers full of newly allocated 672 * inodes. These buffers are only relogged with the XFS_BLI_INODE_BUF 674 * inodes in the buffers will be replayed during recovery. If the 1044 * This is the iodone() function for buffers which have had callbacks 1126 * This is the iodone() function for buffers which have been 1146 * log-committed callbacks to unpin these buffers. Or we may never xfs_buf_iodone()
|
H A D | xfs_buf_item.h | 50 * needed to log buffers. It tracks how many times the lock has been
|
/linux-4.1.27/include/net/caif/ |
H A D | caif_hsi.h | 27 * Decides if HSI buffers should be prefilled with 0xFF pattern for easier 28 * debugging. Both TX and RX buffers will be filled before the transfer.
|
H A D | caif_spi.h | 49 /* Decides if SPI buffers should be prefilled with 0xFF pattern for easier 50 * debugging. Both TX and RX buffers will be filled before the transfer.
|
/linux-4.1.27/arch/s390/appldata/ |
H A D | appldata_mem.c | 56 u64 bufferram; /* memory reserved for buffers, free cache */ 57 u64 cached; /* size of (used) cache, w/o buffers */
|
/linux-4.1.27/drivers/media/pci/meye/ |
H A D | meye.h | 272 /* Maximum number of buffers */ 305 /* list of buffers */ 309 struct kfifo grabq; /* queue for buffers to be grabbed */ 311 struct kfifo doneq; /* queue for grabbed buffers */
|
/linux-4.1.27/drivers/media/platform/s3c-camif/ |
H A D | camif-core.h | 193 * @pending_buf_q: pending (empty) buffers queue head 194 * @active_buf_q: active (being written) buffers queue head 198 * @reqbufs_count: the number of buffers requested 320 * @list: list head for the buffers queue
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
H A D | sunvnet.h | 26 /* VNET packets are sent in buffers with the first 6 bytes skipped
|
/linux-4.1.27/drivers/staging/iio/adc/ |
H A D | ad7606.h | 77 * transfer buffers to live in their own cache lines.
|
/linux-4.1.27/drivers/staging/iio/frequency/ |
H A D | ad9834.h | 66 * transfer buffers to live in their own cache lines.
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | pci-nommu.c | 41 /* Map a set of buffers described by scatterlist in streaming
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | fixmap.h | 32 * these 'compile-time allocated' memory buffers are
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
H A D | qxl_prime.c | 28 * device that might share buffers with qxl */
|
/linux-4.1.27/drivers/char/xillybus/ |
H A D | xillybus_core.c | 81 * buffers' end_offset fields against changes made by IRQ handler (and in 323 struct xilly_buffer **buffers, xilly_get_dma_buffers() 331 if (buffers) { /* Not the message buffer */ xilly_get_dma_buffers() 383 if (buffers) { /* Not the message buffer */ xilly_get_dma_buffers() 386 buffers[i] = this_buffer++; xilly_get_dma_buffers() 486 struct xilly_buffer **buffers = NULL; xilly_setupchannels() local 515 buffers = devm_kcalloc(dev, bufnum, xilly_setupchannels() 518 if (!buffers) xilly_setupchannels() 531 channel->rd_buffers = buffers; xilly_setupchannels() 532 rc = xilly_get_dma_buffers(ep, &rd_alloc, buffers, xilly_setupchannels() 544 channel->wr_buffers = buffers; xilly_setupchannels() 545 rc = xilly_get_dma_buffers(ep, &wr_alloc, buffers, xilly_setupchannels() 846 /* Don't request more than all buffers */ xillybus_read() 1106 while (1) { /* Loop waiting for draining of buffers */ xillybus_myflush() 1608 * we have no leftover buffers pending before wrapping xillybus_release() 2023 /* Bootstrap phase III: Allocate buffers according to IDT */ xillybus_endpoint_discovery() 321 xilly_get_dma_buffers(struct xilly_endpoint *ep, struct xilly_alloc_state *s, struct xilly_buffer **buffers, int bufnum, int bytebufsize) xilly_get_dma_buffers() argument
|
/linux-4.1.27/arch/blackfin/lib/ |
H A D | ins.S | 71 * buffers in/out of FIFOs.
|
/linux-4.1.27/fs/ext2/ |
H A D | file.c | 84 "detected IO error when writing metadata buffers"); ext2_fsync()
|
/linux-4.1.27/fs/gfs2/ |
H A D | meta_io.c | 73 * The page and its buffers are protected by PageWriteback(), so we can gfs2_aspace_writepage() 298 * gfs2_meta_wipe - make inode's buffers so they aren't dirty/pinned anymore 299 * @ip: the inode who owns the buffers 301 * @blen: the number of buffers in the run
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 62 * buffers which the hardware consumes) and free list entries are 78 * Max number of Rx buffers we replenish at a time. Again keep this 79 * modest, allocating buffers isn't cheap either. 257 * Tests specified Free List to see whether the number of buffers 368 * free_tx_desc - reclaims TX descriptors and their buffers 372 * @unmap: whether the buffers should be unmapped for DMA 375 * TX buffers. Called with the TX queue lock held. 424 * @unmap: whether the buffers should be unmapped for DMA 427 * and frees the associated buffers if possible. Called with the TX 464 * free_rx_bufs - free RX buffers on an SGE Free List 466 * @fl: the SGE Free List to free buffers from 467 * @n: how many buffers to free 469 * Release the next @n buffers on an SGE Free List RX queue. The 470 * buffers must be made inaccessible to hardware before calling this 581 * Support for poisoning RX buffers ... 596 * @n: the number of new buffers to allocate 599 * (Re)populate an SGE free-buffer queue with up to @n new packet buffers, 603 * of buffers allocated. If afterwards the queue is found critically low, 617 * Sanity: ensure that the result of adding n Free List buffers refill_fl() 626 * If we support large pages, prefer large buffers and fail over to refill_fl() 708 * buffers, tell the hardware about them and return the number of refill_fl() 709 * buffers which we were able to allocate. refill_fl() 900 * Generates a Scatter/Gather List for the buffers that make up a packet. 1498 * buff! size, which is expected since buffers are at least t4vf_pktgl_to_skb() 1670 * restore_rx_bufs - put back a packet's RX buffers 1680 * We implement the suspension by putting all of the RX buffers associated 1681 * with the current packet back on the original Free List. The buffers 2033 * Normally new RX buffers are added when existing ones are consumed but 2045 * Lists in need of more free buffers. If we find one and it's not sge_rx_timer_cb() 2047 * to refill it. If we're successful in adding enough buffers to push sge_rx_timer_cb() 2633 /* A FL with <= fl_starve_thres buffers is starving and a periodic t4vf_sge_init()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_adminq.c | 127 * This assumes the posted send buffers have already been cleaned 139 * This assumes the posted receive buffers have already been cleaned 148 * i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue 159 * allocate the mapped buffers for the event processing i40e_alloc_arq_bufs() 169 /* allocate the mapped buffers */ i40e_alloc_arq_bufs() 231 /* allocate the mapped buffers */ i40e_alloc_asq_bufs() 346 /* Update tail in the HW to post pre-allocated buffers */ i40e_config_arq_regs() 396 /* allocate buffers in the rings */ i40e_init_asq() 455 /* allocate buffers in the rings */ i40e_init_arq() 500 /* free ring buffers */ i40e_shutdown_asq() 533 /* free ring buffers */ i40e_shutdown_arq() 926 * @e: event info from the receive descriptor, includes any buffers
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_adminq.c | 125 * This assumes the posted send buffers have already been cleaned 137 * This assumes the posted receive buffers have already been cleaned 146 * i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue 157 * allocate the mapped buffers for the event processing i40e_alloc_arq_bufs() 167 /* allocate the mapped buffers */ i40e_alloc_arq_bufs() 229 /* allocate the mapped buffers */ i40e_alloc_asq_bufs() 344 /* Update tail in the HW to post pre-allocated buffers */ i40e_config_arq_regs() 394 /* allocate buffers in the rings */ i40e_init_asq() 453 /* allocate buffers in the rings */ i40e_init_arq() 498 /* free ring buffers */ i40e_shutdown_asq() 531 /* free ring buffers */ i40e_shutdown_arq() 878 * @e: event info from the receive descriptor, includes any buffers
|