/linux-4.1.27/drivers/staging/lustre/lustre/include/ |
H A D | lustre_cfg.h | 124 static inline void lustre_cfg_bufs_set(struct lustre_cfg_bufs *bufs, lustre_cfg_bufs_set() argument 131 if (bufs == NULL) lustre_cfg_bufs_set() 134 if (bufs->lcfg_bufcount <= index) lustre_cfg_bufs_set() 135 bufs->lcfg_bufcount = index + 1; lustre_cfg_bufs_set() 137 bufs->lcfg_buf[index] = buf; lustre_cfg_bufs_set() 138 bufs->lcfg_buflen[index] = buflen; lustre_cfg_bufs_set() 141 static inline void lustre_cfg_bufs_set_string(struct lustre_cfg_bufs *bufs, lustre_cfg_bufs_set_string() argument 145 lustre_cfg_bufs_set(bufs, index, str, str ? strlen(str) + 1 : 0); lustre_cfg_bufs_set_string() 148 static inline void lustre_cfg_bufs_reset(struct lustre_cfg_bufs *bufs, char *name) lustre_cfg_bufs_reset() argument 150 memset((bufs), 0, sizeof(*bufs)); lustre_cfg_bufs_reset() 152 lustre_cfg_bufs_set_string(bufs, 0, name); lustre_cfg_bufs_reset() 173 static inline void lustre_cfg_bufs_init(struct lustre_cfg_bufs *bufs, lustre_cfg_bufs_init() argument 177 bufs->lcfg_bufcount = lcfg->lcfg_bufcount; lustre_cfg_bufs_init() 178 for (i = 0; i < bufs->lcfg_bufcount; i++) { lustre_cfg_bufs_init() 179 bufs->lcfg_buflen[i] = lcfg->lcfg_buflens[i]; lustre_cfg_bufs_init() 180 bufs->lcfg_buf[i] = lustre_cfg_buf(lcfg, i); lustre_cfg_bufs_init() 228 struct lustre_cfg_bufs *bufs) lustre_cfg_new() 234 OBD_ALLOC(lcfg, lustre_cfg_len(bufs->lcfg_bufcount, lustre_cfg_new() 235 bufs->lcfg_buflen)); lustre_cfg_new() 241 lcfg->lcfg_bufcount = bufs->lcfg_bufcount; lustre_cfg_new() 245 lcfg->lcfg_buflens[i] = bufs->lcfg_buflen[i]; lustre_cfg_new() 246 LOGL((char *)bufs->lcfg_buf[i], bufs->lcfg_buflen[i], ptr); lustre_cfg_new() 227 lustre_cfg_new(int cmd, struct lustre_cfg_bufs *bufs) lustre_cfg_new() argument
|
H A D | lustre_net.h | 2399 __u32 version, int opcode, char **bufs, 2403 char **bufs); 2406 int count, __u32 *lengths, char **bufs, 2575 char **bufs); 2577 __u32 *lens, char **bufs); 2579 char **bufs); 2581 __u32 *lens, char **bufs, int flags); 2584 char **bufs, int flags);
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_dma.c | 65 memset(&dev->dma->bufs[i], 0, sizeof(dev->dma->bufs[0])); drm_legacy_dma_setup() 93 if (dma->bufs[i].seg_count) { drm_legacy_dma_takedown() 97 dma->bufs[i].buf_count, drm_legacy_dma_takedown() 98 dma->bufs[i].seg_count); drm_legacy_dma_takedown() 99 for (j = 0; j < dma->bufs[i].seg_count; j++) { drm_legacy_dma_takedown() 100 if (dma->bufs[i].seglist[j]) { drm_legacy_dma_takedown() 101 drm_pci_free(dev, dma->bufs[i].seglist[j]); drm_legacy_dma_takedown() 104 kfree(dma->bufs[i].seglist); drm_legacy_dma_takedown() 106 if (dma->bufs[i].buf_count) { drm_legacy_dma_takedown() 107 for (j = 0; j < dma->bufs[i].buf_count; j++) { drm_legacy_dma_takedown() 108 kfree(dma->bufs[i].buflist[j].dev_private); drm_legacy_dma_takedown() 110 kfree(dma->bufs[i].buflist); drm_legacy_dma_takedown()
|
H A D | drm_info.c | 115 * Called when "/proc/dri/.../bufs" is read. 133 if (dma->bufs[i].buf_count) { drm_bufs_info() 134 seg_pages = dma->bufs[i].seg_count * (1 << dma->bufs[i].page_order); drm_bufs_info() 137 dma->bufs[i].buf_size, drm_bufs_info() 138 dma->bufs[i].buf_count, drm_bufs_info() 140 dma->bufs[i].seg_count, drm_bufs_info()
|
H A D | drm_bufs.c | 664 entry = &dma->bufs[order]; drm_legacy_addbufs_agp() 814 entry = &dma->bufs[order]; drm_legacy_addbufs_pci() 1025 entry = &dma->bufs[order]; drm_legacy_addbufs_sg() 1206 if (dma->bufs[i].buf_count) drm_legacy_infobufs() 1214 if (dma->bufs[i].buf_count) { drm_legacy_infobufs() 1217 struct drm_buf_entry *from = &dma->bufs[i]; drm_legacy_infobufs() 1234 dma->bufs[i].buf_count, drm_legacy_infobufs() 1235 dma->bufs[i].buf_size, drm_legacy_infobufs() 1236 dma->bufs[i].low_mark, drm_legacy_infobufs() 1237 dma->bufs[i].high_mark); drm_legacy_infobufs() 1257 * updates the respective drm_device_dma::bufs entry low and high water mark. 1283 entry = &dma->bufs[order]; drm_legacy_markbufs()
|
H A D | drm_debugfs.c | 51 {"bufs", drm_bufs_info, 0},
|
H A D | drm_prime.c | 748 * @drm_gem_prime_import to import dma-bufs.
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_rq.c | 37 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ(count), GFP_ATOMIC); vnic_rq_alloc_bufs() 38 if (!rq->bufs[i]) vnic_rq_alloc_bufs() 43 buf = rq->bufs[i]; vnic_rq_alloc_bufs() 49 buf->next = rq->bufs[0]; vnic_rq_alloc_bufs() 52 buf->next = rq->bufs[i + 1]; vnic_rq_alloc_bufs() 60 rq->to_use = rq->to_clean = rq->bufs[0]; vnic_rq_alloc_bufs() 75 if (rq->bufs[i]) { vnic_rq_free() 76 kfree(rq->bufs[i]); vnic_rq_free() 77 rq->bufs[i] = NULL; vnic_rq_free() 133 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES(count)] vnic_rq_init_start() 209 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES(count)] vnic_rq_clean()
|
H A D | vnic_wq.c | 37 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ(count), GFP_ATOMIC); vnic_wq_alloc_bufs() 38 if (!wq->bufs[i]) vnic_wq_alloc_bufs() 43 buf = wq->bufs[i]; vnic_wq_alloc_bufs() 49 buf->next = wq->bufs[0]; vnic_wq_alloc_bufs() 53 buf->next = wq->bufs[i + 1]; vnic_wq_alloc_bufs() 63 wq->to_use = wq->to_clean = wq->bufs[0]; vnic_wq_alloc_bufs() 78 if (wq->bufs[i]) { vnic_wq_free() 79 kfree(wq->bufs[i]); vnic_wq_free() 80 wq->bufs[i] = NULL; vnic_wq_free() 135 &wq->bufs[fetch_index / VNIC_WQ_BUF_BLK_ENTRIES(count)] vnic_wq_init_start() 191 wq->to_use = wq->to_clean = wq->bufs[0]; vnic_wq_clean()
|
H A D | vnic_wq.h | 85 struct vnic_wq_buf *bufs[VNIC_WQ_BUF_BLKS_MAX]; member in struct:vnic_wq
|
H A D | vnic_rq.h | 83 struct vnic_rq_buf *bufs[VNIC_RQ_BUF_BLKS_MAX]; member in struct:vnic_rq
|
H A D | enic_main.c | 547 * For all non EOP bufs, os_buf is NULL. enic_queue_wq_skb()
|
/linux-4.1.27/fs/afs/ |
H A D | netdevices.c | 40 int afs_get_ipv4_interfaces(struct afs_interface *bufs, size_t maxbufs, afs_get_ipv4_interfaces() argument 57 bufs[n].address.s_addr = ifa->ifa_address; for_primary_ifa() 58 bufs[n].netmask.s_addr = ifa->ifa_mask; for_primary_ifa() 59 bufs[n].mtu = dev->mtu; for_primary_ifa()
|
/linux-4.1.27/drivers/scsi/fnic/ |
H A D | vnic_rq.c | 37 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ, GFP_ATOMIC); vnic_rq_alloc_bufs() 38 if (!rq->bufs[i]) { vnic_rq_alloc_bufs() 45 buf = rq->bufs[i]; vnic_rq_alloc_bufs() 51 buf->next = rq->bufs[0]; vnic_rq_alloc_bufs() 54 buf->next = rq->bufs[i + 1]; vnic_rq_alloc_bufs() 62 rq->to_use = rq->to_clean = rq->bufs[0]; vnic_rq_alloc_bufs() 78 kfree(rq->bufs[i]); vnic_rq_free() 79 rq->bufs[i] = NULL; vnic_rq_free() 133 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES] vnic_rq_init() 189 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES] vnic_rq_clean()
|
H A D | vnic_wq.c | 37 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ, GFP_ATOMIC); vnic_wq_alloc_bufs() 38 if (!wq->bufs[i]) { vnic_wq_alloc_bufs() 45 buf = wq->bufs[i]; vnic_wq_alloc_bufs() 51 buf->next = wq->bufs[0]; vnic_wq_alloc_bufs() 54 buf->next = wq->bufs[i + 1]; vnic_wq_alloc_bufs() 62 wq->to_use = wq->to_clean = wq->bufs[0]; vnic_wq_alloc_bufs() 77 kfree(wq->bufs[i]); vnic_wq_free() 78 wq->bufs[i] = NULL; vnic_wq_free() 176 wq->to_use = wq->to_clean = wq->bufs[0]; vnic_wq_clean()
|
H A D | vnic_wq.h | 90 struct vnic_wq_buf *bufs[VNIC_WQ_BUF_BLKS_MAX]; member in struct:vnic_wq
|
H A D | vnic_rq.h | 97 struct vnic_rq_buf *bufs[VNIC_RQ_BUF_BLKS_MAX]; member in struct:vnic_rq
|
/linux-4.1.27/sound/core/oss/ |
H A D | io.c | 50 void **bufs = (void**)plugin->extra_data; io_playback_transfer() local 51 if (snd_BUG_ON(!bufs)) io_playback_transfer() 55 bufs[channel] = src_channels[channel].area.addr; io_playback_transfer() 57 bufs[channel] = NULL; io_playback_transfer() 59 return pcm_writev(plugin->plug, bufs, frames); io_playback_transfer() 76 void **bufs = (void**)plugin->extra_data; io_capture_transfer() local 77 if (snd_BUG_ON(!bufs)) io_capture_transfer() 81 bufs[channel] = dst_channels[channel].area.addr; io_capture_transfer() 83 bufs[channel] = NULL; io_capture_transfer() 85 return pcm_readv(plugin->plug, bufs, frames); io_capture_transfer()
|
H A D | pcm_plugin.h | 165 void **bufs, snd_pcm_uframes_t frames, 168 void **bufs, snd_pcm_uframes_t frames,
|
H A D | pcm_oss.c | 1258 snd_pcm_sframes_t snd_pcm_oss_writev3(struct snd_pcm_substream *substream, void **bufs, snd_pcm_uframes_t frames, int in_kernel) snd_pcm_oss_writev3() argument 1278 ret = snd_pcm_lib_writev(substream, (void __user **)bufs, frames); snd_pcm_oss_writev3() 1281 ret = snd_pcm_lib_writev(substream, (void __user **)bufs, frames); snd_pcm_oss_writev3() 1294 snd_pcm_sframes_t snd_pcm_oss_readv3(struct snd_pcm_substream *substream, void **bufs, snd_pcm_uframes_t frames, int in_kernel) snd_pcm_oss_readv3() argument 1318 ret = snd_pcm_lib_readv(substream, (void __user **)bufs, frames); snd_pcm_oss_readv3() 1321 ret = snd_pcm_lib_readv(substream, (void __user **)bufs, frames); snd_pcm_oss_readv3()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf-core.c | 204 if (NULL == q->bufs[i]) videobuf_queue_is_busy() 206 if (q->bufs[i]->map) { videobuf_queue_is_busy() 210 if (q->bufs[i]->state == VIDEOBUF_QUEUED) { videobuf_queue_is_busy() 214 if (q->bufs[i]->state == VIDEOBUF_ACTIVE) { videobuf_queue_is_busy() 246 if (q->bufs[i] && q->bufs[i]->map) { __videobuf_free() 252 if (NULL == q->bufs[i]) __videobuf_free() 254 q->ops->buf_release(q, q->bufs[i]); __videobuf_free() 255 kfree(q->bufs[i]); __videobuf_free() 256 q->bufs[i] = NULL; __videobuf_free() 275 if (NULL == q->bufs[i]) videobuf_queue_cancel() 277 if (q->bufs[i]->state == VIDEOBUF_QUEUED) { videobuf_queue_cancel() 278 list_del(&q->bufs[i]->queue); videobuf_queue_cancel() 279 q->bufs[i]->state = VIDEOBUF_ERROR; videobuf_queue_cancel() 280 wake_up_all(&q->bufs[i]->done); videobuf_queue_cancel() 287 if (NULL == q->bufs[i]) videobuf_queue_cancel() 289 q->ops->buf_release(q, q->bufs[i]); videobuf_queue_cancel() 399 q->bufs[i] = videobuf_alloc_vb(q); __videobuf_mmap_setup() 401 if (NULL == q->bufs[i]) __videobuf_mmap_setup() 404 q->bufs[i]->i = i; __videobuf_mmap_setup() 405 q->bufs[i]->memory = memory; __videobuf_mmap_setup() 406 q->bufs[i]->bsize = bsize; __videobuf_mmap_setup() 409 q->bufs[i]->boff = PAGE_ALIGN(bsize) * i; __videobuf_mmap_setup() 482 dprintk(1, "reqbufs: bufs=%d, size=0x%x [%u pages total]\n", videobuf_reqbufs() 514 if (unlikely(NULL == q->bufs[b->index])) { videobuf_querybuf() 519 videobuf_status(q, b, q->bufs[b->index], q->type); videobuf_querybuf() 555 buf = q->bufs[b->index]; videobuf_qbuf() 983 err = q->ops->buf_prepare(q, q->bufs[i], field); __videobuf_read_start() 986 list_add_tail(&q->bufs[i]->stream, &q->stream); __videobuf_read_start() 990 q->ops->buf_queue(q, q->bufs[i]); __videobuf_read_start() 1004 if (NULL == q->bufs[i]) __videobuf_read_stop() 1006 kfree(q->bufs[i]); __videobuf_read_stop() 1007 q->bufs[i] = NULL; __videobuf_read_stop() 1184 struct videobuf_buffer *buf = q->bufs[i]; videobuf_mmap_mapper()
|
H A D | videobuf-dma-sg.c | 409 if (NULL == q->bufs[i]) videobuf_vm_close() 411 mem = q->bufs[i]->priv; videobuf_vm_close() 417 if (q->bufs[i]->map != map) videobuf_vm_close() 419 q->bufs[i]->map = NULL; videobuf_vm_close() 420 q->bufs[i]->baddr = 0; videobuf_vm_close() 421 q->ops->buf_release(q, q->bufs[i]); videobuf_vm_close() 603 if (buf == q->bufs[first]) { __videobuf_mmap_mapper() 604 size = PAGE_ALIGN(q->bufs[first]->bsize); __videobuf_mmap_mapper() 626 if (NULL == q->bufs[i]) __videobuf_mmap_mapper() 628 q->bufs[i]->map = map; __videobuf_mmap_mapper() 629 q->bufs[i]->baddr = vma->vm_start + size; __videobuf_mmap_mapper() 630 size += PAGE_ALIGN(q->bufs[i]->bsize); __videobuf_mmap_mapper() 639 dprintk(1, "mmap %p: q=%p %08lx-%08lx pgoff %08lx bufs %d-%d\n", __videobuf_mmap_mapper()
|
H A D | videobuf-vmalloc.c | 85 if (NULL == q->bufs[i]) videobuf_vm_close() 88 if (q->bufs[i]->map != map) videobuf_vm_close() 91 mem = q->bufs[i]->priv; videobuf_vm_close() 111 q->bufs[i]->map = NULL; videobuf_vm_close() 112 q->bufs[i]->baddr = 0; videobuf_vm_close()
|
H A D | videobuf-dma-contig.c | 97 if (NULL == q->bufs[i]) videobuf_vm_close() 100 if (q->bufs[i]->map != map) videobuf_vm_close() 103 mem = q->bufs[i]->priv; videobuf_vm_close() 123 q->bufs[i]->map = NULL; videobuf_vm_close() 124 q->bufs[i]->baddr = 0; videobuf_vm_close()
|
H A D | videobuf2-core.c | 294 vb = q->bufs[buffer]; __setup_lengths() 315 vb = q->bufs[q->num_buffers - 1]; __setup_offsets() 323 vb = q->bufs[buffer]; __setup_offsets() 396 q->bufs[q->num_buffers + buffer] = vb; __vb2_queue_alloc() 419 vb = q->bufs[buffer]; __vb2_free_mem() 452 if (q->bufs[buffer] == NULL) __vb2_queue_free() 454 if (q->bufs[buffer]->state == VB2_BUF_STATE_PREPARING) { __vb2_queue_free() 463 struct vb2_buffer *vb = q->bufs[buffer]; __vb2_queue_free() 498 struct vb2_buffer *vb = q->bufs[buffer]; __vb2_queue_free() 537 kfree(q->bufs[buffer]); __vb2_queue_free() 538 q->bufs[buffer] = NULL; __vb2_queue_free() 644 if (__buffer_in_use(q, q->bufs[buffer])) __buffers_in_use() 752 vb = q->bufs[b->index]; vb2_querybuf() 1705 if (q->bufs[b->index] == NULL) { vb2_queue_or_prepare_buf() 1716 return __verify_planes_array(q->bufs[b->index], b); vb2_queue_or_prepare_buf() 1748 vb = q->bufs[b->index]; vb2_prepare_buf() 1813 vb = q->bufs[i]; vb2_start_streaming() 1837 vb = q->bufs[b->index]; vb2_internal_qbuf() 2181 if (q->bufs[i]->state == VB2_BUF_STATE_ACTIVE) __vb2_queue_cancel() 2182 vb2_buffer_done(q->bufs[i], VB2_BUF_STATE_ERROR); __vb2_queue_cancel() 2214 struct vb2_buffer *vb = q->bufs[i]; __vb2_queue_cancel() 2374 vb = q->bufs[buffer]; __find_plane_by_offset() 2429 vb = q->bufs[eb->index]; vb2_expbuf() 2528 vb = q->bufs[buffer]; vb2_mmap() 2578 vb = q->bufs[buffer]; vb2_get_unmapped_area() 2798 struct vb2_fileio_buf bufs[VIDEO_MAX_FRAME]; member in struct:vb2_fileio_data 2869 if (q->bufs[0]->num_planes != 1) { __vb2_init_fileio() 2878 fileio->bufs[i].vaddr = vb2_plane_vaddr(q->bufs[i], 0); __vb2_init_fileio() 2879 if (fileio->bufs[i].vaddr == NULL) { __vb2_init_fileio() 2883 fileio->bufs[i].size = vb2_plane_size(q->bufs[i], 0); __vb2_init_fileio() 2910 fileio->bufs[i].queued = 1; __vb2_init_fileio() 3024 buf = &fileio->bufs[index]; __vb2_perform_fileio() 3031 buf->size = read ? vb2_get_plane_payload(q->bufs[index], 0) __vb2_perform_fileio() 3032 : vb2_plane_size(q->bufs[index], 0); __vb2_perform_fileio() 3040 buf = &fileio->bufs[index]; __vb2_perform_fileio() 3109 buf->size = vb2_plane_size(q->bufs[index], 0); __vb2_perform_fileio() 3199 vb = q->bufs[fileio->b.index]; vb2_thread()
|
/linux-4.1.27/drivers/scsi/arm/ |
H A D | scsi.h | 29 int bufs = SCp->buffers_residual; copy_SCp_to_sg() local 34 BUG_ON(bufs + 1 > max); copy_SCp_to_sg() 38 if (bufs) { copy_SCp_to_sg() 42 for_each_sg(sg_next(SCp->buffer), src_sg, bufs, i) copy_SCp_to_sg() 47 return bufs + 1; copy_SCp_to_sg()
|
H A D | powertec.c | 140 int bufs, map_dir, dma_dir; powertecscsi_dma_setup() local 142 bufs = copy_SCp_to_sg(&info->sg[0], SCp, NR_SG); powertecscsi_dma_setup() 151 dma_map_sg(dev, info->sg, bufs, map_dir); powertecscsi_dma_setup() 154 set_dma_sg(dmach, info->sg, bufs); powertecscsi_dma_setup()
|
H A D | cumana_2.c | 168 int bufs, map_dir, dma_dir, alatch_dir; cumanascsi_2_dma_setup() local 170 bufs = copy_SCp_to_sg(&info->sg[0], SCp, NR_SG); cumanascsi_2_dma_setup() 181 dma_map_sg(dev, info->sg, bufs, map_dir); cumanascsi_2_dma_setup() 184 set_dma_sg(dmach, info->sg, bufs); cumanascsi_2_dma_setup()
|
H A D | eesox.c | 167 int bufs, map_dir, dma_dir; eesoxscsi_dma_setup() local 169 bufs = copy_SCp_to_sg(&info->sg[0], SCp, NR_SG); eesoxscsi_dma_setup() 178 dma_map_sg(dev, info->sg, bufs, map_dir); eesoxscsi_dma_setup() 181 set_dma_sg(dmach, info->sg, bufs); eesoxscsi_dma_setup()
|
/linux-4.1.27/drivers/crypto/caam/ |
H A D | caamrng.c | 76 struct buf_data bufs[2]; member in struct:caam_rng_ctx 95 rng_unmap_buf(jrdev, &ctx->bufs[0]); rng_unmap_ctx() 96 rng_unmap_buf(jrdev, &ctx->bufs[1]); rng_unmap_ctx() 119 struct buf_data *bd = &ctx->bufs[!(to_current ^ ctx->current_buf)]; submit_job() 138 struct buf_data *bd = &ctx->bufs[ctx->current_buf]; caam_read() 220 struct buf_data *bd = &ctx->bufs[buf_id]; rng_create_job_desc() 247 bd = &rng_ctx->bufs[i]; caam_cleanup() 257 struct buf_data *bd = &ctx->bufs[buf_id]; caam_init_buf()
|
/linux-4.1.27/fs/ |
H A D | pipe.c | 245 int bufs = pipe->nrbufs; pipe_read() local 246 if (bufs) { pipe_read() 248 struct pipe_buffer *buf = pipe->bufs + curbuf; pipe_read() 285 pipe->nrbufs = --bufs; pipe_read() 292 if (bufs) /* More to do? */ pipe_read() 364 struct pipe_buffer *buf = pipe->bufs + lastbuf; pipe_write() 387 int bufs; pipe_write() local 395 bufs = pipe->nrbufs; pipe_write() 396 if (bufs < pipe->buffers) { pipe_write() 397 int newbuf = (pipe->curbuf + bufs) & (pipe->buffers-1); pipe_write() 398 struct pipe_buffer *buf = pipe->bufs + newbuf; pipe_write() 434 pipe->nrbufs = ++bufs; pipe_write() 440 if (bufs < pipe->buffers) pipe_write() 488 count += pipe->bufs[buf].len; pipe_ioctl() 593 pipe->bufs = kzalloc(sizeof(struct pipe_buffer) * PIPE_DEF_BUFFERS, GFP_KERNEL); alloc_pipe_info() 594 if (pipe->bufs) { alloc_pipe_info() 612 struct pipe_buffer *buf = pipe->bufs + i; free_pipe_info() 618 kfree(pipe->bufs); free_pipe_info() 963 struct pipe_buffer *bufs; pipe_set_size() local 974 bufs = kcalloc(nr_pages, sizeof(*bufs), GFP_KERNEL | __GFP_NOWARN); pipe_set_size() 975 if (unlikely(!bufs)) pipe_set_size() 994 memcpy(bufs, pipe->bufs + pipe->curbuf, head * sizeof(struct pipe_buffer)); pipe_set_size() 996 memcpy(bufs + head, pipe->bufs, tail * sizeof(struct pipe_buffer)); pipe_set_size() 1000 kfree(pipe->bufs); pipe_set_size() 1001 pipe->bufs = bufs; pipe_set_size()
|
H A D | splice.c | 207 struct pipe_buffer *buf = pipe->bufs + newbuf; splice_to_pipe() 760 struct pipe_buffer *buf = pipe->bufs + pipe->curbuf; splice_from_pipe_feed() 990 struct pipe_buffer *buf = pipe->bufs + idx; iter_file_splice_write() 1024 struct pipe_buffer *buf = pipe->bufs + pipe->curbuf; iter_file_splice_write() 1264 struct pipe_buffer *buf = pipe->bufs + i; splice_direct_to_actor() 1850 ibuf = ipipe->bufs + ipipe->curbuf; splice_pipe_to_pipe() 1852 obuf = opipe->bufs + nbuf; splice_pipe_to_pipe() 1934 ibuf = ipipe->bufs + ((ipipe->curbuf + i) & (ipipe->buffers-1)); link_pipe() 1943 obuf = opipe->bufs + nbuf; link_pipe()
|
/linux-4.1.27/tools/virtio/ |
H A D | virtio_test.c | 154 bool delayed, int bufs) run_test() 168 if (started < bufs) { run_test() 181 /* Flush out completed bufs if any */ run_test() 190 assert(completed <= bufs); run_test() 191 assert(started <= bufs); run_test() 192 if (completed == bufs) run_test() 153 run_test(struct vdev_info *dev, struct vq_info *vq, bool delayed, int bufs) run_test() argument
|
H A D | vringh_test.c | 331 /* Consume bufs. */ parallel_test() 408 /* Consume bufs. */ parallel_test()
|
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/ |
H A D | obd_mount.c | 79 struct lustre_cfg_bufs *bufs; lustre_process_log() local 87 OBD_ALLOC_PTR(bufs); lustre_process_log() 88 if (bufs == NULL) lustre_process_log() 92 lustre_cfg_bufs_reset(bufs, mgc->obd_name); lustre_process_log() 93 lustre_cfg_bufs_set_string(bufs, 1, logname); lustre_process_log() 94 lustre_cfg_bufs_set(bufs, 2, cfg, sizeof(*cfg)); lustre_process_log() 95 lustre_cfg_bufs_set(bufs, 3, &sb, sizeof(sb)); lustre_process_log() 96 lcfg = lustre_cfg_new(LCFG_LOG_START, bufs); lustre_process_log() 100 OBD_FREE_PTR(bufs); lustre_process_log() 121 struct lustre_cfg_bufs bufs; lustre_end_log() local 130 lustre_cfg_bufs_reset(&bufs, mgc->obd_name); lustre_end_log() 131 lustre_cfg_bufs_set_string(&bufs, 1, logname); lustre_end_log() 133 lustre_cfg_bufs_set(&bufs, 2, cfg, sizeof(*cfg)); lustre_end_log() 134 lcfg = lustre_cfg_new(LCFG_LOG_END, &bufs); lustre_end_log() 149 struct lustre_cfg_bufs bufs; do_lcfg() local 156 lustre_cfg_bufs_reset(&bufs, cfgname); do_lcfg() 158 lustre_cfg_bufs_set_string(&bufs, 1, s1); do_lcfg() 160 lustre_cfg_bufs_set_string(&bufs, 2, s2); do_lcfg() 162 lustre_cfg_bufs_set_string(&bufs, 3, s3); do_lcfg() 164 lustre_cfg_bufs_set_string(&bufs, 4, s4); do_lcfg() 166 lcfg = lustre_cfg_new(cmd, &bufs); do_lcfg()
|
H A D | obd_config.c | 991 struct lustre_cfg_bufs *bufs = NULL; lustre_cfg_rename() local 1022 OBD_ALLOC_PTR(bufs); lustre_cfg_rename() 1023 if (bufs == NULL) { lustre_cfg_rename() 1028 lustre_cfg_bufs_reset(bufs, NULL); lustre_cfg_rename() 1029 lustre_cfg_bufs_init(bufs, cfg); lustre_cfg_rename() 1030 lustre_cfg_bufs_set_string(bufs, 1, new_param); lustre_cfg_rename() 1032 new_cfg = lustre_cfg_new(cfg->lcfg_command, bufs); lustre_cfg_rename() 1035 OBD_FREE_PTR(bufs); lustre_cfg_rename() 1400 struct lustre_cfg_bufs bufs; class_config_llog_handler() local 1489 lustre_cfg_bufs_init(&bufs, lcfg); class_config_llog_handler() 1504 lustre_cfg_bufs_set_string(&bufs, 0, inst_name); class_config_llog_handler() 1513 lustre_cfg_bufs_set_string(&bufs, 2, class_config_llog_handler() 1525 lustre_cfg_bufs_set(&bufs, 2, bufs.lcfg_buf[1], class_config_llog_handler() 1526 bufs.lcfg_buflen[1]); class_config_llog_handler() 1527 lustre_cfg_bufs_set(&bufs, 1, bufs.lcfg_buf[0], class_config_llog_handler() 1528 bufs.lcfg_buflen[0]); class_config_llog_handler() 1529 lustre_cfg_bufs_set_string(&bufs, 0, class_config_llog_handler() 1533 lcfg_new = lustre_cfg_new(lcfg->lcfg_command, &bufs); class_config_llog_handler() 1722 struct lustre_cfg_bufs bufs; class_manual_cleanup() local 1738 lustre_cfg_bufs_reset(&bufs, obd->obd_name); class_manual_cleanup() 1739 lustre_cfg_bufs_set_string(&bufs, 1, flags); class_manual_cleanup() 1740 lcfg = lustre_cfg_new(LCFG_CLEANUP, &bufs); class_manual_cleanup()
|
/linux-4.1.27/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 57 struct alx_buffer *txb = &alx->txq.bufs[entry]; alx_free_txbuf() 84 cur_buf = &rxq->bufs[cur]; alx_refill_rx_ring() 116 cur_buf = &rxq->bufs[cur]; alx_refill_rx_ring() 153 skb = txq->bufs[sw_read_idx].skb; alx_clean_tx_irq() 210 rxb = &rxq->bufs[rxq->read_idx]; alx_clean_rx_irq() 391 if (!txq->bufs) alx_free_txring_buf() 397 memset(txq->bufs, 0, alx->tx_ringsz * sizeof(struct alx_buffer)); alx_free_txring_buf() 415 cur_buf = rxq->bufs + i; alx_free_rxring_buf() 512 alx->txq.bufs = kcalloc(alx->tx_ringsz, alx_alloc_descriptors() 515 if (!alx->txq.bufs) alx_alloc_descriptors() 518 alx->rxq.bufs = kcalloc(alx->rx_ringsz, alx_alloc_descriptors() 521 if (!alx->rxq.bufs) alx_alloc_descriptors() 565 kfree(alx->txq.bufs); alx_alloc_descriptors() 566 kfree(alx->rxq.bufs); alx_alloc_descriptors() 593 kfree(alx->txq.bufs); alx_free_rings() 594 kfree(alx->rxq.bufs); alx_free_rings() 1020 dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen); alx_map_tx_skb() 1021 dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma); alx_map_tx_skb() 1042 dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen); alx_map_tx_skb() 1043 dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma); alx_map_tx_skb() 1051 txq->bufs[txq->write_idx].skb = skb; alx_map_tx_skb()
|
H A D | alx.h | 59 struct alx_buffer *bufs; member in struct:alx_rx_queue 69 struct alx_buffer *bufs; member in struct:alx_tx_queue
|
/linux-4.1.27/fs/jbd/ |
H A D | commit.c | 158 static void journal_do_submit_data(struct buffer_head **wbuf, int bufs, journal_do_submit_data() argument 163 for (i = 0; i < bufs; i++) { journal_do_submit_data() 189 int bufs = 0; journal_submit_data_buffers() local 225 journal_do_submit_data(wbuf, bufs, write_op); journal_submit_data_buffers() 226 bufs = 0; journal_submit_data_buffers() 250 wbuf[bufs++] = bh; journal_submit_data_buffers() 254 if (bufs == journal->j_wbufsize) { journal_submit_data_buffers() 258 journal_do_submit_data(wbuf, bufs, write_op); journal_submit_data_buffers() 259 bufs = 0; journal_submit_data_buffers() 285 journal_do_submit_data(wbuf, bufs, write_op); journal_submit_data_buffers() 301 int bufs; journal_commit_transaction() local 534 bufs = 0; journal_commit_transaction() 563 J_ASSERT (bufs == 0); journal_commit_transaction() 586 wbuf[bufs++] = bh; journal_commit_transaction() 633 wbuf[bufs++] = jh2bh(new_jh); journal_commit_transaction() 660 if (bufs == journal->j_wbufsize || journal_commit_transaction() 664 jbd_debug(4, "JBD: Submit %d IOs\n", bufs); journal_commit_transaction() 673 for (i = 0; i < bufs; i++) { journal_commit_transaction() 696 bufs = 0; journal_commit_transaction()
|
H A D | recovery.c | 76 struct buffer_head * bufs[MAXBUF]; do_readahead() local 104 bufs[nbufs++] = bh; do_readahead() 106 ll_rw_block(READ, nbufs, bufs); do_readahead() 107 journal_brelse_array(bufs, nbufs); do_readahead() 115 ll_rw_block(READ, nbufs, bufs); do_readahead() 120 journal_brelse_array(bufs, nbufs); do_readahead()
|
/linux-4.1.27/sound/core/ |
H A D | pcm_compat.c | 403 /* snd_xfern needs remapping of bufs */ 406 u32 bufs; /* this is void **; */ member in struct:snd_xfern32 422 void __user **bufs; snd_pcm_ioctl_xfern_compat() local 432 if (get_user(buf, &data32->bufs) || snd_pcm_ioctl_xfern_compat() 436 bufs = kmalloc(sizeof(void __user *) * ch, GFP_KERNEL); snd_pcm_ioctl_xfern_compat() 437 if (bufs == NULL) snd_pcm_ioctl_xfern_compat() 442 kfree(bufs); snd_pcm_ioctl_xfern_compat() 445 bufs[i] = compat_ptr(ptr); snd_pcm_ioctl_xfern_compat() 449 err = snd_pcm_lib_writev(substream, bufs, frames); snd_pcm_ioctl_xfern_compat() 451 err = snd_pcm_lib_readv(substream, bufs, frames); snd_pcm_ioctl_xfern_compat() 456 kfree(bufs); snd_pcm_ioctl_xfern_compat()
|
H A D | pcm_native.c | 2851 void __user **bufs; snd_pcm_playback_ioctl1() local 2862 bufs = memdup_user(xfern.bufs, snd_pcm_playback_ioctl1() 2864 if (IS_ERR(bufs)) snd_pcm_playback_ioctl1() 2865 return PTR_ERR(bufs); snd_pcm_playback_ioctl1() 2866 result = snd_pcm_lib_writev(substream, bufs, xfern.frames); snd_pcm_playback_ioctl1() 2867 kfree(bufs); snd_pcm_playback_ioctl1() 2931 void *bufs; snd_pcm_capture_ioctl1() local 2942 bufs = memdup_user(xfern.bufs, snd_pcm_capture_ioctl1() 2944 if (IS_ERR(bufs)) snd_pcm_capture_ioctl1() 2945 return PTR_ERR(bufs); snd_pcm_capture_ioctl1() 2946 result = snd_pcm_lib_readv(substream, bufs, xfern.frames); snd_pcm_capture_ioctl1() 2947 kfree(bufs); snd_pcm_capture_ioctl1() 3090 void __user **bufs; snd_pcm_readv() local 3107 bufs = kmalloc(sizeof(void *) * to->nr_segs, GFP_KERNEL); snd_pcm_readv() 3108 if (bufs == NULL) snd_pcm_readv() 3111 bufs[i] = to->iov[i].iov_base; snd_pcm_readv() 3112 result = snd_pcm_lib_readv(substream, bufs, frames); snd_pcm_readv() 3115 kfree(bufs); snd_pcm_readv() 3126 void __user **bufs; snd_pcm_writev() local 3142 bufs = kmalloc(sizeof(void *) * from->nr_segs, GFP_KERNEL); snd_pcm_writev() 3143 if (bufs == NULL) snd_pcm_writev() 3146 bufs[i] = from->iov[i].iov_base; snd_pcm_writev() 3147 result = snd_pcm_lib_writev(substream, bufs, frames); snd_pcm_writev() 3150 kfree(bufs); snd_pcm_writev()
|
H A D | pcm_lib.c | 2156 void __user **bufs = (void __user **)data; snd_pcm_lib_writev_transfer() local 2162 for (c = 0; c < channels; ++c, ++bufs) { snd_pcm_lib_writev_transfer() 2163 if (*bufs == NULL) { snd_pcm_lib_writev_transfer() 2167 char __user *buf = *bufs + samples_to_bytes(runtime, off); snd_pcm_lib_writev_transfer() 2175 for (c = 0; c < channels; ++c, ++bufs) { snd_pcm_lib_writev_transfer() 2177 if (*bufs == NULL) { snd_pcm_lib_writev_transfer() 2180 char __user *buf = *bufs + samples_to_bytes(runtime, off); snd_pcm_lib_writev_transfer() 2190 void __user **bufs, snd_pcm_lib_writev() 2205 return snd_pcm_lib_write1(substream, (unsigned long)bufs, frames, snd_pcm_lib_writev() 2365 void __user **bufs = (void __user **)data; snd_pcm_lib_readv_transfer() local 2369 for (c = 0; c < channels; ++c, ++bufs) { snd_pcm_lib_readv_transfer() 2371 if (*bufs == NULL) snd_pcm_lib_readv_transfer() 2373 buf = *bufs + samples_to_bytes(runtime, off); snd_pcm_lib_readv_transfer() 2379 for (c = 0; c < channels; ++c, ++bufs) { snd_pcm_lib_readv_transfer() 2382 if (*bufs == NULL) snd_pcm_lib_readv_transfer() 2386 buf = *bufs + samples_to_bytes(runtime, off); snd_pcm_lib_readv_transfer() 2395 void __user **bufs, snd_pcm_lib_readv() 2412 return snd_pcm_lib_read1(substream, (unsigned long)bufs, frames, nonblock, snd_pcm_lib_readv_transfer); snd_pcm_lib_readv() 2189 snd_pcm_lib_writev(struct snd_pcm_substream *substream, void __user **bufs, snd_pcm_uframes_t frames) snd_pcm_lib_writev() argument 2394 snd_pcm_lib_readv(struct snd_pcm_substream *substream, void __user **bufs, snd_pcm_uframes_t frames) snd_pcm_lib_readv() argument
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
H A D | qat_algs.c | 664 int i, bufs = 0, sg_nctr = 0; qat_alg_sgl_to_bufl() local 689 bufl->bufers[bufs].addr = dma_map_single(dev, for_each_sg() 693 bufl->bufers[bufs].len = sg->length; for_each_sg() 694 if (unlikely(dma_mapping_error(dev, bufl->bufers[bufs].addr))) for_each_sg() 696 bufs++; for_each_sg() 699 bufl->bufers[bufs].addr = dma_map_single(dev, iv, ivlen, 701 bufl->bufers[bufs].len = ivlen; 702 if (unlikely(dma_mapping_error(dev, bufl->bufers[bufs].addr))) 704 bufs++; 708 int y = sg_nctr + bufs; for_each_sg() 721 bufl->num_bufs = sg_nctr + bufs; 743 for (i = 0; i < bufs; i++) { 748 int y = sg_nctr + bufs; for_each_sg() 761 buflout->num_bufs = sg_nctr + bufs; 775 for (i = 0; i < n + bufs; i++) 786 for (i = bufs; i < n + bufs; i++)
|
/linux-4.1.27/drivers/block/aoe/ |
H A D | aoemain.c | 67 aoeblk_exit(); /* free cache after de-allocating bufs */ aoe_exit()
|
/linux-4.1.27/fs/jbd2/ |
H A D | commit.c | 377 int bufs; jbd2_journal_commit_transaction() local 585 bufs = 0; jbd2_journal_commit_transaction() 617 J_ASSERT (bufs == 0); jbd2_journal_commit_transaction() 641 wbuf[bufs++] = descriptor; jbd2_journal_commit_transaction() 679 jh, &wbuf[bufs], blocknr); jbd2_journal_commit_transaction() 684 jbd2_file_log_bh(&io_bufs, wbuf[bufs]); jbd2_journal_commit_transaction() 698 jbd2_block_tag_csum_set(journal, tag, wbuf[bufs], jbd2_journal_commit_transaction() 702 bufs++; jbd2_journal_commit_transaction() 714 if (bufs == journal->j_wbufsize || jbd2_journal_commit_transaction() 718 jbd_debug(4, "JBD2: Submit %d IOs\n", bufs); jbd2_journal_commit_transaction() 728 for (i = 0; i < bufs; i++) { jbd2_journal_commit_transaction() 746 stats.run.rs_blocks_logged += bufs; jbd2_journal_commit_transaction() 751 bufs = 0; jbd2_journal_commit_transaction()
|
H A D | recovery.c | 77 struct buffer_head * bufs[MAXBUF]; do_readahead() local 105 bufs[nbufs++] = bh; do_readahead() 107 ll_rw_block(READ, nbufs, bufs); do_readahead() 108 journal_brelse_array(bufs, nbufs); do_readahead() 116 ll_rw_block(READ, nbufs, bufs); do_readahead() 121 journal_brelse_array(bufs, nbufs); do_readahead()
|
/linux-4.1.27/include/linux/ |
H A D | pipe_fs_i.h | 44 * @bufs: the circular array of pipe buffers 59 struct pipe_buffer *bufs; member in struct:pipe_inode_info
|
/linux-4.1.27/drivers/media/pci/pt1/ |
H A D | va1j5jf8007t.c | 467 const u8 (*bufs)[2]; va1j5jf8007t_prepare() 477 bufs = va1j5jf8007t_20mhz_prepare_bufs; va1j5jf8007t_prepare() 481 bufs = va1j5jf8007t_25mhz_prepare_bufs; va1j5jf8007t_prepare() 494 memcpy(buf, bufs[i], sizeof(buf)); va1j5jf8007t_prepare()
|
H A D | va1j5jf8007s.c | 648 const u8 (*bufs)[2]; va1j5jf8007s_prepare_2() 657 bufs = va1j5jf8007s_20mhz_prepare_bufs; va1j5jf8007s_prepare_2() 661 bufs = va1j5jf8007s_25mhz_prepare_bufs; va1j5jf8007s_prepare_2() 675 memcpy(buf, bufs[i], sizeof(buf)); va1j5jf8007s_prepare_2()
|
H A D | pt1.c | 66 struct pt1_buffer bufs[PT1_NR_BUFS]; member in struct:pt1_table 325 page = pt1->tables[pt1->table_index].bufs[pt1->buf_index].page; pt1_thread() 392 pt1_cleanup_buffer(pt1, &table->bufs[i]); pt1_cleanup_table() 410 ret = pt1_init_buffer(pt1, &table->bufs[i], &buf_pfn); pt1_init_table() 424 pt1_cleanup_buffer(pt1, &table->bufs[i]); pt1_init_table()
|
/linux-4.1.27/fs/fuse/ |
H A D | dev.c | 1381 struct pipe_buffer *bufs; fuse_dev_splice_read() local 1387 bufs = kmalloc(pipe->buffers * sizeof(struct pipe_buffer), GFP_KERNEL); fuse_dev_splice_read() 1388 if (!bufs) fuse_dev_splice_read() 1392 cs.pipebufs = bufs; fuse_dev_splice_read() 1415 struct pipe_buffer *buf = pipe->bufs + newbuf; fuse_dev_splice_read() 1417 buf->page = bufs[page_nr].page; fuse_dev_splice_read() 1418 buf->offset = bufs[page_nr].offset; fuse_dev_splice_read() 1419 buf->len = bufs[page_nr].len; fuse_dev_splice_read() 1446 page_cache_release(bufs[page_nr].page); fuse_dev_splice_read() 1448 kfree(bufs); fuse_dev_splice_read() 1988 struct pipe_buffer *bufs; fuse_dev_splice_write() local 1998 bufs = kmalloc(pipe->buffers * sizeof(struct pipe_buffer), GFP_KERNEL); fuse_dev_splice_write() 1999 if (!bufs) fuse_dev_splice_write() 2006 rem += pipe->bufs[(pipe->curbuf + idx) & (pipe->buffers - 1)].len; fuse_dev_splice_write() 2021 ibuf = &pipe->bufs[pipe->curbuf]; fuse_dev_splice_write() 2022 obuf = &bufs[nbuf]; fuse_dev_splice_write() 2043 cs.pipebufs = bufs; fuse_dev_splice_write() 2053 struct pipe_buffer *buf = &bufs[idx]; fuse_dev_splice_write() 2057 kfree(bufs); fuse_dev_splice_write()
|
/linux-4.1.27/drivers/media/usb/usbtv/ |
H A D | usbtv-video.c | 301 if (list_empty(&usbtv->bufs)) { usbtv_image_chunk() 308 buf = list_first_entry(&usbtv->bufs, struct usbtv_buf, list); usbtv_image_chunk() 422 while (!list_empty(&usbtv->bufs)) { usbtv_stop() 423 struct usbtv_buf *buf = list_first_entry(&usbtv->bufs, usbtv_stop() 627 list_add_tail(&buf->list, &usbtv->bufs); usbtv_buf_queue() 674 INIT_LIST_HEAD(&usbtv->bufs); usbtv_video_init()
|
H A D | usbtv.h | 82 struct list_head bufs; member in struct:usbtv
|
/linux-4.1.27/drivers/staging/octeon/ |
H A D | ethernet-rx.c | 236 skb_in_hw = USE_SKBUFFS_IN_HW && work->word2.s.bufs == 1; cvm_oct_napi_poll() 277 if (unlikely(work->word2.s.bufs == 0)) { cvm_oct_napi_poll() 293 int segments = work->word2.s.bufs; cvm_oct_napi_poll()
|
H A D | ethernet-tx.c | 617 work->word2.s.bufs = 1; cvm_oct_xmit_pow()
|
H A D | ethernet.c | 200 int segments = work->word2.s.bufs; cvm_oct_free_work()
|
/linux-4.1.27/drivers/scsi/ |
H A D | qlogicpti.h | 339 struct pti_queue_entry *res_cpu; /* Ptr to RESPONSE bufs (CPU) */ 340 struct pti_queue_entry *req_cpu; /* Ptr to REQUEST bufs (CPU) */ 359 __u32 res_dvma; /* Ptr to RESPONSE bufs (DVMA)*/ 360 __u32 req_dvma; /* Ptr to REQUEST bufs (DVMA) */
|
/linux-4.1.27/drivers/scsi/ibmvscsi/ |
H A D | ibmvscsi.h | 41 /* Number of indirect bufs...the list of these has to fit in the
|
/linux-4.1.27/drivers/spi/ |
H A D | spi-fsl-lib.h | 28 /* rx & tx bufs from the spi_transfer */
|
H A D | spi-sirf.c | 149 /* rx & tx bufs from the spi_transfer */
|
/linux-4.1.27/drivers/dma/ |
H A D | dmatest.c | 229 static void dmatest_init_srcs(u8 **bufs, unsigned int start, unsigned int len, dmatest_init_srcs() argument 235 for (; (buf = *bufs); bufs++) { dmatest_init_srcs() 247 static void dmatest_init_dsts(u8 **bufs, unsigned int start, unsigned int len, dmatest_init_dsts() argument 253 for (; (buf = *bufs); bufs++) { dmatest_init_dsts() 286 static unsigned int dmatest_verify(u8 **bufs, unsigned int start, dmatest_verify() argument 297 for (; (buf = *bufs); bufs++) { dmatest_verify()
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
H A D | msm_gem.h | 91 /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc,
|
/linux-4.1.27/drivers/atm/ |
H A D | horizon.h | 315 HDW cell_count; // AAL5 rx bufs 318 HDW partial_crc; // AAL5 rx bufs 319 HDW cell_header; // RAW bufs 357 // cell_buf bufs[2048];
|
H A D | fore200e.c | 984 printk(FORE200E "bsq_audit(%d): queue %d.%d, %d bufs in free list, but freebuf_count = %d\n", bsq_audit() 1021 printk(FORE200E "no more free bufs in queue %d.%d, but freebuf_count = %d\n", fore200e_supply() 2859 " free small bufs, scheme 1:\t%d\n" fore200e_proc_read() 2860 " free large bufs, scheme 1:\t%d\n" fore200e_proc_read() 2861 " free small bufs, scheme 2:\t%d\n" fore200e_proc_read() 2862 " free large bufs, scheme 2:\t%d\n", fore200e_proc_read()
|
H A D | lanai.c | 352 /* Lanai requires 256-byte alignment of DMA bufs */ lanai_buf_allocate()
|
/linux-4.1.27/arch/mips/include/asm/octeon/ |
H A D | cvmx-wqe.h | 62 uint64_t bufs:8; member in struct:__anon2032::__anon2033 193 uint64_t bufs:8; 221 uint64_t bufs:8; member in struct:__anon2032::__anon2035 352 uint64_t bufs:8;
|
H A D | cvmx-helper-util.h | 156 number_buffers = work->word2.s.bufs; cvmx_helper_free_packet_data()
|
/linux-4.1.27/include/drm/ |
H A D | drm_legacy.h | 101 struct drm_buf_entry bufs[DRM_MAX_ORDER + 1]; /**< buffers, grouped by their size order */ member in struct:drm_device_dma 103 struct drm_buf **buflist; /**< Vector of pointers into drm_device_dma::bufs */
|
/linux-4.1.27/drivers/s390/net/ |
H A D | qeth_core_main.c | 325 q->bufs[i].buffer = q->qdio_bufs[i]; qeth_alloc_qdio_queue() 439 if (q->bufs[bidx]->next_pending != NULL) { qeth_cleanup_handled_pending() 440 struct qeth_qdio_out_buffer *head = q->bufs[bidx]; qeth_cleanup_handled_pending() 441 struct qeth_qdio_out_buffer *c = q->bufs[bidx]->next_pending; qeth_cleanup_handled_pending() 466 if (forced_cleanup && (atomic_read(&(q->bufs[bidx]->state)) == qeth_cleanup_handled_pending() 469 q->bufs[bidx]->aob = q->bufstates[bidx].aob; qeth_cleanup_handled_pending() 1281 if (!q->bufs[j]) qeth_clear_outq_buffers() 1284 qeth_clear_output_buffer(q, q->bufs[j], QETH_QDIO_BUF_EMPTY); qeth_clear_outq_buffers() 1286 kmem_cache_free(qeth_qdio_outbuf_cache, q->bufs[j]); qeth_clear_outq_buffers() 1287 q->bufs[j] = NULL; qeth_clear_outq_buffers() 2436 newbuf->next_pending = q->bufs[bidx]; qeth_init_qdio_out_buf() 2438 q->bufs[bidx] = newbuf; qeth_init_qdio_out_buf() 2507 WARN_ON(card->qdio.out_qs[i]->bufs[j] != NULL); qeth_alloc_qdio_buffers() 2523 card->qdio.out_qs[i]->bufs[j]); qeth_alloc_qdio_buffers() 2524 card->qdio.out_qs[i]->bufs[j] = NULL; qeth_alloc_qdio_buffers() 2554 if (card->qdio.in_q->bufs[j].rx_skb) qeth_free_qdio_buffers() 2555 dev_kfree_skb_any(card->qdio.in_q->bufs[j].rx_skb); qeth_free_qdio_buffers() 2874 qeth_init_input_buffer(card, &card->qdio.in_q->bufs[i]); qeth_init_qdio_queues() 2896 card->qdio.out_qs[i]->bufs[j], qeth_init_qdio_queues() 3328 &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q])) { qeth_queue_input_buffer() 3450 buffer = queue->bufs[queue->next_buf_to_fill]; qeth_switch_to_nonpacking_if_needed() 3476 buffer = queue->bufs[queue->next_buf_to_fill]; qeth_flush_buffers_on_no_pci() 3498 buf = queue->bufs[bidx]; qeth_flush_buffers() 3760 buffer = queue->bufs[bidx]; qeth_qdio_output_handler() 4076 buffer = queue->bufs[queue->next_buf_to_fill]; qeth_do_send_packet_fast() 4110 buffer = queue->bufs[queue->next_buf_to_fill]; qeth_do_send_packet() 4132 buffer = queue->bufs[queue->next_buf_to_fill]; qeth_do_send_packet() 4893 virt_to_phys(card->qdio.c_q->bufs[i].buffer); qeth_qdio_establish_cq() 4931 virt_to_phys(card->qdio.in_q->bufs[i].buffer); qeth_qdio_establish() 4955 card->qdio.out_qs[i]->bufs[j]->buffer); qeth_qdio_establish()
|
H A D | qeth_core.h | 438 struct qeth_qdio_buffer bufs[QDIO_MAX_BUFFERS_PER_Q]; member in struct:qeth_qdio_q 464 struct qeth_qdio_out_buffer *bufs[QDIO_MAX_BUFFERS_PER_Q]; member in struct:qeth_qdio_out_q
|
H A D | qeth_l2_main.c | 445 &card->qdio.in_q->bufs[card->rx.b_index], qeth_l2_process_inbound_buffer() 510 &card->qdio.in_q->bufs[card->rx.b_index] qeth_l2_poll() 516 buffer = &card->qdio.in_q->bufs[card->rx.b_index]; qeth_l2_poll() 538 ->bufs[card->rx.b_index] qeth_l2_poll()
|
H A D | qeth_l3_main.c | 1972 &card->qdio.in_q->bufs[card->rx.b_index], qeth_l3_process_inbound_buffer() 2047 &card->qdio.in_q->bufs[card->rx.b_index] qeth_l3_poll() 2053 buffer = &card->qdio.in_q->bufs[card->rx.b_index]; qeth_l3_poll() 2075 ->bufs[card->rx.b_index] qeth_l3_poll()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | pack_generic.c | 192 char **bufs) lustre_init_msg_v2() 204 if (bufs == NULL) lustre_init_msg_v2() 209 char *tmp = bufs[i]; lustre_init_msg_v2() 216 int count, __u32 *lens, char **bufs) lustre_pack_request_v2() 228 lustre_init_msg_v2(req->rq_reqmsg, count, lens, bufs); lustre_pack_request_v2() 234 __u32 *lens, char **bufs) lustre_pack_request() 247 return lustre_pack_request_v2(req, count, lens, bufs); lustre_pack_request() 321 __u32 *lens, char **bufs, int flags) lustre_pack_reply_v2() 353 lustre_init_msg_v2(rs->rs_msg, count, lens, bufs); lustre_pack_reply_v2() 363 char **bufs, int flags) lustre_pack_reply_flags() 378 rc = lustre_pack_reply_v2(req, count, lens, bufs, flags); lustre_pack_reply_flags() 393 char **bufs) lustre_pack_reply() 395 return lustre_pack_reply_flags(req, count, lens, bufs, 0); lustre_pack_reply() 191 lustre_init_msg_v2(struct lustre_msg_v2 *msg, int count, __u32 *lens, char **bufs) lustre_init_msg_v2() argument 215 lustre_pack_request_v2(struct ptlrpc_request *req, int count, __u32 *lens, char **bufs) lustre_pack_request_v2() argument 233 lustre_pack_request(struct ptlrpc_request *req, __u32 magic, int count, __u32 *lens, char **bufs) lustre_pack_request() argument 320 lustre_pack_reply_v2(struct ptlrpc_request *req, int count, __u32 *lens, char **bufs, int flags) lustre_pack_reply_v2() argument 362 lustre_pack_reply_flags(struct ptlrpc_request *req, int count, __u32 *lens, char **bufs, int flags) lustre_pack_reply_flags() argument 392 lustre_pack_reply(struct ptlrpc_request *req, int count, __u32 *lens, char **bufs) lustre_pack_reply() argument
|
H A D | client.c | 580 int count, __u32 *lengths, char **bufs, __ptlrpc_request_bufs_pack() 597 lengths, bufs); __ptlrpc_request_bufs_pack() 647 __u32 version, int opcode, char **bufs, ptlrpc_request_bufs_pack() 655 bufs, ctx); ptlrpc_request_bufs_pack() 813 * are provided by a pointer \a bufs. 819 int count, __u32 *lengths, char **bufs, ptlrpc_prep_req_pool() 830 lengths, bufs, NULL); ptlrpc_prep_req_pool() 844 __u32 *lengths, char **bufs) ptlrpc_prep_req() 846 return ptlrpc_prep_req_pool(imp, version, opcode, count, lengths, bufs, ptlrpc_prep_req() 578 __ptlrpc_request_bufs_pack(struct ptlrpc_request *request, __u32 version, int opcode, int count, __u32 *lengths, char **bufs, struct ptlrpc_cli_ctx *ctx) __ptlrpc_request_bufs_pack() argument 646 ptlrpc_request_bufs_pack(struct ptlrpc_request *request, __u32 version, int opcode, char **bufs, struct ptlrpc_cli_ctx *ctx) ptlrpc_request_bufs_pack() argument 817 ptlrpc_prep_req_pool(struct obd_import *imp, __u32 version, int opcode, int count, __u32 *lengths, char **bufs, struct ptlrpc_request_pool *pool) ptlrpc_prep_req_pool() argument 843 ptlrpc_prep_req(struct obd_import *imp, __u32 version, int opcode, int count, __u32 *lengths, char **bufs) ptlrpc_prep_req() argument
|
H A D | import.c | 639 char *bufs[] = { NULL, ptlrpc_connect_import() local 704 imp->imp_connect_op, bufs, NULL); ptlrpc_connect_import()
|
H A D | lproc_ptlrpc.c | 223 svc_counter_config, "reqbuf_avail", "bufs"); ptlrpc_lprocfs_register()
|
/linux-4.1.27/drivers/media/platform/omap/ |
H A D | omap_vout.c | 899 if (NULL == q->bufs[i]) omap_vout_mmap() 901 if (V4L2_MEMORY_MMAP != q->bufs[i]->memory) omap_vout_mmap() 903 if (q->bufs[i]->boff == (vma->vm_pgoff << PAGE_SHIFT)) omap_vout_mmap() 921 q->bufs[i]->baddr = vma->vm_start; omap_vout_mmap() 1505 if (q->bufs[0] && (V4L2_MEMORY_MMAP == q->bufs[0]->memory)) { vidioc_reqbufs() 1520 } else if (q->bufs[0] && (V4L2_MEMORY_USERPTR == q->bufs[0]->memory)) { vidioc_reqbufs() 1524 kfree(q->bufs[i]); vidioc_reqbufs() 1525 q->bufs[i] = NULL; vidioc_reqbufs() 1564 (q->bufs[buffer->index]->memory != buffer->memory)) { vidioc_qbuf() 1594 vb = q->bufs[b->index]; vidioc_dqbuf()
|
/linux-4.1.27/drivers/tty/ |
H A D | synclink_gt.c | 329 char *bufs; /* virtual address of DMA buffer lists */ member in struct:slgt_info 491 static int alloc_bufs(struct slgt_info *info, struct slgt_desc *bufs, int count); 492 static void free_bufs(struct slgt_info *info, struct slgt_desc *bufs, int count); 1861 struct slgt_desc *bufs = info->rbufs; rx_async() local 1869 while(desc_complete(bufs[end])) { rx_async() 1870 count = desc_count(bufs[end]) - info->rbuf_index; rx_async() 1871 p = bufs[end].buf + info->rbuf_index; rx_async() 3386 info->bufs = pci_zalloc_consistent(info->pdev, DESC_LIST_SIZE, alloc_desc() 3388 if (info->bufs == NULL) alloc_desc() 3391 info->rbufs = (struct slgt_desc*)info->bufs; alloc_desc() 3392 info->tbufs = ((struct slgt_desc*)info->bufs) + info->rbuf_count; alloc_desc() 3428 if (info->bufs != NULL) { free_desc() 3429 pci_free_consistent(info->pdev, DESC_LIST_SIZE, info->bufs, info->bufs_dma_addr); free_desc() 3430 info->bufs = NULL; free_desc() 3436 static int alloc_bufs(struct slgt_info *info, struct slgt_desc *bufs, int count) alloc_bufs() argument 3440 if ((bufs[i].buf = pci_alloc_consistent(info->pdev, DMABUFSIZE, &bufs[i].buf_dma_addr)) == NULL) alloc_bufs() 3442 bufs[i].pbuf = cpu_to_le32((unsigned int)bufs[i].buf_dma_addr); alloc_bufs() 3447 static void free_bufs(struct slgt_info *info, struct slgt_desc *bufs, int count) free_bufs() argument 3451 if (bufs[i].buf == NULL) free_bufs() 3453 pci_free_consistent(info->pdev, DMABUFSIZE, bufs[i].buf, bufs[i].buf_dma_addr); free_bufs() 3454 bufs[i].buf = NULL; free_bufs() 3476 if (info->bufs) { free_dma_bufs()
|
H A D | mips_ejtag_fdc.c | 1252 const char *bufs[1] = { kgdbfdc_wbuf }; kgdbfdc_push_one() local 1258 word = mips_ejtag_fdc_encode(bufs, &kgdbfdc_wbuflen, 1); kgdbfdc_push_one()
|
/linux-4.1.27/drivers/scsi/csiostor/ |
H A D | csio_wr.c | 119 struct csio_dma_buf *buf = &flq->un.fl.bufs[0]; csio_wr_fill_fl() 279 flq->un.fl.bufs = kzalloc(flq->credits * csio_wr_alloc_q() 282 if (!flq->un.fl.bufs) { csio_wr_alloc_q() 284 "Failed to allocate FL queue bufs" csio_wr_alloc_q() 1069 buf = &flq->un.fl.bufs[flq->cidx]; csio_wr_process_fl() 1624 if (!q->un.fl.bufs) csio_wrm_exit() 1627 buf = &q->un.fl.bufs[j]; csio_wrm_exit() 1634 kfree(q->un.fl.bufs); csio_wrm_exit()
|
H A D | csio_wr.h | 374 struct csio_dma_buf *bufs; /* Free list buffer ptr array member in struct:csio_fl
|
H A D | csio_scsi.c | 625 /* set number of ddp bufs used */ csio_setup_ddp()
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
H A D | debug.c | 243 seq_printf(seq, "TX bufs: %d x %d bytes\n", cw1200_status_show() 246 seq_printf(seq, "Used bufs: %d\n", cw1200_status_show()
|
/linux-4.1.27/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-helper-util.c | 101 cvmx_dprintf(" Buffers: %u\n", work->word2.s.bufs); cvmx_helper_dump_packet() 103 if (work->word2.s.bufs == 0) { cvmx_helper_dump_packet()
|
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/ |
H A D | mdc_reint.c | 259 * mdc_create_pack() fills msg->bufs[1] with name and msg->bufs[2] with mdc_create()
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
H A D | c2_vq.c | 74 * VQ Reply bufs are local host memory copies of a
|
/linux-4.1.27/drivers/media/pci/saa7164/ |
H A D | saa7164-buffer.c | 276 /* Don't exceed the device maximum # bufs */ saa7164_buffer_cfg_port()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_init_chip.c | 756 dev_info(&dd->pcidev->dev, "Allocating %d PIO bufs of " ipath_init_chip() 782 ipath_cdbg(VERBOSE, "%d PIO bufs for kernel out of %d total %u " ipath_init_chip() 904 "rcvhdrq and/or egr bufs\n"); ipath_init_chip() 1049 "Allocating %d PIO bufs for kernel leaves " ipath_set_kpiobufs()
|
H A D | ipath_wc_x86_64.c | 134 "mtrr_add() WC for PIO bufs " ipath_enable_wc()
|
H A D | ipath_kernel.h | 98 * rcvegr bufs base, physical, must fit 120 /* number of pio bufs for this port (all procs, if shared) */ 407 /* pio bufs allocated per port */ 409 /* if remainder on bufs/port, ports < extrabuf get 1 extra */ 414 * supports, less gives more pio bufs/port, etc.
|
H A D | ipath_ruc.c | 677 * If no PIO bufs are available, return. An interrupt will ipath_do_send()
|
H A D | ipath_driver.c | 765 * free up port 0 (kernel) rcvhdr, egr bufs, and eventually tid bufs ipath_remove_one() 827 /* can't disarm bufs back-to-back per iba7220 spec */ ipath_disarm_piobufs()
|
H A D | ipath_common.h | 324 /* num bufs in each contiguous set */
|
H A D | ipath_file_ops.c | 1113 dev_info(&dd->pcidev->dev, "FAIL on egr bufs: " mmap_rcvegrbufs()
|
/linux-4.1.27/drivers/virtio/ |
H A D | virtio_ring.c | 627 u16 bufs; virtqueue_enable_cb_delayed() local 638 bufs = (u16)(virtio16_to_cpu(_vq->vdev, vq->vring.avail->idx) - vq->last_used_idx) * 3 / 4; virtqueue_enable_cb_delayed() 639 vring_used_event(&vq->vring) = cpu_to_virtio16(_vq->vdev, vq->last_used_idx + bufs); virtqueue_enable_cb_delayed() 641 if (unlikely((u16)(virtio16_to_cpu(_vq->vdev, vq->vring.used->idx) - vq->last_used_idx) > bufs)) { virtqueue_enable_cb_delayed()
|
/linux-4.1.27/include/media/ |
H A D | videobuf-core.h | 147 struct videobuf_buffer *bufs[VIDEO_MAX_FRAME]; member in struct:videobuf_queue
|
H A D | videobuf2-core.h | 368 * @bufs: videobuf buffer structures 408 struct vb2_buffer *bufs[VIDEO_MAX_FRAME]; member in struct:vb2_queue
|
/linux-4.1.27/sound/oss/dmasound/ |
H A D | dmasound_core.c | 956 static int set_queue_frags(struct sound_queue *sq, int bufs, int size) set_queue_frags() argument 971 if (bufs <= 0) set_queue_frags() 973 if (bufs > sq->numBufs) /* the user is allowed say "don't care" with 0x7fff */ set_queue_frags() 974 bufs = sq->numBufs ; set_queue_frags() 982 sq->max_active = bufs ; set_queue_frags()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib.h | 83 __u64 sps_nopiobufs; /* no pio bufs avail from kernel */ 143 * rcvegr bufs base, physical, must fit 172 /* number of pio bufs for this ctxt (all procs, if shared) */ 876 /* pio bufs allocated per ctxt */ 878 /* if remainder on bufs/ctxt, ctxts < extrabuf get 1 extra */ 882 * supports, less gives more pio bufs/ctxt, etc.
|
H A D | qib_diag.c | 225 * - piobufs (2K and 4K bufs in either order) 228 * - piobufs (2K and 4K bufs in either order)
|
H A D | qib_init.c | 703 "failed to allocate kernel ctxt's rcvhdrq and/or egr bufs\n"); qib_init() 1788 * - piobufs (2K and 4K bufs in either order) init_chip_wc_pat() 1791 * - piobufs (2K and 4K bufs in either order) init_chip_wc_pat()
|
H A D | qib_common.h | 239 /* num bufs in each contiguous set */
|
H A D | qib_file_ops.c | 864 "FAIL on egr bufs: reqlen %lx > actual %lx\n", mmap_rcvegrbufs() 1717 qib_dev_err(dd, "Ctxt%u: would use 4KB bufs, using %u\n", qib_do_user_init()
|
H A D | qib_iba7220.c | 4145 * all the 4k bufs for the kernel. qib_init_7220_variables()
|
H A D | qib_iba7322.c | 6722 * all the 4k bufs for the kernel. qib_init_7322_variables()
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 611 * for the posted rx bufs refcount to become zero handles everything */ iser_rcv_completion() 623 iser_err("posting %d rx bufs err %d\n", count, err); iser_rcv_completion()
|
/linux-4.1.27/drivers/mtd/nand/ |
H A D | mxc_nand.c | 534 int bufs, i; send_page_v1() local 537 bufs = 4; send_page_v1() 539 bufs = 1; send_page_v1() 541 for (i = 0; i < bufs; i++) { send_page_v1()
|
/linux-4.1.27/drivers/media/platform/marvell-ccic/ |
H A D | mcam-core.h | 174 unsigned int buf_seq[MAX_DMA_BUFS]; /* Sequence for individual bufs */
|
H A D | mcam-core.c | 411 * planar formats for now, so no UV bufs). Then either mcam_ctlr_dma_vmalloc()
|
/linux-4.1.27/drivers/net/ |
H A D | ntb_netdev.c | 183 /* Add some empty rx bufs */ ntb_netdev_open()
|
/linux-4.1.27/drivers/staging/slicoss/ |
H A D | slic.h | 109 ushort bottombits; /* to denote num bufs to card */
|
H A D | slicoss.c | 1858 ihcmd->u.slic_buffers.bufs[0].paddrl = SLIC_GET_ADDR_LOW(phys_addr); slic_xmit_build_request() 1859 ihcmd->u.slic_buffers.bufs[0].paddrh = SLIC_GET_ADDR_HIGH(phys_addr); slic_xmit_build_request() 1860 ihcmd->u.slic_buffers.bufs[0].length = skb->len; slic_xmit_build_request() 1862 hcmd->cmdsize = (u32) ((((u64)&ihcmd->u.slic_buffers.bufs[1] - slic_xmit_build_request() 1865 hcmd->cmdsize = (((u32)&ihcmd->u.slic_buffers.bufs[1] - slic_xmit_build_request()
|
H A D | slichw.h | 279 struct slic_host64sg bufs[SLIC_MAX64_BCNT]; member in struct:slic_host64_cmd::__anon10273::__anon10274
|
/linux-4.1.27/drivers/staging/lustre/lustre/mgc/ |
H A D | mgc_request.c | 1120 struct lustre_cfg_bufs bufs; mgc_apply_recover_logs() local 1239 lustre_cfg_bufs_reset(&bufs, obdname); mgc_apply_recover_logs() 1284 lustre_cfg_bufs_set_string(&bufs, 1, params); mgc_apply_recover_logs() 1287 lcfg = lustre_cfg_new(LCFG_PARAM, &bufs); mgc_apply_recover_logs()
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_trans_buf.c | 766 * buffers is that we must not replay dquot bufs when recovering 768 * between usr dquot bufs and grp dquot bufs, because usr and grp quotas
|
H A D | xfs_buf.c | 1491 * Wait for any bufs with callbacks that have been submitted but have not yet
|
H A D | xfs_qm.c | 1179 * Once we're done, we'll log all the dquot bufs. xfs_qm_dqusage_adjust()
|
H A D | xfs_log.c | 1672 * The bdstrat callback function for log bufs. This gives us a central 1673 * place to trap bufs in case we get hit by a log I/O error and need to
|
/linux-4.1.27/drivers/media/usb/dvb-usb/ |
H A D | cxusb.c | 291 u8 buf, bufs[] = { cxusb_aver_power_ctrl() local 304 for (i = 0; i < sizeof(bufs)/sizeof(u8); i += 4/sizeof(u8)) { cxusb_aver_power_ctrl() 306 bufs+i, 4, &buf, 1); cxusb_aver_power_ctrl()
|
/linux-4.1.27/drivers/net/ethernet/apple/ |
H A D | bmac.c | 1418 XXDEBUG(("bmac: free rx bufs\n")); bmac_close() 1425 XXDEBUG(("bmac: free tx bufs\n")); bmac_close() 1432 XXDEBUG(("bmac: all bufs freed\n")); bmac_close()
|
/linux-4.1.27/drivers/media/tuners/ |
H A D | mt20xx.c | 492 tuner_dbg("bufs is: "); mt2050_set_if_freq()
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 454 * set the highest bits. This will break if we really allow exporting of dma-bufs. qxl_release_fence_buffer_objects()
|
/linux-4.1.27/net/caif/ |
H A D | cfctrl.c | 460 /* Fifosize bufs */ cfctrl_recv()
|
/linux-4.1.27/sound/soc/intel/atom/sst/ |
H A D | sst.h | 185 * @bufs: stream buffer list
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_log_format.h | 221 #define XFS_LI_BUF 0x123c /* v2 bufs, variable sized inode bufs */
|
/linux-4.1.27/fs/gfs2/ |
H A D | log.c | 428 * We can have metadata bufs and jdata bufs in the same journal. So each
|
/linux-4.1.27/drivers/media/platform/ti-vpe/ |
H A D | vpe.c | 384 unsigned int bufs_completed; /* bufs done in this batch */ 390 dma_addr_t mv_buf_dma[2]; /* dma addrs of motion vector in/out bufs */ 391 void *mv_buf[2]; /* virtual addrs of motion vector bufs */
|
/linux-4.1.27/drivers/media/pci/pt3/ |
H A D | pt3.c | 697 dev_err(&pt3->pdev->dev, "failed to alloc DMA bufs\n"); pt3_resume()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
H A D | bnad_ethtool.c | 887 /* Fill hardware stats excluding the rxf/txf into ethtool bufs */ bnad_get_ethtool_stats()
|
H A D | bnad.c | 180 * bnad_txcmpl_process : Frees the Tx bufs on Tx completion
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_object.c | 582 * ttm_prime_dmabuf_release - Release method for the dma-bufs we export
|
/linux-4.1.27/drivers/staging/vt6655/ |
H A D | device_main.c | 639 dev_err(&pDevice->pcid->dev, "can not alloc rx bufs\n"); device_init_rd0_ring() 663 dev_err(&pDevice->pcid->dev, "can not alloc rx bufs\n"); device_init_rd1_ring()
|
/linux-4.1.27/include/sound/ |
H A D | pcm.h | 1078 void __user **bufs, snd_pcm_uframes_t frames); 1080 void __user **bufs, snd_pcm_uframes_t frames);
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
H A D | mad.c | 1789 pr_err("Failed allocating qp0 tunnel bufs (%d)\n", ret); create_pv_resources() 1796 pr_err("Failed allocating qp1 tunnel bufs (%d)\n", ret); create_pv_resources()
|
/linux-4.1.27/drivers/media/pci/cx23885/ |
H A D | cx23885-417.c | 1100 data[2] = 4; /* total bufs */ cx23885_initialize_codec()
|
/linux-4.1.27/drivers/net/wireless/ath/ath10k/ |
H A D | htt.h | 1314 /* where HTT SW has processed bufs filled by rx MAC DMA */
|
/linux-4.1.27/drivers/net/wireless/ath/ath5k/ |
H A D | debug.c | 859 " len: %d bufs: %d\n", txq->txq_len, n); read_file_queue()
|
H A D | ath5k.h | 1252 #define ATH5K_TXQ_LEN_MAX (ATH_TXBUF / 4) /* bufs per queue */
|
/linux-4.1.27/drivers/media/platform/s5p-mfc/ |
H A D | s5p_mfc_dec.c | 212 .name = "Minimum number of cap bufs",
|
H A D | s5p_mfc_enc.c | 694 .name = "Minimum number of output bufs",
|
H A D | s5p_mfc_opr_v5.c | 75 mfc_debug(2, "Totals bufs: %d\n", ctx->total_dpb_count); s5p_mfc_alloc_codec_buffers_v5()
|
H A D | s5p_mfc_opr_v6.c | 75 mfc_debug(2, "Totals bufs: %d\n", ctx->total_dpb_count); s5p_mfc_alloc_codec_buffers_v6()
|
/linux-4.1.27/drivers/media/usb/pwc/ |
H A D | pwc-if.c | 678 /* And cleanup any queued bufs!! */ start_streaming()
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
H A D | sunbmac.c | 222 /* Free any skippy bufs left around in the rings. */ bigmac_init_rings()
|
H A D | sunhme.c | 1257 /* Free any skippy bufs left around in the rings. */ happy_meal_init_rings() 1261 /* Now get new skippy bufs for the receive ring. */ happy_meal_init_rings()
|
/linux-4.1.27/drivers/staging/wlan-ng/ |
H A D | p80211netdev.c | 411 netdev_dbg(netdev, "txframe success, no more bufs\n"); p80211knetdev_hard_start_xmit()
|
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/ |
H A D | router.c | 1324 CERROR("Failed to allocate %d router bufs of %d pages\n", lnet_rtrpool_alloc_bufs()
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/lustre/ |
H A D | lustre_idl.h | 3640 * bufs[0] : following are packets of object. 3644 * lens/bufs: other parameters of the update. 3648 * lens/bufs: other parameters of the update. 3652 * lens/bufs: other parameters of the update.
|
/linux-4.1.27/include/uapi/sound/ |
H A D | asound.h | 493 void __user * __user *bufs; member in struct:snd_xfern
|
/linux-4.1.27/drivers/net/ethernet/8390/ |
H A D | lib8390.c | 339 * slot, and then perform some sanity checks. With two Tx bufs, __ei_start_xmit()
|
H A D | axnet_cs.c | 984 * slot, and then perform some sanity checks. With two Tx bufs, axnet_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/freescale/ |
H A D | gianfar.h | 71 /* Number of bytes to align the rx bufs to */
|
/linux-4.1.27/drivers/net/ethernet/i825xx/ |
H A D | lib82596.c | 922 printk(KERN_ERR "%s: Failed to init rx bufs\n", dev->name); i596_open()
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
H A D | htc_pipe.c | 1069 * TODO: the message based HIF architecture allocates net bufs ath6kl_htc_pipe_rx_complete()
|
/linux-4.1.27/drivers/media/platform/soc_camera/ |
H A D | omap1_camera.c | 1416 struct omap1_cam_buf *buf = container_of(icd->vb_vidq.bufs[i], omap1_cam_reqbufs()
|
H A D | pxa_camera.c | 1548 struct pxa_buffer *buf = container_of(icd->vb_vidq.bufs[i], pxa_camera_reqbufs()
|
/linux-4.1.27/drivers/media/usb/cx231xx/ |
H A D | cx231xx-417.c | 1184 data[2] = 4; /* total bufs */ cx231xx_initialize_codec()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | siena_sriov.c | 543 "ERROR: Invalid INIT_EVQ from %s: evq %d bufs %d\n", efx_vfdi_init_evq()
|
/linux-4.1.27/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 934 /* Allocate and initialize tx rings. 16kb = 128 bufs. */ ioc3_alloc_rings()
|
/linux-4.1.27/drivers/net/wireless/ipw2x00/ |
H A D | ipw2200.h | 1216 int rx_bufs_min; /**< minimum number of bufs in Rx queue */
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | radeon_cp.c | 1903 * bufs until freelist code is used. Note this hides a problem with
|
/linux-4.1.27/drivers/tty/serial/ |
H A D | mpsc.c | 1149 } else { /* All tx data copied into ring bufs */ mpsc_copy_tx_data()
|
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge.c | 2202 netdev_err(dev, "slice-%d: alloced only %d small bufs\n", myri10ge_allocate_rings() 2209 netdev_err(dev, "slice-%d: alloced only %d big bufs\n", myri10ge_allocate_rings()
|
/linux-4.1.27/drivers/net/vmxnet3/ |
H A D | vmxnet3_drv.c | 1466 } else { /* subsequent bufs for a pkt is frag */ vmxnet3_rq_init()
|
/linux-4.1.27/drivers/net/wan/ |
H A D | farsync.c | 1185 dbg(DBG_ASS, "intr_rx: Discarding more bufs" fst_recover_rx_error()
|
/linux-4.1.27/drivers/net/ethernet/dec/tulip/ |
H A D | de4x5.c | 859 char *rx_bufs; /* rx bufs on alpha, sparc, ... */
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 1888 /* If rx bufs are exhausted called after 50ms to attempt to refresh */ nv_do_rx_refill()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_scsi.c | 5445 * extra. This list of scsi bufs exists for the lifetime of the driver. lpfc_slave_alloc()
|
H A D | lpfc_init.c | 9736 * Call scsi_free before mem_free since scsi bufs are released to their lpfc_pci_remove_one_s3()
|
/linux-4.1.27/fs/ocfs2/ |
H A D | xattr.c | 1520 /* Value bufs are for value trees */ ocfs2_xa_fill_value_buf()
|