/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
D | iwch_mem.c | 134 int build_phys_page_list(struct ib_phys_buf *buffer_list, in build_phys_page_list() argument 148 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list() 151 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list() 153 *total_size += buffer_list[i].size; in build_phys_page_list() 155 mask |= buffer_list[i].addr; in build_phys_page_list() 157 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list() 159 mask |= buffer_list[i].addr + buffer_list[i].size; in build_phys_page_list() 161 mask |= (buffer_list[i].addr + buffer_list[i].size + in build_phys_page_list() 173 buffer_list[0].size += buffer_list[0].addr & ((1ULL << *shift) - 1); in build_phys_page_list() 174 buffer_list[0].addr &= ~0ull << *shift; in build_phys_page_list() [all …]
|
D | iwch_provider.c | 473 struct ib_phys_buf *buffer_list, in iwch_register_phys_mem() argument 498 if ((*iova_start & ~PAGE_MASK) != (buffer_list[0].addr & ~PAGE_MASK)) { in iwch_register_phys_mem() 504 ((buffer_list[0].addr + buffer_list[0].size) & ~PAGE_MASK)) { in iwch_register_phys_mem() 509 ret = build_phys_page_list(buffer_list, num_phys_buf, iova_start, in iwch_register_phys_mem() 552 struct ib_phys_buf *buffer_list, in iwch_reregister_phys_mem() argument 587 ret = build_phys_page_list(buffer_list, num_phys_buf, in iwch_reregister_phys_mem()
|
D | iwch_provider.h | 349 int build_phys_page_list(struct ib_phys_buf *buffer_list,
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 434 static int build_phys_page_list(struct ib_phys_buf *buffer_list, in build_phys_page_list() argument 445 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list() 448 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list() 450 *total_size += buffer_list[i].size; in build_phys_page_list() 452 mask |= buffer_list[i].addr; in build_phys_page_list() 454 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list() 456 mask |= buffer_list[i].addr + buffer_list[i].size; in build_phys_page_list() 458 mask |= (buffer_list[i].addr + buffer_list[i].size + in build_phys_page_list() 470 buffer_list[0].size += buffer_list[0].addr & ((1ULL << *shift) - 1); in build_phys_page_list() 471 buffer_list[0].addr &= ~0ull << *shift; in build_phys_page_list() [all …]
|
D | iw_cxgb4.h | 983 struct ib_phys_buf *buffer_list, 990 struct ib_phys_buf *buffer_list,
|
/linux-4.1.27/drivers/iio/ |
D | industrialio-buffer.c | 37 return !list_empty(&buf->buffer_list); in iio_buffer_is_active() 180 INIT_LIST_HEAD(&buffer->buffer_list); in iio_buffer_init() 532 list_add(&buffer->buffer_list, &indio_dev->buffer_list); in iio_buffer_activate() 537 list_del_init(&buffer->buffer_list); in iio_buffer_deactivate() 546 if (list_empty(&indio_dev->buffer_list)) in iio_disable_all_buffers() 553 &indio_dev->buffer_list, buffer_list) in iio_disable_all_buffers() 589 if (!list_empty(&indio_dev->buffer_list)) { in __iio_update_buffers() 613 if (list_empty(&indio_dev->buffer_list)) { in __iio_update_buffers() 630 list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { in __iio_update_buffers() 675 list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { in __iio_update_buffers() [all …]
|
D | industrialio-core.c | 998 INIT_LIST_HEAD(&dev->buffer_list); in iio_device_alloc()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_mr.c | 255 int start_index, u64 *buffer_list, int list_len) in __mthca_write_mtt() argument 273 mtt_entry[i + 2] = cpu_to_be64(buffer_list[i] | in __mthca_write_mtt() 291 buffer_list += i; in __mthca_write_mtt() 317 u64 *buffer_list, int list_len) in mthca_tavor_write_mtt_seg() argument 325 mthca_write64_raw(cpu_to_be64(buffer_list[i] | MTHCA_MTT_FLAG_PRESENT), in mthca_tavor_write_mtt_seg() 331 u64 *buffer_list, int list_len) in mthca_arbel_write_mtt_seg() argument 352 mtts[i] = cpu_to_be64(buffer_list[i] | MTHCA_MTT_FLAG_PRESENT); in mthca_arbel_write_mtt_seg() 359 int start_index, u64 *buffer_list, int list_len) in mthca_write_mtt() argument 366 return __mthca_write_mtt(dev, mtt, start_index, buffer_list, list_len); in mthca_write_mtt() 372 buffer_list, chunk); in mthca_write_mtt() [all …]
|
D | mthca_provider.c | 894 struct ib_phys_buf *buffer_list, in mthca_reg_phys_mr() argument 908 mask = buffer_list[0].addr ^ *iova_start; in mthca_reg_phys_mr() 912 mask |= buffer_list[i].addr; in mthca_reg_phys_mr() 914 mask |= buffer_list[i].addr + buffer_list[i].size; in mthca_reg_phys_mr() 916 total_size += buffer_list[i].size; in mthca_reg_phys_mr() 924 buffer_list[0].size += buffer_list[0].addr & ((1ULL << shift) - 1); in mthca_reg_phys_mr() 925 buffer_list[0].addr &= ~0ull << shift; in mthca_reg_phys_mr() 933 npages += (buffer_list[i].size + (1ULL << shift) - 1) >> shift; in mthca_reg_phys_mr() 947 j < (buffer_list[i].size + (1ULL << shift) - 1) >> shift; in mthca_reg_phys_mr() 949 page_list[n++] = buffer_list[i].addr + ((u64) j << shift); in mthca_reg_phys_mr() [all …]
|
D | mthca_dev.h | 470 int start_index, u64 *buffer_list, int list_len); 476 u64 *buffer_list, int buffer_size_shift,
|
/linux-4.1.27/fs/xfs/ |
D | xfs_qm.c | 874 struct list_head *buffer_list) in xfs_qm_dqiter_bufs() argument 924 xfs_buf_delwri_queue(bp, buffer_list); in xfs_qm_dqiter_bufs() 944 struct list_head *buffer_list) in xfs_qm_dqiterate() argument 1019 flags, buffer_list); in xfs_qm_dqiterate() 1222 struct list_head *buffer_list = data; in xfs_qm_flush_one() local 1237 xfs_buf_delwri_queue(bp, buffer_list); in xfs_qm_flush_one() 1256 LIST_HEAD (buffer_list); in xfs_qm_quotacheck() 1278 &buffer_list); in xfs_qm_quotacheck() 1286 &buffer_list); in xfs_qm_quotacheck() 1294 &buffer_list); in xfs_qm_quotacheck() [all …]
|
D | xfs_log_recover.c | 1503 LIST_HEAD(buffer_list); in xlog_recover_reorder_trans() 1513 list_move_tail(&item->ri_list, &buffer_list); in xlog_recover_reorder_trans() 1526 list_move_tail(&item->ri_list, &buffer_list); in xlog_recover_reorder_trans() 1554 if (!list_empty(&buffer_list)) in xlog_recover_reorder_trans() 1555 list_splice(&buffer_list, &trans->r_itemq); in xlog_recover_reorder_trans() 2314 struct list_head *buffer_list, in xlog_recover_buffer_pass2() argument 2415 xfs_buf_delwri_queue(bp, buffer_list); in xlog_recover_buffer_pass2() 2458 struct list_head *buffer_list) in xfs_recover_inode_owner_change() argument 2481 ip->i_ino, buffer_list); in xfs_recover_inode_owner_change() 2489 ip->i_ino, buffer_list); in xfs_recover_inode_owner_change() [all …]
|
D | xfs_dquot_item.c | 143 struct list_head *buffer_list) __releases(&lip->li_ailp->xa_lock) in xfs_qm_dquot_logitem_push() argument 183 if (!xfs_buf_delwri_queue(bp, buffer_list)) in xfs_qm_dquot_logitem_push() 330 struct list_head *buffer_list) in xfs_qm_qoff_logitem_push() argument
|
D | xfs_extfree_item.c | 165 struct list_head *buffer_list) in xfs_efi_item_push() argument 413 struct list_head *buffer_list) in xfs_efd_item_push() argument
|
D | xfs_icreate_item.c | 119 struct list_head *buffer_list) in xfs_icreate_item_push() argument
|
D | xfs_buf.c | 1788 struct list_head *buffer_list, in __xfs_buf_delwri_submit() argument 1796 list_for_each_entry_safe(bp, n, buffer_list, b_list) { in __xfs_buf_delwri_submit() 1859 struct list_head *buffer_list) in xfs_buf_delwri_submit_nowait() argument 1862 return __xfs_buf_delwri_submit(buffer_list, &io_list, false); in xfs_buf_delwri_submit_nowait() 1875 struct list_head *buffer_list) in xfs_buf_delwri_submit() argument 1881 __xfs_buf_delwri_submit(buffer_list, &io_list, true); in xfs_buf_delwri_submit()
|
D | xfs_inode_item.c | 409 struct list_head *buffer_list) in xfs_inode_item_push() argument 457 if (!xfs_buf_delwri_queue(bp, buffer_list)) in xfs_inode_item_push()
|
D | xfs_buf_item.c | 513 struct list_head *buffer_list) in xfs_buf_item_push() argument 546 if (!xfs_buf_delwri_queue(bp, buffer_list)) in xfs_buf_item_push()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_mr.c | 133 struct ib_phys_buf *buffer_list, in ipath_reg_phys_mr() argument 158 mr->mr.map[m]->segs[n].vaddr = (void *) buffer_list[i].addr; in ipath_reg_phys_mr() 159 mr->mr.map[m]->segs[n].length = buffer_list[i].size; in ipath_reg_phys_mr() 160 mr->mr.length += buffer_list[i].size; in ipath_reg_phys_mr()
|
D | ipath_verbs.h | 823 struct ib_phys_buf *buffer_list,
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2_provider.c | 333 struct ib_phys_buf *buffer_list, in c2_reg_phys_mr() argument 354 if (buffer_list[i].addr & ~PAGE_MASK) { in c2_reg_phys_mr() 356 (unsigned int) buffer_list[i].addr); in c2_reg_phys_mr() 360 if (!buffer_list[i].size) { in c2_reg_phys_mr() 365 total_len += buffer_list[i].size; in c2_reg_phys_mr() 366 pbl_depth += ALIGN(buffer_list[i].size, in c2_reg_phys_mr() 381 naddrs = ALIGN(buffer_list[i].size, in c2_reg_phys_mr() 384 page_list[j++] = (buffer_list[i].addr + in c2_reg_phys_mr()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_mr.c | 183 struct ib_phys_buf *buffer_list, in qib_reg_phys_mr() argument 203 mr->mr.map[m]->segs[n].vaddr = (void *) buffer_list[i].addr; in qib_reg_phys_mr() 204 mr->mr.map[m]->segs[n].length = buffer_list[i].size; in qib_reg_phys_mr() 205 mr->mr.length += buffer_list[i].size; in qib_reg_phys_mr()
|
D | qib_verbs.h | 1025 struct ib_phys_buf *buffer_list,
|
/linux-4.1.27/drivers/media/pci/sta2x11/ |
D | sta2x11_vip.c | 146 struct list_head buffer_list; member 250 if (list_empty(&vip->buffer_list)) {/* No available buffer */ in vip_active_buf_next() 254 vip->active = list_first_entry(&vip->buffer_list, in vip_active_buf_next() 320 list_add_tail(&vip_buf->list, &vip->buffer_list); in buffer_queue() 322 vip->active = list_first_entry(&vip->buffer_list, in buffer_queue() 372 list_for_each_entry_safe(vip_buf, node, &vip->buffer_list, list) { in stop_streaming() 870 INIT_LIST_HEAD(&vip->buffer_list); in sta2x11_vip_init_buffer()
|
/linux-4.1.27/include/linux/iio/ |
D | buffer.h | 92 struct list_head buffer_list; member
|
D | iio.h | 484 struct list_head buffer_list; member
|
/linux-4.1.27/fs/xfs/libxfs/ |
D | xfs_ialloc.h | 159 struct list_head *buffer_list,
|
D | xfs_bmap_btree.c | 864 struct list_head *buffer_list) in xfs_bmbt_change_owner() argument 869 ASSERT(tp || buffer_list); in xfs_bmbt_change_owner() 870 ASSERT(!(tp && buffer_list)); in xfs_bmbt_change_owner() 880 error = xfs_btree_change_owner(cur, new_owner, buffer_list); in xfs_bmbt_change_owner()
|
D | xfs_bmap_btree.h | 138 struct list_head *buffer_list);
|
D | xfs_btree.h | 397 struct list_head *buffer_list);
|
D | xfs_btree.c | 3981 struct list_head *buffer_list) in xfs_btree_block_change_owner() argument 4009 xfs_buf_delwri_queue(bp, buffer_list); in xfs_btree_block_change_owner() 4028 struct list_head *buffer_list) in xfs_btree_change_owner() argument 4059 buffer_list); in xfs_btree_change_owner()
|
D | xfs_ialloc.c | 222 struct list_head *buffer_list, in xfs_ialloc_inode_init() argument 342 xfs_buf_delwri_queue(fbuf, buffer_list); in xfs_ialloc_inode_init()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.h | 89 struct ib_phys_buf *buffer_list,
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
D | nes_verbs.c | 2093 struct ib_phys_buf *buffer_list, int num_phys_buf, int acc, in nes_reg_phys_mr() argument 2135 if ((buffer_list[0].addr ^ *iova_start) & ~PAGE_MASK) in nes_reg_phys_mr() 2204 mask = !buffer_list[i].size; in nes_reg_phys_mr() 2206 mask |= buffer_list[i].addr; in nes_reg_phys_mr() 2208 mask |= buffer_list[i].addr + buffer_list[i].size; in nes_reg_phys_mr() 2218 region_length += buffer_list[i].size; in nes_reg_phys_mr() 2220 if ((buffer_list[i-1].addr+PAGE_SIZE) != buffer_list[i].addr) in nes_reg_phys_mr() 2223 vpbl.pbl_vbase[cur_pbl_index].pa_low = cpu_to_le32((u32)buffer_list[i].addr & PAGE_MASK); in nes_reg_phys_mr() 2225 cpu_to_le32((u32)((((u64)buffer_list[i].addr) >> 32))); in nes_reg_phys_mr() 2247 buffer_list[0].addr, pbl_count, (u16)cur_pbl_index, acc, iova_start, in nes_reg_phys_mr()
|
/linux-4.1.27/drivers/tty/ |
D | synclink.c | 220 char *buffer_list; /* virtual address of Rx & Tx buffer lists */ member 3697 info->buffer_list = info->memory_base + info->last_mem_alloc; in mgsl_alloc_buffer_list_memory() 3707 …info->buffer_list = dma_alloc_coherent(NULL, BUFFERLISTSIZE, &info->buffer_list_dma_addr, GFP_KERN… in mgsl_alloc_buffer_list_memory() 3708 if (info->buffer_list == NULL) in mgsl_alloc_buffer_list_memory() 3715 memset( info->buffer_list, 0, BUFFERLISTSIZE ); in mgsl_alloc_buffer_list_memory() 3720 info->rx_buffer_list = (DMABUFFERENTRY *)info->buffer_list; in mgsl_alloc_buffer_list_memory() 3721 info->tx_buffer_list = (DMABUFFERENTRY *)info->buffer_list; in mgsl_alloc_buffer_list_memory() 3777 if (info->buffer_list && info->bus_type != MGSL_BUS_TYPE_PCI) in mgsl_free_buffer_list_memory() 3778 dma_free_coherent(NULL, BUFFERLISTSIZE, info->buffer_list, info->buffer_list_dma_addr); in mgsl_free_buffer_list_memory() 3780 info->buffer_list = NULL; in mgsl_free_buffer_list_memory() [all …]
|
D | synclinkmp.c | 194 char *buffer_list; /* virtual address of Rx & Tx buffer lists */ member 3454 info->buffer_list = info->memory_base + info->port_array[0]->last_mem_alloc; in alloc_buf_list() 3458 memset(info->buffer_list, 0, BUFFERLISTSIZE); in alloc_buf_list() 3463 info->rx_buf_list = (SCADESC *)info->buffer_list; in alloc_buf_list() 3465 info->tx_buf_list = (SCADESC *)info->buffer_list; in alloc_buf_list() 3527 info->buffer_list = NULL; in free_dma_bufs()
|
/linux-4.1.27/drivers/firewire/ |
D | ohci.c | 143 struct list_head buffer_list; member 1075 desc = list_entry(ctx->buffer_list.next, in context_tasklet() 1103 list_move_tail(&old_desc->list, &ctx->buffer_list); in context_tasklet() 1137 list_add_tail(&desc->list, &ctx->buffer_list); in context_add_buffer() 1150 INIT_LIST_HEAD(&ctx->buffer_list); in context_init() 1154 ctx->buffer_tail = list_entry(ctx->buffer_list.next, in context_init() 1181 list_for_each_entry_safe(desc, tmp, &ctx->buffer_list, list) in context_release() 1201 if (desc->list.next == &ctx->buffer_list) { in context_get_descriptors()
|