/linux-4.1.27/drivers/infiniband/core/ |
H A D | umem.c | 54 ib_dma_unmap_sg(dev, umem->sg_head.sgl, __ib_umem_release() 58 for_each_sg(umem->sg_head.sgl, sg, umem->npages, i) { __ib_umem_release() 66 sg_free_table(&umem->sg_head); __ib_umem_release() 183 ret = sg_alloc_table(&umem->sg_head, npages, GFP_KERNEL); ib_umem_get() 188 sg_list_start = umem->sg_head.sgl; ib_umem_get() 215 umem->sg_head.sgl, 326 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) ib_umem_page_count() 355 ret = sg_pcopy_to_buffer(umem->sg_head.sgl, umem->nmap, dst, length, ib_umem_copy_from()
|
/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx.c | 80 * @sg_head: pointer to the first scatter list element to build 85 * This function will start writing nx_sg elements at @sg_head and keep 91 struct nx_sg *nx_build_sg_list(struct nx_sg *sg_head, nx_build_sg_list() argument 121 for (sg = sg_head; sg_len < *len; sg++) { nx_build_sg_list() 139 if ((sg - sg_head) == sgmax) { nx_build_sg_list() 148 /* return the moved sg_head pointer */ nx_build_sg_list()
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 75 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); mlx4_ib_db_map_user()
|
H A D | mr.c | 105 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { mlx4_ib_umem_write_mtt()
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 77 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); mlx5_ib_db_map_user()
|
H A D | mem.c | 78 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { mlx5_ib_cont_pages() 180 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { __mlx5_ib_populate_pas()
|
/linux-4.1.27/include/rdma/ |
H A D | ib_umem.h | 55 struct sg_table sg_head; member in struct:ib_umem
|
/linux-4.1.27/drivers/scsi/ |
H A D | advansys.c | 374 ASC_SG_HEAD *sg_head; member in struct:asc_scsi_q 383 ASC_SG_HEAD *sg_head; member in struct:asc_scsi_req_q 394 ASC_SG_HEAD *sg_head; member in struct:asc_scsi_bios_req_q 2613 printk(" cdbptr 0x%lx, cdb_len %u, sg_head 0x%lx, sg_queue_cnt %u\n", asc_prt_asc_scsi_q() 2615 (ulong)q->sg_head, q->q1.sg_queue_cnt); asc_prt_asc_scsi_q() 2617 if (q->sg_head) { asc_prt_asc_scsi_q() 2618 sgp = q->sg_head; asc_prt_asc_scsi_q() 6862 ASC_SG_HEAD *sg_head; /* Ptr to driver SG request. */ AscIsrChipHalted() local 6908 sg_head = scsiq->sg_head; AscIsrChipHalted() 6941 for (i = 0; i < sg_head->queue_cnt; i++) { AscIsrChipHalted() 6983 (uchar *)&sg_head-> AscIsrChipHalted() 7927 asc_sg_head = kzalloc(sizeof(asc_scsi_q->sg_head) + asc_build_req() 7936 asc_scsi_q->sg_head = asc_sg_head; asc_build_req() 8359 ASC_SG_HEAD *sg_head; AscPutReadySgListQueue() local 8371 sg_head = scsiq->sg_head; AscPutReadySgListQueue() 8374 scsiq->q1.data_addr = (ASC_PADDR) sg_head->sg_list[0].addr; AscPutReadySgListQueue() 8375 scsiq->q1.data_cnt = (ASC_DCNT) sg_head->sg_list[0].bytes; AscPutReadySgListQueue() 8378 * If sg_head->entry_cnt is greater than ASC_MAX_SG_LIST AscPutReadySgListQueue() 8383 if (sg_head->entry_cnt > ASC_MAX_SG_LIST) { AscPutReadySgListQueue() 8398 sg_head->entry_cnt - ASC_MAX_SG_LIST; AscPutReadySgListQueue() 8406 sg_entry_cnt = sg_head->entry_cnt - 1; AscPutReadySgListQueue() 8414 scsiq->q1.sg_queue_cnt = sg_head->queue_cnt; AscPutReadySgListQueue() 8417 for (i = 0; i < sg_head->queue_cnt; i++) { AscPutReadySgListQueue() 8441 if (sg_head->entry_cnt > ASC_MAX_SG_LIST) { AscPutReadySgListQueue() 8473 (uchar *)&sg_head-> AscPutReadySgListQueue() 8508 scsiq->sg_head->queue_cnt = n_q_required - 1; AscSendScsiQueue() 8564 ASC_SG_HEAD *sg_head; AscExeScsiQueue() local 8568 sg_head = scsiq->sg_head; AscExeScsiQueue() 8600 if ((sg_entry_cnt = sg_head->entry_cnt) == 0) { AscExeScsiQueue() 8612 (ADV_PADDR)sg_head->sg_list[0].addr; AscExeScsiQueue() 8614 (ADV_DCNT)sg_head->sg_list[0].bytes; AscExeScsiQueue() 8627 (ADV_DCNT)le32_to_cpu(sg_head->sg_list[i]. AscExeScsiQueue() 8666 (ADV_PADDR)le32_to_cpu(sg_head-> AscExeScsiQueue() 8670 (ADV_DCNT)le32_to_cpu(sg_head-> AscExeScsiQueue() 8687 le32_to_cpu(sg_head-> AscExeScsiQueue() 8693 sg_head-> AscExeScsiQueue() 8702 sg_head->entry_to_copy = sg_head->entry_cnt; AscExeScsiQueue() 8918 kfree(asc_scsi_q.sg_head); asc_execute_scsi_cmnd()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_mr.c | 224 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { ipath_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_mr.c | 269 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { qib_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
H A D | c2_provider.c | 464 for_each_sg(c2mr->umem->sg_head.sgl, sg, c2mr->umem->nmap, entry) { c2_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
H A D | mem.c | 754 for_each_sg(mhp->umem->sg_head.sgl, sg, mhp->umem->nmap, entry) { c4iw_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 1031 for_each_sg(mr->umem->sg_head.sgl, sg, mr->umem->nmap, entry) { mthca_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 661 for_each_sg(mhp->umem->sg_head.sgl, sg, mhp->umem->nmap, entry) { iwch_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
H A D | ehca_mrmw.c | 403 pginfo.u.usr.next_sg = pginfo.u.usr.region->sg_head.sgl; ehca_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.c | 2385 for_each_sg(region->sg_head.sgl, sg, region->nmap, entry) { nes_reg_user_mr() 2615 for_each_sg(region->sg_head.sgl, sg, region->nmap, entry) { nes_reg_user_mr()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 853 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { build_user_pbes()
|