/linux-4.4.14/block/ |
D | bsg-lib.c | 39 kfree(job->request_payload.sg_list); in bsg_destroy_job() 40 kfree(job->reply_payload.sg_list); in bsg_destroy_job() 96 buf->sg_list = kzalloc(sz, GFP_KERNEL); in bsg_map_buffer() 97 if (!buf->sg_list) in bsg_map_buffer() 99 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer() 100 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer() 148 kfree(job->request_payload.sg_list); in bsg_create_job()
|
/linux-4.4.14/drivers/virt/ |
D | fsl_hypervisor.c | 153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local 243 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy() 265 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 266 sg_list[0].target = param.remote_paddr; in ioctl_memcpy() 268 sg_list[0].source = param.remote_paddr; in ioctl_memcpy() 269 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 271 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy() 273 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy() 274 count = param.count - sg_list[0].size; in ioctl_memcpy() 279 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy() [all …]
|
/linux-4.4.14/drivers/staging/rdma/ehca/ |
D | ehca_reqs.c | 85 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); in ehca_write_rwqe() 91 wqe_p->u.all_rcv.sg_list[cnt_ds].vaddr = in ehca_write_rwqe() 92 recv_wr->sg_list[cnt_ds].addr; in ehca_write_rwqe() 93 wqe_p->u.all_rcv.sg_list[cnt_ds].lkey = in ehca_write_rwqe() 94 recv_wr->sg_list[cnt_ds].lkey; in ehca_write_rwqe() 95 wqe_p->u.all_rcv.sg_list[cnt_ds].length = in ehca_write_rwqe() 96 recv_wr->sg_list[cnt_ds].length; in ehca_write_rwqe() 119 struct ib_sge *sge = ud_wr->wr.sg_list; in trace_ud_wr() 175 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); in ehca_write_swqe() 248 wqe_p->u.ud_av.sg_list[idx].vaddr = in ehca_write_swqe() [all …]
|
D | ehca_qes.h | 156 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member 164 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member 168 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES - member 176 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member
|
D | ehca_qp.c | 370 u.nud.sg_list[act_nr_sge]); in ehca_calc_wqe_size()
|
/linux-4.4.14/drivers/scsi/qla2xxx/ |
D | qla_bsg.c | 41 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 46 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 49 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 52 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 183 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 210 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 336 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 343 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() 395 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 397 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() [all …]
|
D | qla_iocb.c | 2042 (bsg_job->request_payload.sg_list))); in qla24xx_els_iocb() 2044 (bsg_job->request_payload.sg_list))); in qla24xx_els_iocb() 2046 (bsg_job->request_payload.sg_list)); in qla24xx_els_iocb() 2049 (bsg_job->reply_payload.sg_list))); in qla24xx_els_iocb() 2051 (bsg_job->reply_payload.sg_list))); in qla24xx_els_iocb() 2053 (bsg_job->reply_payload.sg_list)); in qla24xx_els_iocb() 2090 (bsg_job->request_payload.sg_list))); in qla2x00_ct_iocb() 2092 (bsg_job->request_payload.sg_list))); in qla2x00_ct_iocb() 2096 (bsg_job->reply_payload.sg_list))); in qla2x00_ct_iocb() 2098 (bsg_job->reply_payload.sg_list))); in qla2x00_ct_iocb() [all …]
|
D | qla_mr.c | 3331 for_each_sg(bsg_job->request_payload.sg_list, sg, in qlafx00_fxdisc_iocb() 3394 for_each_sg(bsg_job->reply_payload.sg_list, sg, in qlafx00_fxdisc_iocb()
|
/linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
D | iwch_qp.c | 73 if ((plen + wr->sg_list[i].length) < plen) in build_rdma_send() 76 plen += wr->sg_list[i].length; in build_rdma_send() 77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_send() 78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); in build_rdma_send() 79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); in build_rdma_send() 110 if ((plen + wr->sg_list[i].length) < plen) { in build_rdma_write() 113 plen += wr->sg_list[i].length; in build_rdma_write() 115 cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_write() 117 cpu_to_be32(wr->sg_list[i].length); in build_rdma_write() 119 cpu_to_be64(wr->sg_list[i].addr); in build_rdma_write() [all …]
|
/linux-4.4.14/arch/blackfin/kernel/ |
D | dma-mapping.c | 116 dma_map_sg(struct device *dev, struct scatterlist *sg_list, int nents, in dma_map_sg() argument 122 for_each_sg(sg_list, sg, nents, i) { in dma_map_sg() 131 void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg_list, in dma_sync_sg_for_device() argument 137 for_each_sg(sg_list, sg, nelems, i) { in dma_sync_sg_for_device()
|
/linux-4.4.14/drivers/infiniband/hw/qib/ |
D | qib_ruc.c | 93 ss->sg_list = qp->r_sg_list; in qib_init_sge() 96 if (wqe->sg_list[i].length == 0) in qib_init_sge() 99 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in qib_init_sge() 100 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) in qib_init_sge() 102 qp->r_len += wqe->sg_list[i].length; in qib_init_sge() 112 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in qib_init_sge() 428 sqp->s_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback() 429 sqp->s_sge.sg_list = wqe->sg_list + 1; in qib_ruc_loopback() 466 qp->r_sge.sg_list = NULL; in qib_ruc_loopback() 480 sqp->s_sge.sg_list = NULL; in qib_ruc_loopback() [all …]
|
D | qib_ud.c | 172 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback() 173 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 190 *sge = *ssge.sg_list++; in qib_ud_loopback() 320 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req() 321 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
|
D | qib_srq.c | 83 wqe->sg_list[i] = wr->sg_list[i]; in qib_post_srq_receive() 289 p->sg_list[i] = wqe->sg_list[i]; in qib_modify_srq()
|
D | qib_uc.c | 99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req() 100 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req() 434 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
|
D | qib_verbs.c | 189 *sge = *ss->sg_list++; in qib_copy_sge() 230 *sge = *ss->sg_list++; in qib_skip_sge() 253 struct qib_sge *sg_list = ss->sg_list; in qib_count_sge() local 277 sge = *sg_list++; in qib_count_sge() 315 *sge = *ss->sg_list++; in qib_copy_from_sge() 383 wr->sg_list[0].length < sizeof(u64) || in qib_post_one_send() 384 wr->sg_list[0].addr & (sizeof(u64) - 1))) in qib_post_one_send() 423 u32 length = wr->sg_list[i].length; in qib_post_one_send() 428 ok = qib_lkey_ok(rkt, pd, &wqe->sg_list[j], in qib_post_one_send() 429 &wr->sg_list[i], acc); in qib_post_one_send() [all …]
|
D | qib_rc.c | 49 ss->sge = wqe->sg_list[0]; in restart_sge() 50 ss->sg_list = wqe->sg_list + 1; in restart_sge() 489 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_rc_req() 490 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_rc_req() 1015 struct qib_sge *sge = &wqe->sg_list[i]; in qib_rc_send_complete() 1071 struct qib_sge *sge = &wqe->sg_list[i]; in do_rc_completion() 1207 u64 *vaddr = wqe->sg_list[0].vaddr; in do_rc_ack() 2065 qp->r_sge.sg_list = NULL; in qib_rc_rcv()
|
D | qib_verbs.h | 354 struct qib_sge sg_list[0]; member 365 struct ib_sge sg_list[0]; member 399 struct qib_sge *sg_list; /* next SGE to be used if any */ member 1083 ss->sge = *ss->sg_list++; in qib_put_ss()
|
D | qib_sdma.c | 622 *sge = *ss->sg_list++; in qib_sdma_verbs_send()
|
D | qib_qp.c | 433 struct qib_sge *sge = &wqe->sg_list[i]; in clear_mr_refs()
|
/linux-4.4.14/drivers/staging/rdma/ipath/ |
D | ipath_ruc.c | 130 if (wqe->sg_list[i].length == 0) in ipath_init_sge() 133 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, in ipath_init_sge() 134 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) in ipath_init_sge() 136 *lengthp += wqe->sg_list[i].length; in ipath_init_sge() 209 qp->r_sge.sg_list = qp->r_sg_list; in ipath_get_rwqe() 327 sqp->s_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback() 328 sqp->s_sge.sg_list = wqe->sg_list + 1; in ipath_ruc_loopback() 369 qp->r_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback() 370 qp->r_sge.sg_list = wqe->sg_list + 1; in ipath_ruc_loopback() 413 *sge = *sqp->s_sge.sg_list++; in ipath_ruc_loopback()
|
D | ipath_srq.c | 83 wqe->sg_list[i] = wr->sg_list[i]; in ipath_post_srq_receive() 293 p->sg_list[i] = wqe->sg_list[i]; in ipath_modify_srq()
|
D | ipath_ud.c | 133 rsge.sg_list = qp->r_ud_sg_list; in ipath_ud_loopback() 183 sge = swqe->sg_list; in ipath_ud_loopback() 322 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_ud_req() 323 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_ud_req()
|
D | ipath_uc.c | 98 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_uc_req() 99 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_uc_req() 453 qp->r_sge.sg_list = NULL; in ipath_uc_rcv()
|
D | ipath_keys.c | 226 ss->sg_list = NULL; in ipath_rkey_ok() 263 ss->sg_list = NULL; in ipath_rkey_ok()
|
D | ipath_verbs.c | 189 *sge = *ss->sg_list++; in ipath_copy_sge() 228 *sge = *ss->sg_list++; in ipath_skip_sge() 251 struct ipath_sge *sg_list = ss->sg_list; in ipath_count_sge() local 275 sge = *sg_list++; in ipath_count_sge() 314 *sge = *ss->sg_list++; in ipath_copy_from_sge() 383 wr->sg_list[0].length < sizeof(u64) || in ipath_post_one_send() 384 wr->sg_list[0].addr & (sizeof(u64) - 1))) in ipath_post_one_send() 417 u32 length = wr->sg_list[i].length; in ipath_post_one_send() 422 ok = ipath_lkey_ok(qp, &wqe->sg_list[j], in ipath_post_one_send() 423 &wr->sg_list[i], acc); in ipath_post_one_send() [all …]
|
D | ipath_rc.c | 48 ss->sge = wqe->sg_list[0]; in restart_sge() 49 ss->sg_list = wqe->sg_list + 1; in restart_sge() 461 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_rc_req() 462 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_rc_req() 936 *(u64 *) wqe->sg_list[0].vaddr = val; in do_rc_ack() 1452 e->rdma_sge.sg_list = NULL; in ipath_rc_rcv_error() 1779 qp->r_sge.sg_list = NULL; in ipath_rc_rcv() 1839 e->rdma_sge.sg_list = NULL; in ipath_rc_rcv()
|
D | ipath_verbs.h | 291 struct ipath_sge sg_list[0]; member 302 struct ib_sge sg_list[0]; member 335 struct ipath_sge *sg_list; /* next SGE to be used if any */ member
|
D | ipath_sdma.c | 764 *sge = *ss->sg_list++; in ipath_sdma_verbs_send()
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
D | ruc.c | 111 ss->sg_list = qp->r_sg_list; in init_sge() 114 if (wqe->sg_list[i].length == 0) in init_sge() 117 if (!hfi1_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in init_sge() 118 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) in init_sge() 120 qp->r_len += wqe->sg_list[i].length; in init_sge() 130 struct hfi1_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in init_sge() 450 sqp->s_sge.sge = wqe->sg_list[0]; in ruc_loopback() 451 sqp->s_sge.sg_list = wqe->sg_list + 1; in ruc_loopback() 488 qp->r_sge.sg_list = NULL; in ruc_loopback() 502 sqp->s_sge.sg_list = NULL; in ruc_loopback() [all …]
|
D | srq.c | 100 wqe->sg_list[i] = wr->sg_list[i]; in hfi1_post_srq_receive() 306 p->sg_list[i] = wqe->sg_list[i]; in hfi1_modify_srq()
|
D | ud.c | 194 ssge.sg_list = swqe->sg_list + 1; in ud_loopback() 195 ssge.sge = *swqe->sg_list; in ud_loopback() 212 *sge = *ssge.sg_list++; in ud_loopback() 350 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_ud_req() 351 qp->s_sge.sg_list = wqe->sg_list + 1; in hfi1_make_ud_req()
|
D | verbs.c | 299 *sge = *ss->sg_list++; in hfi1_copy_sge() 340 *sge = *ss->sg_list++; in hfi1_skip_sge() 401 wr->sg_list[0].length < sizeof(u64) || in post_one_send() 402 wr->sg_list[0].addr & (sizeof(u64) - 1))) in post_one_send() 437 u32 length = wr->sg_list[i].length; in post_one_send() 442 ok = hfi1_lkey_ok(rkt, pd, &wqe->sg_list[j], in post_one_send() 443 &wr->sg_list[i], acc); in post_one_send() 468 struct hfi1_sge *sge = &wqe->sg_list[--j]; in post_one_send() 569 wqe->sg_list[i] = wr->sg_list[i]; in post_receive() 724 *sge = *ss->sg_list++; in update_sge() [all …]
|
D | uc.c | 117 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_uc_req() 118 qp->s_sge.sg_list = wqe->sg_list + 1; in hfi1_make_uc_req() 484 qp->r_sge.sg_list = NULL; in hfi1_uc_rcv()
|
D | verbs.h | 361 struct hfi1_sge sg_list[0]; member 372 struct ib_sge sg_list[0]; member 406 struct hfi1_sge *sg_list; /* next SGE to be used if any */ member 1055 ss->sge = *ss->sg_list++; in hfi1_put_ss()
|
D | rc.c | 69 ss->sge = wqe->sg_list[0]; in restart_sge() 70 ss->sg_list = wqe->sg_list + 1; in restart_sge() 517 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_rc_req() 518 qp->s_sge.sg_list = wqe->sg_list + 1; in hfi1_make_rc_req() 1029 struct hfi1_sge *sge = &wqe->sg_list[i]; in hfi1_rc_send_complete() 1086 struct hfi1_sge *sge = &wqe->sg_list[i]; in do_rc_completion() 1238 u64 *vaddr = wqe->sg_list[0].vaddr; in do_rc_ack() 2153 qp->r_sge.sg_list = NULL; in hfi1_rc_rcv()
|
D | qp.c | 418 struct hfi1_sge *sge = &wqe->sg_list[i]; in clear_mr_refs()
|
/linux-4.4.14/drivers/dma/ |
D | imx-dma.c | 165 struct scatterlist *sg_list; member 804 kfree(imxdmac->sg_list); in imxdma_free_chan_resources() 805 imxdmac->sg_list = NULL; in imxdma_free_chan_resources() 879 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic() 881 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic() 883 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic() 886 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic() 889 imxdmac->sg_list[i].page_link = 0; in imxdma_prep_dma_cyclic() 890 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic() 891 imxdmac->sg_list[i].dma_address = dma_addr; in imxdma_prep_dma_cyclic() [all …]
|
/linux-4.4.14/drivers/crypto/caam/ |
D | sg_sw_sec4.h | 73 static inline int sg_count(struct scatterlist *sg_list, int nbytes) in sg_count() argument 75 int sg_nents = sg_nents_for_len(sg_list, nbytes); in sg_count()
|
/linux-4.4.14/drivers/scsi/aacraid/ |
D | commctrl.c | 480 void *sg_list[32]; in aac_send_raw_srb() local 507 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb() 560 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb() 619 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 671 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 725 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 761 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 804 if(copy_to_user(sg_user[i], sg_list[i], byte_count)){ in aac_send_raw_srb() 823 kfree(sg_list[i]); in aac_send_raw_srb()
|
/linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 398 if ((plen + wr->sg_list[i].length) > max) in build_immd() 400 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd() 401 plen += wr->sg_list[i].length; in build_immd() 402 rem = wr->sg_list[i].length; in build_immd() 428 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl() argument 437 if ((plen + sg_list[i].length) < plen) in build_isgl() 439 plen += sg_list[i].length; in build_isgl() 440 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | in build_isgl() 441 sg_list[i].length); in build_isgl() 444 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl() [all …]
|
/linux-4.4.14/include/linux/ |
D | bsg-lib.h | 36 struct scatterlist *sg_list; member
|
D | agp_backend.h | 85 struct scatterlist *sg_list; member
|
D | efi.h | 527 unsigned long sg_list);
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | fsl_hcalls.h | 358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument 370 r5 = (uint32_t) sg_list; in fh_partition_memcpy() 373 r6 = sg_list >> 32; in fh_partition_memcpy()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.h | 120 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member 129 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member 146 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
|
D | ehea_main.c | 503 rwqe->sg_list[0].l_key = pr->recv_mr.lkey; in ehea_refill_rq_def() 504 rwqe->sg_list[0].vaddr = tmp_addr; in ehea_refill_rq_def() 505 rwqe->sg_list[0].len = packet_size; in ehea_refill_rq_def() 1672 struct ehea_vsgentry *sg_list, *sg1entry, *sgentry; in write_swqe2_data() local 1678 sg_list = (struct ehea_vsgentry *)&swqe->u.immdata_desc.sg_list; in write_swqe2_data() 1701 sgentry = &sg_list[i - sg1entry_contains_frag_data]; in write_swqe2_data() 2628 rwqe->sg_list[0].l_key = lkey; in ehea_update_rqs() 2632 rwqe->sg_list[0].vaddr = ehea_map_vaddr(skb->data); in ehea_update_rqs() 2637 rwqe->sg_list[0].l_key = lkey; in ehea_update_rqs() 2641 rwqe->sg_list[0].vaddr = ehea_map_vaddr(skb->data); in ehea_update_rqs()
|
/linux-4.4.14/drivers/staging/rdma/amso1100/ |
D | c2_qp.c | 845 ib_wr->sg_list, in c2_post_send() 868 ib_wr->sg_list, in c2_post_send() 888 cpu_to_be32(ib_wr->sg_list->lkey); in c2_post_send() 890 cpu_to_be64(ib_wr->sg_list->addr); in c2_post_send() 896 cpu_to_be32(ib_wr->sg_list->length); in c2_post_send() 980 ib_wr->sg_list, in c2_post_receive()
|
/linux-4.4.14/drivers/firmware/efi/ |
D | runtime-wrappers.c | 258 unsigned long sg_list) in virt_efi_update_capsule() argument 267 status = efi_call_virt(update_capsule, capsules, count, sg_list); in virt_efi_update_capsule()
|
/linux-4.4.14/drivers/xen/ |
D | efi.c | 223 unsigned long sg_list) in xen_efi_update_capsule() argument 233 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
|
/linux-4.4.14/drivers/scsi/qla4xxx/ |
D | ql4_bsg.c | 63 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_flash() 123 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_flash() 187 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_get_acb_state() 258 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_nvram() 322 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_nvram() 436 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_bsg_get_acb()
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
D | srq.c | 354 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv() 355 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv() 356 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv()
|
D | mad.c | 592 wr.wr.sg_list = &list; in mlx4_ib_send_to_slave() 1139 struct ib_sge sg_list; in mlx4_ib_post_pv_qp_buf() local 1146 sg_list.addr = tun_qp->ring[index].map; in mlx4_ib_post_pv_qp_buf() 1147 sg_list.length = size; in mlx4_ib_post_pv_qp_buf() 1148 sg_list.lkey = ctx->pd->local_dma_lkey; in mlx4_ib_post_pv_qp_buf() 1151 recv_wr.sg_list = &sg_list; in mlx4_ib_post_pv_qp_buf() 1269 wr.wr.sg_list = &list; in mlx4_ib_send_to_wire()
|
D | qp.c | 2164 send_size += wr->wr.sg_list[i].length; in build_sriov_qp0_header() 2287 send_size += wr->wr.sg_list[i].length; in build_mlx_header() 2990 set_data_seg(dseg, wr->sg_list + i); in mlx4_ib_post_send() 3116 scat->lkey = cpu_to_be32(wr->sg_list->lkey); in mlx4_ib_post_recv() 3123 __set_data_seg(scat + i, wr->sg_list + i); in mlx4_ib_post_recv()
|
/linux-4.4.14/drivers/scsi/ |
D | 3w-9xxx.c | 1348 …if (full_command_packet->command.newcommand.sg_list[0].length < scsi_bufflen(tw_dev->srb[request_i… in twa_interrupt() 1349 …scsi_set_resid(cmd, scsi_bufflen(cmd) - full_command_packet->command.newcommand.sg_list[0].length); in twa_interrupt() 1393 newcommand->sg_list[0].address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); in twa_load_sgl() 1394 newcommand->sg_list[0].length = cpu_to_le32(length); in twa_load_sgl() 1855 command_packet->sg_list[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); in DEF_SCSI_QCMD() 1856 command_packet->sg_list[0].length = cpu_to_le32(TW_MIN_SGL_LENGTH); in DEF_SCSI_QCMD() 1863 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sg_dma_address(sg)); in DEF_SCSI_QCMD() 1864 command_packet->sg_list[i].length = cpu_to_le32(sg_dma_len(sg)); in DEF_SCSI_QCMD() 1865 if (command_packet->sg_list[i].address & TW_CPU_TO_SGL(TW_ALIGNMENT_9000_SGL)) { in DEF_SCSI_QCMD() 1876 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sglistarg[i].address); in DEF_SCSI_QCMD() [all …]
|
D | hptiop.h | 222 struct hpt_iopsg sg_list[1]; member 239 struct hpt_iopsg sg_list[1]; member
|
D | 3w-sas.c | 345 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sg_dma_address(sg)); in twl_scsiop_execute_scsi() 346 command_packet->sg_list[i].length = TW_CPU_TO_SGL(sg_dma_len(sg)); in twl_scsiop_execute_scsi() 353 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sglistarg[i].address); in twl_scsiop_execute_scsi() 354 command_packet->sg_list[i].length = TW_CPU_TO_SGL(sglistarg[i].length); in twl_scsiop_execute_scsi() 697 newcommand->sg_list[0].address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); in twl_load_sgl() 698 newcommand->sg_list[0].length = TW_CPU_TO_SGL(length); in twl_load_sgl() 1220 …if (full_command_packet->command.newcommand.sg_list[0].length < scsi_bufflen(tw_dev->srb[request_i… in twl_interrupt() 1221 …scsi_set_resid(cmd, scsi_bufflen(cmd) - full_command_packet->command.newcommand.sg_list[0].length); in twl_interrupt()
|
D | ips.c | 1810 scb->sg_list.enh_list[indx].address_lo = in ips_fill_scb_sg_single() 1812 scb->sg_list.enh_list[indx].address_hi = in ips_fill_scb_sg_single() 1814 scb->sg_list.enh_list[indx].length = cpu_to_le32(e_len); in ips_fill_scb_sg_single() 1816 scb->sg_list.std_list[indx].address = in ips_fill_scb_sg_single() 1818 scb->sg_list.std_list[indx].length = cpu_to_le32(e_len); in ips_fill_scb_sg_single() 1834 IPS_SG_LIST sg_list; in ips_flash_firmware() local 1849 sg_list.list = scb->sg_list.list; in ips_flash_firmware() 1854 scb->sg_list.list = sg_list.list; in ips_flash_firmware() 1907 IPS_SG_LIST sg_list; in ips_usrcmd() local 1916 sg_list.list = scb->sg_list.list; in ips_usrcmd() [all …]
|
D | wd719x.h | 60 struct wd719x_sglist sg_list[WD719X_SG] __aligned(8); /* SG list */ member
|
D | 3w-xxxx.h | 386 TW_SG_Entry sg_list[TW_ATA_PASS_SGL_MAX]; member
|
D | eata_generic.h | 314 struct eata_sg_list *sg_list; member
|
D | 3w-sas.h | 252 TW_SG_Entry_ISO sg_list[TW_LIBERATOR_MAX_SGL_LENGTH]; member
|
D | wd719x.c | 274 offsetof(struct wd719x_scb, sg_list)); in wd719x_queuecommand() 277 scb->sg_list[i].ptr = cpu_to_le32(sg_dma_address(sg)); in wd719x_queuecommand() 278 scb->sg_list[i].length = cpu_to_le32(sg_dma_len(sg)); in wd719x_queuecommand()
|
D | advansys.c | 323 ASC_SG_LIST sg_list[0]; member 368 ASC_SG_LIST sg_list[7]; member 1737 } sg_list[NO_OF_SG_PER_BLOCK]; member 2529 i, (ulong)le32_to_cpu(sgp->sg_list[i].addr), in asc_prt_asc_scsi_q() 2530 (ulong)le32_to_cpu(sgp->sg_list[i].bytes)); in asc_prt_asc_scsi_q() 2568 i, (u32)le32_to_cpu(b->sg_list[i].sg_addr), in asc_prt_adv_sgblock() 2569 (u32)le32_to_cpu(b->sg_list[i].sg_count)); in asc_prt_adv_sgblock() 7603 asc_sg_head->sg_list[sgcnt].addr = in asc_build_req() 7605 asc_sg_head->sg_list[sgcnt].bytes = in asc_build_req() 7705 sg_block->sg_list[i].sg_addr = in adv_get_sglist() [all …]
|
D | dpt_i2o.c | 1713 void *sg_list[pHba->sg_tablesize]; in adpt_i2o_passthru() local 1757 memset(sg_list,0, sizeof(sg_list[0])*pHba->sg_tablesize); in adpt_i2o_passthru() 1785 sg_list[sg_index++] = p; // sglist indexed with input frame, not our internal frame. in adpt_i2o_passthru() 1857 if (copy_to_user((void __user *)(ulong)sg[j].addr_bus,sg_list[j], sg_size)) { in adpt_i2o_passthru() 1858 … printk(KERN_WARNING"%s: Could not copy %p TO user %x\n",pHba->name, sg_list[j], sg[j].addr_bus); in adpt_i2o_passthru() 1886 if(sg_list[--sg_index]) { in adpt_i2o_passthru() 1889 sg_list[sg_index], in adpt_i2o_passthru()
|
D | nsp32.c | 2660 data->sg_list = pci_alloc_consistent(pdev, NSP32_SG_TABLE_SIZE, in nsp32_detect() 2662 if (data->sg_list == NULL) { in nsp32_detect() 2676 .sglun = &(data->sg_list[offset]), in nsp32_detect() 2767 data->sg_list, data->sg_paddr); in nsp32_detect() 2789 if (data->sg_list) { in nsp32_release() 2791 data->sg_list, data->sg_paddr); in nsp32_release()
|
D | 3w-9xxx.h | 521 TW_SG_Entry sg_list[TW_APACHE_MAX_SGL_LENGTH]; member
|
D | eata.c | 742 struct sg_list { struct 811 struct sg_list *sglist; /* pointer to the allocated SG list */ 1346 size_t sz = shost->sg_tablesize *sizeof(struct sg_list); in port_detect() 1634 sizeof(struct sg_list), in map_dma() 1636 cpp->data_len = H2DEV((scsi_sg_count(SCpnt) * sizeof(struct sg_list))); in map_dma()
|
D | ips.h | 1100 IPS_SG_LIST sg_list; member 1126 IPS_SG_LIST *sg_list; member
|
D | u14-34f.c | 544 struct sg_list { struct 585 struct sg_list *sglist; /* pointer to the allocated SG list */ 983 sh[j]->sg_tablesize * sizeof(struct sg_list), in port_detect() 1143 cpp->use_sg * sizeof(struct sg_list), in map_dma()
|
D | nsp32.h | 575 nsp32_sglun *sg_list; /* sglist virtuxal address */ member
|
D | libiscsi_tcp.c | 369 struct scatterlist *sg_list, unsigned int sg_count, in iscsi_segment_seek_sg() argument 377 for_each_sg(sg_list, sg, sg_count, i) { in iscsi_segment_seek_sg()
|
D | scsi_transport_fc.c | 3575 kfree(job->request_payload.sg_list); in fc_destroy_bsgjob() 3576 kfree(job->reply_payload.sg_list); in fc_destroy_bsgjob() 3679 buf->sg_list = kzalloc(sz, GFP_KERNEL); in fc_bsg_map_buffer() 3680 if (!buf->sg_list) in fc_bsg_map_buffer() 3682 sg_init_table(buf->sg_list, req->nr_phys_segments); in fc_bsg_map_buffer() 3683 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in fc_bsg_map_buffer() 3756 kfree(job->request_payload.sg_list); in fc_req_to_bsgjob()
|
D | 3w-xxxx.c | 974 passthru->sg_list[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1; in tw_chrdev_ioctl() 975 passthru->sg_list[0].length = data_buffer_length_adjusted; in tw_chrdev_ioctl()
|
D | hptiop.c | 767 memcpy(scp->sense_buffer, &req->sg_list, SCSI_SENSE_BUFFERSIZE); in hptiop_finish_scsi_req() 1049 sg_count = hptiop_buildsgl(scp, req->sg_list); in hptiop_queuecommand_lck()
|
/linux-4.4.14/drivers/infiniband/core/ |
D | mad.c | 1030 mad_send_wr->sg_list[0].length = hdr_len; in ib_create_send_mad() 1031 mad_send_wr->sg_list[0].lkey = mad_agent->qp->pd->local_dma_lkey; in ib_create_send_mad() 1036 mad_send_wr->sg_list[1].length = data_len; in ib_create_send_mad() 1038 mad_send_wr->sg_list[1].length = mad_size - hdr_len; in ib_create_send_mad() 1040 mad_send_wr->sg_list[1].lkey = mad_agent->qp->pd->local_dma_lkey; in ib_create_send_mad() 1043 mad_send_wr->send_wr.wr.sg_list = mad_send_wr->sg_list; in ib_create_send_mad() 1158 sge = mad_send_wr->sg_list; in ib_send_mad() 2437 mad_send_wr->sg_list[0].length, DMA_TO_DEVICE); in ib_mad_send_done_handler() 2440 mad_send_wr->sg_list[1].length, DMA_TO_DEVICE); in ib_mad_send_done_handler() 2878 struct ib_sge sg_list; in ib_mad_post_receive_mads() local [all …]
|
D | mad_priv.h | 127 struct ib_sge sg_list[IB_MAD_SEND_REQ_MAX_SG]; member
|
D | uverbs_cmd.c | 2581 next->sg_list = (void *) next + in ib_uverbs_post_send() 2583 if (copy_from_user(next->sg_list, in ib_uverbs_post_send() 2593 next->sg_list = NULL; in ib_uverbs_post_send() 2682 next->sg_list = (void *) next + in ib_uverbs_unmarshall_recv() 2684 if (copy_from_user(next->sg_list, in ib_uverbs_unmarshall_recv() 2693 next->sg_list = NULL; in ib_uverbs_unmarshall_recv()
|
/linux-4.4.14/drivers/char/agp/ |
D | intel-gtt.c | 123 static void intel_gtt_unmap_memory(struct scatterlist *sg_list, int num_sg) in intel_gtt_unmap_memory() argument 128 pci_unmap_sg(intel_private.pcidev, sg_list, in intel_gtt_unmap_memory() 131 st.sgl = sg_list; in intel_gtt_unmap_memory() 919 mem->sg_list = st.sgl; in intel_fake_agp_insert_entries() 955 intel_gtt_unmap_memory(mem->sg_list, mem->num_sg); in intel_fake_agp_remove_entries() 956 mem->sg_list = NULL; in intel_fake_agp_remove_entries()
|
/linux-4.4.14/include/scsi/ |
D | libiscsi_tcp.h | 117 struct scatterlist *sg_list, unsigned int sg_count,
|
D | scsi_transport_fc.h | 630 struct scatterlist *sg_list; member
|
/linux-4.4.14/drivers/infiniband/ulp/ipoib/ |
D | ipoib_verbs.c | 225 priv->tx_wr.wr.sg_list = priv->tx_sge; in ipoib_transport_dev_init() 234 priv->rx_wr.sg_list = priv->rx_sge; in ipoib_transport_dev_init()
|
D | ipoib_cm.c | 342 wr->sg_list = sge; in ipoib_cm_init_rx_wr()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
D | srq.c | 458 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx5_ib_post_srq_recv() 459 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx5_ib_post_srq_recv() 460 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx5_ib_post_srq_recv()
|
D | qp.c | 2092 addr = (void *)(unsigned long)(wr->sg_list[i].addr); in set_data_inl_seg() 2093 len = wr->sg_list[i].length; in set_data_inl_seg() 2230 u32 data_len = wr->wr.sg_list->length; in set_sig_data_segment() 2231 u32 data_key = wr->wr.sg_list->lkey; in set_sig_data_segment() 2232 u64 data_va = wr->wr.sg_list->addr; in set_sig_data_segment() 2373 region_len = wr->wr.sg_list->length; in set_sig_umr_wr() 2375 (wr->prot->lkey != wr->wr.sg_list->lkey || in set_sig_umr_wr() 2376 wr->prot->addr != wr->wr.sg_list->addr || in set_sig_umr_wr() 2377 wr->prot->length != wr->wr.sg_list->length)) in set_sig_umr_wr() 2821 if (likely(wr->sg_list[i].length)) { in mlx5_ib_post_send() [all …]
|
D | mr.c | 710 wr->sg_list = sg; in prep_umr_reg_wqe() 942 wr.wr.sg_list = &sg; in mlx5_ib_update_mtt()
|
/linux-4.4.14/drivers/infiniband/hw/nes/ |
D | nes_verbs.c | 3233 ib_wr->sg_list[sge_index].addr); in fill_wqe_sg_send() 3235 ib_wr->sg_list[sge_index].length); in fill_wqe_sg_send() 3238 (ib_wr->sg_list[sge_index].lkey)); in fill_wqe_sg_send() 3242 total_payload_length += ib_wr->sg_list[sge_index].length; in fill_wqe_sg_send() 3326 (ib_wr->sg_list[0].length <= 64)) { in nes_post_send() 3328 (void *)(unsigned long)ib_wr->sg_list[0].addr, ib_wr->sg_list[0].length); in nes_post_send() 3330 ib_wr->sg_list[0].length); in nes_post_send() 3356 (ib_wr->sg_list[0].length <= 64)) { in nes_post_send() 3358 (void *)(unsigned long)ib_wr->sg_list[0].addr, ib_wr->sg_list[0].length); in nes_post_send() 3360 ib_wr->sg_list[0].length); in nes_post_send() [all …]
|
/linux-4.4.14/drivers/crypto/ |
D | atmel-aes.c | 164 struct scatterlist *sg_list; in atmel_aes_sg_length() local 167 sg_list = sg; in atmel_aes_sg_length() 171 len = min(sg_list->length, total); in atmel_aes_sg_length() 176 sg_list = sg_next(sg_list); in atmel_aes_sg_length() 177 if (!sg_list) in atmel_aes_sg_length()
|
D | bfin_crc.c | 102 static struct scatterlist *sg_get(struct scatterlist *sg_list, unsigned int nents, in sg_get() argument 108 for_each_sg(sg_list, sg, nents, i) in sg_get()
|
D | picoxcell_crypto.c | 276 static inline int sg_count(struct scatterlist *sg_list, int nbytes) in sg_count() argument 278 return sg_nents_for_len(sg_list, nbytes); in sg_count()
|
/linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 2015 struct ib_sge *sg_list) in ocrdma_build_sges() argument 2020 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges() 2021 sge[i].addr_lo = sg_list[i].addr; in ocrdma_build_sges() 2022 sge[i].addr_hi = upper_32_bits(sg_list[i].addr); in ocrdma_build_sges() 2023 sge[i].len = sg_list[i].length; in ocrdma_build_sges() 2024 hdr->total_len += sg_list[i].length; in ocrdma_build_sges() 2030 static inline uint32_t ocrdma_sglist_len(struct ib_sge *sg_list, int num_sge) in ocrdma_sglist_len() argument 2035 total_len += sg_list[i].length; in ocrdma_sglist_len() 2049 hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); in ocrdma_build_inline_sges() 2059 (void *)(unsigned long)wr->sg_list[i].addr, in ocrdma_build_inline_sges() [all …]
|
/linux-4.4.14/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 286 write_wr.wr.sg_list = &sge[0]; in send_write() 548 send_wr.sg_list = ctxt->sge; in send_reply()
|
D | svc_rdma_recvfrom.c | 189 read_wr.wr.sg_list = ctxt->sge; in rdma_read_chunk_lcl() 316 read_wr.wr.sg_list = ctxt->sge; in rdma_read_chunk_frmr()
|
D | verbs.c | 1272 send_wr.sg_list = iov; in rpcrdma_ep_post() 1310 recv_wr.sg_list = &rep->rr_rdmabuf->rg_iov; in rpcrdma_ep_post_recv()
|
D | svc_rdma_transport.c | 579 recv_wr.sg_list = &ctxt->sge[0]; in svc_rdma_post_recv() 1351 err_wr.sg_list = ctxt->sge; in svc_rdma_send_error()
|
/linux-4.4.14/net/9p/ |
D | trans_rdma.c | 418 wr.sg_list = &sge; in post_recv() 511 wr.sg_list = &sge; in rdma_request()
|
/linux-4.4.14/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 988 rx_wr->sg_list = &rx_desc->rx_sg; in isert_post_recvm() 1010 rx_wr.sg_list = &rx_desc->rx_sg; in isert_post_recv() 1033 send_wr.sg_list = tx_desc->tx_sg; in isert_post_send() 1104 send_wr->sg_list = &tx_desc->tx_sg[0]; in isert_init_send_wr() 1126 rx_wr.sg_list = &sge; in isert_rdma_post_recvl() 2383 rdma_wr->wr.sg_list = ib_sge; in isert_build_rdma_wr() 2411 rdma_wr->wr.sg_list, rdma_wr->wr.num_sge); in isert_build_rdma_wr() 2674 sig_wr.wr.sg_list = &rdma_wr->ib_sg[DATA]; in isert_reg_sig_mr() 2826 rdma_wr->wr.sg_list = &wr->s_ib_sge; in isert_reg_rdma()
|
/linux-4.4.14/drivers/infiniband/hw/mthca/ |
D | mthca_srq.c | 522 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_srq_recv() 614 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_srq_recv()
|
D | mthca_qp.c | 1737 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_send() 1862 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_receive() 2078 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_send() 2198 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_receive()
|
/linux-4.4.14/net/rds/ |
D | iw_recv.c | 95 recv->r_wr.sg_list = recv->r_sge; in rds_iw_recv_init_ring() 356 wr->sg_list = sge; in rds_iw_recv_init_ack()
|
D | ib_recv.c | 59 recv->r_wr.sg_list = recv->r_sge; in rds_ib_recv_init_ring() 569 wr->sg_list = sge; in rds_ib_recv_init_ack()
|
D | iw_send.c | 142 send->s_send_wr.sg_list = send->s_sge; in rds_iw_send_init_ring()
|
D | ib_send.c | 199 send->s_wr.sg_list = send->s_sge; in rds_ib_send_init_ring()
|
/linux-4.4.14/drivers/block/ |
D | skd_main.c | 4191 struct fit_sg_descriptor *sg_list; in skd_cons_sg_list() local 4194 nbytes = sizeof(*sg_list) * n_sg; in skd_cons_sg_list() 4196 sg_list = pci_alloc_consistent(skdev->pdev, nbytes, ret_dma_addr); in skd_cons_sg_list() 4198 if (sg_list != NULL) { in skd_cons_sg_list() 4202 memset(sg_list, 0, nbytes); in skd_cons_sg_list() 4208 sg_list[i].next_desc_ptr = dma_address + ndp_off; in skd_cons_sg_list() 4210 sg_list[i].next_desc_ptr = 0LL; in skd_cons_sg_list() 4213 return sg_list; in skd_cons_sg_list() 4570 struct fit_sg_descriptor *sg_list, in skd_free_sg_list() argument 4573 if (sg_list != NULL) { in skd_free_sg_list() [all …]
|
/linux-4.4.14/drivers/scsi/lpfc/ |
D | lpfc_bsg.c | 235 sg_miter_start(&miter, bsg_buffers->sg_list, bsg_buffers->sg_cnt, in lpfc_bsg_copy_data() 613 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_rport_els_cmp() 712 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_rport_els() 1335 sg_copy_from_buffer(job->request_payload.sg_list, in lpfc_bsg_hba_get_event() 3076 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_diag_loopback_run() 3241 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_diag_loopback_run() 3369 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_issue_mbox_cmpl() 3563 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_issue_mbox_ext_handle_job() 4311 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_read_ebuf_get() 4363 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_write_ebuf_set() [all …]
|
/linux-4.4.14/drivers/infiniband/ulp/iser/ |
D | iser_verbs.c | 1058 rx_wr.sg_list = &sge; in iser_post_recvl() 1082 rx_wr->sg_list = &rx_desc->rx_sg; in iser_post_recvm() 1119 wr->sg_list = tx_desc->tx_sg; in iser_post_send()
|
D | iser_memory.c | 456 wr->wr.sg_list = &data_reg->sge; in iser_reg_sig_mr()
|
/linux-4.4.14/include/xen/interface/ |
D | platform.h | 199 uint64_t sg_list; /* machine address */ member
|
/linux-4.4.14/drivers/scsi/libfc/ |
D | fc_lport.c | 1985 sg_copy_to_buffer(job->request_payload.sg_list, in fc_lport_els_request() 2009 info->sg = job->reply_payload.sg_list; in fc_lport_els_request() 2046 sg_copy_to_buffer(job->request_payload.sg_list, in fc_lport_ct_request() 2070 info->sg = job->reply_payload.sg_list; in fc_lport_ct_request()
|
/linux-4.4.14/drivers/scsi/aic7xxx/ |
D | aic79xx_core.c | 467 sg = (struct ahd_dma64_seg *)scb->sg_list; in ahd_setup_data_scb() 474 sg = (struct ahd_dma_seg *)scb->sg_list; in ahd_setup_data_scb() 511 return ((uint8_t *)scb->sg_list + sg_offset); in ahd_sg_bus_to_virt() 520 sg_offset = ((uint8_t *)sg - (uint8_t *)scb->sg_list) in ahd_sg_virt_to_bus() 1741 struct ahd_dma64_seg *sg_list; in ahd_dump_sglist() local 1743 sg_list = (struct ahd_dma64_seg*)scb->sg_list; in ahd_dump_sglist() 1748 addr = ahd_le64toh(sg_list[i].addr); in ahd_dump_sglist() 1749 len = ahd_le32toh(sg_list[i].len); in ahd_dump_sglist() 1754 sg_list[i].len & AHD_SG_LEN_MASK, in ahd_dump_sglist() 1755 (sg_list[i].len & AHD_DMA_LAST_SEG) in ahd_dump_sglist() [all …]
|
D | aic7xxx_core.c | 380 return (&scb->sg_list[sg_index]); in ahc_sg_bus_to_virt() 389 sg_index = sg - &scb->sg_list[1]; in ahc_sg_virt_to_bus() 391 return (scb->sg_list_phys + (sg_index * sizeof(*scb->sg_list))); in ahc_sg_virt_to_bus() 417 /*offset*/(scb->sg_list - scb->sg_map->sg_vaddr) in ahc_sync_sglist() 1080 sg = scb->sg_list; in ahc_handle_seqint() 1411 (ahc_le32toh(scb->sg_list[i].len) >> 24 in ahc_handle_seqint() 1413 ahc_le32toh(scb->sg_list[i].addr), in ahc_handle_seqint() 1414 ahc_le32toh(scb->sg_list[i].len) in ahc_handle_seqint() 2125 (ahc_le32toh(scb->sg_list[i].len) >> 24 2127 ahc_le32toh(scb->sg_list[i].addr), [all …]
|
D | aic7xxx_osm.c | 1528 sg = scb->sg_list; in ahc_linux_run_command() 1558 scb->hscb->dataptr = scb->sg_list->addr; in ahc_linux_run_command() 1559 scb->hscb->datacnt = scb->sg_list->len; in ahc_linux_run_command()
|
D | aic7xxx.h | 576 struct ahc_dma_seg *sg_list; member
|
D | aic79xx.h | 617 void *sg_list; member
|
D | aic79xx_osm.c | 1636 void *sg = scb->sg_list; in ahd_linux_run_command()
|
/linux-4.4.14/drivers/scsi/ibmvscsi/ |
D | ibmvfc.c | 1859 req_seg = dma_map_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request() 1867 rsp_seg = dma_map_sg(vhost->dev, job->reply_payload.sg_list, in ibmvfc_bsg_request() 1871 dma_unmap_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request() 1911 mad->iu.cmd.va = cpu_to_be64(sg_dma_address(job->request_payload.sg_list)); in ibmvfc_bsg_request() 1912 mad->iu.cmd.len = cpu_to_be32(sg_dma_len(job->request_payload.sg_list)); in ibmvfc_bsg_request() 1913 mad->iu.rsp.va = cpu_to_be64(sg_dma_address(job->reply_payload.sg_list)); in ibmvfc_bsg_request() 1914 mad->iu.rsp.len = cpu_to_be32(sg_dma_len(job->reply_payload.sg_list)); in ibmvfc_bsg_request() 1943 dma_unmap_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request() 1945 dma_unmap_sg(vhost->dev, job->reply_payload.sg_list, in ibmvfc_bsg_request()
|
/linux-4.4.14/arch/x86/platform/efi/ |
D | efi_64.c | 547 unsigned long count, unsigned long sg_list) in efi_thunk_update_capsule() argument
|
/linux-4.4.14/drivers/s390/scsi/ |
D | zfcp_fc.c | 993 ct_els->req = job->request_payload.sg_list; in zfcp_fc_exec_bsg_job() 994 ct_els->resp = job->reply_payload.sg_list; in zfcp_fc_exec_bsg_job()
|
/linux-4.4.14/drivers/char/ |
D | virtio_console.c | 868 struct sg_list { struct 878 struct sg_list *sgl = sd->u.data; in pipe_to_sg() argument 924 struct sg_list sgl; in port_fops_splice_write()
|
/linux-4.4.14/include/rdma/ |
D | ib_verbs.h | 1085 struct ib_sge *sg_list; member 1179 struct ib_sge *sg_list; member
|
/linux-4.4.14/drivers/scsi/bfa/ |
D | bfad_bsg.c | 3156 sg_copy_to_buffer(job->request_payload.sg_list, in bfad_im_bsg_vendor_request() 3167 sg_copy_from_buffer(job->reply_payload.sg_list, in bfad_im_bsg_vendor_request() 3469 sg_copy_to_buffer(job->request_payload.sg_list, in bfad_im_bsg_els_ct_request() 3530 sg_copy_from_buffer(job->reply_payload.sg_list, in bfad_im_bsg_els_ct_request()
|
/linux-4.4.14/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 788 wr.sg_list = &list; in srpt_post_recv() 824 wr.sg_list = &list; in srpt_post_send() 2811 wr.wr.sg_list = riu->sge; in srpt_perform_rdmas()
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd_cb.c | 169 rx->rx_wrq.sg_list = &rx->rx_sge; in kiblnd_post_rx() 1036 wrq->wr.sg_list = sge; in kiblnd_init_tx_msg() 1101 wrq->wr.sg_list = sge; in kiblnd_init_rdma()
|
/linux-4.4.14/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 511 sg_copy_to_buffer(job->request_payload.sg_list, in mgmt_vendor_specific_fw_cmd()
|
D | be_main.c | 5224 sg_copy_from_buffer(job->reply_payload.sg_list, in beiscsi_bsg_request()
|
/linux-4.4.14/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1754 wr.sg_list = &list; in srp_post_send() 1774 wr.sg_list = &list; in srp_post_recv()
|