Home
last modified time | relevance | path

Searched refs:sg_list (Results 1 – 122 of 122) sorted by relevance

/linux-4.1.27/block/
Dbsg-lib.c39 kfree(job->request_payload.sg_list); in bsg_destroy_job()
40 kfree(job->reply_payload.sg_list); in bsg_destroy_job()
96 buf->sg_list = kzalloc(sz, GFP_KERNEL); in bsg_map_buffer()
97 if (!buf->sg_list) in bsg_map_buffer()
99 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer()
100 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer()
148 kfree(job->request_payload.sg_list); in bsg_create_job()
/linux-4.1.27/drivers/virt/
Dfsl_hypervisor.c153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local
243 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy()
265 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
266 sg_list[0].target = param.remote_paddr; in ioctl_memcpy()
268 sg_list[0].source = param.remote_paddr; in ioctl_memcpy()
269 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
271 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy()
273 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy()
274 count = param.count - sg_list[0].size; in ioctl_memcpy()
279 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy()
[all …]
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c85 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); in ehca_write_rwqe()
91 wqe_p->u.all_rcv.sg_list[cnt_ds].vaddr = in ehca_write_rwqe()
92 recv_wr->sg_list[cnt_ds].addr; in ehca_write_rwqe()
93 wqe_p->u.all_rcv.sg_list[cnt_ds].lkey = in ehca_write_rwqe()
94 recv_wr->sg_list[cnt_ds].lkey; in ehca_write_rwqe()
95 wqe_p->u.all_rcv.sg_list[cnt_ds].length = in ehca_write_rwqe()
96 recv_wr->sg_list[cnt_ds].length; in ehca_write_rwqe()
119 struct ib_sge *sge = send_wr->sg_list; in trace_send_wr_ud()
175 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); in ehca_write_swqe()
247 wqe_p->u.ud_av.sg_list[idx].vaddr = in ehca_write_swqe()
[all …]
Dehca_qes.h156 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member
164 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member
168 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES - member
176 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member
Dehca_qp.c370 u.nud.sg_list[act_nr_sge]); in ehca_calc_wqe_size()
/linux-4.1.27/drivers/scsi/qla2xxx/
Dqla_bsg.c41 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free()
46 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free()
49 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free()
52 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free()
183 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd()
210 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd()
336 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els()
343 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els()
395 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els()
397 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els()
[all …]
Dqla_iocb.c2066 (bsg_job->request_payload.sg_list))); in qla24xx_els_iocb()
2068 (bsg_job->request_payload.sg_list))); in qla24xx_els_iocb()
2070 (bsg_job->request_payload.sg_list)); in qla24xx_els_iocb()
2073 (bsg_job->reply_payload.sg_list))); in qla24xx_els_iocb()
2075 (bsg_job->reply_payload.sg_list))); in qla24xx_els_iocb()
2077 (bsg_job->reply_payload.sg_list)); in qla24xx_els_iocb()
2115 (bsg_job->request_payload.sg_list))); in qla2x00_ct_iocb()
2117 (bsg_job->request_payload.sg_list))); in qla2x00_ct_iocb()
2121 (bsg_job->reply_payload.sg_list))); in qla2x00_ct_iocb()
2123 (bsg_job->reply_payload.sg_list))); in qla2x00_ct_iocb()
[all …]
Dqla_mr.c3335 for_each_sg(bsg_job->request_payload.sg_list, sg, in qlafx00_fxdisc_iocb()
3398 for_each_sg(bsg_job->reply_payload.sg_list, sg, in qlafx00_fxdisc_iocb()
/linux-4.1.27/drivers/crypto/caam/
Dsg_sw_sec4.h59 static inline int __sg_count(struct scatterlist *sg_list, int nbytes, in __sg_count() argument
62 struct scatterlist *sg = sg_list; in __sg_count()
77 static inline int sg_count(struct scatterlist *sg_list, int nbytes, in sg_count() argument
80 int sg_nents = __sg_count(sg_list, nbytes, chained); in sg_count()
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_qp.c73 if ((plen + wr->sg_list[i].length) < plen) in build_rdma_send()
76 plen += wr->sg_list[i].length; in build_rdma_send()
77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_send()
78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); in build_rdma_send()
79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); in build_rdma_send()
110 if ((plen + wr->sg_list[i].length) < plen) { in build_rdma_write()
113 plen += wr->sg_list[i].length; in build_rdma_write()
115 cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_write()
117 cpu_to_be32(wr->sg_list[i].length); in build_rdma_write()
119 cpu_to_be64(wr->sg_list[i].addr); in build_rdma_write()
[all …]
/linux-4.1.27/arch/blackfin/kernel/
Ddma-mapping.c116 dma_map_sg(struct device *dev, struct scatterlist *sg_list, int nents, in dma_map_sg() argument
122 for_each_sg(sg_list, sg, nents, i) { in dma_map_sg()
131 void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg_list, in dma_sync_sg_for_device() argument
137 for_each_sg(sg_list, sg, nelems, i) { in dma_sync_sg_for_device()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_ruc.c92 ss->sg_list = qp->r_sg_list; in qib_init_sge()
95 if (wqe->sg_list[i].length == 0) in qib_init_sge()
98 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in qib_init_sge()
99 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) in qib_init_sge()
101 qp->r_len += wqe->sg_list[i].length; in qib_init_sge()
111 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in qib_init_sge()
427 sqp->s_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback()
428 sqp->s_sge.sg_list = wqe->sg_list + 1; in qib_ruc_loopback()
465 qp->r_sge.sg_list = NULL; in qib_ruc_loopback()
479 sqp->s_sge.sg_list = NULL; in qib_ruc_loopback()
[all …]
Dqib_ud.c172 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback()
173 ssge.sge = *swqe->sg_list; in qib_ud_loopback()
190 *sge = *ssge.sg_list++; in qib_ud_loopback()
320 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req()
321 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
Dqib_srq.c83 wqe->sg_list[i] = wr->sg_list[i]; in qib_post_srq_receive()
289 p->sg_list[i] = wqe->sg_list[i]; in qib_modify_srq()
Dqib_uc.c99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req()
100 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req()
434 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
Dqib_verbs.c189 *sge = *ss->sg_list++; in qib_copy_sge()
230 *sge = *ss->sg_list++; in qib_skip_sge()
253 struct qib_sge *sg_list = ss->sg_list; in qib_count_sge() local
277 sge = *sg_list++; in qib_count_sge()
315 *sge = *ss->sg_list++; in qib_copy_from_sge()
383 wr->sg_list[0].length < sizeof(u64) || in qib_post_one_send()
384 wr->sg_list[0].addr & (sizeof(u64) - 1))) in qib_post_one_send()
407 u32 length = wr->sg_list[i].length; in qib_post_one_send()
412 ok = qib_lkey_ok(rkt, pd, &wqe->sg_list[j], in qib_post_one_send()
413 &wr->sg_list[i], acc); in qib_post_one_send()
[all …]
Dqib_rc.c49 ss->sge = wqe->sg_list[0]; in restart_sge()
50 ss->sg_list = wqe->sg_list + 1; in restart_sge()
487 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_rc_req()
488 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_rc_req()
1013 struct qib_sge *sge = &wqe->sg_list[i]; in qib_rc_send_complete()
1069 struct qib_sge *sge = &wqe->sg_list[i]; in do_rc_completion()
1205 u64 *vaddr = wqe->sg_list[0].vaddr; in do_rc_ack()
2063 qp->r_sge.sg_list = NULL; in qib_rc_rcv()
Dqib_verbs.h346 struct qib_sge sg_list[0]; member
357 struct ib_sge sg_list[0]; member
391 struct qib_sge *sg_list; /* next SGE to be used if any */ member
1071 ss->sge = *ss->sg_list++; in qib_put_ss()
Dqib_sdma.c622 *sge = *ss->sg_list++; in qib_sdma_verbs_send()
Dqib_qp.c433 struct qib_sge *sge = &wqe->sg_list[i]; in clear_mr_refs()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_ruc.c131 if (wqe->sg_list[i].length == 0) in ipath_init_sge()
134 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, in ipath_init_sge()
135 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) in ipath_init_sge()
137 *lengthp += wqe->sg_list[i].length; in ipath_init_sge()
210 qp->r_sge.sg_list = qp->r_sg_list; in ipath_get_rwqe()
328 sqp->s_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
329 sqp->s_sge.sg_list = wqe->sg_list + 1; in ipath_ruc_loopback()
370 qp->r_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
371 qp->r_sge.sg_list = wqe->sg_list + 1; in ipath_ruc_loopback()
414 *sge = *sqp->s_sge.sg_list++; in ipath_ruc_loopback()
Dipath_srq.c83 wqe->sg_list[i] = wr->sg_list[i]; in ipath_post_srq_receive()
293 p->sg_list[i] = wqe->sg_list[i]; in ipath_modify_srq()
Dipath_ud.c134 rsge.sg_list = qp->r_ud_sg_list; in ipath_ud_loopback()
184 sge = swqe->sg_list; in ipath_ud_loopback()
323 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_ud_req()
324 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_ud_req()
Dipath_uc.c98 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_uc_req()
99 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_uc_req()
453 qp->r_sge.sg_list = NULL; in ipath_uc_rcv()
Dipath_keys.c226 ss->sg_list = NULL; in ipath_rkey_ok()
263 ss->sg_list = NULL; in ipath_rkey_ok()
Dipath_verbs.c189 *sge = *ss->sg_list++; in ipath_copy_sge()
228 *sge = *ss->sg_list++; in ipath_skip_sge()
251 struct ipath_sge *sg_list = ss->sg_list; in ipath_count_sge() local
275 sge = *sg_list++; in ipath_count_sge()
314 *sge = *ss->sg_list++; in ipath_copy_from_sge()
383 wr->sg_list[0].length < sizeof(u64) || in ipath_post_one_send()
384 wr->sg_list[0].addr & (sizeof(u64) - 1))) in ipath_post_one_send()
404 u32 length = wr->sg_list[i].length; in ipath_post_one_send()
409 ok = ipath_lkey_ok(qp, &wqe->sg_list[j], in ipath_post_one_send()
410 &wr->sg_list[i], acc); in ipath_post_one_send()
[all …]
Dipath_rc.c48 ss->sge = wqe->sg_list[0]; in restart_sge()
49 ss->sg_list = wqe->sg_list + 1; in restart_sge()
461 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_rc_req()
462 qp->s_sge.sg_list = wqe->sg_list + 1; in ipath_make_rc_req()
936 *(u64 *) wqe->sg_list[0].vaddr = val; in do_rc_ack()
1452 e->rdma_sge.sg_list = NULL; in ipath_rc_rcv_error()
1779 qp->r_sge.sg_list = NULL; in ipath_rc_rcv()
1839 e->rdma_sge.sg_list = NULL; in ipath_rc_rcv()
Dipath_verbs.h285 struct ipath_sge sg_list[0]; member
296 struct ib_sge sg_list[0]; member
329 struct ipath_sge *sg_list; /* next SGE to be used if any */ member
Dipath_sdma.c764 *sge = *ss->sg_list++; in ipath_sdma_verbs_send()
/linux-4.1.27/drivers/dma/
Dimx-dma.c165 struct scatterlist *sg_list; member
804 kfree(imxdmac->sg_list); in imxdma_free_chan_resources()
805 imxdmac->sg_list = NULL; in imxdma_free_chan_resources()
879 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic()
881 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic()
883 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic()
886 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic()
889 imxdmac->sg_list[i].page_link = 0; in imxdma_prep_dma_cyclic()
890 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic()
891 imxdmac->sg_list[i].dma_address = dma_addr; in imxdma_prep_dma_cyclic()
[all …]
/linux-4.1.27/drivers/staging/i2o/
Di2o_config.c535 struct i2o_dma sg_list[SG_TABLESIZE]; in i2o_cfg_passthru32() local
602 memset(sg_list, 0, sizeof(sg_list[0]) * SG_TABLESIZE); in i2o_cfg_passthru32()
635 p = &(sg_list[sg_index]); in i2o_cfg_passthru32()
713 sg_list[j].virt, sg_size)) { in i2o_cfg_passthru32()
716 c->name, sg_list[j].virt, in i2o_cfg_passthru32()
742 i2o_dma_free(&c->pdev->dev, &sg_list[i]); in i2o_cfg_passthru32()
786 struct i2o_dma sg_list[SG_TABLESIZE]; in i2o_cfg_passthru() local
844 memset(sg_list, 0, sizeof(sg_list[0]) * SG_TABLESIZE); in i2o_cfg_passthru()
877 p = &(sg_list[sg_index]); in i2o_cfg_passthru()
952 ((void __user *)sg[j].addr_bus, sg_list[j].virt, in i2o_cfg_passthru()
[all …]
/linux-4.1.27/drivers/scsi/aacraid/
Dcommctrl.c480 void *sg_list[32]; in aac_send_raw_srb() local
507 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb()
560 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb()
619 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
671 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
725 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
761 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
804 if(copy_to_user(sg_user[i], sg_list[i], byte_count)){ in aac_send_raw_srb()
823 kfree(sg_list[i]); in aac_send_raw_srb()
/linux-4.1.27/include/linux/
Dbsg-lib.h36 struct scatterlist *sg_list; member
Dagp_backend.h85 struct scatterlist *sg_list; member
Defi.h523 unsigned long sg_list);
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dqp.c374 if ((plen + wr->sg_list[i].length) > max) in build_immd()
376 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd()
377 plen += wr->sg_list[i].length; in build_immd()
378 rem = wr->sg_list[i].length; in build_immd()
404 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl() argument
413 if ((plen + sg_list[i].length) < plen) in build_isgl()
415 plen += sg_list[i].length; in build_isgl()
416 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | in build_isgl()
417 sg_list[i].length); in build_isgl()
420 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl()
[all …]
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.h120 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
129 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member
146 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
Dehea_main.c503 rwqe->sg_list[0].l_key = pr->recv_mr.lkey; in ehea_refill_rq_def()
504 rwqe->sg_list[0].vaddr = tmp_addr; in ehea_refill_rq_def()
505 rwqe->sg_list[0].len = packet_size; in ehea_refill_rq_def()
1672 struct ehea_vsgentry *sg_list, *sg1entry, *sgentry; in write_swqe2_data() local
1678 sg_list = (struct ehea_vsgentry *)&swqe->u.immdata_desc.sg_list; in write_swqe2_data()
1701 sgentry = &sg_list[i - sg1entry_contains_frag_data]; in write_swqe2_data()
2628 rwqe->sg_list[0].l_key = lkey; in ehea_update_rqs()
2632 rwqe->sg_list[0].vaddr = ehea_map_vaddr(skb->data); in ehea_update_rqs()
2637 rwqe->sg_list[0].l_key = lkey; in ehea_update_rqs()
2641 rwqe->sg_list[0].vaddr = ehea_map_vaddr(skb->data); in ehea_update_rqs()
/linux-4.1.27/arch/powerpc/include/asm/
Dfsl_hcalls.h358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument
370 r5 = (uint32_t) sg_list; in fh_partition_memcpy()
373 r6 = sg_list >> 32; in fh_partition_memcpy()
/linux-4.1.27/drivers/crypto/
Dbfin_crc.c103 static int sg_count(struct scatterlist *sg_list) in sg_count() argument
105 struct scatterlist *sg = sg_list; in sg_count()
108 if (sg_list == NULL) in sg_count()
122 static struct scatterlist *sg_get(struct scatterlist *sg_list, unsigned int nents, in sg_get() argument
128 for_each_sg(sg_list, sg, nents, i) in sg_get()
Datmel-aes.c164 struct scatterlist *sg_list; in atmel_aes_sg_length() local
167 sg_list = sg; in atmel_aes_sg_length()
171 len = min(sg_list->length, total); in atmel_aes_sg_length()
176 sg_list = sg_next(sg_list); in atmel_aes_sg_length()
177 if (!sg_list) in atmel_aes_sg_length()
Dsahara.c282 struct scatterlist *sg_list; in sahara_sg_length() local
285 sg_list = sg; in sahara_sg_length()
288 len = min(sg_list->length, total); in sahara_sg_length()
293 sg_list = sg_next(sg_list); in sahara_sg_length()
294 if (!sg_list) in sahara_sg_length()
Dpicoxcell_crypto.c264 static int sg_count(struct scatterlist *sg_list, int nbytes) in sg_count() argument
266 struct scatterlist *sg = sg_list; in sg_count()
Dtalitos.c1093 static int sg_count(struct scatterlist *sg_list, int nbytes, bool *chained) in sg_count() argument
1095 struct scatterlist *sg = sg_list; in sg_count()
/linux-4.1.27/drivers/crypto/qce/
Ddma.h52 int qce_countsg(struct scatterlist *sg_list, int nbytes, bool *chained);
/linux-4.1.27/drivers/firmware/efi/
Druntime-wrappers.c258 unsigned long sg_list) in virt_efi_update_capsule() argument
267 status = efi_call_virt(update_capsule, capsules, count, sg_list); in virt_efi_update_capsule()
/linux-4.1.27/drivers/xen/
Defi.c223 unsigned long sg_list) in xen_efi_update_capsule() argument
233 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dsrq.c349 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv()
350 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv()
351 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv()
Dmad.c597 wr.sg_list = &list; in mlx4_ib_send_to_slave()
1121 struct ib_sge sg_list; in mlx4_ib_post_pv_qp_buf() local
1128 sg_list.addr = tun_qp->ring[index].map; in mlx4_ib_post_pv_qp_buf()
1129 sg_list.length = size; in mlx4_ib_post_pv_qp_buf()
1130 sg_list.lkey = ctx->mr->lkey; in mlx4_ib_post_pv_qp_buf()
1133 recv_wr.sg_list = &sg_list; in mlx4_ib_post_pv_qp_buf()
1250 wr.sg_list = &list; in mlx4_ib_send_to_wire()
Dqp.c2055 send_size += wr->sg_list[i].length; in build_sriov_qp0_header()
2178 send_size += wr->sg_list[i].length; in build_mlx_header()
2876 set_data_seg(dseg, wr->sg_list + i); in mlx4_ib_post_send()
3002 scat->lkey = cpu_to_be32(wr->sg_list->lkey); in mlx4_ib_post_recv()
3009 __set_data_seg(scat + i, wr->sg_list + i); in mlx4_ib_post_recv()
/linux-4.1.27/drivers/scsi/qla4xxx/
Dql4_bsg.c63 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_flash()
123 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_flash()
187 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_get_acb_state()
258 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_nvram()
322 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_nvram()
436 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_bsg_get_acb()
/linux-4.1.27/drivers/infiniband/hw/amso1100/
Dc2_qp.c845 ib_wr->sg_list, in c2_post_send()
868 ib_wr->sg_list, in c2_post_send()
888 cpu_to_be32(ib_wr->sg_list->lkey); in c2_post_send()
890 cpu_to_be64(ib_wr->sg_list->addr); in c2_post_send()
896 cpu_to_be32(ib_wr->sg_list->length); in c2_post_send()
980 ib_wr->sg_list, in c2_post_receive()
/linux-4.1.27/drivers/infiniband/core/
Dmad.c955 mad_send_wr->sg_list[0].length = hdr_len; in ib_create_send_mad()
956 mad_send_wr->sg_list[0].lkey = mad_agent->mr->lkey; in ib_create_send_mad()
957 mad_send_wr->sg_list[1].length = sizeof(struct ib_mad) - hdr_len; in ib_create_send_mad()
958 mad_send_wr->sg_list[1].lkey = mad_agent->mr->lkey; in ib_create_send_mad()
961 mad_send_wr->send_wr.sg_list = mad_send_wr->sg_list; in ib_create_send_mad()
1076 sge = mad_send_wr->sg_list; in ib_send_mad()
2221 mad_send_wr->sg_list[0].length, DMA_TO_DEVICE); in ib_mad_send_done_handler()
2224 mad_send_wr->sg_list[1].length, DMA_TO_DEVICE); in ib_mad_send_done_handler()
2647 struct ib_sge sg_list; in ib_mad_post_receive_mads() local
2652 sg_list.length = sizeof *mad_priv - sizeof mad_priv->header; in ib_mad_post_receive_mads()
[all …]
Dmad_priv.h129 struct ib_sge sg_list[IB_MAD_SEND_REQ_MAX_SG]; member
Duverbs_cmd.c2301 next->sg_list = (void *) next + in ib_uverbs_post_send()
2303 if (copy_from_user(next->sg_list, in ib_uverbs_post_send()
2313 next->sg_list = NULL; in ib_uverbs_post_send()
2402 next->sg_list = (void *) next + in ib_uverbs_unmarshall_recv()
2404 if (copy_from_user(next->sg_list, in ib_uverbs_unmarshall_recv()
2413 next->sg_list = NULL; in ib_uverbs_unmarshall_recv()
/linux-4.1.27/drivers/scsi/
D3w-9xxx.c1348 …if (full_command_packet->command.newcommand.sg_list[0].length < scsi_bufflen(tw_dev->srb[request_i… in twa_interrupt()
1349 …scsi_set_resid(cmd, scsi_bufflen(cmd) - full_command_packet->command.newcommand.sg_list[0].length); in twa_interrupt()
1393 newcommand->sg_list[0].address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); in twa_load_sgl()
1394 newcommand->sg_list[0].length = cpu_to_le32(length); in twa_load_sgl()
1855 command_packet->sg_list[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); in DEF_SCSI_QCMD()
1856 command_packet->sg_list[0].length = cpu_to_le32(TW_MIN_SGL_LENGTH); in DEF_SCSI_QCMD()
1863 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sg_dma_address(sg)); in DEF_SCSI_QCMD()
1864 command_packet->sg_list[i].length = cpu_to_le32(sg_dma_len(sg)); in DEF_SCSI_QCMD()
1865 if (command_packet->sg_list[i].address & TW_CPU_TO_SGL(TW_ALIGNMENT_9000_SGL)) { in DEF_SCSI_QCMD()
1876 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sglistarg[i].address); in DEF_SCSI_QCMD()
[all …]
Dhptiop.h222 struct hpt_iopsg sg_list[1]; member
239 struct hpt_iopsg sg_list[1]; member
D3w-sas.c345 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sg_dma_address(sg)); in twl_scsiop_execute_scsi()
346 command_packet->sg_list[i].length = TW_CPU_TO_SGL(sg_dma_len(sg)); in twl_scsiop_execute_scsi()
353 command_packet->sg_list[i].address = TW_CPU_TO_SGL(sglistarg[i].address); in twl_scsiop_execute_scsi()
354 command_packet->sg_list[i].length = TW_CPU_TO_SGL(sglistarg[i].length); in twl_scsiop_execute_scsi()
697 newcommand->sg_list[0].address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); in twl_load_sgl()
698 newcommand->sg_list[0].length = TW_CPU_TO_SGL(length); in twl_load_sgl()
1220 …if (full_command_packet->command.newcommand.sg_list[0].length < scsi_bufflen(tw_dev->srb[request_i… in twl_interrupt()
1221 …scsi_set_resid(cmd, scsi_bufflen(cmd) - full_command_packet->command.newcommand.sg_list[0].length); in twl_interrupt()
Dips.c1814 scb->sg_list.enh_list[indx].address_lo = in ips_fill_scb_sg_single()
1816 scb->sg_list.enh_list[indx].address_hi = in ips_fill_scb_sg_single()
1818 scb->sg_list.enh_list[indx].length = cpu_to_le32(e_len); in ips_fill_scb_sg_single()
1820 scb->sg_list.std_list[indx].address = in ips_fill_scb_sg_single()
1822 scb->sg_list.std_list[indx].length = cpu_to_le32(e_len); in ips_fill_scb_sg_single()
1838 IPS_SG_LIST sg_list; in ips_flash_firmware() local
1853 sg_list.list = scb->sg_list.list; in ips_flash_firmware()
1858 scb->sg_list.list = sg_list.list; in ips_flash_firmware()
1911 IPS_SG_LIST sg_list; in ips_usrcmd() local
1920 sg_list.list = scb->sg_list.list; in ips_usrcmd()
[all …]
Dwd719x.h62 struct wd719x_sglist sg_list[WD719X_SG] __aligned(8); /* SG list */ member
D3w-xxxx.h386 TW_SG_Entry sg_list[TW_ATA_PASS_SGL_MAX]; member
Deata_generic.h314 struct eata_sg_list *sg_list; member
D3w-sas.h252 TW_SG_Entry_ISO sg_list[TW_LIBERATOR_MAX_SGL_LENGTH]; member
Dwd719x.c274 offsetof(struct wd719x_scb, sg_list)); in wd719x_queuecommand()
277 scb->sg_list[i].ptr = cpu_to_le32(sg_dma_address(sg)); in wd719x_queuecommand()
278 scb->sg_list[i].length = cpu_to_le32(sg_dma_len(sg)); in wd719x_queuecommand()
Dadvansys.c367 ASC_SG_LIST sg_list[0]; member
423 ASC_SG_LIST sg_list[7]; member
1828 } sg_list[NO_OF_SG_PER_BLOCK]; member
2624 i, (ulong)le32_to_cpu(sgp->sg_list[i].addr), in asc_prt_asc_scsi_q()
2625 (ulong)le32_to_cpu(sgp->sg_list[i].bytes)); in asc_prt_asc_scsi_q()
2663 i, (ulong)b->sg_list[i].sg_addr, in asc_prt_adv_sgblock()
2664 (ulong)b->sg_list[i].sg_count); in asc_prt_adv_sgblock()
6984 sg_list[scsiq->next_sg_index], in AscIsrChipHalted()
7948 asc_sg_head->sg_list[sgcnt].addr = in asc_build_req()
7950 asc_sg_head->sg_list[sgcnt].bytes = in asc_build_req()
[all …]
Ddpt_i2o.c1713 void *sg_list[pHba->sg_tablesize]; in adpt_i2o_passthru() local
1757 memset(sg_list,0, sizeof(sg_list[0])*pHba->sg_tablesize); in adpt_i2o_passthru()
1785 sg_list[sg_index++] = p; // sglist indexed with input frame, not our internal frame. in adpt_i2o_passthru()
1857 if (copy_to_user((void __user *)(ulong)sg[j].addr_bus,sg_list[j], sg_size)) { in adpt_i2o_passthru()
1858 … printk(KERN_WARNING"%s: Could not copy %p TO user %x\n",pHba->name, sg_list[j], sg[j].addr_bus); in adpt_i2o_passthru()
1886 if(sg_list[--sg_index]) { in adpt_i2o_passthru()
1889 sg_list[sg_index], in adpt_i2o_passthru()
Dnsp32.c2661 data->sg_list = pci_alloc_consistent(pdev, NSP32_SG_TABLE_SIZE, in nsp32_detect()
2663 if (data->sg_list == NULL) { in nsp32_detect()
2677 .sglun = &(data->sg_list[offset]), in nsp32_detect()
2768 data->sg_list, data->sg_paddr); in nsp32_detect()
2790 if (data->sg_list) { in nsp32_release()
2792 data->sg_list, data->sg_paddr); in nsp32_release()
D3w-9xxx.h521 TW_SG_Entry sg_list[TW_APACHE_MAX_SGL_LENGTH]; member
Deata.c742 struct sg_list { struct
811 struct sg_list *sglist; /* pointer to the allocated SG list */
1346 size_t sz = shost->sg_tablesize *sizeof(struct sg_list); in port_detect()
1634 sizeof(struct sg_list), in map_dma()
1636 cpp->data_len = H2DEV((scsi_sg_count(SCpnt) * sizeof(struct sg_list))); in map_dma()
Dips.h1100 IPS_SG_LIST sg_list; member
1126 IPS_SG_LIST *sg_list; member
Du14-34f.c544 struct sg_list { struct
585 struct sg_list *sglist; /* pointer to the allocated SG list */
983 sh[j]->sg_tablesize * sizeof(struct sg_list), in port_detect()
1143 cpp->use_sg * sizeof(struct sg_list), in map_dma()
Dnsp32.h575 nsp32_sglun *sg_list; /* sglist virtuxal address */ member
Dlibiscsi_tcp.c369 struct scatterlist *sg_list, unsigned int sg_count, in iscsi_segment_seek_sg() argument
377 for_each_sg(sg_list, sg, sg_count, i) { in iscsi_segment_seek_sg()
Dscsi_transport_fc.c3575 kfree(job->request_payload.sg_list); in fc_destroy_bsgjob()
3576 kfree(job->reply_payload.sg_list); in fc_destroy_bsgjob()
3679 buf->sg_list = kzalloc(sz, GFP_KERNEL); in fc_bsg_map_buffer()
3680 if (!buf->sg_list) in fc_bsg_map_buffer()
3682 sg_init_table(buf->sg_list, req->nr_phys_segments); in fc_bsg_map_buffer()
3683 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in fc_bsg_map_buffer()
3756 kfree(job->request_payload.sg_list); in fc_req_to_bsgjob()
D3w-xxxx.c974 passthru->sg_list[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1; in tw_chrdev_ioctl()
975 passthru->sg_list[0].length = data_buffer_length_adjusted; in tw_chrdev_ioctl()
Dhptiop.c767 memcpy(scp->sense_buffer, &req->sg_list, in hptiop_finish_scsi_req()
1050 sg_count = hptiop_buildsgl(scp, req->sg_list); in hptiop_queuecommand_lck()
/linux-4.1.27/drivers/char/agp/
Dintel-gtt.c123 static void intel_gtt_unmap_memory(struct scatterlist *sg_list, int num_sg) in intel_gtt_unmap_memory() argument
128 pci_unmap_sg(intel_private.pcidev, sg_list, in intel_gtt_unmap_memory()
131 st.sgl = sg_list; in intel_gtt_unmap_memory()
919 mem->sg_list = st.sgl; in intel_fake_agp_insert_entries()
955 intel_gtt_unmap_memory(mem->sg_list, mem->num_sg); in intel_fake_agp_remove_entries()
956 mem->sg_list = NULL; in intel_fake_agp_remove_entries()
/linux-4.1.27/include/scsi/
Dlibiscsi_tcp.h117 struct scatterlist *sg_list, unsigned int sg_count,
Dscsi_transport_fc.h630 struct scatterlist *sg_list; member
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c227 priv->tx_wr.sg_list = priv->tx_sge; in ipoib_transport_dev_init()
236 priv->rx_wr.sg_list = priv->rx_sge; in ipoib_transport_dev_init()
Dipoib_cm.c342 wr->sg_list = sge; in ipoib_cm_init_rx_wr()
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dsrq.c459 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx5_ib_post_srq_recv()
460 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx5_ib_post_srq_recv()
461 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx5_ib_post_srq_recv()
Dqp.c2107 addr = (void *)(unsigned long)(wr->sg_list[i].addr); in set_data_inl_seg()
2108 len = wr->sg_list[i].length; in set_data_inl_seg()
2245 u32 data_len = wr->sg_list->length; in set_sig_data_segment()
2246 u32 data_key = wr->sg_list->lkey; in set_sig_data_segment()
2247 u64 data_va = wr->sg_list->addr; in set_sig_data_segment()
2388 region_len = wr->sg_list->length; in set_sig_umr_wr()
2390 (wr->wr.sig_handover.prot->lkey != wr->sg_list->lkey || in set_sig_umr_wr()
2391 wr->wr.sig_handover.prot->addr != wr->sg_list->addr || in set_sig_umr_wr()
2392 wr->wr.sig_handover.prot->length != wr->sg_list->length)) in set_sig_umr_wr()
2830 if (likely(wr->sg_list[i].length)) { in mlx5_ib_post_send()
[all …]
Dmr.c702 wr->sg_list = sg; in prep_umr_reg_wqe()
933 wr.sg_list = &sg; in mlx5_ib_update_mtt()
/linux-4.1.27/drivers/infiniband/hw/nes/
Dnes_verbs.c3239 ib_wr->sg_list[sge_index].addr); in fill_wqe_sg_send()
3241 ib_wr->sg_list[sge_index].length); in fill_wqe_sg_send()
3244 (ib_wr->sg_list[sge_index].lkey)); in fill_wqe_sg_send()
3248 total_payload_length += ib_wr->sg_list[sge_index].length; in fill_wqe_sg_send()
3332 (ib_wr->sg_list[0].length <= 64)) { in nes_post_send()
3334 (void *)(unsigned long)ib_wr->sg_list[0].addr, ib_wr->sg_list[0].length); in nes_post_send()
3336 ib_wr->sg_list[0].length); in nes_post_send()
3362 (ib_wr->sg_list[0].length <= 64)) { in nes_post_send()
3364 (void *)(unsigned long)ib_wr->sg_list[0].addr, ib_wr->sg_list[0].length); in nes_post_send()
3366 ib_wr->sg_list[0].length); in nes_post_send()
[all …]
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1944 struct ib_sge *sg_list) in ocrdma_build_sges() argument
1949 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges()
1950 sge[i].addr_lo = sg_list[i].addr; in ocrdma_build_sges()
1951 sge[i].addr_hi = upper_32_bits(sg_list[i].addr); in ocrdma_build_sges()
1952 sge[i].len = sg_list[i].length; in ocrdma_build_sges()
1953 hdr->total_len += sg_list[i].length; in ocrdma_build_sges()
1959 static inline uint32_t ocrdma_sglist_len(struct ib_sge *sg_list, int num_sge) in ocrdma_sglist_len() argument
1964 total_len += sg_list[i].length; in ocrdma_sglist_len()
1978 hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); in ocrdma_build_inline_sges()
1988 (void *)(unsigned long)wr->sg_list[i].addr, in ocrdma_build_inline_sges()
[all …]
/linux-4.1.27/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c213 write_wr.sg_list = &sge[0]; in send_write()
473 send_wr.sg_list = ctxt->sge; in send_reply()
Dsvc_rdma_recvfrom.c199 read_wr.sg_list = ctxt->sge; in rdma_read_chunk_lcl()
320 read_wr.sg_list = ctxt->sge; in rdma_read_chunk_frmr()
Dsvc_rdma_transport.c538 recv_wr.sg_list = &ctxt->sge[0]; in svc_rdma_post_recv()
1353 err_wr.sg_list = ctxt->sge; in svc_rdma_send_error()
Dverbs.c1594 send_wr.sg_list = req->rl_send_iov; in rpcrdma_ep_post()
1636 recv_wr.sg_list = &rep->rr_rdmabuf->rg_iov; in rpcrdma_ep_post_recv()
/linux-4.1.27/net/9p/
Dtrans_rdma.c423 wr.sg_list = &sge; in post_recv()
516 wr.sg_list = &sge; in rdma_request()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_verbs.c1015 rx_wr.sg_list = &sge; in iser_post_recvl()
1039 rx_wr->sg_list = &rx_desc->rx_sg; in iser_post_recvm()
1076 send_wr.sg_list = tx_desc->tx_sg; in iser_post_send()
Diser_memory.c678 sig_wr.sg_list = &data_reg->sge; in iser_reg_sig_mr()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_srq.c522 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_srq_recv()
614 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_srq_recv()
Dmthca_qp.c1737 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_send()
1862 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_receive()
2078 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_send()
2198 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_arbel_post_receive()
/linux-4.1.27/net/rds/
Diw_recv.c95 recv->r_wr.sg_list = recv->r_sge; in rds_iw_recv_init_ring()
356 wr->sg_list = sge; in rds_iw_recv_init_ack()
Dib_recv.c59 recv->r_wr.sg_list = recv->r_sge; in rds_ib_recv_init_ring()
525 wr->sg_list = sge; in rds_ib_recv_init_ack()
Diw_send.c142 send->s_wr.sg_list = send->s_sge; in rds_iw_send_init_ring()
Dib_send.c233 send->s_wr.sg_list = send->s_sge; in rds_ib_send_init_ring()
/linux-4.1.27/drivers/scsi/lpfc/
Dlpfc_bsg.c235 sg_miter_start(&miter, bsg_buffers->sg_list, bsg_buffers->sg_cnt, in lpfc_bsg_copy_data()
613 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_rport_els_cmp()
712 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_rport_els()
1337 sg_copy_from_buffer(job->request_payload.sg_list, in lpfc_bsg_hba_get_event()
3081 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_diag_loopback_run()
3246 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_diag_loopback_run()
3378 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_issue_mbox_cmpl()
3572 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_issue_mbox_ext_handle_job()
4320 sg_copy_from_buffer(job->reply_payload.sg_list, in lpfc_bsg_read_ebuf_get()
4376 sg_copy_to_buffer(job->request_payload.sg_list, in lpfc_bsg_write_ebuf_set()
[all …]
/linux-4.1.27/drivers/block/
Dskd_main.c4191 struct fit_sg_descriptor *sg_list; in skd_cons_sg_list() local
4194 nbytes = sizeof(*sg_list) * n_sg; in skd_cons_sg_list()
4196 sg_list = pci_alloc_consistent(skdev->pdev, nbytes, ret_dma_addr); in skd_cons_sg_list()
4198 if (sg_list != NULL) { in skd_cons_sg_list()
4202 memset(sg_list, 0, nbytes); in skd_cons_sg_list()
4208 sg_list[i].next_desc_ptr = dma_address + ndp_off; in skd_cons_sg_list()
4210 sg_list[i].next_desc_ptr = 0LL; in skd_cons_sg_list()
4213 return sg_list; in skd_cons_sg_list()
4570 struct fit_sg_descriptor *sg_list, in skd_free_sg_list() argument
4573 if (sg_list != NULL) { in skd_free_sg_list()
[all …]
/linux-4.1.27/include/xen/interface/
Dplatform.h199 uint64_t sg_list; /* machine address */ member
/linux-4.1.27/drivers/scsi/libfc/
Dfc_lport.c1985 sg_copy_to_buffer(job->request_payload.sg_list, in fc_lport_els_request()
2009 info->sg = job->reply_payload.sg_list; in fc_lport_els_request()
2046 sg_copy_to_buffer(job->request_payload.sg_list, in fc_lport_ct_request()
2070 info->sg = job->reply_payload.sg_list; in fc_lport_ct_request()
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c1019 rx_wr->sg_list = &rx_desc->rx_sg; in isert_post_recv()
1053 send_wr.sg_list = tx_desc->tx_sg; in isert_post_send()
1124 send_wr->sg_list = &tx_desc->tx_sg[0]; in isert_init_send_wr()
1146 rx_wr.sg_list = &sge; in isert_rdma_post_recvl()
2401 send_wr->sg_list = ib_sge; in isert_build_rdma_wr()
2429 send_wr->sg_list, send_wr->num_sge); in isert_build_rdma_wr()
2737 sig_wr.sg_list = &rdma_wr->ib_sg[DATA]; in isert_reg_sig_mr()
2889 send_wr->sg_list = &wr->s_ib_sge; in isert_reg_rdma()
/linux-4.1.27/drivers/scsi/aic7xxx/
Daic79xx_core.c467 sg = (struct ahd_dma64_seg *)scb->sg_list; in ahd_setup_data_scb()
474 sg = (struct ahd_dma_seg *)scb->sg_list; in ahd_setup_data_scb()
511 return ((uint8_t *)scb->sg_list + sg_offset); in ahd_sg_bus_to_virt()
520 sg_offset = ((uint8_t *)sg - (uint8_t *)scb->sg_list) in ahd_sg_virt_to_bus()
1741 struct ahd_dma64_seg *sg_list; in ahd_dump_sglist() local
1743 sg_list = (struct ahd_dma64_seg*)scb->sg_list; in ahd_dump_sglist()
1748 addr = ahd_le64toh(sg_list[i].addr); in ahd_dump_sglist()
1749 len = ahd_le32toh(sg_list[i].len); in ahd_dump_sglist()
1754 sg_list[i].len & AHD_SG_LEN_MASK, in ahd_dump_sglist()
1755 (sg_list[i].len & AHD_DMA_LAST_SEG) in ahd_dump_sglist()
[all …]
Daic7xxx_core.c380 return (&scb->sg_list[sg_index]); in ahc_sg_bus_to_virt()
389 sg_index = sg - &scb->sg_list[1]; in ahc_sg_virt_to_bus()
391 return (scb->sg_list_phys + (sg_index * sizeof(*scb->sg_list))); in ahc_sg_virt_to_bus()
417 /*offset*/(scb->sg_list - scb->sg_map->sg_vaddr) in ahc_sync_sglist()
1080 sg = scb->sg_list; in ahc_handle_seqint()
1411 (ahc_le32toh(scb->sg_list[i].len) >> 24 in ahc_handle_seqint()
1413 ahc_le32toh(scb->sg_list[i].addr), in ahc_handle_seqint()
1414 ahc_le32toh(scb->sg_list[i].len) in ahc_handle_seqint()
2125 (ahc_le32toh(scb->sg_list[i].len) >> 24
2127 ahc_le32toh(scb->sg_list[i].addr),
[all …]
Daic7xxx_osm.c1529 sg = scb->sg_list; in ahc_linux_run_command()
1559 scb->hscb->dataptr = scb->sg_list->addr; in ahc_linux_run_command()
1560 scb->hscb->datacnt = scb->sg_list->len; in ahc_linux_run_command()
Daic7xxx.h576 struct ahc_dma_seg *sg_list; member
Daic79xx.h617 void *sg_list; member
Daic79xx_osm.c1637 void *sg = scb->sg_list; in ahd_linux_run_command()
/linux-4.1.27/drivers/scsi/ibmvscsi/
Dibmvfc.c1859 req_seg = dma_map_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request()
1867 rsp_seg = dma_map_sg(vhost->dev, job->reply_payload.sg_list, in ibmvfc_bsg_request()
1871 dma_unmap_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request()
1911 mad->iu.cmd.va = cpu_to_be64(sg_dma_address(job->request_payload.sg_list)); in ibmvfc_bsg_request()
1912 mad->iu.cmd.len = cpu_to_be32(sg_dma_len(job->request_payload.sg_list)); in ibmvfc_bsg_request()
1913 mad->iu.rsp.va = cpu_to_be64(sg_dma_address(job->reply_payload.sg_list)); in ibmvfc_bsg_request()
1914 mad->iu.rsp.len = cpu_to_be32(sg_dma_len(job->reply_payload.sg_list)); in ibmvfc_bsg_request()
1943 dma_unmap_sg(vhost->dev, job->request_payload.sg_list, in ibmvfc_bsg_request()
1945 dma_unmap_sg(vhost->dev, job->reply_payload.sg_list, in ibmvfc_bsg_request()
/linux-4.1.27/arch/x86/platform/efi/
Defi_64.c547 unsigned long count, unsigned long sg_list) in efi_thunk_update_capsule() argument
/linux-4.1.27/drivers/s390/scsi/
Dzfcp_fc.c993 ct_els->req = job->request_payload.sg_list; in zfcp_fc_exec_bsg_job()
994 ct_els->resp = job->reply_payload.sg_list; in zfcp_fc_exec_bsg_job()
/linux-4.1.27/drivers/char/
Dvirtio_console.c868 struct sg_list { struct
878 struct sg_list *sgl = sd->u.data; in pipe_to_sg() argument
924 struct sg_list sgl; in port_fops_splice_write()
/linux-4.1.27/drivers/crypto/amcc/
Dcrypto4xx_core.c747 static int get_sg_count(struct scatterlist *sg_list, int nbytes) in get_sg_count() argument
749 struct scatterlist *sg = sg_list; in get_sg_count()
/linux-4.1.27/include/rdma/
Dib_verbs.h1031 struct ib_sge *sg_list; member
1090 struct ib_sge *sg_list; member
/linux-4.1.27/drivers/scsi/bfa/
Dbfad_bsg.c3156 sg_copy_to_buffer(job->request_payload.sg_list, in bfad_im_bsg_vendor_request()
3167 sg_copy_from_buffer(job->reply_payload.sg_list, in bfad_im_bsg_vendor_request()
3469 sg_copy_to_buffer(job->request_payload.sg_list, in bfad_im_bsg_els_ct_request()
3530 sg_copy_from_buffer(job->reply_payload.sg_list, in bfad_im_bsg_els_ct_request()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c168 rx->rx_wrq.sg_list = &rx->rx_sge; in kiblnd_post_rx()
1077 wrq->sg_list = sge; in kiblnd_init_tx_msg()
1141 wrq->sg_list = sge; in kiblnd_init_rdma()
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.c789 wr.sg_list = &list; in srpt_post_recv()
825 wr.sg_list = &list; in srpt_post_send()
2818 wr.sg_list = riu->sge; in srpt_perform_rdmas()
/linux-4.1.27/drivers/scsi/be2iscsi/
Dbe_mgmt.c511 sg_copy_to_buffer(job->request_payload.sg_list, in mgmt_vendor_specific_fw_cmd()
Dbe_main.c5156 sg_copy_from_buffer(job->reply_payload.sg_list, in beiscsi_bsg_request()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.c1696 wr.sg_list = &list; in srp_post_send()
1716 wr.sg_list = &list; in srp_post_recv()