Home
last modified time | relevance | path

Searched refs:sge (Results 1 – 99 of 99) sorted by relevance

/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/
Dsge.c254 struct sge { struct
282 static void tx_sched_stop(struct sge *sge) in tx_sched_stop() argument
284 struct sched *s = sge->tx_sched; in tx_sched_stop()
297 unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port, in t1_sched_update_parms() argument
300 struct sched *s = sge->tx_sched; in t1_sched_update_parms()
320 if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) { in t1_sched_update_parms()
343 void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val)
345 struct sched *s = sge->tx_sched;
350 t1_sched_update_parms(sge, i, 0, 0);
357 void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port,
[all …]
Dsge.h71 struct sge;
73 struct sge *t1_sge_create(struct adapter *, struct sge_params *);
74 int t1_sge_configure(struct sge *, struct sge_params *);
75 int t1_sge_set_coalesce_params(struct sge *, struct sge_params *);
76 void t1_sge_destroy(struct sge *);
82 void t1_sge_start(struct sge *);
83 void t1_sge_stop(struct sge *);
84 int t1_sge_intr_error_handler(struct sge *);
85 void t1_sge_intr_enable(struct sge *);
86 void t1_sge_intr_disable(struct sge *);
[all …]
Dcxgb2.c174 t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed); in t1_link_negotiated()
228 t1_sge_start(adapter->sge); in cxgb_up()
239 t1_sge_stop(adapter->sge); in cxgb_down()
467 t = t1_sge_get_intr_counts(adapter->sge); in get_stats()
468 t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss); in get_stats()
716 e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl]; in get_sge_param()
717 e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl]; in get_sge_param()
718 e->tx_pending = adapter->params.sge.cmdQ_size[0]; in get_sge_param()
737 adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending; in set_sge_param()
738 adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending; in set_sge_param()
[all …]
Dcommon.h185 struct sge_params sge; member
224 struct sge;
242 struct sge *sge; member
Dsubr.c219 t1_sge_intr_error_handler(adapter->sge); in fpga_slow_intr()
761 t1_sge_intr_enable(adapter->sge); in t1_interrupts_enable()
793 t1_sge_intr_disable(adapter->sge); in t1_interrupts_disable()
819 t1_sge_intr_clear(adapter->sge); in t1_interrupts_clear()
853 t1_sge_intr_error_handler(adapter->sge); in asic_slow_intr()
982 err = t1_sge_configure(adapter->sge, &adapter->params.sge); in t1_init_hw_modules()
1022 if (adapter->sge) in t1_free_sw_modules()
1023 t1_sge_destroy(adapter->sge); in t1_free_sw_modules()
1059 adapter->sge = t1_sge_create(adapter, &adapter->params.sge); in t1_init_sw_modules()
1060 if (!adapter->sge) { in t1_init_sw_modules()
DMakefile8 cxgb-objs := cxgb2.o espi.o tp.o pm3393.o sge.o subr.o \
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_keys.c122 struct ib_sge *sge, int acc) in ipath_lkey_ok() argument
134 if (sge->lkey == 0) { in ipath_lkey_ok()
143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok()
144 isge->length = sge->length; in ipath_lkey_ok()
145 isge->sge_length = sge->length; in ipath_lkey_ok()
149 mr = rkt->table[(sge->lkey >> (32 - ib_ipath_lkey_table_size))]; in ipath_lkey_ok()
150 if (unlikely(mr == NULL || mr->lkey != sge->lkey || in ipath_lkey_ok()
156 off = sge->addr - mr->user_base; in ipath_lkey_ok()
157 if (unlikely(sge->addr < mr->user_base || in ipath_lkey_ok()
158 off + sge->length > mr->length || in ipath_lkey_ok()
[all …]
Dipath_verbs.c173 struct ipath_sge *sge = &ss->sge; in ipath_copy_sge() local
176 u32 len = sge->length; in ipath_copy_sge()
180 if (len > sge->sge_length) in ipath_copy_sge()
181 len = sge->sge_length; in ipath_copy_sge()
183 memcpy(sge->vaddr, data, len); in ipath_copy_sge()
184 sge->vaddr += len; in ipath_copy_sge()
185 sge->length -= len; in ipath_copy_sge()
186 sge->sge_length -= len; in ipath_copy_sge()
187 if (sge->sge_length == 0) { in ipath_copy_sge()
189 *sge = *ss->sg_list++; in ipath_copy_sge()
[all …]
Dipath_ruc.c134 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, in ipath_init_sge()
265 struct ipath_sge *sge; in ipath_ruc_loopback() local
328 sqp->s_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
370 qp->r_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
385 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in ipath_ruc_loopback()
387 *(u64 *) sqp->s_sge.sge.vaddr = in ipath_ruc_loopback()
390 (u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr, in ipath_ruc_loopback()
399 sge = &sqp->s_sge.sge; in ipath_ruc_loopback()
403 if (len > sge->length) in ipath_ruc_loopback()
404 len = sge->length; in ipath_ruc_loopback()
[all …]
Dipath_ud.c59 struct ipath_sge *sge; in ipath_ud_loopback() local
184 sge = swqe->sg_list; in ipath_ud_loopback()
186 u32 len = sge->length; in ipath_ud_loopback()
190 if (len > sge->sge_length) in ipath_ud_loopback()
191 len = sge->sge_length; in ipath_ud_loopback()
193 ipath_copy_sge(&rsge, sge->vaddr, len); in ipath_ud_loopback()
194 sge->vaddr += len; in ipath_ud_loopback()
195 sge->length -= len; in ipath_ud_loopback()
196 sge->sge_length -= len; in ipath_ud_loopback()
197 if (sge->sge_length == 0) { in ipath_ud_loopback()
[all …]
Dipath_sdma.c670 struct ipath_sge *sge; in ipath_sdma_verbs_send() local
729 sge = &ss->sge; in ipath_sdma_verbs_send()
735 if (len > sge->length) in ipath_sdma_verbs_send()
736 len = sge->length; in ipath_sdma_verbs_send()
737 if (len > sge->sge_length) in ipath_sdma_verbs_send()
738 len = sge->sge_length; in ipath_sdma_verbs_send()
741 addr = dma_map_single(&dd->pcidev->dev, sge->vaddr, dw << 2, in ipath_sdma_verbs_send()
759 sge->vaddr += len; in ipath_sdma_verbs_send()
760 sge->length -= len; in ipath_sdma_verbs_send()
761 sge->sge_length -= len; in ipath_sdma_verbs_send()
[all …]
Dipath_rc.c48 ss->sge = wqe->sg_list[0]; in restart_sge()
129 len = e->rdma_sge.sge.sge_length; in ipath_make_rc_ack()
162 len = qp->s_ack_rdma_sge.sge.sge_length; in ipath_make_rc_ack()
461 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_rc_req()
1439 if (unlikely(offset + len > e->rdma_sge.sge.sge_length)) in ipath_rc_rcv_error()
1454 e->rdma_sge.sge.mr = NULL; in ipath_rc_rcv_error()
1455 e->rdma_sge.sge.vaddr = NULL; in ipath_rc_rcv_error()
1456 e->rdma_sge.sge.length = 0; in ipath_rc_rcv_error()
1457 e->rdma_sge.sge.sge_length = 0; in ipath_rc_rcv_error()
1780 qp->r_sge.sge.mr = NULL; in ipath_rc_rcv()
[all …]
Dipath_uc.c98 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_uc_req()
454 qp->r_sge.sge.mr = NULL; in ipath_uc_rcv()
455 qp->r_sge.sge.vaddr = NULL; in ipath_uc_rcv()
456 qp->r_sge.sge.length = 0; in ipath_uc_rcv()
457 qp->r_sge.sge.sge_length = 0; in ipath_uc_rcv()
Dipath_verbs.h330 struct ipath_sge sge; /* progress state for the current SGE */ member
786 struct ib_sge *sge, int acc);
/linux-4.1.27/drivers/scsi/esas2r/
Desas2r_io.c222 if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) { in esas2r_build_sg_list_sge()
237 sgelen = (u8)((u8 *)sgc->sge.a64.curr in esas2r_build_sg_list_sge()
238 - (u8 *)sgc->sge.a64.last); in esas2r_build_sg_list_sge()
244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge()
247 sgc->sge.a64.curr = in esas2r_build_sg_list_sge()
252 sgc->sge.a64.limit = in esas2r_build_sg_list_sge()
257 sgc->sge.a64.last->length = cpu_to_le32( in esas2r_build_sg_list_sge()
259 sgc->sge.a64.last->address = in esas2r_build_sg_list_sge()
268 if (sgc->sge.a64.chain) { in esas2r_build_sg_list_sge()
269 sgc->sge.a64.chain->length |= in esas2r_build_sg_list_sge()
[all …]
Desas2r_vda.c119 firstsg = rq->vrq->flash.data.file.sge; in esas2r_process_vda_ioctl()
130 firstsg = rq->vrq->cli.sge; in esas2r_process_vda_ioctl()
194 firstsg = rq->vrq->mgt.sge; in esas2r_process_vda_ioctl()
363 data.sge); in esas2r_build_flash_req()
394 struct atto_vda_mgmt_req, sge); in esas2r_build_mgt_req()
396 vrq->sge[0].length = cpu_to_le32(SGE_LAST | length); in esas2r_build_mgt_req()
397 vrq->sge[0].address = cpu_to_le64( in esas2r_build_mgt_req()
432 (u8)offsetof(struct atto_vda_ae_req, sge); in esas2r_build_ae_req()
433 vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length); in esas2r_build_ae_req()
434 vrq->sge[0].address = cpu_to_le64( in esas2r_build_ae_req()
[all …]
Datvda.h504 struct atto_vda_sge sge[1]; member
557 struct atto_vda_sge sge[1]; member
1072 struct atto_vda_sge sge[1]; member
1095 struct atto_vda_sge sge[1]; member
1098 struct atto_vda_sge sge[1]; member
1120 struct atto_vda_sge sge[1]; member
1133 struct atto_vda_sge sge[1]; member
1147 struct atto_vda_sge sge[1]; member
1174 struct atto_vda_sge sge[1]; member
1191 struct atto_vda_sge sge; member
[all …]
Desas2r.h619 } sge; member
1180 sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq in esas2r_sgc_init()
1187 sgc->sge.a64.last = in esas2r_sgc_init()
1188 sgc->sge.a64.curr = first; in esas2r_sgc_init()
1193 sgc->sge.a64.last = in esas2r_sgc_init()
1194 sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0]; in esas2r_sgc_init()
1196 (u8)offsetof(struct atto_vda_scsi_req, u.sge); in esas2r_sgc_init()
1198 sgc->sge.a64.chain = NULL; in esas2r_sgc_init()
Desas2r_flash.c212 esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]); in build_flash_msg()
915 esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge); in esas2r_process_fs_ioctl()
1311 vrq->data.sge[0].length = in esas2r_nvram_write()
1314 vrq->data.sge[0].address = cpu_to_le64( in esas2r_nvram_write()
Desas2r_ioctl.c300 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in smp_ioctl_callback()
353 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in csmi_ioctl_tunnel()
672 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in hba_ioctl_tunnel()
Desas2r_disc.c901 esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge); in esas2r_disc_passthru_dev_addr()
Desas2r_init.c1280 data.sge); in esas2r_format_init_msg()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_verbs.c171 struct qib_sge *sge = &ss->sge; in qib_copy_sge() local
174 u32 len = sge->length; in qib_copy_sge()
178 if (len > sge->sge_length) in qib_copy_sge()
179 len = sge->sge_length; in qib_copy_sge()
181 memcpy(sge->vaddr, data, len); in qib_copy_sge()
182 sge->vaddr += len; in qib_copy_sge()
183 sge->length -= len; in qib_copy_sge()
184 sge->sge_length -= len; in qib_copy_sge()
185 if (sge->sge_length == 0) { in qib_copy_sge()
187 qib_put_mr(sge->mr); in qib_copy_sge()
[all …]
Dqib_keys.c156 struct qib_sge *isge, struct ib_sge *sge, int acc) in qib_lkey_ok() argument
167 if (sge->lkey == 0) { in qib_lkey_ok()
180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok()
181 isge->length = sge->length; in qib_lkey_ok()
182 isge->sge_length = sge->length; in qib_lkey_ok()
188 rkt->table[(sge->lkey >> (32 - ib_qib_lkey_table_size))]); in qib_lkey_ok()
189 if (unlikely(!mr || mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) in qib_lkey_ok()
192 off = sge->addr - mr->user_base; in qib_lkey_ok()
193 if (unlikely(sge->addr < mr->user_base || in qib_lkey_ok()
194 off + sge->length > mr->length || in qib_lkey_ok()
[all …]
Dqib_ruc.c98 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in qib_init_sge()
111 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in qib_init_sge() local
113 qib_put_mr(sge->mr); in qib_init_sge()
360 struct qib_sge *sge; in qib_ruc_loopback() local
427 sqp->s_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback()
460 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, wqe->length, in qib_ruc_loopback()
473 if (unlikely(!qib_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, in qib_ruc_loopback()
481 qp->r_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback()
491 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), in qib_ruc_loopback()
497 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in qib_ruc_loopback()
[all …]
Dqib_ud.c57 struct qib_sge *sge; in qib_ud_loopback() local
173 ssge.sge = *swqe->sg_list; in qib_ud_loopback()
175 sge = &ssge.sge; in qib_ud_loopback()
177 u32 len = sge->length; in qib_ud_loopback()
181 if (len > sge->sge_length) in qib_ud_loopback()
182 len = sge->sge_length; in qib_ud_loopback()
184 qib_copy_sge(&qp->r_sge, sge->vaddr, len, 1); in qib_ud_loopback()
185 sge->vaddr += len; in qib_ud_loopback()
186 sge->length -= len; in qib_ud_loopback()
187 sge->sge_length -= len; in qib_ud_loopback()
[all …]
Dqib_sdma.c538 struct qib_sge *sge; in qib_sdma_verbs_send() local
586 sge = &ss->sge; in qib_sdma_verbs_send()
592 if (len > sge->length) in qib_sdma_verbs_send()
593 len = sge->length; in qib_sdma_verbs_send()
594 if (len > sge->sge_length) in qib_sdma_verbs_send()
595 len = sge->sge_length; in qib_sdma_verbs_send()
598 addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr, in qib_sdma_verbs_send()
617 sge->vaddr += len; in qib_sdma_verbs_send()
618 sge->length -= len; in qib_sdma_verbs_send()
619 sge->sge_length -= len; in qib_sdma_verbs_send()
[all …]
Dqib_uc.c99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req()
441 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, in qib_uc_rcv()
448 qp->r_sge.sge.mr = NULL; in qib_uc_rcv()
449 qp->r_sge.sge.vaddr = NULL; in qib_uc_rcv()
450 qp->r_sge.sge.length = 0; in qib_uc_rcv()
451 qp->r_sge.sge.sge_length = 0; in qib_uc_rcv()
Dqib_rc.c49 ss->sge = wqe->sg_list[0]; in restart_sge()
137 qp->s_ack_rdma_sge.sge = e->rdma_sge; in qib_make_rc_ack()
173 qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; in qib_make_rc_ack()
176 len = qp->s_ack_rdma_sge.sge.sge_length; in qib_make_rc_ack()
487 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_rc_req()
1013 struct qib_sge *sge = &wqe->sg_list[i]; in qib_rc_send_complete() local
1015 qib_put_mr(sge->mr); in qib_rc_send_complete()
1069 struct qib_sge *sge = &wqe->sg_list[i]; in do_rc_completion() local
1071 qib_put_mr(sge->mr); in do_rc_completion()
2070 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, in qib_rc_rcv()
[all …]
Dqib_verbs.h392 struct qib_sge sge; /* progress state for the current SGE */ member
984 struct qib_sge *isge, struct ib_sge *sge, int acc);
986 int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
1069 qib_put_mr(ss->sge.mr); in qib_put_ss()
1071 ss->sge = *ss->sg_list++; in qib_put_ss()
Dqib_qp.c433 struct qib_sge *sge = &wqe->sg_list[i]; in clear_mr_refs() local
435 qib_put_mr(sge->mr); in clear_mr_refs()
/linux-4.1.27/drivers/scsi/csiostor/
Dcsio_wr.c57 csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg) in csio_get_flbuf_size() argument
59 sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A + in csio_get_flbuf_size()
65 csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf) in csio_wr_fl_bufsz() argument
67 return sge->sge_fl_buf_size[buf->paddr & 0xF]; in csio_wr_fl_bufsz()
74 return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ? 128 : 64; in csio_wr_qstat_pgsz()
117 struct csio_sge *sge = &wrm->sge; in csio_wr_fill_fl() local
125 buf->len = sge->sge_fl_buf_size[sreg]; in csio_wr_fill_fl()
1044 struct csio_sge *sge = &wrm->sge; in csio_wr_process_fl() local
1070 bufsz = csio_wr_fl_bufsz(sge, buf); in csio_wr_process_fl()
1091 flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align); in csio_wr_process_fl()
[all …]
Dcsio_wr.h449 struct csio_sge sge; /* SGE params */ member
/linux-4.1.27/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c73 vec->sge[sge_no].iov_base = xdr->head[0].iov_base; in map_xdr()
74 vec->sge[sge_no].iov_len = xdr->head[0].iov_len; in map_xdr()
82 vec->sge[sge_no].iov_base = in map_xdr()
86 vec->sge[sge_no].iov_len = sge_bytes; in map_xdr()
95 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr()
96 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr()
148 struct ib_sge *sge; in send_write() local
164 write_len, vec->sge, vec->count); in send_write()
168 sge = ctxt->sge; in send_write()
173 if (vec->sge[xdr_sge_no].iov_len > bc) in send_write()
[all …]
Dsvc_rdma_recvfrom.c75 min_t(size_t, byte_count, ctxt->sge[0].length); in rdma_build_arg_xdr()
98 bc -= min_t(u32, bc, ctxt->sge[sge_no].length); in rdma_build_arg_xdr()
99 rqstp->rq_arg.buflen += ctxt->sge[sge_no].length; in rdma_build_arg_xdr()
162 ctxt->sge[pno].addr = in rdma_read_chunk_lcl()
168 ctxt->sge[pno].addr); in rdma_read_chunk_lcl()
174 ctxt->sge[pno].lkey = xprt->sc_dma_lkey; in rdma_read_chunk_lcl()
175 ctxt->sge[pno].length = len; in rdma_read_chunk_lcl()
199 read_wr.sg_list = ctxt->sge; in rdma_read_chunk_lcl()
296 ctxt->sge[0].addr = (unsigned long)frmr->kva + *page_offset; in rdma_read_chunk_frmr()
297 ctxt->sge[0].lkey = frmr->mr->lkey; in rdma_read_chunk_frmr()
[all …]
Dsvc_rdma_transport.c120 for (i = 0; i < ctxt->count && ctxt->sge[i].length; i++) { in svc_rdma_unmap_dma()
127 if (ctxt->sge[i].lkey == xprt->sc_dma_lkey) { in svc_rdma_unmap_dma()
130 ctxt->sge[i].addr, in svc_rdma_unmap_dma()
131 ctxt->sge[i].length, in svc_rdma_unmap_dma()
531 ctxt->sge[sge_no].addr = pa; in svc_rdma_post_recv()
532 ctxt->sge[sge_no].length = PAGE_SIZE; in svc_rdma_post_recv()
533 ctxt->sge[sge_no].lkey = xprt->sc_dma_lkey; in svc_rdma_post_recv()
538 recv_wr.sg_list = &ctxt->sge[0]; in svc_rdma_post_recv()
1338 ctxt->sge[0].addr = ib_dma_map_page(xprt->sc_cm_id->device, in svc_rdma_send_error()
1340 if (ib_dma_mapping_error(xprt->sc_cm_id->device, ctxt->sge[0].addr)) { in svc_rdma_send_error()
[all …]
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c170 struct sge *s = &adapter->sge; in fl_mtu_bufsize()
257 const struct sge *s = &adapter->sge; in fl_starving()
340 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl()
455 struct sge *s = &adapter->sge; in get_buf_size()
589 struct sge *s = &adap->sge; in refill_fl()
668 set_bit(q->cntxt_id - adap->sge.egr_start, in refill_fl()
669 adap->sge.starving_fl); in refill_fl()
881 to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; in write_sgl()
895 unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1; in write_sgl()
898 memcpy(sgl->sge, buf, part0); in write_sgl()
[all …]
Dcxgb4_main.c301 struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; in dcb_tx_queue_prio_enable()
614 txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; in fwevtq_handler()
616 if ((u8 *)txq < (u8 *)q->adap->sge.ofldtxq) { in fwevtq_handler()
757 for_each_ofldrxq(&adap->sge, i) in name_msix_vecs()
761 for_each_rdmarxq(&adap->sge, i) in name_msix_vecs()
765 for_each_rdmaciq(&adap->sge, i) in name_msix_vecs()
772 struct sge *s = &adap->sge; in request_msix_queue_irqs()
839 struct sge *s = &adap->sge; in free_msix_queue_irqs()
864 const struct sge_eth_rxq *q = &pi->adapter->sge.ethrxq[pi->first_qset]; in cxgb4_write_rss()
903 static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) in rxq_to_chan()
[all …]
Dcxgb4_debugfs.c1684 int eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4); in sge_qinfo_show()
1685 int toe_entries = DIV_ROUND_UP(adap->sge.ofldqsets, 4); in sge_qinfo_show()
1686 int rdma_entries = DIV_ROUND_UP(adap->sge.rdmaqs, 4); in sge_qinfo_show()
1687 int ciq_entries = DIV_ROUND_UP(adap->sge.rdmaciqs, 4); in sge_qinfo_show()
1712 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[base_qset]; in sge_qinfo_show()
1713 const struct sge_eth_txq *tx = &adap->sge.ethtxq[base_qset]; in sge_qinfo_show()
1714 int n = min(4, adap->sge.ethqsets - 4 * r); in sge_qinfo_show()
1740 adap->sge.counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show()
1748 const struct sge_ofld_rxq *rx = &adap->sge.ofldrxq[toe_idx * 4]; in sge_qinfo_show()
1749 const struct sge_ofld_txq *tx = &adap->sge.ofldtxq[toe_idx * 4]; in sge_qinfo_show()
[all …]
Dcxgb4_ethtool.c193 const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; in collect_sge_port_stats()
194 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset]; in collect_sge_port_stats()
483 const struct sge *s = &pi->adapter->sge; in get_sge_param()
501 struct sge *s = &adapter->sge; in set_sge_param()
535 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_rx_intr_params()
550 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_adaptive_rx_setting()
562 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in get_adaptive_rx_setting()
578 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
582 adap->sge.counter_val[rq->pktcnt_idx] : 0; in get_coalesce()
Dcxgb4.h296 struct sge_params sge; member
602 struct sge { struct
651 #define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++) argument
652 #define for_each_ofldrxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++) argument
653 #define for_each_rdmarxq(sge, i) for (i = 0; i < (sge)->rdmaqs; i++) argument
654 #define for_each_rdmaciq(sge, i) for (i = 0; i < (sge)->rdmaciqs; i++) argument
692 struct sge sge; member
1036 return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; in qtimer_val()
DMakefile7 cxgb4-objs := cxgb4_main.o l2t.o t4_hw.o sge.o clip_tbl.o cxgb4_ethtool.o
Dl2t.c154 TID_QID_V(adap->sge.fw_evtq.abs_id))); in write_l2e()
Dt4_msg.h1054 struct ulptx_sge_pair sge[0]; member
Dt4_hw.c5142 page_shift = adapter->params.sge.hps + 10; in cxgb4_t4_bar2_sge_qregs()
5148 ? adapter->params.sge.eq_qpp in cxgb4_t4_bar2_sge_qregs()
5149 : adapter->params.sge.iq_qpp); in cxgb4_t4_bar2_sge_qregs()
5250 struct sge_params *sge_params = &adapter->params.sge; in t4_init_sge_params()
/linux-4.1.27/net/rds/
Diw_recv.c88 struct ib_sge *sge; in rds_iw_recv_init_ring() local
98 sge = rds_iw_data_sge(ic, recv->r_sge); in rds_iw_recv_init_ring()
99 sge->addr = 0; in rds_iw_recv_init_ring()
100 sge->length = RDS_FRAG_SIZE; in rds_iw_recv_init_ring()
101 sge->lkey = 0; in rds_iw_recv_init_ring()
103 sge = rds_iw_header_sge(ic, recv->r_sge); in rds_iw_recv_init_ring()
104 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_recv_init_ring()
105 sge->length = sizeof(struct rds_header); in rds_iw_recv_init_ring()
106 sge->lkey = 0; in rds_iw_recv_init_ring()
143 struct ib_sge *sge; in rds_iw_recv_refill_one() local
[all …]
Dib_recv.c52 struct ib_sge *sge; in rds_ib_recv_init_ring() local
62 sge = &recv->r_sge[0]; in rds_ib_recv_init_ring()
63 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_recv_init_ring()
64 sge->length = sizeof(struct rds_header); in rds_ib_recv_init_ring()
65 sge->lkey = ic->i_mr->lkey; in rds_ib_recv_init_ring()
67 sge = &recv->r_sge[1]; in rds_ib_recv_init_ring()
68 sge->addr = 0; in rds_ib_recv_init_ring()
69 sge->length = RDS_FRAG_SIZE; in rds_ib_recv_init_ring()
70 sge->lkey = ic->i_mr->lkey; in rds_ib_recv_init_ring()
303 struct ib_sge *sge; in rds_ib_recv_refill_one() local
[all …]
Diw_send.c134 struct ib_sge *sge; in rds_iw_send_init_ring() local
148 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_send_init_ring()
149 sge->lkey = 0; in rds_iw_send_init_ring()
151 sge = rds_iw_header_sge(ic, send->s_sge); in rds_iw_send_init_ring()
152 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_send_init_ring()
153 sge->length = sizeof(struct rds_header); in rds_iw_send_init_ring()
154 sge->lkey = 0; in rds_iw_send_init_ring()
453 struct ib_sge *sge; in rds_iw_xmit_populate_wr() local
465 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_xmit_populate_wr()
466 sge->addr = buffer; in rds_iw_xmit_populate_wr()
[all …]
Diw.h384 rds_iw_header_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_header_sge() argument
386 return &sge[0]; in rds_iw_header_sge()
390 rds_iw_data_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_data_sge() argument
392 return &sge[1]; in rds_iw_data_sge()
Dib_send.c228 struct ib_sge *sge; in rds_ib_send_init_ring() local
236 sge = &send->s_sge[0]; in rds_ib_send_init_ring()
237 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_send_init_ring()
238 sge->length = sizeof(struct rds_header); in rds_ib_send_init_ring()
239 sge->lkey = ic->i_mr->lkey; in rds_ib_send_init_ring()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c323 struct sge *s = &adapter->sge; in request_msix_queue_irqs()
361 struct sge *s = &adapter->sge; in free_msix_queue_irqs()
394 struct sge *s = &adapter->sge; in enable_rx()
417 struct sge *s = &adapter->sge; in quiesce_rx()
475 struct sge *s = &adapter->sge; in fwevtq_handler()
531 struct sge *s = &adapter->sge; in setup_sge_queues()
646 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; in setup_rss()
986 static int closest_timer(const struct sge *s, int us) in closest_timer()
1002 static int closest_thres(const struct sge *s, int thres) in closest_thres()
1027 ? adapter->sge.timer_val[timer_idx] in qtimer_val()
[all …]
Dsge.c264 const struct sge *s = &adapter->sge; in fl_starving()
325 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl()
457 const struct sge *s = &adapter->sge; in get_buf_size()
609 struct sge *s = &adapter->sge; in refill_fl()
717 set_bit(fl->cntxt_id, adapter->sge.starving_fl); in refill_fl()
937 to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; in write_sgl()
951 unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; in write_sgl()
954 memcpy(sgl->sge, buf, part0); in write_sgl()
1193 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit()
1555 struct sge *s = &adapter->sge; in do_gro()
[all …]
Dadapter.h269 struct sge { struct
348 #define for_each_ethrxq(sge, iter) \ argument
349 for (iter = 0; iter < (sge)->ethqsets; iter++)
372 struct sge sge; member
DMakefile7 cxgb4vf-objs := cxgb4vf_main.o t4vf_hw.o sge.o
Dt4vf_common.h214 struct sge_params sge; /* Scatter Gather Engine */ member
Dt4vf_hw.c472 page_shift = adapter->params.sge.sge_vf_hps + 10; in t4_bar2_sge_qregs()
478 ? adapter->params.sge.sge_vf_eq_qpp in t4_bar2_sge_qregs()
479 : adapter->params.sge.sge_vf_iq_qpp); in t4_bar2_sge_qregs()
529 struct sge_params *sge_params = &adapter->params.sge; in t4vf_get_sge_params()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_memory.c396 reg->sge.lkey = device->mr->lkey; in iser_reg_dma()
398 reg->sge.addr = ib_sg_dma_address(device->ib_device, &sg[0]); in iser_reg_dma()
399 reg->sge.length = ib_sg_dma_len(device->ib_device, &sg[0]); in iser_reg_dma()
402 " length=0x%x\n", reg->sge.lkey, reg->rkey, in iser_reg_dma()
403 reg->sge.addr, reg->sge.length); in iser_reg_dma()
474 mem_reg->sge.lkey = fmr->fmr->lkey; in iser_reg_page_vec()
476 mem_reg->sge.addr = page_vec->pages[0] + page_vec->offset; in iser_reg_page_vec()
477 mem_reg->sge.length = page_vec->data_size; in iser_reg_page_vec()
678 sig_wr.sg_list = &data_reg->sge; in iser_reg_sig_mr()
683 sig_wr.wr.sig_handover.prot = &prot_reg->sge; in iser_reg_sig_mr()
[all …]
Diser_initiator.c85 hdr->read_va = cpu_to_be64(mem_reg->sge.addr); in iser_prepare_read_cmd()
89 (unsigned long long)mem_reg->sge.addr); in iser_prepare_read_cmd()
142 hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz); in iser_prepare_write_cmd()
147 (unsigned long long)mem_reg->sge.addr, unsol_sz); in iser_prepare_write_cmd()
153 tx_dsg->addr = mem_reg->sge.addr; in iser_prepare_write_cmd()
155 tx_dsg->lkey = mem_reg->sge.lkey; in iser_prepare_write_cmd()
484 tx_dsg->addr = mem_reg->sge.addr + buf_offset; in iser_send_data_out()
486 tx_dsg->lkey = mem_reg->sge.lkey; in iser_send_data_out()
Diser_verbs.c1007 struct ib_sge sge; in iser_post_recvl() local
1010 sge.addr = iser_conn->login_resp_dma; in iser_post_recvl()
1011 sge.length = ISER_RX_LOGIN_SIZE; in iser_post_recvl()
1012 sge.lkey = ib_conn->device->mr->lkey; in iser_post_recvl()
1015 rx_wr.sg_list = &sge; in iser_post_recvl()
Discsi_iser.h251 struct ib_sge sge; member
/linux-4.1.27/net/9p/
Dtrans_rdma.c408 struct ib_sge sge; in post_recv() local
416 sge.addr = c->busa; in post_recv()
417 sge.length = client->msize; in post_recv()
418 sge.lkey = rdma->lkey; in post_recv()
423 wr.sg_list = &sge; in post_recv()
436 struct ib_sge sge; in rdma_request() local
507 sge.addr = c->busa; in rdma_request()
508 sge.length = c->req->tc->size; in rdma_request()
509 sge.lkey = rdma->lkey; in rdma_request()
516 wr.sg_list = &sge; in rdma_request()
/linux-4.1.27/drivers/scsi/be2iscsi/
Dbe_mgmt.c282 struct be_sge *sge; in mgmt_get_session_info() local
299 sge = nonembedded_sgl(wrb); in mgmt_get_session_info()
309 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); in mgmt_get_session_info()
310 sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); in mgmt_get_session_info()
311 sge->len = cpu_to_le32(nonemb_cmd->size); in mgmt_get_session_info()
428 struct be_sge *sge = nonembedded_sgl(wrb); in mgmt_check_supported_fw() local
448 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); in mgmt_check_supported_fw()
449 sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF); in mgmt_check_supported_fw()
450 sge->len = cpu_to_le32(nonemb_cmd.size); in mgmt_check_supported_fw()
591 struct be_sge *sge; in mgmt_invalidate_icds() local
[all …]
/linux-4.1.27/fs/exofs/
Dore_raid.c273 struct osd_sg_entry *sge; in _ore_add_sg_seg() local
282 sge = per_dev->sglist; in _ore_add_sg_seg()
287 sge->offset = per_dev->offset; in _ore_add_sg_seg()
288 sge->len = per_dev->length; in _ore_add_sg_seg()
300 sge = &per_dev->sglist[per_dev->cur_sg - 1]; in _ore_add_sg_seg()
301 sge->len = per_dev->length - per_dev->last_sgs_total; in _ore_add_sg_seg()
306 struct osd_sg_entry *next_sge = sge + 1; in _ore_add_sg_seg()
309 next_sge->offset = sge->offset + sge->len + cur_len; in _ore_add_sg_seg()
313 } else if (!sge->len) { in _ore_add_sg_seg()
/linux-4.1.27/drivers/ata/
Dsata_sil24.c257 struct sil24_sge sge[SIL24_MAX_SGE]; member
263 struct sil24_sge sge[SIL24_MAX_SGE]; member
780 struct sil24_sge *sge) in sil24_fill_sg() argument
787 sge->addr = cpu_to_le64(sg_dma_address(sg)); in sil24_fill_sg()
788 sge->cnt = cpu_to_le32(sg_dma_len(sg)); in sil24_fill_sg()
789 sge->flags = 0; in sil24_fill_sg()
791 last_sge = sge; in sil24_fill_sg()
792 sge++; in sil24_fill_sg()
849 struct sil24_sge *sge; in sil24_qc_prep() local
856 sge = cb->ata.sge; in sil24_qc_prep()
[all …]
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c679 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
681 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
692 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
694 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
708 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
710 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
1516 ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb); in t3_mgmt_tx()
1770 struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; in t3_offload_tx()
2029 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in rx_eth()
2131 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in lro_add_page()
[all …]
Dadapter.h216 struct sge { struct
245 struct sge sge; member
Dcxgb3_main.c411 adap->sge.qs[qidx]. in request_msix_data_irqs()
414 &adap->sge.qs[qidx]); in request_msix_data_irqs()
418 &adap->sge.qs[qidx]); in request_msix_data_irqs()
438 &adapter->sge.qs[i]); in free_irq_resources()
448 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
461 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
600 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs()
613 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi()
638 if (adap->sge.qs[i].adap) in quiesce_rx()
639 napi_disable(&adap->sge.qs[i].napi); in quiesce_rx()
[all …]
DMakefile8 xgmac.o sge.o l2t.o cxgb3_offload.o aq100x.o
Dcxgb3_offload.c303 spin_lock_irqsave(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl()
306 spin_unlock_irqrestore(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl()
334 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
341 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
345 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
347 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
352 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
358 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
Dcommon.h382 struct sge_params sge; member
Dt3_hw.c3369 min(adapter->params.sge.max_pkt_size, in t3_init_hw()
3372 min(adapter->params.sge.max_pkt_size, 16384U)); in t3_init_hw()
3389 t3_sge_init(adapter, &adapter->params.sge); in t3_init_hw()
3653 t3_sge_prep(adapter, &adapter->params.sge); in t3_prep_adapter()
/linux-4.1.27/drivers/scsi/lpfc/
Dlpfc_mbox.c1678 nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr); in lpfc_sli4_mbx_sge_set()
1679 nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr); in lpfc_sli4_mbx_sge_set()
1680 nembed_sge->sge[sgentry].length = length; in lpfc_sli4_mbx_sge_set()
1693 struct lpfc_mbx_sge *sge) in lpfc_sli4_mbx_sge_get() argument
1699 sge->pa_lo = nembed_sge->sge[sgentry].pa_lo; in lpfc_sli4_mbx_sge_get()
1700 sge->pa_hi = nembed_sge->sge[sgentry].pa_hi; in lpfc_sli4_mbx_sge_get()
1701 sge->length = nembed_sge->sge[sgentry].length; in lpfc_sli4_mbx_sge_get()
1715 struct lpfc_mbx_sge sge; in lpfc_sli4_mbox_cmd_free() local
1736 lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge); in lpfc_sli4_mbox_cmd_free()
1737 phyaddr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_mbox_cmd_free()
[all …]
Dlpfc_hbadisc.c1803 struct lpfc_mbx_sge sge; in lpfc_sli4_fcf_rec_mbox_parse() local
1812 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_fcf_rec_mbox_parse()
1813 phys_addr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_fcf_rec_mbox_parse()
Dlpfc_bsg.c4811 receive_length = nembed_sge->sge[0].length; in lpfc_bsg_issue_mbox()
4823 nembed_sge->sge[0].pa_hi = in lpfc_bsg_issue_mbox()
4826 nembed_sge->sge[0].pa_lo = in lpfc_bsg_issue_mbox()
Dlpfc_hw4.h1024 struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES]; member
Dlpfc_sli.c15942 struct lpfc_mbx_sge sge; in lpfc_sli4_add_fcf_record() local
15973 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_add_fcf_record()
15974 phys_addr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_add_fcf_record()
/linux-4.1.27/drivers/scsi/aacraid/
Daachba.c3352 rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32)); in aac_build_sgraw2()
3353 rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff)); in aac_build_sgraw2()
3355 rio2->sge[i].length = cur_size; in aac_build_sgraw2()
3356 rio2->sge[i].flags = 0; in aac_build_sgraw2()
3373 u32 temp = le32_to_cpu(rio2->sge[i-1].length) - in aac_build_sgraw2()
3375 rio2->sge[i-1].length = cpu_to_le32(temp); in aac_build_sgraw2()
3388 if (rio2->sge[j].length % (i*PAGE_SIZE)) { in aac_build_sgraw2()
3392 nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE)); in aac_build_sgraw2()
3414 struct sge_ieee1212 *sge; in aac_convert_sgraw2() local
3421 sge = kmalloc(nseg_new * sizeof(struct sge_ieee1212), GFP_ATOMIC); in aac_convert_sgraw2()
[all …]
Daacraid.h1476 struct sge_ieee1212 sge[1]; member
/linux-4.1.27/include/linux/sunrpc/
Dsvc_rdma.h85 struct ib_sge sge[RPCSVC_MAXPAGES]; member
119 struct kvec sge[RPCSVC_MAXPAGES]; member
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c119 struct ib_sge *sge = send_wr->sg_list; in trace_send_wr_ud() local
138 u8 *data = __va(sge->addr); in trace_send_wr_ud()
141 idx, j, data, sge->length, sge->lkey); in trace_send_wr_ud()
143 ehca_dmp(data, sge->length, "send_wr#%x sge#%x", in trace_send_wr_ud()
145 sge++; in trace_send_wr_ud()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1943 struct ocrdma_sge *sge, int num_sge, in ocrdma_build_sges() argument
1949 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges()
1950 sge[i].addr_lo = sg_list[i].addr; in ocrdma_build_sges()
1951 sge[i].addr_hi = upper_32_bits(sg_list[i].addr); in ocrdma_build_sges()
1952 sge[i].len = sg_list[i].length; in ocrdma_build_sges()
1956 memset(sge, 0, sizeof(*sge)); in ocrdma_build_sges()
1971 struct ocrdma_sge *sge, in ocrdma_build_inline_sges() argument
1985 dpp_addr = (char *)sge; in ocrdma_build_inline_sges()
1998 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_inline_sges()
2013 struct ocrdma_sge *sge; in ocrdma_build_send() local
[all …]
Docrdma_hw.c1234 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); in ocrdma_mbx_rdma_stats()
1235 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); in ocrdma_mbx_rdma_stats()
1236 mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size; in ocrdma_mbx_rdma_stats()
1282 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); in ocrdma_mbx_get_ctrl_attribs()
1283 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); in ocrdma_mbx_get_ctrl_attribs()
1284 mqe->u.nonemb_req.sge[0].len = dma.size; in ocrdma_mbx_get_ctrl_attribs()
2811 struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge; in ocrdma_mbx_get_dcbx_config()
Docrdma_sli.h267 struct ocrdma_mqe_sge sge[19]; member
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c1133 struct ib_sge sge; in isert_rdma_post_recvl() local
1136 memset(&sge, 0, sizeof(struct ib_sge)); in isert_rdma_post_recvl()
1137 sge.addr = isert_conn->login_req_dma; in isert_rdma_post_recvl()
1138 sge.length = ISER_RX_LOGIN_SIZE; in isert_rdma_post_recvl()
1139 sge.lkey = isert_conn->device->mr->lkey; in isert_rdma_post_recvl()
1142 sge.addr, sge.length, sge.lkey); in isert_rdma_post_recvl()
1146 rx_wr.sg_list = &sge; in isert_rdma_post_recvl()
2577 struct ib_sge *sge) in isert_fast_reg_mr() argument
2589 sge->lkey = device->mr->lkey; in isert_fast_reg_mr()
2590 sge->addr = ib_sg_dma_address(ib_dev, &mem->sg[0]); in isert_fast_reg_mr()
[all …]
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.c1059 kfree(ioctx->rdma_ius[--ioctx->n_rdma].sge); in srpt_unmap_sg_to_ib_sge()
1089 struct ib_sge *sge; in srpt_map_sg_to_ib_sge() local
1173 riu->sge = in srpt_map_sg_to_ib_sge()
1174 kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge()
1176 if (!riu->sge) in srpt_map_sg_to_ib_sge()
1187 riu->sge = kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge()
1189 if (!riu->sge) in srpt_map_sg_to_ib_sge()
1204 sge = riu->sge; in srpt_map_sg_to_ib_sge()
1208 sge->addr = dma_addr; in srpt_map_sg_to_ib_sge()
1209 sge->lkey = ch->sport->sdev->mr->lkey; in srpt_map_sg_to_ib_sge()
[all …]
Dib_srpt.h156 struct ib_sge *sge; member
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c117 struct ib_sge *sge, int id) in ipoib_cm_post_receive_nonsrq() argument
126 sge[i].addr = rx->rx_ring[id].mapping[i]; in ipoib_cm_post_receive_nonsrq()
329 struct ib_sge *sge) in ipoib_cm_init_rx_wr() argument
335 sge[i].lkey = priv->mr->lkey; in ipoib_cm_init_rx_wr()
337 sge[0].length = IPOIB_CM_HEAD_SIZE; in ipoib_cm_init_rx_wr()
339 sge[i].length = PAGE_SIZE; in ipoib_cm_init_rx_wr()
342 wr->sg_list = sge; in ipoib_cm_init_rx_wr()
352 struct ib_sge sge[IPOIB_CM_RX_SG]; in ipoib_cm_nonsrq_init_rx() member
370 ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); in ipoib_cm_nonsrq_init_rx()
392 ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); in ipoib_cm_nonsrq_init_rx()
/linux-4.1.27/drivers/scsi/
Dvmw_pvscsi.c52 struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; member
336 struct PVSCSISGElement *sge; in pvscsi_create_sg() local
340 sge = &ctx->sgl->sge[0]; in pvscsi_create_sg()
342 sge[i].addr = sg_dma_address(sg); in pvscsi_create_sg()
343 sge[i].length = sg_dma_len(sg); in pvscsi_create_sg()
344 sge[i].flags = 0; in pvscsi_create_sg()
/linux-4.1.27/drivers/infiniband/core/
Dmad.c1066 struct ib_sge *sge; in ib_send_mad() local
1076 sge = mad_send_wr->sg_list; in ib_send_mad()
1077 sge[0].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad()
1079 sge[0].length, in ib_send_mad()
1081 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr))) in ib_send_mad()
1084 mad_send_wr->header_mapping = sge[0].addr; in ib_send_mad()
1086 sge[1].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad()
1088 sge[1].length, in ib_send_mad()
1090 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) { in ib_send_mad()
1093 sge[0].length, DMA_TO_DEVICE); in ib_send_mad()
[all …]
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c1055 struct ib_sge *sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_tx_msg() local
1069 sge->lkey = mr->lkey; in kiblnd_init_tx_msg()
1070 sge->addr = tx->tx_msgaddr; in kiblnd_init_tx_msg()
1071 sge->length = nob; in kiblnd_init_tx_msg()
1077 wrq->sg_list = sge; in kiblnd_init_tx_msg()
1091 struct ib_sge *sge = &tx->tx_sge[0]; in kiblnd_init_rdma() local
1132 sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_rdma()
1133 sge->addr = kiblnd_rd_frag_addr(srcrd, srcidx); in kiblnd_init_rdma()
1134 sge->lkey = kiblnd_rd_frag_key(srcrd, srcidx); in kiblnd_init_rdma()
1135 sge->length = wrknob; in kiblnd_init_rdma()
[all …]
/linux-4.1.27/drivers/scsi/bfa/
Dbfa_fcpim.c2461 struct bfi_sge_s *sge, *sgpge; in bfa_ioim_send_ioreq() local
2489 sge = &m->sges[0]; in bfa_ioim_send_ioreq()
2498 sge->sga = *(union bfi_addr_u *) &addr; in bfa_ioim_send_ioreq()
2500 sge->sg_len = pgdlen; in bfa_ioim_send_ioreq()
2501 sge->flags = (ioim->nsges > BFI_SGE_INLINE) ? in bfa_ioim_send_ioreq()
2503 bfa_sge_to_be(sge); in bfa_ioim_send_ioreq()
2504 sge++; in bfa_ioim_send_ioreq()
2545 sge->sga = ioim->sgpg->sgpg_pa; in bfa_ioim_send_ioreq()
2547 sge->sga.a32.addr_lo = 0; in bfa_ioim_send_ioreq()
2548 sge->sga.a32.addr_hi = 0; in bfa_ioim_send_ioreq()
[all …]
Dbfad_bsg.c3197 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sgaddr_cb() local
3200 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sgaddr_cb()
3201 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_req_sgaddr_cb()
3209 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sglen_cb() local
3211 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sglen_cb()
3212 return sge->sg_len; in bfad_fcxp_get_req_sglen_cb()
3219 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sgaddr_cb() local
3222 sge = drv_fcxp->rsp_sge + sgeid; in bfad_fcxp_get_rsp_sgaddr_cb()
3223 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_rsp_sgaddr_cb()
3231 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sglen_cb() local
[all …]
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_cmn.h801 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_free_rx_sge() local
812 sge->addr_hi = 0; in bnx2x_free_rx_sge()
813 sge->addr_lo = 0; in bnx2x_free_rx_sge()
Dbnx2x_cmn.c549 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_alloc_rx_sge() local
568 sge->addr_hi = cpu_to_le32(U64_HI(mapping)); in bnx2x_alloc_rx_sge()
569 sge->addr_lo = cpu_to_le32(U64_LO(mapping)); in bnx2x_alloc_rx_sge()
1331 struct eth_rx_sge *sge; in bnx2x_set_next_page_sgl() local
1333 sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2]; in bnx2x_set_next_page_sgl()
1334 sge->addr_hi = in bnx2x_set_next_page_sgl()
1338 sge->addr_lo = in bnx2x_set_next_page_sgl()
/linux-4.1.27/drivers/message/fusion/
Dmptbase.h334 SGESimple32_t sge[1]; member
338 SGESimple64_t sge[1]; member
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dt4fw_ri_api.h124 struct fw_ri_dsge_pair sge[0]; member
140 struct fw_ri_sge sge[0]; member
Dqp.c410 __be64 *flitp = (__be64 *)isglp->sge; in build_isgl()
/linux-4.1.27/drivers/net/ethernet/emulex/benet/
Dbe_cmds.c705 struct be_sge *sge; in be_wrb_cmd_hdr_prepare() local
716 sge = nonembedded_sgl(wrb); in be_wrb_cmd_hdr_prepare()
717 sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma)); in be_wrb_cmd_hdr_prepare()
718 sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF); in be_wrb_cmd_hdr_prepare()
719 sge->len = cpu_to_le32(mem->size); in be_wrb_cmd_hdr_prepare()
/linux-4.1.27/drivers/net/ethernet/broadcom/
Dcnic_defs.h3097 struct fcoe_bd_ctx sge[3]; member