Home
last modified time | relevance | path

Searched refs:sge (Results 1 – 109 of 109) sorted by relevance

/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb/
Dsge.c254 struct sge { struct
282 static void tx_sched_stop(struct sge *sge) in tx_sched_stop() argument
284 struct sched *s = sge->tx_sched; in tx_sched_stop()
297 unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port, in t1_sched_update_parms() argument
300 struct sched *s = sge->tx_sched; in t1_sched_update_parms()
320 if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) { in t1_sched_update_parms()
343 void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val)
345 struct sched *s = sge->tx_sched;
350 t1_sched_update_parms(sge, i, 0, 0);
357 void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port,
[all …]
Dsge.h71 struct sge;
73 struct sge *t1_sge_create(struct adapter *, struct sge_params *);
74 int t1_sge_configure(struct sge *, struct sge_params *);
75 int t1_sge_set_coalesce_params(struct sge *, struct sge_params *);
76 void t1_sge_destroy(struct sge *);
82 void t1_sge_start(struct sge *);
83 void t1_sge_stop(struct sge *);
84 int t1_sge_intr_error_handler(struct sge *);
85 void t1_sge_intr_enable(struct sge *);
86 void t1_sge_intr_disable(struct sge *);
[all …]
Dcxgb2.c174 t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed); in t1_link_negotiated()
228 t1_sge_start(adapter->sge); in cxgb_up()
239 t1_sge_stop(adapter->sge); in cxgb_down()
467 t = t1_sge_get_intr_counts(adapter->sge); in get_stats()
468 t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss); in get_stats()
716 e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl]; in get_sge_param()
717 e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl]; in get_sge_param()
718 e->tx_pending = adapter->params.sge.cmdQ_size[0]; in get_sge_param()
737 adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending; in set_sge_param()
738 adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending; in set_sge_param()
[all …]
Dcommon.h185 struct sge_params sge; member
224 struct sge;
242 struct sge *sge; member
Dsubr.c219 t1_sge_intr_error_handler(adapter->sge); in fpga_slow_intr()
761 t1_sge_intr_enable(adapter->sge); in t1_interrupts_enable()
793 t1_sge_intr_disable(adapter->sge); in t1_interrupts_disable()
819 t1_sge_intr_clear(adapter->sge); in t1_interrupts_clear()
853 t1_sge_intr_error_handler(adapter->sge); in asic_slow_intr()
982 err = t1_sge_configure(adapter->sge, &adapter->params.sge); in t1_init_hw_modules()
1022 if (adapter->sge) in t1_free_sw_modules()
1023 t1_sge_destroy(adapter->sge); in t1_free_sw_modules()
1059 adapter->sge = t1_sge_create(adapter, &adapter->params.sge); in t1_init_sw_modules()
1060 if (!adapter->sge) { in t1_init_sw_modules()
DMakefile8 cxgb-objs := cxgb2.o espi.o tp.o pm3393.o sge.o subr.o \
/linux-4.4.14/drivers/staging/rdma/ipath/
Dipath_keys.c122 struct ib_sge *sge, int acc) in ipath_lkey_ok() argument
134 if (sge->lkey == 0) { in ipath_lkey_ok()
143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok()
144 isge->length = sge->length; in ipath_lkey_ok()
145 isge->sge_length = sge->length; in ipath_lkey_ok()
149 mr = rkt->table[(sge->lkey >> (32 - ib_ipath_lkey_table_size))]; in ipath_lkey_ok()
150 if (unlikely(mr == NULL || mr->lkey != sge->lkey || in ipath_lkey_ok()
156 off = sge->addr - mr->user_base; in ipath_lkey_ok()
157 if (unlikely(sge->addr < mr->user_base || in ipath_lkey_ok()
158 off + sge->length > mr->length || in ipath_lkey_ok()
[all …]
Dipath_verbs.c173 struct ipath_sge *sge = &ss->sge; in ipath_copy_sge() local
176 u32 len = sge->length; in ipath_copy_sge()
180 if (len > sge->sge_length) in ipath_copy_sge()
181 len = sge->sge_length; in ipath_copy_sge()
183 memcpy(sge->vaddr, data, len); in ipath_copy_sge()
184 sge->vaddr += len; in ipath_copy_sge()
185 sge->length -= len; in ipath_copy_sge()
186 sge->sge_length -= len; in ipath_copy_sge()
187 if (sge->sge_length == 0) { in ipath_copy_sge()
189 *sge = *ss->sg_list++; in ipath_copy_sge()
[all …]
Dipath_ruc.c133 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, in ipath_init_sge()
264 struct ipath_sge *sge; in ipath_ruc_loopback() local
327 sqp->s_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
369 qp->r_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback()
384 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in ipath_ruc_loopback()
386 *(u64 *) sqp->s_sge.sge.vaddr = in ipath_ruc_loopback()
389 (u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr, in ipath_ruc_loopback()
398 sge = &sqp->s_sge.sge; in ipath_ruc_loopback()
402 if (len > sge->length) in ipath_ruc_loopback()
403 len = sge->length; in ipath_ruc_loopback()
[all …]
Dipath_ud.c58 struct ipath_sge *sge; in ipath_ud_loopback() local
183 sge = swqe->sg_list; in ipath_ud_loopback()
185 u32 len = sge->length; in ipath_ud_loopback()
189 if (len > sge->sge_length) in ipath_ud_loopback()
190 len = sge->sge_length; in ipath_ud_loopback()
192 ipath_copy_sge(&rsge, sge->vaddr, len); in ipath_ud_loopback()
193 sge->vaddr += len; in ipath_ud_loopback()
194 sge->length -= len; in ipath_ud_loopback()
195 sge->sge_length -= len; in ipath_ud_loopback()
196 if (sge->sge_length == 0) { in ipath_ud_loopback()
[all …]
Dipath_sdma.c670 struct ipath_sge *sge; in ipath_sdma_verbs_send() local
729 sge = &ss->sge; in ipath_sdma_verbs_send()
735 if (len > sge->length) in ipath_sdma_verbs_send()
736 len = sge->length; in ipath_sdma_verbs_send()
737 if (len > sge->sge_length) in ipath_sdma_verbs_send()
738 len = sge->sge_length; in ipath_sdma_verbs_send()
741 addr = dma_map_single(&dd->pcidev->dev, sge->vaddr, dw << 2, in ipath_sdma_verbs_send()
759 sge->vaddr += len; in ipath_sdma_verbs_send()
760 sge->length -= len; in ipath_sdma_verbs_send()
761 sge->sge_length -= len; in ipath_sdma_verbs_send()
[all …]
Dipath_rc.c48 ss->sge = wqe->sg_list[0]; in restart_sge()
129 len = e->rdma_sge.sge.sge_length; in ipath_make_rc_ack()
162 len = qp->s_ack_rdma_sge.sge.sge_length; in ipath_make_rc_ack()
461 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_rc_req()
1439 if (unlikely(offset + len > e->rdma_sge.sge.sge_length)) in ipath_rc_rcv_error()
1454 e->rdma_sge.sge.mr = NULL; in ipath_rc_rcv_error()
1455 e->rdma_sge.sge.vaddr = NULL; in ipath_rc_rcv_error()
1456 e->rdma_sge.sge.length = 0; in ipath_rc_rcv_error()
1457 e->rdma_sge.sge.sge_length = 0; in ipath_rc_rcv_error()
1780 qp->r_sge.sge.mr = NULL; in ipath_rc_rcv()
[all …]
Dipath_uc.c98 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_uc_req()
454 qp->r_sge.sge.mr = NULL; in ipath_uc_rcv()
455 qp->r_sge.sge.vaddr = NULL; in ipath_uc_rcv()
456 qp->r_sge.sge.length = 0; in ipath_uc_rcv()
457 qp->r_sge.sge.sge_length = 0; in ipath_uc_rcv()
Dipath_verbs.h336 struct ipath_sge sge; /* progress state for the current SGE */ member
794 struct ib_sge *sge, int acc);
/linux-4.4.14/drivers/scsi/esas2r/
Desas2r_io.c222 if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) { in esas2r_build_sg_list_sge()
237 sgelen = (u8)((u8 *)sgc->sge.a64.curr in esas2r_build_sg_list_sge()
238 - (u8 *)sgc->sge.a64.last); in esas2r_build_sg_list_sge()
244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge()
247 sgc->sge.a64.curr = in esas2r_build_sg_list_sge()
252 sgc->sge.a64.limit = in esas2r_build_sg_list_sge()
257 sgc->sge.a64.last->length = cpu_to_le32( in esas2r_build_sg_list_sge()
259 sgc->sge.a64.last->address = in esas2r_build_sg_list_sge()
268 if (sgc->sge.a64.chain) { in esas2r_build_sg_list_sge()
269 sgc->sge.a64.chain->length |= in esas2r_build_sg_list_sge()
[all …]
Desas2r_vda.c119 firstsg = rq->vrq->flash.data.file.sge; in esas2r_process_vda_ioctl()
130 firstsg = rq->vrq->cli.sge; in esas2r_process_vda_ioctl()
194 firstsg = rq->vrq->mgt.sge; in esas2r_process_vda_ioctl()
363 data.sge); in esas2r_build_flash_req()
394 struct atto_vda_mgmt_req, sge); in esas2r_build_mgt_req()
396 vrq->sge[0].length = cpu_to_le32(SGE_LAST | length); in esas2r_build_mgt_req()
397 vrq->sge[0].address = cpu_to_le64( in esas2r_build_mgt_req()
432 (u8)offsetof(struct atto_vda_ae_req, sge); in esas2r_build_ae_req()
433 vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length); in esas2r_build_ae_req()
434 vrq->sge[0].address = cpu_to_le64( in esas2r_build_ae_req()
[all …]
Datvda.h504 struct atto_vda_sge sge[1]; member
557 struct atto_vda_sge sge[1]; member
1072 struct atto_vda_sge sge[1]; member
1095 struct atto_vda_sge sge[1]; member
1098 struct atto_vda_sge sge[1]; member
1120 struct atto_vda_sge sge[1]; member
1133 struct atto_vda_sge sge[1]; member
1147 struct atto_vda_sge sge[1]; member
1174 struct atto_vda_sge sge[1]; member
1191 struct atto_vda_sge sge; member
[all …]
Desas2r.h619 } sge; member
1180 sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq in esas2r_sgc_init()
1187 sgc->sge.a64.last = in esas2r_sgc_init()
1188 sgc->sge.a64.curr = first; in esas2r_sgc_init()
1193 sgc->sge.a64.last = in esas2r_sgc_init()
1194 sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0]; in esas2r_sgc_init()
1196 (u8)offsetof(struct atto_vda_scsi_req, u.sge); in esas2r_sgc_init()
1198 sgc->sge.a64.chain = NULL; in esas2r_sgc_init()
Desas2r_flash.c212 esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]); in build_flash_msg()
915 esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge); in esas2r_process_fs_ioctl()
1311 vrq->data.sge[0].length = in esas2r_nvram_write()
1314 vrq->data.sge[0].address = cpu_to_le64( in esas2r_nvram_write()
Desas2r_ioctl.c300 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in smp_ioctl_callback()
353 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in csmi_ioctl_tunnel()
672 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in hba_ioctl_tunnel()
Desas2r_disc.c901 esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge); in esas2r_disc_passthru_dev_addr()
Desas2r_init.c1280 data.sge); in esas2r_format_init_msg()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dkeys.c180 struct hfi1_sge *isge, struct ib_sge *sge, int acc) in hfi1_lkey_ok() argument
191 if (sge->lkey == 0) { in hfi1_lkey_ok()
203 isge->vaddr = (void *) sge->addr; in hfi1_lkey_ok()
204 isge->length = sge->length; in hfi1_lkey_ok()
205 isge->sge_length = sge->length; in hfi1_lkey_ok()
211 rkt->table[(sge->lkey >> (32 - hfi1_lkey_table_size))]); in hfi1_lkey_ok()
212 if (unlikely(!mr || mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) in hfi1_lkey_ok()
215 off = sge->addr - mr->user_base; in hfi1_lkey_ok()
216 if (unlikely(sge->addr < mr->user_base || in hfi1_lkey_ok()
217 off + sge->length > mr->length || in hfi1_lkey_ok()
[all …]
Druc.c117 if (!hfi1_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in init_sge()
130 struct hfi1_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in init_sge() local
132 hfi1_put_mr(sge->mr); in init_sge()
381 struct hfi1_sge *sge; in ruc_loopback() local
450 sqp->s_sge.sge = wqe->sg_list[0]; in ruc_loopback()
483 if (unlikely(!hfi1_rkey_ok(qp, &qp->r_sge.sge, wqe->length, in ruc_loopback()
496 if (unlikely(!hfi1_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, in ruc_loopback()
504 qp->r_sge.sge = wqe->sg_list[0]; in ruc_loopback()
514 if (unlikely(!hfi1_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), in ruc_loopback()
520 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in ruc_loopback()
[all …]
Dverbs.c281 struct hfi1_sge *sge = &ss->sge; in hfi1_copy_sge() local
284 u32 len = sge->length; in hfi1_copy_sge()
288 if (len > sge->sge_length) in hfi1_copy_sge()
289 len = sge->sge_length; in hfi1_copy_sge()
291 memcpy(sge->vaddr, data, len); in hfi1_copy_sge()
292 sge->vaddr += len; in hfi1_copy_sge()
293 sge->length -= len; in hfi1_copy_sge()
294 sge->sge_length -= len; in hfi1_copy_sge()
295 if (sge->sge_length == 0) { in hfi1_copy_sge()
297 hfi1_put_mr(sge->mr); in hfi1_copy_sge()
[all …]
Dud.c76 struct hfi1_sge *sge; in ud_loopback() local
195 ssge.sge = *swqe->sg_list; in ud_loopback()
197 sge = &ssge.sge; in ud_loopback()
199 u32 len = sge->length; in ud_loopback()
203 if (len > sge->sge_length) in ud_loopback()
204 len = sge->sge_length; in ud_loopback()
206 hfi1_copy_sge(&qp->r_sge, sge->vaddr, len, 1); in ud_loopback()
207 sge->vaddr += len; in ud_loopback()
208 sge->length -= len; in ud_loopback()
209 sge->sge_length -= len; in ud_loopback()
[all …]
Duc.c117 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_uc_req()
491 ok = hfi1_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, in hfi1_uc_rcv()
498 qp->r_sge.sge.mr = NULL; in hfi1_uc_rcv()
499 qp->r_sge.sge.vaddr = NULL; in hfi1_uc_rcv()
500 qp->r_sge.sge.length = 0; in hfi1_uc_rcv()
501 qp->r_sge.sge.sge_length = 0; in hfi1_uc_rcv()
Drc.c69 ss->sge = wqe->sg_list[0]; in restart_sge()
158 qp->s_ack_rdma_sge.sge = e->rdma_sge; in make_rc_ack()
194 qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; in make_rc_ack()
197 len = qp->s_ack_rdma_sge.sge.sge_length; in make_rc_ack()
517 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_rc_req()
1029 struct hfi1_sge *sge = &wqe->sg_list[i]; in hfi1_rc_send_complete() local
1031 hfi1_put_mr(sge->mr); in hfi1_rc_send_complete()
1086 struct hfi1_sge *sge = &wqe->sg_list[i]; in do_rc_completion() local
1088 hfi1_put_mr(sge->mr); in do_rc_completion()
2160 ok = hfi1_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, in hfi1_rc_rcv()
[all …]
Dverbs.h407 struct hfi1_sge sge; /* progress state for the current SGE */ member
971 struct hfi1_sge *isge, struct ib_sge *sge, int acc);
973 int hfi1_rkey_ok(struct hfi1_qp *qp, struct hfi1_sge *sge,
1053 hfi1_put_mr(ss->sge.mr); in hfi1_put_ss()
1055 ss->sge = *ss->sg_list++; in hfi1_put_ss()
Ddiag.c1717 void *addr = temp_ss.sge.vaddr; in snoop_send_pio_handler()
1718 u32 slen = temp_ss.sge.length; in snoop_send_pio_handler()
Dqp.c418 struct hfi1_sge *sge = &wqe->sg_list[i]; in clear_mr_refs() local
420 hfi1_put_mr(sge->mr); in clear_mr_refs()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_verbs.c171 struct qib_sge *sge = &ss->sge; in qib_copy_sge() local
174 u32 len = sge->length; in qib_copy_sge()
178 if (len > sge->sge_length) in qib_copy_sge()
179 len = sge->sge_length; in qib_copy_sge()
181 memcpy(sge->vaddr, data, len); in qib_copy_sge()
182 sge->vaddr += len; in qib_copy_sge()
183 sge->length -= len; in qib_copy_sge()
184 sge->sge_length -= len; in qib_copy_sge()
185 if (sge->sge_length == 0) { in qib_copy_sge()
187 qib_put_mr(sge->mr); in qib_copy_sge()
[all …]
Dqib_keys.c156 struct qib_sge *isge, struct ib_sge *sge, int acc) in qib_lkey_ok() argument
167 if (sge->lkey == 0) { in qib_lkey_ok()
180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok()
181 isge->length = sge->length; in qib_lkey_ok()
182 isge->sge_length = sge->length; in qib_lkey_ok()
188 rkt->table[(sge->lkey >> (32 - ib_qib_lkey_table_size))]); in qib_lkey_ok()
189 if (unlikely(!mr || mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) in qib_lkey_ok()
192 off = sge->addr - mr->user_base; in qib_lkey_ok()
193 if (unlikely(sge->addr < mr->user_base || in qib_lkey_ok()
194 off + sge->length > mr->length || in qib_lkey_ok()
[all …]
Dqib_ruc.c99 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in qib_init_sge()
112 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in qib_init_sge() local
114 qib_put_mr(sge->mr); in qib_init_sge()
361 struct qib_sge *sge; in qib_ruc_loopback() local
428 sqp->s_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback()
461 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, wqe->length, in qib_ruc_loopback()
474 if (unlikely(!qib_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, in qib_ruc_loopback()
482 qp->r_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback()
492 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), in qib_ruc_loopback()
498 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in qib_ruc_loopback()
[all …]
Dqib_ud.c57 struct qib_sge *sge; in qib_ud_loopback() local
173 ssge.sge = *swqe->sg_list; in qib_ud_loopback()
175 sge = &ssge.sge; in qib_ud_loopback()
177 u32 len = sge->length; in qib_ud_loopback()
181 if (len > sge->sge_length) in qib_ud_loopback()
182 len = sge->sge_length; in qib_ud_loopback()
184 qib_copy_sge(&qp->r_sge, sge->vaddr, len, 1); in qib_ud_loopback()
185 sge->vaddr += len; in qib_ud_loopback()
186 sge->length -= len; in qib_ud_loopback()
187 sge->sge_length -= len; in qib_ud_loopback()
[all …]
Dqib_sdma.c538 struct qib_sge *sge; in qib_sdma_verbs_send() local
586 sge = &ss->sge; in qib_sdma_verbs_send()
592 if (len > sge->length) in qib_sdma_verbs_send()
593 len = sge->length; in qib_sdma_verbs_send()
594 if (len > sge->sge_length) in qib_sdma_verbs_send()
595 len = sge->sge_length; in qib_sdma_verbs_send()
598 addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr, in qib_sdma_verbs_send()
617 sge->vaddr += len; in qib_sdma_verbs_send()
618 sge->length -= len; in qib_sdma_verbs_send()
619 sge->sge_length -= len; in qib_sdma_verbs_send()
[all …]
Dqib_uc.c99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req()
441 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, in qib_uc_rcv()
448 qp->r_sge.sge.mr = NULL; in qib_uc_rcv()
449 qp->r_sge.sge.vaddr = NULL; in qib_uc_rcv()
450 qp->r_sge.sge.length = 0; in qib_uc_rcv()
451 qp->r_sge.sge.sge_length = 0; in qib_uc_rcv()
Dqib_rc.c49 ss->sge = wqe->sg_list[0]; in restart_sge()
137 qp->s_ack_rdma_sge.sge = e->rdma_sge; in qib_make_rc_ack()
173 qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; in qib_make_rc_ack()
176 len = qp->s_ack_rdma_sge.sge.sge_length; in qib_make_rc_ack()
489 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_rc_req()
1015 struct qib_sge *sge = &wqe->sg_list[i]; in qib_rc_send_complete() local
1017 qib_put_mr(sge->mr); in qib_rc_send_complete()
1071 struct qib_sge *sge = &wqe->sg_list[i]; in do_rc_completion() local
1073 qib_put_mr(sge->mr); in do_rc_completion()
2072 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, in qib_rc_rcv()
[all …]
Dqib_verbs.h400 struct qib_sge sge; /* progress state for the current SGE */ member
994 struct qib_sge *isge, struct ib_sge *sge, int acc);
996 int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
1081 qib_put_mr(ss->sge.mr); in qib_put_ss()
1083 ss->sge = *ss->sg_list++; in qib_put_ss()
Dqib_qp.c433 struct qib_sge *sge = &wqe->sg_list[i]; in clear_mr_refs() local
435 qib_put_mr(sge->mr); in clear_mr_refs()
/linux-4.4.14/drivers/scsi/csiostor/
Dcsio_wr.c57 csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg) in csio_get_flbuf_size() argument
59 sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A + in csio_get_flbuf_size()
65 csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf) in csio_wr_fl_bufsz() argument
67 return sge->sge_fl_buf_size[buf->paddr & 0xF]; in csio_wr_fl_bufsz()
74 return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ? 128 : 64; in csio_wr_qstat_pgsz()
117 struct csio_sge *sge = &wrm->sge; in csio_wr_fill_fl() local
125 buf->len = sge->sge_fl_buf_size[sreg]; in csio_wr_fill_fl()
1044 struct csio_sge *sge = &wrm->sge; in csio_wr_process_fl() local
1070 bufsz = csio_wr_fl_bufsz(sge, buf); in csio_wr_process_fl()
1091 flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align); in csio_wr_process_fl()
[all …]
Dcsio_wr.h449 struct csio_sge sge; /* SGE params */ member
/linux-4.4.14/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c73 vec->sge[sge_no].iov_base = xdr->head[0].iov_base; in map_xdr()
74 vec->sge[sge_no].iov_len = xdr->head[0].iov_len; in map_xdr()
82 vec->sge[sge_no].iov_base = in map_xdr()
86 vec->sge[sge_no].iov_len = sge_bytes; in map_xdr()
95 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr()
96 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr()
221 struct ib_sge *sge; in send_write() local
237 write_len, vec->sge, vec->count); in send_write()
241 sge = ctxt->sge; in send_write()
246 if (vec->sge[xdr_sge_no].iov_len > bc) in send_write()
[all …]
Dsvc_rdma_recvfrom.c75 min_t(size_t, byte_count, ctxt->sge[0].length); in rdma_build_arg_xdr()
98 bc -= min_t(u32, bc, ctxt->sge[sge_no].length); in rdma_build_arg_xdr()
99 rqstp->rq_arg.buflen += ctxt->sge[sge_no].length; in rdma_build_arg_xdr()
152 ctxt->sge[pno].addr = in rdma_read_chunk_lcl()
158 ctxt->sge[pno].addr); in rdma_read_chunk_lcl()
164 ctxt->sge[pno].lkey = xprt->sc_dma_lkey; in rdma_read_chunk_lcl()
165 ctxt->sge[pno].length = len; in rdma_read_chunk_lcl()
189 read_wr.wr.sg_list = ctxt->sge; in rdma_read_chunk_lcl()
295 ctxt->sge[0].addr = frmr->mr->iova; in rdma_read_chunk_frmr()
296 ctxt->sge[0].lkey = frmr->mr->lkey; in rdma_read_chunk_frmr()
[all …]
Dsvc_rdma_transport.c174 for (i = 0; i < ctxt->count && ctxt->sge[i].length; i++) { in svc_rdma_unmap_dma()
181 if (ctxt->sge[i].lkey == xprt->sc_dma_lkey) { in svc_rdma_unmap_dma()
184 ctxt->sge[i].addr, in svc_rdma_unmap_dma()
185 ctxt->sge[i].length, in svc_rdma_unmap_dma()
572 ctxt->sge[sge_no].addr = pa; in svc_rdma_post_recv()
573 ctxt->sge[sge_no].length = PAGE_SIZE; in svc_rdma_post_recv()
574 ctxt->sge[sge_no].lkey = xprt->sc_dma_lkey; in svc_rdma_post_recv()
579 recv_wr.sg_list = &ctxt->sge[0]; in svc_rdma_post_recv()
1336 ctxt->sge[0].addr = ib_dma_map_page(xprt->sc_cm_id->device, in svc_rdma_send_error()
1338 if (ib_dma_mapping_error(xprt->sc_cm_id->device, ctxt->sge[0].addr)) { in svc_rdma_send_error()
[all …]
/linux-4.4.14/drivers/infiniband/ulp/iser/
Diser_memory.c253 reg->sge.lkey = device->pd->local_dma_lkey; in iser_reg_dma()
255 reg->sge.addr = ib_sg_dma_address(device->ib_device, &sg[0]); in iser_reg_dma()
256 reg->sge.length = ib_sg_dma_len(device->ib_device, &sg[0]); in iser_reg_dma()
259 " length=0x%x\n", reg->sge.lkey, reg->rkey, in iser_reg_dma()
260 reg->sge.addr, reg->sge.length); in iser_reg_dma()
305 reg->sge.lkey = fmr->fmr->lkey; in iser_fast_reg_fmr()
307 reg->sge.addr = page_vec->pages[0] + page_vec->offset; in iser_fast_reg_fmr()
308 reg->sge.length = page_vec->data_size; in iser_fast_reg_fmr()
312 " length=0x%x\n", reg->sge.lkey, reg->rkey, in iser_fast_reg_fmr()
313 reg->sge.addr, reg->sge.length); in iser_fast_reg_fmr()
[all …]
Diser_initiator.c84 hdr->read_va = cpu_to_be64(mem_reg->sge.addr); in iser_prepare_read_cmd()
88 (unsigned long long)mem_reg->sge.addr); in iser_prepare_read_cmd()
140 hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz); in iser_prepare_write_cmd()
145 (unsigned long long)mem_reg->sge.addr, unsol_sz); in iser_prepare_write_cmd()
151 tx_dsg->addr = mem_reg->sge.addr; in iser_prepare_write_cmd()
153 tx_dsg->lkey = mem_reg->sge.lkey; in iser_prepare_write_cmd()
477 tx_dsg->addr = mem_reg->sge.addr + buf_offset; in iser_send_data_out()
479 tx_dsg->lkey = mem_reg->sge.lkey; in iser_send_data_out()
Diser_verbs.c1050 struct ib_sge sge; in iser_post_recvl() local
1053 sge.addr = iser_conn->login_resp_dma; in iser_post_recvl()
1054 sge.length = ISER_RX_LOGIN_SIZE; in iser_post_recvl()
1055 sge.lkey = ib_conn->device->pd->local_dma_lkey; in iser_post_recvl()
1058 rx_wr.sg_list = &sge; in iser_post_recvl()
Discsi_iser.h252 struct ib_sge sge; member
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c160 struct sge *s = &adapter->sge; in fl_mtu_bufsize()
247 const struct sge *s = &adapter->sge; in fl_starving()
330 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl()
445 struct sge *s = &adapter->sge; in get_buf_size()
582 struct sge *s = &adap->sge; in refill_fl()
591 if (test_bit(q->cntxt_id - adap->sge.egr_start, adap->sge.blocked_fl)) in refill_fl()
666 set_bit(q->cntxt_id - adap->sge.egr_start, in refill_fl()
667 adap->sge.starving_fl); in refill_fl()
879 to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; in write_sgl()
893 unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1; in write_sgl()
[all …]
Dcxgb4_main.c312 struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; in dcb_tx_queue_prio_enable()
626 txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; in fwevtq_handler()
628 if ((u8 *)txq < (u8 *)q->adap->sge.ofldtxq) { in fwevtq_handler()
769 for_each_ofldrxq(&adap->sge, i) in name_msix_vecs()
773 for_each_rdmarxq(&adap->sge, i) in name_msix_vecs()
777 for_each_rdmaciq(&adap->sge, i) in name_msix_vecs()
784 struct sge *s = &adap->sge; in request_msix_queue_irqs()
851 struct sge *s = &adap->sge; in free_msix_queue_irqs()
880 rxq = &adapter->sge.ethrxq[pi->first_qset]; in cxgb4_write_rss()
935 static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) in rxq_to_chan()
[all …]
Dcxgb4.h311 struct sge_params sge; member
633 struct sge { struct
678 #define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++) argument
679 #define for_each_ofldrxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++) argument
680 #define for_each_rdmarxq(sge, i) for (i = 0; i < (sge)->rdmaqs; i++) argument
681 #define for_each_rdmaciq(sge, i) for (i = 0; i < (sge)->rdmaciqs; i++) argument
724 struct sge sge; member
1084 return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; in qtimer_val()
Dcxgb4_debugfs.c2247 int eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4); in sge_qinfo_show()
2248 int iscsi_entries = DIV_ROUND_UP(adap->sge.ofldqsets, 4); in sge_qinfo_show()
2249 int rdma_entries = DIV_ROUND_UP(adap->sge.rdmaqs, 4); in sge_qinfo_show()
2250 int ciq_entries = DIV_ROUND_UP(adap->sge.rdmaciqs, 4); in sge_qinfo_show()
2279 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[base_qset]; in sge_qinfo_show()
2280 const struct sge_eth_txq *tx = &adap->sge.ethtxq[base_qset]; in sge_qinfo_show()
2281 int n = min(4, adap->sge.ethqsets - 4 * r); in sge_qinfo_show()
2307 adap->sge.counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show()
2332 &adap->sge.ofldrxq[iscsi_idx * 4]; in sge_qinfo_show()
2334 &adap->sge.ofldtxq[iscsi_idx * 4]; in sge_qinfo_show()
[all …]
Dcxgb4_ethtool.c312 const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; in collect_sge_port_stats()
313 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset]; in collect_sge_port_stats()
697 const struct sge *s = &pi->adapter->sge; in get_sge_param()
715 struct sge *s = &adapter->sge; in set_sge_param()
749 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_rx_intr_params()
764 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_adaptive_rx_setting()
776 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in get_adaptive_rx_setting()
792 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
796 adap->sge.counter_val[rq->pktcnt_idx] : 0; in get_coalesce()
DMakefile7 cxgb4-objs := cxgb4_main.o l2t.o t4_hw.o sge.o clip_tbl.o cxgb4_ethtool.o
Dl2t.c156 TID_QID_V(adap->sge.fw_evtq.abs_id))); in write_l2e()
Dt4_msg.h1141 struct ulptx_sge_pair sge[0]; member
Dt4_hw.c7147 page_shift = adapter->params.sge.hps + 10; in t4_bar2_sge_qregs()
7153 ? adapter->params.sge.eq_qpp in t4_bar2_sge_qregs()
7154 : adapter->params.sge.iq_qpp); in t4_bar2_sge_qregs()
7256 struct sge_params *sge_params = &adapter->params.sge; in t4_init_sge_params()
/linux-4.4.14/net/rds/
Diw_recv.c88 struct ib_sge *sge; in rds_iw_recv_init_ring() local
98 sge = rds_iw_data_sge(ic, recv->r_sge); in rds_iw_recv_init_ring()
99 sge->addr = 0; in rds_iw_recv_init_ring()
100 sge->length = RDS_FRAG_SIZE; in rds_iw_recv_init_ring()
101 sge->lkey = 0; in rds_iw_recv_init_ring()
103 sge = rds_iw_header_sge(ic, recv->r_sge); in rds_iw_recv_init_ring()
104 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_recv_init_ring()
105 sge->length = sizeof(struct rds_header); in rds_iw_recv_init_ring()
106 sge->lkey = 0; in rds_iw_recv_init_ring()
143 struct ib_sge *sge; in rds_iw_recv_refill_one() local
[all …]
Dib_recv.c52 struct ib_sge *sge; in rds_ib_recv_init_ring() local
62 sge = &recv->r_sge[0]; in rds_ib_recv_init_ring()
63 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_recv_init_ring()
64 sge->length = sizeof(struct rds_header); in rds_ib_recv_init_ring()
65 sge->lkey = ic->i_pd->local_dma_lkey; in rds_ib_recv_init_ring()
67 sge = &recv->r_sge[1]; in rds_ib_recv_init_ring()
68 sge->addr = 0; in rds_ib_recv_init_ring()
69 sge->length = RDS_FRAG_SIZE; in rds_ib_recv_init_ring()
70 sge->lkey = ic->i_pd->local_dma_lkey; in rds_ib_recv_init_ring()
303 struct ib_sge *sge; in rds_ib_recv_refill_one() local
[all …]
Diw_send.c134 struct ib_sge *sge; in rds_iw_send_init_ring() local
148 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_send_init_ring()
149 sge->lkey = 0; in rds_iw_send_init_ring()
151 sge = rds_iw_header_sge(ic, send->s_sge); in rds_iw_send_init_ring()
152 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_send_init_ring()
153 sge->length = sizeof(struct rds_header); in rds_iw_send_init_ring()
154 sge->lkey = 0; in rds_iw_send_init_ring()
445 struct ib_sge *sge; in rds_iw_xmit_populate_wr() local
457 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_xmit_populate_wr()
458 sge->addr = buffer; in rds_iw_xmit_populate_wr()
[all …]
Diw.h387 rds_iw_header_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_header_sge() argument
389 return &sge[0]; in rds_iw_header_sge()
393 rds_iw_data_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_data_sge() argument
395 return &sge[1]; in rds_iw_data_sge()
Dib_send.c194 struct ib_sge *sge; in rds_ib_send_init_ring() local
202 sge = &send->s_sge[0]; in rds_ib_send_init_ring()
203 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_send_init_ring()
204 sge->length = sizeof(struct rds_header); in rds_ib_send_init_ring()
205 sge->lkey = ic->i_pd->local_dma_lkey; in rds_ib_send_init_ring()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c323 struct sge *s = &adapter->sge; in request_msix_queue_irqs()
361 struct sge *s = &adapter->sge; in free_msix_queue_irqs()
394 struct sge *s = &adapter->sge; in enable_rx()
417 struct sge *s = &adapter->sge; in quiesce_rx()
475 struct sge *s = &adapter->sge; in fwevtq_handler()
531 struct sge *s = &adapter->sge; in setup_sge_queues()
646 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; in setup_rss()
986 static int closest_timer(const struct sge *s, int us) in closest_timer()
1002 static int closest_thres(const struct sge *s, int thres) in closest_thres()
1027 ? adapter->sge.timer_val[timer_idx] in qtimer_val()
[all …]
Dsge.c264 const struct sge *s = &adapter->sge; in fl_starving()
325 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl()
457 const struct sge *s = &adapter->sge; in get_buf_size()
607 struct sge *s = &adapter->sge; in refill_fl()
715 set_bit(fl->cntxt_id, adapter->sge.starving_fl); in refill_fl()
935 to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; in write_sgl()
949 unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; in write_sgl()
952 memcpy(sgl->sge, buf, part0); in write_sgl()
1203 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit()
1572 struct sge *s = &adapter->sge; in do_gro()
[all …]
Dadapter.h269 struct sge { struct
348 #define for_each_ethrxq(sge, iter) \ argument
349 for (iter = 0; iter < (sge)->ethqsets; iter++)
372 struct sge sge; member
DMakefile7 cxgb4vf-objs := cxgb4vf_main.o t4vf_hw.o sge.o
Dt4vf_common.h221 struct sge_params sge; /* Scatter Gather Engine */ member
Dt4vf_hw.c472 page_shift = adapter->params.sge.sge_vf_hps + 10; in t4vf_bar2_sge_qregs()
478 ? adapter->params.sge.sge_vf_eq_qpp in t4vf_bar2_sge_qregs()
479 : adapter->params.sge.sge_vf_iq_qpp); in t4vf_bar2_sge_qregs()
529 struct sge_params *sge_params = &adapter->params.sge; in t4vf_get_sge_params()
/linux-4.4.14/net/9p/
Dtrans_rdma.c403 struct ib_sge sge; in post_recv() local
411 sge.addr = c->busa; in post_recv()
412 sge.length = client->msize; in post_recv()
413 sge.lkey = rdma->pd->local_dma_lkey; in post_recv()
418 wr.sg_list = &sge; in post_recv()
431 struct ib_sge sge; in rdma_request() local
502 sge.addr = c->busa; in rdma_request()
503 sge.length = c->req->tc->size; in rdma_request()
504 sge.lkey = rdma->pd->local_dma_lkey; in rdma_request()
511 wr.sg_list = &sge; in rdma_request()
/linux-4.4.14/drivers/scsi/be2iscsi/
Dbe_mgmt.c282 struct be_sge *sge; in mgmt_get_session_info() local
299 sge = nonembedded_sgl(wrb); in mgmt_get_session_info()
309 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); in mgmt_get_session_info()
310 sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); in mgmt_get_session_info()
311 sge->len = cpu_to_le32(nonemb_cmd->size); in mgmt_get_session_info()
428 struct be_sge *sge = nonembedded_sgl(wrb); in mgmt_check_supported_fw() local
448 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); in mgmt_check_supported_fw()
449 sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF); in mgmt_check_supported_fw()
450 sge->len = cpu_to_le32(nonemb_cmd.size); in mgmt_check_supported_fw()
591 struct be_sge *sge; in mgmt_invalidate_icds() local
[all …]
/linux-4.4.14/fs/exofs/
Dore_raid.c273 struct osd_sg_entry *sge; in _ore_add_sg_seg() local
282 sge = per_dev->sglist; in _ore_add_sg_seg()
287 sge->offset = per_dev->offset; in _ore_add_sg_seg()
288 sge->len = per_dev->length; in _ore_add_sg_seg()
300 sge = &per_dev->sglist[per_dev->cur_sg - 1]; in _ore_add_sg_seg()
301 sge->len = per_dev->length - per_dev->last_sgs_total; in _ore_add_sg_seg()
306 struct osd_sg_entry *next_sge = sge + 1; in _ore_add_sg_seg()
309 next_sge->offset = sge->offset + sge->len + cur_len; in _ore_add_sg_seg()
313 } else if (!sge->len) { in _ore_add_sg_seg()
/linux-4.4.14/drivers/ata/
Dsata_sil24.c257 struct sil24_sge sge[SIL24_MAX_SGE]; member
263 struct sil24_sge sge[SIL24_MAX_SGE]; member
780 struct sil24_sge *sge) in sil24_fill_sg() argument
787 sge->addr = cpu_to_le64(sg_dma_address(sg)); in sil24_fill_sg()
788 sge->cnt = cpu_to_le32(sg_dma_len(sg)); in sil24_fill_sg()
789 sge->flags = 0; in sil24_fill_sg()
791 last_sge = sge; in sil24_fill_sg()
792 sge++; in sil24_fill_sg()
849 struct sil24_sge *sge; in sil24_qc_prep() local
856 sge = cb->ata.sge; in sil24_qc_prep()
[all …]
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c679 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
681 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
692 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
694 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
708 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset()
710 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset()
1516 ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb); in t3_mgmt_tx()
1770 struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; in t3_offload_tx()
2029 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in rx_eth()
2131 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in lro_add_page()
[all …]
Dadapter.h216 struct sge { struct
245 struct sge sge; member
Dcxgb3_main.c411 adap->sge.qs[qidx]. in request_msix_data_irqs()
414 &adap->sge.qs[qidx]); in request_msix_data_irqs()
418 &adap->sge.qs[qidx]); in request_msix_data_irqs()
438 &adapter->sge.qs[i]); in free_irq_resources()
448 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
461 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
600 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs()
613 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi()
638 if (adap->sge.qs[i].adap) in quiesce_rx()
639 napi_disable(&adap->sge.qs[i].napi); in quiesce_rx()
[all …]
DMakefile8 xgmac.o sge.o l2t.o cxgb3_offload.o aq100x.o
Dcxgb3_offload.c303 spin_lock_irqsave(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl()
306 spin_unlock_irqrestore(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl()
334 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
341 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
345 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
347 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
352 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
358 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
Dcommon.h382 struct sge_params sge; member
Dt3_hw.c3369 min(adapter->params.sge.max_pkt_size, in t3_init_hw()
3372 min(adapter->params.sge.max_pkt_size, 16384U)); in t3_init_hw()
3389 t3_sge_init(adapter, &adapter->params.sge); in t3_init_hw()
3653 t3_sge_prep(adapter, &adapter->params.sge); in t3_prep_adapter()
/linux-4.4.14/include/linux/sunrpc/
Dsvc_rdma.h85 struct ib_sge sge[RPCSVC_MAXPAGES]; member
117 struct kvec sge[RPCSVC_MAXPAGES]; member
/linux-4.4.14/drivers/scsi/aacraid/
Daachba.c3367 rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32)); in aac_build_sgraw2()
3368 rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff)); in aac_build_sgraw2()
3370 rio2->sge[i].length = cur_size; in aac_build_sgraw2()
3371 rio2->sge[i].flags = 0; in aac_build_sgraw2()
3388 u32 temp = le32_to_cpu(rio2->sge[i-1].length) - in aac_build_sgraw2()
3390 rio2->sge[i-1].length = cpu_to_le32(temp); in aac_build_sgraw2()
3403 if (rio2->sge[j].length % (i*PAGE_SIZE)) { in aac_build_sgraw2()
3407 nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE)); in aac_build_sgraw2()
3429 struct sge_ieee1212 *sge; in aac_convert_sgraw2() local
3436 sge = kmalloc(nseg_new * sizeof(struct sge_ieee1212), GFP_ATOMIC); in aac_convert_sgraw2()
[all …]
Daacraid.h1489 struct sge_ieee1212 sge[1]; member
/linux-4.4.14/drivers/scsi/lpfc/
Dlpfc_mbox.c1683 nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr); in lpfc_sli4_mbx_sge_set()
1684 nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr); in lpfc_sli4_mbx_sge_set()
1685 nembed_sge->sge[sgentry].length = length; in lpfc_sli4_mbx_sge_set()
1698 struct lpfc_mbx_sge *sge) in lpfc_sli4_mbx_sge_get() argument
1704 sge->pa_lo = nembed_sge->sge[sgentry].pa_lo; in lpfc_sli4_mbx_sge_get()
1705 sge->pa_hi = nembed_sge->sge[sgentry].pa_hi; in lpfc_sli4_mbx_sge_get()
1706 sge->length = nembed_sge->sge[sgentry].length; in lpfc_sli4_mbx_sge_get()
1720 struct lpfc_mbx_sge sge; in lpfc_sli4_mbox_cmd_free() local
1741 lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge); in lpfc_sli4_mbox_cmd_free()
1742 phyaddr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_mbox_cmd_free()
[all …]
Dlpfc_bsg.c4795 receive_length = nembed_sge->sge[0].length; in lpfc_bsg_issue_mbox()
4807 nembed_sge->sge[0].pa_hi = in lpfc_bsg_issue_mbox()
4810 nembed_sge->sge[0].pa_lo = in lpfc_bsg_issue_mbox()
Dlpfc_hbadisc.c1776 struct lpfc_mbx_sge sge; in lpfc_sli4_fcf_rec_mbox_parse() local
1785 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_fcf_rec_mbox_parse()
Dlpfc_hw4.h1026 struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES]; member
Dlpfc_sli.c15917 struct lpfc_mbx_sge sge; in lpfc_sli4_add_fcf_record() local
15948 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_add_fcf_record()
/linux-4.4.14/drivers/infiniband/ulp/isert/
Dib_isert.c1113 struct ib_sge sge; in isert_rdma_post_recvl() local
1116 memset(&sge, 0, sizeof(struct ib_sge)); in isert_rdma_post_recvl()
1117 sge.addr = isert_conn->login_req_dma; in isert_rdma_post_recvl()
1118 sge.length = ISER_RX_LOGIN_SIZE; in isert_rdma_post_recvl()
1119 sge.lkey = isert_conn->device->pd->local_dma_lkey; in isert_rdma_post_recvl()
1122 sge.addr, sge.length, sge.lkey); in isert_rdma_post_recvl()
1126 rx_wr.sg_list = &sge; in isert_rdma_post_recvl()
2520 struct ib_sge *sge) in isert_fast_reg_mr() argument
2530 sge->lkey = device->pd->local_dma_lkey; in isert_fast_reg_mr()
2531 sge->addr = ib_sg_dma_address(ib_dev, &mem->sg[0]); in isert_fast_reg_mr()
[all …]
/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_reqs.c119 struct ib_sge *sge = ud_wr->wr.sg_list; in trace_ud_wr() local
138 u8 *data = __va(sge->addr); in trace_ud_wr()
141 idx, j, data, sge->length, sge->lkey); in trace_ud_wr()
143 ehca_dmp(data, sge->length, "ud_wr#%x sge#%x", in trace_ud_wr()
145 sge++; in trace_ud_wr()
/linux-4.4.14/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c2014 struct ocrdma_sge *sge, int num_sge, in ocrdma_build_sges() argument
2020 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges()
2021 sge[i].addr_lo = sg_list[i].addr; in ocrdma_build_sges()
2022 sge[i].addr_hi = upper_32_bits(sg_list[i].addr); in ocrdma_build_sges()
2023 sge[i].len = sg_list[i].length; in ocrdma_build_sges()
2027 memset(sge, 0, sizeof(*sge)); in ocrdma_build_sges()
2042 struct ocrdma_sge *sge, in ocrdma_build_inline_sges() argument
2056 dpp_addr = (char *)sge; in ocrdma_build_inline_sges()
2069 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_inline_sges()
2084 struct ocrdma_sge *sge; in ocrdma_build_send() local
[all …]
Docrdma_hw.c1296 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); in ocrdma_mbx_rdma_stats()
1297 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); in ocrdma_mbx_rdma_stats()
1298 mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size; in ocrdma_mbx_rdma_stats()
1344 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); in ocrdma_mbx_get_ctrl_attribs()
1345 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); in ocrdma_mbx_get_ctrl_attribs()
1346 mqe->u.nonemb_req.sge[0].len = dma.size; in ocrdma_mbx_get_ctrl_attribs()
2881 struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge; in ocrdma_mbx_get_dcbx_config()
Docrdma_sli.h284 struct ocrdma_mqe_sge sge[19]; member
/linux-4.4.14/drivers/infiniband/ulp/srpt/
Dib_srpt.c1058 kfree(ioctx->rdma_ius[--ioctx->n_rdma].sge); in srpt_unmap_sg_to_ib_sge()
1088 struct ib_sge *sge; in srpt_map_sg_to_ib_sge() local
1172 riu->sge = in srpt_map_sg_to_ib_sge()
1173 kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge()
1175 if (!riu->sge) in srpt_map_sg_to_ib_sge()
1186 riu->sge = kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge()
1188 if (!riu->sge) in srpt_map_sg_to_ib_sge()
1203 sge = riu->sge; in srpt_map_sg_to_ib_sge()
1207 sge->addr = dma_addr; in srpt_map_sg_to_ib_sge()
1208 sge->lkey = ch->sport->sdev->pd->local_dma_lkey; in srpt_map_sg_to_ib_sge()
[all …]
Dib_srpt.h156 struct ib_sge *sge; member
/linux-4.4.14/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c117 struct ib_sge *sge, int id) in ipoib_cm_post_receive_nonsrq() argument
126 sge[i].addr = rx->rx_ring[id].mapping[i]; in ipoib_cm_post_receive_nonsrq()
329 struct ib_sge *sge) in ipoib_cm_init_rx_wr() argument
335 sge[i].lkey = priv->pd->local_dma_lkey; in ipoib_cm_init_rx_wr()
337 sge[0].length = IPOIB_CM_HEAD_SIZE; in ipoib_cm_init_rx_wr()
339 sge[i].length = PAGE_SIZE; in ipoib_cm_init_rx_wr()
342 wr->sg_list = sge; in ipoib_cm_init_rx_wr()
352 struct ib_sge sge[IPOIB_CM_RX_SG]; in ipoib_cm_nonsrq_init_rx() member
370 ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); in ipoib_cm_nonsrq_init_rx()
392 ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); in ipoib_cm_nonsrq_init_rx()
/linux-4.4.14/drivers/scsi/
Dvmw_pvscsi.c52 struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; member
336 struct PVSCSISGElement *sge; in pvscsi_create_sg() local
340 sge = &ctx->sgl->sge[0]; in pvscsi_create_sg()
342 sge[i].addr = sg_dma_address(sg); in pvscsi_create_sg()
343 sge[i].length = sg_dma_len(sg); in pvscsi_create_sg()
344 sge[i].flags = 0; in pvscsi_create_sg()
/linux-4.4.14/drivers/infiniband/core/
Dmad.c1148 struct ib_sge *sge; in ib_send_mad() local
1158 sge = mad_send_wr->sg_list; in ib_send_mad()
1159 sge[0].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad()
1161 sge[0].length, in ib_send_mad()
1163 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr))) in ib_send_mad()
1166 mad_send_wr->header_mapping = sge[0].addr; in ib_send_mad()
1168 sge[1].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad()
1170 sge[1].length, in ib_send_mad()
1172 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) { in ib_send_mad()
1175 sge[0].length, DMA_TO_DEVICE); in ib_send_mad()
[all …]
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c1014 struct ib_sge *sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_tx_msg() local
1028 sge->lkey = mr->lkey; in kiblnd_init_tx_msg()
1029 sge->addr = tx->tx_msgaddr; in kiblnd_init_tx_msg()
1030 sge->length = nob; in kiblnd_init_tx_msg()
1036 wrq->wr.sg_list = sge; in kiblnd_init_tx_msg()
1050 struct ib_sge *sge = &tx->tx_sge[0]; in kiblnd_init_rdma() local
1091 sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_rdma()
1092 sge->addr = kiblnd_rd_frag_addr(srcrd, srcidx); in kiblnd_init_rdma()
1093 sge->lkey = kiblnd_rd_frag_key(srcrd, srcidx); in kiblnd_init_rdma()
1094 sge->length = wrknob; in kiblnd_init_rdma()
[all …]
/linux-4.4.14/drivers/scsi/bfa/
Dbfa_fcpim.c2461 struct bfi_sge_s *sge, *sgpge; in bfa_ioim_send_ioreq() local
2489 sge = &m->sges[0]; in bfa_ioim_send_ioreq()
2498 sge->sga = *(union bfi_addr_u *) &addr; in bfa_ioim_send_ioreq()
2500 sge->sg_len = pgdlen; in bfa_ioim_send_ioreq()
2501 sge->flags = (ioim->nsges > BFI_SGE_INLINE) ? in bfa_ioim_send_ioreq()
2503 bfa_sge_to_be(sge); in bfa_ioim_send_ioreq()
2504 sge++; in bfa_ioim_send_ioreq()
2545 sge->sga = ioim->sgpg->sgpg_pa; in bfa_ioim_send_ioreq()
2547 sge->sga.a32.addr_lo = 0; in bfa_ioim_send_ioreq()
2548 sge->sga.a32.addr_hi = 0; in bfa_ioim_send_ioreq()
[all …]
Dbfad_bsg.c3197 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sgaddr_cb() local
3200 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sgaddr_cb()
3201 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_req_sgaddr_cb()
3209 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sglen_cb() local
3211 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sglen_cb()
3212 return sge->sg_len; in bfad_fcxp_get_req_sglen_cb()
3219 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sgaddr_cb() local
3222 sge = drv_fcxp->rsp_sge + sgeid; in bfad_fcxp_get_rsp_sgaddr_cb()
3223 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_rsp_sgaddr_cb()
3231 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sglen_cb() local
[all …]
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_cmn.h811 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_free_rx_sge() local
826 sge->addr_hi = 0; in bnx2x_free_rx_sge()
827 sge->addr_lo = 0; in bnx2x_free_rx_sge()
Dbnx2x_cmn.c550 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_alloc_rx_sge() local
584 sge->addr_hi = cpu_to_le32(U64_HI(mapping)); in bnx2x_alloc_rx_sge()
585 sge->addr_lo = cpu_to_le32(U64_LO(mapping)); in bnx2x_alloc_rx_sge()
1351 struct eth_rx_sge *sge; in bnx2x_set_next_page_sgl() local
1353 sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2]; in bnx2x_set_next_page_sgl()
1354 sge->addr_hi = in bnx2x_set_next_page_sgl()
1358 sge->addr_lo = in bnx2x_set_next_page_sgl()
/linux-4.4.14/drivers/message/fusion/
Dmptbase.h334 SGESimple32_t sge[1]; member
338 SGESimple64_t sge[1]; member
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
Dt4fw_ri_api.h124 struct fw_ri_dsge_pair sge[0]; member
140 struct fw_ri_sge sge[0]; member
Dqp.c434 __be64 *flitp = (__be64 *)isglp->sge; in build_isgl()
/linux-4.4.14/drivers/net/ethernet/emulex/benet/
Dbe_cmds.c737 struct be_sge *sge; in be_wrb_cmd_hdr_prepare() local
748 sge = nonembedded_sgl(wrb); in be_wrb_cmd_hdr_prepare()
749 sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma)); in be_wrb_cmd_hdr_prepare()
750 sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF); in be_wrb_cmd_hdr_prepare()
751 sge->len = cpu_to_le32(mem->size); in be_wrb_cmd_hdr_prepare()
/linux-4.4.14/Documentation/DocBook/
Ddevice-drivers.xml.db537 API-mpt-add-sge
538 API-mpt-add-sge-64bit
539 API-mpt-add-sge-64bit-1078
/linux-4.4.14/drivers/net/ethernet/broadcom/
Dcnic_defs.h3097 struct fcoe_bd_ctx sge[3]; member