/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 254 struct sge { struct 282 static void tx_sched_stop(struct sge *sge) in tx_sched_stop() argument 284 struct sched *s = sge->tx_sched; in tx_sched_stop() 297 unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port, in t1_sched_update_parms() argument 300 struct sched *s = sge->tx_sched; in t1_sched_update_parms() 320 if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) { in t1_sched_update_parms() 343 void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val) 345 struct sched *s = sge->tx_sched; 350 t1_sched_update_parms(sge, i, 0, 0); 357 void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port, [all …]
|
D | sge.h | 71 struct sge; 73 struct sge *t1_sge_create(struct adapter *, struct sge_params *); 74 int t1_sge_configure(struct sge *, struct sge_params *); 75 int t1_sge_set_coalesce_params(struct sge *, struct sge_params *); 76 void t1_sge_destroy(struct sge *); 82 void t1_sge_start(struct sge *); 83 void t1_sge_stop(struct sge *); 84 int t1_sge_intr_error_handler(struct sge *); 85 void t1_sge_intr_enable(struct sge *); 86 void t1_sge_intr_disable(struct sge *); [all …]
|
D | cxgb2.c | 174 t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed); in t1_link_negotiated() 228 t1_sge_start(adapter->sge); in cxgb_up() 239 t1_sge_stop(adapter->sge); in cxgb_down() 467 t = t1_sge_get_intr_counts(adapter->sge); in get_stats() 468 t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss); in get_stats() 716 e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl]; in get_sge_param() 717 e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl]; in get_sge_param() 718 e->tx_pending = adapter->params.sge.cmdQ_size[0]; in get_sge_param() 737 adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending; in set_sge_param() 738 adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending; in set_sge_param() [all …]
|
D | common.h | 185 struct sge_params sge; member 224 struct sge; 242 struct sge *sge; member
|
D | subr.c | 219 t1_sge_intr_error_handler(adapter->sge); in fpga_slow_intr() 761 t1_sge_intr_enable(adapter->sge); in t1_interrupts_enable() 793 t1_sge_intr_disable(adapter->sge); in t1_interrupts_disable() 819 t1_sge_intr_clear(adapter->sge); in t1_interrupts_clear() 853 t1_sge_intr_error_handler(adapter->sge); in asic_slow_intr() 982 err = t1_sge_configure(adapter->sge, &adapter->params.sge); in t1_init_hw_modules() 1022 if (adapter->sge) in t1_free_sw_modules() 1023 t1_sge_destroy(adapter->sge); in t1_free_sw_modules() 1059 adapter->sge = t1_sge_create(adapter, &adapter->params.sge); in t1_init_sw_modules() 1060 if (!adapter->sge) { in t1_init_sw_modules()
|
D | Makefile | 8 cxgb-objs := cxgb2.o espi.o tp.o pm3393.o sge.o subr.o \
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_keys.c | 122 struct ib_sge *sge, int acc) in ipath_lkey_ok() argument 134 if (sge->lkey == 0) { in ipath_lkey_ok() 143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok() 144 isge->length = sge->length; in ipath_lkey_ok() 145 isge->sge_length = sge->length; in ipath_lkey_ok() 149 mr = rkt->table[(sge->lkey >> (32 - ib_ipath_lkey_table_size))]; in ipath_lkey_ok() 150 if (unlikely(mr == NULL || mr->lkey != sge->lkey || in ipath_lkey_ok() 156 off = sge->addr - mr->user_base; in ipath_lkey_ok() 157 if (unlikely(sge->addr < mr->user_base || in ipath_lkey_ok() 158 off + sge->length > mr->length || in ipath_lkey_ok() [all …]
|
D | ipath_verbs.c | 173 struct ipath_sge *sge = &ss->sge; in ipath_copy_sge() local 176 u32 len = sge->length; in ipath_copy_sge() 180 if (len > sge->sge_length) in ipath_copy_sge() 181 len = sge->sge_length; in ipath_copy_sge() 183 memcpy(sge->vaddr, data, len); in ipath_copy_sge() 184 sge->vaddr += len; in ipath_copy_sge() 185 sge->length -= len; in ipath_copy_sge() 186 sge->sge_length -= len; in ipath_copy_sge() 187 if (sge->sge_length == 0) { in ipath_copy_sge() 189 *sge = *ss->sg_list++; in ipath_copy_sge() [all …]
|
D | ipath_ruc.c | 134 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, in ipath_init_sge() 265 struct ipath_sge *sge; in ipath_ruc_loopback() local 328 sqp->s_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback() 370 qp->r_sge.sge = wqe->sg_list[0]; in ipath_ruc_loopback() 385 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in ipath_ruc_loopback() 387 *(u64 *) sqp->s_sge.sge.vaddr = in ipath_ruc_loopback() 390 (u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr, in ipath_ruc_loopback() 399 sge = &sqp->s_sge.sge; in ipath_ruc_loopback() 403 if (len > sge->length) in ipath_ruc_loopback() 404 len = sge->length; in ipath_ruc_loopback() [all …]
|
D | ipath_ud.c | 59 struct ipath_sge *sge; in ipath_ud_loopback() local 184 sge = swqe->sg_list; in ipath_ud_loopback() 186 u32 len = sge->length; in ipath_ud_loopback() 190 if (len > sge->sge_length) in ipath_ud_loopback() 191 len = sge->sge_length; in ipath_ud_loopback() 193 ipath_copy_sge(&rsge, sge->vaddr, len); in ipath_ud_loopback() 194 sge->vaddr += len; in ipath_ud_loopback() 195 sge->length -= len; in ipath_ud_loopback() 196 sge->sge_length -= len; in ipath_ud_loopback() 197 if (sge->sge_length == 0) { in ipath_ud_loopback() [all …]
|
D | ipath_sdma.c | 670 struct ipath_sge *sge; in ipath_sdma_verbs_send() local 729 sge = &ss->sge; in ipath_sdma_verbs_send() 735 if (len > sge->length) in ipath_sdma_verbs_send() 736 len = sge->length; in ipath_sdma_verbs_send() 737 if (len > sge->sge_length) in ipath_sdma_verbs_send() 738 len = sge->sge_length; in ipath_sdma_verbs_send() 741 addr = dma_map_single(&dd->pcidev->dev, sge->vaddr, dw << 2, in ipath_sdma_verbs_send() 759 sge->vaddr += len; in ipath_sdma_verbs_send() 760 sge->length -= len; in ipath_sdma_verbs_send() 761 sge->sge_length -= len; in ipath_sdma_verbs_send() [all …]
|
D | ipath_rc.c | 48 ss->sge = wqe->sg_list[0]; in restart_sge() 129 len = e->rdma_sge.sge.sge_length; in ipath_make_rc_ack() 162 len = qp->s_ack_rdma_sge.sge.sge_length; in ipath_make_rc_ack() 461 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_rc_req() 1439 if (unlikely(offset + len > e->rdma_sge.sge.sge_length)) in ipath_rc_rcv_error() 1454 e->rdma_sge.sge.mr = NULL; in ipath_rc_rcv_error() 1455 e->rdma_sge.sge.vaddr = NULL; in ipath_rc_rcv_error() 1456 e->rdma_sge.sge.length = 0; in ipath_rc_rcv_error() 1457 e->rdma_sge.sge.sge_length = 0; in ipath_rc_rcv_error() 1780 qp->r_sge.sge.mr = NULL; in ipath_rc_rcv() [all …]
|
D | ipath_uc.c | 98 qp->s_sge.sge = wqe->sg_list[0]; in ipath_make_uc_req() 454 qp->r_sge.sge.mr = NULL; in ipath_uc_rcv() 455 qp->r_sge.sge.vaddr = NULL; in ipath_uc_rcv() 456 qp->r_sge.sge.length = 0; in ipath_uc_rcv() 457 qp->r_sge.sge.sge_length = 0; in ipath_uc_rcv()
|
D | ipath_verbs.h | 330 struct ipath_sge sge; /* progress state for the current SGE */ member 786 struct ib_sge *sge, int acc);
|
/linux-4.1.27/drivers/scsi/esas2r/ |
D | esas2r_io.c | 222 if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) { in esas2r_build_sg_list_sge() 237 sgelen = (u8)((u8 *)sgc->sge.a64.curr in esas2r_build_sg_list_sge() 238 - (u8 *)sgc->sge.a64.last); in esas2r_build_sg_list_sge() 244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge() 247 sgc->sge.a64.curr = in esas2r_build_sg_list_sge() 252 sgc->sge.a64.limit = in esas2r_build_sg_list_sge() 257 sgc->sge.a64.last->length = cpu_to_le32( in esas2r_build_sg_list_sge() 259 sgc->sge.a64.last->address = in esas2r_build_sg_list_sge() 268 if (sgc->sge.a64.chain) { in esas2r_build_sg_list_sge() 269 sgc->sge.a64.chain->length |= in esas2r_build_sg_list_sge() [all …]
|
D | esas2r_vda.c | 119 firstsg = rq->vrq->flash.data.file.sge; in esas2r_process_vda_ioctl() 130 firstsg = rq->vrq->cli.sge; in esas2r_process_vda_ioctl() 194 firstsg = rq->vrq->mgt.sge; in esas2r_process_vda_ioctl() 363 data.sge); in esas2r_build_flash_req() 394 struct atto_vda_mgmt_req, sge); in esas2r_build_mgt_req() 396 vrq->sge[0].length = cpu_to_le32(SGE_LAST | length); in esas2r_build_mgt_req() 397 vrq->sge[0].address = cpu_to_le64( in esas2r_build_mgt_req() 432 (u8)offsetof(struct atto_vda_ae_req, sge); in esas2r_build_ae_req() 433 vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length); in esas2r_build_ae_req() 434 vrq->sge[0].address = cpu_to_le64( in esas2r_build_ae_req() [all …]
|
D | atvda.h | 504 struct atto_vda_sge sge[1]; member 557 struct atto_vda_sge sge[1]; member 1072 struct atto_vda_sge sge[1]; member 1095 struct atto_vda_sge sge[1]; member 1098 struct atto_vda_sge sge[1]; member 1120 struct atto_vda_sge sge[1]; member 1133 struct atto_vda_sge sge[1]; member 1147 struct atto_vda_sge sge[1]; member 1174 struct atto_vda_sge sge[1]; member 1191 struct atto_vda_sge sge; member [all …]
|
D | esas2r.h | 619 } sge; member 1180 sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq in esas2r_sgc_init() 1187 sgc->sge.a64.last = in esas2r_sgc_init() 1188 sgc->sge.a64.curr = first; in esas2r_sgc_init() 1193 sgc->sge.a64.last = in esas2r_sgc_init() 1194 sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0]; in esas2r_sgc_init() 1196 (u8)offsetof(struct atto_vda_scsi_req, u.sge); in esas2r_sgc_init() 1198 sgc->sge.a64.chain = NULL; in esas2r_sgc_init()
|
D | esas2r_flash.c | 212 esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]); in build_flash_msg() 915 esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge); in esas2r_process_fs_ioctl() 1311 vrq->data.sge[0].length = in esas2r_nvram_write() 1314 vrq->data.sge[0].address = cpu_to_le64( in esas2r_nvram_write()
|
D | esas2r_ioctl.c | 300 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in smp_ioctl_callback() 353 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in csmi_ioctl_tunnel() 672 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in hba_ioctl_tunnel()
|
D | esas2r_disc.c | 901 esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge); in esas2r_disc_passthru_dev_addr()
|
D | esas2r_init.c | 1280 data.sge); in esas2r_format_init_msg()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_verbs.c | 171 struct qib_sge *sge = &ss->sge; in qib_copy_sge() local 174 u32 len = sge->length; in qib_copy_sge() 178 if (len > sge->sge_length) in qib_copy_sge() 179 len = sge->sge_length; in qib_copy_sge() 181 memcpy(sge->vaddr, data, len); in qib_copy_sge() 182 sge->vaddr += len; in qib_copy_sge() 183 sge->length -= len; in qib_copy_sge() 184 sge->sge_length -= len; in qib_copy_sge() 185 if (sge->sge_length == 0) { in qib_copy_sge() 187 qib_put_mr(sge->mr); in qib_copy_sge() [all …]
|
D | qib_keys.c | 156 struct qib_sge *isge, struct ib_sge *sge, int acc) in qib_lkey_ok() argument 167 if (sge->lkey == 0) { in qib_lkey_ok() 180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok() 181 isge->length = sge->length; in qib_lkey_ok() 182 isge->sge_length = sge->length; in qib_lkey_ok() 188 rkt->table[(sge->lkey >> (32 - ib_qib_lkey_table_size))]); in qib_lkey_ok() 189 if (unlikely(!mr || mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) in qib_lkey_ok() 192 off = sge->addr - mr->user_base; in qib_lkey_ok() 193 if (unlikely(sge->addr < mr->user_base || in qib_lkey_ok() 194 off + sge->length > mr->length || in qib_lkey_ok() [all …]
|
D | qib_ruc.c | 98 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, in qib_init_sge() 111 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in qib_init_sge() local 113 qib_put_mr(sge->mr); in qib_init_sge() 360 struct qib_sge *sge; in qib_ruc_loopback() local 427 sqp->s_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback() 460 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, wqe->length, in qib_ruc_loopback() 473 if (unlikely(!qib_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, in qib_ruc_loopback() 481 qp->r_sge.sge = wqe->sg_list[0]; in qib_ruc_loopback() 491 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), in qib_ruc_loopback() 497 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; in qib_ruc_loopback() [all …]
|
D | qib_ud.c | 57 struct qib_sge *sge; in qib_ud_loopback() local 173 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 175 sge = &ssge.sge; in qib_ud_loopback() 177 u32 len = sge->length; in qib_ud_loopback() 181 if (len > sge->sge_length) in qib_ud_loopback() 182 len = sge->sge_length; in qib_ud_loopback() 184 qib_copy_sge(&qp->r_sge, sge->vaddr, len, 1); in qib_ud_loopback() 185 sge->vaddr += len; in qib_ud_loopback() 186 sge->length -= len; in qib_ud_loopback() 187 sge->sge_length -= len; in qib_ud_loopback() [all …]
|
D | qib_sdma.c | 538 struct qib_sge *sge; in qib_sdma_verbs_send() local 586 sge = &ss->sge; in qib_sdma_verbs_send() 592 if (len > sge->length) in qib_sdma_verbs_send() 593 len = sge->length; in qib_sdma_verbs_send() 594 if (len > sge->sge_length) in qib_sdma_verbs_send() 595 len = sge->sge_length; in qib_sdma_verbs_send() 598 addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr, in qib_sdma_verbs_send() 617 sge->vaddr += len; in qib_sdma_verbs_send() 618 sge->length -= len; in qib_sdma_verbs_send() 619 sge->sge_length -= len; in qib_sdma_verbs_send() [all …]
|
D | qib_uc.c | 99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req() 441 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, in qib_uc_rcv() 448 qp->r_sge.sge.mr = NULL; in qib_uc_rcv() 449 qp->r_sge.sge.vaddr = NULL; in qib_uc_rcv() 450 qp->r_sge.sge.length = 0; in qib_uc_rcv() 451 qp->r_sge.sge.sge_length = 0; in qib_uc_rcv()
|
D | qib_rc.c | 49 ss->sge = wqe->sg_list[0]; in restart_sge() 137 qp->s_ack_rdma_sge.sge = e->rdma_sge; in qib_make_rc_ack() 173 qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; in qib_make_rc_ack() 176 len = qp->s_ack_rdma_sge.sge.sge_length; in qib_make_rc_ack() 487 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_rc_req() 1013 struct qib_sge *sge = &wqe->sg_list[i]; in qib_rc_send_complete() local 1015 qib_put_mr(sge->mr); in qib_rc_send_complete() 1069 struct qib_sge *sge = &wqe->sg_list[i]; in do_rc_completion() local 1071 qib_put_mr(sge->mr); in do_rc_completion() 2070 ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, in qib_rc_rcv() [all …]
|
D | qib_verbs.h | 392 struct qib_sge sge; /* progress state for the current SGE */ member 984 struct qib_sge *isge, struct ib_sge *sge, int acc); 986 int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge, 1069 qib_put_mr(ss->sge.mr); in qib_put_ss() 1071 ss->sge = *ss->sg_list++; in qib_put_ss()
|
D | qib_qp.c | 433 struct qib_sge *sge = &wqe->sg_list[i]; in clear_mr_refs() local 435 qib_put_mr(sge->mr); in clear_mr_refs()
|
/linux-4.1.27/drivers/scsi/csiostor/ |
D | csio_wr.c | 57 csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg) in csio_get_flbuf_size() argument 59 sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A + in csio_get_flbuf_size() 65 csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf) in csio_wr_fl_bufsz() argument 67 return sge->sge_fl_buf_size[buf->paddr & 0xF]; in csio_wr_fl_bufsz() 74 return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ? 128 : 64; in csio_wr_qstat_pgsz() 117 struct csio_sge *sge = &wrm->sge; in csio_wr_fill_fl() local 125 buf->len = sge->sge_fl_buf_size[sreg]; in csio_wr_fill_fl() 1044 struct csio_sge *sge = &wrm->sge; in csio_wr_process_fl() local 1070 bufsz = csio_wr_fl_bufsz(sge, buf); in csio_wr_process_fl() 1091 flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align); in csio_wr_process_fl() [all …]
|
D | csio_wr.h | 449 struct csio_sge sge; /* SGE params */ member
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 73 vec->sge[sge_no].iov_base = xdr->head[0].iov_base; in map_xdr() 74 vec->sge[sge_no].iov_len = xdr->head[0].iov_len; in map_xdr() 82 vec->sge[sge_no].iov_base = in map_xdr() 86 vec->sge[sge_no].iov_len = sge_bytes; in map_xdr() 95 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr() 96 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr() 148 struct ib_sge *sge; in send_write() local 164 write_len, vec->sge, vec->count); in send_write() 168 sge = ctxt->sge; in send_write() 173 if (vec->sge[xdr_sge_no].iov_len > bc) in send_write() [all …]
|
D | svc_rdma_recvfrom.c | 75 min_t(size_t, byte_count, ctxt->sge[0].length); in rdma_build_arg_xdr() 98 bc -= min_t(u32, bc, ctxt->sge[sge_no].length); in rdma_build_arg_xdr() 99 rqstp->rq_arg.buflen += ctxt->sge[sge_no].length; in rdma_build_arg_xdr() 162 ctxt->sge[pno].addr = in rdma_read_chunk_lcl() 168 ctxt->sge[pno].addr); in rdma_read_chunk_lcl() 174 ctxt->sge[pno].lkey = xprt->sc_dma_lkey; in rdma_read_chunk_lcl() 175 ctxt->sge[pno].length = len; in rdma_read_chunk_lcl() 199 read_wr.sg_list = ctxt->sge; in rdma_read_chunk_lcl() 296 ctxt->sge[0].addr = (unsigned long)frmr->kva + *page_offset; in rdma_read_chunk_frmr() 297 ctxt->sge[0].lkey = frmr->mr->lkey; in rdma_read_chunk_frmr() [all …]
|
D | svc_rdma_transport.c | 120 for (i = 0; i < ctxt->count && ctxt->sge[i].length; i++) { in svc_rdma_unmap_dma() 127 if (ctxt->sge[i].lkey == xprt->sc_dma_lkey) { in svc_rdma_unmap_dma() 130 ctxt->sge[i].addr, in svc_rdma_unmap_dma() 131 ctxt->sge[i].length, in svc_rdma_unmap_dma() 531 ctxt->sge[sge_no].addr = pa; in svc_rdma_post_recv() 532 ctxt->sge[sge_no].length = PAGE_SIZE; in svc_rdma_post_recv() 533 ctxt->sge[sge_no].lkey = xprt->sc_dma_lkey; in svc_rdma_post_recv() 538 recv_wr.sg_list = &ctxt->sge[0]; in svc_rdma_post_recv() 1338 ctxt->sge[0].addr = ib_dma_map_page(xprt->sc_cm_id->device, in svc_rdma_send_error() 1340 if (ib_dma_mapping_error(xprt->sc_cm_id->device, ctxt->sge[0].addr)) { in svc_rdma_send_error() [all …]
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 170 struct sge *s = &adapter->sge; in fl_mtu_bufsize() 257 const struct sge *s = &adapter->sge; in fl_starving() 340 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl() 455 struct sge *s = &adapter->sge; in get_buf_size() 589 struct sge *s = &adap->sge; in refill_fl() 668 set_bit(q->cntxt_id - adap->sge.egr_start, in refill_fl() 669 adap->sge.starving_fl); in refill_fl() 881 to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; in write_sgl() 895 unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1; in write_sgl() 898 memcpy(sgl->sge, buf, part0); in write_sgl() [all …]
|
D | cxgb4_main.c | 301 struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; in dcb_tx_queue_prio_enable() 614 txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; in fwevtq_handler() 616 if ((u8 *)txq < (u8 *)q->adap->sge.ofldtxq) { in fwevtq_handler() 757 for_each_ofldrxq(&adap->sge, i) in name_msix_vecs() 761 for_each_rdmarxq(&adap->sge, i) in name_msix_vecs() 765 for_each_rdmaciq(&adap->sge, i) in name_msix_vecs() 772 struct sge *s = &adap->sge; in request_msix_queue_irqs() 839 struct sge *s = &adap->sge; in free_msix_queue_irqs() 864 const struct sge_eth_rxq *q = &pi->adapter->sge.ethrxq[pi->first_qset]; in cxgb4_write_rss() 903 static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) in rxq_to_chan() [all …]
|
D | cxgb4_debugfs.c | 1684 int eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4); in sge_qinfo_show() 1685 int toe_entries = DIV_ROUND_UP(adap->sge.ofldqsets, 4); in sge_qinfo_show() 1686 int rdma_entries = DIV_ROUND_UP(adap->sge.rdmaqs, 4); in sge_qinfo_show() 1687 int ciq_entries = DIV_ROUND_UP(adap->sge.rdmaciqs, 4); in sge_qinfo_show() 1712 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[base_qset]; in sge_qinfo_show() 1713 const struct sge_eth_txq *tx = &adap->sge.ethtxq[base_qset]; in sge_qinfo_show() 1714 int n = min(4, adap->sge.ethqsets - 4 * r); in sge_qinfo_show() 1740 adap->sge.counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show() 1748 const struct sge_ofld_rxq *rx = &adap->sge.ofldrxq[toe_idx * 4]; in sge_qinfo_show() 1749 const struct sge_ofld_txq *tx = &adap->sge.ofldtxq[toe_idx * 4]; in sge_qinfo_show() [all …]
|
D | cxgb4_ethtool.c | 193 const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; in collect_sge_port_stats() 194 const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset]; in collect_sge_port_stats() 483 const struct sge *s = &pi->adapter->sge; in get_sge_param() 501 struct sge *s = &adapter->sge; in set_sge_param() 535 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_rx_intr_params() 550 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in set_adaptive_rx_setting() 562 struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; in get_adaptive_rx_setting() 578 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce() 582 adap->sge.counter_val[rq->pktcnt_idx] : 0; in get_coalesce()
|
D | cxgb4.h | 296 struct sge_params sge; member 602 struct sge { struct 651 #define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++) argument 652 #define for_each_ofldrxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++) argument 653 #define for_each_rdmarxq(sge, i) for (i = 0; i < (sge)->rdmaqs; i++) argument 654 #define for_each_rdmaciq(sge, i) for (i = 0; i < (sge)->rdmaciqs; i++) argument 692 struct sge sge; member 1036 return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; in qtimer_val()
|
D | Makefile | 7 cxgb4-objs := cxgb4_main.o l2t.o t4_hw.o sge.o clip_tbl.o cxgb4_ethtool.o
|
D | l2t.c | 154 TID_QID_V(adap->sge.fw_evtq.abs_id))); in write_l2e()
|
D | t4_msg.h | 1054 struct ulptx_sge_pair sge[0]; member
|
D | t4_hw.c | 5142 page_shift = adapter->params.sge.hps + 10; in cxgb4_t4_bar2_sge_qregs() 5148 ? adapter->params.sge.eq_qpp in cxgb4_t4_bar2_sge_qregs() 5149 : adapter->params.sge.iq_qpp); in cxgb4_t4_bar2_sge_qregs() 5250 struct sge_params *sge_params = &adapter->params.sge; in t4_init_sge_params()
|
/linux-4.1.27/net/rds/ |
D | iw_recv.c | 88 struct ib_sge *sge; in rds_iw_recv_init_ring() local 98 sge = rds_iw_data_sge(ic, recv->r_sge); in rds_iw_recv_init_ring() 99 sge->addr = 0; in rds_iw_recv_init_ring() 100 sge->length = RDS_FRAG_SIZE; in rds_iw_recv_init_ring() 101 sge->lkey = 0; in rds_iw_recv_init_ring() 103 sge = rds_iw_header_sge(ic, recv->r_sge); in rds_iw_recv_init_ring() 104 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_recv_init_ring() 105 sge->length = sizeof(struct rds_header); in rds_iw_recv_init_ring() 106 sge->lkey = 0; in rds_iw_recv_init_ring() 143 struct ib_sge *sge; in rds_iw_recv_refill_one() local [all …]
|
D | ib_recv.c | 52 struct ib_sge *sge; in rds_ib_recv_init_ring() local 62 sge = &recv->r_sge[0]; in rds_ib_recv_init_ring() 63 sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_recv_init_ring() 64 sge->length = sizeof(struct rds_header); in rds_ib_recv_init_ring() 65 sge->lkey = ic->i_mr->lkey; in rds_ib_recv_init_ring() 67 sge = &recv->r_sge[1]; in rds_ib_recv_init_ring() 68 sge->addr = 0; in rds_ib_recv_init_ring() 69 sge->length = RDS_FRAG_SIZE; in rds_ib_recv_init_ring() 70 sge->lkey = ic->i_mr->lkey; in rds_ib_recv_init_ring() 303 struct ib_sge *sge; in rds_ib_recv_refill_one() local [all …]
|
D | iw_send.c | 134 struct ib_sge *sge; in rds_iw_send_init_ring() local 148 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_send_init_ring() 149 sge->lkey = 0; in rds_iw_send_init_ring() 151 sge = rds_iw_header_sge(ic, send->s_sge); in rds_iw_send_init_ring() 152 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_iw_send_init_ring() 153 sge->length = sizeof(struct rds_header); in rds_iw_send_init_ring() 154 sge->lkey = 0; in rds_iw_send_init_ring() 453 struct ib_sge *sge; in rds_iw_xmit_populate_wr() local 465 sge = rds_iw_data_sge(ic, send->s_sge); in rds_iw_xmit_populate_wr() 466 sge->addr = buffer; in rds_iw_xmit_populate_wr() [all …]
|
D | iw.h | 384 rds_iw_header_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_header_sge() argument 386 return &sge[0]; in rds_iw_header_sge() 390 rds_iw_data_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_data_sge() argument 392 return &sge[1]; in rds_iw_data_sge()
|
D | ib_send.c | 228 struct ib_sge *sge; in rds_ib_send_init_ring() local 236 sge = &send->s_sge[0]; in rds_ib_send_init_ring() 237 sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); in rds_ib_send_init_ring() 238 sge->length = sizeof(struct rds_header); in rds_ib_send_init_ring() 239 sge->lkey = ic->i_mr->lkey; in rds_ib_send_init_ring()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | cxgb4vf_main.c | 323 struct sge *s = &adapter->sge; in request_msix_queue_irqs() 361 struct sge *s = &adapter->sge; in free_msix_queue_irqs() 394 struct sge *s = &adapter->sge; in enable_rx() 417 struct sge *s = &adapter->sge; in quiesce_rx() 475 struct sge *s = &adapter->sge; in fwevtq_handler() 531 struct sge *s = &adapter->sge; in setup_sge_queues() 646 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; in setup_rss() 986 static int closest_timer(const struct sge *s, int us) in closest_timer() 1002 static int closest_thres(const struct sge *s, int thres) in closest_thres() 1027 ? adapter->sge.timer_val[timer_idx] in qtimer_val() [all …]
|
D | sge.c | 264 const struct sge *s = &adapter->sge; in fl_starving() 325 for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { in unmap_sgl() 457 const struct sge *s = &adapter->sge; in get_buf_size() 609 struct sge *s = &adapter->sge; in refill_fl() 717 set_bit(fl->cntxt_id, adapter->sge.starving_fl); in refill_fl() 937 to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; in write_sgl() 951 unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; in write_sgl() 954 memcpy(sgl->sge, buf, part0); in write_sgl() 1193 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit() 1555 struct sge *s = &adapter->sge; in do_gro() [all …]
|
D | adapter.h | 269 struct sge { struct 348 #define for_each_ethrxq(sge, iter) \ argument 349 for (iter = 0; iter < (sge)->ethqsets; iter++) 372 struct sge sge; member
|
D | Makefile | 7 cxgb4vf-objs := cxgb4vf_main.o t4vf_hw.o sge.o
|
D | t4vf_common.h | 214 struct sge_params sge; /* Scatter Gather Engine */ member
|
D | t4vf_hw.c | 472 page_shift = adapter->params.sge.sge_vf_hps + 10; in t4_bar2_sge_qregs() 478 ? adapter->params.sge.sge_vf_eq_qpp in t4_bar2_sge_qregs() 479 : adapter->params.sge.sge_vf_iq_qpp); in t4_bar2_sge_qregs() 529 struct sge_params *sge_params = &adapter->params.sge; in t4vf_get_sge_params()
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 396 reg->sge.lkey = device->mr->lkey; in iser_reg_dma() 398 reg->sge.addr = ib_sg_dma_address(device->ib_device, &sg[0]); in iser_reg_dma() 399 reg->sge.length = ib_sg_dma_len(device->ib_device, &sg[0]); in iser_reg_dma() 402 " length=0x%x\n", reg->sge.lkey, reg->rkey, in iser_reg_dma() 403 reg->sge.addr, reg->sge.length); in iser_reg_dma() 474 mem_reg->sge.lkey = fmr->fmr->lkey; in iser_reg_page_vec() 476 mem_reg->sge.addr = page_vec->pages[0] + page_vec->offset; in iser_reg_page_vec() 477 mem_reg->sge.length = page_vec->data_size; in iser_reg_page_vec() 678 sig_wr.sg_list = &data_reg->sge; in iser_reg_sig_mr() 683 sig_wr.wr.sig_handover.prot = &prot_reg->sge; in iser_reg_sig_mr() [all …]
|
D | iser_initiator.c | 85 hdr->read_va = cpu_to_be64(mem_reg->sge.addr); in iser_prepare_read_cmd() 89 (unsigned long long)mem_reg->sge.addr); in iser_prepare_read_cmd() 142 hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz); in iser_prepare_write_cmd() 147 (unsigned long long)mem_reg->sge.addr, unsol_sz); in iser_prepare_write_cmd() 153 tx_dsg->addr = mem_reg->sge.addr; in iser_prepare_write_cmd() 155 tx_dsg->lkey = mem_reg->sge.lkey; in iser_prepare_write_cmd() 484 tx_dsg->addr = mem_reg->sge.addr + buf_offset; in iser_send_data_out() 486 tx_dsg->lkey = mem_reg->sge.lkey; in iser_send_data_out()
|
D | iser_verbs.c | 1007 struct ib_sge sge; in iser_post_recvl() local 1010 sge.addr = iser_conn->login_resp_dma; in iser_post_recvl() 1011 sge.length = ISER_RX_LOGIN_SIZE; in iser_post_recvl() 1012 sge.lkey = ib_conn->device->mr->lkey; in iser_post_recvl() 1015 rx_wr.sg_list = &sge; in iser_post_recvl()
|
D | iscsi_iser.h | 251 struct ib_sge sge; member
|
/linux-4.1.27/net/9p/ |
D | trans_rdma.c | 408 struct ib_sge sge; in post_recv() local 416 sge.addr = c->busa; in post_recv() 417 sge.length = client->msize; in post_recv() 418 sge.lkey = rdma->lkey; in post_recv() 423 wr.sg_list = &sge; in post_recv() 436 struct ib_sge sge; in rdma_request() local 507 sge.addr = c->busa; in rdma_request() 508 sge.length = c->req->tc->size; in rdma_request() 509 sge.lkey = rdma->lkey; in rdma_request() 516 wr.sg_list = &sge; in rdma_request()
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 282 struct be_sge *sge; in mgmt_get_session_info() local 299 sge = nonembedded_sgl(wrb); in mgmt_get_session_info() 309 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); in mgmt_get_session_info() 310 sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); in mgmt_get_session_info() 311 sge->len = cpu_to_le32(nonemb_cmd->size); in mgmt_get_session_info() 428 struct be_sge *sge = nonembedded_sgl(wrb); in mgmt_check_supported_fw() local 448 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); in mgmt_check_supported_fw() 449 sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF); in mgmt_check_supported_fw() 450 sge->len = cpu_to_le32(nonemb_cmd.size); in mgmt_check_supported_fw() 591 struct be_sge *sge; in mgmt_invalidate_icds() local [all …]
|
/linux-4.1.27/fs/exofs/ |
D | ore_raid.c | 273 struct osd_sg_entry *sge; in _ore_add_sg_seg() local 282 sge = per_dev->sglist; in _ore_add_sg_seg() 287 sge->offset = per_dev->offset; in _ore_add_sg_seg() 288 sge->len = per_dev->length; in _ore_add_sg_seg() 300 sge = &per_dev->sglist[per_dev->cur_sg - 1]; in _ore_add_sg_seg() 301 sge->len = per_dev->length - per_dev->last_sgs_total; in _ore_add_sg_seg() 306 struct osd_sg_entry *next_sge = sge + 1; in _ore_add_sg_seg() 309 next_sge->offset = sge->offset + sge->len + cur_len; in _ore_add_sg_seg() 313 } else if (!sge->len) { in _ore_add_sg_seg()
|
/linux-4.1.27/drivers/ata/ |
D | sata_sil24.c | 257 struct sil24_sge sge[SIL24_MAX_SGE]; member 263 struct sil24_sge sge[SIL24_MAX_SGE]; member 780 struct sil24_sge *sge) in sil24_fill_sg() argument 787 sge->addr = cpu_to_le64(sg_dma_address(sg)); in sil24_fill_sg() 788 sge->cnt = cpu_to_le32(sg_dma_len(sg)); in sil24_fill_sg() 789 sge->flags = 0; in sil24_fill_sg() 791 last_sge = sge; in sil24_fill_sg() 792 sge++; in sil24_fill_sg() 849 struct sil24_sge *sge; in sil24_qc_prep() local 856 sge = cb->ata.sge; in sil24_qc_prep() [all …]
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 679 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset() 681 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset() 692 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset() 694 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset() 708 spin_lock_irq(&adapter->sge.reg_lock); in t3_free_qset() 710 spin_unlock_irq(&adapter->sge.reg_lock); in t3_free_qset() 1516 ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb); in t3_mgmt_tx() 1770 struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; in t3_offload_tx() 2029 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in rx_eth() 2131 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in lro_add_page() [all …]
|
D | adapter.h | 216 struct sge { struct 245 struct sge sge; member
|
D | cxgb3_main.c | 411 adap->sge.qs[qidx]. in request_msix_data_irqs() 414 &adap->sge.qs[qidx]); in request_msix_data_irqs() 418 &adap->sge.qs[qidx]); in request_msix_data_irqs() 438 &adapter->sge.qs[i]); in free_irq_resources() 448 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies() 461 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity() 600 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs() 613 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi() 638 if (adap->sge.qs[i].adap) in quiesce_rx() 639 napi_disable(&adap->sge.qs[i].napi); in quiesce_rx() [all …]
|
D | Makefile | 8 xgmac.o sge.o l2t.o cxgb3_offload.o aq100x.o
|
D | cxgb3_offload.c | 303 spin_lock_irqsave(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl() 306 spin_unlock_irqrestore(&adapter->sge.reg_lock, flags); in cxgb_rdma_ctl() 334 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl() 341 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl() 345 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl() 347 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl() 352 spin_lock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl() 358 spin_unlock_irq(&adapter->sge.reg_lock); in cxgb_rdma_ctl()
|
D | common.h | 382 struct sge_params sge; member
|
D | t3_hw.c | 3369 min(adapter->params.sge.max_pkt_size, in t3_init_hw() 3372 min(adapter->params.sge.max_pkt_size, 16384U)); in t3_init_hw() 3389 t3_sge_init(adapter, &adapter->params.sge); in t3_init_hw() 3653 t3_sge_prep(adapter, &adapter->params.sge); in t3_prep_adapter()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
D | lpfc_mbox.c | 1678 nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr); in lpfc_sli4_mbx_sge_set() 1679 nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr); in lpfc_sli4_mbx_sge_set() 1680 nembed_sge->sge[sgentry].length = length; in lpfc_sli4_mbx_sge_set() 1693 struct lpfc_mbx_sge *sge) in lpfc_sli4_mbx_sge_get() argument 1699 sge->pa_lo = nembed_sge->sge[sgentry].pa_lo; in lpfc_sli4_mbx_sge_get() 1700 sge->pa_hi = nembed_sge->sge[sgentry].pa_hi; in lpfc_sli4_mbx_sge_get() 1701 sge->length = nembed_sge->sge[sgentry].length; in lpfc_sli4_mbx_sge_get() 1715 struct lpfc_mbx_sge sge; in lpfc_sli4_mbox_cmd_free() local 1736 lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge); in lpfc_sli4_mbox_cmd_free() 1737 phyaddr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_mbox_cmd_free() [all …]
|
D | lpfc_hbadisc.c | 1803 struct lpfc_mbx_sge sge; in lpfc_sli4_fcf_rec_mbox_parse() local 1812 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_fcf_rec_mbox_parse() 1813 phys_addr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_fcf_rec_mbox_parse()
|
D | lpfc_bsg.c | 4811 receive_length = nembed_sge->sge[0].length; in lpfc_bsg_issue_mbox() 4823 nembed_sge->sge[0].pa_hi = in lpfc_bsg_issue_mbox() 4826 nembed_sge->sge[0].pa_lo = in lpfc_bsg_issue_mbox()
|
D | lpfc_hw4.h | 1024 struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES]; member
|
D | lpfc_sli.c | 15942 struct lpfc_mbx_sge sge; in lpfc_sli4_add_fcf_record() local 15973 lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); in lpfc_sli4_add_fcf_record() 15974 phys_addr = getPaddr(sge.pa_hi, sge.pa_lo); in lpfc_sli4_add_fcf_record()
|
/linux-4.1.27/drivers/scsi/aacraid/ |
D | aachba.c | 3352 rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32)); in aac_build_sgraw2() 3353 rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff)); in aac_build_sgraw2() 3355 rio2->sge[i].length = cur_size; in aac_build_sgraw2() 3356 rio2->sge[i].flags = 0; in aac_build_sgraw2() 3373 u32 temp = le32_to_cpu(rio2->sge[i-1].length) - in aac_build_sgraw2() 3375 rio2->sge[i-1].length = cpu_to_le32(temp); in aac_build_sgraw2() 3388 if (rio2->sge[j].length % (i*PAGE_SIZE)) { in aac_build_sgraw2() 3392 nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE)); in aac_build_sgraw2() 3414 struct sge_ieee1212 *sge; in aac_convert_sgraw2() local 3421 sge = kmalloc(nseg_new * sizeof(struct sge_ieee1212), GFP_ATOMIC); in aac_convert_sgraw2() [all …]
|
D | aacraid.h | 1476 struct sge_ieee1212 sge[1]; member
|
/linux-4.1.27/include/linux/sunrpc/ |
D | svc_rdma.h | 85 struct ib_sge sge[RPCSVC_MAXPAGES]; member 119 struct kvec sge[RPCSVC_MAXPAGES]; member
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_reqs.c | 119 struct ib_sge *sge = send_wr->sg_list; in trace_send_wr_ud() local 138 u8 *data = __va(sge->addr); in trace_send_wr_ud() 141 idx, j, data, sge->length, sge->lkey); in trace_send_wr_ud() 143 ehca_dmp(data, sge->length, "send_wr#%x sge#%x", in trace_send_wr_ud() 145 sge++; in trace_send_wr_ud()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 1943 struct ocrdma_sge *sge, int num_sge, in ocrdma_build_sges() argument 1949 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges() 1950 sge[i].addr_lo = sg_list[i].addr; in ocrdma_build_sges() 1951 sge[i].addr_hi = upper_32_bits(sg_list[i].addr); in ocrdma_build_sges() 1952 sge[i].len = sg_list[i].length; in ocrdma_build_sges() 1956 memset(sge, 0, sizeof(*sge)); in ocrdma_build_sges() 1971 struct ocrdma_sge *sge, in ocrdma_build_inline_sges() argument 1985 dpp_addr = (char *)sge; in ocrdma_build_inline_sges() 1998 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_inline_sges() 2013 struct ocrdma_sge *sge; in ocrdma_build_send() local [all …]
|
D | ocrdma_hw.c | 1234 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); in ocrdma_mbx_rdma_stats() 1235 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); in ocrdma_mbx_rdma_stats() 1236 mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size; in ocrdma_mbx_rdma_stats() 1282 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); in ocrdma_mbx_get_ctrl_attribs() 1283 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); in ocrdma_mbx_get_ctrl_attribs() 1284 mqe->u.nonemb_req.sge[0].len = dma.size; in ocrdma_mbx_get_ctrl_attribs() 2811 struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge; in ocrdma_mbx_get_dcbx_config()
|
D | ocrdma_sli.h | 267 struct ocrdma_mqe_sge sge[19]; member
|
/linux-4.1.27/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 1133 struct ib_sge sge; in isert_rdma_post_recvl() local 1136 memset(&sge, 0, sizeof(struct ib_sge)); in isert_rdma_post_recvl() 1137 sge.addr = isert_conn->login_req_dma; in isert_rdma_post_recvl() 1138 sge.length = ISER_RX_LOGIN_SIZE; in isert_rdma_post_recvl() 1139 sge.lkey = isert_conn->device->mr->lkey; in isert_rdma_post_recvl() 1142 sge.addr, sge.length, sge.lkey); in isert_rdma_post_recvl() 1146 rx_wr.sg_list = &sge; in isert_rdma_post_recvl() 2577 struct ib_sge *sge) in isert_fast_reg_mr() argument 2589 sge->lkey = device->mr->lkey; in isert_fast_reg_mr() 2590 sge->addr = ib_sg_dma_address(ib_dev, &mem->sg[0]); in isert_fast_reg_mr() [all …]
|
/linux-4.1.27/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 1059 kfree(ioctx->rdma_ius[--ioctx->n_rdma].sge); in srpt_unmap_sg_to_ib_sge() 1089 struct ib_sge *sge; in srpt_map_sg_to_ib_sge() local 1173 riu->sge = in srpt_map_sg_to_ib_sge() 1174 kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge() 1176 if (!riu->sge) in srpt_map_sg_to_ib_sge() 1187 riu->sge = kmalloc(riu->sge_cnt * sizeof *riu->sge, in srpt_map_sg_to_ib_sge() 1189 if (!riu->sge) in srpt_map_sg_to_ib_sge() 1204 sge = riu->sge; in srpt_map_sg_to_ib_sge() 1208 sge->addr = dma_addr; in srpt_map_sg_to_ib_sge() 1209 sge->lkey = ch->sport->sdev->mr->lkey; in srpt_map_sg_to_ib_sge() [all …]
|
D | ib_srpt.h | 156 struct ib_sge *sge; member
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_cm.c | 117 struct ib_sge *sge, int id) in ipoib_cm_post_receive_nonsrq() argument 126 sge[i].addr = rx->rx_ring[id].mapping[i]; in ipoib_cm_post_receive_nonsrq() 329 struct ib_sge *sge) in ipoib_cm_init_rx_wr() argument 335 sge[i].lkey = priv->mr->lkey; in ipoib_cm_init_rx_wr() 337 sge[0].length = IPOIB_CM_HEAD_SIZE; in ipoib_cm_init_rx_wr() 339 sge[i].length = PAGE_SIZE; in ipoib_cm_init_rx_wr() 342 wr->sg_list = sge; in ipoib_cm_init_rx_wr() 352 struct ib_sge sge[IPOIB_CM_RX_SG]; in ipoib_cm_nonsrq_init_rx() member 370 ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); in ipoib_cm_nonsrq_init_rx() 392 ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); in ipoib_cm_nonsrq_init_rx()
|
/linux-4.1.27/drivers/scsi/ |
D | vmw_pvscsi.c | 52 struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; member 336 struct PVSCSISGElement *sge; in pvscsi_create_sg() local 340 sge = &ctx->sgl->sge[0]; in pvscsi_create_sg() 342 sge[i].addr = sg_dma_address(sg); in pvscsi_create_sg() 343 sge[i].length = sg_dma_len(sg); in pvscsi_create_sg() 344 sge[i].flags = 0; in pvscsi_create_sg()
|
/linux-4.1.27/drivers/infiniband/core/ |
D | mad.c | 1066 struct ib_sge *sge; in ib_send_mad() local 1076 sge = mad_send_wr->sg_list; in ib_send_mad() 1077 sge[0].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad() 1079 sge[0].length, in ib_send_mad() 1081 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr))) in ib_send_mad() 1084 mad_send_wr->header_mapping = sge[0].addr; in ib_send_mad() 1086 sge[1].addr = ib_dma_map_single(mad_agent->device, in ib_send_mad() 1088 sge[1].length, in ib_send_mad() 1090 if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) { in ib_send_mad() 1093 sge[0].length, DMA_TO_DEVICE); in ib_send_mad() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd_cb.c | 1055 struct ib_sge *sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_tx_msg() local 1069 sge->lkey = mr->lkey; in kiblnd_init_tx_msg() 1070 sge->addr = tx->tx_msgaddr; in kiblnd_init_tx_msg() 1071 sge->length = nob; in kiblnd_init_tx_msg() 1077 wrq->sg_list = sge; in kiblnd_init_tx_msg() 1091 struct ib_sge *sge = &tx->tx_sge[0]; in kiblnd_init_rdma() local 1132 sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_rdma() 1133 sge->addr = kiblnd_rd_frag_addr(srcrd, srcidx); in kiblnd_init_rdma() 1134 sge->lkey = kiblnd_rd_frag_key(srcrd, srcidx); in kiblnd_init_rdma() 1135 sge->length = wrknob; in kiblnd_init_rdma() [all …]
|
/linux-4.1.27/drivers/scsi/bfa/ |
D | bfa_fcpim.c | 2461 struct bfi_sge_s *sge, *sgpge; in bfa_ioim_send_ioreq() local 2489 sge = &m->sges[0]; in bfa_ioim_send_ioreq() 2498 sge->sga = *(union bfi_addr_u *) &addr; in bfa_ioim_send_ioreq() 2500 sge->sg_len = pgdlen; in bfa_ioim_send_ioreq() 2501 sge->flags = (ioim->nsges > BFI_SGE_INLINE) ? in bfa_ioim_send_ioreq() 2503 bfa_sge_to_be(sge); in bfa_ioim_send_ioreq() 2504 sge++; in bfa_ioim_send_ioreq() 2545 sge->sga = ioim->sgpg->sgpg_pa; in bfa_ioim_send_ioreq() 2547 sge->sga.a32.addr_lo = 0; in bfa_ioim_send_ioreq() 2548 sge->sga.a32.addr_hi = 0; in bfa_ioim_send_ioreq() [all …]
|
D | bfad_bsg.c | 3197 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sgaddr_cb() local 3200 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sgaddr_cb() 3201 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_req_sgaddr_cb() 3209 struct bfa_sge_s *sge; in bfad_fcxp_get_req_sglen_cb() local 3211 sge = drv_fcxp->req_sge + sgeid; in bfad_fcxp_get_req_sglen_cb() 3212 return sge->sg_len; in bfad_fcxp_get_req_sglen_cb() 3219 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sgaddr_cb() local 3222 sge = drv_fcxp->rsp_sge + sgeid; in bfad_fcxp_get_rsp_sgaddr_cb() 3223 addr = (u64)(size_t) sge->sg_addr; in bfad_fcxp_get_rsp_sgaddr_cb() 3231 struct bfa_sge_s *sge; in bfad_fcxp_get_rsp_sglen_cb() local [all …]
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.h | 801 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_free_rx_sge() local 812 sge->addr_hi = 0; in bnx2x_free_rx_sge() 813 sge->addr_lo = 0; in bnx2x_free_rx_sge()
|
D | bnx2x_cmn.c | 549 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; in bnx2x_alloc_rx_sge() local 568 sge->addr_hi = cpu_to_le32(U64_HI(mapping)); in bnx2x_alloc_rx_sge() 569 sge->addr_lo = cpu_to_le32(U64_LO(mapping)); in bnx2x_alloc_rx_sge() 1331 struct eth_rx_sge *sge; in bnx2x_set_next_page_sgl() local 1333 sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2]; in bnx2x_set_next_page_sgl() 1334 sge->addr_hi = in bnx2x_set_next_page_sgl() 1338 sge->addr_lo = in bnx2x_set_next_page_sgl()
|
/linux-4.1.27/drivers/message/fusion/ |
D | mptbase.h | 334 SGESimple32_t sge[1]; member 338 SGESimple64_t sge[1]; member
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
D | t4fw_ri_api.h | 124 struct fw_ri_dsge_pair sge[0]; member 140 struct fw_ri_sge sge[0]; member
|
D | qp.c | 410 __be64 *flitp = (__be64 *)isglp->sge; in build_isgl()
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
D | be_cmds.c | 705 struct be_sge *sge; in be_wrb_cmd_hdr_prepare() local 716 sge = nonembedded_sgl(wrb); in be_wrb_cmd_hdr_prepare() 717 sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma)); in be_wrb_cmd_hdr_prepare() 718 sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF); in be_wrb_cmd_hdr_prepare() 719 sge->len = cpu_to_le32(mem->size); in be_wrb_cmd_hdr_prepare()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
D | cnic_defs.h | 3097 struct fcoe_bd_ctx sge[3]; member
|