sge 247 drivers/ata/sata_sil24.c struct sil24_sge sge[SIL24_MAX_SGE]; sge 253 drivers/ata/sata_sil24.c struct sil24_sge sge[SIL24_MAX_SGE]; sge 770 drivers/ata/sata_sil24.c struct sil24_sge *sge) sge 777 drivers/ata/sata_sil24.c sge->addr = cpu_to_le64(sg_dma_address(sg)); sge 778 drivers/ata/sata_sil24.c sge->cnt = cpu_to_le32(sg_dma_len(sg)); sge 779 drivers/ata/sata_sil24.c sge->flags = 0; sge 781 drivers/ata/sata_sil24.c last_sge = sge; sge 782 drivers/ata/sata_sil24.c sge++; sge 839 drivers/ata/sata_sil24.c struct sil24_sge *sge; sge 846 drivers/ata/sata_sil24.c sge = cb->ata.sge; sge 860 drivers/ata/sata_sil24.c sge = cb->atapi.sge; sge 876 drivers/ata/sata_sil24.c sil24_fill_sg(qc, sge); sge 28 drivers/crypto/bcm/util.c struct scatterlist **sge, unsigned int *sge_offset) sge 45 drivers/crypto/bcm/util.c *sge = sg; sge 71 drivers/crypto/bcm/util.h struct scatterlist **sge, unsigned int *sge_offset); sge 480 drivers/crypto/chelsio/chcr_algo.c walk->pair = ulp->sge; sge 584 drivers/crypto/chelsio/chcr_algo.c adap->sge.uld_txq_info[CXGB4_TX_CRYPTO]; sge 396 drivers/crypto/chelsio/chcr_ipsec.c q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 444 drivers/crypto/chelsio/chcr_ipsec.c q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 487 drivers/crypto/chelsio/chcr_ipsec.c q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 543 drivers/crypto/chelsio/chcr_ipsec.c struct sge_eth_txq *q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 701 drivers/crypto/chelsio/chcr_ipsec.c q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 1215 drivers/infiniband/core/mad.c struct ib_sge *sge; sge 1226 drivers/infiniband/core/mad.c sge = mad_send_wr->sg_list; sge 1227 drivers/infiniband/core/mad.c sge[0].addr = ib_dma_map_single(mad_agent->device, sge 1229 drivers/infiniband/core/mad.c sge[0].length, sge 1231 drivers/infiniband/core/mad.c if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr))) sge 1234 drivers/infiniband/core/mad.c mad_send_wr->header_mapping = sge[0].addr; sge 1236 drivers/infiniband/core/mad.c sge[1].addr = ib_dma_map_single(mad_agent->device, sge 1238 drivers/infiniband/core/mad.c sge[1].length, sge 1240 drivers/infiniband/core/mad.c if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) { sge 1243 drivers/infiniband/core/mad.c sge[0].length, DMA_TO_DEVICE); sge 1246 drivers/infiniband/core/mad.c mad_send_wr->payload_mapping = sge[1].addr; sge 1267 drivers/infiniband/core/mad.c sge[0].length, DMA_TO_DEVICE); sge 1270 drivers/infiniband/core/mad.c sge[1].length, DMA_TO_DEVICE); sge 113 drivers/infiniband/core/rw.c reg->sge.addr = reg->mr->iova; sge 114 drivers/infiniband/core/rw.c reg->sge.length = reg->mr->length; sge 153 drivers/infiniband/core/rw.c reg->wr.wr.sg_list = ®->sge; sge 167 drivers/infiniband/core/rw.c remote_addr += reg->sge.length; sge 195 drivers/infiniband/core/rw.c struct ib_sge *sge; sge 200 drivers/infiniband/core/rw.c ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); sge 219 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = sge; sge 222 drivers/infiniband/core/rw.c sge->addr = sg_dma_address(sg) + offset; sge 223 drivers/infiniband/core/rw.c sge->length = sg_dma_len(sg) - offset; sge 224 drivers/infiniband/core/rw.c sge->lkey = qp->pd->local_dma_lkey; sge 226 drivers/infiniband/core/rw.c total_len += sge->length; sge 227 drivers/infiniband/core/rw.c sge++; sge 253 drivers/infiniband/core/rw.c ctx->single.sge.lkey = qp->pd->local_dma_lkey; sge 254 drivers/infiniband/core/rw.c ctx->single.sge.addr = sg_dma_address(sg) + offset; sge 255 drivers/infiniband/core/rw.c ctx->single.sge.length = sg_dma_len(sg) - offset; sge 262 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = &ctx->single.sge; sge 439 drivers/infiniband/core/rw.c ctx->reg->sge.addr = ctx->reg->mr->iova; sge 440 drivers/infiniband/core/rw.c ctx->reg->sge.length = ctx->reg->mr->length; sge 442 drivers/infiniband/core/rw.c ctx->reg->sge.length -= ctx->reg->mr->sig_attrs->meta_length; sge 445 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = &ctx->reg->sge; sge 482 drivers/infiniband/core/rw.c reg->sge.lkey = reg->mr->lkey; sge 100 drivers/infiniband/hw/bnxt_re/bnxt_re.h struct bnxt_qplib_sge sge; sge 1844 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_qplib_sge sge; sge 1961 drivers/infiniband/hw/bnxt_re/ib_verbs.c buf = bnxt_qplib_get_qp1_sq_buf(&qp->qplib_qp, &sge); sge 1978 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge.size -= 20; sge 1985 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge.size -= 8; sge 1989 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge.size -= 4; sge 1991 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->sg_list[0].addr = sge.addr; sge 1992 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->sg_list[0].lkey = sge.lkey; sge 1993 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->sg_list[0].size = sge.size; sge 2014 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_qplib_sge ref, sge; sge 2020 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (!bnxt_qplib_get_qp1_rq_buf(&qp->qplib_qp, &sge)) sge 2034 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->sg_list[0].addr = sge.addr; sge 2035 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->sg_list[0].lkey = sge.lkey; sge 2037 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge.size -= wqe->sg_list[0].size; sge 2039 drivers/infiniband/hw/bnxt_re/ib_verbs.c sqp_entry->sge.addr = ref.addr; sge 2040 drivers/infiniband/hw/bnxt_re/ib_verbs.c sqp_entry->sge.lkey = ref.lkey; sge 2041 drivers/infiniband/hw/bnxt_re/ib_verbs.c sqp_entry->sge.size = ref.size; sge 2935 drivers/infiniband/hw/bnxt_re/ib_verbs.c r_sge[1].addr = sqp_entry->sge.addr + offset; sge 2936 drivers/infiniband/hw/bnxt_re/ib_verbs.c r_sge[1].lkey = sqp_entry->sge.lkey; sge 1457 drivers/infiniband/hw/bnxt_re/qplib_fp.c struct bnxt_qplib_sge *sge) sge 1462 drivers/infiniband/hw/bnxt_re/qplib_fp.c memset(sge, 0, sizeof(*sge)); sge 1466 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->addr = (dma_addr_t)(qp->sq_hdr_buf_map + sge 1468 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->lkey = 0xFFFFFFFF; sge 1469 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->size = qp->sq_hdr_buf_size; sge 1470 drivers/infiniband/hw/bnxt_re/qplib_fp.c return qp->sq_hdr_buf + sw_prod * sge->size; sge 1488 drivers/infiniband/hw/bnxt_re/qplib_fp.c struct bnxt_qplib_sge *sge) sge 1493 drivers/infiniband/hw/bnxt_re/qplib_fp.c memset(sge, 0, sizeof(*sge)); sge 1497 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->addr = (dma_addr_t)(qp->rq_hdr_buf_map + sge 1499 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->lkey = 0xFFFFFFFF; sge 1500 drivers/infiniband/hw/bnxt_re/qplib_fp.c sge->size = qp->rq_hdr_buf_size; sge 1501 drivers/infiniband/hw/bnxt_re/qplib_fp.c return qp->rq_hdr_buf + sw_prod * sge->size; sge 534 drivers/infiniband/hw/bnxt_re/qplib_fp.h struct bnxt_qplib_sge *sge); sge 536 drivers/infiniband/hw/bnxt_re/qplib_fp.h struct bnxt_qplib_sge *sge); sge 464 drivers/infiniband/hw/cxgb4/qp.c flitp = (__be64 *)isglp->sge; sge 126 drivers/infiniband/hw/cxgb4/t4fw_ri_api.h struct fw_ri_dsge_pair sge[0]; sge 142 drivers/infiniband/hw/cxgb4/t4fw_ri_api.h struct fw_ri_sge sge[0]; sge 200 drivers/infiniband/hw/hfi1/rc.c qp->s_ack_rdma_sge.sge = e->rdma_sge; sge 249 drivers/infiniband/hw/hfi1/rc.c qp->s_ack_rdma_sge.sge = e->rdma_sge; sge 273 drivers/infiniband/hw/hfi1/rc.c ps->s_txreq->mr = qp->s_ack_rdma_sge.sge.mr; sge 276 drivers/infiniband/hw/hfi1/rc.c len = qp->s_ack_rdma_sge.sge.sge_length; sge 833 drivers/infiniband/hw/hfi1/rc.c qp->s_sge.sge = wqe->sg_list[0]; sge 918 drivers/infiniband/hw/hfi1/rc.c qp->s_sge.sge = wqe->sg_list[0]; sge 3029 drivers/infiniband/hw/hfi1/rc.c ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, sge 3036 drivers/infiniband/hw/hfi1/rc.c qp->r_sge.sge.mr = NULL; sge 3037 drivers/infiniband/hw/hfi1/rc.c qp->r_sge.sge.vaddr = NULL; sge 3038 drivers/infiniband/hw/hfi1/rc.c qp->r_sge.sge.length = 0; sge 3039 drivers/infiniband/hw/hfi1/rc.c qp->r_sge.sge.sge_length = 0; sge 3160 drivers/infiniband/hw/hfi1/rc.c if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), sge 3165 drivers/infiniband/hw/hfi1/rc.c maddr = (atomic64_t *)qp->r_sge.sge.vaddr; sge 3169 drivers/infiniband/hw/hfi1/rc.c (u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr, sge 3172 drivers/infiniband/hw/hfi1/rc.c rvt_put_mr(qp->r_sge.sge.mr); sge 1085 drivers/infiniband/hw/hfi1/tid_rdma.c struct rvt_sge *sge = &ss->sge; sge 1091 drivers/infiniband/hw/hfi1/tid_rdma.c pages[i++] = virt_to_page(sge->vaddr); sge 1093 drivers/infiniband/hw/hfi1/tid_rdma.c sge->vaddr += len; sge 1094 drivers/infiniband/hw/hfi1/tid_rdma.c sge->length -= len; sge 1095 drivers/infiniband/hw/hfi1/tid_rdma.c sge->sge_length -= len; sge 1096 drivers/infiniband/hw/hfi1/tid_rdma.c if (!sge->sge_length) { sge 1098 drivers/infiniband/hw/hfi1/tid_rdma.c *sge = ss->sg_list[req->isge - 1]; sge 1099 drivers/infiniband/hw/hfi1/tid_rdma.c } else if (sge->length == 0 && sge->mr->lkey) { sge 1100 drivers/infiniband/hw/hfi1/tid_rdma.c if (++sge->n >= RVT_SEGSZ) { sge 1101 drivers/infiniband/hw/hfi1/tid_rdma.c ++sge->m; sge 1102 drivers/infiniband/hw/hfi1/tid_rdma.c sge->n = 0; sge 1104 drivers/infiniband/hw/hfi1/tid_rdma.c sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; sge 1105 drivers/infiniband/hw/hfi1/tid_rdma.c sge->length = sge->mr->map[sge->m]->segs[sge->n].length; sge 1720 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.vaddr = req_addr; sge 1721 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.sge_length = req_len; sge 1722 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.length = wpriv->ss.sge.sge_length; sge 1727 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.mr = NULL; sge 1728 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.m = 0; sge 1729 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.sge.n = 0; sge 1732 drivers/infiniband/hw/hfi1/tid_rdma.c wpriv->ss.total_len = wpriv->ss.sge.sge_length; sge 3256 drivers/infiniband/hw/hfi1/tid_rdma.c struct rvt_sge *sge, int num_sge) sge 3260 drivers/infiniband/hw/hfi1/tid_rdma.c for (i = 0; i < num_sge; i++, sge++) { sge 3261 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_sge_check_align(qp, i, sge); sge 3262 drivers/infiniband/hw/hfi1/tid_rdma.c if ((u64)sge->vaddr & ~PAGE_MASK || sge 3263 drivers/infiniband/hw/hfi1/tid_rdma.c sge->sge_length & ~PAGE_MASK) sge 3755 drivers/infiniband/hw/hfi1/tid_rdma.c req->ss.sge = e->rdma_sge; sge 3881 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.vaddr = resp_addr; sge 3882 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.sge_length = resp_len; sge 3883 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.length = epriv->ss.sge.sge_length; sge 3888 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.mr = NULL; sge 3889 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.m = 0; sge 3890 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.sge.n = 0; sge 3893 drivers/infiniband/hw/hfi1/tid_rdma.c epriv->ss.total_len = epriv->ss.sge.sge_length; sge 4320 drivers/infiniband/hw/hfi1/tid_rdma.c ss.sge = e->rdma_sge; sge 5062 drivers/infiniband/hw/hfi1/tid_rdma.c priv->tid_ss.sge = wqe->sg_list[0]; sge 1324 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge), sge 1325 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, sge), sge 1337 drivers/infiniband/hw/hfi1/trace_tid.h __entry->vaddr = (u64)sge->vaddr; sge 1338 drivers/infiniband/hw/hfi1/trace_tid.h __entry->sge_length = sge->sge_length; sge 1352 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge), sge 1353 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, sge) sge 153 drivers/infiniband/hw/hfi1/uc.c qp->s_sge.sge = wqe->sg_list[0]; sge 500 drivers/infiniband/hw/hfi1/uc.c ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, sge 507 drivers/infiniband/hw/hfi1/uc.c qp->r_sge.sge.mr = NULL; sge 508 drivers/infiniband/hw/hfi1/uc.c qp->r_sge.sge.vaddr = NULL; sge 509 drivers/infiniband/hw/hfi1/uc.c qp->r_sge.sge.length = 0; sge 510 drivers/infiniband/hw/hfi1/uc.c qp->r_sge.sge.sge_length = 0; sge 82 drivers/infiniband/hw/hfi1/ud.c struct rvt_sge *sge; sge 221 drivers/infiniband/hw/hfi1/ud.c ssge.sge = *swqe->sg_list; sge 223 drivers/infiniband/hw/hfi1/ud.c sge = &ssge.sge; sge 225 drivers/infiniband/hw/hfi1/ud.c u32 len = rvt_get_sge_length(sge, length); sge 228 drivers/infiniband/hw/hfi1/ud.c rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, len, true, false); sge 555 drivers/infiniband/hw/hfi1/ud.c qp->s_sge.sge = wqe->sg_list[0]; sge 702 drivers/infiniband/hw/hfi1/verbs.c struct rvt_sge sge = ss->sge; sge 708 drivers/infiniband/hw/hfi1/verbs.c len = rvt_get_sge_length(&ss->sge, length); sge 713 drivers/infiniband/hw/hfi1/verbs.c ss->sge.vaddr, sge 723 drivers/infiniband/hw/hfi1/verbs.c ss->sge = sge; sge 1081 drivers/infiniband/hw/hfi1/verbs.c void *addr = ss->sge.vaddr; sge 1082 drivers/infiniband/hw/hfi1/verbs.c u32 slen = rvt_get_sge_length(&ss->sge, len); sge 691 drivers/infiniband/hw/hns/hns_roce_device.h struct hns_roce_sge sge; sge 135 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dseg = get_send_extend_sge(qp, (*sge_ind) & (qp->sge.sge_cnt - 1)); sge 146 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (*sge_ind) & (qp->sge.sge_cnt - 1)); sge 206 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (*sge_ind) & (qp->sge.sge_cnt - 1)); sge 385 drivers/infiniband/hw/hns/hns_roce_hw_v2.c sge_idx & (qp->sge.sge_cnt - 1)); sge 3227 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ilog2((unsigned int)hr_qp->sge.sge_cnt)); sge 3234 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ilog2((unsigned int)hr_qp->sge.sge_cnt) : 0); sge 3932 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (hr_qp->sge.offset) { sge 3935 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hr_qp->sge.offset / page_size, sge 379 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt * sge 383 drivers/infiniband/hw/hns/hns_roce_qp.c if (hr_qp->sge.sge_cnt > hr_dev->caps.max_extend_sg) { sge 386 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt); sge 391 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift = 4; sge 392 drivers/infiniband/hw/hns/hns_roce_qp.c ex_sge_num = hr_qp->sge.sge_cnt; sge 406 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt = ex_sge_num ? sge 407 drivers/infiniband/hw/hns/hns_roce_qp.c max(page_size / (1 << hr_qp->sge.sge_shift), ex_sge_num) : 0; sge 410 drivers/infiniband/hw/hns/hns_roce_qp.c HNS_ROCE_ALOGN_UP((hr_qp->sge.sge_cnt << sge 411 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift), page_size) + sge 417 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.offset = HNS_ROCE_ALOGN_UP( sge 421 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->rq.offset = hr_qp->sge.offset + sge 422 drivers/infiniband/hw/hns/hns_roce_qp.c HNS_ROCE_ALOGN_UP((hr_qp->sge.sge_cnt << sge 423 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift), sge 450 drivers/infiniband/hw/hns/hns_roce_qp.c if (hr_qp->sge.sge_cnt > 0) sge 457 drivers/infiniband/hw/hns/hns_roce_qp.c buf_size = hr_qp->sge.offset - hr_qp->sq.offset; sge 472 drivers/infiniband/hw/hns/hns_roce_qp.c buf_size = hr_qp->rq.offset - hr_qp->sge.offset; sge 477 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.offset / page_size, sge 527 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt * sge 529 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift = 4; sge 534 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt * sge 536 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift = 4; sge 540 drivers/infiniband/hw/hns/hns_roce_qp.c if (hr_qp->sge.sge_cnt > hr_dev->caps.max_extend_sg) { sge 542 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt); sge 599 drivers/infiniband/hw/hns/hns_roce_qp.c if (hr_dev->caps.max_sq_sg > 2 && hr_qp->sge.sge_cnt) { sge 600 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_cnt = max(page_size/(1 << hr_qp->sge.sge_shift), sge 601 drivers/infiniband/hw/hns/hns_roce_qp.c (u32)hr_qp->sge.sge_cnt); sge 602 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.offset = size; sge 603 drivers/infiniband/hw/hns/hns_roce_qp.c size += HNS_ROCE_ALOGN_UP(hr_qp->sge.sge_cnt << sge 604 drivers/infiniband/hw/hns/hns_roce_qp.c hr_qp->sge.sge_shift, page_size); sge 1283 drivers/infiniband/hw/hns/hns_roce_qp.c return hns_roce_buf_offset(&hr_qp->hr_buf, hr_qp->sge.offset + sge 1284 drivers/infiniband/hw/hns/hns_roce_qp.c (n << hr_qp->sge.sge_shift)); sge 205 drivers/infiniband/hw/i40iw/i40iw_uk.c static void i40iw_set_fragment(u64 *wqe, u32 offset, struct i40iw_sge *sge) sge 207 drivers/infiniband/hw/i40iw/i40iw_uk.c if (sge) { sge 208 drivers/infiniband/hw/i40iw/i40iw_uk.c set_64bit_val(wqe, offset, LS_64(sge->tag_off, I40IWQPSQ_FRAG_TO)); sge 210 drivers/infiniband/hw/i40iw/i40iw_uk.c (LS_64(sge->len, I40IWQPSQ_FRAG_LEN) | sge 211 drivers/infiniband/hw/i40iw/i40iw_uk.c LS_64(sge->stag, I40IWQPSQ_FRAG_STAG))); sge 907 drivers/infiniband/hw/i40iw/i40iw_uk.c void i40iw_get_wqe_shift(u32 sge, u32 inline_data, u8 *shift) sge 910 drivers/infiniband/hw/i40iw/i40iw_uk.c if (sge > 1 || inline_data > 16) sge 911 drivers/infiniband/hw/i40iw/i40iw_uk.c *shift = (sge < 4 && inline_data <= 48) ? 1 : 2; sge 427 drivers/infiniband/hw/i40iw/i40iw_user.h void i40iw_get_wqe_shift(u32 sge, u32 inline_data, u8 *shift); sge 1302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); sge 1303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); sge 1304 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size; sge 1350 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); sge 1351 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); sge 1352 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].len = dma.size; sge 2887 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge; sge 288 drivers/infiniband/hw/ocrdma/ocrdma_sli.h struct ocrdma_mqe_sge sge[19]; sge 1906 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge, int num_sge, sge 1912 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge[i].lrkey = sg_list[i].lkey; sge 1913 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge[i].addr_lo = sg_list[i].addr; sge 1914 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge[i].addr_hi = upper_32_bits(sg_list[i].addr); sge 1915 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge[i].len = sg_list[i].length; sge 1919 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c memset(sge, 0, sizeof(*sge)); sge 1934 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge, sge 1948 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c dpp_addr = (char *)sge; sge 1961 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); sge 1976 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge; sge 1981 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge = (struct ocrdma_sge *)(hdr + 2); sge 1984 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge = (struct ocrdma_sge *)(hdr + 1); sge 1987 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size); sge 1996 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge = ext_rw + 1; sge 1999 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size); sge 2013 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge = ext_rw + 1; sge 2017 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); sge 2217 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *sge; sge 2219 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wqe_size = (wr->num_sge * sizeof(*sge)) + sizeof(*rqe); sge 2221 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wqe_size = sizeof(*sge) + sizeof(*rqe); sge 2229 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c sge = (struct ocrdma_sge *)(rqe + 1); sge 2230 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(rqe, sge, wr->num_sge, wr->sg_list); sge 178 drivers/infiniband/hw/qedr/qedr_hsi_rdma.h struct rdma_srq_sge sge; sge 2992 drivers/infiniband/hw/qedr/verbs.c #define RQ_SGE_SET(sge, vaddr, vlength, vflags) \ sge 2994 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(sge->addr, vaddr); \ sge 2995 drivers/infiniband/hw/qedr/verbs.c (sge)->length = cpu_to_le32(vlength); \ sge 2996 drivers/infiniband/hw/qedr/verbs.c (sge)->flags = cpu_to_le32(vflags); \ sge 3005 drivers/infiniband/hw/qedr/verbs.c #define SRQ_SGE_SET(sge, vaddr, vlength, vlkey) \ sge 3007 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(sge->addr, vaddr); \ sge 3008 drivers/infiniband/hw/qedr/verbs.c (sge)->length = cpu_to_le32(vlength); \ sge 3009 drivers/infiniband/hw/qedr/verbs.c (sge)->l_key = cpu_to_le32(vlkey); \ sge 3019 drivers/infiniband/hw/qedr/verbs.c struct rdma_sq_sge *sge = qed_chain_produce(&qp->sq.pbl); sge 3021 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(sge->addr, wr->sg_list[i].addr); sge 3022 drivers/infiniband/hw/qedr/verbs.c sge->l_key = cpu_to_le32(wr->sg_list[i].lkey); sge 3023 drivers/infiniband/hw/qedr/verbs.c sge->length = cpu_to_le32(wr->sg_list[i].length); sge 122 drivers/infiniband/hw/qib/qib_rc.c qp->s_ack_rdma_sge.sge = e->rdma_sge; sge 155 drivers/infiniband/hw/qib/qib_rc.c qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; sge 158 drivers/infiniband/hw/qib/qib_rc.c len = qp->s_ack_rdma_sge.sge.sge_length; sge 447 drivers/infiniband/hw/qib/qib_rc.c qp->s_sge.sge = wqe->sg_list[0]; sge 1908 drivers/infiniband/hw/qib/qib_rc.c ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr, sge 1915 drivers/infiniband/hw/qib/qib_rc.c qp->r_sge.sge.mr = NULL; sge 1916 drivers/infiniband/hw/qib/qib_rc.c qp->r_sge.sge.vaddr = NULL; sge 1917 drivers/infiniband/hw/qib/qib_rc.c qp->r_sge.sge.length = 0; sge 1918 drivers/infiniband/hw/qib/qib_rc.c qp->r_sge.sge.sge_length = 0; sge 2035 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), sge 2040 drivers/infiniband/hw/qib/qib_rc.c maddr = (atomic64_t *) qp->r_sge.sge.vaddr; sge 2044 drivers/infiniband/hw/qib/qib_rc.c (u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr, sge 2047 drivers/infiniband/hw/qib/qib_rc.c rvt_put_mr(qp->r_sge.sge.mr); sge 516 drivers/infiniband/hw/qib/qib_sdma.c struct rvt_sge *sge; sge 565 drivers/infiniband/hw/qib/qib_sdma.c sge = &ss->sge; sge 568 drivers/infiniband/hw/qib/qib_sdma.c u32 len = rvt_get_sge_length(sge, dwords << 2); sge 571 drivers/infiniband/hw/qib/qib_sdma.c addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr, sge 98 drivers/infiniband/hw/qib/qib_uc.c qp->s_sge.sge = wqe->sg_list[0]; sge 425 drivers/infiniband/hw/qib/qib_uc.c ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, sge 432 drivers/infiniband/hw/qib/qib_uc.c qp->r_sge.sge.mr = NULL; sge 433 drivers/infiniband/hw/qib/qib_uc.c qp->r_sge.sge.vaddr = NULL; sge 434 drivers/infiniband/hw/qib/qib_uc.c qp->r_sge.sge.length = 0; sge 435 drivers/infiniband/hw/qib/qib_uc.c qp->r_sge.sge.sge_length = 0; sge 61 drivers/infiniband/hw/qib/qib_ud.c struct rvt_sge *sge; sge 172 drivers/infiniband/hw/qib/qib_ud.c ssge.sge = *swqe->sg_list; sge 174 drivers/infiniband/hw/qib/qib_ud.c sge = &ssge.sge; sge 176 drivers/infiniband/hw/qib/qib_ud.c u32 len = rvt_get_sge_length(sge, length); sge 178 drivers/infiniband/hw/qib/qib_ud.c rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, len, true, false); sge 179 drivers/infiniband/hw/qib/qib_ud.c sge->vaddr += len; sge 180 drivers/infiniband/hw/qib/qib_ud.c sge->length -= len; sge 181 drivers/infiniband/hw/qib/qib_ud.c sge->sge_length -= len; sge 182 drivers/infiniband/hw/qib/qib_ud.c if (sge->sge_length == 0) { sge 184 drivers/infiniband/hw/qib/qib_ud.c *sge = *ssge.sg_list++; sge 185 drivers/infiniband/hw/qib/qib_ud.c } else if (sge->length == 0 && sge->mr->lkey) { sge 186 drivers/infiniband/hw/qib/qib_ud.c if (++sge->n >= RVT_SEGSZ) { sge 187 drivers/infiniband/hw/qib/qib_ud.c if (++sge->m >= sge->mr->mapsz) sge 189 drivers/infiniband/hw/qib/qib_ud.c sge->n = 0; sge 191 drivers/infiniband/hw/qib/qib_ud.c sge->vaddr = sge 192 drivers/infiniband/hw/qib/qib_ud.c sge->mr->map[sge->m]->segs[sge->n].vaddr; sge 193 drivers/infiniband/hw/qib/qib_ud.c sge->length = sge 194 drivers/infiniband/hw/qib/qib_ud.c sge->mr->map[sge->m]->segs[sge->n].length; sge 317 drivers/infiniband/hw/qib/qib_ud.c qp->s_sge.sge = wqe->sg_list[0]; sge 142 drivers/infiniband/hw/qib/qib_verbs.c struct rvt_sge sge = ss->sge; sge 147 drivers/infiniband/hw/qib/qib_verbs.c u32 len = rvt_get_sge_length(&sge, length); sge 149 drivers/infiniband/hw/qib/qib_verbs.c if (((long) sge.vaddr & (sizeof(u32) - 1)) || sge 155 drivers/infiniband/hw/qib/qib_verbs.c sge.vaddr += len; sge 156 drivers/infiniband/hw/qib/qib_verbs.c sge.length -= len; sge 157 drivers/infiniband/hw/qib/qib_verbs.c sge.sge_length -= len; sge 158 drivers/infiniband/hw/qib/qib_verbs.c if (sge.sge_length == 0) { sge 160 drivers/infiniband/hw/qib/qib_verbs.c sge = *sg_list++; sge 161 drivers/infiniband/hw/qib/qib_verbs.c } else if (sge.length == 0 && sge.mr->lkey) { sge 162 drivers/infiniband/hw/qib/qib_verbs.c if (++sge.n >= RVT_SEGSZ) { sge 163 drivers/infiniband/hw/qib/qib_verbs.c if (++sge.m >= sge.mr->mapsz) sge 165 drivers/infiniband/hw/qib/qib_verbs.c sge.n = 0; sge 167 drivers/infiniband/hw/qib/qib_verbs.c sge.vaddr = sge 168 drivers/infiniband/hw/qib/qib_verbs.c sge.mr->map[sge.m]->segs[sge.n].vaddr; sge 169 drivers/infiniband/hw/qib/qib_verbs.c sge.length = sge 170 drivers/infiniband/hw/qib/qib_verbs.c sge.mr->map[sge.m]->segs[sge.n].length; sge 182 drivers/infiniband/hw/qib/qib_verbs.c struct rvt_sge *sge = &ss->sge; sge 185 drivers/infiniband/hw/qib/qib_verbs.c u32 len = rvt_get_sge_length(sge, length); sge 187 drivers/infiniband/hw/qib/qib_verbs.c memcpy(data, sge->vaddr, len); sge 188 drivers/infiniband/hw/qib/qib_verbs.c sge->vaddr += len; sge 189 drivers/infiniband/hw/qib/qib_verbs.c sge->length -= len; sge 190 drivers/infiniband/hw/qib/qib_verbs.c sge->sge_length -= len; sge 191 drivers/infiniband/hw/qib/qib_verbs.c if (sge->sge_length == 0) { sge 193 drivers/infiniband/hw/qib/qib_verbs.c *sge = *ss->sg_list++; sge 194 drivers/infiniband/hw/qib/qib_verbs.c } else if (sge->length == 0 && sge->mr->lkey) { sge 195 drivers/infiniband/hw/qib/qib_verbs.c if (++sge->n >= RVT_SEGSZ) { sge 196 drivers/infiniband/hw/qib/qib_verbs.c if (++sge->m >= sge->mr->mapsz) sge 198 drivers/infiniband/hw/qib/qib_verbs.c sge->n = 0; sge 200 drivers/infiniband/hw/qib/qib_verbs.c sge->vaddr = sge 201 drivers/infiniband/hw/qib/qib_verbs.c sge->mr->map[sge->m]->segs[sge->n].vaddr; sge 202 drivers/infiniband/hw/qib/qib_verbs.c sge->length = sge 203 drivers/infiniband/hw/qib/qib_verbs.c sge->mr->map[sge->m]->segs[sge->n].length; sge 437 drivers/infiniband/hw/qib/qib_verbs.c u32 len = rvt_get_sge_length(&ss->sge, length); sge 441 drivers/infiniband/hw/qib/qib_verbs.c off = (unsigned long)ss->sge.vaddr & (sizeof(u32) - 1); sge 443 drivers/infiniband/hw/qib/qib_verbs.c u32 *addr = (u32 *)((unsigned long)ss->sge.vaddr & sge 474 drivers/infiniband/hw/qib/qib_verbs.c u32 *addr = (u32 *) ss->sge.vaddr; sge 527 drivers/infiniband/hw/qib/qib_verbs.c qib_pio_copy(piobuf, ss->sge.vaddr, w - 1); sge 529 drivers/infiniband/hw/qib/qib_verbs.c last = ((u32 *) ss->sge.vaddr)[w - 1]; sge 534 drivers/infiniband/hw/qib/qib_verbs.c qib_pio_copy(piobuf, ss->sge.vaddr, w); sge 539 drivers/infiniband/hw/qib/qib_verbs.c u32 v = ((u32 *) ss->sge.vaddr)[w]; sge 962 drivers/infiniband/hw/qib/qib_verbs.c if (likely(ss->num_sge == 1 && len <= ss->sge.length && sge 963 drivers/infiniband/hw/qib/qib_verbs.c !((unsigned long)ss->sge.vaddr & (sizeof(u32) - 1)))) { sge 964 drivers/infiniband/hw/qib/qib_verbs.c u32 *addr = (u32 *) ss->sge.vaddr; sge 625 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct pvrdma_sge *sge; sge 790 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge = (struct pvrdma_sge *)(wqe_hdr + 1); sge 793 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->addr = wr->sg_list[i].addr; sge 794 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->length = wr->sg_list[i].length; sge 795 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->lkey = wr->sg_list[i].lkey; sge 796 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge++; sge 835 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct pvrdma_sge *sge; sge 882 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge = (struct pvrdma_sge *)(wqe_hdr + 1); sge 884 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->addr = wr->sg_list[i].addr; sge 885 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->length = wr->sg_list[i].length; sge 886 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->lkey = wr->sg_list[i].lkey; sge 887 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge++; sge 536 drivers/infiniband/sw/rdmavt/mr.c rval = rvt_mr_has_lkey(ss->sge.mr, lkey); sge 880 drivers/infiniband/sw/rdmavt/mr.c struct ib_sge *sge) sge 882 drivers/infiniband/sw/rdmavt/mr.c if (last_sge && sge->lkey == last_sge->mr->lkey && sge 883 drivers/infiniband/sw/rdmavt/mr.c ((uint64_t)(last_sge->vaddr + last_sge->length) == sge->addr)) { sge 884 drivers/infiniband/sw/rdmavt/mr.c if (sge->lkey) { sge 885 drivers/infiniband/sw/rdmavt/mr.c if (unlikely((sge->addr - last_sge->mr->user_base + sge 886 drivers/infiniband/sw/rdmavt/mr.c sge->length > last_sge->mr->length))) sge 889 drivers/infiniband/sw/rdmavt/mr.c last_sge->length += sge->length; sge 891 drivers/infiniband/sw/rdmavt/mr.c last_sge->sge_length += sge->length; sge 892 drivers/infiniband/sw/rdmavt/mr.c trace_rvt_sge_adjacent(last_sge, sge); sge 916 drivers/infiniband/sw/rdmavt/mr.c struct ib_sge *sge, int acc) sge 926 drivers/infiniband/sw/rdmavt/mr.c if (sge->lkey == 0) { sge 931 drivers/infiniband/sw/rdmavt/mr.c if (rvt_sge_adjacent(last_sge, sge)) sge 941 drivers/infiniband/sw/rdmavt/mr.c isge->vaddr = (void *)sge->addr; sge 942 drivers/infiniband/sw/rdmavt/mr.c isge->length = sge->length; sge 943 drivers/infiniband/sw/rdmavt/mr.c isge->sge_length = sge->length; sge 948 drivers/infiniband/sw/rdmavt/mr.c if (rvt_sge_adjacent(last_sge, sge)) sge 951 drivers/infiniband/sw/rdmavt/mr.c mr = rcu_dereference(rkt->table[sge->lkey >> rkt->shift]); sge 959 drivers/infiniband/sw/rdmavt/mr.c mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) sge 962 drivers/infiniband/sw/rdmavt/mr.c off = sge->addr - mr->user_base; sge 963 drivers/infiniband/sw/rdmavt/mr.c if (unlikely(sge->addr < mr->user_base || sge 964 drivers/infiniband/sw/rdmavt/mr.c off + sge->length > mr->length || sge 997 drivers/infiniband/sw/rdmavt/mr.c isge->sge_length = sge->length; sge 1001 drivers/infiniband/sw/rdmavt/mr.c trace_rvt_sge_new(isge, sge); sge 1024 drivers/infiniband/sw/rdmavt/mr.c int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge, sge 1050 drivers/infiniband/sw/rdmavt/mr.c sge->mr = mr; sge 1051 drivers/infiniband/sw/rdmavt/mr.c sge->vaddr = (void *)vaddr; sge 1052 drivers/infiniband/sw/rdmavt/mr.c sge->length = len; sge 1053 drivers/infiniband/sw/rdmavt/mr.c sge->sge_length = len; sge 1054 drivers/infiniband/sw/rdmavt/mr.c sge->m = 0; sge 1055 drivers/infiniband/sw/rdmavt/mr.c sge->n = 0; sge 1101 drivers/infiniband/sw/rdmavt/mr.c sge->mr = mr; sge 1102 drivers/infiniband/sw/rdmavt/mr.c sge->vaddr = mr->map[m]->segs[n].vaddr + off; sge 1103 drivers/infiniband/sw/rdmavt/mr.c sge->length = mr->map[m]->segs[n].length - off; sge 1104 drivers/infiniband/sw/rdmavt/mr.c sge->sge_length = len; sge 1105 drivers/infiniband/sw/rdmavt/mr.c sge->m = m; sge 1106 drivers/infiniband/sw/rdmavt/mr.c sge->n = n; sge 665 drivers/infiniband/sw/rdmavt/qp.c struct rvt_sge *sge = &wqe->sg_list[i]; sge 667 drivers/infiniband/sw/rdmavt/qp.c if (rvt_mr_has_lkey(sge->mr, lkey)) sge 2165 drivers/infiniband/sw/rdmavt/qp.c struct rvt_sge *sge = &wqe->sg_list[--j]; sge 2167 drivers/infiniband/sw/rdmavt/qp.c rvt_put_mr(sge->mr); sge 2291 drivers/infiniband/sw/rdmavt/qp.c static struct ib_sge *rvt_cast_sge(struct rvt_wqe_sge *sge) sge 2299 drivers/infiniband/sw/rdmavt/qp.c return (struct ib_sge *)sge; sge 2324 drivers/infiniband/sw/rdmavt/qp.c ret = rvt_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, sge 2338 drivers/infiniband/sw/rdmavt/qp.c struct rvt_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; sge 2340 drivers/infiniband/sw/rdmavt/qp.c rvt_put_mr(sge->mr); sge 2878 drivers/infiniband/sw/rdmavt/qp.c struct rvt_sge *sge = &ss->sge; sge 2895 drivers/infiniband/sw/rdmavt/qp.c wss_insert(wss, sge->vaddr); sge 2897 drivers/infiniband/sw/rdmavt/qp.c wss_insert(wss, (sge->vaddr + PAGE_SIZE)); sge 2916 drivers/infiniband/sw/rdmavt/qp.c u32 len = rvt_get_sge_length(sge, length); sge 2922 drivers/infiniband/sw/rdmavt/qp.c ((u8 *)sge->vaddr)[i] = ((u8 *)data)[i]; sge 2924 drivers/infiniband/sw/rdmavt/qp.c cacheless_memcpy(sge->vaddr, data, len); sge 2926 drivers/infiniband/sw/rdmavt/qp.c memcpy(sge->vaddr, data, len); sge 2970 drivers/infiniband/sw/rdmavt/qp.c struct rvt_sge *sge; sge 3041 drivers/infiniband/sw/rdmavt/qp.c sqp->s_sge.sge = wqe->sg_list[0]; sge 3105 drivers/infiniband/sw/rdmavt/qp.c if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, wqe->length, sge 3118 drivers/infiniband/sw/rdmavt/qp.c if (unlikely(!rvt_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, sge 3126 drivers/infiniband/sw/rdmavt/qp.c qp->r_sge.sge = wqe->sg_list[0]; sge 3136 drivers/infiniband/sw/rdmavt/qp.c if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64), sge 3142 drivers/infiniband/sw/rdmavt/qp.c maddr = (atomic64_t *)qp->r_sge.sge.vaddr; sge 3144 drivers/infiniband/sw/rdmavt/qp.c *(u64 *)sqp->s_sge.sge.vaddr = sge 3147 drivers/infiniband/sw/rdmavt/qp.c (u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr, sge 3149 drivers/infiniband/sw/rdmavt/qp.c rvt_put_mr(qp->r_sge.sge.mr); sge 3158 drivers/infiniband/sw/rdmavt/qp.c sge = &sqp->s_sge.sge; sge 3160 drivers/infiniband/sw/rdmavt/qp.c u32 len = rvt_get_sge_length(sge, sqp->s_len); sge 3163 drivers/infiniband/sw/rdmavt/qp.c rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, sge 201 drivers/infiniband/sw/rdmavt/rc.c ss->sge = wqe->sg_list[0]; sge 124 drivers/infiniband/sw/rdmavt/trace_mr.h TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge), sge 125 drivers/infiniband/sw/rdmavt/trace_mr.h TP_ARGS(sge, isge), sge 127 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ENTRY(ib_to_rvt(sge->mr->pd->device)) sge 129 drivers/infiniband/sw/rdmavt/trace_mr.h __field(struct rvt_sge *, sge) sge 142 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ASSIGN(ib_to_rvt(sge->mr->pd->device)); sge 143 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->mr = sge->mr; sge 144 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->sge = sge; sge 146 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->vaddr = sge->vaddr; sge 148 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->lkey = sge->mr->lkey; sge 149 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->sge_length = sge->sge_length; sge 150 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->length = sge->length; sge 152 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->m = sge->m; sge 153 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->n = sge->m; sge 154 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->user = ibpd_to_rvtpd(sge->mr->pd)->user; sge 160 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->sge, sge 176 drivers/infiniband/sw/rdmavt/trace_mr.h TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge), sge 177 drivers/infiniband/sw/rdmavt/trace_mr.h TP_ARGS(sge, isge)); sge 181 drivers/infiniband/sw/rdmavt/trace_mr.h TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge), sge 182 drivers/infiniband/sw/rdmavt/trace_mr.h TP_ARGS(sge, isge)); sge 445 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_sge *sge = &dma->sge[dma->cur_sge]; sge 460 drivers/infiniband/sw/rxe/rxe_mr.c if (sge->length && (offset < sge->length)) { sge 461 drivers/infiniband/sw/rxe/rxe_mr.c mem = lookup_mem(pd, access, sge->lkey, lookup_local); sge 471 drivers/infiniband/sw/rxe/rxe_mr.c if (offset >= sge->length) { sge 476 drivers/infiniband/sw/rxe/rxe_mr.c sge++; sge 485 drivers/infiniband/sw/rxe/rxe_mr.c if (sge->length) { sge 486 drivers/infiniband/sw/rxe/rxe_mr.c mem = lookup_mem(pd, access, sge->lkey, sge 497 drivers/infiniband/sw/rxe/rxe_mr.c if (bytes > sge->length - offset) sge 498 drivers/infiniband/sw/rxe/rxe_mr.c bytes = sge->length - offset; sge 501 drivers/infiniband/sw/rxe/rxe_mr.c iova = sge->addr + offset; sge 531 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_sge *sge = &dma->sge[dma->cur_sge]; sge 538 drivers/infiniband/sw/rxe/rxe_mr.c if (offset >= sge->length) { sge 539 drivers/infiniband/sw/rxe/rxe_mr.c sge++; sge 548 drivers/infiniband/sw/rxe/rxe_mr.c if (bytes > sge->length - offset) sge 549 drivers/infiniband/sw/rxe/rxe_mr.c bytes = sge->length - offset; sge 271 drivers/infiniband/sw/rxe/rxe_verbs.c memcpy(recv_wqe->dma.sge, ibwr->sg_list, sge 591 drivers/infiniband/sw/rxe/rxe_verbs.c struct ib_sge *sge; sge 605 drivers/infiniband/sw/rxe/rxe_verbs.c sge = ibwr->sg_list; sge 606 drivers/infiniband/sw/rxe/rxe_verbs.c for (i = 0; i < num_sge; i++, sge++) { sge 607 drivers/infiniband/sw/rxe/rxe_verbs.c memcpy(p, (void *)(uintptr_t)sge->addr, sge 608 drivers/infiniband/sw/rxe/rxe_verbs.c sge->length); sge 610 drivers/infiniband/sw/rxe/rxe_verbs.c p += sge->length; sge 617 drivers/infiniband/sw/rxe/rxe_verbs.c memcpy(wqe->dma.sge, ibwr->sg_list, sge 222 drivers/infiniband/sw/rxe/rxe_verbs.h struct ib_sge sge[RXE_MAX_SGE]; sge 222 drivers/infiniband/sw/siw/siw_mem.c int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[], sge 229 drivers/infiniband/sw/siw/siw_mem.c if (len + off > sge->length) { sge 234 drivers/infiniband/sw/siw/siw_mem.c new = siw_mem_id2obj(sdev, sge->lkey >> 8); sge 236 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey); sge 243 drivers/infiniband/sw/siw/siw_mem.c if (unlikely((*mem)->stag != sge->lkey)) { sge 244 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_mem((*mem), "STag mismatch: 0x%08x\n", sge->lkey); sge 248 drivers/infiniband/sw/siw/siw_mem.c rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len); sge 18 drivers/infiniband/sw/siw/siw_mem.h int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, sge 277 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].length = 0; sge 278 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].laddr = 0; sge 279 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].lkey = 0; sge 479 drivers/infiniband/sw/siw/siw_qp.c rreq->ddp_msn = htonl(wqe->sqe.sge[0].length); sge 484 drivers/infiniband/sw/siw/siw_qp.c rreq->read_size = htonl(wqe->sqe.sge[0].length); sge 485 drivers/infiniband/sw/siw/siw_qp.c rreq->source_stag = htonl(wqe->sqe.sge[0].lkey); sge 487 drivers/infiniband/sw/siw/siw_qp.c cpu_to_be64(wqe->sqe.sge[0].laddr); sge 872 drivers/infiniband/sw/siw/siw_qp.c rreq->sge[0].laddr = sqe->sge[0].laddr; sge 873 drivers/infiniband/sw/siw/siw_qp.c rreq->sge[0].length = sqe->sge[0].length; sge 874 drivers/infiniband/sw/siw/siw_qp.c rreq->sge[0].lkey = sqe->sge[0].lkey; sge 875 drivers/infiniband/sw/siw/siw_qp.c rreq->sge[1].lkey = sqe->sge[1].lkey; sge 913 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].length = irqe->sge[0].length; sge 914 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].laddr = irqe->sge[0].laddr; sge 915 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].lkey = irqe->sge[0].lkey; sge 923 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[1].length = irqe->sge[1].length; sge 955 drivers/infiniband/sw/siw/siw_qp.c if (wqe->sqe.sge[0].length > SIW_MAX_INLINE) { sge 959 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].laddr = (uintptr_t)&wqe->sqe.sge[1]; sge 960 drivers/infiniband/sw/siw/siw_qp.c wqe->sqe.sge[0].lkey = 0; sge 175 drivers/infiniband/sw/siw/siw_qp_rx.c srx->ddp_stag = wqe->sqe.sge[0].lkey; sge 176 drivers/infiniband/sw/siw/siw_qp_rx.c srx->ddp_to = wqe->sqe.sge[0].laddr; sge 366 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->rqe.sge[i].laddr = rqe->sge[i].laddr; sge 367 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->rqe.sge[i].lkey = rqe->sge[i].lkey; sge 368 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->rqe.sge[i].length = rqe->sge[i].length; sge 369 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->bytes += wqe->rqe.sge[i].length; sge 456 drivers/infiniband/sw/siw/siw_qp_rx.c struct siw_sge *sge; sge 459 drivers/infiniband/sw/siw/siw_qp_rx.c sge = &wqe->rqe.sge[frx->sge_idx]; sge 461 drivers/infiniband/sw/siw/siw_qp_rx.c if (!sge->length) { sge 468 drivers/infiniband/sw/siw/siw_qp_rx.c sge_bytes = min(data_bytes, sge->length - frx->sge_off); sge 476 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE, sge 489 drivers/infiniband/sw/siw/siw_qp_rx.c (void *)(uintptr_t)(sge->laddr + frx->sge_off), sge 493 drivers/infiniband/sw/siw/siw_qp_rx.c sge->laddr + frx->sge_off, sge_bytes); sge 496 drivers/infiniband/sw/siw/siw_qp_rx.c sge->laddr + frx->sge_off, sge_bytes); sge 508 drivers/infiniband/sw/siw/siw_qp_rx.c if (frx->sge_off == sge->length) { sge 698 drivers/infiniband/sw/siw/siw_qp_rx.c resp->sge[0].length = length; sge 699 drivers/infiniband/sw/siw/siw_qp_rx.c resp->sge[0].laddr = laddr; sge 700 drivers/infiniband/sw/siw/siw_qp_rx.c resp->sge[0].lkey = lkey; sge 705 drivers/infiniband/sw/siw/siw_qp_rx.c resp->sge[1].length = msn; sge 751 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->sqe.sge[0].laddr = orqe->sge[0].laddr; sge 752 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->sqe.sge[0].lkey = orqe->sge[0].lkey; sge 753 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->sqe.sge[0].length = orqe->sge[0].length; sge 756 drivers/infiniband/sw/siw/siw_qp_rx.c wqe->bytes = orqe->sge[0].length; sge 783 drivers/infiniband/sw/siw/siw_qp_rx.c struct siw_sge *sge; sge 818 drivers/infiniband/sw/siw/siw_qp_rx.c sge = wqe->sqe.sge; /* there is only one */ sge 825 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0, sge 846 drivers/infiniband/sw/siw/siw_qp_rx.c (void *)(uintptr_t)(sge->laddr + wqe->processed), sge 849 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_rx_umem(srx, mem_p->umem, sge->laddr + wqe->processed, sge 853 drivers/infiniband/sw/siw/siw_qp_rx.c sge->laddr + wqe->processed, bytes); sge 1272 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey); sge 43 drivers/infiniband/sw/siw/siw_qp_tx.c struct siw_sge *sge = &wqe->sqe.sge[0]; sge 44 drivers/infiniband/sw/siw/siw_qp_tx.c u32 bytes = sge->length; sge 53 drivers/infiniband/sw/siw/siw_qp_tx.c memcpy(paddr, &wqe->sqe.sge[1], bytes); sge 60 drivers/infiniband/sw/siw/siw_qp_tx.c (const void *)(uintptr_t)sge->laddr, bytes); sge 62 drivers/infiniband/sw/siw/siw_qp_tx.c if (copy_from_user(paddr, u64_to_user_ptr(sge->laddr), sge 66 drivers/infiniband/sw/siw/siw_qp_tx.c unsigned int off = sge->laddr & ~PAGE_MASK; sge 72 drivers/infiniband/sw/siw/siw_qp_tx.c p = siw_get_upage(mem->umem, sge->laddr); sge 74 drivers/infiniband/sw/siw/siw_qp_tx.c p = siw_get_pblpage(mem, sge->laddr, &pbl_idx); sge 91 drivers/infiniband/sw/siw/siw_qp_tx.c sge->laddr + part); sge 94 drivers/infiniband/sw/siw/siw_qp_tx.c sge->laddr + part, sge 137 drivers/infiniband/sw/siw/siw_qp_tx.c c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey); sge 139 drivers/infiniband/sw/siw/siw_qp_tx.c cpu_to_be64(wqe->sqe.sge[0].laddr); sge 142 drivers/infiniband/sw/siw/siw_qp_tx.c c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length); sge 370 drivers/infiniband/sw/siw/siw_qp_tx.c struct siw_sge *sge, unsigned int offset, sge 374 drivers/infiniband/sw/siw/siw_qp_tx.c int sge_bytes = min(sge->length - offset, size); sge 376 drivers/infiniband/sw/siw/siw_qp_tx.c offset = (sge->laddr + offset) & ~PAGE_MASK; sge 386 drivers/infiniband/sw/siw/siw_qp_tx.c sge++; sge 387 drivers/infiniband/sw/siw/siw_qp_tx.c sge_bytes = min(sge->length, size - sent); sge 388 drivers/infiniband/sw/siw/siw_qp_tx.c offset = sge->laddr & ~PAGE_MASK; sge 427 drivers/infiniband/sw/siw/siw_qp_tx.c struct siw_sge *sge = &wqe->sqe.sge[c_tx->sge_idx]; sge 457 drivers/infiniband/sw/siw/siw_qp_tx.c unsigned int sge_len = min(sge->length - sge_off, data_len); sge 458 drivers/infiniband/sw/siw/siw_qp_tx.c unsigned int fp_off = (sge->laddr + sge_off) & ~PAGE_MASK; sge 473 drivers/infiniband/sw/siw/siw_qp_tx.c (void *)(uintptr_t)(sge->laddr + sge_off); sge 494 drivers/infiniband/sw/siw/siw_qp_tx.c mem, sge->laddr + sge_off, sge 498 drivers/infiniband/sw/siw/siw_qp_tx.c sge->laddr + sge_off); sge 526 drivers/infiniband/sw/siw/siw_qp_tx.c u64 va = sge->laddr + sge_off; sge 551 drivers/infiniband/sw/siw/siw_qp_tx.c if (sge_off == sge->length && sge 554 drivers/infiniband/sw/siw/siw_qp_tx.c sge++; sge 582 drivers/infiniband/sw/siw/siw_qp_tx.c rv = siw_0copy_tx(s, page_array, &wqe->sqe.sge[c_tx->sge_idx], sge 640 drivers/infiniband/sw/siw/siw_qp_tx.c sge = &wqe->sqe.sge[c_tx->sge_idx]; sge 641 drivers/infiniband/sw/siw/siw_qp_tx.c sge_unsent = sge->length - c_tx->sge_off; sge 647 drivers/infiniband/sw/siw/siw_qp_tx.c sge++; sge 648 drivers/infiniband/sw/siw/siw_qp_tx.c sge_unsent = sge->length; sge 754 drivers/infiniband/sw/siw/siw_qp_tx.c struct siw_sge *sge = &wqe->sqe.sge[0]; sge 760 drivers/infiniband/sw/siw/siw_qp_tx.c for (i = 0, len = 0; num_sge; num_sge--, i++, sge++) { sge 764 drivers/infiniband/sw/siw/siw_qp_tx.c if (sge->length) { sge 765 drivers/infiniband/sw/siw/siw_qp_tx.c int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0, sge 766 drivers/infiniband/sw/siw/siw_qp_tx.c sge->length); sge 771 drivers/infiniband/sw/siw/siw_qp_tx.c len += sge->length; sge 819 drivers/infiniband/sw/siw/siw_qp_tx.c wqe->bytes = wqe->sqe.sge[0].length; sge 825 drivers/infiniband/sw/siw/siw_qp_tx.c wqe->sqe.sge[0].laddr = sge 826 drivers/infiniband/sw/siw/siw_qp_tx.c (u64)(uintptr_t)&wqe->sqe.sge[1]; sge 660 drivers/infiniband/sw/siw/siw_verbs.c void *kbuf = &sqe->sge[1]; sge 663 drivers/infiniband/sw/siw/siw_verbs.c sqe->sge[0].laddr = (uintptr_t)kbuf; sge 664 drivers/infiniband/sw/siw/siw_verbs.c sqe->sge[0].lkey = 0; sge 682 drivers/infiniband/sw/siw/siw_verbs.c sqe->sge[0].length = bytes > 0 ? bytes : 0; sge 828 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, sqe->sge, sge 861 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1); sge 877 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, &sqe->sge[0], sge 1060 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge); sge 1781 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge); sge 24 drivers/infiniband/sw/siw/siw_verbs.h static inline void siw_copy_sgl(struct ib_sge *sge, struct siw_sge *siw_sge, sge 28 drivers/infiniband/sw/siw/siw_verbs.h siw_sge->laddr = sge->addr; sge 29 drivers/infiniband/sw/siw/siw_verbs.h siw_sge->length = sge->length; sge 30 drivers/infiniband/sw/siw/siw_verbs.h siw_sge->lkey = sge->lkey; sge 33 drivers/infiniband/sw/siw/siw_verbs.h sge++; sge 119 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_sge *sge, int id) sge 127 drivers/infiniband/ulp/ipoib/ipoib_cm.c sge[i].addr = rx->rx_ring[id].mapping[i]; sge 330 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_sge *sge) sge 336 drivers/infiniband/ulp/ipoib/ipoib_cm.c sge[i].lkey = priv->pd->local_dma_lkey; sge 338 drivers/infiniband/ulp/ipoib/ipoib_cm.c sge[0].length = IPOIB_CM_HEAD_SIZE; sge 340 drivers/infiniband/ulp/ipoib/ipoib_cm.c sge[i].length = PAGE_SIZE; sge 343 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->sg_list = sge; sge 353 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_sge sge[IPOIB_CM_RX_SG]; sge 369 drivers/infiniband/ulp/ipoib/ipoib_cm.c ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); sge 391 drivers/infiniband/ulp/ipoib/ipoib_cm.c ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); sge 218 drivers/infiniband/ulp/iser/iscsi_iser.h struct ib_sge sge; sge 297 drivers/infiniband/ulp/iser/iscsi_iser.h struct ib_sge sge; sge 84 drivers/infiniband/ulp/iser/iser_initiator.c hdr->read_va = cpu_to_be64(mem_reg->sge.addr); sge 88 drivers/infiniband/ulp/iser/iser_initiator.c (unsigned long long)mem_reg->sge.addr); sge 142 drivers/infiniband/ulp/iser/iser_initiator.c hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz); sge 147 drivers/infiniband/ulp/iser/iser_initiator.c (unsigned long long)mem_reg->sge.addr, unsol_sz); sge 153 drivers/infiniband/ulp/iser/iser_initiator.c tx_dsg->addr = mem_reg->sge.addr; sge 155 drivers/infiniband/ulp/iser/iser_initiator.c tx_dsg->lkey = mem_reg->sge.lkey; sge 470 drivers/infiniband/ulp/iser/iser_initiator.c tx_dsg->addr = mem_reg->sge.addr + buf_offset; sge 472 drivers/infiniband/ulp/iser/iser_initiator.c tx_dsg->lkey = mem_reg->sge.lkey; sge 196 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.lkey = device->pd->local_dma_lkey; sge 206 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.addr = sg_dma_address(&sg[0]); sge 207 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.length = sg_dma_len(&sg[0]); sge 210 drivers/infiniband/ulp/iser/iser_memory.c " length=0x%x\n", reg->sge.lkey, reg->rkey, sge 211 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.addr, reg->sge.length); sge 258 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.lkey = fmr->fmr->lkey; sge 260 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.addr = page_vec->fake_mr.iova; sge 261 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.length = page_vec->fake_mr.length; sge 265 drivers/infiniband/ulp/iser/iser_memory.c " length=0x%x\n", reg->sge.lkey, reg->rkey, sge 266 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.addr, reg->sge.length); sge 426 drivers/infiniband/ulp/iser/iser_memory.c sig_reg->sge.lkey = mr->lkey; sge 428 drivers/infiniband/ulp/iser/iser_memory.c sig_reg->sge.addr = mr->iova; sge 429 drivers/infiniband/ulp/iser/iser_memory.c sig_reg->sge.length = mr->length; sge 432 drivers/infiniband/ulp/iser/iser_memory.c sig_reg->sge.lkey, sig_reg->rkey, sig_reg->sge.addr, sge 433 drivers/infiniband/ulp/iser/iser_memory.c sig_reg->sge.length); sge 474 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.lkey = mr->lkey; sge 476 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.addr = mr->iova; sge 477 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.length = mr->length; sge 480 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.lkey, reg->rkey, reg->sge.addr, reg->sge.length); sge 979 drivers/infiniband/ulp/iser/iser_verbs.c desc->sge.addr = desc->rsp_dma; sge 980 drivers/infiniband/ulp/iser/iser_verbs.c desc->sge.length = ISER_RX_LOGIN_SIZE; sge 981 drivers/infiniband/ulp/iser/iser_verbs.c desc->sge.lkey = ib_conn->device->pd->local_dma_lkey; sge 985 drivers/infiniband/ulp/iser/iser_verbs.c wr.sg_list = &desc->sge; sge 963 drivers/infiniband/ulp/isert/ib_isert.c struct ib_sge sge; sge 966 drivers/infiniband/ulp/isert/ib_isert.c memset(&sge, 0, sizeof(struct ib_sge)); sge 967 drivers/infiniband/ulp/isert/ib_isert.c sge.addr = isert_conn->login_req_dma; sge 968 drivers/infiniband/ulp/isert/ib_isert.c sge.length = ISER_RX_PAYLOAD_SIZE; sge 969 drivers/infiniband/ulp/isert/ib_isert.c sge.lkey = isert_conn->device->pd->local_dma_lkey; sge 972 drivers/infiniband/ulp/isert/ib_isert.c sge.addr, sge.length, sge.lkey); sge 978 drivers/infiniband/ulp/isert/ib_isert.c rx_wr.sg_list = &sge; sge 1845 drivers/infiniband/ulp/srp/ib_srp.c struct ib_sge *sge = &req->cmd->sge[1]; sge 1854 drivers/infiniband/ulp/srp/ib_srp.c sge[i].addr = sg_dma_address(sg); sge 1855 drivers/infiniband/ulp/srp/ib_srp.c sge[i].length = sg_dma_len(sg); sge 1856 drivers/infiniband/ulp/srp/ib_srp.c sge[i].lkey = target->lkey; sge 2079 drivers/infiniband/ulp/srp/ib_srp.c iu->sge[0].addr = iu->dma; sge 2080 drivers/infiniband/ulp/srp/ib_srp.c iu->sge[0].length = len; sge 2081 drivers/infiniband/ulp/srp/ib_srp.c iu->sge[0].lkey = target->lkey; sge 2087 drivers/infiniband/ulp/srp/ib_srp.c wr.sg_list = &iu->sge[0]; sge 277 drivers/infiniband/ulp/srp/ib_srp.h struct ib_sge sge[SRP_MAX_SGE]; sge 2798 drivers/infiniband/ulp/srpt/ib_srpt.c struct ib_sge sge; sge 2858 drivers/infiniband/ulp/srpt/ib_srpt.c sge.addr = ioctx->ioctx.dma; sge 2859 drivers/infiniband/ulp/srpt/ib_srpt.c sge.length = resp_len; sge 2860 drivers/infiniband/ulp/srpt/ib_srpt.c sge.lkey = sdev->lkey; sge 2865 drivers/infiniband/ulp/srpt/ib_srpt.c send_wr.sg_list = &sge; sge 334 drivers/message/fusion/mptbase.h SGESimple32_t sge[1]; sge 338 drivers/message/fusion/mptbase.h SGESimple64_t sge[1]; sge 550 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; sge 574 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge->addr_hi = cpu_to_le32(U64_HI(mapping)); sge 575 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge->addr_lo = cpu_to_le32(U64_LO(mapping)); sge 1342 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c struct eth_rx_sge *sge; sge 1344 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2]; sge 1345 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge->addr_hi = sge 1349 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge->addr_lo = sge 806 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; sge 821 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h sge->addr_hi = 0; sge 822 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h sge->addr_lo = 0; sge 3097 drivers/net/ethernet/broadcom/cnic_defs.h struct fcoe_bd_ctx sge[3]; sge 190 drivers/net/ethernet/chelsio/cxgb/common.h struct sge_params sge; sge 228 drivers/net/ethernet/chelsio/cxgb/common.h struct sge; sge 246 drivers/net/ethernet/chelsio/cxgb/common.h struct sge *sge; sge 170 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed); sge 224 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sge_start(adapter->sge); sge 235 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sge_stop(adapter->sge); sge 463 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t = t1_sge_get_intr_counts(adapter->sge); sge 464 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss); sge 722 drivers/net/ethernet/chelsio/cxgb/cxgb2.c e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl]; sge 723 drivers/net/ethernet/chelsio/cxgb/cxgb2.c e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl]; sge 724 drivers/net/ethernet/chelsio/cxgb/cxgb2.c e->tx_pending = adapter->params.sge.cmdQ_size[0]; sge 743 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending; sge 744 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending; sge 745 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.cmdQ_size[0] = e->tx_pending; sge 746 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ? sge 755 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.rx_coalesce_usecs = c->rx_coalesce_usecs; sge 756 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.coalesce_enable = c->use_adaptive_rx_coalesce; sge 757 drivers/net/ethernet/chelsio/cxgb/cxgb2.c adapter->params.sge.sample_interval_usecs = c->rate_sample_interval; sge 758 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sge_set_coalesce_params(adapter->sge, &adapter->params.sge); sge 766 drivers/net/ethernet/chelsio/cxgb/cxgb2.c c->rx_coalesce_usecs = adapter->params.sge.rx_coalesce_usecs; sge 767 drivers/net/ethernet/chelsio/cxgb/cxgb2.c c->rate_sample_interval = adapter->params.sge.sample_interval_usecs; sge 768 drivers/net/ethernet/chelsio/cxgb/cxgb2.c c->use_adaptive_rx_coalesce = adapter->params.sge.coalesce_enable; sge 955 drivers/net/ethernet/chelsio/cxgb/cxgb2.c t1_sge_stop(adapter->sge); sge 282 drivers/net/ethernet/chelsio/cxgb/sge.c static void tx_sched_stop(struct sge *sge) sge 284 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 297 drivers/net/ethernet/chelsio/cxgb/sge.c unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port, sge 300 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 320 drivers/net/ethernet/chelsio/cxgb/sge.c if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) { sge 343 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val) sge 345 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 350 drivers/net/ethernet/chelsio/cxgb/sge.c t1_sched_update_parms(sge, i, 0, 0); sge 357 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port, sge 360 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 363 drivers/net/ethernet/chelsio/cxgb/sge.c t1_sched_update_parms(sge, port, 0, 0); sge 371 drivers/net/ethernet/chelsio/cxgb/sge.c static int tx_sched_init(struct sge *sge) sge 381 drivers/net/ethernet/chelsio/cxgb/sge.c tasklet_init(&s->sched_tsk, restart_sched, (unsigned long) sge); sge 382 drivers/net/ethernet/chelsio/cxgb/sge.c sge->tx_sched = s; sge 386 drivers/net/ethernet/chelsio/cxgb/sge.c t1_sched_update_parms(sge, i, 1500, 1000); sge 397 drivers/net/ethernet/chelsio/cxgb/sge.c static inline int sched_update_avail(struct sge *sge) sge 399 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 431 drivers/net/ethernet/chelsio/cxgb/sge.c static struct sk_buff *sched_skb(struct sge *sge, struct sk_buff *skb, sge 434 drivers/net/ethernet/chelsio/cxgb/sge.c struct sched *s = sge->tx_sched; sge 472 drivers/net/ethernet/chelsio/cxgb/sge.c if (update-- && sched_update_avail(sge)) sge 480 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[0]; sge 484 drivers/net/ethernet/chelsio/cxgb/sge.c writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL); sge 525 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_rx_resources(struct sge *sge) sge 527 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 530 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->respQ.entries) { sge 531 drivers/net/ethernet/chelsio/cxgb/sge.c size = sizeof(struct respQ_e) * sge->respQ.size; sge 532 drivers/net/ethernet/chelsio/cxgb/sge.c pci_free_consistent(pdev, size, sge->respQ.entries, sge 533 drivers/net/ethernet/chelsio/cxgb/sge.c sge->respQ.dma_addr); sge 537 drivers/net/ethernet/chelsio/cxgb/sge.c struct freelQ *q = &sge->freelQ[i]; sge 555 drivers/net/ethernet/chelsio/cxgb/sge.c static int alloc_rx_resources(struct sge *sge, struct sge_params *p) sge 557 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 561 drivers/net/ethernet/chelsio/cxgb/sge.c struct freelQ *q = &sge->freelQ[i]; sge 565 drivers/net/ethernet/chelsio/cxgb/sge.c q->dma_offset = sge->rx_pkt_pad ? 0 : NET_IP_ALIGN; sge 584 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[!sge->jumbo_fl].rx_buffer_size = SGE_RX_SM_BUF_SIZE + sge 586 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[!sge->jumbo_fl].dma_offset; sge 590 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[sge->jumbo_fl].rx_buffer_size = size; sge 596 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[!sge->jumbo_fl].recycleq_idx = 0; sge 597 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[sge->jumbo_fl].recycleq_idx = 1; sge 599 drivers/net/ethernet/chelsio/cxgb/sge.c sge->respQ.genbit = 1; sge 600 drivers/net/ethernet/chelsio/cxgb/sge.c sge->respQ.size = SGE_RESPQ_E_N; sge 601 drivers/net/ethernet/chelsio/cxgb/sge.c sge->respQ.credits = 0; sge 602 drivers/net/ethernet/chelsio/cxgb/sge.c size = sizeof(struct respQ_e) * sge->respQ.size; sge 603 drivers/net/ethernet/chelsio/cxgb/sge.c sge->respQ.entries = sge 604 drivers/net/ethernet/chelsio/cxgb/sge.c pci_alloc_consistent(pdev, size, &sge->respQ.dma_addr); sge 605 drivers/net/ethernet/chelsio/cxgb/sge.c if (!sge->respQ.entries) sge 610 drivers/net/ethernet/chelsio/cxgb/sge.c free_rx_resources(sge); sge 617 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_cmdQ_buffers(struct sge *sge, struct cmdQ *q, unsigned int n) sge 620 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 651 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_tx_resources(struct sge *sge) sge 653 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 657 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[i]; sge 661 drivers/net/ethernet/chelsio/cxgb/sge.c free_cmdQ_buffers(sge, q, q->in_use); sge 675 drivers/net/ethernet/chelsio/cxgb/sge.c static int alloc_tx_resources(struct sge *sge, struct sge_params *p) sge 677 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 681 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[i]; sge 709 drivers/net/ethernet/chelsio/cxgb/sge.c sge->cmdQ[0].stop_thres = sge->adapter->params.nports * sge 714 drivers/net/ethernet/chelsio/cxgb/sge.c free_tx_resources(sge); sge 732 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 735 drivers/net/ethernet/chelsio/cxgb/sge.c sge->sge_control |= F_VLAN_XTRACT; sge 737 drivers/net/ethernet/chelsio/cxgb/sge.c sge->sge_control &= ~F_VLAN_XTRACT; sge 739 drivers/net/ethernet/chelsio/cxgb/sge.c writel(sge->sge_control, adapter->regs + A_SG_CONTROL); sge 748 drivers/net/ethernet/chelsio/cxgb/sge.c static void configure_sge(struct sge *sge, struct sge_params *p) sge 750 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *ap = sge->adapter; sge 753 drivers/net/ethernet/chelsio/cxgb/sge.c setup_ring_params(ap, sge->cmdQ[0].dma_addr, sge->cmdQ[0].size, sge 755 drivers/net/ethernet/chelsio/cxgb/sge.c setup_ring_params(ap, sge->cmdQ[1].dma_addr, sge->cmdQ[1].size, sge 757 drivers/net/ethernet/chelsio/cxgb/sge.c setup_ring_params(ap, sge->freelQ[0].dma_addr, sge 758 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[0].size, A_SG_FL0BASELWR, sge 760 drivers/net/ethernet/chelsio/cxgb/sge.c setup_ring_params(ap, sge->freelQ[1].dma_addr, sge 761 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[1].size, A_SG_FL1BASELWR, sge 767 drivers/net/ethernet/chelsio/cxgb/sge.c setup_ring_params(ap, sge->respQ.dma_addr, sge->respQ.size, sge 769 drivers/net/ethernet/chelsio/cxgb/sge.c writel((u32)sge->respQ.size - 1, ap->regs + A_SG_RSPQUEUECREDIT); sge 771 drivers/net/ethernet/chelsio/cxgb/sge.c sge->sge_control = F_CMDQ0_ENABLE | F_CMDQ1_ENABLE | F_FL0_ENABLE | sge 774 drivers/net/ethernet/chelsio/cxgb/sge.c V_RX_PKT_OFFSET(sge->rx_pkt_pad); sge 777 drivers/net/ethernet/chelsio/cxgb/sge.c sge->sge_control |= F_ENABLE_BIG_ENDIAN; sge 781 drivers/net/ethernet/chelsio/cxgb/sge.c sge->intrtimer_nres = SGE_INTRTIMER_NRES * core_ticks_per_usec(ap); sge 783 drivers/net/ethernet/chelsio/cxgb/sge.c t1_sge_set_coalesce_params(sge, p); sge 789 drivers/net/ethernet/chelsio/cxgb/sge.c static inline unsigned int jumbo_payload_capacity(const struct sge *sge) sge 791 drivers/net/ethernet/chelsio/cxgb/sge.c return sge->freelQ[sge->jumbo_fl].rx_buffer_size - sge 792 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[sge->jumbo_fl].dma_offset - sge 799 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_destroy(struct sge *sge) sge 803 drivers/net/ethernet/chelsio/cxgb/sge.c for_each_port(sge->adapter, i) sge 804 drivers/net/ethernet/chelsio/cxgb/sge.c free_percpu(sge->port_stats[i]); sge 806 drivers/net/ethernet/chelsio/cxgb/sge.c kfree(sge->tx_sched); sge 807 drivers/net/ethernet/chelsio/cxgb/sge.c free_tx_resources(sge); sge 808 drivers/net/ethernet/chelsio/cxgb/sge.c free_rx_resources(sge); sge 809 drivers/net/ethernet/chelsio/cxgb/sge.c kfree(sge); sge 824 drivers/net/ethernet/chelsio/cxgb/sge.c static void refill_free_list(struct sge *sge, struct freelQ *q) sge 826 drivers/net/ethernet/chelsio/cxgb/sge.c struct pci_dev *pdev = sge->adapter->pdev; sge 842 drivers/net/ethernet/chelsio/cxgb/sge.c skb_reserve(skb, sge->rx_pkt_pad); sge 870 drivers/net/ethernet/chelsio/cxgb/sge.c static void freelQs_empty(struct sge *sge) sge 872 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 876 drivers/net/ethernet/chelsio/cxgb/sge.c refill_free_list(sge, &sge->freelQ[0]); sge 877 drivers/net/ethernet/chelsio/cxgb/sge.c refill_free_list(sge, &sge->freelQ[1]); sge 879 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->freelQ[0].credits > (sge->freelQ[0].size >> 2) && sge 880 drivers/net/ethernet/chelsio/cxgb/sge.c sge->freelQ[1].credits > (sge->freelQ[1].size >> 2)) { sge 882 drivers/net/ethernet/chelsio/cxgb/sge.c irqholdoff_reg = sge->fixed_intrtimer; sge 886 drivers/net/ethernet/chelsio/cxgb/sge.c irqholdoff_reg = sge->intrtimer_nres; sge 903 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_disable(struct sge *sge) sge 905 drivers/net/ethernet/chelsio/cxgb/sge.c u32 val = readl(sge->adapter->regs + A_PL_ENABLE); sge 907 drivers/net/ethernet/chelsio/cxgb/sge.c writel(val & ~SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE); sge 908 drivers/net/ethernet/chelsio/cxgb/sge.c writel(0, sge->adapter->regs + A_SG_INT_ENABLE); sge 914 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_enable(struct sge *sge) sge 917 drivers/net/ethernet/chelsio/cxgb/sge.c u32 val = readl(sge->adapter->regs + A_PL_ENABLE); sge 919 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->adapter->port[0].dev->hw_features & NETIF_F_TSO) sge 921 drivers/net/ethernet/chelsio/cxgb/sge.c writel(en, sge->adapter->regs + A_SG_INT_ENABLE); sge 922 drivers/net/ethernet/chelsio/cxgb/sge.c writel(val | SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE); sge 928 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_clear(struct sge *sge) sge 930 drivers/net/ethernet/chelsio/cxgb/sge.c writel(SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_CAUSE); sge 931 drivers/net/ethernet/chelsio/cxgb/sge.c writel(0xffffffff, sge->adapter->regs + A_SG_INT_CAUSE); sge 937 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_intr_error_handler(struct sge *sge) sge 939 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 945 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.respQ_empty++; sge 947 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.respQ_overflow++; sge 952 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.freelistQ_empty++; sge 953 drivers/net/ethernet/chelsio/cxgb/sge.c freelQs_empty(sge); sge 956 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.pkt_too_big++; sge 961 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.pkt_mismatch++; sge 971 drivers/net/ethernet/chelsio/cxgb/sge.c const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge) sge 973 drivers/net/ethernet/chelsio/cxgb/sge.c return &sge->stats; sge 976 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_get_port_stats(const struct sge *sge, int port, sge 983 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge_port_stats *st = per_cpu_ptr(sge->port_stats[port], cpu); sge 1288 drivers/net/ethernet/chelsio/cxgb/sge.c static inline void reclaim_completed_tx(struct sge *sge, struct cmdQ *q) sge 1295 drivers/net/ethernet/chelsio/cxgb/sge.c free_cmdQ_buffers(sge, q, reclaim); sge 1306 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = (struct sge *) arg; sge 1307 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 1308 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[0]; sge 1313 drivers/net/ethernet/chelsio/cxgb/sge.c reclaim_completed_tx(sge, q); sge 1317 drivers/net/ethernet/chelsio/cxgb/sge.c while ((skb = sched_skb(sge, NULL, credits)) != NULL) { sge 1352 drivers/net/ethernet/chelsio/cxgb/sge.c static void sge_rx(struct sge *sge, struct freelQ *fl, unsigned int len) sge 1356 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 1360 drivers/net/ethernet/chelsio/cxgb/sge.c skb = get_packet(adapter, fl, len - sge->rx_pkt_pad); sge 1362 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.rx_drops++; sge 1373 drivers/net/ethernet/chelsio/cxgb/sge.c st = this_cpu_ptr(sge->port_stats[p->iff]); sge 1407 drivers/net/ethernet/chelsio/cxgb/sge.c static void restart_tx_queues(struct sge *sge) sge 1409 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adap = sge->adapter; sge 1412 drivers/net/ethernet/chelsio/cxgb/sge.c if (!enough_free_Tx_descs(&sge->cmdQ[0])) sge 1418 drivers/net/ethernet/chelsio/cxgb/sge.c if (test_and_clear_bit(nd->if_port, &sge->stopped_tx_queues) && sge 1420 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.cmdQ_restarted[2]++; sge 1434 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1435 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *cmdq = &sge->cmdQ[0]; sge 1439 drivers/net/ethernet/chelsio/cxgb/sge.c freelQs_empty(sge); sge 1450 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->tx_sched) sge 1451 drivers/net/ethernet/chelsio/cxgb/sge.c tasklet_hi_schedule(&sge->tx_sched->sched_tsk); sge 1456 drivers/net/ethernet/chelsio/cxgb/sge.c if (unlikely(sge->stopped_tx_queues != 0)) sge 1457 drivers/net/ethernet/chelsio/cxgb/sge.c restart_tx_queues(sge); sge 1468 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1469 drivers/net/ethernet/chelsio/cxgb/sge.c struct respQ *q = &sge->respQ; sge 1491 drivers/net/ethernet/chelsio/cxgb/sge.c sge->cmdQ[1].processed += cmdq_processed[1]; sge 1496 drivers/net/ethernet/chelsio/cxgb/sge.c struct freelQ *fl = &sge->freelQ[e->FreelistQid]; sge 1502 drivers/net/ethernet/chelsio/cxgb/sge.c sge_rx(sge, fl, e->BufferLength); sge 1516 drivers/net/ethernet/chelsio/cxgb/sge.c refill_free_list(sge, fl); sge 1518 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.pure_rsps++; sge 1535 drivers/net/ethernet/chelsio/cxgb/sge.c sge->cmdQ[1].processed += cmdq_processed[1]; sge 1542 drivers/net/ethernet/chelsio/cxgb/sge.c const struct respQ *Q = &adapter->sge->respQ; sge 1558 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1559 drivers/net/ethernet/chelsio/cxgb/sge.c struct respQ *q = &sge->respQ; sge 1561 drivers/net/ethernet/chelsio/cxgb/sge.c const struct freelQ *fl = &sge->freelQ[e->FreelistQid]; sge 1587 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.pure_rsps++; sge 1591 drivers/net/ethernet/chelsio/cxgb/sge.c sge->cmdQ[1].processed += cmdq_processed[1]; sge 1608 drivers/net/ethernet/chelsio/cxgb/sge.c writel(adapter->sge->respQ.cidx, sge 1617 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1628 drivers/net/ethernet/chelsio/cxgb/sge.c writel(sge->respQ.cidx, adapter->regs + A_SG_SLEEPING); sge 1641 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.unhandled_irqs++; sge 1662 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1663 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[qid]; sge 1668 drivers/net/ethernet/chelsio/cxgb/sge.c reclaim_completed_tx(sge, q); sge 1679 drivers/net/ethernet/chelsio/cxgb/sge.c set_bit(dev->if_port, &sge->stopped_tx_queues); sge 1680 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.cmdQ_full[2]++; sge 1690 drivers/net/ethernet/chelsio/cxgb/sge.c set_bit(dev->if_port, &sge->stopped_tx_queues); sge 1691 drivers/net/ethernet/chelsio/cxgb/sge.c sge->stats.cmdQ_full[2]++; sge 1697 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->tx_sched && !qid && skb->dev) { sge 1703 drivers/net/ethernet/chelsio/cxgb/sge.c skb = sched_skb(sge, skb, credits); sge 1773 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = adapter->sge; sge 1774 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge_port_stats *st = this_cpu_ptr(sge->port_stats[dev->if_port]); sge 1839 drivers/net/ethernet/chelsio/cxgb/sge.c if ((unlikely(!adapter->sge->espibug_skb[dev->if_port]))) { sge 1842 drivers/net/ethernet/chelsio/cxgb/sge.c adapter->sge->espibug_skb[dev->if_port] = skb; sge 1887 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = from_timer(sge, t, tx_reclaim_timer); sge 1890 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[i]; sge 1895 drivers/net/ethernet/chelsio/cxgb/sge.c reclaim_completed_tx(sge, q); sge 1897 drivers/net/ethernet/chelsio/cxgb/sge.c writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL); sge 1901 drivers/net/ethernet/chelsio/cxgb/sge.c mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD); sge 1907 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_set_coalesce_params(struct sge *sge, struct sge_params *p) sge 1909 drivers/net/ethernet/chelsio/cxgb/sge.c sge->fixed_intrtimer = p->rx_coalesce_usecs * sge 1910 drivers/net/ethernet/chelsio/cxgb/sge.c core_ticks_per_usec(sge->adapter); sge 1911 drivers/net/ethernet/chelsio/cxgb/sge.c writel(sge->fixed_intrtimer, sge->adapter->regs + A_SG_INTRTIMER); sge 1919 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_configure(struct sge *sge, struct sge_params *p) sge 1921 drivers/net/ethernet/chelsio/cxgb/sge.c if (alloc_rx_resources(sge, p)) sge 1923 drivers/net/ethernet/chelsio/cxgb/sge.c if (alloc_tx_resources(sge, p)) { sge 1924 drivers/net/ethernet/chelsio/cxgb/sge.c free_rx_resources(sge); sge 1927 drivers/net/ethernet/chelsio/cxgb/sge.c configure_sge(sge, p); sge 1935 drivers/net/ethernet/chelsio/cxgb/sge.c p->large_buf_capacity = jumbo_payload_capacity(sge); sge 1942 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_stop(struct sge *sge) sge 1945 drivers/net/ethernet/chelsio/cxgb/sge.c writel(0, sge->adapter->regs + A_SG_CONTROL); sge 1946 drivers/net/ethernet/chelsio/cxgb/sge.c readl(sge->adapter->regs + A_SG_CONTROL); /* flush */ sge 1948 drivers/net/ethernet/chelsio/cxgb/sge.c if (is_T2(sge->adapter)) sge 1949 drivers/net/ethernet/chelsio/cxgb/sge.c del_timer_sync(&sge->espibug_timer); sge 1951 drivers/net/ethernet/chelsio/cxgb/sge.c del_timer_sync(&sge->tx_reclaim_timer); sge 1952 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->tx_sched) sge 1953 drivers/net/ethernet/chelsio/cxgb/sge.c tx_sched_stop(sge); sge 1956 drivers/net/ethernet/chelsio/cxgb/sge.c kfree_skb(sge->espibug_skb[i]); sge 1962 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_start(struct sge *sge) sge 1964 drivers/net/ethernet/chelsio/cxgb/sge.c refill_free_list(sge, &sge->freelQ[0]); sge 1965 drivers/net/ethernet/chelsio/cxgb/sge.c refill_free_list(sge, &sge->freelQ[1]); sge 1967 drivers/net/ethernet/chelsio/cxgb/sge.c writel(sge->sge_control, sge->adapter->regs + A_SG_CONTROL); sge 1968 drivers/net/ethernet/chelsio/cxgb/sge.c doorbell_pio(sge->adapter, F_FL0_ENABLE | F_FL1_ENABLE); sge 1969 drivers/net/ethernet/chelsio/cxgb/sge.c readl(sge->adapter->regs + A_SG_CONTROL); /* flush */ sge 1971 drivers/net/ethernet/chelsio/cxgb/sge.c mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD); sge 1973 drivers/net/ethernet/chelsio/cxgb/sge.c if (is_T2(sge->adapter)) sge 1974 drivers/net/ethernet/chelsio/cxgb/sge.c mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout); sge 1982 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = from_timer(sge, t, espibug_timer); sge 1983 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 1994 drivers/net/ethernet/chelsio/cxgb/sge.c struct sk_buff *skb = sge->espibug_skb[i]; sge 2020 drivers/net/ethernet/chelsio/cxgb/sge.c mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout); sge 2025 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = from_timer(sge, t, espibug_timer); sge 2026 drivers/net/ethernet/chelsio/cxgb/sge.c struct adapter *adapter = sge->adapter; sge 2029 drivers/net/ethernet/chelsio/cxgb/sge.c struct sk_buff *skb = sge->espibug_skb[0]; sge 2052 drivers/net/ethernet/chelsio/cxgb/sge.c mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout); sge 2058 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *t1_sge_create(struct adapter *adapter, struct sge_params *p) sge 2060 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *sge = kzalloc(sizeof(*sge), GFP_KERNEL); sge 2063 drivers/net/ethernet/chelsio/cxgb/sge.c if (!sge) sge 2066 drivers/net/ethernet/chelsio/cxgb/sge.c sge->adapter = adapter; sge 2067 drivers/net/ethernet/chelsio/cxgb/sge.c sge->netdev = adapter->port[0].dev; sge 2068 drivers/net/ethernet/chelsio/cxgb/sge.c sge->rx_pkt_pad = t1_is_T1B(adapter) ? 0 : 2; sge 2069 drivers/net/ethernet/chelsio/cxgb/sge.c sge->jumbo_fl = t1_is_T1B(adapter) ? 1 : 0; sge 2072 drivers/net/ethernet/chelsio/cxgb/sge.c sge->port_stats[i] = alloc_percpu(struct sge_port_stats); sge 2073 drivers/net/ethernet/chelsio/cxgb/sge.c if (!sge->port_stats[i]) sge 2077 drivers/net/ethernet/chelsio/cxgb/sge.c timer_setup(&sge->tx_reclaim_timer, sge_tx_reclaim_cb, 0); sge 2079 drivers/net/ethernet/chelsio/cxgb/sge.c if (is_T2(sge->adapter)) { sge 2080 drivers/net/ethernet/chelsio/cxgb/sge.c timer_setup(&sge->espibug_timer, sge 2085 drivers/net/ethernet/chelsio/cxgb/sge.c tx_sched_init(sge); sge 2087 drivers/net/ethernet/chelsio/cxgb/sge.c sge->espibug_timeout = 1; sge 2090 drivers/net/ethernet/chelsio/cxgb/sge.c sge->espibug_timeout = HZ/100; sge 2096 drivers/net/ethernet/chelsio/cxgb/sge.c p->freelQ_size[!sge->jumbo_fl] = SGE_FREEL_SIZE; sge 2097 drivers/net/ethernet/chelsio/cxgb/sge.c p->freelQ_size[sge->jumbo_fl] = SGE_JUMBO_FREEL_SIZE; sge 2098 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->tx_sched) { sge 2099 drivers/net/ethernet/chelsio/cxgb/sge.c if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) sge 2109 drivers/net/ethernet/chelsio/cxgb/sge.c return sge; sge 2112 drivers/net/ethernet/chelsio/cxgb/sge.c free_percpu(sge->port_stats[i]); sge 2115 drivers/net/ethernet/chelsio/cxgb/sge.c kfree(sge); sge 71 drivers/net/ethernet/chelsio/cxgb/sge.h struct sge; sge 73 drivers/net/ethernet/chelsio/cxgb/sge.h struct sge *t1_sge_create(struct adapter *, struct sge_params *); sge 74 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_configure(struct sge *, struct sge_params *); sge 75 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_set_coalesce_params(struct sge *, struct sge_params *); sge 76 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_destroy(struct sge *); sge 82 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_start(struct sge *); sge 83 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_stop(struct sge *); sge 84 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_intr_error_handler(struct sge *); sge 85 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_enable(struct sge *); sge 86 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_disable(struct sge *); sge 87 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_clear(struct sge *); sge 88 drivers/net/ethernet/chelsio/cxgb/sge.h const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge); sge 89 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_get_port_stats(const struct sge *sge, int port, struct sge_port_stats *); sge 90 drivers/net/ethernet/chelsio/cxgb/sge.h unsigned int t1_sched_update_parms(struct sge *, unsigned int, unsigned int, sge 219 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_intr_error_handler(adapter->sge); sge 761 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_intr_enable(adapter->sge); sge 793 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_intr_disable(adapter->sge); sge 819 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_intr_clear(adapter->sge); sge 853 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_intr_error_handler(adapter->sge); sge 982 drivers/net/ethernet/chelsio/cxgb/subr.c err = t1_sge_configure(adapter->sge, &adapter->params.sge); sge 1022 drivers/net/ethernet/chelsio/cxgb/subr.c if (adapter->sge) sge 1023 drivers/net/ethernet/chelsio/cxgb/subr.c t1_sge_destroy(adapter->sge); sge 1059 drivers/net/ethernet/chelsio/cxgb/subr.c adapter->sge = t1_sge_create(adapter, &adapter->params.sge); sge 1060 drivers/net/ethernet/chelsio/cxgb/subr.c if (!adapter->sge) { sge 244 drivers/net/ethernet/chelsio/cxgb3/adapter.h struct sge sge; sge 382 drivers/net/ethernet/chelsio/cxgb3/common.h struct sge_params sge; sge 411 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c adap->sge.qs[qidx]. sge 414 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c &adap->sge.qs[qidx]); sge 418 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c &adap->sge.qs[qidx]); sge 438 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c &adapter->sge.qs[i]); sge 448 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { sge 461 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; sge 597 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct sge_qset *qs = &adap->sge.qs[i]; sge 610 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct sge_qset *qs = &adap->sge.qs[i]; sge 635 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c if (adap->sge.qs[i].adap) sge 636 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c napi_disable(&adap->sge.qs[i].napi); sge 643 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c if (adap->sge.qs[i].adap) sge 644 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c napi_enable(&adap->sge.qs[i].napi); sge 667 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c pi->qs = &adap->sge.qs[pi->first_qset]; sge 672 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c &adap->params.sge.qset[qset_idx], ntxq, dev, sge 1173 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct sge_rspq *q = &adap->sge.qs[i].rspq; sge 1279 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c adap->sge.qs[0].rspq. sge 1660 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c tot += adapter->sge.qs[i].port_stats[idx]; sge 1957 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset]; sge 1989 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c q = &adapter->params.sge.qset[pi->first_qset]; sge 2013 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c qsp = &adapter->params.sge.qset[i]; sge 2014 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c qs = &adapter->sge.qs[i]; sge 2026 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct qset_params *q = adapter->params.sge.qset; sge 2202 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c q = &adapter->params.sge.qset[t.qset_idx]; sge 2220 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c &adapter->sge.qs[t.qset_idx]; sge 2235 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c q = &adapter->params.sge. sge 2278 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c q = &adapter->params.sge.qset[q1 + t.qset_idx]; sge 2623 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct sge_qset *qs = &adapter->sge.qs[qidx]; sge 2771 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct sge_qset *qs = &adapter->sge.qs[0]; sge 303 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_lock_irqsave(&adapter->sge.reg_lock, flags); sge 306 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_unlock_irqrestore(&adapter->sge.reg_lock, flags); sge 334 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_lock_irq(&adapter->sge.reg_lock); sge 341 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_unlock_irq(&adapter->sge.reg_lock); sge 345 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_lock_irq(&adapter->sge.reg_lock); sge 347 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_unlock_irq(&adapter->sge.reg_lock); sge 352 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_lock_irq(&adapter->sge.reg_lock); sge 358 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c spin_unlock_irq(&adapter->sge.reg_lock); sge 683 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_irq(&adapter->sge.reg_lock); sge 685 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock_irq(&adapter->sge.reg_lock); sge 696 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_irq(&adapter->sge.reg_lock); sge 698 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock_irq(&adapter->sge.reg_lock); sge 712 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_irq(&adapter->sge.reg_lock); sge 714 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock_irq(&adapter->sge.reg_lock); sge 1567 drivers/net/ethernet/chelsio/cxgb3/sge.c ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb); sge 1835 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; sge 2094 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); sge 2196 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); sge 2647 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q = &adap->sge.qs[0].rspq; sge 2651 drivers/net/ethernet/chelsio/cxgb3/sge.c if (process_responses(adap, &adap->sge.qs[0], -1)) { sge 2658 drivers/net/ethernet/chelsio/cxgb3/sge.c process_responses(adap, &adap->sge.qs[1], -1)) { sge 2659 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q1 = &adap->sge.qs[1].rspq; sge 2697 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q = &adap->sge.qs[0].rspq; sge 2701 drivers/net/ethernet/chelsio/cxgb3/sge.c new_packets = rspq_check_napi(&adap->sge.qs[0]); sge 2703 drivers/net/ethernet/chelsio/cxgb3/sge.c new_packets += rspq_check_napi(&adap->sge.qs[1]); sge 2735 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q0 = &adap->sge.qs[0].rspq; sge 2736 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q1 = &adap->sge.qs[1].rspq; sge 2773 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_rspq *q0 = &adap->sge.qs[0].rspq; sge 2790 drivers/net/ethernet/chelsio/cxgb3/sge.c process_responses_gts(adap, &adap->sge.qs[1].rspq); sge 2807 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *qs0 = &adap->sge.qs[0]; sge 2825 drivers/net/ethernet/chelsio/cxgb3/sge.c napi_schedule(&adap->sge.qs[1].napi); sge 2970 drivers/net/ethernet/chelsio/cxgb3/sge.c &qs->rspq.lock : &adap->sge.qs[0].rspq.lock; sge 3041 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *q = &adapter->sge.qs[id]; sge 3124 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_irq(&adapter->sge.reg_lock); sge 3168 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock_irq(&adapter->sge.reg_lock); sge 3198 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock_irq(&adapter->sge.reg_lock); sge 3215 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *q = &adap->sge.qs[i]; sge 3238 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *q = &adap->sge.qs[i]; sge 3258 drivers/net/ethernet/chelsio/cxgb3/sge.c t3_free_qset(adap, &adap->sge.qs[i]); sge 3293 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *qs = &adap->sge.qs[i]; sge 3371 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_init(&adap->sge.reg_lock); sge 3403 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c min(adapter->params.sge.max_pkt_size, sge 3406 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c min(adapter->params.sge.max_pkt_size, 16384U)); sge 3423 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_sge_init(adapter, &adapter->params.sge); sge 3687 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_sge_prep(adapter, &adapter->params.sge); sge 2967 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct sge *s = &padap->sge; sge 364 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct sge_params sge; sge 847 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h #define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++) sge 848 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h #define for_each_ofldtxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++) sge 959 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct sge sge; sge 1397 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; sge 2628 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_uld_txq_info *utxq_info = adap->sge.uld_txq_info[uld]; sge 2639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_uld_rxq_info *urxq_info = adap->sge.uld_rxq_info[uld]; sge 2668 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct sge *s = &adap->sge; sge 2670 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4); sge 3012 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c return DIV_ROUND_UP(adap->sge.ethqsets, 4) + sge 3215 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c ssize_t size = (adap->sge.egr_sz + 3) / 4 + sge 3216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c adap->sge.egr_sz / 32 + 2; /* includes ,/\n/\0 */ sge 3223 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c adap->sge.egr_sz, adap->sge.blocked_fl); sge 3237 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c t = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sizeof(long), GFP_KERNEL); sge 3241 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c err = bitmap_parse_user(ubuf, count, t, adap->sge.egr_sz); sge 3247 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c bitmap_copy(adap->sge.blocked_fl, t, adap->sge.egr_sz); sge 244 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; sge 245 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset]; sge 826 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c const struct sge *s = &pi->adapter->sge; sge 844 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge *s = &adapter->sge; sge 878 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; sge 893 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; sge 905 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; sge 921 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c return adap->sge.dbqtimer_tick; sge 933 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c txq = &adap->sge.ethtxq[pi->first_qset]; sge 939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c return adap->sge.dbqtimer_val[txq->dbqtimerix]; sge 953 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge *s = &adap->sge; sge 988 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge *s = &adap->sge; sge 1100 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; sge 1104 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c adap->sge.counter_val[rq->pktcnt_idx] : 0; sge 70 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c QUEUENO_V(adap->sge.fw_evtq.abs_id) | sge 344 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c iq = adapter->sge.ethrxq[pi->first_qset + sge 528 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c t4_mk_filtdelwr(f->tid, fwr, adapter->sge.fw_evtq.abs_id); sge 631 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c FW_FILTER_WR_RX_RPL_IQ_V(adapter->sge.fw_evtq.abs_id)); sge 1198 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c ((adapter->sge.fw_evtq.abs_id << 14) | atid), sge 1208 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c ((adapter->sge.fw_evtq.abs_id << 14) | atid), sge 1501 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c TCB_RSS_INFO_V(adapter->sge.fw_evtq.abs_id), 0, 1); sge 236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; sge 578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; sge 741 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 779 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 841 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c rxq = &adapter->sge.ethrxq[pi->first_qset]; sge 896 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) sge 909 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for (i = 0; i < adap->sge.ingr_sz; i++) { sge 910 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_rspq *q = adap->sge.ingr_map[i]; sge 939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for (i = 0; i < adap->sge.ingr_sz; i++) { sge 940 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_rspq *q = adap->sge.ingr_map[i]; sge 957 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 989 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 1107 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int closest_timer(const struct sge *s, int time) sge 1123 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int closest_thres(const struct sge *s, int thres) sge 1160 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c new_idx = closest_thres(&adap->sge, cnt); sge 1175 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c us = us == 0 ? 6 : closest_timer(&adap->sge, us); sge 1547 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c chan = rxq_to_chan(&adap->sge, queue); sge 1590 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c chan = rxq_to_chan(&adap->sge, queue); sge 2056 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ethrxq(&adap->sge, i) sge 2057 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c disable_txq_db(&adap->sge.ethtxq[i].q); sge 2060 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.uld_txq_info[CXGB4_TX_OFLD]; sge 2063 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ofldtxq(&adap->sge, i) { sge 2071 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c disable_txq_db(&adap->sge.ctrlq[i].q); sge 2078 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ethrxq(&adap->sge, i) sge 2079 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c enable_txq_db(adap, &adap->sge.ethtxq[i].q); sge 2082 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.uld_txq_info[CXGB4_TX_OFLD]; sge 2085 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ofldtxq(&adap->sge, i) { sge 2093 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c enable_txq_db(adap, &adap->sge.ctrlq[i].q); sge 2160 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ethrxq(&adap->sge, i) sge 2161 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sync_txq_pidx(adap, &adap->sge.ethtxq[i].q); sge 2164 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.uld_txq_info[CXGB4_TX_OFLD]; sge 2166 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_ofldtxq(&adap->sge, i) { sge 2174 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sync_txq_pidx(adap, &adap->sge.ctrlq[i].q); sge 3046 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset]; sge 3742 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ret = t4_cfg_pfvf(adap, adap->mbox, adap->pf, 0, adap->sge.egr_sz, 64, sge 4485 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.dbqtimer_tick = val[0]; sge 4487 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ARRAY_SIZE(adap->sge.dbqtimer_val), sge 4488 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.dbqtimer_val); sge 4519 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.egr_start = val[0]; sge 4524 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.ingr_start = val[5]; sge 4551 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.egr_sz = val[0] - adap->sge.egr_start + 1; sge 4552 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.ingr_sz = val[1] - adap->sge.ingr_start + 1; sge 4554 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.egr_map = kcalloc(adap->sge.egr_sz, sge 4555 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sizeof(*adap->sge.egr_map), GFP_KERNEL); sge 4556 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adap->sge.egr_map) { sge 4561 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.ingr_map = kcalloc(adap->sge.ingr_sz, sge 4562 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sizeof(*adap->sge.ingr_map), GFP_KERNEL); sge 4563 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adap->sge.ingr_map) { sge 4571 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.starving_fl = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sge 4573 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adap->sge.starving_fl) { sge 4578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.txq_maperr = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sge 4580 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adap->sge.txq_maperr) { sge 4586 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.blocked_fl = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sge 4588 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adap->sge.blocked_fl) { sge 4919 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adap->sge.egr_map); sge 4920 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adap->sge.ingr_map); sge 4921 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adap->sge.starving_fl); sge 4922 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adap->sge.txq_maperr); sge 4924 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adap->sge.blocked_fl); sge 5079 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 5213 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c while (n < adap->sge.ethqsets) sge 5218 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->sge.ethqsets--; sge 5219 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (adap->sge.ethqsets <= n) sge 5277 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge *s = &adap->sge; sge 5445 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->sge.egr_map); sge 5446 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->sge.ingr_map); sge 5447 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->sge.starving_fl); sge 5448 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->sge.txq_maperr); sge 5450 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->sge.blocked_fl); sge 5841 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c bitmap_zero(adapter->sge.blocked_fl, adapter->sge.egr_sz); sge 173 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c struct sge *s = &adapter->sge; sge 144 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 198 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 214 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 243 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 246 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 271 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c adap->sge.uld_rxq_info[uld_type] = rxq_info; sge 346 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 348 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c adap->sge.uld_rxq_info[uld_type] = NULL; sge 357 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 391 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 407 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 441 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 450 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 482 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 508 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c txq_info = adap->sge.uld_txq_info[tx_uld_type]; sge 514 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c adap->sge.uld_txq_info[tx_uld_type] = NULL; sge 526 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c txq_info = adap->sge.uld_txq_info[tx_uld_type]; sge 563 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c adap->sge.uld_txq_info[tx_uld_type] = txq_info; sge 570 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type]; sge 572 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_txq_info *txq_info = adap->sge.uld_txq_info[tx_uld_type]; sge 583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 611 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge *s = &adap->sge; sge 673 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->udb_density = 1 << adap->params.sge.eq_qpp; sge 674 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->ucq_density = 1 << adap->params.sge.iq_qpp; sge 675 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->sge_host_page_size = 1 << (adap->params.sge.hps + 10); sge 684 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->sge_ingpadboundary = adap->sge.fl_align; sge 685 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->sge_egrstatuspagesize = adap->sge.stat_len; sge 686 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c lld->sge_pktshift = adap->sge.pktshift; sge 154 drivers/net/ethernet/chelsio/cxgb4/l2t.c TID_QID_V(adap->sge.fw_evtq.abs_id))); sge 156 drivers/net/ethernet/chelsio/cxgb4/sched.c txq = &adap->sge.ethtxq[pi->first_qset + p->queue]; sge 201 drivers/net/ethernet/chelsio/cxgb4/sched.c txq = &adap->sge.ethtxq[pi->first_qset + p->queue]; sge 149 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adapter->sge; sge 236 drivers/net/ethernet/chelsio/cxgb4/sge.c const struct sge *s = &adapter->sge; sge 320 drivers/net/ethernet/chelsio/cxgb4/sge.c for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { sge 457 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adapter->sge; sge 594 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 603 drivers/net/ethernet/chelsio/cxgb4/sge.c if (test_bit(q->cntxt_id - adap->sge.egr_start, adap->sge.blocked_fl)) sge 681 drivers/net/ethernet/chelsio/cxgb4/sge.c set_bit(q->cntxt_id - adap->sge.egr_start, sge 682 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.starving_fl); sge 909 drivers/net/ethernet/chelsio/cxgb4/sge.c to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; sge 923 drivers/net/ethernet/chelsio/cxgb4/sge.c unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1; sge 926 drivers/net/ethernet/chelsio/cxgb4/sge.c memcpy(sgl->sge, buf, part0); sge 1420 drivers/net/ethernet/chelsio/cxgb4/sge.c q = &adap->sge.ptptxq; sge 1422 drivers/net/ethernet/chelsio/cxgb4/sge.c q = &adap->sge.ethtxq[qidx + pi->first_qset]; sge 1743 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; sge 2151 drivers/net/ethernet/chelsio/cxgb4/sge.c ret = ctrl_xmit(&adap->sge.ctrlq[0], skb); sge 2209 drivers/net/ethernet/chelsio/cxgb4/sge.c set_bit(q->q.cntxt_id - q->adap->sge.egr_start, sge 2210 drivers/net/ethernet/chelsio/cxgb4/sge.c q->adap->sge.txq_maperr); sge 2438 drivers/net/ethernet/chelsio/cxgb4/sge.c return ctrl_xmit(&adap->sge.ctrlq[idx], skb); sge 2441 drivers/net/ethernet/chelsio/cxgb4/sge.c txq_info = adap->sge.uld_txq_info[tx_uld_type]; sge 2567 drivers/net/ethernet/chelsio/cxgb4/sge.c txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD]; sge 2743 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adapter->sge; sge 2892 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adapter->sge; sge 2934 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &q->adap->sge; sge 3141 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adapter->sge; sge 3301 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_rspq *q = &adap->sge.intrq; sge 3304 drivers/net/ethernet/chelsio/cxgb4/sge.c spin_lock(&adap->sge.intrq_lock); sge 3314 drivers/net/ethernet/chelsio/cxgb4/sge.c qid -= adap->sge.ingr_start; sge 3315 drivers/net/ethernet/chelsio/cxgb4/sge.c napi_schedule(&adap->sge.ingr_map[qid]->napi); sge 3334 drivers/net/ethernet/chelsio/cxgb4/sge.c spin_unlock(&adap->sge.intrq_lock); sge 3388 drivers/net/ethernet/chelsio/cxgb4/sge.c struct adapter *adap = from_timer(adap, t, sge.rx_timer); sge 3389 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3423 drivers/net/ethernet/chelsio/cxgb4/sge.c struct adapter *adap = from_timer(adap, t, sge.tx_timer); sge 3424 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3518 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3630 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.ingr_map[iq->cntxt_id - adap->sge.ingr_start] = iq; sge 3637 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.egr_map[fl->cntxt_id - adap->sge.egr_start] = fl; sge 3716 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.egr_map[id - adap->sge.egr_start] = q; sge 3734 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3819 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3890 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3949 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3962 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 3965 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.ingr_map[rq->cntxt_id - adap->sge.ingr_start] = NULL; sge 4015 drivers/net/ethernet/chelsio/cxgb4/sge.c for (i = 0; i < adap->sge.ethqsets; i++) { sge 4016 drivers/net/ethernet/chelsio/cxgb4/sge.c eq = &adap->sge.ethrxq[i]; sge 4026 drivers/net/ethernet/chelsio/cxgb4/sge.c for (i = 0; i < adap->sge.ethqsets; i++) { sge 4027 drivers/net/ethernet/chelsio/cxgb4/sge.c eq = &adap->sge.ethrxq[i]; sge 4032 drivers/net/ethernet/chelsio/cxgb4/sge.c etq = &adap->sge.ethtxq[i]; sge 4045 drivers/net/ethernet/chelsio/cxgb4/sge.c for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) { sge 4046 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_ctrl_txq *cq = &adap->sge.ctrlq[i]; sge 4057 drivers/net/ethernet/chelsio/cxgb4/sge.c if (adap->sge.fw_evtq.desc) sge 4058 drivers/net/ethernet/chelsio/cxgb4/sge.c free_rspq_fl(adap, &adap->sge.fw_evtq, NULL); sge 4060 drivers/net/ethernet/chelsio/cxgb4/sge.c if (adap->sge.intrq.desc) sge 4061 drivers/net/ethernet/chelsio/cxgb4/sge.c free_rspq_fl(adap, &adap->sge.intrq, NULL); sge 4064 drivers/net/ethernet/chelsio/cxgb4/sge.c etq = &adap->sge.ptptxq; sge 4077 drivers/net/ethernet/chelsio/cxgb4/sge.c memset(adap->sge.egr_map, 0, sge 4078 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.egr_sz * sizeof(*adap->sge.egr_map)); sge 4083 drivers/net/ethernet/chelsio/cxgb4/sge.c adap->sge.ethtxq_rover = 0; sge 4084 drivers/net/ethernet/chelsio/cxgb4/sge.c mod_timer(&adap->sge.rx_timer, jiffies + RX_QCHECK_PERIOD); sge 4085 drivers/net/ethernet/chelsio/cxgb4/sge.c mod_timer(&adap->sge.tx_timer, jiffies + TX_QCHECK_PERIOD); sge 4099 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 4112 drivers/net/ethernet/chelsio/cxgb4/sge.c txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD]; sge 4116 drivers/net/ethernet/chelsio/cxgb4/sge.c for_each_ofldtxq(&adap->sge, i) { sge 4126 drivers/net/ethernet/chelsio/cxgb4/sge.c txq_info = adap->sge.uld_txq_info[CXGB4_TX_CRYPTO]; sge 4130 drivers/net/ethernet/chelsio/cxgb4/sge.c for_each_ofldtxq(&adap->sge, i) { sge 4155 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 4252 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge *s = &adap->sge; sge 200 drivers/net/ethernet/chelsio/cxgb4/smt.c TID_QID_V(adapter->sge.fw_evtq.abs_id))); sge 91 drivers/net/ethernet/chelsio/cxgb4/srq.c TID_QID_V(adap->sge.fw_evtq.abs_id))); sge 9263 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c page_shift = adapter->params.sge.hps + 10; sge 9269 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ? adapter->params.sge.eq_qpp sge 9270 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c : adapter->params.sge.iq_qpp); sge 9372 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c struct sge_params *sge_params = &adapter->params.sge; sge 1509 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h struct ulptx_sge_pair sge[0]; sge 349 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h #define for_each_ethrxq(sge, iter) \ sge 350 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h for (iter = 0; iter < (sge)->ethqsets; iter++) sge 383 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h struct sge sge; sge 384 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 422 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 455 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 478 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 536 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 592 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 707 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; sge 1021 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c static int closest_timer(const struct sge *s, int us) sge 1037 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c static int closest_thres(const struct sge *s, int thres) sge 1062 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c ? adapter->sge.timer_val[timer_idx] sge 1099 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c pktcnt_idx = closest_thres(&adapter->sge, cnt); sge 1118 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c : closest_timer(&adapter->sge, us)); sge 1241 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c rxq = &adapter->sge.ethrxq[pi->first_qset]; sge 1600 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge *s = &pi->adapter->sge; sge 1624 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 1657 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_rspq *rspq = &adapter->sge.ethrxq[pi->first_qset].rspq; sge 1662 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c ? adapter->sge.counter_val[rspq->pktcnt_idx] sge 1679 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c &adapter->sge.ethrxq[pi->first_qset].rspq, sge 1804 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[pi->first_qset]; sge 1805 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; sge 2056 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL); sge 2074 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; sge 2075 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL]; sge 2076 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c int n = min(QPL, adapter->sge.ethqsets - QPL * r); sge 2098 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.counter_val[rxq[qs].rspq.pktcnt_idx]); sge 2111 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_rspq *evtq = &adapter->sge.fw_evtq; sge 2118 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.counter_val[evtq->pktcnt_idx]); sge 2122 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_rspq *intrq = &adapter->sge.intrq; sge 2129 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.counter_val[intrq->pktcnt_idx]); sge 2152 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 + sge 2209 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL); sge 2231 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; sge 2232 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL]; sge 2233 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c int n = min(QPL, adapter->sge.ethqsets - QPL * r); sge 2261 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_rspq *evtq = &adapter->sge.fw_evtq; sge 2269 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_rspq *intrq = &adapter->sge.intrq; sge 2298 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 + sge 2563 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.max_ethqsets = ethqsets; sge 2565 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (adapter->sge.max_ethqsets < adapter->params.nports) { sge 2568 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.max_ethqsets, adapter->params.nports); sge 2569 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->params.nports = adapter->sge.max_ethqsets; sge 2580 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_params *sge_params = &adapter->params.sge; sge 2581 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 2733 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 2760 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c q10g = (adapter->sge.max_ethqsets - n1g) / n10g; sge 2837 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c while (n < adapter->sge.ethqsets) sge 2842 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c adapter->sge.ethqsets--; sge 2843 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (adapter->sge.ethqsets <= n) sge 2870 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge *s = &adapter->sge; sge 264 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const struct sge *s = &adapter->sge; sge 325 drivers/net/ethernet/chelsio/cxgb4vf/sge.c for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { sge 457 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const struct sge *s = &adapter->sge; sge 607 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 715 drivers/net/ethernet/chelsio/cxgb4vf/sge.c set_bit(fl->cntxt_id, adapter->sge.starving_fl); sge 930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; sge 944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; sge 947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c memcpy(sgl->sge, buf, part0); sge 1198 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; sge 1571 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 1623 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 1761 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 1939 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 1944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c spin_lock(&adapter->sge.intrq_lock); sge 2021 drivers/net/ethernet/chelsio/cxgb4vf/sge.c spin_unlock(&adapter->sge.intrq_lock); sge 2068 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct adapter *adapter = from_timer(adapter, t, sge.rx_timer); sge 2069 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2127 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct adapter *adapter = from_timer(adapter, t, sge.tx_timer); sge 2128 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2209 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2223 drivers/net/ethernet/chelsio/cxgb4vf/sge.c rspq != &adapter->sge.intrq) { sge 2225 drivers/net/ethernet/chelsio/cxgb4vf/sge.c intr_dest = adapter->sge.intrq.abs_id; sge 2419 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2515 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2532 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2565 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2572 drivers/net/ethernet/chelsio/cxgb4vf/sge.c for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) { sge 2596 drivers/net/ethernet/chelsio/cxgb4vf/sge.c adapter->sge.ethtxq_rover = 0; sge 2597 drivers/net/ethernet/chelsio/cxgb4vf/sge.c mod_timer(&adapter->sge.rx_timer, jiffies + RX_QCHECK_PERIOD); sge 2598 drivers/net/ethernet/chelsio/cxgb4vf/sge.c mod_timer(&adapter->sge.tx_timer, jiffies + TX_QCHECK_PERIOD); sge 2611 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 2630 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_params *sge_params = &adapter->params.sge; sge 2633 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge *s = &adapter->sge; sge 262 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h struct sge_params sge; /* Scatter Gather Engine */ sge 703 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c sge_control = adapter->params.sge.sge_control; sge 729 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c sge_control2 = adapter->params.sge.sge_control2; sge 784 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c page_shift = adapter->params.sge.sge_vf_hps + 10; sge 790 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c ? adapter->params.sge.sge_vf_eq_qpp sge 791 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c : adapter->params.sge.sge_vf_iq_qpp); sge 850 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c struct sge_params *sge_params = &adapter->params.sge; sge 797 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_sge *sge; sge 808 drivers/net/ethernet/emulex/benet/be_cmds.c sge = nonembedded_sgl(wrb); sge 809 drivers/net/ethernet/emulex/benet/be_cmds.c sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma)); sge 810 drivers/net/ethernet/emulex/benet/be_cmds.c sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF); sge 811 drivers/net/ethernet/emulex/benet/be_cmds.c sge->len = cpu_to_le32(mem->size); sge 138 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_sg_entry *sge = &sgt[i]; sge 145 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c sg_addr = dpaa2_sg_get_addr(sge); sge 150 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c sg_length = dpaa2_sg_get_len(sge); sge 170 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c sg_offset = dpaa2_sg_get_offset(sge); sge 191 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c if (dpaa2_sg_is_final(sge)) sge 55 drivers/net/ethernet/huawei/hinic/hinic_common.c void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len) sge 57 drivers/net/ethernet/huawei/hinic/hinic_common.c sge->hi_addr = upper_32_bits(addr); sge 58 drivers/net/ethernet/huawei/hinic/hinic_common.c sge->lo_addr = lower_32_bits(addr); sge 59 drivers/net/ethernet/huawei/hinic/hinic_common.c sge->len = len; sge 68 drivers/net/ethernet/huawei/hinic/hinic_common.c dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge) sge 70 drivers/net/ethernet/huawei/hinic/hinic_common.c return (dma_addr_t)((((u64)sge->hi_addr) << 32) | sge->lo_addr); sge 25 drivers/net/ethernet/huawei/hinic/hinic_common.h void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len); sge 27 drivers/net/ethernet/huawei/hinic/hinic_common.h dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge); sge 179 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c hinic_set_sge(&sge_resp->sge, buf_out->dma_addr, buf_out->size); sge 236 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size); sge 603 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sq_wqe->buf_descs[i].sge = sges[i]; sge 781 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sges[i] = sq_wqe->buf_descs[i].sge; sge 921 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c u16 cons_idx, struct hinic_sge *sge) sge 926 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sge->hi_addr = be32_to_cpu(rq_wqe->buf_desc.hi_addr); sge 927 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sge->lo_addr = be32_to_cpu(rq_wqe->buf_desc.lo_addr); sge 928 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sge->len = HINIC_RQ_CQE_SGE_GET(len, LEN); sge 939 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c struct hinic_rq_wqe *rq_wqe, struct hinic_sge *sge) sge 955 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c hinic_set_sge(&cqe_sect->sge, cqe_dma, sizeof(*cqe)); sge 957 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c buf_desc->hi_addr = sge->hi_addr; sge 958 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c buf_desc->lo_addr = sge->lo_addr; sge 217 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h u16 cons_idx, struct hinic_sge *sge); sge 220 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h struct hinic_rq_wqe *wqe, struct hinic_sge *sge); sge 334 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h struct hinic_sge sge; sge 353 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h struct hinic_sge sge; sge 403 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h struct hinic_sge sge; sge 430 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h struct hinic_sge sge; sge 201 drivers/net/ethernet/huawei/hinic/hinic_rx.c struct hinic_sge sge; sge 220 drivers/net/ethernet/huawei/hinic/hinic_rx.c hinic_set_sge(&sge, dma_addr, skb->len); sge 229 drivers/net/ethernet/huawei/hinic/hinic_rx.c hinic_rq_prepare_wqe(rxq->rq, prod_idx, rq_wqe, &sge); sge 252 drivers/net/ethernet/huawei/hinic/hinic_rx.c struct hinic_sge sge; sge 259 drivers/net/ethernet/huawei/hinic/hinic_rx.c hinic_rq_get_sge(rq, &hw_wqe->rq_wqe, ci, &sge); sge 263 drivers/net/ethernet/huawei/hinic/hinic_rx.c rx_free_skb(rxq, rq->saved_skb[ci], hinic_sge_to_dma(&sge)); sge 282 drivers/net/ethernet/huawei/hinic/hinic_rx.c struct hinic_sge sge; sge 291 drivers/net/ethernet/huawei/hinic/hinic_rx.c hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge); sge 293 drivers/net/ethernet/huawei/hinic/hinic_rx.c rx_unmap_skb(rxq, hinic_sge_to_dma(&sge)); sge 336 drivers/net/ethernet/huawei/hinic/hinic_rx.c struct hinic_sge sge; sge 358 drivers/net/ethernet/huawei/hinic/hinic_rx.c hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge); sge 360 drivers/net/ethernet/huawei/hinic/hinic_rx.c rx_unmap_skb(rxq, hinic_sge_to_dma(&sge)); sge 366 drivers/net/ethernet/huawei/hinic/hinic_rx.c pkt_len = sge.len; sge 859 drivers/net/ethernet/ibm/ibmvnic.h struct ibmvnic_sge_desc sge; sge 678 drivers/nvme/host/pci.c static void nvme_pci_sgl_set_data(struct nvme_sgl_desc *sge, sge 681 drivers/nvme/host/pci.c sge->addr = cpu_to_le64(sg_dma_address(sg)); sge 682 drivers/nvme/host/pci.c sge->length = cpu_to_le32(sg_dma_len(sg)); sge 683 drivers/nvme/host/pci.c sge->type = NVME_SGL_FMT_DATA_DESC << 4; sge 686 drivers/nvme/host/pci.c static void nvme_pci_sgl_set_seg(struct nvme_sgl_desc *sge, sge 689 drivers/nvme/host/pci.c sge->addr = cpu_to_le64(dma_addr); sge 691 drivers/nvme/host/pci.c sge->length = cpu_to_le32(entries * sizeof(*sge)); sge 692 drivers/nvme/host/pci.c sge->type = NVME_SGL_FMT_LAST_SEG_DESC << 4; sge 694 drivers/nvme/host/pci.c sge->length = cpu_to_le32(PAGE_SIZE); sge 695 drivers/nvme/host/pci.c sge->type = NVME_SGL_FMT_SEG_DESC << 4; sge 59 drivers/nvme/host/rdma.c struct ib_sge sge[1 + NVME_RDMA_MAX_INLINE_SEGMENTS]; sge 1187 drivers/nvme/host/rdma.c struct ib_sge *sge = &req->sge[1]; sge 1191 drivers/nvme/host/rdma.c for (i = 0; i < count; i++, sgl++, sge++) { sge 1192 drivers/nvme/host/rdma.c sge->addr = sg_dma_address(sgl); sge 1193 drivers/nvme/host/rdma.c sge->length = sg_dma_len(sgl); sge 1194 drivers/nvme/host/rdma.c sge->lkey = queue->device->pd->local_dma_lkey; sge 1195 drivers/nvme/host/rdma.c len += sge->length; sge 1343 drivers/nvme/host/rdma.c struct nvme_rdma_qe *qe, struct ib_sge *sge, u32 num_sge, sge 1349 drivers/nvme/host/rdma.c sge->addr = qe->dma; sge 1350 drivers/nvme/host/rdma.c sge->length = sizeof(struct nvme_command), sge 1351 drivers/nvme/host/rdma.c sge->lkey = queue->device->pd->local_dma_lkey; sge 1355 drivers/nvme/host/rdma.c wr.sg_list = sge; sge 1421 drivers/nvme/host/rdma.c struct ib_sge sge; sge 1437 drivers/nvme/host/rdma.c ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL); sge 1782 drivers/nvme/host/rdma.c err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge, sge 35 drivers/nvme/target/rdma.c struct ib_sge sge[NVMET_RDMA_MAX_INLINE_SGE + 1]; sge 218 drivers/nvme/target/rdma.c struct ib_sge *sge; sge 225 drivers/nvme/target/rdma.c sge = &c->sge[1]; sge 227 drivers/nvme/target/rdma.c for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) { sge 228 drivers/nvme/target/rdma.c if (sge->length) sge 229 drivers/nvme/target/rdma.c ib_dma_unmap_page(ndev->device, sge->addr, sge 230 drivers/nvme/target/rdma.c sge->length, DMA_FROM_DEVICE); sge 240 drivers/nvme/target/rdma.c struct ib_sge *sge; sge 250 drivers/nvme/target/rdma.c sge = &c->sge[1]; sge 253 drivers/nvme/target/rdma.c for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) { sge 258 drivers/nvme/target/rdma.c sge->addr = ib_dma_map_page(ndev->device, sge 260 drivers/nvme/target/rdma.c if (ib_dma_mapping_error(ndev->device, sge->addr)) sge 262 drivers/nvme/target/rdma.c sge->length = min_t(int, len, PAGE_SIZE); sge 263 drivers/nvme/target/rdma.c sge->lkey = ndev->pd->local_dma_lkey; sge 264 drivers/nvme/target/rdma.c len -= sge->length; sge 269 drivers/nvme/target/rdma.c for (; i >= 0; i--, sg--, sge--) { sge 270 drivers/nvme/target/rdma.c if (sge->length) sge 271 drivers/nvme/target/rdma.c ib_dma_unmap_page(ndev->device, sge->addr, sge 272 drivers/nvme/target/rdma.c sge->length, DMA_FROM_DEVICE); sge 287 drivers/nvme/target/rdma.c c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd, sge 289 drivers/nvme/target/rdma.c if (ib_dma_mapping_error(ndev->device, c->sge[0].addr)) sge 292 drivers/nvme/target/rdma.c c->sge[0].length = sizeof(*c->nvme_cmd); sge 293 drivers/nvme/target/rdma.c c->sge[0].lkey = ndev->pd->local_dma_lkey; sge 301 drivers/nvme/target/rdma.c c->wr.sg_list = c->sge; sge 307 drivers/nvme/target/rdma.c ib_dma_unmap_single(ndev->device, c->sge[0].addr, sge 321 drivers/nvme/target/rdma.c ib_dma_unmap_single(ndev->device, c->sge[0].addr, sge 461 drivers/nvme/target/rdma.c cmd->sge[0].addr, cmd->sge[0].length, sge 761 drivers/nvme/target/rdma.c cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length, sge 4057 drivers/scsi/aacraid/aachba.c rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32)); sge 4058 drivers/scsi/aacraid/aachba.c rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff)); sge 4060 drivers/scsi/aacraid/aachba.c rio2->sge[i].length = cur_size; sge 4061 drivers/scsi/aacraid/aachba.c rio2->sge[i].flags = 0; sge 4078 drivers/scsi/aacraid/aachba.c u32 temp = le32_to_cpu(rio2->sge[i-1].length) - sge 4080 drivers/scsi/aacraid/aachba.c rio2->sge[i-1].length = cpu_to_le32(temp); sge 4093 drivers/scsi/aacraid/aachba.c if (rio2->sge[j].length % (i*PAGE_SIZE)) { sge 4097 drivers/scsi/aacraid/aachba.c nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE)); sge 4122 drivers/scsi/aacraid/aachba.c struct sge_ieee1212 *sge; sge 4129 drivers/scsi/aacraid/aachba.c sge = kmalloc_array(nseg_new, sizeof(struct sge_ieee1212), GFP_ATOMIC); sge 4130 drivers/scsi/aacraid/aachba.c if (sge == NULL) sge 4134 drivers/scsi/aacraid/aachba.c for (j = 0; j < rio2->sge[i].length / (pages * PAGE_SIZE); ++j) { sge 4135 drivers/scsi/aacraid/aachba.c addr_low = rio2->sge[i].addrLow + j * pages * PAGE_SIZE; sge 4136 drivers/scsi/aacraid/aachba.c sge[pos].addrLow = addr_low; sge 4137 drivers/scsi/aacraid/aachba.c sge[pos].addrHigh = rio2->sge[i].addrHigh; sge 4138 drivers/scsi/aacraid/aachba.c if (addr_low < rio2->sge[i].addrLow) sge 4139 drivers/scsi/aacraid/aachba.c sge[pos].addrHigh++; sge 4140 drivers/scsi/aacraid/aachba.c sge[pos].length = pages * PAGE_SIZE; sge 4141 drivers/scsi/aacraid/aachba.c sge[pos].flags = 0; sge 4145 drivers/scsi/aacraid/aachba.c sge[pos] = rio2->sge[nseg-1]; sge 4146 drivers/scsi/aacraid/aachba.c memcpy(&rio2->sge[1], &sge[1], (nseg_new-1)*sizeof(struct sge_ieee1212)); sge 4148 drivers/scsi/aacraid/aachba.c kfree(sge); sge 4165 drivers/scsi/aacraid/aachba.c struct aac_hba_sgl *sge; sge 4174 drivers/scsi/aacraid/aachba.c sge = &hbacmd->sge[2]; sge 4176 drivers/scsi/aacraid/aachba.c sge = &hbacmd->sge[0]; sge 4183 drivers/scsi/aacraid/aachba.c sge->addr_hi = cpu_to_le32((u32)(addr>>32)); sge 4184 drivers/scsi/aacraid/aachba.c sge->addr_lo = cpu_to_le32((u32)(addr & 0xffffffff)); sge 4186 drivers/scsi/aacraid/aachba.c sge->len = cur_size; sge 4187 drivers/scsi/aacraid/aachba.c sge->flags = 0; sge 4189 drivers/scsi/aacraid/aachba.c sge++; sge 4192 drivers/scsi/aacraid/aachba.c sge--; sge 4197 drivers/scsi/aacraid/aachba.c temp = le32_to_cpu(sge->len) - byte_count sge 4199 drivers/scsi/aacraid/aachba.c sge->len = cpu_to_le32(temp); sge 4205 drivers/scsi/aacraid/aachba.c sge->flags = cpu_to_le32(0x40000000); sge 4208 drivers/scsi/aacraid/aachba.c hbacmd->sge[0].flags = cpu_to_le32(0x80000000); sge 4210 drivers/scsi/aacraid/aachba.c hbacmd->sge[0].addr_hi = (u32)cpu_to_le32(sg_address >> 32); sge 4211 drivers/scsi/aacraid/aachba.c hbacmd->sge[0].addr_lo = sge 225 drivers/scsi/aacraid/aacraid.h struct aac_hba_sgl sge[HBA_MAX_SG_SEPARATE+2]; /* SG list space */ sge 1928 drivers/scsi/aacraid/aacraid.h struct sge_ieee1212 sge[1]; sge 682 drivers/scsi/aacraid/commctrl.c hbacmd->sge[i].addr_hi = cpu_to_le32((u32)(addr>>32)); sge 683 drivers/scsi/aacraid/commctrl.c hbacmd->sge[i].addr_lo = cpu_to_le32( sge 685 drivers/scsi/aacraid/commctrl.c hbacmd->sge[i].len = cpu_to_le32(sg_count[i]); sge 686 drivers/scsi/aacraid/commctrl.c hbacmd->sge[i].flags = 0; sge 691 drivers/scsi/aacraid/commctrl.c hbacmd->sge[usg32->count-1].flags = sge 184 drivers/scsi/aacraid/commsup.c offsetof(struct aac_hba_cmd_req, sge[2]); sge 1264 drivers/scsi/be2iscsi/be_cmds.c struct be_sge *sge = nonembedded_sgl(wrb); sge 1284 drivers/scsi/be2iscsi/be_cmds.c sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); sge 1285 drivers/scsi/be2iscsi/be_cmds.c sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF); sge 1286 drivers/scsi/be2iscsi/be_cmds.c sge->len = cpu_to_le32(nonemb_cmd.size); sge 127 drivers/scsi/be2iscsi/be_mgmt.c struct be_sge *sge; sge 154 drivers/scsi/be2iscsi/be_mgmt.c sge = nonembedded_sgl(wrb); sge 197 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); sge 198 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); sge 199 drivers/scsi/be2iscsi/be_mgmt.c sge->len = cpu_to_le32(nonemb_cmd->size); sge 229 drivers/scsi/be2iscsi/be_mgmt.c struct be_sge *sge; sge 241 drivers/scsi/be2iscsi/be_mgmt.c sge = nonembedded_sgl(wrb); sge 243 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); sge 244 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd->dma)); sge 245 drivers/scsi/be2iscsi/be_mgmt.c sge->len = cpu_to_le32(nonemb_cmd->size); sge 1004 drivers/scsi/be2iscsi/be_mgmt.c struct be_sge *sge; sge 1027 drivers/scsi/be2iscsi/be_mgmt.c sge = nonembedded_sgl(wrb); sge 1033 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); sge 1034 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); sge 1035 drivers/scsi/be2iscsi/be_mgmt.c sge->len = cpu_to_le32(nonemb_cmd->size); sge 1506 drivers/scsi/be2iscsi/be_mgmt.c struct be_sge *sge; sge 1543 drivers/scsi/be2iscsi/be_mgmt.c sge = nonembedded_sgl(wrb); sge 1544 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); sge 1545 drivers/scsi/be2iscsi/be_mgmt.c sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd.dma)); sge 1546 drivers/scsi/be2iscsi/be_mgmt.c sge->len = cpu_to_le32(nonemb_cmd.size); sge 2454 drivers/scsi/bfa/bfa_fcpim.c struct bfi_sge_s *sge, *sgpge; sge 2482 drivers/scsi/bfa/bfa_fcpim.c sge = &m->sges[0]; sge 2491 drivers/scsi/bfa/bfa_fcpim.c sge->sga = *(union bfi_addr_u *) &addr; sge 2493 drivers/scsi/bfa/bfa_fcpim.c sge->sg_len = pgdlen; sge 2494 drivers/scsi/bfa/bfa_fcpim.c sge->flags = (ioim->nsges > BFI_SGE_INLINE) ? sge 2496 drivers/scsi/bfa/bfa_fcpim.c bfa_sge_to_be(sge); sge 2497 drivers/scsi/bfa/bfa_fcpim.c sge++; sge 2538 drivers/scsi/bfa/bfa_fcpim.c sge->sga = ioim->sgpg->sgpg_pa; sge 2540 drivers/scsi/bfa/bfa_fcpim.c sge->sga.a32.addr_lo = 0; sge 2541 drivers/scsi/bfa/bfa_fcpim.c sge->sga.a32.addr_hi = 0; sge 2543 drivers/scsi/bfa/bfa_fcpim.c sge->sg_len = pgdlen; sge 2544 drivers/scsi/bfa/bfa_fcpim.c sge->flags = BFI_SGE_PGDLEN; sge 2545 drivers/scsi/bfa/bfa_fcpim.c bfa_sge_to_be(sge); sge 3184 drivers/scsi/bfa/bfad_bsg.c struct bfa_sge_s *sge; sge 3187 drivers/scsi/bfa/bfad_bsg.c sge = drv_fcxp->req_sge + sgeid; sge 3188 drivers/scsi/bfa/bfad_bsg.c addr = (u64)(size_t) sge->sg_addr; sge 3196 drivers/scsi/bfa/bfad_bsg.c struct bfa_sge_s *sge; sge 3198 drivers/scsi/bfa/bfad_bsg.c sge = drv_fcxp->req_sge + sgeid; sge 3199 drivers/scsi/bfa/bfad_bsg.c return sge->sg_len; sge 3206 drivers/scsi/bfa/bfad_bsg.c struct bfa_sge_s *sge; sge 3209 drivers/scsi/bfa/bfad_bsg.c sge = drv_fcxp->rsp_sge + sgeid; sge 3210 drivers/scsi/bfa/bfad_bsg.c addr = (u64)(size_t) sge->sg_addr; sge 3218 drivers/scsi/bfa/bfad_bsg.c struct bfa_sge_s *sge; sge 3220 drivers/scsi/bfa/bfad_bsg.c sge = drv_fcxp->rsp_sge + sgeid; sge 3221 drivers/scsi/bfa/bfad_bsg.c return sge->sg_len; sge 58 drivers/scsi/csiostor/csio_wr.c csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg) sge 60 drivers/scsi/csiostor/csio_wr.c sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A + sge 66 drivers/scsi/csiostor/csio_wr.c csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf) sge 68 drivers/scsi/csiostor/csio_wr.c return sge->sge_fl_buf_size[buf->paddr & 0xF]; sge 75 drivers/scsi/csiostor/csio_wr.c return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ? 128 : 64; sge 118 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 126 drivers/scsi/csiostor/csio_wr.c buf->len = sge->sge_fl_buf_size[sreg]; sge 1049 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 1075 drivers/scsi/csiostor/csio_wr.c bufsz = csio_wr_fl_bufsz(sge, buf); sge 1096 drivers/scsi/csiostor/csio_wr.c flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align); sge 1312 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 1418 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = fl_align; sge 1435 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 1437 drivers/scsi/csiostor/csio_wr.c csio_sge_thresh_reg = csio_closest_thresh(sge, csio_intr_coalesce_cnt); sge 1444 drivers/scsi/csiostor/csio_wr.c csio_sge_timer_reg = csio_closest_timer(sge, csio_intr_coalesce_time); sge 1457 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 1463 drivers/scsi/csiostor/csio_wr.c sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A); sge 1465 drivers/scsi/csiostor/csio_wr.c ingpad = INGPADBOUNDARY_G(sge->sge_control); sge 1469 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 32; break; sge 1471 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 64; break; sge 1473 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 128; break; sge 1475 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 256; break; sge 1477 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 512; break; sge 1479 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 1024; break; sge 1481 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 2048; break; sge 1483 drivers/scsi/csiostor/csio_wr.c sge->csio_fl_align = 4096; break; sge 1487 drivers/scsi/csiostor/csio_wr.c csio_get_flbuf_size(hw, sge, i); sge 1493 drivers/scsi/csiostor/csio_wr.c sge->timer_val[0] = (uint16_t)csio_core_ticks_to_us(hw, sge 1495 drivers/scsi/csiostor/csio_wr.c sge->timer_val[1] = (uint16_t)csio_core_ticks_to_us(hw, sge 1497 drivers/scsi/csiostor/csio_wr.c sge->timer_val[2] = (uint16_t)csio_core_ticks_to_us(hw, sge 1499 drivers/scsi/csiostor/csio_wr.c sge->timer_val[3] = (uint16_t)csio_core_ticks_to_us(hw, sge 1501 drivers/scsi/csiostor/csio_wr.c sge->timer_val[4] = (uint16_t)csio_core_ticks_to_us(hw, sge 1503 drivers/scsi/csiostor/csio_wr.c sge->timer_val[5] = (uint16_t)csio_core_ticks_to_us(hw, sge 1507 drivers/scsi/csiostor/csio_wr.c sge->counter_val[0] = THRESHOLD_0_G(ingress_rx_threshold); sge 1508 drivers/scsi/csiostor/csio_wr.c sge->counter_val[1] = THRESHOLD_1_G(ingress_rx_threshold); sge 1509 drivers/scsi/csiostor/csio_wr.c sge->counter_val[2] = THRESHOLD_2_G(ingress_rx_threshold); sge 1510 drivers/scsi/csiostor/csio_wr.c sge->counter_val[3] = THRESHOLD_3_G(ingress_rx_threshold); sge 1526 drivers/scsi/csiostor/csio_wr.c struct csio_sge *sge = &wrm->sge; sge 1535 drivers/scsi/csiostor/csio_wr.c sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A); sge 1556 drivers/scsi/csiostor/csio_wr.c csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE2 + sge->csio_fl_align - 1) sge 1557 drivers/scsi/csiostor/csio_wr.c & ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE2_A); sge 1558 drivers/scsi/csiostor/csio_wr.c csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE3 + sge->csio_fl_align - 1) sge 1559 drivers/scsi/csiostor/csio_wr.c & ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE3_A); sge 1567 drivers/scsi/csiostor/csio_wr.c csio_get_flbuf_size(hw, sge, i); sge 1570 drivers/scsi/csiostor/csio_wr.c sge->timer_val[0] = CSIO_SGE_TIMER_VAL_0; sge 1571 drivers/scsi/csiostor/csio_wr.c sge->timer_val[1] = CSIO_SGE_TIMER_VAL_1; sge 1572 drivers/scsi/csiostor/csio_wr.c sge->timer_val[2] = CSIO_SGE_TIMER_VAL_2; sge 1573 drivers/scsi/csiostor/csio_wr.c sge->timer_val[3] = CSIO_SGE_TIMER_VAL_3; sge 1574 drivers/scsi/csiostor/csio_wr.c sge->timer_val[4] = CSIO_SGE_TIMER_VAL_4; sge 1575 drivers/scsi/csiostor/csio_wr.c sge->timer_val[5] = CSIO_SGE_TIMER_VAL_5; sge 1577 drivers/scsi/csiostor/csio_wr.c sge->counter_val[0] = CSIO_SGE_INT_CNT_VAL_0; sge 1578 drivers/scsi/csiostor/csio_wr.c sge->counter_val[1] = CSIO_SGE_INT_CNT_VAL_1; sge 1579 drivers/scsi/csiostor/csio_wr.c sge->counter_val[2] = CSIO_SGE_INT_CNT_VAL_2; sge 1580 drivers/scsi/csiostor/csio_wr.c sge->counter_val[3] = CSIO_SGE_INT_CNT_VAL_3; sge 1582 drivers/scsi/csiostor/csio_wr.c csio_wr_reg32(hw, THRESHOLD_0_V(sge->counter_val[0]) | sge 1583 drivers/scsi/csiostor/csio_wr.c THRESHOLD_1_V(sge->counter_val[1]) | sge 1584 drivers/scsi/csiostor/csio_wr.c THRESHOLD_2_V(sge->counter_val[2]) | sge 1585 drivers/scsi/csiostor/csio_wr.c THRESHOLD_3_V(sge->counter_val[3]), sge 1589 drivers/scsi/csiostor/csio_wr.c TIMERVALUE0_V(csio_us_to_core_ticks(hw, sge->timer_val[0])) | sge 1590 drivers/scsi/csiostor/csio_wr.c TIMERVALUE1_V(csio_us_to_core_ticks(hw, sge->timer_val[1])), sge 1594 drivers/scsi/csiostor/csio_wr.c TIMERVALUE2_V(csio_us_to_core_ticks(hw, sge->timer_val[2])) | sge 1595 drivers/scsi/csiostor/csio_wr.c TIMERVALUE3_V(csio_us_to_core_ticks(hw, sge->timer_val[3])), sge 1599 drivers/scsi/csiostor/csio_wr.c TIMERVALUE4_V(csio_us_to_core_ticks(hw, sge->timer_val[4])) | sge 1600 drivers/scsi/csiostor/csio_wr.c TIMERVALUE5_V(csio_us_to_core_ticks(hw, sge->timer_val[5])), sge 449 drivers/scsi/csiostor/csio_wr.h struct csio_sge sge; /* SGE params */ sge 504 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 557 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1072 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1095 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1098 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1120 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1133 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1147 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1174 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[1]; sge 1191 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge; sge 1210 drivers/scsi/esas2r/atvda.h struct atto_vda_sge sge[2]; sge 619 drivers/scsi/esas2r/esas2r.h } sge; sge 1178 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq sge 1185 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.last = sge 1186 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.curr = first; sge 1191 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.last = sge 1192 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0]; sge 1194 drivers/scsi/esas2r/esas2r.h (u8)offsetof(struct atto_vda_scsi_req, u.sge); sge 1196 drivers/scsi/esas2r/esas2r.h sgc->sge.a64.chain = NULL; sge 901 drivers/scsi/esas2r/esas2r_disc.c esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge); sge 212 drivers/scsi/esas2r/esas2r_flash.c esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]); sge 915 drivers/scsi/esas2r/esas2r_flash.c esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge); sge 1312 drivers/scsi/esas2r/esas2r_flash.c vrq->data.sge[0].length = sge 1315 drivers/scsi/esas2r/esas2r_flash.c vrq->data.sge[0].address = cpu_to_le64( sge 1229 drivers/scsi/esas2r/esas2r_init.c data.sge); sge 222 drivers/scsi/esas2r/esas2r_io.c if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) { sge 237 drivers/scsi/esas2r/esas2r_io.c sgelen = (u8)((u8 *)sgc->sge.a64.curr sge 238 drivers/scsi/esas2r/esas2r_io.c - (u8 *)sgc->sge.a64.last); sge 244 drivers/scsi/esas2r/esas2r_io.c memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); sge 247 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.curr = sge 252 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.limit = sge 257 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.last->length = cpu_to_le32( sge 259 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.last->address = sge 268 drivers/scsi/esas2r/esas2r_io.c if (sgc->sge.a64.chain) { sge 269 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.chain->length |= sge 271 drivers/scsi/esas2r/esas2r_io.c ((u8 *)(sgc->sge.a64. sge 280 drivers/scsi/esas2r/esas2r_io.c sge.a64.last - sge 299 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.chain = sgc->sge.a64.last; sge 306 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.last = sgc->sge.a64.curr; sge 309 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.curr->length = cpu_to_le32(SGE_ADDR_64 | len); sge 310 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.curr->address = cpu_to_le32(addr); sge 311 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.curr++; sge 328 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.last->length |= cpu_to_le32(SGE_LAST); sge 334 drivers/scsi/esas2r/esas2r_io.c if (sgc->sge.a64.chain) { sge 335 drivers/scsi/esas2r/esas2r_io.c sgc->sge.a64.chain->length |= cpu_to_le32( sge 336 drivers/scsi/esas2r/esas2r_io.c ((u8 *)(sgc->sge.a64.curr) - sge 346 drivers/scsi/esas2r/esas2r_io.c ((u16)((u8 *)sgc->sge.a64.last - (u8 *)vrq) sge 409 drivers/scsi/esas2r/esas2r_io.c if (sgc->sge.prd.sge_cnt == 0) { sge 418 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr->ctl_len = cpu_to_le32( sge 420 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr->address = cpu_to_le64(addr); sge 432 drivers/scsi/esas2r/esas2r_io.c if (sgc->sge.prd.chain) { sge 439 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain->ctl_len |= cpu_to_le32( sge 440 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.sgl_max_cnt); sge 467 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain = sgc->sge.prd.curr; sge 469 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain->ctl_len = cpu_to_le32(PRD_CHAIN); sge 470 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain->address = sge 478 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr = sge 482 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.sge_cnt = sgc->sge.prd.sgl_max_cnt - 1; sge 485 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.sge_cnt--; sge 487 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr->ctl_len = cpu_to_le32(PRD_DATA | len); sge 488 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr->address = cpu_to_le64(addr); sge 492 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr++; sge 513 drivers/scsi/esas2r/esas2r_io.c if (sgc->sge.prd.chain) { sge 514 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain->ctl_len |= sge 515 drivers/scsi/esas2r/esas2r_io.c cpu_to_le32(sgc->sge.prd.sgl_max_cnt sge 516 drivers/scsi/esas2r/esas2r_io.c - sgc->sge.prd.sge_cnt sge 614 drivers/scsi/esas2r/esas2r_io.c (struct atto_physical_region_description *)sgc->sge.a64.curr; sge 616 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.sgl_max_cnt = sgl_page_size / sge 623 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.sge_cnt = 0; sge 624 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.chain = NULL; sge 625 drivers/scsi/esas2r/esas2r_io.c sgc->sge.prd.curr = curr_iblk_chn; sge 300 drivers/scsi/esas2r/esas2r_ioctl.c esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); sge 353 drivers/scsi/esas2r/esas2r_ioctl.c esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); sge 672 drivers/scsi/esas2r/esas2r_ioctl.c esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); sge 119 drivers/scsi/esas2r/esas2r_vda.c firstsg = rq->vrq->flash.data.file.sge; sge 130 drivers/scsi/esas2r/esas2r_vda.c firstsg = rq->vrq->cli.sge; sge 194 drivers/scsi/esas2r/esas2r_vda.c firstsg = rq->vrq->mgt.sge; sge 363 drivers/scsi/esas2r/esas2r_vda.c data.sge); sge 394 drivers/scsi/esas2r/esas2r_vda.c struct atto_vda_mgmt_req, sge); sge 396 drivers/scsi/esas2r/esas2r_vda.c vrq->sge[0].length = cpu_to_le32(SGE_LAST | length); sge 397 drivers/scsi/esas2r/esas2r_vda.c vrq->sge[0].address = cpu_to_le64( sge 432 drivers/scsi/esas2r/esas2r_vda.c (u8)offsetof(struct atto_vda_ae_req, sge); sge 433 drivers/scsi/esas2r/esas2r_vda.c vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length); sge 434 drivers/scsi/esas2r/esas2r_vda.c vrq->sge[0].address = cpu_to_le64( sge 461 drivers/scsi/esas2r/esas2r_vda.c vrq->sg_list_offset = (u8)offsetof(struct atto_vda_cli_req, sge); sge 478 drivers/scsi/esas2r/esas2r_vda.c vrq->sg_list_offset = (u8)offsetof(struct atto_vda_ioctl_req, sge); sge 513 drivers/scsi/hisi_sas/hisi_sas.h struct hisi_sas_sge sge[HISI_SAS_SGE_PAGE_CNT]; sge 518 drivers/scsi/hisi_sas/hisi_sas.h struct hisi_sas_sge sge[HISI_SAS_SGE_DIF_PAGE_CNT]; sge 898 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; sge 1678 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; sge 1059 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; sge 1085 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_sge *entry = &sge_dif_page->sge[i]; sge 4939 drivers/scsi/lpfc/lpfc_bsg.c receive_length = nembed_sge->sge[0].length; sge 4951 drivers/scsi/lpfc/lpfc_bsg.c nembed_sge->sge[0].pa_hi = sge 4954 drivers/scsi/lpfc/lpfc_bsg.c nembed_sge->sge[0].pa_lo = sge 1838 drivers/scsi/lpfc/lpfc_hbadisc.c struct lpfc_mbx_sge sge; sge 1847 drivers/scsi/lpfc/lpfc_hbadisc.c lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); sge 1121 drivers/scsi/lpfc/lpfc_hw4.h struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES]; sge 1694 drivers/scsi/lpfc/lpfc_mbox.c nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr); sge 1695 drivers/scsi/lpfc/lpfc_mbox.c nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr); sge 1696 drivers/scsi/lpfc/lpfc_mbox.c nembed_sge->sge[sgentry].length = length; sge 1709 drivers/scsi/lpfc/lpfc_mbox.c struct lpfc_mbx_sge *sge) sge 1715 drivers/scsi/lpfc/lpfc_mbox.c sge->pa_lo = nembed_sge->sge[sgentry].pa_lo; sge 1716 drivers/scsi/lpfc/lpfc_mbox.c sge->pa_hi = nembed_sge->sge[sgentry].pa_hi; sge 1717 drivers/scsi/lpfc/lpfc_mbox.c sge->length = nembed_sge->sge[sgentry].length; sge 1731 drivers/scsi/lpfc/lpfc_mbox.c struct lpfc_mbx_sge sge; sge 1752 drivers/scsi/lpfc/lpfc_mbox.c lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge); sge 1753 drivers/scsi/lpfc/lpfc_mbox.c phyaddr = getPaddr(sge.pa_hi, sge.pa_lo); sge 2030 drivers/scsi/lpfc/lpfc_mbox.c struct lpfc_mbx_sge sge; sge 2056 drivers/scsi/lpfc/lpfc_mbox.c lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); sge 18520 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_mbx_sge sge; sge 18551 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_mbx_sge_get(mboxq, 0, &sge); sge 188 drivers/scsi/myrs.c sgl->sge[0].sge_addr = ctlr_info_addr; sge 189 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ctlr_info.dma_size; sge 247 drivers/scsi/myrs.c sgl->sge[0].sge_addr = ldev_info_addr; sge 248 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ldev_info.dma_size; sge 335 drivers/scsi/myrs.c sgl->sge[0].sge_addr = pdev_info_addr; sge 336 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->pdev_info.dma_size; sge 406 drivers/scsi/myrs.c sgl->sge[0].sge_addr = devmap_addr; sge 407 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->pdev_info.dma_size; sge 442 drivers/scsi/myrs.c sgl->sge[0].sge_addr = event_addr; sge 443 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->get_event.dma_size; sge 470 drivers/scsi/myrs.c sgl->sge[0].sge_addr = cs->fwstat_addr; sge 471 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ctlr_info.dma_size; sge 1725 drivers/scsi/myrs.c hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl); sge 1726 drivers/scsi/myrs.c hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl); sge 1759 drivers/scsi/myrs.c hw_sgl = hw_sge->sge; sge 670 drivers/scsi/myrs.h struct myrs_sge sge[2]; /* Bytes 0-31 */ sge 35 drivers/scsi/qedf/drv_scsi_fw_funcs.c ctx_data_desc->sge[sge_index].sge_addr.lo = val; sge 37 drivers/scsi/qedf/drv_scsi_fw_funcs.c ctx_data_desc->sge[sge_index].sge_addr.hi = val; sge 39 drivers/scsi/qedf/drv_scsi_fw_funcs.c ctx_data_desc->sge[sge_index].sge_len = val; sge 44 drivers/scsi/qedi/qedi_fw_api.c ctx_data_desc->sge[sge_index].sge_addr.lo = val; sge 46 drivers/scsi/qedi/qedi_fw_api.c ctx_data_desc->sge[sge_index].sge_addr.hi = val; sge 48 drivers/scsi/qedi/qedi_fw_api.c ctx_data_desc->sge[sge_index].sge_len = val; sge 191 drivers/scsi/qedi/qedi_iscsi.c struct scsi_sge *sge; sge 195 drivers/scsi/qedi/qedi_iscsi.c sizeof(*sge), sge 52 drivers/scsi/vmw_pvscsi.c struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; sge 333 drivers/scsi/vmw_pvscsi.c struct PVSCSISGElement *sge; sge 337 drivers/scsi/vmw_pvscsi.c sge = &ctx->sgl->sge[0]; sge 339 drivers/scsi/vmw_pvscsi.c sge[i].addr = sg_dma_address(sg); sge 340 drivers/scsi/vmw_pvscsi.c sge[i].length = sg_dma_len(sg); sge 341 drivers/scsi/vmw_pvscsi.c sge[i].flags = 0; sge 283 fs/cifs/smbdirect.c request->sge[i].addr, sge 284 fs/cifs/smbdirect.c request->sge[i].length, sge 504 fs/cifs/smbdirect.c response->sge.addr, sge 505 fs/cifs/smbdirect.c response->sge.length, sge 731 fs/cifs/smbdirect.c request->sge[0].addr = ib_dma_map_single( sge 734 fs/cifs/smbdirect.c if (ib_dma_mapping_error(info->id->device, request->sge[0].addr)) { sge 739 fs/cifs/smbdirect.c request->sge[0].length = sizeof(*packet); sge 740 fs/cifs/smbdirect.c request->sge[0].lkey = info->pd->local_dma_lkey; sge 743 fs/cifs/smbdirect.c info->id->device, request->sge[0].addr, sge 744 fs/cifs/smbdirect.c request->sge[0].length, DMA_TO_DEVICE); sge 750 fs/cifs/smbdirect.c send_wr.sg_list = request->sge; sge 756 fs/cifs/smbdirect.c request->sge[0].addr, sge 757 fs/cifs/smbdirect.c request->sge[0].length, request->sge[0].lkey); sge 768 fs/cifs/smbdirect.c ib_dma_unmap_single(info->id->device, request->sge[0].addr, sge 769 fs/cifs/smbdirect.c request->sge[0].length, DMA_TO_DEVICE); sge 894 fs/cifs/smbdirect.c request->sge[0].addr = ib_dma_map_single(info->id->device, sge 898 fs/cifs/smbdirect.c if (ib_dma_mapping_error(info->id->device, request->sge[0].addr)) { sge 904 fs/cifs/smbdirect.c request->sge[0].length = header_length; sge 905 fs/cifs/smbdirect.c request->sge[0].lkey = info->pd->local_dma_lkey; sge 920 fs/cifs/smbdirect.c request->sge[0].addr, sge 921 fs/cifs/smbdirect.c request->sge[0].length, sge 937 fs/cifs/smbdirect.c i, request->sge[i].addr, request->sge[i].length); sge 940 fs/cifs/smbdirect.c request->sge[i].addr, sge 941 fs/cifs/smbdirect.c request->sge[i].length, sge 949 fs/cifs/smbdirect.c send_wr.sg_list = request->sge; sge 997 fs/cifs/smbdirect.c request->sge[i+1].addr = sge 1001 fs/cifs/smbdirect.c info->id->device, request->sge[i+1].addr)) { sge 1003 fs/cifs/smbdirect.c request->sge[i+1].addr = 0; sge 1006 fs/cifs/smbdirect.c request->sge[i+1].length = sg->length; sge 1007 fs/cifs/smbdirect.c request->sge[i+1].lkey = info->pd->local_dma_lkey; sge 1017 fs/cifs/smbdirect.c if (request->sge[i].addr) sge 1019 fs/cifs/smbdirect.c request->sge[i].addr, sge 1020 fs/cifs/smbdirect.c request->sge[i].length, sge 1095 fs/cifs/smbdirect.c response->sge.addr = ib_dma_map_single( sge 1098 fs/cifs/smbdirect.c if (ib_dma_mapping_error(info->id->device, response->sge.addr)) sge 1101 fs/cifs/smbdirect.c response->sge.length = info->max_receive_size; sge 1102 fs/cifs/smbdirect.c response->sge.lkey = info->pd->local_dma_lkey; sge 1108 fs/cifs/smbdirect.c recv_wr.sg_list = &response->sge; sge 1113 fs/cifs/smbdirect.c ib_dma_unmap_single(info->id->device, response->sge.addr, sge 1114 fs/cifs/smbdirect.c response->sge.length, DMA_FROM_DEVICE); sge 1133 fs/cifs/smbdirect.c rc, response->sge.addr, sge 1134 fs/cifs/smbdirect.c response->sge.length, response->sge.lkey); sge 1275 fs/cifs/smbdirect.c ib_dma_unmap_single(info->id->device, response->sge.addr, sge 1276 fs/cifs/smbdirect.c response->sge.length, DMA_FROM_DEVICE); sge 242 fs/cifs/smbdirect.h struct ib_sge sge[SMBDIRECT_MAX_SGE]; sge 253 fs/cifs/smbdirect.h struct ib_sge sge; sge 96 include/linux/qed/storage_common.h struct scsi_sge sge[4]; sge 232 include/linux/skmsg.h struct scatterlist *sge = sk_msg_elem(msg, msg->sg.start); sge 238 include/linux/skmsg.h msg->data = sg_virt(sge); sge 239 include/linux/skmsg.h msg->data_end = msg->data + sge->length; sge 246 include/linux/skmsg.h struct scatterlist *sge; sge 249 include/linux/skmsg.h sge = sk_msg_elem(msg, msg->sg.end); sge 250 include/linux/skmsg.h sg_set_page(sge, page, len, offset); sge 251 include/linux/skmsg.h sg_unmark_end(sge); sge 566 include/rdma/rdma_vt.h int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge, sge 570 include/rdma/rdma_vt.h struct ib_sge *sge, int acc); sge 120 include/rdma/rdmavt_mr.h struct rvt_sge sge; /* progress state for the current SGE */ sge 138 include/rdma/rdmavt_mr.h rvt_put_mr(ss->sge.mr); sge 140 include/rdma/rdmavt_mr.h ss->sge = *ss->sg_list++; sge 144 include/rdma/rdmavt_mr.h static inline u32 rvt_get_sge_length(struct rvt_sge *sge, u32 length) sge 146 include/rdma/rdmavt_mr.h u32 len = sge->length; sge 150 include/rdma/rdmavt_mr.h if (len > sge->sge_length) sge 151 include/rdma/rdmavt_mr.h len = sge->sge_length; sge 159 include/rdma/rdmavt_mr.h struct rvt_sge *sge = &ss->sge; sge 161 include/rdma/rdmavt_mr.h sge->vaddr += length; sge 162 include/rdma/rdmavt_mr.h sge->length -= length; sge 163 include/rdma/rdmavt_mr.h sge->sge_length -= length; sge 164 include/rdma/rdmavt_mr.h if (sge->sge_length == 0) { sge 166 include/rdma/rdmavt_mr.h rvt_put_mr(sge->mr); sge 168 include/rdma/rdmavt_mr.h *sge = *ss->sg_list++; sge 169 include/rdma/rdmavt_mr.h } else if (sge->length == 0 && sge->mr->lkey) { sge 170 include/rdma/rdmavt_mr.h if (++sge->n >= RVT_SEGSZ) { sge 171 include/rdma/rdmavt_mr.h if (++sge->m >= sge->mr->mapsz) sge 173 include/rdma/rdmavt_mr.h sge->n = 0; sge 175 include/rdma/rdmavt_mr.h sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; sge 176 include/rdma/rdmavt_mr.h sge->length = sge->mr->map[sge->m]->segs[sge->n].length; sge 183 include/rdma/rdmavt_mr.h struct rvt_sge *sge = &ss->sge; sge 186 include/rdma/rdmavt_mr.h u32 len = rvt_get_sge_length(sge, length); sge 587 include/rdma/rdmavt_qp.h struct rvt_sge *sge = &wqe->sg_list[i]; sge 589 include/rdma/rdmavt_qp.h rvt_put_mr(sge->mr); sge 24 include/rdma/rw.h struct ib_sge sge; sge 36 include/rdma/rw.h struct ib_sge sge; sge 129 include/uapi/rdma/rdma_user_rxe.h struct rxe_sge sge[0]; sge 119 include/uapi/rdma/siw-abi.h struct siw_sge sge[SIW_MAX_SGE]; sge 135 include/uapi/rdma/siw-abi.h struct siw_sge sge[SIW_MAX_SGE]; sge 387 net/9p/trans_rdma.c struct ib_sge sge; sge 397 net/9p/trans_rdma.c sge.addr = c->busa; sge 398 net/9p/trans_rdma.c sge.length = client->msize; sge 399 net/9p/trans_rdma.c sge.lkey = rdma->pd->local_dma_lkey; sge 403 net/9p/trans_rdma.c wr.sg_list = &sge; sge 416 net/9p/trans_rdma.c struct ib_sge sge; sge 489 net/9p/trans_rdma.c sge.addr = c->busa; sge 490 net/9p/trans_rdma.c sge.length = c->req->tc.size; sge 491 net/9p/trans_rdma.c sge.lkey = rdma->pd->local_dma_lkey; sge 497 net/9p/trans_rdma.c wr.sg_list = &sge; sge 2224 net/core/filter.c struct scatterlist *sge; sge 2281 net/core/filter.c sge = sk_msg_elem(msg, i); sge 2282 net/core/filter.c from = sg_virt(sge); sge 2283 net/core/filter.c len = sge->length; sge 2288 net/core/filter.c sge->length = 0; sge 2289 net/core/filter.c put_page(sg_page(sge)); sge 2348 net/core/filter.c struct scatterlist sge, nsge, nnsge, rsge = {0}, *psge; sge 2427 net/core/filter.c sge = sk_msg_elem_cpy(msg, i); sge 2430 net/core/filter.c sg_unmark_end(&sge); sge 2440 net/core/filter.c msg->sg.data[i] = sge; sge 2441 net/core/filter.c sge = nsge; sge 2492 net/core/filter.c struct scatterlist tmp, sge; sge 2495 net/core/filter.c sge = sk_msg_elem_cpy(msg, i); sge 2500 net/core/filter.c msg->sg.data[i] = sge; sge 2502 net/core/filter.c sge = tmp; sge 2557 net/core/filter.c struct scatterlist *nsge, *sge = sk_msg_elem(msg, i); sge 2559 net/core/filter.c int b = sge->length - pop - a; sge 2563 net/core/filter.c if (pop < sge->length - a) { sge 2565 net/core/filter.c sge->length = a; sge 2568 net/core/filter.c get_page(sg_page(sge)); sge 2570 net/core/filter.c sg_page(sge), sge 2571 net/core/filter.c b, sge->offset + pop + a); sge 2582 net/core/filter.c sge->length = a; sge 2583 net/core/filter.c orig = sg_page(sge); sge 2584 net/core/filter.c from = sg_virt(sge); sge 2588 net/core/filter.c sg_set_page(sge, page, a + b, 0); sge 2592 net/core/filter.c } else if (pop >= sge->length - a) { sge 2593 net/core/filter.c pop -= (sge->length - a); sge 2594 net/core/filter.c sge->length = a; sge 2616 net/core/filter.c struct scatterlist *sge = sk_msg_elem(msg, i); sge 2618 net/core/filter.c if (pop < sge->length) { sge 2619 net/core/filter.c sge->length -= pop; sge 2620 net/core/filter.c sge->offset += pop; sge 2623 net/core/filter.c pop -= sge->length; sge 33 net/core/skmsg.c struct scatterlist *sge; sge 47 net/core/skmsg.c sge = &msg->sg.data[i]; sge 50 net/core/skmsg.c sg_page(sge) == pfrag->page && sge 51 net/core/skmsg.c sge->offset + sge->length == orig_offset) { sge 52 net/core/skmsg.c sge->length += use; sge 59 net/core/skmsg.c sge = &msg->sg.data[msg->sg.end]; sge 60 net/core/skmsg.c sg_unmark_end(sge); sge 61 net/core/skmsg.c sg_set_page(sge, pfrag->page, use, orig_offset); sge 80 net/core/skmsg.c struct scatterlist *sge = sk_msg_elem(src, i); sge 85 net/core/skmsg.c if (sge->length > off) sge 87 net/core/skmsg.c off -= sge->length; sge 91 net/core/skmsg.c sge = sk_msg_elem(src, i); sge 95 net/core/skmsg.c sge_len = sge->length - off; sge 103 net/core/skmsg.c (sg_page(sge) == sg_page(sgd)) && sge 104 net/core/skmsg.c (sg_virt(sge) + off == sg_virt(sgd) + sgd->length)) { sge 108 net/core/skmsg.c sge_off = sge->offset + off; sge 109 net/core/skmsg.c sk_msg_page_add(dst, sg_page(sge), sge_len, sge_off); sge 120 net/core/skmsg.c sge = sk_msg_elem(src, i); sge 132 net/core/skmsg.c struct scatterlist *sge = sk_msg_elem(msg, i); sge 134 net/core/skmsg.c if (bytes < sge->length) { sge 135 net/core/skmsg.c sge->length -= bytes; sge 136 net/core/skmsg.c sge->offset += bytes; sge 141 net/core/skmsg.c sk_mem_uncharge(sk, sge->length); sge 142 net/core/skmsg.c bytes -= sge->length; sge 143 net/core/skmsg.c sge->length = 0; sge 144 net/core/skmsg.c sge->offset = 0; sge 156 net/core/skmsg.c struct scatterlist *sge = &msg->sg.data[i]; sge 157 net/core/skmsg.c int uncharge = (bytes < sge->length) ? bytes : sge->length; sge 169 net/core/skmsg.c struct scatterlist *sge = sk_msg_elem(msg, i); sge 170 net/core/skmsg.c u32 len = sge->length; sge 175 net/core/skmsg.c put_page(sg_page(sge)); sge 176 net/core/skmsg.c memset(sge, 0, sizeof(*sge)); sge 183 net/core/skmsg.c struct scatterlist *sge = sk_msg_elem(msg, i); sge 187 net/core/skmsg.c msg->sg.size -= sge->length; sge 191 net/core/skmsg.c sge = sk_msg_elem(msg, i); sge 213 net/core/skmsg.c struct scatterlist *sge; sge 217 net/core/skmsg.c sge = sk_msg_elem(msg, i); sge 218 net/core/skmsg.c if (!sge->length) sge 220 net/core/skmsg.c if (bytes < sge->length) { sge 223 net/core/skmsg.c sge->length -= bytes; sge 224 net/core/skmsg.c sge->offset += bytes; sge 229 net/core/skmsg.c msg->sg.size -= sge->length; sge 230 net/core/skmsg.c bytes -= sge->length; sge 360 net/core/skmsg.c struct scatterlist *sge; sge 365 net/core/skmsg.c sge = sk_msg_elem(msg, i); sge 367 net/core/skmsg.c if (msg->sg.copybreak >= sge->length) { sge 372 net/core/skmsg.c sge = sk_msg_elem(msg, i); sge 375 net/core/skmsg.c buf_size = sge->length - msg->sg.copybreak; sge 377 net/core/skmsg.c to = sg_virt(sge) + msg->sg.copybreak; sge 57 net/ipv4/tcp_bpf.c struct scatterlist *sge; sge 67 net/ipv4/tcp_bpf.c sge = sk_msg_elem(msg_rx, i); sge 68 net/ipv4/tcp_bpf.c copy = sge->length; sge 69 net/ipv4/tcp_bpf.c page = sg_page(sge); sge 72 net/ipv4/tcp_bpf.c ret = copy_page_to_iter(page, sge->offset, copy, iter); sge 80 net/ipv4/tcp_bpf.c sge->offset += copy; sge 81 net/ipv4/tcp_bpf.c sge->length -= copy; sge 85 net/ipv4/tcp_bpf.c if (!sge->length) { sge 104 net/ipv4/tcp_bpf.c if (!sge->length && msg_rx->sg.start == msg_rx->sg.end) { sge 168 net/ipv4/tcp_bpf.c struct scatterlist *sge; sge 181 net/ipv4/tcp_bpf.c sge = sk_msg_elem(msg, i); sge 182 net/ipv4/tcp_bpf.c size = (apply && apply_bytes < sge->length) ? sge 183 net/ipv4/tcp_bpf.c apply_bytes : sge->length; sge 193 net/ipv4/tcp_bpf.c if (sge->length) sge 221 net/ipv4/tcp_bpf.c struct scatterlist *sge; sge 229 net/ipv4/tcp_bpf.c sge = sk_msg_elem(msg, msg->sg.start); sge 230 net/ipv4/tcp_bpf.c size = (apply && apply_bytes < sge->length) ? sge 231 net/ipv4/tcp_bpf.c apply_bytes : sge->length; sge 232 net/ipv4/tcp_bpf.c off = sge->offset; sge 233 net/ipv4/tcp_bpf.c page = sg_page(sge); sge 251 net/ipv4/tcp_bpf.c sge->offset += ret; sge 252 net/ipv4/tcp_bpf.c sge->length -= ret; sge 260 net/ipv4/tcp_bpf.c if (!sge->length) { sge 263 net/ipv4/tcp_bpf.c sg_init_table(sge, 1); sge 53 net/rds/ib_recv.c struct ib_sge *sge; sge 63 net/rds/ib_recv.c sge = &recv->r_sge[0]; sge 64 net/rds/ib_recv.c sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header)); sge 65 net/rds/ib_recv.c sge->length = sizeof(struct rds_header); sge 66 net/rds/ib_recv.c sge->lkey = ic->i_pd->local_dma_lkey; sge 68 net/rds/ib_recv.c sge = &recv->r_sge[1]; sge 69 net/rds/ib_recv.c sge->addr = 0; sge 70 net/rds/ib_recv.c sge->length = RDS_FRAG_SIZE; sge 71 net/rds/ib_recv.c sge->lkey = ic->i_pd->local_dma_lkey; sge 311 net/rds/ib_recv.c struct ib_sge *sge; sge 345 net/rds/ib_recv.c sge = &recv->r_sge[0]; sge 346 net/rds/ib_recv.c sge->addr = ic->i_recv_hdrs_dma + (recv - ic->i_recvs) * sizeof(struct rds_header); sge 347 net/rds/ib_recv.c sge->length = sizeof(struct rds_header); sge 349 net/rds/ib_recv.c sge = &recv->r_sge[1]; sge 350 net/rds/ib_recv.c sge->addr = sg_dma_address(&recv->r_frag->f_sg); sge 351 net/rds/ib_recv.c sge->length = sg_dma_len(&recv->r_frag->f_sg); sge 576 net/rds/ib_recv.c struct ib_sge *sge = &ic->i_ack_sge; sge 578 net/rds/ib_recv.c sge->addr = ic->i_ack_dma; sge 579 net/rds/ib_recv.c sge->length = sizeof(struct rds_header); sge 580 net/rds/ib_recv.c sge->lkey = ic->i_pd->local_dma_lkey; sge 582 net/rds/ib_recv.c wr->sg_list = sge; sge 195 net/rds/ib_send.c struct ib_sge *sge; sge 203 net/rds/ib_send.c sge = &send->s_sge[0]; sge 204 net/rds/ib_send.c sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header)); sge 205 net/rds/ib_send.c sge->length = sizeof(struct rds_header); sge 206 net/rds/ib_send.c sge->lkey = ic->i_pd->local_dma_lkey; sge 323 net/smc/smc_tx.c struct ib_sge *sge = sge 328 net/smc/smc_tx.c sge[srcchunk].addr = dma_addr + src_off; sge 329 net/smc/smc_tx.c sge[srcchunk].length = src_len; sge 568 net/sunrpc/xprtrdma/rpc_rdma.c struct ib_sge *sge; sge 577 net/sunrpc/xprtrdma/rpc_rdma.c for (sge = &sc->sc_sges[2]; sc->sc_unmap_count; sge 578 net/sunrpc/xprtrdma/rpc_rdma.c ++sge, --sc->sc_unmap_count) sge 579 net/sunrpc/xprtrdma/rpc_rdma.c ib_dma_unmap_page(sc->sc_device, sge->addr, sge->length, sge 592 net/sunrpc/xprtrdma/rpc_rdma.c struct ib_sge *sge = sc->sc_sges; sge 596 net/sunrpc/xprtrdma/rpc_rdma.c sge->addr = rdmab_addr(rb); sge 597 net/sunrpc/xprtrdma/rpc_rdma.c sge->length = len; sge 598 net/sunrpc/xprtrdma/rpc_rdma.c sge->lkey = rdmab_lkey(rb); sge 600 net/sunrpc/xprtrdma/rpc_rdma.c ib_dma_sync_single_for_device(rdmab_device(rb), sge->addr, sge->length, sge 621 net/sunrpc/xprtrdma/rpc_rdma.c struct ib_sge *sge = sc->sc_sges; sge 631 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].addr = rdmab_addr(rb); sge 632 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].length = xdr->head[0].iov_len; sge 633 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].lkey = rdmab_lkey(rb); sge 634 net/sunrpc/xprtrdma/rpc_rdma.c ib_dma_sync_single_for_device(rdmab_device(rb), sge[sge_no].addr, sge 635 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].length, DMA_TO_DEVICE); sge 676 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].addr = sge 680 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].addr)) sge 682 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].length = len; sge 683 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].lkey = rdmab_lkey(rb); sge 704 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].addr = sge 707 net/sunrpc/xprtrdma/rpc_rdma.c if (ib_dma_mapping_error(rdmab_device(rb), sge[sge_no].addr)) sge 709 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].length = len; sge 710 net/sunrpc/xprtrdma/rpc_rdma.c sge[sge_no].lkey = rdmab_lkey(rb); sge 731 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_dma_maperr(sge[sge_no].addr); sge 430 net/tls/tls_sw.c struct scatterlist *sge; sge 439 net/tls/tls_sw.c sge = sk_msg_elem(msg_en, msg_en->sg.curr); sge 440 net/tls/tls_sw.c sge->offset -= prot->prepend_size; sge 441 net/tls/tls_sw.c sge->length += prot->prepend_size; sge 493 net/tls/tls_sw.c struct scatterlist *sge = sk_msg_elem(msg_en, start); sge 507 net/tls/tls_sw.c sge->offset += prot->prepend_size; sge 508 net/tls/tls_sw.c sge->length -= prot->prepend_size; sge 528 net/tls/tls_sw.c sge->offset -= prot->prepend_size; sge 529 net/tls/tls_sw.c sge->length += prot->prepend_size; sge 551 net/tls/tls_sw.c struct scatterlist *sge, *osge, *nsge; sge 570 net/tls/tls_sw.c sge = sk_msg_elem(msg_opl, i); sge 571 net/tls/tls_sw.c while (apply && sge->length) { sge 572 net/tls/tls_sw.c if (sge->length > apply) { sge 573 net/tls/tls_sw.c u32 len = sge->length - apply; sge 575 net/tls/tls_sw.c get_page(sg_page(sge)); sge 576 net/tls/tls_sw.c sg_set_page(&tmp, sg_page(sge), len, sge 577 net/tls/tls_sw.c sge->offset + apply); sge 578 net/tls/tls_sw.c sge->length = apply; sge 582 net/tls/tls_sw.c apply -= sge->length; sge 583 net/tls/tls_sw.c bytes += sge->length; sge 589 net/tls/tls_sw.c sge = sk_msg_elem(msg_opl, i);