sge               247 drivers/ata/sata_sil24.c 	struct sil24_sge sge[SIL24_MAX_SGE];
sge               253 drivers/ata/sata_sil24.c 	struct sil24_sge sge[SIL24_MAX_SGE];
sge               770 drivers/ata/sata_sil24.c 				 struct sil24_sge *sge)
sge               777 drivers/ata/sata_sil24.c 		sge->addr = cpu_to_le64(sg_dma_address(sg));
sge               778 drivers/ata/sata_sil24.c 		sge->cnt = cpu_to_le32(sg_dma_len(sg));
sge               779 drivers/ata/sata_sil24.c 		sge->flags = 0;
sge               781 drivers/ata/sata_sil24.c 		last_sge = sge;
sge               782 drivers/ata/sata_sil24.c 		sge++;
sge               839 drivers/ata/sata_sil24.c 	struct sil24_sge *sge;
sge               846 drivers/ata/sata_sil24.c 		sge = cb->ata.sge;
sge               860 drivers/ata/sata_sil24.c 		sge = cb->atapi.sge;
sge               876 drivers/ata/sata_sil24.c 		sil24_fill_sg(qc, sge);
sge                28 drivers/crypto/bcm/util.c 		     struct scatterlist **sge, unsigned int *sge_offset)
sge                45 drivers/crypto/bcm/util.c 	*sge = sg;
sge                71 drivers/crypto/bcm/util.h 		     struct scatterlist **sge, unsigned int *sge_offset);
sge               480 drivers/crypto/chelsio/chcr_algo.c 	walk->pair = ulp->sge;
sge               584 drivers/crypto/chelsio/chcr_algo.c 		adap->sge.uld_txq_info[CXGB4_TX_CRYPTO];
sge               396 drivers/crypto/chelsio/chcr_ipsec.c 	q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge               444 drivers/crypto/chelsio/chcr_ipsec.c 	q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge               487 drivers/crypto/chelsio/chcr_ipsec.c 	q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge               543 drivers/crypto/chelsio/chcr_ipsec.c 	struct sge_eth_txq *q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge               701 drivers/crypto/chelsio/chcr_ipsec.c 	q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge              1215 drivers/infiniband/core/mad.c 	struct ib_sge *sge;
sge              1226 drivers/infiniband/core/mad.c 	sge = mad_send_wr->sg_list;
sge              1227 drivers/infiniband/core/mad.c 	sge[0].addr = ib_dma_map_single(mad_agent->device,
sge              1229 drivers/infiniband/core/mad.c 					sge[0].length,
sge              1231 drivers/infiniband/core/mad.c 	if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr)))
sge              1234 drivers/infiniband/core/mad.c 	mad_send_wr->header_mapping = sge[0].addr;
sge              1236 drivers/infiniband/core/mad.c 	sge[1].addr = ib_dma_map_single(mad_agent->device,
sge              1238 drivers/infiniband/core/mad.c 					sge[1].length,
sge              1240 drivers/infiniband/core/mad.c 	if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) {
sge              1243 drivers/infiniband/core/mad.c 				    sge[0].length, DMA_TO_DEVICE);
sge              1246 drivers/infiniband/core/mad.c 	mad_send_wr->payload_mapping = sge[1].addr;
sge              1267 drivers/infiniband/core/mad.c 				    sge[0].length, DMA_TO_DEVICE);
sge              1270 drivers/infiniband/core/mad.c 				    sge[1].length, DMA_TO_DEVICE);
sge               113 drivers/infiniband/core/rw.c 	reg->sge.addr = reg->mr->iova;
sge               114 drivers/infiniband/core/rw.c 	reg->sge.length = reg->mr->length;
sge               153 drivers/infiniband/core/rw.c 		reg->wr.wr.sg_list = &reg->sge;
sge               167 drivers/infiniband/core/rw.c 		remote_addr += reg->sge.length;
sge               195 drivers/infiniband/core/rw.c 	struct ib_sge *sge;
sge               200 drivers/infiniband/core/rw.c 	ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL);
sge               219 drivers/infiniband/core/rw.c 		rdma_wr->wr.sg_list = sge;
sge               222 drivers/infiniband/core/rw.c 			sge->addr = sg_dma_address(sg) + offset;
sge               223 drivers/infiniband/core/rw.c 			sge->length = sg_dma_len(sg) - offset;
sge               224 drivers/infiniband/core/rw.c 			sge->lkey = qp->pd->local_dma_lkey;
sge               226 drivers/infiniband/core/rw.c 			total_len += sge->length;
sge               227 drivers/infiniband/core/rw.c 			sge++;
sge               253 drivers/infiniband/core/rw.c 	ctx->single.sge.lkey = qp->pd->local_dma_lkey;
sge               254 drivers/infiniband/core/rw.c 	ctx->single.sge.addr = sg_dma_address(sg) + offset;
sge               255 drivers/infiniband/core/rw.c 	ctx->single.sge.length = sg_dma_len(sg) - offset;
sge               262 drivers/infiniband/core/rw.c 	rdma_wr->wr.sg_list = &ctx->single.sge;
sge               439 drivers/infiniband/core/rw.c 	ctx->reg->sge.addr = ctx->reg->mr->iova;
sge               440 drivers/infiniband/core/rw.c 	ctx->reg->sge.length = ctx->reg->mr->length;
sge               442 drivers/infiniband/core/rw.c 		ctx->reg->sge.length -= ctx->reg->mr->sig_attrs->meta_length;
sge               445 drivers/infiniband/core/rw.c 	rdma_wr->wr.sg_list = &ctx->reg->sge;
sge               482 drivers/infiniband/core/rw.c 	reg->sge.lkey = reg->mr->lkey;
sge               100 drivers/infiniband/hw/bnxt_re/bnxt_re.h 	struct bnxt_qplib_sge sge;
sge              1844 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_qplib_sge sge;
sge              1961 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	buf = bnxt_qplib_get_qp1_sq_buf(&qp->qplib_qp, &sge);
sge              1978 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			sge.size -= 20;
sge              1985 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			sge.size -= 8;
sge              1989 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			sge.size -= 4;
sge              1991 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->sg_list[0].addr = sge.addr;
sge              1992 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->sg_list[0].lkey = sge.lkey;
sge              1993 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->sg_list[0].size = sge.size;
sge              2014 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_qplib_sge ref, sge;
sge              2020 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (!bnxt_qplib_get_qp1_rq_buf(&qp->qplib_qp, &sge))
sge              2034 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->sg_list[0].addr = sge.addr;
sge              2035 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->sg_list[0].lkey = sge.lkey;
sge              2037 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	sge.size -= wqe->sg_list[0].size;
sge              2039 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	sqp_entry->sge.addr = ref.addr;
sge              2040 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	sqp_entry->sge.lkey = ref.lkey;
sge              2041 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	sqp_entry->sge.size = ref.size;
sge              2935 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	r_sge[1].addr = sqp_entry->sge.addr + offset;
sge              2936 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	r_sge[1].lkey = sqp_entry->sge.lkey;
sge              1457 drivers/infiniband/hw/bnxt_re/qplib_fp.c 				struct bnxt_qplib_sge *sge)
sge              1462 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	memset(sge, 0, sizeof(*sge));
sge              1466 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->addr = (dma_addr_t)(qp->sq_hdr_buf_map +
sge              1468 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->lkey = 0xFFFFFFFF;
sge              1469 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->size = qp->sq_hdr_buf_size;
sge              1470 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		return qp->sq_hdr_buf + sw_prod * sge->size;
sge              1488 drivers/infiniband/hw/bnxt_re/qplib_fp.c 				struct bnxt_qplib_sge *sge)
sge              1493 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	memset(sge, 0, sizeof(*sge));
sge              1497 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->addr = (dma_addr_t)(qp->rq_hdr_buf_map +
sge              1499 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->lkey = 0xFFFFFFFF;
sge              1500 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		sge->size = qp->rq_hdr_buf_size;
sge              1501 drivers/infiniband/hw/bnxt_re/qplib_fp.c 		return qp->rq_hdr_buf + sw_prod * sge->size;
sge               534 drivers/infiniband/hw/bnxt_re/qplib_fp.h 				struct bnxt_qplib_sge *sge);
sge               536 drivers/infiniband/hw/bnxt_re/qplib_fp.h 				struct bnxt_qplib_sge *sge);
sge               464 drivers/infiniband/hw/cxgb4/qp.c 	flitp = (__be64 *)isglp->sge;
sge               126 drivers/infiniband/hw/cxgb4/t4fw_ri_api.h 	struct fw_ri_dsge_pair sge[0];
sge               142 drivers/infiniband/hw/cxgb4/t4fw_ri_api.h 	struct fw_ri_sge sge[0];
sge               200 drivers/infiniband/hw/hfi1/rc.c 			qp->s_ack_rdma_sge.sge = e->rdma_sge;
sge               249 drivers/infiniband/hw/hfi1/rc.c 			qp->s_ack_rdma_sge.sge = e->rdma_sge;
sge               273 drivers/infiniband/hw/hfi1/rc.c 		ps->s_txreq->mr = qp->s_ack_rdma_sge.sge.mr;
sge               276 drivers/infiniband/hw/hfi1/rc.c 		len = qp->s_ack_rdma_sge.sge.sge_length;
sge               833 drivers/infiniband/hw/hfi1/rc.c 					qp->s_sge.sge = wqe->sg_list[0];
sge               918 drivers/infiniband/hw/hfi1/rc.c 			qp->s_sge.sge = wqe->sg_list[0];
sge              3029 drivers/infiniband/hw/hfi1/rc.c 			ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr,
sge              3036 drivers/infiniband/hw/hfi1/rc.c 			qp->r_sge.sge.mr = NULL;
sge              3037 drivers/infiniband/hw/hfi1/rc.c 			qp->r_sge.sge.vaddr = NULL;
sge              3038 drivers/infiniband/hw/hfi1/rc.c 			qp->r_sge.sge.length = 0;
sge              3039 drivers/infiniband/hw/hfi1/rc.c 			qp->r_sge.sge.sge_length = 0;
sge              3160 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64),
sge              3165 drivers/infiniband/hw/hfi1/rc.c 		maddr = (atomic64_t *)qp->r_sge.sge.vaddr;
sge              3169 drivers/infiniband/hw/hfi1/rc.c 			(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
sge              3172 drivers/infiniband/hw/hfi1/rc.c 		rvt_put_mr(qp->r_sge.sge.mr);
sge              1085 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct rvt_sge *sge = &ss->sge;
sge              1091 drivers/infiniband/hw/hfi1/tid_rdma.c 		pages[i++] = virt_to_page(sge->vaddr);
sge              1093 drivers/infiniband/hw/hfi1/tid_rdma.c 		sge->vaddr += len;
sge              1094 drivers/infiniband/hw/hfi1/tid_rdma.c 		sge->length -= len;
sge              1095 drivers/infiniband/hw/hfi1/tid_rdma.c 		sge->sge_length -= len;
sge              1096 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (!sge->sge_length) {
sge              1098 drivers/infiniband/hw/hfi1/tid_rdma.c 				*sge = ss->sg_list[req->isge - 1];
sge              1099 drivers/infiniband/hw/hfi1/tid_rdma.c 		} else if (sge->length == 0 && sge->mr->lkey) {
sge              1100 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (++sge->n >= RVT_SEGSZ) {
sge              1101 drivers/infiniband/hw/hfi1/tid_rdma.c 				++sge->m;
sge              1102 drivers/infiniband/hw/hfi1/tid_rdma.c 				sge->n = 0;
sge              1104 drivers/infiniband/hw/hfi1/tid_rdma.c 			sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
sge              1105 drivers/infiniband/hw/hfi1/tid_rdma.c 			sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
sge              1720 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.vaddr = req_addr;
sge              1721 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.sge_length = req_len;
sge              1722 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.length = wpriv->ss.sge.sge_length;
sge              1727 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.mr = NULL;
sge              1728 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.m = 0;
sge              1729 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.sge.n = 0;
sge              1732 drivers/infiniband/hw/hfi1/tid_rdma.c 	wpriv->ss.total_len = wpriv->ss.sge.sge_length;
sge              3256 drivers/infiniband/hw/hfi1/tid_rdma.c 					struct rvt_sge *sge, int num_sge)
sge              3260 drivers/infiniband/hw/hfi1/tid_rdma.c 	for (i = 0; i < num_sge; i++, sge++) {
sge              3261 drivers/infiniband/hw/hfi1/tid_rdma.c 		trace_hfi1_sge_check_align(qp, i, sge);
sge              3262 drivers/infiniband/hw/hfi1/tid_rdma.c 		if ((u64)sge->vaddr & ~PAGE_MASK ||
sge              3263 drivers/infiniband/hw/hfi1/tid_rdma.c 		    sge->sge_length & ~PAGE_MASK)
sge              3755 drivers/infiniband/hw/hfi1/tid_rdma.c 	req->ss.sge = e->rdma_sge;
sge              3881 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.vaddr = resp_addr;
sge              3882 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.sge_length = resp_len;
sge              3883 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.length = epriv->ss.sge.sge_length;
sge              3888 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.mr = NULL;
sge              3889 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.m = 0;
sge              3890 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.sge.n = 0;
sge              3893 drivers/infiniband/hw/hfi1/tid_rdma.c 	epriv->ss.total_len = epriv->ss.sge.sge_length;
sge              4320 drivers/infiniband/hw/hfi1/tid_rdma.c 			ss.sge = e->rdma_sge;
sge              5062 drivers/infiniband/hw/hfi1/tid_rdma.c 		priv->tid_ss.sge = wqe->sg_list[0];
sge              1324 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge),
sge              1325 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(qp, index, sge),
sge              1337 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->vaddr = (u64)sge->vaddr;
sge              1338 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->sge_length = sge->sge_length;
sge              1352 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge),
sge              1353 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(qp, index, sge)
sge               153 drivers/infiniband/hw/hfi1/uc.c 		qp->s_sge.sge = wqe->sg_list[0];
sge               500 drivers/infiniband/hw/hfi1/uc.c 			ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len,
sge               507 drivers/infiniband/hw/hfi1/uc.c 			qp->r_sge.sge.mr = NULL;
sge               508 drivers/infiniband/hw/hfi1/uc.c 			qp->r_sge.sge.vaddr = NULL;
sge               509 drivers/infiniband/hw/hfi1/uc.c 			qp->r_sge.sge.length = 0;
sge               510 drivers/infiniband/hw/hfi1/uc.c 			qp->r_sge.sge.sge_length = 0;
sge                82 drivers/infiniband/hw/hfi1/ud.c 	struct rvt_sge *sge;
sge               221 drivers/infiniband/hw/hfi1/ud.c 	ssge.sge = *swqe->sg_list;
sge               223 drivers/infiniband/hw/hfi1/ud.c 	sge = &ssge.sge;
sge               225 drivers/infiniband/hw/hfi1/ud.c 		u32 len = rvt_get_sge_length(sge, length);
sge               228 drivers/infiniband/hw/hfi1/ud.c 		rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, len, true, false);
sge               555 drivers/infiniband/hw/hfi1/ud.c 	qp->s_sge.sge = wqe->sg_list[0];
sge               702 drivers/infiniband/hw/hfi1/verbs.c 	struct rvt_sge sge = ss->sge;
sge               708 drivers/infiniband/hw/hfi1/verbs.c 		len = rvt_get_sge_length(&ss->sge, length);
sge               713 drivers/infiniband/hw/hfi1/verbs.c 			ss->sge.vaddr,
sge               723 drivers/infiniband/hw/hfi1/verbs.c 	ss->sge = sge;
sge              1081 drivers/infiniband/hw/hfi1/verbs.c 				void *addr = ss->sge.vaddr;
sge              1082 drivers/infiniband/hw/hfi1/verbs.c 				u32 slen = rvt_get_sge_length(&ss->sge, len);
sge               691 drivers/infiniband/hw/hns/hns_roce_device.h 	struct hns_roce_sge	sge;
sge               135 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	dseg = get_send_extend_sge(qp, (*sge_ind) & (qp->sge.sge_cnt - 1));
sge               146 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					   (*sge_ind) & (qp->sge.sge_cnt - 1));
sge               206 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				     (*sge_ind) & (qp->sge.sge_cnt - 1));
sge               385 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				     sge_idx & (qp->sge.sge_cnt - 1));
sge              3227 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       ilog2((unsigned int)hr_qp->sge.sge_cnt));
sge              3234 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       ilog2((unsigned int)hr_qp->sge.sge_cnt) : 0);
sge              3932 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (hr_qp->sge.offset) {
sge              3935 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  hr_qp->sge.offset / page_size,
sge               379 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt *
sge               383 drivers/infiniband/hw/hns/hns_roce_qp.c 		if (hr_qp->sge.sge_cnt > hr_dev->caps.max_extend_sg) {
sge               386 drivers/infiniband/hw/hns/hns_roce_qp.c 				hr_qp->sge.sge_cnt);
sge               391 drivers/infiniband/hw/hns/hns_roce_qp.c 	hr_qp->sge.sge_shift = 4;
sge               392 drivers/infiniband/hw/hns/hns_roce_qp.c 	ex_sge_num = hr_qp->sge.sge_cnt;
sge               406 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_cnt = ex_sge_num ?
sge               407 drivers/infiniband/hw/hns/hns_roce_qp.c 		   max(page_size / (1 << hr_qp->sge.sge_shift), ex_sge_num) : 0;
sge               410 drivers/infiniband/hw/hns/hns_roce_qp.c 				   HNS_ROCE_ALOGN_UP((hr_qp->sge.sge_cnt <<
sge               411 drivers/infiniband/hw/hns/hns_roce_qp.c 					     hr_qp->sge.sge_shift), page_size) +
sge               417 drivers/infiniband/hw/hns/hns_roce_qp.c 			hr_qp->sge.offset = HNS_ROCE_ALOGN_UP(
sge               421 drivers/infiniband/hw/hns/hns_roce_qp.c 			hr_qp->rq.offset = hr_qp->sge.offset +
sge               422 drivers/infiniband/hw/hns/hns_roce_qp.c 					HNS_ROCE_ALOGN_UP((hr_qp->sge.sge_cnt <<
sge               423 drivers/infiniband/hw/hns/hns_roce_qp.c 						hr_qp->sge.sge_shift),
sge               450 drivers/infiniband/hw/hns/hns_roce_qp.c 	if (hr_qp->sge.sge_cnt > 0)
sge               457 drivers/infiniband/hw/hns/hns_roce_qp.c 		buf_size = hr_qp->sge.offset - hr_qp->sq.offset;
sge               472 drivers/infiniband/hw/hns/hns_roce_qp.c 		buf_size = hr_qp->rq.offset - hr_qp->sge.offset;
sge               477 drivers/infiniband/hw/hns/hns_roce_qp.c 						 hr_qp->sge.offset / page_size,
sge               527 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt *
sge               529 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_shift = 4;
sge               534 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_cnt = roundup_pow_of_two(hr_qp->sq.wqe_cnt *
sge               536 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_shift = 4;
sge               540 drivers/infiniband/hw/hns/hns_roce_qp.c 		if (hr_qp->sge.sge_cnt > hr_dev->caps.max_extend_sg) {
sge               542 drivers/infiniband/hw/hns/hns_roce_qp.c 				hr_qp->sge.sge_cnt);
sge               599 drivers/infiniband/hw/hns/hns_roce_qp.c 	if (hr_dev->caps.max_sq_sg > 2 && hr_qp->sge.sge_cnt) {
sge               600 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.sge_cnt = max(page_size/(1 << hr_qp->sge.sge_shift),
sge               601 drivers/infiniband/hw/hns/hns_roce_qp.c 					(u32)hr_qp->sge.sge_cnt);
sge               602 drivers/infiniband/hw/hns/hns_roce_qp.c 		hr_qp->sge.offset = size;
sge               603 drivers/infiniband/hw/hns/hns_roce_qp.c 		size += HNS_ROCE_ALOGN_UP(hr_qp->sge.sge_cnt <<
sge               604 drivers/infiniband/hw/hns/hns_roce_qp.c 					  hr_qp->sge.sge_shift, page_size);
sge              1283 drivers/infiniband/hw/hns/hns_roce_qp.c 	return hns_roce_buf_offset(&hr_qp->hr_buf, hr_qp->sge.offset +
sge              1284 drivers/infiniband/hw/hns/hns_roce_qp.c 					(n << hr_qp->sge.sge_shift));
sge               205 drivers/infiniband/hw/i40iw/i40iw_uk.c static void i40iw_set_fragment(u64 *wqe, u32 offset, struct i40iw_sge *sge)
sge               207 drivers/infiniband/hw/i40iw/i40iw_uk.c 	if (sge) {
sge               208 drivers/infiniband/hw/i40iw/i40iw_uk.c 		set_64bit_val(wqe, offset, LS_64(sge->tag_off, I40IWQPSQ_FRAG_TO));
sge               210 drivers/infiniband/hw/i40iw/i40iw_uk.c 			      (LS_64(sge->len, I40IWQPSQ_FRAG_LEN) |
sge               211 drivers/infiniband/hw/i40iw/i40iw_uk.c 			       LS_64(sge->stag, I40IWQPSQ_FRAG_STAG)));
sge               907 drivers/infiniband/hw/i40iw/i40iw_uk.c void i40iw_get_wqe_shift(u32 sge, u32 inline_data, u8 *shift)
sge               910 drivers/infiniband/hw/i40iw/i40iw_uk.c 	if (sge > 1 || inline_data > 16)
sge               911 drivers/infiniband/hw/i40iw/i40iw_uk.c 		*shift = (sge < 4 && inline_data <= 48) ? 1 : 2;
sge               427 drivers/infiniband/hw/i40iw/i40iw_user.h void i40iw_get_wqe_shift(u32 sge, u32 inline_data, u8 *shift);
sge              1302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff);
sge              1303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa);
sge              1304 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size;
sge              1350 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff);
sge              1351 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa);
sge              1352 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].len = dma.size;
sge              2887 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge;
sge               288 drivers/infiniband/hw/ocrdma/ocrdma_sli.h 			struct ocrdma_mqe_sge sge[19];
sge              1906 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			      struct ocrdma_sge *sge, int num_sge,
sge              1912 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge[i].lrkey = sg_list[i].lkey;
sge              1913 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge[i].addr_lo = sg_list[i].addr;
sge              1914 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge[i].addr_hi = upper_32_bits(sg_list[i].addr);
sge              1915 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge[i].len = sg_list[i].length;
sge              1919 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		memset(sge, 0, sizeof(*sge));
sge              1934 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				    struct ocrdma_sge *sge,
sge              1948 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		dpp_addr = (char *)sge;
sge              1961 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
sge              1976 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_sge *sge;
sge              1981 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge = (struct ocrdma_sge *)(hdr + 2);
sge              1984 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		sge = (struct ocrdma_sge *)(hdr + 1);
sge              1987 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
sge              1996 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_sge *sge = ext_rw + 1;
sge              1999 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
sge              2013 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_sge *sge = ext_rw + 1;
sge              2017 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
sge              2217 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_sge *sge;
sge              2219 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wqe_size = (wr->num_sge * sizeof(*sge)) + sizeof(*rqe);
sge              2221 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wqe_size = sizeof(*sge) + sizeof(*rqe);
sge              2229 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	sge = (struct ocrdma_sge *)(rqe + 1);
sge              2230 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_build_sges(rqe, sge, wr->num_sge, wr->sg_list);
sge               178 drivers/infiniband/hw/qedr/qedr_hsi_rdma.h 	struct rdma_srq_sge sge;
sge              2992 drivers/infiniband/hw/qedr/verbs.c #define RQ_SGE_SET(sge, vaddr, vlength, vflags)			\
sge              2994 drivers/infiniband/hw/qedr/verbs.c 		DMA_REGPAIR_LE(sge->addr, vaddr);		\
sge              2995 drivers/infiniband/hw/qedr/verbs.c 		(sge)->length = cpu_to_le32(vlength);		\
sge              2996 drivers/infiniband/hw/qedr/verbs.c 		(sge)->flags = cpu_to_le32(vflags);		\
sge              3005 drivers/infiniband/hw/qedr/verbs.c #define SRQ_SGE_SET(sge, vaddr, vlength, vlkey)			\
sge              3007 drivers/infiniband/hw/qedr/verbs.c 		DMA_REGPAIR_LE(sge->addr, vaddr);		\
sge              3008 drivers/infiniband/hw/qedr/verbs.c 		(sge)->length = cpu_to_le32(vlength);		\
sge              3009 drivers/infiniband/hw/qedr/verbs.c 		(sge)->l_key = cpu_to_le32(vlkey);		\
sge              3019 drivers/infiniband/hw/qedr/verbs.c 		struct rdma_sq_sge *sge = qed_chain_produce(&qp->sq.pbl);
sge              3021 drivers/infiniband/hw/qedr/verbs.c 		DMA_REGPAIR_LE(sge->addr, wr->sg_list[i].addr);
sge              3022 drivers/infiniband/hw/qedr/verbs.c 		sge->l_key = cpu_to_le32(wr->sg_list[i].lkey);
sge              3023 drivers/infiniband/hw/qedr/verbs.c 		sge->length = cpu_to_le32(wr->sg_list[i].length);
sge               122 drivers/infiniband/hw/qib/qib_rc.c 			qp->s_ack_rdma_sge.sge = e->rdma_sge;
sge               155 drivers/infiniband/hw/qib/qib_rc.c 		qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr;
sge               158 drivers/infiniband/hw/qib/qib_rc.c 		len = qp->s_ack_rdma_sge.sge.sge_length;
sge               447 drivers/infiniband/hw/qib/qib_rc.c 		qp->s_sge.sge = wqe->sg_list[0];
sge              1908 drivers/infiniband/hw/qib/qib_rc.c 			ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr,
sge              1915 drivers/infiniband/hw/qib/qib_rc.c 			qp->r_sge.sge.mr = NULL;
sge              1916 drivers/infiniband/hw/qib/qib_rc.c 			qp->r_sge.sge.vaddr = NULL;
sge              1917 drivers/infiniband/hw/qib/qib_rc.c 			qp->r_sge.sge.length = 0;
sge              1918 drivers/infiniband/hw/qib/qib_rc.c 			qp->r_sge.sge.sge_length = 0;
sge              2035 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64),
sge              2040 drivers/infiniband/hw/qib/qib_rc.c 		maddr = (atomic64_t *) qp->r_sge.sge.vaddr;
sge              2044 drivers/infiniband/hw/qib/qib_rc.c 			(u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr,
sge              2047 drivers/infiniband/hw/qib/qib_rc.c 		rvt_put_mr(qp->r_sge.sge.mr);
sge               516 drivers/infiniband/hw/qib/qib_sdma.c 	struct rvt_sge *sge;
sge               565 drivers/infiniband/hw/qib/qib_sdma.c 	sge = &ss->sge;
sge               568 drivers/infiniband/hw/qib/qib_sdma.c 		u32 len = rvt_get_sge_length(sge, dwords << 2);
sge               571 drivers/infiniband/hw/qib/qib_sdma.c 		addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr,
sge                98 drivers/infiniband/hw/qib/qib_uc.c 		qp->s_sge.sge = wqe->sg_list[0];
sge               425 drivers/infiniband/hw/qib/qib_uc.c 			ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len,
sge               432 drivers/infiniband/hw/qib/qib_uc.c 			qp->r_sge.sge.mr = NULL;
sge               433 drivers/infiniband/hw/qib/qib_uc.c 			qp->r_sge.sge.vaddr = NULL;
sge               434 drivers/infiniband/hw/qib/qib_uc.c 			qp->r_sge.sge.length = 0;
sge               435 drivers/infiniband/hw/qib/qib_uc.c 			qp->r_sge.sge.sge_length = 0;
sge                61 drivers/infiniband/hw/qib/qib_ud.c 	struct rvt_sge *sge;
sge               172 drivers/infiniband/hw/qib/qib_ud.c 	ssge.sge = *swqe->sg_list;
sge               174 drivers/infiniband/hw/qib/qib_ud.c 	sge = &ssge.sge;
sge               176 drivers/infiniband/hw/qib/qib_ud.c 		u32 len = rvt_get_sge_length(sge, length);
sge               178 drivers/infiniband/hw/qib/qib_ud.c 		rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, len, true, false);
sge               179 drivers/infiniband/hw/qib/qib_ud.c 		sge->vaddr += len;
sge               180 drivers/infiniband/hw/qib/qib_ud.c 		sge->length -= len;
sge               181 drivers/infiniband/hw/qib/qib_ud.c 		sge->sge_length -= len;
sge               182 drivers/infiniband/hw/qib/qib_ud.c 		if (sge->sge_length == 0) {
sge               184 drivers/infiniband/hw/qib/qib_ud.c 				*sge = *ssge.sg_list++;
sge               185 drivers/infiniband/hw/qib/qib_ud.c 		} else if (sge->length == 0 && sge->mr->lkey) {
sge               186 drivers/infiniband/hw/qib/qib_ud.c 			if (++sge->n >= RVT_SEGSZ) {
sge               187 drivers/infiniband/hw/qib/qib_ud.c 				if (++sge->m >= sge->mr->mapsz)
sge               189 drivers/infiniband/hw/qib/qib_ud.c 				sge->n = 0;
sge               191 drivers/infiniband/hw/qib/qib_ud.c 			sge->vaddr =
sge               192 drivers/infiniband/hw/qib/qib_ud.c 				sge->mr->map[sge->m]->segs[sge->n].vaddr;
sge               193 drivers/infiniband/hw/qib/qib_ud.c 			sge->length =
sge               194 drivers/infiniband/hw/qib/qib_ud.c 				sge->mr->map[sge->m]->segs[sge->n].length;
sge               317 drivers/infiniband/hw/qib/qib_ud.c 	qp->s_sge.sge = wqe->sg_list[0];
sge               142 drivers/infiniband/hw/qib/qib_verbs.c 	struct rvt_sge sge = ss->sge;
sge               147 drivers/infiniband/hw/qib/qib_verbs.c 		u32 len = rvt_get_sge_length(&sge, length);
sge               149 drivers/infiniband/hw/qib/qib_verbs.c 		if (((long) sge.vaddr & (sizeof(u32) - 1)) ||
sge               155 drivers/infiniband/hw/qib/qib_verbs.c 		sge.vaddr += len;
sge               156 drivers/infiniband/hw/qib/qib_verbs.c 		sge.length -= len;
sge               157 drivers/infiniband/hw/qib/qib_verbs.c 		sge.sge_length -= len;
sge               158 drivers/infiniband/hw/qib/qib_verbs.c 		if (sge.sge_length == 0) {
sge               160 drivers/infiniband/hw/qib/qib_verbs.c 				sge = *sg_list++;
sge               161 drivers/infiniband/hw/qib/qib_verbs.c 		} else if (sge.length == 0 && sge.mr->lkey) {
sge               162 drivers/infiniband/hw/qib/qib_verbs.c 			if (++sge.n >= RVT_SEGSZ) {
sge               163 drivers/infiniband/hw/qib/qib_verbs.c 				if (++sge.m >= sge.mr->mapsz)
sge               165 drivers/infiniband/hw/qib/qib_verbs.c 				sge.n = 0;
sge               167 drivers/infiniband/hw/qib/qib_verbs.c 			sge.vaddr =
sge               168 drivers/infiniband/hw/qib/qib_verbs.c 				sge.mr->map[sge.m]->segs[sge.n].vaddr;
sge               169 drivers/infiniband/hw/qib/qib_verbs.c 			sge.length =
sge               170 drivers/infiniband/hw/qib/qib_verbs.c 				sge.mr->map[sge.m]->segs[sge.n].length;
sge               182 drivers/infiniband/hw/qib/qib_verbs.c 	struct rvt_sge *sge = &ss->sge;
sge               185 drivers/infiniband/hw/qib/qib_verbs.c 		u32 len = rvt_get_sge_length(sge, length);
sge               187 drivers/infiniband/hw/qib/qib_verbs.c 		memcpy(data, sge->vaddr, len);
sge               188 drivers/infiniband/hw/qib/qib_verbs.c 		sge->vaddr += len;
sge               189 drivers/infiniband/hw/qib/qib_verbs.c 		sge->length -= len;
sge               190 drivers/infiniband/hw/qib/qib_verbs.c 		sge->sge_length -= len;
sge               191 drivers/infiniband/hw/qib/qib_verbs.c 		if (sge->sge_length == 0) {
sge               193 drivers/infiniband/hw/qib/qib_verbs.c 				*sge = *ss->sg_list++;
sge               194 drivers/infiniband/hw/qib/qib_verbs.c 		} else if (sge->length == 0 && sge->mr->lkey) {
sge               195 drivers/infiniband/hw/qib/qib_verbs.c 			if (++sge->n >= RVT_SEGSZ) {
sge               196 drivers/infiniband/hw/qib/qib_verbs.c 				if (++sge->m >= sge->mr->mapsz)
sge               198 drivers/infiniband/hw/qib/qib_verbs.c 				sge->n = 0;
sge               200 drivers/infiniband/hw/qib/qib_verbs.c 			sge->vaddr =
sge               201 drivers/infiniband/hw/qib/qib_verbs.c 				sge->mr->map[sge->m]->segs[sge->n].vaddr;
sge               202 drivers/infiniband/hw/qib/qib_verbs.c 			sge->length =
sge               203 drivers/infiniband/hw/qib/qib_verbs.c 				sge->mr->map[sge->m]->segs[sge->n].length;
sge               437 drivers/infiniband/hw/qib/qib_verbs.c 		u32 len = rvt_get_sge_length(&ss->sge, length);
sge               441 drivers/infiniband/hw/qib/qib_verbs.c 		off = (unsigned long)ss->sge.vaddr & (sizeof(u32) - 1);
sge               443 drivers/infiniband/hw/qib/qib_verbs.c 			u32 *addr = (u32 *)((unsigned long)ss->sge.vaddr &
sge               474 drivers/infiniband/hw/qib/qib_verbs.c 			u32 *addr = (u32 *) ss->sge.vaddr;
sge               527 drivers/infiniband/hw/qib/qib_verbs.c 			qib_pio_copy(piobuf, ss->sge.vaddr, w - 1);
sge               529 drivers/infiniband/hw/qib/qib_verbs.c 			last = ((u32 *) ss->sge.vaddr)[w - 1];
sge               534 drivers/infiniband/hw/qib/qib_verbs.c 			qib_pio_copy(piobuf, ss->sge.vaddr, w);
sge               539 drivers/infiniband/hw/qib/qib_verbs.c 				u32 v = ((u32 *) ss->sge.vaddr)[w];
sge               962 drivers/infiniband/hw/qib/qib_verbs.c 	if (likely(ss->num_sge == 1 && len <= ss->sge.length &&
sge               963 drivers/infiniband/hw/qib/qib_verbs.c 		   !((unsigned long)ss->sge.vaddr & (sizeof(u32) - 1)))) {
sge               964 drivers/infiniband/hw/qib/qib_verbs.c 		u32 *addr = (u32 *) ss->sge.vaddr;
sge               625 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	struct pvrdma_sge *sge;
sge               790 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		sge = (struct pvrdma_sge *)(wqe_hdr + 1);
sge               793 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->addr = wr->sg_list[i].addr;
sge               794 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->length = wr->sg_list[i].length;
sge               795 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->lkey = wr->sg_list[i].lkey;
sge               796 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge++;
sge               835 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	struct pvrdma_sge *sge;
sge               882 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		sge = (struct pvrdma_sge *)(wqe_hdr + 1);
sge               884 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->addr = wr->sg_list[i].addr;
sge               885 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->length = wr->sg_list[i].length;
sge               886 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->lkey = wr->sg_list[i].lkey;
sge               887 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge++;
sge               536 drivers/infiniband/sw/rdmavt/mr.c 	rval = rvt_mr_has_lkey(ss->sge.mr, lkey);
sge               880 drivers/infiniband/sw/rdmavt/mr.c 				    struct ib_sge *sge)
sge               882 drivers/infiniband/sw/rdmavt/mr.c 	if (last_sge && sge->lkey == last_sge->mr->lkey &&
sge               883 drivers/infiniband/sw/rdmavt/mr.c 	    ((uint64_t)(last_sge->vaddr + last_sge->length) == sge->addr)) {
sge               884 drivers/infiniband/sw/rdmavt/mr.c 		if (sge->lkey) {
sge               885 drivers/infiniband/sw/rdmavt/mr.c 			if (unlikely((sge->addr - last_sge->mr->user_base +
sge               886 drivers/infiniband/sw/rdmavt/mr.c 			      sge->length > last_sge->mr->length)))
sge               889 drivers/infiniband/sw/rdmavt/mr.c 			last_sge->length += sge->length;
sge               891 drivers/infiniband/sw/rdmavt/mr.c 		last_sge->sge_length += sge->length;
sge               892 drivers/infiniband/sw/rdmavt/mr.c 		trace_rvt_sge_adjacent(last_sge, sge);
sge               916 drivers/infiniband/sw/rdmavt/mr.c 		struct ib_sge *sge, int acc)
sge               926 drivers/infiniband/sw/rdmavt/mr.c 	if (sge->lkey == 0) {
sge               931 drivers/infiniband/sw/rdmavt/mr.c 		if (rvt_sge_adjacent(last_sge, sge))
sge               941 drivers/infiniband/sw/rdmavt/mr.c 		isge->vaddr = (void *)sge->addr;
sge               942 drivers/infiniband/sw/rdmavt/mr.c 		isge->length = sge->length;
sge               943 drivers/infiniband/sw/rdmavt/mr.c 		isge->sge_length = sge->length;
sge               948 drivers/infiniband/sw/rdmavt/mr.c 	if (rvt_sge_adjacent(last_sge, sge))
sge               951 drivers/infiniband/sw/rdmavt/mr.c 	mr = rcu_dereference(rkt->table[sge->lkey >> rkt->shift]);
sge               959 drivers/infiniband/sw/rdmavt/mr.c 		     mr->lkey != sge->lkey || mr->pd != &pd->ibpd))
sge               962 drivers/infiniband/sw/rdmavt/mr.c 	off = sge->addr - mr->user_base;
sge               963 drivers/infiniband/sw/rdmavt/mr.c 	if (unlikely(sge->addr < mr->user_base ||
sge               964 drivers/infiniband/sw/rdmavt/mr.c 		     off + sge->length > mr->length ||
sge               997 drivers/infiniband/sw/rdmavt/mr.c 	isge->sge_length = sge->length;
sge              1001 drivers/infiniband/sw/rdmavt/mr.c 	trace_rvt_sge_new(isge, sge);
sge              1024 drivers/infiniband/sw/rdmavt/mr.c int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge,
sge              1050 drivers/infiniband/sw/rdmavt/mr.c 		sge->mr = mr;
sge              1051 drivers/infiniband/sw/rdmavt/mr.c 		sge->vaddr = (void *)vaddr;
sge              1052 drivers/infiniband/sw/rdmavt/mr.c 		sge->length = len;
sge              1053 drivers/infiniband/sw/rdmavt/mr.c 		sge->sge_length = len;
sge              1054 drivers/infiniband/sw/rdmavt/mr.c 		sge->m = 0;
sge              1055 drivers/infiniband/sw/rdmavt/mr.c 		sge->n = 0;
sge              1101 drivers/infiniband/sw/rdmavt/mr.c 	sge->mr = mr;
sge              1102 drivers/infiniband/sw/rdmavt/mr.c 	sge->vaddr = mr->map[m]->segs[n].vaddr + off;
sge              1103 drivers/infiniband/sw/rdmavt/mr.c 	sge->length = mr->map[m]->segs[n].length - off;
sge              1104 drivers/infiniband/sw/rdmavt/mr.c 	sge->sge_length = len;
sge              1105 drivers/infiniband/sw/rdmavt/mr.c 	sge->m = m;
sge              1106 drivers/infiniband/sw/rdmavt/mr.c 	sge->n = n;
sge               665 drivers/infiniband/sw/rdmavt/qp.c 		struct rvt_sge *sge = &wqe->sg_list[i];
sge               667 drivers/infiniband/sw/rdmavt/qp.c 		if (rvt_mr_has_lkey(sge->mr, lkey))
sge              2165 drivers/infiniband/sw/rdmavt/qp.c 		struct rvt_sge *sge = &wqe->sg_list[--j];
sge              2167 drivers/infiniband/sw/rdmavt/qp.c 		rvt_put_mr(sge->mr);
sge              2291 drivers/infiniband/sw/rdmavt/qp.c static struct ib_sge *rvt_cast_sge(struct rvt_wqe_sge *sge)
sge              2299 drivers/infiniband/sw/rdmavt/qp.c 	return (struct ib_sge *)sge;
sge              2324 drivers/infiniband/sw/rdmavt/qp.c 		ret = rvt_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge,
sge              2338 drivers/infiniband/sw/rdmavt/qp.c 		struct rvt_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge;
sge              2340 drivers/infiniband/sw/rdmavt/qp.c 		rvt_put_mr(sge->mr);
sge              2878 drivers/infiniband/sw/rdmavt/qp.c 	struct rvt_sge *sge = &ss->sge;
sge              2895 drivers/infiniband/sw/rdmavt/qp.c 			wss_insert(wss, sge->vaddr);
sge              2897 drivers/infiniband/sw/rdmavt/qp.c 				wss_insert(wss, (sge->vaddr + PAGE_SIZE));
sge              2916 drivers/infiniband/sw/rdmavt/qp.c 		u32 len = rvt_get_sge_length(sge, length);
sge              2922 drivers/infiniband/sw/rdmavt/qp.c 				((u8 *)sge->vaddr)[i] = ((u8 *)data)[i];
sge              2924 drivers/infiniband/sw/rdmavt/qp.c 			cacheless_memcpy(sge->vaddr, data, len);
sge              2926 drivers/infiniband/sw/rdmavt/qp.c 			memcpy(sge->vaddr, data, len);
sge              2970 drivers/infiniband/sw/rdmavt/qp.c 	struct rvt_sge *sge;
sge              3041 drivers/infiniband/sw/rdmavt/qp.c 	sqp->s_sge.sge = wqe->sg_list[0];
sge              3105 drivers/infiniband/sw/rdmavt/qp.c 		if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, wqe->length,
sge              3118 drivers/infiniband/sw/rdmavt/qp.c 		if (unlikely(!rvt_rkey_ok(qp, &sqp->s_sge.sge, wqe->length,
sge              3126 drivers/infiniband/sw/rdmavt/qp.c 		qp->r_sge.sge = wqe->sg_list[0];
sge              3136 drivers/infiniband/sw/rdmavt/qp.c 		if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64),
sge              3142 drivers/infiniband/sw/rdmavt/qp.c 		maddr = (atomic64_t *)qp->r_sge.sge.vaddr;
sge              3144 drivers/infiniband/sw/rdmavt/qp.c 		*(u64 *)sqp->s_sge.sge.vaddr =
sge              3147 drivers/infiniband/sw/rdmavt/qp.c 			(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
sge              3149 drivers/infiniband/sw/rdmavt/qp.c 		rvt_put_mr(qp->r_sge.sge.mr);
sge              3158 drivers/infiniband/sw/rdmavt/qp.c 	sge = &sqp->s_sge.sge;
sge              3160 drivers/infiniband/sw/rdmavt/qp.c 		u32 len = rvt_get_sge_length(sge, sqp->s_len);
sge              3163 drivers/infiniband/sw/rdmavt/qp.c 		rvt_copy_sge(qp, &qp->r_sge, sge->vaddr,
sge               201 drivers/infiniband/sw/rdmavt/rc.c 	ss->sge = wqe->sg_list[0];
sge               124 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
sge               125 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_ARGS(sge, isge),
sge               127 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ENTRY(ib_to_rvt(sge->mr->pd->device))
sge               129 drivers/infiniband/sw/rdmavt/trace_mr.h 		__field(struct rvt_sge *, sge)
sge               142 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ASSIGN(ib_to_rvt(sge->mr->pd->device));
sge               143 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->mr = sge->mr;
sge               144 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->sge = sge;
sge               146 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->vaddr = sge->vaddr;
sge               148 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->lkey = sge->mr->lkey;
sge               149 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->sge_length = sge->sge_length;
sge               150 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->length = sge->length;
sge               152 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->m = sge->m;
sge               153 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->n = sge->m;
sge               154 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->user = ibpd_to_rvtpd(sge->mr->pd)->user;
sge               160 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->sge,
sge               176 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
sge               177 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_ARGS(sge, isge));
sge               181 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
sge               182 drivers/infiniband/sw/rdmavt/trace_mr.h 	TP_ARGS(sge, isge));
sge               445 drivers/infiniband/sw/rxe/rxe_mr.c 	struct rxe_sge		*sge	= &dma->sge[dma->cur_sge];
sge               460 drivers/infiniband/sw/rxe/rxe_mr.c 	if (sge->length && (offset < sge->length)) {
sge               461 drivers/infiniband/sw/rxe/rxe_mr.c 		mem = lookup_mem(pd, access, sge->lkey, lookup_local);
sge               471 drivers/infiniband/sw/rxe/rxe_mr.c 		if (offset >= sge->length) {
sge               476 drivers/infiniband/sw/rxe/rxe_mr.c 			sge++;
sge               485 drivers/infiniband/sw/rxe/rxe_mr.c 			if (sge->length) {
sge               486 drivers/infiniband/sw/rxe/rxe_mr.c 				mem = lookup_mem(pd, access, sge->lkey,
sge               497 drivers/infiniband/sw/rxe/rxe_mr.c 		if (bytes > sge->length - offset)
sge               498 drivers/infiniband/sw/rxe/rxe_mr.c 			bytes = sge->length - offset;
sge               501 drivers/infiniband/sw/rxe/rxe_mr.c 			iova = sge->addr + offset;
sge               531 drivers/infiniband/sw/rxe/rxe_mr.c 	struct rxe_sge		*sge	= &dma->sge[dma->cur_sge];
sge               538 drivers/infiniband/sw/rxe/rxe_mr.c 		if (offset >= sge->length) {
sge               539 drivers/infiniband/sw/rxe/rxe_mr.c 			sge++;
sge               548 drivers/infiniband/sw/rxe/rxe_mr.c 		if (bytes > sge->length - offset)
sge               549 drivers/infiniband/sw/rxe/rxe_mr.c 			bytes = sge->length - offset;
sge               271 drivers/infiniband/sw/rxe/rxe_verbs.c 	memcpy(recv_wqe->dma.sge, ibwr->sg_list,
sge               591 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct ib_sge *sge;
sge               605 drivers/infiniband/sw/rxe/rxe_verbs.c 		sge = ibwr->sg_list;
sge               606 drivers/infiniband/sw/rxe/rxe_verbs.c 		for (i = 0; i < num_sge; i++, sge++) {
sge               607 drivers/infiniband/sw/rxe/rxe_verbs.c 			memcpy(p, (void *)(uintptr_t)sge->addr,
sge               608 drivers/infiniband/sw/rxe/rxe_verbs.c 					sge->length);
sge               610 drivers/infiniband/sw/rxe/rxe_verbs.c 			p += sge->length;
sge               617 drivers/infiniband/sw/rxe/rxe_verbs.c 		memcpy(wqe->dma.sge, ibwr->sg_list,
sge               222 drivers/infiniband/sw/rxe/rxe_verbs.h 		struct ib_sge		sge[RXE_MAX_SGE];
sge               222 drivers/infiniband/sw/siw/siw_mem.c int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[],
sge               229 drivers/infiniband/sw/siw/siw_mem.c 	if (len + off > sge->length) {
sge               234 drivers/infiniband/sw/siw/siw_mem.c 		new = siw_mem_id2obj(sdev, sge->lkey >> 8);
sge               236 drivers/infiniband/sw/siw/siw_mem.c 			siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey);
sge               243 drivers/infiniband/sw/siw/siw_mem.c 	if (unlikely((*mem)->stag != sge->lkey)) {
sge               244 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_mem((*mem), "STag mismatch: 0x%08x\n", sge->lkey);
sge               248 drivers/infiniband/sw/siw/siw_mem.c 	rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len);
sge                18 drivers/infiniband/sw/siw/siw_mem.h int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge,
sge               277 drivers/infiniband/sw/siw/siw_qp.c 	wqe->sqe.sge[0].length = 0;
sge               278 drivers/infiniband/sw/siw/siw_qp.c 	wqe->sqe.sge[0].laddr = 0;
sge               279 drivers/infiniband/sw/siw/siw_qp.c 	wqe->sqe.sge[0].lkey = 0;
sge               479 drivers/infiniband/sw/siw/siw_qp.c 				rreq->ddp_msn = htonl(wqe->sqe.sge[0].length);
sge               484 drivers/infiniband/sw/siw/siw_qp.c 				rreq->read_size = htonl(wqe->sqe.sge[0].length);
sge               485 drivers/infiniband/sw/siw/siw_qp.c 				rreq->source_stag = htonl(wqe->sqe.sge[0].lkey);
sge               487 drivers/infiniband/sw/siw/siw_qp.c 					cpu_to_be64(wqe->sqe.sge[0].laddr);
sge               872 drivers/infiniband/sw/siw/siw_qp.c 	rreq->sge[0].laddr = sqe->sge[0].laddr;
sge               873 drivers/infiniband/sw/siw/siw_qp.c 	rreq->sge[0].length = sqe->sge[0].length;
sge               874 drivers/infiniband/sw/siw/siw_qp.c 	rreq->sge[0].lkey = sqe->sge[0].lkey;
sge               875 drivers/infiniband/sw/siw/siw_qp.c 	rreq->sge[1].lkey = sqe->sge[1].lkey;
sge               913 drivers/infiniband/sw/siw/siw_qp.c 			wqe->sqe.sge[0].length = irqe->sge[0].length;
sge               914 drivers/infiniband/sw/siw/siw_qp.c 			wqe->sqe.sge[0].laddr = irqe->sge[0].laddr;
sge               915 drivers/infiniband/sw/siw/siw_qp.c 			wqe->sqe.sge[0].lkey = irqe->sge[0].lkey;
sge               923 drivers/infiniband/sw/siw/siw_qp.c 		wqe->sqe.sge[1].length = irqe->sge[1].length;
sge               955 drivers/infiniband/sw/siw/siw_qp.c 			if (wqe->sqe.sge[0].length > SIW_MAX_INLINE) {
sge               959 drivers/infiniband/sw/siw/siw_qp.c 			wqe->sqe.sge[0].laddr = (uintptr_t)&wqe->sqe.sge[1];
sge               960 drivers/infiniband/sw/siw/siw_qp.c 			wqe->sqe.sge[0].lkey = 0;
sge               175 drivers/infiniband/sw/siw/siw_qp_rx.c 		srx->ddp_stag = wqe->sqe.sge[0].lkey;
sge               176 drivers/infiniband/sw/siw/siw_qp_rx.c 		srx->ddp_to = wqe->sqe.sge[0].laddr;
sge               366 drivers/infiniband/sw/siw/siw_qp_rx.c 				wqe->rqe.sge[i].laddr = rqe->sge[i].laddr;
sge               367 drivers/infiniband/sw/siw/siw_qp_rx.c 				wqe->rqe.sge[i].lkey = rqe->sge[i].lkey;
sge               368 drivers/infiniband/sw/siw/siw_qp_rx.c 				wqe->rqe.sge[i].length = rqe->sge[i].length;
sge               369 drivers/infiniband/sw/siw/siw_qp_rx.c 				wqe->bytes += wqe->rqe.sge[i].length;
sge               456 drivers/infiniband/sw/siw/siw_qp_rx.c 		struct siw_sge *sge;
sge               459 drivers/infiniband/sw/siw/siw_qp_rx.c 		sge = &wqe->rqe.sge[frx->sge_idx];
sge               461 drivers/infiniband/sw/siw/siw_qp_rx.c 		if (!sge->length) {
sge               468 drivers/infiniband/sw/siw/siw_qp_rx.c 		sge_bytes = min(data_bytes, sge->length - frx->sge_off);
sge               476 drivers/infiniband/sw/siw/siw_qp_rx.c 		rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE,
sge               489 drivers/infiniband/sw/siw/siw_qp_rx.c 				(void *)(uintptr_t)(sge->laddr + frx->sge_off),
sge               493 drivers/infiniband/sw/siw/siw_qp_rx.c 					 sge->laddr + frx->sge_off, sge_bytes);
sge               496 drivers/infiniband/sw/siw/siw_qp_rx.c 					sge->laddr + frx->sge_off, sge_bytes);
sge               508 drivers/infiniband/sw/siw/siw_qp_rx.c 		if (frx->sge_off == sge->length) {
sge               698 drivers/infiniband/sw/siw/siw_qp_rx.c 		resp->sge[0].length = length;
sge               699 drivers/infiniband/sw/siw/siw_qp_rx.c 		resp->sge[0].laddr = laddr;
sge               700 drivers/infiniband/sw/siw/siw_qp_rx.c 		resp->sge[0].lkey = lkey;
sge               705 drivers/infiniband/sw/siw/siw_qp_rx.c 		resp->sge[1].length = msn;
sge               751 drivers/infiniband/sw/siw/siw_qp_rx.c 		wqe->sqe.sge[0].laddr = orqe->sge[0].laddr;
sge               752 drivers/infiniband/sw/siw/siw_qp_rx.c 		wqe->sqe.sge[0].lkey = orqe->sge[0].lkey;
sge               753 drivers/infiniband/sw/siw/siw_qp_rx.c 		wqe->sqe.sge[0].length = orqe->sge[0].length;
sge               756 drivers/infiniband/sw/siw/siw_qp_rx.c 		wqe->bytes = orqe->sge[0].length;
sge               783 drivers/infiniband/sw/siw/siw_qp_rx.c 	struct siw_sge *sge;
sge               818 drivers/infiniband/sw/siw/siw_qp_rx.c 	sge = wqe->sqe.sge; /* there is only one */
sge               825 drivers/infiniband/sw/siw/siw_qp_rx.c 		rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0,
sge               846 drivers/infiniband/sw/siw/siw_qp_rx.c 			(void *)(uintptr_t)(sge->laddr + wqe->processed),
sge               849 drivers/infiniband/sw/siw/siw_qp_rx.c 		rv = siw_rx_umem(srx, mem_p->umem, sge->laddr + wqe->processed,
sge               853 drivers/infiniband/sw/siw/siw_qp_rx.c 				sge->laddr + wqe->processed, bytes);
sge              1272 drivers/infiniband/sw/siw/siw_qp_rx.c 			rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey);
sge                43 drivers/infiniband/sw/siw/siw_qp_tx.c 	struct siw_sge *sge = &wqe->sqe.sge[0];
sge                44 drivers/infiniband/sw/siw/siw_qp_tx.c 	u32 bytes = sge->length;
sge                53 drivers/infiniband/sw/siw/siw_qp_tx.c 		memcpy(paddr, &wqe->sqe.sge[1], bytes);
sge                60 drivers/infiniband/sw/siw/siw_qp_tx.c 			       (const void *)(uintptr_t)sge->laddr, bytes);
sge                62 drivers/infiniband/sw/siw/siw_qp_tx.c 			if (copy_from_user(paddr, u64_to_user_ptr(sge->laddr),
sge                66 drivers/infiniband/sw/siw/siw_qp_tx.c 			unsigned int off = sge->laddr & ~PAGE_MASK;
sge                72 drivers/infiniband/sw/siw/siw_qp_tx.c 				p = siw_get_upage(mem->umem, sge->laddr);
sge                74 drivers/infiniband/sw/siw/siw_qp_tx.c 				p = siw_get_pblpage(mem, sge->laddr, &pbl_idx);
sge                91 drivers/infiniband/sw/siw/siw_qp_tx.c 							  sge->laddr + part);
sge                94 drivers/infiniband/sw/siw/siw_qp_tx.c 							    sge->laddr + part,
sge               137 drivers/infiniband/sw/siw/siw_qp_tx.c 		c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey);
sge               139 drivers/infiniband/sw/siw/siw_qp_tx.c 			cpu_to_be64(wqe->sqe.sge[0].laddr);
sge               142 drivers/infiniband/sw/siw/siw_qp_tx.c 		c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length);
sge               370 drivers/infiniband/sw/siw/siw_qp_tx.c 			struct siw_sge *sge, unsigned int offset,
sge               374 drivers/infiniband/sw/siw/siw_qp_tx.c 	int sge_bytes = min(sge->length - offset, size);
sge               376 drivers/infiniband/sw/siw/siw_qp_tx.c 	offset = (sge->laddr + offset) & ~PAGE_MASK;
sge               386 drivers/infiniband/sw/siw/siw_qp_tx.c 			sge++;
sge               387 drivers/infiniband/sw/siw/siw_qp_tx.c 			sge_bytes = min(sge->length, size - sent);
sge               388 drivers/infiniband/sw/siw/siw_qp_tx.c 			offset = sge->laddr & ~PAGE_MASK;
sge               427 drivers/infiniband/sw/siw/siw_qp_tx.c 	struct siw_sge *sge = &wqe->sqe.sge[c_tx->sge_idx];
sge               457 drivers/infiniband/sw/siw/siw_qp_tx.c 		unsigned int sge_len = min(sge->length - sge_off, data_len);
sge               458 drivers/infiniband/sw/siw/siw_qp_tx.c 		unsigned int fp_off = (sge->laddr + sge_off) & ~PAGE_MASK;
sge               473 drivers/infiniband/sw/siw/siw_qp_tx.c 				(void *)(uintptr_t)(sge->laddr + sge_off);
sge               494 drivers/infiniband/sw/siw/siw_qp_tx.c 						mem, sge->laddr + sge_off,
sge               498 drivers/infiniband/sw/siw/siw_qp_tx.c 							  sge->laddr + sge_off);
sge               526 drivers/infiniband/sw/siw/siw_qp_tx.c 				u64 va = sge->laddr + sge_off;
sge               551 drivers/infiniband/sw/siw/siw_qp_tx.c 		if (sge_off == sge->length &&
sge               554 drivers/infiniband/sw/siw/siw_qp_tx.c 			sge++;
sge               582 drivers/infiniband/sw/siw/siw_qp_tx.c 		rv = siw_0copy_tx(s, page_array, &wqe->sqe.sge[c_tx->sge_idx],
sge               640 drivers/infiniband/sw/siw/siw_qp_tx.c 			sge = &wqe->sqe.sge[c_tx->sge_idx];
sge               641 drivers/infiniband/sw/siw/siw_qp_tx.c 			sge_unsent = sge->length - c_tx->sge_off;
sge               647 drivers/infiniband/sw/siw/siw_qp_tx.c 				sge++;
sge               648 drivers/infiniband/sw/siw/siw_qp_tx.c 				sge_unsent = sge->length;
sge               754 drivers/infiniband/sw/siw/siw_qp_tx.c 	struct siw_sge *sge = &wqe->sqe.sge[0];
sge               760 drivers/infiniband/sw/siw/siw_qp_tx.c 	for (i = 0, len = 0; num_sge; num_sge--, i++, sge++) {
sge               764 drivers/infiniband/sw/siw/siw_qp_tx.c 		if (sge->length) {
sge               765 drivers/infiniband/sw/siw/siw_qp_tx.c 			int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0,
sge               766 drivers/infiniband/sw/siw/siw_qp_tx.c 					       sge->length);
sge               771 drivers/infiniband/sw/siw/siw_qp_tx.c 		len += sge->length;
sge               819 drivers/infiniband/sw/siw/siw_qp_tx.c 			wqe->bytes = wqe->sqe.sge[0].length;
sge               825 drivers/infiniband/sw/siw/siw_qp_tx.c 				wqe->sqe.sge[0].laddr =
sge               826 drivers/infiniband/sw/siw/siw_qp_tx.c 					(u64)(uintptr_t)&wqe->sqe.sge[1];
sge               660 drivers/infiniband/sw/siw/siw_verbs.c 	void *kbuf = &sqe->sge[1];
sge               663 drivers/infiniband/sw/siw/siw_verbs.c 	sqe->sge[0].laddr = (uintptr_t)kbuf;
sge               664 drivers/infiniband/sw/siw/siw_verbs.c 	sqe->sge[0].lkey = 0;
sge               682 drivers/infiniband/sw/siw/siw_verbs.c 	sqe->sge[0].length = bytes > 0 ? bytes : 0;
sge               828 drivers/infiniband/sw/siw/siw_verbs.c 				siw_copy_sgl(wr->sg_list, sqe->sge,
sge               861 drivers/infiniband/sw/siw/siw_verbs.c 			siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1);
sge               877 drivers/infiniband/sw/siw/siw_verbs.c 				siw_copy_sgl(wr->sg_list, &sqe->sge[0],
sge              1060 drivers/infiniband/sw/siw/siw_verbs.c 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
sge              1781 drivers/infiniband/sw/siw/siw_verbs.c 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
sge                24 drivers/infiniband/sw/siw/siw_verbs.h static inline void siw_copy_sgl(struct ib_sge *sge, struct siw_sge *siw_sge,
sge                28 drivers/infiniband/sw/siw/siw_verbs.h 		siw_sge->laddr = sge->addr;
sge                29 drivers/infiniband/sw/siw/siw_verbs.h 		siw_sge->length = sge->length;
sge                30 drivers/infiniband/sw/siw/siw_verbs.h 		siw_sge->lkey = sge->lkey;
sge                33 drivers/infiniband/sw/siw/siw_verbs.h 		sge++;
sge               119 drivers/infiniband/ulp/ipoib/ipoib_cm.c 					struct ib_sge *sge, int id)
sge               127 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		sge[i].addr = rx->rx_ring[id].mapping[i];
sge               330 drivers/infiniband/ulp/ipoib/ipoib_cm.c 				struct ib_sge *sge)
sge               336 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		sge[i].lkey = priv->pd->local_dma_lkey;
sge               338 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	sge[0].length = IPOIB_CM_HEAD_SIZE;
sge               340 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		sge[i].length = PAGE_SIZE;
sge               343 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->sg_list = sge;
sge               353 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		struct ib_sge sge[IPOIB_CM_RX_SG];
sge               369 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	ipoib_cm_init_rx_wr(dev, &t->wr, t->sge);
sge               391 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i);
sge               218 drivers/infiniband/ulp/iser/iscsi_iser.h 	struct ib_sge	 sge;
sge               297 drivers/infiniband/ulp/iser/iscsi_iser.h 	struct ib_sge                sge;
sge                84 drivers/infiniband/ulp/iser/iser_initiator.c 	hdr->read_va   = cpu_to_be64(mem_reg->sge.addr);
sge                88 drivers/infiniband/ulp/iser/iser_initiator.c 		 (unsigned long long)mem_reg->sge.addr);
sge               142 drivers/infiniband/ulp/iser/iser_initiator.c 			hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz);
sge               147 drivers/infiniband/ulp/iser/iser_initiator.c 			 (unsigned long long)mem_reg->sge.addr, unsol_sz);
sge               153 drivers/infiniband/ulp/iser/iser_initiator.c 		tx_dsg->addr = mem_reg->sge.addr;
sge               155 drivers/infiniband/ulp/iser/iser_initiator.c 		tx_dsg->lkey = mem_reg->sge.lkey;
sge               470 drivers/infiniband/ulp/iser/iser_initiator.c 	tx_dsg->addr = mem_reg->sge.addr + buf_offset;
sge               472 drivers/infiniband/ulp/iser/iser_initiator.c 	tx_dsg->lkey = mem_reg->sge.lkey;
sge               196 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.lkey = device->pd->local_dma_lkey;
sge               206 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.addr = sg_dma_address(&sg[0]);
sge               207 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.length = sg_dma_len(&sg[0]);
sge               210 drivers/infiniband/ulp/iser/iser_memory.c 		 " length=0x%x\n", reg->sge.lkey, reg->rkey,
sge               211 drivers/infiniband/ulp/iser/iser_memory.c 		 reg->sge.addr, reg->sge.length);
sge               258 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.lkey = fmr->fmr->lkey;
sge               260 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.addr = page_vec->fake_mr.iova;
sge               261 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.length = page_vec->fake_mr.length;
sge               265 drivers/infiniband/ulp/iser/iser_memory.c 		 " length=0x%x\n", reg->sge.lkey, reg->rkey,
sge               266 drivers/infiniband/ulp/iser/iser_memory.c 		 reg->sge.addr, reg->sge.length);
sge               426 drivers/infiniband/ulp/iser/iser_memory.c 	sig_reg->sge.lkey = mr->lkey;
sge               428 drivers/infiniband/ulp/iser/iser_memory.c 	sig_reg->sge.addr = mr->iova;
sge               429 drivers/infiniband/ulp/iser/iser_memory.c 	sig_reg->sge.length = mr->length;
sge               432 drivers/infiniband/ulp/iser/iser_memory.c 		 sig_reg->sge.lkey, sig_reg->rkey, sig_reg->sge.addr,
sge               433 drivers/infiniband/ulp/iser/iser_memory.c 		 sig_reg->sge.length);
sge               474 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.lkey = mr->lkey;
sge               476 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.addr = mr->iova;
sge               477 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.length = mr->length;
sge               480 drivers/infiniband/ulp/iser/iser_memory.c 		 reg->sge.lkey, reg->rkey, reg->sge.addr, reg->sge.length);
sge               979 drivers/infiniband/ulp/iser/iser_verbs.c 	desc->sge.addr = desc->rsp_dma;
sge               980 drivers/infiniband/ulp/iser/iser_verbs.c 	desc->sge.length = ISER_RX_LOGIN_SIZE;
sge               981 drivers/infiniband/ulp/iser/iser_verbs.c 	desc->sge.lkey = ib_conn->device->pd->local_dma_lkey;
sge               985 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.sg_list = &desc->sge;
sge               963 drivers/infiniband/ulp/isert/ib_isert.c 	struct ib_sge sge;
sge               966 drivers/infiniband/ulp/isert/ib_isert.c 	memset(&sge, 0, sizeof(struct ib_sge));
sge               967 drivers/infiniband/ulp/isert/ib_isert.c 	sge.addr = isert_conn->login_req_dma;
sge               968 drivers/infiniband/ulp/isert/ib_isert.c 	sge.length = ISER_RX_PAYLOAD_SIZE;
sge               969 drivers/infiniband/ulp/isert/ib_isert.c 	sge.lkey = isert_conn->device->pd->local_dma_lkey;
sge               972 drivers/infiniband/ulp/isert/ib_isert.c 		sge.addr, sge.length, sge.lkey);
sge               978 drivers/infiniband/ulp/isert/ib_isert.c 	rx_wr.sg_list = &sge;
sge              1845 drivers/infiniband/ulp/srp/ib_srp.c 		struct ib_sge *sge = &req->cmd->sge[1];
sge              1854 drivers/infiniband/ulp/srp/ib_srp.c 			sge[i].addr   = sg_dma_address(sg);
sge              1855 drivers/infiniband/ulp/srp/ib_srp.c 			sge[i].length = sg_dma_len(sg);
sge              1856 drivers/infiniband/ulp/srp/ib_srp.c 			sge[i].lkey   = target->lkey;
sge              2079 drivers/infiniband/ulp/srp/ib_srp.c 	iu->sge[0].addr   = iu->dma;
sge              2080 drivers/infiniband/ulp/srp/ib_srp.c 	iu->sge[0].length = len;
sge              2081 drivers/infiniband/ulp/srp/ib_srp.c 	iu->sge[0].lkey   = target->lkey;
sge              2087 drivers/infiniband/ulp/srp/ib_srp.c 	wr.sg_list    = &iu->sge[0];
sge               277 drivers/infiniband/ulp/srp/ib_srp.h 	struct ib_sge		sge[SRP_MAX_SGE];
sge              2798 drivers/infiniband/ulp/srpt/ib_srpt.c 	struct ib_sge sge;
sge              2858 drivers/infiniband/ulp/srpt/ib_srpt.c 	sge.addr = ioctx->ioctx.dma;
sge              2859 drivers/infiniband/ulp/srpt/ib_srpt.c 	sge.length = resp_len;
sge              2860 drivers/infiniband/ulp/srpt/ib_srpt.c 	sge.lkey = sdev->lkey;
sge              2865 drivers/infiniband/ulp/srpt/ib_srpt.c 	send_wr.sg_list = &sge;
sge               334 drivers/message/fusion/mptbase.h 	SGESimple32_t	 sge[1];
sge               338 drivers/message/fusion/mptbase.h 	SGESimple64_t	 sge[1];
sge               550 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	struct eth_rx_sge *sge = &fp->rx_sge_ring[index];
sge               574 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	sge->addr_hi = cpu_to_le32(U64_HI(mapping));
sge               575 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	sge->addr_lo = cpu_to_le32(U64_LO(mapping));
sge              1342 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		struct eth_rx_sge *sge;
sge              1344 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2];
sge              1345 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		sge->addr_hi =
sge              1349 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		sge->addr_lo =
sge               806 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	struct eth_rx_sge *sge = &fp->rx_sge_ring[index];
sge               821 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	sge->addr_hi = 0;
sge               822 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	sge->addr_lo = 0;
sge              3097 drivers/net/ethernet/broadcom/cnic_defs.h 	struct fcoe_bd_ctx sge[3];
sge               190 drivers/net/ethernet/chelsio/cxgb/common.h 	struct sge_params sge;
sge               228 drivers/net/ethernet/chelsio/cxgb/common.h struct sge;
sge               246 drivers/net/ethernet/chelsio/cxgb/common.h 	struct sge    *sge;
sge               170 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 			t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed);
sge               224 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	t1_sge_start(adapter->sge);
sge               235 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	t1_sge_stop(adapter->sge);
sge               463 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	t = t1_sge_get_intr_counts(adapter->sge);
sge               464 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss);
sge               722 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl];
sge               723 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl];
sge               724 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	e->tx_pending = adapter->params.sge.cmdQ_size[0];
sge               743 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending;
sge               744 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending;
sge               745 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.cmdQ_size[0] = e->tx_pending;
sge               746 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ?
sge               755 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.rx_coalesce_usecs = c->rx_coalesce_usecs;
sge               756 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.coalesce_enable = c->use_adaptive_rx_coalesce;
sge               757 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	adapter->params.sge.sample_interval_usecs = c->rate_sample_interval;
sge               758 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	t1_sge_set_coalesce_params(adapter->sge, &adapter->params.sge);
sge               766 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	c->rx_coalesce_usecs = adapter->params.sge.rx_coalesce_usecs;
sge               767 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	c->rate_sample_interval = adapter->params.sge.sample_interval_usecs;
sge               768 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	c->use_adaptive_rx_coalesce = adapter->params.sge.coalesce_enable;
sge               955 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 		t1_sge_stop(adapter->sge);
sge               282 drivers/net/ethernet/chelsio/cxgb/sge.c static void tx_sched_stop(struct sge *sge)
sge               284 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               297 drivers/net/ethernet/chelsio/cxgb/sge.c unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port,
sge               300 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               320 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) {
sge               343 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val)
sge               345 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               350 drivers/net/ethernet/chelsio/cxgb/sge.c 		t1_sched_update_parms(sge, i, 0, 0);
sge               357 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port,
sge               360 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               363 drivers/net/ethernet/chelsio/cxgb/sge.c 	t1_sched_update_parms(sge, port, 0, 0);
sge               371 drivers/net/ethernet/chelsio/cxgb/sge.c static int tx_sched_init(struct sge *sge)
sge               381 drivers/net/ethernet/chelsio/cxgb/sge.c 	tasklet_init(&s->sched_tsk, restart_sched, (unsigned long) sge);
sge               382 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->tx_sched = s;
sge               386 drivers/net/ethernet/chelsio/cxgb/sge.c 		t1_sched_update_parms(sge, i, 1500, 1000);
sge               397 drivers/net/ethernet/chelsio/cxgb/sge.c static inline int sched_update_avail(struct sge *sge)
sge               399 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               431 drivers/net/ethernet/chelsio/cxgb/sge.c static struct sk_buff *sched_skb(struct sge *sge, struct sk_buff *skb,
sge               434 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sched *s = sge->tx_sched;
sge               472 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (update-- && sched_update_avail(sge))
sge               480 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct cmdQ *q = &sge->cmdQ[0];
sge               484 drivers/net/ethernet/chelsio/cxgb/sge.c 			writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL);
sge               525 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_rx_resources(struct sge *sge)
sge               527 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               530 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->respQ.entries) {
sge               531 drivers/net/ethernet/chelsio/cxgb/sge.c 		size = sizeof(struct respQ_e) * sge->respQ.size;
sge               532 drivers/net/ethernet/chelsio/cxgb/sge.c 		pci_free_consistent(pdev, size, sge->respQ.entries,
sge               533 drivers/net/ethernet/chelsio/cxgb/sge.c 				    sge->respQ.dma_addr);
sge               537 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct freelQ *q = &sge->freelQ[i];
sge               555 drivers/net/ethernet/chelsio/cxgb/sge.c static int alloc_rx_resources(struct sge *sge, struct sge_params *p)
sge               557 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               561 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct freelQ *q = &sge->freelQ[i];
sge               565 drivers/net/ethernet/chelsio/cxgb/sge.c 		q->dma_offset = sge->rx_pkt_pad ? 0 : NET_IP_ALIGN;
sge               584 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->freelQ[!sge->jumbo_fl].rx_buffer_size = SGE_RX_SM_BUF_SIZE +
sge               586 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->freelQ[!sge->jumbo_fl].dma_offset;
sge               590 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->freelQ[sge->jumbo_fl].rx_buffer_size = size;
sge               596 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->freelQ[!sge->jumbo_fl].recycleq_idx = 0;
sge               597 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->freelQ[sge->jumbo_fl].recycleq_idx = 1;
sge               599 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->respQ.genbit = 1;
sge               600 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->respQ.size = SGE_RESPQ_E_N;
sge               601 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->respQ.credits = 0;
sge               602 drivers/net/ethernet/chelsio/cxgb/sge.c 	size = sizeof(struct respQ_e) * sge->respQ.size;
sge               603 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->respQ.entries =
sge               604 drivers/net/ethernet/chelsio/cxgb/sge.c 		pci_alloc_consistent(pdev, size, &sge->respQ.dma_addr);
sge               605 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (!sge->respQ.entries)
sge               610 drivers/net/ethernet/chelsio/cxgb/sge.c 	free_rx_resources(sge);
sge               617 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_cmdQ_buffers(struct sge *sge, struct cmdQ *q, unsigned int n)
sge               620 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               651 drivers/net/ethernet/chelsio/cxgb/sge.c static void free_tx_resources(struct sge *sge)
sge               653 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               657 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct cmdQ *q = &sge->cmdQ[i];
sge               661 drivers/net/ethernet/chelsio/cxgb/sge.c 				free_cmdQ_buffers(sge, q, q->in_use);
sge               675 drivers/net/ethernet/chelsio/cxgb/sge.c static int alloc_tx_resources(struct sge *sge, struct sge_params *p)
sge               677 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               681 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct cmdQ *q = &sge->cmdQ[i];
sge               709 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->cmdQ[0].stop_thres = sge->adapter->params.nports *
sge               714 drivers/net/ethernet/chelsio/cxgb/sge.c 	free_tx_resources(sge);
sge               732 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge               735 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->sge_control |= F_VLAN_XTRACT;
sge               737 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->sge_control &= ~F_VLAN_XTRACT;
sge               739 drivers/net/ethernet/chelsio/cxgb/sge.c 		writel(sge->sge_control, adapter->regs + A_SG_CONTROL);
sge               748 drivers/net/ethernet/chelsio/cxgb/sge.c static void configure_sge(struct sge *sge, struct sge_params *p)
sge               750 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *ap = sge->adapter;
sge               753 drivers/net/ethernet/chelsio/cxgb/sge.c 	setup_ring_params(ap, sge->cmdQ[0].dma_addr, sge->cmdQ[0].size,
sge               755 drivers/net/ethernet/chelsio/cxgb/sge.c 	setup_ring_params(ap, sge->cmdQ[1].dma_addr, sge->cmdQ[1].size,
sge               757 drivers/net/ethernet/chelsio/cxgb/sge.c 	setup_ring_params(ap, sge->freelQ[0].dma_addr,
sge               758 drivers/net/ethernet/chelsio/cxgb/sge.c 			  sge->freelQ[0].size, A_SG_FL0BASELWR,
sge               760 drivers/net/ethernet/chelsio/cxgb/sge.c 	setup_ring_params(ap, sge->freelQ[1].dma_addr,
sge               761 drivers/net/ethernet/chelsio/cxgb/sge.c 			  sge->freelQ[1].size, A_SG_FL1BASELWR,
sge               767 drivers/net/ethernet/chelsio/cxgb/sge.c 	setup_ring_params(ap, sge->respQ.dma_addr, sge->respQ.size,
sge               769 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel((u32)sge->respQ.size - 1, ap->regs + A_SG_RSPQUEUECREDIT);
sge               771 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->sge_control = F_CMDQ0_ENABLE | F_CMDQ1_ENABLE | F_FL0_ENABLE |
sge               774 drivers/net/ethernet/chelsio/cxgb/sge.c 		V_RX_PKT_OFFSET(sge->rx_pkt_pad);
sge               777 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->sge_control |= F_ENABLE_BIG_ENDIAN;
sge               781 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->intrtimer_nres = SGE_INTRTIMER_NRES * core_ticks_per_usec(ap);
sge               783 drivers/net/ethernet/chelsio/cxgb/sge.c 	t1_sge_set_coalesce_params(sge, p);
sge               789 drivers/net/ethernet/chelsio/cxgb/sge.c static inline unsigned int jumbo_payload_capacity(const struct sge *sge)
sge               791 drivers/net/ethernet/chelsio/cxgb/sge.c 	return sge->freelQ[sge->jumbo_fl].rx_buffer_size -
sge               792 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->freelQ[sge->jumbo_fl].dma_offset -
sge               799 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_destroy(struct sge *sge)
sge               803 drivers/net/ethernet/chelsio/cxgb/sge.c 	for_each_port(sge->adapter, i)
sge               804 drivers/net/ethernet/chelsio/cxgb/sge.c 		free_percpu(sge->port_stats[i]);
sge               806 drivers/net/ethernet/chelsio/cxgb/sge.c 	kfree(sge->tx_sched);
sge               807 drivers/net/ethernet/chelsio/cxgb/sge.c 	free_tx_resources(sge);
sge               808 drivers/net/ethernet/chelsio/cxgb/sge.c 	free_rx_resources(sge);
sge               809 drivers/net/ethernet/chelsio/cxgb/sge.c 	kfree(sge);
sge               824 drivers/net/ethernet/chelsio/cxgb/sge.c static void refill_free_list(struct sge *sge, struct freelQ *q)
sge               826 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct pci_dev *pdev = sge->adapter->pdev;
sge               842 drivers/net/ethernet/chelsio/cxgb/sge.c 		skb_reserve(skb, sge->rx_pkt_pad);
sge               870 drivers/net/ethernet/chelsio/cxgb/sge.c static void freelQs_empty(struct sge *sge)
sge               872 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge               876 drivers/net/ethernet/chelsio/cxgb/sge.c 	refill_free_list(sge, &sge->freelQ[0]);
sge               877 drivers/net/ethernet/chelsio/cxgb/sge.c 	refill_free_list(sge, &sge->freelQ[1]);
sge               879 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->freelQ[0].credits > (sge->freelQ[0].size >> 2) &&
sge               880 drivers/net/ethernet/chelsio/cxgb/sge.c 	    sge->freelQ[1].credits > (sge->freelQ[1].size >> 2)) {
sge               882 drivers/net/ethernet/chelsio/cxgb/sge.c 		irqholdoff_reg = sge->fixed_intrtimer;
sge               886 drivers/net/ethernet/chelsio/cxgb/sge.c 		irqholdoff_reg = sge->intrtimer_nres;
sge               903 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_disable(struct sge *sge)
sge               905 drivers/net/ethernet/chelsio/cxgb/sge.c 	u32 val = readl(sge->adapter->regs + A_PL_ENABLE);
sge               907 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(val & ~SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE);
sge               908 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(0, sge->adapter->regs + A_SG_INT_ENABLE);
sge               914 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_enable(struct sge *sge)
sge               917 drivers/net/ethernet/chelsio/cxgb/sge.c 	u32 val = readl(sge->adapter->regs + A_PL_ENABLE);
sge               919 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->adapter->port[0].dev->hw_features & NETIF_F_TSO)
sge               921 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(en, sge->adapter->regs + A_SG_INT_ENABLE);
sge               922 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(val | SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE);
sge               928 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_intr_clear(struct sge *sge)
sge               930 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_CAUSE);
sge               931 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(0xffffffff, sge->adapter->regs + A_SG_INT_CAUSE);
sge               937 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_intr_error_handler(struct sge *sge)
sge               939 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge               945 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.respQ_empty++;
sge               947 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.respQ_overflow++;
sge               952 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.freelistQ_empty++;
sge               953 drivers/net/ethernet/chelsio/cxgb/sge.c 		freelQs_empty(sge);
sge               956 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.pkt_too_big++;
sge               961 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.pkt_mismatch++;
sge               971 drivers/net/ethernet/chelsio/cxgb/sge.c const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge)
sge               973 drivers/net/ethernet/chelsio/cxgb/sge.c 	return &sge->stats;
sge               976 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_get_port_stats(const struct sge *sge, int port,
sge               983 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct sge_port_stats *st = per_cpu_ptr(sge->port_stats[port], cpu);
sge              1288 drivers/net/ethernet/chelsio/cxgb/sge.c static inline void reclaim_completed_tx(struct sge *sge, struct cmdQ *q)
sge              1295 drivers/net/ethernet/chelsio/cxgb/sge.c 		free_cmdQ_buffers(sge, q, reclaim);
sge              1306 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = (struct sge *) arg;
sge              1307 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge              1308 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct cmdQ *q = &sge->cmdQ[0];
sge              1313 drivers/net/ethernet/chelsio/cxgb/sge.c 	reclaim_completed_tx(sge, q);
sge              1317 drivers/net/ethernet/chelsio/cxgb/sge.c 	while ((skb = sched_skb(sge, NULL, credits)) != NULL) {
sge              1352 drivers/net/ethernet/chelsio/cxgb/sge.c static void sge_rx(struct sge *sge, struct freelQ *fl, unsigned int len)
sge              1356 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge              1360 drivers/net/ethernet/chelsio/cxgb/sge.c 	skb = get_packet(adapter, fl, len - sge->rx_pkt_pad);
sge              1362 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.rx_drops++;
sge              1373 drivers/net/ethernet/chelsio/cxgb/sge.c 	st = this_cpu_ptr(sge->port_stats[p->iff]);
sge              1407 drivers/net/ethernet/chelsio/cxgb/sge.c static void restart_tx_queues(struct sge *sge)
sge              1409 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adap = sge->adapter;
sge              1412 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (!enough_free_Tx_descs(&sge->cmdQ[0]))
sge              1418 drivers/net/ethernet/chelsio/cxgb/sge.c 		if (test_and_clear_bit(nd->if_port, &sge->stopped_tx_queues) &&
sge              1420 drivers/net/ethernet/chelsio/cxgb/sge.c 			sge->stats.cmdQ_restarted[2]++;
sge              1434 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1435 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct cmdQ *cmdq = &sge->cmdQ[0];
sge              1439 drivers/net/ethernet/chelsio/cxgb/sge.c 		freelQs_empty(sge);
sge              1450 drivers/net/ethernet/chelsio/cxgb/sge.c 		if (sge->tx_sched)
sge              1451 drivers/net/ethernet/chelsio/cxgb/sge.c 			tasklet_hi_schedule(&sge->tx_sched->sched_tsk);
sge              1456 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (unlikely(sge->stopped_tx_queues != 0))
sge              1457 drivers/net/ethernet/chelsio/cxgb/sge.c 		restart_tx_queues(sge);
sge              1468 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1469 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct respQ *q = &sge->respQ;
sge              1491 drivers/net/ethernet/chelsio/cxgb/sge.c 			sge->cmdQ[1].processed += cmdq_processed[1];
sge              1496 drivers/net/ethernet/chelsio/cxgb/sge.c 			struct freelQ *fl = &sge->freelQ[e->FreelistQid];
sge              1502 drivers/net/ethernet/chelsio/cxgb/sge.c 				sge_rx(sge, fl, e->BufferLength);
sge              1516 drivers/net/ethernet/chelsio/cxgb/sge.c 				refill_free_list(sge, fl);
sge              1518 drivers/net/ethernet/chelsio/cxgb/sge.c 			sge->stats.pure_rsps++;
sge              1535 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->cmdQ[1].processed += cmdq_processed[1];
sge              1542 drivers/net/ethernet/chelsio/cxgb/sge.c 	const struct respQ *Q = &adapter->sge->respQ;
sge              1558 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1559 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct respQ *q = &sge->respQ;
sge              1561 drivers/net/ethernet/chelsio/cxgb/sge.c 	const struct freelQ *fl = &sge->freelQ[e->FreelistQid];
sge              1587 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.pure_rsps++;
sge              1591 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->cmdQ[1].processed += cmdq_processed[1];
sge              1608 drivers/net/ethernet/chelsio/cxgb/sge.c 		writel(adapter->sge->respQ.cidx,
sge              1617 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1628 drivers/net/ethernet/chelsio/cxgb/sge.c 				writel(sge->respQ.cidx, adapter->regs + A_SG_SLEEPING);
sge              1641 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.unhandled_irqs++;
sge              1662 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1663 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct cmdQ *q = &sge->cmdQ[qid];
sge              1668 drivers/net/ethernet/chelsio/cxgb/sge.c 	reclaim_completed_tx(sge, q);
sge              1679 drivers/net/ethernet/chelsio/cxgb/sge.c 			set_bit(dev->if_port, &sge->stopped_tx_queues);
sge              1680 drivers/net/ethernet/chelsio/cxgb/sge.c 			sge->stats.cmdQ_full[2]++;
sge              1690 drivers/net/ethernet/chelsio/cxgb/sge.c 		set_bit(dev->if_port, &sge->stopped_tx_queues);
sge              1691 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->stats.cmdQ_full[2]++;
sge              1697 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->tx_sched && !qid && skb->dev) {
sge              1703 drivers/net/ethernet/chelsio/cxgb/sge.c 		skb = sched_skb(sge, skb, credits);
sge              1773 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = adapter->sge;
sge              1774 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge_port_stats *st = this_cpu_ptr(sge->port_stats[dev->if_port]);
sge              1839 drivers/net/ethernet/chelsio/cxgb/sge.c 		if ((unlikely(!adapter->sge->espibug_skb[dev->if_port]))) {
sge              1842 drivers/net/ethernet/chelsio/cxgb/sge.c 				adapter->sge->espibug_skb[dev->if_port] = skb;
sge              1887 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = from_timer(sge, t, tx_reclaim_timer);
sge              1890 drivers/net/ethernet/chelsio/cxgb/sge.c 		struct cmdQ *q = &sge->cmdQ[i];
sge              1895 drivers/net/ethernet/chelsio/cxgb/sge.c 		reclaim_completed_tx(sge, q);
sge              1897 drivers/net/ethernet/chelsio/cxgb/sge.c 			writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL);
sge              1901 drivers/net/ethernet/chelsio/cxgb/sge.c 	mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD);
sge              1907 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_set_coalesce_params(struct sge *sge, struct sge_params *p)
sge              1909 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->fixed_intrtimer = p->rx_coalesce_usecs *
sge              1910 drivers/net/ethernet/chelsio/cxgb/sge.c 		core_ticks_per_usec(sge->adapter);
sge              1911 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(sge->fixed_intrtimer, sge->adapter->regs + A_SG_INTRTIMER);
sge              1919 drivers/net/ethernet/chelsio/cxgb/sge.c int t1_sge_configure(struct sge *sge, struct sge_params *p)
sge              1921 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (alloc_rx_resources(sge, p))
sge              1923 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (alloc_tx_resources(sge, p)) {
sge              1924 drivers/net/ethernet/chelsio/cxgb/sge.c 		free_rx_resources(sge);
sge              1927 drivers/net/ethernet/chelsio/cxgb/sge.c 	configure_sge(sge, p);
sge              1935 drivers/net/ethernet/chelsio/cxgb/sge.c 	p->large_buf_capacity = jumbo_payload_capacity(sge);
sge              1942 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_stop(struct sge *sge)
sge              1945 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(0, sge->adapter->regs + A_SG_CONTROL);
sge              1946 drivers/net/ethernet/chelsio/cxgb/sge.c 	readl(sge->adapter->regs + A_SG_CONTROL); /* flush */
sge              1948 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (is_T2(sge->adapter))
sge              1949 drivers/net/ethernet/chelsio/cxgb/sge.c 		del_timer_sync(&sge->espibug_timer);
sge              1951 drivers/net/ethernet/chelsio/cxgb/sge.c 	del_timer_sync(&sge->tx_reclaim_timer);
sge              1952 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->tx_sched)
sge              1953 drivers/net/ethernet/chelsio/cxgb/sge.c 		tx_sched_stop(sge);
sge              1956 drivers/net/ethernet/chelsio/cxgb/sge.c 		kfree_skb(sge->espibug_skb[i]);
sge              1962 drivers/net/ethernet/chelsio/cxgb/sge.c void t1_sge_start(struct sge *sge)
sge              1964 drivers/net/ethernet/chelsio/cxgb/sge.c 	refill_free_list(sge, &sge->freelQ[0]);
sge              1965 drivers/net/ethernet/chelsio/cxgb/sge.c 	refill_free_list(sge, &sge->freelQ[1]);
sge              1967 drivers/net/ethernet/chelsio/cxgb/sge.c 	writel(sge->sge_control, sge->adapter->regs + A_SG_CONTROL);
sge              1968 drivers/net/ethernet/chelsio/cxgb/sge.c 	doorbell_pio(sge->adapter, F_FL0_ENABLE | F_FL1_ENABLE);
sge              1969 drivers/net/ethernet/chelsio/cxgb/sge.c 	readl(sge->adapter->regs + A_SG_CONTROL); /* flush */
sge              1971 drivers/net/ethernet/chelsio/cxgb/sge.c 	mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD);
sge              1973 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (is_T2(sge->adapter))
sge              1974 drivers/net/ethernet/chelsio/cxgb/sge.c 		mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
sge              1982 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = from_timer(sge, t, espibug_timer);
sge              1983 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge              1994 drivers/net/ethernet/chelsio/cxgb/sge.c 			struct sk_buff *skb = sge->espibug_skb[i];
sge              2020 drivers/net/ethernet/chelsio/cxgb/sge.c 	mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
sge              2025 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = from_timer(sge, t, espibug_timer);
sge              2026 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct adapter *adapter = sge->adapter;
sge              2029 drivers/net/ethernet/chelsio/cxgb/sge.c 	        struct sk_buff *skb = sge->espibug_skb[0];
sge              2052 drivers/net/ethernet/chelsio/cxgb/sge.c 	mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
sge              2058 drivers/net/ethernet/chelsio/cxgb/sge.c struct sge *t1_sge_create(struct adapter *adapter, struct sge_params *p)
sge              2060 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct sge *sge = kzalloc(sizeof(*sge), GFP_KERNEL);
sge              2063 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (!sge)
sge              2066 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->adapter = adapter;
sge              2067 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->netdev = adapter->port[0].dev;
sge              2068 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->rx_pkt_pad = t1_is_T1B(adapter) ? 0 : 2;
sge              2069 drivers/net/ethernet/chelsio/cxgb/sge.c 	sge->jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
sge              2072 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->port_stats[i] = alloc_percpu(struct sge_port_stats);
sge              2073 drivers/net/ethernet/chelsio/cxgb/sge.c 		if (!sge->port_stats[i])
sge              2077 drivers/net/ethernet/chelsio/cxgb/sge.c 	timer_setup(&sge->tx_reclaim_timer, sge_tx_reclaim_cb, 0);
sge              2079 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (is_T2(sge->adapter)) {
sge              2080 drivers/net/ethernet/chelsio/cxgb/sge.c 		timer_setup(&sge->espibug_timer,
sge              2085 drivers/net/ethernet/chelsio/cxgb/sge.c 			tx_sched_init(sge);
sge              2087 drivers/net/ethernet/chelsio/cxgb/sge.c 		sge->espibug_timeout = 1;
sge              2090 drivers/net/ethernet/chelsio/cxgb/sge.c 			sge->espibug_timeout = HZ/100;
sge              2096 drivers/net/ethernet/chelsio/cxgb/sge.c 	p->freelQ_size[!sge->jumbo_fl] = SGE_FREEL_SIZE;
sge              2097 drivers/net/ethernet/chelsio/cxgb/sge.c 	p->freelQ_size[sge->jumbo_fl] = SGE_JUMBO_FREEL_SIZE;
sge              2098 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->tx_sched) {
sge              2099 drivers/net/ethernet/chelsio/cxgb/sge.c 		if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204)
sge              2109 drivers/net/ethernet/chelsio/cxgb/sge.c 	return sge;
sge              2112 drivers/net/ethernet/chelsio/cxgb/sge.c 		free_percpu(sge->port_stats[i]);
sge              2115 drivers/net/ethernet/chelsio/cxgb/sge.c 	kfree(sge);
sge                71 drivers/net/ethernet/chelsio/cxgb/sge.h struct sge;
sge                73 drivers/net/ethernet/chelsio/cxgb/sge.h struct sge *t1_sge_create(struct adapter *, struct sge_params *);
sge                74 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_configure(struct sge *, struct sge_params *);
sge                75 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_set_coalesce_params(struct sge *, struct sge_params *);
sge                76 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_destroy(struct sge *);
sge                82 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_start(struct sge *);
sge                83 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_stop(struct sge *);
sge                84 drivers/net/ethernet/chelsio/cxgb/sge.h int t1_sge_intr_error_handler(struct sge *);
sge                85 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_enable(struct sge *);
sge                86 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_disable(struct sge *);
sge                87 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_intr_clear(struct sge *);
sge                88 drivers/net/ethernet/chelsio/cxgb/sge.h const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge);
sge                89 drivers/net/ethernet/chelsio/cxgb/sge.h void t1_sge_get_port_stats(const struct sge *sge, int port, struct sge_port_stats *);
sge                90 drivers/net/ethernet/chelsio/cxgb/sge.h unsigned int t1_sched_update_parms(struct sge *, unsigned int, unsigned int,
sge               219 drivers/net/ethernet/chelsio/cxgb/subr.c 		t1_sge_intr_error_handler(adapter->sge);
sge               761 drivers/net/ethernet/chelsio/cxgb/subr.c 	t1_sge_intr_enable(adapter->sge);
sge               793 drivers/net/ethernet/chelsio/cxgb/subr.c 	t1_sge_intr_disable(adapter->sge);
sge               819 drivers/net/ethernet/chelsio/cxgb/subr.c 	t1_sge_intr_clear(adapter->sge);
sge               853 drivers/net/ethernet/chelsio/cxgb/subr.c 		t1_sge_intr_error_handler(adapter->sge);
sge               982 drivers/net/ethernet/chelsio/cxgb/subr.c 	err = t1_sge_configure(adapter->sge, &adapter->params.sge);
sge              1022 drivers/net/ethernet/chelsio/cxgb/subr.c 	if (adapter->sge)
sge              1023 drivers/net/ethernet/chelsio/cxgb/subr.c 		t1_sge_destroy(adapter->sge);
sge              1059 drivers/net/ethernet/chelsio/cxgb/subr.c 	adapter->sge = t1_sge_create(adapter, &adapter->params.sge);
sge              1060 drivers/net/ethernet/chelsio/cxgb/subr.c 	if (!adapter->sge) {
sge               244 drivers/net/ethernet/chelsio/cxgb3/adapter.h 	struct sge sge;
sge               382 drivers/net/ethernet/chelsio/cxgb3/common.h 	struct sge_params sge;
sge               411 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 							  adap->sge.qs[qidx].
sge               414 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 					  &adap->sge.qs[qidx]);
sge               418 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 						 &adap->sge.qs[qidx]);
sge               438 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 				 &adapter->sge.qs[i]);
sge               448 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) {
sge               461 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts;
sge               597 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		struct sge_qset *qs = &adap->sge.qs[i];
sge               610 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		struct sge_qset *qs = &adap->sge.qs[i];
sge               635 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		if (adap->sge.qs[i].adap)
sge               636 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 			napi_disable(&adap->sge.qs[i].napi);
sge               643 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		if (adap->sge.qs[i].adap)
sge               644 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 			napi_enable(&adap->sge.qs[i].napi);
sge               667 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		pi->qs = &adap->sge.qs[pi->first_qset];
sge               672 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 				&adap->params.sge.qset[qset_idx], ntxq, dev,
sge              1173 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		struct sge_rspq *q = &adap->sge.qs[i].rspq;
sge              1279 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 						      adap->sge.qs[0].rspq.
sge              1660 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		tot += adapter->sge.qs[i].port_stats[idx];
sge              1957 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset];
sge              1989 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	q = &adapter->params.sge.qset[pi->first_qset];
sge              2013 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		qsp = &adapter->params.sge.qset[i];
sge              2014 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		qs = &adapter->sge.qs[i];
sge              2026 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	struct qset_params *q = adapter->params.sge.qset;
sge              2202 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		q = &adapter->params.sge.qset[t.qset_idx];
sge              2220 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 				&adapter->sge.qs[t.qset_idx];
sge              2235 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 					q = &adapter->params.sge.
sge              2278 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		q = &adapter->params.sge.qset[q1 + t.qset_idx];
sge              2623 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		struct sge_qset *qs = &adapter->sge.qs[qidx];
sge              2771 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		struct sge_qset *qs = &adapter->sge.qs[0];
sge               303 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_lock_irqsave(&adapter->sge.reg_lock, flags);
sge               306 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_unlock_irqrestore(&adapter->sge.reg_lock, flags);
sge               334 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_lock_irq(&adapter->sge.reg_lock);
sge               341 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_unlock_irq(&adapter->sge.reg_lock);
sge               345 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_lock_irq(&adapter->sge.reg_lock);
sge               347 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_unlock_irq(&adapter->sge.reg_lock);
sge               352 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_lock_irq(&adapter->sge.reg_lock);
sge               358 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		spin_unlock_irq(&adapter->sge.reg_lock);
sge               683 drivers/net/ethernet/chelsio/cxgb3/sge.c 			spin_lock_irq(&adapter->sge.reg_lock);
sge               685 drivers/net/ethernet/chelsio/cxgb3/sge.c 			spin_unlock_irq(&adapter->sge.reg_lock);
sge               696 drivers/net/ethernet/chelsio/cxgb3/sge.c 			spin_lock_irq(&adapter->sge.reg_lock);
sge               698 drivers/net/ethernet/chelsio/cxgb3/sge.c 			spin_unlock_irq(&adapter->sge.reg_lock);
sge               712 drivers/net/ethernet/chelsio/cxgb3/sge.c 		spin_lock_irq(&adapter->sge.reg_lock);
sge               714 drivers/net/ethernet/chelsio/cxgb3/sge.c 		spin_unlock_irq(&adapter->sge.reg_lock);
sge              1567 drivers/net/ethernet/chelsio/cxgb3/sge.c 	ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb);
sge              1835 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_qset *qs = &adap->sge.qs[queue_set(skb)];
sge              2094 drivers/net/ethernet/chelsio/cxgb3/sge.c 	skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
sge              2196 drivers/net/ethernet/chelsio/cxgb3/sge.c 	skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
sge              2647 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_rspq *q = &adap->sge.qs[0].rspq;
sge              2651 drivers/net/ethernet/chelsio/cxgb3/sge.c 	if (process_responses(adap, &adap->sge.qs[0], -1)) {
sge              2658 drivers/net/ethernet/chelsio/cxgb3/sge.c 	    process_responses(adap, &adap->sge.qs[1], -1)) {
sge              2659 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
sge              2697 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_rspq *q = &adap->sge.qs[0].rspq;
sge              2701 drivers/net/ethernet/chelsio/cxgb3/sge.c 	new_packets = rspq_check_napi(&adap->sge.qs[0]);
sge              2703 drivers/net/ethernet/chelsio/cxgb3/sge.c 		new_packets += rspq_check_napi(&adap->sge.qs[1]);
sge              2735 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
sge              2736 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
sge              2773 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
sge              2790 drivers/net/ethernet/chelsio/cxgb3/sge.c 		process_responses_gts(adap, &adap->sge.qs[1].rspq);
sge              2807 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_qset *qs0 = &adap->sge.qs[0];
sge              2825 drivers/net/ethernet/chelsio/cxgb3/sge.c 		napi_schedule(&adap->sge.qs[1].napi);
sge              2970 drivers/net/ethernet/chelsio/cxgb3/sge.c 	       &qs->rspq.lock : &adap->sge.qs[0].rspq.lock;
sge              3041 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_qset *q = &adapter->sge.qs[id];
sge              3124 drivers/net/ethernet/chelsio/cxgb3/sge.c 	spin_lock_irq(&adapter->sge.reg_lock);
sge              3168 drivers/net/ethernet/chelsio/cxgb3/sge.c 	spin_unlock_irq(&adapter->sge.reg_lock);
sge              3198 drivers/net/ethernet/chelsio/cxgb3/sge.c 	spin_unlock_irq(&adapter->sge.reg_lock);
sge              3215 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_qset *q = &adap->sge.qs[i];
sge              3238 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_qset *q = &adap->sge.qs[i];
sge              3258 drivers/net/ethernet/chelsio/cxgb3/sge.c 		t3_free_qset(adap, &adap->sge.qs[i]);
sge              3293 drivers/net/ethernet/chelsio/cxgb3/sge.c 			struct sge_qset *qs = &adap->sge.qs[i];
sge              3371 drivers/net/ethernet/chelsio/cxgb3/sge.c 	spin_lock_init(&adap->sge.reg_lock);
sge              3403 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 				  min(adapter->params.sge.max_pkt_size,
sge              3406 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 			     min(adapter->params.sge.max_pkt_size, 16384U));
sge              3423 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	t3_sge_init(adapter, &adapter->params.sge);
sge              3687 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	t3_sge_prep(adapter, &adapter->params.sge);
sge              2967 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	struct sge *s = &padap->sge;
sge               364 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	struct sge_params sge;
sge               847 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h #define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++)
sge               848 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h #define for_each_ofldtxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++)
sge               959 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	struct sge sge;
sge              1397 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0;
sge              2628 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	const struct sge_uld_txq_info *utxq_info = adap->sge.uld_txq_info[uld];
sge              2639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	const struct sge_uld_rxq_info *urxq_info = adap->sge.uld_rxq_info[uld];
sge              2668 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct sge *s = &adap->sge;
sge              2670 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4);
sge              3012 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	return DIV_ROUND_UP(adap->sge.ethqsets, 4) +
sge              3215 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	ssize_t size = (adap->sge.egr_sz + 3) / 4 +
sge              3216 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			adap->sge.egr_sz / 32 + 2; /* includes ,/\n/\0 */
sge              3223 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		       adap->sge.egr_sz, adap->sge.blocked_fl);
sge              3237 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	t = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sizeof(long), GFP_KERNEL);
sge              3241 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	err = bitmap_parse_user(ubuf, count, t, adap->sge.egr_sz);
sge              3247 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	bitmap_copy(adap->sge.blocked_fl, t, adap->sge.egr_sz);
sge               244 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset];
sge               245 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset];
sge               826 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	const struct sge *s = &pi->adapter->sge;
sge               844 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge *s = &adapter->sge;
sge               878 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
sge               893 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
sge               905 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
sge               921 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	return adap->sge.dbqtimer_tick;
sge               933 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	txq = &adap->sge.ethtxq[pi->first_qset];
sge               939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	return adap->sge.dbqtimer_val[txq->dbqtimerix];
sge               953 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge *s = &adap->sge;
sge               988 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge *s = &adap->sge;
sge              1100 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq;
sge              1104 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 		adap->sge.counter_val[rq->pktcnt_idx] : 0;
sge                70 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 				QUEUENO_V(adap->sge.fw_evtq.abs_id) |
sge               344 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 			iq = adapter->sge.ethrxq[pi->first_qset +
sge               528 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	t4_mk_filtdelwr(f->tid, fwr, adapter->sge.fw_evtq.abs_id);
sge               631 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 		      FW_FILTER_WR_RX_RPL_IQ_V(adapter->sge.fw_evtq.abs_id));
sge              1198 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 				 ((adapter->sge.fw_evtq.abs_id << 14) | atid),
sge              1208 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 				((adapter->sge.fw_evtq.abs_id << 14) | atid),
sge              1501 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 		       TCB_RSS_INFO_V(adapter->sge.fw_evtq.abs_id), 0, 1);
sge               236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset];
sge               578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start];
sge               741 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge               779 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge               841 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	rxq = &adapter->sge.ethrxq[pi->first_qset];
sge               896 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid)
sge               909 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for (i = 0; i < adap->sge.ingr_sz; i++) {
sge               910 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		struct sge_rspq *q = adap->sge.ingr_map[i];
sge               939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for (i = 0; i < adap->sge.ingr_sz; i++) {
sge               940 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		struct sge_rspq *q = adap->sge.ingr_map[i];
sge               957 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge               989 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge              1107 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int closest_timer(const struct sge *s, int time)
sge              1123 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int closest_thres(const struct sge *s, int thres)
sge              1160 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		new_idx = closest_thres(&adap->sge, cnt);
sge              1175 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	us = us == 0 ? 6 : closest_timer(&adap->sge, us);
sge              1547 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	chan = rxq_to_chan(&adap->sge, queue);
sge              1590 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	chan = rxq_to_chan(&adap->sge, queue);
sge              2056 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_ethrxq(&adap->sge, i)
sge              2057 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		disable_txq_db(&adap->sge.ethtxq[i].q);
sge              2060 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			adap->sge.uld_txq_info[CXGB4_TX_OFLD];
sge              2063 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			for_each_ofldtxq(&adap->sge, i) {
sge              2071 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		disable_txq_db(&adap->sge.ctrlq[i].q);
sge              2078 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_ethrxq(&adap->sge, i)
sge              2079 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		enable_txq_db(adap, &adap->sge.ethtxq[i].q);
sge              2082 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			adap->sge.uld_txq_info[CXGB4_TX_OFLD];
sge              2085 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			for_each_ofldtxq(&adap->sge, i) {
sge              2093 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		enable_txq_db(adap, &adap->sge.ctrlq[i].q);
sge              2160 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_ethrxq(&adap->sge, i)
sge              2161 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		sync_txq_pidx(adap, &adap->sge.ethtxq[i].q);
sge              2164 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			adap->sge.uld_txq_info[CXGB4_TX_OFLD];
sge              2166 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			for_each_ofldtxq(&adap->sge, i) {
sge              2174 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		sync_txq_pidx(adap, &adap->sge.ctrlq[i].q);
sge              3046 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset];
sge              3742 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	ret = t4_cfg_pfvf(adap, adap->mbox, adap->pf, 0, adap->sge.egr_sz, 64,
sge              4485 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		adap->sge.dbqtimer_tick = val[0];
sge              4487 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 					    ARRAY_SIZE(adap->sge.dbqtimer_val),
sge              4488 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 					    adap->sge.dbqtimer_val);
sge              4519 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.egr_start = val[0];
sge              4524 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.ingr_start = val[5];
sge              4551 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.egr_sz = val[0] - adap->sge.egr_start + 1;
sge              4552 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.ingr_sz = val[1] - adap->sge.ingr_start + 1;
sge              4554 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.egr_map = kcalloc(adap->sge.egr_sz,
sge              4555 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				    sizeof(*adap->sge.egr_map), GFP_KERNEL);
sge              4556 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adap->sge.egr_map) {
sge              4561 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.ingr_map = kcalloc(adap->sge.ingr_sz,
sge              4562 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				     sizeof(*adap->sge.ingr_map), GFP_KERNEL);
sge              4563 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adap->sge.ingr_map) {
sge              4571 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.starving_fl =	kcalloc(BITS_TO_LONGS(adap->sge.egr_sz),
sge              4573 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adap->sge.starving_fl) {
sge              4578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.txq_maperr = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz),
sge              4580 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adap->sge.txq_maperr) {
sge              4586 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adap->sge.blocked_fl = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz),
sge              4588 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adap->sge.blocked_fl) {
sge              4919 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adap->sge.egr_map);
sge              4920 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adap->sge.ingr_map);
sge              4921 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adap->sge.starving_fl);
sge              4922 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adap->sge.txq_maperr);
sge              4924 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adap->sge.blocked_fl);
sge              5079 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge              5213 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	while (n < adap->sge.ethqsets)
sge              5218 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				adap->sge.ethqsets--;
sge              5219 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				if (adap->sge.ethqsets <= n)
sge              5277 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge *s = &adap->sge;
sge              5445 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->sge.egr_map);
sge              5446 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->sge.ingr_map);
sge              5447 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->sge.starving_fl);
sge              5448 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->sge.txq_maperr);
sge              5450 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->sge.blocked_fl);
sge              5841 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	bitmap_zero(adapter->sge.blocked_fl, adapter->sge.egr_sz);
sge               173 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c 	struct sge *s = &adapter->sge;
sge               144 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge *s = &adap->sge;
sge               198 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               214 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		struct sge *s = &adap->sge;
sge               243 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               246 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		struct sge *s = &adap->sge;
sge               271 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge *s = &adap->sge;
sge               339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	adap->sge.uld_rxq_info[uld_type] = rxq_info;
sge               346 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               348 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	adap->sge.uld_rxq_info[uld_type] = NULL;
sge               357 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               391 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               407 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               441 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               450 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               482 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge *s = &adap->sge;
sge               508 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	txq_info = adap->sge.uld_txq_info[tx_uld_type];
sge               514 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		adap->sge.uld_txq_info[tx_uld_type] = NULL;
sge               526 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	txq_info = adap->sge.uld_txq_info[tx_uld_type];
sge               563 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	adap->sge.uld_txq_info[tx_uld_type] = txq_info;
sge               570 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
sge               572 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge_uld_txq_info *txq_info = adap->sge.uld_txq_info[tx_uld_type];
sge               583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge *s = &adap->sge;
sge               611 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	struct sge *s = &adap->sge;
sge               673 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->udb_density = 1 << adap->params.sge.eq_qpp;
sge               674 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->ucq_density = 1 << adap->params.sge.iq_qpp;
sge               675 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->sge_host_page_size = 1 << (adap->params.sge.hps + 10);
sge               684 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->sge_ingpadboundary = adap->sge.fl_align;
sge               685 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->sge_egrstatuspagesize = adap->sge.stat_len;
sge               686 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 	lld->sge_pktshift = adap->sge.pktshift;
sge               154 drivers/net/ethernet/chelsio/cxgb4/l2t.c 					TID_QID_V(adap->sge.fw_evtq.abs_id)));
sge               156 drivers/net/ethernet/chelsio/cxgb4/sched.c 	txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
sge               201 drivers/net/ethernet/chelsio/cxgb4/sched.c 	txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
sge               149 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adapter->sge;
sge               236 drivers/net/ethernet/chelsio/cxgb4/sge.c 	const struct sge *s = &adapter->sge;
sge               320 drivers/net/ethernet/chelsio/cxgb4/sge.c 	for (p = sgl->sge; nfrags >= 2; nfrags -= 2) {
sge               457 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adapter->sge;
sge               594 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge               603 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (test_bit(q->cntxt_id - adap->sge.egr_start, adap->sge.blocked_fl))
sge               681 drivers/net/ethernet/chelsio/cxgb4/sge.c 		set_bit(q->cntxt_id - adap->sge.egr_start,
sge               682 drivers/net/ethernet/chelsio/cxgb4/sge.c 			adap->sge.starving_fl);
sge               909 drivers/net/ethernet/chelsio/cxgb4/sge.c 	to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge;
sge               923 drivers/net/ethernet/chelsio/cxgb4/sge.c 		unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1;
sge               926 drivers/net/ethernet/chelsio/cxgb4/sge.c 			memcpy(sgl->sge, buf, part0);
sge              1420 drivers/net/ethernet/chelsio/cxgb4/sge.c 		q = &adap->sge.ptptxq;
sge              1422 drivers/net/ethernet/chelsio/cxgb4/sge.c 		q = &adap->sge.ethtxq[qidx + pi->first_qset];
sge              1743 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
sge              2151 drivers/net/ethernet/chelsio/cxgb4/sge.c 	ret = ctrl_xmit(&adap->sge.ctrlq[0], skb);
sge              2209 drivers/net/ethernet/chelsio/cxgb4/sge.c 	set_bit(q->q.cntxt_id - q->adap->sge.egr_start,
sge              2210 drivers/net/ethernet/chelsio/cxgb4/sge.c 		q->adap->sge.txq_maperr);
sge              2438 drivers/net/ethernet/chelsio/cxgb4/sge.c 		return ctrl_xmit(&adap->sge.ctrlq[idx], skb);
sge              2441 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq_info = adap->sge.uld_txq_info[tx_uld_type];
sge              2567 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD];
sge              2743 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adapter->sge;
sge              2892 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adapter->sge;
sge              2934 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &q->adap->sge;
sge              3141 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adapter->sge;
sge              3301 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_rspq *q = &adap->sge.intrq;
sge              3304 drivers/net/ethernet/chelsio/cxgb4/sge.c 	spin_lock(&adap->sge.intrq_lock);
sge              3314 drivers/net/ethernet/chelsio/cxgb4/sge.c 			qid -= adap->sge.ingr_start;
sge              3315 drivers/net/ethernet/chelsio/cxgb4/sge.c 			napi_schedule(&adap->sge.ingr_map[qid]->napi);
sge              3334 drivers/net/ethernet/chelsio/cxgb4/sge.c 	spin_unlock(&adap->sge.intrq_lock);
sge              3388 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct adapter *adap = from_timer(adap, t, sge.rx_timer);
sge              3389 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3423 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct adapter *adap = from_timer(adap, t, sge.tx_timer);
sge              3424 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3518 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3630 drivers/net/ethernet/chelsio/cxgb4/sge.c 	adap->sge.ingr_map[iq->cntxt_id - adap->sge.ingr_start] = iq;
sge              3637 drivers/net/ethernet/chelsio/cxgb4/sge.c 		adap->sge.egr_map[fl->cntxt_id - adap->sge.egr_start] = fl;
sge              3716 drivers/net/ethernet/chelsio/cxgb4/sge.c 	adap->sge.egr_map[id - adap->sge.egr_start] = q;
sge              3734 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3819 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3890 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3949 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3962 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              3965 drivers/net/ethernet/chelsio/cxgb4/sge.c 	adap->sge.ingr_map[rq->cntxt_id - adap->sge.ingr_start] = NULL;
sge              4015 drivers/net/ethernet/chelsio/cxgb4/sge.c 	for (i = 0; i < adap->sge.ethqsets; i++) {
sge              4016 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eq = &adap->sge.ethrxq[i];
sge              4026 drivers/net/ethernet/chelsio/cxgb4/sge.c 	for (i = 0; i < adap->sge.ethqsets; i++) {
sge              4027 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eq = &adap->sge.ethrxq[i];
sge              4032 drivers/net/ethernet/chelsio/cxgb4/sge.c 		etq = &adap->sge.ethtxq[i];
sge              4045 drivers/net/ethernet/chelsio/cxgb4/sge.c 	for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) {
sge              4046 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct sge_ctrl_txq *cq = &adap->sge.ctrlq[i];
sge              4057 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (adap->sge.fw_evtq.desc)
sge              4058 drivers/net/ethernet/chelsio/cxgb4/sge.c 		free_rspq_fl(adap, &adap->sge.fw_evtq, NULL);
sge              4060 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (adap->sge.intrq.desc)
sge              4061 drivers/net/ethernet/chelsio/cxgb4/sge.c 		free_rspq_fl(adap, &adap->sge.intrq, NULL);
sge              4064 drivers/net/ethernet/chelsio/cxgb4/sge.c 		etq = &adap->sge.ptptxq;
sge              4077 drivers/net/ethernet/chelsio/cxgb4/sge.c 	memset(adap->sge.egr_map, 0,
sge              4078 drivers/net/ethernet/chelsio/cxgb4/sge.c 	       adap->sge.egr_sz * sizeof(*adap->sge.egr_map));
sge              4083 drivers/net/ethernet/chelsio/cxgb4/sge.c 	adap->sge.ethtxq_rover = 0;
sge              4084 drivers/net/ethernet/chelsio/cxgb4/sge.c 	mod_timer(&adap->sge.rx_timer, jiffies + RX_QCHECK_PERIOD);
sge              4085 drivers/net/ethernet/chelsio/cxgb4/sge.c 	mod_timer(&adap->sge.tx_timer, jiffies + TX_QCHECK_PERIOD);
sge              4099 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              4112 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD];
sge              4116 drivers/net/ethernet/chelsio/cxgb4/sge.c 			for_each_ofldtxq(&adap->sge, i) {
sge              4126 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq_info = adap->sge.uld_txq_info[CXGB4_TX_CRYPTO];
sge              4130 drivers/net/ethernet/chelsio/cxgb4/sge.c 			for_each_ofldtxq(&adap->sge, i) {
sge              4155 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge              4252 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge *s = &adap->sge;
sge               200 drivers/net/ethernet/chelsio/cxgb4/smt.c 				    TID_QID_V(adapter->sge.fw_evtq.abs_id)));
sge                91 drivers/net/ethernet/chelsio/cxgb4/srq.c 				TID_QID_V(adap->sge.fw_evtq.abs_id)));
sge              9263 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	page_shift = adapter->params.sge.hps + 10;
sge              9269 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		     ? adapter->params.sge.eq_qpp
sge              9270 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		     : adapter->params.sge.iq_qpp);
sge              9372 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	struct sge_params *sge_params = &adapter->params.sge;
sge              1509 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h 	struct ulptx_sge_pair sge[0];
sge               349 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h #define for_each_ethrxq(sge, iter) \
sge               350 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h 	for (iter = 0; iter < (sge)->ethqsets; iter++)
sge               383 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h 	struct sge sge;
sge               384 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               422 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               455 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               478 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               536 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge *s = &adapter->sge;
sge               592 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               707 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset];
sge              1021 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c static int closest_timer(const struct sge *s, int us)
sge              1037 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c static int closest_thres(const struct sge *s, int thres)
sge              1062 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		? adapter->sge.timer_val[timer_idx]
sge              1099 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		pktcnt_idx = closest_thres(&adapter->sge, cnt);
sge              1118 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		     : closest_timer(&adapter->sge, us));
sge              1241 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		rxq = &adapter->sge.ethrxq[pi->first_qset];
sge              1600 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	const struct sge *s = &pi->adapter->sge;
sge              1624 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge              1657 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	const struct sge_rspq *rspq = &adapter->sge.ethrxq[pi->first_qset].rspq;
sge              1662 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		 ? adapter->sge.counter_val[rspq->pktcnt_idx]
sge              1679 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				   &adapter->sge.ethrxq[pi->first_qset].rspq,
sge              1804 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	const struct sge_eth_txq *txq = &adapter->sge.ethtxq[pi->first_qset];
sge              1805 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset];
sge              2056 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL);
sge              2074 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL];
sge              2075 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
sge              2076 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		int n = min(QPL, adapter->sge.ethqsets - QPL * r);
sge              2098 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		   adapter->sge.counter_val[rxq[qs].rspq.pktcnt_idx]);
sge              2111 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_rspq *evtq = &adapter->sge.fw_evtq;
sge              2118 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			   adapter->sge.counter_val[evtq->pktcnt_idx]);
sge              2122 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_rspq *intrq = &adapter->sge.intrq;
sge              2129 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			   adapter->sge.counter_val[intrq->pktcnt_idx]);
sge              2152 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 +
sge              2209 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL);
sge              2231 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL];
sge              2232 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
sge              2233 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		int n = min(QPL, adapter->sge.ethqsets - QPL * r);
sge              2261 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_rspq *evtq = &adapter->sge.fw_evtq;
sge              2269 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_rspq *intrq = &adapter->sge.intrq;
sge              2298 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 +
sge              2563 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	adapter->sge.max_ethqsets = ethqsets;
sge              2565 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	if (adapter->sge.max_ethqsets < adapter->params.nports) {
sge              2568 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			 adapter->sge.max_ethqsets, adapter->params.nports);
sge              2569 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		adapter->params.nports = adapter->sge.max_ethqsets;
sge              2580 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge_params *sge_params = &adapter->params.sge;
sge              2581 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge              2733 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge              2760 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		q10g = (adapter->sge.max_ethqsets - n1g) / n10g;
sge              2837 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	while (n < adapter->sge.ethqsets)
sge              2842 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				adapter->sge.ethqsets--;
sge              2843 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				if (adapter->sge.ethqsets <= n)
sge              2870 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	struct sge *s = &adapter->sge;
sge               264 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	const struct sge *s = &adapter->sge;
sge               325 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	for (p = sgl->sge; nfrags >= 2; nfrags -= 2) {
sge               457 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	const struct sge *s = &adapter->sge;
sge               607 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge               715 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		set_bit(fl->cntxt_id, adapter->sge.starving_fl);
sge               930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge;
sge               944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1;
sge               947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			memcpy(sgl->sge, buf, part0);
sge              1198 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
sge              1571 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              1623 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              1761 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              1939 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              1944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	spin_lock(&adapter->sge.intrq_lock);
sge              2021 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	spin_unlock(&adapter->sge.intrq_lock);
sge              2068 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct adapter *adapter = from_timer(adapter, t, sge.rx_timer);
sge              2069 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2127 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct adapter *adapter = from_timer(adapter, t, sge.tx_timer);
sge              2128 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2209 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2223 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	    rspq != &adapter->sge.intrq) {
sge              2225 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		intr_dest = adapter->sge.intrq.abs_id;
sge              2419 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2515 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2532 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2565 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2572 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) {
sge              2596 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	adapter->sge.ethtxq_rover = 0;
sge              2597 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	mod_timer(&adapter->sge.rx_timer, jiffies + RX_QCHECK_PERIOD);
sge              2598 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	mod_timer(&adapter->sge.tx_timer, jiffies + TX_QCHECK_PERIOD);
sge              2611 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge              2630 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge_params *sge_params = &adapter->params.sge;
sge              2633 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge *s = &adapter->sge;
sge               262 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h 	struct sge_params sge;		/* Scatter Gather Engine */
sge               703 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	sge_control = adapter->params.sge.sge_control;
sge               729 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		sge_control2 = adapter->params.sge.sge_control2;
sge               784 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	page_shift = adapter->params.sge.sge_vf_hps + 10;
sge               790 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		     ? adapter->params.sge.sge_vf_eq_qpp
sge               791 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 		     : adapter->params.sge.sge_vf_iq_qpp);
sge               850 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	struct sge_params *sge_params = &adapter->params.sge;
sge               797 drivers/net/ethernet/emulex/benet/be_cmds.c 	struct be_sge *sge;
sge               808 drivers/net/ethernet/emulex/benet/be_cmds.c 		sge = nonembedded_sgl(wrb);
sge               809 drivers/net/ethernet/emulex/benet/be_cmds.c 		sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma));
sge               810 drivers/net/ethernet/emulex/benet/be_cmds.c 		sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF);
sge               811 drivers/net/ethernet/emulex/benet/be_cmds.c 		sge->len = cpu_to_le32(mem->size);
sge               138 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		struct dpaa2_sg_entry *sge = &sgt[i];
sge               145 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		sg_addr = dpaa2_sg_get_addr(sge);
sge               150 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		sg_length = dpaa2_sg_get_len(sge);
sge               170 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 			sg_offset = dpaa2_sg_get_offset(sge);
sge               191 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		if (dpaa2_sg_is_final(sge))
sge                55 drivers/net/ethernet/huawei/hinic/hinic_common.c void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len)
sge                57 drivers/net/ethernet/huawei/hinic/hinic_common.c 	sge->hi_addr = upper_32_bits(addr);
sge                58 drivers/net/ethernet/huawei/hinic/hinic_common.c 	sge->lo_addr = lower_32_bits(addr);
sge                59 drivers/net/ethernet/huawei/hinic/hinic_common.c 	sge->len  = len;
sge                68 drivers/net/ethernet/huawei/hinic/hinic_common.c dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge)
sge                70 drivers/net/ethernet/huawei/hinic/hinic_common.c 	return (dma_addr_t)((((u64)sge->hi_addr) << 32) | sge->lo_addr);
sge                25 drivers/net/ethernet/huawei/hinic/hinic_common.h void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len);
sge                27 drivers/net/ethernet/huawei/hinic/hinic_common.h dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge);
sge               179 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c 	hinic_set_sge(&sge_resp->sge, buf_out->dma_addr, buf_out->size);
sge               236 drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c 	hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size);
sge               603 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		sq_wqe->buf_descs[i].sge = sges[i];
sge               781 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		sges[i] = sq_wqe->buf_descs[i].sge;
sge               921 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		      u16 cons_idx, struct hinic_sge *sge)
sge               926 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	sge->hi_addr = be32_to_cpu(rq_wqe->buf_desc.hi_addr);
sge               927 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	sge->lo_addr = be32_to_cpu(rq_wqe->buf_desc.lo_addr);
sge               928 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	sge->len = HINIC_RQ_CQE_SGE_GET(len, LEN);
sge               939 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 			  struct hinic_rq_wqe *rq_wqe, struct hinic_sge *sge)
sge               955 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	hinic_set_sge(&cqe_sect->sge, cqe_dma, sizeof(*cqe));
sge               957 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	buf_desc->hi_addr = sge->hi_addr;
sge               958 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	buf_desc->lo_addr = sge->lo_addr;
sge               217 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 		      u16 cons_idx, struct hinic_sge *sge);
sge               220 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 			  struct hinic_rq_wqe *wqe, struct hinic_sge *sge);
sge               334 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h 	struct hinic_sge        sge;
sge               353 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h 	struct hinic_sge        sge;
sge               403 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h 	struct hinic_sge sge;
sge               430 drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h 	struct hinic_sge        sge;
sge               201 drivers/net/ethernet/huawei/hinic/hinic_rx.c 	struct hinic_sge sge;
sge               220 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		hinic_set_sge(&sge, dma_addr, skb->len);
sge               229 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		hinic_rq_prepare_wqe(rxq->rq, prod_idx, rq_wqe, &sge);
sge               252 drivers/net/ethernet/huawei/hinic/hinic_rx.c 	struct hinic_sge sge;
sge               259 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		hinic_rq_get_sge(rq, &hw_wqe->rq_wqe, ci, &sge);
sge               263 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		rx_free_skb(rxq, rq->saved_skb[ci], hinic_sge_to_dma(&sge));
sge               282 drivers/net/ethernet/huawei/hinic/hinic_rx.c 	struct hinic_sge sge;
sge               291 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge);
sge               293 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		rx_unmap_skb(rxq, hinic_sge_to_dma(&sge));
sge               336 drivers/net/ethernet/huawei/hinic/hinic_rx.c 	struct hinic_sge sge;
sge               358 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge);
sge               360 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		rx_unmap_skb(rxq, hinic_sge_to_dma(&sge));
sge               366 drivers/net/ethernet/huawei/hinic/hinic_rx.c 		pkt_len = sge.len;
sge               859 drivers/net/ethernet/ibm/ibmvnic.h 	struct ibmvnic_sge_desc sge;
sge               678 drivers/nvme/host/pci.c static void nvme_pci_sgl_set_data(struct nvme_sgl_desc *sge,
sge               681 drivers/nvme/host/pci.c 	sge->addr = cpu_to_le64(sg_dma_address(sg));
sge               682 drivers/nvme/host/pci.c 	sge->length = cpu_to_le32(sg_dma_len(sg));
sge               683 drivers/nvme/host/pci.c 	sge->type = NVME_SGL_FMT_DATA_DESC << 4;
sge               686 drivers/nvme/host/pci.c static void nvme_pci_sgl_set_seg(struct nvme_sgl_desc *sge,
sge               689 drivers/nvme/host/pci.c 	sge->addr = cpu_to_le64(dma_addr);
sge               691 drivers/nvme/host/pci.c 		sge->length = cpu_to_le32(entries * sizeof(*sge));
sge               692 drivers/nvme/host/pci.c 		sge->type = NVME_SGL_FMT_LAST_SEG_DESC << 4;
sge               694 drivers/nvme/host/pci.c 		sge->length = cpu_to_le32(PAGE_SIZE);
sge               695 drivers/nvme/host/pci.c 		sge->type = NVME_SGL_FMT_SEG_DESC << 4;
sge                59 drivers/nvme/host/rdma.c 	struct ib_sge		sge[1 + NVME_RDMA_MAX_INLINE_SEGMENTS];
sge              1187 drivers/nvme/host/rdma.c 	struct ib_sge *sge = &req->sge[1];
sge              1191 drivers/nvme/host/rdma.c 	for (i = 0; i < count; i++, sgl++, sge++) {
sge              1192 drivers/nvme/host/rdma.c 		sge->addr = sg_dma_address(sgl);
sge              1193 drivers/nvme/host/rdma.c 		sge->length = sg_dma_len(sgl);
sge              1194 drivers/nvme/host/rdma.c 		sge->lkey = queue->device->pd->local_dma_lkey;
sge              1195 drivers/nvme/host/rdma.c 		len += sge->length;
sge              1343 drivers/nvme/host/rdma.c 		struct nvme_rdma_qe *qe, struct ib_sge *sge, u32 num_sge,
sge              1349 drivers/nvme/host/rdma.c 	sge->addr   = qe->dma;
sge              1350 drivers/nvme/host/rdma.c 	sge->length = sizeof(struct nvme_command),
sge              1351 drivers/nvme/host/rdma.c 	sge->lkey   = queue->device->pd->local_dma_lkey;
sge              1355 drivers/nvme/host/rdma.c 	wr.sg_list    = sge;
sge              1421 drivers/nvme/host/rdma.c 	struct ib_sge sge;
sge              1437 drivers/nvme/host/rdma.c 	ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL);
sge              1782 drivers/nvme/host/rdma.c 	err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge,
sge                35 drivers/nvme/target/rdma.c 	struct ib_sge		sge[NVMET_RDMA_MAX_INLINE_SGE + 1];
sge               218 drivers/nvme/target/rdma.c 	struct ib_sge *sge;
sge               225 drivers/nvme/target/rdma.c 	sge = &c->sge[1];
sge               227 drivers/nvme/target/rdma.c 	for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
sge               228 drivers/nvme/target/rdma.c 		if (sge->length)
sge               229 drivers/nvme/target/rdma.c 			ib_dma_unmap_page(ndev->device, sge->addr,
sge               230 drivers/nvme/target/rdma.c 					sge->length, DMA_FROM_DEVICE);
sge               240 drivers/nvme/target/rdma.c 	struct ib_sge *sge;
sge               250 drivers/nvme/target/rdma.c 	sge = &c->sge[1];
sge               253 drivers/nvme/target/rdma.c 	for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
sge               258 drivers/nvme/target/rdma.c 		sge->addr = ib_dma_map_page(ndev->device,
sge               260 drivers/nvme/target/rdma.c 		if (ib_dma_mapping_error(ndev->device, sge->addr))
sge               262 drivers/nvme/target/rdma.c 		sge->length = min_t(int, len, PAGE_SIZE);
sge               263 drivers/nvme/target/rdma.c 		sge->lkey = ndev->pd->local_dma_lkey;
sge               264 drivers/nvme/target/rdma.c 		len -= sge->length;
sge               269 drivers/nvme/target/rdma.c 	for (; i >= 0; i--, sg--, sge--) {
sge               270 drivers/nvme/target/rdma.c 		if (sge->length)
sge               271 drivers/nvme/target/rdma.c 			ib_dma_unmap_page(ndev->device, sge->addr,
sge               272 drivers/nvme/target/rdma.c 					sge->length, DMA_FROM_DEVICE);
sge               287 drivers/nvme/target/rdma.c 	c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd,
sge               289 drivers/nvme/target/rdma.c 	if (ib_dma_mapping_error(ndev->device, c->sge[0].addr))
sge               292 drivers/nvme/target/rdma.c 	c->sge[0].length = sizeof(*c->nvme_cmd);
sge               293 drivers/nvme/target/rdma.c 	c->sge[0].lkey = ndev->pd->local_dma_lkey;
sge               301 drivers/nvme/target/rdma.c 	c->wr.sg_list = c->sge;
sge               307 drivers/nvme/target/rdma.c 	ib_dma_unmap_single(ndev->device, c->sge[0].addr,
sge               321 drivers/nvme/target/rdma.c 	ib_dma_unmap_single(ndev->device, c->sge[0].addr,
sge               461 drivers/nvme/target/rdma.c 		cmd->sge[0].addr, cmd->sge[0].length,
sge               761 drivers/nvme/target/rdma.c 		cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length,
sge              4057 drivers/scsi/aacraid/aachba.c 		rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32));
sge              4058 drivers/scsi/aacraid/aachba.c 		rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff));
sge              4060 drivers/scsi/aacraid/aachba.c 		rio2->sge[i].length = cur_size;
sge              4061 drivers/scsi/aacraid/aachba.c 		rio2->sge[i].flags = 0;
sge              4078 drivers/scsi/aacraid/aachba.c 		u32 temp = le32_to_cpu(rio2->sge[i-1].length) -
sge              4080 drivers/scsi/aacraid/aachba.c 		rio2->sge[i-1].length = cpu_to_le32(temp);
sge              4093 drivers/scsi/aacraid/aachba.c 				if (rio2->sge[j].length % (i*PAGE_SIZE)) {
sge              4097 drivers/scsi/aacraid/aachba.c 				nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE));
sge              4122 drivers/scsi/aacraid/aachba.c 	struct sge_ieee1212 *sge;
sge              4129 drivers/scsi/aacraid/aachba.c 	sge = kmalloc_array(nseg_new, sizeof(struct sge_ieee1212), GFP_ATOMIC);
sge              4130 drivers/scsi/aacraid/aachba.c 	if (sge == NULL)
sge              4134 drivers/scsi/aacraid/aachba.c 		for (j = 0; j < rio2->sge[i].length / (pages * PAGE_SIZE); ++j) {
sge              4135 drivers/scsi/aacraid/aachba.c 			addr_low = rio2->sge[i].addrLow + j * pages * PAGE_SIZE;
sge              4136 drivers/scsi/aacraid/aachba.c 			sge[pos].addrLow = addr_low;
sge              4137 drivers/scsi/aacraid/aachba.c 			sge[pos].addrHigh = rio2->sge[i].addrHigh;
sge              4138 drivers/scsi/aacraid/aachba.c 			if (addr_low < rio2->sge[i].addrLow)
sge              4139 drivers/scsi/aacraid/aachba.c 				sge[pos].addrHigh++;
sge              4140 drivers/scsi/aacraid/aachba.c 			sge[pos].length = pages * PAGE_SIZE;
sge              4141 drivers/scsi/aacraid/aachba.c 			sge[pos].flags = 0;
sge              4145 drivers/scsi/aacraid/aachba.c 	sge[pos] = rio2->sge[nseg-1];
sge              4146 drivers/scsi/aacraid/aachba.c 	memcpy(&rio2->sge[1], &sge[1], (nseg_new-1)*sizeof(struct sge_ieee1212));
sge              4148 drivers/scsi/aacraid/aachba.c 	kfree(sge);
sge              4165 drivers/scsi/aacraid/aachba.c 	struct aac_hba_sgl *sge;
sge              4174 drivers/scsi/aacraid/aachba.c 		sge = &hbacmd->sge[2];
sge              4176 drivers/scsi/aacraid/aachba.c 		sge = &hbacmd->sge[0];
sge              4183 drivers/scsi/aacraid/aachba.c 		sge->addr_hi = cpu_to_le32((u32)(addr>>32));
sge              4184 drivers/scsi/aacraid/aachba.c 		sge->addr_lo = cpu_to_le32((u32)(addr & 0xffffffff));
sge              4186 drivers/scsi/aacraid/aachba.c 		sge->len = cur_size;
sge              4187 drivers/scsi/aacraid/aachba.c 		sge->flags = 0;
sge              4189 drivers/scsi/aacraid/aachba.c 		sge++;
sge              4192 drivers/scsi/aacraid/aachba.c 	sge--;
sge              4197 drivers/scsi/aacraid/aachba.c 		temp = le32_to_cpu(sge->len) - byte_count
sge              4199 drivers/scsi/aacraid/aachba.c 		sge->len = cpu_to_le32(temp);
sge              4205 drivers/scsi/aacraid/aachba.c 		sge->flags = cpu_to_le32(0x40000000);
sge              4208 drivers/scsi/aacraid/aachba.c 		hbacmd->sge[0].flags = cpu_to_le32(0x80000000);
sge              4210 drivers/scsi/aacraid/aachba.c 		hbacmd->sge[0].addr_hi = (u32)cpu_to_le32(sg_address >> 32);
sge              4211 drivers/scsi/aacraid/aachba.c 		hbacmd->sge[0].addr_lo =
sge               225 drivers/scsi/aacraid/aacraid.h 	struct aac_hba_sgl sge[HBA_MAX_SG_SEPARATE+2]; /* SG list space */
sge              1928 drivers/scsi/aacraid/aacraid.h 	struct sge_ieee1212	sge[1];
sge               682 drivers/scsi/aacraid/commctrl.c 			hbacmd->sge[i].addr_hi = cpu_to_le32((u32)(addr>>32));
sge               683 drivers/scsi/aacraid/commctrl.c 			hbacmd->sge[i].addr_lo = cpu_to_le32(
sge               685 drivers/scsi/aacraid/commctrl.c 			hbacmd->sge[i].len = cpu_to_le32(sg_count[i]);
sge               686 drivers/scsi/aacraid/commctrl.c 			hbacmd->sge[i].flags = 0;
sge               691 drivers/scsi/aacraid/commctrl.c 			hbacmd->sge[usg32->count-1].flags =
sge               184 drivers/scsi/aacraid/commsup.c 			offsetof(struct aac_hba_cmd_req, sge[2]);
sge              1264 drivers/scsi/be2iscsi/be_cmds.c 	struct be_sge *sge = nonembedded_sgl(wrb);
sge              1284 drivers/scsi/be2iscsi/be_cmds.c 	sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma));
sge              1285 drivers/scsi/be2iscsi/be_cmds.c 	sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF);
sge              1286 drivers/scsi/be2iscsi/be_cmds.c 	sge->len = cpu_to_le32(nonemb_cmd.size);
sge               127 drivers/scsi/be2iscsi/be_mgmt.c 	struct be_sge *sge;
sge               154 drivers/scsi/be2iscsi/be_mgmt.c 	sge = nonembedded_sgl(wrb);
sge               197 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
sge               198 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF);
sge               199 drivers/scsi/be2iscsi/be_mgmt.c 	sge->len = cpu_to_le32(nonemb_cmd->size);
sge               229 drivers/scsi/be2iscsi/be_mgmt.c 	struct be_sge *sge;
sge               241 drivers/scsi/be2iscsi/be_mgmt.c 	sge = nonembedded_sgl(wrb);
sge               243 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
sge               244 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd->dma));
sge               245 drivers/scsi/be2iscsi/be_mgmt.c 	sge->len = cpu_to_le32(nonemb_cmd->size);
sge              1004 drivers/scsi/be2iscsi/be_mgmt.c 	struct be_sge *sge;
sge              1027 drivers/scsi/be2iscsi/be_mgmt.c 	sge = nonembedded_sgl(wrb);
sge              1033 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
sge              1034 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF);
sge              1035 drivers/scsi/be2iscsi/be_mgmt.c 	sge->len = cpu_to_le32(nonemb_cmd->size);
sge              1506 drivers/scsi/be2iscsi/be_mgmt.c 	struct be_sge *sge;
sge              1543 drivers/scsi/be2iscsi/be_mgmt.c 	sge = nonembedded_sgl(wrb);
sge              1544 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma));
sge              1545 drivers/scsi/be2iscsi/be_mgmt.c 	sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd.dma));
sge              1546 drivers/scsi/be2iscsi/be_mgmt.c 	sge->len = cpu_to_le32(nonemb_cmd.size);
sge              2454 drivers/scsi/bfa/bfa_fcpim.c 	struct bfi_sge_s *sge, *sgpge;
sge              2482 drivers/scsi/bfa/bfa_fcpim.c 	sge = &m->sges[0];
sge              2491 drivers/scsi/bfa/bfa_fcpim.c 			sge->sga = *(union bfi_addr_u *) &addr;
sge              2493 drivers/scsi/bfa/bfa_fcpim.c 			sge->sg_len = pgdlen;
sge              2494 drivers/scsi/bfa/bfa_fcpim.c 			sge->flags = (ioim->nsges > BFI_SGE_INLINE) ?
sge              2496 drivers/scsi/bfa/bfa_fcpim.c 			bfa_sge_to_be(sge);
sge              2497 drivers/scsi/bfa/bfa_fcpim.c 			sge++;
sge              2538 drivers/scsi/bfa/bfa_fcpim.c 		sge->sga = ioim->sgpg->sgpg_pa;
sge              2540 drivers/scsi/bfa/bfa_fcpim.c 		sge->sga.a32.addr_lo = 0;
sge              2541 drivers/scsi/bfa/bfa_fcpim.c 		sge->sga.a32.addr_hi = 0;
sge              2543 drivers/scsi/bfa/bfa_fcpim.c 	sge->sg_len = pgdlen;
sge              2544 drivers/scsi/bfa/bfa_fcpim.c 	sge->flags = BFI_SGE_PGDLEN;
sge              2545 drivers/scsi/bfa/bfa_fcpim.c 	bfa_sge_to_be(sge);
sge              3184 drivers/scsi/bfa/bfad_bsg.c 	struct bfa_sge_s  *sge;
sge              3187 drivers/scsi/bfa/bfad_bsg.c 	sge = drv_fcxp->req_sge + sgeid;
sge              3188 drivers/scsi/bfa/bfad_bsg.c 	addr = (u64)(size_t) sge->sg_addr;
sge              3196 drivers/scsi/bfa/bfad_bsg.c 	struct bfa_sge_s	*sge;
sge              3198 drivers/scsi/bfa/bfad_bsg.c 	sge = drv_fcxp->req_sge + sgeid;
sge              3199 drivers/scsi/bfa/bfad_bsg.c 	return sge->sg_len;
sge              3206 drivers/scsi/bfa/bfad_bsg.c 	struct bfa_sge_s	*sge;
sge              3209 drivers/scsi/bfa/bfad_bsg.c 	sge = drv_fcxp->rsp_sge + sgeid;
sge              3210 drivers/scsi/bfa/bfad_bsg.c 	addr = (u64)(size_t) sge->sg_addr;
sge              3218 drivers/scsi/bfa/bfad_bsg.c 	struct bfa_sge_s	*sge;
sge              3220 drivers/scsi/bfa/bfad_bsg.c 	sge = drv_fcxp->rsp_sge + sgeid;
sge              3221 drivers/scsi/bfa/bfad_bsg.c 	return sge->sg_len;
sge                58 drivers/scsi/csiostor/csio_wr.c csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg)
sge                60 drivers/scsi/csiostor/csio_wr.c 	sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A +
sge                66 drivers/scsi/csiostor/csio_wr.c csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf)
sge                68 drivers/scsi/csiostor/csio_wr.c 	return sge->sge_fl_buf_size[buf->paddr & 0xF];
sge                75 drivers/scsi/csiostor/csio_wr.c 	return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ?  128 : 64;
sge               118 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge               126 drivers/scsi/csiostor/csio_wr.c 		buf->len = sge->sge_fl_buf_size[sreg];
sge              1049 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge              1075 drivers/scsi/csiostor/csio_wr.c 		bufsz = csio_wr_fl_bufsz(sge, buf);
sge              1096 drivers/scsi/csiostor/csio_wr.c 		flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align);
sge              1312 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge              1418 drivers/scsi/csiostor/csio_wr.c 	sge->csio_fl_align = fl_align;
sge              1435 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge              1437 drivers/scsi/csiostor/csio_wr.c 	csio_sge_thresh_reg = csio_closest_thresh(sge, csio_intr_coalesce_cnt);
sge              1444 drivers/scsi/csiostor/csio_wr.c 	csio_sge_timer_reg = csio_closest_timer(sge, csio_intr_coalesce_time);
sge              1457 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge              1463 drivers/scsi/csiostor/csio_wr.c 	sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A);
sge              1465 drivers/scsi/csiostor/csio_wr.c 	ingpad = INGPADBOUNDARY_G(sge->sge_control);
sge              1469 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 32; break;
sge              1471 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 64; break;
sge              1473 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 128; break;
sge              1475 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 256; break;
sge              1477 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 512; break;
sge              1479 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 1024; break;
sge              1481 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 2048; break;
sge              1483 drivers/scsi/csiostor/csio_wr.c 		sge->csio_fl_align = 4096; break;
sge              1487 drivers/scsi/csiostor/csio_wr.c 		csio_get_flbuf_size(hw, sge, i);
sge              1493 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[0] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1495 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[1] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1497 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[2] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1499 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[3] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1501 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[4] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1503 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[5] = (uint16_t)csio_core_ticks_to_us(hw,
sge              1507 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[0] = THRESHOLD_0_G(ingress_rx_threshold);
sge              1508 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[1] = THRESHOLD_1_G(ingress_rx_threshold);
sge              1509 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[2] = THRESHOLD_2_G(ingress_rx_threshold);
sge              1510 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[3] = THRESHOLD_3_G(ingress_rx_threshold);
sge              1526 drivers/scsi/csiostor/csio_wr.c 	struct csio_sge *sge = &wrm->sge;
sge              1535 drivers/scsi/csiostor/csio_wr.c 	sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A);
sge              1556 drivers/scsi/csiostor/csio_wr.c 	csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE2 + sge->csio_fl_align - 1)
sge              1557 drivers/scsi/csiostor/csio_wr.c 		      & ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE2_A);
sge              1558 drivers/scsi/csiostor/csio_wr.c 	csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE3 + sge->csio_fl_align - 1)
sge              1559 drivers/scsi/csiostor/csio_wr.c 		      & ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE3_A);
sge              1567 drivers/scsi/csiostor/csio_wr.c 		csio_get_flbuf_size(hw, sge, i);
sge              1570 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[0] = CSIO_SGE_TIMER_VAL_0;
sge              1571 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[1] = CSIO_SGE_TIMER_VAL_1;
sge              1572 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[2] = CSIO_SGE_TIMER_VAL_2;
sge              1573 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[3] = CSIO_SGE_TIMER_VAL_3;
sge              1574 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[4] = CSIO_SGE_TIMER_VAL_4;
sge              1575 drivers/scsi/csiostor/csio_wr.c 	sge->timer_val[5] = CSIO_SGE_TIMER_VAL_5;
sge              1577 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[0] = CSIO_SGE_INT_CNT_VAL_0;
sge              1578 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[1] = CSIO_SGE_INT_CNT_VAL_1;
sge              1579 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[2] = CSIO_SGE_INT_CNT_VAL_2;
sge              1580 drivers/scsi/csiostor/csio_wr.c 	sge->counter_val[3] = CSIO_SGE_INT_CNT_VAL_3;
sge              1582 drivers/scsi/csiostor/csio_wr.c 	csio_wr_reg32(hw, THRESHOLD_0_V(sge->counter_val[0]) |
sge              1583 drivers/scsi/csiostor/csio_wr.c 		      THRESHOLD_1_V(sge->counter_val[1]) |
sge              1584 drivers/scsi/csiostor/csio_wr.c 		      THRESHOLD_2_V(sge->counter_val[2]) |
sge              1585 drivers/scsi/csiostor/csio_wr.c 		      THRESHOLD_3_V(sge->counter_val[3]),
sge              1589 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE0_V(csio_us_to_core_ticks(hw, sge->timer_val[0])) |
sge              1590 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE1_V(csio_us_to_core_ticks(hw, sge->timer_val[1])),
sge              1594 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE2_V(csio_us_to_core_ticks(hw, sge->timer_val[2])) |
sge              1595 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE3_V(csio_us_to_core_ticks(hw, sge->timer_val[3])),
sge              1599 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE4_V(csio_us_to_core_ticks(hw, sge->timer_val[4])) |
sge              1600 drivers/scsi/csiostor/csio_wr.c 		   TIMERVALUE5_V(csio_us_to_core_ticks(hw, sge->timer_val[5])),
sge               449 drivers/scsi/csiostor/csio_wr.h 	struct csio_sge		sge;		/* SGE params */
sge               504 drivers/scsi/esas2r/atvda.h 	struct atto_vda_sge sge[1];
sge               557 drivers/scsi/esas2r/atvda.h 	struct atto_vda_sge sge[1];
sge              1072 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge[1];
sge              1095 drivers/scsi/esas2r/atvda.h 			struct atto_vda_sge sge[1];
sge              1098 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge[1];
sge              1120 drivers/scsi/esas2r/atvda.h 	struct atto_vda_sge sge[1];
sge              1133 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge[1];
sge              1147 drivers/scsi/esas2r/atvda.h 	struct atto_vda_sge sge[1];
sge              1174 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge[1];
sge              1191 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge;
sge              1210 drivers/scsi/esas2r/atvda.h 		struct atto_vda_sge sge[2];
sge               619 drivers/scsi/esas2r/esas2r.h 	} sge;
sge              1178 drivers/scsi/esas2r/esas2r.h 	sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq
sge              1185 drivers/scsi/esas2r/esas2r.h 		sgc->sge.a64.last =
sge              1186 drivers/scsi/esas2r/esas2r.h 			sgc->sge.a64.curr = first;
sge              1191 drivers/scsi/esas2r/esas2r.h 		sgc->sge.a64.last =
sge              1192 drivers/scsi/esas2r/esas2r.h 			sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0];
sge              1194 drivers/scsi/esas2r/esas2r.h 			(u8)offsetof(struct atto_vda_scsi_req, u.sge);
sge              1196 drivers/scsi/esas2r/esas2r.h 	sgc->sge.a64.chain = NULL;
sge               901 drivers/scsi/esas2r/esas2r_disc.c 	esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge);
sge               212 drivers/scsi/esas2r/esas2r_flash.c 		esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]);
sge               915 drivers/scsi/esas2r/esas2r_flash.c 		esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge);
sge              1312 drivers/scsi/esas2r/esas2r_flash.c 		vrq->data.sge[0].length =
sge              1315 drivers/scsi/esas2r/esas2r_flash.c 		vrq->data.sge[0].address = cpu_to_le64(
sge              1229 drivers/scsi/esas2r/esas2r_init.c 				data.sge);
sge               222 drivers/scsi/esas2r/esas2r_io.c 		if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) {
sge               237 drivers/scsi/esas2r/esas2r_io.c 			sgelen = (u8)((u8 *)sgc->sge.a64.curr
sge               238 drivers/scsi/esas2r/esas2r_io.c 				      - (u8 *)sgc->sge.a64.last);
sge               244 drivers/scsi/esas2r/esas2r_io.c 			memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen);
sge               247 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.a64.curr =
sge               252 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.a64.limit =
sge               257 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.a64.last->length = cpu_to_le32(
sge               259 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.a64.last->address =
sge               268 drivers/scsi/esas2r/esas2r_io.c 			if (sgc->sge.a64.chain) {
sge               269 drivers/scsi/esas2r/esas2r_io.c 				sgc->sge.a64.chain->length |=
sge               271 drivers/scsi/esas2r/esas2r_io.c 						((u8 *)(sgc->sge.a64.
sge               280 drivers/scsi/esas2r/esas2r_io.c 							  sge.a64.last -
sge               299 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.a64.chain = sgc->sge.a64.last;
sge               306 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.a64.last = sgc->sge.a64.curr;
sge               309 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.a64.curr->length = cpu_to_le32(SGE_ADDR_64 | len);
sge               310 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.a64.curr->address = cpu_to_le32(addr);
sge               311 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.a64.curr++;
sge               328 drivers/scsi/esas2r/esas2r_io.c 	sgc->sge.a64.last->length |= cpu_to_le32(SGE_LAST);
sge               334 drivers/scsi/esas2r/esas2r_io.c 	if (sgc->sge.a64.chain) {
sge               335 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.a64.chain->length |= cpu_to_le32(
sge               336 drivers/scsi/esas2r/esas2r_io.c 			((u8 *)(sgc->sge.a64.curr) -
sge               346 drivers/scsi/esas2r/esas2r_io.c 			((u16)((u8 *)sgc->sge.a64.last - (u8 *)vrq)
sge               409 drivers/scsi/esas2r/esas2r_io.c 		if (sgc->sge.prd.sge_cnt == 0) {
sge               418 drivers/scsi/esas2r/esas2r_io.c 				sgc->sge.prd.curr->ctl_len = cpu_to_le32(
sge               420 drivers/scsi/esas2r/esas2r_io.c 				sgc->sge.prd.curr->address = cpu_to_le64(addr);
sge               432 drivers/scsi/esas2r/esas2r_io.c 			if (sgc->sge.prd.chain) {
sge               439 drivers/scsi/esas2r/esas2r_io.c 				sgc->sge.prd.chain->ctl_len |= cpu_to_le32(
sge               440 drivers/scsi/esas2r/esas2r_io.c 					sgc->sge.prd.sgl_max_cnt);
sge               467 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.chain = sgc->sge.prd.curr;
sge               469 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.chain->ctl_len = cpu_to_le32(PRD_CHAIN);
sge               470 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.chain->address =
sge               478 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.curr =
sge               482 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.sge_cnt = sgc->sge.prd.sgl_max_cnt - 1;
sge               485 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.sge_cnt--;
sge               487 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.curr->ctl_len = cpu_to_le32(PRD_DATA | len);
sge               488 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.curr->address = cpu_to_le64(addr);
sge               492 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.curr++;
sge               513 drivers/scsi/esas2r/esas2r_io.c 		if (sgc->sge.prd.chain) {
sge               514 drivers/scsi/esas2r/esas2r_io.c 			sgc->sge.prd.chain->ctl_len |=
sge               515 drivers/scsi/esas2r/esas2r_io.c 				cpu_to_le32(sgc->sge.prd.sgl_max_cnt
sge               516 drivers/scsi/esas2r/esas2r_io.c 					    - sgc->sge.prd.sge_cnt
sge               614 drivers/scsi/esas2r/esas2r_io.c 		(struct atto_physical_region_description *)sgc->sge.a64.curr;
sge               616 drivers/scsi/esas2r/esas2r_io.c 	sgc->sge.prd.sgl_max_cnt = sgl_page_size /
sge               623 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.sge_cnt = 0;
sge               624 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.chain = NULL;
sge               625 drivers/scsi/esas2r/esas2r_io.c 		sgc->sge.prd.curr = curr_iblk_chn;
sge               300 drivers/scsi/esas2r/esas2r_ioctl.c 	esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
sge               353 drivers/scsi/esas2r/esas2r_ioctl.c 	esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
sge               672 drivers/scsi/esas2r/esas2r_ioctl.c 	esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
sge               119 drivers/scsi/esas2r/esas2r_vda.c 		firstsg = rq->vrq->flash.data.file.sge;
sge               130 drivers/scsi/esas2r/esas2r_vda.c 		firstsg = rq->vrq->cli.sge;
sge               194 drivers/scsi/esas2r/esas2r_vda.c 		firstsg = rq->vrq->mgt.sge;
sge               363 drivers/scsi/esas2r/esas2r_vda.c 						   data.sge);
sge               394 drivers/scsi/esas2r/esas2r_vda.c 				struct atto_vda_mgmt_req, sge);
sge               396 drivers/scsi/esas2r/esas2r_vda.c 			vrq->sge[0].length = cpu_to_le32(SGE_LAST | length);
sge               397 drivers/scsi/esas2r/esas2r_vda.c 			vrq->sge[0].address = cpu_to_le64(
sge               432 drivers/scsi/esas2r/esas2r_vda.c 			(u8)offsetof(struct atto_vda_ae_req, sge);
sge               433 drivers/scsi/esas2r/esas2r_vda.c 		vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length);
sge               434 drivers/scsi/esas2r/esas2r_vda.c 		vrq->sge[0].address = cpu_to_le64(
sge               461 drivers/scsi/esas2r/esas2r_vda.c 	vrq->sg_list_offset = (u8)offsetof(struct atto_vda_cli_req, sge);
sge               478 drivers/scsi/esas2r/esas2r_vda.c 	vrq->sg_list_offset = (u8)offsetof(struct atto_vda_ioctl_req, sge);
sge               513 drivers/scsi/hisi_sas/hisi_sas.h 	struct hisi_sas_sge sge[HISI_SAS_SGE_PAGE_CNT];
sge               518 drivers/scsi/hisi_sas/hisi_sas.h 	struct hisi_sas_sge sge[HISI_SAS_SGE_DIF_PAGE_CNT];
sge               898 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
sge              1678 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
sge              1059 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		struct hisi_sas_sge *entry = &sge_page->sge[i];
sge              1085 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		struct hisi_sas_sge *entry = &sge_dif_page->sge[i];
sge              4939 drivers/scsi/lpfc/lpfc_bsg.c 				receive_length = nembed_sge->sge[0].length;
sge              4951 drivers/scsi/lpfc/lpfc_bsg.c 				nembed_sge->sge[0].pa_hi =
sge              4954 drivers/scsi/lpfc/lpfc_bsg.c 				nembed_sge->sge[0].pa_lo =
sge              1838 drivers/scsi/lpfc/lpfc_hbadisc.c 	struct lpfc_mbx_sge sge;
sge              1847 drivers/scsi/lpfc/lpfc_hbadisc.c 	lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
sge              1121 drivers/scsi/lpfc/lpfc_hw4.h 	struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES];
sge              1694 drivers/scsi/lpfc/lpfc_mbox.c 	nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr);
sge              1695 drivers/scsi/lpfc/lpfc_mbox.c 	nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr);
sge              1696 drivers/scsi/lpfc/lpfc_mbox.c 	nembed_sge->sge[sgentry].length = length;
sge              1709 drivers/scsi/lpfc/lpfc_mbox.c 		      struct lpfc_mbx_sge *sge)
sge              1715 drivers/scsi/lpfc/lpfc_mbox.c 	sge->pa_lo = nembed_sge->sge[sgentry].pa_lo;
sge              1716 drivers/scsi/lpfc/lpfc_mbox.c 	sge->pa_hi = nembed_sge->sge[sgentry].pa_hi;
sge              1717 drivers/scsi/lpfc/lpfc_mbox.c 	sge->length = nembed_sge->sge[sgentry].length;
sge              1731 drivers/scsi/lpfc/lpfc_mbox.c 	struct lpfc_mbx_sge sge;
sge              1752 drivers/scsi/lpfc/lpfc_mbox.c 		lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge);
sge              1753 drivers/scsi/lpfc/lpfc_mbox.c 		phyaddr = getPaddr(sge.pa_hi, sge.pa_lo);
sge              2030 drivers/scsi/lpfc/lpfc_mbox.c 	struct lpfc_mbx_sge sge;
sge              2056 drivers/scsi/lpfc/lpfc_mbox.c 	lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
sge              18520 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_mbx_sge sge;
sge              18551 drivers/scsi/lpfc/lpfc_sli.c 	lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
sge               188 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = ctlr_info_addr;
sge               189 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
sge               247 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = ldev_info_addr;
sge               248 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ldev_info.dma_size;
sge               335 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = pdev_info_addr;
sge               336 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
sge               406 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = devmap_addr;
sge               407 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
sge               442 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = event_addr;
sge               443 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->get_event.dma_size;
sge               470 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = cs->fwstat_addr;
sge               471 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
sge              1725 drivers/scsi/myrs.c 		hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl);
sge              1726 drivers/scsi/myrs.c 		hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl);
sge              1759 drivers/scsi/myrs.c 			hw_sgl = hw_sge->sge;
sge               670 drivers/scsi/myrs.h 	struct myrs_sge sge[2]; /* Bytes 0-31 */
sge                35 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		ctx_data_desc->sge[sge_index].sge_addr.lo = val;
sge                37 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		ctx_data_desc->sge[sge_index].sge_addr.hi = val;
sge                39 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		ctx_data_desc->sge[sge_index].sge_len = val;
sge                44 drivers/scsi/qedi/qedi_fw_api.c 		ctx_data_desc->sge[sge_index].sge_addr.lo = val;
sge                46 drivers/scsi/qedi/qedi_fw_api.c 		ctx_data_desc->sge[sge_index].sge_addr.hi = val;
sge                48 drivers/scsi/qedi/qedi_fw_api.c 		ctx_data_desc->sge[sge_index].sge_len = val;
sge               191 drivers/scsi/qedi/qedi_iscsi.c 	struct scsi_sge *sge;
sge               195 drivers/scsi/qedi/qedi_iscsi.c 					 sizeof(*sge),
sge                52 drivers/scsi/vmw_pvscsi.c 	struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT];
sge               333 drivers/scsi/vmw_pvscsi.c 	struct PVSCSISGElement *sge;
sge               337 drivers/scsi/vmw_pvscsi.c 	sge = &ctx->sgl->sge[0];
sge               339 drivers/scsi/vmw_pvscsi.c 		sge[i].addr   = sg_dma_address(sg);
sge               340 drivers/scsi/vmw_pvscsi.c 		sge[i].length = sg_dma_len(sg);
sge               341 drivers/scsi/vmw_pvscsi.c 		sge[i].flags  = 0;
sge               283 fs/cifs/smbdirect.c 			request->sge[i].addr,
sge               284 fs/cifs/smbdirect.c 			request->sge[i].length,
sge               504 fs/cifs/smbdirect.c 		response->sge.addr,
sge               505 fs/cifs/smbdirect.c 		response->sge.length,
sge               731 fs/cifs/smbdirect.c 	request->sge[0].addr = ib_dma_map_single(
sge               734 fs/cifs/smbdirect.c 	if (ib_dma_mapping_error(info->id->device, request->sge[0].addr)) {
sge               739 fs/cifs/smbdirect.c 	request->sge[0].length = sizeof(*packet);
sge               740 fs/cifs/smbdirect.c 	request->sge[0].lkey = info->pd->local_dma_lkey;
sge               743 fs/cifs/smbdirect.c 		info->id->device, request->sge[0].addr,
sge               744 fs/cifs/smbdirect.c 		request->sge[0].length, DMA_TO_DEVICE);
sge               750 fs/cifs/smbdirect.c 	send_wr.sg_list = request->sge;
sge               756 fs/cifs/smbdirect.c 		request->sge[0].addr,
sge               757 fs/cifs/smbdirect.c 		request->sge[0].length, request->sge[0].lkey);
sge               768 fs/cifs/smbdirect.c 	ib_dma_unmap_single(info->id->device, request->sge[0].addr,
sge               769 fs/cifs/smbdirect.c 		request->sge[0].length, DMA_TO_DEVICE);
sge               894 fs/cifs/smbdirect.c 	request->sge[0].addr = ib_dma_map_single(info->id->device,
sge               898 fs/cifs/smbdirect.c 	if (ib_dma_mapping_error(info->id->device, request->sge[0].addr)) {
sge               904 fs/cifs/smbdirect.c 	request->sge[0].length = header_length;
sge               905 fs/cifs/smbdirect.c 	request->sge[0].lkey = info->pd->local_dma_lkey;
sge               920 fs/cifs/smbdirect.c 			    request->sge[0].addr,
sge               921 fs/cifs/smbdirect.c 			    request->sge[0].length,
sge               937 fs/cifs/smbdirect.c 			i, request->sge[i].addr, request->sge[i].length);
sge               940 fs/cifs/smbdirect.c 			request->sge[i].addr,
sge               941 fs/cifs/smbdirect.c 			request->sge[i].length,
sge               949 fs/cifs/smbdirect.c 	send_wr.sg_list = request->sge;
sge               997 fs/cifs/smbdirect.c 		request->sge[i+1].addr =
sge              1001 fs/cifs/smbdirect.c 				info->id->device, request->sge[i+1].addr)) {
sge              1003 fs/cifs/smbdirect.c 			request->sge[i+1].addr = 0;
sge              1006 fs/cifs/smbdirect.c 		request->sge[i+1].length = sg->length;
sge              1007 fs/cifs/smbdirect.c 		request->sge[i+1].lkey = info->pd->local_dma_lkey;
sge              1017 fs/cifs/smbdirect.c 		if (request->sge[i].addr)
sge              1019 fs/cifs/smbdirect.c 					    request->sge[i].addr,
sge              1020 fs/cifs/smbdirect.c 					    request->sge[i].length,
sge              1095 fs/cifs/smbdirect.c 	response->sge.addr = ib_dma_map_single(
sge              1098 fs/cifs/smbdirect.c 	if (ib_dma_mapping_error(info->id->device, response->sge.addr))
sge              1101 fs/cifs/smbdirect.c 	response->sge.length = info->max_receive_size;
sge              1102 fs/cifs/smbdirect.c 	response->sge.lkey = info->pd->local_dma_lkey;
sge              1108 fs/cifs/smbdirect.c 	recv_wr.sg_list = &response->sge;
sge              1113 fs/cifs/smbdirect.c 		ib_dma_unmap_single(info->id->device, response->sge.addr,
sge              1114 fs/cifs/smbdirect.c 				    response->sge.length, DMA_FROM_DEVICE);
sge              1133 fs/cifs/smbdirect.c 		rc, response->sge.addr,
sge              1134 fs/cifs/smbdirect.c 		response->sge.length, response->sge.lkey);
sge              1275 fs/cifs/smbdirect.c 	ib_dma_unmap_single(info->id->device, response->sge.addr,
sge              1276 fs/cifs/smbdirect.c 		response->sge.length, DMA_FROM_DEVICE);
sge               242 fs/cifs/smbdirect.h 	struct ib_sge sge[SMBDIRECT_MAX_SGE];
sge               253 fs/cifs/smbdirect.h 	struct ib_sge sge;
sge                96 include/linux/qed/storage_common.h 	struct scsi_sge sge[4];
sge               232 include/linux/skmsg.h 	struct scatterlist *sge = sk_msg_elem(msg, msg->sg.start);
sge               238 include/linux/skmsg.h 		msg->data = sg_virt(sge);
sge               239 include/linux/skmsg.h 		msg->data_end = msg->data + sge->length;
sge               246 include/linux/skmsg.h 	struct scatterlist *sge;
sge               249 include/linux/skmsg.h 	sge = sk_msg_elem(msg, msg->sg.end);
sge               250 include/linux/skmsg.h 	sg_set_page(sge, page, len, offset);
sge               251 include/linux/skmsg.h 	sg_unmark_end(sge);
sge               566 include/rdma/rdma_vt.h int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge,
sge               570 include/rdma/rdma_vt.h 		struct ib_sge *sge, int acc);
sge               120 include/rdma/rdmavt_mr.h 	struct rvt_sge sge;   /* progress state for the current SGE */
sge               138 include/rdma/rdmavt_mr.h 		rvt_put_mr(ss->sge.mr);
sge               140 include/rdma/rdmavt_mr.h 			ss->sge = *ss->sg_list++;
sge               144 include/rdma/rdmavt_mr.h static inline u32 rvt_get_sge_length(struct rvt_sge *sge, u32 length)
sge               146 include/rdma/rdmavt_mr.h 	u32 len = sge->length;
sge               150 include/rdma/rdmavt_mr.h 	if (len > sge->sge_length)
sge               151 include/rdma/rdmavt_mr.h 		len = sge->sge_length;
sge               159 include/rdma/rdmavt_mr.h 	struct rvt_sge *sge = &ss->sge;
sge               161 include/rdma/rdmavt_mr.h 	sge->vaddr += length;
sge               162 include/rdma/rdmavt_mr.h 	sge->length -= length;
sge               163 include/rdma/rdmavt_mr.h 	sge->sge_length -= length;
sge               164 include/rdma/rdmavt_mr.h 	if (sge->sge_length == 0) {
sge               166 include/rdma/rdmavt_mr.h 			rvt_put_mr(sge->mr);
sge               168 include/rdma/rdmavt_mr.h 			*sge = *ss->sg_list++;
sge               169 include/rdma/rdmavt_mr.h 	} else if (sge->length == 0 && sge->mr->lkey) {
sge               170 include/rdma/rdmavt_mr.h 		if (++sge->n >= RVT_SEGSZ) {
sge               171 include/rdma/rdmavt_mr.h 			if (++sge->m >= sge->mr->mapsz)
sge               173 include/rdma/rdmavt_mr.h 			sge->n = 0;
sge               175 include/rdma/rdmavt_mr.h 		sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
sge               176 include/rdma/rdmavt_mr.h 		sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
sge               183 include/rdma/rdmavt_mr.h 	struct rvt_sge *sge = &ss->sge;
sge               186 include/rdma/rdmavt_mr.h 		u32 len = rvt_get_sge_length(sge, length);
sge               587 include/rdma/rdmavt_qp.h 		struct rvt_sge *sge = &wqe->sg_list[i];
sge               589 include/rdma/rdmavt_qp.h 		rvt_put_mr(sge->mr);
sge                24 include/rdma/rw.h 			struct ib_sge		sge;
sge                36 include/rdma/rw.h 			struct ib_sge		sge;
sge               129 include/uapi/rdma/rdma_user_rxe.h 		struct rxe_sge	sge[0];
sge               119 include/uapi/rdma/siw-abi.h 		struct siw_sge sge[SIW_MAX_SGE];
sge               135 include/uapi/rdma/siw-abi.h 	struct siw_sge sge[SIW_MAX_SGE];
sge               387 net/9p/trans_rdma.c 	struct ib_sge sge;
sge               397 net/9p/trans_rdma.c 	sge.addr = c->busa;
sge               398 net/9p/trans_rdma.c 	sge.length = client->msize;
sge               399 net/9p/trans_rdma.c 	sge.lkey = rdma->pd->local_dma_lkey;
sge               403 net/9p/trans_rdma.c 	wr.sg_list = &sge;
sge               416 net/9p/trans_rdma.c 	struct ib_sge sge;
sge               489 net/9p/trans_rdma.c 	sge.addr = c->busa;
sge               490 net/9p/trans_rdma.c 	sge.length = c->req->tc.size;
sge               491 net/9p/trans_rdma.c 	sge.lkey = rdma->pd->local_dma_lkey;
sge               497 net/9p/trans_rdma.c 	wr.sg_list = &sge;
sge              2224 net/core/filter.c 	struct scatterlist *sge;
sge              2281 net/core/filter.c 		sge = sk_msg_elem(msg, i);
sge              2282 net/core/filter.c 		from = sg_virt(sge);
sge              2283 net/core/filter.c 		len = sge->length;
sge              2288 net/core/filter.c 		sge->length = 0;
sge              2289 net/core/filter.c 		put_page(sg_page(sge));
sge              2348 net/core/filter.c 	struct scatterlist sge, nsge, nnsge, rsge = {0}, *psge;
sge              2427 net/core/filter.c 		sge = sk_msg_elem_cpy(msg, i);
sge              2430 net/core/filter.c 		sg_unmark_end(&sge);
sge              2440 net/core/filter.c 			msg->sg.data[i] = sge;
sge              2441 net/core/filter.c 			sge = nsge;
sge              2492 net/core/filter.c 	struct scatterlist tmp, sge;
sge              2495 net/core/filter.c 	sge = sk_msg_elem_cpy(msg, i);
sge              2500 net/core/filter.c 		msg->sg.data[i] = sge;
sge              2502 net/core/filter.c 		sge = tmp;
sge              2557 net/core/filter.c 		struct scatterlist *nsge, *sge = sk_msg_elem(msg, i);
sge              2559 net/core/filter.c 		int b = sge->length - pop - a;
sge              2563 net/core/filter.c 		if (pop < sge->length - a) {
sge              2565 net/core/filter.c 				sge->length = a;
sge              2568 net/core/filter.c 				get_page(sg_page(sge));
sge              2570 net/core/filter.c 					    sg_page(sge),
sge              2571 net/core/filter.c 					    b, sge->offset + pop + a);
sge              2582 net/core/filter.c 				sge->length = a;
sge              2583 net/core/filter.c 				orig = sg_page(sge);
sge              2584 net/core/filter.c 				from = sg_virt(sge);
sge              2588 net/core/filter.c 				sg_set_page(sge, page, a + b, 0);
sge              2592 net/core/filter.c 		} else if (pop >= sge->length - a) {
sge              2593 net/core/filter.c 			pop -= (sge->length - a);
sge              2594 net/core/filter.c 			sge->length = a;
sge              2616 net/core/filter.c 		struct scatterlist *sge = sk_msg_elem(msg, i);
sge              2618 net/core/filter.c 		if (pop < sge->length) {
sge              2619 net/core/filter.c 			sge->length -= pop;
sge              2620 net/core/filter.c 			sge->offset += pop;
sge              2623 net/core/filter.c 			pop -= sge->length;
sge                33 net/core/skmsg.c 		struct scatterlist *sge;
sge                47 net/core/skmsg.c 		sge = &msg->sg.data[i];
sge                50 net/core/skmsg.c 		    sg_page(sge) == pfrag->page &&
sge                51 net/core/skmsg.c 		    sge->offset + sge->length == orig_offset) {
sge                52 net/core/skmsg.c 			sge->length += use;
sge                59 net/core/skmsg.c 			sge = &msg->sg.data[msg->sg.end];
sge                60 net/core/skmsg.c 			sg_unmark_end(sge);
sge                61 net/core/skmsg.c 			sg_set_page(sge, pfrag->page, use, orig_offset);
sge                80 net/core/skmsg.c 	struct scatterlist *sge = sk_msg_elem(src, i);
sge                85 net/core/skmsg.c 		if (sge->length > off)
sge                87 net/core/skmsg.c 		off -= sge->length;
sge                91 net/core/skmsg.c 		sge = sk_msg_elem(src, i);
sge                95 net/core/skmsg.c 		sge_len = sge->length - off;
sge               103 net/core/skmsg.c 		    (sg_page(sge) == sg_page(sgd)) &&
sge               104 net/core/skmsg.c 		    (sg_virt(sge) + off == sg_virt(sgd) + sgd->length)) {
sge               108 net/core/skmsg.c 			sge_off = sge->offset + off;
sge               109 net/core/skmsg.c 			sk_msg_page_add(dst, sg_page(sge), sge_len, sge_off);
sge               120 net/core/skmsg.c 		sge = sk_msg_elem(src, i);
sge               132 net/core/skmsg.c 		struct scatterlist *sge = sk_msg_elem(msg, i);
sge               134 net/core/skmsg.c 		if (bytes < sge->length) {
sge               135 net/core/skmsg.c 			sge->length -= bytes;
sge               136 net/core/skmsg.c 			sge->offset += bytes;
sge               141 net/core/skmsg.c 		sk_mem_uncharge(sk, sge->length);
sge               142 net/core/skmsg.c 		bytes -= sge->length;
sge               143 net/core/skmsg.c 		sge->length = 0;
sge               144 net/core/skmsg.c 		sge->offset = 0;
sge               156 net/core/skmsg.c 		struct scatterlist *sge = &msg->sg.data[i];
sge               157 net/core/skmsg.c 		int uncharge = (bytes < sge->length) ? bytes : sge->length;
sge               169 net/core/skmsg.c 	struct scatterlist *sge = sk_msg_elem(msg, i);
sge               170 net/core/skmsg.c 	u32 len = sge->length;
sge               175 net/core/skmsg.c 		put_page(sg_page(sge));
sge               176 net/core/skmsg.c 	memset(sge, 0, sizeof(*sge));
sge               183 net/core/skmsg.c 	struct scatterlist *sge = sk_msg_elem(msg, i);
sge               187 net/core/skmsg.c 		msg->sg.size -= sge->length;
sge               191 net/core/skmsg.c 		sge = sk_msg_elem(msg, i);
sge               213 net/core/skmsg.c 	struct scatterlist *sge;
sge               217 net/core/skmsg.c 		sge = sk_msg_elem(msg, i);
sge               218 net/core/skmsg.c 		if (!sge->length)
sge               220 net/core/skmsg.c 		if (bytes < sge->length) {
sge               223 net/core/skmsg.c 			sge->length -= bytes;
sge               224 net/core/skmsg.c 			sge->offset += bytes;
sge               229 net/core/skmsg.c 		msg->sg.size -= sge->length;
sge               230 net/core/skmsg.c 		bytes -= sge->length;
sge               360 net/core/skmsg.c 	struct scatterlist *sge;
sge               365 net/core/skmsg.c 		sge = sk_msg_elem(msg, i);
sge               367 net/core/skmsg.c 		if (msg->sg.copybreak >= sge->length) {
sge               372 net/core/skmsg.c 			sge = sk_msg_elem(msg, i);
sge               375 net/core/skmsg.c 		buf_size = sge->length - msg->sg.copybreak;
sge               377 net/core/skmsg.c 		to = sg_virt(sge) + msg->sg.copybreak;
sge                57 net/ipv4/tcp_bpf.c 		struct scatterlist *sge;
sge                67 net/ipv4/tcp_bpf.c 			sge = sk_msg_elem(msg_rx, i);
sge                68 net/ipv4/tcp_bpf.c 			copy = sge->length;
sge                69 net/ipv4/tcp_bpf.c 			page = sg_page(sge);
sge                72 net/ipv4/tcp_bpf.c 			ret = copy_page_to_iter(page, sge->offset, copy, iter);
sge                80 net/ipv4/tcp_bpf.c 				sge->offset += copy;
sge                81 net/ipv4/tcp_bpf.c 				sge->length -= copy;
sge                85 net/ipv4/tcp_bpf.c 				if (!sge->length) {
sge               104 net/ipv4/tcp_bpf.c 		if (!sge->length && msg_rx->sg.start == msg_rx->sg.end) {
sge               168 net/ipv4/tcp_bpf.c 	struct scatterlist *sge;
sge               181 net/ipv4/tcp_bpf.c 		sge = sk_msg_elem(msg, i);
sge               182 net/ipv4/tcp_bpf.c 		size = (apply && apply_bytes < sge->length) ?
sge               183 net/ipv4/tcp_bpf.c 			apply_bytes : sge->length;
sge               193 net/ipv4/tcp_bpf.c 		if (sge->length)
sge               221 net/ipv4/tcp_bpf.c 	struct scatterlist *sge;
sge               229 net/ipv4/tcp_bpf.c 		sge = sk_msg_elem(msg, msg->sg.start);
sge               230 net/ipv4/tcp_bpf.c 		size = (apply && apply_bytes < sge->length) ?
sge               231 net/ipv4/tcp_bpf.c 			apply_bytes : sge->length;
sge               232 net/ipv4/tcp_bpf.c 		off  = sge->offset;
sge               233 net/ipv4/tcp_bpf.c 		page = sg_page(sge);
sge               251 net/ipv4/tcp_bpf.c 		sge->offset += ret;
sge               252 net/ipv4/tcp_bpf.c 		sge->length -= ret;
sge               260 net/ipv4/tcp_bpf.c 		if (!sge->length) {
sge               263 net/ipv4/tcp_bpf.c 			sg_init_table(sge, 1);
sge                53 net/rds/ib_recv.c 		struct ib_sge *sge;
sge                63 net/rds/ib_recv.c 		sge = &recv->r_sge[0];
sge                64 net/rds/ib_recv.c 		sge->addr = ic->i_recv_hdrs_dma + (i * sizeof(struct rds_header));
sge                65 net/rds/ib_recv.c 		sge->length = sizeof(struct rds_header);
sge                66 net/rds/ib_recv.c 		sge->lkey = ic->i_pd->local_dma_lkey;
sge                68 net/rds/ib_recv.c 		sge = &recv->r_sge[1];
sge                69 net/rds/ib_recv.c 		sge->addr = 0;
sge                70 net/rds/ib_recv.c 		sge->length = RDS_FRAG_SIZE;
sge                71 net/rds/ib_recv.c 		sge->lkey = ic->i_pd->local_dma_lkey;
sge               311 net/rds/ib_recv.c 	struct ib_sge *sge;
sge               345 net/rds/ib_recv.c 	sge = &recv->r_sge[0];
sge               346 net/rds/ib_recv.c 	sge->addr = ic->i_recv_hdrs_dma + (recv - ic->i_recvs) * sizeof(struct rds_header);
sge               347 net/rds/ib_recv.c 	sge->length = sizeof(struct rds_header);
sge               349 net/rds/ib_recv.c 	sge = &recv->r_sge[1];
sge               350 net/rds/ib_recv.c 	sge->addr = sg_dma_address(&recv->r_frag->f_sg);
sge               351 net/rds/ib_recv.c 	sge->length = sg_dma_len(&recv->r_frag->f_sg);
sge               576 net/rds/ib_recv.c 	struct ib_sge *sge = &ic->i_ack_sge;
sge               578 net/rds/ib_recv.c 	sge->addr = ic->i_ack_dma;
sge               579 net/rds/ib_recv.c 	sge->length = sizeof(struct rds_header);
sge               580 net/rds/ib_recv.c 	sge->lkey = ic->i_pd->local_dma_lkey;
sge               582 net/rds/ib_recv.c 	wr->sg_list = sge;
sge               195 net/rds/ib_send.c 		struct ib_sge *sge;
sge               203 net/rds/ib_send.c 		sge = &send->s_sge[0];
sge               204 net/rds/ib_send.c 		sge->addr = ic->i_send_hdrs_dma + (i * sizeof(struct rds_header));
sge               205 net/rds/ib_send.c 		sge->length = sizeof(struct rds_header);
sge               206 net/rds/ib_send.c 		sge->lkey = ic->i_pd->local_dma_lkey;
sge               323 net/smc/smc_tx.c 		struct ib_sge *sge =
sge               328 net/smc/smc_tx.c 			sge[srcchunk].addr = dma_addr + src_off;
sge               329 net/smc/smc_tx.c 			sge[srcchunk].length = src_len;
sge               568 net/sunrpc/xprtrdma/rpc_rdma.c 	struct ib_sge *sge;
sge               577 net/sunrpc/xprtrdma/rpc_rdma.c 	for (sge = &sc->sc_sges[2]; sc->sc_unmap_count;
sge               578 net/sunrpc/xprtrdma/rpc_rdma.c 	     ++sge, --sc->sc_unmap_count)
sge               579 net/sunrpc/xprtrdma/rpc_rdma.c 		ib_dma_unmap_page(sc->sc_device, sge->addr, sge->length,
sge               592 net/sunrpc/xprtrdma/rpc_rdma.c 	struct ib_sge *sge = sc->sc_sges;
sge               596 net/sunrpc/xprtrdma/rpc_rdma.c 	sge->addr = rdmab_addr(rb);
sge               597 net/sunrpc/xprtrdma/rpc_rdma.c 	sge->length = len;
sge               598 net/sunrpc/xprtrdma/rpc_rdma.c 	sge->lkey = rdmab_lkey(rb);
sge               600 net/sunrpc/xprtrdma/rpc_rdma.c 	ib_dma_sync_single_for_device(rdmab_device(rb), sge->addr, sge->length,
sge               621 net/sunrpc/xprtrdma/rpc_rdma.c 	struct ib_sge *sge = sc->sc_sges;
sge               631 net/sunrpc/xprtrdma/rpc_rdma.c 	sge[sge_no].addr = rdmab_addr(rb);
sge               632 net/sunrpc/xprtrdma/rpc_rdma.c 	sge[sge_no].length = xdr->head[0].iov_len;
sge               633 net/sunrpc/xprtrdma/rpc_rdma.c 	sge[sge_no].lkey = rdmab_lkey(rb);
sge               634 net/sunrpc/xprtrdma/rpc_rdma.c 	ib_dma_sync_single_for_device(rdmab_device(rb), sge[sge_no].addr,
sge               635 net/sunrpc/xprtrdma/rpc_rdma.c 				      sge[sge_no].length, DMA_TO_DEVICE);
sge               676 net/sunrpc/xprtrdma/rpc_rdma.c 			sge[sge_no].addr =
sge               680 net/sunrpc/xprtrdma/rpc_rdma.c 						 sge[sge_no].addr))
sge               682 net/sunrpc/xprtrdma/rpc_rdma.c 			sge[sge_no].length = len;
sge               683 net/sunrpc/xprtrdma/rpc_rdma.c 			sge[sge_no].lkey = rdmab_lkey(rb);
sge               704 net/sunrpc/xprtrdma/rpc_rdma.c 		sge[sge_no].addr =
sge               707 net/sunrpc/xprtrdma/rpc_rdma.c 		if (ib_dma_mapping_error(rdmab_device(rb), sge[sge_no].addr))
sge               709 net/sunrpc/xprtrdma/rpc_rdma.c 		sge[sge_no].length = len;
sge               710 net/sunrpc/xprtrdma/rpc_rdma.c 		sge[sge_no].lkey = rdmab_lkey(rb);
sge               731 net/sunrpc/xprtrdma/rpc_rdma.c 	trace_xprtrdma_dma_maperr(sge[sge_no].addr);
sge               430 net/tls/tls_sw.c 	struct scatterlist *sge;
sge               439 net/tls/tls_sw.c 	sge = sk_msg_elem(msg_en, msg_en->sg.curr);
sge               440 net/tls/tls_sw.c 	sge->offset -= prot->prepend_size;
sge               441 net/tls/tls_sw.c 	sge->length += prot->prepend_size;
sge               493 net/tls/tls_sw.c 	struct scatterlist *sge = sk_msg_elem(msg_en, start);
sge               507 net/tls/tls_sw.c 	sge->offset += prot->prepend_size;
sge               508 net/tls/tls_sw.c 	sge->length -= prot->prepend_size;
sge               528 net/tls/tls_sw.c 		sge->offset -= prot->prepend_size;
sge               529 net/tls/tls_sw.c 		sge->length += prot->prepend_size;
sge               551 net/tls/tls_sw.c 	struct scatterlist *sge, *osge, *nsge;
sge               570 net/tls/tls_sw.c 	sge = sk_msg_elem(msg_opl, i);
sge               571 net/tls/tls_sw.c 	while (apply && sge->length) {
sge               572 net/tls/tls_sw.c 		if (sge->length > apply) {
sge               573 net/tls/tls_sw.c 			u32 len = sge->length - apply;
sge               575 net/tls/tls_sw.c 			get_page(sg_page(sge));
sge               576 net/tls/tls_sw.c 			sg_set_page(&tmp, sg_page(sge), len,
sge               577 net/tls/tls_sw.c 				    sge->offset + apply);
sge               578 net/tls/tls_sw.c 			sge->length = apply;
sge               582 net/tls/tls_sw.c 			apply -= sge->length;
sge               583 net/tls/tls_sw.c 			bytes += sge->length;
sge               589 net/tls/tls_sw.c 		sge = sk_msg_elem(msg_opl, i);