scat              350 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_wqe_data_seg *scat = NULL;
scat              390 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		scat = (struct hns_roce_wqe_data_seg *)(ctrl + 1);
scat              393 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			set_data_seg(scat + i, wr->sg_list + i);
scat             3859 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_wqe_data_seg *scat;
scat             3894 drivers/infiniband/hw/mlx4/qp.c 		scat = get_recv_wqe(qp, ind);
scat             3902 drivers/infiniband/hw/mlx4/qp.c 			scat->byte_count =
scat             3905 drivers/infiniband/hw/mlx4/qp.c 			scat->lkey = cpu_to_be32(wr->sg_list->lkey);
scat             3906 drivers/infiniband/hw/mlx4/qp.c 			scat->addr = cpu_to_be64(qp->sqp_proxy_rcv[ind].map);
scat             3907 drivers/infiniband/hw/mlx4/qp.c 			scat++;
scat             3912 drivers/infiniband/hw/mlx4/qp.c 			__set_data_seg(scat + i, wr->sg_list + i);
scat             3915 drivers/infiniband/hw/mlx4/qp.c 			scat[i].byte_count = 0;
scat             3916 drivers/infiniband/hw/mlx4/qp.c 			scat[i].lkey       = cpu_to_be32(MLX4_INVALID_LKEY);
scat             3917 drivers/infiniband/hw/mlx4/qp.c 			scat[i].addr       = 0;
scat              305 drivers/infiniband/hw/mlx4/srq.c 	struct mlx4_wqe_data_seg *scat;
scat              337 drivers/infiniband/hw/mlx4/srq.c 		scat      = (struct mlx4_wqe_data_seg *) (next + 1);
scat              340 drivers/infiniband/hw/mlx4/srq.c 			scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length);
scat              341 drivers/infiniband/hw/mlx4/srq.c 			scat[i].lkey       = cpu_to_be32(wr->sg_list[i].lkey);
scat              342 drivers/infiniband/hw/mlx4/srq.c 			scat[i].addr       = cpu_to_be64(wr->sg_list[i].addr);
scat              346 drivers/infiniband/hw/mlx4/srq.c 			scat[i].byte_count = 0;
scat              347 drivers/infiniband/hw/mlx4/srq.c 			scat[i].lkey       = cpu_to_be32(MLX4_INVALID_LKEY);
scat              348 drivers/infiniband/hw/mlx4/srq.c 			scat[i].addr       = 0;
scat             5359 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_wqe_data_seg *scat;
scat             5395 drivers/infiniband/hw/mlx5/qp.c 		scat = mlx5_frag_buf_get_wqe(&qp->rq.fbc, ind);
scat             5397 drivers/infiniband/hw/mlx5/qp.c 			scat++;
scat             5400 drivers/infiniband/hw/mlx5/qp.c 			set_data_ptr_seg(scat + i, wr->sg_list + i);
scat             5403 drivers/infiniband/hw/mlx5/qp.c 			scat[i].byte_count = 0;
scat             5404 drivers/infiniband/hw/mlx5/qp.c 			scat[i].lkey       = cpu_to_be32(MLX5_INVALID_LKEY);
scat             5405 drivers/infiniband/hw/mlx5/qp.c 			scat[i].addr       = 0;
scat             5409 drivers/infiniband/hw/mlx5/qp.c 			sig = (struct mlx5_rwqe_sig *)scat;
scat              425 drivers/infiniband/hw/mlx5/srq.c 	struct mlx5_wqe_data_seg *scat;
scat              458 drivers/infiniband/hw/mlx5/srq.c 		scat      = (struct mlx5_wqe_data_seg *)(next + 1);
scat              461 drivers/infiniband/hw/mlx5/srq.c 			scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length);
scat              462 drivers/infiniband/hw/mlx5/srq.c 			scat[i].lkey       = cpu_to_be32(wr->sg_list[i].lkey);
scat              463 drivers/infiniband/hw/mlx5/srq.c 			scat[i].addr       = cpu_to_be64(wr->sg_list[i].addr);
scat              467 drivers/infiniband/hw/mlx5/srq.c 			scat[i].byte_count = 0;
scat              468 drivers/infiniband/hw/mlx5/srq.c 			scat[i].lkey       = cpu_to_be32(MLX5_INVALID_LKEY);
scat              469 drivers/infiniband/hw/mlx5/srq.c 			scat[i].addr       = 0;
scat             1644 drivers/infiniband/ulp/srp/ib_srp.c 			  struct srp_request *req, struct scatterlist *scat,
scat             1654 drivers/infiniband/ulp/srp/ib_srp.c 	for_each_sg(scat, sg, count, i) {
scat             1668 drivers/infiniband/ulp/srp/ib_srp.c 			 struct srp_request *req, struct scatterlist *scat,
scat             1675 drivers/infiniband/ulp/srp/ib_srp.c 	state->sg = scat;
scat             1696 drivers/infiniband/ulp/srp/ib_srp.c 			  struct srp_request *req, struct scatterlist *scat,
scat             1703 drivers/infiniband/ulp/srp/ib_srp.c 	for_each_sg(scat, sg, count, i) {
scat             1768 drivers/infiniband/ulp/srp/ib_srp.c 			      struct scatterlist *scat, int count)
scat             1804 drivers/infiniband/ulp/srp/ib_srp.c 	struct scatterlist *scat, *sg;
scat             1830 drivers/infiniband/ulp/srp/ib_srp.c 	scat  = scsi_sglist(scmnd);
scat             1836 drivers/infiniband/ulp/srp/ib_srp.c 	count = ib_dma_map_sg(ibdev, scat, nents, scmnd->sc_data_direction);
scat             1853 drivers/infiniband/ulp/srp/ib_srp.c 		for_each_sg(scat, sg, count, i) {
scat             1876 drivers/infiniband/ulp/srp/ib_srp.c 		buf->va  = cpu_to_be64(sg_dma_address(scat));
scat             1878 drivers/infiniband/ulp/srp/ib_srp.c 		buf->len = cpu_to_be32(sg_dma_len(scat));
scat             1896 drivers/infiniband/ulp/srp/ib_srp.c 		ret = srp_map_sg_fr(&state, ch, req, scat, count);
scat             1898 drivers/infiniband/ulp/srp/ib_srp.c 		ret = srp_map_sg_fmr(&state, ch, req, scat, count);
scat             1900 drivers/infiniband/ulp/srp/ib_srp.c 		ret = srp_map_sg_dma(&state, ch, req, scat, count);
scat             1909 drivers/infiniband/ulp/srp/ib_srp.c 			srp_check_mapping(&state, ch, req, scat, count);
scat              104 net/rds/ib_fmr.c 	struct scatterlist *scat = sg;
scat              122 net/rds/ib_fmr.c 		unsigned int dma_len = sg_dma_len(&scat[i]);
scat              123 net/rds/ib_fmr.c 		u64 dma_addr = sg_dma_address(&scat[i]);
scat              162 net/rds/ib_fmr.c 		unsigned int dma_len = sg_dma_len(&scat[i]);
scat              163 net/rds/ib_fmr.c 		u64 dma_addr = sg_dma_address(&scat[i]);
scat              181 net/rds/ib_fmr.c 	ibmr->sg = scat;
scat              492 net/rds/ib_send.c 	struct scatterlist *scat;
scat              512 net/rds/ib_send.c 		scat = &rm->data.op_sg[sg];
scat              513 net/rds/ib_send.c 		ret = max_t(int, RDS_CONG_MAP_BYTES, scat->length);
scat              621 net/rds/ib_send.c 	scat = &ic->i_data_op->op_sg[rm->data.op_dmasg];
scat              642 net/rds/ib_send.c 		    && scat != &rm->data.op_sg[rm->data.op_count]) {
scat              644 net/rds/ib_send.c 				  sg_dma_len(scat) - rm->data.op_dmaoff);
scat              647 net/rds/ib_send.c 			send->s_sge[1].addr = sg_dma_address(scat);
scat              653 net/rds/ib_send.c 			if (rm->data.op_dmaoff == sg_dma_len(scat)) {
scat              654 net/rds/ib_send.c 				scat++;
scat              695 net/rds/ib_send.c 		 && scat != &rm->data.op_sg[rm->data.op_count]);
scat              703 net/rds/ib_send.c 	if (scat == &rm->data.op_sg[rm->data.op_count]) {
scat              846 net/rds/ib_send.c 	struct scatterlist *scat;
scat              891 net/rds/ib_send.c 	scat = &op->op_sg[0];
scat              895 net/rds/ib_send.c 	for (i = 0; i < work_alloc && scat != &op->op_sg[op->op_count]; i++) {
scat              921 net/rds/ib_send.c 		     scat != &op->op_sg[op->op_count]; j++) {
scat              922 net/rds/ib_send.c 			len = sg_dma_len(scat);
scat              923 net/rds/ib_send.c 			send->s_sge[j].addr = sg_dma_address(scat);
scat              931 net/rds/ib_send.c 			scat++;
scat              945 net/rds/ib_send.c 	if (scat == &op->op_sg[op->op_count]) {
scat               68 net/rds/page.c int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes,
scat               84 net/rds/page.c 			sg_set_page(scat, page, PAGE_SIZE, 0);
scat              103 net/rds/page.c 			sg_set_page(scat, rem->r_page, bytes, rem->r_offset);
scat              104 net/rds/page.c 			get_page(sg_page(scat));
scat              147 net/rds/page.c 		 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset,
scat              148 net/rds/page.c 		 ret ? 0 : scat->length);
scat              884 net/rds/rds.h  int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes,