Searched refs:scat (Results 1 - 12 of 12) sorted by relevance

/linux-4.4.14/net/rds/
H A Dpage.c80 * @scat: Scatter list for message
97 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, rds_page_remainder_alloc() argument
113 sg_set_page(scat, page, PAGE_SIZE, 0); rds_page_remainder_alloc()
132 sg_set_page(scat, rem->r_page, bytes, rem->r_offset); rds_page_remainder_alloc()
133 get_page(sg_page(scat)); rds_page_remainder_alloc()
176 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset, rds_page_remainder_alloc()
177 ret ? 0 : scat->length); rds_page_remainder_alloc()
H A Dib_send.c492 struct scatterlist *scat; rds_ib_xmit() local
512 scat = &rm->data.op_sg[sg]; rds_ib_xmit()
513 ret = max_t(int, RDS_CONG_MAP_BYTES, scat->length); rds_ib_xmit()
621 scat = &ic->i_data_op->op_sg[rm->data.op_dmasg]; rds_ib_xmit()
642 && scat != &rm->data.op_sg[rm->data.op_count]) { rds_ib_xmit()
644 ib_sg_dma_len(dev, scat) - rm->data.op_dmaoff); rds_ib_xmit()
647 send->s_sge[1].addr = ib_sg_dma_address(dev, scat); rds_ib_xmit()
653 if (rm->data.op_dmaoff == ib_sg_dma_len(dev, scat)) { rds_ib_xmit()
654 scat++; rds_ib_xmit()
693 && scat != &rm->data.op_sg[rm->data.op_count]); rds_ib_xmit()
701 if (scat == &rm->data.op_sg[rm->data.op_count]) { rds_ib_xmit()
850 struct scatterlist *scat; rds_ib_xmit_rdma() local
895 scat = &op->op_sg[0]; rds_ib_xmit_rdma()
899 for (i = 0; i < work_alloc && scat != &op->op_sg[op->op_count]; i++) { rds_ib_xmit_rdma()
923 scat != &op->op_sg[op->op_count]; j++) { rds_ib_xmit_rdma()
924 len = ib_sg_dma_len(ic->i_cm_id->device, scat); rds_ib_xmit_rdma()
926 ib_sg_dma_address(ic->i_cm_id->device, scat); rds_ib_xmit_rdma()
934 scat++; rds_ib_xmit_rdma()
948 if (scat == &op->op_sg[op->op_count]) { rds_ib_xmit_rdma()
H A Diw_send.c497 struct scatterlist *scat; rds_iw_xmit() local
619 scat = &rm->data.op_sg[rm->data.op_dmasg]; rds_iw_xmit()
648 for (; i < work_alloc && scat != &rm->data.op_sg[rm->data.op_count]; i++) { rds_iw_xmit()
654 ib_sg_dma_len(dev, scat) - rm->data.op_dmaoff); rds_iw_xmit()
656 ib_sg_dma_address(dev, scat) + rm->data.op_dmaoff, len, rds_iw_xmit()
686 if (rm->data.op_dmaoff == ib_sg_dma_len(dev, scat)) { rds_iw_xmit()
687 scat++; rds_iw_xmit()
728 if (scat == &rm->data.op_sg[rm->data.op_count]) { rds_iw_xmit()
794 struct scatterlist *scat; rds_iw_xmit_rdma() local
855 scat = &op->op_sg[0]; rds_iw_xmit_rdma()
860 for (i = 0; i < work_alloc && scat != &op->op_sg[op->op_count]; i++) { rds_iw_xmit_rdma()
898 scat != &op->op_sg[op->op_count]; j++) { rds_iw_xmit_rdma()
899 len = ib_sg_dma_len(ic->i_cm_id->device, scat); rds_iw_xmit_rdma()
904 send->s_sge[j].addr = ib_sg_dma_address(ic->i_cm_id->device, scat); rds_iw_xmit_rdma()
913 scat++; rds_iw_xmit_rdma()
934 if (scat == &op->op_sg[op->op_count]) rds_iw_xmit_rdma()
H A Dib_rdma.c421 struct scatterlist *scat = sg; rds_ib_map_fmr() local
440 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr()
441 u64 dma_addr = ib_sg_dma_address(dev, &scat[i]); rds_ib_map_fmr()
470 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr()
471 u64 dma_addr = ib_sg_dma_address(dev, &scat[i]); rds_ib_map_fmr()
487 ibmr->sg = scat; rds_ib_map_fmr()
H A Drds.h707 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes,
/linux-4.4.14/drivers/infiniband/hw/mlx4/
H A Dsrq.c319 struct mlx4_wqe_data_seg *scat; mlx4_ib_post_srq_recv() local
351 scat = (struct mlx4_wqe_data_seg *) (next + 1); mlx4_ib_post_srq_recv()
354 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); mlx4_ib_post_srq_recv()
355 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); mlx4_ib_post_srq_recv()
356 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); mlx4_ib_post_srq_recv()
360 scat[i].byte_count = 0; mlx4_ib_post_srq_recv()
361 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); mlx4_ib_post_srq_recv()
362 scat[i].addr = 0; mlx4_ib_post_srq_recv()
H A Dqp.c3071 struct mlx4_wqe_data_seg *scat; mlx4_ib_post_recv() local
3105 scat = get_recv_wqe(qp, ind); mlx4_ib_post_recv()
3113 scat->byte_count = mlx4_ib_post_recv()
3116 scat->lkey = cpu_to_be32(wr->sg_list->lkey); mlx4_ib_post_recv()
3117 scat->addr = cpu_to_be64(qp->sqp_proxy_rcv[ind].map); mlx4_ib_post_recv()
3118 scat++; mlx4_ib_post_recv()
3123 __set_data_seg(scat + i, wr->sg_list + i); mlx4_ib_post_recv()
3126 scat[i].byte_count = 0; mlx4_ib_post_recv()
3127 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); mlx4_ib_post_recv()
3128 scat[i].addr = 0; mlx4_ib_post_recv()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
H A Dsrq.c430 struct mlx5_wqe_data_seg *scat; mlx5_ib_post_srq_recv() local
455 scat = (struct mlx5_wqe_data_seg *)(next + 1); mlx5_ib_post_srq_recv()
458 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); mlx5_ib_post_srq_recv()
459 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); mlx5_ib_post_srq_recv()
460 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); mlx5_ib_post_srq_recv()
464 scat[i].byte_count = 0; mlx5_ib_post_srq_recv()
465 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); mlx5_ib_post_srq_recv()
466 scat[i].addr = 0; mlx5_ib_post_srq_recv()
H A Dqp.c2890 struct mlx5_wqe_data_seg *scat; mlx5_ib_post_recv() local
2915 scat = get_recv_wqe(qp, ind); mlx5_ib_post_recv()
2917 scat++; mlx5_ib_post_recv()
2920 set_data_ptr_seg(scat + i, wr->sg_list + i); mlx5_ib_post_recv()
2923 scat[i].byte_count = 0; mlx5_ib_post_recv()
2924 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); mlx5_ib_post_recv()
2925 scat[i].addr = 0; mlx5_ib_post_recv()
2929 sig = (struct mlx5_rwqe_sig *)scat; mlx5_ib_post_recv()
/linux-4.4.14/drivers/infiniband/ulp/srp/
H A Dib_srp.c1420 struct srp_request *req, struct scatterlist *scat, srp_map_sg_fmr()
1431 for_each_sg(scat, sg, count, i) { for_each_sg()
1447 struct srp_request *req, struct scatterlist *scat, srp_map_sg_fr()
1453 state->sg = scat; srp_map_sg_fr()
1473 struct srp_request *req, struct scatterlist *scat, srp_map_sg_dma()
1482 for_each_sg(scat, sg, count, i) { for_each_sg()
1551 struct scatterlist *scat; srp_map_data() local
1574 scat = scsi_sglist(scmnd); srp_map_data()
1579 count = ib_dma_map_sg(ibdev, scat, nents, scmnd->sc_data_direction); srp_map_data()
1595 buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat)); srp_map_data()
1597 buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat)); srp_map_data()
1614 srp_map_sg_fr(&state, ch, req, scat, count); srp_map_data()
1616 srp_map_sg_fmr(&state, ch, req, scat, count); srp_map_data()
1618 srp_map_sg_dma(&state, ch, req, scat, count); srp_map_data()
1419 srp_map_sg_fmr(struct srp_map_state *state, struct srp_rdma_ch *ch, struct srp_request *req, struct scatterlist *scat, int count) srp_map_sg_fmr() argument
1446 srp_map_sg_fr(struct srp_map_state *state, struct srp_rdma_ch *ch, struct srp_request *req, struct scatterlist *scat, int count) srp_map_sg_fr() argument
1472 srp_map_sg_dma(struct srp_map_state *state, struct srp_rdma_ch *ch, struct srp_request *req, struct scatterlist *scat, int count) srp_map_sg_dma() argument
/linux-4.4.14/drivers/scsi/
H A Dsg.c159 unsigned save_scat_len; /* original length of trunc. scat. element */
/linux-4.4.14/drivers/net/wireless/ath/ath6kl/
H A Dhtc_mbox.c478 "htc tx scat complete len %d entries %d\n", htc_async_tx_scat_complete()

Completed in 405 milliseconds