/linux-4.1.27/net/rds/ |
D | page.c | 97 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, in rds_page_remainder_alloc() argument 113 sg_set_page(scat, page, PAGE_SIZE, 0); in rds_page_remainder_alloc() 132 sg_set_page(scat, rem->r_page, bytes, rem->r_offset); in rds_page_remainder_alloc() 133 get_page(sg_page(scat)); in rds_page_remainder_alloc() 176 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset, in rds_page_remainder_alloc() 177 ret ? 0 : scat->length); in rds_page_remainder_alloc()
|
D | ib_send.c | 534 struct scatterlist *scat; in rds_ib_xmit() local 554 scat = &rm->data.op_sg[sg]; in rds_ib_xmit() 555 ret = max_t(int, RDS_CONG_MAP_BYTES, scat->length); in rds_ib_xmit() 661 scat = &ic->i_data_op->op_sg[sg]; in rds_ib_xmit() 682 && scat != &rm->data.op_sg[rm->data.op_count]) { in rds_ib_xmit() 683 len = min(RDS_FRAG_SIZE, ib_sg_dma_len(dev, scat) - off); in rds_ib_xmit() 686 send->s_sge[1].addr = ib_sg_dma_address(dev, scat) + off; in rds_ib_xmit() 691 if (off == ib_sg_dma_len(dev, scat)) { in rds_ib_xmit() 692 scat++; in rds_ib_xmit() 730 && scat != &rm->data.op_sg[rm->data.op_count]); in rds_ib_xmit() [all …]
|
D | iw_send.c | 505 struct scatterlist *scat; in rds_iw_xmit() local 625 scat = &rm->data.op_sg[sg]; in rds_iw_xmit() 654 for (; i < work_alloc && scat != &rm->data.op_sg[rm->data.op_count]; i++) { in rds_iw_xmit() 659 len = min(RDS_FRAG_SIZE, ib_sg_dma_len(dev, scat) - off); in rds_iw_xmit() 661 ib_sg_dma_address(dev, scat) + off, len, in rds_iw_xmit() 691 if (off == ib_sg_dma_len(dev, scat)) { in rds_iw_xmit() 692 scat++; in rds_iw_xmit() 732 if (scat == &rm->data.op_sg[rm->data.op_count]) { in rds_iw_xmit() 796 struct scatterlist *scat; in rds_iw_xmit_rdma() local 856 scat = &op->op_sg[0]; in rds_iw_xmit_rdma() [all …]
|
D | ib_rdma.c | 371 struct scatterlist *scat = sg; in rds_ib_map_fmr() local 390 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); in rds_ib_map_fmr() 391 u64 dma_addr = ib_sg_dma_address(dev, &scat[i]); in rds_ib_map_fmr() 420 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); in rds_ib_map_fmr() 421 u64 dma_addr = ib_sg_dma_address(dev, &scat[i]); in rds_ib_map_fmr() 437 ibmr->sg = scat; in rds_ib_map_fmr()
|
D | rds.h | 689 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes,
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | srq.c | 314 struct mlx4_wqe_data_seg *scat; in mlx4_ib_post_srq_recv() local 346 scat = (struct mlx4_wqe_data_seg *) (next + 1); in mlx4_ib_post_srq_recv() 349 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv() 350 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv() 351 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv() 355 scat[i].byte_count = 0; in mlx4_ib_post_srq_recv() 356 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_post_srq_recv() 357 scat[i].addr = 0; in mlx4_ib_post_srq_recv()
|
D | qp.c | 2957 struct mlx4_wqe_data_seg *scat; in mlx4_ib_post_recv() local 2991 scat = get_recv_wqe(qp, ind); in mlx4_ib_post_recv() 2999 scat->byte_count = in mlx4_ib_post_recv() 3002 scat->lkey = cpu_to_be32(wr->sg_list->lkey); in mlx4_ib_post_recv() 3003 scat->addr = cpu_to_be64(qp->sqp_proxy_rcv[ind].map); in mlx4_ib_post_recv() 3004 scat++; in mlx4_ib_post_recv() 3009 __set_data_seg(scat + i, wr->sg_list + i); in mlx4_ib_post_recv() 3012 scat[i].byte_count = 0; in mlx4_ib_post_recv() 3013 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_post_recv() 3014 scat[i].addr = 0; in mlx4_ib_post_recv()
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
D | srq.c | 431 struct mlx5_wqe_data_seg *scat; in mlx5_ib_post_srq_recv() local 456 scat = (struct mlx5_wqe_data_seg *)(next + 1); in mlx5_ib_post_srq_recv() 459 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx5_ib_post_srq_recv() 460 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx5_ib_post_srq_recv() 461 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx5_ib_post_srq_recv() 465 scat[i].byte_count = 0; in mlx5_ib_post_srq_recv() 466 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); in mlx5_ib_post_srq_recv() 467 scat[i].addr = 0; in mlx5_ib_post_srq_recv()
|
D | qp.c | 2899 struct mlx5_wqe_data_seg *scat; in mlx5_ib_post_recv() local 2924 scat = get_recv_wqe(qp, ind); in mlx5_ib_post_recv() 2926 scat++; in mlx5_ib_post_recv() 2929 set_data_ptr_seg(scat + i, wr->sg_list + i); in mlx5_ib_post_recv() 2932 scat[i].byte_count = 0; in mlx5_ib_post_recv() 2933 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); in mlx5_ib_post_recv() 2934 scat[i].addr = 0; in mlx5_ib_post_recv() 2938 sig = (struct mlx5_rwqe_sig *)scat; in mlx5_ib_post_recv()
|
/linux-4.1.27/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1455 struct srp_request *req, struct scatterlist *scat, in srp_map_sg() argument 1475 for_each_sg(scat, sg, count, i) { in srp_map_sg() 1510 struct scatterlist *scat; in srp_map_data() local 1532 scat = scsi_sglist(scmnd); in srp_map_data() 1537 count = ib_dma_map_sg(ibdev, scat, nents, scmnd->sc_data_direction); in srp_map_data() 1553 buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat)); in srp_map_data() 1555 buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat)); in srp_map_data() 1571 srp_map_sg(&state, ch, req, scat, count); in srp_map_data()
|
/linux-4.1.27/Documentation/m68k/ |
D | kernel-options.txt | 572 Syntax: atascsi=<can_queue>[,<cmd_per_lun>[,<scat-gat>[,<host-id>[,<tagged>]]]] 606 <scat-gat>:
|