pages_per_mr 89 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pages_per_mr 91 drivers/infiniband/core/rw.c u32 nents = min(sg_cnt, pages_per_mr); pages_per_mr 123 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pages_per_mr 127 drivers/infiniband/core/rw.c ctx->nr_ops = (sg_cnt + pages_per_mr - 1) / pages_per_mr; pages_per_mr 136 drivers/infiniband/core/rw.c u32 nents = min(sg_cnt, pages_per_mr); pages_per_mr 378 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pages_per_mr 383 drivers/infiniband/core/rw.c if (sg_cnt > pages_per_mr || prot_sg_cnt > pages_per_mr) { pages_per_mr 385 drivers/infiniband/core/rw.c sg_cnt, prot_sg_cnt, pages_per_mr); pages_per_mr 492 drivers/infiniband/ulp/iser/iscsi_iser.h unsigned short pages_per_mr; pages_per_mr 254 drivers/infiniband/ulp/iser/iser_initiator.c iser_conn->pages_per_mr)) pages_per_mr 683 drivers/infiniband/ulp/iser/iser_verbs.c iser_conn->pages_per_mr = pages_per_mr 440 drivers/nvme/host/rdma.c int ret, pages_per_mr; pages_per_mr 487 drivers/nvme/host/rdma.c pages_per_mr = nvme_rdma_get_max_fr_pages(ibdev) + 1; pages_per_mr 491 drivers/nvme/host/rdma.c pages_per_mr, 0);