pages_per_mr       89 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pages_per_mr       91 drivers/infiniband/core/rw.c 	u32 nents = min(sg_cnt, pages_per_mr);
pages_per_mr      123 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pages_per_mr      127 drivers/infiniband/core/rw.c 	ctx->nr_ops = (sg_cnt + pages_per_mr - 1) / pages_per_mr;
pages_per_mr      136 drivers/infiniband/core/rw.c 		u32 nents = min(sg_cnt, pages_per_mr);
pages_per_mr      378 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pages_per_mr      383 drivers/infiniband/core/rw.c 	if (sg_cnt > pages_per_mr || prot_sg_cnt > pages_per_mr) {
pages_per_mr      385 drivers/infiniband/core/rw.c 		       sg_cnt, prot_sg_cnt, pages_per_mr);
pages_per_mr      492 drivers/infiniband/ulp/iser/iscsi_iser.h 	unsigned short               pages_per_mr;
pages_per_mr      254 drivers/infiniband/ulp/iser/iser_initiator.c 					   iser_conn->pages_per_mr))
pages_per_mr      683 drivers/infiniband/ulp/iser/iser_verbs.c 	iser_conn->pages_per_mr =
pages_per_mr      440 drivers/nvme/host/rdma.c 	int ret, pages_per_mr;
pages_per_mr      487 drivers/nvme/host/rdma.c 	pages_per_mr = nvme_rdma_get_max_fr_pages(ibdev) + 1;
pages_per_mr      491 drivers/nvme/host/rdma.c 			      pages_per_mr, 0);