Lines Matching refs:rds_iwdev

92 			     struct rds_iw_device **rds_iwdev,  argument
98 *rds_iwdev = NULL;
136 *rds_iwdev = iwdev;
147 static int rds_iw_add_cm_id(struct rds_iw_device *rds_iwdev, struct rdma_cm_id *cm_id) argument
157 spin_lock_irq(&rds_iwdev->spinlock);
158 list_add_tail(&i_cm_id->list, &rds_iwdev->cm_id_list);
159 spin_unlock_irq(&rds_iwdev->spinlock);
164 static void rds_iw_remove_cm_id(struct rds_iw_device *rds_iwdev, argument
169 spin_lock_irq(&rds_iwdev->spinlock);
170 list_for_each_entry(i_cm_id, &rds_iwdev->cm_id_list, list) {
177 spin_unlock_irq(&rds_iwdev->spinlock);
181 int rds_iw_update_cm_id(struct rds_iw_device *rds_iwdev, struct rdma_cm_id *cm_id) argument
193 rds_iw_remove_cm_id(rds_iwdev, cm_id);
195 return rds_iw_add_cm_id(rds_iwdev, cm_id);
198 void rds_iw_add_conn(struct rds_iw_device *rds_iwdev, struct rds_connection *conn) argument
208 spin_lock(&rds_iwdev->spinlock);
209 list_add_tail(&ic->iw_node, &rds_iwdev->conn_list);
210 spin_unlock(&rds_iwdev->spinlock);
213 ic->rds_iwdev = rds_iwdev;
216 void rds_iw_remove_conn(struct rds_iw_device *rds_iwdev, struct rds_connection *conn) argument
223 spin_lock_irq(&rds_iwdev->spinlock);
226 spin_unlock_irq(&rds_iwdev->spinlock);
232 rds_iw_remove_cm_id(ic->rds_iwdev, ic->i_cm_id);
233 ic->rds_iwdev = NULL;
261 static u64 *rds_iw_map_scatterlist(struct rds_iw_device *rds_iwdev, argument
264 struct ib_device *dev = rds_iwdev->dev;
327 ib_dma_unmap_sg(rds_iwdev->dev, sg->list, sg->len, DMA_BIDIRECTIONAL);
334 struct rds_iw_mr_pool *rds_iw_create_mr_pool(struct rds_iw_device *rds_iwdev) argument
344 pool->device = rds_iwdev;
366 void rds_iw_get_mr_info(struct rds_iw_device *rds_iwdev, struct rds_info_rdma_connection *iinfo) argument
368 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
398 static struct rds_iw_mr *rds_iw_alloc_mr(struct rds_iw_device *rds_iwdev) argument
400 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
462 struct rds_iw_device *rds_iwdev = ibmr->device; local
466 ib_dma_sync_sg_for_cpu(rds_iwdev->dev, ibmr->mapping.m_sg.list,
470 ib_dma_sync_sg_for_device(rds_iwdev->dev, ibmr->mapping.m_sg.list,
581 struct rds_iw_device *rds_iwdev; local
583 list_for_each_entry(rds_iwdev, &rds_iw_devices, list) {
584 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
594 struct rds_iw_device *rds_iwdev; local
607 ret = rds_iw_get_device(&src, &dst, &rds_iwdev, &cm_id);
613 if (!rds_iwdev->mr_pool) {
618 ibmr = rds_iw_alloc_mr(rds_iwdev);
623 ibmr->device = rds_iwdev;
625 ret = rds_iw_map_fastreg(rds_iwdev->mr_pool, ibmr, sg, nents);
665 struct rds_iw_device *rds_iwdev = pool->device; local
670 mr = ib_alloc_fast_reg_mr(rds_iwdev->pd, pool->max_message_size);
681 page_list = ib_alloc_fast_reg_page_list(rds_iwdev->dev, pool->max_message_size);
763 struct rds_iw_device *rds_iwdev = pool->device; local
770 dma_pages = rds_iw_map_scatterlist(rds_iwdev, &mapping->m_sg);