Lines Matching refs:rds_iwdev

91 			     struct rds_iw_device **rds_iwdev,  argument
97 *rds_iwdev = NULL;
135 *rds_iwdev = iwdev;
146 static int rds_iw_add_cm_id(struct rds_iw_device *rds_iwdev, struct rdma_cm_id *cm_id) argument
156 spin_lock_irq(&rds_iwdev->spinlock);
157 list_add_tail(&i_cm_id->list, &rds_iwdev->cm_id_list);
158 spin_unlock_irq(&rds_iwdev->spinlock);
163 static void rds_iw_remove_cm_id(struct rds_iw_device *rds_iwdev, argument
168 spin_lock_irq(&rds_iwdev->spinlock);
169 list_for_each_entry(i_cm_id, &rds_iwdev->cm_id_list, list) {
176 spin_unlock_irq(&rds_iwdev->spinlock);
180 int rds_iw_update_cm_id(struct rds_iw_device *rds_iwdev, struct rdma_cm_id *cm_id) argument
192 rds_iw_remove_cm_id(rds_iwdev, cm_id);
194 return rds_iw_add_cm_id(rds_iwdev, cm_id);
197 void rds_iw_add_conn(struct rds_iw_device *rds_iwdev, struct rds_connection *conn) argument
207 spin_lock(&rds_iwdev->spinlock);
208 list_add_tail(&ic->iw_node, &rds_iwdev->conn_list);
209 spin_unlock(&rds_iwdev->spinlock);
212 ic->rds_iwdev = rds_iwdev;
215 void rds_iw_remove_conn(struct rds_iw_device *rds_iwdev, struct rds_connection *conn) argument
222 spin_lock_irq(&rds_iwdev->spinlock);
225 spin_unlock_irq(&rds_iwdev->spinlock);
231 rds_iw_remove_cm_id(ic->rds_iwdev, ic->i_cm_id);
232 ic->rds_iwdev = NULL;
260 static int rds_iw_map_scatterlist(struct rds_iw_device *rds_iwdev, argument
263 struct ib_device *dev = rds_iwdev->dev;
309 ib_dma_unmap_sg(rds_iwdev->dev, sg->list, sg->len, DMA_BIDIRECTIONAL);
315 struct rds_iw_mr_pool *rds_iw_create_mr_pool(struct rds_iw_device *rds_iwdev) argument
325 pool->device = rds_iwdev;
347 void rds_iw_get_mr_info(struct rds_iw_device *rds_iwdev, struct rds_info_rdma_connection *iinfo) argument
349 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
379 static struct rds_iw_mr *rds_iw_alloc_mr(struct rds_iw_device *rds_iwdev) argument
381 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
443 struct rds_iw_device *rds_iwdev = ibmr->device; local
447 ib_dma_sync_sg_for_cpu(rds_iwdev->dev, ibmr->mapping.m_sg.list,
451 ib_dma_sync_sg_for_device(rds_iwdev->dev, ibmr->mapping.m_sg.list,
560 struct rds_iw_device *rds_iwdev; local
562 list_for_each_entry(rds_iwdev, &rds_iw_devices, list) {
563 struct rds_iw_mr_pool *pool = rds_iwdev->mr_pool;
573 struct rds_iw_device *rds_iwdev; local
586 ret = rds_iw_get_device(&src, &dst, &rds_iwdev, &cm_id);
592 if (!rds_iwdev->mr_pool) {
597 ibmr = rds_iw_alloc_mr(rds_iwdev);
602 ibmr->device = rds_iwdev;
604 ret = rds_iw_map_reg(rds_iwdev->mr_pool, ibmr, sg, nents);
644 struct rds_iw_device *rds_iwdev = pool->device; local
648 mr = ib_alloc_mr(rds_iwdev->pd, IB_MR_TYPE_MEM_REG,
731 struct rds_iw_device *rds_iwdev = pool->device; local
738 ret = rds_iw_map_scatterlist(rds_iwdev, &mapping->m_sg);