Lines Matching refs:mapping

52 	struct rds_iw_mapping	mapping;  member
390 ibmr = list_entry(pool->clean_list.next, struct rds_iw_mr, mapping.m_list);
391 list_del_init(&ibmr->mapping.m_list);
439 spin_lock_init(&ibmr->mapping.m_lock);
440 INIT_LIST_HEAD(&ibmr->mapping.m_list);
441 ibmr->mapping.m_mr = ibmr;
466 ib_dma_sync_sg_for_cpu(rds_iwdev->dev, ibmr->mapping.m_sg.list,
467 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL);
470 ib_dma_sync_sg_for_device(rds_iwdev->dev, ibmr->mapping.m_sg.list,
471 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL);
520 list_for_each_entry_safe(ibmr, next, &kill_list, mapping.m_list) {
522 list_del(&ibmr->mapping.m_list);
556 rdsdebug("RDS/IW: free_mr nents %u\n", ibmr->mapping.m_sg.len);
695 static int rds_iw_rdma_build_fastreg(struct rds_iw_mapping *mapping) argument
697 struct rds_iw_mr *ibmr = mapping->m_mr;
708 mapping->m_rkey = ibmr->mr->rkey;
713 f_wr.wr.fast_reg.length = mapping->m_sg.bytes;
714 f_wr.wr.fast_reg.rkey = mapping->m_rkey;
716 f_wr.wr.fast_reg.page_list_len = mapping->m_sg.dma_len;
764 struct rds_iw_mapping *mapping = &ibmr->mapping; local
768 rds_iw_set_scatterlist(&mapping->m_sg, sg, sg_len);
770 dma_pages = rds_iw_map_scatterlist(rds_iwdev, &mapping->m_sg);
777 if (mapping->m_sg.dma_len > pool->max_message_size) {
782 for (i = 0; i < mapping->m_sg.dma_npages; ++i)
785 ret = rds_iw_rdma_build_fastreg(mapping);
806 if (!ibmr->mapping.m_sg.dma_len)
816 list_add_tail(&ibmr->mapping.m_list, &pool->dirty_list);
817 atomic_add(ibmr->mapping.m_sg.len, &pool->free_pinned);
828 struct rds_iw_mapping *mapping, *next; local
852 list_for_each_entry_safe(mapping, next, unmap_list, m_list) {
853 *unpinned += mapping->m_sg.len;
854 list_move(&mapping->m_list, &laundered);