Lines Matching refs:mq

92 	q = &cq->mq;  in c2_cq_clean()
138 ce = c2_mq_consume(&cq->mq); in c2_poll_one()
150 c2_mq_free(&cq->mq); in c2_poll_one()
151 ce = c2_mq_consume(&cq->mq); in c2_poll_one()
196 c2_mq_free(&cq->mq); in c2_poll_one()
230 shared = cq->mq.peer; in c2_arm_cq()
250 ret = !c2_mq_empty(&cq->mq); in c2_arm_cq()
257 static void c2_free_cq_buf(struct c2_dev *c2dev, struct c2_mq *mq) in c2_free_cq_buf() argument
259 dma_free_coherent(&c2dev->pcidev->dev, mq->q_size * mq->msg_size, in c2_free_cq_buf()
260 mq->msg_pool.host, dma_unmap_addr(mq, mapping)); in c2_free_cq_buf()
263 static int c2_alloc_cq_buf(struct c2_dev *c2dev, struct c2_mq *mq, in c2_alloc_cq_buf() argument
272 &mq->host_dma, GFP_KERNEL); in c2_alloc_cq_buf()
276 c2_mq_rep_init(mq, in c2_alloc_cq_buf()
284 dma_unmap_addr_set(mq, mapping, mq->host_dma); in c2_alloc_cq_buf()
304 cq->mq.shared = c2_alloc_mqsp(c2dev, c2dev->kern_mqsp_pool, in c2_init_cq()
305 &cq->mq.shared_dma, GFP_KERNEL); in c2_init_cq()
306 if (!cq->mq.shared) in c2_init_cq()
310 err = c2_alloc_cq_buf(c2dev, &cq->mq, entries + 1, C2_CQ_MSG_SIZE); in c2_init_cq()
324 wr.msg_size = cpu_to_be32(cq->mq.msg_size); in c2_init_cq()
325 wr.depth = cpu_to_be32(cq->mq.q_size); in c2_init_cq()
326 wr.shared_ht = cpu_to_be64(cq->mq.shared_dma); in c2_init_cq()
327 wr.msg_pool = cpu_to_be64(cq->mq.host_dma); in c2_init_cq()
352 cq->mq.index = be32_to_cpu(reply->mq_index); in c2_init_cq()
355 cq->mq.peer = ioremap_nocache(peer_pa, PAGE_SIZE); in c2_init_cq()
356 if (!cq->mq.peer) { in c2_init_cq()
372 cq->cqn = cq->mq.index; in c2_init_cq()
382 c2_free_cq_buf(c2dev, &cq->mq); in c2_init_cq()
384 c2_free_mqsp(cq->mq.shared); in c2_init_cq()
400 c2dev->qptr_array[cq->mq.index] = NULL; in c2_free_cq()
436 c2_free_cq_buf(c2dev, &cq->mq); in c2_free_cq()