Home
last modified time | relevance | path

Searched refs:mcq (Results 1 – 16 of 16) sorted by relevance

/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dcq.c46 static void mlx5_ib_cq_event(struct mlx5_core_cq *mcq, enum mlx5_event type) in mlx5_ib_cq_event() argument
48 struct mlx5_ib_cq *cq = container_of(mcq, struct mlx5_ib_cq, mcq); in mlx5_ib_cq_event()
55 type, mcq->cqn); in mlx5_ib_cq_event()
74 return get_cqe_from_buf(&cq->buf, n, cq->mcq.cqe_sz); in get_cqe()
87 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe()
99 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
425 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in mlx5_poll_one()
427 ++cq->mcq.cons_index; in mlx5_poll_one()
456 cq->mcq.cqn, qpn); in mlx5_poll_one()
490 "Requestor" : "Responder", cq->mcq.cqn); in mlx5_poll_one()
[all …]
Dmlx5_ib.h269 struct mlx5_core_cq mcq; member
446 static inline struct mlx5_ib_cq *to_mibcq(struct mlx5_core_cq *mcq) in to_mibcq() argument
448 return container_of(mcq, struct mlx5_ib_cq, mcq); in to_mibcq()
Dqp.c1001 in->ctx.cqn_recv = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1002 in->ctx.cqn_send = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1007 in->ctx.cqn_recv = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1022 in->ctx.cqn_send = cpu_to_be32(to_mcq(init_attr->send_cq)->mcq.cqn); in create_qp_common()
1025 in->ctx.cqn_recv = cpu_to_be32(to_mcq(init_attr->recv_cq)->mcq.cqn); in create_qp_common()
1061 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1065 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1091 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
1094 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
1288 qp->ibqp.qp_num, qp->mqp.qpn, to_mcq(init_attr->recv_cq)->mcq.cqn, in mlx5_ib_create_qp()
[all …]
Dsrq.c296 in->ctx.pgoff_cqn |= cpu_to_be32(to_mcq(init_attr->ext.xrc.cq)->mcq.cqn); in mlx5_ib_create_srq()
299 in->ctx.pgoff_cqn |= cpu_to_be32(to_mcq(dev->devr.c0)->mcq.cqn); in mlx5_ib_create_srq()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Den_cq.c112 cq->mcq.set_ci_db = cq->wqres.db.db; in mlx4_en_activate_cq()
113 cq->mcq.arm_db = cq->wqres.db.db + 1; in mlx4_en_activate_cq()
114 *cq->mcq.set_ci_db = 0; in mlx4_en_activate_cq()
115 *cq->mcq.arm_db = 0; in mlx4_en_activate_cq()
159 &mdev->priv_uar, cq->wqres.db.dma, &cq->mcq, in mlx4_en_activate_cq()
164 cq->mcq.comp = cq->is_tx ? mlx4_en_tx_irq : mlx4_en_rx_irq; in mlx4_en_activate_cq()
165 cq->mcq.event = mlx4_en_cq_event; in mlx4_en_activate_cq()
173 err = irq_set_affinity_hint(cq->mcq.irq, in mlx4_en_activate_cq()
210 irq_set_affinity_hint(cq->mcq.irq, NULL); in mlx4_en_deactivate_cq()
214 mlx4_cq_free(priv->mdev->dev, &cq->mcq); in mlx4_en_deactivate_cq()
[all …]
Dcq.c63 struct mlx4_cq *mcq, *temp; in mlx4_cq_tasklet_cb() local
69 list_for_each_entry_safe(mcq, temp, &ctx->process_list, tasklet_ctx.list) { in mlx4_cq_tasklet_cb()
70 list_del_init(&mcq->tasklet_ctx.list); in mlx4_cq_tasklet_cb()
71 mcq->tasklet_ctx.comp(mcq); in mlx4_cq_tasklet_cb()
72 if (atomic_dec_and_test(&mcq->refcount)) in mlx4_cq_tasklet_cb()
73 complete(&mcq->free); in mlx4_cq_tasklet_cb()
Den_tx.c389 struct mlx4_cq *mcq = &cq->mcq; in mlx4_en_process_tx_cq() local
396 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq()
473 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
474 mlx4_cq_set_ci(mcq); in mlx4_en_process_tx_cq()
493 void mlx4_en_tx_irq(struct mlx4_cq *mcq) in mlx4_en_tx_irq() argument
495 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_tx_irq()
Den_rx.c445 ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn; in mlx4_en_activate_rx_rings()
777 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq()
782 cq->mcq.cons_index & cq->size)) { in mlx4_en_process_rx_cq()
995 ++cq->mcq.cons_index; in mlx4_en_process_rx_cq()
996 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq()
1004 mlx4_cq_set_ci(&cq->mcq); in mlx4_en_process_rx_cq()
1006 ring->cons = cq->mcq.cons_index; in mlx4_en_process_rx_cq()
1013 void mlx4_en_rx_irq(struct mlx4_cq *mcq) in mlx4_en_rx_irq() argument
1015 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_rx_irq()
Dmlx4_en.h335 struct mlx4_cq mcq; member
760 void mlx4_en_tx_irq(struct mlx4_cq *mcq);
805 void mlx4_en_rx_irq(struct mlx4_cq *mcq);
Den_netdev.c1587 priv->rx_ring[i]->cqn = cq->mcq.cqn; in mlx4_en_start_port()
1630 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, in mlx4_en_start_port()
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dcq.c89 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
94 struct mlx4_ib_cq *mcq = to_mcq(cq); in mlx4_ib_modify_cq() local
97 return mlx4_cq_modify(dev->dev, &mcq->mcq, cq_count, cq_period); in mlx4_ib_modify_cq()
218 cq->mcq.set_ci_db = cq->db.db; in mlx4_ib_create_cq()
219 cq->mcq.arm_db = cq->db.db + 1; in mlx4_ib_create_cq()
220 *cq->mcq.set_ci_db = 0; in mlx4_ib_create_cq()
221 *cq->mcq.arm_db = 0; in mlx4_ib_create_cq()
234 cq->db.dma, &cq->mcq, vector, 0, 0); in mlx4_ib_create_cq()
239 cq->mcq.tasklet_ctx.comp = mlx4_ib_cq_comp; in mlx4_ib_create_cq()
241 cq->mcq.comp = mlx4_ib_cq_comp; in mlx4_ib_create_cq()
[all …]
Dmlx4_ib.h105 struct mlx4_cq mcq; member
583 static inline struct mlx4_ib_cq *to_mibcq(struct mlx4_cq *mcq) in to_mibcq() argument
585 return container_of(mcq, struct mlx4_ib_cq, mcq); in to_mibcq()
Dmain.c2633 struct mlx4_cq *mcq; in mlx4_ib_handle_catas_error() local
2647 if (send_mcq->mcq.comp && in mlx4_ib_handle_catas_error()
2649 if (!send_mcq->mcq.reset_notify_added) { in mlx4_ib_handle_catas_error()
2650 send_mcq->mcq.reset_notify_added = 1; in mlx4_ib_handle_catas_error()
2651 list_add_tail(&send_mcq->mcq.reset_notify, in mlx4_ib_handle_catas_error()
2665 if (recv_mcq->mcq.comp && in mlx4_ib_handle_catas_error()
2667 if (!recv_mcq->mcq.reset_notify_added) { in mlx4_ib_handle_catas_error()
2668 recv_mcq->mcq.reset_notify_added = 1; in mlx4_ib_handle_catas_error()
2669 list_add_tail(&recv_mcq->mcq.reset_notify, in mlx4_ib_handle_catas_error()
2680 list_for_each_entry(mcq, &cq_notify_list, reset_notify) { in mlx4_ib_handle_catas_error()
[all …]
Dqp.c630 struct mlx4_ib_cq *mcq; in create_qp_common() local
853 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
854 list_add_tail(&qp->cq_send_list, &mcq->send_qp_list); in create_qp_common()
855 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
856 list_add_tail(&qp->cq_recv_list, &mcq->recv_qp_list); in create_qp_common()
920 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_lock_cqs()
935 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_unlock_cqs()
1599 context->cqn_send = cpu_to_be32(send_cq->mcq.cqn); in __mlx4_ib_modify_qp()
1600 context->cqn_recv = cpu_to_be32(recv_cq->mcq.cqn); in __mlx4_ib_modify_qp()
Dsrq.c181 to_mcq(init_attr->ext.xrc.cq)->mcq.cqn : 0; in mlx4_ib_create_srq()
/linux-4.1.27/drivers/scsi/lpfc/
Dlpfc_sli.c6704 struct lpfc_queue *mcq; in lpfc_sli4_mbox_completions_pending() local
6713 mcq = phba->sli4_hba.mbx_cq; in lpfc_sli4_mbox_completions_pending()
6714 idx = mcq->hba_index; in lpfc_sli4_mbox_completions_pending()
6715 while (bf_get_le32(lpfc_cqe_valid, mcq->qe[idx].cqe)) { in lpfc_sli4_mbox_completions_pending()
6716 mcqe = (struct lpfc_mcqe *)mcq->qe[idx].cqe; in lpfc_sli4_mbox_completions_pending()
6722 idx = (idx + 1) % mcq->entry_count; in lpfc_sli4_mbox_completions_pending()
6723 if (mcq->hba_index == idx) in lpfc_sli4_mbox_completions_pending()