Lines Matching refs:send_cq

962 		scqe_sz = mlx5_ib_get_cqe_size(dev, init_attr->send_cq);  in create_qp_common()
1013 if (init_attr->send_cq) in create_qp_common()
1014 in->ctx.cqn_send = cpu_to_be32(to_mcq(init_attr->send_cq)->mcq.cqn); in create_qp_common()
1048 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq) in mlx5_ib_lock_cqs() argument
1049 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mlx5_ib_lock_cqs()
1051 if (send_cq) { in mlx5_ib_lock_cqs()
1053 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1054 spin_lock_irq(&send_cq->lock); in mlx5_ib_lock_cqs()
1057 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1058 spin_lock_irq(&send_cq->lock); in mlx5_ib_lock_cqs()
1062 spin_lock_nested(&send_cq->lock, in mlx5_ib_lock_cqs()
1066 spin_lock_irq(&send_cq->lock); in mlx5_ib_lock_cqs()
1071 __acquire(&send_cq->lock); in mlx5_ib_lock_cqs()
1073 __acquire(&send_cq->lock); in mlx5_ib_lock_cqs()
1078 static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq) in mlx5_ib_unlock_cqs() argument
1079 __releases(&send_cq->lock) __releases(&recv_cq->lock) in mlx5_ib_unlock_cqs()
1081 if (send_cq) { in mlx5_ib_unlock_cqs()
1083 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
1085 spin_unlock_irq(&send_cq->lock); in mlx5_ib_unlock_cqs()
1086 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
1088 spin_unlock_irq(&send_cq->lock); in mlx5_ib_unlock_cqs()
1090 spin_unlock(&send_cq->lock); in mlx5_ib_unlock_cqs()
1095 spin_unlock_irq(&send_cq->lock); in mlx5_ib_unlock_cqs()
1098 __release(&send_cq->lock); in mlx5_ib_unlock_cqs()
1102 __release(&send_cq->lock); in mlx5_ib_unlock_cqs()
1112 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq) in get_cqs() argument
1116 *send_cq = NULL; in get_cqs()
1121 *send_cq = to_mcq(qp->ibqp.send_cq); in get_cqs()
1132 *send_cq = to_mcq(qp->ibqp.send_cq); in get_cqs()
1139 *send_cq = NULL; in get_cqs()
1147 struct mlx5_ib_cq *send_cq, *recv_cq; in destroy_qp_common() local
1163 get_cqs(qp, &send_cq, &recv_cq); in destroy_qp_common()
1166 mlx5_ib_lock_cqs(send_cq, recv_cq); in destroy_qp_common()
1169 if (send_cq != recv_cq) in destroy_qp_common()
1170 __mlx5_ib_cq_clean(send_cq, qp->mqp.qpn, NULL); in destroy_qp_common()
1171 mlx5_ib_unlock_cqs(send_cq, recv_cq); in destroy_qp_common()
1249 init_attr->send_cq = NULL; in mlx5_ib_create_qp()
1279 to_mcq(init_attr->send_cq)->mcq.cqn); in mlx5_ib_create_qp()
1558 struct mlx5_ib_cq *send_cq, *recv_cq; in __mlx5_ib_modify_qp() local
1644 get_cqs(qp, &send_cq, &recv_cq); in __mlx5_ib_modify_qp()
1647 context->cqn_send = send_cq ? cpu_to_be32(send_cq->mcq.cqn) : 0; in __mlx5_ib_modify_qp()
1742 if (send_cq != recv_cq) in __mlx5_ib_modify_qp()
1743 mlx5_ib_cq_clean(send_cq, qp->mqp.qpn, NULL); in __mlx5_ib_modify_qp()
2541 if (unlikely(mlx5_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq))) { in begin_wqe()