send_cq           345 drivers/infiniband/core/core_priv.h 	qp->send_cq = attr->send_cq;
send_cq          3154 drivers/infiniband/core/mad.c 	qp_init_attr.send_cq = qp_info->port_priv->cq;
send_cq          1377 drivers/infiniband/core/uverbs_cmd.c 	attr.send_cq       = scq;
send_cq          1435 drivers/infiniband/core/uverbs_cmd.c 		if (attr.send_cq)
send_cq          1436 drivers/infiniband/core/uverbs_cmd.c 			atomic_inc(&attr.send_cq->usecnt);
send_cq          1133 drivers/infiniband/core/verbs.c 	qp->send_cq = qp->recv_cq = NULL;
send_cq          1208 drivers/infiniband/core/verbs.c 	qp->send_cq = qp_init_attr->send_cq;
send_cq          1212 drivers/infiniband/core/verbs.c 	if (qp_init_attr->send_cq)
send_cq          1213 drivers/infiniband/core/verbs.c 		atomic_inc(&qp_init_attr->send_cq->usecnt);
send_cq          1876 drivers/infiniband/core/verbs.c 	scq  = qp->send_cq;
send_cq          2632 drivers/infiniband/core/verbs.c 	struct ib_cq *cq = qp->send_cq;
send_cq          1048 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (qp_init_attr->send_cq) {
send_cq          1049 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		cq = container_of(qp_init_attr->send_cq, struct bnxt_re_cq,
send_cq           713 drivers/infiniband/hw/cxgb3/iwch_provider.c 	schp = get_chp(rhp, ((struct iwch_cq *) attrs->send_cq)->cq.cqid);
send_cq           770 drivers/infiniband/hw/cxgb3/iwch_provider.c 	qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid;
send_cq           242 drivers/infiniband/hw/cxgb4/cq.c 	struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq);
send_cq           987 drivers/infiniband/hw/cxgb4/qp.c 	schp = to_c4iw_cq(qhp->ibqp.send_cq);
send_cq          1665 drivers/infiniband/hw/cxgb4/qp.c 	schp = to_c4iw_cq(qhp->ibqp.send_cq);
send_cq          2134 drivers/infiniband/hw/cxgb4/qp.c 	schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid);
send_cq          2197 drivers/infiniband/hw/cxgb4/qp.c 	qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid;
send_cq           392 drivers/infiniband/hw/efa/efa_verbs.c 	qp_init_attr->send_cq = ibqp->send_cq;
send_cq           707 drivers/infiniband/hw/efa/efa_verbs.c 	create_qp_params.send_cq_idx = to_ecq(init_attr->send_cq)->cq_idx;
send_cq           705 drivers/infiniband/hw/hfi1/qp.c 		   ib_cq_head(qp->ibqp.send_cq),
send_cq           706 drivers/infiniband/hw/hfi1/qp.c 		   ib_cq_tail(qp->ibqp.send_cq),
send_cq          1248 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_lock_cqs(struct hns_roce_cq *send_cq,
send_cq          1250 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq,
send_cq            93 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
send_cq           812 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		hr_qp->ibqp.send_cq	= cq;
send_cq          2651 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       to_hr_cq(ibqp->send_cq)->cqn);
send_cq          2694 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (ibqp->send_cq != ibqp->recv_cq)
send_cq          2695 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq),
send_cq          2796 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       to_hr_cq(ibqp->send_cq)->cqn);
send_cq          2862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       to_hr_cq(ibqp->send_cq)->cqn);
send_cq          3307 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (ibqp->send_cq != ibqp->recv_cq)
send_cq          3308 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq),
send_cq          3606 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_cq *send_cq, *recv_cq;
send_cq          3613 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	send_cq = to_hr_cq(hr_qp->ibqp.send_cq);
send_cq          3616 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	hns_roce_lock_cqs(send_cq, recv_cq);
send_cq          3620 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (send_cq != recv_cq)
send_cq          3621 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			__hns_roce_v1_cq_clean(send_cq, hr_qp->qpn, NULL);
send_cq          3623 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	hns_roce_unlock_cqs(send_cq, recv_cq);
send_cq           280 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
send_cq          3534 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn);
send_cq          3612 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn);
send_cq          4463 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (ibqp->send_cq != ibqp->recv_cq)
send_cq          4464 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq),
send_cq          4653 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_cq *send_cq, *recv_cq;
send_cq          4665 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	send_cq = to_hr_cq(hr_qp->ibqp.send_cq);
send_cq          4668 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	hns_roce_lock_cqs(send_cq, recv_cq);
send_cq          4673 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (send_cq != recv_cq)
send_cq          4674 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			__hns_roce_v2_cq_clean(send_cq, hr_qp->qpn, NULL);
send_cq          4679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	hns_roce_unlock_cqs(send_cq, recv_cq);
send_cq          1234 drivers/infiniband/hw/hns/hns_roce_qp.c void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq)
send_cq          1235 drivers/infiniband/hw/hns/hns_roce_qp.c 		       __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
send_cq          1237 drivers/infiniband/hw/hns/hns_roce_qp.c 	if (send_cq == recv_cq) {
send_cq          1238 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_lock_irq(&send_cq->lock);
send_cq          1240 drivers/infiniband/hw/hns/hns_roce_qp.c 	} else if (send_cq->cqn < recv_cq->cqn) {
send_cq          1241 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_lock_irq(&send_cq->lock);
send_cq          1245 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
send_cq          1249 drivers/infiniband/hw/hns/hns_roce_qp.c void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq,
send_cq          1250 drivers/infiniband/hw/hns/hns_roce_qp.c 			 struct hns_roce_cq *recv_cq) __releases(&send_cq->lock)
send_cq          1253 drivers/infiniband/hw/hns/hns_roce_qp.c 	if (send_cq == recv_cq) {
send_cq          1255 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_unlock_irq(&send_cq->lock);
send_cq          1256 drivers/infiniband/hw/hns/hns_roce_qp.c 	} else if (send_cq->cqn < recv_cq->cqn) {
send_cq          1258 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_unlock_irq(&send_cq->lock);
send_cq          1260 drivers/infiniband/hw/hns/hns_roce_qp.c 		spin_unlock(&send_cq->lock);
send_cq           610 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwqp->iwscq = to_iwcq(init_attr->send_cq);
send_cq           775 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	init_attr->send_cq = iwqp->ibqp.send_cq;
send_cq          1811 drivers/infiniband/hw/mlx4/mad.c 	qp_init_attr.init_attr.send_cq = ctx->cq;
send_cq          3122 drivers/infiniband/hw/mlx4/main.c 			send_mcq = to_mcq(mqp->ibqp.send_cq);
send_cq          3125 drivers/infiniband/hw/mlx4/main.c 			    mqp->ibqp.send_cq->comp_handler) {
send_cq            52 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq,
send_cq            54 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq,
send_cq           752 drivers/infiniband/hw/mlx4/qp.c 	if (init_attr->send_cq || init_attr->cap.max_send_wr) {
send_cq           959 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq),
send_cq           968 drivers/infiniband/hw/mlx4/qp.c 	mcq = to_mcq(init_attr->send_cq);
send_cq           972 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq),
send_cq          1246 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq),
send_cq          1255 drivers/infiniband/hw/mlx4/qp.c 	mcq = to_mcq(init_attr->send_cq);
send_cq          1259 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq),
send_cq          1317 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
send_cq          1318 drivers/infiniband/hw/mlx4/qp.c 	__acquires(&send_cq->lock) __acquires(&recv_cq->lock)
send_cq          1320 drivers/infiniband/hw/mlx4/qp.c 	if (send_cq == recv_cq) {
send_cq          1321 drivers/infiniband/hw/mlx4/qp.c 		spin_lock(&send_cq->lock);
send_cq          1323 drivers/infiniband/hw/mlx4/qp.c 	} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
send_cq          1324 drivers/infiniband/hw/mlx4/qp.c 		spin_lock(&send_cq->lock);
send_cq          1328 drivers/infiniband/hw/mlx4/qp.c 		spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
send_cq          1332 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
send_cq          1333 drivers/infiniband/hw/mlx4/qp.c 	__releases(&send_cq->lock) __releases(&recv_cq->lock)
send_cq          1335 drivers/infiniband/hw/mlx4/qp.c 	if (send_cq == recv_cq) {
send_cq          1337 drivers/infiniband/hw/mlx4/qp.c 		spin_unlock(&send_cq->lock);
send_cq          1338 drivers/infiniband/hw/mlx4/qp.c 	} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
send_cq          1340 drivers/infiniband/hw/mlx4/qp.c 		spin_unlock(&send_cq->lock);
send_cq          1342 drivers/infiniband/hw/mlx4/qp.c 		spin_unlock(&send_cq->lock);
send_cq          1366 drivers/infiniband/hw/mlx4/qp.c 		    struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq)
send_cq          1370 drivers/infiniband/hw/mlx4/qp.c 		*send_cq = to_mcq(to_mxrcd(qp->ibqp.xrcd)->cq);
send_cq          1371 drivers/infiniband/hw/mlx4/qp.c 		*recv_cq = *send_cq;
send_cq          1374 drivers/infiniband/hw/mlx4/qp.c 		*send_cq = to_mcq(qp->ibqp.send_cq);
send_cq          1375 drivers/infiniband/hw/mlx4/qp.c 		*recv_cq = *send_cq;
send_cq          1380 drivers/infiniband/hw/mlx4/qp.c 		*send_cq = (src == MLX4_IB_QP_SRC) ? to_mcq(qp->ibqp.send_cq) :
send_cq          1420 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_cq *send_cq, *recv_cq;
send_cq          1451 drivers/infiniband/hw/mlx4/qp.c 	get_cqs(qp, src, &send_cq, &recv_cq);
send_cq          1454 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_lock_cqs(send_cq, recv_cq);
send_cq          1463 drivers/infiniband/hw/mlx4/qp.c 		if (send_cq != recv_cq)
send_cq          1464 drivers/infiniband/hw/mlx4/qp.c 			__mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL);
send_cq          1469 drivers/infiniband/hw/mlx4/qp.c 	mlx4_ib_unlock_cqs(send_cq, recv_cq);
send_cq          1580 drivers/infiniband/hw/mlx4/qp.c 		init_attr->send_cq = to_mxrcd(init_attr->xrcd)->cq;
send_cq          1585 drivers/infiniband/hw/mlx4/qp.c 		init_attr->recv_cq = init_attr->send_cq;
send_cq          2144 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_cq *send_cq, *recv_cq;
send_cq          2379 drivers/infiniband/hw/mlx4/qp.c 		get_cqs(qp, src_type, &send_cq, &recv_cq);
send_cq          2381 drivers/infiniband/hw/mlx4/qp.c 		send_cq = to_mcq(rwq_ind_tbl->ind_tbl[0]->cq);
send_cq          2382 drivers/infiniband/hw/mlx4/qp.c 		recv_cq = send_cq;
send_cq          2384 drivers/infiniband/hw/mlx4/qp.c 	context->cqn_send = cpu_to_be32(send_cq->mcq.cqn);
send_cq          2603 drivers/infiniband/hw/mlx4/qp.c 			if (send_cq != recv_cq)
send_cq          2604 drivers/infiniband/hw/mlx4/qp.c 				mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL);
send_cq          3587 drivers/infiniband/hw/mlx4/qp.c 		if (mlx4_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
send_cq          4183 drivers/infiniband/hw/mlx4/qp.c 	ib_qp_init_attr.send_cq = ib_qp_init_attr.recv_cq; /* Dummy CQ */
send_cq          4479 drivers/infiniband/hw/mlx4/qp.c 	struct ib_cq *cq = qp->send_cq;
send_cq            75 drivers/infiniband/hw/mlx5/gsi.c 	struct ib_cq *gsi_cq = gsi->ibqp.send_cq;
send_cq           181 drivers/infiniband/hw/mlx5/gsi.c 	hw_init_attr.send_cq = gsi->cq;
send_cq           257 drivers/infiniband/hw/mlx5/gsi.c 		.send_cq = gsi->cq,
send_cq          4484 drivers/infiniband/hw/mlx5/main.c 			send_mcq = to_mcq(mqp->ibqp.send_cq);
send_cq          4487 drivers/infiniband/hw/mlx5/main.c 			    mqp->ibqp.send_cq->comp_handler) {
send_cq          4834 drivers/infiniband/hw/mlx5/main.c 	init_attr->send_cq = cq;
send_cq          4851 drivers/infiniband/hw/mlx5/main.c 	qp->send_cq    = init_attr->send_cq;
send_cq           101 drivers/infiniband/hw/mlx5/qp.c 		    struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq);
send_cq           711 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq,
send_cq           713 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq,
send_cq          1612 drivers/infiniband/hw/mlx5/qp.c 	if (init_attr->create_flags || init_attr->send_cq)
send_cq          1882 drivers/infiniband/hw/mlx5/qp.c 	scqe_sz = mlx5_ib_get_cqe_size(init_attr->send_cq);
send_cq          1955 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_cq *send_cq;
send_cq          2243 drivers/infiniband/hw/mlx5/qp.c 	if (init_attr->send_cq)
send_cq          2244 drivers/infiniband/hw/mlx5/qp.c 		MLX5_SET(qpc, qpc, cqn_snd, to_mcq(init_attr->send_cq)->mcq.cqn);
send_cq          2300 drivers/infiniband/hw/mlx5/qp.c 	get_cqs(init_attr->qp_type, init_attr->send_cq, init_attr->recv_cq,
send_cq          2301 drivers/infiniband/hw/mlx5/qp.c 		&send_cq, &recv_cq);
send_cq          2303 drivers/infiniband/hw/mlx5/qp.c 	mlx5_ib_lock_cqs(send_cq, recv_cq);
send_cq          2310 drivers/infiniband/hw/mlx5/qp.c 	if (send_cq)
send_cq          2311 drivers/infiniband/hw/mlx5/qp.c 		list_add_tail(&qp->cq_send_list, &send_cq->list_send_qp);
send_cq          2314 drivers/infiniband/hw/mlx5/qp.c 	mlx5_ib_unlock_cqs(send_cq, recv_cq);
send_cq          2330 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq)
send_cq          2331 drivers/infiniband/hw/mlx5/qp.c 	__acquires(&send_cq->lock) __acquires(&recv_cq->lock)
send_cq          2333 drivers/infiniband/hw/mlx5/qp.c 	if (send_cq) {
send_cq          2335 drivers/infiniband/hw/mlx5/qp.c 			if (send_cq->mcq.cqn < recv_cq->mcq.cqn)  {
send_cq          2336 drivers/infiniband/hw/mlx5/qp.c 				spin_lock(&send_cq->lock);
send_cq          2339 drivers/infiniband/hw/mlx5/qp.c 			} else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) {
send_cq          2340 drivers/infiniband/hw/mlx5/qp.c 				spin_lock(&send_cq->lock);
send_cq          2344 drivers/infiniband/hw/mlx5/qp.c 				spin_lock_nested(&send_cq->lock,
send_cq          2348 drivers/infiniband/hw/mlx5/qp.c 			spin_lock(&send_cq->lock);
send_cq          2353 drivers/infiniband/hw/mlx5/qp.c 		__acquire(&send_cq->lock);
send_cq          2355 drivers/infiniband/hw/mlx5/qp.c 		__acquire(&send_cq->lock);
send_cq          2360 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq)
send_cq          2361 drivers/infiniband/hw/mlx5/qp.c 	__releases(&send_cq->lock) __releases(&recv_cq->lock)
send_cq          2363 drivers/infiniband/hw/mlx5/qp.c 	if (send_cq) {
send_cq          2365 drivers/infiniband/hw/mlx5/qp.c 			if (send_cq->mcq.cqn < recv_cq->mcq.cqn)  {
send_cq          2367 drivers/infiniband/hw/mlx5/qp.c 				spin_unlock(&send_cq->lock);
send_cq          2368 drivers/infiniband/hw/mlx5/qp.c 			} else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) {
send_cq          2370 drivers/infiniband/hw/mlx5/qp.c 				spin_unlock(&send_cq->lock);
send_cq          2372 drivers/infiniband/hw/mlx5/qp.c 				spin_unlock(&send_cq->lock);
send_cq          2377 drivers/infiniband/hw/mlx5/qp.c 			spin_unlock(&send_cq->lock);
send_cq          2380 drivers/infiniband/hw/mlx5/qp.c 		__release(&send_cq->lock);
send_cq          2384 drivers/infiniband/hw/mlx5/qp.c 		__release(&send_cq->lock);
send_cq          2395 drivers/infiniband/hw/mlx5/qp.c 		    struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq)
send_cq          2399 drivers/infiniband/hw/mlx5/qp.c 		*send_cq = NULL;
send_cq          2404 drivers/infiniband/hw/mlx5/qp.c 		*send_cq = ib_send_cq ? to_mcq(ib_send_cq) : NULL;
send_cq          2416 drivers/infiniband/hw/mlx5/qp.c 		*send_cq = ib_send_cq ? to_mcq(ib_send_cq) : NULL;
send_cq          2422 drivers/infiniband/hw/mlx5/qp.c 		*send_cq = NULL;
send_cq          2435 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_cq *send_cq, *recv_cq;
send_cq          2468 drivers/infiniband/hw/mlx5/qp.c 	get_cqs(qp->ibqp.qp_type, qp->ibqp.send_cq, qp->ibqp.recv_cq,
send_cq          2469 drivers/infiniband/hw/mlx5/qp.c 		&send_cq, &recv_cq);
send_cq          2472 drivers/infiniband/hw/mlx5/qp.c 	mlx5_ib_lock_cqs(send_cq, recv_cq);
send_cq          2475 drivers/infiniband/hw/mlx5/qp.c 	if (send_cq)
send_cq          2484 drivers/infiniband/hw/mlx5/qp.c 		if (send_cq != recv_cq)
send_cq          2485 drivers/infiniband/hw/mlx5/qp.c 			__mlx5_ib_cq_clean(send_cq, base->mqp.qpn,
send_cq          2488 drivers/infiniband/hw/mlx5/qp.c 	mlx5_ib_unlock_cqs(send_cq, recv_cq);
send_cq          2693 drivers/infiniband/hw/mlx5/qp.c 			init_attr->send_cq = NULL;
send_cq          2726 drivers/infiniband/hw/mlx5/qp.c 			    init_attr->send_cq ? to_mcq(init_attr->send_cq)->mcq.cqn : -1);
send_cq          3455 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_cq *send_cq, *recv_cq;
send_cq          3563 drivers/infiniband/hw/mlx5/qp.c 	get_cqs(qp->ibqp.qp_type, qp->ibqp.send_cq, qp->ibqp.recv_cq,
send_cq          3564 drivers/infiniband/hw/mlx5/qp.c 		&send_cq, &recv_cq);
send_cq          3567 drivers/infiniband/hw/mlx5/qp.c 	context->cqn_send = send_cq ? cpu_to_be32(send_cq->mcq.cqn) : 0;
send_cq          3718 drivers/infiniband/hw/mlx5/qp.c 		if (send_cq != recv_cq)
send_cq          3719 drivers/infiniband/hw/mlx5/qp.c 			mlx5_ib_cq_clean(send_cq, base->mqp.qpn, NULL);
send_cq          4902 drivers/infiniband/hw/mlx5/qp.c 	if (unlikely(mlx5_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)))
send_cq          5811 drivers/infiniband/hw/mlx5/qp.c 	qp_init_attr->send_cq = ibqp->send_cq;
send_cq          6441 drivers/infiniband/hw/mlx5/qp.c 	struct ib_cq *cq = qp->send_cq;
send_cq           546 drivers/infiniband/hw/mthca/mthca_dev.h 		   struct mthca_cq *send_cq,
send_cq           555 drivers/infiniband/hw/mthca/mthca_dev.h 		    struct mthca_cq *send_cq,
send_cq           522 drivers/infiniband/hw/mthca/mthca_provider.c 				     to_mcq(init_attr->send_cq),
send_cq           555 drivers/infiniband/hw/mthca/mthca_provider.c 				      to_mcq(init_attr->send_cq),
send_cq           735 drivers/infiniband/hw/mthca/mthca_qp.c 	qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn);
send_cq           837 drivers/infiniband/hw/mthca/mthca_qp.c 		if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
send_cq           838 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL);
send_cq          1161 drivers/infiniband/hw/mthca/mthca_qp.c 				 struct mthca_cq *send_cq,
send_cq          1290 drivers/infiniband/hw/mthca/mthca_qp.c 		   struct mthca_cq *send_cq,
send_cq          1318 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
send_cq          1333 drivers/infiniband/hw/mthca/mthca_qp.c static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
send_cq          1334 drivers/infiniband/hw/mthca/mthca_qp.c 	__acquires(&send_cq->lock) __acquires(&recv_cq->lock)
send_cq          1336 drivers/infiniband/hw/mthca/mthca_qp.c 	if (send_cq == recv_cq) {
send_cq          1337 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_lock_irq(&send_cq->lock);
send_cq          1339 drivers/infiniband/hw/mthca/mthca_qp.c 	} else if (send_cq->cqn < recv_cq->cqn) {
send_cq          1340 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_lock_irq(&send_cq->lock);
send_cq          1344 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
send_cq          1348 drivers/infiniband/hw/mthca/mthca_qp.c static void mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
send_cq          1349 drivers/infiniband/hw/mthca/mthca_qp.c 	__releases(&send_cq->lock) __releases(&recv_cq->lock)
send_cq          1351 drivers/infiniband/hw/mthca/mthca_qp.c 	if (send_cq == recv_cq) {
send_cq          1353 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_unlock_irq(&send_cq->lock);
send_cq          1354 drivers/infiniband/hw/mthca/mthca_qp.c 	} else if (send_cq->cqn < recv_cq->cqn) {
send_cq          1356 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_unlock_irq(&send_cq->lock);
send_cq          1358 drivers/infiniband/hw/mthca/mthca_qp.c 		spin_unlock(&send_cq->lock);
send_cq          1365 drivers/infiniband/hw/mthca/mthca_qp.c 		    struct mthca_cq *send_cq,
send_cq          1402 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
send_cq          1416 drivers/infiniband/hw/mthca/mthca_qp.c 	mthca_lock_cqs(send_cq, recv_cq);
send_cq          1422 drivers/infiniband/hw/mthca/mthca_qp.c 	mthca_unlock_cqs(send_cq, recv_cq);
send_cq          1445 drivers/infiniband/hw/mthca/mthca_qp.c 	struct mthca_cq *send_cq;
send_cq          1448 drivers/infiniband/hw/mthca/mthca_qp.c 	send_cq = to_mcq(qp->ibqp.send_cq);
send_cq          1455 drivers/infiniband/hw/mthca/mthca_qp.c 	mthca_lock_cqs(send_cq, recv_cq);
send_cq          1463 drivers/infiniband/hw/mthca/mthca_qp.c 	mthca_unlock_cqs(send_cq, recv_cq);
send_cq          1479 drivers/infiniband/hw/mthca/mthca_qp.c 		if (send_cq != recv_cq)
send_cq          1480 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_cq_clean(dev, send_cq, qp->qpn, NULL);
send_cq          1654 drivers/infiniband/hw/mthca/mthca_qp.c 		if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
send_cq          1984 drivers/infiniband/hw/mthca/mthca_qp.c 		if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
send_cq          2431 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cq = get_ocrdma_cq(attrs->send_cq);
send_cq          1174 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if ((dev->gsi_sqcq == get_ocrdma_cq(attrs->send_cq)) ||
send_cq          1301 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		dev->gsi_sqcq = get_ocrdma_cq(attrs->send_cq);
send_cq            62 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	dev->gsi_sqcq = get_qedr_cq(attrs->send_cq);
send_cq           139 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cq = get_qedr_cq(attrs->send_cq);
send_cq          1207 drivers/infiniband/hw/qedr/verbs.c 	qp->sq_cq = get_qedr_cq(attrs->send_cq);
send_cq          1529 drivers/infiniband/hw/qedr/verbs.c 	params->sq_cq_id = get_qedr_cq(attrs->send_cq)->icid;
send_cq          1904 drivers/infiniband/hw/qedr/verbs.c 		 get_qedr_cq(attrs->send_cq),
send_cq          1905 drivers/infiniband/hw/qedr/verbs.c 		 get_qedr_cq(attrs->send_cq)->icid,
send_cq           512 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	cq_cnt = (init_attr->send_cq == init_attr->recv_cq) ? 1 : 2;
send_cq            55 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq,
send_cq            58 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	*send_cq = to_vcq(qp->ibqp.send_cq);
send_cq           350 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle;
send_cq           989 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	init_attr->send_cq = qp->ibqp.send_cq;
send_cq          1798 drivers/infiniband/sw/rdmavt/qp.c 	init_attr->send_cq = qp->ibqp.send_cq;
send_cq            92 drivers/infiniband/sw/rxe/rxe_qp.c 	if (!init->recv_cq || !init->send_cq) {
send_cq           344 drivers/infiniband/sw/rxe/rxe_qp.c 	struct rxe_cq *scq = to_rcq(init->send_cq);
send_cq           390 drivers/infiniband/sw/rxe/rxe_qp.c 	init->send_cq			= qp->ibqp.send_cq;
send_cq           356 drivers/infiniband/sw/siw/siw_verbs.c 	scq = to_siw_cq(attrs->send_cq);
send_cq           549 drivers/infiniband/sw/siw/siw_verbs.c 	qp_init_attr->send_cq = base_qp->send_cq;
send_cq           366 drivers/infiniband/ulp/ipoib/ipoib.h 	struct ib_cq	 *send_cq;
send_cq           255 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		.send_cq = priv->recv_cq, /* For drain WR */
send_cq           770 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		rc = ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP |
send_cq          1061 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		.send_cq		= priv->send_cq,
send_cq           439 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc);
send_cq           507 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc);
send_cq           519 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP |
send_cq           649 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP |
send_cq           187 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL,
send_cq           189 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	if (IS_ERR(priv->send_cq)) {
send_cq           197 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	init_attr.send_cq = priv->send_cq;
send_cq           215 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP))
send_cq           241 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	ib_destroy_cq(priv->send_cq);
send_cq           263 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	ib_destroy_cq(priv->send_cq);
send_cq           393 drivers/infiniband/ulp/iser/iser_verbs.c 	init_attr.send_cq	= ib_conn->comp->cq;
send_cq           126 drivers/infiniband/ulp/isert/ib_isert.c 	attr.send_cq = comp->cq;
send_cq           544 drivers/infiniband/ulp/srp/ib_srp.c 	ib_process_cq_direct(ch->send_cq, -1);
send_cq           556 drivers/infiniband/ulp/srp/ib_srp.c 	struct ib_cq *recv_cq, *send_cq;
send_cq           575 drivers/infiniband/ulp/srp/ib_srp.c 	send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size,
send_cq           577 drivers/infiniband/ulp/srp/ib_srp.c 	if (IS_ERR(send_cq)) {
send_cq           578 drivers/infiniband/ulp/srp/ib_srp.c 		ret = PTR_ERR(send_cq);
send_cq           589 drivers/infiniband/ulp/srp/ib_srp.c 	init_attr->send_cq             = send_cq;
send_cq           633 drivers/infiniband/ulp/srp/ib_srp.c 	if (ch->send_cq)
send_cq           634 drivers/infiniband/ulp/srp/ib_srp.c 		ib_free_cq(ch->send_cq);
send_cq           638 drivers/infiniband/ulp/srp/ib_srp.c 	ch->send_cq = send_cq;
send_cq           660 drivers/infiniband/ulp/srp/ib_srp.c 	ib_free_cq(send_cq);
send_cq           708 drivers/infiniband/ulp/srp/ib_srp.c 	ib_free_cq(ch->send_cq);
send_cq           720 drivers/infiniband/ulp/srp/ib_srp.c 	ch->send_cq = ch->recv_cq = NULL;
send_cq          2025 drivers/infiniband/ulp/srp/ib_srp.c 	ib_process_cq_direct(ch->send_cq, -1);
send_cq           155 drivers/infiniband/ulp/srp/ib_srp.h 	struct ib_cq	       *send_cq;
send_cq          1806 drivers/infiniband/ulp/srpt/ib_srpt.c 	qp_init->send_cq = ch->cq;
send_cq           350 drivers/net/ethernet/ibm/ehea/ehea.h 	struct ehea_cq *send_cq;
send_cq           195 drivers/net/ethernet/ibm/ehea/ehea_main.c 				arr[i++].fwh = pr->send_cq->fw_handle;
send_cq           800 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct ehea_cq *send_cq = pr->send_cq;
send_cq           809 drivers/net/ethernet/ibm/ehea/ehea_main.c 	cqe = ehea_poll_cq(send_cq);
send_cq           811 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_inc_cq(send_cq);
send_cq           851 drivers/net/ethernet/ibm/ehea/ehea_main.c 		cqe = ehea_poll_cq(send_cq);
send_cq           854 drivers/net/ethernet/ibm/ehea/ehea_main.c 	ehea_update_feca(send_cq, cqe_counter);
send_cq           889 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_reset_cq_ep(pr->send_cq);
send_cq           891 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_reset_cq_n1(pr->send_cq);
send_cq           894 drivers/net/ethernet/ibm/ehea/ehea_main.c 		cqe_skb = ehea_poll_cq(pr->send_cq);
send_cq          1473 drivers/net/ethernet/ibm/ehea/ehea_main.c 	pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq,
send_cq          1476 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (!pr->send_cq) {
send_cq          1483 drivers/net/ethernet/ibm/ehea/ehea_main.c 			pr->send_cq->attr.act_nr_of_cqes,
send_cq          1508 drivers/net/ethernet/ibm/ehea/ehea_main.c 	init_attr->send_cq_handle = pr->send_cq->fw_handle;
send_cq          1558 drivers/net/ethernet/ibm/ehea/ehea_main.c 	ehea_destroy_cq(pr->send_cq);
send_cq          1575 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_destroy_cq(pr->send_cq);
send_cq           270 drivers/nvme/host/rdma.c 	init_attr.send_cq = queue->ib_cq;
send_cq          1005 drivers/nvme/target/rdma.c 	qp_attr.send_cq = queue->cq;
send_cq          1464 fs/cifs/smbdirect.c 	ib_free_cq(info->send_cq);
send_cq          1657 fs/cifs/smbdirect.c 	info->send_cq = NULL;
send_cq          1659 fs/cifs/smbdirect.c 	info->send_cq =
send_cq          1662 fs/cifs/smbdirect.c 	if (IS_ERR(info->send_cq)) {
send_cq          1663 fs/cifs/smbdirect.c 		info->send_cq = NULL;
send_cq          1685 fs/cifs/smbdirect.c 	qp_attr.send_cq = info->send_cq;
send_cq          1804 fs/cifs/smbdirect.c 	if (info->send_cq)
send_cq          1805 fs/cifs/smbdirect.c 		ib_free_cq(info->send_cq);
send_cq            59 fs/cifs/smbdirect.h 	struct ib_cq *send_cq, *recv_cq;
send_cq          1127 include/rdma/ib_verbs.h 	struct ib_cq	       *send_cq;
send_cq          1714 include/rdma/ib_verbs.h 	struct ib_cq	       *send_cq;
send_cq           821 include/rdma/rdmavt_qp.h 	struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq);
send_cq           936 include/rdma/rdmavt_qp.h static inline u32 ib_cq_tail(struct ib_cq *send_cq)
send_cq           938 include/rdma/rdmavt_qp.h 	struct rvt_cq *cq = ibcq_to_rvtcq(send_cq);
send_cq           940 include/rdma/rdmavt_qp.h 	return ibcq_to_rvtcq(send_cq)->ip ?
send_cq           942 include/rdma/rdmavt_qp.h 	       ibcq_to_rvtcq(send_cq)->kqueue->tail;
send_cq           952 include/rdma/rdmavt_qp.h static inline u32 ib_cq_head(struct ib_cq *send_cq)
send_cq           954 include/rdma/rdmavt_qp.h 	struct rvt_cq *cq = ibcq_to_rvtcq(send_cq);
send_cq           956 include/rdma/rdmavt_qp.h 	return ibcq_to_rvtcq(send_cq)->ip ?
send_cq           958 include/rdma/rdmavt_qp.h 	       ibcq_to_rvtcq(send_cq)->kqueue->head;
send_cq           709 net/9p/trans_rdma.c 	qp_attr.send_cq = rdma->cq;
send_cq           537 net/rds/ib_cm.c 	attr.send_cq = ic->i_send_cq;
send_cq           331 net/smc/smc_ib.c 		.send_cq = lnk->smcibdev->roce_cq_send,
send_cq           479 net/sunrpc/xprtrdma/svc_rdma_transport.c 	qp_attr.send_cq = newxprt->sc_sq_cq;
send_cq           426 net/sunrpc/xprtrdma/verbs.c 	ib_free_cq(ep->rep_attr.send_cq);
send_cq           427 net/sunrpc/xprtrdma/verbs.c 	ep->rep_attr.send_cq = NULL;
send_cq           539 net/sunrpc/xprtrdma/verbs.c 	ep->rep_attr.send_cq = sendcq;
send_cq           598 net/sunrpc/xprtrdma/verbs.c 	if (ep->rep_attr.send_cq)
send_cq           599 net/sunrpc/xprtrdma/verbs.c 		ib_free_cq(ep->rep_attr.send_cq);