send_cq 345 drivers/infiniband/core/core_priv.h qp->send_cq = attr->send_cq; send_cq 3154 drivers/infiniband/core/mad.c qp_init_attr.send_cq = qp_info->port_priv->cq; send_cq 1377 drivers/infiniband/core/uverbs_cmd.c attr.send_cq = scq; send_cq 1435 drivers/infiniband/core/uverbs_cmd.c if (attr.send_cq) send_cq 1436 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&attr.send_cq->usecnt); send_cq 1133 drivers/infiniband/core/verbs.c qp->send_cq = qp->recv_cq = NULL; send_cq 1208 drivers/infiniband/core/verbs.c qp->send_cq = qp_init_attr->send_cq; send_cq 1212 drivers/infiniband/core/verbs.c if (qp_init_attr->send_cq) send_cq 1213 drivers/infiniband/core/verbs.c atomic_inc(&qp_init_attr->send_cq->usecnt); send_cq 1876 drivers/infiniband/core/verbs.c scq = qp->send_cq; send_cq 2632 drivers/infiniband/core/verbs.c struct ib_cq *cq = qp->send_cq; send_cq 1048 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (qp_init_attr->send_cq) { send_cq 1049 drivers/infiniband/hw/bnxt_re/ib_verbs.c cq = container_of(qp_init_attr->send_cq, struct bnxt_re_cq, send_cq 713 drivers/infiniband/hw/cxgb3/iwch_provider.c schp = get_chp(rhp, ((struct iwch_cq *) attrs->send_cq)->cq.cqid); send_cq 770 drivers/infiniband/hw/cxgb3/iwch_provider.c qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid; send_cq 242 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); send_cq 987 drivers/infiniband/hw/cxgb4/qp.c schp = to_c4iw_cq(qhp->ibqp.send_cq); send_cq 1665 drivers/infiniband/hw/cxgb4/qp.c schp = to_c4iw_cq(qhp->ibqp.send_cq); send_cq 2134 drivers/infiniband/hw/cxgb4/qp.c schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid); send_cq 2197 drivers/infiniband/hw/cxgb4/qp.c qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid; send_cq 392 drivers/infiniband/hw/efa/efa_verbs.c qp_init_attr->send_cq = ibqp->send_cq; send_cq 707 drivers/infiniband/hw/efa/efa_verbs.c create_qp_params.send_cq_idx = to_ecq(init_attr->send_cq)->cq_idx; send_cq 705 drivers/infiniband/hw/hfi1/qp.c ib_cq_head(qp->ibqp.send_cq), send_cq 706 drivers/infiniband/hw/hfi1/qp.c ib_cq_tail(qp->ibqp.send_cq), send_cq 1248 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, send_cq 1250 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq, send_cq 93 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { send_cq 812 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hr_qp->ibqp.send_cq = cq; send_cq 2651 drivers/infiniband/hw/hns/hns_roce_hw_v1.c to_hr_cq(ibqp->send_cq)->cqn); send_cq 2694 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (ibqp->send_cq != ibqp->recv_cq) send_cq 2695 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), send_cq 2796 drivers/infiniband/hw/hns/hns_roce_hw_v1.c to_hr_cq(ibqp->send_cq)->cqn); send_cq 2862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c to_hr_cq(ibqp->send_cq)->cqn); send_cq 3307 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (ibqp->send_cq != ibqp->recv_cq) send_cq 3308 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), send_cq 3606 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_cq *send_cq, *recv_cq; send_cq 3613 drivers/infiniband/hw/hns/hns_roce_hw_v1.c send_cq = to_hr_cq(hr_qp->ibqp.send_cq); send_cq 3616 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_lock_cqs(send_cq, recv_cq); send_cq 3620 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (send_cq != recv_cq) send_cq 3621 drivers/infiniband/hw/hns/hns_roce_hw_v1.c __hns_roce_v1_cq_clean(send_cq, hr_qp->qpn, NULL); send_cq 3623 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_unlock_cqs(send_cq, recv_cq); send_cq 280 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { send_cq 3534 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); send_cq 3612 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); send_cq 4463 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (ibqp->send_cq != ibqp->recv_cq) send_cq 4464 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq), send_cq 4653 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_cq *send_cq, *recv_cq; send_cq 4665 drivers/infiniband/hw/hns/hns_roce_hw_v2.c send_cq = to_hr_cq(hr_qp->ibqp.send_cq); send_cq 4668 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_lock_cqs(send_cq, recv_cq); send_cq 4673 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (send_cq != recv_cq) send_cq 4674 drivers/infiniband/hw/hns/hns_roce_hw_v2.c __hns_roce_v2_cq_clean(send_cq, hr_qp->qpn, NULL); send_cq 4679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_unlock_cqs(send_cq, recv_cq); send_cq 1234 drivers/infiniband/hw/hns/hns_roce_qp.c void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) send_cq 1235 drivers/infiniband/hw/hns/hns_roce_qp.c __acquires(&send_cq->lock) __acquires(&recv_cq->lock) send_cq 1237 drivers/infiniband/hw/hns/hns_roce_qp.c if (send_cq == recv_cq) { send_cq 1238 drivers/infiniband/hw/hns/hns_roce_qp.c spin_lock_irq(&send_cq->lock); send_cq 1240 drivers/infiniband/hw/hns/hns_roce_qp.c } else if (send_cq->cqn < recv_cq->cqn) { send_cq 1241 drivers/infiniband/hw/hns/hns_roce_qp.c spin_lock_irq(&send_cq->lock); send_cq 1245 drivers/infiniband/hw/hns/hns_roce_qp.c spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); send_cq 1249 drivers/infiniband/hw/hns/hns_roce_qp.c void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq, send_cq 1250 drivers/infiniband/hw/hns/hns_roce_qp.c struct hns_roce_cq *recv_cq) __releases(&send_cq->lock) send_cq 1253 drivers/infiniband/hw/hns/hns_roce_qp.c if (send_cq == recv_cq) { send_cq 1255 drivers/infiniband/hw/hns/hns_roce_qp.c spin_unlock_irq(&send_cq->lock); send_cq 1256 drivers/infiniband/hw/hns/hns_roce_qp.c } else if (send_cq->cqn < recv_cq->cqn) { send_cq 1258 drivers/infiniband/hw/hns/hns_roce_qp.c spin_unlock_irq(&send_cq->lock); send_cq 1260 drivers/infiniband/hw/hns/hns_roce_qp.c spin_unlock(&send_cq->lock); send_cq 610 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwqp->iwscq = to_iwcq(init_attr->send_cq); send_cq 775 drivers/infiniband/hw/i40iw/i40iw_verbs.c init_attr->send_cq = iwqp->ibqp.send_cq; send_cq 1811 drivers/infiniband/hw/mlx4/mad.c qp_init_attr.init_attr.send_cq = ctx->cq; send_cq 3122 drivers/infiniband/hw/mlx4/main.c send_mcq = to_mcq(mqp->ibqp.send_cq); send_cq 3125 drivers/infiniband/hw/mlx4/main.c mqp->ibqp.send_cq->comp_handler) { send_cq 52 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, send_cq 54 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, send_cq 752 drivers/infiniband/hw/mlx4/qp.c if (init_attr->send_cq || init_attr->cap.max_send_wr) { send_cq 959 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), send_cq 968 drivers/infiniband/hw/mlx4/qp.c mcq = to_mcq(init_attr->send_cq); send_cq 972 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), send_cq 1246 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), send_cq 1255 drivers/infiniband/hw/mlx4/qp.c mcq = to_mcq(init_attr->send_cq); send_cq 1259 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), send_cq 1317 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) send_cq 1318 drivers/infiniband/hw/mlx4/qp.c __acquires(&send_cq->lock) __acquires(&recv_cq->lock) send_cq 1320 drivers/infiniband/hw/mlx4/qp.c if (send_cq == recv_cq) { send_cq 1321 drivers/infiniband/hw/mlx4/qp.c spin_lock(&send_cq->lock); send_cq 1323 drivers/infiniband/hw/mlx4/qp.c } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { send_cq 1324 drivers/infiniband/hw/mlx4/qp.c spin_lock(&send_cq->lock); send_cq 1328 drivers/infiniband/hw/mlx4/qp.c spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); send_cq 1332 drivers/infiniband/hw/mlx4/qp.c static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) send_cq 1333 drivers/infiniband/hw/mlx4/qp.c __releases(&send_cq->lock) __releases(&recv_cq->lock) send_cq 1335 drivers/infiniband/hw/mlx4/qp.c if (send_cq == recv_cq) { send_cq 1337 drivers/infiniband/hw/mlx4/qp.c spin_unlock(&send_cq->lock); send_cq 1338 drivers/infiniband/hw/mlx4/qp.c } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { send_cq 1340 drivers/infiniband/hw/mlx4/qp.c spin_unlock(&send_cq->lock); send_cq 1342 drivers/infiniband/hw/mlx4/qp.c spin_unlock(&send_cq->lock); send_cq 1366 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) send_cq 1370 drivers/infiniband/hw/mlx4/qp.c *send_cq = to_mcq(to_mxrcd(qp->ibqp.xrcd)->cq); send_cq 1371 drivers/infiniband/hw/mlx4/qp.c *recv_cq = *send_cq; send_cq 1374 drivers/infiniband/hw/mlx4/qp.c *send_cq = to_mcq(qp->ibqp.send_cq); send_cq 1375 drivers/infiniband/hw/mlx4/qp.c *recv_cq = *send_cq; send_cq 1380 drivers/infiniband/hw/mlx4/qp.c *send_cq = (src == MLX4_IB_QP_SRC) ? to_mcq(qp->ibqp.send_cq) : send_cq 1420 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_cq *send_cq, *recv_cq; send_cq 1451 drivers/infiniband/hw/mlx4/qp.c get_cqs(qp, src, &send_cq, &recv_cq); send_cq 1454 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_lock_cqs(send_cq, recv_cq); send_cq 1463 drivers/infiniband/hw/mlx4/qp.c if (send_cq != recv_cq) send_cq 1464 drivers/infiniband/hw/mlx4/qp.c __mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL); send_cq 1469 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_unlock_cqs(send_cq, recv_cq); send_cq 1580 drivers/infiniband/hw/mlx4/qp.c init_attr->send_cq = to_mxrcd(init_attr->xrcd)->cq; send_cq 1585 drivers/infiniband/hw/mlx4/qp.c init_attr->recv_cq = init_attr->send_cq; send_cq 2144 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_cq *send_cq, *recv_cq; send_cq 2379 drivers/infiniband/hw/mlx4/qp.c get_cqs(qp, src_type, &send_cq, &recv_cq); send_cq 2381 drivers/infiniband/hw/mlx4/qp.c send_cq = to_mcq(rwq_ind_tbl->ind_tbl[0]->cq); send_cq 2382 drivers/infiniband/hw/mlx4/qp.c recv_cq = send_cq; send_cq 2384 drivers/infiniband/hw/mlx4/qp.c context->cqn_send = cpu_to_be32(send_cq->mcq.cqn); send_cq 2603 drivers/infiniband/hw/mlx4/qp.c if (send_cq != recv_cq) send_cq 2604 drivers/infiniband/hw/mlx4/qp.c mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL); send_cq 3587 drivers/infiniband/hw/mlx4/qp.c if (mlx4_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { send_cq 4183 drivers/infiniband/hw/mlx4/qp.c ib_qp_init_attr.send_cq = ib_qp_init_attr.recv_cq; /* Dummy CQ */ send_cq 4479 drivers/infiniband/hw/mlx4/qp.c struct ib_cq *cq = qp->send_cq; send_cq 75 drivers/infiniband/hw/mlx5/gsi.c struct ib_cq *gsi_cq = gsi->ibqp.send_cq; send_cq 181 drivers/infiniband/hw/mlx5/gsi.c hw_init_attr.send_cq = gsi->cq; send_cq 257 drivers/infiniband/hw/mlx5/gsi.c .send_cq = gsi->cq, send_cq 4484 drivers/infiniband/hw/mlx5/main.c send_mcq = to_mcq(mqp->ibqp.send_cq); send_cq 4487 drivers/infiniband/hw/mlx5/main.c mqp->ibqp.send_cq->comp_handler) { send_cq 4834 drivers/infiniband/hw/mlx5/main.c init_attr->send_cq = cq; send_cq 4851 drivers/infiniband/hw/mlx5/main.c qp->send_cq = init_attr->send_cq; send_cq 101 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq); send_cq 711 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, send_cq 713 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq, send_cq 1612 drivers/infiniband/hw/mlx5/qp.c if (init_attr->create_flags || init_attr->send_cq) send_cq 1882 drivers/infiniband/hw/mlx5/qp.c scqe_sz = mlx5_ib_get_cqe_size(init_attr->send_cq); send_cq 1955 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_cq *send_cq; send_cq 2243 drivers/infiniband/hw/mlx5/qp.c if (init_attr->send_cq) send_cq 2244 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(qpc, qpc, cqn_snd, to_mcq(init_attr->send_cq)->mcq.cqn); send_cq 2300 drivers/infiniband/hw/mlx5/qp.c get_cqs(init_attr->qp_type, init_attr->send_cq, init_attr->recv_cq, send_cq 2301 drivers/infiniband/hw/mlx5/qp.c &send_cq, &recv_cq); send_cq 2303 drivers/infiniband/hw/mlx5/qp.c mlx5_ib_lock_cqs(send_cq, recv_cq); send_cq 2310 drivers/infiniband/hw/mlx5/qp.c if (send_cq) send_cq 2311 drivers/infiniband/hw/mlx5/qp.c list_add_tail(&qp->cq_send_list, &send_cq->list_send_qp); send_cq 2314 drivers/infiniband/hw/mlx5/qp.c mlx5_ib_unlock_cqs(send_cq, recv_cq); send_cq 2330 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq) send_cq 2331 drivers/infiniband/hw/mlx5/qp.c __acquires(&send_cq->lock) __acquires(&recv_cq->lock) send_cq 2333 drivers/infiniband/hw/mlx5/qp.c if (send_cq) { send_cq 2335 drivers/infiniband/hw/mlx5/qp.c if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { send_cq 2336 drivers/infiniband/hw/mlx5/qp.c spin_lock(&send_cq->lock); send_cq 2339 drivers/infiniband/hw/mlx5/qp.c } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { send_cq 2340 drivers/infiniband/hw/mlx5/qp.c spin_lock(&send_cq->lock); send_cq 2344 drivers/infiniband/hw/mlx5/qp.c spin_lock_nested(&send_cq->lock, send_cq 2348 drivers/infiniband/hw/mlx5/qp.c spin_lock(&send_cq->lock); send_cq 2353 drivers/infiniband/hw/mlx5/qp.c __acquire(&send_cq->lock); send_cq 2355 drivers/infiniband/hw/mlx5/qp.c __acquire(&send_cq->lock); send_cq 2360 drivers/infiniband/hw/mlx5/qp.c static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq) send_cq 2361 drivers/infiniband/hw/mlx5/qp.c __releases(&send_cq->lock) __releases(&recv_cq->lock) send_cq 2363 drivers/infiniband/hw/mlx5/qp.c if (send_cq) { send_cq 2365 drivers/infiniband/hw/mlx5/qp.c if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { send_cq 2367 drivers/infiniband/hw/mlx5/qp.c spin_unlock(&send_cq->lock); send_cq 2368 drivers/infiniband/hw/mlx5/qp.c } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { send_cq 2370 drivers/infiniband/hw/mlx5/qp.c spin_unlock(&send_cq->lock); send_cq 2372 drivers/infiniband/hw/mlx5/qp.c spin_unlock(&send_cq->lock); send_cq 2377 drivers/infiniband/hw/mlx5/qp.c spin_unlock(&send_cq->lock); send_cq 2380 drivers/infiniband/hw/mlx5/qp.c __release(&send_cq->lock); send_cq 2384 drivers/infiniband/hw/mlx5/qp.c __release(&send_cq->lock); send_cq 2395 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq) send_cq 2399 drivers/infiniband/hw/mlx5/qp.c *send_cq = NULL; send_cq 2404 drivers/infiniband/hw/mlx5/qp.c *send_cq = ib_send_cq ? to_mcq(ib_send_cq) : NULL; send_cq 2416 drivers/infiniband/hw/mlx5/qp.c *send_cq = ib_send_cq ? to_mcq(ib_send_cq) : NULL; send_cq 2422 drivers/infiniband/hw/mlx5/qp.c *send_cq = NULL; send_cq 2435 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_cq *send_cq, *recv_cq; send_cq 2468 drivers/infiniband/hw/mlx5/qp.c get_cqs(qp->ibqp.qp_type, qp->ibqp.send_cq, qp->ibqp.recv_cq, send_cq 2469 drivers/infiniband/hw/mlx5/qp.c &send_cq, &recv_cq); send_cq 2472 drivers/infiniband/hw/mlx5/qp.c mlx5_ib_lock_cqs(send_cq, recv_cq); send_cq 2475 drivers/infiniband/hw/mlx5/qp.c if (send_cq) send_cq 2484 drivers/infiniband/hw/mlx5/qp.c if (send_cq != recv_cq) send_cq 2485 drivers/infiniband/hw/mlx5/qp.c __mlx5_ib_cq_clean(send_cq, base->mqp.qpn, send_cq 2488 drivers/infiniband/hw/mlx5/qp.c mlx5_ib_unlock_cqs(send_cq, recv_cq); send_cq 2693 drivers/infiniband/hw/mlx5/qp.c init_attr->send_cq = NULL; send_cq 2726 drivers/infiniband/hw/mlx5/qp.c init_attr->send_cq ? to_mcq(init_attr->send_cq)->mcq.cqn : -1); send_cq 3455 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_cq *send_cq, *recv_cq; send_cq 3563 drivers/infiniband/hw/mlx5/qp.c get_cqs(qp->ibqp.qp_type, qp->ibqp.send_cq, qp->ibqp.recv_cq, send_cq 3564 drivers/infiniband/hw/mlx5/qp.c &send_cq, &recv_cq); send_cq 3567 drivers/infiniband/hw/mlx5/qp.c context->cqn_send = send_cq ? cpu_to_be32(send_cq->mcq.cqn) : 0; send_cq 3718 drivers/infiniband/hw/mlx5/qp.c if (send_cq != recv_cq) send_cq 3719 drivers/infiniband/hw/mlx5/qp.c mlx5_ib_cq_clean(send_cq, base->mqp.qpn, NULL); send_cq 4902 drivers/infiniband/hw/mlx5/qp.c if (unlikely(mlx5_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq))) send_cq 5811 drivers/infiniband/hw/mlx5/qp.c qp_init_attr->send_cq = ibqp->send_cq; send_cq 6441 drivers/infiniband/hw/mlx5/qp.c struct ib_cq *cq = qp->send_cq; send_cq 546 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_cq *send_cq, send_cq 555 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_cq *send_cq, send_cq 522 drivers/infiniband/hw/mthca/mthca_provider.c to_mcq(init_attr->send_cq), send_cq 555 drivers/infiniband/hw/mthca/mthca_provider.c to_mcq(init_attr->send_cq), send_cq 735 drivers/infiniband/hw/mthca/mthca_qp.c qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); send_cq 837 drivers/infiniband/hw/mthca/mthca_qp.c if (qp->ibqp.send_cq != qp->ibqp.recv_cq) send_cq 838 drivers/infiniband/hw/mthca/mthca_qp.c mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); send_cq 1161 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_cq *send_cq, send_cq 1290 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_cq *send_cq, send_cq 1318 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, send_cq 1333 drivers/infiniband/hw/mthca/mthca_qp.c static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) send_cq 1334 drivers/infiniband/hw/mthca/mthca_qp.c __acquires(&send_cq->lock) __acquires(&recv_cq->lock) send_cq 1336 drivers/infiniband/hw/mthca/mthca_qp.c if (send_cq == recv_cq) { send_cq 1337 drivers/infiniband/hw/mthca/mthca_qp.c spin_lock_irq(&send_cq->lock); send_cq 1339 drivers/infiniband/hw/mthca/mthca_qp.c } else if (send_cq->cqn < recv_cq->cqn) { send_cq 1340 drivers/infiniband/hw/mthca/mthca_qp.c spin_lock_irq(&send_cq->lock); send_cq 1344 drivers/infiniband/hw/mthca/mthca_qp.c spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); send_cq 1348 drivers/infiniband/hw/mthca/mthca_qp.c static void mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) send_cq 1349 drivers/infiniband/hw/mthca/mthca_qp.c __releases(&send_cq->lock) __releases(&recv_cq->lock) send_cq 1351 drivers/infiniband/hw/mthca/mthca_qp.c if (send_cq == recv_cq) { send_cq 1353 drivers/infiniband/hw/mthca/mthca_qp.c spin_unlock_irq(&send_cq->lock); send_cq 1354 drivers/infiniband/hw/mthca/mthca_qp.c } else if (send_cq->cqn < recv_cq->cqn) { send_cq 1356 drivers/infiniband/hw/mthca/mthca_qp.c spin_unlock_irq(&send_cq->lock); send_cq 1358 drivers/infiniband/hw/mthca/mthca_qp.c spin_unlock(&send_cq->lock); send_cq 1365 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_cq *send_cq, send_cq 1402 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, send_cq 1416 drivers/infiniband/hw/mthca/mthca_qp.c mthca_lock_cqs(send_cq, recv_cq); send_cq 1422 drivers/infiniband/hw/mthca/mthca_qp.c mthca_unlock_cqs(send_cq, recv_cq); send_cq 1445 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_cq *send_cq; send_cq 1448 drivers/infiniband/hw/mthca/mthca_qp.c send_cq = to_mcq(qp->ibqp.send_cq); send_cq 1455 drivers/infiniband/hw/mthca/mthca_qp.c mthca_lock_cqs(send_cq, recv_cq); send_cq 1463 drivers/infiniband/hw/mthca/mthca_qp.c mthca_unlock_cqs(send_cq, recv_cq); send_cq 1479 drivers/infiniband/hw/mthca/mthca_qp.c if (send_cq != recv_cq) send_cq 1480 drivers/infiniband/hw/mthca/mthca_qp.c mthca_cq_clean(dev, send_cq, qp->qpn, NULL); send_cq 1654 drivers/infiniband/hw/mthca/mthca_qp.c if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { send_cq 1984 drivers/infiniband/hw/mthca/mthca_qp.c if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { send_cq 2431 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cq = get_ocrdma_cq(attrs->send_cq); send_cq 1174 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if ((dev->gsi_sqcq == get_ocrdma_cq(attrs->send_cq)) || send_cq 1301 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c dev->gsi_sqcq = get_ocrdma_cq(attrs->send_cq); send_cq 62 drivers/infiniband/hw/qedr/qedr_roce_cm.c dev->gsi_sqcq = get_qedr_cq(attrs->send_cq); send_cq 139 drivers/infiniband/hw/qedr/qedr_roce_cm.c cq = get_qedr_cq(attrs->send_cq); send_cq 1207 drivers/infiniband/hw/qedr/verbs.c qp->sq_cq = get_qedr_cq(attrs->send_cq); send_cq 1529 drivers/infiniband/hw/qedr/verbs.c params->sq_cq_id = get_qedr_cq(attrs->send_cq)->icid; send_cq 1904 drivers/infiniband/hw/qedr/verbs.c get_qedr_cq(attrs->send_cq), send_cq 1905 drivers/infiniband/hw/qedr/verbs.c get_qedr_cq(attrs->send_cq)->icid, send_cq 512 drivers/infiniband/hw/usnic/usnic_ib_verbs.c cq_cnt = (init_attr->send_cq == init_attr->recv_cq) ? 1 : 2; send_cq 55 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq, send_cq 58 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *send_cq = to_vcq(qp->ibqp.send_cq); send_cq 350 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle; send_cq 989 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c init_attr->send_cq = qp->ibqp.send_cq; send_cq 1798 drivers/infiniband/sw/rdmavt/qp.c init_attr->send_cq = qp->ibqp.send_cq; send_cq 92 drivers/infiniband/sw/rxe/rxe_qp.c if (!init->recv_cq || !init->send_cq) { send_cq 344 drivers/infiniband/sw/rxe/rxe_qp.c struct rxe_cq *scq = to_rcq(init->send_cq); send_cq 390 drivers/infiniband/sw/rxe/rxe_qp.c init->send_cq = qp->ibqp.send_cq; send_cq 356 drivers/infiniband/sw/siw/siw_verbs.c scq = to_siw_cq(attrs->send_cq); send_cq 549 drivers/infiniband/sw/siw/siw_verbs.c qp_init_attr->send_cq = base_qp->send_cq; send_cq 366 drivers/infiniband/ulp/ipoib/ipoib.h struct ib_cq *send_cq; send_cq 255 drivers/infiniband/ulp/ipoib/ipoib_cm.c .send_cq = priv->recv_cq, /* For drain WR */ send_cq 770 drivers/infiniband/ulp/ipoib/ipoib_cm.c rc = ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | send_cq 1061 drivers/infiniband/ulp/ipoib/ipoib_cm.c .send_cq = priv->send_cq, send_cq 439 drivers/infiniband/ulp/ipoib/ipoib_ib.c n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); send_cq 507 drivers/infiniband/ulp/ipoib/ipoib_ib.c n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); send_cq 519 drivers/infiniband/ulp/ipoib/ipoib_ib.c if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | send_cq 649 drivers/infiniband/ulp/ipoib/ipoib_ib.c if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | send_cq 187 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL, send_cq 189 drivers/infiniband/ulp/ipoib/ipoib_verbs.c if (IS_ERR(priv->send_cq)) { send_cq 197 drivers/infiniband/ulp/ipoib/ipoib_verbs.c init_attr.send_cq = priv->send_cq; send_cq 215 drivers/infiniband/ulp/ipoib/ipoib_verbs.c if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP)) send_cq 241 drivers/infiniband/ulp/ipoib/ipoib_verbs.c ib_destroy_cq(priv->send_cq); send_cq 263 drivers/infiniband/ulp/ipoib/ipoib_verbs.c ib_destroy_cq(priv->send_cq); send_cq 393 drivers/infiniband/ulp/iser/iser_verbs.c init_attr.send_cq = ib_conn->comp->cq; send_cq 126 drivers/infiniband/ulp/isert/ib_isert.c attr.send_cq = comp->cq; send_cq 544 drivers/infiniband/ulp/srp/ib_srp.c ib_process_cq_direct(ch->send_cq, -1); send_cq 556 drivers/infiniband/ulp/srp/ib_srp.c struct ib_cq *recv_cq, *send_cq; send_cq 575 drivers/infiniband/ulp/srp/ib_srp.c send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size, send_cq 577 drivers/infiniband/ulp/srp/ib_srp.c if (IS_ERR(send_cq)) { send_cq 578 drivers/infiniband/ulp/srp/ib_srp.c ret = PTR_ERR(send_cq); send_cq 589 drivers/infiniband/ulp/srp/ib_srp.c init_attr->send_cq = send_cq; send_cq 633 drivers/infiniband/ulp/srp/ib_srp.c if (ch->send_cq) send_cq 634 drivers/infiniband/ulp/srp/ib_srp.c ib_free_cq(ch->send_cq); send_cq 638 drivers/infiniband/ulp/srp/ib_srp.c ch->send_cq = send_cq; send_cq 660 drivers/infiniband/ulp/srp/ib_srp.c ib_free_cq(send_cq); send_cq 708 drivers/infiniband/ulp/srp/ib_srp.c ib_free_cq(ch->send_cq); send_cq 720 drivers/infiniband/ulp/srp/ib_srp.c ch->send_cq = ch->recv_cq = NULL; send_cq 2025 drivers/infiniband/ulp/srp/ib_srp.c ib_process_cq_direct(ch->send_cq, -1); send_cq 155 drivers/infiniband/ulp/srp/ib_srp.h struct ib_cq *send_cq; send_cq 1806 drivers/infiniband/ulp/srpt/ib_srpt.c qp_init->send_cq = ch->cq; send_cq 350 drivers/net/ethernet/ibm/ehea/ehea.h struct ehea_cq *send_cq; send_cq 195 drivers/net/ethernet/ibm/ehea/ehea_main.c arr[i++].fwh = pr->send_cq->fw_handle; send_cq 800 drivers/net/ethernet/ibm/ehea/ehea_main.c struct ehea_cq *send_cq = pr->send_cq; send_cq 809 drivers/net/ethernet/ibm/ehea/ehea_main.c cqe = ehea_poll_cq(send_cq); send_cq 811 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_inc_cq(send_cq); send_cq 851 drivers/net/ethernet/ibm/ehea/ehea_main.c cqe = ehea_poll_cq(send_cq); send_cq 854 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_update_feca(send_cq, cqe_counter); send_cq 889 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_reset_cq_ep(pr->send_cq); send_cq 891 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_reset_cq_n1(pr->send_cq); send_cq 894 drivers/net/ethernet/ibm/ehea/ehea_main.c cqe_skb = ehea_poll_cq(pr->send_cq); send_cq 1473 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq, send_cq 1476 drivers/net/ethernet/ibm/ehea/ehea_main.c if (!pr->send_cq) { send_cq 1483 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->send_cq->attr.act_nr_of_cqes, send_cq 1508 drivers/net/ethernet/ibm/ehea/ehea_main.c init_attr->send_cq_handle = pr->send_cq->fw_handle; send_cq 1558 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_cq(pr->send_cq); send_cq 1575 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_cq(pr->send_cq); send_cq 270 drivers/nvme/host/rdma.c init_attr.send_cq = queue->ib_cq; send_cq 1005 drivers/nvme/target/rdma.c qp_attr.send_cq = queue->cq; send_cq 1464 fs/cifs/smbdirect.c ib_free_cq(info->send_cq); send_cq 1657 fs/cifs/smbdirect.c info->send_cq = NULL; send_cq 1659 fs/cifs/smbdirect.c info->send_cq = send_cq 1662 fs/cifs/smbdirect.c if (IS_ERR(info->send_cq)) { send_cq 1663 fs/cifs/smbdirect.c info->send_cq = NULL; send_cq 1685 fs/cifs/smbdirect.c qp_attr.send_cq = info->send_cq; send_cq 1804 fs/cifs/smbdirect.c if (info->send_cq) send_cq 1805 fs/cifs/smbdirect.c ib_free_cq(info->send_cq); send_cq 59 fs/cifs/smbdirect.h struct ib_cq *send_cq, *recv_cq; send_cq 1127 include/rdma/ib_verbs.h struct ib_cq *send_cq; send_cq 1714 include/rdma/ib_verbs.h struct ib_cq *send_cq; send_cq 821 include/rdma/rdmavt_qp.h struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); send_cq 936 include/rdma/rdmavt_qp.h static inline u32 ib_cq_tail(struct ib_cq *send_cq) send_cq 938 include/rdma/rdmavt_qp.h struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); send_cq 940 include/rdma/rdmavt_qp.h return ibcq_to_rvtcq(send_cq)->ip ? send_cq 942 include/rdma/rdmavt_qp.h ibcq_to_rvtcq(send_cq)->kqueue->tail; send_cq 952 include/rdma/rdmavt_qp.h static inline u32 ib_cq_head(struct ib_cq *send_cq) send_cq 954 include/rdma/rdmavt_qp.h struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); send_cq 956 include/rdma/rdmavt_qp.h return ibcq_to_rvtcq(send_cq)->ip ? send_cq 958 include/rdma/rdmavt_qp.h ibcq_to_rvtcq(send_cq)->kqueue->head; send_cq 709 net/9p/trans_rdma.c qp_attr.send_cq = rdma->cq; send_cq 537 net/rds/ib_cm.c attr.send_cq = ic->i_send_cq; send_cq 331 net/smc/smc_ib.c .send_cq = lnk->smcibdev->roce_cq_send, send_cq 479 net/sunrpc/xprtrdma/svc_rdma_transport.c qp_attr.send_cq = newxprt->sc_sq_cq; send_cq 426 net/sunrpc/xprtrdma/verbs.c ib_free_cq(ep->rep_attr.send_cq); send_cq 427 net/sunrpc/xprtrdma/verbs.c ep->rep_attr.send_cq = NULL; send_cq 539 net/sunrpc/xprtrdma/verbs.c ep->rep_attr.send_cq = sendcq; send_cq 598 net/sunrpc/xprtrdma/verbs.c if (ep->rep_attr.send_cq) send_cq 599 net/sunrpc/xprtrdma/verbs.c ib_free_cq(ep->rep_attr.send_cq);