cur_qp 2186 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_qp **cur_qp, struct ib_wc *wc) cur_qp 2224 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!*cur_qp || (qpn & HNS_ROCE_CQE_QPN_MASK) != (*cur_qp)->qpn) { cur_qp 2232 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *cur_qp = hr_qp; cur_qp 2235 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wc->qp = &(*cur_qp)->ibqp; cur_qp 2293 drivers/infiniband/hw/hns/hns_roce_hw_v1.c sq_wqe = get_send_wqe(*cur_qp, roce_get_field(cqe->cqe_byte_4, cur_qp 2296 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ((*cur_qp)->sq.wqe_cnt-1)); cur_qp 2321 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wq = &(*cur_qp)->sq; cur_qp 2322 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if ((*cur_qp)->sq_signal_bits) { cur_qp 2368 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wq = &(*cur_qp)->rq; cur_qp 2390 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_qp *cur_qp = NULL; cur_qp 2398 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ret = hns_roce_v1_poll_one(hr_cq, &cur_qp, wc + npolled); cur_qp 2661 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_qp **cur_qp, cur_qp 2671 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr_cnt = wr_num & ((*cur_qp)->rq.wqe_cnt - 1); cur_qp 2673 drivers/infiniband/hw/hns/hns_roce_hw_v2.c sge_list = (*cur_qp)->rq_inl_buf.wqe_list[wr_cnt].sg_list; cur_qp 2674 drivers/infiniband/hw/hns/hns_roce_hw_v2.c sge_num = (*cur_qp)->rq_inl_buf.wqe_list[wr_cnt].sge_cnt; cur_qp 2675 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wqe_buf = get_recv_wqe(*cur_qp, wr_cnt); cur_qp 2695 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_qp **cur_qp, struct ib_wc *wc) cur_qp 2726 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!*cur_qp || (qpn & HNS_ROCE_V2_CQE_QPN_MASK) != (*cur_qp)->qpn) { cur_qp 2734 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *cur_qp = hr_qp; cur_qp 2737 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wc->qp = &(*cur_qp)->ibqp; cur_qp 2741 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wq = &(*cur_qp)->sq; cur_qp 2742 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if ((*cur_qp)->sq_signal_bits) { cur_qp 2757 drivers/infiniband/hw/hns/hns_roce_hw_v2.c } else if ((*cur_qp)->ibqp.srq) { cur_qp 2758 drivers/infiniband/hw/hns/hns_roce_hw_v2.c srq = to_hr_srq((*cur_qp)->ibqp.srq); cur_qp 2766 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wq = &(*cur_qp)->rq; cur_qp 2826 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return hns_roce_v2_modify_qp(&(*cur_qp)->ibqp, cur_qp 2828 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (*cur_qp)->state, IB_QPS_ERR); cur_qp 2929 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = hns_roce_handle_recv_inl_wqe(cqe, cur_qp, wc); cur_qp 2975 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_qp *cur_qp = NULL; cur_qp 2982 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (hns_roce_v2_poll_one(hr_cq, &cur_qp, wc + npolled)) cur_qp 662 drivers/infiniband/hw/mlx4/cq.c struct mlx4_ib_qp **cur_qp, cur_qp 713 drivers/infiniband/hw/mlx4/cq.c if (!*cur_qp || cur_qp 714 drivers/infiniband/hw/mlx4/cq.c (be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) { cur_qp 722 drivers/infiniband/hw/mlx4/cq.c *cur_qp = to_mibqp(mqp); cur_qp 725 drivers/infiniband/hw/mlx4/cq.c wc->qp = &(*cur_qp)->ibqp; cur_qp 737 drivers/infiniband/hw/mlx4/cq.c wq = &(*cur_qp)->sq; cur_qp 738 drivers/infiniband/hw/mlx4/cq.c if (!(*cur_qp)->sq_signal_bits) { cur_qp 744 drivers/infiniband/hw/mlx4/cq.c } else if ((*cur_qp)->ibqp.srq) { cur_qp 745 drivers/infiniband/hw/mlx4/cq.c srq = to_msrq((*cur_qp)->ibqp.srq); cur_qp 755 drivers/infiniband/hw/mlx4/cq.c wq = &(*cur_qp)->rq; cur_qp 840 drivers/infiniband/hw/mlx4/cq.c (*cur_qp)->port) == cur_qp 843 drivers/infiniband/hw/mlx4/cq.c if ((*cur_qp)->mlx4_ib_qp_type & cur_qp 846 drivers/infiniband/hw/mlx4/cq.c use_tunnel_data(*cur_qp, cq, wc, tail, cqe, cur_qp 884 drivers/infiniband/hw/mlx4/cq.c struct mlx4_ib_qp *cur_qp = NULL; cur_qp 896 drivers/infiniband/hw/mlx4/cq.c if (mlx4_ib_poll_one(cq, &cur_qp, wc + npolled)) cur_qp 440 drivers/infiniband/hw/mlx5/cq.c struct mlx5_ib_qp **cur_qp, cur_qp 485 drivers/infiniband/hw/mlx5/cq.c if (!*cur_qp || (qpn != (*cur_qp)->ibqp.qp_num)) { cur_qp 491 drivers/infiniband/hw/mlx5/cq.c *cur_qp = to_mibqp(mqp); cur_qp 494 drivers/infiniband/hw/mlx5/cq.c wc->qp = &(*cur_qp)->ibqp; cur_qp 497 drivers/infiniband/hw/mlx5/cq.c wq = &(*cur_qp)->sq; cur_qp 501 drivers/infiniband/hw/mlx5/cq.c handle_atomics(*cur_qp, cqe64, wq->last_poll, idx); cur_qp 510 drivers/infiniband/hw/mlx5/cq.c handle_responder(wc, cqe64, *cur_qp); cur_qp 525 drivers/infiniband/hw/mlx5/cq.c wq = &(*cur_qp)->sq; cur_qp 533 drivers/infiniband/hw/mlx5/cq.c if ((*cur_qp)->ibqp.srq) { cur_qp 534 drivers/infiniband/hw/mlx5/cq.c srq = to_msrq((*cur_qp)->ibqp.srq); cur_qp 539 drivers/infiniband/hw/mlx5/cq.c wq = &(*cur_qp)->rq; cur_qp 599 drivers/infiniband/hw/mlx5/cq.c struct mlx5_ib_qp *cur_qp = NULL; cur_qp 621 drivers/infiniband/hw/mlx5/cq.c if (mlx5_poll_one(cq, &cur_qp, wc + soft_polled + npolled)) cur_qp 480 drivers/infiniband/hw/mthca/mthca_cq.c struct mthca_qp **cur_qp, cur_qp 514 drivers/infiniband/hw/mthca/mthca_cq.c if (!*cur_qp || be32_to_cpu(cqe->my_qpn) != (*cur_qp)->qpn) { cur_qp 520 drivers/infiniband/hw/mthca/mthca_cq.c *cur_qp = mthca_array_get(&dev->qp_table.qp, cur_qp 523 drivers/infiniband/hw/mthca/mthca_cq.c if (!*cur_qp) { cur_qp 531 drivers/infiniband/hw/mthca/mthca_cq.c entry->qp = &(*cur_qp)->ibqp; cur_qp 534 drivers/infiniband/hw/mthca/mthca_cq.c wq = &(*cur_qp)->sq; cur_qp 535 drivers/infiniband/hw/mthca/mthca_cq.c wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset) cur_qp 537 drivers/infiniband/hw/mthca/mthca_cq.c entry->wr_id = (*cur_qp)->wrid[wqe_index + cur_qp 538 drivers/infiniband/hw/mthca/mthca_cq.c (*cur_qp)->rq.max]; cur_qp 539 drivers/infiniband/hw/mthca/mthca_cq.c } else if ((*cur_qp)->ibqp.srq) { cur_qp 540 drivers/infiniband/hw/mthca/mthca_cq.c struct mthca_srq *srq = to_msrq((*cur_qp)->ibqp.srq); cur_qp 548 drivers/infiniband/hw/mthca/mthca_cq.c wq = &(*cur_qp)->rq; cur_qp 558 drivers/infiniband/hw/mthca/mthca_cq.c entry->wr_id = (*cur_qp)->wrid[wqe_index]; cur_qp 571 drivers/infiniband/hw/mthca/mthca_cq.c handle_error_cqe(dev, cq, *cur_qp, wqe_index, is_send, cur_qp 318 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c static int pvrdma_poll_one(struct pvrdma_cq *cq, struct pvrdma_qp **cur_qp, cur_qp 348 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c *cur_qp = (struct pvrdma_qp *)dev->qp_tbl[cqe->qp & 0xffff]; cur_qp 355 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c wc->qp = &(*cur_qp)->ibqp; cur_qp 385 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c struct pvrdma_qp *cur_qp = NULL; cur_qp 394 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c if (pvrdma_poll_one(cq, &cur_qp, wc + npolled))