/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 378 struct mthca_qp *qp, int wqe_index, int is_send, in handle_error_cqe() argument 466 mthca_free_err_wqe(dev, qp, is_send, wqe_index, &dbd, &new_wqe); in handle_error_cqe() 491 int wqe_index; in mthca_poll_one() local 540 wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset) in mthca_poll_one() 542 entry->wr_id = (*cur_qp)->wrid[wqe_index + in mthca_poll_one() 548 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one() 549 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one() 555 wqe_index = wqe >> wq->wqe_shift; in mthca_poll_one() 561 if (unlikely(wqe_index < 0)) in mthca_poll_one() 562 wqe_index = wq->max - 1; in mthca_poll_one() [all …]
|
/linux-4.1.27/include/linux/mlx4/ |
D | cq.h | 57 __be16 wqe_index; member 66 __be16 wqe_index; member 82 __be16 wqe_index; member
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
D | odp.c | 396 u16 wqe_index = pfault->mpfault.wqe.wqe_index; in mlx5_ib_mr_initiator_pfault_handler() local 411 wqe_index, qp->mqp.qpn); in mlx5_ib_mr_initiator_pfault_handler() 419 if (wqe_index != ctrl_wqe_index) { in mlx5_ib_mr_initiator_pfault_handler() 421 wqe_index, qp->mqp.qpn, in mlx5_ib_mr_initiator_pfault_handler() 430 wqe_index, qp->mqp.qpn, in mlx5_ib_mr_initiator_pfault_handler() 545 u16 wqe_index = pfault->mpfault.wqe.wqe_index; in mlx5_ib_mr_wqe_pfault_handler() local 555 ret = mlx5_ib_read_user_wqe(qp, requestor, wqe_index, buffer, in mlx5_ib_mr_wqe_pfault_handler() 559 -ret, wqe_index, qp->mqp.qpn); in mlx5_ib_mr_wqe_pfault_handler()
|
D | srq.c | 412 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index) in mlx5_ib_free_srq_wqe() argument 420 next->next_wqe_index = cpu_to_be16(wqe_index); in mlx5_ib_free_srq_wqe() 421 srq->tail = wqe_index; in mlx5_ib_free_srq_wqe()
|
D | mlx5_ib.h | 526 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index); 557 int mlx5_ib_read_user_wqe(struct mlx5_ib_qp *qp, int send, int wqe_index,
|
D | qp.c | 121 int mlx5_ib_read_user_wqe(struct mlx5_ib_qp *qp, int send, int wqe_index, in mlx5_ib_read_user_wqe() argument 140 offset = wq->offset + ((wqe_index % wq->wqe_cnt) << wq->wqe_shift); in mlx5_ib_read_user_wqe()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.h | 322 int *wqe_index) in ehea_get_swqe() argument 327 *wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_SQ); in ehea_get_swqe() 339 static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index) in ehea_poll_rq1() argument 343 *wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_RQ1); in ehea_poll_rq1()
|
D | ehea_main.c | 604 int arr_len, int wqe_index) in get_skb_by_index_ll() argument 610 x = wqe_index + 1; in get_skb_by_index_ll() 623 skb = skb_array[wqe_index]; in get_skb_by_index_ll() 624 skb_array[wqe_index] = NULL; in get_skb_by_index_ll() 680 int wqe_index, last_wqe_index, rq, port_reset; in ehea_proc_rwqes() local 685 cqe = ehea_poll_rq1(qp, &wqe_index); in ehea_proc_rwqes() 693 last_wqe_index = wqe_index; in ehea_proc_rwqes() 700 wqe_index); in ehea_proc_rwqes() 752 cqe = ehea_poll_rq1(qp, &wqe_index); in ehea_proc_rwqes() 896 int wqe_index; in ehea_poll() local [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | srq.c | 295 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index) in mlx4_ib_free_srq_wqe() argument 303 next->next_wqe_index = cpu_to_be16(wqe_index); in mlx4_ib_free_srq_wqe() 304 srq->tail = wqe_index; in mlx4_ib_free_srq_wqe()
|
D | cq.c | 502 be32_to_cpu(cqe->my_qpn), be16_to_cpu(cqe->wqe_index), in mlx4_ib_handle_error_cqe() 742 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 749 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 754 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 954 mlx4_ib_free_srq_wqe(srq, be16_to_cpu(cqe->wqe_index)); in __mlx4_ib_cq_clean()
|
D | mlx4_ib.h | 691 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index);
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx5/core/ |
D | qp.c | 144 pfault.wqe.wqe_index = in mlx5_eq_pagefault() 145 be16_to_cpu(pf_eqe->wqe.wqe_index); in mlx5_eq_pagefault() 150 qpn, pfault.wqe.wqe_index); in mlx5_eq_pagefault()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
D | nes_verbs.c | 3646 u32 wqe_index; in nes_poll_cq() local 3671 wqe_index = u32temp & (nesdev->nesadapter->max_qp_wr - 1); in nes_poll_cq() 3705 wrid = (((u64)(cpu_to_le32((u32)nesqp->hwqp.sq_vbase[wqe_index]. in nes_poll_cq() 3707 ((u64)(cpu_to_le32((u32)nesqp->hwqp.sq_vbase[wqe_index]. in nes_poll_cq() 3709 entry->byte_len = le32_to_cpu(nesqp->hwqp.sq_vbase[wqe_index]. in nes_poll_cq() 3712 switch (le32_to_cpu(nesqp->hwqp.sq_vbase[wqe_index]. in nes_poll_cq() 3721 entry->byte_len = le32_to_cpu(nesqp->hwqp.sq_vbase[wqe_index]. in nes_poll_cq() 3739 nesqp->hwqp.sq_tail = (wqe_index+1)&(nesqp->hwqp.sq_size - 1); in nes_poll_cq() 3747 …wrid = ((u64)(le32_to_cpu(nesqp->hwqp.rq_vbase[wqe_index].wqe_words[NES_IWARP_RQ_WQE_COMP_SCRATCH_… in nes_poll_cq() 3748 …((u64)(le32_to_cpu(nesqp->hwqp.rq_vbase[wqe_index].wqe_words[NES_IWARP_RQ_WQE_COMP_SCRATCH_HIGH_ID… in nes_poll_cq() [all …]
|
/linux-4.1.27/include/linux/mlx5/ |
D | qp.h | 396 u16 wqe_index; member
|
D | device.h | 491 __be16 wqe_index; member
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | en_tx.c | 440 new_index = be16_to_cpu(cqe->wqe_index) & size_mask; in mlx4_en_process_tx_cq()
|
D | en_netdev.c | 1626 cq->buf->wqe_index = cpu_to_be16(0xffff); in mlx4_en_start_port()
|