/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
D | cq.c | 100 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn) in mlx4_cq_completion() argument 105 cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_completion() 107 mlx4_dbg(dev, "Completion event for bogus CQ %08x\n", cqn); in mlx4_cq_completion() 116 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type) in mlx4_cq_event() argument 123 cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_event() 130 mlx4_warn(dev, "Async event for bogus CQ %08x\n", cqn); in mlx4_cq_event() 178 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 1); in mlx4_cq_modify() 204 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 0); in mlx4_cq_resize() 211 int __mlx4_cq_alloc_icm(struct mlx4_dev *dev, int *cqn) in __mlx4_cq_alloc_icm() argument 217 *cqn = mlx4_bitmap_alloc(&cq_table->bitmap); in __mlx4_cq_alloc_icm() [all …]
|
D | en_resources.c | 41 int is_tx, int rss, int qpn, int cqn, in mlx4_en_fill_qp_context() argument 70 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context() 71 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
|
D | resource_tracker.c | 1593 static int cq_res_start_move_to(struct mlx4_dev *dev, int slave, int cqn, in cq_res_start_move_to() argument 1602 r = res_tracker_lookup(&tracker->res_tree[RES_CQ], cqn); in cq_res_start_move_to() 1860 int cqn; in cq_alloc_res() local 1869 err = __mlx4_cq_alloc_icm(dev, &cqn); in cq_alloc_res() 1875 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_alloc_res() 1878 __mlx4_cq_free_icm(dev, cqn); in cq_alloc_res() 1882 set_param_l(out_param, cqn); in cq_alloc_res() 2387 int cqn; in cq_free_res() local 2392 cqn = get_param_l(&in_param); in cq_free_res() 2393 err = rem_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_free_res() [all …]
|
D | srq.c | 165 int mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd, in mlx4_srq_alloc() argument 195 srq_context->pg_offset_cqn = cpu_to_be32(cqn & 0xffffff); in mlx4_srq_alloc()
|
D | eq.c | 485 int cqn = -1; in mlx4_eq_int() local 508 cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mlx4_eq_int() 509 mlx4_cq_completion(dev, cqn); in mlx4_eq_int() 665 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); in mlx4_eq_int() 669 be32_to_cpu(eqe->event.cq_err.cqn) in mlx4_eq_int() 684 be32_to_cpu(eqe->event.cq_err.cqn) in mlx4_eq_int() 826 if (!(cqn & ~0xffffff)) in mlx4_eq_int()
|
D | mlx4_en.h | 284 u16 cqn; /* index of port CQ associated with this ring */ member 314 u16 cqn; /* index of port CQ associated with this ring */ member 796 int is_tx, int rss, int qpn, int cqn, int user_prio,
|
D | en_rx.c | 439 ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn; in mlx4_en_activate_rx_rings() 1144 qpn, ring->cqn, -1, context); in mlx4_en_config_rss_qp() 1242 priv->rx_ring[0]->cqn, -1, &context); in mlx4_en_config_rss_steer()
|
D | mlx4.h | 956 int __mlx4_cq_alloc_icm(struct mlx4_dev *dev, int *cqn); 957 void __mlx4_cq_free_icm(struct mlx4_dev *dev, int cqn); 1200 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn); 1201 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type);
|
D | en_tx.c | 202 ring->cqn = cq; in mlx4_en_activate_tx_ring() 214 ring->cqn, user_prio, &ring->context); in mlx4_en_activate_tx_ring()
|
D | en_netdev.c | 1307 i, priv->tx_ring[i]->qpn, priv->tx_ring[i]->cqn, in mlx4_en_tx_timeout() 1595 priv->rx_ring[i]->cqn = cq->mcq.cqn; in mlx4_en_start_port() 1641 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, in mlx4_en_start_port()
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/ |
D | cq.c | 42 void mlx5_cq_completion(struct mlx5_core_dev *dev, u32 cqn) in mlx5_cq_completion() argument 48 cq = radix_tree_lookup(&table->tree, cqn); in mlx5_cq_completion() 54 mlx5_core_warn(dev, "Completion event for bogus CQ 0x%x\n", cqn); in mlx5_cq_completion() 66 void mlx5_cq_event(struct mlx5_core_dev *dev, u32 cqn, int event_type) in mlx5_cq_event() argument 73 cq = radix_tree_lookup(&table->tree, cqn); in mlx5_cq_event() 80 mlx5_core_warn(dev, "Async event for bogus CQ 0x%x\n", cqn); in mlx5_cq_event() 109 cq->cqn = be32_to_cpu(out.cqn) & 0xffffff; in mlx5_core_create_cq() 116 err = radix_tree_insert(&table->tree, cq->cqn, cq); in mlx5_core_create_cq() 125 cq->cqn); in mlx5_core_create_cq() 147 tmp = radix_tree_delete(&table->tree, cq->cqn); in mlx5_core_destroy_cq() [all …]
|
D | eq.c | 202 u32 cqn; in mlx5_eq_int() local 217 cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff; in mlx5_eq_int() 218 mlx5_cq_completion(dev, cqn); in mlx5_eq_int() 267 cqn = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff; in mlx5_eq_int() 269 cqn, eqe->data.cq_err.syndrome); in mlx5_eq_int() 270 mlx5_cq_event(dev, cqn, eqe->type); in mlx5_eq_int()
|
D | en_txrx.c | 104 __func__, mcq->cqn, event); in mlx5e_cq_error_event()
|
D | debugfs.c | 595 &cq->dbg, cq->cqn, cq_fields, in mlx5_debug_cq_add()
|
D | en_main.c | 393 MLX5_SET(rqc, rqc, cqn, rq->cq.mcq.cqn); in mlx5e_enable_rq() 622 MLX5_SET(sqc, sqc, cqn, c->sq[sq->tc].cq.mcq.cqn); in mlx5e_enable_sq()
|
/linux-4.4.14/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 76 __be32 cqn; member 211 mthca_write64(MTHCA_TAVOR_CQ_DB_INC_CI | cq->cqn, incr - 1, in update_cons_index() 222 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn) in mthca_cq_completion() argument 226 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_completion() 229 mthca_warn(dev, "Completion event for bogus CQ %08x\n", cqn); in mthca_cq_completion() 238 void mthca_cq_event(struct mthca_dev *dev, u32 cqn, in mthca_cq_event() argument 246 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_event() 253 mthca_warn(dev, "Async event for bogus CQ %08x\n", cqn); in mthca_cq_event() 302 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean() 389 cq->cqn, cq->cons_index); in handle_error_cqe() [all …]
|
D | mthca_eq.c | 132 __be32 cqn; member 149 __be32 cqn; member 219 static inline void disarm_cq(struct mthca_dev *dev, int eqn, int cqn) in disarm_cq() argument 222 mthca_write64(MTHCA_EQ_DB_DISARM_CQ | eqn, cqn, in disarm_cq() 276 disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mthca_eq_int() 343 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); in mthca_eq_int() 344 mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn), in mthca_eq_int()
|
D | mthca_user.h | 83 __u32 cqn; member
|
D | mthca_provider.h | 205 int cqn; member
|
D | mthca_dev.h | 503 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn); 504 void mthca_cq_event(struct mthca_dev *dev, u32 cqn,
|
D | mthca_qp.c | 719 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); in __mthca_modify_qp() 758 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp() 1319 } else if (send_cq->cqn < recv_cq->cqn) { in mthca_lock_cqs() 1334 } else if (send_cq->cqn < recv_cq->cqn) { in mthca_unlock_cqs()
|
D | mthca_provider.c | 704 if (context && ib_copy_to_udata(udata, &cq->cqn, sizeof (__u32))) { in mthca_create_cq() 807 ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries)); in mthca_resize_cq()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
D | cq.c | 56 type, mcq->cqn); in mlx5_ib_cq_event() 464 cq->mcq.cqn, qpn); in mlx5_poll_one() 498 "Requestor" : "Responder", cq->mcq.cqn); in mlx5_poll_one() 531 cq->mcq.cqn, be32_to_cpu(sig_err_cqe->mkey)); in mlx5_poll_one() 541 cq->mcq.cqn, mr->sig->err_item.key, in mlx5_poll_one() 811 mlx5_ib_dbg(dev, "cqn 0x%x\n", cq->mcq.cqn); in mlx5_ib_create_cq() 817 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof(__u32))) { in mlx5_ib_create_cq() 944 in->cqn = cpu_to_be32(mcq->mcq.cqn); in mlx5_ib_modify_cq() 953 mlx5_ib_warn(dev, "modify cq 0x%x failed\n", mcq->mcq.cqn); in mlx5_ib_modify_cq() 1072 cq->mcq.cqn); in copy_resize_cqes() [all …]
|
D | user.h | 98 __u32 cqn; member
|
D | qp.c | 992 in->ctx.cqn_recv = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common() 993 in->ctx.cqn_send = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common() 998 in->ctx.cqn_recv = cpu_to_be32(to_mcq(devr->c0)->mcq.cqn); in create_qp_common() 1014 in->ctx.cqn_send = cpu_to_be32(to_mcq(init_attr->send_cq)->mcq.cqn); in create_qp_common() 1017 in->ctx.cqn_recv = cpu_to_be32(to_mcq(init_attr->recv_cq)->mcq.cqn); in create_qp_common() 1053 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs() 1057 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs() 1083 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs() 1086 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs() 1278 qp->ibqp.qp_num, qp->mqp.qpn, to_mcq(init_attr->recv_cq)->mcq.cqn, in mlx5_ib_create_qp() [all …]
|
D | srq.c | 295 in->ctx.pgoff_cqn |= cpu_to_be32(to_mcq(init_attr->ext.xrc.cq)->mcq.cqn); in mlx5_ib_create_srq() 298 in->ctx.pgoff_cqn |= cpu_to_be32(to_mcq(dev->devr.c0)->mcq.cqn); in mlx5_ib_create_srq()
|
/linux-4.4.14/include/linux/mlx5/ |
D | cq.h | 41 u32 cqn; member 158 doorbell[1] = cpu_to_be32(cq->cqn); in mlx5_cq_arm()
|
D | device.h | 466 __be32 cqn; member 475 __be32 cqn; member 782 __be32 cqn; member 788 __be32 cqn; member 799 __be32 cqn; member 813 __be32 cqn; member
|
D | mlx5_ifc.h | 1544 u8 cqn[0x18]; member 1884 u8 cqn[0x18]; member 2009 u8 cqn[0x18]; member 2051 u8 cqn[0x18]; member 2105 u8 cqn[0x18]; member 2344 u8 cqn[0x18]; member 3784 u8 cqn[0x18]; member 4304 u8 cqn[0x18]; member 5002 u8 cqn[0x18]; member 5646 u8 cqn[0x18]; member
|
D | driver.h | 751 void mlx5_cq_completion(struct mlx5_core_dev *dev, u32 cqn); 759 void mlx5_cq_event(struct mlx5_core_dev *dev, u32 cqn, int event_type);
|
/linux-4.4.14/drivers/staging/rdma/amso1100/ |
D | c2_cq.c | 46 static struct c2_cq *c2_cq_get(struct c2_dev *c2dev, int cqn) in c2_cq_get() argument 52 cq = c2dev->qptr_array[cqn]; in c2_cq_get() 372 cq->cqn = cq->mq.index; in c2_init_cq() 373 c2dev->qptr_array[cq->cqn] = cq; in c2_init_cq()
|
D | c2_user.h | 69 __u32 cqn; member
|
D | c2_provider.h | 95 int cqn; member
|
D | c2_qp.c | 624 c2_cq_clean(c2dev, qp, send_cq->cqn); in c2_free_qp() 626 c2_cq_clean(c2dev, qp, recv_cq->cqn); in c2_free_qp()
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
D | srq.c | 80 u32 cqn; in mlx4_ib_create_srq() local 185 cqn = (init_attr->srq_type == IB_SRQT_XRC) ? in mlx4_ib_create_srq() 186 to_mcq(init_attr->ext.xrc.cq)->mcq.cqn : 0; in mlx4_ib_create_srq() 190 err = mlx4_srq_alloc(dev->dev, to_mpd(pd)->pdn, cqn, xrcdn, &srq->mtt, in mlx4_ib_create_srq()
|
D | user.h | 80 __u32 cqn; member
|
D | cq.c | 55 "on CQ %06x\n", type, cq->cqn); in mlx4_ib_cq_event() 254 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof (__u32))) { in mlx4_ib_create_cq() 725 cq->mcq.cqn, be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK); in mlx4_ib_poll_one() 743 cq->mcq.cqn, srq_num); in mlx4_ib_poll_one()
|
D | qp.c | 939 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_lock_cqs() 954 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_unlock_cqs() 1706 context->cqn_send = cpu_to_be32(send_cq->mcq.cqn); in __mlx4_ib_modify_qp() 1707 context->cqn_recv = cpu_to_be32(recv_cq->mcq.cqn); in __mlx4_ib_modify_qp()
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlxsw/ |
D | pci.h | 199 MLXSW_ITEM32(pci, eqe, cqn, 0x0C, 8, 7);
|
D | pci.c | 839 u8 cqn; in mlxsw_pci_eq_tasklet() local 855 cqn = mlxsw_pci_eqe_cqn_get(eqe); in mlxsw_pci_eq_tasklet() 856 set_bit(cqn, active_cqns); in mlxsw_pci_eq_tasklet() 873 for_each_set_bit(cqn, active_cqns, cq_count) { in mlxsw_pci_eq_tasklet() 874 q = mlxsw_pci_cq_get(mlxsw_pci, cqn); in mlxsw_pci_eq_tasklet()
|
/linux-4.4.14/include/linux/mlx4/ |
D | cq.h | 158 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn); in mlx4_cq_arm()
|
D | device.h | 721 int cqn; member 872 __be32 cqn; member 889 __be32 cqn; member 1106 int mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcdn,
|