eqe 69 drivers/crypto/hisilicon/qm.c #define QM_EQE_PHASE(eqe) (((eqe)->dw0 >> 16) & 0x1) eqe 443 drivers/crypto/hisilicon/qm.c static struct hisi_qp *qm_to_hisi_qp(struct hisi_qm *qm, struct qm_eqe *eqe) eqe 445 drivers/crypto/hisilicon/qm.c u16 cqn = eqe->dw0 & QM_EQE_CQN_MASK; eqe 492 drivers/crypto/hisilicon/qm.c struct qm_eqe *eqe = qm->eqe + qm->status.eq_head; eqe 496 drivers/crypto/hisilicon/qm.c while (QM_EQE_PHASE(eqe) == qm->status.eqc_phase) { eqe 498 drivers/crypto/hisilicon/qm.c qp = qm_to_hisi_qp(qm, eqe); eqe 504 drivers/crypto/hisilicon/qm.c eqe = qm->eqe; eqe 507 drivers/crypto/hisilicon/qm.c eqe++; eqe 1648 drivers/crypto/hisilicon/qm.c QM_INIT_BUF(qm, eqe, QM_Q_DEPTH); eqe 1658 drivers/crypto/hisilicon/qm.c qm->eqe, (unsigned long)qm->eqe_dma, eqe 140 drivers/crypto/hisilicon/qm.h struct qm_eqe *eqe; eqe 63 drivers/infiniband/hw/mlx4/mad.c #define GET_BLK_PTR_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.block_ptr) eqe 64 drivers/infiniband/hw/mlx4/mad.c #define GET_MASK_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.tbl_entries_mask) eqe 1128 drivers/infiniband/hw/mlx4/mad.c struct mlx4_eqe *eqe) eqe 1130 drivers/infiniband/hw/mlx4/mad.c __propagate_pkey_ev(dev, port_num, GET_BLK_PTR_FROM_EQE(eqe), eqe 1131 drivers/infiniband/hw/mlx4/mad.c GET_MASK_FROM_EQE(eqe)); eqe 1189 drivers/infiniband/hw/mlx4/mad.c struct mlx4_eqe *eqe = &(ew->ib_eqe); eqe 1190 drivers/infiniband/hw/mlx4/mad.c u8 port = eqe->event.port_mgmt_change.port; eqe 1195 drivers/infiniband/hw/mlx4/mad.c switch (eqe->subtype) { eqe 1197 drivers/infiniband/hw/mlx4/mad.c changed_attr = be32_to_cpu(eqe->event.port_mgmt_change.params.port_info.changed_attr); eqe 1202 drivers/infiniband/hw/mlx4/mad.c u16 lid = be16_to_cpu(eqe->event.port_mgmt_change.params.port_info.mstr_sm_lid); eqe 1203 drivers/infiniband/hw/mlx4/mad.c u8 sl = eqe->event.port_mgmt_change.params.port_info.mstr_sm_sl & 0xf; eqe 1217 drivers/infiniband/hw/mlx4/mad.c if (!eqe->event.port_mgmt_change.params.port_info.gid_prefix) eqe 1221 drivers/infiniband/hw/mlx4/mad.c eqe->event.port_mgmt_change.params.port_info.gid_prefix; eqe 1248 drivers/infiniband/hw/mlx4/mad.c propagate_pkey_ev(dev, port, eqe); eqe 1256 drivers/infiniband/hw/mlx4/mad.c tbl_block = GET_BLK_PTR_FROM_EQE(eqe); eqe 1257 drivers/infiniband/hw/mlx4/mad.c change_bitmap = GET_MASK_FROM_EQE(eqe); eqe 1272 drivers/infiniband/hw/mlx4/mad.c eqe->event.port_mgmt_change.params.sl2vl_tbl_change_info.sl2vl_table[jj]; eqe 1281 drivers/infiniband/hw/mlx4/mad.c "Port Management Change event\n", eqe->subtype); eqe 3244 drivers/infiniband/hw/mlx4/main.c struct mlx4_eqe *eqe = NULL; eqe 3261 drivers/infiniband/hw/mlx4/main.c eqe = (struct mlx4_eqe *)param; eqe 3299 drivers/infiniband/hw/mlx4/main.c memcpy(&ew->ib_eqe, eqe, sizeof *eqe); eqe 40 drivers/infiniband/hw/mlx5/cq.c static void mlx5_ib_cq_comp(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe) eqe 305 drivers/infiniband/hw/mlx5/devx.c static u16 get_event_obj_type(unsigned long event_type, struct mlx5_eqe *eqe) eqe 318 drivers/infiniband/hw/mlx5/devx.c return eqe->data.qp_srq.type; eqe 326 drivers/infiniband/hw/mlx5/devx.c return MLX5_GET(affiliated_event_header, &eqe->data, obj_type); eqe 1374 drivers/infiniband/hw/mlx5/devx.c static void devx_cq_comp(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) eqe 1392 drivers/infiniband/hw/mlx5/devx.c dispatch_event_fd(&obj_event->obj_sub_list, eqe); eqe 2269 drivers/infiniband/hw/mlx5/devx.c struct mlx5_eqe *eqe = data; eqe 2283 drivers/infiniband/hw/mlx5/devx.c obj_id = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; eqe 2286 drivers/infiniband/hw/mlx5/devx.c obj_id = be32_to_cpu(eqe->data.xrq_err.type_xrqn) & 0xffffff; eqe 2290 drivers/infiniband/hw/mlx5/devx.c obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; eqe 2293 drivers/infiniband/hw/mlx5/devx.c obj_id = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff; eqe 2296 drivers/infiniband/hw/mlx5/devx.c obj_id = MLX5_GET(affiliated_event_header, &eqe->data, obj_id); eqe 4546 drivers/infiniband/hw/mlx5/main.c static void handle_general_event(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe, eqe 4549 drivers/infiniband/hw/mlx5/main.c u8 port = (eqe->data.port.port >> 4) & 0xf; eqe 4551 drivers/infiniband/hw/mlx5/main.c switch (eqe->sub_type) { eqe 4562 drivers/infiniband/hw/mlx5/main.c static int handle_port_change(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe, eqe 4565 drivers/infiniband/hw/mlx5/main.c u8 port = (eqe->data.port.port >> 4) & 0xf; eqe 4569 drivers/infiniband/hw/mlx5/main.c switch (eqe->sub_type) { eqe 4580 drivers/infiniband/hw/mlx5/main.c ibev->event = (eqe->sub_type == MLX5_PORT_CHANGE_SUBTYPE_ACTIVE) ? eqe 1365 drivers/infiniband/hw/mlx5/odp.c struct mlx5_eqe *eqe; eqe 1368 drivers/infiniband/hw/mlx5/odp.c while ((eqe = mlx5_eq_get_eqe(eq->core, cc))) { eqe 1375 drivers/infiniband/hw/mlx5/odp.c pf_eqe = &eqe->data.page_fault; eqe 1376 drivers/infiniband/hw/mlx5/odp.c pfault->event_subtype = eqe->sub_type; eqe 1381 drivers/infiniband/hw/mlx5/odp.c eqe->sub_type, pfault->bytes_committed); eqe 1383 drivers/infiniband/hw/mlx5/odp.c switch (eqe->sub_type) { eqe 1432 drivers/infiniband/hw/mlx5/odp.c eqe->sub_type); eqe 663 drivers/infiniband/hw/mlx5/srq_cmd.c struct mlx5_eqe *eqe; eqe 672 drivers/infiniband/hw/mlx5/srq_cmd.c eqe = data; eqe 673 drivers/infiniband/hw/mlx5/srq_cmd.c srqn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; eqe 684 drivers/infiniband/hw/mlx5/srq_cmd.c srq->event(srq, eqe->type); eqe 236 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eqe *eqe; eqe 237 drivers/infiniband/hw/mthca/mthca_eq.c eqe = get_eqe(eq, eq->cons_index); eqe 238 drivers/infiniband/hw/mthca/mthca_eq.c return (MTHCA_EQ_ENTRY_OWNER_HW & eqe->owner) ? NULL : eqe; eqe 241 drivers/infiniband/hw/mthca/mthca_eq.c static inline void set_eqe_hw(struct mthca_eqe *eqe) eqe 243 drivers/infiniband/hw/mthca/mthca_eq.c eqe->owner = MTHCA_EQ_ENTRY_OWNER_HW; eqe 262 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eqe *eqe; eqe 267 drivers/infiniband/hw/mthca/mthca_eq.c while ((eqe = next_eqe_sw(eq))) { eqe 274 drivers/infiniband/hw/mthca/mthca_eq.c switch (eqe->type) { eqe 276 drivers/infiniband/hw/mthca/mthca_eq.c disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; eqe 282 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 287 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 292 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 297 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 302 drivers/infiniband/hw/mthca/mthca_eq.c mthca_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) & 0xffffff, eqe 307 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 312 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 317 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 322 drivers/infiniband/hw/mthca/mthca_eq.c mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, eqe 328 drivers/infiniband/hw/mthca/mthca_eq.c be16_to_cpu(eqe->event.cmd.token), eqe 329 drivers/infiniband/hw/mthca/mthca_eq.c eqe->event.cmd.status, eqe 330 drivers/infiniband/hw/mthca/mthca_eq.c be64_to_cpu(eqe->event.cmd.out_param)); eqe 335 drivers/infiniband/hw/mthca/mthca_eq.c (be32_to_cpu(eqe->event.port_change.port) >> 28) & 3, eqe 336 drivers/infiniband/hw/mthca/mthca_eq.c eqe->subtype == 0x4); eqe 341 drivers/infiniband/hw/mthca/mthca_eq.c eqe->event.cq_err.syndrome == 1 ? eqe 343 drivers/infiniband/hw/mthca/mthca_eq.c be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); eqe 344 drivers/infiniband/hw/mthca/mthca_eq.c mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn), eqe 358 drivers/infiniband/hw/mthca/mthca_eq.c eqe->type, eqe->subtype, eq->eqn); eqe 362 drivers/infiniband/hw/mthca/mthca_eq.c set_eqe_hw(eqe); eqe 998 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_eqe eqe; eqe 1006 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eqe = *ptr; eqe 1007 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_le32_to_cpu(&eqe, sizeof(eqe)); eqe 1008 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mcode = (eqe.id_valid & OCRDMA_EQE_MAJOR_CODE_MASK) eqe 1012 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->q.id, eqe.id_valid); eqe 1013 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if ((eqe.id_valid & OCRDMA_EQE_VALID_MASK) == 0) eqe 1020 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if ((eqe.id_valid & OCRDMA_EQE_FOR_CQE_MASK) == 0) { eqe 1021 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cq_id = eqe.id_valid >> OCRDMA_EQE_RESOURCE_ID_SHIFT; eqe 440 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c struct pvrdma_eqe *eqe; eqe 442 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c eqe = get_eqe(dev, head); eqe 444 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c switch (eqe->type) { eqe 453 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c pvrdma_qp_event(dev, eqe->info, eqe->type); eqe 457 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c pvrdma_cq_event(dev, eqe->info, eqe->type); eqe 462 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c pvrdma_srq_event(dev, eqe->info, eqe->type); eqe 472 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c pvrdma_dev_event(dev, eqe->info, eqe->type); eqe 476 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c pvrdma_dev_event(dev, 1, eqe->type); eqe 2786 drivers/net/ethernet/emulex/benet/be_main.c struct be_eq_entry *eqe; eqe 2790 drivers/net/ethernet/emulex/benet/be_main.c eqe = queue_tail_node(&eqo->q); eqe 2791 drivers/net/ethernet/emulex/benet/be_main.c if (eqe->evt == 0) eqe 2795 drivers/net/ethernet/emulex/benet/be_main.c eqe->evt = 0; eqe 921 drivers/net/ethernet/ibm/ehea/ehea_main.c struct ehea_eqe *eqe; eqe 927 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(port->qp_eq); eqe 929 drivers/net/ethernet/ibm/ehea/ehea_main.c while (eqe) { eqe 930 drivers/net/ethernet/ibm/ehea/ehea_main.c qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry); eqe 932 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe->entry, qp_token); eqe 946 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(port->qp_eq); eqe 1134 drivers/net/ethernet/ibm/ehea/ehea_main.c static void ehea_parse_eqe(struct ehea_adapter *adapter, u64 eqe) eqe 1142 drivers/net/ethernet/ibm/ehea/ehea_main.c ec = EHEA_BMASK_GET(NEQE_EVENT_CODE, eqe); eqe 1143 drivers/net/ethernet/ibm/ehea/ehea_main.c portnum = EHEA_BMASK_GET(NEQE_PORTNUM, eqe); eqe 1154 drivers/net/ethernet/ibm/ehea/ehea_main.c if (EHEA_BMASK_GET(NEQE_PORT_UP, eqe)) { eqe 1179 drivers/net/ethernet/ibm/ehea/ehea_main.c if (EHEA_BMASK_GET(NEQE_EXTSWITCH_PORT_UP, eqe)) { eqe 1193 drivers/net/ethernet/ibm/ehea/ehea_main.c if (EHEA_BMASK_GET(NEQE_EXTSWITCH_PRIMARY, eqe)) eqe 1210 drivers/net/ethernet/ibm/ehea/ehea_main.c netdev_err(dev, "unknown event code %x, eqe=0x%llX\n", ec, eqe); eqe 1218 drivers/net/ethernet/ibm/ehea/ehea_main.c struct ehea_eqe *eqe; eqe 1221 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(adapter->neq); eqe 1222 drivers/net/ethernet/ibm/ehea/ehea_main.c pr_debug("eqe=%p\n", eqe); eqe 1224 drivers/net/ethernet/ibm/ehea/ehea_main.c while (eqe) { eqe 1225 drivers/net/ethernet/ibm/ehea/ehea_main.c pr_debug("*eqe=%lx\n", (unsigned long) eqe->entry); eqe 1226 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_parse_eqe(adapter, eqe->entry); eqe 1227 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(adapter->neq); eqe 1228 drivers/net/ethernet/ibm/ehea/ehea_main.c pr_debug("next eqe=%p\n", eqe); eqe 305 drivers/net/ethernet/ibm/ehea/ehea_qmr.c struct ehea_eqe *eqe; eqe 309 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue); eqe 312 drivers/net/ethernet/ibm/ehea/ehea_qmr.c return eqe; eqe 123 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); eqe 124 drivers/net/ethernet/mellanox/mlx4/eq.c return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; eqe 129 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe = eqe 131 drivers/net/ethernet/mellanox/mlx4/eq.c return (!!(eqe->owner & 0x80) ^ eqe 133 drivers/net/ethernet/mellanox/mlx4/eq.c eqe : NULL; eqe 146 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe; eqe 150 drivers/net/ethernet/mellanox/mlx4/eq.c for (eqe = next_slave_event_eqe(slave_eq); eqe; eqe 151 drivers/net/ethernet/mellanox/mlx4/eq.c eqe = next_slave_event_eqe(slave_eq)) { eqe 152 drivers/net/ethernet/mellanox/mlx4/eq.c slave = eqe->slave_id; eqe 154 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->type == MLX4_EVENT_TYPE_PORT_CHANGE && eqe 155 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->subtype == MLX4_PORT_CHANGE_SUBTYPE_DOWN && eqe 169 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->type == MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT && eqe 170 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->subtype == MLX4_DEV_PMC_SUBTYPE_PORT_INFO) { eqe 171 drivers/net/ethernet/mellanox/mlx4/eq.c phys_port = eqe->event.port_mgmt_change.port; eqe 175 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.port_mgmt_change.port = slave_port; eqe 177 drivers/net/ethernet/mellanox/mlx4/eq.c if (mlx4_GEN_EQE(dev, i, eqe)) eqe 181 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.port_mgmt_change.port = phys_port; eqe 184 drivers/net/ethernet/mellanox/mlx4/eq.c if (mlx4_GEN_EQE(dev, slave, eqe)) eqe 194 drivers/net/ethernet/mellanox/mlx4/eq.c static void slave_event(struct mlx4_dev *dev, u8 slave, struct mlx4_eqe *eqe) eqe 211 drivers/net/ethernet/mellanox/mlx4/eq.c memcpy(s_eqe, eqe, sizeof(struct mlx4_eqe) - 1); eqe 224 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe) eqe 233 drivers/net/ethernet/mellanox/mlx4/eq.c slave_event(dev, slave, eqe); eqe 255 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe eqe; eqe 263 drivers/net/ethernet/mellanox/mlx4/eq.c memset(&eqe, 0, sizeof(eqe)); eqe 265 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT; eqe 266 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.subtype = MLX4_DEV_PMC_SUBTYPE_PKEY_TABLE; eqe 267 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.event.port_mgmt_change.port = mlx4_phys_to_slave_port(dev, slave, port); eqe 269 drivers/net/ethernet/mellanox/mlx4/eq.c return mlx4_GEN_EQE(dev, slave, &eqe); eqe 275 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe eqe; eqe 280 drivers/net/ethernet/mellanox/mlx4/eq.c memset(&eqe, 0, sizeof(eqe)); eqe 282 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT; eqe 283 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.subtype = MLX4_DEV_PMC_SUBTYPE_GUID_INFO; eqe 284 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.event.port_mgmt_change.port = mlx4_phys_to_slave_port(dev, slave, port); eqe 286 drivers/net/ethernet/mellanox/mlx4/eq.c return mlx4_GEN_EQE(dev, slave, &eqe); eqe 293 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe eqe; eqe 299 drivers/net/ethernet/mellanox/mlx4/eq.c memset(&eqe, 0, sizeof(eqe)); eqe 301 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.type = MLX4_EVENT_TYPE_PORT_CHANGE; eqe 302 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.subtype = port_subtype_change; eqe 303 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.event.port_change.port = cpu_to_be32(slave_port << 28); eqe 307 drivers/net/ethernet/mellanox/mlx4/eq.c return mlx4_GEN_EQE(dev, slave, &eqe); eqe 434 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe eqe; eqe 436 drivers/net/ethernet/mellanox/mlx4/eq.c memset(&eqe, 0, sizeof(eqe)); eqe 438 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT; eqe 439 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.subtype = MLX4_DEV_PMC_SUBTYPE_PORT_INFO; eqe 440 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.event.port_mgmt_change.port = port; eqe 441 drivers/net/ethernet/mellanox/mlx4/eq.c eqe.event.port_mgmt_change.params.port_info.changed_attr = eqe 444 drivers/net/ethernet/mellanox/mlx4/eq.c slave_event(dev, ALL_SLAVES, &eqe); eqe 497 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe; eqe 512 drivers/net/ethernet/mellanox/mlx4/eq.c while ((eqe = next_eqe_sw(eq, dev->caps.eqe_factor, eqe_size))) { eqe 519 drivers/net/ethernet/mellanox/mlx4/eq.c switch (eqe->type) { eqe 521 drivers/net/ethernet/mellanox/mlx4/eq.c cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; eqe 533 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_dbg(dev, "event %d arrived\n", eqe->type); eqe 538 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.qp.qpn) eqe 542 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eqe 548 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, slave, eqe); eqe 553 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & eqe 554 drivers/net/ethernet/mellanox/mlx4/eq.c 0xffffff, eqe->type); eqe 559 drivers/net/ethernet/mellanox/mlx4/eq.c __func__, be32_to_cpu(eqe->event.srq.srqn), eqe 567 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.srq.srqn) eqe 572 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eqe 576 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->type == eqe 580 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.srq.srqn), eqe 581 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype); eqe 584 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->type == eqe 587 drivers/net/ethernet/mellanox/mlx4/eq.c __func__, eqe->type, eqe 588 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->subtype, slave); eqe 589 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, slave, eqe); eqe 593 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) & eqe 594 drivers/net/ethernet/mellanox/mlx4/eq.c 0xffffff, eqe->type); eqe 599 drivers/net/ethernet/mellanox/mlx4/eq.c be16_to_cpu(eqe->event.cmd.token), eqe 600 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.cmd.status, eqe 601 drivers/net/ethernet/mellanox/mlx4/eq.c be64_to_cpu(eqe->event.cmd.out_param)); eqe 606 drivers/net/ethernet/mellanox/mlx4/eq.c port = be32_to_cpu(eqe->event.port_change.port) >> 28; eqe 608 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->subtype == MLX4_PORT_CHANGE_SUBTYPE_DOWN) { eqe 627 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.port_change.port = eqe 629 drivers/net/ethernet/mellanox/mlx4/eq.c (be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF) eqe 631 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, i, eqe); eqe 641 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.port_change.port = eqe 643 drivers/net/ethernet/mellanox/mlx4/eq.c (be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF) eqe 645 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, i, eqe); eqe 668 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.port_change.port = eqe 670 drivers/net/ethernet/mellanox/mlx4/eq.c (be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF) eqe 672 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, i, eqe); eqe 686 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.cq_err.syndrome == 1 ? eqe 688 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); eqe 692 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.cq_err.cqn) eqe 696 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eqe 702 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, slave, eqe); eqe 707 drivers/net/ethernet/mellanox/mlx4/eq.c be32_to_cpu(eqe->event.cq_err.cqn) eqe 709 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type); eqe 730 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.comm_channel_arm.bit_vec, eqe 731 drivers/net/ethernet/mellanox/mlx4/eq.c sizeof(eqe->event.comm_channel_arm.bit_vec)); eqe 737 drivers/net/ethernet/mellanox/mlx4/eq.c flr_slave = be32_to_cpu(eqe->event.flr_event.slave_id); eqe 767 drivers/net/ethernet/mellanox/mlx4/eq.c if (eqe->subtype == MLX4_FATAL_WARNING_SUBTYPE_WARMING) { eqe 774 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_slave_event(dev, i, eqe); eqe 777 drivers/net/ethernet/mellanox/mlx4/eq.c be16_to_cpu(eqe->event.warming.warning_threshold), eqe 778 drivers/net/ethernet/mellanox/mlx4/eq.c be16_to_cpu(eqe->event.warming.current_temperature)); eqe 781 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eqe 782 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eqe 783 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->slave_id, eqe 784 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eqe->owner & 0x80) ^ eqe 791 drivers/net/ethernet/mellanox/mlx4/eq.c (unsigned long) eqe); eqe 795 drivers/net/ethernet/mellanox/mlx4/eq.c switch (eqe->subtype) { eqe 798 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->event.bad_cable.port); eqe 806 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eqe 807 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eqe 808 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eqe->owner & 0x80) ^ eqe 818 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eqe 819 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eqe 820 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->slave_id, eqe 821 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eqe->owner & 0x80) ^ eqe 1198 drivers/net/ethernet/mellanox/mlx4/mlx4.h int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe); eqe 3338 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe) eqe 3357 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c event_eq = &priv->mfunc.master.slave_state[slave].event_eq[eqe->type]; eqe 3380 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (eqe->type == MLX4_EVENT_TYPE_CMD) { eqe 3382 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c eqe->event.cmd.token = cpu_to_be16(event_eq->token); eqe 3385 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c memcpy(mailbox->buf, (u8 *) eqe, 28); eqe 1429 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_eqe *eqe; eqe 1433 drivers/net/ethernet/mellanox/mlx5/core/cmd.c eqe = data; eqe 1435 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_cmd_comp_handler(dev, be32_to_cpu(eqe->data.cmd.vector), false); eqe 72 drivers/net/ethernet/mellanox/mlx5/core/cq.c struct mlx5_eqe *eqe) eqe 1071 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c struct mlx5_eqe *eqe = data; eqe 1073 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c switch (eqe->sub_type) { eqe 1084 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c eqe->sub_type); eqe 914 drivers/net/ethernet/mellanox/mlx5/core/en.h void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe); eqe 61 drivers/net/ethernet/mellanox/mlx5/core/en/monitor_stats.c unsigned long event, void *eqe) eqe 204 drivers/net/ethernet/mellanox/mlx5/core/en_main.c struct mlx5_eqe *eqe = data; eqe 209 drivers/net/ethernet/mellanox/mlx5/core/en_main.c switch (eqe->sub_type) { eqe 1699 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eqe *eqe = data; eqe 1701 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c switch (eqe->sub_type) { eqe 199 drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) eqe 135 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *eqe; eqe 139 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = next_eqe_sw(eq); eqe 140 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (!eqe) eqe 151 drivers/net/ethernet/mellanox/mlx5/core/eq.c cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff; eqe 156 drivers/net/ethernet/mellanox/mlx5/core/eq.c cq->comp(cq, eqe); eqe 164 drivers/net/ethernet/mellanox/mlx5/core/eq.c } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq))); eqe 201 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *eqe; eqe 207 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = next_eqe_sw(eq); eqe 208 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (!eqe) eqe 218 drivers/net/ethernet/mellanox/mlx5/core/eq.c atomic_notifier_call_chain(&eqt->nh[eqe->type], eqe->type, eqe); eqe 219 drivers/net/ethernet/mellanox/mlx5/core/eq.c atomic_notifier_call_chain(&eqt->nh[MLX5_EVENT_TYPE_NOTIFY_ANY], eqe->type, eqe); eqe 223 drivers/net/ethernet/mellanox/mlx5/core/eq.c } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq))); eqe 233 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *eqe; eqe 237 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = get_eqe(eq, i); eqe 238 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe->owner = MLX5_EQE_OWNER_INIT_VAL; eqe 479 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *eqe; eqe 487 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = data; eqe 489 drivers/net/ethernet/mellanox/mlx5/core/eq.c cqn = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff; eqe 491 drivers/net/ethernet/mellanox/mlx5/core/eq.c cqn, eqe->data.cq_err.syndrome); eqe 736 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *eqe; eqe 738 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = get_eqe(eq, ci & (eq->nent - 1)); eqe 739 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = ((eqe->owner & 1) ^ !!(ci & eq->nent)) ? NULL : eqe; eqe 743 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (eqe) eqe 746 drivers/net/ethernet/mellanox/mlx5/core/eq.c return eqe; eqe 1767 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c struct mlx5_eqe *eqe = data; eqe 1771 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c vport_num = be16_to_cpu(eqe->data.vport_change.vport_num); eqe 142 drivers/net/ethernet/mellanox/mlx5/core/events.c struct mlx5_eqe *eqe = data; eqe 145 drivers/net/ethernet/mellanox/mlx5/core/events.c eqe_type_str(eqe->type), eqe->sub_type); eqe 154 drivers/net/ethernet/mellanox/mlx5/core/events.c struct mlx5_eqe *eqe = data; eqe 158 drivers/net/ethernet/mellanox/mlx5/core/events.c value_lsb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_lsb); eqe 159 drivers/net/ethernet/mellanox/mlx5/core/events.c value_msb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_msb); eqe 216 drivers/net/ethernet/mellanox/mlx5/core/events.c struct mlx5_eqe *eqe = data; eqe 224 drivers/net/ethernet/mellanox/mlx5/core/events.c module_event_eqe = &eqe->data.port_module; eqe 307 drivers/net/ethernet/mellanox/mlx5/core/events.c struct mlx5_eqe *eqe = data; eqe 309 drivers/net/ethernet/mellanox/mlx5/core/events.c switch (eqe->sub_type) { eqe 330 drivers/net/ethernet/mellanox/mlx5/core/events.c struct mlx5_eqe *eqe = data; eqe 333 drivers/net/ethernet/mellanox/mlx5/core/events.c eqe_type_str(eqe->type), eqe->sub_type); eqe 418 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c struct mlx5_eqe *eqe) eqe 151 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c static int fpga_err_event(struct notifier_block *nb, unsigned long event, void *eqe) eqe 155 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c return mlx5_fpga_event(fdev, event, eqe); eqe 158 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c static int fpga_qp_err_event(struct notifier_block *nb, unsigned long event, void *eqe) eqe 162 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c return mlx5_fpga_event(fdev, event, eqe); eqe 314 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c unsigned long event, void *eqe) eqe 316 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c void *data = ((struct mlx5_eqe *)eqe)->data.raw; eqe 472 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c struct mlx5_eqe *eqe = data; eqe 473 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c int pin = eqe->data.pps.pin; eqe 481 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c be64_to_cpu(eqe->data.pps.time_stamp)); eqe 57 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); eqe 59 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; eqe 458 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c struct mlx5_eqe *eqe; eqe 465 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c eqe = data; eqe 467 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c func_id = be16_to_cpu(eqe->data.req_pages.func_id); eqe 468 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c npages = be32_to_cpu(eqe->data.req_pages.num_pages); eqe 469 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c ec_function = be16_to_cpu(eqe->data.req_pages.ec_function) & EC_FUNCTION_MASK; eqe 132 drivers/net/ethernet/mellanox/mlx5/core/qp.c struct mlx5_eqe *eqe; eqe 137 drivers/net/ethernet/mellanox/mlx5/core/qp.c eqe = data; eqe 138 drivers/net/ethernet/mellanox/mlx5/core/qp.c rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; eqe 149 drivers/net/ethernet/mellanox/mlx5/core/qp.c eqe = data; eqe 150 drivers/net/ethernet/mellanox/mlx5/core/qp.c rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; eqe 151 drivers/net/ethernet/mellanox/mlx5/core/qp.c rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN); eqe 161 drivers/net/ethernet/mellanox/mlx5/core/qp.c mlx5_core_dbg(dev, "event (%d) arrived on resource 0x%x\n", eqe->type, rsn); eqe 693 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c struct mlx5_eqe *eqe) eqe 694 drivers/net/ethernet/mellanox/mlxsw/pci.c static void mlxsw_pci_eq_cmd_event(struct mlxsw_pci *mlxsw_pci, char *eqe) eqe 696 drivers/net/ethernet/mellanox/mlxsw/pci.c mlxsw_pci->cmd.comp.status = mlxsw_pci_eqe_cmd_status_get(eqe); eqe 698 drivers/net/ethernet/mellanox/mlxsw/pci.c ((u64) mlxsw_pci_eqe_cmd_out_param_h_get(eqe)) << 32 | eqe 699 drivers/net/ethernet/mellanox/mlxsw/pci.c mlxsw_pci_eqe_cmd_out_param_l_get(eqe); eqe 726 drivers/net/ethernet/mellanox/mlxsw/pci.c char *eqe; eqe 734 drivers/net/ethernet/mellanox/mlxsw/pci.c while ((eqe = mlxsw_pci_eq_sw_eqe_get(q))) { eqe 742 drivers/net/ethernet/mellanox/mlxsw/pci.c mlxsw_pci_eq_cmd_event(mlxsw_pci, eqe); eqe 746 drivers/net/ethernet/mellanox/mlxsw/pci.c cqn = mlxsw_pci_eqe_cqn_get(eqe); eqe 218 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, event_type, 0x0C, 24, 8); eqe 225 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, event_sub_type, 0x0C, 16, 8); eqe 230 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cqn, 0x0C, 8, 7); eqe 235 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, owner, 0x0C, 0, 1); eqe 240 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_token, 0x00, 16, 16); eqe 245 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_status, 0x00, 0, 8); eqe 250 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x04, 0, 32); eqe 255 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x08, 0, 32); eqe 668 drivers/scsi/be2iscsi/be_main.c struct be_eq_entry *eqe; eqe 678 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 681 drivers/scsi/be2iscsi/be_main.c while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] eqe 683 drivers/scsi/be2iscsi/be_main.c if (((eqe->dw[offsetof(struct amap_eq_entry, eqe 688 drivers/scsi/be2iscsi/be_main.c AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0); eqe 690 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 732 drivers/scsi/be2iscsi/be_main.c struct be_eq_entry *eqe; eqe 753 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 757 drivers/scsi/be2iscsi/be_main.c while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] eqe 759 drivers/scsi/be2iscsi/be_main.c if (((eqe->dw[offsetof(struct amap_eq_entry, eqe 764 drivers/scsi/be2iscsi/be_main.c AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0); eqe 766 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 2037 drivers/scsi/be2iscsi/be_main.c struct be_eq_entry *eqe = NULL; eqe 2049 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 2050 drivers/scsi/be2iscsi/be_main.c while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] & eqe 2052 drivers/scsi/be2iscsi/be_main.c AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0); eqe 2054 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 3587 drivers/scsi/be2iscsi/be_main.c struct be_eq_entry *eqe = NULL; eqe 3603 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 3605 drivers/scsi/be2iscsi/be_main.c while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] eqe 3607 drivers/scsi/be2iscsi/be_main.c AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0); eqe 3609 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eqe 87 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe); eqe 326 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe; eqe 331 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_qe(q, q->host_index); eqe 334 drivers/scsi/lpfc/lpfc_sli.c if (bf_get_le32(lpfc_eqe_valid, eqe) != q->qe_valid) eqe 347 drivers/scsi/lpfc/lpfc_sli.c return eqe; eqe 457 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe) eqe 460 drivers/scsi/lpfc/lpfc_sli.c bf_set_le32(lpfc_eqe_valid, eqe, 0); eqe 472 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe; eqe 476 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eqe 477 drivers/scsi/lpfc/lpfc_sli.c while (eqe) { eqe 478 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(phba, eq, eqe); eqe 480 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eqe 491 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe; eqe 497 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eqe 498 drivers/scsi/lpfc/lpfc_sli.c while (eqe) { eqe 499 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_hba_handle_eqe(phba, eq, eqe); eqe 500 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(phba, eq, eqe); eqe 512 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eqe 13532 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_sp_handle_eqe(struct lpfc_hba *phba, struct lpfc_eqe *eqe, eqe 13539 drivers/scsi/lpfc/lpfc_sli.c cqid = bf_get_le32(lpfc_eqe_resource_id, eqe); eqe 14044 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_eqe *eqe) eqe 14050 drivers/scsi/lpfc/lpfc_sli.c if (unlikely(bf_get_le32(lpfc_eqe_major_code, eqe) != 0)) { eqe 14054 drivers/scsi/lpfc/lpfc_sli.c bf_get_le32(lpfc_eqe_major_code, eqe), eqe 14055 drivers/scsi/lpfc/lpfc_sli.c bf_get_le32(lpfc_eqe_minor_code, eqe)); eqe 14060 drivers/scsi/lpfc/lpfc_sli.c cqid = bf_get_le32(lpfc_eqe_resource_id, eqe); eqe 14087 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_sp_handle_eqe(phba, eqe, eqe 50 include/linux/mlx5/cq.h void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe); eqe 58 include/linux/mlx5/cq.h void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe); eqe 6445 include/linux/mlx5/mlx5_ifc.h u8 eqe[64][0x8];