eqe                69 drivers/crypto/hisilicon/qm.c #define QM_EQE_PHASE(eqe)		(((eqe)->dw0 >> 16) & 0x1)
eqe               443 drivers/crypto/hisilicon/qm.c static struct hisi_qp *qm_to_hisi_qp(struct hisi_qm *qm, struct qm_eqe *eqe)
eqe               445 drivers/crypto/hisilicon/qm.c 	u16 cqn = eqe->dw0 & QM_EQE_CQN_MASK;
eqe               492 drivers/crypto/hisilicon/qm.c 	struct qm_eqe *eqe = qm->eqe + qm->status.eq_head;
eqe               496 drivers/crypto/hisilicon/qm.c 	while (QM_EQE_PHASE(eqe) == qm->status.eqc_phase) {
eqe               498 drivers/crypto/hisilicon/qm.c 		qp = qm_to_hisi_qp(qm, eqe);
eqe               504 drivers/crypto/hisilicon/qm.c 			eqe = qm->eqe;
eqe               507 drivers/crypto/hisilicon/qm.c 			eqe++;
eqe              1648 drivers/crypto/hisilicon/qm.c 	QM_INIT_BUF(qm, eqe, QM_Q_DEPTH);
eqe              1658 drivers/crypto/hisilicon/qm.c 		     qm->eqe, (unsigned long)qm->eqe_dma,
eqe               140 drivers/crypto/hisilicon/qm.h 	struct qm_eqe *eqe;
eqe                63 drivers/infiniband/hw/mlx4/mad.c #define GET_BLK_PTR_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.block_ptr)
eqe                64 drivers/infiniband/hw/mlx4/mad.c #define GET_MASK_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.tbl_entries_mask)
eqe              1128 drivers/infiniband/hw/mlx4/mad.c 			      struct mlx4_eqe *eqe)
eqe              1130 drivers/infiniband/hw/mlx4/mad.c 	__propagate_pkey_ev(dev, port_num, GET_BLK_PTR_FROM_EQE(eqe),
eqe              1131 drivers/infiniband/hw/mlx4/mad.c 			    GET_MASK_FROM_EQE(eqe));
eqe              1189 drivers/infiniband/hw/mlx4/mad.c 	struct mlx4_eqe *eqe = &(ew->ib_eqe);
eqe              1190 drivers/infiniband/hw/mlx4/mad.c 	u8 port = eqe->event.port_mgmt_change.port;
eqe              1195 drivers/infiniband/hw/mlx4/mad.c 	switch (eqe->subtype) {
eqe              1197 drivers/infiniband/hw/mlx4/mad.c 		changed_attr = be32_to_cpu(eqe->event.port_mgmt_change.params.port_info.changed_attr);
eqe              1202 drivers/infiniband/hw/mlx4/mad.c 			u16 lid = be16_to_cpu(eqe->event.port_mgmt_change.params.port_info.mstr_sm_lid);
eqe              1203 drivers/infiniband/hw/mlx4/mad.c 			u8 sl = eqe->event.port_mgmt_change.params.port_info.mstr_sm_sl & 0xf;
eqe              1217 drivers/infiniband/hw/mlx4/mad.c 				if (!eqe->event.port_mgmt_change.params.port_info.gid_prefix)
eqe              1221 drivers/infiniband/hw/mlx4/mad.c 						eqe->event.port_mgmt_change.params.port_info.gid_prefix;
eqe              1248 drivers/infiniband/hw/mlx4/mad.c 			propagate_pkey_ev(dev, port, eqe);
eqe              1256 drivers/infiniband/hw/mlx4/mad.c 			tbl_block = GET_BLK_PTR_FROM_EQE(eqe);
eqe              1257 drivers/infiniband/hw/mlx4/mad.c 			change_bitmap = GET_MASK_FROM_EQE(eqe);
eqe              1272 drivers/infiniband/hw/mlx4/mad.c 					eqe->event.port_mgmt_change.params.sl2vl_tbl_change_info.sl2vl_table[jj];
eqe              1281 drivers/infiniband/hw/mlx4/mad.c 			"Port Management Change event\n", eqe->subtype);
eqe              3244 drivers/infiniband/hw/mlx4/main.c 	struct mlx4_eqe *eqe = NULL;
eqe              3261 drivers/infiniband/hw/mlx4/main.c 		eqe = (struct mlx4_eqe *)param;
eqe              3299 drivers/infiniband/hw/mlx4/main.c 		memcpy(&ew->ib_eqe, eqe, sizeof *eqe);
eqe                40 drivers/infiniband/hw/mlx5/cq.c static void mlx5_ib_cq_comp(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe)
eqe               305 drivers/infiniband/hw/mlx5/devx.c static u16 get_event_obj_type(unsigned long event_type, struct mlx5_eqe *eqe)
eqe               318 drivers/infiniband/hw/mlx5/devx.c 		return eqe->data.qp_srq.type;
eqe               326 drivers/infiniband/hw/mlx5/devx.c 		return MLX5_GET(affiliated_event_header, &eqe->data, obj_type);
eqe              1374 drivers/infiniband/hw/mlx5/devx.c static void devx_cq_comp(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe)
eqe              1392 drivers/infiniband/hw/mlx5/devx.c 	dispatch_event_fd(&obj_event->obj_sub_list, eqe);
eqe              2269 drivers/infiniband/hw/mlx5/devx.c 	struct mlx5_eqe *eqe = data;
eqe              2283 drivers/infiniband/hw/mlx5/devx.c 		obj_id = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
eqe              2286 drivers/infiniband/hw/mlx5/devx.c 		obj_id = be32_to_cpu(eqe->data.xrq_err.type_xrqn) & 0xffffff;
eqe              2290 drivers/infiniband/hw/mlx5/devx.c 		obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff;
eqe              2293 drivers/infiniband/hw/mlx5/devx.c 		obj_id = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff;
eqe              2296 drivers/infiniband/hw/mlx5/devx.c 		obj_id = MLX5_GET(affiliated_event_header, &eqe->data, obj_id);
eqe              4546 drivers/infiniband/hw/mlx5/main.c static void handle_general_event(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe,
eqe              4549 drivers/infiniband/hw/mlx5/main.c 	u8 port = (eqe->data.port.port >> 4) & 0xf;
eqe              4551 drivers/infiniband/hw/mlx5/main.c 	switch (eqe->sub_type) {
eqe              4562 drivers/infiniband/hw/mlx5/main.c static int handle_port_change(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe,
eqe              4565 drivers/infiniband/hw/mlx5/main.c 	u8 port = (eqe->data.port.port >> 4) & 0xf;
eqe              4569 drivers/infiniband/hw/mlx5/main.c 	switch (eqe->sub_type) {
eqe              4580 drivers/infiniband/hw/mlx5/main.c 		ibev->event = (eqe->sub_type == MLX5_PORT_CHANGE_SUBTYPE_ACTIVE) ?
eqe              1365 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_eqe *eqe;
eqe              1368 drivers/infiniband/hw/mlx5/odp.c 	while ((eqe = mlx5_eq_get_eqe(eq->core, cc))) {
eqe              1375 drivers/infiniband/hw/mlx5/odp.c 		pf_eqe = &eqe->data.page_fault;
eqe              1376 drivers/infiniband/hw/mlx5/odp.c 		pfault->event_subtype = eqe->sub_type;
eqe              1381 drivers/infiniband/hw/mlx5/odp.c 			    eqe->sub_type, pfault->bytes_committed);
eqe              1383 drivers/infiniband/hw/mlx5/odp.c 		switch (eqe->sub_type) {
eqe              1432 drivers/infiniband/hw/mlx5/odp.c 				     eqe->sub_type);
eqe               663 drivers/infiniband/hw/mlx5/srq_cmd.c 	struct mlx5_eqe *eqe;
eqe               672 drivers/infiniband/hw/mlx5/srq_cmd.c 	eqe = data;
eqe               673 drivers/infiniband/hw/mlx5/srq_cmd.c 	srqn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
eqe               684 drivers/infiniband/hw/mlx5/srq_cmd.c 	srq->event(srq, eqe->type);
eqe               236 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_eqe *eqe;
eqe               237 drivers/infiniband/hw/mthca/mthca_eq.c 	eqe = get_eqe(eq, eq->cons_index);
eqe               238 drivers/infiniband/hw/mthca/mthca_eq.c 	return (MTHCA_EQ_ENTRY_OWNER_HW & eqe->owner) ? NULL : eqe;
eqe               241 drivers/infiniband/hw/mthca/mthca_eq.c static inline void set_eqe_hw(struct mthca_eqe *eqe)
eqe               243 drivers/infiniband/hw/mthca/mthca_eq.c 	eqe->owner =  MTHCA_EQ_ENTRY_OWNER_HW;
eqe               262 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_eqe *eqe;
eqe               267 drivers/infiniband/hw/mthca/mthca_eq.c 	while ((eqe = next_eqe_sw(eq))) {
eqe               274 drivers/infiniband/hw/mthca/mthca_eq.c 		switch (eqe->type) {
eqe               276 drivers/infiniband/hw/mthca/mthca_eq.c 			disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff;
eqe               282 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               287 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               292 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               297 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               302 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) & 0xffffff,
eqe               307 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               312 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               317 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               322 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
eqe               328 drivers/infiniband/hw/mthca/mthca_eq.c 					be16_to_cpu(eqe->event.cmd.token),
eqe               329 drivers/infiniband/hw/mthca/mthca_eq.c 					eqe->event.cmd.status,
eqe               330 drivers/infiniband/hw/mthca/mthca_eq.c 					be64_to_cpu(eqe->event.cmd.out_param));
eqe               335 drivers/infiniband/hw/mthca/mthca_eq.c 				    (be32_to_cpu(eqe->event.port_change.port) >> 28) & 3,
eqe               336 drivers/infiniband/hw/mthca/mthca_eq.c 				    eqe->subtype == 0x4);
eqe               341 drivers/infiniband/hw/mthca/mthca_eq.c 				   eqe->event.cq_err.syndrome == 1 ?
eqe               343 drivers/infiniband/hw/mthca/mthca_eq.c 				   be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff);
eqe               344 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn),
eqe               358 drivers/infiniband/hw/mthca/mthca_eq.c 				   eqe->type, eqe->subtype, eq->eqn);
eqe               362 drivers/infiniband/hw/mthca/mthca_eq.c 		set_eqe_hw(eqe);
eqe               998 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_eqe eqe;
eqe              1006 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eqe = *ptr;
eqe              1007 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_le32_to_cpu(&eqe, sizeof(eqe));
eqe              1008 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		mcode = (eqe.id_valid & OCRDMA_EQE_MAJOR_CODE_MASK)
eqe              1012 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			       eq->q.id, eqe.id_valid);
eqe              1013 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		if ((eqe.id_valid & OCRDMA_EQE_VALID_MASK) == 0)
eqe              1020 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		if ((eqe.id_valid & OCRDMA_EQE_FOR_CQE_MASK) == 0) {
eqe              1021 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			cq_id = eqe.id_valid >> OCRDMA_EQE_RESOURCE_ID_SHIFT;
eqe               440 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 		struct pvrdma_eqe *eqe;
eqe               442 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 		eqe = get_eqe(dev, head);
eqe               444 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 		switch (eqe->type) {
eqe               453 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 			pvrdma_qp_event(dev, eqe->info, eqe->type);
eqe               457 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 			pvrdma_cq_event(dev, eqe->info, eqe->type);
eqe               462 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 			pvrdma_srq_event(dev, eqe->info, eqe->type);
eqe               472 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 			pvrdma_dev_event(dev, eqe->info, eqe->type);
eqe               476 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 			pvrdma_dev_event(dev, 1, eqe->type);
eqe              2786 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_eq_entry *eqe;
eqe              2790 drivers/net/ethernet/emulex/benet/be_main.c 		eqe = queue_tail_node(&eqo->q);
eqe              2791 drivers/net/ethernet/emulex/benet/be_main.c 		if (eqe->evt == 0)
eqe              2795 drivers/net/ethernet/emulex/benet/be_main.c 		eqe->evt = 0;
eqe               921 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct ehea_eqe *eqe;
eqe               927 drivers/net/ethernet/ibm/ehea/ehea_main.c 	eqe = ehea_poll_eq(port->qp_eq);
eqe               929 drivers/net/ethernet/ibm/ehea/ehea_main.c 	while (eqe) {
eqe               930 drivers/net/ethernet/ibm/ehea/ehea_main.c 		qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry);
eqe               932 drivers/net/ethernet/ibm/ehea/ehea_main.c 		       eqe->entry, qp_token);
eqe               946 drivers/net/ethernet/ibm/ehea/ehea_main.c 		eqe = ehea_poll_eq(port->qp_eq);
eqe              1134 drivers/net/ethernet/ibm/ehea/ehea_main.c static void ehea_parse_eqe(struct ehea_adapter *adapter, u64 eqe)
eqe              1142 drivers/net/ethernet/ibm/ehea/ehea_main.c 	ec = EHEA_BMASK_GET(NEQE_EVENT_CODE, eqe);
eqe              1143 drivers/net/ethernet/ibm/ehea/ehea_main.c 	portnum = EHEA_BMASK_GET(NEQE_PORTNUM, eqe);
eqe              1154 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (EHEA_BMASK_GET(NEQE_PORT_UP, eqe)) {
eqe              1179 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (EHEA_BMASK_GET(NEQE_EXTSWITCH_PORT_UP, eqe)) {
eqe              1193 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (EHEA_BMASK_GET(NEQE_EXTSWITCH_PRIMARY, eqe))
eqe              1210 drivers/net/ethernet/ibm/ehea/ehea_main.c 		netdev_err(dev, "unknown event code %x, eqe=0x%llX\n", ec, eqe);
eqe              1218 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct ehea_eqe *eqe;
eqe              1221 drivers/net/ethernet/ibm/ehea/ehea_main.c 	eqe = ehea_poll_eq(adapter->neq);
eqe              1222 drivers/net/ethernet/ibm/ehea/ehea_main.c 	pr_debug("eqe=%p\n", eqe);
eqe              1224 drivers/net/ethernet/ibm/ehea/ehea_main.c 	while (eqe) {
eqe              1225 drivers/net/ethernet/ibm/ehea/ehea_main.c 		pr_debug("*eqe=%lx\n", (unsigned long) eqe->entry);
eqe              1226 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_parse_eqe(adapter, eqe->entry);
eqe              1227 drivers/net/ethernet/ibm/ehea/ehea_main.c 		eqe = ehea_poll_eq(adapter->neq);
eqe              1228 drivers/net/ethernet/ibm/ehea/ehea_main.c 		pr_debug("next eqe=%p\n", eqe);
eqe               305 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	struct ehea_eqe *eqe;
eqe               309 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue);
eqe               312 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	return eqe;
eqe               123 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size);
eqe               124 drivers/net/ethernet/mellanox/mlx4/eq.c 	return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe;
eqe               129 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe *eqe =
eqe               131 drivers/net/ethernet/mellanox/mlx4/eq.c 	return (!!(eqe->owner & 0x80) ^
eqe               133 drivers/net/ethernet/mellanox/mlx4/eq.c 		eqe : NULL;
eqe               146 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe *eqe;
eqe               150 drivers/net/ethernet/mellanox/mlx4/eq.c 	for (eqe = next_slave_event_eqe(slave_eq); eqe;
eqe               151 drivers/net/ethernet/mellanox/mlx4/eq.c 	      eqe = next_slave_event_eqe(slave_eq)) {
eqe               152 drivers/net/ethernet/mellanox/mlx4/eq.c 		slave = eqe->slave_id;
eqe               154 drivers/net/ethernet/mellanox/mlx4/eq.c 		if (eqe->type == MLX4_EVENT_TYPE_PORT_CHANGE &&
eqe               155 drivers/net/ethernet/mellanox/mlx4/eq.c 		    eqe->subtype == MLX4_PORT_CHANGE_SUBTYPE_DOWN &&
eqe               169 drivers/net/ethernet/mellanox/mlx4/eq.c 				if (eqe->type == MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT &&
eqe               170 drivers/net/ethernet/mellanox/mlx4/eq.c 				    eqe->subtype == MLX4_DEV_PMC_SUBTYPE_PORT_INFO) {
eqe               171 drivers/net/ethernet/mellanox/mlx4/eq.c 					phys_port  = eqe->event.port_mgmt_change.port;
eqe               175 drivers/net/ethernet/mellanox/mlx4/eq.c 					eqe->event.port_mgmt_change.port = slave_port;
eqe               177 drivers/net/ethernet/mellanox/mlx4/eq.c 				if (mlx4_GEN_EQE(dev, i, eqe))
eqe               181 drivers/net/ethernet/mellanox/mlx4/eq.c 					eqe->event.port_mgmt_change.port = phys_port;
eqe               184 drivers/net/ethernet/mellanox/mlx4/eq.c 			if (mlx4_GEN_EQE(dev, slave, eqe))
eqe               194 drivers/net/ethernet/mellanox/mlx4/eq.c static void slave_event(struct mlx4_dev *dev, u8 slave, struct mlx4_eqe *eqe)
eqe               211 drivers/net/ethernet/mellanox/mlx4/eq.c 	memcpy(s_eqe, eqe, sizeof(struct mlx4_eqe) - 1);
eqe               224 drivers/net/ethernet/mellanox/mlx4/eq.c 			     struct mlx4_eqe *eqe)
eqe               233 drivers/net/ethernet/mellanox/mlx4/eq.c 	slave_event(dev, slave, eqe);
eqe               255 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe eqe;
eqe               263 drivers/net/ethernet/mellanox/mlx4/eq.c 	memset(&eqe, 0, sizeof(eqe));
eqe               265 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT;
eqe               266 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.subtype = MLX4_DEV_PMC_SUBTYPE_PKEY_TABLE;
eqe               267 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.event.port_mgmt_change.port = mlx4_phys_to_slave_port(dev, slave, port);
eqe               269 drivers/net/ethernet/mellanox/mlx4/eq.c 	return mlx4_GEN_EQE(dev, slave, &eqe);
eqe               275 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe eqe;
eqe               280 drivers/net/ethernet/mellanox/mlx4/eq.c 	memset(&eqe, 0, sizeof(eqe));
eqe               282 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT;
eqe               283 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.subtype = MLX4_DEV_PMC_SUBTYPE_GUID_INFO;
eqe               284 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.event.port_mgmt_change.port = mlx4_phys_to_slave_port(dev, slave, port);
eqe               286 drivers/net/ethernet/mellanox/mlx4/eq.c 	return mlx4_GEN_EQE(dev, slave, &eqe);
eqe               293 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe eqe;
eqe               299 drivers/net/ethernet/mellanox/mlx4/eq.c 	memset(&eqe, 0, sizeof(eqe));
eqe               301 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.type = MLX4_EVENT_TYPE_PORT_CHANGE;
eqe               302 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.subtype = port_subtype_change;
eqe               303 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.event.port_change.port = cpu_to_be32(slave_port << 28);
eqe               307 drivers/net/ethernet/mellanox/mlx4/eq.c 	return mlx4_GEN_EQE(dev, slave, &eqe);
eqe               434 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe eqe;
eqe               436 drivers/net/ethernet/mellanox/mlx4/eq.c 	memset(&eqe, 0, sizeof(eqe));
eqe               438 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.type = MLX4_EVENT_TYPE_PORT_MNG_CHG_EVENT;
eqe               439 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.subtype = MLX4_DEV_PMC_SUBTYPE_PORT_INFO;
eqe               440 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.event.port_mgmt_change.port = port;
eqe               441 drivers/net/ethernet/mellanox/mlx4/eq.c 	eqe.event.port_mgmt_change.params.port_info.changed_attr =
eqe               444 drivers/net/ethernet/mellanox/mlx4/eq.c 	slave_event(dev, ALL_SLAVES, &eqe);
eqe               497 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe *eqe;
eqe               512 drivers/net/ethernet/mellanox/mlx4/eq.c 	while ((eqe = next_eqe_sw(eq, dev->caps.eqe_factor, eqe_size))) {
eqe               519 drivers/net/ethernet/mellanox/mlx4/eq.c 		switch (eqe->type) {
eqe               521 drivers/net/ethernet/mellanox/mlx4/eq.c 			cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff;
eqe               533 drivers/net/ethernet/mellanox/mlx4/eq.c 			mlx4_dbg(dev, "event %d arrived\n", eqe->type);
eqe               538 drivers/net/ethernet/mellanox/mlx4/eq.c 						be32_to_cpu(eqe->event.qp.qpn)
eqe               542 drivers/net/ethernet/mellanox/mlx4/eq.c 						 eqe->type, eqe->subtype,
eqe               548 drivers/net/ethernet/mellanox/mlx4/eq.c 					mlx4_slave_event(dev, slave, eqe);
eqe               553 drivers/net/ethernet/mellanox/mlx4/eq.c 			mlx4_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) &
eqe               554 drivers/net/ethernet/mellanox/mlx4/eq.c 				      0xffffff, eqe->type);
eqe               559 drivers/net/ethernet/mellanox/mlx4/eq.c 				 __func__, be32_to_cpu(eqe->event.srq.srqn),
eqe               567 drivers/net/ethernet/mellanox/mlx4/eq.c 						be32_to_cpu(eqe->event.srq.srqn)
eqe               572 drivers/net/ethernet/mellanox/mlx4/eq.c 						  eqe->type, eqe->subtype,
eqe               576 drivers/net/ethernet/mellanox/mlx4/eq.c 				if (eqe->type ==
eqe               580 drivers/net/ethernet/mellanox/mlx4/eq.c 						  be32_to_cpu(eqe->event.srq.srqn),
eqe               581 drivers/net/ethernet/mellanox/mlx4/eq.c 						  eqe->type, eqe->subtype);
eqe               584 drivers/net/ethernet/mellanox/mlx4/eq.c 					if (eqe->type ==
eqe               587 drivers/net/ethernet/mellanox/mlx4/eq.c 							  __func__, eqe->type,
eqe               588 drivers/net/ethernet/mellanox/mlx4/eq.c 							  eqe->subtype, slave);
eqe               589 drivers/net/ethernet/mellanox/mlx4/eq.c 					mlx4_slave_event(dev, slave, eqe);
eqe               593 drivers/net/ethernet/mellanox/mlx4/eq.c 			mlx4_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) &
eqe               594 drivers/net/ethernet/mellanox/mlx4/eq.c 				       0xffffff, eqe->type);
eqe               599 drivers/net/ethernet/mellanox/mlx4/eq.c 				       be16_to_cpu(eqe->event.cmd.token),
eqe               600 drivers/net/ethernet/mellanox/mlx4/eq.c 				       eqe->event.cmd.status,
eqe               601 drivers/net/ethernet/mellanox/mlx4/eq.c 				       be64_to_cpu(eqe->event.cmd.out_param));
eqe               606 drivers/net/ethernet/mellanox/mlx4/eq.c 			port = be32_to_cpu(eqe->event.port_change.port) >> 28;
eqe               608 drivers/net/ethernet/mellanox/mlx4/eq.c 			if (eqe->subtype == MLX4_PORT_CHANGE_SUBTYPE_DOWN) {
eqe               627 drivers/net/ethernet/mellanox/mlx4/eq.c 							eqe->event.port_change.port =
eqe               629 drivers/net/ethernet/mellanox/mlx4/eq.c 								(be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF)
eqe               631 drivers/net/ethernet/mellanox/mlx4/eq.c 							mlx4_slave_event(dev, i, eqe);
eqe               641 drivers/net/ethernet/mellanox/mlx4/eq.c 							eqe->event.port_change.port =
eqe               643 drivers/net/ethernet/mellanox/mlx4/eq.c 								(be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF)
eqe               645 drivers/net/ethernet/mellanox/mlx4/eq.c 							mlx4_slave_event(dev, i, eqe);
eqe               668 drivers/net/ethernet/mellanox/mlx4/eq.c 							eqe->event.port_change.port =
eqe               670 drivers/net/ethernet/mellanox/mlx4/eq.c 								(be32_to_cpu(eqe->event.port_change.port) & 0xFFFFFFF)
eqe               672 drivers/net/ethernet/mellanox/mlx4/eq.c 							mlx4_slave_event(dev, i, eqe);
eqe               686 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eqe->event.cq_err.syndrome == 1 ?
eqe               688 drivers/net/ethernet/mellanox/mlx4/eq.c 				  be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff);
eqe               692 drivers/net/ethernet/mellanox/mlx4/eq.c 					be32_to_cpu(eqe->event.cq_err.cqn)
eqe               696 drivers/net/ethernet/mellanox/mlx4/eq.c 						 eqe->type, eqe->subtype,
eqe               702 drivers/net/ethernet/mellanox/mlx4/eq.c 					mlx4_slave_event(dev, slave, eqe);
eqe               707 drivers/net/ethernet/mellanox/mlx4/eq.c 				      be32_to_cpu(eqe->event.cq_err.cqn)
eqe               709 drivers/net/ethernet/mellanox/mlx4/eq.c 				      eqe->type);
eqe               730 drivers/net/ethernet/mellanox/mlx4/eq.c 			       eqe->event.comm_channel_arm.bit_vec,
eqe               731 drivers/net/ethernet/mellanox/mlx4/eq.c 			       sizeof(eqe->event.comm_channel_arm.bit_vec));
eqe               737 drivers/net/ethernet/mellanox/mlx4/eq.c 			flr_slave = be32_to_cpu(eqe->event.flr_event.slave_id);
eqe               767 drivers/net/ethernet/mellanox/mlx4/eq.c 			if (eqe->subtype == MLX4_FATAL_WARNING_SUBTYPE_WARMING) {
eqe               774 drivers/net/ethernet/mellanox/mlx4/eq.c 						mlx4_slave_event(dev, i, eqe);
eqe               777 drivers/net/ethernet/mellanox/mlx4/eq.c 					 be16_to_cpu(eqe->event.warming.warning_threshold),
eqe               778 drivers/net/ethernet/mellanox/mlx4/eq.c 					 be16_to_cpu(eqe->event.warming.current_temperature));
eqe               781 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eqe->type, eqe->subtype, eq->eqn,
eqe               782 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eq->cons_index, eqe->owner, eq->nent,
eqe               783 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eqe->slave_id,
eqe               784 drivers/net/ethernet/mellanox/mlx4/eq.c 					  !!(eqe->owner & 0x80) ^
eqe               791 drivers/net/ethernet/mellanox/mlx4/eq.c 					    (unsigned long) eqe);
eqe               795 drivers/net/ethernet/mellanox/mlx4/eq.c 			switch (eqe->subtype) {
eqe               798 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eqe->event.bad_cable.port);
eqe               806 drivers/net/ethernet/mellanox/mlx4/eq.c 					 eqe->type, eqe->subtype, eq->eqn,
eqe               807 drivers/net/ethernet/mellanox/mlx4/eq.c 					 eq->cons_index, eqe->owner, eq->nent,
eqe               808 drivers/net/ethernet/mellanox/mlx4/eq.c 					 !!(eqe->owner & 0x80) ^
eqe               818 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eqe->type, eqe->subtype, eq->eqn,
eqe               819 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eq->cons_index, eqe->owner, eq->nent,
eqe               820 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eqe->slave_id,
eqe               821 drivers/net/ethernet/mellanox/mlx4/eq.c 				  !!(eqe->owner & 0x80) ^
eqe              1198 drivers/net/ethernet/mellanox/mlx4/mlx4.h int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe);
eqe              3338 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe)
eqe              3357 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	event_eq = &priv->mfunc.master.slave_state[slave].event_eq[eqe->type];
eqe              3380 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (eqe->type == MLX4_EVENT_TYPE_CMD) {
eqe              3382 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		eqe->event.cmd.token = cpu_to_be16(event_eq->token);
eqe              3385 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	memcpy(mailbox->buf, (u8 *) eqe, 28);
eqe              1429 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	struct mlx5_eqe *eqe;
eqe              1433 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	eqe = data;
eqe              1435 drivers/net/ethernet/mellanox/mlx5/core/cmd.c 	mlx5_cmd_comp_handler(dev, be32_to_cpu(eqe->data.cmd.vector), false);
eqe                72 drivers/net/ethernet/mellanox/mlx5/core/cq.c 				   struct mlx5_eqe *eqe)
eqe              1071 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c 	struct mlx5_eqe *eqe = data;
eqe              1073 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c 	switch (eqe->sub_type) {
eqe              1084 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c 			      eqe->sub_type);
eqe               914 drivers/net/ethernet/mellanox/mlx5/core/en.h void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe);
eqe                61 drivers/net/ethernet/mellanox/mlx5/core/en/monitor_stats.c 				       unsigned long event, void *eqe)
eqe               204 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	struct mlx5_eqe   *eqe = data;
eqe               209 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	switch (eqe->sub_type) {
eqe              1699 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 		struct mlx5_eqe *eqe = data;
eqe              1701 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c 		switch (eqe->sub_type) {
eqe               199 drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe)
eqe               135 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eqe *eqe;
eqe               139 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = next_eqe_sw(eq);
eqe               140 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	if (!eqe)
eqe               151 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff;
eqe               156 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			cq->comp(cq, eqe);
eqe               164 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	} while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq)));
eqe               201 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eqe *eqe;
eqe               207 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = next_eqe_sw(eq);
eqe               208 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	if (!eqe)
eqe               218 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		atomic_notifier_call_chain(&eqt->nh[eqe->type], eqe->type, eqe);
eqe               219 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		atomic_notifier_call_chain(&eqt->nh[MLX5_EVENT_TYPE_NOTIFY_ANY], eqe->type, eqe);
eqe               223 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	} while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq)));
eqe               233 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eqe *eqe;
eqe               237 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eqe = get_eqe(eq, i);
eqe               238 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eqe->owner = MLX5_EQE_OWNER_INIT_VAL;
eqe               479 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eqe *eqe;
eqe               487 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = data;
eqe               489 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	cqn = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff;
eqe               491 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		       cqn, eqe->data.cq_err.syndrome);
eqe               736 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eqe *eqe;
eqe               738 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = get_eqe(eq, ci & (eq->nent - 1));
eqe               739 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = ((eqe->owner & 1) ^ !!(ci & eq->nent)) ? NULL : eqe;
eqe               743 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	if (eqe)
eqe               746 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	return eqe;
eqe              1767 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	struct mlx5_eqe *eqe = data;
eqe              1771 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	vport_num = be16_to_cpu(eqe->data.vport_change.vport_num);
eqe               142 drivers/net/ethernet/mellanox/mlx5/core/events.c 	struct mlx5_eqe      *eqe      = data;
eqe               145 drivers/net/ethernet/mellanox/mlx5/core/events.c 		      eqe_type_str(eqe->type), eqe->sub_type);
eqe               154 drivers/net/ethernet/mellanox/mlx5/core/events.c 	struct mlx5_eqe      *eqe      = data;
eqe               158 drivers/net/ethernet/mellanox/mlx5/core/events.c 	value_lsb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_lsb);
eqe               159 drivers/net/ethernet/mellanox/mlx5/core/events.c 	value_msb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_msb);
eqe               216 drivers/net/ethernet/mellanox/mlx5/core/events.c 	struct mlx5_eqe      *eqe      = data;
eqe               224 drivers/net/ethernet/mellanox/mlx5/core/events.c 	module_event_eqe = &eqe->data.port_module;
eqe               307 drivers/net/ethernet/mellanox/mlx5/core/events.c 	struct mlx5_eqe         *eqe      = data;
eqe               309 drivers/net/ethernet/mellanox/mlx5/core/events.c 	switch (eqe->sub_type) {
eqe               330 drivers/net/ethernet/mellanox/mlx5/core/events.c 	struct mlx5_eqe      *eqe      = data;
eqe               333 drivers/net/ethernet/mellanox/mlx5/core/events.c 		      eqe_type_str(eqe->type), eqe->sub_type);
eqe               418 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 				       struct mlx5_eqe *eqe)
eqe               151 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c static int fpga_err_event(struct notifier_block *nb, unsigned long event, void *eqe)
eqe               155 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c 	return mlx5_fpga_event(fdev, event, eqe);
eqe               158 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c static int fpga_qp_err_event(struct notifier_block *nb, unsigned long event, void *eqe)
eqe               162 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c 	return mlx5_fpga_event(fdev, event, eqe);
eqe               314 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c 			   unsigned long event, void *eqe)
eqe               316 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.c 	void *data = ((struct mlx5_eqe *)eqe)->data.raw;
eqe               472 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	struct mlx5_eqe *eqe = data;
eqe               473 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	int pin = eqe->data.pps.pin;
eqe               481 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 					be64_to_cpu(eqe->data.pps.time_stamp));
eqe                57 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1));
eqe                59 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe;
eqe               458 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	struct mlx5_eqe *eqe;
eqe               465 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	eqe  = data;
eqe               467 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	func_id = be16_to_cpu(eqe->data.req_pages.func_id);
eqe               468 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	npages  = be32_to_cpu(eqe->data.req_pages.num_pages);
eqe               469 drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c 	ec_function = be16_to_cpu(eqe->data.req_pages.ec_function) & EC_FUNCTION_MASK;
eqe               132 drivers/net/ethernet/mellanox/mlx5/core/qp.c 	struct mlx5_eqe *eqe;
eqe               137 drivers/net/ethernet/mellanox/mlx5/core/qp.c 		eqe = data;
eqe               138 drivers/net/ethernet/mellanox/mlx5/core/qp.c 		rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff;
eqe               149 drivers/net/ethernet/mellanox/mlx5/core/qp.c 		eqe = data;
eqe               150 drivers/net/ethernet/mellanox/mlx5/core/qp.c 		rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff;
eqe               151 drivers/net/ethernet/mellanox/mlx5/core/qp.c 		rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN);
eqe               161 drivers/net/ethernet/mellanox/mlx5/core/qp.c 	mlx5_core_dbg(dev, "event (%d) arrived on resource 0x%x\n", eqe->type, rsn);
eqe               693 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 			   struct mlx5_eqe *eqe)
eqe               694 drivers/net/ethernet/mellanox/mlxsw/pci.c static void mlxsw_pci_eq_cmd_event(struct mlxsw_pci *mlxsw_pci, char *eqe)
eqe               696 drivers/net/ethernet/mellanox/mlxsw/pci.c 	mlxsw_pci->cmd.comp.status = mlxsw_pci_eqe_cmd_status_get(eqe);
eqe               698 drivers/net/ethernet/mellanox/mlxsw/pci.c 		((u64) mlxsw_pci_eqe_cmd_out_param_h_get(eqe)) << 32 |
eqe               699 drivers/net/ethernet/mellanox/mlxsw/pci.c 		mlxsw_pci_eqe_cmd_out_param_l_get(eqe);
eqe               726 drivers/net/ethernet/mellanox/mlxsw/pci.c 	char *eqe;
eqe               734 drivers/net/ethernet/mellanox/mlxsw/pci.c 	while ((eqe = mlxsw_pci_eq_sw_eqe_get(q))) {
eqe               742 drivers/net/ethernet/mellanox/mlxsw/pci.c 			mlxsw_pci_eq_cmd_event(mlxsw_pci, eqe);
eqe               746 drivers/net/ethernet/mellanox/mlxsw/pci.c 			cqn = mlxsw_pci_eqe_cqn_get(eqe);
eqe               218 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, event_type, 0x0C, 24, 8);
eqe               225 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, event_sub_type, 0x0C, 16, 8);
eqe               230 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cqn, 0x0C, 8, 7);
eqe               235 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, owner, 0x0C, 0, 1);
eqe               240 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_token, 0x00, 16, 16);
eqe               245 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_status, 0x00, 0, 8);
eqe               250 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x04, 0, 32);
eqe               255 drivers/net/ethernet/mellanox/mlxsw/pci_hw.h MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x08, 0, 32);
eqe               668 drivers/scsi/be2iscsi/be_main.c 	struct be_eq_entry *eqe;
eqe               678 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eqe               681 drivers/scsi/be2iscsi/be_main.c 	while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
eqe               683 drivers/scsi/be2iscsi/be_main.c 		if (((eqe->dw[offsetof(struct amap_eq_entry,
eqe               688 drivers/scsi/be2iscsi/be_main.c 		AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
eqe               690 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eqe               732 drivers/scsi/be2iscsi/be_main.c 	struct be_eq_entry *eqe;
eqe               753 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eqe               757 drivers/scsi/be2iscsi/be_main.c 	while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
eqe               759 drivers/scsi/be2iscsi/be_main.c 		if (((eqe->dw[offsetof(struct amap_eq_entry,
eqe               764 drivers/scsi/be2iscsi/be_main.c 		AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
eqe               766 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eqe              2037 drivers/scsi/be2iscsi/be_main.c 	struct be_eq_entry *eqe = NULL;
eqe              2049 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eqe              2050 drivers/scsi/be2iscsi/be_main.c 	while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] &
eqe              2052 drivers/scsi/be2iscsi/be_main.c 		AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
eqe              2054 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eqe              3587 drivers/scsi/be2iscsi/be_main.c 	struct be_eq_entry *eqe = NULL;
eqe              3603 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eqe              3605 drivers/scsi/be2iscsi/be_main.c 		while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
eqe              3607 drivers/scsi/be2iscsi/be_main.c 			AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
eqe              3609 drivers/scsi/be2iscsi/be_main.c 			eqe = queue_tail_node(eq);
eqe                87 drivers/scsi/lpfc/lpfc_sli.c 				     struct lpfc_eqe *eqe);
eqe               326 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_eqe *eqe;
eqe               331 drivers/scsi/lpfc/lpfc_sli.c 	eqe = lpfc_sli4_qe(q, q->host_index);
eqe               334 drivers/scsi/lpfc/lpfc_sli.c 	if (bf_get_le32(lpfc_eqe_valid, eqe) != q->qe_valid)
eqe               347 drivers/scsi/lpfc/lpfc_sli.c 	return eqe;
eqe               457 drivers/scsi/lpfc/lpfc_sli.c 			struct lpfc_eqe *eqe)
eqe               460 drivers/scsi/lpfc/lpfc_sli.c 		bf_set_le32(lpfc_eqe_valid, eqe, 0);
eqe               472 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_eqe *eqe;
eqe               476 drivers/scsi/lpfc/lpfc_sli.c 	eqe = lpfc_sli4_eq_get(eq);
eqe               477 drivers/scsi/lpfc/lpfc_sli.c 	while (eqe) {
eqe               478 drivers/scsi/lpfc/lpfc_sli.c 		__lpfc_sli4_consume_eqe(phba, eq, eqe);
eqe               480 drivers/scsi/lpfc/lpfc_sli.c 		eqe = lpfc_sli4_eq_get(eq);
eqe               491 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_eqe *eqe;
eqe               497 drivers/scsi/lpfc/lpfc_sli.c 	eqe = lpfc_sli4_eq_get(eq);
eqe               498 drivers/scsi/lpfc/lpfc_sli.c 	while (eqe) {
eqe               499 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli4_hba_handle_eqe(phba, eq, eqe);
eqe               500 drivers/scsi/lpfc/lpfc_sli.c 		__lpfc_sli4_consume_eqe(phba, eq, eqe);
eqe               512 drivers/scsi/lpfc/lpfc_sli.c 		eqe = lpfc_sli4_eq_get(eq);
eqe              13532 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_sp_handle_eqe(struct lpfc_hba *phba, struct lpfc_eqe *eqe,
eqe              13539 drivers/scsi/lpfc/lpfc_sli.c 	cqid = bf_get_le32(lpfc_eqe_resource_id, eqe);
eqe              14044 drivers/scsi/lpfc/lpfc_sli.c 			 struct lpfc_eqe *eqe)
eqe              14050 drivers/scsi/lpfc/lpfc_sli.c 	if (unlikely(bf_get_le32(lpfc_eqe_major_code, eqe) != 0)) {
eqe              14054 drivers/scsi/lpfc/lpfc_sli.c 				bf_get_le32(lpfc_eqe_major_code, eqe),
eqe              14055 drivers/scsi/lpfc/lpfc_sli.c 				bf_get_le32(lpfc_eqe_minor_code, eqe));
eqe              14060 drivers/scsi/lpfc/lpfc_sli.c 	cqid = bf_get_le32(lpfc_eqe_resource_id, eqe);
eqe              14087 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli4_sp_handle_eqe(phba, eqe,
eqe                50 include/linux/mlx5/cq.h 	void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
eqe                58 include/linux/mlx5/cq.h 		void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
eqe              6445 include/linux/mlx5/mlx5_ifc.h 	u8         eqe[64][0x8];