Searched refs:err_cqe (Results 1 - 4 of 4) sorted by relevance

/linux-4.4.14/drivers/infiniband/hw/cxgb4/
H A Dev.c63 static void dump_err_cqe(struct c4iw_dev *dev, struct t4_cqe *err_cqe) dump_err_cqe() argument
65 __be64 *p = (void *)err_cqe; dump_err_cqe()
70 CQE_QPID(err_cqe), CQE_OPCODE(err_cqe), dump_err_cqe()
71 CQE_STATUS(err_cqe), CQE_TYPE(err_cqe), ntohl(err_cqe->len), dump_err_cqe()
72 CQE_WRID_HI(err_cqe), CQE_WRID_LOW(err_cqe)); dump_err_cqe()
82 if (RQ_TYPE(err_cqe) && (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE || dump_err_cqe()
83 CQE_OPCODE(err_cqe) == FW_RI_READ_RESP)) dump_err_cqe()
84 print_tpte(dev, CQE_WRID_STAG(err_cqe)); dump_err_cqe()
89 struct t4_cqe *err_cqe, post_qp_event()
96 dump_err_cqe(dev, err_cqe); post_qp_event()
118 void c4iw_ev_dispatch(struct c4iw_dev *dev, struct t4_cqe *err_cqe) c4iw_ev_dispatch() argument
125 qhp = get_qhp(dev, CQE_QPID(err_cqe)); c4iw_ev_dispatch()
129 CQE_QPID(err_cqe), c4iw_ev_dispatch()
130 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe), c4iw_ev_dispatch()
131 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe), c4iw_ev_dispatch()
132 CQE_WRID_LOW(err_cqe)); c4iw_ev_dispatch()
137 if (SQ_TYPE(err_cqe)) c4iw_ev_dispatch()
145 cqid, CQE_QPID(err_cqe), c4iw_ev_dispatch()
146 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe), c4iw_ev_dispatch()
147 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe), c4iw_ev_dispatch()
148 CQE_WRID_LOW(err_cqe)); c4iw_ev_dispatch()
158 if (RQ_TYPE(err_cqe) && c4iw_ev_dispatch()
159 (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE)) { c4iw_ev_dispatch()
160 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_REQ_ERR); c4iw_ev_dispatch()
164 switch (CQE_STATUS(err_cqe)) { c4iw_ev_dispatch()
179 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_ACCESS_ERR); c4iw_ev_dispatch()
186 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_DEVICE_FATAL); c4iw_ev_dispatch()
206 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL); c4iw_ev_dispatch()
211 CQE_STATUS(err_cqe), qhp->wq.sq.qid); c4iw_ev_dispatch()
212 post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL); c4iw_ev_dispatch()
87 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) post_qp_event() argument
H A Dqp.c945 static inline void build_term_codes(struct t4_cqe *err_cqe, u8 *layer_type, build_term_codes() argument
954 if (!err_cqe) { build_term_codes()
960 status = CQE_STATUS(err_cqe); build_term_codes()
961 opcode = CQE_OPCODE(err_cqe); build_term_codes()
962 rqtype = RQ_TYPE(err_cqe); build_term_codes()
1081 static void post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, post_terminate() argument
1110 build_term_codes(err_cqe, &term->layer_etype, &term->ecode); post_terminate()
H A Diw_cxgb4.h1013 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);
1020 void c4iw_ev_dispatch(struct c4iw_dev *dev, struct t4_cqe *err_cqe);
/linux-4.4.14/drivers/infiniband/hw/mlx5/
H A Dcq.c415 struct mlx5_err_cqe *err_cqe; mlx5_poll_one() local
494 err_cqe = (struct mlx5_err_cqe *)cqe64; mlx5_poll_one()
495 mlx5_handle_error_cqe(dev, err_cqe, wc); mlx5_poll_one()
500 err_cqe->syndrome, err_cqe->vendor_err_synd); mlx5_poll_one()

Completed in 181 milliseconds