| /linux-4.4.14/drivers/staging/rdma/ipath/ |
| D | ipath_cq.c | 63 if (head >= (unsigned) cq->ibcq.cqe) { in ipath_cq_enter() 64 head = cq->ibcq.cqe; in ipath_cq_enter() 70 if (cq->ibcq.event_handler) { in ipath_cq_enter() 73 ev.device = cq->ibcq.device; in ipath_cq_enter() 74 ev.element.cq = &cq->ibcq; in ipath_cq_enter() 76 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in ipath_cq_enter() 115 to_idev(cq->ibcq.device)->n_wqe_errs++; in ipath_cq_enter() 129 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in ipath_poll_cq() argument 131 struct ipath_cq *cq = to_icq(ibcq); in ipath_poll_cq() 147 if (tail > (u32) cq->ibcq.cqe) in ipath_poll_cq() [all …]
|
| D | ipath_verbs.h | 215 struct ib_cq ibcq; member 676 static inline struct ipath_cq *to_icq(struct ib_cq *ibcq) in to_icq() argument 678 return container_of(ibcq, struct ipath_cq, ibcq); in to_icq() 816 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 823 int ipath_destroy_cq(struct ib_cq *ibcq); 825 int ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 827 int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_cq.c | 66 if (head >= (unsigned) cq->ibcq.cqe) { in qib_cq_enter() 67 head = cq->ibcq.cqe; in qib_cq_enter() 73 if (cq->ibcq.event_handler) { in qib_cq_enter() 76 ev.device = cq->ibcq.device; in qib_cq_enter() 77 ev.element.cq = &cq->ibcq; in qib_cq_enter() 79 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in qib_cq_enter() 136 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in qib_poll_cq() argument 138 struct qib_cq *cq = to_icq(ibcq); in qib_poll_cq() 154 if (tail > (u32) cq->ibcq.cqe) in qib_poll_cq() 155 tail = (u32) cq->ibcq.cqe; in qib_poll_cq() [all …]
|
| D | qib_verbs.h | 270 struct ib_cq ibcq; member 831 static inline struct qib_cq *to_icq(struct ib_cq *ibcq) in to_icq() argument 833 return container_of(ibcq, struct qib_cq, ibcq); in to_icq() 1020 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 1027 int qib_destroy_cq(struct ib_cq *ibcq); 1029 int qib_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 1031 int qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
| /linux-4.4.14/drivers/staging/rdma/hfi1/ |
| D | cq.c | 82 if (head >= (unsigned) cq->ibcq.cqe) { in hfi1_cq_enter() 83 head = cq->ibcq.cqe; in hfi1_cq_enter() 89 if (cq->ibcq.event_handler) { in hfi1_cq_enter() 92 ev.device = cq->ibcq.device; in hfi1_cq_enter() 93 ev.element.cq = &cq->ibcq; in hfi1_cq_enter() 95 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in hfi1_cq_enter() 152 int hfi1_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in hfi1_poll_cq() argument 154 struct hfi1_cq *cq = to_icq(ibcq); in hfi1_poll_cq() 170 if (tail > (u32) cq->ibcq.cqe) in hfi1_poll_cq() 171 tail = (u32) cq->ibcq.cqe; in hfi1_poll_cq() [all …]
|
| D | verbs.h | 282 struct ib_cq ibcq; member 815 static inline struct hfi1_cq *to_icq(struct ib_cq *ibcq) in to_icq() argument 817 return container_of(ibcq, struct hfi1_cq, ibcq); in to_icq() 997 int hfi1_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 1005 int hfi1_destroy_cq(struct ib_cq *ibcq); 1008 struct ib_cq *ibcq, 1011 int hfi1_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
| /linux-4.4.14/drivers/infiniband/hw/mlx4/ |
| D | cq.c | 44 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 51 struct ib_cq *ibcq; in mlx4_ib_cq_event() local 59 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 60 if (ibcq->event_handler) { in mlx4_ib_cq_event() 61 event.device = ibcq->device; in mlx4_ib_cq_event() 63 event.element.cq = ibcq; in mlx4_ib_cq_event() 64 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event() 80 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 84 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe() [all …]
|
| D | mlx4_ib.h | 117 struct ib_cq ibcq; member 630 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 632 return container_of(ibcq, struct mlx4_ib_cq, ibcq); in to_mcq() 717 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 723 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
| /linux-4.4.14/drivers/infiniband/hw/mthca/ |
| D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 259 event.element.cq = &cq->ibcq; in mthca_cq_event() 260 if (cq->ibcq.event_handler) in mthca_cq_event() 261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 295 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean() 297 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 309 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 315 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean() 321 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() [all …]
|
| D | mthca_provider.h | 202 struct ib_cq ibcq; member 324 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 326 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
|
| D | mthca_provider.c | 712 return &cq->ibcq; in mthca_create_cq() 775 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) in mthca_resize_cq() argument 777 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_resize_cq() 778 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq() 789 if (entries == ibcq->cqe + 1) { in mthca_resize_cq() 829 tcqe = cq->ibcq.cqe; in mthca_resize_cq() 831 cq->ibcq.cqe = cq->resize_buf->cqe; in mthca_resize_cq() 843 ibcq->cqe = entries - 1; in mthca_resize_cq()
|
| D | mthca_dev.h | 494 int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
|
| /linux-4.4.14/drivers/infiniband/hw/mlx5/ |
| D | cq.c | 42 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() local 44 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 50 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 51 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() local 60 if (ibcq->event_handler) { in mlx5_ib_cq_event() 63 event.element.cq = ibcq; in mlx5_ib_cq_event() 64 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 85 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 91 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe() 414 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() [all …]
|
| D | mlx5_ib.h | 272 struct ib_cq ibcq; member 451 static inline struct mlx5_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 453 return container_of(ibcq, struct mlx5_ib_cq, ibcq); in to_mcq() 544 int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); 545 int mlx5_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags); 547 int mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 600 int mlx5_ib_get_cqe_size(struct mlx5_ib_dev *dev, struct ib_cq *ibcq);
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
| D | iwch_ev.c | 89 event.device = chp->ibcq.device; in post_qp_event() 91 event.element.cq = &chp->ibcq; in post_qp_event() 99 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 178 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch()
|
| D | iwch_provider.h | 104 struct ib_cq ibcq; member 114 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) in to_iwch_cq() argument 116 return container_of(ibcq, struct iwch_cq, ibcq); in to_iwch_cq() 336 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
| D | iwch_qp.c | 829 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in __flush_qp() 843 (*schp->ibcq.comp_handler)(&schp->ibcq, schp->ibcq.cq_context); in __flush_qp() 865 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 870 (*schp->ibcq.comp_handler)(&schp->ibcq, in flush_qp() 871 schp->ibcq.cq_context); in flush_qp()
|
| D | iwch_cq.c | 195 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in iwch_poll_cq() argument 203 chp = to_iwch_cq(ibcq); in iwch_poll_cq()
|
| D | iwch_provider.c | 200 chp->ibcq.cqe = 1 << chp->cq.size_log2; in iwch_create_cq() 216 iwch_destroy_cq(&chp->ibcq); in iwch_create_cq() 243 iwch_destroy_cq(&chp->ibcq); in iwch_create_cq() 251 return &chp->ibcq; in iwch_create_cq() 302 chp->ibcq.cqe = (1<<chp->cq.size_log2) - 1; in iwch_resize_cq() 322 static int iwch_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in iwch_arm_cq() argument 331 chp = to_iwch_cq(ibcq); in iwch_arm_cq()
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
| D | ev.c | 105 event.device = chp->ibcq.device; in post_qp_event() 107 event.element.cq = &chp->ibcq; in post_qp_event() 114 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
| D | qp.c | 1156 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp() 1157 rchp->ibcq.cq_context); in __flush_qp() 1163 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp() 1164 rchp->ibcq.cq_context); in __flush_qp() 1169 (*schp->ibcq.comp_handler)(&schp->ibcq, in __flush_qp() 1170 schp->ibcq.cq_context); in __flush_qp() 1188 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 1193 (*schp->ibcq.comp_handler)(&schp->ibcq, in flush_qp() 1194 schp->ibcq.cq_context); in flush_qp()
|
| D | iw_cxgb4.h | 413 struct ib_cq ibcq; member 422 static inline struct c4iw_cq *to_c4iw_cq(struct ib_cq *ibcq) in to_c4iw_cq() argument 424 return container_of(ibcq, struct c4iw_cq, ibcq); in to_c4iw_cq() 945 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); 989 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
|
| D | cq.c | 826 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in c4iw_poll_cq() argument 833 chp = to_c4iw_cq(ibcq); in c4iw_poll_cq() 939 chp->ibcq.cqe = entries - 2; in c4iw_create_cq() 984 return &chp->ibcq; in c4iw_create_cq() 1004 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in c4iw_arm_cq() argument 1010 chp = to_c4iw_cq(ibcq); in c4iw_arm_cq()
|
| /linux-4.4.14/drivers/staging/rdma/amso1100/ |
| D | c2_provider.h | 92 struct ib_cq ibcq; member 153 static inline struct c2_cq *to_c2cq(struct ib_cq *ibcq) in to_c2cq() argument 155 return container_of(ibcq, struct c2_cq, ibcq); in to_c2cq()
|
| D | c2_cq.c | 78 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in c2_cq_event() 201 int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in c2_poll_cq() argument 203 struct c2_dev *c2dev = to_c2dev(ibcq->device); in c2_poll_cq() 204 struct c2_cq *cq = to_c2cq(ibcq); in c2_poll_cq() 222 int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) in c2_arm_cq() argument 229 cq = to_c2cq(ibcq); in c2_arm_cq() 300 cq->ibcq.cqe = entries - 1; in c2_init_cq()
|
| D | c2_ae.c | 310 ib_event.element.cq = &cq->ibcq; in c2_ae_event() 313 if (cq->ibcq.event_handler) in c2_ae_event() 314 cq->ibcq.event_handler(&ib_event, in c2_ae_event() 315 cq->ibcq.cq_context); in c2_ae_event()
|
| D | c2.h | 517 extern int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 518 extern int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
|
| D | c2_provider.c | 317 return &cq->ibcq; in c2_create_cq()
|
| /linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma.h | 317 struct ib_cq ibcq; member 474 static inline struct ocrdma_cq *get_ocrdma_cq(struct ib_cq *ibcq) in get_ocrdma_cq() argument 476 return container_of(ibcq, struct ocrdma_cq, ibcq); in get_ocrdma_cq()
|
| D | ocrdma_verbs.c | 1116 return &cq->ibcq; in ocrdma_create_cq() 1124 int ocrdma_resize_cq(struct ib_cq *ibcq, int new_cnt, in ocrdma_resize_cq() argument 1128 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); in ocrdma_resize_cq() 1134 ibcq->cqe = new_cnt; in ocrdma_resize_cq() 1144 struct ocrdma_dev *dev = get_ocrdma_dev(cq->ibcq.device); in ocrdma_flush_cq() 1164 int ocrdma_destroy_cq(struct ib_cq *ibcq) in ocrdma_destroy_cq() argument 1166 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); in ocrdma_destroy_cq() 1168 struct ocrdma_dev *dev = get_ocrdma_dev(ibcq->device); in ocrdma_destroy_cq() 2870 struct ocrdma_dev *dev = get_ocrdma_dev(cq->ibcq.device); in ocrdma_poll_hwcq() 2951 int ocrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in ocrdma_poll_cq() argument [all …]
|
| D | ocrdma_hw.c | 718 ib_evt.element.cq = &cq->ibcq; in ocrdma_dispatch_ibevent() 724 ib_evt.element.cq = &cq->ibcq; in ocrdma_dispatch_ibevent() 784 if (cq->ibcq.event_handler) in ocrdma_dispatch_ibevent() 785 cq->ibcq.event_handler(&ib_evt, cq->ibcq.cq_context); in ocrdma_dispatch_ibevent() 955 if (bcq && bcq->ibcq.comp_handler) { in ocrdma_qp_buddy_cq_handler() 957 (*bcq->ibcq.comp_handler) (&bcq->ibcq, bcq->ibcq.cq_context); in ocrdma_qp_buddy_cq_handler() 974 if (cq->ibcq.comp_handler) { in ocrdma_qp_cq_handler() 976 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in ocrdma_qp_cq_handler()
|
| /linux-4.4.14/drivers/staging/rdma/ehca/ |
| D | ehca_main.c | 554 struct ib_cq *ibcq; in ehca_create_aqp1() local 566 ibcq = ib_create_cq(&shca->ib_device, NULL, NULL, (void *)(-1), in ehca_create_aqp1() 568 if (IS_ERR(ibcq)) { in ehca_create_aqp1() 570 return PTR_ERR(ibcq); in ehca_create_aqp1() 572 sport->ibcq_aqp1 = ibcq; in ehca_create_aqp1() 581 qp_init_attr.send_cq = ibcq; in ehca_create_aqp1() 582 qp_init_attr.recv_cq = ibcq; in ehca_create_aqp1()
|
| /linux-4.4.14/drivers/infiniband/hw/nes/ |
| D | nes.h | 505 static inline struct nes_cq *to_nescq(struct ib_cq *ibcq) in to_nescq() argument 507 return container_of(ibcq, struct nes_cq, ibcq); in to_nescq()
|
| D | nes_verbs.h | 116 struct ib_cq ibcq; member
|
| D | nes_hw.c | 3734 if (nescq->ibcq.event_handler) { in nes_process_iwarp_aeqe() 3735 ibevent.device = nescq->ibcq.device; in nes_process_iwarp_aeqe() 3737 ibevent.element.cq = &nescq->ibcq; in nes_process_iwarp_aeqe() 3738 nescq->ibcq.event_handler(&ibevent, nescq->ibcq.cq_context); in nes_process_iwarp_aeqe() 3763 if (nescq->ibcq.comp_handler) in nes_iwarp_ce_handler() 3764 nescq->ibcq.comp_handler(&nescq->ibcq, nescq->ibcq.cq_context); in nes_iwarp_ce_handler()
|
| D | nes_verbs.c | 1556 nescq->ibcq.cqe = nescq->hw_cq.cq_size - 1; in nes_create_cq() 1764 return &nescq->ibcq; in nes_create_cq() 3608 static int nes_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in nes_poll_cq() argument 3613 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); in nes_poll_cq() 3615 struct nes_cq *nescq = to_nescq(ibcq); in nes_poll_cq() 3783 static int nes_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) in nes_req_notify_cq() argument 3785 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); in nes_req_notify_cq() 3787 struct nes_cq *nescq = to_nescq(ibcq); in nes_req_notify_cq()
|
| /linux-4.4.14/drivers/infiniband/hw/usnic/ |
| D | usnic_ib_verbs.h | 86 int usnic_ib_poll_cq(struct ib_cq *ibcq, int num_entries,
|
| D | usnic_ib_verbs.c | 769 int usnic_ib_poll_cq(struct ib_cq *ibcq, int num_entries, in usnic_ib_poll_cq() argument
|