qp_info 59 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info, qp_info 63 drivers/infiniband/core/mad.c struct ib_device *dev = qp_info->port_priv->device; qp_info 64 drivers/infiniband/core/mad.c u8 pnum = qp_info->port_priv->port_num; qp_info 103 drivers/infiniband/core/mad.c static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info, qp_info 365 drivers/infiniband/core/mad.c if (!port_priv->qp_info[qpn].qp) { qp_info 388 drivers/infiniband/core/mad.c mad_agent_priv->qp_info = &port_priv->qp_info[qpn]; qp_info 395 drivers/infiniband/core/mad.c mad_agent_priv->agent.qp = port_priv->qp_info[qpn].qp; qp_info 502 drivers/infiniband/core/mad.c static int register_snoop_agent(struct ib_mad_qp_info *qp_info, qp_info 509 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 511 drivers/infiniband/core/mad.c for (i = 0; i < qp_info->snoop_table_size; i++) qp_info 512 drivers/infiniband/core/mad.c if (!qp_info->snoop_table[i]) qp_info 515 drivers/infiniband/core/mad.c if (i == qp_info->snoop_table_size) { qp_info 517 drivers/infiniband/core/mad.c new_snoop_table = krealloc(qp_info->snoop_table, qp_info 519 drivers/infiniband/core/mad.c (qp_info->snoop_table_size + 1), qp_info 526 drivers/infiniband/core/mad.c qp_info->snoop_table = new_snoop_table; qp_info 527 drivers/infiniband/core/mad.c qp_info->snoop_table_size++; qp_info 529 drivers/infiniband/core/mad.c qp_info->snoop_table[i] = mad_snoop_priv; qp_info 530 drivers/infiniband/core/mad.c atomic_inc(&qp_info->snoop_count); qp_info 532 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 574 drivers/infiniband/core/mad.c mad_snoop_priv->qp_info = &port_priv->qp_info[qpn]; qp_info 579 drivers/infiniband/core/mad.c mad_snoop_priv->agent.qp = port_priv->qp_info[qpn].qp; qp_info 591 drivers/infiniband/core/mad.c &port_priv->qp_info[qpn], qp_info 633 drivers/infiniband/core/mad.c port_priv = mad_agent_priv->qp_info->port_priv; qp_info 655 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info; qp_info 658 drivers/infiniband/core/mad.c qp_info = mad_snoop_priv->qp_info; qp_info 659 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 660 drivers/infiniband/core/mad.c qp_info->snoop_table[mad_snoop_priv->snoop_index] = NULL; qp_info 661 drivers/infiniband/core/mad.c atomic_dec(&qp_info->snoop_count); qp_info 662 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 709 drivers/infiniband/core/mad.c static void snoop_send(struct ib_mad_qp_info *qp_info, qp_info 718 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 719 drivers/infiniband/core/mad.c for (i = 0; i < qp_info->snoop_table_size; i++) { qp_info 720 drivers/infiniband/core/mad.c mad_snoop_priv = qp_info->snoop_table[i]; qp_info 726 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 730 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 732 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 735 drivers/infiniband/core/mad.c static void snoop_recv(struct ib_mad_qp_info *qp_info, qp_info 743 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 744 drivers/infiniband/core/mad.c for (i = 0; i < qp_info->snoop_table_size; i++) { qp_info 745 drivers/infiniband/core/mad.c mad_snoop_priv = qp_info->snoop_table[i]; qp_info 751 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 755 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->snoop_lock, flags); qp_info 757 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->snoop_lock, flags); qp_info 823 drivers/infiniband/core/mad.c size_t mad_size = port_mad_size(mad_agent_priv->qp_info->port_priv); qp_info 826 drivers/infiniband/core/mad.c bool opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device, qp_info 827 drivers/infiniband/core/mad.c mad_agent_priv->qp_info->port_priv->port_num); qp_info 975 drivers/infiniband/core/mad.c queue_work(mad_agent_priv->qp_info->port_priv->wq, qp_info 1212 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info; qp_info 1220 drivers/infiniband/core/mad.c qp_info = mad_send_wr->mad_agent_priv->qp_info; qp_info 1221 drivers/infiniband/core/mad.c mad_send_wr->mad_list.mad_queue = &qp_info->send_queue; qp_info 1248 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->send_queue.lock, flags); qp_info 1249 drivers/infiniband/core/mad.c if (qp_info->send_queue.count < qp_info->send_queue.max_active) { qp_info 1250 drivers/infiniband/core/mad.c trace_ib_mad_ib_send_mad(mad_send_wr, qp_info); qp_info 1253 drivers/infiniband/core/mad.c list = &qp_info->send_queue.list; qp_info 1256 drivers/infiniband/core/mad.c list = &qp_info->overflow_list; qp_info 1260 drivers/infiniband/core/mad.c qp_info->send_queue.count++; qp_info 1263 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->send_queue.lock, flags); qp_info 1522 drivers/infiniband/core/mad.c port_priv = agent_priv->qp_info->port_priv; qp_info 1586 drivers/infiniband/core/mad.c port_priv = agent_priv->qp_info->port_priv; qp_info 1689 drivers/infiniband/core/mad.c port_priv = agent_priv->qp_info->port_priv; qp_info 1854 drivers/infiniband/core/mad.c const struct ib_mad_qp_info *qp_info, qp_info 1858 drivers/infiniband/core/mad.c u32 qp_num = qp_info->qp->qp_num; qp_info 2096 drivers/infiniband/core/mad.c const struct ib_mad_qp_info *qp_info, qp_info 2137 drivers/infiniband/core/mad.c qp_info->qp->qp_num, qp_info 2183 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info, qp_info 2227 drivers/infiniband/core/mad.c qp_info->qp->qp_num, qp_info 2239 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info, qp_info 2250 drivers/infiniband/core/mad.c return handle_opa_smi(port_priv, qp_info, wc, port_num, recv, qp_info 2253 drivers/infiniband/core/mad.c return handle_ib_smi(port_priv, qp_info, wc, port_num, recv, response); qp_info 2261 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info; qp_info 2282 drivers/infiniband/core/mad.c qp_info = mad_list->mad_queue->qp_info; qp_info 2285 drivers/infiniband/core/mad.c opa = rdma_cap_opa_mad(qp_info->port_priv->device, qp_info 2286 drivers/infiniband/core/mad.c qp_info->port_priv->port_num); qp_info 2311 drivers/infiniband/core/mad.c if (atomic_read(&qp_info->snoop_count)) qp_info 2312 drivers/infiniband/core/mad.c snoop_recv(qp_info, &recv->header.recv_wc, IB_MAD_SNOOP_RECVS); qp_info 2315 drivers/infiniband/core/mad.c if (!validate_mad((const struct ib_mad_hdr *)recv->mad, qp_info, opa)) qp_info 2318 drivers/infiniband/core/mad.c trace_ib_mad_recv_done_handler(qp_info, wc, qp_info 2333 drivers/infiniband/core/mad.c if (handle_smi(port_priv, qp_info, wc, port_num, recv, qp_info 2358 drivers/infiniband/core/mad.c qp_info->qp->qp_num, qp_info 2378 drivers/infiniband/core/mad.c qp_info->qp->qp_num, mad_size, opa); qp_info 2384 drivers/infiniband/core/mad.c ib_mad_post_receive_mads(qp_info, response); qp_info 2387 drivers/infiniband/core/mad.c ib_mad_post_receive_mads(qp_info, recv); qp_info 2408 drivers/infiniband/core/mad.c mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq, qp_info 2443 drivers/infiniband/core/mad.c mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq, qp_info 2513 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info; qp_info 2530 drivers/infiniband/core/mad.c qp_info = send_queue->qp_info; qp_info 2548 drivers/infiniband/core/mad.c mad_list = container_of(qp_info->overflow_list.next, qp_info 2560 drivers/infiniband/core/mad.c if (atomic_read(&qp_info->snoop_count)) qp_info 2561 drivers/infiniband/core/mad.c snoop_send(qp_info, &mad_send_wr->send_buf, &mad_send_wc, qp_info 2566 drivers/infiniband/core/mad.c trace_ib_mad_send_done_resend(queued_send_wr, qp_info); qp_info 2567 drivers/infiniband/core/mad.c ret = ib_post_send(qp_info->qp, &queued_send_wr->send_wr.wr, qp_info 2579 drivers/infiniband/core/mad.c static void mark_sends_for_retry(struct ib_mad_qp_info *qp_info) qp_info 2585 drivers/infiniband/core/mad.c spin_lock_irqsave(&qp_info->send_queue.lock, flags); qp_info 2586 drivers/infiniband/core/mad.c list_for_each_entry(mad_list, &qp_info->send_queue.list, list) { qp_info 2592 drivers/infiniband/core/mad.c spin_unlock_irqrestore(&qp_info->send_queue.lock, flags); qp_info 2600 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info = mad_list->mad_queue->qp_info; qp_info 2614 drivers/infiniband/core/mad.c trace_ib_mad_error_handler(mad_send_wr, qp_info); qp_info 2615 drivers/infiniband/core/mad.c ret = ib_post_send(qp_info->qp, &mad_send_wr->send_wr.wr, qp_info 2628 drivers/infiniband/core/mad.c ret = ib_modify_qp(qp_info->qp, attr, qp_info 2636 drivers/infiniband/core/mad.c mark_sends_for_retry(qp_info); qp_info 2756 drivers/infiniband/core/mad.c opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device, qp_info 2757 drivers/infiniband/core/mad.c mad_agent_priv->qp_info->port_priv->port_num); qp_info 2804 drivers/infiniband/core/mad.c if (atomic_read(&recv_mad_agent->qp_info->snoop_count)) qp_info 2805 drivers/infiniband/core/mad.c snoop_recv(recv_mad_agent->qp_info, qp_info 2822 drivers/infiniband/core/mad.c if (atomic_read(&mad_agent_priv->qp_info->snoop_count)) qp_info 2823 drivers/infiniband/core/mad.c snoop_send(mad_agent_priv->qp_info, qp_info 2895 drivers/infiniband/core/mad.c queue_delayed_work(mad_agent_priv->qp_info-> qp_info 2925 drivers/infiniband/core/mad.c static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info, qp_info 2933 drivers/infiniband/core/mad.c struct ib_mad_queue *recv_queue = &qp_info->recv_queue; qp_info 2936 drivers/infiniband/core/mad.c sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey; qp_info 2949 drivers/infiniband/core/mad.c mad_priv = alloc_mad_private(port_mad_size(qp_info->port_priv), qp_info 2957 drivers/infiniband/core/mad.c sg_list.addr = ib_dma_map_single(qp_info->port_priv->device, qp_info 2961 drivers/infiniband/core/mad.c if (unlikely(ib_dma_mapping_error(qp_info->port_priv->device, qp_info 2976 drivers/infiniband/core/mad.c ret = ib_post_recv(qp_info->qp, &recv_wr, NULL); qp_info 2982 drivers/infiniband/core/mad.c ib_dma_unmap_single(qp_info->port_priv->device, qp_info 2987 drivers/infiniband/core/mad.c dev_err(&qp_info->port_priv->device->dev, qp_info 2999 drivers/infiniband/core/mad.c static void cleanup_recv_queue(struct ib_mad_qp_info *qp_info) qp_info 3005 drivers/infiniband/core/mad.c if (!qp_info->qp) qp_info 3008 drivers/infiniband/core/mad.c while (!list_empty(&qp_info->recv_queue.list)) { qp_info 3010 drivers/infiniband/core/mad.c mad_list = list_entry(qp_info->recv_queue.list.next, qp_info 3021 drivers/infiniband/core/mad.c ib_dma_unmap_single(qp_info->port_priv->device, qp_info 3028 drivers/infiniband/core/mad.c qp_info->recv_queue.count = 0; qp_info 3051 drivers/infiniband/core/mad.c qp = port_priv->qp_info[i].qp; qp_info 3100 drivers/infiniband/core/mad.c if (!port_priv->qp_info[i].qp) qp_info 3103 drivers/infiniband/core/mad.c ret = ib_mad_post_receive_mads(&port_priv->qp_info[i], NULL); qp_info 3117 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info = qp_context; qp_info 3120 drivers/infiniband/core/mad.c dev_err(&qp_info->port_priv->device->dev, qp_info 3122 drivers/infiniband/core/mad.c event->event, qp_info->qp->qp_num); qp_info 3125 drivers/infiniband/core/mad.c static void init_mad_queue(struct ib_mad_qp_info *qp_info, qp_info 3128 drivers/infiniband/core/mad.c mad_queue->qp_info = qp_info; qp_info 3135 drivers/infiniband/core/mad.c struct ib_mad_qp_info *qp_info) qp_info 3137 drivers/infiniband/core/mad.c qp_info->port_priv = port_priv; qp_info 3138 drivers/infiniband/core/mad.c init_mad_queue(qp_info, &qp_info->send_queue); qp_info 3139 drivers/infiniband/core/mad.c init_mad_queue(qp_info, &qp_info->recv_queue); qp_info 3140 drivers/infiniband/core/mad.c INIT_LIST_HEAD(&qp_info->overflow_list); qp_info 3141 drivers/infiniband/core/mad.c spin_lock_init(&qp_info->snoop_lock); qp_info 3142 drivers/infiniband/core/mad.c qp_info->snoop_table = NULL; qp_info 3143 drivers/infiniband/core/mad.c qp_info->snoop_table_size = 0; qp_info 3144 drivers/infiniband/core/mad.c atomic_set(&qp_info->snoop_count, 0); qp_info 3147 drivers/infiniband/core/mad.c static int create_mad_qp(struct ib_mad_qp_info *qp_info, qp_info 3154 drivers/infiniband/core/mad.c qp_init_attr.send_cq = qp_info->port_priv->cq; qp_info 3155 drivers/infiniband/core/mad.c qp_init_attr.recv_cq = qp_info->port_priv->cq; qp_info 3162 drivers/infiniband/core/mad.c qp_init_attr.port_num = qp_info->port_priv->port_num; qp_info 3163 drivers/infiniband/core/mad.c qp_init_attr.qp_context = qp_info; qp_info 3165 drivers/infiniband/core/mad.c qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr); qp_info 3166 drivers/infiniband/core/mad.c if (IS_ERR(qp_info->qp)) { qp_info 3167 drivers/infiniband/core/mad.c dev_err(&qp_info->port_priv->device->dev, qp_info 3170 drivers/infiniband/core/mad.c ret = PTR_ERR(qp_info->qp); qp_info 3174 drivers/infiniband/core/mad.c qp_info->send_queue.max_active = mad_sendq_size; qp_info 3175 drivers/infiniband/core/mad.c qp_info->recv_queue.max_active = mad_recvq_size; qp_info 3182 drivers/infiniband/core/mad.c static void destroy_mad_qp(struct ib_mad_qp_info *qp_info) qp_info 3184 drivers/infiniband/core/mad.c if (!qp_info->qp) qp_info 3187 drivers/infiniband/core/mad.c ib_destroy_qp(qp_info->qp); qp_info 3188 drivers/infiniband/core/mad.c kfree(qp_info->snoop_table); qp_info 3219 drivers/infiniband/core/mad.c init_mad_qp(port_priv, &port_priv->qp_info[0]); qp_info 3220 drivers/infiniband/core/mad.c init_mad_qp(port_priv, &port_priv->qp_info[1]); qp_info 3243 drivers/infiniband/core/mad.c ret = create_mad_qp(&port_priv->qp_info[0], IB_QPT_SMI); qp_info 3247 drivers/infiniband/core/mad.c ret = create_mad_qp(&port_priv->qp_info[1], IB_QPT_GSI); qp_info 3277 drivers/infiniband/core/mad.c destroy_mad_qp(&port_priv->qp_info[1]); qp_info 3279 drivers/infiniband/core/mad.c destroy_mad_qp(&port_priv->qp_info[0]); qp_info 3282 drivers/infiniband/core/mad.c cleanup_recv_queue(&port_priv->qp_info[1]); qp_info 3283 drivers/infiniband/core/mad.c cleanup_recv_queue(&port_priv->qp_info[0]); qp_info 3313 drivers/infiniband/core/mad.c destroy_mad_qp(&port_priv->qp_info[1]); qp_info 3314 drivers/infiniband/core/mad.c destroy_mad_qp(&port_priv->qp_info[0]); qp_info 3317 drivers/infiniband/core/mad.c cleanup_recv_queue(&port_priv->qp_info[1]); qp_info 3318 drivers/infiniband/core/mad.c cleanup_recv_queue(&port_priv->qp_info[0]); qp_info 94 drivers/infiniband/core/mad_priv.h struct ib_mad_qp_info *qp_info; qp_info 115 drivers/infiniband/core/mad_priv.h struct ib_mad_qp_info *qp_info; qp_info 184 drivers/infiniband/core/mad_priv.h struct ib_mad_qp_info *qp_info; qp_info 209 drivers/infiniband/core/mad_priv.h struct ib_mad_qp_info qp_info[IB_MAD_QPS_CORE]; qp_info 106 drivers/infiniband/core/mad_rmpp.c flush_workqueue(agent->qp_info->port_priv->wq); qp_info 405 drivers/infiniband/core/mad_rmpp.c return max(agent->qp_info->recv_queue.max_active >> 3, 1); qp_info 440 drivers/infiniband/core/mad_rmpp.c bool opa = rdma_cap_opa_mad(rmpp_recv->agent->qp_info->port_priv->device, qp_info 441 drivers/infiniband/core/mad_rmpp.c rmpp_recv->agent->qp_info->port_priv->port_num); qp_info 472 drivers/infiniband/core/mad_rmpp.c queue_delayed_work(rmpp_recv->agent->qp_info->port_priv->wq, qp_info 565 drivers/infiniband/core/mad_rmpp.c queue_delayed_work(agent->qp_info->port_priv->wq, qp_info 1225 drivers/infiniband/hw/i40iw/i40iw_utils.c struct i40iw_create_qp_info *qp_info; qp_info 1233 drivers/infiniband/hw/i40iw/i40iw_utils.c qp_info = &cqp_request->info.in.u.qp_create.info; qp_info 1235 drivers/infiniband/hw/i40iw/i40iw_utils.c memset(qp_info, 0, sizeof(*qp_info)); qp_info 1237 drivers/infiniband/hw/i40iw/i40iw_utils.c qp_info->cq_num_valid = true; qp_info 1238 drivers/infiniband/hw/i40iw/i40iw_utils.c qp_info->next_iwarp_state = I40IW_QP_STATE_RTS; qp_info 538 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_create_qp_info *qp_info; qp_info 702 drivers/infiniband/hw/i40iw/i40iw_verbs.c qp_info = &cqp_request->info.in.u.qp_create.info; qp_info 704 drivers/infiniband/hw/i40iw/i40iw_verbs.c memset(qp_info, 0, sizeof(*qp_info)); qp_info 706 drivers/infiniband/hw/i40iw/i40iw_verbs.c qp_info->cq_num_valid = true; qp_info 707 drivers/infiniband/hw/i40iw/i40iw_verbs.c qp_info->next_iwarp_state = I40IW_QP_STATE_IDLE; qp_info 52 drivers/misc/mic/scif/scif_api.c ep->qp_info.qp = kzalloc(sizeof(*ep->qp_info.qp), GFP_KERNEL); qp_info 53 drivers/misc/mic/scif/scif_api.c if (!ep->qp_info.qp) qp_info 71 drivers/misc/mic/scif/scif_api.c kfree(ep->qp_info.qp); qp_info 425 drivers/misc/mic/scif/scif_api.c ep->qp_info.qp = NULL; qp_info 476 drivers/misc/mic/scif/scif_api.c err = scif_setup_qp_connect(ep->qp_info.qp, &ep->qp_info.qp_offset, qp_info 481 drivers/misc/mic/scif/scif_api.c __func__, err, ep->qp_info.qp_offset); qp_info 496 drivers/misc/mic/scif/scif_api.c msg.payload[1] = ep->qp_info.qp_offset; qp_info 519 drivers/misc/mic/scif/scif_api.c ep->qp_info.qp, qp_info 520 drivers/misc/mic/scif/scif_api.c ep->qp_info.gnt_pld); qp_info 684 drivers/misc/mic/scif/scif_api.c ep->qp_info.qp->magic = SCIFEP_MAGIC; qp_info 818 drivers/misc/mic/scif/scif_api.c cep->qp_info.qp = kzalloc(sizeof(*cep->qp_info.qp), GFP_KERNEL); qp_info 819 drivers/misc/mic/scif/scif_api.c if (!cep->qp_info.qp) { qp_info 828 drivers/misc/mic/scif/scif_api.c cep->qp_info.qp->magic = SCIFEP_MAGIC; qp_info 834 drivers/misc/mic/scif/scif_api.c err = scif_setup_qp_accept(cep->qp_info.qp, &cep->qp_info.qp_offset, qp_info 840 drivers/misc/mic/scif/scif_api.c lep, cep, err, cep->qp_info.qp_offset); qp_info 856 drivers/misc/mic/scif/scif_api.c msg.payload[1] = cep->qp_info.qp_offset; qp_info 942 drivers/misc/mic/scif/scif_api.c struct scif_qp *qp = ep->qp_info.qp; qp_info 1005 drivers/misc/mic/scif/scif_api.c struct scif_qp *qp = ep->qp_info.qp; qp_info 1345 drivers/misc/mic/scif/scif_api.c if (scif_rb_count(&ep->qp_info.qp->inbound_q, 1)) qp_info 1348 drivers/misc/mic/scif/scif_api.c if (scif_rb_space(&ep->qp_info.qp->outbound_q)) qp_info 14 drivers/misc/mic/scif/scif_epd.c struct scif_qp *qp = ep->qp_info.qp; qp_info 41 drivers/misc/mic/scif/scif_epd.c struct scif_qp *qp = ep->qp_info.qp; qp_info 173 drivers/misc/mic/scif/scif_epd.c ep->qp_info.gnt_pld = msg->payload[1]; qp_info 102 drivers/misc/mic/scif/scif_epd.h struct scif_endpt_qp_info qp_info; qp_info 760 drivers/scsi/bnx2i/bnx2i.h struct qp_info qp; qp_info 1957 drivers/scsi/bnx2i/bnx2i_hwi.c struct qp_info *qp; qp_info 19 include/trace/events/ib_mad.h struct ib_mad_qp_info *qp_info, qp_info 25 include/trace/events/ib_mad.h struct ib_mad_qp_info *qp_info), qp_info 26 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info), qp_info 58 include/trace/events/ib_mad.h __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num; qp_info 84 include/trace/events/ib_mad.h create_mad_addr_info(wr, qp_info, __entry); qp_info 110 include/trace/events/ib_mad.h struct ib_mad_qp_info *qp_info), qp_info 111 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); qp_info 114 include/trace/events/ib_mad.h struct ib_mad_qp_info *qp_info), qp_info 115 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); qp_info 118 include/trace/events/ib_mad.h struct ib_mad_qp_info *qp_info), qp_info 119 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); qp_info 147 include/trace/events/ib_mad.h __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num; qp_info 185 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_qp_info *qp_info, struct ib_wc *wc, qp_info 187 include/trace/events/ib_mad.h TP_ARGS(qp_info, wc, mad_hdr), qp_info 211 include/trace/events/ib_mad.h __entry->dev_index = qp_info->port_priv->device->index; qp_info 212 include/trace/events/ib_mad.h __entry->port_num = qp_info->port_priv->port_num; qp_info 213 include/trace/events/ib_mad.h __entry->qp_num = qp_info->qp->qp_num; qp_info 227 include/trace/events/ib_mad.h ib_query_pkey(qp_info->port_priv->device, qp_info 228 include/trace/events/ib_mad.h qp_info->port_priv->port_num,