qp_info            59 drivers/infiniband/core/mad.c 			  struct ib_mad_qp_info *qp_info,
qp_info            63 drivers/infiniband/core/mad.c 	struct ib_device *dev = qp_info->port_priv->device;
qp_info            64 drivers/infiniband/core/mad.c 	u8 pnum = qp_info->port_priv->port_num;
qp_info           103 drivers/infiniband/core/mad.c static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info,
qp_info           365 drivers/infiniband/core/mad.c 	if (!port_priv->qp_info[qpn].qp) {
qp_info           388 drivers/infiniband/core/mad.c 	mad_agent_priv->qp_info = &port_priv->qp_info[qpn];
qp_info           395 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.qp = port_priv->qp_info[qpn].qp;
qp_info           502 drivers/infiniband/core/mad.c static int register_snoop_agent(struct ib_mad_qp_info *qp_info,
qp_info           509 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           511 drivers/infiniband/core/mad.c 	for (i = 0; i < qp_info->snoop_table_size; i++)
qp_info           512 drivers/infiniband/core/mad.c 		if (!qp_info->snoop_table[i])
qp_info           515 drivers/infiniband/core/mad.c 	if (i == qp_info->snoop_table_size) {
qp_info           517 drivers/infiniband/core/mad.c 		new_snoop_table = krealloc(qp_info->snoop_table,
qp_info           519 drivers/infiniband/core/mad.c 					   (qp_info->snoop_table_size + 1),
qp_info           526 drivers/infiniband/core/mad.c 		qp_info->snoop_table = new_snoop_table;
qp_info           527 drivers/infiniband/core/mad.c 		qp_info->snoop_table_size++;
qp_info           529 drivers/infiniband/core/mad.c 	qp_info->snoop_table[i] = mad_snoop_priv;
qp_info           530 drivers/infiniband/core/mad.c 	atomic_inc(&qp_info->snoop_count);
qp_info           532 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           574 drivers/infiniband/core/mad.c 	mad_snoop_priv->qp_info = &port_priv->qp_info[qpn];
qp_info           579 drivers/infiniband/core/mad.c 	mad_snoop_priv->agent.qp = port_priv->qp_info[qpn].qp;
qp_info           591 drivers/infiniband/core/mad.c 						&port_priv->qp_info[qpn],
qp_info           633 drivers/infiniband/core/mad.c 	port_priv = mad_agent_priv->qp_info->port_priv;
qp_info           655 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info *qp_info;
qp_info           658 drivers/infiniband/core/mad.c 	qp_info = mad_snoop_priv->qp_info;
qp_info           659 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           660 drivers/infiniband/core/mad.c 	qp_info->snoop_table[mad_snoop_priv->snoop_index] = NULL;
qp_info           661 drivers/infiniband/core/mad.c 	atomic_dec(&qp_info->snoop_count);
qp_info           662 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           709 drivers/infiniband/core/mad.c static void snoop_send(struct ib_mad_qp_info *qp_info,
qp_info           718 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           719 drivers/infiniband/core/mad.c 	for (i = 0; i < qp_info->snoop_table_size; i++) {
qp_info           720 drivers/infiniband/core/mad.c 		mad_snoop_priv = qp_info->snoop_table[i];
qp_info           726 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           730 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           732 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           735 drivers/infiniband/core/mad.c static void snoop_recv(struct ib_mad_qp_info *qp_info,
qp_info           743 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           744 drivers/infiniband/core/mad.c 	for (i = 0; i < qp_info->snoop_table_size; i++) {
qp_info           745 drivers/infiniband/core/mad.c 		mad_snoop_priv = qp_info->snoop_table[i];
qp_info           751 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           755 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&qp_info->snoop_lock, flags);
qp_info           757 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->snoop_lock, flags);
qp_info           823 drivers/infiniband/core/mad.c 	size_t mad_size = port_mad_size(mad_agent_priv->qp_info->port_priv);
qp_info           826 drivers/infiniband/core/mad.c 	bool opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
qp_info           827 drivers/infiniband/core/mad.c 				    mad_agent_priv->qp_info->port_priv->port_num);
qp_info           975 drivers/infiniband/core/mad.c 	queue_work(mad_agent_priv->qp_info->port_priv->wq,
qp_info          1212 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info *qp_info;
qp_info          1220 drivers/infiniband/core/mad.c 	qp_info = mad_send_wr->mad_agent_priv->qp_info;
qp_info          1221 drivers/infiniband/core/mad.c 	mad_send_wr->mad_list.mad_queue = &qp_info->send_queue;
qp_info          1248 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->send_queue.lock, flags);
qp_info          1249 drivers/infiniband/core/mad.c 	if (qp_info->send_queue.count < qp_info->send_queue.max_active) {
qp_info          1250 drivers/infiniband/core/mad.c 		trace_ib_mad_ib_send_mad(mad_send_wr, qp_info);
qp_info          1253 drivers/infiniband/core/mad.c 		list = &qp_info->send_queue.list;
qp_info          1256 drivers/infiniband/core/mad.c 		list = &qp_info->overflow_list;
qp_info          1260 drivers/infiniband/core/mad.c 		qp_info->send_queue.count++;
qp_info          1263 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
qp_info          1522 drivers/infiniband/core/mad.c 	port_priv = agent_priv->qp_info->port_priv;
qp_info          1586 drivers/infiniband/core/mad.c 	port_priv = agent_priv->qp_info->port_priv;
qp_info          1689 drivers/infiniband/core/mad.c 	port_priv = agent_priv->qp_info->port_priv;
qp_info          1854 drivers/infiniband/core/mad.c 			const struct ib_mad_qp_info *qp_info,
qp_info          1858 drivers/infiniband/core/mad.c 	u32 qp_num = qp_info->qp->qp_num;
qp_info          2096 drivers/infiniband/core/mad.c 				     const struct ib_mad_qp_info *qp_info,
qp_info          2137 drivers/infiniband/core/mad.c 				    qp_info->qp->qp_num,
qp_info          2183 drivers/infiniband/core/mad.c 	       struct ib_mad_qp_info *qp_info,
qp_info          2227 drivers/infiniband/core/mad.c 				    qp_info->qp->qp_num,
qp_info          2239 drivers/infiniband/core/mad.c 	   struct ib_mad_qp_info *qp_info,
qp_info          2250 drivers/infiniband/core/mad.c 		return handle_opa_smi(port_priv, qp_info, wc, port_num, recv,
qp_info          2253 drivers/infiniband/core/mad.c 	return handle_ib_smi(port_priv, qp_info, wc, port_num, recv, response);
qp_info          2261 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info *qp_info;
qp_info          2282 drivers/infiniband/core/mad.c 	qp_info = mad_list->mad_queue->qp_info;
qp_info          2285 drivers/infiniband/core/mad.c 	opa = rdma_cap_opa_mad(qp_info->port_priv->device,
qp_info          2286 drivers/infiniband/core/mad.c 			       qp_info->port_priv->port_num);
qp_info          2311 drivers/infiniband/core/mad.c 	if (atomic_read(&qp_info->snoop_count))
qp_info          2312 drivers/infiniband/core/mad.c 		snoop_recv(qp_info, &recv->header.recv_wc, IB_MAD_SNOOP_RECVS);
qp_info          2315 drivers/infiniband/core/mad.c 	if (!validate_mad((const struct ib_mad_hdr *)recv->mad, qp_info, opa))
qp_info          2318 drivers/infiniband/core/mad.c 	trace_ib_mad_recv_done_handler(qp_info, wc,
qp_info          2333 drivers/infiniband/core/mad.c 		if (handle_smi(port_priv, qp_info, wc, port_num, recv,
qp_info          2358 drivers/infiniband/core/mad.c 						    qp_info->qp->qp_num,
qp_info          2378 drivers/infiniband/core/mad.c 				    qp_info->qp->qp_num, mad_size, opa);
qp_info          2384 drivers/infiniband/core/mad.c 		ib_mad_post_receive_mads(qp_info, response);
qp_info          2387 drivers/infiniband/core/mad.c 		ib_mad_post_receive_mads(qp_info, recv);
qp_info          2408 drivers/infiniband/core/mad.c 			mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
qp_info          2443 drivers/infiniband/core/mad.c 		mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
qp_info          2513 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info		*qp_info;
qp_info          2530 drivers/infiniband/core/mad.c 	qp_info = send_queue->qp_info;
qp_info          2548 drivers/infiniband/core/mad.c 		mad_list = container_of(qp_info->overflow_list.next,
qp_info          2560 drivers/infiniband/core/mad.c 	if (atomic_read(&qp_info->snoop_count))
qp_info          2561 drivers/infiniband/core/mad.c 		snoop_send(qp_info, &mad_send_wr->send_buf, &mad_send_wc,
qp_info          2566 drivers/infiniband/core/mad.c 		trace_ib_mad_send_done_resend(queued_send_wr, qp_info);
qp_info          2567 drivers/infiniband/core/mad.c 		ret = ib_post_send(qp_info->qp, &queued_send_wr->send_wr.wr,
qp_info          2579 drivers/infiniband/core/mad.c static void mark_sends_for_retry(struct ib_mad_qp_info *qp_info)
qp_info          2585 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&qp_info->send_queue.lock, flags);
qp_info          2586 drivers/infiniband/core/mad.c 	list_for_each_entry(mad_list, &qp_info->send_queue.list, list) {
qp_info          2592 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
qp_info          2600 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info *qp_info = mad_list->mad_queue->qp_info;
qp_info          2614 drivers/infiniband/core/mad.c 			trace_ib_mad_error_handler(mad_send_wr, qp_info);
qp_info          2615 drivers/infiniband/core/mad.c 			ret = ib_post_send(qp_info->qp, &mad_send_wr->send_wr.wr,
qp_info          2628 drivers/infiniband/core/mad.c 			ret = ib_modify_qp(qp_info->qp, attr,
qp_info          2636 drivers/infiniband/core/mad.c 				mark_sends_for_retry(qp_info);
qp_info          2756 drivers/infiniband/core/mad.c 	opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
qp_info          2757 drivers/infiniband/core/mad.c 			       mad_agent_priv->qp_info->port_priv->port_num);
qp_info          2804 drivers/infiniband/core/mad.c 			if (atomic_read(&recv_mad_agent->qp_info->snoop_count))
qp_info          2805 drivers/infiniband/core/mad.c 				snoop_recv(recv_mad_agent->qp_info,
qp_info          2822 drivers/infiniband/core/mad.c 		if (atomic_read(&mad_agent_priv->qp_info->snoop_count))
qp_info          2823 drivers/infiniband/core/mad.c 			snoop_send(mad_agent_priv->qp_info,
qp_info          2895 drivers/infiniband/core/mad.c 			queue_delayed_work(mad_agent_priv->qp_info->
qp_info          2925 drivers/infiniband/core/mad.c static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info,
qp_info          2933 drivers/infiniband/core/mad.c 	struct ib_mad_queue *recv_queue = &qp_info->recv_queue;
qp_info          2936 drivers/infiniband/core/mad.c 	sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey;
qp_info          2949 drivers/infiniband/core/mad.c 			mad_priv = alloc_mad_private(port_mad_size(qp_info->port_priv),
qp_info          2957 drivers/infiniband/core/mad.c 		sg_list.addr = ib_dma_map_single(qp_info->port_priv->device,
qp_info          2961 drivers/infiniband/core/mad.c 		if (unlikely(ib_dma_mapping_error(qp_info->port_priv->device,
qp_info          2976 drivers/infiniband/core/mad.c 		ret = ib_post_recv(qp_info->qp, &recv_wr, NULL);
qp_info          2982 drivers/infiniband/core/mad.c 			ib_dma_unmap_single(qp_info->port_priv->device,
qp_info          2987 drivers/infiniband/core/mad.c 			dev_err(&qp_info->port_priv->device->dev,
qp_info          2999 drivers/infiniband/core/mad.c static void cleanup_recv_queue(struct ib_mad_qp_info *qp_info)
qp_info          3005 drivers/infiniband/core/mad.c 	if (!qp_info->qp)
qp_info          3008 drivers/infiniband/core/mad.c 	while (!list_empty(&qp_info->recv_queue.list)) {
qp_info          3010 drivers/infiniband/core/mad.c 		mad_list = list_entry(qp_info->recv_queue.list.next,
qp_info          3021 drivers/infiniband/core/mad.c 		ib_dma_unmap_single(qp_info->port_priv->device,
qp_info          3028 drivers/infiniband/core/mad.c 	qp_info->recv_queue.count = 0;
qp_info          3051 drivers/infiniband/core/mad.c 		qp = port_priv->qp_info[i].qp;
qp_info          3100 drivers/infiniband/core/mad.c 		if (!port_priv->qp_info[i].qp)
qp_info          3103 drivers/infiniband/core/mad.c 		ret = ib_mad_post_receive_mads(&port_priv->qp_info[i], NULL);
qp_info          3117 drivers/infiniband/core/mad.c 	struct ib_mad_qp_info	*qp_info = qp_context;
qp_info          3120 drivers/infiniband/core/mad.c 	dev_err(&qp_info->port_priv->device->dev,
qp_info          3122 drivers/infiniband/core/mad.c 		event->event, qp_info->qp->qp_num);
qp_info          3125 drivers/infiniband/core/mad.c static void init_mad_queue(struct ib_mad_qp_info *qp_info,
qp_info          3128 drivers/infiniband/core/mad.c 	mad_queue->qp_info = qp_info;
qp_info          3135 drivers/infiniband/core/mad.c 			struct ib_mad_qp_info *qp_info)
qp_info          3137 drivers/infiniband/core/mad.c 	qp_info->port_priv = port_priv;
qp_info          3138 drivers/infiniband/core/mad.c 	init_mad_queue(qp_info, &qp_info->send_queue);
qp_info          3139 drivers/infiniband/core/mad.c 	init_mad_queue(qp_info, &qp_info->recv_queue);
qp_info          3140 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&qp_info->overflow_list);
qp_info          3141 drivers/infiniband/core/mad.c 	spin_lock_init(&qp_info->snoop_lock);
qp_info          3142 drivers/infiniband/core/mad.c 	qp_info->snoop_table = NULL;
qp_info          3143 drivers/infiniband/core/mad.c 	qp_info->snoop_table_size = 0;
qp_info          3144 drivers/infiniband/core/mad.c 	atomic_set(&qp_info->snoop_count, 0);
qp_info          3147 drivers/infiniband/core/mad.c static int create_mad_qp(struct ib_mad_qp_info *qp_info,
qp_info          3154 drivers/infiniband/core/mad.c 	qp_init_attr.send_cq = qp_info->port_priv->cq;
qp_info          3155 drivers/infiniband/core/mad.c 	qp_init_attr.recv_cq = qp_info->port_priv->cq;
qp_info          3162 drivers/infiniband/core/mad.c 	qp_init_attr.port_num = qp_info->port_priv->port_num;
qp_info          3163 drivers/infiniband/core/mad.c 	qp_init_attr.qp_context = qp_info;
qp_info          3165 drivers/infiniband/core/mad.c 	qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr);
qp_info          3166 drivers/infiniband/core/mad.c 	if (IS_ERR(qp_info->qp)) {
qp_info          3167 drivers/infiniband/core/mad.c 		dev_err(&qp_info->port_priv->device->dev,
qp_info          3170 drivers/infiniband/core/mad.c 		ret = PTR_ERR(qp_info->qp);
qp_info          3174 drivers/infiniband/core/mad.c 	qp_info->send_queue.max_active = mad_sendq_size;
qp_info          3175 drivers/infiniband/core/mad.c 	qp_info->recv_queue.max_active = mad_recvq_size;
qp_info          3182 drivers/infiniband/core/mad.c static void destroy_mad_qp(struct ib_mad_qp_info *qp_info)
qp_info          3184 drivers/infiniband/core/mad.c 	if (!qp_info->qp)
qp_info          3187 drivers/infiniband/core/mad.c 	ib_destroy_qp(qp_info->qp);
qp_info          3188 drivers/infiniband/core/mad.c 	kfree(qp_info->snoop_table);
qp_info          3219 drivers/infiniband/core/mad.c 	init_mad_qp(port_priv, &port_priv->qp_info[0]);
qp_info          3220 drivers/infiniband/core/mad.c 	init_mad_qp(port_priv, &port_priv->qp_info[1]);
qp_info          3243 drivers/infiniband/core/mad.c 		ret = create_mad_qp(&port_priv->qp_info[0], IB_QPT_SMI);
qp_info          3247 drivers/infiniband/core/mad.c 	ret = create_mad_qp(&port_priv->qp_info[1], IB_QPT_GSI);
qp_info          3277 drivers/infiniband/core/mad.c 	destroy_mad_qp(&port_priv->qp_info[1]);
qp_info          3279 drivers/infiniband/core/mad.c 	destroy_mad_qp(&port_priv->qp_info[0]);
qp_info          3282 drivers/infiniband/core/mad.c 	cleanup_recv_queue(&port_priv->qp_info[1]);
qp_info          3283 drivers/infiniband/core/mad.c 	cleanup_recv_queue(&port_priv->qp_info[0]);
qp_info          3313 drivers/infiniband/core/mad.c 	destroy_mad_qp(&port_priv->qp_info[1]);
qp_info          3314 drivers/infiniband/core/mad.c 	destroy_mad_qp(&port_priv->qp_info[0]);
qp_info          3317 drivers/infiniband/core/mad.c 	cleanup_recv_queue(&port_priv->qp_info[1]);
qp_info          3318 drivers/infiniband/core/mad.c 	cleanup_recv_queue(&port_priv->qp_info[0]);
qp_info            94 drivers/infiniband/core/mad_priv.h 	struct ib_mad_qp_info *qp_info;
qp_info           115 drivers/infiniband/core/mad_priv.h 	struct ib_mad_qp_info *qp_info;
qp_info           184 drivers/infiniband/core/mad_priv.h 	struct ib_mad_qp_info *qp_info;
qp_info           209 drivers/infiniband/core/mad_priv.h 	struct ib_mad_qp_info qp_info[IB_MAD_QPS_CORE];
qp_info           106 drivers/infiniband/core/mad_rmpp.c 	flush_workqueue(agent->qp_info->port_priv->wq);
qp_info           405 drivers/infiniband/core/mad_rmpp.c 	return max(agent->qp_info->recv_queue.max_active >> 3, 1);
qp_info           440 drivers/infiniband/core/mad_rmpp.c 	bool opa = rdma_cap_opa_mad(rmpp_recv->agent->qp_info->port_priv->device,
qp_info           441 drivers/infiniband/core/mad_rmpp.c 				    rmpp_recv->agent->qp_info->port_priv->port_num);
qp_info           472 drivers/infiniband/core/mad_rmpp.c 	queue_delayed_work(rmpp_recv->agent->qp_info->port_priv->wq,
qp_info           565 drivers/infiniband/core/mad_rmpp.c 		queue_delayed_work(agent->qp_info->port_priv->wq,
qp_info          1225 drivers/infiniband/hw/i40iw/i40iw_utils.c 	struct i40iw_create_qp_info *qp_info;
qp_info          1233 drivers/infiniband/hw/i40iw/i40iw_utils.c 	qp_info = &cqp_request->info.in.u.qp_create.info;
qp_info          1235 drivers/infiniband/hw/i40iw/i40iw_utils.c 	memset(qp_info, 0, sizeof(*qp_info));
qp_info          1237 drivers/infiniband/hw/i40iw/i40iw_utils.c 	qp_info->cq_num_valid = true;
qp_info          1238 drivers/infiniband/hw/i40iw/i40iw_utils.c 	qp_info->next_iwarp_state = I40IW_QP_STATE_RTS;
qp_info           538 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_create_qp_info *qp_info;
qp_info           702 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	qp_info = &cqp_request->info.in.u.qp_create.info;
qp_info           704 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	memset(qp_info, 0, sizeof(*qp_info));
qp_info           706 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	qp_info->cq_num_valid = true;
qp_info           707 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	qp_info->next_iwarp_state = I40IW_QP_STATE_IDLE;
qp_info            52 drivers/misc/mic/scif/scif_api.c 	ep->qp_info.qp = kzalloc(sizeof(*ep->qp_info.qp), GFP_KERNEL);
qp_info            53 drivers/misc/mic/scif/scif_api.c 	if (!ep->qp_info.qp)
qp_info            71 drivers/misc/mic/scif/scif_api.c 	kfree(ep->qp_info.qp);
qp_info           425 drivers/misc/mic/scif/scif_api.c 	ep->qp_info.qp = NULL;
qp_info           476 drivers/misc/mic/scif/scif_api.c 	err = scif_setup_qp_connect(ep->qp_info.qp, &ep->qp_info.qp_offset,
qp_info           481 drivers/misc/mic/scif/scif_api.c 			__func__, err, ep->qp_info.qp_offset);
qp_info           496 drivers/misc/mic/scif/scif_api.c 	msg.payload[1] = ep->qp_info.qp_offset;
qp_info           519 drivers/misc/mic/scif/scif_api.c 						     ep->qp_info.qp,
qp_info           520 drivers/misc/mic/scif/scif_api.c 						     ep->qp_info.gnt_pld);
qp_info           684 drivers/misc/mic/scif/scif_api.c 	ep->qp_info.qp->magic = SCIFEP_MAGIC;
qp_info           818 drivers/misc/mic/scif/scif_api.c 	cep->qp_info.qp = kzalloc(sizeof(*cep->qp_info.qp), GFP_KERNEL);
qp_info           819 drivers/misc/mic/scif/scif_api.c 	if (!cep->qp_info.qp) {
qp_info           828 drivers/misc/mic/scif/scif_api.c 	cep->qp_info.qp->magic = SCIFEP_MAGIC;
qp_info           834 drivers/misc/mic/scif/scif_api.c 	err = scif_setup_qp_accept(cep->qp_info.qp, &cep->qp_info.qp_offset,
qp_info           840 drivers/misc/mic/scif/scif_api.c 			lep, cep, err, cep->qp_info.qp_offset);
qp_info           856 drivers/misc/mic/scif/scif_api.c 	msg.payload[1] = cep->qp_info.qp_offset;
qp_info           942 drivers/misc/mic/scif/scif_api.c 	struct scif_qp *qp = ep->qp_info.qp;
qp_info          1005 drivers/misc/mic/scif/scif_api.c 	struct scif_qp *qp = ep->qp_info.qp;
qp_info          1345 drivers/misc/mic/scif/scif_api.c 			if (scif_rb_count(&ep->qp_info.qp->inbound_q, 1))
qp_info          1348 drivers/misc/mic/scif/scif_api.c 			if (scif_rb_space(&ep->qp_info.qp->outbound_q))
qp_info            14 drivers/misc/mic/scif/scif_epd.c 	struct scif_qp *qp = ep->qp_info.qp;
qp_info            41 drivers/misc/mic/scif/scif_epd.c 	struct scif_qp *qp = ep->qp_info.qp;
qp_info           173 drivers/misc/mic/scif/scif_epd.c 		ep->qp_info.gnt_pld = msg->payload[1];
qp_info           102 drivers/misc/mic/scif/scif_epd.h 	struct scif_endpt_qp_info qp_info;
qp_info           760 drivers/scsi/bnx2i/bnx2i.h 	struct qp_info qp;
qp_info          1957 drivers/scsi/bnx2i/bnx2i_hwi.c 	struct qp_info *qp;
qp_info            19 include/trace/events/ib_mad.h 			  struct ib_mad_qp_info *qp_info,
qp_info            25 include/trace/events/ib_mad.h 		 struct ib_mad_qp_info *qp_info),
qp_info            26 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info),
qp_info            58 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
qp_info            84 include/trace/events/ib_mad.h 		create_mad_addr_info(wr, qp_info, __entry);
qp_info           110 include/trace/events/ib_mad.h 		 struct ib_mad_qp_info *qp_info),
qp_info           111 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
qp_info           114 include/trace/events/ib_mad.h 		 struct ib_mad_qp_info *qp_info),
qp_info           115 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
qp_info           118 include/trace/events/ib_mad.h 		 struct ib_mad_qp_info *qp_info),
qp_info           119 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
qp_info           147 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
qp_info           185 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_qp_info *qp_info, struct ib_wc *wc,
qp_info           187 include/trace/events/ib_mad.h 	TP_ARGS(qp_info, wc, mad_hdr),
qp_info           211 include/trace/events/ib_mad.h 		__entry->dev_index = qp_info->port_priv->device->index;
qp_info           212 include/trace/events/ib_mad.h 		__entry->port_num = qp_info->port_priv->port_num;
qp_info           213 include/trace/events/ib_mad.h 		__entry->qp_num = qp_info->qp->qp_num;
qp_info           227 include/trace/events/ib_mad.h 		ib_query_pkey(qp_info->port_priv->device,
qp_info           228 include/trace/events/ib_mad.h 			      qp_info->port_priv->port_num,