mad_agent_priv    105 drivers/infiniband/core/mad.c static void cancel_mads(struct ib_mad_agent_private *mad_agent_priv);
mad_agent_priv    237 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv    373 drivers/infiniband/core/mad.c 	mad_agent_priv = kzalloc(sizeof *mad_agent_priv, GFP_KERNEL);
mad_agent_priv    374 drivers/infiniband/core/mad.c 	if (!mad_agent_priv) {
mad_agent_priv    388 drivers/infiniband/core/mad.c 	mad_agent_priv->qp_info = &port_priv->qp_info[qpn];
mad_agent_priv    389 drivers/infiniband/core/mad.c 	mad_agent_priv->reg_req = reg_req;
mad_agent_priv    390 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.rmpp_version = rmpp_version;
mad_agent_priv    391 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.device = device;
mad_agent_priv    392 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.recv_handler = recv_handler;
mad_agent_priv    393 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.send_handler = send_handler;
mad_agent_priv    394 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.context = context;
mad_agent_priv    395 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.qp = port_priv->qp_info[qpn].qp;
mad_agent_priv    396 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.port_num = port_num;
mad_agent_priv    397 drivers/infiniband/core/mad.c 	mad_agent_priv->agent.flags = registration_flags;
mad_agent_priv    398 drivers/infiniband/core/mad.c 	spin_lock_init(&mad_agent_priv->lock);
mad_agent_priv    399 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&mad_agent_priv->send_list);
mad_agent_priv    400 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&mad_agent_priv->wait_list);
mad_agent_priv    401 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&mad_agent_priv->done_list);
mad_agent_priv    402 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&mad_agent_priv->rmpp_list);
mad_agent_priv    403 drivers/infiniband/core/mad.c 	INIT_DELAYED_WORK(&mad_agent_priv->timed_work, timeout_sends);
mad_agent_priv    404 drivers/infiniband/core/mad.c 	INIT_LIST_HEAD(&mad_agent_priv->local_list);
mad_agent_priv    405 drivers/infiniband/core/mad.c 	INIT_WORK(&mad_agent_priv->local_work, local_completions);
mad_agent_priv    406 drivers/infiniband/core/mad.c 	atomic_set(&mad_agent_priv->refcount, 1);
mad_agent_priv    407 drivers/infiniband/core/mad.c 	init_completion(&mad_agent_priv->comp);
mad_agent_priv    409 drivers/infiniband/core/mad.c 	ret2 = ib_mad_agent_security_setup(&mad_agent_priv->agent, qp_type);
mad_agent_priv    419 drivers/infiniband/core/mad.c 	ret2 = xa_alloc_cyclic(&ib_mad_clients, &mad_agent_priv->agent.hi_tid,
mad_agent_priv    420 drivers/infiniband/core/mad.c 			mad_agent_priv, XA_LIMIT(0, (1 << 24) - 1),
mad_agent_priv    445 drivers/infiniband/core/mad.c 			ret2 = add_nonoui_reg_req(mad_reg_req, mad_agent_priv,
mad_agent_priv    461 drivers/infiniband/core/mad.c 			ret2 = add_oui_reg_req(mad_reg_req, mad_agent_priv);
mad_agent_priv    470 drivers/infiniband/core/mad.c 	trace_ib_mad_create_agent(mad_agent_priv);
mad_agent_priv    471 drivers/infiniband/core/mad.c 	return &mad_agent_priv->agent;
mad_agent_priv    474 drivers/infiniband/core/mad.c 	xa_erase(&ib_mad_clients, mad_agent_priv->agent.hi_tid);
mad_agent_priv    476 drivers/infiniband/core/mad.c 	ib_mad_agent_security_cleanup(&mad_agent_priv->agent);
mad_agent_priv    480 drivers/infiniband/core/mad.c 	kfree(mad_agent_priv);
mad_agent_priv    609 drivers/infiniband/core/mad.c static inline void deref_mad_agent(struct ib_mad_agent_private *mad_agent_priv)
mad_agent_priv    611 drivers/infiniband/core/mad.c 	if (atomic_dec_and_test(&mad_agent_priv->refcount))
mad_agent_priv    612 drivers/infiniband/core/mad.c 		complete(&mad_agent_priv->comp);
mad_agent_priv    621 drivers/infiniband/core/mad.c static void unregister_mad_agent(struct ib_mad_agent_private *mad_agent_priv)
mad_agent_priv    626 drivers/infiniband/core/mad.c 	trace_ib_mad_unregister_agent(mad_agent_priv);
mad_agent_priv    632 drivers/infiniband/core/mad.c 	cancel_mads(mad_agent_priv);
mad_agent_priv    633 drivers/infiniband/core/mad.c 	port_priv = mad_agent_priv->qp_info->port_priv;
mad_agent_priv    634 drivers/infiniband/core/mad.c 	cancel_delayed_work(&mad_agent_priv->timed_work);
mad_agent_priv    637 drivers/infiniband/core/mad.c 	remove_mad_reg_req(mad_agent_priv);
mad_agent_priv    639 drivers/infiniband/core/mad.c 	xa_erase(&ib_mad_clients, mad_agent_priv->agent.hi_tid);
mad_agent_priv    642 drivers/infiniband/core/mad.c 	ib_cancel_rmpp_recvs(mad_agent_priv);
mad_agent_priv    644 drivers/infiniband/core/mad.c 	deref_mad_agent(mad_agent_priv);
mad_agent_priv    645 drivers/infiniband/core/mad.c 	wait_for_completion(&mad_agent_priv->comp);
mad_agent_priv    647 drivers/infiniband/core/mad.c 	ib_mad_agent_security_cleanup(&mad_agent_priv->agent);
mad_agent_priv    649 drivers/infiniband/core/mad.c 	kfree(mad_agent_priv->reg_req);
mad_agent_priv    650 drivers/infiniband/core/mad.c 	kfree_rcu(mad_agent_priv, rcu);
mad_agent_priv    679 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv    684 drivers/infiniband/core/mad.c 		mad_agent_priv = container_of(mad_agent,
mad_agent_priv    687 drivers/infiniband/core/mad.c 		unregister_mad_agent(mad_agent_priv);
mad_agent_priv    808 drivers/infiniband/core/mad.c static int handle_outgoing_dr_smp(struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv    819 drivers/infiniband/core/mad.c 	struct ib_device *device = mad_agent_priv->agent.device;
mad_agent_priv    823 drivers/infiniband/core/mad.c 	size_t mad_size = port_mad_size(mad_agent_priv->qp_info->port_priv);
mad_agent_priv    826 drivers/infiniband/core/mad.c 	bool opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
mad_agent_priv    827 drivers/infiniband/core/mad.c 				    mad_agent_priv->qp_info->port_priv->port_num);
mad_agent_priv    833 drivers/infiniband/core/mad.c 		port_num = mad_agent_priv->agent.port_num;
mad_agent_priv    903 drivers/infiniband/core/mad.c 	build_smp_wc(mad_agent_priv->agent.qp,
mad_agent_priv    923 drivers/infiniband/core/mad.c 		    mad_agent_priv->agent.recv_handler) {
mad_agent_priv    925 drivers/infiniband/core/mad.c 			local->recv_mad_agent = mad_agent_priv;
mad_agent_priv    930 drivers/infiniband/core/mad.c 			atomic_inc(&mad_agent_priv->refcount);
mad_agent_priv    939 drivers/infiniband/core/mad.c 		port_priv = ib_get_mad_port(mad_agent_priv->agent.device,
mad_agent_priv    940 drivers/infiniband/core/mad.c 					    mad_agent_priv->agent.port_num);
mad_agent_priv    970 drivers/infiniband/core/mad.c 	atomic_inc(&mad_agent_priv->refcount);
mad_agent_priv    972 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv    973 drivers/infiniband/core/mad.c 	list_add_tail(&local->completion_list, &mad_agent_priv->local_list);
mad_agent_priv    974 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv    975 drivers/infiniband/core/mad.c 	queue_work(mad_agent_priv->qp_info->port_priv->wq,
mad_agent_priv    976 drivers/infiniband/core/mad.c 		   &mad_agent_priv->local_work);
mad_agent_priv   1032 drivers/infiniband/core/mad.c 	rmpp_mad->rmpp_hdr.rmpp_version = send_wr->mad_agent_priv->
mad_agent_priv   1056 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   1063 drivers/infiniband/core/mad.c 	mad_agent_priv = container_of(mad_agent, struct ib_mad_agent_private,
mad_agent_priv   1095 drivers/infiniband/core/mad.c 	mad_send_wr->mad_agent_priv = mad_agent_priv;
mad_agent_priv   1128 drivers/infiniband/core/mad.c 	atomic_inc(&mad_agent_priv->refcount);
mad_agent_priv   1196 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   1199 drivers/infiniband/core/mad.c 	mad_agent_priv = container_of(send_buf->mad_agent,
mad_agent_priv   1206 drivers/infiniband/core/mad.c 	deref_mad_agent(mad_agent_priv);
mad_agent_priv   1220 drivers/infiniband/core/mad.c 	qp_info = mad_send_wr->mad_agent_priv->qp_info;
mad_agent_priv   1282 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   1293 drivers/infiniband/core/mad.c 		mad_agent_priv = mad_send_wr->mad_agent_priv;
mad_agent_priv   1295 drivers/infiniband/core/mad.c 		ret = ib_mad_enforce_security(mad_agent_priv,
mad_agent_priv   1308 drivers/infiniband/core/mad.c 			if (mad_agent_priv->agent.rmpp_version) {
mad_agent_priv   1324 drivers/infiniband/core/mad.c 			ret = handle_outgoing_dr_smp(mad_agent_priv,
mad_agent_priv   1343 drivers/infiniband/core/mad.c 		atomic_inc(&mad_agent_priv->refcount);
mad_agent_priv   1344 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   1346 drivers/infiniband/core/mad.c 			      &mad_agent_priv->send_list);
mad_agent_priv   1347 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   1349 drivers/infiniband/core/mad.c 		if (ib_mad_kernel_rmpp_agent(&mad_agent_priv->agent)) {
mad_agent_priv   1357 drivers/infiniband/core/mad.c 			spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   1359 drivers/infiniband/core/mad.c 			spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   1360 drivers/infiniband/core/mad.c 			atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   1888 drivers/infiniband/core/mad.c static int is_rmpp_data_mad(const struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv   1894 drivers/infiniband/core/mad.c 	return !mad_agent_priv->agent.rmpp_version ||
mad_agent_priv   1895 drivers/infiniband/core/mad.c 		!ib_mad_kernel_rmpp_agent(&mad_agent_priv->agent) ||
mad_agent_priv   1908 drivers/infiniband/core/mad.c static inline int rcv_has_same_gid(const struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv   1915 drivers/infiniband/core/mad.c 	struct ib_device *device = mad_agent_priv->agent.device;
mad_agent_priv   1916 drivers/infiniband/core/mad.c 	u8 port_num = mad_agent_priv->agent.port_num;
mad_agent_priv   1970 drivers/infiniband/core/mad.c ib_find_send_mad(const struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv   1978 drivers/infiniband/core/mad.c 	list_for_each_entry(wr, &mad_agent_priv->wait_list, agent_list) {
mad_agent_priv   1986 drivers/infiniband/core/mad.c 		     rcv_has_same_gid(mad_agent_priv, wr, wc)))
mad_agent_priv   1994 drivers/infiniband/core/mad.c 	list_for_each_entry(wr, &mad_agent_priv->send_list, agent_list) {
mad_agent_priv   1995 drivers/infiniband/core/mad.c 		if (is_rmpp_data_mad(mad_agent_priv, wr->send_buf.mad) &&
mad_agent_priv   2004 drivers/infiniband/core/mad.c 		     rcv_has_same_gid(mad_agent_priv, wr, wc)))
mad_agent_priv   2016 drivers/infiniband/core/mad.c 			      &mad_send_wr->mad_agent_priv->done_list);
mad_agent_priv   2019 drivers/infiniband/core/mad.c static void ib_mad_complete_recv(struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv   2028 drivers/infiniband/core/mad.c 	ret = ib_mad_enforce_security(mad_agent_priv,
mad_agent_priv   2032 drivers/infiniband/core/mad.c 		deref_mad_agent(mad_agent_priv);
mad_agent_priv   2037 drivers/infiniband/core/mad.c 	if (ib_mad_kernel_rmpp_agent(&mad_agent_priv->agent)) {
mad_agent_priv   2038 drivers/infiniband/core/mad.c 		mad_recv_wc = ib_process_rmpp_recv_wc(mad_agent_priv,
mad_agent_priv   2041 drivers/infiniband/core/mad.c 			deref_mad_agent(mad_agent_priv);
mad_agent_priv   2048 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2049 drivers/infiniband/core/mad.c 		mad_send_wr = ib_find_send_mad(mad_agent_priv, mad_recv_wc);
mad_agent_priv   2051 drivers/infiniband/core/mad.c 			spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2052 drivers/infiniband/core/mad.c 			if (!ib_mad_kernel_rmpp_agent(&mad_agent_priv->agent)
mad_agent_priv   2059 drivers/infiniband/core/mad.c 				mad_agent_priv->agent.recv_handler(
mad_agent_priv   2060 drivers/infiniband/core/mad.c 						&mad_agent_priv->agent, NULL,
mad_agent_priv   2062 drivers/infiniband/core/mad.c 				atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   2067 drivers/infiniband/core/mad.c 				deref_mad_agent(mad_agent_priv);
mad_agent_priv   2072 drivers/infiniband/core/mad.c 			spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2075 drivers/infiniband/core/mad.c 			mad_agent_priv->agent.recv_handler(
mad_agent_priv   2076 drivers/infiniband/core/mad.c 					&mad_agent_priv->agent,
mad_agent_priv   2079 drivers/infiniband/core/mad.c 			atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   2087 drivers/infiniband/core/mad.c 		mad_agent_priv->agent.recv_handler(&mad_agent_priv->agent, NULL,
mad_agent_priv   2089 drivers/infiniband/core/mad.c 		deref_mad_agent(mad_agent_priv);
mad_agent_priv   2390 drivers/infiniband/core/mad.c static void adjust_timeout(struct ib_mad_agent_private *mad_agent_priv)
mad_agent_priv   2395 drivers/infiniband/core/mad.c 	if (list_empty(&mad_agent_priv->wait_list)) {
mad_agent_priv   2396 drivers/infiniband/core/mad.c 		cancel_delayed_work(&mad_agent_priv->timed_work);
mad_agent_priv   2398 drivers/infiniband/core/mad.c 		mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
mad_agent_priv   2402 drivers/infiniband/core/mad.c 		if (time_after(mad_agent_priv->timeout,
mad_agent_priv   2404 drivers/infiniband/core/mad.c 			mad_agent_priv->timeout = mad_send_wr->timeout;
mad_agent_priv   2408 drivers/infiniband/core/mad.c 			mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
mad_agent_priv   2409 drivers/infiniband/core/mad.c 					 &mad_agent_priv->timed_work, delay);
mad_agent_priv   2416 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   2421 drivers/infiniband/core/mad.c 	mad_agent_priv = mad_send_wr->mad_agent_priv;
mad_agent_priv   2428 drivers/infiniband/core/mad.c 		list_for_each_prev(list_item, &mad_agent_priv->wait_list) {
mad_agent_priv   2438 drivers/infiniband/core/mad.c 		list_item = &mad_agent_priv->wait_list;
mad_agent_priv   2442 drivers/infiniband/core/mad.c 	if (mad_agent_priv->wait_list.next == &mad_send_wr->agent_list)
mad_agent_priv   2443 drivers/infiniband/core/mad.c 		mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
mad_agent_priv   2444 drivers/infiniband/core/mad.c 				 &mad_agent_priv->timed_work, delay);
mad_agent_priv   2460 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private	*mad_agent_priv;
mad_agent_priv   2464 drivers/infiniband/core/mad.c 	mad_agent_priv = mad_send_wr->mad_agent_priv;
mad_agent_priv   2465 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2466 drivers/infiniband/core/mad.c 	if (ib_mad_kernel_rmpp_agent(&mad_agent_priv->agent)) {
mad_agent_priv   2489 drivers/infiniband/core/mad.c 	adjust_timeout(mad_agent_priv);
mad_agent_priv   2490 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2497 drivers/infiniband/core/mad.c 		mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,
mad_agent_priv   2501 drivers/infiniband/core/mad.c 	deref_mad_agent(mad_agent_priv);
mad_agent_priv   2504 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2532 drivers/infiniband/core/mad.c 	trace_ib_mad_send_done_agent(mad_send_wr->mad_agent_priv);
mad_agent_priv   2643 drivers/infiniband/core/mad.c static void cancel_mads(struct ib_mad_agent_private *mad_agent_priv)
mad_agent_priv   2652 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2654 drivers/infiniband/core/mad.c 				 &mad_agent_priv->send_list, agent_list) {
mad_agent_priv   2662 drivers/infiniband/core/mad.c 	list_splice_init(&mad_agent_priv->wait_list, &cancel_list);
mad_agent_priv   2663 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2673 drivers/infiniband/core/mad.c 		mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,
mad_agent_priv   2675 drivers/infiniband/core/mad.c 		atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   2680 drivers/infiniband/core/mad.c find_send_wr(struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv   2685 drivers/infiniband/core/mad.c 	list_for_each_entry(mad_send_wr, &mad_agent_priv->wait_list,
mad_agent_priv   2691 drivers/infiniband/core/mad.c 	list_for_each_entry(mad_send_wr, &mad_agent_priv->send_list,
mad_agent_priv   2693 drivers/infiniband/core/mad.c 		if (is_rmpp_data_mad(mad_agent_priv,
mad_agent_priv   2704 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   2709 drivers/infiniband/core/mad.c 	mad_agent_priv = container_of(mad_agent, struct ib_mad_agent_private,
mad_agent_priv   2711 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2712 drivers/infiniband/core/mad.c 	mad_send_wr = find_send_wr(mad_agent_priv, send_buf);
mad_agent_priv   2714 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2730 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2744 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   2753 drivers/infiniband/core/mad.c 	mad_agent_priv =
mad_agent_priv   2756 drivers/infiniband/core/mad.c 	opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
mad_agent_priv   2757 drivers/infiniband/core/mad.c 			       mad_agent_priv->qp_info->port_priv->port_num);
mad_agent_priv   2759 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2760 drivers/infiniband/core/mad.c 	while (!list_empty(&mad_agent_priv->local_list)) {
mad_agent_priv   2761 drivers/infiniband/core/mad.c 		local = list_entry(mad_agent_priv->local_list.next,
mad_agent_priv   2765 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2771 drivers/infiniband/core/mad.c 				dev_err(&mad_agent_priv->agent.device->dev,
mad_agent_priv   2822 drivers/infiniband/core/mad.c 		if (atomic_read(&mad_agent_priv->qp_info->snoop_count))
mad_agent_priv   2823 drivers/infiniband/core/mad.c 			snoop_send(mad_agent_priv->qp_info,
mad_agent_priv   2826 drivers/infiniband/core/mad.c 		mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,
mad_agent_priv   2829 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2830 drivers/infiniband/core/mad.c 		atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   2835 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2850 drivers/infiniband/core/mad.c 	if (ib_mad_kernel_rmpp_agent(&mad_send_wr->mad_agent_priv->agent)) {
mad_agent_priv   2869 drivers/infiniband/core/mad.c 			      &mad_send_wr->mad_agent_priv->send_list);
mad_agent_priv   2876 drivers/infiniband/core/mad.c 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv   2881 drivers/infiniband/core/mad.c 	mad_agent_priv = container_of(work, struct ib_mad_agent_private,
mad_agent_priv   2885 drivers/infiniband/core/mad.c 	spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2886 drivers/infiniband/core/mad.c 	while (!list_empty(&mad_agent_priv->wait_list)) {
mad_agent_priv   2887 drivers/infiniband/core/mad.c 		mad_send_wr = list_entry(mad_agent_priv->wait_list.next,
mad_agent_priv   2895 drivers/infiniband/core/mad.c 			queue_delayed_work(mad_agent_priv->qp_info->
mad_agent_priv   2897 drivers/infiniband/core/mad.c 					   &mad_agent_priv->timed_work, delay);
mad_agent_priv   2906 drivers/infiniband/core/mad.c 		spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv   2913 drivers/infiniband/core/mad.c 		mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,
mad_agent_priv   2916 drivers/infiniband/core/mad.c 		atomic_dec(&mad_agent_priv->refcount);
mad_agent_priv   2917 drivers/infiniband/core/mad.c 		spin_lock_irqsave(&mad_agent_priv->lock, flags);
mad_agent_priv   2919 drivers/infiniband/core/mad.c 	spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
mad_agent_priv    125 drivers/infiniband/core/mad_priv.h 	struct ib_mad_agent_private *mad_agent_priv;
mad_agent_priv    215 drivers/infiniband/core/mad_priv.h ib_find_send_mad(const struct ib_mad_agent_private *mad_agent_priv,
mad_agent_priv    745 drivers/infiniband/core/mad_rmpp.c 			      &mad_send_wr->mad_agent_priv->send_list);
mad_agent_priv    853 drivers/infiniband/core/mad_rmpp.c 	struct ib_mad_agent_private *agent = mad_send_wr->mad_agent_priv;
mad_agent_priv     56 include/trace/events/ib_mad.h 		__entry->dev_index = wr->mad_agent_priv->agent.device->index;
mad_agent_priv     57 include/trace/events/ib_mad.h 		__entry->port_num = wr->mad_agent_priv->agent.port_num;
mad_agent_priv     58 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
mad_agent_priv     59 include/trace/events/ib_mad.h 		__entry->agent_priv = wr->mad_agent_priv;
mad_agent_priv    145 include/trace/events/ib_mad.h 		__entry->dev_index = wr->mad_agent_priv->agent.device->index;
mad_agent_priv    146 include/trace/events/ib_mad.h 		__entry->port_num = wr->mad_agent_priv->agent.port_num;
mad_agent_priv    147 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
mad_agent_priv    148 include/trace/events/ib_mad.h 		__entry->agent_priv = wr->mad_agent_priv;