Lines Matching refs:intf
57 static void handle_new_recv_msgs(ipmi_smi_t intf);
58 static void need_waiter(ipmi_smi_t intf);
59 static int handle_one_recv_msg(ipmi_smi_t intf,
109 ipmi_smi_t intf; member
424 void (*null_user_handler)(ipmi_smi_t intf, struct ipmi_recv_msg *msg);
470 #define ipmi_inc_stat(intf, stat) \ argument
471 atomic_inc(&(intf)->stats[IPMI_STAT_ ## stat])
472 #define ipmi_get_stat(intf, stat) \ argument
473 ((unsigned int) atomic_read(&(intf)->stats[IPMI_STAT_ ## stat]))
522 static void clean_up_interface_data(ipmi_smi_t intf) in clean_up_interface_data() argument
528 tasklet_kill(&intf->recv_tasklet); in clean_up_interface_data()
530 free_smi_msg_list(&intf->waiting_rcv_msgs); in clean_up_interface_data()
531 free_recv_msg_list(&intf->waiting_events); in clean_up_interface_data()
537 mutex_lock(&intf->cmd_rcvrs_mutex); in clean_up_interface_data()
539 list_splice_init_rcu(&intf->cmd_rcvrs, &list, synchronize_rcu); in clean_up_interface_data()
540 mutex_unlock(&intf->cmd_rcvrs_mutex); in clean_up_interface_data()
546 if ((intf->seq_table[i].inuse) in clean_up_interface_data()
547 && (intf->seq_table[i].recv_msg)) in clean_up_interface_data()
548 ipmi_free_recv_msg(intf->seq_table[i].recv_msg); in clean_up_interface_data()
554 ipmi_smi_t intf = container_of(ref, struct ipmi_smi, refcount); in intf_free() local
556 clean_up_interface_data(intf); in intf_free()
557 kfree(intf); in intf_free()
562 ipmi_smi_t intf; member
568 ipmi_smi_t intf; in ipmi_smi_watcher_register() local
577 list_for_each_entry(intf, &ipmi_interfaces, link) { in ipmi_smi_watcher_register()
578 if (intf->intf_num == -1) in ipmi_smi_watcher_register()
583 kref_get(&intf->refcount); in ipmi_smi_watcher_register()
584 e->intf = intf; in ipmi_smi_watcher_register()
585 e->intf_num = intf->intf_num; in ipmi_smi_watcher_register()
596 watcher->new_smi(e->intf_num, e->intf->si_dev); in ipmi_smi_watcher_register()
597 kref_put(&e->intf->refcount, intf_free); in ipmi_smi_watcher_register()
610 kref_put(&e->intf->refcount, intf_free); in ipmi_smi_watcher_register()
736 ipmi_smi_t intf = msg->user_msg_data; in deliver_response() local
739 if (intf->null_user_handler) { in deliver_response()
740 intf->null_user_handler(intf, msg); in deliver_response()
741 ipmi_inc_stat(intf, handled_local_responses); in deliver_response()
744 ipmi_inc_stat(intf, unhandled_local_responses); in deliver_response()
769 static int intf_next_seq(ipmi_smi_t intf, in intf_next_seq() argument
780 for (i = intf->curr_seq; (i+1)%IPMI_IPMB_NUM_SEQ != intf->curr_seq; in intf_next_seq()
782 if (!intf->seq_table[i].inuse) in intf_next_seq()
786 if (!intf->seq_table[i].inuse) { in intf_next_seq()
787 intf->seq_table[i].recv_msg = recv_msg; in intf_next_seq()
793 intf->seq_table[i].timeout = MAX_MSG_TIMEOUT; in intf_next_seq()
794 intf->seq_table[i].orig_timeout = timeout; in intf_next_seq()
795 intf->seq_table[i].retries_left = retries; in intf_next_seq()
796 intf->seq_table[i].broadcast = broadcast; in intf_next_seq()
797 intf->seq_table[i].inuse = 1; in intf_next_seq()
798 intf->seq_table[i].seqid = NEXT_SEQID(intf->seq_table[i].seqid); in intf_next_seq()
800 *seqid = intf->seq_table[i].seqid; in intf_next_seq()
801 intf->curr_seq = (i+1)%IPMI_IPMB_NUM_SEQ; in intf_next_seq()
802 need_waiter(intf); in intf_next_seq()
817 static int intf_find_seq(ipmi_smi_t intf, in intf_find_seq() argument
831 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_find_seq()
832 if (intf->seq_table[seq].inuse) { in intf_find_seq()
833 struct ipmi_recv_msg *msg = intf->seq_table[seq].recv_msg; in intf_find_seq()
839 intf->seq_table[seq].inuse = 0; in intf_find_seq()
843 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_find_seq()
850 static int intf_start_seq_timer(ipmi_smi_t intf, in intf_start_seq_timer() argument
861 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_start_seq_timer()
866 if ((intf->seq_table[seq].inuse) in intf_start_seq_timer()
867 && (intf->seq_table[seq].seqid == seqid)) { in intf_start_seq_timer()
868 struct seq_table *ent = &(intf->seq_table[seq]); in intf_start_seq_timer()
872 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_start_seq_timer()
878 static int intf_err_seq(ipmi_smi_t intf, in intf_err_seq() argument
891 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_err_seq()
896 if ((intf->seq_table[seq].inuse) in intf_err_seq()
897 && (intf->seq_table[seq].seqid == seqid)) { in intf_err_seq()
898 struct seq_table *ent = &(intf->seq_table[seq]); in intf_err_seq()
904 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_err_seq()
921 ipmi_smi_t intf; in ipmi_create_user() local
956 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_create_user()
957 if (intf->intf_num == if_num) in ipmi_create_user()
966 kref_get(&intf->refcount); in ipmi_create_user()
971 new_user->intf = intf; in ipmi_create_user()
974 if (!try_module_get(intf->handlers->owner)) { in ipmi_create_user()
979 if (intf->handlers->inc_usecount) { in ipmi_create_user()
980 rv = intf->handlers->inc_usecount(intf->send_info); in ipmi_create_user()
982 module_put(intf->handlers->owner); in ipmi_create_user()
994 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_create_user()
995 list_add_rcu(&new_user->link, &intf->users); in ipmi_create_user()
996 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_create_user()
999 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_create_user()
1000 need_waiter(intf); in ipmi_create_user()
1006 kref_put(&intf->refcount, intf_free); in ipmi_create_user()
1017 ipmi_smi_t intf; in ipmi_get_smi_info() local
1021 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_get_smi_info()
1022 if (intf->intf_num == if_num) in ipmi_get_smi_info()
1031 handlers = intf->handlers; in ipmi_get_smi_info()
1034 rv = handlers->get_smi_info(intf->send_info, data); in ipmi_get_smi_info()
1049 ipmi_smi_t intf = user->intf; in ipmi_destroy_user() local
1058 atomic_dec(&intf->event_waiters); in ipmi_destroy_user()
1061 atomic_dec(&intf->event_waiters); in ipmi_destroy_user()
1064 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_destroy_user()
1068 if (intf->seq_table[i].inuse in ipmi_destroy_user()
1069 && (intf->seq_table[i].recv_msg->user == user)) { in ipmi_destroy_user()
1070 intf->seq_table[i].inuse = 0; in ipmi_destroy_user()
1071 ipmi_free_recv_msg(intf->seq_table[i].recv_msg); in ipmi_destroy_user()
1074 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_destroy_user()
1082 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_destroy_user()
1083 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in ipmi_destroy_user()
1090 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_destroy_user()
1099 if (intf->handlers) { in ipmi_destroy_user()
1100 module_put(intf->handlers->owner); in ipmi_destroy_user()
1101 if (intf->handlers->dec_usecount) in ipmi_destroy_user()
1102 intf->handlers->dec_usecount(intf->send_info); in ipmi_destroy_user()
1106 kref_put(&intf->refcount, intf_free); in ipmi_destroy_user()
1118 *major = user->intf->ipmi_version_major; in ipmi_get_version()
1119 *minor = user->intf->ipmi_version_minor; in ipmi_get_version()
1129 user->intf->channels[channel].address = address; in ipmi_set_my_address()
1140 *address = user->intf->channels[channel].address; in ipmi_get_my_address()
1151 user->intf->channels[channel].lun = LUN & 0x3; in ipmi_set_my_LUN()
1162 *address = user->intf->channels[channel].lun; in ipmi_get_my_LUN()
1172 spin_lock_irqsave(&user->intf->maintenance_mode_lock, flags); in ipmi_get_maintenance_mode()
1173 mode = user->intf->maintenance_mode; in ipmi_get_maintenance_mode()
1174 spin_unlock_irqrestore(&user->intf->maintenance_mode_lock, flags); in ipmi_get_maintenance_mode()
1180 static void maintenance_mode_update(ipmi_smi_t intf) in maintenance_mode_update() argument
1182 if (intf->handlers->set_maintenance_mode) in maintenance_mode_update()
1183 intf->handlers->set_maintenance_mode( in maintenance_mode_update()
1184 intf->send_info, intf->maintenance_mode_enable); in maintenance_mode_update()
1191 ipmi_smi_t intf = user->intf; in ipmi_set_maintenance_mode() local
1193 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in ipmi_set_maintenance_mode()
1194 if (intf->maintenance_mode != mode) { in ipmi_set_maintenance_mode()
1197 intf->maintenance_mode_enable in ipmi_set_maintenance_mode()
1198 = (intf->auto_maintenance_timeout > 0); in ipmi_set_maintenance_mode()
1202 intf->maintenance_mode_enable = false; in ipmi_set_maintenance_mode()
1206 intf->maintenance_mode_enable = true; in ipmi_set_maintenance_mode()
1213 intf->maintenance_mode = mode; in ipmi_set_maintenance_mode()
1215 maintenance_mode_update(intf); in ipmi_set_maintenance_mode()
1218 spin_unlock_irqrestore(&intf->maintenance_mode_lock, flags); in ipmi_set_maintenance_mode()
1227 ipmi_smi_t intf = user->intf; in ipmi_set_gets_events() local
1233 spin_lock_irqsave(&intf->events_lock, flags); in ipmi_set_gets_events()
1240 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_set_gets_events()
1241 need_waiter(intf); in ipmi_set_gets_events()
1243 atomic_dec(&intf->event_waiters); in ipmi_set_gets_events()
1246 if (intf->delivering_events) in ipmi_set_gets_events()
1254 while (user->gets_events && !list_empty(&intf->waiting_events)) { in ipmi_set_gets_events()
1255 list_for_each_entry_safe(msg, msg2, &intf->waiting_events, link) in ipmi_set_gets_events()
1257 intf->waiting_events_count = 0; in ipmi_set_gets_events()
1258 if (intf->event_msg_printed) { in ipmi_set_gets_events()
1261 intf->event_msg_printed = 0; in ipmi_set_gets_events()
1264 intf->delivering_events = 1; in ipmi_set_gets_events()
1265 spin_unlock_irqrestore(&intf->events_lock, flags); in ipmi_set_gets_events()
1273 spin_lock_irqsave(&intf->events_lock, flags); in ipmi_set_gets_events()
1274 intf->delivering_events = 0; in ipmi_set_gets_events()
1278 spin_unlock_irqrestore(&intf->events_lock, flags); in ipmi_set_gets_events()
1284 static struct cmd_rcvr *find_cmd_rcvr(ipmi_smi_t intf, in find_cmd_rcvr() argument
1291 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in find_cmd_rcvr()
1299 static int is_cmd_rcvr_exclusive(ipmi_smi_t intf, in is_cmd_rcvr_exclusive() argument
1306 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in is_cmd_rcvr_exclusive()
1319 ipmi_smi_t intf = user->intf; in ipmi_register_for_cmd() local
1332 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_register_for_cmd()
1334 if (!is_cmd_rcvr_exclusive(intf, netfn, cmd, chans)) { in ipmi_register_for_cmd()
1339 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_register_for_cmd()
1340 need_waiter(intf); in ipmi_register_for_cmd()
1342 list_add_rcu(&rcvr->link, &intf->cmd_rcvrs); in ipmi_register_for_cmd()
1345 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_register_for_cmd()
1358 ipmi_smi_t intf = user->intf; in ipmi_unregister_for_cmd() local
1363 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_unregister_for_cmd()
1367 rcvr = find_cmd_rcvr(intf, netfn, cmd, i); in ipmi_unregister_for_cmd()
1380 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_unregister_for_cmd()
1383 atomic_dec(&intf->event_waiters); in ipmi_unregister_for_cmd()
1486 static struct ipmi_smi_msg *smi_add_send_msg(ipmi_smi_t intf, in smi_add_send_msg() argument
1490 if (intf->curr_msg) { in smi_add_send_msg()
1492 list_add_tail(&smi_msg->link, &intf->hp_xmit_msgs); in smi_add_send_msg()
1494 list_add_tail(&smi_msg->link, &intf->xmit_msgs); in smi_add_send_msg()
1497 intf->curr_msg = smi_msg; in smi_add_send_msg()
1504 static void smi_send(ipmi_smi_t intf, struct ipmi_smi_handlers *handlers, in smi_send() argument
1507 int run_to_completion = intf->run_to_completion; in smi_send()
1510 smi_msg = smi_add_send_msg(intf, smi_msg, priority); in smi_send()
1514 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in smi_send()
1515 smi_msg = smi_add_send_msg(intf, smi_msg, priority); in smi_send()
1516 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in smi_send()
1520 handlers->sender(intf->send_info, smi_msg); in smi_send()
1530 ipmi_smi_t intf, in i_ipmi_request() argument
1569 if (intf->in_shutdown) { in i_ipmi_request()
1595 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1610 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1619 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in i_ipmi_request()
1620 intf->auto_maintenance_timeout in i_ipmi_request()
1622 if (!intf->maintenance_mode in i_ipmi_request()
1623 && !intf->maintenance_mode_enable) { in i_ipmi_request()
1624 intf->maintenance_mode_enable = true; in i_ipmi_request()
1625 maintenance_mode_update(intf); in i_ipmi_request()
1627 spin_unlock_irqrestore(&intf->maintenance_mode_lock, in i_ipmi_request()
1632 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1644 ipmi_inc_stat(intf, sent_local_commands); in i_ipmi_request()
1652 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1657 if (intf->channels[addr->channel].medium in i_ipmi_request()
1659 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1690 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1697 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1709 ipmi_inc_stat(intf, sent_ipmb_responses); in i_ipmi_request()
1722 spin_lock_irqsave(&(intf->seq_lock), flags); in i_ipmi_request()
1728 rv = intf_next_seq(intf, in i_ipmi_request()
1740 spin_unlock_irqrestore(&(intf->seq_lock), in i_ipmi_request()
1745 ipmi_inc_stat(intf, sent_ipmb_commands); in i_ipmi_request()
1774 spin_unlock_irqrestore(&(intf->seq_lock), flags); in i_ipmi_request()
1782 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1787 if ((intf->channels[addr->channel].medium in i_ipmi_request()
1789 && (intf->channels[addr->channel].medium in i_ipmi_request()
1791 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1804 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1811 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1823 ipmi_inc_stat(intf, sent_lan_responses); in i_ipmi_request()
1835 spin_lock_irqsave(&(intf->seq_lock), flags); in i_ipmi_request()
1841 rv = intf_next_seq(intf, in i_ipmi_request()
1853 spin_unlock_irqrestore(&(intf->seq_lock), in i_ipmi_request()
1858 ipmi_inc_stat(intf, sent_lan_commands); in i_ipmi_request()
1886 spin_unlock_irqrestore(&(intf->seq_lock), flags); in i_ipmi_request()
1890 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1904 smi_send(intf, intf->handlers, smi_msg, priority); in i_ipmi_request()
1916 static int check_addr(ipmi_smi_t intf, in check_addr() argument
1923 *lun = intf->channels[addr->channel].lun; in check_addr()
1924 *saddr = intf->channels[addr->channel].address; in check_addr()
1942 rv = check_addr(user->intf, addr, &saddr, &lun); in ipmi_request_settime()
1946 user->intf, in ipmi_request_settime()
1974 rv = check_addr(user->intf, addr, &saddr, &lun); in ipmi_request_supply_msgs()
1978 user->intf, in ipmi_request_supply_msgs()
1995 ipmi_smi_t intf = m->private; in smi_ipmb_proc_show() local
1998 seq_printf(m, "%x", intf->channels[0].address); in smi_ipmb_proc_show()
2000 seq_printf(m, " %x", intf->channels[i].address); in smi_ipmb_proc_show()
2020 ipmi_smi_t intf = m->private; in smi_version_proc_show() local
2023 ipmi_version_major(&intf->bmc->id), in smi_version_proc_show()
2024 ipmi_version_minor(&intf->bmc->id)); in smi_version_proc_show()
2043 ipmi_smi_t intf = m->private; in smi_stats_proc_show() local
2046 ipmi_get_stat(intf, sent_invalid_commands)); in smi_stats_proc_show()
2048 ipmi_get_stat(intf, sent_local_commands)); in smi_stats_proc_show()
2050 ipmi_get_stat(intf, handled_local_responses)); in smi_stats_proc_show()
2052 ipmi_get_stat(intf, unhandled_local_responses)); in smi_stats_proc_show()
2054 ipmi_get_stat(intf, sent_ipmb_commands)); in smi_stats_proc_show()
2056 ipmi_get_stat(intf, sent_ipmb_command_errs)); in smi_stats_proc_show()
2058 ipmi_get_stat(intf, retransmitted_ipmb_commands)); in smi_stats_proc_show()
2060 ipmi_get_stat(intf, timed_out_ipmb_commands)); in smi_stats_proc_show()
2062 ipmi_get_stat(intf, timed_out_ipmb_broadcasts)); in smi_stats_proc_show()
2064 ipmi_get_stat(intf, sent_ipmb_responses)); in smi_stats_proc_show()
2066 ipmi_get_stat(intf, handled_ipmb_responses)); in smi_stats_proc_show()
2068 ipmi_get_stat(intf, invalid_ipmb_responses)); in smi_stats_proc_show()
2070 ipmi_get_stat(intf, unhandled_ipmb_responses)); in smi_stats_proc_show()
2072 ipmi_get_stat(intf, sent_lan_commands)); in smi_stats_proc_show()
2074 ipmi_get_stat(intf, sent_lan_command_errs)); in smi_stats_proc_show()
2076 ipmi_get_stat(intf, retransmitted_lan_commands)); in smi_stats_proc_show()
2078 ipmi_get_stat(intf, timed_out_lan_commands)); in smi_stats_proc_show()
2080 ipmi_get_stat(intf, sent_lan_responses)); in smi_stats_proc_show()
2082 ipmi_get_stat(intf, handled_lan_responses)); in smi_stats_proc_show()
2084 ipmi_get_stat(intf, invalid_lan_responses)); in smi_stats_proc_show()
2086 ipmi_get_stat(intf, unhandled_lan_responses)); in smi_stats_proc_show()
2088 ipmi_get_stat(intf, handled_commands)); in smi_stats_proc_show()
2090 ipmi_get_stat(intf, invalid_commands)); in smi_stats_proc_show()
2092 ipmi_get_stat(intf, unhandled_commands)); in smi_stats_proc_show()
2094 ipmi_get_stat(intf, invalid_events)); in smi_stats_proc_show()
2096 ipmi_get_stat(intf, events)); in smi_stats_proc_show()
2098 ipmi_get_stat(intf, dropped_rexmit_lan_commands)); in smi_stats_proc_show()
2100 ipmi_get_stat(intf, dropped_rexmit_ipmb_commands)); in smi_stats_proc_show()
2420 static void ipmi_bmc_unregister(ipmi_smi_t intf) in ipmi_bmc_unregister() argument
2422 struct bmc_device *bmc = intf->bmc; in ipmi_bmc_unregister()
2424 sysfs_remove_link(&intf->si_dev->kobj, "bmc"); in ipmi_bmc_unregister()
2425 if (intf->my_dev_name) { in ipmi_bmc_unregister()
2426 sysfs_remove_link(&bmc->pdev.dev.kobj, intf->my_dev_name); in ipmi_bmc_unregister()
2427 kfree(intf->my_dev_name); in ipmi_bmc_unregister()
2428 intf->my_dev_name = NULL; in ipmi_bmc_unregister()
2433 intf->bmc = NULL; in ipmi_bmc_unregister()
2437 static int ipmi_bmc_register(ipmi_smi_t intf, int ifnum) in ipmi_bmc_register() argument
2440 struct bmc_device *bmc = intf->bmc; in ipmi_bmc_register()
2462 intf->bmc = old_bmc; in ipmi_bmc_register()
2525 dev_info(intf->si_dev, "Found new BMC (man_id: 0x%6.6x, " in ipmi_bmc_register()
2536 rv = sysfs_create_link(&intf->si_dev->kobj, &bmc->pdev.dev.kobj, "bmc"); in ipmi_bmc_register()
2544 intf->my_dev_name = kasprintf(GFP_KERNEL, "ipmi%d", ifnum); in ipmi_bmc_register()
2545 if (!intf->my_dev_name) { in ipmi_bmc_register()
2553 rv = sysfs_create_link(&bmc->pdev.dev.kobj, &intf->si_dev->kobj, in ipmi_bmc_register()
2554 intf->my_dev_name); in ipmi_bmc_register()
2556 kfree(intf->my_dev_name); in ipmi_bmc_register()
2557 intf->my_dev_name = NULL; in ipmi_bmc_register()
2568 ipmi_bmc_unregister(intf); in ipmi_bmc_register()
2573 send_guid_cmd(ipmi_smi_t intf, int chan) in send_guid_cmd() argument
2587 intf, in send_guid_cmd()
2591 intf, in send_guid_cmd()
2595 intf->channels[0].address, in send_guid_cmd()
2596 intf->channels[0].lun, in send_guid_cmd()
2601 guid_handler(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in guid_handler() argument
2611 intf->bmc->guid_set = 0; in guid_handler()
2616 intf->bmc->guid_set = 0; in guid_handler()
2625 memcpy(intf->bmc->guid, msg->msg.data, 16); in guid_handler()
2626 intf->bmc->guid_set = 1; in guid_handler()
2628 wake_up(&intf->waitq); in guid_handler()
2632 get_guid(ipmi_smi_t intf) in get_guid() argument
2636 intf->bmc->guid_set = 0x2; in get_guid()
2637 intf->null_user_handler = guid_handler; in get_guid()
2638 rv = send_guid_cmd(intf, 0); in get_guid()
2641 intf->bmc->guid_set = 0; in get_guid()
2642 wait_event(intf->waitq, intf->bmc->guid_set != 2); in get_guid()
2643 intf->null_user_handler = NULL; in get_guid()
2647 send_channel_info_cmd(ipmi_smi_t intf, int chan) in send_channel_info_cmd() argument
2663 intf, in send_channel_info_cmd()
2667 intf, in send_channel_info_cmd()
2671 intf->channels[0].address, in send_channel_info_cmd()
2672 intf->channels[0].lun, in send_channel_info_cmd()
2677 channel_handler(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in channel_handler() argument
2696 intf->channels[0].medium in channel_handler()
2698 intf->channels[0].protocol in channel_handler()
2701 intf->curr_channel = IPMI_MAX_CHANNELS; in channel_handler()
2702 wake_up(&intf->waitq); in channel_handler()
2711 chan = intf->curr_channel; in channel_handler()
2712 intf->channels[chan].medium = msg->msg.data[2] & 0x7f; in channel_handler()
2713 intf->channels[chan].protocol = msg->msg.data[3] & 0x1f; in channel_handler()
2716 intf->curr_channel++; in channel_handler()
2717 if (intf->curr_channel >= IPMI_MAX_CHANNELS) in channel_handler()
2718 wake_up(&intf->waitq); in channel_handler()
2720 rv = send_channel_info_cmd(intf, intf->curr_channel); in channel_handler()
2726 " %d: %d\n", intf->curr_channel, rv); in channel_handler()
2728 intf->curr_channel = IPMI_MAX_CHANNELS; in channel_handler()
2729 wake_up(&intf->waitq); in channel_handler()
2736 static void ipmi_poll(ipmi_smi_t intf) in ipmi_poll() argument
2738 if (intf->handlers->poll) in ipmi_poll()
2739 intf->handlers->poll(intf->send_info); in ipmi_poll()
2741 handle_new_recv_msgs(intf); in ipmi_poll()
2746 ipmi_poll(user->intf); in ipmi_poll_interface()
2758 ipmi_smi_t intf; in ipmi_register_smi() local
2778 intf = kzalloc(sizeof(*intf), GFP_KERNEL); in ipmi_register_smi()
2779 if (!intf) in ipmi_register_smi()
2782 intf->ipmi_version_major = ipmi_version_major(device_id); in ipmi_register_smi()
2783 intf->ipmi_version_minor = ipmi_version_minor(device_id); in ipmi_register_smi()
2785 intf->bmc = kzalloc(sizeof(*intf->bmc), GFP_KERNEL); in ipmi_register_smi()
2786 if (!intf->bmc) { in ipmi_register_smi()
2787 kfree(intf); in ipmi_register_smi()
2790 intf->intf_num = -1; /* Mark it invalid for now. */ in ipmi_register_smi()
2791 kref_init(&intf->refcount); in ipmi_register_smi()
2792 intf->bmc->id = *device_id; in ipmi_register_smi()
2793 intf->si_dev = si_dev; in ipmi_register_smi()
2795 intf->channels[j].address = IPMI_BMC_SLAVE_ADDR; in ipmi_register_smi()
2796 intf->channels[j].lun = 2; in ipmi_register_smi()
2799 intf->channels[0].address = slave_addr; in ipmi_register_smi()
2800 INIT_LIST_HEAD(&intf->users); in ipmi_register_smi()
2801 intf->handlers = handlers; in ipmi_register_smi()
2802 intf->send_info = send_info; in ipmi_register_smi()
2803 spin_lock_init(&intf->seq_lock); in ipmi_register_smi()
2805 intf->seq_table[j].inuse = 0; in ipmi_register_smi()
2806 intf->seq_table[j].seqid = 0; in ipmi_register_smi()
2808 intf->curr_seq = 0; in ipmi_register_smi()
2810 mutex_init(&intf->proc_entry_lock); in ipmi_register_smi()
2812 spin_lock_init(&intf->waiting_rcv_msgs_lock); in ipmi_register_smi()
2813 INIT_LIST_HEAD(&intf->waiting_rcv_msgs); in ipmi_register_smi()
2814 tasklet_init(&intf->recv_tasklet, in ipmi_register_smi()
2816 (unsigned long) intf); in ipmi_register_smi()
2817 atomic_set(&intf->watchdog_pretimeouts_to_deliver, 0); in ipmi_register_smi()
2818 spin_lock_init(&intf->xmit_msgs_lock); in ipmi_register_smi()
2819 INIT_LIST_HEAD(&intf->xmit_msgs); in ipmi_register_smi()
2820 INIT_LIST_HEAD(&intf->hp_xmit_msgs); in ipmi_register_smi()
2821 spin_lock_init(&intf->events_lock); in ipmi_register_smi()
2822 atomic_set(&intf->event_waiters, 0); in ipmi_register_smi()
2823 intf->ticks_to_req_ev = IPMI_REQUEST_EV_TIME; in ipmi_register_smi()
2824 INIT_LIST_HEAD(&intf->waiting_events); in ipmi_register_smi()
2825 intf->waiting_events_count = 0; in ipmi_register_smi()
2826 mutex_init(&intf->cmd_rcvrs_mutex); in ipmi_register_smi()
2827 spin_lock_init(&intf->maintenance_mode_lock); in ipmi_register_smi()
2828 INIT_LIST_HEAD(&intf->cmd_rcvrs); in ipmi_register_smi()
2829 init_waitqueue_head(&intf->waitq); in ipmi_register_smi()
2831 atomic_set(&intf->stats[i], 0); in ipmi_register_smi()
2833 intf->proc_dir = NULL; in ipmi_register_smi()
2849 list_add_rcu(&intf->link, &ipmi_interfaces); in ipmi_register_smi()
2851 list_add_tail_rcu(&intf->link, link); in ipmi_register_smi()
2853 rv = handlers->start_processing(send_info, intf); in ipmi_register_smi()
2857 get_guid(intf); in ipmi_register_smi()
2859 if ((intf->ipmi_version_major > 1) in ipmi_register_smi()
2860 || ((intf->ipmi_version_major == 1) in ipmi_register_smi()
2861 && (intf->ipmi_version_minor >= 5))) { in ipmi_register_smi()
2866 intf->null_user_handler = channel_handler; in ipmi_register_smi()
2867 intf->curr_channel = 0; in ipmi_register_smi()
2868 rv = send_channel_info_cmd(intf, 0); in ipmi_register_smi()
2877 wait_event(intf->waitq, in ipmi_register_smi()
2878 intf->curr_channel >= IPMI_MAX_CHANNELS); in ipmi_register_smi()
2879 intf->null_user_handler = NULL; in ipmi_register_smi()
2882 intf->channels[0].medium = IPMI_CHANNEL_MEDIUM_IPMB; in ipmi_register_smi()
2883 intf->channels[0].protocol = IPMI_CHANNEL_PROTOCOL_IPMB; in ipmi_register_smi()
2884 intf->curr_channel = IPMI_MAX_CHANNELS; in ipmi_register_smi()
2888 rv = add_proc_entries(intf, i); in ipmi_register_smi()
2890 rv = ipmi_bmc_register(intf, i); in ipmi_register_smi()
2894 if (intf->proc_dir) in ipmi_register_smi()
2895 remove_proc_entries(intf); in ipmi_register_smi()
2896 intf->handlers = NULL; in ipmi_register_smi()
2897 list_del_rcu(&intf->link); in ipmi_register_smi()
2901 kref_put(&intf->refcount, intf_free); in ipmi_register_smi()
2909 intf->intf_num = i; in ipmi_register_smi()
2912 call_smi_watchers(i, intf->si_dev); in ipmi_register_smi()
2920 static void deliver_smi_err_response(ipmi_smi_t intf, in deliver_smi_err_response() argument
2929 handle_one_recv_msg(intf, msg); in deliver_smi_err_response()
2932 static void cleanup_smi_msgs(ipmi_smi_t intf) in cleanup_smi_msgs() argument
2942 list_splice_tail(&intf->hp_xmit_msgs, &tmplist); in cleanup_smi_msgs()
2943 list_splice_tail(&intf->xmit_msgs, &tmplist); in cleanup_smi_msgs()
2946 while (intf->curr_msg && !list_empty(&intf->waiting_rcv_msgs)) { in cleanup_smi_msgs()
2961 deliver_smi_err_response(intf, msg, IPMI_ERR_UNSPECIFIED); in cleanup_smi_msgs()
2965 ent = &(intf->seq_table[i]); in cleanup_smi_msgs()
2972 int ipmi_unregister_smi(ipmi_smi_t intf) in ipmi_unregister_smi() argument
2975 int intf_num = intf->intf_num; in ipmi_unregister_smi()
2978 ipmi_bmc_unregister(intf); in ipmi_unregister_smi()
2982 intf->intf_num = -1; in ipmi_unregister_smi()
2983 intf->in_shutdown = true; in ipmi_unregister_smi()
2984 list_del_rcu(&intf->link); in ipmi_unregister_smi()
2988 cleanup_smi_msgs(intf); in ipmi_unregister_smi()
2993 list_for_each_entry_rcu(user, &intf->users, link) { in ipmi_unregister_smi()
2994 module_put(intf->handlers->owner); in ipmi_unregister_smi()
2995 if (intf->handlers->dec_usecount) in ipmi_unregister_smi()
2996 intf->handlers->dec_usecount(intf->send_info); in ipmi_unregister_smi()
2999 intf->handlers = NULL; in ipmi_unregister_smi()
3002 remove_proc_entries(intf); in ipmi_unregister_smi()
3012 kref_put(&intf->refcount, intf_free); in ipmi_unregister_smi()
3017 static int handle_ipmb_get_msg_rsp(ipmi_smi_t intf, in handle_ipmb_get_msg_rsp() argument
3029 ipmi_inc_stat(intf, invalid_ipmb_responses); in handle_ipmb_get_msg_rsp()
3047 if (intf_find_seq(intf, in handle_ipmb_get_msg_rsp()
3058 ipmi_inc_stat(intf, unhandled_ipmb_responses); in handle_ipmb_get_msg_rsp()
3074 ipmi_inc_stat(intf, handled_ipmb_responses); in handle_ipmb_get_msg_rsp()
3080 static int handle_ipmb_get_msg_cmd(ipmi_smi_t intf, in handle_ipmb_get_msg_cmd() argument
3094 ipmi_inc_stat(intf, invalid_commands); in handle_ipmb_get_msg_cmd()
3108 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_ipmb_get_msg_cmd()
3118 ipmi_inc_stat(intf, unhandled_commands); in handle_ipmb_get_msg_cmd()
3126 msg->data[6] = intf->channels[msg->rsp[3] & 0xf].address; in handle_ipmb_get_msg_cmd()
3144 if (!intf->in_shutdown) { in handle_ipmb_get_msg_cmd()
3145 smi_send(intf, intf->handlers, msg, 0); in handle_ipmb_get_msg_cmd()
3156 ipmi_inc_stat(intf, handled_commands); in handle_ipmb_get_msg_cmd()
3201 static int handle_lan_get_msg_rsp(ipmi_smi_t intf, in handle_lan_get_msg_rsp() argument
3214 ipmi_inc_stat(intf, invalid_lan_responses); in handle_lan_get_msg_rsp()
3235 if (intf_find_seq(intf, in handle_lan_get_msg_rsp()
3246 ipmi_inc_stat(intf, unhandled_lan_responses); in handle_lan_get_msg_rsp()
3262 ipmi_inc_stat(intf, handled_lan_responses); in handle_lan_get_msg_rsp()
3268 static int handle_lan_get_msg_cmd(ipmi_smi_t intf, in handle_lan_get_msg_cmd() argument
3282 ipmi_inc_stat(intf, invalid_commands); in handle_lan_get_msg_cmd()
3296 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_lan_get_msg_cmd()
3306 ipmi_inc_stat(intf, unhandled_commands); in handle_lan_get_msg_cmd()
3315 ipmi_inc_stat(intf, handled_commands); in handle_lan_get_msg_cmd()
3368 static int handle_oem_get_msg_cmd(ipmi_smi_t intf, in handle_oem_get_msg_cmd() argument
3386 ipmi_inc_stat(intf, invalid_commands); in handle_oem_get_msg_cmd()
3404 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_oem_get_msg_cmd()
3414 ipmi_inc_stat(intf, unhandled_commands); in handle_oem_get_msg_cmd()
3424 ipmi_inc_stat(intf, handled_commands); in handle_oem_get_msg_cmd()
3488 static int handle_read_event_rsp(ipmi_smi_t intf, in handle_read_event_rsp() argument
3500 ipmi_inc_stat(intf, invalid_events); in handle_read_event_rsp()
3511 spin_lock_irqsave(&intf->events_lock, flags); in handle_read_event_rsp()
3513 ipmi_inc_stat(intf, events); in handle_read_event_rsp()
3520 list_for_each_entry_rcu(user, &intf->users, link) { in handle_read_event_rsp()
3556 } else if (intf->waiting_events_count < MAX_EVENTS_IN_QUEUE) { in handle_read_event_rsp()
3573 list_add_tail(&(recv_msg->link), &(intf->waiting_events)); in handle_read_event_rsp()
3574 intf->waiting_events_count++; in handle_read_event_rsp()
3575 } else if (!intf->event_msg_printed) { in handle_read_event_rsp()
3582 intf->event_msg_printed = 1; in handle_read_event_rsp()
3586 spin_unlock_irqrestore(&(intf->events_lock), flags); in handle_read_event_rsp()
3591 static int handle_bmc_rsp(ipmi_smi_t intf, in handle_bmc_rsp() argument
3611 ipmi_inc_stat(intf, unhandled_local_responses); in handle_bmc_rsp()
3616 ipmi_inc_stat(intf, handled_local_responses); in handle_bmc_rsp()
3642 static int handle_one_recv_msg(ipmi_smi_t intf, in handle_one_recv_msg() argument
3731 if (intf->curr_channel < IPMI_MAX_CHANNELS) { in handle_one_recv_msg()
3736 switch (intf->channels[chan].medium) { in handle_one_recv_msg()
3743 requeue = handle_ipmb_get_msg_rsp(intf, msg); in handle_one_recv_msg()
3749 requeue = handle_ipmb_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3760 requeue = handle_lan_get_msg_rsp(intf, msg); in handle_one_recv_msg()
3766 requeue = handle_lan_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3773 if ((intf->channels[chan].medium in handle_one_recv_msg()
3775 && (intf->channels[chan].medium in handle_one_recv_msg()
3777 requeue = handle_oem_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3790 requeue = handle_read_event_rsp(intf, msg); in handle_one_recv_msg()
3793 requeue = handle_bmc_rsp(intf, msg); in handle_one_recv_msg()
3803 static void handle_new_recv_msgs(ipmi_smi_t intf) in handle_new_recv_msgs() argument
3808 int run_to_completion = intf->run_to_completion; in handle_new_recv_msgs()
3812 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3813 while (!list_empty(&intf->waiting_rcv_msgs)) { in handle_new_recv_msgs()
3814 smi_msg = list_entry(intf->waiting_rcv_msgs.next, in handle_new_recv_msgs()
3817 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, in handle_new_recv_msgs()
3819 rv = handle_one_recv_msg(intf, smi_msg); in handle_new_recv_msgs()
3821 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3837 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3843 if (atomic_add_unless(&intf->watchdog_pretimeouts_to_deliver, -1, 0)) { in handle_new_recv_msgs()
3847 list_for_each_entry_rcu(user, &intf->users, link) { in handle_new_recv_msgs()
3859 ipmi_smi_t intf = (ipmi_smi_t) val; in smi_recv_tasklet() local
3860 int run_to_completion = intf->run_to_completion; in smi_recv_tasklet()
3871 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in smi_recv_tasklet()
3872 if (intf->curr_msg == NULL && !intf->in_shutdown) { in smi_recv_tasklet()
3876 if (!list_empty(&intf->hp_xmit_msgs)) in smi_recv_tasklet()
3877 entry = intf->hp_xmit_msgs.next; in smi_recv_tasklet()
3878 else if (!list_empty(&intf->xmit_msgs)) in smi_recv_tasklet()
3879 entry = intf->xmit_msgs.next; in smi_recv_tasklet()
3884 intf->curr_msg = newmsg; in smi_recv_tasklet()
3888 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in smi_recv_tasklet()
3890 intf->handlers->sender(intf->send_info, newmsg); in smi_recv_tasklet()
3892 handle_new_recv_msgs(intf); in smi_recv_tasklet()
3896 void ipmi_smi_msg_received(ipmi_smi_t intf, in ipmi_smi_msg_received() argument
3900 int run_to_completion = intf->run_to_completion; in ipmi_smi_msg_received()
3907 if (intf->in_shutdown) in ipmi_smi_msg_received()
3933 else if ((intf->channels[chan].medium in ipmi_smi_msg_received()
3935 || (intf->channels[chan].medium in ipmi_smi_msg_received()
3937 ipmi_inc_stat(intf, sent_lan_command_errs); in ipmi_smi_msg_received()
3939 ipmi_inc_stat(intf, sent_ipmb_command_errs); in ipmi_smi_msg_received()
3940 intf_err_seq(intf, msg->msgid, msg->rsp[2]); in ipmi_smi_msg_received()
3943 intf_start_seq_timer(intf, msg->msgid); in ipmi_smi_msg_received()
3953 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in ipmi_smi_msg_received()
3954 list_add_tail(&msg->link, &intf->waiting_rcv_msgs); in ipmi_smi_msg_received()
3956 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, in ipmi_smi_msg_received()
3961 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in ipmi_smi_msg_received()
3962 if (msg == intf->curr_msg) in ipmi_smi_msg_received()
3963 intf->curr_msg = NULL; in ipmi_smi_msg_received()
3965 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in ipmi_smi_msg_received()
3968 smi_recv_tasklet((unsigned long) intf); in ipmi_smi_msg_received()
3970 tasklet_schedule(&intf->recv_tasklet); in ipmi_smi_msg_received()
3974 void ipmi_smi_watchdog_pretimeout(ipmi_smi_t intf) in ipmi_smi_watchdog_pretimeout() argument
3976 if (intf->in_shutdown) in ipmi_smi_watchdog_pretimeout()
3979 atomic_set(&intf->watchdog_pretimeouts_to_deliver, 1); in ipmi_smi_watchdog_pretimeout()
3980 tasklet_schedule(&intf->recv_tasklet); in ipmi_smi_watchdog_pretimeout()
3985 smi_from_recv_msg(ipmi_smi_t intf, struct ipmi_recv_msg *recv_msg, in smi_from_recv_msg() argument
4012 static void check_msg_timeout(ipmi_smi_t intf, struct seq_table *ent, in check_msg_timeout() argument
4020 if (intf->in_shutdown) in check_msg_timeout()
4038 ipmi_inc_stat(intf, timed_out_ipmb_broadcasts); in check_msg_timeout()
4040 ipmi_inc_stat(intf, timed_out_lan_commands); in check_msg_timeout()
4042 ipmi_inc_stat(intf, timed_out_ipmb_commands); in check_msg_timeout()
4055 smi_msg = smi_from_recv_msg(intf, ent->recv_msg, slot, in check_msg_timeout()
4059 ipmi_inc_stat(intf, in check_msg_timeout()
4062 ipmi_inc_stat(intf, in check_msg_timeout()
4067 spin_unlock_irqrestore(&intf->seq_lock, *flags); in check_msg_timeout()
4076 handlers = intf->handlers; in check_msg_timeout()
4079 ipmi_inc_stat(intf, in check_msg_timeout()
4082 ipmi_inc_stat(intf, in check_msg_timeout()
4085 smi_send(intf, intf->handlers, smi_msg, 0); in check_msg_timeout()
4089 spin_lock_irqsave(&intf->seq_lock, *flags); in check_msg_timeout()
4093 static unsigned int ipmi_timeout_handler(ipmi_smi_t intf, long timeout_period) in ipmi_timeout_handler() argument
4107 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_timeout_handler()
4109 check_msg_timeout(intf, &(intf->seq_table[i]), in ipmi_timeout_handler()
4112 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_timeout_handler()
4125 if (intf->auto_maintenance_timeout > 0) { in ipmi_timeout_handler()
4126 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in ipmi_timeout_handler()
4127 if (intf->auto_maintenance_timeout > 0) { in ipmi_timeout_handler()
4128 intf->auto_maintenance_timeout in ipmi_timeout_handler()
4130 if (!intf->maintenance_mode in ipmi_timeout_handler()
4131 && (intf->auto_maintenance_timeout <= 0)) { in ipmi_timeout_handler()
4132 intf->maintenance_mode_enable = false; in ipmi_timeout_handler()
4133 maintenance_mode_update(intf); in ipmi_timeout_handler()
4136 spin_unlock_irqrestore(&intf->maintenance_mode_lock, in ipmi_timeout_handler()
4140 tasklet_schedule(&intf->recv_tasklet); in ipmi_timeout_handler()
4145 static void ipmi_request_event(ipmi_smi_t intf) in ipmi_request_event() argument
4148 if (intf->maintenance_mode_enable) in ipmi_request_event()
4151 if (!intf->in_shutdown) in ipmi_request_event()
4152 intf->handlers->request_events(intf->send_info); in ipmi_request_event()
4161 ipmi_smi_t intf; in ipmi_timeout() local
4168 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_timeout()
4171 if (atomic_read(&intf->event_waiters)) { in ipmi_timeout()
4172 intf->ticks_to_req_ev--; in ipmi_timeout()
4173 if (intf->ticks_to_req_ev == 0) { in ipmi_timeout()
4174 ipmi_request_event(intf); in ipmi_timeout()
4175 intf->ticks_to_req_ev = IPMI_REQUEST_EV_TIME; in ipmi_timeout()
4180 lnt += ipmi_timeout_handler(intf, IPMI_TIMEOUT_TIME); in ipmi_timeout()
4183 if (lnt != intf->last_needs_timer && in ipmi_timeout()
4184 intf->handlers->set_need_watch) in ipmi_timeout()
4185 intf->handlers->set_need_watch(intf->send_info, lnt); in ipmi_timeout()
4186 intf->last_needs_timer = lnt; in ipmi_timeout()
4196 static void need_waiter(ipmi_smi_t intf) in need_waiter() argument
4269 static void ipmi_panic_request_and_wait(ipmi_smi_t intf, in ipmi_panic_request_and_wait() argument
4281 intf, in ipmi_panic_request_and_wait()
4285 intf, in ipmi_panic_request_and_wait()
4289 intf->channels[0].address, in ipmi_panic_request_and_wait()
4290 intf->channels[0].lun, in ipmi_panic_request_and_wait()
4295 ipmi_poll(intf); in ipmi_panic_request_and_wait()
4299 static void event_receiver_fetcher(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in event_receiver_fetcher() argument
4306 intf->event_receiver = msg->msg.data[1]; in event_receiver_fetcher()
4307 intf->event_receiver_lun = msg->msg.data[2] & 0x3; in event_receiver_fetcher()
4311 static void device_id_fetcher(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in device_id_fetcher() argument
4321 intf->local_sel_device = (msg->msg.data[6] >> 2) & 1; in device_id_fetcher()
4322 intf->local_event_generator = (msg->msg.data[6] >> 5) & 1; in device_id_fetcher()
4330 ipmi_smi_t intf; in send_panic_events() local
4362 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in send_panic_events()
4363 if (!intf->handlers) in send_panic_events()
4367 intf->run_to_completion = 1; in send_panic_events()
4369 intf->handlers->set_run_to_completion(intf->send_info, 1); in send_panic_events()
4370 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4382 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in send_panic_events()
4387 if (intf->intf_num == -1) in send_panic_events()
4408 intf->local_sel_device = 0; in send_panic_events()
4409 intf->local_event_generator = 0; in send_panic_events()
4410 intf->event_receiver = 0; in send_panic_events()
4417 intf->null_user_handler = device_id_fetcher; in send_panic_events()
4418 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4420 if (intf->local_event_generator) { in send_panic_events()
4426 intf->null_user_handler = event_receiver_fetcher; in send_panic_events()
4427 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4429 intf->null_user_handler = NULL; in send_panic_events()
4436 if (((intf->event_receiver & 1) == 0) in send_panic_events()
4437 && (intf->event_receiver != 0) in send_panic_events()
4438 && (intf->event_receiver != intf->channels[0].address)) { in send_panic_events()
4446 ipmb->lun = intf->event_receiver_lun; in send_panic_events()
4447 ipmb->slave_addr = intf->event_receiver; in send_panic_events()
4448 } else if (intf->local_sel_device) { in send_panic_events()
4475 data[3] = intf->channels[0].address; in send_panic_events()
4484 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4497 ipmi_smi_t intf; in panic_event() local
4504 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in panic_event()
4505 if (!intf->handlers) in panic_event()
4509 intf->run_to_completion = 1; in panic_event()
4510 intf->handlers->set_run_to_completion(intf->send_info, 1); in panic_event()