Lines Matching refs:intf
57 static void handle_new_recv_msgs(ipmi_smi_t intf);
58 static void need_waiter(ipmi_smi_t intf);
59 static int handle_one_recv_msg(ipmi_smi_t intf,
109 ipmi_smi_t intf; member
424 void (*null_user_handler)(ipmi_smi_t intf, struct ipmi_recv_msg *msg);
470 #define ipmi_inc_stat(intf, stat) \ argument
471 atomic_inc(&(intf)->stats[IPMI_STAT_ ## stat])
472 #define ipmi_get_stat(intf, stat) \ argument
473 ((unsigned int) atomic_read(&(intf)->stats[IPMI_STAT_ ## stat]))
522 static void clean_up_interface_data(ipmi_smi_t intf) in clean_up_interface_data() argument
528 tasklet_kill(&intf->recv_tasklet); in clean_up_interface_data()
530 free_smi_msg_list(&intf->waiting_rcv_msgs); in clean_up_interface_data()
531 free_recv_msg_list(&intf->waiting_events); in clean_up_interface_data()
537 mutex_lock(&intf->cmd_rcvrs_mutex); in clean_up_interface_data()
539 list_splice_init_rcu(&intf->cmd_rcvrs, &list, synchronize_rcu); in clean_up_interface_data()
540 mutex_unlock(&intf->cmd_rcvrs_mutex); in clean_up_interface_data()
546 if ((intf->seq_table[i].inuse) in clean_up_interface_data()
547 && (intf->seq_table[i].recv_msg)) in clean_up_interface_data()
548 ipmi_free_recv_msg(intf->seq_table[i].recv_msg); in clean_up_interface_data()
554 ipmi_smi_t intf = container_of(ref, struct ipmi_smi, refcount); in intf_free() local
556 clean_up_interface_data(intf); in intf_free()
557 kfree(intf); in intf_free()
562 ipmi_smi_t intf; member
568 ipmi_smi_t intf; in ipmi_smi_watcher_register() local
577 list_for_each_entry(intf, &ipmi_interfaces, link) { in ipmi_smi_watcher_register()
578 if (intf->intf_num == -1) in ipmi_smi_watcher_register()
583 kref_get(&intf->refcount); in ipmi_smi_watcher_register()
584 e->intf = intf; in ipmi_smi_watcher_register()
585 e->intf_num = intf->intf_num; in ipmi_smi_watcher_register()
596 watcher->new_smi(e->intf_num, e->intf->si_dev); in ipmi_smi_watcher_register()
597 kref_put(&e->intf->refcount, intf_free); in ipmi_smi_watcher_register()
610 kref_put(&e->intf->refcount, intf_free); in ipmi_smi_watcher_register()
736 ipmi_smi_t intf = msg->user_msg_data; in deliver_response() local
739 if (intf->null_user_handler) { in deliver_response()
740 intf->null_user_handler(intf, msg); in deliver_response()
741 ipmi_inc_stat(intf, handled_local_responses); in deliver_response()
744 ipmi_inc_stat(intf, unhandled_local_responses); in deliver_response()
775 static int intf_next_seq(ipmi_smi_t intf, in intf_next_seq() argument
786 for (i = intf->curr_seq; (i+1)%IPMI_IPMB_NUM_SEQ != intf->curr_seq; in intf_next_seq()
788 if (!intf->seq_table[i].inuse) in intf_next_seq()
792 if (!intf->seq_table[i].inuse) { in intf_next_seq()
793 intf->seq_table[i].recv_msg = recv_msg; in intf_next_seq()
799 intf->seq_table[i].timeout = MAX_MSG_TIMEOUT; in intf_next_seq()
800 intf->seq_table[i].orig_timeout = timeout; in intf_next_seq()
801 intf->seq_table[i].retries_left = retries; in intf_next_seq()
802 intf->seq_table[i].broadcast = broadcast; in intf_next_seq()
803 intf->seq_table[i].inuse = 1; in intf_next_seq()
804 intf->seq_table[i].seqid = NEXT_SEQID(intf->seq_table[i].seqid); in intf_next_seq()
806 *seqid = intf->seq_table[i].seqid; in intf_next_seq()
807 intf->curr_seq = (i+1)%IPMI_IPMB_NUM_SEQ; in intf_next_seq()
808 need_waiter(intf); in intf_next_seq()
823 static int intf_find_seq(ipmi_smi_t intf, in intf_find_seq() argument
837 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_find_seq()
838 if (intf->seq_table[seq].inuse) { in intf_find_seq()
839 struct ipmi_recv_msg *msg = intf->seq_table[seq].recv_msg; in intf_find_seq()
845 intf->seq_table[seq].inuse = 0; in intf_find_seq()
849 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_find_seq()
856 static int intf_start_seq_timer(ipmi_smi_t intf, in intf_start_seq_timer() argument
867 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_start_seq_timer()
872 if ((intf->seq_table[seq].inuse) in intf_start_seq_timer()
873 && (intf->seq_table[seq].seqid == seqid)) { in intf_start_seq_timer()
874 struct seq_table *ent = &(intf->seq_table[seq]); in intf_start_seq_timer()
878 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_start_seq_timer()
884 static int intf_err_seq(ipmi_smi_t intf, in intf_err_seq() argument
897 spin_lock_irqsave(&(intf->seq_lock), flags); in intf_err_seq()
902 if ((intf->seq_table[seq].inuse) in intf_err_seq()
903 && (intf->seq_table[seq].seqid == seqid)) { in intf_err_seq()
904 struct seq_table *ent = &(intf->seq_table[seq]); in intf_err_seq()
910 spin_unlock_irqrestore(&(intf->seq_lock), flags); in intf_err_seq()
927 ipmi_smi_t intf; in ipmi_create_user() local
962 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_create_user()
963 if (intf->intf_num == if_num) in ipmi_create_user()
972 kref_get(&intf->refcount); in ipmi_create_user()
977 new_user->intf = intf; in ipmi_create_user()
980 if (!try_module_get(intf->handlers->owner)) { in ipmi_create_user()
985 if (intf->handlers->inc_usecount) { in ipmi_create_user()
986 rv = intf->handlers->inc_usecount(intf->send_info); in ipmi_create_user()
988 module_put(intf->handlers->owner); in ipmi_create_user()
1000 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_create_user()
1001 list_add_rcu(&new_user->link, &intf->users); in ipmi_create_user()
1002 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_create_user()
1005 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_create_user()
1006 need_waiter(intf); in ipmi_create_user()
1012 kref_put(&intf->refcount, intf_free); in ipmi_create_user()
1023 ipmi_smi_t intf; in ipmi_get_smi_info() local
1027 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_get_smi_info()
1028 if (intf->intf_num == if_num) in ipmi_get_smi_info()
1037 handlers = intf->handlers; in ipmi_get_smi_info()
1040 rv = handlers->get_smi_info(intf->send_info, data); in ipmi_get_smi_info()
1055 ipmi_smi_t intf = user->intf; in ipmi_destroy_user() local
1064 atomic_dec(&intf->event_waiters); in ipmi_destroy_user()
1067 atomic_dec(&intf->event_waiters); in ipmi_destroy_user()
1070 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_destroy_user()
1074 if (intf->seq_table[i].inuse in ipmi_destroy_user()
1075 && (intf->seq_table[i].recv_msg->user == user)) { in ipmi_destroy_user()
1076 intf->seq_table[i].inuse = 0; in ipmi_destroy_user()
1077 ipmi_free_recv_msg(intf->seq_table[i].recv_msg); in ipmi_destroy_user()
1080 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_destroy_user()
1088 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_destroy_user()
1089 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in ipmi_destroy_user()
1096 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_destroy_user()
1105 if (intf->handlers) { in ipmi_destroy_user()
1106 module_put(intf->handlers->owner); in ipmi_destroy_user()
1107 if (intf->handlers->dec_usecount) in ipmi_destroy_user()
1108 intf->handlers->dec_usecount(intf->send_info); in ipmi_destroy_user()
1112 kref_put(&intf->refcount, intf_free); in ipmi_destroy_user()
1124 *major = user->intf->ipmi_version_major; in ipmi_get_version()
1125 *minor = user->intf->ipmi_version_minor; in ipmi_get_version()
1135 user->intf->channels[channel].address = address; in ipmi_set_my_address()
1146 *address = user->intf->channels[channel].address; in ipmi_get_my_address()
1157 user->intf->channels[channel].lun = LUN & 0x3; in ipmi_set_my_LUN()
1168 *address = user->intf->channels[channel].lun; in ipmi_get_my_LUN()
1178 spin_lock_irqsave(&user->intf->maintenance_mode_lock, flags); in ipmi_get_maintenance_mode()
1179 mode = user->intf->maintenance_mode; in ipmi_get_maintenance_mode()
1180 spin_unlock_irqrestore(&user->intf->maintenance_mode_lock, flags); in ipmi_get_maintenance_mode()
1186 static void maintenance_mode_update(ipmi_smi_t intf) in maintenance_mode_update() argument
1188 if (intf->handlers->set_maintenance_mode) in maintenance_mode_update()
1189 intf->handlers->set_maintenance_mode( in maintenance_mode_update()
1190 intf->send_info, intf->maintenance_mode_enable); in maintenance_mode_update()
1197 ipmi_smi_t intf = user->intf; in ipmi_set_maintenance_mode() local
1199 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in ipmi_set_maintenance_mode()
1200 if (intf->maintenance_mode != mode) { in ipmi_set_maintenance_mode()
1203 intf->maintenance_mode_enable in ipmi_set_maintenance_mode()
1204 = (intf->auto_maintenance_timeout > 0); in ipmi_set_maintenance_mode()
1208 intf->maintenance_mode_enable = false; in ipmi_set_maintenance_mode()
1212 intf->maintenance_mode_enable = true; in ipmi_set_maintenance_mode()
1219 intf->maintenance_mode = mode; in ipmi_set_maintenance_mode()
1221 maintenance_mode_update(intf); in ipmi_set_maintenance_mode()
1224 spin_unlock_irqrestore(&intf->maintenance_mode_lock, flags); in ipmi_set_maintenance_mode()
1233 ipmi_smi_t intf = user->intf; in ipmi_set_gets_events() local
1239 spin_lock_irqsave(&intf->events_lock, flags); in ipmi_set_gets_events()
1246 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_set_gets_events()
1247 need_waiter(intf); in ipmi_set_gets_events()
1249 atomic_dec(&intf->event_waiters); in ipmi_set_gets_events()
1252 if (intf->delivering_events) in ipmi_set_gets_events()
1260 while (user->gets_events && !list_empty(&intf->waiting_events)) { in ipmi_set_gets_events()
1261 list_for_each_entry_safe(msg, msg2, &intf->waiting_events, link) in ipmi_set_gets_events()
1263 intf->waiting_events_count = 0; in ipmi_set_gets_events()
1264 if (intf->event_msg_printed) { in ipmi_set_gets_events()
1267 intf->event_msg_printed = 0; in ipmi_set_gets_events()
1270 intf->delivering_events = 1; in ipmi_set_gets_events()
1271 spin_unlock_irqrestore(&intf->events_lock, flags); in ipmi_set_gets_events()
1279 spin_lock_irqsave(&intf->events_lock, flags); in ipmi_set_gets_events()
1280 intf->delivering_events = 0; in ipmi_set_gets_events()
1284 spin_unlock_irqrestore(&intf->events_lock, flags); in ipmi_set_gets_events()
1290 static struct cmd_rcvr *find_cmd_rcvr(ipmi_smi_t intf, in find_cmd_rcvr() argument
1297 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in find_cmd_rcvr()
1305 static int is_cmd_rcvr_exclusive(ipmi_smi_t intf, in is_cmd_rcvr_exclusive() argument
1312 list_for_each_entry_rcu(rcvr, &intf->cmd_rcvrs, link) { in is_cmd_rcvr_exclusive()
1325 ipmi_smi_t intf = user->intf; in ipmi_register_for_cmd() local
1338 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_register_for_cmd()
1340 if (!is_cmd_rcvr_exclusive(intf, netfn, cmd, chans)) { in ipmi_register_for_cmd()
1345 if (atomic_inc_return(&intf->event_waiters) == 1) in ipmi_register_for_cmd()
1346 need_waiter(intf); in ipmi_register_for_cmd()
1348 list_add_rcu(&rcvr->link, &intf->cmd_rcvrs); in ipmi_register_for_cmd()
1351 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_register_for_cmd()
1364 ipmi_smi_t intf = user->intf; in ipmi_unregister_for_cmd() local
1369 mutex_lock(&intf->cmd_rcvrs_mutex); in ipmi_unregister_for_cmd()
1373 rcvr = find_cmd_rcvr(intf, netfn, cmd, i); in ipmi_unregister_for_cmd()
1386 mutex_unlock(&intf->cmd_rcvrs_mutex); in ipmi_unregister_for_cmd()
1389 atomic_dec(&intf->event_waiters); in ipmi_unregister_for_cmd()
1492 static struct ipmi_smi_msg *smi_add_send_msg(ipmi_smi_t intf, in smi_add_send_msg() argument
1496 if (intf->curr_msg) { in smi_add_send_msg()
1498 list_add_tail(&smi_msg->link, &intf->hp_xmit_msgs); in smi_add_send_msg()
1500 list_add_tail(&smi_msg->link, &intf->xmit_msgs); in smi_add_send_msg()
1503 intf->curr_msg = smi_msg; in smi_add_send_msg()
1510 static void smi_send(ipmi_smi_t intf, const struct ipmi_smi_handlers *handlers, in smi_send() argument
1513 int run_to_completion = intf->run_to_completion; in smi_send()
1516 smi_msg = smi_add_send_msg(intf, smi_msg, priority); in smi_send()
1520 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in smi_send()
1521 smi_msg = smi_add_send_msg(intf, smi_msg, priority); in smi_send()
1522 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in smi_send()
1526 handlers->sender(intf->send_info, smi_msg); in smi_send()
1536 ipmi_smi_t intf, in i_ipmi_request() argument
1575 if (intf->in_shutdown) { in i_ipmi_request()
1601 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1616 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1625 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in i_ipmi_request()
1626 intf->auto_maintenance_timeout in i_ipmi_request()
1628 if (!intf->maintenance_mode in i_ipmi_request()
1629 && !intf->maintenance_mode_enable) { in i_ipmi_request()
1630 intf->maintenance_mode_enable = true; in i_ipmi_request()
1631 maintenance_mode_update(intf); in i_ipmi_request()
1633 spin_unlock_irqrestore(&intf->maintenance_mode_lock, in i_ipmi_request()
1638 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1650 ipmi_inc_stat(intf, sent_local_commands); in i_ipmi_request()
1658 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1663 if (intf->channels[addr->channel].medium in i_ipmi_request()
1665 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1696 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1703 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1715 ipmi_inc_stat(intf, sent_ipmb_responses); in i_ipmi_request()
1728 spin_lock_irqsave(&(intf->seq_lock), flags); in i_ipmi_request()
1734 rv = intf_next_seq(intf, in i_ipmi_request()
1746 spin_unlock_irqrestore(&(intf->seq_lock), in i_ipmi_request()
1751 ipmi_inc_stat(intf, sent_ipmb_commands); in i_ipmi_request()
1780 spin_unlock_irqrestore(&(intf->seq_lock), flags); in i_ipmi_request()
1788 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1793 if ((intf->channels[addr->channel].medium in i_ipmi_request()
1795 && (intf->channels[addr->channel].medium in i_ipmi_request()
1797 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1810 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1817 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1829 ipmi_inc_stat(intf, sent_lan_responses); in i_ipmi_request()
1841 spin_lock_irqsave(&(intf->seq_lock), flags); in i_ipmi_request()
1847 rv = intf_next_seq(intf, in i_ipmi_request()
1859 spin_unlock_irqrestore(&(intf->seq_lock), in i_ipmi_request()
1864 ipmi_inc_stat(intf, sent_lan_commands); in i_ipmi_request()
1892 spin_unlock_irqrestore(&(intf->seq_lock), flags); in i_ipmi_request()
1896 ipmi_inc_stat(intf, sent_invalid_commands); in i_ipmi_request()
1910 smi_send(intf, intf->handlers, smi_msg, priority); in i_ipmi_request()
1922 static int check_addr(ipmi_smi_t intf, in check_addr() argument
1929 *lun = intf->channels[addr->channel].lun; in check_addr()
1930 *saddr = intf->channels[addr->channel].address; in check_addr()
1948 rv = check_addr(user->intf, addr, &saddr, &lun); in ipmi_request_settime()
1952 user->intf, in ipmi_request_settime()
1980 rv = check_addr(user->intf, addr, &saddr, &lun); in ipmi_request_supply_msgs()
1984 user->intf, in ipmi_request_supply_msgs()
2001 ipmi_smi_t intf = m->private; in smi_ipmb_proc_show() local
2004 seq_printf(m, "%x", intf->channels[0].address); in smi_ipmb_proc_show()
2006 seq_printf(m, " %x", intf->channels[i].address); in smi_ipmb_proc_show()
2026 ipmi_smi_t intf = m->private; in smi_version_proc_show() local
2029 ipmi_version_major(&intf->bmc->id), in smi_version_proc_show()
2030 ipmi_version_minor(&intf->bmc->id)); in smi_version_proc_show()
2049 ipmi_smi_t intf = m->private; in smi_stats_proc_show() local
2052 ipmi_get_stat(intf, sent_invalid_commands)); in smi_stats_proc_show()
2054 ipmi_get_stat(intf, sent_local_commands)); in smi_stats_proc_show()
2056 ipmi_get_stat(intf, handled_local_responses)); in smi_stats_proc_show()
2058 ipmi_get_stat(intf, unhandled_local_responses)); in smi_stats_proc_show()
2060 ipmi_get_stat(intf, sent_ipmb_commands)); in smi_stats_proc_show()
2062 ipmi_get_stat(intf, sent_ipmb_command_errs)); in smi_stats_proc_show()
2064 ipmi_get_stat(intf, retransmitted_ipmb_commands)); in smi_stats_proc_show()
2066 ipmi_get_stat(intf, timed_out_ipmb_commands)); in smi_stats_proc_show()
2068 ipmi_get_stat(intf, timed_out_ipmb_broadcasts)); in smi_stats_proc_show()
2070 ipmi_get_stat(intf, sent_ipmb_responses)); in smi_stats_proc_show()
2072 ipmi_get_stat(intf, handled_ipmb_responses)); in smi_stats_proc_show()
2074 ipmi_get_stat(intf, invalid_ipmb_responses)); in smi_stats_proc_show()
2076 ipmi_get_stat(intf, unhandled_ipmb_responses)); in smi_stats_proc_show()
2078 ipmi_get_stat(intf, sent_lan_commands)); in smi_stats_proc_show()
2080 ipmi_get_stat(intf, sent_lan_command_errs)); in smi_stats_proc_show()
2082 ipmi_get_stat(intf, retransmitted_lan_commands)); in smi_stats_proc_show()
2084 ipmi_get_stat(intf, timed_out_lan_commands)); in smi_stats_proc_show()
2086 ipmi_get_stat(intf, sent_lan_responses)); in smi_stats_proc_show()
2088 ipmi_get_stat(intf, handled_lan_responses)); in smi_stats_proc_show()
2090 ipmi_get_stat(intf, invalid_lan_responses)); in smi_stats_proc_show()
2092 ipmi_get_stat(intf, unhandled_lan_responses)); in smi_stats_proc_show()
2094 ipmi_get_stat(intf, handled_commands)); in smi_stats_proc_show()
2096 ipmi_get_stat(intf, invalid_commands)); in smi_stats_proc_show()
2098 ipmi_get_stat(intf, unhandled_commands)); in smi_stats_proc_show()
2100 ipmi_get_stat(intf, invalid_events)); in smi_stats_proc_show()
2102 ipmi_get_stat(intf, events)); in smi_stats_proc_show()
2104 ipmi_get_stat(intf, dropped_rexmit_lan_commands)); in smi_stats_proc_show()
2106 ipmi_get_stat(intf, dropped_rexmit_ipmb_commands)); in smi_stats_proc_show()
2426 static void ipmi_bmc_unregister(ipmi_smi_t intf) in ipmi_bmc_unregister() argument
2428 struct bmc_device *bmc = intf->bmc; in ipmi_bmc_unregister()
2430 sysfs_remove_link(&intf->si_dev->kobj, "bmc"); in ipmi_bmc_unregister()
2431 if (intf->my_dev_name) { in ipmi_bmc_unregister()
2432 sysfs_remove_link(&bmc->pdev.dev.kobj, intf->my_dev_name); in ipmi_bmc_unregister()
2433 kfree(intf->my_dev_name); in ipmi_bmc_unregister()
2434 intf->my_dev_name = NULL; in ipmi_bmc_unregister()
2439 intf->bmc = NULL; in ipmi_bmc_unregister()
2443 static int ipmi_bmc_register(ipmi_smi_t intf, int ifnum) in ipmi_bmc_register() argument
2446 struct bmc_device *bmc = intf->bmc; in ipmi_bmc_register()
2468 intf->bmc = old_bmc; in ipmi_bmc_register()
2531 dev_info(intf->si_dev, "Found new BMC (man_id: 0x%6.6x, " in ipmi_bmc_register()
2542 rv = sysfs_create_link(&intf->si_dev->kobj, &bmc->pdev.dev.kobj, "bmc"); in ipmi_bmc_register()
2550 intf->my_dev_name = kasprintf(GFP_KERNEL, "ipmi%d", ifnum); in ipmi_bmc_register()
2551 if (!intf->my_dev_name) { in ipmi_bmc_register()
2559 rv = sysfs_create_link(&bmc->pdev.dev.kobj, &intf->si_dev->kobj, in ipmi_bmc_register()
2560 intf->my_dev_name); in ipmi_bmc_register()
2562 kfree(intf->my_dev_name); in ipmi_bmc_register()
2563 intf->my_dev_name = NULL; in ipmi_bmc_register()
2574 ipmi_bmc_unregister(intf); in ipmi_bmc_register()
2579 send_guid_cmd(ipmi_smi_t intf, int chan) in send_guid_cmd() argument
2593 intf, in send_guid_cmd()
2597 intf, in send_guid_cmd()
2601 intf->channels[0].address, in send_guid_cmd()
2602 intf->channels[0].lun, in send_guid_cmd()
2607 guid_handler(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in guid_handler() argument
2617 intf->bmc->guid_set = 0; in guid_handler()
2622 intf->bmc->guid_set = 0; in guid_handler()
2631 memcpy(intf->bmc->guid, msg->msg.data, 16); in guid_handler()
2632 intf->bmc->guid_set = 1; in guid_handler()
2634 wake_up(&intf->waitq); in guid_handler()
2638 get_guid(ipmi_smi_t intf) in get_guid() argument
2642 intf->bmc->guid_set = 0x2; in get_guid()
2643 intf->null_user_handler = guid_handler; in get_guid()
2644 rv = send_guid_cmd(intf, 0); in get_guid()
2647 intf->bmc->guid_set = 0; in get_guid()
2648 wait_event(intf->waitq, intf->bmc->guid_set != 2); in get_guid()
2649 intf->null_user_handler = NULL; in get_guid()
2653 send_channel_info_cmd(ipmi_smi_t intf, int chan) in send_channel_info_cmd() argument
2669 intf, in send_channel_info_cmd()
2673 intf, in send_channel_info_cmd()
2677 intf->channels[0].address, in send_channel_info_cmd()
2678 intf->channels[0].lun, in send_channel_info_cmd()
2683 channel_handler(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in channel_handler() argument
2702 intf->channels[0].medium in channel_handler()
2704 intf->channels[0].protocol in channel_handler()
2707 intf->curr_channel = IPMI_MAX_CHANNELS; in channel_handler()
2708 wake_up(&intf->waitq); in channel_handler()
2717 chan = intf->curr_channel; in channel_handler()
2718 intf->channels[chan].medium = msg->msg.data[2] & 0x7f; in channel_handler()
2719 intf->channels[chan].protocol = msg->msg.data[3] & 0x1f; in channel_handler()
2722 intf->curr_channel++; in channel_handler()
2723 if (intf->curr_channel >= IPMI_MAX_CHANNELS) in channel_handler()
2724 wake_up(&intf->waitq); in channel_handler()
2726 rv = send_channel_info_cmd(intf, intf->curr_channel); in channel_handler()
2732 " %d: %d\n", intf->curr_channel, rv); in channel_handler()
2734 intf->curr_channel = IPMI_MAX_CHANNELS; in channel_handler()
2735 wake_up(&intf->waitq); in channel_handler()
2742 static void ipmi_poll(ipmi_smi_t intf) in ipmi_poll() argument
2744 if (intf->handlers->poll) in ipmi_poll()
2745 intf->handlers->poll(intf->send_info); in ipmi_poll()
2747 handle_new_recv_msgs(intf); in ipmi_poll()
2752 ipmi_poll(user->intf); in ipmi_poll_interface()
2764 ipmi_smi_t intf; in ipmi_register_smi() local
2784 intf = kzalloc(sizeof(*intf), GFP_KERNEL); in ipmi_register_smi()
2785 if (!intf) in ipmi_register_smi()
2788 intf->ipmi_version_major = ipmi_version_major(device_id); in ipmi_register_smi()
2789 intf->ipmi_version_minor = ipmi_version_minor(device_id); in ipmi_register_smi()
2791 intf->bmc = kzalloc(sizeof(*intf->bmc), GFP_KERNEL); in ipmi_register_smi()
2792 if (!intf->bmc) { in ipmi_register_smi()
2793 kfree(intf); in ipmi_register_smi()
2796 intf->intf_num = -1; /* Mark it invalid for now. */ in ipmi_register_smi()
2797 kref_init(&intf->refcount); in ipmi_register_smi()
2798 intf->bmc->id = *device_id; in ipmi_register_smi()
2799 intf->si_dev = si_dev; in ipmi_register_smi()
2801 intf->channels[j].address = IPMI_BMC_SLAVE_ADDR; in ipmi_register_smi()
2802 intf->channels[j].lun = 2; in ipmi_register_smi()
2805 intf->channels[0].address = slave_addr; in ipmi_register_smi()
2806 INIT_LIST_HEAD(&intf->users); in ipmi_register_smi()
2807 intf->handlers = handlers; in ipmi_register_smi()
2808 intf->send_info = send_info; in ipmi_register_smi()
2809 spin_lock_init(&intf->seq_lock); in ipmi_register_smi()
2811 intf->seq_table[j].inuse = 0; in ipmi_register_smi()
2812 intf->seq_table[j].seqid = 0; in ipmi_register_smi()
2814 intf->curr_seq = 0; in ipmi_register_smi()
2816 mutex_init(&intf->proc_entry_lock); in ipmi_register_smi()
2818 spin_lock_init(&intf->waiting_rcv_msgs_lock); in ipmi_register_smi()
2819 INIT_LIST_HEAD(&intf->waiting_rcv_msgs); in ipmi_register_smi()
2820 tasklet_init(&intf->recv_tasklet, in ipmi_register_smi()
2822 (unsigned long) intf); in ipmi_register_smi()
2823 atomic_set(&intf->watchdog_pretimeouts_to_deliver, 0); in ipmi_register_smi()
2824 spin_lock_init(&intf->xmit_msgs_lock); in ipmi_register_smi()
2825 INIT_LIST_HEAD(&intf->xmit_msgs); in ipmi_register_smi()
2826 INIT_LIST_HEAD(&intf->hp_xmit_msgs); in ipmi_register_smi()
2827 spin_lock_init(&intf->events_lock); in ipmi_register_smi()
2828 atomic_set(&intf->event_waiters, 0); in ipmi_register_smi()
2829 intf->ticks_to_req_ev = IPMI_REQUEST_EV_TIME; in ipmi_register_smi()
2830 INIT_LIST_HEAD(&intf->waiting_events); in ipmi_register_smi()
2831 intf->waiting_events_count = 0; in ipmi_register_smi()
2832 mutex_init(&intf->cmd_rcvrs_mutex); in ipmi_register_smi()
2833 spin_lock_init(&intf->maintenance_mode_lock); in ipmi_register_smi()
2834 INIT_LIST_HEAD(&intf->cmd_rcvrs); in ipmi_register_smi()
2835 init_waitqueue_head(&intf->waitq); in ipmi_register_smi()
2837 atomic_set(&intf->stats[i], 0); in ipmi_register_smi()
2839 intf->proc_dir = NULL; in ipmi_register_smi()
2855 list_add_rcu(&intf->link, &ipmi_interfaces); in ipmi_register_smi()
2857 list_add_tail_rcu(&intf->link, link); in ipmi_register_smi()
2859 rv = handlers->start_processing(send_info, intf); in ipmi_register_smi()
2863 get_guid(intf); in ipmi_register_smi()
2865 if ((intf->ipmi_version_major > 1) in ipmi_register_smi()
2866 || ((intf->ipmi_version_major == 1) in ipmi_register_smi()
2867 && (intf->ipmi_version_minor >= 5))) { in ipmi_register_smi()
2872 intf->null_user_handler = channel_handler; in ipmi_register_smi()
2873 intf->curr_channel = 0; in ipmi_register_smi()
2874 rv = send_channel_info_cmd(intf, 0); in ipmi_register_smi()
2883 wait_event(intf->waitq, in ipmi_register_smi()
2884 intf->curr_channel >= IPMI_MAX_CHANNELS); in ipmi_register_smi()
2885 intf->null_user_handler = NULL; in ipmi_register_smi()
2888 intf->channels[0].medium = IPMI_CHANNEL_MEDIUM_IPMB; in ipmi_register_smi()
2889 intf->channels[0].protocol = IPMI_CHANNEL_PROTOCOL_IPMB; in ipmi_register_smi()
2890 intf->curr_channel = IPMI_MAX_CHANNELS; in ipmi_register_smi()
2894 rv = add_proc_entries(intf, i); in ipmi_register_smi()
2896 rv = ipmi_bmc_register(intf, i); in ipmi_register_smi()
2900 if (intf->proc_dir) in ipmi_register_smi()
2901 remove_proc_entries(intf); in ipmi_register_smi()
2902 intf->handlers = NULL; in ipmi_register_smi()
2903 list_del_rcu(&intf->link); in ipmi_register_smi()
2907 kref_put(&intf->refcount, intf_free); in ipmi_register_smi()
2915 intf->intf_num = i; in ipmi_register_smi()
2918 call_smi_watchers(i, intf->si_dev); in ipmi_register_smi()
2926 static void deliver_smi_err_response(ipmi_smi_t intf, in deliver_smi_err_response() argument
2935 handle_one_recv_msg(intf, msg); in deliver_smi_err_response()
2938 static void cleanup_smi_msgs(ipmi_smi_t intf) in cleanup_smi_msgs() argument
2948 list_splice_tail(&intf->hp_xmit_msgs, &tmplist); in cleanup_smi_msgs()
2949 list_splice_tail(&intf->xmit_msgs, &tmplist); in cleanup_smi_msgs()
2952 while (intf->curr_msg && !list_empty(&intf->waiting_rcv_msgs)) { in cleanup_smi_msgs()
2967 deliver_smi_err_response(intf, msg, IPMI_ERR_UNSPECIFIED); in cleanup_smi_msgs()
2971 ent = &(intf->seq_table[i]); in cleanup_smi_msgs()
2978 int ipmi_unregister_smi(ipmi_smi_t intf) in ipmi_unregister_smi() argument
2981 int intf_num = intf->intf_num; in ipmi_unregister_smi()
2984 ipmi_bmc_unregister(intf); in ipmi_unregister_smi()
2988 intf->intf_num = -1; in ipmi_unregister_smi()
2989 intf->in_shutdown = true; in ipmi_unregister_smi()
2990 list_del_rcu(&intf->link); in ipmi_unregister_smi()
2994 cleanup_smi_msgs(intf); in ipmi_unregister_smi()
2999 list_for_each_entry_rcu(user, &intf->users, link) { in ipmi_unregister_smi()
3000 module_put(intf->handlers->owner); in ipmi_unregister_smi()
3001 if (intf->handlers->dec_usecount) in ipmi_unregister_smi()
3002 intf->handlers->dec_usecount(intf->send_info); in ipmi_unregister_smi()
3005 intf->handlers = NULL; in ipmi_unregister_smi()
3008 remove_proc_entries(intf); in ipmi_unregister_smi()
3018 kref_put(&intf->refcount, intf_free); in ipmi_unregister_smi()
3023 static int handle_ipmb_get_msg_rsp(ipmi_smi_t intf, in handle_ipmb_get_msg_rsp() argument
3035 ipmi_inc_stat(intf, invalid_ipmb_responses); in handle_ipmb_get_msg_rsp()
3053 if (intf_find_seq(intf, in handle_ipmb_get_msg_rsp()
3064 ipmi_inc_stat(intf, unhandled_ipmb_responses); in handle_ipmb_get_msg_rsp()
3080 ipmi_inc_stat(intf, handled_ipmb_responses); in handle_ipmb_get_msg_rsp()
3086 static int handle_ipmb_get_msg_cmd(ipmi_smi_t intf, in handle_ipmb_get_msg_cmd() argument
3100 ipmi_inc_stat(intf, invalid_commands); in handle_ipmb_get_msg_cmd()
3114 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_ipmb_get_msg_cmd()
3124 ipmi_inc_stat(intf, unhandled_commands); in handle_ipmb_get_msg_cmd()
3132 msg->data[6] = intf->channels[msg->rsp[3] & 0xf].address; in handle_ipmb_get_msg_cmd()
3150 if (!intf->in_shutdown) { in handle_ipmb_get_msg_cmd()
3151 smi_send(intf, intf->handlers, msg, 0); in handle_ipmb_get_msg_cmd()
3162 ipmi_inc_stat(intf, handled_commands); in handle_ipmb_get_msg_cmd()
3207 static int handle_lan_get_msg_rsp(ipmi_smi_t intf, in handle_lan_get_msg_rsp() argument
3220 ipmi_inc_stat(intf, invalid_lan_responses); in handle_lan_get_msg_rsp()
3241 if (intf_find_seq(intf, in handle_lan_get_msg_rsp()
3252 ipmi_inc_stat(intf, unhandled_lan_responses); in handle_lan_get_msg_rsp()
3268 ipmi_inc_stat(intf, handled_lan_responses); in handle_lan_get_msg_rsp()
3274 static int handle_lan_get_msg_cmd(ipmi_smi_t intf, in handle_lan_get_msg_cmd() argument
3288 ipmi_inc_stat(intf, invalid_commands); in handle_lan_get_msg_cmd()
3302 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_lan_get_msg_cmd()
3312 ipmi_inc_stat(intf, unhandled_commands); in handle_lan_get_msg_cmd()
3321 ipmi_inc_stat(intf, handled_commands); in handle_lan_get_msg_cmd()
3374 static int handle_oem_get_msg_cmd(ipmi_smi_t intf, in handle_oem_get_msg_cmd() argument
3392 ipmi_inc_stat(intf, invalid_commands); in handle_oem_get_msg_cmd()
3410 rcvr = find_cmd_rcvr(intf, netfn, cmd, chan); in handle_oem_get_msg_cmd()
3420 ipmi_inc_stat(intf, unhandled_commands); in handle_oem_get_msg_cmd()
3430 ipmi_inc_stat(intf, handled_commands); in handle_oem_get_msg_cmd()
3494 static int handle_read_event_rsp(ipmi_smi_t intf, in handle_read_event_rsp() argument
3506 ipmi_inc_stat(intf, invalid_events); in handle_read_event_rsp()
3517 spin_lock_irqsave(&intf->events_lock, flags); in handle_read_event_rsp()
3519 ipmi_inc_stat(intf, events); in handle_read_event_rsp()
3526 list_for_each_entry_rcu(user, &intf->users, link) { in handle_read_event_rsp()
3562 } else if (intf->waiting_events_count < MAX_EVENTS_IN_QUEUE) { in handle_read_event_rsp()
3579 list_add_tail(&(recv_msg->link), &(intf->waiting_events)); in handle_read_event_rsp()
3580 intf->waiting_events_count++; in handle_read_event_rsp()
3581 } else if (!intf->event_msg_printed) { in handle_read_event_rsp()
3588 intf->event_msg_printed = 1; in handle_read_event_rsp()
3592 spin_unlock_irqrestore(&(intf->events_lock), flags); in handle_read_event_rsp()
3597 static int handle_bmc_rsp(ipmi_smi_t intf, in handle_bmc_rsp() argument
3617 ipmi_inc_stat(intf, unhandled_local_responses); in handle_bmc_rsp()
3622 ipmi_inc_stat(intf, handled_local_responses); in handle_bmc_rsp()
3648 static int handle_one_recv_msg(ipmi_smi_t intf, in handle_one_recv_msg() argument
3737 if (intf->curr_channel < IPMI_MAX_CHANNELS) { in handle_one_recv_msg()
3742 switch (intf->channels[chan].medium) { in handle_one_recv_msg()
3749 requeue = handle_ipmb_get_msg_rsp(intf, msg); in handle_one_recv_msg()
3755 requeue = handle_ipmb_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3766 requeue = handle_lan_get_msg_rsp(intf, msg); in handle_one_recv_msg()
3772 requeue = handle_lan_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3779 if ((intf->channels[chan].medium in handle_one_recv_msg()
3781 && (intf->channels[chan].medium in handle_one_recv_msg()
3783 requeue = handle_oem_get_msg_cmd(intf, msg); in handle_one_recv_msg()
3796 requeue = handle_read_event_rsp(intf, msg); in handle_one_recv_msg()
3799 requeue = handle_bmc_rsp(intf, msg); in handle_one_recv_msg()
3809 static void handle_new_recv_msgs(ipmi_smi_t intf) in handle_new_recv_msgs() argument
3814 int run_to_completion = intf->run_to_completion; in handle_new_recv_msgs()
3818 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3819 while (!list_empty(&intf->waiting_rcv_msgs)) { in handle_new_recv_msgs()
3820 smi_msg = list_entry(intf->waiting_rcv_msgs.next, in handle_new_recv_msgs()
3823 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, in handle_new_recv_msgs()
3825 rv = handle_one_recv_msg(intf, smi_msg); in handle_new_recv_msgs()
3827 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3843 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, flags); in handle_new_recv_msgs()
3849 if (atomic_add_unless(&intf->watchdog_pretimeouts_to_deliver, -1, 0)) { in handle_new_recv_msgs()
3853 list_for_each_entry_rcu(user, &intf->users, link) { in handle_new_recv_msgs()
3865 ipmi_smi_t intf = (ipmi_smi_t) val; in smi_recv_tasklet() local
3866 int run_to_completion = intf->run_to_completion; in smi_recv_tasklet()
3877 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in smi_recv_tasklet()
3878 if (intf->curr_msg == NULL && !intf->in_shutdown) { in smi_recv_tasklet()
3882 if (!list_empty(&intf->hp_xmit_msgs)) in smi_recv_tasklet()
3883 entry = intf->hp_xmit_msgs.next; in smi_recv_tasklet()
3884 else if (!list_empty(&intf->xmit_msgs)) in smi_recv_tasklet()
3885 entry = intf->xmit_msgs.next; in smi_recv_tasklet()
3890 intf->curr_msg = newmsg; in smi_recv_tasklet()
3894 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in smi_recv_tasklet()
3896 intf->handlers->sender(intf->send_info, newmsg); in smi_recv_tasklet()
3898 handle_new_recv_msgs(intf); in smi_recv_tasklet()
3902 void ipmi_smi_msg_received(ipmi_smi_t intf, in ipmi_smi_msg_received() argument
3906 int run_to_completion = intf->run_to_completion; in ipmi_smi_msg_received()
3913 if (intf->in_shutdown) in ipmi_smi_msg_received()
3939 else if ((intf->channels[chan].medium in ipmi_smi_msg_received()
3941 || (intf->channels[chan].medium in ipmi_smi_msg_received()
3943 ipmi_inc_stat(intf, sent_lan_command_errs); in ipmi_smi_msg_received()
3945 ipmi_inc_stat(intf, sent_ipmb_command_errs); in ipmi_smi_msg_received()
3946 intf_err_seq(intf, msg->msgid, msg->rsp[2]); in ipmi_smi_msg_received()
3949 intf_start_seq_timer(intf, msg->msgid); in ipmi_smi_msg_received()
3959 spin_lock_irqsave(&intf->waiting_rcv_msgs_lock, flags); in ipmi_smi_msg_received()
3960 list_add_tail(&msg->link, &intf->waiting_rcv_msgs); in ipmi_smi_msg_received()
3962 spin_unlock_irqrestore(&intf->waiting_rcv_msgs_lock, in ipmi_smi_msg_received()
3967 spin_lock_irqsave(&intf->xmit_msgs_lock, flags); in ipmi_smi_msg_received()
3972 if (msg == intf->curr_msg) in ipmi_smi_msg_received()
3973 intf->curr_msg = NULL; in ipmi_smi_msg_received()
3975 spin_unlock_irqrestore(&intf->xmit_msgs_lock, flags); in ipmi_smi_msg_received()
3978 smi_recv_tasklet((unsigned long) intf); in ipmi_smi_msg_received()
3980 tasklet_schedule(&intf->recv_tasklet); in ipmi_smi_msg_received()
3984 void ipmi_smi_watchdog_pretimeout(ipmi_smi_t intf) in ipmi_smi_watchdog_pretimeout() argument
3986 if (intf->in_shutdown) in ipmi_smi_watchdog_pretimeout()
3989 atomic_set(&intf->watchdog_pretimeouts_to_deliver, 1); in ipmi_smi_watchdog_pretimeout()
3990 tasklet_schedule(&intf->recv_tasklet); in ipmi_smi_watchdog_pretimeout()
3995 smi_from_recv_msg(ipmi_smi_t intf, struct ipmi_recv_msg *recv_msg, in smi_from_recv_msg() argument
4022 static void check_msg_timeout(ipmi_smi_t intf, struct seq_table *ent, in check_msg_timeout() argument
4030 if (intf->in_shutdown) in check_msg_timeout()
4048 ipmi_inc_stat(intf, timed_out_ipmb_broadcasts); in check_msg_timeout()
4050 ipmi_inc_stat(intf, timed_out_lan_commands); in check_msg_timeout()
4052 ipmi_inc_stat(intf, timed_out_ipmb_commands); in check_msg_timeout()
4065 smi_msg = smi_from_recv_msg(intf, ent->recv_msg, slot, in check_msg_timeout()
4069 ipmi_inc_stat(intf, in check_msg_timeout()
4072 ipmi_inc_stat(intf, in check_msg_timeout()
4077 spin_unlock_irqrestore(&intf->seq_lock, *flags); in check_msg_timeout()
4086 handlers = intf->handlers; in check_msg_timeout()
4089 ipmi_inc_stat(intf, in check_msg_timeout()
4092 ipmi_inc_stat(intf, in check_msg_timeout()
4095 smi_send(intf, handlers, smi_msg, 0); in check_msg_timeout()
4099 spin_lock_irqsave(&intf->seq_lock, *flags); in check_msg_timeout()
4103 static unsigned int ipmi_timeout_handler(ipmi_smi_t intf, long timeout_period) in ipmi_timeout_handler() argument
4117 spin_lock_irqsave(&intf->seq_lock, flags); in ipmi_timeout_handler()
4119 check_msg_timeout(intf, &(intf->seq_table[i]), in ipmi_timeout_handler()
4122 spin_unlock_irqrestore(&intf->seq_lock, flags); in ipmi_timeout_handler()
4135 if (intf->auto_maintenance_timeout > 0) { in ipmi_timeout_handler()
4136 spin_lock_irqsave(&intf->maintenance_mode_lock, flags); in ipmi_timeout_handler()
4137 if (intf->auto_maintenance_timeout > 0) { in ipmi_timeout_handler()
4138 intf->auto_maintenance_timeout in ipmi_timeout_handler()
4140 if (!intf->maintenance_mode in ipmi_timeout_handler()
4141 && (intf->auto_maintenance_timeout <= 0)) { in ipmi_timeout_handler()
4142 intf->maintenance_mode_enable = false; in ipmi_timeout_handler()
4143 maintenance_mode_update(intf); in ipmi_timeout_handler()
4146 spin_unlock_irqrestore(&intf->maintenance_mode_lock, in ipmi_timeout_handler()
4150 tasklet_schedule(&intf->recv_tasklet); in ipmi_timeout_handler()
4155 static void ipmi_request_event(ipmi_smi_t intf) in ipmi_request_event() argument
4158 if (intf->maintenance_mode_enable) in ipmi_request_event()
4161 if (!intf->in_shutdown) in ipmi_request_event()
4162 intf->handlers->request_events(intf->send_info); in ipmi_request_event()
4171 ipmi_smi_t intf; in ipmi_timeout() local
4178 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in ipmi_timeout()
4181 if (atomic_read(&intf->event_waiters)) { in ipmi_timeout()
4182 intf->ticks_to_req_ev--; in ipmi_timeout()
4183 if (intf->ticks_to_req_ev == 0) { in ipmi_timeout()
4184 ipmi_request_event(intf); in ipmi_timeout()
4185 intf->ticks_to_req_ev = IPMI_REQUEST_EV_TIME; in ipmi_timeout()
4190 lnt += ipmi_timeout_handler(intf, IPMI_TIMEOUT_TIME); in ipmi_timeout()
4193 if (lnt != intf->last_needs_timer && in ipmi_timeout()
4194 intf->handlers->set_need_watch) in ipmi_timeout()
4195 intf->handlers->set_need_watch(intf->send_info, lnt); in ipmi_timeout()
4196 intf->last_needs_timer = lnt; in ipmi_timeout()
4206 static void need_waiter(ipmi_smi_t intf) in need_waiter() argument
4279 static void ipmi_panic_request_and_wait(ipmi_smi_t intf, in ipmi_panic_request_and_wait() argument
4291 intf, in ipmi_panic_request_and_wait()
4295 intf, in ipmi_panic_request_and_wait()
4299 intf->channels[0].address, in ipmi_panic_request_and_wait()
4300 intf->channels[0].lun, in ipmi_panic_request_and_wait()
4304 else if (intf->handlers->flush_messages) in ipmi_panic_request_and_wait()
4305 intf->handlers->flush_messages(intf->send_info); in ipmi_panic_request_and_wait()
4308 ipmi_poll(intf); in ipmi_panic_request_and_wait()
4312 static void event_receiver_fetcher(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in event_receiver_fetcher() argument
4319 intf->event_receiver = msg->msg.data[1]; in event_receiver_fetcher()
4320 intf->event_receiver_lun = msg->msg.data[2] & 0x3; in event_receiver_fetcher()
4324 static void device_id_fetcher(ipmi_smi_t intf, struct ipmi_recv_msg *msg) in device_id_fetcher() argument
4334 intf->local_sel_device = (msg->msg.data[6] >> 2) & 1; in device_id_fetcher()
4335 intf->local_event_generator = (msg->msg.data[6] >> 5) & 1; in device_id_fetcher()
4343 ipmi_smi_t intf; in send_panic_events() local
4375 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in send_panic_events()
4376 if (!intf->handlers) in send_panic_events()
4381 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4393 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in send_panic_events()
4398 if (intf->intf_num == -1) in send_panic_events()
4419 intf->local_sel_device = 0; in send_panic_events()
4420 intf->local_event_generator = 0; in send_panic_events()
4421 intf->event_receiver = 0; in send_panic_events()
4428 intf->null_user_handler = device_id_fetcher; in send_panic_events()
4429 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4431 if (intf->local_event_generator) { in send_panic_events()
4437 intf->null_user_handler = event_receiver_fetcher; in send_panic_events()
4438 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4440 intf->null_user_handler = NULL; in send_panic_events()
4447 if (((intf->event_receiver & 1) == 0) in send_panic_events()
4448 && (intf->event_receiver != 0) in send_panic_events()
4449 && (intf->event_receiver != intf->channels[0].address)) { in send_panic_events()
4457 ipmb->lun = intf->event_receiver_lun; in send_panic_events()
4458 ipmb->slave_addr = intf->event_receiver; in send_panic_events()
4459 } else if (intf->local_sel_device) { in send_panic_events()
4486 data[3] = intf->channels[0].address; in send_panic_events()
4495 ipmi_panic_request_and_wait(intf, &addr, &msg); in send_panic_events()
4508 ipmi_smi_t intf; in panic_event() local
4515 list_for_each_entry_rcu(intf, &ipmi_interfaces, link) { in panic_event()
4516 if (!intf->handlers) in panic_event()
4526 if (!spin_trylock(&intf->xmit_msgs_lock)) { in panic_event()
4527 INIT_LIST_HEAD(&intf->xmit_msgs); in panic_event()
4528 INIT_LIST_HEAD(&intf->hp_xmit_msgs); in panic_event()
4530 spin_unlock(&intf->xmit_msgs_lock); in panic_event()
4532 if (!spin_trylock(&intf->waiting_rcv_msgs_lock)) in panic_event()
4533 INIT_LIST_HEAD(&intf->waiting_rcv_msgs); in panic_event()
4535 spin_unlock(&intf->waiting_rcv_msgs_lock); in panic_event()
4537 intf->run_to_completion = 1; in panic_event()
4538 intf->handlers->set_run_to_completion(intf->send_info, 1); in panic_event()