/linux-4.1.27/drivers/scsi/ |
H A D | scsi_error.c | 328 * @work_q: Queue of scsi cmds to process. 331 struct list_head *work_q) scsi_eh_prt_fail_stats() 341 list_for_each_entry(scmd, work_q, eh_entry) { list_for_each_entry() 1133 * @work_q: Queue of commands to process. 1151 int scsi_eh_get_sense(struct list_head *work_q, scsi_eh_get_sense() argument 1158 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe() 1209 return list_empty(work_q); 1247 * @work_q: queue for commands which still need more error recovery 1254 * are still failing to respond are returned to the work_q for more 1258 struct list_head *work_q, scsi_eh_test_devices() 1271 /* Push items back onto work_q */ scsi_eh_test_devices() 1272 list_splice_init(cmd_list, work_q); scsi_eh_test_devices() 1293 list_move_tail(&scmd->eh_entry, work_q); list_for_each_entry_safe() 1296 return list_empty(work_q); 1302 * @work_q: &list_head for pending commands. 1312 static int scsi_eh_abort_cmds(struct list_head *work_q, scsi_eh_abort_cmds() argument 1320 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe() 1325 list_splice_init(&check_list, work_q); list_for_each_entry_safe() 1330 return list_empty(work_q); list_for_each_entry_safe() 1341 list_splice_init(&check_list, work_q); list_for_each_entry_safe() 1342 return list_empty(work_q); list_for_each_entry_safe() 1351 return scsi_eh_test_devices(&check_list, work_q, done_q, 0); 1381 * @work_q: &list_head for pending commands. 1389 struct list_head *work_q, scsi_eh_stu() 1404 list_for_each_entry(scmd, work_q, eh_entry) list_for_each_entry() 1423 work_q, eh_entry) { list_for_each_entry_safe() 1437 return list_empty(work_q); 1444 * @work_q: &list_head for pending commands. 1454 struct list_head *work_q, scsi_eh_bus_device_reset() 1470 list_for_each_entry(scmd, work_q, eh_entry) list_for_each_entry() 1488 work_q, eh_entry) { list_for_each_entry_safe() 1502 return list_empty(work_q); 1508 * @work_q: &list_head for pending commands. 1515 struct list_head *work_q, scsi_eh_target_reset() 1521 list_splice_init(work_q, &tmp_list); scsi_eh_target_reset() 1530 list_splice_init(&check_list, work_q); scsi_eh_target_reset() 1531 list_splice_init(&tmp_list, work_q); scsi_eh_target_reset() 1536 return list_empty(work_q); scsi_eh_target_reset() 1563 list_move(&scmd->eh_entry, work_q); scsi_eh_target_reset() 1567 return scsi_eh_test_devices(&check_list, work_q, done_q, 0); scsi_eh_target_reset() 1573 * @work_q: &list_head for pending commands. 1577 struct list_head *work_q, scsi_eh_bus_reset() 1594 list_splice_init(&check_list, work_q); scsi_eh_bus_reset() 1599 return list_empty(work_q); scsi_eh_bus_reset() 1603 list_for_each_entry(scmd, work_q, eh_entry) { list_for_each_entry() 1622 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe() 1639 return scsi_eh_test_devices(&check_list, work_q, done_q, 0); 1645 * @work_q: &list_head for pending commands. 1649 struct list_head *work_q, scsi_eh_host_reset() 1656 if (!list_empty(work_q)) { scsi_eh_host_reset() 1657 scmd = list_entry(work_q->next, scsi_eh_host_reset() 1667 list_splice_init(work_q, &check_list); scsi_eh_host_reset() 1669 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe() 1679 return scsi_eh_test_devices(&check_list, work_q, done_q, 1); 1684 * @work_q: &list_head for pending commands. 1687 static void scsi_eh_offline_sdevs(struct list_head *work_q, scsi_eh_offline_sdevs() argument 1692 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe() 2055 * @work_q: &list_head for pending commands. 2059 struct list_head *work_q, scsi_eh_ready_devs() 2062 if (!scsi_eh_stu(shost, work_q, done_q)) scsi_eh_ready_devs() 2063 if (!scsi_eh_bus_device_reset(shost, work_q, done_q)) scsi_eh_ready_devs() 2064 if (!scsi_eh_target_reset(shost, work_q, done_q)) scsi_eh_ready_devs() 2065 if (!scsi_eh_bus_reset(shost, work_q, done_q)) scsi_eh_ready_devs() 2066 if (!scsi_eh_host_reset(shost, work_q, done_q)) scsi_eh_ready_devs() 2067 scsi_eh_offline_sdevs(work_q, scsi_eh_ready_devs() 330 scsi_eh_prt_fail_stats(struct Scsi_Host *shost, struct list_head *work_q) scsi_eh_prt_fail_stats() argument 1257 scsi_eh_test_devices(struct list_head *cmd_list, struct list_head *work_q, struct list_head *done_q, int try_stu) scsi_eh_test_devices() argument 1388 scsi_eh_stu(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_stu() argument 1453 scsi_eh_bus_device_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_bus_device_reset() argument 1514 scsi_eh_target_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_target_reset() argument 1576 scsi_eh_bus_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_bus_reset() argument 1648 scsi_eh_host_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_host_reset() argument 2058 scsi_eh_ready_devs(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_ready_devs() argument
|
H A D | hosts.c | 273 shost->work_q = create_singlethread_workqueue( scsi_add_host_with_dma() 275 if (!shost->work_q) { scsi_add_host_with_dma() 289 if (shost->work_q) scsi_add_host_with_dma() 290 destroy_workqueue(shost->work_q); scsi_add_host_with_dma() 320 if (shost->work_q) scsi_host_dev_release() 321 destroy_workqueue(shost->work_q); scsi_host_dev_release() 611 if (unlikely(!shost->work_q)) { scsi_queue_work() 620 return queue_work(shost->work_q, work); scsi_queue_work() 630 if (!shost->work_q) { scsi_flush_work() 638 flush_workqueue(shost->work_q); scsi_flush_work()
|
H A D | scsi_priv.h | 76 struct list_head *work_q, 78 int scsi_eh_get_sense(struct list_head *work_q,
|
H A D | stex.c | 319 struct workqueue_struct *work_q; member in struct:st_hba 857 queue_work(hba->work_q, &hba->reset_work); stex_intr() 955 queue_work(hba->work_q, &hba->reset_work); stex_ss_intr() 1642 hba->work_q = create_singlethread_workqueue(hba->work_q_name); stex_probe() 1643 if (!hba->work_q) { stex_probe() 1685 destroy_workqueue(hba->work_q); stex_probe() 1753 destroy_workqueue(hba->work_q); stex_hba_free()
|
H A D | scsi_transport_fc.c | 443 fc_host->work_q = alloc_workqueue("%s", 0, 0, fc_host->work_q_name); fc_host_setup() 444 if (!fc_host->work_q) fc_host_setup() 454 destroy_workqueue(fc_host->work_q); fc_host_setup() 455 fc_host->work_q = NULL; fc_host_setup() 2442 struct workqueue_struct *work_q; fc_remove_host() local 2473 if (fc_host->work_q) { fc_remove_host() 2474 work_q = fc_host->work_q; fc_remove_host() 2475 fc_host->work_q = NULL; fc_remove_host() 2476 destroy_workqueue(work_q); fc_remove_host() 2481 work_q = fc_host->devloss_work_q; fc_remove_host() 2483 destroy_workqueue(work_q); fc_remove_host() 2538 * if a scan is pending, flush the SCSI Host work_q so that fc_rport_final_delete() 2982 * No need to flush the fc_host work_q's, as all adds are synchronous. fc_remote_port_delete() 2985 * (in fc_rport_final_delete()) we'll flush the scsi host work_q if fc_remote_port_delete()
|
H A D | mvumi.h | 172 struct work_struct work_q; member in struct:mvumi_events_wq
|
H A D | mvumi.c | 1771 container_of(work, struct mvumi_events_wq, work_q); mvumi_scan_events() 1791 INIT_WORK(&mu_ev->work_q, mvumi_scan_events); mvumi_launch_events() 1796 schedule_work(&mu_ev->work_q); mvumi_launch_events()
|
H A D | ipr.c | 1445 schedule_work(&ioa_cfg->work_q); ipr_handle_config_change() 1452 schedule_work(&ioa_cfg->work_q); ipr_handle_config_change() 3264 container_of(work, struct ipr_ioa_cfg, work_q); ipr_worker_thread() 4251 schedule_work(&ioa_cfg->work_q); ipr_alloc_dump() 7033 schedule_work(&ioa_cfg->work_q); ipr_ioa_reset_done() 7057 schedule_work(&ioa_cfg->work_q); ipr_ioa_reset_done() 8335 schedule_work(&ioa_cfg->work_q); ipr_reset_restore_cfg_space() 9586 INIT_WORK(&ioa_cfg->work_q, ipr_worker_thread); ipr_init_ioa_cfg() 10182 flush_work(&ioa_cfg->work_q); __ipr_remove() 10289 schedule_work(&ioa_cfg->work_q); ipr_probe()
|
H A D | ipr.h | 1546 struct work_struct work_q; member in struct:ipr_ioa_cfg
|
/linux-4.1.27/include/scsi/ |
H A D | sas_ata.h | 43 void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q, 70 static inline void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q, sas_ata_eh() argument
|
H A D | fcoe_sysfs.h | 66 struct workqueue_struct *work_q; member in struct:fcoe_ctlr_device
|
H A D | scsi_transport_fc.h | 540 struct workqueue_struct *work_q; member in struct:fc_host_attrs 618 (((struct fc_host_attrs *)(x)->shost_data)->work_q)
|
H A D | scsi_host.h | 680 struct workqueue_struct *work_q; member in struct:Scsi_Host
|
/linux-4.1.27/drivers/scsi/fcoe/ |
H A D | fcoe_sysfs.c | 62 ((x)->work_q) 701 ctlr->work_q = create_singlethread_workqueue( fcoe_ctlr_device_add() 703 if (!ctlr->work_q) fcoe_ctlr_device_add() 725 destroy_workqueue(ctlr->work_q); fcoe_ctlr_device_add() 726 ctlr->work_q = NULL; fcoe_ctlr_device_add() 771 destroy_workqueue(ctlr->work_q); fcoe_ctlr_device_delete() 772 ctlr->work_q = NULL; fcoe_ctlr_device_delete()
|
/linux-4.1.27/drivers/scsi/libsas/ |
H A D | sas_scsi_host.c | 573 static void sas_eh_handle_sas_errors(struct Scsi_Host *shost, struct list_head *work_q) sas_eh_handle_sas_errors() argument 584 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe() 601 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe() 642 sas_scsi_clear_queue_lu(work_q, cmd); list_for_each_entry_safe() 657 sas_scsi_clear_queue_I_T(work_q, dev); list_for_each_entry_safe() 671 sas_scsi_clear_queue_port(work_q, list_for_each_entry_safe() 700 list_splice_tail(&done, work_q); 701 list_splice_tail_init(&ha->eh_ata_q, work_q); 706 list_for_each_entry_safe(cmd, n, work_q, eh_entry)
|
H A D | sas_event.c | 59 struct workqueue_struct *wq = ha->core.shost->work_q; __sas_drain_work()
|
H A D | sas_ata.c | 793 void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q, sas_ata_eh() argument 803 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
H A D | bnad.h | 349 struct workqueue_struct *work_q; member in struct:bnad
|
H A D | bnad.c | 1135 queue_delayed_work(bnad->work_q, &tx_info->tx_cleanup_work, bnad_tx_cleanup() 1158 queue_delayed_work(bnad->work_q, &tx_info->tx_cleanup_work, 0); bnad_cb_tx_cleanup() 1240 queue_work(bnad->work_q, &rx_info->rx_cleanup_work); bnad_cb_rx_cleanup() 3513 bnad->work_q = create_singlethread_workqueue(bnad->wq_name); bnad_init() 3514 if (!bnad->work_q) { bnad_init() 3530 if (bnad->work_q) { bnad_uninit() 3531 flush_workqueue(bnad->work_q); bnad_uninit() 3532 destroy_workqueue(bnad->work_q); bnad_uninit() 3533 bnad->work_q = NULL; bnad_uninit()
|
/linux-4.1.27/drivers/message/fusion/ |
H A D | mptfc.c | 1500 struct workqueue_struct *work_q; mptfc_remove() local 1505 if ((work_q=ioc->fc_rescan_work_q)) { mptfc_remove() 1509 destroy_workqueue(work_q); mptfc_remove()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_83xx_hw.c | 841 flush_workqueue(ahw->mailbox->work_q); qlcnic_83xx_poll_for_mbx_completion() 874 flush_workqueue(mbx->work_q); qlcnic_83xx_issue_cmd() 3800 destroy_workqueue(mbx->work_q); qlcnic_83xx_free_mailbox() 3926 flush_workqueue(mbx->work_q); qlcnic_83xx_detach_mailbox_work() 3947 queue_work(mbx->work_q, &mbx->work); qlcnic_83xx_enqueue_mbx_cmd() 4102 mbx->work_q = create_singlethread_workqueue("qlcnic_mailbox"); qlcnic_83xx_init_mailbox_work() 4103 if (mbx->work_q == NULL) { qlcnic_83xx_init_mailbox_work()
|
H A D | qlcnic_sriov_common.c | 332 flush_workqueue(mbx->work_q); qlcnic_sriov_post_bc_msg()
|
H A D | qlcnic.h | 1093 struct workqueue_struct *work_q; member in struct:qlcnic_mailbox
|
/linux-4.1.27/drivers/scsi/mvsas/ |
H A D | mv_sas.h | 427 struct delayed_work work_q; member in struct:mvs_wq
|
H A D | mv_sas.c | 1891 struct mvs_wq *mwq = container_of(dw, struct mvs_wq, work_q); mvs_work_queue() 1943 MV_INIT_DELAYED_WORK(&mwq->work_q, mvs_work_queue, mwq); mvs_handle_event() 1945 schedule_delayed_work(&mwq->work_q, HZ * 2); mvs_handle_event()
|
H A D | mv_init.c | 158 cancel_delayed_work(&mwq->work_q); mvs_free()
|
/linux-4.1.27/drivers/net/wan/ |
H A D | farsync.c | 573 static void fst_process_tx_work_q(unsigned long work_q); 574 static void fst_process_int_work_q(unsigned long work_q); 607 fst_process_tx_work_q(unsigned long /*void **/work_q) fst_process_tx_work_q() argument 637 fst_process_int_work_q(unsigned long /*void **/work_q) fst_process_int_work_q() argument
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
H A D | nes_hw.h | 1063 struct workqueue_struct *work_q; member in struct:nes_adapter
|
/linux-4.1.27/drivers/scsi/bfa/ |
H A D | bfa_fcs_lport.c | 6618 * We queue the vport delete work to the IM work_q from here. bfa_fcs_vport_free()
|