Searched refs:work_q (Results 1 - 29 of 29) sorted by relevance

/linux-4.1.27/drivers/scsi/
H A Dscsi_error.c328 * @work_q: Queue of scsi cmds to process.
331 struct list_head *work_q) scsi_eh_prt_fail_stats()
341 list_for_each_entry(scmd, work_q, eh_entry) { list_for_each_entry()
1133 * @work_q: Queue of commands to process.
1151 int scsi_eh_get_sense(struct list_head *work_q, scsi_eh_get_sense() argument
1158 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe()
1209 return list_empty(work_q);
1247 * @work_q: queue for commands which still need more error recovery
1254 * are still failing to respond are returned to the work_q for more
1258 struct list_head *work_q, scsi_eh_test_devices()
1271 /* Push items back onto work_q */ scsi_eh_test_devices()
1272 list_splice_init(cmd_list, work_q); scsi_eh_test_devices()
1293 list_move_tail(&scmd->eh_entry, work_q); list_for_each_entry_safe()
1296 return list_empty(work_q);
1302 * @work_q: &list_head for pending commands.
1312 static int scsi_eh_abort_cmds(struct list_head *work_q, scsi_eh_abort_cmds() argument
1320 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe()
1325 list_splice_init(&check_list, work_q); list_for_each_entry_safe()
1330 return list_empty(work_q); list_for_each_entry_safe()
1341 list_splice_init(&check_list, work_q); list_for_each_entry_safe()
1342 return list_empty(work_q); list_for_each_entry_safe()
1351 return scsi_eh_test_devices(&check_list, work_q, done_q, 0);
1381 * @work_q: &list_head for pending commands.
1389 struct list_head *work_q, scsi_eh_stu()
1404 list_for_each_entry(scmd, work_q, eh_entry) list_for_each_entry()
1423 work_q, eh_entry) { list_for_each_entry_safe()
1437 return list_empty(work_q);
1444 * @work_q: &list_head for pending commands.
1454 struct list_head *work_q, scsi_eh_bus_device_reset()
1470 list_for_each_entry(scmd, work_q, eh_entry) list_for_each_entry()
1488 work_q, eh_entry) { list_for_each_entry_safe()
1502 return list_empty(work_q);
1508 * @work_q: &list_head for pending commands.
1515 struct list_head *work_q, scsi_eh_target_reset()
1521 list_splice_init(work_q, &tmp_list); scsi_eh_target_reset()
1530 list_splice_init(&check_list, work_q); scsi_eh_target_reset()
1531 list_splice_init(&tmp_list, work_q); scsi_eh_target_reset()
1536 return list_empty(work_q); scsi_eh_target_reset()
1563 list_move(&scmd->eh_entry, work_q); scsi_eh_target_reset()
1567 return scsi_eh_test_devices(&check_list, work_q, done_q, 0); scsi_eh_target_reset()
1573 * @work_q: &list_head for pending commands.
1577 struct list_head *work_q, scsi_eh_bus_reset()
1594 list_splice_init(&check_list, work_q); scsi_eh_bus_reset()
1599 return list_empty(work_q); scsi_eh_bus_reset()
1603 list_for_each_entry(scmd, work_q, eh_entry) { list_for_each_entry()
1622 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe()
1639 return scsi_eh_test_devices(&check_list, work_q, done_q, 0);
1645 * @work_q: &list_head for pending commands.
1649 struct list_head *work_q, scsi_eh_host_reset()
1656 if (!list_empty(work_q)) { scsi_eh_host_reset()
1657 scmd = list_entry(work_q->next, scsi_eh_host_reset()
1667 list_splice_init(work_q, &check_list); scsi_eh_host_reset()
1669 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe()
1679 return scsi_eh_test_devices(&check_list, work_q, done_q, 1);
1684 * @work_q: &list_head for pending commands.
1687 static void scsi_eh_offline_sdevs(struct list_head *work_q, scsi_eh_offline_sdevs() argument
1692 list_for_each_entry_safe(scmd, next, work_q, eh_entry) { list_for_each_entry_safe()
2055 * @work_q: &list_head for pending commands.
2059 struct list_head *work_q, scsi_eh_ready_devs()
2062 if (!scsi_eh_stu(shost, work_q, done_q)) scsi_eh_ready_devs()
2063 if (!scsi_eh_bus_device_reset(shost, work_q, done_q)) scsi_eh_ready_devs()
2064 if (!scsi_eh_target_reset(shost, work_q, done_q)) scsi_eh_ready_devs()
2065 if (!scsi_eh_bus_reset(shost, work_q, done_q)) scsi_eh_ready_devs()
2066 if (!scsi_eh_host_reset(shost, work_q, done_q)) scsi_eh_ready_devs()
2067 scsi_eh_offline_sdevs(work_q, scsi_eh_ready_devs()
330 scsi_eh_prt_fail_stats(struct Scsi_Host *shost, struct list_head *work_q) scsi_eh_prt_fail_stats() argument
1257 scsi_eh_test_devices(struct list_head *cmd_list, struct list_head *work_q, struct list_head *done_q, int try_stu) scsi_eh_test_devices() argument
1388 scsi_eh_stu(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_stu() argument
1453 scsi_eh_bus_device_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_bus_device_reset() argument
1514 scsi_eh_target_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_target_reset() argument
1576 scsi_eh_bus_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_bus_reset() argument
1648 scsi_eh_host_reset(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_host_reset() argument
2058 scsi_eh_ready_devs(struct Scsi_Host *shost, struct list_head *work_q, struct list_head *done_q) scsi_eh_ready_devs() argument
H A Dhosts.c273 shost->work_q = create_singlethread_workqueue( scsi_add_host_with_dma()
275 if (!shost->work_q) { scsi_add_host_with_dma()
289 if (shost->work_q) scsi_add_host_with_dma()
290 destroy_workqueue(shost->work_q); scsi_add_host_with_dma()
320 if (shost->work_q) scsi_host_dev_release()
321 destroy_workqueue(shost->work_q); scsi_host_dev_release()
611 if (unlikely(!shost->work_q)) { scsi_queue_work()
620 return queue_work(shost->work_q, work); scsi_queue_work()
630 if (!shost->work_q) { scsi_flush_work()
638 flush_workqueue(shost->work_q); scsi_flush_work()
H A Dscsi_priv.h76 struct list_head *work_q,
78 int scsi_eh_get_sense(struct list_head *work_q,
H A Dstex.c319 struct workqueue_struct *work_q; member in struct:st_hba
857 queue_work(hba->work_q, &hba->reset_work); stex_intr()
955 queue_work(hba->work_q, &hba->reset_work); stex_ss_intr()
1642 hba->work_q = create_singlethread_workqueue(hba->work_q_name); stex_probe()
1643 if (!hba->work_q) { stex_probe()
1685 destroy_workqueue(hba->work_q); stex_probe()
1753 destroy_workqueue(hba->work_q); stex_hba_free()
H A Dscsi_transport_fc.c443 fc_host->work_q = alloc_workqueue("%s", 0, 0, fc_host->work_q_name); fc_host_setup()
444 if (!fc_host->work_q) fc_host_setup()
454 destroy_workqueue(fc_host->work_q); fc_host_setup()
455 fc_host->work_q = NULL; fc_host_setup()
2442 struct workqueue_struct *work_q; fc_remove_host() local
2473 if (fc_host->work_q) { fc_remove_host()
2474 work_q = fc_host->work_q; fc_remove_host()
2475 fc_host->work_q = NULL; fc_remove_host()
2476 destroy_workqueue(work_q); fc_remove_host()
2481 work_q = fc_host->devloss_work_q; fc_remove_host()
2483 destroy_workqueue(work_q); fc_remove_host()
2538 * if a scan is pending, flush the SCSI Host work_q so that fc_rport_final_delete()
2982 * No need to flush the fc_host work_q's, as all adds are synchronous. fc_remote_port_delete()
2985 * (in fc_rport_final_delete()) we'll flush the scsi host work_q if fc_remote_port_delete()
H A Dmvumi.h172 struct work_struct work_q; member in struct:mvumi_events_wq
H A Dmvumi.c1771 container_of(work, struct mvumi_events_wq, work_q); mvumi_scan_events()
1791 INIT_WORK(&mu_ev->work_q, mvumi_scan_events); mvumi_launch_events()
1796 schedule_work(&mu_ev->work_q); mvumi_launch_events()
H A Dipr.c1445 schedule_work(&ioa_cfg->work_q); ipr_handle_config_change()
1452 schedule_work(&ioa_cfg->work_q); ipr_handle_config_change()
3264 container_of(work, struct ipr_ioa_cfg, work_q); ipr_worker_thread()
4251 schedule_work(&ioa_cfg->work_q); ipr_alloc_dump()
7033 schedule_work(&ioa_cfg->work_q); ipr_ioa_reset_done()
7057 schedule_work(&ioa_cfg->work_q); ipr_ioa_reset_done()
8335 schedule_work(&ioa_cfg->work_q); ipr_reset_restore_cfg_space()
9586 INIT_WORK(&ioa_cfg->work_q, ipr_worker_thread); ipr_init_ioa_cfg()
10182 flush_work(&ioa_cfg->work_q); __ipr_remove()
10289 schedule_work(&ioa_cfg->work_q); ipr_probe()
H A Dipr.h1546 struct work_struct work_q; member in struct:ipr_ioa_cfg
/linux-4.1.27/include/scsi/
H A Dsas_ata.h43 void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q,
70 static inline void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q, sas_ata_eh() argument
H A Dfcoe_sysfs.h66 struct workqueue_struct *work_q; member in struct:fcoe_ctlr_device
H A Dscsi_transport_fc.h540 struct workqueue_struct *work_q; member in struct:fc_host_attrs
618 (((struct fc_host_attrs *)(x)->shost_data)->work_q)
H A Dscsi_host.h680 struct workqueue_struct *work_q; member in struct:Scsi_Host
/linux-4.1.27/drivers/scsi/fcoe/
H A Dfcoe_sysfs.c62 ((x)->work_q)
701 ctlr->work_q = create_singlethread_workqueue( fcoe_ctlr_device_add()
703 if (!ctlr->work_q) fcoe_ctlr_device_add()
725 destroy_workqueue(ctlr->work_q); fcoe_ctlr_device_add()
726 ctlr->work_q = NULL; fcoe_ctlr_device_add()
771 destroy_workqueue(ctlr->work_q); fcoe_ctlr_device_delete()
772 ctlr->work_q = NULL; fcoe_ctlr_device_delete()
/linux-4.1.27/drivers/scsi/libsas/
H A Dsas_scsi_host.c573 static void sas_eh_handle_sas_errors(struct Scsi_Host *shost, struct list_head *work_q) sas_eh_handle_sas_errors() argument
584 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe()
601 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe()
642 sas_scsi_clear_queue_lu(work_q, cmd); list_for_each_entry_safe()
657 sas_scsi_clear_queue_I_T(work_q, dev); list_for_each_entry_safe()
671 sas_scsi_clear_queue_port(work_q, list_for_each_entry_safe()
700 list_splice_tail(&done, work_q);
701 list_splice_tail_init(&ha->eh_ata_q, work_q);
706 list_for_each_entry_safe(cmd, n, work_q, eh_entry)
H A Dsas_event.c59 struct workqueue_struct *wq = ha->core.shost->work_q; __sas_drain_work()
H A Dsas_ata.c793 void sas_ata_eh(struct Scsi_Host *shost, struct list_head *work_q, sas_ata_eh() argument
803 list_for_each_entry_safe(cmd, n, work_q, eh_entry) { list_for_each_entry_safe()
/linux-4.1.27/drivers/net/ethernet/brocade/bna/
H A Dbnad.h349 struct workqueue_struct *work_q; member in struct:bnad
H A Dbnad.c1135 queue_delayed_work(bnad->work_q, &tx_info->tx_cleanup_work, bnad_tx_cleanup()
1158 queue_delayed_work(bnad->work_q, &tx_info->tx_cleanup_work, 0); bnad_cb_tx_cleanup()
1240 queue_work(bnad->work_q, &rx_info->rx_cleanup_work); bnad_cb_rx_cleanup()
3513 bnad->work_q = create_singlethread_workqueue(bnad->wq_name); bnad_init()
3514 if (!bnad->work_q) { bnad_init()
3530 if (bnad->work_q) { bnad_uninit()
3531 flush_workqueue(bnad->work_q); bnad_uninit()
3532 destroy_workqueue(bnad->work_q); bnad_uninit()
3533 bnad->work_q = NULL; bnad_uninit()
/linux-4.1.27/drivers/message/fusion/
H A Dmptfc.c1500 struct workqueue_struct *work_q; mptfc_remove() local
1505 if ((work_q=ioc->fc_rescan_work_q)) { mptfc_remove()
1509 destroy_workqueue(work_q); mptfc_remove()
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_83xx_hw.c841 flush_workqueue(ahw->mailbox->work_q); qlcnic_83xx_poll_for_mbx_completion()
874 flush_workqueue(mbx->work_q); qlcnic_83xx_issue_cmd()
3800 destroy_workqueue(mbx->work_q); qlcnic_83xx_free_mailbox()
3926 flush_workqueue(mbx->work_q); qlcnic_83xx_detach_mailbox_work()
3947 queue_work(mbx->work_q, &mbx->work); qlcnic_83xx_enqueue_mbx_cmd()
4102 mbx->work_q = create_singlethread_workqueue("qlcnic_mailbox"); qlcnic_83xx_init_mailbox_work()
4103 if (mbx->work_q == NULL) { qlcnic_83xx_init_mailbox_work()
H A Dqlcnic_sriov_common.c332 flush_workqueue(mbx->work_q); qlcnic_sriov_post_bc_msg()
H A Dqlcnic.h1093 struct workqueue_struct *work_q; member in struct:qlcnic_mailbox
/linux-4.1.27/drivers/scsi/mvsas/
H A Dmv_sas.h427 struct delayed_work work_q; member in struct:mvs_wq
H A Dmv_sas.c1891 struct mvs_wq *mwq = container_of(dw, struct mvs_wq, work_q); mvs_work_queue()
1943 MV_INIT_DELAYED_WORK(&mwq->work_q, mvs_work_queue, mwq); mvs_handle_event()
1945 schedule_delayed_work(&mwq->work_q, HZ * 2); mvs_handle_event()
H A Dmv_init.c158 cancel_delayed_work(&mwq->work_q); mvs_free()
/linux-4.1.27/drivers/net/wan/
H A Dfarsync.c573 static void fst_process_tx_work_q(unsigned long work_q);
574 static void fst_process_int_work_q(unsigned long work_q);
607 fst_process_tx_work_q(unsigned long /*void **/work_q) fst_process_tx_work_q() argument
637 fst_process_int_work_q(unsigned long /*void **/work_q) fst_process_int_work_q() argument
/linux-4.1.27/drivers/infiniband/hw/nes/
H A Dnes_hw.h1063 struct workqueue_struct *work_q; member in struct:nes_adapter
/linux-4.1.27/drivers/scsi/bfa/
H A Dbfa_fcs_lport.c6618 * We queue the vport delete work to the IM work_q from here. bfa_fcs_vport_free()

Completed in 1513 milliseconds