dwork 2478 arch/x86/kvm/x86.c struct delayed_work *dwork = to_delayed_work(work); dwork 2479 arch/x86/kvm/x86.c struct kvm_arch *ka = container_of(dwork, struct kvm_arch, dwork 2503 arch/x86/kvm/x86.c struct delayed_work *dwork = to_delayed_work(work); dwork 2504 arch/x86/kvm/x86.c struct kvm_arch *ka = container_of(dwork, struct kvm_arch, dwork 1666 block/blk-core.c int kblockd_mod_delayed_work_on(int cpu, struct delayed_work *dwork, dwork 1669 block/blk-core.c return mod_delayed_work_on(cpu, kblockd_workqueue, dwork, delay); dwork 1632 block/genhd.c struct delayed_work dwork; dwork 1704 block/genhd.c cancel_delayed_work_sync(&disk->ev->dwork); dwork 1726 block/genhd.c &ev->dwork, 0); dwork 1729 block/genhd.c &ev->dwork, intv); dwork 1773 block/genhd.c &ev->dwork, 0); dwork 1838 block/genhd.c struct delayed_work *dwork = to_delayed_work(work); dwork 1839 block/genhd.c struct disk_events *ev = container_of(dwork, struct disk_events, dwork); dwork 1867 block/genhd.c &ev->dwork, intv); dwork 2036 block/genhd.c INIT_DELAYED_WORK(&ev->dwork, disk_events_workfn); dwork 1339 drivers/acpi/nfit/core.c mod_delayed_work(nfit_wq, &acpi_desc->dwork, HZ); dwork 3215 drivers/acpi/nfit/core.c queue_delayed_work(nfit_wq, &acpi_desc->dwork, tmo * HZ); dwork 3239 drivers/acpi/nfit/core.c acpi_desc = container_of(work, typeof(*acpi_desc), dwork.work); dwork 3487 drivers/acpi/nfit/core.c if (work_busy(&acpi_desc->dwork.work)) dwork 3575 drivers/acpi/nfit/core.c INIT_DELAYED_WORK(&acpi_desc->dwork, acpi_nfit_scrub); dwork 3599 drivers/acpi/nfit/core.c cancel_delayed_work_sync(&acpi_desc->dwork); dwork 232 drivers/acpi/nfit/nfit.h struct delayed_work dwork; dwork 1229 drivers/ata/libata-sff.c void ata_sff_queue_delayed_work(struct delayed_work *dwork, unsigned long delay) dwork 1231 drivers/ata/libata-sff.c queue_delayed_work(ata_sff_wq, dwork, delay); dwork 215 drivers/ata/pata_arasan_cf.c struct delayed_work dwork; dwork 552 drivers/ata/pata_arasan_cf.c ata_sff_queue_delayed_work(&acdev->dwork, 1); dwork 576 drivers/ata/pata_arasan_cf.c dwork.work); dwork 586 drivers/ata/pata_arasan_cf.c ata_sff_queue_delayed_work(&acdev->dwork, 1); dwork 667 drivers/ata/pata_arasan_cf.c cancel_delayed_work_sync(&acdev->dwork); dwork 859 drivers/ata/pata_arasan_cf.c INIT_DELAYED_WORK(&acdev->dwork, delayed_finish); dwork 121 drivers/auxdisplay/charlcd.c struct delayed_work *dwork = to_delayed_work(work); dwork 123 drivers/auxdisplay/charlcd.c container_of(dwork, struct charlcd_priv, bl_work); dwork 1878 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct delayed_work *dwork = to_delayed_work(work); dwork 1880 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c container_of(dwork, struct amdkfd_process_info, dwork 1004 drivers/gpu/drm/amd/amdkfd/kfd_process.c struct delayed_work *dwork; dwork 1006 drivers/gpu/drm/amd/amdkfd/kfd_process.c dwork = to_delayed_work(work); dwork 1011 drivers/gpu/drm/amd/amdkfd/kfd_process.c p = container_of(dwork, struct kfd_process, eviction_work); dwork 1039 drivers/gpu/drm/amd/amdkfd/kfd_process.c struct delayed_work *dwork; dwork 1043 drivers/gpu/drm/amd/amdkfd/kfd_process.c dwork = to_delayed_work(work); dwork 1048 drivers/gpu/drm/amd/amdkfd/kfd_process.c p = container_of(dwork, struct kfd_process, restore_work); dwork 3621 drivers/infiniband/hw/i40iw/i40iw_cm.c struct disconn_work *dwork = container_of(work, struct disconn_work, work); dwork 3622 drivers/infiniband/hw/i40iw/i40iw_cm.c struct i40iw_qp *iwqp = dwork->iwqp; dwork 3624 drivers/infiniband/hw/i40iw/i40iw_cm.c kfree(dwork); dwork 1733 drivers/infiniband/hw/i40iw/i40iw_main.c struct l2params_work *dwork = dwork 1735 drivers/infiniband/hw/i40iw/i40iw_main.c struct i40iw_device *iwdev = dwork->iwdev; dwork 1737 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_change_l2params(&iwdev->vsi, &dwork->l2params); dwork 686 drivers/infiniband/hw/mlx5/mlx5_ib.h struct delayed_work dwork; dwork 371 drivers/infiniband/hw/mlx5/mr.c queue_delayed_work(cache->wq, &ent->dwork, dwork 376 drivers/infiniband/hw/mlx5/mr.c queue_delayed_work(cache->wq, &ent->dwork, dwork 401 drivers/infiniband/hw/mlx5/mr.c queue_delayed_work(cache->wq, &ent->dwork, 300 * HZ); dwork 410 drivers/infiniband/hw/mlx5/mr.c ent = container_of(work, struct mlx5_cache_ent, dwork.work); dwork 543 drivers/infiniband/hw/mlx5/mr.c cancel_delayed_work(&ent->dwork); dwork 632 drivers/infiniband/hw/mlx5/mr.c INIT_DELAYED_WORK(&ent->dwork, delayed_cache_work_func); dwork 201 drivers/infiniband/hw/qedr/qedr_iw_cm.c struct qedr_discon_work *dwork = dwork 204 drivers/infiniband/hw/qedr/qedr_iw_cm.c struct qedr_iw_ep *ep = dwork->ep; dwork 218 drivers/infiniband/hw/qedr/qedr_iw_cm.c event.status = dwork->status; dwork 224 drivers/infiniband/hw/qedr/qedr_iw_cm.c if (dwork->status) dwork 240 drivers/infiniband/hw/qedr/qedr_iw_cm.c kfree(dwork); dwork 57 drivers/input/keyboard/qt2160.c struct delayed_work dwork; dwork 202 drivers/input/keyboard/qt2160.c mod_delayed_work(system_wq, &qt2160->dwork, 0); dwork 209 drivers/input/keyboard/qt2160.c schedule_delayed_work(&qt2160->dwork, QT2160_CYCLE_INTERVAL); dwork 215 drivers/input/keyboard/qt2160.c container_of(work, struct qt2160_data, dwork.work); dwork 372 drivers/input/keyboard/qt2160.c INIT_DELAYED_WORK(&qt2160->dwork, qt2160_worker); dwork 445 drivers/input/keyboard/qt2160.c cancel_delayed_work_sync(&qt2160->dwork); dwork 46 drivers/input/keyboard/tca6416-keypad.c struct delayed_work dwork; dwork 135 drivers/input/keyboard/tca6416-keypad.c container_of(work, struct tca6416_keypad_chip, dwork.work); dwork 138 drivers/input/keyboard/tca6416-keypad.c schedule_delayed_work(&chip->dwork, msecs_to_jiffies(100)); dwork 149 drivers/input/keyboard/tca6416-keypad.c schedule_delayed_work(&chip->dwork, msecs_to_jiffies(100)); dwork 161 drivers/input/keyboard/tca6416-keypad.c cancel_delayed_work_sync(&chip->dwork); dwork 232 drivers/input/keyboard/tca6416-keypad.c INIT_DELAYED_WORK(&chip->dwork, tca6416_keys_work_func); dwork 221 drivers/input/mouse/synaptics_i2c.c struct delayed_work dwork; dwork 375 drivers/input/mouse/synaptics_i2c.c mod_delayed_work(system_wq, &touch->dwork, 0); dwork 435 drivers/input/mouse/synaptics_i2c.c container_of(work, struct synaptics_i2c, dwork.work); dwork 451 drivers/input/mouse/synaptics_i2c.c mod_delayed_work(system_wq, &touch->dwork, delay); dwork 464 drivers/input/mouse/synaptics_i2c.c mod_delayed_work(system_wq, &touch->dwork, dwork 477 drivers/input/mouse/synaptics_i2c.c cancel_delayed_work_sync(&touch->dwork); dwork 519 drivers/input/mouse/synaptics_i2c.c INIT_DELAYED_WORK(&touch->dwork, synaptics_i2c_work_handler); dwork 608 drivers/input/mouse/synaptics_i2c.c cancel_delayed_work_sync(&touch->dwork); dwork 626 drivers/input/mouse/synaptics_i2c.c mod_delayed_work(system_wq, &touch->dwork, dwork 116 drivers/input/serio/ps2-gpio.c struct delayed_work *dwork = to_delayed_work(work); dwork 117 drivers/input/serio/ps2-gpio.c struct ps2_gpio_data *drvdata = container_of(dwork, dwork 40 drivers/input/touchscreen/ili210x.c struct delayed_work dwork; dwork 176 drivers/input/touchscreen/ili210x.c dwork.work); dwork 201 drivers/input/touchscreen/ili210x.c schedule_delayed_work(&priv->dwork, dwork 209 drivers/input/touchscreen/ili210x.c schedule_delayed_work(&priv->dwork, 0); dwork 260 drivers/input/touchscreen/ili210x.c cancel_delayed_work_sync(&priv->dwork); dwork 309 drivers/input/touchscreen/ili210x.c INIT_DELAYED_WORK(&priv->dwork, ili210x_work); dwork 811 drivers/media/i2c/ad9389b.c struct delayed_work *dwork = to_delayed_work(work); dwork 813 drivers/media/i2c/ad9389b.c container_of(dwork, struct ad9389b_state, edid_handler); dwork 1490 drivers/media/i2c/adv7511-v4l2.c struct delayed_work *dwork = to_delayed_work(work); dwork 1491 drivers/media/i2c/adv7511-v4l2.c struct adv7511_state *state = container_of(dwork, struct adv7511_state, edid_handler); dwork 525 drivers/media/i2c/adv7604.c struct delayed_work *dwork = to_delayed_work(work); dwork 526 drivers/media/i2c/adv7604.c struct adv76xx_state *state = container_of(dwork, struct adv76xx_state, dwork 707 drivers/media/i2c/adv7842.c struct delayed_work *dwork = to_delayed_work(work); dwork 708 drivers/media/i2c/adv7842.c struct adv7842_state *state = container_of(dwork, dwork 358 drivers/media/i2c/tc358743.c struct delayed_work *dwork = to_delayed_work(work); dwork 359 drivers/media/i2c/tc358743.c struct tc358743_state *state = container_of(dwork, dwork 560 drivers/media/i2c/tda1997x.c struct delayed_work *dwork = to_delayed_work(work); dwork 561 drivers/media/i2c/tda1997x.c struct tda1997x_state *state = container_of(dwork, dwork 1315 drivers/media/platform/aspeed-video.c struct delayed_work *dwork = to_delayed_work(work); dwork 1316 drivers/media/platform/aspeed-video.c struct aspeed_video *video = container_of(dwork, struct aspeed_video, dwork 916 drivers/misc/cardreader/rtsx_pcr.c struct delayed_work *dwork; dwork 922 drivers/misc/cardreader/rtsx_pcr.c dwork = to_delayed_work(work); dwork 923 drivers/misc/cardreader/rtsx_pcr.c pcr = container_of(dwork, struct rtsx_pcr, carddet_work); dwork 1112 drivers/misc/cardreader/rtsx_pcr.c struct delayed_work *dwork = to_delayed_work(work); dwork 1113 drivers/misc/cardreader/rtsx_pcr.c struct rtsx_pcr *pcr = container_of(dwork, struct rtsx_pcr, idle_work); dwork 358 drivers/misc/vmw_balloon.c struct delayed_work dwork; dwork 1349 drivers/misc/vmw_balloon.c mod_delayed_work(system_freezable_wq, &b->dwork, 0); dwork 1478 drivers/misc/vmw_balloon.c struct delayed_work *dwork = to_delayed_work(work); dwork 1479 drivers/misc/vmw_balloon.c struct vmballoon *b = container_of(dwork, struct vmballoon, dwork); dwork 1514 drivers/misc/vmw_balloon.c dwork, round_jiffies_relative(HZ)); dwork 1935 drivers/misc/vmw_balloon.c INIT_DELAYED_WORK(&balloon.dwork, vmballoon_work); dwork 1958 drivers/misc/vmw_balloon.c queue_delayed_work(system_freezable_wq, &balloon.dwork, 0); dwork 1981 drivers/misc/vmw_balloon.c cancel_delayed_work_sync(&balloon.dwork); dwork 581 drivers/net/can/dev.c struct delayed_work *dwork = to_delayed_work(work); dwork 582 drivers/net/can/dev.c struct can_priv *priv = container_of(dwork, struct can_priv, dwork 63 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct delayed_work dwork; dwork 994 drivers/net/ethernet/cavium/thunder/thunder_bgx.c queue_delayed_work(lmac->check_link, &lmac->dwork, HZ * 3); dwork 1002 drivers/net/ethernet/cavium/thunder/thunder_bgx.c lmac = container_of(work, struct lmac, dwork.work); dwork 1043 drivers/net/ethernet/cavium/thunder/thunder_bgx.c queue_delayed_work(lmac->check_link, &lmac->dwork, HZ * 2); dwork 1136 drivers/net/ethernet/cavium/thunder/thunder_bgx.c INIT_DELAYED_WORK(&lmac->dwork, bgx_poll_for_link); dwork 1137 drivers/net/ethernet/cavium/thunder/thunder_bgx.c queue_delayed_work(lmac->check_link, &lmac->dwork, 0); dwork 1150 drivers/net/ethernet/cavium/thunder/thunder_bgx.c cancel_delayed_work_sync(&lmac->dwork); dwork 516 drivers/net/ethernet/freescale/fec_ptp.c struct delayed_work *dwork = to_delayed_work(work); dwork 517 drivers/net/ethernet/freescale/fec_ptp.c struct fec_enet_private *fep = container_of(dwork, struct fec_enet_private, time_keep); dwork 769 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct delayed_work *dwork; dwork 774 drivers/net/ethernet/intel/fm10k/fm10k_pci.c dwork = to_delayed_work(work); dwork 775 drivers/net/ethernet/intel/fm10k/fm10k_pci.c interface = container_of(dwork, struct fm10k_intfc, macvlan_task); dwork 832 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct delayed_work *dwork = container_of(work, struct delayed_work, dwork 834 drivers/net/ethernet/mellanox/mlx5/core/cmd.c struct mlx5_cmd_work_ent *ent = container_of(dwork, dwork 61 drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c struct delayed_work *dwork; dwork 66 drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c dwork = to_delayed_work(work); dwork 67 drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c sagent = container_of(dwork, struct mlx5e_hv_vhca_stats_agent, work); dwork 259 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h struct delayed_work *dwork, dwork 449 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c struct delayed_work *dwork, dwork 454 drivers/net/ethernet/mellanox/mlx5/core/fs_counters.c queue_delayed_work(fc_stats->wq, dwork, delay); dwork 131 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c struct delayed_work *dwork = to_delayed_work(work); dwork 132 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c struct mlx5_clock *clock = container_of(dwork, struct mlx5_clock, dwork 2075 drivers/net/ethernet/mellanox/mlxsw/core.c int mlxsw_core_schedule_dw(struct delayed_work *dwork, unsigned long delay) dwork 2077 drivers/net/ethernet/mellanox/mlxsw/core.c return queue_delayed_work(mlxsw_wq, dwork, delay); dwork 197 drivers/net/ethernet/mellanox/mlxsw/core.h int mlxsw_core_schedule_dw(struct delayed_work *dwork, unsigned long delay); dwork 239 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c struct delayed_work *dwork = to_delayed_work(work); dwork 242 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock = container_of(dwork, struct mlxsw_sp_ptp_clock, overflow_work); dwork 680 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c struct delayed_work *dwork = to_delayed_work(work); dwork 687 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c ptp_state = container_of(dwork, struct mlxsw_sp_ptp_state, ht_gc_dw); dwork 1051 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c struct delayed_work *dwork = to_delayed_work(work); dwork 1055 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c mlxsw_sp_port = container_of(dwork, struct mlxsw_sp_port, dwork 926 drivers/net/hyperv/hyperv_net.h struct delayed_work dwork; dwork 751 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, 0); dwork 1891 drivers/net/hyperv/netvsc_drv.c container_of(w, struct net_device_context, dwork.work); dwork 1902 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); dwork 1920 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, delay); dwork 1983 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); dwork 2287 drivers/net/hyperv/netvsc_drv.c INIT_DELAYED_WORK(&net_device_ctx->dwork, netvsc_link_change); dwork 2396 drivers/net/hyperv/netvsc_drv.c cancel_delayed_work_sync(&ndev_ctx->dwork); dwork 896 drivers/net/phy/phy.c struct delayed_work *dwork = to_delayed_work(work); dwork 898 drivers/net/phy/phy.c container_of(dwork, struct phy_device, state_queue); dwork 1090 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h struct delayed_work dwork; dwork 706 drivers/net/wireless/intel/iwlwifi/mvm/ops.c INIT_DELAYED_WORK(&mvm->tdls_cs.dwork, iwl_mvm_tdls_ch_switch_work); dwork 1910 drivers/net/wireless/intel/iwlwifi/mvm/sta.c cancel_delayed_work(&mvm->tdls_cs.dwork); dwork 287 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c mod_delayed_work(system_wq, &mvm->tdls_cs.dwork, dwork 511 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c mvm = container_of(work, struct iwl_mvm, tdls_cs.dwork.work); dwork 544 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c schedule_delayed_work(&mvm->tdls_cs.dwork, msecs_to_jiffies(delay)); dwork 606 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c mod_delayed_work(system_wq, &mvm->tdls_cs.dwork, dwork 661 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c flush_delayed_work(&mvm->tdls_cs.dwork); dwork 720 drivers/net/wireless/intel/iwlwifi/mvm/tdls.c mod_delayed_work(system_wq, &mvm->tdls_cs.dwork, dwork 18 drivers/net/wireless/ti/wl1251/ps.c struct delayed_work *dwork; dwork 21 drivers/net/wireless/ti/wl1251/ps.c dwork = to_delayed_work(work); dwork 22 drivers/net/wireless/ti/wl1251/ps.c wl = container_of(dwork, struct wl1251, elp_work); dwork 251 drivers/net/wireless/ti/wlcore/main.c struct delayed_work *dwork; dwork 254 drivers/net/wireless/ti/wlcore/main.c dwork = to_delayed_work(work); dwork 255 drivers/net/wireless/ti/wlcore/main.c wl = container_of(dwork, struct wl1271, tx_watchdog_work); dwork 2033 drivers/net/wireless/ti/wlcore/main.c struct delayed_work *dwork; dwork 2039 drivers/net/wireless/ti/wlcore/main.c dwork = to_delayed_work(work); dwork 2040 drivers/net/wireless/ti/wlcore/main.c wlvif = container_of(dwork, struct wl12xx_vif, channel_switch_work); dwork 2073 drivers/net/wireless/ti/wlcore/main.c struct delayed_work *dwork; dwork 2078 drivers/net/wireless/ti/wlcore/main.c dwork = to_delayed_work(work); dwork 2079 drivers/net/wireless/ti/wlcore/main.c wlvif = container_of(dwork, struct wl12xx_vif, connection_loss_work); dwork 2101 drivers/net/wireless/ti/wlcore/main.c struct delayed_work *dwork; dwork 2107 drivers/net/wireless/ti/wlcore/main.c dwork = to_delayed_work(work); dwork 2108 drivers/net/wireless/ti/wlcore/main.c wlvif = container_of(dwork, struct wl12xx_vif, dwork 5740 drivers/net/wireless/ti/wlcore/main.c struct delayed_work *dwork; dwork 5744 drivers/net/wireless/ti/wlcore/main.c dwork = to_delayed_work(work); dwork 5745 drivers/net/wireless/ti/wlcore/main.c wl = container_of(dwork, struct wl1271, roc_complete_work); dwork 22 drivers/net/wireless/ti/wlcore/scan.c struct delayed_work *dwork; dwork 30 drivers/net/wireless/ti/wlcore/scan.c dwork = to_delayed_work(work); dwork 31 drivers/net/wireless/ti/wlcore/scan.c wl = container_of(dwork, struct wl1271, scan_complete_work); dwork 405 drivers/nvdimm/bus.c cancel_delayed_work_sync(&nvdimm->dwork); dwork 494 drivers/nvdimm/dimm_devs.c INIT_DELAYED_WORK(&nvdimm->dwork, nvdimm_security_overwrite_query); dwork 47 drivers/nvdimm/nd-core.h struct delayed_work dwork; dwork 410 drivers/nvdimm/security.c queue_delayed_work(system_wq, &nvdimm->dwork, 0); dwork 443 drivers/nvdimm/security.c queue_delayed_work(system_wq, &nvdimm->dwork, tmo * HZ); dwork 466 drivers/nvdimm/security.c container_of(work, typeof(*nvdimm), dwork.work); dwork 515 drivers/power/supply/charger-manager.c struct delayed_work *dwork = to_delayed_work(work); dwork 516 drivers/power/supply/charger-manager.c struct charger_manager *cm = container_of(dwork, dwork 1135 drivers/regulator/qcom_spmi-regulator.c struct delayed_work *dwork = to_delayed_work(work); dwork 1137 drivers/regulator/qcom_spmi-regulator.c = container_of(dwork, struct spmi_regulator, ocp_work); dwork 155 drivers/s390/block/dasd_alias.c INIT_DELAYED_WORK(&lcu->ruac_data.dwork, lcu_update_work); dwork 274 drivers/s390/block/dasd_alias.c cancel_delayed_work_sync(&lcu->ruac_data.dwork); dwork 536 drivers/s390/block/dasd_alias.c ruac_data = container_of(work, struct read_uac_work_data, dwork.work); dwork 549 drivers/s390/block/dasd_alias.c if (!schedule_delayed_work(&lcu->ruac_data.dwork, 30*HZ)) dwork 597 drivers/s390/block/dasd_alias.c if (!schedule_delayed_work(&lcu->ruac_data.dwork, 0)) dwork 625 drivers/s390/block/dasd_eckd.h struct delayed_work dwork; dwork 39 drivers/staging/greybus/bootrom.c struct delayed_work dwork; dwork 54 drivers/staging/greybus/bootrom.c struct delayed_work *dwork = to_delayed_work(work); dwork 55 drivers/staging/greybus/bootrom.c struct gb_bootrom *bootrom = container_of(dwork, dwork 56 drivers/staging/greybus/bootrom.c struct gb_bootrom, dwork); dwork 93 drivers/staging/greybus/bootrom.c schedule_delayed_work(&bootrom->dwork, msecs_to_jiffies(timeout)); dwork 98 drivers/staging/greybus/bootrom.c cancel_delayed_work_sync(&bootrom->dwork); dwork 444 drivers/staging/greybus/bootrom.c INIT_DELAYED_WORK(&bootrom->dwork, gb_bootrom_timedout); dwork 29 drivers/staging/greybus/fw-download.c struct delayed_work dwork; dwork 131 drivers/staging/greybus/fw-download.c struct delayed_work *dwork = to_delayed_work(work); dwork 132 drivers/staging/greybus/fw-download.c struct fw_request *fw_req = container_of(dwork, dwork 133 drivers/staging/greybus/fw-download.c struct fw_request, dwork); dwork 209 drivers/staging/greybus/fw-download.c INIT_DELAYED_WORK(&fw_req->dwork, fw_request_timedout); dwork 210 drivers/staging/greybus/fw-download.c schedule_delayed_work(&fw_req->dwork, NEXT_REQ_TIMEOUT_J); dwork 301 drivers/staging/greybus/fw-download.c cancel_delayed_work_sync(&fw_req->dwork); dwork 344 drivers/staging/greybus/fw-download.c schedule_delayed_work(&fw_req->dwork, NEXT_REQ_TIMEOUT_J); dwork 377 drivers/staging/greybus/fw-download.c cancel_delayed_work_sync(&fw_req->dwork); dwork 459 drivers/staging/greybus/fw-download.c cancel_delayed_work_sync(&fw_req->dwork); dwork 21 drivers/staging/media/tegra-vde/dmabuf-cache.c struct delayed_work dwork; dwork 52 drivers/staging/media/tegra-vde/dmabuf-cache.c dwork.work); dwork 79 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) dwork 132 drivers/staging/media/tegra-vde/dmabuf-cache.c INIT_DELAYED_WORK(&entry->dwork, tegra_vde_delayed_unmap); dwork 179 drivers/staging/media/tegra-vde/dmabuf-cache.c schedule_delayed_work(&entry->dwork, 5 * HZ); dwork 197 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) dwork 214 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) dwork 40 drivers/staging/most/i2c/i2c.c struct delayed_work dwork; dwork 132 drivers/staging/most/i2c/i2c.c cancel_delayed_work_sync(&dev->rx.dwork); dwork 135 drivers/staging/most/i2c/i2c.c pending_rx_work(&dev->rx.dwork.work); dwork 176 drivers/staging/most/i2c/i2c.c cancel_delayed_work_sync(&dev->rx.dwork); dwork 235 drivers/staging/most/i2c/i2c.c struct hdm_i2c *dev = container_of(work, struct hdm_i2c, rx.dwork.work); dwork 243 drivers/staging/most/i2c/i2c.c schedule_delayed_work(&dev->rx.dwork, dev->rx.delay); dwork 274 drivers/staging/most/i2c/i2c.c schedule_delayed_work(&dev->rx.dwork, 0); dwork 321 drivers/staging/most/i2c/i2c.c INIT_DELAYED_WORK(&dev->rx.dwork, pending_rx_work); dwork 474 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct delayed_work *dwork = to_delayed_work(work); dwork 475 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct ieee80211_device *ieee = container_of(dwork, struct ieee80211_device, softmac_scan_wq); dwork 2205 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct delayed_work *dwork = to_delayed_work(work); dwork 2206 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct ieee80211_device *ieee = container_of(dwork, struct ieee80211_device, start_ibss_wq); dwork 2369 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct delayed_work *dwork = to_delayed_work(work); dwork 2370 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct ieee80211_device *ieee = container_of(dwork, struct ieee80211_device, associate_retry_wq); dwork 3271 drivers/staging/rtl8192u/r8192U_core.c struct delayed_work *dwork = to_delayed_work(work); dwork 3272 drivers/staging/rtl8192u/r8192U_core.c struct r8192_priv *priv = container_of(dwork, dwork 758 drivers/staging/rtl8192u/r8192U_dm.c struct delayed_work *dwork = to_delayed_work(work); dwork 759 drivers/staging/rtl8192u/r8192U_dm.c struct r8192_priv *priv = container_of(dwork, struct r8192_priv, txpower_tracking_wq); dwork 2320 drivers/staging/rtl8192u/r8192U_dm.c struct delayed_work *dwork = to_delayed_work(work); dwork 2321 drivers/staging/rtl8192u/r8192U_dm.c struct r8192_priv *priv = container_of(dwork, struct r8192_priv, rfpath_check_wq); dwork 1642 drivers/staging/rtl8192u/r819xU_phy.c struct delayed_work *dwork = to_delayed_work(work); dwork 1643 drivers/staging/rtl8192u/r819xU_phy.c struct r8192_priv *priv = container_of(dwork, struct r8192_priv, dwork 539 drivers/usb/class/cdc-acm.c schedule_delayed_work(&acm->dwork, HZ / 2); dwork 1358 drivers/usb/class/cdc-acm.c INIT_DELAYED_WORK(&acm->dwork, acm_softint); dwork 1568 drivers/usb/class/cdc-acm.c cancel_delayed_work_sync(&acm->dwork); dwork 1611 drivers/usb/class/cdc-acm.c cancel_delayed_work_sync(&acm->dwork); dwork 116 drivers/usb/class/cdc-acm.h struct delayed_work dwork; /* for cool downs needed in error recovery */ dwork 212 drivers/video/fbdev/hyperv_fb.c struct delayed_work dwork; dwork 372 drivers/video/fbdev/hyperv_fb.c schedule_delayed_work(&par->dwork, HVFB_UPDATE_DELAY); dwork 523 drivers/video/fbdev/hyperv_fb.c struct hvfb_par *par = container_of(w, struct hvfb_par, dwork.work); dwork 530 drivers/video/fbdev/hyperv_fb.c schedule_delayed_work(&par->dwork, HVFB_UPDATE_DELAY); dwork 772 drivers/video/fbdev/hyperv_fb.c INIT_DELAYED_WORK(&par->dwork, hvfb_update_work); dwork 851 drivers/video/fbdev/hyperv_fb.c cancel_delayed_work_sync(&par->dwork); dwork 870 drivers/video/fbdev/hyperv_fb.c cancel_delayed_work_sync(&par->dwork); dwork 156 fs/fs-writeback.c mod_delayed_work(bdi_wq, &wb->dwork, 0); dwork 188 fs/fs-writeback.c mod_delayed_work(bdi_wq, &wb->dwork, 0); dwork 2063 fs/fs-writeback.c struct bdi_writeback, dwork); dwork 1154 fs/jffs2/wbuf.c struct delayed_work *dwork; dwork 1156 fs/jffs2/wbuf.c dwork = to_delayed_work(work); dwork 1157 fs/jffs2/wbuf.c return container_of(dwork, struct jffs2_sb_info, wbuf_dwork); dwork 101 fs/lockd/svc.c struct delayed_work *dwork = to_delayed_work(grace); dwork 102 fs/lockd/svc.c struct lockd_net *ln = container_of(dwork, struct lockd_net, dwork 5336 fs/nfsd/nfs4state.c struct delayed_work *dwork = to_delayed_work(laundry); dwork 5337 fs/nfsd/nfs4state.c struct nfsd_net *nn = container_of(dwork, struct nfsd_net, dwork 167 include/linux/backing-dev-defs.h struct delayed_work dwork; /* work item used for writeback */ dwork 1501 include/linux/blkdev.h int kblockd_mod_delayed_work_on(int cpu, struct delayed_work *dwork, unsigned long delay); dwork 119 include/linux/kthread.h #define KTHREAD_DELAYED_WORK_INIT(dwork, fn) { \ dwork 120 include/linux/kthread.h .work = KTHREAD_WORK_INIT((dwork).work, (fn)), \ dwork 131 include/linux/kthread.h #define DEFINE_KTHREAD_DELAYED_WORK(dwork, fn) \ dwork 132 include/linux/kthread.h struct kthread_delayed_work dwork = \ dwork 133 include/linux/kthread.h KTHREAD_DELAYED_WORK_INIT(dwork, fn) dwork 164 include/linux/kthread.h #define kthread_init_delayed_work(dwork, fn) \ dwork 166 include/linux/kthread.h kthread_init_work(&(dwork)->work, (fn)); \ dwork 167 include/linux/kthread.h timer_setup(&(dwork)->timer, \ dwork 185 include/linux/kthread.h struct kthread_delayed_work *dwork, dwork 189 include/linux/kthread.h struct kthread_delayed_work *dwork, dwork 1852 include/linux/libata.h extern void ata_sff_queue_delayed_work(struct delayed_work *dwork, dwork 188 include/linux/sunrpc/sched.h struct delayed_work dwork; dwork 451 include/linux/workqueue.h struct delayed_work *dwork, unsigned long delay); dwork 464 include/linux/workqueue.h extern bool flush_delayed_work(struct delayed_work *dwork); dwork 465 include/linux/workqueue.h extern bool cancel_delayed_work(struct delayed_work *dwork); dwork 466 include/linux/workqueue.h extern bool cancel_delayed_work_sync(struct delayed_work *dwork); dwork 506 include/linux/workqueue.h struct delayed_work *dwork, dwork 509 include/linux/workqueue.h return queue_delayed_work_on(WORK_CPU_UNBOUND, wq, dwork, delay); dwork 521 include/linux/workqueue.h struct delayed_work *dwork, dwork 524 include/linux/workqueue.h return mod_delayed_work_on(WORK_CPU_UNBOUND, wq, dwork, delay); dwork 593 include/linux/workqueue.h static inline bool schedule_delayed_work_on(int cpu, struct delayed_work *dwork, dwork 596 include/linux/workqueue.h return queue_delayed_work_on(cpu, system_wq, dwork, delay); dwork 607 include/linux/workqueue.h static inline bool schedule_delayed_work(struct delayed_work *dwork, dwork 610 include/linux/workqueue.h return queue_delayed_work(system_wq, dwork, delay); dwork 5359 include/net/mac80211.h struct delayed_work *dwork, dwork 215 kernel/cgroup/cgroup-v1.c struct delayed_work *dwork = to_delayed_work(work); dwork 216 kernel/cgroup/cgroup-v1.c struct cgroup_pidlist *l = container_of(dwork, struct cgroup_pidlist, dwork 226 kernel/cgroup/cgroup-v1.c if (!delayed_work_pending(dwork)) { dwork 844 kernel/kthread.c struct kthread_delayed_work *dwork = from_timer(dwork, t, timer); dwork 845 kernel/kthread.c struct kthread_work *work = &dwork->work; dwork 870 kernel/kthread.c struct kthread_delayed_work *dwork, dwork 873 kernel/kthread.c struct timer_list *timer = &dwork->timer; dwork 874 kernel/kthread.c struct kthread_work *work = &dwork->work; dwork 914 kernel/kthread.c struct kthread_delayed_work *dwork, dwork 917 kernel/kthread.c struct kthread_work *work = &dwork->work; dwork 924 kernel/kthread.c __kthread_queue_delayed_work(worker, dwork, delay); dwork 998 kernel/kthread.c struct kthread_delayed_work *dwork = dwork 1010 kernel/kthread.c del_timer_sync(&dwork->timer); dwork 1051 kernel/kthread.c struct kthread_delayed_work *dwork, dwork 1054 kernel/kthread.c struct kthread_work *work = &dwork->work; dwork 1073 kernel/kthread.c __kthread_queue_delayed_work(worker, dwork, delay); dwork 1145 kernel/kthread.c bool kthread_cancel_delayed_work_sync(struct kthread_delayed_work *dwork) dwork 1147 kernel/kthread.c return __kthread_cancel_work_sync(&dwork->work, true); dwork 3650 kernel/sched/core.c struct delayed_work *dwork = to_delayed_work(work); dwork 3651 kernel/sched/core.c struct tick_work *twork = container_of(dwork, struct tick_work, work); dwork 3700 kernel/sched/core.c queue_delayed_work(system_unbound_wq, dwork, HZ); dwork 412 kernel/sched/psi.c struct delayed_work *dwork; dwork 418 kernel/sched/psi.c dwork = to_delayed_work(work); dwork 419 kernel/sched/psi.c group = container_of(dwork, struct psi_group, avgs_work); dwork 438 kernel/sched/psi.c schedule_delayed_work(dwork, nsecs_to_jiffies( dwork 581 kernel/sched/psi.c struct kthread_delayed_work *dwork; dwork 586 kernel/sched/psi.c dwork = container_of(work, struct kthread_delayed_work, work); dwork 587 kernel/sched/psi.c group = container_of(dwork, struct psi_group, poll_work); dwork 1245 kernel/workqueue.c struct delayed_work *dwork = to_delayed_work(work); dwork 1252 kernel/workqueue.c if (likely(del_timer(&dwork->timer))) dwork 1616 kernel/workqueue.c struct delayed_work *dwork = from_timer(dwork, t, timer); dwork 1619 kernel/workqueue.c __queue_work(dwork->cpu, dwork->wq, &dwork->work); dwork 1624 kernel/workqueue.c struct delayed_work *dwork, unsigned long delay) dwork 1626 kernel/workqueue.c struct timer_list *timer = &dwork->timer; dwork 1627 kernel/workqueue.c struct work_struct *work = &dwork->work; dwork 1641 kernel/workqueue.c __queue_work(cpu, wq, &dwork->work); dwork 1645 kernel/workqueue.c dwork->wq = wq; dwork 1646 kernel/workqueue.c dwork->cpu = cpu; dwork 1667 kernel/workqueue.c struct delayed_work *dwork, unsigned long delay) dwork 1669 kernel/workqueue.c struct work_struct *work = &dwork->work; dwork 1677 kernel/workqueue.c __queue_delayed_work(cpu, wq, dwork, delay); dwork 1705 kernel/workqueue.c struct delayed_work *dwork, unsigned long delay) dwork 1711 kernel/workqueue.c ret = try_to_grab_pending(&dwork->work, true, &flags); dwork 1715 kernel/workqueue.c __queue_delayed_work(cpu, wq, dwork, delay); dwork 3188 kernel/workqueue.c bool flush_delayed_work(struct delayed_work *dwork) dwork 3191 kernel/workqueue.c if (del_timer_sync(&dwork->timer)) dwork 3192 kernel/workqueue.c __queue_work(dwork->cpu, dwork->wq, &dwork->work); dwork 3194 kernel/workqueue.c return flush_work(&dwork->work); dwork 3251 kernel/workqueue.c bool cancel_delayed_work(struct delayed_work *dwork) dwork 3253 kernel/workqueue.c return __cancel_work(&dwork->work, true); dwork 3266 kernel/workqueue.c bool cancel_delayed_work_sync(struct delayed_work *dwork) dwork 3268 kernel/workqueue.c return __cancel_work_timer(&dwork->work, true); dwork 275 mm/backing-dev.c queue_delayed_work(bdi_wq, &wb->dwork, timeout); dwork 309 mm/backing-dev.c INIT_DELAYED_WORK(&wb->dwork, wb_workfn); dwork 363 mm/backing-dev.c mod_delayed_work(bdi_wq, &wb->dwork, 0); dwork 364 mm/backing-dev.c flush_delayed_work(&wb->dwork); dwork 372 mm/backing-dev.c WARN_ON(delayed_work_pending(&wb->dwork)); dwork 882 net/mac80211/util.c struct delayed_work *dwork, dwork 890 net/mac80211/util.c queue_delayed_work(local->workqueue, dwork, delay); dwork 100 net/netfilter/ipset/ip_set_hash_gen.h struct delayed_work dwork; dwork 482 net/netfilter/ipset/ip_set_hash_gen.h cancel_delayed_work_sync(&h->gc.dwork); dwork 595 net/netfilter/ipset/ip_set_hash_gen.h gc = container_of(work, struct htable_gc, dwork.work); dwork 619 net/netfilter/ipset/ip_set_hash_gen.h queue_delayed_work(system_power_efficient_wq, &gc->dwork, next_run); dwork 626 net/netfilter/ipset/ip_set_hash_gen.h INIT_DEFERRABLE_WORK(&gc->dwork, mtype_gc); dwork 627 net/netfilter/ipset/ip_set_hash_gen.h queue_delayed_work(system_power_efficient_wq, &gc->dwork, HZ); dwork 68 net/netfilter/nf_conntrack_core.c struct delayed_work dwork; dwork 1230 net/netfilter/nf_conntrack_core.c gc_work = container_of(work, struct conntrack_gc_work, dwork.work); dwork 1334 net/netfilter/nf_conntrack_core.c queue_delayed_work(system_power_efficient_wq, &gc_work->dwork, next_run); dwork 1339 net/netfilter/nf_conntrack_core.c INIT_DEFERRABLE_WORK(&gc_work->dwork, gc_worker); dwork 2247 net/netfilter/nf_conntrack_core.c cancel_delayed_work_sync(&conntrack_gc_work.dwork); dwork 2524 net/netfilter/nf_conntrack_core.c queue_delayed_work(system_power_efficient_wq, &conntrack_gc_work.dwork, HZ); dwork 92 net/sunrpc/sched.c cancel_delayed_work(&queue->timer_list.dwork); dwork 104 net/sunrpc/sched.c mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires); dwork 263 net/sunrpc/sched.c INIT_DELAYED_WORK(&queue->timer_list.dwork, __rpc_queue_timer_fn); dwork 282 net/sunrpc/sched.c cancel_delayed_work_sync(&queue->timer_list.dwork); dwork 757 net/sunrpc/sched.c timer_list.dwork.work); dwork 169 sound/firewire/bebob/bebob.c container_of(work, struct snd_bebob, dwork.work); dwork 285 sound/firewire/bebob/bebob.c INIT_DEFERRABLE_WORK(&bebob->dwork, do_registration); dwork 290 sound/firewire/bebob/bebob.c snd_fw_schedule_registration(unit, &bebob->dwork); dwork 335 sound/firewire/bebob/bebob.c snd_fw_schedule_registration(unit, &bebob->dwork); dwork 352 sound/firewire/bebob/bebob.c cancel_delayed_work_sync(&bebob->dwork); dwork 87 sound/firewire/bebob/bebob.h struct delayed_work dwork; dwork 137 sound/firewire/dice/dice.c struct snd_dice *dice = container_of(work, struct snd_dice, dwork.work); dwork 227 sound/firewire/dice/dice.c INIT_DEFERRABLE_WORK(&dice->dwork, do_registration); dwork 228 sound/firewire/dice/dice.c snd_fw_schedule_registration(unit, &dice->dwork); dwork 242 sound/firewire/dice/dice.c cancel_delayed_work_sync(&dice->dwork); dwork 259 sound/firewire/dice/dice.c snd_fw_schedule_registration(unit, &dice->dwork); dwork 82 sound/firewire/dice/dice.h struct delayed_work dwork; dwork 54 sound/firewire/digi00x/digi00x.c container_of(work, struct snd_dg00x, dwork.work); dwork 127 sound/firewire/digi00x/digi00x.c INIT_DEFERRABLE_WORK(&dg00x->dwork, do_registration); dwork 128 sound/firewire/digi00x/digi00x.c snd_fw_schedule_registration(unit, &dg00x->dwork); dwork 139 sound/firewire/digi00x/digi00x.c snd_fw_schedule_registration(unit, &dg00x->dwork); dwork 163 sound/firewire/digi00x/digi00x.c cancel_delayed_work_sync(&dg00x->dwork); dwork 41 sound/firewire/digi00x/digi00x.h struct delayed_work dwork; dwork 39 sound/firewire/fireface/ff.c struct snd_ff *ff = container_of(work, struct snd_ff, dwork.work); dwork 107 sound/firewire/fireface/ff.c INIT_DEFERRABLE_WORK(&ff->dwork, do_registration); dwork 108 sound/firewire/fireface/ff.c snd_fw_schedule_registration(unit, &ff->dwork); dwork 119 sound/firewire/fireface/ff.c snd_fw_schedule_registration(unit, &ff->dwork); dwork 136 sound/firewire/fireface/ff.c cancel_work_sync(&ff->dwork.work); dwork 67 sound/firewire/fireface/ff.h struct delayed_work dwork; dwork 202 sound/firewire/fireworks/fireworks.c struct snd_efw *efw = container_of(work, struct snd_efw, dwork.work); dwork 297 sound/firewire/fireworks/fireworks.c INIT_DEFERRABLE_WORK(&efw->dwork, do_registration); dwork 298 sound/firewire/fireworks/fireworks.c snd_fw_schedule_registration(unit, &efw->dwork); dwork 309 sound/firewire/fireworks/fireworks.c snd_fw_schedule_registration(unit, &efw->dwork); dwork 333 sound/firewire/fireworks/fireworks.c cancel_delayed_work_sync(&efw->dwork); dwork 69 sound/firewire/fireworks/fireworks.h struct delayed_work dwork; dwork 85 sound/firewire/lib.c struct delayed_work *dwork) dwork 98 sound/firewire/lib.c mod_delayed_work(system_wq, dwork, delay); dwork 27 sound/firewire/lib.h struct delayed_work *dwork); dwork 64 sound/firewire/motu/motu.c struct snd_motu *motu = container_of(work, struct snd_motu, dwork.work); dwork 137 sound/firewire/motu/motu.c INIT_DEFERRABLE_WORK(&motu->dwork, do_registration); dwork 138 sound/firewire/motu/motu.c snd_fw_schedule_registration(unit, &motu->dwork); dwork 152 sound/firewire/motu/motu.c cancel_delayed_work_sync(&motu->dwork); dwork 169 sound/firewire/motu/motu.c snd_fw_schedule_registration(unit, &motu->dwork); dwork 50 sound/firewire/motu/motu.h struct delayed_work dwork; dwork 184 sound/firewire/oxfw/oxfw.c struct snd_oxfw *oxfw = container_of(work, struct snd_oxfw, dwork.work); dwork 261 sound/firewire/oxfw/oxfw.c INIT_DEFERRABLE_WORK(&oxfw->dwork, do_registration); dwork 262 sound/firewire/oxfw/oxfw.c snd_fw_schedule_registration(unit, &oxfw->dwork); dwork 272 sound/firewire/oxfw/oxfw.c snd_fw_schedule_registration(unit, &oxfw->dwork); dwork 295 sound/firewire/oxfw/oxfw.c cancel_delayed_work_sync(&oxfw->dwork); dwork 44 sound/firewire/oxfw/oxfw.h struct delayed_work dwork; dwork 100 sound/firewire/tascam/tascam.c struct snd_tscm *tscm = container_of(work, struct snd_tscm, dwork.work); dwork 166 sound/firewire/tascam/tascam.c INIT_DEFERRABLE_WORK(&tscm->dwork, do_registration); dwork 167 sound/firewire/tascam/tascam.c snd_fw_schedule_registration(unit, &tscm->dwork); dwork 178 sound/firewire/tascam/tascam.c snd_fw_schedule_registration(unit, &tscm->dwork); dwork 202 sound/firewire/tascam/tascam.c cancel_delayed_work_sync(&tscm->dwork); dwork 74 sound/firewire/tascam/tascam.h struct delayed_work dwork; dwork 343 sound/soc/intel/skylake/skl-messages.c struct delayed_work *dwork; dwork 348 sound/soc/intel/skylake/skl-messages.c dwork = &skl->d0i3.work; dwork 350 sound/soc/intel/skylake/skl-messages.c if (dwork->work.func) { dwork 352 sound/soc/intel/skylake/skl-messages.c flush_delayed_work(dwork); dwork 354 sound/soc/intel/skylake/skl-messages.c cancel_delayed_work_sync(dwork); dwork 69 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h struct delayed_work *dwork,