tx_work 160 drivers/bluetooth/btmtksdio.c struct work_struct tx_work; tx_work 306 drivers/bluetooth/btmtksdio.c tx_work); tx_work 520 drivers/bluetooth/btmtksdio.c schedule_work(&bdev->tx_work); tx_work 653 drivers/bluetooth/btmtksdio.c cancel_work_sync(&bdev->tx_work); tx_work 920 drivers/bluetooth/btmtksdio.c schedule_work(&bdev->tx_work); tx_work 943 drivers/bluetooth/btmtksdio.c INIT_WORK(&bdev->tx_work, btmtksdio_tx_work); tx_work 130 drivers/bluetooth/btmtkuart.c struct work_struct tx_work; tx_work 371 drivers/bluetooth/btmtkuart.c tx_work); tx_work 422 drivers/bluetooth/btmtkuart.c schedule_work(&bdev->tx_work); tx_work 629 drivers/bluetooth/btmtkuart.c cancel_work_sync(&bdev->tx_work); tx_work 989 drivers/bluetooth/btmtkuart.c INIT_WORK(&bdev->tx_work, btmtkuart_tx_work); tx_work 3604 drivers/gpu/drm/drm_dp_mst_topology.c queue_work(system_long_wq, &mgr->tx_work); tx_work 3731 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_topology_mgr *mgr = container_of(work, struct drm_dp_mst_topology_mgr, tx_work); tx_work 3962 drivers/gpu/drm/drm_dp_mst_topology.c INIT_WORK(&mgr->tx_work, drm_dp_tx_work); tx_work 661 drivers/infiniband/sw/siw/siw_qp_rx.c struct siw_wqe *tx_work = tx_wqe(qp); tx_work 682 drivers/infiniband/sw/siw/siw_qp_rx.c if (tx_work->wr_status == SIW_WR_IDLE) { tx_work 687 drivers/infiniband/sw/siw/siw_qp_rx.c tx_work->processed = 0; tx_work 688 drivers/infiniband/sw/siw/siw_qp_rx.c tx_work->mem[0] = NULL; tx_work 689 drivers/infiniband/sw/siw/siw_qp_rx.c tx_work->wr_status = SIW_WR_QUEUED; tx_work 690 drivers/infiniband/sw/siw/siw_qp_rx.c resp = &tx_work->sqe; tx_work 61 drivers/input/serio/ps2-gpio.c struct delayed_work tx_work; tx_work 76 drivers/input/serio/ps2-gpio.c flush_delayed_work(&drvdata->tx_work); tx_work 90 drivers/input/serio/ps2-gpio.c schedule_delayed_work(&drvdata->tx_work, usecs_to_jiffies(200)); tx_work 119 drivers/input/serio/ps2-gpio.c tx_work); tx_work 408 drivers/input/serio/ps2-gpio.c INIT_DELAYED_WORK(&drvdata->tx_work, ps2_gpio_tx_work_fn); tx_work 1534 drivers/net/can/m_can/m_can.c tx_work); tx_work 1564 drivers/net/can/m_can/m_can.c queue_work(cdev->tx_wq, &cdev->tx_work); tx_work 1600 drivers/net/can/m_can/m_can.c INIT_WORK(&cdev->tx_work, m_can_tx_work_queue); tx_work 81 drivers/net/can/m_can/m_can.h struct work_struct tx_work; tx_work 87 drivers/net/can/slcan.c struct work_struct tx_work; /* Flushes transmit buffer */ tx_work 313 drivers/net/can/slcan.c struct slcan *sl = container_of(work, struct slcan, tx_work); tx_work 352 drivers/net/can/slcan.c schedule_work(&sl->tx_work); tx_work 550 drivers/net/can/slcan.c INIT_WORK(&sl->tx_work, slcan_transmit); tx_work 660 drivers/net/can/slcan.c flush_work(&sl->tx_work); tx_work 159 drivers/net/can/spi/hi311x.c struct work_struct tx_work; tx_work 381 drivers/net/can/spi/hi311x.c queue_work(priv->wq, &priv->tx_work); tx_work 576 drivers/net/can/spi/hi311x.c tx_work); tx_work 772 drivers/net/can/spi/hi311x.c INIT_WORK(&priv->tx_work, hi3110_tx_work_handler); tx_work 216 drivers/net/can/spi/mcp251x.c struct work_struct tx_work; tx_work 476 drivers/net/can/spi/mcp251x.c queue_work(priv->wq, &priv->tx_work); tx_work 686 drivers/net/can/spi/mcp251x.c tx_work); tx_work 913 drivers/net/can/spi/mcp251x.c INIT_WORK(&priv->tx_work, mcp251x_tx_work_handler); tx_work 124 drivers/net/ethernet/micrel/ks8851.c struct work_struct tx_work; tx_work 758 drivers/net/ethernet/micrel/ks8851.c struct ks8851_net *ks = container_of(work, struct ks8851_net, tx_work); tx_work 891 drivers/net/ethernet/micrel/ks8851.c flush_work(&ks->tx_work); tx_work 954 drivers/net/ethernet/micrel/ks8851.c schedule_work(&ks->tx_work); tx_work 1485 drivers/net/ethernet/micrel/ks8851.c INIT_WORK(&ks->tx_work, ks8851_tx_work); tx_work 60 drivers/net/ethernet/microchip/enc28j60.c struct work_struct tx_work; tx_work 1298 drivers/net/ethernet/microchip/enc28j60.c schedule_work(&priv->tx_work); tx_work 1306 drivers/net/ethernet/microchip/enc28j60.c container_of(work, struct enc28j60_net, tx_work); tx_work 1561 drivers/net/ethernet/microchip/enc28j60.c INIT_WORK(&priv->tx_work, enc28j60_tx_work_handler); tx_work 54 drivers/net/ethernet/microchip/encx24j600.c struct kthread_work tx_work; tx_work 870 drivers/net/ethernet/microchip/encx24j600.c container_of(ws, struct encx24j600_priv, tx_work); tx_work 889 drivers/net/ethernet/microchip/encx24j600.c kthread_queue_work(&priv->kworker, &priv->tx_work); tx_work 1053 drivers/net/ethernet/microchip/encx24j600.c kthread_init_work(&priv->tx_work, encx24j600_tx_proc); tx_work 2576 drivers/net/ethernet/nvidia/forcedeth.c int tx_work = 0; tx_work 2582 drivers/net/ethernet/nvidia/forcedeth.c (tx_work < limit)) { tx_work 2604 drivers/net/ethernet/nvidia/forcedeth.c tx_work++; tx_work 2624 drivers/net/ethernet/nvidia/forcedeth.c tx_work++; tx_work 2633 drivers/net/ethernet/nvidia/forcedeth.c netdev_completed_queue(np->dev, tx_work, bytes_compl); tx_work 2639 drivers/net/ethernet/nvidia/forcedeth.c return tx_work; tx_work 2646 drivers/net/ethernet/nvidia/forcedeth.c int tx_work = 0; tx_work 2652 drivers/net/ethernet/nvidia/forcedeth.c (tx_work < limit)) { tx_work 2678 drivers/net/ethernet/nvidia/forcedeth.c tx_work++; tx_work 2690 drivers/net/ethernet/nvidia/forcedeth.c netdev_completed_queue(np->dev, tx_work, bytes_cleaned); tx_work 2696 drivers/net/ethernet/nvidia/forcedeth.c return tx_work; tx_work 3750 drivers/net/ethernet/nvidia/forcedeth.c int rx_count, tx_work = 0, rx_work = 0; tx_work 3755 drivers/net/ethernet/nvidia/forcedeth.c tx_work += nv_tx_done(dev, np->tx_ring_size); tx_work 3762 drivers/net/ethernet/nvidia/forcedeth.c tx_work += nv_tx_done_optimized(dev, np->tx_ring_size); tx_work 3779 drivers/net/ethernet/nvidia/forcedeth.c nv_change_interrupt_mode(dev, tx_work + rx_work); tx_work 51 drivers/net/ethernet/qualcomm/qca_uart.c struct work_struct tx_work; /* Flushes transmit buffer */ tx_work 129 drivers/net/ethernet/qualcomm/qca_uart.c struct qcauart *qca = container_of(work, struct qcauart, tx_work); tx_work 167 drivers/net/ethernet/qualcomm/qca_uart.c schedule_work(&qca->tx_work); tx_work 189 drivers/net/ethernet/qualcomm/qca_uart.c flush_work(&qca->tx_work); tx_work 347 drivers/net/ethernet/qualcomm/qca_uart.c INIT_WORK(&qca->tx_work, qcauart_transmit); tx_work 383 drivers/net/ethernet/qualcomm/qca_uart.c cancel_work_sync(&qca->tx_work); tx_work 402 drivers/net/ethernet/qualcomm/qca_uart.c cancel_work_sync(&qca->tx_work); tx_work 170 drivers/net/ethernet/wiznet/w5100.c struct work_struct tx_work; tx_work 821 drivers/net/ethernet/wiznet/w5100.c tx_work); tx_work 840 drivers/net/ethernet/wiznet/w5100.c queue_work(priv->xfer_wq, &priv->tx_work); tx_work 1156 drivers/net/ethernet/wiznet/w5100.c INIT_WORK(&priv->tx_work, w5100_tx_work); tx_work 423 drivers/net/slip/slip.c struct slip *sl = container_of(work, struct slip, tx_work); tx_work 462 drivers/net/slip/slip.c schedule_work(&sl->tx_work); tx_work 769 drivers/net/slip/slip.c INIT_WORK(&sl->tx_work, slip_transmit); tx_work 900 drivers/net/slip/slip.c flush_work(&sl->tx_work); tx_work 57 drivers/net/slip/slip.h struct work_struct tx_work; /* Flushes transmit buffer */ tx_work 785 drivers/net/wireless/ath/ar5523/ar5523.c ieee80211_queue_work(ar->hw, &ar->tx_work); tx_work 885 drivers/net/wireless/ath/ar5523/ar5523.c struct ar5523 *ar = container_of(work, struct ar5523, tx_work); tx_work 1605 drivers/net/wireless/ath/ar5523/ar5523.c INIT_WORK(&ar->tx_work, ar5523_tx_work); tx_work 101 drivers/net/wireless/ath/ar5523/ar5523.h struct work_struct tx_work; tx_work 958 drivers/net/wireless/broadcom/b43/b43.h struct work_struct tx_work; tx_work 1579 drivers/net/wireless/broadcom/b43/dma.c ieee80211_queue_work(dev->wl->hw, &dev->wl->tx_work); tx_work 3582 drivers/net/wireless/broadcom/b43/main.c struct b43_wl *wl = container_of(work, struct b43_wl, tx_work); tx_work 3638 drivers/net/wireless/broadcom/b43/main.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 4336 drivers/net/wireless/broadcom/b43/main.c cancel_work_sync(&wl->tx_work); tx_work 5601 drivers/net/wireless/broadcom/b43/main.c INIT_WORK(&wl->tx_work, b43_tx_work); tx_work 631 drivers/net/wireless/broadcom/b43legacy/b43legacy.h struct work_struct tx_work; tx_work 1238 drivers/net/wireless/broadcom/b43legacy/dma.c ieee80211_queue_work(dev->wl->hw, &dev->wl->tx_work); tx_work 2448 drivers/net/wireless/broadcom/b43legacy/main.c tx_work); tx_work 2501 drivers/net/wireless/broadcom/b43legacy/main.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 2936 drivers/net/wireless/broadcom/b43legacy/main.c cancel_work_sync(&wl->tx_work); tx_work 3831 drivers/net/wireless/broadcom/b43legacy/main.c INIT_WORK(&wl->tx_work, b43legacy_tx_work); tx_work 195 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h struct work_struct tx_work; tx_work 189 drivers/net/wireless/marvell/libertas_tf/main.c queue_work(lbtf_wq, &priv->tx_work); tx_work 200 drivers/net/wireless/marvell/libertas_tf/main.c tx_work); tx_work 290 drivers/net/wireless/marvell/libertas_tf/main.c cancel_work_sync(&priv->tx_work); tx_work 595 drivers/net/wireless/marvell/libertas_tf/main.c INIT_WORK(&priv->tx_work, lbtf_tx_work); tx_work 672 drivers/net/wireless/marvell/libertas_tf/main.c queue_work(lbtf_wq, &priv->tx_work); tx_work 692 drivers/net/wireless/marvell/libertas_tf/main.c queue_work(lbtf_wq, &priv->tx_work); tx_work 352 drivers/net/wireless/ti/wl1251/main.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 446 drivers/net/wireless/ti/wl1251/main.c cancel_work_sync(&wl->tx_work); tx_work 1621 drivers/net/wireless/ti/wl1251/main.c INIT_WORK(&wl->tx_work, wl1251_tx_work); tx_work 332 drivers/net/wireless/ti/wl1251/tx.c struct wl1251 *wl = container_of(work, struct wl1251, tx_work); tx_work 481 drivers/net/wireless/ti/wl1251/tx.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 308 drivers/net/wireless/ti/wl1251/wl1251.h struct work_struct tx_work; tx_work 677 drivers/net/wireless/ti/wlcore/main.c cancel_work_sync(&wl->tx_work); tx_work 690 drivers/net/wireless/ti/wlcore/main.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 1269 drivers/net/wireless/ti/wlcore/main.c ieee80211_queue_work(wl->hw, &wl->tx_work); tx_work 1779 drivers/net/wireless/ti/wlcore/main.c flush_work(&wl->tx_work); tx_work 1942 drivers/net/wireless/ti/wlcore/main.c cancel_work_sync(&wl->tx_work); tx_work 6392 drivers/net/wireless/ti/wlcore/main.c INIT_WORK(&wl->tx_work, wl1271_tx_work); tx_work 854 drivers/net/wireless/ti/wlcore/tx.c struct wl1271 *wl = container_of(work, struct wl1271, tx_work); tx_work 1160 drivers/net/wireless/ti/wlcore/tx.c wl1271_tx_work(&wl->tx_work); tx_work 271 drivers/net/wireless/ti/wlcore/wlcore.h struct work_struct tx_work; tx_work 114 drivers/nfc/st21nfca/dep.c dep_info.tx_work); tx_work 137 drivers/nfc/st21nfca/dep.c schedule_work(&info->dep_info.tx_work); tx_work 669 drivers/nfc/st21nfca/dep.c INIT_WORK(&info->dep_info.tx_work, st21nfca_tx_work); tx_work 680 drivers/nfc/st21nfca/dep.c cancel_work_sync(&info->dep_info.tx_work); tx_work 117 drivers/nfc/st21nfca/st21nfca.h struct work_struct tx_work; tx_work 70 drivers/staging/greybus/uart.c struct work_struct tx_work; tx_work 194 drivers/staging/greybus/uart.c schedule_work(&gb_tty->tx_work); tx_work 241 drivers/staging/greybus/uart.c gb_tty = container_of(work, struct gb_tty, tx_work); tx_work 449 drivers/staging/greybus/uart.c schedule_work(&gb_tty->tx_work); tx_work 768 drivers/staging/greybus/uart.c cancel_work_sync(&gb_tty->tx_work); tx_work 855 drivers/staging/greybus/uart.c INIT_WORK(&gb_tty->tx_work, gb_uart_tx_write_work); tx_work 273 drivers/staging/nvec/nvec.c schedule_work(&nvec->tx_work); tx_work 375 drivers/staging/nvec/nvec.c struct nvec_chip *nvec = container_of(work, struct nvec_chip, tx_work); tx_work 826 drivers/staging/nvec/nvec.c INIT_WORK(&nvec->tx_work, nvec_request_master); tx_work 892 drivers/staging/nvec/nvec.c cancel_work_sync(&nvec->tx_work); tx_work 144 drivers/staging/nvec/nvec.h struct work_struct rx_work, tx_work; tx_work 1402 drivers/staging/qlge/qlge.h struct delayed_work tx_work; tx_work 258 drivers/tty/serial/max310x.c struct work_struct tx_work; tx_work 794 drivers/tty/serial/max310x.c schedule_work(&one->tx_work); tx_work 853 drivers/tty/serial/max310x.c struct max310x_one *one = container_of(ws, struct max310x_one, tx_work); tx_work 1359 drivers/tty/serial/max310x.c INIT_WORK(&s->p[i].tx_work, max310x_tx_proc); tx_work 1427 drivers/tty/serial/max310x.c cancel_work_sync(&s->p[i].tx_work); tx_work 315 drivers/tty/serial/sc16is7xx.c struct kthread_work tx_work; tx_work 743 drivers/tty/serial/sc16is7xx.c struct uart_port *port = &(to_sc16is7xx_one(ws, tx_work)->port); tx_work 830 drivers/tty/serial/sc16is7xx.c kthread_queue_work(&s->kworker, &one->tx_work); tx_work 1281 drivers/tty/serial/sc16is7xx.c kthread_init_work(&s->p[i].tx_work, sc16is7xx_tx_proc); tx_work 575 include/drm/drm_dp_mst_helper.h struct work_struct tx_work; tx_work 43 include/linux/netpoll.h struct delayed_work tx_work; tx_work 359 include/net/bluetooth/hci_core.h struct work_struct tx_work; tx_work 70 include/net/kcm.h struct work_struct tx_work; tx_work 222 include/net/nfc/nci_core.h struct work_struct tx_work; tx_work 156 include/net/tls.h struct tx_work tx_work; tx_work 1561 net/bluetooth/hci_core.c flush_work(&hdev->tx_work); tx_work 1690 net/bluetooth/hci_core.c flush_work(&hdev->tx_work); tx_work 3250 net/bluetooth/hci_core.c INIT_WORK(&hdev->tx_work, hci_tx_work); tx_work 3790 net/bluetooth/hci_core.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 3811 net/bluetooth/hci_core.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 4263 net/bluetooth/hci_core.c struct hci_dev *hdev = container_of(work, struct hci_dev, tx_work); tx_work 3697 net/bluetooth/hci_event.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 3769 net/bluetooth/hci_event.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 1758 net/bluetooth/hci_sock.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 1779 net/bluetooth/hci_sock.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 1802 net/bluetooth/hci_sock.c queue_work(hdev->workqueue, &hdev->tx_work); tx_work 101 net/core/netpoll.c container_of(work, struct netpoll_info, tx_work.work); tx_work 130 net/core/netpoll.c schedule_delayed_work(&npinfo->tx_work, HZ/10); tx_work 357 net/core/netpoll.c schedule_delayed_work(&npinfo->tx_work,0); tx_work 609 net/core/netpoll.c INIT_DELAYED_WORK(&npinfo->tx_work, queue_process); tx_work 784 net/core/netpoll.c cancel_delayed_work(&npinfo->tx_work); tx_work 789 net/core/netpoll.c cancel_delayed_work(&npinfo->tx_work); tx_work 82 net/kcm/kcmsock.c queue_work(kcm_wq, &psock->tx_kcm->tx_work); tx_work 426 net/kcm/kcmsock.c queue_work(kcm_wq, &kcm->tx_work); tx_work 512 net/kcm/kcmsock.c queue_work(kcm_wq, &kcm->tx_work); tx_work 723 net/kcm/kcmsock.c struct kcm_sock *kcm = container_of(w, struct kcm_sock, tx_work); tx_work 1364 net/kcm/kcmsock.c INIT_WORK(&kcm->tx_work, kcm_tx_work); tx_work 1583 net/kcm/kcmsock.c queue_work(kcm_wq, &psock->tx_kcm->tx_work); tx_work 1874 net/kcm/kcmsock.c cancel_work_sync(&kcm->tx_work); tx_work 58 net/mac802154/ieee802154_i.h struct work_struct tx_work; tx_work 100 net/mac802154/main.c INIT_WORK(&local->tx_work, ieee802154_xmit_worker); tx_work 28 net/mac802154/tx.c container_of(work, struct ieee802154_local, tx_work); tx_work 91 net/mac802154/tx.c queue_work(local->workqueue, &local->tx_work); tx_work 61 net/nfc/llcp.h struct work_struct tx_work; tx_work 158 net/nfc/llcp_core.c cancel_work_sync(&local->tx_work); tx_work 717 net/nfc/llcp_core.c tx_work); tx_work 1488 net/nfc/llcp_core.c schedule_work(&local->tx_work); tx_work 1561 net/nfc/llcp_core.c schedule_work(&local->tx_work); tx_work 1583 net/nfc/llcp_core.c INIT_WORK(&local->tx_work, nfc_llcp_tx_work); tx_work 1211 net/nfc/nci/core.c INIT_WORK(&ndev->tx_work, nci_tx_work); tx_work 1424 net/nfc/nci/core.c struct nci_dev *ndev = container_of(work, struct nci_dev, tx_work); tx_work 196 net/nfc/nci/data.c queue_work(ndev->tx_wq, &ndev->tx_work); tx_work 62 net/nfc/nci/ntf.c queue_work(ndev->tx_wq, &ndev->tx_work); tx_work 31 net/nfc/nfc.h struct work_struct tx_work; tx_work 42 net/nfc/nfc.h ((struct sock *) container_of(_tx_work, struct nfc_rawsock, tx_work)) tx_work 160 net/nfc/rawsock.c schedule_work(&nfc_rawsock(sk)->tx_work); tx_work 230 net/nfc/rawsock.c schedule_work(&nfc_rawsock(sk)->tx_work); tx_work 351 net/nfc/rawsock.c INIT_WORK(&nfc_rawsock(sk)->tx_work, rawsock_tx_work); tx_work 233 net/smc/af_smc.c INIT_DELAYED_WORK(&smc->conn.tx_work, smc_tx_work); tx_work 1751 net/smc/af_smc.c mod_delayed_work(system_wq, &smc->conn.tx_work, tx_work 1760 net/smc/af_smc.c mod_delayed_work(system_wq, &smc->conn.tx_work, tx_work 157 net/smc/smc.h struct delayed_work tx_work; /* retry of smc_cdc_msg_send */ tx_work 130 net/smc/smc_close.c cancel_delayed_work_sync(&smc->conn.tx_work); tx_work 141 net/smc/smc_close.c cancel_delayed_work_sync(&smc->conn.tx_work); tx_work 213 net/smc/smc_close.c cancel_delayed_work_sync(&conn->tx_work); tx_work 242 net/smc/smc_close.c cancel_delayed_work_sync(&conn->tx_work); tx_work 358 net/smc/smc_close.c cancel_delayed_work_sync(&conn->tx_work); tx_work 448 net/smc/smc_close.c cancel_delayed_work_sync(&conn->tx_work); tx_work 463 net/smc/smc_close.c cancel_delayed_work_sync(&conn->tx_work); tx_work 230 net/smc/smc_tx.c schedule_delayed_work(&conn->tx_work, tx_work 500 net/smc/smc_tx.c mod_delayed_work(system_wq, &conn->tx_work, tx_work 571 net/smc/smc_tx.c tx_work); tx_work 613 net/smc/smc_tx.c schedule_delayed_work(&conn->tx_work, tx_work 481 net/tls/tls_sw.c schedule_delayed_work(&ctx->tx_work.work, 1); tx_work 1123 net/tls/tls_sw.c cancel_delayed_work(&ctx->tx_work.work); tx_work 1242 net/tls/tls_sw.c cancel_delayed_work(&ctx->tx_work.work); tx_work 2116 net/tls/tls_sw.c cancel_delayed_work_sync(&ctx->tx_work.work); tx_work 2214 net/tls/tls_sw.c struct tx_work *tx_work = container_of(delayed_work, tx_work 2215 net/tls/tls_sw.c struct tx_work, work); tx_work 2216 net/tls/tls_sw.c struct sock *sk = tx_work->sk; tx_work 2243 net/tls/tls_sw.c schedule_delayed_work(&tx_ctx->tx_work.work, 0); tx_work 2315 net/tls/tls_sw.c INIT_DELAYED_WORK(&sw_ctx_tx->tx_work.work, tx_work_handler); tx_work 2316 net/tls/tls_sw.c sw_ctx_tx->tx_work.sk = sk; tx_work 33 net/vmw_vsock/virtio_transport.c struct work_struct tx_work; tx_work 333 net/vmw_vsock/virtio_transport.c container_of(work, struct virtio_vsock, tx_work); tx_work 531 net/vmw_vsock/virtio_transport.c queue_work(virtio_vsock_workqueue, &vsock->tx_work); tx_work 643 net/vmw_vsock/virtio_transport.c INIT_WORK(&vsock->tx_work, virtio_transport_tx_work); tx_work 744 net/vmw_vsock/virtio_transport.c flush_work(&vsock->tx_work);