tx_work           160 drivers/bluetooth/btmtksdio.c 	struct work_struct tx_work;
tx_work           306 drivers/bluetooth/btmtksdio.c 						  tx_work);
tx_work           520 drivers/bluetooth/btmtksdio.c 		schedule_work(&bdev->tx_work);
tx_work           653 drivers/bluetooth/btmtksdio.c 	cancel_work_sync(&bdev->tx_work);
tx_work           920 drivers/bluetooth/btmtksdio.c 	schedule_work(&bdev->tx_work);
tx_work           943 drivers/bluetooth/btmtksdio.c 	INIT_WORK(&bdev->tx_work, btmtksdio_tx_work);
tx_work           130 drivers/bluetooth/btmtkuart.c 	struct work_struct tx_work;
tx_work           371 drivers/bluetooth/btmtkuart.c 						   tx_work);
tx_work           422 drivers/bluetooth/btmtkuart.c 	schedule_work(&bdev->tx_work);
tx_work           629 drivers/bluetooth/btmtkuart.c 	cancel_work_sync(&bdev->tx_work);
tx_work           989 drivers/bluetooth/btmtkuart.c 	INIT_WORK(&bdev->tx_work, btmtkuart_tx_work);
tx_work          3604 drivers/gpu/drm/drm_dp_mst_topology.c 	queue_work(system_long_wq, &mgr->tx_work);
tx_work          3731 drivers/gpu/drm/drm_dp_mst_topology.c 	struct drm_dp_mst_topology_mgr *mgr = container_of(work, struct drm_dp_mst_topology_mgr, tx_work);
tx_work          3962 drivers/gpu/drm/drm_dp_mst_topology.c 	INIT_WORK(&mgr->tx_work, drm_dp_tx_work);
tx_work           661 drivers/infiniband/sw/siw/siw_qp_rx.c 	struct siw_wqe *tx_work = tx_wqe(qp);
tx_work           682 drivers/infiniband/sw/siw/siw_qp_rx.c 	if (tx_work->wr_status == SIW_WR_IDLE) {
tx_work           687 drivers/infiniband/sw/siw/siw_qp_rx.c 		tx_work->processed = 0;
tx_work           688 drivers/infiniband/sw/siw/siw_qp_rx.c 		tx_work->mem[0] = NULL;
tx_work           689 drivers/infiniband/sw/siw/siw_qp_rx.c 		tx_work->wr_status = SIW_WR_QUEUED;
tx_work           690 drivers/infiniband/sw/siw/siw_qp_rx.c 		resp = &tx_work->sqe;
tx_work            61 drivers/input/serio/ps2-gpio.c 	struct delayed_work tx_work;
tx_work            76 drivers/input/serio/ps2-gpio.c 	flush_delayed_work(&drvdata->tx_work);
tx_work            90 drivers/input/serio/ps2-gpio.c 	schedule_delayed_work(&drvdata->tx_work, usecs_to_jiffies(200));
tx_work           119 drivers/input/serio/ps2-gpio.c 						    tx_work);
tx_work           408 drivers/input/serio/ps2-gpio.c 	INIT_DELAYED_WORK(&drvdata->tx_work, ps2_gpio_tx_work_fn);
tx_work          1534 drivers/net/can/m_can/m_can.c 						tx_work);
tx_work          1564 drivers/net/can/m_can/m_can.c 			queue_work(cdev->tx_wq, &cdev->tx_work);
tx_work          1600 drivers/net/can/m_can/m_can.c 		INIT_WORK(&cdev->tx_work, m_can_tx_work_queue);
tx_work            81 drivers/net/can/m_can/m_can.h 	struct work_struct tx_work;
tx_work            87 drivers/net/can/slcan.c 	struct work_struct	tx_work;	/* Flushes transmit buffer   */
tx_work           313 drivers/net/can/slcan.c 	struct slcan *sl = container_of(work, struct slcan, tx_work);
tx_work           352 drivers/net/can/slcan.c 	schedule_work(&sl->tx_work);
tx_work           550 drivers/net/can/slcan.c 	INIT_WORK(&sl->tx_work, slcan_transmit);
tx_work           660 drivers/net/can/slcan.c 	flush_work(&sl->tx_work);
tx_work           159 drivers/net/can/spi/hi311x.c 	struct work_struct tx_work;
tx_work           381 drivers/net/can/spi/hi311x.c 	queue_work(priv->wq, &priv->tx_work);
tx_work           576 drivers/net/can/spi/hi311x.c 						tx_work);
tx_work           772 drivers/net/can/spi/hi311x.c 	INIT_WORK(&priv->tx_work, hi3110_tx_work_handler);
tx_work           216 drivers/net/can/spi/mcp251x.c 	struct work_struct tx_work;
tx_work           476 drivers/net/can/spi/mcp251x.c 	queue_work(priv->wq, &priv->tx_work);
tx_work           686 drivers/net/can/spi/mcp251x.c 						 tx_work);
tx_work           913 drivers/net/can/spi/mcp251x.c 	INIT_WORK(&priv->tx_work, mcp251x_tx_work_handler);
tx_work           124 drivers/net/ethernet/micrel/ks8851.c 	struct work_struct	tx_work;
tx_work           758 drivers/net/ethernet/micrel/ks8851.c 	struct ks8851_net *ks = container_of(work, struct ks8851_net, tx_work);
tx_work           891 drivers/net/ethernet/micrel/ks8851.c 	flush_work(&ks->tx_work);
tx_work           954 drivers/net/ethernet/micrel/ks8851.c 	schedule_work(&ks->tx_work);
tx_work          1485 drivers/net/ethernet/micrel/ks8851.c 	INIT_WORK(&ks->tx_work, ks8851_tx_work);
tx_work            60 drivers/net/ethernet/microchip/enc28j60.c 	struct work_struct tx_work;
tx_work          1298 drivers/net/ethernet/microchip/enc28j60.c 	schedule_work(&priv->tx_work);
tx_work          1306 drivers/net/ethernet/microchip/enc28j60.c 		container_of(work, struct enc28j60_net, tx_work);
tx_work          1561 drivers/net/ethernet/microchip/enc28j60.c 	INIT_WORK(&priv->tx_work, enc28j60_tx_work_handler);
tx_work            54 drivers/net/ethernet/microchip/encx24j600.c 	struct kthread_work       tx_work;
tx_work           870 drivers/net/ethernet/microchip/encx24j600.c 			container_of(ws, struct encx24j600_priv, tx_work);
tx_work           889 drivers/net/ethernet/microchip/encx24j600.c 	kthread_queue_work(&priv->kworker, &priv->tx_work);
tx_work          1053 drivers/net/ethernet/microchip/encx24j600.c 	kthread_init_work(&priv->tx_work, encx24j600_tx_proc);
tx_work          2576 drivers/net/ethernet/nvidia/forcedeth.c 	int tx_work = 0;
tx_work          2582 drivers/net/ethernet/nvidia/forcedeth.c 	       (tx_work < limit)) {
tx_work          2604 drivers/net/ethernet/nvidia/forcedeth.c 				tx_work++;
tx_work          2624 drivers/net/ethernet/nvidia/forcedeth.c 				tx_work++;
tx_work          2633 drivers/net/ethernet/nvidia/forcedeth.c 	netdev_completed_queue(np->dev, tx_work, bytes_compl);
tx_work          2639 drivers/net/ethernet/nvidia/forcedeth.c 	return tx_work;
tx_work          2646 drivers/net/ethernet/nvidia/forcedeth.c 	int tx_work = 0;
tx_work          2652 drivers/net/ethernet/nvidia/forcedeth.c 	       (tx_work < limit)) {
tx_work          2678 drivers/net/ethernet/nvidia/forcedeth.c 			tx_work++;
tx_work          2690 drivers/net/ethernet/nvidia/forcedeth.c 	netdev_completed_queue(np->dev, tx_work, bytes_cleaned);
tx_work          2696 drivers/net/ethernet/nvidia/forcedeth.c 	return tx_work;
tx_work          3750 drivers/net/ethernet/nvidia/forcedeth.c 	int rx_count, tx_work = 0, rx_work = 0;
tx_work          3755 drivers/net/ethernet/nvidia/forcedeth.c 			tx_work += nv_tx_done(dev, np->tx_ring_size);
tx_work          3762 drivers/net/ethernet/nvidia/forcedeth.c 			tx_work += nv_tx_done_optimized(dev, np->tx_ring_size);
tx_work          3779 drivers/net/ethernet/nvidia/forcedeth.c 	nv_change_interrupt_mode(dev, tx_work + rx_work);
tx_work            51 drivers/net/ethernet/qualcomm/qca_uart.c 	struct work_struct tx_work;		/* Flushes transmit buffer   */
tx_work           129 drivers/net/ethernet/qualcomm/qca_uart.c 	struct qcauart *qca = container_of(work, struct qcauart, tx_work);
tx_work           167 drivers/net/ethernet/qualcomm/qca_uart.c 	schedule_work(&qca->tx_work);
tx_work           189 drivers/net/ethernet/qualcomm/qca_uart.c 	flush_work(&qca->tx_work);
tx_work           347 drivers/net/ethernet/qualcomm/qca_uart.c 	INIT_WORK(&qca->tx_work, qcauart_transmit);
tx_work           383 drivers/net/ethernet/qualcomm/qca_uart.c 		cancel_work_sync(&qca->tx_work);
tx_work           402 drivers/net/ethernet/qualcomm/qca_uart.c 	cancel_work_sync(&qca->tx_work);
tx_work           170 drivers/net/ethernet/wiznet/w5100.c 	struct work_struct tx_work;
tx_work           821 drivers/net/ethernet/wiznet/w5100.c 					       tx_work);
tx_work           840 drivers/net/ethernet/wiznet/w5100.c 		queue_work(priv->xfer_wq, &priv->tx_work);
tx_work          1156 drivers/net/ethernet/wiznet/w5100.c 	INIT_WORK(&priv->tx_work, w5100_tx_work);
tx_work           423 drivers/net/slip/slip.c 	struct slip *sl = container_of(work, struct slip, tx_work);
tx_work           462 drivers/net/slip/slip.c 	schedule_work(&sl->tx_work);
tx_work           769 drivers/net/slip/slip.c 	INIT_WORK(&sl->tx_work, slip_transmit);
tx_work           900 drivers/net/slip/slip.c 	flush_work(&sl->tx_work);
tx_work            57 drivers/net/slip/slip.h   struct work_struct	tx_work;	/* Flushes transmit buffer	*/
tx_work           785 drivers/net/wireless/ath/ar5523/ar5523.c 	ieee80211_queue_work(ar->hw, &ar->tx_work);
tx_work           885 drivers/net/wireless/ath/ar5523/ar5523.c 	struct ar5523 *ar = container_of(work, struct ar5523, tx_work);
tx_work          1605 drivers/net/wireless/ath/ar5523/ar5523.c 	INIT_WORK(&ar->tx_work, ar5523_tx_work);
tx_work           101 drivers/net/wireless/ath/ar5523/ar5523.h 	struct work_struct	tx_work;
tx_work           958 drivers/net/wireless/broadcom/b43/b43.h 	struct work_struct tx_work;
tx_work          1579 drivers/net/wireless/broadcom/b43/dma.c 	ieee80211_queue_work(dev->wl->hw, &dev->wl->tx_work);
tx_work          3582 drivers/net/wireless/broadcom/b43/main.c 	struct b43_wl *wl = container_of(work, struct b43_wl, tx_work);
tx_work          3638 drivers/net/wireless/broadcom/b43/main.c 		ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work          4336 drivers/net/wireless/broadcom/b43/main.c 	cancel_work_sync(&wl->tx_work);
tx_work          5601 drivers/net/wireless/broadcom/b43/main.c 	INIT_WORK(&wl->tx_work, b43_tx_work);
tx_work           631 drivers/net/wireless/broadcom/b43legacy/b43legacy.h 	struct work_struct tx_work;
tx_work          1238 drivers/net/wireless/broadcom/b43legacy/dma.c 	ieee80211_queue_work(dev->wl->hw, &dev->wl->tx_work);
tx_work          2448 drivers/net/wireless/broadcom/b43legacy/main.c 				  tx_work);
tx_work          2501 drivers/net/wireless/broadcom/b43legacy/main.c 		ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work          2936 drivers/net/wireless/broadcom/b43legacy/main.c 	cancel_work_sync(&wl->tx_work);
tx_work          3831 drivers/net/wireless/broadcom/b43legacy/main.c 	INIT_WORK(&wl->tx_work, b43legacy_tx_work);
tx_work           195 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h 	struct work_struct tx_work;
tx_work           189 drivers/net/wireless/marvell/libertas_tf/main.c 	queue_work(lbtf_wq, &priv->tx_work);
tx_work           200 drivers/net/wireless/marvell/libertas_tf/main.c 					 tx_work);
tx_work           290 drivers/net/wireless/marvell/libertas_tf/main.c 	cancel_work_sync(&priv->tx_work);
tx_work           595 drivers/net/wireless/marvell/libertas_tf/main.c 	INIT_WORK(&priv->tx_work, lbtf_tx_work);
tx_work           672 drivers/net/wireless/marvell/libertas_tf/main.c 		queue_work(lbtf_wq, &priv->tx_work);
tx_work           692 drivers/net/wireless/marvell/libertas_tf/main.c 			queue_work(lbtf_wq, &priv->tx_work);
tx_work           352 drivers/net/wireless/ti/wl1251/main.c 	ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work           446 drivers/net/wireless/ti/wl1251/main.c 	cancel_work_sync(&wl->tx_work);
tx_work          1621 drivers/net/wireless/ti/wl1251/main.c 	INIT_WORK(&wl->tx_work, wl1251_tx_work);
tx_work           332 drivers/net/wireless/ti/wl1251/tx.c 	struct wl1251 *wl = container_of(work, struct wl1251, tx_work);
tx_work           481 drivers/net/wireless/ti/wl1251/tx.c 		ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work           308 drivers/net/wireless/ti/wl1251/wl1251.h 	struct work_struct tx_work;
tx_work           677 drivers/net/wireless/ti/wlcore/main.c 	cancel_work_sync(&wl->tx_work);
tx_work           690 drivers/net/wireless/ti/wlcore/main.c 		ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work          1269 drivers/net/wireless/ti/wlcore/main.c 		ieee80211_queue_work(wl->hw, &wl->tx_work);
tx_work          1779 drivers/net/wireless/ti/wlcore/main.c 	flush_work(&wl->tx_work);
tx_work          1942 drivers/net/wireless/ti/wlcore/main.c 	cancel_work_sync(&wl->tx_work);
tx_work          6392 drivers/net/wireless/ti/wlcore/main.c 	INIT_WORK(&wl->tx_work, wl1271_tx_work);
tx_work           854 drivers/net/wireless/ti/wlcore/tx.c 	struct wl1271 *wl = container_of(work, struct wl1271, tx_work);
tx_work          1160 drivers/net/wireless/ti/wlcore/tx.c 			wl1271_tx_work(&wl->tx_work);
tx_work           271 drivers/net/wireless/ti/wlcore/wlcore.h 	struct work_struct tx_work;
tx_work           114 drivers/nfc/st21nfca/dep.c 						dep_info.tx_work);
tx_work           137 drivers/nfc/st21nfca/dep.c 	schedule_work(&info->dep_info.tx_work);
tx_work           669 drivers/nfc/st21nfca/dep.c 	INIT_WORK(&info->dep_info.tx_work, st21nfca_tx_work);
tx_work           680 drivers/nfc/st21nfca/dep.c 	cancel_work_sync(&info->dep_info.tx_work);
tx_work           117 drivers/nfc/st21nfca/st21nfca.h 	struct work_struct tx_work;
tx_work            70 drivers/staging/greybus/uart.c 	struct work_struct tx_work;
tx_work           194 drivers/staging/greybus/uart.c 		schedule_work(&gb_tty->tx_work);
tx_work           241 drivers/staging/greybus/uart.c 	gb_tty = container_of(work, struct gb_tty, tx_work);
tx_work           449 drivers/staging/greybus/uart.c 		schedule_work(&gb_tty->tx_work);
tx_work           768 drivers/staging/greybus/uart.c 	cancel_work_sync(&gb_tty->tx_work);
tx_work           855 drivers/staging/greybus/uart.c 	INIT_WORK(&gb_tty->tx_work, gb_uart_tx_write_work);
tx_work           273 drivers/staging/nvec/nvec.c 	schedule_work(&nvec->tx_work);
tx_work           375 drivers/staging/nvec/nvec.c 	struct nvec_chip *nvec = container_of(work, struct nvec_chip, tx_work);
tx_work           826 drivers/staging/nvec/nvec.c 	INIT_WORK(&nvec->tx_work, nvec_request_master);
tx_work           892 drivers/staging/nvec/nvec.c 	cancel_work_sync(&nvec->tx_work);
tx_work           144 drivers/staging/nvec/nvec.h 	struct work_struct rx_work, tx_work;
tx_work          1402 drivers/staging/qlge/qlge.h 	struct delayed_work tx_work;
tx_work           258 drivers/tty/serial/max310x.c 	struct work_struct	tx_work;
tx_work           794 drivers/tty/serial/max310x.c 	schedule_work(&one->tx_work);
tx_work           853 drivers/tty/serial/max310x.c 	struct max310x_one *one = container_of(ws, struct max310x_one, tx_work);
tx_work          1359 drivers/tty/serial/max310x.c 		INIT_WORK(&s->p[i].tx_work, max310x_tx_proc);
tx_work          1427 drivers/tty/serial/max310x.c 		cancel_work_sync(&s->p[i].tx_work);
tx_work           315 drivers/tty/serial/sc16is7xx.c 	struct kthread_work		tx_work;
tx_work           743 drivers/tty/serial/sc16is7xx.c 	struct uart_port *port = &(to_sc16is7xx_one(ws, tx_work)->port);
tx_work           830 drivers/tty/serial/sc16is7xx.c 	kthread_queue_work(&s->kworker, &one->tx_work);
tx_work          1281 drivers/tty/serial/sc16is7xx.c 		kthread_init_work(&s->p[i].tx_work, sc16is7xx_tx_proc);
tx_work           575 include/drm/drm_dp_mst_helper.h 	struct work_struct tx_work;
tx_work            43 include/linux/netpoll.h 	struct delayed_work tx_work;
tx_work           359 include/net/bluetooth/hci_core.h 	struct work_struct	tx_work;
tx_work            70 include/net/kcm.h 	struct work_struct tx_work;
tx_work           222 include/net/nfc/nci_core.h 	struct work_struct	tx_work;
tx_work           156 include/net/tls.h 	struct tx_work tx_work;
tx_work          1561 net/bluetooth/hci_core.c 		flush_work(&hdev->tx_work);
tx_work          1690 net/bluetooth/hci_core.c 	flush_work(&hdev->tx_work);
tx_work          3250 net/bluetooth/hci_core.c 	INIT_WORK(&hdev->tx_work, hci_tx_work);
tx_work          3790 net/bluetooth/hci_core.c 	queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          3811 net/bluetooth/hci_core.c 	queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          4263 net/bluetooth/hci_core.c 	struct hci_dev *hdev = container_of(work, struct hci_dev, tx_work);
tx_work          3697 net/bluetooth/hci_event.c 	queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          3769 net/bluetooth/hci_event.c 	queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          1758 net/bluetooth/hci_sock.c 		queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          1779 net/bluetooth/hci_sock.c 			queue_work(hdev->workqueue, &hdev->tx_work);
tx_work          1802 net/bluetooth/hci_sock.c 		queue_work(hdev->workqueue, &hdev->tx_work);
tx_work           101 net/core/netpoll.c 		container_of(work, struct netpoll_info, tx_work.work);
tx_work           130 net/core/netpoll.c 			schedule_delayed_work(&npinfo->tx_work, HZ/10);
tx_work           357 net/core/netpoll.c 		schedule_delayed_work(&npinfo->tx_work,0);
tx_work           609 net/core/netpoll.c 		INIT_DELAYED_WORK(&npinfo->tx_work, queue_process);
tx_work           784 net/core/netpoll.c 	cancel_delayed_work(&npinfo->tx_work);
tx_work           789 net/core/netpoll.c 	cancel_delayed_work(&npinfo->tx_work);
tx_work            82 net/kcm/kcmsock.c 		queue_work(kcm_wq, &psock->tx_kcm->tx_work);
tx_work           426 net/kcm/kcmsock.c 		queue_work(kcm_wq, &kcm->tx_work);
tx_work           512 net/kcm/kcmsock.c 		queue_work(kcm_wq, &kcm->tx_work);
tx_work           723 net/kcm/kcmsock.c 	struct kcm_sock *kcm = container_of(w, struct kcm_sock, tx_work);
tx_work          1364 net/kcm/kcmsock.c 	INIT_WORK(&kcm->tx_work, kcm_tx_work);
tx_work          1583 net/kcm/kcmsock.c 		queue_work(kcm_wq, &psock->tx_kcm->tx_work);
tx_work          1874 net/kcm/kcmsock.c 	cancel_work_sync(&kcm->tx_work);
tx_work            58 net/mac802154/ieee802154_i.h 	struct work_struct tx_work;
tx_work           100 net/mac802154/main.c 	INIT_WORK(&local->tx_work, ieee802154_xmit_worker);
tx_work            28 net/mac802154/tx.c 		container_of(work, struct ieee802154_local, tx_work);
tx_work            91 net/mac802154/tx.c 		queue_work(local->workqueue, &local->tx_work);
tx_work            61 net/nfc/llcp.h 	struct work_struct	 tx_work;
tx_work           158 net/nfc/llcp_core.c 	cancel_work_sync(&local->tx_work);
tx_work           717 net/nfc/llcp_core.c 						    tx_work);
tx_work          1488 net/nfc/llcp_core.c 	schedule_work(&local->tx_work);
tx_work          1561 net/nfc/llcp_core.c 		schedule_work(&local->tx_work);
tx_work          1583 net/nfc/llcp_core.c 	INIT_WORK(&local->tx_work, nfc_llcp_tx_work);
tx_work          1211 net/nfc/nci/core.c 	INIT_WORK(&ndev->tx_work, nci_tx_work);
tx_work          1424 net/nfc/nci/core.c 	struct nci_dev *ndev = container_of(work, struct nci_dev, tx_work);
tx_work           196 net/nfc/nci/data.c 	queue_work(ndev->tx_wq, &ndev->tx_work);
tx_work            62 net/nfc/nci/ntf.c 		queue_work(ndev->tx_wq, &ndev->tx_work);
tx_work            31 net/nfc/nfc.h  	struct work_struct tx_work;
tx_work            42 net/nfc/nfc.h  	((struct sock *) container_of(_tx_work, struct nfc_rawsock, tx_work))
tx_work           160 net/nfc/rawsock.c 		schedule_work(&nfc_rawsock(sk)->tx_work);
tx_work           230 net/nfc/rawsock.c 		schedule_work(&nfc_rawsock(sk)->tx_work);
tx_work           351 net/nfc/rawsock.c 		INIT_WORK(&nfc_rawsock(sk)->tx_work, rawsock_tx_work);
tx_work           233 net/smc/af_smc.c 	INIT_DELAYED_WORK(&smc->conn.tx_work, smc_tx_work);
tx_work          1751 net/smc/af_smc.c 				mod_delayed_work(system_wq, &smc->conn.tx_work,
tx_work          1760 net/smc/af_smc.c 				mod_delayed_work(system_wq, &smc->conn.tx_work,
tx_work           157 net/smc/smc.h  	struct delayed_work	tx_work;	/* retry of smc_cdc_msg_send */
tx_work           130 net/smc/smc_close.c 		cancel_delayed_work_sync(&smc->conn.tx_work);
tx_work           141 net/smc/smc_close.c 		cancel_delayed_work_sync(&smc->conn.tx_work);
tx_work           213 net/smc/smc_close.c 		cancel_delayed_work_sync(&conn->tx_work);
tx_work           242 net/smc/smc_close.c 		cancel_delayed_work_sync(&conn->tx_work);
tx_work           358 net/smc/smc_close.c 		cancel_delayed_work_sync(&conn->tx_work);
tx_work           448 net/smc/smc_close.c 		cancel_delayed_work_sync(&conn->tx_work);
tx_work           463 net/smc/smc_close.c 		cancel_delayed_work_sync(&conn->tx_work);
tx_work           230 net/smc/smc_tx.c 			schedule_delayed_work(&conn->tx_work,
tx_work           500 net/smc/smc_tx.c 				mod_delayed_work(system_wq, &conn->tx_work,
tx_work           571 net/smc/smc_tx.c 						   tx_work);
tx_work           613 net/smc/smc_tx.c 			schedule_delayed_work(&conn->tx_work,
tx_work           481 net/tls/tls_sw.c 		schedule_delayed_work(&ctx->tx_work.work, 1);
tx_work          1123 net/tls/tls_sw.c 		cancel_delayed_work(&ctx->tx_work.work);
tx_work          1242 net/tls/tls_sw.c 			cancel_delayed_work(&ctx->tx_work.work);
tx_work          2116 net/tls/tls_sw.c 	cancel_delayed_work_sync(&ctx->tx_work.work);
tx_work          2214 net/tls/tls_sw.c 	struct tx_work *tx_work = container_of(delayed_work,
tx_work          2215 net/tls/tls_sw.c 					       struct tx_work, work);
tx_work          2216 net/tls/tls_sw.c 	struct sock *sk = tx_work->sk;
tx_work          2243 net/tls/tls_sw.c 		schedule_delayed_work(&tx_ctx->tx_work.work, 0);
tx_work          2315 net/tls/tls_sw.c 		INIT_DELAYED_WORK(&sw_ctx_tx->tx_work.work, tx_work_handler);
tx_work          2316 net/tls/tls_sw.c 		sw_ctx_tx->tx_work.sk = sk;
tx_work            33 net/vmw_vsock/virtio_transport.c 	struct work_struct tx_work;
tx_work           333 net/vmw_vsock/virtio_transport.c 		container_of(work, struct virtio_vsock, tx_work);
tx_work           531 net/vmw_vsock/virtio_transport.c 	queue_work(virtio_vsock_workqueue, &vsock->tx_work);
tx_work           643 net/vmw_vsock/virtio_transport.c 	INIT_WORK(&vsock->tx_work, virtio_transport_tx_work);
tx_work           744 net/vmw_vsock/virtio_transport.c 	flush_work(&vsock->tx_work);