rx_work 271 drivers/misc/mei/bus.c cldev = container_of(work, struct mei_cl_device, rx_work); rx_work 338 drivers/misc/mei/bus.c schedule_work(&cldev->rx_work); rx_work 364 drivers/misc/mei/bus.c INIT_WORK(&cldev->rx_work, mei_cl_bus_rx_work); rx_work 554 drivers/misc/mei/bus.c cancel_work_sync(&cldev->rx_work); rx_work 403 drivers/net/ethernet/dec/tulip/de2104x.c unsigned rx_work = DE_RX_RING_SIZE; rx_work 407 drivers/net/ethernet/dec/tulip/de2104x.c while (--rx_work) { rx_work 444 drivers/net/ethernet/dec/tulip/de2104x.c rx_work = 100; rx_work 489 drivers/net/ethernet/dec/tulip/de2104x.c if (!rx_work) rx_work 2861 drivers/net/ethernet/nvidia/forcedeth.c int rx_work = 0; rx_work 2867 drivers/net/ethernet/nvidia/forcedeth.c (rx_work < limit)) { rx_work 2952 drivers/net/ethernet/nvidia/forcedeth.c rx_work++; rx_work 2955 drivers/net/ethernet/nvidia/forcedeth.c return rx_work; rx_work 2963 drivers/net/ethernet/nvidia/forcedeth.c int rx_work = 0; rx_work 2969 drivers/net/ethernet/nvidia/forcedeth.c (rx_work < limit)) { rx_work 3041 drivers/net/ethernet/nvidia/forcedeth.c rx_work++; rx_work 3044 drivers/net/ethernet/nvidia/forcedeth.c return rx_work; rx_work 3750 drivers/net/ethernet/nvidia/forcedeth.c int rx_count, tx_work = 0, rx_work = 0; rx_work 3758 drivers/net/ethernet/nvidia/forcedeth.c rx_count = nv_rx_process(dev, budget - rx_work); rx_work 3766 drivers/net/ethernet/nvidia/forcedeth.c budget - rx_work); rx_work 3770 drivers/net/ethernet/nvidia/forcedeth.c rx_count > 0 && (rx_work += rx_count) < budget); rx_work 3779 drivers/net/ethernet/nvidia/forcedeth.c nv_change_interrupt_mode(dev, tx_work + rx_work); rx_work 3801 drivers/net/ethernet/nvidia/forcedeth.c return rx_work; rx_work 3804 drivers/net/ethernet/nvidia/forcedeth.c if (rx_work < budget) { rx_work 3807 drivers/net/ethernet/nvidia/forcedeth.c napi_complete_done(napi, rx_work); rx_work 3811 drivers/net/ethernet/nvidia/forcedeth.c return rx_work; rx_work 168 drivers/net/ethernet/wiznet/w5100.c struct work_struct rx_work; rx_work 887 drivers/net/ethernet/wiznet/w5100.c rx_work); rx_work 937 drivers/net/ethernet/wiznet/w5100.c queue_work(priv->xfer_wq, &priv->rx_work); rx_work 1155 drivers/net/ethernet/wiznet/w5100.c INIT_WORK(&priv->rx_work, w5100_rx_work); rx_work 194 drivers/net/hamradio/dmascc.c struct work_struct rx_work; rx_work 569 drivers/net/hamradio/dmascc.c INIT_WORK(&priv->rx_work, rx_bh); rx_work 1227 drivers/net/hamradio/dmascc.c schedule_work(&priv->rx_work); rx_work 1250 drivers/net/hamradio/dmascc.c struct scc_priv *priv = container_of(ugli_api, struct scc_priv, rx_work); rx_work 246 drivers/net/wireless/ath/carl9170/carl9170.h struct usb_anchor rx_work; rx_work 348 drivers/net/wireless/ath/carl9170/usb.c urb = usb_get_from_anchor(&ar->rx_work); rx_work 411 drivers/net/wireless/ath/carl9170/usb.c usb_anchor_urb(urb, &ar->rx_work); rx_work 551 drivers/net/wireless/ath/carl9170/usb.c usb_scuttle_anchored_urbs(&ar->rx_work); rx_work 599 drivers/net/wireless/ath/carl9170/usb.c usb_scuttle_anchored_urbs(&ar->rx_work); rx_work 1077 drivers/net/wireless/ath/carl9170/usb.c init_usb_anchor(&ar->rx_work); rx_work 181 drivers/net/wireless/marvell/mwifiex/main.c queue_work(adapter->rx_workqueue, &adapter->rx_work); rx_work 1371 drivers/net/wireless/marvell/mwifiex/main.c container_of(work, struct mwifiex_adapter, rx_work); rx_work 1523 drivers/net/wireless/marvell/mwifiex/main.c INIT_WORK(&adapter->rx_work, mwifiex_rx_work_queue); rx_work 1679 drivers/net/wireless/marvell/mwifiex/main.c INIT_WORK(&adapter->rx_work, mwifiex_rx_work_queue); rx_work 899 drivers/net/wireless/marvell/mwifiex/main.h struct work_struct rx_work; rx_work 407 drivers/nfc/nfcmrvl/fw_dnld.c rx_work); rx_work 458 drivers/nfc/nfcmrvl/fw_dnld.c INIT_WORK(&priv->fw_dnld.rx_work, fw_dnld_rx_work); rx_work 485 drivers/nfc/nfcmrvl/fw_dnld.c queue_work(priv->fw_dnld.rx_wq, &priv->fw_dnld.rx_work); rx_work 85 drivers/nfc/nfcmrvl/fw_dnld.h struct work_struct rx_work; rx_work 162 drivers/rapidio/rio_cm.c struct work_struct rx_work; rx_work 582 drivers/rapidio/rio_cm.c struct cm_dev *cm = container_of(work, struct cm_dev, rx_work); rx_work 622 drivers/rapidio/rio_cm.c if (rio_mport_is_running(cm->mport) && !work_pending(&cm->rx_work)) rx_work 623 drivers/rapidio/rio_cm.c queue_work(cm->rx_wq, &cm->rx_work); rx_work 2149 drivers/rapidio/rio_cm.c INIT_WORK(&cm->rx_work, rio_ibmsg_handler); rx_work 109 drivers/rpmsg/qcom_glink_native.c struct work_struct rx_work; rx_work 780 drivers/rpmsg/qcom_glink_native.c schedule_work(&glink->rx_work); rx_work 1513 drivers/rpmsg/qcom_glink_native.c rx_work); rx_work 1571 drivers/rpmsg/qcom_glink_native.c cancel_work_sync(&glink->rx_work); rx_work 1601 drivers/rpmsg/qcom_glink_native.c INIT_WORK(&glink->rx_work, qcom_glink_work); rx_work 25 drivers/soc/qcom/apr.c struct work_struct rx_work; rx_work 96 drivers/soc/qcom/apr.c queue_work(apr->rxwq, &apr->rx_work); rx_work 172 drivers/soc/qcom/apr.c struct apr *apr = container_of(work, struct apr, rx_work); rx_work 345 drivers/soc/qcom/apr.c INIT_WORK(&apr->rx_work, apr_rxwq); rx_work 439 drivers/staging/nvec/nvec.c struct nvec_chip *nvec = container_of(work, struct nvec_chip, rx_work); rx_work 520 drivers/staging/nvec/nvec.c schedule_work(&nvec->rx_work); rx_work 825 drivers/staging/nvec/nvec.c INIT_WORK(&nvec->rx_work, nvec_dispatch); rx_work 891 drivers/staging/nvec/nvec.c cancel_work_sync(&nvec->rx_work); rx_work 144 drivers/staging/nvec/nvec.h struct work_struct rx_work, tx_work; rx_work 50 include/linux/mei_cl_bus.h struct work_struct rx_work; rx_work 357 include/net/bluetooth/hci_core.h struct work_struct rx_work; rx_work 219 include/net/nfc/nci_core.h struct work_struct rx_work; rx_work 1563 net/bluetooth/hci_core.c flush_work(&hdev->rx_work); rx_work 1691 net/bluetooth/hci_core.c flush_work(&hdev->rx_work); rx_work 3248 net/bluetooth/hci_core.c INIT_WORK(&hdev->rx_work, hci_rx_work); rx_work 3511 net/bluetooth/hci_core.c queue_work(hdev->workqueue, &hdev->rx_work); rx_work 3527 net/bluetooth/hci_core.c queue_work(hdev->workqueue, &hdev->rx_work); rx_work 4451 net/bluetooth/hci_core.c struct hci_dev *hdev = container_of(work, struct hci_dev, rx_work); rx_work 62 net/nfc/llcp.h struct work_struct rx_work; rx_work 159 net/nfc/llcp_core.c cancel_work_sync(&local->rx_work); rx_work 1473 net/nfc/llcp_core.c rx_work); rx_work 1497 net/nfc/llcp_core.c schedule_work(&local->rx_work); rx_work 1586 net/nfc/llcp_core.c INIT_WORK(&local->rx_work, nfc_llcp_rx_work); rx_work 596 net/nfc/nci/core.c queue_work(ndev->rx_wq, &ndev->rx_work); rx_work 1203 net/nfc/nci/core.c INIT_WORK(&ndev->rx_work, nci_rx_work); rx_work 1288 net/nfc/nci/core.c queue_work(ndev->rx_wq, &ndev->rx_work); rx_work 1461 net/nfc/nci/core.c struct nci_dev *ndev = container_of(work, struct nci_dev, rx_work); rx_work 34 net/vmw_vsock/virtio_transport.c struct work_struct rx_work; rx_work 206 net/vmw_vsock/virtio_transport.c queue_work(virtio_vsock_workqueue, &vsock->rx_work); rx_work 279 net/vmw_vsock/virtio_transport.c queue_work(virtio_vsock_workqueue, &vsock->rx_work); rx_work 376 net/vmw_vsock/virtio_transport.c container_of(work, struct virtio_vsock, rx_work); rx_work 540 net/vmw_vsock/virtio_transport.c queue_work(virtio_vsock_workqueue, &vsock->rx_work); rx_work 642 net/vmw_vsock/virtio_transport.c INIT_WORK(&vsock->rx_work, virtio_transport_rx_work); rx_work 743 net/vmw_vsock/virtio_transport.c flush_work(&vsock->rx_work);