netdev_work 741 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c struct pvrdma_netdevice_work *netdev_work; netdev_work 744 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work = container_of(work, struct pvrdma_netdevice_work, work); netdev_work 748 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c if ((netdev_work->event == NETDEV_REGISTER) || netdev_work 749 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c (dev->netdev == netdev_work->event_netdev)) { netdev_work 751 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work->event_netdev, netdev_work 752 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work->event); netdev_work 758 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c kfree(netdev_work); netdev_work 765 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c struct pvrdma_netdevice_work *netdev_work; netdev_work 767 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work = kmalloc(sizeof(*netdev_work), GFP_ATOMIC); netdev_work 768 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c if (!netdev_work) netdev_work 771 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c INIT_WORK(&netdev_work->work, pvrdma_netdevice_event_work); netdev_work 772 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work->event_netdev = event_netdev; netdev_work 773 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c netdev_work->event = event; netdev_work 774 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c queue_work(event_wq, &netdev_work->work);