netdev_work       741 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	struct pvrdma_netdevice_work *netdev_work;
netdev_work       744 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	netdev_work = container_of(work, struct pvrdma_netdevice_work, work);
netdev_work       748 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 		if ((netdev_work->event == NETDEV_REGISTER) ||
netdev_work       749 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 		    (dev->netdev == netdev_work->event_netdev)) {
netdev_work       751 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 						      netdev_work->event_netdev,
netdev_work       752 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 						      netdev_work->event);
netdev_work       758 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	kfree(netdev_work);
netdev_work       765 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	struct pvrdma_netdevice_work *netdev_work;
netdev_work       767 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	netdev_work = kmalloc(sizeof(*netdev_work), GFP_ATOMIC);
netdev_work       768 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	if (!netdev_work)
netdev_work       771 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	INIT_WORK(&netdev_work->work, pvrdma_netdevice_event_work);
netdev_work       772 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	netdev_work->event_netdev = event_netdev;
netdev_work       773 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	netdev_work->event = event;
netdev_work       774 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c 	queue_work(event_wq, &netdev_work->work);