ndev_ctx 678 drivers/net/hyperv/netvsc.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 716 drivers/net/hyperv/netvsc.c ndev_ctx->eth_stats.wake_queue++; ndev_ctx 817 drivers/net/hyperv/netvsc.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 862 drivers/net/hyperv/netvsc.c ndev_ctx->eth_stats.stop_queue++; ndev_ctx 866 drivers/net/hyperv/netvsc.c ndev_ctx->eth_stats.stop_queue++; ndev_ctx 878 drivers/net/hyperv/netvsc.c ndev_ctx->eth_stats.wake_queue++; ndev_ctx 905 drivers/net/hyperv/netvsc.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 907 drivers/net/hyperv/netvsc.c = rcu_dereference_bh(ndev_ctx->nvdev); ndev_ctx 908 drivers/net/hyperv/netvsc.c struct hv_device *device = ndev_ctx->device_ctx; ndev_ctx 951 drivers/net/hyperv/netvsc.c ++ndev_ctx->eth_stats.tx_send_full; ndev_ctx 1046 drivers/net/hyperv/netvsc.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 1048 drivers/net/hyperv/netvsc.c ++ndev_ctx->eth_stats.rx_comp_busy; ndev_ctx 64 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 65 drivers/net/hyperv/netvsc_drv.c struct net_device *vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); ndev_ctx 84 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 89 drivers/net/hyperv/netvsc_drv.c vf_netdev = rcu_dereference(ndev_ctx->vf_netdev); ndev_ctx 95 drivers/net/hyperv/netvsc_drv.c nvdev = rcu_dereference(ndev_ctx->nvdev); ndev_ctx 112 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 113 drivers/net/hyperv/netvsc_drv.c struct net_device *vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); ndev_ctx 114 drivers/net/hyperv/netvsc_drv.c struct netvsc_device *nvdev = rtnl_dereference(ndev_ctx->nvdev); ndev_ctx 499 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 509 drivers/net/hyperv/netvsc_drv.c = this_cpu_ptr(ndev_ctx->vf_stats); ndev_ctx 516 drivers/net/hyperv/netvsc_drv.c this_cpu_inc(ndev_ctx->vf_stats->tx_dropped); ndev_ctx 719 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 729 drivers/net/hyperv/netvsc_drv.c ndev_ctx->speed = speed; ndev_ctx 747 drivers/net/hyperv/netvsc_drv.c spin_lock_irqsave(&ndev_ctx->lock, flags); ndev_ctx 748 drivers/net/hyperv/netvsc_drv.c list_add_tail(&event->list, &ndev_ctx->reconfig_events); ndev_ctx 749 drivers/net/hyperv/netvsc_drv.c spin_unlock_irqrestore(&ndev_ctx->lock, flags); ndev_ctx 751 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, 0); ndev_ctx 918 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 919 drivers/net/hyperv/netvsc_drv.c struct hv_device *hdev = ndev_ctx->device_ctx; ndev_ctx 955 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 956 drivers/net/hyperv/netvsc_drv.c struct hv_device *hdev = ndev_ctx->device_ctx; ndev_ctx 1160 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 1167 drivers/net/hyperv/netvsc_drv.c = per_cpu_ptr(ndev_ctx->vf_stats, i); ndev_ctx 1190 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 1191 drivers/net/hyperv/netvsc_drv.c struct netvsc_device *nvdev = rcu_dereference_rtnl(ndev_ctx->nvdev); ndev_ctx 1197 drivers/net/hyperv/netvsc_drv.c per_cpu_ptr(ndev_ctx->vf_stats, i); ndev_ctx 1248 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(net); ndev_ctx 1255 drivers/net/hyperv/netvsc_drv.c nvdev = rcu_dereference(ndev_ctx->nvdev); ndev_ctx 1834 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 1836 drivers/net/hyperv/netvsc_drv.c return ndev_ctx->msg_enable; ndev_ctx 1841 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 1843 drivers/net/hyperv/netvsc_drv.c ndev_ctx->msg_enable = val; ndev_ctx 1890 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = ndev_ctx 1892 drivers/net/hyperv/netvsc_drv.c struct hv_device *device_obj = ndev_ctx->device_ctx; ndev_ctx 1902 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); ndev_ctx 1906 drivers/net/hyperv/netvsc_drv.c net_device = rtnl_dereference(ndev_ctx->nvdev); ndev_ctx 1912 drivers/net/hyperv/netvsc_drv.c next_reconfig = ndev_ctx->last_reconfig + LINKCHANGE_INT; ndev_ctx 1920 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, delay); ndev_ctx 1923 drivers/net/hyperv/netvsc_drv.c ndev_ctx->last_reconfig = jiffies; ndev_ctx 1925 drivers/net/hyperv/netvsc_drv.c spin_lock_irqsave(&ndev_ctx->lock, flags); ndev_ctx 1926 drivers/net/hyperv/netvsc_drv.c if (!list_empty(&ndev_ctx->reconfig_events)) { ndev_ctx 1927 drivers/net/hyperv/netvsc_drv.c event = list_first_entry(&ndev_ctx->reconfig_events, ndev_ctx 1930 drivers/net/hyperv/netvsc_drv.c reschedule = !list_empty(&ndev_ctx->reconfig_events); ndev_ctx 1932 drivers/net/hyperv/netvsc_drv.c spin_unlock_irqrestore(&ndev_ctx->lock, flags); ndev_ctx 1966 drivers/net/hyperv/netvsc_drv.c spin_lock_irqsave(&ndev_ctx->lock, flags); ndev_ctx 1967 drivers/net/hyperv/netvsc_drv.c list_add(&event->list, &ndev_ctx->reconfig_events); ndev_ctx 1968 drivers/net/hyperv/netvsc_drv.c spin_unlock_irqrestore(&ndev_ctx->lock, flags); ndev_ctx 1983 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->dwork, LINKCHANGE_INT); ndev_ctx 2015 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 2017 drivers/net/hyperv/netvsc_drv.c = this_cpu_ptr(ndev_ctx->vf_stats); ndev_ctx 2038 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 2062 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->vf_takeover, VF_TAKEOVER_INT); ndev_ctx 2108 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx ndev_ctx 2110 drivers/net/hyperv/netvsc_drv.c struct net_device *ndev = hv_get_drvdata(ndev_ctx->device_ctx); ndev_ctx 2114 drivers/net/hyperv/netvsc_drv.c schedule_delayed_work(&ndev_ctx->vf_takeover, 0); ndev_ctx 2118 drivers/net/hyperv/netvsc_drv.c vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); ndev_ctx 2131 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx; ndev_ctx 2150 drivers/net/hyperv/netvsc_drv.c list_for_each_entry(ndev_ctx, &netvsc_dev_list, list) { ndev_ctx 2151 drivers/net/hyperv/netvsc_drv.c if (!ndev_ctx->vf_alloc) ndev_ctx 2154 drivers/net/hyperv/netvsc_drv.c if (ndev_ctx->vf_serial == serial) ndev_ctx 2155 drivers/net/hyperv/netvsc_drv.c return hv_get_drvdata(ndev_ctx->device_ctx); ndev_ctx 2384 drivers/net/hyperv/netvsc_drv.c struct net_device_context *ndev_ctx; ndev_ctx 2394 drivers/net/hyperv/netvsc_drv.c ndev_ctx = netdev_priv(net); ndev_ctx 2396 drivers/net/hyperv/netvsc_drv.c cancel_delayed_work_sync(&ndev_ctx->dwork); ndev_ctx 2399 drivers/net/hyperv/netvsc_drv.c nvdev = rtnl_dereference(ndev_ctx->nvdev); ndev_ctx 2407 drivers/net/hyperv/netvsc_drv.c vf_netdev = rtnl_dereference(ndev_ctx->vf_netdev); ndev_ctx 2415 drivers/net/hyperv/netvsc_drv.c list_del(&ndev_ctx->list); ndev_ctx 2421 drivers/net/hyperv/netvsc_drv.c free_percpu(ndev_ctx->vf_stats); ndev_ctx 1088 drivers/net/hyperv/rndis_filter.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 1097 drivers/net/hyperv/rndis_filter.c nvscdev = rcu_dereference_raw(ndev_ctx->nvdev); ndev_ctx 1132 drivers/net/hyperv/rndis_filter.c struct net_device_context *ndev_ctx = netdev_priv(ndev); ndev_ctx 1133 drivers/net/hyperv/rndis_filter.c struct hv_device *hv_dev = ndev_ctx->device_ctx; ndev_ctx 1170 drivers/net/hyperv/rndis_filter.c ndev_ctx->tx_table[i] = i % nvdev->num_chn;