Lines Matching refs:mdev

196 		rc = mlx4_flow_detach(priv->mdev->dev, filter->reg_id);  in mlx4_en_filter_work()
201 rc = mlx4_flow_attach(priv->mdev->dev, &rule, &filter->reg_id); in mlx4_en_filter_work()
267 rc = mlx4_flow_detach(priv->mdev->dev, filter->reg_id); in mlx4_en_filter_free()
349 queue_work(priv->mdev->workqueue, &filter->work); in mlx4_en_filter_rfs()
415 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_vlan_rx_add_vid() local
424 mutex_lock(&mdev->state_lock); in mlx4_en_vlan_rx_add_vid()
425 if (mdev->device_up && priv->port_up) { in mlx4_en_vlan_rx_add_vid()
426 err = mlx4_SET_VLAN_FLTR(mdev->dev, priv); in mlx4_en_vlan_rx_add_vid()
430 if (mlx4_register_vlan(mdev->dev, priv->port, vid, &idx)) in mlx4_en_vlan_rx_add_vid()
432 mutex_unlock(&mdev->state_lock); in mlx4_en_vlan_rx_add_vid()
441 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_vlan_rx_kill_vid() local
449 mutex_lock(&mdev->state_lock); in mlx4_en_vlan_rx_kill_vid()
450 mlx4_unregister_vlan(mdev->dev, priv->port, vid); in mlx4_en_vlan_rx_kill_vid()
452 if (mdev->device_up && priv->port_up) { in mlx4_en_vlan_rx_kill_vid()
453 err = mlx4_SET_VLAN_FLTR(mdev->dev, priv); in mlx4_en_vlan_rx_kill_vid()
457 mutex_unlock(&mdev->state_lock); in mlx4_en_vlan_rx_kill_vid()
478 if (priv->mdev->dev->caps.tunnel_offload_mode != MLX4_TUNNEL_OFFLOAD_MODE_VXLAN || in mlx4_en_tunnel_steer_add()
479 priv->mdev->dev->caps.dmfs_high_steer_mode == MLX4_STEERING_DMFS_A0_STATIC) in mlx4_en_tunnel_steer_add()
482 err = mlx4_tunnel_steer_add(priv->mdev->dev, addr, priv->port, qpn, in mlx4_en_tunnel_steer_add()
496 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_uc_steer_add() local
497 struct mlx4_dev *dev = mdev->dev; in mlx4_en_uc_steer_add()
548 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_uc_steer_release() local
549 struct mlx4_dev *dev = mdev->dev; in mlx4_en_uc_steer_release()
574 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_get_qp() local
575 struct mlx4_dev *dev = mdev->dev; in mlx4_en_get_qp()
631 mlx4_flow_detach(priv->mdev->dev, priv->tunnel_reg_id); in mlx4_en_get_qp()
645 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_put_qp() local
646 struct mlx4_dev *dev = mdev->dev; in mlx4_en_put_qp()
677 mlx4_flow_detach(priv->mdev->dev, priv->tunnel_reg_id); in mlx4_en_put_qp()
691 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_replace_mac() local
692 struct mlx4_dev *dev = mdev->dev; in mlx4_en_replace_mac()
724 mlx4_flow_detach(priv->mdev->dev, priv->tunnel_reg_id); in mlx4_en_replace_mac()
761 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_set_mac() local
769 mutex_lock(&mdev->state_lock); in mlx4_en_set_mac()
774 mutex_unlock(&mdev->state_lock); in mlx4_en_set_mac()
862 queue_work(priv->mdev->workqueue, &priv->rx_mode_task); in mlx4_en_set_rx_mode()
866 struct mlx4_en_dev *mdev) in mlx4_en_set_promisc_mode() argument
876 switch (mdev->dev->caps.steering_mode) { in mlx4_en_set_promisc_mode()
878 err = mlx4_flow_steer_promisc_add(mdev->dev, in mlx4_en_set_promisc_mode()
888 err = mlx4_unicast_promisc_add(mdev->dev, in mlx4_en_set_promisc_mode()
898 err = mlx4_multicast_promisc_add(mdev->dev, in mlx4_en_set_promisc_mode()
908 err = mlx4_SET_PORT_qpn_calc(mdev->dev, in mlx4_en_set_promisc_mode()
918 err = mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, 0, in mlx4_en_set_promisc_mode()
926 struct mlx4_en_dev *mdev) in mlx4_en_clear_promisc_mode() argument
935 switch (mdev->dev->caps.steering_mode) { in mlx4_en_clear_promisc_mode()
937 err = mlx4_flow_steer_promisc_remove(mdev->dev, in mlx4_en_clear_promisc_mode()
946 err = mlx4_unicast_promisc_remove(mdev->dev, in mlx4_en_clear_promisc_mode()
953 err = mlx4_multicast_promisc_remove(mdev->dev, in mlx4_en_clear_promisc_mode()
963 err = mlx4_SET_PORT_qpn_calc(mdev->dev, in mlx4_en_clear_promisc_mode()
974 struct mlx4_en_dev *mdev) in mlx4_en_do_multicast() argument
983 err = mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, 0, in mlx4_en_do_multicast()
990 switch (mdev->dev->caps.steering_mode) { in mlx4_en_do_multicast()
992 err = mlx4_flow_steer_promisc_add(mdev->dev, in mlx4_en_do_multicast()
999 err = mlx4_multicast_promisc_add(mdev->dev, in mlx4_en_do_multicast()
1014 switch (mdev->dev->caps.steering_mode) { in mlx4_en_do_multicast()
1016 err = mlx4_flow_steer_promisc_remove(mdev->dev, in mlx4_en_do_multicast()
1022 err = mlx4_multicast_promisc_remove(mdev->dev, in mlx4_en_do_multicast()
1035 err = mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, 0, in mlx4_en_do_multicast()
1041 mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, ETH_BCAST, in mlx4_en_do_multicast()
1051 mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, in mlx4_en_do_multicast()
1054 err = mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, 0, in mlx4_en_do_multicast()
1065 err = mlx4_multicast_detach(mdev->dev, in mlx4_en_do_multicast()
1074 err = mlx4_flow_detach(priv->mdev->dev, mclist->tunnel_reg_id); in mlx4_en_do_multicast()
1087 err = mlx4_multicast_attach(mdev->dev, in mlx4_en_do_multicast()
1107 struct mlx4_en_dev *mdev) in mlx4_en_do_uc_filter() argument
1147 mlx4_unregister_mac(mdev->dev, priv->port, mac); in mlx4_en_do_uc_filter()
1188 err = mlx4_register_mac(mdev->dev, priv->port, mac); in mlx4_en_do_uc_filter()
1202 mlx4_unregister_mac(mdev->dev, priv->port, mac); in mlx4_en_do_uc_filter()
1230 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_do_set_rx_mode() local
1233 mutex_lock(&mdev->state_lock); in mlx4_en_do_set_rx_mode()
1234 if (!mdev->device_up) { in mlx4_en_do_set_rx_mode()
1244 if (!mlx4_en_QUERY_PORT(mdev, priv->port)) { in mlx4_en_do_set_rx_mode()
1254 mlx4_en_do_uc_filter(priv, dev, mdev); in mlx4_en_do_set_rx_mode()
1259 mlx4_en_set_promisc_mode(priv, mdev); in mlx4_en_do_set_rx_mode()
1265 mlx4_en_clear_promisc_mode(priv, mdev); in mlx4_en_do_set_rx_mode()
1267 mlx4_en_do_multicast(priv, dev, mdev); in mlx4_en_do_set_rx_mode()
1269 mutex_unlock(&mdev->state_lock); in mlx4_en_do_set_rx_mode()
1289 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_tx_timeout() local
1305 queue_work(mdev->workqueue, &priv->watchdog_task); in mlx4_en_tx_timeout()
1433 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_do_get_stats() local
1436 mutex_lock(&mdev->state_lock); in mlx4_en_do_get_stats()
1437 if (mdev->device_up) { in mlx4_en_do_get_stats()
1439 err = mlx4_en_DUMP_ETH_STATS(mdev, priv->port, 0); in mlx4_en_do_get_stats()
1446 queue_delayed_work(mdev->workqueue, &priv->stats_task, STATS_DELAY); in mlx4_en_do_get_stats()
1448 if (mdev->mac_removed[MLX4_MAX_PORTS + 1 - priv->port]) { in mlx4_en_do_get_stats()
1450 mdev->mac_removed[MLX4_MAX_PORTS + 1 - priv->port] = 0; in mlx4_en_do_get_stats()
1452 mutex_unlock(&mdev->state_lock); in mlx4_en_do_get_stats()
1463 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_service_task() local
1465 mutex_lock(&mdev->state_lock); in mlx4_en_service_task()
1466 if (mdev->device_up) { in mlx4_en_service_task()
1467 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS) in mlx4_en_service_task()
1468 mlx4_en_ptp_overflow_check(mdev); in mlx4_en_service_task()
1471 queue_delayed_work(mdev->workqueue, &priv->service_task, in mlx4_en_service_task()
1474 mutex_unlock(&mdev->state_lock); in mlx4_en_service_task()
1481 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_linkstate() local
1484 mutex_lock(&mdev->state_lock); in mlx4_en_linkstate()
1497 mutex_unlock(&mdev->state_lock); in mlx4_en_linkstate()
1503 int numa_node = priv->mdev->dev->numa_node; in mlx4_en_init_affinity_hint()
1521 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_start_port() local
1598 mdev->mac_removed[priv->port] = 0; in mlx4_en_start_port()
1649 err = mlx4_SET_PORT_general(mdev->dev, priv->port, in mlx4_en_start_port()
1661 err = mlx4_SET_PORT_qpn_calc(mdev->dev, priv->port, priv->base_qpn, 0); in mlx4_en_start_port()
1667 if (mdev->dev->caps.tunnel_offload_mode == MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) { in mlx4_en_start_port()
1668 err = mlx4_SET_PORT_VXLAN(mdev->dev, priv->port, VXLAN_STEER_BY_OUTER_MAC, 1); in mlx4_en_start_port()
1678 err = mlx4_INIT_PORT(mdev->dev, priv->port); in mlx4_en_start_port()
1687 if (mlx4_multicast_attach(mdev->dev, &priv->rss_map.indir_qp, mc_list, in mlx4_en_start_port()
1690 mlx4_warn(mdev, "Failed Attaching Broadcast\n"); in mlx4_en_start_port()
1696 queue_work(mdev->workqueue, &priv->rx_mode_task); in mlx4_en_start_port()
1699 if (priv->mdev->dev->caps.tunnel_offload_mode == MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) in mlx4_en_start_port()
1733 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_stop_port() local
1745 mlx4_CLOSE_PORT(mdev->dev, priv->port); in mlx4_en_stop_port()
1760 if (mdev->dev->caps.steering_mode == in mlx4_en_stop_port()
1764 mlx4_flow_steer_promisc_remove(mdev->dev, in mlx4_en_stop_port()
1767 mlx4_flow_steer_promisc_remove(mdev->dev, in mlx4_en_stop_port()
1774 mlx4_unicast_promisc_remove(mdev->dev, priv->base_qpn, in mlx4_en_stop_port()
1779 mlx4_multicast_promisc_remove(mdev->dev, priv->base_qpn, in mlx4_en_stop_port()
1788 mlx4_multicast_detach(mdev->dev, &priv->rss_map.indir_qp, mc_list, in mlx4_en_stop_port()
1793 mlx4_multicast_detach(mdev->dev, &priv->rss_map.indir_qp, in mlx4_en_stop_port()
1796 mlx4_flow_detach(mdev->dev, mclist->tunnel_reg_id); in mlx4_en_stop_port()
1805 mlx4_SET_MCAST_FLTR(mdev->dev, priv->port, 0, 1, MLX4_MCAST_CONFIG); in mlx4_en_stop_port()
1808 if (mdev->dev->caps.steering_mode == in mlx4_en_stop_port()
1813 mlx4_flow_detach(mdev->dev, flow->id); in mlx4_en_stop_port()
1835 if (!(mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_REASSIGN_MAC_EN)) in mlx4_en_stop_port()
1836 mdev->mac_removed[priv->port] = 1; in mlx4_en_stop_port()
1861 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_restart() local
1866 mutex_lock(&mdev->state_lock); in mlx4_en_restart()
1872 mutex_unlock(&mdev->state_lock); in mlx4_en_restart()
1878 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_clear_stats() local
1881 if (mlx4_en_DUMP_ETH_STATS(mdev, priv->port, 1)) in mlx4_en_clear_stats()
1912 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_open() local
1915 mutex_lock(&mdev->state_lock); in mlx4_en_open()
1917 if (!mdev->device_up) { in mlx4_en_open()
1931 mutex_unlock(&mdev->state_lock); in mlx4_en_open()
1939 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_close() local
1943 mutex_lock(&mdev->state_lock); in mlx4_en_close()
1948 mutex_unlock(&mdev->state_lock); in mlx4_en_close()
2011 if (priv->mdev->dev->caps.comp_pool) { in mlx4_en_alloc_resources()
2012 priv->dev->rx_cpu_rmap = alloc_irq_cpu_rmap(priv->mdev->dev->caps.comp_pool); in mlx4_en_alloc_resources()
2043 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_destroy_netdev() local
2052 mlx4_free_hwq_res(mdev->dev, &priv->res, MLX4_EN_PAGE_SIZE); in mlx4_en_destroy_netdev()
2057 flush_workqueue(mdev->workqueue); in mlx4_en_destroy_netdev()
2060 mutex_lock(&mdev->state_lock); in mlx4_en_destroy_netdev()
2061 mdev->pndev[priv->port] = NULL; in mlx4_en_destroy_netdev()
2062 mdev->upper[priv->port] = NULL; in mlx4_en_destroy_netdev()
2063 mutex_unlock(&mdev->state_lock); in mlx4_en_destroy_netdev()
2076 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_change_mtu() local
2089 mutex_lock(&mdev->state_lock); in mlx4_en_change_mtu()
2090 if (!mdev->device_up) { in mlx4_en_change_mtu()
2100 queue_work(mdev->workqueue, &priv->watchdog_task); in mlx4_en_change_mtu()
2103 mutex_unlock(&mdev->state_lock); in mlx4_en_change_mtu()
2111 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_hwtstamp_set() local
2122 if (!(mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS)) in mlx4_en_hwtstamp_set()
2205 ret = mlx4_SET_PORT_fcs_check(priv->mdev->dev, in mlx4_en_set_features()
2240 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_set_vf_mac() local
2246 return mlx4_set_vf_mac(mdev->dev, en_priv->port, queue, mac_u64); in mlx4_en_set_vf_mac()
2252 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_set_vf_vlan() local
2254 return mlx4_set_vf_vlan(mdev->dev, en_priv->port, vf, vlan, qos); in mlx4_en_set_vf_vlan()
2261 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_set_vf_rate() local
2263 return mlx4_set_vf_rate(mdev->dev, en_priv->port, vf, min_tx_rate, in mlx4_en_set_vf_rate()
2270 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_set_vf_spoofchk() local
2272 return mlx4_set_vf_spoofchk(mdev->dev, en_priv->port, vf, setting); in mlx4_en_set_vf_spoofchk()
2278 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_get_vf_config() local
2280 return mlx4_get_vf_config(mdev->dev, en_priv->port, vf, ivf); in mlx4_en_get_vf_config()
2286 struct mlx4_en_dev *mdev = en_priv->mdev; in mlx4_en_set_vf_link_state() local
2288 return mlx4_set_vf_link_state(mdev->dev, en_priv->port, vf, link_state); in mlx4_en_set_vf_link_state()
2296 struct mlx4_dev *mdev = priv->mdev->dev; in mlx4_en_get_phys_port_id() local
2298 u64 phys_port_id = mdev->caps.phys_port_id[priv->port]; in mlx4_en_get_phys_port_id()
2318 ret = mlx4_config_vxlan_port(priv->mdev->dev, priv->vxlan_port); in mlx4_en_add_vxlan_offloads()
2322 ret = mlx4_SET_PORT_VXLAN(priv->mdev->dev, priv->port, in mlx4_en_add_vxlan_offloads()
2344 ret = mlx4_SET_PORT_VXLAN(priv->mdev->dev, priv->port, in mlx4_en_del_vxlan_offloads()
2358 if (priv->mdev->dev->caps.tunnel_offload_mode != MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) in mlx4_en_add_vxlan_port()
2372 queue_work(priv->mdev->workqueue, &priv->vxlan_add_task); in mlx4_en_add_vxlan_port()
2381 if (priv->mdev->dev->caps.tunnel_offload_mode != MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) in mlx4_en_del_vxlan_port()
2393 queue_work(priv->mdev->workqueue, &priv->vxlan_del_task); in mlx4_en_del_vxlan_port()
2412 if (!(priv->mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_QP_RATE_LIMIT)) in mlx4_en_set_tx_maxrate()
2427 err = mlx4_update_qp(priv->mdev->dev, tx_ring->qpn, MLX4_UPDATE_QP_RATE_LIMIT, in mlx4_en_set_tx_maxrate()
2515 struct mlx4_dev *dev = bond->priv->mdev->dev; in mlx4_en_bond_work()
2555 queue_work(priv->mdev->workqueue, &bond->work); in mlx4_en_queue_bond_work()
2564 struct mlx4_en_dev *mdev; in mlx4_en_netdev_event() local
2575 mdev = container_of(this, struct mlx4_en_dev, nb); in mlx4_en_netdev_event()
2576 dev = mdev->dev; in mlx4_en_netdev_event()
2583 if (!port && (mdev->pndev[i] == ndev)) in mlx4_en_netdev_event()
2585 mdev->upper[i] = mdev->pndev[i] ? in mlx4_en_netdev_event()
2586 netdev_master_upper_dev_get(mdev->pndev[i]) : NULL; in mlx4_en_netdev_event()
2588 if (!mdev->upper[i]) in mlx4_en_netdev_event()
2593 if (mdev->upper[i] != mdev->upper[i-1]) in mlx4_en_netdev_event()
2752 int mlx4_en_init_netdev(struct mlx4_en_dev *mdev, int port, in mlx4_en_init_netdev() argument
2769 SET_NETDEV_DEV(dev, &mdev->dev->persist->pdev->dev); in mlx4_en_init_netdev()
2794 priv->mdev = mdev; in mlx4_en_init_netdev()
2795 priv->ddev = &mdev->pdev->dev; in mlx4_en_init_netdev()
2803 priv->num_tx_rings_p_up = mdev->profile.num_tx_rings_p_up; in mlx4_en_init_netdev()
2821 priv->cqe_factor = (mdev->dev->caps.cqe_size == 64) ? 1 : 0; in mlx4_en_init_netdev()
2822 priv->cqe_size = mdev->dev->caps.cqe_size; in mlx4_en_init_netdev()
2826 if (!mlx4_is_slave(priv->mdev->dev)) { in mlx4_en_init_netdev()
2827 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_ETS_CFG) { in mlx4_en_init_netdev()
2840 priv->max_mtu = mdev->dev->caps.eth_mtu_cap[priv->port]; in mlx4_en_init_netdev()
2842 if (mdev->dev->caps.rx_checksum_flags_port[priv->port] & in mlx4_en_init_netdev()
2848 mlx4_en_u64_to_mac(dev->dev_addr, mdev->dev->caps.def_mac[priv->port]); in mlx4_en_init_netdev()
2850 if (mlx4_is_slave(priv->mdev->dev)) { in mlx4_en_init_netdev()
2854 mdev->dev->caps.def_mac[priv->port] = mac_u64; in mlx4_en_init_netdev()
2877 err = mlx4_alloc_hwq_res(mdev->dev, &priv->res, in mlx4_en_init_netdev()
2888 if (mlx4_is_master(priv->mdev->dev)) in mlx4_en_init_netdev()
2902 if (mdev->LSO_support) in mlx4_en_init_netdev()
2914 if (mdev->dev->caps.flags & MLX4_DEV_CAP_FLAG_FCS_KEEP) in mlx4_en_init_netdev()
2917 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_IGNORE_FCS) in mlx4_en_init_netdev()
2920 if (mdev->dev->caps.steering_mode == in mlx4_en_init_netdev()
2922 mdev->dev->caps.dmfs_high_steer_mode != MLX4_STEERING_DMFS_A0_STATIC) in mlx4_en_init_netdev()
2925 if (mdev->dev->caps.steering_mode != MLX4_STEERING_MODE_A0) in mlx4_en_init_netdev()
2929 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_RSS_TOP) { in mlx4_en_init_netdev()
2931 } else if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_RSS_XOR) { in mlx4_en_init_netdev()
2939 if (mdev->dev->caps.tunnel_offload_mode == MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) { in mlx4_en_init_netdev()
2944 mdev->pndev[port] = dev; in mlx4_en_init_netdev()
2945 mdev->upper[port] = NULL; in mlx4_en_init_netdev()
2957 err = mlx4_SET_PORT_general(mdev->dev, priv->port, in mlx4_en_init_netdev()
2967 if (mdev->dev->caps.tunnel_offload_mode == MLX4_TUNNEL_OFFLOAD_MODE_VXLAN) { in mlx4_en_init_netdev()
2968 err = mlx4_SET_PORT_VXLAN(mdev->dev, priv->port, VXLAN_STEER_BY_OUTER_MAC, 1); in mlx4_en_init_netdev()
2978 err = mlx4_INIT_PORT(mdev->dev, priv->port); in mlx4_en_init_netdev()
2983 queue_delayed_work(mdev->workqueue, &priv->stats_task, STATS_DELAY); in mlx4_en_init_netdev()
2985 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS) in mlx4_en_init_netdev()
2986 queue_delayed_work(mdev->workqueue, &priv->service_task, in mlx4_en_init_netdev()
2989 mlx4_en_set_stats_bitmap(mdev->dev, &priv->stats_bitmap, in mlx4_en_init_netdev()
2990 mdev->profile.prof[priv->port].rx_ppp, in mlx4_en_init_netdev()
2991 mdev->profile.prof[priv->port].rx_pause, in mlx4_en_init_netdev()
2992 mdev->profile.prof[priv->port].tx_ppp, in mlx4_en_init_netdev()
2993 mdev->profile.prof[priv->port].tx_pause); in mlx4_en_init_netdev()
3015 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_reset_config() local
3032 mutex_lock(&mdev->state_lock); in mlx4_en_reset_config()
3090 mutex_unlock(&mdev->state_lock); in mlx4_en_reset_config()