Lines Matching refs:vpaths

127 		VXGE_COMPLETE_VPATH_TX(&vdev->vpaths[i].fifo);  in VXGE_COMPLETE_ALL_TX()
137 ring = &vdev->vpaths[i].ring; in VXGE_COMPLETE_ALL_RX()
710 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_add_mac_addr()
738 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
748 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
764 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
860 fifo = &vdev->vpaths[vpath_no].fifo; in vxge_xmit()
1097 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_del_mac_addr()
1146 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1156 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1169 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1188 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1189 list_head = &vdev->vpaths[0].mac_addr_list; in vxge_set_multicast()
1191 (vdev->vpaths[0].mac_addr_cnt - mcast_cnt)) > in vxge_set_multicast()
1192 vdev->vpaths[0].max_mac_addr_cnt) in vxge_set_multicast()
1236 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1258 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1314 struct vxge_vpath *vpath = &vdev->vpaths[vpath_idx]; in vxge_set_mac_addr()
1360 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_enable()
1393 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_disable()
1512 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_reset_vpath()
1589 hw_ring = vdev->vpaths[i].ring.handle; in vxge_config_ci_for_tti_rti()
1596 struct __vxge_hw_fifo *hw_fifo = vdev->vpaths[i].fifo.handle; in vxge_config_ci_for_tti_rti()
1726 if (vdev->vpaths[i].handle) { in do_vxge_reset()
1728 vdev->vpaths[i].handle) in do_vxge_reset()
1749 vxge_restore_vpath_mac_addr(&vdev->vpaths[vp_id]); in do_vxge_reset()
1750 vxge_restore_vpath_vid_table(&vdev->vpaths[vp_id]); in do_vxge_reset()
1766 vxge_hw_vpath_enable(vdev->vpaths[i].handle); in do_vxge_reset()
1768 vxge_hw_vpath_rx_doorbell_init(vdev->vpaths[i].handle); in do_vxge_reset()
1853 ring = &vdev->vpaths[i].ring; in vxge_poll_inta()
1938 "for vpath:%d", vdev->vpaths[0].device_id); in vxge_rth_configure()
1959 vdev->vpaths[index].handle, in vxge_rth_configure()
1966 vdev->vpaths[index].device_id); in vxge_rth_configure()
1982 vpath = &vdev->vpaths[i]; in vxge_reset_all_vpaths()
2013 vpath = &vdev->vpaths[i]; in vxge_close_vpaths()
2034 vpath = &vdev->vpaths[i]; in vxge_open_vpaths()
2281 vxge_hw_vpath_msix_mask(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2282 vxge_hw_vpath_msix_clear(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2285 status = vxge_hw_vpath_alarm_process(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2288 vxge_hw_vpath_msix_unmask(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2404 struct vxge_vpath *vpath = &vdev->vpaths[i]; in vxge_enable_msix()
2497 &vdev->vpaths[vp_idx].fifo); in vxge_add_isr()
2499 &vdev->vpaths[vp_idx].fifo; in vxge_add_isr()
2513 &vdev->vpaths[vp_idx].ring); in vxge_add_isr()
2515 &vdev->vpaths[vp_idx].ring; in vxge_add_isr()
2535 msix_idx += vdev->vpaths[vp_idx].device_id * in vxge_add_isr()
2538 vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2559 &vdev->vpaths[0]); in vxge_add_isr()
2572 msix_idx = (vdev->vpaths[0].handle->vpath->vp_id * in vxge_add_isr()
2574 vxge_hw_vpath_msix_unmask(vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2577 vdev->vxge_entries[intr_cnt].arg = &vdev->vpaths[0]; in vxge_add_isr()
2588 vxge_hw_vpath_tti_ci_set(vdev->vpaths[0].fifo.handle); in vxge_add_isr()
2636 ring = &vdev->vpaths[i].ring; in vxge_poll_vp_lockup()
2651 vpath = &vdev->vpaths[i]; in vxge_poll_vp_lockup()
2760 vpath = &vdev->vpaths[i]; in vxge_open()
2765 vpath = &vdev->vpaths[i]; in vxge_open()
2788 vpath = &vdev->vpaths[i]; in vxge_open()
2810 vpath = &vdev->vpaths[i]; in vxge_open()
2848 vpath = &vdev->vpaths[i]; in vxge_open()
2891 vpath = &vdev->vpaths[i]; in vxge_open()
2913 napi_disable(&vdev->vpaths[i].ring.napi); in vxge_open()
2946 netif_napi_del(&vdev->vpaths[i].ring.napi); in vxge_napi_del_all()
2973 vpath_vector = vxge_mBIT(vdev->vpaths[0].device_id); in do_vxge_close()
3025 napi_disable(&vdev->vpaths[i].ring.napi); in do_vxge_close()
3134 struct vxge_ring_stats *rxstats = &vdev->vpaths[k].ring.stats; in vxge_get_stats64()
3135 struct vxge_fifo_stats *txstats = &vdev->vpaths[k].fifo.stats; in vxge_get_stats64()
3247 vdev->vpaths[i].ring.rx_hwts = vdev->rx_hwts; in vxge_hwtstamp_set()
3334 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_add_vid()
3362 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_kill_vid()
3453 vdev->vpaths = kzalloc((sizeof(struct vxge_vpath)) * in vxge_device_register()
3455 if (!vdev->vpaths) { in vxge_device_register()
3516 kfree(vdev->vpaths); in vxge_device_register()
3547 kfree(vdev->vpaths); in vxge_device_unregister()
3582 vpath = &vdev->vpaths[vpath_idx]; in vxge_callback_crit_err()
4627 vdev->vpaths[j].is_configured = 1; in vxge_probe()
4628 vdev->vpaths[j].device_id = i; in vxge_probe()
4629 vdev->vpaths[j].ring.driver_id = j; in vxge_probe()
4630 vdev->vpaths[j].vdev = vdev; in vxge_probe()
4631 vdev->vpaths[j].max_mac_addr_cnt = max_mac_vpath; in vxge_probe()
4632 memcpy((u8 *)vdev->vpaths[j].macaddr, in vxge_probe()
4637 INIT_LIST_HEAD(&vdev->vpaths[j].mac_addr_list); in vxge_probe()
4639 vdev->vpaths[j].mac_addr_cnt = 0; in vxge_probe()
4640 vdev->vpaths[j].mcast_addr_cnt = 0; in vxge_probe()
4652 macaddr = (u8 *)vdev->vpaths[0].macaddr; in vxge_probe()
4703 memcpy(vdev->ndev->dev_addr, (u8 *)vdev->vpaths[0].macaddr, ETH_ALEN); in vxge_probe()
4717 list_add(&entry->item, &vdev->vpaths[i].mac_addr_list); in vxge_probe()
4718 vdev->vpaths[i].mac_addr_cnt = 1; in vxge_probe()
4755 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_probe()
4798 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_remove()