Lines Matching refs:vpath

25 	status = __vxge_hw_vpath_stats_access(vpath,			\
50 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_wait_receive_idle() local
54 vpath = &hldev->virtual_paths[vp_id]; in vxge_hw_vpath_wait_receive_idle()
55 vp_reg = vpath->vp_reg; in vxge_hw_vpath_wait_receive_idle()
157 vxge_hw_vpath_fw_api(struct __vxge_hw_virtualpath *vpath, u32 action, in vxge_hw_vpath_fw_api() argument
161 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in vxge_hw_vpath_fw_api()
166 spin_lock(&vpath->lock); in vxge_hw_vpath_fw_api()
167 if (!vpath->vp_open) { in vxge_hw_vpath_fw_api()
168 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
196 if (!vpath->vp_open) in vxge_hw_vpath_fw_api()
216 if (vpath->vp_open) in vxge_hw_vpath_fw_api()
217 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
226 struct __vxge_hw_virtualpath *vpath; in vxge_hw_upgrade_read_version() local
229 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_upgrade_read_version()
231 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_upgrade_read_version()
249 struct __vxge_hw_virtualpath *vpath; in vxge_hw_flash_fw() local
253 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_flash_fw()
255 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_flash_fw()
280 struct __vxge_hw_virtualpath *vpath; in vxge_update_fw_image() local
284 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_update_fw_image()
287 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
306 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
377 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_eprom_img_ver_get() local
381 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_vpath_eprom_img_ver_get()
387 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_vpath_eprom_img_ver_get()
426 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_channel_initialize() local
428 vpath = channel->vph->vpath; in __vxge_hw_channel_initialize()
437 vpath->fifoh = (struct __vxge_hw_fifo *)channel; in __vxge_hw_channel_initialize()
442 vpath->ringh = (struct __vxge_hw_ring *)channel; in __vxge_hw_channel_initialize()
804 __vxge_hw_vpath_fw_ver_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_fw_ver_get() argument
814 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_fw_ver_get()
871 __vxge_hw_vpath_card_info_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_card_info_get() argument
883 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
896 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
911 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
930 __vxge_hw_vpath_pci_func_mode_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_func_mode_get() argument
938 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_pci_func_mode_get()
954 __vxge_hw_vpath_addr_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_addr_get() argument
963 status = vxge_hw_vpath_fw_api(vpath, action, in __vxge_hw_vpath_addr_get()
1006 struct __vxge_hw_virtualpath vpath; in vxge_hw_device_hw_info_get() local
1054 spin_lock_init(&vpath.lock); in vxge_hw_device_hw_info_get()
1055 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1056 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1058 status = __vxge_hw_vpath_pci_func_mode_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1062 status = __vxge_hw_vpath_fw_ver_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1066 status = __vxge_hw_vpath_card_info_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1078 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1079 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1081 status = __vxge_hw_vpath_addr_get(&vpath, in vxge_hw_device_hw_info_get()
1413 __vxge_hw_vpath_stats_access(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_access() argument
1420 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_access()
1425 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_access()
1434 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_stats_access()
1448 __vxge_hw_vpath_xmac_tx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_tx_stats_get() argument
1458 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_tx_stats_get()
1464 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_tx_stats_get()
1480 __vxge_hw_vpath_xmac_rx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_rx_stats_get() argument
1489 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_rx_stats_get()
1494 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_rx_stats_get()
1511 __vxge_hw_vpath_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_get() argument
1518 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_get()
1522 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_get()
1582 status = __vxge_hw_vpath_xmac_tx_stats_get(vpath, &hw_stats->tx_stats); in __vxge_hw_vpath_stats_get()
1586 status = __vxge_hw_vpath_xmac_rx_stats_get(vpath, &hw_stats->rx_stats); in __vxge_hw_vpath_stats_get()
2194 hldev = vph->vpath->hldev; in __vxge_hw_channel_allocate()
2195 vp_id = vph->vpath->vp_id; in __vxge_hw_channel_allocate()
2788 struct __vxge_hw_ring *ring = vp->vpath->ringh; in __vxge_hw_ring_delete()
2795 vp->vpath->ringh = NULL; in __vxge_hw_ring_delete()
2824 hldev = vp->vpath->hldev; in __vxge_hw_ring_create()
2825 vp_id = vp->vpath->vp_id; in __vxge_hw_ring_create()
2842 vp->vpath->ringh = ring; in __vxge_hw_ring_create()
2844 ring->vp_reg = vp->vpath->vp_reg; in __vxge_hw_ring_create()
2846 ring->stats = &vp->vpath->sw_stats->ring_stats; in __vxge_hw_ring_create()
2852 ring->tim_rti_cfg1_saved = vp->vpath->tim_rti_cfg1_saved; in __vxge_hw_ring_create()
2853 ring->tim_rti_cfg3_saved = vp->vpath->tim_rti_cfg3_saved; in __vxge_hw_ring_create()
3392 struct __vxge_hw_fifo *fifo = vp->vpath->fifoh; in __vxge_hw_fifo_delete()
3399 vp->vpath->fifoh = NULL; in __vxge_hw_fifo_delete()
3461 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_fifo_create() local
3467 vpath = vp->vpath; in __vxge_hw_fifo_create()
3468 config = &vpath->hldev->config.vp_config[vpath->vp_id].fifo; in __vxge_hw_fifo_create()
3484 vpath->fifoh = fifo; in __vxge_hw_fifo_create()
3485 fifo->nofl_db = vpath->nofl_db; in __vxge_hw_fifo_create()
3487 fifo->vp_id = vpath->vp_id; in __vxge_hw_fifo_create()
3488 fifo->vp_reg = vpath->vp_reg; in __vxge_hw_fifo_create()
3489 fifo->stats = &vpath->sw_stats->fifo_stats; in __vxge_hw_fifo_create()
3495 fifo->tim_tti_cfg1_saved = vpath->tim_tti_cfg1_saved; in __vxge_hw_fifo_create()
3496 fifo->tim_tti_cfg3_saved = vpath->tim_tti_cfg3_saved; in __vxge_hw_fifo_create()
3546 __vxge_hw_mempool_create(vpath->hldev, in __vxge_hw_fifo_create()
3578 __vxge_hw_vpath_pci_read(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_read() argument
3583 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in __vxge_hw_vpath_pci_read()
3624 struct __vxge_hw_virtualpath *vpath; in vxge_hw_device_flick_link_led() local
3633 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_device_flick_link_led()
3636 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_device_flick_link_led()
3671 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_get()
3707 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_set()
3831 if (vp->vpath->hldev->config.rth_it_type in vxge_hw_vpath_rts_rth_itable_set()
3974 struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_mgmt_read() argument
3981 val64 = readq(&vpath->vpmgmt_reg-> in __vxge_hw_vpath_mgmt_read()
3991 vpath->max_mtu = mtu + VXGE_HW_MAC_HEADER_MAX_SIZE; in __vxge_hw_vpath_mgmt_read()
3993 val64 = readq(&vpath->vpmgmt_reg->xmac_vsport_choices_vp); in __vxge_hw_vpath_mgmt_read()
3997 vpath->vsport_number = i; in __vxge_hw_vpath_mgmt_read()
4000 val64 = readq(&vpath->vpmgmt_reg->xgmac_gen_status_vpmgmt_clone); in __vxge_hw_vpath_mgmt_read()
4003 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_UP); in __vxge_hw_vpath_mgmt_read()
4005 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_DOWN); in __vxge_hw_vpath_mgmt_read()
4016 __vxge_hw_vpath_reset_check(struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_reset_check() argument
4021 &vpath->hldev->common_reg->vpath_rst_in_prog, in __vxge_hw_vpath_reset_check()
4023 1 << (16 - vpath->vp_id)), in __vxge_hw_vpath_reset_check()
4024 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_reset_check()
4054 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_sw_reset() local
4056 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_sw_reset()
4058 if (vpath->ringh) { in __vxge_hw_vpath_sw_reset()
4059 status = __vxge_hw_ring_reset(vpath->ringh); in __vxge_hw_vpath_sw_reset()
4064 if (vpath->fifoh) in __vxge_hw_vpath_sw_reset()
4065 status = __vxge_hw_fifo_reset(vpath->fifoh); in __vxge_hw_vpath_sw_reset()
4079 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_prc_configure() local
4083 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_prc_configure()
4084 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_prc_configure()
4085 vp_config = vpath->vp_config; in __vxge_hw_vpath_prc_configure()
4094 val64 = readq(&vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4096 writeq(val64, &vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4100 if (vpath->vp_config->ring.scatter_mode != in __vxge_hw_vpath_prc_configure()
4105 switch (vpath->vp_config->ring.scatter_mode) { in __vxge_hw_vpath_prc_configure()
4125 vpath->ringh) >> 3), &vp_reg->prc_cfg5); in __vxge_hw_vpath_prc_configure()
4153 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_kdfc_configure() local
4156 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_kdfc_configure()
4157 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_kdfc_configure()
4165 vpath->max_kdfc_db = in __vxge_hw_vpath_kdfc_configure()
4169 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in __vxge_hw_vpath_kdfc_configure()
4171 vpath->max_nofl_db = vpath->max_kdfc_db; in __vxge_hw_vpath_kdfc_configure()
4173 if (vpath->max_nofl_db < in __vxge_hw_vpath_kdfc_configure()
4174 ((vpath->vp_config->fifo.memblock_size / in __vxge_hw_vpath_kdfc_configure()
4175 (vpath->vp_config->fifo.max_frags * in __vxge_hw_vpath_kdfc_configure()
4177 vpath->vp_config->fifo.fifo_blocks)) { in __vxge_hw_vpath_kdfc_configure()
4182 (vpath->max_nofl_db*2)-1); in __vxge_hw_vpath_kdfc_configure()
4207 vpath->nofl_db = in __vxge_hw_vpath_kdfc_configure()
4224 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_mac_configure() local
4228 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_mac_configure()
4229 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_mac_configure()
4230 vp_config = vpath->vp_config; in __vxge_hw_vpath_mac_configure()
4233 vpath->vsport_number), &vp_reg->xmac_vsport_choice); in __vxge_hw_vpath_mac_configure()
4254 VXGE_HW_MAC_HEADER_MAX_SIZE) < vpath->max_mtu) in __vxge_hw_vpath_mac_configure()
4260 vpath->max_mtu); in __vxge_hw_vpath_mac_configure()
4291 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_tim_configure() local
4295 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_tim_configure()
4296 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_tim_configure()
4297 config = vpath->vp_config; in __vxge_hw_vpath_tim_configure()
4359 vpath->tim_tti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4417 vpath->tim_tti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4466 vpath->tim_rti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4524 vpath->tim_rti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4554 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_initialize() local
4557 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_initialize()
4563 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_initialize()
4565 status = __vxge_hw_vpath_swapper_set(vpath->vp_reg); in __vxge_hw_vpath_initialize()
4584 status = __vxge_hw_vpath_pci_read(vpath, 1, 0x78, &val32); in __vxge_hw_vpath_initialize()
4613 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_terminate() local
4615 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_terminate()
4617 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) in __vxge_hw_vp_terminate()
4620 VXGE_HW_DEVICE_TIM_INT_MASK_RESET(vpath->hldev->tim_int_mask0, in __vxge_hw_vp_terminate()
4621 vpath->hldev->tim_int_mask1, vpath->vp_id); in __vxge_hw_vp_terminate()
4622 hldev->stats.hw_dev_info_stats.vpath_info[vpath->vp_id] = NULL; in __vxge_hw_vp_terminate()
4627 spin_lock(&vpath->lock); in __vxge_hw_vp_terminate()
4628 vpath->vp_open = VXGE_HW_VP_NOT_OPEN; in __vxge_hw_vp_terminate()
4629 spin_unlock(&vpath->lock); in __vxge_hw_vp_terminate()
4631 vpath->vpmgmt_reg = NULL; in __vxge_hw_vp_terminate()
4632 vpath->nofl_db = NULL; in __vxge_hw_vp_terminate()
4633 vpath->max_mtu = 0; in __vxge_hw_vp_terminate()
4634 vpath->vsport_number = 0; in __vxge_hw_vp_terminate()
4635 vpath->max_kdfc_db = 0; in __vxge_hw_vp_terminate()
4636 vpath->max_nofl_db = 0; in __vxge_hw_vp_terminate()
4637 vpath->ringh = NULL; in __vxge_hw_vp_terminate()
4638 vpath->fifoh = NULL; in __vxge_hw_vp_terminate()
4639 memset(&vpath->vpath_handles, 0, sizeof(struct list_head)); in __vxge_hw_vp_terminate()
4640 vpath->stats_block = NULL; in __vxge_hw_vp_terminate()
4641 vpath->hw_stats = NULL; in __vxge_hw_vp_terminate()
4642 vpath->hw_stats_sav = NULL; in __vxge_hw_vp_terminate()
4643 vpath->sw_stats = NULL; in __vxge_hw_vp_terminate()
4658 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_initialize() local
4666 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_initialize()
4668 spin_lock_init(&vpath->lock); in __vxge_hw_vp_initialize()
4669 vpath->vp_id = vp_id; in __vxge_hw_vp_initialize()
4670 vpath->vp_open = VXGE_HW_VP_OPEN; in __vxge_hw_vp_initialize()
4671 vpath->hldev = hldev; in __vxge_hw_vp_initialize()
4672 vpath->vp_config = config; in __vxge_hw_vp_initialize()
4673 vpath->vp_reg = hldev->vpath_reg[vp_id]; in __vxge_hw_vp_initialize()
4674 vpath->vpmgmt_reg = hldev->vpmgmt_reg[vp_id]; in __vxge_hw_vp_initialize()
4678 status = __vxge_hw_vpath_reset_check(vpath); in __vxge_hw_vp_initialize()
4680 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4684 status = __vxge_hw_vpath_mgmt_read(hldev, vpath); in __vxge_hw_vp_initialize()
4686 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4690 INIT_LIST_HEAD(&vpath->vpath_handles); in __vxge_hw_vp_initialize()
4692 vpath->sw_stats = &hldev->stats.sw_dev_info_stats.vpath_info[vp_id]; in __vxge_hw_vp_initialize()
4714 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_mtu_set() local
4720 vpath = vp->vpath; in vxge_hw_vpath_mtu_set()
4724 if ((new_mtu < VXGE_HW_MIN_MTU) || (new_mtu > vpath->max_mtu)) in vxge_hw_vpath_mtu_set()
4727 val64 = readq(&vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4732 writeq(val64, &vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4734 vpath->vp_config->mtu = new_mtu - VXGE_HW_MAC_HEADER_MAX_SIZE; in vxge_hw_vpath_mtu_set()
4749 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_stats_enable() local
4751 vpath = vp->vpath; in vxge_hw_vpath_stats_enable()
4753 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_stats_enable()
4758 memcpy(vpath->hw_stats_sav, vpath->hw_stats, in vxge_hw_vpath_stats_enable()
4761 status = __vxge_hw_vpath_stats_get(vpath, vpath->hw_stats); in vxge_hw_vpath_stats_enable()
4809 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_open() local
4813 vpath = &hldev->virtual_paths[attr->vp_id]; in vxge_hw_vpath_open()
4815 if (vpath->vp_open == VXGE_HW_VP_OPEN) { in vxge_hw_vpath_open()
4831 vp->vpath = vpath; in vxge_hw_vpath_open()
4833 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in vxge_hw_vpath_open()
4839 if (vpath->vp_config->ring.enable == VXGE_HW_RING_ENABLE) { in vxge_hw_vpath_open()
4847 vpath->fifoh->tx_intr_num = in vxge_hw_vpath_open()
4851 vpath->stats_block = __vxge_hw_blockpool_block_allocate(hldev, in vxge_hw_vpath_open()
4853 if (vpath->stats_block == NULL) { in vxge_hw_vpath_open()
4858 vpath->hw_stats = vpath->stats_block->memblock; in vxge_hw_vpath_open()
4859 memset(vpath->hw_stats, 0, in vxge_hw_vpath_open()
4863 vpath->hw_stats; in vxge_hw_vpath_open()
4865 vpath->hw_stats_sav = in vxge_hw_vpath_open()
4867 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_open()
4870 writeq(vpath->stats_block->dma_addr, &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_open()
4876 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_open()
4878 hldev->vpaths_deployed |= vxge_mBIT(vpath->vp_id); in vxge_hw_vpath_open()
4882 attr->fifo_attr.userdata = vpath->fifoh; in vxge_hw_vpath_open()
4883 attr->ring_attr.userdata = vpath->ringh; in vxge_hw_vpath_open()
4888 if (vpath->ringh != NULL) in vxge_hw_vpath_open()
4891 if (vpath->fifoh != NULL) in vxge_hw_vpath_open()
4912 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_rx_doorbell_init() local
4913 struct __vxge_hw_ring *ring = vpath->ringh; in vxge_hw_vpath_rx_doorbell_init()
4914 struct vxgedev *vdev = netdev_priv(vpath->hldev->ndev); in vxge_hw_vpath_rx_doorbell_init()
4918 new_count = readq(&vpath->vp_reg->rxdmem_size); in vxge_hw_vpath_rx_doorbell_init()
4926 &vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4927 readl(&vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4930 val64 = readq(&vpath->vp_reg->prc_cfg6); in vxge_hw_vpath_rx_doorbell_init()
4975 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_close() local
4977 u32 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_close()
4981 vpath = vp->vpath; in vxge_hw_vpath_close()
4982 devh = vpath->hldev; in vxge_hw_vpath_close()
4984 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_close()
4991 if (!list_empty(&vpath->vpath_handles)) { in vxge_hw_vpath_close()
4992 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_close()
5003 if (vpath->ringh != NULL) in vxge_hw_vpath_close()
5006 if (vpath->fifoh != NULL) in vxge_hw_vpath_close()
5009 if (vpath->stats_block != NULL) in vxge_hw_vpath_close()
5010 __vxge_hw_blockpool_block_free(devh, vpath->stats_block); in vxge_hw_vpath_close()
5028 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_reset() local
5030 vp_id = vpath->vp_id; in vxge_hw_vpath_reset()
5032 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_reset()
5037 status = __vxge_hw_vpath_reset(vpath->hldev, vp_id); in vxge_hw_vpath_reset()
5039 vpath->sw_stats->soft_reset_cnt++; in vxge_hw_vpath_reset()
5052 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_recover_from_reset() local
5057 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_recover_from_reset()
5058 vpath = vp->vpath; in vxge_hw_vpath_recover_from_reset()
5059 hldev = vpath->hldev; in vxge_hw_vpath_recover_from_reset()
5061 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_recover_from_reset()
5066 status = __vxge_hw_vpath_reset_check(vpath); in vxge_hw_vpath_recover_from_reset()
5078 if (vpath->ringh != NULL) in vxge_hw_vpath_recover_from_reset()
5081 memset(vpath->hw_stats, 0, in vxge_hw_vpath_recover_from_reset()
5084 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_recover_from_reset()
5087 writeq(vpath->stats_block->dma_addr, in vxge_hw_vpath_recover_from_reset()
5088 &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_recover_from_reset()
5107 hldev = vp->vpath->hldev; in vxge_hw_vpath_enable()
5110 1 << (16 - vp->vpath->vp_id)); in vxge_hw_vpath_enable()