Lines Matching refs:bna

289 	bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);  in bna_bfi_ucast_req()
305 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_add_req()
321 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_del_req()
336 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_filter_req()
351 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_promisc_req()
376 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_vlan_filter_set()
391 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_vlan_strip_enable()
407 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rit_cfg()
427 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_cfg()
442 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_enable()
491 mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod); in bna_rxf_mchandle_attach()
519 bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle); in bna_rxf_mcast_del()
539 bna_cam_mod_mac_put(bna_mcam_mod_del_q(rxf->rx->bna), mac); in bna_rxf_mcast_cfg_apply()
590 bna_cam_mod_mac_put(bna_mcam_mod_del_q(rxf->rx->bna), mac); in bna_rxf_mcast_cfg_reset()
800 bna_cam_mod_mac_put(bna_ucam_mod_free_q(rxf->rx->bna), mac); in bna_rxf_uninit()
805 bna_cam_mod_mac_put(bna_ucam_mod_free_q(rxf->rx->bna), in bna_rxf_uninit()
813 bna_cam_mod_mac_put(bna_mcam_mod_free_q(rxf->rx->bna), mac); in bna_rxf_uninit()
818 if (rxf->rx->bna->promisc_rid == rxf->rx->rid) in bna_rxf_uninit()
819 rxf->rx->bna->promisc_rid = BFI_INVALID_RID; in bna_rxf_uninit()
820 if (rxf->rx->bna->default_mode_rid == rxf->rx->rid) in bna_rxf_uninit()
821 rxf->rx->bna->default_mode_rid = BFI_INVALID_RID; in bna_rxf_uninit()
873 bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna)); in bna_rx_ucast_set()
882 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_ucast_set()
900 cbfn(rx->bna->bnad, rx); in bna_rx_mcast_add()
904 mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna)); in bna_rx_mcast_add()
912 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mcast_add()
923 struct bna_ucam_mod *ucam_mod = &rx->bna->ucam_mod; in bna_rx_ucast_listset()
972 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_ucast_listset()
992 struct bna_mcam_mod *mcam_mod = &rx->bna->mcam_mod; in bna_rx_mcast_listset()
1044 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mcast_listset()
1074 bna_cam_mod_mac_put(bna_mcam_mod_free_q(rxf->rx->bna), mac); in bna_rx_mcast_delall()
1083 del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rx_mcast_delall()
1088 bna_cam_mod_mac_put(bna_mcam_mod_free_q(rxf->rx->bna), mac); in bna_rx_mcast_delall()
1094 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mcast_delall()
1100 (*cbfn)(rx->bna->bnad, rx); in bna_rx_mcast_delall()
1145 bna_cam_mod_mac_put(bna_ucam_mod_del_q(rxf->rx->bna), mac); in bna_rxf_ucast_cfg_apply()
1185 bna_cam_mod_mac_put(bna_ucam_mod_del_q(rxf->rx->bna), in bna_rxf_ucast_cfg_reset()
1190 bna_cam_mod_mac_put(bna_ucam_mod_del_q(rxf->rx->bna), in bna_rxf_ucast_cfg_reset()
1225 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_cfg_apply() local
1242 bna->promisc_rid = BFI_INVALID_RID; in bna_rxf_promisc_cfg_apply()
1253 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_cfg_reset() local
1261 bna->promisc_rid = BFI_INVALID_RID; in bna_rxf_promisc_cfg_reset()
1339 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_enable() local
1355 bna->promisc_rid = rxf->rx->rid; in bna_rxf_promisc_enable()
1365 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_disable() local
1377 bna->promisc_rid = BFI_INVALID_RID; in bna_rxf_promisc_disable()
1472 (rx)->rx_stall_cbfn((rx)->bna->bnad, (rx)); \
1556 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); in bna_rx_sm_stop_wait()
1593 rx->rx_post_cbfn(rx->bna->bnad, rx); in bna_rx_sm_rxf_start_wait_entry()
1610 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); in bna_rx_sm_rxf_stop_wait()
1663 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular); in bna_rx_sm_started_entry()
1666 bna_ethport_cb_rx_started(&rx->bna->ethport); in bna_rx_sm_started_entry()
1675 bna_ethport_cb_rx_stopped(&rx->bna->ethport); in bna_rx_sm_started()
1681 bna_ethport_cb_rx_stopped(&rx->bna->ethport); in bna_rx_sm_started()
1684 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); in bna_rx_sm_started()
1705 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); in bna_rx_sm_rxf_start_wait()
1815 cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); in bna_bfi_rx_enet_start()
1847 bna_enet_mtu_get(&rx->bna->enet); in bna_bfi_rx_enet_start()
1903 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); in bna_bfi_rx_enet_start()
1917 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); in bna_bfi_rx_enet_stop()
1929 bna_ib_stop(rx->bna, &rxp->cq.ib); in bna_rx_enet_stop()
2165 rx_mod->stop_cbfn(&rx_mod->bna->enet); in bna_rx_mod_cb_rx_stopped_all()
2182 bna_rx_mod_cb_rx_stopped(&rx->bna->rx_mod, rx); in bna_rx_stop()
2185 rx->stop_cbarg = &rx->bna->rx_mod; in bna_rx_stop()
2254 void bna_rx_mod_init(struct bna_rx_mod *rx_mod, struct bna *bna, in bna_rx_mod_init() argument
2262 rx_mod->bna = bna; in bna_rx_mod_init()
2282 for (index = 0; index < bna->ioceth.attr.num_rxp; index++) { in bna_rx_mod_init()
2287 rx_ptr->bna = NULL; in bna_rx_mod_init()
2297 for (index = 0; index < bna->ioceth.attr.num_rxp; index++) { in bna_rx_mod_init()
2305 for (index = 0; index < (bna->ioceth.attr.num_rxp * 2); index++) { in bna_rx_mod_init()
2331 rx_mod->bna = NULL; in bna_rx_mod_uninit()
2343 bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp, in bna_bfi_rx_enet_start_rsp()
2356 rx->bna->pcidev.pci_bar_kva in bna_bfi_rx_enet_start_rsp()
2360 rx->bna->pcidev.pci_bar_kva in bna_bfi_rx_enet_start_rsp()
2365 rx->bna->pcidev.pci_bar_kva in bna_bfi_rx_enet_start_rsp()
2503 bna_rx_create(struct bna *bna, struct bnad *bnad, in bna_rx_create() argument
2509 struct bna_rx_mod *rx_mod = &bna->rx_mod; in bna_rx_create()
2561 rx->bna = bna; in bna_rx_create()
2577 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_STARTED) { in bna_rx_create()
2580 if (!(rx->bna->rx_mod.flags & in bna_rx_create()
2585 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_LOOPBACK) in bna_rx_create()
2644 q0->rcb->bnad = bna->bnad; in bna_rx_create()
2669 q1->rcb->bnad = bna->bnad; in bna_rx_create()
2712 rxp->cq.ccb->bnad = bna->bnad; in bna_rx_create()
2736 struct bna_rx_mod *rx_mod = &rx->bna->rx_mod; in bna_rx_destroy()
2748 rx->rcb_destroy_cbfn(rx->bna->bnad, q0->rcb); in bna_rx_destroy()
2756 rx->rcb_destroy_cbfn(rx->bna->bnad, q1->rcb); in bna_rx_destroy()
2766 rx->ccb_destroy_cbfn(rx->bna->bnad, rxp->cq.ccb); in bna_rx_destroy()
2782 rx->bna = NULL; in bna_rx_destroy()
2804 (*cbfn)(rx->bna->bnad, rx); in bna_rx_disable()
2807 rx->stop_cbarg = rx->bna->bnad; in bna_rx_disable()
2857 if ((rx->bna->promisc_rid != BFI_INVALID_RID) && in bna_rx_mode_set()
2858 (rx->bna->promisc_rid != rxf->rx->rid)) in bna_rx_mode_set()
2862 if (rx->bna->default_mode_rid != BFI_INVALID_RID) in bna_rx_mode_set()
2872 if ((rx->bna->default_mode_rid != BFI_INVALID_RID) && in bna_rx_mode_set()
2873 (rx->bna->default_mode_rid != rxf->rx->rid)) { in bna_rx_mode_set()
2878 if (rx->bna->promisc_rid != BFI_INVALID_RID) in bna_rx_mode_set()
2904 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mode_set()
2907 (*cbfn)(rx->bna->bnad, rx); in bna_rx_mode_set()
2941 bna_rx_dim_reconfig(struct bna *bna, const u32 vector[][BNA_BIAS_T_MAX]) in bna_rx_dim_reconfig() argument
2947 bna->rx_mod.dim_vector[i][j] = vector[i][j]; in bna_rx_dim_reconfig()
2953 struct bna *bna = ccb->cq->rx->bna; in bna_rx_dim_update() local
2994 coalescing_timeo = bna->rx_mod.dim_vector[load][bias]; in bna_rx_dim_update()
3033 cbfn((tx)->bna->bnad, (tx)); \
3154 bna_ib_start(tx->bna, &txq->ib, is_regular); in bna_tx_sm_started_entry()
3156 tx->tx_resume_cbfn(tx->bna->bnad, tx); in bna_tx_sm_started_entry()
3165 tx->tx_stall_cbfn(tx->bna->bnad, tx); in bna_tx_sm_started()
3171 tx->tx_stall_cbfn(tx->bna->bnad, tx); in bna_tx_sm_started()
3172 tx->tx_cleanup_cbfn(tx->bna->bnad, tx); in bna_tx_sm_started()
3197 tx->tx_cleanup_cbfn(tx->bna->bnad, tx); in bna_tx_sm_stop_wait()
3245 tx->tx_stall_cbfn(tx->bna->bnad, tx); in bna_tx_sm_prio_stop_wait_entry()
3260 tx->tx_cleanup_cbfn(tx->bna->bnad, tx); in bna_tx_sm_prio_stop_wait()
3281 tx->tx_cleanup_cbfn(tx->bna->bnad, tx); in bna_tx_sm_prio_cleanup_wait_entry()
3419 bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); in bna_bfi_tx_enet_start()
3433 bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); in bna_bfi_tx_enet_stop()
3445 bna_ib_stop(tx->bna, &txq->ib); in bna_tx_enet_stop()
3510 struct bna_tx_mod *tx_mod = &tx->bna->tx_mod; in bna_tx_free()
3531 tx->bna = NULL; in bna_tx_free()
3570 tx->stop_cbarg = &tx->bna->tx_mod; in bna_tx_stop()
3591 bfa_msgq_rsp_copy(&tx->bna->msgq, (u8 *)cfg_rsp, in bna_bfi_tx_enet_start_rsp()
3602 tx->bna->pcidev.pci_bar_kva in bna_bfi_tx_enet_start_rsp()
3605 tx->bna->pcidev.pci_bar_kva in bna_bfi_tx_enet_start_rsp()
3683 bna_tx_create(struct bna *bna, struct bnad *bnad, in bna_tx_create() argument
3689 struct bna_tx_mod *tx_mod = &bna->tx_mod; in bna_tx_create()
3712 tx->bna = bna; in bna_tx_create()
3746 if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_STARTED) { in bna_tx_create()
3749 if (!(tx->bna->tx_mod.flags & in bna_tx_create()
3754 if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_LOOPBACK) in bna_tx_create()
3810 (tx->tcb_setup_cbfn)(bna->bnad, txq->tcb); in bna_tx_create()
3842 (tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb); in bna_tx_destroy()
3845 tx->bna->tx_mod.rid_mask &= ~(1 << tx->rid); in bna_tx_destroy()
3866 (*cbfn)(tx->bna->bnad, tx); in bna_tx_disable()
3871 tx->stop_cbarg = tx->bna->bnad; in bna_tx_disable()
3898 tx_mod->stop_cbfn(&tx_mod->bna->enet); in bna_tx_mod_cb_tx_stopped_all()
3903 bna_tx_mod_init(struct bna_tx_mod *tx_mod, struct bna *bna, in bna_tx_mod_init() argument
3908 tx_mod->bna = bna; in bna_tx_mod_init()
3921 for (i = 0; i < bna->ioceth.attr.num_txq; i++) { in bna_tx_mod_init()
3949 tx_mod->bna = NULL; in bna_tx_mod_uninit()