Lines Matching refs:txr
252 static inline u32 bnx2_tx_avail(struct bnx2 *bp, struct bnx2_tx_ring_info *txr) in bnx2_tx_avail() argument
262 diff = txr->tx_prod - txr->tx_cons; in bnx2_tx_avail()
700 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_mem() local
702 if (txr->tx_desc_ring) { in bnx2_free_tx_mem()
704 txr->tx_desc_ring, in bnx2_free_tx_mem()
705 txr->tx_desc_mapping); in bnx2_free_tx_mem()
706 txr->tx_desc_ring = NULL; in bnx2_free_tx_mem()
708 kfree(txr->tx_buf_ring); in bnx2_free_tx_mem()
709 txr->tx_buf_ring = NULL; in bnx2_free_tx_mem()
752 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_alloc_tx_mem() local
754 txr->tx_buf_ring = kzalloc(SW_TXBD_RING_SIZE, GFP_KERNEL); in bnx2_alloc_tx_mem()
755 if (txr->tx_buf_ring == NULL) in bnx2_alloc_tx_mem()
758 txr->tx_desc_ring = in bnx2_alloc_tx_mem()
760 &txr->tx_desc_mapping, GFP_KERNEL); in bnx2_alloc_tx_mem()
761 if (txr->tx_desc_ring == NULL) in bnx2_alloc_tx_mem()
2830 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_tx_int() local
2840 sw_cons = txr->tx_cons; in bnx2_tx_int()
2849 tx_buf = &txr->tx_buf_ring[sw_ring_cons]; in bnx2_tx_int()
2880 tx_buf = &txr->tx_buf_ring[BNX2_TX_RING_IDX(sw_cons)]; in bnx2_tx_int()
2900 txr->hw_tx_cons = hw_cons; in bnx2_tx_int()
2901 txr->tx_cons = sw_cons; in bnx2_tx_int()
2911 (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh)) { in bnx2_tx_int()
2914 (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh)) in bnx2_tx_int()
3378 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_has_fast_work() local
3382 (bnx2_get_hw_tx_cons(bnapi) != txr->hw_tx_cons)) in bnx2_has_fast_work()
3472 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_poll_work() local
3475 if (bnx2_get_hw_tx_cons(bnapi) != txr->hw_tx_cons) in bnx2_poll_work()
5102 struct bnx2_tx_ring_info *txr; in bnx2_clear_ring_states() local
5108 txr = &bnapi->tx_ring; in bnx2_clear_ring_states()
5111 txr->tx_cons = 0; in bnx2_clear_ring_states()
5112 txr->hw_tx_cons = 0; in bnx2_clear_ring_states()
5122 bnx2_init_tx_context(struct bnx2 *bp, u32 cid, struct bnx2_tx_ring_info *txr) in bnx2_init_tx_context() argument
5144 val = (u64) txr->tx_desc_mapping >> 32; in bnx2_init_tx_context()
5147 val = (u64) txr->tx_desc_mapping & 0xffffffff; in bnx2_init_tx_context()
5157 struct bnx2_tx_ring_info *txr; in bnx2_init_tx_ring() local
5160 txr = &bnapi->tx_ring; in bnx2_init_tx_ring()
5169 txbd = &txr->tx_desc_ring[BNX2_MAX_TX_DESC_CNT]; in bnx2_init_tx_ring()
5171 txbd->tx_bd_haddr_hi = (u64) txr->tx_desc_mapping >> 32; in bnx2_init_tx_ring()
5172 txbd->tx_bd_haddr_lo = (u64) txr->tx_desc_mapping & 0xffffffff; in bnx2_init_tx_ring()
5174 txr->tx_prod = 0; in bnx2_init_tx_ring()
5175 txr->tx_prod_bseq = 0; in bnx2_init_tx_ring()
5177 txr->tx_bidx_addr = MB_GET_CID_ADDR(cid) + BNX2_L2CTX_TX_HOST_BIDX; in bnx2_init_tx_ring()
5178 txr->tx_bseq_addr = MB_GET_CID_ADDR(cid) + BNX2_L2CTX_TX_HOST_BSEQ; in bnx2_init_tx_ring()
5180 bnx2_init_tx_context(bp, cid, txr); in bnx2_init_tx_ring()
5407 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_skbs() local
5410 if (txr->tx_buf_ring == NULL) in bnx2_free_tx_skbs()
5414 struct bnx2_sw_tx_bd *tx_buf = &txr->tx_buf_ring[j]; in bnx2_free_tx_skbs()
5433 tx_buf = &txr->tx_buf_ring[BNX2_TX_RING_IDX(j)]; in bnx2_free_tx_skbs()
5788 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_run_loopback() local
5793 txr = &tx_napi->tx_ring; in bnx2_run_loopback()
5836 txbd = &txr->tx_desc_ring[BNX2_TX_RING_IDX(txr->tx_prod)]; in bnx2_run_loopback()
5844 txr->tx_prod = BNX2_NEXT_TX_BD(txr->tx_prod); in bnx2_run_loopback()
5845 txr->tx_prod_bseq += pkt_size; in bnx2_run_loopback()
5847 BNX2_WR16(bp, txr->tx_bidx_addr, txr->tx_prod); in bnx2_run_loopback()
5848 BNX2_WR(bp, txr->tx_bseq_addr, txr->tx_prod_bseq); in bnx2_run_loopback()
5862 if (bnx2_get_hw_tx_cons(tx_napi) != txr->tx_prod) in bnx2_run_loopback()
6573 struct bnx2_tx_ring_info *txr; in bnx2_start_xmit() local
6579 txr = &bnapi->tx_ring; in bnx2_start_xmit()
6582 if (unlikely(bnx2_tx_avail(bp, txr) < in bnx2_start_xmit()
6590 prod = txr->tx_prod; in bnx2_start_xmit()
6643 tx_buf = &txr->tx_buf_ring[ring_prod]; in bnx2_start_xmit()
6647 txbd = &txr->tx_desc_ring[ring_prod]; in bnx2_start_xmit()
6663 txbd = &txr->tx_desc_ring[ring_prod]; in bnx2_start_xmit()
6670 dma_unmap_addr_set(&txr->tx_buf_ring[ring_prod], mapping, in bnx2_start_xmit()
6687 txr->tx_prod_bseq += skb->len; in bnx2_start_xmit()
6689 BNX2_WR16(bp, txr->tx_bidx_addr, prod); in bnx2_start_xmit()
6690 BNX2_WR(bp, txr->tx_bseq_addr, txr->tx_prod_bseq); in bnx2_start_xmit()
6694 txr->tx_prod = prod; in bnx2_start_xmit()
6696 if (unlikely(bnx2_tx_avail(bp, txr) <= MAX_SKB_FRAGS)) { in bnx2_start_xmit()
6705 if (bnx2_tx_avail(bp, txr) > bp->tx_wake_thresh) in bnx2_start_xmit()
6715 prod = txr->tx_prod; in bnx2_start_xmit()
6717 tx_buf = &txr->tx_buf_ring[ring_prod]; in bnx2_start_xmit()
6726 tx_buf = &txr->tx_buf_ring[ring_prod]; in bnx2_start_xmit()