Searched refs:tx_next (Results 1 - 10 of 10) sorted by relevance
/linux-4.4.14/drivers/net/ethernet/aeroflex/ |
H A D | greth.c | 326 greth->tx_next = 0; greth_init_rings() 427 bdp = greth->tx_bd_base + greth->tx_next; greth_start_xmit() 435 greth->tx_bufs_length[greth->tx_next] = skb->len & GRETH_BD_LEN; greth_start_xmit() 438 if (greth->tx_next == GRETH_TXBD_NUM_MASK) { greth_start_xmit() 442 greth->tx_next = NEXT_TX(greth->tx_next); greth_start_xmit() 456 static inline u16 greth_num_free_bds(u16 tx_last, u16 tx_next) greth_num_free_bds() argument 458 if (tx_next < tx_last) greth_num_free_bds() 459 return (tx_last - tx_next) - 1; greth_num_free_bds() 461 return GRETH_TXBD_NUM - (tx_next - tx_last) - 1; greth_num_free_bds() 478 if (greth_num_free_bds(tx_last, greth->tx_next) < nr_frags + 1) { greth_start_xmit_gbit() 493 greth->tx_skbuff[greth->tx_next] = skb; greth_start_xmit_gbit() 504 if (greth->tx_next == GRETH_TXBD_NUM_MASK) greth_start_xmit_gbit() 508 bdp = greth->tx_bd_base + greth->tx_next; greth_start_xmit_gbit() 517 curr_tx = NEXT_TX(greth->tx_next); greth_start_xmit_gbit() 556 bdp = greth->tx_bd_base + greth->tx_next; greth_start_xmit_gbit() 561 greth->tx_next = curr_tx; greth_start_xmit_gbit() 569 for (i = 0; greth->tx_next + i != curr_tx; i++) { greth_start_xmit_gbit() 570 bdp = greth->tx_bd_base + greth->tx_next + i; greth_start_xmit_gbit() 687 while (tx_last != greth->tx_next) { greth_clean_tx_gbit() 735 (greth_num_free_bds(tx_last, greth->tx_next) > greth_clean_tx_gbit() 995 if ((greth->gbit_mac && (greth->tx_last != greth->tx_next)) || greth_poll()
|
H A D | greth.h | 108 u16 tx_next; member in struct:greth_private
|
/linux-4.4.14/drivers/net/can/ |
H A D | at91_can.c | 136 unsigned int tx_next; member in struct:at91_priv 269 return (priv->tx_next & get_next_mb_mask(priv)) + get_mb_tx_first(priv); get_tx_next_mb() 274 return (priv->tx_next >> get_next_prio_shift(priv)) & 0xf; get_tx_next_prio() 352 priv->tx_next = priv->tx_echo = 0; at91_setup_mailboxes() 445 * We use the priv->tx_next as counter for the next transmission 449 * priv->tx_next = (prio << get_next_prio_shift(priv)) | 496 * tx_next buffer prio and mailbox equals 0. at91_start_xmit() 501 priv->tx_next++; at91_start_xmit() 504 (priv->tx_next & get_next_mask(priv)) == 0) at91_start_xmit() 829 * We iterate from priv->tx_echo to priv->tx_next and check if the 842 for (/* nix */; (priv->tx_next - priv->tx_echo) > 0; priv->tx_echo++) { at91_irq_tx() 873 if ((priv->tx_next & get_next_mask(priv)) != 0 || at91_irq_tx()
|
/linux-4.4.14/drivers/net/ethernet/synopsys/ |
H A D | dwc_eth_qos.c | 624 size_t tx_next; member in struct:net_local 704 pr_info("tx_free %zu, tx_cur %zu, tx_next %zu\n", lp->tx_free, print_status() 705 lp->tx_cur, lp->tx_next); print_status() 1109 lp->tx_next = 0; dwceqos_descriptor_init() 1957 struct dwceqos_dma_desc *dd = &lp->tx_descs[lp->tx_next]; dwceqos_dmadesc_set_ctx() 1964 lp->tx_next = (lp->tx_next + 1) % DWCEQOS_TX_DCNT; dwceqos_dmadesc_set_ctx() 1998 tx->initial_descriptor = lp->tx_next; dwceqos_tx_prepare() 1999 tx->last_descriptor = lp->tx_next; dwceqos_tx_prepare() 2028 rd = &lp->tx_skb[lp->tx_next]; dwceqos_tx_linear() 2029 dd = &lp->tx_descs[lp->tx_next]; dwceqos_tx_linear() 2065 if (lp->tx_next != tx->initial_descriptor) dwceqos_tx_linear() 2068 tx->last_descriptor = lp->tx_next; dwceqos_tx_linear() 2069 lp->tx_next = (lp->tx_next + 1) % DWCEQOS_TX_DCNT; dwceqos_tx_linear() 2104 rd = &lp->tx_skb[lp->tx_next]; dwceqos_tx_frags() 2107 dd = &lp->tx_descs[lp->tx_next]; dwceqos_tx_frags() 2121 tx->last_descriptor = lp->tx_next; dwceqos_tx_frags() 2122 lp->tx_next = (lp->tx_next + 1) % DWCEQOS_TX_DCNT; dwceqos_tx_frags() 2158 while (i != lp->tx_next) { dwceqos_tx_rollback() 2173 lp->tx_next = tx->initial_descriptor; dwceqos_tx_rollback() 2197 WARN_ON(lp->tx_next != dwceqos_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/octeon/ |
H A D | octeon_mgmt.c | 128 unsigned int tx_next; member in struct:octeon_mgmt 991 p->tx_next = 0; octeon_mgmt_open() 1307 p->tx_ring[p->tx_next] = re.d64; octeon_mgmt_xmit() 1308 p->tx_next = (p->tx_next + 1) % OCTEON_MGMT_TX_RING_SIZE; octeon_mgmt_xmit()
|
/linux-4.4.14/drivers/net/ethernet/aurora/ |
H A D | nb8800.c | 432 next = priv->tx_next; nb8800_xmit() 470 priv->tx_next = next; nb8800_xmit() 499 unsigned int limit = priv->tx_next; nb8800_tx_done() 824 priv->tx_next = 0; nb8800_dma_reset() 894 priv->tx_done == priv->tx_next, nb8800_dma_stop()
|
H A D | nb8800.h | 271 u32 tx_next; member in struct:nb8800_priv
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
H A D | sdma.c | 2146 struct sdma_txreq *tx, *tx_next; sdma_send_txlist() local 2154 list_for_each_entry_safe(tx, tx_next, tx_list, list) { list_for_each_entry_safe() 2182 list_for_each_entry_safe(tx, tx_next, tx_list, list) { list_for_each_entry_safe()
|
/linux-4.4.14/drivers/net/ethernet/marvell/ |
H A D | sky2.c | 2040 sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); sky2_tx_complete() 4575 for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; sky2_debug_show()
|
H A D | sky2.h | 2233 u16 tx_next; /* debug only */ member in struct:sky2_port
|
Completed in 384 milliseconds