Searched refs:cur_tx (Results 1 - 55 of 55) sorted by relevance

/linux-4.4.14/drivers/net/ethernet/amd/
H A Dariadne.c92 int cur_tx, cur_rx; /* The next free ring entry */ member in struct:ariadne_private
125 priv->cur_rx = priv->cur_tx = 0; ariadne_init_ring()
308 while (dirty_tx < priv->cur_tx) { ariadne_interrupt()
345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { ariadne_interrupt()
347 dirty_tx, priv->cur_tx, ariadne_interrupt()
354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { ariadne_interrupt()
576 entry = priv->cur_tx % TX_RING_SIZE; ariadne_start_xmit()
596 priv->cur_tx++; ariadne_start_xmit()
597 if ((priv->cur_tx >= TX_RING_SIZE) && ariadne_start_xmit()
600 netdev_dbg(dev, "*** Subtracting TX_RING_SIZE from cur_tx (%d) and dirty_tx (%d)\n", ariadne_start_xmit()
601 priv->cur_tx, priv->dirty_tx); ariadne_start_xmit()
603 priv->cur_tx -= TX_RING_SIZE; ariadne_start_xmit()
H A Datarilance.c222 int cur_rx, cur_tx; /* The next free ring entry */ member in struct:lance_private
689 lp->cur_rx = lp->cur_tx = 0; lance_init_ring()
747 DPRINTK( 2, ( "Ring data: dirty_tx %d cur_tx %d%s cur_rx %d\n", lance_tx_timeout()
748 lp->dirty_tx, lp->cur_tx, lance_tx_timeout()
814 entry = lp->cur_tx & TX_RING_MOD_MASK; lance_start_xmit()
828 lp->cur_tx++; lance_start_xmit()
829 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { lance_start_xmit()
830 lp->cur_tx -= TX_RING_SIZE; lance_start_xmit()
884 while( dirty_tx < lp->cur_tx) { lance_interrupt()
920 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { lance_interrupt()
923 dirty_tx, lp->cur_tx, lp->tx_full )); lance_interrupt()
929 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) { lance_interrupt()
H A Dlance.c248 int cur_rx, cur_tx; /* The next free ring entry */ member in struct:lance_private
867 lp->cur_rx = lp->cur_tx = 0; lance_init_ring()
930 printk (" Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.", lance_tx_timeout()
931 lp->dirty_tx, lp->cur_tx, netif_queue_stopped(dev) ? " (full)" : "", lance_tx_timeout()
971 entry = lp->cur_tx & TX_RING_MOD_MASK; lance_start_xmit()
1006 lp->cur_tx++; lance_start_xmit()
1012 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) lance_start_xmit()
1051 while (dirty_tx < lp->cur_tx) { lance_interrupt()
1095 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { lance_interrupt()
1097 dirty_tx, lp->cur_tx, lance_interrupt()
1105 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) lance_interrupt()
H A Dpcnet32.c275 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:pcnet32_private
1252 while (dirty_tx != lp->cur_tx) { pcnet32_tx()
1312 delta = (lp->cur_tx - dirty_tx) & (lp->tx_mod_mask + lp->tx_ring_size); pcnet32_tx()
1315 dirty_tx, lp->cur_tx, lp->tx_full); pcnet32_tx()
2307 lp->cur_rx = lp->cur_tx = 0; pcnet32_init_ring()
2410 " Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.", pcnet32_tx_timeout()
2411 lp->dirty_tx, lp->cur_tx, lp->tx_full ? " (full)" : "", pcnet32_tx_timeout()
2458 entry = lp->cur_tx & lp->tx_mod_mask; pcnet32_start_xmit()
2479 lp->cur_tx++; pcnet32_start_xmit()
/linux-4.4.14/drivers/net/arcnet/
H A Darcnet.c473 lp->cur_tx = lp->next_tx = -1; arcnet_open()
609 lp->hw.status(dev), lp->cur_tx, lp->next_tx, skb->len, skb->protocol); arcnet_send_packet()
689 arc_printk(D_DURING, dev, "go_tx: status=%Xh, intmask=%Xh, next_tx=%d, cur_tx=%d\n", go_tx()
690 lp->hw.status(dev), lp->intmask, lp->next_tx, lp->cur_tx); go_tx()
692 if (lp->cur_tx != -1 || lp->next_tx == -1) go_tx()
698 lp->cur_tx = lp->next_tx; go_tx()
702 lp->hw.command(dev, TXcmd | (lp->cur_tx << 3)); go_tx()
728 lp->hw.command(dev, NOTXcmd | (lp->cur_tx << 3)); arcnet_timeout()
745 if (lp->cur_tx == -1) arcnet_timeout()
850 if (lp->cur_tx != -1 && !lp->timed_out) { arcnet_interrupt()
882 if (lp->cur_tx != -1) arcnet_interrupt()
883 release_arcbuf(dev, lp->cur_tx); arcnet_interrupt()
885 lp->cur_tx = -1; arcnet_interrupt()
910 if (lp->cur_tx == -1) arcnet_interrupt()
H A Darc-rawmode.c121 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
H A Dcapmode.c145 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
H A Drfc1051.c215 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
H A Darcdevice.h252 int cur_tx, /* buffer used by current transmit, or -1 */ member in struct:arcnet_local
H A Drfc1201.c483 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
/linux-4.4.14/drivers/net/ethernet/packetengines/
H A Dyellowfin.c326 unsigned int cur_tx, dirty_tx; member in struct:yellowfin_private
689 yp->cur_tx, yp->dirty_tx, yellowfin_tx_timeout()
714 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) yellowfin_tx_timeout()
728 yp->cur_rx = yp->cur_tx = 0; yellowfin_init_ring()
817 "ownership" bit last, and only then increment cur_tx. */ yellowfin_start_xmit()
820 entry = yp->cur_tx % TX_RING_SIZE; yellowfin_start_xmit()
850 yp->cur_tx++; yellowfin_start_xmit()
858 yp->cur_tx++; yellowfin_start_xmit()
860 unsigned next_entry = yp->cur_tx % TX_RING_SIZE; yellowfin_start_xmit()
875 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) yellowfin_start_xmit()
882 yp->cur_tx, entry); yellowfin_start_xmit()
919 for (; yp->cur_tx - yp->dirty_tx > 0; yp->dirty_tx++) { yellowfin_interrupt()
935 yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE - 4) { yellowfin_interrupt()
944 for (dirty_tx = yp->dirty_tx; yp->cur_tx - dirty_tx > 0; yellowfin_interrupt()
996 if (yp->cur_tx - dirty_tx > TX_RING_SIZE) { yellowfin_interrupt()
998 dirty_tx, yp->cur_tx, yp->tx_full); yellowfin_interrupt()
1004 yp->cur_tx - dirty_tx < TX_QUEUE_SIZE - 2) { yellowfin_interrupt()
1202 yp->cur_tx, yp->dirty_tx, yellowfin_close()
H A Dhamachi.c495 unsigned int cur_tx, dirty_tx; member in struct:hamachi_private
998 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++) { hamachi_tx()
1112 hmp->cur_rx = hmp->cur_tx = 0; hamachi_tx_timeout()
1166 hmp->cur_rx = hmp->cur_tx = 0; hamachi_init_ring()
1221 printk(KERN_WARNING "%s: Hamachi transmit queue full at slot %d.\n",dev->name, hmp->cur_tx); hamachi_start_xmit()
1235 entry = hmp->cur_tx % TX_RING_SIZE; hamachi_start_xmit()
1257 hmp->cur_tx++; hamachi_start_xmit()
1276 if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4)) hamachi_start_xmit()
1285 dev->name, hmp->cur_tx, entry); hamachi_start_xmit()
1330 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++){ hamachi_interrupt()
1352 if (hmp->cur_tx - hmp->dirty_tx < TX_RING_SIZE - 4){ hamachi_interrupt()
1662 dev->name, hmp->cur_tx, hmp->dirty_tx, hmp->cur_rx, hmp->dirty_rx); hamachi_close()
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/
H A Dchain_mode.c35 unsigned int entry = priv->cur_tx % txsize; stmmac_jumbo_frm()
57 entry = (++priv->cur_tx) % txsize; stmmac_jumbo_frm()
H A Dring_mode.c35 unsigned int entry = priv->cur_tx % txsize; stmmac_jumbo_frm()
65 entry = (++priv->cur_tx) % txsize; stmmac_jumbo_frm()
H A Dstmmac.h55 unsigned int cur_tx; member in struct:stmmac_priv
H A Dstmmac_main.c204 return priv->dirty_tx + priv->dma_tx_size - priv->cur_tx - 1; stmmac_tx_avail()
230 if ((priv->dirty_tx == priv->cur_tx) && stmmac_enable_eee_mode()
1110 priv->cur_tx = 0; init_dma_desc_rings()
1325 while (priv->dirty_tx != priv->cur_tx) { stmmac_tx_clean()
1357 priv->cur_tx, priv->dirty_tx); stmmac_tx_clean()
1442 priv->cur_tx = 0; stmmac_tx_err()
1972 entry = priv->cur_tx % txsize; stmmac_xmit()
2007 entry = (++priv->cur_tx) % txsize; stmmac_xmit()
2050 priv->cur_tx++; stmmac_xmit()
2054 __func__, (priv->cur_tx % txsize), stmmac_xmit()
3109 priv->cur_tx = 0; stmmac_resume()
/linux-4.4.14/drivers/net/ethernet/smsc/
H A Depic100.c268 unsigned int cur_tx, dirty_tx; member in struct:epic_private
790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); epic_restart()
880 netdev_dbg(dev, "Tx indices: dirty_tx %d, cur_tx %d.\n", epic_tx_timeout()
881 ep->dirty_tx, ep->cur_tx); epic_tx_timeout()
905 ep->dirty_tx = ep->cur_tx = 0; epic_init_ring()
960 free_count = ep->cur_tx - ep->dirty_tx; epic_start_xmit()
961 entry = ep->cur_tx % TX_RING_SIZE; epic_start_xmit()
982 ep->cur_tx++; epic_start_xmit()
1021 unsigned int dirty_tx, cur_tx; epic_tx() local
1027 cur_tx = ep->cur_tx; epic_tx()
1028 for (dirty_tx = ep->dirty_tx; cur_tx - dirty_tx > 0; dirty_tx++) { epic_tx()
1052 if (cur_tx - dirty_tx > TX_RING_SIZE) { epic_tx()
1054 dirty_tx, cur_tx, ep->tx_full); epic_tx()
1059 if (ep->tx_full && cur_tx - dirty_tx < TX_QUEUE_LEN - 4) { epic_tx()
/linux-4.4.14/drivers/net/ethernet/dlink/
H A Dsundance.c398 unsigned int cur_tx, dirty_tx; member in struct:netdev_private
1000 printk(KERN_DEBUG "cur_tx=%d(%02x) dirty_tx=%d(%02x)\n", tx_timeout()
1001 np->cur_tx, np->cur_tx % TX_RING_SIZE, tx_timeout()
1016 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { tx_timeout()
1030 np->cur_rx = np->cur_tx = 0; init_ring()
1078 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; tx_poll()
1081 for (; np->cur_tx - np->cur_task > 0; np->cur_task++) { tx_poll()
1106 entry = np->cur_tx % TX_RING_SIZE; start_tx()
1119 /* Increment cur_tx before tasklet_schedule() */ start_tx()
1120 np->cur_tx++; start_tx()
1126 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 1 && start_tx()
1135 dev->name, np->cur_tx, entry); start_tx()
1173 np->cur_tx = np->dirty_tx = 0; reset_tx()
1271 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { intr_handler()
1297 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { intr_handler()
1317 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { intr_handler()
1839 np->cur_tx = 0; netdev_close()
1852 dev->name, np->cur_tx, np->dirty_tx, np->cur_rx, np->dirty_rx); netdev_close()
H A Ddl2k.c597 np->cur_rx = np->cur_tx = 0; alloc_list()
656 entry = np->cur_tx % TX_RING_SIZE; start_xmit()
693 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; start_xmit()
694 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE start_xmit()
746 if (np->cur_tx != np->old_tx) rio_interrupt()
770 while (entry != np->cur_tx) { rio_free_tx()
798 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE rio_free_tx()
H A Ddl2k.h398 unsigned long cur_tx, old_tx; member in struct:netdev_private
/linux-4.4.14/drivers/net/ethernet/
H A Dfealnx.c403 struct fealnx_desc *cur_tx; member in struct:netdev_private
1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), enable_rxtx()
1280 np->cur_tx = &np->tx_ring[0]; init_ring()
1385 np->cur_tx = &np->tx_ring[0]; reset_tx_descriptors()
1494 long tx_status = np->cur_tx->status; intr_handler()
1495 long tx_control = np->cur_tx->control; intr_handler()
1500 next = np->cur_tx->next_desc_logical; intr_handler()
1537 pci_unmap_single(np->pci_dev, np->cur_tx->buffer, intr_handler()
1538 np->cur_tx->skbuff->len, PCI_DMA_TODEVICE); intr_handler()
1539 dev_kfree_skb_irq(np->cur_tx->skbuff); intr_handler()
1540 np->cur_tx->skbuff = NULL; intr_handler()
1542 if (np->cur_tx->control & TXLD) { intr_handler()
1543 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler()
1546 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler()
1547 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler()
H A Dethoc.c186 * @cur_tx: last send buffer written
208 unsigned int cur_tx; member in struct:ethoc
309 dev->cur_tx = 0; ethoc_init_ring()
526 if (bd.stat & TX_BD_READY || (priv->dty_tx == priv->cur_tx)) { ethoc_tx()
537 (priv->dty_tx == priv->cur_tx)) ethoc_tx()
545 if ((priv->cur_tx - priv->dty_tx) <= (priv->num_tx / 2)) ethoc_tx()
868 entry = priv->cur_tx % priv->num_tx; ethoc_start_xmit()
870 priv->cur_tx++; ethoc_start_xmit()
888 if (priv->cur_tx == (priv->dty_tx + priv->num_tx)) { ethoc_start_xmit()
/linux-4.4.14/drivers/net/ethernet/dec/tulip/
H A Dtulip_core.c334 tp->cur_rx = tp->cur_tx = 0; tulip_up()
367 tp->tx_buffers[tp->cur_tx].skb = NULL; tulip_up()
368 tp->tx_buffers[tp->cur_tx].mapping = mapping; tulip_up()
371 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); tulip_up()
372 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); tulip_up()
373 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); tulip_up()
375 tp->cur_tx++; tulip_up()
677 entry = tp->cur_tx % TX_RING_SIZE; tulip_start_xmit()
685 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ tulip_start_xmit()
687 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { tulip_start_xmit()
689 } else if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE - 2) { tulip_start_xmit()
704 tp->cur_tx++; tulip_start_xmit()
718 for (dirty_tx = tp->dirty_tx ; tp->cur_tx - dirty_tx > 0; tulip_clean_tx_ring()
1142 if (tp->cur_tx - tp->dirty_tx > TX_RING_SIZE - 2) {
1150 entry = tp->cur_tx++ % TX_RING_SIZE;
1161 entry = tp->cur_tx++ % TX_RING_SIZE;
1179 if (tp->cur_tx - tp->dirty_tx >= TX_RING_SIZE - 2)
H A Dwinbond-840.c314 unsigned int cur_tx, dirty_tx; member in struct:netdev_private
834 np->tx_q_bytes = np->dirty_tx = np->cur_tx = 0; init_rxtx_rings()
948 np->cur_tx, np->dirty_tx, np->tx_full, np->tx_q_bytes); tx_timeout()
1008 entry = np->cur_tx % TX_RING_SIZE; start_tx()
1028 * increasing np->cur_tx and setting DescOwned: start_tx()
1029 * - if np->cur_tx is increased first the interrupt start_tx()
1034 * since the np->cur_tx was not yet increased. start_tx()
1037 np->cur_tx++; start_tx()
1046 if (np->cur_tx - np->dirty_tx > TX_QUEUE_LEN || start_tx()
1056 np->cur_tx, entry); start_tx()
1064 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { netdev_tx_done()
1102 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN_RESTART && netdev_tx_done()
1143 np->cur_tx != np->dirty_tx) { intr_handler()
1489 np->cur_tx, np->dirty_tx, netdev_close()
H A Dinterrupt.c588 for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0; tulip_interrupt()
644 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { tulip_interrupt()
647 dirty_tx, tp->cur_tx); tulip_interrupt()
652 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) tulip_interrupt()
H A Dtulip.h422 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:tulip_private
/linux-4.4.14/drivers/net/ethernet/3com/
H A D3c515.c310 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:corkscrew_private
842 vp->cur_tx = vp->dirty_tx = 0; corkscrew_open()
980 vp->cur_tx); corkscrew_timeout()
1013 int entry = vp->cur_tx % TX_RING_SIZE; corkscrew_start_xmit()
1020 if (vp->cur_tx != 0) corkscrew_start_xmit()
1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; corkscrew_start_xmit()
1026 dev->name, vp->cur_tx); corkscrew_start_xmit()
1050 vp->cur_tx++; corkscrew_start_xmit()
1051 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) corkscrew_start_xmit()
1175 while (lp->cur_tx - dirty_tx > 0) { corkscrew_interrupt()
1187 if (lp->tx_full && (lp->cur_tx - dirty_tx <= TX_RING_SIZE - 1)) { corkscrew_interrupt()
H A D3c59x.c604 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:vortex_private
1694 vp->cur_tx = vp->dirty_tx = 0; vortex_up()
1936 if (vp->cur_tx - vp->dirty_tx > 0 && ioread32(ioaddr + DownListPtr) == 0) vortex_tx_timeout()
1939 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { vortex_tx_timeout()
2132 int entry = vp->cur_tx % TX_RING_SIZE; boomerang_start_xmit()
2134 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; boomerang_start_xmit()
2141 dev->name, vp->cur_tx); boomerang_start_xmit()
2153 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { boomerang_start_xmit()
2238 vp->cur_tx++; boomerang_start_xmit()
2241 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) { boomerang_start_xmit()
2447 while (vp->cur_tx - dirty_tx > 0) { boomerang_interrupt()
2482 if (vp->cur_tx - dirty_tx <= TX_RING_SIZE - 1) { boomerang_interrupt()
2826 vp->cur_tx, vp->cur_tx % TX_RING_SIZE); dump_tx_ring()
/linux-4.4.14/drivers/net/ethernet/nuvoton/
H A Dw90p910_ether.c166 unsigned int cur_tx; member in struct:w90p910_ether
480 ether->cur_tx = 0x0; w90p910_reset_mac()
607 txbd = &ether->tdesc->desclist[ether->cur_tx]; w90p910_send_frame()
608 buffer = ether->tdesc->tran_buf[ether->cur_tx]; w90p910_send_frame()
625 if (++ether->cur_tx >= TX_DESC_SIZE) w90p910_send_frame()
626 ether->cur_tx = 0; w90p910_send_frame()
628 txbd = &ether->tdesc->desclist[ether->cur_tx]; w90p910_send_frame()
955 ether->cur_tx = 0x0; w90p910_ether_setup()
/linux-4.4.14/drivers/net/ethernet/natsemi/
H A Dsonic.c291 int entry = lp->cur_tx; sonic_interrupt()
295 /* At this point, cur_tx is the index of a TD that is one of: sonic_interrupt()
341 lp->cur_tx = entry; sonic_interrupt()
685 lp->cur_tx = lp->next_tx = 0; sonic_init()
H A Dnatsemi.c559 unsigned int cur_tx, dirty_tx; member in struct:netdev_private
1973 np->dirty_tx = np->cur_tx = 0; init_ring()
2079 np->dirty_tx = np->cur_tx = 0; reinit_ring()
2094 "ownership" bit last, and only then increment cur_tx. */ start_tx()
2097 entry = np->cur_tx % TX_RING_SIZE; start_tx()
2118 np->cur_tx++; start_tx()
2119 if (np->cur_tx - np->dirty_tx >= TX_QUEUE_LEN - 1) { start_tx()
2121 if (np->cur_tx - np->dirty_tx >= TX_QUEUE_LEN - 1) start_tx()
2134 dev->name, np->cur_tx, entry); start_tx()
2143 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { netdev_tx_done()
2176 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { netdev_tx_done()
3160 dev->name, np->cur_tx, np->dirty_tx, netdev_close()
H A Dsonic.h317 unsigned int cur_tx; /* first unacked transmit packet */ member in struct:sonic_local
/linux-4.4.14/drivers/net/ethernet/adaptec/
H A Dstarfire.c560 unsigned int cur_tx, dirty_tx, reap_tx; member in struct:netdev_private
1144 np->cur_rx = np->cur_tx = np->reap_tx = 0; init_ring()
1193 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { start_tx()
1205 entry = np->cur_tx % TX_RING_SIZE; start_tx()
1243 dev->name, np->cur_tx, np->dirty_tx, start_tx()
1247 np->cur_tx += np->tx_info[entry].used_slots; start_tx()
1251 np->cur_tx += np->tx_info[entry].used_slots; start_tx()
1255 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) start_tx()
1268 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) start_tx()
1367 (np->cur_tx - np->dirty_tx + 4 < TX_RING_SIZE)) { intr_handler()
1906 dev->name, np->cur_tx, np->dirty_tx, netdev_close()
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_main.c195 return queue->dirty_tx + tx_qsize - queue->cur_tx - 1; sxgbe_tx_avail()
427 tx_ring->cur_tx = 0; init_tx_ring()
749 while (tqueue->dirty_tx != tqueue->cur_tx) { sxgbe_tx_queue_clean()
762 __func__, tqueue->cur_tx, tqueue->dirty_tx); sxgbe_tx_queue_clean()
841 tx_ring->cur_tx = 0; sxgbe_restart_tx_queue()
1325 entry = tqueue->cur_tx % tx_rsize; sxgbe_xmit()
1350 entry = (++tqueue->cur_tx) % tx_rsize; sxgbe_xmit()
1372 entry = (++tqueue->cur_tx) % tx_rsize; sxgbe_xmit()
1412 tqueue->cur_tx++; sxgbe_xmit()
1416 __func__, tqueue->cur_tx % tx_rsize, sxgbe_xmit()
H A Dsxgbe_common.h388 unsigned int cur_tx; member in struct:sxgbe_tx_queue
/linux-4.4.14/drivers/net/wireless/
H A Dadm8211.c314 for (dirty_tx = priv->dirty_tx; priv->cur_tx - dirty_tx; dirty_tx++) { adm8211_interrupt_tci()
347 if (priv->cur_tx - dirty_tx < priv->tx_ring_size - 2) adm8211_interrupt_tci()
1468 priv->cur_rx = priv->cur_tx = priv->dirty_tx = 0; adm8211_init_rings()
1631 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size / 2) adm8211_tx_raw()
1636 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size - 2) adm8211_tx_raw()
1639 entry = priv->cur_tx % priv->tx_ring_size; adm8211_tx_raw()
1654 priv->cur_tx++; adm8211_tx_raw()
H A Dadm8211.h547 unsigned int cur_tx, dirty_tx, cur_rx; member in struct:adm8211_priv
/linux-4.4.14/drivers/net/ethernet/renesas/
H A Dravb_main.c232 priv->cur_tx[q] = 0; ravb_ring_format()
433 for (; priv->cur_tx[q] - priv->dirty_tx[q] > 0; priv->dirty_tx[q]++) { ravb_tx_free()
1073 data[i++] = priv->cur_tx[q]; ravb_get_ethtool_stats()
1316 if (priv->cur_tx[q] - priv->dirty_tx[q] > (priv->num_tx_ring[q] - 1) * ravb_start_xmit()
1324 entry = priv->cur_tx[q] % (priv->num_tx_ring[q] * NUM_TX_DESC); ravb_start_xmit()
1381 priv->cur_tx[q] += NUM_TX_DESC; ravb_start_xmit()
1382 if (priv->cur_tx[q] - priv->dirty_tx[q] > ravb_start_xmit()
H A Dsh_eth.c1173 mdp->cur_tx = 0; sh_eth_ring_format()
1421 for (; mdp->cur_tx - mdp->dirty_tx > 0; mdp->dirty_tx++) { sh_eth_txfree()
1694 netdev_err(ndev, "TX error. status=%8.8x cur_tx=%8.8x dirty_tx=%8.8x state=%8.8x EDTRR=%8.8x.\n", sh_eth_error()
1695 intr_status, mdp->cur_tx, mdp->dirty_tx, sh_eth_error()
2198 data[i++] = mdp->cur_tx; sh_eth_get_ethtool_stats()
2389 if ((mdp->cur_tx - mdp->dirty_tx) >= (mdp->num_tx_ring - 4)) { sh_eth_start_xmit()
2402 entry = mdp->cur_tx % mdp->num_tx_ring; sh_eth_start_xmit()
2423 mdp->cur_tx++; sh_eth_start_xmit()
H A Dsh_eth.h514 u32 cur_tx, dirty_tx; member in struct:sh_eth_private
H A Dravb.h802 u32 cur_tx[NUM_TX_QUEUE]; member in struct:ravb_private
/linux-4.4.14/drivers/net/ethernet/sis/
H A Dsis190.c272 u32 cur_tx; member in struct:sis190_private
695 pending = tp->cur_tx - dirty_tx; sis190_tx_interrupt()
808 tp->dirty_tx = tp->dirty_rx = tp->cur_tx = tp->cur_rx = 0; sis190_init_ring_indexes()
1122 tp->cur_tx = tp->dirty_tx = 0; sis190_tx_clear()
1192 entry = tp->cur_tx % NUM_TX_DESC; sis190_start_xmit()
1228 tp->cur_tx++; sis190_start_xmit()
1235 if ((tp->cur_tx - NUM_TX_DESC) == dirty_tx) { sis190_start_xmit()
H A Dsis900.c180 unsigned int cur_tx, dirty_tx; member in struct:sis900_private
1131 sis_priv->dirty_tx = sis_priv->cur_tx = 0; sis900_init_tx_ring()
1558 sis_priv->dirty_tx = sis_priv->cur_tx = 0; sis900_tx_timeout()
1610 entry = sis_priv->cur_tx % NUM_TX_DESC; sis900_start_xmit()
1627 sis_priv->cur_tx ++; sis900_start_xmit()
1628 index_cur_tx = sis_priv->cur_tx; sis900_start_xmit()
1635 /* dirty_tx is met in the cycle of cur_tx, buffer full */ sis900_start_xmit()
1892 for (; sis_priv->dirty_tx != sis_priv->cur_tx; sis_priv->dirty_tx++) { sis900_finish_xmit()
1940 sis_priv->cur_tx - sis_priv->dirty_tx < NUM_TX_DESC - 4) { sis900_finish_xmit()
/linux-4.4.14/drivers/net/ethernet/via/
H A Dvia-rhine.c475 unsigned int cur_tx, dirty_tx; member in struct:rhine_private
1326 rp->dirty_tx = rp->cur_tx = 0; alloc_tbufs()
1783 return (rp->cur_tx - rp->dirty_tx) >= TX_QUEUE_LEN; rhine_tx_queue_full()
1798 entry = rp->cur_tx % TX_RING_SIZE; rhine_start_tx()
1860 rp->cur_tx++; rhine_start_tx()
1862 * Nobody wants cur_tx write to rot for ages after the NIC will have rhine_start_tx()
1889 rp->cur_tx - 1, entry); rhine_start_tx()
1936 unsigned int cur_tx; rhine_tx() local
1941 * driver enforces a value of cur_tx that was relevant when the rhine_tx()
1946 cur_tx = rp->cur_tx; rhine_tx()
1948 while (dirty_tx != cur_tx) { rhine_tx()
2009 /* cur_tx may be optimistically out-of-sync. See rhine_start_tx. */ rhine_tx()
/linux-4.4.14/drivers/net/ethernet/freescale/
H A Dfec_main.c296 (const char *)txq->cur_tx) / fep->bufdesc_size - 1; fec_enet_get_free_txdesc_num()
336 bdp == txq->cur_tx ? 'S' : ' ', fec_dump()
373 struct bufdesc *bdp = txq->cur_tx; fec_enet_txq_submit_frag_skb()
444 bdp = txq->cur_tx; fec_enet_txq_submit_frag_skb()
483 bdp = txq->cur_tx; fec_enet_txq_submit_skb()
563 * cur_tx. fec_enet_txq_submit_skb()
566 txq->cur_tx = bdp; fec_enet_txq_submit_skb()
697 struct bufdesc *bdp = txq->cur_tx; fec_enet_txq_submit_tso()
760 txq->cur_tx = bdp; fec_enet_txq_submit_tso()
842 txq->cur_tx = bdp; fec_enet_bd_init()
1221 while (bdp != READ_ONCE(txq->cur_tx)) { fec_enet_tx_queue()
1222 /* Order the load of cur_tx and cbd_sc */ fec_enet_tx_queue()
1298 if (bdp != txq->cur_tx && fec_enet_tx_queue()
H A Dfec.h446 struct bufdesc *cur_tx; member in struct:fec_enet_priv_tx_q
465 * cur_rx and cur_tx point to the currently available buffer.
467 * controller. The cur_tx and dirty_tx are equal under both completely
H A Dgianfar.h982 * @cur_tx: Next free ring entry
995 struct txbd8 *cur_tx; member in struct:gfar_priv_tx_q
1119 * controller. The cur_tx and dirty_tx are equal under both completely
H A Dgianfar.c186 tx_queue->cur_tx = tx_queue->tx_bd_base; gfar_init_bds()
2392 txbdp = txbdp_start = tx_queue->cur_tx; gfar_start_xmit()
2524 tx_queue->cur_tx = next_txbd(txbdp, base, tx_queue->tx_ring_size); gfar_start_xmit()
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/
H A Dfs_enet-main.c398 fep->dirty_tx = fep->cur_tx = fep->tx_bd_base; fs_init_bds()
556 bdp = fep->cur_tx; fs_enet_start_xmit()
630 fep->cur_tx = bdp; fs_enet_start_xmit()
662 wake = fep->tx_free && !(CBDR_SC(fep->cur_tx) & BD_ENET_TX_READY); fs_timeout()
H A Dfs_enet.h142 cbd_t __iomem *cur_tx; member in struct:fs_enet_private
/linux-4.4.14/drivers/net/ethernet/realtek/
H A D8139too.c592 unsigned long cur_tx; member in struct:rtl8139_private
1460 tp->cur_tx = 0; rtl8139_init_ring()
1655 tp->cur_tx = 0; rtl8139_tx_clear()
1675 tp->cur_tx, tp->dirty_tx); rtl8139_tx_timeout_task()
1727 entry = tp->cur_tx % NUM_TX_DESC; rtl8139_start_xmit()
1751 tp->cur_tx++; rtl8139_start_xmit()
1753 if ((tp->cur_tx - NUM_TX_DESC) == tp->dirty_tx) rtl8139_start_xmit()
1774 tx_left = tp->cur_tx - dirty_tx; rtl8139_tx_interrupt()
1819 if (tp->cur_tx - dirty_tx > NUM_TX_DESC) { rtl8139_tx_interrupt()
1821 dirty_tx, tp->cur_tx); rtl8139_tx_interrupt()
H A Dr8169.c77 (tp->dirty_tx + NUM_TX_DESC - tp->cur_tx)
783 u32 cur_tx; /* Index into the Tx descriptor buffer of next Rx pkt. */ member in struct:rtl8169_private
4961 tp->dirty_tx = tp->cur_tx = tp->cur_rx = 0; rtl8169_init_ring_indexes()
6843 tp->cur_tx = tp->dirty_tx = 0; rtl8169_tx_clear()
6884 entry = tp->cur_tx; rtl8169_xmit_frags()
6923 rtl8169_tx_clear_range(tp, tp->cur_tx + 1, cur_frag); rtl8169_xmit_frags()
7112 unsigned int entry = tp->cur_tx % NUM_TX_DESC; rtl8169_start_xmit()
7172 tp->cur_tx += frags + 1; rtl8169_start_xmit()
7185 * - publish queue status and cur_tx ring index (write barrier) rtl8169_start_xmit()
7264 tx_left = tp->cur_tx - dirty_tx; rtl_tx()
7299 * - refresh cur_tx ring index and queue status (read barrier) rtl_tx()
7315 if (tp->cur_tx != dirty_tx) { rtl_tx()
/linux-4.4.14/drivers/staging/rdma/amso1100/
H A Dc2.c367 c2_port->c2dev->cur_tx = tx_ring->to_use - tx_ring->start; c2_tx_clean()
640 c2_port->tx_ring.start + c2dev->cur_tx; c2_up()
1062 c2dev->cur_tx = 0; c2_probe()
H A Dc2.h295 unsigned int cur_tx; member in struct:c2_dev
/linux-4.4.14/drivers/staging/rdma/hfi1/
H A Dchip.c8146 u64 cur_tx; update_synth_timer() local
8166 cur_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0); update_synth_timer()
8171 dd->unit, cur_tx, cur_rx, dd->last_tx, dd->last_rx); update_synth_timer()
8173 if ((cur_tx < dd->last_tx) || (cur_rx < dd->last_rx)) { update_synth_timer()
8182 total_flits = (cur_tx - dd->last_tx) + (cur_rx - dd->last_rx); update_synth_timer()

Completed in 3817 milliseconds