/linux-4.4.14/drivers/net/ethernet/amd/ |
H A D | ariadne.c | 92 int cur_tx, cur_rx; /* The next free ring entry */ member in struct:ariadne_private 125 priv->cur_rx = priv->cur_tx = 0; ariadne_init_ring() 308 while (dirty_tx < priv->cur_tx) { ariadne_interrupt() 345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { ariadne_interrupt() 347 dirty_tx, priv->cur_tx, ariadne_interrupt() 354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { ariadne_interrupt() 576 entry = priv->cur_tx % TX_RING_SIZE; ariadne_start_xmit() 596 priv->cur_tx++; ariadne_start_xmit() 597 if ((priv->cur_tx >= TX_RING_SIZE) && ariadne_start_xmit() 600 netdev_dbg(dev, "*** Subtracting TX_RING_SIZE from cur_tx (%d) and dirty_tx (%d)\n", ariadne_start_xmit() 601 priv->cur_tx, priv->dirty_tx); ariadne_start_xmit() 603 priv->cur_tx -= TX_RING_SIZE; ariadne_start_xmit()
|
H A D | atarilance.c | 222 int cur_rx, cur_tx; /* The next free ring entry */ member in struct:lance_private 689 lp->cur_rx = lp->cur_tx = 0; lance_init_ring() 747 DPRINTK( 2, ( "Ring data: dirty_tx %d cur_tx %d%s cur_rx %d\n", lance_tx_timeout() 748 lp->dirty_tx, lp->cur_tx, lance_tx_timeout() 814 entry = lp->cur_tx & TX_RING_MOD_MASK; lance_start_xmit() 828 lp->cur_tx++; lance_start_xmit() 829 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { lance_start_xmit() 830 lp->cur_tx -= TX_RING_SIZE; lance_start_xmit() 884 while( dirty_tx < lp->cur_tx) { lance_interrupt() 920 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { lance_interrupt() 923 dirty_tx, lp->cur_tx, lp->tx_full )); lance_interrupt() 929 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) { lance_interrupt()
|
H A D | lance.c | 248 int cur_rx, cur_tx; /* The next free ring entry */ member in struct:lance_private 867 lp->cur_rx = lp->cur_tx = 0; lance_init_ring() 930 printk (" Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.", lance_tx_timeout() 931 lp->dirty_tx, lp->cur_tx, netif_queue_stopped(dev) ? " (full)" : "", lance_tx_timeout() 971 entry = lp->cur_tx & TX_RING_MOD_MASK; lance_start_xmit() 1006 lp->cur_tx++; lance_start_xmit() 1012 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) lance_start_xmit() 1051 while (dirty_tx < lp->cur_tx) { lance_interrupt() 1095 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { lance_interrupt() 1097 dirty_tx, lp->cur_tx, lance_interrupt() 1105 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) lance_interrupt()
|
H A D | pcnet32.c | 275 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:pcnet32_private 1252 while (dirty_tx != lp->cur_tx) { pcnet32_tx() 1312 delta = (lp->cur_tx - dirty_tx) & (lp->tx_mod_mask + lp->tx_ring_size); pcnet32_tx() 1315 dirty_tx, lp->cur_tx, lp->tx_full); pcnet32_tx() 2307 lp->cur_rx = lp->cur_tx = 0; pcnet32_init_ring() 2410 " Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.", pcnet32_tx_timeout() 2411 lp->dirty_tx, lp->cur_tx, lp->tx_full ? " (full)" : "", pcnet32_tx_timeout() 2458 entry = lp->cur_tx & lp->tx_mod_mask; pcnet32_start_xmit() 2479 lp->cur_tx++; pcnet32_start_xmit()
|
/linux-4.4.14/drivers/net/arcnet/ |
H A D | arcnet.c | 473 lp->cur_tx = lp->next_tx = -1; arcnet_open() 609 lp->hw.status(dev), lp->cur_tx, lp->next_tx, skb->len, skb->protocol); arcnet_send_packet() 689 arc_printk(D_DURING, dev, "go_tx: status=%Xh, intmask=%Xh, next_tx=%d, cur_tx=%d\n", go_tx() 690 lp->hw.status(dev), lp->intmask, lp->next_tx, lp->cur_tx); go_tx() 692 if (lp->cur_tx != -1 || lp->next_tx == -1) go_tx() 698 lp->cur_tx = lp->next_tx; go_tx() 702 lp->hw.command(dev, TXcmd | (lp->cur_tx << 3)); go_tx() 728 lp->hw.command(dev, NOTXcmd | (lp->cur_tx << 3)); arcnet_timeout() 745 if (lp->cur_tx == -1) arcnet_timeout() 850 if (lp->cur_tx != -1 && !lp->timed_out) { arcnet_interrupt() 882 if (lp->cur_tx != -1) arcnet_interrupt() 883 release_arcbuf(dev, lp->cur_tx); arcnet_interrupt() 885 lp->cur_tx = -1; arcnet_interrupt() 910 if (lp->cur_tx == -1) arcnet_interrupt()
|
H A D | arc-rawmode.c | 121 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
|
H A D | capmode.c | 145 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
|
H A D | rfc1051.c | 215 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
|
H A D | arcdevice.h | 252 int cur_tx, /* buffer used by current transmit, or -1 */ member in struct:arcnet_local
|
H A D | rfc1201.c | 483 lp->next_tx, lp->cur_tx, bufnum); prepare_tx()
|
/linux-4.4.14/drivers/net/ethernet/packetengines/ |
H A D | yellowfin.c | 326 unsigned int cur_tx, dirty_tx; member in struct:yellowfin_private 689 yp->cur_tx, yp->dirty_tx, yellowfin_tx_timeout() 714 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) yellowfin_tx_timeout() 728 yp->cur_rx = yp->cur_tx = 0; yellowfin_init_ring() 817 "ownership" bit last, and only then increment cur_tx. */ yellowfin_start_xmit() 820 entry = yp->cur_tx % TX_RING_SIZE; yellowfin_start_xmit() 850 yp->cur_tx++; yellowfin_start_xmit() 858 yp->cur_tx++; yellowfin_start_xmit() 860 unsigned next_entry = yp->cur_tx % TX_RING_SIZE; yellowfin_start_xmit() 875 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) yellowfin_start_xmit() 882 yp->cur_tx, entry); yellowfin_start_xmit() 919 for (; yp->cur_tx - yp->dirty_tx > 0; yp->dirty_tx++) { yellowfin_interrupt() 935 yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE - 4) { yellowfin_interrupt() 944 for (dirty_tx = yp->dirty_tx; yp->cur_tx - dirty_tx > 0; yellowfin_interrupt() 996 if (yp->cur_tx - dirty_tx > TX_RING_SIZE) { yellowfin_interrupt() 998 dirty_tx, yp->cur_tx, yp->tx_full); yellowfin_interrupt() 1004 yp->cur_tx - dirty_tx < TX_QUEUE_SIZE - 2) { yellowfin_interrupt() 1202 yp->cur_tx, yp->dirty_tx, yellowfin_close()
|
H A D | hamachi.c | 495 unsigned int cur_tx, dirty_tx; member in struct:hamachi_private 998 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++) { hamachi_tx() 1112 hmp->cur_rx = hmp->cur_tx = 0; hamachi_tx_timeout() 1166 hmp->cur_rx = hmp->cur_tx = 0; hamachi_init_ring() 1221 printk(KERN_WARNING "%s: Hamachi transmit queue full at slot %d.\n",dev->name, hmp->cur_tx); hamachi_start_xmit() 1235 entry = hmp->cur_tx % TX_RING_SIZE; hamachi_start_xmit() 1257 hmp->cur_tx++; hamachi_start_xmit() 1276 if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4)) hamachi_start_xmit() 1285 dev->name, hmp->cur_tx, entry); hamachi_start_xmit() 1330 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++){ hamachi_interrupt() 1352 if (hmp->cur_tx - hmp->dirty_tx < TX_RING_SIZE - 4){ hamachi_interrupt() 1662 dev->name, hmp->cur_tx, hmp->dirty_tx, hmp->cur_rx, hmp->dirty_rx); hamachi_close()
|
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/ |
H A D | chain_mode.c | 35 unsigned int entry = priv->cur_tx % txsize; stmmac_jumbo_frm() 57 entry = (++priv->cur_tx) % txsize; stmmac_jumbo_frm()
|
H A D | ring_mode.c | 35 unsigned int entry = priv->cur_tx % txsize; stmmac_jumbo_frm() 65 entry = (++priv->cur_tx) % txsize; stmmac_jumbo_frm()
|
H A D | stmmac.h | 55 unsigned int cur_tx; member in struct:stmmac_priv
|
H A D | stmmac_main.c | 204 return priv->dirty_tx + priv->dma_tx_size - priv->cur_tx - 1; stmmac_tx_avail() 230 if ((priv->dirty_tx == priv->cur_tx) && stmmac_enable_eee_mode() 1110 priv->cur_tx = 0; init_dma_desc_rings() 1325 while (priv->dirty_tx != priv->cur_tx) { stmmac_tx_clean() 1357 priv->cur_tx, priv->dirty_tx); stmmac_tx_clean() 1442 priv->cur_tx = 0; stmmac_tx_err() 1972 entry = priv->cur_tx % txsize; stmmac_xmit() 2007 entry = (++priv->cur_tx) % txsize; stmmac_xmit() 2050 priv->cur_tx++; stmmac_xmit() 2054 __func__, (priv->cur_tx % txsize), stmmac_xmit() 3109 priv->cur_tx = 0; stmmac_resume()
|
/linux-4.4.14/drivers/net/ethernet/smsc/ |
H A D | epic100.c | 268 unsigned int cur_tx, dirty_tx; member in struct:epic_private 790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); epic_restart() 880 netdev_dbg(dev, "Tx indices: dirty_tx %d, cur_tx %d.\n", epic_tx_timeout() 881 ep->dirty_tx, ep->cur_tx); epic_tx_timeout() 905 ep->dirty_tx = ep->cur_tx = 0; epic_init_ring() 960 free_count = ep->cur_tx - ep->dirty_tx; epic_start_xmit() 961 entry = ep->cur_tx % TX_RING_SIZE; epic_start_xmit() 982 ep->cur_tx++; epic_start_xmit() 1021 unsigned int dirty_tx, cur_tx; epic_tx() local 1027 cur_tx = ep->cur_tx; epic_tx() 1028 for (dirty_tx = ep->dirty_tx; cur_tx - dirty_tx > 0; dirty_tx++) { epic_tx() 1052 if (cur_tx - dirty_tx > TX_RING_SIZE) { epic_tx() 1054 dirty_tx, cur_tx, ep->tx_full); epic_tx() 1059 if (ep->tx_full && cur_tx - dirty_tx < TX_QUEUE_LEN - 4) { epic_tx()
|
/linux-4.4.14/drivers/net/ethernet/dlink/ |
H A D | sundance.c | 398 unsigned int cur_tx, dirty_tx; member in struct:netdev_private 1000 printk(KERN_DEBUG "cur_tx=%d(%02x) dirty_tx=%d(%02x)\n", tx_timeout() 1001 np->cur_tx, np->cur_tx % TX_RING_SIZE, tx_timeout() 1016 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { tx_timeout() 1030 np->cur_rx = np->cur_tx = 0; init_ring() 1078 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; tx_poll() 1081 for (; np->cur_tx - np->cur_task > 0; np->cur_task++) { tx_poll() 1106 entry = np->cur_tx % TX_RING_SIZE; start_tx() 1119 /* Increment cur_tx before tasklet_schedule() */ start_tx() 1120 np->cur_tx++; start_tx() 1126 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 1 && start_tx() 1135 dev->name, np->cur_tx, entry); start_tx() 1173 np->cur_tx = np->dirty_tx = 0; reset_tx() 1271 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { intr_handler() 1297 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { intr_handler() 1317 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { intr_handler() 1839 np->cur_tx = 0; netdev_close() 1852 dev->name, np->cur_tx, np->dirty_tx, np->cur_rx, np->dirty_rx); netdev_close()
|
H A D | dl2k.c | 597 np->cur_rx = np->cur_tx = 0; alloc_list() 656 entry = np->cur_tx % TX_RING_SIZE; start_xmit() 693 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; start_xmit() 694 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE start_xmit() 746 if (np->cur_tx != np->old_tx) rio_interrupt() 770 while (entry != np->cur_tx) { rio_free_tx() 798 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE rio_free_tx()
|
H A D | dl2k.h | 398 unsigned long cur_tx, old_tx; member in struct:netdev_private
|
/linux-4.4.14/drivers/net/ethernet/ |
H A D | fealnx.c | 403 struct fealnx_desc *cur_tx; member in struct:netdev_private 1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), enable_rxtx() 1280 np->cur_tx = &np->tx_ring[0]; init_ring() 1385 np->cur_tx = &np->tx_ring[0]; reset_tx_descriptors() 1494 long tx_status = np->cur_tx->status; intr_handler() 1495 long tx_control = np->cur_tx->control; intr_handler() 1500 next = np->cur_tx->next_desc_logical; intr_handler() 1537 pci_unmap_single(np->pci_dev, np->cur_tx->buffer, intr_handler() 1538 np->cur_tx->skbuff->len, PCI_DMA_TODEVICE); intr_handler() 1539 dev_kfree_skb_irq(np->cur_tx->skbuff); intr_handler() 1540 np->cur_tx->skbuff = NULL; intr_handler() 1542 if (np->cur_tx->control & TXLD) { intr_handler() 1543 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler() 1546 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler() 1547 np->cur_tx = np->cur_tx->next_desc_logical; intr_handler()
|
H A D | ethoc.c | 186 * @cur_tx: last send buffer written 208 unsigned int cur_tx; member in struct:ethoc 309 dev->cur_tx = 0; ethoc_init_ring() 526 if (bd.stat & TX_BD_READY || (priv->dty_tx == priv->cur_tx)) { ethoc_tx() 537 (priv->dty_tx == priv->cur_tx)) ethoc_tx() 545 if ((priv->cur_tx - priv->dty_tx) <= (priv->num_tx / 2)) ethoc_tx() 868 entry = priv->cur_tx % priv->num_tx; ethoc_start_xmit() 870 priv->cur_tx++; ethoc_start_xmit() 888 if (priv->cur_tx == (priv->dty_tx + priv->num_tx)) { ethoc_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/dec/tulip/ |
H A D | tulip_core.c | 334 tp->cur_rx = tp->cur_tx = 0; tulip_up() 367 tp->tx_buffers[tp->cur_tx].skb = NULL; tulip_up() 368 tp->tx_buffers[tp->cur_tx].mapping = mapping; tulip_up() 371 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); tulip_up() 372 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); tulip_up() 373 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); tulip_up() 375 tp->cur_tx++; tulip_up() 677 entry = tp->cur_tx % TX_RING_SIZE; tulip_start_xmit() 685 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ tulip_start_xmit() 687 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { tulip_start_xmit() 689 } else if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE - 2) { tulip_start_xmit() 704 tp->cur_tx++; tulip_start_xmit() 718 for (dirty_tx = tp->dirty_tx ; tp->cur_tx - dirty_tx > 0; tulip_clean_tx_ring() 1142 if (tp->cur_tx - tp->dirty_tx > TX_RING_SIZE - 2) { 1150 entry = tp->cur_tx++ % TX_RING_SIZE; 1161 entry = tp->cur_tx++ % TX_RING_SIZE; 1179 if (tp->cur_tx - tp->dirty_tx >= TX_RING_SIZE - 2)
|
H A D | winbond-840.c | 314 unsigned int cur_tx, dirty_tx; member in struct:netdev_private 834 np->tx_q_bytes = np->dirty_tx = np->cur_tx = 0; init_rxtx_rings() 948 np->cur_tx, np->dirty_tx, np->tx_full, np->tx_q_bytes); tx_timeout() 1008 entry = np->cur_tx % TX_RING_SIZE; start_tx() 1028 * increasing np->cur_tx and setting DescOwned: start_tx() 1029 * - if np->cur_tx is increased first the interrupt start_tx() 1034 * since the np->cur_tx was not yet increased. start_tx() 1037 np->cur_tx++; start_tx() 1046 if (np->cur_tx - np->dirty_tx > TX_QUEUE_LEN || start_tx() 1056 np->cur_tx, entry); start_tx() 1064 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { netdev_tx_done() 1102 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN_RESTART && netdev_tx_done() 1143 np->cur_tx != np->dirty_tx) { intr_handler() 1489 np->cur_tx, np->dirty_tx, netdev_close()
|
H A D | interrupt.c | 588 for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0; tulip_interrupt() 644 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { tulip_interrupt() 647 dirty_tx, tp->cur_tx); tulip_interrupt() 652 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) tulip_interrupt()
|
H A D | tulip.h | 422 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:tulip_private
|
/linux-4.4.14/drivers/net/ethernet/3com/ |
H A D | 3c515.c | 310 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:corkscrew_private 842 vp->cur_tx = vp->dirty_tx = 0; corkscrew_open() 980 vp->cur_tx); corkscrew_timeout() 1013 int entry = vp->cur_tx % TX_RING_SIZE; corkscrew_start_xmit() 1020 if (vp->cur_tx != 0) corkscrew_start_xmit() 1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; corkscrew_start_xmit() 1026 dev->name, vp->cur_tx); corkscrew_start_xmit() 1050 vp->cur_tx++; corkscrew_start_xmit() 1051 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) corkscrew_start_xmit() 1175 while (lp->cur_tx - dirty_tx > 0) { corkscrew_interrupt() 1187 if (lp->tx_full && (lp->cur_tx - dirty_tx <= TX_RING_SIZE - 1)) { corkscrew_interrupt()
|
H A D | 3c59x.c | 604 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member in struct:vortex_private 1694 vp->cur_tx = vp->dirty_tx = 0; vortex_up() 1936 if (vp->cur_tx - vp->dirty_tx > 0 && ioread32(ioaddr + DownListPtr) == 0) vortex_tx_timeout() 1939 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { vortex_tx_timeout() 2132 int entry = vp->cur_tx % TX_RING_SIZE; boomerang_start_xmit() 2134 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; boomerang_start_xmit() 2141 dev->name, vp->cur_tx); boomerang_start_xmit() 2153 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { boomerang_start_xmit() 2238 vp->cur_tx++; boomerang_start_xmit() 2241 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) { boomerang_start_xmit() 2447 while (vp->cur_tx - dirty_tx > 0) { boomerang_interrupt() 2482 if (vp->cur_tx - dirty_tx <= TX_RING_SIZE - 1) { boomerang_interrupt() 2826 vp->cur_tx, vp->cur_tx % TX_RING_SIZE); dump_tx_ring()
|
/linux-4.4.14/drivers/net/ethernet/nuvoton/ |
H A D | w90p910_ether.c | 166 unsigned int cur_tx; member in struct:w90p910_ether 480 ether->cur_tx = 0x0; w90p910_reset_mac() 607 txbd = ðer->tdesc->desclist[ether->cur_tx]; w90p910_send_frame() 608 buffer = ether->tdesc->tran_buf[ether->cur_tx]; w90p910_send_frame() 625 if (++ether->cur_tx >= TX_DESC_SIZE) w90p910_send_frame() 626 ether->cur_tx = 0; w90p910_send_frame() 628 txbd = ðer->tdesc->desclist[ether->cur_tx]; w90p910_send_frame() 955 ether->cur_tx = 0x0; w90p910_ether_setup()
|
/linux-4.4.14/drivers/net/ethernet/natsemi/ |
H A D | sonic.c | 291 int entry = lp->cur_tx; sonic_interrupt() 295 /* At this point, cur_tx is the index of a TD that is one of: sonic_interrupt() 341 lp->cur_tx = entry; sonic_interrupt() 685 lp->cur_tx = lp->next_tx = 0; sonic_init()
|
H A D | natsemi.c | 559 unsigned int cur_tx, dirty_tx; member in struct:netdev_private 1973 np->dirty_tx = np->cur_tx = 0; init_ring() 2079 np->dirty_tx = np->cur_tx = 0; reinit_ring() 2094 "ownership" bit last, and only then increment cur_tx. */ start_tx() 2097 entry = np->cur_tx % TX_RING_SIZE; start_tx() 2118 np->cur_tx++; start_tx() 2119 if (np->cur_tx - np->dirty_tx >= TX_QUEUE_LEN - 1) { start_tx() 2121 if (np->cur_tx - np->dirty_tx >= TX_QUEUE_LEN - 1) start_tx() 2134 dev->name, np->cur_tx, entry); start_tx() 2143 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { netdev_tx_done() 2176 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { netdev_tx_done() 3160 dev->name, np->cur_tx, np->dirty_tx, netdev_close()
|
H A D | sonic.h | 317 unsigned int cur_tx; /* first unacked transmit packet */ member in struct:sonic_local
|
/linux-4.4.14/drivers/net/ethernet/adaptec/ |
H A D | starfire.c | 560 unsigned int cur_tx, dirty_tx, reap_tx; member in struct:netdev_private 1144 np->cur_rx = np->cur_tx = np->reap_tx = 0; init_ring() 1193 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { start_tx() 1205 entry = np->cur_tx % TX_RING_SIZE; start_tx() 1243 dev->name, np->cur_tx, np->dirty_tx, start_tx() 1247 np->cur_tx += np->tx_info[entry].used_slots; start_tx() 1251 np->cur_tx += np->tx_info[entry].used_slots; start_tx() 1255 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) start_tx() 1268 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) start_tx() 1367 (np->cur_tx - np->dirty_tx + 4 < TX_RING_SIZE)) { intr_handler() 1906 dev->name, np->cur_tx, np->dirty_tx, netdev_close()
|
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_main.c | 195 return queue->dirty_tx + tx_qsize - queue->cur_tx - 1; sxgbe_tx_avail() 427 tx_ring->cur_tx = 0; init_tx_ring() 749 while (tqueue->dirty_tx != tqueue->cur_tx) { sxgbe_tx_queue_clean() 762 __func__, tqueue->cur_tx, tqueue->dirty_tx); sxgbe_tx_queue_clean() 841 tx_ring->cur_tx = 0; sxgbe_restart_tx_queue() 1325 entry = tqueue->cur_tx % tx_rsize; sxgbe_xmit() 1350 entry = (++tqueue->cur_tx) % tx_rsize; sxgbe_xmit() 1372 entry = (++tqueue->cur_tx) % tx_rsize; sxgbe_xmit() 1412 tqueue->cur_tx++; sxgbe_xmit() 1416 __func__, tqueue->cur_tx % tx_rsize, sxgbe_xmit()
|
H A D | sxgbe_common.h | 388 unsigned int cur_tx; member in struct:sxgbe_tx_queue
|
/linux-4.4.14/drivers/net/wireless/ |
H A D | adm8211.c | 314 for (dirty_tx = priv->dirty_tx; priv->cur_tx - dirty_tx; dirty_tx++) { adm8211_interrupt_tci() 347 if (priv->cur_tx - dirty_tx < priv->tx_ring_size - 2) adm8211_interrupt_tci() 1468 priv->cur_rx = priv->cur_tx = priv->dirty_tx = 0; adm8211_init_rings() 1631 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size / 2) adm8211_tx_raw() 1636 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size - 2) adm8211_tx_raw() 1639 entry = priv->cur_tx % priv->tx_ring_size; adm8211_tx_raw() 1654 priv->cur_tx++; adm8211_tx_raw()
|
H A D | adm8211.h | 547 unsigned int cur_tx, dirty_tx, cur_rx; member in struct:adm8211_priv
|
/linux-4.4.14/drivers/net/ethernet/renesas/ |
H A D | ravb_main.c | 232 priv->cur_tx[q] = 0; ravb_ring_format() 433 for (; priv->cur_tx[q] - priv->dirty_tx[q] > 0; priv->dirty_tx[q]++) { ravb_tx_free() 1073 data[i++] = priv->cur_tx[q]; ravb_get_ethtool_stats() 1316 if (priv->cur_tx[q] - priv->dirty_tx[q] > (priv->num_tx_ring[q] - 1) * ravb_start_xmit() 1324 entry = priv->cur_tx[q] % (priv->num_tx_ring[q] * NUM_TX_DESC); ravb_start_xmit() 1381 priv->cur_tx[q] += NUM_TX_DESC; ravb_start_xmit() 1382 if (priv->cur_tx[q] - priv->dirty_tx[q] > ravb_start_xmit()
|
H A D | sh_eth.c | 1173 mdp->cur_tx = 0; sh_eth_ring_format() 1421 for (; mdp->cur_tx - mdp->dirty_tx > 0; mdp->dirty_tx++) { sh_eth_txfree() 1694 netdev_err(ndev, "TX error. status=%8.8x cur_tx=%8.8x dirty_tx=%8.8x state=%8.8x EDTRR=%8.8x.\n", sh_eth_error() 1695 intr_status, mdp->cur_tx, mdp->dirty_tx, sh_eth_error() 2198 data[i++] = mdp->cur_tx; sh_eth_get_ethtool_stats() 2389 if ((mdp->cur_tx - mdp->dirty_tx) >= (mdp->num_tx_ring - 4)) { sh_eth_start_xmit() 2402 entry = mdp->cur_tx % mdp->num_tx_ring; sh_eth_start_xmit() 2423 mdp->cur_tx++; sh_eth_start_xmit()
|
H A D | sh_eth.h | 514 u32 cur_tx, dirty_tx; member in struct:sh_eth_private
|
H A D | ravb.h | 802 u32 cur_tx[NUM_TX_QUEUE]; member in struct:ravb_private
|
/linux-4.4.14/drivers/net/ethernet/sis/ |
H A D | sis190.c | 272 u32 cur_tx; member in struct:sis190_private 695 pending = tp->cur_tx - dirty_tx; sis190_tx_interrupt() 808 tp->dirty_tx = tp->dirty_rx = tp->cur_tx = tp->cur_rx = 0; sis190_init_ring_indexes() 1122 tp->cur_tx = tp->dirty_tx = 0; sis190_tx_clear() 1192 entry = tp->cur_tx % NUM_TX_DESC; sis190_start_xmit() 1228 tp->cur_tx++; sis190_start_xmit() 1235 if ((tp->cur_tx - NUM_TX_DESC) == dirty_tx) { sis190_start_xmit()
|
H A D | sis900.c | 180 unsigned int cur_tx, dirty_tx; member in struct:sis900_private 1131 sis_priv->dirty_tx = sis_priv->cur_tx = 0; sis900_init_tx_ring() 1558 sis_priv->dirty_tx = sis_priv->cur_tx = 0; sis900_tx_timeout() 1610 entry = sis_priv->cur_tx % NUM_TX_DESC; sis900_start_xmit() 1627 sis_priv->cur_tx ++; sis900_start_xmit() 1628 index_cur_tx = sis_priv->cur_tx; sis900_start_xmit() 1635 /* dirty_tx is met in the cycle of cur_tx, buffer full */ sis900_start_xmit() 1892 for (; sis_priv->dirty_tx != sis_priv->cur_tx; sis_priv->dirty_tx++) { sis900_finish_xmit() 1940 sis_priv->cur_tx - sis_priv->dirty_tx < NUM_TX_DESC - 4) { sis900_finish_xmit()
|
/linux-4.4.14/drivers/net/ethernet/via/ |
H A D | via-rhine.c | 475 unsigned int cur_tx, dirty_tx; member in struct:rhine_private 1326 rp->dirty_tx = rp->cur_tx = 0; alloc_tbufs() 1783 return (rp->cur_tx - rp->dirty_tx) >= TX_QUEUE_LEN; rhine_tx_queue_full() 1798 entry = rp->cur_tx % TX_RING_SIZE; rhine_start_tx() 1860 rp->cur_tx++; rhine_start_tx() 1862 * Nobody wants cur_tx write to rot for ages after the NIC will have rhine_start_tx() 1889 rp->cur_tx - 1, entry); rhine_start_tx() 1936 unsigned int cur_tx; rhine_tx() local 1941 * driver enforces a value of cur_tx that was relevant when the rhine_tx() 1946 cur_tx = rp->cur_tx; rhine_tx() 1948 while (dirty_tx != cur_tx) { rhine_tx() 2009 /* cur_tx may be optimistically out-of-sync. See rhine_start_tx. */ rhine_tx()
|
/linux-4.4.14/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 296 (const char *)txq->cur_tx) / fep->bufdesc_size - 1; fec_enet_get_free_txdesc_num() 336 bdp == txq->cur_tx ? 'S' : ' ', fec_dump() 373 struct bufdesc *bdp = txq->cur_tx; fec_enet_txq_submit_frag_skb() 444 bdp = txq->cur_tx; fec_enet_txq_submit_frag_skb() 483 bdp = txq->cur_tx; fec_enet_txq_submit_skb() 563 * cur_tx. fec_enet_txq_submit_skb() 566 txq->cur_tx = bdp; fec_enet_txq_submit_skb() 697 struct bufdesc *bdp = txq->cur_tx; fec_enet_txq_submit_tso() 760 txq->cur_tx = bdp; fec_enet_txq_submit_tso() 842 txq->cur_tx = bdp; fec_enet_bd_init() 1221 while (bdp != READ_ONCE(txq->cur_tx)) { fec_enet_tx_queue() 1222 /* Order the load of cur_tx and cbd_sc */ fec_enet_tx_queue() 1298 if (bdp != txq->cur_tx && fec_enet_tx_queue()
|
H A D | fec.h | 446 struct bufdesc *cur_tx; member in struct:fec_enet_priv_tx_q 465 * cur_rx and cur_tx point to the currently available buffer. 467 * controller. The cur_tx and dirty_tx are equal under both completely
|
H A D | gianfar.h | 982 * @cur_tx: Next free ring entry 995 struct txbd8 *cur_tx; member in struct:gfar_priv_tx_q 1119 * controller. The cur_tx and dirty_tx are equal under both completely
|
H A D | gianfar.c | 186 tx_queue->cur_tx = tx_queue->tx_bd_base; gfar_init_bds() 2392 txbdp = txbdp_start = tx_queue->cur_tx; gfar_start_xmit() 2524 tx_queue->cur_tx = next_txbd(txbdp, base, tx_queue->tx_ring_size); gfar_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/ |
H A D | fs_enet-main.c | 398 fep->dirty_tx = fep->cur_tx = fep->tx_bd_base; fs_init_bds() 556 bdp = fep->cur_tx; fs_enet_start_xmit() 630 fep->cur_tx = bdp; fs_enet_start_xmit() 662 wake = fep->tx_free && !(CBDR_SC(fep->cur_tx) & BD_ENET_TX_READY); fs_timeout()
|
H A D | fs_enet.h | 142 cbd_t __iomem *cur_tx; member in struct:fs_enet_private
|
/linux-4.4.14/drivers/net/ethernet/realtek/ |
H A D | 8139too.c | 592 unsigned long cur_tx; member in struct:rtl8139_private 1460 tp->cur_tx = 0; rtl8139_init_ring() 1655 tp->cur_tx = 0; rtl8139_tx_clear() 1675 tp->cur_tx, tp->dirty_tx); rtl8139_tx_timeout_task() 1727 entry = tp->cur_tx % NUM_TX_DESC; rtl8139_start_xmit() 1751 tp->cur_tx++; rtl8139_start_xmit() 1753 if ((tp->cur_tx - NUM_TX_DESC) == tp->dirty_tx) rtl8139_start_xmit() 1774 tx_left = tp->cur_tx - dirty_tx; rtl8139_tx_interrupt() 1819 if (tp->cur_tx - dirty_tx > NUM_TX_DESC) { rtl8139_tx_interrupt() 1821 dirty_tx, tp->cur_tx); rtl8139_tx_interrupt()
|
H A D | r8169.c | 77 (tp->dirty_tx + NUM_TX_DESC - tp->cur_tx) 783 u32 cur_tx; /* Index into the Tx descriptor buffer of next Rx pkt. */ member in struct:rtl8169_private 4961 tp->dirty_tx = tp->cur_tx = tp->cur_rx = 0; rtl8169_init_ring_indexes() 6843 tp->cur_tx = tp->dirty_tx = 0; rtl8169_tx_clear() 6884 entry = tp->cur_tx; rtl8169_xmit_frags() 6923 rtl8169_tx_clear_range(tp, tp->cur_tx + 1, cur_frag); rtl8169_xmit_frags() 7112 unsigned int entry = tp->cur_tx % NUM_TX_DESC; rtl8169_start_xmit() 7172 tp->cur_tx += frags + 1; rtl8169_start_xmit() 7185 * - publish queue status and cur_tx ring index (write barrier) rtl8169_start_xmit() 7264 tx_left = tp->cur_tx - dirty_tx; rtl_tx() 7299 * - refresh cur_tx ring index and queue status (read barrier) rtl_tx() 7315 if (tp->cur_tx != dirty_tx) { rtl_tx()
|
/linux-4.4.14/drivers/staging/rdma/amso1100/ |
H A D | c2.c | 367 c2_port->c2dev->cur_tx = tx_ring->to_use - tx_ring->start; c2_tx_clean() 640 c2_port->tx_ring.start + c2dev->cur_tx; c2_up() 1062 c2dev->cur_tx = 0; c2_probe()
|
H A D | c2.h | 295 unsigned int cur_tx; member in struct:c2_dev
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
H A D | chip.c | 8146 u64 cur_tx; update_synth_timer() local 8166 cur_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0); update_synth_timer() 8171 dd->unit, cur_tx, cur_rx, dd->last_tx, dd->last_rx); update_synth_timer() 8173 if ((cur_tx < dd->last_tx) || (cur_rx < dd->last_rx)) { update_synth_timer() 8182 total_flits = (cur_tx - dd->last_tx) + (cur_rx - dd->last_rx); update_synth_timer()
|