Home
last modified time | relevance | path

Searched refs:TX_RING_SIZE (Results 1 – 49 of 49) sorted by relevance

/linux-4.4.14/drivers/net/ethernet/sun/
Dsungem.h882 #define TX_RING_SIZE 128 macro
885 #if TX_RING_SIZE == 32
887 #elif TX_RING_SIZE == 64
889 #elif TX_RING_SIZE == 128
891 #elif TX_RING_SIZE == 256
893 #elif TX_RING_SIZE == 512
895 #elif TX_RING_SIZE == 1024
897 #elif TX_RING_SIZE == 2048
899 #elif TX_RING_SIZE == 4096
901 #elif TX_RING_SIZE == 8192
[all …]
Dsunhme.h330 #define TX_RING_SIZE 32 /* Must be >16 and <255, multiple of 16 */ macro
333 #if (TX_RING_SIZE < 16 || TX_RING_SIZE > 256 || (TX_RING_SIZE % 16) != 0)
334 #error TX_RING_SIZE holds illegal value
360 #define NEXT_TX(num) (((num) + 1) & (TX_RING_SIZE - 1))
362 #define PREV_TX(num) (((num) - 1) & (TX_RING_SIZE - 1))
366 (hp)->tx_old + (TX_RING_SIZE - 1) - (hp)->tx_new : \
417 struct sk_buff *tx_skbs[TX_RING_SIZE];
Dsunbmac.h250 #define TX_RING_SIZE 256 macro
254 #define NEXT_TX(num) (((num) + 1) & (TX_RING_SIZE - 1))
256 #define PREV_TX(num) (((num) - 1) & (TX_RING_SIZE - 1))
260 (bp)->tx_old + (TX_RING_SIZE - 1) - (bp)->tx_new : \
299 struct sk_buff *tx_skbs[TX_RING_SIZE];
Dsunqe.h290 #define TX_RING_SIZE 16 macro
300 (qp)->tx_old + (TX_RING_SIZE - 1) - (qp)->tx_new : \
325 u8 tx_buf[TX_RING_SIZE][PKT_BUF_SZ];
Dsunqe.c586 txbuf = &qbufs->tx_buf[entry & (TX_RING_SIZE - 1)][0]; in qe_start_xmit()
588 qebuf_offset(tx_buf, (entry & (TX_RING_SIZE - 1))); in qe_start_xmit()
Dsungem.c666 last &= (TX_RING_SIZE - 1); in gem_tx()
992 if (!(entry & ((TX_RING_SIZE>>1)-1))) in gem_intme()
1592 for (i = 0; i < TX_RING_SIZE; i++) { in gem_clean_rings()
1601 int ent = i & (TX_RING_SIZE - 1); in gem_clean_rings()
1655 for (i = 0; i < TX_RING_SIZE; i++) { in gem_init_rings()
Dsunhme.c144 for (i = 0; i < TX_RING_SIZE; i+=4) { in tx_dump_ring()
1216 for (i = 0; i < TX_RING_SIZE; i++) { in happy_meal_clean_rings()
1290 for (i = 0; i < TX_RING_SIZE; i++) in happy_meal_init_rings()
1622 HMD(("tx rsize=%d oreg[%08x], ", (int)TX_RING_SIZE, in happy_meal_init()
1624 hme_write32(hp, etxregs + ETX_RSIZE, (TX_RING_SIZE >> ETX_RSIZE_SHIFT) - 1); in happy_meal_init()
1943 last &= (TX_RING_SIZE - 1); in happy_meal_tx()
Dsunbmac.c203 for (i = 0; i < TX_RING_SIZE; i++) { in bigmac_clean_rings()
248 for (i = 0; i < TX_RING_SIZE; i++) in bigmac_init_rings()
/linux-4.4.14/drivers/net/ethernet/amd/
Dariadne.c80 #define TX_RING_SIZE 5 macro
88 volatile struct TDRE *tx_ring[TX_RING_SIZE];
90 volatile u_short *tx_buff[TX_RING_SIZE];
100 struct TDRE tx_ring[TX_RING_SIZE];
102 u_short tx_buff[TX_RING_SIZE][PKT_BUF_SIZE / sizeof(u_short)];
129 for (i = 0; i < TX_RING_SIZE; i++) { in ariadne_init_ring()
309 int entry = dirty_tx % TX_RING_SIZE; in ariadne_interrupt()
345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { in ariadne_interrupt()
349 dirty_tx += TX_RING_SIZE; in ariadne_interrupt()
354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { in ariadne_interrupt()
[all …]
D7990.h38 #define TX_RING_SIZE (1 << LANCE_LOG_TX_BUFFERS) macro
40 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
86 volatile struct lance_tx_desc btx_ring[TX_RING_SIZE];
89 volatile char tx_buf[TX_RING_SIZE][TX_BUFF_SIZE];
Datarilance.c111 #define TX_RING_SIZE (1 << TX_LOG_RING_SIZE) macro
113 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
157 struct lance_tx_head tx_head[TX_RING_SIZE];
705 for( i = 0; i < TX_RING_SIZE; i++ ) { in lance_init_ring()
756 for( i = 0 ; i < TX_RING_SIZE; i++ ) in lance_tx_timeout()
829 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { in lance_start_xmit()
830 lp->cur_tx -= TX_RING_SIZE; in lance_start_xmit()
831 lp->dirty_tx -= TX_RING_SIZE; in lance_start_xmit()
920 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
924 dirty_tx += TX_RING_SIZE; in lance_interrupt()
[all …]
Dlance.c193 #define TX_RING_SIZE (1 << (LANCE_LOG_TX_BUFFERS)) macro
194 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
238 struct lance_tx_head tx_ring[TX_RING_SIZE];
242 struct sk_buff* tx_skbuff[TX_RING_SIZE];
560 lp->tx_bounce_buffs = kmalloc(PKT_BUF_SZ*TX_RING_SIZE, in lance_probe1()
851 for (i = 0; i < TX_RING_SIZE; i++) { in lance_purge_ring()
888 for (i = 0; i < TX_RING_SIZE; i++) { in lance_init_ring()
937 for (i = 0; i < TX_RING_SIZE; i++) in lance_tx_timeout()
1012 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) in lance_start_xmit()
1095 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
[all …]
Dsun3lance.c93 #define TX_RING_SIZE (1 << TX_LOG_RING_SIZE) macro
95 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
141 struct lance_tx_head tx_head[TX_RING_SIZE];
144 char tx_data[TX_RING_SIZE][PKT_BUF_SZ];
464 for( i = 0; i < TX_RING_SIZE; i++ ) { in lance_init_ring()
551 for( i = 0 ; i < TX_RING_SIZE; i++ ) in lance_start_xmit()
Ddeclance.c155 #define TX_RING_SIZE (1 << (LANCE_LOG_TX_BUFFERS)) macro
156 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
222 struct lance_tx_desc btx_ring[TX_RING_SIZE];
266 char *tx_buf_ptr_cpu[TX_RING_SIZE];
270 uint tx_buf_ptr_lnc[TX_RING_SIZE];
493 for (i = 0; i < TX_RING_SIZE; i++) { in lance_init_ring()
1091 for (i = 0; i < TX_RING_SIZE; i++) { in dec_lance_probe()
1136 for (i = 0; i < TX_RING_SIZE; i++) { in dec_lance_probe()
1167 for (i = 0; i < TX_RING_SIZE; i++) { in dec_lance_probe()
Da2065.c72 #define TX_RING_SIZE (1 << LANCE_LOG_TX_BUFFERS) macro
75 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
97 struct lance_tx_desc btx_ring[TX_RING_SIZE];
100 char tx_buf[TX_RING_SIZE][TX_BUFF_SIZE];
Dsunlance.c180 #define TX_RING_SIZE (1 << (LANCE_LOG_TX_BUFFERS)) macro
181 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
227 struct lance_tx_desc btx_ring[TX_RING_SIZE];
229 u8 tx_buf [TX_RING_SIZE][TX_BUFF_SIZE];
344 for (i = 0; i < TX_RING_SIZE; i++) { in lance_init_ring_dvma()
400 for (i = 0; i < TX_RING_SIZE; i++) { in lance_init_ring_pio()
D7990.c103 for (t = 0; t < TX_RING_SIZE; t++) { \
Dpcnet32.c171 #define TX_RING_SIZE (1 << (PCNET32_LOG_TX_BUFFERS)) macro
1801 lp->tx_ring_size = TX_RING_SIZE; /* default tx ring size */ in pcnet32_probe1()
/linux-4.4.14/drivers/net/ethernet/pasemi/
Dpasemi_mac.h30 #define TX_RING_SIZE 4096 macro
31 #define CS_RING_SIZE (TX_RING_SIZE*2)
110 #define TX_DESC(tx, num) ((tx)->chan.ring_virt[(num) & (TX_RING_SIZE-1)])
111 #define TX_DESC_INFO(tx, num) ((tx)->ring_info[(num) & (TX_RING_SIZE-1)])
Dpasemi_mac.c513 ring->size = TX_RING_SIZE; in pasemi_mac_setup_tx_resources()
515 TX_RING_SIZE, GFP_KERNEL); in pasemi_mac_setup_tx_resources()
520 if (pasemi_dma_alloc_ring(&ring->chan, TX_RING_SIZE)) in pasemi_mac_setup_tx_resources()
526 val |= PAS_DMA_TXCHAN_BASEU_SIZ(TX_RING_SIZE >> 3); in pasemi_mac_setup_tx_resources()
568 limit += TX_RING_SIZE; in pasemi_mac_free_tx_resources()
571 info = &txring->ring_info[(i+1) & (TX_RING_SIZE-1)]; in pasemi_mac_free_tx_resources()
576 (TX_RING_SIZE-1)].dma; in pasemi_mac_free_tx_resources()
889 ring_limit += TX_RING_SIZE; in pasemi_mac_clean_tx()
937 txring->next_to_clean = i & (TX_RING_SIZE-1); in pasemi_mac_clean_tx()
1571 txring->next_to_fill = (fill + nfrags + 1) & (TX_RING_SIZE-1); in pasemi_mac_start_tx()
Dpasemi_mac_ethtool.c114 ering->tx_max_pending = TX_RING_SIZE/2; in pasemi_mac_ethtool_get_ringparam()
/linux-4.4.14/drivers/net/ethernet/packetengines/
Dyellowfin.c73 #define TX_RING_SIZE 16 macro
76 #define STATUS_TOTAL_SIZE TX_RING_SIZE*sizeof(struct tx_status_words)
77 #define TX_TOTAL_SIZE 2*TX_RING_SIZE*sizeof(struct yellowfin_desc)
312 struct sk_buff* tx_skbuff[TX_RING_SIZE];
701 for (i = 0; i < TX_RING_SIZE; i++) in yellowfin_tx_timeout()
760 for (i = 0; i < TX_RING_SIZE; i++) { in yellowfin_init_ring()
764 ((i+1)%TX_RING_SIZE)*sizeof(struct yellowfin_desc)); in yellowfin_init_ring()
771 for (i = 0; i < TX_RING_SIZE; i++) { in yellowfin_init_ring()
797 ((j+1)%(2*TX_RING_SIZE))*sizeof(struct yellowfin_desc)); in yellowfin_init_ring()
820 entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit()
[all …]
Dhamachi.c119 #define TX_RING_SIZE 64 macro
121 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct hamachi_desc)
487 struct sk_buff* tx_skbuff[TX_RING_SIZE];
999 int entry = hmp->dirty_tx % TX_RING_SIZE; in hamachi_tx()
1014 if (entry >= TX_RING_SIZE-1) in hamachi_tx()
1015 hmp->tx_ring[TX_RING_SIZE-1].status_n_length |= in hamachi_tx()
1064 for (i = 0; i < TX_RING_SIZE; i++) in hamachi_tx_timeout()
1087 for (i = 0; i < TX_RING_SIZE; i++){ in hamachi_tx_timeout()
1090 if (i >= TX_RING_SIZE - 1) in hamachi_tx_timeout()
1198 for (i = 0; i < TX_RING_SIZE; i++) { in hamachi_init_ring()
[all …]
/linux-4.4.14/drivers/net/ethernet/dlink/
Ddl2k.h38 #define TX_RING_SIZE 256 macro
39 #define TX_QUEUE_LEN (TX_RING_SIZE - 1) /* Limit ring entries actually used.*/
41 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct netdev_desc)
372 struct sk_buff *tx_skbuff[TX_RING_SIZE];
Dsundance.c68 #define TX_RING_SIZE 32 macro
69 #define TX_QUEUE_LEN (TX_RING_SIZE - 1) /* Limit ring entries actually used. */
72 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct netdev_desc)
374 struct sk_buff* tx_skbuff[TX_RING_SIZE];
988 for (i=0; i<TX_RING_SIZE; i++) { in tx_timeout()
1001 np->cur_tx, np->cur_tx % TX_RING_SIZE, in tx_timeout()
1002 np->dirty_tx, np->dirty_tx % TX_RING_SIZE); in tx_timeout()
1066 for (i = 0; i < TX_RING_SIZE; i++) { in init_ring()
1076 unsigned head = np->cur_task % TX_RING_SIZE; in tx_poll()
1078 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; in tx_poll()
[all …]
Ddl2k.c225 else if (tx_coalesce > TX_RING_SIZE-1) in rio_probe1()
226 tx_coalesce = TX_RING_SIZE - 1; in rio_probe1()
602 for (i = 0; i < TX_RING_SIZE; i++) { in alloc_list()
606 ((i+1)%TX_RING_SIZE) * in alloc_list()
656 entry = np->cur_tx % TX_RING_SIZE; in start_xmit()
693 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit()
694 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit()
760 int entry = np->old_tx % TX_RING_SIZE; in rio_free_tx()
785 entry = (entry + 1) % TX_RING_SIZE; in rio_free_tx()
798 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
[all …]
/linux-4.4.14/drivers/net/ethernet/dec/tulip/
Dtulip_core.c600 for (i = 0; i < TX_RING_SIZE; i++) in tulip_tx_timeout()
656 for (i = 0; i < TX_RING_SIZE; i++) { in tulip_init_ring()
677 entry = tp->cur_tx % TX_RING_SIZE; in tulip_start_xmit()
685 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ in tulip_start_xmit()
687 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { in tulip_start_xmit()
689 } else if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE - 2) { in tulip_start_xmit()
695 if (entry == TX_RING_SIZE-1) in tulip_start_xmit()
720 int entry = dirty_tx % TX_RING_SIZE; in tulip_clean_tx_ring()
819 for (i = 0; i < TX_RING_SIZE; i++) { in tulip_free_ring()
1142 if (tp->cur_tx - tp->dirty_tx > TX_RING_SIZE - 2) { in set_rx_mode()
[all …]
Dinterrupt.c526 int maxtx = TX_RING_SIZE; in tulip_interrupt()
527 int maxoi = TX_RING_SIZE; in tulip_interrupt()
590 int entry = dirty_tx % TX_RING_SIZE; in tulip_interrupt()
644 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { in tulip_interrupt()
648 dirty_tx += TX_RING_SIZE; in tulip_interrupt()
652 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) in tulip_interrupt()
Dwinbond-840.c298 dma_addr_t tx_addr[TX_RING_SIZE];
303 struct sk_buff* tx_skbuff[TX_RING_SIZE];
829 for (i = 0; i < TX_RING_SIZE; i++) { in init_rxtx_rings()
857 for (i = 0; i < TX_RING_SIZE; i++) { in free_rxtx_rings()
943 for (i = 0; i < TX_RING_SIZE; i++) in tx_timeout()
982 sizeof(struct w840_tx_desc)*TX_RING_SIZE, in alloc_ringdesc()
994 sizeof(struct w840_tx_desc)*TX_RING_SIZE, in free_ringdesc()
1008 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1023 if(entry == TX_RING_SIZE-1) in start_tx()
1065 int entry = np->dirty_tx % TX_RING_SIZE; in netdev_tx_done()
[all …]
Dtulip.h326 #define TX_RING_SIZE 32 macro
409 struct ring_info tx_buffers[TX_RING_SIZE];
/linux-4.4.14/drivers/net/wan/
Ddscc4.c164 #define TX_RING_SIZE 32 macro
166 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct TxFD)
201 struct sk_buff *tx_skbuff[TX_RING_SIZE];
431 ((dpriv->tx_current-1)%TX_RING_SIZE)*sizeof(struct TxFD); in dscc4_do_tx()
496 for (i = 0; i < TX_RING_SIZE; i++) { in dscc4_release_ring()
1153 next = dpriv->tx_current%TX_RING_SIZE; in dscc4_start_xmit()
1172 if (!((++dpriv->tx_current - dpriv->tx_dirty)%TX_RING_SIZE)) in dscc4_start_xmit()
1546 if ((dpriv->tx_current - dpriv->tx_dirty)%TX_RING_SIZE) in dscc4_tx_irq()
1572 cur = dpriv->tx_dirty%TX_RING_SIZE; in dscc4_tx_irq()
1648 (dpriv->tx_dirty%TX_RING_SIZE)* in dscc4_tx_irq()
[all …]
/linux-4.4.14/drivers/net/ethernet/3com/
D3c515.c56 #define TX_RING_SIZE 16 macro
306 struct boom_tx_desc tx_ring[TX_RING_SIZE];
309 struct sk_buff *tx_skbuff[TX_RING_SIZE];
845 for (i = 0; i < TX_RING_SIZE; i++) in corkscrew_open()
983 for (i = 0; i < TX_RING_SIZE; i++) { in corkscrew_timeout()
1013 int entry = vp->cur_tx % TX_RING_SIZE; in corkscrew_start_xmit()
1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit()
1051 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) in corkscrew_start_xmit()
1176 int entry = dirty_tx % TX_RING_SIZE; in corkscrew_interrupt()
1187 if (lp->tx_full && (lp->cur_tx - dirty_tx <= TX_RING_SIZE - 1)) { in corkscrew_interrupt()
[all …]
D3c59x.c38 #define TX_RING_SIZE 16 macro
603 struct sk_buff* tx_skbuff[TX_RING_SIZE];
1220 + sizeof(struct boom_tx_desc) * TX_RING_SIZE, in vortex_probe1()
1485 + sizeof(struct boom_tx_desc) * TX_RING_SIZE, in vortex_probe1()
1700 for (i = 0; i < TX_RING_SIZE; i++) in vortex_up()
1937 iowrite32(vp->tx_ring_dma + (vp->dirty_tx % TX_RING_SIZE) * sizeof(struct boom_tx_desc), in vortex_tx_timeout()
1939 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { in vortex_tx_timeout()
2132 int entry = vp->cur_tx % TX_RING_SIZE; in boomerang_start_xmit()
2134 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit()
2153 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { in boomerang_start_xmit()
[all …]
/linux-4.4.14/drivers/net/ethernet/freescale/
Dfec.h335 #define TX_RING_SIZE 512 /* Must be power of two */ macro
436 unsigned char *tx_bounce[TX_RING_SIZE];
437 struct sk_buff *tx_skbuff[TX_RING_SIZE];
Dfec_main.c2746 txq->tx_ring_size = TX_RING_SIZE; in fec_enet_alloc_queue()
/linux-4.4.14/drivers/net/ethernet/via/
Dvia-rhine.c79 #define TX_RING_SIZE 64 macro
80 #define TX_QUEUE_LEN (TX_RING_SIZE - 6) /* Limit ring entries actually used. */
452 struct sk_buff *tx_skbuff[TX_RING_SIZE];
453 dma_addr_t tx_skbuff_dma[TX_RING_SIZE];
456 unsigned char *tx_buf[TX_RING_SIZE];
1166 TX_RING_SIZE * sizeof(struct tx_desc), in alloc_ring()
1175 PKT_BUF_SZ * TX_RING_SIZE, in alloc_ring()
1181 TX_RING_SIZE * sizeof(struct tx_desc), in alloc_ring()
1202 TX_RING_SIZE * sizeof(struct tx_desc), in free_ring()
1207 dma_free_coherent(hwdev, PKT_BUF_SZ * TX_RING_SIZE, in free_ring()
[all …]
/linux-4.4.14/drivers/net/ethernet/adaptec/
Dstarfire.c123 #define TX_RING_SIZE 32 macro
539 struct tx_ring_info tx_info[TX_RING_SIZE];
902 …tx_ring_size = ((sizeof(starfire_tx_desc) * TX_RING_SIZE + QUEUE_ALIGN - 1) / QUEUE_ALIGN) * QUEUE… in netdev_open()
1177 for (i = 0; i < TX_RING_SIZE; i++) in init_ring()
1193 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { in start_tx()
1205 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1213 if (entry >= TX_RING_SIZE - skb_num_frags(skb)) { in start_tx()
1246 np->tx_info[entry].used_slots = TX_RING_SIZE - entry; in start_tx()
1255 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) in start_tx()
1268 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) in start_tx()
[all …]
/linux-4.4.14/drivers/net/ethernet/smsc/
Depic100.c53 #define TX_RING_SIZE 256 macro
56 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct epic_tx_desc)
256 struct sk_buff* tx_skbuff[TX_RING_SIZE];
815 (ep->dirty_tx % TX_RING_SIZE) * sizeof(struct epic_tx_desc)); in epic_restart()
935 for (i = 0; i < TX_RING_SIZE; i++) { in epic_init_ring()
961 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit()
1030 int entry = dirty_tx % TX_RING_SIZE; in epic_tx()
1052 if (cur_tx - dirty_tx > TX_RING_SIZE) { in epic_tx()
1055 dirty_tx += TX_RING_SIZE; in epic_tx()
1324 for (i = 0; i < TX_RING_SIZE; i++) { in epic_close()
Dsmsc9420.c553 for (i = 0; i < TX_RING_SIZE; i++) { in smsc9420_free_tx_ring()
973 pd->tx_ring_tail = (pd->tx_ring_tail + 1) % TX_RING_SIZE; in smsc9420_complete_tx()
985 (((pd->tx_ring_head + 2) % TX_RING_SIZE) == pd->tx_ring_tail); in smsc9420_hard_start_xmit()
1012 if (unlikely(index == (TX_RING_SIZE - 1))) in smsc9420_hard_start_xmit()
1020 pd->tx_ring_head = (pd->tx_ring_head + 1) % TX_RING_SIZE; in smsc9420_hard_start_xmit()
1242 pd->tx_buffers = kmalloc_array(TX_RING_SIZE, in smsc9420_alloc_tx_ring()
1249 for (i = 0; i < TX_RING_SIZE; i++) { in smsc9420_alloc_tx_ring()
1257 pd->tx_ring[TX_RING_SIZE - 1].length = TDES1_TER_; in smsc9420_alloc_tx_ring()
1628 sizeof(struct smsc9420_dma_desc) * TX_RING_SIZE, in smsc9420_probe()
1686 (RX_RING_SIZE + TX_RING_SIZE), pd->rx_ring, pd->rx_dma_addr); in smsc9420_probe()
[all …]
Dsmsc9420.h24 #define TX_RING_SIZE (32) macro
/linux-4.4.14/drivers/net/ethernet/qlogic/qede/
Dqede.h278 #define TX_RING_SIZE BIT(TX_RING_SIZE_POW) macro
279 #define NUM_TX_BDS_MAX (TX_RING_SIZE - 1)
/linux-4.4.14/drivers/net/ethernet/
Dfealnx.c58 #define TX_RING_SIZE 6 macro
60 #define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct fealnx_desc)
1217 for (i = 0; i < TX_RING_SIZE; i++) in fealnx_tx_timeout()
1283 np->free_tx_count = TX_RING_SIZE; in init_ring()
1285 for (i = 0; i < TX_RING_SIZE; i++) { in init_ring()
1388 np->free_tx_count = TX_RING_SIZE; in reset_tx_descriptors()
1390 for (i = 0; i < TX_RING_SIZE; i++) { in reset_tx_descriptors()
1406 np->tx_ring[TX_RING_SIZE - 1].next_desc = np->tx_ring_dma; in reset_tx_descriptors()
1407 np->tx_ring[TX_RING_SIZE - 1].next_desc_logical = &np->tx_ring[0]; in reset_tx_descriptors()
1925 for (i = 0; i < TX_RING_SIZE; i++) { in netdev_close()
/linux-4.4.14/drivers/net/ethernet/cadence/
Dmacb.c41 #define TX_RING_SIZE 128 /* must be power of 2 */ macro
42 #define TX_RING_BYTES (sizeof(struct macb_dma_desc) * TX_RING_SIZE)
45 #define MACB_TX_WAKEUP_THRESH (3 * TX_RING_SIZE / 4)
68 return index & (TX_RING_SIZE - 1); in macb_tx_ring_wrap()
721 TX_RING_SIZE) <= MACB_TX_WAKEUP_THRESH) in macb_tx_interrupt()
1261 if (unlikely(entry == (TX_RING_SIZE - 1))) in macb_tx_map()
1320 if (CIRC_SPACE(queue->tx_head, queue->tx_tail, TX_RING_SIZE) < count) { in macb_start_xmit()
1341 if (CIRC_SPACE(queue->tx_head, queue->tx_tail, TX_RING_SIZE) < 1) in macb_start_xmit()
1480 size = TX_RING_SIZE * sizeof(struct macb_tx_skb); in macb_alloc_consistent()
1512 for (i = 0; i < TX_RING_SIZE; i++) { in gem_init_rings()
[all …]
/linux-4.4.14/drivers/net/ethernet/natsemi/
Dnatsemi.c102 #define TX_RING_SIZE 16 macro
547 struct sk_buff *tx_skbuff[TX_RING_SIZE];
548 dma_addr_t tx_dma[TX_RING_SIZE];
1864 for (i = 0; i < TX_RING_SIZE; i++) { in dump_ring()
1916 sizeof(struct netdev_desc) * (RX_RING_SIZE+TX_RING_SIZE), in alloc_ring()
1974 for (i = 0; i < TX_RING_SIZE; i++) { in init_ring()
1978 *((i+1)%TX_RING_SIZE+RX_RING_SIZE)); in init_ring()
2010 for (i = 0; i < TX_RING_SIZE; i++) { in drain_tx()
2052 sizeof(struct netdev_desc) * (RX_RING_SIZE+TX_RING_SIZE), in free_ring()
2080 for (i=0;i<TX_RING_SIZE;i++) in reinit_ring()
[all …]
/linux-4.4.14/drivers/net/ethernet/i825xx/
D82596.c287 #define TX_RING_SIZE 64 macro
334 struct tx_cmd tx_cmds[TX_RING_SIZE];
335 struct i596_tbd tbds[TX_RING_SIZE];
372 static int max_cmd_backlog = TX_RING_SIZE-1;
1076 if (++lp->next_tx_cmd == TX_RING_SIZE) in i596_start_xmit()
Dlib82596.c271 #define TX_RING_SIZE 32 macro
310 struct tx_cmd tx_cmds[TX_RING_SIZE] __attribute__((aligned(32)));
311 struct i596_tbd tbds[TX_RING_SIZE] __attribute__((aligned(32)));
362 static int max_cmd_backlog = TX_RING_SIZE-1;
998 if (++lp->next_tx_cmd == TX_RING_SIZE) in i596_start_xmit()
/linux-4.4.14/drivers/net/ethernet/renesas/
Dsh_eth.h24 #define TX_RING_SIZE 64 /* Tx ring size */ macro
Dsh_eth.c3107 mdp->num_tx_ring = TX_RING_SIZE; in sh_eth_drv_probe()
/linux-4.4.14/drivers/net/ethernet/alteon/
Dacenic.h448 #define TX_RING_SIZE (MAX_TX_RING_ENTRIES * sizeof(struct tx_desc)) macro