Home
last modified time | relevance | path

Searched refs:tx_skbuff (Results 1 – 41 of 41) sorted by relevance

/linux-4.1.27/drivers/net/ethernet/calxeda/
Dxgmac.c371 struct sk_buff **tx_skbuff; member
755 priv->tx_skbuff = kzalloc(sizeof(struct sk_buff *) * DMA_TX_RING_SZ, in xgmac_dma_desc_rings_init()
757 if (!priv->tx_skbuff) in xgmac_dma_desc_rings_init()
789 kfree(priv->tx_skbuff); in xgmac_dma_desc_rings_init()
825 if (!priv->tx_skbuff) in xgmac_free_tx_skbufs()
829 if (priv->tx_skbuff[i] == NULL) in xgmac_free_tx_skbufs()
841 dev_kfree_skb_any(priv->tx_skbuff[i]); in xgmac_free_tx_skbufs()
842 priv->tx_skbuff[i] = NULL; in xgmac_free_tx_skbufs()
867 kfree(priv->tx_skbuff); in xgmac_free_dma_desc_rings()
868 priv->tx_skbuff = NULL; in xgmac_free_dma_desc_rings()
[all …]
/linux-4.1.27/drivers/net/hippi/
Drrunner.c603 rrpriv->tx_skbuff[i] = NULL; in rr_init1()
1065 if(rrpriv->tx_skbuff[txcon]){ in rr_interrupt()
1070 skb = rrpriv->tx_skbuff[txcon]; in rr_interrupt()
1080 rrpriv->tx_skbuff[txcon] = NULL; in rr_interrupt()
1112 struct sk_buff *skb = rrpriv->tx_skbuff[i]; in rr_raz_tx()
1122 rrpriv->tx_skbuff[i] = NULL; in rr_raz_tx()
1289 if (rrpriv->tx_skbuff[index]){ in rr_dump()
1290 len = min_t(int, 0x80, rrpriv->tx_skbuff[index]->len); in rr_dump()
1295 printk("%02x ", (unsigned char) rrpriv->tx_skbuff[index]->data[i]); in rr_dump()
1300 if (rrpriv->tx_skbuff[cons]){ in rr_dump()
[all …]
Drrunner.h809 struct sk_buff *tx_skbuff[TX_RING_ENTRIES]; member
/linux-4.1.27/drivers/net/ethernet/packetengines/
Dyellowfin.c312 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
761 yp->tx_skbuff[i] = NULL; in yellowfin_init_ring()
773 yp->tx_skbuff[i] = 0; in yellowfin_init_ring()
828 yp->tx_skbuff[entry] = NULL; in yellowfin_start_xmit()
834 yp->tx_skbuff[entry] = skb; in yellowfin_start_xmit()
925 skb = yp->tx_skbuff[entry]; in yellowfin_interrupt()
932 yp->tx_skbuff[entry] = NULL; in yellowfin_interrupt()
962 skb = yp->tx_skbuff[entry]; in yellowfin_interrupt()
990 yp->tx_skbuff[entry] = 0; in yellowfin_interrupt()
1264 if (yp->tx_skbuff[i]) in yellowfin_close()
[all …]
Dhamachi.c487 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
1005 skb = hmp->tx_skbuff[entry]; in hamachi_tx()
1011 hmp->tx_skbuff[entry] = NULL; in hamachi_tx()
1097 skb = hmp->tx_skbuff[i]; in hamachi_tx_timeout()
1102 hmp->tx_skbuff[i] = NULL; in hamachi_tx_timeout()
1199 hmp->tx_skbuff[i] = NULL; in hamachi_init_ring()
1237 hmp->tx_skbuff[entry] = skb; in hamachi_start_xmit()
1336 skb = hmp->tx_skbuff[entry]; in hamachi_interrupt()
1344 hmp->tx_skbuff[entry] = NULL; in hamachi_interrupt()
1719 skb = hmp->tx_skbuff[i]; in hamachi_close()
[all …]
/linux-4.1.27/drivers/net/ethernet/amd/
Dlance.c242 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
852 if (lp->tx_skbuff[i]) { in lance_purge_ring()
853 dev_kfree_skb_any(lp->tx_skbuff[i]); in lance_purge_ring()
854 lp->tx_skbuff[i] = NULL; in lance_purge_ring()
889 lp->tx_skbuff[i] = NULL; in lance_init_ring()
1003 lp->tx_skbuff[entry] = skb; in lance_start_xmit()
1087 if (lp->tx_skbuff[entry]) { in lance_interrupt()
1088 dev_kfree_skb_irq(lp->tx_skbuff[entry]); in lance_interrupt()
1089 lp->tx_skbuff[entry] = NULL; in lance_interrupt()
Dpcnet32.c269 struct sk_buff **tx_skbuff; member
503 kfree(lp->tx_skbuff); in pcnet32_realloc_tx_ring()
515 lp->tx_skbuff = new_skb_list; in pcnet32_realloc_tx_ring()
921 lp->tx_skbuff[x] = skb; in pcnet32_loopback_test()
998 packet = lp->tx_skbuff[x]->data; in pcnet32_loopback_test()
1300 if (lp->tx_skbuff[entry]) { in pcnet32_tx()
1303 lp->tx_skbuff[entry]-> in pcnet32_tx()
1305 dev_kfree_skb_any(lp->tx_skbuff[entry]); in pcnet32_tx()
1306 lp->tx_skbuff[entry] = NULL; in pcnet32_tx()
1988 lp->tx_skbuff = kcalloc(lp->tx_ring_size, sizeof(struct sk_buff *), in pcnet32_alloc_ring()
[all …]
Damd8111e.c242 if(lp->tx_skbuff[i]){ in amd8111e_free_skbs()
243 pci_unmap_single(lp->pci_dev,lp->tx_dma_addr[i], lp->tx_skbuff[i]->len,PCI_DMA_TODEVICE); in amd8111e_free_skbs()
244 dev_kfree_skb (lp->tx_skbuff[i]); in amd8111e_free_skbs()
245 lp->tx_skbuff[i] = NULL; in amd8111e_free_skbs()
665 if (lp->tx_skbuff[tx_index]) { in amd8111e_tx()
667 lp->tx_skbuff[tx_index]->len, in amd8111e_tx()
669 dev_kfree_skb_irq (lp->tx_skbuff[tx_index]); in amd8111e_tx()
670 lp->tx_skbuff[tx_index] = NULL; in amd8111e_tx()
1273 if (lp->tx_skbuff[tx_index]) in amd8111e_tx_queue_avail()
1298 lp->tx_skbuff[tx_index] = skb; in amd8111e_start_xmit()
Damd8111e.h755 struct sk_buff *tx_skbuff[NUM_TX_BUFFERS]; member
/linux-4.1.27/drivers/net/ethernet/dec/tulip/
Dwinbond-840.c303 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
830 np->tx_skbuff[i] = NULL; in init_rxtx_rings()
858 if (np->tx_skbuff[i]) { in free_rxtx_rings()
861 np->tx_skbuff[i]->len, in free_rxtx_rings()
863 dev_kfree_skb(np->tx_skbuff[i]); in free_rxtx_rings()
865 np->tx_skbuff[i] = NULL; in free_rxtx_rings()
1012 np->tx_skbuff[entry] = skb; in start_tx()
1089 np->stats.tx_bytes += np->tx_skbuff[entry]->len; in netdev_tx_done()
1095 np->tx_skbuff[entry]->len, in netdev_tx_done()
1097 np->tx_q_bytes -= np->tx_skbuff[entry]->len; in netdev_tx_done()
[all …]
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/
Dfs_enet-main.c242 skb = fep->tx_skbuff[dirtyidx]; in fs_enet_tx_napi()
293 fep->tx_skbuff[dirtyidx] = NULL; in fs_enet_tx_napi()
435 fep->tx_skbuff[i] = NULL; in fs_init_bds()
453 if ((skb = fep->tx_skbuff[i]) == NULL) in fs_cleanup_bds()
460 fep->tx_skbuff[i] = NULL; in fs_cleanup_bds()
585 fep->tx_skbuff[curidx] = NULL; in fs_enet_start_xmit()
605 fep->tx_skbuff[curidx] = skb; in fs_enet_start_xmit()
1011 fep->tx_skbuff = fep->rx_skbuff + fpi->rx_ring; in fs_enet_probe()
Dfs_enet.h136 struct sk_buff **tx_skbuff; member
/linux-4.1.27/drivers/net/ethernet/dlink/
Dsundance.c374 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
1067 np->tx_skbuff[i] = NULL; in init_ring()
1107 np->tx_skbuff[entry] = skb; in start_tx()
1141 np->tx_skbuff[entry] = NULL; in start_tx()
1163 skb = np->tx_skbuff[i]; in reset_tx()
1169 np->tx_skbuff[i] = NULL; in reset_tx()
1284 skb = np->tx_skbuff[entry]; in intr_handler()
1289 dev_kfree_skb_irq (np->tx_skbuff[entry]); in intr_handler()
1290 np->tx_skbuff[entry] = NULL; in intr_handler()
1303 skb = np->tx_skbuff[entry]; in intr_handler()
[all …]
Ddl2k.c559 np->tx_skbuff[i] = NULL; in alloc_list()
613 np->tx_skbuff[entry] = skb; in start_xmit()
731 skb = np->tx_skbuff[entry]; in rio_free_tx()
740 np->tx_skbuff[entry] = NULL; in rio_free_tx()
1716 skb = np->tx_skbuff[i]; in rio_close()
1721 np->tx_skbuff[i] = NULL; in rio_close()
Ddl2k.h365 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
/linux-4.1.27/drivers/net/ethernet/3com/
D3c515.c309 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
846 vp->tx_skbuff[i] = NULL; in corkscrew_open()
1028 vp->tx_skbuff[entry] = skb; in corkscrew_start_xmit()
1179 if (lp->tx_skbuff[entry]) { in corkscrew_interrupt()
1180 dev_kfree_skb_irq(lp->tx_skbuff[entry]); in corkscrew_interrupt()
1181 lp->tx_skbuff[entry] = NULL; in corkscrew_interrupt()
1459 if (vp->tx_skbuff[i]) { in corkscrew_close()
1460 dev_kfree_skb(vp->tx_skbuff[i]); in corkscrew_close()
1461 vp->tx_skbuff[i] = NULL; in corkscrew_close()
D3c59x.c603 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
1701 vp->tx_skbuff[i] = NULL; in vortex_up()
2156 vp->tx_skbuff[entry] = skb; in boomerang_start_xmit()
2444 if (vp->tx_skbuff[entry]) { in boomerang_interrupt()
2445 struct sk_buff *skb = vp->tx_skbuff[entry]; in boomerang_interrupt()
2458 vp->tx_skbuff[entry] = NULL; in boomerang_interrupt()
2772 if (vp->tx_skbuff[i]) { in vortex_close()
2773 struct sk_buff *skb = vp->tx_skbuff[i]; in vortex_close()
2786 vp->tx_skbuff[i] = NULL; in vortex_close()
/linux-4.1.27/drivers/net/ethernet/freescale/
Dgianfar.c290 tx_queue->tx_skbuff = in gfar_alloc_skb_resources()
292 sizeof(*tx_queue->tx_skbuff), in gfar_alloc_skb_resources()
294 if (!tx_queue->tx_skbuff) in gfar_alloc_skb_resources()
298 tx_queue->tx_skbuff[k] = NULL; in gfar_alloc_skb_resources()
585 priv->tx_queue[i]->tx_skbuff = NULL; in gfar_alloc_tx_queues()
1902 if (!tx_queue->tx_skbuff[i]) in free_skb_tx_queue()
1908 for (j = 0; j < skb_shinfo(tx_queue->tx_skbuff[i])->nr_frags; in free_skb_tx_queue()
1916 dev_kfree_skb_any(tx_queue->tx_skbuff[i]); in free_skb_tx_queue()
1917 tx_queue->tx_skbuff[i] = NULL; in free_skb_tx_queue()
1919 kfree(tx_queue->tx_skbuff); in free_skb_tx_queue()
[all …]
Dfec_main.c335 txq->tx_skbuff[index]); in fec_dump()
546 txq->tx_skbuff[index] = skb; in fec_enet_txq_submit_skb()
753 txq->tx_skbuff[index] = skb; in fec_enet_txq_submit_tso()
843 if (txq->tx_skbuff[i]) { in fec_enet_bd_init()
844 dev_kfree_skb_any(txq->tx_skbuff[i]); in fec_enet_bd_init()
845 txq->tx_skbuff[i] = NULL; in fec_enet_bd_init()
906 if (txq->tx_skbuff[j]) { in fec_enet_reset_skb()
907 dev_kfree_skb_any(txq->tx_skbuff[j]); in fec_enet_reset_skb()
908 txq->tx_skbuff[j] = NULL; in fec_enet_reset_skb()
1223 skb = txq->tx_skbuff[index]; in fec_enet_tx_queue()
[all …]
Ducc_geth.c1893 if (ugeth->tx_skbuff[i][j]) { in ucc_geth_free_tx()
1899 dev_kfree_skb_any(ugeth->tx_skbuff[i][j]); in ucc_geth_free_tx()
1900 ugeth->tx_skbuff[i][j] = NULL; in ucc_geth_free_tx()
1904 kfree(ugeth->tx_skbuff[i]); in ucc_geth_free_tx()
2250 ugeth->tx_skbuff[j] = kmalloc(sizeof(struct sk_buff *) * in ucc_geth_alloc_tx()
2254 if (ugeth->tx_skbuff[j] == NULL) { in ucc_geth_alloc_tx()
2261 ugeth->tx_skbuff[j][i] = NULL; in ucc_geth_alloc_tx()
3103 ugeth->tx_skbuff[txQ][ugeth->skb_curtx[txQ]] = skb; in ucc_geth_start_xmit()
3253 skb = ugeth->tx_skbuff[txQ][ugeth->skb_dirtytx[txQ]]; in ucc_geth_tx()
3261 ugeth->tx_skbuff[txQ][ugeth->skb_dirtytx[txQ]] = NULL; in ucc_geth_tx()
Dfec.h435 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
Ducc_geth.h1209 struct sk_buff **tx_skbuff[NUM_TX_QUEUES]; member
Dgianfar.h996 struct sk_buff **tx_skbuff; member
/linux-4.1.27/drivers/net/ethernet/renesas/
Dsh_eth.c1112 if (mdp->tx_skbuff) { in sh_eth_ring_free()
1114 dev_kfree_skb(mdp->tx_skbuff[i]); in sh_eth_ring_free()
1116 kfree(mdp->tx_skbuff); in sh_eth_ring_free()
1117 mdp->tx_skbuff = NULL; in sh_eth_ring_free()
1182 mdp->tx_skbuff[i] = NULL; in sh_eth_ring_format()
1222 mdp->tx_skbuff = kmalloc_array(mdp->num_tx_ring, in sh_eth_ring_init()
1223 sizeof(*mdp->tx_skbuff), GFP_KERNEL); in sh_eth_ring_init()
1224 if (!mdp->tx_skbuff) { in sh_eth_ring_init()
1424 if (mdp->tx_skbuff[entry]) { in sh_eth_txfree()
1427 dev_kfree_skb_irq(mdp->tx_skbuff[entry]); in sh_eth_txfree()
[all …]
Dsh_eth.h512 struct sk_buff **tx_skbuff; member
/linux-4.1.27/drivers/net/ethernet/smsc/
Depic100.c256 struct sk_buff* tx_skbuff[TX_RING_SIZE]; member
936 ep->tx_skbuff[i] = NULL; in epic_init_ring()
963 ep->tx_skbuff[entry] = skb; in epic_start_xmit()
1039 dev->stats.tx_bytes += ep->tx_skbuff[entry]->len; in epic_tx()
1044 skb = ep->tx_skbuff[entry]; in epic_tx()
1048 ep->tx_skbuff[entry] = NULL; in epic_tx()
1325 skb = ep->tx_skbuff[i]; in epic_close()
1326 ep->tx_skbuff[i] = NULL; in epic_close()
/linux-4.1.27/drivers/net/ethernet/aeroflex/
Dgreth.h102 struct sk_buff *tx_skbuff[GRETH_TXBD_NUM]; member
Dgreth.c190 struct sk_buff *skb = greth->tx_skbuff[greth->tx_last]; in greth_clean_rings()
493 greth->tx_skbuff[greth->tx_next] = skb; in greth_start_xmit_gbit()
522 greth->tx_skbuff[curr_tx] = NULL; in greth_start_xmit_gbit()
689 skb = greth->tx_skbuff[tx_last]; in greth_clean_tx_gbit()
703 greth->tx_skbuff[tx_last] = NULL; in greth_clean_tx_gbit()
/linux-4.1.27/drivers/net/ethernet/stmicro/stmmac/
Dstmmac.h46 struct sk_buff **tx_skbuff; member
Dstmmac_main.c1098 priv->tx_skbuff[i] = NULL; in init_dma_desc_rings()
1150 if (priv->tx_skbuff[i] != NULL) { in dma_free_tx_skbufs()
1151 dev_kfree_skb_any(priv->tx_skbuff[i]); in dma_free_tx_skbufs()
1152 priv->tx_skbuff[i] = NULL; in dma_free_tx_skbufs()
1189 priv->tx_skbuff = kmalloc_array(txsize, sizeof(struct sk_buff *), in alloc_dma_desc_resources()
1191 if (!priv->tx_skbuff) in alloc_dma_desc_resources()
1237 kfree(priv->tx_skbuff); in alloc_dma_desc_resources()
1272 kfree(priv->tx_skbuff); in free_dma_desc_resources()
1320 struct sk_buff *skb = priv->tx_skbuff[entry]; in stmmac_tx_clean()
1371 priv->tx_skbuff[entry] = NULL; in stmmac_tx_clean()
[all …]
Dring_mode.c64 priv->tx_skbuff[entry] = NULL; in stmmac_jumbo_frm()
Dchain_mode.c56 priv->tx_skbuff[entry] = NULL; in stmmac_jumbo_frm()
/linux-4.1.27/drivers/net/ethernet/sis/
Dsis900.c183 struct sk_buff *tx_skbuff[NUM_TX_DESC]; member
1134 sis_priv->tx_skbuff[i] = NULL; in sis900_init_tx_ring()
1560 struct sk_buff *skb = sis_priv->tx_skbuff[i]; in sis900_tx_timeout()
1567 sis_priv->tx_skbuff[i] = NULL; in sis900_tx_timeout()
1611 sis_priv->tx_skbuff[entry] = skb; in sis900_start_xmit()
1619 sis_priv->tx_skbuff[entry] = NULL; in sis900_start_xmit()
1929 skb = sis_priv->tx_skbuff[entry]; in sis900_finish_xmit()
1934 sis_priv->tx_skbuff[entry] = NULL; in sis900_finish_xmit()
1988 skb = sis_priv->tx_skbuff[i]; in sis900_close()
1993 sis_priv->tx_skbuff[i] = NULL; in sis900_close()
/linux-4.1.27/drivers/net/ethernet/via/
Dvia-rhine.c453 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
1290 rp->tx_skbuff[i] = NULL; in alloc_tbufs()
1313 if (rp->tx_skbuff[i]) { in free_tbufs()
1317 rp->tx_skbuff[i]->len, in free_tbufs()
1320 dev_kfree_skb(rp->tx_skbuff[i]); in free_tbufs()
1322 rp->tx_skbuff[i] = NULL; in free_tbufs()
1750 rp->tx_skbuff[entry] = skb; in rhine_start_tx()
1758 rp->tx_skbuff[entry] = NULL; in rhine_start_tx()
1880 skb = rp->tx_skbuff[entry]; in rhine_tx()
1923 rp->tx_skbuff[entry] = NULL; in rhine_tx()
/linux-4.1.27/drivers/net/wan/
Ddscc4.c201 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
495 skbuff = dpriv->tx_skbuff; in dscc4_release_ring()
1157 dpriv->tx_skbuff[next] = skb; in dscc4_start_xmit()
1577 skb = dpriv->tx_skbuff[cur]; in dscc4_tx_irq()
1586 dpriv->tx_skbuff[cur] = NULL; in dscc4_tx_irq()
1904 dpriv->tx_skbuff[last] = skb; in dscc4_init_dummy_skb()
1928 memset(dpriv->tx_skbuff, 0, sizeof(struct sk_buff *)*TX_RING_SIZE); in dscc4_init_ring()
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c416 tx_ring->tx_skbuff = devm_kcalloc(dev, tx_rsize, in init_tx_ring()
419 if (!tx_ring->tx_skbuff) in init_tx_ring()
630 dev_kfree_skb_any(txqueue->tx_skbuff[dma_desc]); in tx_free_ring_skbufs()
631 txqueue->tx_skbuff[dma_desc] = NULL; in tx_free_ring_skbufs()
751 struct sk_buff *skb = tqueue->tx_skbuff[entry]; in sxgbe_tx_queue_clean()
774 tqueue->tx_skbuff[entry] = NULL; in sxgbe_tx_queue_clean()
1333 tqueue->tx_skbuff[entry] = skb; in sxgbe_xmit()
1378 tqueue->tx_skbuff[entry] = NULL; in sxgbe_xmit()
Dsxgbe_common.h385 struct sk_buff **tx_skbuff; member
/linux-4.1.27/drivers/net/ethernet/natsemi/
Dnatsemi.c547 struct sk_buff *tx_skbuff[TX_RING_SIZE]; member
1969 np->tx_skbuff[i] = NULL; in init_ring()
2005 if (np->tx_skbuff[i]) { in drain_tx()
2007 np->tx_dma[i], np->tx_skbuff[i]->len, in drain_tx()
2009 dev_kfree_skb(np->tx_skbuff[i]); in drain_tx()
2012 np->tx_skbuff[i] = NULL; in drain_tx()
2093 np->tx_skbuff[entry] = skb; in start_tx()
2142 dev->stats.tx_bytes += np->tx_skbuff[entry]->len; in netdev_tx_done()
2157 np->tx_skbuff[entry]->len, in netdev_tx_done()
2160 dev_kfree_skb_irq(np->tx_skbuff[entry]); in netdev_tx_done()
[all …]
/linux-4.1.27/drivers/net/ethernet/alteon/
Dacenic.h616 struct tx_ring_info tx_skbuff[MAX_TX_RING_ENTRIES]; member
Dacenic.c2047 info = ap->skb->tx_skbuff + idx; in ace_tx_int()
2328 info = ap->skb->tx_skbuff + i; in ace_close()
2378 info = ap->skb->tx_skbuff + idx; in ace_map_tx_skb()
2467 info = ap->skb->tx_skbuff + idx; in ace_start_xmit()
/linux-4.1.27/drivers/net/ethernet/nvidia/
Dforcedeth.c4580 u8 *rxtx_ring, *rx_skbuff, *tx_skbuff; in nv_set_ringparam() local
4607 tx_skbuff = kmalloc(sizeof(struct nv_skb_map) * ring->tx_pending, GFP_KERNEL); in nv_set_ringparam()
4608 if (!rxtx_ring || !rx_skbuff || !tx_skbuff) { in nv_set_ringparam()
4621 kfree(tx_skbuff); in nv_set_ringparam()
4652 np->tx_skb = (struct nv_skb_map *)tx_skbuff; in nv_set_ringparam()