skbuff 42 arch/um/drivers/slip_kern.c static unsigned short slip_protocol(struct sk_buff *skbuff) skbuff 47 arch/um/drivers/slirp_kern.c static unsigned short slirp_protocol(struct sk_buff *skbuff) skbuff 568 drivers/net/ethernet/cavium/liquidio/lio_core.c void *skbuff, skbuff 577 drivers/net/ethernet/cavium/liquidio/lio_core.c struct sk_buff *skb = (struct sk_buff *)skbuff; skbuff 696 drivers/net/ethernet/cavium/thunder/nicvf_main.c skb = (struct sk_buff *)sq->skbuff[cqe_tx->sqe_ptr]; skbuff 720 drivers/net/ethernet/cavium/thunder/nicvf_main.c sq->skbuff[cqe_tx->sqe_ptr] = (u64)NULL; skbuff 514 drivers/net/ethernet/cavium/thunder/nicvf_queues.c sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); skbuff 515 drivers/net/ethernet/cavium/thunder/nicvf_queues.c if (!sq->skbuff) skbuff 593 drivers/net/ethernet/cavium/thunder/nicvf_queues.c skb = (struct sk_buff *)sq->skbuff[sq->head]; skbuff 621 drivers/net/ethernet/cavium/thunder/nicvf_queues.c kfree(sq->skbuff); skbuff 1195 drivers/net/ethernet/cavium/thunder/nicvf_queues.c skb = (struct sk_buff *)sq->skbuff[sq->head]; skbuff 1352 drivers/net/ethernet/cavium/thunder/nicvf_queues.c sq->skbuff[qentry] = (u64)skb; skbuff 1444 drivers/net/ethernet/cavium/thunder/nicvf_queues.c sq->skbuff[qentry] = (u64)skb; skbuff 1534 drivers/net/ethernet/cavium/thunder/nicvf_queues.c sq->skbuff[hdr_qentry] = (u64)NULL; skbuff 1540 drivers/net/ethernet/cavium/thunder/nicvf_queues.c sq->skbuff[hdr_qentry] = (u64)skb; skbuff 272 drivers/net/ethernet/cavium/thunder/nicvf_queues.h u64 *skbuff; skbuff 244 drivers/net/ethernet/fealnx.c struct sk_buff *skbuff; skbuff 1069 drivers/net/ethernet/fealnx.c while (np->lack_rxbuf->skbuff) skbuff 1072 drivers/net/ethernet/fealnx.c np->lack_rxbuf->skbuff = skb; skbuff 1250 drivers/net/ethernet/fealnx.c np->rx_ring[i].skbuff = NULL; skbuff 1267 drivers/net/ethernet/fealnx.c np->rx_ring[i].skbuff = skb; skbuff 1286 drivers/net/ethernet/fealnx.c np->tx_ring[i].skbuff = NULL; skbuff 1302 drivers/net/ethernet/fealnx.c np->cur_tx_copy->skbuff = skb; skbuff 1331 drivers/net/ethernet/fealnx.c next->skbuff = skb; skbuff 1387 drivers/net/ethernet/fealnx.c if (cur->skbuff) { skbuff 1389 drivers/net/ethernet/fealnx.c cur->skbuff->len, PCI_DMA_TODEVICE); skbuff 1390 drivers/net/ethernet/fealnx.c dev_kfree_skb_any(cur->skbuff); skbuff 1391 drivers/net/ethernet/fealnx.c cur->skbuff = NULL; skbuff 1416 drivers/net/ethernet/fealnx.c if (cur->skbuff) skbuff 1533 drivers/net/ethernet/fealnx.c np->cur_tx->skbuff->len, PCI_DMA_TODEVICE); skbuff 1534 drivers/net/ethernet/fealnx.c dev_consume_skb_irq(np->cur_tx->skbuff); skbuff 1535 drivers/net/ethernet/fealnx.c np->cur_tx->skbuff = NULL; skbuff 1613 drivers/net/ethernet/fealnx.c while (!(np->cur_rx->status & RXOWN) && np->cur_rx->skbuff) { skbuff 1666 drivers/net/ethernet/fealnx.c if (!np->cur_rx->skbuff) { skbuff 1707 drivers/net/ethernet/fealnx.c np->cur_rx->skbuff->data, pkt_len); skbuff 1710 drivers/net/ethernet/fealnx.c skb_put_data(skb, np->cur_rx->skbuff->data, skbuff 1722 drivers/net/ethernet/fealnx.c skb_put(skb = np->cur_rx->skbuff, pkt_len); skbuff 1723 drivers/net/ethernet/fealnx.c np->cur_rx->skbuff = NULL; skbuff 1910 drivers/net/ethernet/fealnx.c struct sk_buff *skb = np->rx_ring[i].skbuff; skbuff 1917 drivers/net/ethernet/fealnx.c np->rx_ring[i].skbuff = NULL; skbuff 1922 drivers/net/ethernet/fealnx.c struct sk_buff *skb = np->tx_ring[i].skbuff; skbuff 1928 drivers/net/ethernet/fealnx.c np->tx_ring[i].skbuff = NULL; skbuff 173 drivers/net/ethernet/ibm/ibmveth.c pool->skbuff = kcalloc(pool->size, sizeof(void *), GFP_KERNEL); skbuff 175 drivers/net/ethernet/ibm/ibmveth.c if (!pool->skbuff) { skbuff 238 drivers/net/ethernet/ibm/ibmveth.c BUG_ON(pool->skbuff[index] != NULL); skbuff 248 drivers/net/ethernet/ibm/ibmveth.c pool->skbuff[index] = skb; skbuff 279 drivers/net/ethernet/ibm/ibmveth.c pool->skbuff[index] = NULL; skbuff 334 drivers/net/ethernet/ibm/ibmveth.c if (pool->skbuff && pool->dma_addr) { skbuff 336 drivers/net/ethernet/ibm/ibmveth.c struct sk_buff *skb = pool->skbuff[i]; skbuff 343 drivers/net/ethernet/ibm/ibmveth.c pool->skbuff[i] = NULL; skbuff 353 drivers/net/ethernet/ibm/ibmveth.c if (pool->skbuff) { skbuff 354 drivers/net/ethernet/ibm/ibmveth.c kfree(pool->skbuff); skbuff 355 drivers/net/ethernet/ibm/ibmveth.c pool->skbuff = NULL; skbuff 371 drivers/net/ethernet/ibm/ibmveth.c skb = adapter->rx_buff_pool[pool].skbuff[index]; skbuff 375 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_buff_pool[pool].skbuff[index] = NULL; skbuff 404 drivers/net/ethernet/ibm/ibmveth.c return adapter->rx_buff_pool[pool].skbuff[index]; skbuff 119 drivers/net/ethernet/ibm/ibmveth.h struct sk_buff **skbuff; skbuff 1456 drivers/net/wireless/ath/ath6kl/txrx.c struct sk_buff *skbuff = NULL; skbuff 1485 drivers/net/wireless/ath/ath6kl/txrx.c while ((skbuff = skb_dequeue(&conn->psq))) { skbuff 1487 drivers/net/wireless/ath/ath6kl/txrx.c ath6kl_data_tx(skbuff, vif->ndev); skbuff 1492 drivers/net/wireless/ath/ath6kl/txrx.c while ((skbuff = skb_dequeue(&conn->apsdq))) { skbuff 1494 drivers/net/wireless/ath/ath6kl/txrx.c ath6kl_data_tx(skbuff, vif->ndev);