skbuff             42 arch/um/drivers/slip_kern.c static unsigned short slip_protocol(struct sk_buff *skbuff)
skbuff             47 arch/um/drivers/slirp_kern.c static unsigned short slirp_protocol(struct sk_buff *skbuff)
skbuff            568 drivers/net/ethernet/cavium/liquidio/lio_core.c 		     void *skbuff,
skbuff            577 drivers/net/ethernet/cavium/liquidio/lio_core.c 	struct sk_buff *skb = (struct sk_buff *)skbuff;
skbuff            696 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	skb = (struct sk_buff *)sq->skbuff[cqe_tx->sqe_ptr];
skbuff            720 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		sq->skbuff[cqe_tx->sqe_ptr] = (u64)NULL;
skbuff            514 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL);
skbuff            515 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	if (!sq->skbuff)
skbuff            593 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		skb = (struct sk_buff *)sq->skbuff[sq->head];
skbuff            621 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	kfree(sq->skbuff);
skbuff           1195 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		skb = (struct sk_buff *)sq->skbuff[sq->head];
skbuff           1352 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		sq->skbuff[qentry] = (u64)skb;
skbuff           1444 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	sq->skbuff[qentry] = (u64)skb;
skbuff           1534 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		sq->skbuff[hdr_qentry] = (u64)NULL;
skbuff           1540 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	sq->skbuff[hdr_qentry] = (u64)skb;
skbuff            272 drivers/net/ethernet/cavium/thunder/nicvf_queues.h 	u64		*skbuff;
skbuff            244 drivers/net/ethernet/fealnx.c 	struct sk_buff *skbuff;
skbuff           1069 drivers/net/ethernet/fealnx.c 		while (np->lack_rxbuf->skbuff)
skbuff           1072 drivers/net/ethernet/fealnx.c 		np->lack_rxbuf->skbuff = skb;
skbuff           1250 drivers/net/ethernet/fealnx.c 		np->rx_ring[i].skbuff = NULL;
skbuff           1267 drivers/net/ethernet/fealnx.c 		np->rx_ring[i].skbuff = skb;
skbuff           1286 drivers/net/ethernet/fealnx.c 		np->tx_ring[i].skbuff = NULL;
skbuff           1302 drivers/net/ethernet/fealnx.c 	np->cur_tx_copy->skbuff = skb;
skbuff           1331 drivers/net/ethernet/fealnx.c 		next->skbuff = skb;
skbuff           1387 drivers/net/ethernet/fealnx.c 		if (cur->skbuff) {
skbuff           1389 drivers/net/ethernet/fealnx.c 				cur->skbuff->len, PCI_DMA_TODEVICE);
skbuff           1390 drivers/net/ethernet/fealnx.c 			dev_kfree_skb_any(cur->skbuff);
skbuff           1391 drivers/net/ethernet/fealnx.c 			cur->skbuff = NULL;
skbuff           1416 drivers/net/ethernet/fealnx.c 		if (cur->skbuff)
skbuff           1533 drivers/net/ethernet/fealnx.c 				np->cur_tx->skbuff->len, PCI_DMA_TODEVICE);
skbuff           1534 drivers/net/ethernet/fealnx.c 			dev_consume_skb_irq(np->cur_tx->skbuff);
skbuff           1535 drivers/net/ethernet/fealnx.c 			np->cur_tx->skbuff = NULL;
skbuff           1613 drivers/net/ethernet/fealnx.c 	while (!(np->cur_rx->status & RXOWN) && np->cur_rx->skbuff) {
skbuff           1666 drivers/net/ethernet/fealnx.c 						if (!np->cur_rx->skbuff) {
skbuff           1707 drivers/net/ethernet/fealnx.c 					np->cur_rx->skbuff->data, pkt_len);
skbuff           1710 drivers/net/ethernet/fealnx.c 				skb_put_data(skb, np->cur_rx->skbuff->data,
skbuff           1722 drivers/net/ethernet/fealnx.c 				skb_put(skb = np->cur_rx->skbuff, pkt_len);
skbuff           1723 drivers/net/ethernet/fealnx.c 				np->cur_rx->skbuff = NULL;
skbuff           1910 drivers/net/ethernet/fealnx.c 		struct sk_buff *skb = np->rx_ring[i].skbuff;
skbuff           1917 drivers/net/ethernet/fealnx.c 			np->rx_ring[i].skbuff = NULL;
skbuff           1922 drivers/net/ethernet/fealnx.c 		struct sk_buff *skb = np->tx_ring[i].skbuff;
skbuff           1928 drivers/net/ethernet/fealnx.c 			np->tx_ring[i].skbuff = NULL;
skbuff            173 drivers/net/ethernet/ibm/ibmveth.c 	pool->skbuff = kcalloc(pool->size, sizeof(void *), GFP_KERNEL);
skbuff            175 drivers/net/ethernet/ibm/ibmveth.c 	if (!pool->skbuff) {
skbuff            238 drivers/net/ethernet/ibm/ibmveth.c 		BUG_ON(pool->skbuff[index] != NULL);
skbuff            248 drivers/net/ethernet/ibm/ibmveth.c 		pool->skbuff[index] = skb;
skbuff            279 drivers/net/ethernet/ibm/ibmveth.c 	pool->skbuff[index] = NULL;
skbuff            334 drivers/net/ethernet/ibm/ibmveth.c 	if (pool->skbuff && pool->dma_addr) {
skbuff            336 drivers/net/ethernet/ibm/ibmveth.c 			struct sk_buff *skb = pool->skbuff[i];
skbuff            343 drivers/net/ethernet/ibm/ibmveth.c 				pool->skbuff[i] = NULL;
skbuff            353 drivers/net/ethernet/ibm/ibmveth.c 	if (pool->skbuff) {
skbuff            354 drivers/net/ethernet/ibm/ibmveth.c 		kfree(pool->skbuff);
skbuff            355 drivers/net/ethernet/ibm/ibmveth.c 		pool->skbuff = NULL;
skbuff            371 drivers/net/ethernet/ibm/ibmveth.c 	skb = adapter->rx_buff_pool[pool].skbuff[index];
skbuff            375 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_buff_pool[pool].skbuff[index] = NULL;
skbuff            404 drivers/net/ethernet/ibm/ibmveth.c 	return adapter->rx_buff_pool[pool].skbuff[index];
skbuff            119 drivers/net/ethernet/ibm/ibmveth.h     struct sk_buff **skbuff;
skbuff           1456 drivers/net/wireless/ath/ath6kl/txrx.c 				struct sk_buff *skbuff = NULL;
skbuff           1485 drivers/net/wireless/ath/ath6kl/txrx.c 				while ((skbuff = skb_dequeue(&conn->psq))) {
skbuff           1487 drivers/net/wireless/ath/ath6kl/txrx.c 					ath6kl_data_tx(skbuff, vif->ndev);
skbuff           1492 drivers/net/wireless/ath/ath6kl/txrx.c 				while ((skbuff = skb_dequeue(&conn->apsdq))) {
skbuff           1494 drivers/net/wireless/ath/ath6kl/txrx.c 					ath6kl_data_tx(skbuff, vif->ndev);