Home
last modified time | relevance | path

Searched refs:skbs (Results 1 – 32 of 32) sorted by relevance

/linux-4.4.14/drivers/isdn/hysdn/
Dhysdn_net.c45 struct sk_buff *skbs[MAX_SKB_BUFFERS]; /* pointers to tx-skbs */ member
92 dev_kfree_skb(nl->skbs[nl->out_idx++]); /* free skb */ in flush_tx_buffers()
126 lp->skbs[lp->in_idx++] = skb; /* add to buffer list */ in net_send_packet()
170 lp->dev->stats.tx_bytes += lp->skbs[lp->out_idx]->len; in hysdn_tx_netack()
172 dev_kfree_skb(lp->skbs[lp->out_idx++]); /* free skb */ in hysdn_tx_netack()
228 return (lp->skbs[lp->out_idx]); /* next packet to send */ in hysdn_tx_netget()
Dhycapi.c126 cinfo->skbs[cinfo->in_idx++] = skb; /* add to buffer list */ in hycapi_sendmsg_internal()
643 kfree_skb(cinfo->skbs[cinfo->out_idx]); /* free skb */ in hycapi_tx_capiack()
644 cinfo->skbs[cinfo->out_idx++] = NULL; in hycapi_tx_capiack()
670 return (cinfo->skbs[cinfo->out_idx]); /* next packet to send */ in hycapi_tx_capiget()
Dhysdn_defs.h205 struct sk_buff *skbs[HYSDN_MAX_CAPI_SKB]; member
/linux-4.4.14/net/mac80211/
Dtx.c880 __skb_queue_tail(&tx->skbs, tmp); in ieee80211_fragment()
922 __skb_queue_tail(&tx->skbs, skb); in ieee80211_tx_h_fragment()
959 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_fragment()
965 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_fragment()
995 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_stats()
1048 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_calculate_duration()
1052 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_calculate_duration()
1053 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); in ieee80211_tx_h_calculate_duration()
1157 __skb_queue_head_init(&tx->skbs); in ieee80211_tx_prepare()
1324 struct sk_buff_head *skbs, in ieee80211_tx_frags() argument
[all …]
Dwpa.c251 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_tkip_encrypt()
482 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_ccmp_encrypt()
709 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_gcmp_encrypt()
934 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_encrypt()
937 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_encrypt()
978 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_256_encrypt()
981 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_256_encrypt()
1123 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_gmac_encrypt()
1126 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_gmac_encrypt()
1222 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_hw_encrypt()
Dwep.c331 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_wep_encrypt()
Dstatus.c961 struct sk_buff_head *skbs) in ieee80211_purge_tx_queue() argument
965 while ((skb = __skb_dequeue(skbs))) in ieee80211_purge_tx_queue()
Dieee80211_i.h167 struct sk_buff_head skbs; member
1636 struct sk_buff_head *skbs);
1882 struct sk_buff_head *skbs);
Dutil.c56 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_set_protected()
414 struct sk_buff_head *skbs) in ieee80211_add_pending_skbs() argument
422 while ((skb = skb_dequeue(skbs))) { in ieee80211_add_pending_skbs()
/linux-4.4.14/drivers/net/ethernet/sfc/
Dselftest.c83 struct sk_buff **skbs; member
421 state->skbs[i] = skb; in efx_begin_loopback()
476 skb = state->skbs[i]; in efx_end_loopback()
532 state->skbs = kcalloc(state->packet_count, in efx_test_loopback()
533 sizeof(state->skbs[0]), GFP_KERNEL); in efx_test_loopback()
534 if (!state->skbs) in efx_test_loopback()
555 kfree(state->skbs); in efx_test_loopback()
/linux-4.4.14/drivers/net/wireless/mediatek/mt7601u/
Ddma.c256 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local
259 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet()
268 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet()
272 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet()
273 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
/linux-4.4.14/Documentation/networking/
Dnetdev-features.txt119 ndo_start_xmit can handle skbs with frags in high memory.
123 Those features say that ndo_start_xmit can handle fragmented skbs:
124 NETIF_F_SG --- paged skbs (skb_shinfo()->frags), NETIF_F_FRAGLIST ---
125 chained skbs (skb->next/prev list).
Dpktgen.txt257 to the same CPU. This reduces cache bouncing when freeing skbs.
Dpacket_mmap.txt593 - PACKET_FANOUT_QM: schedule to socket by skbs recorded queue_mapping
/linux-4.4.14/drivers/net/wireless/iwlwifi/mvm/
Dtx.c685 struct sk_buff_head skbs; in iwl_mvm_rx_tx_cmd_single() local
689 __skb_queue_head_init(&skbs); in iwl_mvm_rx_tx_cmd_single()
694 iwl_trans_reclaim(mvm->trans, txq_id, ssn, &skbs); in iwl_mvm_rx_tx_cmd_single()
696 while (!skb_queue_empty(&skbs)) { in iwl_mvm_rx_tx_cmd_single()
697 struct sk_buff *skb = __skb_dequeue(&skbs); in iwl_mvm_rx_tx_cmd_single()
/linux-4.4.14/drivers/net/wireless/iwlwifi/dvm/
Dtx.c1145 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local
1166 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx()
1194 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs); in iwlagn_rx_reply_tx()
1200 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx()
1271 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx()
1272 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
/linux-4.4.14/drivers/net/wireless/iwlwifi/
Diwl-trans.h578 struct sk_buff_head *skbs);
931 int ssn, struct sk_buff_head *skbs) in iwl_trans_reclaim() argument
936 trans->ops->reclaim(trans, queue, ssn, skbs); in iwl_trans_reclaim()
/linux-4.4.14/drivers/net/ethernet/natsemi/
Dns83820.c415 struct sk_buff *skbs[NR_RX_DESC]; member
538 BUG_ON(NULL != dev->rx_info.skbs[next_empty]); in ns83820_add_rx_skb()
539 dev->rx_info.skbs[next_empty] = skb; in ns83820_add_rx_skb()
802 struct sk_buff *skb = dev->rx_info.skbs[i]; in ns83820_cleanup_rx()
803 dev->rx_info.skbs[i] = NULL; in ns83820_cleanup_rx()
868 skb = info->skbs[next_rx];
869 info->skbs[next_rx] = NULL;
/linux-4.4.14/drivers/net/wireless/iwlegacy/
D3945.c298 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim()
300 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim()
347 info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]); in il3945_hdl_tx()
682 if (txq->skbs) { in il3945_hw_txq_free_tfd()
683 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il3945_hw_txq_free_tfd()
688 txq->skbs[txq->q.read_ptr] = NULL; in il3945_hw_txq_free_tfd()
D4965-mac.c1782 txq->skbs[q->write_ptr] = skb; in il4965_tx_skb()
2490 skb = txq->skbs[txq->q.read_ptr]; in il4965_tx_queue_reclaim()
2501 txq->skbs[txq->q.read_ptr] = NULL; in il4965_tx_queue_reclaim()
2565 info = IEEE80211_SKB_CB(il->txq[scd_flow].skbs[agg->start_idx]); in il4965_tx_status_reply_compressed_ba()
2692 info = IEEE80211_SKB_CB(il->txq[txq_id].skbs[idx]); in il4965_tx_status_reply_tx()
2725 skb = il->txq[txq_id].skbs[idx]; in il4965_tx_status_reply_tx()
2803 skb = txq->skbs[txq->q.read_ptr]; in il4965_hdl_tx()
3965 if (txq->skbs) { in il4965_hw_txq_free_tfd()
3966 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il4965_hw_txq_free_tfd()
3971 txq->skbs[txq->q.read_ptr] = NULL; in il4965_hw_txq_free_tfd()
Dcommon.c2806 kfree(txq->skbs); in il_tx_queue_free()
2807 txq->skbs = NULL; in il_tx_queue_free()
2983 txq->skbs = kcalloc(TFD_QUEUE_SIZE_MAX, in il_tx_queue_alloc()
2986 if (!txq->skbs) { in il_tx_queue_alloc()
2991 txq->skbs = NULL; in il_tx_queue_alloc()
3005 kfree(txq->skbs); in il_tx_queue_alloc()
3006 txq->skbs = NULL; in il_tx_queue_alloc()
Dcommon.h169 struct sk_buff **skbs; member
D3945-mac.c544 txq->skbs[q->write_ptr] = skb; in il3945_tx_skb()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_offload.c480 static int rx_offload_blackhole(struct t3cdev *dev, struct sk_buff **skbs, in rx_offload_blackhole() argument
484 dev_kfree_skb_any(skbs[n]); in rx_offload_blackhole()
1019 static int process_rx(struct t3cdev *dev, struct sk_buff **skbs, int n) in process_rx() argument
1022 struct sk_buff *skb = *skbs++; in process_rx()
Dsge.c1811 struct sk_buff *skbs[], int n) in deliver_partial_bundle() argument
1815 tdev->recv(tdev, skbs, n); in deliver_partial_bundle()
1838 struct sk_buff *skb, *tmp, *skbs[RX_BUNDLE_SIZE]; in ofld_poll() local
1860 skbs[ngathered] = skb; in ofld_poll()
1863 adapter->tdev.recv(&adapter->tdev, skbs, in ofld_poll()
1874 deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); in ofld_poll()
/linux-4.4.14/net/bluetooth/
Dl2cap_core.c64 struct sk_buff_head *skbs, u8 event);
1829 struct sk_buff_head *skbs) in l2cap_streaming_send() argument
1834 BT_DBG("chan %p, skbs %p", chan, skbs); in l2cap_streaming_send()
1839 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send()
2659 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_xmit() argument
2661 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_xmit()
2667 chan->tx_send_head = skb_peek(skbs); in l2cap_tx_state_xmit()
2669 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_tx_state_xmit()
2731 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_wait_f() argument
2733 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_wait_f()
[all …]
/linux-4.4.14/drivers/net/wireless/iwlwifi/pcie/
Dinternal.h436 struct sk_buff_head *skbs);
Dtx.c965 struct sk_buff_head *skbs) in iwl_trans_pcie_reclaim() argument
1003 if (WARN_ON(!skb_queue_empty(skbs))) in iwl_trans_pcie_reclaim()
1013 __skb_queue_tail(skbs, txq->entries[txq->q.read_ptr].skb); in iwl_trans_pcie_reclaim()
/linux-4.4.14/drivers/net/ethernet/sun/
Dcassini.c1857 struct sk_buff **skbs; in cas_tx_ringN() local
1863 skbs = cp->tx_skbs[ring]; in cas_tx_ringN()
1868 struct sk_buff *skb = skbs[entry]; in cas_tx_ringN()
1888 skbs[entry] = NULL; in cas_tx_ringN()
3896 struct sk_buff *skb, **skbs = cp->tx_skbs[ring]; in cas_clean_txd() local
3904 if (skbs[i] == NULL) in cas_clean_txd()
3907 skb = skbs[i]; in cas_clean_txd()
3908 skbs[i] = NULL; in cas_clean_txd()
/linux-4.4.14/drivers/net/ethernet/pasemi/
Dpasemi_mac.c872 struct sk_buff *skbs[TX_CLEAN_BATCHSIZE]; in pasemi_mac_clean_tx() local
929 skbs[descr_count] = skb; in pasemi_mac_clean_tx()
943 pasemi_mac_unmap_tx_skb(mac, nf[i], skbs[i], dmas[i]); in pasemi_mac_clean_tx()
/linux-4.4.14/drivers/net/
DKconfig292 This option enables a monitoring net device for netlink skbs. The
/linux-4.4.14/Documentation/RCU/
DRTFP.txt1665 Using RCU as a pure delay leads to a 2.5x slowdown in skbs in