Searched refs:new_skb (Results 1 - 55 of 55) sorted by relevance

/linux-4.1.27/drivers/atm/
H A Datmtcp.c194 struct sk_buff *new_skb; atmtcp_v_send() local
213 new_skb = atm_alloc_charge(out_vcc,size,GFP_ATOMIC); atmtcp_v_send()
214 if (!new_skb) { atmtcp_v_send()
220 hdr = (void *) skb_put(new_skb,sizeof(struct atmtcp_hdr)); atmtcp_v_send()
224 skb_copy_from_linear_data(skb, skb_put(new_skb, skb->len), skb->len); atmtcp_v_send()
227 out_vcc->push(out_vcc,new_skb); atmtcp_v_send()
287 struct sk_buff *new_skb; atmtcp_c_send() local
307 new_skb = atm_alloc_charge(out_vcc,skb->len,GFP_KERNEL); atmtcp_c_send()
308 if (!new_skb) { atmtcp_c_send()
312 __net_timestamp(new_skb); atmtcp_c_send()
313 skb_copy_from_linear_data(skb, skb_put(new_skb, skb->len), skb->len); atmtcp_c_send()
314 out_vcc->push(out_vcc,new_skb); atmtcp_c_send()
/linux-4.1.27/net/mac80211/
H A Dmesh_ps.c415 struct sk_buff *new_skb, *skb = skb_peek_tail(frames); mpsp_qos_null_append() local
422 new_skb = mps_qos_null_get(sta); mpsp_qos_null_append()
423 if (!new_skb) mpsp_qos_null_append()
432 new_skb->priority = 1; mpsp_qos_null_append()
433 skb_set_queue_mapping(new_skb, IEEE80211_AC_BK); mpsp_qos_null_append()
434 ieee80211_set_qos_hdr(sdata, new_skb); mpsp_qos_null_append()
436 info = IEEE80211_SKB_CB(new_skb); mpsp_qos_null_append()
440 __skb_queue_tail(frames, new_skb); mpsp_qos_null_append()
/linux-4.1.27/drivers/s390/net/
H A Dqeth_l3_main.c2947 struct sk_buff *new_skb = NULL; qeth_l3_hard_start_xmit() local
2985 new_skb = skb; qeth_l3_hard_start_xmit()
2986 if (new_skb->protocol == ETH_P_AF_IUCV) qeth_l3_hard_start_xmit()
2996 new_skb = skb_realloc_headroom(skb, sizeof(struct qeth_hdr_tso) qeth_l3_hard_start_xmit()
2998 if (!new_skb) qeth_l3_hard_start_xmit()
3004 skb_pull(new_skb, ETH_HLEN); qeth_l3_hard_start_xmit()
3007 skb_pull(new_skb, ETH_HLEN); qeth_l3_hard_start_xmit()
3010 if (ipv != 4 && skb_vlan_tag_present(new_skb)) { qeth_l3_hard_start_xmit()
3011 skb_push(new_skb, VLAN_HLEN); qeth_l3_hard_start_xmit()
3012 skb_copy_to_linear_data(new_skb, new_skb->data + 4, 4); qeth_l3_hard_start_xmit()
3013 skb_copy_to_linear_data_offset(new_skb, 4, qeth_l3_hard_start_xmit()
3014 new_skb->data + 8, 4); qeth_l3_hard_start_xmit()
3015 skb_copy_to_linear_data_offset(new_skb, 8, qeth_l3_hard_start_xmit()
3016 new_skb->data + 12, 4); qeth_l3_hard_start_xmit()
3017 tag = (u16 *)(new_skb->data + 12); qeth_l3_hard_start_xmit()
3019 *(tag + 1) = htons(skb_vlan_tag_get(new_skb)); qeth_l3_hard_start_xmit()
3029 if (qeth_l3_tso_elements(new_skb) + 1 > 16) { qeth_l3_hard_start_xmit()
3030 if (skb_linearize(new_skb)) qeth_l3_hard_start_xmit()
3038 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l3_hard_start_xmit()
3041 qeth_l3_fill_header(card, hdr, new_skb, ipv, cast_type); qeth_l3_hard_start_xmit()
3042 qeth_tso_fill_header(card, hdr, new_skb); qeth_l3_hard_start_xmit()
3046 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l3_hard_start_xmit()
3048 qeth_l3_fill_header(card, hdr, new_skb, ipv, qeth_l3_hard_start_xmit()
3051 if (new_skb->protocol == ETH_P_AF_IUCV) qeth_l3_hard_start_xmit()
3052 qeth_l3_fill_af_iucv_hdr(card, hdr, new_skb); qeth_l3_hard_start_xmit()
3054 qeth_l3_fill_header(card, hdr, new_skb, ipv, qeth_l3_hard_start_xmit()
3056 hdr->hdr.l3.length = new_skb->len - data_offset; qeth_l3_hard_start_xmit()
3061 qeth_l3_hdr_csum(card, hdr, new_skb); qeth_l3_hard_start_xmit()
3064 elems = qeth_get_elements_no(card, new_skb, elements_needed); qeth_l3_hard_start_xmit()
3071 nr_frags = skb_shinfo(new_skb)->nr_frags; qeth_l3_hard_start_xmit()
3076 len = ((unsigned long)tcp_hdr(new_skb) + qeth_l3_hard_start_xmit()
3077 tcp_hdr(new_skb)->doff * 4) - qeth_l3_hard_start_xmit()
3078 (unsigned long)new_skb->data; qeth_l3_hard_start_xmit()
3082 if (qeth_hdr_chk_and_bounce(new_skb, &hdr, len)) qeth_l3_hard_start_xmit()
3084 rc = qeth_do_send_packet(card, queue, new_skb, hdr, qeth_l3_hard_start_xmit()
3087 rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, qeth_l3_hard_start_xmit()
3093 if (new_skb != skb) qeth_l3_hard_start_xmit()
3112 if (new_skb != skb) qeth_l3_hard_start_xmit()
3113 dev_kfree_skb_any(new_skb); qeth_l3_hard_start_xmit()
3128 if ((new_skb != skb) && new_skb) qeth_l3_hard_start_xmit()
3129 dev_kfree_skb_any(new_skb); qeth_l3_hard_start_xmit()
H A Dqeth_l2_main.c718 struct sk_buff *new_skb = skb; qeth_l2_hard_start_xmit() local
752 new_skb = skb; qeth_l2_hard_start_xmit()
760 skb_reset_mac_header(new_skb); qeth_l2_hard_start_xmit()
761 qeth_l2_fill_header(card, hdr, new_skb, cast_type); qeth_l2_hard_start_xmit()
762 hdr->hdr.l2.pkt_length = new_skb->len; qeth_l2_hard_start_xmit()
764 skb_mac_header(new_skb), ETH_HLEN); qeth_l2_hard_start_xmit()
767 new_skb = skb_realloc_headroom(skb, qeth_l2_hard_start_xmit()
769 if (!new_skb) qeth_l2_hard_start_xmit()
771 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l2_hard_start_xmit()
773 skb_set_mac_header(new_skb, sizeof(struct qeth_hdr)); qeth_l2_hard_start_xmit()
774 qeth_l2_fill_header(card, hdr, new_skb, cast_type); qeth_l2_hard_start_xmit()
778 elements = qeth_get_elements_no(card, new_skb, elements_needed); qeth_l2_hard_start_xmit()
786 if (qeth_hdr_chk_and_bounce(new_skb, &hdr, qeth_l2_hard_start_xmit()
789 rc = qeth_do_send_packet(card, queue, new_skb, hdr, qeth_l2_hard_start_xmit()
792 rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, qeth_l2_hard_start_xmit()
797 if (new_skb != skb) qeth_l2_hard_start_xmit()
805 if (new_skb != skb) qeth_l2_hard_start_xmit()
806 dev_kfree_skb_any(new_skb); qeth_l2_hard_start_xmit()
821 if ((new_skb != skb) && new_skb) qeth_l2_hard_start_xmit()
822 dev_kfree_skb_any(new_skb); qeth_l2_hard_start_xmit()
H A Dctcm_fsms.c1392 struct sk_buff *new_skb; ctcmpc_chx_rx() local
1418 new_skb = __dev_alloc_skb(ch->max_bufsize, GFP_ATOMIC); ctcmpc_chx_rx()
1420 if (new_skb == NULL) { ctcmpc_chx_rx()
1430 dev_kfree_skb_any(new_skb); ctcmpc_chx_rx()
1434 memcpy(skb_put(new_skb, block_len), ctcmpc_chx_rx()
1436 skb_queue_tail(&ch->io_queue, new_skb); ctcmpc_chx_rx()
1440 memcpy(skb_put(new_skb, len), skb->data, len); ctcmpc_chx_rx()
1441 skb_queue_tail(&ch->io_queue, new_skb); ctcmpc_chx_rx()
/linux-4.1.27/drivers/net/ethernet/adi/
H A Dbfin_mac.c119 struct sk_buff *new_skb; desc_list_init() local
190 new_skb = netdev_alloc_skb(dev, PKT_BUF_SZ + NET_IP_ALIGN); desc_list_init()
191 if (!new_skb) desc_list_init()
194 skb_reserve(new_skb, NET_IP_ALIGN); desc_list_init()
198 blackfin_dcache_invalidate_range((unsigned long)new_skb->head, desc_list_init()
199 (unsigned long)new_skb->end); desc_list_init()
200 r->skb = new_skb; desc_list_init()
212 a->start_addr = (unsigned long)new_skb->data - 2; desc_list_init()
1222 struct sk_buff *skb, *new_skb; bfin_mac_rx() local
1241 new_skb = netdev_alloc_skb(dev, PKT_BUF_SZ + NET_IP_ALIGN); bfin_mac_rx()
1242 if (!new_skb) { bfin_mac_rx()
1247 skb_reserve(new_skb, NET_IP_ALIGN); bfin_mac_rx()
1251 blackfin_dcache_invalidate_range((unsigned long)new_skb->head, bfin_mac_rx()
1252 (unsigned long)new_skb->end); bfin_mac_rx()
1254 current_rx_ptr->skb = new_skb; bfin_mac_rx()
1255 current_rx_ptr->desc_a.start_addr = (unsigned long)new_skb->data - 2; bfin_mac_rx()
/linux-4.1.27/drivers/net/wireless/mwifiex/
H A Duap_txrx.c96 struct sk_buff *new_skb; mwifiex_uap_queue_bridged_pkt() local
160 new_skb = mwifiex_uap_queue_bridged_pkt()
162 if (unlikely(!new_skb)) { mwifiex_uap_queue_bridged_pkt()
164 "Tx: cannot allocate new_skb\n"); mwifiex_uap_queue_bridged_pkt()
171 skb = new_skb; mwifiex_uap_queue_bridged_pkt()
H A Dmain.c714 struct sk_buff *new_skb; mwifiex_hard_start_xmit() local
737 new_skb = mwifiex_hard_start_xmit()
739 if (unlikely(!new_skb)) { mwifiex_hard_start_xmit()
740 dev_err(priv->adapter->dev, "Tx: cannot alloca new_skb\n"); mwifiex_hard_start_xmit()
746 skb = new_skb; mwifiex_hard_start_xmit()
/linux-4.1.27/net/nfc/
H A Ddigital_dep.c180 struct sk_buff *new_skb; digital_send_dep_data_prep() local
185 new_skb = digital_skb_alloc(ddev, ddev->remote_payload_max); digital_send_dep_data_prep()
186 if (!new_skb) { digital_send_dep_data_prep()
193 skb_reserve(new_skb, ddev->tx_headroom + NFC_HEADER_SIZE + digital_send_dep_data_prep()
195 memcpy(skb_put(new_skb, ddev->remote_payload_max), skb->data, digital_send_dep_data_prep()
203 new_skb = skb; digital_send_dep_data_prep()
206 return new_skb; digital_send_dep_data_prep()
217 struct sk_buff *new_skb; digital_recv_dep_data_gather() local
232 new_skb = skb_copy_expand(ddev->chaining_skb, digital_recv_dep_data_gather()
237 if (!new_skb) { digital_recv_dep_data_gather()
243 ddev->chaining_skb = new_skb; digital_recv_dep_data_gather()
H A Dllcp_core.c1364 struct sk_buff *new_skb; nfc_llcp_recv_agf() local
1388 new_skb = nfc_alloc_recv_skb(pdu_len, GFP_KERNEL); nfc_llcp_recv_agf()
1389 if (new_skb == NULL) { nfc_llcp_recv_agf()
1394 memcpy(skb_put(new_skb, pdu_len), skb->data, pdu_len); nfc_llcp_recv_agf()
1396 nfc_llcp_rx_skb(local, new_skb); nfc_llcp_recv_agf()
1398 kfree_skb(new_skb); nfc_llcp_recv_agf()
/linux-4.1.27/drivers/net/ethernet/natsemi/
H A Dsonic.c416 struct sk_buff *new_skb; sonic_rx() local
425 new_skb = netdev_alloc_skb(dev, SONIC_RBSIZE + 2); sonic_rx()
426 if (new_skb == NULL) { sonic_rx()
432 skb_reserve(new_skb, 2); sonic_rx()
434 new_laddr = dma_map_single(lp->device, skb_put(new_skb, SONIC_RBSIZE), sonic_rx()
437 dev_kfree_skb(new_skb); sonic_rx()
455 lp->rx_skb[entry] = new_skb; sonic_rx()
/linux-4.1.27/net/irda/irlan/
H A Dirlan_eth.c170 struct sk_buff *new_skb = irlan_eth_xmit() local
177 if (new_skb == NULL) irlan_eth_xmit()
181 skb = new_skb; irlan_eth_xmit()
/linux-4.1.27/drivers/net/usb/
H A Dsierra_net.c782 struct sk_buff *new_skb; sierra_net_skb_clone() local
785 new_skb = skb_clone(skb, GFP_ATOMIC); sierra_net_skb_clone()
791 if (new_skb) { sierra_net_skb_clone()
792 skb_trim(new_skb, len); sierra_net_skb_clone()
799 return new_skb; sierra_net_skb_clone()
807 struct sk_buff *new_skb; sierra_net_rx_fixup() local
844 new_skb = sierra_net_skb_clone(dev, skb, hh.payload_len.word); sierra_net_rx_fixup()
845 if (new_skb) sierra_net_rx_fixup()
846 usbnet_skb_return(dev, new_skb); sierra_net_rx_fixup()
/linux-4.1.27/net/atm/
H A Dmpc.c688 struct sk_buff *new_skb; mpc_push() local
761 new_skb = skb_realloc_headroom(skb, eg->ctrl_info.DH_length); mpc_push()
764 if (new_skb == NULL) { mpc_push()
768 skb_push(new_skb, eg->ctrl_info.DH_length); /* add MAC header */ mpc_push()
769 skb_copy_to_linear_data(new_skb, eg->ctrl_info.DLL_header, mpc_push()
771 new_skb->protocol = eth_type_trans(new_skb, dev); mpc_push()
772 skb_reset_network_header(new_skb); mpc_push()
774 eg->latest_ip_addr = ip_hdr(new_skb)->saddr; mpc_push()
778 memset(ATM_SKB(new_skb), 0, sizeof(struct atm_skb_data)); mpc_push()
779 netif_rx(new_skb); mpc_push()
/linux-4.1.27/drivers/net/wimax/i2400m/
H A Dusb-rx.c253 struct sk_buff *new_skb; i2400mu_rx() local
263 new_skb = skb_copy_expand(rx_skb, 0, rx_size - rx_skb->len, i2400mu_rx()
265 if (new_skb == NULL) { i2400mu_rx()
274 rx_skb = new_skb; i2400mu_rx()
280 (long) skb_end_offset(new_skb)); i2400mu_rx()
/linux-4.1.27/drivers/net/ppp/
H A Dppp_generic.c1141 struct sk_buff *new_skb; pad_compress_skb() local
1147 new_skb = alloc_skb(new_skb_size, GFP_ATOMIC); pad_compress_skb()
1148 if (!new_skb) { pad_compress_skb()
1154 skb_reserve(new_skb, pad_compress_skb()
1159 new_skb->data, skb->len + 2, pad_compress_skb()
1163 skb = new_skb; pad_compress_skb()
1168 consume_skb(new_skb); pad_compress_skb()
1169 new_skb = skb; pad_compress_skb()
1182 consume_skb(new_skb); pad_compress_skb()
1183 new_skb = NULL; pad_compress_skb()
1185 return new_skb; pad_compress_skb()
1197 struct sk_buff *new_skb; ppp_send_frame() local
1235 new_skb = alloc_skb(skb->len + ppp->dev->hard_header_len - 2, ppp_send_frame()
1237 if (!new_skb) { ppp_send_frame()
1241 skb_reserve(new_skb, ppp->dev->hard_header_len - 2); ppp_send_frame()
1244 new_skb->data + 2, &cp, ppp_send_frame()
1248 consume_skb(new_skb); ppp_send_frame()
1258 skb = new_skb; ppp_send_frame()
H A Dpptp.c208 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom); pptp_xmit() local
209 if (!new_skb) { pptp_xmit()
214 skb_set_owner_w(new_skb, skb->sk); pptp_xmit()
216 skb = new_skb; pptp_xmit()
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/
H A Dfs_enet-main.c491 struct sk_buff *new_skb; tx_skb_align_workaround() local
494 new_skb = netdev_alloc_skb(dev, skb->len + 4); tx_skb_align_workaround()
495 if (!new_skb) tx_skb_align_workaround()
499 skb_align(new_skb, 4); tx_skb_align_workaround()
502 skb_copy_from_linear_data(skb, new_skb->data, skb->len); tx_skb_align_workaround()
503 skb_put(new_skb, skb->len); tx_skb_align_workaround()
508 return new_skb; tx_skb_align_workaround()
/linux-4.1.27/drivers/net/ethernet/sun/
H A Dsunbmac.c825 struct sk_buff *new_skb; bigmac_rx() local
828 new_skb = big_mac_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); bigmac_rx()
829 if (new_skb == NULL) { bigmac_rx()
837 bp->rx_skbs[elem] = new_skb; bigmac_rx()
838 skb_put(new_skb, ETH_FRAME_LEN); bigmac_rx()
839 skb_reserve(new_skb, 34); bigmac_rx()
842 new_skb->data, bigmac_rx()
H A Dsungem.c819 struct sk_buff *new_skb; gem_rx() local
821 new_skb = gem_alloc_skb(dev, RX_BUF_ALLOC_SIZE(gp), GFP_ATOMIC); gem_rx()
822 if (new_skb == NULL) { gem_rx()
829 gp->rx_skbs[entry] = new_skb; gem_rx()
830 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); gem_rx()
832 virt_to_page(new_skb->data), gem_rx()
833 offset_in_page(new_skb->data), gem_rx()
836 skb_reserve(new_skb, RX_OFFSET); gem_rx()
H A Dsunhme.c2029 struct sk_buff *new_skb; happy_meal_rx() local
2033 new_skb = happy_meal_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); happy_meal_rx()
2034 if (new_skb == NULL) { happy_meal_rx()
2038 skb_put(new_skb, (ETH_FRAME_LEN + RX_OFFSET + 4)); happy_meal_rx()
2039 mapping = dma_map_single(hp->dma_dev, new_skb->data, happy_meal_rx()
2043 dev_kfree_skb_any(new_skb); happy_meal_rx()
2049 hp->rx_skbs[elem] = new_skb; happy_meal_rx()
2053 skb_reserve(new_skb, RX_OFFSET); happy_meal_rx()
/linux-4.1.27/drivers/net/ethernet/sgi/
H A Dioc3-eth.c578 struct sk_buff *skb, *new_skb; ioc3_rx() local
600 new_skb = ioc3_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); ioc3_rx()
601 if (!new_skb) { ioc3_rx()
605 new_skb = skb; ioc3_rx()
618 skb_put(new_skb, (1664 + RX_OFFSET)); ioc3_rx()
619 rxb = (struct ioc3_erxbuf *) new_skb->data; ioc3_rx()
620 skb_reserve(new_skb, RX_OFFSET); ioc3_rx()
628 new_skb = skb; ioc3_rx()
636 ip->rx_skbs[n_entry] = new_skb; ioc3_rx()
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
H A Dtxrx.c998 struct sk_buff *new_skb; aggr_slice_amsdu() local
1019 new_skb = aggr_get_free_skb(p_aggr); aggr_slice_amsdu()
1020 if (!new_skb) { aggr_slice_amsdu()
1025 memcpy(new_skb->data, framep, frame_8023_len); aggr_slice_amsdu()
1026 skb_put(new_skb, frame_8023_len); aggr_slice_amsdu()
1027 if (ath6kl_wmi_dot3_2_dix(new_skb)) { aggr_slice_amsdu()
1029 dev_kfree_skb(new_skb); aggr_slice_amsdu()
1033 skb_queue_tail(&rxtid->q, new_skb); aggr_slice_amsdu()
/linux-4.1.27/drivers/net/ethernet/xilinx/
H A Dxilinx_emaclite.c1002 struct sk_buff *new_skb; xemaclite_send() local
1008 new_skb = orig_skb; xemaclite_send()
1011 if (xemaclite_send_data(lp, (u8 *) new_skb->data, len) != 0) { xemaclite_send()
1016 lp->deferred_skb = new_skb; xemaclite_send()
1018 skb_tx_timestamp(new_skb); xemaclite_send()
1024 skb_tx_timestamp(new_skb); xemaclite_send()
1027 dev_consume_skb_any(new_skb); xemaclite_send()
H A Dll_temac_main.c746 struct sk_buff *skb, *new_skb; ll_temac_recv() local
786 new_skb = netdev_alloc_skb_ip_align(ndev, ll_temac_recv()
788 if (!new_skb) { ll_temac_recv()
794 cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, ll_temac_recv()
798 lp->rx_skb[lp->rx_bd_ci] = new_skb; ll_temac_recv()
H A Dxilinx_axienet_main.c731 struct sk_buff *skb, *new_skb; axienet_recv() local
770 new_skb = netdev_alloc_skb_ip_align(ndev, lp->max_frm_size); axienet_recv()
771 if (!new_skb) axienet_recv()
774 cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, axienet_recv()
779 cur_p->sw_id_offset = (u32) new_skb; axienet_recv()
/linux-4.1.27/drivers/net/ethernet/3com/
H A Dtyphoon.c1633 struct sk_buff *skb, *new_skb; typhoon_rx() local
1663 (new_skb = netdev_alloc_skb(tp->dev, pkt_len + 2)) != NULL) { typhoon_rx()
1664 skb_reserve(new_skb, 2); typhoon_rx()
1668 skb_copy_to_linear_data(new_skb, skb->data, pkt_len); typhoon_rx()
1672 skb_put(new_skb, pkt_len); typhoon_rx()
1675 new_skb = skb; typhoon_rx()
1676 skb_put(new_skb, pkt_len); typhoon_rx()
1681 new_skb->protocol = eth_type_trans(new_skb, tp->dev); typhoon_rx()
1688 new_skb->ip_summed = CHECKSUM_UNNECESSARY; typhoon_rx()
1690 skb_checksum_none_assert(new_skb); typhoon_rx()
1693 __vlan_hwaccel_put_tag(new_skb, htons(ETH_P_8021Q), typhoon_rx()
1695 netif_receive_skb(new_skb); typhoon_rx()
/linux-4.1.27/drivers/net/ethernet/ti/
H A Dcpsw.c707 struct sk_buff *new_skb; cpsw_rx_handler() local
733 new_skb = skb; cpsw_rx_handler()
742 new_skb = netdev_alloc_skb_ip_align(ndev, priv->rx_packet_max); cpsw_rx_handler()
743 if (new_skb) { cpsw_rx_handler()
752 new_skb = skb; cpsw_rx_handler()
756 ret = cpdma_chan_submit(priv->rxch, new_skb, new_skb->data, cpsw_rx_handler()
757 skb_tailroom(new_skb), 0); cpsw_rx_handler()
759 dev_kfree_skb_any(new_skb); cpsw_rx_handler()
H A Dtlan.c1523 struct sk_buff *new_skb; tlan_handle_rx_eof() local
1529 new_skb = netdev_alloc_skb_ip_align(dev, tlan_handle_rx_eof()
1531 if (!new_skb) tlan_handle_rx_eof()
1545 pci_map_single(priv->pci_dev, new_skb->data, tlan_handle_rx_eof()
1548 tlan_store_skb(head_list, new_skb); tlan_handle_rx_eof()
/linux-4.1.27/net/netfilter/ipvs/
H A Dip_vs_xmit.c868 struct sk_buff *new_skb = NULL; ip_vs_prepare_tunneled_skb() local
877 new_skb = skb_realloc_headroom(skb, max_headroom); ip_vs_prepare_tunneled_skb()
878 if (!new_skb) ip_vs_prepare_tunneled_skb()
881 skb_set_owner_w(new_skb, skb->sk); ip_vs_prepare_tunneled_skb()
883 skb = new_skb; ip_vs_prepare_tunneled_skb()
/linux-4.1.27/drivers/net/ethernet/amd/
H A Damd8111e.c696 struct sk_buff *skb,*new_skb; amd8111e_rx_poll() local
754 new_skb = netdev_alloc_skb(dev, lp->rx_buff_len); amd8111e_rx_poll()
755 if (!new_skb) { amd8111e_rx_poll()
764 skb_reserve(new_skb, 2); amd8111e_rx_poll()
769 lp->rx_skbuff[rx_index] = new_skb; amd8111e_rx_poll()
771 new_skb->data, amd8111e_rx_poll()
/linux-4.1.27/drivers/net/ethernet/ibm/
H A Dibmveth.c1096 struct sk_buff *skb, *new_skb; ibmveth_poll() local
1103 new_skb = NULL; ibmveth_poll()
1105 new_skb = netdev_alloc_skb(netdev, length); ibmveth_poll()
1107 if (new_skb) { ibmveth_poll()
1108 skb_copy_to_linear_data(new_skb, ibmveth_poll()
1116 skb = new_skb; ibmveth_poll()
/linux-4.1.27/drivers/net/hippi/
H A Drrunner.c1401 struct sk_buff *new_skb; rr_start_xmit() local
1413 if (!(new_skb = dev_alloc_skb(len + 8))) { rr_start_xmit()
1418 skb_reserve(new_skb, 8); rr_start_xmit()
1419 skb_put(new_skb, len); rr_start_xmit()
1420 skb_copy_from_linear_data(skb, new_skb->data, len); rr_start_xmit()
1422 skb = new_skb; rr_start_xmit()
/linux-4.1.27/drivers/net/ethernet/rdc/
H A Dr6040.c530 struct sk_buff *skb_ptr, *new_skb; r6040_rx() local
562 new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE); r6040_rx()
563 if (!new_skb) { r6040_rx()
582 descptr->skb_ptr = new_skb; r6040_rx()
/linux-4.1.27/drivers/block/aoe/
H A Daoecmd.c63 new_skb(ulong len) new_skb() function
186 (skb = new_skb(ETH_ZLEN))) skb_pool_get()
228 f->skb = skb = new_skb(ETH_ZLEN); newtframe()
427 skb = new_skb(sizeof *h + sizeof *ch); aoecmd_cfg_pkts()
/linux-4.1.27/net/irda/irnet/
H A Dirnet_ppp.c880 struct sk_buff * new_skb; irnet_prepare_skb() local
885 new_skb = skb_realloc_headroom(skb, ap->max_header_size + needaddr); irnet_prepare_skb()
891 DABORT(new_skb == NULL, NULL, PPP_ERROR, "Could not realloc skb\n"); irnet_prepare_skb()
894 skb = new_skb; irnet_prepare_skb()
/linux-4.1.27/drivers/net/wireless/rtlwifi/
H A Dpci.c675 struct sk_buff *new_skb, u8 *entry, _rtl_pci_init_one_rxdesc()
684 if (likely(new_skb)) { _rtl_pci_init_one_rxdesc()
685 skb = new_skb; _rtl_pci_init_one_rxdesc()
796 struct sk_buff *new_skb; _rtl_pci_rx_interrupt() local
826 new_skb = dev_alloc_skb(rtlpci->rxbuffersize); _rtl_pci_rx_interrupt()
827 if (unlikely(!new_skb)) _rtl_pci_rx_interrupt()
926 skb = new_skb; _rtl_pci_rx_interrupt()
674 _rtl_pci_init_one_rxdesc(struct ieee80211_hw *hw, struct sk_buff *new_skb, u8 *entry, int rxring_idx, int desc_idx) _rtl_pci_init_one_rxdesc() argument
/linux-4.1.27/drivers/isdn/i4l/
H A Disdn_ppp.c1349 struct sk_buff *new_skb; isdn_ppp_xmit() local
1363 new_skb = alloc_skb(hl + skb->len, GFP_ATOMIC); isdn_ppp_xmit()
1364 if (new_skb) { isdn_ppp_xmit()
1368 skb_reserve(new_skb, hl); isdn_ppp_xmit()
1369 new_skb->dev = skb->dev; isdn_ppp_xmit()
1370 skb_put(new_skb, skb->len); isdn_ppp_xmit()
1373 pktlen = slhc_compress(ipts->slcomp, skb->data, skb->len, new_skb->data, isdn_ppp_xmit()
1377 if (new_skb->data != buf) isdn_ppp_xmit()
1380 skb = new_skb; isdn_ppp_xmit()
1382 dev_kfree_skb(new_skb); isdn_ppp_xmit()
/linux-4.1.27/net/ipv6/
H A Dip6_gre.c621 struct sk_buff *new_skb; ip6gre_xmit2() local
693 new_skb = skb_realloc_headroom(skb, max_headroom); ip6gre_xmit2()
696 if (!new_skb) ip6gre_xmit2()
700 skb_set_owner_w(new_skb, skb->sk); ip6gre_xmit2()
702 skb = new_skb; ip6gre_xmit2()
H A Dip6_tunnel.c1064 struct sk_buff *new_skb; ip6_tnl_xmit2() local
1066 new_skb = skb_realloc_headroom(skb, max_headroom); ip6_tnl_xmit2()
1067 if (!new_skb) ip6_tnl_xmit2()
1071 skb_set_owner_w(new_skb, skb->sk); ip6_tnl_xmit2()
1073 skb = new_skb; ip6_tnl_xmit2()
H A Dsit.c961 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom); ipip6_tunnel_xmit() local
962 if (!new_skb) { ipip6_tunnel_xmit()
969 skb_set_owner_w(new_skb, skb->sk); ipip6_tunnel_xmit()
971 skb = new_skb; ipip6_tunnel_xmit()
/linux-4.1.27/drivers/net/ethernet/realtek/
H A D8139cp.c475 struct sk_buff *skb, *new_skb; cp_rx_poll() local
510 new_skb = napi_alloc_skb(napi, buflen); cp_rx_poll()
511 if (!new_skb) { cp_rx_poll()
516 new_mapping = dma_map_single(&cp->pdev->dev, new_skb->data, buflen, cp_rx_poll()
520 kfree_skb(new_skb); cp_rx_poll()
535 cp->rx_skb[rx_tail] = new_skb; cp_rx_poll()
/linux-4.1.27/drivers/net/ethernet/cisco/enic/
H A Denic_main.c1019 struct sk_buff *new_skb; enic_rxcopybreak() local
1023 new_skb = netdev_alloc_skb_ip_align(netdev, len); enic_rxcopybreak()
1024 if (!new_skb) enic_rxcopybreak()
1028 memcpy(new_skb->data, (*skb)->data, len); enic_rxcopybreak()
1029 *skb = new_skb; enic_rxcopybreak()
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/
H A Dixgb_main.c1971 struct sk_buff *new_skb; ixgb_check_copybreak() local
1976 new_skb = napi_alloc_skb(napi, length); ixgb_check_copybreak()
1977 if (!new_skb) ixgb_check_copybreak()
1980 skb_copy_to_linear_data_offset(new_skb, -NET_IP_ALIGN, ixgb_check_copybreak()
1985 *skb = new_skb; ixgb_check_copybreak()
/linux-4.1.27/drivers/net/wireless/rtl818x/rtl8180/
H A Ddev.c258 struct sk_buff *new_skb = dev_alloc_skb(MAX_RX_SIZE); rtl8180_handle_rx() local
260 if (unlikely(!new_skb)) rtl8180_handle_rx()
264 skb_tail_pointer(new_skb), rtl8180_handle_rx()
268 kfree_skb(new_skb); rtl8180_handle_rx()
325 skb = new_skb; rtl8180_handle_rx()
/linux-4.1.27/drivers/net/ethernet/freescale/
H A Dfec_main.c1331 struct sk_buff *new_skb; fec_enet_copybreak() local
1336 new_skb = netdev_alloc_skb(ndev, length); fec_enet_copybreak()
1337 if (!new_skb) fec_enet_copybreak()
1344 memcpy(new_skb->data, (*skb)->data, length); fec_enet_copybreak()
1346 swap_buffer2(new_skb->data, (*skb)->data, length); fec_enet_copybreak()
1347 *skb = new_skb; fec_enet_copybreak()
/linux-4.1.27/drivers/net/ethernet/via/
H A Dvia-velocity.c2010 struct sk_buff *new_skb; velocity_rx_copy() local
2012 new_skb = netdev_alloc_skb_ip_align(vptr->netdev, pkt_size); velocity_rx_copy()
2013 if (new_skb) { velocity_rx_copy()
2014 new_skb->ip_summed = rx_skb[0]->ip_summed; velocity_rx_copy()
2015 skb_copy_from_linear_data(*rx_skb, new_skb->data, pkt_size); velocity_rx_copy()
2016 *rx_skb = new_skb; velocity_rx_copy()
/linux-4.1.27/drivers/net/cris/
H A Deth_v10.c1291 struct sk_buff *new_skb = dev_alloc_skb(MAX_MEDIA_DATA_SIZE + 2 * L1_CACHE_BYTES); e100_rx() local
1292 if (!new_skb) { e100_rx()
1301 myNextRxDesc->skb = new_skb; e100_rx()
/linux-4.1.27/drivers/net/ethernet/dec/tulip/
H A Duli526x.c857 struct sk_buff *new_skb = NULL; uli526x_rx_packet() local
864 (((new_skb = netdev_alloc_skb(dev, rxlen + 2)) != NULL))) { uli526x_rx_packet()
865 skb = new_skb; uli526x_rx_packet()
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/
H A Dqlge_main.c1640 struct sk_buff *new_skb = NULL; ql_process_mac_rx_skb() local
1644 /* Allocate new_skb and copy */ ql_process_mac_rx_skb()
1645 new_skb = netdev_alloc_skb(qdev->ndev, length + NET_IP_ALIGN); ql_process_mac_rx_skb()
1646 if (new_skb == NULL) { ql_process_mac_rx_skb()
1650 skb_reserve(new_skb, NET_IP_ALIGN); ql_process_mac_rx_skb()
1651 memcpy(skb_put(new_skb, length), skb->data, length); ql_process_mac_rx_skb()
1652 skb = new_skb; ql_process_mac_rx_skb()
/linux-4.1.27/drivers/net/vmxnet3/
H A Dvmxnet3_drv.c1182 struct sk_buff *skb, *new_skb = NULL; vmxnet3_rq_rx_complete() local
1229 new_skb = netdev_alloc_skb_ip_align(adapter->netdev, vmxnet3_rq_rx_complete()
1231 if (new_skb == NULL) { vmxnet3_rq_rx_complete()
1256 rbi->skb = new_skb; vmxnet3_rq_rx_complete()
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/
H A Drtl_core.c2326 struct sk_buff *new_skb; rtl8192_rx_normal() local
2333 new_skb = dev_alloc_skb(priv->rxbuffersize); rtl8192_rx_normal()
2337 if (unlikely(!new_skb)) rtl8192_rx_normal()
2381 skb = new_skb; rtl8192_rx_normal()
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c1220 struct sk_buff *new_skb; bcmgenet_put_tx_csum() local
1230 new_skb = skb_realloc_headroom(skb, sizeof(*status)); bcmgenet_put_tx_csum()
1232 if (!new_skb) { bcmgenet_put_tx_csum()
1237 skb = new_skb; bcmgenet_put_tx_csum()
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/
H A Dnetdev.c1019 struct sk_buff *new_skb = e1000_clean_rx_irq() local
1021 if (new_skb) { e1000_clean_rx_irq()
1022 skb_copy_to_linear_data_offset(new_skb, e1000_clean_rx_irq()
1030 skb = new_skb; e1000_clean_rx_irq()
/linux-4.1.27/drivers/net/ethernet/broadcom/
H A Dtg3.c7788 struct sk_buff *new_skb, *skb = *pskb; tigon3_dma_hwbug_workaround() local
7793 new_skb = skb_copy(skb, GFP_ATOMIC); tigon3_dma_hwbug_workaround()
7797 new_skb = skb_copy_expand(skb, tigon3_dma_hwbug_workaround()
7802 if (!new_skb) { tigon3_dma_hwbug_workaround()
7806 new_addr = pci_map_single(tp->pdev, new_skb->data, new_skb->len, tigon3_dma_hwbug_workaround()
7810 dev_kfree_skb_any(new_skb); tigon3_dma_hwbug_workaround()
7817 tnapi->tx_buffers[*entry].skb = new_skb; tigon3_dma_hwbug_workaround()
7822 new_skb->len, base_flags, tigon3_dma_hwbug_workaround()
7825 dev_kfree_skb_any(new_skb); tigon3_dma_hwbug_workaround()
7832 *pskb = new_skb; tigon3_dma_hwbug_workaround()

Completed in 3270 milliseconds