/linux-4.1.27/drivers/atm/ |
H A D | atmtcp.c | 194 struct sk_buff *new_skb; atmtcp_v_send() local 213 new_skb = atm_alloc_charge(out_vcc,size,GFP_ATOMIC); atmtcp_v_send() 214 if (!new_skb) { atmtcp_v_send() 220 hdr = (void *) skb_put(new_skb,sizeof(struct atmtcp_hdr)); atmtcp_v_send() 224 skb_copy_from_linear_data(skb, skb_put(new_skb, skb->len), skb->len); atmtcp_v_send() 227 out_vcc->push(out_vcc,new_skb); atmtcp_v_send() 287 struct sk_buff *new_skb; atmtcp_c_send() local 307 new_skb = atm_alloc_charge(out_vcc,skb->len,GFP_KERNEL); atmtcp_c_send() 308 if (!new_skb) { atmtcp_c_send() 312 __net_timestamp(new_skb); atmtcp_c_send() 313 skb_copy_from_linear_data(skb, skb_put(new_skb, skb->len), skb->len); atmtcp_c_send() 314 out_vcc->push(out_vcc,new_skb); atmtcp_c_send()
|
/linux-4.1.27/net/mac80211/ |
H A D | mesh_ps.c | 415 struct sk_buff *new_skb, *skb = skb_peek_tail(frames); mpsp_qos_null_append() local 422 new_skb = mps_qos_null_get(sta); mpsp_qos_null_append() 423 if (!new_skb) mpsp_qos_null_append() 432 new_skb->priority = 1; mpsp_qos_null_append() 433 skb_set_queue_mapping(new_skb, IEEE80211_AC_BK); mpsp_qos_null_append() 434 ieee80211_set_qos_hdr(sdata, new_skb); mpsp_qos_null_append() 436 info = IEEE80211_SKB_CB(new_skb); mpsp_qos_null_append() 440 __skb_queue_tail(frames, new_skb); mpsp_qos_null_append()
|
/linux-4.1.27/drivers/s390/net/ |
H A D | qeth_l3_main.c | 2947 struct sk_buff *new_skb = NULL; qeth_l3_hard_start_xmit() local 2985 new_skb = skb; qeth_l3_hard_start_xmit() 2986 if (new_skb->protocol == ETH_P_AF_IUCV) qeth_l3_hard_start_xmit() 2996 new_skb = skb_realloc_headroom(skb, sizeof(struct qeth_hdr_tso) qeth_l3_hard_start_xmit() 2998 if (!new_skb) qeth_l3_hard_start_xmit() 3004 skb_pull(new_skb, ETH_HLEN); qeth_l3_hard_start_xmit() 3007 skb_pull(new_skb, ETH_HLEN); qeth_l3_hard_start_xmit() 3010 if (ipv != 4 && skb_vlan_tag_present(new_skb)) { qeth_l3_hard_start_xmit() 3011 skb_push(new_skb, VLAN_HLEN); qeth_l3_hard_start_xmit() 3012 skb_copy_to_linear_data(new_skb, new_skb->data + 4, 4); qeth_l3_hard_start_xmit() 3013 skb_copy_to_linear_data_offset(new_skb, 4, qeth_l3_hard_start_xmit() 3014 new_skb->data + 8, 4); qeth_l3_hard_start_xmit() 3015 skb_copy_to_linear_data_offset(new_skb, 8, qeth_l3_hard_start_xmit() 3016 new_skb->data + 12, 4); qeth_l3_hard_start_xmit() 3017 tag = (u16 *)(new_skb->data + 12); qeth_l3_hard_start_xmit() 3019 *(tag + 1) = htons(skb_vlan_tag_get(new_skb)); qeth_l3_hard_start_xmit() 3029 if (qeth_l3_tso_elements(new_skb) + 1 > 16) { qeth_l3_hard_start_xmit() 3030 if (skb_linearize(new_skb)) qeth_l3_hard_start_xmit() 3038 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l3_hard_start_xmit() 3041 qeth_l3_fill_header(card, hdr, new_skb, ipv, cast_type); qeth_l3_hard_start_xmit() 3042 qeth_tso_fill_header(card, hdr, new_skb); qeth_l3_hard_start_xmit() 3046 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l3_hard_start_xmit() 3048 qeth_l3_fill_header(card, hdr, new_skb, ipv, qeth_l3_hard_start_xmit() 3051 if (new_skb->protocol == ETH_P_AF_IUCV) qeth_l3_hard_start_xmit() 3052 qeth_l3_fill_af_iucv_hdr(card, hdr, new_skb); qeth_l3_hard_start_xmit() 3054 qeth_l3_fill_header(card, hdr, new_skb, ipv, qeth_l3_hard_start_xmit() 3056 hdr->hdr.l3.length = new_skb->len - data_offset; qeth_l3_hard_start_xmit() 3061 qeth_l3_hdr_csum(card, hdr, new_skb); qeth_l3_hard_start_xmit() 3064 elems = qeth_get_elements_no(card, new_skb, elements_needed); qeth_l3_hard_start_xmit() 3071 nr_frags = skb_shinfo(new_skb)->nr_frags; qeth_l3_hard_start_xmit() 3076 len = ((unsigned long)tcp_hdr(new_skb) + qeth_l3_hard_start_xmit() 3077 tcp_hdr(new_skb)->doff * 4) - qeth_l3_hard_start_xmit() 3078 (unsigned long)new_skb->data; qeth_l3_hard_start_xmit() 3082 if (qeth_hdr_chk_and_bounce(new_skb, &hdr, len)) qeth_l3_hard_start_xmit() 3084 rc = qeth_do_send_packet(card, queue, new_skb, hdr, qeth_l3_hard_start_xmit() 3087 rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, qeth_l3_hard_start_xmit() 3093 if (new_skb != skb) qeth_l3_hard_start_xmit() 3112 if (new_skb != skb) qeth_l3_hard_start_xmit() 3113 dev_kfree_skb_any(new_skb); qeth_l3_hard_start_xmit() 3128 if ((new_skb != skb) && new_skb) qeth_l3_hard_start_xmit() 3129 dev_kfree_skb_any(new_skb); qeth_l3_hard_start_xmit()
|
H A D | qeth_l2_main.c | 718 struct sk_buff *new_skb = skb; qeth_l2_hard_start_xmit() local 752 new_skb = skb; qeth_l2_hard_start_xmit() 760 skb_reset_mac_header(new_skb); qeth_l2_hard_start_xmit() 761 qeth_l2_fill_header(card, hdr, new_skb, cast_type); qeth_l2_hard_start_xmit() 762 hdr->hdr.l2.pkt_length = new_skb->len; qeth_l2_hard_start_xmit() 764 skb_mac_header(new_skb), ETH_HLEN); qeth_l2_hard_start_xmit() 767 new_skb = skb_realloc_headroom(skb, qeth_l2_hard_start_xmit() 769 if (!new_skb) qeth_l2_hard_start_xmit() 771 hdr = (struct qeth_hdr *)skb_push(new_skb, qeth_l2_hard_start_xmit() 773 skb_set_mac_header(new_skb, sizeof(struct qeth_hdr)); qeth_l2_hard_start_xmit() 774 qeth_l2_fill_header(card, hdr, new_skb, cast_type); qeth_l2_hard_start_xmit() 778 elements = qeth_get_elements_no(card, new_skb, elements_needed); qeth_l2_hard_start_xmit() 786 if (qeth_hdr_chk_and_bounce(new_skb, &hdr, qeth_l2_hard_start_xmit() 789 rc = qeth_do_send_packet(card, queue, new_skb, hdr, qeth_l2_hard_start_xmit() 792 rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, qeth_l2_hard_start_xmit() 797 if (new_skb != skb) qeth_l2_hard_start_xmit() 805 if (new_skb != skb) qeth_l2_hard_start_xmit() 806 dev_kfree_skb_any(new_skb); qeth_l2_hard_start_xmit() 821 if ((new_skb != skb) && new_skb) qeth_l2_hard_start_xmit() 822 dev_kfree_skb_any(new_skb); qeth_l2_hard_start_xmit()
|
H A D | ctcm_fsms.c | 1392 struct sk_buff *new_skb; ctcmpc_chx_rx() local 1418 new_skb = __dev_alloc_skb(ch->max_bufsize, GFP_ATOMIC); ctcmpc_chx_rx() 1420 if (new_skb == NULL) { ctcmpc_chx_rx() 1430 dev_kfree_skb_any(new_skb); ctcmpc_chx_rx() 1434 memcpy(skb_put(new_skb, block_len), ctcmpc_chx_rx() 1436 skb_queue_tail(&ch->io_queue, new_skb); ctcmpc_chx_rx() 1440 memcpy(skb_put(new_skb, len), skb->data, len); ctcmpc_chx_rx() 1441 skb_queue_tail(&ch->io_queue, new_skb); ctcmpc_chx_rx()
|
/linux-4.1.27/drivers/net/ethernet/adi/ |
H A D | bfin_mac.c | 119 struct sk_buff *new_skb; desc_list_init() local 190 new_skb = netdev_alloc_skb(dev, PKT_BUF_SZ + NET_IP_ALIGN); desc_list_init() 191 if (!new_skb) desc_list_init() 194 skb_reserve(new_skb, NET_IP_ALIGN); desc_list_init() 198 blackfin_dcache_invalidate_range((unsigned long)new_skb->head, desc_list_init() 199 (unsigned long)new_skb->end); desc_list_init() 200 r->skb = new_skb; desc_list_init() 212 a->start_addr = (unsigned long)new_skb->data - 2; desc_list_init() 1222 struct sk_buff *skb, *new_skb; bfin_mac_rx() local 1241 new_skb = netdev_alloc_skb(dev, PKT_BUF_SZ + NET_IP_ALIGN); bfin_mac_rx() 1242 if (!new_skb) { bfin_mac_rx() 1247 skb_reserve(new_skb, NET_IP_ALIGN); bfin_mac_rx() 1251 blackfin_dcache_invalidate_range((unsigned long)new_skb->head, bfin_mac_rx() 1252 (unsigned long)new_skb->end); bfin_mac_rx() 1254 current_rx_ptr->skb = new_skb; bfin_mac_rx() 1255 current_rx_ptr->desc_a.start_addr = (unsigned long)new_skb->data - 2; bfin_mac_rx()
|
/linux-4.1.27/drivers/net/wireless/mwifiex/ |
H A D | uap_txrx.c | 96 struct sk_buff *new_skb; mwifiex_uap_queue_bridged_pkt() local 160 new_skb = mwifiex_uap_queue_bridged_pkt() 162 if (unlikely(!new_skb)) { mwifiex_uap_queue_bridged_pkt() 164 "Tx: cannot allocate new_skb\n"); mwifiex_uap_queue_bridged_pkt() 171 skb = new_skb; mwifiex_uap_queue_bridged_pkt()
|
H A D | main.c | 714 struct sk_buff *new_skb; mwifiex_hard_start_xmit() local 737 new_skb = mwifiex_hard_start_xmit() 739 if (unlikely(!new_skb)) { mwifiex_hard_start_xmit() 740 dev_err(priv->adapter->dev, "Tx: cannot alloca new_skb\n"); mwifiex_hard_start_xmit() 746 skb = new_skb; mwifiex_hard_start_xmit()
|
/linux-4.1.27/net/nfc/ |
H A D | digital_dep.c | 180 struct sk_buff *new_skb; digital_send_dep_data_prep() local 185 new_skb = digital_skb_alloc(ddev, ddev->remote_payload_max); digital_send_dep_data_prep() 186 if (!new_skb) { digital_send_dep_data_prep() 193 skb_reserve(new_skb, ddev->tx_headroom + NFC_HEADER_SIZE + digital_send_dep_data_prep() 195 memcpy(skb_put(new_skb, ddev->remote_payload_max), skb->data, digital_send_dep_data_prep() 203 new_skb = skb; digital_send_dep_data_prep() 206 return new_skb; digital_send_dep_data_prep() 217 struct sk_buff *new_skb; digital_recv_dep_data_gather() local 232 new_skb = skb_copy_expand(ddev->chaining_skb, digital_recv_dep_data_gather() 237 if (!new_skb) { digital_recv_dep_data_gather() 243 ddev->chaining_skb = new_skb; digital_recv_dep_data_gather()
|
H A D | llcp_core.c | 1364 struct sk_buff *new_skb; nfc_llcp_recv_agf() local 1388 new_skb = nfc_alloc_recv_skb(pdu_len, GFP_KERNEL); nfc_llcp_recv_agf() 1389 if (new_skb == NULL) { nfc_llcp_recv_agf() 1394 memcpy(skb_put(new_skb, pdu_len), skb->data, pdu_len); nfc_llcp_recv_agf() 1396 nfc_llcp_rx_skb(local, new_skb); nfc_llcp_recv_agf() 1398 kfree_skb(new_skb); nfc_llcp_recv_agf()
|
/linux-4.1.27/drivers/net/ethernet/natsemi/ |
H A D | sonic.c | 416 struct sk_buff *new_skb; sonic_rx() local 425 new_skb = netdev_alloc_skb(dev, SONIC_RBSIZE + 2); sonic_rx() 426 if (new_skb == NULL) { sonic_rx() 432 skb_reserve(new_skb, 2); sonic_rx() 434 new_laddr = dma_map_single(lp->device, skb_put(new_skb, SONIC_RBSIZE), sonic_rx() 437 dev_kfree_skb(new_skb); sonic_rx() 455 lp->rx_skb[entry] = new_skb; sonic_rx()
|
/linux-4.1.27/net/irda/irlan/ |
H A D | irlan_eth.c | 170 struct sk_buff *new_skb = irlan_eth_xmit() local 177 if (new_skb == NULL) irlan_eth_xmit() 181 skb = new_skb; irlan_eth_xmit()
|
/linux-4.1.27/drivers/net/usb/ |
H A D | sierra_net.c | 782 struct sk_buff *new_skb; sierra_net_skb_clone() local 785 new_skb = skb_clone(skb, GFP_ATOMIC); sierra_net_skb_clone() 791 if (new_skb) { sierra_net_skb_clone() 792 skb_trim(new_skb, len); sierra_net_skb_clone() 799 return new_skb; sierra_net_skb_clone() 807 struct sk_buff *new_skb; sierra_net_rx_fixup() local 844 new_skb = sierra_net_skb_clone(dev, skb, hh.payload_len.word); sierra_net_rx_fixup() 845 if (new_skb) sierra_net_rx_fixup() 846 usbnet_skb_return(dev, new_skb); sierra_net_rx_fixup()
|
/linux-4.1.27/net/atm/ |
H A D | mpc.c | 688 struct sk_buff *new_skb; mpc_push() local 761 new_skb = skb_realloc_headroom(skb, eg->ctrl_info.DH_length); mpc_push() 764 if (new_skb == NULL) { mpc_push() 768 skb_push(new_skb, eg->ctrl_info.DH_length); /* add MAC header */ mpc_push() 769 skb_copy_to_linear_data(new_skb, eg->ctrl_info.DLL_header, mpc_push() 771 new_skb->protocol = eth_type_trans(new_skb, dev); mpc_push() 772 skb_reset_network_header(new_skb); mpc_push() 774 eg->latest_ip_addr = ip_hdr(new_skb)->saddr; mpc_push() 778 memset(ATM_SKB(new_skb), 0, sizeof(struct atm_skb_data)); mpc_push() 779 netif_rx(new_skb); mpc_push()
|
/linux-4.1.27/drivers/net/wimax/i2400m/ |
H A D | usb-rx.c | 253 struct sk_buff *new_skb; i2400mu_rx() local 263 new_skb = skb_copy_expand(rx_skb, 0, rx_size - rx_skb->len, i2400mu_rx() 265 if (new_skb == NULL) { i2400mu_rx() 274 rx_skb = new_skb; i2400mu_rx() 280 (long) skb_end_offset(new_skb)); i2400mu_rx()
|
/linux-4.1.27/drivers/net/ppp/ |
H A D | ppp_generic.c | 1141 struct sk_buff *new_skb; pad_compress_skb() local 1147 new_skb = alloc_skb(new_skb_size, GFP_ATOMIC); pad_compress_skb() 1148 if (!new_skb) { pad_compress_skb() 1154 skb_reserve(new_skb, pad_compress_skb() 1159 new_skb->data, skb->len + 2, pad_compress_skb() 1163 skb = new_skb; pad_compress_skb() 1168 consume_skb(new_skb); pad_compress_skb() 1169 new_skb = skb; pad_compress_skb() 1182 consume_skb(new_skb); pad_compress_skb() 1183 new_skb = NULL; pad_compress_skb() 1185 return new_skb; pad_compress_skb() 1197 struct sk_buff *new_skb; ppp_send_frame() local 1235 new_skb = alloc_skb(skb->len + ppp->dev->hard_header_len - 2, ppp_send_frame() 1237 if (!new_skb) { ppp_send_frame() 1241 skb_reserve(new_skb, ppp->dev->hard_header_len - 2); ppp_send_frame() 1244 new_skb->data + 2, &cp, ppp_send_frame() 1248 consume_skb(new_skb); ppp_send_frame() 1258 skb = new_skb; ppp_send_frame()
|
H A D | pptp.c | 208 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom); pptp_xmit() local 209 if (!new_skb) { pptp_xmit() 214 skb_set_owner_w(new_skb, skb->sk); pptp_xmit() 216 skb = new_skb; pptp_xmit()
|
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/ |
H A D | fs_enet-main.c | 491 struct sk_buff *new_skb; tx_skb_align_workaround() local 494 new_skb = netdev_alloc_skb(dev, skb->len + 4); tx_skb_align_workaround() 495 if (!new_skb) tx_skb_align_workaround() 499 skb_align(new_skb, 4); tx_skb_align_workaround() 502 skb_copy_from_linear_data(skb, new_skb->data, skb->len); tx_skb_align_workaround() 503 skb_put(new_skb, skb->len); tx_skb_align_workaround() 508 return new_skb; tx_skb_align_workaround()
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
H A D | sunbmac.c | 825 struct sk_buff *new_skb; bigmac_rx() local 828 new_skb = big_mac_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); bigmac_rx() 829 if (new_skb == NULL) { bigmac_rx() 837 bp->rx_skbs[elem] = new_skb; bigmac_rx() 838 skb_put(new_skb, ETH_FRAME_LEN); bigmac_rx() 839 skb_reserve(new_skb, 34); bigmac_rx() 842 new_skb->data, bigmac_rx()
|
H A D | sungem.c | 819 struct sk_buff *new_skb; gem_rx() local 821 new_skb = gem_alloc_skb(dev, RX_BUF_ALLOC_SIZE(gp), GFP_ATOMIC); gem_rx() 822 if (new_skb == NULL) { gem_rx() 829 gp->rx_skbs[entry] = new_skb; gem_rx() 830 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); gem_rx() 832 virt_to_page(new_skb->data), gem_rx() 833 offset_in_page(new_skb->data), gem_rx() 836 skb_reserve(new_skb, RX_OFFSET); gem_rx()
|
H A D | sunhme.c | 2029 struct sk_buff *new_skb; happy_meal_rx() local 2033 new_skb = happy_meal_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); happy_meal_rx() 2034 if (new_skb == NULL) { happy_meal_rx() 2038 skb_put(new_skb, (ETH_FRAME_LEN + RX_OFFSET + 4)); happy_meal_rx() 2039 mapping = dma_map_single(hp->dma_dev, new_skb->data, happy_meal_rx() 2043 dev_kfree_skb_any(new_skb); happy_meal_rx() 2049 hp->rx_skbs[elem] = new_skb; happy_meal_rx() 2053 skb_reserve(new_skb, RX_OFFSET); happy_meal_rx()
|
/linux-4.1.27/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 578 struct sk_buff *skb, *new_skb; ioc3_rx() local 600 new_skb = ioc3_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); ioc3_rx() 601 if (!new_skb) { ioc3_rx() 605 new_skb = skb; ioc3_rx() 618 skb_put(new_skb, (1664 + RX_OFFSET)); ioc3_rx() 619 rxb = (struct ioc3_erxbuf *) new_skb->data; ioc3_rx() 620 skb_reserve(new_skb, RX_OFFSET); ioc3_rx() 628 new_skb = skb; ioc3_rx() 636 ip->rx_skbs[n_entry] = new_skb; ioc3_rx()
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
H A D | txrx.c | 998 struct sk_buff *new_skb; aggr_slice_amsdu() local 1019 new_skb = aggr_get_free_skb(p_aggr); aggr_slice_amsdu() 1020 if (!new_skb) { aggr_slice_amsdu() 1025 memcpy(new_skb->data, framep, frame_8023_len); aggr_slice_amsdu() 1026 skb_put(new_skb, frame_8023_len); aggr_slice_amsdu() 1027 if (ath6kl_wmi_dot3_2_dix(new_skb)) { aggr_slice_amsdu() 1029 dev_kfree_skb(new_skb); aggr_slice_amsdu() 1033 skb_queue_tail(&rxtid->q, new_skb); aggr_slice_amsdu()
|
/linux-4.1.27/drivers/net/ethernet/xilinx/ |
H A D | xilinx_emaclite.c | 1002 struct sk_buff *new_skb; xemaclite_send() local 1008 new_skb = orig_skb; xemaclite_send() 1011 if (xemaclite_send_data(lp, (u8 *) new_skb->data, len) != 0) { xemaclite_send() 1016 lp->deferred_skb = new_skb; xemaclite_send() 1018 skb_tx_timestamp(new_skb); xemaclite_send() 1024 skb_tx_timestamp(new_skb); xemaclite_send() 1027 dev_consume_skb_any(new_skb); xemaclite_send()
|
H A D | ll_temac_main.c | 746 struct sk_buff *skb, *new_skb; ll_temac_recv() local 786 new_skb = netdev_alloc_skb_ip_align(ndev, ll_temac_recv() 788 if (!new_skb) { ll_temac_recv() 794 cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, ll_temac_recv() 798 lp->rx_skb[lp->rx_bd_ci] = new_skb; ll_temac_recv()
|
H A D | xilinx_axienet_main.c | 731 struct sk_buff *skb, *new_skb; axienet_recv() local 770 new_skb = netdev_alloc_skb_ip_align(ndev, lp->max_frm_size); axienet_recv() 771 if (!new_skb) axienet_recv() 774 cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, axienet_recv() 779 cur_p->sw_id_offset = (u32) new_skb; axienet_recv()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
H A D | typhoon.c | 1633 struct sk_buff *skb, *new_skb; typhoon_rx() local 1663 (new_skb = netdev_alloc_skb(tp->dev, pkt_len + 2)) != NULL) { typhoon_rx() 1664 skb_reserve(new_skb, 2); typhoon_rx() 1668 skb_copy_to_linear_data(new_skb, skb->data, pkt_len); typhoon_rx() 1672 skb_put(new_skb, pkt_len); typhoon_rx() 1675 new_skb = skb; typhoon_rx() 1676 skb_put(new_skb, pkt_len); typhoon_rx() 1681 new_skb->protocol = eth_type_trans(new_skb, tp->dev); typhoon_rx() 1688 new_skb->ip_summed = CHECKSUM_UNNECESSARY; typhoon_rx() 1690 skb_checksum_none_assert(new_skb); typhoon_rx() 1693 __vlan_hwaccel_put_tag(new_skb, htons(ETH_P_8021Q), typhoon_rx() 1695 netif_receive_skb(new_skb); typhoon_rx()
|
/linux-4.1.27/drivers/net/ethernet/ti/ |
H A D | cpsw.c | 707 struct sk_buff *new_skb; cpsw_rx_handler() local 733 new_skb = skb; cpsw_rx_handler() 742 new_skb = netdev_alloc_skb_ip_align(ndev, priv->rx_packet_max); cpsw_rx_handler() 743 if (new_skb) { cpsw_rx_handler() 752 new_skb = skb; cpsw_rx_handler() 756 ret = cpdma_chan_submit(priv->rxch, new_skb, new_skb->data, cpsw_rx_handler() 757 skb_tailroom(new_skb), 0); cpsw_rx_handler() 759 dev_kfree_skb_any(new_skb); cpsw_rx_handler()
|
H A D | tlan.c | 1523 struct sk_buff *new_skb; tlan_handle_rx_eof() local 1529 new_skb = netdev_alloc_skb_ip_align(dev, tlan_handle_rx_eof() 1531 if (!new_skb) tlan_handle_rx_eof() 1545 pci_map_single(priv->pci_dev, new_skb->data, tlan_handle_rx_eof() 1548 tlan_store_skb(head_list, new_skb); tlan_handle_rx_eof()
|
/linux-4.1.27/net/netfilter/ipvs/ |
H A D | ip_vs_xmit.c | 868 struct sk_buff *new_skb = NULL; ip_vs_prepare_tunneled_skb() local 877 new_skb = skb_realloc_headroom(skb, max_headroom); ip_vs_prepare_tunneled_skb() 878 if (!new_skb) ip_vs_prepare_tunneled_skb() 881 skb_set_owner_w(new_skb, skb->sk); ip_vs_prepare_tunneled_skb() 883 skb = new_skb; ip_vs_prepare_tunneled_skb()
|
/linux-4.1.27/drivers/net/ethernet/amd/ |
H A D | amd8111e.c | 696 struct sk_buff *skb,*new_skb; amd8111e_rx_poll() local 754 new_skb = netdev_alloc_skb(dev, lp->rx_buff_len); amd8111e_rx_poll() 755 if (!new_skb) { amd8111e_rx_poll() 764 skb_reserve(new_skb, 2); amd8111e_rx_poll() 769 lp->rx_skbuff[rx_index] = new_skb; amd8111e_rx_poll() 771 new_skb->data, amd8111e_rx_poll()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ |
H A D | ibmveth.c | 1096 struct sk_buff *skb, *new_skb; ibmveth_poll() local 1103 new_skb = NULL; ibmveth_poll() 1105 new_skb = netdev_alloc_skb(netdev, length); ibmveth_poll() 1107 if (new_skb) { ibmveth_poll() 1108 skb_copy_to_linear_data(new_skb, ibmveth_poll() 1116 skb = new_skb; ibmveth_poll()
|
/linux-4.1.27/drivers/net/hippi/ |
H A D | rrunner.c | 1401 struct sk_buff *new_skb; rr_start_xmit() local 1413 if (!(new_skb = dev_alloc_skb(len + 8))) { rr_start_xmit() 1418 skb_reserve(new_skb, 8); rr_start_xmit() 1419 skb_put(new_skb, len); rr_start_xmit() 1420 skb_copy_from_linear_data(skb, new_skb->data, len); rr_start_xmit() 1422 skb = new_skb; rr_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/rdc/ |
H A D | r6040.c | 530 struct sk_buff *skb_ptr, *new_skb; r6040_rx() local 562 new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE); r6040_rx() 563 if (!new_skb) { r6040_rx() 582 descptr->skb_ptr = new_skb; r6040_rx()
|
/linux-4.1.27/drivers/block/aoe/ |
H A D | aoecmd.c | 63 new_skb(ulong len) new_skb() function 186 (skb = new_skb(ETH_ZLEN))) skb_pool_get() 228 f->skb = skb = new_skb(ETH_ZLEN); newtframe() 427 skb = new_skb(sizeof *h + sizeof *ch); aoecmd_cfg_pkts()
|
/linux-4.1.27/net/irda/irnet/ |
H A D | irnet_ppp.c | 880 struct sk_buff * new_skb; irnet_prepare_skb() local 885 new_skb = skb_realloc_headroom(skb, ap->max_header_size + needaddr); irnet_prepare_skb() 891 DABORT(new_skb == NULL, NULL, PPP_ERROR, "Could not realloc skb\n"); irnet_prepare_skb() 894 skb = new_skb; irnet_prepare_skb()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/ |
H A D | pci.c | 675 struct sk_buff *new_skb, u8 *entry, _rtl_pci_init_one_rxdesc() 684 if (likely(new_skb)) { _rtl_pci_init_one_rxdesc() 685 skb = new_skb; _rtl_pci_init_one_rxdesc() 796 struct sk_buff *new_skb; _rtl_pci_rx_interrupt() local 826 new_skb = dev_alloc_skb(rtlpci->rxbuffersize); _rtl_pci_rx_interrupt() 827 if (unlikely(!new_skb)) _rtl_pci_rx_interrupt() 926 skb = new_skb; _rtl_pci_rx_interrupt() 674 _rtl_pci_init_one_rxdesc(struct ieee80211_hw *hw, struct sk_buff *new_skb, u8 *entry, int rxring_idx, int desc_idx) _rtl_pci_init_one_rxdesc() argument
|
/linux-4.1.27/drivers/isdn/i4l/ |
H A D | isdn_ppp.c | 1349 struct sk_buff *new_skb; isdn_ppp_xmit() local 1363 new_skb = alloc_skb(hl + skb->len, GFP_ATOMIC); isdn_ppp_xmit() 1364 if (new_skb) { isdn_ppp_xmit() 1368 skb_reserve(new_skb, hl); isdn_ppp_xmit() 1369 new_skb->dev = skb->dev; isdn_ppp_xmit() 1370 skb_put(new_skb, skb->len); isdn_ppp_xmit() 1373 pktlen = slhc_compress(ipts->slcomp, skb->data, skb->len, new_skb->data, isdn_ppp_xmit() 1377 if (new_skb->data != buf) isdn_ppp_xmit() 1380 skb = new_skb; isdn_ppp_xmit() 1382 dev_kfree_skb(new_skb); isdn_ppp_xmit()
|
/linux-4.1.27/net/ipv6/ |
H A D | ip6_gre.c | 621 struct sk_buff *new_skb; ip6gre_xmit2() local 693 new_skb = skb_realloc_headroom(skb, max_headroom); ip6gre_xmit2() 696 if (!new_skb) ip6gre_xmit2() 700 skb_set_owner_w(new_skb, skb->sk); ip6gre_xmit2() 702 skb = new_skb; ip6gre_xmit2()
|
H A D | ip6_tunnel.c | 1064 struct sk_buff *new_skb; ip6_tnl_xmit2() local 1066 new_skb = skb_realloc_headroom(skb, max_headroom); ip6_tnl_xmit2() 1067 if (!new_skb) ip6_tnl_xmit2() 1071 skb_set_owner_w(new_skb, skb->sk); ip6_tnl_xmit2() 1073 skb = new_skb; ip6_tnl_xmit2()
|
H A D | sit.c | 961 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom); ipip6_tunnel_xmit() local 962 if (!new_skb) { ipip6_tunnel_xmit() 969 skb_set_owner_w(new_skb, skb->sk); ipip6_tunnel_xmit() 971 skb = new_skb; ipip6_tunnel_xmit()
|
/linux-4.1.27/drivers/net/ethernet/realtek/ |
H A D | 8139cp.c | 475 struct sk_buff *skb, *new_skb; cp_rx_poll() local 510 new_skb = napi_alloc_skb(napi, buflen); cp_rx_poll() 511 if (!new_skb) { cp_rx_poll() 516 new_mapping = dma_map_single(&cp->pdev->dev, new_skb->data, buflen, cp_rx_poll() 520 kfree_skb(new_skb); cp_rx_poll() 535 cp->rx_skb[rx_tail] = new_skb; cp_rx_poll()
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
H A D | enic_main.c | 1019 struct sk_buff *new_skb; enic_rxcopybreak() local 1023 new_skb = netdev_alloc_skb_ip_align(netdev, len); enic_rxcopybreak() 1024 if (!new_skb) enic_rxcopybreak() 1028 memcpy(new_skb->data, (*skb)->data, len); enic_rxcopybreak() 1029 *skb = new_skb; enic_rxcopybreak()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/ |
H A D | ixgb_main.c | 1971 struct sk_buff *new_skb; ixgb_check_copybreak() local 1976 new_skb = napi_alloc_skb(napi, length); ixgb_check_copybreak() 1977 if (!new_skb) ixgb_check_copybreak() 1980 skb_copy_to_linear_data_offset(new_skb, -NET_IP_ALIGN, ixgb_check_copybreak() 1985 *skb = new_skb; ixgb_check_copybreak()
|
/linux-4.1.27/drivers/net/wireless/rtl818x/rtl8180/ |
H A D | dev.c | 258 struct sk_buff *new_skb = dev_alloc_skb(MAX_RX_SIZE); rtl8180_handle_rx() local 260 if (unlikely(!new_skb)) rtl8180_handle_rx() 264 skb_tail_pointer(new_skb), rtl8180_handle_rx() 268 kfree_skb(new_skb); rtl8180_handle_rx() 325 skb = new_skb; rtl8180_handle_rx()
|
/linux-4.1.27/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 1331 struct sk_buff *new_skb; fec_enet_copybreak() local 1336 new_skb = netdev_alloc_skb(ndev, length); fec_enet_copybreak() 1337 if (!new_skb) fec_enet_copybreak() 1344 memcpy(new_skb->data, (*skb)->data, length); fec_enet_copybreak() 1346 swap_buffer2(new_skb->data, (*skb)->data, length); fec_enet_copybreak() 1347 *skb = new_skb; fec_enet_copybreak()
|
/linux-4.1.27/drivers/net/ethernet/via/ |
H A D | via-velocity.c | 2010 struct sk_buff *new_skb; velocity_rx_copy() local 2012 new_skb = netdev_alloc_skb_ip_align(vptr->netdev, pkt_size); velocity_rx_copy() 2013 if (new_skb) { velocity_rx_copy() 2014 new_skb->ip_summed = rx_skb[0]->ip_summed; velocity_rx_copy() 2015 skb_copy_from_linear_data(*rx_skb, new_skb->data, pkt_size); velocity_rx_copy() 2016 *rx_skb = new_skb; velocity_rx_copy()
|
/linux-4.1.27/drivers/net/cris/ |
H A D | eth_v10.c | 1291 struct sk_buff *new_skb = dev_alloc_skb(MAX_MEDIA_DATA_SIZE + 2 * L1_CACHE_BYTES); e100_rx() local 1292 if (!new_skb) { e100_rx() 1301 myNextRxDesc->skb = new_skb; e100_rx()
|
/linux-4.1.27/drivers/net/ethernet/dec/tulip/ |
H A D | uli526x.c | 857 struct sk_buff *new_skb = NULL; uli526x_rx_packet() local 864 (((new_skb = netdev_alloc_skb(dev, rxlen + 2)) != NULL))) { uli526x_rx_packet() 865 skb = new_skb; uli526x_rx_packet()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/ |
H A D | qlge_main.c | 1640 struct sk_buff *new_skb = NULL; ql_process_mac_rx_skb() local 1644 /* Allocate new_skb and copy */ ql_process_mac_rx_skb() 1645 new_skb = netdev_alloc_skb(qdev->ndev, length + NET_IP_ALIGN); ql_process_mac_rx_skb() 1646 if (new_skb == NULL) { ql_process_mac_rx_skb() 1650 skb_reserve(new_skb, NET_IP_ALIGN); ql_process_mac_rx_skb() 1651 memcpy(skb_put(new_skb, length), skb->data, length); ql_process_mac_rx_skb() 1652 skb = new_skb; ql_process_mac_rx_skb()
|
/linux-4.1.27/drivers/net/vmxnet3/ |
H A D | vmxnet3_drv.c | 1182 struct sk_buff *skb, *new_skb = NULL; vmxnet3_rq_rx_complete() local 1229 new_skb = netdev_alloc_skb_ip_align(adapter->netdev, vmxnet3_rq_rx_complete() 1231 if (new_skb == NULL) { vmxnet3_rq_rx_complete() 1256 rbi->skb = new_skb; vmxnet3_rq_rx_complete()
|
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/ |
H A D | rtl_core.c | 2326 struct sk_buff *new_skb; rtl8192_rx_normal() local 2333 new_skb = dev_alloc_skb(priv->rxbuffersize); rtl8192_rx_normal() 2337 if (unlikely(!new_skb)) rtl8192_rx_normal() 2381 skb = new_skb; rtl8192_rx_normal()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/ |
H A D | bcmgenet.c | 1220 struct sk_buff *new_skb; bcmgenet_put_tx_csum() local 1230 new_skb = skb_realloc_headroom(skb, sizeof(*status)); bcmgenet_put_tx_csum() 1232 if (!new_skb) { bcmgenet_put_tx_csum() 1237 skb = new_skb; bcmgenet_put_tx_csum()
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/ |
H A D | netdev.c | 1019 struct sk_buff *new_skb = e1000_clean_rx_irq() local 1021 if (new_skb) { e1000_clean_rx_irq() 1022 skb_copy_to_linear_data_offset(new_skb, e1000_clean_rx_irq() 1030 skb = new_skb; e1000_clean_rx_irq()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
H A D | tg3.c | 7788 struct sk_buff *new_skb, *skb = *pskb; tigon3_dma_hwbug_workaround() local 7793 new_skb = skb_copy(skb, GFP_ATOMIC); tigon3_dma_hwbug_workaround() 7797 new_skb = skb_copy_expand(skb, tigon3_dma_hwbug_workaround() 7802 if (!new_skb) { tigon3_dma_hwbug_workaround() 7806 new_addr = pci_map_single(tp->pdev, new_skb->data, new_skb->len, tigon3_dma_hwbug_workaround() 7810 dev_kfree_skb_any(new_skb); tigon3_dma_hwbug_workaround() 7817 tnapi->tx_buffers[*entry].skb = new_skb; tigon3_dma_hwbug_workaround() 7822 new_skb->len, base_flags, tigon3_dma_hwbug_workaround() 7825 dev_kfree_skb_any(new_skb); tigon3_dma_hwbug_workaround() 7832 *pskb = new_skb; tigon3_dma_hwbug_workaround()
|