Home
last modified time | relevance | path

Searched refs:skb2 (Results 1 – 78 of 78) sorted by relevance

/linux-4.1.27/drivers/net/usb/
Dcdc_eem.c107 struct sk_buff *skb2 = NULL; in eem_tx_fixup() local
140 skb2 = skb_copy_expand(skb, EEM_HEAD, ETH_FCS_LEN + padlen, flags); in eem_tx_fixup()
141 if (!skb2) in eem_tx_fixup()
145 skb = skb2; in eem_tx_fixup()
183 struct sk_buff *skb2 = NULL; in eem_rx_fixup() local
232 skb2 = skb_clone(skb, GFP_ATOMIC); in eem_rx_fixup()
233 if (unlikely(!skb2)) in eem_rx_fixup()
235 skb_trim(skb2, len); in eem_rx_fixup()
237 skb_push(skb2, 2)); in eem_rx_fixup()
238 eem_linkcmd(dev, skb2); in eem_rx_fixup()
[all …]
Dcx82310_eth.c224 struct sk_buff *skb2; in cx82310_rx_fixup() local
232 skb2 = alloc_skb(len, GFP_ATOMIC); in cx82310_rx_fixup()
233 if (!skb2) in cx82310_rx_fixup()
235 skb_put(skb2, len); in cx82310_rx_fixup()
236 memcpy(skb2->data, (void *)dev->partial_data, in cx82310_rx_fixup()
238 memcpy(skb2->data + dev->partial_len, skb->data, in cx82310_rx_fixup()
240 usbnet_skb_return(dev, skb2); in cx82310_rx_fixup()
275 skb2 = alloc_skb(len, GFP_ATOMIC); in cx82310_rx_fixup()
276 if (!skb2) in cx82310_rx_fixup()
278 skb_put(skb2, len); in cx82310_rx_fixup()
[all …]
Dkalmia.c168 struct sk_buff *skb2 = NULL; in kalmia_tx_fixup() local
191 skb2 = skb_copy_expand(skb, KALMIA_HEADER_LENGTH, in kalmia_tx_fixup()
193 if (!skb2) in kalmia_tx_fixup()
197 skb = skb2; in kalmia_tx_fixup()
250 struct sk_buff *skb2 = NULL; in kalmia_rx_fixup() local
308 skb2 = skb; in kalmia_rx_fixup()
311 skb2 = skb_clone(skb, GFP_ATOMIC); in kalmia_rx_fixup()
312 if (unlikely(!skb2)) in kalmia_rx_fixup()
316 skb_trim(skb2, ether_packet_length); in kalmia_rx_fixup()
322 usbnet_skb_return(dev, skb2); in kalmia_rx_fixup()
Dint51x1.c94 struct sk_buff *skb2; in int51x1_tx_fixup() local
96 skb2 = skb_copy_expand(skb, in int51x1_tx_fixup()
101 if (!skb2) in int51x1_tx_fixup()
103 skb = skb2; in int51x1_tx_fixup()
Drndis_host.c502 struct sk_buff *skb2; in rndis_rx_fixup() local
528 skb2 = skb_clone(skb, GFP_ATOMIC); in rndis_rx_fixup()
529 if (unlikely(!skb2)) in rndis_rx_fixup()
532 skb_trim(skb2, data_len); in rndis_rx_fixup()
533 usbnet_skb_return(dev, skb2); in rndis_rx_fixup()
545 struct sk_buff *skb2; in rndis_tx_fixup() local
566 skb2 = skb_copy_expand(skb, sizeof *hdr, 1, flags); in rndis_tx_fixup()
568 if (unlikely(!skb2)) in rndis_tx_fixup()
569 return skb2; in rndis_tx_fixup()
570 skb = skb2; in rndis_tx_fixup()
Dgl620a.c169 struct sk_buff *skb2; in genelink_tx_fixup() local
170 skb2 = skb_copy_expand(skb, (4 + 4*1) , padlen, flags); in genelink_tx_fixup()
172 skb = skb2; in genelink_tx_fixup()
Dnet1080.c438 struct sk_buff *skb2; in net1080_tx_fixup() local
467 skb2 = skb_copy_expand(skb, in net1080_tx_fixup()
472 if (!skb2) in net1080_tx_fixup()
473 return skb2; in net1080_tx_fixup()
474 skb = skb2; in net1080_tx_fixup()
Dzaurus.c60 struct sk_buff *skb2; in zaurus_tx_fixup() local
68 skb2 = skb_copy_expand(skb, 0, 4 + padlen, flags); in zaurus_tx_fixup()
70 skb = skb2; in zaurus_tx_fixup()
Dsr9700.c460 struct sk_buff *skb2; in sr9700_tx_fixup() local
462 skb2 = skb_copy_expand(skb, SR_TX_OVERHEAD, 0, flags); in sr9700_tx_fixup()
464 skb = skb2; in sr9700_tx_fixup()
Ddm9601.c497 struct sk_buff *skb2; in dm9601_tx_fixup() local
499 skb2 = skb_copy_expand(skb, DM_TX_OVERHEAD, pad, flags); in dm9601_tx_fixup()
501 skb = skb2; in dm9601_tx_fixup()
Dasix_common.c172 struct sk_buff *skb2; in asix_tx_fixup() local
174 skb2 = skb_copy_expand(skb, 4, padlen, flags); in asix_tx_fixup()
176 skb = skb2; in asix_tx_fixup()
Dsr9800.c128 struct sk_buff *skb2; in sr_tx_fixup() local
129 skb2 = skb_copy_expand(skb, 4, padlen, flags); in sr_tx_fixup()
131 skb = skb2; in sr_tx_fixup()
Dsmsc95xx.c1844 struct sk_buff *skb2 = skb_copy_expand(skb, in smsc95xx_tx_fixup() local
1847 skb = skb2; in smsc95xx_tx_fixup()
Dsmsc75xx.c2202 struct sk_buff *skb2 = in smsc75xx_tx_fixup() local
2205 skb = skb2; in smsc75xx_tx_fixup()
/linux-4.1.27/net/bridge/
Dbr_input.c130 struct sk_buff *skb2; in br_handle_frame_finish() local
155 skb2 = NULL; in br_handle_frame_finish()
158 skb2 = skb; in br_handle_frame_finish()
166 skb2 = skb; in br_handle_frame_finish()
174 skb2 = skb; in br_handle_frame_finish()
175 br_multicast_forward(mdst, skb, skb2); in br_handle_frame_finish()
177 if (!skb2) in br_handle_frame_finish()
180 skb2 = skb; in br_handle_frame_finish()
186 skb2 = skb; in br_handle_frame_finish()
194 br_forward(dst->dst, skb, skb2); in br_handle_frame_finish()
[all …]
Dbr_multicast.c1559 struct sk_buff *skb2 = skb; in br_multicast_ipv4_rcv() local
1594 skb2 = skb_clone(skb, GFP_ATOMIC); in br_multicast_ipv4_rcv()
1595 if (!skb2) in br_multicast_ipv4_rcv()
1598 err = pskb_trim_rcsum(skb2, len); in br_multicast_ipv4_rcv()
1603 len -= ip_hdrlen(skb2); in br_multicast_ipv4_rcv()
1604 offset = skb_network_offset(skb2) + ip_hdrlen(skb2); in br_multicast_ipv4_rcv()
1605 __skb_pull(skb2, offset); in br_multicast_ipv4_rcv()
1606 skb_reset_transport_header(skb2); in br_multicast_ipv4_rcv()
1609 if (!pskb_may_pull(skb2, sizeof(*ih))) in br_multicast_ipv4_rcv()
1612 switch (skb2->ip_summed) { in br_multicast_ipv4_rcv()
[all …]
Dbr_forward.c225 struct sk_buff *skb2, bool unicast) in br_flood_forward() argument
227 br_flood(br, skb, skb2, __br_forward, unicast); in br_flood_forward()
289 struct sk_buff *skb, struct sk_buff *skb2) in br_multicast_forward() argument
291 br_multicast_flood(mdst, skb, skb2, __br_forward); in br_multicast_forward()
Dbr_private.h419 struct sk_buff *skb2, bool unicast);
469 struct sk_buff *skb, struct sk_buff *skb2);
572 struct sk_buff *skb2) in br_multicast_forward() argument
/linux-4.1.27/net/sched/
Dact_mirred.c133 struct sk_buff *skb2; in tcf_mirred() local
154 skb2 = skb_act_clone(skb, GFP_ATOMIC, m->tcf_action); in tcf_mirred()
155 if (skb2 == NULL) in tcf_mirred()
160 skb_push(skb2, skb->mac_len); in tcf_mirred()
165 skb2->tc_verd = SET_TC_FROM(skb2->tc_verd, at); in tcf_mirred()
167 skb2->skb_iif = skb->dev->ifindex; in tcf_mirred()
168 skb2->dev = dev; in tcf_mirred()
169 skb_sender_cpu_clear(skb2); in tcf_mirred()
170 err = dev_queue_xmit(skb2); in tcf_mirred()
Dsch_choke.c170 struct sk_buff *skb2) in choke_match_flow() argument
174 if (skb1->protocol != skb2->protocol) in choke_match_flow()
183 if (!choke_skb_cb(skb2)->keys_valid) { in choke_match_flow()
184 choke_skb_cb(skb2)->keys_valid = 1; in choke_match_flow()
185 skb_flow_dissect(skb2, &temp); in choke_match_flow()
186 memcpy(&choke_skb_cb(skb2)->keys, &temp, CHOKE_K_LEN); in choke_match_flow()
190 &choke_skb_cb(skb2)->keys, in choke_match_flow()
Dsch_netem.c409 struct sk_buff *skb2; in netem_enqueue() local
440 if (count > 1 && (skb2 = skb_clone(skb, GFP_ATOMIC)) != NULL) { in netem_enqueue()
445 qdisc_enqueue_root(skb2, rootq); in netem_enqueue()
/linux-4.1.27/drivers/usb/gadget/function/
Df_eem.c352 struct sk_buff *skb2 = NULL; in eem_wrap() local
370 skb2 = skb_copy_expand(skb, EEM_HLEN, ETH_FCS_LEN + padlen, GFP_ATOMIC); in eem_wrap()
372 skb = skb2; in eem_wrap()
407 struct sk_buff *skb2; in eem_unwrap() local
447 skb2 = skb_clone(skb, GFP_ATOMIC); in eem_unwrap()
448 if (unlikely(!skb2)) { in eem_unwrap()
452 skb_trim(skb2, len); in eem_unwrap()
454 skb_push(skb2, 2)); in eem_unwrap()
455 skb_copy_bits(skb2, 0, req->buf, skb2->len); in eem_unwrap()
456 req->length = skb2->len; in eem_unwrap()
[all …]
Df_ncm.c895 struct sk_buff *skb2 = NULL; in package_for_tx() local
929 swap(skb2, ncm->skb_tx_data); in package_for_tx()
936 ntb_iter = (void *) skb_put(skb2, ndp_pad); in package_for_tx()
940 ntb_iter = (void *) skb_put(skb2, ncm->skb_tx_ndp->len); in package_for_tx()
946 ntb_iter = (void *) skb_put(skb2, dgram_idx_len); in package_for_tx()
949 return skb2; in package_for_tx()
956 struct sk_buff *skb2 = NULL; in ncm_wrap_ntb() local
996 skb2 = package_for_tx(ncm); in ncm_wrap_ntb()
997 if (!skb2) in ncm_wrap_ntb()
1071 skb2 = package_for_tx(ncm); in ncm_wrap_ntb()
[all …]
Du_ether.c265 struct sk_buff *skb = req->context, *skb2; in rx_complete() local
293 skb2 = skb_dequeue(&dev->rx_frames); in rx_complete()
294 while (skb2) { in rx_complete()
296 || ETH_HLEN > skb2->len in rx_complete()
297 || skb2->len > VLAN_ETH_FRAME_LEN) { in rx_complete()
300 DBG(dev, "rx length %d\n", skb2->len); in rx_complete()
301 dev_kfree_skb_any(skb2); in rx_complete()
304 skb2->protocol = eth_type_trans(skb2, dev->net); in rx_complete()
306 dev->net->stats.rx_bytes += skb2->len; in rx_complete()
311 status = netif_rx(skb2); in rx_complete()
[all …]
Df_rndis.c375 struct sk_buff *skb2; in rndis_add_header() local
377 skb2 = skb_realloc_headroom(skb, sizeof(struct rndis_packet_msg_type)); in rndis_add_header()
378 rndis_add_hdr(skb2); in rndis_add_header()
381 return skb2; in rndis_add_header()
/linux-4.1.27/drivers/net/team/
Dteam_mode_broadcast.c23 struct sk_buff *skb2; in bc_transmit() local
30 skb2 = skb_clone(skb, GFP_ATOMIC); in bc_transmit()
31 if (skb2) { in bc_transmit()
33 skb2); in bc_transmit()
/linux-4.1.27/net/decnet/netfilter/
Ddn_rtmsg.c66 struct sk_buff *skb2; in dnrmg_send_peer() local
82 skb2 = dnrmg_build_message(skb, &status); in dnrmg_send_peer()
83 if (skb2 == NULL) in dnrmg_send_peer()
85 NETLINK_CB(skb2).dst_group = group; in dnrmg_send_peer()
86 netlink_broadcast(dnrmg, skb2, 0, group, GFP_ATOMIC); in dnrmg_send_peer()
/linux-4.1.27/net/ipv4/
Dtcp_fastopen.c194 struct sk_buff *skb2; in tcp_fastopen_create_child() local
197 skb2 = skb_clone(skb, GFP_ATOMIC); in tcp_fastopen_create_child()
199 skb2 = skb_get(skb); in tcp_fastopen_create_child()
201 if (likely(skb2)) { in tcp_fastopen_create_child()
202 skb_dst_drop(skb2); in tcp_fastopen_create_child()
203 __skb_pull(skb2, tcp_hdrlen(skb)); in tcp_fastopen_create_child()
204 skb_set_owner_r(skb2, child); in tcp_fastopen_create_child()
205 __skb_queue_tail(&child->sk_receive_queue, skb2); in tcp_fastopen_create_child()
Dip_output.c187 struct sk_buff *skb2; in ip_finish_output2() local
189 skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev)); in ip_finish_output2()
190 if (!skb2) { in ip_finish_output2()
195 skb_set_owner_w(skb2, skb->sk); in ip_finish_output2()
197 skb = skb2; in ip_finish_output2()
495 struct sk_buff *skb2; in ip_fragment() local
672 skb2 = alloc_skb(len + hlen + ll_rs, GFP_ATOMIC); in ip_fragment()
673 if (!skb2) { in ip_fragment()
682 ip_copy_metadata(skb2, skb); in ip_fragment()
683 skb_reserve(skb2, ll_rs); in ip_fragment()
[all …]
Dip_input.c175 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_call_ra_chain() local
176 if (skb2) in ip_call_ra_chain()
177 raw_rcv(last, skb2); in ip_call_ra_chain()
Dipmr.c1904 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_forward() local
1906 if (skb2) in ip_mr_forward()
1907 ipmr_queue_xmit(net, mrt, skb2, cache, in ip_mr_forward()
1916 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_forward() local
1918 if (skb2) in ip_mr_forward()
1919 ipmr_queue_xmit(net, mrt, skb2, cache, psend); in ip_mr_forward()
2017 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip_mr_input() local
2019 if (!skb2) in ip_mr_input()
2021 skb = skb2; in ip_mr_input()
2220 struct sk_buff *skb2; in ipmr_get_route() local
[all …]
Dping.c978 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ping_rcv() local
981 if (skb2) in ping_rcv()
982 ping_queue_rcv_skb(sk, skb2); in ping_rcv()
Dtcp_output.c1126 static void tcp_fragment_tstamp(struct sk_buff *skb, struct sk_buff *skb2) in tcp_fragment_tstamp() argument
1131 !before(shinfo->tskey, TCP_SKB_CB(skb2)->seq)) { in tcp_fragment_tstamp()
1132 struct skb_shared_info *shinfo2 = skb_shinfo(skb2); in tcp_fragment_tstamp()
/linux-4.1.27/net/netfilter/
Dnfnetlink_cttimeout.c264 struct sk_buff *skb2; in cttimeout_get_timeout() local
269 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in cttimeout_get_timeout()
270 if (skb2 == NULL) { in cttimeout_get_timeout()
275 ret = ctnl_timeout_fill_info(skb2, NETLINK_CB(skb).portid, in cttimeout_get_timeout()
280 kfree_skb(skb2); in cttimeout_get_timeout()
283 ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, in cttimeout_get_timeout()
441 struct sk_buff *skb2; in cttimeout_default_get() local
457 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in cttimeout_default_get()
458 if (skb2 == NULL) { in cttimeout_default_get()
463 ret = cttimeout_default_fill_info(net, skb2, NETLINK_CB(skb).portid, in cttimeout_default_get()
[all …]
Dnfnetlink_acct.c287 struct sk_buff *skb2; in nfnl_acct_get() local
292 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_acct_get()
293 if (skb2 == NULL) { in nfnl_acct_get()
298 ret = nfnl_acct_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_acct_get()
303 kfree_skb(skb2); in nfnl_acct_get()
306 ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, in nfnl_acct_get()
Dnfnetlink_cthelper.c507 struct sk_buff *skb2; in nfnl_cthelper_get() local
546 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_cthelper_get()
547 if (skb2 == NULL) { in nfnl_cthelper_get()
552 ret = nfnl_cthelper_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_cthelper_get()
557 kfree_skb(skb2); in nfnl_cthelper_get()
561 ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, in nfnl_cthelper_get()
Dnf_tables_api.c514 struct sk_buff *skb2; in nf_tables_gettable() local
536 skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL); in nf_tables_gettable()
537 if (!skb2) in nf_tables_gettable()
540 err = nf_tables_fill_table_info(skb2, net, NETLINK_CB(skb).portid, in nf_tables_gettable()
546 return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); in nf_tables_gettable()
549 kfree_skb(skb2); in nf_tables_gettable()
1066 struct sk_buff *skb2; in nf_tables_getchain() local
1094 skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL); in nf_tables_getchain()
1095 if (!skb2) in nf_tables_getchain()
1098 err = nf_tables_fill_chain_info(skb2, net, NETLINK_CB(skb).portid, in nf_tables_getchain()
[all …]
Dnft_compat.c529 struct sk_buff *skb2; in nfnl_compat_get() local
568 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in nfnl_compat_get()
569 if (skb2 == NULL) in nfnl_compat_get()
573 if (nfnl_compat_fill_info(skb2, NETLINK_CB(skb).portid, in nfnl_compat_get()
579 kfree_skb(skb2); in nfnl_compat_get()
583 ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, in nfnl_compat_get()
Dnf_conntrack_netlink.c1112 struct sk_buff *skb2 = NULL; in ctnetlink_get_conntrack() local
1157 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ctnetlink_get_conntrack()
1158 if (skb2 == NULL) { in ctnetlink_get_conntrack()
1164 err = ctnetlink_fill_info(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, in ctnetlink_get_conntrack()
1171 err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); in ctnetlink_get_conntrack()
1178 kfree_skb(skb2); in ctnetlink_get_conntrack()
2014 struct sk_buff *skb2; in ctnetlink_stat_ct() local
2017 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ctnetlink_stat_ct()
2018 if (skb2 == NULL) in ctnetlink_stat_ct()
2021 err = ctnetlink_stat_ct_fill_info(skb2, NETLINK_CB(skb).portid, in ctnetlink_stat_ct()
[all …]
/linux-4.1.27/net/netfilter/ipset/
Dip_set_core.c1423 struct sk_buff *skb2; in call_ad() local
1432 skb2 = nlmsg_new(payload, GFP_KERNEL); in call_ad()
1433 if (skb2 == NULL) in call_ad()
1435 rep = __nlmsg_put(skb2, NETLINK_CB(skb).portid, in call_ad()
1450 netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); in call_ad()
1611 struct sk_buff *skb2; in ip_set_header() local
1623 skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); in ip_set_header()
1624 if (skb2 == NULL) in ip_set_header()
1627 nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, in ip_set_header()
1631 if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, IPSET_PROTOCOL) || in ip_set_header()
[all …]
/linux-4.1.27/net/caif/
Dcfpkt_skbuff.c342 struct sk_buff *skb2; in cfpkt_split() local
361 skb2 = pkt_to_skb(tmppkt); in cfpkt_split()
364 if (skb2 == NULL) in cfpkt_split()
371 memcpy(skb2->data, split, len2nd); in cfpkt_split()
372 skb2->tail += len2nd; in cfpkt_split()
373 skb2->len += len2nd; in cfpkt_split()
374 skb2->priority = skb->priority; in cfpkt_split()
375 return skb_to_pkt(skb2); in cfpkt_split()
/linux-4.1.27/net/ipv6/
Dip6_tunnel.c576 struct sk_buff *skb2; in ip4ip6_err() local
612 skb2 = skb_clone(skb, GFP_ATOMIC); in ip4ip6_err()
613 if (!skb2) in ip4ip6_err()
616 skb_dst_drop(skb2); in ip4ip6_err()
618 skb_pull(skb2, offset); in ip4ip6_err()
619 skb_reset_network_header(skb2); in ip4ip6_err()
620 eiph = ip_hdr(skb2); in ip4ip6_err()
630 skb2->dev = rt->dst.dev; in ip4ip6_err()
647 skb_dst_set(skb2, &rt->dst); in ip4ip6_err()
650 if (ip_route_input(skb2, eiph->daddr, eiph->saddr, eiph->tos, in ip4ip6_err()
[all …]
Dip6_input.c309 struct sk_buff *skb2; in ip6_mc_input() local
343 skb2 = skb_clone(skb, GFP_ATOMIC); in ip6_mc_input()
345 skb2 = skb; in ip6_mc_input()
349 if (skb2) { in ip6_mc_input()
350 ip6_mr_input(skb2); in ip6_mc_input()
Dsit.c489 struct sk_buff *skb2; in ipip6_err_gen_icmpv6_unreach() local
494 skb2 = skb_clone(skb, GFP_ATOMIC); in ipip6_err_gen_icmpv6_unreach()
496 if (!skb2) in ipip6_err_gen_icmpv6_unreach()
499 skb_dst_drop(skb2); in ipip6_err_gen_icmpv6_unreach()
500 skb_pull(skb2, ihl); in ipip6_err_gen_icmpv6_unreach()
501 skb_reset_network_header(skb2); in ipip6_err_gen_icmpv6_unreach()
503 rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr, NULL, 0, 0); in ipip6_err_gen_icmpv6_unreach()
506 skb2->dev = rt->dst.dev; in ipip6_err_gen_icmpv6_unreach()
508 icmpv6_send(skb2, ICMPV6_DEST_UNREACH, ICMPV6_ADDR_UNREACH, 0); in ipip6_err_gen_icmpv6_unreach()
513 kfree_skb(skb2); in ipip6_err_gen_icmpv6_unreach()
Dip6mr.c2160 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip6_mr_forward() local
2161 if (skb2) in ip6_mr_forward()
2162 ip6mr_forward2(net, mrt, skb2, cache, psend); in ip6_mr_forward()
2306 struct sk_buff *skb2; in ip6mr_get_route() local
2323 skb2 = alloc_skb(sizeof(struct ipv6hdr), GFP_ATOMIC); in ip6mr_get_route()
2324 if (!skb2) { in ip6mr_get_route()
2329 skb_reset_transport_header(skb2); in ip6mr_get_route()
2331 skb_put(skb2, sizeof(struct ipv6hdr)); in ip6mr_get_route()
2332 skb_reset_network_header(skb2); in ip6mr_get_route()
2334 iph = ipv6_hdr(skb2); in ip6mr_get_route()
[all …]
Dip6_output.c180 struct sk_buff *skb2 = skb_realloc_headroom(skb, head_room); in ip6_xmit() local
181 if (!skb2) { in ip6_xmit()
188 skb = skb2; in ip6_xmit()
251 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in ip6_call_ra_chain() local
252 if (skb2) in ip6_call_ra_chain()
253 rawv6_rcv(last, skb2); in ip6_call_ra_chain()
/linux-4.1.27/net/decnet/
Ddn_neigh.c220 struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); in dn_long_output() local
221 if (skb2 == NULL) { in dn_long_output()
227 skb = skb2; in dn_long_output()
267 struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); in dn_short_output() local
268 if (skb2 == NULL) { in dn_short_output()
274 skb = skb2; in dn_short_output()
308 struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); in dn_phase3_output() local
309 if (skb2 == NULL) { in dn_phase3_output()
315 skb = skb2; in dn_phase3_output()
Ddn_nsp_out.c215 struct sk_buff *skb2; in dn_nsp_clone_and_send() local
218 if ((skb2 = skb_clone(skb, gfp)) != NULL) { in dn_nsp_clone_and_send()
222 skb2->sk = skb->sk; in dn_nsp_clone_and_send()
223 dn_nsp_send(skb2); in dn_nsp_clone_and_send()
384 struct sk_buff *skb2, *n, *ack = NULL; in dn_nsp_check_xmit_queue() local
392 skb_queue_walk_safe(q, skb2, n) { in dn_nsp_check_xmit_queue()
393 struct dn_skb_cb *cb2 = DN_SKB_CB(skb2); in dn_nsp_check_xmit_queue()
396 ack = skb2; in dn_nsp_check_xmit_queue()
Ddn_dev.c964 struct sk_buff *skb2 = skb_copy(skb, GFP_ATOMIC); in dn_send_router_hello() local
965 if (skb2) { in dn_send_router_hello()
966 dn_rt_finish_output(skb2, dn_rt_all_end_mcast, src); in dn_send_router_hello()
/linux-4.1.27/net/mac802154/
Drx.c226 struct sk_buff *skb2; in ieee802154_monitors_rx() local
241 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee802154_monitors_rx()
242 if (skb2) { in ieee802154_monitors_rx()
243 skb2->dev = sdata->dev; in ieee802154_monitors_rx()
244 ieee802154_deliver_skb(skb2); in ieee802154_monitors_rx()
/linux-4.1.27/drivers/net/wireless/hostap/
Dhostap_80211_rx.c482 struct sk_buff *skb2; in hostap_rx_frame_mgmt() local
485 skb2 = skb_clone(skb, GFP_ATOMIC); in hostap_rx_frame_mgmt()
486 if (skb2) in hostap_rx_frame_mgmt()
487 hostap_rx(skb2->dev, skb2, rx_stats); in hostap_rx_frame_mgmt()
728 struct sk_buff *skb2 = NULL; in hostap_80211_rx() local
1075 skb2 = skb_clone(skb, GFP_ATOMIC); in hostap_80211_rx()
1076 if (skb2 == NULL) in hostap_80211_rx()
1083 skb2 = skb; in hostap_80211_rx()
1088 if (skb2 != NULL) { in hostap_80211_rx()
1090 skb2->dev = dev; in hostap_80211_rx()
[all …]
/linux-4.1.27/net/atm/
Dlec.c138 struct sk_buff *skb2; in lec_handle_bridge() local
141 skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC); in lec_handle_bridge()
142 if (skb2 == NULL) in lec_handle_bridge()
144 skb2->len = sizeof(struct atmlec_msg); in lec_handle_bridge()
145 mesg = (struct atmlec_msg *)skb2->data; in lec_handle_bridge()
152 atm_force_charge(priv->lecd, skb2->truesize); in lec_handle_bridge()
154 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_handle_bridge()
204 struct sk_buff *skb2; in lec_start_xmit() local
233 skb2 = skb_realloc_headroom(skb, LEC_HEADER_LEN); in lec_start_xmit()
234 if (unlikely(!skb2)) { in lec_start_xmit()
[all …]
Dbr2684.c212 struct sk_buff *skb2 = skb_realloc_headroom(skb, minheadroom); in br2684_xmit_vcc() local
215 if (skb2 == NULL) { in br2684_xmit_vcc()
219 skb = skb2; in br2684_xmit_vcc()
/linux-4.1.27/drivers/net/wireless/ipw2x00/
Dlibipw_rx.c186 struct sk_buff *skb2; in libipw_rx_frame_mgmt() local
189 skb2 = skb_clone(skb, GFP_ATOMIC); in libipw_rx_frame_mgmt()
190 if (skb2) in libipw_rx_frame_mgmt()
191 hostap_rx(skb2->dev, skb2, rx_stats); in libipw_rx_frame_mgmt()
344 struct sk_buff *skb2 = NULL; in libipw_rx() local
779 skb2 = skb_clone(skb, GFP_ATOMIC); in libipw_rx()
780 if (skb2 == NULL) in libipw_rx()
787 skb2 = skb; in libipw_rx()
792 if (skb2 != NULL) { in libipw_rx()
794 skb2->dev = dev; in libipw_rx()
[all …]
/linux-4.1.27/net/ipx/
Daf_ipx.c452 struct sk_buff *skb1 = NULL, *skb2 = NULL; in ipxitf_demux_socket() local
510 skb2 = skb_clone(skb1, GFP_ATOMIC); in ipxitf_demux_socket()
512 skb2 = skb1; in ipxitf_demux_socket()
517 if (!skb2) in ipxitf_demux_socket()
521 ipxitf_def_skb_handler(sock2, skb2); in ipxitf_demux_socket()
537 struct sk_buff *skb2; in ipxitf_adjust_skbuff() local
548 skb2 = alloc_skb(len, GFP_ATOMIC); in ipxitf_adjust_skbuff()
549 if (skb2) { in ipxitf_adjust_skbuff()
550 skb_reserve(skb2, out_offset); in ipxitf_adjust_skbuff()
551 skb_reset_network_header(skb2); in ipxitf_adjust_skbuff()
[all …]
/linux-4.1.27/drivers/net/ethernet/octeon/
Docteon_mgmt.c394 struct sk_buff *skb2; in octeon_mgmt_receive_one() local
429 re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2); in octeon_mgmt_receive_one()
433 skb_put(skb2, re2.s.len); in octeon_mgmt_receive_one()
434 skb_new = skb_copy_expand(skb, 0, skb2->len, in octeon_mgmt_receive_one()
438 if (skb_copy_bits(skb2, 0, skb_tail_pointer(skb_new), in octeon_mgmt_receive_one()
439 skb2->len)) in octeon_mgmt_receive_one()
441 skb_put(skb_new, skb2->len); in octeon_mgmt_receive_one()
443 dev_kfree_skb_any(skb2); in octeon_mgmt_receive_one()
458 dev_kfree_skb_any(skb2); in octeon_mgmt_receive_one()
460 re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2); in octeon_mgmt_receive_one()
[all …]
/linux-4.1.27/drivers/isdn/pcbit/
Ddrv.c472 struct sk_buff *skb2; in pcbit_l3_receive() local
491 if (capi_tdata_resp(chan, &skb2) > 0) in pcbit_l3_receive()
493 skb2, skb2->len); in pcbit_l3_receive()
537 if ((len = capi_disc_req(*(ushort *)(skb->data), &skb2, CAUSE_NOCHAN)) > 0) in pcbit_l3_receive()
538 pcbit_l2_write(dev, MSG_DISC_REQ, refnum, skb2, len); in pcbit_l3_receive()
/linux-4.1.27/drivers/net/wan/
Dz85230.c800 c->skb2 = NULL; in z8530_sync_open()
866 c->skb2 = NULL; in z8530_sync_dma_open()
1061 c->skb2 = NULL; in z8530_sync_txdma_open()
1668 c->skb = c->skb2; in z8530_rx_done()
1680 c->skb2 = dev_alloc_skb(c->mtu); in z8530_rx_done()
1681 if (c->skb2 == NULL) in z8530_rx_done()
1684 skb_put(c->skb2, c->mtu); in z8530_rx_done()
Dz85230.h274 struct sk_buff *skb2; /* Pending buffer */ member
Dhdlc_fr.c309 struct sk_buff *skb2 = skb_realloc_headroom(skb, in fr_hard_header() local
311 if (!skb2) in fr_hard_header()
314 skb = *skb_p = skb2; in fr_hard_header()
/linux-4.1.27/drivers/staging/gdm72xx/
Dgdm_wimax.c62 struct sk_buff *skb2; in gdm_wimax_header() local
64 skb2 = skb_realloc_headroom(skb, HCI_HEADER_SIZE); in gdm_wimax_header()
65 if (skb2 == NULL) in gdm_wimax_header()
68 skb_set_owner_w(skb2, skb->sk); in gdm_wimax_header()
70 skb = skb2; in gdm_wimax_header()
/linux-4.1.27/net/core/
Dskbuff.c268 kmemcheck_annotate_bitfield(&fclones->skb2, flags1); in __alloc_skb()
272 fclones->skb2.fclone = SKB_FCLONE_CLONE; in __alloc_skb()
273 fclones->skb2.pfmemalloc = pfmemalloc; in __alloc_skb()
677 fclones = container_of(skb, struct sk_buff_fclones, skb2); in kfree_skbmem()
998 n = &fclones->skb2; in skb_clone()
1248 struct sk_buff *skb2; in skb_realloc_headroom() local
1252 skb2 = pskb_copy(skb, GFP_ATOMIC); in skb_realloc_headroom()
1254 skb2 = skb_clone(skb, GFP_ATOMIC); in skb_realloc_headroom()
1255 if (skb2 && pskb_expand_head(skb2, SKB_DATA_ALIGN(delta), 0, in skb_realloc_headroom()
1257 kfree_skb(skb2); in skb_realloc_headroom()
[all …]
Ddev.c1806 struct sk_buff *skb2 = NULL; in dev_queue_xmit_nit() local
1820 deliver_skb(skb2, pt_prev, skb->dev); in dev_queue_xmit_nit()
1826 skb2 = skb_clone(skb, GFP_ATOMIC); in dev_queue_xmit_nit()
1827 if (!skb2) in dev_queue_xmit_nit()
1830 net_timestamp_set(skb2); in dev_queue_xmit_nit()
1836 skb_reset_mac_header(skb2); in dev_queue_xmit_nit()
1838 if (skb_network_header(skb2) < skb2->data || in dev_queue_xmit_nit()
1839 skb_network_header(skb2) > skb_tail_pointer(skb2)) { in dev_queue_xmit_nit()
1841 ntohs(skb2->protocol), in dev_queue_xmit_nit()
1843 skb_reset_network_header(skb2); in dev_queue_xmit_nit()
[all …]
/linux-4.1.27/net/netlink/
Daf_netlink.c1969 struct sk_buff *skb, *skb2; member
1996 if (p->skb2 == NULL) { in do_one_broadcast()
1998 p->skb2 = skb_clone(p->skb, p->allocation); in do_one_broadcast()
2000 p->skb2 = skb_get(p->skb); in do_one_broadcast()
2005 skb_orphan(p->skb2); in do_one_broadcast()
2008 if (p->skb2 == NULL) { in do_one_broadcast()
2014 } else if (p->tx_filter && p->tx_filter(sk, p->skb2, p->tx_data)) { in do_one_broadcast()
2015 kfree_skb(p->skb2); in do_one_broadcast()
2016 p->skb2 = NULL; in do_one_broadcast()
2017 } else if (sk_filter(sk, p->skb2)) { in do_one_broadcast()
[all …]
/linux-4.1.27/net/mac80211/
Dstatus.c649 struct sk_buff *skb2; in ieee80211_tx_status() local
885 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee80211_tx_status()
886 if (skb2) { in ieee80211_tx_status()
887 skb2->dev = prev_dev; in ieee80211_tx_status()
888 netif_rx(skb2); in ieee80211_tx_status()
Drx.c431 struct sk_buff *skb, *skb2; in ieee80211_rx_monitor() local
524 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee80211_rx_monitor()
525 if (skb2) { in ieee80211_rx_monitor()
526 skb2->dev = prev_dev; in ieee80211_rx_monitor()
527 netif_receive_skb(skb2); in ieee80211_rx_monitor()
3046 struct sk_buff *skb = rx->skb, *skb2; in ieee80211_rx_cooked_monitor() local
3090 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee80211_rx_cooked_monitor()
3091 if (skb2) { in ieee80211_rx_cooked_monitor()
3092 skb2->dev = prev_dev; in ieee80211_rx_cooked_monitor()
3093 netif_receive_skb(skb2); in ieee80211_rx_cooked_monitor()
Dtx.c1508 struct sk_buff *skb2; in ieee80211_tx_prepare_skb() local
1528 skb2 = __skb_dequeue(&tx.skbs); in ieee80211_tx_prepare_skb()
1529 if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) { in ieee80211_tx_prepare_skb()
1530 ieee80211_free_txskb(hw, skb2); in ieee80211_tx_prepare_skb()
/linux-4.1.27/Documentation/networking/
Dx25-iface.txt79 netif_rx(skb2), skb2 might be delivered to the network layer
/linux-4.1.27/net/llc/
Dllc_conn.c355 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in llc_conn_send_pdus() local
358 if (!skb2) in llc_conn_send_pdus()
360 skb = skb2; in llc_conn_send_pdus()
/linux-4.1.27/drivers/net/wireless/ath/wil6210/
Dtxrx.c951 struct sk_buff *skb2; in wil_find_tx_bcast_2() local
998 skb2 = skb_copy(skb, GFP_ATOMIC); in wil_find_tx_bcast_2()
999 if (skb2) { in wil_find_tx_bcast_2()
1001 wil_set_da_for_vring(wil, skb2, i); in wil_find_tx_bcast_2()
1002 wil_tx_vring(wil, v2, skb2); in wil_find_tx_bcast_2()
/linux-4.1.27/net/key/
Daf_key.c190 static int pfkey_broadcast_one(struct sk_buff *skb, struct sk_buff **skb2, in pfkey_broadcast_one() argument
196 if (*skb2 == NULL) { in pfkey_broadcast_one()
198 *skb2 = skb_clone(skb, allocation); in pfkey_broadcast_one()
200 *skb2 = skb; in pfkey_broadcast_one()
204 if (*skb2 != NULL) { in pfkey_broadcast_one()
206 skb_set_owner_r(*skb2, sk); in pfkey_broadcast_one()
207 skb_queue_tail(&sk->sk_receive_queue, *skb2); in pfkey_broadcast_one()
209 *skb2 = NULL; in pfkey_broadcast_one()
228 struct sk_buff *skb2 = NULL; in pfkey_broadcast() local
247 pfkey_broadcast_one(skb, &skb2, allocation, sk); in pfkey_broadcast()
[all …]
/linux-4.1.27/drivers/net/ethernet/qlogic/
Dqla3xxx.c2043 struct sk_buff *skb1 = NULL, *skb2; in ql_process_macip_rx_intr() local
2065 skb2 = lrg_buf_cb2->skb; in ql_process_macip_rx_intr()
2067 skb_put(skb2, length); /* Just the second buffer length here. */ in ql_process_macip_rx_intr()
2072 prefetch(skb2->data); in ql_process_macip_rx_intr()
2074 skb_checksum_none_assert(skb2); in ql_process_macip_rx_intr()
2081 skb_push(skb2, size), size); in ql_process_macip_rx_intr()
2095 skb2->ip_summed = CHECKSUM_UNNECESSARY; in ql_process_macip_rx_intr()
2098 skb2->protocol = eth_type_trans(skb2, qdev->ndev); in ql_process_macip_rx_intr()
2100 netif_receive_skb(skb2); in ql_process_macip_rx_intr()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/
Dcore.c218 struct sk_buff *skb2; in brcmf_netdev_start_xmit() local
223 skb2 = skb_realloc_headroom(skb, drvr->hdrlen); in brcmf_netdev_start_xmit()
225 skb = skb2; in brcmf_netdev_start_xmit()
/linux-4.1.27/drivers/staging/rtl8192u/ieee80211/
Dieee80211_rx.c250 struct sk_buff *skb2; in ieee80211_rx_frame_mgmt() local
253 skb2 = skb_clone(skb, GFP_ATOMIC); in ieee80211_rx_frame_mgmt()
254 if (skb2) in ieee80211_rx_frame_mgmt()
255 hostap_rx(skb2->dev, skb2, rx_stats); in ieee80211_rx_frame_mgmt()
908 struct sk_buff *skb2 = NULL; in ieee80211_rx() local
/linux-4.1.27/drivers/net/ethernet/tehuti/
Dtehuti.c1209 struct sk_buff *skb, *skb2; in bdx_rx_receive() local
1273 (skb2 = netdev_alloc_skb(priv->ndev, len + NET_IP_ALIGN))) { in bdx_rx_receive()
1274 skb_reserve(skb2, NET_IP_ALIGN); in bdx_rx_receive()
1279 memcpy(skb2->data, skb->data, len); in bdx_rx_receive()
1281 skb = skb2; in bdx_rx_receive()
/linux-4.1.27/net/appletalk/
Dddp.c1679 struct sk_buff *skb2 = skb_copy(skb, GFP_KERNEL); in atalk_sendmsg() local
1681 if (skb2) { in atalk_sendmsg()
1687 aarp_send_ddp(dev, skb2, &usat->sat_addr, NULL); in atalk_sendmsg()
/linux-4.1.27/include/linux/
Dskbuff.h796 struct sk_buff skb2; member
818 fclones->skb2.sk == sk; in skb_fclone_busy()
/linux-4.1.27/drivers/net/bonding/
Dbond_main.c3894 struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); in bond_xmit_broadcast() local
3896 if (!skb2) { in bond_xmit_broadcast()
3901 bond_dev_queue_xmit(bond, skb2, slave->dev); in bond_xmit_broadcast()