skb2 4129 drivers/net/bonding/bond_main.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 4131 drivers/net/bonding/bond_main.c if (!skb2) { skb2 4136 drivers/net/bonding/bond_main.c bond_dev_queue_xmit(bond, skb2, slave->dev); skb2 393 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c struct sk_buff *skb2; skb2 428 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2); skb2 432 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c skb_put(skb2, re2.s.len); skb2 433 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c skb_new = skb_copy_expand(skb, 0, skb2->len, skb2 437 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c if (skb_copy_bits(skb2, 0, skb_tail_pointer(skb_new), skb2 438 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c skb2->len)) skb2 440 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c skb_put(skb_new, skb2->len); skb2 442 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c dev_kfree_skb_any(skb2); skb2 457 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c dev_kfree_skb_any(skb2); skb2 459 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2); skb2 460 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c dev_kfree_skb_any(skb2); skb2 2046 drivers/net/ethernet/qlogic/qla3xxx.c struct sk_buff *skb1 = NULL, *skb2; skb2 2068 drivers/net/ethernet/qlogic/qla3xxx.c skb2 = lrg_buf_cb2->skb; skb2 2070 drivers/net/ethernet/qlogic/qla3xxx.c skb_put(skb2, length); /* Just the second buffer length here. */ skb2 2075 drivers/net/ethernet/qlogic/qla3xxx.c prefetch(skb2->data); skb2 2077 drivers/net/ethernet/qlogic/qla3xxx.c skb_checksum_none_assert(skb2); skb2 2084 drivers/net/ethernet/qlogic/qla3xxx.c skb_push(skb2, size), size); skb2 2098 drivers/net/ethernet/qlogic/qla3xxx.c skb2->ip_summed = CHECKSUM_UNNECESSARY; skb2 2101 drivers/net/ethernet/qlogic/qla3xxx.c skb2->protocol = eth_type_trans(skb2, qdev->ndev); skb2 2103 drivers/net/ethernet/qlogic/qla3xxx.c napi_gro_receive(&qdev->napi, skb2); skb2 1195 drivers/net/ethernet/tehuti/tehuti.c struct sk_buff *skb, *skb2; skb2 1259 drivers/net/ethernet/tehuti/tehuti.c (skb2 = netdev_alloc_skb(priv->ndev, len + NET_IP_ALIGN))) { skb2 1260 drivers/net/ethernet/tehuti/tehuti.c skb_reserve(skb2, NET_IP_ALIGN); skb2 1265 drivers/net/ethernet/tehuti/tehuti.c memcpy(skb2->data, skb->data, len); skb2 1267 drivers/net/ethernet/tehuti/tehuti.c skb = skb2; skb2 1213 drivers/net/gtp.c struct sk_buff *skb2; skb2 1227 drivers/net/gtp.c skb2 = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 1228 drivers/net/gtp.c if (skb2 == NULL) { skb2 1233 drivers/net/gtp.c err = gtp_genl_fill_info(skb2, NETLINK_CB(skb).portid, info->snd_seq, skb2 1239 drivers/net/gtp.c return genlmsg_unicast(genl_info_net(info), skb2, info->snd_portid); skb2 1242 drivers/net/gtp.c kfree_skb(skb2); skb2 19 drivers/net/team/team_mode_broadcast.c struct sk_buff *skb2; skb2 26 drivers/net/team/team_mode_broadcast.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 27 drivers/net/team/team_mode_broadcast.c if (skb2) { skb2 29 drivers/net/team/team_mode_broadcast.c skb2); skb2 250 drivers/net/usb/asix_common.c struct sk_buff *skb2; skb2 252 drivers/net/usb/asix_common.c skb2 = skb_copy_expand(skb, 4, padlen, flags); skb2 254 drivers/net/usb/asix_common.c skb = skb2; skb2 92 drivers/net/usb/cdc_eem.c struct sk_buff *skb2 = NULL; skb2 125 drivers/net/usb/cdc_eem.c skb2 = skb_copy_expand(skb, EEM_HEAD, ETH_FCS_LEN + padlen, flags); skb2 126 drivers/net/usb/cdc_eem.c if (!skb2) skb2 130 drivers/net/usb/cdc_eem.c skb = skb2; skb2 168 drivers/net/usb/cdc_eem.c struct sk_buff *skb2 = NULL; skb2 217 drivers/net/usb/cdc_eem.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 218 drivers/net/usb/cdc_eem.c if (unlikely(!skb2)) skb2 220 drivers/net/usb/cdc_eem.c skb_trim(skb2, len); skb2 222 drivers/net/usb/cdc_eem.c skb_push(skb2, 2)); skb2 223 drivers/net/usb/cdc_eem.c eem_linkcmd(dev, skb2); skb2 288 drivers/net/usb/cdc_eem.c skb2 = skb; skb2 290 drivers/net/usb/cdc_eem.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 291 drivers/net/usb/cdc_eem.c if (unlikely(!skb2)) skb2 302 drivers/net/usb/cdc_eem.c crc = get_unaligned_le32(skb2->data skb2 304 drivers/net/usb/cdc_eem.c crc2 = ~crc32_le(~0, skb2->data, skb2->len skb2 307 drivers/net/usb/cdc_eem.c crc = get_unaligned_be32(skb2->data skb2 311 drivers/net/usb/cdc_eem.c skb_trim(skb2, len - ETH_FCS_LEN); skb2 318 drivers/net/usb/cdc_eem.c dev_kfree_skb_any(skb2); skb2 320 drivers/net/usb/cdc_eem.c usbnet_skb_return(dev, skb2); skb2 213 drivers/net/usb/cx82310_eth.c struct sk_buff *skb2; skb2 221 drivers/net/usb/cx82310_eth.c skb2 = alloc_skb(len, GFP_ATOMIC); skb2 222 drivers/net/usb/cx82310_eth.c if (!skb2) skb2 224 drivers/net/usb/cx82310_eth.c skb_put(skb2, len); skb2 225 drivers/net/usb/cx82310_eth.c memcpy(skb2->data, (void *)dev->partial_data, skb2 227 drivers/net/usb/cx82310_eth.c memcpy(skb2->data + dev->partial_len, skb->data, skb2 229 drivers/net/usb/cx82310_eth.c usbnet_skb_return(dev, skb2); skb2 264 drivers/net/usb/cx82310_eth.c skb2 = alloc_skb(len, GFP_ATOMIC); skb2 265 drivers/net/usb/cx82310_eth.c if (!skb2) skb2 267 drivers/net/usb/cx82310_eth.c skb_put(skb2, len); skb2 268 drivers/net/usb/cx82310_eth.c memcpy(skb2->data, skb->data, len); skb2 270 drivers/net/usb/cx82310_eth.c usbnet_skb_return(dev, skb2); skb2 497 drivers/net/usb/dm9601.c struct sk_buff *skb2; skb2 499 drivers/net/usb/dm9601.c skb2 = skb_copy_expand(skb, DM_TX_OVERHEAD, pad, flags); skb2 501 drivers/net/usb/dm9601.c skb = skb2; skb2 156 drivers/net/usb/gl620a.c struct sk_buff *skb2; skb2 157 drivers/net/usb/gl620a.c skb2 = skb_copy_expand(skb, (4 + 4*1) , padlen, flags); skb2 159 drivers/net/usb/gl620a.c skb = skb2; skb2 83 drivers/net/usb/int51x1.c struct sk_buff *skb2; skb2 85 drivers/net/usb/int51x1.c skb2 = skb_copy_expand(skb, skb2 90 drivers/net/usb/int51x1.c if (!skb2) skb2 92 drivers/net/usb/int51x1.c skb = skb2; skb2 160 drivers/net/usb/kalmia.c struct sk_buff *skb2 = NULL; skb2 183 drivers/net/usb/kalmia.c skb2 = skb_copy_expand(skb, KALMIA_HEADER_LENGTH, skb2 185 drivers/net/usb/kalmia.c if (!skb2) skb2 189 drivers/net/usb/kalmia.c skb = skb2; skb2 242 drivers/net/usb/kalmia.c struct sk_buff *skb2 = NULL; skb2 300 drivers/net/usb/kalmia.c skb2 = skb; skb2 303 drivers/net/usb/kalmia.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 304 drivers/net/usb/kalmia.c if (unlikely(!skb2)) skb2 308 drivers/net/usb/kalmia.c skb_trim(skb2, ether_packet_length); skb2 314 drivers/net/usb/kalmia.c usbnet_skb_return(dev, skb2); skb2 2845 drivers/net/usb/lan78xx.c struct sk_buff *skb2 = NULL; skb2 2849 drivers/net/usb/lan78xx.c skb2 = lan78xx_tx_prep(dev, skb, GFP_ATOMIC); skb2 2852 drivers/net/usb/lan78xx.c if (skb2) { skb2 2853 drivers/net/usb/lan78xx.c skb_queue_tail(&dev->txq_pend, skb2); skb2 3103 drivers/net/usb/lan78xx.c struct sk_buff *skb2; skb2 3138 drivers/net/usb/lan78xx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 3139 drivers/net/usb/lan78xx.c if (unlikely(!skb2)) { skb2 3144 drivers/net/usb/lan78xx.c skb2->len = size; skb2 3145 drivers/net/usb/lan78xx.c skb2->data = packet; skb2 3146 drivers/net/usb/lan78xx.c skb_set_tail_pointer(skb2, size); skb2 3148 drivers/net/usb/lan78xx.c lan78xx_rx_csum_offload(dev, skb2, rx_cmd_a, rx_cmd_b); skb2 3149 drivers/net/usb/lan78xx.c lan78xx_rx_vlan_offload(dev, skb2, rx_cmd_a, rx_cmd_b); skb2 3151 drivers/net/usb/lan78xx.c skb_trim(skb2, skb2->len - 4); /* remove fcs */ skb2 3152 drivers/net/usb/lan78xx.c skb2->truesize = size + sizeof(struct sk_buff); skb2 3154 drivers/net/usb/lan78xx.c lan78xx_skb_return(dev, skb2); skb2 3324 drivers/net/usb/lan78xx.c struct sk_buff *skb, *skb2; skb2 3362 drivers/net/usb/lan78xx.c skb2 = skb_dequeue(tqp); skb2 3363 drivers/net/usb/lan78xx.c if (skb2) { skb2 3364 drivers/net/usb/lan78xx.c length += (skb2->len - TX_OVERHEAD); skb2 3365 drivers/net/usb/lan78xx.c memcpy(skb->data + pos, skb2->data, skb2->len); skb2 3366 drivers/net/usb/lan78xx.c pos += roundup(skb2->len, sizeof(u32)); skb2 3367 drivers/net/usb/lan78xx.c dev_kfree_skb(skb2); skb2 417 drivers/net/usb/net1080.c struct sk_buff *skb2; skb2 446 drivers/net/usb/net1080.c skb2 = skb_copy_expand(skb, skb2 451 drivers/net/usb/net1080.c if (!skb2) skb2 452 drivers/net/usb/net1080.c return skb2; skb2 453 drivers/net/usb/net1080.c skb = skb2; skb2 495 drivers/net/usb/rndis_host.c struct sk_buff *skb2; skb2 521 drivers/net/usb/rndis_host.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 522 drivers/net/usb/rndis_host.c if (unlikely(!skb2)) skb2 525 drivers/net/usb/rndis_host.c skb_trim(skb2, data_len); skb2 526 drivers/net/usb/rndis_host.c usbnet_skb_return(dev, skb2); skb2 538 drivers/net/usb/rndis_host.c struct sk_buff *skb2; skb2 559 drivers/net/usb/rndis_host.c skb2 = skb_copy_expand(skb, sizeof *hdr, 1, flags); skb2 561 drivers/net/usb/rndis_host.c if (unlikely(!skb2)) skb2 562 drivers/net/usb/rndis_host.c return skb2; skb2 563 drivers/net/usb/rndis_host.c skb = skb2; skb2 129 drivers/net/usb/sr9800.c struct sk_buff *skb2; skb2 130 drivers/net/usb/sr9800.c skb2 = skb_copy_expand(skb, 4, padlen, flags); skb2 132 drivers/net/usb/sr9800.c skb = skb2; skb2 48 drivers/net/usb/zaurus.c struct sk_buff *skb2; skb2 56 drivers/net/usb/zaurus.c skb2 = skb_copy_expand(skb, 0, 4 + padlen, flags); skb2 58 drivers/net/usb/zaurus.c skb = skb2; skb2 568 drivers/net/vrf.c struct sk_buff *skb2; skb2 570 drivers/net/vrf.c skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev)); skb2 571 drivers/net/vrf.c if (!skb2) { skb2 576 drivers/net/vrf.c skb_set_owner_w(skb2, skb->sk); skb2 579 drivers/net/vrf.c skb = skb2; skb2 307 drivers/net/wan/hdlc_fr.c struct sk_buff *skb2 = skb_realloc_headroom(skb, skb2 309 drivers/net/wan/hdlc_fr.c if (!skb2) skb2 312 drivers/net/wan/hdlc_fr.c skb = *skb_p = skb2; skb2 795 drivers/net/wan/z85230.c c->skb2 = NULL; skb2 861 drivers/net/wan/z85230.c c->skb2 = NULL; skb2 1056 drivers/net/wan/z85230.c c->skb2 = NULL; skb2 1663 drivers/net/wan/z85230.c c->skb = c->skb2; skb2 1675 drivers/net/wan/z85230.c c->skb2 = dev_alloc_skb(c->mtu); skb2 1676 drivers/net/wan/z85230.c if (c->skb2 == NULL) skb2 1679 drivers/net/wan/z85230.c skb_put(c->skb2, c->mtu); skb2 275 drivers/net/wan/z85230.h struct sk_buff *skb2; /* Pending buffer */ skb2 1557 drivers/net/wireless/ath/wil6210/txrx.c struct sk_buff *skb2; skb2 1609 drivers/net/wireless/ath/wil6210/txrx.c skb2 = skb_copy(skb, GFP_ATOMIC); skb2 1610 drivers/net/wireless/ath/wil6210/txrx.c if (skb2) { skb2 1612 drivers/net/wireless/ath/wil6210/txrx.c wil_set_da_for_vring(wil, skb2, i); skb2 1613 drivers/net/wireless/ath/wil6210/txrx.c wil_tx_ring(wil, vif, v2, skb2); skb2 1615 drivers/net/wireless/ath/wil6210/txrx.c dev_kfree_skb_any(skb2); skb2 182 drivers/net/wireless/intel/ipw2x00/libipw_rx.c struct sk_buff *skb2; skb2 185 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 186 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (skb2) skb2 187 drivers/net/wireless/intel/ipw2x00/libipw_rx.c hostap_rx(skb2->dev, skb2, rx_stats); skb2 340 drivers/net/wireless/intel/ipw2x00/libipw_rx.c struct sk_buff *skb2 = NULL; skb2 775 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 776 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (skb2 == NULL) skb2 783 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb2 = skb; skb2 788 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (skb2 != NULL) { skb2 790 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb2->dev = dev; skb2 791 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb2->protocol = htons(ETH_P_802_3); skb2 792 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb_reset_mac_header(skb2); skb2 793 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb_reset_network_header(skb2); skb2 795 drivers/net/wireless/intel/ipw2x00/libipw_rx.c dev_queue_xmit(skb2); skb2 481 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c struct sk_buff *skb2; skb2 484 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 485 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (skb2) skb2 486 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c hostap_rx(skb2->dev, skb2, rx_stats); skb2 727 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c struct sk_buff *skb2 = NULL; skb2 1074 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 1075 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (skb2 == NULL) skb2 1082 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb2 = skb; skb2 1087 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (skb2 != NULL) { skb2 1089 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb2->dev = dev; skb2 1090 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb2->protocol = cpu_to_be16(ETH_P_802_3); skb2 1091 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb_reset_mac_header(skb2); skb2 1092 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb_reset_network_header(skb2); skb2 1094 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c dev_queue_xmit(skb2); skb2 246 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct sk_buff *skb2; skb2 249 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 250 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (skb2) skb2 251 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c hostap_rx(skb2->dev, skb2, rx_stats); skb2 147 drivers/staging/wilc1000/wilc_mon.c struct sk_buff *skb2; skb2 163 drivers/staging/wilc1000/wilc_mon.c skb2 = dev_alloc_skb(skb->len + sizeof(*cb_hdr)); skb2 164 drivers/staging/wilc1000/wilc_mon.c if (!skb2) skb2 167 drivers/staging/wilc1000/wilc_mon.c skb_put_data(skb2, skb->data, skb->len); skb2 169 drivers/staging/wilc1000/wilc_mon.c cb_hdr = skb_push(skb2, sizeof(*cb_hdr)); skb2 181 drivers/staging/wilc1000/wilc_mon.c skb2->dev = dev; skb2 182 drivers/staging/wilc1000/wilc_mon.c skb_reset_mac_header(skb2); skb2 183 drivers/staging/wilc1000/wilc_mon.c skb2->ip_summed = CHECKSUM_UNNECESSARY; skb2 184 drivers/staging/wilc1000/wilc_mon.c skb2->pkt_type = PACKET_OTHERHOST; skb2 185 drivers/staging/wilc1000/wilc_mon.c skb2->protocol = htons(ETH_P_802_2); skb2 186 drivers/staging/wilc1000/wilc_mon.c memset(skb2->cb, 0, sizeof(skb2->cb)); skb2 188 drivers/staging/wilc1000/wilc_mon.c netif_rx(skb2); skb2 337 drivers/usb/gadget/function/f_eem.c struct sk_buff *skb2 = NULL; skb2 359 drivers/usb/gadget/function/f_eem.c skb2 = skb_copy_expand(skb, EEM_HLEN, ETH_FCS_LEN + padlen, GFP_ATOMIC); skb2 361 drivers/usb/gadget/function/f_eem.c skb = skb2; skb2 396 drivers/usb/gadget/function/f_eem.c struct sk_buff *skb2; skb2 436 drivers/usb/gadget/function/f_eem.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 437 drivers/usb/gadget/function/f_eem.c if (unlikely(!skb2)) { skb2 441 drivers/usb/gadget/function/f_eem.c skb_trim(skb2, len); skb2 443 drivers/usb/gadget/function/f_eem.c skb_push(skb2, 2)); skb2 444 drivers/usb/gadget/function/f_eem.c skb_copy_bits(skb2, 0, req->buf, skb2->len); skb2 445 drivers/usb/gadget/function/f_eem.c req->length = skb2->len; skb2 448 drivers/usb/gadget/function/f_eem.c req->context = skb2; skb2 497 drivers/usb/gadget/function/f_eem.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 498 drivers/usb/gadget/function/f_eem.c if (unlikely(!skb2)) { skb2 502 drivers/usb/gadget/function/f_eem.c skb_trim(skb2, len - ETH_FCS_LEN); skb2 504 drivers/usb/gadget/function/f_eem.c skb3 = skb_copy_expand(skb2, skb2 509 drivers/usb/gadget/function/f_eem.c dev_kfree_skb_any(skb2); skb2 512 drivers/usb/gadget/function/f_eem.c dev_kfree_skb_any(skb2); skb2 959 drivers/usb/gadget/function/f_ncm.c struct sk_buff *skb2 = NULL; skb2 993 drivers/usb/gadget/function/f_ncm.c swap(skb2, ncm->skb_tx_data); skb2 1000 drivers/usb/gadget/function/f_ncm.c skb_put_zero(skb2, ndp_pad); skb2 1003 drivers/usb/gadget/function/f_ncm.c skb_put_data(skb2, ncm->skb_tx_ndp->data, ncm->skb_tx_ndp->len); skb2 1008 drivers/usb/gadget/function/f_ncm.c skb_put_zero(skb2, dgram_idx_len); skb2 1010 drivers/usb/gadget/function/f_ncm.c return skb2; skb2 1017 drivers/usb/gadget/function/f_ncm.c struct sk_buff *skb2 = NULL; skb2 1057 drivers/usb/gadget/function/f_ncm.c skb2 = package_for_tx(ncm); skb2 1058 drivers/usb/gadget/function/f_ncm.c if (!skb2) skb2 1129 drivers/usb/gadget/function/f_ncm.c skb2 = package_for_tx(ncm); skb2 1130 drivers/usb/gadget/function/f_ncm.c if (!skb2) skb2 1134 drivers/usb/gadget/function/f_ncm.c return skb2; skb2 1184 drivers/usb/gadget/function/f_ncm.c struct sk_buff *skb2; skb2 1290 drivers/usb/gadget/function/f_ncm.c skb2 = netdev_alloc_skb_ip_align(ncm->netdev, skb2 1292 drivers/usb/gadget/function/f_ncm.c if (skb2 == NULL) skb2 1294 drivers/usb/gadget/function/f_ncm.c skb_put_data(skb2, skb->data + index, skb2 1297 drivers/usb/gadget/function/f_ncm.c skb_queue_tail(list, skb2); skb2 371 drivers/usb/gadget/function/f_rndis.c struct sk_buff *skb2; skb2 376 drivers/usb/gadget/function/f_rndis.c skb2 = skb_realloc_headroom(skb, sizeof(struct rndis_packet_msg_type)); skb2 377 drivers/usb/gadget/function/f_rndis.c rndis_add_hdr(skb2); skb2 380 drivers/usb/gadget/function/f_rndis.c return skb2; skb2 255 drivers/usb/gadget/function/u_ether.c struct sk_buff *skb = req->context, *skb2; skb2 283 drivers/usb/gadget/function/u_ether.c skb2 = skb_dequeue(&dev->rx_frames); skb2 284 drivers/usb/gadget/function/u_ether.c while (skb2) { skb2 286 drivers/usb/gadget/function/u_ether.c || ETH_HLEN > skb2->len skb2 287 drivers/usb/gadget/function/u_ether.c || skb2->len > GETHER_MAX_ETH_FRAME_LEN) { skb2 290 drivers/usb/gadget/function/u_ether.c DBG(dev, "rx length %d\n", skb2->len); skb2 291 drivers/usb/gadget/function/u_ether.c dev_kfree_skb_any(skb2); skb2 294 drivers/usb/gadget/function/u_ether.c skb2->protocol = eth_type_trans(skb2, dev->net); skb2 296 drivers/usb/gadget/function/u_ether.c dev->net->stats.rx_bytes += skb2->len; skb2 301 drivers/usb/gadget/function/u_ether.c status = netif_rx(skb2); skb2 303 drivers/usb/gadget/function/u_ether.c skb2 = skb_dequeue(&dev->rx_frames); skb2 1065 include/linux/skbuff.h struct sk_buff skb2; skb2 1088 include/linux/skbuff.h fclones->skb2.sk == sk; skb2 1677 net/appletalk/ddp.c struct sk_buff *skb2 = skb_copy(skb, GFP_KERNEL); skb2 1679 net/appletalk/ddp.c if (skb2) { skb2 1685 net/appletalk/ddp.c aarp_send_ddp(dev, skb2, &usat->sat_addr, NULL); skb2 213 net/atm/br2684.c struct sk_buff *skb2 = skb_realloc_headroom(skb, minheadroom); skb2 216 net/atm/br2684.c if (skb2 == NULL) { skb2 220 net/atm/br2684.c skb = skb2; skb2 142 net/atm/lec.c struct sk_buff *skb2; skb2 145 net/atm/lec.c skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC); skb2 146 net/atm/lec.c if (skb2 == NULL) skb2 148 net/atm/lec.c skb2->len = sizeof(struct atmlec_msg); skb2 149 net/atm/lec.c mesg = (struct atmlec_msg *)skb2->data; skb2 156 net/atm/lec.c atm_force_charge(priv->lecd, skb2->truesize); skb2 158 net/atm/lec.c skb_queue_tail(&sk->sk_receive_queue, skb2); skb2 207 net/atm/lec.c struct sk_buff *skb2; skb2 236 net/atm/lec.c skb2 = skb_realloc_headroom(skb, LEC_HEADER_LEN); skb2 237 net/atm/lec.c if (unlikely(!skb2)) { skb2 242 net/atm/lec.c skb = skb2; skb2 266 net/atm/lec.c skb2 = skb_copy_expand(skb, 0, skb2 270 net/atm/lec.c if (skb2 == NULL) { skb2 274 net/atm/lec.c skb = skb2; skb2 304 net/atm/lec.c while (entry && (skb2 = skb_dequeue(&entry->tx_wait))) { skb2 306 net/atm/lec.c lec_send(vcc, skb2); skb2 442 net/atm/lec.c struct sk_buff *skb2; skb2 447 net/atm/lec.c skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC); skb2 448 net/atm/lec.c if (skb2 == NULL) skb2 450 net/atm/lec.c skb2->len = sizeof(struct atmlec_msg); skb2 451 net/atm/lec.c skb_copy_to_linear_data(skb2, mesg, sizeof(*mesg)); skb2 452 net/atm/lec.c atm_force_charge(priv->lecd, skb2->truesize); skb2 454 net/atm/lec.c skb_queue_tail(&sk->sk_receive_queue, skb2); skb2 101 net/bridge/netfilter/nf_conntrack_bridge.c struct sk_buff *skb2; skb2 103 net/bridge/netfilter/nf_conntrack_bridge.c skb2 = ip_frag_next(skb, &state); skb2 104 net/bridge/netfilter/nf_conntrack_bridge.c if (IS_ERR(skb2)) { skb2 105 net/bridge/netfilter/nf_conntrack_bridge.c err = PTR_ERR(skb2); skb2 109 net/bridge/netfilter/nf_conntrack_bridge.c skb2->tstamp = tstamp; skb2 110 net/bridge/netfilter/nf_conntrack_bridge.c err = output(net, sk, data, skb2); skb2 333 net/caif/cfpkt_skbuff.c struct sk_buff *skb2; skb2 352 net/caif/cfpkt_skbuff.c skb2 = pkt_to_skb(tmppkt); skb2 355 net/caif/cfpkt_skbuff.c if (skb2 == NULL) skb2 358 net/caif/cfpkt_skbuff.c skb_put_data(skb2, split, len2nd); skb2 363 net/caif/cfpkt_skbuff.c skb2->priority = skb->priority; skb2 364 net/caif/cfpkt_skbuff.c return skb_to_pkt(skb2); skb2 1940 net/core/dev.c struct sk_buff *skb2 = NULL; skb2 1957 net/core/dev.c deliver_skb(skb2, pt_prev, skb->dev); skb2 1963 net/core/dev.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 1964 net/core/dev.c if (!skb2) skb2 1967 net/core/dev.c net_timestamp_set(skb2); skb2 1973 net/core/dev.c skb_reset_mac_header(skb2); skb2 1975 net/core/dev.c if (skb_network_header(skb2) < skb2->data || skb2 1976 net/core/dev.c skb_network_header(skb2) > skb_tail_pointer(skb2)) { skb2 1978 net/core/dev.c ntohs(skb2->protocol), skb2 1980 net/core/dev.c skb_reset_network_header(skb2); skb2 1983 net/core/dev.c skb2->transport_header = skb2->network_header; skb2 1984 net/core/dev.c skb2->pkt_type = PACKET_OUTGOING; skb2 1994 net/core/dev.c if (!skb_orphan_frags_rx(skb2, GFP_ATOMIC)) skb2 1995 net/core/dev.c pt_prev->func(skb2, skb->dev, pt_prev, skb->dev); skb2 1997 net/core/dev.c kfree_skb(skb2); skb2 249 net/core/skbuff.c fclones->skb2.fclone = SKB_FCLONE_CLONE; skb2 638 net/core/skbuff.c fclones = container_of(skb, struct sk_buff_fclones, skb2); skb2 1442 net/core/skbuff.c n = &fclones->skb2; skb2 1703 net/core/skbuff.c struct sk_buff *skb2; skb2 1707 net/core/skbuff.c skb2 = pskb_copy(skb, GFP_ATOMIC); skb2 1709 net/core/skbuff.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 1710 net/core/skbuff.c if (skb2 && pskb_expand_head(skb2, SKB_DATA_ALIGN(delta), 0, skb2 1712 net/core/skbuff.c kfree_skb(skb2); skb2 1713 net/core/skbuff.c skb2 = NULL; skb2 1716 net/core/skbuff.c return skb2; skb2 4363 net/core/skbuff.c struct sk_buff *skb2; skb2 4367 net/core/skbuff.c skb2 = skb_copy(skb1, GFP_ATOMIC); skb2 4369 net/core/skbuff.c skb2 = skb_copy_expand(skb1, skb2 4373 net/core/skbuff.c if (unlikely(skb2 == NULL)) skb2 4377 net/core/skbuff.c skb_set_owner_w(skb2, skb1->sk); skb2 4382 net/core/skbuff.c skb2->next = skb1->next; skb2 4383 net/core/skbuff.c *skb_p = skb2; skb2 4385 net/core/skbuff.c skb1 = skb2; skb2 969 net/decnet/dn_dev.c struct sk_buff *skb2 = skb_copy(skb, GFP_ATOMIC); skb2 970 net/decnet/dn_dev.c if (skb2) { skb2 971 net/decnet/dn_dev.c dn_rt_finish_output(skb2, dn_rt_all_end_mcast, src); skb2 221 net/decnet/dn_neigh.c struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); skb2 222 net/decnet/dn_neigh.c if (skb2 == NULL) { skb2 228 net/decnet/dn_neigh.c skb = skb2; skb2 269 net/decnet/dn_neigh.c struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); skb2 270 net/decnet/dn_neigh.c if (skb2 == NULL) { skb2 276 net/decnet/dn_neigh.c skb = skb2; skb2 311 net/decnet/dn_neigh.c struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom); skb2 312 net/decnet/dn_neigh.c if (skb2 == NULL) { skb2 318 net/decnet/dn_neigh.c skb = skb2; skb2 207 net/decnet/dn_nsp_out.c struct sk_buff *skb2; skb2 210 net/decnet/dn_nsp_out.c if ((skb2 = skb_clone(skb, gfp)) != NULL) { skb2 214 net/decnet/dn_nsp_out.c skb2->sk = skb->sk; skb2 215 net/decnet/dn_nsp_out.c dn_nsp_send(skb2); skb2 373 net/decnet/dn_nsp_out.c struct sk_buff *skb2, *n, *ack = NULL; skb2 381 net/decnet/dn_nsp_out.c skb_queue_walk_safe(q, skb2, n) { skb2 382 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb2 = DN_SKB_CB(skb2); skb2 385 net/decnet/dn_nsp_out.c ack = skb2; skb2 64 net/decnet/netfilter/dn_rtmsg.c struct sk_buff *skb2; skb2 80 net/decnet/netfilter/dn_rtmsg.c skb2 = dnrmg_build_message(skb, &status); skb2 81 net/decnet/netfilter/dn_rtmsg.c if (skb2 == NULL) skb2 83 net/decnet/netfilter/dn_rtmsg.c NETLINK_CB(skb2).dst_group = group; skb2 84 net/decnet/netfilter/dn_rtmsg.c netlink_broadcast(dnrmg, skb2, 0, group, GFP_ATOMIC); skb2 170 net/ipv4/ip_input.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 171 net/ipv4/ip_input.c if (skb2) skb2 172 net/ipv4/ip_input.c raw_rcv(last, skb2); skb2 201 net/ipv4/ip_output.c struct sk_buff *skb2; skb2 203 net/ipv4/ip_output.c skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev)); skb2 204 net/ipv4/ip_output.c if (!skb2) { skb2 209 net/ipv4/ip_output.c skb_set_owner_w(skb2, skb->sk); skb2 211 net/ipv4/ip_output.c skb = skb2; skb2 685 net/ipv4/ip_output.c struct sk_buff *skb2; skb2 699 net/ipv4/ip_output.c skb2 = alloc_skb(len + state->hlen + state->ll_rs, GFP_ATOMIC); skb2 700 net/ipv4/ip_output.c if (!skb2) skb2 707 net/ipv4/ip_output.c ip_copy_metadata(skb2, skb); skb2 708 net/ipv4/ip_output.c skb_reserve(skb2, state->ll_rs); skb2 709 net/ipv4/ip_output.c skb_put(skb2, len + state->hlen); skb2 710 net/ipv4/ip_output.c skb_reset_network_header(skb2); skb2 711 net/ipv4/ip_output.c skb2->transport_header = skb2->network_header + state->hlen; skb2 719 net/ipv4/ip_output.c skb_set_owner_w(skb2, skb->sk); skb2 725 net/ipv4/ip_output.c skb_copy_from_linear_data(skb, skb_network_header(skb2), state->hlen); skb2 730 net/ipv4/ip_output.c if (skb_copy_bits(skb, state->ptr, skb_transport_header(skb2), len)) skb2 737 net/ipv4/ip_output.c iph = ip_hdr(skb2); skb2 755 net/ipv4/ip_output.c return skb2; skb2 770 net/ipv4/ip_output.c struct sk_buff *skb2; skb2 896 net/ipv4/ip_output.c skb2 = ip_frag_next(skb, &state); skb2 897 net/ipv4/ip_output.c if (IS_ERR(skb2)) { skb2 898 net/ipv4/ip_output.c err = PTR_ERR(skb2); skb2 901 net/ipv4/ip_output.c ip_frag_ipcb(skb, skb2, first_frag, &state); skb2 906 net/ipv4/ip_output.c skb2->tstamp = tstamp; skb2 907 net/ipv4/ip_output.c err = output(net, sk, skb2); skb2 2041 net/ipv4/ipmr.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 2043 net/ipv4/ipmr.c if (skb2) skb2 2045 net/ipv4/ipmr.c skb2, psend); skb2 2053 net/ipv4/ipmr.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 2055 net/ipv4/ipmr.c if (skb2) skb2 2056 net/ipv4/ipmr.c ipmr_queue_xmit(net, mrt, true_vifi, skb2, skb2 2165 net/ipv4/ipmr.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 2167 net/ipv4/ipmr.c if (!skb2) skb2 2169 net/ipv4/ipmr.c skb = skb2; skb2 2278 net/ipv4/ipmr.c struct sk_buff *skb2; skb2 2293 net/ipv4/ipmr.c skb2 = skb_realloc_headroom(skb, sizeof(struct iphdr)); skb2 2294 net/ipv4/ipmr.c if (!skb2) { skb2 2300 net/ipv4/ipmr.c NETLINK_CB(skb2).portid = portid; skb2 2301 net/ipv4/ipmr.c skb_push(skb2, sizeof(struct iphdr)); skb2 2302 net/ipv4/ipmr.c skb_reset_network_header(skb2); skb2 2303 net/ipv4/ipmr.c iph = ip_hdr(skb2); skb2 2308 net/ipv4/ipmr.c err = ipmr_cache_unresolved(mrt, vif, skb2, dev); skb2 974 net/ipv4/ping.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 977 net/ipv4/ping.c if (skb2) skb2 978 net/ipv4/ping.c ping_queue_rcv_skb(sk, skb2); skb2 1263 net/ipv4/tcp_output.c static void tcp_fragment_tstamp(struct sk_buff *skb, struct sk_buff *skb2) skb2 1268 net/ipv4/tcp_output.c !before(shinfo->tskey, TCP_SKB_CB(skb2)->seq)) { skb2 1269 net/ipv4/tcp_output.c struct skb_shared_info *shinfo2 = skb_shinfo(skb2); skb2 1275 net/ipv4/tcp_output.c TCP_SKB_CB(skb2)->txstamp_ack = TCP_SKB_CB(skb)->txstamp_ack; skb2 1280 net/ipv4/tcp_output.c static void tcp_skb_fragment_eor(struct sk_buff *skb, struct sk_buff *skb2) skb2 1282 net/ipv4/tcp_output.c TCP_SKB_CB(skb2)->eor = TCP_SKB_CB(skb)->eor; skb2 617 net/ipv6/icmp.c struct sk_buff *skb2; skb2 627 net/ipv6/icmp.c skb2 = data_len ? skb_copy(skb, GFP_ATOMIC) : skb_clone(skb, GFP_ATOMIC); skb2 629 net/ipv6/icmp.c if (!skb2) skb2 632 net/ipv6/icmp.c skb_dst_drop(skb2); skb2 633 net/ipv6/icmp.c skb_pull(skb2, nhs); skb2 634 net/ipv6/icmp.c skb_reset_network_header(skb2); skb2 636 net/ipv6/icmp.c rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr, NULL, 0, skb2 640 net/ipv6/icmp.c skb2->dev = rt->dst.dev; skb2 648 net/ipv6/icmp.c __skb_push(skb2, nhs); skb2 649 net/ipv6/icmp.c skb_reset_network_header(skb2); skb2 650 net/ipv6/icmp.c memmove(skb2->data, skb2->data + nhs, data_len - nhs); skb2 651 net/ipv6/icmp.c memset(skb2->data + data_len - nhs, 0, nhs); skb2 658 net/ipv6/icmp.c icmp6_send(skb2, ICMPV6_TIME_EXCEED, ICMPV6_EXC_HOPLIMIT, skb2 661 net/ipv6/icmp.c icmp6_send(skb2, ICMPV6_DEST_UNREACH, ICMPV6_ADDR_UNREACH, skb2 666 net/ipv6/icmp.c kfree_skb(skb2); skb2 506 net/ipv6/ip6_input.c struct sk_buff *skb2; skb2 540 net/ipv6/ip6_input.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 542 net/ipv6/ip6_input.c skb2 = skb; skb2 546 net/ipv6/ip6_input.c if (skb2) { skb2 547 net/ipv6/ip6_input.c ip6_mr_input(skb2); skb2 214 net/ipv6/ip6_output.c struct sk_buff *skb2 = skb_realloc_headroom(skb, head_room); skb2 215 net/ipv6/ip6_output.c if (!skb2) { skb2 222 net/ipv6/ip6_output.c skb_set_owner_w(skb2, skb->sk); skb2 224 net/ipv6/ip6_output.c skb = skb2; skb2 314 net/ipv6/ip6_output.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 315 net/ipv6/ip6_output.c if (skb2) skb2 316 net/ipv6/ip6_output.c rawv6_rcv(last, skb2); skb2 566 net/ipv6/ip6_tunnel.c struct sk_buff *skb2; skb2 601 net/ipv6/ip6_tunnel.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 602 net/ipv6/ip6_tunnel.c if (!skb2) skb2 605 net/ipv6/ip6_tunnel.c skb_dst_drop(skb2); skb2 607 net/ipv6/ip6_tunnel.c skb_pull(skb2, offset); skb2 608 net/ipv6/ip6_tunnel.c skb_reset_network_header(skb2); skb2 609 net/ipv6/ip6_tunnel.c eiph = ip_hdr(skb2); skb2 617 net/ipv6/ip6_tunnel.c skb2->dev = rt->dst.dev; skb2 630 net/ipv6/ip6_tunnel.c skb_dst_set(skb2, &rt->dst); skb2 632 net/ipv6/ip6_tunnel.c if (ip_route_input(skb2, eiph->daddr, eiph->saddr, eiph->tos, skb2 633 net/ipv6/ip6_tunnel.c skb2->dev) || skb2 634 net/ipv6/ip6_tunnel.c skb_dst(skb2)->dev->type != ARPHRD_TUNNEL6) skb2 640 net/ipv6/ip6_tunnel.c if (rel_info > dst_mtu(skb_dst(skb2))) skb2 643 net/ipv6/ip6_tunnel.c skb_dst_update_pmtu_no_confirm(skb2, rel_info); skb2 646 net/ipv6/ip6_tunnel.c icmp_send(skb2, rel_type, rel_code, htonl(rel_info)); skb2 649 net/ipv6/ip6_tunnel.c kfree_skb(skb2); skb2 669 net/ipv6/ip6_tunnel.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 671 net/ipv6/ip6_tunnel.c if (!skb2) skb2 674 net/ipv6/ip6_tunnel.c skb_dst_drop(skb2); skb2 675 net/ipv6/ip6_tunnel.c skb_pull(skb2, offset); skb2 676 net/ipv6/ip6_tunnel.c skb_reset_network_header(skb2); skb2 679 net/ipv6/ip6_tunnel.c rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr, skb2 680 net/ipv6/ip6_tunnel.c NULL, 0, skb2, 0); skb2 683 net/ipv6/ip6_tunnel.c skb2->dev = rt->dst.dev; skb2 685 net/ipv6/ip6_tunnel.c icmpv6_send(skb2, rel_type, rel_code, rel_info); skb2 689 net/ipv6/ip6_tunnel.c kfree_skb(skb2); skb2 2161 net/ipv6/ip6mr.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 2162 net/ipv6/ip6mr.c if (skb2) skb2 2163 net/ipv6/ip6mr.c ip6mr_forward2(net, mrt, skb2, psend); skb2 2273 net/ipv6/ip6mr.c struct sk_buff *skb2; skb2 2285 net/ipv6/ip6mr.c skb2 = alloc_skb(sizeof(struct ipv6hdr), GFP_ATOMIC); skb2 2286 net/ipv6/ip6mr.c if (!skb2) { skb2 2291 net/ipv6/ip6mr.c NETLINK_CB(skb2).portid = portid; skb2 2292 net/ipv6/ip6mr.c skb_reset_transport_header(skb2); skb2 2294 net/ipv6/ip6mr.c skb_put(skb2, sizeof(struct ipv6hdr)); skb2 2295 net/ipv6/ip6mr.c skb_reset_network_header(skb2); skb2 2297 net/ipv6/ip6mr.c iph = ipv6_hdr(skb2); skb2 2309 net/ipv6/ip6mr.c err = ip6mr_cache_unresolved(mrt, vif, skb2, dev); skb2 212 net/ipv6/netfilter.c struct sk_buff *skb2; skb2 214 net/ipv6/netfilter.c skb2 = ip6_frag_next(skb, &state); skb2 215 net/ipv6/netfilter.c if (IS_ERR(skb2)) { skb2 216 net/ipv6/netfilter.c err = PTR_ERR(skb2); skb2 220 net/ipv6/netfilter.c skb2->tstamp = tstamp; skb2 221 net/ipv6/netfilter.c err = output(net, sk, data, skb2); skb2 336 net/llc/llc_conn.c struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); skb2 339 net/llc/llc_conn.c if (!skb2) skb2 341 net/llc/llc_conn.c skb = skb2; skb2 3513 net/mac80211/rx.c struct sk_buff *skb = rx->skb, *skb2; skb2 3557 net/mac80211/rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 3558 net/mac80211/rx.c if (skb2) { skb2 3559 net/mac80211/rx.c skb2->dev = prev_dev; skb2 3560 net/mac80211/rx.c netif_receive_skb(skb2); skb2 809 net/mac80211/status.c struct sk_buff *skb2; skb2 843 net/mac80211/status.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 844 net/mac80211/status.c if (skb2) { skb2 845 net/mac80211/status.c skb2->dev = prev_dev; skb2 846 net/mac80211/status.c netif_rx(skb2); skb2 1863 net/mac80211/tx.c struct sk_buff *skb2; skb2 1883 net/mac80211/tx.c skb2 = __skb_dequeue(&tx.skbs); skb2 1884 net/mac80211/tx.c if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) { skb2 1885 net/mac80211/tx.c ieee80211_free_txskb(hw, skb2); skb2 223 net/mac802154/rx.c struct sk_buff *skb2; skb2 238 net/mac802154/rx.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 239 net/mac802154/rx.c if (skb2) { skb2 240 net/mac802154/rx.c skb2->dev = sdata->dev; skb2 241 net/mac802154/rx.c ieee802154_deliver_skb(skb2); skb2 1553 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1562 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(payload, GFP_KERNEL); skb2 1563 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1565 net/netfilter/ipset/ip_set_core.c rep = __nlmsg_put(skb2, NETLINK_CB(skb).portid, skb2 1577 net/netfilter/ipset/ip_set_core.c nlmsg_free(skb2); skb2 1584 net/netfilter/ipset/ip_set_core.c netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, skb2 1711 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1723 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1724 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1727 net/netfilter/ipset/ip_set_core.c nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, skb2 1731 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) || skb2 1732 net/netfilter/ipset/ip_set_core.c nla_put_string(skb2, IPSET_ATTR_SETNAME, set->name) || skb2 1733 net/netfilter/ipset/ip_set_core.c nla_put_string(skb2, IPSET_ATTR_TYPENAME, set->type->name) || skb2 1734 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_FAMILY, set->family) || skb2 1735 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_REVISION, set->revision)) skb2 1737 net/netfilter/ipset/ip_set_core.c nlmsg_end(skb2, nlh2); skb2 1739 net/netfilter/ipset/ip_set_core.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1746 net/netfilter/ipset/ip_set_core.c nlmsg_cancel(skb2, nlh2); skb2 1748 net/netfilter/ipset/ip_set_core.c kfree_skb(skb2); skb2 1766 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1783 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1784 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1787 net/netfilter/ipset/ip_set_core.c nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, skb2 1791 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) || skb2 1792 net/netfilter/ipset/ip_set_core.c nla_put_string(skb2, IPSET_ATTR_TYPENAME, typename) || skb2 1793 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_FAMILY, family) || skb2 1794 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_REVISION, max) || skb2 1795 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_REVISION_MIN, min)) skb2 1797 net/netfilter/ipset/ip_set_core.c nlmsg_end(skb2, nlh2); skb2 1800 net/netfilter/ipset/ip_set_core.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1807 net/netfilter/ipset/ip_set_core.c nlmsg_cancel(skb2, nlh2); skb2 1809 net/netfilter/ipset/ip_set_core.c kfree_skb(skb2); skb2 1825 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1832 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1833 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1836 net/netfilter/ipset/ip_set_core.c nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, skb2 1840 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, IPSET_PROTOCOL)) skb2 1842 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL_MIN, IPSET_PROTOCOL_MIN)) skb2 1844 net/netfilter/ipset/ip_set_core.c nlmsg_end(skb2, nlh2); skb2 1846 net/netfilter/ipset/ip_set_core.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1853 net/netfilter/ipset/ip_set_core.c nlmsg_cancel(skb2, nlh2); skb2 1855 net/netfilter/ipset/ip_set_core.c kfree_skb(skb2); skb2 1867 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1881 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1882 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1885 net/netfilter/ipset/ip_set_core.c nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, skb2 1889 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) || skb2 1890 net/netfilter/ipset/ip_set_core.c nla_put_u8(skb2, IPSET_ATTR_FAMILY, set->family) || skb2 1891 net/netfilter/ipset/ip_set_core.c nla_put_net16(skb2, IPSET_ATTR_INDEX, htons(id))) skb2 1893 net/netfilter/ipset/ip_set_core.c nlmsg_end(skb2, nlh2); skb2 1895 net/netfilter/ipset/ip_set_core.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1902 net/netfilter/ipset/ip_set_core.c nlmsg_cancel(skb2, nlh2); skb2 1904 net/netfilter/ipset/ip_set_core.c kfree_skb(skb2); skb2 1919 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb2; skb2 1936 net/netfilter/ipset/ip_set_core.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1937 net/netfilter/ipset/ip_set_core.c if (!skb2) skb2 1940 net/netfilter/ipset/ip_set_core.c nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0, skb2 1944 net/netfilter/ipset/ip_set_core.c if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) || skb2 1945 net/netfilter/ipset/ip_set_core.c nla_put_string(skb2, IPSET_ATTR_SETNAME, set->name)) skb2 1947 net/netfilter/ipset/ip_set_core.c nlmsg_end(skb2, nlh2); skb2 1949 net/netfilter/ipset/ip_set_core.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1956 net/netfilter/ipset/ip_set_core.c nlmsg_cancel(skb2, nlh2); skb2 1958 net/netfilter/ipset/ip_set_core.c kfree_skb(skb2); skb2 1321 net/netfilter/nf_conntrack_netlink.c struct sk_buff *skb2 = NULL; skb2 1361 net/netfilter/nf_conntrack_netlink.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 1362 net/netfilter/nf_conntrack_netlink.c if (skb2 == NULL) { skb2 1368 net/netfilter/nf_conntrack_netlink.c err = ctnetlink_fill_info(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, skb2 1375 net/netfilter/nf_conntrack_netlink.c err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 1382 net/netfilter/nf_conntrack_netlink.c kfree_skb(skb2); skb2 2321 net/netfilter/nf_conntrack_netlink.c struct sk_buff *skb2; skb2 2324 net/netfilter/nf_conntrack_netlink.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 2325 net/netfilter/nf_conntrack_netlink.c if (skb2 == NULL) skb2 2328 net/netfilter/nf_conntrack_netlink.c err = ctnetlink_stat_ct_fill_info(skb2, NETLINK_CB(skb).portid, skb2 2335 net/netfilter/nf_conntrack_netlink.c err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 2342 net/netfilter/nf_conntrack_netlink.c kfree_skb(skb2); skb2 3038 net/netfilter/nf_conntrack_netlink.c struct sk_buff *skb2; skb2 3087 net/netfilter/nf_conntrack_netlink.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 3088 net/netfilter/nf_conntrack_netlink.c if (skb2 == NULL) { skb2 3094 net/netfilter/nf_conntrack_netlink.c err = ctnetlink_exp_fill_info(skb2, NETLINK_CB(skb).portid, skb2 3101 net/netfilter/nf_conntrack_netlink.c err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 3108 net/netfilter/nf_conntrack_netlink.c kfree_skb(skb2); skb2 720 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 739 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 740 net/netfilter/nf_tables_api.c if (!skb2) skb2 743 net/netfilter/nf_tables_api.c err = nf_tables_fill_table_info(skb2, net, NETLINK_CB(skb).portid, skb2 749 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 752 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 1413 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 1438 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 1439 net/netfilter/nf_tables_api.c if (!skb2) skb2 1442 net/netfilter/nf_tables_api.c err = nf_tables_fill_chain_info(skb2, net, NETLINK_CB(skb).portid, skb2 1448 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 1451 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 2583 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 2617 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 2618 net/netfilter/nf_tables_api.c if (!skb2) skb2 2621 net/netfilter/nf_tables_api.c err = nf_tables_fill_rule_info(skb2, net, NETLINK_CB(skb).portid, skb2 2627 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 2630 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 3490 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 3522 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 3523 net/netfilter/nf_tables_api.c if (skb2 == NULL) skb2 3526 net/netfilter/nf_tables_api.c err = nf_tables_fill_set(skb2, &ctx, set, NFT_MSG_NEWSET, 0); skb2 3530 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 3533 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 5456 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 5490 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 5491 net/netfilter/nf_tables_api.c if (!skb2) skb2 5497 net/netfilter/nf_tables_api.c err = nf_tables_fill_obj_info(skb2, net, NETLINK_CB(skb).portid, skb2 5503 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 5505 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 6139 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 6167 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 6168 net/netfilter/nf_tables_api.c if (!skb2) skb2 6171 net/netfilter/nf_tables_api.c err = nf_tables_fill_flowtable_info(skb2, net, NETLINK_CB(skb).portid, skb2 6178 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 6180 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 6299 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 6306 net/netfilter/nf_tables_api.c skb2 = nlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL); skb2 6307 net/netfilter/nf_tables_api.c if (skb2 == NULL) skb2 6310 net/netfilter/nf_tables_api.c err = nf_tables_fill_gen_info(skb2, net, NETLINK_CB(skb).portid, skb2 6313 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 6317 net/netfilter/nf_tables_api.c nfnetlink_send(skb2, net, NETLINK_CB(skb).portid, NFNLGRP_NFTABLES, skb2 6330 net/netfilter/nf_tables_api.c struct sk_buff *skb2; skb2 6333 net/netfilter/nf_tables_api.c skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC); skb2 6334 net/netfilter/nf_tables_api.c if (skb2 == NULL) skb2 6337 net/netfilter/nf_tables_api.c err = nf_tables_fill_gen_info(skb2, net, NETLINK_CB(skb).portid, skb2 6342 net/netfilter/nf_tables_api.c return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid); skb2 6344 net/netfilter/nf_tables_api.c kfree_skb(skb2); skb2 292 net/netfilter/nfnetlink_acct.c struct sk_buff *skb2; skb2 297 net/netfilter/nfnetlink_acct.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 298 net/netfilter/nfnetlink_acct.c if (skb2 == NULL) { skb2 303 net/netfilter/nfnetlink_acct.c ret = nfnl_acct_fill_info(skb2, NETLINK_CB(skb).portid, skb2 308 net/netfilter/nfnetlink_acct.c kfree_skb(skb2); skb2 311 net/netfilter/nfnetlink_acct.c ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, skb2 622 net/netfilter/nfnetlink_cthelper.c struct sk_buff *skb2; skb2 660 net/netfilter/nfnetlink_cthelper.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 661 net/netfilter/nfnetlink_cthelper.c if (skb2 == NULL) { skb2 666 net/netfilter/nfnetlink_cthelper.c ret = nfnl_cthelper_fill_info(skb2, NETLINK_CB(skb).portid, skb2 671 net/netfilter/nfnetlink_cthelper.c kfree_skb(skb2); skb2 675 net/netfilter/nfnetlink_cthelper.c ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, skb2 263 net/netfilter/nfnetlink_cttimeout.c struct sk_buff *skb2; skb2 268 net/netfilter/nfnetlink_cttimeout.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 269 net/netfilter/nfnetlink_cttimeout.c if (skb2 == NULL) { skb2 274 net/netfilter/nfnetlink_cttimeout.c ret = ctnl_timeout_fill_info(skb2, NETLINK_CB(skb).portid, skb2 279 net/netfilter/nfnetlink_cttimeout.c kfree_skb(skb2); skb2 282 net/netfilter/nfnetlink_cttimeout.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, skb2 431 net/netfilter/nfnetlink_cttimeout.c struct sk_buff *skb2; skb2 487 net/netfilter/nfnetlink_cttimeout.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 488 net/netfilter/nfnetlink_cttimeout.c if (skb2 == NULL) { skb2 493 net/netfilter/nfnetlink_cttimeout.c ret = cttimeout_default_fill_info(net, skb2, NETLINK_CB(skb).portid, skb2 499 net/netfilter/nfnetlink_cttimeout.c kfree_skb(skb2); skb2 503 net/netfilter/nfnetlink_cttimeout.c ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT); skb2 612 net/netfilter/nft_compat.c struct sk_buff *skb2; skb2 654 net/netfilter/nft_compat.c skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); skb2 655 net/netfilter/nft_compat.c if (skb2 == NULL) { skb2 661 net/netfilter/nft_compat.c if (nfnl_compat_fill_info(skb2, NETLINK_CB(skb).portid, skb2 667 net/netfilter/nft_compat.c kfree_skb(skb2); skb2 671 net/netfilter/nft_compat.c ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid, skb2 1398 net/netlink/af_netlink.c struct sk_buff *skb, *skb2; skb2 1434 net/netlink/af_netlink.c if (p->skb2 == NULL) { skb2 1436 net/netlink/af_netlink.c p->skb2 = skb_clone(p->skb, p->allocation); skb2 1438 net/netlink/af_netlink.c p->skb2 = skb_get(p->skb); skb2 1443 net/netlink/af_netlink.c skb_orphan(p->skb2); skb2 1446 net/netlink/af_netlink.c if (p->skb2 == NULL) { skb2 1454 net/netlink/af_netlink.c if (p->tx_filter && p->tx_filter(sk, p->skb2, p->tx_data)) { skb2 1455 net/netlink/af_netlink.c kfree_skb(p->skb2); skb2 1456 net/netlink/af_netlink.c p->skb2 = NULL; skb2 1459 net/netlink/af_netlink.c if (sk_filter(sk, p->skb2)) { skb2 1460 net/netlink/af_netlink.c kfree_skb(p->skb2); skb2 1461 net/netlink/af_netlink.c p->skb2 = NULL; skb2 1464 net/netlink/af_netlink.c NETLINK_CB(p->skb2).nsid = peernet2id(sock_net(sk), p->net); skb2 1465 net/netlink/af_netlink.c if (NETLINK_CB(p->skb2).nsid != NETNSA_NSID_NOT_ASSIGNED) skb2 1466 net/netlink/af_netlink.c NETLINK_CB(p->skb2).nsid_is_set = true; skb2 1467 net/netlink/af_netlink.c val = netlink_broadcast_deliver(sk, p->skb2); skb2 1475 net/netlink/af_netlink.c p->skb2 = NULL; skb2 1502 net/netlink/af_netlink.c info.skb2 = NULL; skb2 1518 net/netlink/af_netlink.c kfree_skb(info.skb2); skb2 1521 net/netlink/af_netlink.c consume_skb(info.skb2); skb2 214 net/sched/act_mirred.c struct sk_buff *skb2 = skb; skb2 261 net/sched/act_mirred.c skb2 = skb_clone(skb, GFP_ATOMIC); skb2 262 net/sched/act_mirred.c if (!skb2) skb2 275 net/sched/act_mirred.c skb_pull_rcsum(skb2, mac_len); skb2 278 net/sched/act_mirred.c skb_push_rcsum(skb2, mac_len); skb2 282 net/sched/act_mirred.c skb2->skb_iif = skb->dev->ifindex; skb2 283 net/sched/act_mirred.c skb2->dev = dev; skb2 287 net/sched/act_mirred.c skb_set_redirected(skb2, skb2->tc_at_ingress); skb2 300 net/sched/act_mirred.c err = dev_queue_xmit(skb2); skb2 302 net/sched/act_mirred.c err = netif_receive_skb(skb2); skb2 156 net/sched/sch_choke.c struct sk_buff *skb2) skb2 160 net/sched/sch_choke.c if (skb1->protocol != skb2->protocol) skb2 169 net/sched/sch_choke.c if (!choke_skb_cb(skb2)->keys_valid) { skb2 170 net/sched/sch_choke.c choke_skb_cb(skb2)->keys_valid = 1; skb2 171 net/sched/sch_choke.c skb_flow_dissect_flow_keys(skb2, &temp, 0); skb2 172 net/sched/sch_choke.c make_flow_keys_digest(&choke_skb_cb(skb2)->keys, &temp); skb2 176 net/sched/sch_choke.c &choke_skb_cb(skb2)->keys, skb2 440 net/sched/sch_netem.c struct sk_buff *skb2; skb2 478 net/sched/sch_netem.c if (count > 1 && (skb2 = skb_clone(skb, GFP_ATOMIC)) != NULL) { skb2 483 net/sched/sch_netem.c rootq->enqueue(skb2, rootq, to_free); skb2 603 net/sched/sch_netem.c skb2 = segs->next; skb2 615 net/sched/sch_netem.c segs = skb2; skb2 79 net/xfrm/xfrm_device.c struct sk_buff *skb2; skb2 149 net/xfrm/xfrm_device.c skb2 = skb; skb2 152 net/xfrm/xfrm_device.c struct sk_buff *nskb = skb2->next; skb2 155 net/xfrm/xfrm_device.c skb_mark_not_on_list(skb2); skb2 157 net/xfrm/xfrm_device.c xo = xfrm_offload(skb2); skb2 160 net/xfrm/xfrm_device.c xfrm_outer_mode_prep(x, skb2); skb2 162 net/xfrm/xfrm_device.c err = x->type_offload->xmit(x, skb2, esp_features); skb2 164 net/xfrm/xfrm_device.c skb2->next = nskb; skb2 167 net/xfrm/xfrm_device.c skb2->next = nskb; skb2 168 net/xfrm/xfrm_device.c kfree_skb_list(skb2); skb2 171 net/xfrm/xfrm_device.c if (skb == skb2) skb2 180 net/xfrm/xfrm_device.c skb_push(skb2, skb2->data - skb_mac_header(skb2)); skb2 183 net/xfrm/xfrm_device.c skb2 = nskb; skb2 184 net/xfrm/xfrm_device.c } while (skb2);