skb2             4129 drivers/net/bonding/bond_main.c 			struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             4131 drivers/net/bonding/bond_main.c 			if (!skb2) {
skb2             4136 drivers/net/bonding/bond_main.c 			bond_dev_queue_xmit(bond, skb2, slave->dev);
skb2              393 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 	struct sk_buff *skb2;
skb2              428 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2);
skb2              432 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			skb_put(skb2,  re2.s.len);
skb2              433 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			skb_new = skb_copy_expand(skb, 0, skb2->len,
skb2              437 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			if (skb_copy_bits(skb2, 0, skb_tail_pointer(skb_new),
skb2              438 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 					  skb2->len))
skb2              440 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			skb_put(skb_new, skb2->len);
skb2              442 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 			dev_kfree_skb_any(skb2);
skb2              457 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 	dev_kfree_skb_any(skb2);
skb2              459 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 		re2.d64 = octeon_mgmt_dequeue_rx_buffer(p, &skb2);
skb2              460 drivers/net/ethernet/cavium/octeon/octeon_mgmt.c 		dev_kfree_skb_any(skb2);
skb2             2046 drivers/net/ethernet/qlogic/qla3xxx.c 	struct sk_buff *skb1 = NULL, *skb2;
skb2             2068 drivers/net/ethernet/qlogic/qla3xxx.c 	skb2 = lrg_buf_cb2->skb;
skb2             2070 drivers/net/ethernet/qlogic/qla3xxx.c 	skb_put(skb2, length);	/* Just the second buffer length here. */
skb2             2075 drivers/net/ethernet/qlogic/qla3xxx.c 	prefetch(skb2->data);
skb2             2077 drivers/net/ethernet/qlogic/qla3xxx.c 	skb_checksum_none_assert(skb2);
skb2             2084 drivers/net/ethernet/qlogic/qla3xxx.c 						 skb_push(skb2, size), size);
skb2             2098 drivers/net/ethernet/qlogic/qla3xxx.c 			skb2->ip_summed = CHECKSUM_UNNECESSARY;
skb2             2101 drivers/net/ethernet/qlogic/qla3xxx.c 	skb2->protocol = eth_type_trans(skb2, qdev->ndev);
skb2             2103 drivers/net/ethernet/qlogic/qla3xxx.c 	napi_gro_receive(&qdev->napi, skb2);
skb2             1195 drivers/net/ethernet/tehuti/tehuti.c 	struct sk_buff *skb, *skb2;
skb2             1259 drivers/net/ethernet/tehuti/tehuti.c 		    (skb2 = netdev_alloc_skb(priv->ndev, len + NET_IP_ALIGN))) {
skb2             1260 drivers/net/ethernet/tehuti/tehuti.c 			skb_reserve(skb2, NET_IP_ALIGN);
skb2             1265 drivers/net/ethernet/tehuti/tehuti.c 			memcpy(skb2->data, skb->data, len);
skb2             1267 drivers/net/ethernet/tehuti/tehuti.c 			skb = skb2;
skb2             1213 drivers/net/gtp.c 	struct sk_buff *skb2;
skb2             1227 drivers/net/gtp.c 	skb2 = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             1228 drivers/net/gtp.c 	if (skb2 == NULL) {
skb2             1233 drivers/net/gtp.c 	err = gtp_genl_fill_info(skb2, NETLINK_CB(skb).portid, info->snd_seq,
skb2             1239 drivers/net/gtp.c 	return genlmsg_unicast(genl_info_net(info), skb2, info->snd_portid);
skb2             1242 drivers/net/gtp.c 	kfree_skb(skb2);
skb2               19 drivers/net/team/team_mode_broadcast.c 	struct sk_buff *skb2;
skb2               26 drivers/net/team/team_mode_broadcast.c 				skb2 = skb_clone(skb, GFP_ATOMIC);
skb2               27 drivers/net/team/team_mode_broadcast.c 				if (skb2) {
skb2               29 drivers/net/team/team_mode_broadcast.c 								   skb2);
skb2              250 drivers/net/usb/asix_common.c 		struct sk_buff *skb2;
skb2              252 drivers/net/usb/asix_common.c 		skb2 = skb_copy_expand(skb, 4, padlen, flags);
skb2              254 drivers/net/usb/asix_common.c 		skb = skb2;
skb2               92 drivers/net/usb/cdc_eem.c 	struct sk_buff	*skb2 = NULL;
skb2              125 drivers/net/usb/cdc_eem.c 	skb2 = skb_copy_expand(skb, EEM_HEAD, ETH_FCS_LEN + padlen, flags);
skb2              126 drivers/net/usb/cdc_eem.c 	if (!skb2)
skb2              130 drivers/net/usb/cdc_eem.c 	skb = skb2;
skb2              168 drivers/net/usb/cdc_eem.c 		struct sk_buff	*skb2 = NULL;
skb2              217 drivers/net/usb/cdc_eem.c 				skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              218 drivers/net/usb/cdc_eem.c 				if (unlikely(!skb2))
skb2              220 drivers/net/usb/cdc_eem.c 				skb_trim(skb2, len);
skb2              222 drivers/net/usb/cdc_eem.c 						skb_push(skb2, 2));
skb2              223 drivers/net/usb/cdc_eem.c 				eem_linkcmd(dev, skb2);
skb2              288 drivers/net/usb/cdc_eem.c 				skb2 = skb;
skb2              290 drivers/net/usb/cdc_eem.c 				skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              291 drivers/net/usb/cdc_eem.c 				if (unlikely(!skb2))
skb2              302 drivers/net/usb/cdc_eem.c 				crc = get_unaligned_le32(skb2->data
skb2              304 drivers/net/usb/cdc_eem.c 				crc2 = ~crc32_le(~0, skb2->data, skb2->len
skb2              307 drivers/net/usb/cdc_eem.c 				crc = get_unaligned_be32(skb2->data
skb2              311 drivers/net/usb/cdc_eem.c 			skb_trim(skb2, len - ETH_FCS_LEN);
skb2              318 drivers/net/usb/cdc_eem.c 				dev_kfree_skb_any(skb2);
skb2              320 drivers/net/usb/cdc_eem.c 				usbnet_skb_return(dev, skb2);
skb2              213 drivers/net/usb/cx82310_eth.c 	struct sk_buff *skb2;
skb2              221 drivers/net/usb/cx82310_eth.c 		skb2 = alloc_skb(len, GFP_ATOMIC);
skb2              222 drivers/net/usb/cx82310_eth.c 		if (!skb2)
skb2              224 drivers/net/usb/cx82310_eth.c 		skb_put(skb2, len);
skb2              225 drivers/net/usb/cx82310_eth.c 		memcpy(skb2->data, (void *)dev->partial_data,
skb2              227 drivers/net/usb/cx82310_eth.c 		memcpy(skb2->data + dev->partial_len, skb->data,
skb2              229 drivers/net/usb/cx82310_eth.c 		usbnet_skb_return(dev, skb2);
skb2              264 drivers/net/usb/cx82310_eth.c 		skb2 = alloc_skb(len, GFP_ATOMIC);
skb2              265 drivers/net/usb/cx82310_eth.c 		if (!skb2)
skb2              267 drivers/net/usb/cx82310_eth.c 		skb_put(skb2, len);
skb2              268 drivers/net/usb/cx82310_eth.c 		memcpy(skb2->data, skb->data, len);
skb2              270 drivers/net/usb/cx82310_eth.c 		usbnet_skb_return(dev, skb2);
skb2              497 drivers/net/usb/dm9601.c 		struct sk_buff *skb2;
skb2              499 drivers/net/usb/dm9601.c 		skb2 = skb_copy_expand(skb, DM_TX_OVERHEAD, pad, flags);
skb2              501 drivers/net/usb/dm9601.c 		skb = skb2;
skb2              156 drivers/net/usb/gl620a.c 		struct sk_buff	*skb2;
skb2              157 drivers/net/usb/gl620a.c 		skb2 = skb_copy_expand(skb, (4 + 4*1) , padlen, flags);
skb2              159 drivers/net/usb/gl620a.c 		skb = skb2;
skb2               83 drivers/net/usb/int51x1.c 		struct sk_buff *skb2;
skb2               85 drivers/net/usb/int51x1.c 		skb2 = skb_copy_expand(skb,
skb2               90 drivers/net/usb/int51x1.c 		if (!skb2)
skb2               92 drivers/net/usb/int51x1.c 		skb = skb2;
skb2              160 drivers/net/usb/kalmia.c 	struct sk_buff *skb2 = NULL;
skb2              183 drivers/net/usb/kalmia.c 	skb2 = skb_copy_expand(skb, KALMIA_HEADER_LENGTH,
skb2              185 drivers/net/usb/kalmia.c 	if (!skb2)
skb2              189 drivers/net/usb/kalmia.c 	skb = skb2;
skb2              242 drivers/net/usb/kalmia.c 		struct sk_buff *skb2 = NULL;
skb2              300 drivers/net/usb/kalmia.c 			skb2 = skb;
skb2              303 drivers/net/usb/kalmia.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              304 drivers/net/usb/kalmia.c 			if (unlikely(!skb2))
skb2              308 drivers/net/usb/kalmia.c 		skb_trim(skb2, ether_packet_length);
skb2              314 drivers/net/usb/kalmia.c 			usbnet_skb_return(dev, skb2);
skb2             2845 drivers/net/usb/lan78xx.c 	struct sk_buff *skb2 = NULL;
skb2             2849 drivers/net/usb/lan78xx.c 		skb2 = lan78xx_tx_prep(dev, skb, GFP_ATOMIC);
skb2             2852 drivers/net/usb/lan78xx.c 	if (skb2) {
skb2             2853 drivers/net/usb/lan78xx.c 		skb_queue_tail(&dev->txq_pend, skb2);
skb2             3103 drivers/net/usb/lan78xx.c 		struct sk_buff *skb2;
skb2             3138 drivers/net/usb/lan78xx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             3139 drivers/net/usb/lan78xx.c 			if (unlikely(!skb2)) {
skb2             3144 drivers/net/usb/lan78xx.c 			skb2->len = size;
skb2             3145 drivers/net/usb/lan78xx.c 			skb2->data = packet;
skb2             3146 drivers/net/usb/lan78xx.c 			skb_set_tail_pointer(skb2, size);
skb2             3148 drivers/net/usb/lan78xx.c 			lan78xx_rx_csum_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
skb2             3149 drivers/net/usb/lan78xx.c 			lan78xx_rx_vlan_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
skb2             3151 drivers/net/usb/lan78xx.c 			skb_trim(skb2, skb2->len - 4); /* remove fcs */
skb2             3152 drivers/net/usb/lan78xx.c 			skb2->truesize = size + sizeof(struct sk_buff);
skb2             3154 drivers/net/usb/lan78xx.c 			lan78xx_skb_return(dev, skb2);
skb2             3324 drivers/net/usb/lan78xx.c 	struct sk_buff *skb, *skb2;
skb2             3362 drivers/net/usb/lan78xx.c 		skb2 = skb_dequeue(tqp);
skb2             3363 drivers/net/usb/lan78xx.c 		if (skb2) {
skb2             3364 drivers/net/usb/lan78xx.c 			length += (skb2->len - TX_OVERHEAD);
skb2             3365 drivers/net/usb/lan78xx.c 			memcpy(skb->data + pos, skb2->data, skb2->len);
skb2             3366 drivers/net/usb/lan78xx.c 			pos += roundup(skb2->len, sizeof(u32));
skb2             3367 drivers/net/usb/lan78xx.c 			dev_kfree_skb(skb2);
skb2              417 drivers/net/usb/net1080.c 	struct sk_buff		*skb2;
skb2              446 drivers/net/usb/net1080.c 	skb2 = skb_copy_expand(skb,
skb2              451 drivers/net/usb/net1080.c 	if (!skb2)
skb2              452 drivers/net/usb/net1080.c 		return skb2;
skb2              453 drivers/net/usb/net1080.c 	skb = skb2;
skb2              495 drivers/net/usb/rndis_host.c 		struct sk_buff		*skb2;
skb2              521 drivers/net/usb/rndis_host.c 		skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              522 drivers/net/usb/rndis_host.c 		if (unlikely(!skb2))
skb2              525 drivers/net/usb/rndis_host.c 		skb_trim(skb2, data_len);
skb2              526 drivers/net/usb/rndis_host.c 		usbnet_skb_return(dev, skb2);
skb2              538 drivers/net/usb/rndis_host.c 	struct sk_buff		*skb2;
skb2              559 drivers/net/usb/rndis_host.c 	skb2 = skb_copy_expand(skb, sizeof *hdr, 1, flags);
skb2              561 drivers/net/usb/rndis_host.c 	if (unlikely(!skb2))
skb2              562 drivers/net/usb/rndis_host.c 		return skb2;
skb2              563 drivers/net/usb/rndis_host.c 	skb = skb2;
skb2              129 drivers/net/usb/sr9800.c 		struct sk_buff *skb2;
skb2              130 drivers/net/usb/sr9800.c 		skb2 = skb_copy_expand(skb, 4, padlen, flags);
skb2              132 drivers/net/usb/sr9800.c 		skb = skb2;
skb2               48 drivers/net/usb/zaurus.c 	struct sk_buff		*skb2;
skb2               56 drivers/net/usb/zaurus.c 	skb2 = skb_copy_expand(skb, 0, 4 + padlen, flags);
skb2               58 drivers/net/usb/zaurus.c 	skb = skb2;
skb2              568 drivers/net/vrf.c 		struct sk_buff *skb2;
skb2              570 drivers/net/vrf.c 		skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev));
skb2              571 drivers/net/vrf.c 		if (!skb2) {
skb2              576 drivers/net/vrf.c 			skb_set_owner_w(skb2, skb->sk);
skb2              579 drivers/net/vrf.c 		skb = skb2;
skb2              307 drivers/net/wan/hdlc_fr.c 			struct sk_buff *skb2 = skb_realloc_headroom(skb,
skb2              309 drivers/net/wan/hdlc_fr.c 			if (!skb2)
skb2              312 drivers/net/wan/hdlc_fr.c 			skb = *skb_p = skb2;
skb2              795 drivers/net/wan/z85230.c 	c->skb2 = NULL;
skb2              861 drivers/net/wan/z85230.c 	c->skb2 = NULL;
skb2             1056 drivers/net/wan/z85230.c 	c->skb2 = NULL;
skb2             1663 drivers/net/wan/z85230.c 		c->skb = c->skb2;
skb2             1675 drivers/net/wan/z85230.c 		c->skb2 = dev_alloc_skb(c->mtu);
skb2             1676 drivers/net/wan/z85230.c 		if (c->skb2 == NULL)
skb2             1679 drivers/net/wan/z85230.c 			skb_put(c->skb2, c->mtu);
skb2              275 drivers/net/wan/z85230.h 	struct sk_buff	*skb2;		/* Pending buffer */
skb2             1557 drivers/net/wireless/ath/wil6210/txrx.c 	struct sk_buff *skb2;
skb2             1609 drivers/net/wireless/ath/wil6210/txrx.c 		skb2 = skb_copy(skb, GFP_ATOMIC);
skb2             1610 drivers/net/wireless/ath/wil6210/txrx.c 		if (skb2) {
skb2             1612 drivers/net/wireless/ath/wil6210/txrx.c 			wil_set_da_for_vring(wil, skb2, i);
skb2             1613 drivers/net/wireless/ath/wil6210/txrx.c 			wil_tx_ring(wil, vif, v2, skb2);
skb2             1615 drivers/net/wireless/ath/wil6210/txrx.c 			dev_kfree_skb_any(skb2);
skb2              182 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			struct sk_buff *skb2;
skb2              185 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              186 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			if (skb2)
skb2              187 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 				hostap_rx(skb2->dev, skb2, rx_stats);
skb2              340 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	struct sk_buff *skb2 = NULL;
skb2              775 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              776 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			if (skb2 == NULL)
skb2              783 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 			skb2 = skb;
skb2              788 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	if (skb2 != NULL) {
skb2              790 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		skb2->dev = dev;
skb2              791 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		skb2->protocol = htons(ETH_P_802_3);
skb2              792 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		skb_reset_mac_header(skb2);
skb2              793 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		skb_reset_network_header(skb2);
skb2              795 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		dev_queue_xmit(skb2);
skb2              481 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			struct sk_buff *skb2;
skb2              484 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              485 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			if (skb2)
skb2              486 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 				hostap_rx(skb2->dev, skb2, rx_stats);
skb2              727 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	struct sk_buff *skb2 = NULL;
skb2             1074 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             1075 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			if (skb2 == NULL)
skb2             1082 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 			skb2 = skb;
skb2             1087 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	if (skb2 != NULL) {
skb2             1089 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		skb2->dev = dev;
skb2             1090 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		skb2->protocol = cpu_to_be16(ETH_P_802_3);
skb2             1091 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		skb_reset_mac_header(skb2);
skb2             1092 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		skb_reset_network_header(skb2);
skb2             1094 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		dev_queue_xmit(skb2);
skb2              246 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			struct sk_buff *skb2;
skb2              249 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              250 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 			if (skb2)
skb2              251 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 				hostap_rx(skb2->dev, skb2, rx_stats);
skb2              147 drivers/staging/wilc1000/wilc_mon.c 	struct sk_buff *skb2;
skb2              163 drivers/staging/wilc1000/wilc_mon.c 		skb2 = dev_alloc_skb(skb->len + sizeof(*cb_hdr));
skb2              164 drivers/staging/wilc1000/wilc_mon.c 		if (!skb2)
skb2              167 drivers/staging/wilc1000/wilc_mon.c 		skb_put_data(skb2, skb->data, skb->len);
skb2              169 drivers/staging/wilc1000/wilc_mon.c 		cb_hdr = skb_push(skb2, sizeof(*cb_hdr));
skb2              181 drivers/staging/wilc1000/wilc_mon.c 		skb2->dev = dev;
skb2              182 drivers/staging/wilc1000/wilc_mon.c 		skb_reset_mac_header(skb2);
skb2              183 drivers/staging/wilc1000/wilc_mon.c 		skb2->ip_summed = CHECKSUM_UNNECESSARY;
skb2              184 drivers/staging/wilc1000/wilc_mon.c 		skb2->pkt_type = PACKET_OTHERHOST;
skb2              185 drivers/staging/wilc1000/wilc_mon.c 		skb2->protocol = htons(ETH_P_802_2);
skb2              186 drivers/staging/wilc1000/wilc_mon.c 		memset(skb2->cb, 0, sizeof(skb2->cb));
skb2              188 drivers/staging/wilc1000/wilc_mon.c 		netif_rx(skb2);
skb2              337 drivers/usb/gadget/function/f_eem.c 	struct sk_buff	*skb2 = NULL;
skb2              359 drivers/usb/gadget/function/f_eem.c 	skb2 = skb_copy_expand(skb, EEM_HLEN, ETH_FCS_LEN + padlen, GFP_ATOMIC);
skb2              361 drivers/usb/gadget/function/f_eem.c 	skb = skb2;
skb2              396 drivers/usb/gadget/function/f_eem.c 		struct sk_buff	*skb2;
skb2              436 drivers/usb/gadget/function/f_eem.c 				skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              437 drivers/usb/gadget/function/f_eem.c 				if (unlikely(!skb2)) {
skb2              441 drivers/usb/gadget/function/f_eem.c 				skb_trim(skb2, len);
skb2              443 drivers/usb/gadget/function/f_eem.c 							skb_push(skb2, 2));
skb2              444 drivers/usb/gadget/function/f_eem.c 				skb_copy_bits(skb2, 0, req->buf, skb2->len);
skb2              445 drivers/usb/gadget/function/f_eem.c 				req->length = skb2->len;
skb2              448 drivers/usb/gadget/function/f_eem.c 				req->context = skb2;
skb2              497 drivers/usb/gadget/function/f_eem.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              498 drivers/usb/gadget/function/f_eem.c 			if (unlikely(!skb2)) {
skb2              502 drivers/usb/gadget/function/f_eem.c 			skb_trim(skb2, len - ETH_FCS_LEN);
skb2              504 drivers/usb/gadget/function/f_eem.c 			skb3 = skb_copy_expand(skb2,
skb2              509 drivers/usb/gadget/function/f_eem.c 				dev_kfree_skb_any(skb2);
skb2              512 drivers/usb/gadget/function/f_eem.c 			dev_kfree_skb_any(skb2);
skb2              959 drivers/usb/gadget/function/f_ncm.c 	struct sk_buff	*skb2 = NULL;
skb2              993 drivers/usb/gadget/function/f_ncm.c 	swap(skb2, ncm->skb_tx_data);
skb2             1000 drivers/usb/gadget/function/f_ncm.c 	skb_put_zero(skb2, ndp_pad);
skb2             1003 drivers/usb/gadget/function/f_ncm.c 	skb_put_data(skb2, ncm->skb_tx_ndp->data, ncm->skb_tx_ndp->len);
skb2             1008 drivers/usb/gadget/function/f_ncm.c 	skb_put_zero(skb2, dgram_idx_len);
skb2             1010 drivers/usb/gadget/function/f_ncm.c 	return skb2;
skb2             1017 drivers/usb/gadget/function/f_ncm.c 	struct sk_buff	*skb2 = NULL;
skb2             1057 drivers/usb/gadget/function/f_ncm.c 			skb2 = package_for_tx(ncm);
skb2             1058 drivers/usb/gadget/function/f_ncm.c 			if (!skb2)
skb2             1129 drivers/usb/gadget/function/f_ncm.c 		skb2 = package_for_tx(ncm);
skb2             1130 drivers/usb/gadget/function/f_ncm.c 		if (!skb2)
skb2             1134 drivers/usb/gadget/function/f_ncm.c 	return skb2;
skb2             1184 drivers/usb/gadget/function/f_ncm.c 	struct sk_buff	*skb2;
skb2             1290 drivers/usb/gadget/function/f_ncm.c 			skb2 = netdev_alloc_skb_ip_align(ncm->netdev,
skb2             1292 drivers/usb/gadget/function/f_ncm.c 			if (skb2 == NULL)
skb2             1294 drivers/usb/gadget/function/f_ncm.c 			skb_put_data(skb2, skb->data + index,
skb2             1297 drivers/usb/gadget/function/f_ncm.c 			skb_queue_tail(list, skb2);
skb2              371 drivers/usb/gadget/function/f_rndis.c 	struct sk_buff *skb2;
skb2              376 drivers/usb/gadget/function/f_rndis.c 	skb2 = skb_realloc_headroom(skb, sizeof(struct rndis_packet_msg_type));
skb2              377 drivers/usb/gadget/function/f_rndis.c 	rndis_add_hdr(skb2);
skb2              380 drivers/usb/gadget/function/f_rndis.c 	return skb2;
skb2              255 drivers/usb/gadget/function/u_ether.c 	struct sk_buff	*skb = req->context, *skb2;
skb2              283 drivers/usb/gadget/function/u_ether.c 		skb2 = skb_dequeue(&dev->rx_frames);
skb2              284 drivers/usb/gadget/function/u_ether.c 		while (skb2) {
skb2              286 drivers/usb/gadget/function/u_ether.c 					|| ETH_HLEN > skb2->len
skb2              287 drivers/usb/gadget/function/u_ether.c 					|| skb2->len > GETHER_MAX_ETH_FRAME_LEN) {
skb2              290 drivers/usb/gadget/function/u_ether.c 				DBG(dev, "rx length %d\n", skb2->len);
skb2              291 drivers/usb/gadget/function/u_ether.c 				dev_kfree_skb_any(skb2);
skb2              294 drivers/usb/gadget/function/u_ether.c 			skb2->protocol = eth_type_trans(skb2, dev->net);
skb2              296 drivers/usb/gadget/function/u_ether.c 			dev->net->stats.rx_bytes += skb2->len;
skb2              301 drivers/usb/gadget/function/u_ether.c 			status = netif_rx(skb2);
skb2              303 drivers/usb/gadget/function/u_ether.c 			skb2 = skb_dequeue(&dev->rx_frames);
skb2             1065 include/linux/skbuff.h 	struct sk_buff	skb2;
skb2             1088 include/linux/skbuff.h 	       fclones->skb2.sk == sk;
skb2             1677 net/appletalk/ddp.c 		struct sk_buff *skb2 = skb_copy(skb, GFP_KERNEL);
skb2             1679 net/appletalk/ddp.c 		if (skb2) {
skb2             1685 net/appletalk/ddp.c 			aarp_send_ddp(dev, skb2, &usat->sat_addr, NULL);
skb2              213 net/atm/br2684.c 		struct sk_buff *skb2 = skb_realloc_headroom(skb, minheadroom);
skb2              216 net/atm/br2684.c 		if (skb2 == NULL) {
skb2              220 net/atm/br2684.c 		skb = skb2;
skb2              142 net/atm/lec.c  		struct sk_buff *skb2;
skb2              145 net/atm/lec.c  		skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC);
skb2              146 net/atm/lec.c  		if (skb2 == NULL)
skb2              148 net/atm/lec.c  		skb2->len = sizeof(struct atmlec_msg);
skb2              149 net/atm/lec.c  		mesg = (struct atmlec_msg *)skb2->data;
skb2              156 net/atm/lec.c  		atm_force_charge(priv->lecd, skb2->truesize);
skb2              158 net/atm/lec.c  		skb_queue_tail(&sk->sk_receive_queue, skb2);
skb2              207 net/atm/lec.c  	struct sk_buff *skb2;
skb2              236 net/atm/lec.c  		skb2 = skb_realloc_headroom(skb, LEC_HEADER_LEN);
skb2              237 net/atm/lec.c  		if (unlikely(!skb2)) {
skb2              242 net/atm/lec.c  		skb = skb2;
skb2              266 net/atm/lec.c  			skb2 = skb_copy_expand(skb, 0,
skb2              270 net/atm/lec.c  			if (skb2 == NULL) {
skb2              274 net/atm/lec.c  			skb = skb2;
skb2              304 net/atm/lec.c  	while (entry && (skb2 = skb_dequeue(&entry->tx_wait))) {
skb2              306 net/atm/lec.c  		lec_send(vcc, skb2);
skb2              442 net/atm/lec.c  			struct sk_buff *skb2;
skb2              447 net/atm/lec.c  			skb2 = alloc_skb(sizeof(struct atmlec_msg), GFP_ATOMIC);
skb2              448 net/atm/lec.c  			if (skb2 == NULL)
skb2              450 net/atm/lec.c  			skb2->len = sizeof(struct atmlec_msg);
skb2              451 net/atm/lec.c  			skb_copy_to_linear_data(skb2, mesg, sizeof(*mesg));
skb2              452 net/atm/lec.c  			atm_force_charge(priv->lecd, skb2->truesize);
skb2              454 net/atm/lec.c  			skb_queue_tail(&sk->sk_receive_queue, skb2);
skb2              101 net/bridge/netfilter/nf_conntrack_bridge.c 		struct sk_buff *skb2;
skb2              103 net/bridge/netfilter/nf_conntrack_bridge.c 		skb2 = ip_frag_next(skb, &state);
skb2              104 net/bridge/netfilter/nf_conntrack_bridge.c 		if (IS_ERR(skb2)) {
skb2              105 net/bridge/netfilter/nf_conntrack_bridge.c 			err = PTR_ERR(skb2);
skb2              109 net/bridge/netfilter/nf_conntrack_bridge.c 		skb2->tstamp = tstamp;
skb2              110 net/bridge/netfilter/nf_conntrack_bridge.c 		err = output(net, sk, data, skb2);
skb2              333 net/caif/cfpkt_skbuff.c 	struct sk_buff *skb2;
skb2              352 net/caif/cfpkt_skbuff.c 	skb2 = pkt_to_skb(tmppkt);
skb2              355 net/caif/cfpkt_skbuff.c 	if (skb2 == NULL)
skb2              358 net/caif/cfpkt_skbuff.c 	skb_put_data(skb2, split, len2nd);
skb2              363 net/caif/cfpkt_skbuff.c 	skb2->priority = skb->priority;
skb2              364 net/caif/cfpkt_skbuff.c 	return skb_to_pkt(skb2);
skb2             1940 net/core/dev.c 	struct sk_buff *skb2 = NULL;
skb2             1957 net/core/dev.c 			deliver_skb(skb2, pt_prev, skb->dev);
skb2             1963 net/core/dev.c 		skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             1964 net/core/dev.c 		if (!skb2)
skb2             1967 net/core/dev.c 		net_timestamp_set(skb2);
skb2             1973 net/core/dev.c 		skb_reset_mac_header(skb2);
skb2             1975 net/core/dev.c 		if (skb_network_header(skb2) < skb2->data ||
skb2             1976 net/core/dev.c 		    skb_network_header(skb2) > skb_tail_pointer(skb2)) {
skb2             1978 net/core/dev.c 					     ntohs(skb2->protocol),
skb2             1980 net/core/dev.c 			skb_reset_network_header(skb2);
skb2             1983 net/core/dev.c 		skb2->transport_header = skb2->network_header;
skb2             1984 net/core/dev.c 		skb2->pkt_type = PACKET_OUTGOING;
skb2             1994 net/core/dev.c 		if (!skb_orphan_frags_rx(skb2, GFP_ATOMIC))
skb2             1995 net/core/dev.c 			pt_prev->func(skb2, skb->dev, pt_prev, skb->dev);
skb2             1997 net/core/dev.c 			kfree_skb(skb2);
skb2              249 net/core/skbuff.c 		fclones->skb2.fclone = SKB_FCLONE_CLONE;
skb2              638 net/core/skbuff.c 		fclones = container_of(skb, struct sk_buff_fclones, skb2);
skb2             1442 net/core/skbuff.c 		n = &fclones->skb2;
skb2             1703 net/core/skbuff.c 	struct sk_buff *skb2;
skb2             1707 net/core/skbuff.c 		skb2 = pskb_copy(skb, GFP_ATOMIC);
skb2             1709 net/core/skbuff.c 		skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             1710 net/core/skbuff.c 		if (skb2 && pskb_expand_head(skb2, SKB_DATA_ALIGN(delta), 0,
skb2             1712 net/core/skbuff.c 			kfree_skb(skb2);
skb2             1713 net/core/skbuff.c 			skb2 = NULL;
skb2             1716 net/core/skbuff.c 	return skb2;
skb2             4363 net/core/skbuff.c 			struct sk_buff *skb2;
skb2             4367 net/core/skbuff.c 				skb2 = skb_copy(skb1, GFP_ATOMIC);
skb2             4369 net/core/skbuff.c 				skb2 = skb_copy_expand(skb1,
skb2             4373 net/core/skbuff.c 			if (unlikely(skb2 == NULL))
skb2             4377 net/core/skbuff.c 				skb_set_owner_w(skb2, skb1->sk);
skb2             4382 net/core/skbuff.c 			skb2->next = skb1->next;
skb2             4383 net/core/skbuff.c 			*skb_p = skb2;
skb2             4385 net/core/skbuff.c 			skb1 = skb2;
skb2              969 net/decnet/dn_dev.c 		struct sk_buff *skb2 = skb_copy(skb, GFP_ATOMIC);
skb2              970 net/decnet/dn_dev.c 		if (skb2) {
skb2              971 net/decnet/dn_dev.c 			dn_rt_finish_output(skb2, dn_rt_all_end_mcast, src);
skb2              221 net/decnet/dn_neigh.c 		struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom);
skb2              222 net/decnet/dn_neigh.c 		if (skb2 == NULL) {
skb2              228 net/decnet/dn_neigh.c 		skb = skb2;
skb2              269 net/decnet/dn_neigh.c 		struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom);
skb2              270 net/decnet/dn_neigh.c 		if (skb2 == NULL) {
skb2              276 net/decnet/dn_neigh.c 		skb = skb2;
skb2              311 net/decnet/dn_neigh.c 		struct sk_buff *skb2 = skb_realloc_headroom(skb, headroom);
skb2              312 net/decnet/dn_neigh.c 		if (skb2 == NULL) {
skb2              318 net/decnet/dn_neigh.c 		skb = skb2;
skb2              207 net/decnet/dn_nsp_out.c 	struct sk_buff *skb2;
skb2              210 net/decnet/dn_nsp_out.c 	if ((skb2 = skb_clone(skb, gfp)) != NULL) {
skb2              214 net/decnet/dn_nsp_out.c 		skb2->sk = skb->sk;
skb2              215 net/decnet/dn_nsp_out.c 		dn_nsp_send(skb2);
skb2              373 net/decnet/dn_nsp_out.c 	struct sk_buff *skb2, *n, *ack = NULL;
skb2              381 net/decnet/dn_nsp_out.c 	skb_queue_walk_safe(q, skb2, n) {
skb2              382 net/decnet/dn_nsp_out.c 		struct dn_skb_cb *cb2 = DN_SKB_CB(skb2);
skb2              385 net/decnet/dn_nsp_out.c 			ack = skb2;
skb2               64 net/decnet/netfilter/dn_rtmsg.c 	struct sk_buff *skb2;
skb2               80 net/decnet/netfilter/dn_rtmsg.c 	skb2 = dnrmg_build_message(skb, &status);
skb2               81 net/decnet/netfilter/dn_rtmsg.c 	if (skb2 == NULL)
skb2               83 net/decnet/netfilter/dn_rtmsg.c 	NETLINK_CB(skb2).dst_group = group;
skb2               84 net/decnet/netfilter/dn_rtmsg.c 	netlink_broadcast(dnrmg, skb2, 0, group, GFP_ATOMIC);
skb2              170 net/ipv4/ip_input.c 				struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              171 net/ipv4/ip_input.c 				if (skb2)
skb2              172 net/ipv4/ip_input.c 					raw_rcv(last, skb2);
skb2              201 net/ipv4/ip_output.c 		struct sk_buff *skb2;
skb2              203 net/ipv4/ip_output.c 		skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev));
skb2              204 net/ipv4/ip_output.c 		if (!skb2) {
skb2              209 net/ipv4/ip_output.c 			skb_set_owner_w(skb2, skb->sk);
skb2              211 net/ipv4/ip_output.c 		skb = skb2;
skb2              685 net/ipv4/ip_output.c 	struct sk_buff *skb2;
skb2              699 net/ipv4/ip_output.c 	skb2 = alloc_skb(len + state->hlen + state->ll_rs, GFP_ATOMIC);
skb2              700 net/ipv4/ip_output.c 	if (!skb2)
skb2              707 net/ipv4/ip_output.c 	ip_copy_metadata(skb2, skb);
skb2              708 net/ipv4/ip_output.c 	skb_reserve(skb2, state->ll_rs);
skb2              709 net/ipv4/ip_output.c 	skb_put(skb2, len + state->hlen);
skb2              710 net/ipv4/ip_output.c 	skb_reset_network_header(skb2);
skb2              711 net/ipv4/ip_output.c 	skb2->transport_header = skb2->network_header + state->hlen;
skb2              719 net/ipv4/ip_output.c 		skb_set_owner_w(skb2, skb->sk);
skb2              725 net/ipv4/ip_output.c 	skb_copy_from_linear_data(skb, skb_network_header(skb2), state->hlen);
skb2              730 net/ipv4/ip_output.c 	if (skb_copy_bits(skb, state->ptr, skb_transport_header(skb2), len))
skb2              737 net/ipv4/ip_output.c 	iph = ip_hdr(skb2);
skb2              755 net/ipv4/ip_output.c 	return skb2;
skb2              770 net/ipv4/ip_output.c 	struct sk_buff *skb2;
skb2              896 net/ipv4/ip_output.c 		skb2 = ip_frag_next(skb, &state);
skb2              897 net/ipv4/ip_output.c 		if (IS_ERR(skb2)) {
skb2              898 net/ipv4/ip_output.c 			err = PTR_ERR(skb2);
skb2              901 net/ipv4/ip_output.c 		ip_frag_ipcb(skb, skb2, first_frag, &state);
skb2              906 net/ipv4/ip_output.c 		skb2->tstamp = tstamp;
skb2              907 net/ipv4/ip_output.c 		err = output(net, sk, skb2);
skb2             2041 net/ipv4/ipmr.c 				struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             2043 net/ipv4/ipmr.c 				if (skb2)
skb2             2045 net/ipv4/ipmr.c 							skb2, psend);
skb2             2053 net/ipv4/ipmr.c 			struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             2055 net/ipv4/ipmr.c 			if (skb2)
skb2             2056 net/ipv4/ipmr.c 				ipmr_queue_xmit(net, mrt, true_vifi, skb2,
skb2             2165 net/ipv4/ipmr.c 			struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             2167 net/ipv4/ipmr.c 			if (!skb2)
skb2             2169 net/ipv4/ipmr.c 			skb = skb2;
skb2             2278 net/ipv4/ipmr.c 		struct sk_buff *skb2;
skb2             2293 net/ipv4/ipmr.c 		skb2 = skb_realloc_headroom(skb, sizeof(struct iphdr));
skb2             2294 net/ipv4/ipmr.c 		if (!skb2) {
skb2             2300 net/ipv4/ipmr.c 		NETLINK_CB(skb2).portid = portid;
skb2             2301 net/ipv4/ipmr.c 		skb_push(skb2, sizeof(struct iphdr));
skb2             2302 net/ipv4/ipmr.c 		skb_reset_network_header(skb2);
skb2             2303 net/ipv4/ipmr.c 		iph = ip_hdr(skb2);
skb2             2308 net/ipv4/ipmr.c 		err = ipmr_cache_unresolved(mrt, vif, skb2, dev);
skb2              974 net/ipv4/ping.c 		struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              977 net/ipv4/ping.c 		if (skb2)
skb2              978 net/ipv4/ping.c 			ping_queue_rcv_skb(sk, skb2);
skb2             1263 net/ipv4/tcp_output.c static void tcp_fragment_tstamp(struct sk_buff *skb, struct sk_buff *skb2)
skb2             1268 net/ipv4/tcp_output.c 	    !before(shinfo->tskey, TCP_SKB_CB(skb2)->seq)) {
skb2             1269 net/ipv4/tcp_output.c 		struct skb_shared_info *shinfo2 = skb_shinfo(skb2);
skb2             1275 net/ipv4/tcp_output.c 		TCP_SKB_CB(skb2)->txstamp_ack = TCP_SKB_CB(skb)->txstamp_ack;
skb2             1280 net/ipv4/tcp_output.c static void tcp_skb_fragment_eor(struct sk_buff *skb, struct sk_buff *skb2)
skb2             1282 net/ipv4/tcp_output.c 	TCP_SKB_CB(skb2)->eor = TCP_SKB_CB(skb)->eor;
skb2              617 net/ipv6/icmp.c 	struct sk_buff *skb2;
skb2              627 net/ipv6/icmp.c 	skb2 = data_len ? skb_copy(skb, GFP_ATOMIC) : skb_clone(skb, GFP_ATOMIC);
skb2              629 net/ipv6/icmp.c 	if (!skb2)
skb2              632 net/ipv6/icmp.c 	skb_dst_drop(skb2);
skb2              633 net/ipv6/icmp.c 	skb_pull(skb2, nhs);
skb2              634 net/ipv6/icmp.c 	skb_reset_network_header(skb2);
skb2              636 net/ipv6/icmp.c 	rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr, NULL, 0,
skb2              640 net/ipv6/icmp.c 		skb2->dev = rt->dst.dev;
skb2              648 net/ipv6/icmp.c 		__skb_push(skb2, nhs);
skb2              649 net/ipv6/icmp.c 		skb_reset_network_header(skb2);
skb2              650 net/ipv6/icmp.c 		memmove(skb2->data, skb2->data + nhs, data_len - nhs);
skb2              651 net/ipv6/icmp.c 		memset(skb2->data + data_len - nhs, 0, nhs);
skb2              658 net/ipv6/icmp.c 		icmp6_send(skb2, ICMPV6_TIME_EXCEED, ICMPV6_EXC_HOPLIMIT,
skb2              661 net/ipv6/icmp.c 		icmp6_send(skb2, ICMPV6_DEST_UNREACH, ICMPV6_ADDR_UNREACH,
skb2              666 net/ipv6/icmp.c 	kfree_skb(skb2);
skb2              506 net/ipv6/ip6_input.c 		struct sk_buff *skb2;
skb2              540 net/ipv6/ip6_input.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              542 net/ipv6/ip6_input.c 			skb2 = skb;
skb2              546 net/ipv6/ip6_input.c 		if (skb2) {
skb2              547 net/ipv6/ip6_input.c 			ip6_mr_input(skb2);
skb2              214 net/ipv6/ip6_output.c 		struct sk_buff *skb2 = skb_realloc_headroom(skb, head_room);
skb2              215 net/ipv6/ip6_output.c 		if (!skb2) {
skb2              222 net/ipv6/ip6_output.c 			skb_set_owner_w(skb2, skb->sk);
skb2              224 net/ipv6/ip6_output.c 		skb = skb2;
skb2              314 net/ipv6/ip6_output.c 				struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              315 net/ipv6/ip6_output.c 				if (skb2)
skb2              316 net/ipv6/ip6_output.c 					rawv6_rcv(last, skb2);
skb2              566 net/ipv6/ip6_tunnel.c 	struct sk_buff *skb2;
skb2              601 net/ipv6/ip6_tunnel.c 	skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              602 net/ipv6/ip6_tunnel.c 	if (!skb2)
skb2              605 net/ipv6/ip6_tunnel.c 	skb_dst_drop(skb2);
skb2              607 net/ipv6/ip6_tunnel.c 	skb_pull(skb2, offset);
skb2              608 net/ipv6/ip6_tunnel.c 	skb_reset_network_header(skb2);
skb2              609 net/ipv6/ip6_tunnel.c 	eiph = ip_hdr(skb2);
skb2              617 net/ipv6/ip6_tunnel.c 	skb2->dev = rt->dst.dev;
skb2              630 net/ipv6/ip6_tunnel.c 		skb_dst_set(skb2, &rt->dst);
skb2              632 net/ipv6/ip6_tunnel.c 		if (ip_route_input(skb2, eiph->daddr, eiph->saddr, eiph->tos,
skb2              633 net/ipv6/ip6_tunnel.c 				   skb2->dev) ||
skb2              634 net/ipv6/ip6_tunnel.c 		    skb_dst(skb2)->dev->type != ARPHRD_TUNNEL6)
skb2              640 net/ipv6/ip6_tunnel.c 		if (rel_info > dst_mtu(skb_dst(skb2)))
skb2              643 net/ipv6/ip6_tunnel.c 		skb_dst_update_pmtu_no_confirm(skb2, rel_info);
skb2              646 net/ipv6/ip6_tunnel.c 	icmp_send(skb2, rel_type, rel_code, htonl(rel_info));
skb2              649 net/ipv6/ip6_tunnel.c 	kfree_skb(skb2);
skb2              669 net/ipv6/ip6_tunnel.c 		struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              671 net/ipv6/ip6_tunnel.c 		if (!skb2)
skb2              674 net/ipv6/ip6_tunnel.c 		skb_dst_drop(skb2);
skb2              675 net/ipv6/ip6_tunnel.c 		skb_pull(skb2, offset);
skb2              676 net/ipv6/ip6_tunnel.c 		skb_reset_network_header(skb2);
skb2              679 net/ipv6/ip6_tunnel.c 		rt = rt6_lookup(dev_net(skb->dev), &ipv6_hdr(skb2)->saddr,
skb2              680 net/ipv6/ip6_tunnel.c 				NULL, 0, skb2, 0);
skb2              683 net/ipv6/ip6_tunnel.c 			skb2->dev = rt->dst.dev;
skb2              685 net/ipv6/ip6_tunnel.c 		icmpv6_send(skb2, rel_type, rel_code, rel_info);
skb2              689 net/ipv6/ip6_tunnel.c 		kfree_skb(skb2);
skb2             2161 net/ipv6/ip6mr.c 				struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             2162 net/ipv6/ip6mr.c 				if (skb2)
skb2             2163 net/ipv6/ip6mr.c 					ip6mr_forward2(net, mrt, skb2, psend);
skb2             2273 net/ipv6/ip6mr.c 		struct sk_buff *skb2;
skb2             2285 net/ipv6/ip6mr.c 		skb2 = alloc_skb(sizeof(struct ipv6hdr), GFP_ATOMIC);
skb2             2286 net/ipv6/ip6mr.c 		if (!skb2) {
skb2             2291 net/ipv6/ip6mr.c 		NETLINK_CB(skb2).portid = portid;
skb2             2292 net/ipv6/ip6mr.c 		skb_reset_transport_header(skb2);
skb2             2294 net/ipv6/ip6mr.c 		skb_put(skb2, sizeof(struct ipv6hdr));
skb2             2295 net/ipv6/ip6mr.c 		skb_reset_network_header(skb2);
skb2             2297 net/ipv6/ip6mr.c 		iph = ipv6_hdr(skb2);
skb2             2309 net/ipv6/ip6mr.c 		err = ip6mr_cache_unresolved(mrt, vif, skb2, dev);
skb2              212 net/ipv6/netfilter.c 		struct sk_buff *skb2;
skb2              214 net/ipv6/netfilter.c 		skb2 = ip6_frag_next(skb, &state);
skb2              215 net/ipv6/netfilter.c 		if (IS_ERR(skb2)) {
skb2              216 net/ipv6/netfilter.c 			err = PTR_ERR(skb2);
skb2              220 net/ipv6/netfilter.c 		skb2->tstamp = tstamp;
skb2              221 net/ipv6/netfilter.c 		err = output(net, sk, data, skb2);
skb2              336 net/llc/llc_conn.c 			struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              339 net/llc/llc_conn.c 			if (!skb2)
skb2              341 net/llc/llc_conn.c 			skb = skb2;
skb2             3513 net/mac80211/rx.c 	struct sk_buff *skb = rx->skb, *skb2;
skb2             3557 net/mac80211/rx.c 			skb2 = skb_clone(skb, GFP_ATOMIC);
skb2             3558 net/mac80211/rx.c 			if (skb2) {
skb2             3559 net/mac80211/rx.c 				skb2->dev = prev_dev;
skb2             3560 net/mac80211/rx.c 				netif_receive_skb(skb2);
skb2              809 net/mac80211/status.c 	struct sk_buff *skb2;
skb2              843 net/mac80211/status.c 				skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              844 net/mac80211/status.c 				if (skb2) {
skb2              845 net/mac80211/status.c 					skb2->dev = prev_dev;
skb2              846 net/mac80211/status.c 					netif_rx(skb2);
skb2             1863 net/mac80211/tx.c 	struct sk_buff *skb2;
skb2             1883 net/mac80211/tx.c 	skb2 = __skb_dequeue(&tx.skbs);
skb2             1884 net/mac80211/tx.c 	if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) {
skb2             1885 net/mac80211/tx.c 		ieee80211_free_txskb(hw, skb2);
skb2              223 net/mac802154/rx.c 	struct sk_buff *skb2;
skb2              238 net/mac802154/rx.c 		skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              239 net/mac802154/rx.c 		if (skb2) {
skb2              240 net/mac802154/rx.c 			skb2->dev = sdata->dev;
skb2              241 net/mac802154/rx.c 			ieee802154_deliver_skb(skb2);
skb2             1553 net/netfilter/ipset/ip_set_core.c 		struct sk_buff *skb2;
skb2             1562 net/netfilter/ipset/ip_set_core.c 		skb2 = nlmsg_new(payload, GFP_KERNEL);
skb2             1563 net/netfilter/ipset/ip_set_core.c 		if (!skb2)
skb2             1565 net/netfilter/ipset/ip_set_core.c 		rep = __nlmsg_put(skb2, NETLINK_CB(skb).portid,
skb2             1577 net/netfilter/ipset/ip_set_core.c 			nlmsg_free(skb2);
skb2             1584 net/netfilter/ipset/ip_set_core.c 		netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid,
skb2             1711 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb2;
skb2             1723 net/netfilter/ipset/ip_set_core.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1724 net/netfilter/ipset/ip_set_core.c 	if (!skb2)
skb2             1727 net/netfilter/ipset/ip_set_core.c 	nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
skb2             1731 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) ||
skb2             1732 net/netfilter/ipset/ip_set_core.c 	    nla_put_string(skb2, IPSET_ATTR_SETNAME, set->name) ||
skb2             1733 net/netfilter/ipset/ip_set_core.c 	    nla_put_string(skb2, IPSET_ATTR_TYPENAME, set->type->name) ||
skb2             1734 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_FAMILY, set->family) ||
skb2             1735 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_REVISION, set->revision))
skb2             1737 net/netfilter/ipset/ip_set_core.c 	nlmsg_end(skb2, nlh2);
skb2             1739 net/netfilter/ipset/ip_set_core.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1746 net/netfilter/ipset/ip_set_core.c 	nlmsg_cancel(skb2, nlh2);
skb2             1748 net/netfilter/ipset/ip_set_core.c 	kfree_skb(skb2);
skb2             1766 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb2;
skb2             1783 net/netfilter/ipset/ip_set_core.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1784 net/netfilter/ipset/ip_set_core.c 	if (!skb2)
skb2             1787 net/netfilter/ipset/ip_set_core.c 	nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
skb2             1791 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) ||
skb2             1792 net/netfilter/ipset/ip_set_core.c 	    nla_put_string(skb2, IPSET_ATTR_TYPENAME, typename) ||
skb2             1793 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_FAMILY, family) ||
skb2             1794 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_REVISION, max) ||
skb2             1795 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_REVISION_MIN, min))
skb2             1797 net/netfilter/ipset/ip_set_core.c 	nlmsg_end(skb2, nlh2);
skb2             1800 net/netfilter/ipset/ip_set_core.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1807 net/netfilter/ipset/ip_set_core.c 	nlmsg_cancel(skb2, nlh2);
skb2             1809 net/netfilter/ipset/ip_set_core.c 	kfree_skb(skb2);
skb2             1825 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb2;
skb2             1832 net/netfilter/ipset/ip_set_core.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1833 net/netfilter/ipset/ip_set_core.c 	if (!skb2)
skb2             1836 net/netfilter/ipset/ip_set_core.c 	nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
skb2             1840 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, IPSET_PROTOCOL))
skb2             1842 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL_MIN, IPSET_PROTOCOL_MIN))
skb2             1844 net/netfilter/ipset/ip_set_core.c 	nlmsg_end(skb2, nlh2);
skb2             1846 net/netfilter/ipset/ip_set_core.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1853 net/netfilter/ipset/ip_set_core.c 	nlmsg_cancel(skb2, nlh2);
skb2             1855 net/netfilter/ipset/ip_set_core.c 	kfree_skb(skb2);
skb2             1867 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb2;
skb2             1881 net/netfilter/ipset/ip_set_core.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1882 net/netfilter/ipset/ip_set_core.c 	if (!skb2)
skb2             1885 net/netfilter/ipset/ip_set_core.c 	nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
skb2             1889 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) ||
skb2             1890 net/netfilter/ipset/ip_set_core.c 	    nla_put_u8(skb2, IPSET_ATTR_FAMILY, set->family) ||
skb2             1891 net/netfilter/ipset/ip_set_core.c 	    nla_put_net16(skb2, IPSET_ATTR_INDEX, htons(id)))
skb2             1893 net/netfilter/ipset/ip_set_core.c 	nlmsg_end(skb2, nlh2);
skb2             1895 net/netfilter/ipset/ip_set_core.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1902 net/netfilter/ipset/ip_set_core.c 	nlmsg_cancel(skb2, nlh2);
skb2             1904 net/netfilter/ipset/ip_set_core.c 	kfree_skb(skb2);
skb2             1919 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb2;
skb2             1936 net/netfilter/ipset/ip_set_core.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1937 net/netfilter/ipset/ip_set_core.c 	if (!skb2)
skb2             1940 net/netfilter/ipset/ip_set_core.c 	nlh2 = start_msg(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
skb2             1944 net/netfilter/ipset/ip_set_core.c 	if (nla_put_u8(skb2, IPSET_ATTR_PROTOCOL, protocol(attr)) ||
skb2             1945 net/netfilter/ipset/ip_set_core.c 	    nla_put_string(skb2, IPSET_ATTR_SETNAME, set->name))
skb2             1947 net/netfilter/ipset/ip_set_core.c 	nlmsg_end(skb2, nlh2);
skb2             1949 net/netfilter/ipset/ip_set_core.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1956 net/netfilter/ipset/ip_set_core.c 	nlmsg_cancel(skb2, nlh2);
skb2             1958 net/netfilter/ipset/ip_set_core.c 	kfree_skb(skb2);
skb2             1321 net/netfilter/nf_conntrack_netlink.c 	struct sk_buff *skb2 = NULL;
skb2             1361 net/netfilter/nf_conntrack_netlink.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             1362 net/netfilter/nf_conntrack_netlink.c 	if (skb2 == NULL) {
skb2             1368 net/netfilter/nf_conntrack_netlink.c 	err = ctnetlink_fill_info(skb2, NETLINK_CB(skb).portid, nlh->nlmsg_seq,
skb2             1375 net/netfilter/nf_conntrack_netlink.c 	err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             1382 net/netfilter/nf_conntrack_netlink.c 	kfree_skb(skb2);
skb2             2321 net/netfilter/nf_conntrack_netlink.c 	struct sk_buff *skb2;
skb2             2324 net/netfilter/nf_conntrack_netlink.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             2325 net/netfilter/nf_conntrack_netlink.c 	if (skb2 == NULL)
skb2             2328 net/netfilter/nf_conntrack_netlink.c 	err = ctnetlink_stat_ct_fill_info(skb2, NETLINK_CB(skb).portid,
skb2             2335 net/netfilter/nf_conntrack_netlink.c 	err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             2342 net/netfilter/nf_conntrack_netlink.c 	kfree_skb(skb2);
skb2             3038 net/netfilter/nf_conntrack_netlink.c 	struct sk_buff *skb2;
skb2             3087 net/netfilter/nf_conntrack_netlink.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2             3088 net/netfilter/nf_conntrack_netlink.c 	if (skb2 == NULL) {
skb2             3094 net/netfilter/nf_conntrack_netlink.c 	err = ctnetlink_exp_fill_info(skb2, NETLINK_CB(skb).portid,
skb2             3101 net/netfilter/nf_conntrack_netlink.c 	err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2             3108 net/netfilter/nf_conntrack_netlink.c 	kfree_skb(skb2);
skb2              720 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2              739 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2              740 net/netfilter/nf_tables_api.c 	if (!skb2)
skb2              743 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_table_info(skb2, net, NETLINK_CB(skb).portid,
skb2              749 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2              752 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             1413 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             1438 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             1439 net/netfilter/nf_tables_api.c 	if (!skb2)
skb2             1442 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_chain_info(skb2, net, NETLINK_CB(skb).portid,
skb2             1448 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             1451 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             2583 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             2617 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             2618 net/netfilter/nf_tables_api.c 	if (!skb2)
skb2             2621 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_rule_info(skb2, net, NETLINK_CB(skb).portid,
skb2             2627 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             2630 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             3490 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             3522 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             3523 net/netfilter/nf_tables_api.c 	if (skb2 == NULL)
skb2             3526 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_set(skb2, &ctx, set, NFT_MSG_NEWSET, 0);
skb2             3530 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             3533 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             5456 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             5490 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             5491 net/netfilter/nf_tables_api.c 	if (!skb2)
skb2             5497 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_obj_info(skb2, net, NETLINK_CB(skb).portid,
skb2             5503 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             5505 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             6139 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             6167 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             6168 net/netfilter/nf_tables_api.c 	if (!skb2)
skb2             6171 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_flowtable_info(skb2, net, NETLINK_CB(skb).portid,
skb2             6178 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             6180 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2             6299 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             6306 net/netfilter/nf_tables_api.c 	skb2 = nlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL);
skb2             6307 net/netfilter/nf_tables_api.c 	if (skb2 == NULL)
skb2             6310 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_gen_info(skb2, net, NETLINK_CB(skb).portid,
skb2             6313 net/netfilter/nf_tables_api.c 		kfree_skb(skb2);
skb2             6317 net/netfilter/nf_tables_api.c 	nfnetlink_send(skb2, net, NETLINK_CB(skb).portid, NFNLGRP_NFTABLES,
skb2             6330 net/netfilter/nf_tables_api.c 	struct sk_buff *skb2;
skb2             6333 net/netfilter/nf_tables_api.c 	skb2 = alloc_skb(NLMSG_GOODSIZE, GFP_ATOMIC);
skb2             6334 net/netfilter/nf_tables_api.c 	if (skb2 == NULL)
skb2             6337 net/netfilter/nf_tables_api.c 	err = nf_tables_fill_gen_info(skb2, net, NETLINK_CB(skb).portid,
skb2             6342 net/netfilter/nf_tables_api.c 	return nlmsg_unicast(nlsk, skb2, NETLINK_CB(skb).portid);
skb2             6344 net/netfilter/nf_tables_api.c 	kfree_skb(skb2);
skb2              292 net/netfilter/nfnetlink_acct.c 		struct sk_buff *skb2;
skb2              297 net/netfilter/nfnetlink_acct.c 		skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2              298 net/netfilter/nfnetlink_acct.c 		if (skb2 == NULL) {
skb2              303 net/netfilter/nfnetlink_acct.c 		ret = nfnl_acct_fill_info(skb2, NETLINK_CB(skb).portid,
skb2              308 net/netfilter/nfnetlink_acct.c 			kfree_skb(skb2);
skb2              311 net/netfilter/nfnetlink_acct.c 		ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid,
skb2              622 net/netfilter/nfnetlink_cthelper.c 	struct sk_buff *skb2;
skb2              660 net/netfilter/nfnetlink_cthelper.c 		skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2              661 net/netfilter/nfnetlink_cthelper.c 		if (skb2 == NULL) {
skb2              666 net/netfilter/nfnetlink_cthelper.c 		ret = nfnl_cthelper_fill_info(skb2, NETLINK_CB(skb).portid,
skb2              671 net/netfilter/nfnetlink_cthelper.c 			kfree_skb(skb2);
skb2              675 net/netfilter/nfnetlink_cthelper.c 		ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid,
skb2              263 net/netfilter/nfnetlink_cttimeout.c 		struct sk_buff *skb2;
skb2              268 net/netfilter/nfnetlink_cttimeout.c 		skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2              269 net/netfilter/nfnetlink_cttimeout.c 		if (skb2 == NULL) {
skb2              274 net/netfilter/nfnetlink_cttimeout.c 		ret = ctnl_timeout_fill_info(skb2, NETLINK_CB(skb).portid,
skb2              279 net/netfilter/nfnetlink_cttimeout.c 			kfree_skb(skb2);
skb2              282 net/netfilter/nfnetlink_cttimeout.c 		ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid,
skb2              431 net/netfilter/nfnetlink_cttimeout.c 	struct sk_buff *skb2;
skb2              487 net/netfilter/nfnetlink_cttimeout.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2              488 net/netfilter/nfnetlink_cttimeout.c 	if (skb2 == NULL) {
skb2              493 net/netfilter/nfnetlink_cttimeout.c 	ret = cttimeout_default_fill_info(net, skb2, NETLINK_CB(skb).portid,
skb2              499 net/netfilter/nfnetlink_cttimeout.c 		kfree_skb(skb2);
skb2              503 net/netfilter/nfnetlink_cttimeout.c 	ret = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).portid, MSG_DONTWAIT);
skb2              612 net/netfilter/nft_compat.c 	struct sk_buff *skb2;
skb2              654 net/netfilter/nft_compat.c 	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
skb2              655 net/netfilter/nft_compat.c 	if (skb2 == NULL) {
skb2              661 net/netfilter/nft_compat.c 	if (nfnl_compat_fill_info(skb2, NETLINK_CB(skb).portid,
skb2              667 net/netfilter/nft_compat.c 		kfree_skb(skb2);
skb2              671 net/netfilter/nft_compat.c 	ret = netlink_unicast(nfnl, skb2, NETLINK_CB(skb).portid,
skb2             1398 net/netlink/af_netlink.c 	struct sk_buff *skb, *skb2;
skb2             1434 net/netlink/af_netlink.c 	if (p->skb2 == NULL) {
skb2             1436 net/netlink/af_netlink.c 			p->skb2 = skb_clone(p->skb, p->allocation);
skb2             1438 net/netlink/af_netlink.c 			p->skb2 = skb_get(p->skb);
skb2             1443 net/netlink/af_netlink.c 			skb_orphan(p->skb2);
skb2             1446 net/netlink/af_netlink.c 	if (p->skb2 == NULL) {
skb2             1454 net/netlink/af_netlink.c 	if (p->tx_filter && p->tx_filter(sk, p->skb2, p->tx_data)) {
skb2             1455 net/netlink/af_netlink.c 		kfree_skb(p->skb2);
skb2             1456 net/netlink/af_netlink.c 		p->skb2 = NULL;
skb2             1459 net/netlink/af_netlink.c 	if (sk_filter(sk, p->skb2)) {
skb2             1460 net/netlink/af_netlink.c 		kfree_skb(p->skb2);
skb2             1461 net/netlink/af_netlink.c 		p->skb2 = NULL;
skb2             1464 net/netlink/af_netlink.c 	NETLINK_CB(p->skb2).nsid = peernet2id(sock_net(sk), p->net);
skb2             1465 net/netlink/af_netlink.c 	if (NETLINK_CB(p->skb2).nsid != NETNSA_NSID_NOT_ASSIGNED)
skb2             1466 net/netlink/af_netlink.c 		NETLINK_CB(p->skb2).nsid_is_set = true;
skb2             1467 net/netlink/af_netlink.c 	val = netlink_broadcast_deliver(sk, p->skb2);
skb2             1475 net/netlink/af_netlink.c 		p->skb2 = NULL;
skb2             1502 net/netlink/af_netlink.c 	info.skb2 = NULL;
skb2             1518 net/netlink/af_netlink.c 		kfree_skb(info.skb2);
skb2             1521 net/netlink/af_netlink.c 	consume_skb(info.skb2);
skb2              214 net/sched/act_mirred.c 	struct sk_buff *skb2 = skb;
skb2              261 net/sched/act_mirred.c 		skb2 = skb_clone(skb, GFP_ATOMIC);
skb2              262 net/sched/act_mirred.c 		if (!skb2)
skb2              275 net/sched/act_mirred.c 			skb_pull_rcsum(skb2, mac_len);
skb2              278 net/sched/act_mirred.c 			skb_push_rcsum(skb2, mac_len);
skb2              282 net/sched/act_mirred.c 	skb2->skb_iif = skb->dev->ifindex;
skb2              283 net/sched/act_mirred.c 	skb2->dev = dev;
skb2              287 net/sched/act_mirred.c 		skb_set_redirected(skb2, skb2->tc_at_ingress);
skb2              300 net/sched/act_mirred.c 		err = dev_queue_xmit(skb2);
skb2              302 net/sched/act_mirred.c 		err = netif_receive_skb(skb2);
skb2              156 net/sched/sch_choke.c 			     struct sk_buff *skb2)
skb2              160 net/sched/sch_choke.c 	if (skb1->protocol != skb2->protocol)
skb2              169 net/sched/sch_choke.c 	if (!choke_skb_cb(skb2)->keys_valid) {
skb2              170 net/sched/sch_choke.c 		choke_skb_cb(skb2)->keys_valid = 1;
skb2              171 net/sched/sch_choke.c 		skb_flow_dissect_flow_keys(skb2, &temp, 0);
skb2              172 net/sched/sch_choke.c 		make_flow_keys_digest(&choke_skb_cb(skb2)->keys, &temp);
skb2              176 net/sched/sch_choke.c 		       &choke_skb_cb(skb2)->keys,
skb2              440 net/sched/sch_netem.c 	struct sk_buff *skb2;
skb2              478 net/sched/sch_netem.c 	if (count > 1 && (skb2 = skb_clone(skb, GFP_ATOMIC)) != NULL) {
skb2              483 net/sched/sch_netem.c 		rootq->enqueue(skb2, rootq, to_free);
skb2              603 net/sched/sch_netem.c 			skb2 = segs->next;
skb2              615 net/sched/sch_netem.c 			segs = skb2;
skb2               79 net/xfrm/xfrm_device.c 	struct sk_buff *skb2;
skb2              149 net/xfrm/xfrm_device.c 	skb2 = skb;
skb2              152 net/xfrm/xfrm_device.c 		struct sk_buff *nskb = skb2->next;
skb2              155 net/xfrm/xfrm_device.c 		skb_mark_not_on_list(skb2);
skb2              157 net/xfrm/xfrm_device.c 		xo = xfrm_offload(skb2);
skb2              160 net/xfrm/xfrm_device.c 		xfrm_outer_mode_prep(x, skb2);
skb2              162 net/xfrm/xfrm_device.c 		err = x->type_offload->xmit(x, skb2, esp_features);
skb2              164 net/xfrm/xfrm_device.c 			skb2->next = nskb;
skb2              167 net/xfrm/xfrm_device.c 			skb2->next = nskb;
skb2              168 net/xfrm/xfrm_device.c 			kfree_skb_list(skb2);
skb2              171 net/xfrm/xfrm_device.c 			if (skb == skb2)
skb2              180 net/xfrm/xfrm_device.c 		skb_push(skb2, skb2->data - skb_mac_header(skb2));
skb2              183 net/xfrm/xfrm_device.c 		skb2 = nskb;
skb2              184 net/xfrm/xfrm_device.c 	} while (skb2);