Home
last modified time | relevance | path

Searched refs:skb_shinfo (Results 1 – 200 of 224) sorted by relevance

12

/linux-4.4.14/net/core/
Dskbuff.c258 shinfo = skb_shinfo(skb); in __alloc_skb()
326 shinfo = skb_shinfo(skb); in __build_skb()
536 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag()
553 skb_drop_list(&skb_shinfo(skb)->frag_list); in skb_drop_fraglist()
576 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data()
720 if (skb_shinfo(skb)->tx_flags & SKBTX_DEV_ZEROCOPY) { in skb_tx_error()
723 uarg = skb_shinfo(skb)->destructor_arg; in skb_tx_error()
726 skb_shinfo(skb)->tx_flags &= ~SKBTX_DEV_ZEROCOPY; in skb_tx_error()
839 atomic_inc(&(skb_shinfo(skb)->dataref)); in __skb_clone()
881 int num_frags = skb_shinfo(skb)->nr_frags; in skb_copy_ubufs()
[all …]
Dtso.c11 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs()
53 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data()
54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data()
77 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start()
78 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
Ddatagram.c376 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_iter()
378 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_iter()
463 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_from_iter()
465 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter()
592 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_and_csum_datagram()
594 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_and_csum_datagram()
Ddev.c2421 skb->len, skb->data_len, skb_shinfo(skb)->gso_size, in skb_warn_bad_offload()
2422 skb_shinfo(skb)->gso_type, skb->ip_summed); in skb_warn_bad_offload()
2437 if (unlikely(skb_shinfo(skb)->gso_size)) { in skb_checksum_help()
2596 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma()
2597 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma()
2608 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma()
2609 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma()
2679 u16 gso_segs = skb_shinfo(skb)->gso_segs; in netif_skb_features()
2842 const struct skb_shared_info *shinfo = skb_shinfo(skb); in qdisc_pkt_len_init()
3084 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_SCHED_TSTAMP)) in __dev_queue_xmit()
[all …]
Dpktgen.c2740 skb_shinfo(skb)->frags[i].page_offset = 0; in pktgen_finalize_skb()
2743 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb()
2746 skb_frag_size_set(&skb_shinfo(skb)->frags[i], frag_len); in pktgen_finalize_skb()
2747 datalen -= skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb()
2748 skb->len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb()
2749 skb->data_len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb()
2751 skb_shinfo(skb)->nr_frags = i; in pktgen_finalize_skb()
3241 int nr_frags = pkt_dev->skb ? skb_shinfo(pkt_dev->skb)->nr_frags : -1; in pktgen_stop_device()
/linux-4.4.14/net/ipv6/
Dudp_offload.c33 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment()
39 int type = skb_shinfo(skb)->gso_type; in udp6_ufo_fragment()
53 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp6_ufo_fragment()
56 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment()
63 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment()
115 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment()
117 fptr->identification = skb_shinfo(skb)->ip6_frag_id; in udp6_ufo_fragment()
163 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp6_gro_complete()
167 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp6_gro_complete()
Dip6_offload.c71 if (unlikely(skb_shinfo(skb)->gso_type & in ipv6_gso_segment()
104 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in ipv6_gso_segment()
289 skb_shinfo(skb)->gso_type |= SKB_GSO_SIT; in sit_gro_complete()
Dreassembly.c441 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip6_frag_reasm()
443 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip6_frag_reasm()
444 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip6_frag_reasm()
480 if (!skb_shinfo(head)->frag_list) in ip6_frag_reasm()
481 skb_shinfo(head)->frag_list = fp; in ip6_frag_reasm()
Dtcpv6_offload.c39 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete()
Dip6_output.c652 frag = skb_shinfo(skb)->frag_list; in ip6_fragment()
1132 skb_shinfo(skb)->gso_size = (mtu - fragheaderlen - in ip6_ufo_append_data()
1134 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip6_ufo_append_data()
1135 skb_shinfo(skb)->ip6_frag_id = ipv6_select_ident(sock_net(sk), in ip6_ufo_append_data()
1454 skb_shinfo(skb)->tx_flags = tx_flags; in __ip6_append_data()
1456 skb_shinfo(skb)->tskey = tskey; in __ip6_append_data()
1515 int i = skb_shinfo(skb)->nr_frags; in __ip6_append_data()
1529 skb_shinfo(skb)->nr_frags = ++i; in __ip6_append_data()
1539 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip6_append_data()
1635 tail_skb = &(skb_shinfo(skb)->frag_list); in __ip6_make_skb()
Doutput_core.c61 skb_shinfo(skb)->ip6_frag_id = htonl(id); in ipv6_proxy_select_ident()
Dip6_input.c115 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ipv6_rcv()
Dudp.c991 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing()
/linux-4.4.14/net/ipv4/
Dtcp_offload.c22 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp()
23 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp()
80 mss = skb_shinfo(skb)->gso_size; in tcp_gso_segment()
86 int type = skb_shinfo(skb)->gso_type; in tcp_gso_segment()
104 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in tcp_gso_segment()
128 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_SW_TSTAMP)) in tcp_gso_segment()
129 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment()
245 mss = skb_shinfo(p)->gso_size; in tcp_gro_receive()
282 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp_gro_complete()
285 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_gro_complete()
[all …]
Dudp_offload.c42 bool need_csum = !!(skb_shinfo(skb)->gso_type & in __skb_udp_tunnel_segment()
44 bool remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM); in __skb_udp_tunnel_segment()
188 (skb_shinfo(skb)->gso_type & in udp4_ufo_fragment()
197 mss = skb_shinfo(skb)->gso_size; in udp4_ufo_fragment()
203 int type = skb_shinfo(skb)->gso_type; in udp4_ufo_fragment()
214 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp4_ufo_fragment()
407 skb_shinfo(skb)->gso_type |= SKB_GSO_TUNNEL_REMCSUM; in udp_gro_complete()
421 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp4_gro_complete()
425 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp4_gro_complete()
Dinet_lro.c154 nr_frags = skb_shinfo(skb)->nr_frags; in lro_init_desc()
156 lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); in lro_init_desc()
228 skb_shinfo(parent)->frag_list = skb; in lro_add_packet()
283 skb_shinfo(lro_desc->parent)->gso_size = lro_desc->mss; in lro_flush()
Dip_output.c452 skb_shinfo(skb)->gso_segs ?: 1); in ip_queue_xmit()
600 frag = skb_shinfo(skb)->frag_list; in ip_do_fragment()
857 skb_shinfo(skb)->gso_size = maxfraglen - fragheaderlen; in ip_ufo_append_data()
858 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip_ufo_append_data()
1013 skb_shinfo(skb)->tx_flags = cork->tx_flags; in __ip_append_data()
1015 skb_shinfo(skb)->tskey = tskey; in __ip_append_data()
1071 int i = skb_shinfo(skb)->nr_frags; in __ip_append_data()
1085 skb_shinfo(skb)->nr_frags = ++i; in __ip_append_data()
1095 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip_append_data()
1246 skb_shinfo(skb)->gso_size = mtu - fragheaderlen; in ip_append_page()
[all …]
Dgre_offload.c31 if (unlikely(skb_shinfo(skb)->gso_type & in gre_gso_segment()
236 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
Dip_fragment.c588 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip_frag_reasm()
590 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip_frag_reasm()
591 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip_frag_reasm()
600 skb_shinfo(head)->frag_list = head->next; in ip_frag_reasm()
Dip_tunnel_core.c80 __ip_select_ident(net, iph, skb_shinfo(skb)->gso_segs ?: 1); in iptunnel_xmit()
165 skb_shinfo(skb)->gso_type |= gso_type_mask; in iptunnel_handle_offloads()
Dtcp_output.c382 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_ecn_send()
991 skb_shinfo(skb)->gso_type = sk->sk_gso_type; in tcp_transmit_skb()
1018 skb_shinfo(skb)->gso_segs = tcp_skb_pcount(skb); in tcp_transmit_skb()
1019 skb_shinfo(skb)->gso_size = tcp_skb_mss(skb); in tcp_transmit_skb()
1117 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_fragment_tstamp()
1121 struct skb_shared_info *shinfo2 = skb_shinfo(skb2); in tcp_fragment_tstamp()
1176 if (!skb_shinfo(skb)->nr_frags && skb->ip_summed != CHECKSUM_PARTIAL) { in tcp_fragment()
1237 shinfo = skb_shinfo(skb); in __pskb_trim_head()
1968 if (!skb_shinfo(skb)->nr_frags) { in tcp_mtu_probe()
2497 if (skb_shinfo(skb)->nr_frags != 0) in tcp_can_collapse()
Dtcp.c435 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_tx_timestamp()
940 i = skb_shinfo(skb)->nr_frags; in do_tcp_sendpages()
950 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages()
955 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in do_tcp_sendpages()
1207 int i = skb_shinfo(skb)->nr_frags; in tcp_sendmsg()
1236 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg()
3027 const struct skb_shared_info *shi = skb_shinfo(skb); in tcp_md5_hash_skb_data()
Dip_input.c421 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ip_rcv()
Dtcp_minisocks.c821 tcp_sk(child)->segs_in += max_t(u16, 1, skb_shinfo(skb)->gso_segs); in tcp_child_process()
Daf_inet.c1210 if (unlikely(skb_shinfo(skb)->gso_type & in inet_gso_segment()
1254 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in inet_gso_segment()
Draw.c380 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_send_hdrinc()
Dip_gre.c473 !(skb_shinfo(skb)->gso_type & in build_header()
/linux-4.4.14/drivers/net/xen-netback/
Dnetback.c369 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_frag_copy()
371 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_frag_copy()
425 int nr_frags = skb_shinfo(skb)->nr_frags; in xenvif_gop_skb()
438 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_skb()
440 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_skb()
449 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb()
459 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb()
485 skb_frag_page(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb()
486 skb_frag_size(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb()
487 skb_shinfo(skb)->frags[i].page_offset, in xenvif_gop_skb()
[all …]
Dinterface.c57 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in xenvif_skb_zerocopy_prepare()
/linux-4.4.14/include/trace/events/
Dnet.h54 __entry->tx_flags = skb_shinfo(skb)->tx_flags;
55 __entry->gso_size = skb_shinfo(skb)->gso_size;
56 __entry->gso_segs = skb_shinfo(skb)->gso_segs;
57 __entry->gso_type = skb_shinfo(skb)->gso_type;
189 __entry->nr_frags = skb_shinfo(skb)->nr_frags;
190 __entry->gso_size = skb_shinfo(skb)->gso_size;
191 __entry->gso_type = skb_shinfo(skb)->gso_type;
/linux-4.4.14/drivers/net/ethernet/sfc/
Dtx.c256 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in efx_skb_copy_bits_to_pio()
257 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in efx_skb_copy_bits_to_pio()
267 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->frag_list); in efx_skb_copy_bits_to_pio()
282 if (skb_shinfo(skb)->nr_frags) { in efx_enqueue_skb_pio()
346 if (skb_shinfo(skb)->gso_size) in efx_enqueue_skb()
411 if (i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb()
413 fragment = &skb_shinfo(skb)->frags[i]; in efx_enqueue_skb()
457 skb_shinfo(skb)->nr_frags + 1); in efx_enqueue_skb()
1152 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet()
1156 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet()
[all …]
Drx.c450 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_packet_gro()
455 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_packet_gro()
503 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_mk_skb()
509 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_mk_skb()
Dnet_driver.h1510 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in efx_xmit_with_hwtstamp()
1514 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in efx_xmit_hwtstamp_pending()
Dptp.c890 if (skb_shinfo(skb)->nr_frags != 0) { in efx_ptp_xmit_skb()
/linux-4.4.14/include/linux/
Dskbuff.h1114 #define skb_shinfo(SKB) ((struct skb_shared_info *)(skb_end_pointer(SKB))) macro
1118 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps()
1223 (atomic_read(&skb_shinfo(skb)->dataref) & SKB_DATAREF_MASK) != 1; in skb_cloned()
1250 dataref = atomic_read(&skb_shinfo(skb)->dataref); in skb_header_cloned()
1268 atomic_add(1 << SKB_DATAREF_SHIFT, &skb_shinfo(skb)->dataref); in skb_header_release()
1281 atomic_set(&skb_shinfo(skb)->dataref, 1 + (1 << SKB_DATAREF_SHIFT)); in __skb_header_release()
1697 for (i = (int)skb_shinfo(skb)->nr_frags - 1; i >= 0; i--) in skb_pagelen()
1698 len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in skb_pagelen()
1718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_fill_page_desc()
1752 skb_shinfo(skb)->nr_frags = i + 1; in skb_fill_page_desc()
[all …]
/linux-4.4.14/drivers/staging/octeon/
Dethernet-tx.c204 if (unlikely(skb_shinfo(skb)->nr_frags > 5)) { in cvm_oct_xmit()
274 if (skb_shinfo(skb)->nr_frags == 0) { in cvm_oct_xmit()
283 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in cvm_oct_xmit()
284 struct skb_frag_struct *fs = skb_shinfo(skb)->frags + i; in cvm_oct_xmit()
293 hw_buffer.s.size = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit()
294 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit()
349 if (unlikely(skb_shinfo(skb)->nr_frags)) { in cvm_oct_xmit()
/linux-4.4.14/net/xfrm/
Dxfrm_ipcomp.c76 if (WARN_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS)) in ipcomp_decompress()
79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress()
100 skb_shinfo(skb)->nr_frags++; in ipcomp_decompress()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/
Den_tx.c191 eseg->mss = cpu_to_be16(skb_shinfo(skb)->gso_size); in mlx5e_sq_xmit()
196 (skb_shinfo(skb)->gso_segs - 1) * ihs; in mlx5e_sq_xmit()
202 !skb_shinfo(skb)->nr_frags; in mlx5e_sq_xmit()
242 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlx5e_sq_xmit()
243 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in mlx5e_sq_xmit()
Den_rx.c114 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in mlx5e_lro_update_hdr()
119 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in mlx5e_lro_update_hdr()
199 skb_shinfo(skb)->gso_size = DIV_ROUND_UP(cqe_bcnt, lro_num_seg); in mlx5e_build_rx_skb()
/linux-4.4.14/drivers/net/ethernet/sun/
Dsunvnet.c1020 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_map()
1021 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_map()
1060 docopy = skb_shinfo(skb)->nr_frags >= ncookies; in vnet_skb_shape()
1061 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_shape()
1062 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_shape()
1123 skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; in vnet_skb_shape()
1124 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in vnet_skb_shape()
1174 gso_size = skb_shinfo(skb)->gso_size; in vnet_handle_offloads()
1175 gso_type = skb_shinfo(skb)->gso_type; in vnet_handle_offloads()
1176 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads()
[all …]
Dsungem.c661 if (skb_shinfo(skb)->nr_frags) { in gem_tx()
662 int last = entry + skb_shinfo(skb)->nr_frags; in gem_tx()
680 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_tx()
1015 if (unlikely(TX_BUFFS_AVAIL(gp) <= (skb_shinfo(skb)->nr_frags + 1))) { in gem_start_xmit()
1027 if (skb_shinfo(skb)->nr_frags == 0) { in gem_start_xmit()
1064 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in gem_start_xmit()
1065 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in gem_start_xmit()
1074 if (frag == skb_shinfo(skb)->nr_frags - 1) in gem_start_xmit()
1600 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_clean_rings()
1609 if (frag != skb_shinfo(skb)->nr_frags) in gem_clean_rings()
Dsunhme.c1225 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_clean_rings()
1239 if (frag != skb_shinfo(skb)->nr_frags) in happy_meal_clean_rings()
1939 if (skb_shinfo(skb)->nr_frags) { in happy_meal_tx()
1942 last = elem + skb_shinfo(skb)->nr_frags; in happy_meal_tx()
1951 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_tx()
2305 if (TX_BUFFS_AVAIL(hp) <= (skb_shinfo(skb)->nr_frags + 1)) { in happy_meal_start_xmit()
2317 if (skb_shinfo(skb)->nr_frags == 0) { in happy_meal_start_xmit()
2343 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_start_xmit()
2344 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in happy_meal_start_xmit()
2356 if (frag == skb_shinfo(skb)->nr_frags - 1) in happy_meal_start_xmit()
Dcassini.c1880 count -= skb_shinfo(skb)->nr_frags + in cas_tx_ringN()
1891 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_tx_ringN()
2005 skb_frag_t *frag = skb_shinfo(skb)->frags; in cas_rx_process_pkt()
2041 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt()
2065 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt()
2778 CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { in cas_xmit_tx_ringN()
2798 nr_frags = skb_shinfo(skb)->nr_frags; in cas_xmit_tx_ringN()
2824 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in cas_xmit_tx_ringN()
3910 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_clean_txd()
3922 if (frag != skb_shinfo(skb)->nr_frags) { in cas_clean_txd()
Dniu.c3280 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, offset, size); in niu_rx_skb_append()
3586 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in release_tx_packet()
3590 skb_frag_size(&skb_shinfo(skb)->frags[i]), in release_tx_packet()
6646 if (niu_tx_avail(rp) <= (skb_shinfo(skb)->nr_frags + 1)) { in niu_start_xmit()
6695 nfg = skb_shinfo(skb)->nr_frags; in niu_start_xmit()
6715 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in niu_start_xmit()
6716 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in niu_start_xmit()
/linux-4.4.14/drivers/net/
Dxen-netfront.c310 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_alloc_rx_buffers()
522 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots()
529 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots()
634 if (skb_shinfo(skb)->gso_size) { in xennet_start_xmit()
642 gso->u.gso.size = skb_shinfo(skb)->gso_size; in xennet_start_xmit()
643 gso->u.gso.type = (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) ? in xennet_start_xmit()
657 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xennet_start_xmit()
658 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit()
859 skb_shinfo(skb)->gso_size = gso->u.gso.size; in xennet_set_skb_gso()
860 skb_shinfo(skb)->gso_type = in xennet_set_skb_gso()
[all …]
Dvirtio_net.c307 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, offset, in page_to_skb()
385 num_skb_frags = skb_shinfo(curr_skb)->nr_frags; in receive_mergeable()
392 skb_shinfo(curr_skb)->frag_list = nskb; in receive_mergeable()
496 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in receive_buf()
499 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in receive_buf()
502 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in receive_buf()
511 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in receive_buf()
513 skb_shinfo(skb)->gso_size = virtio16_to_cpu(vi->vdev, in receive_buf()
515 if (skb_shinfo(skb)->gso_size == 0) { in receive_buf()
521 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in receive_buf()
[all …]
Dmacvtap.c657 skb_shinfo(skb)->gso_size = macvtap16_to_cpu(q, vnet_hdr->gso_size); in macvtap_skb_from_vnet_hdr()
658 skb_shinfo(skb)->gso_type = gso_type; in macvtap_skb_from_vnet_hdr()
661 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in macvtap_skb_from_vnet_hdr()
662 skb_shinfo(skb)->gso_segs = 0; in macvtap_skb_from_vnet_hdr()
674 struct skb_shared_info *sinfo = skb_shinfo(skb); in macvtap_skb_to_vnet_hdr()
821 skb_shinfo(skb)->destructor_arg = m->msg_control; in macvtap_get_user()
822 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in macvtap_get_user()
823 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in macvtap_get_user()
Dtun.c866 sock_tx_timestamp(skb->sk, &skb_shinfo(skb)->tx_flags); in tun_net_xmit()
1223 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in tun_get_user()
1226 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in tun_get_user()
1229 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in tun_get_user()
1238 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tun_get_user()
1240 skb_shinfo(skb)->gso_size = tun16_to_cpu(tun, gso.gso_size); in tun_get_user()
1241 if (skb_shinfo(skb)->gso_size == 0) { in tun_get_user()
1248 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in tun_get_user()
1249 skb_shinfo(skb)->gso_segs = 0; in tun_get_user()
1254 skb_shinfo(skb)->destructor_arg = msg_control; in tun_get_user()
[all …]
/linux-4.4.14/net/ipv6/netfilter/
Dnf_conntrack_reasm.c423 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm()
425 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in nf_ct_frag6_reasm()
426 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in nf_ct_frag6_reasm()
445 skb_shinfo(head)->frag_list = head->next; in nf_ct_frag6_reasm()
478 fp = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm()
/linux-4.4.14/net/ieee802154/6lowpan/
Dreassembly.c271 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in lowpan_frag_reasm()
273 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in lowpan_frag_reasm()
274 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in lowpan_frag_reasm()
294 if (!skb_shinfo(head)->frag_list) in lowpan_frag_reasm()
295 skb_shinfo(head)->frag_list = fp; in lowpan_frag_reasm()
/linux-4.4.14/drivers/net/ethernet/neterion/vxge/
Dvxge-main.h515 #define vxge_tcp_mss(skb) (skb_shinfo(skb)->gso_size)
516 #define vxge_udp_mss(skb) (skb_shinfo(skb)->gso_size)
517 #define vxge_offload_type(skb) (skb_shinfo(skb)->gso_type)
Dvxge-main.c565 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit_compl()
566 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit_compl()
913 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit()
923 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit()
978 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit()
1047 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_tx_term()
1048 frag = &skb_shinfo(skb)->frags[0]; in vxge_tx_term()
/linux-4.4.14/drivers/net/ethernet/cavium/thunder/
Dnicvf_queues.c917 struct skb_shared_info *sh = skb_shinfo(skb); in nicvf_tso_count_subdescs()
963 if (skb_shinfo(skb)->gso_size) { in nicvf_sq_subdesc_required()
968 if (skb_shinfo(skb)->nr_frags) in nicvf_sq_subdesc_required()
969 subdesc_cnt += skb_shinfo(skb)->nr_frags; in nicvf_sq_subdesc_required()
1054 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in nicvf_sq_append_tso()
1136 if (skb_shinfo(skb)->gso_size) in nicvf_sq_append_skb()
1151 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in nicvf_sq_append_skb()
1154 frag = &skb_shinfo(skb)->frags[i]; in nicvf_sq_append_skb()
1225 if (!skb_shinfo(skb)->frag_list) in nicvf_get_rcv_skb()
1226 skb_shinfo(skb)->frag_list = skb_frag; in nicvf_get_rcv_skb()
/linux-4.4.14/drivers/infiniband/hw/nes/
Dnes_nic.c394 wqe_misc |= NES_NIC_SQ_WQE_LSO_ENABLE | (u16)skb_shinfo(skb)->gso_size; in nes_nic_send()
411 if ((skb_shinfo(skb)->nr_frags + 1) > 4) { in nes_nic_send()
413 netdev->name, skb_shinfo(skb)->nr_frags + 2, skb_headlen(skb)); in nes_nic_send()
437 for (skb_fragment_index = 0; skb_fragment_index < skb_shinfo(skb)->nr_frags; in nes_nic_send()
440 &skb_shinfo(skb)->frags[skb_fragment_index]; in nes_nic_send()
445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); in nes_nic_send()
524 nr_frags = skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit()
557 for (tso_frag_count=0; tso_frag_count < skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit()
560 &skb_shinfo(skb)->frags[tso_frag_count]; in nes_netdev_start_xmit()
599 skb_shinfo(skb)->nr_frags, skb_is_gso(skb), skb_shinfo(skb)->gso_size); in nes_netdev_start_xmit()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/
Di40e_txrx.c1035 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps()
1486 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso()
1674 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize()
1675 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize()
1688 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize()
1693 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize()
1695 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize()
1781 gso_segs = skb_shinfo(skb)->gso_segs; in i40evf_tx_map()
1796 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in i40evf_tx_map()
1933 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in i40evf_xmit_descriptor_count()
[all …]
/linux-4.4.14/drivers/net/ethernet/aeroflex/
Dgreth.c104 if (skb_shinfo(skb)->nr_frags == 0) in greth_print_tx_packet()
112 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in greth_print_tx_packet()
115 skb_frag_address(&skb_shinfo(skb)->frags[i]), in greth_print_tx_packet()
116 skb_shinfo(skb)->frags[i].size, true); in greth_print_tx_packet()
191 int nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_rings()
201 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_rings()
474 nr_frags = skb_shinfo(skb)->nr_frags; in greth_start_xmit_gbit()
521 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_start_xmit_gbit()
691 nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_tx_gbit()
718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_tx_gbit()
/linux-4.4.14/net/mpls/
Dmpls_gso.c29 if (unlikely(skb_shinfo(skb)->gso_type & in mpls_gso_segment()
/linux-4.4.14/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_io.c397 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_encap_pkt()
500 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_pkt()
587 nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_map_tx_skb()
599 frag = &skb_shinfo(skb)->frags[i]; in qlcnic_map_tx_skb()
629 int i, nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_unmap_buffers()
676 frag_count = skb_shinfo(skb)->nr_frags + 1; in qlcnic_xmit_frame()
683 delta += skb_frag_size(&skb_shinfo(skb)->frags[i]); in qlcnic_xmit_frame()
688 frag_count = 1 + skb_shinfo(skb)->nr_frags; in qlcnic_xmit_frame()
1337 skb_shinfo(skb)->gso_size = qlcnic_get_lro_sts_mss(sts_data1); in qlcnic_process_lro()
1339 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in qlcnic_process_lro()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/
Dfm10k_main.c309 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in fm10k_add_rx_frag()
813 first->gso_segs = skb_shinfo(skb)->gso_segs; in fm10k_tso()
819 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso()
918 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in fm10k_tx_desc_flags()
919 likely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in fm10k_tx_desc_flags()
1000 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in fm10k_tx_map()
1108 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in fm10k_xmit_frame_ring()
1109 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in fm10k_xmit_frame_ring()
Dfm10k_ptp.c74 skb_shinfo(clone)->tx_flags |= SKBTX_IN_PROGRESS; in fm10k_ts_tx_enqueue()
Dfm10k_netdev.c664 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in fm10k_xmit_frame()
/linux-4.4.14/net/phonet/
Dpep-gprs.c112 skb_shinfo(rskb)->frag_list = skb; in gprs_recv()
120 skb->next = skb_shinfo(skb)->frag_list; in gprs_recv()
Dpep.c1221 skb_shinfo(rskb)->frag_list = skb; in pep_write()
1229 skb->next = skb_shinfo(skb)->frag_list; in pep_write()
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
Di40e_txrx.c1570 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps()
2237 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso()
2259 if (likely(!(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP))) in i40e_tsyn()
2275 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in i40e_tsyn()
2510 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize()
2511 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize()
2524 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize()
2529 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize()
2531 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize()
2589 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_tx_map()
[all …]
Di40e_fcoe.c1218 if (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE) { in i40e_fcoe_tso()
1221 skb_shinfo(skb)->gso_type); in i40e_fcoe_tso()
1243 cd_mss = skb_shinfo(skb)->gso_size; in i40e_fcoe_tso()
/linux-4.4.14/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c283 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_map_tx()
284 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_map_tx()
296 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in ipoib_dma_map_tx()
322 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_unmap_tx()
323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_unmap_tx()
526 priv->tx_wr.mss = skb_shinfo(skb)->gso_size; in post_send()
Dipoib.h516 skb_frag_t *frags = skb_shinfo(skb)->frags; in ipoib_build_sge()
517 int nr_frags = skb_shinfo(skb)->nr_frags; in ipoib_build_sge()
Dipoib_cm.c537 num_frags = skb_shinfo(skb)->nr_frags; in skb_put_frags()
539 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags()
545 --skb_shinfo(skb)->nr_frags; in skb_put_frags()
/linux-4.4.14/drivers/target/tcm_fc/
Dtfc_io.c147 skb_shinfo(fp_skb(fp))->nr_frags, in ft_queue_data_in()
171 (skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN)) in ft_queue_data_in()
/linux-4.4.14/drivers/staging/unisys/visornic/
Dvisornic_main.c221 numfrags = skb_shinfo(skb)->nr_frags; in visor_copy_fragsinfo_from_skb()
234 numfrags = skb_shinfo(skb)->nr_frags; in visor_copy_fragsinfo_from_skb()
263 skb_frag_page(&skb_shinfo(skb)->frags[ii])), in visor_copy_fragsinfo_from_skb()
264 skb_shinfo(skb)->frags[ii]. in visor_copy_fragsinfo_from_skb()
266 skb_shinfo(skb)->frags[ii]. in visor_copy_fragsinfo_from_skb()
280 if (skb_shinfo(skb)->frag_list) { in visor_copy_fragsinfo_from_skb()
284 for (skbinlist = skb_shinfo(skb)->frag_list; skbinlist; in visor_copy_fragsinfo_from_skb()
1261 skb_shinfo(skb)->frag_list = curr; in visornic_rx()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c253 nfrags = skb_shinfo(skb)->nr_frags; in unmap_skb()
257 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]), in unmap_skb()
899 skb_fill_page_desc(newskb, skb_shinfo(newskb)->nr_frags, in get_packet_pg()
946 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1) + 2; in calc_tx_descs()
947 if (skb_shinfo(skb)->gso_size) in calc_tx_descs()
978 nfrags = skb_shinfo(skb)->nr_frags; in make_sgl()
980 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in make_sgl()
1154 tso_info = V_LSO_MSS(skb_shinfo(skb)->gso_size); in write_tx_pkt_wr()
1283 if (skb_shinfo(skb)->gso_size) in t3_eth_xmit()
1544 si = skb_shinfo(skb); in deferred_unmap_destructor()
[all …]
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/
Dlio_main.c1389 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf()
1397 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf()
1437 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf_with_resp()
1445 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf_with_resp()
2529 if (unlikely((skb_shinfo(skb)->tx_flags | SKBTX_IN_PROGRESS) != 0)) { in handle_timestamp()
2791 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) { in liquidio_xmit()
2792 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in liquidio_xmit()
2796 if (skb_shinfo(skb)->nr_frags == 0) { in liquidio_xmit()
2830 cmdsetup.s.u.gatherptrs = (skb_shinfo(skb)->nr_frags + 1); in liquidio_xmit()
2846 frags = skb_shinfo(skb)->nr_frags; in liquidio_xmit()
[all …]
/linux-4.4.14/include/net/
Dudp_tunnel.h114 skb_shinfo(skb)->gso_type |= uh->check ? in udp_tunnel_gro_complete()
Dudp.h146 for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) { in udp_csum()
Dsch_generic.h519 bstats->packets += skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 1; in bstats_update()
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/
Dxgbe-drv.c1167 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in xgbe_prep_tx_tstamp()
1197 packet->mss = skb_shinfo(skb)->gso_size; in xgbe_prep_tso()
1206 packet->tx_packets = skb_shinfo(skb)->gso_segs; in xgbe_prep_tso()
1244 if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { in xgbe_packet_info()
1273 if ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in xgbe_packet_info()
1283 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_packet_info()
1284 frag = &skb_shinfo(skb)->frags[i]; in xgbe_packet_info()
1792 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xgbe_create_skb()
1968 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xgbe_rx_poll()
Dxgbe-desc.c571 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_map_tx_skb()
575 frag = &skb_shinfo(skb)->frags[i]; in xgbe_map_tx_skb()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb/
Dsge.c1117 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in compute_large_page_tx_descs()
1124 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in compute_large_page_tx_descs()
1208 nfrags = skb_shinfo(skb)->nr_frags; in write_tx_descs()
1256 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in write_tx_descs()
1320 count = 1 + skb_shinfo(skb)->nr_frags; in restart_sched()
1674 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx()
1711 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx()
1796 if (skb_shinfo(skb)->gso_size) { in t1_start_xmit()
1811 skb_shinfo(skb)->gso_size)); in t1_start_xmit()
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_cmn.c524 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in bnx2x_set_gro_params()
527 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in bnx2x_set_gro_params()
538 skb_shinfo(skb)->gso_size = len_on_bd - hdrs_len; in bnx2x_set_gro_params()
741 if (skb_shinfo(skb)->gso_size) { in bnx2x_gro_receive()
3452 if (skb_shinfo(skb)->nr_frags >= (MAX_FETCH_BD - num_tso_win_sub)) { in bnx2x_pkt_req_lin()
3454 unsigned short lso_mss = skb_shinfo(skb)->gso_size; in bnx2x_pkt_req_lin()
3457 int num_wnds = skb_shinfo(skb)->nr_frags - wnd_size; in bnx2x_pkt_req_lin()
3479 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]); in bnx2x_pkt_req_lin()
3495 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx + wnd_size - 1]); in bnx2x_pkt_req_lin()
3502 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx]); in bnx2x_pkt_req_lin()
[all …]
/linux-4.4.14/drivers/net/ethernet/apm/xgene/
Dxgene_enet_main.c165 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgene_enet_tx_completion()
166 frag = &skb_shinfo(skb)->frags[i]; in xgene_enet_tx_completion()
219 mss = skb_shinfo(skb)->gso_size; in xgene_enet_work_msg()
223 nr_frags = skb_shinfo(skb)->nr_frags; in xgene_enet_work_msg()
226 len += skb_shinfo(skb)->frags[i].size; in xgene_enet_work_msg()
338 nr_frags = skb_shinfo(skb)->nr_frags; in xgene_enet_setup_tx_desc()
346 frag = &skb_shinfo(skb)->frags[fidx]; in xgene_enet_setup_tx_desc()
/linux-4.4.14/drivers/net/vmxnet3/
Dvmxnet3_drv.c655 struct skb_frag_struct *frag = skb_shinfo(skb)->frags + in vmxnet3_append_frag()
656 skb_shinfo(skb)->nr_frags; in vmxnet3_append_frag()
658 BUG_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS); in vmxnet3_append_frag()
665 skb_shinfo(skb)->nr_frags++; in vmxnet3_append_frag()
752 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vmxnet3_map_pkt()
753 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in vmxnet3_map_pkt()
933 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in txd_estimate()
934 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in txd_estimate()
973 ctx.mss = skb_shinfo(skb)->gso_size; in vmxnet3_tq_xmit()
1424 skb_shinfo(skb)->gso_type = rcd->v4 ? in vmxnet3_rq_rx_complete()
[all …]
/linux-4.4.14/drivers/net/hyperv/
Dnetvsc_drv.c328 int frags = skb_shinfo(skb)->nr_frags; in init_page_array()
349 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in init_page_array()
360 int i, frags = skb_shinfo(skb)->nr_frags; in count_skb_frag_slots()
364 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in count_skb_frag_slots()
613 lso_info->lso_v2_transmit.mss = skb_shinfo(skb)->gso_size; in netvsc_start_xmit()
/linux-4.4.14/net/sctp/
Dulpqueue.c335 struct sk_buff *list = skb_shinfo(f_frag)->frag_list; in sctp_make_reassembled_event()
365 skb_shinfo(new)->frag_list = pos; in sctp_make_reassembled_event()
367 skb_shinfo(f_frag)->frag_list = pos; in sctp_make_reassembled_event()
997 flist = skb_shinfo(skb)->frag_list; in sctp_ulpq_renege_list()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/
Dixgbe_fcoe.c522 if (skb_is_gso(skb) && (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE)) { in ixgbe_fso()
524 skb_shinfo(skb)->gso_type); in ixgbe_fso()
593 skb_shinfo(skb)->gso_size); in ixgbe_fso()
602 mss_l4len_idx = skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_fso()
Dixgbe_main.c1602 skb_shinfo(skb)->gso_size = DIV_ROUND_UP((skb->len - hdr_len), in ixgbe_set_rsc_gso_size()
1604 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in ixgbe_set_rsc_gso_size()
1735 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_pull_tail()
1781 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_dma_sync_frag()
1919 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbe_add_rx_frag()
6984 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbe_tso()
6989 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_tso()
7219 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in ixgbe_tx_map()
7510 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbe_xmit_frame_ring()
7511 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbe_xmit_frame_ring()
[all …]
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c544 struct skb_frag_struct *skb_frags_rx = skb_shinfo(skb)->frags; in mlx4_en_complete_rx_desc()
623 skb_shinfo(skb)->nr_frags = used_frags; in mlx4_en_rx_skb()
631 skb_shinfo(skb)->frags[0].page_offset += pull_len; in mlx4_en_rx_skb()
634 skb_frag_size_sub(&skb_shinfo(skb)->frags[0], pull_len); in mlx4_en_rx_skb()
891 void *va = skb_frag_address(skb_shinfo(gro_skb)->frags); in mlx4_en_process_rx_cq()
900 skb_shinfo(gro_skb)->nr_frags = nr; in mlx4_en_process_rx_cq()
Den_tx.c709 struct skb_shared_info *shinfo = skb_shinfo(skb); in mlx4_en_xmit()
/linux-4.4.14/drivers/staging/rdma/amso1100/
Dc2.c769 if (unlikely(c2_port->tx_avail < (skb_shinfo(skb)->nr_frags + 1))) { in c2_xmit_frame()
799 if (skb_shinfo(skb)->nr_frags) { in c2_xmit_frame()
800 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in c2_xmit_frame()
801 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in c2_xmit_frame()
824 c2_port->tx_avail -= (skb_shinfo(skb)->nr_frags + 1); in c2_xmit_frame()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c287 si = skb_shinfo(skb); in map_skb()
310 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl()
878 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1); in calc_tx_flits()
879 if (skb_shinfo(skb)->gso_size) in calc_tx_flits()
912 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl()
1193 if (!skb_shinfo(skb)->gso_size && (unlikely(skb->len > max_pkt_len))) in t4vf_eth_xmit()
1279 ssi = skb_shinfo(skb); in t4vf_eth_xmit()
1483 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
/linux-4.4.14/drivers/net/usb/
Dcdc-phonet.c166 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete()
172 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete()
Dusbnet.c1275 num_sgs = skb_shinfo(skb)->nr_frags + 1; in build_dma_sg()
1291 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in build_dma_sg()
1292 struct skb_frag_struct *f = &skb_shinfo(skb)->frags[i]; in build_dma_sg()
Dr8152.c1385 if (skb_shinfo(skb)->gso_size) { in r8152_csum_workaround()
1466 u32 mss = skb_shinfo(skb)->gso_size; in r8152_tx_csum()
1984 u32 mss = skb_shinfo(skb)->gso_size; in rtl8152_features_check()
Dax88179_178a.c1428 int mss = skb_shinfo(skb)->gso_size; in ax88179_tx_fixup()
/linux-4.4.14/drivers/net/wireless/ath/wil6210/
Dtxrx.c1228 int nr_frags = skb_shinfo(skb)->nr_frags; in __wil_tx_vring_tso()
1230 int mss = skb_shinfo(skb)->gso_size; /* payload size w/o headers */ in __wil_tx_vring_tso()
1270 gso_type = skb_shinfo(skb)->gso_type & (SKB_GSO_TCPV6 | SKB_GSO_TCPV4); in __wil_tx_vring_tso()
1326 frag = &skb_shinfo(skb)->frags[f]; in __wil_tx_vring_tso()
1500 int nr_frags = skb_shinfo(skb)->nr_frags; in __wil_tx_vring()
1554 &skb_shinfo(skb)->frags[f]; in __wil_tx_vring()
1722 (skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS); in wil_need_txstat()
Ddebugfs.c870 int nr_frags = skb_shinfo(skb)->nr_frags; in wil_seq_print_skb()
879 &skb_shinfo(skb)->frags[i]; in wil_seq_print_skb()
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/
Dfs_enet-main.c528 nr_frags = skb_shinfo(skb)->nr_frags; in fs_enet_start_xmit()
529 frag = skb_shinfo(skb)->frags; in fs_enet_start_xmit()
558 nr_frags = skb_shinfo(skb)->nr_frags; in fs_enet_start_xmit()
586 frag = skb_shinfo(skb)->frags; in fs_enet_start_xmit()
/linux-4.4.14/drivers/net/ethernet/realtek/
D8139cp.c727 this_frag = &skb_shinfo(skb)->frags[frag]; in unwind_tx_frag_mapping()
746 if (TX_BUFFS_AVAIL(cp) <= (skb_shinfo(skb)->nr_frags + 1)) { in cp_start_xmit()
755 mss = skb_shinfo(skb)->gso_size; in cp_start_xmit()
780 if (skb_shinfo(skb)->nr_frags == 0) { in cp_start_xmit()
821 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in cp_start_xmit()
822 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in cp_start_xmit()
841 if (frag == skb_shinfo(skb)->nr_frags - 1) in cp_start_xmit()
1865 if (skb_shinfo(skb)->gso_size > MSSMask) in cp_features_check()
Dr8169.c6879 struct skb_shared_info *info = skb_shinfo(skb); in rtl8169_xmit_frags()
6941 if (skb_shinfo(skb)->gso_size) { in r8169_csum_workaround()
7011 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v1()
7034 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v2()
7121 if (unlikely(!TX_FRAGS_READY_FOR(tp, skb_shinfo(skb)->nr_frags))) { in rtl8169_start_xmit()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c262 si = skb_shinfo(skb); in map_skb()
291 si = skb_shinfo(skb); in unmap_skb()
315 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl()
773 int hdrlen = skb_shinfo(skb)->gso_size ? in is_eth_imm()
810 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1); in calc_tx_flits()
811 if (skb_shinfo(skb)->gso_size) in calc_tx_flits()
856 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl()
1170 if (!skb_shinfo(skb)->gso_size && (unlikely(skb->len > max_pkt_len))) in t4_eth_xmit()
1220 ssi = skb_shinfo(skb); in t4_eth_xmit()
1504 cnt = skb_shinfo(skb)->nr_frags; in calc_tx_flits_ofld()
[all …]
/linux-4.4.14/drivers/block/aoe/
Daoedev.c405 while (atomic_read(&skb_shinfo(skb)->dataref) != 1 && i-- > 0) in skbfree()
415 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in skbfree()
Daoecmd.c181 if (skb && atomic_read(&skb_shinfo(skb)->dataref) == 1) { in skb_pool_get()
235 if (atomic_read(&skb_shinfo(skb)->dataref) != 1) { in newtframe()
244 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in newtframe()
/linux-4.4.14/drivers/scsi/fcoe/
Dfcoe_transport.c308 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in fcoe_fc_crc()
309 frag = &skb_shinfo(skb)->frags[i]; in fcoe_fc_crc()
471 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, in fcoe_get_paged_crc_eof()
Dfcoe.c1644 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in fcoe_xmit()
1697 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in fcoe_xmit()
1698 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in fcoe_xmit()
1700 skb_shinfo(skb)->gso_type = 0; in fcoe_xmit()
1701 skb_shinfo(skb)->gso_size = 0; in fcoe_xmit()
/linux-4.4.14/drivers/net/ethernet/ibm/
Dibmveth.c1037 if (skb_shinfo(skb)->nr_frags > 5 && __skb_linearize(skb)) { in ibmveth_start_xmit()
1108 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ibmveth_start_xmit()
1109 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ibmveth_start_xmit()
1123 mss = (unsigned long)skb_shinfo(skb)->gso_size; in ibmveth_start_xmit()
1132 cpu_to_be16(skb_shinfo(skb)->gso_size); in ibmveth_start_xmit()
1150 for (i = 1; i < skb_shinfo(skb)->nr_frags + 1; i++) in ibmveth_start_xmit()
/linux-4.4.14/drivers/hsi/clients/
Dssi_protocol.c186 BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); in ssip_skb_to_msg()
190 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ssip_skb_to_msg()
193 frag = &skb_shinfo(skb)->frags[i]; in ssip_skb_to_msg()
215 msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags); in ssip_alloc_data()
/linux-4.4.14/drivers/atm/
Deni.c1101 DPRINTK("iovcnt = %d\n",skb_shinfo(skb)->nr_frags); in do_tx()
1102 if (!skb_shinfo(skb)->nr_frags) dma_size += 5; in do_tx()
1103 else dma_size += 5*(skb_shinfo(skb)->nr_frags+1); in do_tx()
1124 if (!skb_shinfo(skb)->nr_frags) in do_tx()
1129 for (i = -1; i < skb_shinfo(skb)->nr_frags; i++) in do_tx()
1136 skb_frag_page(&skb_shinfo(skb)->frags[i]) + in do_tx()
1137 skb_shinfo(skb)->frags[i].page_offset, in do_tx()
1138 skb_frag_size(&skb_shinfo(skb)->frags[i])); in do_tx()
Dhe.c2514 if (skb_shinfo(skb)->nr_frags) { in he_send()
2558 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in he_send()
2559 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in he_send()
Dzatm.c642 EVENT("iovcnt=%d\n",skb_shinfo(skb)->nr_frags,0); in do_tx()
644 if (!skb_shinfo(skb)->nr_frags) { in do_tx()
/linux-4.4.14/drivers/net/ethernet/neterion/
Ds2io.h1137 #define s2io_tcp_mss(skb) skb_shinfo(skb)->gso_size
1138 #define s2io_udp_mss(skb) skb_shinfo(skb)->gso_size
1139 #define s2io_offload_type(skb) skb_shinfo(skb)->gso_type
Ds2io.c2340 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_txdl_getskb()
2344 const skb_frag_t *frag = &skb_shinfo(skb)->frags[j]; in s2io_txdl_getskb()
4154 (__force u64)skb_shinfo(skb)->ip6_frag_id; in s2io_xmit()
4157 (__force u64)skb_shinfo(skb)->ip6_frag_id << 32; in s2io_xmit()
4179 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_xmit()
4182 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in s2io_xmit()
8548 if (skb_shinfo(first)->frag_list) in lro_append_pkt()
8551 skb_shinfo(first)->frag_list = skb; in lro_append_pkt()
/linux-4.4.14/Documentation/networking/
Ddriver.txt21 if (TX_BUFFS_AVAIL(dp) <= (skb_shinfo(skb)->nr_frags + 1)) {
Dtimestamping.txt225 relevant sequence number in skb_shinfo(skb)->tskey. Because an skbuff
439 - In hard_start_xmit(), check if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
444 SKBTX_IN_PROGRESS in skb_shinfo(skb)->tx_flags , e.g. with
446 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
Dnetdev-features.txt124 NETIF_F_SG --- paged skbs (skb_shinfo()->frags), NETIF_F_FRAGLIST ---
/linux-4.4.14/drivers/net/ethernet/freescale/
Dgianfar.c1968 for (j = 0; j < skb_shinfo(tx_queue->tx_skbuff[i])->nr_frags; in free_skb_tx_queue()
2341 do_tstamp = (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in gfar_start_xmit()
2369 nr_frags = skb_shinfo(skb)->nr_frags; in gfar_start_xmit()
2416 frag_len = skb_shinfo(skb)->frags[i].size; in gfar_start_xmit()
2426 &skb_shinfo(skb)->frags[i], in gfar_start_xmit()
2477 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in gfar_start_xmit()
2681 frags = skb_shinfo(skb)->nr_frags; in gfar_clean_tx_ring()
2686 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in gfar_clean_tx_ring()
2700 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring()
2710 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring()
[all …]
Dfec_main.c375 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_frag_skb()
387 this_frag = &skb_shinfo(skb)->frags[frag]; in fec_enet_txq_submit_frag_skb()
394 frag_len = skb_shinfo(skb)->frags[frag].size; in fec_enet_txq_submit_frag_skb()
401 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_frag_skb()
457 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_skb()
520 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_skb()
530 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in fec_enet_txq_submit_skb()
532 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in fec_enet_txq_submit_skb()
724 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in fec_enet_txq_submit_tso()
1261 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS) && in fec_enet_tx_queue()
/linux-4.4.14/drivers/net/ethernet/emulex/benet/
Dbe_main.c684 u64 tx_pkts = skb_shinfo(skb)->gso_segs ? : 1; in be_tx_stats_update()
699 return 1 + (skb_headlen(skb) ? 1 : 0) + skb_shinfo(skb)->nr_frags; in skb_wrb_cnt()
773 wrb_params->lso_mss = skb_shinfo(skb)->gso_size; in be_get_wrb_params_from_skb()
950 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in be_xmit_enqueue()
951 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in be_xmit_enqueue()
2055 skb_shinfo(skb)->nr_frags = 1; in skb_fill_rx_data()
2057 skb_shinfo(skb)->frags[0].page_offset = in skb_fill_rx_data()
2059 skb_frag_size_set(&skb_shinfo(skb)->frags[0], in skb_fill_rx_data()
2083 skb_shinfo(skb)->frags[j].page_offset = in skb_fill_rx_data()
2085 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); in skb_fill_rx_data()
[all …]
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c1291 int nr_frags = skb_shinfo(skb)->nr_frags; in sxgbe_xmit()
1294 u16 cur_mss = skb_shinfo(skb)->gso_size; in sxgbe_xmit()
1304 ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit()
1369 const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num]; in sxgbe_xmit()
1428 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit()
1431 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in sxgbe_xmit()
/linux-4.4.14/drivers/net/ethernet/atheros/atl1e/
Datl1e_main.c1617 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atl1e_cal_tdp_req()
1618 fg_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in atl1e_cal_tdp_req()
1624 (skb_shinfo(skb)->gso_type == SKB_GSO_TCPV6)) { in atl1e_cal_tdp_req()
1652 offload_type = skb_shinfo(skb)->gso_type; in atl1e_tso_csum()
1680 tpd->word3 |= ((skb_shinfo(skb)->gso_size) & in atl1e_tso_csum()
1725 nr_frags = skb_shinfo(skb)->nr_frags; in atl1e_tx_map()
1794 frag = &skb_shinfo(skb)->frags[f]; in atl1e_tx_map()
/linux-4.4.14/drivers/net/ethernet/intel/igbvf/
Dnetdev.c318 if (!skb_shinfo(skb)->nr_frags) { in igbvf_clean_rx_irq()
332 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in igbvf_clean_rx_irq()
830 segs = skb_shinfo(skb)->gso_segs ?: 1; in igbvf_clean_tx_irq()
1973 mss_l4len_idx = (skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT); in igbvf_tso()
2103 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in igbvf_tx_map_adv()
2111 frag = &skb_shinfo(skb)->frags[f]; in igbvf_tx_map_adv()
2242 if (igbvf_maybe_stop_tx(netdev, skb_shinfo(skb)->nr_frags + 4)) { in igbvf_xmit_frame_ring_adv()
/linux-4.4.14/net/packet/
Daf_packet.c1934 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_sendmsg_spkt()
2323 ph = skb_shinfo(skb)->destructor_arg; in tpacket_destruct_skb()
2359 sock_tx_timestamp(&po->sk, &skb_shinfo(skb)->tx_flags); in tpacket_fill_skb()
2360 skb_shinfo(skb)->destructor_arg = ph.raw; in tpacket_fill_skb()
2440 nr_frags = skb_shinfo(skb)->nr_frags; in tpacket_fill_skb()
2771 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_snd()
2796 skb_shinfo(skb)->gso_size = in packet_snd()
2798 skb_shinfo(skb)->gso_type = gso_type; in packet_snd()
2801 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in packet_snd()
2802 skb_shinfo(skb)->gso_segs = 0; in packet_snd()
[all …]
/linux-4.4.14/drivers/net/ethernet/pasemi/
Dpasemi_mac.c302 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in pasemi_mac_unmap_tx_skb()
573 nfrags = skb_shinfo(info->skb)->nr_frags; in pasemi_mac_free_tx_resources()
1398 const int nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_queue_csdesc()
1494 nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_start_tx()
1503 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in pasemi_mac_start_tx()
/linux-4.4.14/drivers/net/ethernet/tile/
Dtilegx.c461 struct skb_shared_info *shtx = skb_shinfo(skb); in tile_tx_timestamp()
1678 struct skb_shared_info *sh = skb_shinfo(skb); in tso_count_edescs()
1724 struct skb_shared_info *sh = skb_shinfo(skb); in tso_headers_prepare()
1833 struct skb_shared_info *sh = skb_shinfo(skb); in tso_egress()
1981 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags()
Dtilepro.c1604 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags()
1677 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_tso()
1851 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx()
/linux-4.4.14/drivers/net/ethernet/synopsys/
Ddwc_eth_qos.c1987 if (skb_is_gso(skb) && skb_shinfo(skb)->gso_size != lp->gso_size) in dwceqos_tx_prepare()
1990 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in dwceqos_tx_prepare()
1991 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in dwceqos_tx_prepare()
2015 if (skb_is_gso(skb) && skb_shinfo(skb)->gso_size != lp->gso_size) { in dwceqos_tx_linear()
2016 dwceqos_dmadesc_set_ctx(lp, skb_shinfo(skb)->gso_size); in dwceqos_tx_linear()
2017 lp->gso_size = skb_shinfo(skb)->gso_size; in dwceqos_tx_linear()
2083 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in dwceqos_tx_frags()
2084 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in dwceqos_tx_frags()
/linux-4.4.14/drivers/net/ethernet/marvell/
Dsky2.c1215 for (i = 0; i < skb_shinfo(re->skb)->nr_frags; i++) in sky2_rx_submit()
1232 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_rx_map_skb()
1233 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_rx_map_skb()
1247 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_map_skb()
1269 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) in sky2_rx_unmap_skb()
1271 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_unmap_skb()
1807 count = (skb_shinfo(skb)->nr_frags + 1) in tx_le_req()
1877 mss = skb_shinfo(skb)->gso_size; in sky2_xmit_frame()
1949 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_xmit_frame()
1950 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_xmit_frame()
[all …]
Dmv643xx_eth.c675 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in has_tiny_unaligned_frags()
676 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in has_tiny_unaligned_frags()
863 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in txq_submit_tso()
913 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_frag_skb()
921 this_frag = &skb_shinfo(skb)->frags[frag]; in txq_submit_frag_skb()
952 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_skb()
Dmvneta.c1698 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in mvneta_tx_tso()
1748 int i, nr_frags = skb_shinfo(skb)->nr_frags; in mvneta_tx_frag_process()
1751 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvneta_tx_frag_process()
1816 frags = skb_shinfo(skb)->nr_frags + 1; in mvneta_tx()
Dskge.c2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame()
2780 if (!skb_shinfo(skb)->nr_frags) /* single buffer i.e. no fragments */ in skge_xmit_frame()
2786 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skge_xmit_frame()
2787 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skge_xmit_frame()
Dmvpp2.c5207 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mvpp2_tx_frag_process()
5208 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvpp2_tx_frag_process()
5226 if (i == (skb_shinfo(skb)->nr_frags - 1)) { in mvpp2_tx_frag_process()
5268 frags = skb_shinfo(skb)->nr_frags + 1; in mvpp2_tx()
/linux-4.4.14/drivers/net/ethernet/adi/
Dbfin_mac.c836 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) { in bfin_tx_hwtstamp()
842 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in bfin_tx_hwtstamp()
1159 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit()
1169 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit()
/linux-4.4.14/drivers/scsi/bnx2fc/
Dbnx2fc_fcoe.c329 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in bnx2fc_xmit()
373 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in bnx2fc_xmit()
374 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in bnx2fc_xmit()
376 skb_shinfo(skb)->gso_type = 0; in bnx2fc_xmit()
377 skb_shinfo(skb)->gso_size = 0; in bnx2fc_xmit()
/linux-4.4.14/drivers/net/ethernet/qlogic/qede/
Dqede_main.c239 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++, bds_consumed++) { in qede_free_tx_pkt()
485 cpu_to_le16(skb_shinfo(skb)->gso_size); in qede_start_xmit()
527 while (tx_data_bd && frag_idx < skb_shinfo(skb)->nr_frags) { in qede_start_xmit()
529 &skb_shinfo(skb)->frags[frag_idx], in qede_start_xmit()
546 for (; frag_idx < skb_shinfo(skb)->nr_frags; frag_idx++, nbd++) { in qede_start_xmit()
553 &skb_shinfo(skb)->frags[frag_idx], in qede_start_xmit()
/linux-4.4.14/net/openvswitch/
Dflow.c528 skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract()
649 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract()
/linux-4.4.14/drivers/net/ethernet/ti/
Dnetcp_core.c686 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in netcp_process_one_rx_packet()
1047 prefetchw(skb_shinfo(skb)); in netcp_tx_map_skb()
1056 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in netcp_tx_map_skb()
1057 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in netcp_tx_map_skb()
1096 if (skb_shinfo(skb)->frag_list) { in netcp_tx_map_skb()
Dcpts.c343 if (!(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in cpts_tx_timestamp()
Dcpsw.c1400 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in cpsw_ndo_start_xmit()
1402 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cpsw_ndo_start_xmit()
/linux-4.4.14/drivers/net/ethernet/tehuti/
Dtehuti.c1503 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_map_skb()
1520 frag = &skb_shinfo(skb)->frags[i]; in bdx_tx_map_skb()
1626 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_transmit()
1646 if (skb_shinfo(skb)->gso_size) { in bdx_tx_transmit()
1647 txd_mss = skb_shinfo(skb)->gso_size; in bdx_tx_transmit()
/linux-4.4.14/drivers/net/ethernet/qlogic/netxen/
Dnetxen_nic_main.c1902 skb_shinfo(skb)->gso_size > 0) { in netxen_tso_check()
1906 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in netxen_tso_check()
2006 nr_frags = skb_shinfo(skb)->nr_frags; in netxen_map_tx_skb()
2018 frag = &skb_shinfo(skb)->frags[i]; in netxen_map_tx_skb()
2071 frag_count = skb_shinfo(skb)->nr_frags + 1; in netxen_nic_xmit_frame()
2079 frag = &skb_shinfo(skb)->frags[i]; in netxen_nic_xmit_frame()
2086 frag_count = 1 + skb_shinfo(skb)->nr_frags; in netxen_nic_xmit_frame()
/linux-4.4.14/drivers/net/ethernet/broadcom/bnxt/
Dbnxt.c192 if (unlikely(free_size < skb_shinfo(skb)->nr_frags + 2)) { in bnxt_start_xmit()
199 last_frag = skb_shinfo(skb)->nr_frags; in bnxt_start_xmit()
249 skb_frag_t *frag = &skb_shinfo(skb)->frags[j]; in bnxt_start_xmit()
323 length = skb_shinfo(skb)->gso_size; in bnxt_start_xmit()
339 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in bnxt_start_xmit()
410 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnxt_start_xmit()
453 skb_frag_size(&skb_shinfo(skb)->frags[j]), in bnxt_tx_int()
711 shinfo = skb_shinfo(skb); in bnxt_rx_pages()
861 skb_shinfo(skb)->gso_size = in bnxt_gro_skb()
863 skb_shinfo(skb)->gso_type = tpa_info->gso_type; in bnxt_gro_skb()
[all …]
/linux-4.4.14/drivers/net/ethernet/atheros/atl1c/
Datl1c_main.c1966 tpd_req = skb_shinfo(skb)->nr_frags + 1; in atl1c_cal_tpd_req()
1972 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in atl1c_cal_tpd_req()
1995 offload_type = skb_shinfo(skb)->gso_type; in atl1c_tso_csum()
2050 (*tpd)->word1 |= (skb_shinfo(skb)->gso_size & TPD_MSS_MASK) << in atl1c_tso_csum()
2114 nr_frags = skb_shinfo(skb)->nr_frags; in atl1c_tx_map()
2164 frag = &skb_shinfo(skb)->frags[f]; in atl1c_tx_map()
/linux-4.4.14/drivers/net/ethernet/cisco/enic/
Denic_main.c433 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_cont()
559 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_tso()
587 unsigned int mss = skb_shinfo(skb)->gso_size; in enic_queue_wq_skb()
652 if (skb_shinfo(skb)->gso_size == 0 && in enic_hard_start_xmit()
653 skb_shinfo(skb)->nr_frags + 1 > ENIC_NON_TSO_MAX_DESC && in enic_hard_start_xmit()
662 skb_shinfo(skb)->nr_frags + ENIC_DESC_MAX_SPLITS) { in enic_hard_start_xmit()
/linux-4.4.14/drivers/net/ethernet/3com/
Dtyphoon.c325 #define skb_tso_size(x) (skb_shinfo(x)->gso_size)
740 numDesc = skb_shinfo(skb)->nr_frags + 1; in typhoon_start_tx()
793 if(skb_shinfo(skb)->nr_frags == 0) { in typhoon_start_tx()
813 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in typhoon_start_tx()
814 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in typhoon_start_tx()
D3c59x.c2170 if (!skb_shinfo(skb)->nr_frags) { in boomerang_start_xmit()
2189 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in boomerang_start_xmit()
2190 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in boomerang_start_xmit()
2214 if (i == skb_shinfo(skb)->nr_frags-1) in boomerang_start_xmit()
2462 for (i=0; i<=skb_shinfo(skb)->nr_frags; i++) in boomerang_interrupt()
2795 for (k=0; k<=skb_shinfo(skb)->nr_frags; k++) in vortex_close()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbevf/
Dixgbevf_main.c812 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbevf_add_rx_frag()
3329 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbevf_tso()
3334 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbevf_tso()
3465 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbevf_tx_map()
3640 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbevf_xmit_frame()
3641 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbevf_xmit_frame()
3643 count += skb_shinfo(skb)->nr_frags; in ixgbevf_xmit_frame()
/linux-4.4.14/drivers/net/ethernet/atheros/atlx/
Datl1.c2125 if (skb_shinfo(skb)->gso_size) { in atl1_tso()
2170 ptpd->word3 |= (skb_shinfo(skb)->gso_size & in atl1_tso()
2220 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_tx_map()
2282 frag = &skb_shinfo(skb)->frags[f]; in atl1_tx_map()
2381 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_xmit_frame()
2383 unsigned int f_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in atl1_xmit_frame()
2388 mss = skb_shinfo(skb)->gso_size; in atl1_xmit_frame()
/linux-4.4.14/drivers/net/ethernet/adaptec/
Dstarfire.c162 #define skb_num_frags(skb) (skb_shinfo(skb)->nr_frags + 1)
1230 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[i - 1]; in start_tx()
1349 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in intr_handler()
1352 skb_frag_size(&skb_shinfo(skb)->frags[i]), in intr_handler()
/linux-4.4.14/drivers/net/ethernet/
Djme.c1981 nr_alloc = skb_shinfo(skb)->nr_frags + 2; in jme_alloc_txdesc()
2059 int i, nr_frags = skb_shinfo(skb)->nr_frags; in jme_map_tx_skb()
2066 frag = &skb_shinfo(skb)->frags[i]; in jme_map_tx_skb()
2097 *mss = cpu_to_le16(skb_shinfo(skb)->gso_size << TXDESC_MSS_SHIFT); in jme_tx_tso()
2207 txbi->nr_desc = skb_shinfo(skb)->nr_frags + 2; in jme_fill_tx_desc()
2282 idx, skb_shinfo(skb)->nr_frags + 2, jiffies); in jme_start_xmit()
/linux-4.4.14/drivers/net/ethernet/atheros/alx/
Dmain.c1026 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in alx_map_tx_skb()
1029 frag = &skb_shinfo(skb)->frags[f]; in alx_map_tx_skb()
1074 int tpdreq = skb_shinfo(skb)->nr_frags + 1; in alx_start_xmit()
/linux-4.4.14/drivers/net/ethernet/intel/ixgb/
Dixgb_main.c1235 mss = skb_shinfo(skb)->gso_size; in ixgb_tso()
1331 unsigned int mss = skb_shinfo(skb)->gso_size; in ixgb_tx_map()
1332 unsigned int nr_frags = skb_shinfo(skb)->nr_frags; in ixgb_tx_map()
1369 frag = &skb_shinfo(skb)->frags[f]; in ixgb_tx_map()
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/
Dstmmac_main.c355 if (likely(!skb || !(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS))) in stmmac_get_tx_hwtstamp()
1952 int nfrags = skb_shinfo(skb)->nr_frags; in stmmac_xmit()
2003 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in stmmac_xmit()
2073 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in stmmac_xmit()
2076 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in stmmac_xmit()
/linux-4.4.14/drivers/net/ethernet/intel/e1000/
De1000_main.c2719 mss = skb_shinfo(skb)->gso_size; in e1000_tso()
2903 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map()
2948 segs = skb_shinfo(skb)->gso_segs ?: 1; in e1000_tx_map()
3136 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame()
3202 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame()
3204 count += TXD_USE_COUNT(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame()
4208 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq()
4217 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq()
/linux-4.4.14/drivers/net/ethernet/xilinx/
Dll_temac_main.c685 num_frag = skb_shinfo(skb)->nr_frags; in temac_start_xmit()
686 frag = &skb_shinfo(skb)->frags[0]; in temac_start_xmit()
Dxilinx_axienet_main.c666 num_frag = skb_shinfo(skb)->nr_frags; in axienet_start_xmit()
698 frag = &skb_shinfo(skb)->frags[ii]; in axienet_start_xmit()
/linux-4.4.14/drivers/net/ethernet/hisilicon/hns/
Dhns_enet.c121 buf_num = skb_shinfo(skb)->nr_frags + 1; in hns_nic_net_xmit_hw()
139 assert(skb_shinfo(skb)->nr_frags == 1); in hns_nic_net_xmit_hw()
159 frag = &skb_shinfo(skb)->frags[i - 1]; in hns_nic_net_xmit_hw()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlxsw/
Dpci.c1528 if (skb_shinfo(skb)->nr_frags > MLXSW_PCI_WQE_SG_ENTRIES - 1) { in mlxsw_pci_skb_transmit()
1554 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlxsw_pci_skb_transmit()
1555 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mlxsw_pci_skb_transmit()
/linux-4.4.14/drivers/net/ethernet/via/
Dvia-velocity.c1738 if (skb_shinfo(skb)->nr_frags > 0) in velocity_free_tx_buf()
2549 if (skb_shinfo(skb)->nr_frags > 6 && __skb_linearize(skb)) { in velocity_xmit()
2554 pktlen = skb_shinfo(skb)->nr_frags == 0 ? in velocity_xmit()
2580 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in velocity_xmit()
2581 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in velocity_xmit()
/linux-4.4.14/drivers/scsi/cxgbi/
Dlibcxgbi.c1143 int frags = skb_shinfo(skb)->nr_frags + in cxgbi_sock_send_pdus()
1155 csk, skb_shinfo(skb)->nr_frags, skb->len, in cxgbi_sock_send_pdus()
2225 skb_shinfo(skb)->nr_frags = tdata->nr_frags; in cxgbi_conn_init_pdu()
2243 i = skb_shinfo(skb)->nr_frags; in cxgbi_conn_init_pdu()
2244 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in cxgbi_conn_init_pdu()
/linux-4.4.14/drivers/net/wireless/iwlwifi/pcie/
Dtx.c1823 skb_shinfo(skb)->nr_frags > IWL_PCIE_MAX_FRAGS && in iwl_trans_pcie_tx()
1914 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in iwl_trans_pcie_tx()
1915 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iwl_trans_pcie_tx()
/linux-4.4.14/net/tipc/
Dmsg.c164 skb_shinfo(head)->frag_list = frag; in tipc_buf_append()
/linux-4.4.14/drivers/net/ethernet/intel/e1000e/
Dnetdev.c1582 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq()
1594 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq()
5357 mss = skb_shinfo(skb)->gso_size; in e1000_tso()
5501 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map()
5529 segs = skb_shinfo(skb)->gso_segs ? : 1; in e1000_tx_map()
5724 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame()
5756 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame()
5758 count += DIV_ROUND_UP(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame()
5803 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in e1000_xmit_frame()
5806 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in e1000_xmit_frame()
/linux-4.4.14/drivers/net/ethernet/alteon/
Dacenic.c2424 if (!skb_shinfo(skb)->nr_frags) { in ace_start_xmit()
2462 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ace_start_xmit()
2463 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ace_start_xmit()
2479 if (i == skb_shinfo(skb)->nr_frags - 1) { in ace_start_xmit()
/linux-4.4.14/drivers/net/ethernet/cadence/
Dmacb.c1169 unsigned int f, nr_frags = skb_shinfo(skb)->nr_frags; in macb_tx_map()
1201 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in macb_tx_map()
1311 nr_frags = skb_shinfo(skb)->nr_frags; in macb_start_xmit()
1313 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in macb_start_xmit()
/linux-4.4.14/drivers/usb/gadget/function/
Df_phonet.c349 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in pn_rx_complete()
/linux-4.4.14/drivers/net/wireless/b43legacy/
Dpio.c197 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in pio_tx_write_fragment()
Ddma.c1024 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in dma_tx_fragment()
/linux-4.4.14/drivers/net/ethernet/tundra/
Dtsi108_eth.c655 int frags = skb_shinfo(skb)->nr_frags + 1; in tsi108_send_packet()
709 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in tsi108_send_packet()
/linux-4.4.14/drivers/net/ethernet/broadcom/
Dbnx2.c2901 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_tx_int()
2960 shinfo = skb_shinfo(skb); in bnx2_reuse_rx_skb_pages()
3087 &skb_shinfo(skb)->frags[i - 1]; in bnx2_rx_skb()
5454 skb_frag_size(&skb_shinfo(skb)->frags[k]), in bnx2_free_tx_skbs()
6601 (skb_shinfo(skb)->nr_frags + 1))) { in bnx2_start_xmit()
6621 if ((mss = skb_shinfo(skb)->gso_size)) { in bnx2_start_xmit()
6629 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) { in bnx2_start_xmit()
6672 last_frag = skb_shinfo(skb)->nr_frags; in bnx2_start_xmit()
6677 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in bnx2_start_xmit()
6746 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_start_xmit()
Dbgmac.c159 nr_frags = skb_shinfo(skb)->nr_frags; in bgmac_dma_tx_add()
183 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bgmac_dma_tx_add()
Dtg3.c6567 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in tg3_tx()
6574 skb_frag_size(&skb_shinfo(skb)->frags[i]), in tg3_tx()
7764 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in tg3_tx_skb_unmap()
7841 return skb_shinfo(skb)->gso_segs < tnapi->tx_pending / 3; in tg3_tso_bug_gso_check()
7853 u32 frag_cnt_est = skb_shinfo(skb)->gso_segs * 3; in tg3_tso_bug()
7917 if (unlikely(budget <= (skb_shinfo(skb)->nr_frags + 1))) { in tg3_start_xmit()
7931 mss = skb_shinfo(skb)->gso_size; in tg3_start_xmit()
8027 if ((unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) && in tg3_start_xmit()
8029 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in tg3_start_xmit()
8049 ((skb_shinfo(skb)->nr_frags == 0) ? TXD_FLAG_END : 0), in tg3_start_xmit()
[all …]
/linux-4.4.14/net/can/
Draw.c758 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_sendmsg()
/linux-4.4.14/net/appletalk/
Dddp.c950 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atalk_sum_skb()
952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in atalk_sum_skb()
/linux-4.4.14/drivers/net/ethernet/intel/igb/
Digb_main.c4691 first->gso_segs = skb_shinfo(skb)->gso_segs; in igb_tso()
4696 mss_l4len_idx |= skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT; in igb_tso()
4887 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in igb_tx_map()
5011 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in igb_xmit_frame_ring()
5012 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in igb_xmit_frame_ring()
5025 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) { in igb_xmit_frame_ring()
5030 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in igb_xmit_frame_ring()
6694 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in igb_add_rx_frag()
/linux-4.4.14/drivers/net/ethernet/nvidia/
Dforcedeth.c2190 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit()
2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit()
2256 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit()
2309 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit()
2337 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit_optimized()
2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit_optimized()
2406 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit_optimized()
2459 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit_optimized()
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/
Dehea_main.c1644 swqe->mss = skb_shinfo(skb)->gso_size; in write_swqe2_immediate()
1676 nfrags = skb_shinfo(skb)->nr_frags; in write_swqe2_data()
1687 frag = &skb_shinfo(skb)->frags[0]; in write_swqe2_data()
1700 frag = &skb_shinfo(skb)->frags[i]; in write_swqe2_data()
/linux-4.4.14/drivers/net/ethernet/myricom/myri10ge/
Dmyri10ge.c1410 frag = skb_shinfo(skb)->frags; in myri10ge_vlan_rx()
1464 rx_frags = skb_shinfo(skb)->frags; in myri10ge_rx_done()
2846 mss = skb_shinfo(skb)->gso_size; in myri10ge_xmit()
2929 frag_cnt = skb_shinfo(skb)->nr_frags; in myri10ge_xmit()
3024 frag = &skb_shinfo(skb)->frags[frag_idx]; in myri10ge_xmit()
/linux-4.4.14/drivers/scsi/libfc/
Dfc_fcp.c651 skb_shinfo(fp_skb(fp))->nr_frags, in fc_fcp_send_data()
672 if ((skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN) && in fc_fcp_send_data()
/linux-4.4.14/drivers/net/ethernet/calxeda/
Dxgmac.c1090 int nfrags = skb_shinfo(skb)->nr_frags; in xgmac_xmit()
1115 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xgmac_xmit()
/linux-4.4.14/net/mac80211/
Dtx.c2261 skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS)) { in ieee80211_build_hdr()
2718 if (skb->sk && skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS) in ieee80211_xmit_fast()
2781 if (skb_shinfo(skb)->gso_size) in ieee80211_xmit_fast()
2783 DIV_ROUND_UP(skb->len, skb_shinfo(skb)->gso_size); in ieee80211_xmit_fast()
/linux-4.4.14/net/wireless/
Dwext-core.c628 skb_shinfo(skb)->frag_list = compskb; in wireless_send_event()
/linux-4.4.14/drivers/net/ethernet/qlogic/qlge/
Dqlge_main.c1318 int frag_cnt = skb_shinfo(skb)->nr_frags; in ql_map_send()
1351 skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_idx]; in ql_map_send()
1519 __skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in ql_process_mac_rx_gro_page()
1527 skb_shinfo(skb)->nr_frags++; in ql_process_mac_rx_gro_page()
2591 mac_iocb_ptr->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in ql_tso()
/linux-4.4.14/drivers/net/ethernet/brocade/bna/
Dbnad.c573 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in bnad_cq_setup_skb_frags()
2844 gso_size = skb_shinfo(skb)->gso_size; in bnad_txq_wi_prepare()
2984 vectors = 1 + skb_shinfo(skb)->nr_frags; in bnad_start_xmit()
3052 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bnad_start_xmit()
/linux-4.4.14/drivers/net/ethernet/renesas/
Dravb_main.c1367 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in ravb_start_xmit()
1404 return (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) ? RAVB_NC : in ravb_select_queue()
/linux-4.4.14/net/netlink/
Daf_netlink.c909 !atomic_dec_return(&(skb_shinfo(skb)->dataref))) in netlink_skb_destructor()
2541 if (unlikely(skb_shinfo(skb)->frag_list)) { in netlink_recvmsg()
2553 data_skb = skb_shinfo(skb)->frag_list; in netlink_recvmsg()
/linux-4.4.14/drivers/net/ethernet/natsemi/
Dns83820.c1074 nr_frags = skb_shinfo(skb)->nr_frags;
1112 frag = skb_shinfo(skb)->frags;
/linux-4.4.14/drivers/s390/net/
Dqeth_core_main.c3821 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in qeth_get_elements_for_frags()
3822 frag = &skb_shinfo(skb)->frags[cnt]; in qeth_get_elements_for_frags()
3906 if (skb_shinfo(skb)->nr_frags) in __qeth_fill_buffer()
3927 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in __qeth_fill_buffer()
3928 frag = &skb_shinfo(skb)->frags[cnt]; in __qeth_fill_buffer()
5258 card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; in qeth_core_get_next_skb()
Dqeth_l3_main.c2799 hdr->ext.mss = skb_shinfo(skb)->gso_size; in qeth_tso_fill_header()
2874 (skb_shinfo(skb)->nr_frags == 0)) { in qeth_l3_hard_start_xmit()
2961 nr_frags = skb_shinfo(new_skb)->nr_frags; in qeth_l3_hard_start_xmit()
/linux-4.4.14/drivers/net/ethernet/qlogic/
Dqla3xxx.c2315 int frag_cnt = (int)skb_shinfo(skb)->nr_frags; in ql_send_map()
2348 skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs]; in ql_send_map()
2474 skb_shinfo(skb)->nr_frags); in ql3xxx_send()
/linux-4.4.14/drivers/net/ethernet/faraday/
Dftgmac100.c466 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, 0, size); in ftgmac100_rx_packet()
/linux-4.4.14/net/
DKconfig41 achieve this, you need to set skb_shinfo(skb)->frag_list to the
/linux-4.4.14/net/iucv/
Daf_iucv.c2246 if (skb_shinfo(list_skb) == skb_shinfo(skb)) { in afiucv_hs_callback_txnotify()
/linux-4.4.14/drivers/net/wireless/mwifiex/
Dmain.c851 skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS && in mwifiex_hard_start_xmit()
/linux-4.4.14/drivers/net/ethernet/altera/
Daltera_tse_main.c572 int nfrags = skb_shinfo(skb)->nr_frags; in tse_start_xmit()
/linux-4.4.14/drivers/net/ethernet/ibm/emac/
Dcore.c1449 int nr_frags = skb_shinfo(skb)->nr_frags; in emac_start_xmit_sg()
1484 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in emac_start_xmit_sg()
/linux-4.4.14/drivers/scsi/cxgbi/cxgb3i/
Dcxgb3i.c408 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); in push_tx_frames()
/linux-4.4.14/drivers/net/phy/
Ddp83640.c1460 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in dp83640_txtstamp()
/linux-4.4.14/drivers/net/ethernet/aurora/
Dnb8800.c267 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in nb8800_receive()
/linux-4.4.14/drivers/net/ethernet/octeon/
Docteon_mgmt.c1282 re.s.tstamp = ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) != 0); in octeon_mgmt_xmit()
/linux-4.4.14/drivers/net/ethernet/xscale/
Dixp4xx_eth.c334 shtx = skb_shinfo(skb); in ixp_tx_timestamp()
/linux-4.4.14/drivers/net/ethernet/broadcom/genet/
Dbcmgenet.c1449 nr_frags = skb_shinfo(skb)->nr_frags; in bcmgenet_xmit()
1490 &skb_shinfo(skb)->frags[i], in bcmgenet_xmit()
/linux-4.4.14/drivers/net/ethernet/agere/
Det131x.c2427 u32 nr_frags = skb_shinfo(skb)->nr_frags + 1; in nic_send_packet()
2428 struct skb_frag_struct *frags = &skb_shinfo(skb)->frags[0]; in nic_send_packet()

12