/linux-4.4.14/net/core/ |
D | skbuff.c | 258 shinfo = skb_shinfo(skb); in __alloc_skb() 326 shinfo = skb_shinfo(skb); in __build_skb() 536 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 553 skb_drop_list(&skb_shinfo(skb)->frag_list); in skb_drop_fraglist() 576 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data() 720 if (skb_shinfo(skb)->tx_flags & SKBTX_DEV_ZEROCOPY) { in skb_tx_error() 723 uarg = skb_shinfo(skb)->destructor_arg; in skb_tx_error() 726 skb_shinfo(skb)->tx_flags &= ~SKBTX_DEV_ZEROCOPY; in skb_tx_error() 839 atomic_inc(&(skb_shinfo(skb)->dataref)); in __skb_clone() 881 int num_frags = skb_shinfo(skb)->nr_frags; in skb_copy_ubufs() [all …]
|
D | tso.c | 11 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs() 53 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data() 54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 77 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start() 78 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
D | datagram.c | 376 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_iter() 378 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_iter() 463 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_from_iter() 465 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter() 592 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_and_csum_datagram() 594 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_and_csum_datagram()
|
D | dev.c | 2421 skb->len, skb->data_len, skb_shinfo(skb)->gso_size, in skb_warn_bad_offload() 2422 skb_shinfo(skb)->gso_type, skb->ip_summed); in skb_warn_bad_offload() 2437 if (unlikely(skb_shinfo(skb)->gso_size)) { in skb_checksum_help() 2596 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma() 2597 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 2608 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma() 2609 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 2679 u16 gso_segs = skb_shinfo(skb)->gso_segs; in netif_skb_features() 2842 const struct skb_shared_info *shinfo = skb_shinfo(skb); in qdisc_pkt_len_init() 3084 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_SCHED_TSTAMP)) in __dev_queue_xmit() [all …]
|
D | pktgen.c | 2740 skb_shinfo(skb)->frags[i].page_offset = 0; in pktgen_finalize_skb() 2743 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb() 2746 skb_frag_size_set(&skb_shinfo(skb)->frags[i], frag_len); in pktgen_finalize_skb() 2747 datalen -= skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2748 skb->len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2749 skb->data_len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2751 skb_shinfo(skb)->nr_frags = i; in pktgen_finalize_skb() 3241 int nr_frags = pkt_dev->skb ? skb_shinfo(pkt_dev->skb)->nr_frags : -1; in pktgen_stop_device()
|
/linux-4.4.14/net/ipv6/ |
D | udp_offload.c | 33 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment() 39 int type = skb_shinfo(skb)->gso_type; in udp6_ufo_fragment() 53 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp6_ufo_fragment() 56 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment() 63 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment() 115 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment() 117 fptr->identification = skb_shinfo(skb)->ip6_frag_id; in udp6_ufo_fragment() 163 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp6_gro_complete() 167 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp6_gro_complete()
|
D | ip6_offload.c | 71 if (unlikely(skb_shinfo(skb)->gso_type & in ipv6_gso_segment() 104 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in ipv6_gso_segment() 289 skb_shinfo(skb)->gso_type |= SKB_GSO_SIT; in sit_gro_complete()
|
D | reassembly.c | 441 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip6_frag_reasm() 443 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip6_frag_reasm() 444 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip6_frag_reasm() 480 if (!skb_shinfo(head)->frag_list) in ip6_frag_reasm() 481 skb_shinfo(head)->frag_list = fp; in ip6_frag_reasm()
|
D | tcpv6_offload.c | 39 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete()
|
D | ip6_output.c | 652 frag = skb_shinfo(skb)->frag_list; in ip6_fragment() 1132 skb_shinfo(skb)->gso_size = (mtu - fragheaderlen - in ip6_ufo_append_data() 1134 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip6_ufo_append_data() 1135 skb_shinfo(skb)->ip6_frag_id = ipv6_select_ident(sock_net(sk), in ip6_ufo_append_data() 1454 skb_shinfo(skb)->tx_flags = tx_flags; in __ip6_append_data() 1456 skb_shinfo(skb)->tskey = tskey; in __ip6_append_data() 1515 int i = skb_shinfo(skb)->nr_frags; in __ip6_append_data() 1529 skb_shinfo(skb)->nr_frags = ++i; in __ip6_append_data() 1539 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip6_append_data() 1635 tail_skb = &(skb_shinfo(skb)->frag_list); in __ip6_make_skb()
|
D | output_core.c | 61 skb_shinfo(skb)->ip6_frag_id = htonl(id); in ipv6_proxy_select_ident()
|
D | ip6_input.c | 115 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ipv6_rcv()
|
D | udp.c | 991 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing()
|
/linux-4.4.14/net/ipv4/ |
D | tcp_offload.c | 22 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp() 23 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 80 mss = skb_shinfo(skb)->gso_size; in tcp_gso_segment() 86 int type = skb_shinfo(skb)->gso_type; in tcp_gso_segment() 104 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in tcp_gso_segment() 128 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_SW_TSTAMP)) in tcp_gso_segment() 129 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 245 mss = skb_shinfo(p)->gso_size; in tcp_gro_receive() 282 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp_gro_complete() 285 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_gro_complete() [all …]
|
D | udp_offload.c | 42 bool need_csum = !!(skb_shinfo(skb)->gso_type & in __skb_udp_tunnel_segment() 44 bool remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM); in __skb_udp_tunnel_segment() 188 (skb_shinfo(skb)->gso_type & in udp4_ufo_fragment() 197 mss = skb_shinfo(skb)->gso_size; in udp4_ufo_fragment() 203 int type = skb_shinfo(skb)->gso_type; in udp4_ufo_fragment() 214 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp4_ufo_fragment() 407 skb_shinfo(skb)->gso_type |= SKB_GSO_TUNNEL_REMCSUM; in udp_gro_complete() 421 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp4_gro_complete() 425 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp4_gro_complete()
|
D | inet_lro.c | 154 nr_frags = skb_shinfo(skb)->nr_frags; in lro_init_desc() 156 lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); in lro_init_desc() 228 skb_shinfo(parent)->frag_list = skb; in lro_add_packet() 283 skb_shinfo(lro_desc->parent)->gso_size = lro_desc->mss; in lro_flush()
|
D | ip_output.c | 452 skb_shinfo(skb)->gso_segs ?: 1); in ip_queue_xmit() 600 frag = skb_shinfo(skb)->frag_list; in ip_do_fragment() 857 skb_shinfo(skb)->gso_size = maxfraglen - fragheaderlen; in ip_ufo_append_data() 858 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip_ufo_append_data() 1013 skb_shinfo(skb)->tx_flags = cork->tx_flags; in __ip_append_data() 1015 skb_shinfo(skb)->tskey = tskey; in __ip_append_data() 1071 int i = skb_shinfo(skb)->nr_frags; in __ip_append_data() 1085 skb_shinfo(skb)->nr_frags = ++i; in __ip_append_data() 1095 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip_append_data() 1246 skb_shinfo(skb)->gso_size = mtu - fragheaderlen; in ip_append_page() [all …]
|
D | gre_offload.c | 31 if (unlikely(skb_shinfo(skb)->gso_type & in gre_gso_segment() 236 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
|
D | ip_fragment.c | 588 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip_frag_reasm() 590 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip_frag_reasm() 591 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip_frag_reasm() 600 skb_shinfo(head)->frag_list = head->next; in ip_frag_reasm()
|
D | ip_tunnel_core.c | 80 __ip_select_ident(net, iph, skb_shinfo(skb)->gso_segs ?: 1); in iptunnel_xmit() 165 skb_shinfo(skb)->gso_type |= gso_type_mask; in iptunnel_handle_offloads()
|
D | tcp_output.c | 382 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_ecn_send() 991 skb_shinfo(skb)->gso_type = sk->sk_gso_type; in tcp_transmit_skb() 1018 skb_shinfo(skb)->gso_segs = tcp_skb_pcount(skb); in tcp_transmit_skb() 1019 skb_shinfo(skb)->gso_size = tcp_skb_mss(skb); in tcp_transmit_skb() 1117 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_fragment_tstamp() 1121 struct skb_shared_info *shinfo2 = skb_shinfo(skb2); in tcp_fragment_tstamp() 1176 if (!skb_shinfo(skb)->nr_frags && skb->ip_summed != CHECKSUM_PARTIAL) { in tcp_fragment() 1237 shinfo = skb_shinfo(skb); in __pskb_trim_head() 1968 if (!skb_shinfo(skb)->nr_frags) { in tcp_mtu_probe() 2497 if (skb_shinfo(skb)->nr_frags != 0) in tcp_can_collapse()
|
D | tcp.c | 435 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_tx_timestamp() 940 i = skb_shinfo(skb)->nr_frags; in do_tcp_sendpages() 950 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages() 955 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in do_tcp_sendpages() 1207 int i = skb_shinfo(skb)->nr_frags; in tcp_sendmsg() 1236 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg() 3027 const struct skb_shared_info *shi = skb_shinfo(skb); in tcp_md5_hash_skb_data()
|
D | ip_input.c | 421 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ip_rcv()
|
D | tcp_minisocks.c | 821 tcp_sk(child)->segs_in += max_t(u16, 1, skb_shinfo(skb)->gso_segs); in tcp_child_process()
|
D | af_inet.c | 1210 if (unlikely(skb_shinfo(skb)->gso_type & in inet_gso_segment() 1254 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in inet_gso_segment()
|
D | raw.c | 380 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_send_hdrinc()
|
D | ip_gre.c | 473 !(skb_shinfo(skb)->gso_type & in build_header()
|
/linux-4.4.14/drivers/net/xen-netback/ |
D | netback.c | 369 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_frag_copy() 371 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_frag_copy() 425 int nr_frags = skb_shinfo(skb)->nr_frags; in xenvif_gop_skb() 438 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_skb() 440 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_skb() 449 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb() 459 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb() 485 skb_frag_page(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb() 486 skb_frag_size(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb() 487 skb_shinfo(skb)->frags[i].page_offset, in xenvif_gop_skb() [all …]
|
D | interface.c | 57 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in xenvif_skb_zerocopy_prepare()
|
/linux-4.4.14/include/trace/events/ |
D | net.h | 54 __entry->tx_flags = skb_shinfo(skb)->tx_flags; 55 __entry->gso_size = skb_shinfo(skb)->gso_size; 56 __entry->gso_segs = skb_shinfo(skb)->gso_segs; 57 __entry->gso_type = skb_shinfo(skb)->gso_type; 189 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 190 __entry->gso_size = skb_shinfo(skb)->gso_size; 191 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
/linux-4.4.14/drivers/net/ethernet/sfc/ |
D | tx.c | 256 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in efx_skb_copy_bits_to_pio() 257 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in efx_skb_copy_bits_to_pio() 267 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->frag_list); in efx_skb_copy_bits_to_pio() 282 if (skb_shinfo(skb)->nr_frags) { in efx_enqueue_skb_pio() 346 if (skb_shinfo(skb)->gso_size) in efx_enqueue_skb() 411 if (i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb() 413 fragment = &skb_shinfo(skb)->frags[i]; in efx_enqueue_skb() 457 skb_shinfo(skb)->nr_frags + 1); in efx_enqueue_skb() 1152 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet() 1156 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet() [all …]
|
D | rx.c | 450 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_packet_gro() 455 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_packet_gro() 503 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_mk_skb() 509 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_mk_skb()
|
D | net_driver.h | 1510 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in efx_xmit_with_hwtstamp() 1514 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in efx_xmit_hwtstamp_pending()
|
D | ptp.c | 890 if (skb_shinfo(skb)->nr_frags != 0) { in efx_ptp_xmit_skb()
|
/linux-4.4.14/include/linux/ |
D | skbuff.h | 1114 #define skb_shinfo(SKB) ((struct skb_shared_info *)(skb_end_pointer(SKB))) macro 1118 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps() 1223 (atomic_read(&skb_shinfo(skb)->dataref) & SKB_DATAREF_MASK) != 1; in skb_cloned() 1250 dataref = atomic_read(&skb_shinfo(skb)->dataref); in skb_header_cloned() 1268 atomic_add(1 << SKB_DATAREF_SHIFT, &skb_shinfo(skb)->dataref); in skb_header_release() 1281 atomic_set(&skb_shinfo(skb)->dataref, 1 + (1 << SKB_DATAREF_SHIFT)); in __skb_header_release() 1697 for (i = (int)skb_shinfo(skb)->nr_frags - 1; i >= 0; i--) in skb_pagelen() 1698 len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in skb_pagelen() 1718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_fill_page_desc() 1752 skb_shinfo(skb)->nr_frags = i + 1; in skb_fill_page_desc() [all …]
|
/linux-4.4.14/drivers/staging/octeon/ |
D | ethernet-tx.c | 204 if (unlikely(skb_shinfo(skb)->nr_frags > 5)) { in cvm_oct_xmit() 274 if (skb_shinfo(skb)->nr_frags == 0) { in cvm_oct_xmit() 283 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in cvm_oct_xmit() 284 struct skb_frag_struct *fs = skb_shinfo(skb)->frags + i; in cvm_oct_xmit() 293 hw_buffer.s.size = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 294 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 349 if (unlikely(skb_shinfo(skb)->nr_frags)) { in cvm_oct_xmit()
|
/linux-4.4.14/net/xfrm/ |
D | xfrm_ipcomp.c | 76 if (WARN_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS)) in ipcomp_decompress() 79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress() 100 skb_shinfo(skb)->nr_frags++; in ipcomp_decompress()
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_tx.c | 191 eseg->mss = cpu_to_be16(skb_shinfo(skb)->gso_size); in mlx5e_sq_xmit() 196 (skb_shinfo(skb)->gso_segs - 1) * ihs; in mlx5e_sq_xmit() 202 !skb_shinfo(skb)->nr_frags; in mlx5e_sq_xmit() 242 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlx5e_sq_xmit() 243 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in mlx5e_sq_xmit()
|
D | en_rx.c | 114 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in mlx5e_lro_update_hdr() 119 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in mlx5e_lro_update_hdr() 199 skb_shinfo(skb)->gso_size = DIV_ROUND_UP(cqe_bcnt, lro_num_seg); in mlx5e_build_rx_skb()
|
/linux-4.4.14/drivers/net/ethernet/sun/ |
D | sunvnet.c | 1020 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_map() 1021 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_map() 1060 docopy = skb_shinfo(skb)->nr_frags >= ncookies; in vnet_skb_shape() 1061 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_shape() 1062 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_shape() 1123 skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; in vnet_skb_shape() 1124 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in vnet_skb_shape() 1174 gso_size = skb_shinfo(skb)->gso_size; in vnet_handle_offloads() 1175 gso_type = skb_shinfo(skb)->gso_type; in vnet_handle_offloads() 1176 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads() [all …]
|
D | sungem.c | 661 if (skb_shinfo(skb)->nr_frags) { in gem_tx() 662 int last = entry + skb_shinfo(skb)->nr_frags; in gem_tx() 680 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_tx() 1015 if (unlikely(TX_BUFFS_AVAIL(gp) <= (skb_shinfo(skb)->nr_frags + 1))) { in gem_start_xmit() 1027 if (skb_shinfo(skb)->nr_frags == 0) { in gem_start_xmit() 1064 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in gem_start_xmit() 1065 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in gem_start_xmit() 1074 if (frag == skb_shinfo(skb)->nr_frags - 1) in gem_start_xmit() 1600 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_clean_rings() 1609 if (frag != skb_shinfo(skb)->nr_frags) in gem_clean_rings()
|
D | sunhme.c | 1225 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_clean_rings() 1239 if (frag != skb_shinfo(skb)->nr_frags) in happy_meal_clean_rings() 1939 if (skb_shinfo(skb)->nr_frags) { in happy_meal_tx() 1942 last = elem + skb_shinfo(skb)->nr_frags; in happy_meal_tx() 1951 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_tx() 2305 if (TX_BUFFS_AVAIL(hp) <= (skb_shinfo(skb)->nr_frags + 1)) { in happy_meal_start_xmit() 2317 if (skb_shinfo(skb)->nr_frags == 0) { in happy_meal_start_xmit() 2343 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_start_xmit() 2344 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in happy_meal_start_xmit() 2356 if (frag == skb_shinfo(skb)->nr_frags - 1) in happy_meal_start_xmit()
|
D | cassini.c | 1880 count -= skb_shinfo(skb)->nr_frags + in cas_tx_ringN() 1891 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_tx_ringN() 2005 skb_frag_t *frag = skb_shinfo(skb)->frags; in cas_rx_process_pkt() 2041 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt() 2065 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt() 2778 CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { in cas_xmit_tx_ringN() 2798 nr_frags = skb_shinfo(skb)->nr_frags; in cas_xmit_tx_ringN() 2824 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in cas_xmit_tx_ringN() 3910 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_clean_txd() 3922 if (frag != skb_shinfo(skb)->nr_frags) { in cas_clean_txd()
|
D | niu.c | 3280 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, offset, size); in niu_rx_skb_append() 3586 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in release_tx_packet() 3590 skb_frag_size(&skb_shinfo(skb)->frags[i]), in release_tx_packet() 6646 if (niu_tx_avail(rp) <= (skb_shinfo(skb)->nr_frags + 1)) { in niu_start_xmit() 6695 nfg = skb_shinfo(skb)->nr_frags; in niu_start_xmit() 6715 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in niu_start_xmit() 6716 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in niu_start_xmit()
|
/linux-4.4.14/drivers/net/ |
D | xen-netfront.c | 310 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_alloc_rx_buffers() 522 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots() 529 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots() 634 if (skb_shinfo(skb)->gso_size) { in xennet_start_xmit() 642 gso->u.gso.size = skb_shinfo(skb)->gso_size; in xennet_start_xmit() 643 gso->u.gso.type = (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) ? in xennet_start_xmit() 657 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xennet_start_xmit() 658 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit() 859 skb_shinfo(skb)->gso_size = gso->u.gso.size; in xennet_set_skb_gso() 860 skb_shinfo(skb)->gso_type = in xennet_set_skb_gso() [all …]
|
D | virtio_net.c | 307 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, offset, in page_to_skb() 385 num_skb_frags = skb_shinfo(curr_skb)->nr_frags; in receive_mergeable() 392 skb_shinfo(curr_skb)->frag_list = nskb; in receive_mergeable() 496 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in receive_buf() 499 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in receive_buf() 502 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in receive_buf() 511 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in receive_buf() 513 skb_shinfo(skb)->gso_size = virtio16_to_cpu(vi->vdev, in receive_buf() 515 if (skb_shinfo(skb)->gso_size == 0) { in receive_buf() 521 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in receive_buf() [all …]
|
D | macvtap.c | 657 skb_shinfo(skb)->gso_size = macvtap16_to_cpu(q, vnet_hdr->gso_size); in macvtap_skb_from_vnet_hdr() 658 skb_shinfo(skb)->gso_type = gso_type; in macvtap_skb_from_vnet_hdr() 661 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in macvtap_skb_from_vnet_hdr() 662 skb_shinfo(skb)->gso_segs = 0; in macvtap_skb_from_vnet_hdr() 674 struct skb_shared_info *sinfo = skb_shinfo(skb); in macvtap_skb_to_vnet_hdr() 821 skb_shinfo(skb)->destructor_arg = m->msg_control; in macvtap_get_user() 822 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in macvtap_get_user() 823 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in macvtap_get_user()
|
D | tun.c | 866 sock_tx_timestamp(skb->sk, &skb_shinfo(skb)->tx_flags); in tun_net_xmit() 1223 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in tun_get_user() 1226 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in tun_get_user() 1229 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in tun_get_user() 1238 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tun_get_user() 1240 skb_shinfo(skb)->gso_size = tun16_to_cpu(tun, gso.gso_size); in tun_get_user() 1241 if (skb_shinfo(skb)->gso_size == 0) { in tun_get_user() 1248 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in tun_get_user() 1249 skb_shinfo(skb)->gso_segs = 0; in tun_get_user() 1254 skb_shinfo(skb)->destructor_arg = msg_control; in tun_get_user() [all …]
|
/linux-4.4.14/net/ipv6/netfilter/ |
D | nf_conntrack_reasm.c | 423 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm() 425 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in nf_ct_frag6_reasm() 426 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in nf_ct_frag6_reasm() 445 skb_shinfo(head)->frag_list = head->next; in nf_ct_frag6_reasm() 478 fp = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm()
|
/linux-4.4.14/net/ieee802154/6lowpan/ |
D | reassembly.c | 271 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in lowpan_frag_reasm() 273 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in lowpan_frag_reasm() 274 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in lowpan_frag_reasm() 294 if (!skb_shinfo(head)->frag_list) in lowpan_frag_reasm() 295 skb_shinfo(head)->frag_list = fp; in lowpan_frag_reasm()
|
/linux-4.4.14/drivers/net/ethernet/neterion/vxge/ |
D | vxge-main.h | 515 #define vxge_tcp_mss(skb) (skb_shinfo(skb)->gso_size) 516 #define vxge_udp_mss(skb) (skb_shinfo(skb)->gso_size) 517 #define vxge_offload_type(skb) (skb_shinfo(skb)->gso_type)
|
D | vxge-main.c | 565 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit_compl() 566 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit_compl() 913 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit() 923 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit() 978 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit() 1047 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_tx_term() 1048 frag = &skb_shinfo(skb)->frags[0]; in vxge_tx_term()
|
/linux-4.4.14/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 917 struct skb_shared_info *sh = skb_shinfo(skb); in nicvf_tso_count_subdescs() 963 if (skb_shinfo(skb)->gso_size) { in nicvf_sq_subdesc_required() 968 if (skb_shinfo(skb)->nr_frags) in nicvf_sq_subdesc_required() 969 subdesc_cnt += skb_shinfo(skb)->nr_frags; in nicvf_sq_subdesc_required() 1054 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in nicvf_sq_append_tso() 1136 if (skb_shinfo(skb)->gso_size) in nicvf_sq_append_skb() 1151 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in nicvf_sq_append_skb() 1154 frag = &skb_shinfo(skb)->frags[i]; in nicvf_sq_append_skb() 1225 if (!skb_shinfo(skb)->frag_list) in nicvf_get_rcv_skb() 1226 skb_shinfo(skb)->frag_list = skb_frag; in nicvf_get_rcv_skb()
|
/linux-4.4.14/drivers/infiniband/hw/nes/ |
D | nes_nic.c | 394 wqe_misc |= NES_NIC_SQ_WQE_LSO_ENABLE | (u16)skb_shinfo(skb)->gso_size; in nes_nic_send() 411 if ((skb_shinfo(skb)->nr_frags + 1) > 4) { in nes_nic_send() 413 netdev->name, skb_shinfo(skb)->nr_frags + 2, skb_headlen(skb)); in nes_nic_send() 437 for (skb_fragment_index = 0; skb_fragment_index < skb_shinfo(skb)->nr_frags; in nes_nic_send() 440 &skb_shinfo(skb)->frags[skb_fragment_index]; in nes_nic_send() 445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); in nes_nic_send() 524 nr_frags = skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit() 557 for (tso_frag_count=0; tso_frag_count < skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit() 560 &skb_shinfo(skb)->frags[tso_frag_count]; in nes_netdev_start_xmit() 599 skb_shinfo(skb)->nr_frags, skb_is_gso(skb), skb_shinfo(skb)->gso_size); in nes_netdev_start_xmit() [all …]
|
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/ |
D | i40e_txrx.c | 1035 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps() 1486 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso() 1674 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize() 1675 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize() 1688 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize() 1693 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize() 1695 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize() 1781 gso_segs = skb_shinfo(skb)->gso_segs; in i40evf_tx_map() 1796 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in i40evf_tx_map() 1933 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in i40evf_xmit_descriptor_count() [all …]
|
/linux-4.4.14/drivers/net/ethernet/aeroflex/ |
D | greth.c | 104 if (skb_shinfo(skb)->nr_frags == 0) in greth_print_tx_packet() 112 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in greth_print_tx_packet() 115 skb_frag_address(&skb_shinfo(skb)->frags[i]), in greth_print_tx_packet() 116 skb_shinfo(skb)->frags[i].size, true); in greth_print_tx_packet() 191 int nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_rings() 201 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_rings() 474 nr_frags = skb_shinfo(skb)->nr_frags; in greth_start_xmit_gbit() 521 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_start_xmit_gbit() 691 nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_tx_gbit() 718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_tx_gbit()
|
/linux-4.4.14/net/mpls/ |
D | mpls_gso.c | 29 if (unlikely(skb_shinfo(skb)->gso_type & in mpls_gso_segment()
|
/linux-4.4.14/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_io.c | 397 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_encap_pkt() 500 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_pkt() 587 nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_map_tx_skb() 599 frag = &skb_shinfo(skb)->frags[i]; in qlcnic_map_tx_skb() 629 int i, nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_unmap_buffers() 676 frag_count = skb_shinfo(skb)->nr_frags + 1; in qlcnic_xmit_frame() 683 delta += skb_frag_size(&skb_shinfo(skb)->frags[i]); in qlcnic_xmit_frame() 688 frag_count = 1 + skb_shinfo(skb)->nr_frags; in qlcnic_xmit_frame() 1337 skb_shinfo(skb)->gso_size = qlcnic_get_lro_sts_mss(sts_data1); in qlcnic_process_lro() 1339 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in qlcnic_process_lro() [all …]
|
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 309 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in fm10k_add_rx_frag() 813 first->gso_segs = skb_shinfo(skb)->gso_segs; in fm10k_tso() 819 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso() 918 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in fm10k_tx_desc_flags() 919 likely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in fm10k_tx_desc_flags() 1000 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in fm10k_tx_map() 1108 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in fm10k_xmit_frame_ring() 1109 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in fm10k_xmit_frame_ring()
|
D | fm10k_ptp.c | 74 skb_shinfo(clone)->tx_flags |= SKBTX_IN_PROGRESS; in fm10k_ts_tx_enqueue()
|
D | fm10k_netdev.c | 664 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in fm10k_xmit_frame()
|
/linux-4.4.14/net/phonet/ |
D | pep-gprs.c | 112 skb_shinfo(rskb)->frag_list = skb; in gprs_recv() 120 skb->next = skb_shinfo(skb)->frag_list; in gprs_recv()
|
D | pep.c | 1221 skb_shinfo(rskb)->frag_list = skb; in pep_write() 1229 skb->next = skb_shinfo(skb)->frag_list; in pep_write()
|
/linux-4.4.14/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 1570 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps() 2237 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso() 2259 if (likely(!(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP))) in i40e_tsyn() 2275 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in i40e_tsyn() 2510 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize() 2511 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize() 2524 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize() 2529 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize() 2531 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize() 2589 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_tx_map() [all …]
|
D | i40e_fcoe.c | 1218 if (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE) { in i40e_fcoe_tso() 1221 skb_shinfo(skb)->gso_type); in i40e_fcoe_tso() 1243 cd_mss = skb_shinfo(skb)->gso_size; in i40e_fcoe_tso()
|
/linux-4.4.14/drivers/infiniband/ulp/ipoib/ |
D | ipoib_ib.c | 283 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_map_tx() 284 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_map_tx() 296 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in ipoib_dma_map_tx() 322 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_unmap_tx() 323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_unmap_tx() 526 priv->tx_wr.mss = skb_shinfo(skb)->gso_size; in post_send()
|
D | ipoib.h | 516 skb_frag_t *frags = skb_shinfo(skb)->frags; in ipoib_build_sge() 517 int nr_frags = skb_shinfo(skb)->nr_frags; in ipoib_build_sge()
|
D | ipoib_cm.c | 537 num_frags = skb_shinfo(skb)->nr_frags; in skb_put_frags() 539 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags() 545 --skb_shinfo(skb)->nr_frags; in skb_put_frags()
|
/linux-4.4.14/drivers/target/tcm_fc/ |
D | tfc_io.c | 147 skb_shinfo(fp_skb(fp))->nr_frags, in ft_queue_data_in() 171 (skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN)) in ft_queue_data_in()
|
/linux-4.4.14/drivers/staging/unisys/visornic/ |
D | visornic_main.c | 221 numfrags = skb_shinfo(skb)->nr_frags; in visor_copy_fragsinfo_from_skb() 234 numfrags = skb_shinfo(skb)->nr_frags; in visor_copy_fragsinfo_from_skb() 263 skb_frag_page(&skb_shinfo(skb)->frags[ii])), in visor_copy_fragsinfo_from_skb() 264 skb_shinfo(skb)->frags[ii]. in visor_copy_fragsinfo_from_skb() 266 skb_shinfo(skb)->frags[ii]. in visor_copy_fragsinfo_from_skb() 280 if (skb_shinfo(skb)->frag_list) { in visor_copy_fragsinfo_from_skb() 284 for (skbinlist = skb_shinfo(skb)->frag_list; skbinlist; in visor_copy_fragsinfo_from_skb() 1261 skb_shinfo(skb)->frag_list = curr; in visornic_rx()
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 253 nfrags = skb_shinfo(skb)->nr_frags; in unmap_skb() 257 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]), in unmap_skb() 899 skb_fill_page_desc(newskb, skb_shinfo(newskb)->nr_frags, in get_packet_pg() 946 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1) + 2; in calc_tx_descs() 947 if (skb_shinfo(skb)->gso_size) in calc_tx_descs() 978 nfrags = skb_shinfo(skb)->nr_frags; in make_sgl() 980 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in make_sgl() 1154 tso_info = V_LSO_MSS(skb_shinfo(skb)->gso_size); in write_tx_pkt_wr() 1283 if (skb_shinfo(skb)->gso_size) in t3_eth_xmit() 1544 si = skb_shinfo(skb); in deferred_unmap_destructor() [all …]
|
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/ |
D | lio_main.c | 1389 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf() 1397 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf() 1437 frags = skb_shinfo(skb)->nr_frags; in free_netsgbuf_with_resp() 1445 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; in free_netsgbuf_with_resp() 2529 if (unlikely((skb_shinfo(skb)->tx_flags | SKBTX_IN_PROGRESS) != 0)) { in handle_timestamp() 2791 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) { in liquidio_xmit() 2792 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in liquidio_xmit() 2796 if (skb_shinfo(skb)->nr_frags == 0) { in liquidio_xmit() 2830 cmdsetup.s.u.gatherptrs = (skb_shinfo(skb)->nr_frags + 1); in liquidio_xmit() 2846 frags = skb_shinfo(skb)->nr_frags; in liquidio_xmit() [all …]
|
/linux-4.4.14/include/net/ |
D | udp_tunnel.h | 114 skb_shinfo(skb)->gso_type |= uh->check ? in udp_tunnel_gro_complete()
|
D | udp.h | 146 for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) { in udp_csum()
|
D | sch_generic.h | 519 bstats->packets += skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 1; in bstats_update()
|
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-drv.c | 1167 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in xgbe_prep_tx_tstamp() 1197 packet->mss = skb_shinfo(skb)->gso_size; in xgbe_prep_tso() 1206 packet->tx_packets = skb_shinfo(skb)->gso_segs; in xgbe_prep_tso() 1244 if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { in xgbe_packet_info() 1273 if ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in xgbe_packet_info() 1283 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_packet_info() 1284 frag = &skb_shinfo(skb)->frags[i]; in xgbe_packet_info() 1792 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xgbe_create_skb() 1968 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xgbe_rx_poll()
|
D | xgbe-desc.c | 571 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_map_tx_skb() 575 frag = &skb_shinfo(skb)->frags[i]; in xgbe_map_tx_skb()
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 1117 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in compute_large_page_tx_descs() 1124 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in compute_large_page_tx_descs() 1208 nfrags = skb_shinfo(skb)->nr_frags; in write_tx_descs() 1256 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in write_tx_descs() 1320 count = 1 + skb_shinfo(skb)->nr_frags; in restart_sched() 1674 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx() 1711 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx() 1796 if (skb_shinfo(skb)->gso_size) { in t1_start_xmit() 1811 skb_shinfo(skb)->gso_size)); in t1_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.c | 524 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in bnx2x_set_gro_params() 527 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in bnx2x_set_gro_params() 538 skb_shinfo(skb)->gso_size = len_on_bd - hdrs_len; in bnx2x_set_gro_params() 741 if (skb_shinfo(skb)->gso_size) { in bnx2x_gro_receive() 3452 if (skb_shinfo(skb)->nr_frags >= (MAX_FETCH_BD - num_tso_win_sub)) { in bnx2x_pkt_req_lin() 3454 unsigned short lso_mss = skb_shinfo(skb)->gso_size; in bnx2x_pkt_req_lin() 3457 int num_wnds = skb_shinfo(skb)->nr_frags - wnd_size; in bnx2x_pkt_req_lin() 3479 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]); in bnx2x_pkt_req_lin() 3495 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx + wnd_size - 1]); in bnx2x_pkt_req_lin() 3502 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx]); in bnx2x_pkt_req_lin() [all …]
|
/linux-4.4.14/drivers/net/ethernet/apm/xgene/ |
D | xgene_enet_main.c | 165 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgene_enet_tx_completion() 166 frag = &skb_shinfo(skb)->frags[i]; in xgene_enet_tx_completion() 219 mss = skb_shinfo(skb)->gso_size; in xgene_enet_work_msg() 223 nr_frags = skb_shinfo(skb)->nr_frags; in xgene_enet_work_msg() 226 len += skb_shinfo(skb)->frags[i].size; in xgene_enet_work_msg() 338 nr_frags = skb_shinfo(skb)->nr_frags; in xgene_enet_setup_tx_desc() 346 frag = &skb_shinfo(skb)->frags[fidx]; in xgene_enet_setup_tx_desc()
|
/linux-4.4.14/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 655 struct skb_frag_struct *frag = skb_shinfo(skb)->frags + in vmxnet3_append_frag() 656 skb_shinfo(skb)->nr_frags; in vmxnet3_append_frag() 658 BUG_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS); in vmxnet3_append_frag() 665 skb_shinfo(skb)->nr_frags++; in vmxnet3_append_frag() 752 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vmxnet3_map_pkt() 753 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in vmxnet3_map_pkt() 933 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in txd_estimate() 934 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in txd_estimate() 973 ctx.mss = skb_shinfo(skb)->gso_size; in vmxnet3_tq_xmit() 1424 skb_shinfo(skb)->gso_type = rcd->v4 ? in vmxnet3_rq_rx_complete() [all …]
|
/linux-4.4.14/drivers/net/hyperv/ |
D | netvsc_drv.c | 328 int frags = skb_shinfo(skb)->nr_frags; in init_page_array() 349 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in init_page_array() 360 int i, frags = skb_shinfo(skb)->nr_frags; in count_skb_frag_slots() 364 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in count_skb_frag_slots() 613 lso_info->lso_v2_transmit.mss = skb_shinfo(skb)->gso_size; in netvsc_start_xmit()
|
/linux-4.4.14/net/sctp/ |
D | ulpqueue.c | 335 struct sk_buff *list = skb_shinfo(f_frag)->frag_list; in sctp_make_reassembled_event() 365 skb_shinfo(new)->frag_list = pos; in sctp_make_reassembled_event() 367 skb_shinfo(f_frag)->frag_list = pos; in sctp_make_reassembled_event() 997 flist = skb_shinfo(skb)->frag_list; in sctp_ulpq_renege_list()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_fcoe.c | 522 if (skb_is_gso(skb) && (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE)) { in ixgbe_fso() 524 skb_shinfo(skb)->gso_type); in ixgbe_fso() 593 skb_shinfo(skb)->gso_size); in ixgbe_fso() 602 mss_l4len_idx = skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_fso()
|
D | ixgbe_main.c | 1602 skb_shinfo(skb)->gso_size = DIV_ROUND_UP((skb->len - hdr_len), in ixgbe_set_rsc_gso_size() 1604 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in ixgbe_set_rsc_gso_size() 1735 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_pull_tail() 1781 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_dma_sync_frag() 1919 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbe_add_rx_frag() 6984 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbe_tso() 6989 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_tso() 7219 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in ixgbe_tx_map() 7510 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbe_xmit_frame_ring() 7511 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbe_xmit_frame_ring() [all …]
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 544 struct skb_frag_struct *skb_frags_rx = skb_shinfo(skb)->frags; in mlx4_en_complete_rx_desc() 623 skb_shinfo(skb)->nr_frags = used_frags; in mlx4_en_rx_skb() 631 skb_shinfo(skb)->frags[0].page_offset += pull_len; in mlx4_en_rx_skb() 634 skb_frag_size_sub(&skb_shinfo(skb)->frags[0], pull_len); in mlx4_en_rx_skb() 891 void *va = skb_frag_address(skb_shinfo(gro_skb)->frags); in mlx4_en_process_rx_cq() 900 skb_shinfo(gro_skb)->nr_frags = nr; in mlx4_en_process_rx_cq()
|
D | en_tx.c | 709 struct skb_shared_info *shinfo = skb_shinfo(skb); in mlx4_en_xmit()
|
/linux-4.4.14/drivers/staging/rdma/amso1100/ |
D | c2.c | 769 if (unlikely(c2_port->tx_avail < (skb_shinfo(skb)->nr_frags + 1))) { in c2_xmit_frame() 799 if (skb_shinfo(skb)->nr_frags) { in c2_xmit_frame() 800 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in c2_xmit_frame() 801 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in c2_xmit_frame() 824 c2_port->tx_avail -= (skb_shinfo(skb)->nr_frags + 1); in c2_xmit_frame()
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 287 si = skb_shinfo(skb); in map_skb() 310 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl() 878 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1); in calc_tx_flits() 879 if (skb_shinfo(skb)->gso_size) in calc_tx_flits() 912 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl() 1193 if (!skb_shinfo(skb)->gso_size && (unlikely(skb->len > max_pkt_len))) in t4vf_eth_xmit() 1279 ssi = skb_shinfo(skb); in t4vf_eth_xmit() 1483 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
|
/linux-4.4.14/drivers/net/usb/ |
D | cdc-phonet.c | 166 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete() 172 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete()
|
D | usbnet.c | 1275 num_sgs = skb_shinfo(skb)->nr_frags + 1; in build_dma_sg() 1291 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in build_dma_sg() 1292 struct skb_frag_struct *f = &skb_shinfo(skb)->frags[i]; in build_dma_sg()
|
D | r8152.c | 1385 if (skb_shinfo(skb)->gso_size) { in r8152_csum_workaround() 1466 u32 mss = skb_shinfo(skb)->gso_size; in r8152_tx_csum() 1984 u32 mss = skb_shinfo(skb)->gso_size; in rtl8152_features_check()
|
D | ax88179_178a.c | 1428 int mss = skb_shinfo(skb)->gso_size; in ax88179_tx_fixup()
|
/linux-4.4.14/drivers/net/wireless/ath/wil6210/ |
D | txrx.c | 1228 int nr_frags = skb_shinfo(skb)->nr_frags; in __wil_tx_vring_tso() 1230 int mss = skb_shinfo(skb)->gso_size; /* payload size w/o headers */ in __wil_tx_vring_tso() 1270 gso_type = skb_shinfo(skb)->gso_type & (SKB_GSO_TCPV6 | SKB_GSO_TCPV4); in __wil_tx_vring_tso() 1326 frag = &skb_shinfo(skb)->frags[f]; in __wil_tx_vring_tso() 1500 int nr_frags = skb_shinfo(skb)->nr_frags; in __wil_tx_vring() 1554 &skb_shinfo(skb)->frags[f]; in __wil_tx_vring() 1722 (skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS); in wil_need_txstat()
|
D | debugfs.c | 870 int nr_frags = skb_shinfo(skb)->nr_frags; in wil_seq_print_skb() 879 &skb_shinfo(skb)->frags[i]; in wil_seq_print_skb()
|
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/ |
D | fs_enet-main.c | 528 nr_frags = skb_shinfo(skb)->nr_frags; in fs_enet_start_xmit() 529 frag = skb_shinfo(skb)->frags; in fs_enet_start_xmit() 558 nr_frags = skb_shinfo(skb)->nr_frags; in fs_enet_start_xmit() 586 frag = skb_shinfo(skb)->frags; in fs_enet_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/realtek/ |
D | 8139cp.c | 727 this_frag = &skb_shinfo(skb)->frags[frag]; in unwind_tx_frag_mapping() 746 if (TX_BUFFS_AVAIL(cp) <= (skb_shinfo(skb)->nr_frags + 1)) { in cp_start_xmit() 755 mss = skb_shinfo(skb)->gso_size; in cp_start_xmit() 780 if (skb_shinfo(skb)->nr_frags == 0) { in cp_start_xmit() 821 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in cp_start_xmit() 822 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in cp_start_xmit() 841 if (frag == skb_shinfo(skb)->nr_frags - 1) in cp_start_xmit() 1865 if (skb_shinfo(skb)->gso_size > MSSMask) in cp_features_check()
|
D | r8169.c | 6879 struct skb_shared_info *info = skb_shinfo(skb); in rtl8169_xmit_frags() 6941 if (skb_shinfo(skb)->gso_size) { in r8169_csum_workaround() 7011 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v1() 7034 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v2() 7121 if (unlikely(!TX_FRAGS_READY_FOR(tp, skb_shinfo(skb)->nr_frags))) { in rtl8169_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 262 si = skb_shinfo(skb); in map_skb() 291 si = skb_shinfo(skb); in unmap_skb() 315 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl() 773 int hdrlen = skb_shinfo(skb)->gso_size ? in is_eth_imm() 810 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1); in calc_tx_flits() 811 if (skb_shinfo(skb)->gso_size) in calc_tx_flits() 856 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl() 1170 if (!skb_shinfo(skb)->gso_size && (unlikely(skb->len > max_pkt_len))) in t4_eth_xmit() 1220 ssi = skb_shinfo(skb); in t4_eth_xmit() 1504 cnt = skb_shinfo(skb)->nr_frags; in calc_tx_flits_ofld() [all …]
|
/linux-4.4.14/drivers/block/aoe/ |
D | aoedev.c | 405 while (atomic_read(&skb_shinfo(skb)->dataref) != 1 && i-- > 0) in skbfree() 415 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in skbfree()
|
D | aoecmd.c | 181 if (skb && atomic_read(&skb_shinfo(skb)->dataref) == 1) { in skb_pool_get() 235 if (atomic_read(&skb_shinfo(skb)->dataref) != 1) { in newtframe() 244 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in newtframe()
|
/linux-4.4.14/drivers/scsi/fcoe/ |
D | fcoe_transport.c | 308 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in fcoe_fc_crc() 309 frag = &skb_shinfo(skb)->frags[i]; in fcoe_fc_crc() 471 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, in fcoe_get_paged_crc_eof()
|
D | fcoe.c | 1644 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in fcoe_xmit() 1697 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in fcoe_xmit() 1698 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in fcoe_xmit() 1700 skb_shinfo(skb)->gso_type = 0; in fcoe_xmit() 1701 skb_shinfo(skb)->gso_size = 0; in fcoe_xmit()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ |
D | ibmveth.c | 1037 if (skb_shinfo(skb)->nr_frags > 5 && __skb_linearize(skb)) { in ibmveth_start_xmit() 1108 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ibmveth_start_xmit() 1109 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ibmveth_start_xmit() 1123 mss = (unsigned long)skb_shinfo(skb)->gso_size; in ibmveth_start_xmit() 1132 cpu_to_be16(skb_shinfo(skb)->gso_size); in ibmveth_start_xmit() 1150 for (i = 1; i < skb_shinfo(skb)->nr_frags + 1; i++) in ibmveth_start_xmit()
|
/linux-4.4.14/drivers/hsi/clients/ |
D | ssi_protocol.c | 186 BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); in ssip_skb_to_msg() 190 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ssip_skb_to_msg() 193 frag = &skb_shinfo(skb)->frags[i]; in ssip_skb_to_msg() 215 msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags); in ssip_alloc_data()
|
/linux-4.4.14/drivers/atm/ |
D | eni.c | 1101 DPRINTK("iovcnt = %d\n",skb_shinfo(skb)->nr_frags); in do_tx() 1102 if (!skb_shinfo(skb)->nr_frags) dma_size += 5; in do_tx() 1103 else dma_size += 5*(skb_shinfo(skb)->nr_frags+1); in do_tx() 1124 if (!skb_shinfo(skb)->nr_frags) in do_tx() 1129 for (i = -1; i < skb_shinfo(skb)->nr_frags; i++) in do_tx() 1136 skb_frag_page(&skb_shinfo(skb)->frags[i]) + in do_tx() 1137 skb_shinfo(skb)->frags[i].page_offset, in do_tx() 1138 skb_frag_size(&skb_shinfo(skb)->frags[i])); in do_tx()
|
D | he.c | 2514 if (skb_shinfo(skb)->nr_frags) { in he_send() 2558 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in he_send() 2559 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in he_send()
|
D | zatm.c | 642 EVENT("iovcnt=%d\n",skb_shinfo(skb)->nr_frags,0); in do_tx() 644 if (!skb_shinfo(skb)->nr_frags) { in do_tx()
|
/linux-4.4.14/drivers/net/ethernet/neterion/ |
D | s2io.h | 1137 #define s2io_tcp_mss(skb) skb_shinfo(skb)->gso_size 1138 #define s2io_udp_mss(skb) skb_shinfo(skb)->gso_size 1139 #define s2io_offload_type(skb) skb_shinfo(skb)->gso_type
|
D | s2io.c | 2340 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_txdl_getskb() 2344 const skb_frag_t *frag = &skb_shinfo(skb)->frags[j]; in s2io_txdl_getskb() 4154 (__force u64)skb_shinfo(skb)->ip6_frag_id; in s2io_xmit() 4157 (__force u64)skb_shinfo(skb)->ip6_frag_id << 32; in s2io_xmit() 4179 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_xmit() 4182 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in s2io_xmit() 8548 if (skb_shinfo(first)->frag_list) in lro_append_pkt() 8551 skb_shinfo(first)->frag_list = skb; in lro_append_pkt()
|
/linux-4.4.14/Documentation/networking/ |
D | driver.txt | 21 if (TX_BUFFS_AVAIL(dp) <= (skb_shinfo(skb)->nr_frags + 1)) {
|
D | timestamping.txt | 225 relevant sequence number in skb_shinfo(skb)->tskey. Because an skbuff 439 - In hard_start_xmit(), check if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) 444 SKBTX_IN_PROGRESS in skb_shinfo(skb)->tx_flags , e.g. with 446 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
|
D | netdev-features.txt | 124 NETIF_F_SG --- paged skbs (skb_shinfo()->frags), NETIF_F_FRAGLIST ---
|
/linux-4.4.14/drivers/net/ethernet/freescale/ |
D | gianfar.c | 1968 for (j = 0; j < skb_shinfo(tx_queue->tx_skbuff[i])->nr_frags; in free_skb_tx_queue() 2341 do_tstamp = (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in gfar_start_xmit() 2369 nr_frags = skb_shinfo(skb)->nr_frags; in gfar_start_xmit() 2416 frag_len = skb_shinfo(skb)->frags[i].size; in gfar_start_xmit() 2426 &skb_shinfo(skb)->frags[i], in gfar_start_xmit() 2477 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in gfar_start_xmit() 2681 frags = skb_shinfo(skb)->nr_frags; in gfar_clean_tx_ring() 2686 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in gfar_clean_tx_ring() 2700 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring() 2710 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring() [all …]
|
D | fec_main.c | 375 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_frag_skb() 387 this_frag = &skb_shinfo(skb)->frags[frag]; in fec_enet_txq_submit_frag_skb() 394 frag_len = skb_shinfo(skb)->frags[frag].size; in fec_enet_txq_submit_frag_skb() 401 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_frag_skb() 457 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_skb() 520 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_skb() 530 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in fec_enet_txq_submit_skb() 532 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in fec_enet_txq_submit_skb() 724 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in fec_enet_txq_submit_tso() 1261 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS) && in fec_enet_tx_queue()
|
/linux-4.4.14/drivers/net/ethernet/emulex/benet/ |
D | be_main.c | 684 u64 tx_pkts = skb_shinfo(skb)->gso_segs ? : 1; in be_tx_stats_update() 699 return 1 + (skb_headlen(skb) ? 1 : 0) + skb_shinfo(skb)->nr_frags; in skb_wrb_cnt() 773 wrb_params->lso_mss = skb_shinfo(skb)->gso_size; in be_get_wrb_params_from_skb() 950 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in be_xmit_enqueue() 951 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in be_xmit_enqueue() 2055 skb_shinfo(skb)->nr_frags = 1; in skb_fill_rx_data() 2057 skb_shinfo(skb)->frags[0].page_offset = in skb_fill_rx_data() 2059 skb_frag_size_set(&skb_shinfo(skb)->frags[0], in skb_fill_rx_data() 2083 skb_shinfo(skb)->frags[j].page_offset = in skb_fill_rx_data() 2085 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); in skb_fill_rx_data() [all …]
|
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_main.c | 1291 int nr_frags = skb_shinfo(skb)->nr_frags; in sxgbe_xmit() 1294 u16 cur_mss = skb_shinfo(skb)->gso_size; in sxgbe_xmit() 1304 ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit() 1369 const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num]; in sxgbe_xmit() 1428 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit() 1431 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in sxgbe_xmit()
|
/linux-4.4.14/drivers/net/ethernet/atheros/atl1e/ |
D | atl1e_main.c | 1617 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atl1e_cal_tdp_req() 1618 fg_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in atl1e_cal_tdp_req() 1624 (skb_shinfo(skb)->gso_type == SKB_GSO_TCPV6)) { in atl1e_cal_tdp_req() 1652 offload_type = skb_shinfo(skb)->gso_type; in atl1e_tso_csum() 1680 tpd->word3 |= ((skb_shinfo(skb)->gso_size) & in atl1e_tso_csum() 1725 nr_frags = skb_shinfo(skb)->nr_frags; in atl1e_tx_map() 1794 frag = &skb_shinfo(skb)->frags[f]; in atl1e_tx_map()
|
/linux-4.4.14/drivers/net/ethernet/intel/igbvf/ |
D | netdev.c | 318 if (!skb_shinfo(skb)->nr_frags) { in igbvf_clean_rx_irq() 332 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in igbvf_clean_rx_irq() 830 segs = skb_shinfo(skb)->gso_segs ?: 1; in igbvf_clean_tx_irq() 1973 mss_l4len_idx = (skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT); in igbvf_tso() 2103 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in igbvf_tx_map_adv() 2111 frag = &skb_shinfo(skb)->frags[f]; in igbvf_tx_map_adv() 2242 if (igbvf_maybe_stop_tx(netdev, skb_shinfo(skb)->nr_frags + 4)) { in igbvf_xmit_frame_ring_adv()
|
/linux-4.4.14/net/packet/ |
D | af_packet.c | 1934 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_sendmsg_spkt() 2323 ph = skb_shinfo(skb)->destructor_arg; in tpacket_destruct_skb() 2359 sock_tx_timestamp(&po->sk, &skb_shinfo(skb)->tx_flags); in tpacket_fill_skb() 2360 skb_shinfo(skb)->destructor_arg = ph.raw; in tpacket_fill_skb() 2440 nr_frags = skb_shinfo(skb)->nr_frags; in tpacket_fill_skb() 2771 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_snd() 2796 skb_shinfo(skb)->gso_size = in packet_snd() 2798 skb_shinfo(skb)->gso_type = gso_type; in packet_snd() 2801 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in packet_snd() 2802 skb_shinfo(skb)->gso_segs = 0; in packet_snd() [all …]
|
/linux-4.4.14/drivers/net/ethernet/pasemi/ |
D | pasemi_mac.c | 302 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in pasemi_mac_unmap_tx_skb() 573 nfrags = skb_shinfo(info->skb)->nr_frags; in pasemi_mac_free_tx_resources() 1398 const int nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_queue_csdesc() 1494 nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_start_tx() 1503 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in pasemi_mac_start_tx()
|
/linux-4.4.14/drivers/net/ethernet/tile/ |
D | tilegx.c | 461 struct skb_shared_info *shtx = skb_shinfo(skb); in tile_tx_timestamp() 1678 struct skb_shared_info *sh = skb_shinfo(skb); in tso_count_edescs() 1724 struct skb_shared_info *sh = skb_shinfo(skb); in tso_headers_prepare() 1833 struct skb_shared_info *sh = skb_shinfo(skb); in tso_egress() 1981 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags()
|
D | tilepro.c | 1604 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags() 1677 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_tso() 1851 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx()
|
/linux-4.4.14/drivers/net/ethernet/synopsys/ |
D | dwc_eth_qos.c | 1987 if (skb_is_gso(skb) && skb_shinfo(skb)->gso_size != lp->gso_size) in dwceqos_tx_prepare() 1990 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in dwceqos_tx_prepare() 1991 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in dwceqos_tx_prepare() 2015 if (skb_is_gso(skb) && skb_shinfo(skb)->gso_size != lp->gso_size) { in dwceqos_tx_linear() 2016 dwceqos_dmadesc_set_ctx(lp, skb_shinfo(skb)->gso_size); in dwceqos_tx_linear() 2017 lp->gso_size = skb_shinfo(skb)->gso_size; in dwceqos_tx_linear() 2083 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in dwceqos_tx_frags() 2084 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in dwceqos_tx_frags()
|
/linux-4.4.14/drivers/net/ethernet/marvell/ |
D | sky2.c | 1215 for (i = 0; i < skb_shinfo(re->skb)->nr_frags; i++) in sky2_rx_submit() 1232 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_rx_map_skb() 1233 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_rx_map_skb() 1247 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_map_skb() 1269 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) in sky2_rx_unmap_skb() 1271 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_unmap_skb() 1807 count = (skb_shinfo(skb)->nr_frags + 1) in tx_le_req() 1877 mss = skb_shinfo(skb)->gso_size; in sky2_xmit_frame() 1949 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_xmit_frame() 1950 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_xmit_frame() [all …]
|
D | mv643xx_eth.c | 675 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in has_tiny_unaligned_frags() 676 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in has_tiny_unaligned_frags() 863 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in txq_submit_tso() 913 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_frag_skb() 921 this_frag = &skb_shinfo(skb)->frags[frag]; in txq_submit_frag_skb() 952 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_skb()
|
D | mvneta.c | 1698 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in mvneta_tx_tso() 1748 int i, nr_frags = skb_shinfo(skb)->nr_frags; in mvneta_tx_frag_process() 1751 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvneta_tx_frag_process() 1816 frags = skb_shinfo(skb)->nr_frags + 1; in mvneta_tx()
|
D | skge.c | 2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame() 2780 if (!skb_shinfo(skb)->nr_frags) /* single buffer i.e. no fragments */ in skge_xmit_frame() 2786 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skge_xmit_frame() 2787 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skge_xmit_frame()
|
D | mvpp2.c | 5207 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mvpp2_tx_frag_process() 5208 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvpp2_tx_frag_process() 5226 if (i == (skb_shinfo(skb)->nr_frags - 1)) { in mvpp2_tx_frag_process() 5268 frags = skb_shinfo(skb)->nr_frags + 1; in mvpp2_tx()
|
/linux-4.4.14/drivers/net/ethernet/adi/ |
D | bfin_mac.c | 836 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) { in bfin_tx_hwtstamp() 842 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in bfin_tx_hwtstamp() 1159 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit() 1169 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit()
|
/linux-4.4.14/drivers/scsi/bnx2fc/ |
D | bnx2fc_fcoe.c | 329 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in bnx2fc_xmit() 373 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in bnx2fc_xmit() 374 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in bnx2fc_xmit() 376 skb_shinfo(skb)->gso_type = 0; in bnx2fc_xmit() 377 skb_shinfo(skb)->gso_size = 0; in bnx2fc_xmit()
|
/linux-4.4.14/drivers/net/ethernet/qlogic/qede/ |
D | qede_main.c | 239 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++, bds_consumed++) { in qede_free_tx_pkt() 485 cpu_to_le16(skb_shinfo(skb)->gso_size); in qede_start_xmit() 527 while (tx_data_bd && frag_idx < skb_shinfo(skb)->nr_frags) { in qede_start_xmit() 529 &skb_shinfo(skb)->frags[frag_idx], in qede_start_xmit() 546 for (; frag_idx < skb_shinfo(skb)->nr_frags; frag_idx++, nbd++) { in qede_start_xmit() 553 &skb_shinfo(skb)->frags[frag_idx], in qede_start_xmit()
|
/linux-4.4.14/net/openvswitch/ |
D | flow.c | 528 skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract() 649 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract()
|
/linux-4.4.14/drivers/net/ethernet/ti/ |
D | netcp_core.c | 686 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in netcp_process_one_rx_packet() 1047 prefetchw(skb_shinfo(skb)); in netcp_tx_map_skb() 1056 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in netcp_tx_map_skb() 1057 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in netcp_tx_map_skb() 1096 if (skb_shinfo(skb)->frag_list) { in netcp_tx_map_skb()
|
D | cpts.c | 343 if (!(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in cpts_tx_timestamp()
|
D | cpsw.c | 1400 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in cpsw_ndo_start_xmit() 1402 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cpsw_ndo_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/tehuti/ |
D | tehuti.c | 1503 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_map_skb() 1520 frag = &skb_shinfo(skb)->frags[i]; in bdx_tx_map_skb() 1626 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_transmit() 1646 if (skb_shinfo(skb)->gso_size) { in bdx_tx_transmit() 1647 txd_mss = skb_shinfo(skb)->gso_size; in bdx_tx_transmit()
|
/linux-4.4.14/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_main.c | 1902 skb_shinfo(skb)->gso_size > 0) { in netxen_tso_check() 1906 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in netxen_tso_check() 2006 nr_frags = skb_shinfo(skb)->nr_frags; in netxen_map_tx_skb() 2018 frag = &skb_shinfo(skb)->frags[i]; in netxen_map_tx_skb() 2071 frag_count = skb_shinfo(skb)->nr_frags + 1; in netxen_nic_xmit_frame() 2079 frag = &skb_shinfo(skb)->frags[i]; in netxen_nic_xmit_frame() 2086 frag_count = 1 + skb_shinfo(skb)->nr_frags; in netxen_nic_xmit_frame()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt.c | 192 if (unlikely(free_size < skb_shinfo(skb)->nr_frags + 2)) { in bnxt_start_xmit() 199 last_frag = skb_shinfo(skb)->nr_frags; in bnxt_start_xmit() 249 skb_frag_t *frag = &skb_shinfo(skb)->frags[j]; in bnxt_start_xmit() 323 length = skb_shinfo(skb)->gso_size; in bnxt_start_xmit() 339 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in bnxt_start_xmit() 410 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnxt_start_xmit() 453 skb_frag_size(&skb_shinfo(skb)->frags[j]), in bnxt_tx_int() 711 shinfo = skb_shinfo(skb); in bnxt_rx_pages() 861 skb_shinfo(skb)->gso_size = in bnxt_gro_skb() 863 skb_shinfo(skb)->gso_type = tpa_info->gso_type; in bnxt_gro_skb() [all …]
|
/linux-4.4.14/drivers/net/ethernet/atheros/atl1c/ |
D | atl1c_main.c | 1966 tpd_req = skb_shinfo(skb)->nr_frags + 1; in atl1c_cal_tpd_req() 1972 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in atl1c_cal_tpd_req() 1995 offload_type = skb_shinfo(skb)->gso_type; in atl1c_tso_csum() 2050 (*tpd)->word1 |= (skb_shinfo(skb)->gso_size & TPD_MSS_MASK) << in atl1c_tso_csum() 2114 nr_frags = skb_shinfo(skb)->nr_frags; in atl1c_tx_map() 2164 frag = &skb_shinfo(skb)->frags[f]; in atl1c_tx_map()
|
/linux-4.4.14/drivers/net/ethernet/cisco/enic/ |
D | enic_main.c | 433 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_cont() 559 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_tso() 587 unsigned int mss = skb_shinfo(skb)->gso_size; in enic_queue_wq_skb() 652 if (skb_shinfo(skb)->gso_size == 0 && in enic_hard_start_xmit() 653 skb_shinfo(skb)->nr_frags + 1 > ENIC_NON_TSO_MAX_DESC && in enic_hard_start_xmit() 662 skb_shinfo(skb)->nr_frags + ENIC_DESC_MAX_SPLITS) { in enic_hard_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/3com/ |
D | typhoon.c | 325 #define skb_tso_size(x) (skb_shinfo(x)->gso_size) 740 numDesc = skb_shinfo(skb)->nr_frags + 1; in typhoon_start_tx() 793 if(skb_shinfo(skb)->nr_frags == 0) { in typhoon_start_tx() 813 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in typhoon_start_tx() 814 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in typhoon_start_tx()
|
D | 3c59x.c | 2170 if (!skb_shinfo(skb)->nr_frags) { in boomerang_start_xmit() 2189 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in boomerang_start_xmit() 2190 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in boomerang_start_xmit() 2214 if (i == skb_shinfo(skb)->nr_frags-1) in boomerang_start_xmit() 2462 for (i=0; i<=skb_shinfo(skb)->nr_frags; i++) in boomerang_interrupt() 2795 for (k=0; k<=skb_shinfo(skb)->nr_frags; k++) in vortex_close()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 812 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbevf_add_rx_frag() 3329 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbevf_tso() 3334 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbevf_tso() 3465 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbevf_tx_map() 3640 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbevf_xmit_frame() 3641 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbevf_xmit_frame() 3643 count += skb_shinfo(skb)->nr_frags; in ixgbevf_xmit_frame()
|
/linux-4.4.14/drivers/net/ethernet/atheros/atlx/ |
D | atl1.c | 2125 if (skb_shinfo(skb)->gso_size) { in atl1_tso() 2170 ptpd->word3 |= (skb_shinfo(skb)->gso_size & in atl1_tso() 2220 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_tx_map() 2282 frag = &skb_shinfo(skb)->frags[f]; in atl1_tx_map() 2381 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_xmit_frame() 2383 unsigned int f_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in atl1_xmit_frame() 2388 mss = skb_shinfo(skb)->gso_size; in atl1_xmit_frame()
|
/linux-4.4.14/drivers/net/ethernet/adaptec/ |
D | starfire.c | 162 #define skb_num_frags(skb) (skb_shinfo(skb)->nr_frags + 1) 1230 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[i - 1]; in start_tx() 1349 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in intr_handler() 1352 skb_frag_size(&skb_shinfo(skb)->frags[i]), in intr_handler()
|
/linux-4.4.14/drivers/net/ethernet/ |
D | jme.c | 1981 nr_alloc = skb_shinfo(skb)->nr_frags + 2; in jme_alloc_txdesc() 2059 int i, nr_frags = skb_shinfo(skb)->nr_frags; in jme_map_tx_skb() 2066 frag = &skb_shinfo(skb)->frags[i]; in jme_map_tx_skb() 2097 *mss = cpu_to_le16(skb_shinfo(skb)->gso_size << TXDESC_MSS_SHIFT); in jme_tx_tso() 2207 txbi->nr_desc = skb_shinfo(skb)->nr_frags + 2; in jme_fill_tx_desc() 2282 idx, skb_shinfo(skb)->nr_frags + 2, jiffies); in jme_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/atheros/alx/ |
D | main.c | 1026 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in alx_map_tx_skb() 1029 frag = &skb_shinfo(skb)->frags[f]; in alx_map_tx_skb() 1074 int tpdreq = skb_shinfo(skb)->nr_frags + 1; in alx_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgb/ |
D | ixgb_main.c | 1235 mss = skb_shinfo(skb)->gso_size; in ixgb_tso() 1331 unsigned int mss = skb_shinfo(skb)->gso_size; in ixgb_tx_map() 1332 unsigned int nr_frags = skb_shinfo(skb)->nr_frags; in ixgb_tx_map() 1369 frag = &skb_shinfo(skb)->frags[f]; in ixgb_tx_map()
|
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/ |
D | stmmac_main.c | 355 if (likely(!skb || !(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS))) in stmmac_get_tx_hwtstamp() 1952 int nfrags = skb_shinfo(skb)->nr_frags; in stmmac_xmit() 2003 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in stmmac_xmit() 2073 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in stmmac_xmit() 2076 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in stmmac_xmit()
|
/linux-4.4.14/drivers/net/ethernet/intel/e1000/ |
D | e1000_main.c | 2719 mss = skb_shinfo(skb)->gso_size; in e1000_tso() 2903 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map() 2948 segs = skb_shinfo(skb)->gso_segs ?: 1; in e1000_tx_map() 3136 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame() 3202 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame() 3204 count += TXD_USE_COUNT(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame() 4208 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq() 4217 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq()
|
/linux-4.4.14/drivers/net/ethernet/xilinx/ |
D | ll_temac_main.c | 685 num_frag = skb_shinfo(skb)->nr_frags; in temac_start_xmit() 686 frag = &skb_shinfo(skb)->frags[0]; in temac_start_xmit()
|
D | xilinx_axienet_main.c | 666 num_frag = skb_shinfo(skb)->nr_frags; in axienet_start_xmit() 698 frag = &skb_shinfo(skb)->frags[ii]; in axienet_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/hisilicon/hns/ |
D | hns_enet.c | 121 buf_num = skb_shinfo(skb)->nr_frags + 1; in hns_nic_net_xmit_hw() 139 assert(skb_shinfo(skb)->nr_frags == 1); in hns_nic_net_xmit_hw() 159 frag = &skb_shinfo(skb)->frags[i - 1]; in hns_nic_net_xmit_hw()
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlxsw/ |
D | pci.c | 1528 if (skb_shinfo(skb)->nr_frags > MLXSW_PCI_WQE_SG_ENTRIES - 1) { in mlxsw_pci_skb_transmit() 1554 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlxsw_pci_skb_transmit() 1555 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mlxsw_pci_skb_transmit()
|
/linux-4.4.14/drivers/net/ethernet/via/ |
D | via-velocity.c | 1738 if (skb_shinfo(skb)->nr_frags > 0) in velocity_free_tx_buf() 2549 if (skb_shinfo(skb)->nr_frags > 6 && __skb_linearize(skb)) { in velocity_xmit() 2554 pktlen = skb_shinfo(skb)->nr_frags == 0 ? in velocity_xmit() 2580 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in velocity_xmit() 2581 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in velocity_xmit()
|
/linux-4.4.14/drivers/scsi/cxgbi/ |
D | libcxgbi.c | 1143 int frags = skb_shinfo(skb)->nr_frags + in cxgbi_sock_send_pdus() 1155 csk, skb_shinfo(skb)->nr_frags, skb->len, in cxgbi_sock_send_pdus() 2225 skb_shinfo(skb)->nr_frags = tdata->nr_frags; in cxgbi_conn_init_pdu() 2243 i = skb_shinfo(skb)->nr_frags; in cxgbi_conn_init_pdu() 2244 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in cxgbi_conn_init_pdu()
|
/linux-4.4.14/drivers/net/wireless/iwlwifi/pcie/ |
D | tx.c | 1823 skb_shinfo(skb)->nr_frags > IWL_PCIE_MAX_FRAGS && in iwl_trans_pcie_tx() 1914 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in iwl_trans_pcie_tx() 1915 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iwl_trans_pcie_tx()
|
/linux-4.4.14/net/tipc/ |
D | msg.c | 164 skb_shinfo(head)->frag_list = frag; in tipc_buf_append()
|
/linux-4.4.14/drivers/net/ethernet/intel/e1000e/ |
D | netdev.c | 1582 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq() 1594 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq() 5357 mss = skb_shinfo(skb)->gso_size; in e1000_tso() 5501 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map() 5529 segs = skb_shinfo(skb)->gso_segs ? : 1; in e1000_tx_map() 5724 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame() 5756 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame() 5758 count += DIV_ROUND_UP(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame() 5803 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in e1000_xmit_frame() 5806 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in e1000_xmit_frame()
|
/linux-4.4.14/drivers/net/ethernet/alteon/ |
D | acenic.c | 2424 if (!skb_shinfo(skb)->nr_frags) { in ace_start_xmit() 2462 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ace_start_xmit() 2463 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ace_start_xmit() 2479 if (i == skb_shinfo(skb)->nr_frags - 1) { in ace_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/cadence/ |
D | macb.c | 1169 unsigned int f, nr_frags = skb_shinfo(skb)->nr_frags; in macb_tx_map() 1201 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in macb_tx_map() 1311 nr_frags = skb_shinfo(skb)->nr_frags; in macb_start_xmit() 1313 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in macb_start_xmit()
|
/linux-4.4.14/drivers/usb/gadget/function/ |
D | f_phonet.c | 349 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in pn_rx_complete()
|
/linux-4.4.14/drivers/net/wireless/b43legacy/ |
D | pio.c | 197 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in pio_tx_write_fragment()
|
D | dma.c | 1024 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in dma_tx_fragment()
|
/linux-4.4.14/drivers/net/ethernet/tundra/ |
D | tsi108_eth.c | 655 int frags = skb_shinfo(skb)->nr_frags + 1; in tsi108_send_packet() 709 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in tsi108_send_packet()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/ |
D | bnx2.c | 2901 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_tx_int() 2960 shinfo = skb_shinfo(skb); in bnx2_reuse_rx_skb_pages() 3087 &skb_shinfo(skb)->frags[i - 1]; in bnx2_rx_skb() 5454 skb_frag_size(&skb_shinfo(skb)->frags[k]), in bnx2_free_tx_skbs() 6601 (skb_shinfo(skb)->nr_frags + 1))) { in bnx2_start_xmit() 6621 if ((mss = skb_shinfo(skb)->gso_size)) { in bnx2_start_xmit() 6629 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) { in bnx2_start_xmit() 6672 last_frag = skb_shinfo(skb)->nr_frags; in bnx2_start_xmit() 6677 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in bnx2_start_xmit() 6746 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_start_xmit()
|
D | bgmac.c | 159 nr_frags = skb_shinfo(skb)->nr_frags; in bgmac_dma_tx_add() 183 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bgmac_dma_tx_add()
|
D | tg3.c | 6567 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in tg3_tx() 6574 skb_frag_size(&skb_shinfo(skb)->frags[i]), in tg3_tx() 7764 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in tg3_tx_skb_unmap() 7841 return skb_shinfo(skb)->gso_segs < tnapi->tx_pending / 3; in tg3_tso_bug_gso_check() 7853 u32 frag_cnt_est = skb_shinfo(skb)->gso_segs * 3; in tg3_tso_bug() 7917 if (unlikely(budget <= (skb_shinfo(skb)->nr_frags + 1))) { in tg3_start_xmit() 7931 mss = skb_shinfo(skb)->gso_size; in tg3_start_xmit() 8027 if ((unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) && in tg3_start_xmit() 8029 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in tg3_start_xmit() 8049 ((skb_shinfo(skb)->nr_frags == 0) ? TXD_FLAG_END : 0), in tg3_start_xmit() [all …]
|
/linux-4.4.14/net/can/ |
D | raw.c | 758 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_sendmsg()
|
/linux-4.4.14/net/appletalk/ |
D | ddp.c | 950 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atalk_sum_skb() 952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in atalk_sum_skb()
|
/linux-4.4.14/drivers/net/ethernet/intel/igb/ |
D | igb_main.c | 4691 first->gso_segs = skb_shinfo(skb)->gso_segs; in igb_tso() 4696 mss_l4len_idx |= skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT; in igb_tso() 4887 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in igb_tx_map() 5011 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in igb_xmit_frame_ring() 5012 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in igb_xmit_frame_ring() 5025 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) { in igb_xmit_frame_ring() 5030 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in igb_xmit_frame_ring() 6694 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in igb_add_rx_frag()
|
/linux-4.4.14/drivers/net/ethernet/nvidia/ |
D | forcedeth.c | 2190 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit() 2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit() 2256 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit() 2309 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit() 2337 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit_optimized() 2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit_optimized() 2406 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit_optimized() 2459 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit_optimized()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
D | ehea_main.c | 1644 swqe->mss = skb_shinfo(skb)->gso_size; in write_swqe2_immediate() 1676 nfrags = skb_shinfo(skb)->nr_frags; in write_swqe2_data() 1687 frag = &skb_shinfo(skb)->frags[0]; in write_swqe2_data() 1700 frag = &skb_shinfo(skb)->frags[i]; in write_swqe2_data()
|
/linux-4.4.14/drivers/net/ethernet/myricom/myri10ge/ |
D | myri10ge.c | 1410 frag = skb_shinfo(skb)->frags; in myri10ge_vlan_rx() 1464 rx_frags = skb_shinfo(skb)->frags; in myri10ge_rx_done() 2846 mss = skb_shinfo(skb)->gso_size; in myri10ge_xmit() 2929 frag_cnt = skb_shinfo(skb)->nr_frags; in myri10ge_xmit() 3024 frag = &skb_shinfo(skb)->frags[frag_idx]; in myri10ge_xmit()
|
/linux-4.4.14/drivers/scsi/libfc/ |
D | fc_fcp.c | 651 skb_shinfo(fp_skb(fp))->nr_frags, in fc_fcp_send_data() 672 if ((skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN) && in fc_fcp_send_data()
|
/linux-4.4.14/drivers/net/ethernet/calxeda/ |
D | xgmac.c | 1090 int nfrags = skb_shinfo(skb)->nr_frags; in xgmac_xmit() 1115 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xgmac_xmit()
|
/linux-4.4.14/net/mac80211/ |
D | tx.c | 2261 skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS)) { in ieee80211_build_hdr() 2718 if (skb->sk && skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS) in ieee80211_xmit_fast() 2781 if (skb_shinfo(skb)->gso_size) in ieee80211_xmit_fast() 2783 DIV_ROUND_UP(skb->len, skb_shinfo(skb)->gso_size); in ieee80211_xmit_fast()
|
/linux-4.4.14/net/wireless/ |
D | wext-core.c | 628 skb_shinfo(skb)->frag_list = compskb; in wireless_send_event()
|
/linux-4.4.14/drivers/net/ethernet/qlogic/qlge/ |
D | qlge_main.c | 1318 int frag_cnt = skb_shinfo(skb)->nr_frags; in ql_map_send() 1351 skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_idx]; in ql_map_send() 1519 __skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in ql_process_mac_rx_gro_page() 1527 skb_shinfo(skb)->nr_frags++; in ql_process_mac_rx_gro_page() 2591 mac_iocb_ptr->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in ql_tso()
|
/linux-4.4.14/drivers/net/ethernet/brocade/bna/ |
D | bnad.c | 573 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in bnad_cq_setup_skb_frags() 2844 gso_size = skb_shinfo(skb)->gso_size; in bnad_txq_wi_prepare() 2984 vectors = 1 + skb_shinfo(skb)->nr_frags; in bnad_start_xmit() 3052 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bnad_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/renesas/ |
D | ravb_main.c | 1367 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in ravb_start_xmit() 1404 return (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) ? RAVB_NC : in ravb_select_queue()
|
/linux-4.4.14/net/netlink/ |
D | af_netlink.c | 909 !atomic_dec_return(&(skb_shinfo(skb)->dataref))) in netlink_skb_destructor() 2541 if (unlikely(skb_shinfo(skb)->frag_list)) { in netlink_recvmsg() 2553 data_skb = skb_shinfo(skb)->frag_list; in netlink_recvmsg()
|
/linux-4.4.14/drivers/net/ethernet/natsemi/ |
D | ns83820.c | 1074 nr_frags = skb_shinfo(skb)->nr_frags; 1112 frag = skb_shinfo(skb)->frags;
|
/linux-4.4.14/drivers/s390/net/ |
D | qeth_core_main.c | 3821 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in qeth_get_elements_for_frags() 3822 frag = &skb_shinfo(skb)->frags[cnt]; in qeth_get_elements_for_frags() 3906 if (skb_shinfo(skb)->nr_frags) in __qeth_fill_buffer() 3927 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in __qeth_fill_buffer() 3928 frag = &skb_shinfo(skb)->frags[cnt]; in __qeth_fill_buffer() 5258 card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; in qeth_core_get_next_skb()
|
D | qeth_l3_main.c | 2799 hdr->ext.mss = skb_shinfo(skb)->gso_size; in qeth_tso_fill_header() 2874 (skb_shinfo(skb)->nr_frags == 0)) { in qeth_l3_hard_start_xmit() 2961 nr_frags = skb_shinfo(new_skb)->nr_frags; in qeth_l3_hard_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/qlogic/ |
D | qla3xxx.c | 2315 int frag_cnt = (int)skb_shinfo(skb)->nr_frags; in ql_send_map() 2348 skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs]; in ql_send_map() 2474 skb_shinfo(skb)->nr_frags); in ql3xxx_send()
|
/linux-4.4.14/drivers/net/ethernet/faraday/ |
D | ftgmac100.c | 466 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, 0, size); in ftgmac100_rx_packet()
|
/linux-4.4.14/net/ |
D | Kconfig | 41 achieve this, you need to set skb_shinfo(skb)->frag_list to the
|
/linux-4.4.14/net/iucv/ |
D | af_iucv.c | 2246 if (skb_shinfo(list_skb) == skb_shinfo(skb)) { in afiucv_hs_callback_txnotify()
|
/linux-4.4.14/drivers/net/wireless/mwifiex/ |
D | main.c | 851 skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS && in mwifiex_hard_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/altera/ |
D | altera_tse_main.c | 572 int nfrags = skb_shinfo(skb)->nr_frags; in tse_start_xmit()
|
/linux-4.4.14/drivers/net/ethernet/ibm/emac/ |
D | core.c | 1449 int nr_frags = skb_shinfo(skb)->nr_frags; in emac_start_xmit_sg() 1484 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in emac_start_xmit_sg()
|
/linux-4.4.14/drivers/scsi/cxgbi/cxgb3i/ |
D | cxgb3i.c | 408 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); in push_tx_frames()
|
/linux-4.4.14/drivers/net/phy/ |
D | dp83640.c | 1460 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in dp83640_txtstamp()
|
/linux-4.4.14/drivers/net/ethernet/aurora/ |
D | nb8800.c | 267 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in nb8800_receive()
|
/linux-4.4.14/drivers/net/ethernet/octeon/ |
D | octeon_mgmt.c | 1282 re.s.tstamp = ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) != 0); in octeon_mgmt_xmit()
|
/linux-4.4.14/drivers/net/ethernet/xscale/ |
D | ixp4xx_eth.c | 334 shtx = skb_shinfo(skb); in ixp_tx_timestamp()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/genet/ |
D | bcmgenet.c | 1449 nr_frags = skb_shinfo(skb)->nr_frags; in bcmgenet_xmit() 1490 &skb_shinfo(skb)->frags[i], in bcmgenet_xmit()
|
/linux-4.4.14/drivers/net/ethernet/agere/ |
D | et131x.c | 2427 u32 nr_frags = skb_shinfo(skb)->nr_frags + 1; in nic_send_packet() 2428 struct skb_frag_struct *frags = &skb_shinfo(skb)->frags[0]; in nic_send_packet()
|