Searched refs:frags (Results 1 - 200 of 213) sorted by relevance

12

/linux-4.1.27/include/net/netns/
H A Dieee802154_6lowpan.h18 struct netns_frags frags; member in struct:netns_ieee802154_lowpan
H A Dipv6.h44 struct netns_frags frags; member in struct:netns_ipv6
88 struct netns_frags frags; member in struct:netns_nf_frag
H A Dipv4.h58 struct netns_frags frags; member in struct:netns_ipv4
/linux-4.1.27/include/linux/ceph/
H A Dceph_frag.h7 * (subset of the number space) can be partitioned into 2^n sub-frags.
18 * can't sort encoded frags numerically. However, it does allow you
19 * to feed encoded frags as values into frag_contains_value.
104 * comparator to sort frags logically, as when traversing the
/linux-4.1.27/drivers/infiniband/hw/nes/
H A Dnes_mgt.c158 skb = fpdu_info->frags[i].skb; nes_download_callback()
159 if (fpdu_info->frags[i].cmplt) { nes_download_callback()
249 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; get_fpdu_info() local
276 memset(&frags, 0, sizeof frags); get_fpdu_info()
278 frags[i].physaddr = cb->busaddr; get_fpdu_info()
279 frags[i].physaddr += skb->data - cb->data_start; get_fpdu_info()
280 frags[i].frag_len = min(tmp_len, skb->len); get_fpdu_info()
281 frags[i].skb = skb; get_fpdu_info()
282 frags[i].cmplt = (skb->len == frags[i].frag_len); get_fpdu_info()
283 frag_tot += frags[i].frag_len; get_fpdu_info()
286 tmp_len -= frags[i].frag_len; get_fpdu_info()
297 skb_unlink(frags[i].skb, &nesqp->pau_list); get_fpdu_info()
298 nes_mgt_free_skb(nesdev, frags[i].skb, PCI_DMA_TODEVICE); get_fpdu_info()
301 frags[0].physaddr = cb->busaddr; get_fpdu_info()
302 frags[0].physaddr += skb->data - cb->data_start; get_fpdu_info()
303 frags[0].frag_len = skb->len; get_fpdu_info()
304 frags[0].skb = skb; get_fpdu_info()
305 frags[0].cmplt = true; get_fpdu_info()
314 frags[0].physaddr = cb->busaddr; get_fpdu_info()
315 frags[0].frag_len = 0; get_fpdu_info()
316 frags[0].skb = skb; get_fpdu_info()
317 frags[0].cmplt = true; get_fpdu_info()
336 cb = (struct nes_rskb_cb *)&frags[0].skb->cb[0]; get_fpdu_info()
343 if (frags[0].cmplt) { get_fpdu_info()
370 memcpy(fpdu_info->frags, frags, sizeof(fpdu_info->frags)); get_fpdu_info()
377 cb = (struct nes_rskb_cb *)&frags[i].skb->cb[0]; get_fpdu_info()
378 skb_pull(frags[i].skb, frags[i].frag_len); get_fpdu_info()
380 if (frags[i].skb->len == 0) { get_fpdu_info()
383 skb_unlink(frags[i].skb, &nesqp->pau_list); get_fpdu_info()
437 u32tmp = (fpdu_info->frags[1].frag_len << 16) | fpdu_info->frags[0].frag_len; forward_fpdus()
441 u32tmp = (fpdu_info->frags[3].frag_len << 16) | fpdu_info->frags[2].frag_len; forward_fpdus()
452 lower_32_bits(fpdu_info->frags[0].physaddr)); forward_fpdus()
454 upper_32_bits(fpdu_info->frags[0].physaddr)); forward_fpdus()
457 lower_32_bits(fpdu_info->frags[1].physaddr)); forward_fpdus()
459 upper_32_bits(fpdu_info->frags[1].physaddr)); forward_fpdus()
462 lower_32_bits(fpdu_info->frags[2].physaddr)); forward_fpdus()
464 upper_32_bits(fpdu_info->frags[2].physaddr)); forward_fpdus()
467 lower_32_bits(fpdu_info->frags[3].physaddr)); forward_fpdus()
469 upper_32_bits(fpdu_info->frags[3].physaddr)); forward_fpdus()
H A Dnes_mgt.h82 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; member in struct:pau_fpdu_info
H A Dnes_nic.c440 &skb_shinfo(skb)->frags[skb_fragment_index]; nes_nic_send()
445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); nes_nic_send()
498 * " (%u frags), tso_size=%u\n", nes_netdev_start_xmit()
551 nes_debug(NES_DBG_NIC_TX, "%s: HNIC SQ full- TSO request has too many frags!\n", nes_netdev_start_xmit()
560 &skb_shinfo(skb)->frags[tso_frag_count]; nes_netdev_start_xmit()
596 " (%u frags), is_gso = %u tso_size=%u\n", nes_netdev_start_xmit()
636 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index])); nes_netdev_start_xmit()
640 tso_wqe_length += skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index++]); nes_netdev_start_xmit()
H A Dnes_hw.h912 struct nes_first_frag *first_frag_vbase; /* virtual address of first frags */
/linux-4.1.27/drivers/staging/unisys/uislib/
H A Duisutils.c104 * struct phys_info frags[])
112 * unsigned int frags_max - input - max len of frags array
113 * struct phys_info frags[] - output - frags array filled in on output
115 * entries filled in frags
/linux-4.1.27/net/ieee802154/6lowpan/
H A Dreassembly.c33 static const char lowpan_frags_cache_name[] = "lowpan-frags";
100 net = container_of(fq->q.net, struct net, ieee802154_lowpan.frags); lowpan_frag_expire()
131 q = inet_frag_find(&ieee802154_lowpan->frags, fq_find()
285 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); lowpan_frag_reasm()
398 .data = &init_net.ieee802154_lowpan.frags.high_thresh,
402 .extra1 = &init_net.ieee802154_lowpan.frags.low_thresh
406 .data = &init_net.ieee802154_lowpan.frags.low_thresh,
411 .extra2 = &init_net.ieee802154_lowpan.frags.high_thresh
415 .data = &init_net.ieee802154_lowpan.frags.timeout,
450 table[0].data = &ieee802154_lowpan->frags.high_thresh; lowpan_frags_ns_sysctl_register()
451 table[0].extra1 = &ieee802154_lowpan->frags.low_thresh; lowpan_frags_ns_sysctl_register()
452 table[0].extra2 = &init_net.ieee802154_lowpan.frags.high_thresh; lowpan_frags_ns_sysctl_register()
453 table[1].data = &ieee802154_lowpan->frags.low_thresh; lowpan_frags_ns_sysctl_register()
454 table[1].extra2 = &ieee802154_lowpan->frags.high_thresh; lowpan_frags_ns_sysctl_register()
455 table[2].data = &ieee802154_lowpan->frags.timeout; lowpan_frags_ns_sysctl_register()
527 ieee802154_lowpan->frags.high_thresh = IPV6_FRAG_HIGH_THRESH; lowpan_frags_init_net()
528 ieee802154_lowpan->frags.low_thresh = IPV6_FRAG_LOW_THRESH; lowpan_frags_init_net()
529 ieee802154_lowpan->frags.timeout = IPV6_FRAG_TIMEOUT; lowpan_frags_init_net()
531 inet_frags_init_net(&ieee802154_lowpan->frags); lowpan_frags_init_net()
542 inet_frags_exit_net(&ieee802154_lowpan->frags, &lowpan_frags); lowpan_frags_exit_net()
/linux-4.1.27/drivers/net/wireless/ath/wil6210/
H A Dtrace.h190 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags),
191 TP_ARGS(vring, index, len, frags),
194 __field(u8, frags)
200 __entry->frags = frags;
204 TP_printk("vring %d index %d len %d frags %d",
205 __entry->vring, __entry->index, __entry->len, __entry->frags)
H A Dtxrx.c1168 &skb_shinfo(skb)->frags[f]; __wil_tx_vring()
1223 nr_frags = f + 1; /* frags mapped + one for skb head */ __wil_tx_vring()
H A Ddebugfs.c814 &skb_shinfo(skb)->frags[i]; wil_seq_print_skb()
/linux-4.1.27/net/ipv4/
H A Dip_fragment.c58 static const char ip_frag_cache_name[] = "ip4-frags";
92 return sum_frag_mem_limit(&net->ipv4.frags); ip_frag_mem()
136 frags); ip4_frag_init()
185 net = container_of(qp->q.net, struct net, ipv4.frags); ip_expire()
250 q = inet_frag_find(&net->ipv4.frags, &ip4_frags, &arg, hash); ip_find()
279 net = container_of(qp->q.net, struct net, ipv4.frags); ip_frag_too_far()
505 struct net *net = container_of(qp->q.net, struct net, ipv4.frags); ip_frag_reasm()
570 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); ip_frag_reasm()
708 .data = &init_net.ipv4.frags.high_thresh,
712 .extra1 = &init_net.ipv4.frags.low_thresh
716 .data = &init_net.ipv4.frags.low_thresh,
721 .extra2 = &init_net.ipv4.frags.high_thresh
725 .data = &init_net.ipv4.frags.timeout,
765 table[0].data = &net->ipv4.frags.high_thresh; ip4_frags_ns_ctl_register()
766 table[0].extra1 = &net->ipv4.frags.low_thresh; ip4_frags_ns_ctl_register()
767 table[0].extra2 = &init_net.ipv4.frags.high_thresh; ip4_frags_ns_ctl_register()
768 table[1].data = &net->ipv4.frags.low_thresh; ip4_frags_ns_ctl_register()
769 table[1].extra2 = &net->ipv4.frags.high_thresh; ip4_frags_ns_ctl_register()
770 table[2].data = &net->ipv4.frags.timeout; ip4_frags_ns_ctl_register()
835 net->ipv4.frags.high_thresh = 4 * 1024 * 1024; ipv4_frags_init_net()
836 net->ipv4.frags.low_thresh = 3 * 1024 * 1024; ipv4_frags_init_net()
842 net->ipv4.frags.timeout = IP_FRAG_TIME; ipv4_frags_init_net()
844 inet_frags_init_net(&net->ipv4.frags); ipv4_frags_init_net()
852 inet_frags_exit_net(&net->ipv4.frags, &ip4_frags); ipv4_frags_exit_net()
H A Dinet_lro.c156 lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); lro_init_desc()
H A Dtcp.c933 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); do_tcp_sendpages()
1212 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); tcp_sendmsg()
1510 * Try to splice more frags tcp_read_sock()
2965 const struct skb_frag_struct *f = &shi->frags[i]; tcp_md5_hash_skb_data()
H A Dudp.c737 struct sk_buff *frags; udp4_hwcsum() local
744 skb_walk_frags(skb, frags) { skb_walk_frags()
745 csum = csum_add(csum, frags->csum); skb_walk_frags()
746 hlen -= frags->len; skb_walk_frags()
H A Dip_output.c1067 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); __ip_append_data()
1281 skb_frag_size_add(&skb_shinfo(skb)->frags[i-1], len); ip_append_page()
H A Dtcp_output.c1250 int size = skb_frag_size(&shinfo->frags[i]); __pskb_trim_head()
1256 shinfo->frags[k] = shinfo->frags[i]; __pskb_trim_head()
1258 shinfo->frags[k].page_offset += eat; __pskb_trim_head()
1259 skb_frag_size_sub(&shinfo->frags[k], eat); __pskb_trim_head()
/linux-4.1.27/net/ipv6/
H A Dreassembly.c63 static const char ip6_frag_cache_name[] = "ip6-frags";
132 struct inet_frags *frags) ip6_expire_frag_queue()
141 inet_frag_kill(&fq->q, frags); ip6_expire_frag_queue()
169 inet_frag_put(&fq->q, frags); ip6_expire_frag_queue()
179 net = container_of(fq->q.net, struct net, ipv6.frags); ip6_frag_expire()
201 q = inet_frag_find(&net->ipv6.frags, &ip6_frags, &arg, hash); fq_find()
382 struct net *net = container_of(fq->q.net, struct net, ipv6.frags); ip6_frag_reasm()
444 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); ip6_frag_reasm()
593 .data = &init_net.ipv6.frags.high_thresh,
597 .extra1 = &init_net.ipv6.frags.low_thresh
601 .data = &init_net.ipv6.frags.low_thresh,
606 .extra2 = &init_net.ipv6.frags.high_thresh
610 .data = &init_net.ipv6.frags.timeout,
642 table[0].data = &net->ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register()
643 table[0].extra1 = &net->ipv6.frags.low_thresh; ip6_frags_ns_sysctl_register()
644 table[0].extra2 = &init_net.ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register()
645 table[1].data = &net->ipv6.frags.low_thresh; ip6_frags_ns_sysctl_register()
646 table[1].extra2 = &net->ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register()
647 table[2].data = &net->ipv6.frags.timeout; ip6_frags_ns_sysctl_register()
713 net->ipv6.frags.high_thresh = IPV6_FRAG_HIGH_THRESH; ipv6_frags_init_net()
714 net->ipv6.frags.low_thresh = IPV6_FRAG_LOW_THRESH; ipv6_frags_init_net()
715 net->ipv6.frags.timeout = IPV6_FRAG_TIMEOUT; ipv6_frags_init_net()
717 inet_frags_init_net(&net->ipv6.frags); ipv6_frags_init_net()
725 inet_frags_exit_net(&net->ipv6.frags, &ip6_frags); ipv6_frags_exit_net()
131 ip6_expire_frag_queue(struct net *net, struct frag_queue *fq, struct inet_frags *frags) ip6_expire_frag_queue() argument
H A Dudp.c990 struct sk_buff *frags = skb_shinfo(skb)->frag_list; udp6_hwcsum_outgoing() local
993 if (!frags) { udp6_hwcsum_outgoing()
1010 csum = csum_add(csum, frags->csum); udp6_hwcsum_outgoing()
1011 } while ((frags = frags->next)); udp6_hwcsum_outgoing()
H A Dip6_output.c1523 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); __ip6_append_data()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c92 struct mlx4_en_rx_alloc *frags, mlx4_en_alloc_frags()
116 frags[i] = ring_alloc[i]; mlx4_en_alloc_frags()
138 struct mlx4_en_rx_alloc *frags, mlx4_en_free_frag()
142 u32 next_frag_end = frags[i].page_offset + 2 * frag_info->frag_stride; mlx4_en_free_frag()
145 if (next_frag_end > frags[i].page_size) mlx4_en_free_frag()
146 dma_unmap_page(priv->ddev, frags[i].dma, frags[i].page_size, mlx4_en_free_frag()
149 if (frags[i].page) mlx4_en_free_frag()
150 put_page(frags[i].page); mlx4_en_free_frag()
166 en_dbg(DRV, priv, " frag %d allocator: - size:%d frags:%d\n", mlx4_en_init_allocator()
241 struct mlx4_en_rx_alloc *frags = ring->rx_info + mlx4_en_prepare_rx_desc() local
244 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc, gfp); mlx4_en_prepare_rx_desc()
262 struct mlx4_en_rx_alloc *frags; mlx4_en_free_rx_desc() local
265 frags = ring->rx_info + (index << priv->log_rx_info); mlx4_en_free_rx_desc()
268 mlx4_en_free_frag(priv, frags, nr); mlx4_en_free_rx_desc()
546 struct mlx4_en_rx_alloc *frags, mlx4_en_complete_rx_desc()
550 struct skb_frag_struct *skb_frags_rx = skb_shinfo(skb)->frags; mlx4_en_complete_rx_desc()
560 if (!frags[nr].page) mlx4_en_complete_rx_desc()
568 __skb_frag_set_page(&skb_frags_rx[nr], frags[nr].page); mlx4_en_complete_rx_desc()
570 skb_frags_rx[nr].page_offset = frags[nr].page_offset; mlx4_en_complete_rx_desc()
572 frags[nr].page = NULL; mlx4_en_complete_rx_desc()
591 struct mlx4_en_rx_alloc *frags, mlx4_en_rx_skb()
609 va = page_address(frags[0].page) + frags[0].page_offset; mlx4_en_rx_skb()
623 used_frags = mlx4_en_complete_rx_desc(priv, rx_desc, frags, mlx4_en_rx_skb()
637 skb_shinfo(skb)->frags[0].page_offset += pull_len; mlx4_en_rx_skb()
640 skb_frag_size_sub(&skb_shinfo(skb)->frags[0], pull_len); mlx4_en_rx_skb()
756 struct mlx4_en_rx_alloc *frags; mlx4_en_process_rx_cq() local
784 frags = ring->rx_info + (index << priv->log_rx_info); mlx4_en_process_rx_cq()
817 ethh = (struct ethhdr *)(page_address(frags[0].page) + mlx4_en_process_rx_cq()
818 frags[0].page_offset); mlx4_en_process_rx_cq()
891 rx_desc, frags, gro_skb,
897 void *va = skb_frag_address(skb_shinfo(gro_skb)->frags);
942 skb = mlx4_en_rx_skb(priv, rx_desc, frags, length);
993 mlx4_en_free_frag(priv, frags, nr);
90 mlx4_en_alloc_frags(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct mlx4_en_rx_alloc *ring_alloc, gfp_t gfp) mlx4_en_alloc_frags() argument
137 mlx4_en_free_frag(struct mlx4_en_priv *priv, struct mlx4_en_rx_alloc *frags, int i) mlx4_en_free_frag() argument
544 mlx4_en_complete_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct sk_buff *skb, int length) mlx4_en_complete_rx_desc() argument
589 mlx4_en_rx_skb(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, unsigned int length) mlx4_en_rx_skb() argument
H A Den_tx.c567 ptr = skb_frag_address_safe(&shinfo->frags[0]); is_inline()
654 skb_frag_size(&shinfo->frags[0])); build_inline_wqe()
675 skb_frag_size(&shinfo->frags[0])); build_inline_wqe()
805 frag = &shinfo->frags[i_frag]; mlx4_en_xmit()
/linux-4.1.27/net/core/
H A Dtso.c47 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; tso_build_data()
70 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; tso_start()
H A Dskbuff.c586 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_coalesce_rx_frag()
633 __skb_frag_unref(&shinfo->frags[i]); skb_release_data()
912 * skb_copy_ubufs - copy userspace skb frags buffers to kernel
917 * It will copy all frags into kernel and drop the reference
935 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; skb_copy_ubufs()
954 /* skb frags release userspace buffers */ skb_copy_ubufs()
960 /* skb frags point to kernel buffers */ skb_copy_ubufs()
963 skb_shinfo(skb)->frags[i].size); skb_copy_ubufs()
1134 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; __pskb_copy_fclone()
1197 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); pskb_expand_head()
1205 /* copy this zero copy skb frags */ pskb_expand_head()
1486 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); ___pskb_trim()
1493 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); ___pskb_trim()
1602 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); __pskb_pull_tail()
1669 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); __pskb_pull_tail()
1675 skb_shinfo(skb)->frags[k] = skb_shinfo(skb)->frags[i]; __pskb_pull_tail()
1677 skb_shinfo(skb)->frags[k].page_offset += eat; __pskb_pull_tail()
1678 skb_frag_size_sub(&skb_shinfo(skb)->frags[k], eat); __pskb_pull_tail()
1730 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; skb_copy_bits()
1912 const skb_frag_t *f = &skb_shinfo(skb)->frags[seg]; __skb_splice_bits()
2017 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_store_bits()
2091 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_checksum()
2181 end = start + skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_copy_and_csum_bits()
2185 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_and_csum_bits()
2264 * to the frags in the source buffer.
2271 * -ENOMEM: couldn't orphan frags of @from due to lack of memory
2318 skb_shinfo(to)->frags[j] = skb_shinfo(from)->frags[i]; skb_zerocopy()
2319 skb_shinfo(to)->frags[j].size = min_t(int, skb_shinfo(to)->frags[j].size, len); skb_zerocopy()
2320 len -= skb_shinfo(to)->frags[j].size; skb_zerocopy()
2528 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; skb_split_inside_header()
2552 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_split_no_header()
2555 skb_shinfo(skb1)->frags[k] = skb_shinfo(skb)->frags[i]; skb_split_no_header()
2567 skb_shinfo(skb1)->frags[0].page_offset += len - pos; skb_split_no_header()
2568 skb_frag_size_sub(&skb_shinfo(skb1)->frags[0], len - pos); skb_split_no_header()
2569 skb_frag_size_set(&skb_shinfo(skb)->frags[i], len - pos); skb_split_no_header()
2617 * If @tgt runs out of frags, the whole operation is aborted.
2623 * specialized skb free'er to handle frags without up-to-date nr_frags.
2636 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift()
2655 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift()
2656 fragto = &skb_shinfo(tgt)->frags[merge]; skb_shift()
2680 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift()
2681 fragto = &skb_shinfo(tgt)->frags[to]; skb_shift()
2708 fragfrom = &skb_shinfo(skb)->frags[0]; skb_shift()
2709 fragto = &skb_shinfo(tgt)->frags[merge]; skb_shift()
2718 skb_shinfo(skb)->frags[to++] = skb_shinfo(skb)->frags[from++]; skb_shift()
2813 frag = &skb_shinfo(st->cur_skb)->frags[st->frag_idx]; skb_seq_read()
3005 skb_frag_t *frag = skb_shinfo(head_skb)->frags; skb_segment()
3055 frag = skb_shinfo(list_skb)->frags; skb_segment()
3131 nskb_frag = skb_shinfo(nskb)->frags; skb_segment()
3145 frag = skb_shinfo(list_skb)->frags; skb_segment()
3156 "skb_segment: too many frags: %u %u\n", skb_segment()
3253 frag = pinfo->frags + nr_frags; skb_gro_receive()
3254 frag2 = skbinfo->frags + i; skb_gro_receive()
3274 skb_frag_t *frag = pinfo->frags + nr_frags; skb_gro_receive()
3292 memcpy(frag + 1, skbinfo->frags, sizeof(*frag) * skbinfo->nr_frags); skb_gro_receive()
3305 skbinfo->frags[0].page_offset += eat; skb_gro_receive()
3306 skb_frag_size_sub(&skbinfo->frags[0], eat); skb_gro_receive()
3383 end = start + skb_frag_size(&skb_shinfo(skb)->frags[i]); __skb_to_sgvec()
3385 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_to_sgvec()
4111 memcpy(skb_shinfo(to)->frags + skb_shinfo(to)->nr_frags, skb_try_coalesce()
4112 skb_shinfo(from)->frags, skb_try_coalesce()
4360 * alloc_skb_with_frags - allocate skb with page frags
4363 * @data_len: needed length in frags
4368 * This can be used to allocate a paged skb, given a maximal order for frags.
H A Dpktgen.c541 " frags: %d delay: %llu clone_skb: %d ifname: %s\n", pktgen_if_show()
972 if (!strcmp(name, "frags")) { pktgen_if_write()
979 sprintf(pg_result, "OK: frags=%u", pkt_dev->nfrags); pktgen_if_write()
2664 int frags = pkt_dev->nfrags; pktgen_finalize_skb() local
2669 if (frags > MAX_SKB_FRAGS) pktgen_finalize_skb()
2670 frags = MAX_SKB_FRAGS; pktgen_finalize_skb()
2671 len = datalen - frags * PAGE_SIZE; pktgen_finalize_skb()
2674 datalen = frags * PAGE_SIZE; pktgen_finalize_skb()
2678 frag_len = (datalen/frags) < PAGE_SIZE ? pktgen_finalize_skb()
2679 (datalen/frags) : PAGE_SIZE; pktgen_finalize_skb()
2692 skb_shinfo(skb)->frags[i].page_offset = 0; pktgen_finalize_skb()
2694 if (i == (frags - 1)) pktgen_finalize_skb()
2695 skb_frag_size_set(&skb_shinfo(skb)->frags[i], pktgen_finalize_skb()
2698 skb_frag_size_set(&skb_shinfo(skb)->frags[i], frag_len); pktgen_finalize_skb()
2699 datalen -= skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb()
2700 skb->len += skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb()
2701 skb->data_len += skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb()
H A Ddatagram.c378 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_datagram_iter()
465 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_datagram_from_iter()
521 * pages and build frags through them.
594 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_and_csum_datagram()
H A Ddev.c2534 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; illegal_highdma()
2546 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; illegal_highdma()
3999 const skb_frag_t *frag0 = &pinfo->frags[0]; skb_gro_reset_offset()
4024 pinfo->frags[0].page_offset += grow; gro_pull_from_frag0()
4025 skb_frag_size_sub(&pinfo->frags[0], grow); gro_pull_from_frag0()
4027 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { gro_pull_from_frag0()
4029 memmove(pinfo->frags, pinfo->frags + 1, gro_pull_from_frag0()
4030 --pinfo->nr_frags * sizeof(pinfo->frags[0])); gro_pull_from_frag0()
/linux-4.1.27/net/ipv6/netfilter/
H A Dnf_conntrack_reasm.c53 static const char nf_frags_cache_name[] = "nf-frags";
72 .data = &init_net.nf_frag.frags.timeout,
79 .data = &init_net.nf_frag.frags.low_thresh,
84 .extra2 = &init_net.nf_frag.frags.high_thresh
88 .data = &init_net.nf_frag.frags.high_thresh,
92 .extra1 = &init_net.nf_frag.frags.low_thresh
109 table[0].data = &net->nf_frag.frags.timeout; nf_ct_frag6_sysctl_register()
110 table[1].data = &net->nf_frag.frags.low_thresh; nf_ct_frag6_sysctl_register()
111 table[1].extra2 = &net->nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register()
112 table[2].data = &net->nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register()
113 table[2].extra1 = &net->nf_frag.frags.low_thresh; nf_ct_frag6_sysctl_register()
114 table[2].extra2 = &init_net.nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register()
185 net = container_of(fq->q.net, struct net, nf_frag.frags); nf_ct_frag6_expire()
209 q = inet_frag_find(&net->nf_frag.frags, &nf_frags, &arg, hash); fq_find()
426 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); nf_ct_frag6_reasm()
652 net->nf_frag.frags.high_thresh = IPV6_FRAG_HIGH_THRESH; nf_ct_net_init()
653 net->nf_frag.frags.low_thresh = IPV6_FRAG_LOW_THRESH; nf_ct_net_init()
654 net->nf_frag.frags.timeout = IPV6_FRAG_TIMEOUT; nf_ct_net_init()
655 inet_frags_init_net(&net->nf_frag.frags); nf_ct_net_init()
663 inet_frags_exit_net(&net->nf_frag.frags, &nf_frags); nf_ct_net_exit()
/linux-4.1.27/drivers/net/xen-netback/
H A Dnetback.c87 * This is sized to avoid pulling headers from the frags for most
431 skb_frag_page(&skb_shinfo(skb)->frags[i]), xenvif_gop_skb()
432 skb_frag_size(&skb_shinfo(skb)->frags[i]), xenvif_gop_skb()
433 skb_shinfo(skb)->frags[i].page_offset, xenvif_gop_skb()
821 skb_frag_t *frags = shinfo->frags; xenvif_get_requests() local
838 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); xenvif_get_requests()
845 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); xenvif_get_requests()
858 frags = shinfo->frags; xenvif_get_requests()
865 frag_set_pending_idx(&frags[shinfo->nr_frags], xenvif_get_requests()
919 frag_get_pending_idx(&shinfo->frags[0]) == pending_idx; xenvif_tx_check_gop()
942 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); xenvif_tx_check_gop()
979 /* Not the first error? Preceding frags already invalidated. */ xenvif_tx_check_gop()
993 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); xenvif_tx_check_gop()
1000 * the first skb's frags xenvif_tx_check_gop()
1004 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); xenvif_tx_check_gop()
1035 skb_frag_t *frag = shinfo->frags + i; xenvif_fill_frags()
1305 frag_set_pending_idx(&skb_shinfo(skb)->frags[0], xenvif_tx_build_gops()
1310 frag_set_pending_idx(&skb_shinfo(skb)->frags[0], xenvif_tx_build_gops()
1340 * frags. Returns 0 or -ENOMEM if can't allocate new pages.
1345 skb_frag_t frags[MAX_SKB_FRAGS]; xenvif_handle_frag_list() local
1354 /* Subtract frags size, we will correct it later */ xenvif_handle_frag_list()
1359 /* create a brand new frags array and coalesce there */ xenvif_handle_frag_list()
1370 put_page(frags[j].page.p); xenvif_handle_frag_list()
1382 frags[i].page.p = page; xenvif_handle_frag_list()
1383 frags[i].page_offset = 0; xenvif_handle_frag_list()
1384 skb_frag_size_set(&frags[i], len); xenvif_handle_frag_list()
1392 /* Release all the original (foreign) frags. */ xenvif_handle_frag_list()
1401 /* Fill the skb with the new (local) frags. */ xenvif_handle_frag_list()
1402 memcpy(skb_shinfo(skb)->frags, frags, i * sizeof(skb_frag_t)); xenvif_handle_frag_list()
1427 * expected to release all the frags which were mapped, xenvif_tx_submit()
H A Dcommon.h87 /* It's possible for an skb to have a maximal number of frags
/linux-4.1.27/drivers/net/wireless/ath/ath10k/
H A Dhtt_tx.c418 struct htt_data_tx_desc_frag *frags; ath10k_htt_tx() local
474 frags = skb_cb->htt.txbuf->frags; ath10k_htt_tx()
476 frags[0].paddr = __cpu_to_le32(skb_cb->paddr); ath10k_htt_tx()
477 frags[0].len = __cpu_to_le32(msdu->len); ath10k_htt_tx()
478 frags[1].paddr = 0; ath10k_htt_tx()
479 frags[1].len = 0; ath10k_htt_tx()
556 sizeof(skb_cb->htt.txbuf->frags); ath10k_htt_tx()
H A Ddebug.c547 "Extra frags on rings 0", pdev->r0_frags); ath10k_fw_stats_fill()
549 "Extra frags on rings 1", pdev->r1_frags); ath10k_fw_stats_fill()
551 "Extra frags on rings 2", pdev->r2_frags); ath10k_fw_stats_fill()
553 "Extra frags on rings 3", pdev->r3_frags); ath10k_fw_stats_fill()
H A Dhtt.h914 /* Extra frags on rings 0-3 */
1238 struct htt_data_tx_desc_frag frags[2]; member in struct:ath10k_htt_txbuf
H A Dpci.c212 /* NB: 50% of src nentries, since tx has 2 frags */
/linux-4.1.27/net/batman-adv/
H A Dfragmentation.h36 * Returns true if the frags entry has timed out, false otherwise.
/linux-4.1.27/fs/jffs2/
H A Dnodelist.c105 this->node->frags--; jffs2_obsolete_node_frag()
106 if (!this->node->frags) { jffs2_obsolete_node_frag()
107 /* The node has no valid frags left. It's totally obsoleted */ jffs2_obsolete_node_frag()
113 dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) REF_NORMAL. frags is %d\n", jffs2_obsolete_node_frag()
114 ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); jffs2_obsolete_node_frag()
185 because there are no frags with offset greater than it. no_overlapping_node()
201 because there are no frags with offset greater than it. no_overlapping_node()
286 this->node->frags++; jffs2_add_frag_to_fragtree()
374 newfrag->node->frags = 1; jffs2_add_full_dnode_to_inode()
571 if (frag->node && !(--frag->node->frags)) { rbtree_postorder_for_each_entry_safe()
H A Dgc.c520 if (nrfrags == frag->node->frags) jffs2_garbage_collect_live()
1008 if(fn->frags > 1) { jffs2_garbage_collect_hole()
1110 D1(if(unlikely(fn->frags <= 1)) { jffs2_garbage_collect_hole()
1112 __func__, fn->frags, je32_to_cpu(ri.version), jffs2_garbage_collect_hole()
1125 new_fn->frags++; jffs2_garbage_collect_hole()
1126 fn->frags--; jffs2_garbage_collect_hole()
1129 if (fn->frags) { jffs2_garbage_collect_hole()
1130 pr_warn("%s(): Old node still has frags!\n", __func__); jffs2_garbage_collect_hole()
1133 if (!new_fn->frags) { jffs2_garbage_collect_hole()
1134 pr_warn("%s(): New node has no frags!\n", __func__); jffs2_garbage_collect_hole()
1166 adjacent frags if the block containing them is already considered jffs2_garbage_collect_dnode()
H A Ddebug.c87 if (fn->frags > 1) { __jffs2_dbg_fragtree_paranoia_check_nolock()
88 JFFS2_ERROR("REF_PRISTINE node at 0x%08x had %d frags. Tell dwmw2.\n", __jffs2_dbg_fragtree_paranoia_check_nolock()
89 ref_offset(fn->raw), fn->frags); __jffs2_dbg_fragtree_paranoia_check_nolock()
H A Dread.c170 frags, we read it twice. Don't do that. */ jffs2_read_inode_range()
H A Ddir.c291 /* FIXME: If you care. We'd need to use frags for the target jffs2_symlink()
H A Dnodelist.h218 uint32_t frags; /* Number of fragments which currently refer member in struct:jffs2_full_dnode
H A Dwrite.c191 fn->frags = 0; jffs2_write_dnode()
H A Dreadinode.c211 * with frags -- we can have multiple entries starting at the same
/linux-4.1.27/drivers/net/ethernet/tile/
H A Dtilepro.c378 * 268 bytes with 4K pages (since the frags[] array needs 18 entries).
1597 * Prepare the "frags" info for the resulting LEPP command.
1599 * If needed, flush the memory used by the frags.
1601 static unsigned int tile_net_tx_frags(lepp_frag_t *frags, tile_net_tx_frags() argument
1617 frags[n].cpa_lo = cpa; tile_net_tx_frags()
1618 frags[n].cpa_hi = cpa >> 32; tile_net_tx_frags()
1619 frags[n].length = b_len; tile_net_tx_frags()
1620 frags[n].hash_for_home = hash_default; tile_net_tx_frags()
1626 skb_frag_t *f = &sh->frags[i]; tile_net_tx_frags()
1641 frags[n].cpa_lo = cpa; tile_net_tx_frags()
1642 frags[n].cpa_hi = cpa >> 32; tile_net_tx_frags()
1643 frags[n].length = skb_frag_size(f); tile_net_tx_frags()
1644 frags[n].hash_for_home = hash_for_home; tile_net_tx_frags()
1659 * and "sh->frags" will contain "skb->data_len" bytes of payload, and
1663 * "data" containing the header and payload, with "frags" being empty.
1717 /* Analyze the "frags". */ tile_net_tx_tso()
1719 tile_net_tx_frags(cmd->frags, skb, data + sh_len, b_len); tile_net_tx_tso()
1721 /* The size of the command, including frags and header. */ tile_net_tx_tso()
1761 memcpy(&cmd->frags[num_frags], data, sh_len); tile_net_tx_tso()
1861 lepp_frag_t frags[1 + MAX_SKB_FRAGS]; tile_net_tx() local
1906 /* ISSUE: Does not dump the "frags". */ tile_net_tx()
1917 num_frags = tile_net_tx_frags(frags, skb, data, skb_headlen(skb)); tile_net_tx()
1924 .cpa_lo = frags[i].cpa_lo, tile_net_tx()
1925 .cpa_hi = frags[i].cpa_hi, tile_net_tx()
1926 .length = frags[i].length, tile_net_tx()
1927 .hash_for_home = frags[i].hash_for_home, tile_net_tx()
2264 /* We support TSO iff the HV supports sufficient frags. */ tile_net_setup()
H A Dtilegx.c1671 * "data" containing the header and payload, with "frags" being empty.
1698 f_size = skb_frag_size(&sh->frags[f_id]); tso_count_edescs()
1802 f_size = skb_frag_size(&sh->frags[f_id]); tso_headers_prepare()
1879 f_size = skb_frag_size(&sh->frags[f_id]); tso_egress()
1880 f_data = tile_net_frag_buf(&sh->frags[f_id]); tso_egress()
1973 /* Analyze the body and frags for a transmit request. */ tile_net_tx_frags()
1974 static unsigned int tile_net_tx_frags(struct frag *frags, tile_net_tx_frags() argument
1983 frags[n].buf = b_data; tile_net_tx_frags()
1984 frags[n++].length = b_len; tile_net_tx_frags()
1988 skb_frag_t *f = &sh->frags[i]; tile_net_tx_frags()
1989 frags[n].buf = tile_net_frag_buf(f); tile_net_tx_frags()
1990 frags[n++].length = skb_frag_size(f); tile_net_tx_frags()
2011 struct frag frags[MAX_FRAGS]; tile_net_tx() local
2021 num_edescs = tile_net_tx_frags(frags, skb, data, skb_headlen(skb)); tile_net_tx()
2028 edesc.xfer_size = frags[i].length; tile_net_tx()
2029 edesc.va = va_to_tile_io_addr(frags[i].buf); tile_net_tx()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/
H A Dsocklnd.h215 int tx_niov; /* # packet iovec frags */
216 struct kvec *tx_iov; /* packet iovec frags */
217 int tx_nkiov; /* # packet page frags */
222 lnet_kiov_t *tx_kiov; /* packet page frags */
244 * header, or up to LNET_MAX_IOV frags of payload of either type. */
286 int ksnc_rx_niov; /* # iovec frags */
287 struct kvec *ksnc_rx_iov; /* the iovec frags */
288 int ksnc_rx_nkiov; /* # page frags */
289 lnet_kiov_t *ksnc_rx_kiov; /* the page frags */
H A Dsocklnd_lib-linux.c161 #warning "XXX risk of kmap deadlock on multiple frags..." ksocknal_lib_send_kiov()
316 #warning "XXX risk of kmap deadlock on multiple frags..." ksocknal_lib_recv_kiov()
H A Dsocklnd_cb.c705 /* Ensure the frags we've been given EXACTLY match the number of ksocknal_queue_tx_locked()
707 * size parameters passed to them and just look at the frags. ksocknal_queue_tx_locked()
940 CDEBUG(D_NET, "sending %u bytes in %d frags to %s\n", ksocknal_send()
/linux-4.1.27/sound/soc/txx9/
H A Dtxx9aclc.c95 dmadata->frags = 2; txx9aclc_pcm_prepare()
98 dmadata->frags = dmadata->buffer_bytes / dmadata->period_bytes; txx9aclc_pcm_prepare()
176 dmadata->frag_count = NR_DMA_CHAIN % dmadata->frags; txx9aclc_dma_tasklet()
196 dmadata->frag_count %= dmadata->frags; txx9aclc_dma_tasklet()
H A Dtxx9aclc.h58 int frags; member in struct:txx9aclc_dmadata
/linux-4.1.27/drivers/media/pci/ttpci/
H A Dav7110_av.c584 p->frags = 0; av7110_p2t_init()
594 p->frags = 0; clear_p2t()
598 static int find_pes_header(u8 const *buf, long int length, int *frags) find_pes_header() argument
603 *frags = 0; find_pes_header()
632 *frags = 1; find_pes_header()
635 *frags = 2; find_pes_header()
639 *frags = 3; find_pes_header()
653 if (p->frags){ av7110_p2t_write()
655 switch(p->frags) { av7110_p2t_write()
700 p->frags = 0; av7110_p2t_write()
704 c2 = find_pes_header(buf + c, length - c, &p->frags); av7110_p2t_write()
718 c2 = find_pes_header(buf + c + add, length - c - add, &p->frags); av7110_p2t_write()
H A Dav7110.h59 int frags; member in struct:av7110_p2t
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c273 end = &si->frags[si->nr_frags]; map_skb()
275 for (fp = si->frags; fp < end; fp++) { map_skb()
284 while (fp-- > si->frags) map_skb()
302 end = &si->frags[si->nr_frags]; unmap_skb()
303 for (fp = si->frags; fp < end; fp++) unmap_skb()
868 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); write_sgl()
884 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl()
885 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); write_sgl()
890 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl()
1699 __skb_fill_page_desc(skb, 0, gl->frags[0].page, copy_frags()
1700 gl->frags[0].offset + offset, copy_frags()
1701 gl->frags[0].size - offset); copy_frags()
1704 __skb_fill_page_desc(skb, i, gl->frags[i].page, copy_frags()
1705 gl->frags[i].offset, copy_frags()
1706 gl->frags[i].size); copy_frags()
1709 get_page(gl->frags[gl->nfrags - 1].page); copy_frags()
1765 for (p = gl->frags, n = gl->nfrags - 1; n--; p++) t4_pktgl_free()
1928 * @frags: number of FL buffers to restore
1940 int frags) restore_rx_bufs()
1944 while (frags--) { restore_rx_bufs()
1950 d->page = si->frags[frags].page; restore_rx_bufs()
2019 u32 len = ntohl(rc->pldbuflen_qid), bufsz, frags; process_responses() local
2031 for (frags = 0, fp = si.frags; ; frags++, fp++) { process_responses()
2051 si.va = page_address(si.frags[0].page) + process_responses()
2052 si.frags[0].offset; process_responses()
2055 si.nfrags = frags + 1; process_responses()
2060 restore_rx_bufs(&si, &rxq->fl, frags); process_responses()
1939 restore_rx_bufs(const struct pkt_gl *si, struct sge_fl *q, int frags) restore_rx_bufs() argument
H A Dcxgb4.h469 struct page_frag frags[MAX_SKB_FRAGS]; member in struct:pkt_gl
/linux-4.1.27/fs/ufs/
H A Dufs_fs.h261 #define ufs_fragstoblks(frags) ((frags) >> uspi->s_fpbshift)
307 __fs32 cs_nffree; /* number of free frags */
313 __fs64 cs_nffree; /* number of free frags */
322 __u64 cs_nffree; /* number of free frags */
366 __fs32 fs_frag; /* number of frags in a block in fs */
375 __fs32 fs_fshift; /* ``numfrags'' calc number of frags */
548 __fs32 cg_frsum[UFS_MAXFRAG]; /* counts of available frags */
590 __fs32 cg_frsum[8]; /* counts of available frags */
913 __fs64 cs_nffree; /* number of free frags */
H A Dutil.h354 * Determine the number of available frags given a
H A Dsuper.c192 pr_info(" cs_nffree(Num of free frags): %llu\n", ufs_print_super_stuff()
/linux-4.1.27/drivers/net/hyperv/
H A Dnetvsc_drv.c285 int frags = skb_shinfo(skb)->nr_frags; init_page_array() local
305 for (i = 0; i < frags; i++) { init_page_array()
306 skb_frag_t *frag = skb_shinfo(skb)->frags + i; init_page_array()
317 int i, frags = skb_shinfo(skb)->nr_frags; count_skb_frag_slots() local
320 for (i = 0; i < frags; i++) { count_skb_frag_slots()
321 skb_frag_t *frag = skb_shinfo(skb)->frags + i; count_skb_frag_slots()
/linux-4.1.27/drivers/net/wireless/ath/ath9k/
H A Dcommon-debug.h38 * @rx_too_many_frags_err: Frames dropped due to too-many-frags received.
/linux-4.1.27/net/sunrpc/auth_gss/
H A Dgss_krb5_crypto.c490 struct scatterlist frags[4]; member in struct:decryptor_desc
505 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, decryptor()
516 sg_mark_end(&desc->frags[desc->fragno - 1]); decryptor()
518 ret = crypto_blkcipher_decrypt_iv(&desc->desc, desc->frags, decryptor()
519 desc->frags, thislen); decryptor()
523 sg_init_table(desc->frags, 4); decryptor()
526 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, decryptor()
553 sg_init_table(desc.frags, 4); gss_decrypt_xdr_buf()
808 sg_init_table(desc.frags, 4); gss_krb5_aes_decrypt()
/linux-4.1.27/include/linux/
H A Dinet_lro.h97 * headers in generated skb when using frags */
H A Disdn_ppp.h108 struct sk_buff * frags; /* fragments sl list -- use skb->next */ member in struct:__anon11864
H A Dskbuff.h198 * Since GRO uses frags we allocate at least 16 regardless of page
283 * all frags to avoid possible bad checksum
338 skb_frag_t frags[MAX_SKB_FRAGS]; member in struct:skb_shared_info
1571 len += skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_pagelen()
1591 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_fill_page_desc()
2096 * skb_orphan_frags - orphan the frags contained in a buffer
2097 * @skb: buffer to orphan frags from
2285 __skb_frag_ref(&skb_shinfo(skb)->frags[f]); skb_frag_ref()
2308 __skb_frag_unref(&skb_shinfo(skb)->frags[f]); skb_frag_unref()
2362 __skb_frag_set_page(&skb_shinfo(skb)->frags[f], page); skb_frag_set_page()
2525 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; skb_can_coalesce()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c288 end = &si->frags[si->nr_frags]; map_skb()
289 for (fp = si->frags; fp < end; fp++) { map_skb()
298 while (fp-- > si->frags) map_skb()
924 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); write_sgl()
940 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl()
941 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); write_sgl()
946 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl()
1350 "eth_xmit: ndesc %u, credits %u, pidx %u, len %u, frags %u", t4vf_eth_xmit()
1463 __skb_fill_page_desc(skb, 0, gl->frags[0].page, copy_frags()
1464 gl->frags[0].offset + offset, copy_frags()
1465 gl->frags[0].size - offset); copy_frags()
1468 __skb_fill_page_desc(skb, i, gl->frags[i].page, copy_frags()
1469 gl->frags[i].offset, copy_frags()
1470 gl->frags[i].size); copy_frags()
1473 get_page(gl->frags[gl->nfrags - 1].page); copy_frags()
1539 put_page(gl->frags[frag].page); t4vf_pktgl_free()
1690 int frags) restore_rx_bufs()
1694 while (frags--) { restore_rx_bufs()
1700 sdesc->page = gl->frags[frags].page; restore_rx_bufs()
1785 for (frag = 0, fp = gl.frags; /**/; frag++, fp++) { process_responses()
1808 gl.va = (page_address(gl.frags[0].page) + process_responses()
1809 gl.frags[0].offset); process_responses()
1689 restore_rx_bufs(const struct pkt_gl *gl, struct sge_fl *fl, int frags) restore_rx_bufs() argument
H A Dadapter.h152 struct page_frag frags[MAX_SKB_FRAGS]; member in struct:pkt_gl
/linux-4.1.27/sound/pci/
H A Dintel8x0m.c156 #define ICH_MAX_FRAGS 32 /* max hw frags */
179 int frags; member in struct:ichdev
413 ichdev->frags = 2; snd_intel8x0m_setup_periods()
426 ichdev->frags = ichdev->size / ichdev->fragsize; snd_intel8x0m_setup_periods()
431 ichdev->lvi_frag = ICH_REG_LVI_MASK % ichdev->frags; snd_intel8x0m_setup_periods()
435 "lvi_frag = %i, frags = %i, period_size = 0x%x, period_size1 = 0x%x\n", snd_intel8x0m_setup_periods()
436 ichdev->lvi_frag, ichdev->frags, ichdev->fragsize, snd_intel8x0m_setup_periods()
475 ichdev->lvi_frag %= ichdev->frags; snd_intel8x0m_update()
H A Dintel8x0.c228 #define ICH_MAX_FRAGS 32 /* max hw frags */
368 int frags; member in struct:ichdev
699 ichdev->frags = 2; snd_intel8x0_setup_periods()
714 ichdev->frags = ichdev->size / ichdev->fragsize; snd_intel8x0_setup_periods()
719 ichdev->lvi_frag = ICH_REG_LVI_MASK % ichdev->frags; snd_intel8x0_setup_periods()
723 "lvi_frag = %i, frags = %i, period_size = 0x%x, period_size1 = 0x%x\n", snd_intel8x0_setup_periods()
724 ichdev->lvi_frag, ichdev->frags, ichdev->fragsize, snd_intel8x0_setup_periods()
788 ichdev->lvi_frag %= ichdev->frags; snd_intel8x0_update()
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
H A Dipoib_cm.c80 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, ipoib_cm_dma_unmap_rx() argument
87 for (i = 0; i < frags; ++i) ipoib_cm_dma_unmap_rx()
142 int id, int frags, ipoib_cm_alloc_rx_skb()
167 for (i = 0; i < frags; i++) { ipoib_cm_alloc_rx_skb()
539 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_put_frags()
567 int frags; ipoib_cm_handle_rx_wc() local
641 frags = PAGE_ALIGN(wc->byte_len - min(wc->byte_len, ipoib_cm_handle_rx_wc()
644 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, ipoib_cm_handle_rx_wc()
656 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); ipoib_cm_handle_rx_wc()
657 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof *mapping); ipoib_cm_handle_rx_wc()
140 ipoib_cm_alloc_rx_skb(struct net_device *dev, struct ipoib_cm_rx_buf *rx_ring, int id, int frags, u64 mapping[IPOIB_CM_RX_SG], gfp_t gfp) ipoib_cm_alloc_rx_skb() argument
H A Dipoib_ib.c285 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ipoib_dma_map_tx()
297 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; ipoib_dma_map_tx()
323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ipoib_dma_unmap_tx()
519 skb_frag_t *frags = skb_shinfo(skb)->frags; post_send() local
532 priv->tx_sge[i + off].length = skb_frag_size(&frags[i]); post_send()
/linux-4.1.27/block/partitions/
H A Dldm.c1288 * @frags: Linked list of VBLK fragments
1296 static bool ldm_frag_add (const u8 *data, int size, struct list_head *frags) ldm_frag_add() argument
1302 BUG_ON (!data || !frags); ldm_frag_add()
1321 list_for_each (item, frags) { list_for_each()
1338 list_add_tail (&f->list, frags);
1378 * @frags: Linked list of VBLK fragments
1387 static bool ldm_frag_commit (struct list_head *frags, struct ldmdb *ldb) ldm_frag_commit() argument
1392 BUG_ON (!frags || !ldb); ldm_frag_commit()
1394 list_for_each (item, frags) { list_for_each()
1428 LIST_HEAD (frags); ldm_get_vblks()
1455 if (!ldm_frag_add (data, size, &frags)) ldm_get_vblks()
1464 result = ldm_frag_commit (&frags, ldb); /* Failures, already logged */ ldm_get_vblks()
1468 ldm_frag_free (&frags); ldm_get_vblks()
/linux-4.1.27/drivers/net/ethernet/qlogic/
H A Dqla3xxx.c2261 * more than two frags are given. Each address list has 5 addr/len pairs.
2262 * The 5th pair in each OAL is used to point to the next OAL if more frags
2263 * are coming. That is why the frags:segment count ratio is not linear.
2265 static int ql_get_seg_count(struct ql3_adapter *qdev, unsigned short frags) ql_get_seg_count() argument
2270 if (frags <= 2) ql_get_seg_count()
2271 return frags + 1; ql_get_seg_count()
2272 else if (frags <= 6) ql_get_seg_count()
2273 return frags + 2; ql_get_seg_count()
2274 else if (frags <= 10) ql_get_seg_count()
2275 return frags + 3; ql_get_seg_count()
2276 else if (frags <= 14) ql_get_seg_count()
2277 return frags + 4; ql_get_seg_count()
2278 else if (frags <= 18) ql_get_seg_count()
2279 return frags + 5; ql_get_seg_count()
2350 skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs]; ql_send_map()
2391 "PCI mapping frags failed with error: %d\n", ql_send_map()
H A Dqla3xxx.h1014 * 1 thru 4 point to frags
/linux-4.1.27/drivers/scsi/cxgbi/
H A Dlibcxgbi.c1139 int frags = skb_shinfo(skb)->nr_frags + cxgbi_sock_send_pdus() local
1149 if (frags >= SKB_WR_LIST_SIZE) { cxgbi_sock_send_pdus()
1150 pr_err("csk 0x%p, frags %d, %u,%u >%u.\n", cxgbi_sock_send_pdus()
2039 unsigned int dlen, struct page_frag *frags, sgl_read_to_frags()
2064 if (i && page == frags[i - 1].page && sgl_read_to_frags()
2066 frags[i - 1].offset + frags[i - 1].size) { sgl_read_to_frags()
2067 frags[i - 1].size += copy; sgl_read_to_frags()
2075 frags[i].page = page; sgl_read_to_frags()
2076 frags[i].offset = sg->offset + sgoffset; sgl_read_to_frags()
2077 frags[i].size = copy; sgl_read_to_frags()
2186 tdata->frags, MAX_PDU_FRAGS); cxgbi_conn_init_pdu()
2197 struct page_frag *frag = tdata->frags; cxgbi_conn_init_pdu()
2216 tdata->frags[i].page, cxgbi_conn_init_pdu()
2217 tdata->frags[i].offset, cxgbi_conn_init_pdu()
2218 tdata->frags[i].size); cxgbi_conn_init_pdu()
2038 sgl_read_to_frags(struct scatterlist *sg, unsigned int sgoffset, unsigned int dlen, struct page_frag *frags, int frag_max) sgl_read_to_frags() argument
H A Dlibcxgbi.h594 struct page_frag frags[MAX_PDU_FRAGS]; member in struct:cxgbi_task_data
/linux-4.1.27/arch/blackfin/mach-bf609/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/sound/pci/lola/
H A Dlola_pcm.c328 if (str->frags >= LOLA_MAX_BDL_ENTRIES) setup_bdle()
344 str->frags++; setup_bdle()
367 str->frags = 0; lola_setup_periods()
454 lola_dsd_write(chip, str->dsd, LVI, str->frags - 1); lola_setup_controller()
H A Dlola.h302 unsigned int frags; member in struct:lola_stream
/linux-4.1.27/drivers/net/
H A Dxen-netfront.c311 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); xennet_alloc_rx_buffers()
477 int i, frags = skb_shinfo(skb)->nr_frags; xennet_count_skb_slots() local
482 for (i = 0; i < frags; i++) { xennet_count_skb_slots()
483 skb_frag_t *frag = skb_shinfo(skb)->frags + i; xennet_count_skb_slots()
605 /* Requests for all the frags. */ xennet_start_xmit()
607 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; xennet_start_xmit()
832 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; xennet_fill_frags()
970 skb_shinfo(skb)->frags[0].page_offset = rx->offset; xennet_poll()
971 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); xennet_poll()
1095 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); xennet_release_rx_bufs()
/linux-4.1.27/net/sctp/
H A Doutqueue.c1103 sctp_sack_variable_t *frags; sctp_sack_update_unack_data() local
1109 frags = sack->variable; sctp_sack_update_unack_data()
1111 unack_data -= ((ntohs(frags[i].gab.end) - sctp_sack_update_unack_data()
1112 ntohs(frags[i].gab.start) + 1)); sctp_sack_update_unack_data()
1130 sctp_sack_variable_t *frags = sack->variable; sctp_outq_sack() local
1186 highest_tsn += ntohs(frags[gap_ack_blocks - 1].gab.end);
1639 sctp_sack_variable_t *frags; sctp_acked() local
1658 frags = sack->variable; sctp_acked()
1661 if (TSN_lte(ntohs(frags[i].gab.start), gap) && sctp_acked()
1662 TSN_lte(gap, ntohs(frags[i].gab.end))) sctp_acked()
/linux-4.1.27/drivers/isdn/i4l/
H A Disdn_ppp.c1595 lp->netdev->pb->frags = NULL; isdn_ppp_mp_init()
1684 if ((frag = mp->frags) == NULL || MP_LT(newseq, MP_SEQ(frag))) { isdn_ppp_mp_receive()
1686 mp->frags = frag = newfrag; isdn_ppp_mp_receive()
1750 if (mp->frags == frag) isdn_ppp_mp_receive()
1751 mp->frags = nextf; isdn_ppp_mp_receive()
1770 mp->frags = nextf; isdn_ppp_mp_receive()
1810 mp->frags = isdn_ppp_mp_discard(mp, start, nextf); isdn_ppp_mp_receive()
1819 if (mp->frags == NULL) isdn_ppp_mp_receive()
1820 mp->frags = frag; isdn_ppp_mp_receive()
1827 frag = mp->frags->next; isdn_ppp_mp_receive()
1828 isdn_ppp_mp_free_skb(mp, mp->frags); isdn_ppp_mp_receive()
1829 mp->frags = frag; isdn_ppp_mp_receive()
1837 struct sk_buff *frag = lp->netdev->pb->frags; isdn_ppp_mp_cleanup()
1844 lp->netdev->pb->frags = NULL; isdn_ppp_mp_cleanup()
/linux-4.1.27/drivers/net/ethernet/aeroflex/
H A Dgreth.c115 skb_frag_address(&skb_shinfo(skb)->frags[i]), greth_print_tx_packet()
116 skb_shinfo(skb)->frags[i].size, true); greth_print_tx_packet()
201 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_clean_rings()
521 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_start_xmit_gbit()
718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_clean_tx_gbit()
/linux-4.1.27/arch/tile/include/hv/
H A Ddrv_xgbe_intf.h439 /** The length of the 'frags' array that follows this struct. */
442 /** The actual frags. */
443 lepp_frag_t frags[0 /* Variable-sized; num_frags entries. */]; member in struct:__anon2852::__anon2853::__anon2860
446 * The packet header template logically follows frags[],
459 /** Maximum number of frags for one TSO command. This is adapted from
/linux-4.1.27/drivers/net/ethernet/emulex/benet/
H A Dbe.h264 /* Struct to remember the pages posted for rx frags */
277 u32 rx_drops_no_frags; /* HW has no fetched frags */
313 bool rx_post_starved; /* Zero rx frags have been posted to BE */
H A Dbe_hw.h158 /* Number of rx frags posted */
H A Dbe_main.c932 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; be_xmit_enqueue()
1815 skb_shinfo(skb)->frags[0].page_offset = skb_fill_rx_data()
1817 skb_frag_size_set(&skb_shinfo(skb)->frags[0], skb_fill_rx_data()
1830 /* More frags present for this completion */ skb_fill_rx_data()
1836 /* Coalesce all frags from the same physical page in one slot */ skb_fill_rx_data()
1841 skb_shinfo(skb)->frags[j].page_offset = skb_fill_rx_data()
1843 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); skb_fill_rx_data()
1849 skb_frag_size_add(&skb_shinfo(skb)->frags[j], curr_frag_len); skb_fill_rx_data()
1918 /* Coalesce all frags from the same physical page in one slot */ be_rx_compl_process_gro()
1923 skb_shinfo(skb)->frags[j].page_offset = be_rx_compl_process_gro()
1925 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); be_rx_compl_process_gro()
1929 skb_frag_size_add(&skb_shinfo(skb)->frags[j], curr_frag_len); be_rx_compl_process_gro()
/linux-4.1.27/drivers/net/ethernet/neterion/vxge/
H A Dvxge-traffic.c1476 priv->frags = 0; vxge_hw_fifo_txdl_reserve()
1519 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; vxge_hw_fifo_txdl_buffer_set()
1529 if (txdl_priv->frags) { vxge_hw_fifo_txdl_buffer_set()
1531 (txdl_priv->frags - 1); vxge_hw_fifo_txdl_buffer_set()
1542 txdl_priv->frags++; vxge_hw_fifo_txdl_buffer_set()
1549 * @frags: Number of contiguous buffers that are part of a single
1569 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); vxge_hw_fifo_txdl_post()
1578 txdl_priv->frags - 1, vxge_hw_fifo_txdl_post()
H A Dvxge-main.c566 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit_compl()
923 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit()
978 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit()
1048 frag = &skb_shinfo(skb)->frags[0]; vxge_tx_term()
4017 "%s: Max frags : %d", vdev->ndev->name, vxge_print_parm()
H A Dvxge-config.h1044 * @frags: Total number of fragments (that is, contiguous data buffers)
1081 u32 frags; member in struct:__vxge_hw_fifo_txdl_priv
/linux-4.1.27/drivers/message/fusion/
H A Dmptctl.c128 static MptSge_t *kbuf_alloc_2_sgl(int bytes, u32 dir, int sge_offset, int *frags,
157 // Define max sg LIST bytes ( == (#frags + #chains) * 8 bytes each)
1018 * Outputs: frags - number of scatter gather elements
1025 kbuf_alloc_2_sgl(int bytes, u32 sgdir, int sge_offset, int *frags, kbuf_alloc_2_sgl() argument
1042 *frags = 0; kbuf_alloc_2_sgl()
1088 printk(MYIOC_s_WARN_FMT "-SG: (freeing %d frags)\n", kbuf_alloc_2_sgl()
1115 printk(MYIOC_s_WARN_FMT "(freeing %d frags)\n", ioc->name, numfrags); kbuf_alloc_2_sgl()
1123 "too many SG frags! :-(\n", ioc->name); kbuf_alloc_2_sgl()
1124 printk(MYIOC_s_WARN_FMT "-SG: (freeing %d frags)\n", kbuf_alloc_2_sgl()
1133 *frags = numfrags; kbuf_alloc_2_sgl()
1137 "%d SG frags generated!\n", ioc->name, numfrags)); kbuf_alloc_2_sgl()
/linux-4.1.27/drivers/net/ethernet/sfc/
H A Dtx.c257 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; efx_skb_copy_bits_to_pio()
413 fragment = &skb_shinfo(skb)->frags[i]; efx_enqueue_skb()
1282 skb_shinfo(skb)->frags + frag_i); efx_enqueue_skb_tso()
1302 skb_shinfo(skb)->frags + frag_i); efx_enqueue_skb_tso()
/linux-4.1.27/drivers/staging/unisys/common-spar/include/channels/
H A Diochannel.h139 * NOTE 2: In Linux, frags array in skb is currently allocated to be
463 int num_frags; /* number of fragments in frags containing data */
464 struct phys_info frags[MAX_PHYS_INFO]; /* physical page information for member in struct:net_pkt_xmt
482 * The full packet is described in frags but the ethernet header is
/linux-4.1.27/include/scsi/
H A Dfc_frame.h59 /* Max number of skb frags allowed, reserving one for fcoe_crc_eof page */
/linux-4.1.27/arch/blackfin/mach-bf518/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/arch/blackfin/mach-bf538/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/net/rds/
H A Dib_recv.c198 /* Recycle inc after freeing attached frags */ rds_ib_inc_free()
208 /* Free attached frags */ rds_ib_inc_free()
408 * We want to recycle several types of recv allocations, like incs and frags.
854 * Usually the frags make their way on to incs and are then freed as rds_ib_process_recv()
H A Drecv.c190 * a message have been dropped. We assume that frags that arrive rds_recv_incoming()
193 * from the sending conn until all their frags are sent. rds_recv_incoming()
H A Diw_recv.c67 * hold page references in their frags.
685 * Usually the frags make their way on to incs and are then freed as rds_iw_process_recv()
/linux-4.1.27/net/xfrm/
H A Dxfrm_ipcomp.c79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; ipcomp_decompress()
/linux-4.1.27/sound/soc/au1x/
H A Ddbdma2.c215 DBG("bits %d frags %d frag_bytes %d is_rx %d\n", params->msbits, au1xpsc_pcm_hw_params()
/linux-4.1.27/drivers/net/ethernet/marvell/
H A Dmvneta.c1699 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; mvneta_tx_frag_process()
1753 int frags = 0; mvneta_tx() local
1760 frags = mvneta_tx_tso(skb, dev, txq); mvneta_tx()
1764 frags = skb_shinfo(skb)->nr_frags + 1; mvneta_tx()
1779 frags = 0; mvneta_tx()
1783 if (frags == 1) { mvneta_tx()
1802 frags = 0; mvneta_tx()
1808 if (frags > 0) { mvneta_tx()
1812 txq->count += frags; mvneta_tx()
1813 mvneta_txq_pend_desc_add(pp, txq, frags); mvneta_tx()
H A Dmvpp2.c5126 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; mvpp2_tx_frag_process()
5177 int frags = 0; mvpp2_tx() local
5186 frags = skb_shinfo(skb)->nr_frags + 1; mvpp2_tx()
5189 if (mvpp2_aggr_desc_num_check(port->priv, aggr_txq, frags) || mvpp2_tx()
5191 txq_pcpu, frags)) { mvpp2_tx()
5192 frags = 0; mvpp2_tx()
5205 frags = 0; mvpp2_tx()
5213 if (frags == 1) { mvpp2_tx()
5227 frags = 0; mvpp2_tx()
5232 txq_pcpu->reserved_num -= frags; mvpp2_tx()
5233 txq_pcpu->count += frags; mvpp2_tx()
5234 aggr_txq->count += frags; mvpp2_tx()
5238 mvpp2_aggr_txq_pend_desc_add(port, frags); mvpp2_tx()
5246 if (frags > 0) { mvpp2_tx()
H A Dsky2.c1233 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; sky2_rx_map_skb()
1247 skb_frag_size(&skb_shinfo(skb)->frags[i]), sky2_rx_map_skb()
1271 skb_frag_size(&skb_shinfo(skb)->frags[i]), sky2_rx_unmap_skb()
1950 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; sky2_xmit_frame()
2527 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_put_frags()
H A Dmv643xx_eth.c676 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; has_tiny_unaligned_frags()
889 this_frag = &skb_shinfo(skb)->frags[frag]; txq_submit_frag_skb()
/linux-4.1.27/drivers/net/ethernet/tundra/
H A Dtsi108_eth.c655 int frags = skb_shinfo(skb)->nr_frags + 1; tsi108_send_packet() local
677 if (data->txfree - frags < MAX_SKB_FRAGS + 1) { tsi108_send_packet()
683 for (i = 0; i < frags; i++) { tsi108_send_packet()
709 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; tsi108_send_packet()
718 if (i == frags - 1) tsi108_send_packet()
/linux-4.1.27/drivers/scsi/cxgbi/cxgb3i/
H A Dcxgb3i.c408 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); push_tx_frames() local
409 int wrs_needed = skb_wrs[frags]; push_tx_frames()
414 WARN_ON(frags >= SKB_WR_LIST_SIZE || wrs_needed < 1); push_tx_frames()
419 csk, skb->len, skb->data_len, frags, push_tx_frames()
434 csk, skb->len, skb->data_len, frags, skb->csum, push_tx_frames()
/linux-4.1.27/sound/pci/hda/
H A Dhda_controller.c144 azx_sd_writew(chip, azx_dev, SD_LVI, azx_dev->frags - 1); azx_setup_controller()
301 if (azx_dev->frags >= AZX_MAX_BDL_ENTRIES) setup_bdle()
323 azx_dev->frags++; setup_bdle()
351 azx_dev->frags = 0; azx_setup_periods()
1407 azx_dev->frags = 0; azx_load_dsp_prepare()
H A Dhda_controller.h195 unsigned int frags; /* number for period in the play buffer */ member in struct:azx_dev
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/
H A Dlib-move.c183 /* skip complete frags before 'doffset' */ lnet_copy_iov2iov()
192 /* skip complete frags before 'soffset' */ lnet_copy_iov2iov()
243 return 0; /* no frags */ lnet_extract_iov()
246 while (offset >= src->iov_len) { /* skip initial frags */ lnet_extract_iov()
525 return 0; /* no frags */ lnet_extract_kiov()
528 while (offset >= src->kiov_len) { /* skip initial frags */ lnet_extract_kiov()
/linux-4.1.27/arch/blackfin/mach-bf537/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/drivers/net/ethernet/freescale/
H A Dgianfar.c2343 frag_len = skb_shinfo(skb)->frags[i].size; gfar_start_xmit()
2353 &skb_shinfo(skb)->frags[i], gfar_start_xmit()
2612 int frags = 0, nr_txbds = 0; gfar_clean_tx_ring() local
2627 frags = skb_shinfo(skb)->nr_frags; gfar_clean_tx_ring()
2633 nr_txbds = frags + 2; gfar_clean_tx_ring()
2635 nr_txbds = frags + 1; gfar_clean_tx_ring()
2671 for (i = 0; i < frags; i++) { gfar_clean_tx_ring()
H A Dfec_main.c383 this_frag = &skb_shinfo(skb)->frags[frag]; fec_enet_txq_submit_frag_skb()
390 frag_len = skb_shinfo(skb)->frags[frag].size; fec_enet_txq_submit_frag_skb()
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_main.c320 /* allocate a skb to store the frags */ fm10k_fetch_rx_buffer()
533 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; fm10k_pull_tail()
1008 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { fm10k_tx_map()
1120 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); fm10k_xmit_frame_ring()
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
H A Di40e_txrx.c1643 /* try the simple math, if we have too many frags per segment */ i40e_chk_linearize()
1649 frag = &skb_shinfo(skb)->frags[0]; i40e_chk_linearize()
1720 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { i40e_tx_map()
1892 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); i40e_xmit_descriptor_count()
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_io.c599 frag = &skb_shinfo(skb)->frags[i]; qlcnic_map_tx_skb()
678 /* 14 frags supported for normal packet and qlcnic_xmit_frame()
679 * 32 frags supported for TSO packet qlcnic_xmit_frame()
683 delta += skb_frag_size(&skb_shinfo(skb)->frags[i]); qlcnic_xmit_frame()
/linux-4.1.27/drivers/net/vmxnet3/
H A Dvmxnet3_ethtool.c59 { " too many frags", offsetof(struct vmxnet3_tq_driver_stats,
H A Dvmxnet3_drv.c644 struct skb_frag_struct *frag = skb_shinfo(skb)->frags + vmxnet3_append_frag()
740 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; vmxnet3_map_pkt()
914 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; txd_estimate()
1281 * processing all the following non-sop frags. vmxnet3_rq_rx_complete()
/linux-4.1.27/arch/blackfin/mach-bf527/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/arch/blackfin/mach-bf548/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/
H A Dqlge_main.c1280 * element and there is more than 6 frags, ql_unmap_send()
1347 * When this happens, the remainder of the frags will be stored ql_map_send()
1351 skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_idx]; ql_map_send()
1407 "PCI mapping frags failed with error: %d.\n", ql_map_send()
1429 * we pass in the number of frags that mapped successfully ql_map_send()
1896 * frags: There are 18 max frags and our small ql_build_rx_skb()
2081 * skb, tack it on frags, and send it up. ql_process_mac_rx_intr()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
H A Do2iblnd.h423 kib_rdma_frag_t rd_frags[0]; /* buffer frags */
506 #define IBLND_REJECT_RDMA_FRAGS 6 /* Fatal: peer's rdma frags can't match mine */
H A Do2iblnd_cb.c1107 CERROR("Src buffer exhausted: %d frags\n", srcidx); kiblnd_init_rdma()
1113 CERROR("Dst buffer exhausted: %d frags\n", dstidx); kiblnd_init_rdma()
1119 CERROR("RDMA too fragmented for %s (%d): %d/%d src %d/%d dst frags\n", kiblnd_init_rdma()
1476 CDEBUG(D_NET, "sending %d bytes in %d frags to %s\n", kiblnd_send()
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/
H A Dbnx2x_cmn.h1241 /* gro frags per page */ bnx2x_mtu_allows_gro()
1245 * 1. Number of frags should not grow above MAX_SKB_FRAGS bnx2x_mtu_allows_gro()
H A Dbnx2x_cmn.c250 /* now free frags */ bnx2x_free_tx_pkt()
3414 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]); bnx2x_pkt_req_lin()
3430 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx + wnd_size - 1]); bnx2x_pkt_req_lin()
3437 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx]); bnx2x_pkt_req_lin()
3892 nbd = 2; /* start_bd + pbd + frags (updated when pages are mapped) */ bnx2x_start_xmit()
4042 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; bnx2x_start_xmit()
H A Dbnx2x.h764 * Frag BDs - describes pages for frags
/linux-4.1.27/include/net/
H A Dipv6.h323 return sum_frag_mem_limit(&net->ipv6.frags); ip6_frag_mem()
535 struct inet_frags *frags);
/linux-4.1.27/drivers/atm/
H A Deni.c1136 skb_frag_page(&skb_shinfo(skb)->frags[i]) + do_tx()
1137 skb_shinfo(skb)->frags[i].page_offset, do_tx()
1138 skb_frag_size(&skb_shinfo(skb)->frags[i])); do_tx()
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
H A Di40e_txrx.c2434 /* try the simple math, if we have too many frags per segment */ i40e_chk_linearize()
2440 frag = &skb_shinfo(skb)->frags[0]; i40e_chk_linearize()
2517 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { i40e_tx_map()
2660 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); i40e_xmit_descriptor_count()
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/
H A Dnetxen_nic_main.c2018 frag = &skb_shinfo(skb)->frags[i]; netxen_map_tx_skb()
2073 /* 14 frags supported for normal packet and netxen_nic_xmit_frame()
2074 * 32 frags supported for TSO packet netxen_nic_xmit_frame()
2079 frag = &skb_shinfo(skb)->frags[i]; netxen_nic_xmit_frame()
/linux-4.1.27/drivers/staging/octeon/
H A Dethernet-tx.c300 struct skb_frag_struct *fs = skb_shinfo(skb)->frags + i; cvm_oct_xmit()
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-desc.c568 frag = &skb_shinfo(skb)->frags[i]; xgbe_map_tx_skb()
H A Dxgbe.h147 * - Maximum number of SKB frags
H A Dxgbe-drv.c1352 frag = &skb_shinfo(skb)->frags[i]; xgbe_packet_info()
/linux-4.1.27/arch/blackfin/mach-bf533/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/arch/blackfin/mach-bf561/include/mach/
H A Danomaly.h4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
/linux-4.1.27/drivers/net/ethernet/ibm/
H A Dibmveth.c1024 /* Map the frags */ ibmveth_start_xmit()
1026 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ibmveth_start_xmit()
/linux-4.1.27/drivers/net/ethernet/pasemi/
H A Dpasemi_mac.c302 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; pasemi_mac_unmap_tx_skb()
1503 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; pasemi_mac_start_tx()
/linux-4.1.27/drivers/net/ethernet/broadcom/
H A Dbnx2.c2883 skb_frag_size(&skb_shinfo(skb)->frags[i]), bnx2_tx_int()
2935 * last one in the frags array, so we need to recycle that page bnx2_reuse_rx_skb_pages()
2944 page = skb_frag_page(&shinfo->frags[shinfo->nr_frags]); bnx2_reuse_rx_skb_pages()
2945 __skb_frag_set_page(&shinfo->frags[shinfo->nr_frags], NULL); bnx2_reuse_rx_skb_pages()
3069 &skb_shinfo(skb)->frags[i - 1]; bnx2_rx_skb()
5436 skb_frag_size(&skb_shinfo(skb)->frags[k]), bnx2_free_tx_skbs()
6659 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; bnx2_start_xmit()
6728 skb_frag_size(&skb_shinfo(skb)->frags[i]), bnx2_start_xmit()
H A Dbgmac.c183 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; bgmac_dma_tx_add()
/linux-4.1.27/sound/oss/dmasound/
H A Ddmasound_core.c1024 partial frags can be queued for output. sq_ioctl()
1122 frags values using SNDCTL_DSP_GETBLKSIZE or similar */ sq_ioctl()
/linux-4.1.27/drivers/net/ethernet/rocker/
H A Drocker.c3943 struct rocker_tlv *frags; rocker_port_xmit() local
3956 frags = rocker_tlv_nest_start(desc_info, ROCKER_TLV_TX_FRAGS); rocker_port_xmit()
3957 if (!frags) rocker_port_xmit()
3967 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; rocker_port_xmit()
3975 rocker_tlv_nest_end(desc_info, frags); rocker_port_xmit()
3989 rocker_tlv_nest_cancel(desc_info, frags); rocker_port_xmit()
/linux-4.1.27/drivers/net/ethernet/agere/
H A Det131x.c2428 struct skb_frag_struct *frags = &skb_shinfo(skb)->frags[0]; nic_send_packet() local
2490 desc[frag].len_vlan = frags[i - 1].size; nic_send_packet()
2492 &frags[i - 1], nic_send_packet()
2494 frags[i - 1].size, nic_send_packet()
/linux-4.1.27/drivers/net/ethernet/sun/
H A Dcassini.c39 * -- driver appends the rest of the data pages as frags to skbuffs
135 #define RX_COPY_ALWAYS 0 /* if 0, use frags */
2005 skb_frag_t *frag = skb_shinfo(skb)->frags; cas_rx_process_pkt()
2007 /* normal or jumbo packets. we use frags */ cas_rx_process_pkt()
2824 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; cas_xmit_tx_ringN()
H A Dsunvnet.c1021 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; vnet_skb_map()
1062 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; vnet_skb_shape()
H A Dsungem.c1065 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; gem_start_xmit()
H A Dsunhme.c2344 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; happy_meal_start_xmit()
/linux-4.1.27/drivers/net/ethernet/tehuti/
H A Dtehuti.c1520 frag = &skb_shinfo(skb)->frags[i]; bdx_tx_map_skb()
1539 /* init_txd_sizes - precalculate sizes of descriptors for skbs up to 16 frags
1540 * number of frags is used as index to fetch correct descriptors size,
/linux-4.1.27/drivers/net/ethernet/cadence/
H A Dmacb.c836 netdev_vdbg(bp->dev, "macb_rx_frame frags %u - %u (len %u)\n", macb_rx_frame()
1155 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; macb_tx_map()
1267 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); macb_start_xmit()
/linux-4.1.27/drivers/net/ethernet/realtek/
H A Dr8169.c6816 const skb_frag_t *frag = info->frags + cur_frag; rtl8169_xmit_frags()
7049 int frags; rtl8169_start_xmit() local
7078 frags = rtl8169_xmit_frags(tp, skb, opts); rtl8169_start_xmit()
7079 if (frags < 0) rtl8169_start_xmit()
7081 else if (frags) rtl8169_start_xmit()
7102 tp->cur_tx += frags + 1; rtl8169_start_xmit()
H A D8139cp.c725 this_frag = &skb_shinfo(skb)->frags[frag]; unwind_tx_frag_mapping()
811 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; cp_start_xmit()
/linux-4.1.27/drivers/net/ethernet/adaptec/
H A Dstarfire.c1230 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[i - 1]; start_tx()
1352 skb_frag_size(&skb_shinfo(skb)->frags[i]), intr_handler()
/linux-4.1.27/drivers/net/ethernet/cisco/enic/
H A Denic_main.c363 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { enic_queue_wq_skb_cont()
489 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { enic_queue_wq_skb_tso()
/linux-4.1.27/drivers/net/ethernet/ti/
H A Dnetcp_core.c1038 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; netcp_tx_map_skb()
1054 dev_err(netcp->ndev_dev, "out of TX desc for frags\n"); netcp_tx_map_skb()
/linux-4.1.27/drivers/net/ethernet/atheros/atl1e/
H A Datl1e_main.c1618 fg_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); atl1e_cal_tdp_req()
1794 frag = &skb_shinfo(skb)->frags[f]; atl1e_tx_map()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/
H A Dsge.c1124 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; compute_large_page_tx_descs()
1256 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; write_tx_descs()
/linux-4.1.27/net/appletalk/
H A Dddp.c949 /* checksum stuff in frags */ atalk_sum_skb()
952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; atalk_sum_skb()
/linux-4.1.27/drivers/infiniband/hw/amso1100/
H A Dc2.c801 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; c2_xmit_frame()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
H A Dsge.c257 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]), unmap_skb()
980 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; make_sgl()
1546 pci_unmap_page(dui->pdev, *p++, skb_frag_size(&si->frags[i]), deferred_unmap_destructor()
2100 rx_frag = skb_shinfo(skb)->frags; lro_add_page()
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/
H A Dfs_enet-main.c569 frag = skb_shinfo(skb)->frags; fs_enet_start_xmit()
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/
H A Dixgbevf_main.c666 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbevf_pull_tail()
865 /* allocate a skb to store the frags */ ixgbevf_fetch_rx_buffer()
3453 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbevf_tx_map()
3629 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); ixgbevf_xmit_frame()
/linux-4.1.27/drivers/net/ethernet/xilinx/
H A Dll_temac_main.c686 frag = &skb_shinfo(skb)->frags[0]; temac_start_xmit()
H A Dxilinx_axienet_main.c694 frag = &skb_shinfo(skb)->frags[ii]; axienet_start_xmit()
/linux-4.1.27/drivers/scsi/be2iscsi/
H A Dbe_cmds.c1372 "BC_%d : FW CMD to map iscsi frags failed.\n"); be_cmd_iscsi_post_sgl_pages()
/linux-4.1.27/drivers/scsi/fcoe/
H A Dfcoe_transport.c309 frag = &skb_shinfo(skb)->frags[i]; fcoe_fc_crc()
H A Dfcoe.c1645 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; fcoe_xmit()
/linux-4.1.27/drivers/staging/i2o/
H A Di2o_config.c300 // it fails if you try and send frags out of order i2o_cfg_swdl()
/linux-4.1.27/drivers/staging/lustre/include/linux/lnet/
H A Dlib-types.h272 unsigned int md_niov; /* # frags */
/linux-4.1.27/drivers/s390/net/
H A Dqeth_core_main.c3870 frag = &skb_shinfo(skb)->frags[cnt]; qeth_get_elements_for_frags()
3976 frag = &skb_shinfo(skb)->frags[cnt]; __qeth_fill_buffer()
5719 {"tx sg frags"},
5721 {"rx sg frags"},
/linux-4.1.27/drivers/net/ethernet/atheros/alx/
H A Dmain.c1029 frag = &skb_shinfo(skb)->frags[f]; alx_map_tx_skb()
/linux-4.1.27/drivers/hsi/clients/
H A Dssi_protocol.c193 frag = &skb_shinfo(skb)->frags[i]; ssip_skb_to_msg()
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
H A Dehea_main.c1687 frag = &skb_shinfo(skb)->frags[0]; write_swqe2_data()
1700 frag = &skb_shinfo(skb)->frags[i]; write_swqe2_data()
/linux-4.1.27/drivers/net/ethernet/packetengines/
H A Dhamachi.c181 #define IP_MF 0x2000 /* IP more frags from <netinet/ip.h> */
1525 /* don't worry about frags */ hamachi_rx()
/linux-4.1.27/drivers/net/ethernet/atheros/atlx/
H A Datl1.c2282 frag = &skb_shinfo(skb)->frags[f]; atl1_tx_map()
2383 unsigned int f_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); atl1_xmit_frame()
/linux-4.1.27/drivers/net/ethernet/nvidia/
H A Dforcedeth.c2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); nv_start_xmit()
2256 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; nv_start_xmit()
2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); nv_start_xmit_optimized()
2406 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; nv_start_xmit_optimized()
/linux-4.1.27/fs/ceph/
H A Ddir.c441 /* more frags? */ ceph_readdir()
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/
H A Dmyri10ge.c1410 frag = skb_shinfo(skb)->frags; myri10ge_vlan_rx()
1464 rx_frags = skb_shinfo(skb)->frags; myri10ge_rx_done()
3024 frag = &skb_shinfo(skb)->frags[frag_idx]; myri10ge_xmit()
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_main.c1695 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbe_pull_tail()
1741 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbe_dma_sync_frag()
1932 /* allocate a skb to store the frags */ ixgbe_fetch_rx_buffer()
7038 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { ixgbe_tx_map()
7301 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); ixgbe_xmit_frame_ring()
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/
H A Dnetdev.c5352 frag = &skb_shinfo(skb)->frags[f]; e1000_tx_map()
5581 * frags into skb->data e1000_xmit_frame()
5609 count += DIV_ROUND_UP(skb_frag_size(&skb_shinfo(skb)->frags[f]), e1000_xmit_frame()
/linux-4.1.27/drivers/net/ethernet/intel/igb/
H A Digb_main.c4891 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { igb_tx_map()
5018 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); igb_xmit_frame_ring()
6708 /* allocate a skb to store the frags */ igb_fetch_rx_buffer()
6840 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; igb_pull_tail()
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/
H A Dixgb_main.c1369 frag = &skb_shinfo(skb)->frags[f]; ixgb_tx_map()
/linux-4.1.27/drivers/net/ethernet/natsemi/
H A Dns83820.c1112 frag = skb_shinfo(skb)->frags; ns83820_hard_start_xmit()
/linux-4.1.27/drivers/net/usb/
H A Dusbnet.c1275 struct skb_frag_struct *f = &skb_shinfo(skb)->frags[i]; build_dma_sg()
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_main.c1369 const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num]; sxgbe_xmit()
/linux-4.1.27/drivers/net/fddi/skfp/
H A Dhwmtm.c1832 * determine the frag count and the virt addresses of the frags smt_send_mbuf()
/linux-4.1.27/drivers/scsi/bnx2fc/
H A Dbnx2fc_fcoe.c329 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; bnx2fc_xmit()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/
H A Dd11.h1602 S_SEQ_NUM_FRAG, /* seq num for frags (at the start of MSDU) */
/linux-4.1.27/drivers/net/ethernet/calxeda/
H A Dxgmac.c1115 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; xgmac_xmit()
/linux-4.1.27/net/netfilter/ipvs/
H A Dip_vs_core.c426 * IPv6 frags, only the first hit here. ip_vs_schedule()
/linux-4.1.27/drivers/net/ethernet/intel/e1000/
H A De1000_main.c2908 frag = &skb_shinfo(skb)->frags[f]; e1000_tx_map()
3209 count += TXD_USE_COUNT(skb_frag_size(&skb_shinfo(skb)->frags[f]), e1000_xmit_frame()
/linux-4.1.27/drivers/net/ethernet/3com/
H A Dtyphoon.c814 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; typhoon_start_tx()
/linux-4.1.27/drivers/net/ethernet/alteon/
H A Dacenic.c2463 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ace_start_xmit()
/linux-4.1.27/drivers/net/ethernet/hp/
H A Dhp100.c1004 * Each pdl is 508 bytes long. (63 frags * 4 bytes for address and hp100_mmuinit()
/linux-4.1.27/drivers/net/ethernet/ibm/emac/
H A Dcore.c1484 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; emac_start_xmit_sg()
/linux-4.1.27/drivers/net/ethernet/intel/igbvf/
H A Dnetdev.c2110 frag = &skb_shinfo(skb)->frags[f]; igbvf_tx_map_adv()
/linux-4.1.27/drivers/net/ethernet/
H A Djme.c2064 frag = &skb_shinfo(skb)->frags[i]; jme_map_tx_skb()
/linux-4.1.27/drivers/net/ethernet/stmicro/stmmac/
H A Dstmmac_main.c1995 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; stmmac_xmit()
/linux-4.1.27/drivers/net/ethernet/via/
H A Dvia-velocity.c2599 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; velocity_xmit()
/linux-4.1.27/drivers/net/ethernet/atheros/atl1c/
H A Datl1c_main.c2155 frag = &skb_shinfo(skb)->frags[f]; atl1c_tx_map()
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c1342 &skb_shinfo(skb)->frags[i], bcmgenet_xmit()

Completed in 8172 milliseconds

12