/linux-4.1.27/include/net/netns/ |
H A D | ieee802154_6lowpan.h | 18 struct netns_frags frags; member in struct:netns_ieee802154_lowpan
|
H A D | ipv6.h | 44 struct netns_frags frags; member in struct:netns_ipv6 88 struct netns_frags frags; member in struct:netns_nf_frag
|
H A D | ipv4.h | 58 struct netns_frags frags; member in struct:netns_ipv4
|
/linux-4.1.27/include/linux/ceph/ |
H A D | ceph_frag.h | 7 * (subset of the number space) can be partitioned into 2^n sub-frags. 18 * can't sort encoded frags numerically. However, it does allow you 19 * to feed encoded frags as values into frag_contains_value. 104 * comparator to sort frags logically, as when traversing the
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
H A D | nes_mgt.c | 158 skb = fpdu_info->frags[i].skb; nes_download_callback() 159 if (fpdu_info->frags[i].cmplt) { nes_download_callback() 249 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; get_fpdu_info() local 276 memset(&frags, 0, sizeof frags); get_fpdu_info() 278 frags[i].physaddr = cb->busaddr; get_fpdu_info() 279 frags[i].physaddr += skb->data - cb->data_start; get_fpdu_info() 280 frags[i].frag_len = min(tmp_len, skb->len); get_fpdu_info() 281 frags[i].skb = skb; get_fpdu_info() 282 frags[i].cmplt = (skb->len == frags[i].frag_len); get_fpdu_info() 283 frag_tot += frags[i].frag_len; get_fpdu_info() 286 tmp_len -= frags[i].frag_len; get_fpdu_info() 297 skb_unlink(frags[i].skb, &nesqp->pau_list); get_fpdu_info() 298 nes_mgt_free_skb(nesdev, frags[i].skb, PCI_DMA_TODEVICE); get_fpdu_info() 301 frags[0].physaddr = cb->busaddr; get_fpdu_info() 302 frags[0].physaddr += skb->data - cb->data_start; get_fpdu_info() 303 frags[0].frag_len = skb->len; get_fpdu_info() 304 frags[0].skb = skb; get_fpdu_info() 305 frags[0].cmplt = true; get_fpdu_info() 314 frags[0].physaddr = cb->busaddr; get_fpdu_info() 315 frags[0].frag_len = 0; get_fpdu_info() 316 frags[0].skb = skb; get_fpdu_info() 317 frags[0].cmplt = true; get_fpdu_info() 336 cb = (struct nes_rskb_cb *)&frags[0].skb->cb[0]; get_fpdu_info() 343 if (frags[0].cmplt) { get_fpdu_info() 370 memcpy(fpdu_info->frags, frags, sizeof(fpdu_info->frags)); get_fpdu_info() 377 cb = (struct nes_rskb_cb *)&frags[i].skb->cb[0]; get_fpdu_info() 378 skb_pull(frags[i].skb, frags[i].frag_len); get_fpdu_info() 380 if (frags[i].skb->len == 0) { get_fpdu_info() 383 skb_unlink(frags[i].skb, &nesqp->pau_list); get_fpdu_info() 437 u32tmp = (fpdu_info->frags[1].frag_len << 16) | fpdu_info->frags[0].frag_len; forward_fpdus() 441 u32tmp = (fpdu_info->frags[3].frag_len << 16) | fpdu_info->frags[2].frag_len; forward_fpdus() 452 lower_32_bits(fpdu_info->frags[0].physaddr)); forward_fpdus() 454 upper_32_bits(fpdu_info->frags[0].physaddr)); forward_fpdus() 457 lower_32_bits(fpdu_info->frags[1].physaddr)); forward_fpdus() 459 upper_32_bits(fpdu_info->frags[1].physaddr)); forward_fpdus() 462 lower_32_bits(fpdu_info->frags[2].physaddr)); forward_fpdus() 464 upper_32_bits(fpdu_info->frags[2].physaddr)); forward_fpdus() 467 lower_32_bits(fpdu_info->frags[3].physaddr)); forward_fpdus() 469 upper_32_bits(fpdu_info->frags[3].physaddr)); forward_fpdus()
|
H A D | nes_mgt.h | 82 struct pau_fpdu_frag frags[MAX_FPDU_FRAGS]; member in struct:pau_fpdu_info
|
H A D | nes_nic.c | 440 &skb_shinfo(skb)->frags[skb_fragment_index]; nes_nic_send() 445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); nes_nic_send() 498 * " (%u frags), tso_size=%u\n", nes_netdev_start_xmit() 551 nes_debug(NES_DBG_NIC_TX, "%s: HNIC SQ full- TSO request has too many frags!\n", nes_netdev_start_xmit() 560 &skb_shinfo(skb)->frags[tso_frag_count]; nes_netdev_start_xmit() 596 " (%u frags), is_gso = %u tso_size=%u\n", nes_netdev_start_xmit() 636 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index])); nes_netdev_start_xmit() 640 tso_wqe_length += skb_frag_size(&skb_shinfo(skb)->frags[tso_frag_index++]); nes_netdev_start_xmit()
|
H A D | nes_hw.h | 912 struct nes_first_frag *first_frag_vbase; /* virtual address of first frags */
|
/linux-4.1.27/drivers/staging/unisys/uislib/ |
H A D | uisutils.c | 104 * struct phys_info frags[]) 112 * unsigned int frags_max - input - max len of frags array 113 * struct phys_info frags[] - output - frags array filled in on output 115 * entries filled in frags
|
/linux-4.1.27/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 33 static const char lowpan_frags_cache_name[] = "lowpan-frags"; 100 net = container_of(fq->q.net, struct net, ieee802154_lowpan.frags); lowpan_frag_expire() 131 q = inet_frag_find(&ieee802154_lowpan->frags, fq_find() 285 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); lowpan_frag_reasm() 398 .data = &init_net.ieee802154_lowpan.frags.high_thresh, 402 .extra1 = &init_net.ieee802154_lowpan.frags.low_thresh 406 .data = &init_net.ieee802154_lowpan.frags.low_thresh, 411 .extra2 = &init_net.ieee802154_lowpan.frags.high_thresh 415 .data = &init_net.ieee802154_lowpan.frags.timeout, 450 table[0].data = &ieee802154_lowpan->frags.high_thresh; lowpan_frags_ns_sysctl_register() 451 table[0].extra1 = &ieee802154_lowpan->frags.low_thresh; lowpan_frags_ns_sysctl_register() 452 table[0].extra2 = &init_net.ieee802154_lowpan.frags.high_thresh; lowpan_frags_ns_sysctl_register() 453 table[1].data = &ieee802154_lowpan->frags.low_thresh; lowpan_frags_ns_sysctl_register() 454 table[1].extra2 = &ieee802154_lowpan->frags.high_thresh; lowpan_frags_ns_sysctl_register() 455 table[2].data = &ieee802154_lowpan->frags.timeout; lowpan_frags_ns_sysctl_register() 527 ieee802154_lowpan->frags.high_thresh = IPV6_FRAG_HIGH_THRESH; lowpan_frags_init_net() 528 ieee802154_lowpan->frags.low_thresh = IPV6_FRAG_LOW_THRESH; lowpan_frags_init_net() 529 ieee802154_lowpan->frags.timeout = IPV6_FRAG_TIMEOUT; lowpan_frags_init_net() 531 inet_frags_init_net(&ieee802154_lowpan->frags); lowpan_frags_init_net() 542 inet_frags_exit_net(&ieee802154_lowpan->frags, &lowpan_frags); lowpan_frags_exit_net()
|
/linux-4.1.27/drivers/net/wireless/ath/wil6210/ |
H A D | trace.h | 190 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags), 191 TP_ARGS(vring, index, len, frags), 194 __field(u8, frags) 200 __entry->frags = frags; 204 TP_printk("vring %d index %d len %d frags %d", 205 __entry->vring, __entry->index, __entry->len, __entry->frags)
|
H A D | txrx.c | 1168 &skb_shinfo(skb)->frags[f]; __wil_tx_vring() 1223 nr_frags = f + 1; /* frags mapped + one for skb head */ __wil_tx_vring()
|
H A D | debugfs.c | 814 &skb_shinfo(skb)->frags[i]; wil_seq_print_skb()
|
/linux-4.1.27/net/ipv4/ |
H A D | ip_fragment.c | 58 static const char ip_frag_cache_name[] = "ip4-frags"; 92 return sum_frag_mem_limit(&net->ipv4.frags); ip_frag_mem() 136 frags); ip4_frag_init() 185 net = container_of(qp->q.net, struct net, ipv4.frags); ip_expire() 250 q = inet_frag_find(&net->ipv4.frags, &ip4_frags, &arg, hash); ip_find() 279 net = container_of(qp->q.net, struct net, ipv4.frags); ip_frag_too_far() 505 struct net *net = container_of(qp->q.net, struct net, ipv4.frags); ip_frag_reasm() 570 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); ip_frag_reasm() 708 .data = &init_net.ipv4.frags.high_thresh, 712 .extra1 = &init_net.ipv4.frags.low_thresh 716 .data = &init_net.ipv4.frags.low_thresh, 721 .extra2 = &init_net.ipv4.frags.high_thresh 725 .data = &init_net.ipv4.frags.timeout, 765 table[0].data = &net->ipv4.frags.high_thresh; ip4_frags_ns_ctl_register() 766 table[0].extra1 = &net->ipv4.frags.low_thresh; ip4_frags_ns_ctl_register() 767 table[0].extra2 = &init_net.ipv4.frags.high_thresh; ip4_frags_ns_ctl_register() 768 table[1].data = &net->ipv4.frags.low_thresh; ip4_frags_ns_ctl_register() 769 table[1].extra2 = &net->ipv4.frags.high_thresh; ip4_frags_ns_ctl_register() 770 table[2].data = &net->ipv4.frags.timeout; ip4_frags_ns_ctl_register() 835 net->ipv4.frags.high_thresh = 4 * 1024 * 1024; ipv4_frags_init_net() 836 net->ipv4.frags.low_thresh = 3 * 1024 * 1024; ipv4_frags_init_net() 842 net->ipv4.frags.timeout = IP_FRAG_TIME; ipv4_frags_init_net() 844 inet_frags_init_net(&net->ipv4.frags); ipv4_frags_init_net() 852 inet_frags_exit_net(&net->ipv4.frags, &ip4_frags); ipv4_frags_exit_net()
|
H A D | inet_lro.c | 156 lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); lro_init_desc()
|
H A D | tcp.c | 933 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); do_tcp_sendpages() 1212 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); tcp_sendmsg() 1510 * Try to splice more frags tcp_read_sock() 2965 const struct skb_frag_struct *f = &shi->frags[i]; tcp_md5_hash_skb_data()
|
H A D | udp.c | 737 struct sk_buff *frags; udp4_hwcsum() local 744 skb_walk_frags(skb, frags) { skb_walk_frags() 745 csum = csum_add(csum, frags->csum); skb_walk_frags() 746 hlen -= frags->len; skb_walk_frags()
|
H A D | ip_output.c | 1067 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); __ip_append_data() 1281 skb_frag_size_add(&skb_shinfo(skb)->frags[i-1], len); ip_append_page()
|
H A D | tcp_output.c | 1250 int size = skb_frag_size(&shinfo->frags[i]); __pskb_trim_head() 1256 shinfo->frags[k] = shinfo->frags[i]; __pskb_trim_head() 1258 shinfo->frags[k].page_offset += eat; __pskb_trim_head() 1259 skb_frag_size_sub(&shinfo->frags[k], eat); __pskb_trim_head()
|
/linux-4.1.27/net/ipv6/ |
H A D | reassembly.c | 63 static const char ip6_frag_cache_name[] = "ip6-frags"; 132 struct inet_frags *frags) ip6_expire_frag_queue() 141 inet_frag_kill(&fq->q, frags); ip6_expire_frag_queue() 169 inet_frag_put(&fq->q, frags); ip6_expire_frag_queue() 179 net = container_of(fq->q.net, struct net, ipv6.frags); ip6_frag_expire() 201 q = inet_frag_find(&net->ipv6.frags, &ip6_frags, &arg, hash); fq_find() 382 struct net *net = container_of(fq->q.net, struct net, ipv6.frags); ip6_frag_reasm() 444 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); ip6_frag_reasm() 593 .data = &init_net.ipv6.frags.high_thresh, 597 .extra1 = &init_net.ipv6.frags.low_thresh 601 .data = &init_net.ipv6.frags.low_thresh, 606 .extra2 = &init_net.ipv6.frags.high_thresh 610 .data = &init_net.ipv6.frags.timeout, 642 table[0].data = &net->ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register() 643 table[0].extra1 = &net->ipv6.frags.low_thresh; ip6_frags_ns_sysctl_register() 644 table[0].extra2 = &init_net.ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register() 645 table[1].data = &net->ipv6.frags.low_thresh; ip6_frags_ns_sysctl_register() 646 table[1].extra2 = &net->ipv6.frags.high_thresh; ip6_frags_ns_sysctl_register() 647 table[2].data = &net->ipv6.frags.timeout; ip6_frags_ns_sysctl_register() 713 net->ipv6.frags.high_thresh = IPV6_FRAG_HIGH_THRESH; ipv6_frags_init_net() 714 net->ipv6.frags.low_thresh = IPV6_FRAG_LOW_THRESH; ipv6_frags_init_net() 715 net->ipv6.frags.timeout = IPV6_FRAG_TIMEOUT; ipv6_frags_init_net() 717 inet_frags_init_net(&net->ipv6.frags); ipv6_frags_init_net() 725 inet_frags_exit_net(&net->ipv6.frags, &ip6_frags); ipv6_frags_exit_net() 131 ip6_expire_frag_queue(struct net *net, struct frag_queue *fq, struct inet_frags *frags) ip6_expire_frag_queue() argument
|
H A D | udp.c | 990 struct sk_buff *frags = skb_shinfo(skb)->frag_list; udp6_hwcsum_outgoing() local 993 if (!frags) { udp6_hwcsum_outgoing() 1010 csum = csum_add(csum, frags->csum); udp6_hwcsum_outgoing() 1011 } while ((frags = frags->next)); udp6_hwcsum_outgoing()
|
H A D | ip6_output.c | 1523 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); __ip6_append_data()
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_rx.c | 92 struct mlx4_en_rx_alloc *frags, mlx4_en_alloc_frags() 116 frags[i] = ring_alloc[i]; mlx4_en_alloc_frags() 138 struct mlx4_en_rx_alloc *frags, mlx4_en_free_frag() 142 u32 next_frag_end = frags[i].page_offset + 2 * frag_info->frag_stride; mlx4_en_free_frag() 145 if (next_frag_end > frags[i].page_size) mlx4_en_free_frag() 146 dma_unmap_page(priv->ddev, frags[i].dma, frags[i].page_size, mlx4_en_free_frag() 149 if (frags[i].page) mlx4_en_free_frag() 150 put_page(frags[i].page); mlx4_en_free_frag() 166 en_dbg(DRV, priv, " frag %d allocator: - size:%d frags:%d\n", mlx4_en_init_allocator() 241 struct mlx4_en_rx_alloc *frags = ring->rx_info + mlx4_en_prepare_rx_desc() local 244 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc, gfp); mlx4_en_prepare_rx_desc() 262 struct mlx4_en_rx_alloc *frags; mlx4_en_free_rx_desc() local 265 frags = ring->rx_info + (index << priv->log_rx_info); mlx4_en_free_rx_desc() 268 mlx4_en_free_frag(priv, frags, nr); mlx4_en_free_rx_desc() 546 struct mlx4_en_rx_alloc *frags, mlx4_en_complete_rx_desc() 550 struct skb_frag_struct *skb_frags_rx = skb_shinfo(skb)->frags; mlx4_en_complete_rx_desc() 560 if (!frags[nr].page) mlx4_en_complete_rx_desc() 568 __skb_frag_set_page(&skb_frags_rx[nr], frags[nr].page); mlx4_en_complete_rx_desc() 570 skb_frags_rx[nr].page_offset = frags[nr].page_offset; mlx4_en_complete_rx_desc() 572 frags[nr].page = NULL; mlx4_en_complete_rx_desc() 591 struct mlx4_en_rx_alloc *frags, mlx4_en_rx_skb() 609 va = page_address(frags[0].page) + frags[0].page_offset; mlx4_en_rx_skb() 623 used_frags = mlx4_en_complete_rx_desc(priv, rx_desc, frags, mlx4_en_rx_skb() 637 skb_shinfo(skb)->frags[0].page_offset += pull_len; mlx4_en_rx_skb() 640 skb_frag_size_sub(&skb_shinfo(skb)->frags[0], pull_len); mlx4_en_rx_skb() 756 struct mlx4_en_rx_alloc *frags; mlx4_en_process_rx_cq() local 784 frags = ring->rx_info + (index << priv->log_rx_info); mlx4_en_process_rx_cq() 817 ethh = (struct ethhdr *)(page_address(frags[0].page) + mlx4_en_process_rx_cq() 818 frags[0].page_offset); mlx4_en_process_rx_cq() 891 rx_desc, frags, gro_skb, 897 void *va = skb_frag_address(skb_shinfo(gro_skb)->frags); 942 skb = mlx4_en_rx_skb(priv, rx_desc, frags, length); 993 mlx4_en_free_frag(priv, frags, nr); 90 mlx4_en_alloc_frags(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct mlx4_en_rx_alloc *ring_alloc, gfp_t gfp) mlx4_en_alloc_frags() argument 137 mlx4_en_free_frag(struct mlx4_en_priv *priv, struct mlx4_en_rx_alloc *frags, int i) mlx4_en_free_frag() argument 544 mlx4_en_complete_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct sk_buff *skb, int length) mlx4_en_complete_rx_desc() argument 589 mlx4_en_rx_skb(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, unsigned int length) mlx4_en_rx_skb() argument
|
H A D | en_tx.c | 567 ptr = skb_frag_address_safe(&shinfo->frags[0]); is_inline() 654 skb_frag_size(&shinfo->frags[0])); build_inline_wqe() 675 skb_frag_size(&shinfo->frags[0])); build_inline_wqe() 805 frag = &shinfo->frags[i_frag]; mlx4_en_xmit()
|
/linux-4.1.27/net/core/ |
H A D | tso.c | 47 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; tso_build_data() 70 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; tso_start()
|
H A D | skbuff.c | 586 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_coalesce_rx_frag() 633 __skb_frag_unref(&shinfo->frags[i]); skb_release_data() 912 * skb_copy_ubufs - copy userspace skb frags buffers to kernel 917 * It will copy all frags into kernel and drop the reference 935 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; skb_copy_ubufs() 954 /* skb frags release userspace buffers */ skb_copy_ubufs() 960 /* skb frags point to kernel buffers */ skb_copy_ubufs() 963 skb_shinfo(skb)->frags[i].size); skb_copy_ubufs() 1134 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; __pskb_copy_fclone() 1197 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); pskb_expand_head() 1205 /* copy this zero copy skb frags */ pskb_expand_head() 1486 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); ___pskb_trim() 1493 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); ___pskb_trim() 1602 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); __pskb_pull_tail() 1669 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); __pskb_pull_tail() 1675 skb_shinfo(skb)->frags[k] = skb_shinfo(skb)->frags[i]; __pskb_pull_tail() 1677 skb_shinfo(skb)->frags[k].page_offset += eat; __pskb_pull_tail() 1678 skb_frag_size_sub(&skb_shinfo(skb)->frags[k], eat); __pskb_pull_tail() 1730 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; skb_copy_bits() 1912 const skb_frag_t *f = &skb_shinfo(skb)->frags[seg]; __skb_splice_bits() 2017 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_store_bits() 2091 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_checksum() 2181 end = start + skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_copy_and_csum_bits() 2185 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_and_csum_bits() 2264 * to the frags in the source buffer. 2271 * -ENOMEM: couldn't orphan frags of @from due to lack of memory 2318 skb_shinfo(to)->frags[j] = skb_shinfo(from)->frags[i]; skb_zerocopy() 2319 skb_shinfo(to)->frags[j].size = min_t(int, skb_shinfo(to)->frags[j].size, len); skb_zerocopy() 2320 len -= skb_shinfo(to)->frags[j].size; skb_zerocopy() 2528 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; skb_split_inside_header() 2552 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_split_no_header() 2555 skb_shinfo(skb1)->frags[k] = skb_shinfo(skb)->frags[i]; skb_split_no_header() 2567 skb_shinfo(skb1)->frags[0].page_offset += len - pos; skb_split_no_header() 2568 skb_frag_size_sub(&skb_shinfo(skb1)->frags[0], len - pos); skb_split_no_header() 2569 skb_frag_size_set(&skb_shinfo(skb)->frags[i], len - pos); skb_split_no_header() 2617 * If @tgt runs out of frags, the whole operation is aborted. 2623 * specialized skb free'er to handle frags without up-to-date nr_frags. 2636 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift() 2655 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift() 2656 fragto = &skb_shinfo(tgt)->frags[merge]; skb_shift() 2680 fragfrom = &skb_shinfo(skb)->frags[from]; skb_shift() 2681 fragto = &skb_shinfo(tgt)->frags[to]; skb_shift() 2708 fragfrom = &skb_shinfo(skb)->frags[0]; skb_shift() 2709 fragto = &skb_shinfo(tgt)->frags[merge]; skb_shift() 2718 skb_shinfo(skb)->frags[to++] = skb_shinfo(skb)->frags[from++]; skb_shift() 2813 frag = &skb_shinfo(st->cur_skb)->frags[st->frag_idx]; skb_seq_read() 3005 skb_frag_t *frag = skb_shinfo(head_skb)->frags; skb_segment() 3055 frag = skb_shinfo(list_skb)->frags; skb_segment() 3131 nskb_frag = skb_shinfo(nskb)->frags; skb_segment() 3145 frag = skb_shinfo(list_skb)->frags; skb_segment() 3156 "skb_segment: too many frags: %u %u\n", skb_segment() 3253 frag = pinfo->frags + nr_frags; skb_gro_receive() 3254 frag2 = skbinfo->frags + i; skb_gro_receive() 3274 skb_frag_t *frag = pinfo->frags + nr_frags; skb_gro_receive() 3292 memcpy(frag + 1, skbinfo->frags, sizeof(*frag) * skbinfo->nr_frags); skb_gro_receive() 3305 skbinfo->frags[0].page_offset += eat; skb_gro_receive() 3306 skb_frag_size_sub(&skbinfo->frags[0], eat); skb_gro_receive() 3383 end = start + skb_frag_size(&skb_shinfo(skb)->frags[i]); __skb_to_sgvec() 3385 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_to_sgvec() 4111 memcpy(skb_shinfo(to)->frags + skb_shinfo(to)->nr_frags, skb_try_coalesce() 4112 skb_shinfo(from)->frags, skb_try_coalesce() 4360 * alloc_skb_with_frags - allocate skb with page frags 4363 * @data_len: needed length in frags 4368 * This can be used to allocate a paged skb, given a maximal order for frags.
|
H A D | pktgen.c | 541 " frags: %d delay: %llu clone_skb: %d ifname: %s\n", pktgen_if_show() 972 if (!strcmp(name, "frags")) { pktgen_if_write() 979 sprintf(pg_result, "OK: frags=%u", pkt_dev->nfrags); pktgen_if_write() 2664 int frags = pkt_dev->nfrags; pktgen_finalize_skb() local 2669 if (frags > MAX_SKB_FRAGS) pktgen_finalize_skb() 2670 frags = MAX_SKB_FRAGS; pktgen_finalize_skb() 2671 len = datalen - frags * PAGE_SIZE; pktgen_finalize_skb() 2674 datalen = frags * PAGE_SIZE; pktgen_finalize_skb() 2678 frag_len = (datalen/frags) < PAGE_SIZE ? pktgen_finalize_skb() 2679 (datalen/frags) : PAGE_SIZE; pktgen_finalize_skb() 2692 skb_shinfo(skb)->frags[i].page_offset = 0; pktgen_finalize_skb() 2694 if (i == (frags - 1)) pktgen_finalize_skb() 2695 skb_frag_size_set(&skb_shinfo(skb)->frags[i], pktgen_finalize_skb() 2698 skb_frag_size_set(&skb_shinfo(skb)->frags[i], frag_len); pktgen_finalize_skb() 2699 datalen -= skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb() 2700 skb->len += skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb() 2701 skb->data_len += skb_frag_size(&skb_shinfo(skb)->frags[i]); pktgen_finalize_skb()
|
H A D | datagram.c | 378 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_datagram_iter() 465 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_datagram_from_iter() 521 * pages and build frags through them. 594 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_copy_and_csum_datagram()
|
H A D | dev.c | 2534 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; illegal_highdma() 2546 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; illegal_highdma() 3999 const skb_frag_t *frag0 = &pinfo->frags[0]; skb_gro_reset_offset() 4024 pinfo->frags[0].page_offset += grow; gro_pull_from_frag0() 4025 skb_frag_size_sub(&pinfo->frags[0], grow); gro_pull_from_frag0() 4027 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { gro_pull_from_frag0() 4029 memmove(pinfo->frags, pinfo->frags + 1, gro_pull_from_frag0() 4030 --pinfo->nr_frags * sizeof(pinfo->frags[0])); gro_pull_from_frag0()
|
/linux-4.1.27/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 53 static const char nf_frags_cache_name[] = "nf-frags"; 72 .data = &init_net.nf_frag.frags.timeout, 79 .data = &init_net.nf_frag.frags.low_thresh, 84 .extra2 = &init_net.nf_frag.frags.high_thresh 88 .data = &init_net.nf_frag.frags.high_thresh, 92 .extra1 = &init_net.nf_frag.frags.low_thresh 109 table[0].data = &net->nf_frag.frags.timeout; nf_ct_frag6_sysctl_register() 110 table[1].data = &net->nf_frag.frags.low_thresh; nf_ct_frag6_sysctl_register() 111 table[1].extra2 = &net->nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register() 112 table[2].data = &net->nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register() 113 table[2].extra1 = &net->nf_frag.frags.low_thresh; nf_ct_frag6_sysctl_register() 114 table[2].extra2 = &init_net.nf_frag.frags.high_thresh; nf_ct_frag6_sysctl_register() 185 net = container_of(fq->q.net, struct net, nf_frag.frags); nf_ct_frag6_expire() 209 q = inet_frag_find(&net->nf_frag.frags, &nf_frags, &arg, hash); fq_find() 426 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); nf_ct_frag6_reasm() 652 net->nf_frag.frags.high_thresh = IPV6_FRAG_HIGH_THRESH; nf_ct_net_init() 653 net->nf_frag.frags.low_thresh = IPV6_FRAG_LOW_THRESH; nf_ct_net_init() 654 net->nf_frag.frags.timeout = IPV6_FRAG_TIMEOUT; nf_ct_net_init() 655 inet_frags_init_net(&net->nf_frag.frags); nf_ct_net_init() 663 inet_frags_exit_net(&net->nf_frag.frags, &nf_frags); nf_ct_net_exit()
|
/linux-4.1.27/drivers/net/xen-netback/ |
H A D | netback.c | 87 * This is sized to avoid pulling headers from the frags for most 431 skb_frag_page(&skb_shinfo(skb)->frags[i]), xenvif_gop_skb() 432 skb_frag_size(&skb_shinfo(skb)->frags[i]), xenvif_gop_skb() 433 skb_shinfo(skb)->frags[i].page_offset, xenvif_gop_skb() 821 skb_frag_t *frags = shinfo->frags; xenvif_get_requests() local 838 start = (frag_get_pending_idx(&shinfo->frags[0]) == pending_idx); xenvif_get_requests() 845 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); xenvif_get_requests() 858 frags = shinfo->frags; xenvif_get_requests() 865 frag_set_pending_idx(&frags[shinfo->nr_frags], xenvif_get_requests() 919 frag_get_pending_idx(&shinfo->frags[0]) == pending_idx; xenvif_tx_check_gop() 942 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); xenvif_tx_check_gop() 979 /* Not the first error? Preceding frags already invalidated. */ xenvif_tx_check_gop() 993 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); xenvif_tx_check_gop() 1000 * the first skb's frags xenvif_tx_check_gop() 1004 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); xenvif_tx_check_gop() 1035 skb_frag_t *frag = shinfo->frags + i; xenvif_fill_frags() 1305 frag_set_pending_idx(&skb_shinfo(skb)->frags[0], xenvif_tx_build_gops() 1310 frag_set_pending_idx(&skb_shinfo(skb)->frags[0], xenvif_tx_build_gops() 1340 * frags. Returns 0 or -ENOMEM if can't allocate new pages. 1345 skb_frag_t frags[MAX_SKB_FRAGS]; xenvif_handle_frag_list() local 1354 /* Subtract frags size, we will correct it later */ xenvif_handle_frag_list() 1359 /* create a brand new frags array and coalesce there */ xenvif_handle_frag_list() 1370 put_page(frags[j].page.p); xenvif_handle_frag_list() 1382 frags[i].page.p = page; xenvif_handle_frag_list() 1383 frags[i].page_offset = 0; xenvif_handle_frag_list() 1384 skb_frag_size_set(&frags[i], len); xenvif_handle_frag_list() 1392 /* Release all the original (foreign) frags. */ xenvif_handle_frag_list() 1401 /* Fill the skb with the new (local) frags. */ xenvif_handle_frag_list() 1402 memcpy(skb_shinfo(skb)->frags, frags, i * sizeof(skb_frag_t)); xenvif_handle_frag_list() 1427 * expected to release all the frags which were mapped, xenvif_tx_submit()
|
H A D | common.h | 87 /* It's possible for an skb to have a maximal number of frags
|
/linux-4.1.27/drivers/net/wireless/ath/ath10k/ |
H A D | htt_tx.c | 418 struct htt_data_tx_desc_frag *frags; ath10k_htt_tx() local 474 frags = skb_cb->htt.txbuf->frags; ath10k_htt_tx() 476 frags[0].paddr = __cpu_to_le32(skb_cb->paddr); ath10k_htt_tx() 477 frags[0].len = __cpu_to_le32(msdu->len); ath10k_htt_tx() 478 frags[1].paddr = 0; ath10k_htt_tx() 479 frags[1].len = 0; ath10k_htt_tx() 556 sizeof(skb_cb->htt.txbuf->frags); ath10k_htt_tx()
|
H A D | debug.c | 547 "Extra frags on rings 0", pdev->r0_frags); ath10k_fw_stats_fill() 549 "Extra frags on rings 1", pdev->r1_frags); ath10k_fw_stats_fill() 551 "Extra frags on rings 2", pdev->r2_frags); ath10k_fw_stats_fill() 553 "Extra frags on rings 3", pdev->r3_frags); ath10k_fw_stats_fill()
|
H A D | htt.h | 914 /* Extra frags on rings 0-3 */ 1238 struct htt_data_tx_desc_frag frags[2]; member in struct:ath10k_htt_txbuf
|
H A D | pci.c | 212 /* NB: 50% of src nentries, since tx has 2 frags */
|
/linux-4.1.27/net/batman-adv/ |
H A D | fragmentation.h | 36 * Returns true if the frags entry has timed out, false otherwise.
|
/linux-4.1.27/fs/jffs2/ |
H A D | nodelist.c | 105 this->node->frags--; jffs2_obsolete_node_frag() 106 if (!this->node->frags) { jffs2_obsolete_node_frag() 107 /* The node has no valid frags left. It's totally obsoleted */ jffs2_obsolete_node_frag() 113 dbg_fragtree2("marking old node @0x%08x (0x%04x-0x%04x) REF_NORMAL. frags is %d\n", jffs2_obsolete_node_frag() 114 ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); jffs2_obsolete_node_frag() 185 because there are no frags with offset greater than it. no_overlapping_node() 201 because there are no frags with offset greater than it. no_overlapping_node() 286 this->node->frags++; jffs2_add_frag_to_fragtree() 374 newfrag->node->frags = 1; jffs2_add_full_dnode_to_inode() 571 if (frag->node && !(--frag->node->frags)) { rbtree_postorder_for_each_entry_safe()
|
H A D | gc.c | 520 if (nrfrags == frag->node->frags) jffs2_garbage_collect_live() 1008 if(fn->frags > 1) { jffs2_garbage_collect_hole() 1110 D1(if(unlikely(fn->frags <= 1)) { jffs2_garbage_collect_hole() 1112 __func__, fn->frags, je32_to_cpu(ri.version), jffs2_garbage_collect_hole() 1125 new_fn->frags++; jffs2_garbage_collect_hole() 1126 fn->frags--; jffs2_garbage_collect_hole() 1129 if (fn->frags) { jffs2_garbage_collect_hole() 1130 pr_warn("%s(): Old node still has frags!\n", __func__); jffs2_garbage_collect_hole() 1133 if (!new_fn->frags) { jffs2_garbage_collect_hole() 1134 pr_warn("%s(): New node has no frags!\n", __func__); jffs2_garbage_collect_hole() 1166 adjacent frags if the block containing them is already considered jffs2_garbage_collect_dnode()
|
H A D | debug.c | 87 if (fn->frags > 1) { __jffs2_dbg_fragtree_paranoia_check_nolock() 88 JFFS2_ERROR("REF_PRISTINE node at 0x%08x had %d frags. Tell dwmw2.\n", __jffs2_dbg_fragtree_paranoia_check_nolock() 89 ref_offset(fn->raw), fn->frags); __jffs2_dbg_fragtree_paranoia_check_nolock()
|
H A D | read.c | 170 frags, we read it twice. Don't do that. */ jffs2_read_inode_range()
|
H A D | dir.c | 291 /* FIXME: If you care. We'd need to use frags for the target jffs2_symlink()
|
H A D | nodelist.h | 218 uint32_t frags; /* Number of fragments which currently refer member in struct:jffs2_full_dnode
|
H A D | write.c | 191 fn->frags = 0; jffs2_write_dnode()
|
H A D | readinode.c | 211 * with frags -- we can have multiple entries starting at the same
|
/linux-4.1.27/drivers/net/ethernet/tile/ |
H A D | tilepro.c | 378 * 268 bytes with 4K pages (since the frags[] array needs 18 entries). 1597 * Prepare the "frags" info for the resulting LEPP command. 1599 * If needed, flush the memory used by the frags. 1601 static unsigned int tile_net_tx_frags(lepp_frag_t *frags, tile_net_tx_frags() argument 1617 frags[n].cpa_lo = cpa; tile_net_tx_frags() 1618 frags[n].cpa_hi = cpa >> 32; tile_net_tx_frags() 1619 frags[n].length = b_len; tile_net_tx_frags() 1620 frags[n].hash_for_home = hash_default; tile_net_tx_frags() 1626 skb_frag_t *f = &sh->frags[i]; tile_net_tx_frags() 1641 frags[n].cpa_lo = cpa; tile_net_tx_frags() 1642 frags[n].cpa_hi = cpa >> 32; tile_net_tx_frags() 1643 frags[n].length = skb_frag_size(f); tile_net_tx_frags() 1644 frags[n].hash_for_home = hash_for_home; tile_net_tx_frags() 1659 * and "sh->frags" will contain "skb->data_len" bytes of payload, and 1663 * "data" containing the header and payload, with "frags" being empty. 1717 /* Analyze the "frags". */ tile_net_tx_tso() 1719 tile_net_tx_frags(cmd->frags, skb, data + sh_len, b_len); tile_net_tx_tso() 1721 /* The size of the command, including frags and header. */ tile_net_tx_tso() 1761 memcpy(&cmd->frags[num_frags], data, sh_len); tile_net_tx_tso() 1861 lepp_frag_t frags[1 + MAX_SKB_FRAGS]; tile_net_tx() local 1906 /* ISSUE: Does not dump the "frags". */ tile_net_tx() 1917 num_frags = tile_net_tx_frags(frags, skb, data, skb_headlen(skb)); tile_net_tx() 1924 .cpa_lo = frags[i].cpa_lo, tile_net_tx() 1925 .cpa_hi = frags[i].cpa_hi, tile_net_tx() 1926 .length = frags[i].length, tile_net_tx() 1927 .hash_for_home = frags[i].hash_for_home, tile_net_tx() 2264 /* We support TSO iff the HV supports sufficient frags. */ tile_net_setup()
|
H A D | tilegx.c | 1671 * "data" containing the header and payload, with "frags" being empty. 1698 f_size = skb_frag_size(&sh->frags[f_id]); tso_count_edescs() 1802 f_size = skb_frag_size(&sh->frags[f_id]); tso_headers_prepare() 1879 f_size = skb_frag_size(&sh->frags[f_id]); tso_egress() 1880 f_data = tile_net_frag_buf(&sh->frags[f_id]); tso_egress() 1973 /* Analyze the body and frags for a transmit request. */ tile_net_tx_frags() 1974 static unsigned int tile_net_tx_frags(struct frag *frags, tile_net_tx_frags() argument 1983 frags[n].buf = b_data; tile_net_tx_frags() 1984 frags[n++].length = b_len; tile_net_tx_frags() 1988 skb_frag_t *f = &sh->frags[i]; tile_net_tx_frags() 1989 frags[n].buf = tile_net_frag_buf(f); tile_net_tx_frags() 1990 frags[n++].length = skb_frag_size(f); tile_net_tx_frags() 2011 struct frag frags[MAX_FRAGS]; tile_net_tx() local 2021 num_edescs = tile_net_tx_frags(frags, skb, data, skb_headlen(skb)); tile_net_tx() 2028 edesc.xfer_size = frags[i].length; tile_net_tx() 2029 edesc.va = va_to_tile_io_addr(frags[i].buf); tile_net_tx()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/ |
H A D | socklnd.h | 215 int tx_niov; /* # packet iovec frags */ 216 struct kvec *tx_iov; /* packet iovec frags */ 217 int tx_nkiov; /* # packet page frags */ 222 lnet_kiov_t *tx_kiov; /* packet page frags */ 244 * header, or up to LNET_MAX_IOV frags of payload of either type. */ 286 int ksnc_rx_niov; /* # iovec frags */ 287 struct kvec *ksnc_rx_iov; /* the iovec frags */ 288 int ksnc_rx_nkiov; /* # page frags */ 289 lnet_kiov_t *ksnc_rx_kiov; /* the page frags */
|
H A D | socklnd_lib-linux.c | 161 #warning "XXX risk of kmap deadlock on multiple frags..." ksocknal_lib_send_kiov() 316 #warning "XXX risk of kmap deadlock on multiple frags..." ksocknal_lib_recv_kiov()
|
H A D | socklnd_cb.c | 705 /* Ensure the frags we've been given EXACTLY match the number of ksocknal_queue_tx_locked() 707 * size parameters passed to them and just look at the frags. ksocknal_queue_tx_locked() 940 CDEBUG(D_NET, "sending %u bytes in %d frags to %s\n", ksocknal_send()
|
/linux-4.1.27/sound/soc/txx9/ |
H A D | txx9aclc.c | 95 dmadata->frags = 2; txx9aclc_pcm_prepare() 98 dmadata->frags = dmadata->buffer_bytes / dmadata->period_bytes; txx9aclc_pcm_prepare() 176 dmadata->frag_count = NR_DMA_CHAIN % dmadata->frags; txx9aclc_dma_tasklet() 196 dmadata->frag_count %= dmadata->frags; txx9aclc_dma_tasklet()
|
H A D | txx9aclc.h | 58 int frags; member in struct:txx9aclc_dmadata
|
/linux-4.1.27/drivers/media/pci/ttpci/ |
H A D | av7110_av.c | 584 p->frags = 0; av7110_p2t_init() 594 p->frags = 0; clear_p2t() 598 static int find_pes_header(u8 const *buf, long int length, int *frags) find_pes_header() argument 603 *frags = 0; find_pes_header() 632 *frags = 1; find_pes_header() 635 *frags = 2; find_pes_header() 639 *frags = 3; find_pes_header() 653 if (p->frags){ av7110_p2t_write() 655 switch(p->frags) { av7110_p2t_write() 700 p->frags = 0; av7110_p2t_write() 704 c2 = find_pes_header(buf + c, length - c, &p->frags); av7110_p2t_write() 718 c2 = find_pes_header(buf + c + add, length - c - add, &p->frags); av7110_p2t_write()
|
H A D | av7110.h | 59 int frags; member in struct:av7110_p2t
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 273 end = &si->frags[si->nr_frags]; map_skb() 275 for (fp = si->frags; fp < end; fp++) { map_skb() 284 while (fp-- > si->frags) map_skb() 302 end = &si->frags[si->nr_frags]; unmap_skb() 303 for (fp = si->frags; fp < end; fp++) unmap_skb() 868 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); write_sgl() 884 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl() 885 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); write_sgl() 890 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl() 1699 __skb_fill_page_desc(skb, 0, gl->frags[0].page, copy_frags() 1700 gl->frags[0].offset + offset, copy_frags() 1701 gl->frags[0].size - offset); copy_frags() 1704 __skb_fill_page_desc(skb, i, gl->frags[i].page, copy_frags() 1705 gl->frags[i].offset, copy_frags() 1706 gl->frags[i].size); copy_frags() 1709 get_page(gl->frags[gl->nfrags - 1].page); copy_frags() 1765 for (p = gl->frags, n = gl->nfrags - 1; n--; p++) t4_pktgl_free() 1928 * @frags: number of FL buffers to restore 1940 int frags) restore_rx_bufs() 1944 while (frags--) { restore_rx_bufs() 1950 d->page = si->frags[frags].page; restore_rx_bufs() 2019 u32 len = ntohl(rc->pldbuflen_qid), bufsz, frags; process_responses() local 2031 for (frags = 0, fp = si.frags; ; frags++, fp++) { process_responses() 2051 si.va = page_address(si.frags[0].page) + process_responses() 2052 si.frags[0].offset; process_responses() 2055 si.nfrags = frags + 1; process_responses() 2060 restore_rx_bufs(&si, &rxq->fl, frags); process_responses() 1939 restore_rx_bufs(const struct pkt_gl *si, struct sge_fl *q, int frags) restore_rx_bufs() argument
|
H A D | cxgb4.h | 469 struct page_frag frags[MAX_SKB_FRAGS]; member in struct:pkt_gl
|
/linux-4.1.27/fs/ufs/ |
H A D | ufs_fs.h | 261 #define ufs_fragstoblks(frags) ((frags) >> uspi->s_fpbshift) 307 __fs32 cs_nffree; /* number of free frags */ 313 __fs64 cs_nffree; /* number of free frags */ 322 __u64 cs_nffree; /* number of free frags */ 366 __fs32 fs_frag; /* number of frags in a block in fs */ 375 __fs32 fs_fshift; /* ``numfrags'' calc number of frags */ 548 __fs32 cg_frsum[UFS_MAXFRAG]; /* counts of available frags */ 590 __fs32 cg_frsum[8]; /* counts of available frags */ 913 __fs64 cs_nffree; /* number of free frags */
|
H A D | util.h | 354 * Determine the number of available frags given a
|
H A D | super.c | 192 pr_info(" cs_nffree(Num of free frags): %llu\n", ufs_print_super_stuff()
|
/linux-4.1.27/drivers/net/hyperv/ |
H A D | netvsc_drv.c | 285 int frags = skb_shinfo(skb)->nr_frags; init_page_array() local 305 for (i = 0; i < frags; i++) { init_page_array() 306 skb_frag_t *frag = skb_shinfo(skb)->frags + i; init_page_array() 317 int i, frags = skb_shinfo(skb)->nr_frags; count_skb_frag_slots() local 320 for (i = 0; i < frags; i++) { count_skb_frag_slots() 321 skb_frag_t *frag = skb_shinfo(skb)->frags + i; count_skb_frag_slots()
|
/linux-4.1.27/drivers/net/wireless/ath/ath9k/ |
H A D | common-debug.h | 38 * @rx_too_many_frags_err: Frames dropped due to too-many-frags received.
|
/linux-4.1.27/net/sunrpc/auth_gss/ |
H A D | gss_krb5_crypto.c | 490 struct scatterlist frags[4]; member in struct:decryptor_desc 505 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, decryptor() 516 sg_mark_end(&desc->frags[desc->fragno - 1]); decryptor() 518 ret = crypto_blkcipher_decrypt_iv(&desc->desc, desc->frags, decryptor() 519 desc->frags, thislen); decryptor() 523 sg_init_table(desc->frags, 4); decryptor() 526 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, decryptor() 553 sg_init_table(desc.frags, 4); gss_decrypt_xdr_buf() 808 sg_init_table(desc.frags, 4); gss_krb5_aes_decrypt()
|
/linux-4.1.27/include/linux/ |
H A D | inet_lro.h | 97 * headers in generated skb when using frags */
|
H A D | isdn_ppp.h | 108 struct sk_buff * frags; /* fragments sl list -- use skb->next */ member in struct:__anon11864
|
H A D | skbuff.h | 198 * Since GRO uses frags we allocate at least 16 regardless of page 283 * all frags to avoid possible bad checksum 338 skb_frag_t frags[MAX_SKB_FRAGS]; member in struct:skb_shared_info 1571 len += skb_frag_size(&skb_shinfo(skb)->frags[i]); skb_pagelen() 1591 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; __skb_fill_page_desc() 2096 * skb_orphan_frags - orphan the frags contained in a buffer 2097 * @skb: buffer to orphan frags from 2285 __skb_frag_ref(&skb_shinfo(skb)->frags[f]); skb_frag_ref() 2308 __skb_frag_unref(&skb_shinfo(skb)->frags[f]); skb_frag_unref() 2362 __skb_frag_set_page(&skb_shinfo(skb)->frags[f], page); skb_frag_set_page() 2525 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i - 1]; skb_can_coalesce()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 288 end = &si->frags[si->nr_frags]; map_skb() 289 for (fp = si->frags; fp < end; fp++) { map_skb() 298 while (fp-- > si->frags) map_skb() 924 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); write_sgl() 940 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl() 941 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); write_sgl() 946 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); write_sgl() 1350 "eth_xmit: ndesc %u, credits %u, pidx %u, len %u, frags %u", t4vf_eth_xmit() 1463 __skb_fill_page_desc(skb, 0, gl->frags[0].page, copy_frags() 1464 gl->frags[0].offset + offset, copy_frags() 1465 gl->frags[0].size - offset); copy_frags() 1468 __skb_fill_page_desc(skb, i, gl->frags[i].page, copy_frags() 1469 gl->frags[i].offset, copy_frags() 1470 gl->frags[i].size); copy_frags() 1473 get_page(gl->frags[gl->nfrags - 1].page); copy_frags() 1539 put_page(gl->frags[frag].page); t4vf_pktgl_free() 1690 int frags) restore_rx_bufs() 1694 while (frags--) { restore_rx_bufs() 1700 sdesc->page = gl->frags[frags].page; restore_rx_bufs() 1785 for (frag = 0, fp = gl.frags; /**/; frag++, fp++) { process_responses() 1808 gl.va = (page_address(gl.frags[0].page) + process_responses() 1809 gl.frags[0].offset); process_responses() 1689 restore_rx_bufs(const struct pkt_gl *gl, struct sge_fl *fl, int frags) restore_rx_bufs() argument
|
H A D | adapter.h | 152 struct page_frag frags[MAX_SKB_FRAGS]; member in struct:pkt_gl
|
/linux-4.1.27/sound/pci/ |
H A D | intel8x0m.c | 156 #define ICH_MAX_FRAGS 32 /* max hw frags */ 179 int frags; member in struct:ichdev 413 ichdev->frags = 2; snd_intel8x0m_setup_periods() 426 ichdev->frags = ichdev->size / ichdev->fragsize; snd_intel8x0m_setup_periods() 431 ichdev->lvi_frag = ICH_REG_LVI_MASK % ichdev->frags; snd_intel8x0m_setup_periods() 435 "lvi_frag = %i, frags = %i, period_size = 0x%x, period_size1 = 0x%x\n", snd_intel8x0m_setup_periods() 436 ichdev->lvi_frag, ichdev->frags, ichdev->fragsize, snd_intel8x0m_setup_periods() 475 ichdev->lvi_frag %= ichdev->frags; snd_intel8x0m_update()
|
H A D | intel8x0.c | 228 #define ICH_MAX_FRAGS 32 /* max hw frags */ 368 int frags; member in struct:ichdev 699 ichdev->frags = 2; snd_intel8x0_setup_periods() 714 ichdev->frags = ichdev->size / ichdev->fragsize; snd_intel8x0_setup_periods() 719 ichdev->lvi_frag = ICH_REG_LVI_MASK % ichdev->frags; snd_intel8x0_setup_periods() 723 "lvi_frag = %i, frags = %i, period_size = 0x%x, period_size1 = 0x%x\n", snd_intel8x0_setup_periods() 724 ichdev->lvi_frag, ichdev->frags, ichdev->fragsize, snd_intel8x0_setup_periods() 788 ichdev->lvi_frag %= ichdev->frags; snd_intel8x0_update()
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 80 static void ipoib_cm_dma_unmap_rx(struct ipoib_dev_priv *priv, int frags, ipoib_cm_dma_unmap_rx() argument 87 for (i = 0; i < frags; ++i) ipoib_cm_dma_unmap_rx() 142 int id, int frags, ipoib_cm_alloc_rx_skb() 167 for (i = 0; i < frags; i++) { ipoib_cm_alloc_rx_skb() 539 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_put_frags() 567 int frags; ipoib_cm_handle_rx_wc() local 641 frags = PAGE_ALIGN(wc->byte_len - min(wc->byte_len, ipoib_cm_handle_rx_wc() 644 newskb = ipoib_cm_alloc_rx_skb(dev, rx_ring, wr_id, frags, ipoib_cm_handle_rx_wc() 656 ipoib_cm_dma_unmap_rx(priv, frags, rx_ring[wr_id].mapping); ipoib_cm_handle_rx_wc() 657 memcpy(rx_ring[wr_id].mapping, mapping, (frags + 1) * sizeof *mapping); ipoib_cm_handle_rx_wc() 140 ipoib_cm_alloc_rx_skb(struct net_device *dev, struct ipoib_cm_rx_buf *rx_ring, int id, int frags, u64 mapping[IPOIB_CM_RX_SG], gfp_t gfp) ipoib_cm_alloc_rx_skb() argument
|
H A D | ipoib_ib.c | 285 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ipoib_dma_map_tx() 297 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; ipoib_dma_map_tx() 323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ipoib_dma_unmap_tx() 519 skb_frag_t *frags = skb_shinfo(skb)->frags; post_send() local 532 priv->tx_sge[i + off].length = skb_frag_size(&frags[i]); post_send()
|
/linux-4.1.27/block/partitions/ |
H A D | ldm.c | 1288 * @frags: Linked list of VBLK fragments 1296 static bool ldm_frag_add (const u8 *data, int size, struct list_head *frags) ldm_frag_add() argument 1302 BUG_ON (!data || !frags); ldm_frag_add() 1321 list_for_each (item, frags) { list_for_each() 1338 list_add_tail (&f->list, frags); 1378 * @frags: Linked list of VBLK fragments 1387 static bool ldm_frag_commit (struct list_head *frags, struct ldmdb *ldb) ldm_frag_commit() argument 1392 BUG_ON (!frags || !ldb); ldm_frag_commit() 1394 list_for_each (item, frags) { list_for_each() 1428 LIST_HEAD (frags); ldm_get_vblks() 1455 if (!ldm_frag_add (data, size, &frags)) ldm_get_vblks() 1464 result = ldm_frag_commit (&frags, ldb); /* Failures, already logged */ ldm_get_vblks() 1468 ldm_frag_free (&frags); ldm_get_vblks()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/ |
H A D | qla3xxx.c | 2261 * more than two frags are given. Each address list has 5 addr/len pairs. 2262 * The 5th pair in each OAL is used to point to the next OAL if more frags 2263 * are coming. That is why the frags:segment count ratio is not linear. 2265 static int ql_get_seg_count(struct ql3_adapter *qdev, unsigned short frags) ql_get_seg_count() argument 2270 if (frags <= 2) ql_get_seg_count() 2271 return frags + 1; ql_get_seg_count() 2272 else if (frags <= 6) ql_get_seg_count() 2273 return frags + 2; ql_get_seg_count() 2274 else if (frags <= 10) ql_get_seg_count() 2275 return frags + 3; ql_get_seg_count() 2276 else if (frags <= 14) ql_get_seg_count() 2277 return frags + 4; ql_get_seg_count() 2278 else if (frags <= 18) ql_get_seg_count() 2279 return frags + 5; ql_get_seg_count() 2350 skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs]; ql_send_map() 2391 "PCI mapping frags failed with error: %d\n", ql_send_map()
|
H A D | qla3xxx.h | 1014 * 1 thru 4 point to frags
|
/linux-4.1.27/drivers/scsi/cxgbi/ |
H A D | libcxgbi.c | 1139 int frags = skb_shinfo(skb)->nr_frags + cxgbi_sock_send_pdus() local 1149 if (frags >= SKB_WR_LIST_SIZE) { cxgbi_sock_send_pdus() 1150 pr_err("csk 0x%p, frags %d, %u,%u >%u.\n", cxgbi_sock_send_pdus() 2039 unsigned int dlen, struct page_frag *frags, sgl_read_to_frags() 2064 if (i && page == frags[i - 1].page && sgl_read_to_frags() 2066 frags[i - 1].offset + frags[i - 1].size) { sgl_read_to_frags() 2067 frags[i - 1].size += copy; sgl_read_to_frags() 2075 frags[i].page = page; sgl_read_to_frags() 2076 frags[i].offset = sg->offset + sgoffset; sgl_read_to_frags() 2077 frags[i].size = copy; sgl_read_to_frags() 2186 tdata->frags, MAX_PDU_FRAGS); cxgbi_conn_init_pdu() 2197 struct page_frag *frag = tdata->frags; cxgbi_conn_init_pdu() 2216 tdata->frags[i].page, cxgbi_conn_init_pdu() 2217 tdata->frags[i].offset, cxgbi_conn_init_pdu() 2218 tdata->frags[i].size); cxgbi_conn_init_pdu() 2038 sgl_read_to_frags(struct scatterlist *sg, unsigned int sgoffset, unsigned int dlen, struct page_frag *frags, int frag_max) sgl_read_to_frags() argument
|
H A D | libcxgbi.h | 594 struct page_frag frags[MAX_PDU_FRAGS]; member in struct:cxgbi_task_data
|
/linux-4.1.27/arch/blackfin/mach-bf609/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/sound/pci/lola/ |
H A D | lola_pcm.c | 328 if (str->frags >= LOLA_MAX_BDL_ENTRIES) setup_bdle() 344 str->frags++; setup_bdle() 367 str->frags = 0; lola_setup_periods() 454 lola_dsd_write(chip, str->dsd, LVI, str->frags - 1); lola_setup_controller()
|
H A D | lola.h | 302 unsigned int frags; member in struct:lola_stream
|
/linux-4.1.27/drivers/net/ |
H A D | xen-netfront.c | 311 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); xennet_alloc_rx_buffers() 477 int i, frags = skb_shinfo(skb)->nr_frags; xennet_count_skb_slots() local 482 for (i = 0; i < frags; i++) { xennet_count_skb_slots() 483 skb_frag_t *frag = skb_shinfo(skb)->frags + i; xennet_count_skb_slots() 605 /* Requests for all the frags. */ xennet_start_xmit() 607 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; xennet_start_xmit() 832 skb_frag_t *nfrag = &skb_shinfo(nskb)->frags[0]; xennet_fill_frags() 970 skb_shinfo(skb)->frags[0].page_offset = rx->offset; xennet_poll() 971 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); xennet_poll() 1095 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); xennet_release_rx_bufs()
|
/linux-4.1.27/net/sctp/ |
H A D | outqueue.c | 1103 sctp_sack_variable_t *frags; sctp_sack_update_unack_data() local 1109 frags = sack->variable; sctp_sack_update_unack_data() 1111 unack_data -= ((ntohs(frags[i].gab.end) - sctp_sack_update_unack_data() 1112 ntohs(frags[i].gab.start) + 1)); sctp_sack_update_unack_data() 1130 sctp_sack_variable_t *frags = sack->variable; sctp_outq_sack() local 1186 highest_tsn += ntohs(frags[gap_ack_blocks - 1].gab.end); 1639 sctp_sack_variable_t *frags; sctp_acked() local 1658 frags = sack->variable; sctp_acked() 1661 if (TSN_lte(ntohs(frags[i].gab.start), gap) && sctp_acked() 1662 TSN_lte(gap, ntohs(frags[i].gab.end))) sctp_acked()
|
/linux-4.1.27/drivers/isdn/i4l/ |
H A D | isdn_ppp.c | 1595 lp->netdev->pb->frags = NULL; isdn_ppp_mp_init() 1684 if ((frag = mp->frags) == NULL || MP_LT(newseq, MP_SEQ(frag))) { isdn_ppp_mp_receive() 1686 mp->frags = frag = newfrag; isdn_ppp_mp_receive() 1750 if (mp->frags == frag) isdn_ppp_mp_receive() 1751 mp->frags = nextf; isdn_ppp_mp_receive() 1770 mp->frags = nextf; isdn_ppp_mp_receive() 1810 mp->frags = isdn_ppp_mp_discard(mp, start, nextf); isdn_ppp_mp_receive() 1819 if (mp->frags == NULL) isdn_ppp_mp_receive() 1820 mp->frags = frag; isdn_ppp_mp_receive() 1827 frag = mp->frags->next; isdn_ppp_mp_receive() 1828 isdn_ppp_mp_free_skb(mp, mp->frags); isdn_ppp_mp_receive() 1829 mp->frags = frag; isdn_ppp_mp_receive() 1837 struct sk_buff *frag = lp->netdev->pb->frags; isdn_ppp_mp_cleanup() 1844 lp->netdev->pb->frags = NULL; isdn_ppp_mp_cleanup()
|
/linux-4.1.27/drivers/net/ethernet/aeroflex/ |
H A D | greth.c | 115 skb_frag_address(&skb_shinfo(skb)->frags[i]), greth_print_tx_packet() 116 skb_shinfo(skb)->frags[i].size, true); greth_print_tx_packet() 201 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_clean_rings() 521 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_start_xmit_gbit() 718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; greth_clean_tx_gbit()
|
/linux-4.1.27/arch/tile/include/hv/ |
H A D | drv_xgbe_intf.h | 439 /** The length of the 'frags' array that follows this struct. */ 442 /** The actual frags. */ 443 lepp_frag_t frags[0 /* Variable-sized; num_frags entries. */]; member in struct:__anon2852::__anon2853::__anon2860 446 * The packet header template logically follows frags[], 459 /** Maximum number of frags for one TSO command. This is adapted from
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
H A D | be.h | 264 /* Struct to remember the pages posted for rx frags */ 277 u32 rx_drops_no_frags; /* HW has no fetched frags */ 313 bool rx_post_starved; /* Zero rx frags have been posted to BE */
|
H A D | be_hw.h | 158 /* Number of rx frags posted */
|
H A D | be_main.c | 932 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; be_xmit_enqueue() 1815 skb_shinfo(skb)->frags[0].page_offset = skb_fill_rx_data() 1817 skb_frag_size_set(&skb_shinfo(skb)->frags[0], skb_fill_rx_data() 1830 /* More frags present for this completion */ skb_fill_rx_data() 1836 /* Coalesce all frags from the same physical page in one slot */ skb_fill_rx_data() 1841 skb_shinfo(skb)->frags[j].page_offset = skb_fill_rx_data() 1843 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); skb_fill_rx_data() 1849 skb_frag_size_add(&skb_shinfo(skb)->frags[j], curr_frag_len); skb_fill_rx_data() 1918 /* Coalesce all frags from the same physical page in one slot */ be_rx_compl_process_gro() 1923 skb_shinfo(skb)->frags[j].page_offset = be_rx_compl_process_gro() 1925 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); be_rx_compl_process_gro() 1929 skb_frag_size_add(&skb_shinfo(skb)->frags[j], curr_frag_len); be_rx_compl_process_gro()
|
/linux-4.1.27/drivers/net/ethernet/neterion/vxge/ |
H A D | vxge-traffic.c | 1476 priv->frags = 0; vxge_hw_fifo_txdl_reserve() 1519 txdp = (struct vxge_hw_fifo_txd *)txdlh + txdl_priv->frags; vxge_hw_fifo_txdl_buffer_set() 1529 if (txdl_priv->frags) { vxge_hw_fifo_txdl_buffer_set() 1531 (txdl_priv->frags - 1); vxge_hw_fifo_txdl_buffer_set() 1542 txdl_priv->frags++; vxge_hw_fifo_txdl_buffer_set() 1549 * @frags: Number of contiguous buffers that are part of a single 1569 txdp_last = (struct vxge_hw_fifo_txd *)txdlh + (txdl_priv->frags - 1); vxge_hw_fifo_txdl_post() 1578 txdl_priv->frags - 1, vxge_hw_fifo_txdl_post()
|
H A D | vxge-main.c | 566 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit_compl() 923 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit() 978 frag = &skb_shinfo(skb)->frags[0]; vxge_xmit() 1048 frag = &skb_shinfo(skb)->frags[0]; vxge_tx_term() 4017 "%s: Max frags : %d", vdev->ndev->name, vxge_print_parm()
|
H A D | vxge-config.h | 1044 * @frags: Total number of fragments (that is, contiguous data buffers) 1081 u32 frags; member in struct:__vxge_hw_fifo_txdl_priv
|
/linux-4.1.27/drivers/message/fusion/ |
H A D | mptctl.c | 128 static MptSge_t *kbuf_alloc_2_sgl(int bytes, u32 dir, int sge_offset, int *frags, 157 // Define max sg LIST bytes ( == (#frags + #chains) * 8 bytes each) 1018 * Outputs: frags - number of scatter gather elements 1025 kbuf_alloc_2_sgl(int bytes, u32 sgdir, int sge_offset, int *frags, kbuf_alloc_2_sgl() argument 1042 *frags = 0; kbuf_alloc_2_sgl() 1088 printk(MYIOC_s_WARN_FMT "-SG: (freeing %d frags)\n", kbuf_alloc_2_sgl() 1115 printk(MYIOC_s_WARN_FMT "(freeing %d frags)\n", ioc->name, numfrags); kbuf_alloc_2_sgl() 1123 "too many SG frags! :-(\n", ioc->name); kbuf_alloc_2_sgl() 1124 printk(MYIOC_s_WARN_FMT "-SG: (freeing %d frags)\n", kbuf_alloc_2_sgl() 1133 *frags = numfrags; kbuf_alloc_2_sgl() 1137 "%d SG frags generated!\n", ioc->name, numfrags)); kbuf_alloc_2_sgl()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | tx.c | 257 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; efx_skb_copy_bits_to_pio() 413 fragment = &skb_shinfo(skb)->frags[i]; efx_enqueue_skb() 1282 skb_shinfo(skb)->frags + frag_i); efx_enqueue_skb_tso() 1302 skb_shinfo(skb)->frags + frag_i); efx_enqueue_skb_tso()
|
/linux-4.1.27/drivers/staging/unisys/common-spar/include/channels/ |
H A D | iochannel.h | 139 * NOTE 2: In Linux, frags array in skb is currently allocated to be 463 int num_frags; /* number of fragments in frags containing data */ 464 struct phys_info frags[MAX_PHYS_INFO]; /* physical page information for member in struct:net_pkt_xmt 482 * The full packet is described in frags but the ethernet header is
|
/linux-4.1.27/include/scsi/ |
H A D | fc_frame.h | 59 /* Max number of skb frags allowed, reserving one for fcoe_crc_eof page */
|
/linux-4.1.27/arch/blackfin/mach-bf518/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/arch/blackfin/mach-bf538/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/net/rds/ |
H A D | ib_recv.c | 198 /* Recycle inc after freeing attached frags */ rds_ib_inc_free() 208 /* Free attached frags */ rds_ib_inc_free() 408 * We want to recycle several types of recv allocations, like incs and frags. 854 * Usually the frags make their way on to incs and are then freed as rds_ib_process_recv()
|
H A D | recv.c | 190 * a message have been dropped. We assume that frags that arrive rds_recv_incoming() 193 * from the sending conn until all their frags are sent. rds_recv_incoming()
|
H A D | iw_recv.c | 67 * hold page references in their frags. 685 * Usually the frags make their way on to incs and are then freed as rds_iw_process_recv()
|
/linux-4.1.27/net/xfrm/ |
H A D | xfrm_ipcomp.c | 79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; ipcomp_decompress()
|
/linux-4.1.27/sound/soc/au1x/ |
H A D | dbdma2.c | 215 DBG("bits %d frags %d frag_bytes %d is_rx %d\n", params->msbits, au1xpsc_pcm_hw_params()
|
/linux-4.1.27/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 1699 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; mvneta_tx_frag_process() 1753 int frags = 0; mvneta_tx() local 1760 frags = mvneta_tx_tso(skb, dev, txq); mvneta_tx() 1764 frags = skb_shinfo(skb)->nr_frags + 1; mvneta_tx() 1779 frags = 0; mvneta_tx() 1783 if (frags == 1) { mvneta_tx() 1802 frags = 0; mvneta_tx() 1808 if (frags > 0) { mvneta_tx() 1812 txq->count += frags; mvneta_tx() 1813 mvneta_txq_pend_desc_add(pp, txq, frags); mvneta_tx()
|
H A D | mvpp2.c | 5126 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; mvpp2_tx_frag_process() 5177 int frags = 0; mvpp2_tx() local 5186 frags = skb_shinfo(skb)->nr_frags + 1; mvpp2_tx() 5189 if (mvpp2_aggr_desc_num_check(port->priv, aggr_txq, frags) || mvpp2_tx() 5191 txq_pcpu, frags)) { mvpp2_tx() 5192 frags = 0; mvpp2_tx() 5205 frags = 0; mvpp2_tx() 5213 if (frags == 1) { mvpp2_tx() 5227 frags = 0; mvpp2_tx() 5232 txq_pcpu->reserved_num -= frags; mvpp2_tx() 5233 txq_pcpu->count += frags; mvpp2_tx() 5234 aggr_txq->count += frags; mvpp2_tx() 5238 mvpp2_aggr_txq_pend_desc_add(port, frags); mvpp2_tx() 5246 if (frags > 0) { mvpp2_tx()
|
H A D | sky2.c | 1233 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; sky2_rx_map_skb() 1247 skb_frag_size(&skb_shinfo(skb)->frags[i]), sky2_rx_map_skb() 1271 skb_frag_size(&skb_shinfo(skb)->frags[i]), sky2_rx_unmap_skb() 1950 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; sky2_xmit_frame() 2527 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; skb_put_frags()
|
H A D | mv643xx_eth.c | 676 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; has_tiny_unaligned_frags() 889 this_frag = &skb_shinfo(skb)->frags[frag]; txq_submit_frag_skb()
|
/linux-4.1.27/drivers/net/ethernet/tundra/ |
H A D | tsi108_eth.c | 655 int frags = skb_shinfo(skb)->nr_frags + 1; tsi108_send_packet() local 677 if (data->txfree - frags < MAX_SKB_FRAGS + 1) { tsi108_send_packet() 683 for (i = 0; i < frags; i++) { tsi108_send_packet() 709 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; tsi108_send_packet() 718 if (i == frags - 1) tsi108_send_packet()
|
/linux-4.1.27/drivers/scsi/cxgbi/cxgb3i/ |
H A D | cxgb3i.c | 408 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); push_tx_frames() local 409 int wrs_needed = skb_wrs[frags]; push_tx_frames() 414 WARN_ON(frags >= SKB_WR_LIST_SIZE || wrs_needed < 1); push_tx_frames() 419 csk, skb->len, skb->data_len, frags, push_tx_frames() 434 csk, skb->len, skb->data_len, frags, skb->csum, push_tx_frames()
|
/linux-4.1.27/sound/pci/hda/ |
H A D | hda_controller.c | 144 azx_sd_writew(chip, azx_dev, SD_LVI, azx_dev->frags - 1); azx_setup_controller() 301 if (azx_dev->frags >= AZX_MAX_BDL_ENTRIES) setup_bdle() 323 azx_dev->frags++; setup_bdle() 351 azx_dev->frags = 0; azx_setup_periods() 1407 azx_dev->frags = 0; azx_load_dsp_prepare()
|
H A D | hda_controller.h | 195 unsigned int frags; /* number for period in the play buffer */ member in struct:azx_dev
|
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/ |
H A D | lib-move.c | 183 /* skip complete frags before 'doffset' */ lnet_copy_iov2iov() 192 /* skip complete frags before 'soffset' */ lnet_copy_iov2iov() 243 return 0; /* no frags */ lnet_extract_iov() 246 while (offset >= src->iov_len) { /* skip initial frags */ lnet_extract_iov() 525 return 0; /* no frags */ lnet_extract_kiov() 528 while (offset >= src->kiov_len) { /* skip initial frags */ lnet_extract_kiov()
|
/linux-4.1.27/arch/blackfin/mach-bf537/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/drivers/net/ethernet/freescale/ |
H A D | gianfar.c | 2343 frag_len = skb_shinfo(skb)->frags[i].size; gfar_start_xmit() 2353 &skb_shinfo(skb)->frags[i], gfar_start_xmit() 2612 int frags = 0, nr_txbds = 0; gfar_clean_tx_ring() local 2627 frags = skb_shinfo(skb)->nr_frags; gfar_clean_tx_ring() 2633 nr_txbds = frags + 2; gfar_clean_tx_ring() 2635 nr_txbds = frags + 1; gfar_clean_tx_ring() 2671 for (i = 0; i < frags; i++) { gfar_clean_tx_ring()
|
H A D | fec_main.c | 383 this_frag = &skb_shinfo(skb)->frags[frag]; fec_enet_txq_submit_frag_skb() 390 frag_len = skb_shinfo(skb)->frags[frag].size; fec_enet_txq_submit_frag_skb()
|
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_main.c | 320 /* allocate a skb to store the frags */ fm10k_fetch_rx_buffer() 533 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; fm10k_pull_tail() 1008 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { fm10k_tx_map() 1120 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); fm10k_xmit_frame_ring()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_txrx.c | 1643 /* try the simple math, if we have too many frags per segment */ i40e_chk_linearize() 1649 frag = &skb_shinfo(skb)->frags[0]; i40e_chk_linearize() 1720 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { i40e_tx_map() 1892 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); i40e_xmit_descriptor_count()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_io.c | 599 frag = &skb_shinfo(skb)->frags[i]; qlcnic_map_tx_skb() 678 /* 14 frags supported for normal packet and qlcnic_xmit_frame() 679 * 32 frags supported for TSO packet qlcnic_xmit_frame() 683 delta += skb_frag_size(&skb_shinfo(skb)->frags[i]); qlcnic_xmit_frame()
|
/linux-4.1.27/drivers/net/vmxnet3/ |
H A D | vmxnet3_ethtool.c | 59 { " too many frags", offsetof(struct vmxnet3_tq_driver_stats,
|
H A D | vmxnet3_drv.c | 644 struct skb_frag_struct *frag = skb_shinfo(skb)->frags + vmxnet3_append_frag() 740 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; vmxnet3_map_pkt() 914 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; txd_estimate() 1281 * processing all the following non-sop frags. vmxnet3_rq_rx_complete()
|
/linux-4.1.27/arch/blackfin/mach-bf527/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/arch/blackfin/mach-bf548/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/ |
H A D | qlge_main.c | 1280 * element and there is more than 6 frags, ql_unmap_send() 1347 * When this happens, the remainder of the frags will be stored ql_map_send() 1351 skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_idx]; ql_map_send() 1407 "PCI mapping frags failed with error: %d.\n", ql_map_send() 1429 * we pass in the number of frags that mapped successfully ql_map_send() 1896 * frags: There are 18 max frags and our small ql_build_rx_skb() 2081 * skb, tack it on frags, and send it up. ql_process_mac_rx_intr()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
H A D | o2iblnd.h | 423 kib_rdma_frag_t rd_frags[0]; /* buffer frags */ 506 #define IBLND_REJECT_RDMA_FRAGS 6 /* Fatal: peer's rdma frags can't match mine */
|
H A D | o2iblnd_cb.c | 1107 CERROR("Src buffer exhausted: %d frags\n", srcidx); kiblnd_init_rdma() 1113 CERROR("Dst buffer exhausted: %d frags\n", dstidx); kiblnd_init_rdma() 1119 CERROR("RDMA too fragmented for %s (%d): %d/%d src %d/%d dst frags\n", kiblnd_init_rdma() 1476 CDEBUG(D_NET, "sending %d bytes in %d frags to %s\n", kiblnd_send()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_cmn.h | 1241 /* gro frags per page */ bnx2x_mtu_allows_gro() 1245 * 1. Number of frags should not grow above MAX_SKB_FRAGS bnx2x_mtu_allows_gro()
|
H A D | bnx2x_cmn.c | 250 /* now free frags */ bnx2x_free_tx_pkt() 3414 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]); bnx2x_pkt_req_lin() 3430 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx + wnd_size - 1]); bnx2x_pkt_req_lin() 3437 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx]); bnx2x_pkt_req_lin() 3892 nbd = 2; /* start_bd + pbd + frags (updated when pages are mapped) */ bnx2x_start_xmit() 4042 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; bnx2x_start_xmit()
|
H A D | bnx2x.h | 764 * Frag BDs - describes pages for frags
|
/linux-4.1.27/include/net/ |
H A D | ipv6.h | 323 return sum_frag_mem_limit(&net->ipv6.frags); ip6_frag_mem() 535 struct inet_frags *frags);
|
/linux-4.1.27/drivers/atm/ |
H A D | eni.c | 1136 skb_frag_page(&skb_shinfo(skb)->frags[i]) + do_tx() 1137 skb_shinfo(skb)->frags[i].page_offset, do_tx() 1138 skb_frag_size(&skb_shinfo(skb)->frags[i])); do_tx()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.c | 2434 /* try the simple math, if we have too many frags per segment */ i40e_chk_linearize() 2440 frag = &skb_shinfo(skb)->frags[0]; i40e_chk_linearize() 2517 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { i40e_tx_map() 2660 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); i40e_xmit_descriptor_count()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_main.c | 2018 frag = &skb_shinfo(skb)->frags[i]; netxen_map_tx_skb() 2073 /* 14 frags supported for normal packet and netxen_nic_xmit_frame() 2074 * 32 frags supported for TSO packet netxen_nic_xmit_frame() 2079 frag = &skb_shinfo(skb)->frags[i]; netxen_nic_xmit_frame()
|
/linux-4.1.27/drivers/staging/octeon/ |
H A D | ethernet-tx.c | 300 struct skb_frag_struct *fs = skb_shinfo(skb)->frags + i; cvm_oct_xmit()
|
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 568 frag = &skb_shinfo(skb)->frags[i]; xgbe_map_tx_skb()
|
H A D | xgbe.h | 147 * - Maximum number of SKB frags
|
H A D | xgbe-drv.c | 1352 frag = &skb_shinfo(skb)->frags[i]; xgbe_packet_info()
|
/linux-4.1.27/arch/blackfin/mach-bf533/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/arch/blackfin/mach-bf561/include/mach/ |
H A D | anomaly.h | 4 * svn://sources.blackfin.uclinux.org/toolchain/trunk/proc-defs/header-frags/
|
/linux-4.1.27/drivers/net/ethernet/ibm/ |
H A D | ibmveth.c | 1024 /* Map the frags */ ibmveth_start_xmit() 1026 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ibmveth_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/pasemi/ |
H A D | pasemi_mac.c | 302 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; pasemi_mac_unmap_tx_skb() 1503 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; pasemi_mac_start_tx()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
H A D | bnx2.c | 2883 skb_frag_size(&skb_shinfo(skb)->frags[i]), bnx2_tx_int() 2935 * last one in the frags array, so we need to recycle that page bnx2_reuse_rx_skb_pages() 2944 page = skb_frag_page(&shinfo->frags[shinfo->nr_frags]); bnx2_reuse_rx_skb_pages() 2945 __skb_frag_set_page(&shinfo->frags[shinfo->nr_frags], NULL); bnx2_reuse_rx_skb_pages() 3069 &skb_shinfo(skb)->frags[i - 1]; bnx2_rx_skb() 5436 skb_frag_size(&skb_shinfo(skb)->frags[k]), bnx2_free_tx_skbs() 6659 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; bnx2_start_xmit() 6728 skb_frag_size(&skb_shinfo(skb)->frags[i]), bnx2_start_xmit()
|
H A D | bgmac.c | 183 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; bgmac_dma_tx_add()
|
/linux-4.1.27/sound/oss/dmasound/ |
H A D | dmasound_core.c | 1024 partial frags can be queued for output. sq_ioctl() 1122 frags values using SNDCTL_DSP_GETBLKSIZE or similar */ sq_ioctl()
|
/linux-4.1.27/drivers/net/ethernet/rocker/ |
H A D | rocker.c | 3943 struct rocker_tlv *frags; rocker_port_xmit() local 3956 frags = rocker_tlv_nest_start(desc_info, ROCKER_TLV_TX_FRAGS); rocker_port_xmit() 3957 if (!frags) rocker_port_xmit() 3967 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; rocker_port_xmit() 3975 rocker_tlv_nest_end(desc_info, frags); rocker_port_xmit() 3989 rocker_tlv_nest_cancel(desc_info, frags); rocker_port_xmit()
|
/linux-4.1.27/drivers/net/ethernet/agere/ |
H A D | et131x.c | 2428 struct skb_frag_struct *frags = &skb_shinfo(skb)->frags[0]; nic_send_packet() local 2490 desc[frag].len_vlan = frags[i - 1].size; nic_send_packet() 2492 &frags[i - 1], nic_send_packet() 2494 frags[i - 1].size, nic_send_packet()
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
H A D | cassini.c | 39 * -- driver appends the rest of the data pages as frags to skbuffs 135 #define RX_COPY_ALWAYS 0 /* if 0, use frags */ 2005 skb_frag_t *frag = skb_shinfo(skb)->frags; cas_rx_process_pkt() 2007 /* normal or jumbo packets. we use frags */ cas_rx_process_pkt() 2824 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; cas_xmit_tx_ringN()
|
H A D | sunvnet.c | 1021 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; vnet_skb_map() 1062 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; vnet_skb_shape()
|
H A D | sungem.c | 1065 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; gem_start_xmit()
|
H A D | sunhme.c | 2344 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; happy_meal_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/tehuti/ |
H A D | tehuti.c | 1520 frag = &skb_shinfo(skb)->frags[i]; bdx_tx_map_skb() 1539 /* init_txd_sizes - precalculate sizes of descriptors for skbs up to 16 frags 1540 * number of frags is used as index to fetch correct descriptors size,
|
/linux-4.1.27/drivers/net/ethernet/cadence/ |
H A D | macb.c | 836 netdev_vdbg(bp->dev, "macb_rx_frame frags %u - %u (len %u)\n", macb_rx_frame() 1155 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; macb_tx_map() 1267 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); macb_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/realtek/ |
H A D | r8169.c | 6816 const skb_frag_t *frag = info->frags + cur_frag; rtl8169_xmit_frags() 7049 int frags; rtl8169_start_xmit() local 7078 frags = rtl8169_xmit_frags(tp, skb, opts); rtl8169_start_xmit() 7079 if (frags < 0) rtl8169_start_xmit() 7081 else if (frags) rtl8169_start_xmit() 7102 tp->cur_tx += frags + 1; rtl8169_start_xmit()
|
H A D | 8139cp.c | 725 this_frag = &skb_shinfo(skb)->frags[frag]; unwind_tx_frag_mapping() 811 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; cp_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/adaptec/ |
H A D | starfire.c | 1230 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[i - 1]; start_tx() 1352 skb_frag_size(&skb_shinfo(skb)->frags[i]), intr_handler()
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
H A D | enic_main.c | 363 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { enic_queue_wq_skb_cont() 489 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { enic_queue_wq_skb_tso()
|
/linux-4.1.27/drivers/net/ethernet/ti/ |
H A D | netcp_core.c | 1038 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; netcp_tx_map_skb() 1054 dev_err(netcp->ndev_dev, "out of TX desc for frags\n"); netcp_tx_map_skb()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1e/ |
H A D | atl1e_main.c | 1618 fg_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); atl1e_cal_tdp_req() 1794 frag = &skb_shinfo(skb)->frags[f]; atl1e_tx_map()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
H A D | sge.c | 1124 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; compute_large_page_tx_descs() 1256 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; write_tx_descs()
|
/linux-4.1.27/net/appletalk/ |
H A D | ddp.c | 949 /* checksum stuff in frags */ atalk_sum_skb() 952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; atalk_sum_skb()
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
H A D | c2.c | 801 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; c2_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | sge.c | 257 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]), unmap_skb() 980 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; make_sgl() 1546 pci_unmap_page(dui->pdev, *p++, skb_frag_size(&si->frags[i]), deferred_unmap_destructor() 2100 rx_frag = skb_shinfo(skb)->frags; lro_add_page()
|
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/ |
H A D | fs_enet-main.c | 569 frag = skb_shinfo(skb)->frags; fs_enet_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/ |
H A D | ixgbevf_main.c | 666 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbevf_pull_tail() 865 /* allocate a skb to store the frags */ ixgbevf_fetch_rx_buffer() 3453 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbevf_tx_map() 3629 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); ixgbevf_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/xilinx/ |
H A D | ll_temac_main.c | 686 frag = &skb_shinfo(skb)->frags[0]; temac_start_xmit()
|
H A D | xilinx_axienet_main.c | 694 frag = &skb_shinfo(skb)->frags[ii]; axienet_start_xmit()
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
H A D | be_cmds.c | 1372 "BC_%d : FW CMD to map iscsi frags failed.\n"); be_cmd_iscsi_post_sgl_pages()
|
/linux-4.1.27/drivers/scsi/fcoe/ |
H A D | fcoe_transport.c | 309 frag = &skb_shinfo(skb)->frags[i]; fcoe_fc_crc()
|
H A D | fcoe.c | 1645 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; fcoe_xmit()
|
/linux-4.1.27/drivers/staging/i2o/ |
H A D | i2o_config.c | 300 // it fails if you try and send frags out of order i2o_cfg_swdl()
|
/linux-4.1.27/drivers/staging/lustre/include/linux/lnet/ |
H A D | lib-types.h | 272 unsigned int md_niov; /* # frags */
|
/linux-4.1.27/drivers/s390/net/ |
H A D | qeth_core_main.c | 3870 frag = &skb_shinfo(skb)->frags[cnt]; qeth_get_elements_for_frags() 3976 frag = &skb_shinfo(skb)->frags[cnt]; __qeth_fill_buffer() 5719 {"tx sg frags"}, 5721 {"rx sg frags"},
|
/linux-4.1.27/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 1029 frag = &skb_shinfo(skb)->frags[f]; alx_map_tx_skb()
|
/linux-4.1.27/drivers/hsi/clients/ |
H A D | ssi_protocol.c | 193 frag = &skb_shinfo(skb)->frags[i]; ssip_skb_to_msg()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_main.c | 1687 frag = &skb_shinfo(skb)->frags[0]; write_swqe2_data() 1700 frag = &skb_shinfo(skb)->frags[i]; write_swqe2_data()
|
/linux-4.1.27/drivers/net/ethernet/packetengines/ |
H A D | hamachi.c | 181 #define IP_MF 0x2000 /* IP more frags from <netinet/ip.h> */ 1525 /* don't worry about frags */ hamachi_rx()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atlx/ |
H A D | atl1.c | 2282 frag = &skb_shinfo(skb)->frags[f]; atl1_tx_map() 2383 unsigned int f_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); atl1_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); nv_start_xmit() 2256 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; nv_start_xmit() 2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); nv_start_xmit_optimized() 2406 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; nv_start_xmit_optimized()
|
/linux-4.1.27/fs/ceph/ |
H A D | dir.c | 441 /* more frags? */ ceph_readdir()
|
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge.c | 1410 frag = skb_shinfo(skb)->frags; myri10ge_vlan_rx() 1464 rx_frags = skb_shinfo(skb)->frags; myri10ge_rx_done() 3024 frag = &skb_shinfo(skb)->frags[frag_idx]; myri10ge_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_main.c | 1695 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbe_pull_tail() 1741 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; ixgbe_dma_sync_frag() 1932 /* allocate a skb to store the frags */ ixgbe_fetch_rx_buffer() 7038 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { ixgbe_tx_map() 7301 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); ixgbe_xmit_frame_ring()
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/ |
H A D | netdev.c | 5352 frag = &skb_shinfo(skb)->frags[f]; e1000_tx_map() 5581 * frags into skb->data e1000_xmit_frame() 5609 count += DIV_ROUND_UP(skb_frag_size(&skb_shinfo(skb)->frags[f]), e1000_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/intel/igb/ |
H A D | igb_main.c | 4891 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { igb_tx_map() 5018 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); igb_xmit_frame_ring() 6708 /* allocate a skb to store the frags */ igb_fetch_rx_buffer() 6840 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; igb_pull_tail()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/ |
H A D | ixgb_main.c | 1369 frag = &skb_shinfo(skb)->frags[f]; ixgb_tx_map()
|
/linux-4.1.27/drivers/net/ethernet/natsemi/ |
H A D | ns83820.c | 1112 frag = skb_shinfo(skb)->frags; ns83820_hard_start_xmit()
|
/linux-4.1.27/drivers/net/usb/ |
H A D | usbnet.c | 1275 struct skb_frag_struct *f = &skb_shinfo(skb)->frags[i]; build_dma_sg()
|
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_main.c | 1369 const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num]; sxgbe_xmit()
|
/linux-4.1.27/drivers/net/fddi/skfp/ |
H A D | hwmtm.c | 1832 * determine the frag count and the virt addresses of the frags smt_send_mbuf()
|
/linux-4.1.27/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_fcoe.c | 329 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; bnx2fc_xmit()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/ |
H A D | d11.h | 1602 S_SEQ_NUM_FRAG, /* seq num for frags (at the start of MSDU) */
|
/linux-4.1.27/drivers/net/ethernet/calxeda/ |
H A D | xgmac.c | 1115 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; xgmac_xmit()
|
/linux-4.1.27/net/netfilter/ipvs/ |
H A D | ip_vs_core.c | 426 * IPv6 frags, only the first hit here. ip_vs_schedule()
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000/ |
H A D | e1000_main.c | 2908 frag = &skb_shinfo(skb)->frags[f]; e1000_tx_map() 3209 count += TXD_USE_COUNT(skb_frag_size(&skb_shinfo(skb)->frags[f]), e1000_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
H A D | typhoon.c | 814 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; typhoon_start_tx()
|
/linux-4.1.27/drivers/net/ethernet/alteon/ |
H A D | acenic.c | 2463 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; ace_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/hp/ |
H A D | hp100.c | 1004 * Each pdl is 508 bytes long. (63 frags * 4 bytes for address and hp100_mmuinit()
|
/linux-4.1.27/drivers/net/ethernet/ibm/emac/ |
H A D | core.c | 1484 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; emac_start_xmit_sg()
|
/linux-4.1.27/drivers/net/ethernet/intel/igbvf/ |
H A D | netdev.c | 2110 frag = &skb_shinfo(skb)->frags[f]; igbvf_tx_map_adv()
|
/linux-4.1.27/drivers/net/ethernet/ |
H A D | jme.c | 2064 frag = &skb_shinfo(skb)->frags[i]; jme_map_tx_skb()
|
/linux-4.1.27/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_main.c | 1995 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; stmmac_xmit()
|
/linux-4.1.27/drivers/net/ethernet/via/ |
H A D | via-velocity.c | 2599 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; velocity_xmit()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1c/ |
H A D | atl1c_main.c | 2155 frag = &skb_shinfo(skb)->frags[f]; atl1c_tx_map()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/ |
H A D | bcmgenet.c | 1342 &skb_shinfo(skb)->frags[i], bcmgenet_xmit()
|