xdpf 2497 drivers/net/ethernet/broadcom/bnxt/bnxt.c xdp_return_frame(tx_buf->xdpf); xdpf 2499 drivers/net/ethernet/broadcom/bnxt/bnxt.c tx_buf->xdpf = NULL; xdpf 685 drivers/net/ethernet/broadcom/bnxt/bnxt.h struct xdp_frame *xdpf; xdpf 60 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c struct xdp_frame *xdpf) xdpf 66 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c tx_buf->xdpf = xdpf; xdpf 91 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c xdp_return_frame(tx_buf->xdpf); xdpf 93 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c tx_buf->xdpf = NULL; xdpf 724 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c xdp_return_frame(swa->xdp.xdpf); xdpf 1847 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct xdp_frame *xdpf) xdpf 1865 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c if (xdpf->headroom < needed_headroom) xdpf 1875 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c buffer_start = xdpf->data - needed_headroom; xdpf 1878 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c if (aligned_start >= xdpf->data - xdpf->headroom) xdpf 1884 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c swa->xdp.dma_size = xdpf->data + xdpf->len - buffer_start; xdpf 1885 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c swa->xdp.xdpf = xdpf; xdpf 1896 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c dpaa2_fd_set_offset(&fd, xdpf->data - buffer_start); xdpf 1897 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c dpaa2_fd_set_len(&fd, xdpf->len); xdpf 1933 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct xdp_frame *xdpf = frames[i]; xdpf 1935 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c err = dpaa2_eth_xdp_xmit_frame(net_dev, xdpf); xdpf 1937 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c xdp_return_frame_rx_napi(xdpf); xdpf 116 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h struct xdp_frame *xdpf; xdpf 608 drivers/net/ethernet/intel/i40e/i40e_txrx.c xdp_return_frame(tx_buffer->xdpf); xdpf 814 drivers/net/ethernet/intel/i40e/i40e_txrx.c xdp_return_frame(tx_buf->xdpf); xdpf 2178 drivers/net/ethernet/intel/i40e/i40e_txrx.c static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf, xdpf 2183 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct xdp_frame *xdpf = convert_to_xdp_frame(xdp); xdpf 2185 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (unlikely(!xdpf)) xdpf 2188 drivers/net/ethernet/intel/i40e/i40e_txrx.c return i40e_xmit_xdp_ring(xdpf, xdp_ring); xdpf 3503 drivers/net/ethernet/intel/i40e/i40e_txrx.c static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf, xdpf 3509 drivers/net/ethernet/intel/i40e/i40e_txrx.c void *data = xdpf->data; xdpf 3510 drivers/net/ethernet/intel/i40e/i40e_txrx.c u32 size = xdpf->len; xdpf 3524 drivers/net/ethernet/intel/i40e/i40e_txrx.c tx_bi->xdpf = xdpf; xdpf 3726 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct xdp_frame *xdpf = frames[i]; xdpf 3729 drivers/net/ethernet/intel/i40e/i40e_txrx.c err = i40e_xmit_xdp_ring(xdpf, xdp_ring); xdpf 3731 drivers/net/ethernet/intel/i40e/i40e_txrx.c xdp_return_frame_rx_napi(xdpf); xdpf 285 drivers/net/ethernet/intel/i40e/i40e_txrx.h struct xdp_frame *xdpf; xdpf 705 drivers/net/ethernet/intel/i40e/i40e_xsk.c xdp_return_frame(tx_bi->xdpf); xdpf 747 drivers/net/ethernet/intel/i40e/i40e_xsk.c if (tx_bi->xdpf) xdpf 752 drivers/net/ethernet/intel/i40e/i40e_xsk.c tx_bi->xdpf = NULL; xdpf 851 drivers/net/ethernet/intel/i40e/i40e_xsk.c if (tx_bi->xdpf) xdpf 856 drivers/net/ethernet/intel/i40e/i40e_xsk.c tx_bi->xdpf = NULL; xdpf 216 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct xdp_frame *xdpf; xdpf 1154 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp_return_frame(tx_buffer->xdpf); xdpf 2202 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct xdp_frame *xdpf; xdpf 2218 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdpf = convert_to_xdp_frame(xdp); xdpf 2219 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (unlikely(!xdpf)) { xdpf 2223 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c result = ixgbe_xmit_xdp_ring(adapter, xdpf); xdpf 5991 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp_return_frame(tx_buffer->xdpf); xdpf 8549 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct xdp_frame *xdpf) xdpf 8558 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c len = xdpf->len; xdpf 8563 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c dma = dma_map_single(ring->dev, xdpf->data, len, DMA_TO_DEVICE); xdpf 8578 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c tx_buffer->xdpf = xdpf; xdpf 10354 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct xdp_frame *xdpf = frames[i]; xdpf 10357 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c err = ixgbe_xmit_xdp_ring(adapter, xdpf); xdpf 10359 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp_return_frame_rx_napi(xdpf); xdpf 16 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h struct xdp_frame *xdpf); xdpf 149 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c struct xdp_frame *xdpf; xdpf 164 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c xdpf = convert_to_xdp_frame(xdp); xdpf 165 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (unlikely(!xdpf)) { xdpf 169 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c result = ixgbe_xmit_xdp_ring(adapter, xdpf); xdpf 602 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c tx_bi->xdpf = NULL; xdpf 633 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c xdp_return_frame(tx_bi->xdpf); xdpf 660 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (tx_bi->xdpf) xdpf 665 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c tx_bi->xdpf = NULL; xdpf 739 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (tx_bi->xdpf) xdpf 744 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c tx_bi->xdpf = NULL; xdpf 464 drivers/net/ethernet/mellanox/mlx5/core/en.h struct xdp_frame *xdpf; xdpf 64 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c struct xdp_frame *xdpf; xdpf 67 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdpf = convert_to_xdp_frame(xdp); xdpf 68 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c if (unlikely(!xdpf)) xdpf 71 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdptxd.data = xdpf->data; xdpf 72 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdptxd.len = xdpf->len; xdpf 92 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdp_return_frame(xdpf); xdpf 97 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdpi.frame.xdpf = xdpf; xdpf 108 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c dma_addr = di->addr + (xdpf->data - (void *)xdpf); xdpf 370 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdpi.frame.xdpf->len, DMA_TO_DEVICE); xdpf 371 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdp_return_frame(xdpi.frame.xdpf); xdpf 495 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c struct xdp_frame *xdpf = frames[i]; xdpf 499 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdptxd.data = xdpf->data; xdpf 500 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdptxd.len = xdpf->len; xdpf 505 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdp_return_frame_rx_napi(xdpf); xdpf 511 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdpi.frame.xdpf = xdpf; xdpf 517 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c xdp_return_frame_rx_napi(xdpf); xdpf 271 drivers/net/ethernet/socionext/netsec.c struct xdp_frame *xdpf; xdpf 664 drivers/net/ethernet/socionext/netsec.c xdp_return_frame(desc->xdpf); xdpf 806 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].xdpf = buf; xdpf 814 drivers/net/ethernet/socionext/netsec.c struct xdp_frame *xdpf, bool is_ndo) xdpf 818 drivers/net/ethernet/socionext/netsec.c struct page *page = virt_to_page(xdpf->data); xdpf 836 drivers/net/ethernet/socionext/netsec.c dma_handle = dma_map_single(priv->dev, xdpf->data, xdpf->len, xdpf 850 drivers/net/ethernet/socionext/netsec.c dma_handle = page_pool_get_dma_addr(page) + xdpf->headroom + xdpf 851 drivers/net/ethernet/socionext/netsec.c sizeof(*xdpf); xdpf 852 drivers/net/ethernet/socionext/netsec.c dma_sync_single_for_device(priv->dev, dma_handle, xdpf->len, xdpf 858 drivers/net/ethernet/socionext/netsec.c tx_desc.addr = xdpf->data; xdpf 859 drivers/net/ethernet/socionext/netsec.c tx_desc.len = xdpf->len; xdpf 861 drivers/net/ethernet/socionext/netsec.c netsec_set_tx_de(priv, tx_ring, &tx_ctrl, &tx_desc, xdpf); xdpf 869 drivers/net/ethernet/socionext/netsec.c struct xdp_frame *xdpf = convert_to_xdp_frame(xdp); xdpf 872 drivers/net/ethernet/socionext/netsec.c if (unlikely(!xdpf)) xdpf 876 drivers/net/ethernet/socionext/netsec.c ret = netsec_xdp_queue_one(priv, xdpf, false); xdpf 1752 drivers/net/ethernet/socionext/netsec.c struct xdp_frame *xdpf = frames[i]; xdpf 1755 drivers/net/ethernet/socionext/netsec.c err = netsec_xdp_queue_one(priv, xdpf, true); xdpf 1757 drivers/net/ethernet/socionext/netsec.c xdp_return_frame_rx_napi(xdpf); xdpf 358 drivers/net/ethernet/ti/cpsw.c static void *cpsw_xdpf_to_handle(struct xdp_frame *xdpf) xdpf 360 drivers/net/ethernet/ti/cpsw.c return (void *)((unsigned long)xdpf | BIT(0)); xdpf 376 drivers/net/ethernet/ti/cpsw.c struct xdp_frame *xdpf; xdpf 383 drivers/net/ethernet/ti/cpsw.c xdpf = cpsw_handle_to_xdpf(token); xdpf 384 drivers/net/ethernet/ti/cpsw.c xmeta = (void *)xdpf + CPSW_XMETA_OFFSET; xdpf 387 drivers/net/ethernet/ti/cpsw.c xdp_return_frame(xdpf); xdpf 450 drivers/net/ethernet/ti/cpsw.c static int cpsw_xdp_tx_frame(struct cpsw_priv *priv, struct xdp_frame *xdpf, xdpf 459 drivers/net/ethernet/ti/cpsw.c xmeta = (void *)xdpf + CPSW_XMETA_OFFSET; xdpf 467 drivers/net/ethernet/ti/cpsw.c dma += xdpf->headroom + sizeof(struct xdp_frame); xdpf 468 drivers/net/ethernet/ti/cpsw.c ret = cpdma_chan_submit_mapped(txch, cpsw_xdpf_to_handle(xdpf), xdpf 469 drivers/net/ethernet/ti/cpsw.c dma, xdpf->len, port); xdpf 471 drivers/net/ethernet/ti/cpsw.c if (sizeof(*xmeta) > xdpf->headroom) { xdpf 472 drivers/net/ethernet/ti/cpsw.c xdp_return_frame_rx_napi(xdpf); xdpf 476 drivers/net/ethernet/ti/cpsw.c ret = cpdma_chan_submit(txch, cpsw_xdpf_to_handle(xdpf), xdpf 477 drivers/net/ethernet/ti/cpsw.c xdpf->data, xdpf->len, port); xdpf 482 drivers/net/ethernet/ti/cpsw.c xdp_return_frame_rx_napi(xdpf); xdpf 494 drivers/net/ethernet/ti/cpsw.c struct xdp_frame *xdpf; xdpf 512 drivers/net/ethernet/ti/cpsw.c xdpf = convert_to_xdp_frame(xdp); xdpf 513 drivers/net/ethernet/ti/cpsw.c if (unlikely(!xdpf)) xdpf 516 drivers/net/ethernet/ti/cpsw.c cpsw_xdp_tx_frame(priv, xdpf, page); xdpf 2378 drivers/net/ethernet/ti/cpsw.c struct xdp_frame *xdpf; xdpf 2385 drivers/net/ethernet/ti/cpsw.c xdpf = frames[i]; xdpf 2386 drivers/net/ethernet/ti/cpsw.c if (xdpf->len < CPSW_MIN_PACKET_SIZE) { xdpf 2387 drivers/net/ethernet/ti/cpsw.c xdp_return_frame_rx_napi(xdpf); xdpf 2392 drivers/net/ethernet/ti/cpsw.c if (cpsw_xdp_tx_frame(priv, xdpf, NULL)) xdpf 657 drivers/net/tun.c struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); xdpf 659 drivers/net/tun.c xdp_return_frame(xdpf); xdpf 2234 drivers/net/tun.c struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); xdpf 2236 drivers/net/tun.c ret = tun_put_user_xdp(tun, tfile, xdpf, to); xdpf 2237 drivers/net/tun.c xdp_return_frame(xdpf); xdpf 2620 drivers/net/tun.c struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); xdpf 2622 drivers/net/tun.c return xdpf->len; xdpf 448 drivers/net/virtio_net.c struct xdp_frame *xdpf) xdpf 454 drivers/net/virtio_net.c if (unlikely(xdpf->metasize > 0)) xdpf 457 drivers/net/virtio_net.c if (unlikely(xdpf->headroom < vi->hdr_len)) xdpf 461 drivers/net/virtio_net.c xdpf->data -= vi->hdr_len; xdpf 463 drivers/net/virtio_net.c hdr = xdpf->data; xdpf 465 drivers/net/virtio_net.c xdpf->len += vi->hdr_len; xdpf 467 drivers/net/virtio_net.c sg_init_one(sq->sg, xdpf->data, xdpf->len); xdpf 469 drivers/net/virtio_net.c err = virtqueue_add_outbuf(sq->vq, sq->sg, 1, xdp_to_ptr(xdpf), xdpf 533 drivers/net/virtio_net.c struct xdp_frame *xdpf = frames[i]; xdpf 535 drivers/net/virtio_net.c err = __virtnet_xdp_xmit_one(vi, sq, xdpf); xdpf 537 drivers/net/virtio_net.c xdp_return_frame_rx_napi(xdpf); xdpf 655 drivers/net/virtio_net.c struct xdp_frame *xdpf; xdpf 701 drivers/net/virtio_net.c xdpf = convert_to_xdp_frame(&xdp); xdpf 702 drivers/net/virtio_net.c if (unlikely(!xdpf)) xdpf 704 drivers/net/virtio_net.c err = virtnet_xdp_xmit(dev, 1, &xdpf, 0); xdpf 803 drivers/net/virtio_net.c struct xdp_frame *xdpf; xdpf 874 drivers/net/virtio_net.c xdpf = convert_to_xdp_frame(&xdp); xdpf 875 drivers/net/virtio_net.c if (unlikely(!xdpf)) xdpf 877 drivers/net/virtio_net.c err = virtnet_xdp_xmit(dev, 1, &xdpf, 0); xdpf 201 drivers/vhost/net.c struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); xdpf 203 drivers/vhost/net.c return xdpf->len; xdpf 128 include/net/xdp.h void xdp_return_frame(struct xdp_frame *xdpf); xdpf 129 include/net/xdp.h void xdp_return_frame_rx_napi(struct xdp_frame *xdpf); xdpf 138 include/net/xdp.h static inline void xdp_release_frame(struct xdp_frame *xdpf) xdpf 140 include/net/xdp.h struct xdp_mem_info *mem = &xdpf->mem; xdpf 144 include/net/xdp.h __xdp_release_frame(xdpf->data, mem); xdpf 164 kernel/bpf/cpumap.c struct xdp_frame *xdpf, xdpf 172 kernel/bpf/cpumap.c hard_start_headroom = sizeof(struct xdp_frame) + xdpf->headroom; xdpf 191 kernel/bpf/cpumap.c frame_size = SKB_DATA_ALIGN(xdpf->len + hard_start_headroom) + xdpf 194 kernel/bpf/cpumap.c pkt_data_start = xdpf->data - hard_start_headroom; xdpf 200 kernel/bpf/cpumap.c __skb_put(skb, xdpf->len); xdpf 201 kernel/bpf/cpumap.c if (xdpf->metasize) xdpf 202 kernel/bpf/cpumap.c skb_metadata_set(skb, xdpf->metasize); xdpf 205 kernel/bpf/cpumap.c skb->protocol = eth_type_trans(skb, xdpf->dev_rx); xdpf 214 kernel/bpf/cpumap.c xdp_release_frame(xdpf); xdpf 217 kernel/bpf/cpumap.c xdp_scrub_frame(xdpf); xdpf 229 kernel/bpf/cpumap.c struct xdp_frame *xdpf; xdpf 231 kernel/bpf/cpumap.c while ((xdpf = ptr_ring_consume(ring))) xdpf 232 kernel/bpf/cpumap.c if (WARN_ON_ONCE(xdpf)) xdpf 233 kernel/bpf/cpumap.c xdp_return_frame(xdpf); xdpf 308 kernel/bpf/cpumap.c struct xdp_frame *xdpf = frames[i]; xdpf 312 kernel/bpf/cpumap.c skb = cpu_map_build_skb(rcpu, xdpf, skb); xdpf 314 kernel/bpf/cpumap.c xdp_return_frame(xdpf); xdpf 617 kernel/bpf/cpumap.c struct xdp_frame *xdpf = bq->q[i]; xdpf 620 kernel/bpf/cpumap.c err = __ptr_ring_produce(q, xdpf); xdpf 624 kernel/bpf/cpumap.c xdp_return_frame_rx_napi(xdpf); xdpf 626 kernel/bpf/cpumap.c xdp_return_frame(xdpf); xdpf 643 kernel/bpf/cpumap.c static int bq_enqueue(struct bpf_cpu_map_entry *rcpu, struct xdp_frame *xdpf) xdpf 660 kernel/bpf/cpumap.c bq->q[bq->count++] = xdpf; xdpf 671 kernel/bpf/cpumap.c struct xdp_frame *xdpf; xdpf 673 kernel/bpf/cpumap.c xdpf = convert_to_xdp_frame(xdp); xdpf 674 kernel/bpf/cpumap.c if (unlikely(!xdpf)) xdpf 678 kernel/bpf/cpumap.c xdpf->dev_rx = dev_rx; xdpf 680 kernel/bpf/cpumap.c bq_enqueue(rcpu, xdpf); xdpf 361 kernel/bpf/devmap.c struct xdp_frame *xdpf = bq->q[i]; xdpf 363 kernel/bpf/devmap.c prefetch(xdpf); xdpf 386 kernel/bpf/devmap.c struct xdp_frame *xdpf = bq->q[i]; xdpf 390 kernel/bpf/devmap.c xdp_return_frame_rx_napi(xdpf); xdpf 392 kernel/bpf/devmap.c xdp_return_frame(xdpf); xdpf 436 kernel/bpf/devmap.c static int bq_enqueue(struct bpf_dtab_netdev *obj, struct xdp_frame *xdpf, xdpf 453 kernel/bpf/devmap.c bq->q[bq->count++] = xdpf; xdpf 465 kernel/bpf/devmap.c struct xdp_frame *xdpf; xdpf 475 kernel/bpf/devmap.c xdpf = convert_to_xdp_frame(xdp); xdpf 476 kernel/bpf/devmap.c if (unlikely(!xdpf)) xdpf 479 kernel/bpf/devmap.c return bq_enqueue(dst, xdpf, dev_rx); xdpf 3467 net/core/filter.c struct xdp_frame *xdpf; xdpf 3478 net/core/filter.c xdpf = convert_to_xdp_frame(xdp); xdpf 3479 net/core/filter.c if (unlikely(!xdpf)) xdpf 3482 net/core/filter.c sent = dev->netdev_ops->ndo_xdp_xmit(dev, 1, &xdpf, XDP_XMIT_FLUSH); xdpf 403 net/core/xdp.c void xdp_return_frame(struct xdp_frame *xdpf) xdpf 405 net/core/xdp.c __xdp_return(xdpf->data, &xdpf->mem, false, 0); xdpf 409 net/core/xdp.c void xdp_return_frame_rx_napi(struct xdp_frame *xdpf) xdpf 411 net/core/xdp.c __xdp_return(xdpf->data, &xdpf->mem, true, 0); xdpf 471 net/core/xdp.c struct xdp_frame *xdpf; xdpf 479 net/core/xdp.c if (sizeof(*xdpf) + totsize > PAGE_SIZE) xdpf 487 net/core/xdp.c xdpf = addr; xdpf 488 net/core/xdp.c memset(xdpf, 0, sizeof(*xdpf)); xdpf 490 net/core/xdp.c addr += sizeof(*xdpf); xdpf 494 net/core/xdp.c xdpf->data = addr + metasize; xdpf 495 net/core/xdp.c xdpf->len = totsize - metasize; xdpf 496 net/core/xdp.c xdpf->headroom = 0; xdpf 497 net/core/xdp.c xdpf->metasize = metasize; xdpf 498 net/core/xdp.c xdpf->mem.type = MEM_TYPE_PAGE_ORDER0; xdpf 501 net/core/xdp.c return xdpf;