xdpf             2497 drivers/net/ethernet/broadcom/bnxt/bnxt.c 				xdp_return_frame(tx_buf->xdpf);
xdpf             2499 drivers/net/ethernet/broadcom/bnxt/bnxt.c 				tx_buf->xdpf = NULL;
xdpf              685 drivers/net/ethernet/broadcom/bnxt/bnxt.h 		struct xdp_frame	*xdpf;
xdpf               60 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 				     struct xdp_frame *xdpf)
xdpf               66 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 	tx_buf->xdpf = xdpf;
xdpf               91 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 			xdp_return_frame(tx_buf->xdpf);
xdpf               93 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 			tx_buf->xdpf = NULL;
xdpf              724 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		xdp_return_frame(swa->xdp.xdpf);
xdpf             1847 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 				    struct xdp_frame *xdpf)
xdpf             1865 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	if (xdpf->headroom < needed_headroom)
xdpf             1875 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	buffer_start = xdpf->data - needed_headroom;
xdpf             1878 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	if (aligned_start >= xdpf->data - xdpf->headroom)
xdpf             1884 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	swa->xdp.dma_size = xdpf->data + xdpf->len - buffer_start;
xdpf             1885 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	swa->xdp.xdpf = xdpf;
xdpf             1896 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	dpaa2_fd_set_offset(&fd, xdpf->data - buffer_start);
xdpf             1897 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	dpaa2_fd_set_len(&fd, xdpf->len);
xdpf             1933 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		struct xdp_frame *xdpf = frames[i];
xdpf             1935 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		err = dpaa2_eth_xdp_xmit_frame(net_dev, xdpf);
xdpf             1937 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              116 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h 			struct xdp_frame *xdpf;
xdpf              608 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			xdp_return_frame(tx_buffer->xdpf);
xdpf              814 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			xdp_return_frame(tx_buf->xdpf);
xdpf             2178 drivers/net/ethernet/intel/i40e/i40e_txrx.c static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
xdpf             2183 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	struct xdp_frame *xdpf = convert_to_xdp_frame(xdp);
xdpf             2185 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	if (unlikely(!xdpf))
xdpf             2188 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	return i40e_xmit_xdp_ring(xdpf, xdp_ring);
xdpf             3503 drivers/net/ethernet/intel/i40e/i40e_txrx.c static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf,
xdpf             3509 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	void *data = xdpf->data;
xdpf             3510 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	u32 size = xdpf->len;
xdpf             3524 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	tx_bi->xdpf = xdpf;
xdpf             3726 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		struct xdp_frame *xdpf = frames[i];
xdpf             3729 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		err = i40e_xmit_xdp_ring(xdpf, xdp_ring);
xdpf             3731 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              285 drivers/net/ethernet/intel/i40e/i40e_txrx.h 		struct xdp_frame *xdpf;
xdpf              705 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	xdp_return_frame(tx_bi->xdpf);
xdpf              747 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		if (tx_bi->xdpf)
xdpf              752 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		tx_bi->xdpf = NULL;
xdpf              851 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		if (tx_bi->xdpf)
xdpf              856 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		tx_bi->xdpf = NULL;
xdpf              216 drivers/net/ethernet/intel/ixgbe/ixgbe.h 		struct xdp_frame *xdpf;
xdpf             1154 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			xdp_return_frame(tx_buffer->xdpf);
xdpf             2202 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	struct xdp_frame *xdpf;
xdpf             2218 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		xdpf = convert_to_xdp_frame(xdp);
xdpf             2219 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (unlikely(!xdpf)) {
xdpf             2223 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		result = ixgbe_xmit_xdp_ring(adapter, xdpf);
xdpf             5991 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			xdp_return_frame(tx_buffer->xdpf);
xdpf             8549 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			struct xdp_frame *xdpf)
xdpf             8558 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	len = xdpf->len;
xdpf             8563 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	dma = dma_map_single(ring->dev, xdpf->data, len, DMA_TO_DEVICE);
xdpf             8578 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	tx_buffer->xdpf = xdpf;
xdpf             10354 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct xdp_frame *xdpf = frames[i];
xdpf             10357 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		err = ixgbe_xmit_xdp_ring(adapter, xdpf);
xdpf             10359 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			xdp_return_frame_rx_napi(xdpf);
xdpf               16 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h 			struct xdp_frame *xdpf);
xdpf              149 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	struct xdp_frame *xdpf;
xdpf              164 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		xdpf = convert_to_xdp_frame(xdp);
xdpf              165 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		if (unlikely(!xdpf)) {
xdpf              169 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		result = ixgbe_xmit_xdp_ring(adapter, xdpf);
xdpf              602 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		tx_bi->xdpf = NULL;
xdpf              633 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	xdp_return_frame(tx_bi->xdpf);
xdpf              660 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		if (tx_bi->xdpf)
xdpf              665 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		tx_bi->xdpf = NULL;
xdpf              739 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		if (tx_bi->xdpf)
xdpf              744 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		tx_bi->xdpf = NULL;
xdpf              464 drivers/net/ethernet/mellanox/mlx5/core/en.h 			struct xdp_frame *xdpf;
xdpf               64 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	struct xdp_frame *xdpf;
xdpf               67 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	xdpf = convert_to_xdp_frame(xdp);
xdpf               68 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	if (unlikely(!xdpf))
xdpf               71 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	xdptxd.data = xdpf->data;
xdpf               72 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 	xdptxd.len  = xdpf->len;
xdpf               92 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 			xdp_return_frame(xdpf);
xdpf               97 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		xdpi.frame.xdpf     = xdpf;
xdpf              108 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		dma_addr = di->addr + (xdpf->data - (void *)xdpf);
xdpf              370 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 					 xdpi.frame.xdpf->len, DMA_TO_DEVICE);
xdpf              371 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 			xdp_return_frame(xdpi.frame.xdpf);
xdpf              495 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		struct xdp_frame *xdpf = frames[i];
xdpf              499 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		xdptxd.data = xdpf->data;
xdpf              500 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		xdptxd.len = xdpf->len;
xdpf              505 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              511 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 		xdpi.frame.xdpf     = xdpf;
xdpf              517 drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              271 drivers/net/ethernet/socionext/netsec.c 		struct xdp_frame *xdpf;
xdpf              664 drivers/net/ethernet/socionext/netsec.c 			xdp_return_frame(desc->xdpf);
xdpf              806 drivers/net/ethernet/socionext/netsec.c 		dring->desc[idx].xdpf = buf;
xdpf              814 drivers/net/ethernet/socionext/netsec.c 				struct xdp_frame *xdpf, bool is_ndo)
xdpf              818 drivers/net/ethernet/socionext/netsec.c 	struct page *page = virt_to_page(xdpf->data);
xdpf              836 drivers/net/ethernet/socionext/netsec.c 		dma_handle = dma_map_single(priv->dev, xdpf->data, xdpf->len,
xdpf              850 drivers/net/ethernet/socionext/netsec.c 		dma_handle = page_pool_get_dma_addr(page) + xdpf->headroom +
xdpf              851 drivers/net/ethernet/socionext/netsec.c 			sizeof(*xdpf);
xdpf              852 drivers/net/ethernet/socionext/netsec.c 		dma_sync_single_for_device(priv->dev, dma_handle, xdpf->len,
xdpf              858 drivers/net/ethernet/socionext/netsec.c 	tx_desc.addr = xdpf->data;
xdpf              859 drivers/net/ethernet/socionext/netsec.c 	tx_desc.len = xdpf->len;
xdpf              861 drivers/net/ethernet/socionext/netsec.c 	netsec_set_tx_de(priv, tx_ring, &tx_ctrl, &tx_desc, xdpf);
xdpf              869 drivers/net/ethernet/socionext/netsec.c 	struct xdp_frame *xdpf = convert_to_xdp_frame(xdp);
xdpf              872 drivers/net/ethernet/socionext/netsec.c 	if (unlikely(!xdpf))
xdpf              876 drivers/net/ethernet/socionext/netsec.c 	ret = netsec_xdp_queue_one(priv, xdpf, false);
xdpf             1752 drivers/net/ethernet/socionext/netsec.c 		struct xdp_frame *xdpf = frames[i];
xdpf             1755 drivers/net/ethernet/socionext/netsec.c 		err = netsec_xdp_queue_one(priv, xdpf, true);
xdpf             1757 drivers/net/ethernet/socionext/netsec.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              358 drivers/net/ethernet/ti/cpsw.c static void *cpsw_xdpf_to_handle(struct xdp_frame *xdpf)
xdpf              360 drivers/net/ethernet/ti/cpsw.c 	return (void *)((unsigned long)xdpf | BIT(0));
xdpf              376 drivers/net/ethernet/ti/cpsw.c 	struct xdp_frame	*xdpf;
xdpf              383 drivers/net/ethernet/ti/cpsw.c 		xdpf = cpsw_handle_to_xdpf(token);
xdpf              384 drivers/net/ethernet/ti/cpsw.c 		xmeta = (void *)xdpf + CPSW_XMETA_OFFSET;
xdpf              387 drivers/net/ethernet/ti/cpsw.c 		xdp_return_frame(xdpf);
xdpf              450 drivers/net/ethernet/ti/cpsw.c static int cpsw_xdp_tx_frame(struct cpsw_priv *priv, struct xdp_frame *xdpf,
xdpf              459 drivers/net/ethernet/ti/cpsw.c 	xmeta = (void *)xdpf + CPSW_XMETA_OFFSET;
xdpf              467 drivers/net/ethernet/ti/cpsw.c 		dma += xdpf->headroom + sizeof(struct xdp_frame);
xdpf              468 drivers/net/ethernet/ti/cpsw.c 		ret = cpdma_chan_submit_mapped(txch, cpsw_xdpf_to_handle(xdpf),
xdpf              469 drivers/net/ethernet/ti/cpsw.c 					       dma, xdpf->len, port);
xdpf              471 drivers/net/ethernet/ti/cpsw.c 		if (sizeof(*xmeta) > xdpf->headroom) {
xdpf              472 drivers/net/ethernet/ti/cpsw.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              476 drivers/net/ethernet/ti/cpsw.c 		ret = cpdma_chan_submit(txch, cpsw_xdpf_to_handle(xdpf),
xdpf              477 drivers/net/ethernet/ti/cpsw.c 					xdpf->data, xdpf->len, port);
xdpf              482 drivers/net/ethernet/ti/cpsw.c 		xdp_return_frame_rx_napi(xdpf);
xdpf              494 drivers/net/ethernet/ti/cpsw.c 	struct xdp_frame *xdpf;
xdpf              512 drivers/net/ethernet/ti/cpsw.c 		xdpf = convert_to_xdp_frame(xdp);
xdpf              513 drivers/net/ethernet/ti/cpsw.c 		if (unlikely(!xdpf))
xdpf              516 drivers/net/ethernet/ti/cpsw.c 		cpsw_xdp_tx_frame(priv, xdpf, page);
xdpf             2378 drivers/net/ethernet/ti/cpsw.c 	struct xdp_frame *xdpf;
xdpf             2385 drivers/net/ethernet/ti/cpsw.c 		xdpf = frames[i];
xdpf             2386 drivers/net/ethernet/ti/cpsw.c 		if (xdpf->len < CPSW_MIN_PACKET_SIZE) {
xdpf             2387 drivers/net/ethernet/ti/cpsw.c 			xdp_return_frame_rx_napi(xdpf);
xdpf             2392 drivers/net/ethernet/ti/cpsw.c 		if (cpsw_xdp_tx_frame(priv, xdpf, NULL))
xdpf              657 drivers/net/tun.c 		struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr);
xdpf              659 drivers/net/tun.c 		xdp_return_frame(xdpf);
xdpf             2234 drivers/net/tun.c 		struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr);
xdpf             2236 drivers/net/tun.c 		ret = tun_put_user_xdp(tun, tfile, xdpf, to);
xdpf             2237 drivers/net/tun.c 		xdp_return_frame(xdpf);
xdpf             2620 drivers/net/tun.c 			struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr);
xdpf             2622 drivers/net/tun.c 			return xdpf->len;
xdpf              448 drivers/net/virtio_net.c 				   struct xdp_frame *xdpf)
xdpf              454 drivers/net/virtio_net.c 	if (unlikely(xdpf->metasize > 0))
xdpf              457 drivers/net/virtio_net.c 	if (unlikely(xdpf->headroom < vi->hdr_len))
xdpf              461 drivers/net/virtio_net.c 	xdpf->data -= vi->hdr_len;
xdpf              463 drivers/net/virtio_net.c 	hdr = xdpf->data;
xdpf              465 drivers/net/virtio_net.c 	xdpf->len   += vi->hdr_len;
xdpf              467 drivers/net/virtio_net.c 	sg_init_one(sq->sg, xdpf->data, xdpf->len);
xdpf              469 drivers/net/virtio_net.c 	err = virtqueue_add_outbuf(sq->vq, sq->sg, 1, xdp_to_ptr(xdpf),
xdpf              533 drivers/net/virtio_net.c 		struct xdp_frame *xdpf = frames[i];
xdpf              535 drivers/net/virtio_net.c 		err = __virtnet_xdp_xmit_one(vi, sq, xdpf);
xdpf              537 drivers/net/virtio_net.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              655 drivers/net/virtio_net.c 		struct xdp_frame *xdpf;
xdpf              701 drivers/net/virtio_net.c 			xdpf = convert_to_xdp_frame(&xdp);
xdpf              702 drivers/net/virtio_net.c 			if (unlikely(!xdpf))
xdpf              704 drivers/net/virtio_net.c 			err = virtnet_xdp_xmit(dev, 1, &xdpf, 0);
xdpf              803 drivers/net/virtio_net.c 		struct xdp_frame *xdpf;
xdpf              874 drivers/net/virtio_net.c 			xdpf = convert_to_xdp_frame(&xdp);
xdpf              875 drivers/net/virtio_net.c 			if (unlikely(!xdpf))
xdpf              877 drivers/net/virtio_net.c 			err = virtnet_xdp_xmit(dev, 1, &xdpf, 0);
xdpf              201 drivers/vhost/net.c 		struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr);
xdpf              203 drivers/vhost/net.c 		return xdpf->len;
xdpf              128 include/net/xdp.h void xdp_return_frame(struct xdp_frame *xdpf);
xdpf              129 include/net/xdp.h void xdp_return_frame_rx_napi(struct xdp_frame *xdpf);
xdpf              138 include/net/xdp.h static inline void xdp_release_frame(struct xdp_frame *xdpf)
xdpf              140 include/net/xdp.h 	struct xdp_mem_info *mem = &xdpf->mem;
xdpf              144 include/net/xdp.h 		__xdp_release_frame(xdpf->data, mem);
xdpf              164 kernel/bpf/cpumap.c 					 struct xdp_frame *xdpf,
xdpf              172 kernel/bpf/cpumap.c 	hard_start_headroom = sizeof(struct xdp_frame) +  xdpf->headroom;
xdpf              191 kernel/bpf/cpumap.c 	frame_size = SKB_DATA_ALIGN(xdpf->len + hard_start_headroom) +
xdpf              194 kernel/bpf/cpumap.c 	pkt_data_start = xdpf->data - hard_start_headroom;
xdpf              200 kernel/bpf/cpumap.c 	__skb_put(skb, xdpf->len);
xdpf              201 kernel/bpf/cpumap.c 	if (xdpf->metasize)
xdpf              202 kernel/bpf/cpumap.c 		skb_metadata_set(skb, xdpf->metasize);
xdpf              205 kernel/bpf/cpumap.c 	skb->protocol = eth_type_trans(skb, xdpf->dev_rx);
xdpf              214 kernel/bpf/cpumap.c 	xdp_release_frame(xdpf);
xdpf              217 kernel/bpf/cpumap.c 	xdp_scrub_frame(xdpf);
xdpf              229 kernel/bpf/cpumap.c 	struct xdp_frame *xdpf;
xdpf              231 kernel/bpf/cpumap.c 	while ((xdpf = ptr_ring_consume(ring)))
xdpf              232 kernel/bpf/cpumap.c 		if (WARN_ON_ONCE(xdpf))
xdpf              233 kernel/bpf/cpumap.c 			xdp_return_frame(xdpf);
xdpf              308 kernel/bpf/cpumap.c 			struct xdp_frame *xdpf = frames[i];
xdpf              312 kernel/bpf/cpumap.c 			skb = cpu_map_build_skb(rcpu, xdpf, skb);
xdpf              314 kernel/bpf/cpumap.c 				xdp_return_frame(xdpf);
xdpf              617 kernel/bpf/cpumap.c 		struct xdp_frame *xdpf = bq->q[i];
xdpf              620 kernel/bpf/cpumap.c 		err = __ptr_ring_produce(q, xdpf);
xdpf              624 kernel/bpf/cpumap.c 				xdp_return_frame_rx_napi(xdpf);
xdpf              626 kernel/bpf/cpumap.c 				xdp_return_frame(xdpf);
xdpf              643 kernel/bpf/cpumap.c static int bq_enqueue(struct bpf_cpu_map_entry *rcpu, struct xdp_frame *xdpf)
xdpf              660 kernel/bpf/cpumap.c 	bq->q[bq->count++] = xdpf;
xdpf              671 kernel/bpf/cpumap.c 	struct xdp_frame *xdpf;
xdpf              673 kernel/bpf/cpumap.c 	xdpf = convert_to_xdp_frame(xdp);
xdpf              674 kernel/bpf/cpumap.c 	if (unlikely(!xdpf))
xdpf              678 kernel/bpf/cpumap.c 	xdpf->dev_rx = dev_rx;
xdpf              680 kernel/bpf/cpumap.c 	bq_enqueue(rcpu, xdpf);
xdpf              361 kernel/bpf/devmap.c 		struct xdp_frame *xdpf = bq->q[i];
xdpf              363 kernel/bpf/devmap.c 		prefetch(xdpf);
xdpf              386 kernel/bpf/devmap.c 		struct xdp_frame *xdpf = bq->q[i];
xdpf              390 kernel/bpf/devmap.c 			xdp_return_frame_rx_napi(xdpf);
xdpf              392 kernel/bpf/devmap.c 			xdp_return_frame(xdpf);
xdpf              436 kernel/bpf/devmap.c static int bq_enqueue(struct bpf_dtab_netdev *obj, struct xdp_frame *xdpf,
xdpf              453 kernel/bpf/devmap.c 	bq->q[bq->count++] = xdpf;
xdpf              465 kernel/bpf/devmap.c 	struct xdp_frame *xdpf;
xdpf              475 kernel/bpf/devmap.c 	xdpf = convert_to_xdp_frame(xdp);
xdpf              476 kernel/bpf/devmap.c 	if (unlikely(!xdpf))
xdpf              479 kernel/bpf/devmap.c 	return bq_enqueue(dst, xdpf, dev_rx);
xdpf             3467 net/core/filter.c 	struct xdp_frame *xdpf;
xdpf             3478 net/core/filter.c 	xdpf = convert_to_xdp_frame(xdp);
xdpf             3479 net/core/filter.c 	if (unlikely(!xdpf))
xdpf             3482 net/core/filter.c 	sent = dev->netdev_ops->ndo_xdp_xmit(dev, 1, &xdpf, XDP_XMIT_FLUSH);
xdpf              403 net/core/xdp.c void xdp_return_frame(struct xdp_frame *xdpf)
xdpf              405 net/core/xdp.c 	__xdp_return(xdpf->data, &xdpf->mem, false, 0);
xdpf              409 net/core/xdp.c void xdp_return_frame_rx_napi(struct xdp_frame *xdpf)
xdpf              411 net/core/xdp.c 	__xdp_return(xdpf->data, &xdpf->mem, true, 0);
xdpf              471 net/core/xdp.c 	struct xdp_frame *xdpf;
xdpf              479 net/core/xdp.c 	if (sizeof(*xdpf) + totsize > PAGE_SIZE)
xdpf              487 net/core/xdp.c 	xdpf = addr;
xdpf              488 net/core/xdp.c 	memset(xdpf, 0, sizeof(*xdpf));
xdpf              490 net/core/xdp.c 	addr += sizeof(*xdpf);
xdpf              494 net/core/xdp.c 	xdpf->data = addr + metasize;
xdpf              495 net/core/xdp.c 	xdpf->len = totsize - metasize;
xdpf              496 net/core/xdp.c 	xdpf->headroom = 0;
xdpf              497 net/core/xdp.c 	xdpf->metasize = metasize;
xdpf              498 net/core/xdp.c 	xdpf->mem.type = MEM_TYPE_PAGE_ORDER0;
xdpf              501 net/core/xdp.c 	return xdpf;