pgs                28 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		dma = dma_map_page_attrs(dev, umem->pgs[i], 0, PAGE_SIZE,
pgs                31 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		dma = dma_map_page_attrs(dev, umem->pgs[i], 0, PAGE_SIZE,
pgs                16 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c 		dma_addr_t dma = dma_map_page(dev, umem->pgs[i], 0, PAGE_SIZE,
pgs                56 include/net/xdp_sock.h 	struct page **pgs;
pgs               187 net/xdp/xdp_umem.c 		if (PageHighMem(umem->pgs[i]))
pgs               197 net/xdp/xdp_umem.c 		if (PageHighMem(umem->pgs[i]))
pgs               198 net/xdp/xdp_umem.c 			addr = vmap(&umem->pgs[i], 1, VM_MAP, PAGE_KERNEL);
pgs               200 net/xdp/xdp_umem.c 			addr = page_address(umem->pgs[i]);
pgs               215 net/xdp/xdp_umem.c 	put_user_pages_dirty_lock(umem->pgs, umem->npgs, true);
pgs               217 net/xdp/xdp_umem.c 	kfree(umem->pgs);
pgs               218 net/xdp/xdp_umem.c 	umem->pgs = NULL;
pgs               288 net/xdp/xdp_umem.c 	umem->pgs = kcalloc(umem->npgs, sizeof(*umem->pgs),
pgs               290 net/xdp/xdp_umem.c 	if (!umem->pgs)
pgs               295 net/xdp/xdp_umem.c 			      gup_flags | FOLL_LONGTERM, &umem->pgs[0], NULL);
pgs               312 net/xdp/xdp_umem.c 	kfree(umem->pgs);
pgs               313 net/xdp/xdp_umem.c 	umem->pgs = NULL;
pgs               399 net/xdp/xdp_umem.c 	umem->pgs = NULL;
pgs               599 net/xdp/xsk.c  	struct xdp_umem_page *pgs = umem->pages;
pgs               604 net/xdp/xsk.c  			(pgs[i].dma + PAGE_SIZE == pgs[i + 1].dma) :
pgs               605 net/xdp/xsk.c  			(pgs[i].addr + PAGE_SIZE == pgs[i + 1].addr);
pgs               606 net/xdp/xsk.c  		pgs[i].addr += is_contig << XSK_NEXT_PG_CONTIG_SHIFT;