pgs 28 drivers/net/ethernet/intel/i40e/i40e_xsk.c dma = dma_map_page_attrs(dev, umem->pgs[i], 0, PAGE_SIZE, pgs 31 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c dma = dma_map_page_attrs(dev, umem->pgs[i], 0, PAGE_SIZE, pgs 16 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c dma_addr_t dma = dma_map_page(dev, umem->pgs[i], 0, PAGE_SIZE, pgs 56 include/net/xdp_sock.h struct page **pgs; pgs 187 net/xdp/xdp_umem.c if (PageHighMem(umem->pgs[i])) pgs 197 net/xdp/xdp_umem.c if (PageHighMem(umem->pgs[i])) pgs 198 net/xdp/xdp_umem.c addr = vmap(&umem->pgs[i], 1, VM_MAP, PAGE_KERNEL); pgs 200 net/xdp/xdp_umem.c addr = page_address(umem->pgs[i]); pgs 215 net/xdp/xdp_umem.c put_user_pages_dirty_lock(umem->pgs, umem->npgs, true); pgs 217 net/xdp/xdp_umem.c kfree(umem->pgs); pgs 218 net/xdp/xdp_umem.c umem->pgs = NULL; pgs 288 net/xdp/xdp_umem.c umem->pgs = kcalloc(umem->npgs, sizeof(*umem->pgs), pgs 290 net/xdp/xdp_umem.c if (!umem->pgs) pgs 295 net/xdp/xdp_umem.c gup_flags | FOLL_LONGTERM, &umem->pgs[0], NULL); pgs 312 net/xdp/xdp_umem.c kfree(umem->pgs); pgs 313 net/xdp/xdp_umem.c umem->pgs = NULL; pgs 399 net/xdp/xdp_umem.c umem->pgs = NULL; pgs 599 net/xdp/xsk.c struct xdp_umem_page *pgs = umem->pages; pgs 604 net/xdp/xsk.c (pgs[i].dma + PAGE_SIZE == pgs[i + 1].dma) : pgs 605 net/xdp/xsk.c (pgs[i].addr + PAGE_SIZE == pgs[i + 1].addr); pgs 606 net/xdp/xsk.c pgs[i].addr += is_contig << XSK_NEXT_PG_CONTIG_SHIFT;