remap_pfn         334 arch/x86/xen/setup.c         unsigned long start_pfn, unsigned long size, unsigned long remap_pfn)
remap_pfn         347 arch/x86/xen/setup.c 	for (ident_pfn_iter = start_pfn, remap_pfn_iter = remap_pfn;
remap_pfn         389 arch/x86/xen/setup.c 	unsigned long remap_pfn)
remap_pfn         395 arch/x86/xen/setup.c 	if (remap_pfn == 0)
remap_pfn         396 arch/x86/xen/setup.c 		remap_pfn = nr_pages;
remap_pfn         413 arch/x86/xen/setup.c 		remap_range_size = xen_find_pfn_range(&remap_pfn);
remap_pfn         424 arch/x86/xen/setup.c 		xen_do_set_identity_and_remap_chunk(cur_pfn, size, remap_pfn);
remap_pfn         428 arch/x86/xen/setup.c 		remap_pfn += size;
remap_pfn         440 arch/x86/xen/setup.c 	return remap_pfn;
remap_pfn          40 drivers/gpu/drm/i915/i915_mm.c 	struct remap_pfn *r = data;
remap_pfn          63 drivers/gpu/drm/i915/i915_mm.c 	struct remap_pfn r;
remap_pfn          75 drivers/gpu/drm/i915/i915_mm.c 	err = apply_to_page_range(r.mm, addr, size, remap_pfn, &r);
remap_pfn         275 drivers/xen/xlate_mmu.c 	struct remap_pfn *r = data;
remap_pfn         288 drivers/xen/xlate_mmu.c 	struct remap_pfn r = {
remap_pfn        1903 mm/memory.c    	unsigned long remap_pfn = pfn;
remap_pfn        1930 mm/memory.c    	err = track_pfn_remap(vma, &prot, remap_pfn, addr, PAGE_ALIGN(size));
remap_pfn        1949 mm/memory.c    		untrack_pfn(vma, remap_pfn, PAGE_ALIGN(size));