remap_pfn 334 arch/x86/xen/setup.c unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) remap_pfn 347 arch/x86/xen/setup.c for (ident_pfn_iter = start_pfn, remap_pfn_iter = remap_pfn; remap_pfn 389 arch/x86/xen/setup.c unsigned long remap_pfn) remap_pfn 395 arch/x86/xen/setup.c if (remap_pfn == 0) remap_pfn 396 arch/x86/xen/setup.c remap_pfn = nr_pages; remap_pfn 413 arch/x86/xen/setup.c remap_range_size = xen_find_pfn_range(&remap_pfn); remap_pfn 424 arch/x86/xen/setup.c xen_do_set_identity_and_remap_chunk(cur_pfn, size, remap_pfn); remap_pfn 428 arch/x86/xen/setup.c remap_pfn += size; remap_pfn 440 arch/x86/xen/setup.c return remap_pfn; remap_pfn 40 drivers/gpu/drm/i915/i915_mm.c struct remap_pfn *r = data; remap_pfn 63 drivers/gpu/drm/i915/i915_mm.c struct remap_pfn r; remap_pfn 75 drivers/gpu/drm/i915/i915_mm.c err = apply_to_page_range(r.mm, addr, size, remap_pfn, &r); remap_pfn 275 drivers/xen/xlate_mmu.c struct remap_pfn *r = data; remap_pfn 288 drivers/xen/xlate_mmu.c struct remap_pfn r = { remap_pfn 1903 mm/memory.c unsigned long remap_pfn = pfn; remap_pfn 1930 mm/memory.c err = track_pfn_remap(vma, &prot, remap_pfn, addr, PAGE_ALIGN(size)); remap_pfn 1949 mm/memory.c untrack_pfn(vma, remap_pfn, PAGE_ALIGN(size));