Home
last modified time | relevance | path

Searched refs:page_to_pfn (Results 1 – 194 of 194) sorted by relevance

/linux-4.4.14/drivers/xen/
Dbiomerge.c10 unsigned long bfn1 = pfn_to_bfn(page_to_pfn(vec1->bv_page)); in xen_biovec_phys_mergeable()
11 unsigned long bfn2 = pfn_to_bfn(page_to_pfn(vec2->bv_page)); in xen_biovec_phys_mergeable()
Dballoon.c458 unsigned long pfn = page_to_pfn(page); in increase_reservation()
536 unsigned long pfn = page_to_pfn(page); in decrease_reservation()
662 ret = xen_alloc_p2m_entry(page_to_pfn(page)); in alloc_xenballooned_pages()
Dxlate_mmu.c97 pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot)); in remap_pte_fn()
Dgntdev.c283 pfn_to_kaddr(page_to_pfn(map->pages[i])); in map_grant_pages()
299 pfn_to_kaddr(page_to_pfn(map->pages[i])); in map_grant_pages()
339 uint8_t *tmp = pfn_to_kaddr(page_to_pfn(map->pages[pgno])); in __unmap_grant_pages()
Dgrant-table.c330 entry->ref, page_to_pfn(entry->page)); in gnttab_handle_deferred()
378 what, ref, page ? page_to_pfn(page) : -1); in gnttab_add_deferred()
Dxen-scsiback.c250 unsigned long pfn = page_to_pfn(page); in vaddr_page()
/linux-4.4.14/arch/microblaze/include/asm/
Dpage.h153 # define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
154 # define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
158 # define page_to_virt(page) (pfn_to_virt(page_to_pfn(page)))
159 # define page_to_phys(page) (pfn_to_phys(page_to_pfn(page)))
/linux-4.4.14/include/trace/events/
Dkmem.h185 __entry->pfn = page_to_pfn(page);
207 __entry->pfn = page_to_pfn(page);
232 __entry->pfn = page ? page_to_pfn(page) : -1UL;
259 __entry->pfn = page ? page_to_pfn(page) : -1UL;
303 __entry->pfn = page ? page_to_pfn(page) : -1UL;
333 __entry->pfn = page_to_pfn(page);
Dpagemap.h45 __entry->pfn = page_to_pfn(page);
76 __entry->pfn = page_to_pfn(page);
Dfilemap.h28 __entry->pfn = page_to_pfn(page);
Dvmscan.h344 __entry->pfn = page_to_pfn(page);
/linux-4.4.14/arch/arm/mm/
Dflush.c143 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access()
152 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access()
237 flush_pfn_alias(page_to_pfn(page), in __flush_dcache_page()
267 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); in __flush_dcache_aliases()
399 pfn = page_to_pfn(page); in __flush_anon_page()
Dcopypage-xsc3.c80 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage()
Dcopypage-feroceon.c77 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage()
Dcopypage-v4wb.c57 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage()
Ddma-mapping.c82 return pfn_to_dma(dev, page_to_pfn(page)) + offset; in arm_dma_map_page()
89 return pfn_to_dma(dev, page_to_pfn(page)) + offset; in arm_coherent_dma_map_page()
235 phys_addr_t base = __pfn_to_phys(page_to_pfn(page)); in __dma_clear_buffer()
666 *handle = pfn_to_dma(dev, page_to_pfn(page)); in __dma_alloc()
799 pfn = page_to_pfn(page) + offset / PAGE_SIZE; in dma_cache_maint_page()
881 pfn = page_to_pfn(page) + off / PAGE_SIZE; in __dma_page_dev_to_cpu()
1262 unsigned int next_pfn = page_to_pfn(pages[i]) + 1; in __iommu_create_mapping()
1267 if (page_to_pfn(pages[j]) != next_pfn) in __iommu_create_mapping()
/linux-4.4.14/include/linux/
Dpageblock-flags.h81 get_pfnblock_flags_mask(page, page_to_pfn(page), \
85 set_pfnblock_flags_mask(page, flags, page_to_pfn(page), \
Dswapops.h108 page_to_pfn(page)); in make_migration_entry()
176 return swp_entry(SWP_HWPOISON, page_to_pfn(page)); in make_hwpoison_entry()
Dmmzone.h82 get_pfnblock_flags_mask(page, page_to_pfn(page), \
Dmm.h85 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
908 return __va(PFN_PHYS(page_to_pfn(page))); in lowmem_page_address()
/linux-4.4.14/arch/sparc/include/asm/
Dcacheflush_32.h23 flush_cache_page(vma, vaddr, page_to_pfn(page));\
28 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Dcacheflush_64.h59 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
66 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
Dpage.h4 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable_32.h285 return __pte((page_to_pfn(page) << (PAGE_SHIFT-4)) | pgprot_val(pgprot)); in mk_pte()
Dpgtable_64.h244 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
253 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.4.14/drivers/gpu/drm/gma500/
Dgtt.c109 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_insert()
114 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_insert()
143 pte = psb_gtt_mask_pte(page_to_pfn(dev_priv->scratch_page), in psb_gtt_remove()
183 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_roll()
188 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_roll()
548 pfn_base = page_to_pfn(dev_priv->scratch_page); in psb_gtt_init()
Dmmu.c146 PSB_WSGX32(page_to_pfn(pd->p) << PAGE_SHIFT, offset); in psb_mmu_set_pd_context()
196 pd->invalid_pde = psb_mmu_mask_pte(page_to_pfn(pd->dummy_pt), in psb_mmu_alloc_pd()
198 pd->invalid_pte = psb_mmu_mask_pte(page_to_pfn(pd->dummy_page), in psb_mmu_alloc_pd()
353 v[index] = (page_to_pfn(pt->p) << 12) | pd->pd_mask; in psb_mmu_pt_alloc_map_lock()
435 return page_to_pfn(pd->p) << PAGE_SHIFT; in psb_get_default_pd_addr()
747 pte = psb_mmu_mask_pte(page_to_pfn(*pages++), in psb_mmu_insert_pages()
Dgem.c214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault()
/linux-4.4.14/arch/avr32/include/asm/
Dpage.h82 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
89 #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + PHYS_PFN_OFFSET) macro
Dpgtable.h280 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/include/asm-generic/
Dpage.h87 #define page_to_virt(page) pfn_to_virt(page_to_pfn(page))
90 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dmemory_model.h78 #define page_to_pfn __page_to_pfn macro
Dpgtable.h599 #define my_zero_pfn(addr) page_to_pfn(ZERO_PAGE(addr))
/linux-4.4.14/arch/alpha/include/asm/
Dmmzone.h83 pfn = page_to_pfn(page) << 32; \
101 (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h214 pte_val(pte) = (page_to_pfn(page) << 32) | pgprot_val(pgprot); \
/linux-4.4.14/arch/mn10300/include/asm/
Dpage.h109 #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + __pfn_disp) macro
119 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgalloc.h28 set_pmd(pmd, __pmd((page_to_pfn(pte) << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
Dpgtable.h391 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/mm/
Dpage_isolation.c25 pfn = page_to_pfn(page); in set_migratetype_isolate()
101 page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); in unset_migratetype_isolate()
105 if (pfn_valid_within(page_to_pfn(buddy)) && in unset_migratetype_isolate()
Dmadvise.c358 page_to_pfn(p), start); in madvise_hwpoison()
365 page_to_pfn(p), start); in madvise_hwpoison()
367 memory_failure(page_to_pfn(p), 0, MF_COUNT_INCREASED); in madvise_hwpoison()
Dpage_ext.c104 unsigned long pfn = page_to_pfn(page); in lookup_page_ext()
181 unsigned long pfn = page_to_pfn(page); in lookup_page_ext()
Dmmzone.c78 if (page_to_pfn(page) != pfn) in memmap_valid_within()
Dmemory-failure.c309 page_to_pfn(p), tsk->comm); in add_to_kill()
902 page_to_pfn(page)); in get_hwpoison_page()
1589 unsigned long pfn = page_to_pfn(page); in soft_offline_huge_page()
1647 unsigned long pfn = page_to_pfn(page); in __soft_offline_page()
1744 unsigned long pfn = page_to_pfn(page); in soft_offline_page()
Dframe_vector.c177 nums[i] = page_to_pfn(pages[i]); in frame_vector_to_pfns()
Dksm.c886 flush_cache_page(vma, addr, page_to_pfn(page)); in write_protect_page()
1164 nid = get_kpfn_nid(page_to_pfn(page)); in stable_tree_search()
1267 kpfn = page_to_pfn(kpage); in stable_tree_insert()
1350 nid = get_kpfn_nid(page_to_pfn(page)); in unstable_tree_search_insert()
1987 VM_BUG_ON_PAGE(stable_node->kpfn != page_to_pfn(oldpage), oldpage); in ksm_migrate_page()
1988 stable_node->kpfn = page_to_pfn(newpage); in ksm_migrate_page()
Dpage_alloc.c353 unsigned long pfn = page_to_pfn(page); in page_outside_zone_boundaries()
374 if (!pfn_valid_within(page_to_pfn(page))) in page_is_consistent()
434 current->comm, page_to_pfn(page)); in bad_page()
603 if (!pfn_valid_within(page_to_pfn(buddy))) in page_is_buddy()
739 if ((order < MAX_ORDER-2) && pfn_valid_within(page_to_pfn(buddy))) { in __free_one_page()
846 __free_one_page(page, page_to_pfn(page), zone, 0, mt); in free_pcppages_bulk()
1014 unsigned long pfn = page_to_pfn(page); in __free_pages_ok()
1512 if (!pfn_valid_within(page_to_pfn(page))) { in move_freepages()
1538 start_pfn = page_to_pfn(page); in move_freepages_block()
2031 pfn = page_to_pfn(list_entry(curr, struct page, lru)); in mark_free_pages()
[all …]
Dsparse.c635 ret = (struct page *)pfn_to_kaddr(page_to_pfn(page)); in __kmalloc_section_memmap()
670 maps_section_nr = pfn_to_section_nr(page_to_pfn(page)); in free_map_bootmem()
Dinternal.h343 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next()
Dcma.c448 pfn = page_to_pfn(pages); in cma_release()
Dcompaction.c48 unsigned long pfn = page_to_pfn(page); in release_freepages()
275 pfn = page_to_pfn(page); in update_pageblock_skip()
Drmap.c786 if (pte_present(*pte) && page_to_pfn(page) == pte_pfn(*pte)) { in __page_check_address()
1338 flush_cache_page(vma, address, page_to_pfn(page)); in try_to_unmap_one()
Dmemory_hotplug.c1342 BUG_ON(page_to_pfn(page) & (pageblock_nr_pages - 1)); in next_active_pageblock()
1451 pfn = page_to_pfn(head) + (1<<compound_order(head)) - 1; in do_migrate_range()
Dgup.c1009 flush_cache_page(vma, addr, page_to_pfn(page)); in get_dump_page()
Dhugetlb.c1024 free_contig_range(page_to_pfn(page), 1 << order); in free_gigantic_page()
1329 compound_idx = page_to_pfn(page) - page_to_pfn(page_head); in __basepage_index()
Dzsmalloc.c811 obj = page_to_pfn(page) << OBJ_INDEX_BITS; in location_to_obj()
Dnommu.c302 return page_to_pfn(virt_to_page(addr)); in vmalloc_to_pfn()
DKconfig128 pfn_to_page and page_to_pfn operations. This is the most
Dvmalloc.c270 return page_to_pfn(vmalloc_to_page(vmalloc_addr)); in vmalloc_to_pfn()
Dmemory.c129 zero_pfn = page_to_pfn(ZERO_PAGE(0)); in init_zero_pfn()
/linux-4.4.14/arch/score/include/asm/
Dpage.h71 #define page_to_virt(page) (pfn_to_virt(page_to_pfn(page)))
73 #define page_to_phys(page) (pfn_to_phys(page_to_pfn(page)))
Dcacheflush.h45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Dpgtable.h99 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.4.14/arch/metag/include/asm/
Dpage.h109 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
111 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h154 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/m68k/include/asm/
Dpage_mm.h155 pfn_to_virt(page_to_pfn(page)); \
164 #define page_to_pfn(_page) ({ \ macro
Dcacheflush_mm.h264 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
272 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
Dvirtconvert.h33 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpage_no.h32 #define page_to_pfn(page) virt_to_pfn(page_to_virt(page)) macro
Dsun3_pgtable.h102 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dmcf_pgtable.h156 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dmotorola_pgtable.h100 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/nios2/mm/
Dcacheflush.c92 page_to_pfn(page)); in flush_aliases()
253 flush_cache_page(vma, user_vaddr, page_to_pfn(page)); in copy_from_user_page()
264 flush_cache_page(vma, user_vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.4.14/arch/parisc/include/asm/
Dcacheflush.h94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
Dpage.h167 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h396 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/tile/include/asm/
Dpgalloc.h60 set_pmd(pmd, ptfn_pmd(HV_CPA_TO_PTFN(PFN_PHYS(page_to_pfn(page))), in pmd_populate()
91 return pfn_to_kaddr(page_to_pfn(pte_alloc_one(mm, address))); in pte_alloc_one_kernel()
Dpage.h320 #define page_to_pa(page) ((phys_addr_t)(page_to_pfn(page)) << PAGE_SHIFT)
322 #define page_to_virt(page) pfn_to_kaddr(page_to_pfn(page))
Dpgtable.h303 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
482 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
Dio.h39 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.4.14/arch/xtensa/include/asm/
Dpage.h182 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
184 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h291 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.4.14/lib/
Dscatterlist.c406 if (page_to_pfn(pages[i]) != page_to_pfn(pages[i - 1]) + 1) in sg_alloc_table_from_pages()
421 if (page_to_pfn(pages[j]) != in sg_alloc_table_from_pages()
422 page_to_pfn(pages[j - 1]) + 1) in sg_alloc_table_from_pages()
Ddma-debug.c583 cln = (phys_addr_t) page_to_pfn(page) << CACHELINE_PER_PAGE_SHIFT; in debug_dma_assert_idle()
1282 entry->pfn = page_to_pfn(page); in debug_dma_map_page()
1379 entry->pfn = page_to_pfn(sg_page(s)); in debug_dma_map_sg()
1430 .pfn = page_to_pfn(sg_page(s)), in debug_dma_unmap_sg()
1466 entry->pfn = page_to_pfn(virt_to_page(virt)); in debug_dma_alloc_coherent()
1482 .pfn = page_to_pfn(virt_to_page(virt)), in debug_dma_free_coherent()
1591 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_cpu()
1624 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_device()
/linux-4.4.14/drivers/staging/android/ion/
Dion_carveout_heap.c65 ion_phys_addr_t paddr = PFN_PHYS(page_to_pfn(page)); in ion_carveout_heap_phys()
114 ion_phys_addr_t paddr = PFN_PHYS(page_to_pfn(page)); in ion_carveout_heap_free()
Dion_heap.c95 ret = remap_pfn_range(vma, addr, page_to_pfn(page), len, in ion_heap_map_user()
Dion.c960 pfn = page_to_pfn(ion_buffer_page(buffer->pages[vmf->pgoff])); in ion_vm_fault()
/linux-4.4.14/include/xen/
Dpage.h20 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
/linux-4.4.14/arch/ia64/include/asm/
Dpage.h106 # define page_to_pfn(page) ((unsigned long) (page - vmem_map)) macro
123 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h255 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/block/
Dbounce.c192 if (page_to_pfn(from.bv_page) > queue_bounce_pfn(q)) in __blk_queue_bounce()
202 if (page_to_pfn(page) <= queue_bounce_pfn(q)) in __blk_queue_bounce()
/linux-4.4.14/arch/frv/include/asm/
Dpgalloc.h26 __set_pmd((PMD), page_to_pfn(PAGE) << PAGE_SHIFT | _PAGE_TABLE); \
Dpgtable.h418 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/um/include/asm/
Dpgalloc.h18 ((unsigned long long)page_to_pfn(pte) << \
Dpgtable.h274 #define page_to_phys(page) pfn_to_phys((pfn_t) page_to_pfn(page))
/linux-4.4.14/arch/unicore32/include/asm/
Dmemory.h66 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgalloc.h108 page_to_pfn(ptep) << PAGE_SHIFT | _PAGE_USER_TABLE); in pmd_populate()
Dpgtable.h230 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.4.14/arch/sparc/mm/
Diommu.c186 ioptex = bit_map_string_get(&iommu->usemap, npages, page_to_pfn(page)); in iommu_get_one()
195 iopte_val(*iopte) = MKIOPTE(page_to_pfn(page), IOPERM); in iommu_get_one()
367 MKIOPTE(page_to_pfn(virt_to_page(page)), ioperm_noc); in iommu_map_dma_area()
Dsrmmu.c129 ptp = page_to_pfn(ptep) << (PAGE_SHIFT-4); /* watch for overflow */ in pmd_populate()
365 p = page_to_pfn(pte) << PAGE_SHIFT; /* Physical address */ in pte_free()
Dinit_64.c1684 unsigned long phys_start = page_to_pfn(page) << PAGE_SHIFT; in __kernel_map_pages()
/linux-4.4.14/kernel/power/
Dsnapshot.c892 memory_bm_set_bit(free_pages_map, page_to_pfn(page)); in swsusp_set_page_free()
898 memory_bm_test_bit(free_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_free()
904 memory_bm_clear_bit(free_pages_map, page_to_pfn(page)); in swsusp_unset_page_free()
910 memory_bm_set_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_set_page_forbidden()
916 memory_bm_test_bit(forbidden_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_forbidden()
922 memory_bm_clear_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_unset_page_forbidden()
1388 memory_bm_set_bit(&copy_bm, page_to_pfn(page)); in preallocate_image_pages()
1783 memory_bm_set_bit(bm, page_to_pfn(page)); in alloc_highmem_pages()
1826 memory_bm_set_bit(copy_bm, page_to_pfn(page)); in swsusp_alloc()
2192 memory_bm_set_bit(bm, page_to_pfn(page)); in prepare_highmem_image()
/linux-4.4.14/arch/openrisc/include/asm/
Dpgalloc.h37 ((unsigned long)page_to_pfn(pte) << in pmd_populate()
Dpage.h90 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.4.14/drivers/base/
Ddma-mapping.c253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap()
316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap()
/linux-4.4.14/drivers/misc/
Dvmw_balloon.c470 unsigned long pfn = page_to_pfn(b->page); in vmballoon_send_batched_lock()
518 unsigned long pfn = page_to_pfn(b->page); in vmballoon_send_batched_unlock()
604 locked = vmballoon_send_lock_page(b, page_to_pfn(page), &hv_status, in vmballoon_lock_page()
705 if (!vmballoon_send_unlock_page(b, page_to_pfn(page), target)) { in vmballoon_unlock_page()
789 (u64)page_to_pfn(p) << PAGE_SHIFT); in vmballoon_add_batched_page()
/linux-4.4.14/arch/mips/mm/
Dpgtable-64.c83 pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); in mk_pmd()
Ddma-default.c214 pfn = page_to_pfn(virt_to_page((void *)addr)); in mips_dma_mmap()
Dinit.c200 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.4.14/arch/arm64/include/asm/
Dmemory.h84 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgtable.h368 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
439 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot)
/linux-4.4.14/arch/powerpc/include/asm/
Dpgalloc-32.h27 (pmd_val(*(pmd)) = (page_to_pfn(pte) << PAGE_SHIFT) | _PMD_PRESENT)
Dpgtable.h78 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dio.h798 #define page_to_phys(page) ((phys_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.4.14/arch/hexagon/include/asm/
Dpage.h141 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgalloc.h103 set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) | in pmd_populate()
Dpgtable.h309 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/x86/xen/
Dgrant-table.c145 pfns[i] = page_to_pfn(pages[i]); in xlated_setup_gnttab_pages()
Dp2m.c738 pfn = page_to_pfn(pages[i]); in set_foreign_p2m_mapping()
763 unsigned long mfn = __pfn_to_mfn(page_to_pfn(pages[i])); in clear_foreign_p2m_mapping()
764 unsigned long pfn = page_to_pfn(pages[i]); in clear_foreign_p2m_mapping()
Dmmu.c753 unsigned long pfn = page_to_pfn(page); in xen_pin_page()
895 unsigned long pfn = page_to_pfn(page); in xen_unpin_page()
/linux-4.4.14/arch/tile/mm/
Dhomecache.c191 pte = pfn_pte(page_to_pfn(page), PAGE_KERNEL); in homecache_finv_map_page()
230 sim_validate_lines_evicted(PFN_PHYS(page_to_pfn(page)), PAGE_SIZE); in homecache_finv_page_internal()
/linux-4.4.14/kernel/
Dkexec_core.c356 pfn = page_to_pfn(pages); in kimage_alloc_normal_control_pages()
637 addr = page_to_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
652 if (page_to_pfn(page) > in kimage_alloc_page()
657 addr = page_to_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
736 result = kimage_add_page(image, page_to_pfn(page) in kimage_load_normal_segment()
Dkexec_file.c560 zero_buf = __va(page_to_pfn(ZERO_PAGE(0)) << PAGE_SHIFT); in kexec_calculate_store_digests()
/linux-4.4.14/arch/arm/kernel/
Dmachine_kexec.c164 page_to_pfn(image->control_code_page) << PAGE_SHIFT; in machine_kexec()
/linux-4.4.14/arch/s390/include/asm/
Dpage.h147 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h1627 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/x86/kernel/
Dmachine_kexec_64.c231 start_pgtable = page_to_pfn(image->control_code_page) << PAGE_SHIFT; in machine_kexec_prepare()
294 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec()
Dmachine_kexec_32.c227 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec()
/linux-4.4.14/arch/x86/include/asm/
Dpgalloc.h72 unsigned long pfn = page_to_pfn(pte); in pmd_populate()
Dio.h144 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h552 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
803 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/arc/include/asm/
Dio.h44 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h278 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.4.14/arch/sh/include/asm/
Dpage.h168 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable_32.h389 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/arch/m32r/include/asm/
Dio.h77 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h279 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), pgprot)
/linux-4.4.14/arch/arm/include/asm/
Dmemory.h126 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgtable-3level.h258 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
Dpgtable.h213 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.4.14/init/
Dmain.c624 page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) { in start_kernel()
626 page_to_pfn(virt_to_page((void *)initrd_start)), in start_kernel()
/linux-4.4.14/drivers/net/ethernet/tile/
Dtilegx.c1120 addr = pfn_to_kaddr(page_to_pfn(page)); in alloc_percpu_mpipe_resources()
1136 addr = pfn_to_kaddr(page_to_pfn(page)); in alloc_percpu_mpipe_resources()
1391 headers = pfn_to_kaddr(page_to_pfn(headers_page)); in tile_net_init_egress()
1403 edescs = pfn_to_kaddr(page_to_pfn(edescs_page)); in tile_net_init_egress()
1414 equeue = pfn_to_kaddr(page_to_pfn(equeue_page)); in tile_net_init_egress()
1633 unsigned long pfn = page_to_pfn(skb_frag_page(f)); in tile_net_frag_buf()
Dtilepro.c1624 unsigned long pfn = page_to_pfn(skb_frag_page(f)); in tile_net_tx_frags()
/linux-4.4.14/drivers/gpu/drm/ttm/
Dttm_memory.c564 if (glob->zone_dma32 && page_to_pfn(page) > 0x00100000UL) in ttm_mem_global_alloc_page()
579 if (glob->zone_dma32 && page_to_pfn(page) > 0x00100000UL) in ttm_mem_global_free_page()
Dttm_bo_vm.c228 pfn = page_to_pfn(page); in ttm_bo_vm_fault()
/linux-4.4.14/arch/nios2/include/asm/
Dpgtable.h241 #define mk_pte(page, prot) (pfn_pte(page_to_pfn(page), prot))
/linux-4.4.14/arch/x86/mm/
Dpageattr.c630 paravirt_alloc_pte(&init_mm, page_to_pfn(base)); in __split_large_page()
1745 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in _set_pages_array()
1771 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in _set_pages_array()
1820 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in set_pages_array_wb()
Dpgtable.c58 paravirt_release_pte(page_to_pfn(pte)); in ___pte_free_tlb()
/linux-4.4.14/arch/powerpc/mm/
Dinit_64.c181 if (pfn_valid(page_to_pfn((struct page *)start))) in vmemmap_populated()
Dmmu_context_iommu.c136 mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT; in mm_iommu_get()
Dmem.c426 __flush_dcache_icache_phys(page_to_pfn(page) << PAGE_SHIFT); in flush_dcache_icache_page()
Dpgtable_64.c782 return pfn_pmd(page_to_pfn(page), pgprot); in mk_pmd()
/linux-4.4.14/fs/proc/
Dpage.c140 } else if (is_zero_pfn(page_to_pfn(page))) in stable_page_flags()
Dvmcore.c281 buf = __va((page_to_pfn(page) << PAGE_SHIFT)); in mmap_vmcore_fault()
/linux-4.4.14/arch/sh/mm/
Dcache.c74 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.4.14/arch/powerpc/kernel/
Ddma.c188 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent()
/linux-4.4.14/arch/metag/kernel/
Ddma.c357 page_to_pfn(c->vm_pages) + off, in dma_mmap()
/linux-4.4.14/arch/mips/include/asm/
Dio.h168 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h431 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.4.14/drivers/gpu/drm/exynos/
Dexynos_drm_gem.c492 pfn = page_to_pfn(exynos_gem->pages[page_offset]); in exynos_drm_gem_fault()
/linux-4.4.14/drivers/staging/unisys/visornic/
Dvisornic_main.c243 page_to_pfn(virt_to_page(skb->data + offset)); in visor_copy_fragsinfo_from_skb()
262 count = add_physinfo_entries(page_to_pfn( in visor_copy_fragsinfo_from_skb()
430 cmdrsp->net.rcvpost.frag.pi_pfn = page_to_pfn(virt_to_page(skb->data)); in post_skb()
/linux-4.4.14/drivers/gpu/drm/
Ddrm_gem.c495 (page_to_pfn(p) >= 0x00100000UL)); in drm_gem_get_pages()
Ddrm_vm.c631 page_to_pfn(virt_to_page(map->handle)), in drm_mmap_locked()
/linux-4.4.14/drivers/virtio/
Dvirtio_balloon.c93 unsigned long pfn = page_to_pfn(page); in page_to_balloon_pfn()
/linux-4.4.14/drivers/block/
Dbrd.c392 *pfn = page_to_pfn(page); in brd_direct_access()
/linux-4.4.14/arch/tile/kernel/
Dpci-dma.c144 PFN_PHYS(page_to_pfn(page)) + offset, size); in __dma_prep_page()
/linux-4.4.14/arch/arm/common/
Ddmabounce.c321 dma_addr = pfn_to_dma(dev, page_to_pfn(page)) + offset; in dmabounce_map_page()
/linux-4.4.14/sound/pci/emu10k1/
Dmemory.c476 if (!p || (page_to_pfn(p) & ~(emu->dma_mask >> PAGE_SHIFT))) { in synth_alloc_pages()
/linux-4.4.14/arch/ia64/mm/
Dinit.c501 args->nid, args->zone, page_to_pfn(map_start), in virtual_memmap_init()
/linux-4.4.14/drivers/gpu/drm/msm/
Dmsm_gem.c220 pfn = page_to_pfn(pages[pgoff]); in msm_gem_fault()
/linux-4.4.14/drivers/block/xen-blkback/
Dblkback.c174 #define vaddr(page) ((unsigned long)pfn_to_kaddr(page_to_pfn(page)))
293 (unsigned long) pfn_to_kaddr(page_to_pfn( in free_persistent_gnts()
/linux-4.4.14/virt/kvm/
Dkvm_main.c1321 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast()
1371 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow()
1565 kvm_release_pfn_clean(page_to_pfn(page)); in kvm_release_page_clean()
1580 kvm_release_pfn_dirty(page_to_pfn(page)); in kvm_release_page_dirty()
/linux-4.4.14/drivers/lguest/
Dpage_tables.c194 return page_to_pfn(page); in get_pfn()
/linux-4.4.14/arch/parisc/mm/
Dinit.c611 set_max_mapnr(page_to_pfn(virt_to_page(high_memory - 1)) + 1); in mem_init()
/linux-4.4.14/arch/powerpc/kvm/
Dbook3s_64_mmu_hv.c67 hpt = (unsigned long)pfn_to_kaddr(page_to_pfn(page)); in kvmppc_alloc_hpt()
532 pfn = page_to_pfn(page); in kvmppc_book3s_hv_page_fault()
/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_mrmw.c1864 pgaddr = page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user1()
1903 u64 pgaddr = page_to_pfn(sg_page(*sg)) << PAGE_SHIFT; in ehca_check_kpages_per_ate()
1933 pgaddr = (page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user2()
/linux-4.4.14/drivers/vfio/
Dvfio_iommu_type1.c238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn()
/linux-4.4.14/drivers/mmc/host/
Dusdhi6rol0.c384 host->pg.page, page_to_pfn(host->pg.page), host->pg.mapped, in usdhi6_sg_map()
507 host->pg.page, page_to_pfn(host->pg.page), host->pg.mapped, in usdhi6_sg_advance()
/linux-4.4.14/drivers/hv/
Dhv_balloon.c1127 page_to_pfn(pg); in alloc_balloon_pages()
/linux-4.4.14/arch/sparc/kernel/
Dldc.c2054 unsigned long base = page_to_pfn(sg_page(sg)) << PAGE_SHIFT; in sg_count_one()
2117 fill_cookies(&state, page_to_pfn(sg_page(s)) << PAGE_SHIFT, in ldc_map_sg()
/linux-4.4.14/drivers/scsi/
Dstorvsc_drv.c1469 page_to_pfn(sg_page((cur_sgl))); in storvsc_queuecommand()
/linux-4.4.14/drivers/net/hyperv/
Dnetvsc_drv.c305 pb[j].pfn = page_to_pfn(page); in fill_pg_buf()
/linux-4.4.14/drivers/gpu/drm/omapdrm/
Domap_gem.c379 pfn = page_to_pfn(omap_obj->pages[pgoff]); in fault_1d()
/linux-4.4.14/fs/
Dexec.c252 flush_cache_page(bprm->vma, pos, page_to_pfn(page)); in flush_arg_page()
/linux-4.4.14/arch/x86/kvm/
Dsvm.c673 wrmsrl(MSR_VM_HSAVE_PA, page_to_pfn(sd->save_area) << PAGE_SHIFT); in svm_hardware_enable()
883 iopm_base = page_to_pfn(iopm_pages) << PAGE_SHIFT; in svm_hardware_setup()
1180 svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT; in svm_create_vcpu()
Dmmu.c2662 sp->role.level, gfn, page_to_pfn(pages[i]), in direct_pte_prefetch_many()
/linux-4.4.14/security/selinux/
Dselinuxfs.c262 page_to_pfn(status), in sel_mmap_handle_status()
/linux-4.4.14/drivers/net/xen-netback/
Dnetback.c113 return page_to_pfn(queue->mmap_pages[idx]); in idx_to_pfn()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dfile_ops.c844 paddr = page_to_pfn(page) << PAGE_SHIFT; in kvirt_to_phys()
/linux-4.4.14/drivers/gpu/drm/i915/
Di915_gem.c2312 if (!i || page_to_pfn(page) != last_pfn + 1) { in i915_gem_object_get_pages_gtt()
2320 last_pfn = page_to_pfn(page); in i915_gem_object_get_pages_gtt()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_file_ops.c93 paddr = page_to_pfn(page) << PAGE_SHIFT; in cvt_kvaddr()
/linux-4.4.14/drivers/net/ethernet/natsemi/
Dns83820.c1164 (long long)buf, (long) page_to_pfn(frag->page),
/linux-4.4.14/drivers/staging/rdma/ipath/
Dipath_file_ops.c89 paddr = page_to_pfn(page) << PAGE_SHIFT; in cvt_kvaddr()
/linux-4.4.14/drivers/staging/comedi/
Dcomedi_fops.c2240 page_to_pfn(virt_to_page(buf->virt_addr)), in comedi_mmap()
/linux-4.4.14/drivers/iommu/
Dintel-iommu.c167 return mm_to_dma_pfn(page_to_pfn(pg)); in page_to_dma_pfn()