Home
last modified time | relevance | path

Searched refs:page_to_pfn (Results 1 – 197 of 197) sorted by relevance

/linux-4.1.27/drivers/xen/
Dbiomerge.c9 unsigned long mfn1 = pfn_to_mfn(page_to_pfn(vec1->bv_page)); in xen_biovec_phys_mergeable()
10 unsigned long mfn2 = pfn_to_mfn(page_to_pfn(vec2->bv_page)); in xen_biovec_phys_mergeable()
Dxlate_mmu.c81 unsigned long pfn = page_to_pfn(page); in remap_pte_fn()
135 pfn = page_to_pfn(pages[i]); in xen_xlate_unmap_gfn_range()
Dtmem.c176 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_put_page()
191 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_get_page()
290 unsigned long pfn = page_to_pfn(page); in tmem_frontswap_store()
316 unsigned long pfn = page_to_pfn(page); in tmem_frontswap_load()
Dballoon.c355 frame_list[i] = page_to_pfn(page); in increase_reservation()
369 pfn = page_to_pfn(page); in increase_reservation()
429 frame_list[i] = page_to_pfn(page); in decrease_reservation()
Dgntdev.c283 pfn_to_kaddr(page_to_pfn(map->pages[i])); in map_grant_pages()
299 pfn_to_kaddr(page_to_pfn(map->pages[i])); in map_grant_pages()
339 uint8_t *tmp = pfn_to_kaddr(page_to_pfn(map->pages[pgno])); in __unmap_grant_pages()
Dgrant-table.c331 entry->ref, page_to_pfn(entry->page)); in gnttab_handle_deferred()
379 what, ref, page ? page_to_pfn(page) : -1); in gnttab_add_deferred()
Dgntalloc.c145 pfn_to_mfn(page_to_pfn(gref->page)), readonly); in add_grefs()
Dxen-scsiback.c257 unsigned long pfn = page_to_pfn(page); in vaddr_page()
/linux-4.1.27/arch/microblaze/include/asm/
Dpage.h153 # define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
154 # define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
158 # define page_to_virt(page) (pfn_to_virt(page_to_pfn(page)))
159 # define page_to_phys(page) (pfn_to_phys(page_to_pfn(page)))
/linux-4.1.27/include/trace/events/
Dkmem.h185 __entry->pfn = page_to_pfn(page);
207 __entry->pfn = page_to_pfn(page);
232 __entry->pfn = page ? page_to_pfn(page) : -1UL;
259 __entry->pfn = page ? page_to_pfn(page) : -1UL;
303 __entry->pfn = page ? page_to_pfn(page) : -1UL;
333 __entry->pfn = page_to_pfn(page);
Dpagemap.h45 __entry->pfn = page_to_pfn(page);
76 __entry->pfn = page_to_pfn(page);
Dfilemap.h28 __entry->pfn = page_to_pfn(page);
Dvmscan.h344 __entry->pfn = page_to_pfn(page);
/linux-4.1.27/arch/arm/mm/
Dflush.c128 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access()
137 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access()
222 flush_pfn_alias(page_to_pfn(page), in __flush_dcache_page()
252 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); in __flush_dcache_aliases()
384 pfn = page_to_pfn(page); in __flush_anon_page()
Dcopypage-feroceon.c77 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage()
Dcopypage-xsc3.c80 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage()
Dcopypage-v4wb.c57 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage()
Ddma-mapping.c81 return pfn_to_dma(dev, page_to_pfn(page)) + offset; in arm_dma_map_page()
88 return pfn_to_dma(dev, page_to_pfn(page)) + offset; in arm_coherent_dma_map_page()
231 phys_addr_t base = __pfn_to_phys(page_to_pfn(page)); in __dma_clear_buffer()
658 *handle = pfn_to_dma(dev, page_to_pfn(page)); in __dma_alloc()
789 pfn = page_to_pfn(page) + offset / PAGE_SIZE; in dma_cache_maint_page()
871 pfn = page_to_pfn(page) + off / PAGE_SIZE; in __dma_page_dev_to_cpu()
1250 unsigned int next_pfn = page_to_pfn(pages[i]) + 1; in __iommu_create_mapping()
1255 if (page_to_pfn(pages[j]) != next_pfn) in __iommu_create_mapping()
/linux-4.1.27/include/linux/
Dpageblock-flags.h81 get_pfnblock_flags_mask(page, page_to_pfn(page), \
85 set_pfnblock_flags_mask(page, flags, page_to_pfn(page), \
Dswapops.h108 page_to_pfn(page)); in make_migration_entry()
173 return swp_entry(SWP_HWPOISON, page_to_pfn(page)); in make_hwpoison_entry()
Dmmzone.h82 get_pfnblock_flags_mask(page, page_to_pfn(page), \
Dmm.h83 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
925 return __va(PFN_PHYS(page_to_pfn(page))); in lowmem_page_address()
/linux-4.1.27/arch/sparc/include/asm/
Dcacheflush_32.h23 flush_cache_page(vma, vaddr, page_to_pfn(page));\
28 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Dcacheflush_64.h59 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
66 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
Dpage.h4 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable_32.h285 return __pte((page_to_pfn(page) << (PAGE_SHIFT-4)) | pgprot_val(pgprot)); in mk_pte()
Dpgtable_64.h244 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
253 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.1.27/drivers/gpu/drm/gma500/
Dgtt.c109 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_insert()
114 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_insert()
143 pte = psb_gtt_mask_pte(page_to_pfn(dev_priv->scratch_page), in psb_gtt_remove()
183 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_roll()
188 pte = psb_gtt_mask_pte(page_to_pfn(r->pages[i]), in psb_gtt_roll()
548 pfn_base = page_to_pfn(dev_priv->scratch_page); in psb_gtt_init()
Dmmu.c146 PSB_WSGX32(page_to_pfn(pd->p) << PAGE_SHIFT, offset); in psb_mmu_set_pd_context()
196 pd->invalid_pde = psb_mmu_mask_pte(page_to_pfn(pd->dummy_pt), in psb_mmu_alloc_pd()
198 pd->invalid_pte = psb_mmu_mask_pte(page_to_pfn(pd->dummy_page), in psb_mmu_alloc_pd()
353 v[index] = (page_to_pfn(pt->p) << 12) | pd->pd_mask; in psb_mmu_pt_alloc_map_lock()
435 return page_to_pfn(pd->p) << PAGE_SHIFT; in psb_get_default_pd_addr()
747 pte = psb_mmu_mask_pte(page_to_pfn(*pages++), in psb_mmu_insert_pages()
Dgem.c214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault()
/linux-4.1.27/arch/avr32/include/asm/
Dpage.h82 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
89 #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + PHYS_PFN_OFFSET) macro
Dpgtable.h280 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/mn10300/include/asm/
Dpage.h109 #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + __pfn_disp) macro
119 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgalloc.h28 set_pmd(pmd, __pmd((page_to_pfn(pte) << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
Dpgtable.h391 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/include/asm-generic/
Dpage.h87 #define page_to_virt(page) pfn_to_virt(page_to_pfn(page))
90 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dmemory_model.h72 #define page_to_pfn __page_to_pfn macro
Dpgtable.h546 #define my_zero_pfn(addr) page_to_pfn(ZERO_PAGE(addr))
/linux-4.1.27/arch/alpha/include/asm/
Dmmzone.h83 pfn = page_to_pfn(page) << 32; \
101 (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h214 pte_val(pte) = (page_to_pfn(page) << 32) | pgprot_val(pgprot); \
/linux-4.1.27/arch/score/include/asm/
Dpage.h71 #define page_to_virt(page) (pfn_to_virt(page_to_pfn(page)))
73 #define page_to_phys(page) (pfn_to_phys(page_to_pfn(page)))
Dcacheflush.h45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Dpgtable.h99 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/arch/metag/include/asm/
Dpage.h109 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
111 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h154 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/mm/
Dpage_isolation.c24 pfn = page_to_pfn(page); in set_migratetype_isolate()
100 page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); in unset_migratetype_isolate()
104 if (pfn_valid_within(page_to_pfn(buddy)) && in unset_migratetype_isolate()
Dmadvise.c356 page_to_pfn(p), start); in madvise_hwpoison()
363 page_to_pfn(p), start); in madvise_hwpoison()
365 memory_failure(page_to_pfn(p), 0, MF_COUNT_INCREASED); in madvise_hwpoison()
Dpage_ext.c100 unsigned long pfn = page_to_pfn(page); in lookup_page_ext()
177 unsigned long pfn = page_to_pfn(page); in lookup_page_ext()
Dmmzone.c78 if (page_to_pfn(page) != pfn) in memmap_valid_within()
Dmm_init.c138 BUG_ON(page_to_pfn(page) != pfn); in mminit_verify_page_links()
Dksm.c885 flush_cache_page(vma, addr, page_to_pfn(page)); in write_protect_page()
1167 nid = get_kpfn_nid(page_to_pfn(page)); in stable_tree_search()
1260 kpfn = page_to_pfn(kpage); in stable_tree_insert()
1331 nid = get_kpfn_nid(page_to_pfn(page)); in unstable_tree_search_insert()
1966 VM_BUG_ON_PAGE(stable_node->kpfn != page_to_pfn(oldpage), oldpage); in ksm_migrate_page()
1967 stable_node->kpfn = page_to_pfn(newpage); in ksm_migrate_page()
Dmemory-failure.c318 page_to_pfn(p), tsk->comm); in add_to_kill()
1574 unsigned long pfn = page_to_pfn(page); in soft_offline_huge_page()
1633 unsigned long pfn = page_to_pfn(page); in __soft_offline_page()
1733 unsigned long pfn = page_to_pfn(page); in soft_offline_page()
Dpage_alloc.c253 unsigned long pfn = page_to_pfn(page); in page_outside_zone_boundaries()
274 if (!pfn_valid_within(page_to_pfn(page))) in page_is_consistent()
334 current->comm, page_to_pfn(page)); in bad_page()
520 if (!pfn_valid_within(page_to_pfn(buddy))) in page_is_buddy()
656 if ((order < MAX_ORDER-2) && pfn_valid_within(page_to_pfn(buddy))) { in __free_one_page()
760 __free_one_page(page, page_to_pfn(page), zone, 0, mt); in free_pcppages_bulk()
842 unsigned long pfn = page_to_pfn(page); in __free_pages_ok()
1112 if (!pfn_valid_within(page_to_pfn(page))) { in move_freepages()
1139 start_pfn = page_to_pfn(page); in move_freepages_block()
1550 pfn = page_to_pfn(list_entry(curr, struct page, lru)); in mark_free_pages()
[all …]
Dsparse.c635 ret = (struct page *)pfn_to_kaddr(page_to_pfn(page)); in __kmalloc_section_memmap()
670 maps_section_nr = pfn_to_section_nr(page_to_pfn(page)); in free_map_bootmem()
Dinternal.h323 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next()
Dcma.c444 pfn = page_to_pfn(pages); in cma_release()
Dcompaction.c59 unsigned long pfn = page_to_pfn(page); in release_freepages()
280 pfn = page_to_pfn(page); in update_pageblock_skip()
Drmap.c680 if (pte_present(*pte) && page_to_pfn(page) == pte_pfn(*pte)) { in __page_check_address()
1215 flush_cache_page(vma, address, page_to_pfn(page)); in try_to_unmap_one()
Dmemory_hotplug.c1303 BUG_ON(page_to_pfn(page) & (pageblock_nr_pages - 1)); in next_active_pageblock()
1405 pfn = page_to_pfn(head) + (1<<compound_order(head)) - 1; in do_migrate_range()
Dgup.c963 flush_cache_page(vma, addr, page_to_pfn(page)); in get_dump_page()
Dhugetlb.c776 free_contig_range(page_to_pfn(page), 1 << order); in free_gigantic_page()
1085 compound_idx = page_to_pfn(page) - page_to_pfn(page_head); in __basepage_index()
Dzsmalloc.c811 obj = page_to_pfn(page) << OBJ_INDEX_BITS; in location_to_obj()
Dnommu.c318 return page_to_pfn(virt_to_page(addr)); in vmalloc_to_pfn()
DKconfig128 pfn_to_page and page_to_pfn operations. This is the most
Dvmalloc.c268 return page_to_pfn(vmalloc_to_page(vmalloc_addr)); in vmalloc_to_pfn()
Dmemory.c128 zero_pfn = page_to_pfn(ZERO_PAGE(0)); in init_zero_pfn()
/linux-4.1.27/arch/m68k/include/asm/
Dpage_mm.h155 pfn_to_virt(page_to_pfn(page)); \
164 #define page_to_pfn(_page) ({ \ macro
Dcacheflush_mm.h264 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
272 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
Dvirtconvert.h33 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpage_no.h32 #define page_to_pfn(page) virt_to_pfn(page_to_virt(page)) macro
Dsun3_pgtable.h102 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dmotorola_pgtable.h100 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dmcf_pgtable.h156 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/parisc/include/asm/
Dcacheflush.h94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
Dpage.h156 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h378 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/xtensa/include/asm/
Dpage.h182 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
184 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Ddma-mapping.h77 return (dma_addr_t)(page_to_pfn(page)) * PAGE_SIZE + offset; in dma_map_page()
Dpgtable.h287 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/arch/tile/include/asm/
Dpgalloc.h60 set_pmd(pmd, ptfn_pmd(HV_CPA_TO_PTFN(PFN_PHYS(page_to_pfn(page))), in pmd_populate()
91 return pfn_to_kaddr(page_to_pfn(pte_alloc_one(mm, address))); in pte_alloc_one_kernel()
Dpage.h318 #define page_to_pa(page) ((phys_addr_t)(page_to_pfn(page)) << PAGE_SHIFT)
320 #define page_to_virt(page) pfn_to_kaddr(page_to_pfn(page))
Dpgtable.h303 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
482 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
Dio.h39 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.1.27/arch/nios2/mm/
Dcacheflush.c108 page_to_pfn(page)); in flush_aliases()
269 flush_cache_page(vma, user_vaddr, page_to_pfn(page)); in copy_from_user_page()
280 flush_cache_page(vma, user_vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.1.27/include/xen/
Dpage.h8 return pfn_to_mfn(page_to_pfn(page)); in page_to_mfn()
/linux-4.1.27/lib/
Dscatterlist.c378 if (page_to_pfn(pages[i]) != page_to_pfn(pages[i - 1]) + 1) in sg_alloc_table_from_pages()
393 if (page_to_pfn(pages[j]) != in sg_alloc_table_from_pages()
394 page_to_pfn(pages[j - 1]) + 1) in sg_alloc_table_from_pages()
Ddma-debug.c583 cln = (phys_addr_t) page_to_pfn(page) << CACHELINE_PER_PAGE_SHIFT; in debug_dma_assert_idle()
1274 entry->pfn = page_to_pfn(page); in debug_dma_map_page()
1371 entry->pfn = page_to_pfn(sg_page(s)); in debug_dma_map_sg()
1422 .pfn = page_to_pfn(sg_page(s)), in debug_dma_unmap_sg()
1458 entry->pfn = page_to_pfn(virt_to_page(virt)); in debug_dma_alloc_coherent()
1474 .pfn = page_to_pfn(virt_to_page(virt)), in debug_dma_free_coherent()
1583 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_cpu()
1616 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_device()
/linux-4.1.27/drivers/staging/android/ion/
Dion_carveout_heap.c65 ion_phys_addr_t paddr = PFN_PHYS(page_to_pfn(page)); in ion_carveout_heap_phys()
114 ion_phys_addr_t paddr = PFN_PHYS(page_to_pfn(page)); in ion_carveout_heap_free()
Dion_heap.c95 ret = remap_pfn_range(vma, addr, page_to_pfn(page), len, in ion_heap_map_user()
Dion.c958 pfn = page_to_pfn(ion_buffer_page(buffer->pages[vmf->pgoff])); in ion_vm_fault()
/linux-4.1.27/arch/ia64/include/asm/
Dpage.h106 # define page_to_pfn(page) ((unsigned long) (page - vmem_map)) macro
123 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h255 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/frv/include/asm/
Dpgalloc.h26 __set_pmd((PMD), page_to_pfn(PAGE) << PAGE_SHIFT | _PAGE_TABLE); \
Dpgtable.h418 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/um/include/asm/
Dpgalloc.h18 ((unsigned long long)page_to_pfn(pte) << \
Dpgtable.h274 #define page_to_phys(page) pfn_to_phys((pfn_t) page_to_pfn(page))
/linux-4.1.27/block/
Dbounce.c211 if (page_to_pfn(from.bv_page) > queue_bounce_pfn(q)) in __blk_queue_bounce()
221 if (page_to_pfn(page) <= queue_bounce_pfn(q) && !force) in __blk_queue_bounce()
Dblk-merge.c54 high = page_to_pfn(bv.bv_page) > queue_bounce_pfn(q); in __blk_recalc_rq_segments()
/linux-4.1.27/arch/sparc/mm/
Diommu.c186 ioptex = bit_map_string_get(&iommu->usemap, npages, page_to_pfn(page)); in iommu_get_one()
195 iopte_val(*iopte) = MKIOPTE(page_to_pfn(page), IOPERM); in iommu_get_one()
367 MKIOPTE(page_to_pfn(virt_to_page(page)), ioperm_noc); in iommu_map_dma_area()
Dsrmmu.c129 ptp = page_to_pfn(ptep) << (PAGE_SHIFT-4); /* watch for overflow */ in pmd_populate()
365 p = page_to_pfn(pte) << PAGE_SHIFT; /* Physical address */ in pte_free()
Dinit_64.c1627 unsigned long phys_start = page_to_pfn(page) << PAGE_SHIFT; in __kernel_map_pages()
/linux-4.1.27/arch/unicore32/include/asm/
Dmemory.h72 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgalloc.h108 page_to_pfn(ptep) << PAGE_SHIFT | _PAGE_USER_TABLE); in pmd_populate()
Dpgtable.h230 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/kernel/power/
Dsnapshot.c892 memory_bm_set_bit(free_pages_map, page_to_pfn(page)); in swsusp_set_page_free()
898 memory_bm_test_bit(free_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_free()
904 memory_bm_clear_bit(free_pages_map, page_to_pfn(page)); in swsusp_unset_page_free()
910 memory_bm_set_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_set_page_forbidden()
916 memory_bm_test_bit(forbidden_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_forbidden()
922 memory_bm_clear_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_unset_page_forbidden()
1388 memory_bm_set_bit(&copy_bm, page_to_pfn(page)); in preallocate_image_pages()
1783 memory_bm_set_bit(bm, page_to_pfn(page)); in alloc_highmem_pages()
1826 memory_bm_set_bit(copy_bm, page_to_pfn(page)); in swsusp_alloc()
2192 memory_bm_set_bit(bm, page_to_pfn(page)); in prepare_highmem_image()
/linux-4.1.27/arch/openrisc/include/asm/
Dpgalloc.h37 ((unsigned long)page_to_pfn(pte) << in pmd_populate()
Dpage.h90 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.1.27/arch/arc/include/asm/
Dio.h25 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
/linux-4.1.27/drivers/base/
Ddma-mapping.c253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap()
316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap()
/linux-4.1.27/arch/mips/mm/
Dpgtable-64.c83 pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); in mk_pmd()
Dinit.c196 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.1.27/arch/hexagon/include/asm/
Dpage.h141 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgalloc.h103 set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) | in pmd_populate()
Dpgtable.h309 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/powerpc/include/asm/
Dpgalloc-32.h27 (pmd_val(*(pmd)) = (page_to_pfn(pte) << PAGE_SHIFT) | _PMD_PRESENT)
Dpgtable.h78 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
Dio.h797 #define page_to_phys(page) ((phys_addr_t)page_to_pfn(page) << PAGE_SHIFT)
/linux-4.1.27/arch/arm64/include/asm/
Dmemory.h92 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgtable.h306 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
377 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot)
/linux-4.1.27/arch/x86/xen/
Dgrant-table.c145 pfns[i] = page_to_pfn(pages[i]); in xlated_setup_gnttab_pages()
Dp2m.c690 pfn = page_to_pfn(pages[i]); in set_foreign_p2m_mapping()
715 unsigned long mfn = __pfn_to_mfn(page_to_pfn(pages[i])); in clear_foreign_p2m_mapping()
716 unsigned long pfn = page_to_pfn(pages[i]); in clear_foreign_p2m_mapping()
Dmmu.c752 unsigned long pfn = page_to_pfn(page); in xen_pin_page()
894 unsigned long pfn = page_to_pfn(page); in xen_unpin_page()
/linux-4.1.27/arch/tile/mm/
Dhomecache.c191 pte = pfn_pte(page_to_pfn(page), PAGE_KERNEL); in homecache_finv_map_page()
230 sim_validate_lines_evicted(PFN_PHYS(page_to_pfn(page)), PAGE_SIZE); in homecache_finv_page_internal()
/linux-4.1.27/arch/arm/kernel/
Dmachine_kexec.c164 page_to_pfn(image->control_code_page) << PAGE_SHIFT; in machine_kexec()
/linux-4.1.27/arch/arm/include/asm/xen/
Dpage-coherent.h38 bool local = PFN_DOWN(dev_addr) == page_to_pfn(page); in xen_dma_map_page()
/linux-4.1.27/arch/s390/include/asm/
Dpage.h147 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h1552 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/m32r/include/asm/
Dio.h75 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h279 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), pgprot)
/linux-4.1.27/arch/x86/kernel/
Dmachine_kexec_64.c229 start_pgtable = page_to_pfn(image->control_code_page) << PAGE_SHIFT; in machine_kexec_prepare()
292 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec()
Dmachine_kexec_32.c227 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec()
/linux-4.1.27/arch/x86/include/asm/
Dpgalloc.h72 unsigned long pfn = page_to_pfn(pte); in pmd_populate()
Dio.h142 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h526 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
776 #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/sh/include/asm/
Dpage.h167 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
Dpgtable_32.h389 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/arch/arm/include/asm/
Dmemory.h132 #define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
Dpgtable-3level.h258 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
Dpgtable.h213 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/drivers/misc/
Dvmw_balloon.c434 locked = vmballoon_send_lock_page(b, page_to_pfn(page), &hv_status); in vmballoon_reserve_page()
471 if (!vmballoon_send_unlock_page(b, page_to_pfn(page))) in vmballoon_release_page()
/linux-4.1.27/init/
Dmain.c619 page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) { in start_kernel()
621 page_to_pfn(virt_to_page((void *)initrd_start)), in start_kernel()
/linux-4.1.27/fs/proc/
Dpage.c134 } else if (is_zero_pfn(page_to_pfn(page))) in stable_page_flags()
Dvmcore.c281 buf = __va((page_to_pfn(page) << PAGE_SHIFT)); in mmap_vmcore_fault()
/linux-4.1.27/drivers/net/ethernet/tile/
Dtilegx.c1119 addr = pfn_to_kaddr(page_to_pfn(page)); in alloc_percpu_mpipe_resources()
1135 addr = pfn_to_kaddr(page_to_pfn(page)); in alloc_percpu_mpipe_resources()
1390 headers = pfn_to_kaddr(page_to_pfn(headers_page)); in tile_net_init_egress()
1402 edescs = pfn_to_kaddr(page_to_pfn(edescs_page)); in tile_net_init_egress()
1413 equeue = pfn_to_kaddr(page_to_pfn(equeue_page)); in tile_net_init_egress()
1632 unsigned long pfn = page_to_pfn(skb_frag_page(f)); in tile_net_frag_buf()
Dtilepro.c1627 unsigned long pfn = page_to_pfn(skb_frag_page(f)); in tile_net_tx_frags()
/linux-4.1.27/drivers/gpu/drm/ttm/
Dttm_memory.c564 if (glob->zone_dma32 && page_to_pfn(page) > 0x00100000UL) in ttm_mem_global_alloc_page()
579 if (glob->zone_dma32 && page_to_pfn(page) > 0x00100000UL) in ttm_mem_global_free_page()
Dttm_bo_vm.c228 pfn = page_to_pfn(page); in ttm_bo_vm_fault()
/linux-4.1.27/kernel/
Dkexec.c713 pfn = page_to_pfn(pages); in kimage_alloc_normal_control_pages()
995 addr = page_to_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
1010 if (page_to_pfn(page) > in kimage_alloc_page()
1015 addr = page_to_pfn(page) << PAGE_SHIFT; in kimage_alloc_page()
1097 result = kimage_add_page(image, page_to_pfn(page) in kimage_load_normal_segment()
2192 zero_buf = __va(page_to_pfn(ZERO_PAGE(0)) << PAGE_SHIFT); in kexec_calculate_store_digests()
/linux-4.1.27/arch/nios2/include/asm/
Dpgtable.h241 #define mk_pte(page, prot) (pfn_pte(page_to_pfn(page), prot))
/linux-4.1.27/arch/powerpc/kernel/
Ddma.c120 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent()
Dvdso.c146 printk("kpg: %p (c:%d,f:%08lx)", __va(page_to_pfn(pg) << PAGE_SHIFT), in dump_one_vdso_page()
150 printk(" upg: %p (c:%d,f:%08lx)", __va(page_to_pfn(upg) in dump_one_vdso_page()
/linux-4.1.27/drivers/xen/xenbus/
Dxenbus_client.c623 unsigned long pfn = page_to_pfn(node->hvm.pages[i]); in xenbus_map_ring_valloc_hvm()
813 addrs[i] = (unsigned long)pfn_to_kaddr(page_to_pfn(node->hvm.pages[i])); in xenbus_unmap_ring_vfree_hvm()
/linux-4.1.27/arch/x86/mm/
Dpageattr.c613 paravirt_alloc_pte(&init_mm, page_to_pfn(base)); in __split_large_page()
1692 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in _set_pages_array()
1714 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in _set_pages_array()
1757 start = page_to_pfn(pages[i]) << PAGE_SHIFT; in set_pages_array_wb()
Dpgtable.c58 paravirt_release_pte(page_to_pfn(pte)); in ___pte_free_tlb()
/linux-4.1.27/arch/powerpc/mm/
Dinit_64.c181 if (pfn_valid(page_to_pfn((struct page *)start))) in vmemmap_populated()
Dmem.c427 __flush_dcache_icache_phys(page_to_pfn(page) << PAGE_SHIFT); in flush_dcache_icache_page()
Dpgtable_64.c797 return pfn_pmd(page_to_pfn(page), pgprot); in mk_pmd()
/linux-4.1.27/arch/metag/kernel/
Ddma.c357 page_to_pfn(c->vm_pages) + off, in dma_mmap()
/linux-4.1.27/arch/sh/mm/
Dcache.c74 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
/linux-4.1.27/arch/mips/include/asm/
Dio.h168 #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
Dpgtable.h429 #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/linux-4.1.27/drivers/scsi/
Dxen-scsifront.c423 mfn = pfn_to_mfn(page_to_pfn(page)); in map_data_for_request()
457 mfn = pfn_to_mfn(page_to_pfn(page)); in map_data_for_request()
Dstorvsc_drv.c1667 page_to_pfn(sg_page((cur_sgl))); in storvsc_queuecommand()
/linux-4.1.27/drivers/gpu/drm/
Ddrm_gem.c495 (page_to_pfn(p) >= 0x00100000UL)); in drm_gem_get_pages()
Ddrm_vm.c631 page_to_pfn(virt_to_page(map->handle)), in drm_mmap_locked()
/linux-4.1.27/drivers/block/
Dbrd.c385 *pfn = page_to_pfn(page); in brd_direct_access()
Dxen-blkfront.c209 gnt_list_entry->pfn = page_to_pfn(granted_page); in fill_grant_buffer()
517 pfn = page_to_pfn(indirect_page); in blkif_queue_request()
525 gnt_list_entry = get_grant(&gref_head, page_to_pfn(sg_page(sg)), info); in blkif_queue_request()
/linux-4.1.27/drivers/virtio/
Dvirtio_balloon.c93 unsigned long pfn = page_to_pfn(page); in page_to_balloon_pfn()
/linux-4.1.27/arch/tile/kernel/
Dpci-dma.c144 PFN_PHYS(page_to_pfn(page)) + offset, size); in __dma_prep_page()
/linux-4.1.27/arch/arm/common/
Ddmabounce.c321 dma_addr = pfn_to_dma(dev, page_to_pfn(page)) + offset; in dmabounce_map_page()
/linux-4.1.27/drivers/block/xen-blkback/
Dblkback.c167 #define vaddr(page) ((unsigned long)pfn_to_kaddr(page_to_pfn(page)))
286 (unsigned long) pfn_to_kaddr(page_to_pfn( in free_persistent_gnts()
/linux-4.1.27/arch/ia64/mm/
Dinit.c508 args->nid, args->zone, page_to_pfn(map_start), in virtual_memmap_init()
/linux-4.1.27/sound/pci/emu10k1/
Dmemory.c476 if (!p || (page_to_pfn(p) & ~(emu->dma_mask >> PAGE_SHIFT))) { in synth_alloc_pages()
/linux-4.1.27/virt/kvm/
Dkvm_main.c1235 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast()
1285 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow()
1476 kvm_release_pfn_clean(page_to_pfn(page)); in kvm_release_page_clean()
1491 kvm_release_pfn_dirty(page_to_pfn(page)); in kvm_release_page_dirty()
/linux-4.1.27/drivers/gpu/drm/msm/
Dmsm_gem.c220 pfn = page_to_pfn(pages[pgoff]); in msm_gem_fault()
/linux-4.1.27/drivers/lguest/
Dpage_tables.c194 return page_to_pfn(page); in get_pfn()
/linux-4.1.27/arch/parisc/mm/
Dinit.c601 set_max_mapnr(page_to_pfn(virt_to_page(high_memory - 1)) + 1); in mem_init()
/linux-4.1.27/arch/powerpc/kvm/
Dbook3s_64_mmu_hv.c67 hpt = (unsigned long)pfn_to_kaddr(page_to_pfn(page)); in kvmppc_alloc_hpt()
531 pfn = page_to_pfn(page); in kvmppc_book3s_hv_page_fault()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_mrmw.c1864 pgaddr = page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user1()
1903 u64 pgaddr = page_to_pfn(sg_page(*sg)) << PAGE_SHIFT; in ehca_check_kpages_per_ate()
1933 pgaddr = (page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user2()
/linux-4.1.27/drivers/vfio/
Dvfio_iommu_type1.c238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn()
/linux-4.1.27/drivers/net/hyperv/
Dnetvsc_drv.c262 pb[j].pfn = page_to_pfn(page); in fill_pg_buf()
/linux-4.1.27/drivers/mmc/host/
Dusdhi6rol0.c384 host->pg.page, page_to_pfn(host->pg.page), host->pg.mapped, in usdhi6_sg_map()
507 host->pg.page, page_to_pfn(host->pg.page), host->pg.mapped, in usdhi6_sg_advance()
/linux-4.1.27/drivers/hv/
Dhv_balloon.c1123 page_to_pfn(pg); in alloc_balloon_pages()
/linux-4.1.27/arch/sparc/kernel/
Dldc.c2054 unsigned long base = page_to_pfn(sg_page(sg)) << PAGE_SHIFT; in sg_count_one()
2116 fill_cookies(&state, page_to_pfn(sg_page(&sg[i])) << PAGE_SHIFT, in ldc_map_sg()
/linux-4.1.27/drivers/gpu/drm/omapdrm/
Domap_gem.c379 pfn = page_to_pfn(omap_obj->pages[pgoff]); in fault_1d()
/linux-4.1.27/fs/
Dexec.c246 flush_cache_page(bprm->vma, pos, page_to_pfn(page)); in flush_arg_page()
/linux-4.1.27/arch/x86/kvm/
Dsvm.c671 wrmsrl(MSR_VM_HSAVE_PA, page_to_pfn(sd->save_area) << PAGE_SHIFT); in svm_hardware_enable()
881 iopm_base = page_to_pfn(iopm_pages) << PAGE_SHIFT; in svm_hardware_setup()
1262 svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT; in svm_create_vcpu()
Dmmu.c2686 sp->role.level, gfn, page_to_pfn(pages[i]), in direct_pte_prefetch_many()
/linux-4.1.27/security/selinux/
Dselinuxfs.c262 page_to_pfn(status), in sel_mmap_handle_status()
/linux-4.1.27/drivers/net/xen-netback/
Dnetback.c113 return page_to_pfn(queue->mmap_pages[idx]); in idx_to_pfn()
/linux-4.1.27/drivers/net/
Dxen-netfront.c311 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers()
/linux-4.1.27/drivers/gpu/drm/i915/
Di915_gem.c2103 if (!i || page_to_pfn(page) != last_pfn + 1) { in i915_gem_object_get_pages_gtt()
2111 last_pfn = page_to_pfn(page); in i915_gem_object_get_pages_gtt()
/linux-4.1.27/drivers/net/ethernet/natsemi/
Dns83820.c1164 (long long)buf, (long) page_to_pfn(frag->page),
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_file_ops.c93 paddr = page_to_pfn(page) << PAGE_SHIFT; in cvt_kvaddr()
/linux-4.1.27/drivers/staging/comedi/
Dcomedi_fops.c2206 page_to_pfn(virt_to_page(buf->virt_addr)), in comedi_mmap()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_file_ops.c89 paddr = page_to_pfn(page) << PAGE_SHIFT; in cvt_kvaddr()
/linux-4.1.27/drivers/iommu/
Dintel-iommu.c162 return mm_to_dma_pfn(page_to_pfn(pg)); in page_to_dma_pfn()