/linux-4.4.14/arch/x86/xen/ |
D | p2m.c | 124 static inline unsigned p2m_top_index(unsigned long pfn) in p2m_top_index() argument 126 BUG_ON(pfn >= MAX_P2M_PFN); in p2m_top_index() 127 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); in p2m_top_index() 130 static inline unsigned p2m_mid_index(unsigned long pfn) in p2m_mid_index() argument 132 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; in p2m_mid_index() 135 static inline unsigned p2m_index(unsigned long pfn) in p2m_index() argument 137 return pfn % P2M_PER_PAGE; in p2m_index() 172 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) in p2m_init_identity() argument 177 p2m[i] = IDENTITY_FRAME(pfn + i); in p2m_init_identity() 210 unsigned long pfn, mfn; in xen_build_mfn_list_list() local [all …]
|
D | setup.c | 163 unsigned long __ref xen_chk_extra_mem(unsigned long pfn) in xen_chk_extra_mem() argument 168 if (pfn >= xen_extra_mem[i].start_pfn && in xen_chk_extra_mem() 169 pfn < xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns) in xen_chk_extra_mem() 173 return IDENTITY_FRAME(pfn); in xen_chk_extra_mem() 181 unsigned long pfn, pfn_s, pfn_e; in xen_inv_extra_mem() local 189 for (pfn = pfn_s; pfn < pfn_e; pfn++) in xen_inv_extra_mem() 190 set_phys_to_machine(pfn, INVALID_P2M_ENTRY); in xen_inv_extra_mem() 256 unsigned long pfn, end; in xen_set_identity_and_release_chunk() local 263 for (pfn = start_pfn; pfn < end; pfn++) { in xen_set_identity_and_release_chunk() 264 unsigned long mfn = pfn_to_mfn(pfn); in xen_set_identity_and_release_chunk() [all …]
|
D | mmu.c | 371 unsigned long pfn = mfn_to_pfn(mfn); in pte_mfn_to_pfn() local 374 if (unlikely(pfn == ~0)) in pte_mfn_to_pfn() 377 val = ((pteval_t)pfn << PAGE_SHIFT) | flags; in pte_mfn_to_pfn() 386 unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT; in pte_pfn_to_mfn() local 391 mfn = __pfn_to_mfn(pfn); in pte_pfn_to_mfn() 393 mfn = pfn; in pte_pfn_to_mfn() 729 static void xen_do_pin(unsigned level, unsigned long pfn) in xen_do_pin() argument 734 op.arg1.mfn = pfn_to_mfn(pfn); in xen_do_pin() 753 unsigned long pfn = page_to_pfn(page); in xen_pin_page() local 784 pfn_pte(pfn, PAGE_KERNEL_RO), in xen_pin_page() [all …]
|
D | mmu.h | 14 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 16 void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags);
|
D | enlighten.c | 490 unsigned long pfn; in set_aliased_prot() local 497 pfn = pte_pfn(*ptep); in set_aliased_prot() 498 page = pfn_to_page(pfn); in set_aliased_prot() 500 pte = pfn_pte(pfn, prot); in set_aliased_prot() 532 void *av = __va(PFN_PHYS(pfn)); in set_aliased_prot() 608 unsigned long pfn, mfn; in xen_load_gdt() local 621 pfn = pte_pfn(*ptep); in xen_load_gdt() 622 mfn = pfn_to_mfn(pfn); in xen_load_gdt() 623 virt = __va(PFN_PHYS(pfn)); in xen_load_gdt() 656 unsigned long pfn, mfn; in xen_load_gdt_boot() local [all …]
|
/linux-4.4.14/arch/x86/include/asm/xen/ |
D | page.h | 46 extern int xen_alloc_p2m_entry(unsigned long pfn); 48 extern unsigned long get_phys_to_machine(unsigned long pfn); 49 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn); 50 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 85 static inline unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 89 if (pfn < xen_p2m_size) in __pfn_to_mfn() 90 mfn = xen_p2m_addr[pfn]; in __pfn_to_mfn() 91 else if (unlikely(pfn < xen_max_p2m_pfn)) in __pfn_to_mfn() 92 return get_phys_to_machine(pfn); in __pfn_to_mfn() 94 return IDENTITY_FRAME(pfn); in __pfn_to_mfn() [all …]
|
D | interface_32.h | 99 #define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20)) argument
|
/linux-4.4.14/mm/ |
D | page_isolation.c | 16 unsigned long flags, pfn; in set_migratetype_isolate() local 25 pfn = page_to_pfn(page); in set_migratetype_isolate() 26 arg.start_pfn = pfn; in set_migratetype_isolate() 133 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument 137 if (pfn_valid_within(pfn + i)) in __first_valid_page() 141 return pfn_to_page(pfn + i); in __first_valid_page() 161 unsigned long pfn; in start_isolate_page_range() local 168 for (pfn = start_pfn; in start_isolate_page_range() 169 pfn < end_pfn; in start_isolate_page_range() 170 pfn += pageblock_nr_pages) { in start_isolate_page_range() [all …]
|
D | page_ext.c | 104 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local 119 offset = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext() 181 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local 182 struct mem_section *section = __pfn_to_section(pfn); in lookup_page_ext() 193 return section->page_ext + pfn; in lookup_page_ext() 215 static int __meminit init_section_page_ext(unsigned long pfn, int nid) in init_section_page_ext() argument 221 section = __pfn_to_section(pfn); in init_section_page_ext() 245 pfn &= PAGE_SECTION_MASK; in init_section_page_ext() 246 section->page_ext = base - pfn; in init_section_page_ext() 266 static void __free_page_ext(unsigned long pfn) in __free_page_ext() argument [all …]
|
D | page_owner.c | 87 print_page_owner(char __user *buf, size_t count, unsigned long pfn, in print_page_owner() argument 110 pageblock_mt = get_pfnblock_migratetype(page, pfn); in print_page_owner() 114 pfn, in print_page_owner() 115 pfn >> pageblock_order, in print_page_owner() 156 unsigned long pfn; in read_page_owner() local 164 pfn = min_low_pfn + *ppos; in read_page_owner() 167 while (!pfn_valid(pfn) && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0) in read_page_owner() 168 pfn++; in read_page_owner() 173 for (; pfn < max_pfn; pfn++) { in read_page_owner() 178 if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0 && !pfn_valid(pfn)) { in read_page_owner() [all …]
|
D | memory-failure.c | 182 unsigned long pfn, struct page *page, int flags) in kill_proc() argument 189 pfn, t->comm, t->pid); in kill_proc() 326 int fail, struct page *page, unsigned long pfn, in kill_procs() argument 341 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 352 pfn, page, flags) < 0) in kill_procs() 355 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 556 static int me_kernel(struct page *p, unsigned long pfn) in me_kernel() argument 564 static int me_unknown(struct page *p, unsigned long pfn) in me_unknown() argument 566 printk(KERN_ERR "MCE %#lx: Unknown page state\n", pfn); in me_unknown() 573 static int me_pagecache_clean(struct page *p, unsigned long pfn) in me_pagecache_clean() argument [all …]
|
D | memory_hotplug.c | 247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 271 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 282 if (pfn_valid(pfn) && (pfn_to_nid(pfn) == node)) in register_page_bootmem_info_node() 283 register_page_bootmem_info_section(pfn); in register_page_bootmem_info_node() 330 unsigned long pfn; in fix_zone_id() local 332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id() 333 set_page_links(pfn_to_page(pfn), zid, nid, pfn); in fix_zone_id() 449 unsigned long flags, pfn; in __add_zone() local 466 for (pfn = phys_start_pfn; pfn < phys_start_pfn + nr_pages; pfn++) { in __add_zone() [all …]
|
D | page_idle.c | 30 static struct page *page_idle_get_page(unsigned long pfn) in page_idle_get_page() argument 35 if (!pfn_valid(pfn)) in page_idle_get_page() 38 page = pfn_to_page(pfn); in page_idle_get_page() 121 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 127 pfn = pos * BITS_PER_BYTE; in page_idle_bitmap_read() 128 if (pfn >= max_pfn) in page_idle_bitmap_read() 131 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 135 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 136 bit = pfn % BITMAP_CHUNK_BITS; in page_idle_bitmap_read() 139 page = page_idle_get_page(pfn); in page_idle_bitmap_read() [all …]
|
D | cma.c | 84 static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, in cma_clear_bitmap() argument 89 bitmap_no = (pfn - cma->base_pfn) >> cma->order_per_bit; in cma_clear_bitmap() 100 unsigned long base_pfn = cma->base_pfn, pfn = base_pfn; in cma_activate_area() local 109 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area() 110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area() 115 base_pfn = pfn; in cma_activate_area() 116 for (j = pageblock_nr_pages; j; --j, pfn++) { in cma_activate_area() 117 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area() 124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area() 367 unsigned long pfn = -1; in cma_alloc() local [all …]
|
D | compaction.c | 48 unsigned long pfn = page_to_pfn(page); in release_freepages() local 51 if (pfn > high_pfn) in release_freepages() 52 high_pfn = pfn; in release_freepages() 215 unsigned long pfn; in __reset_isolation_suitable() local 220 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable() 225 if (!pfn_valid(pfn)) in __reset_isolation_suitable() 228 page = pfn_to_page(pfn); in __reset_isolation_suitable() 262 unsigned long pfn; in update_pageblock_skip() local 275 pfn = page_to_pfn(page); in update_pageblock_skip() 279 if (pfn > zone->compact_cached_migrate_pfn[0]) in update_pageblock_skip() [all …]
|
D | page_alloc.c | 276 static inline bool __meminit early_page_uninitialised(unsigned long pfn) in early_page_uninitialised() argument 278 if (pfn >= NODE_DATA(early_pfn_to_nid(pfn))->first_deferred_pfn) in early_page_uninitialised() 284 static inline bool early_page_nid_uninitialised(unsigned long pfn, int nid) in early_page_nid_uninitialised() argument 286 if (pfn >= NODE_DATA(nid)->first_deferred_pfn) in early_page_nid_uninitialised() 297 unsigned long pfn, unsigned long zone_end, in update_defer_init() argument 307 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in update_defer_init() 308 pgdat->first_deferred_pfn = pfn; in update_defer_init() 319 static inline bool early_page_uninitialised(unsigned long pfn) in early_page_uninitialised() argument 324 static inline bool early_page_nid_uninitialised(unsigned long pfn, int nid) in early_page_nid_uninitialised() argument 330 unsigned long pfn, unsigned long zone_end, in update_defer_init() argument [all …]
|
D | hwpoison-inject.c | 15 unsigned long pfn = val; in hwpoison_inject() local 23 if (!pfn_valid(pfn)) in hwpoison_inject() 26 p = pfn_to_page(pfn); in hwpoison_inject() 55 pr_info("Injecting memory failure at pfn %#lx\n", pfn); in hwpoison_inject() 56 return memory_failure(pfn, 18, MF_COUNT_INCREASED); in hwpoison_inject()
|
D | vmstat.c | 1002 unsigned long pfn; in pagetypeinfo_showblockcount_print() local 1007 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print() 1010 if (!pfn_valid(pfn)) in pagetypeinfo_showblockcount_print() 1013 page = pfn_to_page(pfn); in pagetypeinfo_showblockcount_print() 1016 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print() 1054 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print() local 1055 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() 1061 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print() 1068 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 1069 if (!pfn_valid(pfn)) { in pagetypeinfo_showmixedcount_print() [all …]
|
D | sparse.c | 172 unsigned long pfn; in memory_present() local 176 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 177 unsigned long section = pfn_to_section_nr(pfn); in memory_present() 197 unsigned long pfn; in node_memmap_size_bytes() local 201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes() 202 if (nid != early_pfn_to_nid(pfn)) in node_memmap_size_bytes() 205 if (pfn_present(pfn)) in node_memmap_size_bytes()
|
D | memory.c | 753 unsigned long pfn = pte_pfn(pte); in vm_normal_page() local 762 if (!is_zero_pfn(pfn)) in vm_normal_page() 771 if (!pfn_valid(pfn)) in vm_normal_page() 777 if (pfn == vma->vm_pgoff + off) in vm_normal_page() 784 if (is_zero_pfn(pfn)) in vm_normal_page() 787 if (unlikely(pfn > highest_memmap_pfn)) { in vm_normal_page() 797 return pfn_to_page(pfn); in vm_normal_page() 804 unsigned long pfn = pmd_pfn(pmd); in vm_normal_page_pmd() local 813 if (!pfn_valid(pfn)) in vm_normal_page_pmd() 819 if (pfn == vma->vm_pgoff + off) in vm_normal_page_pmd() [all …]
|
/linux-4.4.14/arch/arm/xen/ |
D | p2m.c | 21 unsigned long pfn; member 42 if (new->pfn == entry->pfn) in xen_add_phys_to_mach_entry() 45 if (new->pfn < entry->pfn) in xen_add_phys_to_mach_entry() 57 __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); in xen_add_phys_to_mach_entry() 62 unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 71 if (entry->pfn <= pfn && in __pfn_to_mfn() 72 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn() 74 return entry->mfn + (pfn - entry->pfn); in __pfn_to_mfn() 76 if (pfn < entry->pfn) in __pfn_to_mfn() 119 bool __set_phys_to_machine_multi(unsigned long pfn, in __set_phys_to_machine_multi() argument [all …]
|
/linux-4.4.14/include/asm-generic/ |
D | memory_model.h | 15 #define arch_pfn_to_nid(pfn) pfn_to_nid(pfn) argument 19 #define arch_local_page_offset(pfn, nid) \ argument 20 ((pfn) - NODE_DATA(nid)->node_start_pfn) 30 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) argument 35 #define __pfn_to_page(pfn) \ argument 36 ({ unsigned long __pfn = (pfn); \ 51 #define __pfn_to_page(pfn) (vmemmap + (pfn)) argument 65 #define __pfn_to_page(pfn) \ argument 66 ({ unsigned long __pfn = (pfn); \ 76 #define __pfn_to_phys(pfn) PFN_PHYS(pfn) argument
|
D | page.h | 84 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 93 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr) argument
|
D | pgtable.h | 546 unsigned long pfn, unsigned long addr, in track_pfn_remap() argument 557 unsigned long pfn) in track_pfn_insert() argument 577 unsigned long pfn, unsigned long size) in untrack_pfn() argument 582 unsigned long pfn, unsigned long addr, 585 unsigned long pfn); 587 extern void untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, 592 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument 595 unsigned long offset_from_zero_pfn = pfn - zero_pfn; in is_zero_pfn() 602 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument 605 return pfn == zero_pfn; in is_zero_pfn()
|
/linux-4.4.14/include/trace/events/ |
D | cma.h | 12 TP_PROTO(unsigned long pfn, const struct page *page, 15 TP_ARGS(pfn, page, count, align), 18 __field(unsigned long, pfn) 25 __entry->pfn = pfn; 32 __entry->pfn, 40 TP_PROTO(unsigned long pfn, const struct page *page, 43 TP_ARGS(pfn, page, count), 46 __field(unsigned long, pfn) 52 __entry->pfn = pfn; 58 __entry->pfn,
|
D | kmem.h | 180 __field( unsigned long, pfn ) 185 __entry->pfn = page_to_pfn(page); 190 pfn_to_page(__entry->pfn), 191 __entry->pfn, 202 __field( unsigned long, pfn ) 207 __entry->pfn = page_to_pfn(page); 212 pfn_to_page(__entry->pfn), 213 __entry->pfn, 225 __field( unsigned long, pfn ) 232 __entry->pfn = page ? page_to_pfn(page) : -1UL; [all …]
|
D | pagemap.h | 38 __field(unsigned long, pfn ) 45 __entry->pfn = page_to_pfn(page); 53 __entry->pfn, 71 __field(unsigned long, pfn ) 76 __entry->pfn = page_to_pfn(page); 80 TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
|
D | filemap.h | 21 __field(unsigned long, pfn) 28 __entry->pfn = page_to_pfn(page); 40 pfn_to_page(__entry->pfn), 41 __entry->pfn,
|
D | xen.h | 328 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned), 329 TP_ARGS(mm, pfn, level, pinned), 332 __field(unsigned long, pfn) 337 __entry->pfn = pfn; 341 __entry->mm, __entry->pfn, __entry->level, 346 TP_PROTO(unsigned long pfn, unsigned level, bool pinned), 347 TP_ARGS(pfn, level, pinned), 349 __field(unsigned long, pfn) 353 TP_fast_assign(__entry->pfn = pfn; 357 __entry->pfn, __entry->level,
|
/linux-4.4.14/arch/m32r/include/asm/ |
D | mmzone.h | 16 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument 27 #define pfn_valid(pfn) (1) argument 29 #define pfn_valid(pfn) ((pfn) < num_physpages) argument 36 static __inline__ int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 41 if (pfn >= node_start_pfn(node) && pfn < node_end_pfn(node)) in pfn_to_nid() 47 static __inline__ struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 49 return(NODE_DATA(pfn_to_nid(pfn))); in pfn_to_pgdat()
|
D | pgtable-2level.h | 71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument 72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
|
D | cacheflush.h | 14 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument 36 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument 50 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
|
/linux-4.4.14/arch/parisc/include/asm/ |
D | mmzone.h | 33 #define pfn_is_io(pfn) ((pfn & (0xf0000000UL >> PAGE_SHIFT)) == (0xf0000000UL >> PAGE_SHIFT)) argument 36 #define pfn_is_io(pfn) ((pfn & (0xf000000000000000UL >> PAGE_SHIFT)) == (0xf000000000000000UL >> PA… argument 39 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 43 if (unlikely(pfn_is_io(pfn))) in pfn_to_nid() 46 i = pfn >> PFNNID_SHIFT; in pfn_to_nid() 52 static inline int pfn_valid(int pfn) in pfn_valid() argument 54 int nid = pfn_to_nid(pfn); in pfn_valid() 57 return (pfn < node_end_pfn(nid)); in pfn_valid()
|
/linux-4.4.14/arch/arm/mach-gemini/ |
D | mm.c | 23 .pfn =__phys_to_pfn(GEMINI_GLOBAL_BASE), 28 .pfn = __phys_to_pfn(GEMINI_UART_BASE), 33 .pfn = __phys_to_pfn(GEMINI_TIMER_BASE), 38 .pfn = __phys_to_pfn(GEMINI_INTERRUPT_BASE), 43 .pfn = __phys_to_pfn(GEMINI_POWER_CTRL_BASE), 48 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(0)), 53 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(1)), 58 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(2)), 63 .pfn = __phys_to_pfn(GEMINI_FLASH_CTRL_BASE), 68 .pfn = __phys_to_pfn(GEMINI_DRAM_CTRL_BASE), [all …]
|
/linux-4.4.14/arch/arm/include/asm/xen/ |
D | page.h | 14 #define phys_to_machine_mapping_valid(pfn) (1) argument 42 unsigned long __pfn_to_mfn(unsigned long pfn); 46 static inline unsigned long pfn_to_gfn(unsigned long pfn) in pfn_to_gfn() argument 48 return pfn; in pfn_to_gfn() 57 static inline unsigned long pfn_to_bfn(unsigned long pfn) in pfn_to_bfn() argument 62 mfn = __pfn_to_mfn(pfn); in pfn_to_bfn() 67 return pfn; in pfn_to_bfn() 105 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 106 bool __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn, 109 static inline bool set_phys_to_machine(unsigned long pfn, unsigned long mfn) in set_phys_to_machine() argument [all …]
|
D | page-coherent.h | 64 unsigned long pfn = PFN_DOWN(handle); in xen_dma_unmap_page() local 73 if (pfn_valid(pfn)) { in xen_dma_unmap_page() 83 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_cpu() local 84 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_cpu() 94 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_device() local 95 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_device()
|
/linux-4.4.14/arch/s390/kernel/ |
D | suspend.c | 99 void page_key_read(unsigned long *pfn) in page_key_read() argument 103 addr = (unsigned long) page_address(pfn_to_page(*pfn)); in page_key_read() 104 *(unsigned char *) pfn = (unsigned char) page_get_storage_key(addr); in page_key_read() 111 void page_key_memorize(unsigned long *pfn) in page_key_memorize() argument 113 page_key_wp->data[page_key_wx] = *(unsigned char *) pfn; in page_key_memorize() 114 *(unsigned char *) pfn = 0; in page_key_memorize() 137 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 145 if (pfn <= LC_PAGES) in pfn_is_nosave() 147 if (pfn >= nosave_begin_pfn && pfn < nosave_end_pfn) in pfn_is_nosave() 150 if (pfn >= stext_pfn && pfn <= eshared_pfn) in pfn_is_nosave() [all …]
|
D | crash_dump.c | 132 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, size_t csize, in copy_oldmem_page() argument 139 src = (pfn << PAGE_SHIFT) + offset; in copy_oldmem_page() 153 unsigned long from, unsigned long pfn, in remap_oldmem_pfn_range_kdump() argument 159 if (pfn < OLDMEM_SIZE >> PAGE_SHIFT) { in remap_oldmem_pfn_range_kdump() 160 size_old = min(size, OLDMEM_SIZE - (pfn << PAGE_SHIFT)); in remap_oldmem_pfn_range_kdump() 162 pfn + (OLDMEM_BASE >> PAGE_SHIFT), in remap_oldmem_pfn_range_kdump() 168 pfn += size_old >> PAGE_SHIFT; in remap_oldmem_pfn_range_kdump() 170 return remap_pfn_range(vma, from, pfn, size, prot); in remap_oldmem_pfn_range_kdump() 181 unsigned long pfn, in remap_oldmem_pfn_range_zfcpdump() argument 187 if (pfn < hsa_end >> PAGE_SHIFT) { in remap_oldmem_pfn_range_zfcpdump() [all …]
|
/linux-4.4.14/arch/arm/mach-omap2/ |
D | io.c | 74 .pfn = __phys_to_pfn(L3_24XX_PHYS), 80 .pfn = __phys_to_pfn(L4_24XX_PHYS), 90 .pfn = __phys_to_pfn(DSP_MEM_2420_PHYS), 96 .pfn = __phys_to_pfn(DSP_IPI_2420_PHYS), 102 .pfn = __phys_to_pfn(DSP_MMU_2420_PHYS), 114 .pfn = __phys_to_pfn(L4_WK_243X_PHYS), 120 .pfn = __phys_to_pfn(OMAP243X_GPMC_PHYS), 126 .pfn = __phys_to_pfn(OMAP243X_SDRC_PHYS), 132 .pfn = __phys_to_pfn(OMAP243X_SMS_PHYS), 144 .pfn = __phys_to_pfn(L3_34XX_PHYS), [all …]
|
/linux-4.4.14/arch/x86/power/ |
D | hibernate_32.c | 83 unsigned long pfn; in resume_physical_mapping_init() local 91 pfn = 0; in resume_physical_mapping_init() 98 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 102 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 110 set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC)); in resume_physical_mapping_init() 111 pfn += PTRS_PER_PTE; in resume_physical_mapping_init() 120 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init() 121 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 124 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init() 167 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument [all …]
|
D | hibernate_64.c | 103 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 107 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/alpha/include/asm/ |
D | mmzone.h | 34 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument 81 unsigned long pfn; \ 83 pfn = page_to_pfn(page) << 32; \ 84 pte_val(pte) = pfn | pgprot_val(pgprot); \ 103 #define pfn_to_nid(pfn) pa_to_nid(((u64)(pfn) << PAGE_SHIFT)) argument 104 #define pfn_valid(pfn) \ argument 105 (((pfn) - node_start_pfn(pfn_to_nid(pfn))) < \ 106 node_spanned_pages(pfn_to_nid(pfn))) \
|
D | pgtable.h | 194 #define PHYS_TWIDDLE(pfn) \ argument 195 ((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \ 196 ? ((pfn) ^= KSEG_PFN) : (pfn)) 198 #define PHYS_TWIDDLE(pfn) (pfn) argument
|
/linux-4.4.14/arch/unicore32/mm/ |
D | ioremap.c | 102 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 118 set_pmd(pmd, __pmd(__pfn_to_phys(pfn) | type->prot_sect)); in remap_area_sections() 119 pfn += SZ_4M >> PAGE_SHIFT; in remap_area_sections() 129 void __iomem *__uc32_ioremap_pfn_caller(unsigned long pfn, in __uc32_ioremap_pfn_caller() argument 140 if (pfn >= 0x100000 && (__pfn_to_phys(pfn) & ~SECTION_MASK)) in __uc32_ioremap_pfn_caller() 146 if (pfn_valid(pfn)) { in __uc32_ioremap_pfn_caller() 168 if (!((__pfn_to_phys(pfn) | size | addr) & ~PMD_MASK)) { in __uc32_ioremap_pfn_caller() 170 err = remap_area_sections(addr, pfn, size, type); in __uc32_ioremap_pfn_caller() 172 err = ioremap_page_range(addr, addr + size, __pfn_to_phys(pfn), in __uc32_ioremap_pfn_caller() 189 unsigned long pfn = __phys_to_pfn(phys_addr); in __uc32_ioremap_caller() local [all …]
|
D | mmu.c | 81 unsigned long pfn; member 164 unsigned long end, unsigned long pfn, in alloc_init_pte() argument 169 set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); in alloc_init_pte() 170 pfn++; in alloc_init_pte() 217 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping() 225 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping() 231 phys = (unsigned long)__pfn_to_phys(md->pfn); in create_mapping() 237 __pfn_to_phys(md->pfn), addr); in create_mapping() 368 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init() 378 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init() [all …]
|
/linux-4.4.14/arch/unicore32/kernel/ |
D | hibernate.c | 75 unsigned long pfn; in resume_physical_mapping_init() local 83 pfn = 0; in resume_physical_mapping_init() 90 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 96 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 107 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init() 108 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 111 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init() 145 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/x86/include/asm/ |
D | mmzone_32.h | 33 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid() 42 static inline int pfn_valid(int pfn) in pfn_valid() argument 44 int nid = pfn_to_nid(pfn); in pfn_valid() 47 return (pfn < node_end_pfn(nid)); in pfn_valid() 51 #define early_pfn_valid(pfn) pfn_valid((pfn)) argument
|
D | pgalloc.h | 15 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pte() argument 16 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pmd() argument 17 static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, in paravirt_alloc_pmd_clone() argument 19 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pud() argument 20 static inline void paravirt_release_pte(unsigned long pfn) {} in paravirt_release_pte() argument 21 static inline void paravirt_release_pmd(unsigned long pfn) {} in paravirt_release_pmd() argument 22 static inline void paravirt_release_pud(unsigned long pfn) {} in paravirt_release_pud() argument 72 unsigned long pfn = page_to_pfn(pte); in pmd_populate() local 74 paravirt_alloc_pte(mm, pfn); in pmd_populate() 75 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
|
D | page_64.h | 34 #define pfn_valid(pfn) ((pfn) < max_pfn) argument
|
D | page_32.h | 18 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
D | highmem.h | 69 void *kmap_atomic_pfn(unsigned long pfn); 70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
|
/linux-4.4.14/kernel/power/ |
D | snapshot.c | 634 static int memory_bm_find_bit(struct memory_bitmap *bm, unsigned long pfn, in memory_bm_find_bit() argument 643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 666 if (((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) in memory_bm_find_bit() 670 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; in memory_bm_find_bit() 685 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; in memory_bm_find_bit() 689 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; in memory_bm_find_bit() 694 static void memory_bm_set_bit(struct memory_bitmap *bm, unsigned long pfn) in memory_bm_set_bit() argument 700 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit() 705 static int mem_bm_set_bit_check(struct memory_bitmap *bm, unsigned long pfn) in mem_bm_set_bit_check() argument [all …]
|
/linux-4.4.14/arch/arm/mm/ |
D | fault-armv.c | 41 unsigned long pfn, pte_t *ptep) in do_adjust_pte() argument 56 flush_cache_page(vma, address, pfn); in do_adjust_pte() 57 outer_flush_range((pfn << PAGE_SHIFT), in do_adjust_pte() 58 (pfn << PAGE_SHIFT) + PAGE_SIZE); in do_adjust_pte() 93 unsigned long pfn) in adjust_pte() argument 123 ret = do_adjust_pte(vma, address, pfn, pte); in adjust_pte() 133 unsigned long addr, pte_t *ptep, unsigned long pfn) in make_coherent() argument 160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); in make_coherent() 164 do_adjust_pte(vma, addr, pfn, ptep); in make_coherent() 183 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local [all …]
|
D | flush.c | 41 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument 46 set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL)); in flush_pfn_alias() 55 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument 61 set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL)); in flush_icache_alias() 101 void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) in flush_cache_page() argument 104 vivt_flush_cache_page(vma, user_addr, pfn); in flush_cache_page() 109 flush_pfn_alias(pfn, user_addr); in flush_cache_page() 118 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument 119 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument 275 unsigned long pfn; in __sync_icache_dcache() local [all …]
|
D | ioremap.c | 186 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 204 pmd[0] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections() 205 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections() 206 pmd[1] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections() 207 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections() 218 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument 238 super_pmd_val = __pfn_to_phys(pfn) | type->prot_sect | in remap_area_supersections() 240 super_pmd_val |= ((pfn >> (32 - PAGE_SHIFT)) & 0xf) << 20; in remap_area_supersections() 251 pfn += SUPERSECTION_SIZE >> PAGE_SHIFT; in remap_area_supersections() 258 static void __iomem * __arm_ioremap_pfn_caller(unsigned long pfn, in __arm_ioremap_pfn_caller() argument [all …]
|
D | mmu.c | 701 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 704 if (!pfn_valid(pfn)) in phys_mem_access_prot() 738 unsigned long end, unsigned long pfn, in alloc_init_pte() argument 743 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0); in alloc_init_pte() 744 pfn++; in alloc_init_pte() 829 phys = __pfn_to_phys(md->pfn); in create_36bit_mapping() 834 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() 846 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() 850 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { in create_36bit_mapping() 852 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() [all …]
|
D | mmap.c | 215 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 217 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range() 231 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 233 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 235 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
D | highmem.c | 129 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument 133 struct page *page = pfn_to_page(pfn); in kmap_atomic_pfn() 146 set_fixmap_pte(idx, pfn_pte(pfn, kmap_prot)); in kmap_atomic_pfn()
|
/linux-4.4.14/include/linux/ |
D | mmzone.h | 85 static inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn) in get_pfnblock_migratetype() argument 88 return get_pfnblock_flags_mask(page, pfn, PB_migrate_end, in get_pfnblock_migratetype() 554 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 556 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 986 static inline unsigned long early_pfn_to_nid(unsigned long pfn) in early_pfn_to_nid() argument 993 #define pfn_to_nid(pfn) (0) argument 1019 #define pfn_to_section_nr(pfn) ((pfn) >> PFN_SECTION_SHIFT) argument 1022 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK) argument 1023 #define SECTION_ALIGN_DOWN(pfn) ((pfn) & PAGE_SECTION_MASK) argument 1121 static inline struct mem_section *__pfn_to_section(unsigned long pfn) in __pfn_to_section() argument [all …]
|
D | io-mapping.h | 88 unsigned long pfn; in io_mapping_map_atomic_wc() local 92 pfn = (unsigned long) (phys_addr >> PAGE_SHIFT); in io_mapping_map_atomic_wc() 93 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc()
|
D | crash_dump.h | 22 unsigned long from, unsigned long pfn, 83 extern int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn));
|
D | pageblock-flags.h | 69 unsigned long pfn, 75 unsigned long pfn,
|
D | suspend.h | 516 void page_key_read(unsigned long *pfn); 517 void page_key_memorize(unsigned long *pfn); 533 static inline void page_key_read(unsigned long *pfn) {} in page_key_read() argument 534 static inline void page_key_memorize(unsigned long *pfn) {} in page_key_memorize() argument
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | page.h | 98 extern int ia64_pfn_valid (unsigned long pfn); 100 # define ia64_pfn_valid(pfn) 1 argument 107 # define pfn_to_page(pfn) (vmem_map + (pfn)) argument 116 # define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn)) argument 120 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn)) argument 125 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
D | mmzone.h | 20 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 23 int nid = paddr_to_nid(pfn << PAGE_SHIFT); in pfn_to_nid()
|
/linux-4.4.14/arch/x86/kvm/ |
D | iommu.c | 50 pfn_t pfn; in kvm_pin_pages() local 52 pfn = gfn_to_pfn_memslot(slot, gfn); in kvm_pin_pages() 56 if (is_error_noslot_pfn(pfn)) in kvm_pin_pages() 57 return pfn; in kvm_pin_pages() 62 return pfn; in kvm_pin_pages() 65 static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages) in kvm_unpin_pages() argument 70 kvm_release_pfn_clean(pfn + i); in kvm_unpin_pages() 76 pfn_t pfn; in kvm_iommu_map_pages() local 123 pfn = kvm_pin_pages(slot, gfn, page_size >> PAGE_SHIFT); in kvm_iommu_map_pages() 124 if (is_error_noslot_pfn(pfn)) { in kvm_iommu_map_pages() [all …]
|
D | mmu.c | 262 pfn_t pfn, unsigned access) in set_mmio_spte() argument 264 if (unlikely(is_noslot_pfn(pfn))) { in set_mmio_spte() 590 pfn_t pfn; in mmu_spte_clear_track_bits() local 601 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits() 608 WARN_ON(!kvm_is_reserved_pfn(pfn) && !page_count(pfn_to_page(pfn))); in mmu_spte_clear_track_bits() 611 kvm_set_pfn_accessed(pfn); in mmu_spte_clear_track_bits() 613 kvm_set_pfn_dirty(pfn); in mmu_spte_clear_track_bits() 2479 static bool kvm_is_mmio_pfn(pfn_t pfn) in kvm_is_mmio_pfn() argument 2481 if (pfn_valid(pfn)) in kvm_is_mmio_pfn() 2482 return !is_zero_pfn(pfn) && PageReserved(pfn_to_page(pfn)); in kvm_is_mmio_pfn() [all …]
|
D | mmu_audit.c | 100 pfn_t pfn; in audit_mappings() local 117 pfn = kvm_vcpu_gfn_to_pfn_atomic(vcpu, gfn); in audit_mappings() 119 if (is_error_pfn(pfn)) in audit_mappings() 122 hpa = pfn << PAGE_SHIFT; in audit_mappings() 125 "ent %llxn", vcpu->arch.mmu.root_level, pfn, in audit_mappings()
|
D | paging_tmpl.h | 459 pfn_t pfn; in FNAME() local 469 pfn = pte_prefetch_gfn_to_pfn(vcpu, gfn, in FNAME() 471 if (is_error_pfn(pfn)) in FNAME() 479 gfn, pfn, true, true); in FNAME() 554 pfn_t pfn, bool map_writable, bool prefault) in FNAME() 626 it.level, gw->gfn, pfn, prefault, map_writable); in FNAME() 634 kvm_release_pfn_clean(pfn); in FNAME() 699 pfn_t pfn; in FNAME() local 757 if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, in FNAME() 762 walker.gfn, pfn, walker.pte_access, &r)) in FNAME() [all …]
|
/linux-4.4.14/arch/tile/mm/ |
D | init.c | 93 pte_t *get_prealloc_pte(unsigned long pfn) in get_prealloc_pte() argument 95 int node = pfn_to_nid(pfn); in get_prealloc_pte() 96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)); in get_prealloc_pte() 98 BUG_ON(pfn >= num_l2_ptes[node]); in get_prealloc_pte() 99 return &l2_ptes[node][pfn]; in get_prealloc_pte() 408 unsigned long address, pfn; in kernel_physical_mapping_init() local 453 pfn = node_start_pfn[i]; in kernel_physical_mapping_init() 456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init() 458 address = (unsigned long) pfn_to_kaddr(pfn); in kernel_physical_mapping_init() 459 while (pfn < end_pfn) { in kernel_physical_mapping_init() [all …]
|
D | highmem.c | 270 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument 272 return kmap_atomic(pfn_to_page(pfn)); in kmap_atomic_pfn() 274 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument 276 return kmap_atomic_prot(pfn_to_page(pfn), prot); in kmap_atomic_prot_pfn()
|
/linux-4.4.14/arch/x86/mm/ |
D | init_32.c | 261 unsigned long pfn; in kernel_physical_mapping_init() local 292 pfn = start_pfn; in kernel_physical_mapping_init() 293 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init() 298 if (pfn >= end_pfn) in kernel_physical_mapping_init() 301 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init() 306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 308 unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; in kernel_physical_mapping_init() 325 pfn &= PMD_MASK >> PAGE_SHIFT; in kernel_physical_mapping_init() 326 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init() 335 set_pmd(pmd, pfn_pmd(pfn, init_prot)); in kernel_physical_mapping_init() [all …]
|
D | pat.c | 423 u64 pfn; in reserve_ram_pages_type() local 437 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 440 page = pfn_to_page(pfn); in reserve_ram_pages_type() 455 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 456 page = pfn_to_page(pfn); in reserve_ram_pages_type() 465 u64 pfn; in free_ram_pages_type() local 467 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type() 468 page = pfn_to_page(pfn); in free_ram_pages_type() 689 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 697 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument [all …]
|
D | numa_32.c | 49 unsigned long pfn; in memory_present() local 57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present() 59 printk(KERN_CONT "%lx ", pfn); in memory_present()
|
D | init.c | 85 unsigned long pfn; in alloc_low_pages() local 106 pfn = ret >> PAGE_SHIFT; in alloc_low_pages() 108 pfn = pgt_buf_end; in alloc_low_pages() 111 pfn << PAGE_SHIFT, (pgt_buf_end << PAGE_SHIFT) - 1); in alloc_low_pages() 117 adr = __va((pfn + i) << PAGE_SHIFT); in alloc_low_pages() 121 return __va(pfn << PAGE_SHIFT); in alloc_low_pages() 269 unsigned long pfn; in split_mem_range() local 275 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 283 if (pfn == 0) in split_mem_range() 286 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() [all …]
|
D | pageattr.c | 38 unsigned long pfn; member 252 unsigned long pfn) in static_protections() argument 261 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) in static_protections() 277 if (within(pfn, __pa_symbol(__start_rodata) >> PAGE_SHIFT, in static_protections() 477 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn, old_pfn; in try_preserve_large_page() local 556 pfn = old_pfn + ((address & (psize - 1)) >> PAGE_SHIFT); in try_preserve_large_page() 557 cpa->pfn = pfn; in try_preserve_large_page() 559 new_prot = static_protections(req_prot, address, pfn); in try_preserve_large_page() 567 pfn = old_pfn; in try_preserve_large_page() 568 for (i = 0; i < (psize >> PAGE_SHIFT); i++, addr += PAGE_SIZE, pfn++) { in try_preserve_large_page() [all …]
|
D | iomap_32.c | 57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument 68 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); in kmap_atomic_prot_pfn() 78 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in iomap_atomic_prot_pfn() argument 91 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); in iomap_atomic_prot_pfn()
|
D | pageattr-test.c | 136 unsigned long pfn = prandom_u32() % max_pfn_mapped; in pageattr_test() local 138 addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT); in pageattr_test() 140 len[i] = min_t(unsigned long, len[i], max_pfn_mapped - pfn - 1); in pageattr_test() 164 if (test_bit(pfn + k, bm)) { in pageattr_test() 168 __set_bit(pfn + k, bm); in pageattr_test()
|
/linux-4.4.14/fs/proc/ |
D | page.c | 34 unsigned long pfn; in kpagecount_read() local 38 pfn = src / KPMSIZE; in kpagecount_read() 44 if (pfn_valid(pfn)) in kpagecount_read() 45 ppage = pfn_to_page(pfn); in kpagecount_read() 58 pfn++; in kpagecount_read() 202 unsigned long pfn; in kpageflags_read() local 205 pfn = src / KPMSIZE; in kpageflags_read() 211 if (pfn_valid(pfn)) in kpageflags_read() 212 ppage = pfn_to_page(pfn); in kpageflags_read() 221 pfn++; in kpageflags_read() [all …]
|
D | vmcore.c | 51 static int (*oldmem_pfn_is_ram)(unsigned long pfn); 53 int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn)) in register_oldmem_pfn_is_ram() argument 69 static int pfn_is_ram(unsigned long pfn) in pfn_is_ram() argument 71 int (*fn)(unsigned long pfn); in pfn_is_ram() 82 ret = fn(pfn); in pfn_is_ram() 91 unsigned long pfn, offset; in read_from_oldmem() local 99 pfn = (unsigned long)(*ppos / PAGE_SIZE); in read_from_oldmem() 108 if (pfn_is_ram(pfn) == 0) in read_from_oldmem() 111 tmp = copy_oldmem_page(pfn, buf, nr_bytes, in read_from_oldmem() 120 ++pfn; in read_from_oldmem() [all …]
|
/linux-4.4.14/arch/frv/include/asm/ |
D | page.h | 43 #define devmem_is_allowed(pfn) 1 argument 48 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 55 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument 58 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
|
/linux-4.4.14/arch/x86/kernel/ |
D | crash_dump_32.c | 17 static inline bool is_crashed_pfn_valid(unsigned long pfn) in is_crashed_pfn_valid() argument 27 return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn; in is_crashed_pfn_valid() 50 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 58 if (!is_crashed_pfn_valid(pfn)) in copy_oldmem_page() 61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
|
D | crash_dump_64.c | 26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.4.14/arch/powerpc/kvm/ |
D | e500_mmu_host.c | 166 pfn_t pfn; in kvmppc_map_magic() local 168 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic() 169 get_page(pfn_to_page(pfn)); in kvmppc_map_magic() 177 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_map_magic() 249 pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() argument 251 ref->pfn = pfn; in kvmppc_e500_ref_setup() 258 kvm_set_pfn_accessed(pfn); in kvmppc_e500_ref_setup() 261 kvm_set_pfn_dirty(pfn); in kvmppc_e500_ref_setup() 268 trace_kvm_booke206_ref_release(ref->pfn, ref->flags); in kvmppc_e500_ref_release() 312 pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe() local [all …]
|
D | trace_booke.h | 137 TP_PROTO(__u64 pfn, __u32 flags), 138 TP_ARGS(pfn, flags), 141 __field( __u64, pfn ) 146 __entry->pfn = pfn; 151 __entry->pfn, __entry->flags)
|
D | book3s_64_mmu_host.c | 101 unsigned long pfn; in kvmppc_mmu_map_page() local 108 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page() 109 if (is_error_noslot_pfn(pfn)) { in kvmppc_mmu_map_page() 115 hpaddr = pfn << PAGE_SHIFT; in kvmppc_mmu_map_page() 135 kvm_set_pfn_accessed(pfn); in kvmppc_mmu_map_page() 140 kvm_set_pfn_dirty(pfn); in kvmppc_mmu_map_page() 146 kvmppc_mmu_flush_icache(pfn); in kvmppc_mmu_map_page() 200 cpte->pfn = pfn; in kvmppc_mmu_map_page() 209 kvm_release_pfn_clean(pfn); in kvmppc_mmu_map_page()
|
D | trace_pr.h | 70 __field( u64, pfn ) 79 __entry->pfn = pte->pfn; 89 __entry->host_vpn, __entry->pfn, __entry->eaddr, 99 __field( u64, pfn ) 108 __entry->pfn = pte->pfn; 118 __entry->host_vpn, __entry->pfn, __entry->eaddr,
|
/linux-4.4.14/arch/s390/pci/ |
D | pci_mmio.c | 15 unsigned long *pfn) in get_pfn() argument 28 ret = follow_pfn(vma, user_addr, pfn); in get_pfn() 40 unsigned long pfn; in SYSCALL_DEFINE3() local 55 ret = get_pfn(mmio_addr, VM_WRITE, &pfn); in SYSCALL_DEFINE3() 58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3() 80 unsigned long pfn; in SYSCALL_DEFINE3() local 95 ret = get_pfn(mmio_addr, VM_READ, &pfn); in SYSCALL_DEFINE3() 98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
|
/linux-4.4.14/arch/tile/include/asm/ |
D | page.h | 155 #define __pfn_to_highbits(pfn) ((pfn) >> (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)) argument 255 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument 257 return vbase_map[__pfn_to_highbits(pfn)] + (pfn << PAGE_SHIFT); in pfn_to_kaddr() 262 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys() local 263 return ((phys_addr_t)pfn << PAGE_SHIFT) + in virt_to_phys() 286 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument 288 return (void *)((pfn << PAGE_SHIFT) + PAGE_OFFSET); in pfn_to_kaddr() 313 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 315 return pfn < max_mapnr; in pfn_valid()
|
D | mmzone.h | 38 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 40 return highbits_to_node[__pfn_to_highbits(pfn)]; in pfn_to_nid() 45 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 47 int nid = pfn_to_nid(pfn); in pfn_valid() 50 return (pfn < node_end_pfn(nid)); in pfn_valid()
|
D | highmem.h | 64 void *kmap_atomic_pfn(unsigned long pfn); 65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
|
/linux-4.4.14/arch/microblaze/include/asm/ |
D | page.h | 142 extern int page_is_ram(unsigned long pfn); 145 # define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument 148 # define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument 165 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && \ argument 166 ((pfn) <= (min_low_pfn + max_mapnr))) 170 # define pfn_valid(pfn) ((pfn) < (max_mapnr + ARCH_PFN_OFFSET)) argument
|
D | cacheflush.h | 92 #define flush_cache_page(vma, vmaddr, pfn) \ argument 93 flush_dcache_range(pfn << PAGE_SHIFT, (pfn << PAGE_SHIFT) + PAGE_SIZE);
|
/linux-4.4.14/arch/arm/include/asm/ |
D | dma-mapping.h | 61 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument 64 pfn -= dev->dma_pfn_offset; in pfn_to_dma() 65 return (dma_addr_t)__pfn_to_bus(pfn); in pfn_to_dma() 70 unsigned long pfn = __bus_to_pfn(addr); in dma_to_pfn() local 73 pfn += dev->dma_pfn_offset; in dma_to_pfn() 75 return pfn; in dma_to_pfn() 81 unsigned long pfn = dma_to_pfn(dev, addr); in dma_to_virt() local 83 return phys_to_virt(__pfn_to_phys(pfn)); in dma_to_virt() 98 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument 100 return __arch_pfn_to_dma(dev, pfn); in pfn_to_dma()
|
D | kvm_mmu.h | 185 static inline void __coherent_cache_guest_page(struct kvm_vcpu *vcpu, pfn_t pfn, in __coherent_cache_guest_page() argument 215 void *va = kmap_atomic_pfn(pfn); in __coherent_cache_guest_page() 225 pfn++; in __coherent_cache_guest_page() 249 pfn_t pfn = pmd_pfn(pmd); in __kvm_flush_dcache_pmd() local 252 void *va = kmap_atomic_pfn(pfn); in __kvm_flush_dcache_pmd() 256 pfn++; in __kvm_flush_dcache_pmd()
|
/linux-4.4.14/arch/metag/include/asm/ |
D | mmzone.h | 10 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 15 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid() 21 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 23 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
|
D | page.h | 115 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_pfn) argument 118 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.4.14/arch/m68k/include/asm/ |
D | page_no.h | 26 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 31 #define pfn_to_page(pfn) virt_to_page(pfn_to_virt(pfn)) argument 33 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
D | page_mm.h | 127 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 158 #define pfn_to_page(pfn) ({ \ argument 159 unsigned long __pfn = (pfn); \ 172 #define pfn_valid(pfn) virt_addr_valid(pfn_to_virt(pfn)) argument
|
/linux-4.4.14/arch/sh/include/asm/ |
D | mmzone.h | 12 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 17 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid() 23 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 25 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
|
D | page.h | 167 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 180 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
|
/linux-4.4.14/arch/powerpc/mm/ |
D | mem.c | 79 int page_is_ram(unsigned long pfn) in page_is_ram() argument 82 return pfn < max_pfn; in page_is_ram() 84 unsigned long paddr = (pfn << PAGE_SHIFT); in page_is_ram() 94 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 98 return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); in phys_mem_access_prot() 100 if (!page_is_ram(pfn)) in phys_mem_access_prot() 341 unsigned long pfn, highmem_mapnr; in mem_init() local 344 for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { in mem_init() 345 phys_addr_t paddr = (phys_addr_t)pfn << PAGE_SHIFT; in mem_init() 346 struct page *page = pfn_to_page(pfn); in mem_init() [all …]
|
/linux-4.4.14/drivers/base/ |
D | node.c | 362 static int __init_refok get_nid_for_pfn(unsigned long pfn) in get_nid_for_pfn() argument 366 if (!pfn_valid_within(pfn)) in get_nid_for_pfn() 370 return early_pfn_to_nid(pfn); in get_nid_for_pfn() 372 page = pfn_to_page(pfn); in get_nid_for_pfn() 375 return pfn_to_nid(pfn); in get_nid_for_pfn() 382 unsigned long pfn, sect_start_pfn, sect_end_pfn; in register_mem_sect_under_node() local 392 for (pfn = sect_start_pfn; pfn <= sect_end_pfn; pfn++) { in register_mem_sect_under_node() 399 if (!pfn_present(pfn)) { in register_mem_sect_under_node() 400 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node() 405 page_nid = get_nid_for_pfn(pfn); in register_mem_sect_under_node() [all …]
|
D | memory.c | 127 unsigned long i, pfn; in show_mem_removable() local 134 pfn = section_nr_to_pfn(mem->start_section_nr + i); in show_mem_removable() 135 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable() 192 unsigned long pfn = start_pfn; in pages_correctly_reserved() local 199 for (i = 0; i < sections_per_block; i++, pfn += PAGES_PER_SECTION) { in pages_correctly_reserved() 200 if (WARN_ON_ONCE(!pfn_valid(pfn))) in pages_correctly_reserved() 202 page = pfn_to_page(pfn); in pages_correctly_reserved() 210 pfn_to_section_nr(pfn), j); in pages_correctly_reserved() 494 u64 pfn; in store_soft_offline_page() local 497 if (kstrtoull(buf, 0, &pfn) < 0) in store_soft_offline_page() [all …]
|
D | dma-mapping.c | 253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap() local 263 pfn + off, in dma_common_mmap() 310 unsigned long pfn; in dma_common_contiguous_remap() local 316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap() 317 pages[i] = pfn_to_page(pfn + i); in dma_common_contiguous_remap()
|
/linux-4.4.14/arch/score/include/asm/ |
D | page.h | 65 #define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument 68 #define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument 77 #define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn)) argument
|
/linux-4.4.14/arch/arm/mach-omap1/ |
D | io.c | 34 .pfn = __phys_to_pfn(OMAP1_IO_PHYS), 44 .pfn = __phys_to_pfn(OMAP7XX_DSP_START), 49 .pfn = __phys_to_pfn(OMAP7XX_DSPREG_START), 60 .pfn = __phys_to_pfn(OMAP1510_DSP_START), 65 .pfn = __phys_to_pfn(OMAP1510_DSPREG_START), 76 .pfn = __phys_to_pfn(OMAP16XX_DSP_START), 81 .pfn = __phys_to_pfn(OMAP16XX_DSPREG_START),
|
/linux-4.4.14/drivers/xen/ |
D | balloon.c | 276 unsigned long pfn = res->start >> PAGE_SHIFT; in additional_memory_resource() local 278 if (pfn > limit) { in additional_memory_resource() 280 pfn, limit); in additional_memory_resource() 345 unsigned long pfn, i; in reserve_additional_memory() local 347 pfn = PFN_DOWN(resource->start); in reserve_additional_memory() 349 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) { in reserve_additional_memory() 458 unsigned long pfn = page_to_pfn(page); in increase_reservation() local 460 set_phys_to_machine(pfn, frame_list[i]); in increase_reservation() 466 (unsigned long)__va(pfn << PAGE_SHIFT), in increase_reservation() 536 unsigned long pfn = page_to_pfn(page); in decrease_reservation() local [all …]
|
D | grant-table.c | 394 int gnttab_grant_foreign_transfer(domid_t domid, unsigned long pfn) in gnttab_grant_foreign_transfer() argument 401 gnttab_grant_foreign_transfer_ref(ref, domid, pfn); in gnttab_grant_foreign_transfer() 408 unsigned long pfn) in gnttab_grant_foreign_transfer_ref() argument 410 gnttab_interface->update_entry(ref, domid, pfn, GTF_accept_transfer); in gnttab_grant_foreign_transfer_ref() 637 xen_pfn_t *pfn; in gnttab_setup_auto_xlat_frames() local 651 pfn = kcalloc(max_nr_gframes, sizeof(pfn[0]), GFP_KERNEL); in gnttab_setup_auto_xlat_frames() 652 if (!pfn) { in gnttab_setup_auto_xlat_frames() 657 pfn[i] = XEN_PFN_DOWN(addr) + i; in gnttab_setup_auto_xlat_frames() 660 xen_auto_xlat_grant_frames.pfn = pfn; in gnttab_setup_auto_xlat_frames() 671 kfree(xen_auto_xlat_grant_frames.pfn); in gnttab_free_auto_xlat_frames() [all …]
|
/linux-4.4.14/arch/parisc/kernel/ |
D | cache.c | 75 #define pfn_va(pfn) __va(PFN_PHYS(pfn)) argument 80 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local 86 if (!pfn_valid(pfn)) in update_mmu_cache() 89 page = pfn_to_page(pfn); in update_mmu_cache() 91 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache() 94 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache() 542 unsigned long pfn; in flush_cache_mm() local 546 pfn = pte_pfn(*ptep); in flush_cache_mm() 547 if (!pfn_valid(pfn)) in flush_cache_mm() 549 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm() [all …]
|
/linux-4.4.14/arch/arm64/mm/ |
D | mmap.c | 114 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 116 return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK); in valid_mmap_phys_addr_range() 129 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 131 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 133 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
/linux-4.4.14/drivers/vfio/ |
D | vfio_iommu_type1.c | 191 static bool is_invalid_reserved_pfn(unsigned long pfn) in is_invalid_reserved_pfn() argument 193 if (pfn_valid(pfn)) { in is_invalid_reserved_pfn() 195 struct page *tail = pfn_to_page(pfn); in is_invalid_reserved_pfn() 219 static int put_pfn(unsigned long pfn, int prot) in put_pfn() argument 221 if (!is_invalid_reserved_pfn(pfn)) { in put_pfn() 222 struct page *page = pfn_to_page(pfn); in put_pfn() 231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument 238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn() 247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn() 248 if (is_invalid_reserved_pfn(*pfn)) in vaddr_get_pfn() [all …]
|
/linux-4.4.14/arch/mn10300/include/asm/ |
D | page.h | 107 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 108 #define pfn_to_page(pfn) (mem_map + ((pfn) - __pfn_disp)) argument 111 #define pfn_valid(pfn) \ argument 113 unsigned long __pfn = (pfn) - __pfn_disp; \
|
D | pgtable.h | 317 #define __pfn_addr(pfn) ((pfn) << PAGE_SHIFT) argument 318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot)) argument 319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot)) argument 480 #define MK_IOSPACE_PFN(space, pfn) (pfn) argument 481 #define GET_IOSPACE(pfn) 0 argument 482 #define GET_PFN(pfn) (pfn) argument
|
/linux-4.4.14/arch/arm/mach-footbridge/ |
D | common.c | 143 .pfn = __phys_to_pfn(DC21285_ARMCSR_BASE), 157 .pfn = __phys_to_pfn(DC21285_PCI_MEM), 162 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_0_CONFIG), 167 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_1_CONFIG), 172 .pfn = __phys_to_pfn(DC21285_PCI_IACK), 256 unsigned long __pfn_to_bus(unsigned long pfn) in __pfn_to_bus() argument 258 return __pfn_to_phys(pfn) + (fb_bus_sdram_offset() - PHYS_OFFSET); in __pfn_to_bus()
|
/linux-4.4.14/arch/arm/mach-exynos/ |
D | exynos.c | 41 .pfn = __phys_to_pfn(EXYNOS4_PA_SROMC), 46 .pfn = __phys_to_pfn(EXYNOS4_PA_CMU), 51 .pfn = __phys_to_pfn(EXYNOS4_PA_COREPERI), 56 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC0), 61 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC1), 70 .pfn = __phys_to_pfn(EXYNOS5_PA_SROMC), 75 .pfn = __phys_to_pfn(EXYNOS5_PA_CMU), 135 iodesc.pfn = __phys_to_pfn(be32_to_cpu(reg[0])); in exynos_fdt_map_chipid()
|
/linux-4.4.14/arch/frv/mm/ |
D | init.c | 123 unsigned long pfn; in mem_init() local 125 for (pfn = get_num_physpages() - 1; in mem_init() 126 pfn >= num_mappedpages; pfn--) in mem_init() 127 free_highmem_page(&mem_map[pfn]); in mem_init()
|
/linux-4.4.14/arch/arm/mach-tegra/ |
D | io.c | 36 .pfn = __phys_to_pfn(IO_PPSB_PHYS), 42 .pfn = __phys_to_pfn(IO_APB_PHYS), 48 .pfn = __phys_to_pfn(IO_CPU_PHYS), 54 .pfn = __phys_to_pfn(IO_IRAM_PHYS),
|
/linux-4.4.14/arch/arm/kvm/ |
D | mmu.c | 101 static bool kvm_is_device_pfn(unsigned long pfn) in kvm_is_device_pfn() argument 103 return !pfn_valid(pfn); in kvm_is_device_pfn() 452 unsigned long end, unsigned long pfn, in create_hyp_pte_mappings() argument 461 kvm_set_pte(pte, pfn_pte(pfn, prot)); in create_hyp_pte_mappings() 464 pfn++; in create_hyp_pte_mappings() 469 unsigned long end, unsigned long pfn, in create_hyp_pmd_mappings() argument 495 create_hyp_pte_mappings(pmd, addr, next, pfn, prot); in create_hyp_pmd_mappings() 496 pfn += (next - addr) >> PAGE_SHIFT; in create_hyp_pmd_mappings() 503 unsigned long end, unsigned long pfn, in create_hyp_pud_mappings() argument 527 ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); in create_hyp_pud_mappings() [all …]
|
/linux-4.4.14/arch/avr32/include/asm/ |
D | pgtable-2level.h | 42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument 43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
|
D | page.h | 88 #define pfn_to_page(pfn) (mem_map + ((pfn) - PHYS_PFN_OFFSET)) argument 90 #define pfn_valid(pfn) ((pfn) >= PHYS_PFN_OFFSET && (pfn) < (PHYS_PFN_OFFSET + max_mapnr)) argument
|
/linux-4.4.14/arch/mips/include/asm/ |
D | page.h | 195 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 199 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 205 return pfn >= pfn_offset && pfn < max_mapnr; in pfn_valid() 214 #define pfn_valid(pfn) \ argument 216 unsigned long __pfn = (pfn); \
|
D | pgtable-32.h | 110 pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument 114 pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | in pfn_pte() 116 pte.pte_high = (pfn << _PFN_SHIFT) | in pfn_pte() 127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument 130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
|
D | pgtable-64.h | 229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument 232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument 233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
|
D | mmzone.h | 13 #define pfn_to_nid(pfn) pa_to_nid((pfn) << PAGE_SHIFT) argument
|
/linux-4.4.14/arch/microblaze/mm/ |
D | init.c | 76 unsigned long pfn; in highmem_setup() local 78 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { in highmem_setup() 79 struct page *page = pfn_to_page(pfn); in highmem_setup() 82 if (!memblock_is_reserved(pfn << PAGE_SHIFT)) in highmem_setup() 267 int page_is_ram(unsigned long pfn) in page_is_ram() argument 269 return __range_ok(pfn, 0); in page_is_ram() 272 int page_is_ram(unsigned long pfn) in page_is_ram() argument 274 return pfn < max_low_pfn; in page_is_ram()
|
D | consistent.c | 204 unsigned long pfn; in consistent_free() local 207 pfn = pte_pfn(*ptep); in consistent_free() 209 if (pfn_valid(pfn)) { in consistent_free() 210 page = pfn_to_page(pfn); in consistent_free()
|
/linux-4.4.14/arch/arm/mach-realview/ |
D | realview_pbx.c | 53 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 58 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_CPU_BASE), 63 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_DIST_BASE), 68 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 73 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER0_1_BASE), 78 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER2_3_BASE), 85 .pfn = __phys_to_pfn(REALVIEW_PBX_PCI_BASE), 93 .pfn = __phys_to_pfn(REALVIEW_PBX_UART0_BASE), 103 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_SCU_BASE), 108 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_GIC_DIST_BASE), [all …]
|
D | realview_pb1176.c | 56 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 61 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_CPU_BASE), 66 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_DIST_BASE), 71 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_CPU_BASE), 76 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_DIST_BASE), 81 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 86 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER0_1_BASE), 91 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER2_3_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_PB1176_L220_BASE), 103 .pfn = __phys_to_pfn(REALVIEW_PB1176_UART0_BASE),
|
D | realview_pba8.c | 51 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 56 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_CPU_BASE), 61 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_DIST_BASE), 66 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 71 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER0_1_BASE), 76 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER2_3_BASE), 83 .pfn = __phys_to_pfn(REALVIEW_PBA8_PCI_BASE), 91 .pfn = __phys_to_pfn(REALVIEW_PBA8_UART0_BASE),
|
D | realview_pb11mp.c | 54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 59 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_CPU_BASE), 64 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_DIST_BASE), 69 .pfn = __phys_to_pfn(REALVIEW_TC11MP_PRIV_MEM_BASE), 74 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 79 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER0_1_BASE), 84 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER2_3_BASE), 89 .pfn = __phys_to_pfn(REALVIEW_TC11MP_L220_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_PB11MP_UART0_BASE),
|
D | realview_eb.c | 54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 59 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_CPU_BASE), 64 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_DIST_BASE), 69 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 74 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER0_1_BASE), 79 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER2_3_BASE), 86 .pfn = __phys_to_pfn(REALVIEW_EB_UART0_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_EB11MP_PRIV_MEM_BASE), 101 .pfn = __phys_to_pfn(REALVIEW_EB11MP_L220_BASE),
|
/linux-4.4.14/arch/um/include/asm/ |
D | pgtable-2level.h | 41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) argument 42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot)) argument
|
D | page.h | 113 #define pfn_to_phys(pfn) ((phys_t) ((pfn) << PAGE_SHIFT)) argument 115 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
/linux-4.4.14/arch/nios2/include/asm/ |
D | page.h | 89 # define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 90 # define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && \ argument 91 (pfn) < max_mapnr)
|
/linux-4.4.14/arch/powerpc/kernel/ |
D | dma.c | 32 u64 pfn = (dev->coherent_dma_mask >> PAGE_SHIFT) + 1; in get_pfn_limit() local 37 pfn = min_t(u64, pfn, sd->max_direct_dma_addr >> PAGE_SHIFT); in get_pfn_limit() 40 return pfn; in get_pfn_limit() 80 u64 pfn = get_pfn_limit(dev); in __dma_direct_alloc_coherent() local 89 zone = dma_pfn_limit_to_zone(pfn); in __dma_direct_alloc_coherent() 92 __func__, pfn); in __dma_direct_alloc_coherent() 182 unsigned long pfn; in dma_direct_mmap_coherent() local 186 pfn = __dma_get_coherent_pfn((unsigned long)cpu_addr); in dma_direct_mmap_coherent() 188 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent() 191 pfn + vma->vm_pgoff, in dma_direct_mmap_coherent()
|
D | suspend.c | 18 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 22 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/sparc/include/asm/ |
D | pgtable_32.h | 262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) argument 405 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument 406 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument 407 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) argument 413 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument 418 offset = ((unsigned long long) GET_PFN(pfn)) << PAGE_SHIFT; in io_remap_pfn_range() 419 space = GET_IOSPACE(pfn); in io_remap_pfn_range()
|
D | pgtable_64.h | 237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument 239 unsigned long paddr = pfn << PAGE_SHIFT; in pfn_pte() 817 unsigned long pfn; in __pmd_page() local 819 pfn = pte_pfn(pte); in __pmd_page() 821 return ((unsigned long) __va(pfn << PAGE_SHIFT)); in __pmd_page() 987 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument 988 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument 989 #define GET_PFN(pfn) (pfn & 0x0fffffffffffffffUL) argument 995 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument 998 unsigned long offset = GET_PFN(pfn) << PAGE_SHIFT; in io_remap_pfn_range() [all …]
|
D | page_32.h | 132 #define pfn_valid(pfn) (((pfn) >= (pfn_base)) && (((pfn)-(pfn_base)) < max_mapnr)) argument
|
/linux-4.4.14/arch/arm/mach-cns3xxx/ |
D | core.c | 32 .pfn = __phys_to_pfn(CNS3XXX_TC11MP_SCU_BASE), 37 .pfn = __phys_to_pfn(CNS3XXX_TIMER1_2_3_BASE), 42 .pfn = __phys_to_pfn(CNS3XXX_MISC_BASE), 47 .pfn = __phys_to_pfn(CNS3XXX_PM_BASE), 53 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_HOST_BASE), 58 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG0_BASE), 63 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG1_BASE), 68 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_HOST_BASE), 73 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG0_BASE), 78 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG1_BASE),
|
/linux-4.4.14/arch/arm/mach-s3c64xx/ |
D | common.c | 108 .pfn = __phys_to_pfn(S3C64XX_PA_SYSCON), 113 .pfn = __phys_to_pfn(S3C64XX_PA_SROM), 118 .pfn = __phys_to_pfn(S3C_PA_UART), 123 .pfn = __phys_to_pfn(S3C64XX_PA_VIC0), 128 .pfn = __phys_to_pfn(S3C64XX_PA_VIC1), 133 .pfn = __phys_to_pfn(S3C_PA_TIMER), 138 .pfn = __phys_to_pfn(S3C64XX_PA_GPIO), 143 .pfn = __phys_to_pfn(S3C64XX_PA_MODEM), 148 .pfn = __phys_to_pfn(S3C64XX_PA_WATCHDOG), 153 .pfn = __phys_to_pfn(S3C64XX_PA_USB_HSPHY),
|
/linux-4.4.14/arch/arm/mach-integrator/ |
D | integrator_cp.c | 69 .pfn = __phys_to_pfn(INTEGRATOR_HDR_BASE), 74 .pfn = __phys_to_pfn(INTEGRATOR_CT_BASE), 79 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE), 84 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE), 89 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE), 94 .pfn = __phys_to_pfn(INTEGRATOR_CP_GPIO_BASE), 99 .pfn = __phys_to_pfn(INTEGRATOR_CP_SIC_BASE),
|
D | integrator_ap.c | 85 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE), 90 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE), 95 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE), 100 .pfn = __phys_to_pfn(INTEGRATOR_AP_GPIO_BASE),
|
/linux-4.4.14/arch/sh/kernel/ |
D | swsusp.c | 22 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | crash_dump.c | 25 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.4.14/arch/sparc/power/ |
D | hibernate.c | 21 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 26 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/arm/mach-spear/ |
D | spear13xx.c | 63 .pfn = __phys_to_pfn(PERIP_GRP2_BASE), 68 .pfn = __phys_to_pfn(PERIP_GRP1_BASE), 73 .pfn = __phys_to_pfn(A9SM_AND_MPMC_BASE), 78 .pfn = __phys_to_pfn(L2CC_BASE),
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | page.h | 129 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && (pfn) < max_mapnr) argument 134 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 411 extern int page_is_ram(unsigned long pfn); 412 extern int devmem_is_allowed(unsigned long pfn);
|
D | pte-hash64-64k.h | 97 #define remap_4k_pfn(vma, addr, pfn, prot) \ argument 98 (WARN_ON(((pfn) >= (1UL << (64 - PTE_RPN_SHIFT)))) ? -EINVAL : \ 99 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, \
|
/linux-4.4.14/arch/mips/power/ |
D | cpu.c | 37 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 42 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/arm/mach-s3c24xx/ |
D | mach-smdk2443.c | 53 .pfn = __phys_to_pfn(S3C2410_CS2), 58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 63 .pfn = __phys_to_pfn(S3C2410_CS2), 68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
D | mach-smdk2440.c | 53 .pfn = __phys_to_pfn(S3C2410_CS2), 58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 63 .pfn = __phys_to_pfn(S3C2410_CS2), 68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
D | mach-bast.c | 83 .pfn = PA_CS2(BAST_PA_ISAIO), 88 .pfn = PA_CS3(BAST_PA_ISAIO), 95 .pfn = __phys_to_pfn(BAST_PA_CTRL1), 100 .pfn = __phys_to_pfn(BAST_PA_CTRL2), 105 .pfn = __phys_to_pfn(BAST_PA_CTRL3), 110 .pfn = __phys_to_pfn(BAST_PA_CTRL4), 117 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQREQ), 122 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQRAW), 127 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQMASK),
|
D | mach-vr1000.c | 72 .pfn = PA_CS2(BAST_PA_ISAIO), 77 .pfn = PA_CS3(BAST_PA_ISAIO), 85 .pfn = __phys_to_pfn(VR1000_PA_CTRL1), 90 .pfn = __phys_to_pfn(VR1000_PA_CTRL2), 95 .pfn = __phys_to_pfn(VR1000_PA_CTRL3), 100 .pfn = __phys_to_pfn(VR1000_PA_CTRL4),
|
D | mach-osiris.c | 65 .pfn = __phys_to_pfn(S3C2410_CS5), 70 .pfn = __phys_to_pfn(S3C2410_CS5), 79 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL0), 84 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL1), 89 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL2), 94 .pfn = __phys_to_pfn(OSIRIS_PA_IDREG),
|
D | mach-smdk2416.c | 65 .pfn = __phys_to_pfn(S3C2410_CS2), 70 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 75 .pfn = __phys_to_pfn(S3C2410_CS2), 80 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
/linux-4.4.14/arch/arm/mach-ebsa110/ |
D | core.c | 78 .pfn = __phys_to_pfn(TRICK4_PHYS), 83 .pfn = __phys_to_pfn(TRICK3_PHYS), 88 .pfn = __phys_to_pfn(TRICK1_PHYS), 93 .pfn = __phys_to_pfn(TRICK0_PHYS), 103 .pfn = __phys_to_pfn(ISAIO_PHYS), 108 .pfn = __phys_to_pfn(ISAMEM_PHYS),
|
/linux-4.4.14/arch/score/mm/ |
D | cache.c | 82 unsigned long pfn, addr; in __update_cache() local 85 pfn = pte_pfn(pte); in __update_cache() 86 if (unlikely(!pfn_valid(pfn))) in __update_cache() 88 page = pfn_to_page(pfn); in __update_cache() 212 unsigned long addr, unsigned long pfn) in flush_cache_page() argument 215 unsigned long kaddr = 0xa0000000 | (pfn << PAGE_SHIFT); in flush_cache_page()
|
/linux-4.4.14/arch/openrisc/include/asm/ |
D | page.h | 83 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 92 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
/linux-4.4.14/arch/tile/gxio/ |
D | kiorpc.c | 34 unsigned long pfn; in iorpc_ioremap() local 56 pfn = pte_pfn(mmio_base) + PFN_DOWN(offset); in iorpc_ioremap() 58 return ioremap_prot(PFN_PHYS(pfn), size, prot); in iorpc_ioremap()
|
/linux-4.4.14/arch/unicore32/include/asm/ |
D | io.h | 63 static inline int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 65 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 67 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
D | mem.c | 56 u64 pfn; in mlx5_ib_cont_pages() local 80 pfn = sg_dma_address(sg) >> page_shift; in mlx5_ib_cont_pages() 83 tmp = (unsigned long)pfn; in mlx5_ib_cont_pages() 87 base = pfn; in mlx5_ib_cont_pages() 90 if (base + p != pfn) { in mlx5_ib_cont_pages() 95 base = pfn; in mlx5_ib_cont_pages()
|
/linux-4.4.14/arch/hexagon/include/asm/ |
D | page.h | 112 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument 144 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.4.14/arch/arm/kernel/ |
D | hibernate.c | 27 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 32 return (pfn >= nosave_begin_pfn) && (pfn <= nosave_end_pfn); in pfn_is_nosave()
|
D | crash_dump.c | 33 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.4.14/drivers/misc/ |
D | vmw_balloon.c | 445 static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_lock_page() argument 451 pfn32 = (u32)pfn; in vmballoon_send_lock_page() 452 if (pfn32 != pfn) in vmballoon_send_lock_page() 457 *hv_status = status = VMWARE_BALLOON_CMD(LOCK, pfn, dummy, *target); in vmballoon_send_lock_page() 461 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_lock_page() 470 unsigned long pfn = page_to_pfn(b->page); in vmballoon_send_batched_lock() local 475 status = VMWARE_BALLOON_CMD(BATCHED_2M_LOCK, pfn, num_pages, in vmballoon_send_batched_lock() 478 status = VMWARE_BALLOON_CMD(BATCHED_LOCK, pfn, num_pages, in vmballoon_send_batched_lock() 484 pr_debug("%s - batch ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_batched_lock() 493 static bool vmballoon_send_unlock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_unlock_page() argument [all …]
|
/linux-4.4.14/arch/arm/mach-ep93xx/ |
D | ts72xx.c | 35 .pfn = __phys_to_pfn(TS72XX_MODEL_PHYS_BASE), 40 .pfn = __phys_to_pfn(TS72XX_OPTIONS_PHYS_BASE), 45 .pfn = __phys_to_pfn(TS72XX_OPTIONS2_PHYS_BASE), 50 .pfn = __phys_to_pfn(TS72XX_RTC_INDEX_PHYS_BASE), 55 .pfn = __phys_to_pfn(TS72XX_RTC_DATA_PHYS_BASE),
|
/linux-4.4.14/drivers/char/ |
D | mem.c | 55 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 62 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument 64 u64 from = ((u64)pfn) << PAGE_SHIFT; in range_is_allowed() 69 if (!devmem_is_allowed(pfn)) { in range_is_allowed() 76 pfn++; in range_is_allowed() 81 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument 224 unsigned long pfn, unsigned long size, pgprot_t *vma_prot) in phys_mem_access_prot_allowed() argument 265 static pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 269 phys_addr_t offset = pfn << PAGE_SHIFT; in phys_mem_access_prot() 357 unsigned long pfn; in mmap_kmem() local [all …]
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 628 static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) in ehea_update_busmap() argument 641 start_section = (pfn * PAGE_SIZE) / EHEA_SECTSIZE; in ehea_update_busmap() 671 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_add_sect_bmap() argument 676 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); in ehea_add_sect_bmap() 681 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_rem_sect_bmap() argument 686 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); in ehea_rem_sect_bmap() 691 static int ehea_is_hugepage(unsigned long pfn) in ehea_is_hugepage() argument 695 if (pfn & EHEA_HUGEPAGE_PFN_MASK) in ehea_is_hugepage() 698 page_order = compound_order(pfn_to_page(pfn)); in ehea_is_hugepage() 709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local [all …]
|
/linux-4.4.14/arch/arm/mach-sa1100/ |
D | nanoengine.c | 70 .pfn = __phys_to_pfn(0x10000000), 76 .pfn = __phys_to_pfn(NANO_PCI_MEM_RW_PHYS), 82 .pfn = __phys_to_pfn(NANO_PCI_CONFIG_SPACE_PHYS),
|
/linux-4.4.14/lib/ |
D | ioremap.c | 57 u64 pfn; in ioremap_pte_range() local 59 pfn = phys_addr >> PAGE_SHIFT; in ioremap_pte_range() 65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range() 66 pfn++; in ioremap_pte_range()
|
/linux-4.4.14/arch/mips/include/asm/mach-ip27/ |
D | kernel-entry-init.h | 51 dsrl t1, 12 # 4K pfn 52 dsrl t2, 12 # 4K pfn 53 dsll t1, 6 # Get pfn into place 54 dsll t2, 6 # Get pfn into place
|
/linux-4.4.14/arch/mips/mm/ |
D | cache.c | 31 unsigned long pfn); 131 unsigned long pfn, addr; in __update_cache() local 134 pfn = pte_pfn(pte); in __update_cache() 135 if (unlikely(!pfn_valid(pfn))) in __update_cache() 137 page = pfn_to_page(pfn); in __update_cache()
|
D | ioremap.c | 23 unsigned long pfn; in remap_area_pte() local 32 pfn = phys_addr >> PAGE_SHIFT; in remap_area_pte() 38 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte() 40 pfn++; in remap_area_pte()
|
/linux-4.4.14/tools/testing/selftests/vm/ |
D | transhuge-stress.c | 110 int64_t pfn; in main() local 112 pfn = allocate_transhuge(p); in main() 114 if (pfn < 0) { in main() 117 size_t idx = pfn >> (HPAGE_SHIFT - PAGE_SHIFT); in main()
|
/linux-4.4.14/arch/xtensa/mm/ |
D | init.c | 238 unsigned long pfn; in bootmem_init() local 258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; in bootmem_init() 259 if (pfn < min_low_pfn) in bootmem_init() 260 min_low_pfn = pfn; in bootmem_init() 261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; in bootmem_init() 262 if (pfn > max_pfn) in bootmem_init() 263 max_pfn = pfn; in bootmem_init()
|
D | cache.c | 200 unsigned long pfn) in local_flush_cache_page() argument 204 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in local_flush_cache_page() 216 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local 219 if (!pfn_valid(pfn)) in update_mmu_cache() 222 page = pfn_to_page(pfn); in update_mmu_cache()
|
/linux-4.4.14/include/ras/ |
D | ras_event.h | 295 TP_PROTO(unsigned long pfn, 299 TP_ARGS(pfn, type, result), 302 __field(unsigned long, pfn) 308 __entry->pfn = pfn; 314 __entry->pfn,
|
/linux-4.4.14/arch/arm/mach-lpc32xx/ |
D | common.c | 168 .pfn = __phys_to_pfn(LPC32XX_AHB0_START), 174 .pfn = __phys_to_pfn(LPC32XX_AHB1_START), 180 .pfn = __phys_to_pfn(LPC32XX_FABAPB_START), 186 .pfn = __phys_to_pfn(LPC32XX_IRAM_BASE),
|
/linux-4.4.14/arch/s390/include/asm/ |
D | numa.h | 18 int numa_pfn_to_nid(unsigned long pfn); 29 static inline int numa_pfn_to_nid(unsigned long pfn) in numa_pfn_to_nid() argument
|
/linux-4.4.14/drivers/gpu/drm/gma500/ |
D | gem.c | 179 unsigned long pfn; in psb_gem_fault() local 212 pfn = (dev_priv->stolen_base + r->offset) >> PAGE_SHIFT; in psb_gem_fault() 214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault() 215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn); in psb_gem_fault()
|
/linux-4.4.14/arch/nios2/mm/ |
D | ioremap.c | 27 unsigned long pfn; in remap_area_pte() local 37 pfn = PFN_DOWN(phys_addr); in remap_area_pte() 43 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte() 45 pfn++; in remap_area_pte()
|
/linux-4.4.14/drivers/edac/ |
D | cell_edac.c | 38 unsigned long address, pfn, offset, syndrome; in cell_edac_count_ce() local 47 pfn = address >> PAGE_SHIFT; in cell_edac_count_ce() 53 csrow->first_page + pfn, offset, syndrome, in cell_edac_count_ce() 61 unsigned long address, pfn, offset; in cell_edac_count_ue() local 70 pfn = address >> PAGE_SHIFT; in cell_edac_count_ue() 75 csrow->first_page + pfn, offset, 0, in cell_edac_count_ue()
|
/linux-4.4.14/arch/ia64/kernel/ |
D | crash_dump.c | 34 copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page()
|
/linux-4.4.14/fs/ |
D | dax.c | 45 unsigned long pfn; in dax_clear_blocks() local 48 count = bdev_direct_access(bdev, sector, &addr, &pfn, size); in dax_clear_blocks() 74 unsigned long pfn; in dax_get_addr() local 76 return bdev_direct_access(bh->b_bdev, sector, addr, &pfn, bh->b_size); in dax_get_addr() 299 unsigned long pfn; in dax_insert_mapping() local 318 error = bdev_direct_access(bh->b_bdev, sector, &addr, &pfn, bh->b_size); in dax_insert_mapping() 331 error = vm_insert_mixed(vma, vaddr, pfn); in dax_insert_mapping() 541 unsigned long pfn; in __dax_pmd_fault() local 627 length = bdev_direct_access(bh.b_bdev, sector, &kaddr, &pfn, in __dax_pmd_fault() 633 if ((length < PMD_SIZE) || (pfn & PG_PMD_COLOUR)) in __dax_pmd_fault() [all …]
|
/linux-4.4.14/drivers/scsi/csiostor/ |
D | csio_mb.c | 281 (FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(hw->pfn)); in csio_mb_ldst() 449 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_alloc() 505 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_write() 628 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_free() 663 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_alloc() 711 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_write() 814 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_free() 1158 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_debug_cmd_handler() 1159 uint32_t data_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_DATA_A); in csio_mb_debug_cmd_handler() 1192 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_issue() [all …]
|
/linux-4.4.14/Documentation/virtual/kvm/ |
D | locking.txt | 31 1): The mapping from gfn to pfn 32 The mapping from gfn to pfn may be changed since we can only ensure the pfn 63 to pin gfn to pfn, because after gfn_to_pfn_atomic(): 64 - We have held the refcount of pfn that means the pfn can not be freed and 66 - The pfn is writable that means it can not be shared between different gfns 106 kvm_set_pfn_accessed(spte.pfn); 108 kvm_set_pfn_dirty(spte.pfn);
|
/linux-4.4.14/virt/kvm/ |
D | kvm_main.c | 114 static void kvm_release_pfn_dirty(pfn_t pfn); 122 bool kvm_is_reserved_pfn(pfn_t pfn) in kvm_is_reserved_pfn() argument 124 if (pfn_valid(pfn)) in kvm_is_reserved_pfn() 125 return PageReserved(pfn_to_page(pfn)); in kvm_is_reserved_pfn() 1303 bool write_fault, bool *writable, pfn_t *pfn) in hva_to_pfn_fast() argument 1321 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast() 1336 bool *writable, pfn_t *pfn) in hva_to_pfn_slow() argument 1371 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow() 1404 pfn_t pfn = 0; in hva_to_pfn() local 1410 if (hva_to_pfn_fast(addr, atomic, async, write_fault, writable, &pfn)) in hva_to_pfn() [all …]
|
/linux-4.4.14/arch/arm/mach-mmp/ |
D | common.c | 30 .pfn = __phys_to_pfn(APB_PHYS_BASE), 35 .pfn = __phys_to_pfn(AXI_PHYS_BASE),
|
/linux-4.4.14/arch/microblaze/kernel/ |
D | dma.c | 166 unsigned long pfn; in dma_direct_mmap_coherent() local 173 pfn = consistent_virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent() 175 pfn = virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent() 177 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent()
|
/linux-4.4.14/arch/mips/kernel/ |
D | crash_dump.c | 26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
|
/linux-4.4.14/tools/vm/ |
D | page-types.c | 290 unsigned long pfn; in pagemap_pfn() local 293 pfn = PM_PFRAME(val); in pagemap_pfn() 295 pfn = 0; in pagemap_pfn() 297 return pfn; in pagemap_pfn() 622 unsigned long pfn; in walk_vma() local 632 pfn = pagemap_pfn(buf[i]); in walk_vma() 633 if (pfn) in walk_vma() 634 walk_pfn(index + i, pfn, 1, buf[i]); in walk_vma() 854 unsigned long nr_pages, pfn, i; in walk_file() local 906 pfn = pagemap_pfn(buf[i]); in walk_file() [all …]
|
/linux-4.4.14/arch/mn10300/mm/ |
D | pgtable.c | 35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument 45 if (pfn & (PTRS_PER_PTE-1)) { /* pfn is misaligned */ in set_pmd_pfn() 56 set_pmd(pmd, pfn_pmd(pfn, flags)); in set_pmd_pfn()
|
/linux-4.4.14/drivers/gpu/drm/ttm/ |
D | ttm_bo_vm.c | 93 unsigned long pfn; in ttm_bo_vm_fault() local 216 pfn = ((bo->mem.bus.base + bo->mem.bus.offset) >> PAGE_SHIFT) + page_offset; in ttm_bo_vm_fault() 228 pfn = page_to_pfn(page); in ttm_bo_vm_fault() 232 ret = vm_insert_mixed(&cvma, address, pfn); in ttm_bo_vm_fault() 234 ret = vm_insert_pfn(&cvma, address, pfn); in ttm_bo_vm_fault()
|
/linux-4.4.14/arch/arm/mach-versatile/ |
D | core.c | 134 .pfn = __phys_to_pfn(VERSATILE_SYS_BASE), 139 .pfn = __phys_to_pfn(VERSATILE_SIC_BASE), 144 .pfn = __phys_to_pfn(VERSATILE_VIC_BASE), 149 .pfn = __phys_to_pfn(VERSATILE_SCTL_BASE), 156 .pfn = __phys_to_pfn(VERSATILE_IB2_BASE), 164 .pfn = __phys_to_pfn(VERSATILE_UART0_BASE), 172 .pfn = __phys_to_pfn(VERSATILE_PCI_CORE_BASE), 177 .pfn = __phys_to_pfn(VERSATILE_PCI_BASE), 182 .pfn = __phys_to_pfn(VERSATILE_PCI_CFG_BASE),
|
/linux-4.4.14/drivers/iommu/ |
D | iova.c | 301 struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn) in find_iova() argument 313 if ((pfn >= iova->pfn_lo) && (pfn <= iova->pfn_hi)) { in find_iova() 324 if (pfn < iova->pfn_lo) in find_iova() 326 else if (pfn > iova->pfn_lo) in find_iova() 362 free_iova(struct iova_domain *iovad, unsigned long pfn) in free_iova() argument 364 struct iova *iova = find_iova(iovad, pfn); in free_iova()
|
/linux-4.4.14/arch/arm/mach-rpc/ |
D | riscpc.c | 71 .pfn = __phys_to_pfn(SCREEN_START), 76 .pfn = __phys_to_pfn(IO_START), 81 .pfn = __phys_to_pfn(EASI_START),
|
/linux-4.4.14/arch/arm/mach-ks8695/include/mach/ |
D | memory.h | 34 #define __arch_pfn_to_dma(dev, pfn) \ argument 35 ({ dma_addr_t __dma = __pfn_to_phys(pfn); \
|
/linux-4.4.14/arch/arm/mach-omap1/include/mach/ |
D | memory.h | 30 #define __arch_pfn_to_dma(dev, pfn) \ argument 31 ({ dma_addr_t __dma = __pfn_to_phys(pfn); \
|