Home
last modified time | relevance | path

Searched refs:pfn (Results 1 – 200 of 568) sorted by relevance

123

/linux-4.4.14/arch/x86/xen/
Dp2m.c124 static inline unsigned p2m_top_index(unsigned long pfn) in p2m_top_index() argument
126 BUG_ON(pfn >= MAX_P2M_PFN); in p2m_top_index()
127 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); in p2m_top_index()
130 static inline unsigned p2m_mid_index(unsigned long pfn) in p2m_mid_index() argument
132 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; in p2m_mid_index()
135 static inline unsigned p2m_index(unsigned long pfn) in p2m_index() argument
137 return pfn % P2M_PER_PAGE; in p2m_index()
172 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) in p2m_init_identity() argument
177 p2m[i] = IDENTITY_FRAME(pfn + i); in p2m_init_identity()
210 unsigned long pfn, mfn; in xen_build_mfn_list_list() local
[all …]
Dsetup.c163 unsigned long __ref xen_chk_extra_mem(unsigned long pfn) in xen_chk_extra_mem() argument
168 if (pfn >= xen_extra_mem[i].start_pfn && in xen_chk_extra_mem()
169 pfn < xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns) in xen_chk_extra_mem()
173 return IDENTITY_FRAME(pfn); in xen_chk_extra_mem()
181 unsigned long pfn, pfn_s, pfn_e; in xen_inv_extra_mem() local
189 for (pfn = pfn_s; pfn < pfn_e; pfn++) in xen_inv_extra_mem()
190 set_phys_to_machine(pfn, INVALID_P2M_ENTRY); in xen_inv_extra_mem()
256 unsigned long pfn, end; in xen_set_identity_and_release_chunk() local
263 for (pfn = start_pfn; pfn < end; pfn++) { in xen_set_identity_and_release_chunk()
264 unsigned long mfn = pfn_to_mfn(pfn); in xen_set_identity_and_release_chunk()
[all …]
Dmmu.c371 unsigned long pfn = mfn_to_pfn(mfn); in pte_mfn_to_pfn() local
374 if (unlikely(pfn == ~0)) in pte_mfn_to_pfn()
377 val = ((pteval_t)pfn << PAGE_SHIFT) | flags; in pte_mfn_to_pfn()
386 unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT; in pte_pfn_to_mfn() local
391 mfn = __pfn_to_mfn(pfn); in pte_pfn_to_mfn()
393 mfn = pfn; in pte_pfn_to_mfn()
729 static void xen_do_pin(unsigned level, unsigned long pfn) in xen_do_pin() argument
734 op.arg1.mfn = pfn_to_mfn(pfn); in xen_do_pin()
753 unsigned long pfn = page_to_pfn(page); in xen_pin_page() local
784 pfn_pte(pfn, PAGE_KERNEL_RO), in xen_pin_page()
[all …]
Dmmu.h14 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
16 void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags);
Denlighten.c490 unsigned long pfn; in set_aliased_prot() local
497 pfn = pte_pfn(*ptep); in set_aliased_prot()
498 page = pfn_to_page(pfn); in set_aliased_prot()
500 pte = pfn_pte(pfn, prot); in set_aliased_prot()
532 void *av = __va(PFN_PHYS(pfn)); in set_aliased_prot()
608 unsigned long pfn, mfn; in xen_load_gdt() local
621 pfn = pte_pfn(*ptep); in xen_load_gdt()
622 mfn = pfn_to_mfn(pfn); in xen_load_gdt()
623 virt = __va(PFN_PHYS(pfn)); in xen_load_gdt()
656 unsigned long pfn, mfn; in xen_load_gdt_boot() local
[all …]
/linux-4.4.14/arch/x86/include/asm/xen/
Dpage.h46 extern int xen_alloc_p2m_entry(unsigned long pfn);
48 extern unsigned long get_phys_to_machine(unsigned long pfn);
49 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn);
50 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
85 static inline unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument
89 if (pfn < xen_p2m_size) in __pfn_to_mfn()
90 mfn = xen_p2m_addr[pfn]; in __pfn_to_mfn()
91 else if (unlikely(pfn < xen_max_p2m_pfn)) in __pfn_to_mfn()
92 return get_phys_to_machine(pfn); in __pfn_to_mfn()
94 return IDENTITY_FRAME(pfn); in __pfn_to_mfn()
[all …]
Dinterface_32.h99 #define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20)) argument
/linux-4.4.14/mm/
Dpage_isolation.c16 unsigned long flags, pfn; in set_migratetype_isolate() local
25 pfn = page_to_pfn(page); in set_migratetype_isolate()
26 arg.start_pfn = pfn; in set_migratetype_isolate()
133 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument
137 if (pfn_valid_within(pfn + i)) in __first_valid_page()
141 return pfn_to_page(pfn + i); in __first_valid_page()
161 unsigned long pfn; in start_isolate_page_range() local
168 for (pfn = start_pfn; in start_isolate_page_range()
169 pfn < end_pfn; in start_isolate_page_range()
170 pfn += pageblock_nr_pages) { in start_isolate_page_range()
[all …]
Dpage_ext.c104 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local
119 offset = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext()
181 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local
182 struct mem_section *section = __pfn_to_section(pfn); in lookup_page_ext()
193 return section->page_ext + pfn; in lookup_page_ext()
215 static int __meminit init_section_page_ext(unsigned long pfn, int nid) in init_section_page_ext() argument
221 section = __pfn_to_section(pfn); in init_section_page_ext()
245 pfn &= PAGE_SECTION_MASK; in init_section_page_ext()
246 section->page_ext = base - pfn; in init_section_page_ext()
266 static void __free_page_ext(unsigned long pfn) in __free_page_ext() argument
[all …]
Dpage_owner.c87 print_page_owner(char __user *buf, size_t count, unsigned long pfn, in print_page_owner() argument
110 pageblock_mt = get_pfnblock_migratetype(page, pfn); in print_page_owner()
114 pfn, in print_page_owner()
115 pfn >> pageblock_order, in print_page_owner()
156 unsigned long pfn; in read_page_owner() local
164 pfn = min_low_pfn + *ppos; in read_page_owner()
167 while (!pfn_valid(pfn) && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0) in read_page_owner()
168 pfn++; in read_page_owner()
173 for (; pfn < max_pfn; pfn++) { in read_page_owner()
178 if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0 && !pfn_valid(pfn)) { in read_page_owner()
[all …]
Dmemory-failure.c182 unsigned long pfn, struct page *page, int flags) in kill_proc() argument
189 pfn, t->comm, t->pid); in kill_proc()
326 int fail, struct page *page, unsigned long pfn, in kill_procs() argument
341 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs()
352 pfn, page, flags) < 0) in kill_procs()
355 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs()
556 static int me_kernel(struct page *p, unsigned long pfn) in me_kernel() argument
564 static int me_unknown(struct page *p, unsigned long pfn) in me_unknown() argument
566 printk(KERN_ERR "MCE %#lx: Unknown page state\n", pfn); in me_unknown()
573 static int me_pagecache_clean(struct page *p, unsigned long pfn) in me_pagecache_clean() argument
[all …]
Dmemory_hotplug.c247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local
271 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node()
275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
282 if (pfn_valid(pfn) && (pfn_to_nid(pfn) == node)) in register_page_bootmem_info_node()
283 register_page_bootmem_info_section(pfn); in register_page_bootmem_info_node()
330 unsigned long pfn; in fix_zone_id() local
332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id()
333 set_page_links(pfn_to_page(pfn), zid, nid, pfn); in fix_zone_id()
449 unsigned long flags, pfn; in __add_zone() local
466 for (pfn = phys_start_pfn; pfn < phys_start_pfn + nr_pages; pfn++) { in __add_zone()
[all …]
Dpage_idle.c30 static struct page *page_idle_get_page(unsigned long pfn) in page_idle_get_page() argument
35 if (!pfn_valid(pfn)) in page_idle_get_page()
38 page = pfn_to_page(pfn); in page_idle_get_page()
121 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local
127 pfn = pos * BITS_PER_BYTE; in page_idle_bitmap_read()
128 if (pfn >= max_pfn) in page_idle_bitmap_read()
131 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read()
135 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read()
136 bit = pfn % BITMAP_CHUNK_BITS; in page_idle_bitmap_read()
139 page = page_idle_get_page(pfn); in page_idle_bitmap_read()
[all …]
Dcma.c84 static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, in cma_clear_bitmap() argument
89 bitmap_no = (pfn - cma->base_pfn) >> cma->order_per_bit; in cma_clear_bitmap()
100 unsigned long base_pfn = cma->base_pfn, pfn = base_pfn; in cma_activate_area() local
109 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area()
110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area()
115 base_pfn = pfn; in cma_activate_area()
116 for (j = pageblock_nr_pages; j; --j, pfn++) { in cma_activate_area()
117 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area()
124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area()
367 unsigned long pfn = -1; in cma_alloc() local
[all …]
Dcompaction.c48 unsigned long pfn = page_to_pfn(page); in release_freepages() local
51 if (pfn > high_pfn) in release_freepages()
52 high_pfn = pfn; in release_freepages()
215 unsigned long pfn; in __reset_isolation_suitable() local
220 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable()
225 if (!pfn_valid(pfn)) in __reset_isolation_suitable()
228 page = pfn_to_page(pfn); in __reset_isolation_suitable()
262 unsigned long pfn; in update_pageblock_skip() local
275 pfn = page_to_pfn(page); in update_pageblock_skip()
279 if (pfn > zone->compact_cached_migrate_pfn[0]) in update_pageblock_skip()
[all …]
Dpage_alloc.c276 static inline bool __meminit early_page_uninitialised(unsigned long pfn) in early_page_uninitialised() argument
278 if (pfn >= NODE_DATA(early_pfn_to_nid(pfn))->first_deferred_pfn) in early_page_uninitialised()
284 static inline bool early_page_nid_uninitialised(unsigned long pfn, int nid) in early_page_nid_uninitialised() argument
286 if (pfn >= NODE_DATA(nid)->first_deferred_pfn) in early_page_nid_uninitialised()
297 unsigned long pfn, unsigned long zone_end, in update_defer_init() argument
307 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in update_defer_init()
308 pgdat->first_deferred_pfn = pfn; in update_defer_init()
319 static inline bool early_page_uninitialised(unsigned long pfn) in early_page_uninitialised() argument
324 static inline bool early_page_nid_uninitialised(unsigned long pfn, int nid) in early_page_nid_uninitialised() argument
330 unsigned long pfn, unsigned long zone_end, in update_defer_init() argument
[all …]
Dhwpoison-inject.c15 unsigned long pfn = val; in hwpoison_inject() local
23 if (!pfn_valid(pfn)) in hwpoison_inject()
26 p = pfn_to_page(pfn); in hwpoison_inject()
55 pr_info("Injecting memory failure at pfn %#lx\n", pfn); in hwpoison_inject()
56 return memory_failure(pfn, 18, MF_COUNT_INCREASED); in hwpoison_inject()
Dvmstat.c1002 unsigned long pfn; in pagetypeinfo_showblockcount_print() local
1007 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
1010 if (!pfn_valid(pfn)) in pagetypeinfo_showblockcount_print()
1013 page = pfn_to_page(pfn); in pagetypeinfo_showblockcount_print()
1016 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print()
1054 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print() local
1055 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print()
1061 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print()
1068 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print()
1069 if (!pfn_valid(pfn)) { in pagetypeinfo_showmixedcount_print()
[all …]
Dsparse.c172 unsigned long pfn; in memory_present() local
176 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
177 unsigned long section = pfn_to_section_nr(pfn); in memory_present()
197 unsigned long pfn; in node_memmap_size_bytes() local
201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
202 if (nid != early_pfn_to_nid(pfn)) in node_memmap_size_bytes()
205 if (pfn_present(pfn)) in node_memmap_size_bytes()
Dmemory.c753 unsigned long pfn = pte_pfn(pte); in vm_normal_page() local
762 if (!is_zero_pfn(pfn)) in vm_normal_page()
771 if (!pfn_valid(pfn)) in vm_normal_page()
777 if (pfn == vma->vm_pgoff + off) in vm_normal_page()
784 if (is_zero_pfn(pfn)) in vm_normal_page()
787 if (unlikely(pfn > highest_memmap_pfn)) { in vm_normal_page()
797 return pfn_to_page(pfn); in vm_normal_page()
804 unsigned long pfn = pmd_pfn(pmd); in vm_normal_page_pmd() local
813 if (!pfn_valid(pfn)) in vm_normal_page_pmd()
819 if (pfn == vma->vm_pgoff + off) in vm_normal_page_pmd()
[all …]
/linux-4.4.14/arch/arm/xen/
Dp2m.c21 unsigned long pfn; member
42 if (new->pfn == entry->pfn) in xen_add_phys_to_mach_entry()
45 if (new->pfn < entry->pfn) in xen_add_phys_to_mach_entry()
57 __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); in xen_add_phys_to_mach_entry()
62 unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument
71 if (entry->pfn <= pfn && in __pfn_to_mfn()
72 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn()
74 return entry->mfn + (pfn - entry->pfn); in __pfn_to_mfn()
76 if (pfn < entry->pfn) in __pfn_to_mfn()
119 bool __set_phys_to_machine_multi(unsigned long pfn, in __set_phys_to_machine_multi() argument
[all …]
/linux-4.4.14/include/asm-generic/
Dmemory_model.h15 #define arch_pfn_to_nid(pfn) pfn_to_nid(pfn) argument
19 #define arch_local_page_offset(pfn, nid) \ argument
20 ((pfn) - NODE_DATA(nid)->node_start_pfn)
30 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) argument
35 #define __pfn_to_page(pfn) \ argument
36 ({ unsigned long __pfn = (pfn); \
51 #define __pfn_to_page(pfn) (vmemmap + (pfn)) argument
65 #define __pfn_to_page(pfn) \ argument
66 ({ unsigned long __pfn = (pfn); \
76 #define __pfn_to_phys(pfn) PFN_PHYS(pfn) argument
Dpage.h84 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
93 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr) argument
Dpgtable.h546 unsigned long pfn, unsigned long addr, in track_pfn_remap() argument
557 unsigned long pfn) in track_pfn_insert() argument
577 unsigned long pfn, unsigned long size) in untrack_pfn() argument
582 unsigned long pfn, unsigned long addr,
585 unsigned long pfn);
587 extern void untrack_pfn(struct vm_area_struct *vma, unsigned long pfn,
592 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument
595 unsigned long offset_from_zero_pfn = pfn - zero_pfn; in is_zero_pfn()
602 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument
605 return pfn == zero_pfn; in is_zero_pfn()
/linux-4.4.14/include/trace/events/
Dcma.h12 TP_PROTO(unsigned long pfn, const struct page *page,
15 TP_ARGS(pfn, page, count, align),
18 __field(unsigned long, pfn)
25 __entry->pfn = pfn;
32 __entry->pfn,
40 TP_PROTO(unsigned long pfn, const struct page *page,
43 TP_ARGS(pfn, page, count),
46 __field(unsigned long, pfn)
52 __entry->pfn = pfn;
58 __entry->pfn,
Dkmem.h180 __field( unsigned long, pfn )
185 __entry->pfn = page_to_pfn(page);
190 pfn_to_page(__entry->pfn),
191 __entry->pfn,
202 __field( unsigned long, pfn )
207 __entry->pfn = page_to_pfn(page);
212 pfn_to_page(__entry->pfn),
213 __entry->pfn,
225 __field( unsigned long, pfn )
232 __entry->pfn = page ? page_to_pfn(page) : -1UL;
[all …]
Dpagemap.h38 __field(unsigned long, pfn )
45 __entry->pfn = page_to_pfn(page);
53 __entry->pfn,
71 __field(unsigned long, pfn )
76 __entry->pfn = page_to_pfn(page);
80 TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
Dfilemap.h21 __field(unsigned long, pfn)
28 __entry->pfn = page_to_pfn(page);
40 pfn_to_page(__entry->pfn),
41 __entry->pfn,
Dxen.h328 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned),
329 TP_ARGS(mm, pfn, level, pinned),
332 __field(unsigned long, pfn)
337 __entry->pfn = pfn;
341 __entry->mm, __entry->pfn, __entry->level,
346 TP_PROTO(unsigned long pfn, unsigned level, bool pinned),
347 TP_ARGS(pfn, level, pinned),
349 __field(unsigned long, pfn)
353 TP_fast_assign(__entry->pfn = pfn;
357 __entry->pfn, __entry->level,
/linux-4.4.14/arch/m32r/include/asm/
Dmmzone.h16 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument
27 #define pfn_valid(pfn) (1) argument
29 #define pfn_valid(pfn) ((pfn) < num_physpages) argument
36 static __inline__ int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
41 if (pfn >= node_start_pfn(node) && pfn < node_end_pfn(node)) in pfn_to_nid()
47 static __inline__ struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
49 return(NODE_DATA(pfn_to_nid(pfn))); in pfn_to_pgdat()
Dpgtable-2level.h71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
Dcacheflush.h14 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
36 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
50 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
/linux-4.4.14/arch/parisc/include/asm/
Dmmzone.h33 #define pfn_is_io(pfn) ((pfn & (0xf0000000UL >> PAGE_SHIFT)) == (0xf0000000UL >> PAGE_SHIFT)) argument
36 #define pfn_is_io(pfn) ((pfn & (0xf000000000000000UL >> PAGE_SHIFT)) == (0xf000000000000000UL >> PA… argument
39 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
43 if (unlikely(pfn_is_io(pfn))) in pfn_to_nid()
46 i = pfn >> PFNNID_SHIFT; in pfn_to_nid()
52 static inline int pfn_valid(int pfn) in pfn_valid() argument
54 int nid = pfn_to_nid(pfn); in pfn_valid()
57 return (pfn < node_end_pfn(nid)); in pfn_valid()
/linux-4.4.14/arch/arm/mach-gemini/
Dmm.c23 .pfn =__phys_to_pfn(GEMINI_GLOBAL_BASE),
28 .pfn = __phys_to_pfn(GEMINI_UART_BASE),
33 .pfn = __phys_to_pfn(GEMINI_TIMER_BASE),
38 .pfn = __phys_to_pfn(GEMINI_INTERRUPT_BASE),
43 .pfn = __phys_to_pfn(GEMINI_POWER_CTRL_BASE),
48 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(0)),
53 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(1)),
58 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(2)),
63 .pfn = __phys_to_pfn(GEMINI_FLASH_CTRL_BASE),
68 .pfn = __phys_to_pfn(GEMINI_DRAM_CTRL_BASE),
[all …]
/linux-4.4.14/arch/arm/include/asm/xen/
Dpage.h14 #define phys_to_machine_mapping_valid(pfn) (1) argument
42 unsigned long __pfn_to_mfn(unsigned long pfn);
46 static inline unsigned long pfn_to_gfn(unsigned long pfn) in pfn_to_gfn() argument
48 return pfn; in pfn_to_gfn()
57 static inline unsigned long pfn_to_bfn(unsigned long pfn) in pfn_to_bfn() argument
62 mfn = __pfn_to_mfn(pfn); in pfn_to_bfn()
67 return pfn; in pfn_to_bfn()
105 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
106 bool __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn,
109 static inline bool set_phys_to_machine(unsigned long pfn, unsigned long mfn) in set_phys_to_machine() argument
[all …]
Dpage-coherent.h64 unsigned long pfn = PFN_DOWN(handle); in xen_dma_unmap_page() local
73 if (pfn_valid(pfn)) { in xen_dma_unmap_page()
83 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_cpu() local
84 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_cpu()
94 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_device() local
95 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_device()
/linux-4.4.14/arch/s390/kernel/
Dsuspend.c99 void page_key_read(unsigned long *pfn) in page_key_read() argument
103 addr = (unsigned long) page_address(pfn_to_page(*pfn)); in page_key_read()
104 *(unsigned char *) pfn = (unsigned char) page_get_storage_key(addr); in page_key_read()
111 void page_key_memorize(unsigned long *pfn) in page_key_memorize() argument
113 page_key_wp->data[page_key_wx] = *(unsigned char *) pfn; in page_key_memorize()
114 *(unsigned char *) pfn = 0; in page_key_memorize()
137 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
145 if (pfn <= LC_PAGES) in pfn_is_nosave()
147 if (pfn >= nosave_begin_pfn && pfn < nosave_end_pfn) in pfn_is_nosave()
150 if (pfn >= stext_pfn && pfn <= eshared_pfn) in pfn_is_nosave()
[all …]
Dcrash_dump.c132 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, size_t csize, in copy_oldmem_page() argument
139 src = (pfn << PAGE_SHIFT) + offset; in copy_oldmem_page()
153 unsigned long from, unsigned long pfn, in remap_oldmem_pfn_range_kdump() argument
159 if (pfn < OLDMEM_SIZE >> PAGE_SHIFT) { in remap_oldmem_pfn_range_kdump()
160 size_old = min(size, OLDMEM_SIZE - (pfn << PAGE_SHIFT)); in remap_oldmem_pfn_range_kdump()
162 pfn + (OLDMEM_BASE >> PAGE_SHIFT), in remap_oldmem_pfn_range_kdump()
168 pfn += size_old >> PAGE_SHIFT; in remap_oldmem_pfn_range_kdump()
170 return remap_pfn_range(vma, from, pfn, size, prot); in remap_oldmem_pfn_range_kdump()
181 unsigned long pfn, in remap_oldmem_pfn_range_zfcpdump() argument
187 if (pfn < hsa_end >> PAGE_SHIFT) { in remap_oldmem_pfn_range_zfcpdump()
[all …]
/linux-4.4.14/arch/arm/mach-omap2/
Dio.c74 .pfn = __phys_to_pfn(L3_24XX_PHYS),
80 .pfn = __phys_to_pfn(L4_24XX_PHYS),
90 .pfn = __phys_to_pfn(DSP_MEM_2420_PHYS),
96 .pfn = __phys_to_pfn(DSP_IPI_2420_PHYS),
102 .pfn = __phys_to_pfn(DSP_MMU_2420_PHYS),
114 .pfn = __phys_to_pfn(L4_WK_243X_PHYS),
120 .pfn = __phys_to_pfn(OMAP243X_GPMC_PHYS),
126 .pfn = __phys_to_pfn(OMAP243X_SDRC_PHYS),
132 .pfn = __phys_to_pfn(OMAP243X_SMS_PHYS),
144 .pfn = __phys_to_pfn(L3_34XX_PHYS),
[all …]
/linux-4.4.14/arch/x86/power/
Dhibernate_32.c83 unsigned long pfn; in resume_physical_mapping_init() local
91 pfn = 0; in resume_physical_mapping_init()
98 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
102 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
110 set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC)); in resume_physical_mapping_init()
111 pfn += PTRS_PER_PTE; in resume_physical_mapping_init()
120 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init()
121 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
124 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init()
167 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
[all …]
Dhibernate_64.c103 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
107 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.4.14/arch/alpha/include/asm/
Dmmzone.h34 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument
81 unsigned long pfn; \
83 pfn = page_to_pfn(page) << 32; \
84 pte_val(pte) = pfn | pgprot_val(pgprot); \
103 #define pfn_to_nid(pfn) pa_to_nid(((u64)(pfn) << PAGE_SHIFT)) argument
104 #define pfn_valid(pfn) \ argument
105 (((pfn) - node_start_pfn(pfn_to_nid(pfn))) < \
106 node_spanned_pages(pfn_to_nid(pfn))) \
Dpgtable.h194 #define PHYS_TWIDDLE(pfn) \ argument
195 ((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \
196 ? ((pfn) ^= KSEG_PFN) : (pfn))
198 #define PHYS_TWIDDLE(pfn) (pfn) argument
/linux-4.4.14/arch/unicore32/mm/
Dioremap.c102 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument
118 set_pmd(pmd, __pmd(__pfn_to_phys(pfn) | type->prot_sect)); in remap_area_sections()
119 pfn += SZ_4M >> PAGE_SHIFT; in remap_area_sections()
129 void __iomem *__uc32_ioremap_pfn_caller(unsigned long pfn, in __uc32_ioremap_pfn_caller() argument
140 if (pfn >= 0x100000 && (__pfn_to_phys(pfn) & ~SECTION_MASK)) in __uc32_ioremap_pfn_caller()
146 if (pfn_valid(pfn)) { in __uc32_ioremap_pfn_caller()
168 if (!((__pfn_to_phys(pfn) | size | addr) & ~PMD_MASK)) { in __uc32_ioremap_pfn_caller()
170 err = remap_area_sections(addr, pfn, size, type); in __uc32_ioremap_pfn_caller()
172 err = ioremap_page_range(addr, addr + size, __pfn_to_phys(pfn), in __uc32_ioremap_pfn_caller()
189 unsigned long pfn = __phys_to_pfn(phys_addr); in __uc32_ioremap_caller() local
[all …]
Dmmu.c81 unsigned long pfn; member
164 unsigned long end, unsigned long pfn, in alloc_init_pte() argument
169 set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); in alloc_init_pte()
170 pfn++; in alloc_init_pte()
217 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
225 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
231 phys = (unsigned long)__pfn_to_phys(md->pfn); in create_mapping()
237 __pfn_to_phys(md->pfn), addr); in create_mapping()
368 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init()
378 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init()
[all …]
/linux-4.4.14/arch/unicore32/kernel/
Dhibernate.c75 unsigned long pfn; in resume_physical_mapping_init() local
83 pfn = 0; in resume_physical_mapping_init()
90 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
96 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
107 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init()
108 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
111 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init()
145 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
/linux-4.4.14/arch/x86/include/asm/
Dmmzone_32.h33 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid()
42 static inline int pfn_valid(int pfn) in pfn_valid() argument
44 int nid = pfn_to_nid(pfn); in pfn_valid()
47 return (pfn < node_end_pfn(nid)); in pfn_valid()
51 #define early_pfn_valid(pfn) pfn_valid((pfn)) argument
Dpgalloc.h15 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pte() argument
16 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pmd() argument
17 static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, in paravirt_alloc_pmd_clone() argument
19 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pud() argument
20 static inline void paravirt_release_pte(unsigned long pfn) {} in paravirt_release_pte() argument
21 static inline void paravirt_release_pmd(unsigned long pfn) {} in paravirt_release_pmd() argument
22 static inline void paravirt_release_pud(unsigned long pfn) {} in paravirt_release_pud() argument
72 unsigned long pfn = page_to_pfn(pte); in pmd_populate() local
74 paravirt_alloc_pte(mm, pfn); in pmd_populate()
75 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
Dpage_64.h34 #define pfn_valid(pfn) ((pfn) < max_pfn) argument
Dpage_32.h18 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
Dhighmem.h69 void *kmap_atomic_pfn(unsigned long pfn);
70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
/linux-4.4.14/kernel/power/
Dsnapshot.c634 static int memory_bm_find_bit(struct memory_bitmap *bm, unsigned long pfn, in memory_bm_find_bit() argument
643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit()
650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit()
666 if (((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) in memory_bm_find_bit()
670 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; in memory_bm_find_bit()
685 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; in memory_bm_find_bit()
689 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; in memory_bm_find_bit()
694 static void memory_bm_set_bit(struct memory_bitmap *bm, unsigned long pfn) in memory_bm_set_bit() argument
700 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit()
705 static int mem_bm_set_bit_check(struct memory_bitmap *bm, unsigned long pfn) in mem_bm_set_bit_check() argument
[all …]
/linux-4.4.14/arch/arm/mm/
Dfault-armv.c41 unsigned long pfn, pte_t *ptep) in do_adjust_pte() argument
56 flush_cache_page(vma, address, pfn); in do_adjust_pte()
57 outer_flush_range((pfn << PAGE_SHIFT), in do_adjust_pte()
58 (pfn << PAGE_SHIFT) + PAGE_SIZE); in do_adjust_pte()
93 unsigned long pfn) in adjust_pte() argument
123 ret = do_adjust_pte(vma, address, pfn, pte); in adjust_pte()
133 unsigned long addr, pte_t *ptep, unsigned long pfn) in make_coherent() argument
160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); in make_coherent()
164 do_adjust_pte(vma, addr, pfn, ptep); in make_coherent()
183 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
[all …]
Dflush.c41 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument
46 set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL)); in flush_pfn_alias()
55 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument
61 set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL)); in flush_icache_alias()
101 void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) in flush_cache_page() argument
104 vivt_flush_cache_page(vma, user_addr, pfn); in flush_cache_page()
109 flush_pfn_alias(pfn, user_addr); in flush_cache_page()
118 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument
119 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument
275 unsigned long pfn; in __sync_icache_dcache() local
[all …]
Dioremap.c186 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument
204 pmd[0] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections()
205 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections()
206 pmd[1] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections()
207 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections()
218 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument
238 super_pmd_val = __pfn_to_phys(pfn) | type->prot_sect | in remap_area_supersections()
240 super_pmd_val |= ((pfn >> (32 - PAGE_SHIFT)) & 0xf) << 20; in remap_area_supersections()
251 pfn += SUPERSECTION_SIZE >> PAGE_SHIFT; in remap_area_supersections()
258 static void __iomem * __arm_ioremap_pfn_caller(unsigned long pfn, in __arm_ioremap_pfn_caller() argument
[all …]
Dmmu.c701 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
704 if (!pfn_valid(pfn)) in phys_mem_access_prot()
738 unsigned long end, unsigned long pfn, in alloc_init_pte() argument
743 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0); in alloc_init_pte()
744 pfn++; in alloc_init_pte()
829 phys = __pfn_to_phys(md->pfn); in create_36bit_mapping()
834 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
846 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
850 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { in create_36bit_mapping()
852 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
[all …]
Dmmap.c215 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
217 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range()
231 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
233 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
235 if (!page_is_ram(pfn)) in devmem_is_allowed()
Dhighmem.c129 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument
133 struct page *page = pfn_to_page(pfn); in kmap_atomic_pfn()
146 set_fixmap_pte(idx, pfn_pte(pfn, kmap_prot)); in kmap_atomic_pfn()
/linux-4.4.14/include/linux/
Dmmzone.h85 static inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn) in get_pfnblock_migratetype() argument
88 return get_pfnblock_flags_mask(page, pfn, PB_migrate_end, in get_pfnblock_migratetype()
554 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument
556 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn()
986 static inline unsigned long early_pfn_to_nid(unsigned long pfn) in early_pfn_to_nid() argument
993 #define pfn_to_nid(pfn) (0) argument
1019 #define pfn_to_section_nr(pfn) ((pfn) >> PFN_SECTION_SHIFT) argument
1022 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK) argument
1023 #define SECTION_ALIGN_DOWN(pfn) ((pfn) & PAGE_SECTION_MASK) argument
1121 static inline struct mem_section *__pfn_to_section(unsigned long pfn) in __pfn_to_section() argument
[all …]
Dio-mapping.h88 unsigned long pfn; in io_mapping_map_atomic_wc() local
92 pfn = (unsigned long) (phys_addr >> PAGE_SHIFT); in io_mapping_map_atomic_wc()
93 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc()
Dcrash_dump.h22 unsigned long from, unsigned long pfn,
83 extern int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn));
Dpageblock-flags.h69 unsigned long pfn,
75 unsigned long pfn,
Dsuspend.h516 void page_key_read(unsigned long *pfn);
517 void page_key_memorize(unsigned long *pfn);
533 static inline void page_key_read(unsigned long *pfn) {} in page_key_read() argument
534 static inline void page_key_memorize(unsigned long *pfn) {} in page_key_memorize() argument
/linux-4.4.14/arch/ia64/include/asm/
Dpage.h98 extern int ia64_pfn_valid (unsigned long pfn);
100 # define ia64_pfn_valid(pfn) 1 argument
107 # define pfn_to_page(pfn) (vmem_map + (pfn)) argument
116 # define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn)) argument
120 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn)) argument
125 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
Dmmzone.h20 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
23 int nid = paddr_to_nid(pfn << PAGE_SHIFT); in pfn_to_nid()
/linux-4.4.14/arch/x86/kvm/
Diommu.c50 pfn_t pfn; in kvm_pin_pages() local
52 pfn = gfn_to_pfn_memslot(slot, gfn); in kvm_pin_pages()
56 if (is_error_noslot_pfn(pfn)) in kvm_pin_pages()
57 return pfn; in kvm_pin_pages()
62 return pfn; in kvm_pin_pages()
65 static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages) in kvm_unpin_pages() argument
70 kvm_release_pfn_clean(pfn + i); in kvm_unpin_pages()
76 pfn_t pfn; in kvm_iommu_map_pages() local
123 pfn = kvm_pin_pages(slot, gfn, page_size >> PAGE_SHIFT); in kvm_iommu_map_pages()
124 if (is_error_noslot_pfn(pfn)) { in kvm_iommu_map_pages()
[all …]
Dmmu.c262 pfn_t pfn, unsigned access) in set_mmio_spte() argument
264 if (unlikely(is_noslot_pfn(pfn))) { in set_mmio_spte()
590 pfn_t pfn; in mmu_spte_clear_track_bits() local
601 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits()
608 WARN_ON(!kvm_is_reserved_pfn(pfn) && !page_count(pfn_to_page(pfn))); in mmu_spte_clear_track_bits()
611 kvm_set_pfn_accessed(pfn); in mmu_spte_clear_track_bits()
613 kvm_set_pfn_dirty(pfn); in mmu_spte_clear_track_bits()
2479 static bool kvm_is_mmio_pfn(pfn_t pfn) in kvm_is_mmio_pfn() argument
2481 if (pfn_valid(pfn)) in kvm_is_mmio_pfn()
2482 return !is_zero_pfn(pfn) && PageReserved(pfn_to_page(pfn)); in kvm_is_mmio_pfn()
[all …]
Dmmu_audit.c100 pfn_t pfn; in audit_mappings() local
117 pfn = kvm_vcpu_gfn_to_pfn_atomic(vcpu, gfn); in audit_mappings()
119 if (is_error_pfn(pfn)) in audit_mappings()
122 hpa = pfn << PAGE_SHIFT; in audit_mappings()
125 "ent %llxn", vcpu->arch.mmu.root_level, pfn, in audit_mappings()
Dpaging_tmpl.h459 pfn_t pfn; in FNAME() local
469 pfn = pte_prefetch_gfn_to_pfn(vcpu, gfn, in FNAME()
471 if (is_error_pfn(pfn)) in FNAME()
479 gfn, pfn, true, true); in FNAME()
554 pfn_t pfn, bool map_writable, bool prefault) in FNAME()
626 it.level, gw->gfn, pfn, prefault, map_writable); in FNAME()
634 kvm_release_pfn_clean(pfn); in FNAME()
699 pfn_t pfn; in FNAME() local
757 if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, in FNAME()
762 walker.gfn, pfn, walker.pte_access, &r)) in FNAME()
[all …]
/linux-4.4.14/arch/tile/mm/
Dinit.c93 pte_t *get_prealloc_pte(unsigned long pfn) in get_prealloc_pte() argument
95 int node = pfn_to_nid(pfn); in get_prealloc_pte()
96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)); in get_prealloc_pte()
98 BUG_ON(pfn >= num_l2_ptes[node]); in get_prealloc_pte()
99 return &l2_ptes[node][pfn]; in get_prealloc_pte()
408 unsigned long address, pfn; in kernel_physical_mapping_init() local
453 pfn = node_start_pfn[i]; in kernel_physical_mapping_init()
456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init()
458 address = (unsigned long) pfn_to_kaddr(pfn); in kernel_physical_mapping_init()
459 while (pfn < end_pfn) { in kernel_physical_mapping_init()
[all …]
Dhighmem.c270 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument
272 return kmap_atomic(pfn_to_page(pfn)); in kmap_atomic_pfn()
274 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument
276 return kmap_atomic_prot(pfn_to_page(pfn), prot); in kmap_atomic_prot_pfn()
/linux-4.4.14/arch/x86/mm/
Dinit_32.c261 unsigned long pfn; in kernel_physical_mapping_init() local
292 pfn = start_pfn; in kernel_physical_mapping_init()
293 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init()
298 if (pfn >= end_pfn) in kernel_physical_mapping_init()
301 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init()
306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init()
308 unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; in kernel_physical_mapping_init()
325 pfn &= PMD_MASK >> PAGE_SHIFT; in kernel_physical_mapping_init()
326 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init()
335 set_pmd(pmd, pfn_pmd(pfn, init_prot)); in kernel_physical_mapping_init()
[all …]
Dpat.c423 u64 pfn; in reserve_ram_pages_type() local
437 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
440 page = pfn_to_page(pfn); in reserve_ram_pages_type()
455 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
456 page = pfn_to_page(pfn); in reserve_ram_pages_type()
465 u64 pfn; in free_ram_pages_type() local
467 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type()
468 page = pfn_to_page(pfn); in free_ram_pages_type()
689 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
697 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
[all …]
Dnuma_32.c49 unsigned long pfn; in memory_present() local
57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present()
59 printk(KERN_CONT "%lx ", pfn); in memory_present()
Dinit.c85 unsigned long pfn; in alloc_low_pages() local
106 pfn = ret >> PAGE_SHIFT; in alloc_low_pages()
108 pfn = pgt_buf_end; in alloc_low_pages()
111 pfn << PAGE_SHIFT, (pgt_buf_end << PAGE_SHIFT) - 1); in alloc_low_pages()
117 adr = __va((pfn + i) << PAGE_SHIFT); in alloc_low_pages()
121 return __va(pfn << PAGE_SHIFT); in alloc_low_pages()
269 unsigned long pfn; in split_mem_range() local
275 pfn = start_pfn = PFN_DOWN(start); in split_mem_range()
283 if (pfn == 0) in split_mem_range()
286 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
[all …]
Dpageattr.c38 unsigned long pfn; member
252 unsigned long pfn) in static_protections() argument
261 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) in static_protections()
277 if (within(pfn, __pa_symbol(__start_rodata) >> PAGE_SHIFT, in static_protections()
477 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn, old_pfn; in try_preserve_large_page() local
556 pfn = old_pfn + ((address & (psize - 1)) >> PAGE_SHIFT); in try_preserve_large_page()
557 cpa->pfn = pfn; in try_preserve_large_page()
559 new_prot = static_protections(req_prot, address, pfn); in try_preserve_large_page()
567 pfn = old_pfn; in try_preserve_large_page()
568 for (i = 0; i < (psize >> PAGE_SHIFT); i++, addr += PAGE_SIZE, pfn++) { in try_preserve_large_page()
[all …]
Diomap_32.c57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument
68 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); in kmap_atomic_prot_pfn()
78 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in iomap_atomic_prot_pfn() argument
91 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); in iomap_atomic_prot_pfn()
Dpageattr-test.c136 unsigned long pfn = prandom_u32() % max_pfn_mapped; in pageattr_test() local
138 addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT); in pageattr_test()
140 len[i] = min_t(unsigned long, len[i], max_pfn_mapped - pfn - 1); in pageattr_test()
164 if (test_bit(pfn + k, bm)) { in pageattr_test()
168 __set_bit(pfn + k, bm); in pageattr_test()
/linux-4.4.14/fs/proc/
Dpage.c34 unsigned long pfn; in kpagecount_read() local
38 pfn = src / KPMSIZE; in kpagecount_read()
44 if (pfn_valid(pfn)) in kpagecount_read()
45 ppage = pfn_to_page(pfn); in kpagecount_read()
58 pfn++; in kpagecount_read()
202 unsigned long pfn; in kpageflags_read() local
205 pfn = src / KPMSIZE; in kpageflags_read()
211 if (pfn_valid(pfn)) in kpageflags_read()
212 ppage = pfn_to_page(pfn); in kpageflags_read()
221 pfn++; in kpageflags_read()
[all …]
Dvmcore.c51 static int (*oldmem_pfn_is_ram)(unsigned long pfn);
53 int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn)) in register_oldmem_pfn_is_ram() argument
69 static int pfn_is_ram(unsigned long pfn) in pfn_is_ram() argument
71 int (*fn)(unsigned long pfn); in pfn_is_ram()
82 ret = fn(pfn); in pfn_is_ram()
91 unsigned long pfn, offset; in read_from_oldmem() local
99 pfn = (unsigned long)(*ppos / PAGE_SIZE); in read_from_oldmem()
108 if (pfn_is_ram(pfn) == 0) in read_from_oldmem()
111 tmp = copy_oldmem_page(pfn, buf, nr_bytes, in read_from_oldmem()
120 ++pfn; in read_from_oldmem()
[all …]
/linux-4.4.14/arch/frv/include/asm/
Dpage.h43 #define devmem_is_allowed(pfn) 1 argument
48 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
55 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
58 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
/linux-4.4.14/arch/x86/kernel/
Dcrash_dump_32.c17 static inline bool is_crashed_pfn_valid(unsigned long pfn) in is_crashed_pfn_valid() argument
27 return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn; in is_crashed_pfn_valid()
50 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
58 if (!is_crashed_pfn_valid(pfn)) in copy_oldmem_page()
61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
Dcrash_dump_64.c26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
/linux-4.4.14/arch/powerpc/kvm/
De500_mmu_host.c166 pfn_t pfn; in kvmppc_map_magic() local
168 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic()
169 get_page(pfn_to_page(pfn)); in kvmppc_map_magic()
177 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_map_magic()
249 pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() argument
251 ref->pfn = pfn; in kvmppc_e500_ref_setup()
258 kvm_set_pfn_accessed(pfn); in kvmppc_e500_ref_setup()
261 kvm_set_pfn_dirty(pfn); in kvmppc_e500_ref_setup()
268 trace_kvm_booke206_ref_release(ref->pfn, ref->flags); in kvmppc_e500_ref_release()
312 pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe() local
[all …]
Dtrace_booke.h137 TP_PROTO(__u64 pfn, __u32 flags),
138 TP_ARGS(pfn, flags),
141 __field( __u64, pfn )
146 __entry->pfn = pfn;
151 __entry->pfn, __entry->flags)
Dbook3s_64_mmu_host.c101 unsigned long pfn; in kvmppc_mmu_map_page() local
108 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page()
109 if (is_error_noslot_pfn(pfn)) { in kvmppc_mmu_map_page()
115 hpaddr = pfn << PAGE_SHIFT; in kvmppc_mmu_map_page()
135 kvm_set_pfn_accessed(pfn); in kvmppc_mmu_map_page()
140 kvm_set_pfn_dirty(pfn); in kvmppc_mmu_map_page()
146 kvmppc_mmu_flush_icache(pfn); in kvmppc_mmu_map_page()
200 cpte->pfn = pfn; in kvmppc_mmu_map_page()
209 kvm_release_pfn_clean(pfn); in kvmppc_mmu_map_page()
Dtrace_pr.h70 __field( u64, pfn )
79 __entry->pfn = pte->pfn;
89 __entry->host_vpn, __entry->pfn, __entry->eaddr,
99 __field( u64, pfn )
108 __entry->pfn = pte->pfn;
118 __entry->host_vpn, __entry->pfn, __entry->eaddr,
/linux-4.4.14/arch/s390/pci/
Dpci_mmio.c15 unsigned long *pfn) in get_pfn() argument
28 ret = follow_pfn(vma, user_addr, pfn); in get_pfn()
40 unsigned long pfn; in SYSCALL_DEFINE3() local
55 ret = get_pfn(mmio_addr, VM_WRITE, &pfn); in SYSCALL_DEFINE3()
58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
80 unsigned long pfn; in SYSCALL_DEFINE3() local
95 ret = get_pfn(mmio_addr, VM_READ, &pfn); in SYSCALL_DEFINE3()
98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
/linux-4.4.14/arch/tile/include/asm/
Dpage.h155 #define __pfn_to_highbits(pfn) ((pfn) >> (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)) argument
255 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument
257 return vbase_map[__pfn_to_highbits(pfn)] + (pfn << PAGE_SHIFT); in pfn_to_kaddr()
262 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys() local
263 return ((phys_addr_t)pfn << PAGE_SHIFT) + in virt_to_phys()
286 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument
288 return (void *)((pfn << PAGE_SHIFT) + PAGE_OFFSET); in pfn_to_kaddr()
313 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
315 return pfn < max_mapnr; in pfn_valid()
Dmmzone.h38 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
40 return highbits_to_node[__pfn_to_highbits(pfn)]; in pfn_to_nid()
45 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
47 int nid = pfn_to_nid(pfn); in pfn_valid()
50 return (pfn < node_end_pfn(nid)); in pfn_valid()
Dhighmem.h64 void *kmap_atomic_pfn(unsigned long pfn);
65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
/linux-4.4.14/arch/microblaze/include/asm/
Dpage.h142 extern int page_is_ram(unsigned long pfn);
145 # define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument
148 # define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument
165 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && \ argument
166 ((pfn) <= (min_low_pfn + max_mapnr)))
170 # define pfn_valid(pfn) ((pfn) < (max_mapnr + ARCH_PFN_OFFSET)) argument
Dcacheflush.h92 #define flush_cache_page(vma, vmaddr, pfn) \ argument
93 flush_dcache_range(pfn << PAGE_SHIFT, (pfn << PAGE_SHIFT) + PAGE_SIZE);
/linux-4.4.14/arch/arm/include/asm/
Ddma-mapping.h61 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument
64 pfn -= dev->dma_pfn_offset; in pfn_to_dma()
65 return (dma_addr_t)__pfn_to_bus(pfn); in pfn_to_dma()
70 unsigned long pfn = __bus_to_pfn(addr); in dma_to_pfn() local
73 pfn += dev->dma_pfn_offset; in dma_to_pfn()
75 return pfn; in dma_to_pfn()
81 unsigned long pfn = dma_to_pfn(dev, addr); in dma_to_virt() local
83 return phys_to_virt(__pfn_to_phys(pfn)); in dma_to_virt()
98 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument
100 return __arch_pfn_to_dma(dev, pfn); in pfn_to_dma()
Dkvm_mmu.h185 static inline void __coherent_cache_guest_page(struct kvm_vcpu *vcpu, pfn_t pfn, in __coherent_cache_guest_page() argument
215 void *va = kmap_atomic_pfn(pfn); in __coherent_cache_guest_page()
225 pfn++; in __coherent_cache_guest_page()
249 pfn_t pfn = pmd_pfn(pmd); in __kvm_flush_dcache_pmd() local
252 void *va = kmap_atomic_pfn(pfn); in __kvm_flush_dcache_pmd()
256 pfn++; in __kvm_flush_dcache_pmd()
/linux-4.4.14/arch/metag/include/asm/
Dmmzone.h10 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
15 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid()
21 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
23 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
Dpage.h115 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_pfn) argument
118 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.4.14/arch/m68k/include/asm/
Dpage_no.h26 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
31 #define pfn_to_page(pfn) virt_to_page(pfn_to_virt(pfn)) argument
33 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
Dpage_mm.h127 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
158 #define pfn_to_page(pfn) ({ \ argument
159 unsigned long __pfn = (pfn); \
172 #define pfn_valid(pfn) virt_addr_valid(pfn_to_virt(pfn)) argument
/linux-4.4.14/arch/sh/include/asm/
Dmmzone.h12 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
17 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid()
23 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
25 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
Dpage.h167 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
180 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
/linux-4.4.14/arch/powerpc/mm/
Dmem.c79 int page_is_ram(unsigned long pfn) in page_is_ram() argument
82 return pfn < max_pfn; in page_is_ram()
84 unsigned long paddr = (pfn << PAGE_SHIFT); in page_is_ram()
94 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
98 return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); in phys_mem_access_prot()
100 if (!page_is_ram(pfn)) in phys_mem_access_prot()
341 unsigned long pfn, highmem_mapnr; in mem_init() local
344 for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { in mem_init()
345 phys_addr_t paddr = (phys_addr_t)pfn << PAGE_SHIFT; in mem_init()
346 struct page *page = pfn_to_page(pfn); in mem_init()
[all …]
/linux-4.4.14/drivers/base/
Dnode.c362 static int __init_refok get_nid_for_pfn(unsigned long pfn) in get_nid_for_pfn() argument
366 if (!pfn_valid_within(pfn)) in get_nid_for_pfn()
370 return early_pfn_to_nid(pfn); in get_nid_for_pfn()
372 page = pfn_to_page(pfn); in get_nid_for_pfn()
375 return pfn_to_nid(pfn); in get_nid_for_pfn()
382 unsigned long pfn, sect_start_pfn, sect_end_pfn; in register_mem_sect_under_node() local
392 for (pfn = sect_start_pfn; pfn <= sect_end_pfn; pfn++) { in register_mem_sect_under_node()
399 if (!pfn_present(pfn)) { in register_mem_sect_under_node()
400 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node()
405 page_nid = get_nid_for_pfn(pfn); in register_mem_sect_under_node()
[all …]
Dmemory.c127 unsigned long i, pfn; in show_mem_removable() local
134 pfn = section_nr_to_pfn(mem->start_section_nr + i); in show_mem_removable()
135 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable()
192 unsigned long pfn = start_pfn; in pages_correctly_reserved() local
199 for (i = 0; i < sections_per_block; i++, pfn += PAGES_PER_SECTION) { in pages_correctly_reserved()
200 if (WARN_ON_ONCE(!pfn_valid(pfn))) in pages_correctly_reserved()
202 page = pfn_to_page(pfn); in pages_correctly_reserved()
210 pfn_to_section_nr(pfn), j); in pages_correctly_reserved()
494 u64 pfn; in store_soft_offline_page() local
497 if (kstrtoull(buf, 0, &pfn) < 0) in store_soft_offline_page()
[all …]
Ddma-mapping.c253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap() local
263 pfn + off, in dma_common_mmap()
310 unsigned long pfn; in dma_common_contiguous_remap() local
316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap()
317 pages[i] = pfn_to_page(pfn + i); in dma_common_contiguous_remap()
/linux-4.4.14/arch/score/include/asm/
Dpage.h65 #define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument
68 #define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument
77 #define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn)) argument
/linux-4.4.14/arch/arm/mach-omap1/
Dio.c34 .pfn = __phys_to_pfn(OMAP1_IO_PHYS),
44 .pfn = __phys_to_pfn(OMAP7XX_DSP_START),
49 .pfn = __phys_to_pfn(OMAP7XX_DSPREG_START),
60 .pfn = __phys_to_pfn(OMAP1510_DSP_START),
65 .pfn = __phys_to_pfn(OMAP1510_DSPREG_START),
76 .pfn = __phys_to_pfn(OMAP16XX_DSP_START),
81 .pfn = __phys_to_pfn(OMAP16XX_DSPREG_START),
/linux-4.4.14/drivers/xen/
Dballoon.c276 unsigned long pfn = res->start >> PAGE_SHIFT; in additional_memory_resource() local
278 if (pfn > limit) { in additional_memory_resource()
280 pfn, limit); in additional_memory_resource()
345 unsigned long pfn, i; in reserve_additional_memory() local
347 pfn = PFN_DOWN(resource->start); in reserve_additional_memory()
349 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) { in reserve_additional_memory()
458 unsigned long pfn = page_to_pfn(page); in increase_reservation() local
460 set_phys_to_machine(pfn, frame_list[i]); in increase_reservation()
466 (unsigned long)__va(pfn << PAGE_SHIFT), in increase_reservation()
536 unsigned long pfn = page_to_pfn(page); in decrease_reservation() local
[all …]
Dgrant-table.c394 int gnttab_grant_foreign_transfer(domid_t domid, unsigned long pfn) in gnttab_grant_foreign_transfer() argument
401 gnttab_grant_foreign_transfer_ref(ref, domid, pfn); in gnttab_grant_foreign_transfer()
408 unsigned long pfn) in gnttab_grant_foreign_transfer_ref() argument
410 gnttab_interface->update_entry(ref, domid, pfn, GTF_accept_transfer); in gnttab_grant_foreign_transfer_ref()
637 xen_pfn_t *pfn; in gnttab_setup_auto_xlat_frames() local
651 pfn = kcalloc(max_nr_gframes, sizeof(pfn[0]), GFP_KERNEL); in gnttab_setup_auto_xlat_frames()
652 if (!pfn) { in gnttab_setup_auto_xlat_frames()
657 pfn[i] = XEN_PFN_DOWN(addr) + i; in gnttab_setup_auto_xlat_frames()
660 xen_auto_xlat_grant_frames.pfn = pfn; in gnttab_setup_auto_xlat_frames()
671 kfree(xen_auto_xlat_grant_frames.pfn); in gnttab_free_auto_xlat_frames()
[all …]
/linux-4.4.14/arch/parisc/kernel/
Dcache.c75 #define pfn_va(pfn) __va(PFN_PHYS(pfn)) argument
80 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
86 if (!pfn_valid(pfn)) in update_mmu_cache()
89 page = pfn_to_page(pfn); in update_mmu_cache()
91 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache()
94 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache()
542 unsigned long pfn; in flush_cache_mm() local
546 pfn = pte_pfn(*ptep); in flush_cache_mm()
547 if (!pfn_valid(pfn)) in flush_cache_mm()
549 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm()
[all …]
/linux-4.4.14/arch/arm64/mm/
Dmmap.c114 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
116 return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK); in valid_mmap_phys_addr_range()
129 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
131 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
133 if (!page_is_ram(pfn)) in devmem_is_allowed()
/linux-4.4.14/drivers/vfio/
Dvfio_iommu_type1.c191 static bool is_invalid_reserved_pfn(unsigned long pfn) in is_invalid_reserved_pfn() argument
193 if (pfn_valid(pfn)) { in is_invalid_reserved_pfn()
195 struct page *tail = pfn_to_page(pfn); in is_invalid_reserved_pfn()
219 static int put_pfn(unsigned long pfn, int prot) in put_pfn() argument
221 if (!is_invalid_reserved_pfn(pfn)) { in put_pfn()
222 struct page *page = pfn_to_page(pfn); in put_pfn()
231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument
238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn()
247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn()
248 if (is_invalid_reserved_pfn(*pfn)) in vaddr_get_pfn()
[all …]
/linux-4.4.14/arch/mn10300/include/asm/
Dpage.h107 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
108 #define pfn_to_page(pfn) (mem_map + ((pfn) - __pfn_disp)) argument
111 #define pfn_valid(pfn) \ argument
113 unsigned long __pfn = (pfn) - __pfn_disp; \
Dpgtable.h317 #define __pfn_addr(pfn) ((pfn) << PAGE_SHIFT) argument
318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot)) argument
319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot)) argument
480 #define MK_IOSPACE_PFN(space, pfn) (pfn) argument
481 #define GET_IOSPACE(pfn) 0 argument
482 #define GET_PFN(pfn) (pfn) argument
/linux-4.4.14/arch/arm/mach-footbridge/
Dcommon.c143 .pfn = __phys_to_pfn(DC21285_ARMCSR_BASE),
157 .pfn = __phys_to_pfn(DC21285_PCI_MEM),
162 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_0_CONFIG),
167 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_1_CONFIG),
172 .pfn = __phys_to_pfn(DC21285_PCI_IACK),
256 unsigned long __pfn_to_bus(unsigned long pfn) in __pfn_to_bus() argument
258 return __pfn_to_phys(pfn) + (fb_bus_sdram_offset() - PHYS_OFFSET); in __pfn_to_bus()
/linux-4.4.14/arch/arm/mach-exynos/
Dexynos.c41 .pfn = __phys_to_pfn(EXYNOS4_PA_SROMC),
46 .pfn = __phys_to_pfn(EXYNOS4_PA_CMU),
51 .pfn = __phys_to_pfn(EXYNOS4_PA_COREPERI),
56 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC0),
61 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC1),
70 .pfn = __phys_to_pfn(EXYNOS5_PA_SROMC),
75 .pfn = __phys_to_pfn(EXYNOS5_PA_CMU),
135 iodesc.pfn = __phys_to_pfn(be32_to_cpu(reg[0])); in exynos_fdt_map_chipid()
/linux-4.4.14/arch/frv/mm/
Dinit.c123 unsigned long pfn; in mem_init() local
125 for (pfn = get_num_physpages() - 1; in mem_init()
126 pfn >= num_mappedpages; pfn--) in mem_init()
127 free_highmem_page(&mem_map[pfn]); in mem_init()
/linux-4.4.14/arch/arm/mach-tegra/
Dio.c36 .pfn = __phys_to_pfn(IO_PPSB_PHYS),
42 .pfn = __phys_to_pfn(IO_APB_PHYS),
48 .pfn = __phys_to_pfn(IO_CPU_PHYS),
54 .pfn = __phys_to_pfn(IO_IRAM_PHYS),
/linux-4.4.14/arch/arm/kvm/
Dmmu.c101 static bool kvm_is_device_pfn(unsigned long pfn) in kvm_is_device_pfn() argument
103 return !pfn_valid(pfn); in kvm_is_device_pfn()
452 unsigned long end, unsigned long pfn, in create_hyp_pte_mappings() argument
461 kvm_set_pte(pte, pfn_pte(pfn, prot)); in create_hyp_pte_mappings()
464 pfn++; in create_hyp_pte_mappings()
469 unsigned long end, unsigned long pfn, in create_hyp_pmd_mappings() argument
495 create_hyp_pte_mappings(pmd, addr, next, pfn, prot); in create_hyp_pmd_mappings()
496 pfn += (next - addr) >> PAGE_SHIFT; in create_hyp_pmd_mappings()
503 unsigned long end, unsigned long pfn, in create_hyp_pud_mappings() argument
527 ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); in create_hyp_pud_mappings()
[all …]
/linux-4.4.14/arch/avr32/include/asm/
Dpgtable-2level.h42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
Dpage.h88 #define pfn_to_page(pfn) (mem_map + ((pfn) - PHYS_PFN_OFFSET)) argument
90 #define pfn_valid(pfn) ((pfn) >= PHYS_PFN_OFFSET && (pfn) < (PHYS_PFN_OFFSET + max_mapnr)) argument
/linux-4.4.14/arch/mips/include/asm/
Dpage.h195 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
199 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
205 return pfn >= pfn_offset && pfn < max_mapnr; in pfn_valid()
214 #define pfn_valid(pfn) \ argument
216 unsigned long __pfn = (pfn); \
Dpgtable-32.h110 pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument
114 pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | in pfn_pte()
116 pte.pte_high = (pfn << _PFN_SHIFT) | in pfn_pte()
127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument
130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
Dpgtable-64.h229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument
232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
Dmmzone.h13 #define pfn_to_nid(pfn) pa_to_nid((pfn) << PAGE_SHIFT) argument
/linux-4.4.14/arch/microblaze/mm/
Dinit.c76 unsigned long pfn; in highmem_setup() local
78 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { in highmem_setup()
79 struct page *page = pfn_to_page(pfn); in highmem_setup()
82 if (!memblock_is_reserved(pfn << PAGE_SHIFT)) in highmem_setup()
267 int page_is_ram(unsigned long pfn) in page_is_ram() argument
269 return __range_ok(pfn, 0); in page_is_ram()
272 int page_is_ram(unsigned long pfn) in page_is_ram() argument
274 return pfn < max_low_pfn; in page_is_ram()
Dconsistent.c204 unsigned long pfn; in consistent_free() local
207 pfn = pte_pfn(*ptep); in consistent_free()
209 if (pfn_valid(pfn)) { in consistent_free()
210 page = pfn_to_page(pfn); in consistent_free()
/linux-4.4.14/arch/arm/mach-realview/
Drealview_pbx.c53 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
58 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_CPU_BASE),
63 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_DIST_BASE),
68 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
73 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER0_1_BASE),
78 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER2_3_BASE),
85 .pfn = __phys_to_pfn(REALVIEW_PBX_PCI_BASE),
93 .pfn = __phys_to_pfn(REALVIEW_PBX_UART0_BASE),
103 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_SCU_BASE),
108 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_GIC_DIST_BASE),
[all …]
Drealview_pb1176.c56 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
61 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_CPU_BASE),
66 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_DIST_BASE),
71 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_CPU_BASE),
76 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_DIST_BASE),
81 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
86 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER0_1_BASE),
91 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER2_3_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_PB1176_L220_BASE),
103 .pfn = __phys_to_pfn(REALVIEW_PB1176_UART0_BASE),
Drealview_pba8.c51 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
56 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_CPU_BASE),
61 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_DIST_BASE),
66 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
71 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER0_1_BASE),
76 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER2_3_BASE),
83 .pfn = __phys_to_pfn(REALVIEW_PBA8_PCI_BASE),
91 .pfn = __phys_to_pfn(REALVIEW_PBA8_UART0_BASE),
Drealview_pb11mp.c54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
59 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_CPU_BASE),
64 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_DIST_BASE),
69 .pfn = __phys_to_pfn(REALVIEW_TC11MP_PRIV_MEM_BASE),
74 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
79 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER0_1_BASE),
84 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER2_3_BASE),
89 .pfn = __phys_to_pfn(REALVIEW_TC11MP_L220_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_PB11MP_UART0_BASE),
Drealview_eb.c54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
59 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_CPU_BASE),
64 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_DIST_BASE),
69 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
74 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER0_1_BASE),
79 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER2_3_BASE),
86 .pfn = __phys_to_pfn(REALVIEW_EB_UART0_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_EB11MP_PRIV_MEM_BASE),
101 .pfn = __phys_to_pfn(REALVIEW_EB11MP_L220_BASE),
/linux-4.4.14/arch/um/include/asm/
Dpgtable-2level.h41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) argument
42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot)) argument
Dpage.h113 #define pfn_to_phys(pfn) ((phys_t) ((pfn) << PAGE_SHIFT)) argument
115 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
/linux-4.4.14/arch/nios2/include/asm/
Dpage.h89 # define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
90 # define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && \ argument
91 (pfn) < max_mapnr)
/linux-4.4.14/arch/powerpc/kernel/
Ddma.c32 u64 pfn = (dev->coherent_dma_mask >> PAGE_SHIFT) + 1; in get_pfn_limit() local
37 pfn = min_t(u64, pfn, sd->max_direct_dma_addr >> PAGE_SHIFT); in get_pfn_limit()
40 return pfn; in get_pfn_limit()
80 u64 pfn = get_pfn_limit(dev); in __dma_direct_alloc_coherent() local
89 zone = dma_pfn_limit_to_zone(pfn); in __dma_direct_alloc_coherent()
92 __func__, pfn); in __dma_direct_alloc_coherent()
182 unsigned long pfn; in dma_direct_mmap_coherent() local
186 pfn = __dma_get_coherent_pfn((unsigned long)cpu_addr); in dma_direct_mmap_coherent()
188 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent()
191 pfn + vma->vm_pgoff, in dma_direct_mmap_coherent()
Dsuspend.c18 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
22 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.4.14/arch/sparc/include/asm/
Dpgtable_32.h262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) argument
405 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument
406 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument
407 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) argument
413 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument
418 offset = ((unsigned long long) GET_PFN(pfn)) << PAGE_SHIFT; in io_remap_pfn_range()
419 space = GET_IOSPACE(pfn); in io_remap_pfn_range()
Dpgtable_64.h237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument
239 unsigned long paddr = pfn << PAGE_SHIFT; in pfn_pte()
817 unsigned long pfn; in __pmd_page() local
819 pfn = pte_pfn(pte); in __pmd_page()
821 return ((unsigned long) __va(pfn << PAGE_SHIFT)); in __pmd_page()
987 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument
988 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument
989 #define GET_PFN(pfn) (pfn & 0x0fffffffffffffffUL) argument
995 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument
998 unsigned long offset = GET_PFN(pfn) << PAGE_SHIFT; in io_remap_pfn_range()
[all …]
Dpage_32.h132 #define pfn_valid(pfn) (((pfn) >= (pfn_base)) && (((pfn)-(pfn_base)) < max_mapnr)) argument
/linux-4.4.14/arch/arm/mach-cns3xxx/
Dcore.c32 .pfn = __phys_to_pfn(CNS3XXX_TC11MP_SCU_BASE),
37 .pfn = __phys_to_pfn(CNS3XXX_TIMER1_2_3_BASE),
42 .pfn = __phys_to_pfn(CNS3XXX_MISC_BASE),
47 .pfn = __phys_to_pfn(CNS3XXX_PM_BASE),
53 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_HOST_BASE),
58 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG0_BASE),
63 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG1_BASE),
68 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_HOST_BASE),
73 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG0_BASE),
78 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG1_BASE),
/linux-4.4.14/arch/arm/mach-s3c64xx/
Dcommon.c108 .pfn = __phys_to_pfn(S3C64XX_PA_SYSCON),
113 .pfn = __phys_to_pfn(S3C64XX_PA_SROM),
118 .pfn = __phys_to_pfn(S3C_PA_UART),
123 .pfn = __phys_to_pfn(S3C64XX_PA_VIC0),
128 .pfn = __phys_to_pfn(S3C64XX_PA_VIC1),
133 .pfn = __phys_to_pfn(S3C_PA_TIMER),
138 .pfn = __phys_to_pfn(S3C64XX_PA_GPIO),
143 .pfn = __phys_to_pfn(S3C64XX_PA_MODEM),
148 .pfn = __phys_to_pfn(S3C64XX_PA_WATCHDOG),
153 .pfn = __phys_to_pfn(S3C64XX_PA_USB_HSPHY),
/linux-4.4.14/arch/arm/mach-integrator/
Dintegrator_cp.c69 .pfn = __phys_to_pfn(INTEGRATOR_HDR_BASE),
74 .pfn = __phys_to_pfn(INTEGRATOR_CT_BASE),
79 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE),
84 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE),
89 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE),
94 .pfn = __phys_to_pfn(INTEGRATOR_CP_GPIO_BASE),
99 .pfn = __phys_to_pfn(INTEGRATOR_CP_SIC_BASE),
Dintegrator_ap.c85 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE),
90 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE),
95 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE),
100 .pfn = __phys_to_pfn(INTEGRATOR_AP_GPIO_BASE),
/linux-4.4.14/arch/sh/kernel/
Dswsusp.c22 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
Dcrash_dump.c25 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
/linux-4.4.14/arch/sparc/power/
Dhibernate.c21 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
26 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.4.14/arch/arm/mach-spear/
Dspear13xx.c63 .pfn = __phys_to_pfn(PERIP_GRP2_BASE),
68 .pfn = __phys_to_pfn(PERIP_GRP1_BASE),
73 .pfn = __phys_to_pfn(A9SM_AND_MPMC_BASE),
78 .pfn = __phys_to_pfn(L2CC_BASE),
/linux-4.4.14/arch/powerpc/include/asm/
Dpage.h129 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && (pfn) < max_mapnr) argument
134 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
411 extern int page_is_ram(unsigned long pfn);
412 extern int devmem_is_allowed(unsigned long pfn);
Dpte-hash64-64k.h97 #define remap_4k_pfn(vma, addr, pfn, prot) \ argument
98 (WARN_ON(((pfn) >= (1UL << (64 - PTE_RPN_SHIFT)))) ? -EINVAL : \
99 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, \
/linux-4.4.14/arch/mips/power/
Dcpu.c37 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
42 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.4.14/arch/arm/mach-s3c24xx/
Dmach-smdk2443.c53 .pfn = __phys_to_pfn(S3C2410_CS2),
58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
63 .pfn = __phys_to_pfn(S3C2410_CS2),
68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
Dmach-smdk2440.c53 .pfn = __phys_to_pfn(S3C2410_CS2),
58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
63 .pfn = __phys_to_pfn(S3C2410_CS2),
68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
Dmach-bast.c83 .pfn = PA_CS2(BAST_PA_ISAIO),
88 .pfn = PA_CS3(BAST_PA_ISAIO),
95 .pfn = __phys_to_pfn(BAST_PA_CTRL1),
100 .pfn = __phys_to_pfn(BAST_PA_CTRL2),
105 .pfn = __phys_to_pfn(BAST_PA_CTRL3),
110 .pfn = __phys_to_pfn(BAST_PA_CTRL4),
117 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQREQ),
122 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQRAW),
127 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQMASK),
Dmach-vr1000.c72 .pfn = PA_CS2(BAST_PA_ISAIO),
77 .pfn = PA_CS3(BAST_PA_ISAIO),
85 .pfn = __phys_to_pfn(VR1000_PA_CTRL1),
90 .pfn = __phys_to_pfn(VR1000_PA_CTRL2),
95 .pfn = __phys_to_pfn(VR1000_PA_CTRL3),
100 .pfn = __phys_to_pfn(VR1000_PA_CTRL4),
Dmach-osiris.c65 .pfn = __phys_to_pfn(S3C2410_CS5),
70 .pfn = __phys_to_pfn(S3C2410_CS5),
79 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL0),
84 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL1),
89 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL2),
94 .pfn = __phys_to_pfn(OSIRIS_PA_IDREG),
Dmach-smdk2416.c65 .pfn = __phys_to_pfn(S3C2410_CS2),
70 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
75 .pfn = __phys_to_pfn(S3C2410_CS2),
80 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
/linux-4.4.14/arch/arm/mach-ebsa110/
Dcore.c78 .pfn = __phys_to_pfn(TRICK4_PHYS),
83 .pfn = __phys_to_pfn(TRICK3_PHYS),
88 .pfn = __phys_to_pfn(TRICK1_PHYS),
93 .pfn = __phys_to_pfn(TRICK0_PHYS),
103 .pfn = __phys_to_pfn(ISAIO_PHYS),
108 .pfn = __phys_to_pfn(ISAMEM_PHYS),
/linux-4.4.14/arch/score/mm/
Dcache.c82 unsigned long pfn, addr; in __update_cache() local
85 pfn = pte_pfn(pte); in __update_cache()
86 if (unlikely(!pfn_valid(pfn))) in __update_cache()
88 page = pfn_to_page(pfn); in __update_cache()
212 unsigned long addr, unsigned long pfn) in flush_cache_page() argument
215 unsigned long kaddr = 0xa0000000 | (pfn << PAGE_SHIFT); in flush_cache_page()
/linux-4.4.14/arch/openrisc/include/asm/
Dpage.h83 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
92 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
/linux-4.4.14/arch/tile/gxio/
Dkiorpc.c34 unsigned long pfn; in iorpc_ioremap() local
56 pfn = pte_pfn(mmio_base) + PFN_DOWN(offset); in iorpc_ioremap()
58 return ioremap_prot(PFN_PHYS(pfn), size, prot); in iorpc_ioremap()
/linux-4.4.14/arch/unicore32/include/asm/
Dio.h63 static inline int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
65 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
67 if (!page_is_ram(pfn)) in devmem_is_allowed()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Dmem.c56 u64 pfn; in mlx5_ib_cont_pages() local
80 pfn = sg_dma_address(sg) >> page_shift; in mlx5_ib_cont_pages()
83 tmp = (unsigned long)pfn; in mlx5_ib_cont_pages()
87 base = pfn; in mlx5_ib_cont_pages()
90 if (base + p != pfn) { in mlx5_ib_cont_pages()
95 base = pfn; in mlx5_ib_cont_pages()
/linux-4.4.14/arch/hexagon/include/asm/
Dpage.h112 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
144 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.4.14/arch/arm/kernel/
Dhibernate.c27 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
32 return (pfn >= nosave_begin_pfn) && (pfn <= nosave_end_pfn); in pfn_is_nosave()
Dcrash_dump.c33 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page()
/linux-4.4.14/drivers/misc/
Dvmw_balloon.c445 static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_lock_page() argument
451 pfn32 = (u32)pfn; in vmballoon_send_lock_page()
452 if (pfn32 != pfn) in vmballoon_send_lock_page()
457 *hv_status = status = VMWARE_BALLOON_CMD(LOCK, pfn, dummy, *target); in vmballoon_send_lock_page()
461 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_lock_page()
470 unsigned long pfn = page_to_pfn(b->page); in vmballoon_send_batched_lock() local
475 status = VMWARE_BALLOON_CMD(BATCHED_2M_LOCK, pfn, num_pages, in vmballoon_send_batched_lock()
478 status = VMWARE_BALLOON_CMD(BATCHED_LOCK, pfn, num_pages, in vmballoon_send_batched_lock()
484 pr_debug("%s - batch ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_batched_lock()
493 static bool vmballoon_send_unlock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_unlock_page() argument
[all …]
/linux-4.4.14/arch/arm/mach-ep93xx/
Dts72xx.c35 .pfn = __phys_to_pfn(TS72XX_MODEL_PHYS_BASE),
40 .pfn = __phys_to_pfn(TS72XX_OPTIONS_PHYS_BASE),
45 .pfn = __phys_to_pfn(TS72XX_OPTIONS2_PHYS_BASE),
50 .pfn = __phys_to_pfn(TS72XX_RTC_INDEX_PHYS_BASE),
55 .pfn = __phys_to_pfn(TS72XX_RTC_DATA_PHYS_BASE),
/linux-4.4.14/drivers/char/
Dmem.c55 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
62 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
64 u64 from = ((u64)pfn) << PAGE_SHIFT; in range_is_allowed()
69 if (!devmem_is_allowed(pfn)) { in range_is_allowed()
76 pfn++; in range_is_allowed()
81 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
224 unsigned long pfn, unsigned long size, pgprot_t *vma_prot) in phys_mem_access_prot_allowed() argument
265 static pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
269 phys_addr_t offset = pfn << PAGE_SHIFT; in phys_mem_access_prot()
357 unsigned long pfn; in mmap_kmem() local
[all …]
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.c628 static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) in ehea_update_busmap() argument
641 start_section = (pfn * PAGE_SIZE) / EHEA_SECTSIZE; in ehea_update_busmap()
671 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_add_sect_bmap() argument
676 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); in ehea_add_sect_bmap()
681 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_rem_sect_bmap() argument
686 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); in ehea_rem_sect_bmap()
691 static int ehea_is_hugepage(unsigned long pfn) in ehea_is_hugepage() argument
695 if (pfn & EHEA_HUGEPAGE_PFN_MASK) in ehea_is_hugepage()
698 page_order = compound_order(pfn_to_page(pfn)); in ehea_is_hugepage()
709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local
[all …]
/linux-4.4.14/arch/arm/mach-sa1100/
Dnanoengine.c70 .pfn = __phys_to_pfn(0x10000000),
76 .pfn = __phys_to_pfn(NANO_PCI_MEM_RW_PHYS),
82 .pfn = __phys_to_pfn(NANO_PCI_CONFIG_SPACE_PHYS),
/linux-4.4.14/lib/
Dioremap.c57 u64 pfn; in ioremap_pte_range() local
59 pfn = phys_addr >> PAGE_SHIFT; in ioremap_pte_range()
65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range()
66 pfn++; in ioremap_pte_range()
/linux-4.4.14/arch/mips/include/asm/mach-ip27/
Dkernel-entry-init.h51 dsrl t1, 12 # 4K pfn
52 dsrl t2, 12 # 4K pfn
53 dsll t1, 6 # Get pfn into place
54 dsll t2, 6 # Get pfn into place
/linux-4.4.14/arch/mips/mm/
Dcache.c31 unsigned long pfn);
131 unsigned long pfn, addr; in __update_cache() local
134 pfn = pte_pfn(pte); in __update_cache()
135 if (unlikely(!pfn_valid(pfn))) in __update_cache()
137 page = pfn_to_page(pfn); in __update_cache()
Dioremap.c23 unsigned long pfn; in remap_area_pte() local
32 pfn = phys_addr >> PAGE_SHIFT; in remap_area_pte()
38 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte()
40 pfn++; in remap_area_pte()
/linux-4.4.14/tools/testing/selftests/vm/
Dtranshuge-stress.c110 int64_t pfn; in main() local
112 pfn = allocate_transhuge(p); in main()
114 if (pfn < 0) { in main()
117 size_t idx = pfn >> (HPAGE_SHIFT - PAGE_SHIFT); in main()
/linux-4.4.14/arch/xtensa/mm/
Dinit.c238 unsigned long pfn; in bootmem_init() local
258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; in bootmem_init()
259 if (pfn < min_low_pfn) in bootmem_init()
260 min_low_pfn = pfn; in bootmem_init()
261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; in bootmem_init()
262 if (pfn > max_pfn) in bootmem_init()
263 max_pfn = pfn; in bootmem_init()
Dcache.c200 unsigned long pfn) in local_flush_cache_page() argument
204 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in local_flush_cache_page()
216 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
219 if (!pfn_valid(pfn)) in update_mmu_cache()
222 page = pfn_to_page(pfn); in update_mmu_cache()
/linux-4.4.14/include/ras/
Dras_event.h295 TP_PROTO(unsigned long pfn,
299 TP_ARGS(pfn, type, result),
302 __field(unsigned long, pfn)
308 __entry->pfn = pfn;
314 __entry->pfn,
/linux-4.4.14/arch/arm/mach-lpc32xx/
Dcommon.c168 .pfn = __phys_to_pfn(LPC32XX_AHB0_START),
174 .pfn = __phys_to_pfn(LPC32XX_AHB1_START),
180 .pfn = __phys_to_pfn(LPC32XX_FABAPB_START),
186 .pfn = __phys_to_pfn(LPC32XX_IRAM_BASE),
/linux-4.4.14/arch/s390/include/asm/
Dnuma.h18 int numa_pfn_to_nid(unsigned long pfn);
29 static inline int numa_pfn_to_nid(unsigned long pfn) in numa_pfn_to_nid() argument
/linux-4.4.14/drivers/gpu/drm/gma500/
Dgem.c179 unsigned long pfn; in psb_gem_fault() local
212 pfn = (dev_priv->stolen_base + r->offset) >> PAGE_SHIFT; in psb_gem_fault()
214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault()
215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn); in psb_gem_fault()
/linux-4.4.14/arch/nios2/mm/
Dioremap.c27 unsigned long pfn; in remap_area_pte() local
37 pfn = PFN_DOWN(phys_addr); in remap_area_pte()
43 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte()
45 pfn++; in remap_area_pte()
/linux-4.4.14/drivers/edac/
Dcell_edac.c38 unsigned long address, pfn, offset, syndrome; in cell_edac_count_ce() local
47 pfn = address >> PAGE_SHIFT; in cell_edac_count_ce()
53 csrow->first_page + pfn, offset, syndrome, in cell_edac_count_ce()
61 unsigned long address, pfn, offset; in cell_edac_count_ue() local
70 pfn = address >> PAGE_SHIFT; in cell_edac_count_ue()
75 csrow->first_page + pfn, offset, 0, in cell_edac_count_ue()
/linux-4.4.14/arch/ia64/kernel/
Dcrash_dump.c34 copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page()
/linux-4.4.14/fs/
Ddax.c45 unsigned long pfn; in dax_clear_blocks() local
48 count = bdev_direct_access(bdev, sector, &addr, &pfn, size); in dax_clear_blocks()
74 unsigned long pfn; in dax_get_addr() local
76 return bdev_direct_access(bh->b_bdev, sector, addr, &pfn, bh->b_size); in dax_get_addr()
299 unsigned long pfn; in dax_insert_mapping() local
318 error = bdev_direct_access(bh->b_bdev, sector, &addr, &pfn, bh->b_size); in dax_insert_mapping()
331 error = vm_insert_mixed(vma, vaddr, pfn); in dax_insert_mapping()
541 unsigned long pfn; in __dax_pmd_fault() local
627 length = bdev_direct_access(bh.b_bdev, sector, &kaddr, &pfn, in __dax_pmd_fault()
633 if ((length < PMD_SIZE) || (pfn & PG_PMD_COLOUR)) in __dax_pmd_fault()
[all …]
/linux-4.4.14/drivers/scsi/csiostor/
Dcsio_mb.c281 (FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(hw->pfn)); in csio_mb_ldst()
449 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_alloc()
505 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_write()
628 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_free()
663 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_alloc()
711 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_write()
814 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_free()
1158 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_debug_cmd_handler()
1159 uint32_t data_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_DATA_A); in csio_mb_debug_cmd_handler()
1192 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_issue()
[all …]
/linux-4.4.14/Documentation/virtual/kvm/
Dlocking.txt31 1): The mapping from gfn to pfn
32 The mapping from gfn to pfn may be changed since we can only ensure the pfn
63 to pin gfn to pfn, because after gfn_to_pfn_atomic():
64 - We have held the refcount of pfn that means the pfn can not be freed and
66 - The pfn is writable that means it can not be shared between different gfns
106 kvm_set_pfn_accessed(spte.pfn);
108 kvm_set_pfn_dirty(spte.pfn);
/linux-4.4.14/virt/kvm/
Dkvm_main.c114 static void kvm_release_pfn_dirty(pfn_t pfn);
122 bool kvm_is_reserved_pfn(pfn_t pfn) in kvm_is_reserved_pfn() argument
124 if (pfn_valid(pfn)) in kvm_is_reserved_pfn()
125 return PageReserved(pfn_to_page(pfn)); in kvm_is_reserved_pfn()
1303 bool write_fault, bool *writable, pfn_t *pfn) in hva_to_pfn_fast() argument
1321 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast()
1336 bool *writable, pfn_t *pfn) in hva_to_pfn_slow() argument
1371 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow()
1404 pfn_t pfn = 0; in hva_to_pfn() local
1410 if (hva_to_pfn_fast(addr, atomic, async, write_fault, writable, &pfn)) in hva_to_pfn()
[all …]
/linux-4.4.14/arch/arm/mach-mmp/
Dcommon.c30 .pfn = __phys_to_pfn(APB_PHYS_BASE),
35 .pfn = __phys_to_pfn(AXI_PHYS_BASE),
/linux-4.4.14/arch/microblaze/kernel/
Ddma.c166 unsigned long pfn; in dma_direct_mmap_coherent() local
173 pfn = consistent_virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent()
175 pfn = virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent()
177 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent()
/linux-4.4.14/arch/mips/kernel/
Dcrash_dump.c26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
/linux-4.4.14/tools/vm/
Dpage-types.c290 unsigned long pfn; in pagemap_pfn() local
293 pfn = PM_PFRAME(val); in pagemap_pfn()
295 pfn = 0; in pagemap_pfn()
297 return pfn; in pagemap_pfn()
622 unsigned long pfn; in walk_vma() local
632 pfn = pagemap_pfn(buf[i]); in walk_vma()
633 if (pfn) in walk_vma()
634 walk_pfn(index + i, pfn, 1, buf[i]); in walk_vma()
854 unsigned long nr_pages, pfn, i; in walk_file() local
906 pfn = pagemap_pfn(buf[i]); in walk_file()
[all …]
/linux-4.4.14/arch/mn10300/mm/
Dpgtable.c35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument
45 if (pfn & (PTRS_PER_PTE-1)) { /* pfn is misaligned */ in set_pmd_pfn()
56 set_pmd(pmd, pfn_pmd(pfn, flags)); in set_pmd_pfn()
/linux-4.4.14/drivers/gpu/drm/ttm/
Dttm_bo_vm.c93 unsigned long pfn; in ttm_bo_vm_fault() local
216 pfn = ((bo->mem.bus.base + bo->mem.bus.offset) >> PAGE_SHIFT) + page_offset; in ttm_bo_vm_fault()
228 pfn = page_to_pfn(page); in ttm_bo_vm_fault()
232 ret = vm_insert_mixed(&cvma, address, pfn); in ttm_bo_vm_fault()
234 ret = vm_insert_pfn(&cvma, address, pfn); in ttm_bo_vm_fault()
/linux-4.4.14/arch/arm/mach-versatile/
Dcore.c134 .pfn = __phys_to_pfn(VERSATILE_SYS_BASE),
139 .pfn = __phys_to_pfn(VERSATILE_SIC_BASE),
144 .pfn = __phys_to_pfn(VERSATILE_VIC_BASE),
149 .pfn = __phys_to_pfn(VERSATILE_SCTL_BASE),
156 .pfn = __phys_to_pfn(VERSATILE_IB2_BASE),
164 .pfn = __phys_to_pfn(VERSATILE_UART0_BASE),
172 .pfn = __phys_to_pfn(VERSATILE_PCI_CORE_BASE),
177 .pfn = __phys_to_pfn(VERSATILE_PCI_BASE),
182 .pfn = __phys_to_pfn(VERSATILE_PCI_CFG_BASE),
/linux-4.4.14/drivers/iommu/
Diova.c301 struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn) in find_iova() argument
313 if ((pfn >= iova->pfn_lo) && (pfn <= iova->pfn_hi)) { in find_iova()
324 if (pfn < iova->pfn_lo) in find_iova()
326 else if (pfn > iova->pfn_lo) in find_iova()
362 free_iova(struct iova_domain *iovad, unsigned long pfn) in free_iova() argument
364 struct iova *iova = find_iova(iovad, pfn); in free_iova()
/linux-4.4.14/arch/arm/mach-rpc/
Driscpc.c71 .pfn = __phys_to_pfn(SCREEN_START),
76 .pfn = __phys_to_pfn(IO_START),
81 .pfn = __phys_to_pfn(EASI_START),
/linux-4.4.14/arch/arm/mach-ks8695/include/mach/
Dmemory.h34 #define __arch_pfn_to_dma(dev, pfn) \ argument
35 ({ dma_addr_t __dma = __pfn_to_phys(pfn); \
/linux-4.4.14/arch/arm/mach-omap1/include/mach/
Dmemory.h30 #define __arch_pfn_to_dma(dev, pfn) \ argument
31 ({ dma_addr_t __dma = __pfn_to_phys(pfn); \

123