Home
last modified time | relevance | path

Searched refs:end_pfn (Results 1 – 73 of 73) sorted by relevance

/linux-4.1.27/arch/x86/mm/
Dinit.c187 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument
190 if (start_pfn < end_pfn) { in save_mr()
194 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr()
265 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local
281 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
283 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
285 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
287 if (end_pfn > limit_pfn) in split_mem_range()
288 end_pfn = limit_pfn; in split_mem_range()
289 if (start_pfn < end_pfn) { in split_mem_range()
[all …]
Dnuma_32.c65 unsigned long end_pfn) in node_memmap_size_bytes() argument
67 unsigned long nr_pages = end_pfn - start_pfn; in node_memmap_size_bytes()
Dinit_32.c258 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local
269 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init()
298 if (pfn >= end_pfn) in kernel_physical_mapping_init()
306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init()
346 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init()
432 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument
439 start_pfn, end_pfn); in add_highpages_with_active_regions()
441 start_pfn, end_pfn); in add_highpages_with_active_regions()
Dnuma_emulation.c34 unsigned long end_pfn = PFN_DOWN(end); in mem_hole_size() local
36 if (start_pfn < end_pfn) in mem_hole_size()
37 return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn)); in mem_hole_size()
Dpat.c304 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram() local
317 if (start_pfn < end_pfn) { in pat_pagerange_is_ram()
318 ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, in pat_pagerange_is_ram()
Dinit_64.c677 unsigned long end_pfn = PFN_UP(start + size); in update_end_of_memory_vars() local
679 if (end_pfn > max_pfn) { in update_end_of_memory_vars()
680 max_pfn = end_pfn; in update_end_of_memory_vars()
681 max_low_pfn = end_pfn; in update_end_of_memory_vars()
/linux-4.1.27/arch/mips/loongson/loongson-3/
Dnuma.c128 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local
144 end_pfn = start_pfn + node_psize; in szmem()
149 start_pfn, end_pfn, num_physpages); in szmem()
153 PFN_PHYS(end_pfn - start_pfn), node); in szmem()
158 end_pfn = start_pfn + node_psize; in szmem()
163 start_pfn, end_pfn, num_physpages); in szmem()
167 PFN_PHYS(end_pfn - start_pfn), node); in szmem()
185 unsigned long start_pfn, end_pfn, freepfn; in node_mem_init() local
191 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init()
196 node, start_pfn, end_pfn, freepfn); in node_mem_init()
[all …]
/linux-4.1.27/mm/
Dpage_isolation.c157 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument
165 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in start_isolate_page_range()
168 pfn < end_pfn; in start_isolate_page_range()
190 int undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument
196 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in undo_isolate_page_range()
198 pfn < end_pfn; in undo_isolate_page_range()
215 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument
220 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock()
257 if (pfn < end_pfn) in __test_page_isolated_in_pageblock()
262 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument
[all …]
Dmemory_hotplug.c247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local
272 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
289 unsigned long end_pfn) in grow_zone_span() argument
299 zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - in grow_zone_span()
306 unsigned long end_pfn) in resize_zone() argument
310 if (end_pfn - start_pfn) { in resize_zone()
312 zone->spanned_pages = end_pfn - start_pfn; in resize_zone()
326 unsigned long end_pfn) in fix_zone_id() argument
332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id()
[all …]
Dcompaction.c103 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument
109 end_pfn--; in pageblock_pfn_to_page()
111 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in pageblock_pfn_to_page()
119 end_page = pfn_to_page(end_pfn); in pageblock_pfn_to_page()
218 unsigned long end_pfn = zone_end_pfn(zone); in __reset_isolation_suitable() local
223 zone->compact_cached_free_pfn = end_pfn; in __reset_isolation_suitable()
227 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable()
401 unsigned long end_pfn, in isolate_freepages_block() argument
414 for (; blockpfn < end_pfn; blockpfn++, cursor++) { in isolate_freepages_block()
504 if (strict && blockpfn < end_pfn) in isolate_freepages_block()
[all …]
Dpage_alloc.c1136 unsigned long start_pfn, end_pfn; in move_freepages_block() local
1143 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block()
1148 if (!zone_spans_pfn(zone, end_pfn)) in move_freepages_block()
4084 static int pageblock_is_reserved(unsigned long start_pfn, unsigned long end_pfn) in pageblock_is_reserved() argument
4088 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in pageblock_is_reserved()
4104 unsigned long start_pfn, pfn, end_pfn, block_end_pfn; in setup_zone_migrate_reserve() local
4117 end_pfn = zone_end_pfn(zone); in setup_zone_migrate_reserve()
4137 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in setup_zone_migrate_reserve()
4154 block_end_pfn = min(pfn + pageblock_nr_pages, end_pfn); in setup_zone_migrate_reserve()
4201 unsigned long end_pfn = start_pfn + size; in memmap_init_zone() local
[all …]
Dsparse.c145 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
156 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits()
159 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits()
160 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits()
163 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits()
165 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits()
195 unsigned long end_pfn) in node_memmap_size_bytes() argument
200 mminit_validate_memmodel_limits(&start_pfn, &end_pfn); in node_memmap_size_bytes()
201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
Dnobootmem.c105 unsigned long end_pfn = min_t(unsigned long, in __free_memory_core() local
108 if (start_pfn > end_pfn) in __free_memory_core()
111 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core()
113 return end_pfn - start_pfn; in __free_memory_core()
Dinternal.h200 unsigned long start_pfn, unsigned long end_pfn);
203 unsigned long low_pfn, unsigned long end_pfn);
393 unsigned long *end_pfn);
396 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
Dpage_owner.c212 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone() local
223 for (; pfn < end_pfn; ) { in init_pages_in_zone()
230 block_end_pfn = min(block_end_pfn, end_pfn); in init_pages_in_zone()
Dpage_ext.c364 unsigned long start_pfn, end_pfn; in page_ext_init() local
367 end_pfn = node_end_pfn(nid); in page_ext_init()
373 for (pfn = start_pfn; pfn < end_pfn; in page_ext_init()
Dmemblock.c1337 unsigned long start_pfn, end_pfn; in memblock_mem_size() local
1341 end_pfn = memblock_region_memory_end_pfn(r); in memblock_mem_size()
1343 end_pfn = min_t(unsigned long, end_pfn, limit_pfn); in memblock_mem_size()
1344 pages += end_pfn - start_pfn; in memblock_mem_size()
1417 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument
1426 *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); in memblock_search_pfn_nid()
Dvmstat.c982 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print() local
985 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
1033 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() local
1046 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print()
1053 block_end_pfn = min(block_end_pfn, end_pfn); in pagetypeinfo_showmixedcount_print()
Dbootmem.c783 unsigned long end_pfn; in __alloc_bootmem_node_high() local
789 end_pfn = pgdat_end_pfn(pgdat); in __alloc_bootmem_node_high()
791 if (end_pfn > MAX_DMA32_PFN + (128 >> (20 - PAGE_SHIFT)) && in __alloc_bootmem_node_high()
Dhugetlb.c782 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_gigantic_page() local
783 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE); in __alloc_gigantic_page()
789 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_gigantic() local
792 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_gigantic()
1194 void dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_huge_pages() argument
1202 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) in dissolve_free_huge_pages()
Dksm.c1992 unsigned long end_pfn) in ksm_check_stable_tree() argument
2004 stable_node->kpfn < end_pfn) { in ksm_check_stable_tree()
2019 stable_node->kpfn < end_pfn) in ksm_check_stable_tree()
Dkmemleak.c1350 unsigned long end_pfn = node_end_pfn(i); in kmemleak_scan() local
1353 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
/linux-4.1.27/arch/sh/mm/
Dnuma.c30 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local
37 end_pfn = end >> PAGE_SHIFT; in setup_bootmem_node()
44 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node()
53 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
56 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node()
60 start_pfn, end_pfn); in setup_bootmem_node()
62 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
Dinit.c195 unsigned long start_pfn, end_pfn; in allocate_pgdat() local
200 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat()
204 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat()
219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat()
225 unsigned long end_pfn; in bootmem_init_one_node() local
234 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node()
242 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node()
244 free_bootmem_with_active_regions(nid, end_pfn); in bootmem_init_one_node()
270 unsigned long start_pfn, end_pfn; in do_init_bootmem() local
272 end_pfn = memblock_region_memory_end_pfn(reg); in do_init_bootmem()
[all …]
/linux-4.1.27/arch/metag/mm/
Dnuma.c33 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local
40 end_pfn = end >> PAGE_SHIFT; in setup_bootmem_node()
45 PFN_PHYS(end_pfn - start_pfn), in setup_bootmem_node()
56 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
59 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node()
63 start_pfn, end_pfn); in setup_bootmem_node()
65 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
Dinit.c111 unsigned long start_pfn, end_pfn; in allocate_pgdat() local
116 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat()
120 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat()
136 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat()
142 unsigned long end_pfn; in bootmem_init_one_node() local
151 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node()
153 if (end_pfn > max_low_pfn) in bootmem_init_one_node()
154 end_pfn = max_low_pfn; in bootmem_init_one_node()
157 total_pages = bootmem_bootmap_pages(end_pfn - p->node_start_pfn); in bootmem_init_one_node()
163 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node()
[all …]
/linux-4.1.27/include/trace/events/
Dcompaction.h16 unsigned long end_pfn,
20 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken),
24 __field(unsigned long, end_pfn)
31 __entry->end_pfn = end_pfn;
38 __entry->end_pfn,
47 unsigned long end_pfn,
51 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
58 unsigned long end_pfn,
62 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
/linux-4.1.27/arch/unicore32/mm/
Dinit.c132 unsigned long end_pfn) in uc32_bootmem_init() argument
143 boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in uc32_bootmem_init()
145 __pfn_to_phys(end_pfn)); in uc32_bootmem_init()
153 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); in uc32_bootmem_init()
160 if (end >= end_pfn) in uc32_bootmem_init()
161 end = end_pfn; in uc32_bootmem_init()
173 if (end >= end_pfn) in uc32_bootmem_init()
174 end = end_pfn; in uc32_bootmem_init()
317 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument
326 end_pg = pfn_to_page(end_pfn); in free_memmap()
/linux-4.1.27/arch/sparc/mm/
Dinit_32.c78 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local
80 if (end_pfn <= max_low_pfn) in calc_highpages()
86 nr += end_pfn - start_pfn; in calc_highpages()
276 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument
281 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region()
284 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region()
325 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local
327 if (end_pfn <= highstart_pfn) in mem_init()
333 map_high_region(start_pfn, end_pfn); in mem_init()
Dinit_64.c872 unsigned long start_pfn, end_pfn; in allocate_node_data() local
889 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_node_data()
891 p->node_spanned_pages = end_pfn - start_pfn; in allocate_node_data()
1346 unsigned long end_pfn; in bootmem_init() local
1348 end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; in bootmem_init()
1349 max_pfn = max_low_pfn = end_pfn; in bootmem_init()
1363 return end_pfn; in bootmem_init()
2018 unsigned long end_pfn, shift, phys_base; in paging_init() local
2186 last_valid_pfn = end_pfn = bootmem_init(phys_base); in paging_init()
2210 max_zone_pfns[ZONE_NORMAL] = end_pfn; in paging_init()
/linux-4.1.27/arch/mips/sgi-ip27/
Dip27-memory.c393 unsigned long start_pfn, end_pfn; in node_mem_init() local
395 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init()
405 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init()
413 start_pfn, end_pfn); in node_mem_init()
414 free_bootmem_with_active_regions(node, end_pfn); in node_mem_init()
467 unsigned long start_pfn, end_pfn; in paging_init() local
469 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in paging_init()
471 if (end_pfn > max_low_pfn) in paging_init()
472 max_low_pfn = end_pfn; in paging_init()
/linux-4.1.27/include/linux/
Dpage-isolation.h51 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
59 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
65 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
Dmemory_hotplug.h266 extern int walk_memory_range(unsigned long start_pfn, unsigned long end_pfn,
Dmemblock.h158 unsigned long *end_pfn);
Dhugetlb.h444 unsigned long end_pfn);
Dmm.h1737 unsigned long end_pfn);
1739 unsigned long end_pfn);
1741 unsigned long *start_pfn, unsigned long *end_pfn);
Dmmzone.h696 unsigned long end_pfn; member
/linux-4.1.27/arch/s390/kernel/
Dearly.c82 unsigned int i, stext_pfn, eshared_pfn, end_pfn, min_size; in create_kernel_nss() local
115 end_pfn = PFN_UP(__pa(&_end)); in create_kernel_nss()
116 min_size = end_pfn << 2; in create_kernel_nss()
121 eshared_pfn - 1, eshared_pfn, end_pfn); in create_kernel_nss()
201 unsigned long end_pfn, init_pfn; in init_kernel_storage_key() local
203 end_pfn = PFN_UP(__pa(&_end)); in init_kernel_storage_key()
205 for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++) in init_kernel_storage_key()
/linux-4.1.27/arch/sh/kernel/
Dswsusp.c25 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local
27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
Dsetup.c192 unsigned long end_pfn) in __add_active_range() argument
200 end = end_pfn << PAGE_SHIFT; in __add_active_range()
209 start_pfn, end_pfn); in __add_active_range()
233 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
/linux-4.1.27/arch/x86/xen/
Dsetup.c224 unsigned long end_pfn, unsigned long nr_pages, unsigned long *released) in xen_set_identity_and_release_chunk() argument
229 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk()
232 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk()
251 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk()
360 unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, in xen_set_identity_and_remap_chunk() argument
365 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk()
406 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk()
440 unsigned long end_pfn = PFN_UP(end); in xen_set_identity_and_remap() local
443 end_pfn = PFN_UP(entry->addr); in xen_set_identity_and_remap()
445 if (start_pfn < end_pfn) in xen_set_identity_and_remap()
[all …]
/linux-4.1.27/arch/x86/kernel/
De820.c764 unsigned long end_pfn; in e820_end_pfn() local
774 end_pfn = (ei->addr + ei->size) >> PAGE_SHIFT; in e820_end_pfn()
778 if (end_pfn > limit_pfn) { in e820_end_pfn()
782 if (end_pfn > last_pfn) in e820_end_pfn()
783 last_pfn = end_pfn; in e820_end_pfn()
1110 unsigned long start_pfn, end_pfn; in memblock_find_dma_reserve() local
1120 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { in memblock_find_dma_reserve()
1122 end_pfn = min(end_pfn, MAX_DMA_PFN); in memblock_find_dma_reserve()
1123 nr_pages += end_pfn - start_pfn; in memblock_find_dma_reserve()
1128 end_pfn = min_t(unsigned long, PFN_DOWN(end), MAX_DMA_PFN); in memblock_find_dma_reserve()
[all …]
Damd_gart_64.c738 unsigned long start_pfn, end_pfn; in gart_iommu_init() local
769 end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); in gart_iommu_init()
772 if (!pfn_range_is_mapped(start_pfn, end_pfn)) in gart_iommu_init()
773 init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); in gart_iommu_init()
/linux-4.1.27/arch/mn10300/kernel/
Dsetup.c95 unsigned long kstart_pfn, start_pfn, free_pfn, end_pfn; in setup_arch() local
128 end_pfn = PFN_DOWN(__pa(memory_end)); in setup_arch()
133 end_pfn); in setup_arch()
140 PFN_PHYS(end_pfn - free_pfn)); in setup_arch()
/linux-4.1.27/arch/powerpc/mm/
Dnuma.c90 static int __init fake_numa_create_new_node(unsigned long end_pfn, in fake_numa_create_new_node() argument
121 if ((end_pfn << PAGE_SHIFT) > mem) { in fake_numa_create_new_node()
819 unsigned long start_pfn, end_pfn; in setup_nonnuma() local
830 end_pfn = memblock_region_memory_end_pfn(reg); in setup_nonnuma()
832 fake_numa_create_new_node(end_pfn, &nid); in setup_nonnuma()
834 PFN_PHYS(end_pfn - start_pfn), in setup_nonnuma()
915 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument
917 u64 spanned_pages = end_pfn - start_pfn; in setup_node_data()
926 (end_pfn << PAGE_SHIFT) - 1); in setup_node_data()
969 unsigned long start_pfn, end_pfn; in initmem_init() local
[all …]
Dmem.c174 unsigned long end_pfn = start_pfn + nr_pages; in walk_system_ram_range() local
180 tend = min(end_pfn, memblock_region_memory_end_pfn(reg)); in walk_system_ram_range()
/linux-4.1.27/arch/alpha/mm/
Dnuma.c307 unsigned long end_pfn = bdata->node_low_pfn; in paging_init() local
309 if (dma_local_pfn >= end_pfn - start_pfn) in paging_init()
310 zones_size[ZONE_DMA] = end_pfn - start_pfn; in paging_init()
313 zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn; in paging_init()
/linux-4.1.27/arch/unicore32/kernel/
Dhibernate.c148 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local
150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/x86/include/asm/
Dmtrr.h49 extern int mtrr_trim_uncached_memory(unsigned long end_pfn);
80 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
Dhighmem.h76 unsigned long end_pfn);
Dpage_types.h61 bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
/linux-4.1.27/kernel/power/
Dsnapshot.c290 unsigned long end_pfn; /* Zone end page frame + 1 */ member
444 zone->end_pfn = end; in create_zone_bm_rtree()
643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit()
650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit()
808 pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; in memory_bm_next_pfn()
830 unsigned long end_pfn; member
842 __register_nosave_region(unsigned long start_pfn, unsigned long end_pfn, in __register_nosave_region() argument
847 if (start_pfn >= end_pfn) in __register_nosave_region()
854 if (region->end_pfn == start_pfn) { in __register_nosave_region()
855 region->end_pfn = end_pfn; in __register_nosave_region()
[all …]
/linux-4.1.27/drivers/hv/
Dhv_balloon.c440 unsigned long end_pfn; member
713 if ((start_pfn >= has->end_pfn)) in pfn_covered()
719 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered()
720 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered()
728 has->end_pfn += new_inc; in pfn_covered()
767 if ((start_pfn >= has->end_pfn)) in handle_pg_range()
796 if ((has->ha_end_pfn < has->end_pfn) && (pfn_cnt > 0)) { in handle_pg_range()
804 size = (has->end_pfn - has->ha_end_pfn); in handle_pg_range()
854 ha_region->end_pfn = rg_start + rg_size; in process_hot_add()
/linux-4.1.27/arch/parisc/mm/
Dinit.c243 unsigned long end_pfn; in setup_bootmem() local
247 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; in setup_bootmem()
250 hole_pages = pmem_ranges[i].start_pfn - end_pfn; in setup_bootmem()
252 pmem_holes[npmem_holes].start_pfn = end_pfn; in setup_bootmem()
254 end_pfn += hole_pages; in setup_bootmem()
256 end_pfn += pmem_ranges[i].pages; in setup_bootmem()
259 pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn; in setup_bootmem()
/linux-4.1.27/arch/microblaze/mm/
Dinit.c190 unsigned long start_pfn, end_pfn; in setup_memory() local
193 end_pfn = memblock_region_memory_end_pfn(reg); in setup_memory()
195 (end_pfn - start_pfn) << PAGE_SHIFT, in setup_memory()
/linux-4.1.27/fs/proc/
Dkcore.c250 unsigned long end_pfn; in kcore_update_ram() local
255 end_pfn = 0; in kcore_update_ram()
259 if (end_pfn < node_end) in kcore_update_ram()
260 end_pfn = node_end; in kcore_update_ram()
263 ret = walk_system_ram_range(0, end_pfn, &head, kclist_add_private); in kcore_update_ram()
/linux-4.1.27/arch/sh/include/asm/
Dmmzone.h42 unsigned long end_pfn);
/linux-4.1.27/arch/tile/mm/
Dinit.c443 unsigned long end_pfn = node_lowmem_end_pfn[i]; in kernel_physical_mapping_init() local
445 unsigned long end_pfn = node_end_pfn[i]; in kernel_physical_mapping_init()
451 end_huge_pfn = end_pfn - (HPAGE_SIZE >> PAGE_SHIFT); in kernel_physical_mapping_init()
456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init()
459 while (pfn < end_pfn) { in kernel_physical_mapping_init()
/linux-4.1.27/drivers/base/
Dmemory.c386 unsigned long start_pfn, end_pfn; in show_valid_zones() local
392 end_pfn = start_pfn + nr_pages; in show_valid_zones()
396 if (!test_pages_in_a_zone(start_pfn, end_pfn)) in show_valid_zones()
403 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones()
Dnode.c459 unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_spanned_pages; in link_mem_sections() local
464 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in link_mem_sections()
/linux-4.1.27/arch/tile/kernel/
Dsetup.c1670 insert_ram_resource(u64 start_pfn, u64 end_pfn, bool reserved) in insert_ram_resource() argument
1678 res->end = (end_pfn << PAGE_SHIFT) - 1; in insert_ram_resource()
1704 u64 end_pfn = node_end_pfn[i]; in request_standard_resources() local
1708 end_pfn > pci_reserve_start_pfn) { in request_standard_resources()
1709 if (end_pfn > pci_reserve_end_pfn) in request_standard_resources()
1711 end_pfn, 0); in request_standard_resources()
1712 end_pfn = pci_reserve_start_pfn; in request_standard_resources()
1715 insert_ram_resource(start_pfn, end_pfn, 0); in request_standard_resources()
Dpci_gx.c984 unsigned long end_pfn = node_end_pfn[j]; in pcibios_init() local
985 unsigned long nr_pages = end_pfn - start_pfn; in pcibios_init()
/linux-4.1.27/arch/arm64/mm/
Dinit.c209 static inline void free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument
218 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
/linux-4.1.27/arch/blackfin/kernel/
Dsetup.c738 unsigned long start_pfn, end_pfn; in setup_bootmem_allocator() local
759 end_pfn = memory_end >> PAGE_SHIFT; in setup_bootmem_allocator()
767 start_pfn, end_pfn); in setup_bootmem_allocator()
780 if (curr_pfn >= end_pfn) in setup_bootmem_allocator()
788 if (last_pfn > end_pfn) in setup_bootmem_allocator()
789 last_pfn = end_pfn; in setup_bootmem_allocator()
/linux-4.1.27/arch/x86/platform/efi/
Defi.c554 u64 start_pfn, end_pfn, end; in old_map_region() local
561 end_pfn = PFN_UP(end); in old_map_region()
563 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in old_map_region()
/linux-4.1.27/arch/arm/mm/
Dinit.c332 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument
341 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
/linux-4.1.27/arch/x86/kernel/cpu/mtrr/
Dcleanup.c871 int __init mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
963 if (range[i].end < end_pfn) in mtrr_trim_uncached_memory()
965 end_pfn); in mtrr_trim_uncached_memory()
/linux-4.1.27/kernel/
Dresource.c457 unsigned long pfn, end_pfn; in walk_system_ram_range() local
468 end_pfn = (res.end + 1) >> PAGE_SHIFT; in walk_system_ram_range()
469 if (end_pfn > pfn) in walk_system_ram_range()
470 ret = (*func)(pfn, end_pfn - pfn, arg); in walk_system_ram_range()
Dmodule.c1713 unsigned long end_pfn = PFN_DOWN((unsigned long)end); in set_page_attributes() local
1715 if (end_pfn > begin_pfn) in set_page_attributes()
1716 set(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_page_attributes()
1726 unsigned long end_pfn; in set_section_ro_nx() local
1743 end_pfn = PFN_UP((unsigned long)base + total_size); in set_section_ro_nx()
1744 if (end_pfn > begin_pfn) in set_section_ro_nx()
1745 set_memory_nx(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_section_ro_nx()
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.c709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local
717 end_pfn = initial_pfn + total_nr_pages; in ehea_create_busmap_callback()
720 while (pfn < end_pfn) { in ehea_create_busmap_callback()
/linux-4.1.27/drivers/iommu/
Dintel-iommu.c2036 unsigned long nr_superpages, end_pfn; in __domain_mapping() local
2042 end_pfn = iov_pfn + nr_superpages * lvl_pages - 1; in __domain_mapping()
2048 dma_pte_free_pagetable(domain, iov_pfn, end_pfn); in __domain_mapping()
2487 unsigned long start_pfn, end_pfn; in si_domain_init() local
2490 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in si_domain_init()
2492 PFN_PHYS(start_pfn), PFN_PHYS(end_pfn)); in si_domain_init()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_mrmw.c2369 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehca_create_busmap_callback() local
2376 end_pfn = initial_pfn + total_nr_pages; in ehca_create_busmap_callback()
2379 while (pfn < end_pfn) { in ehca_create_busmap_callback()
/linux-4.1.27/Documentation/virtual/uml/
DUserModeLinux-HOWTO.txt3746 bootmap_size = init_bootmem(start_pfn, end_pfn - start_pfn);