/linux-4.1.27/arch/x86/mm/ |
D | init.c | 187 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument 190 if (start_pfn < end_pfn) { in save_mr() 194 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 265 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 281 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 283 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 285 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 287 if (end_pfn > limit_pfn) in split_mem_range() 288 end_pfn = limit_pfn; in split_mem_range() 289 if (start_pfn < end_pfn) { in split_mem_range() [all …]
|
D | numa_32.c | 65 unsigned long end_pfn) in node_memmap_size_bytes() argument 67 unsigned long nr_pages = end_pfn - start_pfn; in node_memmap_size_bytes()
|
D | init_32.c | 258 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local 269 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init() 298 if (pfn >= end_pfn) in kernel_physical_mapping_init() 306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 346 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init() 432 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument 439 start_pfn, end_pfn); in add_highpages_with_active_regions() 441 start_pfn, end_pfn); in add_highpages_with_active_regions()
|
D | numa_emulation.c | 34 unsigned long end_pfn = PFN_DOWN(end); in mem_hole_size() local 36 if (start_pfn < end_pfn) in mem_hole_size() 37 return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn)); in mem_hole_size()
|
D | pat.c | 304 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram() local 317 if (start_pfn < end_pfn) { in pat_pagerange_is_ram() 318 ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, in pat_pagerange_is_ram()
|
D | init_64.c | 677 unsigned long end_pfn = PFN_UP(start + size); in update_end_of_memory_vars() local 679 if (end_pfn > max_pfn) { in update_end_of_memory_vars() 680 max_pfn = end_pfn; in update_end_of_memory_vars() 681 max_low_pfn = end_pfn; in update_end_of_memory_vars()
|
/linux-4.1.27/arch/mips/loongson/loongson-3/ |
D | numa.c | 128 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local 144 end_pfn = start_pfn + node_psize; in szmem() 149 start_pfn, end_pfn, num_physpages); in szmem() 153 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 158 end_pfn = start_pfn + node_psize; in szmem() 163 start_pfn, end_pfn, num_physpages); in szmem() 167 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 185 unsigned long start_pfn, end_pfn, freepfn; in node_mem_init() local 191 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 196 node, start_pfn, end_pfn, freepfn); in node_mem_init() [all …]
|
/linux-4.1.27/mm/ |
D | page_isolation.c | 157 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 165 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in start_isolate_page_range() 168 pfn < end_pfn; in start_isolate_page_range() 190 int undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 196 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in undo_isolate_page_range() 198 pfn < end_pfn; in undo_isolate_page_range() 215 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument 220 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock() 257 if (pfn < end_pfn) in __test_page_isolated_in_pageblock() 262 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument [all …]
|
D | memory_hotplug.c | 247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 272 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 289 unsigned long end_pfn) in grow_zone_span() argument 299 zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - in grow_zone_span() 306 unsigned long end_pfn) in resize_zone() argument 310 if (end_pfn - start_pfn) { in resize_zone() 312 zone->spanned_pages = end_pfn - start_pfn; in resize_zone() 326 unsigned long end_pfn) in fix_zone_id() argument 332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id() [all …]
|
D | compaction.c | 103 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 109 end_pfn--; in pageblock_pfn_to_page() 111 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in pageblock_pfn_to_page() 119 end_page = pfn_to_page(end_pfn); in pageblock_pfn_to_page() 218 unsigned long end_pfn = zone_end_pfn(zone); in __reset_isolation_suitable() local 223 zone->compact_cached_free_pfn = end_pfn; in __reset_isolation_suitable() 227 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable() 401 unsigned long end_pfn, in isolate_freepages_block() argument 414 for (; blockpfn < end_pfn; blockpfn++, cursor++) { in isolate_freepages_block() 504 if (strict && blockpfn < end_pfn) in isolate_freepages_block() [all …]
|
D | page_alloc.c | 1136 unsigned long start_pfn, end_pfn; in move_freepages_block() local 1143 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block() 1148 if (!zone_spans_pfn(zone, end_pfn)) in move_freepages_block() 4084 static int pageblock_is_reserved(unsigned long start_pfn, unsigned long end_pfn) in pageblock_is_reserved() argument 4088 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in pageblock_is_reserved() 4104 unsigned long start_pfn, pfn, end_pfn, block_end_pfn; in setup_zone_migrate_reserve() local 4117 end_pfn = zone_end_pfn(zone); in setup_zone_migrate_reserve() 4137 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in setup_zone_migrate_reserve() 4154 block_end_pfn = min(pfn + pageblock_nr_pages, end_pfn); in setup_zone_migrate_reserve() 4201 unsigned long end_pfn = start_pfn + size; in memmap_init_zone() local [all …]
|
D | sparse.c | 145 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument 156 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 159 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 160 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 163 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 165 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 195 unsigned long end_pfn) in node_memmap_size_bytes() argument 200 mminit_validate_memmodel_limits(&start_pfn, &end_pfn); in node_memmap_size_bytes() 201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
|
D | nobootmem.c | 105 unsigned long end_pfn = min_t(unsigned long, in __free_memory_core() local 108 if (start_pfn > end_pfn) in __free_memory_core() 111 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core() 113 return end_pfn - start_pfn; in __free_memory_core()
|
D | internal.h | 200 unsigned long start_pfn, unsigned long end_pfn); 203 unsigned long low_pfn, unsigned long end_pfn); 393 unsigned long *end_pfn); 396 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
|
D | page_owner.c | 212 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone() local 223 for (; pfn < end_pfn; ) { in init_pages_in_zone() 230 block_end_pfn = min(block_end_pfn, end_pfn); in init_pages_in_zone()
|
D | page_ext.c | 364 unsigned long start_pfn, end_pfn; in page_ext_init() local 367 end_pfn = node_end_pfn(nid); in page_ext_init() 373 for (pfn = start_pfn; pfn < end_pfn; in page_ext_init()
|
D | memblock.c | 1337 unsigned long start_pfn, end_pfn; in memblock_mem_size() local 1341 end_pfn = memblock_region_memory_end_pfn(r); in memblock_mem_size() 1343 end_pfn = min_t(unsigned long, end_pfn, limit_pfn); in memblock_mem_size() 1344 pages += end_pfn - start_pfn; in memblock_mem_size() 1417 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument 1426 *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); in memblock_search_pfn_nid()
|
D | vmstat.c | 982 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print() local 985 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print() 1033 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() local 1046 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 1053 block_end_pfn = min(block_end_pfn, end_pfn); in pagetypeinfo_showmixedcount_print()
|
D | bootmem.c | 783 unsigned long end_pfn; in __alloc_bootmem_node_high() local 789 end_pfn = pgdat_end_pfn(pgdat); in __alloc_bootmem_node_high() 791 if (end_pfn > MAX_DMA32_PFN + (128 >> (20 - PAGE_SHIFT)) && in __alloc_bootmem_node_high()
|
D | hugetlb.c | 782 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_gigantic_page() local 783 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE); in __alloc_gigantic_page() 789 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_gigantic() local 792 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_gigantic() 1194 void dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_huge_pages() argument 1202 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) in dissolve_free_huge_pages()
|
D | ksm.c | 1992 unsigned long end_pfn) in ksm_check_stable_tree() argument 2004 stable_node->kpfn < end_pfn) { in ksm_check_stable_tree() 2019 stable_node->kpfn < end_pfn) in ksm_check_stable_tree()
|
D | kmemleak.c | 1350 unsigned long end_pfn = node_end_pfn(i); in kmemleak_scan() local 1353 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
|
/linux-4.1.27/arch/sh/mm/ |
D | numa.c | 30 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 37 end_pfn = end >> PAGE_SHIFT; in setup_bootmem_node() 44 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 53 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node() 56 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node() 60 start_pfn, end_pfn); in setup_bootmem_node() 62 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
|
D | init.c | 195 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 200 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 204 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat() 219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 225 unsigned long end_pfn; in bootmem_init_one_node() local 234 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node() 242 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node() 244 free_bootmem_with_active_regions(nid, end_pfn); in bootmem_init_one_node() 270 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 272 end_pfn = memblock_region_memory_end_pfn(reg); in do_init_bootmem() [all …]
|
/linux-4.1.27/arch/metag/mm/ |
D | numa.c | 33 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 40 end_pfn = end >> PAGE_SHIFT; in setup_bootmem_node() 45 PFN_PHYS(end_pfn - start_pfn), in setup_bootmem_node() 56 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node() 59 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node() 63 start_pfn, end_pfn); in setup_bootmem_node() 65 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
|
D | init.c | 111 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 116 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 120 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat() 136 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 142 unsigned long end_pfn; in bootmem_init_one_node() local 151 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node() 153 if (end_pfn > max_low_pfn) in bootmem_init_one_node() 154 end_pfn = max_low_pfn; in bootmem_init_one_node() 157 total_pages = bootmem_bootmap_pages(end_pfn - p->node_start_pfn); in bootmem_init_one_node() 163 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node() [all …]
|
/linux-4.1.27/include/trace/events/ |
D | compaction.h | 16 unsigned long end_pfn, 20 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 24 __field(unsigned long, end_pfn) 31 __entry->end_pfn = end_pfn; 38 __entry->end_pfn, 47 unsigned long end_pfn, 51 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 58 unsigned long end_pfn, 62 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
|
/linux-4.1.27/arch/unicore32/mm/ |
D | init.c | 132 unsigned long end_pfn) in uc32_bootmem_init() argument 143 boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in uc32_bootmem_init() 145 __pfn_to_phys(end_pfn)); in uc32_bootmem_init() 153 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); in uc32_bootmem_init() 160 if (end >= end_pfn) in uc32_bootmem_init() 161 end = end_pfn; in uc32_bootmem_init() 173 if (end >= end_pfn) in uc32_bootmem_init() 174 end = end_pfn; in uc32_bootmem_init() 317 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 326 end_pg = pfn_to_page(end_pfn); in free_memmap()
|
/linux-4.1.27/arch/sparc/mm/ |
D | init_32.c | 78 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 80 if (end_pfn <= max_low_pfn) in calc_highpages() 86 nr += end_pfn - start_pfn; in calc_highpages() 276 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 281 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 284 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 325 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 327 if (end_pfn <= highstart_pfn) in mem_init() 333 map_high_region(start_pfn, end_pfn); in mem_init()
|
D | init_64.c | 872 unsigned long start_pfn, end_pfn; in allocate_node_data() local 889 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_node_data() 891 p->node_spanned_pages = end_pfn - start_pfn; in allocate_node_data() 1346 unsigned long end_pfn; in bootmem_init() local 1348 end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; in bootmem_init() 1349 max_pfn = max_low_pfn = end_pfn; in bootmem_init() 1363 return end_pfn; in bootmem_init() 2018 unsigned long end_pfn, shift, phys_base; in paging_init() local 2186 last_valid_pfn = end_pfn = bootmem_init(phys_base); in paging_init() 2210 max_zone_pfns[ZONE_NORMAL] = end_pfn; in paging_init()
|
/linux-4.1.27/arch/mips/sgi-ip27/ |
D | ip27-memory.c | 393 unsigned long start_pfn, end_pfn; in node_mem_init() local 395 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 405 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 413 start_pfn, end_pfn); in node_mem_init() 414 free_bootmem_with_active_regions(node, end_pfn); in node_mem_init() 467 unsigned long start_pfn, end_pfn; in paging_init() local 469 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in paging_init() 471 if (end_pfn > max_low_pfn) in paging_init() 472 max_low_pfn = end_pfn; in paging_init()
|
/linux-4.1.27/include/linux/ |
D | page-isolation.h | 51 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 59 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 65 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
D | memory_hotplug.h | 266 extern int walk_memory_range(unsigned long start_pfn, unsigned long end_pfn,
|
D | memblock.h | 158 unsigned long *end_pfn);
|
D | hugetlb.h | 444 unsigned long end_pfn);
|
D | mm.h | 1737 unsigned long end_pfn); 1739 unsigned long end_pfn); 1741 unsigned long *start_pfn, unsigned long *end_pfn);
|
D | mmzone.h | 696 unsigned long end_pfn; member
|
/linux-4.1.27/arch/s390/kernel/ |
D | early.c | 82 unsigned int i, stext_pfn, eshared_pfn, end_pfn, min_size; in create_kernel_nss() local 115 end_pfn = PFN_UP(__pa(&_end)); in create_kernel_nss() 116 min_size = end_pfn << 2; in create_kernel_nss() 121 eshared_pfn - 1, eshared_pfn, end_pfn); in create_kernel_nss() 201 unsigned long end_pfn, init_pfn; in init_kernel_storage_key() local 203 end_pfn = PFN_UP(__pa(&_end)); in init_kernel_storage_key() 205 for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++) in init_kernel_storage_key()
|
/linux-4.1.27/arch/sh/kernel/ |
D | swsusp.c | 25 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | setup.c | 192 unsigned long end_pfn) in __add_active_range() argument 200 end = end_pfn << PAGE_SHIFT; in __add_active_range() 209 start_pfn, end_pfn); in __add_active_range() 233 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
/linux-4.1.27/arch/x86/xen/ |
D | setup.c | 224 unsigned long end_pfn, unsigned long nr_pages, unsigned long *released) in xen_set_identity_and_release_chunk() argument 229 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 232 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk() 251 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk() 360 unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, in xen_set_identity_and_remap_chunk() argument 365 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 406 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk() 440 unsigned long end_pfn = PFN_UP(end); in xen_set_identity_and_remap() local 443 end_pfn = PFN_UP(entry->addr); in xen_set_identity_and_remap() 445 if (start_pfn < end_pfn) in xen_set_identity_and_remap() [all …]
|
/linux-4.1.27/arch/x86/kernel/ |
D | e820.c | 764 unsigned long end_pfn; in e820_end_pfn() local 774 end_pfn = (ei->addr + ei->size) >> PAGE_SHIFT; in e820_end_pfn() 778 if (end_pfn > limit_pfn) { in e820_end_pfn() 782 if (end_pfn > last_pfn) in e820_end_pfn() 783 last_pfn = end_pfn; in e820_end_pfn() 1110 unsigned long start_pfn, end_pfn; in memblock_find_dma_reserve() local 1120 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { in memblock_find_dma_reserve() 1122 end_pfn = min(end_pfn, MAX_DMA_PFN); in memblock_find_dma_reserve() 1123 nr_pages += end_pfn - start_pfn; in memblock_find_dma_reserve() 1128 end_pfn = min_t(unsigned long, PFN_DOWN(end), MAX_DMA_PFN); in memblock_find_dma_reserve() [all …]
|
D | amd_gart_64.c | 738 unsigned long start_pfn, end_pfn; in gart_iommu_init() local 769 end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); in gart_iommu_init() 772 if (!pfn_range_is_mapped(start_pfn, end_pfn)) in gart_iommu_init() 773 init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); in gart_iommu_init()
|
/linux-4.1.27/arch/mn10300/kernel/ |
D | setup.c | 95 unsigned long kstart_pfn, start_pfn, free_pfn, end_pfn; in setup_arch() local 128 end_pfn = PFN_DOWN(__pa(memory_end)); in setup_arch() 133 end_pfn); in setup_arch() 140 PFN_PHYS(end_pfn - free_pfn)); in setup_arch()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | numa.c | 90 static int __init fake_numa_create_new_node(unsigned long end_pfn, in fake_numa_create_new_node() argument 121 if ((end_pfn << PAGE_SHIFT) > mem) { in fake_numa_create_new_node() 819 unsigned long start_pfn, end_pfn; in setup_nonnuma() local 830 end_pfn = memblock_region_memory_end_pfn(reg); in setup_nonnuma() 832 fake_numa_create_new_node(end_pfn, &nid); in setup_nonnuma() 834 PFN_PHYS(end_pfn - start_pfn), in setup_nonnuma() 915 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 917 u64 spanned_pages = end_pfn - start_pfn; in setup_node_data() 926 (end_pfn << PAGE_SHIFT) - 1); in setup_node_data() 969 unsigned long start_pfn, end_pfn; in initmem_init() local [all …]
|
D | mem.c | 174 unsigned long end_pfn = start_pfn + nr_pages; in walk_system_ram_range() local 180 tend = min(end_pfn, memblock_region_memory_end_pfn(reg)); in walk_system_ram_range()
|
/linux-4.1.27/arch/alpha/mm/ |
D | numa.c | 307 unsigned long end_pfn = bdata->node_low_pfn; in paging_init() local 309 if (dma_local_pfn >= end_pfn - start_pfn) in paging_init() 310 zones_size[ZONE_DMA] = end_pfn - start_pfn; in paging_init() 313 zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn; in paging_init()
|
/linux-4.1.27/arch/unicore32/kernel/ |
D | hibernate.c | 148 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | mtrr.h | 49 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 80 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
|
D | highmem.h | 76 unsigned long end_pfn);
|
D | page_types.h | 61 bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
|
/linux-4.1.27/kernel/power/ |
D | snapshot.c | 290 unsigned long end_pfn; /* Zone end page frame + 1 */ member 444 zone->end_pfn = end; in create_zone_bm_rtree() 643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 808 pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; in memory_bm_next_pfn() 830 unsigned long end_pfn; member 842 __register_nosave_region(unsigned long start_pfn, unsigned long end_pfn, in __register_nosave_region() argument 847 if (start_pfn >= end_pfn) in __register_nosave_region() 854 if (region->end_pfn == start_pfn) { in __register_nosave_region() 855 region->end_pfn = end_pfn; in __register_nosave_region() [all …]
|
/linux-4.1.27/drivers/hv/ |
D | hv_balloon.c | 440 unsigned long end_pfn; member 713 if ((start_pfn >= has->end_pfn)) in pfn_covered() 719 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered() 720 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered() 728 has->end_pfn += new_inc; in pfn_covered() 767 if ((start_pfn >= has->end_pfn)) in handle_pg_range() 796 if ((has->ha_end_pfn < has->end_pfn) && (pfn_cnt > 0)) { in handle_pg_range() 804 size = (has->end_pfn - has->ha_end_pfn); in handle_pg_range() 854 ha_region->end_pfn = rg_start + rg_size; in process_hot_add()
|
/linux-4.1.27/arch/parisc/mm/ |
D | init.c | 243 unsigned long end_pfn; in setup_bootmem() local 247 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; in setup_bootmem() 250 hole_pages = pmem_ranges[i].start_pfn - end_pfn; in setup_bootmem() 252 pmem_holes[npmem_holes].start_pfn = end_pfn; in setup_bootmem() 254 end_pfn += hole_pages; in setup_bootmem() 256 end_pfn += pmem_ranges[i].pages; in setup_bootmem() 259 pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn; in setup_bootmem()
|
/linux-4.1.27/arch/microblaze/mm/ |
D | init.c | 190 unsigned long start_pfn, end_pfn; in setup_memory() local 193 end_pfn = memblock_region_memory_end_pfn(reg); in setup_memory() 195 (end_pfn - start_pfn) << PAGE_SHIFT, in setup_memory()
|
/linux-4.1.27/fs/proc/ |
D | kcore.c | 250 unsigned long end_pfn; in kcore_update_ram() local 255 end_pfn = 0; in kcore_update_ram() 259 if (end_pfn < node_end) in kcore_update_ram() 260 end_pfn = node_end; in kcore_update_ram() 263 ret = walk_system_ram_range(0, end_pfn, &head, kclist_add_private); in kcore_update_ram()
|
/linux-4.1.27/arch/sh/include/asm/ |
D | mmzone.h | 42 unsigned long end_pfn);
|
/linux-4.1.27/arch/tile/mm/ |
D | init.c | 443 unsigned long end_pfn = node_lowmem_end_pfn[i]; in kernel_physical_mapping_init() local 445 unsigned long end_pfn = node_end_pfn[i]; in kernel_physical_mapping_init() 451 end_huge_pfn = end_pfn - (HPAGE_SIZE >> PAGE_SHIFT); in kernel_physical_mapping_init() 456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init() 459 while (pfn < end_pfn) { in kernel_physical_mapping_init()
|
/linux-4.1.27/drivers/base/ |
D | memory.c | 386 unsigned long start_pfn, end_pfn; in show_valid_zones() local 392 end_pfn = start_pfn + nr_pages; in show_valid_zones() 396 if (!test_pages_in_a_zone(start_pfn, end_pfn)) in show_valid_zones() 403 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones()
|
D | node.c | 459 unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_spanned_pages; in link_mem_sections() local 464 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in link_mem_sections()
|
/linux-4.1.27/arch/tile/kernel/ |
D | setup.c | 1670 insert_ram_resource(u64 start_pfn, u64 end_pfn, bool reserved) in insert_ram_resource() argument 1678 res->end = (end_pfn << PAGE_SHIFT) - 1; in insert_ram_resource() 1704 u64 end_pfn = node_end_pfn[i]; in request_standard_resources() local 1708 end_pfn > pci_reserve_start_pfn) { in request_standard_resources() 1709 if (end_pfn > pci_reserve_end_pfn) in request_standard_resources() 1711 end_pfn, 0); in request_standard_resources() 1712 end_pfn = pci_reserve_start_pfn; in request_standard_resources() 1715 insert_ram_resource(start_pfn, end_pfn, 0); in request_standard_resources()
|
D | pci_gx.c | 984 unsigned long end_pfn = node_end_pfn[j]; in pcibios_init() local 985 unsigned long nr_pages = end_pfn - start_pfn; in pcibios_init()
|
/linux-4.1.27/arch/arm64/mm/ |
D | init.c | 209 static inline void free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 218 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | setup.c | 738 unsigned long start_pfn, end_pfn; in setup_bootmem_allocator() local 759 end_pfn = memory_end >> PAGE_SHIFT; in setup_bootmem_allocator() 767 start_pfn, end_pfn); in setup_bootmem_allocator() 780 if (curr_pfn >= end_pfn) in setup_bootmem_allocator() 788 if (last_pfn > end_pfn) in setup_bootmem_allocator() 789 last_pfn = end_pfn; in setup_bootmem_allocator()
|
/linux-4.1.27/arch/x86/platform/efi/ |
D | efi.c | 554 u64 start_pfn, end_pfn, end; in old_map_region() local 561 end_pfn = PFN_UP(end); in old_map_region() 563 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in old_map_region()
|
/linux-4.1.27/arch/arm/mm/ |
D | init.c | 332 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 341 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
|
/linux-4.1.27/arch/x86/kernel/cpu/mtrr/ |
D | cleanup.c | 871 int __init mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument 963 if (range[i].end < end_pfn) in mtrr_trim_uncached_memory() 965 end_pfn); in mtrr_trim_uncached_memory()
|
/linux-4.1.27/kernel/ |
D | resource.c | 457 unsigned long pfn, end_pfn; in walk_system_ram_range() local 468 end_pfn = (res.end + 1) >> PAGE_SHIFT; in walk_system_ram_range() 469 if (end_pfn > pfn) in walk_system_ram_range() 470 ret = (*func)(pfn, end_pfn - pfn, arg); in walk_system_ram_range()
|
D | module.c | 1713 unsigned long end_pfn = PFN_DOWN((unsigned long)end); in set_page_attributes() local 1715 if (end_pfn > begin_pfn) in set_page_attributes() 1716 set(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_page_attributes() 1726 unsigned long end_pfn; in set_section_ro_nx() local 1743 end_pfn = PFN_UP((unsigned long)base + total_size); in set_section_ro_nx() 1744 if (end_pfn > begin_pfn) in set_section_ro_nx() 1745 set_memory_nx(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_section_ro_nx()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local 717 end_pfn = initial_pfn + total_nr_pages; in ehea_create_busmap_callback() 720 while (pfn < end_pfn) { in ehea_create_busmap_callback()
|
/linux-4.1.27/drivers/iommu/ |
D | intel-iommu.c | 2036 unsigned long nr_superpages, end_pfn; in __domain_mapping() local 2042 end_pfn = iov_pfn + nr_superpages * lvl_pages - 1; in __domain_mapping() 2048 dma_pte_free_pagetable(domain, iov_pfn, end_pfn); in __domain_mapping() 2487 unsigned long start_pfn, end_pfn; in si_domain_init() local 2490 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in si_domain_init() 2492 PFN_PHYS(start_pfn), PFN_PHYS(end_pfn)); in si_domain_init()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_mrmw.c | 2369 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehca_create_busmap_callback() local 2376 end_pfn = initial_pfn + total_nr_pages; in ehca_create_busmap_callback() 2379 while (pfn < end_pfn) { in ehca_create_busmap_callback()
|
/linux-4.1.27/Documentation/virtual/uml/ |
D | UserModeLinux-HOWTO.txt | 3746 bootmap_size = init_bootmem(start_pfn, end_pfn - start_pfn);
|