/linux-4.4.14/arch/x86/mm/ |
D | init.c | 190 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument 193 if (start_pfn < end_pfn) { in save_mr() 197 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 268 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 284 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 286 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 288 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 290 if (end_pfn > limit_pfn) in split_mem_range() 291 end_pfn = limit_pfn; in split_mem_range() 292 if (start_pfn < end_pfn) { in split_mem_range() [all …]
|
D | numa_32.c | 65 unsigned long end_pfn) in node_memmap_size_bytes() argument 67 unsigned long nr_pages = end_pfn - start_pfn; in node_memmap_size_bytes()
|
D | init_32.c | 258 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local 269 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init() 298 if (pfn >= end_pfn) in kernel_physical_mapping_init() 306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 346 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init() 432 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument 439 start_pfn, end_pfn); in add_highpages_with_active_regions() 441 start_pfn, end_pfn); in add_highpages_with_active_regions()
|
D | numa_emulation.c | 34 unsigned long end_pfn = PFN_DOWN(end); in mem_hole_size() local 36 if (start_pfn < end_pfn) in mem_hole_size() 37 return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn)); in mem_hole_size()
|
D | pat.c | 386 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram() local 399 if (start_pfn < end_pfn) { in pat_pagerange_is_ram() 400 ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, in pat_pagerange_is_ram()
|
D | init_64.c | 677 unsigned long end_pfn = PFN_UP(start + size); in update_end_of_memory_vars() local 679 if (end_pfn > max_pfn) { in update_end_of_memory_vars() 680 max_pfn = end_pfn; in update_end_of_memory_vars() 681 max_low_pfn = end_pfn; in update_end_of_memory_vars()
|
/linux-4.4.14/arch/mips/loongson64/loongson-3/ |
D | numa.c | 128 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local 144 end_pfn = start_pfn + node_psize; in szmem() 149 start_pfn, end_pfn, num_physpages); in szmem() 153 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 158 end_pfn = start_pfn + node_psize; in szmem() 163 start_pfn, end_pfn, num_physpages); in szmem() 167 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 185 unsigned long start_pfn, end_pfn, freepfn; in node_mem_init() local 191 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 196 node, start_pfn, end_pfn, freepfn); in node_mem_init() [all …]
|
/linux-4.4.14/mm/ |
D | page_isolation.c | 158 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 166 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in start_isolate_page_range() 169 pfn < end_pfn; in start_isolate_page_range() 191 int undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 197 BUG_ON((end_pfn) & (pageblock_nr_pages - 1)); in undo_isolate_page_range() 199 pfn < end_pfn; in undo_isolate_page_range() 216 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument 221 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock() 240 if (pfn < end_pfn) in __test_page_isolated_in_pageblock() 245 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument [all …]
|
D | memory_hotplug.c | 247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 272 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 289 unsigned long end_pfn) in grow_zone_span() argument 299 zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - in grow_zone_span() 306 unsigned long end_pfn) in resize_zone() argument 310 if (end_pfn - start_pfn) { in resize_zone() 312 zone->spanned_pages = end_pfn - start_pfn; in resize_zone() 326 unsigned long end_pfn) in fix_zone_id() argument 332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id() [all …]
|
D | page_idle.c | 121 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 131 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 132 if (end_pfn > max_pfn) in page_idle_bitmap_read() 133 end_pfn = ALIGN(max_pfn, BITMAP_CHUNK_BITS); in page_idle_bitmap_read() 135 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 166 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 176 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 177 if (end_pfn > max_pfn) in page_idle_bitmap_write() 178 end_pfn = ALIGN(max_pfn, BITMAP_CHUNK_BITS); in page_idle_bitmap_write() 180 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
|
D | compaction.c | 92 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 98 end_pfn--; in pageblock_pfn_to_page() 100 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in pageblock_pfn_to_page() 108 end_page = pfn_to_page(end_pfn); in pageblock_pfn_to_page() 214 unsigned long end_pfn = zone_end_pfn(zone); in __reset_isolation_suitable() local 220 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable() 396 unsigned long end_pfn, in isolate_freepages_block() argument 409 for (; blockpfn < end_pfn; blockpfn++, cursor++) { in isolate_freepages_block() 510 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block() 511 blockpfn = end_pfn; in isolate_freepages_block() [all …]
|
D | sparse.c | 145 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument 156 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 159 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 160 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 163 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 165 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 195 unsigned long end_pfn) in node_memmap_size_bytes() argument 200 mminit_validate_memmodel_limits(&start_pfn, &end_pfn); in node_memmap_size_bytes() 201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
|
D | page_alloc.c | 957 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local 959 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 1174 unsigned long pfn, end_pfn; in deferred_init_memmap() local 1180 end_pfn = min(walk_end, zone_end_pfn(zone)); in deferred_init_memmap() 1187 for (; pfn < end_pfn; pfn++) { in deferred_init_memmap() 1245 first_init_pfn = max(end_pfn, first_init_pfn); in deferred_init_memmap() 1535 unsigned long start_pfn, end_pfn; in move_freepages_block() local 1542 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block() 1547 if (!zone_spans_pfn(zone, end_pfn)) in move_freepages_block() 4492 unsigned long end_pfn = start_pfn + size; in memmap_init_zone() local [all …]
|
D | internal.h | 221 unsigned long start_pfn, unsigned long end_pfn); 224 unsigned long low_pfn, unsigned long end_pfn); 405 unsigned long *end_pfn); 408 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
|
D | nobootmem.c | 114 unsigned long end_pfn = min_t(unsigned long, in __free_memory_core() local 117 if (start_pfn > end_pfn) in __free_memory_core() 120 __free_pages_memory(start_pfn, end_pfn); in __free_memory_core() 122 return end_pfn - start_pfn; in __free_memory_core()
|
D | page_owner.c | 219 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone() local 230 for (; pfn < end_pfn; ) { in init_pages_in_zone() 237 block_end_pfn = min(block_end_pfn, end_pfn); in init_pages_in_zone()
|
D | page_ext.c | 368 unsigned long start_pfn, end_pfn; in page_ext_init() local 371 end_pfn = node_end_pfn(nid); in page_ext_init() 377 for (pfn = start_pfn; pfn < end_pfn; in page_ext_init()
|
D | memblock.c | 1444 unsigned long start_pfn, end_pfn; in memblock_mem_size() local 1448 end_pfn = memblock_region_memory_end_pfn(r); in memblock_mem_size() 1450 end_pfn = min_t(unsigned long, end_pfn, limit_pfn); in memblock_mem_size() 1451 pages += end_pfn - start_pfn; in memblock_mem_size() 1524 unsigned long *start_pfn, unsigned long *end_pfn) in memblock_search_pfn_nid() argument 1533 *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); in memblock_search_pfn_nid()
|
D | bootmem.c | 790 unsigned long end_pfn; in __alloc_bootmem_node_high() local 796 end_pfn = pgdat_end_pfn(pgdat); in __alloc_bootmem_node_high() 798 if (end_pfn > MAX_DMA32_PFN + (128 >> (20 - PAGE_SHIFT)) && in __alloc_bootmem_node_high()
|
D | vmstat.c | 1004 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print() local 1007 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print() 1055 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() local 1068 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 1075 block_end_pfn = min(block_end_pfn, end_pfn); in pagetypeinfo_showmixedcount_print()
|
D | ksm.c | 2013 unsigned long end_pfn) in ksm_check_stable_tree() argument 2025 stable_node->kpfn < end_pfn) { in ksm_check_stable_tree() 2040 stable_node->kpfn < end_pfn) in ksm_check_stable_tree()
|
D | hugetlb.c | 1030 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_gigantic_page() local 1031 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE); in __alloc_gigantic_page() 1037 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_gigantic() local 1040 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_gigantic() 1434 void dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) in dissolve_free_huge_pages() argument 1442 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) in dissolve_free_huge_pages()
|
D | kmemleak.c | 1384 unsigned long end_pfn = node_end_pfn(i); in kmemleak_scan() local 1387 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in kmemleak_scan()
|
/linux-4.4.14/arch/sh/mm/ |
D | numa.c | 30 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 37 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 44 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 53 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node() 56 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node() 60 start_pfn, end_pfn); in setup_bootmem_node() 62 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
|
D | init.c | 195 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 200 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 204 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat() 219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 225 unsigned long end_pfn; in bootmem_init_one_node() local 234 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node() 242 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node() 244 free_bootmem_with_active_regions(nid, end_pfn); in bootmem_init_one_node() 270 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 272 end_pfn = memblock_region_memory_end_pfn(reg); in do_init_bootmem() [all …]
|
/linux-4.4.14/arch/metag/mm/ |
D | numa.c | 33 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 40 end_pfn = end >> PAGE_SHIFT; in setup_bootmem_node() 45 PFN_PHYS(end_pfn - start_pfn), in setup_bootmem_node() 56 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node() 59 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in setup_bootmem_node() 63 start_pfn, end_pfn); in setup_bootmem_node() 65 free_bootmem_with_active_regions(nid, end_pfn); in setup_bootmem_node()
|
D | init.c | 111 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 116 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 120 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); in allocate_pgdat() 136 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 142 unsigned long end_pfn; in bootmem_init_one_node() local 151 end_pfn = pgdat_end_pfn(p); in bootmem_init_one_node() 153 if (end_pfn > max_low_pfn) in bootmem_init_one_node() 154 end_pfn = max_low_pfn; in bootmem_init_one_node() 157 total_pages = bootmem_bootmap_pages(end_pfn - p->node_start_pfn); in bootmem_init_one_node() 163 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); in bootmem_init_one_node() [all …]
|
/linux-4.4.14/arch/unicore32/mm/ |
D | init.c | 132 unsigned long end_pfn) in uc32_bootmem_init() argument 143 boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); in uc32_bootmem_init() 145 __pfn_to_phys(end_pfn)); in uc32_bootmem_init() 153 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); in uc32_bootmem_init() 160 if (end >= end_pfn) in uc32_bootmem_init() 161 end = end_pfn; in uc32_bootmem_init() 173 if (end >= end_pfn) in uc32_bootmem_init() 174 end = end_pfn; in uc32_bootmem_init() 317 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 326 end_pg = pfn_to_page(end_pfn); in free_memmap()
|
/linux-4.4.14/include/trace/events/ |
D | compaction.h | 72 unsigned long end_pfn, 76 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 80 __field(unsigned long, end_pfn) 87 __entry->end_pfn = end_pfn; 94 __entry->end_pfn, 103 unsigned long end_pfn, 107 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 114 unsigned long end_pfn, 118 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
|
/linux-4.4.14/arch/sparc/mm/ |
D | init_32.c | 78 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 80 if (end_pfn <= max_low_pfn) in calc_highpages() 86 nr += end_pfn - start_pfn; in calc_highpages() 276 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 281 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 284 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 325 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 327 if (end_pfn <= highstart_pfn) in mem_init() 333 map_high_region(start_pfn, end_pfn); in mem_init()
|
D | init_64.c | 862 unsigned long start_pfn, end_pfn; in allocate_node_data() local 879 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_node_data() 881 p->node_spanned_pages = end_pfn - start_pfn; in allocate_node_data() 1403 unsigned long end_pfn; in bootmem_init() local 1405 end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; in bootmem_init() 1406 max_pfn = max_low_pfn = end_pfn; in bootmem_init() 1420 return end_pfn; in bootmem_init() 2077 unsigned long end_pfn, shift, phys_base; in paging_init() local 2245 last_valid_pfn = end_pfn = bootmem_init(phys_base); in paging_init() 2269 max_zone_pfns[ZONE_NORMAL] = end_pfn; in paging_init()
|
/linux-4.4.14/arch/s390/numa/ |
D | numa.c | 106 unsigned long start_pfn, end_pfn; in numa_setup_memory() local 111 end_pfn = 0; in numa_setup_memory() 115 if (t_end > end_pfn) in numa_setup_memory() 116 end_pfn = t_end; in numa_setup_memory() 118 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in numa_setup_memory()
|
/linux-4.4.14/arch/mips/sgi-ip27/ |
D | ip27-memory.c | 393 unsigned long start_pfn, end_pfn; in node_mem_init() local 395 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 405 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 413 start_pfn, end_pfn); in node_mem_init() 414 free_bootmem_with_active_regions(node, end_pfn); in node_mem_init() 467 unsigned long start_pfn, end_pfn; in paging_init() local 469 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in paging_init() 471 if (end_pfn > max_low_pfn) in paging_init() 472 max_low_pfn = end_pfn; in paging_init()
|
/linux-4.4.14/include/linux/ |
D | page-isolation.h | 51 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 59 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 65 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
D | memory_hotplug.h | 267 extern int walk_memory_range(unsigned long start_pfn, unsigned long end_pfn,
|
D | memblock.h | 189 unsigned long *end_pfn);
|
D | hugetlb.h | 458 unsigned long end_pfn);
|
D | mm.h | 1776 unsigned long end_pfn); 1778 unsigned long end_pfn); 1780 unsigned long *start_pfn, unsigned long *end_pfn);
|
/linux-4.4.14/arch/s390/kernel/ |
D | early.c | 83 unsigned int i, stext_pfn, eshared_pfn, end_pfn, min_size; in create_kernel_nss() local 116 end_pfn = PFN_UP(__pa(&_end)); in create_kernel_nss() 117 min_size = end_pfn << 2; in create_kernel_nss() 122 eshared_pfn - 1, eshared_pfn, end_pfn); in create_kernel_nss() 202 unsigned long end_pfn, init_pfn; in init_kernel_storage_key() local 204 end_pfn = PFN_UP(__pa(&_end)); in init_kernel_storage_key() 206 for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++) in init_kernel_storage_key()
|
/linux-4.4.14/arch/x86/xen/ |
D | setup.c | 254 unsigned long end_pfn, unsigned long nr_pages) in xen_set_identity_and_release_chunk() argument 259 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 262 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk() 281 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk() 389 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_set_identity_and_remap_chunk() argument 394 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 436 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk() 445 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_count_remap_pages() argument 451 return remap_pages + min(end_pfn, nr_pages) - start_pfn; in xen_count_remap_pages() 455 unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, in xen_foreach_remap_area() argument [all …]
|
/linux-4.4.14/arch/sh/kernel/ |
D | swsusp.c | 25 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | setup.c | 192 unsigned long end_pfn) in __add_active_range() argument 200 end = end_pfn << PAGE_SHIFT; in __add_active_range() 209 start_pfn, end_pfn); in __add_active_range() 233 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
/linux-4.4.14/arch/x86/kernel/ |
D | e820.c | 765 unsigned long end_pfn; in e820_end_pfn() local 775 end_pfn = (ei->addr + ei->size) >> PAGE_SHIFT; in e820_end_pfn() 779 if (end_pfn > limit_pfn) { in e820_end_pfn() 783 if (end_pfn > last_pfn) in e820_end_pfn() 784 last_pfn = end_pfn; in e820_end_pfn() 1130 unsigned long start_pfn, end_pfn; in memblock_find_dma_reserve() local 1140 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { in memblock_find_dma_reserve() 1142 end_pfn = min(end_pfn, MAX_DMA_PFN); in memblock_find_dma_reserve() 1143 nr_pages += end_pfn - start_pfn; in memblock_find_dma_reserve() 1149 end_pfn = min_t(unsigned long, PFN_DOWN(end), MAX_DMA_PFN); in memblock_find_dma_reserve() [all …]
|
D | amd_gart_64.c | 738 unsigned long start_pfn, end_pfn; in gart_iommu_init() local 769 end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); in gart_iommu_init() 772 if (!pfn_range_is_mapped(start_pfn, end_pfn)) in gart_iommu_init() 773 init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); in gart_iommu_init()
|
/linux-4.4.14/arch/mn10300/kernel/ |
D | setup.c | 95 unsigned long kstart_pfn, start_pfn, free_pfn, end_pfn; in setup_arch() local 128 end_pfn = PFN_DOWN(__pa(memory_end)); in setup_arch() 133 end_pfn); in setup_arch() 140 PFN_PHYS(end_pfn - free_pfn)); in setup_arch()
|
/linux-4.4.14/drivers/iommu/ |
D | dma-iommu.c | 91 unsigned long order, base_pfn, end_pfn; in iommu_dma_init_domain() local 99 end_pfn = (base + size - 1) >> order; in iommu_dma_init_domain() 111 end_pfn = min_t(unsigned long, end_pfn, in iommu_dma_init_domain() 119 end_pfn < iovad->dma_32bit_pfn) { in iommu_dma_init_domain() 123 iovad->dma_32bit_pfn = end_pfn; in iommu_dma_init_domain() 125 init_iova_domain(iovad, 1UL << order, base_pfn, end_pfn); in iommu_dma_init_domain()
|
D | intel-iommu.c | 2174 unsigned long nr_superpages, end_pfn; in __domain_mapping() local 2180 end_pfn = iov_pfn + nr_superpages * lvl_pages - 1; in __domain_mapping() 2186 dma_pte_free_pagetable(domain, iov_pfn, end_pfn); in __domain_mapping() 2623 unsigned long start_pfn, end_pfn; in si_domain_init() local 2626 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in si_domain_init() 2628 PFN_PHYS(start_pfn), PFN_PHYS(end_pfn)); in si_domain_init()
|
/linux-4.4.14/arch/powerpc/mm/ |
D | numa.c | 90 static int __init fake_numa_create_new_node(unsigned long end_pfn, in fake_numa_create_new_node() argument 121 if ((end_pfn << PAGE_SHIFT) > mem) { in fake_numa_create_new_node() 826 unsigned long start_pfn, end_pfn; in setup_nonnuma() local 837 end_pfn = memblock_region_memory_end_pfn(reg); in setup_nonnuma() 839 fake_numa_create_new_node(end_pfn, &nid); in setup_nonnuma() 841 PFN_PHYS(end_pfn - start_pfn), in setup_nonnuma() 922 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 924 u64 spanned_pages = end_pfn - start_pfn; in setup_node_data() 933 (end_pfn << PAGE_SHIFT) - 1); in setup_node_data() 976 unsigned long start_pfn, end_pfn; in initmem_init() local [all …]
|
D | mem.c | 174 unsigned long end_pfn = start_pfn + nr_pages; in walk_system_ram_range() local 180 tend = min(end_pfn, memblock_region_memory_end_pfn(reg)); in walk_system_ram_range()
|
/linux-4.4.14/arch/alpha/mm/ |
D | numa.c | 307 unsigned long end_pfn = bdata->node_low_pfn; in paging_init() local 309 if (dma_local_pfn >= end_pfn - start_pfn) in paging_init() 310 zones_size[ZONE_DMA] = end_pfn - start_pfn; in paging_init() 313 zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn; in paging_init()
|
/linux-4.4.14/arch/unicore32/kernel/ |
D | hibernate.c | 148 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
/linux-4.4.14/arch/x86/include/asm/ |
D | mtrr.h | 49 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 79 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
|
D | highmem.h | 75 unsigned long end_pfn);
|
D | page_types.h | 66 bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
|
/linux-4.4.14/kernel/power/ |
D | snapshot.c | 290 unsigned long end_pfn; /* Zone end page frame + 1 */ member 444 zone->end_pfn = end; in create_zone_bm_rtree() 643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 808 pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; in memory_bm_next_pfn() 830 unsigned long end_pfn; member 842 __register_nosave_region(unsigned long start_pfn, unsigned long end_pfn, in __register_nosave_region() argument 847 if (start_pfn >= end_pfn) in __register_nosave_region() 854 if (region->end_pfn == start_pfn) { in __register_nosave_region() 855 region->end_pfn = end_pfn; in __register_nosave_region() [all …]
|
/linux-4.4.14/arch/parisc/mm/ |
D | init.c | 245 unsigned long end_pfn; in setup_bootmem() local 249 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; in setup_bootmem() 252 hole_pages = pmem_ranges[i].start_pfn - end_pfn; in setup_bootmem() 254 pmem_holes[npmem_holes].start_pfn = end_pfn; in setup_bootmem() 256 end_pfn += hole_pages; in setup_bootmem() 258 end_pfn += pmem_ranges[i].pages; in setup_bootmem() 261 pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn; in setup_bootmem()
|
/linux-4.4.14/drivers/hv/ |
D | hv_balloon.c | 442 unsigned long end_pfn; member 717 if ((start_pfn >= has->end_pfn)) in pfn_covered() 723 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered() 724 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered() 732 has->end_pfn += new_inc; in pfn_covered() 771 if ((start_pfn >= has->end_pfn)) in handle_pg_range() 800 if ((has->ha_end_pfn < has->end_pfn) && (pfn_cnt > 0)) { in handle_pg_range() 808 size = (has->end_pfn - has->ha_end_pfn); in handle_pg_range() 858 ha_region->end_pfn = rg_start + rg_size; in process_hot_add()
|
/linux-4.4.14/arch/microblaze/mm/ |
D | init.c | 190 unsigned long start_pfn, end_pfn; in setup_memory() local 193 end_pfn = memblock_region_memory_end_pfn(reg); in setup_memory() 195 (end_pfn - start_pfn) << PAGE_SHIFT, in setup_memory()
|
/linux-4.4.14/fs/proc/ |
D | kcore.c | 250 unsigned long end_pfn; in kcore_update_ram() local 255 end_pfn = 0; in kcore_update_ram() 259 if (end_pfn < node_end) in kcore_update_ram() 260 end_pfn = node_end; in kcore_update_ram() 263 ret = walk_system_ram_range(0, end_pfn, &head, kclist_add_private); in kcore_update_ram()
|
/linux-4.4.14/arch/sh/include/asm/ |
D | mmzone.h | 42 unsigned long end_pfn);
|
/linux-4.4.14/arch/tile/mm/ |
D | init.c | 443 unsigned long end_pfn = node_lowmem_end_pfn[i]; in kernel_physical_mapping_init() local 445 unsigned long end_pfn = node_end_pfn[i]; in kernel_physical_mapping_init() 451 end_huge_pfn = end_pfn - (HPAGE_SIZE >> PAGE_SHIFT); in kernel_physical_mapping_init() 456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init() 459 while (pfn < end_pfn) { in kernel_physical_mapping_init()
|
/linux-4.4.14/drivers/base/ |
D | memory.c | 390 unsigned long start_pfn, end_pfn; in show_valid_zones() local 396 end_pfn = start_pfn + nr_pages; in show_valid_zones() 400 if (!test_pages_in_a_zone(start_pfn, end_pfn)) in show_valid_zones() 407 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones()
|
D | node.c | 463 unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_spanned_pages; in link_mem_sections() local 468 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in link_mem_sections()
|
/linux-4.4.14/arch/tile/kernel/ |
D | setup.c | 1670 insert_ram_resource(u64 start_pfn, u64 end_pfn, bool reserved) in insert_ram_resource() argument 1678 res->end = (end_pfn << PAGE_SHIFT) - 1; in insert_ram_resource() 1704 u64 end_pfn = node_end_pfn[i]; in request_standard_resources() local 1708 end_pfn > pci_reserve_start_pfn) { in request_standard_resources() 1709 if (end_pfn > pci_reserve_end_pfn) in request_standard_resources() 1711 end_pfn, 0); in request_standard_resources() 1712 end_pfn = pci_reserve_start_pfn; in request_standard_resources() 1715 insert_ram_resource(start_pfn, end_pfn, 0); in request_standard_resources()
|
D | pci_gx.c | 984 unsigned long end_pfn = node_end_pfn[j]; in pcibios_init() local 985 unsigned long nr_pages = end_pfn - start_pfn; in pcibios_init()
|
/linux-4.4.14/arch/arm64/mm/ |
D | init.c | 210 static inline void free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 219 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
|
/linux-4.4.14/arch/blackfin/kernel/ |
D | setup.c | 738 unsigned long start_pfn, end_pfn; in setup_bootmem_allocator() local 759 end_pfn = memory_end >> PAGE_SHIFT; in setup_bootmem_allocator() 767 start_pfn, end_pfn); in setup_bootmem_allocator() 780 if (curr_pfn >= end_pfn) in setup_bootmem_allocator() 788 if (last_pfn > end_pfn) in setup_bootmem_allocator() 789 last_pfn = end_pfn; in setup_bootmem_allocator()
|
/linux-4.4.14/tools/perf/scripts/python/ |
D | compaction-times.py | 220 common_callchain, start_pfn, end_pfn, nr_scanned, nr_taken): argument 227 common_callchain, start_pfn, end_pfn, nr_scanned, nr_taken): argument
|
/linux-4.4.14/arch/x86/platform/efi/ |
D | efi.c | 580 u64 start_pfn, end_pfn, end; in old_map_region() local 587 end_pfn = PFN_UP(end); in old_map_region() 589 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in old_map_region()
|
/linux-4.4.14/arch/x86/kernel/cpu/mtrr/ |
D | cleanup.c | 872 int __init mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument 964 if (range[i].end < end_pfn) in mtrr_trim_uncached_memory() 966 end_pfn); in mtrr_trim_uncached_memory()
|
/linux-4.4.14/arch/arm/mm/ |
D | init.c | 334 free_memmap(unsigned long start_pfn, unsigned long end_pfn) in free_memmap() argument 343 end_pg = pfn_to_page(end_pfn - 1) + 1; in free_memmap()
|
/linux-4.4.14/kernel/ |
D | resource.c | 457 unsigned long pfn, end_pfn; in walk_system_ram_range() local 468 end_pfn = (res.end + 1) >> PAGE_SHIFT; in walk_system_ram_range() 469 if (end_pfn > pfn) in walk_system_ram_range() 470 ret = (*func)(pfn, end_pfn - pfn, arg); in walk_system_ram_range()
|
D | module.c | 1883 unsigned long end_pfn = PFN_DOWN((unsigned long)end); in set_page_attributes() local 1885 if (end_pfn > begin_pfn) in set_page_attributes() 1886 set(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_page_attributes() 1896 unsigned long end_pfn; in set_section_ro_nx() local 1913 end_pfn = PFN_UP((unsigned long)base + total_size); in set_section_ro_nx() 1914 if (end_pfn > begin_pfn) in set_section_ro_nx() 1915 set_memory_nx(begin_pfn << PAGE_SHIFT, end_pfn - begin_pfn); in set_section_ro_nx()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local 717 end_pfn = initial_pfn + total_nr_pages; in ehea_create_busmap_callback() 720 while (pfn < end_pfn) { in ehea_create_busmap_callback()
|
/linux-4.4.14/drivers/staging/rdma/ehca/ |
D | ehca_mrmw.c | 2369 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehca_create_busmap_callback() local 2376 end_pfn = initial_pfn + total_nr_pages; in ehca_create_busmap_callback() 2379 while (pfn < end_pfn) { in ehca_create_busmap_callback()
|
/linux-4.4.14/Documentation/virtual/uml/ |
D | UserModeLinux-HOWTO.txt | 3746 bootmap_size = init_bootmem(start_pfn, end_pfn - start_pfn);
|