/linux-4.4.14/arch/x86/mm/ |
H A D | highmem_32.c | 118 unsigned long zone_start_pfn, zone_end_pfn; for_each_zone() local 123 zone_start_pfn = zone->zone_start_pfn; for_each_zone() 124 zone_end_pfn = zone_start_pfn + zone->spanned_pages; for_each_zone() 128 zone->name, nid, zone_start_pfn, zone_end_pfn); for_each_zone() 130 add_highpages_with_active_regions(nid, zone_start_pfn, for_each_zone()
|
/linux-4.4.14/arch/c6x/mm/ |
H A D | init.c | 53 pgdat->node_zones[ZONE_NORMAL].zone_start_pfn = paging_init()
|
/linux-4.4.14/mm/ |
H A D | memory_hotplug.c | 296 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) grow_zone_span() 297 zone->zone_start_pfn = start_pfn; grow_zone_span() 300 zone->zone_start_pfn; grow_zone_span() 311 zone->zone_start_pfn = start_pfn; resize_zone() 318 zone->zone_start_pfn = 0; resize_zone() 364 if (start_pfn > z2->zone_start_pfn) move_pfn_range_left() 367 if (end_pfn <= z2->zone_start_pfn) move_pfn_range_left() 372 z1_start_pfn = z1->zone_start_pfn; move_pfn_range_left() 403 if (z1->zone_start_pfn > start_pfn) move_pfn_range_right() 418 resize_zone(z1, z1->zone_start_pfn, start_pfn); move_pfn_range_right() 587 unsigned long zone_start_pfn = zone->zone_start_pfn; shrink_zone_span() local 595 if (zone_start_pfn == start_pfn) { shrink_zone_span() 598 * shrink zone->zone_start_pfn and zone->zone_spanned_pages. shrink_zone_span() 605 zone->zone_start_pfn = pfn; shrink_zone_span() 615 pfn = find_biggest_section_pfn(nid, zone, zone_start_pfn, shrink_zone_span() 618 zone->spanned_pages = pfn - zone_start_pfn + 1; shrink_zone_span() 627 pfn = zone_start_pfn; shrink_zone_span() 647 zone->zone_start_pfn = 0; shrink_zone_span() 1215 if (movable_zone->zone_start_pfn <= start_pfn) should_add_memory_movable()
|
H A D | page_owner.c | 218 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; init_pages_in_zone() 223 pfn = zone->zone_start_pfn; init_pages_in_zone()
|
H A D | page_alloc.c | 358 start_pfn = zone->zone_start_pfn; page_outside_zone_boundaries() 937 if (pfn >= zone->zone_start_pfn && pfn < zone_end_pfn(zone)) init_reserved_page() 1184 if (pfn < zone->zone_start_pfn) deferred_init_memmap() 1185 pfn = zone->zone_start_pfn; deferred_init_memmap() 2019 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) mark_free_pages() 4768 unsigned long zone_start_pfn, init_currently_empty_zone() 4778 zone->zone_start_pfn = zone_start_pfn; init_currently_empty_zone() 4784 zone_start_pfn, (zone_start_pfn + size)); init_currently_empty_zone() 4922 unsigned long *zone_start_pfn, adjust_zone_range_for_zone_movable() 4929 *zone_start_pfn = zone_movable_pfn[nid]; adjust_zone_range_for_zone_movable() 4934 } else if (*zone_start_pfn < zone_movable_pfn[nid] && adjust_zone_range_for_zone_movable() 4939 } else if (*zone_start_pfn >= zone_movable_pfn[nid]) adjust_zone_range_for_zone_movable() 4940 *zone_start_pfn = *zone_end_pfn; adjust_zone_range_for_zone_movable() 4954 unsigned long zone_start_pfn, zone_end_pfn; zone_spanned_pages_in_node() local 4961 zone_start_pfn = arch_zone_lowest_possible_pfn[zone_type]; zone_spanned_pages_in_node() 4965 &zone_start_pfn, &zone_end_pfn); zone_spanned_pages_in_node() 4968 if (zone_end_pfn < node_start_pfn || zone_start_pfn > node_end_pfn) zone_spanned_pages_in_node() 4973 zone_start_pfn = max(zone_start_pfn, node_start_pfn); zone_spanned_pages_in_node() 4976 return zone_end_pfn - zone_start_pfn; zone_spanned_pages_in_node() 5021 unsigned long zone_start_pfn, zone_end_pfn; zone_absent_pages_in_node() local 5027 zone_start_pfn = clamp(node_start_pfn, zone_low, zone_high); zone_absent_pages_in_node() 5032 &zone_start_pfn, &zone_end_pfn); zone_absent_pages_in_node() 5033 return __absent_pages_in_range(nid, zone_start_pfn, zone_end_pfn); zone_absent_pages_in_node() 5101 static unsigned long __init usemap_size(unsigned long zone_start_pfn, unsigned long zonesize) usemap_size() argument 5105 zonesize += zone_start_pfn & (pageblock_nr_pages-1); usemap_size() 5116 unsigned long zone_start_pfn, setup_usemap() 5119 unsigned long usemapsize = usemap_size(zone_start_pfn, zonesize); setup_usemap() 5128 unsigned long zone_start_pfn, unsigned long zonesize) {} 5200 unsigned long zone_start_pfn = pgdat->node_start_pfn; free_area_init_core() local 5280 setup_usemap(pgdat, zone, zone_start_pfn, size); free_area_init_core() 5281 ret = init_currently_empty_zone(zone, zone_start_pfn, size); free_area_init_core() 5283 memmap_init(size, nid, j, zone_start_pfn); free_area_init_core() 5284 zone_start_pfn += size; free_area_init_core() 6441 pfn = pfn - round_down(zone->zone_start_pfn, pageblock_nr_pages); pfn_to_bitidx() 4767 init_currently_empty_zone(struct zone *zone, unsigned long zone_start_pfn, unsigned long size) init_currently_empty_zone() argument 4918 adjust_zone_range_for_zone_movable(int nid, unsigned long zone_type, unsigned long node_start_pfn, unsigned long node_end_pfn, unsigned long *zone_start_pfn, unsigned long *zone_end_pfn) adjust_zone_range_for_zone_movable() argument 5114 setup_usemap(struct pglist_data *pgdat, struct zone *zone, unsigned long zone_start_pfn, unsigned long zonesize) setup_usemap() argument 5127 setup_usemap(struct pglist_data *pgdat, struct zone *zone, unsigned long zone_start_pfn, unsigned long zonesize) setup_usemap() argument
|
H A D | vmstat.c | 1003 unsigned long start_pfn = zone->zone_start_pfn; pagetypeinfo_showblockcount_print() 1054 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; pagetypeinfo_showmixedcount_print() 1061 pfn = zone->zone_start_pfn; pagetypeinfo_showmixedcount_print() 1263 zone->zone_start_pfn,
|
H A D | compaction.c | 201 zone->compact_cached_migrate_pfn[0] = zone->zone_start_pfn; reset_cached_positions() 202 zone->compact_cached_migrate_pfn[1] = zone->zone_start_pfn; reset_cached_positions() 213 unsigned long start_pfn = zone->zone_start_pfn; __reset_isolation_suitable() 1334 unsigned long start_pfn = zone->zone_start_pfn; compact_zone()
|
H A D | hugetlb.c | 1076 pfn = ALIGN(z->zone_start_pfn, nr_pages); alloc_gigantic_page()
|
/linux-4.4.14/include/linux/ |
H A D | mmzone.h | 385 /* zone_start_pfn == zone_start_paddr >> PAGE_SHIFT */ 386 unsigned long zone_start_pfn; member in struct:zone 391 * spanned_pages = zone_end_pfn - zone_start_pfn; 410 * zone_start_pfn and spanned_pages are protected by span_seqlock. 551 return zone->zone_start_pfn + zone->spanned_pages; zone_end_pfn() 556 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); zone_spans_pfn()
|
H A D | mm.h | 1676 unsigned long zone_start_pfn, unsigned long *zholes_size);
|
/linux-4.4.14/kernel/power/ |
H A D | snapshot.c | 524 zone_start = zone->zone_start_pfn; for_each_populated_zone() 1122 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) for_each_populated_zone() 1186 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) for_each_populated_zone() 1279 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) for_each_populated_zone() 2016 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) for_each_populated_zone()
|
/linux-4.4.14/drivers/base/ |
H A D | memory.c | 414 if (start_pfn == zone->zone_start_pfn) show_valid_zones()
|
/linux-4.4.14/arch/tile/mm/ |
H A D | init.c | 704 start = z->zone_start_pfn; for_each_zone()
|