Home
last modified time | relevance | path

Searched refs:PAGES_PER_SECTION (Results 1 – 18 of 18) sorted by relevance

/linux-4.4.14/arch/x86/mm/
Dnuma_32.c55 start = round_down(start, PAGES_PER_SECTION); in memory_present()
56 end = round_up(end, PAGES_PER_SECTION); in memory_present()
57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present()
Dnuma.c543 if (pfn_align && pfn_align < PAGES_PER_SECTION) { in numa_register_memblks()
546 PFN_PHYS(PAGES_PER_SECTION) >> 20); in numa_register_memblks()
/linux-4.4.14/mm/
Dsparse.c176 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
206 nr_pages += PAGES_PER_SECTION; in node_memmap_size_bytes()
377 map = alloc_remap(nid, sizeof(struct page) * PAGES_PER_SECTION); in sparse_mem_map_populate()
381 size = PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); in sparse_mem_map_populate()
394 unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; in sparse_mem_maps_populate_node()
606 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in __kfree_section_memmap()
614 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in free_map_bootmem()
623 unsigned long memmap_size = sizeof(struct page) * PAGES_PER_SECTION; in __kmalloc_section_memmap()
652 get_order(sizeof(struct page) * PAGES_PER_SECTION)); in __kfree_section_memmap()
[all …]
Dpage_ext.c226 table_size = sizeof(struct page_ext) * PAGES_PER_SECTION; in init_section_page_ext()
259 table_size = sizeof(struct page_ext) * PAGES_PER_SECTION; in free_page_ext()
299 for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) { in online_page_ext()
308 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext()
322 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in offline_page_ext()
378 pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) { in page_ext_init()
Dmemory_hotplug.c202 mapsize = sizeof(struct page) * PAGES_PER_SECTION; in register_page_bootmem_info_section()
233 register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION); in register_page_bootmem_info_section()
275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
446 int nr_pages = PAGES_PER_SECTION; in __add_zone()
538 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SECTION) { in find_smallest_section_pfn()
566 for (; pfn >= start_pfn; pfn -= PAGES_PER_SECTION) { in find_biggest_section_pfn()
628 for (; pfn < zone_end_pfn; pfn += PAGES_PER_SECTION) { in shrink_zone_span()
696 for (; pfn < pgdat_end_pfn; pfn += PAGES_PER_SECTION) { in shrink_pgdat_span()
721 int nr_pages = PAGES_PER_SECTION; in __remove_zone()
777 BUG_ON(nr_pages % PAGES_PER_SECTION); in __remove_pages()
[all …]
Dsparse-vmemmap.c185 map = pfn_to_page(pnum * PAGES_PER_SECTION); in sparse_mem_map_populate()
187 end = (unsigned long)(map + PAGES_PER_SECTION); in sparse_mem_map_populate()
201 unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; in sparse_mem_maps_populate_node()
Dpage_alloc.c307 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in update_defer_init()
6438 pfn &= (PAGES_PER_SECTION-1); in pfn_to_bitidx()
/linux-4.4.14/arch/x86/include/asm/
Dmmzone_32.h29 #define PAGES_PER_SECTION (MAX_NR_PAGES/MAX_SECTIONS) macro
36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid()
/linux-4.4.14/drivers/base/
Dmemory.c135 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable()
199 for (i = 0; i < sections_per_block; i++, pfn += PAGES_PER_SECTION) { in pages_correctly_reserved()
204 for (j = 0; j < PAGES_PER_SECTION; j++) { in pages_correctly_reserved()
228 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_action()
391 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in show_valid_zones()
455 unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; in memory_probe_store()
467 PAGES_PER_SECTION << PAGE_SHIFT); in memory_probe_store()
Dnode.c391 sect_end_pfn += PAGES_PER_SECTION - 1; in register_mem_sect_under_node()
400 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node()
401 PAGES_PER_SECTION) - 1; in register_mem_sect_under_node()
440 sect_end_pfn = sect_start_pfn + PAGES_PER_SECTION - 1; in unregister_mem_sect_under_nodes()
468 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in link_mem_sections()
/linux-4.4.14/arch/arm64/mm/
Dinit.c252 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
271 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap()
272 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
/linux-4.4.14/include/linux/
Dmemory.h121 #define CONFIG_MEM_BLOCK_SIZE (PAGES_PER_SECTION<<PAGE_SHIFT)
Dmmzone.h1009 #define PAGES_PER_SECTION (1UL << PFN_SECTION_SHIFT) macro
1010 #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1))
1022 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
/linux-4.4.14/arch/arm/mm/
Dinit.c381 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
407 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap()
409 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
/linux-4.4.14/drivers/xen/
Dballoon.c266 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in additional_memory_resource()
320 balloon_hotplug = round_up(credit, PAGES_PER_SECTION); in reserve_additional_memory()
/linux-4.4.14/arch/powerpc/mm/
Dinit_64.c180 for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) in vmemmap_populated()
/linux-4.4.14/drivers/nvdimm/
Dnd.h33 ND_PFN_ALIGN = PAGES_PER_SECTION * PAGE_SIZE,
/linux-4.4.14/arch/powerpc/platforms/pseries/
Dhotplug-memory.c210 rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in lmb_is_removable()