PAGES_PER_SECTION 374 arch/arm/mm/init.c ALIGN(prev_end, PAGES_PER_SECTION)); PAGES_PER_SECTION 400 arch/arm/mm/init.c if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) PAGES_PER_SECTION 402 arch/arm/mm/init.c ALIGN(prev_end, PAGES_PER_SECTION)); PAGES_PER_SECTION 504 arch/arm64/mm/init.c start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); PAGES_PER_SECTION 523 arch/arm64/mm/init.c if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) PAGES_PER_SECTION 524 arch/arm64/mm/init.c free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); PAGES_PER_SECTION 96 arch/powerpc/mm/init_64.c for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) PAGES_PER_SECTION 368 arch/powerpc/platforms/pseries/hotplug-memory.c rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION); PAGES_PER_SECTION 37 arch/x86/include/asm/mmzone_32.h return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); PAGES_PER_SECTION 1519 arch/x86/mm/init_64.c if (end - start < PAGES_PER_SECTION * sizeof(struct page)) PAGES_PER_SECTION 559 arch/x86/mm/numa.c if (pfn_align && pfn_align < PAGES_PER_SECTION) { PAGES_PER_SECTION 562 arch/x86/mm/numa.c PFN_PHYS(PAGES_PER_SECTION) >> 20); PAGES_PER_SECTION 54 arch/x86/mm/numa_32.c start = round_down(start, PAGES_PER_SECTION); PAGES_PER_SECTION 55 arch/x86/mm/numa_32.c end = round_up(end, PAGES_PER_SECTION); PAGES_PER_SECTION 56 arch/x86/mm/numa_32.c for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 57 arch/x86/mm/numa_32.c physnode_map[pfn / PAGES_PER_SECTION] = nid; PAGES_PER_SECTION 191 drivers/base/memory.c section_nr, pfn, pfn + PAGES_PER_SECTION); PAGES_PER_SECTION 195 drivers/base/memory.c section_nr, pfn, pfn + PAGES_PER_SECTION); PAGES_PER_SECTION 199 drivers/base/memory.c section_nr, pfn, pfn + PAGES_PER_SECTION); PAGES_PER_SECTION 202 drivers/base/memory.c pfn += PAGES_PER_SECTION; PAGES_PER_SECTION 217 drivers/base/memory.c unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; PAGES_PER_SECTION 383 drivers/base/memory.c unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; PAGES_PER_SECTION 478 drivers/base/memory.c unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; PAGES_PER_SECTION 779 drivers/base/node.c pfn = round_down(pfn + PAGES_PER_SECTION, PAGES_PER_SECTION 780 drivers/base/node.c PAGES_PER_SECTION) - 1; PAGES_PER_SECTION 263 drivers/xen/balloon.c PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); PAGES_PER_SECTION 304 drivers/xen/balloon.c balloon_hotplug = round_up(credit, PAGES_PER_SECTION); PAGES_PER_SECTION 123 include/linux/memory.h #define CONFIG_MEM_BLOCK_SIZE (PAGES_PER_SECTION<<PAGE_SHIFT) PAGES_PER_SECTION 1138 include/linux/mmzone.h #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1)) PAGES_PER_SECTION 1156 include/linux/mmzone.h #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK) PAGES_PER_SECTION 184 mm/memory_hotplug.c mapsize = sizeof(struct page) * PAGES_PER_SECTION; PAGES_PER_SECTION 213 mm/memory_hotplug.c register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION); PAGES_PER_SECTION 241 mm/memory_hotplug.c for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 271 mm/memory_hotplug.c min_align = PAGES_PER_SECTION; PAGES_PER_SECTION 315 mm/memory_hotplug.c pfns = min(nr_pages, PAGES_PER_SECTION PAGES_PER_SECTION 535 mm/memory_hotplug.c pfns = min(nr_pages, PAGES_PER_SECTION PAGES_PER_SECTION 1225 mm/memory_hotplug.c pfn = sec_end_pfn, sec_end_pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 427 mm/page_alloc.c if ((nr_initialised > PAGES_PER_SECTION) && PAGES_PER_SECTION 428 mm/page_alloc.c (pfn & (PAGES_PER_SECTION - 1)) == 0) { PAGES_PER_SECTION 462 mm/page_alloc.c pfn &= (PAGES_PER_SECTION-1); PAGES_PER_SECTION 1846 mm/page_alloc.c unsigned long nr_pages_needed = ALIGN(1 << order, PAGES_PER_SECTION); PAGES_PER_SECTION 1900 mm/page_alloc.c if ((first_deferred_pfn ^ spfn) < PAGES_PER_SECTION) PAGES_PER_SECTION 6973 mm/page_alloc.c round_up(max_pfn, PAGES_PER_SECTION)); PAGES_PER_SECTION 232 mm/page_ext.c table_size = page_ext_size * PAGES_PER_SECTION; PAGES_PER_SECTION 265 mm/page_ext.c table_size = page_ext_size * PAGES_PER_SECTION; PAGES_PER_SECTION 306 mm/page_ext.c for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 315 mm/page_ext.c for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) PAGES_PER_SECTION 329 mm/page_ext.c for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) PAGES_PER_SECTION 385 mm/page_ext.c pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) { PAGES_PER_SECTION 243 mm/sparse.c pfns = min(nr_pages, PAGES_PER_SECTION PAGES_PER_SECTION 277 mm/sparse.c for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 442 mm/sparse.c return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); PAGES_PER_SECTION 448 mm/sparse.c return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); PAGES_PER_SECTION 548 mm/sparse.c map = __populate_section_memmap(pfn, PAGES_PER_SECTION, PAGES_PER_SECTION 613 mm/sparse.c for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 632 mm/sparse.c for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { PAGES_PER_SECTION 667 mm/sparse.c unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); PAGES_PER_SECTION 676 mm/sparse.c unsigned long memmap_size = sizeof(struct page) * PAGES_PER_SECTION; PAGES_PER_SECTION 703 mm/sparse.c get_order(sizeof(struct page) * PAGES_PER_SECTION)); PAGES_PER_SECTION 712 mm/sparse.c nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) PAGES_PER_SECTION 851 mm/sparse.c if (nr_pages < PAGES_PER_SECTION && early_section(ms))