PAGES_PER_SECTION  374 arch/arm/mm/init.c 				 ALIGN(prev_end, PAGES_PER_SECTION));
PAGES_PER_SECTION  400 arch/arm/mm/init.c 	if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION))
PAGES_PER_SECTION  402 arch/arm/mm/init.c 			    ALIGN(prev_end, PAGES_PER_SECTION));
PAGES_PER_SECTION  504 arch/arm64/mm/init.c 		start = min(start, ALIGN(prev_end, PAGES_PER_SECTION));
PAGES_PER_SECTION  523 arch/arm64/mm/init.c 	if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION))
PAGES_PER_SECTION  524 arch/arm64/mm/init.c 		free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION));
PAGES_PER_SECTION   96 arch/powerpc/mm/init_64.c 	for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page)))
PAGES_PER_SECTION  368 arch/powerpc/platforms/pseries/hotplug-memory.c 		rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION);
PAGES_PER_SECTION   37 arch/x86/include/asm/mmzone_32.h 	return((int) physnode_map[(pfn) / PAGES_PER_SECTION]);
PAGES_PER_SECTION 1519 arch/x86/mm/init_64.c 	if (end - start < PAGES_PER_SECTION * sizeof(struct page))
PAGES_PER_SECTION  559 arch/x86/mm/numa.c 	if (pfn_align && pfn_align < PAGES_PER_SECTION) {
PAGES_PER_SECTION  562 arch/x86/mm/numa.c 		       PFN_PHYS(PAGES_PER_SECTION) >> 20);
PAGES_PER_SECTION   54 arch/x86/mm/numa_32.c 	start = round_down(start, PAGES_PER_SECTION);
PAGES_PER_SECTION   55 arch/x86/mm/numa_32.c 	end = round_up(end, PAGES_PER_SECTION);
PAGES_PER_SECTION   56 arch/x86/mm/numa_32.c 	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION   57 arch/x86/mm/numa_32.c 		physnode_map[pfn / PAGES_PER_SECTION] = nid;
PAGES_PER_SECTION  191 drivers/base/memory.c 				section_nr, pfn, pfn + PAGES_PER_SECTION);
PAGES_PER_SECTION  195 drivers/base/memory.c 				section_nr, pfn, pfn + PAGES_PER_SECTION);
PAGES_PER_SECTION  199 drivers/base/memory.c 				section_nr, pfn, pfn + PAGES_PER_SECTION);
PAGES_PER_SECTION  202 drivers/base/memory.c 		pfn += PAGES_PER_SECTION;
PAGES_PER_SECTION  217 drivers/base/memory.c 	unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
PAGES_PER_SECTION  383 drivers/base/memory.c 	unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
PAGES_PER_SECTION  478 drivers/base/memory.c 	unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block;
PAGES_PER_SECTION  779 drivers/base/node.c 			pfn = round_down(pfn + PAGES_PER_SECTION,
PAGES_PER_SECTION  780 drivers/base/node.c 					 PAGES_PER_SECTION) - 1;
PAGES_PER_SECTION  263 drivers/xen/balloon.c 				PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL);
PAGES_PER_SECTION  304 drivers/xen/balloon.c 	balloon_hotplug = round_up(credit, PAGES_PER_SECTION);
PAGES_PER_SECTION  123 include/linux/memory.h #define CONFIG_MEM_BLOCK_SIZE	(PAGES_PER_SECTION<<PAGE_SHIFT)
PAGES_PER_SECTION 1138 include/linux/mmzone.h #define PAGE_SECTION_MASK	(~(PAGES_PER_SECTION-1))
PAGES_PER_SECTION 1156 include/linux/mmzone.h #define SECTION_ALIGN_UP(pfn)	(((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
PAGES_PER_SECTION  184 mm/memory_hotplug.c 	mapsize = sizeof(struct page) * PAGES_PER_SECTION;
PAGES_PER_SECTION  213 mm/memory_hotplug.c 	register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION);
PAGES_PER_SECTION  241 mm/memory_hotplug.c 	for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  271 mm/memory_hotplug.c 		min_align = PAGES_PER_SECTION;
PAGES_PER_SECTION  315 mm/memory_hotplug.c 		pfns = min(nr_pages, PAGES_PER_SECTION
PAGES_PER_SECTION  535 mm/memory_hotplug.c 		pfns = min(nr_pages, PAGES_PER_SECTION
PAGES_PER_SECTION 1225 mm/memory_hotplug.c 	     pfn = sec_end_pfn, sec_end_pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  427 mm/page_alloc.c 	if ((nr_initialised > PAGES_PER_SECTION) &&
PAGES_PER_SECTION  428 mm/page_alloc.c 	    (pfn & (PAGES_PER_SECTION - 1)) == 0) {
PAGES_PER_SECTION  462 mm/page_alloc.c 	pfn &= (PAGES_PER_SECTION-1);
PAGES_PER_SECTION 1846 mm/page_alloc.c 	unsigned long nr_pages_needed = ALIGN(1 << order, PAGES_PER_SECTION);
PAGES_PER_SECTION 1900 mm/page_alloc.c 		if ((first_deferred_pfn ^ spfn) < PAGES_PER_SECTION)
PAGES_PER_SECTION 6973 mm/page_alloc.c 				round_up(max_pfn, PAGES_PER_SECTION));
PAGES_PER_SECTION  232 mm/page_ext.c  	table_size = page_ext_size * PAGES_PER_SECTION;
PAGES_PER_SECTION  265 mm/page_ext.c  		table_size = page_ext_size * PAGES_PER_SECTION;
PAGES_PER_SECTION  306 mm/page_ext.c  	for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  315 mm/page_ext.c  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
PAGES_PER_SECTION  329 mm/page_ext.c  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
PAGES_PER_SECTION  385 mm/page_ext.c  			pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) {
PAGES_PER_SECTION  243 mm/sparse.c    		pfns = min(nr_pages, PAGES_PER_SECTION
PAGES_PER_SECTION  277 mm/sparse.c    	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  442 mm/sparse.c    	return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE);
PAGES_PER_SECTION  448 mm/sparse.c    	return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION);
PAGES_PER_SECTION  548 mm/sparse.c    		map = __populate_section_memmap(pfn, PAGES_PER_SECTION,
PAGES_PER_SECTION  613 mm/sparse.c    	for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  632 mm/sparse.c    	for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
PAGES_PER_SECTION  667 mm/sparse.c    	unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION);
PAGES_PER_SECTION  676 mm/sparse.c    	unsigned long memmap_size = sizeof(struct page) * PAGES_PER_SECTION;
PAGES_PER_SECTION  703 mm/sparse.c    			   get_order(sizeof(struct page) * PAGES_PER_SECTION));
PAGES_PER_SECTION  712 mm/sparse.c    	nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page))
PAGES_PER_SECTION  851 mm/sparse.c    	if (nr_pages < PAGES_PER_SECTION && early_section(ms))