/linux-4.4.14/include/linux/ |
D | mmzone.h | 24 #define MAX_ORDER 11 macro 26 #define MAX_ORDER CONFIG_FORCE_MAX_ZONEORDER macro 28 #define MAX_ORDER_NR_PAGES (1 << (MAX_ORDER - 1)) 73 for (order = 0; order < MAX_ORDER; order++) \ 479 struct free_area free_area[MAX_ORDER]; 1015 #if (MAX_ORDER - 1 + PAGE_SHIFT) > SECTION_SIZE_BITS 1016 #error Allocator MAX_ORDER exceeds SECTION_SIZE
|
D | slab.h | 192 #define KMALLOC_SHIFT_HIGH ((MAX_ORDER + PAGE_SHIFT - 1) <= 25 ? \ 193 (MAX_ORDER + PAGE_SHIFT - 1) : 25) 206 #define KMALLOC_SHIFT_MAX (MAX_ORDER + PAGE_SHIFT)
|
D | pageblock-flags.h | 59 #define pageblock_order (MAX_ORDER-1)
|
D | huge_mm.h | 122 #if HPAGE_PMD_ORDER >= MAX_ORDER
|
D | hugetlb.h | 407 return huge_page_order(h) >= MAX_ORDER; in hstate_is_gigantic()
|
/linux-4.4.14/arch/metag/kernel/ |
D | setup.c | 368 max_pfn = max_low_pfn + ((1 << MAX_ORDER) - 1); in setup_arch() 369 max_pfn &= ~((1 << MAX_ORDER) - 1); in setup_arch() 373 if (min_low_pfn & ((1 << MAX_ORDER) - 1)) { in setup_arch() 383 1 << (MAX_ORDER + PAGE_SHIFT), in setup_arch()
|
/linux-4.4.14/mm/ |
D | page_alloc.c | 523 if (kstrtoul(buf, 10, &res) < 0 || res > MAX_ORDER / 2) { in debug_guardpage_minorder_setup() 667 max_order = min_t(unsigned int, MAX_ORDER, pageblock_order + 1); in __free_one_page() 676 page_idx = pfn & ((1 << MAX_ORDER) - 1); in __free_one_page() 703 if (max_order < MAX_ORDER) { in __free_one_page() 739 if ((order < MAX_ORDER-2) && pfn_valid_within(page_to_pfn(buddy))) { in __free_one_page() 1120 __free_pages_boot_core(page, pfn, MAX_ORDER-1); in deferred_free_range() 1290 if (pageblock_order >= MAX_ORDER) { in init_cma_reserved_pageblock() 1295 __free_pages(p, MAX_ORDER - 1); in init_cma_reserved_pageblock() 1438 for (current_order = order; current_order < MAX_ORDER; ++current_order) { in __rmqueue_smallest() 1720 for (order = 0; order < MAX_ORDER; order++) { in unreserve_highatomic_pageblock() [all …]
|
D | cma.c | 186 alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); in cma_init_reserved_mem() 270 (phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order)); in cma_declare_contiguous()
|
D | vmstat.c | 644 for (order = 0; order < MAX_ORDER; order++) { in fill_contig_page_info() 941 for (order = 0; order < MAX_ORDER; ++order) in frag_show_print() 966 for (order = 0; order < MAX_ORDER; ++order) { in pagetypeinfo_showfree_print() 989 for (order = 0; order < MAX_ORDER; ++order) in pagetypeinfo_showfree() 1599 for (order = 0; order < MAX_ORDER; ++order) { in unusable_show_print() 1661 for (order = 0; order < MAX_ORDER; ++order) { in extfrag_show_print()
|
D | compaction.c | 439 if (likely(comp_order < MAX_ORDER)) { in isolate_freepages_block() 732 if (freepage_order > 0 && freepage_order < MAX_ORDER) in isolate_migratepages_block() 762 if (likely(comp_order < MAX_ORDER)) in isolate_migratepages_block() 1227 for (order = cc->order; order < MAX_ORDER; order++) { in __compact_finished()
|
D | page_isolation.c | 101 page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); in unset_migratetype_isolate()
|
D | page_owner.c | 191 if (freepage_order < MAX_ORDER) in read_page_owner()
|
D | nobootmem.c | 99 order = min(MAX_ORDER - 1UL, __ffs(start)); in __free_pages_memory()
|
D | hugetlb.c | 1328 if (compound_order(page_head) >= MAX_ORDER) in __basepage_index() 1987 if (unlikely(order > (MAX_ORDER - 1))) in prep_compound_huge_page() 2751 if (hugetlb_max_hstate && parsed_hstate->order >= MAX_ORDER) in hugetlb_nrpages_setup()
|
D | memory_hotplug.c | 1349 if ((order < MAX_ORDER) && (order >= pageblock_order)) in next_active_pageblock()
|
D | slub.c | 3088 order = slab_order(size, 1, MAX_ORDER, 1, reserved); in calculate_order() 3089 if (order < MAX_ORDER) in calculate_order() 3503 slub_max_order = min(slub_max_order, MAX_ORDER - 1); in setup_slub_max_order()
|
D | slab.c | 595 min(slab_max_order, MAX_ORDER - 1); in slab_max_order_setup()
|
/linux-4.4.14/arch/m32r/mm/ |
D | discontig.c | 41 const unsigned long zone_alignment = 1UL << (MAX_ORDER - 1); in mem_prof_init()
|
/linux-4.4.14/arch/ia64/mm/ |
D | hugetlbpage.c | 181 size >= (1UL << PAGE_SHIFT << MAX_ORDER)) { in hugetlb_setup_sz()
|
/linux-4.4.14/arch/sparc/mm/ |
D | tsb.c | 342 if (max_tsb_size > (PAGE_SIZE << MAX_ORDER)) in tsb_grow() 343 max_tsb_size = (PAGE_SIZE << MAX_ORDER); in tsb_grow()
|
/linux-4.4.14/drivers/base/ |
D | dma-contiguous.c | 245 phys_addr_t align = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); in rmem_cma_setup()
|
/linux-4.4.14/fs/ramfs/ |
D | file-nommu.c | 77 if (unlikely(order >= MAX_ORDER)) in ramfs_nommu_expand_for_mapping()
|
/linux-4.4.14/arch/um/kernel/ |
D | um_arch.c | 310 max_physmem &= ~((1 << (PAGE_SHIFT + MAX_ORDER)) - 1); in linux_main()
|
/linux-4.4.14/kernel/events/ |
D | ring_buffer.c | 439 if (order > MAX_ORDER) in rb_alloc_aux_page() 440 order = MAX_ORDER; in rb_alloc_aux_page()
|
/linux-4.4.14/arch/tile/mm/ |
D | init.c | 669 if (order >= MAX_ORDER) in init_free_pfn_range() 670 order = MAX_ORDER-1; in init_free_pfn_range()
|
D | pgtable.c | 71 for (order = 0; order < MAX_ORDER; order++) { in show_mem()
|
/linux-4.4.14/drivers/misc/mic/scif/ |
D | scif_rma.h | 404 if (align && get_order(align) < MAX_ORDER) in scif_zalloc()
|
D | scif_api.c | 1111 int chunk_len = min(len, (1 << (MAX_ORDER + PAGE_SHIFT - 1))); in scif_user_send() 1172 int chunk_len = min(len, (1 << (MAX_ORDER + PAGE_SHIFT - 1))); in scif_user_recv()
|
/linux-4.4.14/drivers/of/ |
D | of_reserved_mem.c | 131 align = max(align, (phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order)); in __reserved_mem_alloc_size()
|
/linux-4.4.14/arch/alpha/mm/ |
D | numa.c | 161 node_min_pfn &= ~((1UL << (MAX_ORDER-1))-1); in setup_memory_node()
|
/linux-4.4.14/Documentation/networking/ |
D | packet_mmap.txt | 272 region allocated by __get_free_pages is determined by the MAX_ORDER macro. More 275 PAGE_SIZE << MAX_ORDER 278 In a 2.4/i386 kernel MAX_ORDER is 10 279 In a 2.6/i386 kernel MAX_ORDER is 11 285 /usr/include/linux/mmzone.h to get PAGE_SIZE MAX_ORDER declarations. 331 <max-order> : is the value defined with MAX_ORDER
|
/linux-4.4.14/drivers/misc/genwqe/ |
D | card_utils.c | 220 if (get_order(size) > MAX_ORDER) in __genwqe_alloc_consistent() 317 if (get_order(sgl->sgl_size) > MAX_ORDER) { in genwqe_alloc_sync_sgl()
|
D | card_dev.c | 449 if (get_order(vsize) > MAX_ORDER) in genwqe_mmap()
|
/linux-4.4.14/drivers/iommu/ |
D | dma-iommu.c | 197 unsigned int order = MAX_ORDER; in __iommu_dma_alloc_pages()
|
/linux-4.4.14/security/integrity/ima/ |
D | ima_crypto.c | 51 if (order >= MAX_ORDER) in param_set_bufsize()
|
/linux-4.4.14/drivers/irqchip/ |
D | irq-gic-v3-its.c | 865 if (order >= MAX_ORDER) { in its_alloc_tables() 866 order = MAX_ORDER - 1; in its_alloc_tables()
|
/linux-4.4.14/arch/s390/kernel/ |
D | setup.c | 867 memblock_trim_memory(1UL << (MAX_ORDER - 1 + PAGE_SHIFT)); in setup_arch()
|
/linux-4.4.14/arch/powerpc/mm/ |
D | hugetlbpage.c | 418 else if (mmu_psize_to_shift(i) < (MAX_ORDER + PAGE_SHIFT)) in reserve_hugetlb_gpages()
|
/linux-4.4.14/arch/sparc/kernel/ |
D | pci_sun4v.c | 144 if (unlikely(order >= MAX_ORDER)) in dma_4v_alloc_coherent()
|
D | traps_64.c | 871 for (order = 0; order < MAX_ORDER; order++) { in cheetah_ecache_flush_init()
|
/linux-4.4.14/kernel/ |
D | kexec_core.c | 1407 VMCOREINFO_LENGTH(zone.free_area, MAX_ORDER); in crash_save_vmcoreinfo_init()
|
/linux-4.4.14/drivers/video/fbdev/vermilion/ |
D | vermilion.c | 213 while (requested > (PAGE_SIZE << order) && order < MAX_ORDER) in vmlfb_alloc_vram()
|
/linux-4.4.14/arch/ia64/ |
D | Kconfig | 304 int "MAX_ORDER (11 - 17)" if !HUGETLB_PAGE
|
/linux-4.4.14/drivers/md/ |
D | dm-bufio.c | 62 #define DM_BUFIO_BLOCK_SIZE_GFP_LIMIT (PAGE_SIZE << (MAX_ORDER - 1))
|
/linux-4.4.14/arch/arm64/ |
D | Kconfig | 588 MAX_ORDER = (PMD_SHIFT - PAGE_SHIFT) + 1 => PAGE_SHIFT - 2
|
/linux-4.4.14/tools/perf/Documentation/ |
D | intel-pt.txt | 459 buffer sizes are limited to powers of 2 up to 4MiB (MAX_ORDER). In order to
|
/linux-4.4.14/Documentation/ |
D | kernel-parameters.txt | 844 possible value is MAX_ORDER/2. Setting this parameter
|