/linux-4.1.27/include/linux/ |
D | hugetlb_cgroup.h | 33 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in hugetlb_cgroup_from_page() 43 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in set_hugetlb_cgroup()
|
D | mm_types.h | 141 unsigned long compound_order; member
|
D | hugetlb.h | 419 return size_to_hstate(PAGE_SIZE << compound_order(page)); in page_hstate()
|
D | pagemap.h | 391 return page->index << compound_order(page); in page_to_pgoff()
|
D | mm.h | 591 static inline unsigned int compound_order(struct page *page) in compound_order() function 595 return page[1].compound_order; in compound_order() 600 page[1].compound_order = order; in set_compound_order()
|
/linux-4.1.27/arch/arm/mm/ |
D | flush.c | 195 size_t page_size = PAGE_SIZE << compound_order(page); in __flush_dcache_page() 200 for (i = 0; i < (1 << compound_order(page)); i++) { in __flush_dcache_page() 206 for (i = 0; i < (1 << compound_order(page)); i++) { in __flush_dcache_page()
|
/linux-4.1.27/mm/kasan/ |
D | kasan.c | 306 PAGE_SIZE << compound_order(page), in kasan_poison_slab() 370 redzone_end = (unsigned long)ptr + (PAGE_SIZE << compound_order(page)); in kasan_kmalloc_large() 399 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), in kasan_kfree() 409 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), in kasan_kfree_large()
|
/linux-4.1.27/drivers/staging/android/ion/ |
D | ion_system_heap.c | 83 unsigned int order = compound_order(page); in free_buffer_page() 149 size_remaining -= PAGE_SIZE << compound_order(page); in ion_system_heap_allocate() 150 max_order = compound_order(page); in ion_system_heap_allocate() 162 sg_set_page(sg, page, PAGE_SIZE << compound_order(page), 0); in ion_system_heap_allocate()
|
D | ion_page_pool.c | 99 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free()
|
/linux-4.1.27/mm/ |
D | memory-failure.c | 205 si.si_addr_lsb = compound_order(compound_head(page)) + PAGE_SHIFT; in kill_proc() 1068 int nr_pages = 1 << compound_order(hpage); in set_page_hwpoison_huge_page() 1076 int nr_pages = 1 << compound_order(hpage); in clear_page_hwpoison_huge_page() 1133 nr_pages = 1 << compound_order(hpage); in memory_failure() 1456 nr_pages = 1 << compound_order(page); in unpoison_memory() 1620 atomic_long_add(1 << compound_order(hpage), in soft_offline_huge_page() 1761 atomic_long_add(1 << compound_order(hpage), in soft_offline_page()
|
D | slob.c | 500 __free_pages(sp, compound_order(sp)); in kfree() 517 return PAGE_SIZE << compound_order(sp); in ksize()
|
D | memory_hotplug.c | 1380 1 << compound_order(page)) - 1; in scan_movable_pages() 1405 pfn = page_to_pfn(head) + (1<<compound_order(head)) - 1; in do_migrate_range() 1406 if (compound_order(head) > PFN_SECTION_SHIFT) { in do_migrate_range() 1411 move_pages -= 1 << compound_order(head); in do_migrate_range()
|
D | hugetlb_cgroup.c | 126 nr_pages = 1 << compound_order(page); in hugetlb_cgroup_move_parent()
|
D | slub.c | 790 length = (PAGE_SIZE << compound_order(page)) - s->reserved; in slab_pad_check() 874 maxobj = order_objects(compound_order(page), s->size, s->reserved); in check_slab() 924 max_objects = order_objects(compound_order(page), s->size, s->reserved); in on_freelist() 1426 order = compound_order(page); in new_slab() 1457 int order = compound_order(page); in __free_slab() 1469 kmemcheck_free_shadow(page, compound_order(page)); in __free_slab() 1507 int order = compound_order(page); in free_slab() 3385 return PAGE_SIZE << compound_order(page); in __ksize() 3415 __free_kmem_pages(page, compound_order(page)); in kfree()
|
D | madvise.c | 343 compound_order(compound_head(p))) { in madvise_hwpoison()
|
D | compaction.c | 745 low_pfn += (1 << compound_order(page)) - 1; in isolate_migratepages_block() 770 low_pfn += (1 << compound_order(page)) - 1; in isolate_migratepages_block()
|
D | migrate.c | 1615 VM_BUG_ON_PAGE(compound_order(page) && !PageTransHuge(page), page); in numamigrate_isolate_page() 1618 if (!migrate_balanced_pgdat(pgdat, 1UL << compound_order(page))) in numamigrate_isolate_page()
|
D | memcontrol.c | 5491 nr_pages <<= compound_order(page); in mem_cgroup_try_charge() 5550 nr_pages <<= compound_order(page); in mem_cgroup_commit_charge() 5592 nr_pages <<= compound_order(page); in mem_cgroup_cancel_charge() 5665 nr_pages <<= compound_order(page); in uncharge_list()
|
D | huge_memory.c | 196 __free_pages(zero_page, compound_order(zero_page)); in get_huge_zero_page() 228 __free_pages(zero_page, compound_order(zero_page)); in shrink_huge_zero_page_scan()
|
D | page_alloc.c | 362 __free_pages_ok(page, compound_order(page)); in free_compound_page() 807 VM_BUG_ON_PAGE(compound && compound_order(page) != order, page); in free_pages_prepare() 6288 iter = round_up(iter + 1, 1<<compound_order(page)) - 1; in has_unmovable_pages()
|
D | hugetlb.c | 1084 if (compound_order(page_head) >= MAX_ORDER) in __basepage_index() 1089 return (index << compound_order(page_head)) + compound_idx; in __basepage_index()
|
D | nommu.c | 148 return PAGE_SIZE << compound_order(page); in kobjsize()
|
/linux-4.1.27/arch/arm64/mm/ |
D | flush.c | 83 PAGE_SIZE << compound_order(page)); in __sync_icache_dcache()
|
/linux-4.1.27/drivers/target/tcm_fc/ |
D | tfc_io.c | 158 PAGE_SIZE << compound_order(page); in ft_queue_data_in()
|
/linux-4.1.27/arch/ia64/mm/ |
D | init.c | 65 flush_icache_range(addr, addr + (PAGE_SIZE << compound_order(page))); in __ia64_sync_icache_dcache()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | hugetlbpage.c | 945 for (i = 0; i < (1UL << compound_order(page)); i++) { in flush_dcache_icache_hugepage()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 698 page_order = compound_order(pfn_to_page(pfn)); in ehea_is_hugepage()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | book3s_64_mmu_hv.c | 534 pte_size <<= compound_order(page); in kvmppc_book3s_hv_page_fault()
|
/linux-4.1.27/drivers/net/xen-netback/ |
D | netback.c | 281 BUG_ON(size + offset > PAGE_SIZE<<compound_order(page)); in xenvif_gop_frag_copy()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_mrmw.c | 2358 page_order = compound_order(pfn_to_page(pfn)); in ehca_is_hugepage()
|