/linux-4.1.27/include/linux/ |
H A D | hugetlb_cgroup.h | 33 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) hugetlb_cgroup_from_page() 43 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) set_hugetlb_cgroup()
|
H A D | hugetlb.h | 419 return size_to_hstate(PAGE_SIZE << compound_order(page)); page_hstate()
|
H A D | pagemap.h | 391 return page->index << compound_order(page); page_to_pgoff()
|
H A D | mm_types.h | 141 unsigned long compound_order; member in struct:page::__anon12169::__anon12171
|
H A D | mm.h | 591 static inline unsigned int compound_order(struct page *page) compound_order() function 595 return page[1].compound_order; compound_order() 600 page[1].compound_order = order; set_compound_order()
|
/linux-4.1.27/drivers/staging/android/ion/ |
H A D | ion_system_heap.c | 83 unsigned int order = compound_order(page); free_buffer_page() 149 size_remaining -= PAGE_SIZE << compound_order(page); ion_system_heap_allocate() 150 max_order = compound_order(page); ion_system_heap_allocate() 162 sg_set_page(sg, page, PAGE_SIZE << compound_order(page), 0); ion_system_heap_allocate()
|
H A D | ion_page_pool.c | 99 BUG_ON(pool->order != compound_order(page)); ion_page_pool_free()
|
/linux-4.1.27/mm/kasan/ |
H A D | kasan.c | 306 PAGE_SIZE << compound_order(page), kasan_poison_slab() 370 redzone_end = (unsigned long)ptr + (PAGE_SIZE << compound_order(page)); kasan_kmalloc_large() 399 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), kasan_kfree() 409 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), kasan_kfree_large()
|
/linux-4.1.27/arch/arm/mm/ |
H A D | flush.c | 195 size_t page_size = PAGE_SIZE << compound_order(page); __flush_dcache_page() 200 for (i = 0; i < (1 << compound_order(page)); i++) { __flush_dcache_page() 206 for (i = 0; i < (1 << compound_order(page)); i++) { __flush_dcache_page()
|
/linux-4.1.27/arch/arm64/mm/ |
H A D | flush.c | 83 PAGE_SIZE << compound_order(page)); __sync_icache_dcache()
|
/linux-4.1.27/mm/ |
H A D | memory-failure.c | 205 si.si_addr_lsb = compound_order(compound_head(page)) + PAGE_SHIFT; kill_proc() 1068 int nr_pages = 1 << compound_order(hpage); set_page_hwpoison_huge_page() 1076 int nr_pages = 1 << compound_order(hpage); clear_page_hwpoison_huge_page() 1127 * so nr_pages should be 1 << compound_order. OTOH when errors are on memory_failure() 1133 nr_pages = 1 << compound_order(hpage); memory_failure() 1456 nr_pages = 1 << compound_order(page); unpoison_memory() 1620 atomic_long_add(1 << compound_order(hpage), soft_offline_huge_page() 1761 atomic_long_add(1 << compound_order(hpage), soft_offline_page()
|
H A D | slob.c | 500 __free_pages(sp, compound_order(sp)); kfree() 517 return PAGE_SIZE << compound_order(sp); ksize()
|
H A D | hugetlb_cgroup.c | 126 nr_pages = 1 << compound_order(page); hugetlb_cgroup_move_parent()
|
H A D | madvise.c | 343 compound_order(compound_head(p))) { madvise_hwpoison()
|
H A D | memory_hotplug.c | 1380 1 << compound_order(page)) - 1; scan_movable_pages() 1405 pfn = page_to_pfn(head) + (1<<compound_order(head)) - 1; do_migrate_range() 1406 if (compound_order(head) > PFN_SECTION_SHIFT) { do_migrate_range() 1411 move_pages -= 1 << compound_order(head); do_migrate_range()
|
H A D | slub.c | 790 length = (PAGE_SIZE << compound_order(page)) - s->reserved; slab_pad_check() 874 maxobj = order_objects(compound_order(page), s->size, s->reserved); check_slab() 924 max_objects = order_objects(compound_order(page), s->size, s->reserved); on_freelist() 1426 order = compound_order(page); new_slab() 1457 int order = compound_order(page); __free_slab() 1469 kmemcheck_free_shadow(page, compound_order(page)); __free_slab() 1507 int order = compound_order(page); free_slab() 3385 return PAGE_SIZE << compound_order(page); __ksize() 3415 __free_kmem_pages(page, compound_order(page)); kfree()
|
H A D | compaction.c | 737 * compound_order() without preventing THP from splitting the isolate_migratepages_block() 745 low_pfn += (1 << compound_order(page)) - 1; isolate_migratepages_block() 770 low_pfn += (1 << compound_order(page)) - 1; isolate_migratepages_block()
|
H A D | migrate.c | 1615 VM_BUG_ON_PAGE(compound_order(page) && !PageTransHuge(page), page); numamigrate_isolate_page() 1618 if (!migrate_balanced_pgdat(pgdat, 1UL << compound_order(page))) numamigrate_isolate_page()
|
H A D | huge_memory.c | 196 __free_pages(zero_page, compound_order(zero_page)); get_huge_zero_page() 228 __free_pages(zero_page, compound_order(zero_page)); shrink_huge_zero_page_scan()
|
H A D | memcontrol.c | 5491 nr_pages <<= compound_order(page); mem_cgroup_try_charge() 5550 nr_pages <<= compound_order(page); mem_cgroup_commit_charge() 5592 nr_pages <<= compound_order(page); mem_cgroup_cancel_charge() 5665 nr_pages <<= compound_order(page); uncharge_list()
|
H A D | page_alloc.c | 362 __free_pages_ok(page, compound_order(page)); free_compound_page() 807 VM_BUG_ON_PAGE(compound && compound_order(page) != order, page); free_pages_prepare() 6288 iter = round_up(iter + 1, 1<<compound_order(page)) - 1; has_unmovable_pages()
|
H A D | nommu.c | 148 return PAGE_SIZE << compound_order(page); kobjsize()
|
H A D | hugetlb.c | 1084 if (compound_order(page_head) >= MAX_ORDER) __basepage_index() 1089 return (index << compound_order(page_head)) + compound_idx; __basepage_index()
|
/linux-4.1.27/drivers/target/tcm_fc/ |
H A D | tfc_io.c | 158 PAGE_SIZE << compound_order(page); ft_queue_data_in()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 698 page_order = compound_order(pfn_to_page(pfn)); ehea_is_hugepage()
|
/linux-4.1.27/arch/ia64/mm/ |
H A D | init.c | 65 flush_icache_range(addr, addr + (PAGE_SIZE << compound_order(page))); __ia64_sync_icache_dcache()
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | hugetlbpage.c | 945 for (i = 0; i < (1UL << compound_order(page)); i++) { flush_dcache_icache_hugepage()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | book3s_64_mmu_hv.c | 534 pte_size <<= compound_order(page); kvmppc_book3s_hv_page_fault()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
H A D | ehca_mrmw.c | 2358 page_order = compound_order(pfn_to_page(pfn)); ehca_is_hugepage()
|
/linux-4.1.27/drivers/net/xen-netback/ |
H A D | netback.c | 281 BUG_ON(size + offset > PAGE_SIZE<<compound_order(page)); xenvif_gop_frag_copy()
|