Searched refs:compound_order (Results 1 - 30 of 30) sorted by relevance

/linux-4.1.27/include/linux/
H A Dhugetlb_cgroup.h33 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) hugetlb_cgroup_from_page()
43 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) set_hugetlb_cgroup()
H A Dhugetlb.h419 return size_to_hstate(PAGE_SIZE << compound_order(page)); page_hstate()
H A Dpagemap.h391 return page->index << compound_order(page); page_to_pgoff()
H A Dmm_types.h141 unsigned long compound_order; member in struct:page::__anon12169::__anon12171
H A Dmm.h591 static inline unsigned int compound_order(struct page *page) compound_order() function
595 return page[1].compound_order; compound_order()
600 page[1].compound_order = order; set_compound_order()
/linux-4.1.27/drivers/staging/android/ion/
H A Dion_system_heap.c83 unsigned int order = compound_order(page); free_buffer_page()
149 size_remaining -= PAGE_SIZE << compound_order(page); ion_system_heap_allocate()
150 max_order = compound_order(page); ion_system_heap_allocate()
162 sg_set_page(sg, page, PAGE_SIZE << compound_order(page), 0); ion_system_heap_allocate()
H A Dion_page_pool.c99 BUG_ON(pool->order != compound_order(page)); ion_page_pool_free()
/linux-4.1.27/mm/kasan/
H A Dkasan.c306 PAGE_SIZE << compound_order(page), kasan_poison_slab()
370 redzone_end = (unsigned long)ptr + (PAGE_SIZE << compound_order(page)); kasan_kmalloc_large()
399 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), kasan_kfree()
409 kasan_poison_shadow(ptr, PAGE_SIZE << compound_order(page), kasan_kfree_large()
/linux-4.1.27/arch/arm/mm/
H A Dflush.c195 size_t page_size = PAGE_SIZE << compound_order(page); __flush_dcache_page()
200 for (i = 0; i < (1 << compound_order(page)); i++) { __flush_dcache_page()
206 for (i = 0; i < (1 << compound_order(page)); i++) { __flush_dcache_page()
/linux-4.1.27/arch/arm64/mm/
H A Dflush.c83 PAGE_SIZE << compound_order(page)); __sync_icache_dcache()
/linux-4.1.27/mm/
H A Dmemory-failure.c205 si.si_addr_lsb = compound_order(compound_head(page)) + PAGE_SHIFT; kill_proc()
1068 int nr_pages = 1 << compound_order(hpage); set_page_hwpoison_huge_page()
1076 int nr_pages = 1 << compound_order(hpage); clear_page_hwpoison_huge_page()
1127 * so nr_pages should be 1 << compound_order. OTOH when errors are on memory_failure()
1133 nr_pages = 1 << compound_order(hpage); memory_failure()
1456 nr_pages = 1 << compound_order(page); unpoison_memory()
1620 atomic_long_add(1 << compound_order(hpage), soft_offline_huge_page()
1761 atomic_long_add(1 << compound_order(hpage), soft_offline_page()
H A Dslob.c500 __free_pages(sp, compound_order(sp)); kfree()
517 return PAGE_SIZE << compound_order(sp); ksize()
H A Dhugetlb_cgroup.c126 nr_pages = 1 << compound_order(page); hugetlb_cgroup_move_parent()
H A Dmadvise.c343 compound_order(compound_head(p))) { madvise_hwpoison()
H A Dmemory_hotplug.c1380 1 << compound_order(page)) - 1; scan_movable_pages()
1405 pfn = page_to_pfn(head) + (1<<compound_order(head)) - 1; do_migrate_range()
1406 if (compound_order(head) > PFN_SECTION_SHIFT) { do_migrate_range()
1411 move_pages -= 1 << compound_order(head); do_migrate_range()
H A Dslub.c790 length = (PAGE_SIZE << compound_order(page)) - s->reserved; slab_pad_check()
874 maxobj = order_objects(compound_order(page), s->size, s->reserved); check_slab()
924 max_objects = order_objects(compound_order(page), s->size, s->reserved); on_freelist()
1426 order = compound_order(page); new_slab()
1457 int order = compound_order(page); __free_slab()
1469 kmemcheck_free_shadow(page, compound_order(page)); __free_slab()
1507 int order = compound_order(page); free_slab()
3385 return PAGE_SIZE << compound_order(page); __ksize()
3415 __free_kmem_pages(page, compound_order(page)); kfree()
H A Dcompaction.c737 * compound_order() without preventing THP from splitting the isolate_migratepages_block()
745 low_pfn += (1 << compound_order(page)) - 1; isolate_migratepages_block()
770 low_pfn += (1 << compound_order(page)) - 1; isolate_migratepages_block()
H A Dmigrate.c1615 VM_BUG_ON_PAGE(compound_order(page) && !PageTransHuge(page), page); numamigrate_isolate_page()
1618 if (!migrate_balanced_pgdat(pgdat, 1UL << compound_order(page))) numamigrate_isolate_page()
H A Dhuge_memory.c196 __free_pages(zero_page, compound_order(zero_page)); get_huge_zero_page()
228 __free_pages(zero_page, compound_order(zero_page)); shrink_huge_zero_page_scan()
H A Dmemcontrol.c5491 nr_pages <<= compound_order(page); mem_cgroup_try_charge()
5550 nr_pages <<= compound_order(page); mem_cgroup_commit_charge()
5592 nr_pages <<= compound_order(page); mem_cgroup_cancel_charge()
5665 nr_pages <<= compound_order(page); uncharge_list()
H A Dpage_alloc.c362 __free_pages_ok(page, compound_order(page)); free_compound_page()
807 VM_BUG_ON_PAGE(compound && compound_order(page) != order, page); free_pages_prepare()
6288 iter = round_up(iter + 1, 1<<compound_order(page)) - 1; has_unmovable_pages()
H A Dnommu.c148 return PAGE_SIZE << compound_order(page); kobjsize()
H A Dhugetlb.c1084 if (compound_order(page_head) >= MAX_ORDER) __basepage_index()
1089 return (index << compound_order(page_head)) + compound_idx; __basepage_index()
/linux-4.1.27/drivers/target/tcm_fc/
H A Dtfc_io.c158 PAGE_SIZE << compound_order(page); ft_queue_data_in()
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.c698 page_order = compound_order(pfn_to_page(pfn)); ehea_is_hugepage()
/linux-4.1.27/arch/ia64/mm/
H A Dinit.c65 flush_icache_range(addr, addr + (PAGE_SIZE << compound_order(page))); __ia64_sync_icache_dcache()
/linux-4.1.27/arch/powerpc/mm/
H A Dhugetlbpage.c945 for (i = 0; i < (1UL << compound_order(page)); i++) { flush_dcache_icache_hugepage()
/linux-4.1.27/arch/powerpc/kvm/
H A Dbook3s_64_mmu_hv.c534 pte_size <<= compound_order(page); kvmppc_book3s_hv_page_fault()
/linux-4.1.27/drivers/infiniband/hw/ehca/
H A Dehca_mrmw.c2358 page_order = compound_order(pfn_to_page(pfn)); ehca_is_hugepage()
/linux-4.1.27/drivers/net/xen-netback/
H A Dnetback.c281 BUG_ON(size + offset > PAGE_SIZE<<compound_order(page)); xenvif_gop_frag_copy()

Completed in 2592 milliseconds