Searched refs:hpage (Results 1 – 6 of 6) sorted by relevance
/linux-4.1.27/mm/ |
D | memory-failure.c | 778 struct page *hpage = compound_head(p); in me_huge_page() local 789 if (!(page_mapping(hpage) || PageAnon(hpage))) { in me_huge_page() 790 res = dequeue_hwpoisoned_huge_page(hpage); in me_huge_page() 929 struct page *hpage = *hpagep; in hwpoison_user_mappings() local 938 if (!(PageLRU(hpage) || PageHuge(p))) in hwpoison_user_mappings() 945 if (!page_mapped(hpage)) in hwpoison_user_mappings() 965 mapping = page_mapping(hpage); in hwpoison_user_mappings() 966 if (!(flags & MF_MUST_KILL) && !PageDirty(hpage) && mapping && in hwpoison_user_mappings() 968 if (page_mkclean(hpage)) { in hwpoison_user_mappings() 969 SetPageDirty(hpage); in hwpoison_user_mappings() [all …]
|
D | hwpoison-inject.c | 17 struct page *hpage; in hwpoison_inject() local 27 hpage = compound_head(p); in hwpoison_inject() 31 if (!get_page_unless_zero(hpage)) in hwpoison_inject() 37 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject() 38 shake_page(hpage, 0); in hwpoison_inject() 42 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject() 51 lock_page(hpage); in hwpoison_inject() 52 err = hwpoison_filter(hpage); in hwpoison_inject() 53 unlock_page(hpage); in hwpoison_inject() 61 put_page(hpage); in hwpoison_inject()
|
D | huge_memory.c | 2316 static bool khugepaged_prealloc_page(struct page **hpage, bool *wait) in khugepaged_prealloc_page() argument 2318 if (IS_ERR(*hpage)) { in khugepaged_prealloc_page() 2323 *hpage = NULL; in khugepaged_prealloc_page() 2325 } else if (*hpage) { in khugepaged_prealloc_page() 2326 put_page(*hpage); in khugepaged_prealloc_page() 2327 *hpage = NULL; in khugepaged_prealloc_page() 2334 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument 2338 VM_BUG_ON_PAGE(*hpage, *hpage); in khugepaged_alloc_page() 2348 *hpage = alloc_pages_exact_node(node, gfp, HPAGE_PMD_ORDER); in khugepaged_alloc_page() 2349 if (unlikely(!*hpage)) { in khugepaged_alloc_page() [all …]
|
D | migrate.c | 1003 struct page *hpage, int force, in unmap_and_move_huge_page() argument 1019 if (!hugepage_migration_supported(page_hstate(hpage))) { in unmap_and_move_huge_page() 1020 putback_active_hugepage(hpage); in unmap_and_move_huge_page() 1024 new_hpage = get_new_page(hpage, private, &result); in unmap_and_move_huge_page() 1030 if (!trylock_page(hpage)) { in unmap_and_move_huge_page() 1033 lock_page(hpage); in unmap_and_move_huge_page() 1036 if (PageAnon(hpage)) in unmap_and_move_huge_page() 1037 anon_vma = page_get_anon_vma(hpage); in unmap_and_move_huge_page() 1039 if (page_mapped(hpage)) { in unmap_and_move_huge_page() 1040 try_to_unmap(hpage, in unmap_and_move_huge_page() [all …]
|
D | hugetlb.c | 3915 int dequeue_hwpoisoned_huge_page(struct page *hpage) in dequeue_hwpoisoned_huge_page() argument 3917 struct hstate *h = page_hstate(hpage); in dequeue_hwpoisoned_huge_page() 3918 int nid = page_to_nid(hpage); in dequeue_hwpoisoned_huge_page() 3926 if (!page_huge_active(hpage) && !page_count(hpage)) { in dequeue_hwpoisoned_huge_page() 3933 list_del_init(&hpage->lru); in dequeue_hwpoisoned_huge_page() 3934 set_page_refcounted(hpage); in dequeue_hwpoisoned_huge_page()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | book3s_pr.c | 490 struct page *hpage; in kvmppc_patch_dcbz() local 495 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz() 496 if (is_error_page(hpage)) in kvmppc_patch_dcbz() 503 get_page(hpage); in kvmppc_patch_dcbz() 504 page = kmap_atomic(hpage); in kvmppc_patch_dcbz() 512 put_page(hpage); in kvmppc_patch_dcbz()
|