Home
last modified time | relevance | path

Searched refs:hpage (Results 1 – 6 of 6) sorted by relevance

/linux-4.1.27/mm/
Dmemory-failure.c778 struct page *hpage = compound_head(p); in me_huge_page() local
789 if (!(page_mapping(hpage) || PageAnon(hpage))) { in me_huge_page()
790 res = dequeue_hwpoisoned_huge_page(hpage); in me_huge_page()
929 struct page *hpage = *hpagep; in hwpoison_user_mappings() local
938 if (!(PageLRU(hpage) || PageHuge(p))) in hwpoison_user_mappings()
945 if (!page_mapped(hpage)) in hwpoison_user_mappings()
965 mapping = page_mapping(hpage); in hwpoison_user_mappings()
966 if (!(flags & MF_MUST_KILL) && !PageDirty(hpage) && mapping && in hwpoison_user_mappings()
968 if (page_mkclean(hpage)) { in hwpoison_user_mappings()
969 SetPageDirty(hpage); in hwpoison_user_mappings()
[all …]
Dhwpoison-inject.c17 struct page *hpage; in hwpoison_inject() local
27 hpage = compound_head(p); in hwpoison_inject()
31 if (!get_page_unless_zero(hpage)) in hwpoison_inject()
37 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject()
38 shake_page(hpage, 0); in hwpoison_inject()
42 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject()
51 lock_page(hpage); in hwpoison_inject()
52 err = hwpoison_filter(hpage); in hwpoison_inject()
53 unlock_page(hpage); in hwpoison_inject()
61 put_page(hpage); in hwpoison_inject()
Dhuge_memory.c2316 static bool khugepaged_prealloc_page(struct page **hpage, bool *wait) in khugepaged_prealloc_page() argument
2318 if (IS_ERR(*hpage)) { in khugepaged_prealloc_page()
2323 *hpage = NULL; in khugepaged_prealloc_page()
2325 } else if (*hpage) { in khugepaged_prealloc_page()
2326 put_page(*hpage); in khugepaged_prealloc_page()
2327 *hpage = NULL; in khugepaged_prealloc_page()
2334 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument
2338 VM_BUG_ON_PAGE(*hpage, *hpage); in khugepaged_alloc_page()
2348 *hpage = alloc_pages_exact_node(node, gfp, HPAGE_PMD_ORDER); in khugepaged_alloc_page()
2349 if (unlikely(!*hpage)) { in khugepaged_alloc_page()
[all …]
Dmigrate.c1003 struct page *hpage, int force, in unmap_and_move_huge_page() argument
1019 if (!hugepage_migration_supported(page_hstate(hpage))) { in unmap_and_move_huge_page()
1020 putback_active_hugepage(hpage); in unmap_and_move_huge_page()
1024 new_hpage = get_new_page(hpage, private, &result); in unmap_and_move_huge_page()
1030 if (!trylock_page(hpage)) { in unmap_and_move_huge_page()
1033 lock_page(hpage); in unmap_and_move_huge_page()
1036 if (PageAnon(hpage)) in unmap_and_move_huge_page()
1037 anon_vma = page_get_anon_vma(hpage); in unmap_and_move_huge_page()
1039 if (page_mapped(hpage)) { in unmap_and_move_huge_page()
1040 try_to_unmap(hpage, in unmap_and_move_huge_page()
[all …]
Dhugetlb.c3915 int dequeue_hwpoisoned_huge_page(struct page *hpage) in dequeue_hwpoisoned_huge_page() argument
3917 struct hstate *h = page_hstate(hpage); in dequeue_hwpoisoned_huge_page()
3918 int nid = page_to_nid(hpage); in dequeue_hwpoisoned_huge_page()
3926 if (!page_huge_active(hpage) && !page_count(hpage)) { in dequeue_hwpoisoned_huge_page()
3933 list_del_init(&hpage->lru); in dequeue_hwpoisoned_huge_page()
3934 set_page_refcounted(hpage); in dequeue_hwpoisoned_huge_page()
/linux-4.1.27/arch/powerpc/kvm/
Dbook3s_pr.c490 struct page *hpage; in kvmppc_patch_dcbz() local
495 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz()
496 if (is_error_page(hpage)) in kvmppc_patch_dcbz()
503 get_page(hpage); in kvmppc_patch_dcbz()
504 page = kmap_atomic(hpage); in kvmppc_patch_dcbz()
512 put_page(hpage); in kvmppc_patch_dcbz()