Lines Matching refs:vm_area_struct
54 void reset_vma_resv_huge_pages(struct vm_area_struct *vma);
64 int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *);
65 long follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *,
66 struct page **, struct vm_area_struct **,
68 void unmap_hugepage_range(struct vm_area_struct *,
71 struct vm_area_struct *vma,
74 void __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma,
81 int hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
84 struct vm_area_struct *vma,
95 struct vm_area_struct *vma,
121 unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
126 static inline void reset_vma_resv_huge_pages(struct vm_area_struct *vma) in reset_vma_resv_huge_pages()
165 static inline unsigned long hugetlb_change_protection(struct vm_area_struct *vma, in hugetlb_change_protection()
172 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range_final()
179 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range()
333 struct page *alloc_huge_page(struct vm_area_struct *vma,
336 struct page *alloc_huge_page_noerr(struct vm_area_struct *vma,
376 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
386 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
388 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
423 static inline pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, in arch_make_huge_pte()