Lines Matching refs:vm_area_struct
51 void reset_vma_resv_huge_pages(struct vm_area_struct *vma);
61 int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *);
62 long follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *,
63 struct page **, struct vm_area_struct **,
65 void unmap_hugepage_range(struct vm_area_struct *,
68 struct vm_area_struct *vma,
71 void __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma,
78 int hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
81 struct vm_area_struct *vma,
111 unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
116 static inline void reset_vma_resv_huge_pages(struct vm_area_struct *vma) in reset_vma_resv_huge_pages()
155 static inline unsigned long hugetlb_change_protection(struct vm_area_struct *vma, in hugetlb_change_protection()
162 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range_final()
169 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range()
324 struct page *alloc_huge_page_noerr(struct vm_area_struct *vma,
362 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
372 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
374 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
409 static inline pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, in arch_make_huge_pte()