Lines Matching refs:address

37 static int follow_pfn_pte(struct vm_area_struct *vma, unsigned long address,  in follow_pfn_pte()  argument
52 set_pte_at(vma->vm_mm, address, pte, entry); in follow_pfn_pte()
53 update_mmu_cache(vma, address, pte); in follow_pfn_pte()
62 unsigned long address, pmd_t *pmd, unsigned int flags) in follow_page_pte() argument
73 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); in follow_page_pte()
90 migration_entry_wait(mm, pmd, address); in follow_page_pte()
100 page = vm_normal_page(vma, address, pte); in follow_page_pte()
113 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte()
178 unsigned long address, unsigned int flags, in follow_page_mask() argument
190 page = follow_huge_addr(mm, address, flags & FOLL_WRITE); in follow_page_mask()
196 pgd = pgd_offset(mm, address); in follow_page_mask()
200 pud = pud_offset(pgd, address); in follow_page_mask()
204 page = follow_huge_pud(mm, address, pud, flags); in follow_page_mask()
212 pmd = pmd_offset(pud, address); in follow_page_mask()
216 page = follow_huge_pmd(mm, address, pmd, flags); in follow_page_mask()
225 split_huge_page_pmd(vma, address, pmd); in follow_page_mask()
226 return follow_page_pte(vma, address, pmd, flags); in follow_page_mask()
234 page = follow_trans_huge_pmd(vma, address, in follow_page_mask()
243 return follow_page_pte(vma, address, pmd, flags); in follow_page_mask()
246 static int get_gate_page(struct mm_struct *mm, unsigned long address, in get_gate_page() argument
259 if (address > TASK_SIZE) in get_gate_page()
260 pgd = pgd_offset_k(address); in get_gate_page()
262 pgd = pgd_offset_gate(mm, address); in get_gate_page()
264 pud = pud_offset(pgd, address); in get_gate_page()
266 pmd = pmd_offset(pud, address); in get_gate_page()
270 pte = pte_offset_map(pmd, address); in get_gate_page()
276 *page = vm_normal_page(*vma, address, *pte); in get_gate_page()
296 unsigned long address, unsigned int *flags, int *nonblocking) in faultin_page() argument
307 (stack_guard_page_start(vma, address) || in faultin_page()
308 stack_guard_page_end(vma, address + PAGE_SIZE))) in faultin_page()
321 ret = handle_mm_fault(mm, vma, address, fault_flags); in faultin_page()
588 unsigned long address, unsigned int fault_flags) in fixup_user_fault() argument
594 vma = find_extend_vma(mm, address); in fixup_user_fault()
595 if (!vma || address < vma->vm_start) in fixup_user_fault()
602 ret = handle_mm_fault(mm, vma, address, fault_flags); in fixup_user_fault()