ptep 29 arch/alpha/include/asm/pgtable.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 226 arch/alpha/include/asm/pgtable.h extern inline void pmd_set(pmd_t * pmdp, pte_t * ptep) ptep 227 arch/alpha/include/asm/pgtable.h { pmd_val(*pmdp) = _PAGE_TABLE | ((((unsigned long) ptep) - PAGE_OFFSET) << (32-PAGE_SHIFT)); } ptep 249 arch/alpha/include/asm/pgtable.h extern inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 251 arch/alpha/include/asm/pgtable.h pte_val(*ptep) = 0; ptep 330 arch/alpha/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 42 arch/arc/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t ptep) ptep 44 arch/arc/include/asm/pgalloc.h pmd_set(pmd, (pte_t *) ptep); ptep 124 arch/arc/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, pgtable_t ptep) ptep 126 arch/arc/include/asm/pgalloc.h pgtable_pte_page_dtor(virt_to_page(ptep)); ptep 127 arch/arc/include/asm/pgalloc.h free_pages((unsigned long)ptep, __get_order_pte()); ptep 265 arch/arc/include/asm/pgtable.h static inline void pmd_set(pmd_t *pmdp, pte_t *ptep) ptep 267 arch/arc/include/asm/pgtable.h pmd_val(*pmdp) = (unsigned long)ptep; ptep 272 arch/arc/include/asm/pgtable.h #define pte_clear(mm, addr, ptep) set_pte_at(mm, addr, ptep, __pte(0)) ptep 329 arch/arc/include/asm/pgtable.h pte_t *ptep, pte_t pteval) ptep 331 arch/arc/include/asm/pgtable.h set_pte(ptep, pteval); ptep 366 arch/arc/include/asm/pgtable.h pte_t *ptep); ptep 515 arch/arc/mm/tlb.c void create_tlb(struct vm_area_struct *vma, unsigned long vaddr, pte_t *ptep) ptep 557 arch/arc/mm/tlb.c pte_val(*ptep) |= (_PAGE_PRESENT | _PAGE_ACCESSED); ptep 564 arch/arc/mm/tlb.c pd0 = vaddr | asid_or_sasid | (pte_val(*ptep) & PTE_BITS_IN_PD0); ptep 573 arch/arc/mm/tlb.c rwx = pte_val(*ptep) & PTE_BITS_RWX; ptep 575 arch/arc/mm/tlb.c if (pte_val(*ptep) & _PAGE_GLOBAL) ptep 580 arch/arc/mm/tlb.c pd1 = rwx | (pte_val(*ptep) & PTE_BITS_NON_RWX_IN_PD1); ptep 597 arch/arc/mm/tlb.c pte_t *ptep) ptep 600 arch/arc/mm/tlb.c phys_addr_t paddr = pte_val(*ptep) & PAGE_MASK; ptep 601 arch/arc/mm/tlb.c struct page *page = pfn_to_page(pte_pfn(*ptep)); ptep 603 arch/arc/mm/tlb.c create_tlb(vma, vaddr, ptep); ptep 21 arch/arm/include/asm/hugetlb-3level.h static inline pte_t huge_ptep_get(pte_t *ptep) ptep 23 arch/arm/include/asm/hugetlb-3level.h pte_t retval = *ptep; ptep 69 arch/arm/include/asm/kvm_mmu.h #define kvm_mk_pmd(ptep) __pmd(__pa(ptep) | PMD_TYPE_TABLE) ptep 194 arch/arm/include/asm/kvm_mmu.h #define kvm_pte_table_empty(kvm, ptep) kvm_page_empty(ptep) ptep 198 arch/arm/include/asm/kvm_mmu.h #define hyp_pte_table_empty(ptep) kvm_page_empty(ptep) ptep 129 arch/arm/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) ptep 134 arch/arm/include/asm/pgalloc.h __pmd_populate(pmdp, __pa(ptep), _PAGE_KERNEL_TABLE); ptep 138 arch/arm/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmdp, pgtable_t ptep) ptep 148 arch/arm/include/asm/pgalloc.h __pmd_populate(pmdp, page_to_phys(ptep), prot); ptep 212 arch/arm/include/asm/pgtable-2level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,pte,ext) ptep 193 arch/arm/include/asm/pgtable-3level.h #define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,__pte(pte_val(pte)|(ext))) ptep 217 arch/arm/include/asm/pgtable.h #define pte_clear(mm,addr,ptep) set_pte_ext(ptep, __pte(0), 0) ptep 256 arch/arm/include/asm/pgtable.h pte_t *ptep, pte_t pteval) ptep 266 arch/arm/include/asm/pgtable.h set_pte_ext(ptep, pteval, ext); ptep 70 arch/arm/include/asm/proc-fns.h void (*set_pte_ext)(pte_t *ptep, pte_t pte); ptep 72 arch/arm/include/asm/proc-fns.h void (*set_pte_ext)(pte_t *ptep, pte_t pte, unsigned int ext); ptep 92 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte); ptep 94 arch/arm/include/asm/proc-fns.h extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext); ptep 56 arch/arm/include/asm/stage2_pgtable.h #define stage2_pte_table_empty(kvm, ptep) kvm_page_empty(ptep) ptep 633 arch/arm/include/asm/tlbflush.h pte_t *ptep); ptep 636 arch/arm/include/asm/tlbflush.h unsigned long addr, pte_t *ptep) ptep 11 arch/arm/kernel/efi.c static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data) ptep 14 arch/arm/kernel/efi.c pte_t pte = *ptep; ptep 20 arch/arm/kernel/efi.c set_pte_ext(ptep, pte, PTE_EXT_NG); ptep 23 arch/arm/lib/uaccess_with_memcpy.c pin_page_for_write(const void __user *_addr, pte_t **ptep, spinlock_t **ptlp) ptep 63 arch/arm/lib/uaccess_with_memcpy.c *ptep = NULL; ptep 78 arch/arm/lib/uaccess_with_memcpy.c *ptep = pte; ptep 38 arch/arm/mm/fault-armv.c unsigned long pfn, pte_t *ptep) ptep 40 arch/arm/mm/fault-armv.c pte_t entry = *ptep; ptep 58 arch/arm/mm/fault-armv.c set_pte_at(vma->vm_mm, address, ptep, entry); ptep 130 arch/arm/mm/fault-armv.c unsigned long addr, pte_t *ptep, unsigned long pfn) ptep 161 arch/arm/mm/fault-armv.c do_adjust_pte(vma, addr, pfn, ptep); ptep 178 arch/arm/mm/fault-armv.c pte_t *ptep) ptep 180 arch/arm/mm/fault-armv.c unsigned long pfn = pte_pfn(*ptep); ptep 200 arch/arm/mm/fault-armv.c make_coherent(mapping, vma, addr, ptep, pfn); ptep 21 arch/arm/mm/highmem.c pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); ptep 23 arch/arm/mm/highmem.c set_pte_ext(ptep, pte, 0); ptep 29 arch/arm/mm/highmem.c pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); ptep 31 arch/arm/mm/highmem.c return *ptep; ptep 28 arch/arm/mm/mm.h pte_t *ptep = pte_offset_kernel(top_pmd, va); ptep 29 arch/arm/mm/mm.h set_pte_ext(ptep, pte, 0); ptep 35 arch/arm/mm/mm.h pte_t *ptep = pte_offset_kernel(top_pmd, va); ptep 36 arch/arm/mm/mm.h return *ptep; ptep 17 arch/arm/mm/pageattr.c static int change_page_range(pte_t *ptep, unsigned long addr, void *data) ptep 20 arch/arm/mm/pageattr.c pte_t pte = *ptep; ptep 25 arch/arm/mm/pageattr.c set_pte_ext(ptep, pte, 0); ptep 21 arch/arm64/include/asm/hugetlb.h static inline pte_t huge_ptep_get(pte_t *ptep) ptep 23 arch/arm64/include/asm/hugetlb.h return READ_ONCE(*ptep); ptep 42 arch/arm64/include/asm/hugetlb.h pte_t *ptep, pte_t pte); ptep 45 arch/arm64/include/asm/hugetlb.h unsigned long addr, pte_t *ptep, ptep 49 arch/arm64/include/asm/hugetlb.h unsigned long addr, pte_t *ptep); ptep 52 arch/arm64/include/asm/hugetlb.h unsigned long addr, pte_t *ptep); ptep 55 arch/arm64/include/asm/hugetlb.h unsigned long addr, pte_t *ptep); ptep 58 arch/arm64/include/asm/hugetlb.h pte_t *ptep, unsigned long sz); ptep 60 arch/arm64/include/asm/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz); ptep 170 arch/arm64/include/asm/kvm_mmu.h #define kvm_mk_pmd(ptep) \ ptep 171 arch/arm64/include/asm/kvm_mmu.h __pmd(__phys_to_pmd_val(__pa(ptep)) | PMD_TYPE_TABLE) ptep 224 arch/arm64/include/asm/kvm_mmu.h static inline void kvm_set_s2pte_readonly(pte_t *ptep) ptep 228 arch/arm64/include/asm/kvm_mmu.h pteval = READ_ONCE(pte_val(*ptep)); ptep 233 arch/arm64/include/asm/kvm_mmu.h pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval); ptep 237 arch/arm64/include/asm/kvm_mmu.h static inline bool kvm_s2pte_readonly(pte_t *ptep) ptep 239 arch/arm64/include/asm/kvm_mmu.h return (READ_ONCE(pte_val(*ptep)) & PTE_S2_RDWR) == PTE_S2_RDONLY; ptep 242 arch/arm64/include/asm/kvm_mmu.h static inline bool kvm_s2pte_exec(pte_t *ptep) ptep 244 arch/arm64/include/asm/kvm_mmu.h return !(READ_ONCE(pte_val(*ptep)) & PTE_S2_XN); ptep 287 arch/arm64/include/asm/kvm_mmu.h #define hyp_pte_table_empty(ptep) kvm_page_empty(ptep) ptep 95 arch/arm64/include/asm/pgalloc.h static inline void __pmd_populate(pmd_t *pmdp, phys_addr_t ptep, ptep 98 arch/arm64/include/asm/pgalloc.h set_pmd(pmdp, __pmd(__phys_to_pmd_val(ptep) | prot)); ptep 106 arch/arm64/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) ptep 111 arch/arm64/include/asm/pgalloc.h __pmd_populate(pmdp, __pa(ptep), PMD_TYPE_TABLE); ptep 115 arch/arm64/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmdp, pgtable_t ptep) ptep 117 arch/arm64/include/asm/pgalloc.h __pmd_populate(pmdp, page_to_phys(ptep), PMD_TYPE_TABLE); ptep 70 arch/arm64/include/asm/pgtable.h #define pte_clear(mm,addr,ptep) set_pte(ptep, __pte(0)) ptep 211 arch/arm64/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte) ptep 213 arch/arm64/include/asm/pgtable.h WRITE_ONCE(*ptep, pte); ptep 243 arch/arm64/include/asm/pgtable.h static inline void __check_racy_pte_update(struct mm_struct *mm, pte_t *ptep, ptep 251 arch/arm64/include/asm/pgtable.h old_pte = READ_ONCE(*ptep); ptep 272 arch/arm64/include/asm/pgtable.h pte_t *ptep, pte_t pte) ptep 277 arch/arm64/include/asm/pgtable.h __check_racy_pte_update(mm, ptep, pte); ptep 279 arch/arm64/include/asm/pgtable.h set_pte(ptep, pte); ptep 676 arch/arm64/include/asm/pgtable.h unsigned long address, pte_t *ptep, ptep 703 arch/arm64/include/asm/pgtable.h static inline int __ptep_test_and_clear_young(pte_t *ptep) ptep 707 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); ptep 711 arch/arm64/include/asm/pgtable.h pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep), ptep 720 arch/arm64/include/asm/pgtable.h pte_t *ptep) ptep 722 arch/arm64/include/asm/pgtable.h return __ptep_test_and_clear_young(ptep); ptep 727 arch/arm64/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 729 arch/arm64/include/asm/pgtable.h int young = ptep_test_and_clear_young(vma, address, ptep); ptep 758 arch/arm64/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 760 arch/arm64/include/asm/pgtable.h return __pte(xchg_relaxed(&pte_val(*ptep), 0)); ptep 777 arch/arm64/include/asm/pgtable.h static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep) ptep 781 arch/arm64/include/asm/pgtable.h pte = READ_ONCE(*ptep); ptep 791 arch/arm64/include/asm/pgtable.h pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep), ptep 847 arch/arm64/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 215 arch/arm64/include/asm/stage2_pgtable.h static inline bool stage2_pte_table_empty(struct kvm *kvm, pte_t *ptep) ptep 217 arch/arm64/include/asm/stage2_pgtable.h return kvm_page_empty(ptep); ptep 85 arch/arm64/kernel/efi.c static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data) ptep 88 arch/arm64/kernel/efi.c pte_t pte = READ_ONCE(*ptep); ptep 94 arch/arm64/kernel/efi.c set_pte(ptep, pte); ptep 208 arch/arm64/kernel/hibernate.c pte_t *ptep; ptep 247 arch/arm64/kernel/hibernate.c ptep = allocator(mask); ptep 248 arch/arm64/kernel/hibernate.c if (!ptep) { ptep 252 arch/arm64/kernel/hibernate.c pmd_populate_kernel(&init_mm, pmdp, ptep); ptep 255 arch/arm64/kernel/hibernate.c ptep = pte_offset_kernel(pmdp, dst_addr); ptep 256 arch/arm64/kernel/hibernate.c set_pte(ptep, pfn_pte(virt_to_pfn(dst), PAGE_KERNEL_EXEC)); ptep 299 arch/arm64/mm/dump.c pte_t *ptep = pte_offset_kernel(pmdp, start); ptep 302 arch/arm64/mm/dump.c note_page(st, addr, 4, READ_ONCE(pte_val(*ptep))); ptep 303 arch/arm64/mm/dump.c } while (ptep++, addr += PAGE_SIZE, addr != end); ptep 161 arch/arm64/mm/fault.c pte_t *ptep, pte; ptep 178 arch/arm64/mm/fault.c ptep = pte_offset_map(pmdp, addr); ptep 179 arch/arm64/mm/fault.c pte = READ_ONCE(*ptep); ptep 181 arch/arm64/mm/fault.c pte_unmap(ptep); ptep 198 arch/arm64/mm/fault.c unsigned long address, pte_t *ptep, ptep 202 arch/arm64/mm/fault.c pte_t pte = READ_ONCE(*ptep); ptep 223 arch/arm64/mm/fault.c pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval); ptep 67 arch/arm64/mm/hugetlbpage.c pte_t *ptep, size_t *pgsize) ptep 76 arch/arm64/mm/hugetlbpage.c if ((pte_t *)pmdp == ptep) { ptep 119 arch/arm64/mm/hugetlbpage.c pte_t *ptep, ptep 123 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep); ptep 127 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, addr += pgsize, ptep++) { ptep 128 arch/arm64/mm/hugetlbpage.c pte_t pte = ptep_get_and_clear(mm, addr, ptep); ptep 160 arch/arm64/mm/hugetlbpage.c pte_t *ptep, ptep 167 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, addr += pgsize, ptep++) ptep 168 arch/arm64/mm/hugetlbpage.c pte_clear(mm, addr, ptep); ptep 174 arch/arm64/mm/hugetlbpage.c pte_t *ptep, pte_t pte) ptep 189 arch/arm64/mm/hugetlbpage.c set_pte_at(mm, addr, ptep, pte); ptep 193 arch/arm64/mm/hugetlbpage.c ncontig = find_num_contig(mm, addr, ptep, &pgsize); ptep 198 arch/arm64/mm/hugetlbpage.c clear_flush(mm, addr, ptep, pgsize, ncontig); ptep 200 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn += dpfn) ptep 201 arch/arm64/mm/hugetlbpage.c set_pte_at(mm, addr, ptep, pfn_pte(pfn, hugeprot)); ptep 205 arch/arm64/mm/hugetlbpage.c pte_t *ptep, pte_t pte, unsigned long sz) ptep 212 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, ptep++) ptep 213 arch/arm64/mm/hugetlbpage.c set_pte(ptep, pte); ptep 222 arch/arm64/mm/hugetlbpage.c pte_t *ptep = NULL; ptep 230 arch/arm64/mm/hugetlbpage.c ptep = (pte_t *)pudp; ptep 244 arch/arm64/mm/hugetlbpage.c ptep = pte_alloc_map(mm, pmdp, addr); ptep 248 arch/arm64/mm/hugetlbpage.c ptep = huge_pmd_share(mm, addr, pudp); ptep 250 arch/arm64/mm/hugetlbpage.c ptep = (pte_t *)pmd_alloc(mm, pudp, addr); ptep 257 arch/arm64/mm/hugetlbpage.c return ptep; ptep 314 arch/arm64/mm/hugetlbpage.c pte_t *ptep, unsigned long sz) ptep 321 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, addr += pgsize, ptep++) ptep 322 arch/arm64/mm/hugetlbpage.c pte_clear(mm, addr, ptep); ptep 326 arch/arm64/mm/hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 330 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep); ptep 333 arch/arm64/mm/hugetlbpage.c return ptep_get_and_clear(mm, addr, ptep); ptep 335 arch/arm64/mm/hugetlbpage.c ncontig = find_num_contig(mm, addr, ptep, &pgsize); ptep 337 arch/arm64/mm/hugetlbpage.c return get_clear_flush(mm, addr, ptep, pgsize, ncontig); ptep 349 arch/arm64/mm/hugetlbpage.c static int __cont_access_flags_changed(pte_t *ptep, pte_t pte, int ncontig) ptep 353 arch/arm64/mm/hugetlbpage.c if (pte_write(pte) != pte_write(huge_ptep_get(ptep))) ptep 357 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep + i); ptep 370 arch/arm64/mm/hugetlbpage.c unsigned long addr, pte_t *ptep, ptep 380 arch/arm64/mm/hugetlbpage.c return ptep_set_access_flags(vma, addr, ptep, pte, dirty); ptep 382 arch/arm64/mm/hugetlbpage.c ncontig = find_num_contig(vma->vm_mm, addr, ptep, &pgsize); ptep 385 arch/arm64/mm/hugetlbpage.c if (!__cont_access_flags_changed(ptep, pte, ncontig)) ptep 388 arch/arm64/mm/hugetlbpage.c orig_pte = get_clear_flush(vma->vm_mm, addr, ptep, pgsize, ncontig); ptep 398 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn += dpfn) ptep 399 arch/arm64/mm/hugetlbpage.c set_pte_at(vma->vm_mm, addr, ptep, pfn_pte(pfn, hugeprot)); ptep 405 arch/arm64/mm/hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 413 arch/arm64/mm/hugetlbpage.c if (!pte_cont(READ_ONCE(*ptep))) { ptep 414 arch/arm64/mm/hugetlbpage.c ptep_set_wrprotect(mm, addr, ptep); ptep 418 arch/arm64/mm/hugetlbpage.c ncontig = find_num_contig(mm, addr, ptep, &pgsize); ptep 421 arch/arm64/mm/hugetlbpage.c pte = get_clear_flush(mm, addr, ptep, pgsize, ncontig); ptep 427 arch/arm64/mm/hugetlbpage.c for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn += dpfn) ptep 428 arch/arm64/mm/hugetlbpage.c set_pte_at(mm, addr, ptep, pfn_pte(pfn, hugeprot)); ptep 432 arch/arm64/mm/hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 437 arch/arm64/mm/hugetlbpage.c if (!pte_cont(READ_ONCE(*ptep))) { ptep 438 arch/arm64/mm/hugetlbpage.c ptep_clear_flush(vma, addr, ptep); ptep 442 arch/arm64/mm/hugetlbpage.c ncontig = find_num_contig(vma->vm_mm, addr, ptep, &pgsize); ptep 443 arch/arm64/mm/hugetlbpage.c clear_flush(vma->vm_mm, addr, ptep, pgsize, ncontig); ptep 104 arch/arm64/mm/kasan_init.c pte_t *ptep = kasan_pte_offset(pmdp, addr, node, early); ptep 113 arch/arm64/mm/kasan_init.c set_pte(ptep, pfn_pte(__phys_to_pfn(page_phys), PAGE_KERNEL)); ptep 114 arch/arm64/mm/kasan_init.c } while (ptep++, addr = next, addr != end && pte_none(READ_ONCE(*ptep))); ptep 143 arch/arm64/mm/mmu.c pte_t *ptep; ptep 145 arch/arm64/mm/mmu.c ptep = pte_set_fixmap_offset(pmdp, addr); ptep 147 arch/arm64/mm/mmu.c pte_t old_pte = READ_ONCE(*ptep); ptep 149 arch/arm64/mm/mmu.c set_pte(ptep, pfn_pte(__phys_to_pfn(phys), prot)); ptep 156 arch/arm64/mm/mmu.c READ_ONCE(pte_val(*ptep)))); ptep 159 arch/arm64/mm/mmu.c } while (ptep++, addr += PAGE_SIZE, addr != end); ptep 696 arch/arm64/mm/mmu.c pte_t *ptep, pte; ptep 721 arch/arm64/mm/mmu.c ptep = pte_offset_kernel(pmdp, addr); ptep 722 arch/arm64/mm/mmu.c pte = READ_ONCE(*ptep); ptep 868 arch/arm64/mm/mmu.c pte_t *ptep; ptep 872 arch/arm64/mm/mmu.c ptep = fixmap_pte(addr); ptep 875 arch/arm64/mm/mmu.c set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, flags)); ptep 877 arch/arm64/mm/mmu.c pte_clear(&init_mm, addr, ptep); ptep 22 arch/arm64/mm/pageattr.c static int change_page_range(pte_t *ptep, unsigned long addr, void *data) ptep 25 arch/arm64/mm/pageattr.c pte_t pte = READ_ONCE(*ptep); ptep 30 arch/arm64/mm/pageattr.c set_pte(ptep, pte); ptep 203 arch/arm64/mm/pageattr.c pte_t *ptep; ptep 227 arch/arm64/mm/pageattr.c ptep = pte_offset_kernel(pmdp, addr); ptep 228 arch/arm64/mm/pageattr.c return pte_valid(READ_ONCE(*ptep)); ptep 51 arch/c6x/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pteval) set_pte(ptep, pteval) ptep 37 arch/csky/abiv1/cacheflush.c pte_t *ptep) ptep 39 arch/csky/abiv1/cacheflush.c unsigned long pfn = pte_pfn(*ptep); ptep 47 arch/csky/include/asm/pgtable.h #define pte_clear(mm, addr, ptep) set_pte((ptep), \ ptep 125 arch/csky/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pteval) set_pte(ptep, pteval) ptep 74 arch/hexagon/include/asm/cacheflush.h unsigned long address, pte_t *ptep) ptep 179 arch/hexagon/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 185 arch/hexagon/include/asm/pgtable.h *ptep = pteval; ptep 207 arch/hexagon/include/asm/pgtable.h pte_t *ptep) ptep 209 arch/hexagon/include/asm/pgtable.h pte_val(*ptep) = _NULL_PTE; ptep 407 arch/hexagon/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte) set_pte(ptep, pte) ptep 26 arch/ia64/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 330 arch/ia64/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 337 arch/ia64/include/asm/pgtable.h (!pte_present(*ptep) || ptep 338 arch/ia64/include/asm/pgtable.h pte_pfn(*ptep) != pte_pfn(pteval))) ptep 341 arch/ia64/include/asm/pgtable.h *ptep = pteval; ptep 344 arch/ia64/include/asm/pgtable.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 410 arch/ia64/include/asm/pgtable.h ptep_test_and_clear_young (struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) ptep 413 arch/ia64/include/asm/pgtable.h if (!pte_young(*ptep)) ptep 415 arch/ia64/include/asm/pgtable.h return test_and_clear_bit(_PAGE_A_BIT, ptep); ptep 417 arch/ia64/include/asm/pgtable.h pte_t pte = *ptep; ptep 420 arch/ia64/include/asm/pgtable.h set_pte_at(vma->vm_mm, addr, ptep, pte_mkold(pte)); ptep 426 arch/ia64/include/asm/pgtable.h ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 429 arch/ia64/include/asm/pgtable.h return __pte(xchg((long *) ptep, 0)); ptep 431 arch/ia64/include/asm/pgtable.h pte_t pte = *ptep; ptep 432 arch/ia64/include/asm/pgtable.h pte_clear(mm, addr, ptep); ptep 438 arch/ia64/include/asm/pgtable.h ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 444 arch/ia64/include/asm/pgtable.h old = pte_val(*ptep); ptep 446 arch/ia64/include/asm/pgtable.h } while (cmpxchg((unsigned long *) ptep, old, new) != old); ptep 448 arch/ia64/include/asm/pgtable.h pte_t old_pte = *ptep; ptep 449 arch/ia64/include/asm/pgtable.h set_pte_at(mm, addr, ptep, pte_wrprotect(old_pte)); ptep 459 arch/ia64/include/asm/pgtable.h #define update_mmu_cache(vma, address, ptep) do { } while (0) ptep 34 arch/ia64/mm/fault.c pte_t *ptep, pte; ptep 48 arch/ia64/mm/fault.c ptep = pte_offset_kernel(pmd, address); ptep 49 arch/ia64/mm/fault.c if (!ptep) ptep 52 arch/ia64/mm/fault.c pte = *ptep; ptep 91 arch/ia64/mm/hugetlbpage.c pte_t *ptep; ptep 96 arch/ia64/mm/hugetlbpage.c ptep = huge_pte_offset(mm, addr, HPAGE_SIZE); ptep 97 arch/ia64/mm/hugetlbpage.c if (!ptep || pte_none(*ptep)) ptep 99 arch/ia64/mm/hugetlbpage.c page = pte_page(*ptep); ptep 165 arch/m68k/include/asm/mcf_pgtable.h #define pmd_set(pmdp, ptep) do {} while (0) ptep 186 arch/m68k/include/asm/mcf_pgtable.h pte_t *ptep) ptep 188 arch/m68k/include/asm/mcf_pgtable.h pte_val(*ptep) = 0; ptep 364 arch/m68k/include/asm/mcf_pgtable.h pte_t *ptep; ptep 369 arch/m68k/include/asm/mcf_pgtable.h ptep = pte_offset_kernel(pmdp, addr); ptep 370 arch/m68k/include/asm/mcf_pgtable.h *ptep = pte_mknocache(*ptep); ptep 380 arch/m68k/include/asm/mcf_pgtable.h pte_t *ptep; ptep 385 arch/m68k/include/asm/mcf_pgtable.h ptep = pte_offset_kernel(pmdp, addr); ptep 386 arch/m68k/include/asm/mcf_pgtable.h *ptep = pte_mkcache(*ptep); ptep 109 arch/m68k/include/asm/motorola_pgtable.h static inline void pmd_set(pmd_t *pmdp, pte_t *ptep) ptep 111 arch/m68k/include/asm/motorola_pgtable.h unsigned long ptbl = virt_to_phys(ptep) | _PAGE_TABLE | _PAGE_ACCESSED; ptep 132 arch/m68k/include/asm/motorola_pgtable.h #define pte_clear(mm,addr,ptep) ({ pte_val(*(ptep)) = 0; }) ptep 243 arch/m68k/include/asm/motorola_pgtable.h pte_t *ptep; ptep 247 arch/m68k/include/asm/motorola_pgtable.h ptep = pte_offset_kernel(pmdp, addr); ptep 248 arch/m68k/include/asm/motorola_pgtable.h *ptep = pte_mknocache(*ptep); ptep 259 arch/m68k/include/asm/motorola_pgtable.h pte_t *ptep; ptep 263 arch/m68k/include/asm/motorola_pgtable.h ptep = pte_offset_kernel(pmdp, addr); ptep 264 arch/m68k/include/asm/motorola_pgtable.h *ptep = pte_mkcache(*ptep); ptep 29 arch/m68k/include/asm/pgtable_mm.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 133 arch/m68k/include/asm/pgtable_mm.h unsigned long address, pte_t *ptep) ptep 111 arch/m68k/include/asm/sun3_pgtable.h #define pmd_set(pmdp,ptep) do {} while (0) ptep 125 arch/m68k/include/asm/sun3_pgtable.h static inline void pte_clear (struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 127 arch/m68k/include/asm/sun3_pgtable.h pte_val (*ptep) = 0; ptep 34 arch/m68k/mm/sun3kmap.c pte_t ptep; ptep 36 arch/m68k/mm/sun3kmap.c ptep = pfn_pte(phys >> PAGE_SHIFT, PAGE_KERNEL); ptep 37 arch/m68k/mm/sun3kmap.c pte = pte_val(ptep); ptep 28 arch/m68k/sun3/dvma.c pte_t ptep; ptep 33 arch/m68k/sun3/dvma.c ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); ptep 34 arch/m68k/sun3/dvma.c pte = pte_val(ptep); ptep 298 arch/microblaze/include/asm/pgtable.h #define pte_clear(mm, addr, ptep) \ ptep 299 arch/microblaze/include/asm/pgtable.h do { set_pte_at((mm), (addr), (ptep), __pte(0)); } while (0) ptep 421 arch/microblaze/include/asm/pgtable.h pte_t *ptep, pte_t pte) ptep 423 arch/microblaze/include/asm/pgtable.h *ptep = pte; ptep 427 arch/microblaze/include/asm/pgtable.h pte_t *ptep, pte_t pte) ptep 429 arch/microblaze/include/asm/pgtable.h *ptep = pte; ptep 434 arch/microblaze/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 436 arch/microblaze/include/asm/pgtable.h return (pte_update(ptep, _PAGE_ACCESSED, 0) & _PAGE_ACCESSED) != 0; ptep 440 arch/microblaze/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 442 arch/microblaze/include/asm/pgtable.h return (pte_update(ptep, \ ptep 448 arch/microblaze/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 450 arch/microblaze/include/asm/pgtable.h return __pte(pte_update(ptep, ~_PAGE_HASHPTE, 0)); ptep 460 arch/microblaze/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 462 arch/microblaze/include/asm/pgtable.h pte_update(ptep, 0, _PAGE_DIRTY); ptep 42 arch/microblaze/include/asm/tlbflush.h #define update_mmu_cache(vma, addr, ptep) do { } while (0) ptep 164 arch/microblaze/kernel/signal.c pte_t *ptep; ptep 203 arch/microblaze/kernel/signal.c ptep = pte_offset_map(pmdp, address); ptep 204 arch/microblaze/kernel/signal.c if (pte_present(*ptep)) { ptep 205 arch/microblaze/kernel/signal.c address = (unsigned long) page_address(pte_page(*ptep)); ptep 213 arch/microblaze/kernel/signal.c pte_unmap(ptep); ptep 188 arch/microblaze/mm/pgtable.c static int get_pteptr(struct mm_struct *mm, unsigned long addr, pte_t **ptep) ptep 202 arch/microblaze/mm/pgtable.c *ptep = pte; ptep 42 arch/mips/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 45 arch/mips/include/asm/hugetlb.h pte_t pte = *ptep; ptep 48 arch/mips/include/asm/hugetlb.h set_pte_at(mm, addr, ptep, clear); ptep 54 arch/mips/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 69 arch/mips/include/asm/hugetlb.h pte_t *ptep, pte_t pte, ptep 72 arch/mips/include/asm/hugetlb.h int changed = !pte_same(*ptep, pte); ptep 75 arch/mips/include/asm/hugetlb.h set_pte_at(vma->vm_mm, addr, ptep, pte); ptep 134 arch/mips/include/asm/pgtable.h pte_t *ptep, pte_t pteval); ptep 147 arch/mips/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pte) ptep 149 arch/mips/include/asm/pgtable.h ptep->pte_high = pte.pte_high; ptep 151 arch/mips/include/asm/pgtable.h ptep->pte_low = pte.pte_low; ptep 158 arch/mips/include/asm/pgtable.h pte_t *buddy = ptep_buddy(ptep); ptep 171 arch/mips/include/asm/pgtable.h static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 178 arch/mips/include/asm/pgtable.h if (ptep_buddy(ptep)->pte_high & _PAGE_GLOBAL) ptep 181 arch/mips/include/asm/pgtable.h if (ptep_buddy(ptep)->pte_low & _PAGE_GLOBAL) ptep 185 arch/mips/include/asm/pgtable.h set_pte_at(mm, addr, ptep, null); ptep 199 arch/mips/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 201 arch/mips/include/asm/pgtable.h *ptep = pteval; ptep 204 arch/mips/include/asm/pgtable.h pte_t *buddy = ptep_buddy(ptep); ptep 218 arch/mips/include/asm/pgtable.h static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 223 arch/mips/include/asm/pgtable.h if (pte_val(*ptep_buddy(ptep)) & _PAGE_GLOBAL) ptep 224 arch/mips/include/asm/pgtable.h set_pte_at(mm, addr, ptep, __pte(_PAGE_GLOBAL)); ptep 227 arch/mips/include/asm/pgtable.h set_pte_at(mm, addr, ptep, __pte(0)); ptep 233 arch/mips/include/asm/pgtable.h pte_t *ptep, pte_t pteval) ptep 240 arch/mips/include/asm/pgtable.h if (pte_present(*ptep) && (pte_pfn(*ptep) == pte_pfn(pteval))) ptep 245 arch/mips/include/asm/pgtable.h set_pte(ptep, pteval); ptep 478 arch/mips/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 480 arch/mips/include/asm/pgtable.h pte_t pte = *ptep; ptep 22 arch/mips/include/asm/tlbex.h void build_update_entries(u32 **p, unsigned int tmp, unsigned int ptep); ptep 616 arch/mips/kvm/mmu.c pte_t *ptep; ptep 624 arch/mips/kvm/mmu.c ptep = kvm_mips_pte_for_gpa(kvm, NULL, gpa); ptep 625 arch/mips/kvm/mmu.c if (!ptep || !pte_present(*ptep)) { ptep 631 arch/mips/kvm/mmu.c if (!pte_young(*ptep)) { ptep 632 arch/mips/kvm/mmu.c set_pte(ptep, pte_mkyoung(*ptep)); ptep 633 arch/mips/kvm/mmu.c pfn = pte_pfn(*ptep); ptep 637 arch/mips/kvm/mmu.c if (write_fault && !pte_dirty(*ptep)) { ptep 638 arch/mips/kvm/mmu.c if (!pte_write(*ptep)) { ptep 644 arch/mips/kvm/mmu.c set_pte(ptep, pte_mkdirty(*ptep)); ptep 645 arch/mips/kvm/mmu.c pfn = pte_pfn(*ptep); ptep 651 arch/mips/kvm/mmu.c *out_entry = *ptep; ptep 653 arch/mips/kvm/mmu.c *out_buddy = *ptep_buddy(ptep); ptep 695 arch/mips/kvm/mmu.c pte_t *ptep, entry, old_pte; ptep 753 arch/mips/kvm/mmu.c ptep = kvm_mips_pte_for_gpa(kvm, memcache, gpa); ptep 768 arch/mips/kvm/mmu.c old_pte = *ptep; ptep 769 arch/mips/kvm/mmu.c set_pte(ptep, entry); ptep 773 arch/mips/kvm/mmu.c *out_entry = *ptep; ptep 775 arch/mips/kvm/mmu.c *out_buddy = *ptep_buddy(ptep); ptep 810 arch/mips/kvm/mmu.c pte_t *ptep; ptep 815 arch/mips/kvm/mmu.c ptep = kvm_mips_walk_pgd(pgdp, NULL, addr); ptep 816 arch/mips/kvm/mmu.c if (ptep) { ptep 817 arch/mips/kvm/mmu.c ptep[0] = pfn_pte(0, __pgprot(0)); ptep 818 arch/mips/kvm/mmu.c ptep[1] = pfn_pte(0, __pgprot(0)); ptep 823 arch/mips/kvm/mmu.c ptep = kvm_mips_walk_pgd(pgdp, NULL, addr); ptep 824 arch/mips/kvm/mmu.c if (ptep) { ptep 825 arch/mips/kvm/mmu.c ptep[0] = pfn_pte(0, __pgprot(0)); ptep 826 arch/mips/kvm/mmu.c ptep[1] = pfn_pte(0, __pgprot(0)); ptep 1101 arch/mips/kvm/mmu.c pte_t *ptep; ptep 1103 arch/mips/kvm/mmu.c ptep = kvm_trap_emul_pte_for_gva(vcpu, badvaddr); ptep 1104 arch/mips/kvm/mmu.c if (!ptep) { ptep 1111 arch/mips/kvm/mmu.c *ptep = pte_mkyoung(pte_mkdirty(pfn_pte(pfn, PAGE_SHARED))); ptep 246 arch/mips/mm/c-r3k.c pte_t *ptep; ptep 258 arch/mips/mm/c-r3k.c ptep = pte_offset(pmdp, addr); ptep 261 arch/mips/mm/c-r3k.c if (!(pte_val(*ptep) & _PAGE_PRESENT)) ptep 655 arch/mips/mm/c-r4k.c pte_t *ptep; ptep 669 arch/mips/mm/c-r4k.c ptep = pte_offset(pmdp, addr); ptep 675 arch/mips/mm/c-r4k.c if (!(pte_present(*ptep))) ptep 678 arch/mips/mm/c-r4k.c if ((mm == current->active_mm) && (pte_val(*ptep) & _PAGE_VALID)) ptep 175 arch/mips/mm/c-tx39.c pte_t *ptep; ptep 188 arch/mips/mm/c-tx39.c ptep = pte_offset(pmdp, page); ptep 194 arch/mips/mm/c-tx39.c if (!(pte_val(*ptep) & _PAGE_PRESENT)) ptep 203 arch/mips/mm/c-tx39.c if ((mm == current->active_mm) && (pte_val(*ptep) & _PAGE_VALID)) { ptep 300 arch/mips/mm/tlb-r4k.c pte_t *ptep; ptep 331 arch/mips/mm/tlb-r4k.c ptep = (pte_t *)pmdp; ptep 332 arch/mips/mm/tlb-r4k.c lo = pte_to_entrylo(pte_val(*ptep)); ptep 346 arch/mips/mm/tlb-r4k.c ptep = pte_offset_map(pmdp, address); ptep 350 arch/mips/mm/tlb-r4k.c write_c0_entrylo0(pte_to_entrylo(ptep->pte_high)); ptep 352 arch/mips/mm/tlb-r4k.c writex_c0_entrylo0(ptep->pte_low & _PFNX_MASK); ptep 353 arch/mips/mm/tlb-r4k.c ptep++; ptep 354 arch/mips/mm/tlb-r4k.c write_c0_entrylo1(pte_to_entrylo(ptep->pte_high)); ptep 356 arch/mips/mm/tlb-r4k.c writex_c0_entrylo1(ptep->pte_low & _PFNX_MASK); ptep 358 arch/mips/mm/tlb-r4k.c write_c0_entrylo0(ptep->pte_high); ptep 359 arch/mips/mm/tlb-r4k.c ptep++; ptep 360 arch/mips/mm/tlb-r4k.c write_c0_entrylo1(ptep->pte_high); ptep 363 arch/mips/mm/tlb-r4k.c write_c0_entrylo0(pte_to_entrylo(pte_val(*ptep++))); ptep 364 arch/mips/mm/tlb-r4k.c write_c0_entrylo1(pte_to_entrylo(pte_val(*ptep))); ptep 1048 arch/mips/mm/tlbex.c void build_update_entries(u32 **p, unsigned int tmp, unsigned int ptep) ptep 1060 arch/mips/mm/tlbex.c uasm_i_lw(p, tmp, pte_off_even, ptep); /* even pte */ ptep 1065 arch/mips/mm/tlbex.c uasm_i_lw(p, tmp, 0, ptep); ptep 1070 arch/mips/mm/tlbex.c uasm_i_lw(p, tmp, pte_off_odd, ptep); /* odd pte */ ptep 1075 arch/mips/mm/tlbex.c uasm_i_lw(p, tmp, sizeof(pte_t), ptep); ptep 1082 arch/mips/mm/tlbex.c UASM_i_LW(p, tmp, pte_off_even, ptep); /* get even pte */ ptep 1083 arch/mips/mm/tlbex.c UASM_i_LW(p, ptep, pte_off_odd, ptep); /* get odd pte */ ptep 1090 arch/mips/mm/tlbex.c build_convert_pte_to_entrylo(p, ptep); ptep 1095 arch/mips/mm/tlbex.c UASM_i_MTC0(p, ptep, C0_ENTRYLO1); /* load it */ ptep 44 arch/nds32/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmdp, pte_t * ptep) ptep 46 arch/nds32/include/asm/pgalloc.h unsigned long pte_ptr = (unsigned long)ptep; ptep 60 arch/nds32/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t * pmdp, pgtable_t ptep) ptep 66 arch/nds32/include/asm/pgalloc.h pmdval = page_to_pfn(ptep) << PAGE_SHIFT | _PAGE_USER_TABLE; ptep 191 arch/nds32/include/asm/pgtable.h #define pte_clear(mm,addr,ptep) set_pte_at((mm),(addr),(ptep), __pte(0)) ptep 205 arch/nds32/include/asm/pgtable.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 224 arch/nds32/include/asm/pgtable.h static inline void set_pte(pte_t * ptep, pte_t pte) ptep 227 arch/nds32/include/asm/pgtable.h *ptep = pte; ptep 229 arch/nds32/include/asm/pgtable.h __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (ptep):"memory"); ptep 317 arch/nds32/include/asm/pgtable.h static inline pmd_t __mk_pmd(pte_t * ptep, unsigned long prot) ptep 319 arch/nds32/include/asm/pgtable.h unsigned long ptr = (unsigned long)ptep; ptep 40 arch/nds32/mm/highmem.c pte_t *ptep; ptep 52 arch/nds32/mm/highmem.c ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); ptep 53 arch/nds32/mm/highmem.c set_pte(ptep, pte); ptep 68 arch/nds32/mm/highmem.c pte_t *ptep; ptep 72 arch/nds32/mm/highmem.c ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); ptep 73 arch/nds32/mm/highmem.c set_pte(ptep, 0); ptep 20 arch/nds32/mm/proc.c pte_t *ptep, pte; ptep 24 arch/nds32/mm/proc.c ptep = pte_offset_map(pmd, addr); ptep 25 arch/nds32/mm/proc.c pte = *ptep; ptep 37 arch/nds32/mm/proc.c pte_t *ptep, pte; ptep 45 arch/nds32/mm/proc.c ptep = pte_offset_map(pmd, addr); ptep 46 arch/nds32/mm/proc.c pte = *ptep; ptep 205 arch/nios2/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 207 arch/nios2/include/asm/pgtable.h *ptep = pteval; ptep 211 arch/nios2/include/asm/pgtable.h pte_t *ptep, pte_t pteval) ptep 216 arch/nios2/include/asm/pgtable.h set_pte(ptep, pteval); ptep 228 arch/nios2/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 234 arch/nios2/include/asm/pgtable.h set_pte_at(mm, addr, ptep, null); ptep 201 arch/nios2/mm/cacheflush.c unsigned long address, pte_t *ptep) ptep 203 arch/nios2/mm/cacheflush.c pte_t pte = *ptep; ptep 50 arch/openrisc/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pteval) set_pte(ptep, pteval) ptep 363 arch/openrisc/include/asm/pgtable.h static inline void pmd_set(pmd_t *pmdp, pte_t *ptep) ptep 365 arch/openrisc/include/asm/pgtable.h pmd_val(*pmdp) = _KERNPG_TABLE | (unsigned long) ptep; ptep 9 arch/parisc/include/asm/hugetlb.h pte_t *ptep, pte_t pte); ptep 13 arch/parisc/include/asm/hugetlb.h pte_t *ptep); ptep 38 arch/parisc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 44 arch/parisc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep); ptep 48 arch/parisc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep, ptep 89 arch/parisc/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pteval) \ ptep 94 arch/parisc/include/asm/pgtable.h old_pte = *ptep; \ ptep 95 arch/parisc/include/asm/pgtable.h set_pte(ptep, pteval); \ ptep 498 arch/parisc/include/asm/pgtable.h static inline int ptep_test_and_clear_young(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) ptep 503 arch/parisc/include/asm/pgtable.h if (!pte_young(*ptep)) ptep 507 arch/parisc/include/asm/pgtable.h pte = *ptep; ptep 512 arch/parisc/include/asm/pgtable.h set_pte(ptep, pte_mkold(pte)); ptep 519 arch/parisc/include/asm/pgtable.h static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 525 arch/parisc/include/asm/pgtable.h old_pte = *ptep; ptep 526 arch/parisc/include/asm/pgtable.h set_pte(ptep, __pte(0)); ptep 533 arch/parisc/include/asm/pgtable.h static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 537 arch/parisc/include/asm/pgtable.h set_pte(ptep, pte_wrprotect(*ptep)); ptep 87 arch/parisc/kernel/cache.c update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) ptep 89 arch/parisc/kernel/cache.c unsigned long pfn = pte_pfn(*ptep); ptep 534 arch/parisc/kernel/cache.c pte_t *ptep = NULL; ptep 541 arch/parisc/kernel/cache.c ptep = pte_offset_map(pmd, addr); ptep 544 arch/parisc/kernel/cache.c return ptep; ptep 579 arch/parisc/kernel/cache.c pte_t *ptep = get_ptep(pgd, addr); ptep 580 arch/parisc/kernel/cache.c if (!ptep) ptep 582 arch/parisc/kernel/cache.c pfn = pte_pfn(*ptep); ptep 620 arch/parisc/kernel/cache.c pte_t *ptep = get_ptep(pgd, addr); ptep 621 arch/parisc/kernel/cache.c if (!ptep) ptep 623 arch/parisc/kernel/cache.c pfn = pte_pfn(*ptep); ptep 118 arch/parisc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) ptep 127 arch/parisc/mm/hugetlbpage.c set_pte(ptep, entry); ptep 128 arch/parisc/mm/hugetlbpage.c ptep++; ptep 138 arch/parisc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) ptep 143 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, entry); ptep 149 arch/parisc/mm/hugetlbpage.c pte_t *ptep) ptep 155 arch/parisc/mm/hugetlbpage.c entry = *ptep; ptep 156 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, __pte(0)); ptep 164 arch/parisc/mm/hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 170 arch/parisc/mm/hugetlbpage.c old_pte = *ptep; ptep 171 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, pte_wrprotect(old_pte)); ptep 176 arch/parisc/mm/hugetlbpage.c unsigned long addr, pte_t *ptep, ptep 184 arch/parisc/mm/hugetlbpage.c changed = !pte_same(*ptep, pte); ptep 186 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, pte); ptep 215 arch/powerpc/include/asm/book3s/32/pgtable.h #define pte_clear(mm, addr, ptep) \ ptep 216 arch/powerpc/include/asm/book3s/32/pgtable.h do { pte_update(ptep, ~_PAGE_HASHPTE, 0); } while (0) ptep 239 arch/powerpc/include/asm/book3s/32/pgtable.h extern void flush_hash_entry(struct mm_struct *mm, pte_t *ptep, ptep 304 arch/powerpc/include/asm/book3s/32/pgtable.h static inline int __ptep_test_and_clear_young(unsigned int context, unsigned long addr, pte_t *ptep) ptep 307 arch/powerpc/include/asm/book3s/32/pgtable.h old = pte_update(ptep, _PAGE_ACCESSED, 0); ptep 309 arch/powerpc/include/asm/book3s/32/pgtable.h unsigned long ptephys = __pa(ptep) & PAGE_MASK; ptep 319 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep) ptep 321 arch/powerpc/include/asm/book3s/32/pgtable.h return __pte(pte_update(ptep, ~_PAGE_HASHPTE, 0)); ptep 326 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep) ptep 328 arch/powerpc/include/asm/book3s/32/pgtable.h pte_update(ptep, _PAGE_RW, 0); ptep 332 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep, pte_t entry, ptep 339 arch/powerpc/include/asm/book3s/32/pgtable.h pte_update(ptep, 0, set); ptep 526 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep, pte_t pte, int percpu) ptep 536 arch/powerpc/include/asm/book3s/32/pgtable.h *ptep = __pte((pte_val(*ptep) & _PAGE_HASHPTE) ptep 539 arch/powerpc/include/asm/book3s/32/pgtable.h pte_update(ptep, ~_PAGE_HASHPTE, pte_val(pte)); ptep 551 arch/powerpc/include/asm/book3s/32/pgtable.h *ptep = __pte((pte_val(*ptep) & _PAGE_HASHPTE) ptep 555 arch/powerpc/include/asm/book3s/32/pgtable.h if (pte_val(*ptep) & _PAGE_HASHPTE) ptep 556 arch/powerpc/include/asm/book3s/32/pgtable.h flush_hash_entry(mm, ptep, addr); ptep 561 arch/powerpc/include/asm/book3s/32/pgtable.h : "=m" (*ptep), "=m" (*((unsigned char *)ptep+4)) ptep 570 arch/powerpc/include/asm/book3s/32/pgtable.h *ptep = __pte((pte_val(*ptep) & _PAGE_HASHPTE) ptep 94 arch/powerpc/include/asm/book3s/64/hash-4k.h static inline unsigned long pte_set_hidx(pte_t *ptep, real_pte_t rpte, ptep 79 arch/powerpc/include/asm/book3s/64/hash-64k.h static inline real_pte_t __real_pte(pte_t pte, pte_t *ptep, int offset) ptep 93 arch/powerpc/include/asm/book3s/64/hash-64k.h hidxp = (unsigned long *)(ptep + offset); ptep 119 arch/powerpc/include/asm/book3s/64/hash-64k.h static inline unsigned long pte_set_hidx(pte_t *ptep, real_pte_t rpte, ptep 123 arch/powerpc/include/asm/book3s/64/hash-64k.h unsigned long *hidxp = (unsigned long *)(ptep + offset); ptep 147 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, unsigned long pte, int huge); ptep 152 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, unsigned long clr, ptep 167 arch/powerpc/include/asm/book3s/64/hash.h : "=&r" (old_be), "=&r" (tmp_be), "=m" (*ptep) ptep 168 arch/powerpc/include/asm/book3s/64/hash.h : "r" (ptep), "r" (cpu_to_be64(clr)), "m" (*ptep), ptep 177 arch/powerpc/include/asm/book3s/64/hash.h hpte_need_flush(mm, addr, ptep, old, huge); ptep 185 arch/powerpc/include/asm/book3s/64/hash.h static inline void hash__ptep_set_access_flags(pte_t *ptep, pte_t entry) ptep 201 arch/powerpc/include/asm/book3s/64/hash.h :"=&r" (old), "=&r" (tmp), "=m" (*ptep) ptep 202 arch/powerpc/include/asm/book3s/64/hash.h :"r" (val), "r" (ptep), "m" (*ptep), "r" (cpu_to_be64(H_PAGE_BUSY)) ptep 225 arch/powerpc/include/asm/book3s/64/hash.h pte_t *ptep, pte_t pte, int percpu) ptep 231 arch/powerpc/include/asm/book3s/64/hash.h *ptep = pte; ptep 17 arch/powerpc/include/asm/book3s/64/hugetlb.h unsigned long addr, pte_t *ptep, ptep 58 arch/powerpc/include/asm/book3s/64/hugetlb.h unsigned long addr, pte_t *ptep); ptep 62 arch/powerpc/include/asm/book3s/64/hugetlb.h unsigned long addr, pte_t *ptep, ptep 457 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned long vsid, pte_t *ptep, unsigned long trap, ptep 460 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned long vsid, pte_t *ptep, unsigned long trap, ptep 470 arch/powerpc/include/asm/book3s/64/mmu-hash.h pte_t *ptep, unsigned long trap, unsigned long flags, ptep 354 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, unsigned long clr, ptep 358 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__pte_update(mm, addr, ptep, clr, set, huge); ptep 359 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__pte_update(mm, addr, ptep, clr, set, huge); ptep 372 arch/powerpc/include/asm/book3s/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 376 arch/powerpc/include/asm/book3s/64/pgtable.h if ((pte_raw(*ptep) & cpu_to_be64(_PAGE_ACCESSED | H_PAGE_HASHPTE)) == 0) ptep 378 arch/powerpc/include/asm/book3s/64/pgtable.h old = pte_update(mm, addr, ptep, _PAGE_ACCESSED, 0, 0); ptep 428 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep) ptep 430 arch/powerpc/include/asm/book3s/64/pgtable.h if (__pte_write(*ptep)) ptep 431 arch/powerpc/include/asm/book3s/64/pgtable.h pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 0); ptep 432 arch/powerpc/include/asm/book3s/64/pgtable.h else if (unlikely(pte_savedwrite(*ptep))) ptep 433 arch/powerpc/include/asm/book3s/64/pgtable.h pte_update(mm, addr, ptep, 0, _PAGE_PRIVILEGED, 0); ptep 438 arch/powerpc/include/asm/book3s/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 444 arch/powerpc/include/asm/book3s/64/pgtable.h if (__pte_write(*ptep)) ptep 445 arch/powerpc/include/asm/book3s/64/pgtable.h pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 1); ptep 446 arch/powerpc/include/asm/book3s/64/pgtable.h else if (unlikely(pte_savedwrite(*ptep))) ptep 447 arch/powerpc/include/asm/book3s/64/pgtable.h pte_update(mm, addr, ptep, 0, _PAGE_PRIVILEGED, 1); ptep 452 arch/powerpc/include/asm/book3s/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 454 arch/powerpc/include/asm/book3s/64/pgtable.h unsigned long old = pte_update(mm, addr, ptep, ~0UL, 0, 0); ptep 461 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, int full) ptep 468 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__ptep_get_and_clear_full(mm, addr, ptep, full); ptep 470 arch/powerpc/include/asm/book3s/64/pgtable.h return ptep_get_and_clear(mm, addr, ptep); ptep 475 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t * ptep) ptep 477 arch/powerpc/include/asm/book3s/64/pgtable.h pte_update(mm, addr, ptep, ~0UL, 0, 0); ptep 790 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, pte_t entry, ptep 795 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__ptep_set_access_flags(vma, ptep, entry, ptep 797 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__ptep_set_access_flags(ptep, entry); ptep 816 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, pte_t pte, int percpu) ptep 819 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__set_pte_at(mm, addr, ptep, pte, percpu); ptep 820 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__set_pte_at(mm, addr, ptep, pte, percpu); ptep 124 arch/powerpc/include/asm/book3s/64/radix.h extern void radix__ptep_set_access_flags(struct vm_area_struct *vma, pte_t *ptep, ptep 129 arch/powerpc/include/asm/book3s/64/radix.h unsigned long addr, pte_t *ptep, ptep 132 arch/powerpc/include/asm/book3s/64/radix.h static inline unsigned long __radix_pte_update(pte_t *ptep, unsigned long clr, ptep 143 arch/powerpc/include/asm/book3s/64/radix.h : "=&r" (old_be), "=&r" (tmp_be), "=m" (*ptep) ptep 144 arch/powerpc/include/asm/book3s/64/radix.h : "r" (ptep), "r" (cpu_to_be64(set)), "r" (cpu_to_be64(clr)) ptep 152 arch/powerpc/include/asm/book3s/64/radix.h pte_t *ptep, unsigned long clr, ptep 158 arch/powerpc/include/asm/book3s/64/radix.h old_pte = __radix_pte_update(ptep, clr, set); ptep 167 arch/powerpc/include/asm/book3s/64/radix.h pte_t *ptep, int full) ptep 172 arch/powerpc/include/asm/book3s/64/radix.h old_pte = pte_val(*ptep); ptep 173 arch/powerpc/include/asm/book3s/64/radix.h *ptep = __pte(0); ptep 175 arch/powerpc/include/asm/book3s/64/radix.h old_pte = radix__pte_update(mm, addr, ptep, ~0ul, 0, 0); ptep 191 arch/powerpc/include/asm/book3s/64/radix.h pte_t *ptep, pte_t pte, int percpu) ptep 193 arch/powerpc/include/asm/book3s/64/radix.h *ptep = pte; ptep 16 arch/powerpc/include/asm/book3s/pgtable.h extern void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, ptep 22 arch/powerpc/include/asm/book3s/pgtable.h pte_t *ptep, pte_t entry, int dirty); ptep 38 arch/powerpc/include/asm/book3s/pgtable.h void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep); ptep 41 arch/powerpc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 44 arch/powerpc/include/asm/hugetlb.h return __pte(pte_update(mm, addr, ptep, ~0UL, 0, 1)); ptep 46 arch/powerpc/include/asm/hugetlb.h return __pte(pte_update(ptep, ~0UL, 0)); ptep 52 arch/powerpc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 54 arch/powerpc/include/asm/hugetlb.h huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); ptep 60 arch/powerpc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep, ptep 420 arch/powerpc/include/asm/kvm_book3s_64.h static inline pte_t kvmppc_read_update_linux_pte(pte_t *ptep, int writing) ptep 428 arch/powerpc/include/asm/kvm_book3s_64.h old_pte = READ_ONCE(*ptep); ptep 444 arch/powerpc/include/asm/kvm_book3s_64.h if (pte_xchg(ptep, old_pte, new_pte)) ptep 163 arch/powerpc/include/asm/nohash/32/pgtable.h #define pte_clear(mm, addr, ptep) \ ptep 164 arch/powerpc/include/asm/nohash/32/pgtable.h do { pte_update(ptep, ~0, 0); } while (0) ptep 289 arch/powerpc/include/asm/nohash/32/pgtable.h static inline int __ptep_test_and_clear_young(unsigned int context, unsigned long addr, pte_t *ptep) ptep 292 arch/powerpc/include/asm/nohash/32/pgtable.h old = pte_update(ptep, _PAGE_ACCESSED, 0); ptep 300 arch/powerpc/include/asm/nohash/32/pgtable.h pte_t *ptep) ptep 302 arch/powerpc/include/asm/nohash/32/pgtable.h return __pte(pte_update(ptep, ~0, 0)); ptep 307 arch/powerpc/include/asm/nohash/32/pgtable.h pte_t *ptep) ptep 312 arch/powerpc/include/asm/nohash/32/pgtable.h pte_update(ptep, clr, set); ptep 316 arch/powerpc/include/asm/nohash/32/pgtable.h pte_t *ptep, pte_t entry, ptep 325 arch/powerpc/include/asm/nohash/32/pgtable.h pte_update(ptep, clr, set); ptep 210 arch/powerpc/include/asm/nohash/64/pgtable.h pte_t *ptep, unsigned long clr, ptep 223 arch/powerpc/include/asm/nohash/64/pgtable.h : "=&r" (old), "=&r" (tmp), "=m" (*ptep) ptep 224 arch/powerpc/include/asm/nohash/64/pgtable.h : "r" (ptep), "r" (clr), "m" (*ptep), "r" (set) ptep 227 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long old = pte_val(*ptep); ptep 228 arch/powerpc/include/asm/nohash/64/pgtable.h *ptep = __pte((old & ~clr) | set); ptep 243 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 247 arch/powerpc/include/asm/nohash/64/pgtable.h if (pte_young(*ptep)) ptep 249 arch/powerpc/include/asm/nohash/64/pgtable.h old = pte_update(mm, addr, ptep, _PAGE_ACCESSED, 0, 0); ptep 262 arch/powerpc/include/asm/nohash/64/pgtable.h pte_t *ptep) ptep 265 arch/powerpc/include/asm/nohash/64/pgtable.h if ((pte_val(*ptep) & _PAGE_RW) == 0) ptep 268 arch/powerpc/include/asm/nohash/64/pgtable.h pte_update(mm, addr, ptep, _PAGE_RW, 0, 0); ptep 273 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 275 arch/powerpc/include/asm/nohash/64/pgtable.h if ((pte_val(*ptep) & _PAGE_RW) == 0) ptep 278 arch/powerpc/include/asm/nohash/64/pgtable.h pte_update(mm, addr, ptep, _PAGE_RW, 0, 1); ptep 291 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long addr, pte_t *ptep) ptep 293 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long old = pte_update(mm, addr, ptep, ~0UL, 0, 0); ptep 298 arch/powerpc/include/asm/nohash/64/pgtable.h pte_t * ptep) ptep 300 arch/powerpc/include/asm/nohash/64/pgtable.h pte_update(mm, addr, ptep, ~0UL, 0, 0); ptep 306 arch/powerpc/include/asm/nohash/64/pgtable.h pte_t *ptep, pte_t entry, ptep 321 arch/powerpc/include/asm/nohash/64/pgtable.h :"=&r" (old), "=&r" (tmp), "=m" (*ptep) ptep 322 arch/powerpc/include/asm/nohash/64/pgtable.h :"r" (bits), "r" (ptep), "m" (*ptep) ptep 325 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long old = pte_val(*ptep); ptep 326 arch/powerpc/include/asm/nohash/64/pgtable.h *ptep = __pte(old | bits); ptep 184 arch/powerpc/include/asm/nohash/pgtable.h extern void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, ptep 193 arch/powerpc/include/asm/nohash/pgtable.h pte_t *ptep, pte_t pte, int percpu) ptep 205 arch/powerpc/include/asm/nohash/pgtable.h : "=m" (*ptep), "=m" (*((unsigned char *)ptep+4)) ptep 213 arch/powerpc/include/asm/nohash/pgtable.h ptep->pte = ptep->pte1 = ptep->pte2 = ptep->pte3 = pte_val(pte); ptep 215 arch/powerpc/include/asm/nohash/pgtable.h *ptep = pte; ptep 231 arch/powerpc/include/asm/nohash/pgtable.h pte_t *ptep, pte_t entry, int dirty); ptep 303 arch/powerpc/include/asm/nohash/pgtable.h void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep); ptep 306 arch/powerpc/include/asm/nohash/pgtable.h void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) {} ptep 81 arch/powerpc/include/asm/pgtable-be-types.h static inline bool pte_xchg(pte_t *ptep, pte_t old, pte_t new) ptep 83 arch/powerpc/include/asm/pgtable-be-types.h unsigned long *p = (unsigned long *)ptep; ptep 61 arch/powerpc/include/asm/pgtable-types.h static inline bool pte_xchg(pte_t *ptep, pte_t old, pte_t new) ptep 63 arch/powerpc/include/asm/pgtable-types.h unsigned long *p = (unsigned long *)ptep; ptep 44 arch/powerpc/include/asm/tlb.h extern void flush_hash_entry(struct mm_struct *mm, pte_t *ptep, ptep 47 arch/powerpc/include/asm/tlb.h static inline void __tlb_remove_tlb_entry(struct mmu_gather *tlb, pte_t *ptep, ptep 51 arch/powerpc/include/asm/tlb.h if (pte_val(*ptep) & _PAGE_HASHPTE) ptep 52 arch/powerpc/include/asm/tlb.h flush_hash_entry(tlb->mm, ptep, address); ptep 355 arch/powerpc/kernel/eeh.c pte_t *ptep; ptep 364 arch/powerpc/kernel/eeh.c ptep = find_init_mm_pte(token, &hugepage_shift); ptep 365 arch/powerpc/kernel/eeh.c if (!ptep) ptep 368 arch/powerpc/kernel/eeh.c pa = pte_pfn(*ptep); ptep 68 arch/powerpc/kernel/io-workarounds.c pte_t *ptep; ptep 77 arch/powerpc/kernel/io-workarounds.c ptep = find_init_mm_pte(vaddr, &hugepage_shift); ptep 78 arch/powerpc/kernel/io-workarounds.c if (ptep == NULL) ptep 82 arch/powerpc/kernel/io-workarounds.c paddr = pte_pfn(*ptep) << PAGE_SHIFT; ptep 30 arch/powerpc/kernel/mce_power.c pte_t *ptep; ptep 41 arch/powerpc/kernel/mce_power.c ptep = __find_linux_pte(mm->pgd, addr, NULL, &shift); ptep 43 arch/powerpc/kernel/mce_power.c if (!ptep || pte_special(*ptep)) { ptep 49 arch/powerpc/kernel/mce_power.c pfn = pte_pfn(*ptep); ptep 52 arch/powerpc/kernel/mce_power.c pfn = pte_pfn(__pte(pte_val(*ptep) | (addr & rpnmask))); ptep 617 arch/powerpc/kvm/book3s_64_mmu_hv.c pte_t *ptep, pte; ptep 624 arch/powerpc/kvm/book3s_64_mmu_hv.c ptep = find_current_mm_pte(current->mm->pgd, ptep 626 arch/powerpc/kvm/book3s_64_mmu_hv.c if (ptep) { ptep 627 arch/powerpc/kvm/book3s_64_mmu_hv.c pte = kvmppc_read_update_linux_pte(ptep, 1); ptep 338 arch/powerpc/kvm/book3s_64_mmu_radix.c static unsigned long kvmppc_radix_update_pte(struct kvm *kvm, pte_t *ptep, ptep 342 arch/powerpc/kvm/book3s_64_mmu_radix.c return __radix_pte_update(ptep, clr, set); ptep 346 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep, pte_t pte) ptep 348 arch/powerpc/kvm/book3s_64_mmu_radix.c radix__set_pte_at(kvm->mm, addr, ptep, pte, 0); ptep 359 arch/powerpc/kvm/book3s_64_mmu_radix.c static void kvmppc_pte_free(pte_t *ptep) ptep 361 arch/powerpc/kvm/book3s_64_mmu_radix.c kmem_cache_free(kvm_pte_cache, ptep); ptep 571 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep, *new_ptep = NULL; ptep 708 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = pte_offset_kernel(pmd, gpa); ptep 709 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte_present(*ptep)) { ptep 711 arch/powerpc/kvm/book3s_64_mmu_radix.c if (pte_raw(*ptep) == pte_raw(pte)) { ptep 716 arch/powerpc/kvm/book3s_64_mmu_radix.c WARN_ON_ONCE((pte_val(*ptep) ^ pte_val(pte)) & ptep 718 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_update_pte(kvm, ptep, 0, pte_val(pte), gpa, 0); ptep 722 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_set_pte_at(kvm, gpa, ptep, pte); ptep 743 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 758 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(pgtable, gpa, NULL, &shift); ptep 759 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep) && (!writing || pte_write(*ptep))) { ptep 760 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_radix_update_pte(kvm, ptep, 0, pgflags, gpa, shift); ptep 778 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t pte, *ptep; ptep 817 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift); ptep 822 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!ptep) { ptep 828 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = *ptep; ptep 971 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 975 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift); ptep 976 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep)) ptep 977 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_unmap_pte(kvm, ptep, gpa, shift, memslot, ptep 986 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 992 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift); ptep 993 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep) && pte_young(*ptep)) { ptep 994 arch/powerpc/kvm/book3s_64_mmu_radix.c old = kvmppc_radix_update_pte(kvm, ptep, _PAGE_ACCESSED, 0, ptep 1011 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 1016 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift); ptep 1017 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep) && pte_young(*ptep)) ptep 1028 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 1033 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift); ptep 1034 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep) && pte_dirty(*ptep)) { ptep 1039 arch/powerpc/kvm/book3s_64_mmu_radix.c old = kvmppc_radix_update_pte(kvm, ptep, _PAGE_DIRTY, 0, ptep 1081 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 1088 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift); ptep 1089 arch/powerpc/kvm/book3s_64_mmu_radix.c if (ptep && pte_present(*ptep)) ptep 1090 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmppc_unmap_pte(kvm, ptep, gpa, shift, memslot, ptep 1202 arch/powerpc/kvm/book3s_64_mmu_radix.c pte_t *ptep; ptep 1302 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = pte_offset_kernel(&pmd, gpa); ptep 1303 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = pte_val(READ_ONCE(*ptep)); ptep 444 arch/powerpc/kvm/book3s_64_vio_hv.c pte_t *ptep, pte; ptep 456 arch/powerpc/kvm/book3s_64_vio_hv.c ptep = __find_linux_pte(vcpu->arch.pgdir, ua, NULL, &shift); ptep 457 arch/powerpc/kvm/book3s_64_vio_hv.c if (!ptep || !pte_present(*ptep)) ptep 459 arch/powerpc/kvm/book3s_64_vio_hv.c pte = *ptep; ptep 798 arch/powerpc/kvm/book3s_hv_nested.c pte_t *ptep; ptep 807 arch/powerpc/kvm/book3s_hv_nested.c ptep = __find_linux_pte(gp->shadow_pgtable, gpa, NULL, &shift); ptep 814 arch/powerpc/kvm/book3s_hv_nested.c if (ptep && pte_present(*ptep) && ((pte_val(*ptep) & mask) == hpa)) { ptep 815 arch/powerpc/kvm/book3s_hv_nested.c __radix_pte_update(ptep, clr, set); ptep 848 arch/powerpc/kvm/book3s_hv_nested.c pte_t *ptep; ptep 857 arch/powerpc/kvm/book3s_hv_nested.c ptep = __find_linux_pte(gp->shadow_pgtable, gpa, NULL, &shift); ptep 859 arch/powerpc/kvm/book3s_hv_nested.c if (ptep && pte_present(*ptep) && ((pte_val(*ptep) & mask) == hpa)) ptep 860 arch/powerpc/kvm/book3s_hv_nested.c kvmppc_unmap_pte(kvm, ptep, gpa, shift, NULL, gp->shadow_lpid); ptep 920 arch/powerpc/kvm/book3s_hv_nested.c pte_t *ptep; ptep 924 arch/powerpc/kvm/book3s_hv_nested.c ptep = __find_linux_pte(gp->shadow_pgtable, gpa, NULL, &shift); ptep 927 arch/powerpc/kvm/book3s_hv_nested.c if (ptep && pte_present(*ptep)) { ptep 928 arch/powerpc/kvm/book3s_hv_nested.c kvmppc_unmap_pte(kvm, ptep, gpa, shift, NULL, gp->shadow_lpid); ptep 210 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte_t *ptep; ptep 261 arch/powerpc/kvm/book3s_hv_rm_mmu.c ptep = __find_linux_pte(pgdir, hva, NULL, &hpage_shift); ptep 262 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (ptep) { ptep 279 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = kvmppc_read_update_linux_pte(ptep, writing); ptep 304 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (ptep && !hpte_cache_flags_ok(ptel, is_ci)) { ptep 899 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte_t *ptep, pte; ptep 911 arch/powerpc/kvm/book3s_hv_rm_mmu.c ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift); ptep 912 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (!ptep) ptep 914 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = kvmppc_read_update_linux_pte(ptep, writing); ptep 336 arch/powerpc/kvm/e500_mmu_host.c pte_t *ptep; ptep 477 arch/powerpc/kvm/e500_mmu_host.c ptep = find_linux_pte(pgdir, hva, NULL, NULL); ptep 478 arch/powerpc/kvm/e500_mmu_host.c if (ptep) { ptep 479 arch/powerpc/kvm/e500_mmu_host.c pte_t pte = READ_ONCE(*ptep); ptep 107 arch/powerpc/lib/code-patching.c pte_t *ptep; ptep 124 arch/powerpc/lib/code-patching.c ptep = pte_offset_kernel(pmdp, addr); ptep 125 arch/powerpc/lib/code-patching.c if (unlikely(!ptep)) ptep 128 arch/powerpc/lib/code-patching.c pr_devel("clearing mm %p, pte %p, addr %lx\n", &init_mm, ptep, addr); ptep 133 arch/powerpc/lib/code-patching.c pte_clear(&init_mm, addr, ptep); ptep 320 arch/powerpc/mm/book3s32/mmu.c pte_t *ptep) ptep 330 arch/powerpc/mm/book3s32/mmu.c if (!pte_young(*ptep) || address >= TASK_SIZE) ptep 35 arch/powerpc/mm/book3s32/tlb.c void flush_hash_entry(struct mm_struct *mm, pte_t *ptep, unsigned long addr) ptep 40 arch/powerpc/mm/book3s32/tlb.c ptephys = __pa(ptep) & PAGE_MASK; ptep 20 arch/powerpc/mm/book3s64/hash_4k.c pte_t *ptep, unsigned long trap, unsigned long flags, ptep 34 arch/powerpc/mm/book3s64/hash_4k.c pte_t pte = READ_ONCE(*ptep); ptep 51 arch/powerpc/mm/book3s64/hash_4k.c } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); ptep 58 arch/powerpc/mm/book3s64/hash_4k.c rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); ptep 114 arch/powerpc/mm/book3s64/hash_4k.c *ptep = __pte(old_pte); ptep 120 arch/powerpc/mm/book3s64/hash_4k.c new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); ptep 122 arch/powerpc/mm/book3s64/hash_4k.c *ptep = __pte(new_pte & ~H_PAGE_BUSY); ptep 37 arch/powerpc/mm/book3s64/hash_64k.c pte_t *ptep, unsigned long trap, unsigned long flags, ptep 52 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); ptep 69 arch/powerpc/mm/book3s64/hash_64k.c } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); ptep 89 arch/powerpc/mm/book3s64/hash_64k.c rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); ptep 129 arch/powerpc/mm/book3s64/hash_64k.c *ptep = __pte(new_pte & ~H_PAGE_BUSY); ptep 210 arch/powerpc/mm/book3s64/hash_64k.c *ptep = __pte(old_pte); ptep 216 arch/powerpc/mm/book3s64/hash_64k.c new_pte |= pte_set_hidx(ptep, rpte, subpg_index, slot, PTRS_PER_PTE); ptep 219 arch/powerpc/mm/book3s64/hash_64k.c *ptep = __pte(new_pte & ~H_PAGE_BUSY); ptep 224 arch/powerpc/mm/book3s64/hash_64k.c unsigned long vsid, pte_t *ptep, unsigned long trap, ptep 238 arch/powerpc/mm/book3s64/hash_64k.c pte_t pte = READ_ONCE(*ptep); ptep 261 arch/powerpc/mm/book3s64/hash_64k.c } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); ptep 264 arch/powerpc/mm/book3s64/hash_64k.c rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); ptep 322 arch/powerpc/mm/book3s64/hash_64k.c *ptep = __pte(old_pte); ptep 329 arch/powerpc/mm/book3s64/hash_64k.c new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); ptep 331 arch/powerpc/mm/book3s64/hash_64k.c *ptep = __pte(new_pte & ~H_PAGE_BUSY); ptep 26 arch/powerpc/mm/book3s64/hash_hugetlbpage.c pte_t *ptep, unsigned long trap, unsigned long flags, ptep 54 arch/powerpc/mm/book3s64/hash_hugetlbpage.c old_pte = pte_val(*ptep); ptep 69 arch/powerpc/mm/book3s64/hash_hugetlbpage.c } while(!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); ptep 80 arch/powerpc/mm/book3s64/hash_hugetlbpage.c rpte = __real_pte(__pte(old_pte), ptep, offset); ptep 116 arch/powerpc/mm/book3s64/hash_hugetlbpage.c *ptep = __pte(old_pte); ptep 122 arch/powerpc/mm/book3s64/hash_hugetlbpage.c new_pte |= pte_set_hidx(ptep, rpte, 0, slot, offset); ptep 128 arch/powerpc/mm/book3s64/hash_hugetlbpage.c *ptep = __pte(new_pte & ~H_PAGE_BUSY); ptep 133 arch/powerpc/mm/book3s64/hash_hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 141 arch/powerpc/mm/book3s64/hash_hugetlbpage.c pte_val = pte_update(vma->vm_mm, addr, ptep, ptep 148 arch/powerpc/mm/book3s64/hash_hugetlbpage.c pte_t *ptep, pte_t old_pte, pte_t pte) ptep 152 arch/powerpc/mm/book3s64/hash_hugetlbpage.c return radix__huge_ptep_modify_prot_commit(vma, addr, ptep, ptep 154 arch/powerpc/mm/book3s64/hash_hugetlbpage.c set_huge_pte_at(vma->vm_mm, addr, ptep, pte); ptep 153 arch/powerpc/mm/book3s64/hash_pgtable.c pte_t *ptep; ptep 164 arch/powerpc/mm/book3s64/hash_pgtable.c ptep = pte_alloc_kernel(pmdp, ea); ptep 165 arch/powerpc/mm/book3s64/hash_pgtable.c if (!ptep) ptep 167 arch/powerpc/mm/book3s64/hash_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); ptep 42 arch/powerpc/mm/book3s64/hash_tlb.c pte_t *ptep, unsigned long pte, int huge) ptep 98 arch/powerpc/mm/book3s64/hash_tlb.c rpte = __real_pte(__pte(pte), ptep, offset); ptep 215 arch/powerpc/mm/book3s64/hash_tlb.c pte_t *ptep = find_current_mm_pte(mm->pgd, start, &is_thp, ptep 219 arch/powerpc/mm/book3s64/hash_tlb.c if (ptep == NULL) ptep 221 arch/powerpc/mm/book3s64/hash_tlb.c pte = pte_val(*ptep); ptep 227 arch/powerpc/mm/book3s64/hash_tlb.c hpte_do_hugepage_flush(mm, start, (pmd_t *)ptep, pte); ptep 229 arch/powerpc/mm/book3s64/hash_tlb.c hpte_need_flush(mm, start, ptep, pte, hugepage_shift); ptep 1251 arch/powerpc/mm/book3s64/hash_utils.c pte_t *ptep; ptep 1325 arch/powerpc/mm/book3s64/hash_utils.c ptep = find_linux_pte(pgdir, ea, &is_thp, &hugeshift); ptep 1326 arch/powerpc/mm/book3s64/hash_utils.c if (ptep == NULL || !pte_present(*ptep)) { ptep 1339 arch/powerpc/mm/book3s64/hash_utils.c if (!check_pte_access(access, pte_val(*ptep))) { ptep 1347 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_thp(ea, access, vsid, (pmd_t *)ptep, ptep 1351 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_huge(ea, access, vsid, ptep, trap, ptep 1370 arch/powerpc/mm/book3s64/hash_utils.c DBG_LOW(" i-pte: %016lx\n", pte_val(*ptep)); ptep 1372 arch/powerpc/mm/book3s64/hash_utils.c DBG_LOW(" i-pte: %016lx %016lx\n", pte_val(*ptep), ptep 1373 arch/powerpc/mm/book3s64/hash_utils.c pte_val(*(ptep + PTRS_PER_PTE))); ptep 1378 arch/powerpc/mm/book3s64/hash_utils.c if ((pte_val(*ptep) & H_PAGE_4K_PFN) && psize == MMU_PAGE_64K) { ptep 1387 arch/powerpc/mm/book3s64/hash_utils.c if (mmu_ci_restrictions && psize == MMU_PAGE_64K && pte_ci(*ptep)) { ptep 1412 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_64K(ea, access, vsid, ptep, trap, ptep 1421 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_4K(ea, access, vsid, ptep, trap, ptep 1431 arch/powerpc/mm/book3s64/hash_utils.c psize, pte_val(*ptep)); ptep 1433 arch/powerpc/mm/book3s64/hash_utils.c DBG_LOW(" o-pte: %016lx\n", pte_val(*ptep)); ptep 1435 arch/powerpc/mm/book3s64/hash_utils.c DBG_LOW(" o-pte: %016lx %016lx\n", pte_val(*ptep), ptep 1436 arch/powerpc/mm/book3s64/hash_utils.c pte_val(*(ptep + PTRS_PER_PTE))); ptep 1527 arch/powerpc/mm/book3s64/hash_utils.c pte_t *ptep; ptep 1560 arch/powerpc/mm/book3s64/hash_utils.c ptep = find_current_mm_pte(pgdir, ea, NULL, &hugepage_shift); ptep 1561 arch/powerpc/mm/book3s64/hash_utils.c if (!ptep) ptep 1572 arch/powerpc/mm/book3s64/hash_utils.c if ((pte_val(*ptep) & H_PAGE_4K_PFN) || pte_ci(*ptep)) ptep 1583 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_64K(ea, access, vsid, ptep, trap, ptep 1587 arch/powerpc/mm/book3s64/hash_utils.c rc = __hash_page_4K(ea, access, vsid, ptep, trap, update_flags, ptep 1597 arch/powerpc/mm/book3s64/hash_utils.c pte_val(*ptep)); ptep 1611 arch/powerpc/mm/book3s64/hash_utils.c pte_t *ptep) ptep 1626 arch/powerpc/mm/book3s64/hash_utils.c if (!pte_young(*ptep) || address >= TASK_SIZE) ptep 1660 arch/powerpc/mm/book3s64/hash_utils.c pte_t *ptep; ptep 1668 arch/powerpc/mm/book3s64/hash_utils.c ptep = find_linux_pte(mm->pgd, address, NULL, NULL); ptep 1669 arch/powerpc/mm/book3s64/hash_utils.c if (ptep) ptep 1670 arch/powerpc/mm/book3s64/hash_utils.c pkey = pte_to_pkey_bits(pte_val(READ_ONCE(*ptep))); ptep 421 arch/powerpc/mm/book3s64/pgtable.c pte_t *ptep) ptep 430 arch/powerpc/mm/book3s64/pgtable.c pte_val = pte_update(vma->vm_mm, addr, ptep, _PAGE_PRESENT, _PAGE_INVALID, 0); ptep 437 arch/powerpc/mm/book3s64/pgtable.c pte_t *ptep, pte_t old_pte, pte_t pte) ptep 441 arch/powerpc/mm/book3s64/pgtable.c ptep, old_pte, pte); ptep 442 arch/powerpc/mm/book3s64/pgtable.c set_pte_at(vma->vm_mm, addr, ptep, pte); ptep 96 arch/powerpc/mm/book3s64/radix_hugetlbpage.c unsigned long addr, pte_t *ptep, ptep 109 arch/powerpc/mm/book3s64/radix_hugetlbpage.c set_huge_pte_at(vma->vm_mm, addr, ptep, pte); ptep 68 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep; ptep 78 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = (pte_t *)pudp; ptep 88 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pmdp_ptep(pmdp); ptep 92 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = early_alloc_pgtable(PAGE_SIZE, nid, ptep 94 arch/powerpc/mm/book3s64/radix_pgtable.c pmd_populate_kernel(&init_mm, pmdp, ptep); ptep 96 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pte_offset_kernel(pmdp, ea); ptep 99 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pfn, flags)); ptep 118 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep; ptep 142 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = (pte_t *)pudp; ptep 149 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pmdp_ptep(pmdp); ptep 152 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pte_alloc_kernel(pmdp, ea); ptep 153 arch/powerpc/mm/book3s64/radix_pgtable.c if (!ptep) ptep 157 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pfn, flags)); ptep 177 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep; ptep 191 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = (pte_t *)pudp; ptep 198 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pmdp_ptep(pmdp); ptep 201 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pte_alloc_kernel(pmdp, idx); ptep 202 arch/powerpc/mm/book3s64/radix_pgtable.c if (!ptep) ptep 205 arch/powerpc/mm/book3s64/radix_pgtable.c radix__pte_update(&init_mm, idx, ptep, clear, 0, 0); ptep 986 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep; ptep 1001 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = (pte_t *) pgtable; ptep 1002 arch/powerpc/mm/book3s64/radix_pgtable.c *ptep = __pte(0); ptep 1003 arch/powerpc/mm/book3s64/radix_pgtable.c ptep++; ptep 1004 arch/powerpc/mm/book3s64/radix_pgtable.c *ptep = __pte(0); ptep 1032 arch/powerpc/mm/book3s64/radix_pgtable.c void radix__ptep_set_access_flags(struct vm_area_struct *vma, pte_t *ptep, ptep 1039 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long change = pte_val(entry) ^ pte_val(*ptep); ptep 1047 arch/powerpc/mm/book3s64/radix_pgtable.c old_pte = __radix_pte_update(ptep, _PAGE_PRESENT, _PAGE_INVALID); ptep 1053 arch/powerpc/mm/book3s64/radix_pgtable.c __radix_pte_update(ptep, _PAGE_INVALID, new_pte); ptep 1055 arch/powerpc/mm/book3s64/radix_pgtable.c __radix_pte_update(ptep, 0, set); ptep 1067 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long addr, pte_t *ptep, ptep 1081 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(mm, addr, ptep, pte); ptep 1102 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep = (pte_t *)pud; ptep 1108 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(&init_mm, 0 /* radix unused */, ptep, new_pud); ptep 1149 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t *ptep = (pte_t *)pmd; ptep 1155 arch/powerpc/mm/book3s64/radix_pgtable.c set_pte_at(&init_mm, 0 /* radix unused */, ptep, new_pmd); ptep 500 arch/powerpc/mm/hugetlbpage.c pte_t *ptep; ptep 515 arch/powerpc/mm/hugetlbpage.c ptep = hugepte_offset(hpd, address, pdshift); ptep 516 arch/powerpc/mm/hugetlbpage.c if (pte_present(*ptep)) { ptep 518 arch/powerpc/mm/hugetlbpage.c page = pte_page(*ptep); ptep 523 arch/powerpc/mm/hugetlbpage.c if (is_hugetlb_entry_migration(*ptep)) { ptep 525 arch/powerpc/mm/hugetlbpage.c __migration_entry_wait(mm, ptep, ptl); ptep 23 arch/powerpc/mm/kasan/kasan_init_32.c static void kasan_populate_pte(pte_t *ptep, pgprot_t prot) ptep 29 arch/powerpc/mm/kasan/kasan_init_32.c for (i = 0; i < PTRS_PER_PTE; i++, ptep++) ptep 30 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); ptep 122 arch/powerpc/mm/kasan/kasan_init_32.c pte_t *ptep = pte_offset_kernel(pmd, k_cur); ptep 124 arch/powerpc/mm/kasan/kasan_init_32.c if ((pte_val(*ptep) & PTE_RPN_MASK) != pa) ptep 127 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, k_cur, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); ptep 192 arch/powerpc/mm/nohash/book3e_hugetlbpage.c void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) ptep 195 arch/powerpc/mm/nohash/book3e_hugetlbpage.c book3e_hugetlb_preload(vma, address, *ptep); ptep 78 arch/powerpc/mm/nohash/book3e_pgtable.c pte_t *ptep; ptep 89 arch/powerpc/mm/nohash/book3e_pgtable.c ptep = pte_alloc_kernel(pmdp, ea); ptep 90 arch/powerpc/mm/nohash/book3e_pgtable.c if (!ptep) ptep 107 arch/powerpc/mm/nohash/book3e_pgtable.c ptep = early_alloc_pgtable(PAGE_SIZE); ptep 108 arch/powerpc/mm/nohash/book3e_pgtable.c pmd_populate_kernel(&init_mm, pmdp, ptep); ptep 110 arch/powerpc/mm/nohash/book3e_pgtable.c ptep = pte_offset_kernel(pmdp, ea); ptep 112 arch/powerpc/mm/nohash/book3e_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); ptep 179 arch/powerpc/mm/pgtable.c void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, ptep 186 arch/powerpc/mm/pgtable.c VM_WARN_ON(pte_hw_valid(*ptep) && !pte_protnone(*ptep)); ptep 198 arch/powerpc/mm/pgtable.c __set_pte_at(mm, addr, ptep, pte, 0); ptep 209 arch/powerpc/mm/pgtable.c pte_t *ptep, pte_t entry, int dirty) ptep 213 arch/powerpc/mm/pgtable.c changed = !pte_same(*(ptep), entry); ptep 216 arch/powerpc/mm/pgtable.c __ptep_set_access_flags(vma, ptep, entry, ptep 224 arch/powerpc/mm/pgtable.c unsigned long addr, pte_t *ptep, ptep 233 arch/powerpc/mm/pgtable.c ptep_set_access_flags(vma, addr, ptep, pte, dirty); ptep 239 arch/powerpc/mm/pgtable.c changed = !pte_same(*(ptep), pte); ptep 247 arch/powerpc/mm/pgtable.c assert_spin_locked(huge_pte_lockptr(h, vma->vm_mm, ptep)); ptep 257 arch/powerpc/mm/pgtable.c __ptep_set_access_flags(vma, ptep, pte, addr, psize); ptep 51 arch/powerpc/mm/pgtable_32.c pte_t *ptep = early_alloc_pgtable(PTE_FRAG_SIZE); ptep 53 arch/powerpc/mm/pgtable_32.c pmd_populate_kernel(&init_mm, pmdp, ptep); ptep 133 arch/powerpc/mm/pgtable_32.c get_pteptr(struct mm_struct *mm, unsigned long addr, pte_t **ptep, pmd_t **pmdp) ptep 150 arch/powerpc/mm/pgtable_32.c *ptep = pte; ptep 115 arch/powerpc/perf/callchain.c pte_t *ptep, pte; ptep 127 arch/powerpc/perf/callchain.c ptep = find_current_mm_pte(pgdir, addr, NULL, &shift); ptep 128 arch/powerpc/perf/callchain.c if (!ptep) ptep 136 arch/powerpc/perf/callchain.c pte = READ_ONCE(*ptep); ptep 3103 arch/powerpc/xmon/xmon.c static void format_pte(void *ptep, unsigned long pte) ptep 3107 arch/powerpc/xmon/xmon.c printf("ptep @ 0x%016lx = 0x%016lx\n", (unsigned long)ptep, pte); ptep 3126 arch/powerpc/xmon/xmon.c pte_t *ptep; ptep 3195 arch/powerpc/xmon/xmon.c ptep = pte_offset_map(pmdp, addr); ptep 3196 arch/powerpc/xmon/xmon.c if (pte_none(*ptep)) { ptep 3201 arch/powerpc/xmon/xmon.c format_pte(ptep, pte_val(*ptep)); ptep 303 arch/riscv/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 326 arch/riscv/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 328 arch/riscv/include/asm/pgtable.h *ptep = pteval; ptep 334 arch/riscv/include/asm/pgtable.h unsigned long addr, pte_t *ptep, pte_t pteval) ptep 339 arch/riscv/include/asm/pgtable.h set_pte(ptep, pteval); ptep 343 arch/riscv/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 345 arch/riscv/include/asm/pgtable.h set_pte_at(mm, addr, ptep, __pte(0)); ptep 350 arch/riscv/include/asm/pgtable.h unsigned long address, pte_t *ptep, ptep 353 arch/riscv/include/asm/pgtable.h if (!pte_same(*ptep, entry)) ptep 354 arch/riscv/include/asm/pgtable.h set_pte_at(vma->vm_mm, address, ptep, entry); ptep 364 arch/riscv/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 366 arch/riscv/include/asm/pgtable.h return __pte(atomic_long_xchg((atomic_long_t *)ptep, 0)); ptep 372 arch/riscv/include/asm/pgtable.h pte_t *ptep) ptep 374 arch/riscv/include/asm/pgtable.h if (!pte_young(*ptep)) ptep 376 arch/riscv/include/asm/pgtable.h return test_and_clear_bit(_PAGE_ACCESSED_OFFSET, &pte_val(*ptep)); ptep 381 arch/riscv/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 383 arch/riscv/include/asm/pgtable.h atomic_long_and(~(unsigned long)_PAGE_WRITE, (atomic_long_t *)ptep); ptep 388 arch/riscv/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 405 arch/riscv/include/asm/pgtable.h return ptep_test_and_clear_young(vma, address, ptep); ptep 164 arch/riscv/mm/init.c pte_t *ptep; ptep 168 arch/riscv/mm/init.c ptep = &fixmap_pte[pte_index(addr)]; ptep 171 arch/riscv/mm/init.c set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, prot)); ptep 173 arch/riscv/mm/init.c pte_clear(&init_mm, addr, ptep); ptep 199 arch/riscv/mm/init.c static void __init create_pte_mapping(pte_t *ptep, ptep 207 arch/riscv/mm/init.c if (pte_none(ptep[pte_index])) ptep 208 arch/riscv/mm/init.c ptep[pte_index] = pfn_pte(PFN_DOWN(pa), prot); ptep 249 arch/riscv/mm/init.c pte_t *ptep; ptep 262 arch/riscv/mm/init.c ptep = get_pte_virt(pte_phys); ptep 263 arch/riscv/mm/init.c memset(ptep, 0, PAGE_SIZE); ptep 266 arch/riscv/mm/init.c ptep = get_pte_virt(pte_phys); ptep 269 arch/riscv/mm/init.c create_pte_mapping(ptep, va, pa, sz, prot); ptep 19 arch/s390/include/asm/hugetlb.h pte_t *ptep, pte_t pte); ptep 20 arch/s390/include/asm/hugetlb.h pte_t huge_ptep_get(pte_t *ptep); ptep 22 arch/s390/include/asm/hugetlb.h unsigned long addr, pte_t *ptep); ptep 51 arch/s390/include/asm/hugetlb.h pte_t *ptep, unsigned long sz) ptep 53 arch/s390/include/asm/hugetlb.h if ((pte_val(*ptep) & _REGION_ENTRY_TYPE_MASK) == _REGION_ENTRY_TYPE_R3) ptep 54 arch/s390/include/asm/hugetlb.h pte_val(*ptep) = _REGION3_ENTRY_EMPTY; ptep 56 arch/s390/include/asm/hugetlb.h pte_val(*ptep) = _SEGMENT_ENTRY_EMPTY; ptep 60 arch/s390/include/asm/hugetlb.h unsigned long address, pte_t *ptep) ptep 62 arch/s390/include/asm/hugetlb.h huge_ptep_get_and_clear(vma->vm_mm, address, ptep); ptep 66 arch/s390/include/asm/hugetlb.h unsigned long addr, pte_t *ptep, ptep 69 arch/s390/include/asm/hugetlb.h int changed = !pte_same(huge_ptep_get(ptep), pte); ptep 71 arch/s390/include/asm/hugetlb.h huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); ptep 72 arch/s390/include/asm/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); ptep 78 arch/s390/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 80 arch/s390/include/asm/hugetlb.h pte_t pte = huge_ptep_get_and_clear(mm, addr, ptep); ptep 81 arch/s390/include/asm/hugetlb.h set_huge_pte_at(mm, addr, ptep, pte_wrprotect(pte)); ptep 48 arch/s390/include/asm/pgtable.h #define update_mmu_cache(vma, address, ptep) do { } while (0) ptep 49 arch/s390/include/asm/pgtable.h #define update_mmu_cache_pmd(vma, address, ptep) do { } while (0) ptep 913 arch/s390/include/asm/pgtable.h static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 915 arch/s390/include/asm/pgtable.h pte_val(*ptep) = _PAGE_INVALID; ptep 1006 arch/s390/include/asm/pgtable.h static __always_inline void __ptep_ipte(unsigned long address, pte_t *ptep, ptep 1010 arch/s390/include/asm/pgtable.h unsigned long pto = (unsigned long) ptep; ptep 1016 arch/s390/include/asm/pgtable.h : "+m" (*ptep) : [r1] "a" (pto), [r2] "a" (address), ptep 1030 arch/s390/include/asm/pgtable.h pte_t *ptep, int local) ptep 1032 arch/s390/include/asm/pgtable.h unsigned long pto = (unsigned long) ptep; ptep 1061 arch/s390/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 1063 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; ptep 1065 arch/s390/include/asm/pgtable.h pte = ptep_xchg_direct(vma->vm_mm, addr, ptep, pte_mkold(pte)); ptep 1071 arch/s390/include/asm/pgtable.h unsigned long address, pte_t *ptep) ptep 1073 arch/s390/include/asm/pgtable.h return ptep_test_and_clear_young(vma, address, ptep); ptep 1078 arch/s390/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 1080 arch/s390/include/asm/pgtable.h return ptep_xchg_lazy(mm, addr, ptep, __pte(_PAGE_INVALID)); ptep 1090 arch/s390/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 1092 arch/s390/include/asm/pgtable.h return ptep_xchg_direct(vma->vm_mm, addr, ptep, __pte(_PAGE_INVALID)); ptep 1105 arch/s390/include/asm/pgtable.h pte_t *ptep, int full) ptep 1108 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; ptep 1109 arch/s390/include/asm/pgtable.h *ptep = __pte(_PAGE_INVALID); ptep 1112 arch/s390/include/asm/pgtable.h return ptep_xchg_lazy(mm, addr, ptep, __pte(_PAGE_INVALID)); ptep 1117 arch/s390/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 1119 arch/s390/include/asm/pgtable.h pte_t pte = *ptep; ptep 1122 arch/s390/include/asm/pgtable.h ptep_xchg_lazy(mm, addr, ptep, pte_wrprotect(pte)); ptep 1127 arch/s390/include/asm/pgtable.h unsigned long addr, pte_t *ptep, ptep 1130 arch/s390/include/asm/pgtable.h if (pte_same(*ptep, entry)) ptep 1132 arch/s390/include/asm/pgtable.h ptep_xchg_direct(vma->vm_mm, addr, ptep, entry); ptep 1140 arch/s390/include/asm/pgtable.h pte_t *ptep, pte_t entry); ptep 1141 arch/s390/include/asm/pgtable.h void ptep_set_notify(struct mm_struct *mm, unsigned long addr, pte_t *ptep); ptep 1143 arch/s390/include/asm/pgtable.h pte_t *ptep, unsigned long bits); ptep 1145 arch/s390/include/asm/pgtable.h pte_t *ptep, int prot, unsigned long bit); ptep 1147 arch/s390/include/asm/pgtable.h pte_t *ptep , int reset); ptep 1148 arch/s390/include/asm/pgtable.h void ptep_zap_key(struct mm_struct *mm, unsigned long addr, pte_t *ptep); ptep 1151 arch/s390/include/asm/pgtable.h void ptep_unshadow_pte(struct mm_struct *mm, unsigned long saddr, pte_t *ptep); ptep 1154 arch/s390/include/asm/pgtable.h pte_t *ptep); ptep 1180 arch/s390/include/asm/pgtable.h pte_t *ptep, pte_t entry) ptep 1185 arch/s390/include/asm/pgtable.h ptep_set_pte_at(mm, addr, ptep, entry); ptep 1187 arch/s390/include/asm/pgtable.h *ptep = entry; ptep 678 arch/s390/mm/gmap.c pte_t *ptep; ptep 686 arch/s390/mm/gmap.c ptep = get_locked_pte(gmap->mm, vmaddr, &ptl); ptep 687 arch/s390/mm/gmap.c if (likely(ptep)) ptep 688 arch/s390/mm/gmap.c ptep_zap_unused(gmap->mm, vmaddr, ptep, 0); ptep 689 arch/s390/mm/gmap.c pte_unmap_unlock(ptep, ptl); ptep 1008 arch/s390/mm/gmap.c pte_t *ptep; ptep 1015 arch/s390/mm/gmap.c ptep = pte_alloc_map_lock(gmap->mm, pmdp, gaddr, &ptl); ptep 1016 arch/s390/mm/gmap.c if (!ptep) ptep 1022 arch/s390/mm/gmap.c rc = ptep_force_prot(gmap->mm, gaddr, ptep, prot, pbits); ptep 1133 arch/s390/mm/gmap.c pte_t *ptep, pte; ptep 1141 arch/s390/mm/gmap.c ptep = gmap_pte_op_walk(gmap, gaddr, &ptl); ptep 1142 arch/s390/mm/gmap.c if (ptep) { ptep 1143 arch/s390/mm/gmap.c pte = *ptep; ptep 1148 arch/s390/mm/gmap.c pte_val(*ptep) |= _PAGE_YOUNG; ptep 1212 arch/s390/mm/gmap.c pte_t *ptep; ptep 1231 arch/s390/mm/gmap.c ptep = gmap_pte_op_walk(parent, paddr, &ptl); ptep 1232 arch/s390/mm/gmap.c if (ptep) { ptep 1234 arch/s390/mm/gmap.c rc = ptep_force_prot(parent->mm, paddr, ptep, PROT_READ, ptep 2464 arch/s390/mm/gmap.c pte_t *ptep; ptep 2476 arch/s390/mm/gmap.c ptep = pte_alloc_map_lock(gmap->mm, pmdp, vmaddr, &ptl); ptep 2477 arch/s390/mm/gmap.c if (!ptep) ptep 2479 arch/s390/mm/gmap.c if (ptep_test_and_clear_uc(gmap->mm, vmaddr, ptep)) ptep 2518 arch/s390/mm/gmap.c pte_t *ptep; ptep 2521 arch/s390/mm/gmap.c ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); ptep 2522 arch/s390/mm/gmap.c if (is_zero_pfn(pte_pfn(*ptep))) ptep 2523 arch/s390/mm/gmap.c ptep_xchg_direct(walk->mm, addr, ptep, __pte(_PAGE_INVALID)); ptep 2524 arch/s390/mm/gmap.c pte_unmap_unlock(ptep, ptl); ptep 153 arch/s390/mm/hugetlbpage.c pte_t *ptep, pte_t pte) ptep 162 arch/s390/mm/hugetlbpage.c if ((pte_val(*ptep) & _REGION_ENTRY_TYPE_MASK) == _REGION_ENTRY_TYPE_R3) { ptep 170 arch/s390/mm/hugetlbpage.c pte_val(*ptep) = rste; ptep 173 arch/s390/mm/hugetlbpage.c pte_t huge_ptep_get(pte_t *ptep) ptep 175 arch/s390/mm/hugetlbpage.c return __rste_to_pte(pte_val(*ptep)); ptep 179 arch/s390/mm/hugetlbpage.c unsigned long addr, pte_t *ptep) ptep 181 arch/s390/mm/hugetlbpage.c pte_t pte = huge_ptep_get(ptep); ptep 182 arch/s390/mm/hugetlbpage.c pmd_t *pmdp = (pmd_t *) ptep; ptep 183 arch/s390/mm/hugetlbpage.c pud_t *pudp = (pud_t *) ptep; ptep 185 arch/s390/mm/hugetlbpage.c if ((pte_val(*ptep) & _REGION_ENTRY_TYPE_MASK) == _REGION_ENTRY_TYPE_R3) ptep 87 arch/s390/mm/pageattr.c pte_t *ptep, new; ptep 89 arch/s390/mm/pageattr.c ptep = pte_offset(pmdp, addr); ptep 91 arch/s390/mm/pageattr.c new = *ptep; ptep 102 arch/s390/mm/pageattr.c pgt_set((unsigned long *)ptep, pte_val(new), addr, CRDTE_DTT_PAGE); ptep 103 arch/s390/mm/pageattr.c ptep++; ptep 113 arch/s390/mm/pageattr.c pte_t *pt_dir, *ptep; ptep 126 arch/s390/mm/pageattr.c ptep = pt_dir; ptep 128 arch/s390/mm/pageattr.c pte_val(*ptep) = pte_addr | prot; ptep 130 arch/s390/mm/pageattr.c ptep++; ptep 30 arch/s390/mm/pgtable.c pte_t *ptep, int nodat) ptep 43 arch/s390/mm/pgtable.c __ptep_ipte(addr, ptep, opt, asce, IPTE_LOCAL); ptep 45 arch/s390/mm/pgtable.c __ptep_ipte(addr, ptep, 0, 0, IPTE_LOCAL); ptep 50 arch/s390/mm/pgtable.c pte_t *ptep, int nodat) ptep 63 arch/s390/mm/pgtable.c __ptep_ipte(addr, ptep, opt, asce, IPTE_GLOBAL); ptep 65 arch/s390/mm/pgtable.c __ptep_ipte(addr, ptep, 0, 0, IPTE_GLOBAL); ptep 70 arch/s390/mm/pgtable.c unsigned long addr, pte_t *ptep, ptep 75 arch/s390/mm/pgtable.c old = *ptep; ptep 81 arch/s390/mm/pgtable.c ptep_ipte_local(mm, addr, ptep, nodat); ptep 83 arch/s390/mm/pgtable.c ptep_ipte_global(mm, addr, ptep, nodat); ptep 89 arch/s390/mm/pgtable.c unsigned long addr, pte_t *ptep, ptep 94 arch/s390/mm/pgtable.c old = *ptep; ptep 100 arch/s390/mm/pgtable.c pte_val(*ptep) |= _PAGE_INVALID; ptep 103 arch/s390/mm/pgtable.c ptep_ipte_global(mm, addr, ptep, nodat); ptep 108 arch/s390/mm/pgtable.c static inline pgste_t pgste_get_lock(pte_t *ptep) ptep 121 arch/s390/mm/pgtable.c : "=&d" (old), "=&d" (new), "=Q" (ptep[PTRS_PER_PTE]) ptep 122 arch/s390/mm/pgtable.c : "Q" (ptep[PTRS_PER_PTE]) : "cc", "memory"); ptep 127 arch/s390/mm/pgtable.c static inline void pgste_set_unlock(pte_t *ptep, pgste_t pgste) ptep 133 arch/s390/mm/pgtable.c : "=Q" (ptep[PTRS_PER_PTE]) ptep 134 arch/s390/mm/pgtable.c : "d" (pgste_val(pgste)), "Q" (ptep[PTRS_PER_PTE]) ptep 139 arch/s390/mm/pgtable.c static inline pgste_t pgste_get(pte_t *ptep) ptep 143 arch/s390/mm/pgtable.c pgste = *(unsigned long *)(ptep + PTRS_PER_PTE); ptep 148 arch/s390/mm/pgtable.c static inline void pgste_set(pte_t *ptep, pgste_t pgste) ptep 151 arch/s390/mm/pgtable.c *(pgste_t *)(ptep + PTRS_PER_PTE) = pgste; ptep 176 arch/s390/mm/pgtable.c static inline void pgste_set_key(pte_t *ptep, pgste_t pgste, pte_t entry, ptep 185 arch/s390/mm/pgtable.c VM_BUG_ON(!(pte_val(*ptep) & _PAGE_INVALID)); ptep 198 arch/s390/mm/pgtable.c static inline pgste_t pgste_set_pte(pte_t *ptep, pgste_t pgste, pte_t entry) ptep 217 arch/s390/mm/pgtable.c *ptep = entry; ptep 223 arch/s390/mm/pgtable.c pte_t *ptep, pgste_t pgste) ptep 231 arch/s390/mm/pgtable.c ptep_notify(mm, addr, ptep, bits); ptep 238 arch/s390/mm/pgtable.c unsigned long addr, pte_t *ptep) ptep 243 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 244 arch/s390/mm/pgtable.c pgste = pgste_pte_notify(mm, addr, ptep, pgste); ptep 250 arch/s390/mm/pgtable.c unsigned long addr, pte_t *ptep, ptep 255 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, new, mm); ptep 262 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, new); ptep 263 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 265 arch/s390/mm/pgtable.c *ptep = new; ptep 271 arch/s390/mm/pgtable.c pte_t *ptep, pte_t new) ptep 278 arch/s390/mm/pgtable.c pgste = ptep_xchg_start(mm, addr, ptep); ptep 280 arch/s390/mm/pgtable.c old = ptep_flush_direct(mm, addr, ptep, nodat); ptep 281 arch/s390/mm/pgtable.c old = ptep_xchg_commit(mm, addr, ptep, pgste, old, new); ptep 288 arch/s390/mm/pgtable.c pte_t *ptep, pte_t new) ptep 295 arch/s390/mm/pgtable.c pgste = ptep_xchg_start(mm, addr, ptep); ptep 297 arch/s390/mm/pgtable.c old = ptep_flush_lazy(mm, addr, ptep, nodat); ptep 298 arch/s390/mm/pgtable.c old = ptep_xchg_commit(mm, addr, ptep, pgste, old, new); ptep 305 arch/s390/mm/pgtable.c pte_t *ptep) ptep 313 arch/s390/mm/pgtable.c pgste = ptep_xchg_start(mm, addr, ptep); ptep 315 arch/s390/mm/pgtable.c old = ptep_flush_lazy(mm, addr, ptep, nodat); ptep 318 arch/s390/mm/pgtable.c pgste_set(ptep, pgste); ptep 324 arch/s390/mm/pgtable.c pte_t *ptep, pte_t old_pte, pte_t pte) ptep 332 arch/s390/mm/pgtable.c pgste = pgste_get(ptep); ptep 333 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, pte, mm); ptep 334 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, pte); ptep 335 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 337 arch/s390/mm/pgtable.c *ptep = pte; ptep 536 arch/s390/mm/pgtable.c pte_t *ptep; ptep 549 arch/s390/mm/pgtable.c ptep = (pte_t *) pgtable; ptep 550 arch/s390/mm/pgtable.c pte_val(*ptep) = _PAGE_INVALID; ptep 551 arch/s390/mm/pgtable.c ptep++; ptep 552 arch/s390/mm/pgtable.c pte_val(*ptep) = _PAGE_INVALID; ptep 559 arch/s390/mm/pgtable.c pte_t *ptep, pte_t entry) ptep 565 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 567 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, entry, mm); ptep 568 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, entry); ptep 569 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 573 arch/s390/mm/pgtable.c void ptep_set_notify(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 578 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 580 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 596 arch/s390/mm/pgtable.c pte_t *ptep, int prot, unsigned long bit) ptep 602 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 603 arch/s390/mm/pgtable.c entry = *ptep; ptep 609 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 615 arch/s390/mm/pgtable.c ptep_flush_direct(mm, addr, ptep, nodat); ptep 620 arch/s390/mm/pgtable.c ptep_flush_direct(mm, addr, ptep, nodat); ptep 625 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, entry); ptep 626 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 657 arch/s390/mm/pgtable.c void ptep_unshadow_pte(struct mm_struct *mm, unsigned long saddr, pte_t *ptep) ptep 662 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 665 arch/s390/mm/pgtable.c ptep_flush_direct(mm, saddr, ptep, nodat); ptep 667 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, __pte(_PAGE_INVALID)); ptep 668 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 684 arch/s390/mm/pgtable.c pte_t *ptep, int reset) ptep 692 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 694 arch/s390/mm/pgtable.c pte = *ptep; ptep 699 arch/s390/mm/pgtable.c pte_clear(mm, addr, ptep); ptep 703 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 707 arch/s390/mm/pgtable.c void ptep_zap_key(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 714 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 717 arch/s390/mm/pgtable.c ptev = pte_val(*ptep); ptep 720 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 728 arch/s390/mm/pgtable.c pte_t *ptep) ptep 735 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 738 arch/s390/mm/pgtable.c pte = *ptep; ptep 740 arch/s390/mm/pgtable.c pgste = pgste_pte_notify(mm, addr, ptep, pgste); ptep 742 arch/s390/mm/pgtable.c ptep_ipte_global(mm, addr, ptep, nodat); ptep 747 arch/s390/mm/pgtable.c *ptep = pte; ptep 749 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 761 arch/s390/mm/pgtable.c pte_t *ptep; ptep 786 arch/s390/mm/pgtable.c ptep = pte_alloc_map_lock(mm, pmdp, addr, &ptl); ptep 787 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 790 arch/s390/mm/pgtable.c new = old = pgste_get_lock(ptep); ptep 796 arch/s390/mm/pgtable.c if (!(pte_val(*ptep) & _PAGE_INVALID)) { ptep 799 arch/s390/mm/pgtable.c paddr = pte_val(*ptep) & PAGE_MASK; ptep 813 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); ptep 814 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 863 arch/s390/mm/pgtable.c pte_t *ptep; ptep 885 arch/s390/mm/pgtable.c ptep = pte_alloc_map_lock(mm, pmdp, addr, &ptl); ptep 886 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 889 arch/s390/mm/pgtable.c new = old = pgste_get_lock(ptep); ptep 893 arch/s390/mm/pgtable.c if (!(pte_val(*ptep) & _PAGE_INVALID)) { ptep 894 arch/s390/mm/pgtable.c paddr = pte_val(*ptep) & PAGE_MASK; ptep 905 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); ptep 906 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 918 arch/s390/mm/pgtable.c pte_t *ptep; ptep 941 arch/s390/mm/pgtable.c ptep = pte_alloc_map_lock(mm, pmdp, addr, &ptl); ptep 942 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 945 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 947 arch/s390/mm/pgtable.c paddr = pte_val(*ptep) & PAGE_MASK; ptep 948 arch/s390/mm/pgtable.c if (!(pte_val(*ptep) & _PAGE_INVALID)) ptep 952 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 953 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 976 arch/s390/mm/pgtable.c pte_t *ptep; ptep 982 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); ptep 983 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 985 arch/s390/mm/pgtable.c pgste = pgste_get_lock(ptep); ptep 988 arch/s390/mm/pgtable.c *oldpte = pte_val(*ptep); ptep 1002 arch/s390/mm/pgtable.c if (pte_val(*ptep) & _PAGE_INVALID) ptep 1008 arch/s390/mm/pgtable.c if (pte_val(*ptep) & _PAGE_INVALID) ptep 1013 arch/s390/mm/pgtable.c if (!(pte_val(*ptep) & _PAGE_INVALID)) { ptep 1037 arch/s390/mm/pgtable.c if (!(pte_val(*ptep) & _PAGE_INVALID)) { ptep 1055 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, pgste); ptep 1056 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 1076 arch/s390/mm/pgtable.c pte_t *ptep; ptep 1078 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); ptep 1079 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 1081 arch/s390/mm/pgtable.c new = pgste_get_lock(ptep); ptep 1086 arch/s390/mm/pgtable.c pgste_set_unlock(ptep, new); ptep 1087 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 1103 arch/s390/mm/pgtable.c pte_t *ptep; ptep 1105 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); ptep 1106 arch/s390/mm/pgtable.c if (unlikely(!ptep)) ptep 1108 arch/s390/mm/pgtable.c *pgstep = pgste_val(pgste_get(ptep)); ptep 1109 arch/s390/mm/pgtable.c pte_unmap_unlock(ptep, ptl); ptep 31 arch/sh/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 135 arch/sh/include/asm/pgtable.h update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) ptep 137 arch/sh/include/asm/pgtable.h pte_t pte = *ptep; ptep 300 arch/sh/include/asm/pgtable_32.h static inline void set_pte(pte_t *ptep, pte_t pte) ptep 302 arch/sh/include/asm/pgtable_32.h ptep->pte_high = pte.pte_high; ptep 304 arch/sh/include/asm/pgtable_32.h ptep->pte_low = pte.pte_low; ptep 310 arch/sh/include/asm/pgtable_32.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 41 arch/sh/include/asm/pgtable_64.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 17 arch/sparc/include/asm/hugetlb.h pte_t *ptep, pte_t pte); ptep 21 arch/sparc/include/asm/hugetlb.h pte_t *ptep); ptep 31 arch/sparc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 37 arch/sparc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep) ptep 39 arch/sparc/include/asm/hugetlb.h pte_t old_pte = *ptep; ptep 40 arch/sparc/include/asm/hugetlb.h set_huge_pte_at(mm, addr, ptep, pte_wrprotect(old_pte)); ptep 45 arch/sparc/include/asm/hugetlb.h unsigned long addr, pte_t *ptep, ptep 48 arch/sparc/include/asm/hugetlb.h int changed = !pte_same(*ptep, pte); ptep 50 arch/sparc/include/asm/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); ptep 53 arch/sparc/include/asm/pgalloc_32.h void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, struct page *ptep); ptep 56 arch/sparc/include/asm/pgalloc_32.h void pmd_set(pmd_t *pmdp, pte_t *ptep); ptep 116 arch/sparc/include/asm/pgtable_32.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 118 arch/sparc/include/asm/pgtable_32.h srmmu_swap((unsigned long *)ptep, pte_val(pteval)); ptep 121 arch/sparc/include/asm/pgtable_32.h #define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval) ptep 155 arch/sparc/include/asm/pgtable_32.h static inline void __pte_clear(pte_t *ptep) ptep 157 arch/sparc/include/asm/pgtable_32.h set_pte(ptep, __pte(0)); ptep 160 arch/sparc/include/asm/pgtable_32.h static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 162 arch/sparc/include/asm/pgtable_32.h __pte_clear(ptep); ptep 345 arch/sparc/include/asm/pgtable_32.h #define update_mmu_cache(vma, address, ptep) do { } while (0) ptep 828 arch/sparc/include/asm/pgtable_64.h static inline void pmd_set(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) ptep 830 arch/sparc/include/asm/pgtable_64.h unsigned long val = __pa((unsigned long) (ptep)); ptep 920 arch/sparc/include/asm/pgtable_64.h pte_t *ptep, pte_t orig, int fullmm, ptep 924 arch/sparc/include/asm/pgtable_64.h pte_t *ptep, pte_t orig, int fullmm, ptep 934 arch/sparc/include/asm/pgtable_64.h tlb_batch_add(mm, vaddr, ptep, orig, fullmm, hugepage_shift); ptep 948 arch/sparc/include/asm/pgtable_64.h pte_t *ptep, pte_t pte, int fullmm) ptep 950 arch/sparc/include/asm/pgtable_64.h pte_t orig = *ptep; ptep 952 arch/sparc/include/asm/pgtable_64.h *ptep = pte; ptep 953 arch/sparc/include/asm/pgtable_64.h maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT); ptep 956 arch/sparc/include/asm/pgtable_64.h #define set_pte_at(mm,addr,ptep,pte) \ ptep 957 arch/sparc/include/asm/pgtable_64.h __set_pte_at((mm), (addr), (ptep), (pte), 0) ptep 959 arch/sparc/include/asm/pgtable_64.h #define pte_clear(mm,addr,ptep) \ ptep 960 arch/sparc/include/asm/pgtable_64.h set_pte_at((mm), (addr), (ptep), __pte(0UL)) ptep 963 arch/sparc/include/asm/pgtable_64.h #define pte_clear_not_present_full(mm,addr,ptep,fullmm) \ ptep 964 arch/sparc/include/asm/pgtable_64.h __set_pte_at((mm), (addr), (ptep), __pte(0UL), (fullmm)) ptep 28 arch/sparc/include/asm/tlb_64.h #define __tlb_remove_tlb_entry(tlb, ptep, address) do { } while (0) ptep 300 arch/sparc/kernel/signal32.c pte_t *ptep, pte; ptep 328 arch/sparc/kernel/signal32.c ptep = pte_offset_map(pmdp, address); ptep 329 arch/sparc/kernel/signal32.c pte = *ptep; ptep 342 arch/sparc/kernel/signal32.c pte_unmap(ptep); ptep 85 arch/sparc/mm/fault_64.c pte_t *ptep, pte; ptep 114 arch/sparc/mm/fault_64.c ptep = pte_offset_map(pmdp, tpc); ptep 115 arch/sparc/mm/fault_64.c pte = *ptep; ptep 125 arch/sparc/mm/fault_64.c pte_unmap(ptep); ptep 321 arch/sparc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) ptep 339 arch/sparc/mm/hugetlbpage.c if (!pte_present(*ptep) && pte_present(entry)) ptep 343 arch/sparc/mm/hugetlbpage.c orig = *ptep; ptep 347 arch/sparc/mm/hugetlbpage.c ptep[i] = __pte(pte_val(entry) + (i << shift)); ptep 349 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr, ptep, orig, 0, orig_shift); ptep 352 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr + REAL_HPAGE_SIZE, ptep, orig, 0, ptep 357 arch/sparc/mm/hugetlbpage.c pte_t *ptep) ptep 363 arch/sparc/mm/hugetlbpage.c entry = *ptep; ptep 382 arch/sparc/mm/hugetlbpage.c ptep[i] = __pte(0UL); ptep 384 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr, ptep, entry, 0, orig_shift); ptep 387 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr + REAL_HPAGE_SIZE, ptep, entry, 0, ptep 416 arch/sparc/mm/init_64.c void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) ptep 421 arch/sparc/mm/init_64.c pte_t pte = *ptep; ptep 535 arch/sparc/mm/init_64.c pte_t *ptep = pte_offset_kernel(pmdp, kaddr); ptep 537 arch/sparc/mm/init_64.c paddr = pte_val(*ptep) & mask; ptep 243 arch/sparc/mm/io-unit.c pte_t *ptep; ptep 248 arch/sparc/mm/io-unit.c ptep = pte_offset_map(pmdp, addr); ptep 250 arch/sparc/mm/io-unit.c set_pte(ptep, mk_pte(virt_to_page(page), dvma_prot)); ptep 347 arch/sparc/mm/iommu.c pte_t *ptep; ptep 358 arch/sparc/mm/iommu.c ptep = pte_offset_map(pmdp, addr); ptep 360 arch/sparc/mm/iommu.c set_pte(ptep, mk_pte(virt_to_page(page), dvma_prot)); ptep 137 arch/sparc/mm/srmmu.c void pmd_set(pmd_t *pmdp, pte_t *ptep) ptep 142 arch/sparc/mm/srmmu.c ptp = __nocache_pa(ptep) >> 4; ptep 149 arch/sparc/mm/srmmu.c void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, struct page *ptep) ptep 154 arch/sparc/mm/srmmu.c ptp = page_to_pfn(ptep) << (PAGE_SHIFT-4); /* watch for overflow */ ptep 520 arch/sparc/mm/srmmu.c pte_t *ptep; ptep 526 arch/sparc/mm/srmmu.c ptep = pte_offset_kernel(pmdp, virt_addr); ptep 536 arch/sparc/mm/srmmu.c set_pte(ptep, __pte(tmp)); ptep 555 arch/sparc/mm/srmmu.c pte_t *ptep; ptep 559 arch/sparc/mm/srmmu.c ptep = pte_offset_kernel(pmdp, virt_addr); ptep 562 arch/sparc/mm/srmmu.c __pte_clear(ptep); ptep 697 arch/sparc/mm/srmmu.c pte_t *ptep; ptep 711 arch/sparc/mm/srmmu.c ptep = __srmmu_get_nocache(PTE_SIZE, PTE_SIZE); ptep 712 arch/sparc/mm/srmmu.c if (ptep == NULL) ptep 714 arch/sparc/mm/srmmu.c memset(__nocache_fix(ptep), 0, PTE_SIZE); ptep 715 arch/sparc/mm/srmmu.c pmd_set(__nocache_fix(pmdp), ptep); ptep 728 arch/sparc/mm/srmmu.c pte_t *ptep; ptep 741 arch/sparc/mm/srmmu.c ptep = __srmmu_get_nocache(PTE_SIZE, ptep 743 arch/sparc/mm/srmmu.c if (ptep == NULL) ptep 745 arch/sparc/mm/srmmu.c memset(ptep, 0, PTE_SIZE); ptep 746 arch/sparc/mm/srmmu.c pmd_set(pmdp, ptep); ptep 783 arch/sparc/mm/srmmu.c pte_t *ptep; ptep 828 arch/sparc/mm/srmmu.c ptep = __srmmu_get_nocache(PTE_SIZE, PTE_SIZE); ptep 829 arch/sparc/mm/srmmu.c if (ptep == NULL) ptep 831 arch/sparc/mm/srmmu.c memset(__nocache_fix(ptep), 0, PTE_SIZE); ptep 832 arch/sparc/mm/srmmu.c pmd_set(__nocache_fix(pmdp), ptep); ptep 847 arch/sparc/mm/srmmu.c ptep = pte_offset_kernel(__nocache_fix(pmdp), start); ptep 848 arch/sparc/mm/srmmu.c *(pte_t *)__nocache_fix(ptep) = __pte(probed); ptep 114 arch/sparc/mm/tlb.c pte_t *ptep, pte_t orig, int fullmm, ptep 349 arch/um/include/asm/pgtable.h #define update_mmu_cache(vma,address,ptep) do ; while (0) ptep 366 arch/um/include/asm/pgtable.h #define kpte_clear_flush(ptep, vaddr) \ ptep 368 arch/um/include/asm/pgtable.h pte_clear(&init_mm, (vaddr), (ptep)); \ ptep 36 arch/unicore32/include/asm/cpu-single.h extern void cpu_set_pte(pte_t *ptep, pte_t pte); ptep 68 arch/unicore32/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, pte_t *ptep) ptep 70 arch/unicore32/include/asm/pgalloc.h unsigned long pte_ptr = (unsigned long)ptep; ptep 80 arch/unicore32/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t *pmdp, pgtable_t ptep) ptep 83 arch/unicore32/include/asm/pgalloc.h page_to_pfn(ptep) << PAGE_SHIFT | _PAGE_USER_TABLE); ptep 155 arch/unicore32/include/asm/pgtable.h #define pte_clear(mm, addr, ptep) set_pte(ptep, __pte(0)) ptep 164 arch/unicore32/include/asm/pgtable.h #define set_pte(ptep, pte) cpu_set_pte(ptep, pte) ptep 166 arch/unicore32/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pteval) \ ptep 168 arch/unicore32/include/asm/pgtable.h set_pte(ptep, pteval); \ ptep 185 arch/unicore32/include/asm/tlbflush.h unsigned long addr, pte_t *ptep); ptep 490 arch/unicore32/mm/mmu.c pte_t *ptep) ptep 492 arch/unicore32/mm/mmu.c unsigned long pfn = pte_pfn(*ptep); ptep 376 arch/x86/include/asm/io.h extern bool is_early_ioremap_ptep(pte_t *ptep); ptep 414 arch/x86/include/asm/paravirt.h pte_t *ptep) ptep 418 arch/x86/include/asm/paravirt.h ret = PVOP_CALL3(pteval_t, mmu.ptep_modify_prot_start, vma, addr, ptep); ptep 424 arch/x86/include/asm/paravirt.h pte_t *ptep, pte_t old_pte, pte_t pte) ptep 429 arch/x86/include/asm/paravirt.h pv_ops.mmu.ptep_modify_prot_commit(vma, addr, ptep, pte); ptep 432 arch/x86/include/asm/paravirt.h vma, addr, ptep, pte.pte); ptep 435 arch/x86/include/asm/paravirt.h static inline void set_pte(pte_t *ptep, pte_t pte) ptep 438 arch/x86/include/asm/paravirt.h PVOP_VCALL3(mmu.set_pte, ptep, pte.pte, (u64)pte.pte >> 32); ptep 440 arch/x86/include/asm/paravirt.h PVOP_VCALL2(mmu.set_pte, ptep, pte.pte); ptep 444 arch/x86/include/asm/paravirt.h pte_t *ptep, pte_t pte) ptep 448 arch/x86/include/asm/paravirt.h pv_ops.mmu.set_pte_at(mm, addr, ptep, pte); ptep 450 arch/x86/include/asm/paravirt.h PVOP_VCALL4(mmu.set_pte_at, mm, addr, ptep, pte.pte); ptep 570 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte) ptep 572 arch/x86/include/asm/paravirt.h PVOP_VCALL3(mmu.set_pte_atomic, ptep, pte.pte, pte.pte >> 32); ptep 576 arch/x86/include/asm/paravirt.h pte_t *ptep) ptep 578 arch/x86/include/asm/paravirt.h PVOP_VCALL3(mmu.pte_clear, mm, addr, ptep); ptep 586 arch/x86/include/asm/paravirt.h static inline void set_pte_atomic(pte_t *ptep, pte_t pte) ptep 588 arch/x86/include/asm/paravirt.h set_pte(ptep, pte); ptep 592 arch/x86/include/asm/paravirt.h pte_t *ptep) ptep 594 arch/x86/include/asm/paravirt.h set_pte_at(mm, addr, ptep, __pte(0)); ptep 248 arch/x86/include/asm/paravirt_types.h void (*set_pte)(pte_t *ptep, pte_t pteval); ptep 250 arch/x86/include/asm/paravirt_types.h pte_t *ptep, pte_t pteval); ptep 254 arch/x86/include/asm/paravirt_types.h pte_t *ptep); ptep 256 arch/x86/include/asm/paravirt_types.h pte_t *ptep, pte_t pte); ptep 266 arch/x86/include/asm/paravirt_types.h void (*set_pte_atomic)(pte_t *ptep, pte_t pteval); ptep 268 arch/x86/include/asm/paravirt_types.h pte_t *ptep); ptep 15 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte(pte_t *ptep , pte_t pte) ptep 17 arch/x86/include/asm/pgtable-2level.h *ptep = pte; ptep 29 arch/x86/include/asm/pgtable-2level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) ptep 31 arch/x86/include/asm/pgtable-2level.h native_set_pte(ptep, pte); ptep 30 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte(pte_t *ptep, pte_t pte) ptep 32 arch/x86/include/asm/pgtable-3level.h ptep->pte_high = pte.pte_high; ptep 34 arch/x86/include/asm/pgtable-3level.h ptep->pte_low = pte.pte_low; ptep 91 arch/x86/include/asm/pgtable-3level.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) ptep 93 arch/x86/include/asm/pgtable-3level.h set_64bit((unsigned long long *)(ptep), native_pte_val(pte)); ptep 115 arch/x86/include/asm/pgtable-3level.h pte_t *ptep) ptep 117 arch/x86/include/asm/pgtable-3level.h ptep->pte_low = 0; ptep 119 arch/x86/include/asm/pgtable-3level.h ptep->pte_high = 0; ptep 151 arch/x86/include/asm/pgtable-3level.h static inline pte_t native_ptep_get_and_clear(pte_t *ptep) ptep 155 arch/x86/include/asm/pgtable-3level.h res.pte = (pteval_t)arch_atomic64_xchg((atomic64_t *)ptep, 0); ptep 63 arch/x86/include/asm/pgtable.h #define set_pte(ptep, pte) native_set_pte(ptep, pte) ptep 64 arch/x86/include/asm/pgtable.h #define set_pte_at(mm, addr, ptep, pte) native_set_pte_at(mm, addr, ptep, pte) ptep 66 arch/x86/include/asm/pgtable.h #define set_pte_atomic(ptep, pte) \ ptep 67 arch/x86/include/asm/pgtable.h native_set_pte_atomic(ptep, pte) ptep 92 arch/x86/include/asm/pgtable.h #define pte_clear(mm, addr, ptep) native_pte_clear(mm, addr, ptep) ptep 1058 arch/x86/include/asm/pgtable.h static inline pte_t native_local_ptep_get_and_clear(pte_t *ptep) ptep 1060 arch/x86/include/asm/pgtable.h pte_t res = *ptep; ptep 1063 arch/x86/include/asm/pgtable.h native_pte_clear(NULL, 0, ptep); ptep 1084 arch/x86/include/asm/pgtable.h pte_t *ptep , pte_t pte) ptep 1086 arch/x86/include/asm/pgtable.h native_set_pte(ptep, pte); ptep 1112 arch/x86/include/asm/pgtable.h unsigned long address, pte_t *ptep, ptep 1117 arch/x86/include/asm/pgtable.h unsigned long addr, pte_t *ptep); ptep 1121 arch/x86/include/asm/pgtable.h unsigned long address, pte_t *ptep); ptep 1125 arch/x86/include/asm/pgtable.h pte_t *ptep) ptep 1127 arch/x86/include/asm/pgtable.h pte_t pte = native_ptep_get_and_clear(ptep); ptep 1133 arch/x86/include/asm/pgtable.h unsigned long addr, pte_t *ptep, ptep 1142 arch/x86/include/asm/pgtable.h pte = native_local_ptep_get_and_clear(ptep); ptep 1144 arch/x86/include/asm/pgtable.h pte = ptep_get_and_clear(mm, addr, ptep); ptep 1151 arch/x86/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 1153 arch/x86/include/asm/pgtable.h clear_bit(_PAGE_BIT_RW, (unsigned long *)&ptep->pte); ptep 1332 arch/x86/include/asm/pgtable.h unsigned long addr, pte_t *ptep) ptep 60 arch/x86/include/asm/pgtable_32.h #define kpte_clear_flush(ptep, vaddr) \ ptep 62 arch/x86/include/asm/pgtable_32.h pte_clear(&init_mm, (vaddr), (ptep)); \ ptep 59 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte(pte_t *ptep, pte_t pte) ptep 61 arch/x86/include/asm/pgtable_64.h WRITE_ONCE(*ptep, pte); ptep 65 arch/x86/include/asm/pgtable_64.h pte_t *ptep) ptep 67 arch/x86/include/asm/pgtable_64.h native_set_pte(ptep, native_make_pte(0)); ptep 70 arch/x86/include/asm/pgtable_64.h static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) ptep 72 arch/x86/include/asm/pgtable_64.h native_set_pte(ptep, pte); ptep 7 arch/x86/include/asm/tlb.h #define __tlb_remove_tlb_entry(tlb, ptep, address) do { } while (0) ptep 794 arch/x86/kernel/alternative.c pte_t pte, *ptep; ptep 831 arch/x86/kernel/alternative.c ptep = get_locked_pte(poking_mm, poking_addr, &ptl); ptep 836 arch/x86/kernel/alternative.c VM_BUG_ON(!ptep); ptep 839 arch/x86/kernel/alternative.c set_pte_at(poking_mm, poking_addr, ptep, pte); ptep 843 arch/x86/kernel/alternative.c set_pte_at(poking_mm, poking_addr + PAGE_SIZE, ptep + 1, pte); ptep 862 arch/x86/kernel/alternative.c pte_clear(poking_mm, poking_addr, ptep); ptep 864 arch/x86/kernel/alternative.c pte_clear(poking_mm, poking_addr + PAGE_SIZE, ptep + 1); ptep 887 arch/x86/kernel/alternative.c pte_unmap_unlock(ptep, ptl); ptep 78 arch/x86/kernel/head32.c pte_t pte, *ptep; ptep 92 arch/x86/kernel/head32.c ptep = (pte_t *)__pa(__brk_base); ptep 97 arch/x86/kernel/head32.c SET_PL2(pl2, (unsigned long)ptep | PDE_IDENT_ATTR); ptep 104 arch/x86/kernel/head32.c *ptep = pte; ptep 106 arch/x86/kernel/head32.c ptep++; ptep 117 arch/x86/kernel/head32.c *ptr = (unsigned long)ptep + PAGE_OFFSET; ptep 232 arch/x86/kernel/ldt.c pte_t pte, *ptep; ptep 242 arch/x86/kernel/ldt.c ptep = get_locked_pte(mm, va, &ptl); ptep 243 arch/x86/kernel/ldt.c if (!ptep) ptep 254 arch/x86/kernel/ldt.c set_pte_at(mm, va, ptep, pte); ptep 255 arch/x86/kernel/ldt.c pte_unmap_unlock(ptep, ptl); ptep 282 arch/x86/kernel/ldt.c pte_t *ptep; ptep 285 arch/x86/kernel/ldt.c ptep = get_locked_pte(mm, va, &ptl); ptep 286 arch/x86/kernel/ldt.c pte_clear(mm, va, ptep); ptep 287 arch/x86/kernel/ldt.c pte_unmap_unlock(ptep, ptl); ptep 1884 arch/x86/kvm/mmu.c pte_t *ptep = (pte_t *)data; ptep 1887 arch/x86/kvm/mmu.c WARN_ON(pte_huge(*ptep)); ptep 1888 arch/x86/kvm/mmu.c new_pfn = pte_pfn(*ptep); ptep 1897 arch/x86/kvm/mmu.c if (pte_write(*ptep)) { ptep 713 arch/x86/mm/init.c pte_t *ptep; ptep 736 arch/x86/mm/init.c ptep = get_locked_pte(poking_mm, poking_addr, &ptl); ptep 737 arch/x86/mm/init.c BUG_ON(!ptep); ptep 738 arch/x86/mm/init.c pte_unmap_unlock(ptep, ptl); ptep 818 arch/x86/mm/ioremap.c bool __init is_early_ioremap_ptep(pte_t *ptep) ptep 820 arch/x86/mm/ioremap.c return ptep >= &bm_pte[0] && ptep < &bm_pte[PAGE_SIZE/sizeof(pte_t)]; ptep 480 arch/x86/mm/pgtable.c unsigned long address, pte_t *ptep, ptep 483 arch/x86/mm/pgtable.c int changed = !pte_same(*ptep, entry); ptep 486 arch/x86/mm/pgtable.c set_pte(ptep, entry); ptep 535 arch/x86/mm/pgtable.c unsigned long addr, pte_t *ptep) ptep 539 arch/x86/mm/pgtable.c if (pte_young(*ptep)) ptep 541 arch/x86/mm/pgtable.c (unsigned long *) &ptep->pte); ptep 572 arch/x86/mm/pgtable.c unsigned long address, pte_t *ptep) ptep 587 arch/x86/mm/pgtable.c return ptep_test_and_clear_young(vma, address, ptep); ptep 351 arch/x86/xen/enlighten_pv.c pte_t *ptep; ptep 357 arch/x86/xen/enlighten_pv.c ptep = lookup_address((unsigned long)v, &level); ptep 358 arch/x86/xen/enlighten_pv.c BUG_ON(ptep == NULL); ptep 360 arch/x86/xen/enlighten_pv.c pfn = pte_pfn(*ptep); ptep 456 arch/x86/xen/enlighten_pv.c pte_t *ptep; ptep 470 arch/x86/xen/enlighten_pv.c ptep = lookup_address(va, &level); ptep 471 arch/x86/xen/enlighten_pv.c BUG_ON(ptep == NULL); ptep 473 arch/x86/xen/enlighten_pv.c pfn = pte_pfn(*ptep); ptep 20 arch/x86/xen/mmu.h pte_t xen_ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep); ptep 22 arch/x86/xen/mmu.h pte_t *ptep, pte_t pte); ptep 260 arch/x86/xen/mmu_pv.c static bool xen_batched_set_pte(pte_t *ptep, pte_t pteval) ptep 269 arch/x86/xen/mmu_pv.c u.ptr = virt_to_machine(ptep).maddr | MMU_NORMAL_PT_UPDATE; ptep 278 arch/x86/xen/mmu_pv.c static inline void __xen_set_pte(pte_t *ptep, pte_t pteval) ptep 280 arch/x86/xen/mmu_pv.c if (!xen_batched_set_pte(ptep, pteval)) { ptep 290 arch/x86/xen/mmu_pv.c u.ptr = virt_to_machine(ptep).maddr | MMU_NORMAL_PT_UPDATE; ptep 296 arch/x86/xen/mmu_pv.c static void xen_set_pte(pte_t *ptep, pte_t pteval) ptep 298 arch/x86/xen/mmu_pv.c trace_xen_mmu_set_pte(ptep, pteval); ptep 299 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pteval); ptep 303 arch/x86/xen/mmu_pv.c pte_t *ptep, pte_t pteval) ptep 305 arch/x86/xen/mmu_pv.c trace_xen_mmu_set_pte_at(mm, addr, ptep, pteval); ptep 306 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pteval); ptep 310 arch/x86/xen/mmu_pv.c unsigned long addr, pte_t *ptep) ptep 313 arch/x86/xen/mmu_pv.c trace_xen_mmu_ptep_modify_prot_start(vma->vm_mm, addr, ptep, *ptep); ptep 314 arch/x86/xen/mmu_pv.c return *ptep; ptep 318 arch/x86/xen/mmu_pv.c pte_t *ptep, pte_t pte) ptep 322 arch/x86/xen/mmu_pv.c trace_xen_mmu_ptep_modify_prot_commit(vma->vm_mm, addr, ptep, pte); ptep 325 arch/x86/xen/mmu_pv.c u.ptr = virt_to_machine(ptep).maddr | MMU_PT_UPDATE_PRESERVE_AD; ptep 443 arch/x86/xen/mmu_pv.c static void xen_set_pte_atomic(pte_t *ptep, pte_t pte) ptep 445 arch/x86/xen/mmu_pv.c trace_xen_mmu_set_pte_atomic(ptep, pte); ptep 446 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pte); ptep 449 arch/x86/xen/mmu_pv.c static void xen_pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 451 arch/x86/xen/mmu_pv.c trace_xen_mmu_pte_clear(mm, addr, ptep); ptep 452 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, native_make_pte(0)); ptep 1562 arch/x86/xen/mmu_pv.c static void __init xen_set_pte_init(pte_t *ptep, pte_t pte) ptep 1567 arch/x86/xen/mmu_pv.c && pte_val_ma(*ptep) & _PAGE_PRESENT) ptep 1568 arch/x86/xen/mmu_pv.c pte = __pte_ma(((pte_val_ma(*ptep) & _PAGE_RW) | ~_PAGE_RW) & ptep 1571 arch/x86/xen/mmu_pv.c __xen_set_pte(ptep, pte); ptep 2691 arch/x86/xen/mmu_pv.c static int remap_area_pfn_pte_fn(pte_t *ptep, unsigned long addr, void *data) ptep 2705 arch/x86/xen/mmu_pv.c rmd->mmu_update->ptr = virt_to_machine(ptep).maddr; ptep 220 arch/x86/xen/p2m.c pte_t *ptep; ptep 248 arch/x86/xen/p2m.c ptep = lookup_address((unsigned long)(xen_p2m_addr + pfn), ptep 250 arch/x86/xen/p2m.c BUG_ON(!ptep || level != PG_LEVEL_4K); ptep 251 arch/x86/xen/p2m.c mfn = pte_mfn(*ptep); ptep 252 arch/x86/xen/p2m.c ptep = (pte_t *)((unsigned long)ptep & ~(PAGE_SIZE - 1)); ptep 258 arch/x86/xen/p2m.c if (ptep == p2m_missing_pte || ptep == p2m_identity_pte) { ptep 336 arch/x86/xen/p2m.c pte_t *ptep; ptep 388 arch/x86/xen/p2m.c ptep = populate_extra_pte((unsigned long)(p2m + pfn)); ptep 389 arch/x86/xen/p2m.c set_pte(ptep, ptep 398 arch/x86/xen/p2m.c ptep = populate_extra_pte((unsigned long)(p2m + pfn)); ptep 399 arch/x86/xen/p2m.c set_pte(ptep, ptep 405 arch/x86/xen/p2m.c ptep = (type == P2M_TYPE_MISSING) ? ptep 410 arch/x86/xen/p2m.c set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); ptep 441 arch/x86/xen/p2m.c pte_t *ptep; ptep 451 arch/x86/xen/p2m.c ptep = lookup_address((unsigned long)(xen_p2m_addr + pfn), &level); ptep 452 arch/x86/xen/p2m.c BUG_ON(!ptep || level != PG_LEVEL_4K); ptep 459 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) ptep 539 arch/x86/xen/p2m.c pte_t *ptep, *pte_pg; ptep 545 arch/x86/xen/p2m.c ptep = lookup_address(addr, &level); ptep 546 arch/x86/xen/p2m.c BUG_ON(!ptep || level != PG_LEVEL_4K); ptep 547 arch/x86/xen/p2m.c pte_pg = (pte_t *)((unsigned long)ptep & ~(PAGE_SIZE - 1)); ptep 551 arch/x86/xen/p2m.c ptep = alloc_p2m_pmd(addr, pte_pg); ptep 552 arch/x86/xen/p2m.c if (!ptep) ptep 589 arch/x86/xen/p2m.c p2m_pfn = pte_pfn(READ_ONCE(*ptep)); ptep 606 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == p2m_pfn) { ptep 609 arch/x86/xen/p2m.c set_pte(ptep, ptep 656 arch/x86/xen/p2m.c pte_t *ptep; ptep 671 arch/x86/xen/p2m.c ptep = lookup_address((unsigned long)(xen_p2m_addr + pfn), &level); ptep 672 arch/x86/xen/p2m.c BUG_ON(!ptep || level != PG_LEVEL_4K); ptep 674 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing))) ptep 677 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) ptep 19 arch/xtensa/include/asm/pgalloc.h #define pmd_populate_kernel(mm, pmdp, ptep) \ ptep 20 arch/xtensa/include/asm/pgalloc.h (pmd_val(*(pmdp)) = ((unsigned long)ptep)) ptep 38 arch/xtensa/include/asm/pgalloc.h pte_t *ptep; ptep 41 arch/xtensa/include/asm/pgalloc.h ptep = (pte_t *)__get_free_page(GFP_KERNEL); ptep 42 arch/xtensa/include/asm/pgalloc.h if (!ptep) ptep 45 arch/xtensa/include/asm/pgalloc.h pte_clear(NULL, 0, ptep + i); ptep 46 arch/xtensa/include/asm/pgalloc.h return ptep; ptep 259 arch/xtensa/include/asm/pgtable.h #define pte_clear(mm,addr,ptep) \ ptep 260 arch/xtensa/include/asm/pgtable.h do { update_pte(ptep, __pte(_PAGE_CA_INVALID | _PAGE_USER)); } while (0) ptep 310 arch/xtensa/include/asm/pgtable.h static inline void update_pte(pte_t *ptep, pte_t pteval) ptep 312 arch/xtensa/include/asm/pgtable.h *ptep = pteval; ptep 314 arch/xtensa/include/asm/pgtable.h __asm__ __volatile__ ("dhwb %0, 0" :: "a" (ptep)); ptep 322 arch/xtensa/include/asm/pgtable.h set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pteval) ptep 324 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pteval); ptep 327 arch/xtensa/include/asm/pgtable.h static inline void set_pte(pte_t *ptep, pte_t pteval) ptep 329 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pteval); ptep 342 arch/xtensa/include/asm/pgtable.h pte_t *ptep) ptep 344 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; ptep 347 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pte_mkold(pte)); ptep 352 arch/xtensa/include/asm/pgtable.h ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 354 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; ptep 355 arch/xtensa/include/asm/pgtable.h pte_clear(mm, addr, ptep); ptep 360 arch/xtensa/include/asm/pgtable.h ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) ptep 362 arch/xtensa/include/asm/pgtable.h pte_t pte = *ptep; ptep 363 arch/xtensa/include/asm/pgtable.h update_pte(ptep, pte_wrprotect(pte)); ptep 432 arch/xtensa/include/asm/pgtable.h unsigned long address, pte_t *ptep); ptep 211 arch/xtensa/mm/cache.c update_mmu_cache(struct vm_area_struct * vma, unsigned long addr, pte_t *ptep) ptep 213 arch/xtensa/mm/cache.c unsigned long pfn = pte_pfn(*ptep); ptep 296 drivers/iommu/io-pgtable-arm-v7s.c static void __arm_v7s_pte_sync(arm_v7s_iopte *ptep, int num_entries, ptep 302 drivers/iommu/io-pgtable-arm-v7s.c dma_sync_single_for_device(cfg->iommu_dev, __arm_v7s_dma_addr(ptep), ptep 303 drivers/iommu/io-pgtable-arm-v7s.c num_entries * sizeof(*ptep), DMA_TO_DEVICE); ptep 305 drivers/iommu/io-pgtable-arm-v7s.c static void __arm_v7s_set_pte(arm_v7s_iopte *ptep, arm_v7s_iopte pte, ptep 311 drivers/iommu/io-pgtable-arm-v7s.c ptep[i] = pte; ptep 313 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_pte_sync(ptep, num_entries, cfg); ptep 414 drivers/iommu/io-pgtable-arm-v7s.c int lvl, int num_entries, arm_v7s_iopte *ptep) ptep 421 drivers/iommu/io-pgtable-arm-v7s.c if (ARM_V7S_PTE_IS_TABLE(ptep[i], lvl)) { ptep 429 drivers/iommu/io-pgtable-arm-v7s.c tblp = ptep - ARM_V7S_LVL_IDX(iova, lvl); ptep 433 drivers/iommu/io-pgtable-arm-v7s.c } else if (ptep[i]) { ptep 445 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_set_pte(ptep, pte, num_entries, cfg); ptep 450 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep, ptep 467 drivers/iommu/io-pgtable-arm-v7s.c old = cmpxchg_relaxed(ptep, curr, new); ptep 468 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_pte_sync(ptep, 1, cfg); ptep 475 drivers/iommu/io-pgtable-arm-v7s.c int lvl, arm_v7s_iopte *ptep) ptep 482 drivers/iommu/io-pgtable-arm-v7s.c ptep += ARM_V7S_LVL_IDX(iova, lvl); ptep 487 drivers/iommu/io-pgtable-arm-v7s.c lvl, num_entries, ptep); ptep 494 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); ptep 500 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_install_table(cptep, ptep, 0, cfg); ptep 505 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_pte_sync(ptep, 1, cfg); ptep 569 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep) ptep 577 drivers/iommu/io-pgtable-arm-v7s.c pte = *ptep; ptep 581 drivers/iommu/io-pgtable-arm-v7s.c ptep -= idx & (ARM_V7S_CONT_PAGES - 1); ptep 584 drivers/iommu/io-pgtable-arm-v7s.c ptep[i] = pte + i * size; ptep 586 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_pte_sync(ptep, ARM_V7S_CONT_PAGES, &iop->cfg); ptep 597 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep) ptep 623 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg); ptep 641 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep) ptep 652 drivers/iommu/io-pgtable-arm-v7s.c ptep += idx; ptep 654 drivers/iommu/io-pgtable-arm-v7s.c pte[i] = READ_ONCE(ptep[i]); ptep 673 drivers/iommu/io-pgtable-arm-v7s.c pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep); ptep 681 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_set_pte(ptep, 0, num_entries, &iop->cfg); ptep 688 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte[i], lvl, data); ptep 689 drivers/iommu/io-pgtable-arm-v7s.c __arm_v7s_free_table(ptep, lvl + 1, data); ptep 709 drivers/iommu/io-pgtable-arm-v7s.c ptep); ptep 713 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte[0], lvl, data); ptep 714 drivers/iommu/io-pgtable-arm-v7s.c return __arm_v7s_unmap(data, gather, iova, size, lvl + 1, ptep); ptep 732 drivers/iommu/io-pgtable-arm-v7s.c arm_v7s_iopte *ptep = data->pgd, pte; ptep 737 drivers/iommu/io-pgtable-arm-v7s.c ptep += ARM_V7S_LVL_IDX(iova, ++lvl); ptep 738 drivers/iommu/io-pgtable-arm-v7s.c pte = READ_ONCE(*ptep); ptep 739 drivers/iommu/io-pgtable-arm-v7s.c ptep = iopte_deref(pte, lvl, data); ptep 278 drivers/iommu/io-pgtable-arm.c static void __arm_lpae_sync_pte(arm_lpae_iopte *ptep, ptep 281 drivers/iommu/io-pgtable-arm.c dma_sync_single_for_device(cfg->iommu_dev, __arm_lpae_dma_addr(ptep), ptep 282 drivers/iommu/io-pgtable-arm.c sizeof(*ptep), DMA_TO_DEVICE); ptep 285 drivers/iommu/io-pgtable-arm.c static void __arm_lpae_set_pte(arm_lpae_iopte *ptep, arm_lpae_iopte pte, ptep 288 drivers/iommu/io-pgtable-arm.c *ptep = pte; ptep 291 drivers/iommu/io-pgtable-arm.c __arm_lpae_sync_pte(ptep, cfg); ptep 297 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep); ptep 301 drivers/iommu/io-pgtable-arm.c int lvl, arm_lpae_iopte *ptep) ptep 318 drivers/iommu/io-pgtable-arm.c __arm_lpae_set_pte(ptep, pte, &data->iop.cfg); ptep 324 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep) ptep 326 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = *ptep; ptep 340 drivers/iommu/io-pgtable-arm.c tblp = ptep - ARM_LPAE_LVL_IDX(iova, lvl, data); ptep 347 drivers/iommu/io-pgtable-arm.c __arm_lpae_init_pte(data, paddr, prot, lvl, ptep); ptep 352 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep, ptep 369 drivers/iommu/io-pgtable-arm.c old = cmpxchg64_relaxed(ptep, curr, new); ptep 375 drivers/iommu/io-pgtable-arm.c __arm_lpae_sync_pte(ptep, cfg); ptep 377 drivers/iommu/io-pgtable-arm.c WRITE_ONCE(*ptep, new | ARM_LPAE_PTE_SW_SYNC); ptep 384 drivers/iommu/io-pgtable-arm.c int lvl, arm_lpae_iopte *ptep) ptep 392 drivers/iommu/io-pgtable-arm.c ptep += ARM_LPAE_LVL_IDX(iova, lvl, data); ptep 396 drivers/iommu/io-pgtable-arm.c return arm_lpae_init_pte(data, iova, paddr, prot, lvl, ptep); ptep 403 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); ptep 409 drivers/iommu/io-pgtable-arm.c pte = arm_lpae_install_table(cptep, ptep, 0, cfg); ptep 413 drivers/iommu/io-pgtable-arm.c __arm_lpae_sync_pte(ptep, cfg); ptep 482 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep = data->pgd; ptep 495 drivers/iommu/io-pgtable-arm.c ret = __arm_lpae_map(data, iova, paddr, size, prot, lvl, ptep); ptep 506 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep) ptep 516 drivers/iommu/io-pgtable-arm.c start = ptep; ptep 520 drivers/iommu/io-pgtable-arm.c end = ptep; ptep 522 drivers/iommu/io-pgtable-arm.c end = (void *)ptep + table_size; ptep 524 drivers/iommu/io-pgtable-arm.c while (ptep != end) { ptep 525 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = *ptep++; ptep 548 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep) ptep 578 drivers/iommu/io-pgtable-arm.c pte = arm_lpae_install_table(tablep, ptep, blk_pte, cfg); ptep 601 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep) ptep 610 drivers/iommu/io-pgtable-arm.c ptep += ARM_LPAE_LVL_IDX(iova, lvl, data); ptep 611 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); ptep 617 drivers/iommu/io-pgtable-arm.c __arm_lpae_set_pte(ptep, 0, &iop->cfg); ptep 623 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); ptep 624 drivers/iommu/io-pgtable-arm.c __arm_lpae_free_pgtable(data, lvl + 1, ptep); ptep 643 drivers/iommu/io-pgtable-arm.c lvl + 1, ptep); ptep 647 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); ptep 648 drivers/iommu/io-pgtable-arm.c return __arm_lpae_unmap(data, gather, iova, size, lvl + 1, ptep); ptep 655 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte *ptep = data->pgd; ptep 661 drivers/iommu/io-pgtable-arm.c return __arm_lpae_unmap(data, gather, iova, size, lvl, ptep); ptep 668 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte, *ptep = data->pgd; ptep 673 drivers/iommu/io-pgtable-arm.c if (!ptep) ptep 677 drivers/iommu/io-pgtable-arm.c ptep += ARM_LPAE_LVL_IDX(iova, lvl, data); ptep 678 drivers/iommu/io-pgtable-arm.c pte = READ_ONCE(*ptep); ptep 689 drivers/iommu/io-pgtable-arm.c ptep = iopte_deref(pte, data); ptep 96 drivers/xen/xlate_mmu.c static int remap_pte_fn(pte_t *ptep, unsigned long addr, void *data) ptep 138 drivers/xen/xlate_mmu.c set_pte_at(info->vma->vm_mm, addr, ptep, pte); ptep 273 drivers/xen/xlate_mmu.c static int remap_pfn_fn(pte_t *ptep, unsigned long addr, void *data) ptep 279 drivers/xen/xlate_mmu.c set_pte_at(r->mm, addr, ptep, pte); ptep 780 fs/dax.c pte_t pte, *ptep = NULL; ptep 802 fs/dax.c &ptep, &pmdp, &ptl)) ptep 830 fs/dax.c if (pfn != pte_pfn(*ptep)) ptep 832 fs/dax.c if (!pte_dirty(*ptep) && !pte_write(*ptep)) ptep 836 fs/dax.c pte = ptep_clear_flush(vma, address, ptep); ptep 839 fs/dax.c set_pte_at(vma->vm_mm, address, ptep, pte); ptep 841 fs/dax.c pte_unmap_unlock(ptep, ptl); ptep 1457 fs/proc/task_mmu.c static int pagemap_hugetlb_range(pte_t *ptep, unsigned long hmask, ptep 1470 fs/proc/task_mmu.c pte = huge_ptep_get(ptep); ptep 234 fs/userfaultfd.c pte_t *ptep, pte; ptep 239 fs/userfaultfd.c ptep = huge_pte_offset(mm, address, vma_mmu_pagesize(vma)); ptep 241 fs/userfaultfd.c if (!ptep) ptep 245 fs/userfaultfd.c pte = huge_ptep_get(ptep); ptep 37 include/asm-generic/hugetlb.h pte_t *ptep, unsigned long sz) ptep 39 include/asm-generic/hugetlb.h pte_clear(mm, addr, ptep); ptep 54 include/asm-generic/hugetlb.h pte_t *ptep, pte_t pte) ptep 56 include/asm-generic/hugetlb.h set_pte_at(mm, addr, ptep, pte); ptep 62 include/asm-generic/hugetlb.h unsigned long addr, pte_t *ptep) ptep 64 include/asm-generic/hugetlb.h return ptep_get_and_clear(mm, addr, ptep); ptep 70 include/asm-generic/hugetlb.h unsigned long addr, pte_t *ptep) ptep 72 include/asm-generic/hugetlb.h ptep_clear_flush(vma, addr, ptep); ptep 107 include/asm-generic/hugetlb.h unsigned long addr, pte_t *ptep) ptep 109 include/asm-generic/hugetlb.h ptep_set_wrprotect(mm, addr, ptep); ptep 115 include/asm-generic/hugetlb.h unsigned long addr, pte_t *ptep, ptep 118 include/asm-generic/hugetlb.h return ptep_set_access_flags(vma, addr, ptep, pte, dirty); ptep 123 include/asm-generic/hugetlb.h static inline pte_t huge_ptep_get(pte_t *ptep) ptep 125 include/asm-generic/hugetlb.h return *ptep; ptep 31 include/asm-generic/pgtable.h unsigned long address, pte_t *ptep, ptep 64 include/asm-generic/pgtable.h pte_t *ptep) ptep 66 include/asm-generic/pgtable.h pte_t pte = *ptep; ptep 71 include/asm-generic/pgtable.h set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); ptep 103 include/asm-generic/pgtable.h unsigned long address, pte_t *ptep); ptep 127 include/asm-generic/pgtable.h pte_t *ptep) ptep 129 include/asm-generic/pgtable.h pte_t pte = *ptep; ptep 130 include/asm-generic/pgtable.h pte_clear(mm, address, ptep); ptep 181 include/asm-generic/pgtable.h unsigned long address, pte_t *ptep, ptep 185 include/asm-generic/pgtable.h pte = ptep_get_and_clear(mm, address, ptep); ptep 198 include/asm-generic/pgtable.h pte_t *ptep, ptep 201 include/asm-generic/pgtable.h pte_clear(mm, address, ptep); ptep 208 include/asm-generic/pgtable.h pte_t *ptep); ptep 222 include/asm-generic/pgtable.h static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep) ptep 224 include/asm-generic/pgtable.h pte_t old_pte = *ptep; ptep 225 include/asm-generic/pgtable.h set_pte_at(mm, address, ptep, pte_wrprotect(old_pte)); ptep 411 include/asm-generic/pgtable.h #define set_pte_safe(ptep, pte) \ ptep 413 include/asm-generic/pgtable.h WARN_ON_ONCE(pte_present(*ptep) && !pte_same(*ptep, pte)); \ ptep 414 include/asm-generic/pgtable.h set_pte(ptep, pte); \ ptep 611 include/asm-generic/pgtable.h pte_t *ptep) ptep 618 include/asm-generic/pgtable.h return ptep_get_and_clear(vma->vm_mm, addr, ptep); ptep 623 include/asm-generic/pgtable.h pte_t *ptep, pte_t pte) ptep 629 include/asm-generic/pgtable.h set_pte_at(vma->vm_mm, addr, ptep, pte); ptep 649 include/asm-generic/pgtable.h pte_t *ptep) ptep 651 include/asm-generic/pgtable.h return __ptep_modify_prot_start(vma, addr, ptep); ptep 660 include/asm-generic/pgtable.h pte_t *ptep, pte_t old_pte, pte_t pte) ptep 662 include/asm-generic/pgtable.h __ptep_modify_prot_commit(vma, addr, ptep, pte); ptep 499 include/asm-generic/tlb.h #define __tlb_remove_tlb_entry(tlb, ptep, address) do { } while (0) ptep 509 include/asm-generic/tlb.h #define tlb_remove_tlb_entry(tlb, ptep, address) \ ptep 513 include/asm-generic/tlb.h __tlb_remove_tlb_entry(tlb, ptep, address); \ ptep 516 include/asm-generic/tlb.h #define tlb_remove_huge_tlb_entry(h, tlb, ptep, address) \ ptep 524 include/asm-generic/tlb.h __tlb_remove_tlb_entry(tlb, ptep, address); \ ptep 576 include/asm-generic/tlb.h #define pte_free_tlb(tlb, ptep, address) \ ptep 581 include/asm-generic/tlb.h __pte_free_tlb(tlb, ptep, address); \ ptep 122 include/linux/hugetlb.h int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep); ptep 156 include/linux/hugetlb.h pte_t *ptep) ptep 568 include/linux/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz) ptep 570 include/linux/hugetlb.h set_huge_pte_at(mm, addr, ptep, pte); ptep 577 include/linux/hugetlb.h unsigned long addr, pte_t *ptep) ptep 579 include/linux/hugetlb.h return huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); ptep 586 include/linux/hugetlb.h unsigned long addr, pte_t *ptep, ptep 589 include/linux/hugetlb.h set_huge_pte_at(vma->vm_mm, addr, ptep, pte); ptep 737 include/linux/hugetlb.h pte_t *ptep, pte_t pte, unsigned long sz) ptep 1755 include/linux/mm.h pte_t *ptep; ptep 1756 include/linux/mm.h __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl)); ptep 1757 include/linux/mm.h return ptep; ptep 204 include/linux/swapops.h extern void __migration_entry_wait(struct mm_struct *mm, pte_t *ptep, ptep 229 include/linux/swapops.h static inline void __migration_entry_wait(struct mm_struct *mm, pte_t *ptep, ptep 135 include/trace/events/xen.h TP_PROTO(pte_t *ptep, pte_t pteval), ptep 136 include/trace/events/xen.h TP_ARGS(ptep, pteval), ptep 138 include/trace/events/xen.h __field(pte_t *, ptep) ptep 141 include/trace/events/xen.h TP_fast_assign(__entry->ptep = ptep; ptep 144 include/trace/events/xen.h __entry->ptep, ptep 151 include/trace/events/xen.h TP_PROTO(pte_t *ptep, pte_t pteval), \ ptep 152 include/trace/events/xen.h TP_ARGS(ptep, pteval)) ptep 158 include/trace/events/xen.h pte_t *ptep, pte_t pteval), ptep 159 include/trace/events/xen.h TP_ARGS(mm, addr, ptep, pteval), ptep 163 include/trace/events/xen.h __field(pte_t *, ptep) ptep 168 include/trace/events/xen.h __entry->ptep = ptep; ptep 171 include/trace/events/xen.h __entry->mm, __entry->addr, __entry->ptep, ptep 197 include/trace/events/xen.h TP_PROTO(struct mm_struct *mm, unsigned long addr, pte_t *ptep), ptep 198 include/trace/events/xen.h TP_ARGS(mm, addr, ptep), ptep 202 include/trace/events/xen.h __field(pte_t *, ptep) ptep 206 include/trace/events/xen.h __entry->ptep = ptep), ptep 208 include/trace/events/xen.h __entry->mm, __entry->addr, __entry->ptep) ptep 280 include/trace/events/xen.h pte_t *ptep, pte_t pteval), ptep 281 include/trace/events/xen.h TP_ARGS(mm, addr, ptep, pteval), ptep 285 include/trace/events/xen.h __field(pte_t *, ptep) ptep 290 include/trace/events/xen.h __entry->ptep = ptep; ptep 293 include/trace/events/xen.h __entry->mm, __entry->addr, __entry->ptep, ptep 300 include/trace/events/xen.h pte_t *ptep, pte_t pteval), \ ptep 301 include/trace/events/xen.h TP_ARGS(mm, addr, ptep, pteval)) ptep 189 mm/gup.c pte_t *ptep, pte; ptep 195 mm/gup.c ptep = pte_offset_map_lock(mm, pmd, address, &ptl); ptep 196 mm/gup.c pte = *ptep; ptep 211 mm/gup.c pte_unmap_unlock(ptep, ptl); ptep 218 mm/gup.c pte_unmap_unlock(ptep, ptl); ptep 245 mm/gup.c ret = follow_pfn_pte(vma, address, ptep, flags); ptep 254 mm/gup.c pte_unmap_unlock(ptep, ptl); ptep 308 mm/gup.c pte_unmap_unlock(ptep, ptl); ptep 311 mm/gup.c pte_unmap_unlock(ptep, ptl); ptep 1777 mm/gup.c static inline pte_t gup_get_pte(pte_t *ptep) ptep 1782 mm/gup.c pte.pte_low = ptep->pte_low; ptep 1784 mm/gup.c pte.pte_high = ptep->pte_high; ptep 1786 mm/gup.c } while (unlikely(pte.pte_low != ptep->pte_low)); ptep 1794 mm/gup.c static inline pte_t gup_get_pte(pte_t *ptep) ptep 1796 mm/gup.c return READ_ONCE(*ptep); ptep 1831 mm/gup.c pte_t *ptep, *ptem; ptep 1833 mm/gup.c ptem = ptep = pte_offset_map(&pmd, addr); ptep 1835 mm/gup.c pte_t pte = gup_get_pte(ptep); ptep 1867 mm/gup.c if (unlikely(pte_val(pte) != pte_val(*ptep))) { ptep 1878 mm/gup.c } while (ptep++, addr += PAGE_SIZE, addr != end); ptep 1990 mm/gup.c static int gup_hugepte(pte_t *ptep, unsigned long sz, unsigned long addr, ptep 2003 mm/gup.c pte = READ_ONCE(*ptep); ptep 2029 mm/gup.c if (unlikely(pte_val(pte) != pte_val(*ptep))) { ptep 2045 mm/gup.c pte_t *ptep; ptep 2049 mm/gup.c ptep = hugepte_offset(hugepd, addr, pdshift); ptep 2052 mm/gup.c if (!gup_hugepte(ptep, sz, addr, end, flags, pages, nr)) ptep 2054 mm/gup.c } while (ptep++, addr = next, addr != end); ptep 455 mm/hmm.c unsigned long end, pmd_t *pmdp, pte_t *ptep, ptep 462 mm/hmm.c pte_t pte = *ptep; ptep 509 mm/hmm.c pte_unmap(ptep); ptep 547 mm/hmm.c pte_unmap(ptep); ptep 561 mm/hmm.c pte_t *ptep; ptep 617 mm/hmm.c ptep = pte_offset_map(pmdp, addr); ptep 619 mm/hmm.c for (; addr < end; addr += PAGE_SIZE, ptep++, i++) { ptep 622 mm/hmm.c r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, &pfns[i]); ptep 639 mm/hmm.c pte_unmap(ptep - 1); ptep 3384 mm/hugetlb.c unsigned long address, pte_t *ptep) ptep 3388 mm/hugetlb.c entry = huge_pte_mkwrite(huge_pte_mkdirty(huge_ptep_get(ptep))); ptep 3389 mm/hugetlb.c if (huge_ptep_set_access_flags(vma, address, ptep, entry, 1)) ptep 3390 mm/hugetlb.c update_mmu_cache(vma, address, ptep); ptep 3525 mm/hugetlb.c pte_t *ptep; ptep 3553 mm/hugetlb.c ptep = huge_pte_offset(mm, address, sz); ptep 3554 mm/hugetlb.c if (!ptep) ptep 3557 mm/hugetlb.c ptl = huge_pte_lock(h, mm, ptep); ptep 3558 mm/hugetlb.c if (huge_pmd_unshare(mm, &address, ptep)) { ptep 3567 mm/hugetlb.c pte = huge_ptep_get(ptep); ptep 3578 mm/hugetlb.c huge_pte_clear(mm, address, ptep, sz); ptep 3602 mm/hugetlb.c pte = huge_ptep_get_and_clear(mm, address, ptep); ptep 3603 mm/hugetlb.c tlb_remove_huge_tlb_entry(h, tlb, ptep, address); ptep 3728 mm/hugetlb.c unsigned long address, pte_t *ptep, ptep 3739 mm/hugetlb.c pte = huge_ptep_get(ptep); ptep 3747 mm/hugetlb.c set_huge_ptep_writable(vma, haddr, ptep); ptep 3787 mm/hugetlb.c ptep = huge_pte_offset(mm, haddr, huge_page_size(h)); ptep 3788 mm/hugetlb.c if (likely(ptep && ptep 3789 mm/hugetlb.c pte_same(huge_ptep_get(ptep), pte))) ptep 3824 mm/hugetlb.c ptep = huge_pte_offset(mm, haddr, huge_page_size(h)); ptep 3825 mm/hugetlb.c if (likely(ptep && pte_same(huge_ptep_get(ptep), pte))) { ptep 3829 mm/hugetlb.c huge_ptep_clear_flush(vma, haddr, ptep); ptep 3831 mm/hugetlb.c set_huge_pte_at(mm, haddr, ptep, ptep 3910 mm/hugetlb.c unsigned long address, pte_t *ptep, unsigned int flags) ptep 3988 mm/hugetlb.c ptl = huge_pte_lock(h, mm, ptep); ptep 3989 mm/hugetlb.c if (!huge_pte_none(huge_ptep_get(ptep))) { ptep 4046 mm/hugetlb.c ptl = huge_pte_lock(h, mm, ptep); ptep 4052 mm/hugetlb.c if (!huge_pte_none(huge_ptep_get(ptep))) ptep 4062 mm/hugetlb.c set_huge_pte_at(mm, haddr, ptep, new_pte); ptep 4067 mm/hugetlb.c ret = hugetlb_cow(mm, vma, address, ptep, page, ptl); ptep 4122 mm/hugetlb.c pte_t *ptep, entry; ptep 4134 mm/hugetlb.c ptep = huge_pte_offset(mm, haddr, huge_page_size(h)); ptep 4135 mm/hugetlb.c if (ptep) { ptep 4136 mm/hugetlb.c entry = huge_ptep_get(ptep); ptep 4138 mm/hugetlb.c migration_entry_wait_huge(vma, mm, ptep); ptep 4144 mm/hugetlb.c ptep = huge_pte_alloc(mm, haddr, huge_page_size(h)); ptep 4145 mm/hugetlb.c if (!ptep) ptep 4160 mm/hugetlb.c entry = huge_ptep_get(ptep); ptep 4162 mm/hugetlb.c ret = hugetlb_no_page(mm, vma, mapping, idx, address, ptep, flags); ptep 4199 mm/hugetlb.c ptl = huge_pte_lock(h, mm, ptep); ptep 4202 mm/hugetlb.c if (unlikely(!pte_same(entry, huge_ptep_get(ptep)))) ptep 4221 mm/hugetlb.c ret = hugetlb_cow(mm, vma, address, ptep, ptep 4228 mm/hugetlb.c if (huge_ptep_set_access_flags(vma, haddr, ptep, entry, ptep 4230 mm/hugetlb.c update_mmu_cache(vma, haddr, ptep); ptep 4558 mm/hugetlb.c pte_t *ptep; ptep 4581 mm/hugetlb.c ptep = huge_pte_offset(mm, address, huge_page_size(h)); ptep 4582 mm/hugetlb.c if (!ptep) ptep 4584 mm/hugetlb.c ptl = huge_pte_lock(h, mm, ptep); ptep 4585 mm/hugetlb.c if (huge_pmd_unshare(mm, &address, ptep)) { ptep 4591 mm/hugetlb.c pte = huge_ptep_get(ptep); ptep 4604 mm/hugetlb.c set_huge_swap_pte_at(mm, address, ptep, ptep 4614 mm/hugetlb.c old_pte = huge_ptep_modify_prot_start(vma, address, ptep); ptep 4617 mm/hugetlb.c huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte); ptep 4940 mm/hugetlb.c int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep) ptep 4946 mm/hugetlb.c BUG_ON(page_count(virt_to_page(ptep)) == 0); ptep 4947 mm/hugetlb.c if (page_count(virt_to_page(ptep)) == 1) ptep 4951 mm/hugetlb.c put_page(virt_to_page(ptep)); ptep 4963 mm/hugetlb.c int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep) ptep 1124 mm/ksm.c pte_t *ptep; ptep 1143 mm/ksm.c ptep = pte_offset_map_lock(mm, pmd, addr, &ptl); ptep 1144 mm/ksm.c if (!pte_same(*ptep, orig_pte)) { ptep 1145 mm/ksm.c pte_unmap_unlock(ptep, ptl); ptep 1169 mm/ksm.c flush_cache_page(vma, addr, pte_pfn(*ptep)); ptep 1176 mm/ksm.c ptep_clear_flush(vma, addr, ptep); ptep 1177 mm/ksm.c set_pte_at_notify(mm, addr, ptep, newpte); ptep 1184 mm/ksm.c pte_unmap_unlock(ptep, ptl); ptep 4140 mm/memory.c pte_t *ptep; ptep 4186 mm/memory.c ptep = pte_offset_map_lock(mm, pmd, address, ptlp); ptep 4187 mm/memory.c if (!pte_present(*ptep)) ptep 4189 mm/memory.c *ptepp = ptep; ptep 4192 mm/memory.c pte_unmap_unlock(ptep, *ptlp); ptep 4240 mm/memory.c pte_t *ptep; ptep 4245 mm/memory.c ret = follow_pte(vma->vm_mm, address, &ptep, &ptl); ptep 4248 mm/memory.c *pfn = pte_pfn(*ptep); ptep 4249 mm/memory.c pte_unmap_unlock(ptep, ptl); ptep 4260 mm/memory.c pte_t *ptep, pte; ptep 4266 mm/memory.c if (follow_pte(vma->vm_mm, address, &ptep, &ptl)) ptep 4268 mm/memory.c pte = *ptep; ptep 4278 mm/memory.c pte_unmap_unlock(ptep, ptl); ptep 307 mm/migrate.c void __migration_entry_wait(struct mm_struct *mm, pte_t *ptep, ptep 315 mm/migrate.c pte = *ptep; ptep 332 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 336 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 343 mm/migrate.c pte_t *ptep = pte_offset_map(pmd, address); ptep 344 mm/migrate.c __migration_entry_wait(mm, ptep, ptl); ptep 2198 mm/migrate.c pte_t *ptep; ptep 2243 mm/migrate.c ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); ptep 2246 mm/migrate.c for (; addr < end; addr += PAGE_SIZE, ptep++) { ptep 2252 mm/migrate.c pte = *ptep; ptep 2316 mm/migrate.c ptep_get_and_clear(mm, addr, ptep); ptep 2324 mm/migrate.c set_pte_at(mm, addr, ptep, swp_pte); ptep 2343 mm/migrate.c pte_unmap_unlock(ptep - 1, ptl); ptep 2719 mm/migrate.c pte_t *ptep; ptep 2781 mm/migrate.c ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); ptep 2783 mm/migrate.c if (pte_present(*ptep)) { ptep 2784 mm/migrate.c unsigned long pfn = pte_pfn(*ptep); ptep 2787 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 2792 mm/migrate.c } else if (!pte_none(*ptep)) { ptep 2793 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 2803 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 2816 mm/migrate.c flush_cache_page(vma, addr, pte_pfn(*ptep)); ptep 2817 mm/migrate.c ptep_clear_flush_notify(vma, addr, ptep); ptep 2818 mm/migrate.c set_pte_at_notify(mm, addr, ptep, entry); ptep 2819 mm/migrate.c update_mmu_cache(vma, addr, ptep); ptep 2822 mm/migrate.c set_pte_at(mm, addr, ptep, entry); ptep 2823 mm/migrate.c update_mmu_cache(vma, addr, ptep); ptep 2826 mm/migrate.c pte_unmap_unlock(ptep, ptl); ptep 127 mm/mincore.c pte_t *ptep; ptep 143 mm/mincore.c ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); ptep 144 mm/mincore.c for (; addr != end; ptep++, addr += PAGE_SIZE) { ptep 145 mm/mincore.c pte_t pte = *ptep; ptep 173 mm/mincore.c pte_unmap_unlock(ptep - 1, ptl); ptep 56 mm/pgtable-generic.c unsigned long address, pte_t *ptep, ptep 59 mm/pgtable-generic.c int changed = !pte_same(*ptep, entry); ptep 61 mm/pgtable-generic.c set_pte_at(vma->vm_mm, address, ptep, entry); ptep 70 mm/pgtable-generic.c unsigned long address, pte_t *ptep) ptep 73 mm/pgtable-generic.c young = ptep_test_and_clear_young(vma, address, ptep); ptep 82 mm/pgtable-generic.c pte_t *ptep) ptep 86 mm/pgtable-generic.c pte = ptep_get_and_clear(mm, address, ptep); ptep 278 mm/vmalloc.c pte_t *ptep, pte; ptep 309 mm/vmalloc.c ptep = pte_offset_map(pmd, addr); ptep 310 mm/vmalloc.c pte = *ptep; ptep 313 mm/vmalloc.c pte_unmap(ptep); ptep 92 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t *ptep; ptep 108 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, vm->pgd) + pgd_index(vm, vaddr) * 8; ptep 109 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!*ptep) { ptep 110 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep = vm_phy_page_alloc(vm, KVM_GUEST_PAGE_TABLE_MIN_PADDR, pgd_memslot); ptep 111 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep |= 3; ptep 116 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, vaddr) * 8; ptep 117 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!*ptep) { ptep 118 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep = vm_phy_page_alloc(vm, KVM_GUEST_PAGE_TABLE_MIN_PADDR, pgd_memslot); ptep 119 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep |= 3; ptep 123 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pmd_index(vm, vaddr) * 8; ptep 124 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!*ptep) { ptep 125 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep = vm_phy_page_alloc(vm, KVM_GUEST_PAGE_TABLE_MIN_PADDR, pgd_memslot); ptep 126 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep |= 3; ptep 130 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, vaddr) * 8; ptep 136 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep = paddr | 3; ptep 137 tools/testing/selftests/kvm/lib/aarch64/processor.c *ptep |= (attr_idx << 2) | (1 << 10) /* Access Flag */; ptep 150 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t *ptep; ptep 155 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, vm->pgd) + pgd_index(vm, gva) * 8; ptep 156 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!ptep) ptep 161 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, gva) * 8; ptep 162 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!ptep) ptep 166 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pmd_index(vm, gva) * 8; ptep 167 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!ptep) ptep 171 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, gva) * 8; ptep 172 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!ptep) ptep 179 tools/testing/selftests/kvm/lib/aarch64/processor.c return pte_addr(vm, *ptep) + (gva & (vm->page_size - 1)); ptep 191 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t pte, *ptep; ptep 197 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pte); ptep 198 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!*ptep) ptep 200 tools/testing/selftests/kvm/lib/aarch64/processor.c printf("%*s%s: %lx: %lx at %p\n", indent, "", type[level], pte, *ptep, ptep); ptep 201 tools/testing/selftests/kvm/lib/aarch64/processor.c pte_dump(stream, vm, indent + 1, pte_addr(vm, *ptep), level + 1); ptep 209 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t pgd, *ptep; ptep 215 tools/testing/selftests/kvm/lib/aarch64/processor.c ptep = addr_gpa2hva(vm, pgd); ptep 216 tools/testing/selftests/kvm/lib/aarch64/processor.c if (!*ptep) ptep 218 tools/testing/selftests/kvm/lib/aarch64/processor.c printf("%*spgd: %lx: %lx at %p\n", indent, "", pgd, *ptep, ptep); ptep 219 tools/testing/selftests/kvm/lib/aarch64/processor.c pte_dump(stream, vm, indent + 1, pte_addr(vm, *ptep), level); ptep 189 virt/kvm/arm/mmu.c static inline void kvm_set_pte(pte_t *ptep, pte_t new_pte) ptep 191 virt/kvm/arm/mmu.c WRITE_ONCE(*ptep, new_pte); ptep 201 virt/kvm/arm/mmu.c static inline void kvm_pmd_populate(pmd_t *pmdp, pte_t *ptep) ptep 203 virt/kvm/arm/mmu.c kvm_set_pmd(pmdp, kvm_mk_pmd(ptep)); ptep 1170 virt/kvm/arm/mmu.c pte_t *ptep; ptep 1194 virt/kvm/arm/mmu.c ptep = pte_offset_kernel(pmdp, addr); ptep 1195 virt/kvm/arm/mmu.c if (!ptep || pte_none(*ptep) || !pte_present(*ptep)) ptep 1198 virt/kvm/arm/mmu.c *ptepp = ptep; ptep 1206 virt/kvm/arm/mmu.c pte_t *ptep; ptep 1209 virt/kvm/arm/mmu.c found = stage2_get_leaf_entry(kvm, addr, &pudp, &pmdp, &ptep); ptep 1218 virt/kvm/arm/mmu.c return kvm_s2pte_exec(ptep);